Spaces:
Running
Let's Talk about AI
Hello, here is an open space for everyone to talk, share, ask and show anything about AI.
Has anyone pre-trained LLM model from scratch ? If yes then share your experience, things to consider while training, notes, tips etc.
Hi i am also intrested into LLM Model , i am about to start this reserach from next week please give any inputs
Hi i am also intrested into LLM Model , i am about to start this reserach from next week please give any inputs
Hey @Shashank2k3 , if you want your own LLM model, first you need huge data. You can start with fine tuning already available good LLM models like Gemma, Phi, LLAMA, mistral etc with your dataset. Start with small models of sizes like 4 to 7B parameters. For pre-training LLM from scratch you need enormous data, good resources like heavy duty GPUs and CPUs and also have knowledge of training techniques, NLP, etc . You can always brainstorm with ChatGPT to get more knowledge.
Hey @kalashshah19 , thanks for the input! I already have a solid foundation in these areas from my Bachelor's degree in AIML, and now Iโm looking to dive deeper into the world of LLMs.
Hey @kalashshah19 , thanks for the input! I already have a solid foundation in these areas from my Bachelor's degree in AIML, and now Iโm looking to dive deeper into the world of LLMs.
Great !
Yupp so what you guys do, i mean profession!!!
Yupp so what you guys do, i mean profession!!!
I am an Associate Data Scientist at Casepoint.
What about you ?
under Mahabharatha collection, could you please change the 5th column name from translation to english_text ? and 6th column is actually translation which you have named it correctly.
A great work indeed.. thanks a lot. much needed
I remember I have named one column transliteration in each dataset. Also, thank you for your kind words
Very soon my HF account will complete the firsr year. So, will it retain the 100 GB private storage, zero-GPU access, and inference capabilities after completing my first year as a user?
After investing a lot of days and nights and passionately researching about our Sanatana Dharma, I'm excited to tell that I have over 150k Sanskrit Verses of different books with Translation, Transliteration and Shlokas in my collection.
My collection has around 10 books including Mahabharat, Ramayana, Gita, Markadeye Purana, Devi Mahatmya, Yoga Vasistha and so on.
Please share it with everyone and please upvote me for my work. More to come soon.
Link to Collection : https://huggingface.co/collections/JDhruv14/sanatana-dharma
Awesome job bro. You can fine tune models based on this and also publish them on HF referencing the same dataset.
Also publish everything on Kaggle.
Very soon my HF account will complete the firsr year. So, will it retain the 100 GB private storage, zero-GPU access, and inference capabilities after completing my first year as a user?
Don't know, haven't completed 1 year yet.
hey everyone,
read my blog on NPU and openVINO toolkit
Happy New year
hey everyone,
read my blog on NPU and openVINO toolkit
Sure !
Happy New year
Happy New Year !