Dataset Viewer
Auto-converted to Parquet Duplicate
audio
audioduration (s)
5.22
512
text
stringlengths
54
5.01k
Hi everyone welcome back to another video in our channel Neural Hacks with Vasant.
So in this video we'll be seeing about how to create your own local LLM.
Okay, so we have chat GPT right? So we'll be having a similar interface like chat GPT.
Like you know, you can ask your questions by loading your llm of your own so if you don't know we have already uh fine-tuned lama 2 on our own even that model should be suitable with this framework okay but unfortunately i don't have the gpu uh enough g GPU to load that kind of a heavy model okay so since this is just ...
So any of these generation models can be used. So these are some running scripts.
But the main thing to notice is to use GPUs, you need to install NVIDIA container toolkit.
We also recommend using NVIDIA drivers with cuda version 11.8 or higher okay so first you need to uh set up your system with uh so first you need to uh set up your system with just a minute.
Sorry I was keeping the wrong screen. So like I said here is the text generation inference and here you have you are seeing right like they are using a rust background okay and here are the optimized architectures and they are supporting other models also like i said by using uh these methods okay so you can also load ...
Okay for this you need to have CUDA toolkit installed.
Okay and then you need code nn and then you'll also need to install visual studio community okay these all needs to be installed okay so that is very important so if that is done the next step would be to install this nvidia container toolkit because we'll be running everything in docker without docker also you can do ...
You can just provide your model. And then volume, you can just have it like this itself and followed by this sorry you need to install that sorry run that docker run command okay I'll add pseudo here and you can have any port here it was defaultly given as 8080 in the github repo but you can also have your own port lik...
So you can just open a bash or else even your WSL terminal and you can execute this command.
Okay. So npm install and npm run dev. For you, it will take some time because i've installed my npm packages so it it will come faster for me but for you guys it will take some time you need not worry like it is taking so time so much time or you need not worry about that okay so now we'll execute this but before movin...
README.md exists but content is empty.
Downloads last month
5