@saqibsidd18: #pakistani #foryou #foryoupage #foryourpage #fyp #fyppppppppppppppppppppppp #fy #fyfyfyfyfyfyfyfyfyfyfyfyfyfyfyfyfyfy #fyfyfyfy #browntiktok #brown #desi #blowthisup #relatable #trending #funny #muslim

Saqib
Saqib
Open In TikTok:
Region: GB
Saturday 24 August 2024 03:08:16 GMT
13921
319
1
78

Music

Download

Comments

pablo_escobar1020
 Pablo Escobar  :
😂 ✨Congratulations ✨
2024-08-24 04:56:18
0
To see more videos from user @saqibsidd18, please go to the Tikwm homepage.

Other Videos

Cursor composer with a local LLM Deepseek R1 70B Distillation? I will show you how! Prerequisites: Git Installed. Python 3 installed. Ngrok installed with api key from their website. Requires cursor pro subscription. Install and setup oobabooga/text-generation-webui `git clone https://github.com/oobabooga/text-generation-webui.git` Start the Text Generation WebUI with the api enabled. Linux: `./start_linux.sh --api` Mac: `./start_macos.sh --api` Windows: `.\start_windows.bat --api` WSL: `.\start_wsl.bat --api` Grab the model you want off of hugging face. You need ~46 GB of ram to run the model in this video. I recommend running it on GPU's but unified Macs should be able to run the model as well. If you're not sure what models your computer can run figure out how much ram / or video ram you have available and google the best r1 model/distillation to use with the amount of ram you have available. There are also different models for GPU's and CPU's. Be sure to get a model that can run on GPU hardware if you have VRAM or CPU if you have a unified mac or regular ram. I do not advise running the model on Intel or AMD cpu as they are not very quick or efficient at inference. The model I am using is the DeepSeek R1 70B distillation at a 4 bit quantization.  Find the model you want to try out on hugging face. Once you find the model you want to copy the name of it. This is the bit of it after `huggingface.co/` in the URL. `unsloth/DeepSeek-R1-Distill-Llama-70B-bnb-4bit`in my case. Open your browser and navigate to the text generation web ui: `http://127.0.0.1:7860` Click on Model on the left side navigation. Find the Download model or LoRA section. In the blank text field below paste the name of your model from hugging face. Again for me its `unsloth/DeepSeek-R1-Distill-Llama-70B-bnb-4bit`. Then click download. Depending on your connection speed this might take awhile. Once your model is downloaded. In the upper left corner go to the model dropdown. Select your model. Select the device you have ram on and slide the slider to the amount of ram per device. In my case I have 2 GPU's with 23500 RAM selected. Next hit Load up next to the model dropdown. Next head over to a terminal and open a Ngrok tunnel to port 5000 `ngrok http 5000` This will give you a publically accessible url. Copy the url ngrok provides. Once the model is loaded jump over to cursor.  Goto Settings Select Open AI API Key and enable. Then select Override OpenAI Base URL: Paste in your url from the ngrok command and append `/v1` to the end.  So if your URL is `https://xxx-xxx-x-xxx-xxx.ngrok-free.app` you would put `https://xxx-xxx-x-xxx-xxx.ngrok-free.app/v1` into the base url. Put in a fake open ai api key: `sk-test` Click verify You are now ready to open composer, chat. Select any model from OpenAI and text-generation-webui will use the model it currently has loaded to respond. #nvda #deepseek #llm #cursor #windsurf #development #strawberry #aidevelopment #softwaredevelopment #ide #vscode #localllm
Cursor composer with a local LLM Deepseek R1 70B Distillation? I will show you how! Prerequisites: Git Installed. Python 3 installed. Ngrok installed with api key from their website. Requires cursor pro subscription. Install and setup oobabooga/text-generation-webui `git clone https://github.com/oobabooga/text-generation-webui.git` Start the Text Generation WebUI with the api enabled. Linux: `./start_linux.sh --api` Mac: `./start_macos.sh --api` Windows: `.\start_windows.bat --api` WSL: `.\start_wsl.bat --api` Grab the model you want off of hugging face. You need ~46 GB of ram to run the model in this video. I recommend running it on GPU's but unified Macs should be able to run the model as well. If you're not sure what models your computer can run figure out how much ram / or video ram you have available and google the best r1 model/distillation to use with the amount of ram you have available. There are also different models for GPU's and CPU's. Be sure to get a model that can run on GPU hardware if you have VRAM or CPU if you have a unified mac or regular ram. I do not advise running the model on Intel or AMD cpu as they are not very quick or efficient at inference. The model I am using is the DeepSeek R1 70B distillation at a 4 bit quantization. Find the model you want to try out on hugging face. Once you find the model you want to copy the name of it. This is the bit of it after `huggingface.co/` in the URL. `unsloth/DeepSeek-R1-Distill-Llama-70B-bnb-4bit`in my case. Open your browser and navigate to the text generation web ui: `http://127.0.0.1:7860` Click on Model on the left side navigation. Find the Download model or LoRA section. In the blank text field below paste the name of your model from hugging face. Again for me its `unsloth/DeepSeek-R1-Distill-Llama-70B-bnb-4bit`. Then click download. Depending on your connection speed this might take awhile. Once your model is downloaded. In the upper left corner go to the model dropdown. Select your model. Select the device you have ram on and slide the slider to the amount of ram per device. In my case I have 2 GPU's with 23500 RAM selected. Next hit Load up next to the model dropdown. Next head over to a terminal and open a Ngrok tunnel to port 5000 `ngrok http 5000` This will give you a publically accessible url. Copy the url ngrok provides. Once the model is loaded jump over to cursor. Goto Settings Select Open AI API Key and enable. Then select Override OpenAI Base URL: Paste in your url from the ngrok command and append `/v1` to the end. So if your URL is `https://xxx-xxx-x-xxx-xxx.ngrok-free.app` you would put `https://xxx-xxx-x-xxx-xxx.ngrok-free.app/v1` into the base url. Put in a fake open ai api key: `sk-test` Click verify You are now ready to open composer, chat. Select any model from OpenAI and text-generation-webui will use the model it currently has loaded to respond. #nvda #deepseek #llm #cursor #windsurf #development #strawberry #aidevelopment #softwaredevelopment #ide #vscode #localllm

About