You ’ve probably learn the bombilation about DeepSeek R1 . It ’s an undetermined - root AI model beingcompared to top - level proprietary model like OpenAI ’s o1 . More than that , it ’s a logical thinking model meaning it habituate a chain of thought processes to analyze problems and its own answer logically and then lento get at an answer . This approach helps the AI model sire more accurate reception , clear complex interrogative sentence that require serious abstract thought skills . As it ’s an open - source framework , for the first meter , you’re able to install a abstract thought AI manikin on your PC and run it offline . No need to worry about privacy .
In this templet , I ’ll show you how to set up DeepSeek R1 locally , even if this is your first time and you ’re fresh to running AI models . The gradation are the same for Mac , Windows , or Linux .
Models You Can Install and Pre-requisites
DeepSeek R1is useable in unlike size . While lean the largest 671B parameter model is n’t executable for most motorcar , smaller , distilled version can be installedlocally on your PCs . Note that running AI models topically is resource - intensive demand storage space , RAM , and GPU power . Each fashion model has specific hardware requirements and here ’s a quick overview :
It is well if you have more RAM . In fact , we recommend you add potent RAM if possible to get better results .
Pro Tip : Just start out and blur about which R1 model to install . we recommend you try installing thesmallest 1.5B parameters model(the first one in the table above ) — it ’s lightweight and easy to prove .
How to Install DeepSeek R1 Locally
There are different ways to instal and hunt down the DeepSeek models locally on your computing machine . We will share a few easy I here .
Pro Tip : We recommendOllama and Chatbox methodsif you are just start out and want an easy way to install DeepSeek R1 model or any AI model for that affair .
Method 1: Installing R1 Using Ollama and Chatbox
This is the easiest way to get go , even for beginners .
pace 1 : Install Ollama
1 . Go to theOllama websiteand download the installer for your operating arrangement ( Mac , Windows , or Linux ) . launch the installer and follow the on - screen door instructions .
2 . Once installed , capable Terminalto substantiate it ’s working . Copy - paste the command below .
You should see a version number seem , which entail Ollama is quick to apply .
Step 2 : Download and execute the DeepSeek R1 Model
1.Give the comply command in Terminal . Replace [ framework size ] with the model of the AI you need to instal . For example , for 1.5B parametric quantity model , run this command : ollama run deepseek - r1:1.5b .
2.Wait for the model to download . You ’ll see progress in the depot .
3.Once downloaded , the model will begin run for . you may interact with it directly from the Terminal . displace forward , you may use the same Terminal program line to chat with DeepSeek R1 AI model .
Now , we will show how to install Chatbox for a drug user - friendly port .
footprint 3 : set up Chatbox
1.Download Chatbox from itsofficial website . Install and open up the app . You ’ll see a simple , user - friendly user interface .
2 . In Chatbox , go toSettingsby clicking on the cog icon in the sidebar .
3.Set the Model Provider toOllama .
4.Set the API host to :
5.Select the DeepSeek R1 model ( e.g. ,deepseek - r1:1.5b ) from the dropdown menu .
6.HitSaveand part chatting .
Method 2: Using Ollama and Docker
This method is majuscule if you need to pass the model in a Docker container .
Step 1 : Install Docker
1.Go to theDocker websiteand downloadDocker Desktop for your O . Install Docker by keep up the on - screen education .
2.launch the app and lumber in with the military service .
3.Type the command below in the Terminal to play it .
You should see a version number appear mean that the Docker is installed .
measure 2 : Pull the Open WebUI picture
1.Open your terminus and type :
2.This will download the necessary data file for the user interface .
Step 3 : start the Docker ContainerandOpen WebUI
1.Start the Docker container with pertinacious information storage and map ports by running :
2.Wait a few endorsement for the container to jump .
3.Open your web internet browser and go to :
4.Create an accounting as prompted , and you ’ll be redirected to the main interface . At this peak , there will be no manakin available for selection .
footmark 5 : Set Up Ollama and Integrate DeepSeek R1
1.Visit theOllama websiteand download / install it .
2.In the terminal , download the desired DeepSeek R1 manikin by type :
3.Refresh the Open WebUI Sir Frederick Handley Page in your internet browser . You ’ll see the download DeepSeek R1 mannikin ( e.g. ,deepseek - r1:8b ) in the role model list .
4.Select the model and start up jaw .
Method 3: Using LM Studio
Works not bad if you do n’t require to practice the Terminal to interact with DeepSeek topically . However , LM Studio presently only supports Qwen 7B and 8B models . So if you require to establish 1.5B or even mellow models like 32B , this method will not work for you .
1.Download LM Studio from itsofficial web site . instal and launch the program program .
2.select thesearch iconin the sidebar and search for the DeepSeek R1 mannikin ( e.g. , deepseek - r1 - distill - llama-8b ) .
3.ClickDownloadand wait for the cognitive operation to complete .
4.Once downloaded , cluck on thesearch barat the top of the LM Studio home page .
5.Select the downloaded model and debase it .
6.That ’s it . case yourpromptin the text box and hitEnter . The model will generate a response .
Final Thoughts
run DeepSeek R1 locally offers privacy , price savings , and the flexibility to customize your AI apparatus .
If you ’re new to this , start with Ollama and Chatbox for a simple setup . Docker is ideal for user conversant with containerization , while LM Studio works best for those quash last mastery . try on a smaller model like the 8B or 1.5B to get started , and descale up as you go .