The DeepSeek just released the DeepSeek R1, and it’s Open Source, so let’s running it locally with ollama
First, I got one Mac Pro
Setup Environment
Install the monitor tools
- asitop show the Mem/GPU/CPU usage and power
- htop show the Mem/CPU usage
brew install pipx
pipx install asitop
brew install htop
Install ollama
brew install ollama
Run DeepSeek R1
ollama serve&
ollama run deepseek-r1:70b
More DeepSeek R1 models
The CPU Power around 80W
ollama API
ollama will start one local server
- http://localhost:11434
So we can integrate it with some 3rd party Apps that support ollama, here is the App list that support ollama, and the API is compatible with OpenAI API, so we can use any App to do the integration that support Open AI API.
- https://github.com/ollama/ollama?tab=readme-ov-file#web–desktop
LobeChat with ollama
Refer the document
- https://lobehub.com/docs/usage/providers/ollama
On Mac, need set the environment, the restart the ollama
launchctl setenv OLLAMA_ORIGINS "*"
If need to build support bind to 0.0.0.0
OLLAMA_HOST=0.0.0.0 ollama serve &
so one line command for quick start
launchctl setenv OLLAMA_ORIGINS "*" && OLLAMA_HOST=0.0.0.0 ollama serve
Run DeepSeek R1 70b with ollama