Get a VPS server with a 15% discount
Sale ends in
00Days
:
00Hours
:
00Minutes
:
00Seconds

Deploying DeepSeek on Your Own Server in a Few Clicks

DeepSeek AI is a powerful open AI model that can run without requiring a GPU. Combined with Ollama, it allows you to run AI locally with full control over performance and privacy.

Hosting DeepSeek on your own server provides a high level of security, eliminating the risk of data interception through an API. The official DeepSeek servers are often overloaded. Local deployment lets you use only your own AI resources without sharing them with other users.

Please note that DeepSeek is a third-party product. SpaceCore Solution LTD. is not responsible for the operation of this software.

 

What specifications are required?

Let’s compare models with different requirements. Each model can run both on CPU and GPU. However, since we are using a server, this article will focus on installing and running the model on CPU resources.

Model versionSizeRAMRecommended planCapabilities
deepseek-r1:1.5b1.1 GB4 GBHi-CPU GalaxySimple tasks, ideal for testing and trial runs.
deepseek-r1:7b4.7 GB10 GBHi-CPU OrionWriting and translating texts. Developing simple code.
deepseek-r1:14b9 GB20 GBHi-CPU PulsarAdvanced capabilities for development and copywriting. Excellent balance of speed and functionality.
deepseek-r1:32b19 GB40 GBHi-CPU InfinityPerformance comparable to ChatGPT o1 mini. In-depth data analysis.
deepseek-r1:70b42 GB85 GBMinimum dedicated serverHigh-level computations for business tasks. Deep data analysis and complex development.
deepseek-r1:671b720 GB768 GBHigh-end dedicated server on demandThe main and most advanced DeepSeek R1 model. Provides computational capabilities comparable to the latest ChatGPT models. We recommend hosting it on a powerful dedicated server with NVMe drives.
 
 

Installing DeepSeek 14B

We will install the 14B model as an example, since it provides high performance with moderate resource usage. The instructions are suitable for any available model, and you can install a different version if needed.

 

For installation we use the Hi-CPU Pulsar plan with Ubuntu 22.04, which is an ideal choice for DeepSeek: Hi-CPU Pulsar with Ubuntu 22.04.

Update the repositories and system packages to the latest version.

sudo apt update && sudo apt upgrade -y

 

Install Ollama, a software manager required to deploy DeepSeek.

curl -fsSL https://ollama.com/install.sh | sh && sudo systemctl start ollama

 

After the installation is complete, run the command to download the required DeepSeek model.

ollama pull deepseek-r1:14b
 

deepseek-r1:14b is the name of the selected model. Change it if you want to install a different version, for example deepseek-r1:32b.

Installing this model will take approximately 2 minutes when using a Hi-CPU Pulsar server, thanks to the high network speed. Then run the command to start the DeepSeek model.

ollama run deepseek-r1:14b

 

A command prompt for entering messages will appear. Great — you can now chat with the AI!

 

You can also send a one-off command, for example:

ollama run deepseek-r1:14b "What is heavier: 1 kg of iron or 1 kg of feathers?"
 
 

To generate a public key and enable API access, run the command:

ollama serve

 

To check the installed models and their status, run:

ollama list
 

Conclusion

We recommend hosting DeepSeek R1 models on servers with sufficient RAM. For stable operation, it is best to rent servers with at least a small memory reserve and fast NVMe drives.

The plans listed in the table are a perfect fit for hosting DeepSeek AI. We guarantee the quality of SpaceCore servers. Entrust your server hosting to us and build an infrastructure for reliable and convenient use of DeepSeek in your business.