Home > Resources > Tech

Local Deployment of DeepSeek: A Step-by-Step Guide to Setting Up

update: Feb 18, 2025

Deploying DeepSeek locally can be a rewarding experience, allowing you to harness the power of this advanced AI model without relying on external servers. This guide will walk you through the process of setting up DeepSeek on your local machine, ensuring you can make the most of its capabilities.

 PopAi launches DeepSeek R1!

Understanding the DeepSeek Model

Before diving into the setup, it’s important to understand the differences between the official DeepSeek R1 model and the locally deployable versions. The official R1 model is the “complete version,” utilizing MLA and MoE architectures with 671 billion parameters, activating 37 billion during inference. It has been trained using the GRPO reinforcement learning algorithm.

DeepSeek-R1-1.5B – the best version of DeepSeek I recommend

What I recommend is the DeepSeek-R1-1.5B version. Why? Because it’s a lightweight model with only 1.5 billion parameters. Sounds pretty “mini,” right? But don’t underestimate it—it’s a “small but powerful” model. It only requires 3GB of GPU memory to run, which means even if your computer isn’t high-end, you can easily handle it. Plus, it performs exceptionally well in mathematical reasoning, even surpassing GPT-4 and Claude 3.5 in some benchmark tests.

steps for locally deploying DeepSeek.

Step 1: Choose Your DeepSeek Model

To begin, decide which version of the DeepSeek model you want to deploy locally. I recommend is the DeepSeek-R1-1.5B version. Of course, if your computer has higher specs, you can try other versions.

Step 2: Download and Install Ollama – a large language model tool that supports DeepSeek

Ollama is a tool for running and managing large language models (LLMs) locally. It simplifies the process of setting up and running these models, which can be complex and resource-intensive.

Ollama allows users to run DeepSeek models locally, avoiding the latency and costs associated with using cloud services.

Visit the official website: https://ollama.com/ to download.

The installation file, OllamaSetup.exe, is approximately 745MB.

Double-click the OllamaSetup.exe to begin the installation.

Download and Install Ollama - a large language model tool that supports DeepSeek

Step 3: Verify if Ollama is installed successfully.

In the command line, type the command ollama -v. If the version number appears as shown, it means the installation was successful.

Verify if Ollama is installed successfully

Step 4: Pull the DeepSeek model through Ollama.

To pull the DeepSeek model through Ollama, you can use the following command in the command line:

ollama run deepseek-r1:1.5b

This will initiate the download of the DeepSeek model via Ollama. Make sure that you have a stable internet connection during the process.

After 5 to 6 minutes, seeing the word “success” indicates that DeepSeek R1 has been successfully installed, and you can then start conversing with DeepSeek.

DeepSeek 1.5B – the entire model size is 1.1 GB.

  • 1.5B: Suitable for lightweight tasks, such as simple interactions on edge devices (like smartwatches, IoT devices), small-scale intelligent Q&A systems, etc. Currently, the smallest open-source version.
  • 671B: Primarily used for large-scale cloud inference, suitable for research analysis, data mining, and other complex tasks that require processing massive amounts of data. Currently, the most powerful open-source version.
  • More versions can be found here: https://ollama.com/library/deepseek-r1.
DeepSeek 1.5B - the entire model size is 1.1 GB.

Step 5: Install WebUI – Interact with DeepSeek through a graphical interface

The installed DeepSeek can interact using the command prompt. However, using the command prompt to converse with DeepSeek is not very user-friendly. For a better experience, we can install a WebUI, and here we use the browser plugin: Page Assist. After starting the Ollama service, press the shortcut key [ctrl + shift + L] to open the WebUI page.

Hardware Requirements for Different Versions of the DeepSeek Model

Model VersionParametersMemory Requirement (FP16)Recommended GPU (Single Card)Multi-Card SupportQuantization SupportSuitable Scenarios
DeepSeek-R1-1.5B1.5 Billion3GBGTX 1650 (4GB VRAM)Not RequiredSupportedLow-resource device deployment (Raspberry Pi, old laptops), real-time text generation, embedded systems
DeepSeek-R1-7B7 Billion14GBRTX 3070/4060 (8GB VRAM)OptionalSupportedMedium-complexity tasks (text summarization, translation), lightweight multi-turn dialogue systems
DeepSeek-R1-8B8 Billion16GBRTX 4070 (12GB VRAM)OptionalSupportedTasks requiring higher precision (code generation, logical reasoning)
DeepSeek-R1-14B14 Billion32GBRTX 4090/A5000 (16GB VRAM)RecommendedSupportedEnterprise-level complex tasks (contract analysis, report generation), long text understanding and generation
DeepSeek-R1-32B32 Billion64GBA100 40GB (24GB VRAM)RecommendedSupportedHigh-precision specialized tasks (medical/legal consulting), multi-modal task preprocessing
DeepSeek-R1-70B70 Billion140GB2x A100 80GB/4x RTX 4090 (multi-card parallel)RequiredSupportedResearch institutions/large enterprises (financial forecasting, large-scale data analysis), high-complexity generation tasks
DeepSeek-671B671 Billion512GB+ (Extremely high single-card memory demand, usually requires multi-node distributed training)8x A100/H100 (server cluster)RequiredSupportedNational-scale/mega-scale AI research (climate modeling, genome analysis), general AI (AGI) exploration

Practical Application Scenarios: What Can DeepSeek-R1-1.5B Do?

Despite its small size, DeepSeek-R1-1.5B is not “weak” at all. It is very suitable for lightweight tasks such as:

  1. Intelligent Customer Service: In small businesses or personal projects, it can quickly answer common customer questions, improving service efficiency.
  2. Language Learning: You can use it to practice language expression, such as inputting a Chinese sentence and having it generate an English translation.
  3. Creative Writing: If you are a writer or copywriter, it can help you quickly generate creative snippets or draft copy.

Conclusion

Do you think deploying the DeepSeek-R1-1.5B model locally is super simple? By following the steps above, you can have a “smart assistant” on your computer.

Moreover, this model not only runs quickly but can also play a significant role in many scenarios.

Deploying DeepSeek locally offers a unique opportunity to explore its advanced features without the limitations of server constraints. By following this guide, you can set up and run DeepSeek on your own hardware, unlocking its potential for a wide range of applications.

Start Using PopAi Today

Suggested Content

More >

SELECT SQL_CALC_FOUND_ROWS DISTINCT wp_posts.*, SUM( COALESCE( pvc.count, 0 ) ) AS post_views FROM wp_posts LEFT JOIN wp_term_relationships ON (wp_posts.ID = wp_term_relationships.object_id) LEFT JOIN wp_post_views pvc ON pvc.id = wp_posts.ID AND pvc.type = 4 WHERE 1=1 AND ( wp_posts.post_date > '2024-11-23 04:54:34' ) AND ( wp_term_relationships.term_taxonomy_id IN (107) ) AND wp_posts.post_type = 'post' AND ((wp_posts.post_status = 'publish')) GROUP BY wp_posts.ID, wp_term_relationships.term_taxonomy_id HAVING post_views > 0 ORDER BY post_views DESC LIMIT 0, 6