Build a Local DeepSeek‑R1 Large Model Service with Ollama – Intro to AI LLMs

This guide walks through installing Ollama on Windows, configuring the OLLAMA_MODELS path, downloading the 7‑b DeepSeek‑R1 model, running it locally, and accessing it via a browser using the Page Assist extension, providing step‑by‑step commands, screenshots, and tips for offline setups.

Fun with Large Models
Fun with Large Models
Fun with Large Models
Build a Local DeepSeek‑R1 Large Model Service with Ollama – Intro to AI LLMs

Overview

Ollama is a local large‑model manager that installs, runs and manages LLMs on Windows, macOS and Linux. It provides a local HTTP API (port 11434) for application integration.

1. Install Ollama

Download the installer from https://ollama.com/ (Windows 11 version shown). Run OllamaSetup.exe and follow the wizard. After installation, open a command prompt and execute ollama; the UI appears, confirming a successful start.

2. Configure model storage

By default models are stored in C:\Users\<username>\.ollama\models. To place models on another drive, set the environment variable OLLAMA_MODELS to the desired folder and restart Ollama.

3. Download a model

Search for deepseek-r1 on the Ollama website. The model has 7 B parameters (~4.7 GB). Run: ollama run deepseek-r1 The command pulls the model files into the blobs directory and creates a manifest directory with metadata.

4. Run and test the model

After the download completes, the model runs automatically. In the prompt, type a simple query such as “1+1?”. The response includes a think tag that shows the model’s reasoning steps.

5. Common Ollama commands

ollama -h        # show help
ollama list      # list installed models
ollama restart   # restart the service
ollama update    # update Ollama
ollama serve     # run as a background service
ollama pull <model>   # download without running
ollama run <model>    # run (downloads if missing)
ollama rm <model>     # delete a local model

6. Access the model via a web UI

Ollama listens on http://127.0.0.1:11434. Install the Chrome extension “Page Assist” (or an equivalent client) and set the Ollama URL to the same address. The extension detects the running model, lets you select it, and provides a chat‑style interface for interaction.

7. Offline deployment

On a machine with internet access, install Ollama and pull the required model. Copy the entire Ollama installation folder and the directory referenced by OLLAMA_MODELS to the offline machine. Install Ollama there, set OLLAMA_MODELS to the copied folder, and run the model without network connectivity.

DeepSeek-R1windowsAI Model DeploymentOllamalocal LLMPage Assist
Fun with Large Models
Written by

Fun with Large Models

Master's graduate from Beijing Institute of Technology, published four top‑journal papers, previously worked as a developer at ByteDance and Alibaba. Currently researching large models at a major state‑owned enterprise. Committed to sharing concise, practical AI large‑model development experience, believing that AI large models will become as essential as PCs in the future. Let's start experimenting now!

0 followers
Reader feedback

How this landed with the community

Sign in to like

Rate this article

Was this worth your time?

Sign in to rate
Discussion

0 Comments

Thoughtful readers leave field notes, pushback, and hard-won operational detail here.