Running AI Locally
One of ClawDesk's most powerful features is the ability to run AI models directly on your computer. This means:
- No internet needed — Works on an airplane, in a cabin, anywhere
- Completely free — No API keys, no subscriptions, no per-message costs
- Completely private — Your conversations never leave your machine
- No rate limits — Chat as much as you want
What Does "Running AI Locally" Mean?
When you use ChatGPT or Claude in a browser, your messages travel over the internet to powerful computers owned by those companies. They process your message and send the response back.
Running AI locally means downloading the AI brain (called a "model") onto your own computer. Your computer does all the thinking, and nothing ever goes to the internet.
Do I Have the Right Computer?
Local AI needs some computing power. Here's a simple guide:
Minimum Requirements
| Component | Minimum | Recommended | Best |
|---|---|---|---|
| RAM | 8 GB | 16 GB | 32+ GB |
| Storage | 10 GB free | 50 GB free | 100+ GB free |
| CPU | Any modern CPU | Apple M1+ / Recent Intel/AMD | Apple M2+ / Latest CPUs |
| GPU | Not required | NVIDIA GPU (6+ GB VRAM) | NVIDIA RTX (12+ GB VRAM) |
What Can Your Computer Run?
| Your Setup | Models You Can Run | Quality |
|---|---|---|
| 8 GB RAM, no GPU | Small models (1-3B parameters) | Basic conversations, simple tasks |
| 16 GB RAM, no GPU | Medium models (7-8B parameters) | Good conversations, coding help |
| 16 GB RAM + GPU | Medium-large models (7-13B) | Very good quality |
| 32+ GB RAM or GPU with 12+ GB | Large models (34-70B) | Excellent, near cloud quality |
| Apple M1/M2/M3 Mac | Thanks to unified memory, even 8GB Macs run 7B models well | Good to excellent |
Apple Silicon Macs (M1/M2/M3/M4) are especially good at running local AI because they share memory between the CPU and GPU. A Mac with 16 GB of RAM can run models that would need a dedicated GPU on other systems.
Method 1: ClawDesk Built-in Local Models
The easiest way — no extra software needed.
Step 1: Open the Local Models Page
In ClawDesk, click "Local Models" in the sidebar.
Step 2: Check Your System
ClawDesk automatically detects your hardware:
- CPU type and speed
- Amount of RAM
- GPU (if any)
- Available storage space
Step 3: Browse Recommended Models
Based on your hardware, ClawDesk recommends models that will run well on your machine:
| Model | Size | What It's Good At | Min RAM |
|---|---|---|---|
| Llama 3.1 8B | ~4.5 GB | General chat, writing | 8 GB |
| Mistral 7B | ~4 GB | Fast conversations | 8 GB |
| Code Llama 7B | ~3.8 GB | Programming help | 8 GB |
| Phi-3 Mini | ~2.3 GB | Quick, lightweight tasks | 4 GB |
| Llama 3.1 70B | ~40 GB | Near cloud quality | 64 GB |
Step 4: Download a Model
- Find a model you want to try
- Click the Download button
- Wait for the download (this can take a few minutes depending on model size)
- Models are stored in
~/.clawdesk/models/
Step 5: Start the Model
- After downloading, click Start next to the model
- ClawDesk launches a local inference server
- A green status indicator shows the model is running
Step 6: Use It in Chat
- Go to the Chat page
- In the model dropdown, select "Local (Built-in)"
- Choose your running model
- Start chatting — everything happens on your computer!
Method 2: Using Ollama
Ollama is a popular tool for running local AI. ClawDesk integrates with it seamlessly.
Step 1: Install Ollama
macOS:
brew install ollama
Windows / Linux: Download from ollama.com/download
Step 2: Download a Model
Open your terminal (or command prompt) and run:
ollama pull llama3.1
This downloads the Llama 3.1 model. You can replace llama3.1 with any model from ollama.com/library.
Step 3: Start Ollama
Ollama usually starts automatically. If not:
ollama serve
Step 4: Connect ClawDesk
- In ClawDesk, go to Settings → Providers
- Find Ollama and click Configure
- Base URL:
http://localhost:11434(this is the default) - Click Save
- Your Ollama models now appear in the chat model dropdown
Recommended Ollama Models
# General purpose — great all-rounder
ollama pull llama3.1
# For coding
ollama pull codellama
# Very fast, lightweight
ollama pull mistral
# Google's model
ollama pull gemma2
# For creative writing
ollama pull llama3.1:70b # Need 64GB RAM
Tips for Best Performance
1. Close Other Apps
Local AI uses a lot of memory. Close browser tabs and other heavy apps while using local models.
2. Start with Small Models
If you're not sure about your hardware, start with a small model like Phi-3 or Mistral 7B. If it runs smoothly, try a larger one.
3. Watch Your Temperature
On laptops, local AI can make your computer warm. Make sure your laptop has good ventilation. If it gets too hot, switch to a smaller model.
4. GPU Acceleration
If you have an NVIDIA GPU, make sure you have the latest drivers installed. ClawDesk and Ollama will automatically use your GPU for faster responses.
5. Model Quality vs. Speed
Comparison: Local vs. Cloud
| Aspect | Local AI | Cloud AI (Claude, GPT) |
|---|---|---|
| Cost | Free | Pay per use |
| Privacy | 100% private | Data goes to servers |
| Internet | Not needed | Required |
| Speed | Depends on your hardware | Usually very fast |
| Quality | Good (can be excellent with big models) | Excellent |
| Setup | Download model (one time) | Get API key (instant) |
| Context length | Limited by RAM | Very large (128K-1M+ tokens) |
When Should You Use Each?
Use Local AI when:
- You're chatting casually or brainstorming
- You're offline or have slow internet
- You care deeply about privacy
- You don't want any costs
- You're processing sensitive documents
Use Cloud AI when:
- You need the best possible quality
- You're working on complex coding tasks
- You're processing very long documents
- You need the latest AI capabilities
- Speed is critical and you have a slower computer
Troubleshooting
| Problem | Solution |
|---|---|
| Model is very slow | Try a smaller model, close other apps, check if GPU is being used |
| "Not enough memory" | Choose a smaller model or add more RAM |
| Ollama won't start | Run ollama serve in terminal, check if port 11434 is free |
| Model gives poor answers | Try a larger model, or use cloud AI for complex tasks |
| Download stuck | Check internet connection, try again, check disk space |
| Computer fan is loud | Normal with local AI — ensure good ventilation |
Next Steps
- Desktop App Tour → — Explore all of ClawDesk's features
- Understanding Agents → — Create specialized AI assistants using local models
- Privacy & Security → — Learn how ClawDesk protects your data