ollama-setup
Configure auto-configure Ollama when user needs local LLM deployment, free AI alternatives, or wants to eliminate hosted API costs. Trigger phrases: "install ollama", "local AI", "free LLM", "self-hosted AI", "replace OpenAI", "no API costs". Use when appropriate context detected. Trigger with relevant phrases based on skill purpose.
Best use case
ollama-setup is best used when you need a repeatable AI agent workflow instead of a one-off prompt.
Configure auto-configure Ollama when user needs local LLM deployment, free AI alternatives, or wants to eliminate hosted API costs. Trigger phrases: "install ollama", "local AI", "free LLM", "self-hosted AI", "replace OpenAI", "no API costs". Use when appropriate context detected. Trigger with relevant phrases based on skill purpose.
Teams using ollama-setup should expect a more consistent output, faster repeated execution, less prompt rewriting.
When to use this skill
- You want a reusable workflow that can be run more than once with consistent structure.
When not to use this skill
- You only need a quick one-off answer and do not need a reusable workflow.
- You cannot install or maintain the underlying files, dependencies, or repository context.
Installation
Claude Code / Cursor / Codex
Manual Installation
- Download SKILL.md from GitHub
- Place it in
.claude/skills/ollama-setup/SKILL.mdinside your project - Restart your AI agent — it will auto-discover the skill
How ollama-setup Compares
| Feature / Agent | ollama-setup | Standard Approach |
|---|---|---|
| Platform Support | Not specified | Limited / Varies |
| Context Awareness | High | Baseline |
| Installation Complexity | Unknown | N/A |
Frequently Asked Questions
What does this skill do?
Configure auto-configure Ollama when user needs local LLM deployment, free AI alternatives, or wants to eliminate hosted API costs. Trigger phrases: "install ollama", "local AI", "free LLM", "self-hosted AI", "replace OpenAI", "no API costs". Use when appropriate context detected. Trigger with relevant phrases based on skill purpose.
Where can I find the source code?
You can find the source code on GitHub using the link provided at the top of the page.
SKILL.md Source
# Ollama Setup
## Overview
Auto-configure Ollama for local LLM deployment, eliminating hosted API costs and enabling offline AI inference. This skill handles system assessment, model selection based on available hardware (RAM, GPU), installation across macOS/Linux/Docker, and integration with Python, Node.js, and REST API clients.
## Prerequisites
- macOS 12+, Linux (Ubuntu 20.04+, Fedora 36+), or Docker runtime
- Minimum 8 GB RAM for 7B parameter models; 16 GB for 13B models; 32 GB+ for 70B models
- Optional: NVIDIA GPU with CUDA drivers for accelerated inference (`nvidia-smi` to verify)
- Optional: Apple Silicon (M1/M2/M3) for Metal-accelerated inference on macOS
- Disk space: 4-40 GB depending on model size (quantized weights)
- Package manager: `brew` (macOS), `curl` (Linux), or `docker` (containerized)
## Instructions
1. Detect the host operating system and available hardware using `uname -s`, `free -h` (Linux) or `vm_stat` (macOS), and `nvidia-smi` (if GPU present)
2. Select appropriate models based on available RAM:
- **8 GB**: llama3.2:7b (4 GB), mistral:7b (4 GB), phi3:14b (8 GB)
- **16 GB**: codellama:13b (7 GB), mixtral:8x7b (26 GB quantized)
- **32 GB+**: llama3.2:70b (40 GB), codellama:34b (20 GB)
3. Install Ollama using the platform-appropriate method:
- macOS: `brew install ollama && brew services start ollama`
- Linux: `curl -fsSL https://ollama.com/install.sh | sh && sudo systemctl start ollama`
- Docker: `docker run -d -v ollama:/root/.ollama -p 11434:11434 --name ollama ollama/ollama`
4. Pull the recommended model: `ollama pull llama3.2`
5. Verify the installation by listing available models (`ollama list`) and running a test prompt (`ollama run llama3.2 "Say hello"`)
6. Confirm the REST API is accessible: `curl http://localhost:11434/api/tags`
7. Configure integration with the target application using the appropriate client library (Python `ollama`, Node.js `ollama`, or raw HTTP)
8. Set up GPU acceleration if NVIDIA or Apple Silicon hardware is detected
9. Configure model persistence and cache directory if non-default storage location is required
10. Validate end-to-end inference latency and throughput for the selected model
See `${CLAUDE_SKILL_DIR}/references/skill-workflow.md` for the detailed workflow with code snippets.
## Output
- Ollama installation confirmed and running as a system service or Docker container
- Selected model(s) pulled and cached locally with verified inference capability
- REST API endpoint accessible at `http://localhost:11434`
- Integration code snippet for the target language (Python, Node.js, or cURL)
- Hardware assessment report: OS, RAM, GPU availability, recommended models
- Performance baseline: tokens per second for the selected model on local hardware
## Error Handling
| Error | Cause | Solution |
|-------|-------|----------|
| `ollama: command not found` | Installation incomplete or PATH not updated | Re-run install script; restart shell session; verify `/usr/local/bin/ollama` exists |
| Model pull fails with timeout | Network connectivity issue or Ollama registry unreachable | Check internet connection; retry with `ollama pull --insecure` behind corporate proxy |
| Out of memory during inference | Model size exceeds available RAM | Switch to a smaller quantized model (e.g., 7B instead of 13B); close memory-intensive applications |
| GPU not detected | CUDA drivers missing or incompatible version | Install CUDA toolkit >= 11.8; verify with `nvidia-smi`; restart Ollama service after driver install |
| Port 11434 already in use | Another service occupying the default Ollama port | Stop conflicting service; or set `OLLAMA_HOST=0.0.0.0:11435` environment variable |
See `${CLAUDE_SKILL_DIR}/references/errors.md` for additional error scenarios.
## Examples
**Scenario 1: Developer Workstation Setup** -- Install Ollama on a macOS M2 machine with 16 GB RAM. Pull codellama:13b for code generation tasks. Integrate with a Python FastAPI application using the `ollama` Python package. Expected throughput: 30-50 tokens/second on Apple Silicon.
**Scenario 2: Air-Gapped Server Deployment** -- Install Ollama on an offline Ubuntu server via pre-downloaded binary. Transfer model weights via USB. Configure as a systemd service with auto-restart. Serve llama3.2:7b via REST API for internal team use.
**Scenario 3: Docker-Based CI Pipeline** -- Run Ollama in a Docker container as part of a CI/CD pipeline for automated code review. Pull mistral:7b, expose the API on port 11434, and integrate with a Node.js test harness that sends code diffs for analysis.
## Resources
- [Ollama Official Documentation](https://ollama.com) -- installation, model library, API reference
- [Ollama Model Library](https://ollama.com/library) -- available models with size and capability details
- [Ollama Python Client](https://github.com/ollama/ollama-python) -- Python SDK for local inference
- [Ollama JavaScript Client](https://github.com/ollama/ollama-js) -- Node.js SDK for local inference
- Hardware sizing guide: RAM requirements by model parameter count and quantization levelRelated Skills
websocket-handler-setup
Websocket Handler Setup - Auto-activating skill for Backend Development. Triggers on: websocket handler setup, websocket handler setup Part of the Backend Development skill category.
vpc-network-setup
Vpc Network Setup - Auto-activating skill for GCP Skills. Triggers on: vpc network setup, vpc network setup Part of the GCP Skills skill category.
tensorflow-serving-setup
Tensorflow Serving Setup - Auto-activating skill for ML Deployment. Triggers on: tensorflow serving setup, tensorflow serving setup Part of the ML Deployment skill category.
sqs-queue-setup
Sqs Queue Setup - Auto-activating skill for AWS Skills. Triggers on: sqs queue setup, sqs queue setup Part of the AWS Skills skill category.
spy-setup-helper
Spy Setup Helper - Auto-activating skill for Test Automation. Triggers on: spy setup helper, spy setup helper Part of the Test Automation skill category.
spike-test-setup
Spike Test Setup - Auto-activating skill for Performance Testing. Triggers on: spike test setup, spike test setup Part of the Performance Testing skill category.
sla-monitor-setup
Sla Monitor Setup - Auto-activating skill for Enterprise Workflows. Triggers on: sla monitor setup, sla monitor setup Part of the Enterprise Workflows skill category.
server-sent-events-setup
Server Sent Events Setup - Auto-activating skill for API Integration. Triggers on: server sent events setup, server sent events setup Part of the API Integration skill category.
react-context-setup
React Context Setup - Auto-activating skill for Frontend Development. Triggers on: react context setup, react context setup Part of the Frontend Development skill category.
rds-instance-setup
Rds Instance Setup - Auto-activating skill for AWS Skills. Triggers on: rds instance setup, rds instance setup Part of the AWS Skills skill category.
rabbitmq-queue-setup
Rabbitmq Queue Setup - Auto-activating skill for Backend Development. Triggers on: rabbitmq queue setup, rabbitmq queue setup Part of the Backend Development skill category.
pre-commit-hook-setup
Pre Commit Hook Setup - Auto-activating skill for DevOps Basics. Triggers on: pre commit hook setup, pre commit hook setup Part of the DevOps Basics skill category.