Getting Started
- Install Ollama from ollama.com with the one-line installer for macOS, Linux, or Windows.
- Run your first model with a single command:
ollama run llama3.2to download and start chatting. - Browse available models at ollama.com/library and pull any model with
ollama pull model-name. - Use the OpenAI-compatible API at
localhost:11434to integrate local models into your applications.
Key Features
- One-command setup downloads and runs any supported model with a single
ollama runcommand. - Extensive model library hosts hundreds of models including Llama, Mistral, Gemma, Phi, Qwen, and DeepSeek.
- OpenAI-compatible API serves models locally with an API that works as a drop-in replacement for OpenAI.
- Automatic quantization optimizes models for your hardware, running efficiently on consumer GPUs and Apple Silicon.
- Modelfile customization lets you create custom model configurations with system prompts, parameters, and templates.
- Cross-platform support runs natively on macOS (Apple Silicon), Linux (NVIDIA/AMD), and Windows.
// related tools
Jan
AI / Local & Self-Hosted
Open-source ChatGPT alternative that runs 100% offline
oss
web git
LM Studio
AI / Local & Self-Hosted
Desktop app for running local LLMs with a clean GUI
free
web
Aider
AI / AI Coding Tools
Terminal-based AI pair programmer that edits code in your git repo
oss
web git