ollama
3
总安装量
2
周安装量
#60266
全站排名
安装命令
npx skills add https://github.com/g1joshi/agent-skills --skill ollama
Agent 安装分布
amp
2
gemini-cli
2
claude-code
2
github-copilot
2
kimi-cli
2
Skill 文档
Ollama
Ollama makes running LLMs locally as easy as docker run. 2025 updates include Windows/AMD support, Multimodal input, and Tool Calling.
When to Use
- Local Development: Coding without wifi or API costs.
- Privacy: Processing sensitive documents on-device.
- Integration: Works with LangChain, LlamaIndex, and Obsidian natively.
Core Concepts
Modelfile
Docker-like file to define a custom model (System prompt + Base model).
FROM llama3
SYSTEM You are Mario from Super Mario Bros.
API
Ollama runs a local server (localhost:11434) compatible with OpenAI SDK.
Best Practices (2025)
Do:
- Use high-speed RAM: Local LLM speed depends on memory bandwidth.
- Use Quantized Models:
q4_k_mis the sweet spot for speed/quality balance. - Unload:
ollama stopwhen done to free VRAM for games/rendering.
Don’t:
- Don’t expect GPT-4 level: Smaller local models (8B) are smart but lack deep reasoning.