Ollama Ollama is an 🏷️#open-source platform for running 📝Large Language Models (LLMs) locally on personal or server hardware. It bundles model weights, configuration, and runtime into a single Modelfile, providing a unified CLI and REST API for downloading and querying models entirely on-device. Compatible with macOS, Linux, and Windows, Ollama supports a curated library of popular models including @Llama Llama, 📝Llama Llama,📝LLaMA📝Mistral, and 📝Llama Llama, Qwen, as well as custom fine-tuned models in GGUF format. Its OpenAI-compatible API simplifies migration from cloud-based inference, making it widely adopted for privacy-preserving and offline AI workflows.
Contexts
- This is the 📝Root Memo for all things 🏷️#ollama.
