Ollama
API for running large language models locally.
APIs
Ollama API
Ollama provides a REST API for running and managing large language models locally. The API supports text generation, chat completions, embeddings, model management, and streamin...
Ollama OpenAI Compatibility API
Ollama provides compatibility with parts of the OpenAI API, allowing existing applications built for OpenAI to connect to locally-running models through Ollama. Supported endpoi...
Ollama Anthropic Compatibility API
Ollama provides compatibility with the Anthropic Messages API, enabling tools like Claude Code to work with locally-running open-source models. Supports messages, streaming, sys...
Ollama Cloud API
Ollama Cloud provides cloud-hosted inference for large language models, giving access to larger models and faster responses without requiring a powerful local GPU. Cloud models ...