Ollama API
Ollama provides a REST API for running and managing large language models locally. The API supports text generation, chat completions, embeddings, model management, and streaming responses. It serves as the primary interface for interacting with models running on the Ollama inference engine at localhost:11434.
Documentation
Documentation
https://docs.ollama.com/
Authentication
https://docs.ollama.com/api/authentication
Documentation
https://docs.ollama.com/api/generate
Documentation
https://docs.ollama.com/api/chat
Documentation
https://docs.ollama.com/api/embed
Documentation
https://docs.ollama.com/api/tags
Documentation
https://docs.ollama.com/api/ps
Documentation
https://docs.ollama.com/api/create
Documentation
https://docs.ollama.com/api/pull
Documentation
https://docs.ollama.com/api/push
Documentation
https://docs.ollama.com/api/copy
Documentation
https://docs.ollama.com/api/delete
Documentation
https://docs.ollama.com/api/show
Documentation
https://docs.ollama.com/api/streaming
Documentation
https://docs.ollama.com/api/errors
Documentation
https://docs.ollama.com/api/usage
Documentation
https://docs.ollama.com/api/blobs
Documentation
https://docs.ollama.com/api/version