Ollama Cloud API

Ollama Cloud provides cloud-hosted inference for large language models, giving access to larger models and faster responses without requiring a powerful local GPU. Cloud models are accessed through the same API interface as local models, with requests encrypted in transit and no storage of prompts or outputs.

API entry from apis.yml

apis.yml Raw ↑
aid: ollama:ollama-cloud-api
name: Ollama Cloud API
tags:
- Cloud
- Inference
- Large Language Models
humanURL: https://docs.ollama.com/cloud
baseURL: https://ollama.com/api
properties:
- url: https://docs.ollama.com/cloud
  type: Documentation
- url: https://ollama.com/cloud
  type: Getting Started
- url: https://ollama.com/pricing
  type: Pricing
- url: https://ollama.com/settings/keys
  type: Authentication
- url: https://ollama.com/search?c=cloud
  type: Models
description: Ollama Cloud provides cloud-hosted inference for large language models, giving access to
  larger models and faster responses without requiring a powerful local GPU. Cloud models are accessed
  through the same API interface as local models, with requests encrypted in transit and no storage of
  prompts or outputs.