Spin Serverless AI API

The Spin Serverless AI API enables Spin components to run AI inference using built-in language model support (Llama 2, CodeLlama, etc.) via the Spin SDK's infer() function. Components must declare the ai_models they need in spin.toml. Supported on Fermyon Cloud and SpinKube deployments.

API entry from apis.yml

apis.yml Raw ↑
aid: spin:spin-serverless-ai-api
name: Spin Serverless AI API
description: The Spin Serverless AI API enables Spin components to run AI inference using built-in language
  model support (Llama 2, CodeLlama, etc.) via the Spin SDK's infer() function. Components must declare
  the ai_models they need in spin.toml. Supported on Fermyon Cloud and SpinKube deployments.
humanURL: https://spinframework.dev/v3/serverless-ai-tutorial
tags:
- AI
- LLM
- Machine Learning
- WebAssembly
properties:
- type: Documentation
  url: https://spinframework.dev/v3/serverless-ai-tutorial
- type: GitHubRepository
  url: https://github.com/fermyon/spin