ποΈ /a2a - A2A Agent Gateway
3 items
ποΈ /assistants
OpenAI has deprecated the Assistants API. It will shut down on August 26, 2026.
ποΈ /audio
2 items
ποΈ /batches
2 items
ποΈ /containers
Manage OpenAI code interpreter containers (sessions) for executing code in isolated environments.
ποΈ /containers/files
Manage files within Code Interpreter containers. Files are created automatically when code interpreter generates outputs (charts, CSVs, images, etc.).
ποΈ /chat/completions
4 items
ποΈ /completions
Overview
ποΈ /converse
Call Bedrock's /converse endpoint through LiteLLM Proxy.
ποΈ /embeddings
Quick Start
ποΈ /files
2 items
ποΈ /fine_tuning
2 items
ποΈ /generateContent
Use LiteLLM to call Google AI's generateContent endpoints for text generation, multimodal interactions, and streaming responses.
ποΈ /guardrails/apply_guardrail
Use this endpoint to directly call a guardrail configured on your LiteLLM instance. This is useful when you have services that need to directly call a guardrail.
ποΈ /invoke
Call Bedrock's /invoke endpoint through LiteLLM Proxy.
ποΈ /interactions
| Feature | Supported | Notes |
ποΈ /images
3 items
ποΈ /videos
| Feature | Supported |
ποΈ /vector_stores/\{vector_store_id\}/files
Vector store files represent the individual files that live inside a vector store.
ποΈ /mcp - Model Context Protocol
5 items
ποΈ /v1/messages
Use LiteLLM to call all your LLM APIs in the Anthropic v1/messages format.
ποΈ /v1/messages/count_tokens
Overview
ποΈ /moderations
Usage
ποΈ /ocr
| Feature | Supported |
ποΈ Pass-through Endpoints (Anthropic SDK, etc.)
14 items
ποΈ /rag
2 items
ποΈ /realtime
Use this to loadbalance across Azure + OpenAI.
ποΈ /rerank
LiteLLM Follows the cohere api request / response for the rerank api
ποΈ /responses
LiteLLM provides an endpoint in the spec of OpenAI's /responses API
ποΈ /search
10 items
ποΈ /skills - Anthropic Skills API
| Feature | Supported |
ποΈ /vector_stores
2 items