๐๏ธ /chat/completions
3 items
๐๏ธ /responses [Beta]
LiteLLM provides a BETA endpoint in the spec of OpenAI's /responses API
๐๏ธ /completions
Usage
๐๏ธ /embeddings
Quick Start
๐๏ธ /v1/messages
Use LiteLLM to call all your LLM APIs in the Anthropic v1/messages format.
๐๏ธ /mcp [BETA] - Model Context Protocol
LiteLLM Proxy provides an MCP Gateway that allows you to use a fixed endpoint for all MCP tools and control MCP access by Key, Team.
๐๏ธ Google AI generateContent
Use LiteLLM to call Google AI's generateContent endpoints for text generation, multimodal interactions, and streaming responses.
๐๏ธ /images
3 items
๐๏ธ /audio
2 items
๐๏ธ Pass-through Endpoints (Anthropic SDK, etc.)
12 items
๐๏ธ /rerank
LiteLLM Follows the cohere api request / response for the rerank api
๐๏ธ /assistants
Covers Threads, Messages, Assistants.
๐๏ธ /files
2 items
๐๏ธ /batches
2 items
๐๏ธ /realtime
Use this to loadbalance across Azure + OpenAI.
๐๏ธ /fine_tuning
2 items
๐๏ธ /moderations
Usage
๐๏ธ /guardrails/apply_guardrail
Use this endpoint to directly call a guardrail configured on your LiteLLM instance. This is useful when you have services that need to directly call a guardrail.