๐๏ธ /chat/completions
3 items
๐๏ธ /responses [Beta]
LiteLLM provides a BETA endpoint in the spec of OpenAI's /responses API
๐๏ธ /completions
Usage
๐๏ธ /embeddings
Quick Start
๐๏ธ /v1/messages
Use LiteLLM to call all your LLM APIs in the Anthropic v1/messages format.
๐๏ธ /mcp [BETA] - Model Context Protocol
LiteLLM Proxy provides an MCP Gateway that allows you to use a fixed endpoint for all MCP tools and control MCP access by Key, Team.
๐๏ธ /images
3 items
๐๏ธ /audio
2 items
๐๏ธ Pass-through Endpoints (Anthropic SDK, etc.)
12 items
๐๏ธ /rerank
LiteLLM Follows the cohere api request / response for the rerank api
๐๏ธ /assistants
Covers Threads, Messages, Assistants.
๐๏ธ /files
2 items
๐๏ธ /batches
2 items
๐๏ธ /realtime
Use this to loadbalance across Azure + OpenAI.
๐๏ธ /fine_tuning
2 items
๐๏ธ /moderations
Usage
๐๏ธ /guardrails/apply_guardrail
Use this endpoint to directly call a guardrail configured on your LiteLLM instance. This is useful when you have services that need to directly call a guardrail.