Generation/Completion/Chat Completion Models
OpenAI Chat Completion Models​
| Model Name | Function Call | Required OS Variables |
|---|---|---|
| gpt-3.5-turbo | completion('gpt-3.5-turbo', messages) | os.environ['OPENAI_API_KEY'] |
| gpt-3.5-turbo-16k | completion('gpt-3.5-turbo-16k', messages) | os.environ['OPENAI_API_KEY'] |
| gpt-3.5-turbo-16k-0613 | completion('gpt-3.5-turbo-16k-0613', messages) | os.environ['OPENAI_API_KEY'] |
| gpt-4 | completion('gpt-4', messages) | os.environ['OPENAI_API_KEY'] |
| gpt-5-pro | completion('gpt-5-pro', messages) | os.environ['OPENAI_API_KEY'] |
Azure OpenAI Chat Completion Models​
For Azure calls add the azure/ prefix to model. If your azure deployment name is gpt-v-2 set model = azure/gpt-v-2
| Model Name | Function Call | Required OS Variables |
|---|---|---|
| gpt-3.5-turbo | completion('azure/gpt-3.5-turbo-deployment', messages) | os.environ['AZURE_API_KEY'],os.environ['AZURE_API_BASE'],os.environ['AZURE_API_VERSION'] |
| gpt-4 | completion('azure/gpt-4-deployment', messages) | os.environ['AZURE_API_KEY'],os.environ['AZURE_API_BASE'],os.environ['AZURE_API_VERSION'] |
OpenAI Text Completion Models​
| Model Name | Function Call | Required OS Variables |
|---|---|---|
| text-davinci-003 | completion('text-davinci-003', messages) | os.environ['OPENAI_API_KEY'] |
Cohere Models​
| Model Name | Function Call | Required OS Variables |
|---|---|---|
| command-nightly | completion('command-nightly', messages) | os.environ['COHERE_API_KEY'] |
Anthropic Models​
| Model Name | Function Call | Required OS Variables |
|---|---|---|
| claude-instant-1 | completion('claude-instant-1', messages) | os.environ['ANTHROPIC_API_KEY'] |
| claude-2 | completion('claude-2', messages) | os.environ['ANTHROPIC_API_KEY'] |
Hugging Face Inference API​
All text2text-generation and text-generation models are supported by liteLLM. You can use any text model from Hugging Face with the following steps:
- Copy the
model repoURL from Hugging Face and set it as themodelparameter in the completion call. - Set
hugging_faceparameter toTrue. - Make sure to set the hugging face API key
Here are some examples of supported models: Note that the models mentioned in the table are examples, and you can use any text model available on Hugging Face by following the steps above.
| Model Name | Function Call | Required OS Variables |
|---|---|---|
| stabilityai/stablecode-completion-alpha-3b-4k | completion(model="stabilityai/stablecode-completion-alpha-3b-4k", messages=messages, hugging_face=True) | os.environ['HF_TOKEN'] |
| bigcode/starcoder | completion(model="bigcode/starcoder", messages=messages, hugging_face=True) | os.environ['HF_TOKEN'] |
| google/flan-t5-xxl | completion(model="google/flan-t5-xxl", messages=messages, hugging_face=True) | os.environ['HF_TOKEN'] |
| google/flan-t5-large | completion(model="google/flan-t5-large", messages=messages, hugging_face=True) | os.environ['HF_TOKEN'] |
OpenRouter Completion Models​
All the text models from OpenRouter are supported by liteLLM.
| Model Name | Function Call | Required OS Variables |
|---|---|---|
| openai/gpt-3.5-turbo | completion('openai/gpt-3.5-turbo', messages) | os.environ['OR_SITE_URL'],os.environ['OR_APP_NAME'],os.environ['OR_API_KEY'] |
| openai/gpt-3.5-turbo-16k | completion('openai/gpt-3.5-turbo-16k', messages) | os.environ['OR_SITE_URL'],os.environ['OR_APP_NAME'],os.environ['OR_API_KEY'] |
| openai/gpt-4 | completion('openai/gpt-4', messages) | os.environ['OR_SITE_URL'],os.environ['OR_APP_NAME'],os.environ['OR_API_KEY'] |
| openai/gpt-4-32k | completion('openai/gpt-4-32k', messages) | os.environ['OR_SITE_URL'],os.environ['OR_APP_NAME'],os.environ['OR_API_KEY'] |
| anthropic/claude-2 | completion('anthropic/claude-2', messages) | os.environ['OR_SITE_URL'],os.environ['OR_APP_NAME'],os.environ['OR_API_KEY'] |
| anthropic/claude-instant-v1 | completion('anthropic/claude-instant-v1', messages) | os.environ['OR_SITE_URL'],os.environ['OR_APP_NAME'],os.environ['OR_API_KEY'] |
| google/palm-2-chat-bison | completion('google/palm-2-chat-bison', messages) | os.environ['OR_SITE_URL'],os.environ['OR_APP_NAME'],os.environ['OR_API_KEY'] |
| google/palm-2-codechat-bison | completion('google/palm-2-codechat-bison', messages) | os.environ['OR_SITE_URL'],os.environ['OR_APP_NAME'],os.environ['OR_API_KEY'] |
| meta-llama/llama-2-13b-chat | completion('meta-llama/llama-2-13b-chat', messages) | os.environ['OR_SITE_URL'],os.environ['OR_APP_NAME'],os.environ['OR_API_KEY'] |
| meta-llama/llama-2-70b-chat | completion('meta-llama/llama-2-70b-chat', messages) | os.environ['OR_SITE_URL'],os.environ['OR_APP_NAME'],os.environ['OR_API_KEY'] |
Novita AI Completion Models​
🚨 LiteLLM supports ALL Novita AI models, send model=novita/<your-novita-model> to send it to Novita AI. See all Novita AI models here
| Model Name | Function Call | Required OS Variables |
|---|---|---|
| novita/deepseek/deepseek-r1 | completion('novita/deepseek/deepseek-r1', messages) | os.environ['NOVITA_API_KEY'] |
| novita/deepseek/deepseek_v3 | completion('novita/deepseek/deepseek_v3', messages) | os.environ['NOVITA_API_KEY'] |
| novita/meta-llama/llama-3.3-70b-instruct | completion('novita/meta-llama/llama-3.3-70b-instruct', messages) | os.environ['NOVITA_API_KEY'] |
| novita/meta-llama/llama-3.1-8b-instruct | completion('novita/meta-llama/llama-3.1-8b-instruct', messages) | os.environ['NOVITA_API_KEY'] |
| novita/meta-llama/llama-3.1-8b-instruct-max | completion('novita/meta-llama/llama-3.1-8b-instruct-max', messages) | os.environ['NOVITA_API_KEY'] |
| novita/meta-llama/llama-3.1-70b-instruct | completion('novita/meta-llama/llama-3.1-70b-instruct', messages) | os.environ['NOVITA_API_KEY'] |
| novita/meta-llama/llama-3-8b-instruct | completion('novita/meta-llama/llama-3-8b-instruct', messages) | os.environ['NOVITA_API_KEY'] |
| novita/meta-llama/llama-3-70b-instruct | completion('novita/meta-llama/llama-3-70b-instruct', messages) | os.environ['NOVITA_API_KEY'] |
| novita/meta-llama/llama-3.2-1b-instruct | completion('novita/meta-llama/llama-3.2-1b-instruct', messages) | os.environ['NOVITA_API_KEY'] |
| novita/meta-llama/llama-3.2-11b-vision-instruct | completion('novita/meta-llama/llama-3.2-11b-vision-instruct', messages) | os.environ['NOVITA_API_KEY'] |
| novita/meta-llama/llama-3.2-3b-instruct | completion('novita/meta-llama/llama-3.2-3b-instruct', messages) | os.environ['NOVITA_API_KEY'] |
| novita/gryphe/mythomax-l2-13b | completion('novita/gryphe/mythomax-l2-13b', messages) | os.environ['NOVITA_API_KEY'] |
| novita/google/gemma-2-9b-it | completion('novita/google/gemma-2-9b-it', messages) | os.environ['NOVITA_API_KEY'] |
| novita/mistralai/mistral-nemo | completion('novita/mistralai/mistral-nemo', messages) | os.environ['NOVITA_API_KEY'] |
| novita/mistralai/mistral-7b-instruct | completion('novita/mistralai/mistral-7b-instruct', messages) | os.environ['NOVITA_API_KEY'] |
| novita/qwen/qwen-2.5-72b-instruct | completion('novita/qwen/qwen-2.5-72b-instruct', messages) | os.environ['NOVITA_API_KEY'] |
| novita/qwen/qwen-2-vl-72b-instruct | completion('novita/qwen/qwen-2-vl-72b-instruct', messages) | os.environ['NOVITA_API_KEY'] |