Skip to main content

🔑 LiteLLM Keys (Access Claude-2, Llama2-70b, etc.)

Use this if you're trying to add support for new LLMs and need access for testing. We provide a free $10 community-key for testing all providers on LiteLLM:

usage (community-key)

import os
from litellm import completion

## set ENV variables
os.environ["OPENAI_API_KEY"] = "your-api-key"
os.environ["COHERE_API_KEY"] = "your-api-key"

messages = [{ "content": "Hello, how are you?","role": "user"}]

# openai call
response = completion(model="gpt-3.5-turbo", messages=messages)

# cohere call
response = completion("command-nightly", messages)

Need a dedicated key? Email us @ krrish@berri.ai

Supported Models for LiteLLM Key

These are the models that currently work with the "sk-litellm-.." keys.

For a complete list of models/providers that you can call with LiteLLM, check out our provider list

  • OpenAI models - OpenAI docs
    • gpt-4
    • gpt-3.5-turbo
    • gpt-3.5-turbo-16k
  • Llama2 models - TogetherAI docs
    • togethercomputer/llama-2-70b-chat
    • togethercomputer/llama-2-70b
    • togethercomputer/LLaMA-2-7B-32K
    • togethercomputer/Llama-2-7B-32K-Instruct
    • togethercomputer/llama-2-7b
    • togethercomputer/CodeLlama-34b
    • WizardLM/WizardCoder-Python-34B-V1.0
    • NousResearch/Nous-Hermes-Llama2-13b
  • Falcon models - TogetherAI docs
    • togethercomputer/falcon-40b-instruct
    • togethercomputer/falcon-7b-instruct
  • Jurassic/AI21 models - AI21 docs
    • j2-ultra
    • j2-mid
    • j2-light
  • NLP Cloud models - NLPCloud docs
    • dolpin
    • chatdolphin
  • Anthropic models - Anthropic docs
    • claude-2
    • claude-instant-v1

For OpenInterpreter

This was initially built for the Open Interpreter community. If you're trying to use this feature in there, here's how you can do it:
Note: You will need to clone and modify the Github repo, until this PR is merged.

git clone https://github.com/krrishdholakia/open-interpreter-litellm-fork

To run it do:

poetry build 

# call gpt-4 - always add 'litellm_proxy/' in front of the model name
poetry run interpreter --model litellm_proxy/gpt-4

# call llama-70b - always add 'litellm_proxy/' in front of the model name
poetry run interpreter --model litellm_proxy/togethercomputer/llama-2-70b-chat

# call claude-2 - always add 'litellm_proxy/' in front of the model name
poetry run interpreter --model litellm_proxy/claude-2

And that's it!

Now you can call any model you like!

Want us to add more models? Let us know!