Skip to main content

Lambda AI

Overviewโ€‹

PropertyDetails
DescriptionLambda AI provides access to a wide range of open-source language models through their cloud GPU infrastructure, optimized for inference at scale.
Provider Route on LiteLLMlambda_ai/
Link to Provider DocLambda AI API Documentation โ†—
Base URLhttps://api.lambda.ai/v1
Supported Operations/chat/completions


https://docs.lambda.ai/api

We support ALL Lambda AI models, just set lambda_ai/ as a prefix when sending completion requests

Available Modelsโ€‹

Lambda AI offers a diverse selection of state-of-the-art open-source models:

Large Language Modelsโ€‹

ModelDescriptionContext Window
lambda_ai/llama3.3-70b-instruct-fp8Llama 3.3 70B with FP8 quantization8,192 tokens
lambda_ai/llama3.1-405b-instruct-fp8Llama 3.1 405B with FP8 quantization8,192 tokens
lambda_ai/llama3.1-70b-instruct-fp8Llama 3.1 70B with FP8 quantization8,192 tokens
lambda_ai/llama3.1-8b-instructLlama 3.1 8B instruction-tuned8,192 tokens
lambda_ai/llama3.1-nemotron-70b-instruct-fp8Llama 3.1 Nemotron 70B8,192 tokens

DeepSeek Modelsโ€‹

ModelDescriptionContext Window
lambda_ai/deepseek-llama3.3-70bDeepSeek Llama 3.3 70B8,192 tokens
lambda_ai/deepseek-r1-0528DeepSeek R1 05288,192 tokens
lambda_ai/deepseek-r1-671bDeepSeek R1 671B8,192 tokens
lambda_ai/deepseek-v3-0324DeepSeek V3 03248,192 tokens

Hermes Modelsโ€‹

ModelDescriptionContext Window
lambda_ai/hermes3-405bHermes 3 405B8,192 tokens
lambda_ai/hermes3-70bHermes 3 70B8,192 tokens
lambda_ai/hermes3-8bHermes 3 8B8,192 tokens

Coding Modelsโ€‹

ModelDescriptionContext Window
lambda_ai/qwen25-coder-32b-instructQwen 2.5 Coder 32B8,192 tokens
lambda_ai/qwen3-32b-fp8Qwen 3 32B with FP88,192 tokens

Vision Modelsโ€‹

ModelDescriptionContext Window
lambda_ai/llama3.2-11b-vision-instructLlama 3.2 11B with vision capabilities8,192 tokens

Specialized Modelsโ€‹

ModelDescriptionContext Window
lambda_ai/llama-4-maverick-17b-128e-instruct-fp8Llama 4 Maverick with 128k context131,072 tokens
lambda_ai/llama-4-scout-17b-16e-instructLlama 4 Scout with 16k context16,384 tokens
lambda_ai/lfm-40bLFM 40B model8,192 tokens
lambda_ai/lfm-7bLFM 7B model8,192 tokens

Required Variablesโ€‹

Environment Variables
os.environ["LAMBDA_API_KEY"] = ""  # your Lambda AI API key

Usage - LiteLLM Python SDKโ€‹

Non-streamingโ€‹

Lambda AI Non-streaming Completion
import os
import litellm
from litellm import completion

os.environ["LAMBDA_API_KEY"] = "" # your Lambda AI API key

messages = [{"content": "Hello, how are you?", "role": "user"}]

# Lambda AI call
response = completion(
model="lambda_ai/llama3.1-8b-instruct",
messages=messages
)

print(response)

Streamingโ€‹

Lambda AI Streaming Completion
import os
import litellm
from litellm import completion

os.environ["LAMBDA_API_KEY"] = "" # your Lambda AI API key

messages = [{"content": "Write a short story about AI", "role": "user"}]

# Lambda AI call with streaming
response = completion(
model="lambda_ai/llama3.1-70b-instruct-fp8",
messages=messages,
stream=True
)

for chunk in response:
print(chunk)

Vision/Multimodal Supportโ€‹

The Llama 3.2 Vision model supports image inputs:

Lambda AI Vision/Multimodal
import os
import litellm
from litellm import completion

os.environ["LAMBDA_API_KEY"] = "" # your Lambda AI API key

messages = [{
"role": "user",
"content": [
{
"type": "text",
"text": "What's in this image?"
},
{
"type": "image_url",
"image_url": {
"url": "https://example.com/image.jpg"
}
}
]
}]

# Lambda AI vision model call
response = completion(
model="lambda_ai/llama3.2-11b-vision-instruct",
messages=messages
)

print(response)

Function Callingโ€‹

Lambda AI models support function calling:

Lambda AI Function Calling
import os
import litellm
from litellm import completion

os.environ["LAMBDA_API_KEY"] = "" # your Lambda AI API key

# Define tools
tools = [{
"type": "function",
"function": {
"name": "get_weather",
"description": "Get the current weather in a location",
"parameters": {
"type": "object",
"properties": {
"location": {
"type": "string",
"description": "The city and state, e.g. San Francisco, CA"
}
},
"required": ["location"]
}
}
}]

messages = [{"role": "user", "content": "What's the weather in Boston?"}]

# Lambda AI call with function calling
response = completion(
model="lambda_ai/hermes3-70b",
messages=messages,
tools=tools,
tool_choice="auto"
)

print(response)

Usage - LiteLLM Proxy Serverโ€‹

config.yaml
model_list:
- model_name: llama-8b
litellm_params:
model: lambda_ai/llama3.1-8b-instruct
api_key: os.environ/LAMBDA_API_KEY
- model_name: deepseek-70b
litellm_params:
model: lambda_ai/deepseek-llama3.3-70b
api_key: os.environ/LAMBDA_API_KEY
- model_name: hermes-405b
litellm_params:
model: lambda_ai/hermes3-405b
api_key: os.environ/LAMBDA_API_KEY
- model_name: qwen-coder
litellm_params:
model: lambda_ai/qwen25-coder-32b-instruct
api_key: os.environ/LAMBDA_API_KEY

Custom API Baseโ€‹

If you need to use a custom API base URL:

Custom API Base
import os
import litellm
from litellm import completion

# Using environment variable
os.environ["LAMBDA_API_BASE"] = "https://custom.lambda-api.com/v1"
os.environ["LAMBDA_API_KEY"] = "" # your API key

# Or pass directly
response = completion(
model="lambda_ai/llama3.1-8b-instruct",
messages=[{"content": "Hello!", "role": "user"}],
api_base="https://custom.lambda-api.com/v1",
api_key="your-api-key"
)

Supported OpenAI Parametersโ€‹

Lambda AI supports all standard OpenAI parameters since it's fully OpenAI-compatible:

  • temperature
  • max_tokens
  • top_p
  • frequency_penalty
  • presence_penalty
  • stop
  • n
  • stream
  • tools
  • tool_choice
  • response_format
  • seed
  • user
  • logit_bias

Example with parameters:

Lambda AI with Parameters
response = completion(
model="lambda_ai/hermes3-405b",
messages=[{"content": "Explain quantum computing", "role": "user"}],
temperature=0.7,
max_tokens=500,
top_p=0.9,
frequency_penalty=0.2,
presence_penalty=0.1
)