Llama API¶
Llama API is a Meta-hosted API service that helps you integrate Llama models into your applications quickly and efficiently.
Llama API provides access to Llama models through a simple API interface, with inference provided by Meta, so you can focus on building AI-powered solutions without managing your own inference infrastructure.
With Llama API, you get access to state-of-the-art AI capabilities through a developer-friendly interface designed for simplicity and performance.
Installation¶
Llama API is configured as an optional dependency in Strands Agents. To install, run:
pip install strands-agents[llamaapi]
Usage¶
After installing llamaapi
, you can import and initialize Strands Agents' Llama API provider as follows:
from strands import Agent
from strands.models.llamaapi import LlamaAPIModel
from strands_tools import calculator
model = LlamaAPIModel(
client_args={
"api_key": "<KEY>",
},
# **model_config
model_id="Llama-4-Maverick-17B-128E-Instruct-FP8",
)
agent = Agent(model=model, tools=[calculator])
response = agent("What is 2+2")
print(response)
Configuration¶
Client Configuration¶
The client_args
configure the underlying LlamaAPI client. For a complete list of available arguments, please refer to the LlamaAPI docs.
Model Configuration¶
The model_config
configures the underlying model selected for inference. The supported configurations are:
Parameter | Description | Example | Options |
---|---|---|---|
model_id |
ID of a model to use | Llama-4-Maverick-17B-128E-Instruct-FP8 |
reference |
repetition_penalty |
Controls the likelyhood and generating repetitive responses. (minimum: 1, maximum: 2, default: 1) | 1 |
reference |
temperature |
Controls randomness of the response by setting a temperature. | 0.7 |
reference |
top_p |
Controls diversity of the response by setting a probability threshold when choosing the next token. | 0.9 |
reference |
max_completion_tokens |
The maximum number of tokens to generate. | 4096 |
reference |
top_k |
Only sample from the top K options for each subsequent token. | 10 |
reference |
Troubleshooting¶
Module Not Found¶
If you encounter the error ModuleNotFoundError: No module named 'llamaapi'
, this means you haven't installed the llamaapi
dependency in your environment. To fix, run pip install strands-agents[llamaapi]
.