strands.models.llamaapi
Llama API model provider.
LlamaAPIModel
Section titled “LlamaAPIModel”class LlamaAPIModel(Model)Defined in: src/strands/models/llamaapi.py:31
Llama API model provider implementation.
LlamaConfig
Section titled “LlamaConfig”class LlamaConfig(TypedDict)Defined in: src/strands/models/llamaapi.py:34
Configuration options for Llama API models.
Attributes:
model_id- Model ID (e.g., “Llama-4-Maverick-17B-128E-Instruct-FP8”).repetition_penalty- Repetition penalty.temperature- Temperature.top_p- Top-p.max_completion_tokens- Maximum completion tokens.top_k- Top-k.
__init__
Section titled “__init__”def __init__(*, client_args: dict[str, Any] | None = None, **model_config: Unpack[LlamaConfig]) -> NoneDefined in: src/strands/models/llamaapi.py:53
Initialize provider instance.
Arguments:
client_args- Arguments for the Llama API client.**model_config- Configuration options for the Llama API model.
update_config
Section titled “update_config”@overridedef update_config(**model_config: Unpack[LlamaConfig]) -> NoneDefined in: src/strands/models/llamaapi.py:75
Update the Llama API Model configuration with the provided arguments.
Arguments:
**model_config- Configuration overrides.
get_config
Section titled “get_config”@overridedef get_config() -> LlamaConfigDefined in: src/strands/models/llamaapi.py:85
Get the Llama API model configuration.
Returns:
The Llama API model configuration.
format_request
Section titled “format_request”def format_request(messages: Messages, tool_specs: list[ToolSpec] | None = None, system_prompt: str | None = None) -> dict[str, Any]Defined in: src/strands/models/llamaapi.py:215
Format a Llama API chat streaming request.
Arguments:
messages- List of message objects to be processed by the model.tool_specs- List of tool specifications to make available to the model.system_prompt- System prompt to provide context to the model.
Returns:
An Llama API chat streaming request.
Raises:
TypeError- If a message contains a content block type that cannot be converted to a LlamaAPI-compatible format.
format_chunk
Section titled “format_chunk”def format_chunk(event: dict[str, Any]) -> StreamEventDefined in: src/strands/models/llamaapi.py:261
Format the Llama API model response events into standardized message chunks.
Arguments:
event- A response event from the model.
Returns:
The formatted chunk.
stream
Section titled “stream”@overrideasync def stream(messages: Messages, tool_specs: list[ToolSpec] | None = None, system_prompt: str | None = None, *, tool_choice: ToolChoice | None = None, **kwargs: Any) -> AsyncGenerator[StreamEvent, None]Defined in: src/strands/models/llamaapi.py:335
Stream conversation with the LlamaAPI model.
Arguments:
messages- List of message objects to be processed by the model.tool_specs- List of tool specifications to make available to the model.system_prompt- System prompt to provide context to the model.tool_choice- Selection strategy for tool invocation. Note: This parameter is accepted for interface consistency but is currently ignored for this model provider.**kwargs- Additional keyword arguments for future extensibility.
Yields:
Formatted message chunks from the model.
Raises:
ModelThrottledException- When the model service is throttling requests from the client.
structured_output
Section titled “structured_output”@overridedef structured_output( output_model: type[T], prompt: Messages, system_prompt: str | None = None, **kwargs: Any) -> AsyncGenerator[dict[str, T | Any], None]Defined in: src/strands/models/llamaapi.py:425
Get structured output from the model.
Arguments:
output_model- The output model to use for the agent.prompt- The prompt messages to use for the agent.system_prompt- System prompt to provide context to the model.**kwargs- Additional keyword arguments for future extensibility.
Yields:
Model events with the last being the structured output.
Raises:
NotImplementedError- Structured output is not currently supported for LlamaAPI models.