Create Response
Create an LLM response.
Supports both streaming and non-streaming modes via the stream parameter.
Request
This endpoint expects an object.
input
Conversation history
model
Model ID in format ‘provider/model-name’. Optional if a routing policy is configured.
tools
Available tools
max_tokens
Maximum tokens to generate
temperature
Sampling temperature
top_p
Nucleus sampling parameter
stop
Stop sequences
response_format
Response format
stream
Whether to stream the response
project_id
Optional project ID (UUID) to associate with this request
routing_policy_id
Override the default routing policy (for testing)
include_routing_metadata
Include detailed routing metadata in response
Response
Successful Response