POST
/
agent
/
{agent-app-id}
/
responses
Agent App
curl --request POST \
  --url https://{controlPlaneURL}/api/llm/agent/{agent-app-id}/responses \
  --header 'Authorization: Bearer <token>' \
  --header 'Content-Type: application/json' \
  --data '{
  "messages": [
    {
      "role": "system",
      "content": "<string>",
      "name": "<string>"
    }
  ],
  "stream": true,
  "iteration_limit": 5
}'
{
  "id": "<string>",
  "object": "chat.completion.chunk",
  "created": 123,
  "model": "<string>",
  "service_tier": "<string>",
  "system_fingerprint": "<string>",
  "choices": [
    {
      "index": 123,
      "delta": {
        "content": "<string>",
        "role": "<string>",
        "refusal": "<any>"
      },
      "logprobs": "<any>",
      "finish_reason": "<string>"
    }
  ]
}

Authorizations

Authorization
string
header
required

Bearer authentication header of the form Bearer <token>, where <token> is your auth token.

Path Parameters

agent-app-id
string
required

The ID of the agent app to use for configuration

Body

application/json

Chat parameters. The model, MCP servers, system prompt, guardrails, and model parameters (temperature, max_tokens, etc.) are autofilled from agent app configuration.

iteration_limit
integer
default:5
required

Maximum number of iterations to perform with tool calls. Must be a positive integer. Defaults to 5.

Required range: 0 < x <= 20
messages
(System Message · object | User Message · object | Assistant Message · object | Function Message · object | Tool Message · object | Developer Message · object)[]

A list of messages comprising the conversation history

stream
boolean | null

Whether to stream responses

Response

Chat completions response with tool calls using agent app configuration

object
enum<string>
required
Available options:
chat.completion.chunk
created
number
required
model
string
required
choices
object[]
required
id
string
service_tier
string
system_fingerprint
string