x-tfy-provider-name
header to the name of the provider integration you’re using.
For example, if you’re using an OpenAI provider integration named my-openai-provider
, you’ll set the x-tfy-provider-name
header to my-openai-provider
.
from openai import OpenAI
client = OpenAI(
base_url="http://{controlPlaneUrl}/api/llm",
api_key="your_truefoundry_api_key",
default_headers={"x-tfy-provider-name": "my-openai-provider"}
)
response = client.responses.create(
model="my-openai-provider/gpt-3-5-turbo",
input=[{"role": "user", "content": "Your prompt here"}]
)
print(response)
response = client.responses.create(
model="my-openai-provider/gpt-4o",
input=[
{"role": "user", "content": "Your prompt here"},
{
"role": "user",
"content": [
{
"type": "input_image",
"image_url": "your_image_url"
}
]
}
]
)
response_id = response.id
{
"id": "resp_6847fa2670f8819887e2d14c08bdc5a305d8dc9b22b6f0dd",
"created_at": 1749547558,
"error": null,
"incomplete_details": null,
"instructions": null,
"metadata": {},
"model": "gpt-4o-mini-2024-07-18",
"object": "response",
"output": [
{
"id": "msg_6847fa35e8c48198a5120ee3d38c353105d8dc9b22b6f0dd",
"content": [
{
"annotations": [],
"text": "It seems you might be looking to start a discussion or ask a question! How can I assist you today?",
"type": "output_text"
}
],
"role": "assistant",
"status": "completed",
"type": "message"
}
],
"parallel_tool_calls": true,
"temperature": 1,
"tool_choice": "auto",
"tools": [],
"top_p": 1,
"max_output_tokens": null,
"previous_response_id": null,
"reasoning": {
"effort": null,
"generate_summary": null,
"summary": null
},
"status": "completed",
"text": {
"format": {
"type": "text"
}
},
"truncation": "disabled",
"usage": {
"input_tokens": 10,
"input_tokens_details": {
"cached_tokens": 0
},
"output_tokens": 23,
"output_tokens_details": {
"reasoning_tokens": 0
},
"total_tokens": 33
},
"user": null,
"background": false,
"service_tier": "default",
"store": true,
"provider": "openai"
}
response = client.responses.retrieve(
response_id=response_id
)
print(response)
{
"id": "resp_6847fa2670f8819887e2d14c08bdc5a305d8dc9b22b6f0dd",
"created_at": 1749547558,
"error": null,
"incomplete_details": null,
"instructions": null,
"metadata": {},
"model": "gpt-4o-mini-2024-07-18",
"object": "response",
"output": [
{
"id": "msg_6847fa35e8c48198a5120ee3d38c353105d8dc9b22b6f0dd",
"content": [
{
"annotations": [],
"text": "It seems you might be looking to start a discussion or ask a question! How can I assist you today?",
"type": "output_text"
}
],
"role": "assistant",
"status": "completed",
"type": "message"
}
],
"parallel_tool_calls": true,
"temperature": 1,
"tool_choice": "auto",
"tools": [],
"top_p": 1,
"max_output_tokens": null,
"previous_response_id": null,
"reasoning": {
"effort": null,
"generate_summary": null,
"summary": null
},
"status": "completed",
"text": {
"format": {
"type": "text"
}
},
"truncation": "disabled",
"usage": {
"input_tokens": 10,
"input_tokens_details": {
"cached_tokens": 0
},
"output_tokens": 23,
"output_tokens_details": {
"reasoning_tokens": 0
},
"total_tokens": 33
},
"user": null,
"background": false,
"service_tier": "default",
"store": true,
"provider": "openai"
}
delete_result = client.responses.delete(
response_id=response_id
)
Was this page helpful?