The chat completions API supports all OpenAI compatible parameters to fine-tune model behavior and response characteristics.
Basic Parameters
Here’s how to use common parameters:
from openai import OpenAI
client = OpenAI(
api_key="your_truefoundry_api_key",
base_url="<truefoundry-base-url>/api/llm/api/inference/openai"
)
response = client.chat.completions.create(
model="openai-main/gpt-4o-mini",
messages=[{"role": "user", "content": "Hello, how are you?"}],
temperature=0.7, # Controls randomness (0.0 to 1.0)
max_tokens=100, # Maximum number of tokens to generate
top_p=0.9, # Nucleus sampling parameter
frequency_penalty=0.0, # Reduces repetition
presence_penalty=0.0, # Encourages new topics
stop=["\n", "Human:"] # Stop sequences
)
print(response.choices[0].message.content)
You can pass any of the parameters supported by OpenAI’s API. You can refer to the OpenAI API Reference for more details.
Some models from different vendors (or maybe even same vendors) do not support all the parameters supported by OpenAI’s API. For example, temperature is not supported by o
series models like o3-mini.