gpt-4
, gpt-4o-mini
), but may experience compatibility issues with Truefoundry’s fully qualified model names (like openai-main/gpt-4
or azure-openai/gpt-4
).
When Langflow encounters these fully qualified model names directly, it may not function as expected due to internal processing differences.
The Solution: Load balancing configuration allows you to:
gpt-4
)openai-main/gpt-4
)gpt-4o-mini
)https://{controlPlaneUrl}/api/llm/api/inference/openai
{controlPlaneUrl}
with your actual Truefoundry control plane URL.
gpt-4
)https://{controlPlaneUrl}/api/llm/api/inference/openai
gpt-4
, your requests get load-balanced according to your configuration. In the example configuration above, any request to gpt-4
will be routed to openai-main/gpt-4
with 100% of the traffic.
You can create more sophisticated routing rules with multiple targets and different weights: