Skip to main content

Guide

If you want to use a language model provider that is not supported directly by Onyx, you can configure a custom inference provider.
Your custom provider must provide OpenAI-compatible API endpoints.
1

Setup your Custom Inference Provider

Determine your provider’s API base URL. It should look something like https://yourprovider.com/v1 or http://localhost:12345/v1.
2

Navigate to Language Models

Access the Admin Panel from your user profile icon, then navigate to Configuration → Language Models.
3

Configure Custom Inference Provider

Select Add Custom LLM Provider from the available providers.Give your provider a Display Name.Enter your model’s Provider Name.
The Provider Name must match Litellm’s list of supported providers.
In this example, the provider name is vertex_ai.Custom inference provider name
4

Configure Optional Fields and Models

Enter the provider’s Base URL.Fill out the other optional fields if applicable.In the Model Configurations section, enter each model you want to make available through this provider.
5

Configure Default and Fast Models

The Default Model is selected automatically for new custom Agents and Chat sessions.Designating a Fast Model is optional. This Fast Model is used behind the scenes for quick operations such as evaluating the type of message, generating different queries (query expansion), and naming the chat session.
If you select a Fast Model, make sure it is a relatively quick and cost-effective model.
6

Designate Provider Access

Lastly, you may select whether or not the provider is public to all users in Onyx.If set to private, the provider’s models will be available to Admins and User Groups you explicitly assign the provider to.