POST
/
model-select
curl --request POST \
  --url https://app.irona.ai/api/v1/model-select \
  --header 'Authorization: Bearer <token>' \
  --header 'Content-Type: application/json' \
  --data '{
  "messages": [
    {
      "role": "system",
      "content": "<string>"
    }
  ],
  "llm_providers": [
    {
      "name": "<string>",
      "price": 123,
      "latency": 123,
      "attributes": {}
    }
  ],
  "tools": [
    {
      "name": "<string>",
      "description": "<string>",
      "parameters": {}
    }
  ],
  "max_model_depth": 123,
  "tradeoff": "<string>",
  "preference_id": "<string>",
  "hash_content": false,
  "previous_session": "<string>"
}'
{
  "llm_label": "<string>"
}

Example usage:

curl -X POST https://api.irona.ai/api/v1/model-select \
        -H "Authorization: Bearer <IAI_API_Key>" \
        -H "Content-Type: application/json" \
        -d '{
  "messages": [
        {"role": "system", "content": "You are a world class software developer."},
        {"role": "assistant", "content": "How can I help you today?"},
        {"role": "user", "content": "Write a merge sort in Python."}
  ],
  "llm_providers": [
    {
      "provider": "openai",
      "model": "gpt-4-1106-preview"
    },
    {
      "provider": "anthropic",
      "model": "claude-3-opus-20240229"
    }
  ],
}'

Authorizations

Authorization
string
header
required

Bearer authentication header of the form Bearer <token>, where <token> is your auth token.

Body

application/json
messages
object[]
required

The messages array in OpenAI format.

llm_providers
object[]
required

A list of LLM providers you want to route between. You can optionally define their custom attributes, such as price and latency.

tools
object[]

A list of tools the model may call. Currently, only functions are supported as a tool. Use this to provide a list of functions the model may generate JSON inputs for. A max of 128 functions are supported.

max_model_depth
integer

The maximum depth considered in the ranked LLM providers list.

tradeoff
string

Additional routing tradeoffs. Valid values are "cost" and "latency".

preference_id
string

The preference ID created on the Not Diamond dashboard. Setting this value will change the ranking algorithm based on the preferences set.

hash_content
boolean
default:false

Whether to hash the content of messages.

previous_session
string

The session ID from a previous API call. This allows you to link recommendations such that the recommendation engine knows which requests belong in the same group.

Response

200
application/json
Successful response
llm_label
string
required

Label of the LLM you should call.