POST
/
v1
/
models
/
rank
curl --request POST \
  --url https://api.pulze.ai/v1/models/rank \
  --header 'Authorization: Bearer <token>' \
  --header 'Content-Type: application/json' \
  --data '{
  "model": "<string>",
  "max_tokens": 123,
  "temperature": 0.5,
  "top_p": 123,
  "tools": [
    {
      "type": "function",
      "function": {
        "name": "<string>",
        "description": "<string>",
        "parameters": {
          "type": "object",
          "properties": {},
          "required": [
            "<string>"
          ]
        }
      }
    }
  ],
  "tool_choice": "none",
  "n": 2,
  "stream": true,
  "logprobs": 2,
  "stop": "<string>",
  "presence_penalty": 0,
  "frequency_penalty": 0,
  "best_of": 2,
  "logit_bias": {},
  "response_format": {
    "type": "text"
  },
  "messages": [
    {
      "role": "user",
      "content": "<string>",
      "tool_calls": [
        {
          "function": {
            "arguments": "<string>",
            "name": "<string>"
          },
          "id": "<string>",
          "type": "function"
        }
      ]
    }
  ],
  "assistant_id": "3c90c3cc-0d44-4b50-8888-8dd25736052a",
  "assistant_version_id": "3c90c3cc-0d44-4b50-8888-8dd25736052a",
  "conversation_id": "3c90c3cc-0d44-4b50-8888-8dd25736052a",
  "parent_request_id": "3c90c3cc-0d44-4b50-8888-8dd25736052a",
  "plugins": [
    "legacy-file-search"
  ],
  "images": [
    "<string>"
  ],
  "files": [
    "<string>"
  ]
}'
{
  "models": [],
  "hits": [
    {
      "id": "<string>",
      "category": "<string>",
      "similarity": 123
    }
  ]
}

This endpoint supports:

Weights

Authorizations

Authorization
string
header
required

Bearer authentication header of the form Bearer <token>, where <token> is your auth token.

Body

application/json

Used after validation is completed -- therefore no validation required here

model
string | null
default:pulze

https://docs.pulze.ai/overview/models Specify the model you'd like Pulze to use. (optional). Can be the full model name, or a subset for multi-matching.

Defaults to our dynamic routing, i.e. best model for this request.

max_tokens
integer | null

The maximum number of tokens that the response can contain.

temperature
number | null
default:1

Optionally specify the temperature for this request only. Leave empty to allow Pulze to guess it for you.

Required range: 0 <= x <= 1
top_p
number | null
default:1

https://octo.ai/docs/text-gen-solution/rest-api#input-parameters A value between 0.0 and 1.0 that controls the probability of the model generating a particular token.

tools
object[] | null
tool_choice
default:none
Available options:
none,
auto
n
integer | null

How many completions to generate for each prompt. @default 1

Required range: x >= 1
stream
boolean | null
default:false

Specify if you want the response to be streamed or to be returned as a standard HTTP request. Currently we only support streaming for OpenAI-compatible models.

logprobs
integer | null

COMING SOON https://platform.openai.com/docs/api-reference/completions/create#completions/create-logprobs Include the log probabilities on the logprobs most likely tokens, as well the chosen tokens.

Required range: 0 <= x <= 5
stop
default:

Stop responding when this sequence of characters is generated. Leave empty to allow the model to decide.

presence_penalty
number | null
Required range: -2 <= x <= 2
frequency_penalty
number | null
Required range: -2 <= x <= 2
best_of
integer | null

The number of responses to generate. Out of those, it will return the best n.

Required range: x >= 1
logit_bias
object | null

COMING SOON https://platform.openai.com/docs/api-reference/completions/create#completions/create-logit_bias Modify the likelihood of specified tokens appearing in the completion.

See here for a detailed explanation on how to use: https://help.openai.com/en/articles/5247780-using-logit-bias-to-define-token-probability

response_format
object | null

https://platform.openai.com/docs/api-reference/chat/create#chat-create-response_format An object specifying the format that the model must output. Must be one of "text" or "json_object". Important: when using JSON mode, you must also instruct the model to produce JSON yourself via a system or user message. To help ensure you don't forget, the API will throw an error if the string "JSON" does not appear somewhere in the context.

messages
object[] | null
assistant_id
string | null

The Assistant ID to use for this request

assistant_version_id
string | null

The Assistant version ID to use for this request

conversation_id
string | null

The Conversation ID to use for this request

parent_request_id
string | null

The Parent Request ID to use for this request

plugins
string[]

The list of plugins to enable for the request

images
string[] | null

Images to be analyzed

files
string[] | null

A list of file urls that should be included. Images, audio, etc.

Response

200
application/json
Successful Response
models
object[]

A list of the best models for the given prompt+config, sorted by overall score descending.

hits
object[] | null

A list of most similar scored prompts if any