POST
/
v1
/
chat
/
completions

Authorizations

Authorization
string
header
required

Bearer authentication header of the form Bearer <token>, where <token> is your auth token.

Body

application/json
messages
object[]
required

The conversation sent (with or without history) (for a /chat/completions request)

best_of
integer | null

The number of responses to generate. Out of those, it will return the best n.

Required range: x > 1
frequency_penalty
number | null

https://platform.openai.com/docs/api-reference/completions/create#completions/create-frequency_penalty Increase the model's likelihood to not repeat tokens/words

Required range: -2 < x < 2
logit_bias
object | null

COMING SOON https://platform.openai.com/docs/api-reference/completions/create#completions/create-logit_bias Modify the likelihood of specified tokens appearing in the completion.

See here for a detailed explanation on how to use: https://help.openai.com/en/articles/5247780-using-logit-bias-to-define-token-probability

logprobs
integer | null

COMING SOON https://platform.openai.com/docs/api-reference/completions/create#completions/create-logprobs Include the log probabilities on the logprobs most likely tokens, as well the chosen tokens.

Required range: 0 < x < 5
max_tokens
integer | null

The maximum number of tokens that the response can contain.

model
string | null
default:
pulze

https://docs.pulze.ai/overview/models Specify the model you'd like Pulze to use. (optional). Can be the full model name, or a subset for multi-matching.

Defaults to our dynamic routing, i.e. best model for this request.

n
integer | null

How many completions to generate for each prompt. @default 1

Required range: x > 1
plugins
string[]

The list of plugins to enable for the request

presence_penalty
number | null

https://platform.openai.com/docs/api-reference/completions/create#completions/create-presence_penalty Increase the model's likelihood to talk about new topics

Required range: -2 < x < 2
response_format
object | null

https://platform.openai.com/docs/api-reference/chat/create#chat-create-response_format An object specifying the format that the model must output. Must be one of "text" or "json_object". Important: when using JSON mode, you must also instruct the model to produce JSON yourself via a system or user message. To help ensure you don't forget, the API will throw an error if the string "JSON" does not appear somewhere in the context.

stop
default:

Stop responding when this sequence of characters is generated. Leave empty to allow the model to decide.

stream
boolean | null
default:
false

Specify if you want the response to be streamed or to be returned as a standard HTTP request. Currently we only support streaming for OpenAI-compatible models.

temperature
number | null
default:
1

Optionally specify the temperature for this request only. Leave empty to allow Pulze to guess it for you.

Required range: 0 < x < 1
tool_choice
default:
none
Available options:
none,
auto
tools
object[] | null
top_p
number | null
default:
1

https://octo.ai/docs/text-gen-solution/rest-api#input-parameters A value between 0.0 and 1.0 that controls the probability of the model generating a particular token.

Response

200 - application/json

The response returned to the user by the Chat Completions endpoint

choices
object[]
required
model
string
required

The fully qualified model name used by PulzeEngine

object
enum<string>
required

The type of response object

Available options:
text_completion,
chat.completion
created
integer
default:
0

Creation timestamp -- in milliseconds (!)

id
string

This ID gets generated by the database when we save the request

metadata
object

Metadata of the response

usage
object

Tokens used