Chat Completions
Perform a Chat Completion request.
Authorizations
Bearer authentication header of the form Bearer <token>
, where <token>
is your auth token.
Body
The prompt text sent (for a /completions request)
The conversation sent (with or without history) (for a /chat/completions request)
https://docs.pulze.ai/overview/models Specify the model you'd like Pulze to use. (optional). Can be the full model name, or a subset for multi-matching.
Defaults to our dynamic routing, i.e. best model for this request.
COMING SOON
The maximum number of tokens that the response can contain.
Optionally specify the temperature for this request only. Leave empty to allow Pulze to guess it for you.
0 < x < 1
https://platform.openai.com/docs/api-reference/completions/create#completions/create-top_p An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass
none
, auto
How many completions to generate for each prompt. @default 1
x > 1
** COMING SOON ** Specify if you want the response to be streamed or to be returned as a standard HTTP request
COMING SOON https://platform.openai.com/docs/api-reference/completions/create#completions/create-logprobs Include the log probabilities on the logprobs most likely tokens, as well the chosen tokens.
0 < x < 5
Stop responding when this sequence of characters is generated. Leave empty to allow the model to decide.
https://platform.openai.com/docs/api-reference/completions/create#completions/create-presence_penalty Increase the model's likelihood to talk about new topics
-2 < x < 2
https://platform.openai.com/docs/api-reference/completions/create#completions/create-frequency_penalty Increase the model's likelihood to not repeat tokens/words
-2 < x < 2
The number of responses to generate. Out of those, it will return the best n
.
COMING SOON https://platform.openai.com/docs/api-reference/completions/create#completions/create-logit_bias Modify the likelihood of specified tokens appearing in the completion.
See here for a detailed explanation on how to use: https://help.openai.com/en/articles/5247780-using-logit-bias-to-define-token-probability
Response
The response returned to the user by the Chat Completions endpoint
The fully qualified model name used by PulzeEngine
The type of response object
text_completion
, chat.completion
Creation timestamp
Metadata of the response
This ID gets generated by the database when we save the request
Tokens used
Was this page helpful?