Getting Started
Spaces Guide
Developer Guide
API REFERENCE
- Spaces
- Data
- Chat
- Assistants
- Documents
PULZE ACADEMY
Update Space And Models
Update an Space’s configuration.
Authorizations
Bearer authentication header of the form Bearer <token>
, where <token>
is your auth token.
Body
Used to update the benchmark model
1
The maximum cost allowed for a request. Only works with compounded requests that require multiple LLM calls. If the value is reached, it will exit with an exception.
x > 0.0001
If an LLM call fails, how many times should Pulze retry the call to the same LLM? There will be a maximum of N+1 calls (original + N retries)
0 < x < 3
If an LLM call fails, how many other models should Pulze try, chosen by quality descending? It will be a maximum of N+1 models (original + N other models)
0 < x < 5
Optimize the internal / intermediate LLM requests, for a big gain in speed and cost savings, at the cost of a potential, and very slight, penalty on quality. The final request ("SYNTHESIZE") is always performed using your original settings.
0
, 1
The level of privacy for a given request
0 = (UNSUPPORTED -- public logs)
1 = Log request, response and all of its metadata (Normal mode)
2 = Do not log neither the request prompt nor the response text. Logs are still visible, and all of the request metadata accessible. Retrievable as a log. (TBD)
3 = Do not log at all. Internally, a minimal representation may be stored for billing: model name, tokens used, which app it belongs to, and timestamp. Not retrievable as a log. (TBD)
1
, 2
, 3
Prompt ID that we will use for requests
The app this sandbox relates to
Prioritizes cost when selecting the most optimized models for your use case.
Prioritizes latency and reduces the time delay between submitting a request and receiving the response.
Prioritizes the quality and readability of the generated responses.
Response
The Auth0ID of the creator of the app
Compare the results of LLMs against this model (for speed, quality, cost...)
The user (auth0_id) who created the model
When the model was added. Auto-populated in DB
The app_id that has access to this model (if only one)
A (usually 0) cost added on top of a request. Some models charge per request, not only per token
The cost of a completion token, in USD
The max_tokens for this model
This determines if the model will be available + pre-selected when users create new apps.
A description of the model
The ID of this model
True if the model is of type Chat Completions, False if it's a Text Completion model.
Whether it's fine-tuned or not
True if the model complies with GDPR
True if the model is open source
True if the model is publicly accessible to all
Model has been created and shared by Pulze
Whether it's rag-tuned or not
Test models are only used for testing and do not perform any LLM requests
The name of the model. Can belong to many providers
When the model was updated. Auto-populated in DB
The fully qualified (namespaced) model name
The org_id that has acccess to this model
The unit of billing for this model
tokens
, characters
The cost of a prompt token, in USD
True if the model supports function
/tool
call
True if the model supports json
-formatted responses
True if the model supports n
and best_of
-- i.e, multiple responses
True if the model supports frequency_penalty
and presence_penalty
True if the model supports streaming responses
True if the model supports image recognition (vision)
The most recent data this model has been trained with
A URL to the model's page or more informatino
Store the name of the model the API requires
Extra model settings inferred from namespace
For models whose deprecation date is known (past or future), to show errors and deny service, or show warnings
The owner of the model. Sometimes, for a provider/model combination, many instances exist, trained on different data
The user (auth0_id) who created the model
When the model was added. Auto-populated in DB
The app_id that has access to this model (if only one)
A (usually 0) cost added on top of a request. Some models charge per request, not only per token
The cost of a completion token, in USD
The max_tokens for this model
This determines if the model will be available + pre-selected when users create new apps.
A description of the model
The ID of this model
True if the model is of type Chat Completions, False if it's a Text Completion model.
Whether it's fine-tuned or not
True if the model complies with GDPR
True if the model is open source
True if the model is publicly accessible to all
Model has been created and shared by Pulze
Whether it's rag-tuned or not
Test models are only used for testing and do not perform any LLM requests
The name of the model. Can belong to many providers
When the model was updated. Auto-populated in DB
The fully qualified (namespaced) model name
The org_id that has acccess to this model
The unit of billing for this model
tokens
, characters
The cost of a prompt token, in USD
True if the model supports function
/tool
call
True if the model supports json
-formatted responses
True if the model supports n
and best_of
-- i.e, multiple responses
True if the model supports frequency_penalty
and presence_penalty
True if the model supports streaming responses
True if the model supports image recognition (vision)
The most recent data this model has been trained with
A URL to the model's page or more informatino
Store the name of the model the API requires
Extra model settings inferred from namespace
For models whose deprecation date is known (past or future), to show errors and deny service, or show warnings
The owner of the model. Sometimes, for a provider/model combination, many instances exist, trained on different data
The ID of parent, in case it's not a base model
The ID of prompt, used for this model
The provider for the model.
The ID of parent, in case it's not a base model
The ID of prompt, used for this model
The provider for the model.
Compare the results of LLMs against this model (for speed, quality, cost...)
A Failover chain, when enabled, skips the SMART router and instead calls the failover chain in order.
The hashed value of the API Key which we store in our database
True if the app is active (not soft-deleted, etc.), false otherwise
The last characters of the API key
The Auth0ID of the last person to modify the table
A name for this app
The Prompt associated to the app, if any
1 - 200
1 - 60
Reason for decline
A logo for this app
A description for this app
The Org to which this App belongs
The maximum cost allowed for a request. Only works with compounded requests that require multiple LLM calls. If the value is reached, it will exit with an exception.
x > 0.0001
If an LLM call fails, how many times should Pulze retry the call to the same LLM? There will be a maximum of N+1 calls (original + N retries)
0 < x < 3
If an LLM call fails, how many other models should Pulze try, chosen by quality descending? It will be a maximum of N+1 models (original + N other models)
0 < x < 5
Optimize the internal / intermediate LLM requests, for a big gain in speed and cost savings, at the cost of a potential, and very slight, penalty on quality. The final request ("SYNTHESIZE") is always performed using your original settings.
0
, 1
The level of privacy for a given request
0 = (UNSUPPORTED -- public logs)
1 = Log request, response and all of its metadata (Normal mode)
2 = Do not log neither the request prompt nor the response text. Logs are still visible, and all of the request metadata accessible. Retrievable as a log. (TBD)
3 = Do not log at all. Internally, a minimal representation may be stored for billing: model name, tokens used, which app it belongs to, and timestamp. Not retrievable as a log. (TBD)
1
, 2
, 3
Prompt ID that we will use for requests
If any, the ID of the prompt associated with this app
The app might be a (testing) sandbox of a different App.
If true, the app will use the custom data of the parent app
Prioritizes cost when selecting the most optimized models for your use case.
Prioritizes latency and reduces the time delay between submitting a request and receiving the response.
Prioritizes the quality and readability of the generated responses.
The max_tokens for this model
A description of the model
Used to uniquely target models when we enable/disable them
The name of the model. Can belong to many providers
The fully qualified (namespaced) model name
The most recent data this model has been trained with
A URL to the model's page or more informatino
Whether the model is active for the app.
Store the name of the model the API requires
Extra model settings inferred from namespace
For models whose deprecation date is known (past or future), to show errors and deny service, or show warnings
Whether the model is active for the org.
The owner of the model. Sometimes, for a provider/model combination, many instances exist, trained on different data
The user (auth0_id) who created the model
When the model was added. Auto-populated in DB
The app_id that has access to this model (if only one)
A (usually 0) cost added on top of a request. Some models charge per request, not only per token
The cost of a completion token, in USD
The max_tokens for this model
This determines if the model will be available + pre-selected when users create new apps.
A description of the model
The ID of this model
True if the model is of type Chat Completions, False if it's a Text Completion model.
Whether it's fine-tuned or not
True if the model complies with GDPR
True if the model is open source
True if the model is publicly accessible to all
Model has been created and shared by Pulze
Whether it's rag-tuned or not
Test models are only used for testing and do not perform any LLM requests
The name of the model. Can belong to many providers
When the model was updated. Auto-populated in DB
The fully qualified (namespaced) model name
The org_id that has acccess to this model
The unit of billing for this model
tokens
, characters
The cost of a prompt token, in USD
True if the model supports function
/tool
call
True if the model supports json
-formatted responses
True if the model supports n
and best_of
-- i.e, multiple responses
True if the model supports frequency_penalty
and presence_penalty
True if the model supports streaming responses
True if the model supports image recognition (vision)
The most recent data this model has been trained with
A URL to the model's page or more informatino
Store the name of the model the API requires
Extra model settings inferred from namespace
For models whose deprecation date is known (past or future), to show errors and deny service, or show warnings
The owner of the model. Sometimes, for a provider/model combination, many instances exist, trained on different data
The user (auth0_id) who created the model
When the model was added. Auto-populated in DB
The app_id that has access to this model (if only one)
A (usually 0) cost added on top of a request. Some models charge per request, not only per token
The cost of a completion token, in USD
The max_tokens for this model
This determines if the model will be available + pre-selected when users create new apps.
A description of the model
The ID of this model
True if the model is of type Chat Completions, False if it's a Text Completion model.
Whether it's fine-tuned or not
True if the model complies with GDPR
True if the model is open source
True if the model is publicly accessible to all
Model has been created and shared by Pulze
Whether it's rag-tuned or not
Test models are only used for testing and do not perform any LLM requests
The name of the model. Can belong to many providers
When the model was updated. Auto-populated in DB
The fully qualified (namespaced) model name
The org_id that has acccess to this model
The unit of billing for this model
tokens
, characters
The cost of a prompt token, in USD
True if the model supports function
/tool
call
True if the model supports json
-formatted responses
True if the model supports n
and best_of
-- i.e, multiple responses
True if the model supports frequency_penalty
and presence_penalty
True if the model supports streaming responses
True if the model supports image recognition (vision)
The most recent data this model has been trained with
A URL to the model's page or more informatino
Store the name of the model the API requires
Extra model settings inferred from namespace
For models whose deprecation date is known (past or future), to show errors and deny service, or show warnings
The owner of the model. Sometimes, for a provider/model combination, many instances exist, trained on different data
The ID of parent, in case it's not a base model
The ID of prompt, used for this model
The provider for the model.
The ID of parent, in case it's not a base model
The ID of prompt, used for this model
The provider for the model.
The ID of parent, in case it's not a base model
The ID of prompt, used for this model
The provider for the model.
The max_tokens for this model
A description of the model
Used to uniquely target models when we enable/disable them
The name of the model. Can belong to many providers
The fully qualified (namespaced) model name
The most recent data this model has been trained with
A URL to the model's page or more informatino
Whether the model is active for the app.
Store the name of the model the API requires
Extra model settings inferred from namespace
For models whose deprecation date is known (past or future), to show errors and deny service, or show warnings
Whether the model is active for the org.
The owner of the model. Sometimes, for a provider/model combination, many instances exist, trained on different data
The user (auth0_id) who created the model
When the model was added. Auto-populated in DB
The app_id that has access to this model (if only one)
A (usually 0) cost added on top of a request. Some models charge per request, not only per token
The cost of a completion token, in USD
The max_tokens for this model
This determines if the model will be available + pre-selected when users create new apps.
A description of the model
The ID of this model
True if the model is of type Chat Completions, False if it's a Text Completion model.
Whether it's fine-tuned or not
True if the model complies with GDPR
True if the model is open source
True if the model is publicly accessible to all
Model has been created and shared by Pulze
Whether it's rag-tuned or not
Test models are only used for testing and do not perform any LLM requests
The name of the model. Can belong to many providers
When the model was updated. Auto-populated in DB
The fully qualified (namespaced) model name
The org_id that has acccess to this model
The unit of billing for this model
tokens
, characters
The cost of a prompt token, in USD
True if the model supports function
/tool
call
True if the model supports json
-formatted responses
True if the model supports n
and best_of
-- i.e, multiple responses
True if the model supports frequency_penalty
and presence_penalty
True if the model supports streaming responses
True if the model supports image recognition (vision)
The most recent data this model has been trained with
A URL to the model's page or more informatino
Store the name of the model the API requires
Extra model settings inferred from namespace
For models whose deprecation date is known (past or future), to show errors and deny service, or show warnings
The owner of the model. Sometimes, for a provider/model combination, many instances exist, trained on different data
The user (auth0_id) who created the model
When the model was added. Auto-populated in DB
The app_id that has access to this model (if only one)
A (usually 0) cost added on top of a request. Some models charge per request, not only per token
The cost of a completion token, in USD
The max_tokens for this model
This determines if the model will be available + pre-selected when users create new apps.
A description of the model
The ID of this model
True if the model is of type Chat Completions, False if it's a Text Completion model.
Whether it's fine-tuned or not
True if the model complies with GDPR
True if the model is open source
True if the model is publicly accessible to all
Model has been created and shared by Pulze
Whether it's rag-tuned or not
Test models are only used for testing and do not perform any LLM requests
The name of the model. Can belong to many providers
When the model was updated. Auto-populated in DB
The fully qualified (namespaced) model name
The org_id that has acccess to this model
The unit of billing for this model
tokens
, characters
The cost of a prompt token, in USD
True if the model supports function
/tool
call
True if the model supports json
-formatted responses
True if the model supports n
and best_of
-- i.e, multiple responses
True if the model supports frequency_penalty
and presence_penalty
True if the model supports streaming responses
True if the model supports image recognition (vision)
The most recent data this model has been trained with
A URL to the model's page or more informatino
Store the name of the model the API requires
Extra model settings inferred from namespace
For models whose deprecation date is known (past or future), to show errors and deny service, or show warnings
The owner of the model. Sometimes, for a provider/model combination, many instances exist, trained on different data
The ID of parent, in case it's not a base model
The ID of prompt, used for this model
The provider for the model.
The ID of parent, in case it's not a base model
The ID of prompt, used for this model
The provider for the model.
The ID of parent, in case it's not a base model
The ID of prompt, used for this model
The provider for the model.
The user (auth0_id) who created the model
When the model was added. Auto-populated in DB
The app_id that has access to this model (if only one)
A (usually 0) cost added on top of a request. Some models charge per request, not only per token
The cost of a completion token, in USD
The max_tokens for this model
This determines if the model will be available + pre-selected when users create new apps.
A description of the model
The ID of this model
True if the model is of type Chat Completions, False if it's a Text Completion model.
Whether it's fine-tuned or not
True if the model complies with GDPR
True if the model is open source
True if the model is publicly accessible to all
Model has been created and shared by Pulze
Whether it's rag-tuned or not
Test models are only used for testing and do not perform any LLM requests
The name of the model. Can belong to many providers
When the model was updated. Auto-populated in DB
The fully qualified (namespaced) model name
The org_id that has acccess to this model
The unit of billing for this model
tokens
, characters
The cost of a prompt token, in USD
True if the model supports function
/tool
call
True if the model supports json
-formatted responses
True if the model supports n
and best_of
-- i.e, multiple responses
True if the model supports frequency_penalty
and presence_penalty
True if the model supports streaming responses
True if the model supports image recognition (vision)
The most recent data this model has been trained with
A URL to the model's page or more informatino
Store the name of the model the API requires
Extra model settings inferred from namespace
For models whose deprecation date is known (past or future), to show errors and deny service, or show warnings
The owner of the model. Sometimes, for a provider/model combination, many instances exist, trained on different data
The user (auth0_id) who created the model
When the model was added. Auto-populated in DB
The app_id that has access to this model (if only one)
A (usually 0) cost added on top of a request. Some models charge per request, not only per token
The cost of a completion token, in USD
The max_tokens for this model
This determines if the model will be available + pre-selected when users create new apps.
A description of the model
The ID of this model
True if the model is of type Chat Completions, False if it's a Text Completion model.
Whether it's fine-tuned or not
True if the model complies with GDPR
True if the model is open source
True if the model is publicly accessible to all
Model has been created and shared by Pulze
Whether it's rag-tuned or not
Test models are only used for testing and do not perform any LLM requests
The name of the model. Can belong to many providers
When the model was updated. Auto-populated in DB
The fully qualified (namespaced) model name
The org_id that has acccess to this model
The unit of billing for this model
tokens
, characters
The cost of a prompt token, in USD
True if the model supports function
/tool
call
True if the model supports json
-formatted responses
True if the model supports n
and best_of
-- i.e, multiple responses
True if the model supports frequency_penalty
and presence_penalty
True if the model supports streaming responses
True if the model supports image recognition (vision)
The most recent data this model has been trained with
A URL to the model's page or more informatino
Store the name of the model the API requires
Extra model settings inferred from namespace
For models whose deprecation date is known (past or future), to show errors and deny service, or show warnings
The owner of the model. Sometimes, for a provider/model combination, many instances exist, trained on different data
The user (auth0_id) who created the model
When the model was added. Auto-populated in DB
The app_id that has access to this model (if only one)
A (usually 0) cost added on top of a request. Some models charge per request, not only per token
The cost of a completion token, in USD
The max_tokens for this model
This determines if the model will be available + pre-selected when users create new apps.
A description of the model
The ID of this model
True if the model is of type Chat Completions, False if it's a Text Completion model.
Whether it's fine-tuned or not
True if the model complies with GDPR
True if the model is open source
True if the model is publicly accessible to all
Model has been created and shared by Pulze
Whether it's rag-tuned or not
Test models are only used for testing and do not perform any LLM requests
The name of the model. Can belong to many providers
When the model was updated. Auto-populated in DB
The fully qualified (namespaced) model name
The org_id that has acccess to this model
The unit of billing for this model
tokens
, characters
The cost of a prompt token, in USD
True if the model supports function
/tool
call
True if the model supports json
-formatted responses
True if the model supports n
and best_of
-- i.e, multiple responses
True if the model supports frequency_penalty
and presence_penalty
True if the model supports streaming responses
True if the model supports image recognition (vision)
The most recent data this model has been trained with
A URL to the model's page or more informatino
Store the name of the model the API requires
Extra model settings inferred from namespace
For models whose deprecation date is known (past or future), to show errors and deny service, or show warnings
The owner of the model. Sometimes, for a provider/model combination, many instances exist, trained on different data
The ID of parent, in case it's not a base model
The ID of prompt, used for this model
The provider for the model.
The ID of parent, in case it's not a base model
The ID of prompt, used for this model
The provider for the model.
The ID of parent, in case it's not a base model
The ID of prompt, used for this model
The provider for the model.
The Auth0ID of the creator of the app
Compare the results of LLMs against this model (for speed, quality, cost...)
The user (auth0_id) who created the model
When the model was added. Auto-populated in DB
The app_id that has access to this model (if only one)
A (usually 0) cost added on top of a request. Some models charge per request, not only per token
The cost of a completion token, in USD
The max_tokens for this model
This determines if the model will be available + pre-selected when users create new apps.
A description of the model
The ID of this model
True if the model is of type Chat Completions, False if it's a Text Completion model.
Whether it's fine-tuned or not
True if the model complies with GDPR
True if the model is open source
True if the model is publicly accessible to all
Model has been created and shared by Pulze
Whether it's rag-tuned or not
Test models are only used for testing and do not perform any LLM requests
The name of the model. Can belong to many providers
When the model was updated. Auto-populated in DB
The fully qualified (namespaced) model name
The org_id that has acccess to this model
The unit of billing for this model
tokens
, characters
The cost of a prompt token, in USD
True if the model supports function
/tool
call
True if the model supports json
-formatted responses
True if the model supports n
and best_of
-- i.e, multiple responses
True if the model supports frequency_penalty
and presence_penalty
True if the model supports streaming responses
True if the model supports image recognition (vision)
The most recent data this model has been trained with
A URL to the model's page or more informatino
Store the name of the model the API requires
Extra model settings inferred from namespace
For models whose deprecation date is known (past or future), to show errors and deny service, or show warnings
The owner of the model. Sometimes, for a provider/model combination, many instances exist, trained on different data
The user (auth0_id) who created the model
When the model was added. Auto-populated in DB
The app_id that has access to this model (if only one)
A (usually 0) cost added on top of a request. Some models charge per request, not only per token
The cost of a completion token, in USD
The max_tokens for this model
This determines if the model will be available + pre-selected when users create new apps.
A description of the model
The ID of this model
True if the model is of type Chat Completions, False if it's a Text Completion model.
Whether it's fine-tuned or not
True if the model complies with GDPR
True if the model is open source
True if the model is publicly accessible to all
Model has been created and shared by Pulze
Whether it's rag-tuned or not
Test models are only used for testing and do not perform any LLM requests
The name of the model. Can belong to many providers
When the model was updated. Auto-populated in DB
The fully qualified (namespaced) model name
The org_id that has acccess to this model
The unit of billing for this model
tokens
, characters
The cost of a prompt token, in USD
True if the model supports function
/tool
call
True if the model supports json
-formatted responses
True if the model supports n
and best_of
-- i.e, multiple responses
True if the model supports frequency_penalty
and presence_penalty
True if the model supports streaming responses
True if the model supports image recognition (vision)
The most recent data this model has been trained with
A URL to the model's page or more informatino
Store the name of the model the API requires
Extra model settings inferred from namespace
For models whose deprecation date is known (past or future), to show errors and deny service, or show warnings
The owner of the model. Sometimes, for a provider/model combination, many instances exist, trained on different data
The ID of parent, in case it's not a base model
The ID of prompt, used for this model
The provider for the model.
The ID of parent, in case it's not a base model
The ID of prompt, used for this model
The provider for the model.
Compare the results of LLMs against this model (for speed, quality, cost...)
A Failover chain, when enabled, skips the SMART router and instead calls the failover chain in order.
The hashed value of the API Key which we store in our database
True if the app is active (not soft-deleted, etc.), false otherwise
The last characters of the API key
The Auth0ID of the last person to modify the table
A name for this app
The Prompt associated to the app, if any
1 - 200
1 - 60
Reason for decline
A logo for this app
A description for this app
The Org to which this App belongs
The maximum cost allowed for a request. Only works with compounded requests that require multiple LLM calls. If the value is reached, it will exit with an exception.
x > 0.0001
If an LLM call fails, how many times should Pulze retry the call to the same LLM? There will be a maximum of N+1 calls (original + N retries)
0 < x < 3
If an LLM call fails, how many other models should Pulze try, chosen by quality descending? It will be a maximum of N+1 models (original + N other models)
0 < x < 5
Optimize the internal / intermediate LLM requests, for a big gain in speed and cost savings, at the cost of a potential, and very slight, penalty on quality. The final request ("SYNTHESIZE") is always performed using your original settings.
0
, 1
The level of privacy for a given request
0 = (UNSUPPORTED -- public logs)
1 = Log request, response and all of its metadata (Normal mode)
2 = Do not log neither the request prompt nor the response text. Logs are still visible, and all of the request metadata accessible. Retrievable as a log. (TBD)
3 = Do not log at all. Internally, a minimal representation may be stored for billing: model name, tokens used, which app it belongs to, and timestamp. Not retrievable as a log. (TBD)
1
, 2
, 3
Prompt ID that we will use for requests
If any, the ID of the prompt associated with this app
The app might be a (testing) sandbox of a different App.
If true, the app will use the custom data of the parent app
Prioritizes cost when selecting the most optimized models for your use case.
Prioritizes latency and reduces the time delay between submitting a request and receiving the response.
Prioritizes the quality and readability of the generated responses.
The max_tokens for this model
A description of the model
Used to uniquely target models when we enable/disable them
The name of the model. Can belong to many providers
The fully qualified (namespaced) model name
The most recent data this model has been trained with
A URL to the model's page or more informatino
Whether the model is active for the app.
Store the name of the model the API requires
Extra model settings inferred from namespace
For models whose deprecation date is known (past or future), to show errors and deny service, or show warnings
Whether the model is active for the org.
The owner of the model. Sometimes, for a provider/model combination, many instances exist, trained on different data
The user (auth0_id) who created the model
When the model was added. Auto-populated in DB
The app_id that has access to this model (if only one)
A (usually 0) cost added on top of a request. Some models charge per request, not only per token
The cost of a completion token, in USD
The max_tokens for this model
This determines if the model will be available + pre-selected when users create new apps.
A description of the model
The ID of this model
True if the model is of type Chat Completions, False if it's a Text Completion model.
Whether it's fine-tuned or not
True if the model complies with GDPR
True if the model is open source
True if the model is publicly accessible to all
Model has been created and shared by Pulze
Whether it's rag-tuned or not
Test models are only used for testing and do not perform any LLM requests
The name of the model. Can belong to many providers
When the model was updated. Auto-populated in DB
The fully qualified (namespaced) model name
The org_id that has acccess to this model
The unit of billing for this model
tokens
, characters
The cost of a prompt token, in USD
True if the model supports function
/tool
call
True if the model supports json
-formatted responses
True if the model supports n
and best_of
-- i.e, multiple responses
True if the model supports frequency_penalty
and presence_penalty
True if the model supports streaming responses
True if the model supports image recognition (vision)
The most recent data this model has been trained with
A URL to the model's page or more informatino
Store the name of the model the API requires
Extra model settings inferred from namespace
For models whose deprecation date is known (past or future), to show errors and deny service, or show warnings
The owner of the model. Sometimes, for a provider/model combination, many instances exist, trained on different data
The user (auth0_id) who created the model
When the model was added. Auto-populated in DB
The app_id that has access to this model (if only one)
A (usually 0) cost added on top of a request. Some models charge per request, not only per token
The cost of a completion token, in USD
The max_tokens for this model
This determines if the model will be available + pre-selected when users create new apps.
A description of the model
The ID of this model
True if the model is of type Chat Completions, False if it's a Text Completion model.
Whether it's fine-tuned or not
True if the model complies with GDPR
True if the model is open source
True if the model is publicly accessible to all
Model has been created and shared by Pulze
Whether it's rag-tuned or not
Test models are only used for testing and do not perform any LLM requests
The name of the model. Can belong to many providers
When the model was updated. Auto-populated in DB
The fully qualified (namespaced) model name
The org_id that has acccess to this model
The unit of billing for this model
tokens
, characters
The cost of a prompt token, in USD
True if the model supports function
/tool
call
True if the model supports json
-formatted responses
True if the model supports n
and best_of
-- i.e, multiple responses
True if the model supports frequency_penalty
and presence_penalty
True if the model supports streaming responses
True if the model supports image recognition (vision)
The most recent data this model has been trained with
A URL to the model's page or more informatino
Store the name of the model the API requires
Extra model settings inferred from namespace
For models whose deprecation date is known (past or future), to show errors and deny service, or show warnings
The owner of the model. Sometimes, for a provider/model combination, many instances exist, trained on different data
The ID of parent, in case it's not a base model
The ID of prompt, used for this model
The provider for the model.
The ID of parent, in case it's not a base model
The ID of prompt, used for this model
The provider for the model.
The ID of parent, in case it's not a base model
The ID of prompt, used for this model
The provider for the model.
The max_tokens for this model
A description of the model
Used to uniquely target models when we enable/disable them
The name of the model. Can belong to many providers
The fully qualified (namespaced) model name
The most recent data this model has been trained with
A URL to the model's page or more informatino
Whether the model is active for the app.
Store the name of the model the API requires
Extra model settings inferred from namespace
For models whose deprecation date is known (past or future), to show errors and deny service, or show warnings
Whether the model is active for the org.
The owner of the model. Sometimes, for a provider/model combination, many instances exist, trained on different data
The user (auth0_id) who created the model
When the model was added. Auto-populated in DB
The app_id that has access to this model (if only one)
A (usually 0) cost added on top of a request. Some models charge per request, not only per token
The cost of a completion token, in USD
The max_tokens for this model
This determines if the model will be available + pre-selected when users create new apps.
A description of the model
The ID of this model
True if the model is of type Chat Completions, False if it's a Text Completion model.
Whether it's fine-tuned or not
True if the model complies with GDPR
True if the model is open source
True if the model is publicly accessible to all
Model has been created and shared by Pulze
Whether it's rag-tuned or not
Test models are only used for testing and do not perform any LLM requests
The name of the model. Can belong to many providers
When the model was updated. Auto-populated in DB
The fully qualified (namespaced) model name
The org_id that has acccess to this model
The unit of billing for this model
tokens
, characters
The cost of a prompt token, in USD
True if the model supports function
/tool
call
True if the model supports json
-formatted responses
True if the model supports n
and best_of
-- i.e, multiple responses
True if the model supports frequency_penalty
and presence_penalty
True if the model supports streaming responses
True if the model supports image recognition (vision)
The most recent data this model has been trained with
A URL to the model's page or more informatino
Store the name of the model the API requires
Extra model settings inferred from namespace
For models whose deprecation date is known (past or future), to show errors and deny service, or show warnings
The owner of the model. Sometimes, for a provider/model combination, many instances exist, trained on different data
The user (auth0_id) who created the model
When the model was added. Auto-populated in DB
The app_id that has access to this model (if only one)
A (usually 0) cost added on top of a request. Some models charge per request, not only per token
The cost of a completion token, in USD
The max_tokens for this model
This determines if the model will be available + pre-selected when users create new apps.
A description of the model
The ID of this model
True if the model is of type Chat Completions, False if it's a Text Completion model.
Whether it's fine-tuned or not
True if the model complies with GDPR
True if the model is open source
True if the model is publicly accessible to all
Model has been created and shared by Pulze
Whether it's rag-tuned or not
Test models are only used for testing and do not perform any LLM requests
The name of the model. Can belong to many providers
When the model was updated. Auto-populated in DB
The fully qualified (namespaced) model name
The org_id that has acccess to this model
The unit of billing for this model
tokens
, characters
The cost of a prompt token, in USD
True if the model supports function
/tool
call
True if the model supports json
-formatted responses
True if the model supports n
and best_of
-- i.e, multiple responses
True if the model supports frequency_penalty
and presence_penalty
True if the model supports streaming responses
True if the model supports image recognition (vision)
The most recent data this model has been trained with
A URL to the model's page or more informatino
Store the name of the model the API requires
Extra model settings inferred from namespace
For models whose deprecation date is known (past or future), to show errors and deny service, or show warnings
The owner of the model. Sometimes, for a provider/model combination, many instances exist, trained on different data
The ID of parent, in case it's not a base model
The ID of prompt, used for this model
The provider for the model.
The ID of parent, in case it's not a base model
The ID of prompt, used for this model
The provider for the model.
The ID of parent, in case it's not a base model
The ID of prompt, used for this model
The provider for the model.
The user (auth0_id) who created the model
When the model was added. Auto-populated in DB
The app_id that has access to this model (if only one)
A (usually 0) cost added on top of a request. Some models charge per request, not only per token
The cost of a completion token, in USD
The max_tokens for this model
This determines if the model will be available + pre-selected when users create new apps.
A description of the model
The ID of this model
True if the model is of type Chat Completions, False if it's a Text Completion model.
Whether it's fine-tuned or not
True if the model complies with GDPR
True if the model is open source
True if the model is publicly accessible to all
Model has been created and shared by Pulze
Whether it's rag-tuned or not
Test models are only used for testing and do not perform any LLM requests
The name of the model. Can belong to many providers
When the model was updated. Auto-populated in DB
The fully qualified (namespaced) model name
The org_id that has acccess to this model
The unit of billing for this model
tokens
, characters
The cost of a prompt token, in USD
True if the model supports function
/tool
call
True if the model supports json
-formatted responses
True if the model supports n
and best_of
-- i.e, multiple responses
True if the model supports frequency_penalty
and presence_penalty
True if the model supports streaming responses
True if the model supports image recognition (vision)
The most recent data this model has been trained with
A URL to the model's page or more informatino
Store the name of the model the API requires
Extra model settings inferred from namespace
For models whose deprecation date is known (past or future), to show errors and deny service, or show warnings
The owner of the model. Sometimes, for a provider/model combination, many instances exist, trained on different data
The user (auth0_id) who created the model
When the model was added. Auto-populated in DB
The app_id that has access to this model (if only one)
A (usually 0) cost added on top of a request. Some models charge per request, not only per token
The cost of a completion token, in USD
The max_tokens for this model
This determines if the model will be available + pre-selected when users create new apps.
A description of the model
The ID of this model
True if the model is of type Chat Completions, False if it's a Text Completion model.
Whether it's fine-tuned or not
True if the model complies with GDPR
True if the model is open source
True if the model is publicly accessible to all
Model has been created and shared by Pulze
Whether it's rag-tuned or not
Test models are only used for testing and do not perform any LLM requests
The name of the model. Can belong to many providers
When the model was updated. Auto-populated in DB
The fully qualified (namespaced) model name
The org_id that has acccess to this model
The unit of billing for this model
tokens
, characters
The cost of a prompt token, in USD
True if the model supports function
/tool
call
True if the model supports json
-formatted responses
True if the model supports n
and best_of
-- i.e, multiple responses
True if the model supports frequency_penalty
and presence_penalty
True if the model supports streaming responses
True if the model supports image recognition (vision)
The most recent data this model has been trained with
A URL to the model's page or more informatino
Store the name of the model the API requires
Extra model settings inferred from namespace
For models whose deprecation date is known (past or future), to show errors and deny service, or show warnings
The owner of the model. Sometimes, for a provider/model combination, many instances exist, trained on different data
The user (auth0_id) who created the model
When the model was added. Auto-populated in DB
The app_id that has access to this model (if only one)
A (usually 0) cost added on top of a request. Some models charge per request, not only per token
The cost of a completion token, in USD
The max_tokens for this model
This determines if the model will be available + pre-selected when users create new apps.
A description of the model
The ID of this model
True if the model is of type Chat Completions, False if it's a Text Completion model.
Whether it's fine-tuned or not
True if the model complies with GDPR
True if the model is open source
True if the model is publicly accessible to all
Model has been created and shared by Pulze
Whether it's rag-tuned or not
Test models are only used for testing and do not perform any LLM requests
The name of the model. Can belong to many providers
When the model was updated. Auto-populated in DB
The fully qualified (namespaced) model name
The org_id that has acccess to this model
The unit of billing for this model
tokens
, characters
The cost of a prompt token, in USD
True if the model supports function
/tool
call
True if the model supports json
-formatted responses
True if the model supports n
and best_of
-- i.e, multiple responses
True if the model supports frequency_penalty
and presence_penalty
True if the model supports streaming responses
True if the model supports image recognition (vision)
The most recent data this model has been trained with
A URL to the model's page or more informatino
Store the name of the model the API requires
Extra model settings inferred from namespace
For models whose deprecation date is known (past or future), to show errors and deny service, or show warnings
The owner of the model. Sometimes, for a provider/model combination, many instances exist, trained on different data
The ID of parent, in case it's not a base model
The ID of prompt, used for this model
The provider for the model.
The ID of parent, in case it's not a base model
The ID of prompt, used for this model
The provider for the model.
The ID of parent, in case it's not a base model
The ID of prompt, used for this model
The provider for the model.
outdated
Was this page helpful?