Authorizations
Bearer authentication header of the form Bearer <token>
, where <token>
is your auth token.
Body
<provider>/<model> to use in the query. If <model> is unique across providers, can supply <model> only
"openai/gpt-4o-mini"
"gpt-4o-mini"
A list of messages comprising the conversation so far.
Key-value pairs of metadata for the request. Identify end-users associated with the query with the waystone
metadata. All other fields are metadata for OpenAI requests and are not used by Waystone.
{
"waystone": "{\n\t\t\t\t\t\t\"user\": {\n\t\t\t\t\t\t\t\"id\": \"user123\",\n\t\t\t\t\t\t\t\"metadata\": { \"email\": \"user@example.com\", \"name\": \"John Doe\" }\n\t\t\t\t\t\t}\n\t\t\t\t\t}"
}
{
"waystone": "{\n\t\t\t\t\t\t\"user\": \"user123\",\n\t\t\t\t\t\t\"group\": { \"id\": \"group123\", \"metadata\": { \"name\": \"Group Name\" } }\n\t\t\t\t\t}"
}
For OpenAI: Whether or not to store the output of this chat completion request in OpenAI
If set to true, the model response data will be streamed to the client as it is generated
For OpenAI: A unique identifier representing your end-user. NOT used by Waystone.
What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. We generally recommend altering this or top_p but not both.
0 <= x <= 2
Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model's likelihood to repeat the same line verbatim.
-2 <= x <= 2
An upper bound for the number of tokens that can be generated for a completion, including visible output tokens and reasoning tokens.
How many chat completion choices to generate for each input message. Note that you will be charged based on the number of generated tokens across all of the choices. Keep n as 1 to minimize costs.