Creates a model response. Provide text or image inputs to generate text or JSON outputs. Have the model call your own custom code or use built-in tools like web search or file search to use your own data as input for the model’s response.
Bearer authentication header of the form Bearer <token>, where <token> is your auth token.
Key-value pairs of metadata for the request. Identify end-users associated with the query with the waystone metadata. All other fields are metadata for OpenAI requests and are not used by Waystone.
For OpenAI: Whether or not to store the output of this chat completion request in OpenAI
If set to true, the model response data will be streamed to the client as it is generated
For OpenAI: A unique identifier representing your end-user. NOT used by Waystone.
What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. We generally recommend altering this or top_p but not both.
0 <= x <= 2Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text so far, decreasing the model's likelihood to repeat the same line verbatim.
-2 <= x <= 2An upper bound for the number of tokens that can be generated for a completion, including visible output tokens and reasoning tokens.
How many chat completion choices to generate for each input message. Note that you will be charged based on the number of generated tokens across all of the choices. Keep n as 1 to minimize costs.
Successful response
The response is of type object.