The content of the message.
The role of the sender (e.g., 'user' or 'assistant').
user - userassistant - assistantThe ID of the project to use.
If set, partial message deltas will be sent, like in ChatGPT.
Optionalfrequency_Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency.
Optionallogit_JSON object that maps tokens to an associated bias value from -100 to 100.
Optionalmax_The maximum number of tokens to generate in the chat completion.
OptionalmodelID of the model to use. See the model endpoint compatibility table for details.
Optionalpresence_Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the text so far.
OptionalrepositoriesOptions for Retrieval Augmented Generation (RAG). Will override launched model settings
Optionalids?: number[]The IDs of the repositories to use.
Optionallimit?: numberOptionalsimilarity_Optionalresponse_An object specifying the format that the model must output.
OptionalseedThis feature is in Beta. If specified, our system will make a best effort to sample deterministically.
Optionalsession_The ID of the session to use. It helps to track the chat history.
OptionalstopUp to 4 sequences where the API will stop generating further tokens.
Optionalsystem_The system prompt to use.
OptionaltemperatureWhat sampling temperature to use, between 0 and 2.
OptionaltoolsA list of tools the model may call. Currently, only functions are supported as a tool.
Optionaltop_An alternative to sampling with temperature, called nucleus sampling.
OptionaluserA unique identifier representing your end-user.
A list of messages comprising the conversation so far.