Create a chat completion for given messages with streaming support
JWT token for authentication
Model UUID for the request
Array of messages in the conversation
Model identifier
llama3_3 Whether to stream the response
Sampling temperature
0 <= x <= 2Maximum number of tokens to generate
1 <= x <= 4096Nucleus sampling parameter
0 <= x <= 1Sequences where the API will stop generating