Create a chat completion for given messages with streaming support
Documentation Index
Fetch the complete documentation index at: https://docs.simplismart.ai/llms.txt
Use this file to discover all available pages before exploring further.
JWT token for authentication
Array of messages in the conversation
Model identifier
meta-llama/Meta-Llama-3.1-8B-Instruct Whether to stream the response
Sampling temperature
0 <= x <= 2Maximum number of tokens to generate
1 <= x <= 4096Nucleus sampling parameter
0 <= x <= 1Sequences where the API will stop generating