Detailed guide for configuring LLM training parameters, including input/output features, quantization settings, and prompt templating for combining multiple data columns.
The training configuration provides a flexible way to define the inputs, outputs, and other advanced settings for your model.
Max Sequence Length (input and output features) : The ‘max_sequence_length’ parameter in the config refers to the maximum tokens in the prompt and response respectively.
If Quantisation of the model is not required (for smaller models like 3B, 2B etc) we can remove the quantization key from the advanced configuration, so that all the parameters are used in full precision which yields better accuracy (however training time would increase).
Ensure that the names in the input and output features match exactly with the corresponding columns in the dataset. (e.g., “question” for input and “answer” for output)