You can view the supported models here.
Upload your training dataset. Refer to Dataset Preparation for more info.
The supported file format is JSONL
Please note that pre-processing is not handled by our training suite.
The data should be provided in the required format (JSONL), including tasks such as image resizing, rotation, or any necessary transformations (e.g., preprocessing PDFs or images).
Update the training parameters based on your requirements for the training job.
Here is a short explanation of these parameters:
Gradient Accumulation Steps: Number of steps to accumulate gradients before updating model weights.
Learning Rate: Controls how much the model adjusts its weights during training.
Batch Size: Number of samples processed together in one training iteration. Available GPU RAM dictates the maximum batch size.
Trainer Epochs: Total number of times the model trains on the complete dataset.
Adapter Alpha: The weight changes applied to the original model weights are scaled by a factor, determined as alpha divided by the rank to balance adaptation and original model knowledge.
Adapter R: The integer rank of the update matrices. A lower rank creates smaller update matrices, reducing the number of trainable parameters. Alpha recommended to be twice as rank
Adapter Dropout: The dropout probability for the LoRA layers, used to prevent overfitting. A value of 0.1 (10%) indicates a 10% chance for each neuron to be dropped during training.
Update Training Configuration:
The training configuration provides a flexible way to define the inputs, outputs, and other advanced settings for your model. Below is a sample configuration:
Once the configuration is updated, start the training job and monitor its progress in the Recent Jobs section of the UI. Keep track of metrics, logs, and any intermediate results to ensure the training meets your requirements.
After the model is trained, you can also deploy the LoRA via Simplismart. Click here to know how.
You can view the supported models here.
Upload your training dataset. Refer to Dataset Preparation for more info.
The supported file format is JSONL
Please note that pre-processing is not handled by our training suite.
The data should be provided in the required format (JSONL), including tasks such as image resizing, rotation, or any necessary transformations (e.g., preprocessing PDFs or images).
Update the training parameters based on your requirements for the training job.
Here is a short explanation of these parameters:
Gradient Accumulation Steps: Number of steps to accumulate gradients before updating model weights.
Learning Rate: Controls how much the model adjusts its weights during training.
Batch Size: Number of samples processed together in one training iteration. Available GPU RAM dictates the maximum batch size.
Trainer Epochs: Total number of times the model trains on the complete dataset.
Adapter Alpha: The weight changes applied to the original model weights are scaled by a factor, determined as alpha divided by the rank to balance adaptation and original model knowledge.
Adapter R: The integer rank of the update matrices. A lower rank creates smaller update matrices, reducing the number of trainable parameters. Alpha recommended to be twice as rank
Adapter Dropout: The dropout probability for the LoRA layers, used to prevent overfitting. A value of 0.1 (10%) indicates a 10% chance for each neuron to be dropped during training.
Update Training Configuration:
The training configuration provides a flexible way to define the inputs, outputs, and other advanced settings for your model. Below is a sample configuration:
Once the configuration is updated, start the training job and monitor its progress in the Recent Jobs section of the UI. Keep track of metrics, logs, and any intermediate results to ensure the training meets your requirements.
After the model is trained, you can also deploy the LoRA via Simplismart. Click here to know how.