Model Compilation
Optimise a Model
Steps to optimise and add a new model to your repository
Enter Model Details
- Model name: Provide a name for your model.
- Source: Specify the source of the model. (e.g., HuggingFace repository)
- Path: Enter the path to the model file or directory (e.g., openai/whisper-large-v3-turbo.
Getting your model path from HuggingFace
Getting your model path from HuggingFace
- Visit huggingface.co.
- Use the search bar to find the desired model. (e.g., “whisper-large”)
- Click on the model you want from the search results. (e.g., openai/whisper-large-v3-turbo)
- Copy the model path displayed at the top of the page (e.g., openai/whisper-large-v3-turbo) for use.
The model path on HuggingFace follows the format: creator/model-slug.
Note: Only instruct-style models are supported in the model compilation step for LLMs. These are typically chat-optimized models and are often identified by the suffix -Instruct
in their names (e.g., meta-llama/Llama-3.2-3B-Instruct
).
Base models such as meta-llama/Llama-3.2-3B
(without the -Instruct
suffix) are not supported.
Optimizing Infrastructure
- Configure the infrastructure to optimize the model’s performance, such as selecting the appropriate compute resources and optimization techniques.
Configuration
- Finalize the model’s configuration by setting any additional parameters or preferences required for deployment.