Enter Model Details

  • Model name: Provide a name for your model.
  • Source: Specify the source of the model. (e.g., HuggingFace repository)
  • Path: Enter the path to the model file or directory (e.g., openai/whisper-large-v3-turbo.

Note: Only instruct-style models are supported in the model compilation step for LLMs. These are typically chat-optimized models and are often identified by the suffix -Instruct in their names (e.g., meta-llama/Llama-3.2-3B-Instruct).

Base models such as meta-llama/Llama-3.2-3B (without the -Instruct suffix) are not supported.

Optimizing Infrastructure

  • Configure the infrastructure to optimize the model’s performance, such as selecting the appropriate compute resources and optimization techniques.


Configuration

  • Finalize the model’s configuration by setting any additional parameters or preferences required for deployment.