Documentation Index
Fetch the complete documentation index at: https://docs.simplismart.ai/llms.txt
Use this file to discover all available pages before exploring further.
Manage model repositories using simplismart model-repos.
List Model Repos
Lists model repositories with optional filtering.
simplismart model-repos list [options]
| Option | Description |
|---|
--offset N | Pagination offset (default: 0) |
--count N | Page size (default: 5) |
--model-id UUID | Fetch specific model repo by ID |
--name NAME | Filter by model name (contains match) |
--status STATUS | Filter by status |
--model-type TYPE | Filter by model type |
--created-by EMAIL | Filter by creator email |
Expected output
{
"limit": 5,
"offset": 0,
"count": 50,
"results": [
{
"uuid": "<MODEL_REPO_ID>",
"name": "whisper-nemo-diarization",
"source_type": "docker_hub",
"source_url": "simplismart/whisper-nemo-diarization:latest",
"is_byom": true,
"accelerator": null,
"runtime_gpus": 1,
"byom": {
"image": "simplismart/MODEL-NAME:latest",
"registry": "simplismart/REGISTRY-NAME",
"tag": "latest"
},
"secrets": {
"source_secret": {
"uuid": "<SECRET_ID>",
"name": "<SECRET_NAME>"
}
},
"status": "SUCCESS",
"model_type": "byom",
"env": {},
"created_at": "2026-03-02T11:52:16.925151Z",
"updated_at": "2026-03-02T11:52:16.925162Z",
"org_id": "<ORG_ID>",
"healthcheck": {
"path": "/health",
"port": 8000,
"periodSeconds": 10,
"timeoutSeconds": 5,
"initialDelaySeconds": 30
},
"ports": {
"http": {
"port": 8000
}
},
"metrics_path": [],
"deployment_custom_configuration": {
"command": []
}
}
]
}
Status Options
| Value | Description |
|---|
SUCCESS | Model repo is ready |
FAILED | Model repo creation failed |
DELETED | Model repo has been deleted |
PROGRESSING | Model repo is being created |
Example:
# List all model repos
simplismart model-repos list --offset 0 --count 5
# Filter by status and name
simplismart model-repos list --status SUCCESS --name vision
# Get specific model repo
simplismart model-repos list --model-id <MODEL_REPO_ID>
Get Model Repo
Gets details of a specific model repository.
simplismart model-repos get [options]
| Option | Description |
|---|
--model-id UUID | Model repo UUID (required) |
Example:
simplismart model-repos get --model-id <MODEL_REPO_ID>
Expected output
{
"uuid": "<MODEL_REPO_ID>",
"name": "whisper-nemo-diarization",
"source_type": "docker_hub",
"source_url": "simplismart/MODEL_NAME:latest",
"is_byom": true,
"accelerator": null,
"runtime_gpus": 1,
"byom": {
"image": "simplismart/MODEL-NAME:latest",
"registry": "simplismart/REGISTRY-NAME",
"tag": "latest"
},
"secrets": {
"source_secret": {
"uuid": "<SECRET_ID>",
"name": "<SECRET_NAME>"
}
},
"status": "SUCCESS",
"model_type": "byom",
"env": {},
"created_at": "2026-03-02T11:52:16.925151Z",
"updated_at": "2026-03-02T11:52:16.925162Z",
"org_id": "<ORG_ID>",
"healthcheck": {
"path": "/health",
"port": 8000,
"periodSeconds": 10,
"timeoutSeconds": 5,
"initialDelaySeconds": 30
},
"ports": {
"http": {
"port": 8000
}
},
"metrics_path": [],
"deployment_custom_configuration": {
"command": []
}
}
Get Model Profiles
Generates/resolves the model profile from a source type and path. Returns the optimization combinations available for that model (quantizations, machines, and parallelism options).
simplismart model-repos profiles [options]
| Option | Description |
|---|
--type TYPE | Source type (required) - e.g., hf, s3, azure_blob |
--path PATH | Source path (required) |
--secret-id UUID | Secret UUID (optional) |
Example:
simplismart model-repos profiles --type hf --path meta-llama/Llama-3.1-8B-Instruct
Expected output
{
"profile": {
"model_info": {
"type": "llm",
"architecture": "LlamaForCausalLM",
"param_size": "8030261248",
"default_quantization": "float16"
},
"quantizations": {
"float16": {
"machines": [
"nvidia-a100-40gb",
"nvidia-a100-80gb",
"nvidia-h100",
"nvidia-l40s"
]
},
"fp8": {
"machines": [
"nvidia-h100",
"nvidia-h100-pcie",
"nvidia-h100-sxm5"
]
},
"int4": {
"machines": [
"nvidia-a100-40gb",
"nvidia-a10g",
"nvidia-h100",
"nvidia-l4",
"nvidia-l40s",
"nvidia-tesla-t4"
]
}
},
"combinations": [
{
"quantization": "float16",
"machine": "nvidia-h100",
"tensor_parallelism": [1, 2, 4, 8],
"data_parallelism": [1],
"pipeline_parallelism": [1],
"expert_parallelism": [1]
}
]
}
}
Output: JSON with a profile object containing:
- model_info —
type, architecture, param_size, default_quantization
- quantizations — map of quantization (e.g.
float16, fp8, int4) to list of supported machines
- combinations — list of valid (quantization, machine, parallelism) combinations; each has
quantization, machine, tensor_parallelism, data_parallelism, pipeline_parallelism, expert_parallelism (each a list of allowed values)
Bring Your Own Container Model Repo
Bring your own container.
simplismart model-repos create-container [options]
| Option | Description |
|---|
--name NAME | Model repo name (required) |
--source-type TYPE | Source type (required) - docker_hub, depot, nvidiadockersecret |
--runtime-gpus N | Runtime GPU count (required) |
--source-secret UUID | Source secret UUID (required for registry sources) |
--registry-path PATH | Registry path/repo (required) |
--docker-tag TAG | Image tag (required) |
--env KEY=VALUE | Environment variable (can be repeated) |
--healthcheck JSON/@file | JSON or file path for healthcheck |
--ports JSON/@file | JSON or file path for ports |
--metrics-path JSON/@file | JSON array or file path for metrics paths |
--deployment-custom-configuration JSON/@file | Custom deployment config |
Example:
simplismart model-repos create-container \
--name vision-container \
--source-type docker_hub \
--runtime-gpus 1 \
--source-secret <SECRET_ID> \
--registry-path org/vision \
--docker-tag v1.2.3
Create Private Compile Model Repo
Creates a private compile model repository: the platform compiles the model from a source (e.g. Hugging Face) using config files for model, optimisation, and pipeline.
simplismart model-repos create-private-compile [options]
| Option | Description |
|---|
--name NAME | Model repo name (required) |
--source-type TYPE | Source type (required) - e.g. huggingface |
--source-url URL | Source path/URL (required) - e.g. HF repo id |
--mode MODE | Compilation mode (default: public_hf) - e.g. public_hf, private_hf, aws, gcp, public_url, simplismart |
--model-class CLASS | Model class (required) - e.g. LlamaForCausalLM |
--accelerator-type TYPE | Accelerator type (required) - e.g. nvidia-h100 |
--accelerator-count N | Accelerator count (default: 0) |
--org-id UUID | Org UUID (optional if inferred from PG token) |
--cloud-account UUID | Cloud account UUID |
--source-secret UUID | Secret UUID for source access |
--lora-secret UUID | LoRA secret UUID |
--description TEXT | Description |
--short-description TEXT | Short description |
--dropdown-description TEXT | Dropdown description |
--machine-type TYPE | Machine type |
--region REGION | Region |
--resource-group GROUP | Resource group |
--processing-mode MODE | One of: SYNC, ASYNC, REALTIME_ASYNC |
--model-family FAMILY | Model family |
--env KEY=VALUE | Environment variable (repeatable) |
--model-config JSON/@file | Model config |
--optimisation-config JSON/@file | Optimisation config |
--pipeline-config JSON/@file | Pipeline config |
--tasks JSON/@file | Tasks list |
--tags JSON/@file | Tags object |
--output-metadata JSON/@file | Output metadata |
--additional-details JSON/@file | Additional details |
Example:
simplismart model-repos create-private-compile \
--name llama-sdk-cli \
--source-type huggingface \
--source-url meta-llama/Llama-3.2-1B-Instruct \
--model-class LlamaForCausalLM \
--accelerator-type nvidia-h100 \
--model-config @examples/private-compile-sample/model_config.json \
--optimisation-config @examples/private-compile-sample/optimisation_config.json \
--pipeline-config @examples/private-compile-sample/pipeline_config.json
Config files: See the Python SDK Model Repos for the structure of model_config.json, optimisation_config.json, and pipeline_config.json.
Example output
{
"uuid": "<MODEL_REPO_ID>",
"created_at": "2026-03-02T16:07:31.776954Z",
"updated_at": "2026-03-02T16:07:31.776967Z",
"org": "<ORG_ID>",
"is_public": false,
"is_deployable": true,
"name": "llama-sdk-cli",
"is_deleted": false,
"deleted_at": null,
"avatar": {
"uuid": "<AVATAR_ID>",
"image_url": "<IMG_URL>",
"font_color": "#000000",
"background_color": "#f3f3f3"
},
"model_class": "LlamaForCausalLM",
"master_model_class": "llm",
"env": {},
"model_type": "platform",
"tags": [],
"short_description": "",
"source_type": "huggingface",
"source_url": "meta-llama/Llama-3.2-1B-Instruct",
"gpu_count": 1,
"is_dynamo": false,
"status": "PENDING",
"logs_pod_hostname": null
}
Full examples are in the Simplismart SDK example repo.
Delete Model Repo
Deletes a model repository.
simplismart model-repos delete [options]
| Option | Description |
|---|
--model-id UUID | Model repo UUID (required) |
Example:
simplismart model-repos delete --model-id <MODEL_REPO_ID>
Example Output
{"status": "deleted", "model_id": "<MODEL_REPO_ID>"}