assets/models/system/Mixtral-8x7B-v0-1/spec.yaml (44 lines of code) (raw):

$schema: https://azuremlschemas.azureedge.net/latest/model.schema.json name: mistralai-Mixtral-8x7B-v01 path: ./ properties: SharedComputeCapacityEnabled: true SHA: 985aa055896a8f943d4a9f2572e6ea1341823841 inference-min-sku-spec: 40|8|672|2900 inference-recommended-sku: Standard_ND40rs_v2, Standard_ND96amsr_A100_v4, Standard_ND96asr_v4 finetune-min-sku-spec: 40|2|440|128 finetune-recommended-sku: Standard_ND40rs_v2, Standard_NC48ads_A100_v4, Standard_NC96ads_A100_v4, Standard_ND96amsr_A100_v4, Standard_ND96asr_v4 finetuning-tasks: text-generation languages: EN tags: Featured: '' SharedComputeCapacityEnabled: '' hiddenlayerscanned: "" disable-batch: 'true' huggingface_model_id: mistralai/Mixtral-8x7B-v0.1 inference_compute_allow_list: [ Standard_ND40rs_v2, Standard_ND96amsr_A100_v4, Standard_ND96asr_v4, ] inference_supported_envs: - vllm finetune_compute_allow_list: - Standard_ND40rs_v2 - Standard_NC48ads_A100_v4 - Standard_NC96ads_A100_v4 - Standard_ND96asr_v4 - Standard_ND96amsr_A100_v4 model_specific_defaults: max_seq_length: 2048 apply_lora: "true" apply_deepspeed: "true" deepspeed_stage: "3" precision: "16" ignore_mismatched_sizes: "false" license: apache-2.0 task: text-generation author: "Mistral AI" benchmark: "quality" version: 15