cli/foundation-models/system/distillation/summarization/distillation_summarization.yaml (39 lines of code) (raw):

type: distillation name: "Distillation-summarization-llama" description: "Distill student model using a teacher model" experiment_name: "Distillation-summarization" # Data Generation Properties data_generation_type: label_generation data_generation_task_type: summarization # Input data training_data: type: uri_file path: ./train_summarization.jsonl validation_data: type: uri_file path: ./validation_summarization.jsonl # Teacher model serverless endpoint information teacher_model_endpoint_connection: type: serverless name: Meta-Llama-3-1-405B-Instruct-vkn endpoint: https://Meta-Llama-3-1-405B-Instruct-vkn.westus3.models.ai.azure.com/chat/completions api_key: EXAMPLE_API_KEY # Model ID student_model: azureml://registries/azureml-meta/models/Meta-Llama-3.1-8B-Instruct/versions/2 # Output distilled model outputs: registered_model: type: mlflow_model name: llama-summarization-distilled # Teacher model related properties (OPTIONAL) teacher_model_settings: inference_parameters: temperature: 0.8 max_tokens: 1024 top_p: 0.95 endpoint_request_settings: request_batch_size: 10 min_endpoint_success_ratio: 0.7 # System prompt settings (OPTIONAL) prompt_settings: enable_chain_of_density: true max_len_summary: 80 # For finetuning (OPTIONAL) hyperparameters: learning_rate_multiplier: "0.2" n_epochs: "5" batch_size: "2" # Resource for Data Generation Step (OPTIONAL) resources: instance_type: Standard_D2_v2