promptflow/deployment/deployment.yaml (30 lines of code) (raw):

$schema: https://azuremlschemas.azureedge.net/latest/managedOnlineDeployment.schema.json name: blue endpoint_name: tobi-care-ms-promptflow-prod # model: azureml:tobi-llm-chat-model:1 # You can also specify model files path inline model: path: ../code properties: # endpoint detail UI Test tab needs this property to know it's from prompt flow azureml.promptflow.source_flow_id: chat-with-web-model # Following are properties only for chat flow # endpoint detail UI Test tab needs this property to know it's a chat flow azureml.promptflow.mode: chat # endpoint detail UI Test tab needs this property to know which is the input column for chat flow azureml.promptflow.chat_input: question # endpoint detail UI Test tab needs this property to know which is the output column for chat flow azureml.promptflow.chat_output: answer environment: image: mcr.microsoft.com/azureml/promptflow/promptflow-runtime:latest # inference config is used to build a serving container for online deployments inference_config: liveness_route: path: /health port: 8080 readiness_route: path: /health port: 8080 scoring_route: path: /score port: 8080 instance_type: Standard_DS3_v2 instance_count: 1 request_settings: request_timeout_ms: 90000 environment_variables: # "compute" mode is the default mode, if you want to deploy to serving mode, you need to set this env variable to "serving" PROMPTFLOW_RUN_MODE: serving # for pulling connections from workspace PRT_CONFIG_OVERRIDE: deployment.subscription_id=<sub-id>,deployment.resource_group=<resource-group>,deployment.workspace_name=<workspace-name>,deployment.endpoint_name=<endpoint-name>,deployment.deployment_name=<endpoint-deployment-name> # (Optional) When there are multiple fields in the response, using this env variable will filter the fields to expose in the response. # For example, if there are 2 flow outputs: "answer", "context", and I only want to have "answer" in the endpoint response, I can set this env variable to '["answer"]'. # If you don't set this environment, by default all flow outputs will be included in the endpoint response. PROMPTFLOW_RESPONSE_INCLUDED_FIELDS: '["answer"]'