3_llmops-aifoundry/3_2_prototyping/chat/flow.dag.yaml (41 lines of code) (raw):
$schema: https://azuremlschemas.azureedge.net/promptflow/latest/Flow.schema.json
environment:
python_requirements_txt: requirements.txt
inputs:
question:
type: string
is_chat_input: true
default: What is the capital of France?
outputs:
phi35_answer:
type: string
reference: ${phi35.output}
is_chat_output: false
gpt4o_answer:
type: string
reference: ${gpt4o.output}
is_chat_output: true
nodes:
- name: phi35
type: python
source:
type: code
path: phi35_finetuned.py
inputs:
connection: replace with your connection name
input_data: ${inputs.question}
- name: gpt4o
type: llm
source:
type: code
path: chat.jinja2
inputs:
deployment_name: gpt-4o
temperature: 0.7
top_p: 1
max_tokens: 256
question: ${inputs.question}
connection: replace with your connection name
api: chat
module: promptflow.tools.aoai
use_variants: false