repo_id stringlengths 15 132 | file_path stringlengths 34 176 | content stringlengths 2 3.52M | __index_level_0__ int64 0 0 |
|---|---|---|---|
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/eval_flow_with_composite_image/flow.dag.yaml | inputs:
image_list:
type: list
default:
- data:image/jpg;path: logo.jpg
- data:image/png;path: logo_2.png
image_dict:
type: object
default:
image_1:
data:image/jpg;path: logo.jpg
image_2:
data:image/png;path: logo_2.png
outputs:
output:
type: list
refere... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/eval_flow_with_composite_image/inputs.jsonl | {"image_list":[{"data:image/jpg;path": "logo.jpg"},{"data:image/jpg;path": "logo_2.png"}],"image_dict":{"image_1": {"data:image/jpg;path": "logo.jpg"},"image_2": {"data:image/png;path": "logo_2.png"}}}
{"image_list":[{"data:image/jpg;path": "logo_2.png"},{"data:image/jpg;path": "logo.jpg"}],"image_dict":{"image_1": {"d... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/partial_fail/data.jsonl | {"key": "no"}
{"key": "raise"}
{"key": "matter"}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/partial_fail/flow.dag.yaml | inputs:
key:
type: string
outputs:
output:
type: string
reference: ${print_env.output.value}
nodes:
- name: print_env
type: python
source:
type: code
path: print_env.py
inputs:
key: ${inputs.key}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/partial_fail/print_env.py | import os
from promptflow import tool
@tool
def get_env_var(key: str):
if key == "raise":
raise Exception("expected raise!")
print(os.environ.get(key))
# get from env var
return {"value": os.environ.get(key)}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/all_depedencies_bypassed_with_activate_met/pass_through.py | from promptflow import tool
@tool
def pass_through(input1: str="Execution") -> str:
return input1 | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/all_depedencies_bypassed_with_activate_met/flow.dag.yaml | inputs:
text:
type: string
default: hi
outputs:
output:
type: string
reference: ${nodeB.output}
nodes:
- name: nodeA
type: python
source:
type: code
path: pass_through.py
inputs:
input1: ${inputs.text}
activate:
when: ${inputs.text}
is: world
- name: nodeB
type: python
... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/all_depedencies_bypassed_with_activate_met/inputs.json | {
"text": "hi"
} | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/all_depedencies_bypassed_with_activate_met/expected_result.json | [
{
"expected_node_count": 2,
"expected_outputs": {
"output": "Execution"
},
"expected_bypassed_nodes": [
"nodeA"
]
}
] | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/flow_with_dict_input/print_val.py | from promptflow import tool
@tool
def print_val(val, origin_val):
print(val)
print(origin_val)
if not isinstance(origin_val, dict):
raise TypeError(f"key must be a dict, got {type(origin_val)}")
return val
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/flow_with_dict_input/get_dict_val.py | from promptflow import tool
@tool
def get_dict_val(key):
# get from env var
print(key)
if not isinstance(key, dict):
raise TypeError(f"key must be a dict, got {type(key)}")
return {"value": f"{key}: {type(key)}", "origin_value": key}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/flow_with_dict_input/flow.dag.yaml | inputs:
key:
type: object
outputs:
output:
type: string
reference: ${get_dict_val.output.value}
nodes:
- name: get_dict_val
type: python
source:
type: code
path: get_dict_val.py
inputs:
key: ${inputs.key}
- name: print_val
type: python
source:
type: code
path: print_val.py
... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/flow_with_dict_input | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/flow_with_dict_input/.promptflow/flow.tools.json | {
"package": {},
"code": {
"print_val.py": {
"name": "print_val.py",
"type": "python",
"inputs": {
"key": {
"type": [
"object"
]
}
},
"source": "pri... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/unordered_nodes_with_skip/flow.dag.yaml | name: node_wrong_order
inputs:
text:
type: string
skip:
type: bool
outputs:
result:
type: string
reference: ${third_node}
nodes:
- name: third_node
type: python
source:
type: code
path: test.py
inputs:
text: ${second_node}
- name: first_node
type: python
source:
type: cod... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/eval-classification-accuracy/data.jsonl | {"groundtruth": "App","prediction": "App"}
{"groundtruth": "Channel","prediction": "Channel"}
{"groundtruth": "Academic","prediction": "Academic"}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/eval-classification-accuracy/calculate_accuracy.py | from typing import List
from promptflow import log_metric, tool
@tool
def calculate_accuracy(grades: List[str]):
result = []
for index in range(len(grades)):
grade = grades[index]
result.append(grade)
# calculate accuracy for each variant
accuracy = round((result.count("Correct") / l... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/eval-classification-accuracy/grade.py | from promptflow import tool
@tool
def grade(groundtruth: str, prediction: str):
return "Correct" if groundtruth.lower() == prediction.lower() else "Incorrect"
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/eval-classification-accuracy/flow.dag.yaml | $schema: https://azuremlschemas.azureedge.net/promptflow/latest/Flow.schema.json
inputs:
groundtruth:
type: string
description: Please specify the groundtruth column, which contains the true label
to the outputs that your flow produces.
default: APP
prediction:
type: string
description: Pl... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/failed_flow/hello.py | import os
import openai
from dotenv import load_dotenv
from promptflow import tool
# The inputs section will change based on the arguments of the tool function, after you save the code
# Adding type to arguments and return value will help the system show the types properly
# Please update the function name/signature ... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/failed_flow/flow.dag.yaml | inputs:
text:
type: string
outputs:
output_prompt:
type: string
reference: ${echo_my_prompt.output}
nodes:
- inputs:
text: ${inputs.text}
name: echo_my_prompt
type: python
source:
type: code
path: hello.py
node_variants: {}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/external_files/fetch_text_content_from_url.py | import bs4
import requests
from promptflow import tool
@tool
def fetch_text_content_from_url(url: str):
# Send a request to the URL
try:
headers = {
"User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/113.0.0.0 Safari/537.36 Edg/113.0.177... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/external_files/convert_to_dict.py | import json
from promptflow import tool
@tool
def convert_to_dict(input_str: str):
try:
return json.loads(input_str)
except Exception as e:
print("input is not valid, error: {}".format(e))
return {"category": "None", "evidence": "None"}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/external_files/summarize_text_content.jinja2 | Please summarize the following text in one paragraph. 100 words.
Do not add any information that is not in the text.
Text: {{text}}
Summary:
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/flow_with_user_output/print_val.py | import sys
from promptflow import tool
@tool
def get_val(key):
# get from env var
print(key)
print("user log")
print("error log", file=sys.stderr) | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/flow_with_user_output/flow.dag.yaml | inputs:
key:
type: object
outputs:
output:
type: string
reference: ${print_val.output.value}
nodes:
- name: print_val
type: python
source:
type: code
path: print_val.py
inputs:
key: ${inputs.key}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/flow_with_user_output | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/flow_with_user_output/.promptflow/flow.tools.json | {
"package": {},
"code": {
"print_val.py": {
"name": "print_val.py",
"type": "python",
"inputs": {
"key": {
"type": [
"object"
]
}
},
"source": "pri... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/activate_condition_always_met/pass_through.py | from promptflow import tool
@tool
def pass_through(input1: str) -> str:
return 'hello ' + input1
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/activate_condition_always_met/flow.dag.yaml | inputs:
text:
type: string
default: hello
outputs:
output:
type: string
reference: ${nodeC.output}
nodes:
- name: nodeA
type: python
source:
type: code
path: pass_through.py
inputs:
input1: ${inputs.text}
activate:
when: ${inputs.text}
is: hi
- name: nodeB
type: python
... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/activate_condition_always_met/summary_result.py | from promptflow import tool
@tool
def summary_result(input1: str="Node A not executed.", input2: str="Node B not executed.") -> str:
return input1 + ' ' + input2
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/activate_condition_always_met/inputs.json | {
"text": "hello"
} | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/activate_condition_always_met/expected_result.json | [
{
"expected_node_count": 3,
"expected_outputs": {
"output": "Node A not executed. Node B not executed."
},
"expected_bypassed_nodes": [
"nodeA",
"nodeB"
]
}
] | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/basic_with_builtin_llm_node/hello.jinja2 | system:
You are a assistant which can write code. Response should only contain code.
user:
Write a simple {{text}} program that displays the greeting message when executed. | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/basic_with_builtin_llm_node/flow.dag.yaml | $schema: https://azuremlschemas.azureedge.net/promptflow/latest/Flow.schema.json
inputs:
text:
type: string
default: Python Hello World!
outputs:
output:
type: string
reference: ${llm.output}
nodes:
- name: hello_prompt
type: prompt
inputs:
text: ${inputs.text}
source:
type: code
p... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/flow_with_dict_input_with_variant/print_val.py | from promptflow import tool
from promptflow.connections import CustomConnection
@tool
def get_val(key, conn: CustomConnection):
# get from env var
print(key)
if not isinstance(key, dict):
raise TypeError(f"key must be a dict, got {type(key)}")
return {"value": f"{key}: {type(key)}"}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/flow_with_dict_input_with_variant/flow.dag.yaml | inputs:
key:
type: object
outputs:
output:
type: string
reference: ${print_val.output.value}
nodes:
- name: print_val
use_variants: true
type: python
source:
type: code
path: print_val.py
node_variants:
print_val:
default_variant_id: variant1
variants:
variant1:
nod... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/flow_with_environment_variables/flow.dag.yaml | environment_variables:
env1: 2
env2: spawn
env3:
- 1
- 2
- 3
- 4
- 5
env4:
a: 1
b: "2"
inputs:
key:
type: string
outputs:
output:
type: string
reference: ${print_env.output.value}
nodes:
- name: print_env
type: python
source:
type: code
path: print_env.py
inputs:
... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/flow_with_environment_variables/inputs.jsonl | {"text": "env1"}
{"text": "env2"}
{"text": "env3"}
{"text": "env4"}
{"text": "env5"}
{"text": "env10"} | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/flow_with_environment_variables/print_env.py | import os
from promptflow import tool
@tool
def get_env_var(key: str):
print(os.environ.get(key))
# get from env var
return {"value": os.environ.get(key)}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/chat_flow_without_defined_chat_history/show_answer.py | from promptflow import tool
@tool
def show_answer(chat_answer: str):
print("print:", chat_answer)
return chat_answer
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/chat_flow_without_defined_chat_history/chat.jinja2 | system:
You are a helpful assistant.
{% for item in chat_history %}
user:
{{item.inputs.question}}
assistant:
{{item.outputs.answer}}
{% endfor %}
user:
{{question}} | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/chat_flow_without_defined_chat_history/flow.dag.yaml | inputs:
chat_history:
type: list
is_chat_history: false
question:
type: string
is_chat_input: true
default: What is ChatGPT?
outputs:
answer:
type: string
reference: ${show_answer.output}
is_chat_output: true
nodes:
- inputs:
deployment_name: gpt-35-turbo
max_tokens: "256"
... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/generator_tools/echo.py | from promptflow import tool
from char_generator import character_generator
@tool
def echo(text):
"""Echo the input string."""
echo_text = "Echo - " + "".join(character_generator(text))
return echo_text | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/generator_tools/flow.dag.yaml | inputs:
text:
type: string
outputs:
answer:
type: string
reference: ${echo.output}
nodes:
- name: echo
type: python
source:
type: code
path: echo.py
inputs:
text: ${inputs.text}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/generator_tools/char_generator.py | from promptflow import tool
@tool
def character_generator(text: str):
"""Generate characters from a string."""
for char in text:
yield char | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/basic-with-connection/hello.py | import json
import os
import openai
from openai.version import VERSION as OPENAI_VERSION
from promptflow import tool
from promptflow.connections import AzureOpenAIConnection
from promptflow.tools.common import render_jinja_template, parse_chat
# The inputs section will change based on the arguments of the tool funct... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/basic-with-connection/data.jsonl | {"text": "Hello World!"}
{"text": "Hello PromptFlow!"}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/basic-with-connection/hello.jinja2 | {# Please replace the template with your own prompt. #}
system:
You task is to generate what I ask
user:
Write a simple {{text}} program that displays the greeting message when executed.
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/basic-with-connection/flow.dag.yaml | inputs:
text:
type: string
outputs:
output_prompt:
type: string
reference: ${echo_my_prompt.output}
nodes:
- inputs:
text: ${inputs.text}
name: hello_prompt
type: prompt
source:
type: code
path: hello.jinja2
- inputs:
prompt: ${hello_prompt.output}
deployment_name: gpt-35-turbo... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/basic-with-connection/inputs.json | {}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/basic-with-connection/samples.json | [{"text": "Hello World!"}]
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/sample_flow_with_functions/run_function.py | from promptflow import tool
import json
def get_current_weather(location, unit="fahrenheit"):
"""Get the current weather in a given location"""
weather_info = {
"location": location,
"temperature": "72",
"unit": unit,
"forecast": ["sunny", "windy"],
}
return weather_inf... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/sample_flow_with_functions/use_functions_with_chat_models.jinja2 | system:
Don't make assumptions about what values to plug into functions. Ask for clarification if a user request is ambiguous.
{% for item in chat_history %}
user:
{{item.inputs.question}}
{% if 'function_call' in item.outputs.llm_output %}
assistant:
Function generation requested, function = {{item.outputs.llm_outpu... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/sample_flow_with_functions/flow.dag.yaml | id: use_functions_with_chat_models
name: Use Functions with Chat Models
inputs:
chat_history:
type: list
default:
- inputs:
question: What is the weather like in Boston?
outputs:
answer: '{"forecast":["sunny","windy"],"location":"Boston","temperature":"72","unit":"fahrenheit"}'
... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/sample_flow_with_functions/samples.json | [
{
"question": "How about London next week?"
}
] | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/conditional_flow_with_activate/investigation_steps.jinja2 | system:
You are a helpful assistant.
user:
When an incident occurs, there have 3 ways to deal with it, please choose one.
1. {{first_method}}
2. {{second_method}}
3. {{third_method}} | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/conditional_flow_with_activate/incident_info_extractor.py | from promptflow import tool
@tool
def extract_incident_info(incident: dict) -> str:
retriever_type = ["icm", "tsg", "kql"]
return {
"retriever": retriever_type[incident["incident_id"]],
"incident_content": incident["incident_content"]
} | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/conditional_flow_with_activate/tsg_retriever.py | from promptflow import tool
@tool
def tsg_retriever(content: str) -> str:
return "TSG: " + content | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/conditional_flow_with_activate/job_info_extractor.py | from promptflow import tool
@tool
def extract_job_info(incident_content: str) -> str:
print(f"Incident: {incident_content}")
return "Execute job info extractor"
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/conditional_flow_with_activate/investigation_method.py | from promptflow import tool
@tool
def choose_investigation_method(method1="Skip job info extractor", method2="Skip incident info extractor"):
method = {}
if method1:
method["first"] = method1
if method2:
method["second"] = method2
return method
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/conditional_flow_with_activate/retriever_summary.py | from promptflow import tool
@tool
def retriever_summary(summary) -> str:
print(f"Summary: {summary}")
return "Execute incident info extractor"
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/conditional_flow_with_activate/incident_id_extractor.py | from promptflow import tool
@tool
def extract_incident_id(incident_content: str, incident_id: int):
if incident_id >= 0 and incident_id < 3:
return {
"has_incident_id": True,
"incident_id": incident_id,
"incident_content": incident_content
}
return {
... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/conditional_flow_with_activate/icm_retriever.py | from promptflow import tool
@tool
def icm_retriever(content: str) -> str:
return "ICM: " + content | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/conditional_flow_with_activate/expected_status_summary.json | {
"incident_id_extractor.completed": 3,
"job_info_extractor.completed": 1,
"job_info_extractor.bypassed": 2,
"incident_info_extractor.completed": 2,
"incident_info_extractor.bypassed": 1,
"icm_retriever.completed": 1,
"icm_retriever.bypassed": 2,
"tsg_retriever.completed": 1,
"tsg_re... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/conditional_flow_with_activate/flow.dag.yaml | id: template_standard_flow
name: Template Standard Flow
inputs:
incident_content:
type: string
incident_id:
type: int
outputs:
investigation_method:
type: string
reference: ${investigation_method.output}
nodes:
- name: incident_id_extractor
type: python
source:
type: code
path: inciden... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/conditional_flow_with_activate/inputs.json | [
{
"incident_id": 1,
"incident_content": "Incident 418856448 : Stale App Deployment for App promptflow"
},
{
"incident_id": 3,
"incident_content": "Incident 418856448 : Stale App Deployment for App promptflow"
},
{
"incident_id": 0,
"incident_content"... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/conditional_flow_with_activate/expected_result.json | [
{
"expected_node_count": 9,
"expected_outputs":{
"investigation_method": {
"first": "Skip job info extractor",
"second": "Execute incident info extractor"
}
},
"expected_bypassed_nodes":["job_info_extractor", "icm_retriever"]
... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/conditional_flow_with_activate/kql_tsg_retriever.py | from promptflow import tool
@tool
def kql_retriever(content: str) -> str:
return "KQL: " + content | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/custom_connection_flow/flow.dag.yaml | inputs:
key:
type: string
outputs:
output:
type: string
reference: ${print_env.output.value}
nodes:
- name: print_env
type: python
source:
type: code
path: print_env.py
inputs:
key: ${inputs.key}
connection: custom_connection
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/custom_connection_flow/print_env.py | import os
from promptflow import tool
from promptflow.connections import CustomConnection
@tool
def get_env_var(key: str, connection: CustomConnection):
# get from env var
return {"value": os.environ.get(key)}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/custom_connection_flow | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/custom_connection_flow/.promptflow/flow.tools.json | {
"package": {},
"code": {
"print_env.py": {
"type": "python",
"inputs": {
"key": {
"type": [
"string"
]
}
},
"function": "get_env_var"
}
}
}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_run_operations_TestFlowRun_test_input_mapping_with_dict.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 azure-ai-ml/1.12.1 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/3.10.13 (Windows-... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_run_operations_TestFlowRun_test_run_data_not_provided.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 promptflow/0.0.1 azure-ai-ml/1.12.1 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_telemetry_TestTelemetry_test_sdk_telemetry_ua.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 prompt-flow-extension/1.0.0 promptflow/0.0.1 azure-ai-ml/1.12.0
azsdk-python-mgmt-machinele... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_run_operations_TestFlowRun_test_run_bulk_with_registry_flow.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 azure-ai-ml/1.12.1 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/3.10.13 (Windows-... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_connection_operations_TestConnectionOperations_test_list_connection_spec.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 azure-ai-ml/1.10.0 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/3.10.13 (Windows-... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_flow_operations_TestFlow_test_create_flow.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 azure-ai-ml/1.12.1 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/3.10.13 (Windows-... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_run_operations_TestFlowRun_test_download_run.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 azure-ai-ml/1.12.1 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/3.10.13 (Windows-... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_run_operations_TestFlowRun_test_show_run.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 azure-ai-ml/1.12.1 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/3.10.13 (Windows-... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_run_operations_TestFlowRun_test_eager_flow_crud.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 azure-ai-ml/1.12.0 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/3.11.5 (Windows-1... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_run_operations_TestFlowRun_test_flow_id_in_submission.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 promptflow/0.0.1 azure-ai-ml/1.12.1 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_run_operations_TestFlowRun_test_run_bulk_from_yaml.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 azure-ai-ml/1.12.1 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/3.10.13 (Windows-... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_flow_operations_TestFlow_test_list_flows.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 azure-ai-ml/1.12.1 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/3.10.13 (Windows-... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_cli_with_azure_TestCliWithAzure_test_basic_flow_run_bulk_without_env.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 azure-ai-ml/1.12.1 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/3.10.13 (Windows-... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_run_operations_TestFlowRun_test_eager_flow_cancel.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 azure-ai-ml/1.12.0 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/3.11.5 (Windows-1... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_run_operations_TestFlowRun_test_pf_run_with_env_var.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 promptflow/0.0.1 azure-ai-ml/1.12.1 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_cli_with_azure_TestCliWithAzure_test_run_file_with_set.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 azure-ai-ml/1.12.1 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/3.10.13 (Windows-... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_run_operations_TestFlowRun_test_requirements_in_additional_includes.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 azure-ai-ml/1.12.0 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/3.11.5 (Windows-1... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_run_operations_TestFlowRun_test_list_runs.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 promptflow/0.0.1 azure-ai-ml/1.12.1 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_run_operations_TestFlowRun_test_update_run.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 azure-ai-ml/1.12.1 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/3.10.13 (Windows-... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_run_operations_TestFlowRun_test_run_with_remote_data.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 azure-ai-ml/1.12.1 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/3.10.13 (Windows-... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_run_operations_TestFlowRun_test_basic_evaluation_without_data.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 azure-ai-ml/1.12.1 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/3.10.13 (Windows-... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_run_operations_TestFlowRun_test_run_submission_exception.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 promptflow/0.0.1 azure-ai-ml/1.12.1 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_cli_with_azure_TestCliWithAzure_test_run_with_remote_data.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 azure-ai-ml/1.12.1 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/3.10.13 (Windows-... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_run_operations_TestFlowRun_test_stream_run_logs.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 azure-ai-ml/1.12.1 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/3.10.13 (Windows-... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_run_operations_TestFlowRun_test_auto_resolve_requirements.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 azure-ai-ml/1.12.0 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/3.11.5 (Windows-1... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_run_operations_TestFlowRun_test_show_metrics.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 azure-ai-ml/1.12.1 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/3.10.13 (Windows-... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_connection_operations_TestConnectionOperations_test_connection_get_create_delete.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 promptflow/0.0.1 azure-ai-ml/1.12.1 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_run_operations_TestFlowRun_test_run_with_env_overwrite.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 azure-ai-ml/1.12.1 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/3.10.13 (Windows-... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_run_operations_TestFlowRun_test_cancel_run.yaml | interactions:
- request:
body: null
headers:
Accept:
- application/json
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
User-Agent:
- promptflow-sdk/0.0.1 azure-ai-ml/1.12.1 azsdk-python-mgmt-machinelearningservices/0.1.0
Python/3.10.13 (Windows-... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs | promptflow_repo/promptflow/src/promptflow/tests/test_configs/recordings/test_azure_cli_perf_TestAzureCliPerf_test_run_restore.yaml | interactions:
- request:
body: '{"hidden": false}'
headers:
Accept:
- '*/*'
Accept-Encoding:
- gzip, deflate
Connection:
- keep-alive
Content-Length:
- '17'
Content-Type:
- application/json
User-Agent:
- python-requests/2.31.0
method: P... | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.