repo_id stringlengths 15 132 | file_path stringlengths 34 176 | content stringlengths 2 3.52M | __index_level_0__ int64 0 0 |
|---|---|---|---|
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/chat-with-assistant-no-file/get_calorie_by_jogging.py | import random
import time
from promptflow import tool
@tool
def get_calorie_by_jogging(duration: float, temperature: float):
"""Estimate the calories burned by jogging based on duration and temperature.
:param duration: the length of the jogging in hours.
:type duration: float
:param temperature: th... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/chat-with-assistant-no-file/add_message_and_run.py | import asyncio
import json
from openai import AsyncOpenAI
from openai.types.beta.threads import MessageContentImageFile, MessageContentText
from promptflow import tool, trace
from promptflow.connections import OpenAIConnection
from promptflow.contracts.multimedia import Image
from promptflow.contracts.types import As... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/chat-with-assistant-no-file/flow.dag.yaml | environment:
python_requirements_txt: requirements.txt
version: 2
inputs:
chat_history:
type: list
is_chat_history: true
default: []
question:
type: string
is_chat_input: true
default: I am going to swim today for 30 min in Guangzhou city, how much
calories will I burn?
assistant_i... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/activate_flow/print_input.py | from promptflow import tool
@tool
def print_input(input: str) -> str:
return input
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/activate_flow/flow.dag.yaml | inputs:
text:
type: string
default: world
outputs:
output1:
type: string
reference: ${nodeC.output}
output2:
type: string
reference: ${nodeD.output}
nodes:
- name: nodeA
type: python
source:
type: code
path: print_input.py
inputs:
input: ${inputs.text}
activate:
whe... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/simple_flow_with_python_tool/inputs.jsonl | {"num": "hello"} | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/simple_flow_with_python_tool/divide_num.py | from promptflow import tool
@tool
def divide_num(num: int) -> int:
return (int)(num / 2)
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/simple_flow_with_python_tool/flow.dag.yaml | inputs:
num:
type: int
outputs:
content:
type: string
reference: ${divide_num.output}
nodes:
- name: divide_num
type: python
source:
type: code
path: divide_num.py
inputs:
num: ${inputs.num}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/default_input/samples.json | [
{
"text": "text_1"
}
] | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/default_input/test_print_aggregation.py | from typing import List
from promptflow import tool
@tool
def test_print_input(input_str: List[str], input_bool: List[bool], input_list: List[List], input_dict: List[dict]):
assert input_bool[0] == False
assert input_list[0] == []
assert input_dict[0] == {}
print(input_str)
return input_str | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/default_input/flow.dag.yaml |
inputs:
input_str:
type: string
default: input value from default
input_bool:
type: bool
default: False
input_list:
type: list
default: []
input_dict:
type: object
default: {}
outputs:
output:
type: string
reference: ${test_print_input.output}
nodes:
- name: test_print... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/default_input/test_print_input.py | from promptflow import tool
@tool
def test_print_input(input_str: str, input_bool: bool, input_list: list, input_dict: dict):
assert not input_bool
assert input_list == []
assert input_dict == {}
print(input_str)
return input_str | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/script_with_import/fail.py | from aaa import bbb # noqa: F401 | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/script_with_import/flow.dag.yaml | inputs:
text:
type: string
outputs:
output:
type: string
reference: ${node1.output}
nodes:
- name: node1
type: python
source:
type: code
path: dummy_utils/main.py
inputs:
x: ${inputs.text}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/script_with_import | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/script_with_import/dummy_utils/util_tool.py | from promptflow import tool
@tool
def passthrough(x: str):
return x
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/script_with_import | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/script_with_import/dummy_utils/main.meta.json | {
"name": "main",
"type": "python",
"inputs": {
"x": {
"type": [
"string"
]
}
},
"source": "dummy_utils/main.py",
"function": "main"
} | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/script_with_import | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/script_with_import/dummy_utils/main.py | from promptflow import tool
from dummy_utils.util_tool import passthrough
@tool
def main(x: str):
return passthrough(x)
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/flow_with_dict_input_with_variant/flow.dag.yaml | inputs:
key:
type: object
outputs:
output:
type: string
reference: ${print_val.output.value}
nodes:
- name: print_val
use_variants: true
type: python
source:
type: code
path: print_val.py
node_variants:
print_val:
default_variant_id: variant1
variants:
variant1:
nod... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/flow_with_dict_input_with_variant/print_val.py | from promptflow import tool
from promptflow.connections import CustomConnection
@tool
def get_val(key, conn: CustomConnection):
# get from env var
print(key)
if not isinstance(key, dict):
raise TypeError(f"key must be a dict, got {type(key)}")
return {"value": f"{key}: {type(key)}"}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/chat_flow_with_stream_output/chat.jinja2 | system:
You are a helpful assistant.
{% for item in chat_history %}
user:
{{item.inputs.question}}
assistant:
{{item.outputs.answer}}
{% endfor %}
user:
{{question}} | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/chat_flow_with_stream_output/flow.dag.yaml | inputs:
chat_history:
type: list
is_chat_history: true
question:
type: string
is_chat_input: true
default: What is ChatGPT?
outputs:
answer:
type: string
reference: ${chat_node.output}
is_chat_output: true
nodes:
- inputs:
deployment_name: gpt-35-turbo
max_tokens: "256"
... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/web_classification_no_variants_unordered/samples.json | [
{
"line_number": 0,
"variant_id": "variant_0",
"groundtruth": "App",
"prediction": "App"
}
]
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/web_classification_no_variants_unordered/convert_to_dict.py | import json
from promptflow import tool
@tool
def convert_to_dict(input_str: str):
try:
return json.loads(input_str)
except Exception as e:
print("input is not valid, error: {}".format(e))
return {"category": "None", "evidence": "None"}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/web_classification_no_variants_unordered/fetch_text_content_from_url.py | import bs4
import requests
from promptflow import tool
@tool
def fetch_text_content_from_url(url: str):
# Send a request to the URL
try:
headers = {
"User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/113.0.0.0 Safari/537.36 Edg/113.0.177... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/web_classification_no_variants_unordered/classify_with_llm.jinja2 | Your task is to classify a given url into one of the following types:
Movie, App, Academic, Channel, Profile, PDF or None based on the text content information.
The classification will be based on the url, the webpage text content summary, or both.
Here are a few examples:
{% for ex in examples %}
URL: {{ex.url}}
Text... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/web_classification_no_variants_unordered/summarize_text_content__variant_1.jinja2 | Please summarize some keywords of this paragraph and have some details of each keywords.
Do not add any information that is not in the text.
Text: {{text}}
Summary:
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/web_classification_no_variants_unordered/prepare_examples.py | from promptflow import tool
@tool
def prepare_examples():
return [
{
"url": "https://play.google.com/store/apps/details?id=com.spotify.music",
"text_content": "Spotify is a free music and podcast streaming app with millions of songs, albums, and original podcasts. It also offers au... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/web_classification_no_variants_unordered/flow.dag.yaml | inputs:
url:
type: string
default: https://www.microsoft.com/en-us/d/xbox-wireless-controller-stellar-shift-special-edition/94fbjc7h0h6h
outputs:
category:
type: string
reference: ${convert_to_dict.output.category}
evidence:
type: string
reference: ${convert_to_dict.output.evidence}
nodes:... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/web_classification_no_variants_unordered/summarize_text_content.jinja2 | Please summarize the following text in one paragraph. 100 words.
Do not add any information that is not in the text.
Text: {{text}}
Summary:
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/external_files/convert_to_dict.py | import json
from promptflow import tool
@tool
def convert_to_dict(input_str: str):
try:
return json.loads(input_str)
except Exception as e:
print("input is not valid, error: {}".format(e))
return {"category": "None", "evidence": "None"}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/external_files/fetch_text_content_from_url.py | import bs4
import requests
from promptflow import tool
@tool
def fetch_text_content_from_url(url: str):
# Send a request to the URL
try:
headers = {
"User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/113.0.0.0 Safari/537.36 Edg/113.0.177... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/external_files/summarize_text_content.jinja2 | Please summarize the following text in one paragraph. 100 words.
Do not add any information that is not in the text.
Text: {{text}}
Summary:
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/openai_chat_api_flow/samples.json | {
"question": "What is the capital of the United States of America?",
"chat_history": []
}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/openai_chat_api_flow/inputs.jsonl | {"question": "What is the capital of the United States of America?", "chat_history": [], "stream": true}
{"question": "What is the capital of the United States of America?", "chat_history": [], "stream": false}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/openai_chat_api_flow/chat.py | import openai
from openai.version import VERSION as OPENAI_VERSION
from typing import List
from promptflow import tool
from promptflow.connections import AzureOpenAIConnection
IS_LEGACY_OPENAI = OPENAI_VERSION.startswith("0.")
def get_client(connection: AzureOpenAIConnection):
api_key = connection.api_key
c... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/openai_chat_api_flow/flow.dag.yaml | inputs:
question:
type: string
chat_history:
type: list
stream:
type: bool
outputs:
answer:
type: string
reference: ${chat.output}
nodes:
- name: chat
type: python
source:
type: code
path: chat.py
inputs:
question: ${inputs.question}
chat_history: ${inputs.chat_history}... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/flow_with_environment/requirements | tensorflow | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/flow_with_environment/flow.dag.yaml | inputs:
key:
type: string
outputs:
output:
type: string
reference: ${print_env.output.value}
nodes:
- name: print_env
type: python
source:
type: code
path: print_env.py
inputs:
key: ${inputs.key}
environment:
python_requirements_txt: requirements
image: python:3.8-slim
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/flow_with_environment/print_env.py | import os
from promptflow import tool
@tool
def get_env_var(key: str):
print(os.environ.get(key))
# get from env var
return {"value": os.environ.get(key)}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/flow_with_environment | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/flow_with_environment/.promptflow/flow.tools.json | {
"package": {},
"code": {
"print_env.py": {
"type": "python",
"inputs": {
"key": {
"type": [
"string"
]
}
},
"function": "get_env_var"
}
}
}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/web_classification_input_dir/details.jsonl | {"url": "https://www.youtube.com/watch?v=o5ZQyXaAv1g", "answer": "Channel", "evidence": "Url"}
{"url": "https://www.youtube.com/watch?v=o5ZQyXaAv1g", "answer": "Channel", "evidence": "Url"}
{"url": "https://www.youtube.com/watch?v=o5ZQyXaAv1g", "answer": "Channel", "evidence": "Url"}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/assistant-with-file/assistant_definition.yaml | model: gpt-4-1106-preview
instructions: You are a helpful assistant.
tools:
- type: code_interpreter
- type: function
source:
type: code
path: get_stock_eod_price.py
tool_type: python
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/assistant-with-file/README.md | # Stock EOD Price Analyzer
This sample demonstrates how the PromptFlow Assistant tool help with time series data (stock EOD price) retrieval, plot and consolidation.
Tools used in this flow:
- `get_or_create_thread` tool, python tool, used to provide assistant thread information if absent
- `add_message_and_run` tool... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/assistant-with-file/get_or_create_thread.py | from openai import AsyncOpenAI
from promptflow import tool
from promptflow.connections import OpenAIConnection
@tool
async def get_or_create_thread(conn: OpenAIConnection, thread_id: str):
if thread_id:
return thread_id
cli = AsyncOpenAI(api_key=conn.api_key, organization=conn.organization)
threa... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/assistant-with-file/requirements.txt | promptflow
promptflow-tools | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/assistant-with-file/stock_price.csv | Date,A,B
2023-03-15,100.25,110.50
2023-03-16,102.75,114.35
2023-03-17,101.60,120.10
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/assistant-with-file/add_message_and_run.py | import asyncio
import json
from openai import AsyncOpenAI
from openai.types.beta.threads import MessageContentImageFile, MessageContentText
from promptflow import tool, trace
from promptflow.connections import OpenAIConnection
from promptflow.contracts.multimedia import Image
from promptflow.contracts.types import As... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/assistant-with-file/flow.dag.yaml | version: 2
inputs:
assistant_input:
type: list
default:
- type: text
text: The provided file contains end-of-day (EOD) stock prices for companies A
and B across various dates in March. However, it does not include the
EOD stock prices for Company C.
- type: file_path
file_p... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/assistant-with-file/get_stock_eod_price.py | import random
import time
from promptflow import tool
@tool
def get_stock_eod_price(date: str, company: str):
"""Get the stock end of day price by date and symbol.
:param date: the date of the stock price. e.g. 2021-01-01
:type date: str
:param company: the company name like A, B, C
:type compan... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/chat_flow_with_defined_chat_history/show_answer.py | from promptflow import tool
@tool
def show_answer(chat_answer: str):
print("print:", chat_answer)
return chat_answer
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/chat_flow_with_defined_chat_history/chat.jinja2 | system:
You are a helpful assistant.
{% for item in chat_history %}
user:
{{item.inputs.question}}
assistant:
{{item.outputs.answer}}
{% endfor %}
user:
{{question}} | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/chat_flow_with_defined_chat_history/flow.dag.yaml | inputs:
user_chat_history:
type: list
is_chat_history: true
question:
type: string
is_chat_input: true
default: What is ChatGPT?
outputs:
answer:
type: string
reference: ${show_answer.output}
is_chat_output: true
nodes:
- inputs:
deployment_name: gpt-35-turbo
max_tokens: "2... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/llm_connection_override/connection_arm_template.json | {
"id": "/subscriptions/xxxx/resourceGroups/xxx/providers/Microsoft.MachineLearningServices/workspaces/xxx/connections/azure_open_ai_connection",
"name": "azure_open_ai_connection",
"type": "Microsoft.MachineLearningServices/workspaces/connections",
"properties": {
"authType": "ApiKey",
... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/llm_connection_override/conn_tool.py | from promptflow import tool
from promptflow.connections import AzureOpenAIConnection
@tool
def conn_tool(conn: AzureOpenAIConnection):
assert isinstance(conn, AzureOpenAIConnection)
return conn.api_base | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/llm_connection_override/flow.dag.yaml | inputs: {}
outputs:
output:
type: string
reference: ${conn_node.output}
nodes:
- name: conn_node
type: python
source:
type: code
path: conn_tool.py
inputs:
conn: aoai connection
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/chat_flow_with_multi_output_invalid/chat.jinja2 | system:
You are a helpful assistant.
{% for item in chat_history %}
user:
{{item.inputs.question}}
assistant:
{{item.outputs.answer}}
{% endfor %}
user:
{{question}} | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/chat_flow_with_multi_output_invalid/flow.dag.yaml | inputs:
chat_history:
type: list
question:
type: string
is_chat_input: true
default: What is ChatGPT?
outputs:
answer:
type: string
reference: ${chat_node.output}
is_chat_output: true
multi_answer:
type: string
reference: ${chat_node.output}
is_chat_output: true
nodes:
- ... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/basic_with_builtin_llm_node/flow.dag.yaml | $schema: https://azuremlschemas.azureedge.net/promptflow/latest/Flow.schema.json
inputs:
text:
type: string
default: Python Hello World!
outputs:
output:
type: string
reference: ${llm.output}
nodes:
- name: hello_prompt
type: prompt
inputs:
text: ${inputs.text}
source:
type: code
p... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/basic_with_builtin_llm_node/hello.jinja2 | system:
You are a assistant which can write code. Response should only contain code.
user:
Write a simple {{text}} program that displays the greeting message when executed. | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/flow_with_custom_connection/hello.py | from promptflow import tool
from promptflow.connections import CustomConnection
@tool
def my_python_tool(text: str, connection: CustomConnection) -> dict:
return connection._to_dict()
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/flow_with_custom_connection/flow.dag.yaml | inputs:
text:
type: string
outputs:
output:
type: object
reference: ${hello_node.output}
nodes:
- inputs:
text: ${inputs.text}
connection: basic_custom_connection
name: hello_node
type: python
source:
type: code
path: hello.py
node_variants: {}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/flow_with_package_tool_with_custom_strong_type_connection/data.jsonl | {"text": "Hello World!"}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/flow_with_package_tool_with_custom_strong_type_connection/flow.dag.yaml | inputs:
text:
type: string
default: Hello!
outputs:
out:
type: string
reference: ${My_First_Tool_00f8.output}
nodes:
- name: My_Second_Tool_usi3
type: python
source:
type: package
tool: my_tool_package.tools.my_tool_2.MyTool.my_tool
inputs:
connection: custom_strong_type_connection... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/meta_files/samples.json | [
{
"line_number": 0,
"variant_id": "variant_0",
"groundtruth": "App",
"prediction": "App"
}
]
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/meta_files/remote_fs.meta.yaml | $schema: https://azuremlschemas.azureedge.net/latest/flow.schema.json
name: classification_accuracy_eval
type: evaluate
path: azureml://datastores/workspaceworkingdirectory/paths/Users/wanhan/my_flow_snapshot/flow.dag.yaml
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/meta_files/remote_flow_short_path.meta.yaml | $schema: https://azuremlschemas.azureedge.net/latest/flow.schema.json
name: classification_accuracy_eval
display_name: Classification Accuracy Evaluation
type: evaluate
path: azureml://datastores/workspaceworkingdirectory/paths/Users/wanhan/a/flow.dag.yaml
description: Measuring the performance of a classification syst... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/meta_files/flow.dag.yaml | inputs:
line_number:
type: int
variant_id:
type: string
groundtruth:
type: string
description: Please specify the groundtruth column, which contains the true label
to the outputs that your flow produces.
prediction:
type: string
description: Please specify the prediction column, wh... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/meta_files/flow.meta.yaml | $schema: https://azuremlschemas.azureedge.net/latest/flow.schema.json
name: web_classificiation_flow_3
display_name: Web Classification
type: standard
description: Create flows that use large language models to classify URLs into multiple categories.
path: ./flow.dag.yaml
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/mod-n | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/mod-n/two/mod_two.py | from promptflow import tool
@tool
def mod_two(number: int):
if number % 2 != 0:
raise Exception("cannot mod 2!")
return {"value": number}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/mod-n | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/mod-n/two/flow.dag.yaml | inputs:
number:
type: int
outputs:
output:
type: int
reference: ${mod_two.output.value}
nodes:
- name: mod_two
type: python
source:
type: code
path: mod_two.py
inputs:
number: ${inputs.number}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/mod-n/two | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/mod-n/two/.promptflow/flow.tools.json | {
"code": {
"mod_two.py": {
"type": "python",
"inputs": {
"number": {
"type": [
"int"
]
}
},
"source": "mod_two.py",
"function": "mod_two"
}
... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/mod-n | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/mod-n/three/flow.dag.yaml | inputs:
number:
type: int
outputs:
output:
type: int
reference: ${mod_three.output.value}
nodes:
- name: mod_three
type: python
source:
type: code
path: mod_three.py
inputs:
number: ${inputs.number}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/mod-n | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/mod-n/three/mod_three.py | from promptflow import tool
@tool
def mod_three(number: int):
if number % 3 != 0:
raise Exception("cannot mod 3!")
return {"value": number}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/mod-n/three | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/mod-n/three/.promptflow/flow.tools.json | {
"code": {
"mod_three.py": {
"type": "python",
"inputs": {
"number": {
"type": [
"int"
]
}
},
"source": "mod_three.py",
"function": "mod_three"
... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/activate_condition_always_met/inputs.json | {
"text": "hello"
} | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/activate_condition_always_met/pass_through.py | from promptflow import tool
@tool
def pass_through(input1: str) -> str:
return 'hello ' + input1
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/activate_condition_always_met/expected_result.json | [
{
"expected_node_count": 3,
"expected_outputs": {
"output": "Node A not executed. Node B not executed."
},
"expected_bypassed_nodes": [
"nodeA",
"nodeB"
]
}
] | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/activate_condition_always_met/flow.dag.yaml | inputs:
text:
type: string
default: hello
outputs:
output:
type: string
reference: ${nodeC.output}
nodes:
- name: nodeA
type: python
source:
type: code
path: pass_through.py
inputs:
input1: ${inputs.text}
activate:
when: ${inputs.text}
is: hi
- name: nodeB
type: python
... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/activate_condition_always_met/summary_result.py | from promptflow import tool
@tool
def summary_result(input1: str="Node A not executed.", input2: str="Node B not executed.") -> str:
return input1 + ' ' + input2
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/python_tool_with_simple_image_without_default/pick_an_image.py | import random
from promptflow.contracts.multimedia import Image
from promptflow import tool
@tool
def pick_an_image(image_1: Image, image_2: Image) -> Image:
if random.choice([True, False]):
return image_1
else:
return image_2
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/python_tool_with_simple_image_without_default/flow.dag.yaml | inputs:
image_1:
type: image
image_2:
type: image
outputs:
output:
type: image
reference: ${python_node.output}
nodes:
- name: python_node
type: python
source:
type: code
path: pick_an_image.py
inputs:
image_1: ${inputs.image_1}
image_2: ${inputs.image_2}
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/python_tool_with_invalid_default_value/pick_an_image.py | import random
from promptflow.contracts.multimedia import Image
from promptflow import tool
@tool
def pick_an_image(image_1: Image, image_2: Image) -> Image:
if random.choice([True, False]):
return image_1
else:
return image_2
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/python_tool_with_invalid_default_value/flow.dag.yaml | inputs:
image:
type: image
default: ""
outputs:
output:
type: image
reference: ${python_node_2.output}
nodes:
- name: python_node
type: python
source:
type: code
path: pick_an_image.py
inputs:
image_1: ${inputs.image}
image_2: logo_2.png
- name: python_node_2
type: python
s... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/web_classification_with_invalid_additional_include/flow.dag.yaml | inputs:
url:
type: string
default: https://www.microsoft.com/en-us/d/xbox-wireless-controller-stellar-shift-special-edition/94fbjc7h0h6h
outputs:
category:
type: string
reference: ${convert_to_dict.output.category}
evidence:
type: string
reference: ${convert_to_dict.output.evidence}
nodes:... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/activate_with_no_inputs/inputs.json | {
"text": "world"
} | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/activate_with_no_inputs/expected_result.json | [
{
"expected_node_count": 2,
"expected_outputs":{
"text": "hello world"
},
"expected_bypassed_nodes":[]
}
] | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/activate_with_no_inputs/flow.dag.yaml | inputs:
text:
type: string
outputs:
text:
type: string
reference: ${node_a.output}
nodes:
- name: node_a
type: python
source:
type: code
path: node_a.py
inputs:
input1: ${inputs.text}
- name: node_b
type: python
source:
type: code
path: node_b.py
inputs: {}
activate:
... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/activate_with_no_inputs/node_a.py | from promptflow import tool
@tool
def my_python_tool(input1: str) -> str:
return 'hello ' + input1
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/activate_with_no_inputs/node_b.py | from promptflow import tool
@tool
def my_python_tool():
print("Avtivate")
return 'Executing...'
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/tool_with_assistant_definition/echo.py | from promptflow import tool
@tool
def echo(message: str):
"""This tool is used to echo the message back.
:param message: The message to echo.
:type message: str
"""
return message
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/tool_with_assistant_definition/assistant_definition.yaml | model: mock_model
instructions: mock_instructions
tools:
- type: function
tool_type: python
source:
type: code
path: echo.py
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/tool_with_assistant_definition/test_assistant_definition.py | from promptflow import tool
from promptflow.contracts.types import AssistantDefinition
@tool
def test_assistant_definition(message: str, assistant_definition: AssistantDefinition):
assert assistant_definition.model == "mock_model"
assert assistant_definition.instructions == "mock_instructions"
invoker = a... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/tool_with_assistant_definition/flow.dag.yaml | inputs:
message:
type: string
default: Hello World!
outputs:
output:
type: object
reference: ${test_assistant_definition.output}
nodes:
- name: test_assistant_definition
type: python
source:
type: code
path: test_assistant_definition.py
inputs:
message: ${inputs.message}
assist... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/one_line_of_bulktest_timeout/samples.json | [{"idx": 1}, {"idx": 4}, {"idx": 10}] | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/one_line_of_bulktest_timeout/my_python_tool.py | from promptflow import tool
import random
@tool
def my_python_tool(idx: int) -> int:
return idx | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/one_line_of_bulktest_timeout/my_python_tool_with_failed_line.py | from promptflow import tool
import random
import time
@tool
def my_python_tool_with_failed_line(idx: int, mod=5) -> int:
if idx % mod == 0:
while True:
time.sleep(60)
return idx
| 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/one_line_of_bulktest_timeout/samples_all_timeout.json | [{"idx": 5}, {"idx": 5}] | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/one_line_of_bulktest_timeout/expected_status_summary.json | {
"__pf__.nodes.my_python_tool.completed": 3,
"__pf__.nodes.my_python_tool_with_failed_line.completed": 2,
"__pf__.nodes.my_python_tool_with_failed_line.failed": 1,
"__pf__.lines.completed": 2,
"__pf__.lines.failed": 1
} | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/one_line_of_bulktest_timeout/flow.dag.yaml | inputs:
idx:
type: int
outputs:
output:
type: int
reference: ${my_python_tool_with_failed_line.output}
nodes:
- name: my_python_tool
type: python
source:
type: code
path: my_python_tool.py
inputs:
idx: ${inputs.idx}
- name: my_python_tool_with_failed_line
type: python
source:
t... | 0 |
promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows | promptflow_repo/promptflow/src/promptflow/tests/test_configs/flows/all_depedencies_bypassed_with_activate_met/inputs.json | {
"text": "hi"
} | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.