task_id stringlengths 11 16 | level stringclasses 4 values | prompt stringlengths 17 253 | available_tools listlengths 4 8 | expected_trace stringlengths 173 2.74k | expected_final_answer stringlengths 18 1.74k | num_steps int32 1 6 | num_tools_offered int32 4 8 | category stringclasses 9 values | pattern stringlengths 0 29 |
|---|---|---|---|---|---|---|---|---|---|
L0_node_0001 | L0_node | Can you add a specified number of days, hours, and/or minutes to a date. Returns the resulting date in ISO format? The parameters are date: 2025-02-14, minutes: 87. | [
"add_duration",
"case_convert",
"retrieve_memory",
"data_filter"
] | {"steps": [{"step_id": "step_1", "tool_name": "add_duration", "arguments": {"date": "2025-02-14", "minutes": 87}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"result_date": "2025-02-14T01:27:00Z"} | 1 | 4 | external_services | |
L0_node_0002 | L0_node | Can you add a specified number of days, hours, and/or minutes to a date. Returns the resulting date in ISO format? The parameters are date: 2026-03-01, hours: 72. | [
"add_duration",
"json_extract",
"format_number",
"retrieve_memory"
] | {"steps": [{"step_id": "step_1", "tool_name": "add_duration", "arguments": {"date": "2026-03-01", "hours": 72}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"result_date": "2026-03-04T00:00:00Z"} | 1 | 4 | external_services | |
L0_node_0003 | L0_node | Please encode a plain-text string to its Base64 representation. | [
"base64_encode",
"knowledge_base_query",
"business_days_between",
"read_file"
] | {"steps": [{"step_id": "step_1", "tool_name": "base64_encode", "arguments": {"text": "The food industry is experiencing disruption from plant-based alternatives and lab-grown meat. Consumer preferences are shifting rapidly."}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"encoded": "VGhlIGZvb2QgaW5kdXN0cnkgaXMgZXhwZXJpZW5jaW5nIGRpc3J1cHRpb24gZnJvbSBwbGFudC1iYXNlZCBhbHRlcm5hdGl2ZXMgYW5kIGxhYi1ncm93biBtZWF0LiBDb25zdW1lciBwcmVmZXJlbmNlcyBhcmUgc2hpZnRpbmcgcmFwaWRseS4="} | 1 | 4 | text_processing | |
L0_node_0004 | L0_node | Please encode a plain-text string to its Base64 representation. | [
"base64_encode",
"standard_deviation",
"percentile",
"compare_texts"
] | {"steps": [{"step_id": "step_1", "tool_name": "base64_encode", "arguments": {"text": "Quantum computing promises to revolutionize how we process information. Unlike classical bits, quantum bits can exist in multiple states simultaneously."}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"encoded": "UXVhbnR1bSBjb21wdXRpbmcgcHJvbWlzZXMgdG8gcmV2b2x1dGlvbml6ZSBob3cgd2UgcHJvY2VzcyBpbmZvcm1hdGlvbi4gVW5saWtlIGNsYXNzaWNhbCBiaXRzLCBxdWFudHVtIGJpdHMgY2FuIGV4aXN0IGluIG11bHRpcGxlIHN0YXRlcyBzaW11bHRhbmVvdXNseS4="} | 1 | 4 | text_processing | |
L0_node_0005 | L0_node | Please count the number of business days (Monday–Friday) between two dates, exclusive of the start date and inclusive of the end date. Also returns the total calendar days — specifically, start date: 2026-02-01, end date: 2026-02-14. | [
"business_days_between",
"list_memories",
"round_number",
"regex_match"
] | {"steps": [{"step_id": "step_1", "tool_name": "business_days_between", "arguments": {"start_date": "2026-02-01", "end_date": "2026-02-14"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"business_days": 10, "total_days": 13} | 1 | 4 | external_services | |
L0_node_0006 | L0_node | Please calculate the difference between two dates in days, weeks, months, or years — specifically, date1: 2025-11-15, date2: 2026-02-01. | [
"calculate_date_diff",
"string_replace",
"clamp_value",
"get_weather"
] | {"steps": [{"step_id": "step_1", "tool_name": "calculate_date_diff", "arguments": {"date1": "2025-11-15", "date2": "2026-02-01"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"date1": "2025-11-15", "date2": "2026-02-01", "days": 78, "weeks": 11.1, "months": 2.6, "years": 0.21, "date1_is_before": true} | 1 | 4 | time_scheduling | |
L0_node_0007 | L0_node | I need you to calculate the difference between two dates in days, weeks, months, or years. Here are the details: date1: 2026-04-01, date2: 2026-02-14. | [
"calculate_date_diff",
"compare_texts",
"string_replace",
"get_exchange_rate"
] | {"steps": [{"step_id": "step_1", "tool_name": "calculate_date_diff", "arguments": {"date1": "2026-04-01", "date2": "2026-02-14"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"date1": "2026-04-01", "date2": "2026-02-14", "days": 46, "weeks": 6.6, "months": 1.5, "years": 0.13, "date1_is_before": false} | 1 | 4 | time_scheduling | |
L0_node_0008 | L0_node | Compute the value of 177 - 61. | [
"calculator",
"string_replace",
"validate_email",
"get_session_context"
] | {"steps": [{"step_id": "step_1", "tool_name": "calculator", "arguments": {"expression": "177 - 61"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"expression": "177 - 61", "result": 116.0} | 1 | 4 | computation | |
L0_node_0009 | L0_node | What is 234 - 89? | [
"calculator",
"transform_format",
"schedule_meeting",
"extract_domain"
] | {"steps": [{"step_id": "step_1", "tool_name": "calculator", "arguments": {"expression": "234 - 89"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"expression": "234 - 89", "result": 145.0} | 1 | 4 | computation | |
L0_node_0010 | L0_node | I need you to convert text to a different case: upper, lower, title, camelCase, or snake_case. Here are the details: target case: lower. | [
"case_convert",
"get_session_context",
"slugify",
"moving_average"
] | {"steps": [{"step_id": "step_1", "tool_name": "case_convert", "arguments": {"text": "The global economy is undergoing a period of significant transformation. Digital technologies are reshaping traditional business models.", "target_case": "lower"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"converted": "the global economy is undergoing a period of significant transformation. digital technologies are reshaping traditional business models."} | 1 | 4 | text_processing | |
L0_node_0011 | L0_node | I need you to convert text to a different case: upper, lower, title, camelCase, or snake_case. Here are the details: target case: camel. | [
"case_convert",
"next_occurrence",
"database_query",
"time_since"
] | {"steps": [{"step_id": "step_1", "tool_name": "case_convert", "arguments": {"text": "Fusion energy research has achieved significant milestones. Commercial fusion power could provide virtually unlimited clean energy.", "target_case": "camel"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"converted": "fusionEnergyResearchHasAchievedSignificantMilestones.CommercialFusionPowerCouldProvideVirtuallyUnlimitedCleanEnergy."} | 1 | 4 | text_processing | |
L0_node_0012 | L0_node | Can you check whether a URL is reachable and return its HTTP status code and estimated response time? The parameters are url: https://news.com. | [
"check_url_status",
"case_convert",
"extract_entities",
"text_to_number"
] | {"steps": [{"step_id": "step_1", "tool_name": "check_url_status", "arguments": {"url": "https://news.com"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"url": "https://news.com", "status": "reachable", "status_code": 200, "response_time_ms": 485} | 1 | 4 | information_retrieval | |
L0_node_0013 | L0_node | Check whether a URL is reachable and return its HTTP status code and estimated response time with the following: url: https://blog.dev. | [
"check_url_status",
"generate_report",
"data_filter",
"http_request"
] | {"steps": [{"step_id": "step_1", "tool_name": "check_url_status", "arguments": {"url": "https://blog.dev"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"url": "https://blog.dev", "status": "client_error", "status_code": 403, "response_time_ms": 200} | 1 | 4 | information_retrieval | |
L0_node_0014 | L0_node | Clamp a number so it falls within the specified minimum and maximum range. Returns the clamped value and whether clamping was applied with the following: value: 86.25, min val: 16.13, max val: 16.84. | [
"clamp_value",
"word_count",
"number_to_text",
"compare_texts"
] | {"steps": [{"step_id": "step_1", "tool_name": "clamp_value", "arguments": {"value": 86.25, "min_val": 16.13, "max_val": 16.84}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"clamped": 16.84, "was_clamped": true} | 1 | 4 | computation | |
L0_node_0015 | L0_node | Please classify a piece of text into one or more predefined categories based on its content. | [
"classify_text",
"word_count",
"compress_data",
"parse_html"
] | {"steps": [{"step_id": "step_1", "tool_name": "classify_text", "arguments": {"text": "Microplastics have been found in every environment on Earth. Research into their health effects is intensifying.", "categories": ["item_0", "item_1"]}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"text_preview": "microplastics have been found in every environment on earth. research into their health effects is i...", "predicted_category": "item_0", "confidence": 0.0, "all_scores": {"item_0": 0.0, "item_1": 0.0}} | 1 | 4 | text_processing | |
L0_node_0016 | L0_node | Classify a piece of text into one or more predefined categories based on its content. | [
"classify_text",
"number_to_text",
"get_weather",
"extract_numbers"
] | {"steps": [{"step_id": "step_1", "tool_name": "classify_text", "arguments": {"text": "Quantum computing promises to revolutionize how we process information. Unlike classical bits, quantum bits can exist in multiple states simultaneously.", "categories": ["item_0", "item_1", "item_2", "item_3"]}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"text_preview": "quantum computing promises to revolutionize how we process information. unlike classical bits, quant...", "predicted_category": "item_0", "confidence": 0.0, "all_scores": {"item_0": 0.0, "item_1": 0.0, "item_2": 0.0, "item_3": 0.0}} | 1 | 4 | text_processing | |
L0_node_0017 | L0_node | I need you to compare two pieces of text and return their similarity score, common keywords, and differences. | [
"compare_texts",
"add_duration",
"regex_match",
"round_number"
] | {"steps": [{"step_id": "step_1", "tool_name": "compare_texts", "arguments": {"text1": "Renewable energy adoption is accelerating globally. Solar and wind power costs have decreased dramatically over the past decade.", "text2": "Biodegradable materials are replacing traditional plastics in packaging. Consumer demand is driving innovation in sustainable materials."}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"similarity_score": 0.032, "common_words_count": 1, "text1_unique_words": 17, "text2_unique_words": 13, "common_keywords": ["is"], "text1_length": 128, "text2_length": 136} | 1 | 4 | text_processing | |
L0_node_0018 | L0_node | Compare two pieces of text and return their similarity score, common keywords, and differences. | [
"compare_texts",
"normalize_data",
"create_invoice",
"create_notification"
] | {"steps": [{"step_id": "step_1", "tool_name": "compare_texts", "arguments": {"text1": "Biodiversity loss is accelerating at an unprecedented rate. Conservation efforts must be scaled up significantly to protect endangered species.", "text2": "The metaverse concept is evolving from gaming to business applications. Virtual reality meetings and digital workspaces are becoming mainstream."}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"similarity_score": 0.056, "common_words_count": 2, "text1_unique_words": 17, "text2_unique_words": 17, "common_keywords": ["is", "to"], "text1_length": 143, "text2_length": 144} | 1 | 4 | text_processing | |
L0_node_0019 | L0_node | Compress a data string using a specified compression algorithm. Returns the compressed size, original size, and compression ratio with the following: data: sample_data_246, algorithm: gzip. | [
"compress_data",
"transform_format",
"summarize_text",
"data_aggregate"
] | {"steps": [{"step_id": "step_1", "tool_name": "compress_data", "arguments": {"data": "sample_data_246", "algorithm": "gzip"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"compressed_size": 23, "original_size": 15, "ratio": 1.5333, "algorithm": "gzip"} | 1 | 4 | communication | |
L0_node_0020 | L0_node | Can you compress a data string using a specified compression algorithm. Returns the compressed size, original size, and compression ratio? The parameters are data: sample_data_779, algorithm: lz4. | [
"compress_data",
"check_url_status",
"parse_date",
"base64_encode"
] | {"steps": [{"step_id": "step_1", "tool_name": "compress_data", "arguments": {"data": "sample_data_779", "algorithm": "lz4"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"compressed_size": 23, "original_size": 15, "ratio": 1.5333, "algorithm": "lz4"} | 1 | 4 | communication | |
L0_node_0021 | L0_node | Create a calendar event with a title, date, duration, and optional list of attendees. Returns a confirmation with the event ID with the following: title: Retrospective, date: 2026-02-14, duration minutes: 91. | [
"create_calendar_event",
"data_sort",
"schedule_meeting",
"correlation"
] | {"steps": [{"step_id": "step_1", "tool_name": "create_calendar_event", "arguments": {"title": "Retrospective", "date": "2026-02-14", "duration_minutes": 91, "attendees": ["item_0", "item_1", "item_2"]}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"event_id": "evt_2c4e30a2bd53", "title": "Retrospective", "date": "2026-02-14", "duration_minutes": 91, "attendees": ["item_0", "item_1", "item_2"], "confirmation": "Calendar event 'Retrospective' created for 2026-02-14."} | 1 | 4 | communication | |
L0_node_0022 | L0_node | Can you create a calendar event with a title, date, duration, and optional list of attendees. Returns a confirmation with the event ID? The parameters are title: Marketing Sync, date: 2026-01-01, duration minutes: 93. | [
"create_calendar_event",
"merge_data",
"ip_geolocation",
"create_contact"
] | {"steps": [{"step_id": "step_1", "tool_name": "create_calendar_event", "arguments": {"title": "Marketing Sync", "date": "2026-01-01", "duration_minutes": 93, "attendees": ["item_0", "item_1", "item_2", "item_3", "item_4"]}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"event_id": "evt_7f9fe9e7a5d3", "title": "Marketing Sync", "date": "2026-01-01", "duration_minutes": 93, "attendees": ["item_0", "item_1", "item_2", "item_3", "item_4"], "confirmation": "Calendar event 'Marketing Sync' created for 2026-01-01."} | 1 | 4 | communication | |
L0_node_0023 | L0_node | Can you create a new contact entry with a name, email address, and optional phone number? The parameters are name: sample_name_675, email: nick@marketing.brand, phone: sample_phone_670. | [
"create_contact",
"url_parse",
"format_date",
"statistical_analysis"
] | {"steps": [{"step_id": "step_1", "tool_name": "create_contact", "arguments": {"name": "sample_name_675", "email": "nick@marketing.brand", "phone": "sample_phone_670"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"contact_id": "con_6e199e2e4173", "name": "sample_name_675", "email": "nick@marketing.brand", "phone": "sample_phone_670", "confirmation": "Contact 'sample_name_675' created."} | 1 | 4 | communication | |
L0_node_0024 | L0_node | I need you to create a new contact entry with a name, email address, and optional phone number. Here are the details: name: sample_name_255, email: olivia@sales.deal. | [
"create_contact",
"get_location_info",
"knowledge_base_query",
"create_calendar_event"
] | {"steps": [{"step_id": "step_1", "tool_name": "create_contact", "arguments": {"name": "sample_name_255", "email": "olivia@sales.deal"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"contact_id": "con_259c4c78c73c", "name": "sample_name_255", "email": "olivia@sales.deal", "phone": null, "confirmation": "Contact 'sample_name_255' created."} | 1 | 4 | communication | |
L0_node_0025 | L0_node | Can you create a notification or alert with a title and message, optionally with a priority level? The parameters are title: Cross-Team Alignment, message: CI/CD pipeline completed in 3m 42s. | [
"create_notification",
"spell_check",
"web_page_fetch",
"ip_geolocation"
] | {"steps": [{"step_id": "step_1", "tool_name": "create_notification", "arguments": {"title": "Cross-Team Alignment", "message": "CI/CD pipeline completed in 3m 42s"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"status": "created", "notification_id": "notif_25a181a23638", "title": "Cross-Team Alignment", "priority": "normal", "timestamp": "2026-02-22T12:00:00"} | 1 | 4 | communication | |
L0_node_0026 | L0_node | Create a notification or alert with a title and message, optionally with a priority level with the following: title: Onboarding Orientation, message: Automated backup verification passed, priority: normal. | [
"create_notification",
"round_number",
"write_file",
"data_sort"
] | {"steps": [{"step_id": "step_1", "tool_name": "create_notification", "arguments": {"title": "Onboarding Orientation", "message": "Automated backup verification passed", "priority": "normal"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"status": "created", "notification_id": "notif_13f776b4ab23", "title": "Onboarding Orientation", "priority": "normal", "timestamp": "2026-02-22T12:00:00"} | 1 | 4 | communication | |
L0_node_0027 | L0_node | I need you to create a spreadsheet with a title, column headers, and data rows. Returns a spreadsheet ID and row count. Here are the details: title: Budget Review. | [
"create_spreadsheet",
"statistical_analysis",
"log_event",
"data_aggregate"
] | {"steps": [{"step_id": "step_1", "tool_name": "create_spreadsheet", "arguments": {"title": "Budget Review", "headers": ["item_0", "item_1", "item_2", "item_3", "item_4"], "rows": ["item_0", "item_1", "item_2"]}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"spreadsheet_id": "sht_757ab09676fa", "title": "Budget Review", "columns": 5, "row_count": 3, "headers": ["item_0", "item_1", "item_2", "item_3", "item_4"], "confirmation": "Spreadsheet 'Budget Review' created with 3 rows and 5 columns."} | 1 | 4 | communication | |
L0_node_0028 | L0_node | Create a spreadsheet with a title, column headers, and data rows. Returns a spreadsheet ID and row count with the following: title: Marketing Sync. | [
"create_spreadsheet",
"get_weekday",
"summarize_text",
"merge_data"
] | {"steps": [{"step_id": "step_1", "tool_name": "create_spreadsheet", "arguments": {"title": "Marketing Sync", "headers": ["item_0", "item_1", "item_2"], "rows": ["item_0", "item_1", "item_2", "item_3", "item_4"]}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"spreadsheet_id": "sht_c38040eb64a3", "title": "Marketing Sync", "columns": 3, "row_count": 5, "headers": ["item_0", "item_1", "item_2"], "confirmation": "Spreadsheet 'Marketing Sync' created with 5 rows and 3 columns."} | 1 | 4 | communication | |
L0_node_0029 | L0_node | Please create a task or to-do item with a title, description, optional due date, and priority — specifically, title: Marketing Sync, description: sample_description_210, due date: 2026-05-01. | [
"create_task",
"url_parse",
"min_max",
"create_calendar_event"
] | {"steps": [{"step_id": "step_1", "tool_name": "create_task", "arguments": {"title": "Marketing Sync", "description": "sample_description_210", "due_date": "2026-05-01"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"status": "created", "task_id": "task_dcbecb0e6809", "title": "Marketing Sync", "description": "sample_description_210", "due_date": "2026-05-01", "priority": "medium", "created_at": "2026-02-22T12:00:00"} | 1 | 4 | communication | |
L0_node_0030 | L0_node | Please create a task or to-do item with a title, description, optional due date, and priority — specifically, title: Vendor Evaluation, description: sample_description_551. | [
"create_task",
"standard_deviation",
"execute_python",
"web_search"
] | {"steps": [{"step_id": "step_1", "tool_name": "create_task", "arguments": {"title": "Vendor Evaluation", "description": "sample_description_551"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"status": "created", "task_id": "task_69df17868447", "title": "Vendor Evaluation", "description": "sample_description_551", "due_date": null, "priority": "medium", "created_at": "2026-02-22T12:00:00"} | 1 | 4 | communication | |
L0_node_0031 | L0_node | Aggregate a list of records by a group key, computing sum, average, count, min, or max for a specified value field with the following: group by: sample_group_by_721, value field: sample_value_field_542. | [
"data_aggregate",
"generate_url",
"list_memories",
"min_max"
] | {"steps": [{"step_id": "step_1", "tool_name": "data_aggregate", "arguments": {"items": ["item_0", "item_1", "item_2", "item_3"], "group_by": "sample_group_by_721", "value_field": "sample_value_field_542"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"operation": "sum", "group_by": "sample_group_by_721", "value_field": "sample_value_field_542", "groups": {}, "total_records": 4, "total_groups": 0} | 1 | 4 | computation | |
L0_node_0032 | L0_node | Please aggregate a list of records by a group key, computing sum, average, count, min, or max for a specified value field — specifically, group by: sample_group_by_761, value field: sample_value_field_289, operation: count. | [
"data_aggregate",
"read_file",
"word_count",
"string_replace"
] | {"steps": [{"step_id": "step_1", "tool_name": "data_aggregate", "arguments": {"items": ["item_0", "item_1", "item_2"], "group_by": "sample_group_by_761", "value_field": "sample_value_field_289", "operation": "count"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"operation": "count", "group_by": "sample_group_by_761", "value_field": "sample_value_field_289", "groups": {}, "total_records": 3, "total_groups": 0} | 1 | 4 | computation | |
L0_node_0033 | L0_node | Filter a list of items based on a condition. For numbers: greater_than, less_than, equals. For strings: contains, starts_with, ends_with with the following: condition: equals, value: sample_value_650. | [
"data_filter",
"number_to_text",
"lookup_entity",
"calculator"
] | {"steps": [{"step_id": "step_1", "tool_name": "data_filter", "arguments": {"items": ["item_0", "item_1", "item_2", "item_3", "item_4"], "condition": "equals", "value": "sample_value_650"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"filtered": [], "count": 0, "original_count": 5, "condition": "equals", "value": "sample_value_650"} | 1 | 4 | computation | |
L0_node_0034 | L0_node | Can you filter a list of items based on a condition. For numbers: greater_than, less_than, equals. For strings: contains, starts_with, ends_with? The parameters are condition: ends_with, value: sample_value_166. | [
"data_filter",
"case_convert",
"mask_pii",
"list_memories"
] | {"steps": [{"step_id": "step_1", "tool_name": "data_filter", "arguments": {"items": ["item_0", "item_1", "item_2", "item_3"], "condition": "ends_with", "value": "sample_value_166"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"filtered": [], "count": 0, "original_count": 4, "condition": "ends_with", "value": "sample_value_166"} | 1 | 4 | computation | |
L0_node_0035 | L0_node | I need you to sort a list of items (numbers or strings) in ascending or descending order. Here are the details: key: project_notes. | [
"data_sort",
"transcribe_audio",
"sentiment_analysis",
"list_memories"
] | {"steps": [{"step_id": "step_1", "tool_name": "data_sort", "arguments": {"items": ["item_0", "item_1"], "key": "project_notes"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"sorted": ["item_0", "item_1"], "order": "ascending", "count": 2} | 1 | 4 | computation | |
L0_node_0036 | L0_node | Sort a list of items (numbers or strings) in ascending or descending order with the following: key: recipe_collection. | [
"data_sort",
"word_count",
"web_search",
"base64_decode"
] | {"steps": [{"step_id": "step_1", "tool_name": "data_sort", "arguments": {"items": ["item_0", "item_1", "item_2", "item_3", "item_4"], "key": "recipe_collection"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"sorted": ["item_0", "item_1", "item_2", "item_3", "item_4"], "order": "ascending", "count": 5} | 1 | 4 | computation | |
L0_node_0037 | L0_node | Please query a structured database table. Supports filtering, sorting, and aggregation on predefined datasets (countries, cities, movies, books) — specifically, table: movies, filter field: sample_filter_field_768, filter value: sample_filter_value_903. | [
"database_query",
"case_convert",
"extract_domain",
"merge_data"
] | {"steps": [{"step_id": "step_1", "tool_name": "database_query", "arguments": {"table": "movies", "filter_field": "sample_filter_field_768", "filter_value": "sample_filter_value_903", "sort_by": "sample_sort_by_802", "limit": 86}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"table": "movies", "results": [], "count": 0, "error": "Table not found"} | 1 | 4 | information_retrieval | |
L0_node_0038 | L0_node | Please query a structured database table. Supports filtering, sorting, and aggregation on predefined datasets (countries, cities, movies, books) — specifically, table: cities, filter field: sample_filter_field_880, filter op: greater_than. | [
"database_query",
"linear_regression",
"lookup_entity",
"data_sort"
] | {"steps": [{"step_id": "step_1", "tool_name": "database_query", "arguments": {"table": "cities", "filter_field": "sample_filter_field_880", "filter_op": "greater_than", "filter_value": "sample_filter_value_841", "sort_by": "sample_sort_by_676"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"table": "cities", "results": [], "count": 0} | 1 | 4 | information_retrieval | |
L0_node_0039 | L0_node | Detect the language of the input text. Uses common function-word heuristics to identify English, French, German, Spanish, Italian, Portuguese, and Dutch. | [
"detect_language",
"hash_text",
"store_memory",
"correlation"
] | {"steps": [{"step_id": "step_1", "tool_name": "detect_language", "arguments": {"text": "The education sector is embracing technology-enhanced learning. Online platforms and AI tutors are making education more accessible."}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"language": "English", "language_code": "en", "confidence": 0.8} | 1 | 4 | text_processing | |
L0_node_0040 | L0_node | Can you detect the language of the input text. Uses common function-word heuristics to identify English, French, German, Spanish, Italian, Portuguese, and Dutch? | [
"detect_language",
"database_query",
"schedule_meeting",
"add_duration"
] | {"steps": [{"step_id": "step_1", "tool_name": "detect_language", "arguments": {"text": "Gene editing technologies like CRISPR offer enormous potential for treating genetic diseases. Ethical considerations remain important."}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"language": "English", "language_code": "en", "confidence": 1.0} | 1 | 4 | text_processing | |
L0_node_0041 | L0_node | Can you perform a DNS lookup for a domain and return the resolved IP address and record type? The parameters are domain: sample_domain_447. | [
"dns_lookup",
"compare_texts",
"log_event",
"is_business_day"
] | {"steps": [{"step_id": "step_1", "tool_name": "dns_lookup", "arguments": {"domain": "sample_domain_447"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"domain": "sample_domain_447", "ip_address": "117.137.50.116", "record_type": "A", "ttl_seconds": 3424} | 1 | 4 | information_retrieval | |
L0_node_0042 | L0_node | Perform a DNS lookup for a domain and return the resolved IP address and record type with the following: domain: sample_domain_261. | [
"dns_lookup",
"calculator",
"format_number",
"number_to_text"
] | {"steps": [{"step_id": "step_1", "tool_name": "dns_lookup", "arguments": {"domain": "sample_domain_261"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"domain": "sample_domain_261", "ip_address": "77.12.18.40", "record_type": "A", "ttl_seconds": 2784} | 1 | 4 | information_retrieval | |
L0_node_0043 | L0_node | Please uRL encode or decode a text string. Encodes special characters for safe use in URLs, or decodes percent-encoded strings back to readable text — specifically, action: encode. | [
"encode_url",
"clamp_value",
"number_to_text",
"data_sort"
] | {"steps": [{"step_id": "step_1", "tool_name": "encode_url", "arguments": {"text": "Neuroscience breakthroughs are improving our understanding of brain function. Brain-computer interfaces could transform healthcare.", "action": "encode"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"result": "Neuroscience%20breakthroughs%20are%20improving%20our%20understanding%20of%20brain%20function.%20Brain-computer%20interfaces%20could%20transform%20healthcare."} | 1 | 4 | information_retrieval | |
L0_node_0044 | L0_node | Please uRL encode or decode a text string. Encodes special characters for safe use in URLs, or decodes percent-encoded strings back to readable text — specifically, action: encode. | [
"encode_url",
"truncate_text",
"round_number",
"send_webhook"
] | {"steps": [{"step_id": "step_1", "tool_name": "encode_url", "arguments": {"text": "The Internet of Things connects billions of devices worldwide. Smart home technology is becoming standard in new construction.", "action": "encode"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"result": "The%20Internet%20of%20Things%20connects%20billions%20of%20devices%20worldwide.%20Smart%20home%20technology%20is%20becoming%20standard%20in%20new%20construction."} | 1 | 4 | information_retrieval | |
L0_node_0045 | L0_node | Can you encrypt a text string using a specified encryption method. Returns the simulated encrypted output as a base64 string? The parameters are method: rsa. | [
"encrypt_text",
"list_memories",
"parse_date",
"translate_text"
] | {"steps": [{"step_id": "step_1", "tool_name": "encrypt_text", "arguments": {"text": "The education sector is embracing technology-enhanced learning. Online platforms and AI tutors are making education more accessible.", "method": "rsa"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"encrypted": "sBIJK3hty3oxA54csWwdY9xTiUBfjTxcc1ypiAaYwfk=", "method": "rsa", "original_length": 132, "encrypted_length": 44} | 1 | 4 | communication | |
L0_node_0046 | L0_node | Can you encrypt a text string using a specified encryption method. Returns the simulated encrypted output as a base64 string? The parameters are method: aes256. | [
"encrypt_text",
"convert_timezone",
"summarize_text",
"round_number"
] | {"steps": [{"step_id": "step_1", "tool_name": "encrypt_text", "arguments": {"text": "Neuroscience breakthroughs are improving our understanding of brain function. Brain-computer interfaces could transform healthcare.", "method": "aes256"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"encrypted": "nrRSS4U2c37qWog1HIA1B3UV/LwKKukaaUz4xAtdLzk=", "method": "aes256", "original_length": 131, "encrypted_length": 44} | 1 | 4 | communication | |
L0_node_0047 | L0_node | Can you extract the domain, subdomain, and top-level domain (TLD) from a given URL? The parameters are url: https://news.com. | [
"extract_domain",
"generate_summary_stats",
"base64_decode",
"read_file"
] | {"steps": [{"step_id": "step_1", "tool_name": "extract_domain", "arguments": {"url": "https://news.com"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"domain": "news", "subdomain": "", "tld": "com"} | 1 | 4 | information_retrieval | |
L0_node_0048 | L0_node | Please extract the domain, subdomain, and top-level domain (TLD) from a given URL — specifically, url: https://blog.dev. | [
"extract_domain",
"data_aggregate",
"lookup_entity",
"log_event"
] | {"steps": [{"step_id": "step_1", "tool_name": "extract_domain", "arguments": {"url": "https://blog.dev"}, "depends_on": [], "output_key": null}], "final_answer_source": "step_1"} | {"domain": "blog", "subdomain": "", "tld": "dev"} | 1 | 4 | information_retrieval | |
L1_chain_0049 | L1_chain | Check the weather in Berlin and convert the temperature to Fahrenheit. | [
"get_weather",
"unit_convert",
"word_count",
"normalize_data",
"transform_format"
] | {"steps": [{"step_id": "step_1", "tool_name": "get_weather", "arguments": {"city": "Berlin"}, "depends_on": [], "output_key": "weather"}, {"step_id": "step_2", "tool_name": "unit_convert", "arguments": {"value": 36, "from_unit": "celsius", "to_unit": "fahrenheit"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"original_value": 36.0, "from_unit": "celsius", "converted_value": 96.8, "to_unit": "fahrenheit"} | 2 | 5 | chain | retrieve-transform |
L1_chain_0050 | L1_chain | Check the weather in Cairo and convert the temperature to Fahrenheit. | [
"get_weather",
"unit_convert",
"create_task",
"calculator",
"database_query"
] | {"steps": [{"step_id": "step_1", "tool_name": "get_weather", "arguments": {"city": "Cairo"}, "depends_on": [], "output_key": "weather"}, {"step_id": "step_2", "tool_name": "unit_convert", "arguments": {"value": 33, "from_unit": "celsius", "to_unit": "fahrenheit"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"original_value": 33.0, "from_unit": "celsius", "converted_value": 91.4, "to_unit": "fahrenheit"} | 2 | 5 | chain | retrieve-transform |
L1_chain_0051 | L1_chain | What is the temperature in Helsinki in Fahrenheit? | [
"get_weather",
"unit_convert",
"set_reminder",
"number_to_text",
"string_replace"
] | {"steps": [{"step_id": "step_1", "tool_name": "get_weather", "arguments": {"city": "Helsinki"}, "depends_on": [], "output_key": "weather"}, {"step_id": "step_2", "tool_name": "unit_convert", "arguments": {"value": 30, "from_unit": "celsius", "to_unit": "fahrenheit"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"original_value": 30.0, "from_unit": "celsius", "converted_value": 86.0, "to_unit": "fahrenheit"} | 2 | 5 | chain | retrieve-transform |
L1_chain_0052 | L1_chain | Find information about "digital twin technology" and give me a summary. | [
"web_search",
"summarize_text",
"split_text",
"slugify",
"spell_check"
] | {"steps": [{"step_id": "step_1", "tool_name": "web_search", "arguments": {"query": "digital twin technology"}, "depends_on": [], "output_key": "results"}, {"step_id": "step_2", "tool_name": "summarize_text", "arguments": {"text": "[{\"title\": \"Digital Twin Technology \\u2014 Wikipedia Overview\", \"url\": \"https://en.wikipedia.org/article/40290\", \"snippet\": \"A comprehensive overview of digital twin technology covering key concepts, recent developments, and practical applications.\", \"source\": \"Wikipedia\"}, {\"title\": \"Digital Twin Technology \\u2014 Academic Analysis\", \"url\": \"https://sciencedirect.com/article/35383\", \"snippet\": \"Expert guide to digital twin technology with detailed explanations, examples, and best practices for "}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"original_length": 500, "summary": "[{\"title\": \"Digital Twin Technology \\u2014 Wikipedia Overview\", \"url\": \"https://en.wikipedia.org/article/40290\", \"snippet\": \"A comprehensive overview of digital twin technology covering key concepts, recent developments, and practical applications.\", \"source\": \"Wikipedia\"}, {\"title\": \"Digital Twin Technology \\u2014 Academic Analysis\", \"url\": \"https://sciencedirect.com/article/35383\", \"snippet\": \"Expert guide to digital twin technology with detailed explanations, examples, and best practices for", "summary_length": 499, "sentences_used": 1, "compression_ratio": 1.0} | 2 | 5 | chain | retrieve-process |
L1_chain_0053 | L1_chain | Search the web for "autonomous vehicle regulations", then summarize the findings. | [
"web_search",
"summarize_text",
"merge_data",
"next_occurrence",
"base64_encode"
] | {"steps": [{"step_id": "step_1", "tool_name": "web_search", "arguments": {"query": "autonomous vehicle regulations"}, "depends_on": [], "output_key": "results"}, {"step_id": "step_2", "tool_name": "summarize_text", "arguments": {"text": "[{\"title\": \"Autonomous Vehicle Regulations \\u2014 News Overview\", \"url\": \"https://stackoverflow.com/article/41076\", \"snippet\": \"Latest research findings on autonomous vehicle regulations from leading institutions and peer-reviewed publications.\", \"source\": \"News\"}, {\"title\": \"Autonomous Vehicle Regulations \\u2014 News Analysis\", \"url\": \"https://stackoverflow.com/article/93556\", \"snippet\": \"Latest research findings on autonomous vehicle regulations from leading institutions and peer-reviewed publ"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"original_length": 500, "summary": "[{\"title\": \"Autonomous Vehicle Regulations \\u2014 News Overview\", \"url\": \"https://stackoverflow.com/article/41076\", \"snippet\": \"Latest research findings on autonomous vehicle regulations from leading institutions and peer-reviewed publications.\", \"source\": \"News\"}, {\"title\": \"Autonomous Vehicle Regulations \\u2014 News Analysis\", \"url\": \"https://stackoverflow.com/article/93556\", \"snippet\": \"Latest research findings on autonomous vehicle regulations from leading institutions and peer-reviewed publ", "summary_length": 500, "sentences_used": 1, "compression_ratio": 1.0} | 2 | 5 | chain | retrieve-process |
L1_chain_0054 | L1_chain | Search for "water purification methods" and summarize the results. | [
"web_search",
"summarize_text",
"extract_links",
"dns_lookup",
"add_duration"
] | {"steps": [{"step_id": "step_1", "tool_name": "web_search", "arguments": {"query": "water purification methods"}, "depends_on": [], "output_key": "results"}, {"step_id": "step_2", "tool_name": "summarize_text", "arguments": {"text": "[{\"title\": \"Water Purification Methods \\u2014 Blog Overview\", \"url\": \"https://nytimes.com/article/12187\", \"snippet\": \"water purification methods: analysis of current trends, methodologies, and future directions in the field.\", \"source\": \"Blog\"}, {\"title\": \"Water Purification Methods \\u2014 Wikipedia Analysis\", \"url\": \"https://github.com/article/59325\", \"snippet\": \"A comprehensive overview of water purification methods covering key concepts, recent developments, and practical applications.\", \"sou"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"original_length": 500, "summary": "[{\"title\": \"Water Purification Methods \\u2014 Blog Overview\", \"url\": \"https://nytimes.com/article/12187\", \"snippet\": \"water purification methods: analysis of current trends, methodologies, and future directions in the field.\", \"source\": \"Blog\"}, {\"title\": \"Water Purification Methods \\u2014 Wikipedia Analysis\", \"url\": \"https://github.com/article/59325\", \"snippet\": \"A comprehensive overview of water purification methods covering key concepts, recent developments, and practical applications.\", \"sou", "summary_length": 500, "sentences_used": 1, "compression_ratio": 1.0} | 2 | 5 | chain | retrieve-process |
L1_chain_0055 | L1_chain | What is V's stock price in NZD? | [
"get_stock_price",
"get_exchange_rate",
"translate_text",
"extract_domain",
"slugify"
] | {"steps": [{"step_id": "step_1", "tool_name": "get_stock_price", "arguments": {"symbol": "V"}, "depends_on": [], "output_key": "stock"}, {"step_id": "step_2", "tool_name": "get_exchange_rate", "arguments": {"from_currency": "USD", "to_currency": "NZD", "amount": 113}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"from_currency": "USD", "to_currency": "NZD", "rate": 1.0, "amount": 113.0, "converted_amount": 113.0} | 2 | 5 | chain | retrieve-convert |
L1_chain_0056 | L1_chain | What is UBER's stock price in SEK? | [
"get_stock_price",
"get_exchange_rate",
"summarize_text",
"database_query",
"get_weekday"
] | {"steps": [{"step_id": "step_1", "tool_name": "get_stock_price", "arguments": {"symbol": "UBER"}, "depends_on": [], "output_key": "stock"}, {"step_id": "step_2", "tool_name": "get_exchange_rate", "arguments": {"from_currency": "USD", "to_currency": "SEK", "amount": 133}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"from_currency": "USD", "to_currency": "SEK", "rate": 10.42, "amount": 133.0, "converted_amount": 1385.86} | 2 | 5 | chain | retrieve-convert |
L1_chain_0057 | L1_chain | Get the price of SQ and convert it to CAD. | [
"get_stock_price",
"get_exchange_rate",
"hash_text",
"store_memory",
"summarize_text"
] | {"steps": [{"step_id": "step_1", "tool_name": "get_stock_price", "arguments": {"symbol": "SQ"}, "depends_on": [], "output_key": "stock"}, {"step_id": "step_2", "tool_name": "get_exchange_rate", "arguments": {"from_currency": "USD", "to_currency": "CAD", "amount": 539}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"from_currency": "USD", "to_currency": "CAD", "rate": 1.36, "amount": 539.0, "converted_amount": 733.04} | 2 | 5 | chain | retrieve-convert |
L1_chain_0058 | L1_chain | Find named entities in "Rosalind Franklin's X-ray crystallography work was done at King's College London..." and determine the sentiment. | [
"extract_entities",
"sentiment_analysis",
"extract_links",
"generate_summary_stats",
"calculate_date_diff"
] | {"steps": [{"step_id": "step_1", "tool_name": "extract_entities", "arguments": {"text": "Rosalind Franklin's X-ray crystallography work was done at King's College London in 1952."}, "depends_on": [], "output_key": "entities"}, {"step_id": "step_2", "tool_name": "sentiment_analysis", "arguments": {"text": "Rosalind Franklin's X-ray crystallography work was done at King's College London in 1952."}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"text_length": 89, "sentiment": "neutral", "confidence": 0.5, "positive_signals": 0, "negative_signals": 0} | 2 | 5 | chain | analyze-classify |
L1_chain_0059 | L1_chain | Identify entities in "Final exam schedule: email registrar@university.edu or visit room 301 in Admin B...", then classify its sentiment. | [
"extract_entities",
"sentiment_analysis",
"send_webhook",
"text_to_number",
"retrieve_memory"
] | {"steps": [{"step_id": "step_1", "tool_name": "extract_entities", "arguments": {"text": "Final exam schedule: email registrar@university.edu or visit room 301 in Admin Building."}, "depends_on": [], "output_key": "entities"}, {"step_id": "step_2", "tool_name": "sentiment_analysis", "arguments": {"text": "Final exam schedule: email registrar@university.edu or visit room 301 in Admin Building."}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"text_length": 88, "sentiment": "neutral", "confidence": 0.5, "positive_signals": 0, "negative_signals": 0} | 2 | 5 | chain | analyze-classify |
L1_chain_0060 | L1_chain | Extract entities from "Registration closes on 06/30/2026. Contact events@conference.org for more inform..." and analyze the sentiment. | [
"extract_entities",
"sentiment_analysis",
"parse_html",
"set_reminder",
"mask_pii"
] | {"steps": [{"step_id": "step_1", "tool_name": "extract_entities", "arguments": {"text": "Registration closes on 06/30/2026. Contact events@conference.org for more information."}, "depends_on": [], "output_key": "entities"}, {"step_id": "step_2", "tool_name": "sentiment_analysis", "arguments": {"text": "Registration closes on 06/30/2026. Contact events@conference.org for more information."}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"text_length": 86, "sentiment": "neutral", "confidence": 0.5, "positive_signals": 0, "negative_signals": 0} | 2 | 5 | chain | analyze-classify |
L1_chain_0061 | L1_chain | Look up weather in Manila, then email the results to yuki@international.global. | [
"get_weather",
"send_email",
"send_webhook",
"string_replace",
"web_search"
] | {"steps": [{"step_id": "step_1", "tool_name": "get_weather", "arguments": {"city": "Manila"}, "depends_on": [], "output_key": "weather"}, {"step_id": "step_2", "tool_name": "send_email", "arguments": {"to": "yuki@international.global", "subject": "Weather in Manila", "body": "{\"city\": \"Manila\", \"temperature_celsius\": 26, \"humidity_percent\": 31, \"wind_speed_kmh\": 41, \"condition\": \"foggy\", \"feels_like_celsius\": 24}"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"status": "sent", "message_id": "msg_a7ab2625ae69", "to": "yuki@international.global", "subject": "Weather in Manila", "body_preview": "{\"city\": \"Manila\", \"temperature_celsius\": 26, \"humidity_percent\": 31, \"wind_speed_kmh\": 41, \"conditi...", "cc": null, "timestamp": "2026-02-22T12:00:00"} | 2 | 5 | chain | retrieve-send |
L1_chain_0062 | L1_chain | Get the weather in Chicago and email the report to olivia@sales.deal. | [
"get_weather",
"send_email",
"split_text",
"round_number",
"web_search"
] | {"steps": [{"step_id": "step_1", "tool_name": "get_weather", "arguments": {"city": "Chicago"}, "depends_on": [], "output_key": "weather"}, {"step_id": "step_2", "tool_name": "send_email", "arguments": {"to": "olivia@sales.deal", "subject": "Weather in Chicago", "body": "{\"city\": \"Chicago\", \"temperature_celsius\": -5, \"humidity_percent\": 30, \"wind_speed_kmh\": 30, \"condition\": \"snowy\", \"feels_like_celsius\": -7}"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"status": "sent", "message_id": "msg_f032ff081cd7", "to": "olivia@sales.deal", "subject": "Weather in Chicago", "body_preview": "{\"city\": \"Chicago\", \"temperature_celsius\": -5, \"humidity_percent\": 30, \"wind_speed_kmh\": 30, \"condit...", "cc": null, "timestamp": "2026-02-22T12:00:00"} | 2 | 5 | chain | retrieve-send |
L1_chain_0063 | L1_chain | Check Chicago's weather and send it via email to karen@legal.law. | [
"get_weather",
"send_email",
"list_memories",
"data_filter",
"next_occurrence"
] | {"steps": [{"step_id": "step_1", "tool_name": "get_weather", "arguments": {"city": "Chicago"}, "depends_on": [], "output_key": "weather"}, {"step_id": "step_2", "tool_name": "send_email", "arguments": {"to": "karen@legal.law", "subject": "Weather in Chicago", "body": "{\"city\": \"Chicago\", \"temperature_celsius\": -5, \"humidity_percent\": 30, \"wind_speed_kmh\": 30, \"condition\": \"snowy\", \"feels_like_celsius\": -7}"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"status": "sent", "message_id": "msg_4e1a5a4074bf", "to": "karen@legal.law", "subject": "Weather in Chicago", "body_preview": "{\"city\": \"Chicago\", \"temperature_celsius\": -5, \"humidity_percent\": 30, \"wind_speed_kmh\": 30, \"condit...", "cc": null, "timestamp": "2026-02-22T12:00:00"} | 2 | 5 | chain | retrieve-send |
L1_chain_0064 | L1_chain | Read the file at /data/employees.csv and summarize its contents. | [
"read_file",
"summarize_text",
"spell_check",
"format_number",
"business_days_between"
] | {"steps": [{"step_id": "step_1", "tool_name": "read_file", "arguments": {"path": "/data/employees.csv"}, "depends_on": [], "output_key": "content"}, {"step_id": "step_2", "tool_name": "summarize_text", "arguments": {"text": "name,department,salary,years\nAlice,Engineering,95000,5\nBob,Marketing,72000,3\nCharlie,Engineering,105000,8\nDiana,Sales,68000,2\nEve,Engineering,88000,4\nFrank,Marketing,76000,6\nGrace,Sales,71000,3\nHenry,Engineering,112000,10"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"original_length": 221, "summary": "name,department,salary,years\nAlice,Engineering,95000,5\nBob,Marketing,72000,3\nCharlie,Engineering,105000,8\nDiana,Sales,68000,2\nEve,Engineering,88000,4\nFrank,Marketing,76000,6\nGrace,Sales,71000,3\nHenry,Engineering,112000,10", "summary_length": 221, "sentences_used": 1, "compression_ratio": 1.0} | 2 | 5 | chain | read-process |
L1_chain_0065 | L1_chain | Search for "smart city infrastructure" and classify the results into categories. | [
"web_search",
"classify_text",
"text_to_number",
"spell_check",
"dns_lookup"
] | {"steps": [{"step_id": "step_1", "tool_name": "web_search", "arguments": {"query": "smart city infrastructure"}, "depends_on": [], "output_key": "results"}, {"step_id": "step_2", "tool_name": "classify_text", "arguments": {"text": "[{\"title\": \"Smart City Infrastructure \\u2014 News Overview\", \"url\": \"https://stackoverflow.com/article/32326\", \"snippet\": \"Latest research findings on smart city infrastructure from leading institutions and peer-reviewed publications.\", \"source\": \"News\"}, {\"title\": \"Smart City Infrastructure \\u2014 ", "categories": ["science", "technology", "politics"]}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"text_preview": "[{\"title\": \"smart city infrastructure \\u2014 news overview\", \"url\": \"https://stackoverflow.com/artic...", "predicted_category": "science", "confidence": 0.0, "all_scores": {"science": 0.0, "technology": 0.0, "politics": 0.0}} | 2 | 5 | chain | retrieve-classify |
L1_chain_0066 | L1_chain | Search for "digital twin technology" and classify the results into categories. | [
"web_search",
"classify_text",
"check_url_status",
"extract_entities",
"detect_language"
] | {"steps": [{"step_id": "step_1", "tool_name": "web_search", "arguments": {"query": "digital twin technology"}, "depends_on": [], "output_key": "results"}, {"step_id": "step_2", "tool_name": "classify_text", "arguments": {"text": "[{\"title\": \"Digital Twin Technology \\u2014 Wikipedia Overview\", \"url\": \"https://en.wikipedia.org/article/40290\", \"snippet\": \"A comprehensive overview of digital twin technology covering key concepts, recent developments, and practical applications.\", \"source\": \"Wikipedia\"}, {\"title\": \"Digital Twin T", "categories": ["sports", "education", "environment"]}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"text_preview": "[{\"title\": \"digital twin technology \\u2014 wikipedia overview\", \"url\": \"https://en.wikipedia.org/art...", "predicted_category": "sports", "confidence": 0.0, "all_scores": {"sports": 0.0, "education": 0.0, "environment": 0.0}} | 2 | 5 | chain | retrieve-classify |
L1_chain_0067 | L1_chain | Find information about "autonomous vehicle regulations" and categorize it. | [
"web_search",
"classify_text",
"min_max",
"extract_links",
"spell_check"
] | {"steps": [{"step_id": "step_1", "tool_name": "web_search", "arguments": {"query": "autonomous vehicle regulations"}, "depends_on": [], "output_key": "results"}, {"step_id": "step_2", "tool_name": "classify_text", "arguments": {"text": "[{\"title\": \"Autonomous Vehicle Regulations \\u2014 News Overview\", \"url\": \"https://stackoverflow.com/article/41076\", \"snippet\": \"Latest research findings on autonomous vehicle regulations from leading institutions and peer-reviewed publications.\", \"source\": \"News\"}, {\"title\": \"Autonomous Vehicle Regu", "categories": ["science", "technology", "politics"]}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"text_preview": "[{\"title\": \"autonomous vehicle regulations \\u2014 news overview\", \"url\": \"https://stackoverflow.com/...", "predicted_category": "science", "confidence": 0.0, "all_scores": {"science": 0.0, "technology": 0.0, "politics": 0.0}} | 2 | 5 | chain | retrieve-classify |
L1_chain_0068 | L1_chain | Count words in "The Internet of Things connects billions of devices worldwid...", then format the number with commas. | [
"word_count",
"format_number",
"compress_data",
"parse_html",
"http_request"
] | {"steps": [{"step_id": "step_1", "tool_name": "word_count", "arguments": {"text": "The Internet of Things connects billions of devices worldwide. Smart home technology is becoming standard in new construction."}, "depends_on": [], "output_key": "wc"}, {"step_id": "step_2", "tool_name": "format_number", "arguments": {"value": 18, "format": "comma"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"formatted": "18.00"} | 2 | 5 | chain | count-format |
L1_chain_0069 | L1_chain | How many words are in "Urban farming and vertical agriculture are addressing food s..."? Give me the count formatted nicely. | [
"word_count",
"format_number",
"create_invoice",
"write_file",
"web_page_fetch"
] | {"steps": [{"step_id": "step_1", "tool_name": "word_count", "arguments": {"text": "Urban farming and vertical agriculture are addressing food security in cities. These innovations reduce transportation costs and carbon emissions."}, "depends_on": [], "output_key": "wc"}, {"step_id": "step_2", "tool_name": "format_number", "arguments": {"value": 19, "format": "comma"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"formatted": "19.00"} | 2 | 5 | chain | count-format |
L1_chain_0070 | L1_chain | Count words in "Ocean acidification threatens marine ecosystems worldwide. C...", then format the number with commas. | [
"word_count",
"format_number",
"create_task",
"detect_language",
"linear_regression"
] | {"steps": [{"step_id": "step_1", "tool_name": "word_count", "arguments": {"text": "Ocean acidification threatens marine ecosystems worldwide. Coral reefs are particularly vulnerable to changes in water chemistry."}, "depends_on": [], "output_key": "wc"}, {"step_id": "step_2", "tool_name": "format_number", "arguments": {"value": 16, "format": "comma"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"formatted": "16.00"} | 2 | 5 | chain | count-format |
L1_chain_0071 | L1_chain | Pull keywords from "Microplastics have been found in every environment on Earth....", then do a web search for the primary keyword. | [
"keyword_extract",
"web_search",
"join_texts",
"get_exchange_rate",
"create_contact"
] | {"steps": [{"step_id": "step_1", "tool_name": "keyword_extract", "arguments": {"text": "Microplastics have been found in every environment on Earth. Research into their health effects is intensifying."}, "depends_on": [], "output_key": "keywords"}, {"step_id": "step_2", "tool_name": "web_search", "arguments": {"query": "microplastics"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"query": "microplastics", "results": [{"title": "Microplastics — Blog Overview", "url": "https://nytimes.com/article/72137", "snippet": "microplastics: analysis of current trends, methodologies, and future directions in the field.", "source": "Blog"}, {"title": "Microplastics — Blog Analysis", "url": "https://nytimes.com/article/81047", "snippet": "microplastics: analysis of current trends, methodologies, and future directions in the field.", "source": "Blog"}, {"title": "Microplastics — Wikipedia Guide", "url": "https://en.wikipedia.org/article/33320", "snippet": "A comprehensive overview of microplastics covering key concepts, recent developments, and practical applications.", "source": "Wikipedia"}, {"title": "Microplastics — Forum Discussion", "url": "https://reuters.com/article/13829", "snippet": "Comparing different approaches to microplastics — strengths, limitations, and real-world performance.", "source": "Forum"}, {"title": "Microplastics — Blog Report", "url": "https://nature.com/article/12052", "snippet": "microplastics: analysis of current trends, methodologies, and future directions in the field.", "source": "Blog"}], "total": 5} | 2 | 5 | chain | extract-search |
L1_chain_0072 | L1_chain | Find the main keywords in "Biodegradable materials are replacing traditional plastics i..." and look up the most important one online. | [
"keyword_extract",
"web_search",
"regex_match",
"retrieve_memory",
"create_task"
] | {"steps": [{"step_id": "step_1", "tool_name": "keyword_extract", "arguments": {"text": "Biodegradable materials are replacing traditional plastics in packaging. Consumer demand is driving innovation in sustainable materials."}, "depends_on": [], "output_key": "keywords"}, {"step_id": "step_2", "tool_name": "web_search", "arguments": {"query": "materials"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"query": "materials", "results": [{"title": "Materials — News Overview", "url": "https://arxiv.org/article/40731", "snippet": "Latest research findings on materials from leading institutions and peer-reviewed publications.", "source": "News"}, {"title": "Materials — News Analysis", "url": "https://arxiv.org/article/92341", "snippet": "Latest research findings on materials from leading institutions and peer-reviewed publications.", "source": "News"}, {"title": "Materials — Blog Guide", "url": "https://nature.com/article/21492", "snippet": "materials: analysis of current trends, methodologies, and future directions in the field.", "source": "Blog"}, {"title": "Materials — News Discussion", "url": "https://arxiv.org/article/63961", "snippet": "Latest research findings on materials from leading institutions and peer-reviewed publications.", "source": "News"}, {"title": "Materials — Academic Report", "url": "https://bbc.com/article/68498", "snippet": "Expert guide to materials with detailed explanations, examples, and best practices for practitioners.", "source": "Academic"}], "total": 5} | 2 | 5 | chain | extract-search |
L1_chain_0073 | L1_chain | Pull keywords from "Biodegradable materials are replacing traditional plastics i...", then do a web search for the primary keyword. | [
"keyword_extract",
"web_search",
"compare_texts",
"extract_entities",
"base64_decode"
] | {"steps": [{"step_id": "step_1", "tool_name": "keyword_extract", "arguments": {"text": "Biodegradable materials are replacing traditional plastics in packaging. Consumer demand is driving innovation in sustainable materials."}, "depends_on": [], "output_key": "keywords"}, {"step_id": "step_2", "tool_name": "web_search", "arguments": {"query": "materials"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"query": "materials", "results": [{"title": "Materials — News Overview", "url": "https://arxiv.org/article/40731", "snippet": "Latest research findings on materials from leading institutions and peer-reviewed publications.", "source": "News"}, {"title": "Materials — News Analysis", "url": "https://arxiv.org/article/92341", "snippet": "Latest research findings on materials from leading institutions and peer-reviewed publications.", "source": "News"}, {"title": "Materials — Blog Guide", "url": "https://nature.com/article/21492", "snippet": "materials: analysis of current trends, methodologies, and future directions in the field.", "source": "Blog"}, {"title": "Materials — News Discussion", "url": "https://arxiv.org/article/63961", "snippet": "Latest research findings on materials from leading institutions and peer-reviewed publications.", "source": "News"}, {"title": "Materials — Academic Report", "url": "https://bbc.com/article/68498", "snippet": "Expert guide to materials with detailed explanations, examples, and best practices for practitioners.", "source": "Academic"}], "total": 5} | 2 | 5 | chain | extract-search |
L1_chain_0074 | L1_chain | Detect the language of "Bonjour, comment allez-vous aujourd'hui?..." and then translate it to English. | [
"detect_language",
"translate_text",
"json_extract",
"generate_summary_stats",
"time_since"
] | {"steps": [{"step_id": "step_1", "tool_name": "detect_language", "arguments": {"text": "Bonjour, comment allez-vous aujourd'hui?"}, "depends_on": [], "output_key": "detected"}, {"step_id": "step_2", "tool_name": "translate_text", "arguments": {"text": "Bonjour, comment allez-vous aujourd'hui?", "from_language": "und", "to_language": "en"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"original_text": "Bonjour, comment allez-vous aujourd'hui?", "translated_text": "Bonjour, comment allez-vous aujourd'hui? (English translation)", "from_language": "und", "to_language": "en"} | 2 | 5 | chain | detect-translate |
L1_chain_0075 | L1_chain | What language is "Donde esta la biblioteca mas cercana?..." in? Translate it to English. | [
"detect_language",
"translate_text",
"unit_convert",
"web_page_fetch",
"spell_check"
] | {"steps": [{"step_id": "step_1", "tool_name": "detect_language", "arguments": {"text": "Donde esta la biblioteca mas cercana?"}, "depends_on": [], "output_key": "detected"}, {"step_id": "step_2", "tool_name": "translate_text", "arguments": {"text": "Donde esta la biblioteca mas cercana?", "from_language": "fr", "to_language": "en"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"original_text": "Donde esta la biblioteca mas cercana?", "translated_text": "Donde esta la biblioteca mas cercana? (English translation)", "from_language": "fr", "to_language": "en"} | 2 | 5 | chain | detect-translate |
L1_chain_0076 | L1_chain | What language is "Eu gostaria de reservar uma mesa para dois...." in? Translate it to English. | [
"detect_language",
"translate_text",
"set_reminder",
"correlation",
"percentile"
] | {"steps": [{"step_id": "step_1", "tool_name": "detect_language", "arguments": {"text": "Eu gostaria de reservar uma mesa para dois."}, "depends_on": [], "output_key": "detected"}, {"step_id": "step_2", "tool_name": "translate_text", "arguments": {"text": "Eu gostaria de reservar uma mesa para dois.", "from_language": "es", "to_language": "en"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"original_text": "Eu gostaria de reservar uma mesa para dois.", "translated_text": "Eu gostaria de reservar uma mesa para dois. (English translation)", "from_language": "es", "to_language": "en"} | 2 | 5 | chain | detect-translate |
L1_chain_0077 | L1_chain | Find the % change from 383.0 to 195.77 and present it in a formatted style. | [
"percentage_change",
"format_number",
"web_search",
"data_aggregate",
"create_contact"
] | {"steps": [{"step_id": "step_1", "tool_name": "percentage_change", "arguments": {"old_value": 383.0, "new_value": 195.77}, "depends_on": [], "output_key": "pct"}, {"step_id": "step_2", "tool_name": "format_number", "arguments": {"value": -48.8851, "format": "percent"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"formatted": "-4888.51%"} | 2 | 5 | chain | compute-format |
L1_chain_0078 | L1_chain | Compute the percentage difference from 19.5 to 274.32 and display it formatted. | [
"percentage_change",
"format_number",
"write_file",
"json_extract",
"data_sort"
] | {"steps": [{"step_id": "step_1", "tool_name": "percentage_change", "arguments": {"old_value": 19.5, "new_value": 274.32}, "depends_on": [], "output_key": "pct"}, {"step_id": "step_2", "tool_name": "format_number", "arguments": {"value": 1306.7692, "format": "percent"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"formatted": "130676.92%"} | 2 | 5 | chain | compute-format |
L1_chain_0079 | L1_chain | What is the percentage change from 392.8 to 329.36? Format the result as a percentage. | [
"percentage_change",
"format_number",
"merge_data",
"get_directions",
"create_notification"
] | {"steps": [{"step_id": "step_1", "tool_name": "percentage_change", "arguments": {"old_value": 392.8, "new_value": 329.36}, "depends_on": [], "output_key": "pct"}, {"step_id": "step_2", "tool_name": "format_number", "arguments": {"value": -16.1507, "format": "percent"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"formatted": "-1615.07%"} | 2 | 5 | chain | compute-format |
L1_chain_0080 | L1_chain | Find the numbers in "Budget allocated $5.2 million for research, $3.1 million for..." and run a statistical analysis. | [
"extract_numbers",
"statistical_analysis",
"merge_data",
"time_since",
"create_invoice"
] | {"steps": [{"step_id": "step_1", "tool_name": "extract_numbers", "arguments": {"text": "Budget allocated $5.2 million for research, $3.1 million for development, and $1.7 million for marketing."}, "depends_on": [], "output_key": "extracted"}, {"step_id": "step_2", "tool_name": "statistical_analysis", "arguments": {"numbers": [5.2, 3.1, 1.7], "operation": "summary"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"count": 3, "mean": 3.3333, "median": 3.1, "min": 1.7, "max": 5.2, "sum": 10.0, "stdev": 1.7616} | 2 | 5 | chain | extract-analyze |
L1_chain_0081 | L1_chain | Find the numbers in "Population grew from 8.3 million in 2010 to 9.1 million in 2..." and run a statistical analysis. | [
"extract_numbers",
"statistical_analysis",
"create_contact",
"slugify",
"database_query"
] | {"steps": [{"step_id": "step_1", "tool_name": "extract_numbers", "arguments": {"text": "Population grew from 8.3 million in 2010 to 9.1 million in 2020, an increase of 800,000."}, "depends_on": [], "output_key": "extracted"}, {"step_id": "step_2", "tool_name": "statistical_analysis", "arguments": {"numbers": [8.3, 201, 0, 9.1, 202, 0, 800000], "operation": "summary"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"count": 7, "mean": 114345.7714, "median": 9.1, "min": 0.0, "max": 800000.0, "sum": 800420.4, "stdev": 302345.11} | 2 | 5 | chain | extract-analyze |
L1_chain_0082 | L1_chain | Find the numbers in "Inflation hit 6.2 percent in 2022, up from 1.4 percent the p..." and run a statistical analysis. | [
"extract_numbers",
"statistical_analysis",
"convert_timezone",
"parse_html",
"split_text"
] | {"steps": [{"step_id": "step_1", "tool_name": "extract_numbers", "arguments": {"text": "Inflation hit 6.2 percent in 2022, up from 1.4 percent the previous year."}, "depends_on": [], "output_key": "extracted"}, {"step_id": "step_2", "tool_name": "statistical_analysis", "arguments": {"numbers": [6.2, 202, 2, 1.4], "operation": "summary"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"count": 4, "mean": 52.9, "median": 4.1, "min": 1.4, "max": 202.0, "sum": 211.6, "stdev": 99.4229} | 2 | 5 | chain | extract-analyze |
L1_chain_0083 | L1_chain | What day of the week is "Jan 1, 2027"? Parse the date first. | [
"parse_date",
"get_weekday",
"normalize_data",
"add_duration",
"classify_text"
] | {"steps": [{"step_id": "step_1", "tool_name": "parse_date", "arguments": {"date_string": "Jan 1, 2027"}, "depends_on": [], "output_key": "parsed"}, {"step_id": "step_2", "tool_name": "get_weekday", "arguments": {"date": "2027-01-01"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"weekday": "Friday", "day_number": 4} | 2 | 5 | chain | parse-lookup |
L1_chain_0084 | L1_chain | Convert "March 15, 2026" to a standard date, then find out the weekday. | [
"parse_date",
"get_weekday",
"sentiment_analysis",
"word_count",
"spell_check"
] | {"steps": [{"step_id": "step_1", "tool_name": "parse_date", "arguments": {"date_string": "March 15, 2026"}, "depends_on": [], "output_key": "parsed"}, {"step_id": "step_2", "tool_name": "get_weekday", "arguments": {"date": "2026-03-15"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"weekday": "Sunday", "day_number": 6} | 2 | 5 | chain | parse-lookup |
L1_chain_0085 | L1_chain | What day of the week is "05/05/2026"? Parse the date first. | [
"parse_date",
"get_weekday",
"execute_python",
"compare_texts",
"format_number"
] | {"steps": [{"step_id": "step_1", "tool_name": "parse_date", "arguments": {"date_string": "05/05/2026"}, "depends_on": [], "output_key": "parsed"}, {"step_id": "step_2", "tool_name": "get_weekday", "arguments": {"date": "2026-05-05"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"weekday": "Tuesday", "day_number": 1} | 2 | 5 | chain | parse-lookup |
L1_chain_0086 | L1_chain | Open /data/employees.csv and pull out the most important keywords. | [
"read_file",
"keyword_extract",
"add_duration",
"slugify",
"dns_lookup"
] | {"steps": [{"step_id": "step_1", "tool_name": "read_file", "arguments": {"path": "/data/employees.csv"}, "depends_on": [], "output_key": "content"}, {"step_id": "step_2", "tool_name": "keyword_extract", "arguments": {"text": "name,department,salary,years\nAlice,Engineering,95000,5\nBob,Marketing,72000,3\nCharlie,Engineering,105000,8\nDiana,Sales,68000,2\nEve,Engineering,88000,4\nFrank,Marketing,76000,6\nGrace,Sales,71000,3\nHenry,Engineering,112000,10"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"keywords": ["engineering", "marketing", "sales", "name", "department"], "scores": [1.0, 0.5, 0.5, 0.25, 0.25]} | 2 | 5 | chain | read-extract |
L1_chain_0087 | L1_chain | Read the file at /data/config.json and extract the keywords from its contents. | [
"read_file",
"keyword_extract",
"encode_url",
"extract_entities",
"rss_feed_parse"
] | {"steps": [{"step_id": "step_1", "tool_name": "read_file", "arguments": {"path": "/data/config.json"}, "depends_on": [], "output_key": "content"}, {"step_id": "step_2", "tool_name": "keyword_extract", "arguments": {"text": "{\"app_name\": \"DataPipeline\", \"version\": \"2.1.0\", \"max_workers\": 4, \"timeout_seconds\": 30, \"features\": {\"caching\": true, \"logging\": true, \"metrics\": false}}"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"keywords": ["true", "datapipeline", "version", "features", "caching"], "scores": [1.0, 0.5, 0.5, 0.5, 0.5]} | 2 | 5 | chain | read-extract |
L1_chain_0088 | L1_chain | Read /data/report.txt, then identify the key topics and keywords. | [
"read_file",
"keyword_extract",
"retrieve_memory",
"generate_summary_stats",
"base64_decode"
] | {"steps": [{"step_id": "step_1", "tool_name": "read_file", "arguments": {"path": "/data/report.txt"}, "depends_on": [], "output_key": "content"}, {"step_id": "step_2", "tool_name": "keyword_extract", "arguments": {"text": "Q1 2026 Sales Report\n\nTotal revenue: $2.4M\nGrowth: 15% YoY\nTop product: Enterprise Plan ($1.2M)\nNew customers: 340\nChurn rate: 2.1%\n\nKey highlights:\n- Enterprise segment grew 28%\n- APAC region exceeded targets by 12%\n- Customer satisfaction score: 4.6/5.0"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"keywords": ["enterprise", "sales", "report", "total", "revenue"], "scores": [1.0, 0.5, 0.5, 0.5, 0.5]} | 2 | 5 | chain | read-extract |
L1_chain_0089 | L1_chain | Read the file at /data/config.json and count how many words it contains. | [
"read_file",
"word_count",
"log_event",
"url_parse",
"percentile"
] | {"steps": [{"step_id": "step_1", "tool_name": "read_file", "arguments": {"path": "/data/config.json"}, "depends_on": [], "output_key": "content"}, {"step_id": "step_2", "tool_name": "word_count", "arguments": {"text": "{\"app_name\": \"DataPipeline\", \"version\": \"2.1.0\", \"max_workers\": 4, \"timeout_seconds\": 30, \"features\": {\"caching\": true, \"logging\": true, \"metrics\": false}}"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"words": 15, "characters": 155, "sentences": 2} | 2 | 5 | chain | read-count |
L1_chain_0090 | L1_chain | How many words are in the file at /data/report.txt? | [
"read_file",
"word_count",
"generate_report",
"add_duration",
"generate_summary_stats"
] | {"steps": [{"step_id": "step_1", "tool_name": "read_file", "arguments": {"path": "/data/report.txt"}, "depends_on": [], "output_key": "content"}, {"step_id": "step_2", "tool_name": "word_count", "arguments": {"text": "Q1 2026 Sales Report\n\nTotal revenue: $2.4M\nGrowth: 15% YoY\nTop product: Enterprise Plan ($1.2M)\nNew customers: 340\nChurn rate: 2.1%\n\nKey highlights:\n- Enterprise segment grew 28%\n- APAC region exceeded targets by 12%\n- Customer satisfaction score: 4.6/5.0"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"words": 40, "characters": 255, "sentences": 5} | 2 | 5 | chain | read-count |
L1_chain_0091 | L1_chain | How many words are in the file at /data/config.json? | [
"read_file",
"word_count",
"truncate_text",
"business_days_between",
"generate_summary_stats"
] | {"steps": [{"step_id": "step_1", "tool_name": "read_file", "arguments": {"path": "/data/config.json"}, "depends_on": [], "output_key": "content"}, {"step_id": "step_2", "tool_name": "word_count", "arguments": {"text": "{\"app_name\": \"DataPipeline\", \"version\": \"2.1.0\", \"max_workers\": 4, \"timeout_seconds\": 30, \"features\": {\"caching\": true, \"logging\": true, \"metrics\": false}}"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"words": 15, "characters": 155, "sentences": 2} | 2 | 5 | chain | read-count |
L1_chain_0092 | L1_chain | Look up the weather in Beijing, then log this as an info event. | [
"get_weather",
"log_event",
"calculate_date_diff",
"generate_url",
"retrieve_memory"
] | {"steps": [{"step_id": "step_1", "tool_name": "get_weather", "arguments": {"city": "Beijing"}, "depends_on": [], "output_key": "weather"}, {"step_id": "step_2", "tool_name": "log_event", "arguments": {"event_type": "weather_check", "message": "Weather in Beijing: cloudy, 19°C", "severity": "info"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"log_id": "log_065f87e80706", "event_type": "weather_check", "message": "Weather in Beijing: cloudy, 19°C", "severity": "info", "timestamp": "2026-02-22T12:00:00"} | 2 | 5 | chain | retrieve-log |
L1_chain_0093 | L1_chain | Fetch Manila's weather and record it in the event log. | [
"get_weather",
"log_event",
"set_reminder",
"generate_summary_stats",
"case_convert"
] | {"steps": [{"step_id": "step_1", "tool_name": "get_weather", "arguments": {"city": "Manila"}, "depends_on": [], "output_key": "weather"}, {"step_id": "step_2", "tool_name": "log_event", "arguments": {"event_type": "weather_check", "message": "Weather in Manila: foggy, 26°C", "severity": "info"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"log_id": "log_fe8e21a90258", "event_type": "weather_check", "message": "Weather in Manila: foggy, 26°C", "severity": "info", "timestamp": "2026-02-22T12:00:00"} | 2 | 5 | chain | retrieve-log |
L1_chain_0094 | L1_chain | Look up the weather in Lisbon, then log this as an info event. | [
"get_weather",
"log_event",
"correlation",
"web_search",
"data_filter"
] | {"steps": [{"step_id": "step_1", "tool_name": "get_weather", "arguments": {"city": "Lisbon"}, "depends_on": [], "output_key": "weather"}, {"step_id": "step_2", "tool_name": "log_event", "arguments": {"event_type": "weather_check", "message": "Weather in Lisbon: snowy, 16°C", "severity": "info"}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"log_id": "log_499c52eb4b19", "event_type": "weather_check", "message": "Weather in Lisbon: snowy, 16°C", "severity": "info", "timestamp": "2026-02-22T12:00:00"} | 2 | 5 | chain | retrieve-log |
L1_chain_0095 | L1_chain | Evaluate 295 * 59 and give me the answer rounded to 3 places. | [
"calculator",
"round_number",
"standard_deviation",
"create_task",
"log_event"
] | {"steps": [{"step_id": "step_1", "tool_name": "calculator", "arguments": {"expression": "295 * 59"}, "depends_on": [], "output_key": "calc"}, {"step_id": "step_2", "tool_name": "round_number", "arguments": {"value": 17405.0, "decimals": 3}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"rounded": 17405.0} | 2 | 5 | chain | compute-round |
L1_chain_0096 | L1_chain | Compute 881 * 43, then round the answer to 3 decimals. | [
"calculator",
"round_number",
"generate_image",
"extract_domain",
"web_page_fetch"
] | {"steps": [{"step_id": "step_1", "tool_name": "calculator", "arguments": {"expression": "881 * 43"}, "depends_on": [], "output_key": "calc"}, {"step_id": "step_2", "tool_name": "round_number", "arguments": {"value": 37883.0, "decimals": 3}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"rounded": 37883.0} | 2 | 5 | chain | compute-round |
L1_chain_0097 | L1_chain | Compute 358 + 37, then round the answer to 0 decimals. | [
"calculator",
"round_number",
"data_filter",
"merge_data",
"join_texts"
] | {"steps": [{"step_id": "step_1", "tool_name": "calculator", "arguments": {"expression": "358 + 37"}, "depends_on": [], "output_key": "calc"}, {"step_id": "step_2", "tool_name": "round_number", "arguments": {"value": 395.0, "decimals": 0}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"rounded": 395.0} | 2 | 5 | chain | compute-round |
L1_chain_0098 | L1_chain | Get the numeric values in "The garden covers 2.5 acres with over 300 species of plants ..." and compute their range. | [
"extract_numbers",
"min_max",
"summarize_text",
"check_url_status",
"transform_format"
] | {"steps": [{"step_id": "step_1", "tool_name": "extract_numbers", "arguments": {"text": "The garden covers 2.5 acres with over 300 species of plants and 15 fountains."}, "depends_on": [], "output_key": "extracted"}, {"step_id": "step_2", "tool_name": "min_max", "arguments": {"numbers": [2.5, 300, 15]}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"min": 2.5, "max": 300.0, "range": 297.5} | 2 | 5 | chain | extract-aggregate |
L1_chain_0099 | L1_chain | Get the numeric values in "The satellite orbits at 35,786 km altitude completing 1 revo..." and compute their range. | [
"extract_numbers",
"min_max",
"create_calendar_event",
"encode_url",
"create_notification"
] | {"steps": [{"step_id": "step_1", "tool_name": "extract_numbers", "arguments": {"text": "The satellite orbits at 35,786 km altitude completing 1 revolution every 24 hours."}, "depends_on": [], "output_key": "extracted"}, {"step_id": "step_2", "tool_name": "min_max", "arguments": {"numbers": [35786, 1, 24]}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"min": 1.0, "max": 35786.0, "range": 35785.0} | 2 | 5 | chain | extract-aggregate |
L1_chain_0100 | L1_chain | Pull out the numbers from "The recipe calls for 2.5 cups of flour, 3 eggs, and 175 gram...", then identify the smallest and largest. | [
"extract_numbers",
"min_max",
"write_file",
"text_similarity",
"format_number"
] | {"steps": [{"step_id": "step_1", "tool_name": "extract_numbers", "arguments": {"text": "The recipe calls for 2.5 cups of flour, 3 eggs, and 175 grams of sugar."}, "depends_on": [], "output_key": "extracted"}, {"step_id": "step_2", "tool_name": "min_max", "arguments": {"numbers": [2.5, 3, 175]}, "depends_on": ["step_1"], "output_key": null}], "final_answer_source": "step_2"} | {"min": 2.5, "max": 175.0, "range": 172.5} | 2 | 5 | chain | extract-aggregate |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.