Add files using upload-large-folder tool
Browse files- test_online/config_0.json +13 -0
- test_online/config_100.json +13 -0
- test_online/config_102.json +13 -0
- test_online/config_105.json +13 -0
- test_online/config_115.json +13 -0
- test_online/config_135.json +13 -0
- test_online/config_136.json +13 -0
- test_online/config_142.json +13 -0
- test_online/config_147.json +13 -0
- test_online/config_150.json +13 -0
- test_online/config_151.json +13 -0
- test_online/config_156.json +13 -0
- test_online/config_178.json +13 -0
- test_online/config_18.json +13 -0
- test_online/config_196.json +13 -0
- test_online/config_197.json +13 -0
- test_online/config_200.json +13 -0
- test_online/config_204.json +13 -0
- test_online/config_212.json +13 -0
- test_online/config_214.json +13 -0
- test_online/config_22.json +13 -0
- test_online/config_228.json +13 -0
- test_online/config_229.json +13 -0
- test_online/config_23.json +13 -0
- test_online/config_234.json +13 -0
- test_online/config_25.json +13 -0
- test_online/config_262.json +13 -0
- test_online/config_265.json +13 -0
- test_online/config_268.json +13 -0
- test_online/config_286.json +13 -0
- test_online/config_287.json +13 -0
- test_online/config_296.json +13 -0
- test_online/config_297.json +13 -0
- test_online/config_298.json +13 -0
- test_online/config_31.json +13 -0
- test_online/config_49.json +13 -0
- test_online/config_5.json +13 -0
- test_online/config_54.json +13 -0
- test_online/config_6.json +13 -0
- test_online/config_60.json +13 -0
- test_online/config_65.json +13 -0
- test_online/config_68.json +13 -0
- test_online/config_69.json +13 -0
- test_online/config_7.json +13 -0
- test_online/config_72.json +13 -0
- test_online/config_76.json +13 -0
- test_online/config_8.json +13 -0
- test_online/config_85.json +13 -0
- test_online/config_92.json +13 -0
- test_online/config_97.json +13 -0
test_online/config_0.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Find the store location and hours of the closest Gamestop to zip code 90028 and set it as the home store on Gamestop.",
|
| 3 |
+
"task_id": 0,
|
| 4 |
+
"start_url": "https://www.gamestop.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"medium"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_100.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Show me community posts about pregnancy fever from the past 30 days.",
|
| 3 |
+
"task_id": 100,
|
| 4 |
+
"start_url": "https://www.babycenter.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"medium"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_102.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Search for adoptable dogs near 21122 zip code.",
|
| 3 |
+
"task_id": 102,
|
| 4 |
+
"start_url": "https://www.adoptapet.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"easy"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_105.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Find the lowest 27\"-32\" Samsung or LG computer monitors nearby which have 4k, IPS display.",
|
| 3 |
+
"task_id": 105,
|
| 4 |
+
"start_url": "https://www.google.com/shopping?udm=28",
|
| 5 |
+
"sites": [
|
| 6 |
+
"medium"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_115.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Find a personal trainer service at 10040 for a 25-year-old client aiming to build muscle.",
|
| 3 |
+
"task_id": 115,
|
| 4 |
+
"start_url": "https://www.thumbtack.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"medium"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_135.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Look for the largest hunting land for auction in Kansas high plain region with mineral rights posted in the last seven days.",
|
| 3 |
+
"task_id": 135,
|
| 4 |
+
"start_url": "https://www.landwatch.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"medium"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_136.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Calculate the shipping cost for 4 pound package from Texas to New York.",
|
| 3 |
+
"task_id": 136,
|
| 4 |
+
"start_url": "https://www.fedex.com/en-us/home.html",
|
| 5 |
+
"sites": [
|
| 6 |
+
"medium"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_142.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "See the prediction about the girl child's height, whose current height at seven years is 4 feet and whose weight is 55 lbs, her mother is 5 feet 2, and her father is 5 feet 8.",
|
| 3 |
+
"task_id": 142,
|
| 4 |
+
"start_url": "https://www.babycenter.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"hard"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_147.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Find cats available for adoption within 10 miles of zip code 94587, Young or adult-age cats, sorted by Oldest Addition.",
|
| 3 |
+
"task_id": 147,
|
| 4 |
+
"start_url": "https://www.petfinder.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"medium"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_150.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Show Teen Driver Safety program information.",
|
| 3 |
+
"task_id": 150,
|
| 4 |
+
"start_url": "https://www.dmv.virginia.gov/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"easy"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_151.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Check the interaction between Novolin N and Novolin R.",
|
| 3 |
+
"task_id": 151,
|
| 4 |
+
"start_url": "https://www.drugs.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"medium"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_156.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Search for the ovulation calculator and enter Mar 1 as the first date of the period and calculate the date of ovulation and pregnancy test day.",
|
| 3 |
+
"task_id": 156,
|
| 4 |
+
"start_url": "https://www.webmd.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"hard"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_178.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Search for a beginner\u2019s course in computer science that includes advertisement skills.",
|
| 3 |
+
"task_id": 178,
|
| 4 |
+
"start_url": "https://www.coursera.org/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"medium"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_18.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Show me the options for a roundtrip leaving from Las Vegas on flexible dates on the interactive map on united.",
|
| 3 |
+
"task_id": 18,
|
| 4 |
+
"start_url": "https://www.united.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"hard"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_196.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Find the way to give a gift to UM-Dearborn.",
|
| 3 |
+
"task_id": 196,
|
| 4 |
+
"start_url": "https://umich.edu/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"easy"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_197.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "See the monthly forecast for Atlanta, GA.",
|
| 3 |
+
"task_id": 197,
|
| 4 |
+
"start_url": "https://www.accuweather.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"easy"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_200.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Show me the rules and cancellation for Alley Spring.",
|
| 3 |
+
"task_id": 200,
|
| 4 |
+
"start_url": "https://www.recreation.gov/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"easy"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_204.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Please find graduate-level computer science courses scheduled on Tuesdays starting time from 2:00 to 6:00 PM in the Fall 2023 semester.",
|
| 3 |
+
"task_id": 204,
|
| 4 |
+
"start_url": "https://www.berkeley.edu/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"medium"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_212.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Look for the newest refrigerator that is 34-36 inches wide, priced between $1,000 and $2,000, and has a customer review rating of 4 stars or higher.",
|
| 3 |
+
"task_id": 212,
|
| 4 |
+
"start_url": "https://www.costco.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"medium"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_214.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Add a 5-piece Tenders Combo to my bag with Sweet Corn as the side, Sweet Tea as the drink, and both Honey BBQ and Honey Mustard sauces. Select the store closest to Zip code 10001 for pick-up tomorrow at 12:00 PM.",
|
| 3 |
+
"task_id": 214,
|
| 4 |
+
"start_url": "https://www.kfc.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"hard"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_22.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Find out what to do when I lose an item on a bus on us.megabus.",
|
| 3 |
+
"task_id": 22,
|
| 4 |
+
"start_url": "https://us.megabus.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"easy"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_228.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Identify the open issue with the most comments in the first trending open-source repository this week.",
|
| 3 |
+
"task_id": 228,
|
| 4 |
+
"start_url": "https://github.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"medium"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_229.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Find the first commit submitted by NielsRogge to the official repository of the SAM2 model.",
|
| 3 |
+
"task_id": 229,
|
| 4 |
+
"start_url": "https://github.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"medium"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_23.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Browse the page with event planning tips on Eventbrite.",
|
| 3 |
+
"task_id": 23,
|
| 4 |
+
"start_url": "https://www.eventbrite.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"easy"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_234.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Open the page for the first Best Paper Award video recording of talks from ICLR 2016.",
|
| 3 |
+
"task_id": 234,
|
| 4 |
+
"start_url": "https://iclr.cc/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"easy"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_25.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Find technical specs for the latest Macbook Air on Apple.",
|
| 3 |
+
"task_id": 25,
|
| 4 |
+
"start_url": "https://www.apple.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"easy"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_262.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Filter handbags to evening bags that are blue, and polyester and cost less than $100.",
|
| 3 |
+
"task_id": 262,
|
| 4 |
+
"start_url": "https://www.macys.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"medium"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_265.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Search for regular weekday jobs around 14810 that I can start within two weeks or three.",
|
| 3 |
+
"task_id": 265,
|
| 4 |
+
"start_url": "https://hiring.amazon.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"hard"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_268.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Check the most recent full-time medical health and safety jobs, requiring 1-3 years of industry experience available in the US.",
|
| 3 |
+
"task_id": 268,
|
| 4 |
+
"start_url": "https://www.amazon.jobs/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"medium"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_286.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Estimate the federal income tax I would owe on $158,500 of taxable income in ZIP code 97007, filing as single.",
|
| 3 |
+
"task_id": 286,
|
| 4 |
+
"start_url": "https://smartasset.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"medium"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_287.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Create a meme with a frog as the background and leave the only text with \"Enjoy your life\".",
|
| 3 |
+
"task_id": 287,
|
| 4 |
+
"start_url": "https://imgur.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"medium"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_296.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Find and open an animal learning course on YouTube Kids for my 6-year-old without login in. As a parent born in 1992, I would prefer not to enable search.",
|
| 3 |
+
"task_id": 296,
|
| 4 |
+
"start_url": "https://www.youtube.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"hard"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_297.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Estimate the total cost (with basic support) of using 5 million input tokens and 5 million output tokens each for GPT-4o and GPT-4o Mini, both deployed in the US/EU Data Zones under Standard (On-Demand) in the East US region.",
|
| 3 |
+
"task_id": 297,
|
| 4 |
+
"start_url": "https://azure.microsoft.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"hard"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_298.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Show me the result of a proton emission decay for a Beryllium nucleus with 6 protons and 4 neutrons in the simulation.",
|
| 3 |
+
"task_id": 298,
|
| 4 |
+
"start_url": "https://phet.colorado.edu/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"hard"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_31.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Find a help page about buying tickets on seatgeek.",
|
| 3 |
+
"task_id": 31,
|
| 4 |
+
"start_url": "https://seatgeek.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"easy"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_49.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Find a permanent job in Logistics within 20 miles of New York, zip 11005, in the middle-income range for a high school diploma holder.",
|
| 3 |
+
"task_id": 49,
|
| 4 |
+
"start_url": "https://ohiomeansjobs.ohio.gov/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"hard"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_5.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Find a 2022 Tesla Model 3 on CarMax.",
|
| 3 |
+
"task_id": 5,
|
| 4 |
+
"start_url": "https://www.carmax.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"medium"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_54.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Find a premier real estate agent in St Augustine, FL.",
|
| 3 |
+
"task_id": 54,
|
| 4 |
+
"start_url": "https://www.redfin.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"easy"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_6.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Get the report from the final environmental impact statement for the Jamaica Bus Depot expansion on new.mta.info.",
|
| 3 |
+
"task_id": 6,
|
| 4 |
+
"start_url": "https://new.mta.info/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"medium"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_60.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Look for reviews of a Nest Hello Video Doorbell and filter by 1-star ratings.",
|
| 3 |
+
"task_id": 60,
|
| 4 |
+
"start_url": "https://www.google.com/shopping?udm=28",
|
| 5 |
+
"sites": [
|
| 6 |
+
"easy"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_65.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "View the speakers that are bluetooth and wireless and filter the results to only show models that are on sale and cost less than $50.",
|
| 3 |
+
"task_id": 65,
|
| 4 |
+
"start_url": "https://www.bestbuy.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"medium"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_68.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Find the latest climate news.",
|
| 3 |
+
"task_id": 68,
|
| 4 |
+
"start_url": "https://www.theweathernetwork.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"easy"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_69.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Find a staffed FedEx location near zip code 10019 to return a package.",
|
| 3 |
+
"task_id": 69,
|
| 4 |
+
"start_url": "https://www.fedex.com/en-us/home.html",
|
| 5 |
+
"sites": [
|
| 6 |
+
"medium"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_7.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Find the used 2012-2013 Honda Crosstour with the lowest mileage for under $25,000 near zip code 49102 on CarGurus.",
|
| 3 |
+
"task_id": 7,
|
| 4 |
+
"start_url": "https://www.cargurus.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"hard"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_72.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Add a $50 Uber gift card to the cart.",
|
| 3 |
+
"task_id": 72,
|
| 4 |
+
"start_url": "https://www.bestbuy.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"easy"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_76.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Browse apartments with at least 2 bedrooms and 2 bathrooms and a max price of $4000 per month.",
|
| 3 |
+
"task_id": 76,
|
| 4 |
+
"start_url": "https://craigslist.org/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"easy"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_8.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Find a walkthrough for the game \"The Legend of Zelda: Breath of the Wild\" on ign.",
|
| 3 |
+
"task_id": 8,
|
| 4 |
+
"start_url": "https://www.ign.com/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"medium"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_85.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Browse couches for sale, sort by cheapest, and search in titles only.",
|
| 3 |
+
"task_id": 85,
|
| 4 |
+
"start_url": "https://craigslist.org/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"medium"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_92.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Look for the best rated BBB accredited charity near 12023.",
|
| 3 |
+
"task_id": 92,
|
| 4 |
+
"start_url": "https://www.bbb.org/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"medium"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|
test_online/config_97.json
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"intent": "Check if a visa is required to work in the UK for longer than 6 months in Healthcare as an American citizen.",
|
| 3 |
+
"task_id": 97,
|
| 4 |
+
"start_url": "https://www.gov.uk/",
|
| 5 |
+
"sites": [
|
| 6 |
+
"hard"
|
| 7 |
+
],
|
| 8 |
+
"eval": {
|
| 9 |
+
"eval_types": [
|
| 10 |
+
"LLM_eval"
|
| 11 |
+
]
|
| 12 |
+
}
|
| 13 |
+
}
|