Datasets:
Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0041/final_frame.png +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0041/first_frame.png +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0041/ground_truth.mp4 +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0048/final_frame.png +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0048/first_frame.png +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0048/ground_truth.mp4 +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0065/original/question.json +14 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0070/final_frame.png +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0070/first_frame.png +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0070/ground_truth.mp4 +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0077/final_frame.png +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0077/first_frame.png +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0077/ground_truth.mp4 +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0083/final_frame.png +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0083/first_frame.png +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0083/ground_truth.mp4 +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0084/final_frame.png +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0084/first_frame.png +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0084/ground_truth.mp4 +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0110/original/question.json +14 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0117/original/question.json +14 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0119/original/question.json +14 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0121/original/question.json +14 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0126/original/question.json +14 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0128/original/question.json +14 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0175/original/question.json +14 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0242/final_frame.png +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0242/first_frame.png +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0242/ground_truth.mp4 +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0245/final_frame.png +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0245/first_frame.png +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0245/ground_truth.mp4 +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0257/original/question.json +14 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0273/final_frame.png +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0273/first_frame.png +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0273/ground_truth.mp4 +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0287/final_frame.png +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0287/first_frame.png +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0287/ground_truth.mp4 +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0289/final_frame.png +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0289/first_frame.png +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0289/ground_truth.mp4 +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0313/original/question.json +14 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0314/original/question.json +14 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0322/original/question.json +14 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0325/original/question.json +14 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0347/original/question.json +14 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0378/original/question.json +14 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0413/final_frame.png +3 -0
- M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0413/first_frame.png +3 -0
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0041/final_frame.png
ADDED
|
Git LFS Details
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0041/first_frame.png
ADDED
|
Git LFS Details
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0041/ground_truth.mp4
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8db8f802ed9689012e33b73d9ddf9c78b290f8fd2d4f6a9f9f6b0fb3d80b7c77
|
| 3 |
+
size 25855
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0048/final_frame.png
ADDED
|
Git LFS Details
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0048/first_frame.png
ADDED
|
Git LFS Details
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0048/ground_truth.mp4
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3b55c9298266d3ba4c2013fdc9e6e9c486826d995edd30dc3ebc1ba1d5a9a612
|
| 3 |
+
size 22895
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0065/original/question.json
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"dataset": "ScienceQA",
|
| 3 |
+
"source_id": "train_0000368",
|
| 4 |
+
"question": "Select the chemical formula for this molecule.",
|
| 5 |
+
"choices": {
|
| 6 |
+
"A": "SiH4",
|
| 7 |
+
"B": "SiHe",
|
| 8 |
+
"C": "SiHe4",
|
| 9 |
+
"D": "Si2H4"
|
| 10 |
+
},
|
| 11 |
+
"answer": "A",
|
| 12 |
+
"original_image_filename": "train_0000368.png",
|
| 13 |
+
"has_embedded_choices": false
|
| 14 |
+
}
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0070/final_frame.png
ADDED
|
Git LFS Details
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0070/first_frame.png
ADDED
|
Git LFS Details
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0070/ground_truth.mp4
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7f877976543217c91a57f28cbb7d8fbf7f4241e21d2c5481b44fa688340f4b53
|
| 3 |
+
size 25303
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0077/final_frame.png
ADDED
|
Git LFS Details
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0077/first_frame.png
ADDED
|
Git LFS Details
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0077/ground_truth.mp4
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fcaede55ab54713a47c96a1252abd65d4ff273ec418fda004c1e3871cc918be4
|
| 3 |
+
size 33728
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0083/final_frame.png
ADDED
|
Git LFS Details
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0083/first_frame.png
ADDED
|
Git LFS Details
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0083/ground_truth.mp4
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b3297dcf143f342db6cb3c3e00618348ef08e712778608149f57331c4065c380
|
| 3 |
+
size 23505
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0084/final_frame.png
ADDED
|
Git LFS Details
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0084/first_frame.png
ADDED
|
Git LFS Details
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0084/ground_truth.mp4
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4b6ad7f7faf29dc89934361783502dcde4484367e2639d1718da8d611cfd6e84
|
| 3 |
+
size 28579
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0110/original/question.json
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"dataset": "ScienceQA",
|
| 3 |
+
"source_id": "train_0000610",
|
| 4 |
+
"question": "Which of these states is farthest north?",
|
| 5 |
+
"choices": {
|
| 6 |
+
"A": "West Virginia",
|
| 7 |
+
"B": "Tennessee",
|
| 8 |
+
"C": "Florida",
|
| 9 |
+
"D": "Louisiana"
|
| 10 |
+
},
|
| 11 |
+
"answer": "A",
|
| 12 |
+
"original_image_filename": "train_0000610.png",
|
| 13 |
+
"has_embedded_choices": false
|
| 14 |
+
}
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0117/original/question.json
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"dataset": "ScienceQA",
|
| 3 |
+
"source_id": "train_0000639",
|
| 4 |
+
"question": "Which country is highlighted?",
|
| 5 |
+
"choices": {
|
| 6 |
+
"A": "Tuvalu",
|
| 7 |
+
"B": "Palau",
|
| 8 |
+
"C": "the Federated States of Micronesia",
|
| 9 |
+
"D": "Australia"
|
| 10 |
+
},
|
| 11 |
+
"answer": "B",
|
| 12 |
+
"original_image_filename": "train_0000639.png",
|
| 13 |
+
"has_embedded_choices": false
|
| 14 |
+
}
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0119/original/question.json
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"dataset": "ScienceQA",
|
| 3 |
+
"source_id": "train_0000649",
|
| 4 |
+
"question": "Which country is highlighted?",
|
| 5 |
+
"choices": {
|
| 6 |
+
"A": "Tuvalu",
|
| 7 |
+
"B": "the Federated States of Micronesia",
|
| 8 |
+
"C": "Palau",
|
| 9 |
+
"D": "Fiji"
|
| 10 |
+
},
|
| 11 |
+
"answer": "C",
|
| 12 |
+
"original_image_filename": "train_0000649.png",
|
| 13 |
+
"has_embedded_choices": false
|
| 14 |
+
}
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0121/original/question.json
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"dataset": "ScienceQA",
|
| 3 |
+
"source_id": "train_0000651",
|
| 4 |
+
"question": "What is the capital of New York?",
|
| 5 |
+
"choices": {
|
| 6 |
+
"A": "Carson City",
|
| 7 |
+
"B": "Albany",
|
| 8 |
+
"C": "Annapolis",
|
| 9 |
+
"D": "Augusta"
|
| 10 |
+
},
|
| 11 |
+
"answer": "B",
|
| 12 |
+
"original_image_filename": "train_0000651.png",
|
| 13 |
+
"has_embedded_choices": false
|
| 14 |
+
}
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0126/original/question.json
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"dataset": "ScienceQA",
|
| 3 |
+
"source_id": "train_0000677",
|
| 4 |
+
"question": "What is the name of the colony shown?",
|
| 5 |
+
"choices": {
|
| 6 |
+
"A": "Georgia",
|
| 7 |
+
"B": "Alabama",
|
| 8 |
+
"C": "Virginia",
|
| 9 |
+
"D": "Michigan"
|
| 10 |
+
},
|
| 11 |
+
"answer": "A",
|
| 12 |
+
"original_image_filename": "train_0000677.png",
|
| 13 |
+
"has_embedded_choices": false
|
| 14 |
+
}
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0128/original/question.json
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"dataset": "ScienceQA",
|
| 3 |
+
"source_id": "train_0000698",
|
| 4 |
+
"question": "Which of these cities is marked on the map?",
|
| 5 |
+
"choices": {
|
| 6 |
+
"A": "Atlanta",
|
| 7 |
+
"B": "Houston",
|
| 8 |
+
"C": "New Orleans",
|
| 9 |
+
"D": "Memphis"
|
| 10 |
+
},
|
| 11 |
+
"answer": "D",
|
| 12 |
+
"original_image_filename": "train_0000698.png",
|
| 13 |
+
"has_embedded_choices": false
|
| 14 |
+
}
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0175/original/question.json
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"dataset": "ScienceQA",
|
| 3 |
+
"source_id": "train_0000962",
|
| 4 |
+
"question": "Which ocean is highlighted?",
|
| 5 |
+
"choices": {
|
| 6 |
+
"A": "the Arctic Ocean",
|
| 7 |
+
"B": "the Southern Ocean",
|
| 8 |
+
"C": "the Pacific Ocean",
|
| 9 |
+
"D": "the Atlantic Ocean"
|
| 10 |
+
},
|
| 11 |
+
"answer": "D",
|
| 12 |
+
"original_image_filename": "train_0000962.png",
|
| 13 |
+
"has_embedded_choices": false
|
| 14 |
+
}
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0242/final_frame.png
ADDED
|
Git LFS Details
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0242/first_frame.png
ADDED
|
Git LFS Details
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0242/ground_truth.mp4
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c76a2b3e548ada2d810072b2710d99b73bf1df5c8eeda00f00b45f80ab095e4a
|
| 3 |
+
size 25638
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0245/final_frame.png
ADDED
|
Git LFS Details
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0245/first_frame.png
ADDED
|
Git LFS Details
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0245/ground_truth.mp4
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:807ac66c3c570234515b78437ed8949e8606922cf37287c8610cc1be7cb278bd
|
| 3 |
+
size 24468
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0257/original/question.json
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"dataset": "ScienceQA",
|
| 3 |
+
"source_id": "train_0001387",
|
| 4 |
+
"question": "What is the capital of Wyoming?",
|
| 5 |
+
"choices": {
|
| 6 |
+
"A": "Albuquerque",
|
| 7 |
+
"B": "Cheyenne",
|
| 8 |
+
"C": "Olympia",
|
| 9 |
+
"D": "Madison"
|
| 10 |
+
},
|
| 11 |
+
"answer": "B",
|
| 12 |
+
"original_image_filename": "train_0001387.png",
|
| 13 |
+
"has_embedded_choices": false
|
| 14 |
+
}
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0273/final_frame.png
ADDED
|
Git LFS Details
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0273/first_frame.png
ADDED
|
Git LFS Details
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0273/ground_truth.mp4
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a822946d7a2ce833abda500cae371024fce0fc844182e57e3e4c0919ddc18c0f
|
| 3 |
+
size 20315
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0287/final_frame.png
ADDED
|
Git LFS Details
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0287/first_frame.png
ADDED
|
Git LFS Details
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0287/ground_truth.mp4
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:267017f1b4ceb14af8cdedfd0ba27d82e97b1a3af816773a6b16e79a4986c9ae
|
| 3 |
+
size 21764
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0289/final_frame.png
ADDED
|
Git LFS Details
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0289/first_frame.png
ADDED
|
Git LFS Details
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0289/ground_truth.mp4
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0f434a8a09c6b195cc94ac635b56f9bc53a85619b521ffc28c0ab534dd60283c
|
| 3 |
+
size 23615
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0313/original/question.json
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"dataset": "ScienceQA",
|
| 3 |
+
"source_id": "train_0001653",
|
| 4 |
+
"question": "What is the expected ratio of offspring with pale orange cheeks to offspring with bright orange cheeks? Choose the most likely ratio.",
|
| 5 |
+
"choices": {
|
| 6 |
+
"A": "2:2",
|
| 7 |
+
"B": "1:3",
|
| 8 |
+
"C": "0:4",
|
| 9 |
+
"D": "3:1"
|
| 10 |
+
},
|
| 11 |
+
"answer": "B",
|
| 12 |
+
"original_image_filename": "train_0001653.png",
|
| 13 |
+
"has_embedded_choices": false
|
| 14 |
+
}
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0314/original/question.json
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"dataset": "ScienceQA",
|
| 3 |
+
"source_id": "train_0001663",
|
| 4 |
+
"question": "Which of these states is farthest north?",
|
| 5 |
+
"choices": {
|
| 6 |
+
"A": "Mississippi",
|
| 7 |
+
"B": "Indiana",
|
| 8 |
+
"C": "Maryland",
|
| 9 |
+
"D": "Montana"
|
| 10 |
+
},
|
| 11 |
+
"answer": "D",
|
| 12 |
+
"original_image_filename": "train_0001663.png",
|
| 13 |
+
"has_embedded_choices": false
|
| 14 |
+
}
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0322/original/question.json
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"dataset": "ScienceQA",
|
| 3 |
+
"source_id": "train_0001702",
|
| 4 |
+
"question": "What is the capital of Colorado?",
|
| 5 |
+
"choices": {
|
| 6 |
+
"A": "Helena",
|
| 7 |
+
"B": "Denver",
|
| 8 |
+
"C": "Colorado Springs",
|
| 9 |
+
"D": "Kansas City"
|
| 10 |
+
},
|
| 11 |
+
"answer": "B",
|
| 12 |
+
"original_image_filename": "train_0001702.png",
|
| 13 |
+
"has_embedded_choices": false
|
| 14 |
+
}
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0325/original/question.json
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"dataset": "ScienceQA",
|
| 3 |
+
"source_id": "train_0001712",
|
| 4 |
+
"question": "What is the name of the colony shown?",
|
| 5 |
+
"choices": {
|
| 6 |
+
"A": "Rhode Island",
|
| 7 |
+
"B": "Ohio",
|
| 8 |
+
"C": "Connecticut",
|
| 9 |
+
"D": "Maine"
|
| 10 |
+
},
|
| 11 |
+
"answer": "C",
|
| 12 |
+
"original_image_filename": "train_0001712.png",
|
| 13 |
+
"has_embedded_choices": false
|
| 14 |
+
}
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0347/original/question.json
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"dataset": "ScienceQA",
|
| 3 |
+
"source_id": "train_0001855",
|
| 4 |
+
"question": "Which i in row B?",
|
| 5 |
+
"choices": {
|
| 6 |
+
"A": "the pond",
|
| 7 |
+
"B": "the gas station",
|
| 8 |
+
"C": "the grocery store",
|
| 9 |
+
"D": "the school"
|
| 10 |
+
},
|
| 11 |
+
"answer": "A",
|
| 12 |
+
"original_image_filename": "train_0001855.png",
|
| 13 |
+
"has_embedded_choices": false
|
| 14 |
+
}
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0378/original/question.json
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"dataset": "ScienceQA",
|
| 3 |
+
"source_id": "train_0002000",
|
| 4 |
+
"question": "What is the capital of Minnesota?",
|
| 5 |
+
"choices": {
|
| 6 |
+
"A": "Saint Paul",
|
| 7 |
+
"B": "Fort Wayne",
|
| 8 |
+
"C": "Indianapolis",
|
| 9 |
+
"D": "Columbus"
|
| 10 |
+
},
|
| 11 |
+
"answer": "A",
|
| 12 |
+
"original_image_filename": "train_0002000.png",
|
| 13 |
+
"has_embedded_choices": false
|
| 14 |
+
}
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0413/final_frame.png
ADDED
|
Git LFS Details
|
M-2_scienceqa_data-generator/scienceqa_task/scienceqa_0413/first_frame.png
ADDED
|
Git LFS Details
|