Upload folder using huggingface_hub
Browse files- long_bench/dureader/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e.incomplete_info.lock +0 -0
- long_bench/dureader/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/dataset_info.json +1 -0
- long_bench/dureader/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/long_bench-test.arrow +3 -0
- long_bench/dureader/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e_builder.lock +0 -0
- long_bench/gov_report/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e.incomplete_info.lock +0 -0
- long_bench/gov_report/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/dataset_info.json +1 -0
- long_bench/gov_report/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/long_bench-test.arrow +3 -0
- long_bench/gov_report/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e_builder.lock +0 -0
- long_bench/hotpotqa/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e.incomplete_info.lock +0 -0
- long_bench/hotpotqa/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/dataset_info.json +1 -0
- long_bench/hotpotqa/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/long_bench-test.arrow +3 -0
- long_bench/hotpotqa/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e_builder.lock +0 -0
- long_bench/lcc/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e.incomplete_info.lock +0 -0
- long_bench/lcc/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/dataset_info.json +1 -0
- long_bench/lcc/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/long_bench-test.arrow +3 -0
- long_bench/lcc/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e_builder.lock +0 -0
- long_bench/multifieldqa_en/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e.incomplete_info.lock +0 -0
- long_bench/multifieldqa_en/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/dataset_info.json +1 -0
- long_bench/multifieldqa_en/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/long_bench-test.arrow +3 -0
- long_bench/multifieldqa_en/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e_builder.lock +0 -0
- long_bench/multifieldqa_zh/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e.incomplete_info.lock +0 -0
- long_bench/multifieldqa_zh/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/dataset_info.json +1 -0
- long_bench/multifieldqa_zh/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/long_bench-test.arrow +3 -0
- long_bench/multifieldqa_zh/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e_builder.lock +0 -0
- long_bench/musique/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e.incomplete_info.lock +0 -0
- long_bench/musique/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/dataset_info.json +1 -0
- long_bench/musique/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/long_bench-test.arrow +3 -0
- long_bench/musique/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e_builder.lock +0 -0
- long_bench/narrativeqa/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e.incomplete_info.lock +0 -0
- long_bench/narrativeqa/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/dataset_info.json +1 -0
- long_bench/narrativeqa/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/long_bench-test.arrow +3 -0
- long_bench/narrativeqa/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e_builder.lock +0 -0
- long_bench/passage_retrieval_en/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e.incomplete_info.lock +0 -0
- long_bench/passage_retrieval_en/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/dataset_info.json +1 -0
- long_bench/passage_retrieval_en/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/long_bench-test.arrow +3 -0
- long_bench/passage_retrieval_en/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e_builder.lock +0 -0
- long_bench/samsum/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e.incomplete_info.lock +0 -0
- long_bench/samsum/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/dataset_info.json +1 -0
- long_bench/samsum/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/long_bench-test.arrow +3 -0
- long_bench/samsum/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e_builder.lock +0 -0
long_bench/dureader/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e.incomplete_info.lock
ADDED
|
File without changes
|
long_bench/dureader/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/dataset_info.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"description": "LongBench is a comprehensive benchmark for multilingual and multi-task purposes, with the goal to fully measure and evaluate the ability of pre-trained language models to understand long text. This dataset consists of twenty different tasks, covering key long-text application scenarios such as multi-document QA, single-document QA, summarization, few-shot learning, synthetic tasks, and code completion.\n", "citation": "", "homepage": "https://github.com/THUDM/LongBench", "license": "", "features": {"input": {"dtype": "string", "_type": "Value"}, "context": {"dtype": "string", "_type": "Value"}, "answers": [{"dtype": "string", "_type": "Value"}], "length": {"dtype": "int32", "_type": "Value"}, "dataset": {"dtype": "string", "_type": "Value"}, "language": {"dtype": "string", "_type": "Value"}, "all_classes": [{"dtype": "string", "_type": "Value"}], "_id": {"dtype": "string", "_type": "Value"}}, "builder_name": "long_bench", "dataset_name": "long_bench", "config_name": "dureader", "version": {"version_str": "1.0.0", "major": 1, "minor": 0, "patch": 0}, "splits": {"test": {"name": "test", "num_bytes": 8212951, "num_examples": 200, "dataset_name": "long_bench"}}, "download_checksums": {"https://huggingface.co/datasets/THUDM/LongBench/resolve/main/data.zip": {"num_bytes": 113932529, "checksum": null}}, "download_size": 113932529, "dataset_size": 8212951, "size_in_bytes": 122145480}
|
long_bench/dureader/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/long_bench-test.arrow
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2b6ad615b1a3d5c4296ec475c76904926d6585d07c123da339ee8d80e6f2d878
|
| 3 |
+
size 8215560
|
long_bench/dureader/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e_builder.lock
ADDED
|
File without changes
|
long_bench/gov_report/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e.incomplete_info.lock
ADDED
|
File without changes
|
long_bench/gov_report/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/dataset_info.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"description": "LongBench is a comprehensive benchmark for multilingual and multi-task purposes, with the goal to fully measure and evaluate the ability of pre-trained language models to understand long text. This dataset consists of twenty different tasks, covering key long-text application scenarios such as multi-document QA, single-document QA, summarization, few-shot learning, synthetic tasks, and code completion.\n", "citation": "", "homepage": "https://github.com/THUDM/LongBench", "license": "", "features": {"input": {"dtype": "string", "_type": "Value"}, "context": {"dtype": "string", "_type": "Value"}, "answers": [{"dtype": "string", "_type": "Value"}], "length": {"dtype": "int32", "_type": "Value"}, "dataset": {"dtype": "string", "_type": "Value"}, "language": {"dtype": "string", "_type": "Value"}, "all_classes": [{"dtype": "string", "_type": "Value"}], "_id": {"dtype": "string", "_type": "Value"}}, "builder_name": "long_bench", "dataset_name": "long_bench", "config_name": "gov_report", "version": {"version_str": "1.0.0", "major": 1, "minor": 0, "patch": 0}, "splits": {"test": {"name": "test", "num_bytes": 11593569, "num_examples": 200, "dataset_name": "long_bench"}}, "download_checksums": {"https://huggingface.co/datasets/THUDM/LongBench/resolve/main/data.zip": {"num_bytes": 113932529, "checksum": null}}, "download_size": 113932529, "dataset_size": 11593569, "size_in_bytes": 125526098}
|
long_bench/gov_report/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/long_bench-test.arrow
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f36bc9e3b2a7c9488ce5ffd10cf4ffab9624de4096ae54360c75fec19b9a1bc0
|
| 3 |
+
size 11596176
|
long_bench/gov_report/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e_builder.lock
ADDED
|
File without changes
|
long_bench/hotpotqa/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e.incomplete_info.lock
ADDED
|
File without changes
|
long_bench/hotpotqa/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/dataset_info.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"description": "LongBench is a comprehensive benchmark for multilingual and multi-task purposes, with the goal to fully measure and evaluate the ability of pre-trained language models to understand long text. This dataset consists of twenty different tasks, covering key long-text application scenarios such as multi-document QA, single-document QA, summarization, few-shot learning, synthetic tasks, and code completion.\n", "citation": "", "homepage": "https://github.com/THUDM/LongBench", "license": "", "features": {"input": {"dtype": "string", "_type": "Value"}, "context": {"dtype": "string", "_type": "Value"}, "answers": [{"dtype": "string", "_type": "Value"}], "length": {"dtype": "int32", "_type": "Value"}, "dataset": {"dtype": "string", "_type": "Value"}, "language": {"dtype": "string", "_type": "Value"}, "all_classes": [{"dtype": "string", "_type": "Value"}], "_id": {"dtype": "string", "_type": "Value"}}, "builder_name": "long_bench", "dataset_name": "long_bench", "config_name": "hotpotqa", "version": {"version_str": "1.0.0", "major": 1, "minor": 0, "patch": 0}, "splits": {"test": {"name": "test", "num_bytes": 11379153, "num_examples": 200, "dataset_name": "long_bench"}}, "download_checksums": {"https://huggingface.co/datasets/THUDM/LongBench/resolve/main/data.zip": {"num_bytes": 113932529, "checksum": null}}, "download_size": 113932529, "dataset_size": 11379153, "size_in_bytes": 125311682}
|
long_bench/hotpotqa/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/long_bench-test.arrow
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:483ea94037bafdeea11a7c2e27601f1fd7324ca6a73f183c96a2fa3e61c198f8
|
| 3 |
+
size 11381768
|
long_bench/hotpotqa/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e_builder.lock
ADDED
|
File without changes
|
long_bench/lcc/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e.incomplete_info.lock
ADDED
|
File without changes
|
long_bench/lcc/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/dataset_info.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"description": "LongBench is a comprehensive benchmark for multilingual and multi-task purposes, with the goal to fully measure and evaluate the ability of pre-trained language models to understand long text. This dataset consists of twenty different tasks, covering key long-text application scenarios such as multi-document QA, single-document QA, summarization, few-shot learning, synthetic tasks, and code completion.\n", "citation": "", "homepage": "https://github.com/THUDM/LongBench", "license": "", "features": {"input": {"dtype": "string", "_type": "Value"}, "context": {"dtype": "string", "_type": "Value"}, "answers": [{"dtype": "string", "_type": "Value"}], "length": {"dtype": "int32", "_type": "Value"}, "dataset": {"dtype": "string", "_type": "Value"}, "language": {"dtype": "string", "_type": "Value"}, "all_classes": [{"dtype": "string", "_type": "Value"}], "_id": {"dtype": "string", "_type": "Value"}}, "builder_name": "long_bench", "dataset_name": "long_bench", "config_name": "lcc", "version": {"version_str": "1.0.0", "major": 1, "minor": 0, "patch": 0}, "splits": {"test": {"name": "test", "num_bytes": 6878988, "num_examples": 500, "dataset_name": "long_bench"}}, "download_checksums": {"https://huggingface.co/datasets/THUDM/LongBench/resolve/main/data.zip": {"num_bytes": 113932529, "checksum": null}}, "download_size": 113932529, "dataset_size": 6878988, "size_in_bytes": 120811517}
|
long_bench/lcc/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/long_bench-test.arrow
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bd11ddb2a6a69ff225c7add90d446754792826a6f86691cdbf7e1c22cf29e114
|
| 3 |
+
size 6882792
|
long_bench/lcc/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e_builder.lock
ADDED
|
File without changes
|
long_bench/multifieldqa_en/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e.incomplete_info.lock
ADDED
|
File without changes
|
long_bench/multifieldqa_en/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/dataset_info.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"description": "LongBench is a comprehensive benchmark for multilingual and multi-task purposes, with the goal to fully measure and evaluate the ability of pre-trained language models to understand long text. This dataset consists of twenty different tasks, covering key long-text application scenarios such as multi-document QA, single-document QA, summarization, few-shot learning, synthetic tasks, and code completion.\n", "citation": "", "homepage": "https://github.com/THUDM/LongBench", "license": "", "features": {"input": {"dtype": "string", "_type": "Value"}, "context": {"dtype": "string", "_type": "Value"}, "answers": [{"dtype": "string", "_type": "Value"}], "length": {"dtype": "int32", "_type": "Value"}, "dataset": {"dtype": "string", "_type": "Value"}, "language": {"dtype": "string", "_type": "Value"}, "all_classes": [{"dtype": "string", "_type": "Value"}], "_id": {"dtype": "string", "_type": "Value"}}, "builder_name": "long_bench", "dataset_name": "long_bench", "config_name": "multifieldqa_en", "version": {"version_str": "1.0.0", "major": 1, "minor": 0, "patch": 0}, "splits": {"test": {"name": "test", "num_bytes": 4427988, "num_examples": 150, "dataset_name": "long_bench"}}, "download_checksums": {"https://huggingface.co/datasets/THUDM/LongBench/resolve/main/data.zip": {"num_bytes": 113932529, "checksum": null}}, "download_size": 113932529, "dataset_size": 4427988, "size_in_bytes": 118360517}
|
long_bench/multifieldqa_en/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/long_bench-test.arrow
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:eab137b627143b0b1c5d71b5e6c97a9d1109bb84d91fc974f2d3067ad191a06f
|
| 3 |
+
size 4430408
|
long_bench/multifieldqa_en/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e_builder.lock
ADDED
|
File without changes
|
long_bench/multifieldqa_zh/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e.incomplete_info.lock
ADDED
|
File without changes
|
long_bench/multifieldqa_zh/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/dataset_info.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"description": "LongBench is a comprehensive benchmark for multilingual and multi-task purposes, with the goal to fully measure and evaluate the ability of pre-trained language models to understand long text. This dataset consists of twenty different tasks, covering key long-text application scenarios such as multi-document QA, single-document QA, summarization, few-shot learning, synthetic tasks, and code completion.\n", "citation": "", "homepage": "https://github.com/THUDM/LongBench", "license": "", "features": {"input": {"dtype": "string", "_type": "Value"}, "context": {"dtype": "string", "_type": "Value"}, "answers": [{"dtype": "string", "_type": "Value"}], "length": {"dtype": "int32", "_type": "Value"}, "dataset": {"dtype": "string", "_type": "Value"}, "language": {"dtype": "string", "_type": "Value"}, "all_classes": [{"dtype": "string", "_type": "Value"}], "_id": {"dtype": "string", "_type": "Value"}}, "builder_name": "long_bench", "dataset_name": "long_bench", "config_name": "multifieldqa_zh", "version": {"version_str": "1.0.0", "major": 1, "minor": 0, "patch": 0}, "splits": {"test": {"name": "test", "num_bytes": 3541307, "num_examples": 200, "dataset_name": "long_bench"}}, "download_checksums": {"https://huggingface.co/datasets/THUDM/LongBench/resolve/main/data.zip": {"num_bytes": 113932529, "checksum": null}}, "download_size": 113932529, "dataset_size": 3541307, "size_in_bytes": 117473836}
|
long_bench/multifieldqa_zh/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/long_bench-test.arrow
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:399af8603e2793697946e4297fb9e16416efb501e28b315519ea0d587389d988
|
| 3 |
+
size 3543912
|
long_bench/multifieldqa_zh/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e_builder.lock
ADDED
|
File without changes
|
long_bench/musique/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e.incomplete_info.lock
ADDED
|
File without changes
|
long_bench/musique/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/dataset_info.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"description": "LongBench is a comprehensive benchmark for multilingual and multi-task purposes, with the goal to fully measure and evaluate the ability of pre-trained language models to understand long text. This dataset consists of twenty different tasks, covering key long-text application scenarios such as multi-document QA, single-document QA, summarization, few-shot learning, synthetic tasks, and code completion.\n", "citation": "", "homepage": "https://github.com/THUDM/LongBench", "license": "", "features": {"input": {"dtype": "string", "_type": "Value"}, "context": {"dtype": "string", "_type": "Value"}, "answers": [{"dtype": "string", "_type": "Value"}], "length": {"dtype": "int32", "_type": "Value"}, "dataset": {"dtype": "string", "_type": "Value"}, "language": {"dtype": "string", "_type": "Value"}, "all_classes": [{"dtype": "string", "_type": "Value"}], "_id": {"dtype": "string", "_type": "Value"}}, "builder_name": "long_bench", "dataset_name": "long_bench", "config_name": "musique", "version": {"version_str": "1.0.0", "major": 1, "minor": 0, "patch": 0}, "splits": {"test": {"name": "test", "num_bytes": 13965034, "num_examples": 200, "dataset_name": "long_bench"}}, "download_checksums": {"https://huggingface.co/datasets/THUDM/LongBench/resolve/main/data.zip": {"num_bytes": 113932529, "checksum": null}}, "download_size": 113932529, "dataset_size": 13965034, "size_in_bytes": 127897563}
|
long_bench/musique/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/long_bench-test.arrow
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:da6776b779813085083cd3f87135e3f06ddc47b617a9787a01595ad2467e66c8
|
| 3 |
+
size 13967648
|
long_bench/musique/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e_builder.lock
ADDED
|
File without changes
|
long_bench/narrativeqa/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e.incomplete_info.lock
ADDED
|
File without changes
|
long_bench/narrativeqa/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/dataset_info.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"description": "LongBench is a comprehensive benchmark for multilingual and multi-task purposes, with the goal to fully measure and evaluate the ability of pre-trained language models to understand long text. This dataset consists of twenty different tasks, covering key long-text application scenarios such as multi-document QA, single-document QA, summarization, few-shot learning, synthetic tasks, and code completion.\n", "citation": "", "homepage": "https://github.com/THUDM/LongBench", "license": "", "features": {"input": {"dtype": "string", "_type": "Value"}, "context": {"dtype": "string", "_type": "Value"}, "answers": [{"dtype": "string", "_type": "Value"}], "length": {"dtype": "int32", "_type": "Value"}, "dataset": {"dtype": "string", "_type": "Value"}, "language": {"dtype": "string", "_type": "Value"}, "all_classes": [{"dtype": "string", "_type": "Value"}], "_id": {"dtype": "string", "_type": "Value"}}, "builder_name": "long_bench", "dataset_name": "long_bench", "config_name": "narrativeqa", "version": {"version_str": "1.0.0", "major": 1, "minor": 0, "patch": 0}, "splits": {"test": {"name": "test", "num_bytes": 21682299, "num_examples": 200, "dataset_name": "long_bench"}}, "download_checksums": {"https://huggingface.co/datasets/THUDM/LongBench/resolve/main/data.zip": {"num_bytes": 113932529, "checksum": null}}, "download_size": 113932529, "dataset_size": 21682299, "size_in_bytes": 135614828}
|
long_bench/narrativeqa/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/long_bench-test.arrow
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4ce89fa82e1b8ef55139a3e0aa282014258d25bd5d82264713dc944ad8a1a04e
|
| 3 |
+
size 21684904
|
long_bench/narrativeqa/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e_builder.lock
ADDED
|
File without changes
|
long_bench/passage_retrieval_en/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e.incomplete_info.lock
ADDED
|
File without changes
|
long_bench/passage_retrieval_en/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/dataset_info.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"description": "LongBench is a comprehensive benchmark for multilingual and multi-task purposes, with the goal to fully measure and evaluate the ability of pre-trained language models to understand long text. This dataset consists of twenty different tasks, covering key long-text application scenarios such as multi-document QA, single-document QA, summarization, few-shot learning, synthetic tasks, and code completion.\n", "citation": "", "homepage": "https://github.com/THUDM/LongBench", "license": "", "features": {"input": {"dtype": "string", "_type": "Value"}, "context": {"dtype": "string", "_type": "Value"}, "answers": [{"dtype": "string", "_type": "Value"}], "length": {"dtype": "int32", "_type": "Value"}, "dataset": {"dtype": "string", "_type": "Value"}, "language": {"dtype": "string", "_type": "Value"}, "all_classes": [{"dtype": "string", "_type": "Value"}], "_id": {"dtype": "string", "_type": "Value"}}, "builder_name": "long_bench", "dataset_name": "long_bench", "config_name": "passage_retrieval_en", "version": {"version_str": "1.0.0", "major": 1, "minor": 0, "patch": 0}, "splits": {"test": {"name": "test", "num_bytes": 11234534, "num_examples": 200, "dataset_name": "long_bench"}}, "download_checksums": {"https://huggingface.co/datasets/THUDM/LongBench/resolve/main/data.zip": {"num_bytes": 113932529, "checksum": null}}, "download_size": 113932529, "dataset_size": 11234534, "size_in_bytes": 125167063}
|
long_bench/passage_retrieval_en/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/long_bench-test.arrow
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c3344a9db3c321c5132fae6db1cd2f266b7e090d041f7642913836bc8e4d478e
|
| 3 |
+
size 11237136
|
long_bench/passage_retrieval_en/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e_builder.lock
ADDED
|
File without changes
|
long_bench/samsum/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e.incomplete_info.lock
ADDED
|
File without changes
|
long_bench/samsum/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/dataset_info.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"description": "LongBench is a comprehensive benchmark for multilingual and multi-task purposes, with the goal to fully measure and evaluate the ability of pre-trained language models to understand long text. This dataset consists of twenty different tasks, covering key long-text application scenarios such as multi-document QA, single-document QA, summarization, few-shot learning, synthetic tasks, and code completion.\n", "citation": "", "homepage": "https://github.com/THUDM/LongBench", "license": "", "features": {"input": {"dtype": "string", "_type": "Value"}, "context": {"dtype": "string", "_type": "Value"}, "answers": [{"dtype": "string", "_type": "Value"}], "length": {"dtype": "int32", "_type": "Value"}, "dataset": {"dtype": "string", "_type": "Value"}, "language": {"dtype": "string", "_type": "Value"}, "all_classes": [{"dtype": "string", "_type": "Value"}], "_id": {"dtype": "string", "_type": "Value"}}, "builder_name": "long_bench", "dataset_name": "long_bench", "config_name": "samsum", "version": {"version_str": "1.0.0", "major": 1, "minor": 0, "patch": 0}, "splits": {"test": {"name": "test", "num_bytes": 6968716, "num_examples": 200, "dataset_name": "long_bench"}}, "download_checksums": {"https://huggingface.co/datasets/THUDM/LongBench/resolve/main/data.zip": {"num_bytes": 113932529, "checksum": null}}, "download_size": 113932529, "dataset_size": 6968716, "size_in_bytes": 120901245}
|
long_bench/samsum/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e/long_bench-test.arrow
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b8cd87cf5261d23b46a996227e1f707357a59f69d1674a8cbb8e5cc00a4cbc45
|
| 3 |
+
size 6971328
|
long_bench/samsum/1.0.0/4a916a4bde5c3481ac49b84d5dde69a9d2eefcd67f884ef65b3d97ee7cc91f3e_builder.lock
ADDED
|
File without changes
|