datasetId
stringlengths
2
117
card
stringlengths
19
1.01M
Ashe/ceshi
--- license: afl-3.0 ---
shahxeebhassan/UrduAssitant-llama2-2k
--- dataset_info: features: - name: text dtype: string splits: - name: train num_bytes: 2055296 num_examples: 2000 download_size: 982598 dataset_size: 2055296 configs: - config_name: default data_files: - split: train path: data/train-* ---
Timbrt/SciOL-CI
--- license: cc-by-4.0 language: - en size_categories: - 10M<n<100M pretty_name: Scientific Openly-Licensed Publications - Caption Images --- # Scientific Openly-Licensed Publications This repository contains companion material for the following [publication](https://openaccess.thecvf.com/content/WACV2024/papers/Tarsi_SciOL_and_MuLMS-Img_Introducing_a_Large-Scale_Multimodal_Scientific_Dataset_and_WACV_2024_paper.pdf): > Tim Tarsi, Heike Adel, Jan Hendrik Metzen, Dan Zhang, Matteo Finco, Annemarie Friedrich. **SciOL and MuLMS-Img: Introducing A Large-Scale Multimodal Scientific Dataset and Models for Image-Text Tasks in the Scientific Domain.** WACV 2024. Please cite this paper if using the dataset, and direct any questions regarding the dataset to [Tim Tarsi](mailto:tim.tarsi@gmail.com) ## Summary Scientific Openly-Licensed Publications (SciOL) is the largest openly-licensed pre-training corpus for multimodal models in the scientific domain, covering multiple sciences including materials science, physics, and computer science. It consists of over 2.7M scientific scientific publications converted into semi-structured data. SciOL contains over 18 Million figure-caption pairs. **Note: This repository only contains the figures and captions of SciOL. For the textual data see:** [SciOL-text](https://huggingface.co/datasets/Timbrt/SciOL-text) ## Data Format We provide the data in the webdataset format, e.g., captions in plain text files and group and compress them together with the images. Each tar file contains 1000 images and captions. Corresponding figures and captions have the same filename (excluding extention). We split the data into a train, test and dev set. ## Citation If you use our dataset in your work, please cite our paper: ``` @InProceedings{Tarsi_2024_WACV, author = {Tarsi, Tim and Adel, Heike and Metzen, Jan Hendrik and Zhang, Dan and Finco, Matteo and Friedrich, Annemarie}, title = {SciOL and MuLMS-Img: Introducing a Large-Scale Multimodal Scientific Dataset and Models for Image-Text Tasks in the Scientific Domain}, booktitle = {Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV)}, month = {January}, year = {2024}, pages = {4560-4571} } ``` ## License The SciOL corpus is released under the [CC BY 4.0](https://creativecommons.org/licenses/by/4.0/) license.
AlienKevin/yue-cmn-eng
--- language: - eng - yue - cmn task_categories: - translation task_ids: [] config_names: - eng-yue - eng-cmn - cmn-yue dataset_info: - config_name: eng-yue features: - name: translation dtype: translation: languages: - eng - yue splits: - name: test num_examples: 1500 - name: train num_examples: 53333 - config_name: eng-cmn features: - name: translation dtype: translation: languages: - eng - cmn splits: - name: test num_examples: 1500 - name: train num_examples: 47135 - config_name: cmn-yue features: - name: translation dtype: translation: languages: - cmn - yue splits: - name: test num_examples: 1500 - name: train num_examples: 11504 configs: - config_name: eng-yue data_files: - split: test path: eng-yue/test-* - split: train path: eng-yue/train-* - config_name: eng-cmn data_files: - split: test path: eng-cmn/test-* - split: train path: eng-cmn/train-* - config_name: cmn-yue data_files: - split: test path: cmn-yue/test-* - split: train path: cmn-yue/train-* ---
Julia0408/ucf101_julia
--- task_categories: - token-classification ---
luisrguerra/MetaMathQA-40K-GPT3.5
--- license: cc-by-sa-4.0 --- MetaMathQA-40K adapted to the GPT3.5 dataset format in JSONL for Fine-tuning. Following the following model: {"messages": [{"role": "system", "content": ""}, {"role": "user", "content": ""}, {"role": "assistant", "content": ""}]}
Seongill/Trivia_missing_5_small
--- dataset_info: features: - name: question dtype: string - name: answers sequence: string - name: ctxs list: - name: hasanswer dtype: bool - name: id dtype: string - name: score dtype: float64 - name: text dtype: string - name: title dtype: string - name: has_answer dtype: bool splits: - name: train num_bytes: 13673733.0 num_examples: 3771 download_size: 8296647 dataset_size: 13673733.0 configs: - config_name: default data_files: - split: train path: data/train-* ---
bys2058/test2
--- dataset_info: features: - name: image dtype: image - name: image_caption dtype: string splits: - name: train num_bytes: 632594898.5 num_examples: 1292 download_size: 632383145 dataset_size: 632594898.5 configs: - config_name: default data_files: - split: train path: data/train-* --- # Dataset Card for "test2" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
ChilleD/SVAMP
--- license: mit task_categories: - text-generation language: - en size_categories: - n<1K ---
316usman/thematic4a_rr
--- dataset_info: features: - name: text dtype: string - name: document_url dtype: string - name: source_url dtype: string - name: num_tokens dtype: int64 splits: - name: train num_bytes: 48735032.78893786 num_examples: 77497 download_size: 17957783 dataset_size: 48735032.78893786 configs: - config_name: default data_files: - split: train path: data/train-* ---
Trelis/touch-rugby-rules-unsupervised
--- task_categories: - text-generation language: - en tags: - fine-tuning - touch rugby size_categories: - n<1K --- # Touch Rugby Rules Dataset train.csv is taken from the [International Touch Website](https://cdn.internationaltouch.org/public/FIT%205th%20Edition%20Rulebook.pdf) All text is chunked to a length of 250 tokens, aiming to keep sentences whole where possible. For educational and non-commercial use only.
silatus/1k_Website_Screenshots_and_Metadata
--- license: cc-by-nc-sa-4.0 task_categories: - text-to-image - image-classification - image-segmentation language: - en tags: - screenshots - metadata - websites - webpages pretty_name: 1000 Website Screenshots with Metadata size_categories: - 1K<n<10K --- # Dataset Card for 1000 Website Screenshots with Metadata ## Dataset Description - **Homepage:** [silatus.com](https://silatus.com/datasets) - **Point of Contact:** [datasets@silatus.com](mailto:datasets@silatus.com) ### Dataset Summary Silatus is sharing, for free, a segment of a dataset that we are using to train a generative AI model for text-to-mockup conversions. This dataset was collected in December 2022 and early January 2023, so it contains very recent data from 1,000 of the world's most popular websites. You can get our larger 10,000 website dataset for free at: [https://silatus.com/datasets](https://silatus.com/datasets) This dataset includes: **High-res screenshots** - 1024x1024px - Loaded Javascript - Loaded Images **Text metadata** - Site title - Navbar content - Full page text data - Page description **Visual metadata** - Content (images, videos, inputs, buttons) absolute & relative positions - Color profile - Base font
open-llm-leaderboard/details_rwitz__go-bruins-v2
--- pretty_name: Evaluation run of rwitz/go-bruins-v2 dataset_summary: "Dataset automatically created during the evaluation run of model\ \ [rwitz/go-bruins-v2](https://huggingface.co/rwitz/go-bruins-v2) on the [Open LLM\ \ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\ \nThe dataset is composed of 63 configuration, each one coresponding to one of the\ \ evaluated task.\n\nThe dataset has been created from 2 run(s). Each run can be\ \ found as a specific split in each configuration, the split being named using the\ \ timestamp of the run.The \"train\" split is always pointing to the latest results.\n\ \nAn additional configuration \"results\" store all the aggregated results of the\ \ run (and is used to compute and display the aggregated metrics on the [Open LLM\ \ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\ \nTo load the details from a run, you can for instance do the following:\n```python\n\ from datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_rwitz__go-bruins-v2\"\ ,\n\t\"harness_winogrande_5\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\n\ These are the [latest results from run 2023-12-10T05:42:16.717744](https://huggingface.co/datasets/open-llm-leaderboard/details_rwitz__go-bruins-v2/blob/main/results_2023-12-10T05-42-16.717744.json)(note\ \ that their might be results for other tasks in the repos if successive evals didn't\ \ cover the same tasks. You find each in the results and the \"latest\" split for\ \ each eval):\n\n```python\n{\n \"all\": {\n \"acc\": 0.6521685007083396,\n\ \ \"acc_stderr\": 0.03205721368340006,\n \"acc_norm\": 0.6521344188001463,\n\ \ \"acc_norm_stderr\": 0.032717447545898726,\n \"mc1\": 0.4369645042839657,\n\ \ \"mc1_stderr\": 0.017363844503195974,\n \"mc2\": 0.5970340702765861,\n\ \ \"mc2_stderr\": 0.015540536389561436\n },\n \"harness|arc:challenge|25\"\ : {\n \"acc\": 0.6697952218430034,\n \"acc_stderr\": 0.013743085603760424,\n\ \ \"acc_norm\": 0.6979522184300341,\n \"acc_norm_stderr\": 0.01341751914471641\n\ \ },\n \"harness|hellaswag|10\": {\n \"acc\": 0.6937860983867755,\n\ \ \"acc_stderr\": 0.004599776866717491,\n \"acc_norm\": 0.8705437163911571,\n\ \ \"acc_norm_stderr\": 0.003350181812941604\n },\n \"harness|hendrycksTest-abstract_algebra|5\"\ : {\n \"acc\": 0.29,\n \"acc_stderr\": 0.045604802157206845,\n \ \ \"acc_norm\": 0.29,\n \"acc_norm_stderr\": 0.045604802157206845\n \ \ },\n \"harness|hendrycksTest-anatomy|5\": {\n \"acc\": 0.6666666666666666,\n\ \ \"acc_stderr\": 0.04072314811876837,\n \"acc_norm\": 0.6666666666666666,\n\ \ \"acc_norm_stderr\": 0.04072314811876837\n },\n \"harness|hendrycksTest-astronomy|5\"\ : {\n \"acc\": 0.6973684210526315,\n \"acc_stderr\": 0.037385206761196686,\n\ \ \"acc_norm\": 0.6973684210526315,\n \"acc_norm_stderr\": 0.037385206761196686\n\ \ },\n \"harness|hendrycksTest-business_ethics|5\": {\n \"acc\": 0.66,\n\ \ \"acc_stderr\": 0.04760952285695238,\n \"acc_norm\": 0.66,\n \ \ \"acc_norm_stderr\": 0.04760952285695238\n },\n \"harness|hendrycksTest-clinical_knowledge|5\"\ : {\n \"acc\": 0.7245283018867924,\n \"acc_stderr\": 0.027495663683724057,\n\ \ \"acc_norm\": 0.7245283018867924,\n \"acc_norm_stderr\": 0.027495663683724057\n\ \ },\n \"harness|hendrycksTest-college_biology|5\": {\n \"acc\": 0.75,\n\ \ \"acc_stderr\": 0.03621034121889507,\n \"acc_norm\": 0.75,\n \ \ \"acc_norm_stderr\": 0.03621034121889507\n },\n \"harness|hendrycksTest-college_chemistry|5\"\ : {\n \"acc\": 0.45,\n \"acc_stderr\": 0.05,\n \"acc_norm\"\ : 0.45,\n \"acc_norm_stderr\": 0.05\n },\n \"harness|hendrycksTest-college_computer_science|5\"\ : {\n \"acc\": 0.56,\n \"acc_stderr\": 0.04988876515698589,\n \ \ \"acc_norm\": 0.56,\n \"acc_norm_stderr\": 0.04988876515698589\n \ \ },\n \"harness|hendrycksTest-college_mathematics|5\": {\n \"acc\": 0.35,\n\ \ \"acc_stderr\": 0.04793724854411019,\n \"acc_norm\": 0.35,\n \ \ \"acc_norm_stderr\": 0.04793724854411019\n },\n \"harness|hendrycksTest-college_medicine|5\"\ : {\n \"acc\": 0.6589595375722543,\n \"acc_stderr\": 0.036146654241808254,\n\ \ \"acc_norm\": 0.6589595375722543,\n \"acc_norm_stderr\": 0.036146654241808254\n\ \ },\n \"harness|hendrycksTest-college_physics|5\": {\n \"acc\": 0.4411764705882353,\n\ \ \"acc_stderr\": 0.049406356306056595,\n \"acc_norm\": 0.4411764705882353,\n\ \ \"acc_norm_stderr\": 0.049406356306056595\n },\n \"harness|hendrycksTest-computer_security|5\"\ : {\n \"acc\": 0.76,\n \"acc_stderr\": 0.04292346959909283,\n \ \ \"acc_norm\": 0.76,\n \"acc_norm_stderr\": 0.04292346959909283\n \ \ },\n \"harness|hendrycksTest-conceptual_physics|5\": {\n \"acc\": 0.6,\n\ \ \"acc_stderr\": 0.03202563076101735,\n \"acc_norm\": 0.6,\n \ \ \"acc_norm_stderr\": 0.03202563076101735\n },\n \"harness|hendrycksTest-econometrics|5\"\ : {\n \"acc\": 0.47368421052631576,\n \"acc_stderr\": 0.046970851366478626,\n\ \ \"acc_norm\": 0.47368421052631576,\n \"acc_norm_stderr\": 0.046970851366478626\n\ \ },\n \"harness|hendrycksTest-electrical_engineering|5\": {\n \"acc\"\ : 0.5241379310344828,\n \"acc_stderr\": 0.0416180850350153,\n \"acc_norm\"\ : 0.5241379310344828,\n \"acc_norm_stderr\": 0.0416180850350153\n },\n\ \ \"harness|hendrycksTest-elementary_mathematics|5\": {\n \"acc\": 0.42592592592592593,\n\ \ \"acc_stderr\": 0.025467149045469553,\n \"acc_norm\": 0.42592592592592593,\n\ \ \"acc_norm_stderr\": 0.025467149045469553\n },\n \"harness|hendrycksTest-formal_logic|5\"\ : {\n \"acc\": 0.5158730158730159,\n \"acc_stderr\": 0.044698818540726076,\n\ \ \"acc_norm\": 0.5158730158730159,\n \"acc_norm_stderr\": 0.044698818540726076\n\ \ },\n \"harness|hendrycksTest-global_facts|5\": {\n \"acc\": 0.33,\n\ \ \"acc_stderr\": 0.047258156262526045,\n \"acc_norm\": 0.33,\n \ \ \"acc_norm_stderr\": 0.047258156262526045\n },\n \"harness|hendrycksTest-high_school_biology|5\"\ : {\n \"acc\": 0.7709677419354839,\n \"acc_stderr\": 0.02390491431178265,\n\ \ \"acc_norm\": 0.7709677419354839,\n \"acc_norm_stderr\": 0.02390491431178265\n\ \ },\n \"harness|hendrycksTest-high_school_chemistry|5\": {\n \"acc\"\ : 0.4876847290640394,\n \"acc_stderr\": 0.035169204442208966,\n \"\ acc_norm\": 0.4876847290640394,\n \"acc_norm_stderr\": 0.035169204442208966\n\ \ },\n \"harness|hendrycksTest-high_school_computer_science|5\": {\n \ \ \"acc\": 0.7,\n \"acc_stderr\": 0.046056618647183814,\n \"acc_norm\"\ : 0.7,\n \"acc_norm_stderr\": 0.046056618647183814\n },\n \"harness|hendrycksTest-high_school_european_history|5\"\ : {\n \"acc\": 0.7818181818181819,\n \"acc_stderr\": 0.03225078108306289,\n\ \ \"acc_norm\": 0.7818181818181819,\n \"acc_norm_stderr\": 0.03225078108306289\n\ \ },\n \"harness|hendrycksTest-high_school_geography|5\": {\n \"acc\"\ : 0.7878787878787878,\n \"acc_stderr\": 0.029126522834586818,\n \"\ acc_norm\": 0.7878787878787878,\n \"acc_norm_stderr\": 0.029126522834586818\n\ \ },\n \"harness|hendrycksTest-high_school_government_and_politics|5\": {\n\ \ \"acc\": 0.8911917098445595,\n \"acc_stderr\": 0.022473253332768763,\n\ \ \"acc_norm\": 0.8911917098445595,\n \"acc_norm_stderr\": 0.022473253332768763\n\ \ },\n \"harness|hendrycksTest-high_school_macroeconomics|5\": {\n \ \ \"acc\": 0.6692307692307692,\n \"acc_stderr\": 0.023854795680971128,\n\ \ \"acc_norm\": 0.6692307692307692,\n \"acc_norm_stderr\": 0.023854795680971128\n\ \ },\n \"harness|hendrycksTest-high_school_mathematics|5\": {\n \"\ acc\": 0.34074074074074073,\n \"acc_stderr\": 0.028897748741131154,\n \ \ \"acc_norm\": 0.34074074074074073,\n \"acc_norm_stderr\": 0.028897748741131154\n\ \ },\n \"harness|hendrycksTest-high_school_microeconomics|5\": {\n \ \ \"acc\": 0.6680672268907563,\n \"acc_stderr\": 0.03058869701378364,\n \ \ \"acc_norm\": 0.6680672268907563,\n \"acc_norm_stderr\": 0.03058869701378364\n\ \ },\n \"harness|hendrycksTest-high_school_physics|5\": {\n \"acc\"\ : 0.3443708609271523,\n \"acc_stderr\": 0.038796870240733264,\n \"\ acc_norm\": 0.3443708609271523,\n \"acc_norm_stderr\": 0.038796870240733264\n\ \ },\n \"harness|hendrycksTest-high_school_psychology|5\": {\n \"acc\"\ : 0.8495412844036697,\n \"acc_stderr\": 0.015328563932669237,\n \"\ acc_norm\": 0.8495412844036697,\n \"acc_norm_stderr\": 0.015328563932669237\n\ \ },\n \"harness|hendrycksTest-high_school_statistics|5\": {\n \"acc\"\ : 0.5324074074074074,\n \"acc_stderr\": 0.03402801581358966,\n \"\ acc_norm\": 0.5324074074074074,\n \"acc_norm_stderr\": 0.03402801581358966\n\ \ },\n \"harness|hendrycksTest-high_school_us_history|5\": {\n \"acc\"\ : 0.8186274509803921,\n \"acc_stderr\": 0.027044621719474082,\n \"\ acc_norm\": 0.8186274509803921,\n \"acc_norm_stderr\": 0.027044621719474082\n\ \ },\n \"harness|hendrycksTest-high_school_world_history|5\": {\n \"\ acc\": 0.8143459915611815,\n \"acc_stderr\": 0.025310495376944863,\n \ \ \"acc_norm\": 0.8143459915611815,\n \"acc_norm_stderr\": 0.025310495376944863\n\ \ },\n \"harness|hendrycksTest-human_aging|5\": {\n \"acc\": 0.695067264573991,\n\ \ \"acc_stderr\": 0.030898610882477515,\n \"acc_norm\": 0.695067264573991,\n\ \ \"acc_norm_stderr\": 0.030898610882477515\n },\n \"harness|hendrycksTest-human_sexuality|5\"\ : {\n \"acc\": 0.7786259541984732,\n \"acc_stderr\": 0.036412970813137276,\n\ \ \"acc_norm\": 0.7786259541984732,\n \"acc_norm_stderr\": 0.036412970813137276\n\ \ },\n \"harness|hendrycksTest-international_law|5\": {\n \"acc\":\ \ 0.7768595041322314,\n \"acc_stderr\": 0.03800754475228733,\n \"\ acc_norm\": 0.7768595041322314,\n \"acc_norm_stderr\": 0.03800754475228733\n\ \ },\n \"harness|hendrycksTest-jurisprudence|5\": {\n \"acc\": 0.7870370370370371,\n\ \ \"acc_stderr\": 0.0395783547198098,\n \"acc_norm\": 0.7870370370370371,\n\ \ \"acc_norm_stderr\": 0.0395783547198098\n },\n \"harness|hendrycksTest-logical_fallacies|5\"\ : {\n \"acc\": 0.7607361963190185,\n \"acc_stderr\": 0.0335195387952127,\n\ \ \"acc_norm\": 0.7607361963190185,\n \"acc_norm_stderr\": 0.0335195387952127\n\ \ },\n \"harness|hendrycksTest-machine_learning|5\": {\n \"acc\": 0.42857142857142855,\n\ \ \"acc_stderr\": 0.04697113923010212,\n \"acc_norm\": 0.42857142857142855,\n\ \ \"acc_norm_stderr\": 0.04697113923010212\n },\n \"harness|hendrycksTest-management|5\"\ : {\n \"acc\": 0.7961165048543689,\n \"acc_stderr\": 0.0398913985953177,\n\ \ \"acc_norm\": 0.7961165048543689,\n \"acc_norm_stderr\": 0.0398913985953177\n\ \ },\n \"harness|hendrycksTest-marketing|5\": {\n \"acc\": 0.8803418803418803,\n\ \ \"acc_stderr\": 0.021262719400406957,\n \"acc_norm\": 0.8803418803418803,\n\ \ \"acc_norm_stderr\": 0.021262719400406957\n },\n \"harness|hendrycksTest-medical_genetics|5\"\ : {\n \"acc\": 0.72,\n \"acc_stderr\": 0.045126085985421276,\n \ \ \"acc_norm\": 0.72,\n \"acc_norm_stderr\": 0.045126085985421276\n \ \ },\n \"harness|hendrycksTest-miscellaneous|5\": {\n \"acc\": 0.8237547892720306,\n\ \ \"acc_stderr\": 0.013625556907993457,\n \"acc_norm\": 0.8237547892720306,\n\ \ \"acc_norm_stderr\": 0.013625556907993457\n },\n \"harness|hendrycksTest-moral_disputes|5\"\ : {\n \"acc\": 0.7341040462427746,\n \"acc_stderr\": 0.023786203255508287,\n\ \ \"acc_norm\": 0.7341040462427746,\n \"acc_norm_stderr\": 0.023786203255508287\n\ \ },\n \"harness|hendrycksTest-moral_scenarios|5\": {\n \"acc\": 0.4324022346368715,\n\ \ \"acc_stderr\": 0.016568971233548606,\n \"acc_norm\": 0.4324022346368715,\n\ \ \"acc_norm_stderr\": 0.016568971233548606\n },\n \"harness|hendrycksTest-nutrition|5\"\ : {\n \"acc\": 0.7254901960784313,\n \"acc_stderr\": 0.025553169991826524,\n\ \ \"acc_norm\": 0.7254901960784313,\n \"acc_norm_stderr\": 0.025553169991826524\n\ \ },\n \"harness|hendrycksTest-philosophy|5\": {\n \"acc\": 0.6816720257234726,\n\ \ \"acc_stderr\": 0.02645722506781103,\n \"acc_norm\": 0.6816720257234726,\n\ \ \"acc_norm_stderr\": 0.02645722506781103\n },\n \"harness|hendrycksTest-prehistory|5\"\ : {\n \"acc\": 0.7407407407407407,\n \"acc_stderr\": 0.02438366553103545,\n\ \ \"acc_norm\": 0.7407407407407407,\n \"acc_norm_stderr\": 0.02438366553103545\n\ \ },\n \"harness|hendrycksTest-professional_accounting|5\": {\n \"\ acc\": 0.4787234042553192,\n \"acc_stderr\": 0.029800481645628693,\n \ \ \"acc_norm\": 0.4787234042553192,\n \"acc_norm_stderr\": 0.029800481645628693\n\ \ },\n \"harness|hendrycksTest-professional_law|5\": {\n \"acc\": 0.4634941329856584,\n\ \ \"acc_stderr\": 0.012736153390214961,\n \"acc_norm\": 0.4634941329856584,\n\ \ \"acc_norm_stderr\": 0.012736153390214961\n },\n \"harness|hendrycksTest-professional_medicine|5\"\ : {\n \"acc\": 0.6801470588235294,\n \"acc_stderr\": 0.02833295951403121,\n\ \ \"acc_norm\": 0.6801470588235294,\n \"acc_norm_stderr\": 0.02833295951403121\n\ \ },\n \"harness|hendrycksTest-professional_psychology|5\": {\n \"\ acc\": 0.6781045751633987,\n \"acc_stderr\": 0.01890101532209309,\n \ \ \"acc_norm\": 0.6781045751633987,\n \"acc_norm_stderr\": 0.01890101532209309\n\ \ },\n \"harness|hendrycksTest-public_relations|5\": {\n \"acc\": 0.6727272727272727,\n\ \ \"acc_stderr\": 0.0449429086625209,\n \"acc_norm\": 0.6727272727272727,\n\ \ \"acc_norm_stderr\": 0.0449429086625209\n },\n \"harness|hendrycksTest-security_studies|5\"\ : {\n \"acc\": 0.726530612244898,\n \"acc_stderr\": 0.02853556033712844,\n\ \ \"acc_norm\": 0.726530612244898,\n \"acc_norm_stderr\": 0.02853556033712844\n\ \ },\n \"harness|hendrycksTest-sociology|5\": {\n \"acc\": 0.8606965174129353,\n\ \ \"acc_stderr\": 0.024484487162913973,\n \"acc_norm\": 0.8606965174129353,\n\ \ \"acc_norm_stderr\": 0.024484487162913973\n },\n \"harness|hendrycksTest-us_foreign_policy|5\"\ : {\n \"acc\": 0.86,\n \"acc_stderr\": 0.0348735088019777,\n \ \ \"acc_norm\": 0.86,\n \"acc_norm_stderr\": 0.0348735088019777\n },\n\ \ \"harness|hendrycksTest-virology|5\": {\n \"acc\": 0.5481927710843374,\n\ \ \"acc_stderr\": 0.03874371556587953,\n \"acc_norm\": 0.5481927710843374,\n\ \ \"acc_norm_stderr\": 0.03874371556587953\n },\n \"harness|hendrycksTest-world_religions|5\"\ : {\n \"acc\": 0.8421052631578947,\n \"acc_stderr\": 0.027966785859160893,\n\ \ \"acc_norm\": 0.8421052631578947,\n \"acc_norm_stderr\": 0.027966785859160893\n\ \ },\n \"harness|truthfulqa:mc|0\": {\n \"mc1\": 0.4369645042839657,\n\ \ \"mc1_stderr\": 0.017363844503195974,\n \"mc2\": 0.5970340702765861,\n\ \ \"mc2_stderr\": 0.015540536389561436\n },\n \"harness|winogrande|5\"\ : {\n \"acc\": 0.8145224940805051,\n \"acc_stderr\": 0.010923965303140505\n\ \ },\n \"harness|gsm8k|5\": {\n \"acc\": 0.6967399545109931,\n \ \ \"acc_stderr\": 0.0126615026634187\n }\n}\n```" repo_url: https://huggingface.co/rwitz/go-bruins-v2 leaderboard_url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard point_of_contact: clementine@hf.co configs: - config_name: harness_arc_challenge_25 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|arc:challenge|25_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|arc:challenge|25_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|arc:challenge|25_2023-12-10T05-42-16.717744.parquet' - config_name: harness_gsm8k_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|gsm8k|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|gsm8k|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|gsm8k|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hellaswag_10 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hellaswag|10_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hellaswag|10_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hellaswag|10_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-abstract_algebra|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-anatomy|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-astronomy|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-business_ethics|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-college_biology|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-college_chemistry|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-college_computer_science|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-college_mathematics|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-college_medicine|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-college_physics|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-computer_security|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-conceptual_physics|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-econometrics|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-electrical_engineering|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-formal_logic|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-global_facts|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-high_school_biology|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-high_school_european_history|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-high_school_geography|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-high_school_physics|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-high_school_psychology|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-high_school_statistics|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-high_school_us_history|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-high_school_world_history|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-human_aging|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-human_sexuality|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-international_law|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-jurisprudence|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-logical_fallacies|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-machine_learning|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-management|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-marketing|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-medical_genetics|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-miscellaneous|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-moral_disputes|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-moral_scenarios|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-nutrition|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-philosophy|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-prehistory|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-professional_accounting|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-professional_law|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-professional_medicine|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-professional_psychology|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-public_relations|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-security_studies|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-sociology|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-virology|5_2023-12-10T05-36-09.275219.parquet' - '**/details_harness|hendrycksTest-world_religions|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-abstract_algebra|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-anatomy|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-astronomy|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-business_ethics|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-college_biology|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-college_chemistry|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-college_computer_science|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-college_mathematics|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-college_medicine|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-college_physics|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-computer_security|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-conceptual_physics|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-econometrics|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-electrical_engineering|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-formal_logic|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-global_facts|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_biology|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_european_history|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_geography|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_physics|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_psychology|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_statistics|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_us_history|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_world_history|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-human_aging|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-human_sexuality|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-international_law|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-jurisprudence|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-logical_fallacies|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-machine_learning|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-management|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-marketing|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-medical_genetics|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-miscellaneous|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-moral_disputes|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-moral_scenarios|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-nutrition|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-philosophy|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-prehistory|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-professional_accounting|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-professional_law|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-professional_medicine|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-professional_psychology|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-public_relations|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-security_studies|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-sociology|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-virology|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-world_religions|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-abstract_algebra|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-anatomy|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-astronomy|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-business_ethics|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-college_biology|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-college_chemistry|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-college_computer_science|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-college_mathematics|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-college_medicine|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-college_physics|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-computer_security|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-conceptual_physics|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-econometrics|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-electrical_engineering|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-formal_logic|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-global_facts|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_biology|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_european_history|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_geography|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_physics|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_psychology|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_statistics|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_us_history|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-high_school_world_history|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-human_aging|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-human_sexuality|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-international_law|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-jurisprudence|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-logical_fallacies|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-machine_learning|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-management|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-marketing|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-medical_genetics|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-miscellaneous|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-moral_disputes|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-moral_scenarios|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-nutrition|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-philosophy|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-prehistory|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-professional_accounting|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-professional_law|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-professional_medicine|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-professional_psychology|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-public_relations|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-security_studies|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-sociology|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-virology|5_2023-12-10T05-42-16.717744.parquet' - '**/details_harness|hendrycksTest-world_religions|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_abstract_algebra_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-abstract_algebra|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-abstract_algebra|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-abstract_algebra|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_anatomy_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-anatomy|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-anatomy|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-anatomy|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_astronomy_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-astronomy|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-astronomy|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-astronomy|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_business_ethics_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-business_ethics|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-business_ethics|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-business_ethics|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_clinical_knowledge_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_college_biology_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-college_biology|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-college_biology|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-college_biology|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_college_chemistry_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-college_chemistry|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-college_chemistry|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-college_chemistry|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_college_computer_science_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-college_computer_science|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-college_computer_science|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-college_computer_science|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_college_mathematics_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-college_mathematics|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-college_mathematics|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-college_mathematics|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_college_medicine_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-college_medicine|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-college_medicine|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-college_medicine|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_college_physics_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-college_physics|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-college_physics|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-college_physics|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_computer_security_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-computer_security|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-computer_security|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-computer_security|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_conceptual_physics_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-conceptual_physics|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-conceptual_physics|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-conceptual_physics|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_econometrics_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-econometrics|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-econometrics|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-econometrics|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_electrical_engineering_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-electrical_engineering|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-electrical_engineering|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-electrical_engineering|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_elementary_mathematics_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_formal_logic_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-formal_logic|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-formal_logic|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-formal_logic|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_global_facts_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-global_facts|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-global_facts|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-global_facts|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_high_school_biology_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-high_school_biology|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-high_school_biology|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_biology|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_high_school_chemistry_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_high_school_computer_science_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_high_school_european_history_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-high_school_european_history|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-high_school_european_history|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_european_history|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_high_school_geography_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-high_school_geography|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-high_school_geography|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_geography|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_high_school_government_and_politics_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_high_school_macroeconomics_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_high_school_mathematics_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_high_school_microeconomics_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_high_school_physics_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-high_school_physics|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-high_school_physics|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_physics|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_high_school_psychology_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-high_school_psychology|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-high_school_psychology|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_psychology|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_high_school_statistics_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-high_school_statistics|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-high_school_statistics|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_statistics|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_high_school_us_history_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-high_school_us_history|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-high_school_us_history|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_us_history|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_high_school_world_history_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-high_school_world_history|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-high_school_world_history|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_world_history|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_human_aging_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-human_aging|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-human_aging|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-human_aging|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_human_sexuality_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-human_sexuality|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-human_sexuality|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-human_sexuality|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_international_law_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-international_law|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-international_law|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-international_law|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_jurisprudence_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-jurisprudence|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-jurisprudence|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-jurisprudence|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_logical_fallacies_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-logical_fallacies|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-logical_fallacies|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-logical_fallacies|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_machine_learning_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-machine_learning|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-machine_learning|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-machine_learning|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_management_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-management|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-management|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-management|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_marketing_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-marketing|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-marketing|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-marketing|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_medical_genetics_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-medical_genetics|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-medical_genetics|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-medical_genetics|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_miscellaneous_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-miscellaneous|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-miscellaneous|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-miscellaneous|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_moral_disputes_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-moral_disputes|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-moral_disputes|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-moral_disputes|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_moral_scenarios_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-moral_scenarios|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-moral_scenarios|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-moral_scenarios|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_nutrition_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-nutrition|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-nutrition|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-nutrition|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_philosophy_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-philosophy|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-philosophy|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-philosophy|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_prehistory_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-prehistory|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-prehistory|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-prehistory|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_professional_accounting_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-professional_accounting|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-professional_accounting|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-professional_accounting|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_professional_law_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-professional_law|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-professional_law|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-professional_law|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_professional_medicine_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-professional_medicine|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-professional_medicine|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-professional_medicine|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_professional_psychology_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-professional_psychology|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-professional_psychology|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-professional_psychology|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_public_relations_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-public_relations|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-public_relations|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-public_relations|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_security_studies_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-security_studies|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-security_studies|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-security_studies|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_sociology_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-sociology|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-sociology|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-sociology|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_us_foreign_policy_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_virology_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-virology|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-virology|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-virology|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_hendrycksTest_world_religions_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|hendrycksTest-world_religions|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|hendrycksTest-world_religions|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|hendrycksTest-world_religions|5_2023-12-10T05-42-16.717744.parquet' - config_name: harness_truthfulqa_mc_0 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|truthfulqa:mc|0_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|truthfulqa:mc|0_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|truthfulqa:mc|0_2023-12-10T05-42-16.717744.parquet' - config_name: harness_winogrande_5 data_files: - split: 2023_12_10T05_36_09.275219 path: - '**/details_harness|winogrande|5_2023-12-10T05-36-09.275219.parquet' - split: 2023_12_10T05_42_16.717744 path: - '**/details_harness|winogrande|5_2023-12-10T05-42-16.717744.parquet' - split: latest path: - '**/details_harness|winogrande|5_2023-12-10T05-42-16.717744.parquet' - config_name: results data_files: - split: 2023_12_10T05_36_09.275219 path: - results_2023-12-10T05-36-09.275219.parquet - split: 2023_12_10T05_42_16.717744 path: - results_2023-12-10T05-42-16.717744.parquet - split: latest path: - results_2023-12-10T05-42-16.717744.parquet --- # Dataset Card for Evaluation run of rwitz/go-bruins-v2 ## Dataset Description - **Homepage:** - **Repository:** https://huggingface.co/rwitz/go-bruins-v2 - **Paper:** - **Leaderboard:** https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard - **Point of Contact:** clementine@hf.co ### Dataset Summary Dataset automatically created during the evaluation run of model [rwitz/go-bruins-v2](https://huggingface.co/rwitz/go-bruins-v2) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard). The dataset is composed of 63 configuration, each one coresponding to one of the evaluated task. The dataset has been created from 2 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results. An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the aggregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)). To load the details from a run, you can for instance do the following: ```python from datasets import load_dataset data = load_dataset("open-llm-leaderboard/details_rwitz__go-bruins-v2", "harness_winogrande_5", split="train") ``` ## Latest results These are the [latest results from run 2023-12-10T05:42:16.717744](https://huggingface.co/datasets/open-llm-leaderboard/details_rwitz__go-bruins-v2/blob/main/results_2023-12-10T05-42-16.717744.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval): ```python { "all": { "acc": 0.6521685007083396, "acc_stderr": 0.03205721368340006, "acc_norm": 0.6521344188001463, "acc_norm_stderr": 0.032717447545898726, "mc1": 0.4369645042839657, "mc1_stderr": 0.017363844503195974, "mc2": 0.5970340702765861, "mc2_stderr": 0.015540536389561436 }, "harness|arc:challenge|25": { "acc": 0.6697952218430034, "acc_stderr": 0.013743085603760424, "acc_norm": 0.6979522184300341, "acc_norm_stderr": 0.01341751914471641 }, "harness|hellaswag|10": { "acc": 0.6937860983867755, "acc_stderr": 0.004599776866717491, "acc_norm": 0.8705437163911571, "acc_norm_stderr": 0.003350181812941604 }, "harness|hendrycksTest-abstract_algebra|5": { "acc": 0.29, "acc_stderr": 0.045604802157206845, "acc_norm": 0.29, "acc_norm_stderr": 0.045604802157206845 }, "harness|hendrycksTest-anatomy|5": { "acc": 0.6666666666666666, "acc_stderr": 0.04072314811876837, "acc_norm": 0.6666666666666666, "acc_norm_stderr": 0.04072314811876837 }, "harness|hendrycksTest-astronomy|5": { "acc": 0.6973684210526315, "acc_stderr": 0.037385206761196686, "acc_norm": 0.6973684210526315, "acc_norm_stderr": 0.037385206761196686 }, "harness|hendrycksTest-business_ethics|5": { "acc": 0.66, "acc_stderr": 0.04760952285695238, "acc_norm": 0.66, "acc_norm_stderr": 0.04760952285695238 }, "harness|hendrycksTest-clinical_knowledge|5": { "acc": 0.7245283018867924, "acc_stderr": 0.027495663683724057, "acc_norm": 0.7245283018867924, "acc_norm_stderr": 0.027495663683724057 }, "harness|hendrycksTest-college_biology|5": { "acc": 0.75, "acc_stderr": 0.03621034121889507, "acc_norm": 0.75, "acc_norm_stderr": 0.03621034121889507 }, "harness|hendrycksTest-college_chemistry|5": { "acc": 0.45, "acc_stderr": 0.05, "acc_norm": 0.45, "acc_norm_stderr": 0.05 }, "harness|hendrycksTest-college_computer_science|5": { "acc": 0.56, "acc_stderr": 0.04988876515698589, "acc_norm": 0.56, "acc_norm_stderr": 0.04988876515698589 }, "harness|hendrycksTest-college_mathematics|5": { "acc": 0.35, "acc_stderr": 0.04793724854411019, "acc_norm": 0.35, "acc_norm_stderr": 0.04793724854411019 }, "harness|hendrycksTest-college_medicine|5": { "acc": 0.6589595375722543, "acc_stderr": 0.036146654241808254, "acc_norm": 0.6589595375722543, "acc_norm_stderr": 0.036146654241808254 }, "harness|hendrycksTest-college_physics|5": { "acc": 0.4411764705882353, "acc_stderr": 0.049406356306056595, "acc_norm": 0.4411764705882353, "acc_norm_stderr": 0.049406356306056595 }, "harness|hendrycksTest-computer_security|5": { "acc": 0.76, "acc_stderr": 0.04292346959909283, "acc_norm": 0.76, "acc_norm_stderr": 0.04292346959909283 }, "harness|hendrycksTest-conceptual_physics|5": { "acc": 0.6, "acc_stderr": 0.03202563076101735, "acc_norm": 0.6, "acc_norm_stderr": 0.03202563076101735 }, "harness|hendrycksTest-econometrics|5": { "acc": 0.47368421052631576, "acc_stderr": 0.046970851366478626, "acc_norm": 0.47368421052631576, "acc_norm_stderr": 0.046970851366478626 }, "harness|hendrycksTest-electrical_engineering|5": { "acc": 0.5241379310344828, "acc_stderr": 0.0416180850350153, "acc_norm": 0.5241379310344828, "acc_norm_stderr": 0.0416180850350153 }, "harness|hendrycksTest-elementary_mathematics|5": { "acc": 0.42592592592592593, "acc_stderr": 0.025467149045469553, "acc_norm": 0.42592592592592593, "acc_norm_stderr": 0.025467149045469553 }, "harness|hendrycksTest-formal_logic|5": { "acc": 0.5158730158730159, "acc_stderr": 0.044698818540726076, "acc_norm": 0.5158730158730159, "acc_norm_stderr": 0.044698818540726076 }, "harness|hendrycksTest-global_facts|5": { "acc": 0.33, "acc_stderr": 0.047258156262526045, "acc_norm": 0.33, "acc_norm_stderr": 0.047258156262526045 }, "harness|hendrycksTest-high_school_biology|5": { "acc": 0.7709677419354839, "acc_stderr": 0.02390491431178265, "acc_norm": 0.7709677419354839, "acc_norm_stderr": 0.02390491431178265 }, "harness|hendrycksTest-high_school_chemistry|5": { "acc": 0.4876847290640394, "acc_stderr": 0.035169204442208966, "acc_norm": 0.4876847290640394, "acc_norm_stderr": 0.035169204442208966 }, "harness|hendrycksTest-high_school_computer_science|5": { "acc": 0.7, "acc_stderr": 0.046056618647183814, "acc_norm": 0.7, "acc_norm_stderr": 0.046056618647183814 }, "harness|hendrycksTest-high_school_european_history|5": { "acc": 0.7818181818181819, "acc_stderr": 0.03225078108306289, "acc_norm": 0.7818181818181819, "acc_norm_stderr": 0.03225078108306289 }, "harness|hendrycksTest-high_school_geography|5": { "acc": 0.7878787878787878, "acc_stderr": 0.029126522834586818, "acc_norm": 0.7878787878787878, "acc_norm_stderr": 0.029126522834586818 }, "harness|hendrycksTest-high_school_government_and_politics|5": { "acc": 0.8911917098445595, "acc_stderr": 0.022473253332768763, "acc_norm": 0.8911917098445595, "acc_norm_stderr": 0.022473253332768763 }, "harness|hendrycksTest-high_school_macroeconomics|5": { "acc": 0.6692307692307692, "acc_stderr": 0.023854795680971128, "acc_norm": 0.6692307692307692, "acc_norm_stderr": 0.023854795680971128 }, "harness|hendrycksTest-high_school_mathematics|5": { "acc": 0.34074074074074073, "acc_stderr": 0.028897748741131154, "acc_norm": 0.34074074074074073, "acc_norm_stderr": 0.028897748741131154 }, "harness|hendrycksTest-high_school_microeconomics|5": { "acc": 0.6680672268907563, "acc_stderr": 0.03058869701378364, "acc_norm": 0.6680672268907563, "acc_norm_stderr": 0.03058869701378364 }, "harness|hendrycksTest-high_school_physics|5": { "acc": 0.3443708609271523, "acc_stderr": 0.038796870240733264, "acc_norm": 0.3443708609271523, "acc_norm_stderr": 0.038796870240733264 }, "harness|hendrycksTest-high_school_psychology|5": { "acc": 0.8495412844036697, "acc_stderr": 0.015328563932669237, "acc_norm": 0.8495412844036697, "acc_norm_stderr": 0.015328563932669237 }, "harness|hendrycksTest-high_school_statistics|5": { "acc": 0.5324074074074074, "acc_stderr": 0.03402801581358966, "acc_norm": 0.5324074074074074, "acc_norm_stderr": 0.03402801581358966 }, "harness|hendrycksTest-high_school_us_history|5": { "acc": 0.8186274509803921, "acc_stderr": 0.027044621719474082, "acc_norm": 0.8186274509803921, "acc_norm_stderr": 0.027044621719474082 }, "harness|hendrycksTest-high_school_world_history|5": { "acc": 0.8143459915611815, "acc_stderr": 0.025310495376944863, "acc_norm": 0.8143459915611815, "acc_norm_stderr": 0.025310495376944863 }, "harness|hendrycksTest-human_aging|5": { "acc": 0.695067264573991, "acc_stderr": 0.030898610882477515, "acc_norm": 0.695067264573991, "acc_norm_stderr": 0.030898610882477515 }, "harness|hendrycksTest-human_sexuality|5": { "acc": 0.7786259541984732, "acc_stderr": 0.036412970813137276, "acc_norm": 0.7786259541984732, "acc_norm_stderr": 0.036412970813137276 }, "harness|hendrycksTest-international_law|5": { "acc": 0.7768595041322314, "acc_stderr": 0.03800754475228733, "acc_norm": 0.7768595041322314, "acc_norm_stderr": 0.03800754475228733 }, "harness|hendrycksTest-jurisprudence|5": { "acc": 0.7870370370370371, "acc_stderr": 0.0395783547198098, "acc_norm": 0.7870370370370371, "acc_norm_stderr": 0.0395783547198098 }, "harness|hendrycksTest-logical_fallacies|5": { "acc": 0.7607361963190185, "acc_stderr": 0.0335195387952127, "acc_norm": 0.7607361963190185, "acc_norm_stderr": 0.0335195387952127 }, "harness|hendrycksTest-machine_learning|5": { "acc": 0.42857142857142855, "acc_stderr": 0.04697113923010212, "acc_norm": 0.42857142857142855, "acc_norm_stderr": 0.04697113923010212 }, "harness|hendrycksTest-management|5": { "acc": 0.7961165048543689, "acc_stderr": 0.0398913985953177, "acc_norm": 0.7961165048543689, "acc_norm_stderr": 0.0398913985953177 }, "harness|hendrycksTest-marketing|5": { "acc": 0.8803418803418803, "acc_stderr": 0.021262719400406957, "acc_norm": 0.8803418803418803, "acc_norm_stderr": 0.021262719400406957 }, "harness|hendrycksTest-medical_genetics|5": { "acc": 0.72, "acc_stderr": 0.045126085985421276, "acc_norm": 0.72, "acc_norm_stderr": 0.045126085985421276 }, "harness|hendrycksTest-miscellaneous|5": { "acc": 0.8237547892720306, "acc_stderr": 0.013625556907993457, "acc_norm": 0.8237547892720306, "acc_norm_stderr": 0.013625556907993457 }, "harness|hendrycksTest-moral_disputes|5": { "acc": 0.7341040462427746, "acc_stderr": 0.023786203255508287, "acc_norm": 0.7341040462427746, "acc_norm_stderr": 0.023786203255508287 }, "harness|hendrycksTest-moral_scenarios|5": { "acc": 0.4324022346368715, "acc_stderr": 0.016568971233548606, "acc_norm": 0.4324022346368715, "acc_norm_stderr": 0.016568971233548606 }, "harness|hendrycksTest-nutrition|5": { "acc": 0.7254901960784313, "acc_stderr": 0.025553169991826524, "acc_norm": 0.7254901960784313, "acc_norm_stderr": 0.025553169991826524 }, "harness|hendrycksTest-philosophy|5": { "acc": 0.6816720257234726, "acc_stderr": 0.02645722506781103, "acc_norm": 0.6816720257234726, "acc_norm_stderr": 0.02645722506781103 }, "harness|hendrycksTest-prehistory|5": { "acc": 0.7407407407407407, "acc_stderr": 0.02438366553103545, "acc_norm": 0.7407407407407407, "acc_norm_stderr": 0.02438366553103545 }, "harness|hendrycksTest-professional_accounting|5": { "acc": 0.4787234042553192, "acc_stderr": 0.029800481645628693, "acc_norm": 0.4787234042553192, "acc_norm_stderr": 0.029800481645628693 }, "harness|hendrycksTest-professional_law|5": { "acc": 0.4634941329856584, "acc_stderr": 0.012736153390214961, "acc_norm": 0.4634941329856584, "acc_norm_stderr": 0.012736153390214961 }, "harness|hendrycksTest-professional_medicine|5": { "acc": 0.6801470588235294, "acc_stderr": 0.02833295951403121, "acc_norm": 0.6801470588235294, "acc_norm_stderr": 0.02833295951403121 }, "harness|hendrycksTest-professional_psychology|5": { "acc": 0.6781045751633987, "acc_stderr": 0.01890101532209309, "acc_norm": 0.6781045751633987, "acc_norm_stderr": 0.01890101532209309 }, "harness|hendrycksTest-public_relations|5": { "acc": 0.6727272727272727, "acc_stderr": 0.0449429086625209, "acc_norm": 0.6727272727272727, "acc_norm_stderr": 0.0449429086625209 }, "harness|hendrycksTest-security_studies|5": { "acc": 0.726530612244898, "acc_stderr": 0.02853556033712844, "acc_norm": 0.726530612244898, "acc_norm_stderr": 0.02853556033712844 }, "harness|hendrycksTest-sociology|5": { "acc": 0.8606965174129353, "acc_stderr": 0.024484487162913973, "acc_norm": 0.8606965174129353, "acc_norm_stderr": 0.024484487162913973 }, "harness|hendrycksTest-us_foreign_policy|5": { "acc": 0.86, "acc_stderr": 0.0348735088019777, "acc_norm": 0.86, "acc_norm_stderr": 0.0348735088019777 }, "harness|hendrycksTest-virology|5": { "acc": 0.5481927710843374, "acc_stderr": 0.03874371556587953, "acc_norm": 0.5481927710843374, "acc_norm_stderr": 0.03874371556587953 }, "harness|hendrycksTest-world_religions|5": { "acc": 0.8421052631578947, "acc_stderr": 0.027966785859160893, "acc_norm": 0.8421052631578947, "acc_norm_stderr": 0.027966785859160893 }, "harness|truthfulqa:mc|0": { "mc1": 0.4369645042839657, "mc1_stderr": 0.017363844503195974, "mc2": 0.5970340702765861, "mc2_stderr": 0.015540536389561436 }, "harness|winogrande|5": { "acc": 0.8145224940805051, "acc_stderr": 0.010923965303140505 }, "harness|gsm8k|5": { "acc": 0.6967399545109931, "acc_stderr": 0.0126615026634187 } } ``` ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions [More Information Needed]
Reza8848/MUFFIN_68k
--- language: - en size_categories: - 10K<n<100K --- <img src="https://cdn-uploads.huggingface.co/production/uploads/6434a6e8ea46c009904c617e/J_4FHXmtM6TuRnN3aL06y.png" width="38" height="38"> This is the training dataset of **[MUFFIN](https://arxiv.org/abs/2312.02436)** (**Mu**lti-**F**aceted **In**structions). Please refer to our project website for more details: [Website](https://renzelou.github.io/Muffin/) ## JSON Format The download data can be read as a Python list. In this list, each elemental Python dictionary has one input text. This input text has multiple task instructions and the corresponding outputs. ```json [ { "input": "XXX", "instances": [ { "instruction": "III", "output": "YYY" }, { "instruction": "III", "output": "YYY" } ] } , { "input": "XXX", "instances": [ { "instruction": "III", "output": "YYY" } ] } ] ``` ## Data Statistics There are a total of 1,463 input texts, where each input is equipped with multiple task instructions (~46.48 instructions per input), resulting in **68,014** training instances in total. The detailed statistics are shown below: <div style="text-align:center"><img src="https://cdn-uploads.huggingface.co/production/uploads/6434a6e8ea46c009904c617e/hcQjRr1TqX08C4tMnEQaZ.png" alt="statistics.png" width="500"/></div> ## 🥳 Citation Please kindly cite our paper if you use our dataset: ```bibtex @inproceedings{Lou2023MUFFIN, title={{MUFFIN}: Curating Multi-Faceted Instructions for Improving Instruction Following}, author={Renze Lou and Kai Zhang and Jian Xie and Yuxuan Sun and Janice Ahn and Hanzi Xu and Yu su and Wenpeng Yin}, booktitle={The Twelfth International Conference on Learning Representations}, year={2024}, url={https://openreview.net/forum?id=1vrS1zwekw} } ```
jjjlyn/prohunt-240219
--- license: apache-2.0 --- # Skills - Wanted (2024) - Job Planet (2024) - General (GPT-4 formatted) # JDs - General (GPT-4 formatted)
seokochin/shopify_QnA
--- license: apache-2.0 ---
Quake24/sumTwitter
--- license: apache-2.0 ---
nitrosocke/arcane-diffusion-dataset
--- license: creativeml-openrail-m --- # Arcane Diffusion Dataset Dataset containing the 75 images used to train the [Arcane Diffusion](https://huggingface.co/nitrosocke/Arcane-Diffusion) model. Settings for training: ```class prompt: illustration style instance prompt: illustration arcane style learning rate: 5e-6 lr scheduler: constant num class images: 1000 max train steps: 5000 ```
awettig/Pile-YoutubeSubtitles-0.5B-8K-opt
--- dataset_info: features: - name: input_ids sequence: int32 - name: attention_mask sequence: int8 - name: labels sequence: int64 splits: - name: train num_bytes: 6499875083 num_examples: 61035 - name: test num_bytes: 64969880 num_examples: 610 download_size: 1557923609 dataset_size: 6564844963 --- # Dataset Card for "Pile-YoutubeSubtitles-0.5B-8K-opt" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
results-sd-v1-5-sd-v2-1-if-v1-0-karlo/d042346d
--- dataset_info: features: - name: result dtype: string - name: id dtype: int64 splits: - name: train num_bytes: 182 num_examples: 10 download_size: 1342 dataset_size: 182 --- # Dataset Card for "d042346d" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
nateraw/fuego-20230213-200026-76714f
--- tags: - fuego fuego: id: 20230213-200026-76714f status: done script: main.py requirements_file: requirements.txt space_id: nateraw/fuego-20230213-200026-76714f space_hardware: t4-small github_repo_id: pytorch/examples github_repo_branch: main github_repo_sha: e4e8da8467d55d28920dbd137261d82255f68c71 ---
w95/megachat-zephyr
--- license: mit task_categories: - conversational - text-generation language: - en size_categories: - 1M<n<10M pretty_name: MegaChat ---
Columbia-NLP/SlimOrca-Dedup-multiturn
--- dataset_info: features: - name: messages list: - name: content dtype: string - name: role dtype: string splits: - name: train_sft num_bytes: 543422894 num_examples: 40397 download_size: 296215447 dataset_size: 543422894 configs: - config_name: default data_files: - split: train_sft path: data/train_sft-* ---
kotoba-speech/ThuVienThanhPhoBacGiang_tscribed_testing
--- dataset_info: features: - name: audio dtype: audio - name: text dtype: string - name: duration dtype: float64 - name: ratio dtype: float64 - name: videoid dtype: string - name: key dtype: string - name: dataset_id dtype: string - name: lang dtype: 'null' - name: start dtype: float64 - name: end dtype: float64 splits: - name: train num_bytes: 925385845.589 num_examples: 1773 download_size: 853956087 dataset_size: 925385845.589 configs: - config_name: default data_files: - split: train path: data/train-* ---
ranWang/UN_Historical_PDF_Article_Text_Corpus
--- dataset_info: features: - name: zh dtype: string - name: ar dtype: string - name: en dtype: string - name: fr dtype: string - name: es dtype: string - name: ru dtype: string - name: record dtype: string splits: - name: train num_bytes: 40392264949 num_examples: 142933 - name: randomTest num_bytes: 791067667 num_examples: 2839 download_size: 4512532002 dataset_size: 41183332616 --- # python ```python dataset = load_dataset("ranWang/UN_Historical_PDF_Article_Text_Corpus", split="train") or dataset = load_dataset("ranWang/UN_Historical_PDF_Article_Text_Corpus", split="randomTest") lang_list = ["ar", "en", "es", "fr", "ru", "zh"] for row in dataset: # 获取pdf文章内容 for lang in lang_list: # type == str lang_match_file_content = row[lang] # 如果按页分割 lang_match_file_pages_content = lang_match_file_content.split("\n----\n") ```
sridharps2/llama2_finetuned_chatbot
--- dataset_info: features: - name: text dtype: string splits: - name: train num_bytes: 15808615 num_examples: 9846 download_size: 9205969 dataset_size: 15808615 configs: - config_name: default data_files: - split: train path: data/train-* ---
ZhangShenao/0.00045_idpo_noreplacerej_decalpha_dataset
--- dataset_info: features: - name: prompt dtype: string - name: prompt_id dtype: string - name: messages list: - name: content dtype: string - name: role dtype: string - name: score_chosen dtype: float64 - name: score_rejected dtype: float64 - name: chosen list: - name: content dtype: string - name: role dtype: string - name: rejected list: - name: content dtype: string - name: role dtype: string - name: is_better dtype: bool splits: - name: test_prefs_1 num_bytes: 13659360 num_examples: 2000 - name: train_prefs_1 num_bytes: 140517334 num_examples: 20378 download_size: 85759145 dataset_size: 154176694 configs: - config_name: default data_files: - split: test_prefs_1 path: data/test_prefs_1-* - split: train_prefs_1 path: data/train_prefs_1-* --- # Dataset Card for "0.00045_idpo_noreplacerej_decalpha_dataset" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
Sakonii/nepalitext-language-model-dataset
--- annotations_creators: - no-annotation language_creators: - found - other language: - ne license: - cc0-1.0 multilinguality: - monolingual source_datasets: - extended|oscar - extended|cc100 task_categories: - text-generation task_ids: - language-modeling pretty_name: nepalitext-language-model-dataset --- # Dataset Card for "nepalitext-language-model-dataset" ### Dataset Summary "NepaliText" language modeling dataset is a collection of over 13 million Nepali text sequences (phrases/sentences/paragraphs) extracted by combining the datasets: [OSCAR](https://huggingface.co/datasets/oscar) , [cc100](https://huggingface.co/datasets/cc100) and a set of scraped Nepali articles on Wikipedia. ### Supported Tasks and Leaderboards This dataset is intended to pre-train language models and word representations on Nepali Language. ### Languages The data is focused on Nepali language, but may have instances of other languages as well. ## Dataset Structure ### Data Instances An example: ``` {'text': 'घरेलु मैदानमा भएको च्याम्पियन्स लिगको दोस्रो लेगमा एथ्लेटिको मड्रिडले आर्सनललाई एक शून्यले हराउँदै समग्रमा दुई एकको अग्रताका साथ फाइनलमा प्रवेश गरेको हो ।\n'} ``` ### Data Fields The data fields are: - `text`: a `string` feature. ### Data Splits train|test| ----:|---:| 13141222|268189| ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations The dataset does not contain any additional annotations. #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information Being extracted and scraped from variety of internet sources, Personal and sensitive information might be present. This must be considered before training deep learning models, specially in the case of text-generation models. ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions Thanks to [@Sakonii](https://github.com/Sakonii) for adding this dataset.
betajuned/SQUADelektrounila
--- task_categories: - question-answering language: - id tags: - not-for-all-audiences size_categories: - n<1K ---
Aliquip/sd-prompttastic
--- license: cc0-1.0 ---
CyberHarem/asahi_rokka_bangdreamdai2ki
--- license: mit task_categories: - text-to-image tags: - art - not-for-all-audiences size_categories: - n<1K --- # Dataset of asahi_rokka/朝日六花 (BanG Dream! Dai 2-ki) This is the dataset of asahi_rokka/朝日六花 (BanG Dream! Dai 2-ki), containing 87 images and their tags. The core tags of this character are `green_eyes, long_hair, bangs, blue_hair, hair_between_eyes, hair_ornament`, which are pruned in this dataset. Images are crawled from many sites (e.g. danbooru, pixiv, zerochan ...), the auto-crawling system is powered by [DeepGHS Team](https://github.com/deepghs)([huggingface organization](https://huggingface.co/deepghs)). ## List of Packages | Name | Images | Size | Download | Type | Description | |:-----------------|---------:|:-----------|:-----------------------------------------------------------------------------------------------------------------------------|:-----------|:---------------------------------------------------------------------| | raw | 87 | 114.99 MiB | [Download](https://huggingface.co/datasets/CyberHarem/asahi_rokka_bangdreamdai2ki/resolve/main/dataset-raw.zip) | Waifuc-Raw | Raw data with meta information (min edge aligned to 1400 if larger). | | 800 | 87 | 68.01 MiB | [Download](https://huggingface.co/datasets/CyberHarem/asahi_rokka_bangdreamdai2ki/resolve/main/dataset-800.zip) | IMG+TXT | dataset with the shorter side not exceeding 800 pixels. | | stage3-p480-800 | 197 | 137.41 MiB | [Download](https://huggingface.co/datasets/CyberHarem/asahi_rokka_bangdreamdai2ki/resolve/main/dataset-stage3-p480-800.zip) | IMG+TXT | 3-stage cropped dataset with the area not less than 480x480 pixels. | | 1200 | 87 | 100.48 MiB | [Download](https://huggingface.co/datasets/CyberHarem/asahi_rokka_bangdreamdai2ki/resolve/main/dataset-1200.zip) | IMG+TXT | dataset with the shorter side not exceeding 1200 pixels. | | stage3-p480-1200 | 197 | 190.07 MiB | [Download](https://huggingface.co/datasets/CyberHarem/asahi_rokka_bangdreamdai2ki/resolve/main/dataset-stage3-p480-1200.zip) | IMG+TXT | 3-stage cropped dataset with the area not less than 480x480 pixels. | ### Load Raw Dataset with Waifuc We provide raw dataset (including tagged images) for [waifuc](https://deepghs.github.io/waifuc/main/tutorials/installation/index.html) loading. If you need this, just run the following code ```python import os import zipfile from huggingface_hub import hf_hub_download from waifuc.source import LocalSource # download raw archive file zip_file = hf_hub_download( repo_id='CyberHarem/asahi_rokka_bangdreamdai2ki', repo_type='dataset', filename='dataset-raw.zip', ) # extract files to your directory dataset_dir = 'dataset_dir' os.makedirs(dataset_dir, exist_ok=True) with zipfile.ZipFile(zip_file, 'r') as zf: zf.extractall(dataset_dir) # load the dataset with waifuc source = LocalSource(dataset_dir) for item in source: print(item.image, item.meta['filename'], item.meta['tags']) ``` ## List of Clusters List of tag clustering result, maybe some outfits can be mined here. ### Raw Text Version | # | Samples | Img-1 | Img-2 | Img-3 | Img-4 | Img-5 | Tags | |----:|----------:|:--------------------------------|:--------------------------------|:--------------------------------|:--------------------------------|:--------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| | 0 | 9 | ![](samples/0/clu0-sample0.png) | ![](samples/0/clu0-sample1.png) | ![](samples/0/clu0-sample2.png) | ![](samples/0/clu0-sample3.png) | ![](samples/0/clu0-sample4.png) | 1girl, electric_guitar, holding_instrument, solo, looking_at_viewer, playing_instrument, black_shirt, blush, jewelry, standing, open_mouth, plectrum | | 1 | 7 | ![](samples/1/clu1-sample0.png) | ![](samples/1/clu1-sample1.png) | ![](samples/1/clu1-sample2.png) | ![](samples/1/clu1-sample3.png) | ![](samples/1/clu1-sample4.png) | 1girl, black-framed_eyewear, glasses, hair_over_shoulder, looking_at_viewer, solo, white_background, blush, hair_scrunchie, simple_background, red_scrunchie, standing, breasts, closed_mouth, collarbone, floral_print, low_ponytail, open_mouth, print_dress, shirt, smile | | 2 | 11 | ![](samples/2/clu2-sample0.png) | ![](samples/2/clu2-sample1.png) | ![](samples/2/clu2-sample2.png) | ![](samples/2/clu2-sample3.png) | ![](samples/2/clu2-sample4.png) | 1girl, solo, blazer, glasses, long_sleeves, school_uniform, black-framed_eyewear, collared_shirt, grey_jacket, looking_at_viewer, hair_over_shoulder, hair_scrunchie, white_shirt, blush, red_scrunchie, upper_body, green_necktie, plaid_skirt, pleated_skirt, star_(symbol), closed_mouth, diagonal-striped_necktie, diagonal_stripes, electric_guitar, eyewear_removed, green_skirt, holding_eyewear, holding_instrument, low_ponytail, simple_background, smile, socks, white_background | ### Table Version | # | Samples | Img-1 | Img-2 | Img-3 | Img-4 | Img-5 | 1girl | electric_guitar | holding_instrument | solo | looking_at_viewer | playing_instrument | black_shirt | blush | jewelry | standing | open_mouth | plectrum | black-framed_eyewear | glasses | hair_over_shoulder | white_background | hair_scrunchie | simple_background | red_scrunchie | breasts | closed_mouth | collarbone | floral_print | low_ponytail | print_dress | shirt | smile | blazer | long_sleeves | school_uniform | collared_shirt | grey_jacket | white_shirt | upper_body | green_necktie | plaid_skirt | pleated_skirt | star_(symbol) | diagonal-striped_necktie | diagonal_stripes | eyewear_removed | green_skirt | holding_eyewear | socks | |----:|----------:|:--------------------------------|:--------------------------------|:--------------------------------|:--------------------------------|:--------------------------------|:--------|:------------------|:---------------------|:-------|:--------------------|:---------------------|:--------------|:--------|:----------|:-----------|:-------------|:-----------|:-----------------------|:----------|:---------------------|:-------------------|:-----------------|:--------------------|:----------------|:----------|:---------------|:-------------|:---------------|:---------------|:--------------|:--------|:--------|:---------|:---------------|:-----------------|:-----------------|:--------------|:--------------|:-------------|:----------------|:--------------|:----------------|:----------------|:---------------------------|:-------------------|:------------------|:--------------|:------------------|:--------| | 0 | 9 | ![](samples/0/clu0-sample0.png) | ![](samples/0/clu0-sample1.png) | ![](samples/0/clu0-sample2.png) | ![](samples/0/clu0-sample3.png) | ![](samples/0/clu0-sample4.png) | X | X | X | X | X | X | X | X | X | X | X | X | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | 1 | 7 | ![](samples/1/clu1-sample0.png) | ![](samples/1/clu1-sample1.png) | ![](samples/1/clu1-sample2.png) | ![](samples/1/clu1-sample3.png) | ![](samples/1/clu1-sample4.png) | X | | | X | X | | | X | | X | X | | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | | | | | | | | | | | | | | | | | | | 2 | 11 | ![](samples/2/clu2-sample0.png) | ![](samples/2/clu2-sample1.png) | ![](samples/2/clu2-sample2.png) | ![](samples/2/clu2-sample3.png) | ![](samples/2/clu2-sample4.png) | X | X | X | X | X | | | X | | | | | X | X | X | X | X | X | X | | X | | | X | | | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X |
berkgungor/llama2_golf
--- license: llama2 ---
Abzu/dolly_wizard_codepy
--- dataset_info: features: - name: prompt dtype: string - name: response dtype: string splits: - name: train num_bytes: 110461778 num_examples: 92945 - name: test num_bytes: 12334691 num_examples: 10328 download_size: 68491380 dataset_size: 122796469 license: cc task_categories: - text-generation language: - en --- # Dataset Card for "dolly_wizard_codepy" This is a mix of datasets from: - Abzu/dolly_hhrlhf - Abzu/Wizard - Abzu/CodeAlpacaPython Which are versions of: - mosaicml/dolly_hhrlhf - ehartford/WizardLM_alpaca_evol_instruct_70k_unfiltered - HuggingFaceH4/CodeAlpaca_20K [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
Joe02/Nasipasuta_refs
--- license: other ---
salma-remyx/hf_objdet_test
--- dataset_info: features: - name: image dtype: image - name: objects struct: - name: bbox sequence: sequence: float64 - name: categories sequence: int64 splits: - name: train num_bytes: 7545187.0 num_examples: 16 download_size: 7548342 dataset_size: 7545187.0 --- # Dataset Card for "hf_objdet_test" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
AIGym/function_calling_v2
--- dataset_info: features: - name: userPrompt dtype: string - name: assistantResponse dtype: string - name: category dtype: string - name: sub_category dtype: string - name: title dtype: string - name: functionList dtype: string splits: - name: train num_bytes: 21177077 num_examples: 14736 - name: test num_bytes: 2920783 num_examples: 2010 download_size: 3529336 dataset_size: 24097860 configs: - config_name: default data_files: - split: train path: data/train-* - split: test path: data/test-* ---
Vinnyyw/Miacolucci
--- license: openrail ---
sethapun/arithmetic_2all_1to50
--- dataset_info: features: - name: expression dtype: string - name: answer dtype: float64 - name: label dtype: class_label: names: '0': 'false' '1': 'true' splits: - name: train num_bytes: 57306 num_examples: 2000 - name: validation num_bytes: 11472 num_examples: 400 download_size: 23365 dataset_size: 68778 --- # Dataset Card for "arithmetic_2all_1to50" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
Vinnyyw/Anylive
--- license: openrail ---
liuyanchen1015/MULTI_VALUE_mrpc_non_coordinated_obj_subj
--- dataset_info: features: - name: sentence1 dtype: string - name: sentence2 dtype: string - name: label dtype: int64 - name: idx dtype: int64 - name: value_score dtype: int64 splits: - name: test num_bytes: 101291 num_examples: 378 - name: train num_bytes: 240558 num_examples: 891 - name: validation num_bytes: 24738 num_examples: 94 download_size: 245547 dataset_size: 366587 --- # Dataset Card for "MULTI_VALUE_mrpc_non_coordinated_obj_subj" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
adamwatters/roblox-guy
--- license: openrail ---
GoldenTanuki/shinmegaten
--- license: other ---
Ghostdofaroest/fernando
--- license: openrail ---
liuyanchen1015/MULTI_VALUE_rte_chaining_main_verbs
--- dataset_info: features: - name: sentence1 dtype: string - name: sentence2 dtype: string - name: label dtype: int64 - name: idx dtype: int64 - name: value_score dtype: int64 splits: - name: test num_bytes: 13509 num_examples: 30 - name: train num_bytes: 18902 num_examples: 37 download_size: 30925 dataset_size: 32411 --- # Dataset Card for "MULTI_VALUE_rte_chaining_main_verbs" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
kursathalat/rating_ds
--- size_categories: n<1K tags: - rlfh - argilla - human-feedback --- # Dataset Card for rating_ds This dataset has been created with [Argilla](https://docs.argilla.io). As shown in the sections below, this dataset can be loaded into Argilla as explained in [Load with Argilla](#load-with-argilla), or used directly with the `datasets` library in [Load with `datasets`](#load-with-datasets). ## Dataset Description - **Homepage:** https://argilla.io - **Repository:** https://github.com/argilla-io/argilla - **Paper:** - **Leaderboard:** - **Point of Contact:** ### Dataset Summary This dataset contains: * A dataset configuration file conforming to the Argilla dataset format named `argilla.yaml`. This configuration file will be used to configure the dataset when using the `FeedbackDataset.from_huggingface` method in Argilla. * Dataset records in a format compatible with HuggingFace `datasets`. These records will be loaded automatically when using `FeedbackDataset.from_huggingface` and can be loaded independently using the `datasets` library via `load_dataset`. * The [annotation guidelines](#annotation-guidelines) that have been used for building and curating the dataset, if they've been defined in Argilla. ### Load with Argilla To load with Argilla, you'll just need to install Argilla as `pip install argilla --upgrade` and then use the following code: ```python import argilla as rg ds = rg.FeedbackDataset.from_huggingface("kursathalat/rating_ds") ``` ### Load with `datasets` To load this dataset with `datasets`, you'll just need to install `datasets` as `pip install datasets --upgrade` and then use the following code: ```python from datasets import load_dataset ds = load_dataset("kursathalat/rating_ds") ``` ### Supported Tasks and Leaderboards This dataset can contain [multiple fields, questions and responses](https://docs.argilla.io/en/latest/conceptual_guides/data_model.html#feedback-dataset) so it can be used for different NLP tasks, depending on the configuration. The dataset structure is described in the [Dataset Structure section](#dataset-structure). There are no leaderboards associated with this dataset. ### Languages [More Information Needed] ## Dataset Structure ### Data in Argilla The dataset is created in Argilla with: **fields**, **questions**, **suggestions**, **metadata**, **vectors**, and **guidelines**. The **fields** are the dataset records themselves, for the moment just text fields are supported. These are the ones that will be used to provide responses to the questions. | Field Name | Title | Type | Required | Markdown | | ---------- | ----- | ---- | -------- | -------- | | sentence1 | Sentence1 | text | True | False | | sentence2 | Sentence2 | text | True | False | The **questions** are the questions that will be asked to the annotators. They can be of different types, such as rating, text, label_selection, multi_label_selection, or ranking. | Question Name | Title | Type | Required | Description | Values/Labels | | ------------- | ----- | ---- | -------- | ----------- | ------------- | | similarity | Similarity | rating | True | N/A | [1, 2, 3, 4, 5, 6, 7] | The **suggestions** are human or machine generated recommendations for each question to assist the annotator during the annotation process, so those are always linked to the existing questions, and named appending "-suggestion" and "-suggestion-metadata" to those, containing the value/s of the suggestion and its metadata, respectively. So on, the possible values are the same as in the table above, but the column name is appended with "-suggestion" and the metadata is appended with "-suggestion-metadata". The **metadata** is a dictionary that can be used to provide additional information about the dataset record. This can be useful to provide additional context to the annotators, or to provide additional information about the dataset record itself. For example, you can use this to provide a link to the original source of the dataset record, or to provide additional information about the dataset record itself, such as the author, the date, or the source. The metadata is always optional, and can be potentially linked to the `metadata_properties` defined in the dataset configuration file in `argilla.yaml`. | Metadata Name | Title | Type | Values | Visible for Annotators | | ------------- | ----- | ---- | ------ | ---------------------- | The **guidelines**, are optional as well, and are just a plain string that can be used to provide instructions to the annotators. Find those in the [annotation guidelines](#annotation-guidelines) section. ### Data Instances An example of a dataset instance in Argilla looks as follows: ```json { "external_id": null, "fields": { "sentence1": " \"If you don\u0027t wear BROWN AND ORANGE...YOU DON\u0027T MATTER!\" We need a tshirt with that on it asap! ", "sentence2": " \"If you don\u0027t weer BROWN AND ORANGE...YOU DON\u0027T MATTER!\" We need e tshirt with thet on it esep! " }, "metadata": {}, "responses": [ { "status": "submitted", "user_id": "464c365f-9fd7-4db9-a06f-42c320c54547", "values": { "similarity": { "value": 2 } } }, { "status": "submitted", "user_id": "525f4274-ebb4-4aee-a116-f8b422b2e2b4", "values": { "similarity": { "value": 2 } } }, { "status": "submitted", "user_id": "156c274f-d22e-406e-a1ec-77cd734d008a", "values": { "similarity": { "value": 3 } } } ], "suggestions": [], "vectors": {} } ``` While the same record in HuggingFace `datasets` looks as follows: ```json { "external_id": null, "metadata": "{}", "sentence1": " \"If you don\u0027t wear BROWN AND ORANGE...YOU DON\u0027T MATTER!\" We need a tshirt with that on it asap! ", "sentence2": " \"If you don\u0027t weer BROWN AND ORANGE...YOU DON\u0027T MATTER!\" We need e tshirt with thet on it esep! ", "similarity": [ { "status": "submitted", "user_id": "464c365f-9fd7-4db9-a06f-42c320c54547", "value": 2 }, { "status": "submitted", "user_id": "525f4274-ebb4-4aee-a116-f8b422b2e2b4", "value": 2 }, { "status": "submitted", "user_id": "156c274f-d22e-406e-a1ec-77cd734d008a", "value": 3 } ], "similarity-suggestion": null, "similarity-suggestion-metadata": { "agent": null, "score": null, "type": null } } ``` ### Data Fields Among the dataset fields, we differentiate between the following: * **Fields:** These are the dataset records themselves, for the moment just text fields are supported. These are the ones that will be used to provide responses to the questions. * **sentence1** is of type `text`. * **sentence2** is of type `text`. * **Questions:** These are the questions that will be asked to the annotators. They can be of different types, such as `RatingQuestion`, `TextQuestion`, `LabelQuestion`, `MultiLabelQuestion`, and `RankingQuestion`. * **similarity** is of type `rating` with the following allowed values [1, 2, 3, 4, 5, 6, 7]. * **Suggestions:** As of Argilla 1.13.0, the suggestions have been included to provide the annotators with suggestions to ease or assist during the annotation process. Suggestions are linked to the existing questions, are always optional, and contain not just the suggestion itself, but also the metadata linked to it, if applicable. * (optional) **similarity-suggestion** is of type `rating` with the following allowed values [1, 2, 3, 4, 5, 6, 7]. Additionally, we also have two more fields that are optional and are the following: * **metadata:** This is an optional field that can be used to provide additional information about the dataset record. This can be useful to provide additional context to the annotators, or to provide additional information about the dataset record itself. For example, you can use this to provide a link to the original source of the dataset record, or to provide additional information about the dataset record itself, such as the author, the date, or the source. The metadata is always optional, and can be potentially linked to the `metadata_properties` defined in the dataset configuration file in `argilla.yaml`. * **external_id:** This is an optional field that can be used to provide an external ID for the dataset record. This can be useful if you want to link the dataset record to an external resource, such as a database or a file. ### Data Splits The dataset contains a single split, which is `train`. ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation guidelines This is a sentence similarity dataset that contains two sentences. Please rate the similarity between the two sentences. #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions [More Information Needed]
tyzhu/squad_qa_baseline_v5_full_recite_ans_sent_random_permute_rerun_4
--- configs: - config_name: default data_files: - split: train path: data/train-* - split: validation path: data/validation-* dataset_info: features: - name: id dtype: string - name: title dtype: string - name: context dtype: string - name: question dtype: string - name: answers sequence: - name: text dtype: string - name: answer_start dtype: int32 - name: answer dtype: string - name: context_id dtype: string - name: inputs dtype: string - name: targets dtype: string splits: - name: train num_bytes: 2996506.0 num_examples: 2385 - name: validation num_bytes: 395889 num_examples: 300 download_size: 842977 dataset_size: 3392395.0 --- # Dataset Card for "squad_qa_baseline_v5_full_recite_ans_sent_random_permute_rerun_4" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
harinarayan/my_dataset_01
--- dataset_info: features: - name: image_file dtype: string - name: caption dtype: string splits: - name: train num_bytes: 2422 num_examples: 20 download_size: 2850 dataset_size: 2422 configs: - config_name: default data_files: - split: train path: data/train-* --- # Dataset Card for "my_dataset_01" This is a dataset for captioning graph images [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
timm/objectnet-in1k
--- dataset_info: features: - name: image dtype: image - name: label dtype: class_label: names: '0': air_freshener '1': alarm_clock '2': backpack '3': baking_sheet '4': banana '5': band_aid '6': baseball_bat '7': baseball_glove '8': basket '9': bathrobe '10': battery '11': bed_sheet '12': beer_bottle '13': beer_can '14': belt '15': bench '16': bicycle '17': bike_pump '18': bills_money '19': binder_closed '20': biscuits '21': blanket '22': blender '23': blouse '24': board_game '25': book_closed '26': bookend '27': boots '28': bottle_cap '29': bottle_opener '30': bottle_stopper '31': box '32': bracelet '33': bread_knife '34': bread_loaf '35': briefcase '36': brooch '37': broom '38': bucket '39': butchers_knife '40': butter '41': button '42': calendar '43': can_opener '44': candle '45': canned_food '46': cd_case '47': cellphone '48': cellphone_case '49': cellphone_charger '50': cereal '51': chair '52': cheese '53': chess_piece '54': chocolate '55': chopstick '56': clothes_hamper '57': clothes_hanger '58': coaster '59': coffee_beans '60': coffee_french_press '61': coffee_grinder '62': coffee_machine '63': coffee_table '64': coin_money '65': comb '66': combination_lock '67': computer_mouse '68': contact_lens_case '69': cooking_oil_bottle '70': cork '71': cutting_board '72': deodorant '73': desk_lamp '74': detergent '75': dish_soap '76': document_folder_closed '77': dog_bed '78': doormat '79': drawer_open '80': dress '81': dress_pants '82': dress_shirt '83': dress_shoe_men '84': dress_shoe_women '85': drill '86': drinking_cup '87': drinking_straw '88': drying_rack_for_clothes '89': drying_rack_for_dishes '90': dust_pan '91': dvd_player '92': earbuds '93': earring '94': egg '95': egg_carton '96': envelope '97': eraser_white_board '98': extension_cable '99': eyeglasses '100': fan '101': figurine_or_statue '102': first_aid_kit '103': flashlight '104': floss_container '105': flour_container '106': fork '107': frying_pan '108': full_sized_towel '109': glue_container '110': hair_brush '111': hair_dryer '112': hairclip '113': hairtie '114': hammer '115': hand_mirror '116': hand_towel_or_rag '117': handbag '118': hat '119': headphones_over_ear '120': helmet '121': honey_container '122': ice '123': ice_cube_tray '124': iron_for_clothes '125': ironing_board '126': jam '127': jar '128': jeans '129': kettle '130': key_chain '131': keyboard '132': ladle '133': lampshade '134': laptop_charger '135': laptop_open '136': leaf '137': leggings '138': lemon '139': letter_opener '140': lettuce '141': light_bulb '142': lighter '143': lipstick '144': loofah '145': magazine '146': makeup '147': makeup_brush '148': marker '149': match '150': measuring_cup '151': microwave '152': milk '153': mixing_salad_bowl '154': monitor '155': mouse_pad '156': mouthwash '157': mug '158': multitool '159': nail_clippers '160': nail_fastener '161': nail_file '162': nail_polish '163': napkin '164': necklace '165': newspaper '166': night_light '167': nightstand '168': notebook '169': notepad '170': nut_for_screw '171': orange '172': oven_mitts '173': padlock '174': paint_can '175': paintbrush '176': paper '177': paper_bag '178': paper_plates '179': paper_towel '180': paperclip '181': peeler '182': pen '183': pencil '184': pepper_shaker '185': pet_food_container '186': phone_landline '187': photograph_printed '188': pill_bottle '189': pill_organizer '190': pillow '191': pitcher '192': placemat '193': plastic_bag '194': plastic_cup '195': plastic_wrap '196': plate '197': playing_cards '198': pliers '199': plunger '200': pop_can '201': portable_heater '202': poster '203': power_bar '204': power_cable '205': printer '206': raincoat '207': rake '208': razor '209': receipt '210': remote_control '211': removable_blade '212': ribbon '213': ring '214': rock '215': rolling_pin '216': ruler '217': running_shoe '218': safety_pin '219': salt_shaker '220': sandal '221': scarf '222': scissors '223': screw '224': scrub_brush '225': sewing_kit '226': shampoo_bottle '227': shoelace '228': shorts '229': shovel '230': skateboard '231': skirt '232': sleeping_bag '233': slipper '234': soap_bar '235': soap_dispenser '236': sock '237': soup_bowl '238': spatula '239': speaker '240': sponge '241': spoon '242': spray_bottle '243': squeegee '244': squeeze_bottle '245': standing_lamp '246': stapler '247': step_stool '248': still_camera '249': stopper_sink_tub '250': strainer '251': stuffed_animal '252': sugar_container '253': suit_jacket '254': suitcase '255': sunglasses '256': sweater '257': swimming_trunks '258': t-shirt '259': table_knife '260': tablecloth '261': tablet_ipad '262': tanktop '263': tape '264': tape_measure '265': tarp '266': teabag '267': teapot '268': tennis_racket '269': thermometer '270': thermos '271': throw_pillow '272': tie '273': tissue '274': toaster '275': toilet_paper_roll '276': tomato '277': tongs '278': toothbrush '279': toothpaste '280': tote_bag '281': toy '282': trash_bag '283': trash_bin '284': travel_case '285': tray '286': trophy '287': tv '288': tweezers '289': umbrella '290': usb_cable '291': usb_flash_drive '292': vacuum_cleaner '293': vase '294': video_camera '295': walker '296': walking_cane '297': wallet '298': watch '299': water_bottle '300': water_filter '301': webcam '302': weight_exercise '303': weight_scale '304': wheel '305': whisk '306': whistle '307': wine_bottle '308': wine_glass '309': winter_glove '310': wok '311': wrench '312': ziploc_bag - name: imagenet_labels sequence: int64 - name: imagenet_synsets sequence: string splits: - name: test num_bytes: 45447504721.55699 num_examples: 18574 download_size: 46256058381 dataset_size: 45447504721.55699 configs: - config_name: default data_files: - split: test path: data/test-* extra_gated_prompt: 'By clicking on “Access repository” below, you also agree to ObjectNet Terms: ObjectNet is free to use for both research and commercial applications. The authors own the source images and allow their use under a license derived from Creative Commons Attribution 4.0 with only two additional clauses. 1. ObjectNet may never be used to tune the parameters of any model. 2. Any individual images from ObjectNet may only be posted to the web including their 1 pixel red border. If you are using ObjectNet, please cite our work, the citation appears at the bottom of this page. Any derivative of ObjectNet must contain attribution as well.' --- # ObjectNet (ImageNet-1k Overlapping) A webp (lossless) encoded version of [ObjectNet-1.0](https://objectnet.dev/index.html) at original resolution, containing only the images for the 113 classes that overlap with ImageNet-1k classes. ## License / Usage Terms ObjectNet is free to use for both research and commercial applications. The authors own the source images and allow their use under a license derived from Creative Commons Attribution 4.0 with only two additional clauses. 1. **ObjectNet may never be used to tune the parameters of any model.** 2. **Any individual images from ObjectNet may only be posted to the web including their 1 pixel red border**. If you are using ObjectNet, please cite our work, the citation appears at the bottom of this page. Any derivative of ObjectNet must contain attribution as well. ## About ObjectNet What is ObjectNet? * A new kind of vision dataset borrowing the idea of controls from other areas of science. * No training set, only a test set! Put your vision system through its paces. * Collected to intentionally show objects from new viewpoints on new backgrounds. * 50,000 image test set, same as ImageNet, with controls for rotation, background, and viewpoint. * 313 object classes with 113 overlapping ImageNet * Large performance drop, what you can expect from vision systems in the real world! * Robust to fine-tuning and a very difficult transfer learning problem ## Why the Red Borders / How do I recognize if an image is in ObjectNet? As training sets become huge, the risk that test and training sets overlap is serious. We provide ObjectNet with a 2 pixel red border around each image which must be removed before performing inference. The ObjectNet license requires that if you post images from ObjectNet to the web, you include this border. Any time you see an image with a solid 2 pixel red border, that's an indication it's in someone's test set and you should be careful about training on it. Reverse image search will allow you to figure out which test set it is from. NOTE: original ObjectNet PNG files actually have a 2 pixel red border while their descriptions say 1. ## Preprocessing Steps for This timm Version 1. Re-encode PNG images with lossless WebP (~32% reduction in size), keeping red border. 2. Add `imagenet_labels` and `imagenet_synsets` consisting of lists of ImageNet-1k classes that overlap with ObjectNet class. 3. Remove all ObjectNet image classes without ImageNet-1k labels. ## Citation ```bibtex @incollection{NIPS2019_9142, title = {ObjectNet: A large-scale bias-controlled dataset for pushing the limits of object recognition models}, author = {Barbu, Andrei and Mayo, David and Alverio, Julian and Luo, William and Wang, Christopher and Gutfreund, Dan and Tenenbaum, Josh and Katz, Boris}, booktitle = {Advances in Neural Information Processing Systems 32}, editor = {H. Wallach and H. Larochelle and A. Beygelzimer and F. d\textquotesingle Alch\'{e}-Buc and E. Fox and R. Garnett}, pages = {9448--9458}, year = {2019}, publisher = {Curran Associates, Inc.}, url = {http://papers.nips.cc/paper/9142-objectnet-a-large-scale-bias-controlled-dataset-for-pushing-the-limits-of-object-recognition-models.pdf} } ```
avalonai/indonesian_corpus_eli5_id
--- dataset_info: features: - name: corpora dtype: string splits: - name: train num_bytes: 517534186 num_examples: 272634 download_size: 263288246 dataset_size: 517534186 configs: - config_name: default data_files: - split: train path: data/train-* ---
marcones/vozmarcones
--- license: openrail ---
Charles333/lama_chat_2.0_fintuned_json_data
--- license: apache-2.0 ---
CyberHarem/blacknight_arknights
--- license: mit task_categories: - text-to-image tags: - art - not-for-all-audiences size_categories: - n<1K --- # Dataset of blacknight/ブラックナイト/夜半 (Arknights) This is the dataset of blacknight/ブラックナイト/夜半 (Arknights), containing 56 images and their tags. The core tags of this character are `long_hair, black_hair, multicolored_hair, breasts, horns, pointy_ears, blue_hair, braid, large_breasts, brown_eyes, mole, tail, very_long_hair`, which are pruned in this dataset. Images are crawled from many sites (e.g. danbooru, pixiv, zerochan ...), the auto-crawling system is powered by [DeepGHS Team](https://github.com/deepghs)([huggingface organization](https://huggingface.co/deepghs)). ## List of Packages | Name | Images | Size | Download | Type | Description | |:-----------------|---------:|:-----------|:----------------------------------------------------------------------------------------------------------------------|:-----------|:---------------------------------------------------------------------| | raw | 56 | 114.35 MiB | [Download](https://huggingface.co/datasets/CyberHarem/blacknight_arknights/resolve/main/dataset-raw.zip) | Waifuc-Raw | Raw data with meta information (min edge aligned to 1400 if larger). | | 1200 | 56 | 91.29 MiB | [Download](https://huggingface.co/datasets/CyberHarem/blacknight_arknights/resolve/main/dataset-1200.zip) | IMG+TXT | dataset with the shorter side not exceeding 1200 pixels. | | stage3-p480-1200 | 150 | 179.09 MiB | [Download](https://huggingface.co/datasets/CyberHarem/blacknight_arknights/resolve/main/dataset-stage3-p480-1200.zip) | IMG+TXT | 3-stage cropped dataset with the area not less than 480x480 pixels. | ### Load Raw Dataset with Waifuc We provide raw dataset (including tagged images) for [waifuc](https://deepghs.github.io/waifuc/main/tutorials/installation/index.html) loading. If you need this, just run the following code ```python import os import zipfile from huggingface_hub import hf_hub_download from waifuc.source import LocalSource # download raw archive file zip_file = hf_hub_download( repo_id='CyberHarem/blacknight_arknights', repo_type='dataset', filename='dataset-raw.zip', ) # extract files to your directory dataset_dir = 'dataset_dir' os.makedirs(dataset_dir, exist_ok=True) with zipfile.ZipFile(zip_file, 'r') as zf: zf.extractall(dataset_dir) # load the dataset with waifuc source = LocalSource(dataset_dir) for item in source: print(item.image, item.meta['filename'], item.meta['tags']) ``` ## List of Clusters List of tag clustering result, maybe some outfits can be mined here. ### Raw Text Version | # | Samples | Img-1 | Img-2 | Img-3 | Img-4 | Img-5 | Tags | |----:|----------:|:--------------------------------|:--------------------------------|:--------------------------------|:--------------------------------|:--------------------------------|:-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| | 0 | 11 | ![](samples/0/clu0-sample0.png) | ![](samples/0/clu0-sample1.png) | ![](samples/0/clu0-sample2.png) | ![](samples/0/clu0-sample3.png) | ![](samples/0/clu0-sample4.png) | 1girl, black_gloves, solo, bare_shoulders, black_dress, looking_at_viewer, ponytail, simple_background, white_background, earrings, holding, streaked_hair, black_footwear, black_nails, elbow_gloves, hair_between_eyes, nail_polish, official_alternate_costume, partially_fingerless_gloves, full_body, smile, boots, closed_mouth, sleeveless_dress, thigh_strap, weapon | | 1 | 16 | ![](samples/1/clu1-sample0.png) | ![](samples/1/clu1-sample1.png) | ![](samples/1/clu1-sample2.png) | ![](samples/1/clu1-sample3.png) | ![](samples/1/clu1-sample4.png) | 1girl, bare_shoulders, cleavage, official_alternate_costume, solo, black_one-piece_swimsuit, casual_one-piece_swimsuit, looking_at_viewer, thigh_strap, detached_sleeves, hair_ornament, necklace, nail_polish, parted_lips, white_flower, closed_mouth, collarbone, simple_background, white_background, covered_navel, hair_between_eyes, holding, mole_under_eye, sitting | ### Table Version | # | Samples | Img-1 | Img-2 | Img-3 | Img-4 | Img-5 | 1girl | black_gloves | solo | bare_shoulders | black_dress | looking_at_viewer | ponytail | simple_background | white_background | earrings | holding | streaked_hair | black_footwear | black_nails | elbow_gloves | hair_between_eyes | nail_polish | official_alternate_costume | partially_fingerless_gloves | full_body | smile | boots | closed_mouth | sleeveless_dress | thigh_strap | weapon | cleavage | black_one-piece_swimsuit | casual_one-piece_swimsuit | detached_sleeves | hair_ornament | necklace | parted_lips | white_flower | collarbone | covered_navel | mole_under_eye | sitting | |----:|----------:|:--------------------------------|:--------------------------------|:--------------------------------|:--------------------------------|:--------------------------------|:--------|:---------------|:-------|:-----------------|:--------------|:--------------------|:-----------|:--------------------|:-------------------|:-----------|:----------|:----------------|:-----------------|:--------------|:---------------|:--------------------|:--------------|:-----------------------------|:------------------------------|:------------|:--------|:--------|:---------------|:-------------------|:--------------|:---------|:-----------|:---------------------------|:----------------------------|:-------------------|:----------------|:-----------|:--------------|:---------------|:-------------|:----------------|:-----------------|:----------| | 0 | 11 | ![](samples/0/clu0-sample0.png) | ![](samples/0/clu0-sample1.png) | ![](samples/0/clu0-sample2.png) | ![](samples/0/clu0-sample3.png) | ![](samples/0/clu0-sample4.png) | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | | | | | | | | | | | | | | 1 | 16 | ![](samples/1/clu1-sample0.png) | ![](samples/1/clu1-sample1.png) | ![](samples/1/clu1-sample2.png) | ![](samples/1/clu1-sample3.png) | ![](samples/1/clu1-sample4.png) | X | | X | X | | X | | X | X | | X | | | | | X | X | X | | | | | X | | X | | X | X | X | X | X | X | X | X | X | X | X | X |
Charitarth/dac-sdc-2024
--- language: - en size_categories: - 1K<n<10K task_categories: - object-detection pretty_name: DAC SDC 2024 dataset_info: features: - name: image dtype: image - name: objects struct: - name: bbox sequence: sequence: int64 - name: categories sequence: int64 - name: segmentation sequence: sequence: float64 splits: - name: train num_bytes: 3223931152.0 num_examples: 10000 download_size: 3336168589 dataset_size: 3223931152.0 configs: - config_name: default data_files: - split: train path: data/train-* --- Information on this dataset can be found on the [contest website](https://pku-sec-lab.github.io/dac-sdc-2024/info/). The only modifications done between the original dataset and here is that images with the _1.jpg suffix have been removed.
Pablao0948/Freya
--- license: openrail ---
livinNector/ner_indic_glue
--- dataset_info: features: - name: tokens sequence: string - name: ner_tags sequence: class_label: names: '0': O '1': B-PER '2': I-PER '3': B-ORG '4': I-ORG '5': B-LOC '6': I-LOC splits: - name: train num_bytes: 4687618 num_examples: 20466 - name: validation num_bytes: 591516 num_examples: 2586 - name: test num_bytes: 585925 num_examples: 2611 download_size: 1381725 dataset_size: 5865059 --- # Dataset Card for "ner_indic_glue" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
anan-2024/twitter_dataset_1713184472
--- dataset_info: features: - name: id dtype: string - name: tweet_content dtype: string - name: user_name dtype: string - name: user_id dtype: string - name: created_at dtype: string - name: url dtype: string - name: favourite_count dtype: int64 - name: scraped_at dtype: string - name: image_urls dtype: string splits: - name: train num_bytes: 75299 num_examples: 201 download_size: 46372 dataset_size: 75299 configs: - config_name: default data_files: - split: train path: data/train-* ---
breno30/Cava
--- license: openrail ---
liuyanchen1015/MULTI_VALUE_mnli_non_coordinated_obj_subj
--- dataset_info: features: - name: premise dtype: string - name: hypothesis dtype: string - name: label dtype: int64 - name: idx dtype: int64 - name: score dtype: int64 splits: - name: dev_matched num_bytes: 557438 num_examples: 2662 - name: dev_mismatched num_bytes: 614716 num_examples: 2766 - name: test_matched num_bytes: 597781 num_examples: 2713 - name: test_mismatched num_bytes: 610200 num_examples: 2773 - name: train num_bytes: 23464192 num_examples: 107927 download_size: 16238926 dataset_size: 25844327 --- # Dataset Card for "MULTI_VALUE_mnli_non_coordinated_obj_subj" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
hllj/vi_grade_school_math_mcq
--- license: mit task_categories: - text-generation - text2text-generation - multiple-choice language: - vi pretty_name: Vietnamese Grade School Math Dataset size_categories: - 1K<n<10K tags: - vietnamese - math - reasoning --- # Dataset Card for Vietnamese Grade School Math Dataset ## Dataset Description - Repository: - Paper: - Point of Contact: vanhop3499@gmail.com ### Dataset Summary The dataset includes multiple-choice math exercises for elementary school students from grades 1 to 5 in Vietnam. ### Supported Tasks and Leaderboards ### Languages The majority of the data is in Vietnamese. ## Dataset Structure ### Data Instances The data includes information about the page paths we crawled and some text that has been post-processed. The structure will be presented as follows: ```python { "id": "f9decb7530da8097ebca80315928825e", "question": "Câu 2: Trang 21 - sgk toán lớp 5\nMột gia đình gồm 3 người (bố, mẹ và một con). Bình quân thu nhập hàng tháng 800 000 đồng mỗi người. Nếu gia đình đó có thêm một con nữa mà tổng thu nhập của gia đình không thay đổi thì bình quân thu nhập hàng tháng của mỗi người giảm đi bao nhiêu tiền?", "explanation": "Tổng thu hập bình quân một tháng của gia đình đó là:\n800000 x 3 = 2400000 ( đồng)\nSau khi thêm một người, thu nhập trung bình của một người trong gia đình là:\n2400000 : 4 = 600000 ( đồng)\nVậy so với trước đó, thu nhập bình quân mỗi tháng của một người đã giảm đi:\n800000 - 600000 = 200000 ( đồng)\nĐáp án: 200000 đồng.", "choices": [ "A. 180000 đồng.", "B. 250000 đồng.", "C. 220000 đồng.", "D. 200000 đồng." ], "answer": "D. 200000 đồng." } ``` ### Data Fields Data fields include: - id: id of an question instance. - question: Multiple choice question text - explanation: explanation for how to find the answer to the question - choices: 4 choices A, B, C, D for the question. - answer: one of 4 choices, the final answer. ## Dataset Creation ### Curation Rationale The data set is built based on developing a model capable of reasoning and solving elementary school math problems, as well as providing mathematical knowledge in the Vietnamese elementary school environment. ### Source Data Data was crawled on khoahoc.vietjack.com, we selected data from grades 1 to 5, selected all exams for maths. ## Considerations for Using the Data ### Social Impact of Dataset We believe that efforts in finding data sources will be an opportunity for future artificial intelligence models to develop and have better reasoning capabilities. ### Discussion of Biases ### Other Known Limitations The current data has not been cleaned too well, and there are many incomplete data samples including images and some post-processed tags. ## Additional Information
aicyd/my-test
--- license: apache-2.0 ---
Junity/UmaMusume-TokaiTeio-Dataset
--- license: openrail --- # 作者 https://space.bilibili.com/317820045
autoevaluate/autoeval-eval-cnn_dailymail-3.0.0-84482e-60145145396
--- type: predictions tags: - autotrain - evaluation datasets: - cnn_dailymail eval_info: task: summarization model: gauravkoradiya/T5-Fintuned-on-cnn_daily_mail metrics: [] dataset_name: cnn_dailymail dataset_config: 3.0.0 dataset_split: test col_mapping: text: article target: highlights --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Summarization * Model: gauravkoradiya/T5-Fintuned-on-cnn_daily_mail * Dataset: cnn_dailymail * Config: 3.0.0 * Split: test To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@SINI RAJ P](https://huggingface.co/SINI RAJ P) for evaluating this model.
ilmariky/WikiQA-100-fi
--- language: - fi license: - gpl-3.0 multilinguality: - monolingual size_categories: - n<1k task_categories: - question-answering task_ids: - extractive-qa pretty_name: WikiQA-100-fi tags: - question-generation train-eval-index: - config: plain_text task: question-answering task_id: extractive_question_answering splits: train_split: train eval_split: validation col_mapping: question: question context: context answers: text: text answer_start: answer_start --- # Dataset Card for "WikiQA-100-fi" ### Dataset Summary WikiQA-100-fi dataset contains 100 questions related to Finnish Wikipedia articles. The dataset is in the [SQuAD](https://rajpurkar.github.io/SQuAD-explorer/) format, and there are 10 questions for each category identified by the authors of SQuAD. Unlike SQuAD2.0, WikiQA-100-fi contains only answerable questions. The dataset is tiny compared to actual QA test sets, but it still gives an impression of the models' performance on purely native text data collected by a native speaker. The dataset was originally created as an evaluation set for models that had been mostly fine-tuned with automatically translated QA data. More information about the dataset and models created with it can be found [here](https://helda.helsinki.fi/handle/10138/344973). ## Dataset Structure ### Data Instances Example data: ``` { "title": "Folksonomia", "paragraphs": [ { "qas": [ { "question": "Minkälaista sisältöä käyttäjät voivat luokitella folksonomian avulla?", "id": "6t4ufel624", "answers": [ { "text": "www-sivuja, valokuvia ja linkkejä", "answer_start": 155 } ], "is_impossible": false } ], "context": "Folksonomia (engl. folksonomy) on yhteisöllisesti tuotettu, avoin luokittelujärjestelmä, jonka avulla internet-käyttäjät voivat luokitella sisältöä, kuten www-sivuja, valokuvia ja linkkejä. Etymologisesti folksonomia on peräisin sanojen \"folk\" (suom. väki) ja \"taxonomy\" (suom. taksonomia) leikkimielisestä yhdistelmästä." } ] } ``` ### Data Fields #### plain_text - `id`: a `string` feature. - `title`: a `string` feature. - `context`: a `string` feature. - `question`: a `string` feature. - `answers`: a dictionary feature containing: - `text`: a `string` feature. - `answer_start`: a `int32` feature. ### Data Splits | name | test| |----------|----:| |plain_text| 100| ### Citation Information ``` @MastersThesis{3241c198b3f147faacbc6d8b64ed9419, author = "Kylli{\"a}inen, {Ilmari}", title = "Neural Factoid Question Answering and Question Generation for Finnish", language = "en", address = "Helsinki, Finland", school = "University of Helsinki", year = "2022", month = "jun", day = "15", url = "https://helda.helsinki.fi/handle/10138/344973" } ```
autoevaluate/autoeval-staging-eval-project-fa97c361-989b-438c-bd2b-73aa1588c214-5654
--- type: predictions tags: - autotrain - evaluation datasets: - glue eval_info: task: natural_language_inference model: autoevaluate/natural-language-inference metrics: [] dataset_name: glue dataset_config: mrpc dataset_split: validation col_mapping: text1: sentence1 text2: sentence2 target: label --- # Dataset Card for AutoTrain Evaluator This repository contains model predictions generated by [AutoTrain](https://huggingface.co/autotrain) for the following task and dataset: * Task: Natural Language Inference * Model: autoevaluate/natural-language-inference * Dataset: glue * Config: mrpc * Split: validation To run new evaluation jobs, visit Hugging Face's [automatic model evaluator](https://huggingface.co/spaces/autoevaluate/model-evaluator). ## Contributions Thanks to [@lewtun](https://huggingface.co/lewtun) for evaluating this model.
Coaso/test
--- license: cc-by-sa-3.0 ---
zolak/twitter_dataset_79_1713105708
--- dataset_info: features: - name: id dtype: string - name: tweet_content dtype: string - name: user_name dtype: string - name: user_id dtype: string - name: created_at dtype: string - name: url dtype: string - name: favourite_count dtype: int64 - name: scraped_at dtype: string - name: image_urls dtype: string splits: - name: train num_bytes: 2863659 num_examples: 7045 download_size: 1420506 dataset_size: 2863659 configs: - config_name: default data_files: - split: train path: data/train-* ---
open-llm-leaderboard/details_LTC-AI-Labs__L2-7b-Base-WVG-Uncensored
--- pretty_name: Evaluation run of LTC-AI-Labs/L2-7b-Base-WVG-Uncensored dataset_summary: "Dataset automatically created during the evaluation run of model\ \ [LTC-AI-Labs/L2-7b-Base-WVG-Uncensored](https://huggingface.co/LTC-AI-Labs/L2-7b-Base-WVG-Uncensored)\ \ on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\ \nThe dataset is composed of 64 configuration, each one coresponding to one of the\ \ evaluated task.\n\nThe dataset has been created from 2 run(s). Each run can be\ \ found as a specific split in each configuration, the split being named using the\ \ timestamp of the run.The \"train\" split is always pointing to the latest results.\n\ \nAn additional configuration \"results\" store all the aggregated results of the\ \ run (and is used to compute and display the agregated metrics on the [Open LLM\ \ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\ \nTo load the details from a run, you can for instance do the following:\n```python\n\ from datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_LTC-AI-Labs__L2-7b-Base-WVG-Uncensored\"\ ,\n\t\"harness_winogrande_5\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\n\ These are the [latest results from run 2023-10-23T14:27:39.647282](https://huggingface.co/datasets/open-llm-leaderboard/details_LTC-AI-Labs__L2-7b-Base-WVG-Uncensored/blob/main/results_2023-10-23T14-27-39.647282.json)(note\ \ that their might be results for other tasks in the repos if successive evals didn't\ \ cover the same tasks. You find each in the results and the \"latest\" split for\ \ each eval):\n\n```python\n{\n \"all\": {\n \"em\": 0.001363255033557047,\n\ \ \"em_stderr\": 0.0003778609196461104,\n \"f1\": 0.05628670302013431,\n\ \ \"f1_stderr\": 0.0013136825712446111,\n \"acc\": 0.4109445880229037,\n\ \ \"acc_stderr\": 0.00959942926369336\n },\n \"harness|drop|3\": {\n\ \ \"em\": 0.001363255033557047,\n \"em_stderr\": 0.0003778609196461104,\n\ \ \"f1\": 0.05628670302013431,\n \"f1_stderr\": 0.0013136825712446111\n\ \ },\n \"harness|gsm8k|5\": {\n \"acc\": 0.07050796057619409,\n \ \ \"acc_stderr\": 0.0070515438139836135\n },\n \"harness|winogrande|5\"\ : {\n \"acc\": 0.7513812154696132,\n \"acc_stderr\": 0.012147314713403107\n\ \ }\n}\n```" repo_url: https://huggingface.co/LTC-AI-Labs/L2-7b-Base-WVG-Uncensored leaderboard_url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard point_of_contact: clementine@hf.co configs: - config_name: harness_arc_challenge_25 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|arc:challenge|25_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|arc:challenge|25_2023-10-03T10-58-44.594405.parquet' - config_name: harness_drop_3 data_files: - split: 2023_10_23T14_27_39.647282 path: - '**/details_harness|drop|3_2023-10-23T14-27-39.647282.parquet' - split: latest path: - '**/details_harness|drop|3_2023-10-23T14-27-39.647282.parquet' - config_name: harness_gsm8k_5 data_files: - split: 2023_10_23T14_27_39.647282 path: - '**/details_harness|gsm8k|5_2023-10-23T14-27-39.647282.parquet' - split: latest path: - '**/details_harness|gsm8k|5_2023-10-23T14-27-39.647282.parquet' - config_name: harness_hellaswag_10 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hellaswag|10_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hellaswag|10_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-abstract_algebra|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-anatomy|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-astronomy|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-business_ethics|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-college_biology|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-college_chemistry|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-college_computer_science|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-college_mathematics|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-college_medicine|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-college_physics|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-computer_security|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-conceptual_physics|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-econometrics|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-electrical_engineering|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-formal_logic|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-global_facts|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_biology|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_european_history|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_geography|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_physics|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_psychology|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_statistics|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_us_history|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_world_history|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-human_aging|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-human_sexuality|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-international_law|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-jurisprudence|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-logical_fallacies|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-machine_learning|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-management|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-marketing|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-medical_genetics|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-miscellaneous|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-moral_disputes|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-moral_scenarios|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-nutrition|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-philosophy|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-prehistory|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-professional_accounting|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-professional_law|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-professional_medicine|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-professional_psychology|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-public_relations|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-security_studies|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-sociology|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-virology|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-world_religions|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-abstract_algebra|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-anatomy|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-astronomy|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-business_ethics|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-college_biology|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-college_chemistry|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-college_computer_science|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-college_mathematics|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-college_medicine|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-college_physics|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-computer_security|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-conceptual_physics|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-econometrics|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-electrical_engineering|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-formal_logic|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-global_facts|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_biology|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_european_history|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_geography|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_physics|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_psychology|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_statistics|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_us_history|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-high_school_world_history|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-human_aging|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-human_sexuality|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-international_law|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-jurisprudence|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-logical_fallacies|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-machine_learning|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-management|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-marketing|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-medical_genetics|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-miscellaneous|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-moral_disputes|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-moral_scenarios|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-nutrition|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-philosophy|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-prehistory|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-professional_accounting|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-professional_law|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-professional_medicine|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-professional_psychology|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-public_relations|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-security_studies|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-sociology|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-virology|5_2023-10-03T10-58-44.594405.parquet' - '**/details_harness|hendrycksTest-world_religions|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_abstract_algebra_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-abstract_algebra|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-abstract_algebra|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_anatomy_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-anatomy|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-anatomy|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_astronomy_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-astronomy|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-astronomy|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_business_ethics_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-business_ethics|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-business_ethics|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_clinical_knowledge_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-clinical_knowledge|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_college_biology_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-college_biology|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-college_biology|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_college_chemistry_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-college_chemistry|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-college_chemistry|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_college_computer_science_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-college_computer_science|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-college_computer_science|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_college_mathematics_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-college_mathematics|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-college_mathematics|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_college_medicine_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-college_medicine|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-college_medicine|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_college_physics_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-college_physics|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-college_physics|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_computer_security_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-computer_security|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-computer_security|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_conceptual_physics_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-conceptual_physics|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-conceptual_physics|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_econometrics_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-econometrics|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-econometrics|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_electrical_engineering_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-electrical_engineering|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-electrical_engineering|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_elementary_mathematics_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-elementary_mathematics|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_formal_logic_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-formal_logic|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-formal_logic|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_global_facts_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-global_facts|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-global_facts|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_high_school_biology_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-high_school_biology|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_biology|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_high_school_chemistry_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_chemistry|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_high_school_computer_science_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_computer_science|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_high_school_european_history_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-high_school_european_history|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_european_history|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_high_school_geography_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-high_school_geography|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_geography|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_high_school_government_and_politics_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_high_school_macroeconomics_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_high_school_mathematics_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_mathematics|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_high_school_microeconomics_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_microeconomics|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_high_school_physics_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-high_school_physics|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_physics|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_high_school_psychology_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-high_school_psychology|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_psychology|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_high_school_statistics_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-high_school_statistics|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_statistics|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_high_school_us_history_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-high_school_us_history|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_us_history|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_high_school_world_history_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-high_school_world_history|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_world_history|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_human_aging_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-human_aging|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-human_aging|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_human_sexuality_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-human_sexuality|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-human_sexuality|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_international_law_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-international_law|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-international_law|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_jurisprudence_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-jurisprudence|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-jurisprudence|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_logical_fallacies_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-logical_fallacies|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-logical_fallacies|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_machine_learning_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-machine_learning|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-machine_learning|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_management_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-management|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-management|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_marketing_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-marketing|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-marketing|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_medical_genetics_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-medical_genetics|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-medical_genetics|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_miscellaneous_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-miscellaneous|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-miscellaneous|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_moral_disputes_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-moral_disputes|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-moral_disputes|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_moral_scenarios_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-moral_scenarios|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-moral_scenarios|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_nutrition_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-nutrition|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-nutrition|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_philosophy_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-philosophy|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-philosophy|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_prehistory_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-prehistory|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-prehistory|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_professional_accounting_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-professional_accounting|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-professional_accounting|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_professional_law_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-professional_law|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-professional_law|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_professional_medicine_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-professional_medicine|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-professional_medicine|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_professional_psychology_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-professional_psychology|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-professional_psychology|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_public_relations_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-public_relations|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-public_relations|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_security_studies_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-security_studies|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-security_studies|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_sociology_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-sociology|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-sociology|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_us_foreign_policy_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-us_foreign_policy|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_virology_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-virology|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-virology|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_hendrycksTest_world_religions_5 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|hendrycksTest-world_religions|5_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|hendrycksTest-world_religions|5_2023-10-03T10-58-44.594405.parquet' - config_name: harness_truthfulqa_mc_0 data_files: - split: 2023_10_03T10_58_44.594405 path: - '**/details_harness|truthfulqa:mc|0_2023-10-03T10-58-44.594405.parquet' - split: latest path: - '**/details_harness|truthfulqa:mc|0_2023-10-03T10-58-44.594405.parquet' - config_name: harness_winogrande_5 data_files: - split: 2023_10_23T14_27_39.647282 path: - '**/details_harness|winogrande|5_2023-10-23T14-27-39.647282.parquet' - split: latest path: - '**/details_harness|winogrande|5_2023-10-23T14-27-39.647282.parquet' - config_name: results data_files: - split: 2023_10_03T10_58_44.594405 path: - results_2023-10-03T10-58-44.594405.parquet - split: 2023_10_23T14_27_39.647282 path: - results_2023-10-23T14-27-39.647282.parquet - split: latest path: - results_2023-10-23T14-27-39.647282.parquet --- # Dataset Card for Evaluation run of LTC-AI-Labs/L2-7b-Base-WVG-Uncensored ## Dataset Description - **Homepage:** - **Repository:** https://huggingface.co/LTC-AI-Labs/L2-7b-Base-WVG-Uncensored - **Paper:** - **Leaderboard:** https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard - **Point of Contact:** clementine@hf.co ### Dataset Summary Dataset automatically created during the evaluation run of model [LTC-AI-Labs/L2-7b-Base-WVG-Uncensored](https://huggingface.co/LTC-AI-Labs/L2-7b-Base-WVG-Uncensored) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard). The dataset is composed of 64 configuration, each one coresponding to one of the evaluated task. The dataset has been created from 2 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results. An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the agregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)). To load the details from a run, you can for instance do the following: ```python from datasets import load_dataset data = load_dataset("open-llm-leaderboard/details_LTC-AI-Labs__L2-7b-Base-WVG-Uncensored", "harness_winogrande_5", split="train") ``` ## Latest results These are the [latest results from run 2023-10-23T14:27:39.647282](https://huggingface.co/datasets/open-llm-leaderboard/details_LTC-AI-Labs__L2-7b-Base-WVG-Uncensored/blob/main/results_2023-10-23T14-27-39.647282.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval): ```python { "all": { "em": 0.001363255033557047, "em_stderr": 0.0003778609196461104, "f1": 0.05628670302013431, "f1_stderr": 0.0013136825712446111, "acc": 0.4109445880229037, "acc_stderr": 0.00959942926369336 }, "harness|drop|3": { "em": 0.001363255033557047, "em_stderr": 0.0003778609196461104, "f1": 0.05628670302013431, "f1_stderr": 0.0013136825712446111 }, "harness|gsm8k|5": { "acc": 0.07050796057619409, "acc_stderr": 0.0070515438139836135 }, "harness|winogrande|5": { "acc": 0.7513812154696132, "acc_stderr": 0.012147314713403107 } } ``` ### Supported Tasks and Leaderboards [More Information Needed] ### Languages [More Information Needed] ## Dataset Structure ### Data Instances [More Information Needed] ### Data Fields [More Information Needed] ### Data Splits [More Information Needed] ## Dataset Creation ### Curation Rationale [More Information Needed] ### Source Data #### Initial Data Collection and Normalization [More Information Needed] #### Who are the source language producers? [More Information Needed] ### Annotations #### Annotation process [More Information Needed] #### Who are the annotators? [More Information Needed] ### Personal and Sensitive Information [More Information Needed] ## Considerations for Using the Data ### Social Impact of Dataset [More Information Needed] ### Discussion of Biases [More Information Needed] ### Other Known Limitations [More Information Needed] ## Additional Information ### Dataset Curators [More Information Needed] ### Licensing Information [More Information Needed] ### Citation Information [More Information Needed] ### Contributions [More Information Needed]
KanziRAY222/fuego-20230406-115747-446647
--- tags: - fuego fuego: id: 20230406-115747-446647 status: preparing script: make_a_video.py space_id: KanziRAY222/fuego-20230406-115747-446647 space_hardware: cpu-basic github_repo_id: lucidrains/make-a-video-pytorch github_repo_branch: main github_repo_sha: 4430a87da4474b2cec58fe3e0e1c581a328472e5 ---
kpriyanshu256/MultiTabQA-multitable_pretraining-Salesforce-codet5-base_train-latex-128000
--- dataset_info: features: - name: input_ids sequence: sequence: int32 - name: attention_mask sequence: sequence: int8 - name: labels sequence: sequence: int64 splits: - name: train num_bytes: 13336000 num_examples: 1000 download_size: 997181 dataset_size: 13336000 configs: - config_name: default data_files: - split: train path: data/train-* ---
CyberHarem/totoki_airi_idolmastercinderellagirls
--- license: mit task_categories: - text-to-image tags: - art - not-for-all-audiences size_categories: - n<1K --- # Dataset of totoki_airi/十時愛梨 (THE iDOLM@STER: Cinderella Girls) This is the dataset of totoki_airi/十時愛梨 (THE iDOLM@STER: Cinderella Girls), containing 500 images and their tags. The core tags of this character are `brown_hair, brown_eyes, breasts, twintails, large_breasts, bangs`, which are pruned in this dataset. Images are crawled from many sites (e.g. danbooru, pixiv, zerochan ...), the auto-crawling system is powered by [DeepGHS Team](https://github.com/deepghs)([huggingface organization](https://huggingface.co/deepghs)). ## List of Packages | Name | Images | Size | Download | Type | Description | |:-----------------|---------:|:-----------|:---------------------------------------------------------------------------------------------------------------------------------------|:-----------|:---------------------------------------------------------------------| | raw | 500 | 654.27 MiB | [Download](https://huggingface.co/datasets/CyberHarem/totoki_airi_idolmastercinderellagirls/resolve/main/dataset-raw.zip) | Waifuc-Raw | Raw data with meta information (min edge aligned to 1400 if larger). | | 800 | 500 | 375.41 MiB | [Download](https://huggingface.co/datasets/CyberHarem/totoki_airi_idolmastercinderellagirls/resolve/main/dataset-800.zip) | IMG+TXT | dataset with the shorter side not exceeding 800 pixels. | | stage3-p480-800 | 1209 | 805.52 MiB | [Download](https://huggingface.co/datasets/CyberHarem/totoki_airi_idolmastercinderellagirls/resolve/main/dataset-stage3-p480-800.zip) | IMG+TXT | 3-stage cropped dataset with the area not less than 480x480 pixels. | | 1200 | 500 | 580.34 MiB | [Download](https://huggingface.co/datasets/CyberHarem/totoki_airi_idolmastercinderellagirls/resolve/main/dataset-1200.zip) | IMG+TXT | dataset with the shorter side not exceeding 1200 pixels. | | stage3-p480-1200 | 1209 | 1.12 GiB | [Download](https://huggingface.co/datasets/CyberHarem/totoki_airi_idolmastercinderellagirls/resolve/main/dataset-stage3-p480-1200.zip) | IMG+TXT | 3-stage cropped dataset with the area not less than 480x480 pixels. | ### Load Raw Dataset with Waifuc We provide raw dataset (including tagged images) for [waifuc](https://deepghs.github.io/waifuc/main/tutorials/installation/index.html) loading. If you need this, just run the following code ```python import os import zipfile from huggingface_hub import hf_hub_download from waifuc.source import LocalSource # download raw archive file zip_file = hf_hub_download( repo_id='CyberHarem/totoki_airi_idolmastercinderellagirls', repo_type='dataset', filename='dataset-raw.zip', ) # extract files to your directory dataset_dir = 'dataset_dir' os.makedirs(dataset_dir, exist_ok=True) with zipfile.ZipFile(zip_file, 'r') as zf: zf.extractall(dataset_dir) # load the dataset with waifuc source = LocalSource(dataset_dir) for item in source: print(item.image, item.meta['filename'], item.meta['tags']) ``` ## List of Clusters List of tag clustering result, maybe some outfits can be mined here. ### Raw Text Version | # | Samples | Img-1 | Img-2 | Img-3 | Img-4 | Img-5 | Tags | |----:|----------:|:----------------------------------|:----------------------------------|:----------------------------------|:----------------------------------|:----------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| | 0 | 7 | ![](samples/0/clu0-sample0.png) | ![](samples/0/clu0-sample1.png) | ![](samples/0/clu0-sample2.png) | ![](samples/0/clu0-sample3.png) | ![](samples/0/clu0-sample4.png) | 1girl, blush, looking_at_viewer, nipples, navel, smile, solo, completely_nude, open_mouth, sweat, female_pubic_hair, hair_ornament, barefoot, breast_hold, heart_censor, pussy, simple_background, white_background | | 1 | 13 | ![](samples/1/clu1-sample0.png) | ![](samples/1/clu1-sample1.png) | ![](samples/1/clu1-sample2.png) | ![](samples/1/clu1-sample3.png) | ![](samples/1/clu1-sample4.png) | 1girl, bikini, cleavage, solo, blush, looking_at_viewer, smile, collarbone, navel, open_mouth, simple_background, sweat, white_background | | 2 | 22 | ![](samples/2/clu2-sample0.png) | ![](samples/2/clu2-sample1.png) | ![](samples/2/clu2-sample2.png) | ![](samples/2/clu2-sample3.png) | ![](samples/2/clu2-sample4.png) | 1girl, blush, heart_necklace, solo, collarbone, looking_at_viewer, cleavage, simple_background, white_background, open_mouth, hair_scrunchie, bare_shoulders, off_shoulder, smile, upper_body, striped_shirt, long_sleeves, sweat | | 3 | 8 | ![](samples/3/clu3-sample0.png) | ![](samples/3/clu3-sample1.png) | ![](samples/3/clu3-sample2.png) | ![](samples/3/clu3-sample3.png) | ![](samples/3/clu3-sample4.png) | 1girl, blush, looking_at_viewer, solo, smile, cleavage, heart_necklace | | 4 | 7 | ![](samples/4/clu4-sample0.png) | ![](samples/4/clu4-sample1.png) | ![](samples/4/clu4-sample2.png) | ![](samples/4/clu4-sample3.png) | ![](samples/4/clu4-sample4.png) | 1girl, blush, cleavage, looking_at_viewer, solo, fur_trim, open_mouth, rabbit_ears, :d, bare_shoulders, upper_body, heart_necklace, simple_background, crop_top, fake_animal_ears, hair_ornament, navel, sidelocks, white_background, white_gloves | | 5 | 8 | ![](samples/5/clu5-sample0.png) | ![](samples/5/clu5-sample1.png) | ![](samples/5/clu5-sample2.png) | ![](samples/5/clu5-sample3.png) | ![](samples/5/clu5-sample4.png) | 1girl, blush, solo, looking_at_viewer, black_thighhighs, open_mouth, striped, dress, sitting, :d, necklace, pantyshot, short_hair, skirt | | 6 | 13 | ![](samples/6/clu6-sample0.png) | ![](samples/6/clu6-sample1.png) | ![](samples/6/clu6-sample2.png) | ![](samples/6/clu6-sample3.png) | ![](samples/6/clu6-sample4.png) | 1girl, solo, blush, cleavage, dress, looking_at_viewer, bare_shoulders, open_mouth, white_gloves, :d, heart, tiara, bow, hair_ribbon, jewelry, microphone, frills | | 7 | 5 | ![](samples/7/clu7-sample0.png) | ![](samples/7/clu7-sample1.png) | ![](samples/7/clu7-sample2.png) | ![](samples/7/clu7-sample3.png) | ![](samples/7/clu7-sample4.png) | 1girl, cleavage, detached_collar, frills, maid_headdress, open_mouth, solo, :d, apron, hair_ribbon, looking_at_viewer, puffy_short_sleeves, blush, long_hair, red_dress, upper_body, bowtie, drill_hair, fruit, hairclip, heart, jewelry, plaid, striped | | 8 | 10 | ![](samples/8/clu8-sample0.png) | ![](samples/8/clu8-sample1.png) | ![](samples/8/clu8-sample2.png) | ![](samples/8/clu8-sample3.png) | ![](samples/8/clu8-sample4.png) | 1boy, 1girl, blush, hetero, nipples, penis, solo_focus, navel, sex, vaginal, looking_at_viewer, open_mouth, completely_nude, girl_on_top, long_hair, smile, cowgirl_position, mosaic_censoring, pov, pussy, spread_legs, sweat, cum, hair_ornament, scrunchie | | 9 | 9 | ![](samples/9/clu9-sample0.png) | ![](samples/9/clu9-sample1.png) | ![](samples/9/clu9-sample2.png) | ![](samples/9/clu9-sample3.png) | ![](samples/9/clu9-sample4.png) | 1boy, 1girl, blush, hetero, nipples, paizuri, penis, solo_focus, heart, smile, huge_breasts, nude, one_eye_closed, bar_censor, ejaculation, facial, looking_at_viewer, open_mouth, pov, sweat | | 10 | 7 | ![](samples/10/clu10-sample0.png) | ![](samples/10/clu10-sample1.png) | ![](samples/10/clu10-sample2.png) | ![](samples/10/clu10-sample3.png) | ![](samples/10/clu10-sample4.png) | 1girl, blue_skirt, long_hair, sleeveless_shirt, beret, bracelet, smile, solo, white_shirt, blush, looking_at_viewer, open_mouth, belt, black_headwear, frills, plaid_skirt, collarbone, earrings, heart_necklace, neck_ribbon | | 11 | 7 | ![](samples/11/clu11-sample0.png) | ![](samples/11/clu11-sample1.png) | ![](samples/11/clu11-sample2.png) | ![](samples/11/clu11-sample3.png) | ![](samples/11/clu11-sample4.png) | 1girl, fishnet_pantyhose, playboy_bunny, rabbit_ears, strapless_leotard, bare_shoulders, black_bowtie, blush, detached_collar, fake_animal_ears, looking_at_viewer, red_leotard, solo, simple_background, cleavage, frills, jewelry, long_hair, open_mouth, brown_pantyhose, cowboy_shot, nipples, smile, white_background | ### Table Version | # | Samples | Img-1 | Img-2 | Img-3 | Img-4 | Img-5 | 1girl | blush | looking_at_viewer | nipples | navel | smile | solo | completely_nude | open_mouth | sweat | female_pubic_hair | hair_ornament | barefoot | breast_hold | heart_censor | pussy | simple_background | white_background | bikini | cleavage | collarbone | heart_necklace | hair_scrunchie | bare_shoulders | off_shoulder | upper_body | striped_shirt | long_sleeves | fur_trim | rabbit_ears | :d | crop_top | fake_animal_ears | sidelocks | white_gloves | black_thighhighs | striped | dress | sitting | necklace | pantyshot | short_hair | skirt | heart | tiara | bow | hair_ribbon | jewelry | microphone | frills | detached_collar | maid_headdress | apron | puffy_short_sleeves | long_hair | red_dress | bowtie | drill_hair | fruit | hairclip | plaid | 1boy | hetero | penis | solo_focus | sex | vaginal | girl_on_top | cowgirl_position | mosaic_censoring | pov | spread_legs | cum | scrunchie | paizuri | huge_breasts | nude | one_eye_closed | bar_censor | ejaculation | facial | blue_skirt | sleeveless_shirt | beret | bracelet | white_shirt | belt | black_headwear | plaid_skirt | earrings | neck_ribbon | fishnet_pantyhose | playboy_bunny | strapless_leotard | black_bowtie | red_leotard | brown_pantyhose | cowboy_shot | |----:|----------:|:----------------------------------|:----------------------------------|:----------------------------------|:----------------------------------|:----------------------------------|:--------|:--------|:--------------------|:----------|:--------|:--------|:-------|:------------------|:-------------|:--------|:--------------------|:----------------|:-----------|:--------------|:---------------|:--------|:--------------------|:-------------------|:---------|:-----------|:-------------|:-----------------|:-----------------|:-----------------|:---------------|:-------------|:----------------|:---------------|:-----------|:--------------|:-----|:-----------|:-------------------|:------------|:---------------|:-------------------|:----------|:--------|:----------|:-----------|:------------|:-------------|:--------|:--------|:--------|:------|:--------------|:----------|:-------------|:---------|:------------------|:-----------------|:--------|:----------------------|:------------|:------------|:---------|:-------------|:--------|:-----------|:--------|:-------|:---------|:--------|:-------------|:------|:----------|:--------------|:-------------------|:-------------------|:------|:--------------|:------|:------------|:----------|:---------------|:-------|:-----------------|:-------------|:--------------|:---------|:-------------|:-------------------|:--------|:-----------|:--------------|:-------|:-----------------|:--------------|:-----------|:--------------|:--------------------|:----------------|:--------------------|:---------------|:--------------|:------------------|:--------------| | 0 | 7 | ![](samples/0/clu0-sample0.png) | ![](samples/0/clu0-sample1.png) | ![](samples/0/clu0-sample2.png) | ![](samples/0/clu0-sample3.png) | ![](samples/0/clu0-sample4.png) | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | 1 | 13 | ![](samples/1/clu1-sample0.png) | ![](samples/1/clu1-sample1.png) | ![](samples/1/clu1-sample2.png) | ![](samples/1/clu1-sample3.png) | ![](samples/1/clu1-sample4.png) | X | X | X | | X | X | X | | X | X | | | | | | | X | X | X | X | X | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | 2 | 22 | ![](samples/2/clu2-sample0.png) | ![](samples/2/clu2-sample1.png) | ![](samples/2/clu2-sample2.png) | ![](samples/2/clu2-sample3.png) | ![](samples/2/clu2-sample4.png) | X | X | X | | | X | X | | X | X | | | | | | | X | X | | X | X | X | X | X | X | X | X | X | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | 3 | 8 | ![](samples/3/clu3-sample0.png) | ![](samples/3/clu3-sample1.png) | ![](samples/3/clu3-sample2.png) | ![](samples/3/clu3-sample3.png) | ![](samples/3/clu3-sample4.png) | X | X | X | | | X | X | | | | | | | | | | | | | X | | X | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | 4 | 7 | ![](samples/4/clu4-sample0.png) | ![](samples/4/clu4-sample1.png) | ![](samples/4/clu4-sample2.png) | ![](samples/4/clu4-sample3.png) | ![](samples/4/clu4-sample4.png) | X | X | X | | X | | X | | X | | | X | | | | | X | X | | X | | X | | X | | X | | | X | X | X | X | X | X | X | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | 5 | 8 | ![](samples/5/clu5-sample0.png) | ![](samples/5/clu5-sample1.png) | ![](samples/5/clu5-sample2.png) | ![](samples/5/clu5-sample3.png) | ![](samples/5/clu5-sample4.png) | X | X | X | | | | X | | X | | | | | | | | | | | | | | | | | | | | | | X | | | | | X | X | X | X | X | X | X | X | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | 6 | 13 | ![](samples/6/clu6-sample0.png) | ![](samples/6/clu6-sample1.png) | ![](samples/6/clu6-sample2.png) | ![](samples/6/clu6-sample3.png) | ![](samples/6/clu6-sample4.png) | X | X | X | | | | X | | X | | | | | | | | | | | X | | | | X | | | | | | | X | | | | X | | | X | | | | | | X | X | X | X | X | X | X | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | 7 | 5 | ![](samples/7/clu7-sample0.png) | ![](samples/7/clu7-sample1.png) | ![](samples/7/clu7-sample2.png) | ![](samples/7/clu7-sample3.png) | ![](samples/7/clu7-sample4.png) | X | X | X | | | | X | | X | | | | | | | | | | | X | | | | | | X | | | | | X | | | | | | X | | | | | | | X | | | X | X | | X | X | X | X | X | X | X | X | X | X | X | X | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | 8 | 10 | ![](samples/8/clu8-sample0.png) | ![](samples/8/clu8-sample1.png) | ![](samples/8/clu8-sample2.png) | ![](samples/8/clu8-sample3.png) | ![](samples/8/clu8-sample4.png) | X | X | X | X | X | X | | X | X | X | | X | | | | X | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | X | | | | | | | X | X | X | X | X | X | X | X | X | X | X | X | X | | | | | | | | | | | | | | | | | | | | | | | | | | 9 | 9 | ![](samples/9/clu9-sample0.png) | ![](samples/9/clu9-sample1.png) | ![](samples/9/clu9-sample2.png) | ![](samples/9/clu9-sample3.png) | ![](samples/9/clu9-sample4.png) | X | X | X | X | | X | | | X | X | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | X | | | | | | | | | | | | | | | | | | X | X | X | X | | | | | | X | | | | X | X | X | X | X | X | X | | | | | | | | | | | | | | | | | | | 10 | 7 | ![](samples/10/clu10-sample0.png) | ![](samples/10/clu10-sample1.png) | ![](samples/10/clu10-sample2.png) | ![](samples/10/clu10-sample3.png) | ![](samples/10/clu10-sample4.png) | X | X | X | | | X | X | | X | | | | | | | | | | | | X | X | | | | | | | | | | | | | | | | | | | | | | | | | | | | X | | | | | X | | | | | | | | | | | | | | | | | | | | | | | | | | | X | X | X | X | X | X | X | X | X | X | | | | | | | | | 11 | 7 | ![](samples/11/clu11-sample0.png) | ![](samples/11/clu11-sample1.png) | ![](samples/11/clu11-sample2.png) | ![](samples/11/clu11-sample3.png) | ![](samples/11/clu11-sample4.png) | X | X | X | X | | X | X | | X | | | | | | | | X | X | | X | | | | X | | | | | | X | | | X | | | | | | | | | | | | | | | X | | X | X | | | | X | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | X | X | X | X | X | X | X |
Elliot4AI/databricksdatabricks-dolly-15k-chinese
--- license: cc-by-sa-3.0 task_categories: - question-answering - text-generation language: - zh tags: - biology - music - climate size_categories: - 10K<n<100K --- # Dataset Summary ## 🏡🏡🏡🏡Fine-tune Dataset:中文数据集🏡🏡🏡🏡 😀😀😀😀😀😀😀😀 这个数据集是databricks/databricks-dolly-15k的中文版本,是直接翻译过来,没有经过人为检查语法。 对databricks/databricks-dolly-15k的描述,请看他的dataset card。 😀😀😀😀😀😀😀😀 This data set is the Chinese version of databricks/databricks-dolly-15k, which is directly translated without human-checked grammar. For a description of databricks/databricks-dolly-15k, see its dataset card.
nlp-guild/intent-recognition-biomedical
--- license: mit --- [source](https://github.com/wangle1218/KBQA-for-Diagnosis/tree/main/nlu/bert_intent_recognition/data)
youndukn/ROMANTIC_TEXT
--- license: apache-2.0 ---
rungalileo/sst2_tiny_subset_32
--- dataset_info: features: - name: idx dtype: int32 - name: sentence dtype: string - name: label dtype: class_label: names: '0': negative '1': positive splits: - name: train num_bytes: 2433 num_examples: 32 - name: validation num_bytes: 2433 num_examples: 32 - name: test num_bytes: 2433 num_examples: 32 download_size: 11709 dataset_size: 7299 --- # Dataset Card for "sst2_tiny_subset_32" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
open-llm-leaderboard/details_ChaoticNeutrals__Eris_Remix_DPO_7B
--- pretty_name: Evaluation run of ChaoticNeutrals/Eris_Remix_DPO_7B dataset_summary: "Dataset automatically created during the evaluation run of model\ \ [ChaoticNeutrals/Eris_Remix_DPO_7B](https://huggingface.co/ChaoticNeutrals/Eris_Remix_DPO_7B)\ \ on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).\n\ \nThe dataset is composed of 63 configuration, each one coresponding to one of the\ \ evaluated task.\n\nThe dataset has been created from 1 run(s). Each run can be\ \ found as a specific split in each configuration, the split being named using the\ \ timestamp of the run.The \"train\" split is always pointing to the latest results.\n\ \nAn additional configuration \"results\" store all the aggregated results of the\ \ run (and is used to compute and display the aggregated metrics on the [Open LLM\ \ Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)).\n\ \nTo load the details from a run, you can for instance do the following:\n```python\n\ from datasets import load_dataset\ndata = load_dataset(\"open-llm-leaderboard/details_ChaoticNeutrals__Eris_Remix_DPO_7B\"\ ,\n\t\"harness_winogrande_5\",\n\tsplit=\"train\")\n```\n\n## Latest results\n\n\ These are the [latest results from run 2024-03-07T11:43:51.829104](https://huggingface.co/datasets/open-llm-leaderboard/details_ChaoticNeutrals__Eris_Remix_DPO_7B/blob/main/results_2024-03-07T11-43-51.829104.json)(note\ \ that their might be results for other tasks in the repos if successive evals didn't\ \ cover the same tasks. You find each in the results and the \"latest\" split for\ \ each eval):\n\n```python\n{\n \"all\": {\n \"acc\": 0.6583336356900076,\n\ \ \"acc_stderr\": 0.031977540942622035,\n \"acc_norm\": 0.6579180688874828,\n\ \ \"acc_norm_stderr\": 0.03264643438189693,\n \"mc1\": 0.5385556915544676,\n\ \ \"mc1_stderr\": 0.017451384104637452,\n \"mc2\": 0.6891701037064589,\n\ \ \"mc2_stderr\": 0.01500545106241752\n },\n \"harness|arc:challenge|25\"\ : {\n \"acc\": 0.6988054607508533,\n \"acc_stderr\": 0.01340674176784763,\n\ \ \"acc_norm\": 0.7244027303754266,\n \"acc_norm_stderr\": 0.01305716965576184\n\ \ },\n \"harness|hellaswag|10\": {\n \"acc\": 0.7110137422824139,\n\ \ \"acc_stderr\": 0.0045236511840162615,\n \"acc_norm\": 0.8803027285401315,\n\ \ \"acc_norm_stderr\": 0.0032394379508979255\n },\n \"harness|hendrycksTest-abstract_algebra|5\"\ : {\n \"acc\": 0.34,\n \"acc_stderr\": 0.04760952285695235,\n \ \ \"acc_norm\": 0.34,\n \"acc_norm_stderr\": 0.04760952285695235\n \ \ },\n \"harness|hendrycksTest-anatomy|5\": {\n \"acc\": 0.6592592592592592,\n\ \ \"acc_stderr\": 0.040943762699967926,\n \"acc_norm\": 0.6592592592592592,\n\ \ \"acc_norm_stderr\": 0.040943762699967926\n },\n \"harness|hendrycksTest-astronomy|5\"\ : {\n \"acc\": 0.7236842105263158,\n \"acc_stderr\": 0.03639057569952928,\n\ \ \"acc_norm\": 0.7236842105263158,\n \"acc_norm_stderr\": 0.03639057569952928\n\ \ },\n \"harness|hendrycksTest-business_ethics|5\": {\n \"acc\": 0.64,\n\ \ \"acc_stderr\": 0.04824181513244218,\n \"acc_norm\": 0.64,\n \ \ \"acc_norm_stderr\": 0.04824181513244218\n },\n \"harness|hendrycksTest-clinical_knowledge|5\"\ : {\n \"acc\": 0.720754716981132,\n \"acc_stderr\": 0.027611163402399715,\n\ \ \"acc_norm\": 0.720754716981132,\n \"acc_norm_stderr\": 0.027611163402399715\n\ \ },\n \"harness|hendrycksTest-college_biology|5\": {\n \"acc\": 0.7708333333333334,\n\ \ \"acc_stderr\": 0.03514697467862388,\n \"acc_norm\": 0.7708333333333334,\n\ \ \"acc_norm_stderr\": 0.03514697467862388\n },\n \"harness|hendrycksTest-college_chemistry|5\"\ : {\n \"acc\": 0.47,\n \"acc_stderr\": 0.050161355804659205,\n \ \ \"acc_norm\": 0.47,\n \"acc_norm_stderr\": 0.050161355804659205\n \ \ },\n \"harness|hendrycksTest-college_computer_science|5\": {\n \"\ acc\": 0.53,\n \"acc_stderr\": 0.050161355804659205,\n \"acc_norm\"\ : 0.53,\n \"acc_norm_stderr\": 0.050161355804659205\n },\n \"harness|hendrycksTest-college_mathematics|5\"\ : {\n \"acc\": 0.31,\n \"acc_stderr\": 0.04648231987117316,\n \ \ \"acc_norm\": 0.31,\n \"acc_norm_stderr\": 0.04648231987117316\n \ \ },\n \"harness|hendrycksTest-college_medicine|5\": {\n \"acc\": 0.6589595375722543,\n\ \ \"acc_stderr\": 0.03614665424180826,\n \"acc_norm\": 0.6589595375722543,\n\ \ \"acc_norm_stderr\": 0.03614665424180826\n },\n \"harness|hendrycksTest-college_physics|5\"\ : {\n \"acc\": 0.4215686274509804,\n \"acc_stderr\": 0.04913595201274498,\n\ \ \"acc_norm\": 0.4215686274509804,\n \"acc_norm_stderr\": 0.04913595201274498\n\ \ },\n \"harness|hendrycksTest-computer_security|5\": {\n \"acc\":\ \ 0.79,\n \"acc_stderr\": 0.04093601807403326,\n \"acc_norm\": 0.79,\n\ \ \"acc_norm_stderr\": 0.04093601807403326\n },\n \"harness|hendrycksTest-conceptual_physics|5\"\ : {\n \"acc\": 0.5872340425531914,\n \"acc_stderr\": 0.03218471141400351,\n\ \ \"acc_norm\": 0.5872340425531914,\n \"acc_norm_stderr\": 0.03218471141400351\n\ \ },\n \"harness|hendrycksTest-econometrics|5\": {\n \"acc\": 0.5263157894736842,\n\ \ \"acc_stderr\": 0.046970851366478626,\n \"acc_norm\": 0.5263157894736842,\n\ \ \"acc_norm_stderr\": 0.046970851366478626\n },\n \"harness|hendrycksTest-electrical_engineering|5\"\ : {\n \"acc\": 0.5793103448275863,\n \"acc_stderr\": 0.0411391498118926,\n\ \ \"acc_norm\": 0.5793103448275863,\n \"acc_norm_stderr\": 0.0411391498118926\n\ \ },\n \"harness|hendrycksTest-elementary_mathematics|5\": {\n \"acc\"\ : 0.43386243386243384,\n \"acc_stderr\": 0.02552503438247489,\n \"\ acc_norm\": 0.43386243386243384,\n \"acc_norm_stderr\": 0.02552503438247489\n\ \ },\n \"harness|hendrycksTest-formal_logic|5\": {\n \"acc\": 0.48412698412698413,\n\ \ \"acc_stderr\": 0.04469881854072606,\n \"acc_norm\": 0.48412698412698413,\n\ \ \"acc_norm_stderr\": 0.04469881854072606\n },\n \"harness|hendrycksTest-global_facts|5\"\ : {\n \"acc\": 0.35,\n \"acc_stderr\": 0.047937248544110196,\n \ \ \"acc_norm\": 0.35,\n \"acc_norm_stderr\": 0.047937248544110196\n \ \ },\n \"harness|hendrycksTest-high_school_biology|5\": {\n \"acc\"\ : 0.7903225806451613,\n \"acc_stderr\": 0.023157879349083525,\n \"\ acc_norm\": 0.7903225806451613,\n \"acc_norm_stderr\": 0.023157879349083525\n\ \ },\n \"harness|hendrycksTest-high_school_chemistry|5\": {\n \"acc\"\ : 0.5123152709359606,\n \"acc_stderr\": 0.035169204442208966,\n \"\ acc_norm\": 0.5123152709359606,\n \"acc_norm_stderr\": 0.035169204442208966\n\ \ },\n \"harness|hendrycksTest-high_school_computer_science|5\": {\n \ \ \"acc\": 0.68,\n \"acc_stderr\": 0.04688261722621505,\n \"acc_norm\"\ : 0.68,\n \"acc_norm_stderr\": 0.04688261722621505\n },\n \"harness|hendrycksTest-high_school_european_history|5\"\ : {\n \"acc\": 0.7757575757575758,\n \"acc_stderr\": 0.032568666616811015,\n\ \ \"acc_norm\": 0.7757575757575758,\n \"acc_norm_stderr\": 0.032568666616811015\n\ \ },\n \"harness|hendrycksTest-high_school_geography|5\": {\n \"acc\"\ : 0.8131313131313131,\n \"acc_stderr\": 0.027772533334218967,\n \"\ acc_norm\": 0.8131313131313131,\n \"acc_norm_stderr\": 0.027772533334218967\n\ \ },\n \"harness|hendrycksTest-high_school_government_and_politics|5\": {\n\ \ \"acc\": 0.8911917098445595,\n \"acc_stderr\": 0.022473253332768776,\n\ \ \"acc_norm\": 0.8911917098445595,\n \"acc_norm_stderr\": 0.022473253332768776\n\ \ },\n \"harness|hendrycksTest-high_school_macroeconomics|5\": {\n \ \ \"acc\": 0.6692307692307692,\n \"acc_stderr\": 0.023854795680971125,\n\ \ \"acc_norm\": 0.6692307692307692,\n \"acc_norm_stderr\": 0.023854795680971125\n\ \ },\n \"harness|hendrycksTest-high_school_mathematics|5\": {\n \"\ acc\": 0.37037037037037035,\n \"acc_stderr\": 0.02944316932303154,\n \ \ \"acc_norm\": 0.37037037037037035,\n \"acc_norm_stderr\": 0.02944316932303154\n\ \ },\n \"harness|hendrycksTest-high_school_microeconomics|5\": {\n \ \ \"acc\": 0.680672268907563,\n \"acc_stderr\": 0.030283995525884396,\n \ \ \"acc_norm\": 0.680672268907563,\n \"acc_norm_stderr\": 0.030283995525884396\n\ \ },\n \"harness|hendrycksTest-high_school_physics|5\": {\n \"acc\"\ : 0.36423841059602646,\n \"acc_stderr\": 0.03929111781242742,\n \"\ acc_norm\": 0.36423841059602646,\n \"acc_norm_stderr\": 0.03929111781242742\n\ \ },\n \"harness|hendrycksTest-high_school_psychology|5\": {\n \"acc\"\ : 0.8385321100917431,\n \"acc_stderr\": 0.015776239256163227,\n \"\ acc_norm\": 0.8385321100917431,\n \"acc_norm_stderr\": 0.015776239256163227\n\ \ },\n \"harness|hendrycksTest-high_school_statistics|5\": {\n \"acc\"\ : 0.5138888888888888,\n \"acc_stderr\": 0.034086558679777494,\n \"\ acc_norm\": 0.5138888888888888,\n \"acc_norm_stderr\": 0.034086558679777494\n\ \ },\n \"harness|hendrycksTest-high_school_us_history|5\": {\n \"acc\"\ : 0.8676470588235294,\n \"acc_stderr\": 0.023784297520918853,\n \"\ acc_norm\": 0.8676470588235294,\n \"acc_norm_stderr\": 0.023784297520918853\n\ \ },\n \"harness|hendrycksTest-high_school_world_history|5\": {\n \"\ acc\": 0.8185654008438819,\n \"acc_stderr\": 0.02508596114457966,\n \ \ \"acc_norm\": 0.8185654008438819,\n \"acc_norm_stderr\": 0.02508596114457966\n\ \ },\n \"harness|hendrycksTest-human_aging|5\": {\n \"acc\": 0.6860986547085202,\n\ \ \"acc_stderr\": 0.031146796482972465,\n \"acc_norm\": 0.6860986547085202,\n\ \ \"acc_norm_stderr\": 0.031146796482972465\n },\n \"harness|hendrycksTest-human_sexuality|5\"\ : {\n \"acc\": 0.8244274809160306,\n \"acc_stderr\": 0.03336820338476073,\n\ \ \"acc_norm\": 0.8244274809160306,\n \"acc_norm_stderr\": 0.03336820338476073\n\ \ },\n \"harness|hendrycksTest-international_law|5\": {\n \"acc\":\ \ 0.7768595041322314,\n \"acc_stderr\": 0.03800754475228733,\n \"\ acc_norm\": 0.7768595041322314,\n \"acc_norm_stderr\": 0.03800754475228733\n\ \ },\n \"harness|hendrycksTest-jurisprudence|5\": {\n \"acc\": 0.7870370370370371,\n\ \ \"acc_stderr\": 0.0395783547198098,\n \"acc_norm\": 0.7870370370370371,\n\ \ \"acc_norm_stderr\": 0.0395783547198098\n },\n \"harness|hendrycksTest-logical_fallacies|5\"\ : {\n \"acc\": 0.7730061349693251,\n \"acc_stderr\": 0.03291099578615769,\n\ \ \"acc_norm\": 0.7730061349693251,\n \"acc_norm_stderr\": 0.03291099578615769\n\ \ },\n \"harness|hendrycksTest-machine_learning|5\": {\n \"acc\": 0.4732142857142857,\n\ \ \"acc_stderr\": 0.047389751192741546,\n \"acc_norm\": 0.4732142857142857,\n\ \ \"acc_norm_stderr\": 0.047389751192741546\n },\n \"harness|hendrycksTest-management|5\"\ : {\n \"acc\": 0.7572815533980582,\n \"acc_stderr\": 0.04245022486384495,\n\ \ \"acc_norm\": 0.7572815533980582,\n \"acc_norm_stderr\": 0.04245022486384495\n\ \ },\n \"harness|hendrycksTest-marketing|5\": {\n \"acc\": 0.8803418803418803,\n\ \ \"acc_stderr\": 0.021262719400406964,\n \"acc_norm\": 0.8803418803418803,\n\ \ \"acc_norm_stderr\": 0.021262719400406964\n },\n \"harness|hendrycksTest-medical_genetics|5\"\ : {\n \"acc\": 0.7,\n \"acc_stderr\": 0.046056618647183814,\n \ \ \"acc_norm\": 0.7,\n \"acc_norm_stderr\": 0.046056618647183814\n \ \ },\n \"harness|hendrycksTest-miscellaneous|5\": {\n \"acc\": 0.8275862068965517,\n\ \ \"acc_stderr\": 0.013507943909371802,\n \"acc_norm\": 0.8275862068965517,\n\ \ \"acc_norm_stderr\": 0.013507943909371802\n },\n \"harness|hendrycksTest-moral_disputes|5\"\ : {\n \"acc\": 0.7369942196531792,\n \"acc_stderr\": 0.023703099525258172,\n\ \ \"acc_norm\": 0.7369942196531792,\n \"acc_norm_stderr\": 0.023703099525258172\n\ \ },\n \"harness|hendrycksTest-moral_scenarios|5\": {\n \"acc\": 0.4145251396648045,\n\ \ \"acc_stderr\": 0.016476342210254,\n \"acc_norm\": 0.4145251396648045,\n\ \ \"acc_norm_stderr\": 0.016476342210254\n },\n \"harness|hendrycksTest-nutrition|5\"\ : {\n \"acc\": 0.7254901960784313,\n \"acc_stderr\": 0.025553169991826524,\n\ \ \"acc_norm\": 0.7254901960784313,\n \"acc_norm_stderr\": 0.025553169991826524\n\ \ },\n \"harness|hendrycksTest-philosophy|5\": {\n \"acc\": 0.7138263665594855,\n\ \ \"acc_stderr\": 0.02567025924218893,\n \"acc_norm\": 0.7138263665594855,\n\ \ \"acc_norm_stderr\": 0.02567025924218893\n },\n \"harness|hendrycksTest-prehistory|5\"\ : {\n \"acc\": 0.7469135802469136,\n \"acc_stderr\": 0.024191808600712992,\n\ \ \"acc_norm\": 0.7469135802469136,\n \"acc_norm_stderr\": 0.024191808600712992\n\ \ },\n \"harness|hendrycksTest-professional_accounting|5\": {\n \"\ acc\": 0.4929078014184397,\n \"acc_stderr\": 0.02982449855912901,\n \ \ \"acc_norm\": 0.4929078014184397,\n \"acc_norm_stderr\": 0.02982449855912901\n\ \ },\n \"harness|hendrycksTest-professional_law|5\": {\n \"acc\": 0.4745762711864407,\n\ \ \"acc_stderr\": 0.012753716929101004,\n \"acc_norm\": 0.4745762711864407,\n\ \ \"acc_norm_stderr\": 0.012753716929101004\n },\n \"harness|hendrycksTest-professional_medicine|5\"\ : {\n \"acc\": 0.6727941176470589,\n \"acc_stderr\": 0.028501452860396556,\n\ \ \"acc_norm\": 0.6727941176470589,\n \"acc_norm_stderr\": 0.028501452860396556\n\ \ },\n \"harness|hendrycksTest-professional_psychology|5\": {\n \"\ acc\": 0.6748366013071896,\n \"acc_stderr\": 0.01895088677080631,\n \ \ \"acc_norm\": 0.6748366013071896,\n \"acc_norm_stderr\": 0.01895088677080631\n\ \ },\n \"harness|hendrycksTest-public_relations|5\": {\n \"acc\": 0.6727272727272727,\n\ \ \"acc_stderr\": 0.0449429086625209,\n \"acc_norm\": 0.6727272727272727,\n\ \ \"acc_norm_stderr\": 0.0449429086625209\n },\n \"harness|hendrycksTest-security_studies|5\"\ : {\n \"acc\": 0.7346938775510204,\n \"acc_stderr\": 0.028263889943784603,\n\ \ \"acc_norm\": 0.7346938775510204,\n \"acc_norm_stderr\": 0.028263889943784603\n\ \ },\n \"harness|hendrycksTest-sociology|5\": {\n \"acc\": 0.845771144278607,\n\ \ \"acc_stderr\": 0.025538433368578337,\n \"acc_norm\": 0.845771144278607,\n\ \ \"acc_norm_stderr\": 0.025538433368578337\n },\n \"harness|hendrycksTest-us_foreign_policy|5\"\ : {\n \"acc\": 0.86,\n \"acc_stderr\": 0.0348735088019777,\n \ \ \"acc_norm\": 0.86,\n \"acc_norm_stderr\": 0.0348735088019777\n },\n\ \ \"harness|hendrycksTest-virology|5\": {\n \"acc\": 0.5542168674698795,\n\ \ \"acc_stderr\": 0.03869543323472101,\n \"acc_norm\": 0.5542168674698795,\n\ \ \"acc_norm_stderr\": 0.03869543323472101\n },\n \"harness|hendrycksTest-world_religions|5\"\ : {\n \"acc\": 0.8245614035087719,\n \"acc_stderr\": 0.02917088550072766,\n\ \ \"acc_norm\": 0.8245614035087719,\n \"acc_norm_stderr\": 0.02917088550072766\n\ \ },\n \"harness|truthfulqa:mc|0\": {\n \"mc1\": 0.5385556915544676,\n\ \ \"mc1_stderr\": 0.017451384104637452,\n \"mc2\": 0.6891701037064589,\n\ \ \"mc2_stderr\": 0.01500545106241752\n },\n \"harness|winogrande|5\"\ : {\n \"acc\": 0.8476716653512234,\n \"acc_stderr\": 0.010099208246065586\n\ \ },\n \"harness|gsm8k|5\": {\n \"acc\": 0.6884003032600455,\n \ \ \"acc_stderr\": 0.012757375376754941\n }\n}\n```" repo_url: https://huggingface.co/ChaoticNeutrals/Eris_Remix_DPO_7B leaderboard_url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard point_of_contact: clementine@hf.co configs: - config_name: harness_arc_challenge_25 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|arc:challenge|25_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|arc:challenge|25_2024-03-07T11-43-51.829104.parquet' - config_name: harness_gsm8k_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|gsm8k|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|gsm8k|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hellaswag_10 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hellaswag|10_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hellaswag|10_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-abstract_algebra|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-anatomy|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-astronomy|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-business_ethics|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-clinical_knowledge|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-college_biology|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-college_chemistry|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-college_computer_science|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-college_mathematics|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-college_medicine|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-college_physics|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-computer_security|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-conceptual_physics|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-econometrics|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-electrical_engineering|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-elementary_mathematics|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-formal_logic|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-global_facts|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_biology|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_chemistry|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_computer_science|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_european_history|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_geography|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_mathematics|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_microeconomics|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_physics|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_psychology|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_statistics|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_us_history|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_world_history|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-human_aging|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-human_sexuality|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-international_law|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-jurisprudence|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-logical_fallacies|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-machine_learning|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-management|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-marketing|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-medical_genetics|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-miscellaneous|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-moral_disputes|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-moral_scenarios|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-nutrition|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-philosophy|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-prehistory|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-professional_accounting|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-professional_law|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-professional_medicine|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-professional_psychology|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-public_relations|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-security_studies|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-sociology|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-us_foreign_policy|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-virology|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-world_religions|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-abstract_algebra|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-anatomy|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-astronomy|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-business_ethics|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-clinical_knowledge|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-college_biology|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-college_chemistry|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-college_computer_science|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-college_mathematics|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-college_medicine|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-college_physics|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-computer_security|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-conceptual_physics|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-econometrics|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-electrical_engineering|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-elementary_mathematics|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-formal_logic|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-global_facts|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_biology|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_chemistry|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_computer_science|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_european_history|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_geography|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_mathematics|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_microeconomics|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_physics|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_psychology|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_statistics|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_us_history|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-high_school_world_history|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-human_aging|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-human_sexuality|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-international_law|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-jurisprudence|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-logical_fallacies|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-machine_learning|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-management|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-marketing|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-medical_genetics|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-miscellaneous|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-moral_disputes|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-moral_scenarios|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-nutrition|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-philosophy|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-prehistory|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-professional_accounting|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-professional_law|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-professional_medicine|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-professional_psychology|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-public_relations|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-security_studies|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-sociology|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-us_foreign_policy|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-virology|5_2024-03-07T11-43-51.829104.parquet' - '**/details_harness|hendrycksTest-world_religions|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_abstract_algebra_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-abstract_algebra|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-abstract_algebra|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_anatomy_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-anatomy|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-anatomy|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_astronomy_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-astronomy|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-astronomy|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_business_ethics_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-business_ethics|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-business_ethics|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_clinical_knowledge_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-clinical_knowledge|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-clinical_knowledge|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_college_biology_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-college_biology|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-college_biology|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_college_chemistry_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-college_chemistry|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-college_chemistry|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_college_computer_science_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-college_computer_science|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-college_computer_science|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_college_mathematics_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-college_mathematics|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-college_mathematics|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_college_medicine_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-college_medicine|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-college_medicine|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_college_physics_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-college_physics|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-college_physics|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_computer_security_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-computer_security|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-computer_security|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_conceptual_physics_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-conceptual_physics|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-conceptual_physics|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_econometrics_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-econometrics|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-econometrics|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_electrical_engineering_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-electrical_engineering|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-electrical_engineering|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_elementary_mathematics_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-elementary_mathematics|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-elementary_mathematics|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_formal_logic_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-formal_logic|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-formal_logic|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_global_facts_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-global_facts|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-global_facts|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_high_school_biology_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-high_school_biology|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_biology|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_high_school_chemistry_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-high_school_chemistry|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_chemistry|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_high_school_computer_science_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-high_school_computer_science|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_computer_science|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_high_school_european_history_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-high_school_european_history|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_european_history|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_high_school_geography_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-high_school_geography|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_geography|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_high_school_government_and_politics_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_government_and_politics|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_high_school_macroeconomics_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_macroeconomics|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_high_school_mathematics_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-high_school_mathematics|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_mathematics|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_high_school_microeconomics_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-high_school_microeconomics|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_microeconomics|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_high_school_physics_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-high_school_physics|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_physics|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_high_school_psychology_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-high_school_psychology|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_psychology|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_high_school_statistics_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-high_school_statistics|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_statistics|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_high_school_us_history_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-high_school_us_history|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_us_history|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_high_school_world_history_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-high_school_world_history|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-high_school_world_history|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_human_aging_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-human_aging|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-human_aging|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_human_sexuality_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-human_sexuality|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-human_sexuality|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_international_law_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-international_law|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-international_law|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_jurisprudence_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-jurisprudence|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-jurisprudence|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_logical_fallacies_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-logical_fallacies|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-logical_fallacies|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_machine_learning_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-machine_learning|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-machine_learning|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_management_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-management|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-management|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_marketing_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-marketing|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-marketing|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_medical_genetics_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-medical_genetics|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-medical_genetics|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_miscellaneous_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-miscellaneous|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-miscellaneous|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_moral_disputes_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-moral_disputes|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-moral_disputes|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_moral_scenarios_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-moral_scenarios|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-moral_scenarios|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_nutrition_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-nutrition|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-nutrition|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_philosophy_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-philosophy|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-philosophy|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_prehistory_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-prehistory|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-prehistory|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_professional_accounting_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-professional_accounting|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-professional_accounting|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_professional_law_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-professional_law|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-professional_law|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_professional_medicine_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-professional_medicine|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-professional_medicine|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_professional_psychology_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-professional_psychology|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-professional_psychology|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_public_relations_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-public_relations|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-public_relations|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_security_studies_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-security_studies|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-security_studies|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_sociology_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-sociology|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-sociology|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_us_foreign_policy_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-us_foreign_policy|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-us_foreign_policy|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_virology_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-virology|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-virology|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_hendrycksTest_world_religions_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|hendrycksTest-world_religions|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|hendrycksTest-world_religions|5_2024-03-07T11-43-51.829104.parquet' - config_name: harness_truthfulqa_mc_0 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|truthfulqa:mc|0_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|truthfulqa:mc|0_2024-03-07T11-43-51.829104.parquet' - config_name: harness_winogrande_5 data_files: - split: 2024_03_07T11_43_51.829104 path: - '**/details_harness|winogrande|5_2024-03-07T11-43-51.829104.parquet' - split: latest path: - '**/details_harness|winogrande|5_2024-03-07T11-43-51.829104.parquet' - config_name: results data_files: - split: 2024_03_07T11_43_51.829104 path: - results_2024-03-07T11-43-51.829104.parquet - split: latest path: - results_2024-03-07T11-43-51.829104.parquet --- # Dataset Card for Evaluation run of ChaoticNeutrals/Eris_Remix_DPO_7B <!-- Provide a quick summary of the dataset. --> Dataset automatically created during the evaluation run of model [ChaoticNeutrals/Eris_Remix_DPO_7B](https://huggingface.co/ChaoticNeutrals/Eris_Remix_DPO_7B) on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard). The dataset is composed of 63 configuration, each one coresponding to one of the evaluated task. The dataset has been created from 1 run(s). Each run can be found as a specific split in each configuration, the split being named using the timestamp of the run.The "train" split is always pointing to the latest results. An additional configuration "results" store all the aggregated results of the run (and is used to compute and display the aggregated metrics on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)). To load the details from a run, you can for instance do the following: ```python from datasets import load_dataset data = load_dataset("open-llm-leaderboard/details_ChaoticNeutrals__Eris_Remix_DPO_7B", "harness_winogrande_5", split="train") ``` ## Latest results These are the [latest results from run 2024-03-07T11:43:51.829104](https://huggingface.co/datasets/open-llm-leaderboard/details_ChaoticNeutrals__Eris_Remix_DPO_7B/blob/main/results_2024-03-07T11-43-51.829104.json)(note that their might be results for other tasks in the repos if successive evals didn't cover the same tasks. You find each in the results and the "latest" split for each eval): ```python { "all": { "acc": 0.6583336356900076, "acc_stderr": 0.031977540942622035, "acc_norm": 0.6579180688874828, "acc_norm_stderr": 0.03264643438189693, "mc1": 0.5385556915544676, "mc1_stderr": 0.017451384104637452, "mc2": 0.6891701037064589, "mc2_stderr": 0.01500545106241752 }, "harness|arc:challenge|25": { "acc": 0.6988054607508533, "acc_stderr": 0.01340674176784763, "acc_norm": 0.7244027303754266, "acc_norm_stderr": 0.01305716965576184 }, "harness|hellaswag|10": { "acc": 0.7110137422824139, "acc_stderr": 0.0045236511840162615, "acc_norm": 0.8803027285401315, "acc_norm_stderr": 0.0032394379508979255 }, "harness|hendrycksTest-abstract_algebra|5": { "acc": 0.34, "acc_stderr": 0.04760952285695235, "acc_norm": 0.34, "acc_norm_stderr": 0.04760952285695235 }, "harness|hendrycksTest-anatomy|5": { "acc": 0.6592592592592592, "acc_stderr": 0.040943762699967926, "acc_norm": 0.6592592592592592, "acc_norm_stderr": 0.040943762699967926 }, "harness|hendrycksTest-astronomy|5": { "acc": 0.7236842105263158, "acc_stderr": 0.03639057569952928, "acc_norm": 0.7236842105263158, "acc_norm_stderr": 0.03639057569952928 }, "harness|hendrycksTest-business_ethics|5": { "acc": 0.64, "acc_stderr": 0.04824181513244218, "acc_norm": 0.64, "acc_norm_stderr": 0.04824181513244218 }, "harness|hendrycksTest-clinical_knowledge|5": { "acc": 0.720754716981132, "acc_stderr": 0.027611163402399715, "acc_norm": 0.720754716981132, "acc_norm_stderr": 0.027611163402399715 }, "harness|hendrycksTest-college_biology|5": { "acc": 0.7708333333333334, "acc_stderr": 0.03514697467862388, "acc_norm": 0.7708333333333334, "acc_norm_stderr": 0.03514697467862388 }, "harness|hendrycksTest-college_chemistry|5": { "acc": 0.47, "acc_stderr": 0.050161355804659205, "acc_norm": 0.47, "acc_norm_stderr": 0.050161355804659205 }, "harness|hendrycksTest-college_computer_science|5": { "acc": 0.53, "acc_stderr": 0.050161355804659205, "acc_norm": 0.53, "acc_norm_stderr": 0.050161355804659205 }, "harness|hendrycksTest-college_mathematics|5": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "harness|hendrycksTest-college_medicine|5": { "acc": 0.6589595375722543, "acc_stderr": 0.03614665424180826, "acc_norm": 0.6589595375722543, "acc_norm_stderr": 0.03614665424180826 }, "harness|hendrycksTest-college_physics|5": { "acc": 0.4215686274509804, "acc_stderr": 0.04913595201274498, "acc_norm": 0.4215686274509804, "acc_norm_stderr": 0.04913595201274498 }, "harness|hendrycksTest-computer_security|5": { "acc": 0.79, "acc_stderr": 0.04093601807403326, "acc_norm": 0.79, "acc_norm_stderr": 0.04093601807403326 }, "harness|hendrycksTest-conceptual_physics|5": { "acc": 0.5872340425531914, "acc_stderr": 0.03218471141400351, "acc_norm": 0.5872340425531914, "acc_norm_stderr": 0.03218471141400351 }, "harness|hendrycksTest-econometrics|5": { "acc": 0.5263157894736842, "acc_stderr": 0.046970851366478626, "acc_norm": 0.5263157894736842, "acc_norm_stderr": 0.046970851366478626 }, "harness|hendrycksTest-electrical_engineering|5": { "acc": 0.5793103448275863, "acc_stderr": 0.0411391498118926, "acc_norm": 0.5793103448275863, "acc_norm_stderr": 0.0411391498118926 }, "harness|hendrycksTest-elementary_mathematics|5": { "acc": 0.43386243386243384, "acc_stderr": 0.02552503438247489, "acc_norm": 0.43386243386243384, "acc_norm_stderr": 0.02552503438247489 }, "harness|hendrycksTest-formal_logic|5": { "acc": 0.48412698412698413, "acc_stderr": 0.04469881854072606, "acc_norm": 0.48412698412698413, "acc_norm_stderr": 0.04469881854072606 }, "harness|hendrycksTest-global_facts|5": { "acc": 0.35, "acc_stderr": 0.047937248544110196, "acc_norm": 0.35, "acc_norm_stderr": 0.047937248544110196 }, "harness|hendrycksTest-high_school_biology|5": { "acc": 0.7903225806451613, "acc_stderr": 0.023157879349083525, "acc_norm": 0.7903225806451613, "acc_norm_stderr": 0.023157879349083525 }, "harness|hendrycksTest-high_school_chemistry|5": { "acc": 0.5123152709359606, "acc_stderr": 0.035169204442208966, "acc_norm": 0.5123152709359606, "acc_norm_stderr": 0.035169204442208966 }, "harness|hendrycksTest-high_school_computer_science|5": { "acc": 0.68, "acc_stderr": 0.04688261722621505, "acc_norm": 0.68, "acc_norm_stderr": 0.04688261722621505 }, "harness|hendrycksTest-high_school_european_history|5": { "acc": 0.7757575757575758, "acc_stderr": 0.032568666616811015, "acc_norm": 0.7757575757575758, "acc_norm_stderr": 0.032568666616811015 }, "harness|hendrycksTest-high_school_geography|5": { "acc": 0.8131313131313131, "acc_stderr": 0.027772533334218967, "acc_norm": 0.8131313131313131, "acc_norm_stderr": 0.027772533334218967 }, "harness|hendrycksTest-high_school_government_and_politics|5": { "acc": 0.8911917098445595, "acc_stderr": 0.022473253332768776, "acc_norm": 0.8911917098445595, "acc_norm_stderr": 0.022473253332768776 }, "harness|hendrycksTest-high_school_macroeconomics|5": { "acc": 0.6692307692307692, "acc_stderr": 0.023854795680971125, "acc_norm": 0.6692307692307692, "acc_norm_stderr": 0.023854795680971125 }, "harness|hendrycksTest-high_school_mathematics|5": { "acc": 0.37037037037037035, "acc_stderr": 0.02944316932303154, "acc_norm": 0.37037037037037035, "acc_norm_stderr": 0.02944316932303154 }, "harness|hendrycksTest-high_school_microeconomics|5": { "acc": 0.680672268907563, "acc_stderr": 0.030283995525884396, "acc_norm": 0.680672268907563, "acc_norm_stderr": 0.030283995525884396 }, "harness|hendrycksTest-high_school_physics|5": { "acc": 0.36423841059602646, "acc_stderr": 0.03929111781242742, "acc_norm": 0.36423841059602646, "acc_norm_stderr": 0.03929111781242742 }, "harness|hendrycksTest-high_school_psychology|5": { "acc": 0.8385321100917431, "acc_stderr": 0.015776239256163227, "acc_norm": 0.8385321100917431, "acc_norm_stderr": 0.015776239256163227 }, "harness|hendrycksTest-high_school_statistics|5": { "acc": 0.5138888888888888, "acc_stderr": 0.034086558679777494, "acc_norm": 0.5138888888888888, "acc_norm_stderr": 0.034086558679777494 }, "harness|hendrycksTest-high_school_us_history|5": { "acc": 0.8676470588235294, "acc_stderr": 0.023784297520918853, "acc_norm": 0.8676470588235294, "acc_norm_stderr": 0.023784297520918853 }, "harness|hendrycksTest-high_school_world_history|5": { "acc": 0.8185654008438819, "acc_stderr": 0.02508596114457966, "acc_norm": 0.8185654008438819, "acc_norm_stderr": 0.02508596114457966 }, "harness|hendrycksTest-human_aging|5": { "acc": 0.6860986547085202, "acc_stderr": 0.031146796482972465, "acc_norm": 0.6860986547085202, "acc_norm_stderr": 0.031146796482972465 }, "harness|hendrycksTest-human_sexuality|5": { "acc": 0.8244274809160306, "acc_stderr": 0.03336820338476073, "acc_norm": 0.8244274809160306, "acc_norm_stderr": 0.03336820338476073 }, "harness|hendrycksTest-international_law|5": { "acc": 0.7768595041322314, "acc_stderr": 0.03800754475228733, "acc_norm": 0.7768595041322314, "acc_norm_stderr": 0.03800754475228733 }, "harness|hendrycksTest-jurisprudence|5": { "acc": 0.7870370370370371, "acc_stderr": 0.0395783547198098, "acc_norm": 0.7870370370370371, "acc_norm_stderr": 0.0395783547198098 }, "harness|hendrycksTest-logical_fallacies|5": { "acc": 0.7730061349693251, "acc_stderr": 0.03291099578615769, "acc_norm": 0.7730061349693251, "acc_norm_stderr": 0.03291099578615769 }, "harness|hendrycksTest-machine_learning|5": { "acc": 0.4732142857142857, "acc_stderr": 0.047389751192741546, "acc_norm": 0.4732142857142857, "acc_norm_stderr": 0.047389751192741546 }, "harness|hendrycksTest-management|5": { "acc": 0.7572815533980582, "acc_stderr": 0.04245022486384495, "acc_norm": 0.7572815533980582, "acc_norm_stderr": 0.04245022486384495 }, "harness|hendrycksTest-marketing|5": { "acc": 0.8803418803418803, "acc_stderr": 0.021262719400406964, "acc_norm": 0.8803418803418803, "acc_norm_stderr": 0.021262719400406964 }, "harness|hendrycksTest-medical_genetics|5": { "acc": 0.7, "acc_stderr": 0.046056618647183814, "acc_norm": 0.7, "acc_norm_stderr": 0.046056618647183814 }, "harness|hendrycksTest-miscellaneous|5": { "acc": 0.8275862068965517, "acc_stderr": 0.013507943909371802, "acc_norm": 0.8275862068965517, "acc_norm_stderr": 0.013507943909371802 }, "harness|hendrycksTest-moral_disputes|5": { "acc": 0.7369942196531792, "acc_stderr": 0.023703099525258172, "acc_norm": 0.7369942196531792, "acc_norm_stderr": 0.023703099525258172 }, "harness|hendrycksTest-moral_scenarios|5": { "acc": 0.4145251396648045, "acc_stderr": 0.016476342210254, "acc_norm": 0.4145251396648045, "acc_norm_stderr": 0.016476342210254 }, "harness|hendrycksTest-nutrition|5": { "acc": 0.7254901960784313, "acc_stderr": 0.025553169991826524, "acc_norm": 0.7254901960784313, "acc_norm_stderr": 0.025553169991826524 }, "harness|hendrycksTest-philosophy|5": { "acc": 0.7138263665594855, "acc_stderr": 0.02567025924218893, "acc_norm": 0.7138263665594855, "acc_norm_stderr": 0.02567025924218893 }, "harness|hendrycksTest-prehistory|5": { "acc": 0.7469135802469136, "acc_stderr": 0.024191808600712992, "acc_norm": 0.7469135802469136, "acc_norm_stderr": 0.024191808600712992 }, "harness|hendrycksTest-professional_accounting|5": { "acc": 0.4929078014184397, "acc_stderr": 0.02982449855912901, "acc_norm": 0.4929078014184397, "acc_norm_stderr": 0.02982449855912901 }, "harness|hendrycksTest-professional_law|5": { "acc": 0.4745762711864407, "acc_stderr": 0.012753716929101004, "acc_norm": 0.4745762711864407, "acc_norm_stderr": 0.012753716929101004 }, "harness|hendrycksTest-professional_medicine|5": { "acc": 0.6727941176470589, "acc_stderr": 0.028501452860396556, "acc_norm": 0.6727941176470589, "acc_norm_stderr": 0.028501452860396556 }, "harness|hendrycksTest-professional_psychology|5": { "acc": 0.6748366013071896, "acc_stderr": 0.01895088677080631, "acc_norm": 0.6748366013071896, "acc_norm_stderr": 0.01895088677080631 }, "harness|hendrycksTest-public_relations|5": { "acc": 0.6727272727272727, "acc_stderr": 0.0449429086625209, "acc_norm": 0.6727272727272727, "acc_norm_stderr": 0.0449429086625209 }, "harness|hendrycksTest-security_studies|5": { "acc": 0.7346938775510204, "acc_stderr": 0.028263889943784603, "acc_norm": 0.7346938775510204, "acc_norm_stderr": 0.028263889943784603 }, "harness|hendrycksTest-sociology|5": { "acc": 0.845771144278607, "acc_stderr": 0.025538433368578337, "acc_norm": 0.845771144278607, "acc_norm_stderr": 0.025538433368578337 }, "harness|hendrycksTest-us_foreign_policy|5": { "acc": 0.86, "acc_stderr": 0.0348735088019777, "acc_norm": 0.86, "acc_norm_stderr": 0.0348735088019777 }, "harness|hendrycksTest-virology|5": { "acc": 0.5542168674698795, "acc_stderr": 0.03869543323472101, "acc_norm": 0.5542168674698795, "acc_norm_stderr": 0.03869543323472101 }, "harness|hendrycksTest-world_religions|5": { "acc": 0.8245614035087719, "acc_stderr": 0.02917088550072766, "acc_norm": 0.8245614035087719, "acc_norm_stderr": 0.02917088550072766 }, "harness|truthfulqa:mc|0": { "mc1": 0.5385556915544676, "mc1_stderr": 0.017451384104637452, "mc2": 0.6891701037064589, "mc2_stderr": 0.01500545106241752 }, "harness|winogrande|5": { "acc": 0.8476716653512234, "acc_stderr": 0.010099208246065586 }, "harness|gsm8k|5": { "acc": 0.6884003032600455, "acc_stderr": 0.012757375376754941 } } ``` ## Dataset Details ### Dataset Description <!-- Provide a longer summary of what this dataset is. --> - **Curated by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] ### Dataset Sources [optional] <!-- Provide the basic links for the dataset. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the dataset is intended to be used. --> ### Direct Use <!-- This section describes suitable use cases for the dataset. --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. --> [More Information Needed] ## Dataset Structure <!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. --> [More Information Needed] ## Dataset Creation ### Curation Rationale <!-- Motivation for the creation of this dataset. --> [More Information Needed] ### Source Data <!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). --> #### Data Collection and Processing <!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. --> [More Information Needed] #### Who are the source data producers? <!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. --> [More Information Needed] ### Annotations [optional] <!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. --> #### Annotation process <!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. --> [More Information Needed] #### Who are the annotators? <!-- This section describes the people or systems who created the annotations. --> [More Information Needed] #### Personal and Sensitive Information <!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users should be made aware of the risks, biases and limitations of the dataset. More information needed for further recommendations. ## Citation [optional] <!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Dataset Card Authors [optional] [More Information Needed] ## Dataset Card Contact [More Information Needed]
shabani1/image-text-demo
--- dataset_info: features: - name: text dtype: string - name: image dtype: image splits: - name: train num_bytes: 38726344.0 num_examples: 21 download_size: 38646616 dataset_size: 38726344.0 configs: - config_name: default data_files: - split: train path: data/train-* ---
CyberHarem/suzutsuki_kantaicollection
--- license: mit task_categories: - text-to-image tags: - art - not-for-all-audiences size_categories: - n<1K --- # Dataset of suzutsuki/涼月 (Kantai Collection) This is the dataset of suzutsuki/涼月 (Kantai Collection), containing 500 images and their tags. The core tags of this character are `long_hair, one_side_up, headband, grey_hair, breasts, hair_between_eyes, grey_eyes, white_hair, blue_eyes`, which are pruned in this dataset. Images are crawled from many sites (e.g. danbooru, pixiv, zerochan ...), the auto-crawling system is powered by [DeepGHS Team](https://github.com/deepghs)([huggingface organization](https://huggingface.co/deepghs)). ## List of Packages | Name | Images | Size | Download | Type | Description | |:-----------------|---------:|:-----------|:----------------------------------------------------------------------------------------------------------------------------|:-----------|:---------------------------------------------------------------------| | raw | 500 | 602.34 MiB | [Download](https://huggingface.co/datasets/CyberHarem/suzutsuki_kantaicollection/resolve/main/dataset-raw.zip) | Waifuc-Raw | Raw data with meta information (min edge aligned to 1400 if larger). | | 800 | 500 | 369.73 MiB | [Download](https://huggingface.co/datasets/CyberHarem/suzutsuki_kantaicollection/resolve/main/dataset-800.zip) | IMG+TXT | dataset with the shorter side not exceeding 800 pixels. | | stage3-p480-800 | 1224 | 819.26 MiB | [Download](https://huggingface.co/datasets/CyberHarem/suzutsuki_kantaicollection/resolve/main/dataset-stage3-p480-800.zip) | IMG+TXT | 3-stage cropped dataset with the area not less than 480x480 pixels. | | 1200 | 500 | 544.68 MiB | [Download](https://huggingface.co/datasets/CyberHarem/suzutsuki_kantaicollection/resolve/main/dataset-1200.zip) | IMG+TXT | dataset with the shorter side not exceeding 1200 pixels. | | stage3-p480-1200 | 1224 | 1.08 GiB | [Download](https://huggingface.co/datasets/CyberHarem/suzutsuki_kantaicollection/resolve/main/dataset-stage3-p480-1200.zip) | IMG+TXT | 3-stage cropped dataset with the area not less than 480x480 pixels. | ### Load Raw Dataset with Waifuc We provide raw dataset (including tagged images) for [waifuc](https://deepghs.github.io/waifuc/main/tutorials/installation/index.html) loading. If you need this, just run the following code ```python import os import zipfile from huggingface_hub import hf_hub_download from waifuc.source import LocalSource # download raw archive file zip_file = hf_hub_download( repo_id='CyberHarem/suzutsuki_kantaicollection', repo_type='dataset', filename='dataset-raw.zip', ) # extract files to your directory dataset_dir = 'dataset_dir' os.makedirs(dataset_dir, exist_ok=True) with zipfile.ZipFile(zip_file, 'r') as zf: zf.extractall(dataset_dir) # load the dataset with waifuc source = LocalSource(dataset_dir) for item in source: print(item.image, item.meta['filename'], item.meta['tags']) ``` ## List of Clusters List of tag clustering result, maybe some outfits can be mined here. ### Raw Text Version | # | Samples | Img-1 | Img-2 | Img-3 | Img-4 | Img-5 | Tags | |----:|----------:|:--------------------------------|:--------------------------------|:--------------------------------|:--------------------------------|:--------------------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| | 0 | 16 | ![](samples/0/clu0-sample0.png) | ![](samples/0/clu0-sample1.png) | ![](samples/0/clu0-sample2.png) | ![](samples/0/clu0-sample3.png) | ![](samples/0/clu0-sample4.png) | 1girl, black_bikini, solo, cleavage, hachimaki, grey_jacket, looking_at_viewer, clothes_writing, black_headband, navel, blush, smile, collarbone, medium_breasts, simple_background, cowboy_shot, large_breasts, white_background, open_jacket, sarong | | 1 | 7 | ![](samples/1/clu1-sample0.png) | ![](samples/1/clu1-sample1.png) | ![](samples/1/clu1-sample2.png) | ![](samples/1/clu1-sample3.png) | ![](samples/1/clu1-sample4.png) | 1girl, black_bikini, cleavage, simple_background, solo, white_background, black_headband, medium_breasts, twitter_username, dated, grey_jacket, hachimaki, looking_at_viewer, one-hour_drawing_challenge, blush, clothes_writing, collarbone, upper_body | | 2 | 15 | ![](samples/2/clu2-sample0.png) | ![](samples/2/clu2-sample1.png) | ![](samples/2/clu2-sample2.png) | ![](samples/2/clu2-sample3.png) | ![](samples/2/clu2-sample4.png) | 1girl, black_sailor_collar, hachimaki, serafuku, solo, upper_body, white_bodysuit, clothes_writing, grey_jacket, looking_at_viewer, simple_background, black_headband, white_background, anchor_symbol, white_gloves, white_neckerchief | | 3 | 24 | ![](samples/3/clu3-sample0.png) | ![](samples/3/clu3-sample1.png) | ![](samples/3/clu3-sample2.png) | ![](samples/3/clu3-sample3.png) | ![](samples/3/clu3-sample4.png) | 1girl, clothes_writing, pleated_skirt, white_bodysuit, white_skirt, hachimaki, serafuku, solo, simple_background, white_gloves, black_sailor_collar, miniskirt, white_background, black_headband, looking_at_viewer, grey_jacket, cowboy_shot, pantyhose, white_necktie, anchor_symbol, smile | | 4 | 9 | ![](samples/4/clu4-sample0.png) | ![](samples/4/clu4-sample1.png) | ![](samples/4/clu4-sample2.png) | ![](samples/4/clu4-sample3.png) | ![](samples/4/clu4-sample4.png) | black_gloves, black_skirt, clothes_writing, grey_neckerchief, hachimaki, microskirt, pleated_skirt, serafuku, short_sleeves, white_headband, white_sailor_collar, 1girl, shawl, closed_mouth, half_gloves, grey_thighhighs, smile, cowboy_shot, solo_focus | | 5 | 5 | ![](samples/5/clu5-sample0.png) | ![](samples/5/clu5-sample1.png) | ![](samples/5/clu5-sample2.png) | ![](samples/5/clu5-sample3.png) | ![](samples/5/clu5-sample4.png) | 1girl, grey_neckerchief, hachimaki, serafuku, solo, upper_body, white_background, white_headband, white_sailor_collar, black_gloves, clothes_writing, simple_background, closed_mouth, half_gloves, smile | | 6 | 5 | ![](samples/6/clu6-sample0.png) | ![](samples/6/clu6-sample1.png) | ![](samples/6/clu6-sample2.png) | ![](samples/6/clu6-sample3.png) | ![](samples/6/clu6-sample4.png) | 1girl, detached_collar, fake_animal_ears, playboy_bunny, rabbit_ears, rabbit_tail, solo, alternate_costume, blush, fake_tail, looking_at_viewer, simple_background, white_background, white_gloves, black_leotard, cleavage, cowboy_shot, medium_breasts, white_pantyhose, elbow_gloves, jacket, smile, wrist_cuffs | | 7 | 8 | ![](samples/7/clu7-sample0.png) | ![](samples/7/clu7-sample1.png) | ![](samples/7/clu7-sample2.png) | ![](samples/7/clu7-sample3.png) | ![](samples/7/clu7-sample4.png) | 1boy, hachimaki, hetero, 1girl, large_breasts, nipples, penis, clothes_writing, solo_focus, blush, bar_censor, paizuri, sex, sweat, black_bikini, cum, mosaic_censoring, smile, vaginal | | 8 | 9 | ![](samples/8/clu8-sample0.png) | ![](samples/8/clu8-sample1.png) | ![](samples/8/clu8-sample2.png) | ![](samples/8/clu8-sample3.png) | ![](samples/8/clu8-sample4.png) | 1girl, medium_breasts, solo, blush, panties, cowboy_shot, navel, simple_background, underwear_only, bra, closed_mouth, looking_at_viewer, white_background, cleavage, collarbone | | 9 | 5 | ![](samples/9/clu9-sample0.png) | ![](samples/9/clu9-sample1.png) | ![](samples/9/clu9-sample2.png) | ![](samples/9/clu9-sample3.png) | ![](samples/9/clu9-sample4.png) | 1girl, black_dress, enmaided, frilled_apron, maid_headdress, solo, white_apron, cowboy_shot, maid_apron, looking_at_viewer, one-hour_drawing_challenge, puffy_sleeves, simple_background, smile, artist_logo, dated, long_sleeves, white_gloves, white_thighhighs | ### Table Version | # | Samples | Img-1 | Img-2 | Img-3 | Img-4 | Img-5 | 1girl | black_bikini | solo | cleavage | hachimaki | grey_jacket | looking_at_viewer | clothes_writing | black_headband | navel | blush | smile | collarbone | medium_breasts | simple_background | cowboy_shot | large_breasts | white_background | open_jacket | sarong | twitter_username | dated | one-hour_drawing_challenge | upper_body | black_sailor_collar | serafuku | white_bodysuit | anchor_symbol | white_gloves | white_neckerchief | pleated_skirt | white_skirt | miniskirt | pantyhose | white_necktie | black_gloves | black_skirt | grey_neckerchief | microskirt | short_sleeves | white_headband | white_sailor_collar | shawl | closed_mouth | half_gloves | grey_thighhighs | solo_focus | detached_collar | fake_animal_ears | playboy_bunny | rabbit_ears | rabbit_tail | alternate_costume | fake_tail | black_leotard | white_pantyhose | elbow_gloves | jacket | wrist_cuffs | 1boy | hetero | nipples | penis | bar_censor | paizuri | sex | sweat | cum | mosaic_censoring | vaginal | panties | underwear_only | bra | black_dress | enmaided | frilled_apron | maid_headdress | white_apron | maid_apron | puffy_sleeves | artist_logo | long_sleeves | white_thighhighs | |----:|----------:|:--------------------------------|:--------------------------------|:--------------------------------|:--------------------------------|:--------------------------------|:--------|:---------------|:-------|:-----------|:------------|:--------------|:--------------------|:------------------|:-----------------|:--------|:--------|:--------|:-------------|:-----------------|:--------------------|:--------------|:----------------|:-------------------|:--------------|:---------|:-------------------|:--------|:-----------------------------|:-------------|:----------------------|:-----------|:-----------------|:----------------|:---------------|:--------------------|:----------------|:--------------|:------------|:------------|:----------------|:---------------|:--------------|:-------------------|:-------------|:----------------|:-----------------|:----------------------|:--------|:---------------|:--------------|:------------------|:-------------|:------------------|:-------------------|:----------------|:--------------|:--------------|:--------------------|:------------|:----------------|:------------------|:---------------|:---------|:--------------|:-------|:---------|:----------|:--------|:-------------|:----------|:------|:--------|:------|:-------------------|:----------|:----------|:-----------------|:------|:--------------|:-----------|:----------------|:-----------------|:--------------|:-------------|:----------------|:--------------|:---------------|:-------------------| | 0 | 16 | ![](samples/0/clu0-sample0.png) | ![](samples/0/clu0-sample1.png) | ![](samples/0/clu0-sample2.png) | ![](samples/0/clu0-sample3.png) | ![](samples/0/clu0-sample4.png) | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | X | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | 1 | 7 | ![](samples/1/clu1-sample0.png) | ![](samples/1/clu1-sample1.png) | ![](samples/1/clu1-sample2.png) | ![](samples/1/clu1-sample3.png) | ![](samples/1/clu1-sample4.png) | X | X | X | X | X | X | X | X | X | | X | | X | X | X | | | X | | | X | X | X | X | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | 2 | 15 | ![](samples/2/clu2-sample0.png) | ![](samples/2/clu2-sample1.png) | ![](samples/2/clu2-sample2.png) | ![](samples/2/clu2-sample3.png) | ![](samples/2/clu2-sample4.png) | X | | X | | X | X | X | X | X | | | | | | X | | | X | | | | | | X | X | X | X | X | X | X | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | 3 | 24 | ![](samples/3/clu3-sample0.png) | ![](samples/3/clu3-sample1.png) | ![](samples/3/clu3-sample2.png) | ![](samples/3/clu3-sample3.png) | ![](samples/3/clu3-sample4.png) | X | | X | | X | X | X | X | X | | | X | | | X | X | | X | | | | | | | X | X | X | X | X | | X | X | X | X | X | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | 4 | 9 | ![](samples/4/clu4-sample0.png) | ![](samples/4/clu4-sample1.png) | ![](samples/4/clu4-sample2.png) | ![](samples/4/clu4-sample3.png) | ![](samples/4/clu4-sample4.png) | X | | | | X | | | X | | | | X | | | | X | | | | | | | | | | X | | | | | X | | | | | X | X | X | X | X | X | X | X | X | X | X | X | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | 5 | 5 | ![](samples/5/clu5-sample0.png) | ![](samples/5/clu5-sample1.png) | ![](samples/5/clu5-sample2.png) | ![](samples/5/clu5-sample3.png) | ![](samples/5/clu5-sample4.png) | X | | X | | X | | | X | | | | X | | | X | | | X | | | | | | X | | X | | | | | | | | | | X | | X | | | X | X | | X | X | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | 6 | 5 | ![](samples/6/clu6-sample0.png) | ![](samples/6/clu6-sample1.png) | ![](samples/6/clu6-sample2.png) | ![](samples/6/clu6-sample3.png) | ![](samples/6/clu6-sample4.png) | X | | X | X | | | X | | | | X | X | | X | X | X | | X | | | | | | | | | | | X | | | | | | | | | | | | | | | | | | | X | X | X | X | X | X | X | X | X | X | X | X | | | | | | | | | | | | | | | | | | | | | | | | | | 7 | 8 | ![](samples/7/clu7-sample0.png) | ![](samples/7/clu7-sample1.png) | ![](samples/7/clu7-sample2.png) | ![](samples/7/clu7-sample3.png) | ![](samples/7/clu7-sample4.png) | X | X | | | X | | | X | | | X | X | | | | | X | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | X | | | | | | | | | | | | | X | X | X | X | X | X | X | X | X | X | X | | | | | | | | | | | | | | | 8 | 9 | ![](samples/8/clu8-sample0.png) | ![](samples/8/clu8-sample1.png) | ![](samples/8/clu8-sample2.png) | ![](samples/8/clu8-sample3.png) | ![](samples/8/clu8-sample4.png) | X | | X | X | | | X | | | X | X | | X | X | X | X | | X | | | | | | | | | | | | | | | | | | | | | | | | | | X | | | | | | | | | | | | | | | | | | | | | | | | | | | X | X | X | | | | | | | | | | | | 9 | 5 | ![](samples/9/clu9-sample0.png) | ![](samples/9/clu9-sample1.png) | ![](samples/9/clu9-sample2.png) | ![](samples/9/clu9-sample3.png) | ![](samples/9/clu9-sample4.png) | X | | X | | | | X | | | | | X | | | X | X | | | | | | X | X | | | | | | X | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | X | X | X | X | X | X | X | X | X | X |
nguyenthanhdo/dolphin_mqa_details
--- dataset_info: features: - name: instruction dtype: string - name: input dtype: string - name: output dtype: string splits: - name: train num_bytes: 26369871.746988524 num_examples: 15037 download_size: 10922205 dataset_size: 26369871.746988524 configs: - config_name: default data_files: - split: train path: data/train-* --- # Dataset Card for "dolphin_mqa_details" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
jaiganesan/real_splitted_image_200
--- dataset_info: features: - name: image dtype: image - name: text sequence: string splits: - name: train num_bytes: 26260519.0 num_examples: 200 download_size: 26234337 dataset_size: 26260519.0 configs: - config_name: default data_files: - split: train path: data/train-* ---
Juanid14317/Urdu1111SentimentAnalysis
--- dataset_info: features: - name: text dtype: string - name: label dtype: int64 - name: __index_level_0__ dtype: int64 splits: - name: train num_bytes: 151146.02409638555 num_examples: 896 - name: test num_bytes: 16868.975903614457 num_examples: 100 download_size: 98809 dataset_size: 168015.0 configs: - config_name: default data_files: - split: train path: data/train-* - split: test path: data/test-* ---
LibraTree/cqa_spec
--- license: mit ---
Hack90/ncbi_genbank_part_5
--- configs: - config_name: default data_files: - split: train path: data/train-* dataset_info: features: - name: id dtype: string - name: sequence dtype: string - name: name dtype: string - name: description dtype: string - name: features dtype: int64 - name: seq_length dtype: int64 splits: - name: train num_bytes: 20489073390 num_examples: 33130 download_size: 9505723044 dataset_size: 20489073390 --- # Dataset Card for "ncbi_genbank_part_5" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
jtatman/open-instruct-uncensored-alpaca
--- dataset_info: features: - name: user dtype: string - name: assistant dtype: string - name: text dtype: string splits: - name: train num_bytes: 2040635964 num_examples: 1255224 download_size: 922350127 dataset_size: 2040635964 configs: - config_name: default data_files: - split: train path: data/train-* --- # Dataset Card for "open-instruct-uncensored-alpaca" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
Zildembergue/locutor
--- license: openrail ---
anlp/sentence_anno
--- configs: - config_name: default data_files: - split: train path: data/train-* dataset_info: features: - name: sentences sequence: string - name: new_gt sequence: string splits: - name: train num_bytes: 1201528 num_examples: 990 download_size: 244599 dataset_size: 1201528 --- # Dataset Card for "sentence_anno" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
praveenku32k/Transformer_kaggle_data
--- license: mit ---
Multimodal-Fatima/OxfordPets_train
--- dataset_info: features: - name: image dtype: image - name: label dtype: class_label: names: '0': abyssinian '1': american bulldog '2': american pit bull terrier '3': basset hound '4': beagle '5': bengal '6': birman '7': bombay '8': boxer '9': british shorthair '10': chihuahua '11': egyptian mau '12': english cocker spaniel '13': english setter '14': german shorthaired '15': great pyrenees '16': havanese '17': japanese chin '18': keeshond '19': leonberger '20': maine coon '21': miniature pinscher '22': newfoundland '23': persian '24': pomeranian '25': pug '26': ragdoll '27': russian blue '28': saint bernard '29': samoyed '30': scottish terrier '31': shiba inu '32': siamese '33': sphynx '34': staffordshire bull terrier '35': wheaten terrier '36': yorkshire terrier - name: species dtype: class_label: names: '0': Cat '1': Dog - name: id dtype: int64 - name: clip_tags_ViT_L_14 sequence: string - name: blip_caption dtype: string - name: LLM_Description_opt175b_downstream_tasks_ViT_L_14 sequence: string - name: LLM_Description_gpt3_downstream_tasks_ViT_L_14 sequence: string - name: clip_tags_ViT_L_14_ensemble_specific dtype: string - name: clip_tags_ViT_L_14_simple_specific dtype: string - name: LLM_Description_gpt3_downstream_tasks_visual_genome_ViT_L_14 sequence: string - name: clip_tags_ViT_L_14with_openai_classes sequence: string - name: clip_tags_ViT_L_14_wo_openai_classes sequence: string - name: clip_tags_ViT_L_14_with_openai_classes sequence: string - name: Attributes_ViT_L_14_text_davinci_003 sequence: string - name: Attributes_ViT_L_14_text_davinci_003_full sequence: string - name: Attributes_ViT_L_14_text_davinci_003_oxfordpets sequence: string - name: clip_tags_ViT_B_16_simple_specific dtype: string - name: clip_tags_ViT_B_16_ensemble_specific dtype: string - name: clip_tags_ViT_B_32_simple_specific dtype: string - name: clip_tags_ViT_B_32_ensemble_specific dtype: string - name: Attributes_ViT_B_16_descriptors_text_davinci_003_full sequence: string - name: Attributes_LAION_ViT_H_14_2B_descriptors_text_davinci_003_full sequence: string - name: clip_tags_LAION_ViT_H_14_2B_simple_specific dtype: string - name: clip_tags_LAION_ViT_H_14_2B_ensemble_specific dtype: string splits: - name: train num_bytes: 386730161.36 num_examples: 3680 download_size: 378295172 dataset_size: 386730161.36 --- # Dataset Card for "OxfordPets_train" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
AdapterOcean/med_alpaca_standardized_cluster_63
--- dataset_info: features: - name: text dtype: string - name: conversation_id dtype: int64 - name: embedding sequence: float64 - name: cluster dtype: int64 splits: - name: train num_bytes: 51371332 num_examples: 5574 download_size: 13852216 dataset_size: 51371332 configs: - config_name: default data_files: - split: train path: data/train-* --- # Dataset Card for "med_alpaca_standardized_cluster_63" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
louisbrulenaudet/code-legion-honneur-medaille-militaire-ordre-national-merite
--- license: apache-2.0 language: - fr multilinguality: - monolingual tags: - finetuning - legal - french law - droit français - Code de la Légion d'honneur, de la Médaille militaire et de l'ordre national du Mérite source_datasets: - original pretty_name: Code de la Légion d'honneur, de la Médaille militaire et de l'ordre national du Mérite task_categories: - text-generation - table-question-answering - summarization - text-retrieval - question-answering - text-classification size_categories: - 1K<n<10K --- # Code de la Légion d'honneur, de la Médaille militaire et de l'ordre national du Mérite, non-instruct (2024-04-15) This project focuses on fine-tuning pre-trained language models to create efficient and accurate models for legal practice. Fine-tuning is the process of adapting a pre-trained model to perform specific tasks or cater to particular domains. It involves adjusting the model's parameters through a further round of training on task-specific or domain-specific data. While conventional fine-tuning strategies involve supervised learning with labeled data, instruction-based fine-tuning introduces a more structured and interpretable approach. Instruction-based fine-tuning leverages the power of human-provided instructions to guide the model's behavior. These instructions can be in the form of text prompts, prompts with explicit task descriptions, or a combination of both. This approach allows for a more controlled and context-aware interaction with the LLM, making it adaptable to a multitude of specialized tasks. Instruction-based fine-tuning significantly enhances the performance of LLMs in the following ways: - Task-Specific Adaptation: LLMs, when fine-tuned with specific instructions, exhibit remarkable adaptability to diverse tasks. They can switch seamlessly between translation, summarization, and question-answering, guided by the provided instructions. - Reduced Ambiguity: Traditional LLMs might generate ambiguous or contextually inappropriate responses. Instruction-based fine-tuning allows for a clearer and more context-aware generation, reducing the likelihood of nonsensical outputs. - Efficient Knowledge Transfer: Instructions can encapsulate domain-specific knowledge, enabling LLMs to benefit from expert guidance. This knowledge transfer is particularly valuable in fields like tax practice, law, medicine, and more. - Interpretability: Instruction-based fine-tuning also makes LLM behavior more interpretable. Since the instructions are human-readable, it becomes easier to understand and control model outputs. - Adaptive Behavior: LLMs, post instruction-based fine-tuning, exhibit adaptive behavior that is responsive to both explicit task descriptions and implicit cues within the provided text. ## Concurrent reading of the LegalKit To use all the legal data published on LegalKit, you can use this code snippet: ```python # -*- coding: utf-8 -*- import concurrent.futures import os import datasets from tqdm.notebook import tqdm def dataset_loader( name:str, streaming:bool=True ) -> datasets.Dataset: """ Helper function to load a single dataset in parallel. Parameters ---------- name : str Name of the dataset to be loaded. streaming : bool, optional Determines if datasets are streamed. Default is True. Returns ------- dataset : datasets.Dataset Loaded dataset object. Raises ------ Exception If an error occurs during dataset loading. """ try: return datasets.load_dataset( name, split="train", streaming=streaming ) except Exception as exc: logging.error(f"Error loading dataset {name}: {exc}") return None def load_datasets( req:list, streaming:bool=True ) -> list: """ Downloads datasets specified in a list and creates a list of loaded datasets. Parameters ---------- req : list A list containing the names of datasets to be downloaded. streaming : bool, optional Determines if datasets are streamed. Default is True. Returns ------- datasets_list : list A list containing loaded datasets as per the requested names provided in 'req'. Raises ------ Exception If an error occurs during dataset loading or processing. Examples -------- >>> datasets = load_datasets(["dataset1", "dataset2"], streaming=False) """ datasets_list = [] with concurrent.futures.ThreadPoolExecutor() as executor: future_to_dataset = {executor.submit(dataset_loader, name): name for name in req} for future in tqdm(concurrent.futures.as_completed(future_to_dataset), total=len(req)): name = future_to_dataset[future] try: dataset = future.result() if dataset: datasets_list.append(dataset) except Exception as exc: logging.error(f"Error processing dataset {name}: {exc}") return datasets_list req = [ "louisbrulenaudet/code-artisanat", "louisbrulenaudet/code-action-sociale-familles", # ... ] datasets_list = load_datasets( req=req, streaming=True ) dataset = datasets.concatenate_datasets( datasets_list ) ``` ## Dataset generation This JSON file is a list of dictionaries, each dictionary contains the following fields: - `instruction`: `string`, presenting the instruction linked to the element. - `input`: `string`, signifying the input details for the element. - `output`: `string`, indicating the output information for the element. - `start`: `string`, the date of entry into force of the article. - `expiration`: `string`, the date of expiration of the article. - `num`: `string`, the id of the article. We used the following list of instructions for generating the dataset: ```python instructions = [ "Compose l'intégralité de l'article sous forme écrite.", "Écris la totalité du contenu de l'article.", "Formule la totalité du texte présent dans l'article.", "Produis l'intégralité de l'article en écriture.", "Développe l'article dans son ensemble par écrit.", "Génère l'ensemble du texte contenu dans l'article.", "Formule le contenu intégral de l'article en entier.", "Rédige la totalité du texte de l'article en entier.", "Compose l'intégralité du contenu textuel de l'article.", "Rédige l'ensemble du texte qui constitue l'article.", "Formule l'article entier dans son contenu écrit.", "Composez l'intégralité de l'article sous forme écrite.", "Écrivez la totalité du contenu de l'article.", "Formulez la totalité du texte présent dans l'article.", "Développez l'article dans son ensemble par écrit.", "Générez l'ensemble du texte contenu dans l'article.", "Formulez le contenu intégral de l'article en entier.", "Rédigez la totalité du texte de l'article en entier.", "Composez l'intégralité du contenu textuel de l'article.", "Écrivez l'article dans son intégralité en termes de texte.", "Rédigez l'ensemble du texte qui constitue l'article.", "Formulez l'article entier dans son contenu écrit.", "Composer l'intégralité de l'article sous forme écrite.", "Écrire la totalité du contenu de l'article.", "Formuler la totalité du texte présent dans l'article.", "Produire l'intégralité de l'article en écriture.", "Développer l'article dans son ensemble par écrit.", "Générer l'ensemble du texte contenu dans l'article.", "Formuler le contenu intégral de l'article en entier.", "Rédiger la totalité du texte de l'article en entier.", "Composer l'intégralité du contenu textuel de l'article.", "Rédiger l'ensemble du texte qui constitue l'article.", "Formuler l'article entier dans son contenu écrit.", "Quelles sont les dispositions de l'article ?", "Quelles dispositions sont incluses dans l'article ?", "Quelles sont les dispositions énoncées dans l'article ?", "Quel est le texte intégral de l'article ?", "Quelle est la lettre de l'article ?" ] ``` ## Feedback If you have any feedback, please reach out at [louisbrulenaudet@icloud.com](mailto:louisbrulenaudet@icloud.com).
khalidalt/model-written-evals
--- annotations_creators: - machine-generated language: - en language_creators: - machine-generated license: - cc-by-4.0 multilinguality: - monolingual pretty_name: Evaluations from "Discovering Language Model Behaviors with Model-Written Evaluations" size_categories: - 100K<n<1M source_datasets: - original tags: - gender bias - social bias - AI safety - personality - politics task_categories: - multiple-choice - zero-shot-classification - question-answering task_ids: - multiple-choice-qa - multiple-choice-coreference-resolution --- # Model-Written Evaluation Datasets This repository includes datasets written by language models, used in the paper "Discovering Language Model Behaviors with Model-Written Evaluations." The evaluations in this dataset were designed for dialogue agents, such as models fine-tuned to respond to user utterances or pretrained language models prompted to simulate a dialogue agent's behavior. However, the data can be adapted to test various other types of models as well. The dataset consis of each of the following: 1. persona: Datasets designed to evaluate models on different aspects of their behavior, such as their political and religious views, personality traits, moral beliefs, and willingness to pursue potentially risky objectives (e.g., self-preservation or power-seeking). 2. sycophancy: Datasets created to assess models based on their tendency to echo a user's perspective when presented with various questions in fields like philosophy, NLP research, and politics. 3. winogenerated: An extended version of the Winogender Dataset (Rudinger et al., 2018) generated by models. The dataset includes occupation titles generated specifically for this dataset, alongside occupation gender statistics from the Bureau of Labor Statistics. 4. advanced-ai-risk: Datasets evaluating models on behaviors associated with potential catastrophic risks posed by advanced AI systems. These datasets were generated in a few-shot manner. Please see the cited paper for additional details on the datasets. **Disclaimer**: As discussed in the paper, some data contains content that includes social biases and stereotypes. The data may also contain other forms of harmful or offensive content. The views expressed in the data do not reflect the views of Anthropic or any of its employees. ## Bibtex Citation If you would like to cite this work or data, you may use the following bibtex citation: ``` @misc{perez2022discovering, doi = {10.48550/ARXIV.2212.09251}, url = {https://arxiv.org/abs/2212.09251}, author = {Perez, Ethan and Ringer, Sam and Lukošiūtė, Kamilė and Nguyen, Karina and Chen, Edwin and Heiner, Scott and Pettit, Craig and Olsson, Catherine and Kundu, Sandipan and Kadavath, Saurav and Jones, Andy and Chen, Anna and Mann, Ben and Israel, Brian and Seethor, Bryan and McKinnon, Cameron and Olah, Christopher and Yan, Da and Amodei, Daniela and Amodei, Dario and Drain, Dawn and Li, Dustin and Tran-Johnson, Eli and Khundadze, Guro and Kernion, Jackson and Landis, James and Kerr, Jamie and Mueller, Jared and Hyun, Jeeyoon and Landau, Joshua and Ndousse, Kamal and Goldberg, Landon and Lovitt, Liane and Lucas, Martin and Sellitto, Michael and Zhang, Miranda and Kingsland, Neerav and Elhage, Nelson and Joseph, Nicholas and Mercado, Noemí and DasSarma, Nova and Rausch, Oliver and Larson, Robin and McCandlish, Sam and Johnston, Scott and Kravec, Shauna and {El Showk}, Sheer and Lanham, Tamera and Telleen-Lawton, Timothy and Brown, Tom and Henighan, Tom and Hume, Tristan and Bai, Yuntao and Hatfield-Dodds, Zac and Clark, Jack and Bowman, Samuel R. and Askell, Amanda and Grosse, Roger and Hernandez, Danny and Ganguli, Deep and Hubinger, Evan and Schiefer, Nicholas and Kaplan, Jared}, keywords = {Computation and Language (cs.CL), Artificial Intelligence (cs.AI), Machine Learning (cs.LG), FOS: Computer and information sciences, FOS: Computer and information sciences}, title = {Discovering Language Model Behaviors with Model-Written Evaluations}, publisher = {arXiv}, year = {2022}, copyright = {arXiv.org perpetual, non-exclusive license} } ```
irds/nfcorpus_test_video
--- pretty_name: '`nfcorpus/test/video`' viewer: false source_datasets: ['irds/nfcorpus'] task_categories: - text-retrieval --- # Dataset Card for `nfcorpus/test/video` The `nfcorpus/test/video` dataset, provided by the [ir-datasets](https://ir-datasets.com/) package. For more information about the dataset, see the [documentation](https://ir-datasets.com/nfcorpus#nfcorpus/test/video). # Data This dataset provides: - `queries` (i.e., topics); count=102 - `qrels`: (relevance assessments); count=3,108 - For `docs`, use [`irds/nfcorpus`](https://huggingface.co/datasets/irds/nfcorpus) ## Usage ```python from datasets import load_dataset queries = load_dataset('irds/nfcorpus_test_video', 'queries') for record in queries: record # {'query_id': ..., 'title': ..., 'desc': ...} qrels = load_dataset('irds/nfcorpus_test_video', 'qrels') for record in qrels: record # {'query_id': ..., 'doc_id': ..., 'relevance': ...} ``` Note that calling `load_dataset` will download the dataset (or provide access instructions when it's not public) and make a copy of the data in 🤗 Dataset format. ## Citation Information ``` @inproceedings{Boteva2016Nfcorpus, title="A Full-Text Learning to Rank Dataset for Medical Information Retrieval", author = "Vera Boteva and Demian Gholipour and Artem Sokolov and Stefan Riezler", booktitle = "Proceedings of the European Conference on Information Retrieval ({ECIR})", location = "Padova, Italy", publisher = "Springer", year = 2016 } ```
tyzhu/lmind_nq_train1000_eval500_v1_docidx
--- configs: - config_name: default data_files: - split: train_qa path: data/train_qa-* - split: train_recite_qa path: data/train_recite_qa-* - split: eval_qa path: data/eval_qa-* - split: eval_recite_qa path: data/eval_recite_qa-* - split: all_docs path: data/all_docs-* - split: all_docs_eval path: data/all_docs_eval-* - split: train path: data/train-* - split: validation path: data/validation-* dataset_info: features: - name: answers struct: - name: answer_start sequence: 'null' - name: text sequence: string - name: inputs dtype: string - name: targets dtype: string splits: - name: train_qa num_bytes: 115608 num_examples: 1000 - name: train_recite_qa num_bytes: 755758 num_examples: 1000 - name: eval_qa num_bytes: 58285 num_examples: 500 - name: eval_recite_qa num_bytes: 377880 num_examples: 500 - name: all_docs num_bytes: 950316 num_examples: 1462 - name: all_docs_eval num_bytes: 950216 num_examples: 1462 - name: train num_bytes: 950316 num_examples: 1462 - name: validation num_bytes: 950216 num_examples: 1462 download_size: 3228354 dataset_size: 5108595 --- # Dataset Card for "lmind_nq_train1000_eval500_v1_docidx" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
parsak/lima-tr-alpacastyle
--- dataset_info: features: - name: source dtype: string - name: instruction dtype: string - name: output dtype: string - name: prompt dtype: string splits: - name: train num_bytes: 6336911 num_examples: 1030 - name: test num_bytes: 131570 num_examples: 300 download_size: 3510588 dataset_size: 6468481 configs: - config_name: default data_files: - split: train path: data/train-* - split: test path: data/test-* license: mit task_categories: - question-answering - text-generation language: - tr pretty_name: Lima - Turkish - Alpaca Style size_categories: - 1K<n<10K ---
dumyy/token-classification-brand
--- dataset_info: features: - name: tokens sequence: string - name: ner_tags sequence: class_label: names: '0': '0' '1': B-BRAND '2': I-BRAND - name: spans dtype: string splits: - name: train num_bytes: 462 num_examples: 3 - name: val num_bytes: 462 num_examples: 3 download_size: 4670 dataset_size: 924 configs: - config_name: default data_files: - split: train path: data/train-* - split: val path: data/val-* --- # Dataset Card for "token-classification-brand" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
HPGomes/MJTalking
--- license: openrail ---
Yukang/dpo_debug
--- license: apache-2.0 ---
minh21/COVID-QA-Chunk-64-testset-biencoder-data-90_10
--- dataset_info: features: - name: question dtype: string - name: answer dtype: string - name: context_chunks sequence: string - name: document_id dtype: int64 - name: id dtype: int64 - name: context dtype: string splits: - name: train num_bytes: 13595044 num_examples: 203 download_size: 459357 dataset_size: 13595044 configs: - config_name: default data_files: - split: train path: data/train-* --- # Dataset Card for "COVID-QA-Chunk-64-testset-biencoder-data-90_10" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
cr7Por/my_controlnet_feat4
--- dataset_info: features: - name: image dtype: image - name: image_caption dtype: string - name: feat_beit dtype: image - name: feat_resnet dtype: image - name: feat_convnext dtype: image - name: feat_vitmae dtype: image - name: validline dtype: string splits: - name: train num_bytes: 171271858.0 num_examples: 435 download_size: 170826843 dataset_size: 171271858.0 --- # Dataset Card for "my_controlnet_feat4" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
hnchen/Session-search
--- license: afl-3.0 ---
iohadrubin/gorilla_openfunctions_yaml_train
--- dataset_info: features: - name: text dtype: string - name: functions dtype: string - name: output dtype: string splits: - name: train num_bytes: 26568497 num_examples: 12125 download_size: 10636054 dataset_size: 26568497 configs: - config_name: default data_files: - split: train path: data/train-* ---