Datasets:
Formats:
csv
Languages:
English
Size:
< 1K
ArXiv:
Tags:
benchmark
llm-evaluation
large-language-models
large-language-model
large-multimodal-models
llm-training
DOI:
License:
| license: mit | |
| task_categories: | |
| - text2text-generation | |
| - text-generation | |
| - text-retrieval | |
| - question-answering | |
| language: | |
| - en | |
| tags: | |
| - benchmark | |
| - llm-evaluation | |
| - large-language-models | |
| - large-language-model | |
| - large-multimodal-models | |
| - llm-training | |
| - foundation-models | |
| - machine-learning | |
| - deep-learning | |
| configs: | |
| - config_name: all_responses | |
| data_files: "AllResponses.csv" | |
| - config_name: clean_responses | |
| data_files: "CleanResponses.csv" | |
| - config_name: additional_data | |
| data_files: "KeyQuestions.csv" | |
| --- | |
| MSEval Dataset: | |
| --- | |
| A benchmark designed to facilitate evaluation and modify the behavior of a foundation model through different existing techniques in the context of material selection for conceptual design. | |
| The data is collected by conducting a survey of experts in the field of material selection. The same questions mentioned in keyquestions.csv are asked to experts. | |
| This can be used to evaluate a Language model performance and its spread compared to a human evaluation. | |
| To get into a more detailed explanation - use this link [https://arxiv.org/abs/2407.09719v1] | |
| --- | |
| # Overview | |
| We introduce MSEval, a benchmark derived from survey results of experts in the field of material selection. | |
| The MixEval consists of two files: `CleanResponses` and `AllResponses`. Below presents the dataset file tree: | |
| ``` | |
| MSEval | |
| │ | |
| ├── AllResponses.csv | |
| └── CleanResponses.csv | |
| └── KeyQuestions.csv | |
| ``` | |
| # Dataset Usage | |
| An example use of the dataset using the datasets library is shown in https://github.com/cmudrc/MSEval | |
| To use this dataset using pandas: | |
| ``` | |
| import pandas as pd | |
| df = pd.read_csv("hf://datasets/cmudrc/Material_Selection_Eval/AllResponses.csv") | |
| ``` | |
| Replace AllResponses with CleanResponses and KeyQuestions in the pathname if required. | |
| # Citation | |
| If you found the dataset useful, please cite: | |
| ```bibtex | |
| @misc{jain2024msevaldatasetmaterialselection, | |
| title={MSEval: A Dataset for Material Selection in Conceptual Design to Evaluate Algorithmic Models}, | |
| author={Yash Patawari Jain and Daniele Grandi and Allin Groom and Brandon Cramer and Christopher McComb}, | |
| year={2024}, | |
| eprint={2407.09719}, | |
| archivePrefix={arXiv}, | |
| primaryClass={cs.LG}, | |
| url={https://arxiv.org/abs/2407.09719}, | |
| } | |
| ``` | |
| license: mit | |
| --- | |