Jerry999 commited on
Commit
6dc5544
·
verified ·
1 Parent(s): ba37176

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +69 -0
README.md ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ # Example metadata to be added to a dataset card.
3
+ # Full dataset card template at https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/datasetcard_template.md
4
+ language:
5
+ - en
6
+ license: apache-2.0 # Example: apache-2.0 or any license from https://hf.co/docs/hub/repositories-licenses
7
+ tags:
8
+ - RAG
9
+ - model card generation
10
+ - responsible AI
11
+
12
+ configs: # Optional. This can be used to pass additional parameters to the dataset loader, such as `data_files`, `data_dir`, and any builder-specific parameters
13
+ - config_name: model_card # Name of the dataset subset, if applicable. Example: default
14
+ data_files:
15
+ - split: test
16
+ path: model_card_test.csv
17
+ - split: whole
18
+ path: model_card_whole.csv
19
+ - config_name: data_card
20
+ data_files:
21
+ - split: whole
22
+ path: data_card_whole.csv
23
+
24
+ ---
25
+
26
+
27
+ # Automatic Generation of Model and Data Cards: A Step Towards Responsible AI
28
+
29
+ The work has been accepted to NAACL 2024 Oral.
30
+
31
+ **Abstract**: In an era of model and data proliferation in machine learning/AI especially marked by the rapid advancement of open-sourced technologies, there arises a critical need for standardized consistent documentation. Our work addresses the information incompleteness in current human-written model and data cards. We propose an automated generation approach using Large Language Models (LLMs). Our key contributions include the establishment of CardBench, a comprehensive dataset aggregated from over 4.8k model cards and 1.4k data cards, coupled with the development of the CardGen pipeline comprising a two-step retrieval process. Our approach exhibits enhanced completeness, objectivity, and faithfulness in generated model and data cards, a significant step in responsible AI documentation practices ensuring better accountability and traceability.
32
+
33
+ **Paper Arxiv**: https://arxiv.org/abs/2405.06258
34
+
35
+ **ACL Anthology**: https://aclanthology.org/2024.naacl-long.110/
36
+
37
+ **Repository and Code**: https://github.com/jiarui-liu/AutomatedModelCardGeneration
38
+
39
+ **Dataset descriptions**:
40
+ - `model_card_test.csv`: Contains the test set used for model card generation.
41
+ - `model_card_whole.csv`: Represents the complete dataset excluding the test set.
42
+ - `data_card_whole.csv`: Represents the complete dataset for data card generation.
43
+ - **Additional files**: Other included files may be useful for reproducing our work.
44
+
45
+ Disclaimer: Please forgive me for not creating this data card exactly as described in our paper. We promise to give it some extra love and polish when we have more time! 🫠
46
+
47
+ **Citation**: If you find our work useful, please cite as follows :)
48
+
49
+ ```
50
+ @inproceedings{liu-etal-2024-automatic,
51
+ title = "Automatic Generation of Model and Data Cards: A Step Towards Responsible {AI}",
52
+ author = "Liu, Jiarui and
53
+ Li, Wenkai and
54
+ Jin, Zhijing and
55
+ Diab, Mona",
56
+ editor = "Duh, Kevin and
57
+ Gomez, Helena and
58
+ Bethard, Steven",
59
+ booktitle = "Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)",
60
+ month = jun,
61
+ year = "2024",
62
+ address = "Mexico City, Mexico",
63
+ publisher = "Association for Computational Linguistics",
64
+ url = "https://aclanthology.org/2024.naacl-long.110",
65
+ doi = "10.18653/v1/2024.naacl-long.110",
66
+ pages = "1975--1997",
67
+ abstract = "In an era of model and data proliferation in machine learning/AI especially marked by the rapid advancement of open-sourced technologies, there arises a critical need for standardized consistent documentation. Our work addresses the information incompleteness in current human-written model and data cards. We propose an automated generation approach using Large Language Models (LLMs). Our key contributions include the establishment of CardBench, a comprehensive dataset aggregated from over 4.8k model cards and 1.4k data cards, coupled with the development of the CardGen pipeline comprising a two-step retrieval process. Our approach exhibits enhanced completeness, objectivity, and faithfulness in generated model and data cards, a significant step in responsible AI documentation practices ensuring better accountability and traceability.",
68
+ }
69
+ ```