|
|
--- |
|
|
|
|
|
|
|
|
language: |
|
|
- en |
|
|
license: apache-2.0 |
|
|
tags: |
|
|
- RAG |
|
|
- model card generation |
|
|
- responsible AI |
|
|
|
|
|
configs: |
|
|
- config_name: model_card |
|
|
data_files: |
|
|
- split: test |
|
|
path: model_card_test.csv |
|
|
- split: whole |
|
|
path: model_card_whole.csv |
|
|
- config_name: data_card |
|
|
data_files: |
|
|
- split: whole |
|
|
path: data_card_whole.csv |
|
|
|
|
|
--- |
|
|
|
|
|
|
|
|
# Automatic Generation of Model and Data Cards: A Step Towards Responsible AI |
|
|
|
|
|
The work has been accepted to NAACL 2024 Oral. |
|
|
|
|
|
**Abstract**: In an era of model and data proliferation in machine learning/AI especially marked by the rapid advancement of open-sourced technologies, there arises a critical need for standardized consistent documentation. Our work addresses the information incompleteness in current human-written model and data cards. We propose an automated generation approach using Large Language Models (LLMs). Our key contributions include the establishment of CardBench, a comprehensive dataset aggregated from over 4.8k model cards and 1.4k data cards, coupled with the development of the CardGen pipeline comprising a two-step retrieval process. Our approach exhibits enhanced completeness, objectivity, and faithfulness in generated model and data cards, a significant step in responsible AI documentation practices ensuring better accountability and traceability. |
|
|
|
|
|
**Paper Arxiv**: https://arxiv.org/abs/2405.06258 |
|
|
|
|
|
**ACL Anthology**: https://aclanthology.org/2024.naacl-long.110/ |
|
|
|
|
|
**Repository and Code**: https://github.com/jiarui-liu/AutomatedModelCardGeneration |
|
|
|
|
|
**Dataset descriptions**: |
|
|
- `model_card_test.csv`: Contains the test set used for model card generation. We collected the model cards and data cards from the HuggingFace page as of October 1, 2023. |
|
|
- `model_card_whole.csv`: Represents the complete dataset excluding the test set. |
|
|
- `data_card_whole.csv`: Represents the complete dataset for data card generation. |
|
|
- **Additional files**: Other included files may be useful for reproducing our work. |
|
|
|
|
|
Disclaimer: Please forgive me for not creating this data card as described in our paper. We promise to give it some extra love and polish when we have more time! 🫠 |
|
|
|
|
|
**Citation**: If you find our work useful, please cite as follows :) |
|
|
|
|
|
``` |
|
|
@inproceedings{liu-etal-2024-automatic, |
|
|
title = "Automatic Generation of Model and Data Cards: A Step Towards Responsible {AI}", |
|
|
author = "Liu, Jiarui and |
|
|
Li, Wenkai and |
|
|
Jin, Zhijing and |
|
|
Diab, Mona", |
|
|
editor = "Duh, Kevin and |
|
|
Gomez, Helena and |
|
|
Bethard, Steven", |
|
|
booktitle = "Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)", |
|
|
month = jun, |
|
|
year = "2024", |
|
|
address = "Mexico City, Mexico", |
|
|
publisher = "Association for Computational Linguistics", |
|
|
url = "https://aclanthology.org/2024.naacl-long.110", |
|
|
doi = "10.18653/v1/2024.naacl-long.110", |
|
|
pages = "1975--1997", |
|
|
abstract = "In an era of model and data proliferation in machine learning/AI especially marked by the rapid advancement of open-sourced technologies, there arises a critical need for standardized consistent documentation. Our work addresses the information incompleteness in current human-written model and data cards. We propose an automated generation approach using Large Language Models (LLMs). Our key contributions include the establishment of CardBench, a comprehensive dataset aggregated from over 4.8k model cards and 1.4k data cards, coupled with the development of the CardGen pipeline comprising a two-step retrieval process. Our approach exhibits enhanced completeness, objectivity, and faithfulness in generated model and data cards, a significant step in responsible AI documentation practices ensuring better accountability and traceability.", |
|
|
} |
|
|
``` |
|
|
|