Spaces:
Running
Running
Update README.md
Browse files
README.md
CHANGED
|
@@ -5,13 +5,14 @@ colorFrom: pink
|
|
| 5 |
colorTo: indigo
|
| 6 |
sdk: static
|
| 7 |
pinned: false
|
|
|
|
|
|
|
| 8 |
---
|
| 9 |
-
# Compressed
|
| 10 |
-
|
| 11 |
|
|
|
|
| 12 |
The models are prepared by [Visual Informatics Group @ University of Texas at Austin (VITA-group)](https://vita-group.github.io/) and
|
| 13 |
[Center for Applied Scientific Computing](https://computing.llnl.gov/casc) at [LLNL](https://www.llnl.gov/).
|
| 14 |
-
Credits to Ajay Jaiswal, Jinhao Duan, Zhenyu Zhang, Zhangheng Li, Lu Yin, Shiwei Liu and Junyuan Hong.
|
| 15 |
|
| 16 |
License: [MIT License](https://opensource.org/license/mit/)
|
| 17 |
|
|
@@ -96,6 +97,18 @@ tokenizer.decode(outputs[0])
|
|
| 96 |
|
| 97 |
If you are using models in this hub, please consider citing our papers.
|
| 98 |
```bibtex
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 99 |
@article{jaiswal2023emergence,
|
| 100 |
title={The Emergence of Essential Sparsity in Large Pre-trained Models: The Weights that Matter},
|
| 101 |
author={Jaiswal, Ajay and Liu, Shiwei and Chen, Tianlong and Wang, Zhangyang},
|
|
@@ -110,6 +123,9 @@ If you are using models in this hub, please consider citing our papers.
|
|
| 110 |
}
|
| 111 |
```
|
| 112 |
|
|
|
|
|
|
|
|
|
|
| 113 |
|
| 114 |
-
For any question, please contact [Junyuan Hong](mailto:jyhong@utexas.edu).
|
| 115 |
|
|
|
|
|
|
| 5 |
colorTo: indigo
|
| 6 |
sdk: static
|
| 7 |
pinned: false
|
| 8 |
+
license: mit
|
| 9 |
+
short_description: Compressed Large Language Models
|
| 10 |
---
|
| 11 |
+
# Compressed Large Language Models
|
|
|
|
| 12 |
|
| 13 |
+
This repo contains compressed LLMs used in the [Decoding Compressed Trust](https://decoding-comp-trust.github.io/) project.
|
| 14 |
The models are prepared by [Visual Informatics Group @ University of Texas at Austin (VITA-group)](https://vita-group.github.io/) and
|
| 15 |
[Center for Applied Scientific Computing](https://computing.llnl.gov/casc) at [LLNL](https://www.llnl.gov/).
|
|
|
|
| 16 |
|
| 17 |
License: [MIT License](https://opensource.org/license/mit/)
|
| 18 |
|
|
|
|
| 97 |
|
| 98 |
If you are using models in this hub, please consider citing our papers.
|
| 99 |
```bibtex
|
| 100 |
+
@article{hong2024comptrust,
|
| 101 |
+
title={Decoding Compressed Trust: Scrutinizing the Trustworthiness of Efficient LLMs Under Compression},
|
| 102 |
+
author={Hong, Junyuan and Duan, Jinhao and Zhang, Chenhui and Li, Zhangheng
|
| 103 |
+
and Xie, Chulin and Lieberman, Kelsey and Diffenderfer, James
|
| 104 |
+
and Bartoldson, Brian and Jaiswal, Ajay and Xu, Kaidi and Kailkhura, Bhavya
|
| 105 |
+
and Hendrycks, Dan and Song, Dawn and Wang, Zhangyang and Bo Li},
|
| 106 |
+
journal={arXiv},
|
| 107 |
+
year={2024}
|
| 108 |
+
}
|
| 109 |
+
```
|
| 110 |
+
Some of the models were used in previous publications.
|
| 111 |
+
```bibtex
|
| 112 |
@article{jaiswal2023emergence,
|
| 113 |
title={The Emergence of Essential Sparsity in Large Pre-trained Models: The Weights that Matter},
|
| 114 |
author={Jaiswal, Ajay and Liu, Shiwei and Chen, Tianlong and Wang, Zhangyang},
|
|
|
|
| 123 |
}
|
| 124 |
```
|
| 125 |
|
| 126 |
+
## Acknowlegement
|
| 127 |
+
|
| 128 |
+
Main credits to Ajay Jaiswal, Jinhao Duan, Zhangheng Li and Junyuan Hong. We also appreciate Zhenyu Zhang, Lu Yin, and Shiwei Liu in some preparations.
|
| 129 |
|
|
|
|
| 130 |
|
| 131 |
+
For any question, please contact [Junyuan Hong](mailto:jyhong@utexas.edu).
|