nielsr HF Staff commited on
Commit
8212f08
·
verified ·
1 Parent(s): 1779648

Improve model card: add pipeline tag, library name, update license, and paper reference

Browse files

This PR enhances the model card for the quantized DeepSeek-R1 model by the AutoRound algorithm.

It addresses the following:
- **Added `pipeline_tag`**: The `text-generation` pipeline tag is added to the metadata, improving discoverability on the Hugging Face Hub.
- **Added `library_name`**: The `transformers` library is added as `library_name` to the metadata, enabling the automated "how to use" widget on the model page, as supported by the provided sample code and project documentation.
- **Updated `license`**: The `license` metadata is updated to `apache-2.0`, aligning with the license of the original `deepseek-ai/DeepSeek-R1` model and the `intel/auto-round` project.
- **Updated Paper Reference**: The model card content and the "Cite" section are updated to correctly reference the latest paper, "[SignRoundV2: Closing the Performance Gap in Extremely Low-Bit Post-Training Quantization for LLMs](https://huggingface.co/papers/2512.04746)".
- **Prominent GitHub Link**: A direct link to the [AutoRound GitHub repository](https://github.com/intel/auto-round) is added near the beginning of the model card for easy access.

These changes significantly improve the model card's clarity, accuracy, and usability for the Hugging Face community.

Files changed (1) hide show
  1. README.md +18 -7
README.md CHANGED
@@ -1,16 +1,20 @@
1
  ---
2
- datasets:
3
- - NeelNanda/pile-10k
4
  base_model:
5
  - deepseek-ai/DeepSeek-R1
 
 
 
 
 
 
6
 
 
7
 
8
-
9
- ---
10
 
11
  ## Model Details
12
 
13
- This model is an int2 model with group_size 64 and symmetric quantization of [deepseek-ai/DeepSeek-R1](https://huggingface.co/deepseek-ai/DeepSeek-R1) generated by [intel/auto-round](https://github.com/intel/auto-round) algorithm. Some layers are fallback to 4/16 bits. Refer to Section "Generate the model" for more details of mixed bits setting.
14
 
15
  Please follow the license of the original model. This model could **NOT** run on other severing frameworks.
16
 
@@ -439,6 +443,13 @@ The license on this model does not constitute legal advice. We are not responsib
439
 
440
  ## Cite
441
 
442
- @article{cheng2023optimize, title={Optimize weight rounding via signed gradient descent for the quantization of llms}, author={Cheng, Wenhua and Zhang, Weiwei and Shen, Haihao and Cai, Yiyang and He, Xin and Lv, Kaokao and Liu, Yi}, journal={arXiv preprint arXiv:2309.05516}, year={2023} }
 
 
 
 
 
 
 
443
 
444
- [arxiv](https://arxiv.org/abs/2309.05516) [github](https://github.com/intel/auto-round)
 
1
  ---
 
 
2
  base_model:
3
  - deepseek-ai/DeepSeek-R1
4
+ datasets:
5
+ - NeelNanda/pile-10k
6
+ pipeline_tag: text-generation
7
+ library_name: transformers
8
+ license: apache-2.0
9
+ ---
10
 
11
+ This model is an int2 model with group_size 64 and symmetric quantization of [deepseek-ai/DeepSeek-R1](https://huggingface.co/deepseek-ai/DeepSeek-R1), generated by the **SignRoundV2** algorithm described in the paper [SignRoundV2: Closing the Performance Gap in Extremely Low-Bit Post-Training Quantization for LLMs](https://huggingface.co/papers/2512.04746).
12
 
13
+ For more details on the AutoRound project and its implementation, see the [GitHub repository](https://github.com/intel/auto-round).
 
14
 
15
  ## Model Details
16
 
17
+ Some layers are fallback to 4/16 bits. Refer to Section "Generate the model" for more details of mixed bits setting.
18
 
19
  Please follow the license of the original model. This model could **NOT** run on other severing frameworks.
20
 
 
443
 
444
  ## Cite
445
 
446
+ ```bibtex
447
+ @article{cheng2025signroundv2,
448
+ title={SignRoundV2: Closing the Performance Gap in Extremely Low-Bit Post-Training Quantization for LLMs},
449
+ author={Cheng, Wenhua and Zhang, Weiwei and Guo, Heng and Shen, Haihao},
450
+ journal={arXiv preprint arXiv:2512.04746},
451
+ year={2025}
452
+ }
453
+ ```
454
 
455
+ [arxiv](https://arxiv.org/abs/2512.04746)