nielsr HF Staff commited on
Commit
6f5fcbd
·
verified ·
1 Parent(s): a5c804b

Improve model card: paper link, project link, remove misleading library_name for GGUF

Browse files

This PR improves the model card for `openbmb/BitCPM4-1B-GGUF` by:

- Updating the paper link to point to the official Hugging Face paper page: https://huggingface.co/papers/2506.07900.
- Adding a link to the project's Hugging Face collection (project page) for better overview: https://huggingface.co/collections/openbmb/minicpm4-6841ab29d180257e940baa9b.
- Removing the misleading `library_name: transformers` metadata, as this specific GGUF model is intended for use with `llama.cpp` (as demonstrated in the usage example) and not the `transformers` library for direct inference.
- Adding a descriptive main title to the model card for clarity.
- Updating the bibtex entry to include the URL to the paper.

Files changed (1) hide show
  1. README.md +19 -13
README.md CHANGED
@@ -1,25 +1,31 @@
1
  ---
2
- license: apache-2.0
3
  language:
4
  - zh
5
  - en
 
6
  pipeline_tag: text-generation
7
- library_name: transformers
8
  ---
 
9
  <div align="center">
10
- <img src="https://github.com/OpenBMB/MiniCPM/blob/main/assets/minicpm_logo.png?raw=true" width="500em" ></img>
11
  </div>
12
 
 
 
 
 
13
  <p align="center">
14
  <a href="https://github.com/OpenBMB/MiniCPM/" target="_blank">GitHub Repo</a> |
15
- <a href="https://github.com/OpenBMB/MiniCPM/tree/main/report/MiniCPM_4_Technical_Report.pdf" target="_blank">Technical Report</a>
 
 
16
  </p>
17
  <p align="center">
18
  👋 Join us on <a href="https://discord.gg/3cGQn9b3YM" target="_blank">Discord</a> and <a href="https://github.com/OpenBMB/MiniCPM/blob/main/assets/wechat.jpg" target="_blank">WeChat</a>
19
  </p>
20
 
21
  ## What's New
22
- - [2025.06.06] **MiniCPM4** series are released! This model achieves ultimate efficiency improvements while maintaining optimal performance at the same scale! It can achieve over 5x generation acceleration on typical end-side chips! You can find technical report [here](https://github.com/OpenBMB/MiniCPM/tree/main/report/MiniCPM_4_Technical_Report.pdf).🔥🔥🔥
23
 
24
  ## MiniCPM4 Series
25
  MiniCPM4 series are highly efficient large language models (LLMs) designed explicitly for end-side devices, which achieves this efficiency through systematic innovation in four key dimensions: model architecture, training data, training algorithms, and inference systems.
@@ -55,21 +61,21 @@ BitCPM4's performance is comparable with other full-precision models in same mod
55
  ![Benchmark of BitCPM](https://raw.githubusercontent.com/OpenBMB/MiniCPM/refs/heads/main/assets/minicpm4/bitcpm4-benchmark.png)
56
 
57
  ## Statement
58
- - As a language model, MiniCPM generates content by learning from a vast amount of text.
59
- - However, it does not possess the ability to comprehend or express personal opinions or value judgments.
60
- - Any content generated by MiniCPM does not represent the viewpoints or positions of the model developers.
61
  - Therefore, when using content generated by MiniCPM, users should take full responsibility for evaluating and verifying it on their own.
62
 
63
  ## LICENSE
64
- - This repository and MiniCPM models are released under the [Apache-2.0](https://github.com/OpenBMB/MiniCPM/blob/main/LICENSE) License.
65
 
66
  ## Citation
67
- - Please cite our [paper](https://github.com/OpenBMB/MiniCPM/tree/main/report/MiniCPM_4_Technical_Report.pdf) if you find our work valuable.
68
 
69
  ```bibtex
70
  @article{minicpm4,
71
  title={{MiniCPM4}: Ultra-Efficient LLMs on End Devices},
72
  author={MiniCPM Team},
73
- year={2025}
74
- }
75
- ```
 
1
  ---
 
2
  language:
3
  - zh
4
  - en
5
+ license: apache-2.0
6
  pipeline_tag: text-generation
 
7
  ---
8
+
9
  <div align="center">
10
+ <img src="https://github.com/OpenBMB/MiniCPM/blob/main/assets/minicpm_logo.png?raw=true" width="500em" ></img>
11
  </div>
12
 
13
+ # MiniCPM4: Ultra-Efficient LLMs on End Devices - BitCPM4-1B-GGUF
14
+
15
+ This repository contains the `BitCPM4-1B-GGUF` model, part of the MiniCPM4 series, as presented in [MiniCPM4: Ultra-Efficient LLMs on End Devices](https://huggingface.co/papers/2506.07900).
16
+
17
  <p align="center">
18
  <a href="https://github.com/OpenBMB/MiniCPM/" target="_blank">GitHub Repo</a> |
19
+ <a href="https://huggingface.co/papers/2506.07900" target="_blank">Paper</a> |
20
+ <a href="https://github.com/OpenBMB/MiniCPM/tree/main/report/MiniCPM_4_Technical_Report.pdf" target="_blank">Technical Report</a> |
21
+ <a href="https://huggingface.co/collections/openbmb/minicpm4-6841ab29d180257e940baa9b" target="_blank">Project Page</a>
22
  </p>
23
  <p align="center">
24
  👋 Join us on <a href="https://discord.gg/3cGQn9b3YM" target="_blank">Discord</a> and <a href="https://github.com/OpenBMB/MiniCPM/blob/main/assets/wechat.jpg" target="_blank">WeChat</a>
25
  </p>
26
 
27
  ## What's New
28
+ - [2025.06.06] **MiniCPM4** series are released! This model achieves ultimate efficiency improvements while maintaining optimal performance at the same scale! It can achieve over 5x generation acceleration on typical end-side chips! You can find technical report [here](https://huggingface.co/papers/2506.07900).🔥🔥🔥
29
 
30
  ## MiniCPM4 Series
31
  MiniCPM4 series are highly efficient large language models (LLMs) designed explicitly for end-side devices, which achieves this efficiency through systematic innovation in four key dimensions: model architecture, training data, training algorithms, and inference systems.
 
61
  ![Benchmark of BitCPM](https://raw.githubusercontent.com/OpenBMB/MiniCPM/refs/heads/main/assets/minicpm4/bitcpm4-benchmark.png)
62
 
63
  ## Statement
64
+ - As a language model, MiniCPM generates content by learning from a vast amount of text.
65
+ - However, it does not possess the ability to comprehend or express personal opinions or value judgments.
66
+ - Any content generated by MiniCPM does not represent the viewpoints or positions of the model developers.
67
  - Therefore, when using content generated by MiniCPM, users should take full responsibility for evaluating and verifying it on their own.
68
 
69
  ## LICENSE
70
+ - This repository and MiniCPM models are released under the [Apache-2.0](https://github.com/OpenBMB/MiniCPM/blob/main/LICENSE) License.
71
 
72
  ## Citation
73
+ - Please cite our [paper](https://huggingface.co/papers/2506.07900) if you find our work valuable.
74
 
75
  ```bibtex
76
  @article{minicpm4,
77
  title={{MiniCPM4}: Ultra-Efficient LLMs on End Devices},
78
  author={MiniCPM Team},
79
+ year={2025},
80
+ url={https://huggingface.co/papers/2506.07900},
81
+ }