Improve model card: Add paper/project links and `code-generation`, `datasets`, `metrics` tags
#1
by
nielsr HF Staff - opened
This PR enhances the model card for MiniCPM4-8B by:
- Adding direct links to the official Hugging Face paper page (MiniCPM4: Ultra-Efficient LLMs on End Devices) and the Hugging Face collection project page (
https://huggingface.co/collections/openbmb/minicpm4-6841ab29d180257e940baa9b) in the header section. Existing links pointing to the GitHub technical report PDF in "What's New" and "Citation" are updated to the HF paper link. The BibTeX entry is also updated with the arXiv journal. - Introducing the
code-generationtag to the metadata, as supported by the model's performance on code benchmarks (HumanEval+, MBPP+, LiveCodeBench) and the "MCP-Code-Executor" application mentioned in the GitHub README. - Adding the
datasetstag foropenbmb/Ultra-FineWeband themetricstag foraccuracy, as both are explicitly mentioned and demonstrated in the model's documentation. - Restructuring the "Usage" section to align with the original GitHub README's presentation of
InfLLM v2andrope_scalingdetails under the Hugging Face inference, rather than theCPM.cusection. This ensures consistency and clarity for users.
These improvements will make the model more discoverable, accurately categorized, and provide a clearer usage guide on the Hugging Face Hub.