Other
Transformers
Safetensors
PyTorch
kvzap
nvidia

Add pipeline tag and improve model card

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +4 -3
README.md CHANGED
@@ -1,8 +1,9 @@
1
  ---
2
- license: apache-2.0
3
  datasets:
4
  - nvidia/Nemotron-Pretraining-Dataset-sample
5
  library_name: transformers
 
 
6
  tags:
7
  - nvidia
8
  - pytorch
@@ -13,7 +14,7 @@ track_downloads: true
13
 
14
 
15
  [![License](https://img.shields.io/badge/License-Apache%202.0-green.svg)](https://www.apache.org/licenses/LICENSE-2.0)
16
- [![GitHub](https://img.shields.io/badge/GitHub-kvpress-blue?logo=github)](https://github.com/NVIDIA/kvpress/kvzap)
17
  [![KVzap collection](https://img.shields.io/badge/🤗%20Hugging%20Face-Collection-orange)](https://huggingface.co/collections/nvidia/kvzap)
18
  [![arXiv](https://img.shields.io/badge/arXiv-2601.07891-b31b1b.svg)](https://arxiv.org/abs/2601.07891)
19
 
@@ -21,4 +22,4 @@ track_downloads: true
21
  [KVzap](https://arxiv.org/abs/2601.07891) is a KV cache pruning method aiming to accelerate LLM inference in both prefilling and decoding. It applies a lightweight model to the hidden states to predict importance scores for every KV pair and prunes the ones with a score below a given threshold.
22
 
23
 
24
- KVzap is trained as a fast approximation of [KVzip+](https://arxiv.org/abs/2505.23416), using 1.2M samples from [Nemotron-Pretraining-Dataset-sample](https://huggingface.co/datasets/nvidia/Nemotron-Pretraining-Dataset-sample). Training code is available in the kvpress repository ([source](https://github.com/NVIDIA/kvpress/blob/main/kvzap)).
 
1
  ---
 
2
  datasets:
3
  - nvidia/Nemotron-Pretraining-Dataset-sample
4
  library_name: transformers
5
+ license: apache-2.0
6
+ pipeline_tag: other
7
  tags:
8
  - nvidia
9
  - pytorch
 
14
 
15
 
16
  [![License](https://img.shields.io/badge/License-Apache%202.0-green.svg)](https://www.apache.org/licenses/LICENSE-2.0)
17
+ [![GitHub](https://img.shields.io/badge/GitHub-kvpress-blue?logo=github)](https://github.com/NVIDIA/kvpress/tree/main/kvzap)
18
  [![KVzap collection](https://img.shields.io/badge/🤗%20Hugging%20Face-Collection-orange)](https://huggingface.co/collections/nvidia/kvzap)
19
  [![arXiv](https://img.shields.io/badge/arXiv-2601.07891-b31b1b.svg)](https://arxiv.org/abs/2601.07891)
20
 
 
22
  [KVzap](https://arxiv.org/abs/2601.07891) is a KV cache pruning method aiming to accelerate LLM inference in both prefilling and decoding. It applies a lightweight model to the hidden states to predict importance scores for every KV pair and prunes the ones with a score below a given threshold.
23
 
24
 
25
+ KVzap is trained as a fast approximation of [KVzip+](https://arxiv.org/abs/2505.23416), using 1.2M samples from [Nemotron-Pretraining-Dataset-sample](https://huggingface.co/datasets/nvidia/Nemotron-Pretraining-Dataset-sample). Training code is available in the kvpress repository ([source](https://github.com/NVIDIA/kvpress/blob/main/kvzap)).