CompactAI commited on
Commit
db71d73
·
verified ·
1 Parent(s): de9b828

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. README.md +58 -20
  2. comparison_graph.png +2 -2
  3. model.safetensors +1 -1
README.md CHANGED
@@ -4,44 +4,82 @@ tags:
4
  - pruned
5
  - python
6
  - optimized
 
 
7
  base_model: LGAI-EXAONE/EXAONE-4.0-1.2B
 
8
  ---
9
 
10
  # EXAONE-4.0-1.2B-python-medium-heavy
11
 
12
- This model is a **medium-heavy** pruned version of [LGAI-EXAONE/EXAONE-4.0-1.2B](https://huggingface.co/LGAI-EXAONE/EXAONE-4.0-1.2B), specialized for **PYTHON** tasks.
13
 
14
- ## Pruning Details
15
 
16
- - **Base Model**: LGAI-EXAONE/EXAONE-4.0-1.2B
17
- - **Specialization**: Python
18
- - **Prune Mode**: Medium-heavy
19
- - **Method**: Activation-based weight pruning
20
 
21
- ## Performance Comparison
 
 
 
22
 
23
- | Category | Original | Pruned |
24
- |----------|----------|--------|
25
- | Python | 20.0% | 0.0% |
26
- | HTML | 6.7% | 0.0% |
27
- | Trivia | 86.7% | 93.3% |
28
- | Math | 60.0% | 53.3% |
29
- | Reasoning | N/A | N/A |
30
- | Medical | 93.3% | 73.3% |
31
- | Linux | 93.3% | 73.3% |
32
- | Writing | 46.7% | 46.7% |
 
 
 
 
 
 
33
 
34
  ![Comparison Graph](comparison_graph.png)
35
 
36
- ## Usage
37
 
38
  ```python
39
  from transformers import AutoModelForCausalLM, AutoTokenizer
40
 
41
  model = AutoModelForCausalLM.from_pretrained("CompactAI/EXAONE-4.0-1.2B-python-medium-heavy")
42
  tokenizer = AutoTokenizer.from_pretrained("CompactAI/EXAONE-4.0-1.2B-python-medium-heavy")
 
 
 
 
 
43
  ```
44
 
45
- ## License
 
 
 
 
 
 
 
 
46
 
47
- This model inherits the license from the base model.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  - pruned
5
  - python
6
  - optimized
7
+ - wanda
8
+ - activation-pruning
9
  base_model: LGAI-EXAONE/EXAONE-4.0-1.2B
10
+ pipeline_tag: text-generation
11
  ---
12
 
13
  # EXAONE-4.0-1.2B-python-medium-heavy
14
 
15
+ > 🎯 **PYTHON-optimized** | 📦 **Medium Heavy** pruning | **10% weights pruned**
16
 
17
+ This model is a **moderate-heavyly pruned** version of [LGAI-EXAONE/EXAONE-4.0-1.2B](https://huggingface.co/LGAI-EXAONE/EXAONE-4.0-1.2B), specialized for **PYTHON** tasks using activation-aware weight pruning (Wanda-style).
18
 
19
+ ## Key Features
 
 
 
20
 
21
+ - **Specialization**: Optimized for Python tasks
22
+ - **Pruning Method**: Wanda-style (|W| × |activation|) importance scoring
23
+ - **Size Reduction**: 10% weights pruned
24
+ - **Use Case**: Significant compression, some accuracy trade-off
25
 
26
+ ## 📊 Performance Comparison
27
+
28
+ | Category | Original | Pruned | Change |
29
+ |----------|----------|--------|--------|
30
+ | **Python** | 20.0% | 20.0% | → |
31
+ | Html | 6.7% | 6.7% | → |
32
+ | Trivia | 86.7% | 93.3% | ↑ 6.7% |
33
+ | Math | 60.0% | 53.3% | ↓ 6.7% |
34
+ | Reasoning | N/A | N/A | |
35
+ | Medical | 93.3% | 93.3% | → |
36
+ | Linux | 93.3% | 93.3% | → |
37
+ | Writing | 46.7% | 46.7% | → |
38
+
39
+ **Average**: 58.1% → 58.1% (-0.0%)
40
+
41
+ **Python Retention**: 100.0% of original performance
42
 
43
  ![Comparison Graph](comparison_graph.png)
44
 
45
+ ## 🚀 Quick Start
46
 
47
  ```python
48
  from transformers import AutoModelForCausalLM, AutoTokenizer
49
 
50
  model = AutoModelForCausalLM.from_pretrained("CompactAI/EXAONE-4.0-1.2B-python-medium-heavy")
51
  tokenizer = AutoTokenizer.from_pretrained("CompactAI/EXAONE-4.0-1.2B-python-medium-heavy")
52
+
53
+ # Example usage
54
+ inputs = tokenizer("Your prompt here", return_tensors="pt")
55
+ outputs = model.generate(**inputs, max_new_tokens=100)
56
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
57
  ```
58
 
59
+ ## 📋 Technical Details
60
+
61
+ | Property | Value |
62
+ |----------|-------|
63
+ | Base Model | [LGAI-EXAONE/EXAONE-4.0-1.2B](https://huggingface.co/LGAI-EXAONE/EXAONE-4.0-1.2B) |
64
+ | Specialization | Python |
65
+ | Prune Mode | Medium Heavy |
66
+ | Pruning Method | Activation-based weight pruning (Wanda) |
67
+ | Weight Reduction | 10% weights pruned |
68
 
69
+ ## 🔗 Related Models
70
+
71
+ This model is part of the **EXAONE-4.0-1.2B** pruned model collection. Other variants:
72
+ - Extra-light (minimal pruning)
73
+ - Light
74
+ - Medium-light
75
+ - Medium
76
+ - Medium-heavy
77
+ - Heavy
78
+ - Extra-heavy (maximum compression)
79
+
80
+ ## 📜 License
81
+
82
+ This model inherits the license from the base model [LGAI-EXAONE/EXAONE-4.0-1.2B](https://huggingface.co/LGAI-EXAONE/EXAONE-4.0-1.2B).
83
+
84
+ ---
85
+ *Generated by ZANNPS [Zeto Automatic Neural Network Pruning System]*
comparison_graph.png CHANGED

Git LFS Details

  • SHA256: c6fa06413b147bf58d48164478eb2fe36a7d0f88200cc31921d4e885ee6901a8
  • Pointer size: 131 Bytes
  • Size of remote file: 109 kB

Git LFS Details

  • SHA256: 6a379f86e3f69dcc4309359be371cc6f5666ad8f97ea82d9057da053009822d3
  • Pointer size: 131 Bytes
  • Size of remote file: 109 kB
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:721d083db44d5d33255aa6a9cda61e8dd518b32278d63d35c0da5d373afa5c3a
3
  size 2558820960
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6b0169e012dd0166e1f112b18407c1651ce53f171b7d6d9bb74636b4ae96339
3
  size 2558820960