llaa33219's picture
Upload README.md with huggingface_hub
eb32ca4 verified
metadata
license: apache-2.0
base_model: upstage/Solar-Open-100B
tags:
  - pruned
  - compressed
  - llm

Solar-Open-100B-pruned-5pct

This model is a pruned version of upstage/Solar-Open-100B.

Pruning Details

Property Value
Original Model upstage/Solar-Open-100B
Original Parameters 47.92B
Pruned Parameters 7.77B
Compression Ratio 0.1622 (6.2x smaller)
Strategy Layer + FFN (Recommended)
Importance Metric magnitude
Layers Removed 26
FFN Reduction 0.00%

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("llaa33219/Solar-Open-100B-pruned-5pct")
tokenizer = AutoTokenizer.from_pretrained("llaa33219/Solar-Open-100B-pruned-5pct")

Notes

This model was created using structured pruning techniques including:

  • Layer pruning (removing entire transformer layers)
  • FFN dimension pruning (reducing intermediate layer sizes)

The pruning was based on magnitude importance scoring to preserve the most important weights.