YAML Metadata Warning:empty or missing yaml metadata in repo card
Check out the documentation for more information.
---
library_name: pytorch
tags:
- llama
- causal-lm
- text-generation
- pruning
- knowledge-distillation
- speedup
license: apache-2.0
dataset: slimpajama-test
pipeline_tag: text-generation
---
hawada/Llama-3.2-1B-h100-slim
This repository contains two variants:
- Gated student (with learned pruning gates) – requires custom code.
- Slim student (post-prune/export) – loads with standard HF APIs plus this repo’s custom code.
Inference (LLaMA slim)
from transformers import AutoModelForCausalLM, AutoTokenizer
tok = AutoTokenizer.from_pretrained('hawada/Llama-3.2-1B-h100-slim')
mdl = AutoModelForCausalLM.from_pretrained('hawada/Llama-3.2-1B-h100-slim', torch_dtype='auto').eval()
x = tok('Hello', return_tensors='pt')
print(tok.decode(mdl.generate(**x, max_new_tokens=16)[0]))
Notes
- The gated repo includes lightweight custom code (adapters/…, core/…) needed to attach/load gates.
- The slim model is exported for efficient inference.
Training metadata
{
"base_id": "meta-llama/Llama-3.2-1B",
"variant": "slim-export",
"repo_slim": "hawada/Llama-3.2-1B-h100-slim"
}
- Downloads last month
- 2
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support