minpeter commited on
Commit
ebb29ac
·
verified ·
1 Parent(s): d25f674

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +132 -0
README.md ADDED
@@ -0,0 +1,132 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: hyperclovax
4
+ license_link: https://huggingface.co/naver-hyperclovax/HyperCLOVAX-SEED-Think-32B/blob/main/LICENSE
5
+ library_name: transformers
6
+ base_model: naver-hyperclovax/HyperCLOVAX-SEED-Think-32B
7
+ tags:
8
+ - llama
9
+ - text-generation
10
+ - korean
11
+ - reasoning
12
+ language:
13
+ - ko
14
+ - en
15
+ pipeline_tag: text-generation
16
+ ---
17
+
18
+ # HyperCLOVAX-SEED-Text-Think-32B
19
+
20
+ **Extracted text-only LLM from [naver-hyperclovax/HyperCLOVAX-SEED-Think-32B](https://huggingface.co/naver-hyperclovax/HyperCLOVAX-SEED-Think-32B)**
21
+
22
+ This model contains only the language model component extracted from the original Vision-Language Model (VLM). The vision encoder and multimodal projector have been removed, making it a pure text-to-text model compatible with standard LLaMA inference pipelines.
23
+
24
+ ## Model Details
25
+
26
+ | Property | Value |
27
+ |----------|-------|
28
+ | Architecture | LlamaForCausalLM |
29
+ | Parameters | ~33B |
30
+ | Hidden Size | 5120 |
31
+ | Layers | 72 |
32
+ | Attention Heads | 40 |
33
+ | KV Heads | 8 (GQA) |
34
+ | Intermediate Size | 24192 |
35
+ | Context Length | 128K |
36
+ | Vocab Size | 128,256 |
37
+ | Precision | bfloat16 |
38
+ | RoPE Theta | 50,000,000 |
39
+
40
+ ## What Was Extracted
41
+
42
+ The original VLM consists of:
43
+ - **Vision Encoder**: Qwen2.5-VL based (~600M params) - **removed**
44
+ - **MM Projector**: Multimodal projection layers - **removed**
45
+ - **Language Model**: HyperCLOVAX LLM (~33B params) - **extracted** ✓
46
+
47
+ Only the `model.language_model.*` weights were extracted and remapped to standard LLaMA format.
48
+
49
+ ## Usage
50
+
51
+ ### With Transformers
52
+
53
+ ```python
54
+ from transformers import AutoModelForCausalLM, AutoTokenizer
55
+
56
+ model_id = "minpeter/HyperCLOVAX-SEED-Text-Think-32B-hf"
57
+
58
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
59
+ model = AutoModelForCausalLM.from_pretrained(
60
+ model_id,
61
+ torch_dtype="bfloat16",
62
+ device_map="auto"
63
+ )
64
+
65
+ messages = [{"role": "user", "content": "What is the capital of South Korea?"}]
66
+ inputs = tokenizer.apply_chat_template(messages, return_tensors="pt", add_generation_prompt=True)
67
+ outputs = model.generate(inputs.to(model.device), max_new_tokens=512)
68
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
69
+ ```
70
+
71
+ ### With vLLM
72
+
73
+ ```bash
74
+ vllm serve minpeter/HyperCLOVAX-SEED-Text-Think-32B-hf \
75
+ --dtype bfloat16 \
76
+ --tensor-parallel-size 2
77
+ ```
78
+
79
+ ```python
80
+ from openai import OpenAI
81
+
82
+ client = OpenAI(base_url="http://localhost:8000/v1", api_key="dummy")
83
+ response = client.chat.completions.create(
84
+ model="minpeter/HyperCLOVAX-SEED-Text-Think-32B-hf",
85
+ messages=[{"role": "user", "content": "안녕하세요! 한국어로 대화할 수 있나요?"}]
86
+ )
87
+ print(response.choices[0].message.content)
88
+ ```
89
+
90
+ ## Thinking Mode
91
+
92
+ The model supports a "thinking mode" for complex reasoning tasks. Use the `<|thinking|>` token to trigger extended reasoning:
93
+
94
+ ```python
95
+ messages = [
96
+ {"role": "user", "content": "Solve this step by step: If x + 2y = 10 and 3x - y = 5, find x and y."}
97
+ ]
98
+ # The model may produce <|thinking|>...</|thinking|> blocks with its reasoning process
99
+ ```
100
+
101
+ ## Hardware Requirements
102
+
103
+ - **Minimum**: 2x NVIDIA A100 40GB (with tensor parallelism)
104
+ - **Recommended**: 2x NVIDIA A100 80GB or 4x NVIDIA A6000
105
+
106
+ ## Limitations
107
+
108
+ - This is a **text-only** model. It cannot process images or videos.
109
+ - The model inherits any limitations from the original HyperCLOVAX-SEED-Think-32B.
110
+ - Optimized primarily for Korean and English.
111
+
112
+ ## License
113
+
114
+ This model inherits the [HyperCLOVAX license](https://huggingface.co/naver-hyperclovax/HyperCLOVAX-SEED-Think-32B/blob/main/LICENSE) from the original model.
115
+
116
+ ## Citation
117
+
118
+ If you use this model, please cite the original:
119
+
120
+ ```bibtex
121
+ @misc{hyperclovax-seed-think-32b,
122
+ title={HyperCLOVA X SEED Think 32B},
123
+ author={NAVER Cloud},
124
+ year={2025},
125
+ url={https://huggingface.co/naver-hyperclovax/HyperCLOVAX-SEED-Think-32B}
126
+ }
127
+ ```
128
+
129
+ ## Acknowledgments
130
+
131
+ - Original model by [NAVER Cloud HyperCLOVA X](https://huggingface.co/naver-hyperclovax)
132
+ - Extraction performed to enable text-only inference without vision dependencies