bobber commited on
Commit
3b5c2b9
·
verified ·
1 Parent(s): 95848e5

Update to routangseng-voice-4b with matching system prompt and UI

Browse files
Files changed (3) hide show
  1. README.md +9 -6
  2. app.py +17 -14
  3. requirements.txt +4 -11
README.md CHANGED
@@ -1,16 +1,19 @@
1
  ---
2
- title: routangseng-chat-gpu
3
  emoji: 🥩
4
- colorFrom: indigo
5
- colorTo: pink
6
  sdk: gradio
7
- sdk_version: 5.46.1
8
- python_version: 3.10.13
9
  app_file: app.py
10
  pinned: false
 
11
  suggested_hardware: zero-a10g
12
  ---
13
 
14
  # 肉糖生 Chat GPU
15
 
16
- Gradio + ZeroGPU + Transformers runtime bootstrap for `bobber/routangseng-qwen35-0.8b-abliterated`.
 
 
 
 
1
  ---
2
+ title: 肉糖生 Chat GPU
3
  emoji: 🥩
4
+ colorFrom: red
5
+ colorTo: purple
6
  sdk: gradio
7
+ sdk_version: "5.23.0"
 
8
  app_file: app.py
9
  pinned: false
10
+ license: apache-2.0
11
  suggested_hardware: zero-a10g
12
  ---
13
 
14
  # 肉糖生 Chat GPU
15
 
16
+ Qwen3.5-4B voice LoRA fine-tuned model running on ZeroGPU.
17
+
18
+ - Model: `bobber/routangseng-voice-4b`
19
+ - Style: 结论先行,不和稀泥
app.py CHANGED
@@ -9,9 +9,9 @@ import gradio as gr
9
  import spaces
10
  import torch
11
 
12
- MODEL_ID = "bobber/routangseng-qwen35-0.8b-abliterated"
13
- SYSTEM_PROMPT = "你是一个中文内容创作者,表达理性、结构化、接地气结论展开并保持多轮对话一致性。"
14
- MAX_TOKENS = 512
15
  TEMPERATURE = 0.7
16
  TOP_P = 0.9
17
 
@@ -103,7 +103,7 @@ def load_model():
103
  _tokenizer = AutoTokenizer.from_pretrained(MODEL_ID, trust_remote_code=False)
104
  _model = AutoModelForCausalLM.from_pretrained(
105
  MODEL_ID,
106
- torch_dtype=torch.float16,
107
  device_map="auto",
108
  trust_remote_code=False,
109
  )
@@ -148,26 +148,29 @@ def chat_fn(message, history):
148
  thread.join()
149
 
150
 
151
- with gr.Blocks(theme=gr.themes.Soft()) as demo:
 
 
 
 
 
152
  gr.Markdown(
153
  """
154
  # 🥩 肉糖生 Chat
155
 
156
- Qwen3.5-0.8B SFT · ZeroGPU · Transformers
 
 
157
  """
158
  )
159
- gr.Markdown(
160
- "首次启动可能会慢一些:Space 需要在运行时引导安装最新 Transformers 主线版本,以获得 Qwen3.5 支持。"
161
- )
162
  gr.ChatInterface(
163
  fn=chat_fn,
164
  type="messages",
165
- title="肉糖生 · Qwen3.5-0.8B Abliterated SFT",
166
- description="基于 bobber/routangseng-qwen35-0.8b-abliterated",
167
  examples=[
168
- "用三句话介绍你自己。",
169
- "帮我写一段中文产品介绍,风格理性、接地气。",
170
- "怎么理解机器人+建筑行业的结合?",
 
171
  ],
172
  fill_height=True,
173
  )
 
9
  import spaces
10
  import torch
11
 
12
+ MODEL_ID = "bobber/routangseng-voice-4b"
13
+ SYSTEM_PROMPT = "你是肉糖生,一个接地气的中文时政分析。风格:结论先行再用结构化分析展开;敢于质疑主流叙事不和稀泥;用类比和现实例子把复杂问题透;语气直率但逻辑严密。回答时先给核心判断,再分层拆解最后给出预测或建议。"
14
+ MAX_TOKENS = 1024
15
  TEMPERATURE = 0.7
16
  TOP_P = 0.9
17
 
 
103
  _tokenizer = AutoTokenizer.from_pretrained(MODEL_ID, trust_remote_code=False)
104
  _model = AutoModelForCausalLM.from_pretrained(
105
  MODEL_ID,
106
+ torch_dtype=torch.bfloat16,
107
  device_map="auto",
108
  trust_remote_code=False,
109
  )
 
148
  thread.join()
149
 
150
 
151
+ CSS = """
152
+ .gradio-container { max-width: 900px !important; }
153
+ footer { display: none !important; }
154
+ """
155
+
156
+ with gr.Blocks(theme=gr.themes.Soft(), css=CSS) as demo:
157
  gr.Markdown(
158
  """
159
  # 🥩 肉糖生 Chat
160
 
161
+ **Qwen3.5-4B Voice LoRA** · ZeroGPU · 结论先行,不和稀泥
162
+
163
+ <small>基于 `bobber/routangseng-voice-4b` · 首次启动需安装最新 Transformers(约30秒)</small>
164
  """
165
  )
 
 
 
166
  gr.ChatInterface(
167
  fn=chat_fn,
168
  type="messages",
 
 
169
  examples=[
170
+ "白领工作都被AI不断代替,现在学生还在传统方式积累白领知识,这不是学了个寂寞嘛?",
171
+ "为什么很多国家的年轻人不想生孩子?这个趋势能逆转吗?",
172
+ "中美关系未来五年会怎么从结构性矛盾的角度讲讲。",
173
+ "怎么理解"小共同体"和"大共同体"的关系?",
174
  ],
175
  fill_height=True,
176
  )
requirements.txt CHANGED
@@ -1,11 +1,4 @@
1
- accelerate>=1.0.0
2
- regex
3
- tokenizers
4
- safetensors
5
- sentencepiece
6
- filelock
7
- packaging
8
- requests
9
- pyyaml
10
- numpy
11
- tqdm
 
1
+ gradio>=4.0
2
+ torch
3
+ spaces
4
+ peft