Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -127,7 +127,6 @@ def _safe_load_lm(model_id: str, device: str) -> AutoModelForCausalLM:
|
|
| 127 |
return m
|
| 128 |
|
| 129 |
def load_models(force_device: str | None = None):
|
| 130 |
-
"""本地:加载并缓存模型(无 spaces/ZeroGPU)"""
|
| 131 |
global model, codec_model, asr_model, tokenizer
|
| 132 |
|
| 133 |
logger.info(f"Using device: {device}")
|
|
@@ -157,7 +156,7 @@ def load_models(force_device: str | None = None):
|
|
| 157 |
|
| 158 |
load_models()
|
| 159 |
|
| 160 |
-
@
|
| 161 |
def inference_batch_transformers(
|
| 162 |
lm: AutoModelForCausalLM,
|
| 163 |
codec: XCodec2Model,
|
|
|
|
| 127 |
return m
|
| 128 |
|
| 129 |
def load_models(force_device: str | None = None):
|
|
|
|
| 130 |
global model, codec_model, asr_model, tokenizer
|
| 131 |
|
| 132 |
logger.info(f"Using device: {device}")
|
|
|
|
| 156 |
|
| 157 |
load_models()
|
| 158 |
|
| 159 |
+
@spaces.GPU
|
| 160 |
def inference_batch_transformers(
|
| 161 |
lm: AutoModelForCausalLM,
|
| 162 |
codec: XCodec2Model,
|