Spaces:
Running on Zero
Running on Zero
Sync backends: Fix infinite loop and prompt echo
Browse files
LightOnOCR-1B-Demo
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
Subproject commit
|
|
|
|
| 1 |
+
Subproject commit 6da9394d3f671f38eed2456ac31b270c476e8020
|
backends/pytorch_backend.py
CHANGED
|
@@ -93,12 +93,14 @@ class PyTorchBackend(OCRBackend):
|
|
| 93 |
if 'pixel_values' in inputs:
|
| 94 |
inputs['pixel_values'] = inputs['pixel_values'].to(self.dtype)
|
| 95 |
|
| 96 |
-
# Configure generation parameters
|
| 97 |
do_sample = temperature > 0.0
|
| 98 |
gen_kwargs = {
|
| 99 |
"max_new_tokens": max_tokens,
|
| 100 |
"pad_token_id": self.processor.tokenizer.eos_token_id,
|
| 101 |
-
"
|
|
|
|
|
|
|
| 102 |
}
|
| 103 |
|
| 104 |
if do_sample:
|
|
|
|
| 93 |
if 'pixel_values' in inputs:
|
| 94 |
inputs['pixel_values'] = inputs['pixel_values'].to(self.dtype)
|
| 95 |
|
| 96 |
+
# Configure generation parameters (aggressive anti-repetition for HF Space)
|
| 97 |
do_sample = temperature > 0.0
|
| 98 |
gen_kwargs = {
|
| 99 |
"max_new_tokens": max_tokens,
|
| 100 |
"pad_token_id": self.processor.tokenizer.eos_token_id,
|
| 101 |
+
"eos_token_id": self.processor.tokenizer.eos_token_id,
|
| 102 |
+
"repetition_penalty": 1.5, # Increased from 1.2
|
| 103 |
+
"early_stopping": True,
|
| 104 |
}
|
| 105 |
|
| 106 |
if do_sample:
|
hf_space/backends/pytorch_backend.py
CHANGED
|
@@ -93,12 +93,14 @@ class PyTorchBackend(OCRBackend):
|
|
| 93 |
if 'pixel_values' in inputs:
|
| 94 |
inputs['pixel_values'] = inputs['pixel_values'].to(self.dtype)
|
| 95 |
|
| 96 |
-
# Configure generation parameters
|
| 97 |
do_sample = temperature > 0.0
|
| 98 |
gen_kwargs = {
|
| 99 |
"max_new_tokens": max_tokens,
|
| 100 |
"pad_token_id": self.processor.tokenizer.eos_token_id,
|
| 101 |
-
"
|
|
|
|
|
|
|
| 102 |
}
|
| 103 |
|
| 104 |
if do_sample:
|
|
|
|
| 93 |
if 'pixel_values' in inputs:
|
| 94 |
inputs['pixel_values'] = inputs['pixel_values'].to(self.dtype)
|
| 95 |
|
| 96 |
+
# Configure generation parameters (aggressive anti-repetition for HF Space)
|
| 97 |
do_sample = temperature > 0.0
|
| 98 |
gen_kwargs = {
|
| 99 |
"max_new_tokens": max_tokens,
|
| 100 |
"pad_token_id": self.processor.tokenizer.eos_token_id,
|
| 101 |
+
"eos_token_id": self.processor.tokenizer.eos_token_id,
|
| 102 |
+
"repetition_penalty": 1.5, # Increased from 1.2
|
| 103 |
+
"early_stopping": True,
|
| 104 |
}
|
| 105 |
|
| 106 |
if do_sample:
|