DocUA commited on
Commit
ff2c62c
·
1 Parent(s): 112b381

Sync backends: Fix infinite loop and prompt echo

Browse files
LightOnOCR-1B-Demo CHANGED
@@ -1 +1 @@
1
- Subproject commit 1378a75c56d8f6216571ade8ba579eefca11e38e
 
1
+ Subproject commit 6da9394d3f671f38eed2456ac31b270c476e8020
backends/pytorch_backend.py CHANGED
@@ -93,12 +93,14 @@ class PyTorchBackend(OCRBackend):
93
  if 'pixel_values' in inputs:
94
  inputs['pixel_values'] = inputs['pixel_values'].to(self.dtype)
95
 
96
- # Configure generation parameters
97
  do_sample = temperature > 0.0
98
  gen_kwargs = {
99
  "max_new_tokens": max_tokens,
100
  "pad_token_id": self.processor.tokenizer.eos_token_id,
101
- "repetition_penalty": 1.2,
 
 
102
  }
103
 
104
  if do_sample:
 
93
  if 'pixel_values' in inputs:
94
  inputs['pixel_values'] = inputs['pixel_values'].to(self.dtype)
95
 
96
+ # Configure generation parameters (aggressive anti-repetition for HF Space)
97
  do_sample = temperature > 0.0
98
  gen_kwargs = {
99
  "max_new_tokens": max_tokens,
100
  "pad_token_id": self.processor.tokenizer.eos_token_id,
101
+ "eos_token_id": self.processor.tokenizer.eos_token_id,
102
+ "repetition_penalty": 1.5, # Increased from 1.2
103
+ "early_stopping": True,
104
  }
105
 
106
  if do_sample:
hf_space/backends/pytorch_backend.py CHANGED
@@ -93,12 +93,14 @@ class PyTorchBackend(OCRBackend):
93
  if 'pixel_values' in inputs:
94
  inputs['pixel_values'] = inputs['pixel_values'].to(self.dtype)
95
 
96
- # Configure generation parameters
97
  do_sample = temperature > 0.0
98
  gen_kwargs = {
99
  "max_new_tokens": max_tokens,
100
  "pad_token_id": self.processor.tokenizer.eos_token_id,
101
- "repetition_penalty": 1.2,
 
 
102
  }
103
 
104
  if do_sample:
 
93
  if 'pixel_values' in inputs:
94
  inputs['pixel_values'] = inputs['pixel_values'].to(self.dtype)
95
 
96
+ # Configure generation parameters (aggressive anti-repetition for HF Space)
97
  do_sample = temperature > 0.0
98
  gen_kwargs = {
99
  "max_new_tokens": max_tokens,
100
  "pad_token_id": self.processor.tokenizer.eos_token_id,
101
+ "eos_token_id": self.processor.tokenizer.eos_token_id,
102
+ "repetition_penalty": 1.5, # Increased from 1.2
103
+ "early_stopping": True,
104
  }
105
 
106
  if do_sample: