Spaces:
Sleeping
Sleeping
Removed streaming
Browse files
invert.py
CHANGED
|
@@ -84,7 +84,7 @@ def _build_mask_token_ids(tokenizer: AutoTokenizer) -> list[int]:
|
|
| 84 |
|
| 85 |
mask_ids = set()
|
| 86 |
for s in _MASK_STRINGS:
|
| 87 |
-
tokens = tokenizer.encode(s, add_special_tokens=False)
|
| 88 |
if len(tokens) == 1:
|
| 89 |
mask_ids.add(tokens[0])
|
| 90 |
if tokenizer.eos_token_id is not None:
|
|
@@ -244,7 +244,7 @@ def beam_search(
|
|
| 244 |
Best candidate found during search.
|
| 245 |
"""
|
| 246 |
prefix, suffix = get_chat_format(tokenizer)
|
| 247 |
-
prompt_tokens = tokenizer.encode(prompt, add_special_tokens=False)
|
| 248 |
mask_ids = _build_mask_token_ids(tokenizer)
|
| 249 |
|
| 250 |
candidates = [Candidate()]
|
|
|
|
| 84 |
|
| 85 |
mask_ids = set()
|
| 86 |
for s in _MASK_STRINGS:
|
| 87 |
+
tokens = list(tokenizer.encode(s, add_special_tokens=False))
|
| 88 |
if len(tokens) == 1:
|
| 89 |
mask_ids.add(tokens[0])
|
| 90 |
if tokenizer.eos_token_id is not None:
|
|
|
|
| 244 |
Best candidate found during search.
|
| 245 |
"""
|
| 246 |
prefix, suffix = get_chat_format(tokenizer)
|
| 247 |
+
prompt_tokens = list(tokenizer.encode(prompt, add_special_tokens=False))
|
| 248 |
mask_ids = _build_mask_token_ids(tokenizer)
|
| 249 |
|
| 250 |
candidates = [Candidate()]
|
model.py
CHANGED
|
@@ -82,7 +82,7 @@ def get_chat_format(tokenizer: AutoTokenizer) -> tuple[list[int], list[int]]:
|
|
| 82 |
[{"role": "user", "content": "hello"}],
|
| 83 |
add_generation_prompt=True,
|
| 84 |
)
|
| 85 |
-
marker_tokens = tokenizer.encode("hello", add_special_tokens=False)
|
| 86 |
|
| 87 |
# Find where the marker content appears in the full template
|
| 88 |
marker_len = len(marker_tokens)
|
|
|
|
| 82 |
[{"role": "user", "content": "hello"}],
|
| 83 |
add_generation_prompt=True,
|
| 84 |
)
|
| 85 |
+
marker_tokens = list(tokenizer.encode("hello", add_special_tokens=False))
|
| 86 |
|
| 87 |
# Find where the marker content appears in the full template
|
| 88 |
marker_len = len(marker_tokens)
|