| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| import torch |
| import os |
| from transformers import AutoTokenizer |
| from tqdm import tqdm |
|
|
| def stream_docs(file_path, delimiter="<|end_of_text|>"): |
| buffer = "" |
| with open(file_path, 'r', encoding='utf-8') as f: |
| while True: |
| chunk = f.read(1024 * 1024) |
| if not chunk: |
| if buffer.strip(): yield buffer |
| break |
| buffer += chunk |
| while delimiter in buffer: |
| doc, buffer = buffer.split(delimiter, 1) |
| if doc.strip(): yield doc |
|
|
| def tokenize_with_overlap( |
| input_file="jirack_base_dataset.txt", |
| |
| model_id=".", |
| chunk_size=2000, |
| max_length=8192, |
| overlap_size=512, |
| output_prefix="jirack_overlap_data" |
| ): |
| print(f"📥 Загрузка токенизатора: {model_id}") |
| tokenizer = AutoTokenizer.from_pretrained(model_id) |
| |
| |
| if tokenizer.pad_token_id is None: |
| if tokenizer.eos_token_id is not None: |
| tokenizer.pad_token_id = tokenizer.eos_token_id |
| else: |
| tokenizer.pad_token_id = 128004 |
| |
| pad_id = tokenizer.pad_token_id |
| print(f"🛠 Используемый Pad Token ID: {pad_id}") |
|
|
| stride = max_length - overlap_size |
| input_ids_buffer = [] |
| labels_buffer = [] |
| chunk_idx = 0 |
|
|
| def save_chunk(ids, labels, idx): |
| if not ids: return |
| filename = f"{output_prefix}_{idx}.pt" |
| torch.save({ |
| "input_ids": torch.stack(ids).to(torch.int64), |
| "labels": torch.stack(labels).to(torch.int64) |
| }, filename) |
| print(f"\n💾 Сохранен чанк {idx}: {filename} ({len(ids)} строк)") |
|
|
| print(f"🔄 Нарезка 36GB файла. Окно: {max_length}, Нахлест: {overlap_size}") |
|
|
| for doc in tqdm(stream_docs(input_file), desc="Processing"): |
| try: |
| text = doc.strip() |
| if not text: continue |
|
|
| full_text = f"<|begin_of_text|>{text}<|end_of_text|>" |
| full_ids = tokenizer.encode(full_text, add_special_tokens=False) |
| |
| if not full_ids: continue |
|
|
| |
| windows = [] |
| if len(full_ids) <= max_length: |
| windows.append(full_ids) |
| else: |
| for i in range(0, len(full_ids), stride): |
| w = full_ids[i : i + max_length] |
| if len(w) > 10: |
| windows.append(w) |
|
|
| for w in windows: |
| ids = list(w) |
| lbs = list(w) |
|
|
| if len(ids) < max_length: |
| pad_len = max_length - len(ids) |
| |
| ids += [pad_id] * pad_len |
| lbs += [-100] * pad_len |
| |
| input_ids_buffer.append(torch.tensor(ids, dtype=torch.int64)) |
| labels_buffer.append(torch.tensor(lbs, dtype=torch.int64)) |
|
|
| if len(input_ids_buffer) >= chunk_size: |
| save_chunk(input_ids_buffer, labels_buffer, chunk_idx) |
| chunk_idx += 1 |
| input_ids_buffer, labels_buffer = [], [] |
|
|
| except Exception as e: |
| |
| print(f"\n⚠️ Ошибка: {e}") |
| continue |
|
|
| if input_ids_buffer: |
| save_chunk(input_ids_buffer, labels_buffer, chunk_idx) |
|
|
| if __name__ == "__main__": |
| tokenize_with_overlap() |
|
|