Upload vlm-streaming-sft-unsloth-qwen.py with huggingface_hub
Browse files
vlm-streaming-sft-unsloth-qwen.py
CHANGED
|
@@ -300,11 +300,16 @@ def main():
|
|
| 300 |
train_data = list(dataset.take(500)) # Take enough samples for training
|
| 301 |
print(f" Loaded {len(train_data)} samples")
|
| 302 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 303 |
trainer = SFTTrainer(
|
| 304 |
model=model,
|
| 305 |
-
tokenizer=tokenizer,
|
| 306 |
-
data_collator=UnslothVisionDataCollator(model, tokenizer),
|
| 307 |
train_dataset=train_data,
|
|
|
|
|
|
|
| 308 |
args=training_config,
|
| 309 |
)
|
| 310 |
|
|
|
|
| 300 |
train_data = list(dataset.take(500)) # Take enough samples for training
|
| 301 |
print(f" Loaded {len(train_data)} samples")
|
| 302 |
|
| 303 |
+
# Get the actual tokenizer object (Qwen may return processor-like object)
|
| 304 |
+
actual_tokenizer = getattr(tokenizer, 'tokenizer', tokenizer)
|
| 305 |
+
print(f" Tokenizer type: {type(tokenizer)}")
|
| 306 |
+
print(f" Actual tokenizer type: {type(actual_tokenizer)}")
|
| 307 |
+
|
| 308 |
trainer = SFTTrainer(
|
| 309 |
model=model,
|
|
|
|
|
|
|
| 310 |
train_dataset=train_data,
|
| 311 |
+
processing_class=actual_tokenizer,
|
| 312 |
+
data_collator=UnslothVisionDataCollator(model, tokenizer),
|
| 313 |
args=training_config,
|
| 314 |
)
|
| 315 |
|