Upload vlm-streaming-sft-unsloth-qwen.py with huggingface_hub
Browse files
vlm-streaming-sft-unsloth-qwen.py
CHANGED
|
@@ -296,21 +296,16 @@ def main():
|
|
| 296 |
)
|
| 297 |
|
| 298 |
# Convert streaming dataset to list (required for Qwen3-VL per Unsloth docs)
|
| 299 |
-
# "Using map kicks in dataset standardization which can be complicated"
|
| 300 |
print(" Converting streaming dataset to list...")
|
| 301 |
train_data = list(dataset.take(500)) # Take enough samples for training
|
| 302 |
print(f" Loaded {len(train_data)} samples")
|
| 303 |
|
| 304 |
-
#
|
| 305 |
-
actual_tokenizer = getattr(tokenizer, 'tokenizer', tokenizer)
|
| 306 |
-
print(f" Tokenizer type: {type(tokenizer)}")
|
| 307 |
-
print(f" Actual tokenizer type: {type(actual_tokenizer)}")
|
| 308 |
-
|
| 309 |
trainer = SFTTrainer(
|
| 310 |
model=model,
|
| 311 |
-
|
| 312 |
-
processing_class=actual_tokenizer,
|
| 313 |
data_collator=UnslothVisionDataCollator(model, tokenizer),
|
|
|
|
| 314 |
args=training_config,
|
| 315 |
)
|
| 316 |
|
|
|
|
| 296 |
)
|
| 297 |
|
| 298 |
# Convert streaming dataset to list (required for Qwen3-VL per Unsloth docs)
|
|
|
|
| 299 |
print(" Converting streaming dataset to list...")
|
| 300 |
train_data = list(dataset.take(500)) # Take enough samples for training
|
| 301 |
print(f" Loaded {len(train_data)} samples")
|
| 302 |
|
| 303 |
+
# Use older 'tokenizer=' parameter (not processing_class) - required for Unsloth VLM
|
|
|
|
|
|
|
|
|
|
|
|
|
| 304 |
trainer = SFTTrainer(
|
| 305 |
model=model,
|
| 306 |
+
tokenizer=tokenizer, # Full processor, not processor.tokenizer
|
|
|
|
| 307 |
data_collator=UnslothVisionDataCollator(model, tokenizer),
|
| 308 |
+
train_dataset=train_data,
|
| 309 |
args=training_config,
|
| 310 |
)
|
| 311 |
|