davanstrien HF Staff commited on
Commit
773eb1a
·
verified ·
1 Parent(s): 82163cc

Upload vlm-streaming-sft-unsloth-qwen.py with huggingface_hub

Browse files
Files changed (1) hide show
  1. vlm-streaming-sft-unsloth-qwen.py +3 -8
vlm-streaming-sft-unsloth-qwen.py CHANGED
@@ -296,21 +296,16 @@ def main():
296
  )
297
 
298
  # Convert streaming dataset to list (required for Qwen3-VL per Unsloth docs)
299
- # "Using map kicks in dataset standardization which can be complicated"
300
  print(" Converting streaming dataset to list...")
301
  train_data = list(dataset.take(500)) # Take enough samples for training
302
  print(f" Loaded {len(train_data)} samples")
303
 
304
- # Get the actual tokenizer object (Qwen may return processor-like object)
305
- actual_tokenizer = getattr(tokenizer, 'tokenizer', tokenizer)
306
- print(f" Tokenizer type: {type(tokenizer)}")
307
- print(f" Actual tokenizer type: {type(actual_tokenizer)}")
308
-
309
  trainer = SFTTrainer(
310
  model=model,
311
- train_dataset=train_data,
312
- processing_class=actual_tokenizer,
313
  data_collator=UnslothVisionDataCollator(model, tokenizer),
 
314
  args=training_config,
315
  )
316
 
 
296
  )
297
 
298
  # Convert streaming dataset to list (required for Qwen3-VL per Unsloth docs)
 
299
  print(" Converting streaming dataset to list...")
300
  train_data = list(dataset.take(500)) # Take enough samples for training
301
  print(f" Loaded {len(train_data)} samples")
302
 
303
+ # Use older 'tokenizer=' parameter (not processing_class) - required for Unsloth VLM
 
 
 
 
304
  trainer = SFTTrainer(
305
  model=model,
306
+ tokenizer=tokenizer, # Full processor, not processor.tokenizer
 
307
  data_collator=UnslothVisionDataCollator(model, tokenizer),
308
+ train_dataset=train_data,
309
  args=training_config,
310
  )
311