KevinKeller commited on
Commit
dd32d27
·
verified ·
1 Parent(s): 239cbc2

Upload train_question_generator.py with huggingface_hub

Browse files
Files changed (1) hide show
  1. train_question_generator.py +12 -4
train_question_generator.py CHANGED
@@ -1,5 +1,5 @@
1
  # /// script
2
- # dependencies = ["trl>=0.17.0", "peft>=0.15.0", "datasets", "transformers", "accelerate", "bitsandbytes"]
3
  # ///
4
 
5
  import os
@@ -9,6 +9,13 @@ from trl import SFTTrainer, SFTConfig
9
  from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig
10
  import torch
11
 
 
 
 
 
 
 
 
12
  print("Loading dataset...")
13
  dataset = load_dataset("KevinKeller/cognitive-question-generator-v1")
14
  train_dataset = dataset["train"]
@@ -18,7 +25,7 @@ print(f"Train samples: {len(train_dataset)}")
18
  if eval_dataset:
19
  print(f"Eval samples: {len(eval_dataset)}")
20
 
21
- # Using Qwen2.5-7B for question generation (good reasoning capabilities)
22
  print("Loading model: Qwen/Qwen2.5-7B-Instruct...")
23
  model_id = "Qwen/Qwen2.5-7B-Instruct"
24
 
@@ -30,7 +37,8 @@ bnb_config = BitsAndBytesConfig(
30
  )
31
 
32
  tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
33
- tokenizer.pad_token = tokenizer.eos_token
 
34
 
35
  model = AutoModelForCausalLM.from_pretrained(
36
  model_id,
@@ -75,7 +83,7 @@ trainer = SFTTrainer(
75
  train_dataset=train_dataset,
76
  eval_dataset=eval_dataset,
77
  peft_config=peft_config,
78
- processing_class=tokenizer,
79
  args=training_args,
80
  )
81
 
 
1
  # /// script
2
+ # dependencies = ["trl>=0.12.0", "peft>=0.13.0", "datasets", "transformers>=4.45.0", "accelerate", "bitsandbytes", "huggingface_hub"]
3
  # ///
4
 
5
  import os
 
9
  from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig
10
  import torch
11
 
12
+ # Authenticate
13
+ from huggingface_hub import login
14
+ hf_token = os.environ.get("HF_TOKEN")
15
+ if hf_token:
16
+ login(token=hf_token)
17
+ print("Authenticated with HuggingFace")
18
+
19
  print("Loading dataset...")
20
  dataset = load_dataset("KevinKeller/cognitive-question-generator-v1")
21
  train_dataset = dataset["train"]
 
25
  if eval_dataset:
26
  print(f"Eval samples: {len(eval_dataset)}")
27
 
28
+ # Using Qwen2.5-7B for question generation
29
  print("Loading model: Qwen/Qwen2.5-7B-Instruct...")
30
  model_id = "Qwen/Qwen2.5-7B-Instruct"
31
 
 
37
  )
38
 
39
  tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
40
+ if tokenizer.pad_token is None:
41
+ tokenizer.pad_token = tokenizer.eos_token
42
 
43
  model = AutoModelForCausalLM.from_pretrained(
44
  model_id,
 
83
  train_dataset=train_dataset,
84
  eval_dataset=eval_dataset,
85
  peft_config=peft_config,
86
+ tokenizer=tokenizer,
87
  args=training_args,
88
  )
89