Upload train_ministral_n8n.py with huggingface_hub
Browse files- train_ministral_n8n.py +5 -4
train_ministral_n8n.py
CHANGED
|
@@ -2,7 +2,7 @@
|
|
| 2 |
# requires-python = ">=3.10"
|
| 3 |
# dependencies = [
|
| 4 |
# "torch>=2.0.0",
|
| 5 |
-
# "transformers>=4.
|
| 6 |
# "datasets>=2.14.0",
|
| 7 |
# "accelerate>=0.24.0",
|
| 8 |
# "peft>=0.7.0",
|
|
@@ -10,6 +10,8 @@
|
|
| 10 |
# "bitsandbytes>=0.41.0",
|
| 11 |
# "huggingface_hub>=0.20.0",
|
| 12 |
# "tokenizers>=0.21.0",
|
|
|
|
|
|
|
| 13 |
# "trackio",
|
| 14 |
# ]
|
| 15 |
# ///
|
|
@@ -37,9 +39,8 @@ MAX_SEQ_LENGTH = 4096 # n8n workflows can be long
|
|
| 37 |
trackio.init(project="ministral-3b-n8n-sft")
|
| 38 |
|
| 39 |
print(f"Loading tokenizer from {MODEL_NAME}...")
|
| 40 |
-
# Ministral uses
|
| 41 |
-
|
| 42 |
-
tokenizer = PreTrainedTokenizerFast.from_pretrained(MODEL_NAME)
|
| 43 |
if tokenizer.pad_token is None:
|
| 44 |
tokenizer.pad_token = tokenizer.eos_token
|
| 45 |
|
|
|
|
| 2 |
# requires-python = ">=3.10"
|
| 3 |
# dependencies = [
|
| 4 |
# "torch>=2.0.0",
|
| 5 |
+
# "transformers>=4.51.0",
|
| 6 |
# "datasets>=2.14.0",
|
| 7 |
# "accelerate>=0.24.0",
|
| 8 |
# "peft>=0.7.0",
|
|
|
|
| 10 |
# "bitsandbytes>=0.41.0",
|
| 11 |
# "huggingface_hub>=0.20.0",
|
| 12 |
# "tokenizers>=0.21.0",
|
| 13 |
+
# "protobuf",
|
| 14 |
+
# "mistral-common>=1.5.0",
|
| 15 |
# "trackio",
|
| 16 |
# ]
|
| 17 |
# ///
|
|
|
|
| 39 |
trackio.init(project="ministral-3b-n8n-sft")
|
| 40 |
|
| 41 |
print(f"Loading tokenizer from {MODEL_NAME}...")
|
| 42 |
+
# Ministral uses Tekken tokenizer - use AutoTokenizer with trust_remote_code
|
| 43 |
+
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, trust_remote_code=True)
|
|
|
|
| 44 |
if tokenizer.pad_token is None:
|
| 45 |
tokenizer.pad_token = tokenizer.eos_token
|
| 46 |
|