Spaces:
Sleeping
Sleeping
Yusuf
commited on
Commit
·
7ef5142
1
Parent(s):
355c513
fix: restore run training script
Browse files
trainingModel/run_training.py
CHANGED
|
@@ -1,18 +1,15 @@
|
|
| 1 |
import os
|
| 2 |
import numpy as np
|
|
|
|
| 3 |
from clearml import Task, Dataset
|
| 4 |
from datasets import load_dataset
|
|
|
|
| 5 |
|
| 6 |
-
|
| 7 |
-
|
| 8 |
-
|
| 9 |
-
raise RuntimeError("No tasks found in project 'Small Group Project'")
|
| 10 |
|
| 11 |
-
dp_tasks = [t for t in all_tasks if t.name == "Data Preparation"]
|
| 12 |
-
if not dp_tasks:
|
| 13 |
-
raise RuntimeError("No 'Data Preparation' tasks found in this project!")
|
| 14 |
|
| 15 |
-
<<<<<<< HEAD
|
| 16 |
# -------------- Load Data --------------
|
| 17 |
|
| 18 |
all_tasks = Task.get_tasks(project_name="Small Group Project")
|
|
@@ -28,12 +25,6 @@ latest_task = max(dp_tasks, key=lambda t: t.id)
|
|
| 28 |
DYNAMIC_TASK_ID = latest_task.id
|
| 29 |
DATA_PREP = Task.get_task(task_id=DYNAMIC_TASK_ID)
|
| 30 |
|
| 31 |
-
=======
|
| 32 |
-
latest_task = max(dp_tasks, key=lambda t: t.id)
|
| 33 |
-
DYNAMIC_TASK_ID = latest_task.id
|
| 34 |
-
DATA_PREP = Task.get_task(task_id=DYNAMIC_TASK_ID)
|
| 35 |
-
|
| 36 |
-
>>>>>>> 20050ad82ebca27a376e15837a7abf79fca23e98
|
| 37 |
# Dataset ID
|
| 38 |
config_objects = DATA_PREP.get_configuration_objects()
|
| 39 |
raw_meta = config_objects["Dataset Metadata"]
|
|
@@ -43,12 +34,7 @@ dataset_id = raw_meta.split("=")[1].strip().replace('"', "")
|
|
| 43 |
subset_clearml = Dataset.get(dataset_id=dataset_id)
|
| 44 |
local_folder = subset_clearml.get_local_copy()
|
| 45 |
|
| 46 |
-
<<<<<<< HEAD
|
| 47 |
subset_indices = np.load(os.path.join(local_folder, "subset_indices.npy"))
|
| 48 |
-
=======
|
| 49 |
-
subset_indices_path = os.path.join(local_folder, "subset_indices.npy")
|
| 50 |
-
subset_indices = np.load(subset_indices_path)
|
| 51 |
-
>>>>>>> 20050ad82ebca27a376e15837a7abf79fca23e98
|
| 52 |
|
| 53 |
# Load Dataset Parameters
|
| 54 |
data_params = DATA_PREP.get_parameters()
|
|
@@ -130,15 +116,8 @@ device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
|
| 130 |
|
| 131 |
# ------- Train the model (on subset for now) -------
|
| 132 |
|
| 133 |
-
<<<<<<< HEAD
|
| 134 |
print("\n--- Starting Model Training on Subset ---")
|
| 135 |
training_metrics = train_model(
|
| 136 |
-
=======
|
| 137 |
-
#When calling this function, the model should be trained on the given dataset
|
| 138 |
-
|
| 139 |
-
print("\n--- Starting Model Training on Subset ---")
|
| 140 |
-
train_model(
|
| 141 |
-
>>>>>>> 20050ad82ebca27a376e15837a7abf79fca23e98
|
| 142 |
model=model,
|
| 143 |
train_loader=subset_loaders['train'],
|
| 144 |
val_loader=subset_loaders['val'],
|
|
@@ -147,7 +126,6 @@ train_model(
|
|
| 147 |
lr=training_config["learning_rate"],
|
| 148 |
save_path=training_config["save_path"],
|
| 149 |
)
|
| 150 |
-
<<<<<<< HEAD
|
| 151 |
|
| 152 |
|
| 153 |
# ----------- Log metrics to ClearML -----------
|
|
@@ -168,6 +146,4 @@ training_logger.report_single_value("best_val_accuracy", training_metrics["best_
|
|
| 168 |
training_task.upload_artifact("best_model", training_config["save_path"])
|
| 169 |
|
| 170 |
print("\nTraining complete.")
|
| 171 |
-
training_task.close()
|
| 172 |
-
=======
|
| 173 |
-
>>>>>>> 20050ad82ebca27a376e15837a7abf79fca23e98
|
|
|
|
| 1 |
import os
|
| 2 |
import numpy as np
|
| 3 |
+
|
| 4 |
from clearml import Task, Dataset
|
| 5 |
from datasets import load_dataset
|
| 6 |
+
from dataPrep.helpers.transforms_loaders import make_dataset_loaders
|
| 7 |
|
| 8 |
+
import torch
|
| 9 |
+
from models.modelOne import modelOne
|
| 10 |
+
from trainingModel.Training import train_model
|
|
|
|
| 11 |
|
|
|
|
|
|
|
|
|
|
| 12 |
|
|
|
|
| 13 |
# -------------- Load Data --------------
|
| 14 |
|
| 15 |
all_tasks = Task.get_tasks(project_name="Small Group Project")
|
|
|
|
| 25 |
DYNAMIC_TASK_ID = latest_task.id
|
| 26 |
DATA_PREP = Task.get_task(task_id=DYNAMIC_TASK_ID)
|
| 27 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 28 |
# Dataset ID
|
| 29 |
config_objects = DATA_PREP.get_configuration_objects()
|
| 30 |
raw_meta = config_objects["Dataset Metadata"]
|
|
|
|
| 34 |
subset_clearml = Dataset.get(dataset_id=dataset_id)
|
| 35 |
local_folder = subset_clearml.get_local_copy()
|
| 36 |
|
|
|
|
| 37 |
subset_indices = np.load(os.path.join(local_folder, "subset_indices.npy"))
|
|
|
|
|
|
|
|
|
|
|
|
|
| 38 |
|
| 39 |
# Load Dataset Parameters
|
| 40 |
data_params = DATA_PREP.get_parameters()
|
|
|
|
| 116 |
|
| 117 |
# ------- Train the model (on subset for now) -------
|
| 118 |
|
|
|
|
| 119 |
print("\n--- Starting Model Training on Subset ---")
|
| 120 |
training_metrics = train_model(
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 121 |
model=model,
|
| 122 |
train_loader=subset_loaders['train'],
|
| 123 |
val_loader=subset_loaders['val'],
|
|
|
|
| 126 |
lr=training_config["learning_rate"],
|
| 127 |
save_path=training_config["save_path"],
|
| 128 |
)
|
|
|
|
| 129 |
|
| 130 |
|
| 131 |
# ----------- Log metrics to ClearML -----------
|
|
|
|
| 146 |
training_task.upload_artifact("best_model", training_config["save_path"])
|
| 147 |
|
| 148 |
print("\nTraining complete.")
|
| 149 |
+
training_task.close()
|
|
|
|
|
|