Spaces:
Sleeping
Sleeping
Merge pull request #6 from K23064919/ops/clone-models
Browse files- requirements.txt +12 -22
- trainingModel/run_training.py +6 -30
requirements.txt
CHANGED
|
@@ -1,29 +1,19 @@
|
|
| 1 |
|
| 2 |
# Core dependencies
|
| 3 |
-
torch
|
| 4 |
-
torchvision
|
| 5 |
-
|
| 6 |
-
numpy
|
| 7 |
-
Pillow
|
|
|
|
| 8 |
|
| 9 |
-
# For model deployment and tracking
|
| 10 |
-
huggingface-hub>=0.19.0
|
| 11 |
-
clearml>=1.14.0
|
| 12 |
-
|
| 13 |
-
# Optional: for advanced features
|
| 14 |
-
datasets>=2.14.0 # For loading PlantVillage dataset from HuggingFace
|
| 15 |
-
# -- Data prep requirements --
|
| 16 |
# Data Handling & Analysis
|
| 17 |
-
|
| 18 |
-
|
| 19 |
-
datasets
|
| 20 |
|
| 21 |
# Visualization
|
| 22 |
-
matplotlib
|
| 23 |
|
| 24 |
-
#
|
| 25 |
-
|
| 26 |
-
|
| 27 |
-
|
| 28 |
-
# Experiment Tracking
|
| 29 |
-
clearml
|
|
|
|
| 1 |
|
| 2 |
# Core dependencies
|
| 3 |
+
torch==2.2.2
|
| 4 |
+
torchvision==0.17.2
|
| 5 |
+
torcheval==0.0.7
|
| 6 |
+
numpy==1.26.4
|
| 7 |
+
Pillow==10.3.0
|
| 8 |
+
gradio==4.19.0
|
| 9 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 10 |
# Data Handling & Analysis
|
| 11 |
+
pandas==2.2.2
|
| 12 |
+
datasets==2.18.0
|
|
|
|
| 13 |
|
| 14 |
# Visualization
|
| 15 |
+
matplotlib==3.8.4
|
| 16 |
|
| 17 |
+
# For model deployment and tracking
|
| 18 |
+
huggingface-hub==0.23.0
|
| 19 |
+
clearml==2.0.2
|
|
|
|
|
|
|
|
|
trainingModel/run_training.py
CHANGED
|
@@ -1,18 +1,15 @@
|
|
| 1 |
import os
|
| 2 |
import numpy as np
|
|
|
|
| 3 |
from clearml import Task, Dataset
|
| 4 |
from datasets import load_dataset
|
|
|
|
| 5 |
|
| 6 |
-
|
| 7 |
-
|
| 8 |
-
|
| 9 |
-
raise RuntimeError("No tasks found in project 'Small Group Project'")
|
| 10 |
|
| 11 |
-
dp_tasks = [t for t in all_tasks if t.name == "Data Preparation"]
|
| 12 |
-
if not dp_tasks:
|
| 13 |
-
raise RuntimeError("No 'Data Preparation' tasks found in this project!")
|
| 14 |
|
| 15 |
-
<<<<<<< HEAD
|
| 16 |
# -------------- Load Data --------------
|
| 17 |
|
| 18 |
all_tasks = Task.get_tasks(project_name="Small Group Project")
|
|
@@ -28,12 +25,6 @@ latest_task = max(dp_tasks, key=lambda t: t.id)
|
|
| 28 |
DYNAMIC_TASK_ID = latest_task.id
|
| 29 |
DATA_PREP = Task.get_task(task_id=DYNAMIC_TASK_ID)
|
| 30 |
|
| 31 |
-
=======
|
| 32 |
-
latest_task = max(dp_tasks, key=lambda t: t.id)
|
| 33 |
-
DYNAMIC_TASK_ID = latest_task.id
|
| 34 |
-
DATA_PREP = Task.get_task(task_id=DYNAMIC_TASK_ID)
|
| 35 |
-
|
| 36 |
-
>>>>>>> 20050ad82ebca27a376e15837a7abf79fca23e98
|
| 37 |
# Dataset ID
|
| 38 |
config_objects = DATA_PREP.get_configuration_objects()
|
| 39 |
raw_meta = config_objects["Dataset Metadata"]
|
|
@@ -43,12 +34,7 @@ dataset_id = raw_meta.split("=")[1].strip().replace('"', "")
|
|
| 43 |
subset_clearml = Dataset.get(dataset_id=dataset_id)
|
| 44 |
local_folder = subset_clearml.get_local_copy()
|
| 45 |
|
| 46 |
-
<<<<<<< HEAD
|
| 47 |
subset_indices = np.load(os.path.join(local_folder, "subset_indices.npy"))
|
| 48 |
-
=======
|
| 49 |
-
subset_indices_path = os.path.join(local_folder, "subset_indices.npy")
|
| 50 |
-
subset_indices = np.load(subset_indices_path)
|
| 51 |
-
>>>>>>> 20050ad82ebca27a376e15837a7abf79fca23e98
|
| 52 |
|
| 53 |
# Load Dataset Parameters
|
| 54 |
data_params = DATA_PREP.get_parameters()
|
|
@@ -130,15 +116,8 @@ device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
|
| 130 |
|
| 131 |
# ------- Train the model (on subset for now) -------
|
| 132 |
|
| 133 |
-
<<<<<<< HEAD
|
| 134 |
print("\n--- Starting Model Training on Subset ---")
|
| 135 |
training_metrics = train_model(
|
| 136 |
-
=======
|
| 137 |
-
#When calling this function, the model should be trained on the given dataset
|
| 138 |
-
|
| 139 |
-
print("\n--- Starting Model Training on Subset ---")
|
| 140 |
-
train_model(
|
| 141 |
-
>>>>>>> 20050ad82ebca27a376e15837a7abf79fca23e98
|
| 142 |
model=model,
|
| 143 |
train_loader=subset_loaders['train'],
|
| 144 |
val_loader=subset_loaders['val'],
|
|
@@ -147,7 +126,6 @@ train_model(
|
|
| 147 |
lr=training_config["learning_rate"],
|
| 148 |
save_path=training_config["save_path"],
|
| 149 |
)
|
| 150 |
-
<<<<<<< HEAD
|
| 151 |
|
| 152 |
|
| 153 |
# ----------- Log metrics to ClearML -----------
|
|
@@ -168,6 +146,4 @@ training_logger.report_single_value("best_val_accuracy", training_metrics["best_
|
|
| 168 |
training_task.upload_artifact("best_model", training_config["save_path"])
|
| 169 |
|
| 170 |
print("\nTraining complete.")
|
| 171 |
-
training_task.close()
|
| 172 |
-
=======
|
| 173 |
-
>>>>>>> 20050ad82ebca27a376e15837a7abf79fca23e98
|
|
|
|
| 1 |
import os
|
| 2 |
import numpy as np
|
| 3 |
+
|
| 4 |
from clearml import Task, Dataset
|
| 5 |
from datasets import load_dataset
|
| 6 |
+
from dataPrep.helpers.transforms_loaders import make_dataset_loaders
|
| 7 |
|
| 8 |
+
import torch
|
| 9 |
+
from models.modelOne import modelOne
|
| 10 |
+
from trainingModel.Training import train_model
|
|
|
|
| 11 |
|
|
|
|
|
|
|
|
|
|
| 12 |
|
|
|
|
| 13 |
# -------------- Load Data --------------
|
| 14 |
|
| 15 |
all_tasks = Task.get_tasks(project_name="Small Group Project")
|
|
|
|
| 25 |
DYNAMIC_TASK_ID = latest_task.id
|
| 26 |
DATA_PREP = Task.get_task(task_id=DYNAMIC_TASK_ID)
|
| 27 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 28 |
# Dataset ID
|
| 29 |
config_objects = DATA_PREP.get_configuration_objects()
|
| 30 |
raw_meta = config_objects["Dataset Metadata"]
|
|
|
|
| 34 |
subset_clearml = Dataset.get(dataset_id=dataset_id)
|
| 35 |
local_folder = subset_clearml.get_local_copy()
|
| 36 |
|
|
|
|
| 37 |
subset_indices = np.load(os.path.join(local_folder, "subset_indices.npy"))
|
|
|
|
|
|
|
|
|
|
|
|
|
| 38 |
|
| 39 |
# Load Dataset Parameters
|
| 40 |
data_params = DATA_PREP.get_parameters()
|
|
|
|
| 116 |
|
| 117 |
# ------- Train the model (on subset for now) -------
|
| 118 |
|
|
|
|
| 119 |
print("\n--- Starting Model Training on Subset ---")
|
| 120 |
training_metrics = train_model(
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 121 |
model=model,
|
| 122 |
train_loader=subset_loaders['train'],
|
| 123 |
val_loader=subset_loaders['val'],
|
|
|
|
| 126 |
lr=training_config["learning_rate"],
|
| 127 |
save_path=training_config["save_path"],
|
| 128 |
)
|
|
|
|
| 129 |
|
| 130 |
|
| 131 |
# ----------- Log metrics to ClearML -----------
|
|
|
|
| 146 |
training_task.upload_artifact("best_model", training_config["save_path"])
|
| 147 |
|
| 148 |
print("\nTraining complete.")
|
| 149 |
+
training_task.close()
|
|
|
|
|
|