| """ |
| metrics.py |
| |
| Utility classes defining a Metrics container and multiple Trackers to enable model/stage-specific logging to various |
| endpoints (e.g., JSONL local logs, Weights & Biases). |
| """ |
|
|
| import time |
| from collections import defaultdict, deque |
| from pathlib import Path |
| from typing import Any, Dict, Optional, Protocol, Tuple, Union |
|
|
| import jsonlines |
| import numpy as np |
| import torch |
| import wandb |
|
|
| from prismatic.overwatch import initialize_overwatch |
|
|
| |
| overwatch = initialize_overwatch(__name__) |
|
|
|
|
| |
| class Tracker(Protocol): |
| def write_hyperparameters(self) -> None: ... |
|
|
| def write(self, global_step: int, metrics: Dict[str, Union[int, float]]) -> None: ... |
|
|
| def finalize(self) -> None: ... |
|
|
|
|
| |
| class JSONLinesTracker: |
| def __init__(self, run_id: str, run_dir: Path, hparams: Dict[str, Any]) -> None: |
| self.run_id, self.run_dir, self.hparams = run_id, run_dir, hparams |
|
|
| @overwatch.rank_zero_only |
| def write_hyperparameters(self) -> None: |
| with jsonlines.open(self.run_dir / "run-metrics.jsonl", mode="w", sort_keys=True) as js_tracker: |
| js_tracker.write({"run_id": self.run_id, "hparams": self.hparams}) |
|
|
| @overwatch.rank_zero_only |
| def write(self, _: int, metrics: Dict[str, Union[int, float]]) -> None: |
| with jsonlines.open(self.run_dir / f"{self.run_id}.jsonl", mode="a", sort_keys=True) as js_tracker: |
| js_tracker.write(metrics) |
|
|
| def finalize(self) -> None: |
| return |
|
|
|
|
| class WeightsBiasesTracker: |
| def __init__( |
| self, |
| run_id: str, |
| run_dir: Path, |
| hparams: Dict[str, Any], |
| project: str = "prismatic", |
| entity: Optional[str] = None, |
| group: str = "align", |
| ) -> None: |
| self.run_id, self.run_dir, self.hparams = run_id, run_dir, hparams |
|
|
| |
| self.project, self.entity, self.group, self.wandb_dir = project, entity, group, self.run_dir |
|
|
| |
| self.initialize() |
|
|
| @overwatch.rank_zero_only |
| def initialize(self) -> None: |
| wandb.init( |
| name=self.run_id, |
| dir=self.wandb_dir, |
| config=self.hparams, |
| project=self.project, |
| entity=self.entity, |
| group=self.group, |
| ) |
|
|
| @overwatch.rank_zero_only |
| def write_hyperparameters(self) -> None: |
| wandb.config = self.hparams |
|
|
| @overwatch.rank_zero_only |
| def write(self, global_step: int, metrics: Dict[str, Union[int, float]]) -> None: |
| wandb.log(metrics, step=global_step) |
|
|
| @staticmethod |
| def finalize() -> None: |
| if overwatch.is_rank_zero(): |
| wandb.finish() |
|
|
| |
| time.sleep(210) |
|
|
|
|
| |
|
|
|
|
| class Metrics: |
| def __init__( |
| self, |
| active_trackers: Tuple[str, ...], |
| run_id: str, |
| run_dir: Path, |
| hparams: Dict[str, Any], |
| stage: str, |
| wandb_project: str = "prismatic", |
| wandb_entity: Optional[str] = None, |
| grad_accumulation_steps: int = 1, |
| window_size: int = 128, |
| ) -> None: |
| self.run_id, self.run_dir, self.hparams, self.stage = run_id, run_dir, hparams, stage |
|
|
| |
| self.trackers = [] |
| for tracker_type in active_trackers: |
| if tracker_type == "jsonl": |
| tracker = JSONLinesTracker(run_id, run_dir, hparams) |
| elif tracker_type == "wandb": |
| tracker = WeightsBiasesTracker( |
| run_id, run_dir, hparams, project=wandb_project, entity=wandb_entity, group=self.stage |
| ) |
| else: |
| raise ValueError(f"Tracker with type `{tracker_type} is not supported!") |
|
|
| |
| tracker.write_hyperparameters() |
| self.trackers.append(tracker) |
|
|
| |
| self.global_step, self.start_time, self.step_start_time = 0, time.time(), time.time() |
| self.state = { |
| "loss_raw": deque(maxlen=grad_accumulation_steps), |
| "loss": deque(maxlen=window_size), |
| "step_time": deque(maxlen=window_size), |
| "lr": [], |
| } |
|
|
| def log(self, global_step: int, metrics: Dict[str, Union[int, float]]) -> None: |
| for tracker in self.trackers: |
| tracker.write(global_step, metrics) |
|
|
| def get_status(self, loss: Optional[torch.Tensor] = None) -> str: |
| lr = self.state["lr"][-1] if len(self.state["lr"]) > 0 else 0 |
| if loss is None: |
| return f"=>> [Global Step] {self.global_step:06d} =>> LR :: {lr:.6f}" |
|
|
| |
| return f"=>> [Global Step] {self.global_step:06d} =>> LR :: {lr:.6f} -- Loss :: {loss:.4f}" |
|
|
| def commit( |
| self, *, global_step: Optional[int] = None, lr: Optional[float] = None, update_step_time: bool = False, **kwargs |
| ) -> None: |
| """Update all metrics in `self.state` by iterating through special positional arguments & kwargs.""" |
| if global_step is not None: |
| self.global_step = global_step |
|
|
| |
| if not overwatch.is_rank_zero(): |
| return |
|
|
| |
| if lr is not None: |
| self.state["lr"].append(lr) |
|
|
| if update_step_time: |
| self.state["step_time"].append(time.time() - self.step_start_time) |
| self.step_start_time = time.time() |
|
|
| |
| for key, value in kwargs.items(): |
| if key == "loss": |
| loss_val = value.detach() |
| self.state["loss_raw"].append(loss_val) |
| self.state["loss"].append(loss_val) |
| else: |
| self.state[key].append(value.detach()) |
|
|
| @overwatch.rank_zero_only |
| def push(self) -> str: |
| |
| loss_raw = torch.stack(list(self.state["loss_raw"])).mean().item() |
| loss = torch.stack(list(self.state["loss"])).mean().item() |
| step_time, lr = np.mean(list(self.state["step_time"])), self.state["lr"][-1] |
| status = self.get_status(loss) |
|
|
| |
| prefix = self.stage.capitalize() |
| self.log( |
| self.global_step, |
| metrics={ |
| f"{prefix}/Step": self.global_step, |
| f"{prefix}/Loss": loss, |
| f"{prefix}/Loss (Raw)": loss_raw, |
| f"{prefix}/Learning Rate": lr, |
| f"{prefix}/Step Time": step_time, |
| }, |
| ) |
| return status |
|
|
| def finalize(self) -> str: |
| for tracker in self.trackers: |
| tracker.finalize() |
|
|
|
|
| class VLAMetrics: |
| def __init__( |
| self, |
| active_trackers: Tuple[str, ...], |
| run_id: str, |
| run_dir: Path, |
| hparams: Dict[str, Any], |
| wandb_project: str = "openvla", |
| wandb_entity: Optional[str] = "stanford-voltron", |
| grad_accumulation_steps: int = 1, |
| window_size: int = 1, |
| resume_step: Optional[int] = None, |
| resume_epoch: Optional[int] = None, |
| ) -> None: |
| self.run_id, self.run_dir, self.hparams = run_id, run_dir, hparams |
|
|
| |
| self.trackers = [] |
| for tracker_type in active_trackers: |
| if tracker_type == "jsonl": |
| tracker = JSONLinesTracker(run_id, run_dir, hparams) |
| elif tracker_type == "wandb": |
| tracker = WeightsBiasesTracker( |
| run_id, run_dir, hparams, project=wandb_project, entity=wandb_entity, group="vla-train" |
| ) |
| else: |
| raise ValueError(f"Tracker with type `{tracker_type} is not supported!") |
|
|
| |
| tracker.write_hyperparameters() |
| self.trackers.append(tracker) |
|
|
| |
| self.global_step = 0 if resume_step is None else resume_step |
| self.epoch = 0 if resume_epoch is None else resume_epoch |
| self.start_time, self.step_start_time = time.time(), time.time() |
| self.state = { |
| "loss_raw": deque(maxlen=grad_accumulation_steps), |
| "loss": deque(maxlen=window_size), |
| "l1_loss": deque(maxlen=window_size), |
| "action_accuracy": deque(maxlen=window_size), |
| "step_time": deque(maxlen=window_size), |
| "lr": [], |
| } |
|
|
| |
| self.dataset_trackers = defaultdict(lambda: VLAMetrics([], "", "", {})) |
| self.window_size = window_size |
|
|
| def log(self, global_step: int, metrics: Dict[str, Union[int, float]]) -> None: |
| for tracker in self.trackers: |
| tracker.write(global_step, metrics) |
|
|
| def get_status(self, loss: Optional[torch.Tensor] = None) -> str: |
| lr = self.state["lr"][-1] if len(self.state["lr"]) > 0 else 0 |
| if loss is None: |
| return f"=>> [Epoch {self.epoch:03d}] Global Step {self.global_step:06d} =>> LR :: {lr:.6f}" |
|
|
| |
| return f"=>> [Epoch {self.epoch:03d}] Global Step {self.global_step:06d} =>> LR :: {lr:.6f} - Loss :: {loss:.4f}" |
|
|
| def commit( |
| self, |
| *, |
| global_step: Optional[int] = None, |
| epoch: Optional[int] = None, |
| lr: Optional[float] = None, |
| update_step_time: bool = False, |
| **kwargs, |
| ) -> None: |
| """Update all metrics in `self.state` by iterating through special positional arguments & kwargs.""" |
| if global_step is not None: |
| self.global_step = global_step |
|
|
| if epoch is not None: |
| self.epoch = epoch |
|
|
| |
| if not overwatch.is_rank_zero(): |
| return |
|
|
| |
| if lr is not None: |
| self.state["lr"].append(lr) |
|
|
| if update_step_time: |
| self.state["step_time"].append(time.time() - self.step_start_time) |
| self.step_start_time = time.time() |
|
|
| |
| for key, value in kwargs.items(): |
| if key == "loss": |
| loss_val = value.detach() |
| self.state["loss_raw"].append(loss_val) |
| self.state["loss"].append(loss_val) |
| else: |
| self.state[key].append(value.detach()) |
|
|
| def commit_for_dataset(self, dataset_name: str, **kwargs) -> None: |
| self.dataset_trackers[dataset_name].commit(**kwargs) |
|
|
| @overwatch.rank_zero_only |
| def push(self, prefix="VLA Train") -> str: |
| |
| loss_raw = torch.stack(list(self.state["loss_raw"])).mean().item() |
| loss = torch.stack(list(self.state["loss"])).mean().item() |
| l1_loss = torch.stack(list(self.state["l1_loss"])).mean().item() |
| action_accuracy = torch.stack(list(self.state["action_accuracy"])).mean().item() |
| step_time, lr = np.mean(list(self.state["step_time"])), self.state["lr"][-1] |
| status = self.get_status(loss) |
|
|
| |
| dataset_metrics = {} |
| for ds, tracker in self.dataset_trackers.items(): |
| dataset_metrics.update( |
| { |
| f"{ds}/L1 Loss": torch.stack(list(tracker.state["l1_loss"])).mean().item(), |
| f"{ds}/Action Token Accuracy": torch.stack(list(tracker.state["action_accuracy"])).mean().item(), |
| } |
| ) |
|
|
| |
| self.log( |
| self.global_step, |
| metrics={ |
| f"{prefix}/Step": self.global_step, |
| f"{prefix}/Epoch": self.epoch, |
| f"{prefix}/Loss": loss, |
| f"{prefix}/L1 Loss": l1_loss, |
| f"{prefix}/Action Token Accuracy": action_accuracy, |
| f"{prefix}/Loss (Raw)": loss_raw, |
| f"{prefix}/Learning Rate": lr, |
| f"{prefix}/Step Time": step_time, |
| **dataset_metrics, |
| }, |
| ) |
| return status |
|
|
| def finalize(self) -> str: |
| for tracker in self.trackers: |
| tracker.finalize() |
|
|