| # Tracking | |
| There are a large number of experiment tracking API's available, however getting them all to work with in a multi-processing environment can oftentimes be complex. | |
| 🤗 Accelerate provides a general tracking API that can be used to log useful items during your script through [`Accelerator.log`] | |
| ## Integrated Trackers | |
| Currently `Accelerate` supports seven trackers out-of-the-box: | |
| - TensorBoard | |
| - WandB | |
| - CometML | |
| - Aim | |
| - MLFlow | |
| - ClearML | |
| - DVCLive | |
| To use any of them, pass in the selected type(s) to the `log_with` parameter in [`Accelerate`]: | |
| ```python | |
| from accelerate import Accelerator | |
| from accelerate.utils import LoggerType | |
| accelerator = Accelerator(log_with="all") # For all available trackers in the environment | |
| accelerator = Accelerator(log_with="wandb") | |
| accelerator = Accelerator(log_with=["wandb", LoggerType.TENSORBOARD]) | |
| ``` | |
| At the start of your experiment [`Accelerator.init_trackers`] should be used to setup your project, and potentially add any experiment hyperparameters to be logged: | |
| ```python | |
| hps = {"num_iterations": 5, "learning_rate": 1e-2} | |
| accelerator.init_trackers("my_project", config=hps) | |
| ``` | |
| When you are ready to log any data, [`Accelerator.log`] should be used. | |
| A `step` can also be passed in to correlate the data with a particular step in the training loop. | |
| ```python | |
| accelerator.log({"train_loss": 1.12, "valid_loss": 0.8}, step=1) | |
| ``` | |
| Once you've finished training, make sure to run [`Accelerator.end_training`] so that all the trackers can run their finish functionalities if they have any. | |
| ```python | |
| accelerator.end_training() | |
| ``` | |
| A full example is below: | |
| ```python | |
| from accelerate import Accelerator | |
| accelerator = Accelerator(log_with="all") | |
| config = { | |
| "num_iterations": 5, | |
| "learning_rate": 1e-2, | |
| "loss_function": str(my_loss_function), | |
| } | |
| accelerator.init_trackers("example_project", config=config) | |
| my_model, my_optimizer, my_training_dataloader = accelerate.prepare(my_model, my_optimizer, my_training_dataloader) | |
| device = accelerator.device | |
| my_model.to(device) | |
| for iteration in config["num_iterations"]: | |
| for step, batch in my_training_dataloader: | |
| my_optimizer.zero_grad() | |
| inputs, targets = batch | |
| inputs = inputs.to(device) | |
| targets = targets.to(device) | |
| outputs = my_model(inputs) | |
| loss = my_loss_function(outputs, targets) | |
| accelerator.backward(loss) | |
| my_optimizer.step() | |
| accelerator.log({"training_loss": loss}, step=step) | |
| accelerator.end_training() | |
| ``` | |
| If a tracker requires a directory to save data to, such as `TensorBoard`, then pass the directory path to `project_dir`. The `project_dir` parameter is useful | |
| when there are other configurations to be combined with in the [`~utils.ProjectConfiguration`] data class. For example, you can save the TensorBoard data to `project_dir` and everything else can be logged in the `logging_dir` parameter of [`~utils.ProjectConfiguration`: | |
| ```python | |
| accelerator = Accelerator(log_with="tensorboard", project_dir=".") | |
| # use with ProjectConfiguration | |
| config = ProjectConfiguration(project_dir=".", logging_dir="another/directory") | |
| accelerator = Accelerator(log_with="tensorboard", project_config=config) | |
| ``` | |
| ## Implementing Custom Trackers | |
| To implement a new tracker to be used in `Accelerator`, a new one can be made through implementing the [`GeneralTracker`] class. | |
| Every tracker must implement three functions and have three properties: | |
| - `__init__`: | |
| - Should store a `run_name` and initialize the tracker API of the integrated library. | |
| - If a tracker stores their data locally (such as TensorBoard), a `logging_dir` parameter can be added. | |
| - `store_init_configuration`: | |
| - Should take in a `values` dictionary and store them as a one-time experiment configuration | |
| - `log`: | |
| - Should take in a `values` dictionary and a `step`, and should log them to the run | |
| - `name` (`str`): | |
| - A unique string name for the tracker, such as `"wandb"` for the wandb tracker. | |
| - This will be used for interacting with this tracker specifically | |
| - `requires_logging_directory` (`bool`): | |
| - Whether a `logging_dir` is needed for this particular tracker and if it uses one. | |
| - `tracker`: | |
| - This should be implemented as a `@property` function | |
| - Should return the internal tracking mechanism the library uses, such as the `run` object for `wandb`. | |
| Each method should also utilize the [`state.PartialState`] class if the logger should only be executed on the main process for instance. | |
| A brief example can be seen below with an integration with Weights and Biases, containing only the relevant information and logging just on | |
| the main process: | |
| ```python | |
| from accelerate.tracking import GeneralTracker, on_main_process | |
| from typing import Optional | |
| import wandb | |
| class MyCustomTracker(GeneralTracker): | |
| name = "wandb" | |
| requires_logging_directory = False | |
| @on_main_process | |
| def __init__(self, run_name: str): | |
| self.run_name = run_name | |
| run = wandb.init(self.run_name) | |
| @property | |
| def tracker(self): | |
| return self.run.run | |
| @on_main_process | |
| def store_init_configuration(self, values: dict): | |
| wandb.config(values) | |
| @on_main_process | |
| def log(self, values: dict, step: Optional[int] = None): | |
| wandb.log(values, step=step) | |
| ``` | |
| When you are ready to build your `Accelerator` object, pass in an **instance** of your tracker to [`Accelerator.log_with`] to have it automatically | |
| be used with the API: | |
| ```python | |
| tracker = MyCustomTracker("some_run_name") | |
| accelerator = Accelerator(log_with=tracker) | |
| ``` | |
| These also can be mixed with existing trackers, including with `"all"`: | |
| ```python | |
| tracker = MyCustomTracker("some_run_name") | |
| accelerator = Accelerator(log_with=[tracker, "all"]) | |
| ``` | |
| ## Accessing the internal tracker | |
| If some custom interactions with a tracker might be wanted directly, you can quickly access one using the | |
| [`Accelerator.get_tracker`] method. Just pass in the string corresponding to a tracker's `.name` attribute | |
| and it will return that tracker on the main process. | |
| This example shows doing so with wandb: | |
| ```python | |
| wandb_tracker = accelerator.get_tracker("wandb") | |
| ``` | |
| From there you can interact with `wandb`'s `run` object like normal: | |
| ```python | |
| wandb_run.log_artifact(some_artifact_to_log) | |
| ``` | |
| <Tip> | |
| Trackers built in Accelerate will automatically execute on the correct process, | |
| so if a tracker is only meant to be ran on the main process it will do so | |
| automatically. | |
| </Tip> | |
| If you want to truly remove Accelerate's wrapping entirely, you can | |
| achieve the same outcome with: | |
| ```python | |
| wandb_tracker = accelerator.get_tracker("wandb", unwrap=True) | |
| with accelerator.on_main_process: | |
| wandb_tracker.log_artifact(some_artifact_to_log) | |
| ``` | |
| ## When a wrapper cannot work | |
| If a library has an API that does not follow a strict `.log` with an overall dictionary such as Neptune.AI, logging can be done manually under an `if accelerator.is_main_process` statement: | |
| ```diff | |
| from accelerate import Accelerator | |
| + import neptune.new as neptune | |
| accelerator = Accelerator() | |
| + run = neptune.init(...) | |
| my_model, my_optimizer, my_training_dataloader = accelerate.prepare(my_model, my_optimizer, my_training_dataloader) | |
| device = accelerator.device | |
| my_model.to(device) | |
| for iteration in config["num_iterations"]: | |
| for batch in my_training_dataloader: | |
| my_optimizer.zero_grad() | |
| inputs, targets = batch | |
| inputs = inputs.to(device) | |
| targets = targets.to(device) | |
| outputs = my_model(inputs) | |
| loss = my_loss_function(outputs, targets) | |
| total_loss += loss | |
| accelerator.backward(loss) | |
| my_optimizer.step() | |
| + if accelerator.is_main_process: | |
| + run["logs/training/batch/loss"].log(loss) | |
| ``` | |