|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
from __future__ import annotations |
|
|
|
|
|
import copy |
|
|
import os |
|
|
import pickle |
|
|
import tempfile |
|
|
from types import ModuleType |
|
|
from typing import Any, Hashable |
|
|
|
|
|
import torch |
|
|
from torch.serialization import DEFAULT_PROTOCOL |
|
|
|
|
|
from monai.config.type_definitions import PathLike |
|
|
|
|
|
__all__ = ["StateCacher"] |
|
|
|
|
|
|
|
|
class StateCacher: |
|
|
"""Class to cache and retrieve the state of an object. |
|
|
|
|
|
Objects can either be stored in memory or on disk. If stored on disk, they can be |
|
|
stored in a given directory, or alternatively a temporary location will be used. |
|
|
|
|
|
If necessary/possible, restored objects will be returned to their original device. |
|
|
|
|
|
Example: |
|
|
|
|
|
>>> state_cacher = StateCacher(memory_cache, cache_dir=cache_dir) |
|
|
>>> state_cacher.store("model", model.state_dict()) |
|
|
>>> model.load_state_dict(state_cacher.retrieve("model")) |
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
in_memory: bool, |
|
|
cache_dir: PathLike | None = None, |
|
|
allow_overwrite: bool = True, |
|
|
pickle_module: ModuleType = pickle, |
|
|
pickle_protocol: int = DEFAULT_PROTOCOL, |
|
|
) -> None: |
|
|
"""Constructor. |
|
|
|
|
|
Args: |
|
|
in_memory: boolean to determine if the object will be cached in memory or on |
|
|
disk. |
|
|
cache_dir: directory for data to be cached if `in_memory==False`. Defaults |
|
|
to using a temporary directory. Any created files will be deleted during |
|
|
the `StateCacher`'s destructor. |
|
|
allow_overwrite: allow the cache to be overwritten. If set to `False`, an |
|
|
error will be thrown if a matching already exists in the list of cached |
|
|
objects. |
|
|
pickle_module: module used for pickling metadata and objects, default to `pickle`. |
|
|
this arg is used by `torch.save`, for more details, please check: |
|
|
https://pytorch.org/docs/stable/generated/torch.save.html#torch.save. |
|
|
pickle_protocol: can be specified to override the default protocol, default to `2`. |
|
|
this arg is used by `torch.save`, for more details, please check: |
|
|
https://pytorch.org/docs/stable/generated/torch.save.html#torch.save. |
|
|
|
|
|
""" |
|
|
self.in_memory = in_memory |
|
|
self.cache_dir = tempfile.gettempdir() if cache_dir is None else cache_dir |
|
|
if not os.path.isdir(self.cache_dir): |
|
|
raise ValueError("Given `cache_dir` is not a valid directory.") |
|
|
|
|
|
self.allow_overwrite = allow_overwrite |
|
|
self.pickle_module = pickle_module |
|
|
self.pickle_protocol = pickle_protocol |
|
|
self.cached: dict = {} |
|
|
|
|
|
def store( |
|
|
self, key: Hashable, data_obj: Any, pickle_module: ModuleType | None = None, pickle_protocol: int | None = None |
|
|
) -> None: |
|
|
""" |
|
|
Store a given object with the given key name. |
|
|
|
|
|
Args: |
|
|
key: key of the data object to store. |
|
|
data_obj: data object to store. |
|
|
pickle_module: module used for pickling metadata and objects, default to `self.pickle_module`. |
|
|
this arg is used by `torch.save`, for more details, please check: |
|
|
https://pytorch.org/docs/stable/generated/torch.save.html#torch.save. |
|
|
pickle_protocol: can be specified to override the default protocol, default to `self.pickle_protocol`. |
|
|
this arg is used by `torch.save`, for more details, please check: |
|
|
https://pytorch.org/docs/stable/generated/torch.save.html#torch.save. |
|
|
|
|
|
""" |
|
|
if key in self.cached and not self.allow_overwrite: |
|
|
raise RuntimeError("Cached key already exists and overwriting is disabled.") |
|
|
if self.in_memory: |
|
|
self.cached.update({key: {"obj": copy.deepcopy(data_obj)}}) |
|
|
else: |
|
|
fn = os.path.join(self.cache_dir, f"state_{key}_{id(self)}.pt") |
|
|
self.cached.update({key: {"obj": fn}}) |
|
|
torch.save( |
|
|
obj=data_obj, |
|
|
f=fn, |
|
|
pickle_module=self.pickle_module if pickle_module is None else pickle_module, |
|
|
pickle_protocol=self.pickle_protocol if pickle_protocol is None else pickle_protocol, |
|
|
) |
|
|
|
|
|
if hasattr(data_obj, "device"): |
|
|
self.cached[key]["device"] = data_obj.device |
|
|
|
|
|
def retrieve(self, key: Hashable) -> Any: |
|
|
"""Retrieve the object stored under a given key name.""" |
|
|
if key not in self.cached: |
|
|
raise KeyError(f"Target {key} was not cached.") |
|
|
|
|
|
if self.in_memory: |
|
|
return self.cached[key]["obj"] |
|
|
|
|
|
fn = self.cached[key]["obj"] |
|
|
if not os.path.exists(fn): |
|
|
raise RuntimeError(f"Failed to load state in {fn}. File doesn't exist anymore.") |
|
|
data_obj = torch.load(fn, map_location=lambda storage, location: storage) |
|
|
|
|
|
if "device" in self.cached[key]: |
|
|
data_obj = data_obj.to(self.cached[key]["device"]) |
|
|
return data_obj |
|
|
|
|
|
def __del__(self): |
|
|
"""If necessary, delete any cached files existing in `cache_dir`.""" |
|
|
if not self.in_memory: |
|
|
for k in self.cached: |
|
|
if os.path.exists(self.cached[k]["obj"]): |
|
|
os.remove(self.cached[k]["obj"]) |
|
|
|