|
|
import os |
|
|
from functools import partial |
|
|
|
|
|
import torch |
|
|
from safetensors.torch import load_file as safe_load_file |
|
|
from transformers.utils import ( |
|
|
SAFE_WEIGHTS_INDEX_NAME, |
|
|
SAFE_WEIGHTS_NAME, |
|
|
WEIGHTS_INDEX_NAME, |
|
|
WEIGHTS_NAME, |
|
|
) |
|
|
from transformers.utils.hub import cached_file, get_checkpoint_shard_files |
|
|
|
|
|
|
|
|
def state_dict_from_pretrained(model_name, device=None, dtype=None): |
|
|
|
|
|
mapped_device = "cpu" if dtype not in [torch.float32, None] else device |
|
|
is_sharded = False |
|
|
load_safe = False |
|
|
resolved_archive_file = None |
|
|
|
|
|
weights_path = os.path.join(model_name, WEIGHTS_NAME) |
|
|
weights_index_path = os.path.join(model_name, WEIGHTS_INDEX_NAME) |
|
|
safe_weights_path = os.path.join(model_name, SAFE_WEIGHTS_NAME) |
|
|
safe_weights_index_path = os.path.join(model_name, SAFE_WEIGHTS_INDEX_NAME) |
|
|
|
|
|
if os.path.isfile(weights_path): |
|
|
resolved_archive_file = cached_file( |
|
|
model_name, WEIGHTS_NAME, _raise_exceptions_for_missing_entries=False |
|
|
) |
|
|
elif os.path.isfile(weights_index_path): |
|
|
resolved_archive_file = cached_file( |
|
|
model_name, WEIGHTS_INDEX_NAME, _raise_exceptions_for_missing_entries=False |
|
|
) |
|
|
is_sharded = True |
|
|
elif os.path.isfile(safe_weights_path): |
|
|
resolved_archive_file = cached_file( |
|
|
model_name, SAFE_WEIGHTS_NAME, _raise_exceptions_for_missing_entries=False |
|
|
) |
|
|
load_safe = True |
|
|
elif os.path.isfile(safe_weights_index_path): |
|
|
resolved_archive_file = cached_file( |
|
|
model_name, SAFE_WEIGHTS_INDEX_NAME, _raise_exceptions_for_missing_entries=False |
|
|
) |
|
|
is_sharded = True |
|
|
load_safe = True |
|
|
else: |
|
|
resolved_archive_file = cached_file(model_name, WEIGHTS_NAME, |
|
|
_raise_exceptions_for_missing_entries=False) |
|
|
if resolved_archive_file is None: |
|
|
resolved_archive_file = cached_file(model_name, WEIGHTS_INDEX_NAME, |
|
|
_raise_exceptions_for_missing_entries=False) |
|
|
if resolved_archive_file is not None: |
|
|
is_sharded = True |
|
|
|
|
|
if resolved_archive_file is None: |
|
|
raise EnvironmentError(f"Model name {model_name} was not found.") |
|
|
|
|
|
if load_safe: |
|
|
loader = partial(safe_load_file, device=mapped_device) |
|
|
else: |
|
|
loader = partial(torch.load, map_location=mapped_device) |
|
|
|
|
|
if is_sharded: |
|
|
|
|
|
|
|
|
resolved_archive_file, sharded_metadata = get_checkpoint_shard_files( |
|
|
model_name, resolved_archive_file |
|
|
) |
|
|
state_dict = {} |
|
|
for sharded_file in resolved_archive_file: |
|
|
state_dict.update(loader(sharded_file)) |
|
|
else: |
|
|
state_dict = loader(resolved_archive_file) |
|
|
|
|
|
if dtype is not None: |
|
|
state_dict = {k: v.to(dtype=dtype) for k, v in state_dict.items()} |
|
|
state_dict = {k: v.to(device=device) for k, v in state_dict.items()} |
|
|
return state_dict |
|
|
|