| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | """PyTorch BERT model.""" |
| |
|
| | from __future__ import absolute_import |
| | from __future__ import division |
| | from __future__ import print_function |
| |
|
| | import os |
| | import copy |
| | import json |
| | import logging |
| | import tarfile |
| | import tempfile |
| | import shutil |
| | import torch |
| | from .file_utils import cached_path |
| |
|
| | logger = logging.getLogger(__name__) |
| |
|
| | class PretrainedConfig(object): |
| |
|
| | pretrained_model_archive_map = {} |
| | config_name = "" |
| | weights_name = "" |
| |
|
| | @classmethod |
| | def get_config(cls, pretrained_model_name, cache_dir, type_vocab_size, state_dict, task_config=None): |
| | archive_file = os.path.join(os.path.dirname(os.path.abspath(__file__)), pretrained_model_name) |
| | if os.path.exists(archive_file) is False: |
| | if pretrained_model_name in cls.pretrained_model_archive_map: |
| | archive_file = cls.pretrained_model_archive_map[pretrained_model_name] |
| | else: |
| | archive_file = pretrained_model_name |
| |
|
| | |
| | try: |
| | resolved_archive_file = cached_path(archive_file, cache_dir=cache_dir) |
| | except FileNotFoundError: |
| | if task_config is None or task_config.local_rank == 0: |
| | logger.error( |
| | "Model name '{}' was not found in model name list. " |
| | "We assumed '{}' was a path or url but couldn't find any file " |
| | "associated to this path or url.".format( |
| | pretrained_model_name, |
| | archive_file)) |
| | return None |
| | if resolved_archive_file == archive_file: |
| | if task_config is None or task_config.local_rank == 0: |
| | logger.info("loading archive file {}".format(archive_file)) |
| | else: |
| | if task_config is None or task_config.local_rank == 0: |
| | logger.info("loading archive file {} from cache at {}".format( |
| | archive_file, resolved_archive_file)) |
| | tempdir = None |
| | if os.path.isdir(resolved_archive_file): |
| | serialization_dir = resolved_archive_file |
| | else: |
| | |
| | tempdir = tempfile.mkdtemp() |
| | if task_config is None or task_config.local_rank == 0: |
| | logger.info("extracting archive file {} to temp dir {}".format( |
| | resolved_archive_file, tempdir)) |
| | with tarfile.open(resolved_archive_file, 'r:gz') as archive: |
| | archive.extractall(tempdir) |
| | serialization_dir = tempdir |
| | |
| | config_file = os.path.join(serialization_dir, cls.config_name) |
| | config = cls.from_json_file(config_file) |
| | config.type_vocab_size = type_vocab_size |
| | if task_config is None or task_config.local_rank == 0: |
| | logger.info("Model config {}".format(config)) |
| |
|
| | if state_dict is None: |
| | weights_path = os.path.join(serialization_dir, cls.weights_name) |
| | if os.path.exists(weights_path): |
| | state_dict = torch.load(weights_path, map_location='cpu') |
| | else: |
| | if task_config is None or task_config.local_rank == 0: |
| | logger.info("Weight doesn't exsits. {}".format(weights_path)) |
| |
|
| | if tempdir: |
| | |
| | shutil.rmtree(tempdir) |
| |
|
| | return config, state_dict |
| |
|
| | @classmethod |
| | def from_dict(cls, json_object): |
| | """Constructs a `BertConfig` from a Python dictionary of parameters.""" |
| | config = cls(vocab_size_or_config_json_file=-1) |
| | for key, value in json_object.items(): |
| | config.__dict__[key] = value |
| | return config |
| |
|
| | @classmethod |
| | def from_json_file(cls, json_file): |
| | """Constructs a `BertConfig` from a json file of parameters.""" |
| | with open(json_file, "r", encoding='utf-8') as reader: |
| | text = reader.read() |
| | return cls.from_dict(json.loads(text)) |
| |
|
| | def __repr__(self): |
| | return str(self.to_json_string()) |
| |
|
| | def to_dict(self): |
| | """Serializes this instance to a Python dictionary.""" |
| | output = copy.deepcopy(self.__dict__) |
| | return output |
| |
|
| | def to_json_string(self): |
| | """Serializes this instance to a JSON string.""" |
| | return json.dumps(self.to_dict(), indent=2, sort_keys=True) + "\n" |