| | import math
|
| | import os
|
| | import requests
|
| | from torch.hub import download_url_to_file, get_dir
|
| | from tqdm import tqdm
|
| | from urllib.parse import urlparse
|
| |
|
| | from .misc import sizeof_fmt
|
| |
|
| |
|
| | def download_file_from_google_drive(file_id, save_path):
|
| | """Download files from google drive.
|
| | Ref:
|
| | https://stackoverflow.com/questions/25010369/wget-curl-large-file-from-google-drive # noqa E501
|
| | Args:
|
| | file_id (str): File id.
|
| | save_path (str): Save path.
|
| | """
|
| |
|
| | session = requests.Session()
|
| | URL = 'https://docs.google.com/uc?export=download'
|
| | params = {'id': file_id}
|
| |
|
| | response = session.get(URL, params=params, stream=True)
|
| | token = get_confirm_token(response)
|
| | if token:
|
| | params['confirm'] = token
|
| | response = session.get(URL, params=params, stream=True)
|
| |
|
| |
|
| | response_file_size = session.get(URL, params=params, stream=True, headers={'Range': 'bytes=0-2'})
|
| | print(response_file_size)
|
| | if 'Content-Range' in response_file_size.headers:
|
| | file_size = int(response_file_size.headers['Content-Range'].split('/')[1])
|
| | else:
|
| | file_size = None
|
| |
|
| | save_response_content(response, save_path, file_size)
|
| |
|
| |
|
| | def get_confirm_token(response):
|
| | for key, value in response.cookies.items():
|
| | if key.startswith('download_warning'):
|
| | return value
|
| | return None
|
| |
|
| |
|
| | def save_response_content(response, destination, file_size=None, chunk_size=32768):
|
| | if file_size is not None:
|
| | pbar = tqdm(total=math.ceil(file_size / chunk_size), unit='chunk')
|
| |
|
| | readable_file_size = sizeof_fmt(file_size)
|
| | else:
|
| | pbar = None
|
| |
|
| | with open(destination, 'wb') as f:
|
| | downloaded_size = 0
|
| | for chunk in response.iter_content(chunk_size):
|
| | downloaded_size += chunk_size
|
| | if pbar is not None:
|
| | pbar.update(1)
|
| | pbar.set_description(f'Download {sizeof_fmt(downloaded_size)} / {readable_file_size}')
|
| | if chunk:
|
| | f.write(chunk)
|
| | if pbar is not None:
|
| | pbar.close()
|
| |
|
| |
|
| | def load_file_from_url(url, model_dir=None, progress=True, file_name=None):
|
| | """Load file form http url, will download models if necessary.
|
| | Ref:https://github.com/1adrianb/face-alignment/blob/master/face_alignment/utils.py
|
| | Args:
|
| | url (str): URL to be downloaded.
|
| | model_dir (str): The path to save the downloaded model. Should be a full path. If None, use pytorch hub_dir.
|
| | Default: None.
|
| | progress (bool): Whether to show the download progress. Default: True.
|
| | file_name (str): The downloaded file name. If None, use the file name in the url. Default: None.
|
| | Returns:
|
| | str: The path to the downloaded file.
|
| | """
|
| | if model_dir is None:
|
| | hub_dir = get_dir()
|
| | model_dir = os.path.join(hub_dir, 'checkpoints')
|
| |
|
| | os.makedirs(model_dir, exist_ok=True)
|
| |
|
| | parts = urlparse(url)
|
| | filename = os.path.basename(parts.path)
|
| | if file_name is not None:
|
| | filename = file_name
|
| | cached_file = os.path.abspath(os.path.join(model_dir, filename))
|
| | if not os.path.exists(cached_file):
|
| | print(f'Downloading: "{url}" to {cached_file}\n')
|
| | download_url_to_file(url, cached_file, hash_prefix=None, progress=progress)
|
| | return cached_file |