| import math |
| import os |
| import requests |
| from torch.hub import download_url_to_file, get_dir |
| from tqdm import tqdm |
| from urllib.parse import urlparse |
|
|
| from .misc import sizeof_fmt |
|
|
|
|
| def download_file_from_google_drive(file_id, save_path): |
| """Download files from google drive. |
| Ref: |
| https://stackoverflow.com/questions/25010369/wget-curl-large-file-from-google-drive # noqa E501 |
| Args: |
| file_id (str): File id. |
| save_path (str): Save path. |
| """ |
|
|
| session = requests.Session() |
| URL = 'https://docs.google.com/uc?export=download' |
| params = {'id': file_id} |
|
|
| response = session.get(URL, params=params, stream=True) |
| token = get_confirm_token(response) |
| if token: |
| params['confirm'] = token |
| response = session.get(URL, params=params, stream=True) |
|
|
| |
| response_file_size = session.get(URL, params=params, stream=True, headers={'Range': 'bytes=0-2'}) |
| print(response_file_size) |
| if 'Content-Range' in response_file_size.headers: |
| file_size = int(response_file_size.headers['Content-Range'].split('/')[1]) |
| else: |
| file_size = None |
|
|
| save_response_content(response, save_path, file_size) |
|
|
|
|
| def get_confirm_token(response): |
| for key, value in response.cookies.items(): |
| if key.startswith('download_warning'): |
| return value |
| return None |
|
|
|
|
| def save_response_content(response, destination, file_size=None, chunk_size=32768): |
| if file_size is not None: |
| pbar = tqdm(total=math.ceil(file_size / chunk_size), unit='chunk') |
|
|
| readable_file_size = sizeof_fmt(file_size) |
| else: |
| pbar = None |
|
|
| with open(destination, 'wb') as f: |
| downloaded_size = 0 |
| for chunk in response.iter_content(chunk_size): |
| downloaded_size += chunk_size |
| if pbar is not None: |
| pbar.update(1) |
| pbar.set_description(f'Download {sizeof_fmt(downloaded_size)} / {readable_file_size}') |
| if chunk: |
| f.write(chunk) |
| if pbar is not None: |
| pbar.close() |
|
|
|
|
| def load_file_from_url(url, model_dir=None, progress=True, file_name=None): |
| """Load file form http url, will download models if necessary. |
| Ref:https://github.com/1adrianb/face-alignment/blob/master/face_alignment/utils.py |
| Args: |
| url (str): URL to be downloaded. |
| model_dir (str): The path to save the downloaded model. Should be a full path. If None, use pytorch hub_dir. |
| Default: None. |
| progress (bool): Whether to show the download progress. Default: True. |
| file_name (str): The downloaded file name. If None, use the file name in the url. Default: None. |
| Returns: |
| str: The path to the downloaded file. |
| """ |
| if model_dir is None: |
| hub_dir = get_dir() |
| model_dir = os.path.join(hub_dir, 'checkpoints') |
|
|
| os.makedirs(model_dir, exist_ok=True) |
|
|
| parts = urlparse(url) |
| filename = os.path.basename(parts.path) |
| if file_name is not None: |
| filename = file_name |
| cached_file = os.path.abspath(os.path.join(model_dir, filename)) |
| if not os.path.exists(cached_file): |
| print(f'Downloading: "{url}" to {cached_file}\n') |
| download_url_to_file(url, cached_file, hash_prefix=None, progress=progress) |
| return cached_file |