repo
stringlengths
7
90
file_url
stringlengths
81
315
file_path
stringlengths
4
228
content
stringlengths
0
32.8k
language
stringclasses
1 value
license
stringclasses
7 values
commit_sha
stringlengths
40
40
retrieved_at
stringdate
2026-01-04 14:38:15
2026-01-05 02:33:18
truncated
bool
2 classes
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/extractor/youtube/pot/__init__.py
yt_dlp/extractor/youtube/pot/__init__.py
# Trigger import of built-in providers from ._builtin.memory_cache import MemoryLRUPCP as _MemoryLRUPCP # noqa: F401 from ._builtin.webpo_cachespec import WebPoPCSP as _WebPoPCSP # noqa: F401
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/extractor/youtube/pot/_provider.py
yt_dlp/extractor/youtube/pot/_provider.py
from __future__ import annotations import abc import enum import functools from yt_dlp.extractor.common import InfoExtractor from yt_dlp.utils import NO_DEFAULT, bug_reports_message, classproperty, traverse_obj from yt_dlp.version import __version__ # xxx: these could be generalized outside YoutubeIE eventually class IEContentProviderLogger(abc.ABC): class LogLevel(enum.IntEnum): TRACE = 0 DEBUG = 10 INFO = 20 WARNING = 30 ERROR = 40 @classmethod def _missing_(cls, value): if isinstance(value, str): value = value.upper() if value in dir(cls): return cls[value] return cls.INFO log_level = LogLevel.INFO @abc.abstractmethod def trace(self, message: str): pass @abc.abstractmethod def debug(self, message: str, *, once=False): pass @abc.abstractmethod def info(self, message: str): pass @abc.abstractmethod def warning(self, message: str, *, once=False): pass @abc.abstractmethod def error(self, message: str, cause=None): pass class IEContentProviderError(Exception): def __init__(self, msg=None, expected=False): super().__init__(msg) self.expected = expected class IEContentProvider(abc.ABC): PROVIDER_VERSION: str = '0.0.0' BUG_REPORT_LOCATION: str = '(developer has not provided a bug report location)' def __init__( self, ie: InfoExtractor, logger: IEContentProviderLogger, settings: dict[str, list[str]], *_, **__, ): self.ie = ie self.settings = settings or {} self.logger = logger super().__init__() @classmethod def __init_subclass__(cls, *, suffix=None, **kwargs): if suffix: cls._PROVIDER_KEY_SUFFIX = suffix return super().__init_subclass__(**kwargs) @classproperty def PROVIDER_NAME(cls) -> str: return cls.__name__[:-len(cls._PROVIDER_KEY_SUFFIX)] @classproperty def BUG_REPORT_MESSAGE(cls): return f'please report this issue to the provider developer at {cls.BUG_REPORT_LOCATION} .' @classproperty def PROVIDER_KEY(cls) -> str: assert hasattr(cls, '_PROVIDER_KEY_SUFFIX'), 'Content Provider implementation must define a suffix for the provider key' assert cls.__name__.endswith(cls._PROVIDER_KEY_SUFFIX), f'Class name must end with "{cls._PROVIDER_KEY_SUFFIX}"' return cls.__name__[:-len(cls._PROVIDER_KEY_SUFFIX)] @abc.abstractmethod def is_available(self) -> bool: """ Check if the provider is available (e.g. all required dependencies are available) This is used to determine if the provider should be used and to provide debug information. IMPORTANT: This method should not make any network requests or perform any expensive operations. It is called multiple times. """ raise NotImplementedError def close(self): # noqa: B027 pass def _configuration_arg(self, key, default=NO_DEFAULT, *, casesense=False): """ @returns A list of values for the setting given by "key" or "default" if no such key is present @param default The default value to return when the key is not present (default: []) @param casesense When false, the values are converted to lower case """ return configuration_arg(self.settings, key, default=default, casesense=casesense) class BuiltinIEContentProvider(IEContentProvider, abc.ABC): PROVIDER_VERSION = __version__ BUG_REPORT_MESSAGE = bug_reports_message(before='') def configuration_arg(config, key, default=NO_DEFAULT, *, casesense=False): """ @returns A list of values for the setting given by "key" or "default" if no such key is present @param config The configuration dictionary @param default The default value to return when the key is not present (default: []) @param casesense When false, the values are converted to lower case """ val = traverse_obj(config, key) if val is None: return [] if default is NO_DEFAULT else default return list(val) if casesense else [x.lower() for x in val] def register_provider_generic( provider, base_class, registry, ): """Generic function to register a provider class""" assert issubclass(provider, base_class), f'{provider} must be a subclass of {base_class.__name__}' assert provider.PROVIDER_KEY not in registry, f'{base_class.__name__} {provider.PROVIDER_KEY} already registered' registry[provider.PROVIDER_KEY] = provider return provider def register_preference_generic( base_class, registry, *providers, ): """Generic function to register a preference for a provider""" assert all(issubclass(provider, base_class) for provider in providers) def outer(preference): @functools.wraps(preference) def inner(provider, *args, **kwargs): if not providers or isinstance(provider, providers): return preference(provider, *args, **kwargs) return 0 registry.add(inner) return preference return outer
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/extractor/youtube/pot/cache.py
yt_dlp/extractor/youtube/pot/cache.py
"""PUBLIC API""" from __future__ import annotations import abc import dataclasses import enum import typing from yt_dlp.extractor.youtube.pot._provider import ( IEContentProvider, IEContentProviderError, register_preference_generic, register_provider_generic, ) from yt_dlp.extractor.youtube.pot._registry import ( _pot_cache_provider_preferences, _pot_cache_providers, _pot_pcs_providers, ) from yt_dlp.extractor.youtube.pot.provider import PoTokenRequest class PoTokenCacheProviderError(IEContentProviderError): """An error occurred while fetching a PO Token""" class PoTokenCacheProvider(IEContentProvider, abc.ABC, suffix='PCP'): @abc.abstractmethod def get(self, key: str) -> str | None: pass @abc.abstractmethod def store(self, key: str, value: str, expires_at: int): pass @abc.abstractmethod def delete(self, key: str): pass class CacheProviderWritePolicy(enum.Enum): WRITE_ALL = enum.auto() # Write to all cache providers WRITE_FIRST = enum.auto() # Write to only the first cache provider @dataclasses.dataclass class PoTokenCacheSpec: key_bindings: dict[str, str | None] default_ttl: int write_policy: CacheProviderWritePolicy = CacheProviderWritePolicy.WRITE_ALL # Internal _provider: PoTokenCacheSpecProvider | None = None class PoTokenCacheSpecProvider(IEContentProvider, abc.ABC, suffix='PCSP'): def is_available(self) -> bool: return True @abc.abstractmethod def generate_cache_spec(self, request: PoTokenRequest) -> PoTokenCacheSpec | None: """Generate a cache spec for the given request""" pass def register_provider(provider: type[PoTokenCacheProvider]): """Register a PoTokenCacheProvider class""" return register_provider_generic( provider=provider, base_class=PoTokenCacheProvider, registry=_pot_cache_providers.value, ) def register_spec(provider: type[PoTokenCacheSpecProvider]): """Register a PoTokenCacheSpecProvider class""" return register_provider_generic( provider=provider, base_class=PoTokenCacheSpecProvider, registry=_pot_pcs_providers.value, ) def register_preference( *providers: type[PoTokenCacheProvider]) -> typing.Callable[[CacheProviderPreference], CacheProviderPreference]: """Register a preference for a PoTokenCacheProvider""" return register_preference_generic( PoTokenCacheProvider, _pot_cache_provider_preferences.value, *providers, ) if typing.TYPE_CHECKING: CacheProviderPreference = typing.Callable[[PoTokenCacheProvider, PoTokenRequest], int]
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/extractor/youtube/pot/provider.py
yt_dlp/extractor/youtube/pot/provider.py
"""PUBLIC API""" from __future__ import annotations import abc import copy import dataclasses import enum import functools import typing import urllib.parse from yt_dlp.cookies import YoutubeDLCookieJar from yt_dlp.extractor.youtube.pot._provider import ( IEContentProvider, IEContentProviderError, register_preference_generic, register_provider_generic, ) from yt_dlp.extractor.youtube.pot._registry import _pot_providers, _ptp_preferences from yt_dlp.networking import Request, Response from yt_dlp.utils import traverse_obj from yt_dlp.utils.networking import HTTPHeaderDict __all__ = [ 'ExternalRequestFeature', 'PoTokenContext', 'PoTokenProvider', 'PoTokenProviderError', 'PoTokenProviderRejectedRequest', 'PoTokenRequest', 'PoTokenResponse', 'provider_bug_report_message', 'register_preference', 'register_provider', ] class PoTokenContext(enum.Enum): GVS = 'gvs' PLAYER = 'player' SUBS = 'subs' @dataclasses.dataclass class PoTokenRequest: # YouTube parameters context: PoTokenContext innertube_context: InnertubeContext innertube_host: str | None = None session_index: str | None = None player_url: str | None = None is_authenticated: bool = False video_webpage: str | None = None internal_client_name: str | None = None # Content binding parameters visitor_data: str | None = None data_sync_id: str | None = None video_id: str | None = None # Internal, YouTube experiment on whether to bind GVS PO Token to video_id. _gvs_bind_to_video_id: bool = False # Networking parameters request_cookiejar: YoutubeDLCookieJar = dataclasses.field(default_factory=YoutubeDLCookieJar) request_proxy: str | None = None request_headers: HTTPHeaderDict = dataclasses.field(default_factory=HTTPHeaderDict) request_timeout: float | None = None request_source_address: str | None = None request_verify_tls: bool = True # Generate a new token, do not used a cached token # The token should still be cached for future requests bypass_cache: bool = False def copy(self): return dataclasses.replace( self, request_headers=HTTPHeaderDict(self.request_headers), innertube_context=copy.deepcopy(self.innertube_context), ) @dataclasses.dataclass class PoTokenResponse: po_token: str expires_at: int | None = None class PoTokenProviderRejectedRequest(IEContentProviderError): """Reject the PoTokenRequest (cannot handle the request)""" class PoTokenProviderError(IEContentProviderError): """An error occurred while fetching a PO Token""" class ExternalRequestFeature(enum.Enum): PROXY_SCHEME_HTTP = enum.auto() PROXY_SCHEME_HTTPS = enum.auto() PROXY_SCHEME_SOCKS4 = enum.auto() PROXY_SCHEME_SOCKS4A = enum.auto() PROXY_SCHEME_SOCKS5 = enum.auto() PROXY_SCHEME_SOCKS5H = enum.auto() SOURCE_ADDRESS = enum.auto() DISABLE_TLS_VERIFICATION = enum.auto() class PoTokenProvider(IEContentProvider, abc.ABC, suffix='PTP'): # Set to None to disable the check _SUPPORTED_CONTEXTS: tuple[PoTokenContext] | None = () # Innertube Client Name. # For example, "WEB", "ANDROID", "TVHTML5". # For a list of WebPO client names, see yt_dlp.extractor.youtube.pot.utils.WEBPO_CLIENTS. # Also see yt_dlp.extractor.youtube._base.INNERTUBE_CLIENTS # for a list of client names currently supported by the YouTube extractor. _SUPPORTED_CLIENTS: tuple[str] | None = () # If making external requests to websites (i.e. to youtube.com) # using another library or service (i.e., not _request_webpage), # add the request features that are supported. # If only using _request_webpage to make external requests, set this to None. _SUPPORTED_EXTERNAL_REQUEST_FEATURES: tuple[ExternalRequestFeature] | None = () def __validate_request(self, request: PoTokenRequest): if not self.is_available(): raise PoTokenProviderRejectedRequest(f'{self.PROVIDER_NAME} is not available') # Validate request using built-in settings if ( self._SUPPORTED_CONTEXTS is not None and request.context not in self._SUPPORTED_CONTEXTS ): raise PoTokenProviderRejectedRequest( f'PO Token Context "{request.context}" is not supported by {self.PROVIDER_NAME}') if self._SUPPORTED_CLIENTS is not None: client_name = traverse_obj( request.innertube_context, ('client', 'clientName')) if client_name not in self._SUPPORTED_CLIENTS: raise PoTokenProviderRejectedRequest( f'Client "{client_name}" is not supported by {self.PROVIDER_NAME}. ' f'Supported clients: {", ".join(self._SUPPORTED_CLIENTS) or "none"}') self.__validate_external_request_features(request) @functools.cached_property def _supported_proxy_schemes(self): return { scheme: feature for scheme, feature in { 'http': ExternalRequestFeature.PROXY_SCHEME_HTTP, 'https': ExternalRequestFeature.PROXY_SCHEME_HTTPS, 'socks4': ExternalRequestFeature.PROXY_SCHEME_SOCKS4, 'socks4a': ExternalRequestFeature.PROXY_SCHEME_SOCKS4A, 'socks5': ExternalRequestFeature.PROXY_SCHEME_SOCKS5, 'socks5h': ExternalRequestFeature.PROXY_SCHEME_SOCKS5H, }.items() if feature in (self._SUPPORTED_EXTERNAL_REQUEST_FEATURES or []) } def __validate_external_request_features(self, request: PoTokenRequest): if self._SUPPORTED_EXTERNAL_REQUEST_FEATURES is None: return if request.request_proxy: scheme = urllib.parse.urlparse(request.request_proxy).scheme if scheme.lower() not in self._supported_proxy_schemes: raise PoTokenProviderRejectedRequest( f'External requests by "{self.PROVIDER_NAME}" provider do not ' f'support proxy scheme "{scheme}". Supported proxy schemes: ' f'{", ".join(self._supported_proxy_schemes) or "none"}') if ( request.request_source_address and ExternalRequestFeature.SOURCE_ADDRESS not in self._SUPPORTED_EXTERNAL_REQUEST_FEATURES ): raise PoTokenProviderRejectedRequest( f'External requests by "{self.PROVIDER_NAME}" provider ' f'do not support setting source address') if ( not request.request_verify_tls and ExternalRequestFeature.DISABLE_TLS_VERIFICATION not in self._SUPPORTED_EXTERNAL_REQUEST_FEATURES ): raise PoTokenProviderRejectedRequest( f'External requests by "{self.PROVIDER_NAME}" provider ' f'do not support ignoring TLS certificate failures') def request_pot(self, request: PoTokenRequest) -> PoTokenResponse: self.__validate_request(request) return self._real_request_pot(request) @abc.abstractmethod def _real_request_pot(self, request: PoTokenRequest) -> PoTokenResponse: """To be implemented by subclasses""" pass # Helper functions def _request_webpage(self, request: Request, pot_request: PoTokenRequest | None = None, note=None, **kwargs) -> Response: """Make a request using the internal HTTP Client. Use this instead of calling requests, urllib3 or other HTTP client libraries directly! YouTube cookies will be automatically applied if this request is made to YouTube. @param request: The request to make @param pot_request: The PoTokenRequest to use. Request parameters will be merged from it. @param note: Custom log message to display when making the request. Set to `False` to disable logging. Tips: - Disable proxy (e.g. if calling local service): Request(..., proxies={'all': None}) - Set request timeout: Request(..., extensions={'timeout': 5.0}) """ req = request.copy() # Merge some ctx request settings into the request # Most of these will already be used by the configured ydl instance, # however, the YouTube extractor may override some. if pot_request is not None: req.headers = HTTPHeaderDict(pot_request.request_headers, req.headers) req.proxies = req.proxies or ({'all': pot_request.request_proxy} if pot_request.request_proxy else {}) if pot_request.request_cookiejar is not None: req.extensions['cookiejar'] = req.extensions.get('cookiejar', pot_request.request_cookiejar) if note is not False: self.logger.info(str(note) if note else 'Requesting webpage') return self.ie._downloader.urlopen(req) def register_provider(provider: type[PoTokenProvider]): """Register a PoTokenProvider class""" return register_provider_generic( provider=provider, base_class=PoTokenProvider, registry=_pot_providers.value, ) def provider_bug_report_message(provider: IEContentProvider, before=';'): msg = provider.BUG_REPORT_MESSAGE before = before.rstrip() if not before or before.endswith(('.', '!', '?')): msg = msg[0].title() + msg[1:] return f'{before} {msg}' if before else msg def register_preference(*providers: type[PoTokenProvider]) -> typing.Callable[[Preference], Preference]: """Register a preference for a PoTokenProvider""" return register_preference_generic( PoTokenProvider, _ptp_preferences.value, *providers, ) if typing.TYPE_CHECKING: Preference = typing.Callable[[PoTokenProvider, PoTokenRequest], int] __all__.append('Preference') # Barebones innertube context. There may be more fields. class ClientInfo(typing.TypedDict, total=False): hl: str | None gl: str | None remoteHost: str | None deviceMake: str | None deviceModel: str | None visitorData: str | None userAgent: str | None clientName: str clientVersion: str osName: str | None osVersion: str | None class InnertubeContext(typing.TypedDict, total=False): client: ClientInfo request: dict user: dict
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/extractor/youtube/pot/_registry.py
yt_dlp/extractor/youtube/pot/_registry.py
from yt_dlp.globals import Indirect _pot_providers = Indirect({}) _ptp_preferences = Indirect(set()) _pot_pcs_providers = Indirect({}) _pot_cache_providers = Indirect({}) _pot_cache_provider_preferences = Indirect(set()) _pot_memory_cache = Indirect({})
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/extractor/youtube/pot/_builtin/webpo_cachespec.py
yt_dlp/extractor/youtube/pot/_builtin/webpo_cachespec.py
from __future__ import annotations from yt_dlp.extractor.youtube.pot._provider import BuiltinIEContentProvider from yt_dlp.extractor.youtube.pot.cache import ( CacheProviderWritePolicy, PoTokenCacheSpec, PoTokenCacheSpecProvider, register_spec, ) from yt_dlp.extractor.youtube.pot.provider import ( PoTokenRequest, ) from yt_dlp.extractor.youtube.pot.utils import ContentBindingType, get_webpo_content_binding from yt_dlp.utils import traverse_obj @register_spec class WebPoPCSP(PoTokenCacheSpecProvider, BuiltinIEContentProvider): PROVIDER_NAME = 'webpo' def generate_cache_spec(self, request: PoTokenRequest) -> PoTokenCacheSpec | None: bind_to_visitor_id = self._configuration_arg( 'bind_to_visitor_id', default=['true'])[0] == 'true' content_binding, content_binding_type = get_webpo_content_binding( request, bind_to_visitor_id=bind_to_visitor_id) if not content_binding or not content_binding_type: return None write_policy = CacheProviderWritePolicy.WRITE_ALL if content_binding_type == ContentBindingType.VIDEO_ID: write_policy = CacheProviderWritePolicy.WRITE_FIRST return PoTokenCacheSpec( key_bindings={ 't': 'webpo', 'cb': content_binding, 'cbt': content_binding_type.value, 'ip': traverse_obj(request.innertube_context, ('client', 'remoteHost')), 'sa': request.request_source_address, 'px': request.request_proxy, }, # Integrity token response usually states it has a ttl of 12 hours (43200 seconds). # We will default to 6 hours to be safe. default_ttl=21600, write_policy=write_policy, )
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/extractor/youtube/pot/_builtin/memory_cache.py
yt_dlp/extractor/youtube/pot/_builtin/memory_cache.py
from __future__ import annotations import datetime as dt import typing from threading import Lock from yt_dlp.extractor.youtube.pot._provider import BuiltinIEContentProvider from yt_dlp.extractor.youtube.pot._registry import _pot_memory_cache from yt_dlp.extractor.youtube.pot.cache import ( PoTokenCacheProvider, register_preference, register_provider, ) def initialize_global_cache(max_size: int): if _pot_memory_cache.value.get('cache') is None: _pot_memory_cache.value['cache'] = {} _pot_memory_cache.value['lock'] = Lock() _pot_memory_cache.value['max_size'] = max_size if _pot_memory_cache.value['max_size'] != max_size: raise ValueError('Cannot change max_size of initialized global memory cache') return ( _pot_memory_cache.value['cache'], _pot_memory_cache.value['lock'], _pot_memory_cache.value['max_size'], ) @register_provider class MemoryLRUPCP(PoTokenCacheProvider, BuiltinIEContentProvider): PROVIDER_NAME = 'memory' DEFAULT_CACHE_SIZE = 25 def __init__( self, *args, initialize_cache: typing.Callable[[int], tuple[dict[str, tuple[str, int]], Lock, int]] = initialize_global_cache, **kwargs, ): super().__init__(*args, **kwargs) self.cache, self.lock, self.max_size = initialize_cache(self.DEFAULT_CACHE_SIZE) def is_available(self) -> bool: return True def get(self, key: str) -> str | None: with self.lock: if key not in self.cache: return None value, expires_at = self.cache.pop(key) if expires_at < int(dt.datetime.now(dt.timezone.utc).timestamp()): return None self.cache[key] = (value, expires_at) return value def store(self, key: str, value: str, expires_at: int): with self.lock: if expires_at < int(dt.datetime.now(dt.timezone.utc).timestamp()): return if key in self.cache: self.cache.pop(key) self.cache[key] = (value, expires_at) if len(self.cache) > self.max_size: oldest_key = next(iter(self.cache)) self.cache.pop(oldest_key) def delete(self, key: str): with self.lock: self.cache.pop(key, None) @register_preference(MemoryLRUPCP) def memorylru_preference(*_, **__): # Memory LRU Cache SHOULD be the highest priority return 10000
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/extractor/youtube/pot/_builtin/__init__.py
yt_dlp/extractor/youtube/pot/_builtin/__init__.py
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/compat/_legacy.py
yt_dlp/compat/_legacy.py
""" Do not use! """ import base64 import collections import ctypes import getpass import html.entities import html.parser import http.client import http.cookiejar import http.cookies import http.server import itertools import os import shlex import shutil import socket import struct import subprocess import tokenize import urllib.error import urllib.parse import urllib.request import xml.etree.ElementTree as etree # isort: split import asyncio # noqa: F401 import re # noqa: F401 from asyncio import run as compat_asyncio_run # noqa: F401 from re import Pattern as compat_Pattern # noqa: F401 from re import match as compat_Match # noqa: F401 from . import compat_expanduser, compat_HTMLParseError from .compat_utils import passthrough_module from ..dependencies import brotli as compat_brotli # noqa: F401 from ..dependencies import websockets as compat_websockets # noqa: F401 from ..dependencies.Cryptodome import AES as compat_pycrypto_AES # noqa: F401 from ..networking.exceptions import HTTPError as compat_HTTPError passthrough_module(__name__, '...utils', ('windows_enable_vt_mode',)) # compat_ctypes_WINFUNCTYPE = ctypes.WINFUNCTYPE # will not work since ctypes.WINFUNCTYPE does not exist in UNIX machines def compat_ctypes_WINFUNCTYPE(*args, **kwargs): return ctypes.WINFUNCTYPE(*args, **kwargs) def compat_setenv(key, value, env=os.environ): env[key] = value compat_base64_b64decode = base64.b64decode compat_basestring = str compat_casefold = str.casefold compat_chr = chr compat_collections_abc = collections.abc compat_cookiejar = compat_http_cookiejar = http.cookiejar compat_cookiejar_Cookie = compat_http_cookiejar_Cookie = http.cookiejar.Cookie compat_cookies = compat_http_cookies = http.cookies compat_cookies_SimpleCookie = compat_http_cookies_SimpleCookie = http.cookies.SimpleCookie compat_etree_Element = compat_xml_etree_ElementTree_Element = etree.Element compat_etree_register_namespace = compat_xml_etree_register_namespace = etree.register_namespace compat_filter = filter compat_get_terminal_size = shutil.get_terminal_size compat_getenv = os.getenv compat_getpass = compat_getpass_getpass = getpass.getpass compat_html_entities = html.entities compat_html_entities_html5 = html.entities.html5 compat_html_parser_HTMLParseError = compat_HTMLParseError compat_HTMLParser = compat_html_parser_HTMLParser = html.parser.HTMLParser compat_http_client = http.client compat_http_server = http.server compat_input = input compat_integer_types = (int, ) compat_itertools_count = itertools.count compat_kwargs = lambda kwargs: kwargs compat_map = map compat_numeric_types = (int, float, complex) compat_os_path_expanduser = compat_expanduser compat_os_path_realpath = os.path.realpath compat_print = print compat_shlex_split = shlex.split compat_socket_create_connection = socket.create_connection compat_Struct = struct.Struct compat_struct_pack = struct.pack compat_struct_unpack = struct.unpack compat_subprocess_get_DEVNULL = lambda: subprocess.DEVNULL compat_tokenize_tokenize = tokenize.tokenize compat_urllib_error = urllib.error compat_urllib_HTTPError = compat_HTTPError compat_urllib_parse = urllib.parse compat_urllib_parse_parse_qs = urllib.parse.parse_qs compat_urllib_parse_quote = urllib.parse.quote compat_urllib_parse_quote_plus = urllib.parse.quote_plus compat_urllib_parse_unquote_plus = urllib.parse.unquote_plus compat_urllib_parse_unquote_to_bytes = urllib.parse.unquote_to_bytes compat_urllib_parse_urlunparse = urllib.parse.urlunparse compat_urllib_request = urllib.request compat_urllib_request_DataHandler = urllib.request.DataHandler compat_urllib_response = urllib.response compat_urlretrieve = compat_urllib_request_urlretrieve = urllib.request.urlretrieve compat_xml_parse_error = compat_xml_etree_ElementTree_ParseError = etree.ParseError compat_xpath = lambda xpath: xpath compat_zip = zip workaround_optparse_bug9161 = lambda: None compat_str = str compat_b64decode = base64.b64decode compat_urlparse = urllib.parse compat_parse_qs = urllib.parse.parse_qs compat_urllib_parse_unquote = urllib.parse.unquote compat_urllib_parse_urlencode = urllib.parse.urlencode compat_urllib_parse_urlparse = urllib.parse.urlparse legacy = []
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/compat/shutil.py
yt_dlp/compat/shutil.py
# flake8: noqa: F405 from shutil import * # noqa: F403 from .compat_utils import passthrough_module passthrough_module(__name__, 'shutil') del passthrough_module import sys if sys.platform.startswith('freebsd'): import errno import os import shutil # Workaround for PermissionError when using restricted ACL mode on FreeBSD def copy2(src, dst, *args, **kwargs): if os.path.isdir(dst): dst = os.path.join(dst, os.path.basename(src)) shutil.copyfile(src, dst, *args, **kwargs) try: shutil.copystat(src, dst, *args, **kwargs) except PermissionError as e: if e.errno != getattr(errno, 'EPERM', None): raise return dst def move(*args, copy_function=copy2, **kwargs): return shutil.move(*args, copy_function=copy_function, **kwargs)
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/compat/_deprecated.py
yt_dlp/compat/_deprecated.py
"""Deprecated - New code should avoid these""" import warnings from .compat_utils import passthrough_module # XXX: Implement this the same way as other DeprecationWarnings without circular import passthrough_module(__name__, '.._legacy', callback=lambda attr: warnings.warn( DeprecationWarning(f'{__name__}.{attr} is deprecated'), stacklevel=6)) del passthrough_module import functools # noqa: F401 import os compat_os_name = os.name compat_realpath = os.path.realpath def compat_shlex_quote(s): from ..utils import shell_quote return shell_quote(s)
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/compat/__init__.py
yt_dlp/compat/__init__.py
import datetime as dt import os import xml.etree.ElementTree as etree from .compat_utils import passthrough_module passthrough_module(__name__, '._deprecated') del passthrough_module # HTMLParseError has been deprecated in Python 3.3 and removed in # Python 3.5. Introducing dummy exception for Python >3.5 for compatible # and uniform cross-version exception handling class compat_HTMLParseError(ValueError): pass class _TreeBuilder(etree.TreeBuilder): def doctype(self, name, pubid, system): pass def compat_etree_fromstring(text): return etree.XML(text, parser=etree.XMLParser(target=_TreeBuilder())) def compat_ord(c): return c if isinstance(c, int) else ord(c) def compat_datetime_from_timestamp(timestamp): # Calling dt.datetime.fromtimestamp with negative timestamps throws error in Windows # Ref: https://github.com/yt-dlp/yt-dlp/issues/5185, https://github.com/python/cpython/issues/81708, # https://github.com/yt-dlp/yt-dlp/issues/6706#issuecomment-1496842642 return (dt.datetime.fromtimestamp(0, dt.timezone.utc) + dt.timedelta(seconds=timestamp)) # Python 3.8+ does not honor %HOME% on windows, but this breaks compatibility with youtube-dl # See https://github.com/yt-dlp/yt-dlp/issues/792 # https://docs.python.org/3/library/os.path.html#os.path.expanduser if os.name in ('nt', 'ce'): def compat_expanduser(path): HOME = os.environ.get('HOME') if not HOME: return os.path.expanduser(path) elif not path.startswith('~'): return path i = path.replace('\\', '/', 1).find('/') # ~user if i < 0: i = len(path) userhome = os.path.join(os.path.dirname(HOME), path[1:i]) if i > 1 else HOME return userhome + path[i:] else: compat_expanduser = os.path.expanduser def urllib_req_to_req(urllib_request): """Convert urllib Request to a networking Request""" from ..networking import Request from ..utils.networking import HTTPHeaderDict return Request( urllib_request.get_full_url(), data=urllib_request.data, method=urllib_request.get_method(), headers=HTTPHeaderDict(urllib_request.headers, urllib_request.unredirected_hdrs), extensions={'timeout': urllib_request.timeout} if hasattr(urllib_request, 'timeout') else None)
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/compat/imghdr.py
yt_dlp/compat/imghdr.py
def what(file=None, h=None): """Detect format of image (Currently supports jpeg, png, webp, gif only) Ref: https://github.com/python/cpython/blob/3.11/Lib/imghdr.py Ref: https://www.w3.org/Graphics/JPEG/itu-t81.pdf """ if h is None: with open(file, 'rb') as f: h = f.read(12) if h.startswith(b'RIFF') and h.startswith(b'WEBP', 8): return 'webp' if h.startswith(b'\x89PNG'): return 'png' if h.startswith(b'\xFF\xD8\xFF'): return 'jpeg' if h.startswith(b'GIF'): return 'gif' return None
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/compat/compat_utils.py
yt_dlp/compat/compat_utils.py
import collections import contextlib import functools import importlib import sys import types _NO_ATTRIBUTE = object() _Package = collections.namedtuple('Package', ('name', 'version')) def get_package_info(module): return _Package( name=getattr(module, '_yt_dlp__identifier', module.__name__), version=str(next(filter(None, ( getattr(module, attr, None) for attr in ('_yt_dlp__version', '__version__', 'version_string', 'version') )), None))) def _is_package(module): return '__path__' in vars(module) def _is_dunder(name): return name.startswith('__') and name.endswith('__') class EnhancedModule(types.ModuleType): def __bool__(self): return vars(self).get('__bool__', lambda: True)() def __getattribute__(self, attr): try: ret = super().__getattribute__(attr) except AttributeError: if _is_dunder(attr): raise getter = getattr(self, '__getattr__', None) if not getter: raise ret = getter(attr) return ret.fget() if isinstance(ret, property) else ret def passthrough_module(parent, child, allowed_attributes=(..., ), *, callback=lambda _: None): """Passthrough parent module into a child module, creating the parent if necessary""" def __getattr__(attr): if _is_package(parent): with contextlib.suppress(ModuleNotFoundError): return importlib.import_module(f'.{attr}', parent.__name__) ret = from_child(attr) if ret is _NO_ATTRIBUTE: raise AttributeError(f'module {parent.__name__} has no attribute {attr}') callback(attr) return ret @functools.cache def from_child(attr): nonlocal child if attr not in allowed_attributes: if ... not in allowed_attributes or _is_dunder(attr): return _NO_ATTRIBUTE if isinstance(child, str): child = importlib.import_module(child, parent.__name__) if _is_package(child): with contextlib.suppress(ImportError): return passthrough_module(f'{parent.__name__}.{attr}', importlib.import_module(f'.{attr}', child.__name__)) with contextlib.suppress(AttributeError): return getattr(child, attr) return _NO_ATTRIBUTE parent = sys.modules.get(parent, types.ModuleType(parent)) parent.__class__ = EnhancedModule parent.__getattr__ = __getattr__ return parent
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/compat/urllib/request.py
yt_dlp/compat/urllib/request.py
# flake8: noqa: F405 from urllib.request import * # noqa: F403 from ..compat_utils import passthrough_module passthrough_module(__name__, 'urllib.request') del passthrough_module import os if os.name == 'nt': # On older Python versions, proxies are extracted from Windows registry erroneously. [1] # If the https proxy in the registry does not have a scheme, urllib will incorrectly add https:// to it. [2] # It is unlikely that the user has actually set it to be https, so we should be fine to safely downgrade # it to http on these older Python versions to avoid issues # This also applies for ftp proxy type, as ftp:// proxy scheme is not supported. # 1: https://github.com/python/cpython/issues/86793 # 2: https://github.com/python/cpython/blob/51f1ae5ceb0673316c4e4b0175384e892e33cc6e/Lib/urllib/request.py#L2683-L2698 import sys from urllib.request import getproxies_environment, getproxies_registry def getproxies_registry_patched(): proxies = getproxies_registry() if sys.version_info < (3, 10, 5): # https://docs.python.org/3.10/whatsnew/changelog.html#python-3-10-5-final for scheme in ('https', 'ftp'): if scheme in proxies and proxies[scheme].startswith(f'{scheme}://'): proxies[scheme] = 'http' + proxies[scheme][len(scheme):] return proxies def getproxies(): return getproxies_environment() or getproxies_registry_patched() del os
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/compat/urllib/__init__.py
yt_dlp/compat/urllib/__init__.py
# flake8: noqa: F405 from urllib import * # noqa: F403 del request # noqa: F821 from . import request # noqa: F401 from ..compat_utils import passthrough_module passthrough_module(__name__, 'urllib') del passthrough_module
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/dependencies/Cryptodome.py
yt_dlp/dependencies/Cryptodome.py
from ..compat.compat_utils import passthrough_module try: import Cryptodome as _parent except ImportError: try: import Crypto as _parent except (ImportError, SyntaxError): # Old Crypto gives SyntaxError in newer Python _parent = passthrough_module(__name__, 'no_Cryptodome') __bool__ = lambda: False del passthrough_module __version__ = '' AES = PKCS1_v1_5 = Blowfish = PKCS1_OAEP = SHA1 = CMAC = RSA = None try: if _parent.__name__ == 'Cryptodome': from Cryptodome import __version__ from Cryptodome.Cipher import AES, PKCS1_OAEP, Blowfish, PKCS1_v1_5 from Cryptodome.Hash import CMAC, SHA1 from Cryptodome.PublicKey import RSA elif _parent.__name__ == 'Crypto': from Crypto import __version__ from Crypto.Cipher import AES, PKCS1_OAEP, Blowfish, PKCS1_v1_5 # noqa: F401 from Crypto.Hash import CMAC, SHA1 # noqa: F401 from Crypto.PublicKey import RSA # noqa: F401 except (ImportError, OSError): __version__ = f'broken {__version__}'.strip() _yt_dlp__identifier = _parent.__name__ if AES and _yt_dlp__identifier == 'Crypto': try: # In pycrypto, mode defaults to ECB. See: # https://www.pycryptodome.org/en/latest/src/vs_pycrypto.html#:~:text=not%20have%20ECB%20as%20default%20mode AES.new(b'abcdefghijklmnop') except TypeError: _yt_dlp__identifier = 'pycrypto'
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/dependencies/__init__.py
yt_dlp/dependencies/__init__.py
# flake8: noqa: F401 """Imports all optional dependencies for the project. An attribute "_yt_dlp__identifier" may be inserted into the module if it uses an ambiguous namespace""" try: import brotlicffi as brotli except ImportError: try: import brotli except ImportError: brotli = None try: import certifi except ImportError: certifi = None else: from os.path import exists as _path_exists # The certificate may not be bundled in executable if not _path_exists(certifi.where()): certifi = None try: import mutagen except ImportError: mutagen = None secretstorage = None try: import secretstorage _SECRETSTORAGE_UNAVAILABLE_REASON = None except ImportError: _SECRETSTORAGE_UNAVAILABLE_REASON = ( 'as the `secretstorage` module is not installed. ' 'Please install by running `python3 -m pip install secretstorage`') except Exception as _err: _SECRETSTORAGE_UNAVAILABLE_REASON = f'as the `secretstorage` module could not be initialized. {_err}' try: import sqlite3 # We need to get the underlying `sqlite` version, see https://github.com/yt-dlp/yt-dlp/issues/8152 sqlite3._yt_dlp__version = sqlite3.sqlite_version except ImportError: # although sqlite3 is part of the standard library, it is possible to compile Python without # sqlite support. See: https://github.com/yt-dlp/yt-dlp/issues/544 sqlite3 = None try: import websockets except ImportError: websockets = None try: import urllib3 except ImportError: urllib3 = None try: import requests except ImportError: requests = None try: import xattr # xattr or pyxattr except ImportError: xattr = None else: if hasattr(xattr, 'set'): # pyxattr xattr._yt_dlp__identifier = 'pyxattr' try: import curl_cffi except ImportError: curl_cffi = None from . import Cryptodome try: import yt_dlp_ejs except ImportError: yt_dlp_ejs = None all_dependencies = {k: v for k, v in globals().items() if not k.startswith('_')} available_dependencies = {k: v for k, v in all_dependencies.items() if v} # Deprecated Cryptodome_AES = Cryptodome.AES __all__ = [ 'all_dependencies', 'available_dependencies', *all_dependencies.keys(), ]
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/utils/_jsruntime.py
yt_dlp/utils/_jsruntime.py
from __future__ import annotations import abc import dataclasses import functools import os.path import sys from ._utils import _get_exe_version_output, detect_exe_version, version_tuple _FALLBACK_PATHEXT = ('.COM', '.EXE', '.BAT', '.CMD') def _find_exe(basename: str) -> str: if os.name != 'nt': return basename paths: list[str] = [] # binary dir if getattr(sys, 'frozen', False): paths.append(os.path.dirname(sys.executable)) # cwd paths.append(os.getcwd()) # PATH items if path := os.environ.get('PATH'): paths.extend(filter(None, path.split(os.path.pathsep))) pathext = os.environ.get('PATHEXT') if pathext is None: exts = _FALLBACK_PATHEXT else: exts = tuple(ext for ext in pathext.split(os.pathsep) if ext) visited = [] for path in map(os.path.realpath, paths): normed = os.path.normcase(path) if normed in visited: continue visited.append(normed) for ext in exts: binary = os.path.join(path, f'{basename}{ext}') if os.access(binary, os.F_OK | os.X_OK) and not os.path.isdir(binary): return binary return basename def _determine_runtime_path(path, basename): if not path: return _find_exe(basename) if os.path.isdir(path): return os.path.join(path, basename) return path @dataclasses.dataclass(frozen=True) class JsRuntimeInfo: name: str path: str version: str version_tuple: tuple[int, ...] supported: bool = True class JsRuntime(abc.ABC): def __init__(self, path=None): self._path = path @functools.cached_property def info(self) -> JsRuntimeInfo | None: return self._info() @abc.abstractmethod def _info(self) -> JsRuntimeInfo | None: raise NotImplementedError class DenoJsRuntime(JsRuntime): MIN_SUPPORTED_VERSION = (2, 0, 0) def _info(self): path = _determine_runtime_path(self._path, 'deno') out = _get_exe_version_output(path, ['--version']) if not out: return None version = detect_exe_version(out, r'^deno (\S+)', 'unknown') vt = version_tuple(version, lenient=True) return JsRuntimeInfo( name='deno', path=path, version=version, version_tuple=vt, supported=vt >= self.MIN_SUPPORTED_VERSION) class BunJsRuntime(JsRuntime): MIN_SUPPORTED_VERSION = (1, 0, 31) def _info(self): path = _determine_runtime_path(self._path, 'bun') out = _get_exe_version_output(path, ['--version']) if not out: return None version = detect_exe_version(out, r'^(\S+)', 'unknown') vt = version_tuple(version, lenient=True) return JsRuntimeInfo( name='bun', path=path, version=version, version_tuple=vt, supported=vt >= self.MIN_SUPPORTED_VERSION) class NodeJsRuntime(JsRuntime): MIN_SUPPORTED_VERSION = (20, 0, 0) def _info(self): path = _determine_runtime_path(self._path, 'node') out = _get_exe_version_output(path, ['--version']) if not out: return None version = detect_exe_version(out, r'^v(\S+)', 'unknown') vt = version_tuple(version, lenient=True) return JsRuntimeInfo( name='node', path=path, version=version, version_tuple=vt, supported=vt >= self.MIN_SUPPORTED_VERSION) class QuickJsRuntime(JsRuntime): MIN_SUPPORTED_VERSION = (2023, 12, 9) def _info(self): path = _determine_runtime_path(self._path, 'qjs') # quickjs does not have --version and --help returns a status code of 1 out = _get_exe_version_output(path, ['--help'], ignore_return_code=True) if not out: return None is_ng = 'QuickJS-ng' in out version = detect_exe_version(out, r'^QuickJS(?:-ng)?\s+version\s+(\S+)', 'unknown') vt = version_tuple(version, lenient=True) if is_ng: return JsRuntimeInfo( name='quickjs-ng', path=path, version=version, version_tuple=vt, supported=vt > (0,)) return JsRuntimeInfo( name='quickjs', path=path, version=version, version_tuple=vt, supported=vt >= self.MIN_SUPPORTED_VERSION)
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/utils/_legacy.py
yt_dlp/utils/_legacy.py
"""No longer used and new code should not use. Exists only for API compat.""" import platform import struct import sys import urllib.error import urllib.parse import urllib.request import zlib from ._utils import Popen, decode_base_n, preferredencoding from .traversal import traverse_obj from ..dependencies import certifi, websockets from ..networking._helper import make_ssl_context from ..networking._urllib import HTTPHandler # isort: split from .networking import escape_rfc3986 # noqa: F401 from .networking import normalize_url as escape_url from .networking import random_user_agent, std_headers # noqa: F401 from ..cookies import YoutubeDLCookieJar # noqa: F401 from ..networking._urllib import PUTRequest # noqa: F401 from ..networking._urllib import SUPPORTED_ENCODINGS, HEADRequest # noqa: F401 from ..networking._urllib import ProxyHandler as PerRequestProxyHandler # noqa: F401 from ..networking._urllib import RedirectHandler as YoutubeDLRedirectHandler # noqa: F401 from ..networking._urllib import ( # noqa: F401 make_socks_conn_class, update_Request, ) from ..networking.exceptions import HTTPError, network_exceptions # noqa: F401 has_certifi = bool(certifi) has_websockets = bool(websockets) def load_plugins(name, suffix, namespace): from ..plugins import load_plugins ret = load_plugins(name, suffix) namespace.update(ret) return ret def traverse_dict(dictn, keys, casesense=True): return traverse_obj(dictn, keys, casesense=casesense, is_user_input=True, traverse_string=True) def decode_base(value, digits): return decode_base_n(value, table=digits) def platform_name(): """ Returns the platform name as a str """ return platform.platform() def get_subprocess_encoding(): if sys.platform == 'win32' and sys.getwindowsversion()[0] >= 5: # For subprocess calls, encode with locale encoding # Refer to http://stackoverflow.com/a/9951851/35070 encoding = preferredencoding() else: encoding = sys.getfilesystemencoding() if encoding is None: encoding = 'utf-8' return encoding # UNUSED # Based on png2str() written by @gdkchan and improved by @yokrysty # Originally posted at https://github.com/ytdl-org/youtube-dl/issues/9706 def decode_png(png_data): # Reference: https://www.w3.org/TR/PNG/ header = png_data[8:] if png_data[:8] != b'\x89PNG\x0d\x0a\x1a\x0a' or header[4:8] != b'IHDR': raise OSError('Not a valid PNG file.') int_map = {1: '>B', 2: '>H', 4: '>I'} unpack_integer = lambda x: struct.unpack(int_map[len(x)], x)[0] chunks = [] while header: length = unpack_integer(header[:4]) header = header[4:] chunk_type = header[:4] header = header[4:] chunk_data = header[:length] header = header[length:] header = header[4:] # Skip CRC chunks.append({ 'type': chunk_type, 'length': length, 'data': chunk_data, }) ihdr = chunks[0]['data'] width = unpack_integer(ihdr[:4]) height = unpack_integer(ihdr[4:8]) idat = b'' for chunk in chunks: if chunk['type'] == b'IDAT': idat += chunk['data'] if not idat: raise OSError('Unable to read PNG data.') decompressed_data = bytearray(zlib.decompress(idat)) stride = width * 3 pixels = [] def _get_pixel(idx): x = idx % stride y = idx // stride return pixels[y][x] for y in range(height): base_pos = y * (1 + stride) filter_type = decompressed_data[base_pos] current_row = [] pixels.append(current_row) for x in range(stride): color = decompressed_data[1 + base_pos + x] basex = y * stride + x left = 0 up = 0 if x > 2: left = _get_pixel(basex - 3) if y > 0: up = _get_pixel(basex - stride) if filter_type == 1: # Sub color = (color + left) & 0xff elif filter_type == 2: # Up color = (color + up) & 0xff elif filter_type == 3: # Average color = (color + ((left + up) >> 1)) & 0xff elif filter_type == 4: # Paeth a = left b = up c = 0 if x > 2 and y > 0: c = _get_pixel(basex - stride - 3) p = a + b - c pa = abs(p - a) pb = abs(p - b) pc = abs(p - c) if pa <= pb and pa <= pc: color = (color + a) & 0xff elif pb <= pc: color = (color + b) & 0xff else: color = (color + c) & 0xff current_row.append(color) return width, height, pixels def register_socks_protocols(): # "Register" SOCKS protocols # In Python < 2.6.5, urlsplit() suffers from bug https://bugs.python.org/issue7904 # URLs with protocols not in urlparse.uses_netloc are not handled correctly for scheme in ('socks', 'socks4', 'socks4a', 'socks5'): if scheme not in urllib.parse.uses_netloc: urllib.parse.uses_netloc.append(scheme) def handle_youtubedl_headers(headers): filtered_headers = headers if 'Youtubedl-no-compression' in filtered_headers: filtered_headers = {k: v for k, v in filtered_headers.items() if k.lower() != 'accept-encoding'} del filtered_headers['Youtubedl-no-compression'] return filtered_headers def request_to_url(req): if isinstance(req, urllib.request.Request): return req.get_full_url() else: return req def sanitized_Request(url, *args, **kwargs): from ..utils import extract_basic_auth, sanitize_url url, auth_header = extract_basic_auth(escape_url(sanitize_url(url))) if auth_header is not None: headers = args[1] if len(args) >= 2 else kwargs.setdefault('headers', {}) headers['Authorization'] = auth_header return urllib.request.Request(url, *args, **kwargs) class YoutubeDLHandler(HTTPHandler): def __init__(self, params, *args, **kwargs): self._params = params super().__init__(*args, **kwargs) YoutubeDLHTTPSHandler = YoutubeDLHandler class YoutubeDLCookieProcessor(urllib.request.HTTPCookieProcessor): def __init__(self, cookiejar=None): urllib.request.HTTPCookieProcessor.__init__(self, cookiejar) def http_response(self, request, response): return urllib.request.HTTPCookieProcessor.http_response(self, request, response) https_request = urllib.request.HTTPCookieProcessor.http_request https_response = http_response def make_HTTPS_handler(params, **kwargs): return YoutubeDLHTTPSHandler(params, context=make_ssl_context( verify=not params.get('nocheckcertificate'), client_certificate=params.get('client_certificate'), client_certificate_key=params.get('client_certificate_key'), client_certificate_password=params.get('client_certificate_password'), legacy_support=params.get('legacyserverconnect'), use_certifi='no-certifi' not in params.get('compat_opts', []), ), **kwargs) def process_communicate_or_kill(p, *args, **kwargs): return Popen.communicate_or_kill(p, *args, **kwargs) def encodeFilename(s, for_subprocess=False): assert isinstance(s, str) return s def decodeFilename(b, for_subprocess=False): return b def decodeArgument(b): return b def decodeOption(optval): if optval is None: return optval if isinstance(optval, bytes): optval = optval.decode(preferredencoding()) assert isinstance(optval, str) return optval def error_to_compat_str(err): return str(err)
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/utils/networking.py
yt_dlp/utils/networking.py
from __future__ import annotations import collections import collections.abc import random import typing import urllib.parse import urllib.request if typing.TYPE_CHECKING: T = typing.TypeVar('T') from ._utils import NO_DEFAULT, remove_start, format_field from .traversal import traverse_obj def random_user_agent(): USER_AGENT_TMPL = 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/{} Safari/537.36' # Target versions released within the last ~6 months CHROME_MAJOR_VERSION_RANGE = (137, 143) return USER_AGENT_TMPL.format(f'{random.randint(*CHROME_MAJOR_VERSION_RANGE)}.0.0.0') class HTTPHeaderDict(dict): """ Store and access keys case-insensitively. The constructor can take multiple dicts, in which keys in the latter are prioritised. Retains a case sensitive mapping of the headers, which can be accessed via `.sensitive()`. """ def __new__(cls, *args: typing.Any, **kwargs: typing.Any) -> typing.Self: obj = dict.__new__(cls, *args, **kwargs) obj.__sensitive_map = {} return obj def __init__(self, /, *args, **kwargs): super().__init__() self.__sensitive_map = {} for dct in filter(None, args): self.update(dct) if kwargs: self.update(kwargs) def sensitive(self, /) -> dict[str, str]: return { self.__sensitive_map[key]: value for key, value in self.items() } def __contains__(self, key: str, /) -> bool: return super().__contains__(key.title() if isinstance(key, str) else key) def __delitem__(self, key: str, /) -> None: key = key.title() del self.__sensitive_map[key] super().__delitem__(key) def __getitem__(self, key, /) -> str: return super().__getitem__(key.title()) def __ior__(self, other, /): if isinstance(other, type(self)): other = other.sensitive() if isinstance(other, dict): self.update(other) return return NotImplemented def __or__(self, other, /) -> typing.Self: if isinstance(other, type(self)): other = other.sensitive() if isinstance(other, dict): return type(self)(self.sensitive(), other) return NotImplemented def __ror__(self, other, /) -> typing.Self: if isinstance(other, type(self)): other = other.sensitive() if isinstance(other, dict): return type(self)(other, self.sensitive()) return NotImplemented def __setitem__(self, key: str, value, /) -> None: if isinstance(value, bytes): value = value.decode('latin-1') key_title = key.title() self.__sensitive_map[key_title] = key super().__setitem__(key_title, str(value).strip()) def clear(self, /) -> None: self.__sensitive_map.clear() super().clear() def copy(self, /) -> typing.Self: return type(self)(self.sensitive()) @typing.overload def get(self, key: str, /) -> str | None: ... @typing.overload def get(self, key: str, /, default: T) -> str | T: ... def get(self, key, /, default=NO_DEFAULT): key = key.title() if default is NO_DEFAULT: return super().get(key) return super().get(key, default) @typing.overload def pop(self, key: str, /) -> str: ... @typing.overload def pop(self, key: str, /, default: T) -> str | T: ... def pop(self, key, /, default=NO_DEFAULT): key = key.title() if default is NO_DEFAULT: self.__sensitive_map.pop(key) return super().pop(key) self.__sensitive_map.pop(key, default) return super().pop(key, default) def popitem(self) -> tuple[str, str]: self.__sensitive_map.popitem() return super().popitem() @typing.overload def setdefault(self, key: str, /) -> str: ... @typing.overload def setdefault(self, key: str, /, default) -> str: ... def setdefault(self, key, /, default=None) -> str: key = key.title() if key in self.__sensitive_map: return super().__getitem__(key) self[key] = default or '' return self[key] def update(self, other, /, **kwargs) -> None: if isinstance(other, type(self)): other = other.sensitive() if isinstance(other, collections.abc.Mapping): for key, value in other.items(): self[key] = value elif hasattr(other, 'keys'): for key in other.keys(): # noqa: SIM118 self[key] = other[key] else: for key, value in other: self[key] = value for key, value in kwargs.items(): self[key] = value std_headers = HTTPHeaderDict({ 'User-Agent': random_user_agent(), 'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8', 'Accept-Language': 'en-us,en;q=0.5', 'Sec-Fetch-Mode': 'navigate', }) def clean_proxies(proxies: dict, headers: HTTPHeaderDict): req_proxy = headers.pop('Ytdl-Request-Proxy', None) if req_proxy: proxies.clear() # XXX: compat: Ytdl-Request-Proxy takes preference over everything, including NO_PROXY proxies['all'] = req_proxy for proxy_key, proxy_url in proxies.items(): if proxy_url == '__noproxy__': proxies[proxy_key] = None continue if proxy_key == 'no': # special case continue if proxy_url is not None: # Ensure proxies without a scheme are http. try: proxy_scheme = urllib.request._parse_proxy(proxy_url)[0] except ValueError: # Ignore invalid proxy URLs. Sometimes these may be introduced through environment # variables unrelated to proxy settings - e.g. Colab `COLAB_LANGUAGE_SERVER_PROXY`. # If the proxy is going to be used, the Request Handler proxy validation will handle it. continue if proxy_scheme is None: proxies[proxy_key] = 'http://' + remove_start(proxy_url, '//') replace_scheme = { 'socks5': 'socks5h', # compat: socks5 was treated as socks5h 'socks': 'socks4', # compat: non-standard } if proxy_scheme in replace_scheme: proxies[proxy_key] = urllib.parse.urlunparse( urllib.parse.urlparse(proxy_url)._replace(scheme=replace_scheme[proxy_scheme])) def clean_headers(headers: HTTPHeaderDict): if 'Youtubedl-No-Compression' in headers: # compat del headers['Youtubedl-No-Compression'] headers['Accept-Encoding'] = 'identity' headers.pop('Ytdl-socks-proxy', None) def remove_dot_segments(path): # Implements RFC3986 5.2.4 remote_dot_segments # Pseudo-code: https://tools.ietf.org/html/rfc3986#section-5.2.4 # https://github.com/urllib3/urllib3/blob/ba49f5c4e19e6bca6827282feb77a3c9f937e64b/src/urllib3/util/url.py#L263 output = [] segments = path.split('/') for s in segments: if s == '.': continue elif s == '..': if output: output.pop() else: output.append(s) if not segments[0] and (not output or output[0]): output.insert(0, '') if segments[-1] in ('.', '..'): output.append('') return '/'.join(output) def escape_rfc3986(s): """Escape non-ASCII characters as suggested by RFC 3986""" return urllib.parse.quote(s, b"%/;:@&=+$,!~*'()?#[]") def normalize_url(url): """Normalize URL as suggested by RFC 3986""" url_parsed = urllib.parse.urlparse(url) return url_parsed._replace( netloc=url_parsed.netloc.encode('idna').decode('ascii'), path=escape_rfc3986(remove_dot_segments(url_parsed.path)), params=escape_rfc3986(url_parsed.params), query=escape_rfc3986(url_parsed.query), fragment=escape_rfc3986(url_parsed.fragment), ).geturl() def select_proxy(url, proxies): """Unified proxy selector for all backends""" url_components = urllib.parse.urlparse(url) if 'no' in proxies: hostport = url_components.hostname + format_field(url_components.port, None, ':%s') if urllib.request.proxy_bypass_environment(hostport, {'no': proxies['no']}): return elif urllib.request.proxy_bypass(hostport): # check system settings return return traverse_obj(proxies, url_components.scheme or 'http', 'all')
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/utils/_deprecated.py
yt_dlp/utils/_deprecated.py
"""Deprecated - New code should avoid these""" import base64 import hashlib import hmac import json import warnings from ..compat.compat_utils import passthrough_module # XXX: Implement this the same way as other DeprecationWarnings without circular import passthrough_module(__name__, '.._legacy', callback=lambda attr: warnings.warn( DeprecationWarning(f'{__name__}.{attr} is deprecated'), stacklevel=6)) del passthrough_module import re import struct def bytes_to_intlist(bs): if not bs: return [] if isinstance(bs[0], int): # Python 3 return list(bs) else: return [ord(c) for c in bs] def intlist_to_bytes(xs): if not xs: return b'' return struct.pack('%dB' % len(xs), *xs) def jwt_encode_hs256(payload_data, key, headers={}): header_data = { 'alg': 'HS256', 'typ': 'JWT', } if headers: header_data.update(headers) header_b64 = base64.b64encode(json.dumps(header_data).encode()) payload_b64 = base64.b64encode(json.dumps(payload_data).encode()) h = hmac.new(key.encode(), header_b64 + b'.' + payload_b64, hashlib.sha256) signature_b64 = base64.b64encode(h.digest()) return header_b64 + b'.' + payload_b64 + b'.' + signature_b64 compiled_regex_type = type(re.compile(''))
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/utils/progress.py
yt_dlp/utils/progress.py
from __future__ import annotations import bisect import threading import time class ProgressCalculator: # Time to calculate the speed over (seconds) SAMPLING_WINDOW = 3 # Minimum timeframe before to sample next downloaded bytes (seconds) SAMPLING_RATE = 0.05 # Time before showing eta (seconds) GRACE_PERIOD = 1 def __init__(self, initial: int): self._initial = initial or 0 self.downloaded = self._initial self.elapsed: float = 0 self.speed = SmoothValue(0, smoothing=0.7) self.eta = SmoothValue(None, smoothing=0.9) self._total = 0 self._start_time = time.monotonic() self._last_update = self._start_time self._lock = threading.Lock() self._thread_sizes: dict[int, int] = {} self._times = [self._start_time] self._downloaded = [self.downloaded] @property def total(self): return self._total @total.setter def total(self, value: int | None): with self._lock: if value is not None and value < self.downloaded: value = self.downloaded self._total = value def thread_reset(self): current_thread = threading.get_ident() with self._lock: self._thread_sizes[current_thread] = 0 def update(self, size: int | None): if not size: return current_thread = threading.get_ident() with self._lock: last_size = self._thread_sizes.get(current_thread, 0) self._thread_sizes[current_thread] = size self._update(size - last_size) def _update(self, size: int): current_time = time.monotonic() self.downloaded += size self.elapsed = current_time - self._start_time if self.total is not None and self.downloaded > self.total: self._total = self.downloaded if self._last_update + self.SAMPLING_RATE > current_time: return self._last_update = current_time self._times.append(current_time) self._downloaded.append(self.downloaded) offset = bisect.bisect_left(self._times, current_time - self.SAMPLING_WINDOW) del self._times[:offset] del self._downloaded[:offset] if len(self._times) < 2: self.speed.reset() self.eta.reset() return download_time = current_time - self._times[0] if not download_time: return self.speed.set((self.downloaded - self._downloaded[0]) / download_time) if self.total and self.speed.value and self.elapsed > self.GRACE_PERIOD: self.eta.set((self.total - self.downloaded) / self.speed.value) else: self.eta.reset() class SmoothValue: def __init__(self, initial: float | None, smoothing: float): self.value = self.smooth = self._initial = initial self._smoothing = smoothing def set(self, value: float): self.value = value if self.smooth is None: self.smooth = self.value else: self.smooth = (1 - self._smoothing) * value + self._smoothing * self.smooth def reset(self): self.value = self.smooth = self._initial
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/utils/__init__.py
yt_dlp/utils/__init__.py
# flake8: noqa: F403 from ..compat.compat_utils import passthrough_module passthrough_module(__name__, '._deprecated') del passthrough_module # isort: off from .traversal import * from ._utils import * from ._utils import _configuration_args, _get_exe_version_output # noqa: F401
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/utils/_utils.py
yt_dlp/utils/_utils.py
import base64 import binascii import calendar import codecs import collections import collections.abc import contextlib import datetime as dt import email.header import email.utils import enum import errno import functools import hashlib import hmac import html.entities import html.parser import inspect import io import itertools import json import locale import math import mimetypes import netrc import operator import os import platform import random import re import shlex import socket import ssl import struct import subprocess import sys import tempfile import time import traceback import types import unicodedata import urllib.error import urllib.parse import urllib.request import xml.etree.ElementTree from . import traversal from ..compat import ( compat_datetime_from_timestamp, compat_etree_fromstring, compat_expanduser, compat_HTMLParseError, ) from ..dependencies import xattr from ..globals import IN_CLI, WINDOWS_VT_MODE __name__ = __name__.rsplit('.', 1)[0] # noqa: A001 # Pretend to be the parent module class NO_DEFAULT: pass def IDENTITY(x): return x ENGLISH_MONTH_NAMES = [ 'January', 'February', 'March', 'April', 'May', 'June', 'July', 'August', 'September', 'October', 'November', 'December'] MONTH_NAMES = { 'en': ENGLISH_MONTH_NAMES, 'fr': [ 'janvier', 'février', 'mars', 'avril', 'mai', 'juin', 'juillet', 'août', 'septembre', 'octobre', 'novembre', 'décembre'], # these follow the genitive grammatical case (dopełniacz) # some websites might be using nominative, which will require another month list # https://en.wikibooks.org/wiki/Polish/Noun_cases 'pl': ['stycznia', 'lutego', 'marca', 'kwietnia', 'maja', 'czerwca', 'lipca', 'sierpnia', 'września', 'października', 'listopada', 'grudnia'], } # From https://github.com/python/cpython/blob/3.11/Lib/email/_parseaddr.py#L36-L42 TIMEZONE_NAMES = { 'UT': 0, 'UTC': 0, 'GMT': 0, 'Z': 0, 'AST': -4, 'ADT': -3, # Atlantic (used in Canada) 'EST': -5, 'EDT': -4, # Eastern 'CST': -6, 'CDT': -5, # Central 'MST': -7, 'MDT': -6, # Mountain 'PST': -8, 'PDT': -7, # Pacific } # needed for sanitizing filenames in restricted mode ACCENT_CHARS = dict(zip('ÂÃÄÀÁÅÆÇÈÉÊËÌÍÎÏÐÑÒÓÔÕÖŐØŒÙÚÛÜŰÝÞßàáâãäåæçèéêëìíîïðñòóôõöőøœùúûüűýþÿ', itertools.chain('AAAAAA', ['AE'], 'CEEEEIIIIDNOOOOOOO', ['OE'], 'UUUUUY', ['TH', 'ss'], 'aaaaaa', ['ae'], 'ceeeeiiiionooooooo', ['oe'], 'uuuuuy', ['th'], 'y'), strict=True)) DATE_FORMATS = ( '%d %B %Y', '%d %b %Y', '%B %d %Y', '%B %dst %Y', '%B %dnd %Y', '%B %drd %Y', '%B %dth %Y', '%b %d %Y', '%b %dst %Y', '%b %dnd %Y', '%b %drd %Y', '%b %dth %Y', '%b %dst %Y %I:%M', '%b %dnd %Y %I:%M', '%b %drd %Y %I:%M', '%b %dth %Y %I:%M', '%Y %m %d', '%Y-%m-%d', '%Y.%m.%d.', '%Y/%m/%d', '%Y/%m/%d %H:%M', '%Y/%m/%d %H:%M:%S', '%Y%m%d%H%M', '%Y%m%d%H%M%S', '%Y%m%d', '%Y-%m-%d %H:%M', '%Y-%m-%d %H:%M:%S', '%Y-%m-%d %H:%M:%S.%f', '%Y-%m-%d %H:%M:%S:%f', '%d.%m.%Y %H:%M', '%d.%m.%Y %H.%M', '%Y-%m-%dT%H:%M:%SZ', '%Y-%m-%dT%H:%M:%S.%fZ', '%Y-%m-%dT%H:%M:%S.%f0Z', '%Y-%m-%dT%H:%M:%S', '%Y-%m-%dT%H:%M:%S.%f', '%Y-%m-%dT%H:%M', '%b %d %Y at %H:%M', '%b %d %Y at %H:%M:%S', '%B %d %Y at %H:%M', '%B %d %Y at %H:%M:%S', '%H:%M %d-%b-%Y', ) DATE_FORMATS_DAY_FIRST = list(DATE_FORMATS) DATE_FORMATS_DAY_FIRST.extend([ '%d-%m-%Y', '%d.%m.%Y', '%d.%m.%y', '%d/%m/%Y', '%d/%m/%y', '%d/%m/%Y %H:%M:%S', '%d-%m-%Y %H:%M', '%H:%M %d/%m/%Y', ]) DATE_FORMATS_MONTH_FIRST = list(DATE_FORMATS) DATE_FORMATS_MONTH_FIRST.extend([ '%m-%d-%Y', '%m.%d.%Y', '%m/%d/%Y', '%m/%d/%y', '%m/%d/%Y %H:%M:%S', ]) PACKED_CODES_RE = r"}\('(.+)',(\d+),(\d+),'([^']+)'\.split\('\|'\)" JSON_LD_RE = r'(?is)<script[^>]+type=(["\']?)application/ld\+json\1[^>]*>\s*(?P<json_ld>{.+?}|\[.+?\])\s*</script>' NUMBER_RE = r'\d+(?:\.\d+)?' @functools.cache def preferredencoding(): """Get preferred encoding. Returns the best encoding scheme for the system, based on locale.getpreferredencoding() and some further tweaks. """ try: pref = locale.getpreferredencoding() 'TEST'.encode(pref) except Exception: pref = 'UTF-8' return pref def write_json_file(obj, fn): """ Encode obj as JSON and write it to fn, atomically if possible """ tf = tempfile.NamedTemporaryFile( prefix=f'{os.path.basename(fn)}.', dir=os.path.dirname(fn), suffix='.tmp', delete=False, mode='w', encoding='utf-8') try: with tf: json.dump(obj, tf, ensure_ascii=False) if sys.platform == 'win32': # Need to remove existing file on Windows, else os.rename raises # WindowsError or FileExistsError. with contextlib.suppress(OSError): os.unlink(fn) with contextlib.suppress(OSError): mask = os.umask(0) os.umask(mask) os.chmod(tf.name, 0o666 & ~mask) os.rename(tf.name, fn) except Exception: with contextlib.suppress(OSError): os.remove(tf.name) raise def partial_application(func): sig = inspect.signature(func) required_args = [ param.name for param in sig.parameters.values() if param.kind in (inspect.Parameter.POSITIONAL_ONLY, inspect.Parameter.POSITIONAL_OR_KEYWORD) if param.default is inspect.Parameter.empty ] @functools.wraps(func) def wrapped(*args, **kwargs): if set(required_args[len(args):]).difference(kwargs): return functools.partial(func, *args, **kwargs) return func(*args, **kwargs) return wrapped def find_xpath_attr(node, xpath, key, val=None): """ Find the xpath xpath[@key=val] """ assert re.match(r'^[a-zA-Z_-]+$', key) expr = xpath + (f'[@{key}]' if val is None else f"[@{key}='{val}']") return node.find(expr) # On python2.6 the xml.etree.ElementTree.Element methods don't support # the namespace parameter def xpath_with_ns(path, ns_map): components = [c.split(':') for c in path.split('/')] replaced = [] for c in components: if len(c) == 1: replaced.append(c[0]) else: ns, tag = c replaced.append(f'{{{ns_map[ns]}}}{tag}') return '/'.join(replaced) def xpath_element(node, xpath, name=None, fatal=False, default=NO_DEFAULT): def _find_xpath(xpath): return node.find(xpath) if isinstance(xpath, str): n = _find_xpath(xpath) else: for xp in xpath: n = _find_xpath(xp) if n is not None: break if n is None: if default is not NO_DEFAULT: return default elif fatal: name = xpath if name is None else name raise ExtractorError(f'Could not find XML element {name}') else: return None return n def xpath_text(node, xpath, name=None, fatal=False, default=NO_DEFAULT): n = xpath_element(node, xpath, name, fatal=fatal, default=default) if n is None or n == default: return n if n.text is None: if default is not NO_DEFAULT: return default elif fatal: name = xpath if name is None else name raise ExtractorError(f'Could not find XML element\'s text {name}') else: return None return n.text def xpath_attr(node, xpath, key, name=None, fatal=False, default=NO_DEFAULT): n = find_xpath_attr(node, xpath, key) if n is None: if default is not NO_DEFAULT: return default elif fatal: name = f'{xpath}[@{key}]' if name is None else name raise ExtractorError(f'Could not find XML attribute {name}') else: return None return n.attrib[key] def get_element_by_id(id, html, **kwargs): """Return the content of the tag with the specified ID in the passed HTML document""" return get_element_by_attribute('id', id, html, **kwargs) def get_element_html_by_id(id, html, **kwargs): """Return the html of the tag with the specified ID in the passed HTML document""" return get_element_html_by_attribute('id', id, html, **kwargs) def get_element_by_class(class_name, html): """Return the content of the first tag with the specified class in the passed HTML document""" retval = get_elements_by_class(class_name, html) return retval[0] if retval else None def get_element_html_by_class(class_name, html): """Return the html of the first tag with the specified class in the passed HTML document""" retval = get_elements_html_by_class(class_name, html) return retval[0] if retval else None def get_element_by_attribute(attribute, value, html, **kwargs): retval = get_elements_by_attribute(attribute, value, html, **kwargs) return retval[0] if retval else None def get_element_html_by_attribute(attribute, value, html, **kargs): retval = get_elements_html_by_attribute(attribute, value, html, **kargs) return retval[0] if retval else None def get_elements_by_class(class_name, html, **kargs): """Return the content of all tags with the specified class in the passed HTML document as a list""" return get_elements_by_attribute( 'class', rf'[^\'"]*(?<=[\'"\s]){re.escape(class_name)}(?=[\'"\s])[^\'"]*', html, escape_value=False) def get_elements_html_by_class(class_name, html): """Return the html of all tags with the specified class in the passed HTML document as a list""" return get_elements_html_by_attribute( 'class', rf'[^\'"]*(?<=[\'"\s]){re.escape(class_name)}(?=[\'"\s])[^\'"]*', html, escape_value=False) def get_elements_by_attribute(*args, **kwargs): """Return the content of the tag with the specified attribute in the passed HTML document""" return [content for content, _ in get_elements_text_and_html_by_attribute(*args, **kwargs)] def get_elements_html_by_attribute(*args, **kwargs): """Return the html of the tag with the specified attribute in the passed HTML document""" return [whole for _, whole in get_elements_text_and_html_by_attribute(*args, **kwargs)] def get_elements_text_and_html_by_attribute(attribute, value, html, *, tag=r'[\w:.-]+', escape_value=True): """ Return the text (content) and the html (whole) of the tag with the specified attribute in the passed HTML document """ if not value: return quote = '' if re.match(r'''[\s"'`=<>]''', value) else '?' value = re.escape(value) if escape_value else value partial_element_re = rf'''(?x) <(?P<tag>{tag}) (?:\s(?:[^>"']|"[^"]*"|'[^']*')*)? \s{re.escape(attribute)}\s*=\s*(?P<_q>['"]{quote})(?-x:{value})(?P=_q) ''' for m in re.finditer(partial_element_re, html): content, whole = get_element_text_and_html_by_tag(m.group('tag'), html[m.start():]) yield ( unescapeHTML(re.sub(r'^(?P<q>["\'])(?P<content>.*)(?P=q)$', r'\g<content>', content, flags=re.DOTALL)), whole, ) class HTMLBreakOnClosingTagParser(html.parser.HTMLParser): """ HTML parser which raises HTMLBreakOnClosingTagException upon reaching the closing tag for the first opening tag it has encountered, and can be used as a context manager """ class HTMLBreakOnClosingTagException(Exception): pass def __init__(self): self.tagstack = collections.deque() html.parser.HTMLParser.__init__(self) def __enter__(self): return self def __exit__(self, *_): self.close() def close(self): # handle_endtag does not return upon raising HTMLBreakOnClosingTagException, # so data remains buffered; we no longer have any interest in it, thus # override this method to discard it pass def handle_starttag(self, tag, _): self.tagstack.append(tag) def handle_endtag(self, tag): if not self.tagstack: raise compat_HTMLParseError('no tags in the stack') while self.tagstack: inner_tag = self.tagstack.pop() if inner_tag == tag: break else: raise compat_HTMLParseError(f'matching opening tag for closing {tag} tag not found') if not self.tagstack: raise self.HTMLBreakOnClosingTagException # XXX: This should be far less strict def get_element_text_and_html_by_tag(tag, html): """ For the first element with the specified tag in the passed HTML document return its' content (text) and the whole element (html) """ def find_or_raise(haystack, needle, exc): try: return haystack.index(needle) except ValueError: raise exc closing_tag = f'</{tag}>' whole_start = find_or_raise( html, f'<{tag}', compat_HTMLParseError(f'opening {tag} tag not found')) content_start = find_or_raise( html[whole_start:], '>', compat_HTMLParseError(f'malformed opening {tag} tag')) content_start += whole_start + 1 with HTMLBreakOnClosingTagParser() as parser: parser.feed(html[whole_start:content_start]) if not parser.tagstack or parser.tagstack[0] != tag: raise compat_HTMLParseError(f'parser did not match opening {tag} tag') offset = content_start while offset < len(html): next_closing_tag_start = find_or_raise( html[offset:], closing_tag, compat_HTMLParseError(f'closing {tag} tag not found')) next_closing_tag_end = next_closing_tag_start + len(closing_tag) try: parser.feed(html[offset:offset + next_closing_tag_end]) offset += next_closing_tag_end except HTMLBreakOnClosingTagParser.HTMLBreakOnClosingTagException: return html[content_start:offset + next_closing_tag_start], \ html[whole_start:offset + next_closing_tag_end] raise compat_HTMLParseError('unexpected end of html') class HTMLAttributeParser(html.parser.HTMLParser): """Trivial HTML parser to gather the attributes for a single element""" def __init__(self): self.attrs = {} html.parser.HTMLParser.__init__(self) def handle_starttag(self, tag, attrs): self.attrs = dict(attrs) raise compat_HTMLParseError('done') class HTMLListAttrsParser(html.parser.HTMLParser): """HTML parser to gather the attributes for the elements of a list""" def __init__(self): html.parser.HTMLParser.__init__(self) self.items = [] self._level = 0 def handle_starttag(self, tag, attrs): if tag == 'li' and self._level == 0: self.items.append(dict(attrs)) self._level += 1 def handle_endtag(self, tag): self._level -= 1 def extract_attributes(html_element): """Given a string for an HTML element such as <el a="foo" B="bar" c="&98;az" d=boz empty= noval entity="&amp;" sq='"' dq="'" > Decode and return a dictionary of attributes. { 'a': 'foo', 'b': 'bar', c: 'baz', d: 'boz', 'empty': '', 'noval': None, 'entity': '&', 'sq': '"', 'dq': '\'' }. """ parser = HTMLAttributeParser() with contextlib.suppress(compat_HTMLParseError): parser.feed(html_element) parser.close() return parser.attrs def parse_list(webpage): """Given a string for an series of HTML <li> elements, return a dictionary of their attributes""" parser = HTMLListAttrsParser() parser.feed(webpage) parser.close() return parser.items def clean_html(html): """Clean an HTML snippet into a readable string""" if html is None: # Convenience for sanitizing descriptions etc. return html html = re.sub(r'\s+', ' ', html) html = re.sub(r'(?u)\s?<\s?br\s?/?\s?>\s?', '\n', html) html = re.sub(r'(?u)<\s?/\s?p\s?>\s?<\s?p[^>]*>', '\n', html) # Strip html tags html = re.sub('<.*?>', '', html) # Replace html entities html = unescapeHTML(html) return html.strip() class LenientJSONDecoder(json.JSONDecoder): # TODO: Write tests def __init__(self, *args, transform_source=None, ignore_extra=False, close_objects=0, **kwargs): self.transform_source, self.ignore_extra = transform_source, ignore_extra self._close_attempts = 2 * close_objects super().__init__(*args, **kwargs) @staticmethod def _close_object(err): doc = err.doc[:err.pos] # We need to add comma first to get the correct error message if err.msg.startswith('Expecting \',\''): return doc + ',' elif not doc.endswith(','): return if err.msg.startswith('Expecting property name'): return doc[:-1] + '}' elif err.msg.startswith('Expecting value'): return doc[:-1] + ']' def decode(self, s): if self.transform_source: s = self.transform_source(s) for attempt in range(self._close_attempts + 1): try: if self.ignore_extra: return self.raw_decode(s.lstrip())[0] return super().decode(s) except json.JSONDecodeError as e: if e.pos is None: raise elif attempt < self._close_attempts: s = self._close_object(e) if s is not None: continue raise type(e)(f'{e.msg} in {s[e.pos - 10:e.pos + 10]!r}', s, e.pos) assert False, 'Too many attempts to decode JSON' def sanitize_open(filename, open_mode): """Try to open the given filename, and slightly tweak it if this fails. Attempts to open the given filename. If this fails, it tries to change the filename slightly, step by step, until it's either able to open it or it fails and raises a final exception, like the standard open() function. It returns the tuple (stream, definitive_file_name). """ if filename == '-': if sys.platform == 'win32': import msvcrt # stdout may be any IO stream, e.g. when using contextlib.redirect_stdout with contextlib.suppress(io.UnsupportedOperation): msvcrt.setmode(sys.stdout.fileno(), os.O_BINARY) return (sys.stdout.buffer if hasattr(sys.stdout, 'buffer') else sys.stdout, filename) for attempt in range(2): try: try: if sys.platform == 'win32': # FIXME: An exclusive lock also locks the file from being read. # Since windows locks are mandatory, don't lock the file on windows (for now). # Ref: https://github.com/yt-dlp/yt-dlp/issues/3124 raise LockingUnsupportedError stream = locked_file(filename, open_mode, block=False).__enter__() except OSError: stream = open(filename, open_mode) return stream, filename except OSError as err: if attempt or err.errno in (errno.EACCES,): raise old_filename, filename = filename, sanitize_path(filename) if old_filename == filename: raise def timeconvert(timestr): """Convert RFC 2822 defined time string into system timestamp""" timestamp = None timetuple = email.utils.parsedate_tz(timestr) if timetuple is not None: timestamp = email.utils.mktime_tz(timetuple) return timestamp def sanitize_filename(s, restricted=False, is_id=NO_DEFAULT): """Sanitizes a string so it could be used as part of a filename. @param restricted Use a stricter subset of allowed characters @param is_id Whether this is an ID that should be kept unchanged if possible. If unset, yt-dlp's new sanitization rules are in effect """ if s == '': return '' def replace_insane(char): if restricted and char in ACCENT_CHARS: return ACCENT_CHARS[char] elif not restricted and char == '\n': return '\0 ' elif is_id is NO_DEFAULT and not restricted and char in '"*:<>?|/\\': # Replace with their full-width unicode counterparts return {'/': '\u29F8', '\\': '\u29f9'}.get(char, chr(ord(char) + 0xfee0)) elif char == '?' or ord(char) < 32 or ord(char) == 127: return '' elif char == '"': return '' if restricted else '\'' elif char == ':': return '\0_\0-' if restricted else '\0 \0-' elif char in '\\/|*<>': return '\0_' if restricted and (char in '!&\'()[]{}$;`^,#' or char.isspace() or ord(char) > 127): return '' if unicodedata.category(char)[0] in 'CM' else '\0_' return char # Replace look-alike Unicode glyphs if restricted and (is_id is NO_DEFAULT or not is_id): s = unicodedata.normalize('NFKC', s) s = re.sub(r'[0-9]+(?::[0-9]+)+', lambda m: m.group(0).replace(':', '_'), s) # Handle timestamps result = ''.join(map(replace_insane, s)) if is_id is NO_DEFAULT: result = re.sub(r'(\0.)(?:(?=\1)..)+', r'\1', result) # Remove repeated substitute chars STRIP_RE = r'(?:\0.|[ _-])*' result = re.sub(f'^\0.{STRIP_RE}|{STRIP_RE}\0.$', '', result) # Remove substitute chars from start/end result = result.replace('\0', '') or '_' if not is_id: while '__' in result: result = result.replace('__', '_') result = result.strip('_') # Common case of "Foreign band name - English song title" if restricted and result.startswith('-_'): result = result[2:] if result.startswith('-'): result = '_' + result[len('-'):] result = result.lstrip('.') if not result: result = '_' return result def _sanitize_path_parts(parts): sanitized_parts = [] for part in parts: if not part or part == '.': continue elif part == '..': if sanitized_parts and sanitized_parts[-1] != '..': sanitized_parts.pop() else: sanitized_parts.append('..') continue # Replace invalid segments with `#` # - trailing dots and spaces (`asdf...` => `asdf..#`) # - invalid chars (`<>` => `##`) sanitized_part = re.sub(r'[/<>:"\|\\?\*]|[\s.]$', '#', part) sanitized_parts.append(sanitized_part) return sanitized_parts def sanitize_path(s, force=False): """Sanitizes and normalizes path on Windows""" if sys.platform != 'win32': if not force: return s root = '/' if s.startswith('/') else '' path = '/'.join(_sanitize_path_parts(s.split('/'))) return root + path if root or path else '.' normed = s.replace('/', '\\') if normed.startswith('\\\\'): # UNC path (`\\SERVER\SHARE`) or device path (`\\.`, `\\?`) parts = normed.split('\\') root = '\\'.join(parts[:4]) + '\\' parts = parts[4:] elif normed[1:2] == ':': # absolute path or drive relative path offset = 3 if normed[2:3] == '\\' else 2 root = normed[:offset] parts = normed[offset:].split('\\') else: # relative/drive root relative path root = '\\' if normed[:1] == '\\' else '' parts = normed.split('\\') path = '\\'.join(_sanitize_path_parts(parts)) return root + path if root or path else '.' def sanitize_url(url, *, scheme='http'): # Prepend protocol-less URLs with `http:` scheme in order to mitigate # the number of unwanted failures due to missing protocol if url is None: return elif url.startswith('//'): return f'{scheme}:{url}' # Fix some common typos seen so far COMMON_TYPOS = ( # https://github.com/ytdl-org/youtube-dl/issues/15649 (r'^httpss://', r'https://'), # https://bx1.be/lives/direct-tv/ (r'^rmtp([es]?)://', r'rtmp\1://'), ) for mistake, fixup in COMMON_TYPOS: if re.match(mistake, url): return re.sub(mistake, fixup, url) return url def extract_basic_auth(url): parts = urllib.parse.urlsplit(url) if parts.username is None: return url, None url = urllib.parse.urlunsplit(parts._replace(netloc=( parts.hostname if parts.port is None else f'{parts.hostname}:{parts.port}'))) auth_payload = base64.b64encode( ('{}:{}'.format(parts.username, parts.password or '')).encode()) return url, f'Basic {auth_payload.decode()}' def expand_path(s): """Expand shell variables and ~""" return os.path.expandvars(compat_expanduser(s)) def orderedSet(iterable, *, lazy=False): """Remove all duplicates from the input iterable""" def _iter(): seen = [] # Do not use set since the items can be unhashable for x in iterable: if x not in seen: seen.append(x) yield x return _iter() if lazy else list(_iter()) def _htmlentity_transform(entity_with_semicolon): """Transforms an HTML entity to a character.""" entity = entity_with_semicolon[:-1] # Known non-numeric HTML entity if entity in html.entities.name2codepoint: return chr(html.entities.name2codepoint[entity]) # TODO: HTML5 allows entities without a semicolon. # E.g. '&Eacuteric' should be decoded as 'Éric'. if entity_with_semicolon in html.entities.html5: return html.entities.html5[entity_with_semicolon] mobj = re.match(r'#(x[0-9a-fA-F]+|[0-9]+)', entity) if mobj is not None: numstr = mobj.group(1) if numstr.startswith('x'): base = 16 numstr = f'0{numstr}' else: base = 10 # See https://github.com/ytdl-org/youtube-dl/issues/7518 with contextlib.suppress(ValueError): return chr(int(numstr, base)) # Unknown entity in name, return its literal representation return f'&{entity};' def unescapeHTML(s): if s is None: return None assert isinstance(s, str) return re.sub( r'&([^&;]+;)', lambda m: _htmlentity_transform(m.group(1)), s) def escapeHTML(text): return ( text .replace('&', '&amp;') .replace('<', '&lt;') .replace('>', '&gt;') .replace('"', '&quot;') .replace("'", '&#39;') ) class netrc_from_content(netrc.netrc): def __init__(self, content): self.hosts, self.macros = {}, {} with io.StringIO(content) as stream: self._parse('-', stream, False) class Popen(subprocess.Popen): if sys.platform == 'win32': _startupinfo = subprocess.STARTUPINFO() _startupinfo.dwFlags |= subprocess.STARTF_USESHOWWINDOW else: _startupinfo = None @staticmethod def _fix_pyinstaller_issues(env): if not hasattr(sys, '_MEIPASS'): return # Force spawning independent subprocesses for exes bundled with PyInstaller>=6.10 # Ref: https://pyinstaller.org/en/v6.10.0/CHANGES.html#incompatible-changes # https://github.com/yt-dlp/yt-dlp/issues/11259 env['PYINSTALLER_RESET_ENVIRONMENT'] = '1' # Restore LD_LIBRARY_PATH when using PyInstaller # Ref: https://pyinstaller.org/en/v6.10.0/runtime-information.html#ld-library-path-libpath-considerations # https://github.com/yt-dlp/yt-dlp/issues/4573 def _fix(key): orig = env.get(f'{key}_ORIG') if orig is None: env.pop(key, None) else: env[key] = orig _fix('LD_LIBRARY_PATH') # Linux _fix('DYLD_LIBRARY_PATH') # macOS def __init__(self, args, *remaining, env=None, text=False, shell=False, **kwargs): if env is None: env = os.environ.copy() self._fix_pyinstaller_issues(env) self.__text_mode = kwargs.get('encoding') or kwargs.get('errors') or text or kwargs.get('universal_newlines') if text is True: kwargs['universal_newlines'] = True # For 3.6 compatibility kwargs.setdefault('encoding', 'utf-8') kwargs.setdefault('errors', 'replace') if os.name == 'nt' and kwargs.get('executable') is None: # Must apply shell escaping if we are trying to run a batch file # These conditions should be very specific to limit impact if not shell and isinstance(args, list) and args and args[0].lower().endswith(('.bat', '.cmd')): shell = True if shell: if not isinstance(args, str): args = shell_quote(args, shell=True) shell = False # Set variable for `cmd.exe` newline escaping (see `utils.shell_quote`) env['='] = '"^\n\n"' args = f'{self.__comspec()} /Q /S /D /V:OFF /E:ON /C "{args}"' super().__init__(args, *remaining, env=env, shell=shell, **kwargs, startupinfo=self._startupinfo) def __comspec(self): comspec = os.environ.get('ComSpec') or os.path.join( os.environ.get('SystemRoot', ''), 'System32', 'cmd.exe') if os.path.isabs(comspec): return comspec raise FileNotFoundError('shell not found: neither %ComSpec% nor %SystemRoot% is set') def communicate_or_kill(self, *args, **kwargs): try: return self.communicate(*args, **kwargs) except BaseException: # Including KeyboardInterrupt self.kill(timeout=None) raise def kill(self, *, timeout=0): super().kill() if timeout != 0: self.wait(timeout=timeout) @classmethod def run(cls, *args, timeout=None, **kwargs): with cls(*args, **kwargs) as proc: default = '' if proc.__text_mode else b'' stdout, stderr = proc.communicate_or_kill(timeout=timeout) return stdout or default, stderr or default, proc.returncode def encodeArgument(s): # Legacy code that uses byte strings # Uncomment the following line after fixing all post processors # assert isinstance(s, str), 'Internal error: %r should be of type %r, is %r' % (s, str, type(s)) return s if isinstance(s, str) else s.decode('ascii') _timetuple = collections.namedtuple('Time', ('hours', 'minutes', 'seconds', 'milliseconds')) def timetuple_from_msec(msec): secs, msec = divmod(msec, 1000) mins, secs = divmod(secs, 60) hrs, mins = divmod(mins, 60) return _timetuple(hrs, mins, secs, msec) def formatSeconds(secs, delim=':', msec=False): time = timetuple_from_msec(secs * 1000) if time.hours: ret = '%d%s%02d%s%02d' % (time.hours, delim, time.minutes, delim, time.seconds) elif time.minutes: ret = '%d%s%02d' % (time.minutes, delim, time.seconds) else: ret = '%d' % time.seconds return '%s.%03d' % (ret, time.milliseconds) if msec else ret def bug_reports_message(before=';'): from ..update import REPOSITORY msg = (f'please report this issue on https://github.com/{REPOSITORY}/issues?q= , ' 'filling out the appropriate issue template. Confirm you are on the latest version using yt-dlp -U') before = before.rstrip() if not before or before.endswith(('.', '!', '?')): msg = msg[0].title() + msg[1:] return (before + ' ' if before else '') + msg class YoutubeDLError(Exception): """Base exception for YoutubeDL errors.""" msg = None def __init__(self, msg=None): if msg is not None: self.msg = msg elif self.msg is None: self.msg = type(self).__name__ super().__init__(self.msg) class ExtractorError(YoutubeDLError): """Error during info extraction.""" def __init__(self, msg, tb=None, expected=False, cause=None, video_id=None, ie=None): """ tb, if given, is the original traceback (so that it can be printed out). If expected is set, this is a normal error message and most likely not a bug in yt-dlp. """ from ..networking.exceptions import network_exceptions if sys.exc_info()[0] in network_exceptions: expected = True
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
true
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/utils/traversal.py
yt_dlp/utils/traversal.py
from __future__ import annotations import collections import collections.abc import contextlib import functools import http.cookies import inspect import itertools import re import typing import xml.etree.ElementTree from ._utils import ( IDENTITY, NO_DEFAULT, ExtractorError, LazyList, deprecation_warning, get_elements_html_by_class, get_elements_html_by_attribute, get_elements_by_attribute, get_element_by_class, get_element_html_by_attribute, get_element_by_attribute, get_element_html_by_id, get_element_by_id, get_element_html_by_class, get_elements_by_class, get_element_text_and_html_by_tag, is_iterable_like, try_call, url_or_none, variadic, ) def traverse_obj( obj, *paths, default=NO_DEFAULT, expected_type=None, get_all=True, casesense=True, is_user_input=NO_DEFAULT, traverse_string=False): """ Safely traverse nested `dict`s and `Iterable`s >>> obj = [{}, {"key": "value"}] >>> traverse_obj(obj, (1, "key")) 'value' Each of the provided `paths` is tested and the first producing a valid result will be returned. The next path will also be tested if the path branched but no results could be found. Supported values for traversal are `Mapping`, `Iterable`, `re.Match`, `xml.etree.ElementTree` (xpath) and `http.cookies.Morsel`. Unhelpful values (`{}`, `None`) are treated as the absence of a value and discarded. The paths will be wrapped in `variadic`, so that `'key'` is conveniently the same as `('key', )`. The keys in the path can be one of: - `None`: Return the current object. - `set`: Requires the only item in the set to be a type or function, like `{type}`/`{type, type, ...}`/`{func}`. If a `type`, return only values of this type. If a function, returns `func(obj)`. - `str`/`int`: Return `obj[key]`. For `re.Match`, return `obj.group(key)`. - `slice`: Branch out and return all values in `obj[key]`. - `Ellipsis`: Branch out and return a list of all values. - `tuple`/`list`: Branch out and return a list of all matching values. Read as: `[traverse_obj(obj, branch) for branch in branches]`. - `function`: Branch out and return values filtered by the function. Read as: `[value for key, value in obj if function(key, value)]`. For `Iterable`s, `key` is the index of the value. For `re.Match`es, `key` is the group number (0 = full match) as well as additionally any group names, if given. - `dict`: Transform the current object and return a matching dict. Read as: `{key: traverse_obj(obj, path) for key, path in dct.items()}`. - `any`-builtin: Take the first matching object and return it, resetting branching. - `all`-builtin: Take all matching objects and return them as a list, resetting branching. - `filter`-builtin: Return the value if it is truthy, `None` otherwise. `tuple`, `list`, and `dict` all support nested paths and branches. @params paths Paths by which to traverse. @param default Value to return if the paths do not match. If the last key in the path is a `dict`, it will apply to each value inside the dict instead, depth first. Try to avoid if using nested `dict` keys. @param expected_type If a `type`, only accept final values of this type. If any other callable, try to call the function on each result. If the last key in the path is a `dict`, it will apply to each value inside the dict instead, recursively. This does respect branching paths. @param get_all If `False`, return the first matching result, otherwise all matching ones. @param casesense If `False`, consider string dictionary keys as case insensitive. `traverse_string` is only meant to be used by YoutubeDL.prepare_outtmpl and is not part of the API @param traverse_string Whether to traverse into objects as strings. If `True`, any non-compatible object will first be converted into a string and then traversed into. The return value of that path will be a string instead, not respecting any further branching. @returns The result of the object traversal. If successful, `get_all=True`, and the path branches at least once, then a list of results is returned instead. If no `default` is given and the last path branches, a `list` of results is always returned. If a path ends on a `dict` that result will always be a `dict`. """ if is_user_input is not NO_DEFAULT: deprecation_warning('The is_user_input parameter is deprecated and no longer works') casefold = lambda k: k.casefold() if isinstance(k, str) else k if isinstance(expected_type, type): type_test = lambda val: val if isinstance(val, expected_type) else None else: type_test = lambda val: try_call(expected_type or IDENTITY, args=(val,)) def apply_key(key, obj, is_last): branching = False result = None if obj is None and traverse_string: if key is ... or callable(key) or isinstance(key, slice): branching = True result = () elif key is None: result = obj elif isinstance(key, set): item = next(iter(key)) if len(key) > 1 or isinstance(item, type): assert all(isinstance(item, type) for item in key) if isinstance(obj, tuple(key)): result = obj else: result = try_call(item, args=(obj,)) elif isinstance(key, (list, tuple)): branching = True result = itertools.chain.from_iterable( apply_path(obj, branch, is_last)[0] for branch in key) elif key is ...: branching = True if isinstance(obj, http.cookies.Morsel): obj = dict(obj, key=obj.key, value=obj.value) if isinstance(obj, collections.abc.Mapping): result = obj.values() elif is_iterable_like(obj) or isinstance(obj, xml.etree.ElementTree.Element): result = obj elif isinstance(obj, re.Match): result = obj.groups() elif traverse_string: branching = False result = str(obj) else: result = () elif callable(key): branching = True if isinstance(obj, http.cookies.Morsel): obj = dict(obj, key=obj.key, value=obj.value) if isinstance(obj, collections.abc.Mapping): iter_obj = obj.items() elif is_iterable_like(obj) or isinstance(obj, xml.etree.ElementTree.Element): iter_obj = enumerate(obj) elif isinstance(obj, re.Match): iter_obj = itertools.chain( enumerate((obj.group(), *obj.groups())), obj.groupdict().items()) elif traverse_string: branching = False iter_obj = enumerate(str(obj)) else: iter_obj = () result = (v for k, v in iter_obj if try_call(key, args=(k, v))) if not branching: # string traversal result = ''.join(result) elif isinstance(key, dict): iter_obj = ((k, _traverse_obj(obj, v, False, is_last)) for k, v in key.items()) result = { k: v if v is not None else default for k, v in iter_obj if v is not None or default is not NO_DEFAULT } or None elif isinstance(obj, collections.abc.Mapping): if isinstance(obj, http.cookies.Morsel): obj = dict(obj, key=obj.key, value=obj.value) result = (try_call(obj.get, args=(key,)) if casesense or try_call(obj.__contains__, args=(key,)) else next((v for k, v in obj.items() if casefold(k) == key), None)) elif isinstance(obj, re.Match): if isinstance(key, int) or casesense: with contextlib.suppress(IndexError): result = obj.group(key) elif isinstance(key, str): result = next((v for k, v in obj.groupdict().items() if casefold(k) == key), None) elif isinstance(key, (int, slice)): if is_iterable_like(obj, (collections.abc.Sequence, xml.etree.ElementTree.Element)): branching = isinstance(key, slice) with contextlib.suppress(IndexError): result = obj[key] elif traverse_string: with contextlib.suppress(IndexError): result = str(obj)[key] elif isinstance(obj, xml.etree.ElementTree.Element) and isinstance(key, str): xpath, _, special = key.rpartition('/') if not special.startswith('@') and not special.endswith('()'): xpath = key special = None # Allow abbreviations of relative paths, absolute paths error if xpath.startswith('/'): xpath = f'.{xpath}' elif xpath and not xpath.startswith('./'): xpath = f'./{xpath}' def apply_specials(element): if special is None: return element if special == '@': return element.attrib if special.startswith('@'): return try_call(element.attrib.get, args=(special[1:],)) if special == 'text()': return element.text raise SyntaxError(f'apply_specials is missing case for {special!r}') if xpath: result = list(map(apply_specials, obj.iterfind(xpath))) else: result = apply_specials(obj) return branching, result if branching else (result,) def lazy_last(iterable): iterator = iter(iterable) prev = next(iterator, NO_DEFAULT) if prev is NO_DEFAULT: return for item in iterator: yield False, prev prev = item yield True, prev def apply_path(start_obj, path, test_type): objs = (start_obj,) has_branched = False key = None for last, key in lazy_last(variadic(path, (str, bytes, dict, set))): if not casesense and isinstance(key, str): key = key.casefold() if key in (any, all): has_branched = False filtered_objs = (obj for obj in objs if obj not in (None, {})) if key is any: objs = (next(filtered_objs, None),) else: objs = (list(filtered_objs),) continue if key is filter: objs = filter(None, objs) continue if __debug__ and callable(key): # Verify function signature inspect.signature(key).bind(None, None) new_objs = [] for obj in objs: branching, results = apply_key(key, obj, last) has_branched |= branching new_objs.append(results) objs = itertools.chain.from_iterable(new_objs) if test_type and not isinstance(key, (dict, list, tuple)): objs = map(type_test, objs) return objs, has_branched, isinstance(key, dict) def _traverse_obj(obj, path, allow_empty, test_type): results, has_branched, is_dict = apply_path(obj, path, test_type) results = LazyList(item for item in results if item not in (None, {})) if get_all and has_branched: if results: return results.exhaust() if allow_empty: return [] if default is NO_DEFAULT else default return None return results[0] if results else {} if allow_empty and is_dict else None for index, path in enumerate(paths, 1): is_last = index == len(paths) try: result = _traverse_obj(obj, path, is_last, True) if result is not None: return result except _RequiredError as e: if is_last: # Reraise to get cleaner stack trace raise ExtractorError(e.orig_msg, expected=e.expected) from None return None if default is NO_DEFAULT else default def value(value, /): return lambda _: value def require(name, /, *, expected=False): def func(value): if value is None: raise _RequiredError(f'Unable to extract {name}', expected=expected) return value return func class _RequiredError(ExtractorError): pass @typing.overload def subs_list_to_dict(*, lang: str | None = 'und', ext: str | None = None) -> collections.abc.Callable[[list[dict]], dict[str, list[dict]]]: ... @typing.overload def subs_list_to_dict(subs: list[dict] | None, /, *, lang: str | None = 'und', ext: str | None = None) -> dict[str, list[dict]]: ... def subs_list_to_dict(subs: list[dict] | None = None, /, *, lang='und', ext=None): """ Convert subtitles from a traversal into a subtitle dict. The path should have an `all` immediately before this function. Arguments: `ext` The default value for `ext` in the subtitle dict In the dict you can set the following additional items: `id` The subtitle id to sort the dict into `quality` The sort order for each subtitle """ if subs is None: return functools.partial(subs_list_to_dict, lang=lang, ext=ext) result = collections.defaultdict(list) for sub in subs: if not url_or_none(sub.get('url')) and not sub.get('data'): continue sub_id = sub.pop('id', None) if not isinstance(sub_id, str): if not lang: continue sub_id = lang sub_ext = sub.get('ext') if not isinstance(sub_ext, str): if not ext: sub.pop('ext', None) else: sub['ext'] = ext result[sub_id].append(sub) result = dict(result) for subs in result.values(): subs.sort(key=lambda x: x.pop('quality', 0) or 0) return result @typing.overload def find_element(*, attr: str, value: str, tag: str | None = None, html=False, regex=False): ... @typing.overload def find_element(*, cls: str, html=False): ... @typing.overload def find_element(*, id: str, tag: str | None = None, html=False, regex=False): ... @typing.overload def find_element(*, tag: str, html=False, regex=False): ... def find_element(*, tag=None, id=None, cls=None, attr=None, value=None, html=False, regex=False): # deliberately using `id=` and `cls=` for ease of readability assert tag or id or cls or (attr and value), 'One of tag, id, cls or (attr AND value) is required' ANY_TAG = r'[\w:.-]+' if attr and value: assert not cls, 'Cannot match both attr and cls' assert not id, 'Cannot match both attr and id' func = get_element_html_by_attribute if html else get_element_by_attribute return functools.partial(func, attr, value, tag=tag or ANY_TAG, escape_value=not regex) elif cls: assert not id, 'Cannot match both cls and id' assert tag is None, 'Cannot match both cls and tag' assert not regex, 'Cannot use regex with cls' func = get_element_html_by_class if html else get_element_by_class return functools.partial(func, cls) elif id: func = get_element_html_by_id if html else get_element_by_id return functools.partial(func, id, tag=tag or ANY_TAG, escape_value=not regex) index = int(bool(html)) return lambda html: get_element_text_and_html_by_tag(tag, html)[index] @typing.overload def find_elements(*, cls: str, html=False): ... @typing.overload def find_elements(*, attr: str, value: str, tag: str | None = None, html=False, regex=False): ... def find_elements(*, tag=None, cls=None, attr=None, value=None, html=False, regex=False): # deliberately using `cls=` for ease of readability assert cls or (attr and value), 'One of cls or (attr AND value) is required' if attr and value: assert not cls, 'Cannot match both attr and cls' func = get_elements_html_by_attribute if html else get_elements_by_attribute return functools.partial(func, attr, value, tag=tag or r'[\w:.-]+', escape_value=not regex) assert not tag, 'Cannot match both cls and tag' assert not regex, 'Cannot use regex with cls' func = get_elements_html_by_class if html else get_elements_by_class return functools.partial(func, cls) def trim_str(*, start=None, end=None): def trim(s): if s is None: return None start_idx = 0 if start and s.startswith(start): start_idx = len(start) if end and s.endswith(end): return s[start_idx:-len(end)] return s[start_idx:] return trim def unpack(func, **kwargs): @functools.wraps(func) def inner(items): return func(*items, **kwargs) return inner def get_first(obj, *paths, **kwargs): return traverse_obj(obj, *((..., *variadic(keys)) for keys in paths), **kwargs, get_all=False) def dict_get(d, key_or_keys, default=None, skip_false_values=True): for val in map(d.get, variadic(key_or_keys)): if val is not None and (val or not skip_false_values): return val return default
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/utils/jslib/devalue.py
yt_dlp/utils/jslib/devalue.py
from __future__ import annotations import array import base64 import datetime as dt import math import re from .._utils import parse_iso8601 TYPE_CHECKING = False if TYPE_CHECKING: import collections.abc import typing T = typing.TypeVar('T') _ARRAY_TYPE_LOOKUP = { 'Int8Array': 'b', 'Uint8Array': 'B', 'Uint8ClampedArray': 'B', 'Int16Array': 'h', 'Uint16Array': 'H', 'Int32Array': 'i', 'Uint32Array': 'I', 'Float32Array': 'f', 'Float64Array': 'd', 'BigInt64Array': 'l', 'BigUint64Array': 'L', 'ArrayBuffer': 'B', } def parse_iter(parsed: typing.Any, /, *, revivers: dict[str, collections.abc.Callable[[list], typing.Any]] | None = None): # based on https://github.com/Rich-Harris/devalue/blob/f3fd2aa93d79f21746555671f955a897335edb1b/src/parse.js resolved = { -1: None, -2: None, -3: math.nan, -4: math.inf, -5: -math.inf, -6: -0.0, } if isinstance(parsed, int) and not isinstance(parsed, bool): if parsed not in resolved or parsed == -2: raise ValueError('invalid integer input') return resolved[parsed] elif not isinstance(parsed, list): raise ValueError('expected int or list as input') elif not parsed: raise ValueError('expected a non-empty list as input') if revivers is None: revivers = {} return_value = [None] stack: list[tuple] = [(return_value, 0, 0)] while stack: target, index, source = stack.pop() if isinstance(source, tuple): name, source, reviver = source try: resolved[source] = target[index] = reviver(target[index]) except Exception as error: yield TypeError(f'failed to parse {source} as {name!r}: {error}') resolved[source] = target[index] = None continue if source in resolved: target[index] = resolved[source] continue # guard against Python negative indexing if source < 0: yield IndexError(f'invalid index: {source!r}') continue try: value = parsed[source] except IndexError as error: yield error continue if isinstance(value, list): if value and isinstance(value[0], str): # TODO: implement zips `strict=True` if reviver := revivers.get(value[0]): if value[1] == source: # XXX: avoid infinite loop yield IndexError(f'{value[0]!r} cannot point to itself (index: {source})') continue # inverse order: resolve index, revive value stack.append((target, index, (value[0], value[1], reviver))) stack.append((target, index, value[1])) continue elif value[0] == 'Date': try: result = dt.datetime.fromtimestamp(parse_iso8601(value[1]), tz=dt.timezone.utc) except Exception: yield ValueError(f'invalid date: {value[1]!r}') result = None elif value[0] == 'Set': result = [None] * (len(value) - 1) for offset, new_source in enumerate(value[1:]): stack.append((result, offset, new_source)) elif value[0] == 'Map': result = [] for key, new_source in zip(*(iter(value[1:]),) * 2, strict=True): pair = [None, None] stack.append((pair, 0, key)) stack.append((pair, 1, new_source)) result.append(pair) elif value[0] == 'RegExp': # XXX: use jsinterp to translate regex flags # currently ignores `value[2]` result = re.compile(value[1]) elif value[0] == 'Object': result = value[1] elif value[0] == 'BigInt': result = int(value[1]) elif value[0] == 'null': result = {} for key, new_source in zip(*(iter(value[1:]),) * 2, strict=True): stack.append((result, key, new_source)) elif value[0] in _ARRAY_TYPE_LOOKUP: typecode = _ARRAY_TYPE_LOOKUP[value[0]] data = base64.b64decode(value[1]) result = array.array(typecode, data).tolist() else: yield TypeError(f'invalid type at {source}: {value[0]!r}') result = None else: result = len(value) * [None] for offset, new_source in enumerate(value): stack.append((result, offset, new_source)) elif isinstance(value, dict): result = {} for key, new_source in value.items(): stack.append((result, key, new_source)) else: result = value target[index] = resolved[source] = result return return_value[0] def parse(parsed: typing.Any, /, *, revivers: dict[str, collections.abc.Callable[[typing.Any], typing.Any]] | None = None): generator = parse_iter(parsed, revivers=revivers) while True: try: raise generator.send(None) except StopIteration as error: return error.value
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/utils/jslib/__init__.py
yt_dlp/utils/jslib/__init__.py
# Utility functions for handling web input based on commonly used JavaScript libraries
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/networking/_urllib.py
yt_dlp/networking/_urllib.py
from __future__ import annotations import functools import http.client import io import ssl import urllib.error import urllib.parse import urllib.request import urllib.response import zlib from urllib.request import ( DataHandler, FileHandler, FTPHandler, HTTPCookieProcessor, HTTPDefaultErrorHandler, HTTPErrorProcessor, UnknownHandler, ) from ._helper import ( InstanceStoreMixin, add_accept_encoding_header, create_connection, create_socks_proxy_socket, get_redirect_method, make_socks_proxy_opts, ) from .common import Features, RequestHandler, Response, register_rh from .exceptions import ( CertificateVerifyError, HTTPError, IncompleteRead, ProxyError, RequestError, SSLError, TransportError, ) from ..dependencies import brotli from ..socks import ProxyError as SocksProxyError from ..utils import update_url_query from ..utils.networking import normalize_url, select_proxy SUPPORTED_ENCODINGS = ['gzip', 'deflate'] CONTENT_DECODE_ERRORS = [zlib.error, OSError] if brotli: SUPPORTED_ENCODINGS.append('br') CONTENT_DECODE_ERRORS.append(brotli.error) def _create_http_connection(http_class, source_address, *args, **kwargs): hc = http_class(*args, **kwargs) if hasattr(hc, '_create_connection'): hc._create_connection = create_connection if source_address is not None: hc.source_address = (source_address, 0) return hc class HTTPHandler(urllib.request.AbstractHTTPHandler): """Handler for HTTP requests and responses. This class, when installed with an OpenerDirector, automatically adds the standard headers to every HTTP request and handles gzipped, deflated and brotli responses from web servers. Part of this code was copied from: http://techknack.net/python-urllib2-handlers/ Andrew Rowls, the author of that code, agreed to release it to the public domain. """ def __init__(self, context=None, source_address=None, *args, **kwargs): super().__init__(*args, **kwargs) self._source_address = source_address self._context = context @staticmethod def _make_conn_class(base, req): conn_class = base socks_proxy = req.headers.pop('Ytdl-socks-proxy', None) if socks_proxy: conn_class = make_socks_conn_class(conn_class, socks_proxy) return conn_class def http_open(self, req): conn_class = self._make_conn_class(http.client.HTTPConnection, req) return self.do_open(functools.partial( _create_http_connection, conn_class, self._source_address), req) def https_open(self, req): conn_class = self._make_conn_class(http.client.HTTPSConnection, req) return self.do_open( functools.partial( _create_http_connection, conn_class, self._source_address), req, context=self._context) @staticmethod def deflate(data): if not data: return data try: return zlib.decompress(data, -zlib.MAX_WBITS) except zlib.error: return zlib.decompress(data) @staticmethod def brotli(data): if not data: return data return brotli.decompress(data) @staticmethod def gz(data): # There may be junk added the end of the file # We ignore it by only ever decoding a single gzip payload if not data: return data return zlib.decompress(data, wbits=zlib.MAX_WBITS | 16) def http_request(self, req): # According to RFC 3986, URLs can not contain non-ASCII characters, however this is not # always respected by websites, some tend to give out URLs with non percent-encoded # non-ASCII characters (see telemb.py, ard.py [#3412]) # urllib chokes on URLs with non-ASCII characters (see http://bugs.python.org/issue3991) # To work around aforementioned issue we will replace request's original URL with # percent-encoded one # Since redirects are also affected (e.g. http://www.southpark.de/alle-episoden/s18e09) # the code of this workaround has been moved here from YoutubeDL.urlopen() url = req.get_full_url() url_escaped = normalize_url(url) # Substitute URL if any change after escaping if url != url_escaped: req = update_Request(req, url=url_escaped) return super().do_request_(req) def http_response(self, req, resp): old_resp = resp # Content-Encoding header lists the encodings in order that they were applied [1]. # To decompress, we simply do the reverse. # [1]: https://datatracker.ietf.org/doc/html/rfc9110#name-content-encoding decoded_response = None for encoding in (e.strip() for e in reversed(resp.headers.get('Content-encoding', '').split(','))): if encoding == 'gzip': decoded_response = self.gz(decoded_response or resp.read()) elif encoding == 'deflate': decoded_response = self.deflate(decoded_response or resp.read()) elif encoding == 'br' and brotli: decoded_response = self.brotli(decoded_response or resp.read()) if decoded_response is not None: resp = urllib.request.addinfourl(io.BytesIO(decoded_response), old_resp.headers, old_resp.url, old_resp.code) resp.msg = old_resp.msg # Percent-encode redirect URL of Location HTTP header to satisfy RFC 3986 (see # https://github.com/ytdl-org/youtube-dl/issues/6457). if 300 <= resp.code < 400: location = resp.headers.get('Location') if location: # As of RFC 2616 default charset is iso-8859-1 that is respected by Python 3 location = location.encode('iso-8859-1').decode() location_escaped = normalize_url(location) if location != location_escaped: del resp.headers['Location'] resp.headers['Location'] = location_escaped return resp https_request = http_request https_response = http_response def make_socks_conn_class(base_class, socks_proxy): assert issubclass(base_class, ( http.client.HTTPConnection, http.client.HTTPSConnection)) proxy_args = make_socks_proxy_opts(socks_proxy) class SocksConnection(base_class): _create_connection = create_connection def connect(self): self.sock = create_connection( (proxy_args['addr'], proxy_args['port']), timeout=self.timeout, source_address=self.source_address, _create_socket_func=functools.partial( create_socks_proxy_socket, (self.host, self.port), proxy_args)) if isinstance(self, http.client.HTTPSConnection): self.sock = self._context.wrap_socket(self.sock, server_hostname=self.host) return SocksConnection class RedirectHandler(urllib.request.HTTPRedirectHandler): """YoutubeDL redirect handler The code is based on HTTPRedirectHandler implementation from CPython [1]. This redirect handler fixes and improves the logic to better align with RFC7261 and what browsers tend to do [2][3] 1. https://github.com/python/cpython/blob/master/Lib/urllib/request.py 2. https://datatracker.ietf.org/doc/html/rfc7231 3. https://github.com/python/cpython/issues/91306 """ http_error_301 = http_error_303 = http_error_307 = http_error_308 = urllib.request.HTTPRedirectHandler.http_error_302 def redirect_request(self, req, fp, code, msg, headers, newurl): if code not in (301, 302, 303, 307, 308): raise urllib.error.HTTPError(req.full_url, code, msg, headers, fp) new_data = req.data # Technically the Cookie header should be in unredirected_hdrs, # however in practice some may set it in normal headers anyway. # We will remove it here to prevent any leaks. remove_headers = ['Cookie'] new_method = get_redirect_method(req.get_method(), code) # only remove payload if method changed (e.g. POST to GET) if new_method != req.get_method(): new_data = None remove_headers.extend(['Content-Length', 'Content-Type']) new_headers = {k: v for k, v in req.headers.items() if k.title() not in remove_headers} return urllib.request.Request( newurl, headers=new_headers, origin_req_host=req.origin_req_host, unverifiable=True, method=new_method, data=new_data) class ProxyHandler(urllib.request.BaseHandler): handler_order = 100 def __init__(self, proxies=None): self.proxies = proxies # Set default handlers for scheme in ('http', 'https', 'ftp'): setattr(self, f'{scheme}_open', lambda r, meth=self.proxy_open: meth(r)) def proxy_open(self, req): proxy = select_proxy(req.get_full_url(), self.proxies) if proxy is None: return if urllib.parse.urlparse(proxy).scheme.lower() in ('socks4', 'socks4a', 'socks5', 'socks5h'): req.add_header('Ytdl-socks-proxy', proxy) # yt-dlp's http/https handlers do wrapping the socket with socks return None return urllib.request.ProxyHandler.proxy_open( self, req, proxy, None) class PUTRequest(urllib.request.Request): def get_method(self): return 'PUT' class HEADRequest(urllib.request.Request): def get_method(self): return 'HEAD' def update_Request(req, url=None, data=None, headers=None, query=None): req_headers = req.headers.copy() req_headers.update(headers or {}) req_data = data if data is not None else req.data req_url = update_url_query(url or req.get_full_url(), query) req_get_method = req.get_method() if req_get_method == 'HEAD': req_type = HEADRequest elif req_get_method == 'PUT': req_type = PUTRequest else: req_type = urllib.request.Request new_req = req_type( req_url, data=req_data, headers=req_headers, origin_req_host=req.origin_req_host, unverifiable=req.unverifiable) if hasattr(req, 'timeout'): new_req.timeout = req.timeout return new_req class UrllibResponseAdapter(Response): """ HTTP Response adapter class for urllib addinfourl and http.client.HTTPResponse """ def __init__(self, res: http.client.HTTPResponse | urllib.response.addinfourl): # addinfourl: In Python 3.9+, .status was introduced and .getcode() was deprecated [1] # HTTPResponse: .getcode() was deprecated, .status always existed [2] # 1. https://docs.python.org/3/library/urllib.request.html#urllib.response.addinfourl.getcode # 2. https://docs.python.org/3.10/library/http.client.html#http.client.HTTPResponse.status super().__init__( fp=res, headers=res.headers, url=res.url, status=getattr(res, 'status', None) or res.getcode(), reason=getattr(res, 'reason', None)) def read(self, amt=None): if self.closed: return b'' try: data = self.fp.read(amt) underlying = getattr(self.fp, 'fp', None) if isinstance(self.fp, http.client.HTTPResponse) and underlying is None: # http.client.HTTPResponse automatically closes itself when fully read self.close() elif isinstance(self.fp, urllib.response.addinfourl) and underlying is not None: # urllib's addinfourl does not close the underlying fp automatically when fully read if isinstance(underlying, io.BytesIO): # data URLs or in-memory responses (e.g. gzip/deflate/brotli decoded) if underlying.tell() >= len(underlying.getbuffer()): self.close() elif isinstance(underlying, io.BufferedReader) and amt is None: # file URLs. # XXX: this will not mark the response as closed if it was fully read with amt. self.close() elif underlying is not None and underlying.closed: # Catch-all for any cases where underlying file is closed self.close() return data except Exception as e: handle_response_read_exceptions(e) raise e def handle_sslerror(e: ssl.SSLError): if not isinstance(e, ssl.SSLError): return if isinstance(e, ssl.SSLCertVerificationError): raise CertificateVerifyError(cause=e) from e raise SSLError(cause=e) from e def handle_response_read_exceptions(e): if isinstance(e, http.client.IncompleteRead): raise IncompleteRead(partial=len(e.partial), cause=e, expected=e.expected) from e elif isinstance(e, ssl.SSLError): handle_sslerror(e) elif isinstance(e, (OSError, EOFError, http.client.HTTPException, *CONTENT_DECODE_ERRORS)): # OSErrors raised here should mostly be network related raise TransportError(cause=e) from e @register_rh class UrllibRH(RequestHandler, InstanceStoreMixin): _SUPPORTED_URL_SCHEMES = ('http', 'https', 'data', 'ftp') _SUPPORTED_PROXY_SCHEMES = ('http', 'socks4', 'socks4a', 'socks5', 'socks5h') _SUPPORTED_FEATURES = (Features.NO_PROXY, Features.ALL_PROXY) RH_NAME = 'urllib' def __init__(self, *, enable_file_urls: bool = False, **kwargs): super().__init__(**kwargs) self.enable_file_urls = enable_file_urls if self.enable_file_urls: self._SUPPORTED_URL_SCHEMES = (*self._SUPPORTED_URL_SCHEMES, 'file') def _check_extensions(self, extensions): super()._check_extensions(extensions) extensions.pop('cookiejar', None) extensions.pop('timeout', None) extensions.pop('legacy_ssl', None) def _create_instance(self, proxies, cookiejar, legacy_ssl_support=None): opener = urllib.request.OpenerDirector() handlers = [ ProxyHandler(proxies), HTTPHandler( debuglevel=int(bool(self.verbose)), context=self._make_sslcontext(legacy_ssl_support=legacy_ssl_support), source_address=self.source_address), HTTPCookieProcessor(cookiejar), DataHandler(), UnknownHandler(), HTTPDefaultErrorHandler(), FTPHandler(), HTTPErrorProcessor(), RedirectHandler(), ] if self.enable_file_urls: handlers.append(FileHandler()) for handler in handlers: opener.add_handler(handler) # Delete the default user-agent header, which would otherwise apply in # cases where our custom HTTP handler doesn't come into play # (See https://github.com/ytdl-org/youtube-dl/issues/1309 for details) opener.addheaders = [] return opener def _prepare_headers(self, _, headers): add_accept_encoding_header(headers, SUPPORTED_ENCODINGS) def _send(self, request): headers = self._get_headers(request) urllib_req = urllib.request.Request( url=request.url, data=request.data, headers=headers, method=request.method, ) opener = self._get_instance( proxies=self._get_proxies(request), cookiejar=self._get_cookiejar(request), legacy_ssl_support=request.extensions.get('legacy_ssl'), ) try: res = opener.open(urllib_req, timeout=self._calculate_timeout(request)) except urllib.error.HTTPError as e: if isinstance(e.fp, (http.client.HTTPResponse, urllib.response.addinfourl)): # Prevent file object from being closed when urllib.error.HTTPError is destroyed. e._closer.close_called = True raise HTTPError(UrllibResponseAdapter(e.fp), redirect_loop='redirect error' in str(e)) from e raise # unexpected except urllib.error.URLError as e: cause = e.reason # NOTE: cause may be a string # proxy errors if 'tunnel connection failed' in str(cause).lower() or isinstance(cause, SocksProxyError): raise ProxyError(cause=e) from e handle_response_read_exceptions(cause) raise TransportError(cause=e) from e except (http.client.InvalidURL, ValueError) as e: # Validation errors # http.client.HTTPConnection raises ValueError in some validation cases # such as if request method contains illegal control characters [1] # 1. https://github.com/python/cpython/blob/987b712b4aeeece336eed24fcc87a950a756c3e2/Lib/http/client.py#L1256 raise RequestError(cause=e) from e except Exception as e: handle_response_read_exceptions(e) raise # unexpected return UrllibResponseAdapter(res)
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/networking/impersonate.py
yt_dlp/networking/impersonate.py
from __future__ import annotations import re from abc import ABC from dataclasses import dataclass from types import NoneType from typing import Any from .common import RequestHandler, register_preference, Request from .exceptions import UnsupportedRequest from ..utils import classproperty, join_nonempty from ..utils.networking import std_headers, HTTPHeaderDict @dataclass(order=True, frozen=True) class ImpersonateTarget: """ A target for browser impersonation. Parameters: @param client: the client to impersonate @param version: the client version to impersonate @param os: the client OS to impersonate @param os_version: the client OS version to impersonate Note: None is used to indicate to match any. """ client: str | None = None version: str | None = None os: str | None = None os_version: str | None = None def __post_init__(self): if self.version and not self.client: raise ValueError('client is required if version is set') if self.os_version and not self.os: raise ValueError('os is required if os_version is set') def __contains__(self, target: ImpersonateTarget): if not isinstance(target, ImpersonateTarget): return False return ( (self.client is None or target.client is None or self.client == target.client) and (self.version is None or target.version is None or self.version == target.version) and (self.os is None or target.os is None or self.os == target.os) and (self.os_version is None or target.os_version is None or self.os_version == target.os_version) ) def __str__(self): return f'{join_nonempty(self.client, self.version)}:{join_nonempty(self.os, self.os_version)}'.rstrip(':') @classmethod def from_str(cls, target: str): mobj = re.fullmatch(r'(?:(?P<client>[^:-]+)(?:-(?P<version>[^:-]+))?)?(?::(?:(?P<os>[^:-]+)(?:-(?P<os_version>[^:-]+))?)?)?', target) if not mobj: raise ValueError(f'Invalid impersonate target "{target}"') return cls(**mobj.groupdict()) class ImpersonateRequestHandler(RequestHandler, ABC): """ Base class for request handlers that support browser impersonation. This provides a method for checking the validity of the impersonate extension, which can be used in _check_extensions. Impersonate targets consist of a client, version, os and os_ver. See the ImpersonateTarget class for more details. The following may be defined: - `_SUPPORTED_IMPERSONATE_TARGET_MAP`: a dict mapping supported targets to custom object. Any Request with an impersonate target not in this list will raise an UnsupportedRequest. Set to None to disable this check. Note: Entries are in order of preference Parameters: @param impersonate: the default impersonate target to use for requests. Set to None to disable impersonation. """ _SUPPORTED_IMPERSONATE_TARGET_MAP: dict[ImpersonateTarget, Any] = {} def __init__(self, *, impersonate: ImpersonateTarget = None, **kwargs): super().__init__(**kwargs) self.impersonate = impersonate def _check_impersonate_target(self, target: ImpersonateTarget): assert isinstance(target, (ImpersonateTarget, NoneType)) if target is None or not self.supported_targets: return if not self.is_supported_target(target): raise UnsupportedRequest(f'Unsupported impersonate target: {target}') def _check_extensions(self, extensions): super()._check_extensions(extensions) if 'impersonate' in extensions: self._check_impersonate_target(extensions.get('impersonate')) def _validate(self, request): super()._validate(request) self._check_impersonate_target(self.impersonate) def _resolve_target(self, target: ImpersonateTarget | None): """Resolve a target to a supported target.""" if target is None: return for supported_target in self.supported_targets: if target in supported_target: if self.verbose: self._logger.stdout( f'{self.RH_NAME}: resolved impersonate target {target} to {supported_target}') return supported_target @classproperty def supported_targets(cls) -> tuple[ImpersonateTarget, ...]: return tuple(cls._SUPPORTED_IMPERSONATE_TARGET_MAP.keys()) def is_supported_target(self, target: ImpersonateTarget): assert isinstance(target, ImpersonateTarget) return self._resolve_target(target) is not None def _get_request_target(self, request): """Get the requested target for the request""" return self._resolve_target(request.extensions.get('impersonate') or self.impersonate) def _prepare_impersonate_headers(self, request: Request, headers: HTTPHeaderDict) -> None: # noqa: B027 """Additional operations to prepare headers before building. To be extended by subclasses. @param request: Request object @param headers: Merged headers to prepare """ def _get_impersonate_headers(self, request: Request) -> dict[str, str]: """ Get headers for external impersonation use. Subclasses may define a _prepare_impersonate_headers method to modify headers after merge but before building. """ headers = self._merge_headers(request.headers) if self._get_request_target(request) is not None: # remove all headers present in std_headers # TODO: change this to not depend on std_headers for k, v in std_headers.items(): if headers.get(k) == v: headers.pop(k) self._prepare_impersonate_headers(request, headers) if request.extensions.get('keep_header_casing'): return headers.sensitive() return dict(headers) @register_preference(ImpersonateRequestHandler) def impersonate_preference(rh, request): if request.extensions.get('impersonate') or rh.impersonate: return 1000 return 0
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/networking/common.py
yt_dlp/networking/common.py
from __future__ import annotations import abc import copy import enum import functools import io import typing import urllib.parse import urllib.request import urllib.response from collections.abc import Iterable, Mapping from email.message import Message from http import HTTPStatus from types import NoneType from ._helper import make_ssl_context, wrap_request_errors from .exceptions import ( NoSupportingHandlers, RequestError, TransportError, UnsupportedRequest, ) from ..cookies import YoutubeDLCookieJar from ..utils import ( bug_reports_message, classproperty, deprecation_warning, error_to_str, update_url_query, ) from ..utils.networking import HTTPHeaderDict, normalize_url DEFAULT_TIMEOUT = 20 def register_preference(*handlers: type[RequestHandler]): assert all(issubclass(handler, RequestHandler) for handler in handlers) def outer(preference: Preference): @functools.wraps(preference) def inner(handler, *args, **kwargs): if not handlers or isinstance(handler, handlers): return preference(handler, *args, **kwargs) return 0 _RH_PREFERENCES.add(inner) return inner return outer class RequestDirector: """RequestDirector class Helper class that, when given a request, forward it to a RequestHandler that supports it. Preference functions in the form of func(handler, request) -> int can be registered into the `preferences` set. These are used to sort handlers in order of preference. @param logger: Logger instance. @param verbose: Print debug request information to stdout. """ def __init__(self, logger, verbose=False): self.handlers: dict[str, RequestHandler] = {} self.preferences: set[Preference] = set() self.logger = logger # TODO(Grub4k): default logger self.verbose = verbose def close(self): for handler in self.handlers.values(): handler.close() self.handlers.clear() def add_handler(self, handler: RequestHandler): """Add a handler. If a handler of the same RH_KEY exists, it will overwrite it""" assert isinstance(handler, RequestHandler), 'handler must be a RequestHandler' self.handlers[handler.RH_KEY] = handler def _get_handlers(self, request: Request) -> list[RequestHandler]: """Sorts handlers by preference, given a request""" preferences = { rh: sum(pref(rh, request) for pref in self.preferences) for rh in self.handlers.values() } self._print_verbose('Handler preferences for this request: {}'.format(', '.join( f'{rh.RH_NAME}={pref}' for rh, pref in preferences.items()))) return sorted(self.handlers.values(), key=preferences.get, reverse=True) def _print_verbose(self, msg): if self.verbose: self.logger.stdout(f'director: {msg}') def send(self, request: Request) -> Response: """ Passes a request onto a suitable RequestHandler """ if not self.handlers: raise RequestError('No request handlers configured') assert isinstance(request, Request) unexpected_errors = [] unsupported_errors = [] for handler in self._get_handlers(request): self._print_verbose(f'Checking if "{handler.RH_NAME}" supports this request.') try: handler.validate(request) except UnsupportedRequest as e: self._print_verbose( f'"{handler.RH_NAME}" cannot handle this request (reason: {error_to_str(e)})') unsupported_errors.append(e) continue self._print_verbose(f'Sending request via "{handler.RH_NAME}"') try: response = handler.send(request) except RequestError: raise except Exception as e: self.logger.error( f'[{handler.RH_NAME}] Unexpected error: {error_to_str(e)}{bug_reports_message()}', is_error=False) unexpected_errors.append(e) continue assert isinstance(response, Response) return response raise NoSupportingHandlers(unsupported_errors, unexpected_errors) _REQUEST_HANDLERS = {} def register_rh(handler): """Register a RequestHandler class""" assert issubclass(handler, RequestHandler), f'{handler} must be a subclass of RequestHandler' assert handler.RH_KEY not in _REQUEST_HANDLERS, f'RequestHandler {handler.RH_KEY} already registered' _REQUEST_HANDLERS[handler.RH_KEY] = handler return handler class Features(enum.Enum): ALL_PROXY = enum.auto() NO_PROXY = enum.auto() class RequestHandler(abc.ABC): """Request Handler class Request handlers are class that, given a Request, process the request from start to finish and return a Response. Concrete subclasses need to redefine the _send(request) method, which handles the underlying request logic and returns a Response. RH_NAME class variable may contain a display name for the RequestHandler. By default, this is generated from the class name. The concrete request handler MUST have "RH" as the suffix in the class name. All exceptions raised by a RequestHandler should be an instance of RequestError. Any other exception raised will be treated as a handler issue. If a Request is not supported by the handler, an UnsupportedRequest should be raised with a reason. By default, some checks are done on the request in _validate() based on the following class variables: - `_SUPPORTED_URL_SCHEMES`: a tuple of supported url schemes. Any Request with an url scheme not in this list will raise an UnsupportedRequest. - `_SUPPORTED_PROXY_SCHEMES`: a tuple of support proxy url schemes. Any Request that contains a proxy url with an url scheme not in this list will raise an UnsupportedRequest. - `_SUPPORTED_FEATURES`: a tuple of supported features, as defined in Features enum. The above may be set to None to disable the checks. Parameters: @param logger: logger instance @param headers: HTTP Headers to include when sending requests. @param cookiejar: Cookiejar to use for requests. @param timeout: Socket timeout to use when sending requests. @param proxies: Proxies to use for sending requests. @param source_address: Client-side IP address to bind to for requests. @param verbose: Print debug request and traffic information to stdout. @param prefer_system_certs: Whether to prefer system certificates over other means (e.g. certifi). @param client_cert: SSL client certificate configuration. dict with {client_certificate, client_certificate_key, client_certificate_password} @param verify: Verify SSL certificates @param legacy_ssl_support: Enable legacy SSL options such as legacy server connect and older cipher support. Some configuration options may be available for individual Requests too. In this case, either the Request configuration option takes precedence or they are merged. Requests may have additional optional parameters defined as extensions. RequestHandler subclasses may choose to support custom extensions. If an extension is supported, subclasses should extend _check_extensions(extensions) to pop and validate the extension. - Extensions left in `extensions` are treated as unsupported and UnsupportedRequest will be raised. The following extensions are defined for RequestHandler: - `cookiejar`: Cookiejar to use for this request. - `timeout`: socket timeout to use for this request. - `legacy_ssl`: Enable legacy SSL options for this request. See legacy_ssl_support. - `keep_header_casing`: Keep the casing of headers when sending the request. To enable these, add extensions.pop('<extension>', None) to _check_extensions Apart from the url protocol, proxies dict may contain the following keys: - `all`: proxy to use for all protocols. Used as a fallback if no proxy is set for a specific protocol. - `no`: comma seperated list of hostnames (optionally with port) to not use a proxy for. Note: a RequestHandler may not support these, as defined in `_SUPPORTED_FEATURES`. """ _SUPPORTED_URL_SCHEMES = () _SUPPORTED_PROXY_SCHEMES = () _SUPPORTED_FEATURES = () def __init__( self, *, logger, # TODO(Grub4k): default logger headers: HTTPHeaderDict = None, cookiejar: YoutubeDLCookieJar = None, timeout: float | int | None = None, proxies: dict | None = None, source_address: str | None = None, verbose: bool = False, prefer_system_certs: bool = False, client_cert: dict[str, str | None] | None = None, verify: bool = True, legacy_ssl_support: bool = False, **_, ): self._logger = logger self.headers = headers or {} self.cookiejar = cookiejar if cookiejar is not None else YoutubeDLCookieJar() self.timeout = float(timeout or DEFAULT_TIMEOUT) self.proxies = proxies or {} self.source_address = source_address self.verbose = verbose self.prefer_system_certs = prefer_system_certs self._client_cert = client_cert or {} self.verify = verify self.legacy_ssl_support = legacy_ssl_support super().__init__() def _make_sslcontext(self, legacy_ssl_support=None): return make_ssl_context( verify=self.verify, legacy_support=legacy_ssl_support if legacy_ssl_support is not None else self.legacy_ssl_support, use_certifi=not self.prefer_system_certs, **self._client_cert, ) def _merge_headers(self, request_headers): return HTTPHeaderDict(self.headers, request_headers) def _prepare_headers(self, request: Request, headers: HTTPHeaderDict) -> None: # noqa: B027 """Additional operations to prepare headers before building. To be extended by subclasses. @param request: Request object @param headers: Merged headers to prepare """ def _get_headers(self, request: Request) -> dict[str, str]: """ Get headers for external use. Subclasses may define a _prepare_headers method to modify headers after merge but before building. """ headers = self._merge_headers(request.headers) self._prepare_headers(request, headers) if request.extensions.get('keep_header_casing'): return headers.sensitive() return dict(headers) def _calculate_timeout(self, request): return float(request.extensions.get('timeout') or self.timeout) def _get_cookiejar(self, request): cookiejar = request.extensions.get('cookiejar') return self.cookiejar if cookiejar is None else cookiejar def _get_proxies(self, request): return (request.proxies or self.proxies).copy() def _check_url_scheme(self, request: Request): scheme = urllib.parse.urlparse(request.url).scheme.lower() if self._SUPPORTED_URL_SCHEMES is not None and scheme not in self._SUPPORTED_URL_SCHEMES: raise UnsupportedRequest(f'Unsupported url scheme: "{scheme}"') return scheme # for further processing def _check_proxies(self, proxies): for proxy_key, proxy_url in proxies.items(): if proxy_url is None: continue if proxy_key == 'no': if self._SUPPORTED_FEATURES is not None and Features.NO_PROXY not in self._SUPPORTED_FEATURES: raise UnsupportedRequest('"no" proxy is not supported') continue if ( proxy_key == 'all' and self._SUPPORTED_FEATURES is not None and Features.ALL_PROXY not in self._SUPPORTED_FEATURES ): raise UnsupportedRequest('"all" proxy is not supported') # Unlikely this handler will use this proxy, so ignore. # This is to allow a case where a proxy may be set for a protocol # for one handler in which such protocol (and proxy) is not supported by another handler. if self._SUPPORTED_URL_SCHEMES is not None and proxy_key not in (*self._SUPPORTED_URL_SCHEMES, 'all'): continue if self._SUPPORTED_PROXY_SCHEMES is None: # Skip proxy scheme checks continue try: if urllib.request._parse_proxy(proxy_url)[0] is None: # Scheme-less proxies are not supported raise UnsupportedRequest(f'Proxy "{proxy_url}" missing scheme') except ValueError as e: # parse_proxy may raise on some invalid proxy urls such as "/a/b/c" raise UnsupportedRequest(f'Invalid proxy url "{proxy_url}": {e}') scheme = urllib.parse.urlparse(proxy_url).scheme.lower() if scheme not in self._SUPPORTED_PROXY_SCHEMES: raise UnsupportedRequest(f'Unsupported proxy type: "{scheme}"') def _check_extensions(self, extensions): """Check extensions for unsupported extensions. Subclasses should extend this.""" assert isinstance(extensions.get('cookiejar'), (YoutubeDLCookieJar, NoneType)) assert isinstance(extensions.get('timeout'), (float, int, NoneType)) assert isinstance(extensions.get('legacy_ssl'), (bool, NoneType)) assert isinstance(extensions.get('keep_header_casing'), (bool, NoneType)) def _validate(self, request): self._check_url_scheme(request) self._check_proxies(request.proxies or self.proxies) extensions = request.extensions.copy() self._check_extensions(extensions) if extensions: # TODO: add support for optional extensions raise UnsupportedRequest(f'Unsupported extensions: {", ".join(extensions.keys())}') @wrap_request_errors def validate(self, request: Request): if not isinstance(request, Request): raise TypeError('Expected an instance of Request') self._validate(request) @wrap_request_errors def send(self, request: Request) -> Response: if not isinstance(request, Request): raise TypeError('Expected an instance of Request') return self._send(request) @abc.abstractmethod def _send(self, request: Request): """Handle a request from start to finish. Redefine in subclasses.""" pass def close(self): # noqa: B027 pass @classproperty def RH_NAME(cls): return cls.__name__[:-2] @classproperty def RH_KEY(cls): assert cls.__name__.endswith('RH'), 'RequestHandler class names must end with "RH"' return cls.__name__[:-2] def __enter__(self): return self def __exit__(self, *args): self.close() class Request: """ Represents a request to be made. Partially backwards-compatible with urllib.request.Request. @param url: url to send. Will be sanitized. @param data: payload data to send. Must be bytes, iterable of bytes, a file-like object or None @param headers: headers to send. @param proxies: proxy dict mapping of proto:proxy to use for the request and any redirects. @param query: URL query parameters to update the url with. @param method: HTTP method to use. If no method specified, will use POST if payload data is present else GET @param extensions: Dictionary of Request extensions to add, as supported by handlers. """ def __init__( self, url: str, data: RequestData = None, headers: typing.Mapping | None = None, proxies: dict | None = None, query: dict | None = None, method: str | None = None, extensions: dict | None = None, ): self._headers = HTTPHeaderDict() self._data = None if query: url = update_url_query(url, query) self.url = url self.method = method if headers: self.headers = headers self.data = data # note: must be done after setting headers self.proxies = proxies or {} self.extensions = extensions or {} @property def url(self): return self._url @url.setter def url(self, url): if not isinstance(url, str): raise TypeError('url must be a string') elif url.startswith('//'): url = 'http:' + url self._url = normalize_url(url) @property def method(self): return self._method or ('POST' if self.data is not None else 'GET') @method.setter def method(self, method): if method is None: self._method = None elif isinstance(method, str): self._method = method.upper() else: raise TypeError('method must be a string') @property def data(self): return self._data @data.setter def data(self, data: RequestData): # Try catch some common mistakes if data is not None and ( not isinstance(data, (bytes, io.IOBase, Iterable)) or isinstance(data, (str, Mapping)) ): raise TypeError('data must be bytes, iterable of bytes, or a file-like object') if data == self._data and self._data is None: self.headers.pop('Content-Length', None) # https://docs.python.org/3/library/urllib.request.html#urllib.request.Request.data if data != self._data: if self._data is not None: self.headers.pop('Content-Length', None) self._data = data if self._data is None: self.headers.pop('Content-Type', None) if 'Content-Type' not in self.headers and self._data is not None: self.headers['Content-Type'] = 'application/x-www-form-urlencoded' @property def headers(self) -> HTTPHeaderDict: return self._headers @headers.setter def headers(self, new_headers: Mapping): """Replaces headers of the request. If not a HTTPHeaderDict, it will be converted to one.""" if isinstance(new_headers, HTTPHeaderDict): self._headers = new_headers elif isinstance(new_headers, Mapping): self._headers = HTTPHeaderDict(new_headers) else: raise TypeError('headers must be a mapping') def update(self, url=None, data=None, headers=None, query=None, extensions=None): self.data = data if data is not None else self.data self.headers.update(headers or {}) self.extensions.update(extensions or {}) self.url = update_url_query(url or self.url, query or {}) def copy(self): return self.__class__( url=self.url, headers=copy.deepcopy(self.headers), proxies=copy.deepcopy(self.proxies), data=self._data, extensions=copy.copy(self.extensions), method=self._method, ) HEADRequest = functools.partial(Request, method='HEAD') PATCHRequest = functools.partial(Request, method='PATCH') PUTRequest = functools.partial(Request, method='PUT') class Response(io.IOBase): """ Base class for HTTP response adapters. By default, it provides a basic wrapper for a file-like response object. Interface partially backwards-compatible with addinfourl and http.client.HTTPResponse. @param fp: Original, file-like, response. @param url: URL that this is a response of. @param headers: response headers. @param status: Response HTTP status code. Default is 200 OK. @param reason: HTTP status reason. Will use built-in reasons based on status code if not provided. @param extensions: Dictionary of handler-specific response extensions. """ def __init__( self, fp: io.IOBase, url: str, headers: Mapping[str, str], status: int = 200, reason: str | None = None, extensions: dict | None = None, ): self.fp = fp self.headers = Message() for name, value in headers.items(): self.headers.add_header(name, value) self.status = status self.url = url try: self.reason = reason or HTTPStatus(status).phrase except ValueError: self.reason = None self.extensions = extensions or {} def readable(self): return self.fp.readable() def read(self, amt: int | None = None) -> bytes: # Expected errors raised here should be of type RequestError or subclasses. # Subclasses should redefine this method with more precise error handling. try: res = self.fp.read(amt) if self.fp.closed: self.close() return res except Exception as e: raise TransportError(cause=e) from e def close(self): if not self.fp.closed: self.fp.close() return super().close() def get_header(self, name, default=None): """Get header for name. If there are multiple matching headers, return all seperated by comma.""" headers = self.headers.get_all(name) if not headers: return default if name.title() == 'Set-Cookie': # Special case, only get the first one # https://www.rfc-editor.org/rfc/rfc9110.html#section-5.3-4.1 return headers[0] return ', '.join(headers) # The following methods are for compatability reasons and are deprecated @property def code(self): deprecation_warning('Response.code is deprecated, use Response.status', stacklevel=2) return self.status def getcode(self): deprecation_warning('Response.getcode() is deprecated, use Response.status', stacklevel=2) return self.status def geturl(self): deprecation_warning('Response.geturl() is deprecated, use Response.url', stacklevel=2) return self.url def info(self): deprecation_warning('Response.info() is deprecated, use Response.headers', stacklevel=2) return self.headers def getheader(self, name, default=None): deprecation_warning('Response.getheader() is deprecated, use Response.get_header', stacklevel=2) return self.get_header(name, default) if typing.TYPE_CHECKING: RequestData = bytes | Iterable[bytes] | typing.IO | None Preference = typing.Callable[[RequestHandler, Request], int] _RH_PREFERENCES: set[Preference] = set()
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/networking/websocket.py
yt_dlp/networking/websocket.py
from __future__ import annotations import abc from .common import RequestHandler, Response class WebSocketResponse(Response): def send(self, message: bytes | str): """ Send a message to the server. @param message: The message to send. A string (str) is sent as a text frame, bytes is sent as a binary frame. """ raise NotImplementedError def recv(self): raise NotImplementedError class WebSocketRequestHandler(RequestHandler, abc.ABC): pass
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/networking/exceptions.py
yt_dlp/networking/exceptions.py
from __future__ import annotations import typing from ..utils import YoutubeDLError if typing.TYPE_CHECKING: from .common import RequestHandler, Response class RequestError(YoutubeDLError): def __init__( self, msg: str | None = None, cause: Exception | str | None = None, handler: RequestHandler = None, ): self.handler = handler self.cause = cause if not msg and cause: msg = str(cause) super().__init__(msg) class UnsupportedRequest(RequestError): """raised when a handler cannot handle a request""" pass class NoSupportingHandlers(RequestError): """raised when no handlers can support a request for various reasons""" def __init__(self, unsupported_errors: list[UnsupportedRequest], unexpected_errors: list[Exception]): self.unsupported_errors = unsupported_errors or [] self.unexpected_errors = unexpected_errors or [] # Print a quick summary of the errors err_handler_map = {} for err in unsupported_errors: err_handler_map.setdefault(err.msg, []).append(err.handler.RH_NAME) reason_str = ', '.join([f'{msg} ({", ".join(handlers)})' for msg, handlers in err_handler_map.items()]) if unexpected_errors: reason_str = ' + '.join(filter(None, [reason_str, f'{len(unexpected_errors)} unexpected error(s)'])) err_str = 'Unable to handle request' if reason_str: err_str += f': {reason_str}' super().__init__(msg=err_str) class TransportError(RequestError): """Network related errors""" class HTTPError(RequestError): def __init__(self, response: Response, redirect_loop=False): self.response = response self.status = response.status self.reason = response.reason self.redirect_loop = redirect_loop msg = f'HTTP Error {response.status}: {response.reason}' if redirect_loop: msg += ' (redirect loop detected)' super().__init__(msg=msg) def close(self): self.response.close() def __repr__(self): return f'<HTTPError {self.status}: {self.reason}>' class IncompleteRead(TransportError): def __init__(self, partial: int, expected: int | None = None, **kwargs): self.partial = partial self.expected = expected msg = f'{partial} bytes read' if expected is not None: msg += f', {expected} more expected' super().__init__(msg=msg, **kwargs) def __repr__(self): return f'<IncompleteRead: {self.msg}>' class SSLError(TransportError): pass class CertificateVerifyError(SSLError): """Raised when certificate validated has failed""" pass class ProxyError(TransportError): pass network_exceptions = (HTTPError, TransportError)
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/networking/_requests.py
yt_dlp/networking/_requests.py
from __future__ import annotations import functools import http.client import logging import re import warnings from ..dependencies import brotli, requests, urllib3 from ..utils import bug_reports_message, int_or_none, variadic from ..utils.networking import normalize_url, select_proxy if requests is None: raise ImportError('requests module is not installed') if urllib3 is None: raise ImportError('urllib3 module is not installed') urllib3_version = tuple(int_or_none(x, default=0) for x in urllib3.__version__.split('.')) if urllib3_version < (2, 0, 2): urllib3._yt_dlp__version = f'{urllib3.__version__} (unsupported)' raise ImportError('Only urllib3 >= 2.0.2 is supported') if requests.__build__ < 0x023202: requests._yt_dlp__version = f'{requests.__version__} (unsupported)' raise ImportError('Only requests >= 2.32.2 is supported') import requests.adapters import requests.utils import urllib3.connection import urllib3.exceptions import urllib3.util from ._helper import ( InstanceStoreMixin, add_accept_encoding_header, create_connection, create_socks_proxy_socket, get_redirect_method, make_socks_proxy_opts, ) from .common import ( Features, RequestHandler, Response, register_preference, register_rh, ) from .exceptions import ( CertificateVerifyError, HTTPError, IncompleteRead, ProxyError, RequestError, SSLError, TransportError, ) from ..socks import ProxyError as SocksProxyError SUPPORTED_ENCODINGS = [ 'gzip', 'deflate', ] if brotli is not None: SUPPORTED_ENCODINGS.append('br') ''' Override urllib3's behavior to not convert lower-case percent-encoded characters to upper-case during url normalization process. RFC3986 defines that the lower or upper case percent-encoded hexidecimal characters are equivalent and normalizers should convert them to uppercase for consistency [1]. However, some sites may have an incorrect implementation where they provide a percent-encoded url that is then compared case-sensitively.[2] While this is a very rare case, since urllib does not do this normalization step, it is best to avoid it in requests too for compatability reasons. 1: https://tools.ietf.org/html/rfc3986#section-2.1 2: https://github.com/streamlink/streamlink/pull/4003 ''' class Urllib3PercentREOverride: def __init__(self, r: re.Pattern): self.re = r # pass through all other attribute calls to the original re def __getattr__(self, item): return self.re.__getattribute__(item) def subn(self, repl, string, *args, **kwargs): return string, self.re.subn(repl, string, *args, **kwargs)[1] # urllib3 >= 1.25.8 uses subn: # https://github.com/urllib3/urllib3/commit/a2697e7c6b275f05879b60f593c5854a816489f0 import urllib3.util.url if hasattr(urllib3.util.url, '_PERCENT_RE'): # was 'PERCENT_RE' in urllib3 < 2.0.0 urllib3.util.url._PERCENT_RE = Urllib3PercentREOverride(urllib3.util.url._PERCENT_RE) else: warnings.warn('Failed to patch _PERCENT_RE in urllib3 (does the attribute exist?)' + bug_reports_message()) # Requests will not automatically handle no_proxy by default # due to buggy no_proxy handling with proxy dict [1]. # 1. https://github.com/psf/requests/issues/5000 requests.adapters.select_proxy = select_proxy class RequestsResponseAdapter(Response): def __init__(self, res: requests.models.Response): super().__init__( fp=res.raw, headers=res.headers, url=res.url, status=res.status_code, reason=res.reason) self._requests_response = res def _real_read(self, amt: int | None = None) -> bytes: # Work around issue with `.read(amt)` then `.read()` # See: https://github.com/urllib3/urllib3/issues/3636 if amt is None: # compat: py3.9: Python 3.9 preallocates the whole read buffer, read in chunks read_chunk = functools.partial(self.fp.read, 1 << 20, decode_content=True) return b''.join(iter(read_chunk, b'')) # Interact with urllib3 response directly. return self.fp.read(amt, decode_content=True) def read(self, amt: int | None = None): try: data = self._real_read(amt) if self.fp.closed: self.close() return data # See urllib3.response.HTTPResponse.read() for exceptions raised on read except urllib3.exceptions.SSLError as e: raise SSLError(cause=e) from e except urllib3.exceptions.ProtocolError as e: # IncompleteRead is always contained within ProtocolError # See urllib3.response.HTTPResponse._error_catcher() ir_err = next( (err for err in (e.__context__, e.__cause__, *variadic(e.args)) if isinstance(err, http.client.IncompleteRead)), None) if ir_err is not None: # `urllib3.exceptions.IncompleteRead` is subclass of `http.client.IncompleteRead` # but uses an `int` for its `partial` property. partial = ir_err.partial if isinstance(ir_err.partial, int) else len(ir_err.partial) raise IncompleteRead(partial=partial, expected=ir_err.expected) from e raise TransportError(cause=e) from e except urllib3.exceptions.HTTPError as e: # catch-all for any other urllib3 response exceptions raise TransportError(cause=e) from e class RequestsHTTPAdapter(requests.adapters.HTTPAdapter): def __init__(self, ssl_context=None, proxy_ssl_context=None, source_address=None, **kwargs): self._pm_args = {} if ssl_context: self._pm_args['ssl_context'] = ssl_context if source_address: self._pm_args['source_address'] = (source_address, 0) self._proxy_ssl_context = proxy_ssl_context or ssl_context super().__init__(**kwargs) def init_poolmanager(self, *args, **kwargs): return super().init_poolmanager(*args, **kwargs, **self._pm_args) def proxy_manager_for(self, proxy, **proxy_kwargs): extra_kwargs = {} if not proxy.lower().startswith('socks') and self._proxy_ssl_context: extra_kwargs['proxy_ssl_context'] = self._proxy_ssl_context return super().proxy_manager_for(proxy, **proxy_kwargs, **self._pm_args, **extra_kwargs) # Skip `requests` internal verification; we use our own SSLContext def cert_verify(*args, **kwargs): pass # requests 2.32.2+: Reimplementation without `_urllib3_request_context` def get_connection_with_tls_context(self, request, verify, proxies=None, cert=None): url = urllib3.util.parse_url(request.url).url manager = self.poolmanager if proxy := select_proxy(url, proxies): manager = self.proxy_manager_for(proxy) return manager.connection_from_url(url) class RequestsSession(requests.sessions.Session): """ Ensure unified redirect method handling with our urllib redirect handler. """ def rebuild_method(self, prepared_request, response): new_method = get_redirect_method(prepared_request.method, response.status_code) # HACK: requests removes headers/body on redirect unless code was a 307/308. if new_method == prepared_request.method: response._real_status_code = response.status_code response.status_code = 308 prepared_request.method = new_method # Requests fails to resolve dot segments on absolute redirect locations # See: https://github.com/yt-dlp/yt-dlp/issues/9020 prepared_request.url = normalize_url(prepared_request.url) def rebuild_auth(self, prepared_request, response): # HACK: undo status code change from rebuild_method, if applicable. # rebuild_auth runs after requests would remove headers/body based on status code if hasattr(response, '_real_status_code'): response.status_code = response._real_status_code del response._real_status_code return super().rebuild_auth(prepared_request, response) class Urllib3LoggingFilter(logging.Filter): def filter(self, record): # Ignore HTTP request messages since HTTPConnection prints those return record.msg != '%s://%s:%s "%s %s %s" %s %s' class Urllib3LoggingHandler(logging.Handler): """Redirect urllib3 logs to our logger""" def __init__(self, logger, *args, **kwargs): super().__init__(*args, **kwargs) self._logger = logger def emit(self, record): try: msg = self.format(record) if record.levelno >= logging.ERROR: self._logger.error(msg) else: self._logger.stdout(msg) except Exception: self.handleError(record) @register_rh class RequestsRH(RequestHandler, InstanceStoreMixin): """Requests RequestHandler https://github.com/psf/requests """ _SUPPORTED_URL_SCHEMES = ('http', 'https') _SUPPORTED_ENCODINGS = tuple(SUPPORTED_ENCODINGS) _SUPPORTED_PROXY_SCHEMES = ('http', 'https', 'socks4', 'socks4a', 'socks5', 'socks5h') _SUPPORTED_FEATURES = (Features.NO_PROXY, Features.ALL_PROXY) RH_NAME = 'requests' def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) # Forward urllib3 debug messages to our logger logger = logging.getLogger('urllib3') self.__logging_handler = Urllib3LoggingHandler(logger=self._logger) self.__logging_handler.setFormatter(logging.Formatter('requests: %(message)s')) self.__logging_handler.addFilter(Urllib3LoggingFilter()) logger.addHandler(self.__logging_handler) # TODO: Use a logger filter to suppress pool reuse warning instead logger.setLevel(logging.ERROR) if self.verbose: # Setting this globally is not ideal, but is easier than hacking with urllib3. # It could technically be problematic for scripts embedding yt-dlp. # However, it is unlikely debug traffic is used in that context in a way this will cause problems. urllib3.connection.HTTPConnection.debuglevel = 1 logger.setLevel(logging.DEBUG) # this is expected if we are using --no-check-certificate urllib3.disable_warnings(urllib3.exceptions.InsecureRequestWarning) def close(self): self._clear_instances() # Remove the logging handler that contains a reference to our logger # See: https://github.com/yt-dlp/yt-dlp/issues/8922 logging.getLogger('urllib3').removeHandler(self.__logging_handler) def _check_extensions(self, extensions): super()._check_extensions(extensions) extensions.pop('cookiejar', None) extensions.pop('timeout', None) extensions.pop('legacy_ssl', None) extensions.pop('keep_header_casing', None) def _create_instance(self, cookiejar, legacy_ssl_support=None): session = RequestsSession() http_adapter = RequestsHTTPAdapter( ssl_context=self._make_sslcontext(legacy_ssl_support=legacy_ssl_support), source_address=self.source_address, max_retries=urllib3.util.retry.Retry(False), ) session.adapters.clear() session.headers = requests.models.CaseInsensitiveDict() session.mount('https://', http_adapter) session.mount('http://', http_adapter) session.cookies = cookiejar session.trust_env = False # no need, we already load proxies from env return session def _prepare_headers(self, _, headers): add_accept_encoding_header(headers, SUPPORTED_ENCODINGS) headers.setdefault('Connection', 'keep-alive') def _send(self, request): headers = self._get_headers(request) max_redirects_exceeded = False session = self._get_instance( cookiejar=self._get_cookiejar(request), legacy_ssl_support=request.extensions.get('legacy_ssl'), ) try: requests_res = session.request( method=request.method, url=request.url, data=request.data, headers=headers, timeout=self._calculate_timeout(request), proxies=self._get_proxies(request), allow_redirects=True, stream=True, ) except requests.exceptions.TooManyRedirects as e: max_redirects_exceeded = True requests_res = e.response except requests.exceptions.SSLError as e: if 'CERTIFICATE_VERIFY_FAILED' in str(e): raise CertificateVerifyError(cause=e) from e raise SSLError(cause=e) from e except requests.exceptions.ProxyError as e: raise ProxyError(cause=e) from e except (requests.exceptions.ConnectionError, requests.exceptions.Timeout) as e: raise TransportError(cause=e) from e except urllib3.exceptions.HTTPError as e: # Catch any urllib3 exceptions that may leak through raise TransportError(cause=e) from e except requests.exceptions.RequestException as e: # Miscellaneous Requests exceptions. May not necessary be network related e.g. InvalidURL raise RequestError(cause=e) from e res = RequestsResponseAdapter(requests_res) if not 200 <= res.status < 300: raise HTTPError(res, redirect_loop=max_redirects_exceeded) return res @register_preference(RequestsRH) def requests_preference(rh, request): return 100 # Use our socks proxy implementation with requests to avoid an extra dependency. class SocksHTTPConnection(urllib3.connection.HTTPConnection): def __init__(self, _socks_options, *args, **kwargs): # must use _socks_options to pass PoolKey checks self._proxy_args = _socks_options super().__init__(*args, **kwargs) def _new_conn(self): try: return create_connection( address=(self._proxy_args['addr'], self._proxy_args['port']), timeout=self.timeout, source_address=self.source_address, _create_socket_func=functools.partial( create_socks_proxy_socket, (self.host, self.port), self._proxy_args)) except TimeoutError as e: raise urllib3.exceptions.ConnectTimeoutError( self, f'Connection to {self.host} timed out. (connect timeout={self.timeout})') from e except SocksProxyError as e: raise urllib3.exceptions.ProxyError(str(e), e) from e except OSError as e: raise urllib3.exceptions.NewConnectionError( self, f'Failed to establish a new connection: {e}') from e class SocksHTTPSConnection(SocksHTTPConnection, urllib3.connection.HTTPSConnection): pass class SocksHTTPConnectionPool(urllib3.HTTPConnectionPool): ConnectionCls = SocksHTTPConnection class SocksHTTPSConnectionPool(urllib3.HTTPSConnectionPool): ConnectionCls = SocksHTTPSConnection class SocksProxyManager(urllib3.PoolManager): def __init__(self, socks_proxy, username=None, password=None, num_pools=10, headers=None, **connection_pool_kw): connection_pool_kw['_socks_options'] = make_socks_proxy_opts(socks_proxy) super().__init__(num_pools, headers, **connection_pool_kw) self.pool_classes_by_scheme = { 'http': SocksHTTPConnectionPool, 'https': SocksHTTPSConnectionPool, } requests.adapters.SOCKSProxyManager = SocksProxyManager
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/networking/_curlcffi.py
yt_dlp/networking/_curlcffi.py
from __future__ import annotations import io import itertools import math import re import urllib.parse from ._helper import InstanceStoreMixin from ..utils.networking import select_proxy from .common import ( Features, Request, Response, register_preference, register_rh, ) from .exceptions import ( CertificateVerifyError, HTTPError, IncompleteRead, ProxyError, SSLError, TransportError, ) from .impersonate import ImpersonateRequestHandler, ImpersonateTarget from ..dependencies import curl_cffi, certifi from ..utils import int_or_none if curl_cffi is None: raise ImportError('curl_cffi is not installed') curl_cffi_version = tuple(map(int, re.split(r'[^\d]+', curl_cffi.__version__)[:3])) if curl_cffi_version != (0, 5, 10) and not (0, 10) <= curl_cffi_version < (0, 14): curl_cffi._yt_dlp__version = f'{curl_cffi.__version__} (unsupported)' raise ImportError('Only curl_cffi versions 0.5.10, 0.10.x, 0.11.x, 0.12.x, 0.13.x are supported') import curl_cffi.requests from curl_cffi.const import CurlECode, CurlOpt class CurlCFFIResponseReader(io.IOBase): def __init__(self, response: curl_cffi.requests.Response): self._response = response self._iterator = response.iter_content() self._buffer = b'' self.bytes_read = 0 def readable(self): return True def read(self, size=None): exception_raised = True try: while self._iterator and (size is None or len(self._buffer) < size): chunk = next(self._iterator, None) if chunk is None: self._iterator = None break self._buffer += chunk self.bytes_read += len(chunk) if size is None: size = len(self._buffer) data = self._buffer[:size] self._buffer = self._buffer[size:] # "free" the curl instance if the response is fully read. # curl_cffi doesn't do this automatically and only allows one open response per thread if not self._iterator and not self._buffer: self.close() exception_raised = False return data finally: if exception_raised: self.close() def close(self): if not self.closed: self._response.close() self._buffer = b'' super().close() class CurlCFFIResponseAdapter(Response): fp: CurlCFFIResponseReader def __init__(self, response: curl_cffi.requests.Response): super().__init__( fp=CurlCFFIResponseReader(response), headers=response.headers, url=response.url, status=response.status_code) def read(self, amt=None): try: res = self.fp.read(amt) if self.fp.closed: self.close() return res except curl_cffi.requests.errors.RequestsError as e: if e.code == CurlECode.PARTIAL_FILE: content_length = e.response and int_or_none(e.response.headers.get('Content-Length')) raise IncompleteRead( partial=self.fp.bytes_read, expected=content_length - self.fp.bytes_read if content_length is not None else None, cause=e) from e raise TransportError(cause=e) from e # See: https://github.com/lexiforest/curl_cffi?tab=readme-ov-file#supported-impersonate-browsers # https://github.com/lexiforest/curl-impersonate?tab=readme-ov-file#supported-browsers BROWSER_TARGETS: dict[tuple[int, ...], dict[str, ImpersonateTarget]] = { (0, 5): { 'chrome99': ImpersonateTarget('chrome', '99', 'windows', '10'), 'chrome99_android': ImpersonateTarget('chrome', '99', 'android', '12'), 'chrome100': ImpersonateTarget('chrome', '100', 'windows', '10'), 'chrome101': ImpersonateTarget('chrome', '101', 'windows', '10'), 'chrome104': ImpersonateTarget('chrome', '104', 'windows', '10'), 'chrome107': ImpersonateTarget('chrome', '107', 'windows', '10'), 'chrome110': ImpersonateTarget('chrome', '110', 'windows', '10'), 'edge99': ImpersonateTarget('edge', '99', 'windows', '10'), 'edge101': ImpersonateTarget('edge', '101', 'windows', '10'), 'safari153': ImpersonateTarget('safari', '15.3', 'macos', '11'), 'safari155': ImpersonateTarget('safari', '15.5', 'macos', '12'), }, (0, 7): { 'chrome116': ImpersonateTarget('chrome', '116', 'windows', '10'), 'chrome119': ImpersonateTarget('chrome', '119', 'macos', '14'), 'chrome120': ImpersonateTarget('chrome', '120', 'macos', '14'), 'chrome123': ImpersonateTarget('chrome', '123', 'macos', '14'), 'chrome124': ImpersonateTarget('chrome', '124', 'macos', '14'), 'safari170': ImpersonateTarget('safari', '17.0', 'macos', '14'), 'safari172_ios': ImpersonateTarget('safari', '17.2', 'ios', '17.2'), }, (0, 9): { 'safari153': ImpersonateTarget('safari', '15.3', 'macos', '14'), 'safari155': ImpersonateTarget('safari', '15.5', 'macos', '14'), 'chrome119': ImpersonateTarget('chrome', '119', 'macos', '14'), 'chrome120': ImpersonateTarget('chrome', '120', 'macos', '14'), 'chrome123': ImpersonateTarget('chrome', '123', 'macos', '14'), 'chrome124': ImpersonateTarget('chrome', '124', 'macos', '14'), 'chrome131': ImpersonateTarget('chrome', '131', 'macos', '14'), 'chrome131_android': ImpersonateTarget('chrome', '131', 'android', '14'), 'chrome133a': ImpersonateTarget('chrome', '133', 'macos', '15'), 'firefox133': ImpersonateTarget('firefox', '133', 'macos', '14'), 'safari180': ImpersonateTarget('safari', '18.0', 'macos', '15'), 'safari180_ios': ImpersonateTarget('safari', '18.0', 'ios', '18.0'), }, (0, 10): { 'firefox135': ImpersonateTarget('firefox', '135', 'macos', '14'), }, (0, 11): { 'tor145': ImpersonateTarget('tor', '14.5', 'macos', '14'), 'safari184': ImpersonateTarget('safari', '18.4', 'macos', '15'), 'safari184_ios': ImpersonateTarget('safari', '18.4', 'ios', '18.4'), 'chrome136': ImpersonateTarget('chrome', '136', 'macos', '15'), }, (0, 12): { 'safari260': ImpersonateTarget('safari', '26.0', 'macos', '26'), 'safari260_ios': ImpersonateTarget('safari', '26.0', 'ios', '26.0'), }, } # Needed for curl_cffi < 0.11 # See: https://github.com/lexiforest/curl_cffi/commit/d2f15c7a31506a08d217fcc04ae7570c39f5f5bb _TARGETS_COMPAT_LOOKUP = { 'safari153': 'safari15_3', 'safari155': 'safari15_5', 'safari170': 'safari17_0', 'safari172_ios': 'safari17_2_ios', 'safari180': 'safari18_0', 'safari180_ios': 'safari18_0_ios', } @register_rh class CurlCFFIRH(ImpersonateRequestHandler, InstanceStoreMixin): RH_NAME = 'curl_cffi' _SUPPORTED_URL_SCHEMES = ('http', 'https') _SUPPORTED_FEATURES = (Features.NO_PROXY, Features.ALL_PROXY) _SUPPORTED_PROXY_SCHEMES = ('http', 'https', 'socks4', 'socks4a', 'socks5', 'socks5h') _SUPPORTED_IMPERSONATE_TARGET_MAP = { target: ( name if curl_cffi_version >= (0, 11) else _TARGETS_COMPAT_LOOKUP.get(name, name) if curl_cffi_version >= (0, 9) else curl_cffi.requests.BrowserType[_TARGETS_COMPAT_LOOKUP.get(name, name)] ) for name, target in dict(sorted(itertools.chain.from_iterable( targets.items() for version, targets in BROWSER_TARGETS.items() if curl_cffi_version >= version ), key=lambda x: ( # deprioritize mobile targets since they give very different behavior x[1].os not in ('ios', 'android'), # prioritize tor < edge < firefox < safari < chrome ('tor', 'edge', 'firefox', 'safari', 'chrome').index(x[1].client), # prioritize newest version float(x[1].version) if x[1].version else 0, # group by os name x[1].os, ), reverse=True)).items() } def _create_instance(self, cookiejar=None): return curl_cffi.requests.Session(cookies=cookiejar) def _check_extensions(self, extensions): super()._check_extensions(extensions) extensions.pop('impersonate', None) extensions.pop('cookiejar', None) extensions.pop('timeout', None) # CurlCFFIRH ignores legacy ssl options currently. # Impersonation generally uses a looser SSL configuration than urllib/requests. extensions.pop('legacy_ssl', None) def send(self, request: Request) -> Response: target = self._get_request_target(request) try: response = super().send(request) except HTTPError as e: e.response.extensions['impersonate'] = target raise response.extensions['impersonate'] = target return response def _send(self, request: Request): max_redirects_exceeded = False session: curl_cffi.requests.Session = self._get_instance( cookiejar=self._get_cookiejar(request) if 'cookie' not in request.headers else None) if self.verbose: session.curl.setopt(CurlOpt.VERBOSE, 1) proxies = self._get_proxies(request) if 'no' in proxies: session.curl.setopt(CurlOpt.NOPROXY, proxies['no']) proxies.pop('no', None) # curl doesn't support per protocol proxies, so we select the one that matches the request protocol proxy = select_proxy(request.url, proxies=proxies) if proxy: session.curl.setopt(CurlOpt.PROXY, proxy) scheme = urllib.parse.urlparse(request.url).scheme.lower() if scheme != 'http': # Enable HTTP CONNECT for HTTPS urls. # Don't use CONNECT for http for compatibility with urllib behaviour. # See: https://curl.se/libcurl/c/CURLOPT_HTTPPROXYTUNNEL.html session.curl.setopt(CurlOpt.HTTPPROXYTUNNEL, 1) # curl_cffi does not currently set these for proxies session.curl.setopt(CurlOpt.PROXY_CAINFO, certifi.where()) if not self.verify: session.curl.setopt(CurlOpt.PROXY_SSL_VERIFYPEER, 0) session.curl.setopt(CurlOpt.PROXY_SSL_VERIFYHOST, 0) headers = self._get_impersonate_headers(request) if self._client_cert: session.curl.setopt(CurlOpt.SSLCERT, self._client_cert['client_certificate']) client_certificate_key = self._client_cert.get('client_certificate_key') client_certificate_password = self._client_cert.get('client_certificate_password') if client_certificate_key: session.curl.setopt(CurlOpt.SSLKEY, client_certificate_key) if client_certificate_password: session.curl.setopt(CurlOpt.KEYPASSWD, client_certificate_password) timeout = self._calculate_timeout(request) # set CURLOPT_LOW_SPEED_LIMIT and CURLOPT_LOW_SPEED_TIME to act as a read timeout. [1] # This is required only for 0.5.10 [2] # Note: CURLOPT_LOW_SPEED_TIME is in seconds, so we need to round up to the nearest second. [3] # [1] https://unix.stackexchange.com/a/305311 # [2] https://github.com/yifeikong/curl_cffi/issues/156 # [3] https://curl.se/libcurl/c/CURLOPT_LOW_SPEED_TIME.html session.curl.setopt(CurlOpt.LOW_SPEED_LIMIT, 1) # 1 byte per second session.curl.setopt(CurlOpt.LOW_SPEED_TIME, math.ceil(timeout)) try: curl_response = session.request( method=request.method, url=request.url, headers=headers, data=request.data, verify=self.verify, max_redirects=5, timeout=(timeout, timeout), impersonate=self._SUPPORTED_IMPERSONATE_TARGET_MAP.get( self._get_request_target(request)), interface=self.source_address, stream=True, ) except curl_cffi.requests.errors.RequestsError as e: if e.code == CurlECode.PEER_FAILED_VERIFICATION: raise CertificateVerifyError(cause=e) from e elif e.code == CurlECode.SSL_CONNECT_ERROR: raise SSLError(cause=e) from e elif e.code == CurlECode.TOO_MANY_REDIRECTS: max_redirects_exceeded = True curl_response = e.response elif ( e.code == CurlECode.PROXY or (e.code == CurlECode.RECV_ERROR and 'CONNECT' in str(e)) ): raise ProxyError(cause=e) from e else: raise TransportError(cause=e) from e response = CurlCFFIResponseAdapter(curl_response) if not 200 <= response.status < 300: raise HTTPError(response, redirect_loop=max_redirects_exceeded) return response @register_preference(CurlCFFIRH) def curl_cffi_preference(rh, request): return -100
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/networking/_helper.py
yt_dlp/networking/_helper.py
from __future__ import annotations import contextlib import functools import os import socket import ssl import sys import typing import urllib.parse import urllib.request from .exceptions import RequestError from ..dependencies import certifi from ..socks import ProxyType, sockssocket if typing.TYPE_CHECKING: from collections.abc import Iterable from ..utils.networking import HTTPHeaderDict def ssl_load_certs(context: ssl.SSLContext, use_certifi=True): if certifi and use_certifi: context.load_verify_locations(cafile=certifi.where()) else: try: context.load_default_certs() # Work around the issue in load_default_certs when there are bad certificates. See: # https://github.com/yt-dlp/yt-dlp/issues/1060, # https://bugs.python.org/issue35665, https://bugs.python.org/issue45312 except ssl.SSLError: # enum_certificates is not present in mingw python. See https://github.com/yt-dlp/yt-dlp/issues/1151 if sys.platform == 'win32' and hasattr(ssl, 'enum_certificates'): for storename in ('CA', 'ROOT'): ssl_load_windows_store_certs(context, storename) context.set_default_verify_paths() def ssl_load_windows_store_certs(ssl_context, storename): # Code adapted from _load_windows_store_certs in https://github.com/python/cpython/blob/main/Lib/ssl.py try: certs = [cert for cert, encoding, trust in ssl.enum_certificates(storename) if encoding == 'x509_asn' and ( trust is True or ssl.Purpose.SERVER_AUTH.oid in trust)] except PermissionError: return for cert in certs: with contextlib.suppress(ssl.SSLError): ssl_context.load_verify_locations(cadata=cert) def make_socks_proxy_opts(socks_proxy): url_components = urllib.parse.urlparse(socks_proxy) if url_components.scheme.lower() == 'socks5': socks_type = ProxyType.SOCKS5 rdns = False elif url_components.scheme.lower() == 'socks5h': socks_type = ProxyType.SOCKS5 rdns = True elif url_components.scheme.lower() == 'socks4': socks_type = ProxyType.SOCKS4 rdns = False elif url_components.scheme.lower() == 'socks4a': socks_type = ProxyType.SOCKS4A rdns = True else: raise ValueError(f'Unknown SOCKS proxy version: {url_components.scheme.lower()}') def unquote_if_non_empty(s): if not s: return s return urllib.parse.unquote_plus(s) return { 'proxytype': socks_type, 'addr': url_components.hostname, 'port': url_components.port or 1080, 'rdns': rdns, 'username': unquote_if_non_empty(url_components.username), 'password': unquote_if_non_empty(url_components.password), } def get_redirect_method(method, status): """Unified redirect method handling""" # A 303 must either use GET or HEAD for subsequent request # https://datatracker.ietf.org/doc/html/rfc7231#section-6.4.4 if status == 303 and method != 'HEAD': method = 'GET' # 301 and 302 redirects are commonly turned into a GET from a POST # for subsequent requests by browsers, so we'll do the same. # https://datatracker.ietf.org/doc/html/rfc7231#section-6.4.2 # https://datatracker.ietf.org/doc/html/rfc7231#section-6.4.3 if status in (301, 302) and method == 'POST': method = 'GET' return method def make_ssl_context( verify=True, client_certificate=None, client_certificate_key=None, client_certificate_password=None, legacy_support=False, use_certifi=True, ): context = ssl.SSLContext(ssl.PROTOCOL_TLS_CLIENT) context.check_hostname = verify context.verify_mode = ssl.CERT_REQUIRED if verify else ssl.CERT_NONE # OpenSSL 1.1.1+ Python 3.8+ keylog file if hasattr(context, 'keylog_filename'): context.keylog_filename = os.environ.get('SSLKEYLOGFILE') or None # Some servers may reject requests if ALPN extension is not sent. See: # https://github.com/python/cpython/issues/85140 # https://github.com/yt-dlp/yt-dlp/issues/3878 with contextlib.suppress(NotImplementedError): context.set_alpn_protocols(['http/1.1']) if verify: ssl_load_certs(context, use_certifi) if legacy_support: context.options |= 4 # SSL_OP_LEGACY_SERVER_CONNECT context.set_ciphers('DEFAULT') # compat elif ssl.OPENSSL_VERSION_INFO >= (1, 1, 1) and not ssl.OPENSSL_VERSION.startswith('LibreSSL'): # Use the default SSL ciphers and minimum TLS version settings from Python 3.10 [1]. # This is to ensure consistent behavior across Python versions and libraries, and help avoid fingerprinting # in some situations [2][3]. # Python 3.10 only supports OpenSSL 1.1.1+ [4]. Because this change is likely # untested on older versions, we only apply this to OpenSSL 1.1.1+ to be safe. # LibreSSL is excluded until further investigation due to cipher support issues [5][6]. # 1. https://github.com/python/cpython/commit/e983252b516edb15d4338b0a47631b59ef1e2536 # 2. https://github.com/yt-dlp/yt-dlp/issues/4627 # 3. https://github.com/yt-dlp/yt-dlp/pull/5294 # 4. https://peps.python.org/pep-0644/ # 5. https://peps.python.org/pep-0644/#libressl-support # 6. https://github.com/yt-dlp/yt-dlp/commit/5b9f253fa0aee996cf1ed30185d4b502e00609c4#commitcomment-89054368 context.set_ciphers( '@SECLEVEL=2:ECDH+AESGCM:ECDH+CHACHA20:ECDH+AES:DHE+AES:!aNULL:!eNULL:!aDSS:!SHA1:!AESCCM') context.minimum_version = ssl.TLSVersion.TLSv1_2 if client_certificate: try: context.load_cert_chain( client_certificate, keyfile=client_certificate_key, password=client_certificate_password) except ssl.SSLError: raise RequestError('Unable to load client certificate') if getattr(context, 'post_handshake_auth', None) is not None: context.post_handshake_auth = True return context class InstanceStoreMixin: def __init__(self, **kwargs): self.__instances = [] super().__init__(**kwargs) # So that both MRO works @staticmethod def _create_instance(**kwargs): raise NotImplementedError def _get_instance(self, **kwargs): for key, instance in self.__instances: if key == kwargs: return instance instance = self._create_instance(**kwargs) self.__instances.append((kwargs, instance)) return instance def _close_instance(self, instance): if callable(getattr(instance, 'close', None)): instance.close() def _clear_instances(self): for _, instance in self.__instances: self._close_instance(instance) self.__instances.clear() def add_accept_encoding_header(headers: HTTPHeaderDict, supported_encodings: Iterable[str]): if 'Accept-Encoding' not in headers: headers['Accept-Encoding'] = ', '.join(supported_encodings) or 'identity' def wrap_request_errors(func): @functools.wraps(func) def wrapper(self, *args, **kwargs): try: return func(self, *args, **kwargs) except RequestError as e: if e.handler is None: e.handler = self raise return wrapper def _socket_connect(ip_addr, timeout, source_address): af, socktype, proto, _canonname, sa = ip_addr sock = socket.socket(af, socktype, proto) try: if timeout is not socket._GLOBAL_DEFAULT_TIMEOUT: sock.settimeout(timeout) if source_address: sock.bind(source_address) sock.connect(sa) return sock except OSError: sock.close() raise def create_socks_proxy_socket(dest_addr, proxy_args, proxy_ip_addr, timeout, source_address): af, socktype, proto, _canonname, sa = proxy_ip_addr sock = sockssocket(af, socktype, proto) try: connect_proxy_args = proxy_args.copy() connect_proxy_args.update({'addr': sa[0], 'port': sa[1]}) sock.setproxy(**connect_proxy_args) if timeout is not socket._GLOBAL_DEFAULT_TIMEOUT: sock.settimeout(timeout) if source_address: sock.bind(source_address) sock.connect(dest_addr) return sock except OSError: sock.close() raise def create_connection( address, timeout=socket._GLOBAL_DEFAULT_TIMEOUT, source_address=None, *, _create_socket_func=_socket_connect, ): # Work around socket.create_connection() which tries all addresses from getaddrinfo() including IPv6. # This filters the addresses based on the given source_address. # Based on: https://github.com/python/cpython/blob/main/Lib/socket.py#L810 host, port = address ip_addrs = socket.getaddrinfo(host, port, 0, socket.SOCK_STREAM) if not ip_addrs: raise OSError('getaddrinfo returns an empty list') if source_address is not None: af = socket.AF_INET if ':' not in source_address[0] else socket.AF_INET6 ip_addrs = [addr for addr in ip_addrs if addr[0] == af] if not ip_addrs: raise OSError( f'No remote IPv{4 if af == socket.AF_INET else 6} addresses available for connect. ' f'Can\'t use "{source_address[0]}" as source address') err = None for ip_addr in ip_addrs: try: sock = _create_socket_func(ip_addr, timeout, source_address) # Explicitly break __traceback__ reference cycle # https://bugs.python.org/issue36820 err = None return sock except OSError as e: err = e try: raise err finally: # Explicitly break __traceback__ reference cycle # https://bugs.python.org/issue36820 err = None
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/networking/_websockets.py
yt_dlp/networking/_websockets.py
from __future__ import annotations import contextlib import functools import io import logging import ssl import sys from ._helper import ( create_connection, create_socks_proxy_socket, make_socks_proxy_opts, ) from ..utils.networking import select_proxy from .common import Features, Response, register_rh from .exceptions import ( CertificateVerifyError, HTTPError, ProxyError, RequestError, SSLError, TransportError, ) from .websocket import WebSocketRequestHandler, WebSocketResponse from ..dependencies import websockets from ..socks import ProxyError as SocksProxyError from ..utils import int_or_none if not websockets: raise ImportError('websockets is not installed') import websockets.version websockets_version = tuple(map(int_or_none, websockets.version.version.split('.'))) if websockets_version < (13, 0): websockets._yt_dlp__version = f'{websockets.version.version} (unsupported)' raise ImportError('Only websockets>=13.0 is supported') import websockets.sync.client from websockets.uri import parse_uri # In websockets Connection, recv_exc and recv_events_exc are defined # after the recv events handler thread is started [1]. # On our CI using PyPy, in some cases a race condition may occur # where the recv events handler thread tries to use these attributes before they are defined [2]. # 1: https://github.com/python-websockets/websockets/blame/de768cf65e7e2b1a3b67854fb9e08816a5ff7050/src/websockets/sync/connection.py#L93 # 2: "AttributeError: 'ClientConnection' object has no attribute 'recv_events_exc'. Did you mean: 'recv_events'?" import websockets.sync.connection # isort: split with contextlib.suppress(Exception): websockets.sync.connection.Connection.recv_exc = None class WebsocketsResponseAdapter(WebSocketResponse): def __init__(self, ws: websockets.sync.client.ClientConnection, url): super().__init__( fp=io.BytesIO(ws.response.body or b''), url=url, headers=ws.response.headers, status=ws.response.status_code, reason=ws.response.reason_phrase, ) self._ws = ws def close(self): self._ws.close() super().close() def send(self, message): # https://websockets.readthedocs.io/en/stable/reference/sync/client.html#websockets.sync.client.ClientConnection.send try: return self._ws.send(message) except (websockets.exceptions.WebSocketException, RuntimeError, TimeoutError) as e: raise TransportError(cause=e) from e except SocksProxyError as e: raise ProxyError(cause=e) from e except TypeError as e: raise RequestError(cause=e) from e def recv(self): # https://websockets.readthedocs.io/en/stable/reference/sync/client.html#websockets.sync.client.ClientConnection.recv try: return self._ws.recv() except SocksProxyError as e: raise ProxyError(cause=e) from e except (websockets.exceptions.WebSocketException, RuntimeError, TimeoutError) as e: raise TransportError(cause=e) from e @register_rh class WebsocketsRH(WebSocketRequestHandler): """ Websockets request handler https://websockets.readthedocs.io https://github.com/python-websockets/websockets """ _SUPPORTED_URL_SCHEMES = ('wss', 'ws') _SUPPORTED_PROXY_SCHEMES = ('socks4', 'socks4a', 'socks5', 'socks5h') _SUPPORTED_FEATURES = (Features.ALL_PROXY, Features.NO_PROXY) RH_NAME = 'websockets' def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) self.__logging_handlers = {} for name in ('websockets.client', 'websockets.server'): logger = logging.getLogger(name) handler = logging.StreamHandler(stream=sys.stdout) handler.setFormatter(logging.Formatter(f'{self.RH_NAME}: %(message)s')) self.__logging_handlers[name] = handler logger.addHandler(handler) if self.verbose: logger.setLevel(logging.DEBUG) def _check_extensions(self, extensions): super()._check_extensions(extensions) extensions.pop('timeout', None) extensions.pop('cookiejar', None) extensions.pop('legacy_ssl', None) extensions.pop('keep_header_casing', None) def close(self): # Remove the logging handler that contains a reference to our logger # See: https://github.com/yt-dlp/yt-dlp/issues/8922 for name, handler in self.__logging_handlers.items(): logging.getLogger(name).removeHandler(handler) def _prepare_headers(self, request, headers): if 'cookie' not in headers: cookiejar = self._get_cookiejar(request) cookie_header = cookiejar.get_cookie_header(request.url) if cookie_header: headers['cookie'] = cookie_header def _send(self, request): timeout = self._calculate_timeout(request) headers = self._get_headers(request) wsuri = parse_uri(request.url) create_conn_kwargs = { 'source_address': (self.source_address, 0) if self.source_address else None, 'timeout': timeout, } proxy = select_proxy(request.url, self._get_proxies(request)) try: if proxy: socks_proxy_options = make_socks_proxy_opts(proxy) sock = create_connection( address=(socks_proxy_options['addr'], socks_proxy_options['port']), _create_socket_func=functools.partial( create_socks_proxy_socket, (wsuri.host, wsuri.port), socks_proxy_options), **create_conn_kwargs, ) else: sock = create_connection( address=(wsuri.host, wsuri.port), **create_conn_kwargs, ) ssl_ctx = self._make_sslcontext(legacy_ssl_support=request.extensions.get('legacy_ssl')) conn = websockets.sync.client.connect( sock=sock, uri=request.url, additional_headers=headers, open_timeout=timeout, user_agent_header=None, ssl=ssl_ctx if wsuri.secure else None, close_timeout=0, # not ideal, but prevents yt-dlp hanging ) return WebsocketsResponseAdapter(conn, url=request.url) # Exceptions as per https://websockets.readthedocs.io/en/stable/reference/sync/client.html except SocksProxyError as e: raise ProxyError(cause=e) from e except websockets.exceptions.InvalidURI as e: raise RequestError(cause=e) from e except ssl.SSLCertVerificationError as e: raise CertificateVerifyError(cause=e) from e except ssl.SSLError as e: raise SSLError(cause=e) from e except websockets.exceptions.InvalidStatus as e: raise HTTPError( Response( fp=io.BytesIO(e.response.body), url=request.url, headers=e.response.headers, status=e.response.status_code, reason=e.response.reason_phrase), ) from e except (OSError, TimeoutError, websockets.exceptions.WebSocketException) as e: raise TransportError(cause=e) from e
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/networking/__init__.py
yt_dlp/networking/__init__.py
# flake8: noqa: F401 import warnings from .common import ( HEADRequest, PATCHRequest, PUTRequest, Request, RequestDirector, RequestHandler, Response, ) # isort: split # TODO: all request handlers should be safely imported from . import _urllib from ..utils import bug_reports_message try: from . import _requests except ImportError: pass except Exception as e: warnings.warn(f'Failed to import "requests" request handler: {e}' + bug_reports_message()) try: from . import _websockets except ImportError: pass except Exception as e: warnings.warn(f'Failed to import "websockets" request handler: {e}' + bug_reports_message()) try: from . import _curlcffi except ImportError: pass except Exception as e: warnings.warn(f'Failed to import "curl_cffi" request handler: {e}' + bug_reports_message())
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/postprocessor/modify_chapters.py
yt_dlp/postprocessor/modify_chapters.py
import copy import heapq import os from .common import PostProcessor from .ffmpeg import FFmpegPostProcessor, FFmpegSubtitlesConvertorPP from .sponsorblock import SponsorBlockPP from ..utils import PostProcessingError, orderedSet, prepend_extension _TINY_CHAPTER_DURATION = 1 DEFAULT_SPONSORBLOCK_CHAPTER_TITLE = '[SponsorBlock]: %(category_names)l' class ModifyChaptersPP(FFmpegPostProcessor): def __init__(self, downloader, remove_chapters_patterns=None, remove_sponsor_segments=None, remove_ranges=None, *, sponsorblock_chapter_title=DEFAULT_SPONSORBLOCK_CHAPTER_TITLE, force_keyframes=False): FFmpegPostProcessor.__init__(self, downloader) self._remove_chapters_patterns = set(remove_chapters_patterns or []) self._remove_sponsor_segments = set(remove_sponsor_segments or []) - set(SponsorBlockPP.NON_SKIPPABLE_CATEGORIES.keys()) self._ranges_to_remove = set(remove_ranges or []) self._sponsorblock_chapter_title = sponsorblock_chapter_title self._force_keyframes = force_keyframes @PostProcessor._restrict_to(images=False) def run(self, info): self._fixup_chapters(info) # Chapters must be preserved intact when downloading multiple formats of the same video. chapters, sponsor_chapters = self._mark_chapters_to_remove( copy.deepcopy(info.get('chapters')) or [], copy.deepcopy(info.get('sponsorblock_chapters')) or []) if not chapters and not sponsor_chapters: return [], info real_duration = self._get_real_video_duration(info['filepath']) if not chapters: chapters = [{'start_time': 0, 'end_time': info.get('duration') or real_duration, 'title': info['title']}] info['chapters'], cuts = self._remove_marked_arrange_sponsors(chapters + sponsor_chapters) if not cuts: return [], info elif not info['chapters']: self.report_warning('You have requested to remove the entire video, which is not possible') return [], info original_duration, info['duration'] = info.get('duration'), info['chapters'][-1]['end_time'] if self._duration_mismatch(real_duration, original_duration, 1): if not self._duration_mismatch(real_duration, info['duration']): self.to_screen(f'Skipping {self.pp_key()} since the video appears to be already cut') return [], info if not info.get('__real_download'): raise PostProcessingError('Cannot cut video since the real and expected durations mismatch. ' 'Different chapters may have already been removed') else: self.write_debug('Expected and actual durations mismatch') concat_opts = self._make_concat_opts(cuts, real_duration) self.write_debug('Concat spec = {}'.format(', '.join(f'{c.get("inpoint", 0.0)}-{c.get("outpoint", "inf")}' for c in concat_opts))) def remove_chapters(file, is_sub): return file, self.remove_chapters(file, cuts, concat_opts, self._force_keyframes and not is_sub) in_out_files = [remove_chapters(info['filepath'], False)] in_out_files.extend(remove_chapters(in_file, True) for in_file in self._get_supported_subs(info)) # Renaming should only happen after all files are processed files_to_remove = [] for in_file, out_file in in_out_files: mtime = os.stat(in_file).st_mtime uncut_file = prepend_extension(in_file, 'uncut') os.replace(in_file, uncut_file) os.replace(out_file, in_file) self.try_utime(in_file, mtime, mtime) files_to_remove.append(uncut_file) return files_to_remove, info def _mark_chapters_to_remove(self, chapters, sponsor_chapters): if self._remove_chapters_patterns: warn_no_chapter_to_remove = True if not chapters: self.to_screen('Chapter information is unavailable') warn_no_chapter_to_remove = False for c in chapters: if any(regex.search(c['title']) for regex in self._remove_chapters_patterns): c['remove'] = True warn_no_chapter_to_remove = False if warn_no_chapter_to_remove: self.to_screen('There are no chapters matching the regex') if self._remove_sponsor_segments: warn_no_chapter_to_remove = True if not sponsor_chapters: self.to_screen('SponsorBlock information is unavailable') warn_no_chapter_to_remove = False for c in sponsor_chapters: if c['category'] in self._remove_sponsor_segments: c['remove'] = True warn_no_chapter_to_remove = False if warn_no_chapter_to_remove: self.to_screen('There are no matching SponsorBlock chapters') sponsor_chapters.extend({ 'start_time': start, 'end_time': end, 'category': 'manually_removed', '_categories': [('manually_removed', start, end, 'Manually removed')], 'remove': True, } for start, end in self._ranges_to_remove) return chapters, sponsor_chapters def _get_supported_subs(self, info): for sub in (info.get('requested_subtitles') or {}).values(): sub_file = sub.get('filepath') # The file might have been removed by --embed-subs if not sub_file or not os.path.exists(sub_file): continue ext = sub['ext'] if ext not in FFmpegSubtitlesConvertorPP.SUPPORTED_EXTS: self.report_warning(f'Cannot remove chapters from external {ext} subtitles; "{sub_file}" is now out of sync') continue # TODO: create __real_download for subs? yield sub_file def _remove_marked_arrange_sponsors(self, chapters): # Store cuts separately, since adjacent and overlapping cuts must be merged. cuts = [] def append_cut(c): assert 'remove' in c, 'Not a cut is appended to cuts' last_to_cut = cuts[-1] if cuts else None if last_to_cut and last_to_cut['end_time'] >= c['start_time']: last_to_cut['end_time'] = max(last_to_cut['end_time'], c['end_time']) else: cuts.append(c) return len(cuts) - 1 def excess_duration(c): # Cuts that are completely within the chapter reduce chapters' duration. # Since cuts can overlap, excess duration may be less that the sum of cuts' durations. # To avoid that, chapter stores the index to the fist cut within the chapter, # instead of storing excess duration. append_cut ensures that subsequent cuts (if any) # will be merged with previous ones (if necessary). cut_idx, excess = c.pop('cut_idx', len(cuts)), 0 while cut_idx < len(cuts): cut = cuts[cut_idx] if cut['start_time'] >= c['end_time']: break if cut['end_time'] > c['start_time']: excess += min(cut['end_time'], c['end_time']) excess -= max(cut['start_time'], c['start_time']) cut_idx += 1 return excess new_chapters = [] def append_chapter(c): assert 'remove' not in c, 'Cut is appended to chapters' length = c['end_time'] - c['start_time'] - excess_duration(c) # Chapter is completely covered by cuts or sponsors. if length <= 0: return start = new_chapters[-1]['end_time'] if new_chapters else 0 c.update(start_time=start, end_time=start + length) new_chapters.append(c) # Turn into a priority queue, index is a tie breaker. # Plain stack sorted by start_time is not enough: after splitting the chapter, # the part returned to the stack is not guaranteed to have start_time # less than or equal to the that of the stack's head. chapters = [(c['start_time'], i, c) for i, c in enumerate(chapters)] heapq.heapify(chapters) _, cur_i, cur_chapter = heapq.heappop(chapters) while chapters: _, i, c = heapq.heappop(chapters) # Non-overlapping chapters or cuts can be appended directly. However, # adjacent non-overlapping cuts must be merged, which is handled by append_cut. if cur_chapter['end_time'] <= c['start_time']: (append_chapter if 'remove' not in cur_chapter else append_cut)(cur_chapter) cur_i, cur_chapter = i, c continue # Eight possibilities for overlapping chapters: (cut, cut), (cut, sponsor), # (cut, normal), (sponsor, cut), (normal, cut), (sponsor, sponsor), # (sponsor, normal), and (normal, sponsor). There is no (normal, normal): # normal chapters are assumed not to overlap. if 'remove' in cur_chapter: # (cut, cut): adjust end_time. if 'remove' in c: cur_chapter['end_time'] = max(cur_chapter['end_time'], c['end_time']) # (cut, sponsor/normal): chop the beginning of the later chapter # (if it's not completely hidden by the cut). Push to the priority queue # to restore sorting by start_time: with beginning chopped, c may actually # start later than the remaining chapters from the queue. elif cur_chapter['end_time'] < c['end_time']: c['start_time'] = cur_chapter['end_time'] c['_was_cut'] = True heapq.heappush(chapters, (c['start_time'], i, c)) # (sponsor/normal, cut). elif 'remove' in c: cur_chapter['_was_cut'] = True # Chop the end of the current chapter if the cut is not contained within it. # Chopping the end doesn't break start_time sorting, no PQ push is necessary. if cur_chapter['end_time'] <= c['end_time']: cur_chapter['end_time'] = c['start_time'] append_chapter(cur_chapter) cur_i, cur_chapter = i, c continue # Current chapter contains the cut within it. If the current chapter is # a sponsor chapter, check whether the categories before and after the cut differ. if '_categories' in cur_chapter: after_c = dict(cur_chapter, start_time=c['end_time'], _categories=[]) cur_cats = [] for cat_start_end in cur_chapter['_categories']: if cat_start_end[1] < c['start_time']: cur_cats.append(cat_start_end) if cat_start_end[2] > c['end_time']: after_c['_categories'].append(cat_start_end) cur_chapter['_categories'] = cur_cats if cur_chapter['_categories'] != after_c['_categories']: # Categories before and after the cut differ: push the after part to PQ. heapq.heappush(chapters, (after_c['start_time'], cur_i, after_c)) cur_chapter['end_time'] = c['start_time'] append_chapter(cur_chapter) cur_i, cur_chapter = i, c continue # Either sponsor categories before and after the cut are the same or # we're dealing with a normal chapter. Just register an outstanding cut: # subsequent append_chapter will reduce the duration. cur_chapter.setdefault('cut_idx', append_cut(c)) # (sponsor, normal): if a normal chapter is not completely overlapped, # chop the beginning of it and push it to PQ. elif '_categories' in cur_chapter and '_categories' not in c: if cur_chapter['end_time'] < c['end_time']: c['start_time'] = cur_chapter['end_time'] c['_was_cut'] = True heapq.heappush(chapters, (c['start_time'], i, c)) # (normal, sponsor) and (sponsor, sponsor) else: assert '_categories' in c, 'Normal chapters overlap' cur_chapter['_was_cut'] = True c['_was_cut'] = True # Push the part after the sponsor to PQ. if cur_chapter['end_time'] > c['end_time']: # deepcopy to make categories in after_c and cur_chapter/c refer to different lists. after_c = dict(copy.deepcopy(cur_chapter), start_time=c['end_time']) heapq.heappush(chapters, (after_c['start_time'], cur_i, after_c)) # Push the part after the overlap to PQ. elif c['end_time'] > cur_chapter['end_time']: after_cur = dict(copy.deepcopy(c), start_time=cur_chapter['end_time']) heapq.heappush(chapters, (after_cur['start_time'], cur_i, after_cur)) c['end_time'] = cur_chapter['end_time'] # (sponsor, sponsor): merge categories in the overlap. if '_categories' in cur_chapter: c['_categories'] = cur_chapter['_categories'] + c['_categories'] # Inherit the cuts that the current chapter has accumulated within it. if 'cut_idx' in cur_chapter: c['cut_idx'] = cur_chapter['cut_idx'] cur_chapter['end_time'] = c['start_time'] append_chapter(cur_chapter) cur_i, cur_chapter = i, c (append_chapter if 'remove' not in cur_chapter else append_cut)(cur_chapter) return self._remove_tiny_rename_sponsors(new_chapters), cuts def _remove_tiny_rename_sponsors(self, chapters): new_chapters = [] for i, c in enumerate(chapters): # Merge with the previous/next if the chapter is tiny. # Only tiny chapters resulting from a cut can be skipped. # Chapters that were already tiny in the original list will be preserved. if (('_was_cut' in c or '_categories' in c) and c['end_time'] - c['start_time'] < _TINY_CHAPTER_DURATION): if not new_chapters: # Prepend tiny chapter to the next one if possible. if i < len(chapters) - 1: chapters[i + 1]['start_time'] = c['start_time'] continue else: old_c = new_chapters[-1] if i < len(chapters) - 1: next_c = chapters[i + 1] # Not a typo: key names in old_c and next_c are really different. prev_is_sponsor = 'categories' in old_c next_is_sponsor = '_categories' in next_c # Preferentially prepend tiny normals to normals and sponsors to sponsors. if (('_categories' not in c and prev_is_sponsor and not next_is_sponsor) or ('_categories' in c and not prev_is_sponsor and next_is_sponsor)): next_c['start_time'] = c['start_time'] continue old_c['end_time'] = c['end_time'] continue c.pop('_was_cut', None) cats = c.pop('_categories', None) if cats: category, _, _, category_name = min(cats, key=lambda c: c[2] - c[1]) c.update({ 'category': category, 'categories': orderedSet(x[0] for x in cats), 'name': category_name, 'category_names': orderedSet(x[3] for x in cats), }) c['title'] = self._downloader.evaluate_outtmpl(self._sponsorblock_chapter_title, c.copy()) # Merge identically named sponsors. if (new_chapters and 'categories' in new_chapters[-1] and new_chapters[-1]['title'] == c['title']): new_chapters[-1]['end_time'] = c['end_time'] continue new_chapters.append(c) return new_chapters def remove_chapters(self, filename, ranges_to_cut, concat_opts, force_keyframes=False): in_file = filename out_file = prepend_extension(in_file, 'temp') if force_keyframes: in_file = self.force_keyframes(in_file, (t for c in ranges_to_cut for t in (c['start_time'], c['end_time']))) self.to_screen(f'Removing chapters from {filename}') self.concat_files([in_file] * len(concat_opts), out_file, concat_opts) if in_file != filename: self._delete_downloaded_files(in_file, msg=None) return out_file @staticmethod def _make_concat_opts(chapters_to_remove, duration): opts = [{}] for s in chapters_to_remove: # Do not create 0 duration chunk at the beginning. if s['start_time'] == 0: opts[-1]['inpoint'] = f'{s["end_time"]:.6f}' continue opts[-1]['outpoint'] = f'{s["start_time"]:.6f}' # Do not create 0 duration chunk at the end. if s['end_time'] < duration: opts.append({'inpoint': f'{s["end_time"]:.6f}'}) return opts
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/postprocessor/sponsorblock.py
yt_dlp/postprocessor/sponsorblock.py
import hashlib import json import re import urllib.parse from .ffmpeg import FFmpegPostProcessor class SponsorBlockPP(FFmpegPostProcessor): # https://wiki.sponsor.ajay.app/w/Types EXTRACTORS = { 'Youtube': 'YouTube', } POI_CATEGORIES = { 'poi_highlight': 'Highlight', } NON_SKIPPABLE_CATEGORIES = { **POI_CATEGORIES, 'chapter': 'Chapter', } CATEGORIES = { 'sponsor': 'Sponsor', 'intro': 'Intermission/Intro Animation', 'outro': 'Endcards/Credits', 'selfpromo': 'Unpaid/Self Promotion', 'preview': 'Preview/Recap', 'filler': 'Filler Tangent', 'interaction': 'Interaction Reminder', 'music_offtopic': 'Non-Music Section', 'hook': 'Hook/Greetings', **NON_SKIPPABLE_CATEGORIES, } def __init__(self, downloader, categories=None, api='https://sponsor.ajay.app'): FFmpegPostProcessor.__init__(self, downloader) self._categories = tuple(categories or self.CATEGORIES.keys()) self._API_URL = api if re.match('https?://', api) else 'https://' + api def run(self, info): extractor = info['extractor_key'] if extractor not in self.EXTRACTORS: self.to_screen(f'SponsorBlock is not supported for {extractor}') return [], info self.to_screen('Fetching SponsorBlock segments') info['sponsorblock_chapters'] = self._get_sponsor_chapters(info, info.get('duration')) return [], info def _get_sponsor_chapters(self, info, duration): segments = self._get_sponsor_segments(info['id'], self.EXTRACTORS[info['extractor_key']]) def duration_filter(s): start_end = s['segment'] # Ignore entire video segments (https://wiki.sponsor.ajay.app/w/Types). if start_end == (0, 0): return False # Ignore milliseconds difference at the start. if start_end[0] <= 1: start_end[0] = 0 # Make POI chapters 1 sec so that we can properly mark them if s['category'] in self.POI_CATEGORIES: start_end[1] += 1 # Ignore milliseconds difference at the end. # Never allow the segment to exceed the video. if duration and duration - start_end[1] <= 1: start_end[1] = duration # SponsorBlock duration may be absent or it may deviate from the real one. diff = abs(duration - s['videoDuration']) if s['videoDuration'] else 0 return diff < 1 or (diff < 5 and diff / (start_end[1] - start_end[0]) < 0.05) duration_match = [s for s in segments if duration_filter(s)] if len(duration_match) != len(segments): self.report_warning('Some SponsorBlock segments are from a video of different duration, maybe from an old version of this video') def to_chapter(s): (start, end), cat = s['segment'], s['category'] title = s['description'] if cat == 'chapter' else self.CATEGORIES[cat] return { 'start_time': start, 'end_time': end, 'category': cat, 'title': title, 'type': s['actionType'], '_categories': [(cat, start, end, title)], } sponsor_chapters = [to_chapter(s) for s in duration_match] if not sponsor_chapters: self.to_screen('No matching segments were found in the SponsorBlock database') else: self.to_screen(f'Found {len(sponsor_chapters)} segments in the SponsorBlock database') return sponsor_chapters def _get_sponsor_segments(self, video_id, service): video_hash = hashlib.sha256(video_id.encode('ascii')).hexdigest() # SponsorBlock API recommends using first 4 hash characters. url = f'{self._API_URL}/api/skipSegments/{video_hash[:4]}?' + urllib.parse.urlencode({ 'service': service, 'categories': json.dumps(self._categories), 'actionTypes': json.dumps(['skip', 'poi', 'chapter']), }) for d in self._download_json(url) or []: if d['videoID'] == video_id: return d['segments'] return []
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/postprocessor/metadataparser.py
yt_dlp/postprocessor/metadataparser.py
import re from .common import PostProcessor from ..utils import Namespace, filter_dict, function_with_repr class MetadataParserPP(PostProcessor): def __init__(self, downloader, actions): super().__init__(downloader) self._actions = [] for f in actions: action, *args = f assert action in self.Actions self._actions.append(action(self, *args)) @classmethod def validate_action(cls, action, *data): """Each action can be: (Actions.INTERPRET, from, to) OR (Actions.REPLACE, field, search, replace) """ if action not in cls.Actions: raise ValueError(f'{action!r} is not a valid action') action(cls, *data) # So this can raise error to validate @staticmethod def field_to_template(tmpl): if re.match(r'[a-zA-Z_]+$', tmpl): return f'%({tmpl})s' from ..YoutubeDL import YoutubeDL err = YoutubeDL.validate_outtmpl(tmpl) if err: raise err return tmpl @staticmethod def format_to_regex(fmt): r""" Converts a string like '%(title)s - %(artist)s' to a regex like '(?P<title>.+)\ \-\ (?P<artist>.+)' """ if not re.search(r'%\(\w+\)s', fmt): return fmt lastpos = 0 regex = '' # replace %(..)s with regex group and escape other string parts for match in re.finditer(r'%\((\w+)\)s', fmt): regex += re.escape(fmt[lastpos:match.start()]) regex += rf'(?P<{match.group(1)}>.+)' lastpos = match.end() if lastpos < len(fmt): regex += re.escape(fmt[lastpos:]) return regex def run(self, info): for f in self._actions: f(info) return [], info @function_with_repr def interpretter(self, inp, out): def f(info): data_to_parse = self._downloader.evaluate_outtmpl(template, info) self.write_debug(f'Searching for {out_re.pattern!r} in {template!r}') match = out_re.search(data_to_parse) if match is None: self.to_screen(f'Could not interpret {inp!r} as {out!r}') return for attribute, value in filter_dict(match.groupdict()).items(): info[attribute] = value self.to_screen(f'Parsed {attribute} from {template!r}: {value!r}') template = self.field_to_template(inp) out_re = re.compile(self.format_to_regex(out)) return f @function_with_repr def replacer(self, field, search, replace): def f(info): val = info.get(field) if val is None: self.to_screen(f'Video does not have a {field}') return elif not isinstance(val, str): self.report_warning(f'Cannot replace in field {field} since it is a {type(val).__name__}') return self.write_debug(f'Replacing all {search!r} in {field} with {replace!r}') info[field], n = search_re.subn(replace, val) if n: self.to_screen(f'Changed {field} to: {info[field]}') else: self.to_screen(f'Did not find {search!r} in {field}') search_re = re.compile(search) return f Actions = Namespace(INTERPRET=interpretter, REPLACE=replacer) class MetadataFromFieldPP(MetadataParserPP): @classmethod def to_action(cls, f): match = re.match(r'(?s)(?P<in>.*?)(?<!\\):(?P<out>.+)$', f) if match is None: raise ValueError(f'it should be FROM:TO, not {f!r}') return ( cls.Actions.INTERPRET, match.group('in').replace('\\:', ':'), match.group('out'), ) def __init__(self, downloader, formats): super().__init__(downloader, [self.to_action(f) for f in formats]) # Deprecated class MetadataFromTitlePP(MetadataParserPP): def __init__(self, downloader, titleformat): super().__init__(downloader, [(self.Actions.INTERPRET, 'title', titleformat)]) self.deprecation_warning( 'yt_dlp.postprocessor.MetadataFromTitlePP is deprecated ' 'and may be removed in a future version. Use yt_dlp.postprocessor.MetadataFromFieldPP instead')
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/postprocessor/common.py
yt_dlp/postprocessor/common.py
import functools import json import os from ..networking import Request from ..networking.exceptions import HTTPError, network_exceptions from ..utils import ( PostProcessingError, RetryManager, _configuration_args, deprecation_warning, ) from ..utils._utils import _ProgressState class PostProcessorMetaClass(type): @staticmethod def run_wrapper(func): @functools.wraps(func) def run(self, info, *args, **kwargs): info_copy = self._copy_infodict(info) self._hook_progress({'status': 'started'}, info_copy) ret = func(self, info, *args, **kwargs) if ret is not None: _, info = ret self._hook_progress({'status': 'finished'}, info_copy) return ret return run def __new__(cls, name, bases, attrs): if 'run' in attrs: attrs['run'] = cls.run_wrapper(attrs['run']) return type.__new__(cls, name, bases, attrs) class PostProcessor(metaclass=PostProcessorMetaClass): """Post Processor class. PostProcessor objects can be added to downloaders with their add_post_processor() method. When the downloader has finished a successful download, it will take its internal chain of PostProcessors and start calling the run() method on each one of them, first with an initial argument and then with the returned value of the previous PostProcessor. PostProcessor objects follow a "mutual registration" process similar to InfoExtractor objects. Optionally PostProcessor can use a list of additional command-line arguments with self._configuration_args. """ _downloader = None def __init__(self, downloader=None): self._progress_hooks = [] self.add_progress_hook(self.report_progress) self.set_downloader(downloader) self.PP_NAME = self.pp_key() @classmethod def pp_key(cls): name = cls.__name__[:-2] return name[6:] if name[:6].lower() == 'ffmpeg' else name def to_screen(self, text, prefix=True, *args, **kwargs): if self._downloader: tag = f'[{self.PP_NAME}] ' if prefix else '' return self._downloader.to_screen(f'{tag}{text}', *args, **kwargs) def report_warning(self, text, *args, **kwargs): if self._downloader: return self._downloader.report_warning(text, *args, **kwargs) def deprecation_warning(self, msg): warn = getattr(self._downloader, 'deprecation_warning', deprecation_warning) return warn(msg, stacklevel=1) def deprecated_feature(self, msg): if self._downloader: return self._downloader.deprecated_feature(msg) return deprecation_warning(msg, stacklevel=1) def report_error(self, text, *args, **kwargs): self.deprecation_warning('"yt_dlp.postprocessor.PostProcessor.report_error" is deprecated. ' 'raise "yt_dlp.utils.PostProcessingError" instead') if self._downloader: return self._downloader.report_error(text, *args, **kwargs) def write_debug(self, text, *args, **kwargs): if self._downloader: return self._downloader.write_debug(text, *args, **kwargs) def _delete_downloaded_files(self, *files_to_delete, **kwargs): if self._downloader: return self._downloader._delete_downloaded_files(*files_to_delete, **kwargs) for filename in set(filter(None, files_to_delete)): os.remove(filename) def get_param(self, name, default=None, *args, **kwargs): if self._downloader: return self._downloader.params.get(name, default, *args, **kwargs) return default def set_downloader(self, downloader): """Sets the downloader for this PP.""" self._downloader = downloader for ph in getattr(downloader, '_postprocessor_hooks', []): self.add_progress_hook(ph) def _copy_infodict(self, info_dict): return getattr(self._downloader, '_copy_infodict', dict)(info_dict) @staticmethod def _restrict_to(*, video=True, audio=True, images=True, simulated=True): allowed = {'video': video, 'audio': audio, 'images': images} def decorator(func): @functools.wraps(func) def wrapper(self, info): if not simulated and (self.get_param('simulate') or self.get_param('skip_download')): return [], info format_type = ( 'video' if info.get('vcodec') != 'none' else 'audio' if info.get('acodec') != 'none' else 'images') if allowed[format_type]: return func(self, info) else: self.to_screen(f'Skipping {format_type}') return [], info return wrapper return decorator def run(self, information): """Run the PostProcessor. The "information" argument is a dictionary like the ones composed by InfoExtractors. The only difference is that this one has an extra field called "filepath" that points to the downloaded file. This method returns a tuple, the first element is a list of the files that can be deleted, and the second of which is the updated information. In addition, this method may raise a PostProcessingError exception if post processing fails. """ return [], information # by default, keep file and do nothing def try_utime(self, path, atime, mtime, errnote='Cannot update utime of file'): try: os.utime(path, (atime, mtime)) except Exception: self.report_warning(errnote) def _configuration_args(self, exe, *args, **kwargs): return _configuration_args( self.pp_key(), self.get_param('postprocessor_args'), exe, *args, **kwargs) def _hook_progress(self, status, info_dict): if not self._progress_hooks: return status.update({ 'info_dict': info_dict, 'postprocessor': self.pp_key(), }) for ph in self._progress_hooks: ph(status) def add_progress_hook(self, ph): # See YoutubeDl.py (search for postprocessor_hooks) for a description of this interface self._progress_hooks.append(ph) def report_progress(self, s): s['_default_template'] = '%(postprocessor)s %(status)s' % s # noqa: UP031 if not self._downloader: return progress_dict = s.copy() progress_dict.pop('info_dict') progress_dict = {'info': s['info_dict'], 'progress': progress_dict} progress_template = self.get_param('progress_template', {}) tmpl = progress_template.get('postprocess') if tmpl: self._downloader.to_screen( self._downloader.evaluate_outtmpl(tmpl, progress_dict), quiet=False) self._downloader.to_console_title(self._downloader.evaluate_outtmpl( progress_template.get('postprocess-title') or 'yt-dlp %(progress._default_template)s', progress_dict), _ProgressState.from_dict(s), s.get('_percent')) def _retry_download(self, err, count, retries): # While this is not an extractor, it behaves similar to one and # so obey extractor_retries and "--retry-sleep extractor" RetryManager.report_retry(err, count, retries, info=self.to_screen, warn=self.report_warning, sleep_func=self.get_param('retry_sleep_functions', {}).get('extractor')) def _download_json(self, url, *, expected_http_errors=(404,)): self.write_debug(f'{self.PP_NAME} query: {url}') for retry in RetryManager(self.get_param('extractor_retries', 3), self._retry_download): try: rsp = self._downloader.urlopen(Request(url)) except network_exceptions as e: if isinstance(e, HTTPError) and e.status in expected_http_errors: return None retry.error = PostProcessingError(f'Unable to communicate with {self.PP_NAME} API: {e}') continue return json.loads(rsp.read().decode(rsp.headers.get_param('charset') or 'utf-8')) class AudioConversionError(PostProcessingError): # Deprecated pass
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/postprocessor/ffmpeg.py
yt_dlp/postprocessor/ffmpeg.py
import collections import contextvars import functools import itertools import json import os import re import subprocess import time from .common import PostProcessor from ..compat import imghdr from ..utils import ( MEDIA_EXTENSIONS, ISO639Utils, Popen, PostProcessingError, _get_exe_version_output, deprecation_warning, detect_exe_version, determine_ext, dfxp2srt, encodeArgument, filter_dict, float_or_none, is_outdated_version, orderedSet, prepend_extension, replace_extension, shell_quote, traverse_obj, variadic, write_json_file, ) EXT_TO_OUT_FORMATS = { 'aac': 'adts', 'flac': 'flac', 'm4a': 'ipod', 'mka': 'matroska', 'mkv': 'matroska', 'mpg': 'mpeg', 'ogv': 'ogg', 'ts': 'mpegts', 'wma': 'asf', 'wmv': 'asf', 'weba': 'webm', 'vtt': 'webvtt', } ACODECS = { # name: (ext, encoder, opts) 'mp3': ('mp3', 'libmp3lame', ()), 'aac': ('m4a', 'aac', ('-f', 'adts')), 'm4a': ('m4a', 'aac', ('-bsf:a', 'aac_adtstoasc')), 'opus': ('opus', 'libopus', ()), 'vorbis': ('ogg', 'libvorbis', ()), 'flac': ('flac', 'flac', ()), 'alac': ('m4a', None, ('-acodec', 'alac')), 'wav': ('wav', None, ('-f', 'wav')), } def create_mapping_re(supported): return re.compile(r'{0}(?:/{0})*$'.format(r'(?:\s*\w+\s*>)?\s*(?:{})\s*'.format('|'.join(supported)))) def resolve_mapping(source, mapping): """ Get corresponding item from a mapping string like 'A>B/C>D/E' @returns (target, error_message) """ for pair in mapping.lower().split('/'): kv = pair.split('>', 1) if len(kv) == 1 or kv[0].strip() == source: target = kv[-1].strip() if target == source: return target, f'already is in target format {source}' return target, None return None, f'could not find a mapping for {source}' class FFmpegPostProcessorError(PostProcessingError): pass class FFmpegPostProcessor(PostProcessor): _ffmpeg_location = contextvars.ContextVar('ffmpeg_location', default=None) def __init__(self, downloader=None): PostProcessor.__init__(self, downloader) self._paths = self._determine_executables() @staticmethod def get_versions_and_features(downloader=None): pp = FFmpegPostProcessor(downloader) return pp._versions, pp._features @staticmethod def get_versions(downloader=None): return FFmpegPostProcessor.get_versions_and_features(downloader)[0] def _determine_executables(self): programs = ['ffmpeg', 'ffprobe'] location = self.get_param('ffmpeg_location', self._ffmpeg_location.get()) if location is None: return {p: p for p in programs} if not os.path.exists(location): self.report_warning( f'ffmpeg-location {location} does not exist! Continuing without ffmpeg', only_once=True) return {} elif os.path.isdir(location): dirname, basename, filename = location, None, None else: filename = os.path.basename(location) basename = next((p for p in programs if p in filename), 'ffmpeg') dirname = os.path.dirname(os.path.abspath(location)) paths = {p: os.path.join(dirname, p) for p in programs} if basename and basename in filename: for p in programs: path = os.path.join(dirname, filename.replace(basename, p)) if os.path.exists(path): paths[p] = path if basename: paths[basename] = location return paths _version_cache, _features_cache = {None: None}, {} def _get_ffmpeg_version(self, prog): path = self._paths.get(prog) if path in self._version_cache: return self._version_cache[path], self._features_cache.get(path, {}) out = _get_exe_version_output(path, ['-bsfs']) ver = detect_exe_version(out) if out else False if ver: regexs = [ r'(?:\d+:)?([0-9.]+)-[0-9]+ubuntu[0-9.]+$', # Ubuntu, see [1] r'n([0-9.]+)$', # Arch Linux # 1. http://www.ducea.com/2006/06/17/ubuntu-package-version-naming-explanation/ ] for regex in regexs: mobj = re.match(regex, ver) if mobj: ver = mobj.group(1) self._version_cache[path] = ver if prog != 'ffmpeg' or not out: return ver, {} mobj = re.search(r'(?m)^\s+libavformat\s+(?:[0-9. ]+)\s+/\s+(?P<runtime>[0-9. ]+)', out) lavf_runtime_version = mobj.group('runtime').replace(' ', '') if mobj else None self._features_cache[path] = features = { 'fdk': '--enable-libfdk-aac' in out, 'setts': 'setts' in out.splitlines(), 'needs_adtstoasc': is_outdated_version(lavf_runtime_version, '57.56.100', False), } return ver, features @property def _versions(self): return filter_dict({self.basename: self._version, self.probe_basename: self._probe_version}) @functools.cached_property def basename(self): _ = self._version # run property return self.basename @functools.cached_property def probe_basename(self): _ = self._probe_version # run property return self.probe_basename def _get_version(self, kind): executables = (kind, ) basename, version, features = next(filter( lambda x: x[1], ((p, *self._get_ffmpeg_version(p)) for p in executables)), (None, None, {})) if kind == 'ffmpeg': self.basename, self._features = basename, features else: self.probe_basename = basename return version @functools.cached_property def _version(self): return self._get_version('ffmpeg') @functools.cached_property def _probe_version(self): return self._get_version('ffprobe') @property def available(self): # If we return that ffmpeg is available, then the basename property *must* be run # (as doing so has side effects), and its value can never be None # See: https://github.com/yt-dlp/yt-dlp/issues/12829 return self.basename is not None @property def executable(self): return self._paths.get(self.basename) @property def probe_available(self): return self.probe_basename is not None @property def probe_executable(self): return self._paths.get(self.probe_basename) @staticmethod def stream_copy_opts(copy=True, *, ext=None): yield from ('-map', '0') # Don't copy Apple TV chapters track, bin_data # See https://github.com/yt-dlp/yt-dlp/issues/2, #19042, #19024, https://trac.ffmpeg.org/ticket/6016 yield from ('-dn', '-ignore_unknown') if copy: yield from ('-c', 'copy') if ext in ('mp4', 'mov', 'm4a'): yield from ('-c:s', 'mov_text') def check_version(self): if not self.available: raise FFmpegPostProcessorError('ffmpeg not found. Please install or provide the path using --ffmpeg-location') required_version = '1.0' if is_outdated_version(self._version, required_version): self.report_warning(f'Your copy of {self.basename} is outdated, update {self.basename} ' f'to version {required_version} or newer if you encounter any errors') def get_audio_codec(self, path): if not self.probe_available and not self.available: raise PostProcessingError('ffprobe and ffmpeg not found. Please install or provide the path using --ffmpeg-location') try: if self.probe_available: cmd = [ self.probe_executable, encodeArgument('-show_streams')] else: cmd = [ self.executable, encodeArgument('-i')] cmd.append(self._ffmpeg_filename_argument(path)) self.write_debug(f'{self.basename} command line: {shell_quote(cmd)}') stdout, stderr, returncode = Popen.run( cmd, text=True, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE) if returncode != (0 if self.probe_available else 1): return None except OSError: return None output = stdout if self.probe_available else stderr if self.probe_available: audio_codec = None for line in output.split('\n'): if line.startswith('codec_name='): audio_codec = line.split('=')[1].strip() elif line.strip() == 'codec_type=audio' and audio_codec is not None: return audio_codec else: # Stream #FILE_INDEX:STREAM_INDEX[STREAM_ID](LANGUAGE): CODEC_TYPE: CODEC_NAME mobj = re.search( r'Stream\s*#\d+:\d+(?:\[0x[0-9a-f]+\])?(?:\([a-z]{3}\))?:\s*Audio:\s*([0-9a-z]+)', output) if mobj: return mobj.group(1) return None def get_metadata_object(self, path, opts=[]): if self.probe_basename != 'ffprobe': if self.probe_available: self.report_warning('Only ffprobe is supported for metadata extraction') raise PostProcessingError('ffprobe not found. Please install or provide the path using --ffmpeg-location') self.check_version() cmd = [ self.probe_executable, encodeArgument('-hide_banner'), encodeArgument('-show_format'), encodeArgument('-show_streams'), encodeArgument('-print_format'), encodeArgument('json'), ] cmd += opts cmd.append(self._ffmpeg_filename_argument(path)) self.write_debug(f'ffprobe command line: {shell_quote(cmd)}') stdout, _, _ = Popen.run(cmd, text=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE, stdin=subprocess.PIPE) return json.loads(stdout) def get_stream_number(self, path, keys, value): streams = self.get_metadata_object(path)['streams'] num = next( (i for i, stream in enumerate(streams) if traverse_obj(stream, keys, casesense=False) == value), None) return num, len(streams) def _fixup_chapters(self, info): last_chapter = traverse_obj(info, ('chapters', -1)) if last_chapter and not last_chapter.get('end_time'): last_chapter['end_time'] = self._get_real_video_duration(info['filepath']) def _get_real_video_duration(self, filepath, fatal=True): try: duration = float_or_none( traverse_obj(self.get_metadata_object(filepath), ('format', 'duration'))) if not duration: raise PostProcessingError('ffprobe returned empty duration') return duration except PostProcessingError as e: if fatal: raise PostProcessingError(f'Unable to determine video duration: {e.msg}') def _duration_mismatch(self, d1, d2, tolerance=2): if not d1 or not d2: return None # The duration is often only known to nearest second. So there can be <1sec disparity natually. # Further excuse an additional <1sec difference. return abs(d1 - d2) > tolerance def run_ffmpeg_multiple_files(self, input_paths, out_path, opts, **kwargs): return self.real_run_ffmpeg( [(path, []) for path in input_paths], [(out_path, opts)], **kwargs) def real_run_ffmpeg(self, input_path_opts, output_path_opts, *, expected_retcodes=(0,)): self.check_version() oldest_mtime = min( os.stat(path).st_mtime for path, _ in input_path_opts if path) cmd = [self.executable, encodeArgument('-y')] # avconv does not have repeat option if self.basename == 'ffmpeg': cmd += [encodeArgument('-loglevel'), encodeArgument('repeat+info')] def make_args(file, args, name, number): keys = [f'_{name}{number}', f'_{name}'] if name == 'o': args += ['-movflags', '+faststart'] if number == 1: keys.append('') args += self._configuration_args(self.basename, keys) if name == 'i': args.append('-i') return ( [encodeArgument(arg) for arg in args] + [self._ffmpeg_filename_argument(file)]) for arg_type, path_opts in (('i', input_path_opts), ('o', output_path_opts)): cmd += itertools.chain.from_iterable( make_args(path, list(opts), arg_type, i + 1) for i, (path, opts) in enumerate(path_opts) if path) self.write_debug(f'ffmpeg command line: {shell_quote(cmd)}') _, stderr, returncode = Popen.run( cmd, text=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE, stdin=subprocess.PIPE) if returncode not in variadic(expected_retcodes): self.write_debug(stderr) raise FFmpegPostProcessorError(stderr.strip().splitlines()[-1]) for out_path, _ in output_path_opts: if out_path: self.try_utime(out_path, oldest_mtime, oldest_mtime) return stderr def run_ffmpeg(self, path, out_path, opts, **kwargs): return self.run_ffmpeg_multiple_files([path], out_path, opts, **kwargs) @staticmethod def _ffmpeg_filename_argument(fn): # Always use 'file:' because the filename may contain ':' (ffmpeg # interprets that as a protocol) or can start with '-' (-- is broken in # ffmpeg, see https://ffmpeg.org/trac/ffmpeg/ticket/2127 for details) # Also leave '-' intact in order not to break streaming to stdout. if fn.startswith(('http://', 'https://')): return fn return 'file:' + fn if fn != '-' else fn @staticmethod def _quote_for_ffmpeg(string): # See https://ffmpeg.org/ffmpeg-utils.html#toc-Quoting-and-escaping # A sequence of '' produces '\'''\''; # final replace removes the empty '' between \' \'. string = string.replace("'", r"'\''").replace("'''", "'") # Handle potential ' at string boundaries. string = string[1:] if string[0] == "'" else "'" + string return string[:-1] if string[-1] == "'" else string + "'" def force_keyframes(self, filename, timestamps): timestamps = orderedSet(timestamps) if timestamps[0] == 0: timestamps = timestamps[1:] keyframe_file = prepend_extension(filename, 'keyframes.temp') self.to_screen(f'Re-encoding "{filename}" with appropriate keyframes') self.run_ffmpeg(filename, keyframe_file, [ *self.stream_copy_opts(False, ext=determine_ext(filename)), '-force_key_frames', ','.join(f'{t:.6f}' for t in timestamps)]) return keyframe_file def concat_files(self, in_files, out_file, concat_opts=None): """ Use concat demuxer to concatenate multiple files having identical streams. Only inpoint, outpoint, and duration concat options are supported. See https://ffmpeg.org/ffmpeg-formats.html#concat-1 for details """ concat_file = f'{out_file}.concat' self.write_debug(f'Writing concat spec to {concat_file}') with open(concat_file, 'w', encoding='utf-8') as f: f.writelines(self._concat_spec(in_files, concat_opts)) out_flags = list(self.stream_copy_opts(ext=determine_ext(out_file))) self.real_run_ffmpeg( [(concat_file, ['-hide_banner', '-nostdin', '-f', 'concat', '-safe', '0'])], [(out_file, out_flags)]) self._delete_downloaded_files(concat_file) @classmethod def _concat_spec(cls, in_files, concat_opts=None): if concat_opts is None: concat_opts = [{}] * len(in_files) yield 'ffconcat version 1.0\n' for file, opts in zip(in_files, concat_opts, strict=True): yield f'file {cls._quote_for_ffmpeg(cls._ffmpeg_filename_argument(file))}\n' # Iterate explicitly to yield the following directives in order, ignoring the rest. for directive in 'inpoint', 'outpoint', 'duration': if directive in opts: yield f'{directive} {opts[directive]}\n' class FFmpegExtractAudioPP(FFmpegPostProcessor): COMMON_AUDIO_EXTS = (*MEDIA_EXTENSIONS.common_audio, 'wma') SUPPORTED_EXTS = tuple(ACODECS.keys()) FORMAT_RE = create_mapping_re(('best', *SUPPORTED_EXTS)) def __init__(self, downloader=None, preferredcodec=None, preferredquality=None, nopostoverwrites=False): FFmpegPostProcessor.__init__(self, downloader) self.mapping = preferredcodec or 'best' self._preferredquality = float_or_none(preferredquality) self._nopostoverwrites = nopostoverwrites def _quality_args(self, codec): if self._preferredquality is None: return [] elif self._preferredquality > 10: return ['-b:a', f'{self._preferredquality}k'] limits = { 'libmp3lame': (10, 0), 'libvorbis': (0, 10), # FFmpeg's AAC encoder does not have an upper limit for the value of -q:a. # Experimentally, with values over 4, bitrate changes were minimal or non-existent 'aac': (0.1, 4), 'libfdk_aac': (1, 5), }.get(codec) if not limits: return [] q = limits[1] + (limits[0] - limits[1]) * (self._preferredquality / 10) if codec == 'libfdk_aac': return ['-vbr', f'{int(q)}'] return ['-q:a', f'{q}'] def run_ffmpeg(self, path, out_path, codec, more_opts): if codec is None: acodec_opts = [] else: acodec_opts = ['-acodec', codec] opts = ['-vn', *acodec_opts, *more_opts] try: FFmpegPostProcessor.run_ffmpeg(self, path, out_path, opts) except FFmpegPostProcessorError as err: raise PostProcessingError(f'audio conversion failed: {err.msg}') @PostProcessor._restrict_to(images=False) def run(self, information): orig_path = path = information['filepath'] target_format, _skip_msg = resolve_mapping(information['ext'], self.mapping) if target_format == 'best' and information['ext'] in self.COMMON_AUDIO_EXTS: target_format, _skip_msg = None, 'the file is already in a common audio format' if not target_format: self.to_screen(f'Not converting audio {orig_path}; {_skip_msg}') return [], information filecodec = self.get_audio_codec(path) if filecodec is None: raise PostProcessingError('WARNING: unable to obtain file audio codec with ffprobe') if filecodec == 'aac' and target_format in ('m4a', 'best'): # Lossless, but in another container extension, _, more_opts, acodec = *ACODECS['m4a'], 'copy' elif target_format == 'best' or target_format == filecodec: # Lossless if possible try: extension, _, more_opts, acodec = *ACODECS[filecodec], 'copy' except KeyError: extension, acodec, more_opts = ACODECS['mp3'] else: # We convert the audio (lossy if codec is lossy) extension, acodec, more_opts = ACODECS[target_format] if acodec == 'aac' and self._features.get('fdk'): acodec, more_opts = 'libfdk_aac', [] more_opts = list(more_opts) if acodec != 'copy': more_opts = self._quality_args(acodec) temp_path = new_path = replace_extension(path, extension, information['ext']) if new_path == path: if acodec == 'copy': self.to_screen(f'Not converting audio {orig_path}; file is already in target format {target_format}') return [], information orig_path = prepend_extension(path, 'orig') temp_path = prepend_extension(path, 'temp') if (self._nopostoverwrites and os.path.exists(new_path) and os.path.exists(orig_path)): self.to_screen(f'Post-process file {new_path} exists, skipping') return [], information self.to_screen(f'Destination: {new_path}') self.run_ffmpeg(path, temp_path, acodec, more_opts) os.replace(path, orig_path) os.replace(temp_path, new_path) information['filepath'] = new_path information['ext'] = extension # Try to update the date time for extracted audio file. if information.get('filetime') is not None: self.try_utime( new_path, time.time(), information['filetime'], errnote='Cannot update utime of audio file') return [orig_path], information class FFmpegVideoConvertorPP(FFmpegPostProcessor): SUPPORTED_EXTS = ( *sorted((*MEDIA_EXTENSIONS.common_video, 'gif')), *sorted((*MEDIA_EXTENSIONS.common_audio, 'aac', 'vorbis')), ) FORMAT_RE = create_mapping_re(SUPPORTED_EXTS) _ACTION = 'converting' def __init__(self, downloader=None, preferedformat=None): super().__init__(downloader) self.mapping = preferedformat @staticmethod def _options(target_ext): yield from FFmpegPostProcessor.stream_copy_opts(False) if target_ext == 'avi': yield from ('-c:v', 'libxvid', '-vtag', 'XVID') @PostProcessor._restrict_to(images=False) def run(self, info): filename, source_ext = info['filepath'], info['ext'].lower() target_ext, _skip_msg = resolve_mapping(source_ext, self.mapping) if _skip_msg: self.to_screen(f'Not {self._ACTION} media file "{filename}"; {_skip_msg}') return [], info outpath = replace_extension(filename, target_ext, source_ext) self.to_screen(f'{self._ACTION.title()} video from {source_ext} to {target_ext}; Destination: {outpath}') self.run_ffmpeg(filename, outpath, self._options(target_ext)) info['filepath'] = outpath info['format'] = info['ext'] = target_ext return [filename], info class FFmpegVideoRemuxerPP(FFmpegVideoConvertorPP): _ACTION = 'remuxing' @staticmethod def _options(target_ext): return FFmpegPostProcessor.stream_copy_opts() class FFmpegEmbedSubtitlePP(FFmpegPostProcessor): SUPPORTED_EXTS = ('mp4', 'mov', 'm4a', 'webm', 'mkv', 'mka') def __init__(self, downloader=None, already_have_subtitle=False): super().__init__(downloader) self._already_have_subtitle = already_have_subtitle @PostProcessor._restrict_to(images=False) def run(self, info): if info['ext'] not in self.SUPPORTED_EXTS: self.to_screen(f'Subtitles can only be embedded in {", ".join(self.SUPPORTED_EXTS)} files') return [], info subtitles = info.get('requested_subtitles') if not subtitles: self.to_screen('There aren\'t any subtitles to embed') return [], info filename = info['filepath'] # Disabled temporarily. There needs to be a way to override this # in case of duration actually mismatching in extractor # See: https://github.com/yt-dlp/yt-dlp/issues/1870, https://github.com/yt-dlp/yt-dlp/issues/1385 ''' if info.get('duration') and not info.get('__real_download') and self._duration_mismatch( self._get_real_video_duration(filename, False), info['duration']): self.to_screen(f'Skipping {self.pp_key()} since the real and expected durations mismatch') return [], info ''' ext = info['ext'] sub_langs, sub_names, sub_filenames = [], [], [] webm_vtt_warn = False mp4_ass_warn = False for lang, sub_info in subtitles.items(): if not os.path.exists(sub_info.get('filepath', '')): self.report_warning(f'Skipping embedding {lang} subtitle because the file is missing') continue sub_ext = sub_info['ext'] if sub_ext == 'json': self.report_warning('JSON subtitles cannot be embedded') elif ext != 'webm' or (ext == 'webm' and sub_ext == 'vtt'): sub_langs.append(lang) sub_names.append(sub_info.get('name')) sub_filenames.append(sub_info['filepath']) else: if not webm_vtt_warn and ext == 'webm' and sub_ext != 'vtt': webm_vtt_warn = True self.report_warning('Only WebVTT subtitles can be embedded in webm files') if not mp4_ass_warn and ext == 'mp4' and sub_ext == 'ass': mp4_ass_warn = True self.report_warning('ASS subtitles cannot be properly embedded in mp4 files; expect issues') if not sub_langs: return [], info input_files = [filename, *sub_filenames] opts = [ *self.stream_copy_opts(ext=info['ext']), # Don't copy the existing subtitles, we may be running the # postprocessor a second time '-map', '-0:s', ] for i, (lang, name) in enumerate(zip(sub_langs, sub_names, strict=True)): opts.extend(['-map', f'{i + 1}:0']) lang_code = ISO639Utils.short2long(lang) or lang opts.extend([f'-metadata:s:s:{i}', f'language={lang_code}']) if name: opts.extend([f'-metadata:s:s:{i}', f'handler_name={name}', f'-metadata:s:s:{i}', f'title={name}']) temp_filename = prepend_extension(filename, 'temp') self.to_screen(f'Embedding subtitles in "{filename}"') self.run_ffmpeg_multiple_files(input_files, temp_filename, opts) os.replace(temp_filename, filename) files_to_delete = [] if self._already_have_subtitle else sub_filenames return files_to_delete, info class FFmpegMetadataPP(FFmpegPostProcessor): def __init__(self, downloader, add_metadata=True, add_chapters=True, add_infojson='if_exists'): FFmpegPostProcessor.__init__(self, downloader) self._add_metadata = add_metadata self._add_chapters = add_chapters self._add_infojson = add_infojson @staticmethod def _options(target_ext): audio_only = target_ext == 'm4a' yield from FFmpegPostProcessor.stream_copy_opts(not audio_only) if audio_only: yield from ('-vn', '-acodec', 'copy') @PostProcessor._restrict_to(images=False) def run(self, info): self._fixup_chapters(info) filename, metadata_filename = info['filepath'], None files_to_delete, options = [], [] if self._add_chapters and info.get('chapters'): metadata_filename = replace_extension(filename, 'meta') options.extend(self._get_chapter_opts(info['chapters'], metadata_filename)) files_to_delete.append(metadata_filename) if self._add_metadata: options.extend(self._get_metadata_opts(info)) if self._add_infojson: if info['ext'] in ('mkv', 'mka'): infojson_filename = info.get('infojson_filename') options.extend(self._get_infojson_opts(info, infojson_filename)) if not infojson_filename: files_to_delete.append(info.get('infojson_filename')) elif self._add_infojson is True: self.to_screen('The info-json can only be attached to mkv/mka files') if not options: self.to_screen('There isn\'t any metadata to add') return [], info temp_filename = prepend_extension(filename, 'temp') self.to_screen(f'Adding metadata to "{filename}"') self.run_ffmpeg_multiple_files( (filename, metadata_filename), temp_filename, itertools.chain(self._options(info['ext']), *options)) self._delete_downloaded_files(*files_to_delete) os.replace(temp_filename, filename) return [], info @staticmethod def _get_chapter_opts(chapters, metadata_filename): with open(metadata_filename, 'w', encoding='utf-8') as f: def ffmpeg_escape(text): return re.sub(r'([\\=;#\n])', r'\\\1', text) metadata_file_content = ';FFMETADATA1\n' for chapter in chapters: metadata_file_content += '[CHAPTER]\nTIMEBASE=1/1000\n' metadata_file_content += 'START=%d\n' % (chapter['start_time'] * 1000) metadata_file_content += 'END=%d\n' % (chapter['end_time'] * 1000) chapter_title = chapter.get('title') if chapter_title: metadata_file_content += f'title={ffmpeg_escape(chapter_title)}\n' f.write(metadata_file_content) yield ('-map_metadata', '1') def _get_metadata_opts(self, info): meta_prefix = 'meta' metadata = collections.defaultdict(dict) def add(meta_list, info_list=None): value = next(( info[key] for key in [f'{meta_prefix}_', *variadic(info_list or meta_list)] if info.get(key) is not None), None) if value not in ('', None): value = ', '.join(map(str, variadic(value))) value = value.replace('\0', '') # nul character cannot be passed in command line metadata['common'].update(dict.fromkeys(variadic(meta_list), value)) # Info on media metadata/metadata supported by ffmpeg: # https://wiki.multimedia.cx/index.php/FFmpeg_Metadata # https://kdenlive.org/en/project/adding-meta-data-to-mp4-video/ # https://kodi.wiki/view/Video_file_tagging add('title', ('track', 'title')) add('date', 'upload_date') add(('description', 'synopsis'), 'description') add(('purl', 'comment'), 'webpage_url') add('track', 'track_number') add('artist', ('artist', 'artists', 'creator', 'creators', 'uploader', 'uploader_id')) add('composer', ('composer', 'composers')) add('genre', ('genre', 'genres', 'categories', 'tags')) add('album', ('album', 'series')) add('album_artist', ('album_artist', 'album_artists')) add('disc', 'disc_number') add('show', 'series') add('season_number') add('episode_id', ('episode', 'episode_id')) add('episode_sort', 'episode_number') if 'embed-metadata' in self.get_param('compat_opts', []): add('comment', 'description') metadata['common'].pop('synopsis', None) meta_regex = rf'{re.escape(meta_prefix)}(?P<i>\d+)?_(?P<key>.+)' for key, value in info.items(): mobj = re.fullmatch(meta_regex, key) if value is not None and mobj: metadata[mobj.group('i') or 'common'][mobj.group('key')] = value.replace('\0', '') # Write id3v1 metadata also since Windows Explorer can't handle id3v2 tags yield ('-write_id3v1', '1') for name, value in metadata['common'].items(): yield ('-metadata', f'{name}={value}') stream_idx = 0 for fmt in info.get('requested_formats') or [info]: stream_count = 2 if 'none' not in (fmt.get('vcodec'), fmt.get('acodec')) else 1 lang = ISO639Utils.short2long(fmt.get('language') or '') or fmt.get('language') for i in range(stream_idx, stream_idx + stream_count): if lang: metadata[str(i)].setdefault('language', lang) for name, value in metadata[str(i)].items(): yield (f'-metadata:s:{i}', f'{name}={value}') stream_idx += stream_count def _get_infojson_opts(self, info, infofn): if not infofn or not os.path.exists(infofn): if self._add_infojson is not True: return infofn = infofn or '%s.temp' % ( self._downloader.prepare_filename(info, 'infojson') or replace_extension(self._downloader.prepare_filename(info), 'info.json', info['ext'])) if not self._downloader._ensure_dir_exists(infofn): return self.write_debug(f'Writing info-json to: {infofn}')
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
true
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/postprocessor/movefilesafterdownload.py
yt_dlp/postprocessor/movefilesafterdownload.py
import os from .common import PostProcessor from ..compat import shutil from ..utils import ( PostProcessingError, make_dir, ) class MoveFilesAfterDownloadPP(PostProcessor): def __init__(self, downloader=None, downloaded=True): PostProcessor.__init__(self, downloader) self._downloaded = downloaded @classmethod def pp_key(cls): return 'MoveFiles' def run(self, info): dl_path, dl_name = os.path.split(info['filepath']) finaldir = info.get('__finaldir', dl_path) finalpath = os.path.join(finaldir, dl_name) if self._downloaded: info['__files_to_move'][info['filepath']] = finalpath make_newfilename = lambda old: os.path.join(finaldir, os.path.basename(old)) for oldfile, newfile in info['__files_to_move'].items(): if not newfile: newfile = make_newfilename(oldfile) if os.path.abspath(oldfile) == os.path.abspath(newfile): continue if not os.path.exists(oldfile): self.report_warning(f'File "{oldfile}" cannot be found') continue if os.path.exists(newfile): if self.get_param('overwrites', True): self.report_warning(f'Replacing existing file "{newfile}"') os.remove(newfile) else: self.report_warning( f'Cannot move file "{oldfile}" out of temporary directory since "{newfile}" already exists. ') continue make_dir(newfile, PostProcessingError) self.to_screen(f'Moving file "{oldfile}" to "{newfile}"') shutil.move(oldfile, newfile) # os.rename cannot move between volumes info['filepath'] = finalpath return [], info
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/postprocessor/xattrpp.py
yt_dlp/postprocessor/xattrpp.py
import os import sys from .common import PostProcessor from ..utils import ( PostProcessingError, XAttrMetadataError, XAttrUnavailableError, hyphenate_date, write_xattr, ) class XAttrMetadataPP(PostProcessor): """Set extended attributes on downloaded file (if xattr support is found) More info about extended attributes for media: http://freedesktop.org/wiki/CommonExtendedAttributes/ http://www.freedesktop.org/wiki/PhreedomDraft/ http://dublincore.org/documents/usageguide/elements.shtml TODO: * capture youtube keywords and put them in 'user.dublincore.subject' (comma-separated) * figure out which xattrs can be used for 'duration', 'thumbnail', 'resolution' """ XATTR_MAPPING = { 'user.xdg.referrer.url': 'webpage_url', 'user.dublincore.title': 'title', 'user.dublincore.date': 'upload_date', 'user.dublincore.contributor': 'uploader', 'user.dublincore.format': 'format', # We do this last because it may get us close to the xattr limits # (e.g., 4kB on ext4), and we don't want to have the other ones fail 'user.dublincore.description': 'description', # 'user.xdg.comment': 'description', 'com.apple.metadata:kMDItemWhereFroms': 'webpage_url', } APPLE_PLIST_TEMPLATE = '''<?xml version="1.0" encoding="UTF-8"?> <!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd"> <plist version="1.0"> <array> \t<string>%s</string> </array> </plist>''' def run(self, info): mtime = os.stat(info['filepath']).st_mtime self.to_screen('Writing metadata to file\'s xattrs') for xattrname, infoname in self.XATTR_MAPPING.items(): try: value = info.get(infoname) if value: if infoname == 'upload_date': value = hyphenate_date(value) elif xattrname == 'com.apple.metadata:kMDItemWhereFroms': # Colon in xattr name throws errors on Windows/NTFS and Linux if sys.platform != 'darwin': continue value = self.APPLE_PLIST_TEMPLATE % value write_xattr(info['filepath'], xattrname, value.encode()) except XAttrUnavailableError as e: raise PostProcessingError(str(e)) except XAttrMetadataError as e: if e.reason == 'NO_SPACE': self.report_warning( 'There\'s no disk space left, disk quota exceeded or filesystem xattr limit exceeded. ' f'Extended attribute "{xattrname}" was not written.') elif e.reason == 'VALUE_TOO_LONG': self.report_warning(f'Unable to write extended attribute "{xattrname}" due to too long values.') else: tip = ('You need to use NTFS' if os.name == 'nt' else 'You may have to enable them in your "/etc/fstab"') raise PostProcessingError(f'This filesystem doesn\'t support extended attributes. {tip}') self.try_utime(info['filepath'], mtime, mtime) return [], info
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/postprocessor/__init__.py
yt_dlp/postprocessor/__init__.py
# flake8: noqa: F401 from .common import PostProcessor from .embedthumbnail import EmbedThumbnailPP from .exec import ExecAfterDownloadPP, ExecPP from .ffmpeg import ( FFmpegConcatPP, FFmpegCopyStreamPP, FFmpegEmbedSubtitlePP, FFmpegExtractAudioPP, FFmpegFixupDuplicateMoovPP, FFmpegFixupDurationPP, FFmpegFixupM3u8PP, FFmpegFixupM4aPP, FFmpegFixupStretchedPP, FFmpegFixupTimestampPP, FFmpegMergerPP, FFmpegMetadataPP, FFmpegPostProcessor, FFmpegSplitChaptersPP, FFmpegSubtitlesConvertorPP, FFmpegThumbnailsConvertorPP, FFmpegVideoConvertorPP, FFmpegVideoRemuxerPP, ) from .metadataparser import ( MetadataFromFieldPP, MetadataFromTitlePP, MetadataParserPP, ) from .modify_chapters import ModifyChaptersPP from .movefilesafterdownload import MoveFilesAfterDownloadPP from .sponsorblock import SponsorBlockPP from .xattrpp import XAttrMetadataPP from ..globals import plugin_pps, postprocessors from ..plugins import PACKAGE_NAME, register_plugin_spec, PluginSpec from ..utils import deprecation_warning def __getattr__(name): lookup = plugin_pps.value if name in lookup: deprecation_warning( f'Importing a plugin Post-Processor from {__name__} is deprecated. ' f'Please import {PACKAGE_NAME}.postprocessor.{name} instead.') return lookup[name] raise AttributeError(f'module {__name__!r} has no attribute {name!r}') def get_postprocessor(key): return postprocessors.value[key + 'PP'] register_plugin_spec(PluginSpec( module_name='postprocessor', suffix='PP', destination=postprocessors, plugin_destination=plugin_pps, )) _default_pps = { name: value for name, value in globals().items() if name.endswith('PP') or name in ('FFmpegPostProcessor', 'PostProcessor') } postprocessors.value.update(_default_pps) __all__ = list(_default_pps.values())
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/postprocessor/embedthumbnail.py
yt_dlp/postprocessor/embedthumbnail.py
import base64 import os import re import subprocess from .common import PostProcessor from .ffmpeg import FFmpegPostProcessor, FFmpegThumbnailsConvertorPP from ..compat import imghdr from ..dependencies import mutagen from ..utils import ( Popen, PostProcessingError, check_executable, encodeArgument, prepend_extension, shell_quote, ) if mutagen: from mutagen.flac import FLAC, Picture from mutagen.mp4 import MP4, MP4Cover from mutagen.oggopus import OggOpus from mutagen.oggvorbis import OggVorbis class EmbedThumbnailPPError(PostProcessingError): pass class EmbedThumbnailPP(FFmpegPostProcessor): def __init__(self, downloader=None, already_have_thumbnail=False): FFmpegPostProcessor.__init__(self, downloader) self._already_have_thumbnail = already_have_thumbnail def _get_thumbnail_resolution(self, filename, thumbnail_dict): def guess(): width, height = thumbnail_dict.get('width'), thumbnail_dict.get('height') if width and height: return width, height try: size_regex = r',\s*(?P<w>\d+)x(?P<h>\d+)\s*[,\[]' size_result = self.run_ffmpeg(filename, None, ['-hide_banner'], expected_retcodes=(1,)) mobj = re.search(size_regex, size_result) if mobj is None: return guess() except PostProcessingError as err: self.report_warning(f'unable to find the thumbnail resolution; {err}') return guess() return int(mobj.group('w')), int(mobj.group('h')) def _report_run(self, exe, filename): self.to_screen(f'{exe}: Adding thumbnail to "{filename}"') @PostProcessor._restrict_to(images=False) def run(self, info): filename = info['filepath'] temp_filename = prepend_extension(filename, 'temp') if not info.get('thumbnails'): self.to_screen('There aren\'t any thumbnails to embed') return [], info idx = next((-i for i, t in enumerate(info['thumbnails'][::-1], 1) if t.get('filepath')), None) if idx is None: self.to_screen('There are no thumbnails on disk') return [], info thumbnail_filename = info['thumbnails'][idx]['filepath'] if not os.path.exists(thumbnail_filename): self.report_warning('Skipping embedding the thumbnail because the file is missing.') return [], info # Correct extension for WebP file with wrong extension (see #25687, #25717) convertor = FFmpegThumbnailsConvertorPP(self._downloader) convertor.fixup_webp(info, idx) original_thumbnail = thumbnail_filename = info['thumbnails'][idx]['filepath'] # Convert unsupported thumbnail formats (see #25687, #25717) # PNG is preferred since JPEG is lossy thumbnail_ext = os.path.splitext(thumbnail_filename)[1][1:] if info['ext'] not in ('mkv', 'mka') and thumbnail_ext not in ('jpg', 'jpeg', 'png'): thumbnail_filename = convertor.convert_thumbnail(thumbnail_filename, 'png') thumbnail_ext = 'png' mtime = os.stat(filename).st_mtime success = True if info['ext'] == 'mp3': options = [ '-c', 'copy', '-map', '0:0', '-map', '1:0', '-write_id3v1', '1', '-id3v2_version', '3', '-metadata:s:v', 'title=Album cover', '-metadata:s:v', 'comment=Cover (front)'] self._report_run('ffmpeg', filename) self.run_ffmpeg_multiple_files([filename, thumbnail_filename], temp_filename, options) elif info['ext'] in ['mkv', 'mka']: options = list(self.stream_copy_opts()) mimetype = f'image/{thumbnail_ext.replace("jpg", "jpeg")}' old_stream, new_stream = self.get_stream_number( filename, ('tags', 'mimetype'), mimetype) if old_stream is not None: options.extend(['-map', f'-0:{old_stream}']) new_stream -= 1 options.extend([ '-attach', self._ffmpeg_filename_argument(thumbnail_filename), f'-metadata:s:{new_stream}', f'mimetype={mimetype}', f'-metadata:s:{new_stream}', f'filename=cover.{thumbnail_ext}']) self._report_run('ffmpeg', filename) self.run_ffmpeg(filename, temp_filename, options) elif info['ext'] in ['m4a', 'mp4', 'm4v', 'mov']: prefer_atomicparsley = 'embed-thumbnail-atomicparsley' in self.get_param('compat_opts', []) # Method 1: Use mutagen if not mutagen or prefer_atomicparsley: success = False else: self._report_run('mutagen', filename) f = {'jpeg': MP4Cover.FORMAT_JPEG, 'png': MP4Cover.FORMAT_PNG} try: with open(thumbnail_filename, 'rb') as thumbfile: thumb_data = thumbfile.read() type_ = imghdr.what(h=thumb_data) if not type_: raise ValueError('could not determine image type') elif type_ not in f: raise ValueError(f'incompatible image type: {type_}') meta = MP4(filename) # NOTE: the 'covr' atom is a non-standard MPEG-4 atom, # Apple iTunes 'M4A' files include the 'moov.udta.meta.ilst' atom. meta.tags['covr'] = [MP4Cover(data=thumb_data, imageformat=f[type_])] meta.save() temp_filename = filename except Exception as err: self.report_warning(f'unable to embed using mutagen; {err}') success = False # Method 2: Use AtomicParsley if not success: success = True atomicparsley = next(( # libatomicparsley.so : See https://github.com/xibr/ytdlp-lazy/issues/1 x for x in ['AtomicParsley', 'atomicparsley', 'libatomicparsley.so'] if check_executable(x, ['-v'])), None) if atomicparsley is None: self.to_screen('Neither mutagen nor AtomicParsley was found. Falling back to ffmpeg') success = False else: if not prefer_atomicparsley: self.to_screen('mutagen was not found. Falling back to AtomicParsley') cmd = [atomicparsley, filename, encodeArgument('--artwork'), thumbnail_filename, encodeArgument('-o'), temp_filename] cmd += [encodeArgument(o) for o in self._configuration_args('AtomicParsley')] self._report_run('atomicparsley', filename) self.write_debug(f'AtomicParsley command line: {shell_quote(cmd)}') stdout, stderr, returncode = Popen.run(cmd, text=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE) if returncode: self.report_warning(f'Unable to embed thumbnails using AtomicParsley; {stderr.strip()}') success = False # for formats that don't support thumbnails (like 3gp) AtomicParsley # won't create to the temporary file elif 'No changes' in stdout: self.report_warning('The file format doesn\'t support embedding a thumbnail') success = False # Method 3: Use ffmpeg+ffprobe # Thumbnails attached using this method doesn't show up as cover in some cases # See https://github.com/yt-dlp/yt-dlp/issues/2125, https://github.com/yt-dlp/yt-dlp/issues/411 if not success: success = True try: options = [*self.stream_copy_opts(), '-map', '1'] old_stream, new_stream = self.get_stream_number( filename, ('disposition', 'attached_pic'), 1) if old_stream is not None: options.extend(['-map', f'-0:{old_stream}']) new_stream -= 1 options.extend([f'-disposition:{new_stream}', 'attached_pic']) self._report_run('ffmpeg', filename) self.run_ffmpeg_multiple_files([filename, thumbnail_filename], temp_filename, options) except PostProcessingError as err: success = False raise EmbedThumbnailPPError(f'Unable to embed using ffprobe & ffmpeg; {err}') elif info['ext'] in ['ogg', 'opus', 'flac']: if not mutagen: raise EmbedThumbnailPPError('module mutagen was not found. Please install using `python3 -m pip install mutagen`') self._report_run('mutagen', filename) f = {'opus': OggOpus, 'flac': FLAC, 'ogg': OggVorbis}[info['ext']](filename) pic = Picture() pic.mime = f'image/{imghdr.what(thumbnail_filename)}' with open(thumbnail_filename, 'rb') as thumbfile: pic.data = thumbfile.read() pic.type = 3 # front cover res = self._get_thumbnail_resolution(thumbnail_filename, info['thumbnails'][idx]) if res is not None: pic.width, pic.height = res if info['ext'] == 'flac': f.add_picture(pic) else: # https://wiki.xiph.org/VorbisComment#METADATA_BLOCK_PICTURE f['METADATA_BLOCK_PICTURE'] = base64.b64encode(pic.write()).decode('ascii') f.save() temp_filename = filename else: raise EmbedThumbnailPPError('Supported filetypes for thumbnail embedding are: mp3, mkv/mka, ogg/opus/flac, m4a/mp4/m4v/mov') if success and temp_filename != filename: os.replace(temp_filename, filename) self.try_utime(filename, mtime, mtime) converted = original_thumbnail != thumbnail_filename self._delete_downloaded_files( thumbnail_filename if converted or not self._already_have_thumbnail else None, original_thumbnail if converted and not self._already_have_thumbnail else None, info=info) return [], info
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/postprocessor/exec.py
yt_dlp/postprocessor/exec.py
from .common import PostProcessor from ..utils import Popen, PostProcessingError, shell_quote, variadic class ExecPP(PostProcessor): def __init__(self, downloader, exec_cmd): PostProcessor.__init__(self, downloader) self.exec_cmd = variadic(exec_cmd) def parse_cmd(self, cmd, info): tmpl, tmpl_dict = self._downloader.prepare_outtmpl(cmd, info) if tmpl_dict: # if there are no replacements, tmpl_dict = {} return self._downloader.escape_outtmpl(tmpl) % tmpl_dict filepath = info.get('filepath', info.get('_filename')) # If video, and no replacements are found, replace {} for backard compatibility if filepath: if '{}' not in cmd: cmd += ' {}' cmd = cmd.replace('{}', shell_quote(filepath, shell=True)) return cmd def run(self, info): for tmpl in self.exec_cmd: cmd = self.parse_cmd(tmpl, info) self.to_screen(f'Executing command: {cmd}') _, _, return_code = Popen.run(cmd, shell=True) if return_code != 0: raise PostProcessingError(f'Command returned error code {return_code}') return [], info # Deprecated class ExecAfterDownloadPP(ExecPP): def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) self.deprecation_warning( 'yt_dlp.postprocessor.ExecAfterDownloadPP is deprecated ' 'and may be removed in a future version. Use yt_dlp.postprocessor.ExecPP instead')
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/downloader/niconico.py
yt_dlp/downloader/niconico.py
import json import threading import time from .common import FileDownloader from .external import FFmpegFD from ..networking import Request from ..networking.websocket import WebSocketResponse from ..utils import DownloadError, str_or_none, truncate_string from ..utils.traversal import traverse_obj class NiconicoLiveFD(FileDownloader): """ Downloads niconico live without being stopped """ def real_download(self, filename, info_dict): video_id = info_dict['id'] opts = info_dict['downloader_options'] quality, ws_extractor, ws_url = opts['max_quality'], opts['ws'], opts['ws_url'] dl = FFmpegFD(self.ydl, self.params or {}) new_info_dict = info_dict.copy() new_info_dict['protocol'] = 'm3u8' def communicate_ws(reconnect): # Support --load-info-json as if it is a reconnect attempt if reconnect or not isinstance(ws_extractor, WebSocketResponse): ws = self.ydl.urlopen(Request( ws_url, headers={'Origin': 'https://live.nicovideo.jp'})) if self.ydl.params.get('verbose', False): self.write_debug('Sending startWatching request') ws.send(json.dumps({ 'data': { 'reconnect': True, 'room': { 'commentable': True, 'protocol': 'webSocket', }, 'stream': { 'accessRightMethod': 'single_cookie', 'chasePlay': False, 'latency': 'high', 'protocol': 'hls', 'quality': quality, }, }, 'type': 'startWatching', })) else: ws = ws_extractor with ws: while True: recv = ws.recv() if not recv: continue data = json.loads(recv) if not data or not isinstance(data, dict): continue if data.get('type') == 'ping': ws.send(r'{"type":"pong"}') ws.send(r'{"type":"keepSeat"}') elif data.get('type') == 'disconnect': self.write_debug(data) return True elif data.get('type') == 'error': self.write_debug(data) message = traverse_obj(data, ('body', 'code', {str_or_none}), default=recv) return DownloadError(message) elif self.ydl.params.get('verbose', False): self.write_debug(f'Server response: {truncate_string(recv, 100)}') def ws_main(): reconnect = False while True: try: ret = communicate_ws(reconnect) if ret is True: return except BaseException as e: self.to_screen( f'[niconico:live] {video_id}: Connection error occured, reconnecting after 10 seconds: {e}') time.sleep(10) continue finally: reconnect = True thread = threading.Thread(target=ws_main, daemon=True) thread.start() return dl.download(filename, new_info_dict)
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/downloader/fragment.py
yt_dlp/downloader/fragment.py
import concurrent.futures import contextlib import json import math import os import struct import time from .common import FileDownloader from .http import HttpFD from ..aes import aes_cbc_decrypt_bytes, unpad_pkcs7 from ..networking import Request from ..networking.exceptions import HTTPError, IncompleteRead from ..utils import DownloadError, RetryManager, traverse_obj from ..utils.networking import HTTPHeaderDict from ..utils.progress import ProgressCalculator class HttpQuietDownloader(HttpFD): def to_screen(self, *args, **kargs): pass to_console_title = to_screen class FragmentFD(FileDownloader): """ A base file downloader class for fragmented media (e.g. f4m/m3u8 manifests). Available options: fragment_retries: Number of times to retry a fragment for HTTP error (DASH and hlsnative only). Default is 0 for API, but 10 for CLI skip_unavailable_fragments: Skip unavailable fragments (DASH and hlsnative only) keep_fragments: Keep downloaded fragments on disk after downloading is finished concurrent_fragment_downloads: The number of threads to use for native hls and dash downloads _no_ytdl_file: Don't use .ytdl file For each incomplete fragment download yt-dlp keeps on disk a special bookkeeping file with download state and metadata (in future such files will be used for any incomplete download handled by yt-dlp). This file is used to properly handle resuming, check download file consistency and detect potential errors. The file has a .ytdl extension and represents a standard JSON file of the following format: extractor: Dictionary of extractor related data. TBD. downloader: Dictionary of downloader related data. May contain following data: current_fragment: Dictionary with current (being downloaded) fragment data: index: 0-based index of current fragment among all fragments fragment_count: Total count of fragments This feature is experimental and file format may change in future. """ def report_retry_fragment(self, err, frag_index, count, retries): self.deprecation_warning('yt_dlp.downloader.FragmentFD.report_retry_fragment is deprecated. ' 'Use yt_dlp.downloader.FileDownloader.report_retry instead') return self.report_retry(err, count, retries, frag_index) def report_skip_fragment(self, frag_index, err=None): err = f' {err};' if err else '' self.to_screen(f'[download]{err} Skipping fragment {frag_index:d} ...') def _prepare_url(self, info_dict, url): headers = info_dict.get('http_headers') return Request(url, None, headers) if headers else url def _prepare_and_start_frag_download(self, ctx, info_dict): self._prepare_frag_download(ctx) self._start_frag_download(ctx, info_dict) def __do_ytdl_file(self, ctx): return ctx['live'] is not True and ctx['tmpfilename'] != '-' and not self.params.get('_no_ytdl_file') def _read_ytdl_file(self, ctx): assert 'ytdl_corrupt' not in ctx stream, _ = self.sanitize_open(self.ytdl_filename(ctx['filename']), 'r') try: ytdl_data = json.loads(stream.read()) ctx['fragment_index'] = ytdl_data['downloader']['current_fragment']['index'] if 'extra_state' in ytdl_data['downloader']: ctx['extra_state'] = ytdl_data['downloader']['extra_state'] except Exception: ctx['ytdl_corrupt'] = True finally: stream.close() def _write_ytdl_file(self, ctx): frag_index_stream, _ = self.sanitize_open(self.ytdl_filename(ctx['filename']), 'w') try: downloader = { 'current_fragment': { 'index': ctx['fragment_index'], }, } if 'extra_state' in ctx: downloader['extra_state'] = ctx['extra_state'] if ctx.get('fragment_count') is not None: downloader['fragment_count'] = ctx['fragment_count'] frag_index_stream.write(json.dumps({'downloader': downloader})) finally: frag_index_stream.close() def _download_fragment(self, ctx, frag_url, info_dict, headers=None, request_data=None): fragment_filename = '%s-Frag%d' % (ctx['tmpfilename'], ctx['fragment_index']) fragment_info_dict = { 'url': frag_url, 'http_headers': headers or info_dict.get('http_headers'), 'request_data': request_data, 'ctx_id': ctx.get('ctx_id'), } frag_resume_len = 0 if ctx['dl'].params.get('continuedl', True): frag_resume_len = self.filesize_or_none(self.temp_name(fragment_filename)) fragment_info_dict['frag_resume_len'] = ctx['frag_resume_len'] = frag_resume_len success, _ = ctx['dl'].download(fragment_filename, fragment_info_dict) if not success: return False if fragment_info_dict.get('filetime'): ctx['fragment_filetime'] = fragment_info_dict.get('filetime') ctx['fragment_filename_sanitized'] = fragment_filename return True def _read_fragment(self, ctx): if not ctx.get('fragment_filename_sanitized'): return None try: down, frag_sanitized = self.sanitize_open(ctx['fragment_filename_sanitized'], 'rb') except FileNotFoundError: if ctx.get('live'): return None raise ctx['fragment_filename_sanitized'] = frag_sanitized frag_content = down.read() down.close() return frag_content def _append_fragment(self, ctx, frag_content): try: ctx['dest_stream'].write(frag_content) ctx['dest_stream'].flush() finally: if self.__do_ytdl_file(ctx): self._write_ytdl_file(ctx) if not self.params.get('keep_fragments', False): self.try_remove(ctx['fragment_filename_sanitized']) del ctx['fragment_filename_sanitized'] def _prepare_frag_download(self, ctx): if not ctx.setdefault('live', False): total_frags_str = '%d' % ctx['total_frags'] ad_frags = ctx.get('ad_frags', 0) if ad_frags: total_frags_str += ' (not including %d ad)' % ad_frags else: total_frags_str = 'unknown (live)' self.to_screen(f'[{self.FD_NAME}] Total fragments: {total_frags_str}') self.report_destination(ctx['filename']) dl = HttpQuietDownloader(self.ydl, { **self.params, 'noprogress': True, 'test': False, 'sleep_interval': 0, 'max_sleep_interval': 0, 'sleep_interval_subtitles': 0, }) tmpfilename = self.temp_name(ctx['filename']) open_mode = 'wb' # Establish possible resume length resume_len = self.filesize_or_none(tmpfilename) if resume_len > 0: open_mode = 'ab' # Should be initialized before ytdl file check ctx.update({ 'tmpfilename': tmpfilename, 'fragment_index': 0, }) if self.__do_ytdl_file(ctx): ytdl_file_exists = os.path.isfile(self.ytdl_filename(ctx['filename'])) continuedl = self.params.get('continuedl', True) if continuedl and ytdl_file_exists: self._read_ytdl_file(ctx) is_corrupt = ctx.get('ytdl_corrupt') is True is_inconsistent = ctx['fragment_index'] > 0 and resume_len == 0 if is_corrupt or is_inconsistent: message = ( '.ytdl file is corrupt' if is_corrupt else 'Inconsistent state of incomplete fragment download') self.report_warning( f'{message}. Restarting from the beginning ...') ctx['fragment_index'] = resume_len = 0 if 'ytdl_corrupt' in ctx: del ctx['ytdl_corrupt'] self._write_ytdl_file(ctx) else: if not continuedl: if ytdl_file_exists: self._read_ytdl_file(ctx) ctx['fragment_index'] = resume_len = 0 self._write_ytdl_file(ctx) assert ctx['fragment_index'] == 0 dest_stream, tmpfilename = self.sanitize_open(tmpfilename, open_mode) ctx.update({ 'dl': dl, 'dest_stream': dest_stream, 'tmpfilename': tmpfilename, # Total complete fragments downloaded so far in bytes 'complete_frags_downloaded_bytes': resume_len, }) def _start_frag_download(self, ctx, info_dict): resume_len = ctx['complete_frags_downloaded_bytes'] total_frags = ctx['total_frags'] ctx_id = ctx.get('ctx_id') # Stores the download progress, updated by the progress hook state = { 'status': 'downloading', 'downloaded_bytes': resume_len, 'fragment_index': ctx['fragment_index'], 'fragment_count': total_frags, 'filename': ctx['filename'], 'tmpfilename': ctx['tmpfilename'], } ctx['started'] = time.time() progress = ProgressCalculator(resume_len) def frag_progress_hook(s): if s['status'] not in ('downloading', 'finished'): return if not total_frags and ctx.get('fragment_count'): state['fragment_count'] = ctx['fragment_count'] if ctx_id is not None and s.get('ctx_id') != ctx_id: return state['max_progress'] = ctx.get('max_progress') state['progress_idx'] = ctx.get('progress_idx') state['elapsed'] = progress.elapsed frag_total_bytes = s.get('total_bytes') or 0 s['fragment_info_dict'] = s.pop('info_dict', {}) # XXX: Fragment resume is not accounted for here if not ctx['live']: estimated_size = ( (ctx['complete_frags_downloaded_bytes'] + frag_total_bytes) / (state['fragment_index'] + 1) * total_frags) progress.total = estimated_size progress.update(s.get('downloaded_bytes')) state['total_bytes_estimate'] = progress.total else: progress.update(s.get('downloaded_bytes')) if s['status'] == 'finished': state['fragment_index'] += 1 ctx['fragment_index'] = state['fragment_index'] progress.thread_reset() state['downloaded_bytes'] = ctx['complete_frags_downloaded_bytes'] = progress.downloaded state['speed'] = ctx['speed'] = progress.speed.smooth state['eta'] = progress.eta.smooth self._hook_progress(state, info_dict) ctx['dl'].add_progress_hook(frag_progress_hook) return ctx['started'] def _finish_frag_download(self, ctx, info_dict): ctx['dest_stream'].close() if self.__do_ytdl_file(ctx): self.try_remove(self.ytdl_filename(ctx['filename'])) elapsed = time.time() - ctx['started'] to_file = ctx['tmpfilename'] != '-' if to_file: downloaded_bytes = self.filesize_or_none(ctx['tmpfilename']) else: downloaded_bytes = ctx['complete_frags_downloaded_bytes'] if not downloaded_bytes: if to_file: self.try_remove(ctx['tmpfilename']) self.report_error('The downloaded file is empty') return False elif to_file: self.try_rename(ctx['tmpfilename'], ctx['filename']) filetime = ctx.get('fragment_filetime') if self.params.get('updatetime') and filetime: with contextlib.suppress(Exception): os.utime(ctx['filename'], (time.time(), filetime)) self._hook_progress({ 'downloaded_bytes': downloaded_bytes, 'total_bytes': downloaded_bytes, 'filename': ctx['filename'], 'status': 'finished', 'elapsed': elapsed, 'ctx_id': ctx.get('ctx_id'), 'max_progress': ctx.get('max_progress'), 'progress_idx': ctx.get('progress_idx'), }, info_dict) return True def _prepare_external_frag_download(self, ctx): if 'live' not in ctx: ctx['live'] = False if not ctx['live']: total_frags_str = '%d' % ctx['total_frags'] ad_frags = ctx.get('ad_frags', 0) if ad_frags: total_frags_str += ' (not including %d ad)' % ad_frags else: total_frags_str = 'unknown (live)' self.to_screen(f'[{self.FD_NAME}] Total fragments: {total_frags_str}') tmpfilename = self.temp_name(ctx['filename']) # Should be initialized before ytdl file check ctx.update({ 'tmpfilename': tmpfilename, 'fragment_index': 0, }) def decrypter(self, info_dict): _key_cache = {} def _get_key(url): if url not in _key_cache: _key_cache[url] = self.ydl.urlopen(self._prepare_url(info_dict, url)).read() return _key_cache[url] def decrypt_fragment(fragment, frag_content): if frag_content is None: return decrypt_info = fragment.get('decrypt_info') if not decrypt_info or decrypt_info['METHOD'] != 'AES-128': return frag_content iv = decrypt_info.get('IV') or struct.pack('>8xq', fragment['media_sequence']) decrypt_info['KEY'] = (decrypt_info.get('KEY') or _get_key(traverse_obj(info_dict, ('hls_aes', 'uri')) or decrypt_info['URI'])) # Don't decrypt the content in tests since the data is explicitly truncated and it's not to a valid block # size (see https://github.com/ytdl-org/youtube-dl/pull/27660). Tests only care that the correct data downloaded, # not what it decrypts to. if self.params.get('test', False): return frag_content return unpad_pkcs7(aes_cbc_decrypt_bytes(frag_content, decrypt_info['KEY'], iv)) return decrypt_fragment def download_and_append_fragments_multiple(self, *args, **kwargs): """ @params (ctx1, fragments1, info_dict1), (ctx2, fragments2, info_dict2), ... all args must be either tuple or list """ interrupt_trigger = [True] max_progress = len(args) if max_progress == 1: return self.download_and_append_fragments(*args[0], **kwargs) max_workers = self.params.get('concurrent_fragment_downloads', 1) if max_progress > 1: self._prepare_multiline_status(max_progress) is_live = any(traverse_obj(args, (..., 2, 'is_live'))) def thread_func(idx, ctx, fragments, info_dict, tpe): ctx['max_progress'] = max_progress ctx['progress_idx'] = idx return self.download_and_append_fragments( ctx, fragments, info_dict, **kwargs, tpe=tpe, interrupt_trigger=interrupt_trigger) class FTPE(concurrent.futures.ThreadPoolExecutor): # has to stop this or it's going to wait on the worker thread itself def __exit__(self, exc_type, exc_val, exc_tb): pass if os.name == 'nt': def future_result(future): while True: try: return future.result(0.1) except KeyboardInterrupt: raise except concurrent.futures.TimeoutError: continue else: def future_result(future): return future.result() def interrupt_trigger_iter(fg): for f in fg: if not interrupt_trigger[0]: break yield f spins = [] for idx, (ctx, fragments, info_dict) in enumerate(args): tpe = FTPE(math.ceil(max_workers / max_progress)) job = tpe.submit(thread_func, idx, ctx, interrupt_trigger_iter(fragments), info_dict, tpe) spins.append((tpe, job)) result = True for tpe, job in spins: try: result = result and future_result(job) except KeyboardInterrupt: interrupt_trigger[0] = False finally: tpe.shutdown(wait=True) if not interrupt_trigger[0] and not is_live: raise KeyboardInterrupt # we expect the user wants to stop and DO WANT the preceding postprocessors to run; # so returning a intermediate result here instead of KeyboardInterrupt on live return result def download_and_append_fragments( self, ctx, fragments, info_dict, *, is_fatal=(lambda idx: False), pack_func=(lambda content, idx: content), finish_func=None, tpe=None, interrupt_trigger=(True, )): if not self.params.get('skip_unavailable_fragments', True): is_fatal = lambda _: True def download_fragment(fragment, ctx): if not interrupt_trigger[0]: return frag_index = ctx['fragment_index'] = fragment['frag_index'] ctx['last_error'] = None headers = HTTPHeaderDict(info_dict.get('http_headers')) byte_range = fragment.get('byte_range') if byte_range: headers['Range'] = 'bytes=%d-%d' % (byte_range['start'], byte_range['end'] - 1) # Never skip the first fragment fatal = is_fatal(fragment.get('index') or (frag_index - 1)) def error_callback(err, count, retries): if fatal and count > retries: ctx['dest_stream'].close() self.report_retry(err, count, retries, frag_index, fatal) ctx['last_error'] = err for retry in RetryManager(self.params.get('fragment_retries'), error_callback): try: ctx['fragment_count'] = fragment.get('fragment_count') if not self._download_fragment( ctx, fragment['url'], info_dict, headers, info_dict.get('request_data')): return except (HTTPError, IncompleteRead) as err: retry.error = err continue except DownloadError: # has own retry settings if fatal: raise def append_fragment(frag_content, frag_index, ctx): if frag_content: self._append_fragment(ctx, pack_func(frag_content, frag_index)) elif not is_fatal(frag_index - 1): self.report_skip_fragment(frag_index, 'fragment not found') else: ctx['dest_stream'].close() self.report_error(f'fragment {frag_index} not found, unable to continue') return False return True decrypt_fragment = self.decrypter(info_dict) max_workers = math.ceil( self.params.get('concurrent_fragment_downloads', 1) / ctx.get('max_progress', 1)) if max_workers > 1: def _download_fragment(fragment): ctx_copy = ctx.copy() download_fragment(fragment, ctx_copy) return fragment, fragment['frag_index'], ctx_copy.get('fragment_filename_sanitized') with tpe or concurrent.futures.ThreadPoolExecutor(max_workers) as pool: try: for fragment, frag_index, frag_filename in pool.map(_download_fragment, fragments): ctx.update({ 'fragment_filename_sanitized': frag_filename, 'fragment_index': frag_index, }) if not append_fragment(decrypt_fragment(fragment, self._read_fragment(ctx)), frag_index, ctx): return False except KeyboardInterrupt: self._finish_multiline_status() self.report_error( 'Interrupted by user. Waiting for all threads to shutdown...', is_error=False, tb=False) pool.shutdown(wait=False) raise else: for fragment in fragments: if not interrupt_trigger[0]: break try: download_fragment(fragment, ctx) result = append_fragment( decrypt_fragment(fragment, self._read_fragment(ctx)), fragment['frag_index'], ctx) except KeyboardInterrupt: if info_dict.get('is_live'): break raise if not result: return False if finish_func is not None: ctx['dest_stream'].write(finish_func()) ctx['dest_stream'].flush() return self._finish_frag_download(ctx, info_dict)
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/downloader/hls.py
yt_dlp/downloader/hls.py
import binascii import io import re import urllib.parse from . import get_suitable_downloader from .external import FFmpegFD from .fragment import FragmentFD from .. import webvtt from ..dependencies import Cryptodome from ..utils import ( bug_reports_message, parse_m3u8_attributes, remove_start, traverse_obj, update_url_query, urljoin, ) from ..utils._utils import _request_dump_filename class HlsFD(FragmentFD): """ Download segments in a m3u8 manifest. External downloaders can take over the fragment downloads by supporting the 'm3u8_frag_urls' protocol and re-defining 'supports_manifest' function """ FD_NAME = 'hlsnative' @staticmethod def _has_drm(manifest): # TODO: https://github.com/yt-dlp/yt-dlp/pull/5039 return bool(re.search('|'.join(( r'#EXT-X-(?:SESSION-)?KEY:.*?URI="skd://', # Apple FairPlay r'#EXT-X-(?:SESSION-)?KEY:.*?KEYFORMAT="com\.apple\.streamingkeydelivery"', # Apple FairPlay r'#EXT-X-(?:SESSION-)?KEY:.*?KEYFORMAT="com\.microsoft\.playready"', # Microsoft PlayReady r'#EXT-X-FAXS-CM:', # Adobe Flash Access )), manifest)) @classmethod def can_download(cls, manifest, info_dict, allow_unplayable_formats=False): UNSUPPORTED_FEATURES = [ # r'#EXT-X-BYTERANGE', # playlists composed of byte ranges of media files [2] # Live streams heuristic does not always work (e.g. geo restricted to Germany # http://hls-geo.daserste.de/i/videoportal/Film/c_620000/622873/format,716451,716457,716450,716458,716459,.mp4.csmil/index_4_av.m3u8?null=0) # r'#EXT-X-MEDIA-SEQUENCE:(?!0$)', # live streams [3] # This heuristic also is not correct since segments may not be appended as well. # Twitch vods of finished streams have EXT-X-PLAYLIST-TYPE:EVENT despite # no segments will definitely be appended to the end of the playlist. # r'#EXT-X-PLAYLIST-TYPE:EVENT', # media segments may be appended to the end of # # event media playlists [4] # r'#EXT-X-MAP:', # media initialization [5] # 1. https://tools.ietf.org/html/draft-pantos-http-live-streaming-17#section-4.3.2.4 # 2. https://tools.ietf.org/html/draft-pantos-http-live-streaming-17#section-4.3.2.2 # 3. https://tools.ietf.org/html/draft-pantos-http-live-streaming-17#section-4.3.3.2 # 4. https://tools.ietf.org/html/draft-pantos-http-live-streaming-17#section-4.3.3.5 # 5. https://tools.ietf.org/html/draft-pantos-http-live-streaming-17#section-4.3.2.5 ] if not allow_unplayable_formats: UNSUPPORTED_FEATURES += [ r'#EXT-X-KEY:METHOD=(?!NONE|AES-128)', # encrypted streams [1], but not necessarily DRM ] def check_results(): yield not info_dict.get('is_live') for feature in UNSUPPORTED_FEATURES: yield not re.search(feature, manifest) if not allow_unplayable_formats: yield not cls._has_drm(manifest) return all(check_results()) def real_download(self, filename, info_dict): man_url = info_dict['url'] s = info_dict.get('hls_media_playlist_data') if s: self.to_screen(f'[{self.FD_NAME}] Using m3u8 manifest from extracted info') else: self.to_screen(f'[{self.FD_NAME}] Downloading m3u8 manifest') urlh = self.ydl.urlopen(self._prepare_url(info_dict, man_url)) man_url = urlh.url s_bytes = urlh.read() if self.params.get('write_pages'): dump_filename = _request_dump_filename( man_url, info_dict['id'], None, trim_length=self.params.get('trim_file_name')) self.to_screen(f'[{self.FD_NAME}] Saving request to {dump_filename}') with open(dump_filename, 'wb') as outf: outf.write(s_bytes) s = s_bytes.decode('utf-8', 'ignore') can_download, message = self.can_download(s, info_dict, self.params.get('allow_unplayable_formats')), None if can_download: has_ffmpeg = FFmpegFD.available() if not Cryptodome.AES and '#EXT-X-KEY:METHOD=AES-128' in s: # Even if pycryptodomex isn't available, force HlsFD for m3u8s that won't work with ffmpeg ffmpeg_can_dl = not traverse_obj(info_dict, (( 'extra_param_to_segment_url', 'extra_param_to_key_url', 'hls_media_playlist_data', ('hls_aes', ('uri', 'key', 'iv')), ), any)) message = 'The stream has AES-128 encryption and {} available'.format( 'neither ffmpeg nor pycryptodomex are' if ffmpeg_can_dl and not has_ffmpeg else 'pycryptodomex is not') if has_ffmpeg and ffmpeg_can_dl: can_download = False else: message += '; decryption will be performed natively, but will be extremely slow' elif info_dict.get('extractor_key') == 'Generic' and re.search(r'(?m)#EXT-X-MEDIA-SEQUENCE:(?!0$)', s): install_ffmpeg = '' if has_ffmpeg else 'install ffmpeg and ' message = ('Live HLS streams are not supported by the native downloader. If this is a livestream, ' f'please {install_ffmpeg}add "--downloader ffmpeg --hls-use-mpegts" to your command') if not can_download: if self._has_drm(s) and not self.params.get('allow_unplayable_formats'): if info_dict.get('has_drm') and self.params.get('test'): self.to_screen(f'[{self.FD_NAME}] This format is DRM protected', skip_eol=True) else: self.report_error( 'This format is DRM protected; Try selecting another format with --format or ' 'add --check-formats to automatically fallback to the next best format', tb=False) return False message = message or 'Unsupported features have been detected' fd = FFmpegFD(self.ydl, self.params) self.report_warning(f'{message}; extraction will be delegated to {fd.get_basename()}') return fd.real_download(filename, info_dict) elif message: self.report_warning(message) is_webvtt = info_dict['ext'] == 'vtt' if is_webvtt: real_downloader = None # Packing the fragments is not currently supported for external downloader else: real_downloader = get_suitable_downloader( info_dict, self.params, None, protocol='m3u8_frag_urls', to_stdout=(filename == '-')) if real_downloader and not real_downloader.supports_manifest(s): real_downloader = None if real_downloader: self.to_screen(f'[{self.FD_NAME}] Fragment downloads will be delegated to {real_downloader.get_basename()}') def is_ad_fragment_start(s): return ((s.startswith('#ANVATO-SEGMENT-INFO') and 'type=ad' in s) or (s.startswith('#UPLYNK-SEGMENT') and s.endswith(',ad'))) def is_ad_fragment_end(s): return ((s.startswith('#ANVATO-SEGMENT-INFO') and 'type=master' in s) or (s.startswith('#UPLYNK-SEGMENT') and s.endswith(',segment'))) fragments = [] media_frags = 0 ad_frags = 0 ad_frag_next = False for line in s.splitlines(): line = line.strip() if not line: continue if line.startswith('#'): if is_ad_fragment_start(line): ad_frag_next = True elif is_ad_fragment_end(line): ad_frag_next = False continue if ad_frag_next: ad_frags += 1 continue media_frags += 1 ctx = { 'filename': filename, 'total_frags': media_frags, 'ad_frags': ad_frags, } if real_downloader: self._prepare_external_frag_download(ctx) else: self._prepare_and_start_frag_download(ctx, info_dict) extra_state = ctx.setdefault('extra_state', {}) format_index = info_dict.get('format_index') extra_segment_query = None if extra_param_to_segment_url := info_dict.get('extra_param_to_segment_url'): extra_segment_query = urllib.parse.parse_qs(extra_param_to_segment_url) extra_key_query = None if extra_param_to_key_url := info_dict.get('extra_param_to_key_url'): extra_key_query = urllib.parse.parse_qs(extra_param_to_key_url) i = 0 media_sequence = 0 decrypt_info = {'METHOD': 'NONE'} external_aes_key = traverse_obj(info_dict, ('hls_aes', 'key')) if external_aes_key: external_aes_key = binascii.unhexlify(remove_start(external_aes_key, '0x')) assert len(external_aes_key) in (16, 24, 32), 'Invalid length for HLS AES-128 key' external_aes_iv = traverse_obj(info_dict, ('hls_aes', 'iv')) if external_aes_iv: external_aes_iv = binascii.unhexlify(remove_start(external_aes_iv, '0x').zfill(32)) byte_range = {} byte_range_offset = 0 discontinuity_count = 0 frag_index = 0 ad_frag_next = False for line in s.splitlines(): line = line.strip() if line: if not line.startswith('#'): if format_index is not None and discontinuity_count != format_index: continue if ad_frag_next: continue frag_index += 1 if frag_index <= ctx['fragment_index']: continue frag_url = urljoin(man_url, line) if extra_segment_query: frag_url = update_url_query(frag_url, extra_segment_query) fragments.append({ 'frag_index': frag_index, 'url': frag_url, 'decrypt_info': decrypt_info, 'byte_range': byte_range, 'media_sequence': media_sequence, }) media_sequence += 1 # If the byte_range is truthy, reset it after appending a fragment that uses it if byte_range: byte_range_offset = byte_range['end'] byte_range = {} elif line.startswith('#EXT-X-MAP'): if format_index is not None and discontinuity_count != format_index: continue if frag_index > 0: self.report_error( 'Initialization fragment found after media fragments, unable to download') return False frag_index += 1 map_info = parse_m3u8_attributes(line[11:]) frag_url = urljoin(man_url, map_info.get('URI')) if extra_segment_query: frag_url = update_url_query(frag_url, extra_segment_query) map_byte_range = {} if map_info.get('BYTERANGE'): splitted_byte_range = map_info.get('BYTERANGE').split('@') sub_range_start = int(splitted_byte_range[1]) if len(splitted_byte_range) == 2 else 0 map_byte_range = { 'start': sub_range_start, 'end': sub_range_start + int(splitted_byte_range[0]), } fragments.append({ 'frag_index': frag_index, 'url': frag_url, 'decrypt_info': decrypt_info, 'byte_range': map_byte_range, 'media_sequence': media_sequence, }) media_sequence += 1 elif line.startswith('#EXT-X-KEY'): decrypt_url = decrypt_info.get('URI') decrypt_info = parse_m3u8_attributes(line[11:]) if decrypt_info['METHOD'] == 'AES-128': if external_aes_iv: decrypt_info['IV'] = external_aes_iv elif 'IV' in decrypt_info: decrypt_info['IV'] = binascii.unhexlify(decrypt_info['IV'][2:].zfill(32)) if external_aes_key: decrypt_info['KEY'] = external_aes_key else: decrypt_info['URI'] = urljoin(man_url, decrypt_info['URI']) if extra_key_query or extra_segment_query: # Fall back to extra_segment_query to key for backwards compat decrypt_info['URI'] = update_url_query( decrypt_info['URI'], extra_key_query or extra_segment_query) if decrypt_url != decrypt_info['URI']: decrypt_info['KEY'] = None elif line.startswith('#EXT-X-MEDIA-SEQUENCE'): media_sequence = int(line[22:]) elif line.startswith('#EXT-X-BYTERANGE'): splitted_byte_range = line[17:].split('@') sub_range_start = int(splitted_byte_range[1]) if len(splitted_byte_range) == 2 else byte_range_offset byte_range = { 'start': sub_range_start, 'end': sub_range_start + int(splitted_byte_range[0]), } elif is_ad_fragment_start(line): ad_frag_next = True elif is_ad_fragment_end(line): ad_frag_next = False elif line.startswith('#EXT-X-DISCONTINUITY'): discontinuity_count += 1 i += 1 # We only download the first fragment during the test if self.params.get('test', False): fragments = [fragments[0] if fragments else None] if real_downloader: info_dict['fragments'] = fragments fd = real_downloader(self.ydl, self.params) # TODO: Make progress updates work without hooking twice # for ph in self._progress_hooks: # fd.add_progress_hook(ph) return fd.real_download(filename, info_dict) if is_webvtt: def pack_fragment(frag_content, frag_index): output = io.StringIO() adjust = 0 overflow = False mpegts_last = None for block in webvtt.parse_fragment(frag_content): if isinstance(block, webvtt.CueBlock): extra_state['webvtt_mpegts_last'] = mpegts_last if overflow: extra_state['webvtt_mpegts_adjust'] += 1 overflow = False block.start += adjust block.end += adjust dedup_window = extra_state.setdefault('webvtt_dedup_window', []) ready = [] i = 0 is_new = True while i < len(dedup_window): wcue = dedup_window[i] wblock = webvtt.CueBlock.from_json(wcue) i += 1 if wblock.hinges(block): wcue['end'] = block.end is_new = False continue if wblock == block: is_new = False continue if wblock.end > block.start: continue ready.append(wblock) i -= 1 del dedup_window[i] if is_new: dedup_window.append(block.as_json) for block in ready: block.write_into(output) # we only emit cues once they fall out of the duplicate window continue elif isinstance(block, webvtt.Magic): # take care of MPEG PES timestamp overflow if block.mpegts is None: block.mpegts = 0 extra_state.setdefault('webvtt_mpegts_adjust', 0) block.mpegts += extra_state['webvtt_mpegts_adjust'] << 33 if block.mpegts < extra_state.get('webvtt_mpegts_last', 0): overflow = True block.mpegts += 1 << 33 mpegts_last = block.mpegts if frag_index == 1: extra_state['webvtt_mpegts'] = block.mpegts or 0 extra_state['webvtt_local'] = block.local or 0 # XXX: block.local = block.mpegts = None ? else: if block.mpegts is not None and block.local is not None: adjust = ( (block.mpegts - extra_state.get('webvtt_mpegts', 0)) - (block.local - extra_state.get('webvtt_local', 0)) ) continue elif isinstance(block, webvtt.HeaderBlock): if frag_index != 1: # XXX: this should probably be silent as well # or verify that all segments contain the same data self.report_warning(bug_reports_message( f'Discarding a {type(block).__name__} block found in the middle of the stream; ' 'if the subtitles display incorrectly,')) continue block.write_into(output) return output.getvalue().encode() def fin_fragments(): dedup_window = extra_state.get('webvtt_dedup_window') if not dedup_window: return b'' output = io.StringIO() for cue in dedup_window: webvtt.CueBlock.from_json(cue).write_into(output) return output.getvalue().encode() if len(fragments) == 1: self.download_and_append_fragments(ctx, fragments, info_dict) else: self.download_and_append_fragments( ctx, fragments, info_dict, pack_func=pack_fragment, finish_func=fin_fragments) else: return self.download_and_append_fragments(ctx, fragments, info_dict)
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/downloader/http.py
yt_dlp/downloader/http.py
import os import random import time from .common import FileDownloader from ..networking import Request from ..networking.exceptions import ( CertificateVerifyError, HTTPError, TransportError, ) from ..utils import ( ContentTooShortError, RetryManager, ThrottledDownload, int_or_none, parse_http_range, try_call, ) from ..utils.networking import HTTPHeaderDict class HttpFD(FileDownloader): def real_download(self, filename, info_dict): url = info_dict['url'] request_data = info_dict.get('request_data', None) request_extensions = {} impersonate_target = self._get_impersonate_target(info_dict) if impersonate_target is not None: request_extensions['impersonate'] = impersonate_target class DownloadContext(dict): __getattr__ = dict.get __setattr__ = dict.__setitem__ __delattr__ = dict.__delitem__ ctx = DownloadContext() ctx.filename = filename ctx.tmpfilename = self.temp_name(filename) ctx.stream = None # Disable compression headers = HTTPHeaderDict({'Accept-Encoding': 'identity'}, info_dict.get('http_headers')) is_test = self.params.get('test', False) chunk_size = self._TEST_FILE_SIZE if is_test else ( self.params.get('http_chunk_size') or info_dict.get('downloader_options', {}).get('http_chunk_size') or 0) ctx.open_mode = 'wb' ctx.resume_len = 0 ctx.block_size = self.params.get('buffersize', 1024) ctx.start_time = time.time() # parse given Range req_start, req_end, _ = parse_http_range(headers.get('Range')) if self.params.get('continuedl', True): # Establish possible resume length if os.path.isfile(ctx.tmpfilename): ctx.resume_len = os.path.getsize(ctx.tmpfilename) ctx.is_resume = ctx.resume_len > 0 class SucceedDownload(Exception): pass class RetryDownload(Exception): def __init__(self, source_error): self.source_error = source_error class NextFragment(Exception): pass def establish_connection(): ctx.chunk_size = (random.randint(int(chunk_size * 0.95), chunk_size) if not is_test and chunk_size else chunk_size) if ctx.resume_len > 0: range_start = ctx.resume_len if req_start is not None: # offset the beginning of Range to be within request range_start += req_start if ctx.is_resume: self.report_resuming_byte(ctx.resume_len) ctx.open_mode = 'ab' elif req_start is not None: range_start = req_start elif ctx.chunk_size > 0: range_start = 0 else: range_start = None ctx.is_resume = False if ctx.chunk_size: chunk_aware_end = range_start + ctx.chunk_size - 1 # we're not allowed to download outside Range range_end = chunk_aware_end if req_end is None else min(chunk_aware_end, req_end) elif req_end is not None: # there's no need for chunked downloads, so download until the end of Range range_end = req_end else: range_end = None if try_call(lambda: range_start > range_end): ctx.resume_len = 0 ctx.open_mode = 'wb' raise RetryDownload(Exception(f'Conflicting range. (start={range_start} > end={range_end})')) if try_call(lambda: range_end >= ctx.content_len): range_end = ctx.content_len - 1 request = Request(url, request_data, headers, extensions=request_extensions) has_range = range_start is not None if has_range: request.headers['Range'] = f'bytes={int(range_start)}-{int_or_none(range_end) or ""}' # Establish connection try: ctx.data = self.ydl.urlopen(request) # When trying to resume, Content-Range HTTP header of response has to be checked # to match the value of requested Range HTTP header. This is due to a webservers # that don't support resuming and serve a whole file with no Content-Range # set in response despite of requested Range (see # https://github.com/ytdl-org/youtube-dl/issues/6057#issuecomment-126129799) if has_range: content_range = ctx.data.headers.get('Content-Range') content_range_start, content_range_end, content_len = parse_http_range(content_range) # Content-Range is present and matches requested Range, resume is possible if range_start == content_range_start and ( # Non-chunked download not ctx.chunk_size # Chunked download and requested piece or # its part is promised to be served or content_range_end == range_end or content_len < range_end): ctx.content_len = content_len if content_len or req_end: ctx.data_len = min(content_len or req_end, req_end or content_len) - (req_start or 0) return # Content-Range is either not present or invalid. Assuming remote webserver is # trying to send the whole file, resume is not possible, so wiping the local file # and performing entire redownload elif range_start > 0: self.report_unable_to_resume() ctx.resume_len = 0 ctx.open_mode = 'wb' ctx.data_len = ctx.content_len = int_or_none(ctx.data.headers.get('Content-length', None)) except HTTPError as err: if err.status == 416: # Unable to resume (requested range not satisfiable) try: # Open the connection again without the range header ctx.data = self.ydl.urlopen( Request(url, request_data, headers)) content_length = ctx.data.headers['Content-Length'] except HTTPError as err: if err.status < 500 or err.status >= 600: raise else: # Examine the reported length if (content_length is not None and (ctx.resume_len - 100 < int(content_length) < ctx.resume_len + 100)): # The file had already been fully downloaded. # Explanation to the above condition: in issue #175 it was revealed that # YouTube sometimes adds or removes a few bytes from the end of the file, # changing the file size slightly and causing problems for some users. So # I decided to implement a suggested change and consider the file # completely downloaded if the file size differs less than 100 bytes from # the one in the hard drive. self.report_file_already_downloaded(ctx.filename) self.try_rename(ctx.tmpfilename, ctx.filename) self._hook_progress({ 'filename': ctx.filename, 'status': 'finished', 'downloaded_bytes': ctx.resume_len, 'total_bytes': ctx.resume_len, }, info_dict) raise SucceedDownload else: # The length does not match, we start the download over self.report_unable_to_resume() ctx.resume_len = 0 ctx.open_mode = 'wb' return elif err.status < 500 or err.status >= 600: # Unexpected HTTP error raise raise RetryDownload(err) except CertificateVerifyError: raise except TransportError as err: raise RetryDownload(err) def close_stream(): if ctx.stream is not None: if ctx.tmpfilename != '-': ctx.stream.close() ctx.stream = None def download(): data_len = ctx.data.headers.get('Content-length') if ctx.data.headers.get('Content-encoding'): # Content-encoding is present, Content-length is not reliable anymore as we are # doing auto decompression. (See: https://github.com/yt-dlp/yt-dlp/pull/6176) data_len = None # Range HTTP header may be ignored/unsupported by a webserver # (e.g. extractor/scivee.py, extractor/bambuser.py). # However, for a test we still would like to download just a piece of a file. # To achieve this we limit data_len to _TEST_FILE_SIZE and manually control # block size when downloading a file. if is_test and (data_len is None or int(data_len) > self._TEST_FILE_SIZE): data_len = self._TEST_FILE_SIZE if data_len is not None: data_len = int(data_len) + ctx.resume_len min_data_len = self.params.get('min_filesize') max_data_len = self.params.get('max_filesize') if min_data_len is not None and data_len < min_data_len: self.to_screen( f'\r[download] File is smaller than min-filesize ({data_len} bytes < {min_data_len} bytes). Aborting.') return False if max_data_len is not None and data_len > max_data_len: self.to_screen( f'\r[download] File is larger than max-filesize ({data_len} bytes > {max_data_len} bytes). Aborting.') return False byte_counter = 0 + ctx.resume_len block_size = ctx.block_size start = time.time() # measure time over whole while-loop, so slow_down() and best_block_size() work together properly now = None # needed for slow_down() in the first loop run before = start # start measuring def retry(e): close_stream() if ctx.tmpfilename == '-': ctx.resume_len = byte_counter else: try: ctx.resume_len = os.path.getsize(ctx.tmpfilename) except FileNotFoundError: ctx.resume_len = 0 raise RetryDownload(e) while True: try: # Download and write data_block = ctx.data.read(block_size if not is_test else min(block_size, data_len - byte_counter)) except TransportError as err: retry(err) byte_counter += len(data_block) # exit loop when download is finished if len(data_block) == 0: break # Open destination file just in time if ctx.stream is None: try: ctx.stream, ctx.tmpfilename = self.sanitize_open( ctx.tmpfilename, ctx.open_mode) assert ctx.stream is not None ctx.filename = self.undo_temp_name(ctx.tmpfilename) self.report_destination(ctx.filename) except OSError as err: self.report_error(f'unable to open for writing: {err}') return False try: ctx.stream.write(data_block) except OSError as err: self.to_stderr('\n') self.report_error(f'unable to write data: {err}') return False # Apply rate limit self.slow_down(start, now, byte_counter - ctx.resume_len) # end measuring of one loop run now = time.time() after = now # Adjust block size if not self.params.get('noresizebuffer', False): block_size = self.best_block_size(after - before, len(data_block)) before = after # Progress message speed = self.calc_speed(start, now, byte_counter - ctx.resume_len) if ctx.data_len is None: eta = None else: eta = self.calc_eta(start, time.time(), ctx.data_len - ctx.resume_len, byte_counter - ctx.resume_len) self._hook_progress({ 'status': 'downloading', 'downloaded_bytes': byte_counter, 'total_bytes': ctx.data_len, 'tmpfilename': ctx.tmpfilename, 'filename': ctx.filename, 'eta': eta, 'speed': speed, 'elapsed': now - ctx.start_time, 'ctx_id': info_dict.get('ctx_id'), }, info_dict) if data_len is not None and byte_counter == data_len: break if speed and speed < (self.params.get('throttledratelimit') or 0): # The speed must stay below the limit for 3 seconds # This prevents raising error when the speed temporarily goes down if ctx.throttle_start is None: ctx.throttle_start = now elif now - ctx.throttle_start > 3: if ctx.stream is not None and ctx.tmpfilename != '-': ctx.stream.close() raise ThrottledDownload elif speed: ctx.throttle_start = None if ctx.stream is None: self.to_stderr('\n') self.report_error('Did not get any data blocks') return False if not is_test and ctx.chunk_size and ctx.content_len is not None and byte_counter < ctx.content_len: ctx.resume_len = byte_counter raise NextFragment if ctx.tmpfilename != '-': ctx.stream.close() if data_len is not None and byte_counter != data_len: err = ContentTooShortError(byte_counter, int(data_len)) retry(err) self.try_rename(ctx.tmpfilename, ctx.filename) # Update file modification time if self.params.get('updatetime'): info_dict['filetime'] = self.try_utime(ctx.filename, ctx.data.headers.get('last-modified', None)) self._hook_progress({ 'downloaded_bytes': byte_counter, 'total_bytes': byte_counter, 'filename': ctx.filename, 'status': 'finished', 'elapsed': time.time() - ctx.start_time, 'ctx_id': info_dict.get('ctx_id'), }, info_dict) return True for retry in RetryManager(self.params.get('retries'), self.report_retry): try: establish_connection() return download() except RetryDownload as err: retry.error = err.source_error continue except NextFragment: retry.error = None retry.attempt -= 1 continue except SucceedDownload: return True except: # noqa: E722 close_stream() raise return False
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/downloader/common.py
yt_dlp/downloader/common.py
import contextlib import errno import functools import os import random import re import threading import time from ..minicurses import ( BreaklineStatusPrinter, MultilineLogger, MultilinePrinter, QuietMultilinePrinter, ) from ..utils import ( IDENTITY, NO_DEFAULT, LockingUnsupportedError, Namespace, RetryManager, classproperty, deprecation_warning, format_bytes, join_nonempty, parse_bytes, remove_start, sanitize_open, shell_quote, timeconvert, timetuple_from_msec, try_call, ) from ..utils._utils import _ProgressState class FileDownloader: """File Downloader class. File downloader objects are the ones responsible of downloading the actual video file and writing it to disk. File downloaders accept a lot of parameters. In order not to saturate the object constructor with arguments, it receives a dictionary of options instead. Available options: verbose: Print additional info to stdout. quiet: Do not print messages to stdout. ratelimit: Download speed limit, in bytes/sec. throttledratelimit: Assume the download is being throttled below this speed (bytes/sec) retries: Number of times to retry for expected network errors. Default is 0 for API, but 10 for CLI file_access_retries: Number of times to retry on file access error (default: 3) buffersize: Size of download buffer in bytes. noresizebuffer: Do not automatically resize the download buffer. continuedl: Try to continue downloads if possible. noprogress: Do not print the progress bar. nopart: Do not use temporary .part files. updatetime: Use the Last-modified header to set output file timestamps. test: Download only first bytes to test the downloader. min_filesize: Skip files smaller than this size max_filesize: Skip files larger than this size progress_delta: The minimum time between progress output, in seconds external_downloader_args: A dictionary of downloader keys (in lower case) and a list of additional command-line arguments for the executable. Use 'default' as the name for arguments to be passed to all downloaders. For compatibility with youtube-dl, a single list of args can also be used hls_use_mpegts: Use the mpegts container for HLS videos. http_chunk_size: Size of a chunk for chunk-based HTTP downloading. May be useful for bypassing bandwidth throttling imposed by a webserver (experimental) progress_template: See YoutubeDL.py retry_sleep_functions: See YoutubeDL.py Subclasses of this one must re-define the real_download method. """ _TEST_FILE_SIZE = 10241 params = None def __init__(self, ydl, params): """Create a FileDownloader object with the given options.""" self._set_ydl(ydl) self._progress_hooks = [] self.params = params self._prepare_multiline_status() self.add_progress_hook(self.report_progress) if self.params.get('progress_delta'): self._progress_delta_lock = threading.Lock() self._progress_delta_time = time.monotonic() def _set_ydl(self, ydl): self.ydl = ydl for func in ( 'deprecation_warning', 'deprecated_feature', 'report_error', 'report_file_already_downloaded', 'report_warning', 'to_console_title', 'to_stderr', 'trouble', 'write_debug', ): if not hasattr(self, func): setattr(self, func, getattr(ydl, func)) def to_screen(self, *args, **kargs): self.ydl.to_screen(*args, quiet=self.params.get('quiet'), **kargs) __to_screen = to_screen @classproperty def FD_NAME(cls): return re.sub(r'(?<=[a-z])(?=[A-Z])', '_', cls.__name__[:-2]).lower() @staticmethod def format_seconds(seconds): if seconds is None: return ' Unknown' time = timetuple_from_msec(seconds * 1000) if time.hours > 99: return '--:--:--' return '%02d:%02d:%02d' % time[:-1] @classmethod def format_eta(cls, seconds): return f'{remove_start(cls.format_seconds(seconds), "00:"):>8s}' @staticmethod def calc_percent(byte_counter, data_len): if data_len is None: return None return float(byte_counter) / float(data_len) * 100.0 @staticmethod def format_percent(percent): return ' N/A%' if percent is None else f'{percent:>5.1f}%' @classmethod def calc_eta(cls, start_or_rate, now_or_remaining, total=NO_DEFAULT, current=NO_DEFAULT): if total is NO_DEFAULT: rate, remaining = start_or_rate, now_or_remaining if None in (rate, remaining): return None return int(float(remaining) / rate) start, now = start_or_rate, now_or_remaining if total is None: return None if now is None: now = time.time() rate = cls.calc_speed(start, now, current) return rate and int((float(total) - float(current)) / rate) @staticmethod def calc_speed(start, now, bytes): dif = now - start if bytes == 0 or dif < 0.001: # One millisecond return None return float(bytes) / dif @staticmethod def format_speed(speed): return ' Unknown B/s' if speed is None else f'{format_bytes(speed):>10s}/s' @staticmethod def format_retries(retries): return 'inf' if retries == float('inf') else int(retries) @staticmethod def filesize_or_none(unencoded_filename): if os.path.isfile(unencoded_filename): return os.path.getsize(unencoded_filename) return 0 @staticmethod def best_block_size(elapsed_time, bytes): new_min = max(bytes / 2.0, 1.0) new_max = min(max(bytes * 2.0, 1.0), 4194304) # Do not surpass 4 MB if elapsed_time < 0.001: return int(new_max) rate = bytes / elapsed_time if rate > new_max: return int(new_max) if rate < new_min: return int(new_min) return int(rate) @staticmethod def parse_bytes(bytestr): """Parse a string indicating a byte quantity into an integer.""" deprecation_warning('yt_dlp.FileDownloader.parse_bytes is deprecated and ' 'may be removed in the future. Use yt_dlp.utils.parse_bytes instead') return parse_bytes(bytestr) def slow_down(self, start_time, now, byte_counter): """Sleep if the download speed is over the rate limit.""" rate_limit = self.params.get('ratelimit') if rate_limit is None or byte_counter == 0: return if now is None: now = time.time() elapsed = now - start_time if elapsed <= 0.0: return speed = float(byte_counter) / elapsed if speed > rate_limit: sleep_time = float(byte_counter) / rate_limit - elapsed if sleep_time > 0: time.sleep(sleep_time) def temp_name(self, filename): """Returns a temporary filename for the given filename.""" if self.params.get('nopart', False) or filename == '-' or \ (os.path.exists(filename) and not os.path.isfile(filename)): return filename return filename + '.part' def undo_temp_name(self, filename): if filename.endswith('.part'): return filename[:-len('.part')] return filename def ytdl_filename(self, filename): return filename + '.ytdl' def wrap_file_access(action, *, fatal=False): def error_callback(err, count, retries, *, fd): return RetryManager.report_retry( err, count, retries, info=fd.__to_screen, warn=lambda e: (time.sleep(0.01), fd.to_screen(f'[download] Unable to {action} file: {e}')), error=None if fatal else lambda e: fd.report_error(f'Unable to {action} file: {e}'), sleep_func=fd.params.get('retry_sleep_functions', {}).get('file_access')) def wrapper(self, func, *args, **kwargs): for retry in RetryManager(self.params.get('file_access_retries', 3), error_callback, fd=self): try: return func(self, *args, **kwargs) except OSError as err: if err.errno in (errno.EACCES, errno.EINVAL): retry.error = err continue retry.error_callback(err, 1, 0) return functools.partial(functools.partialmethod, wrapper) @wrap_file_access('open', fatal=True) def sanitize_open(self, filename, open_mode): f, filename = sanitize_open(filename, open_mode) if not getattr(f, 'locked', None): self.write_debug(f'{LockingUnsupportedError.msg}. Proceeding without locking', only_once=True) return f, filename @wrap_file_access('remove') def try_remove(self, filename): if os.path.isfile(filename): os.remove(filename) @wrap_file_access('rename') def try_rename(self, old_filename, new_filename): if old_filename == new_filename: return os.replace(old_filename, new_filename) def try_utime(self, filename, last_modified_hdr): """Try to set the last-modified time of the given file.""" if last_modified_hdr is None: return if not os.path.isfile(filename): return timestr = last_modified_hdr if timestr is None: return filetime = timeconvert(timestr) if filetime is None: return filetime # Ignore obviously invalid dates if filetime == 0: return with contextlib.suppress(Exception): os.utime(filename, (time.time(), filetime)) return filetime def report_destination(self, filename): """Report destination filename.""" self.to_screen('[download] Destination: ' + filename) def _prepare_multiline_status(self, lines=1): if self.params.get('noprogress'): self._multiline = QuietMultilinePrinter() elif self.ydl.params.get('logger'): self._multiline = MultilineLogger(self.ydl.params['logger'], lines) elif self.params.get('progress_with_newline'): self._multiline = BreaklineStatusPrinter(self.ydl._out_files.out, lines) else: self._multiline = MultilinePrinter(self.ydl._out_files.out, lines, not self.params.get('quiet')) self._multiline.allow_colors = self.ydl._allow_colors.out and self.ydl._allow_colors.out != 'no_color' self._multiline._HAVE_FULLCAP = self.ydl._allow_colors.out def _finish_multiline_status(self): self._multiline.end() ProgressStyles = Namespace( downloaded_bytes='light blue', percent='light blue', eta='yellow', speed='green', elapsed='bold white', total_bytes='', total_bytes_estimate='', ) def _report_progress_status(self, s, default_template): for name, style in self.ProgressStyles.items_: name = f'_{name}_str' if name not in s: continue s[name] = self._format_progress(s[name], style) s['_default_template'] = default_template % s progress_dict = s.copy() progress_dict.pop('info_dict') progress_dict = {'info': s['info_dict'], 'progress': progress_dict} progress_template = self.params.get('progress_template', {}) self._multiline.print_at_line(self.ydl.evaluate_outtmpl( progress_template.get('download') or '[download] %(progress._default_template)s', progress_dict), s.get('progress_idx') or 0) self.to_console_title(self.ydl.evaluate_outtmpl( progress_template.get('download-title') or 'yt-dlp %(progress._default_template)s', progress_dict), _ProgressState.from_dict(s), s.get('_percent')) def _format_progress(self, *args, **kwargs): return self.ydl._format_text( self._multiline.stream, self._multiline.allow_colors, *args, **kwargs) def report_progress(self, s): def with_fields(*tups, default=''): for *fields, tmpl in tups: if all(s.get(f) is not None for f in fields): return tmpl return default _format_bytes = lambda k: f'{format_bytes(s.get(k)):>10s}' if s['status'] == 'finished': if self.params.get('noprogress'): self.to_screen('[download] Download completed') speed = try_call(lambda: s['total_bytes'] / s['elapsed']) s.update({ 'speed': speed, '_speed_str': self.format_speed(speed).strip(), '_total_bytes_str': _format_bytes('total_bytes'), '_elapsed_str': self.format_seconds(s.get('elapsed')), '_percent': 100.0, '_percent_str': self.format_percent(100), }) self._report_progress_status(s, join_nonempty( '100%%', with_fields(('total_bytes', 'of %(_total_bytes_str)s')), with_fields(('elapsed', 'in %(_elapsed_str)s')), with_fields(('speed', 'at %(_speed_str)s')), delim=' ')) if s['status'] != 'downloading': return if update_delta := self.params.get('progress_delta'): with self._progress_delta_lock: if time.monotonic() < self._progress_delta_time: return self._progress_delta_time += update_delta progress = try_call( lambda: 100 * s['downloaded_bytes'] / s['total_bytes'], lambda: 100 * s['downloaded_bytes'] / s['total_bytes_estimate'], lambda: s['downloaded_bytes'] == 0 and 0) s.update({ '_eta_str': self.format_eta(s.get('eta')).strip(), '_speed_str': self.format_speed(s.get('speed')), '_percent': progress, '_percent_str': self.format_percent(progress), '_total_bytes_str': _format_bytes('total_bytes'), '_total_bytes_estimate_str': _format_bytes('total_bytes_estimate'), '_downloaded_bytes_str': _format_bytes('downloaded_bytes'), '_elapsed_str': self.format_seconds(s.get('elapsed')), }) msg_template = with_fields( ('total_bytes', '%(_percent_str)s of %(_total_bytes_str)s at %(_speed_str)s ETA %(_eta_str)s'), ('total_bytes_estimate', '%(_percent_str)s of ~%(_total_bytes_estimate_str)s at %(_speed_str)s ETA %(_eta_str)s'), ('downloaded_bytes', 'elapsed', '%(_downloaded_bytes_str)s at %(_speed_str)s (%(_elapsed_str)s)'), ('downloaded_bytes', '%(_downloaded_bytes_str)s at %(_speed_str)s'), default='%(_percent_str)s at %(_speed_str)s ETA %(_eta_str)s') msg_template += with_fields( ('fragment_index', 'fragment_count', ' (frag %(fragment_index)s/%(fragment_count)s)'), ('fragment_index', ' (frag %(fragment_index)s)')) self._report_progress_status(s, msg_template) def report_resuming_byte(self, resume_len): """Report attempt to resume at given byte.""" self.to_screen(f'[download] Resuming download at byte {resume_len}') def report_retry(self, err, count, retries, frag_index=NO_DEFAULT, fatal=True): """Report retry""" is_frag = False if frag_index is NO_DEFAULT else 'fragment' RetryManager.report_retry( err, count, retries, info=self.__to_screen, warn=lambda msg: self.__to_screen(f'[download] Got error: {msg}'), error=IDENTITY if not fatal else lambda e: self.report_error(f'\r[download] Got error: {e}'), sleep_func=self.params.get('retry_sleep_functions', {}).get(is_frag or 'http'), suffix=f'fragment{"s" if frag_index is None else f" {frag_index}"}' if is_frag else None) def report_unable_to_resume(self): """Report it was impossible to resume download.""" self.to_screen('[download] Unable to resume') @staticmethod def supports_manifest(manifest): """ Whether the downloader can download the fragments from the manifest. Redefine in subclasses if needed. """ pass def download(self, filename, info_dict, subtitle=False): """Download to a filename using the info from info_dict Return True on success and False otherwise """ nooverwrites_and_exists = ( not self.params.get('overwrites', True) and os.path.exists(filename) ) if not hasattr(filename, 'write'): continuedl_and_exists = ( self.params.get('continuedl', True) and os.path.isfile(filename) and not self.params.get('nopart', False) ) # Check file already present if filename != '-' and (nooverwrites_and_exists or continuedl_and_exists): self.report_file_already_downloaded(filename) self._hook_progress({ 'filename': filename, 'status': 'finished', 'total_bytes': os.path.getsize(filename), }, info_dict) self._finish_multiline_status() return True, False sleep_note = '' if subtitle: sleep_interval = self.params.get('sleep_interval_subtitles') or 0 else: min_sleep_interval = self.params.get('sleep_interval') or 0 max_sleep_interval = self.params.get('max_sleep_interval') or 0 requested_formats = info_dict.get('requested_formats') or [info_dict] if available_at := max(f.get('available_at') or 0 for f in requested_formats): forced_sleep_interval = available_at - int(time.time()) if forced_sleep_interval > min_sleep_interval: sleep_note = 'as required by the site' min_sleep_interval = forced_sleep_interval if forced_sleep_interval > max_sleep_interval: max_sleep_interval = forced_sleep_interval sleep_interval = random.uniform( min_sleep_interval, max_sleep_interval or min_sleep_interval) if sleep_interval > 0: self.to_screen(f'[download] Sleeping {sleep_interval:.2f} seconds {sleep_note}...') time.sleep(sleep_interval) ret = self.real_download(filename, info_dict) self._finish_multiline_status() return ret, True def real_download(self, filename, info_dict): """Real download process. Redefine in subclasses.""" raise NotImplementedError('This method must be implemented by subclasses') def _hook_progress(self, status, info_dict): # Ideally we want to make a copy of the dict, but that is too slow status['info_dict'] = info_dict # youtube-dl passes the same status object to all the hooks. # Some third party scripts seems to be relying on this. # So keep this behavior if possible for ph in self._progress_hooks: ph(status) def add_progress_hook(self, ph): # See YoutubeDl.py (search for progress_hooks) for a description of # this interface self._progress_hooks.append(ph) def _debug_cmd(self, args, exe=None): if not self.params.get('verbose', False): return if exe is None: exe = os.path.basename(args[0]) self.write_debug(f'{exe} command line: {shell_quote(args)}') def _get_impersonate_target(self, info_dict): impersonate = info_dict.get('impersonate') if impersonate is None: return None available_target, requested_targets = self.ydl._parse_impersonate_targets(impersonate) if available_target: return available_target elif requested_targets: self.report_warning(self.ydl._unavailable_targets_message(requested_targets)) return None
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/downloader/fc2.py
yt_dlp/downloader/fc2.py
import threading from .common import FileDownloader from .external import FFmpegFD class FC2LiveFD(FileDownloader): """ Downloads FC2 live without being stopped. <br> Note, this is not a part of public API, and will be removed without notice. DO NOT USE """ def real_download(self, filename, info_dict): ws = info_dict['ws'] heartbeat_lock = threading.Lock() heartbeat_state = [None, 1] def heartbeat(): if heartbeat_state[1] < 0: return try: heartbeat_state[1] += 1 ws.send('{"name":"heartbeat","arguments":{},"id":%d}' % heartbeat_state[1]) except Exception: self.to_screen('[fc2:live] Heartbeat failed') with heartbeat_lock: heartbeat_state[0] = threading.Timer(30, heartbeat) heartbeat_state[0]._daemonic = True heartbeat_state[0].start() heartbeat() new_info_dict = info_dict.copy() new_info_dict.update({ 'ws': None, 'protocol': 'live_ffmpeg', }) try: return FFmpegFD(self.ydl, self.params or {}).download(filename, new_info_dict) finally: # stop heartbeating heartbeat_state[1] = -1
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/downloader/f4m.py
yt_dlp/downloader/f4m.py
import base64 import io import itertools import struct import time import urllib.parse from .fragment import FragmentFD from ..compat import compat_etree_fromstring from ..networking.exceptions import HTTPError from ..utils import fix_xml_ampersands, xpath_text class DataTruncatedError(Exception): pass class FlvReader(io.BytesIO): """ Reader for Flv files The file format is documented in https://www.adobe.com/devnet/f4v.html """ def read_bytes(self, n): data = self.read(n) if len(data) < n: raise DataTruncatedError( 'FlvReader error: need %d bytes while only %d bytes got' % ( n, len(data))) return data # Utility functions for reading numbers and strings def read_unsigned_long_long(self): return struct.unpack('!Q', self.read_bytes(8))[0] def read_unsigned_int(self): return struct.unpack('!I', self.read_bytes(4))[0] def read_unsigned_char(self): return struct.unpack('!B', self.read_bytes(1))[0] def read_string(self): res = b'' while True: char = self.read_bytes(1) if char == b'\x00': break res += char return res def read_box_info(self): """ Read a box and return the info as a tuple: (box_size, box_type, box_data) """ real_size = size = self.read_unsigned_int() box_type = self.read_bytes(4) header_end = 8 if size == 1: real_size = self.read_unsigned_long_long() header_end = 16 return real_size, box_type, self.read_bytes(real_size - header_end) def read_asrt(self): # version self.read_unsigned_char() # flags self.read_bytes(3) quality_entry_count = self.read_unsigned_char() # QualityEntryCount for _ in range(quality_entry_count): self.read_string() segment_run_count = self.read_unsigned_int() segments = [] for _ in range(segment_run_count): first_segment = self.read_unsigned_int() fragments_per_segment = self.read_unsigned_int() segments.append((first_segment, fragments_per_segment)) return { 'segment_run': segments, } def read_afrt(self): # version self.read_unsigned_char() # flags self.read_bytes(3) # time scale self.read_unsigned_int() quality_entry_count = self.read_unsigned_char() # QualitySegmentUrlModifiers for _ in range(quality_entry_count): self.read_string() fragments_count = self.read_unsigned_int() fragments = [] for _ in range(fragments_count): first = self.read_unsigned_int() first_ts = self.read_unsigned_long_long() duration = self.read_unsigned_int() if duration == 0: discontinuity_indicator = self.read_unsigned_char() else: discontinuity_indicator = None fragments.append({ 'first': first, 'ts': first_ts, 'duration': duration, 'discontinuity_indicator': discontinuity_indicator, }) return { 'fragments': fragments, } def read_abst(self): # version self.read_unsigned_char() # flags self.read_bytes(3) self.read_unsigned_int() # BootstrapinfoVersion # Profile,Live,Update,Reserved flags = self.read_unsigned_char() live = flags & 0x20 != 0 # time scale self.read_unsigned_int() # CurrentMediaTime self.read_unsigned_long_long() # SmpteTimeCodeOffset self.read_unsigned_long_long() self.read_string() # MovieIdentifier server_count = self.read_unsigned_char() # ServerEntryTable for _ in range(server_count): self.read_string() quality_count = self.read_unsigned_char() # QualityEntryTable for _ in range(quality_count): self.read_string() # DrmData self.read_string() # MetaData self.read_string() segments_count = self.read_unsigned_char() segments = [] for _ in range(segments_count): _box_size, box_type, box_data = self.read_box_info() assert box_type == b'asrt' segment = FlvReader(box_data).read_asrt() segments.append(segment) fragments_run_count = self.read_unsigned_char() fragments = [] for _ in range(fragments_run_count): _box_size, box_type, box_data = self.read_box_info() assert box_type == b'afrt' fragments.append(FlvReader(box_data).read_afrt()) return { 'segments': segments, 'fragments': fragments, 'live': live, } def read_bootstrap_info(self): _, box_type, box_data = self.read_box_info() assert box_type == b'abst' return FlvReader(box_data).read_abst() def read_bootstrap_info(bootstrap_bytes): return FlvReader(bootstrap_bytes).read_bootstrap_info() def build_fragments_list(boot_info): """ Return a list of (segment, fragment) for each fragment in the video """ res = [] segment_run_table = boot_info['segments'][0] fragment_run_entry_table = boot_info['fragments'][0]['fragments'] first_frag_number = fragment_run_entry_table[0]['first'] fragments_counter = itertools.count(first_frag_number) for segment, fragments_count in segment_run_table['segment_run']: # In some live HDS streams (e.g. Rai), `fragments_count` is # abnormal and causing out-of-memory errors. It's OK to change the # number of fragments for live streams as they are updated periodically if fragments_count == 4294967295 and boot_info['live']: fragments_count = 2 for _ in range(fragments_count): res.append((segment, next(fragments_counter))) if boot_info['live']: res = res[-2:] return res def write_unsigned_int(stream, val): stream.write(struct.pack('!I', val)) def write_unsigned_int_24(stream, val): stream.write(struct.pack('!I', val)[1:]) def write_flv_header(stream): """Writes the FLV header to stream""" # FLV header stream.write(b'FLV\x01') stream.write(b'\x05') stream.write(b'\x00\x00\x00\x09') stream.write(b'\x00\x00\x00\x00') def write_metadata_tag(stream, metadata): """Writes optional metadata tag to stream""" SCRIPT_TAG = b'\x12' FLV_TAG_HEADER_LEN = 11 if metadata: stream.write(SCRIPT_TAG) write_unsigned_int_24(stream, len(metadata)) stream.write(b'\x00\x00\x00\x00\x00\x00\x00') stream.write(metadata) write_unsigned_int(stream, FLV_TAG_HEADER_LEN + len(metadata)) def remove_encrypted_media(media): return list(filter(lambda e: 'drmAdditionalHeaderId' not in e.attrib and 'drmAdditionalHeaderSetId' not in e.attrib, media)) def _add_ns(prop, ver=1): return '{http://ns.adobe.com/f4m/%d.0}%s' % (ver, prop) def get_base_url(manifest): base_url = xpath_text( manifest, [_add_ns('baseURL'), _add_ns('baseURL', 2)], 'base URL', default=None) if base_url: base_url = base_url.strip() return base_url class F4mFD(FragmentFD): """ A downloader for f4m manifests or AdobeHDS. """ def _get_unencrypted_media(self, doc): media = doc.findall(_add_ns('media')) if not media: self.report_error('No media found') if not self.params.get('allow_unplayable_formats'): for e in (doc.findall(_add_ns('drmAdditionalHeader')) + doc.findall(_add_ns('drmAdditionalHeaderSet'))): # If id attribute is missing it's valid for all media nodes # without drmAdditionalHeaderId or drmAdditionalHeaderSetId attribute if 'id' not in e.attrib: self.report_error('Missing ID in f4m DRM') media = remove_encrypted_media(media) if not media: self.report_error('Unsupported DRM') return media def _get_bootstrap_from_url(self, bootstrap_url): bootstrap = self.ydl.urlopen(bootstrap_url).read() return read_bootstrap_info(bootstrap) def _update_live_fragments(self, bootstrap_url, latest_fragment): fragments_list = [] retries = 30 while (not fragments_list) and (retries > 0): boot_info = self._get_bootstrap_from_url(bootstrap_url) fragments_list = build_fragments_list(boot_info) fragments_list = [f for f in fragments_list if f[1] > latest_fragment] if not fragments_list: # Retry after a while time.sleep(5.0) retries -= 1 if not fragments_list: self.report_error('Failed to update fragments') return fragments_list def _parse_bootstrap_node(self, node, base_url): # Sometimes non empty inline bootstrap info can be specified along # with bootstrap url attribute (e.g. dummy inline bootstrap info # contains whitespace characters in [1]). We will prefer bootstrap # url over inline bootstrap info when present. # 1. http://live-1-1.rutube.ru/stream/1024/HDS/SD/C2NKsS85HQNckgn5HdEmOQ/1454167650/S-s604419906/move/four/dirs/upper/1024-576p.f4m bootstrap_url = node.get('url') if bootstrap_url: bootstrap_url = urllib.parse.urljoin( base_url, bootstrap_url) boot_info = self._get_bootstrap_from_url(bootstrap_url) else: bootstrap_url = None bootstrap = base64.b64decode(node.text) boot_info = read_bootstrap_info(bootstrap) return boot_info, bootstrap_url def real_download(self, filename, info_dict): man_url = info_dict['url'] requested_bitrate = info_dict.get('tbr') self.to_screen(f'[{self.FD_NAME}] Downloading f4m manifest') urlh = self.ydl.urlopen(self._prepare_url(info_dict, man_url)) man_url = urlh.url # Some manifests may be malformed, e.g. prosiebensat1 generated manifests # (see https://github.com/ytdl-org/youtube-dl/issues/6215#issuecomment-121704244 # and https://github.com/ytdl-org/youtube-dl/issues/7823) manifest = fix_xml_ampersands(urlh.read().decode('utf-8', 'ignore')).strip() doc = compat_etree_fromstring(manifest) formats = [(int(f.attrib.get('bitrate', -1)), f) for f in self._get_unencrypted_media(doc)] if requested_bitrate is None or len(formats) == 1: # get the best format formats = sorted(formats, key=lambda f: f[0]) _, media = formats[-1] else: _, media = next(filter( lambda f: int(f[0]) == requested_bitrate, formats)) # Prefer baseURL for relative URLs as per 11.2 of F4M 3.0 spec. man_base_url = get_base_url(doc) or man_url base_url = urllib.parse.urljoin(man_base_url, media.attrib['url']) bootstrap_node = doc.find(_add_ns('bootstrapInfo')) boot_info, bootstrap_url = self._parse_bootstrap_node( bootstrap_node, man_base_url) live = boot_info['live'] metadata_node = media.find(_add_ns('metadata')) if metadata_node is not None: metadata = base64.b64decode(metadata_node.text) else: metadata = None fragments_list = build_fragments_list(boot_info) test = self.params.get('test', False) if test: # We only download the first fragment fragments_list = fragments_list[:1] total_frags = len(fragments_list) # For some akamai manifests we'll need to add a query to the fragment url akamai_pv = xpath_text(doc, _add_ns('pv-2.0')) ctx = { 'filename': filename, 'total_frags': total_frags, 'live': bool(live), } self._prepare_frag_download(ctx) dest_stream = ctx['dest_stream'] if ctx['complete_frags_downloaded_bytes'] == 0: write_flv_header(dest_stream) if not live: write_metadata_tag(dest_stream, metadata) base_url_parsed = urllib.parse.urlparse(base_url) self._start_frag_download(ctx, info_dict) frag_index = 0 while fragments_list: seg_i, frag_i = fragments_list.pop(0) frag_index += 1 if frag_index <= ctx['fragment_index']: continue name = 'Seg%d-Frag%d' % (seg_i, frag_i) query = [] if base_url_parsed.query: query.append(base_url_parsed.query) if akamai_pv: query.append(akamai_pv.strip(';')) if info_dict.get('extra_param_to_segment_url'): query.append(info_dict['extra_param_to_segment_url']) url_parsed = base_url_parsed._replace(path=base_url_parsed.path + name, query='&'.join(query)) try: success = self._download_fragment(ctx, url_parsed.geturl(), info_dict) if not success: return False down_data = self._read_fragment(ctx) reader = FlvReader(down_data) while True: try: _, box_type, box_data = reader.read_box_info() except DataTruncatedError: if test: # In tests, segments may be truncated, and thus # FlvReader may not be able to parse the whole # chunk. If so, write the segment as is # See https://github.com/ytdl-org/youtube-dl/issues/9214 dest_stream.write(down_data) break raise if box_type == b'mdat': self._append_fragment(ctx, box_data) break except HTTPError as err: if live and (err.status == 404 or err.status == 410): # We didn't keep up with the live window. Continue # with the next available fragment. msg = 'Fragment %d unavailable' % frag_i self.report_warning(msg) fragments_list = [] else: raise if not fragments_list and not test and live and bootstrap_url: fragments_list = self._update_live_fragments(bootstrap_url, frag_i) total_frags += len(fragments_list) if fragments_list and (fragments_list[0][1] > frag_i + 1): msg = 'Missed %d fragments' % (fragments_list[0][1] - (frag_i + 1)) self.report_warning(msg) return self._finish_frag_download(ctx, info_dict)
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/downloader/bunnycdn.py
yt_dlp/downloader/bunnycdn.py
import hashlib import random import threading from .common import FileDownloader from . import HlsFD from ..networking import Request from ..networking.exceptions import network_exceptions class BunnyCdnFD(FileDownloader): """ Downloads from BunnyCDN with required pings Note, this is not a part of public API, and will be removed without notice. DO NOT USE """ def real_download(self, filename, info_dict): self.to_screen(f'[{self.FD_NAME}] Downloading from BunnyCDN') fd = HlsFD(self.ydl, self.params) stop_event = threading.Event() ping_thread = threading.Thread(target=self.ping_thread, args=(stop_event,), kwargs=info_dict['_bunnycdn_ping_data']) ping_thread.start() try: return fd.real_download(filename, info_dict) finally: stop_event.set() def ping_thread(self, stop_event, url, headers, secret, context_id): # Site sends ping every 4 seconds, but this throttles the download. Pinging every 2 seconds seems to work. ping_interval = 2 # Hard coded resolution as it doesn't seem to matter res = 1080 paused = 'false' current_time = 0 while not stop_event.wait(ping_interval): current_time += ping_interval time = current_time + round(random.random(), 6) md5_hash = hashlib.md5(f'{secret}_{context_id}_{time}_{paused}_{res}'.encode()).hexdigest() ping_url = f'{url}?hash={md5_hash}&time={time}&paused={paused}&resolution={res}' try: self.ydl.urlopen(Request(ping_url, headers=headers)).read() except network_exceptions as e: self.to_screen(f'[{self.FD_NAME}] Ping failed: {e}')
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/downloader/websocket.py
yt_dlp/downloader/websocket.py
import asyncio import contextlib import os import signal import threading from .common import FileDownloader from .external import FFmpegFD from ..dependencies import websockets class FFmpegSinkFD(FileDownloader): """ A sink to ffmpeg for downloading fragments in any form """ def real_download(self, filename, info_dict): info_copy = info_dict.copy() info_copy['url'] = '-' async def call_conn(proc, stdin): try: await self.real_connection(stdin, info_dict) except OSError: pass finally: with contextlib.suppress(OSError): stdin.flush() stdin.close() os.kill(os.getpid(), signal.SIGINT) class FFmpegStdinFD(FFmpegFD): @classmethod def get_basename(cls): return FFmpegFD.get_basename() def on_process_started(self, proc, stdin): thread = threading.Thread(target=asyncio.run, daemon=True, args=(call_conn(proc, stdin), )) thread.start() return FFmpegStdinFD(self.ydl, self.params or {}).download(filename, info_copy) async def real_connection(self, sink, info_dict): """ Override this in subclasses """ raise NotImplementedError('This method must be implemented by subclasses') class WebSocketFragmentFD(FFmpegSinkFD): async def real_connection(self, sink, info_dict): async with websockets.connect(info_dict['url'], extra_headers=info_dict.get('http_headers', {})) as ws: while True: recv = await ws.recv() if isinstance(recv, str): recv = recv.encode('utf8') sink.write(recv)
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/downloader/mhtml.py
yt_dlp/downloader/mhtml.py
import io import quopri import re import uuid from .fragment import FragmentFD from ..compat import imghdr from ..utils import escapeHTML, formatSeconds, srt_subtitles_timecode, urljoin from ..version import __version__ as YT_DLP_VERSION class MhtmlFD(FragmentFD): _STYLESHEET = '''\ html, body { margin: 0; padding: 0; height: 100vh; } html { overflow-y: scroll; scroll-snap-type: y mandatory; } body { scroll-snap-type: y mandatory; display: flex; flex-flow: column; } body > figure { max-width: 100vw; max-height: 100vh; scroll-snap-align: center; } body > figure > figcaption { text-align: center; height: 2.5em; } body > figure > img { display: block; margin: auto; max-width: 100%; max-height: calc(100vh - 5em); } ''' _STYLESHEET = re.sub(r'\s+', ' ', _STYLESHEET) _STYLESHEET = re.sub(r'\B \B|(?<=[\w\-]) (?=[^\w\-])|(?<=[^\w\-]) (?=[\w\-])', '', _STYLESHEET) @staticmethod def _escape_mime(s): return '=?utf-8?Q?' + (b''.join( bytes((b,)) if b >= 0x20 else b'=%02X' % b for b in quopri.encodestring(s.encode(), header=True) )).decode('us-ascii') + '?=' def _gen_cid(self, i, fragment, frag_boundary): return f'{i}.{frag_boundary}@yt-dlp.github.io.invalid' def _gen_stub(self, *, fragments, frag_boundary, title): output = io.StringIO() output.write( '<!DOCTYPE html>' '<html>' '<head>' f'<meta name="generator" content="yt-dlp {escapeHTML(YT_DLP_VERSION)}">' f'<title>{escapeHTML(title)}</title>' f'<style>{self._STYLESHEET}</style>' '<body>') t0 = 0 for i, frag in enumerate(fragments): output.write('<figure>') try: t1 = t0 + frag['duration'] output.write(( '<figcaption>Slide #{num}: {t0} – {t1} (duration: {duration})</figcaption>' ).format( num=i + 1, t0=srt_subtitles_timecode(t0), t1=srt_subtitles_timecode(t1), duration=formatSeconds(frag['duration'], msec=True), )) except (KeyError, ValueError, TypeError): t1 = None output.write(f'<figcaption>Slide #{i + 1}</figcaption>') output.write(f'<img src="cid:{self._gen_cid(i, frag, frag_boundary)}">') output.write('</figure>') t0 = t1 return output.getvalue() def real_download(self, filename, info_dict): fragment_base_url = info_dict.get('fragment_base_url') fragments = info_dict['fragments'][:1] if self.params.get( 'test', False) else info_dict['fragments'] title = info_dict.get('title', info_dict['format_id']) origin = info_dict.get('webpage_url', info_dict['url']) ctx = { 'filename': filename, 'total_frags': len(fragments), } self._prepare_and_start_frag_download(ctx, info_dict) extra_state = ctx.setdefault('extra_state', { 'header_written': False, 'mime_boundary': str(uuid.uuid4()).replace('-', ''), }) frag_boundary = extra_state['mime_boundary'] if not extra_state['header_written']: stub = self._gen_stub( fragments=fragments, frag_boundary=frag_boundary, title=title, ) ctx['dest_stream'].write(( 'MIME-Version: 1.0\r\n' 'From: <nowhere@yt-dlp.github.io.invalid>\r\n' 'To: <nowhere@yt-dlp.github.io.invalid>\r\n' f'Subject: {self._escape_mime(title)}\r\n' 'Content-type: multipart/related; ' f'boundary="{frag_boundary}"; ' 'type="text/html"\r\n' f'X.yt-dlp.Origin: {origin}\r\n' '\r\n' f'--{frag_boundary}\r\n' 'Content-Type: text/html; charset=utf-8\r\n' f'Content-Length: {len(stub)}\r\n' '\r\n' f'{stub}\r\n').encode()) extra_state['header_written'] = True for i, fragment in enumerate(fragments): if (i + 1) <= ctx['fragment_index']: continue fragment_url = fragment.get('url') if not fragment_url: assert fragment_base_url fragment_url = urljoin(fragment_base_url, fragment['path']) success = self._download_fragment(ctx, fragment_url, info_dict) if not success: continue frag_content = self._read_fragment(ctx) frag_header = io.BytesIO() frag_header.write( b'--%b\r\n' % frag_boundary.encode('us-ascii')) frag_header.write( b'Content-ID: <%b>\r\n' % self._gen_cid(i, fragment, frag_boundary).encode('us-ascii')) frag_header.write( b'Content-type: %b\r\n' % f'image/{imghdr.what(h=frag_content) or "jpeg"}'.encode()) frag_header.write( b'Content-length: %u\r\n' % len(frag_content)) frag_header.write( b'Content-location: %b\r\n' % fragment_url.encode('us-ascii')) frag_header.write( b'X.yt-dlp.Duration: %f\r\n' % fragment['duration']) frag_header.write(b'\r\n') self._append_fragment( ctx, frag_header.getvalue() + frag_content + b'\r\n') ctx['dest_stream'].write( b'--%b--\r\n\r\n' % frag_boundary.encode('us-ascii')) return self._finish_frag_download(ctx, info_dict)
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/downloader/dash.py
yt_dlp/downloader/dash.py
import time import urllib.parse from . import get_suitable_downloader from .fragment import FragmentFD from ..utils import ReExtractInfo, update_url_query, urljoin class DashSegmentsFD(FragmentFD): """ Download segments in a DASH manifest. External downloaders can take over the fragment downloads by supporting the 'dash_frag_urls' protocol """ FD_NAME = 'dashsegments' def real_download(self, filename, info_dict): if 'http_dash_segments_generator' in info_dict['protocol'].split('+'): real_downloader = None # No external FD can support --live-from-start else: if info_dict.get('is_live'): self.report_error('Live DASH videos are not supported') real_downloader = get_suitable_downloader( info_dict, self.params, None, protocol='dash_frag_urls', to_stdout=(filename == '-')) real_start = time.time() requested_formats = [{**info_dict, **fmt} for fmt in info_dict.get('requested_formats', [])] args = [] for fmt in requested_formats or [info_dict]: # Re-extract if --load-info-json is used and 'fragments' was originally a generator # See https://github.com/yt-dlp/yt-dlp/issues/13906 if isinstance(fmt['fragments'], str): raise ReExtractInfo('the stream needs to be re-extracted', expected=True) try: fragment_count = 1 if self.params.get('test') else len(fmt['fragments']) except TypeError: fragment_count = None ctx = { 'filename': fmt.get('filepath') or filename, 'live': 'is_from_start' if fmt.get('is_from_start') else fmt.get('is_live'), 'total_frags': fragment_count, } if real_downloader: self._prepare_external_frag_download(ctx) else: self._prepare_and_start_frag_download(ctx, fmt) ctx['start'] = real_start extra_query = None extra_param_to_segment_url = info_dict.get('extra_param_to_segment_url') if extra_param_to_segment_url: extra_query = urllib.parse.parse_qs(extra_param_to_segment_url) fragments_to_download = self._get_fragments(fmt, ctx, extra_query) if real_downloader: self.to_screen( f'[{self.FD_NAME}] Fragment downloads will be delegated to {real_downloader.get_basename()}') info_dict['fragments'] = list(fragments_to_download) fd = real_downloader(self.ydl, self.params) return fd.real_download(filename, info_dict) args.append([ctx, fragments_to_download, fmt]) return self.download_and_append_fragments_multiple(*args, is_fatal=lambda idx: idx == 0) def _resolve_fragments(self, fragments, ctx): fragments = fragments(ctx) if callable(fragments) else fragments return [next(iter(fragments))] if self.params.get('test') else fragments def _get_fragments(self, fmt, ctx, extra_query): fragment_base_url = fmt.get('fragment_base_url') fragments = self._resolve_fragments(fmt['fragments'], ctx) frag_index = 0 for i, fragment in enumerate(fragments): frag_index += 1 if frag_index <= ctx['fragment_index']: continue fragment_url = fragment.get('url') if not fragment_url: assert fragment_base_url fragment_url = urljoin(fragment_base_url, fragment['path']) if extra_query: fragment_url = update_url_query(fragment_url, extra_query) yield { 'frag_index': frag_index, 'fragment_count': fragment.get('fragment_count'), 'index': i, 'url': fragment_url, }
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/downloader/rtsp.py
yt_dlp/downloader/rtsp.py
import os import subprocess from .common import FileDownloader from ..utils import check_executable class RtspFD(FileDownloader): def real_download(self, filename, info_dict): url = info_dict['url'] self.report_destination(filename) tmpfilename = self.temp_name(filename) if check_executable('mplayer', ['-h']): args = [ 'mplayer', '-really-quiet', '-vo', 'null', '-vc', 'dummy', '-dumpstream', '-dumpfile', tmpfilename, url] elif check_executable('mpv', ['-h']): args = [ 'mpv', '-really-quiet', '--vo=null', '--stream-dump=' + tmpfilename, url] else: self.report_error('MMS or RTSP download detected but neither "mplayer" nor "mpv" could be run. Please install one') return False self._debug_cmd(args) retval = subprocess.call(args) if retval == 0: fsize = os.path.getsize(tmpfilename) self.to_screen(f'\r[{args[0]}] {fsize} bytes') self.try_rename(tmpfilename, filename) self._hook_progress({ 'downloaded_bytes': fsize, 'total_bytes': fsize, 'filename': filename, 'status': 'finished', }, info_dict) return True else: self.to_stderr('\n') self.report_error('%s exited with code %d' % (args[0], retval)) return False
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/downloader/external.py
yt_dlp/downloader/external.py
import enum import functools import json import os import re import subprocess import sys import tempfile import time import uuid from .fragment import FragmentFD from ..networking import Request from ..postprocessor.ffmpeg import EXT_TO_OUT_FORMATS, FFmpegPostProcessor from ..utils import ( Popen, RetryManager, _configuration_args, check_executable, classproperty, cli_bool_option, cli_option, cli_valueless_option, determine_ext, encodeArgument, find_available_port, remove_end, traverse_obj, ) class Features(enum.Enum): TO_STDOUT = enum.auto() MULTIPLE_FORMATS = enum.auto() class ExternalFD(FragmentFD): SUPPORTED_PROTOCOLS = ('http', 'https', 'ftp', 'ftps') SUPPORTED_FEATURES = () _CAPTURE_STDERR = True def real_download(self, filename, info_dict): self.report_destination(filename) tmpfilename = self.temp_name(filename) self._cookies_tempfile = None try: started = time.time() retval = self._call_downloader(tmpfilename, info_dict) except KeyboardInterrupt: if not info_dict.get('is_live'): raise # Live stream downloading cancellation should be considered as # correct and expected termination thus all postprocessing # should take place retval = 0 self.to_screen(f'[{self.get_basename()}] Interrupted by user') finally: if self._cookies_tempfile: self.try_remove(self._cookies_tempfile) if retval == 0: status = { 'filename': filename, 'status': 'finished', 'elapsed': time.time() - started, } if filename != '-': fsize = os.path.getsize(tmpfilename) self.try_rename(tmpfilename, filename) status.update({ 'downloaded_bytes': fsize, 'total_bytes': fsize, }) self._hook_progress(status, info_dict) return True else: self.to_stderr('\n') self.report_error('%s exited with code %d' % ( self.get_basename(), retval)) return False @classmethod def get_basename(cls): return cls.__name__[:-2].lower() @classproperty def EXE_NAME(cls): return cls.get_basename() @functools.cached_property def exe(self): return self.EXE_NAME @classmethod def available(cls, path=None): path = check_executable( cls.EXE_NAME if path in (None, cls.get_basename()) else path, [cls.AVAILABLE_OPT]) if not path: return False cls.exe = path return path @classmethod def supports(cls, info_dict): return all(( not info_dict.get('to_stdout') or Features.TO_STDOUT in cls.SUPPORTED_FEATURES, '+' not in info_dict['protocol'] or Features.MULTIPLE_FORMATS in cls.SUPPORTED_FEATURES, not traverse_obj(info_dict, ('hls_aes', ...), 'extra_param_to_segment_url', 'extra_param_to_key_url'), all(proto in cls.SUPPORTED_PROTOCOLS for proto in info_dict['protocol'].split('+')), )) @classmethod def can_download(cls, info_dict, path=None): return cls.available(path) and cls.supports(info_dict) def _option(self, command_option, param): return cli_option(self.params, command_option, param) def _bool_option(self, command_option, param, true_value='true', false_value='false', separator=None): return cli_bool_option(self.params, command_option, param, true_value, false_value, separator) def _valueless_option(self, command_option, param, expected_value=True): return cli_valueless_option(self.params, command_option, param, expected_value) def _configuration_args(self, keys=None, *args, **kwargs): return _configuration_args( self.get_basename(), self.params.get('external_downloader_args'), self.EXE_NAME, keys, *args, **kwargs) def _write_cookies(self): if not self.ydl.cookiejar.filename: tmp_cookies = tempfile.NamedTemporaryFile(suffix='.cookies', delete=False) tmp_cookies.close() self._cookies_tempfile = tmp_cookies.name self.to_screen(f'[download] Writing temporary cookies file to "{self._cookies_tempfile}"') # real_download resets _cookies_tempfile; if it's None then save() will write to cookiejar.filename self.ydl.cookiejar.save(self._cookies_tempfile) return self.ydl.cookiejar.filename or self._cookies_tempfile def _call_downloader(self, tmpfilename, info_dict): """ Either overwrite this or implement _make_cmd """ cmd = [encodeArgument(a) for a in self._make_cmd(tmpfilename, info_dict)] self._debug_cmd(cmd) if 'fragments' not in info_dict: _, stderr, returncode = self._call_process(cmd, info_dict) if returncode and stderr: self.to_stderr(stderr) return returncode skip_unavailable_fragments = self.params.get('skip_unavailable_fragments', True) retry_manager = RetryManager(self.params.get('fragment_retries'), self.report_retry, frag_index=None, fatal=not skip_unavailable_fragments) for retry in retry_manager: _, stderr, returncode = self._call_process(cmd, info_dict) if not returncode: break # TODO: Decide whether to retry based on error code # https://aria2.github.io/manual/en/html/aria2c.html#exit-status if stderr: self.to_stderr(stderr) retry.error = Exception() continue if not skip_unavailable_fragments and retry_manager.error: return -1 decrypt_fragment = self.decrypter(info_dict) dest, _ = self.sanitize_open(tmpfilename, 'wb') for frag_index, fragment in enumerate(info_dict['fragments']): fragment_filename = f'{tmpfilename}-Frag{frag_index}' try: src, _ = self.sanitize_open(fragment_filename, 'rb') except OSError as err: if skip_unavailable_fragments and frag_index > 1: self.report_skip_fragment(frag_index, err) continue self.report_error(f'Unable to open fragment {frag_index}; {err}') return -1 dest.write(decrypt_fragment(fragment, src.read())) src.close() if not self.params.get('keep_fragments', False): self.try_remove(fragment_filename) dest.close() self.try_remove(f'{tmpfilename}.frag.urls') return 0 def _call_process(self, cmd, info_dict): return Popen.run(cmd, text=True, stderr=subprocess.PIPE if self._CAPTURE_STDERR else None) class CurlFD(ExternalFD): AVAILABLE_OPT = '-V' _CAPTURE_STDERR = False # curl writes the progress to stderr def _make_cmd(self, tmpfilename, info_dict): cmd = [self.exe, '--location', '-o', tmpfilename, '--compressed'] cookie_header = self.ydl.cookiejar.get_cookie_header(info_dict['url']) if cookie_header: cmd += ['--cookie', cookie_header] if info_dict.get('http_headers') is not None: for key, val in info_dict['http_headers'].items(): cmd += ['--header', f'{key}: {val}'] cmd += self._bool_option('--continue-at', 'continuedl', '-', '0') cmd += self._valueless_option('--silent', 'noprogress') cmd += self._valueless_option('--verbose', 'verbose') cmd += self._option('--limit-rate', 'ratelimit') retry = self._option('--retry', 'retries') if len(retry) == 2: if retry[1] in ('inf', 'infinite'): retry[1] = '2147483647' cmd += retry cmd += self._option('--max-filesize', 'max_filesize') cmd += self._option('--interface', 'source_address') cmd += self._option('--proxy', 'proxy') cmd += self._valueless_option('--insecure', 'nocheckcertificate') cmd += self._configuration_args() cmd += ['--', info_dict['url']] return cmd class AxelFD(ExternalFD): AVAILABLE_OPT = '-V' def _make_cmd(self, tmpfilename, info_dict): cmd = [self.exe, '-o', tmpfilename] if info_dict.get('http_headers') is not None: for key, val in info_dict['http_headers'].items(): cmd += ['-H', f'{key}: {val}'] cookie_header = self.ydl.cookiejar.get_cookie_header(info_dict['url']) if cookie_header: cmd += ['-H', f'Cookie: {cookie_header}', '--max-redirect=0'] cmd += self._configuration_args() cmd += ['--', info_dict['url']] return cmd class WgetFD(ExternalFD): AVAILABLE_OPT = '--version' def _make_cmd(self, tmpfilename, info_dict): cmd = [self.exe, '-O', tmpfilename, '-nv', '--compression=auto'] if self.ydl.cookiejar.get_cookie_header(info_dict['url']): cmd += ['--load-cookies', self._write_cookies()] if info_dict.get('http_headers') is not None: for key, val in info_dict['http_headers'].items(): cmd += ['--header', f'{key}: {val}'] cmd += self._option('--limit-rate', 'ratelimit') retry = self._option('--tries', 'retries') if len(retry) == 2: if retry[1] in ('inf', 'infinite'): retry[1] = '0' cmd += retry cmd += self._option('--bind-address', 'source_address') proxy = self.params.get('proxy') if proxy: for var in ('http_proxy', 'https_proxy'): cmd += ['--execute', f'{var}={proxy}'] cmd += self._valueless_option('--no-check-certificate', 'nocheckcertificate') cmd += self._configuration_args() cmd += ['--', info_dict['url']] return cmd class Aria2cFD(ExternalFD): AVAILABLE_OPT = '-v' SUPPORTED_PROTOCOLS = ('http', 'https', 'ftp', 'ftps', 'dash_frag_urls', 'm3u8_frag_urls') @staticmethod def supports_manifest(manifest): UNSUPPORTED_FEATURES = [ r'#EXT-X-BYTERANGE', # playlists composed of byte ranges of media files [1] # 1. https://tools.ietf.org/html/draft-pantos-http-live-streaming-17#section-4.3.2.2 ] check_results = (not re.search(feature, manifest) for feature in UNSUPPORTED_FEATURES) return all(check_results) @staticmethod def _aria2c_filename(fn): return fn if os.path.isabs(fn) else f'.{os.path.sep}{fn}' def _call_downloader(self, tmpfilename, info_dict): # FIXME: Disabled due to https://github.com/yt-dlp/yt-dlp/issues/5931 if False and 'no-external-downloader-progress' not in self.params.get('compat_opts', []): info_dict['__rpc'] = { 'port': find_available_port() or 19190, 'secret': str(uuid.uuid4()), } return super()._call_downloader(tmpfilename, info_dict) def _make_cmd(self, tmpfilename, info_dict): cmd = [self.exe, '-c', '--no-conf', '--console-log-level=warn', '--summary-interval=0', '--download-result=hide', '--http-accept-gzip=true', '--file-allocation=none', '-x16', '-j16', '-s16'] if 'fragments' in info_dict: cmd += ['--allow-overwrite=true', '--allow-piece-length-change=true'] else: cmd += ['--min-split-size', '1M'] if self.ydl.cookiejar.get_cookie_header(info_dict['url']): cmd += [f'--load-cookies={self._write_cookies()}'] if info_dict.get('http_headers') is not None: for key, val in info_dict['http_headers'].items(): cmd += ['--header', f'{key}: {val}'] cmd += self._option('--max-overall-download-limit', 'ratelimit') cmd += self._option('--interface', 'source_address') cmd += self._option('--all-proxy', 'proxy') cmd += self._bool_option('--check-certificate', 'nocheckcertificate', 'false', 'true', '=') cmd += self._bool_option('--remote-time', 'updatetime', 'true', 'false', '=') cmd += self._bool_option('--show-console-readout', 'noprogress', 'false', 'true', '=') cmd += self._configuration_args() if '__rpc' in info_dict: cmd += [ '--enable-rpc', f'--rpc-listen-port={info_dict["__rpc"]["port"]}', f'--rpc-secret={info_dict["__rpc"]["secret"]}'] # aria2c strips out spaces from the beginning/end of filenames and paths. # We work around this issue by adding a "./" to the beginning of the # filename and relative path, and adding a "/" at the end of the path. # See: https://github.com/yt-dlp/yt-dlp/issues/276 # https://github.com/ytdl-org/youtube-dl/issues/20312 # https://github.com/aria2/aria2/issues/1373 dn = os.path.dirname(tmpfilename) if dn: cmd += ['--dir', self._aria2c_filename(dn) + os.path.sep] if 'fragments' not in info_dict: cmd += ['--out', self._aria2c_filename(os.path.basename(tmpfilename))] cmd += ['--auto-file-renaming=false'] if 'fragments' in info_dict: cmd += ['--uri-selector=inorder'] url_list_file = f'{tmpfilename}.frag.urls' url_list = [] for frag_index, fragment in enumerate(info_dict['fragments']): fragment_filename = f'{os.path.basename(tmpfilename)}-Frag{frag_index}' url_list.append('{}\n\tout={}'.format(fragment['url'], self._aria2c_filename(fragment_filename))) stream, _ = self.sanitize_open(url_list_file, 'wb') stream.write('\n'.join(url_list).encode()) stream.close() cmd += ['-i', self._aria2c_filename(url_list_file)] else: cmd += ['--', info_dict['url']] return cmd def aria2c_rpc(self, rpc_port, rpc_secret, method, params=()): # Does not actually need to be UUID, just unique sanitycheck = str(uuid.uuid4()) d = json.dumps({ 'jsonrpc': '2.0', 'id': sanitycheck, 'method': method, 'params': [f'token:{rpc_secret}', *params], }).encode() request = Request( f'http://localhost:{rpc_port}/jsonrpc', data=d, headers={ 'Content-Type': 'application/json', 'Content-Length': f'{len(d)}', }, proxies={'all': None}) with self.ydl.urlopen(request) as r: resp = json.load(r) assert resp.get('id') == sanitycheck, 'Something went wrong with RPC server' return resp['result'] def _call_process(self, cmd, info_dict): if '__rpc' not in info_dict: return super()._call_process(cmd, info_dict) send_rpc = functools.partial(self.aria2c_rpc, info_dict['__rpc']['port'], info_dict['__rpc']['secret']) started = time.time() fragmented = 'fragments' in info_dict frag_count = len(info_dict['fragments']) if fragmented else 1 status = { 'filename': info_dict.get('_filename'), 'status': 'downloading', 'elapsed': 0, 'downloaded_bytes': 0, 'fragment_count': frag_count if fragmented else None, 'fragment_index': 0 if fragmented else None, } self._hook_progress(status, info_dict) def get_stat(key, *obj, average=False): val = tuple(filter(None, map(float, traverse_obj(obj, (..., ..., key))))) or [0] return sum(val) / (len(val) if average else 1) with Popen(cmd, text=True, stdout=subprocess.DEVNULL, stderr=subprocess.PIPE) as p: # Add a small sleep so that RPC client can receive response, # or the connection stalls infinitely time.sleep(0.2) retval = p.poll() while retval is None: # We don't use tellStatus as we won't know the GID without reading stdout # Ref: https://aria2.github.io/manual/en/html/aria2c.html#aria2.tellActive active = send_rpc('aria2.tellActive') completed = send_rpc('aria2.tellStopped', [0, frag_count]) downloaded = get_stat('totalLength', completed) + get_stat('completedLength', active) speed = get_stat('downloadSpeed', active) total = frag_count * get_stat('totalLength', active, completed, average=True) if total < downloaded: total = None status.update({ 'downloaded_bytes': int(downloaded), 'speed': speed, 'total_bytes': None if fragmented else total, 'total_bytes_estimate': total, 'eta': (total - downloaded) / (speed or 1), 'fragment_index': min(frag_count, len(completed) + 1) if fragmented else None, 'elapsed': time.time() - started, }) self._hook_progress(status, info_dict) if not active and len(completed) >= frag_count: send_rpc('aria2.shutdown') retval = p.wait() break time.sleep(0.1) retval = p.poll() return '', p.stderr.read(), retval class HttpieFD(ExternalFD): AVAILABLE_OPT = '--version' EXE_NAME = 'http' def _make_cmd(self, tmpfilename, info_dict): cmd = ['http', '--download', '--output', tmpfilename, info_dict['url']] if info_dict.get('http_headers') is not None: for key, val in info_dict['http_headers'].items(): cmd += [f'{key}:{val}'] # httpie 3.1.0+ removes the Cookie header on redirect, so this should be safe for now. [1] # If we ever need cookie handling for redirects, we can export the cookiejar into a session. [2] # 1: https://github.com/httpie/httpie/security/advisories/GHSA-9w4w-cpc8-h2fq # 2: https://httpie.io/docs/cli/sessions cookie_header = self.ydl.cookiejar.get_cookie_header(info_dict['url']) if cookie_header: cmd += [f'Cookie:{cookie_header}'] return cmd class FFmpegFD(ExternalFD): SUPPORTED_PROTOCOLS = ('http', 'https', 'ftp', 'ftps', 'm3u8', 'm3u8_native', 'rtsp', 'rtmp', 'rtmp_ffmpeg', 'mms', 'http_dash_segments') SUPPORTED_FEATURES = (Features.TO_STDOUT, Features.MULTIPLE_FORMATS) @classmethod def available(cls, path=None): # TODO: Fix path for ffmpeg # Fixme: This may be wrong when --ffmpeg-location is used return FFmpegPostProcessor().available def on_process_started(self, proc, stdin): """ Override this in subclasses """ pass @classmethod def can_merge_formats(cls, info_dict, params): return ( info_dict.get('requested_formats') and info_dict.get('protocol') and not params.get('allow_unplayable_formats') and 'no-direct-merge' not in params.get('compat_opts', []) and cls.can_download(info_dict)) def _call_downloader(self, tmpfilename, info_dict): ffpp = FFmpegPostProcessor(downloader=self) if not ffpp.available: self.report_error('m3u8 download detected but ffmpeg could not be found. Please install') return False ffpp.check_version() args = [ffpp.executable, '-y'] for log_level in ('quiet', 'verbose'): if self.params.get(log_level, False): args += ['-loglevel', log_level] break if not self.params.get('verbose'): args += ['-hide_banner'] env = None proxy = self.params.get('proxy') if proxy: if not re.match(r'[\da-zA-Z]+://', proxy): proxy = f'http://{proxy}' if proxy.startswith('socks'): self.report_warning( f'{self.get_basename()} does not support SOCKS proxies. Downloading is likely to fail. ' 'Consider adding --hls-prefer-native to your command.') # Since December 2015 ffmpeg supports -http_proxy option (see # http://git.videolan.org/?p=ffmpeg.git;a=commit;h=b4eb1f29ebddd60c41a2eb39f5af701e38e0d3fd) # We could switch to the following code if we are able to detect version properly # args += ['-http_proxy', proxy] env = os.environ.copy() env['HTTP_PROXY'] = proxy env['http_proxy'] = proxy start_time, end_time = info_dict.get('section_start') or 0, info_dict.get('section_end') fallback_input_args = traverse_obj(info_dict, ('downloader_options', 'ffmpeg_args', ...)) selected_formats = info_dict.get('requested_formats') or [info_dict] for i, fmt in enumerate(selected_formats): is_http = re.match(r'https?://', fmt['url']) cookies = self.ydl.cookiejar.get_cookies_for_url(fmt['url']) if is_http else [] if cookies: args.extend(['-cookies', ''.join( f'{cookie.name}={cookie.value}; path={cookie.path}; domain={cookie.domain};\r\n' for cookie in cookies)]) if fmt.get('http_headers') and is_http: # Trailing \r\n after each HTTP header is important to prevent warning from ffmpeg: # [http @ 00000000003d2fa0] No trailing CRLF found in HTTP header. args.extend(['-headers', ''.join(f'{key}: {val}\r\n' for key, val in fmt['http_headers'].items())]) if start_time: args += ['-ss', str(start_time)] if end_time: args += ['-t', str(end_time - start_time)] protocol = fmt.get('protocol') if protocol == 'rtmp': player_url = fmt.get('player_url') page_url = fmt.get('page_url') app = fmt.get('app') play_path = fmt.get('play_path') tc_url = fmt.get('tc_url') flash_version = fmt.get('flash_version') live = fmt.get('rtmp_live', False) conn = fmt.get('rtmp_conn') if player_url is not None: args += ['-rtmp_swfverify', player_url] if page_url is not None: args += ['-rtmp_pageurl', page_url] if app is not None: args += ['-rtmp_app', app] if play_path is not None: args += ['-rtmp_playpath', play_path] if tc_url is not None: args += ['-rtmp_tcurl', tc_url] if flash_version is not None: args += ['-rtmp_flashver', flash_version] if live: args += ['-rtmp_live', 'live'] if isinstance(conn, list): for entry in conn: args += ['-rtmp_conn', entry] elif isinstance(conn, str): args += ['-rtmp_conn', conn] elif protocol == 'http_dash_segments' and info_dict.get('is_live'): # ffmpeg may try to read past the latest available segments for # live DASH streams unless we pass `-re`. In modern ffmpeg, this # is an alias of `-readrate 1`, but `-readrate` was not added # until ffmpeg 5.0, so we must stick to using `-re` args += ['-re'] url = fmt['url'] if self.params.get('enable_file_urls') and url.startswith('file:'): # The default protocol_whitelist is 'file,crypto,data' when reading local m3u8 URLs, # so only local segments can be read unless we also include 'http,https,tcp,tls' args += ['-protocol_whitelist', 'file,crypto,data,http,https,tcp,tls'] # ffmpeg incorrectly handles 'file:' URLs by only removing the # 'file:' prefix and treating the rest as if it's a normal filepath. # FFmpegPostProcessor also depends on this behavior, so we need to fixup the URLs: # - On Windows/Cygwin, replace 'file:///' and 'file://localhost/' with 'file:' # - On *nix, replace 'file://localhost/' with 'file:/' # Ref: https://github.com/yt-dlp/yt-dlp/issues/13781 # https://trac.ffmpeg.org/ticket/2702 url = re.sub(r'^file://(?:localhost)?/', 'file:' if os.name == 'nt' else 'file:/', url) args += traverse_obj(fmt, ('downloader_options', 'ffmpeg_args', ...)) or fallback_input_args args += [*self._configuration_args((f'_i{i + 1}', '_i')), '-i', url] if not (start_time or end_time) or not self.params.get('force_keyframes_at_cuts'): args += ['-c', 'copy'] if info_dict.get('requested_formats') or protocol == 'http_dash_segments': for i, fmt in enumerate(selected_formats): stream_number = fmt.get('manifest_stream_number', 0) args.extend(['-map', f'{i}:{stream_number}']) if self.params.get('test', False): args += ['-fs', str(self._TEST_FILE_SIZE)] ext = info_dict['ext'] if protocol in ('m3u8', 'm3u8_native'): use_mpegts = (tmpfilename == '-') or self.params.get('hls_use_mpegts') if use_mpegts is None: use_mpegts = info_dict.get('is_live') if use_mpegts: args += ['-f', 'mpegts'] else: args += ['-f', 'mp4'] if (ffpp.basename == 'ffmpeg' and ffpp._features.get('needs_adtstoasc')) and (not info_dict.get('acodec') or info_dict['acodec'].split('.')[0] in ('aac', 'mp4a')): args += ['-bsf:a', 'aac_adtstoasc'] elif protocol == 'rtmp': args += ['-f', 'flv'] elif ext == 'mp4' and tmpfilename == '-': args += ['-f', 'mpegts'] elif ext == 'unknown_video': ext = determine_ext(remove_end(tmpfilename, '.part')) if ext == 'unknown_video': self.report_warning( 'The video format is unknown and cannot be downloaded by ffmpeg. ' 'Explicitly set the extension in the filename to attempt download in that format') else: self.report_warning(f'The video format is unknown. Trying to download as {ext} according to the filename') args += ['-f', EXT_TO_OUT_FORMATS.get(ext, ext)] else: args += ['-f', EXT_TO_OUT_FORMATS.get(ext, ext)] args += traverse_obj(info_dict, ('downloader_options', 'ffmpeg_args_out', ...)) args += self._configuration_args(('_o1', '_o', '')) args = [encodeArgument(opt) for opt in args] args.append(ffpp._ffmpeg_filename_argument(tmpfilename)) self._debug_cmd(args) piped = any(fmt['url'] in ('-', 'pipe:') for fmt in selected_formats) with Popen(args, stdin=subprocess.PIPE, env=env) as proc: if piped: self.on_process_started(proc, proc.stdin) try: retval = proc.wait() except BaseException as e: # subprocces.run would send the SIGKILL signal to ffmpeg and the # mp4 file couldn't be played, but if we ask ffmpeg to quit it # produces a file that is playable (this is mostly useful for live # streams). Note that Windows is not affected and produces playable # files (see https://github.com/ytdl-org/youtube-dl/issues/8300). if isinstance(e, KeyboardInterrupt) and sys.platform != 'win32' and not piped: proc.communicate_or_kill(b'q') else: proc.kill(timeout=None) raise return retval _BY_NAME = { klass.get_basename(): klass for name, klass in globals().items() if name.endswith('FD') and name not in ('ExternalFD', 'FragmentFD') } def list_external_downloaders(): return sorted(_BY_NAME.keys()) def get_external_downloader(external_downloader): """ Given the name of the executable, see whether we support the given downloader """ bn = os.path.splitext(os.path.basename(external_downloader))[0] return _BY_NAME.get(bn) or next(( klass for klass in _BY_NAME.values() if klass.EXE_NAME in bn ), None)
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/downloader/ism.py
yt_dlp/downloader/ism.py
import binascii import io import struct import time from .fragment import FragmentFD from ..networking.exceptions import HTTPError from ..utils import RetryManager u8 = struct.Struct('>B') u88 = struct.Struct('>Bx') u16 = struct.Struct('>H') u1616 = struct.Struct('>Hxx') u32 = struct.Struct('>I') u64 = struct.Struct('>Q') s88 = struct.Struct('>bx') s16 = struct.Struct('>h') s1616 = struct.Struct('>hxx') s32 = struct.Struct('>i') unity_matrix = (s32.pack(0x10000) + s32.pack(0) * 3) * 2 + s32.pack(0x40000000) TRACK_ENABLED = 0x1 TRACK_IN_MOVIE = 0x2 TRACK_IN_PREVIEW = 0x4 SELF_CONTAINED = 0x1 def box(box_type, payload): return u32.pack(8 + len(payload)) + box_type + payload def full_box(box_type, version, flags, payload): return box(box_type, u8.pack(version) + u32.pack(flags)[1:] + payload) def write_piff_header(stream, params): track_id = params['track_id'] fourcc = params['fourcc'] duration = params['duration'] timescale = params.get('timescale', 10000000) language = params.get('language', 'und') height = params.get('height', 0) width = params.get('width', 0) stream_type = params['stream_type'] creation_time = modification_time = int(time.time()) ftyp_payload = b'isml' # major brand ftyp_payload += u32.pack(1) # minor version ftyp_payload += b'piff' + b'iso2' # compatible brands stream.write(box(b'ftyp', ftyp_payload)) # File Type Box mvhd_payload = u64.pack(creation_time) mvhd_payload += u64.pack(modification_time) mvhd_payload += u32.pack(timescale) mvhd_payload += u64.pack(duration) mvhd_payload += s1616.pack(1) # rate mvhd_payload += s88.pack(1) # volume mvhd_payload += u16.pack(0) # reserved mvhd_payload += u32.pack(0) * 2 # reserved mvhd_payload += unity_matrix mvhd_payload += u32.pack(0) * 6 # pre defined mvhd_payload += u32.pack(0xffffffff) # next track id moov_payload = full_box(b'mvhd', 1, 0, mvhd_payload) # Movie Header Box tkhd_payload = u64.pack(creation_time) tkhd_payload += u64.pack(modification_time) tkhd_payload += u32.pack(track_id) # track id tkhd_payload += u32.pack(0) # reserved tkhd_payload += u64.pack(duration) tkhd_payload += u32.pack(0) * 2 # reserved tkhd_payload += s16.pack(0) # layer tkhd_payload += s16.pack(0) # alternate group tkhd_payload += s88.pack(1 if stream_type == 'audio' else 0) # volume tkhd_payload += u16.pack(0) # reserved tkhd_payload += unity_matrix tkhd_payload += u1616.pack(width) tkhd_payload += u1616.pack(height) trak_payload = full_box(b'tkhd', 1, TRACK_ENABLED | TRACK_IN_MOVIE | TRACK_IN_PREVIEW, tkhd_payload) # Track Header Box mdhd_payload = u64.pack(creation_time) mdhd_payload += u64.pack(modification_time) mdhd_payload += u32.pack(timescale) mdhd_payload += u64.pack(duration) mdhd_payload += u16.pack(((ord(language[0]) - 0x60) << 10) | ((ord(language[1]) - 0x60) << 5) | (ord(language[2]) - 0x60)) mdhd_payload += u16.pack(0) # pre defined mdia_payload = full_box(b'mdhd', 1, 0, mdhd_payload) # Media Header Box hdlr_payload = u32.pack(0) # pre defined if stream_type == 'audio': # handler type hdlr_payload += b'soun' hdlr_payload += u32.pack(0) * 3 # reserved hdlr_payload += b'SoundHandler\0' # name elif stream_type == 'video': hdlr_payload += b'vide' hdlr_payload += u32.pack(0) * 3 # reserved hdlr_payload += b'VideoHandler\0' # name elif stream_type == 'text': hdlr_payload += b'subt' hdlr_payload += u32.pack(0) * 3 # reserved hdlr_payload += b'SubtitleHandler\0' # name else: assert False mdia_payload += full_box(b'hdlr', 0, 0, hdlr_payload) # Handler Reference Box if stream_type == 'audio': smhd_payload = s88.pack(0) # balance smhd_payload += u16.pack(0) # reserved media_header_box = full_box(b'smhd', 0, 0, smhd_payload) # Sound Media Header elif stream_type == 'video': vmhd_payload = u16.pack(0) # graphics mode vmhd_payload += u16.pack(0) * 3 # opcolor media_header_box = full_box(b'vmhd', 0, 1, vmhd_payload) # Video Media Header elif stream_type == 'text': media_header_box = full_box(b'sthd', 0, 0, b'') # Subtitle Media Header else: assert False minf_payload = media_header_box dref_payload = u32.pack(1) # entry count dref_payload += full_box(b'url ', 0, SELF_CONTAINED, b'') # Data Entry URL Box dinf_payload = full_box(b'dref', 0, 0, dref_payload) # Data Reference Box minf_payload += box(b'dinf', dinf_payload) # Data Information Box stsd_payload = u32.pack(1) # entry count sample_entry_payload = u8.pack(0) * 6 # reserved sample_entry_payload += u16.pack(1) # data reference index if stream_type == 'audio': sample_entry_payload += u32.pack(0) * 2 # reserved sample_entry_payload += u16.pack(params.get('channels', 2)) sample_entry_payload += u16.pack(params.get('bits_per_sample', 16)) sample_entry_payload += u16.pack(0) # pre defined sample_entry_payload += u16.pack(0) # reserved sample_entry_payload += u1616.pack(params['sampling_rate']) if fourcc == 'AACL': sample_entry_box = box(b'mp4a', sample_entry_payload) if fourcc == 'EC-3': sample_entry_box = box(b'ec-3', sample_entry_payload) elif stream_type == 'video': sample_entry_payload += u16.pack(0) # pre defined sample_entry_payload += u16.pack(0) # reserved sample_entry_payload += u32.pack(0) * 3 # pre defined sample_entry_payload += u16.pack(width) sample_entry_payload += u16.pack(height) sample_entry_payload += u1616.pack(0x48) # horiz resolution 72 dpi sample_entry_payload += u1616.pack(0x48) # vert resolution 72 dpi sample_entry_payload += u32.pack(0) # reserved sample_entry_payload += u16.pack(1) # frame count sample_entry_payload += u8.pack(0) * 32 # compressor name sample_entry_payload += u16.pack(0x18) # depth sample_entry_payload += s16.pack(-1) # pre defined codec_private_data = binascii.unhexlify(params['codec_private_data'].encode()) if fourcc in ('H264', 'AVC1'): sps, pps = codec_private_data.split(u32.pack(1))[1:] avcc_payload = u8.pack(1) # configuration version avcc_payload += sps[1:4] # avc profile indication + profile compatibility + avc level indication avcc_payload += u8.pack(0xfc | (params.get('nal_unit_length_field', 4) - 1)) # complete representation (1) + reserved (11111) + length size minus one avcc_payload += u8.pack(1) # reserved (0) + number of sps (0000001) avcc_payload += u16.pack(len(sps)) avcc_payload += sps avcc_payload += u8.pack(1) # number of pps avcc_payload += u16.pack(len(pps)) avcc_payload += pps sample_entry_payload += box(b'avcC', avcc_payload) # AVC Decoder Configuration Record sample_entry_box = box(b'avc1', sample_entry_payload) # AVC Simple Entry else: assert False elif stream_type == 'text': if fourcc == 'TTML': sample_entry_payload += b'http://www.w3.org/ns/ttml\0' # namespace sample_entry_payload += b'\0' # schema location sample_entry_payload += b'\0' # auxilary mime types(??) sample_entry_box = box(b'stpp', sample_entry_payload) else: assert False else: assert False stsd_payload += sample_entry_box stbl_payload = full_box(b'stsd', 0, 0, stsd_payload) # Sample Description Box stts_payload = u32.pack(0) # entry count stbl_payload += full_box(b'stts', 0, 0, stts_payload) # Decoding Time to Sample Box stsc_payload = u32.pack(0) # entry count stbl_payload += full_box(b'stsc', 0, 0, stsc_payload) # Sample To Chunk Box stco_payload = u32.pack(0) # entry count stbl_payload += full_box(b'stco', 0, 0, stco_payload) # Chunk Offset Box minf_payload += box(b'stbl', stbl_payload) # Sample Table Box mdia_payload += box(b'minf', minf_payload) # Media Information Box trak_payload += box(b'mdia', mdia_payload) # Media Box moov_payload += box(b'trak', trak_payload) # Track Box mehd_payload = u64.pack(duration) mvex_payload = full_box(b'mehd', 1, 0, mehd_payload) # Movie Extends Header Box trex_payload = u32.pack(track_id) # track id trex_payload += u32.pack(1) # default sample description index trex_payload += u32.pack(0) # default sample duration trex_payload += u32.pack(0) # default sample size trex_payload += u32.pack(0) # default sample flags mvex_payload += full_box(b'trex', 0, 0, trex_payload) # Track Extends Box moov_payload += box(b'mvex', mvex_payload) # Movie Extends Box stream.write(box(b'moov', moov_payload)) # Movie Box def extract_box_data(data, box_sequence): data_reader = io.BytesIO(data) while True: box_size = u32.unpack(data_reader.read(4))[0] box_type = data_reader.read(4) if box_type == box_sequence[0]: box_data = data_reader.read(box_size - 8) if len(box_sequence) == 1: return box_data return extract_box_data(box_data, box_sequence[1:]) data_reader.seek(box_size - 8, 1) class IsmFD(FragmentFD): """ Download segments in a ISM manifest """ def real_download(self, filename, info_dict): segments = info_dict['fragments'][:1] if self.params.get( 'test', False) else info_dict['fragments'] ctx = { 'filename': filename, 'total_frags': len(segments), } self._prepare_and_start_frag_download(ctx, info_dict) extra_state = ctx.setdefault('extra_state', { 'ism_track_written': False, }) skip_unavailable_fragments = self.params.get('skip_unavailable_fragments', True) frag_index = 0 for segment in segments: frag_index += 1 if frag_index <= ctx['fragment_index']: continue retry_manager = RetryManager(self.params.get('fragment_retries'), self.report_retry, frag_index=frag_index, fatal=not skip_unavailable_fragments) for retry in retry_manager: try: success = self._download_fragment(ctx, segment['url'], info_dict) if not success: return False frag_content = self._read_fragment(ctx) if not extra_state['ism_track_written']: tfhd_data = extract_box_data(frag_content, [b'moof', b'traf', b'tfhd']) info_dict['_download_params']['track_id'] = u32.unpack(tfhd_data[4:8])[0] write_piff_header(ctx['dest_stream'], info_dict['_download_params']) extra_state['ism_track_written'] = True self._append_fragment(ctx, frag_content) except HTTPError as err: retry.error = err continue if retry_manager.error: if not skip_unavailable_fragments: return False self.report_skip_fragment(frag_index) return self._finish_frag_download(ctx, info_dict)
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/downloader/__init__.py
yt_dlp/downloader/__init__.py
from ..utils import NO_DEFAULT, determine_protocol def get_suitable_downloader(info_dict, params={}, default=NO_DEFAULT, protocol=None, to_stdout=False): info_dict['protocol'] = determine_protocol(info_dict) info_copy = info_dict.copy() info_copy['to_stdout'] = to_stdout protocols = (protocol or info_copy['protocol']).split('+') downloaders = [_get_suitable_downloader(info_copy, proto, params, default) for proto in protocols] if set(downloaders) == {FFmpegFD} and FFmpegFD.can_merge_formats(info_copy, params): return FFmpegFD elif (set(downloaders) == {DashSegmentsFD} and not (to_stdout and len(protocols) > 1) and set(protocols) == {'http_dash_segments_generator'}): return DashSegmentsFD elif len(downloaders) == 1: return downloaders[0] return None # Some of these require get_suitable_downloader from .common import FileDownloader from .dash import DashSegmentsFD from .external import FFmpegFD, get_external_downloader from .f4m import F4mFD from .fc2 import FC2LiveFD from .hls import HlsFD from .http import HttpFD from .ism import IsmFD from .mhtml import MhtmlFD from .niconico import NiconicoLiveFD from .rtmp import RtmpFD from .rtsp import RtspFD from .websocket import WebSocketFragmentFD from .youtube_live_chat import YoutubeLiveChatFD from .bunnycdn import BunnyCdnFD PROTOCOL_MAP = { 'rtmp': RtmpFD, 'rtmpe': RtmpFD, 'rtmp_ffmpeg': FFmpegFD, 'm3u8_native': HlsFD, 'm3u8': FFmpegFD, 'mms': RtspFD, 'rtsp': RtspFD, 'f4m': F4mFD, 'http_dash_segments': DashSegmentsFD, 'http_dash_segments_generator': DashSegmentsFD, 'ism': IsmFD, 'mhtml': MhtmlFD, 'niconico_live': NiconicoLiveFD, 'fc2_live': FC2LiveFD, 'websocket_frag': WebSocketFragmentFD, 'youtube_live_chat': YoutubeLiveChatFD, 'youtube_live_chat_replay': YoutubeLiveChatFD, 'bunnycdn': BunnyCdnFD, } def shorten_protocol_name(proto, simplify=False): short_protocol_names = { 'm3u8_native': 'm3u8', 'm3u8': 'm3u8F', 'rtmp_ffmpeg': 'rtmpF', 'http_dash_segments': 'dash', 'http_dash_segments_generator': 'dashG', 'websocket_frag': 'WSfrag', } if simplify: short_protocol_names.update({ 'https': 'http', 'ftps': 'ftp', 'm3u8': 'm3u8', # Reverse above m3u8 mapping 'm3u8_native': 'm3u8', 'http_dash_segments_generator': 'dash', 'rtmp_ffmpeg': 'rtmp', 'm3u8_frag_urls': 'm3u8', 'dash_frag_urls': 'dash', }) return short_protocol_names.get(proto, proto) def _get_suitable_downloader(info_dict, protocol, params, default): """Get the downloader class that can handle the info dict.""" if default is NO_DEFAULT: default = HttpFD if (info_dict.get('section_start') or info_dict.get('section_end')) and FFmpegFD.can_download(info_dict): return FFmpegFD info_dict['protocol'] = protocol downloaders = params.get('external_downloader') external_downloader = ( downloaders if isinstance(downloaders, str) or downloaders is None else downloaders.get(shorten_protocol_name(protocol, True), downloaders.get('default'))) if external_downloader is None: if info_dict['to_stdout'] and FFmpegFD.can_merge_formats(info_dict, params): return FFmpegFD elif external_downloader.lower() != 'native' and info_dict.get('impersonate') is None: ed = get_external_downloader(external_downloader) if ed.can_download(info_dict, external_downloader): return ed if protocol == 'http_dash_segments': if info_dict.get('is_live') and (external_downloader or '').lower() != 'native': return FFmpegFD if protocol in ('m3u8', 'm3u8_native'): if info_dict.get('is_live'): return FFmpegFD elif (external_downloader or '').lower() == 'native': return HlsFD elif protocol == 'm3u8_native' and get_suitable_downloader( info_dict, params, None, protocol='m3u8_frag_urls', to_stdout=info_dict['to_stdout']): return HlsFD elif params.get('hls_prefer_native') is True: return HlsFD elif params.get('hls_prefer_native') is False: return FFmpegFD return PROTOCOL_MAP.get(protocol, default) __all__ = [ 'FileDownloader', 'get_suitable_downloader', 'shorten_protocol_name', ]
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/downloader/youtube_live_chat.py
yt_dlp/downloader/youtube_live_chat.py
import json import time from .fragment import FragmentFD from ..networking.exceptions import HTTPError from ..utils import ( RegexNotFoundError, RetryManager, dict_get, int_or_none, try_get, ) from ..utils.networking import HTTPHeaderDict class YoutubeLiveChatFD(FragmentFD): """ Downloads YouTube live chats fragment by fragment """ def real_download(self, filename, info_dict): video_id = info_dict['video_id'] self.to_screen(f'[{self.FD_NAME}] Downloading live chat') if not self.params.get('skip_download') and info_dict['protocol'] == 'youtube_live_chat': self.report_warning('Live chat download runs until the livestream ends. ' 'If you wish to download the video simultaneously, run a separate yt-dlp instance') test = self.params.get('test', False) ctx = { 'filename': filename, 'live': True, 'total_frags': None, } from ..extractor.youtube import YoutubeBaseInfoExtractor ie = YoutubeBaseInfoExtractor(self.ydl) start_time = int(time.time() * 1000) def dl_fragment(url, data=None, headers=None): http_headers = HTTPHeaderDict(info_dict.get('http_headers'), headers) return self._download_fragment(ctx, url, info_dict, http_headers, data) def parse_actions_replay(live_chat_continuation): offset = continuation_id = click_tracking_params = None processed_fragment = bytearray() for action in live_chat_continuation.get('actions', []): if 'replayChatItemAction' in action: replay_chat_item_action = action['replayChatItemAction'] offset = int(replay_chat_item_action['videoOffsetTimeMsec']) processed_fragment.extend( json.dumps(action, ensure_ascii=False).encode() + b'\n') if offset is not None: continuation = try_get( live_chat_continuation, lambda x: x['continuations'][0]['liveChatReplayContinuationData'], dict) if continuation: continuation_id = continuation.get('continuation') click_tracking_params = continuation.get('clickTrackingParams') self._append_fragment(ctx, processed_fragment) return continuation_id, offset, click_tracking_params def try_refresh_replay_beginning(live_chat_continuation): # choose the second option that contains the unfiltered live chat replay refresh_continuation = try_get( live_chat_continuation, lambda x: x['header']['liveChatHeaderRenderer']['viewSelector']['sortFilterSubMenuRenderer']['subMenuItems'][1]['continuation']['reloadContinuationData'], dict) if refresh_continuation: # no data yet but required to call _append_fragment self._append_fragment(ctx, b'') refresh_continuation_id = refresh_continuation.get('continuation') offset = 0 click_tracking_params = refresh_continuation.get('trackingParams') return refresh_continuation_id, offset, click_tracking_params return parse_actions_replay(live_chat_continuation) live_offset = 0 def parse_actions_live(live_chat_continuation): nonlocal live_offset continuation_id = click_tracking_params = None processed_fragment = bytearray() for action in live_chat_continuation.get('actions', []): timestamp = self.parse_live_timestamp(action) if timestamp is not None: live_offset = timestamp - start_time # compatibility with replay format pseudo_action = { 'replayChatItemAction': {'actions': [action]}, 'videoOffsetTimeMsec': str(live_offset), 'isLive': True, } processed_fragment.extend( json.dumps(pseudo_action, ensure_ascii=False).encode() + b'\n') continuation_data_getters = [ lambda x: x['continuations'][0]['invalidationContinuationData'], lambda x: x['continuations'][0]['timedContinuationData'], ] continuation_data = try_get(live_chat_continuation, continuation_data_getters, dict) if continuation_data: continuation_id = continuation_data.get('continuation') click_tracking_params = continuation_data.get('clickTrackingParams') timeout_ms = int_or_none(continuation_data.get('timeoutMs')) if timeout_ms is not None: time.sleep(timeout_ms / 1000) self._append_fragment(ctx, processed_fragment) return continuation_id, live_offset, click_tracking_params def download_and_parse_fragment(url, frag_index, request_data=None, headers=None): for retry in RetryManager(self.params.get('fragment_retries'), self.report_retry, frag_index=frag_index): try: success = dl_fragment(url, request_data, headers) if not success: return False, None, None, None raw_fragment = self._read_fragment(ctx) try: data = ie.extract_yt_initial_data(video_id, raw_fragment.decode('utf-8', 'replace')) except RegexNotFoundError: data = None if not data: data = json.loads(raw_fragment) live_chat_continuation = try_get( data, lambda x: x['continuationContents']['liveChatContinuation'], dict) or {} func = ((info_dict['protocol'] == 'youtube_live_chat' and parse_actions_live) or (frag_index == 1 and try_refresh_replay_beginning) or parse_actions_replay) return (True, *func(live_chat_continuation)) except HTTPError as err: retry.error = err continue return False, None, None, None self._prepare_and_start_frag_download(ctx, info_dict) success = dl_fragment(info_dict['url']) if not success: return False raw_fragment = self._read_fragment(ctx) try: data = ie.extract_yt_initial_data(video_id, raw_fragment.decode('utf-8', 'replace')) except RegexNotFoundError: return False continuation_id = try_get( data, lambda x: x['contents']['twoColumnWatchNextResults']['conversationBar']['liveChatRenderer']['continuations'][0]['reloadContinuationData']['continuation']) # no data yet but required to call _append_fragment self._append_fragment(ctx, b'') ytcfg = ie.extract_ytcfg(video_id, raw_fragment.decode('utf-8', 'replace')) if not ytcfg: return False api_key = try_get(ytcfg, lambda x: x['INNERTUBE_API_KEY']) innertube_context = try_get(ytcfg, lambda x: x['INNERTUBE_CONTEXT']) if not api_key or not innertube_context: return False visitor_data = try_get(innertube_context, lambda x: x['client']['visitorData'], str) if info_dict['protocol'] == 'youtube_live_chat_replay': url = 'https://www.youtube.com/youtubei/v1/live_chat/get_live_chat_replay?key=' + api_key chat_page_url = 'https://www.youtube.com/live_chat_replay?continuation=' + continuation_id elif info_dict['protocol'] == 'youtube_live_chat': url = 'https://www.youtube.com/youtubei/v1/live_chat/get_live_chat?key=' + api_key chat_page_url = 'https://www.youtube.com/live_chat?continuation=' + continuation_id frag_index = offset = 0 click_tracking_params = None while continuation_id is not None: frag_index += 1 request_data = { 'context': innertube_context, 'continuation': continuation_id, } if frag_index > 1: request_data['currentPlayerState'] = {'playerOffsetMs': str(max(offset - 5000, 0))} if click_tracking_params: request_data['context']['clickTracking'] = {'clickTrackingParams': click_tracking_params} headers = ie.generate_api_headers(ytcfg=ytcfg, visitor_data=visitor_data) headers.update({'content-type': 'application/json'}) fragment_request_data = json.dumps(request_data, ensure_ascii=False).encode() + b'\n' success, continuation_id, offset, click_tracking_params = download_and_parse_fragment( url, frag_index, fragment_request_data, headers) else: success, continuation_id, offset, click_tracking_params = download_and_parse_fragment( chat_page_url, frag_index) if not success: return False if test: break return self._finish_frag_download(ctx, info_dict) @staticmethod def parse_live_timestamp(action): action_content = dict_get( action, ['addChatItemAction', 'addLiveChatTickerItemAction', 'addBannerToLiveChatCommand']) if not isinstance(action_content, dict): return None item = dict_get(action_content, ['item', 'bannerRenderer']) if not isinstance(item, dict): return None renderer = dict_get(item, [ # text 'liveChatTextMessageRenderer', 'liveChatPaidMessageRenderer', 'liveChatMembershipItemRenderer', 'liveChatPaidStickerRenderer', # ticker 'liveChatTickerPaidMessageItemRenderer', 'liveChatTickerSponsorItemRenderer', # banner 'liveChatBannerRenderer', ]) if not isinstance(renderer, dict): return None parent_item_getters = [ lambda x: x['showItemEndpoint']['showLiveChatItemEndpoint']['renderer'], lambda x: x['contents'], ] parent_item = try_get(renderer, parent_item_getters, dict) if parent_item: renderer = dict_get(parent_item, [ 'liveChatTextMessageRenderer', 'liveChatPaidMessageRenderer', 'liveChatMembershipItemRenderer', 'liveChatPaidStickerRenderer', ]) if not isinstance(renderer, dict): return None return int_or_none(renderer.get('timestampUsec'), 1000)
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/downloader/rtmp.py
yt_dlp/downloader/rtmp.py
import os import re import subprocess import time from .common import FileDownloader from ..utils import ( Popen, check_executable, encodeArgument, get_exe_version, ) def rtmpdump_version(): return get_exe_version( 'rtmpdump', ['--help'], r'(?i)RTMPDump\s*v?([0-9a-zA-Z._-]+)') class RtmpFD(FileDownloader): def real_download(self, filename, info_dict): def run_rtmpdump(args): start = time.time() resume_percent = None resume_downloaded_data_len = None proc = Popen(args, stderr=subprocess.PIPE) cursor_in_new_line = True proc_stderr_closed = False try: while not proc_stderr_closed: # read line from stderr line = '' while True: char = proc.stderr.read(1) if not char: proc_stderr_closed = True break if char in [b'\r', b'\n']: break line += char.decode('ascii', 'replace') if not line: # proc_stderr_closed is True continue mobj = re.search(r'([0-9]+\.[0-9]{3}) kB / [0-9]+\.[0-9]{2} sec \(([0-9]{1,2}\.[0-9])%\)', line) if mobj: downloaded_data_len = int(float(mobj.group(1)) * 1024) percent = float(mobj.group(2)) if not resume_percent: resume_percent = percent resume_downloaded_data_len = downloaded_data_len time_now = time.time() eta = self.calc_eta(start, time_now, 100 - resume_percent, percent - resume_percent) speed = self.calc_speed(start, time_now, downloaded_data_len - resume_downloaded_data_len) data_len = None if percent > 0: data_len = int(downloaded_data_len * 100 / percent) self._hook_progress({ 'status': 'downloading', 'downloaded_bytes': downloaded_data_len, 'total_bytes_estimate': data_len, 'tmpfilename': tmpfilename, 'filename': filename, 'eta': eta, 'elapsed': time_now - start, 'speed': speed, }, info_dict) cursor_in_new_line = False else: # no percent for live streams mobj = re.search(r'([0-9]+\.[0-9]{3}) kB / [0-9]+\.[0-9]{2} sec', line) if mobj: downloaded_data_len = int(float(mobj.group(1)) * 1024) time_now = time.time() speed = self.calc_speed(start, time_now, downloaded_data_len) self._hook_progress({ 'downloaded_bytes': downloaded_data_len, 'tmpfilename': tmpfilename, 'filename': filename, 'status': 'downloading', 'elapsed': time_now - start, 'speed': speed, }, info_dict) cursor_in_new_line = False elif self.params.get('verbose', False): if not cursor_in_new_line: self.to_screen('') cursor_in_new_line = True self.to_screen('[rtmpdump] ' + line) if not cursor_in_new_line: self.to_screen('') return proc.wait() except BaseException: # Including KeyboardInterrupt proc.kill(timeout=None) raise url = info_dict['url'] player_url = info_dict.get('player_url') page_url = info_dict.get('page_url') app = info_dict.get('app') play_path = info_dict.get('play_path') tc_url = info_dict.get('tc_url') flash_version = info_dict.get('flash_version') live = info_dict.get('rtmp_live', False) conn = info_dict.get('rtmp_conn') protocol = info_dict.get('rtmp_protocol') real_time = info_dict.get('rtmp_real_time', False) no_resume = info_dict.get('no_resume', False) continue_dl = self.params.get('continuedl', True) self.report_destination(filename) tmpfilename = self.temp_name(filename) test = self.params.get('test', False) # Check for rtmpdump first if not check_executable('rtmpdump', ['-h']): self.report_error('RTMP download detected but "rtmpdump" could not be run. Please install') return False # Download using rtmpdump. rtmpdump returns exit code 2 when # the connection was interrupted and resuming appears to be # possible. This is part of rtmpdump's normal usage, AFAIK. basic_args = [ 'rtmpdump', '--verbose', '-r', url, '-o', tmpfilename] if player_url is not None: basic_args += ['--swfVfy', player_url] if page_url is not None: basic_args += ['--pageUrl', page_url] if app is not None: basic_args += ['--app', app] if play_path is not None: basic_args += ['--playpath', play_path] if tc_url is not None: basic_args += ['--tcUrl', tc_url] if test: basic_args += ['--stop', '1'] if flash_version is not None: basic_args += ['--flashVer', flash_version] if live: basic_args += ['--live'] if isinstance(conn, list): for entry in conn: basic_args += ['--conn', entry] elif isinstance(conn, str): basic_args += ['--conn', conn] if protocol is not None: basic_args += ['--protocol', protocol] if real_time: basic_args += ['--realtime'] args = basic_args if not no_resume and continue_dl and not live: args += ['--resume'] if not live and continue_dl: args += ['--skip', '1'] args = [encodeArgument(a) for a in args] self._debug_cmd(args, exe='rtmpdump') RD_SUCCESS = 0 RD_FAILED = 1 RD_INCOMPLETE = 2 RD_NO_CONNECT = 3 started = time.time() try: retval = run_rtmpdump(args) except KeyboardInterrupt: if not info_dict.get('is_live'): raise retval = RD_SUCCESS self.to_screen('\n[rtmpdump] Interrupted by user') if retval == RD_NO_CONNECT: self.report_error('[rtmpdump] Could not connect to RTMP server.') return False while retval in (RD_INCOMPLETE, RD_FAILED) and not test and not live: prevsize = os.path.getsize(tmpfilename) self.to_screen(f'[rtmpdump] Downloaded {prevsize} bytes') time.sleep(5.0) # This seems to be needed args = [*basic_args, '--resume'] if retval == RD_FAILED: args += ['--skip', '1'] args = [encodeArgument(a) for a in args] retval = run_rtmpdump(args) cursize = os.path.getsize(tmpfilename) if prevsize == cursize and retval == RD_FAILED: break # Some rtmp streams seem abort after ~ 99.8%. Don't complain for those if prevsize == cursize and retval == RD_INCOMPLETE and cursize > 1024: self.to_screen('[rtmpdump] Could not download the whole video. This can happen for some advertisements.') retval = RD_SUCCESS break if retval == RD_SUCCESS or (test and retval == RD_INCOMPLETE): fsize = os.path.getsize(tmpfilename) self.to_screen(f'[rtmpdump] Downloaded {fsize} bytes') self.try_rename(tmpfilename, filename) self._hook_progress({ 'downloaded_bytes': fsize, 'total_bytes': fsize, 'filename': filename, 'status': 'finished', 'elapsed': time.time() - started, }, info_dict) return True else: self.to_stderr('\n') self.report_error('rtmpdump exited with code %d' % retval) return False
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/__pyinstaller/hook-yt_dlp.py
yt_dlp/__pyinstaller/hook-yt_dlp.py
import sys from PyInstaller.utils.hooks import collect_submodules, collect_data_files def pycryptodome_module(): try: import Cryptodome # noqa: F401 except ImportError: try: import Crypto # noqa: F401 print('WARNING: Using Crypto since Cryptodome is not available. ' 'Install with: python3 -m pip install pycryptodomex', file=sys.stderr) return 'Crypto' except ImportError: pass return 'Cryptodome' def get_hidden_imports(): yield from ('yt_dlp.compat._legacy', 'yt_dlp.compat._deprecated') yield from ('yt_dlp.utils._legacy', 'yt_dlp.utils._deprecated') yield pycryptodome_module() # Only `websockets` is required, others are collected just in case for module in ('websockets', 'requests', 'urllib3'): yield from collect_submodules(module) # These are auto-detected, but explicitly add them just in case yield from ('mutagen', 'brotli', 'certifi', 'secretstorage', 'curl_cffi') hiddenimports = list(get_hidden_imports()) print(f'Adding imports: {hiddenimports}') excludedimports = ['youtube_dl', 'youtube_dlc', 'test', 'ytdlp_plugins', 'devscripts', 'bundle'] datas = collect_data_files('curl_cffi', includes=['cacert.pem']) datas += collect_data_files('yt_dlp_ejs', includes=['**/*.js'])
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
yt-dlp/yt-dlp
https://github.com/yt-dlp/yt-dlp/blob/5a481d65fa99862110bb84d10a2f15f0cb47cab3/yt_dlp/__pyinstaller/__init__.py
yt_dlp/__pyinstaller/__init__.py
import os def get_hook_dirs(): return [os.path.dirname(__file__)]
python
Unlicense
5a481d65fa99862110bb84d10a2f15f0cb47cab3
2026-01-04T14:38:15.430780Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/example_scenes.py
example_scenes.py
from manimlib import * import numpy as np # To watch one of these scenes, run the following: # manimgl example_scenes.py OpeningManimExample # Use -s to skip to the end and just save the final frame # Use -w to write the animation to a file # Use -o to write it to a file and open it once done # Use -n <number> to skip ahead to the n'th animation of a scene. class OpeningManimExample(Scene): def construct(self): intro_words = Text(""" The original motivation for manim was to better illustrate mathematical functions as transformations. """) intro_words.to_edge(UP) self.play(Write(intro_words)) self.wait(2) # Linear transform grid = NumberPlane((-10, 10), (-5, 5)) matrix = [[1, 1], [0, 1]] linear_transform_words = VGroup( Text("This is what the matrix"), IntegerMatrix(matrix), Text("looks like") ) linear_transform_words.arrange(RIGHT) linear_transform_words.to_edge(UP) linear_transform_words.set_backstroke(width=5) self.play( ShowCreation(grid), FadeTransform(intro_words, linear_transform_words) ) self.wait() self.play(grid.animate.apply_matrix(matrix), run_time=3) self.wait() # Complex map c_grid = ComplexPlane() moving_c_grid = c_grid.copy() moving_c_grid.prepare_for_nonlinear_transform() c_grid.set_stroke(BLUE_E, 1) c_grid.add_coordinate_labels(font_size=24) complex_map_words = TexText(""" Or thinking of the plane as $\\mathds{C}$,\\\\ this is the map $z \\rightarrow z^2$ """) complex_map_words.to_corner(UR) complex_map_words.set_backstroke(width=5) self.play( FadeOut(grid), Write(c_grid, run_time=3), FadeIn(moving_c_grid), FadeTransform(linear_transform_words, complex_map_words), ) self.wait() self.play( moving_c_grid.animate.apply_complex_function(lambda z: z**2), run_time=6, ) self.wait(2) class AnimatingMethods(Scene): def construct(self): grid = Tex(R"\pi").get_grid(10, 10, height=4) self.add(grid) # You can animate the application of mobject methods with the # ".animate" syntax: self.play(grid.animate.shift(LEFT)) # Both of those will interpolate between the mobject's initial # state and whatever happens when you apply that method. # For this example, calling grid.shift(LEFT) would shift the # grid one unit to the left, but both of the previous calls to # "self.play" animate that motion. # The same applies for any method, including those setting colors. self.play(grid.animate.set_color(YELLOW)) self.wait() self.play(grid.animate.set_submobject_colors_by_gradient(BLUE, GREEN)) self.wait() self.play(grid.animate.set_height(TAU - MED_SMALL_BUFF)) self.wait() # The method Mobject.apply_complex_function lets you apply arbitrary # complex functions, treating the points defining the mobject as # complex numbers. self.play(grid.animate.apply_complex_function(np.exp), run_time=5) self.wait() # Even more generally, you could apply Mobject.apply_function, # which takes in functions form R^3 to R^3 self.play( grid.animate.apply_function( lambda p: [ p[0] + 0.5 * math.sin(p[1]), p[1] + 0.5 * math.sin(p[0]), p[2] ] ), run_time=5, ) self.wait() class TextExample(Scene): def construct(self): # To run this scene properly, you should have "Consolas" font in your computer # for full usage, you can see https://github.com/3b1b/manim/pull/680 text = Text("Here is a text", font="Consolas", font_size=90) difference = Text( """ The most important difference between Text and TexText is that\n you can change the font more easily, but can't use the LaTeX grammar """, font="Arial", font_size=24, # t2c is a dict that you can choose color for different text t2c={"Text": BLUE, "TexText": BLUE, "LaTeX": ORANGE} ) VGroup(text, difference).arrange(DOWN, buff=1) self.play(Write(text)) self.play(FadeIn(difference, UP)) self.wait(3) fonts = Text( "And you can also set the font according to different words", font="Arial", t2f={"font": "Consolas", "words": "Consolas"}, t2c={"font": BLUE, "words": GREEN} ) fonts.set_width(FRAME_WIDTH - 1) slant = Text( "And the same as slant and weight", font="Consolas", t2s={"slant": ITALIC}, t2w={"weight": BOLD}, t2c={"slant": ORANGE, "weight": RED} ) VGroup(fonts, slant).arrange(DOWN, buff=0.8) self.play(FadeOut(text), FadeOut(difference, shift=DOWN)) self.play(Write(fonts)) self.wait() self.play(Write(slant)) self.wait() class TexTransformExample(Scene): def construct(self): # Tex to color map t2c = { "A": BLUE, "B": TEAL, "C": GREEN, } # Configuration to pass along to each Tex mobject kw = dict(font_size=72, t2c=t2c) lines = VGroup( Tex("A^2 + B^2 = C^2", **kw), Tex("A^2 = C^2 - B^2", **kw), Tex("A^2 = (C + B)(C - B)", **kw), Tex(R"A = \sqrt{(C + B)(C - B)}", **kw), ) lines.arrange(DOWN, buff=LARGE_BUFF) self.add(lines[0]) # The animation TransformMatchingStrings will line up parts # of the source and target which have matching substring strings. # Here, giving it a little path_arc makes each part rotate into # their final positions, which feels appropriate for the idea of # rearranging an equation self.play( TransformMatchingStrings( lines[0].copy(), lines[1], # matched_keys specifies which substring should # line up. If it's not specified, the animation # will align the longest matching substrings. # In this case, the substring "^2 = C^2" would # trip it up matched_keys=["A^2", "B^2", "C^2"], # When you want a substring from the source # to go to a non-equal substring from the target, # use the key map. key_map={"+": "-"}, path_arc=90 * DEG, ), ) self.wait() self.play(TransformMatchingStrings( lines[1].copy(), lines[2], matched_keys=["A^2"] )) self.wait() self.play( TransformMatchingStrings( lines[2].copy(), lines[3], key_map={"2": R"\sqrt"}, path_arc=-30 * DEG, ), ) self.wait(2) self.play(LaggedStartMap(FadeOut, lines, shift=2 * RIGHT)) # TransformMatchingShapes will try to line up all pieces of a # source mobject with those of a target, regardless of the # what Mobject type they are. source = Text("the morse code", height=1) target = Text("here come dots", height=1) saved_source = source.copy() self.play(Write(source)) self.wait() kw = dict(run_time=3, path_arc=PI / 2) self.play(TransformMatchingShapes(source, target, **kw)) self.wait() self.play(TransformMatchingShapes(target, saved_source, **kw)) self.wait() class TexIndexing(Scene): def construct(self): # You can index into Tex mobject (or other StringMobjects) by substrings equation = Tex(R"e^{\pi i} = -1", font_size=144) self.add(equation) self.play(FlashAround(equation["e"])) self.wait() self.play(Indicate(equation[R"\pi"])) self.wait() self.play(TransformFromCopy( equation[R"e^{\pi i}"].copy().set_opacity(0.5), equation["-1"], path_arc=-PI / 2, run_time=3 )) self.play(FadeOut(equation)) # Or regular expressions equation = Tex("A^2 + B^2 = C^2", font_size=144) self.play(Write(equation)) for part in equation[re.compile(r"\w\^2")]: self.play(FlashAround(part)) self.wait() self.play(FadeOut(equation)) # Indexing by substrings like this may not work when # the order in which Latex draws symbols does not match # the order in which they show up in the string. # For example, here the infinity is drawn before the sigma # so we don't get the desired behavior. equation = Tex(R"\sum_{n = 1}^\infty \frac{1}{n^2} = \frac{\pi^2}{6}", font_size=72) self.play(FadeIn(equation)) self.play(equation[R"\infty"].animate.set_color(RED)) # Doesn't hit the infinity self.wait() self.play(FadeOut(equation)) # However you can always fix this by explicitly passing in # a string you might want to isolate later. Also, using # \over instead of \frac helps to avoid the issue for fractions equation = Tex( R"\sum_{n = 1}^\infty {1 \over n^2} = {\pi^2 \over 6}", # Explicitly mark "\infty" as a substring you might want to access isolate=[R"\infty"], font_size=72 ) self.play(FadeIn(equation)) self.play(equation[R"\infty"].animate.set_color(RED)) # Got it! self.wait() self.play(FadeOut(equation)) class UpdatersExample(Scene): def construct(self): square = Square() square.set_fill(BLUE_E, 1) # On all frames, the constructor Brace(square, UP) will # be called, and the mobject brace will set its data to match # that of the newly constructed object brace = always_redraw(Brace, square, UP) label = TexText("Width = 0.00") number = label.make_number_changeable("0.00") # This ensures that the method deicmal.next_to(square) # is called on every frame label.always.next_to(brace, UP) # You could also write the following equivalent line # label.add_updater(lambda m: m.next_to(brace, UP)) # If the argument itself might change, you can use f_always, # for which the arguments following the initial Mobject method # should be functions returning arguments to that method. # The following line ensures thst decimal.set_value(square.get_y()) # is called every frame number.f_always.set_value(square.get_width) # You could also write the following equivalent line # number.add_updater(lambda m: m.set_value(square.get_width())) self.add(square, brace, label) # Notice that the brace and label track with the square self.play( square.animate.scale(2), rate_func=there_and_back, run_time=2, ) self.wait() self.play( square.animate.set_width(5, stretch=True), run_time=3, ) self.wait() self.play( square.animate.set_width(2), run_time=3 ) self.wait() # In general, you can always call Mobject.add_updater, and pass in # a function that you want to be called on every frame. The function # should take in either one argument, the mobject, or two arguments, # the mobject and the amount of time since the last frame. now = self.time w0 = square.get_width() square.add_updater( lambda m: m.set_width(w0 * math.sin(self.time - now) + w0) ) self.wait(4 * PI) class CoordinateSystemExample(Scene): def construct(self): axes = Axes( # x-axis ranges from -1 to 10, with a default step size of 1 x_range=(-1, 10), # y-axis ranges from -2 to 2 with a step size of 0.5 y_range=(-2, 2, 0.5), # The axes will be stretched so as to match the specified # height and width height=6, width=10, # Axes is made of two NumberLine mobjects. You can specify # their configuration with axis_config axis_config=dict( stroke_color=GREY_A, stroke_width=2, numbers_to_exclude=[0], ), # Alternatively, you can specify configuration for just one # of them, like this. y_axis_config=dict( big_tick_numbers=[-2, 2], ) ) # Keyword arguments of add_coordinate_labels can be used to # configure the DecimalNumber mobjects which it creates and # adds to the axes axes.add_coordinate_labels( font_size=20, num_decimal_places=1, ) self.add(axes) # Axes descends from the CoordinateSystem class, meaning # you can call call axes.coords_to_point, abbreviated to # axes.c2p, to associate a set of coordinates with a point, # like so: dot = Dot(color=RED) dot.move_to(axes.c2p(0, 0)) self.play(FadeIn(dot, scale=0.5)) self.play(dot.animate.move_to(axes.c2p(3, 2))) self.wait() self.play(dot.animate.move_to(axes.c2p(5, 0.5))) self.wait() # Similarly, you can call axes.point_to_coords, or axes.p2c # print(axes.p2c(dot.get_center())) # We can draw lines from the axes to better mark the coordinates # of a given point. # Here, the always_redraw command means that on each new frame # the lines will be redrawn h_line = always_redraw(lambda: axes.get_h_line(dot.get_left())) v_line = always_redraw(lambda: axes.get_v_line(dot.get_bottom())) self.play( ShowCreation(h_line), ShowCreation(v_line), ) self.play(dot.animate.move_to(axes.c2p(3, -2))) self.wait() self.play(dot.animate.move_to(axes.c2p(1, 1))) self.wait() # If we tie the dot to a particular set of coordinates, notice # that as we move the axes around it respects the coordinate # system defined by them. f_always(dot.move_to, lambda: axes.c2p(1, 1)) self.play( axes.animate.scale(0.75).to_corner(UL), run_time=2, ) self.wait() self.play(FadeOut(VGroup(axes, dot, h_line, v_line))) # Other coordinate systems you can play around with include # ThreeDAxes, NumberPlane, and ComplexPlane. class GraphExample(Scene): def construct(self): axes = Axes((-3, 10), (-1, 8), height=6) axes.add_coordinate_labels() self.play(Write(axes, lag_ratio=0.01, run_time=1)) # Axes.get_graph will return the graph of a function sin_graph = axes.get_graph( lambda x: 2 * math.sin(x), color=BLUE, ) # By default, it draws it so as to somewhat smoothly interpolate # between sampled points (x, f(x)). If the graph is meant to have # a corner, though, you can set use_smoothing to False relu_graph = axes.get_graph( lambda x: max(x, 0), use_smoothing=False, color=YELLOW, ) # For discontinuous functions, you can specify the point of # discontinuity so that it does not try to draw over the gap. step_graph = axes.get_graph( lambda x: 2.0 if x > 3 else 1.0, discontinuities=[3], color=GREEN, ) # Axes.get_graph_label takes in either a string or a mobject. # If it's a string, it treats it as a LaTeX expression. By default # it places the label next to the graph near the right side, and # has it match the color of the graph sin_label = axes.get_graph_label(sin_graph, "\\sin(x)") relu_label = axes.get_graph_label(relu_graph, Text("ReLU")) step_label = axes.get_graph_label(step_graph, Text("Step"), x=4) self.play( ShowCreation(sin_graph), FadeIn(sin_label, RIGHT), ) self.wait(2) self.play( ReplacementTransform(sin_graph, relu_graph), FadeTransform(sin_label, relu_label), ) self.wait() self.play( ReplacementTransform(relu_graph, step_graph), FadeTransform(relu_label, step_label), ) self.wait() parabola = axes.get_graph(lambda x: 0.25 * x**2) parabola.set_stroke(BLUE) self.play( FadeOut(step_graph), FadeOut(step_label), ShowCreation(parabola) ) self.wait() # You can use axes.input_to_graph_point, abbreviated # to axes.i2gp, to find a particular point on a graph dot = Dot(color=RED) dot.move_to(axes.i2gp(2, parabola)) self.play(FadeIn(dot, scale=0.5)) # A value tracker lets us animate a parameter, usually # with the intent of having other mobjects update based # on the parameter x_tracker = ValueTracker(2) dot.add_updater(lambda d: d.move_to(axes.i2gp(x_tracker.get_value(), parabola))) self.play(x_tracker.animate.set_value(4), run_time=3) self.play(x_tracker.animate.set_value(-2), run_time=3) self.wait() class TexAndNumbersExample(Scene): def construct(self): axes = Axes((-3, 3), (-3, 3), unit_size=1) axes.to_edge(DOWN) axes.add_coordinate_labels(font_size=16) circle = Circle(radius=2) circle.set_stroke(YELLOW, 3) circle.move_to(axes.get_origin()) self.add(axes, circle) # When numbers show up in tex, they can be readily # replaced with DecimalMobjects so that methods like # get_value and set_value can be called on them, and # animations like ChangeDecimalToValue can be called # on them. tex = Tex("x^2 + y^2 = 4.00") tex.next_to(axes, UP, buff=0.5) value = tex.make_number_changeable("4.00") # This will tie the right hand side of our equation to # the square of the radius of the circle value.add_updater(lambda v: v.set_value(circle.get_radius()**2)) self.add(tex) text = Text(""" You can manipulate numbers in Tex mobjects """, font_size=30) text.next_to(tex, RIGHT, buff=1.5) arrow = Arrow(text, tex) self.add(text, arrow) self.play( circle.animate.set_height(2.0), run_time=4, rate_func=there_and_back, ) # By default, tex.make_number_changeable replaces the first occurrence # of the number,but by passing replace_all=True it replaces all and # returns a group of the results exponents = tex.make_number_changeable("2", replace_all=True) self.play( LaggedStartMap( FlashAround, exponents, lag_ratio=0.2, buff=0.1, color=RED ), exponents.animate.set_color(RED) ) def func(x, y): # Switch from manim coords to axes coords xa, ya = axes.point_to_coords(np.array([x, y, 0])) return xa**4 + ya**4 - 4 new_curve = ImplicitFunction(func) new_curve.match_style(circle) circle.rotate(angle_of_vector(new_curve.get_start())) # Align value.clear_updaters() self.play( *(ChangeDecimalToValue(exp, 4) for exp in exponents), ReplacementTransform(circle.copy(), new_curve), circle.animate.set_stroke(width=1, opacity=0.5), ) class SurfaceExample(ThreeDScene): def construct(self): surface_text = Text("For 3d scenes, try using surfaces") surface_text.fix_in_frame() surface_text.to_edge(UP) self.add(surface_text) self.wait(0.1) torus1 = Torus(r1=1, r2=1) torus2 = Torus(r1=3, r2=1) sphere = Sphere(radius=3, resolution=torus1.resolution) # You can texture a surface with up to two images, which will # be interpreted as the side towards the light, and away from # the light. These can be either urls, or paths to a local file # in whatever you've set as the image directory in # the custom_config.yml file # day_texture = "EarthTextureMap" # night_texture = "NightEarthTextureMap" day_texture = "https://upload.wikimedia.org/wikipedia/commons/thumb/4/4d/Whole_world_-_land_and_oceans.jpg/1280px-Whole_world_-_land_and_oceans.jpg" night_texture = "https://upload.wikimedia.org/wikipedia/commons/thumb/b/ba/The_earth_at_night.jpg/1280px-The_earth_at_night.jpg" surfaces = [ TexturedSurface(surface, day_texture, night_texture) for surface in [sphere, torus1, torus2] ] for mob in surfaces: mob.shift(IN) mob.mesh = SurfaceMesh(mob) mob.mesh.set_stroke(BLUE, 1, opacity=0.5) surface = surfaces[0] self.play( FadeIn(surface), ShowCreation(surface.mesh, lag_ratio=0.01, run_time=3), ) for mob in surfaces: mob.add(mob.mesh) surface.save_state() self.play(Rotate(surface, PI / 2), run_time=2) for mob in surfaces[1:]: mob.rotate(PI / 2) self.play( Transform(surface, surfaces[1]), run_time=3 ) self.play( Transform(surface, surfaces[2]), # Move camera frame during the transition self.frame.animate.increment_phi(-10 * DEG), self.frame.animate.increment_theta(-20 * DEG), run_time=3 ) # Add ambient rotation self.frame.add_updater(lambda m, dt: m.increment_theta(-0.1 * dt)) # Play around with where the light is light_text = Text("You can move around the light source") light_text.move_to(surface_text) light_text.fix_in_frame() self.play(FadeTransform(surface_text, light_text)) light = self.camera.light_source light_dot = GlowDot(color=WHITE, radius=0.5) light_dot.always.move_to(light) self.add(light, light_dot) light.save_state() self.play(light.animate.move_to(3 * IN), run_time=5) self.play(light.animate.shift(10 * OUT), run_time=5) drag_text = Text("Try moving the mouse while pressing d or f") drag_text.move_to(light_text) drag_text.fix_in_frame() self.play(FadeTransform(light_text, drag_text)) self.wait() class InteractiveDevelopment(Scene): def construct(self): circle = Circle() circle.set_fill(BLUE, opacity=0.5) circle.set_stroke(BLUE_E, width=4) square = Square() self.play(ShowCreation(square)) self.wait() # This opens an iPython terminal where you can keep writing # lines as if they were part of this construct method. # In particular, 'square', 'circle' and 'self' will all be # part of the local namespace in that terminal. self.embed() # Try copying and pasting some of the lines below into # the interactive shell self.play(ReplacementTransform(square, circle)) self.wait() self.play(circle.animate.stretch(4, 0)) self.play(Rotate(circle, 90 * DEG)) self.play(circle.animate.shift(2 * RIGHT).scale(0.25)) text = Text(""" In general, using the interactive shell is very helpful when developing new scenes """) self.play(Write(text)) # In the interactive shell, you can just type # play, add, remove, clear, wait, save_state and restore, # instead of self.play, self.add, self.remove, etc. # To interact with the window, type touch(). You can then # scroll in the window, or zoom by holding down 'z' while scrolling, # and change camera perspective by holding down 'd' while moving # the mouse. Press 'r' to reset to the standard camera position. # Press 'q' to stop interacting with the window and go back to # typing new commands into the shell. # In principle you can customize a scene to be responsive to # mouse and keyboard interactions always(circle.move_to, self.mouse_point) class ControlsExample(Scene): drag_to_pan = False def setup(self): self.textbox = Textbox() self.checkbox = Checkbox() self.color_picker = ColorSliders() self.panel = ControlPanel( Text("Text", font_size=24), self.textbox, Line(), Text("Show/Hide Text", font_size=24), self.checkbox, Line(), Text("Color of Text", font_size=24), self.color_picker ) self.add(self.panel) def construct(self): text = Text("text", font_size=96) def text_updater(old_text): assert(isinstance(old_text, Text)) new_text = Text(self.textbox.get_value(), font_size=old_text.font_size) # new_text.align_data_and_family(old_text) new_text.move_to(old_text) if self.checkbox.get_value(): new_text.set_fill( color=self.color_picker.get_picked_color(), opacity=self.color_picker.get_picked_opacity() ) else: new_text.set_opacity(0) old_text.become(new_text) text.add_updater(text_updater) self.add(MotionMobject(text)) self.textbox.set_value("Manim") # self.wait(60) # self.embed() # See https://github.com/3b1b/videos for many, many more
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/setup.py
setup.py
import setuptools setuptools.setup()
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/typing.py
manimlib/typing.py
from typing import TYPE_CHECKING if TYPE_CHECKING: from typing import Union, Tuple, Annotated, Literal, Iterable, Dict from colour import Color import numpy as np import re try: from typing_extensions import Self except ImportError: from typing import Self # Abbreviations for a common types ManimColor = Union[str, Color, None] RangeSpecifier = Tuple[float, float, float] | Tuple[float, float] Span = tuple[int, int] SingleSelector = Union[ str, re.Pattern, tuple[Union[int, None], Union[int, None]], ] Selector = Union[SingleSelector, Iterable[SingleSelector]] UniformDict = Dict[str, float | bool | np.ndarray | tuple] # These are various alternate names for np.ndarray meant to specify # certain shapes. # # In theory, these annotations could be used to check arrays sizes # at runtime, but at the moment nothing actually uses them, and # the names are here primarily to enhance readibility and allow # for some stronger type checking if numpy has stronger typing # in the future FloatArray = np.ndarray[int, np.dtype[np.float64]] Vect2 = Annotated[FloatArray, Literal[2]] Vect3 = Annotated[FloatArray, Literal[3]] Vect4 = Annotated[FloatArray, Literal[4]] VectN = Annotated[FloatArray, Literal["N"]] Matrix3x3 = Annotated[FloatArray, Literal[3, 3]] VectArray = Annotated[FloatArray, Literal["N", 1]] Vect2Array = Annotated[FloatArray, Literal["N", 2]] Vect3Array = Annotated[FloatArray, Literal["N", 3]] Vect4Array = Annotated[FloatArray, Literal["N", 4]] VectNArray = Annotated[FloatArray, Literal["N", "M"]]
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/extract_scene.py
manimlib/extract_scene.py
from __future__ import annotations import copy import inspect import sys from manimlib.module_loader import ModuleLoader from manimlib.config import manim_config from manimlib.logger import log from manimlib.scene.interactive_scene import InteractiveScene from manimlib.scene.scene import Scene from typing import TYPE_CHECKING if TYPE_CHECKING: Module = importlib.util.types.ModuleType from typing import Optional from addict import Dict class BlankScene(InteractiveScene): def construct(self): exec(manim_config.universal_import_line) self.embed() def is_child_scene(obj, module): if not inspect.isclass(obj): return False if not issubclass(obj, Scene): return False if obj == Scene: return False if not obj.__module__.startswith(module.__name__): return False return True def prompt_user_for_choice(scene_classes): name_to_class = {} max_digits = len(str(len(scene_classes))) for idx, scene_class in enumerate(scene_classes, start=1): name = scene_class.__name__ print(f"{str(idx).zfill(max_digits)}: {name}") name_to_class[name] = scene_class try: user_input = input("\nSelect which scene to render (by name or number): ") return [ name_to_class[split_str] if not split_str.isnumeric() else scene_classes[int(split_str) - 1] for split_str in user_input.replace(" ", "").split(",") ] except IndexError: log.error("Invalid scene number") sys.exit(2) except KeyError: log.error("Invalid scene name") sys.exit(2) except EOFError: sys.exit(1) def compute_total_frames(scene_class, scene_config): """ When a scene is being written to file, a copy of the scene is run with skip_animations set to true so as to count how many frames it will require. This allows for a total progress bar on rendering, and also allows runtime errors to be exposed preemptively for long running scenes. """ pre_config = copy.deepcopy(scene_config) pre_config["file_writer_config"]["write_to_movie"] = False pre_config["file_writer_config"]["save_last_frame"] = False pre_config["file_writer_config"]["quiet"] = True pre_config["skip_animations"] = True pre_scene = scene_class(**pre_config) pre_scene.run() total_time = pre_scene.time - pre_scene.skip_time return int(total_time * manim_config.camera.fps) def scene_from_class(scene_class, scene_config: Dict, run_config: Dict): fw_config = manim_config.file_writer if fw_config.write_to_movie and run_config.prerun: scene_config.file_writer_config.total_frames = compute_total_frames(scene_class, scene_config) return scene_class(**scene_config) def note_missing_scenes(arg_names, module_names): for name in arg_names: if name not in module_names: log.error(f"No scene named {name} found") def get_scenes_to_render(all_scene_classes: list, scene_config: Dict, run_config: Dict): if run_config["write_all"] or len(all_scene_classes) == 1: classes_to_run = all_scene_classes else: name_to_class = {sc.__name__: sc for sc in all_scene_classes} classes_to_run = [name_to_class.get(name) for name in run_config.scene_names] classes_to_run = list(filter(lambda x: x, classes_to_run)) # Remove Nones note_missing_scenes(run_config.scene_names, name_to_class.keys()) if len(classes_to_run) == 0: classes_to_run = prompt_user_for_choice(all_scene_classes) return [ scene_from_class(scene_class, scene_config, run_config) for scene_class in classes_to_run ] def get_scene_classes(module: Optional[Module]): if module is None: # If no module was passed in, just play the blank scene return [BlankScene] if hasattr(module, "SCENES_IN_ORDER"): return module.SCENES_IN_ORDER else: return [ member[1] for member in inspect.getmembers( module, lambda x: is_child_scene(x, module) ) ] def get_indent(code_lines: list[str], line_number: int) -> str: """ Find the indent associated with a given line of python code, as a string of spaces """ # Find most recent non-empty line try: line = next(filter(lambda line: line.strip(), code_lines[line_number - 1::-1])) except StopIteration: return "" # Either return its leading spaces, or add for if it ends with colon n_spaces = len(line) - len(line.lstrip()) if line.endswith(":"): n_spaces += 4 return n_spaces * " " def insert_embed_line_to_module(module: Module, run_config: Dict) -> None: """ This is hacky, but convenient. When user includes the argument "-e", it will try to recreate a file that inserts the line `self.embed()` into the end of the scene's construct method. If there is an argument passed in, it will insert the line after the last line in the sourcefile which includes that string. """ lines = inspect.getsource(module).splitlines() line_number = run_config.embed_line # Add the relevant embed line to the code indent = get_indent(lines, line_number) lines.insert(line_number, indent + "self.embed()") new_code = "\n".join(lines) # When the user executes the `-e <line_number>` command # without specifying scene_names, the nearest class name above # `<line_number>` will be automatically used as 'scene_names'. if not run_config.scene_names: classes = list(filter(lambda line: line.startswith("class"), lines[:line_number])) if classes: from re import search scene_name = search(r"(\w+)\(", classes[-1]) run_config.update(scene_names=[scene_name.group(1)]) else: log.error(f"No 'class' found above {line_number}!") # Execute the code, which presumably redefines the user's # scene to include this embed line, within the relevant module. code_object = compile(new_code, module.__name__, 'exec') exec(code_object, module.__dict__) def get_module(run_config: Dict) -> Module: module = ModuleLoader.get_module(run_config.file_name, run_config.is_reload) if run_config.embed_line: insert_embed_line_to_module(module, run_config) return module def main(scene_config: Dict, run_config: Dict): module = get_module(run_config) all_scene_classes = get_scene_classes(module) scenes = get_scenes_to_render(all_scene_classes, scene_config, run_config) if len(scenes) == 0: print("No scenes found to run") return scenes
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/constants.py
manimlib/constants.py
from __future__ import annotations import numpy as np from typing import TYPE_CHECKING if TYPE_CHECKING: from typing import List from manimlib.typing import ManimColor, Vect3 # See manimlib/default_config.yml from manimlib.config import manim_config DEFAULT_RESOLUTION: tuple[int, int] = manim_config.camera.resolution DEFAULT_PIXEL_WIDTH: int = DEFAULT_RESOLUTION[0] DEFAULT_PIXEL_HEIGHT: int = DEFAULT_RESOLUTION[1] # Sizes relevant to default camera frame ASPECT_RATIO: float = DEFAULT_PIXEL_WIDTH / DEFAULT_PIXEL_HEIGHT FRAME_HEIGHT: float = manim_config.sizes.frame_height FRAME_WIDTH: float = FRAME_HEIGHT * ASPECT_RATIO FRAME_SHAPE: tuple[float, float] = (FRAME_WIDTH, FRAME_HEIGHT) FRAME_Y_RADIUS: float = FRAME_HEIGHT / 2 FRAME_X_RADIUS: float = FRAME_WIDTH / 2 # Helpful values for positioning mobjects SMALL_BUFF: float = manim_config.sizes.small_buff MED_SMALL_BUFF: float = manim_config.sizes.med_small_buff MED_LARGE_BUFF: float = manim_config.sizes.med_large_buff LARGE_BUFF: float = manim_config.sizes.large_buff DEFAULT_MOBJECT_TO_EDGE_BUFF: float = manim_config.sizes.default_mobject_to_edge_buff DEFAULT_MOBJECT_TO_MOBJECT_BUFF: float = manim_config.sizes.default_mobject_to_mobject_buff # Standard vectors ORIGIN: Vect3 = np.array([0., 0., 0.]) UP: Vect3 = np.array([0., 1., 0.]) DOWN: Vect3 = np.array([0., -1., 0.]) RIGHT: Vect3 = np.array([1., 0., 0.]) LEFT: Vect3 = np.array([-1., 0., 0.]) IN: Vect3 = np.array([0., 0., -1.]) OUT: Vect3 = np.array([0., 0., 1.]) X_AXIS: Vect3 = np.array([1., 0., 0.]) Y_AXIS: Vect3 = np.array([0., 1., 0.]) Z_AXIS: Vect3 = np.array([0., 0., 1.]) NULL_POINTS = np.array([[0., 0., 0.]]) # Useful abbreviations for diagonals UL: Vect3 = UP + LEFT UR: Vect3 = UP + RIGHT DL: Vect3 = DOWN + LEFT DR: Vect3 = DOWN + RIGHT TOP: Vect3 = FRAME_Y_RADIUS * UP BOTTOM: Vect3 = FRAME_Y_RADIUS * DOWN LEFT_SIDE: Vect3 = FRAME_X_RADIUS * LEFT RIGHT_SIDE: Vect3 = FRAME_X_RADIUS * RIGHT # Angles PI: float = np.pi TAU: float = 2 * PI DEG: float = TAU / 360 DEGREES = DEG # Many older animations use the full name # Nice to have a constant for readability # when juxtaposed with expressions like 30 * DEG RADIANS: float = 1 # Related to Text NORMAL: str = "NORMAL" ITALIC: str = "ITALIC" OBLIQUE: str = "OBLIQUE" BOLD: str = "BOLD" DEFAULT_STROKE_WIDTH: float = manim_config.vmobject.default_stroke_width # Colors BLUE_E: ManimColor = manim_config.colors.blue_e BLUE_D: ManimColor = manim_config.colors.blue_d BLUE_C: ManimColor = manim_config.colors.blue_c BLUE_B: ManimColor = manim_config.colors.blue_b BLUE_A: ManimColor = manim_config.colors.blue_a TEAL_E: ManimColor = manim_config.colors.teal_e TEAL_D: ManimColor = manim_config.colors.teal_d TEAL_C: ManimColor = manim_config.colors.teal_c TEAL_B: ManimColor = manim_config.colors.teal_b TEAL_A: ManimColor = manim_config.colors.teal_a GREEN_E: ManimColor = manim_config.colors.green_e GREEN_D: ManimColor = manim_config.colors.green_d GREEN_C: ManimColor = manim_config.colors.green_c GREEN_B: ManimColor = manim_config.colors.green_b GREEN_A: ManimColor = manim_config.colors.green_a YELLOW_E: ManimColor = manim_config.colors.yellow_e YELLOW_D: ManimColor = manim_config.colors.yellow_d YELLOW_C: ManimColor = manim_config.colors.yellow_c YELLOW_B: ManimColor = manim_config.colors.yellow_b YELLOW_A: ManimColor = manim_config.colors.yellow_a GOLD_E: ManimColor = manim_config.colors.gold_e GOLD_D: ManimColor = manim_config.colors.gold_d GOLD_C: ManimColor = manim_config.colors.gold_c GOLD_B: ManimColor = manim_config.colors.gold_b GOLD_A: ManimColor = manim_config.colors.gold_a RED_E: ManimColor = manim_config.colors.red_e RED_D: ManimColor = manim_config.colors.red_d RED_C: ManimColor = manim_config.colors.red_c RED_B: ManimColor = manim_config.colors.red_b RED_A: ManimColor = manim_config.colors.red_a MAROON_E: ManimColor = manim_config.colors.maroon_e MAROON_D: ManimColor = manim_config.colors.maroon_d MAROON_C: ManimColor = manim_config.colors.maroon_c MAROON_B: ManimColor = manim_config.colors.maroon_b MAROON_A: ManimColor = manim_config.colors.maroon_a PURPLE_E: ManimColor = manim_config.colors.purple_e PURPLE_D: ManimColor = manim_config.colors.purple_d PURPLE_C: ManimColor = manim_config.colors.purple_c PURPLE_B: ManimColor = manim_config.colors.purple_b PURPLE_A: ManimColor = manim_config.colors.purple_a GREY_E: ManimColor = manim_config.colors.grey_e GREY_D: ManimColor = manim_config.colors.grey_d GREY_C: ManimColor = manim_config.colors.grey_c GREY_B: ManimColor = manim_config.colors.grey_b GREY_A: ManimColor = manim_config.colors.grey_a WHITE: ManimColor = manim_config.colors.white BLACK: ManimColor = manim_config.colors.black GREY_BROWN: ManimColor = manim_config.colors.grey_brown DARK_BROWN: ManimColor = manim_config.colors.dark_brown LIGHT_BROWN: ManimColor = manim_config.colors.light_brown PINK: ManimColor = manim_config.colors.pink LIGHT_PINK: ManimColor = manim_config.colors.light_pink GREEN_SCREEN: ManimColor = manim_config.colors.green_screen ORANGE: ManimColor = manim_config.colors.orange PURE_RED: ManimColor = manim_config.colors.pure_red PURE_GREEN: ManimColor = manim_config.colors.pure_green PURE_BLUE: ManimColor = manim_config.colors.pure_blue MANIM_COLORS: List[ManimColor] = list(manim_config.colors.values()) # Abbreviated names for the "median" colors BLUE: ManimColor = BLUE_C TEAL: ManimColor = TEAL_C GREEN: ManimColor = GREEN_C YELLOW: ManimColor = YELLOW_C GOLD: ManimColor = GOLD_C RED: ManimColor = RED_C MAROON: ManimColor = MAROON_C PURPLE: ManimColor = PURPLE_C GREY: ManimColor = GREY_C COLORMAP_3B1B: List[ManimColor] = [BLUE_E, GREEN, YELLOW, RED] # Default mobject colors should be configurable just like background color # DEFAULT_MOBJECT_COLOR is mainly for text, tex, line, etc... mobjects. Default is WHITE # DEFAULT_LIGHT_COLOR is mainly for things like axes, arrows, annulus and other lightly colored mobjects. Default is GREY_B DEFAULT_MOBJECT_COLOR: ManimColor = manim_config.mobject.default_mobject_color or WHITE DEFAULT_LIGHT_COLOR: ManimColor = manim_config.mobject.default_light_color or GREY_B DEFAULT_VMOBJECT_STROKE_COLOR : ManimColor = manim_config.vmobject.default_stroke_color or GREY_A DEFAULT_VMOBJECT_FILL_COLOR : ManimColor = manim_config.vmobject.default_fill_color or GREY_C
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/logger.py
manimlib/logger.py
import logging from rich.logging import RichHandler __all__ = ["log"] FORMAT = "%(message)s" logging.basicConfig( level=logging.WARNING, format=FORMAT, datefmt="[%X]", handlers=[RichHandler()] ) log = logging.getLogger("manimgl")
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/__main__.py
manimlib/__main__.py
#!/usr/bin/env python from addict import Dict from manimlib import __version__ from manimlib.config import manim_config from manimlib.config import parse_cli import manimlib.extract_scene from manimlib.utils.cache import clear_cache from manimlib.window import Window from IPython.terminal.embed import KillEmbedded from typing import TYPE_CHECKING if TYPE_CHECKING: from argparse import Namespace def run_scenes(): """ Runs the scenes in a loop and detects when a scene reload is requested. """ # Create a new dict to be able to upate without # altering global configuration scene_config = Dict(manim_config.scene) run_config = manim_config.run if run_config.show_in_window: # Create a reusable window window = Window(**manim_config.window) scene_config.update(window=window) while True: try: # Blocking call since a scene may init an IPython shell() scenes = manimlib.extract_scene.main(scene_config, run_config) for scene in scenes: scene.run() return except KillEmbedded: # Requested via the `exit_raise` IPython runline magic # by means of the reload_scene() command pass except KeyboardInterrupt: break def main(): """ Main entry point for ManimGL. """ print(f"ManimGL \033[32mv{__version__}\033[0m") args = parse_cli() if args.version and args.file is None: return if args.clear_cache: clear_cache() run_scenes() if __name__ == "__main__": main()
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/module_loader.py
manimlib/module_loader.py
from __future__ import annotations import builtins import importlib import os import sys import sysconfig from manimlib.config import manim_config from manimlib.logger import log Module = importlib.util.types.ModuleType class ModuleLoader: """ Utility class to load a module from a file and handle its imports. Most parts of this class are only needed for the reload functionality, while the `get_module` method is the main entry point to import a module. """ @staticmethod def get_module(file_name: str | None, is_during_reload=False) -> Module | None: """ Imports a module from a file and returns it. During reload (when the user calls `reload()` in the IPython shell), we also track the imported modules and reload them as well (they would be cached otherwise). See the reload_manager where the reload parameter is set. Note that `exec_module()` is called twice when reloading a module: 1. In exec_module_and_track_imports to track the imports 2. Here to actually execute the module again with the respective imported modules reloaded. """ if file_name is None: return None module_name = file_name.replace(os.sep, ".").replace(".py", "") spec = importlib.util.spec_from_file_location(module_name, file_name) module = importlib.util.module_from_spec(spec) if is_during_reload: imported_modules = ModuleLoader._exec_module_and_track_imports(spec, module) reloaded_modules_tracker = set() ModuleLoader._reload_modules(imported_modules, reloaded_modules_tracker) spec.loader.exec_module(module) return module @staticmethod def _exec_module_and_track_imports(spec, module: Module) -> set[str]: """ Executes the given module (imports it) and returns all the modules that are imported during its execution. This is achieved by replacing the __import__ function with a custom one that tracks the imported modules. At the end, the original __import__ built-in function is restored. """ imported_modules: set[str] = set() original_import = builtins.__import__ def tracked_import(name, globals=None, locals=None, fromlist=(), level=0): """ Custom __import__ function that does exactly the same as the original one, but also tracks the imported modules by means of adding their names to a set. """ result = original_import(name, globals, locals, fromlist, level) imported_modules.add(name) return result builtins.__import__ = tracked_import try: module_name = module.__name__ log.debug('Reloading module "%s"', module_name) spec.loader.exec_module(module) finally: builtins.__import__ = original_import return imported_modules @staticmethod def _reload_modules(modules: set[str], reloaded_modules_tracker: set[str]): """ Out of the given modules, reloads the ones that were not already imported. We skip modules that are not user-defined (see `is_user_defined_module()`). """ for mod in modules: if mod in reloaded_modules_tracker: continue if not ModuleLoader._is_user_defined_module(mod): continue module = sys.modules[mod] ModuleLoader._deep_reload(module, reloaded_modules_tracker) reloaded_modules_tracker.add(mod) @staticmethod def _is_user_defined_module(mod: str) -> bool: """ Returns whether the given module is user-defined or not. A module is considered user-defined if - it is not part of the standard library - AND it is not an external library (site-packages or dist-packages) """ if mod not in sys.modules: return False if mod in sys.builtin_module_names: return False module = sys.modules[mod] module_path = getattr(module, "__file__", None) if module_path is None: return False module_path = os.path.abspath(module_path) # External libraries (site-packages or dist-packages), e.g. numpy if "site-packages" in module_path or "dist-packages" in module_path: return False # Standard lib standard_lib_path = sysconfig.get_path("stdlib") if module_path.startswith(standard_lib_path): return False return True @staticmethod def _deep_reload(module: Module, reloaded_modules_tracker: set[str]): """ Recursively reloads modules imported by the given module. Only user-defined modules are reloaded, see `is_user_defined_module()`. """ ignore_manimlib_modules = manim_config.ignore_manimlib_modules_on_reload if ignore_manimlib_modules and module.__name__.startswith("manimlib"): return if module.__name__.startswith("manimlib.config"): # We don't want to reload global manim_config return if not hasattr(module, "__dict__"): return # Prevent reloading the same module multiple times if module.__name__ in reloaded_modules_tracker: return reloaded_modules_tracker.add(module.__name__) # Recurse for all imported modules for _attr_name, attr_value in module.__dict__.items(): if isinstance(attr_value, Module): if ModuleLoader._is_user_defined_module(attr_value.__name__): ModuleLoader._deep_reload(attr_value, reloaded_modules_tracker) # Also reload modules that are part of a class or function # e.g. when importing `from custom_module import CustomClass` elif hasattr(attr_value, "__module__"): attr_module_name = attr_value.__module__ if ModuleLoader._is_user_defined_module(attr_module_name): attr_module = sys.modules[attr_module_name] ModuleLoader._deep_reload(attr_module, reloaded_modules_tracker) # Reload log.debug('Reloading module "%s"', module.__name__) importlib.reload(module)
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/config.py
manimlib/config.py
from __future__ import annotations import argparse import colour import importlib import inspect import os import sys import yaml from pathlib import Path from ast import literal_eval from addict import Dict from manimlib.logger import log from manimlib.utils.dict_ops import merge_dicts_recursively from typing import TYPE_CHECKING if TYPE_CHECKING: from argparse import Namespace from typing import Optional def initialize_manim_config() -> Dict: """ Return default configuration for various classes in manim, such as Scene, Window, Camera, and SceneFileWriter, as well as configuration determining how the scene is run (e.g. written to file or previewed in window). The result is initially on the contents of default_config.yml in the manimlib directory, which can be further updated by a custom configuration file custom_config.yml. It is further updated based on command line argument. """ args = parse_cli() global_defaults_file = os.path.join(get_manim_dir(), "manimlib", "default_config.yml") config = Dict(merge_dicts_recursively( load_yaml(global_defaults_file), load_yaml("custom_config.yml"), # From current working directory load_yaml(args.config_file) if args.config_file else dict(), )) log.setLevel(args.log_level or config["log_level"]) update_directory_config(config) update_window_config(config, args) update_camera_config(config, args) update_file_writer_config(config, args) update_scene_config(config, args) update_run_config(config, args) update_embed_config(config, args) return config def parse_cli(): try: parser = argparse.ArgumentParser() module_location = parser.add_mutually_exclusive_group() module_location.add_argument( "file", nargs="?", help="Path to file holding the python code for the scene", ) parser.add_argument( "scene_names", nargs="*", help="Name of the Scene class you want to see", ) parser.add_argument( "-w", "--write_file", action="store_true", help="Render the scene as a movie file", ) parser.add_argument( "-s", "--skip_animations", action="store_true", help="Save the last frame", ) parser.add_argument( "-l", "--low_quality", action="store_true", help="Render at 480p", ) parser.add_argument( "-m", "--medium_quality", action="store_true", help="Render at 720p", ) parser.add_argument( "--hd", action="store_true", help="Render at a 1080p", ) parser.add_argument( "--uhd", action="store_true", help="Render at a 4k", ) parser.add_argument( "-f", "--full_screen", action="store_true", help="Show window in full screen", ) parser.add_argument( "-p", "--presenter_mode", action="store_true", help="Scene will stay paused during wait calls until " + \ "space bar or right arrow is hit, like a slide show" ) parser.add_argument( "-i", "--gif", action="store_true", help="Save the video as gif", ) parser.add_argument( "-t", "--transparent", action="store_true", help="Render to a movie file with an alpha channel", ) parser.add_argument( "--vcodec", help="Video codec to use with ffmpeg", ) parser.add_argument( "--pix_fmt", help="Pixel format to use for the output of ffmpeg, defaults to `yuv420p`", ) parser.add_argument( "-q", "--quiet", action="store_true", help="", ) parser.add_argument( "-a", "--write_all", action="store_true", help="Write all the scenes from a file", ) parser.add_argument( "-o", "--open", action="store_true", help="Automatically open the saved file once its done", ) parser.add_argument( "--finder", action="store_true", help="Show the output file in finder", ) parser.add_argument( "--subdivide", action="store_true", help="Divide the output animation into individual movie files " + "for each animation", ) parser.add_argument( "--file_name", help="Name for the movie or image file", ) parser.add_argument( "-n", "--start_at_animation_number", help="Start rendering not from the first animation, but " + \ "from another, specified by its index. If you pass " + \ "in two comma separated values, e.g. \"3,6\", it will end " + \ "the rendering at the second value", ) parser.add_argument( "-e", "--embed", metavar="LINE_NUMBER", help="Adds a breakpoint at the inputted file dropping into an " + \ "interactive iPython session at that point of the code." ) parser.add_argument( "-r", "--resolution", help="Resolution, passed as \"WxH\", e.g. \"1920x1080\"", ) parser.add_argument( "--fps", help="Frame rate, as an integer", type=int, ) parser.add_argument( "-c", "--color", help="Background color", ) parser.add_argument( "--leave_progress_bars", action="store_true", help="Leave progress bars displayed in terminal", ) parser.add_argument( "--show_animation_progress", action="store_true", help="Show progress bar for each animation", ) parser.add_argument( "--prerun", action="store_true", help="Calculate total framecount, to display in a progress bar, by doing " + \ "an initial run of the scene which skips animations." ) parser.add_argument( "--video_dir", help="Directory to write video", ) parser.add_argument( "--config_file", help="Path to the custom configuration file", ) parser.add_argument( "-v", "--version", action="store_true", help="Display the version of manimgl" ) parser.add_argument( "--log-level", help="Level of messages to Display, can be DEBUG / INFO / WARNING / ERROR / CRITICAL" ) parser.add_argument( "--clear-cache", action="store_true", help="Erase the cache used for Tex and Text Mobjects" ) parser.add_argument( "--autoreload", action="store_true", help="Automatically reload Python modules to pick up code changes " + "across different files", ) args = parser.parse_args() args.write_file = any([args.write_file, args.open, args.finder]) return args except argparse.ArgumentError as err: log.error(str(err)) sys.exit(2) def update_directory_config(config: Dict): dir_config = config.directories base = dir_config.base for key, subdir in dir_config.subdirs.items(): dir_config[key] = os.path.join(base, subdir) def update_window_config(config: Dict, args: Namespace): window_config = config.window for key in "position", "size": if window_config.get(key): window_config[key] = literal_eval(window_config[key]) if args.full_screen: window_config.full_screen = True def update_camera_config(config: Dict, args: Namespace): camera_config = config.camera arg_resolution = get_resolution_from_args(args, config.resolution_options) camera_config.resolution = arg_resolution or literal_eval(camera_config.resolution) if args.fps: camera_config.fps = args.fps if args.color: try: camera_config.background_color = colour.Color(args.color) except Exception as err: log.error("Please use a valid color") log.error(err) sys.exit(2) if args.transparent: camera_config.background_opacity = 0.0 def update_file_writer_config(config: Dict, args: Namespace): file_writer_config = config.file_writer file_writer_config.update( write_to_movie=(not args.skip_animations and args.write_file), subdivide_output=args.subdivide, save_last_frame=(args.skip_animations and args.write_file), png_mode=("RGBA" if args.transparent else "RGB"), movie_file_extension=(get_file_ext(args)), output_directory=get_output_directory(args, config), file_name=args.file_name, open_file_upon_completion=args.open, show_file_location_upon_completion=args.finder, quiet=args.quiet, ) if args.vcodec: file_writer_config.video_codec = args.vcodec elif args.transparent: file_writer_config.video_codec = 'prores_ks' file_writer_config.pixel_format = '' elif args.gif: file_writer_config.video_codec = '' if args.pix_fmt: file_writer_config.pixel_format = args.pix_fmt def update_scene_config(config: Dict, args: Namespace): scene_config = config.scene start, end = get_animations_numbers(args) scene_config.update( # Note, Scene.__init__ makes use of both manimlib.camera and # manimlib.file_writer below, so the arguments here are just for # any future specifications beyond what the global configuration holds camera_config=dict(), file_writer_config=dict(), skip_animations=args.skip_animations, start_at_animation_number=start, end_at_animation_number=end, presenter_mode=args.presenter_mode, ) if args.leave_progress_bars: scene_config.leave_progress_bars = True if args.show_animation_progress: scene_config.show_animation_progress = True def update_run_config(config: Dict, args: Namespace): config.run = Dict( file_name=args.file, embed_line=(int(args.embed) if args.embed is not None else None), is_reload=False, prerun=args.prerun, scene_names=args.scene_names, quiet=args.quiet or args.write_all, write_all=args.write_all, show_in_window=not args.write_file ) def update_embed_config(config: Dict, args: Namespace): if args.autoreload: config.embed.autoreload = True # Helpers for the functions above def load_yaml(file_path: str): try: with open(file_path, "r") as file: return yaml.safe_load(file) or {} except FileNotFoundError: return {} def get_manim_dir(): manimlib_module = importlib.import_module("manimlib") manimlib_dir = os.path.dirname(inspect.getabsfile(manimlib_module)) return os.path.abspath(os.path.join(manimlib_dir, "..")) def get_resolution_from_args(args: Optional[Namespace], resolution_options: dict) -> Optional[tuple[int, int]]: if args.resolution: return tuple(map(int, args.resolution.split("x"))) if args.low_quality: return literal_eval(resolution_options["low"]) if args.medium_quality: return literal_eval(resolution_options["med"]) if args.hd: return literal_eval(resolution_options["high"]) if args.uhd: return literal_eval(resolution_options["4k"]) return None def get_file_ext(args: Namespace) -> str: if args.transparent: file_ext = ".mov" elif args.gif: file_ext = ".gif" else: file_ext = ".mp4" return file_ext def get_animations_numbers(args: Namespace) -> tuple[int | None, int | None]: stan = args.start_at_animation_number if stan is None: return (None, None) elif "," in stan: return tuple(map(int, stan.split(","))) else: return int(stan), None def get_output_directory(args: Namespace, config: Dict) -> str: dir_config = config.directories out_dir = args.video_dir or dir_config.output if dir_config.mirror_module_path and args.file: file_path = Path(args.file).absolute() if str(file_path).startswith(dir_config.removed_mirror_prefix): rel_path = file_path.relative_to(dir_config.removed_mirror_prefix) rel_path = Path(str(rel_path).lstrip("_")) else: rel_path = file_path.stem out_dir = Path(out_dir, rel_path).with_suffix("") return out_dir # Create global configuration manim_config: Dict = initialize_manim_config()
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/__init__.py
manimlib/__init__.py
import pkg_resources __version__ = pkg_resources.get_distribution("manimgl").version from typing import TYPE_CHECKING if TYPE_CHECKING: from manimlib.typing import * from manimlib.constants import * from manimlib.window import * from manimlib.animation.animation import * from manimlib.animation.composition import * from manimlib.animation.creation import * from manimlib.animation.fading import * from manimlib.animation.growing import * from manimlib.animation.indication import * from manimlib.animation.movement import * from manimlib.animation.numbers import * from manimlib.animation.rotation import * from manimlib.animation.specialized import * from manimlib.animation.transform import * from manimlib.animation.transform_matching_parts import * from manimlib.animation.update import * from manimlib.camera.camera import * from manimlib.mobject.boolean_ops import * from manimlib.mobject.changing import * from manimlib.mobject.coordinate_systems import * from manimlib.mobject.frame import * from manimlib.mobject.functions import * from manimlib.mobject.geometry import * from manimlib.mobject.interactive import * from manimlib.mobject.matrix import * from manimlib.mobject.mobject import * from manimlib.mobject.mobject_update_utils import * from manimlib.mobject.number_line import * from manimlib.mobject.numbers import * from manimlib.mobject.probability import * from manimlib.mobject.shape_matchers import * from manimlib.mobject.svg.brace import * from manimlib.mobject.svg.drawings import * from manimlib.mobject.svg.string_mobject import * from manimlib.mobject.svg.svg_mobject import * from manimlib.mobject.svg.special_tex import * from manimlib.mobject.svg.tex_mobject import * from manimlib.mobject.svg.text_mobject import * from manimlib.mobject.three_dimensions import * from manimlib.mobject.types.dot_cloud import * from manimlib.mobject.types.image_mobject import * from manimlib.mobject.types.point_cloud_mobject import * from manimlib.mobject.types.surface import * from manimlib.mobject.types.vectorized_mobject import * from manimlib.mobject.value_tracker import * from manimlib.mobject.vector_field import * from manimlib.scene.interactive_scene import * from manimlib.scene.scene import * from manimlib.utils.bezier import * from manimlib.utils.cache import * from manimlib.utils.color import * from manimlib.utils.dict_ops import * from manimlib.utils.debug import * from manimlib.utils.directories import * from manimlib.utils.file_ops import * from manimlib.utils.images import * from manimlib.utils.iterables import * from manimlib.utils.paths import * from manimlib.utils.rate_functions import * from manimlib.utils.simple_functions import * from manimlib.utils.shaders import * from manimlib.utils.sounds import * from manimlib.utils.space_ops import * from manimlib.utils.tex import *
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/shader_wrapper.py
manimlib/shader_wrapper.py
from __future__ import annotations import copy import os import re import OpenGL.GL as gl import moderngl import numpy as np from functools import lru_cache from manimlib.config import parse_cli from manimlib.config import manim_config from manimlib.utils.shaders import get_shader_code_from_file from manimlib.utils.shaders import get_shader_program from manimlib.utils.shaders import image_path_to_texture from manimlib.utils.shaders import set_program_uniform from typing import TYPE_CHECKING if TYPE_CHECKING: from typing import Optional, Tuple, Iterable from manimlib.typing import UniformDict from moderngl.vertex_array import VertexArray from moderngl.framebuffer import Framebuffer # Mobjects that should be rendered with # the same shader will be organized and # clumped together based on keeping track # of a dict holding all the relevant information # to that shader class ShaderWrapper(object): def __init__( self, ctx: moderngl.context.Context, vert_data: np.ndarray, shader_folder: Optional[str] = None, mobject_uniforms: Optional[UniformDict] = None, # A dictionary mapping names of uniform variables texture_paths: Optional[dict[str, str]] = None, # A dictionary mapping names to filepaths for textures. depth_test: bool = False, render_primitive: int = moderngl.TRIANGLE_STRIP, code_replacements: dict[str, str] = dict(), ): self.ctx = ctx self.vert_data = vert_data self.vert_attributes = vert_data.dtype.names self.shader_folder = shader_folder self.depth_test = depth_test self.render_primitive = render_primitive self.texture_paths = texture_paths or dict() self.program_uniform_mirror: UniformDict = dict() self.bind_to_mobject_uniforms(mobject_uniforms or dict()) self.init_program_code() for old, new in code_replacements.items(): self.replace_code(old, new) self.init_program() self.init_textures() self.init_vertex_objects() self.refresh_id() def __deepcopy__(self, memo): # Don't allow deepcopies, e.g. if the mobject with this ShaderWrapper as an # attribute gets copies. Returning None means the parent object with this ShaderWrapper # as an attribute should smoothly handle this case. return None def init_program_code(self) -> None: def get_code(name: str) -> str | None: return get_shader_code_from_file( os.path.join(self.shader_folder, f"{name}.glsl") ) self.program_code: dict[str, str | None] = { "vertex_shader": get_code("vert"), "geometry_shader": get_code("geom"), "fragment_shader": get_code("frag"), } def init_program(self): if not self.shader_folder: self.program = None self.vert_format = None self.programs = [] return self.program = get_shader_program(self.ctx, **self.program_code) self.vert_format = moderngl.detect_format(self.program, self.vert_attributes) self.programs = [self.program] def init_textures(self): self.texture_names_to_ids = dict() self.textures = [] for name, path in self.texture_paths.items(): self.add_texture(name, image_path_to_texture(path, self.ctx)) def init_vertex_objects(self): self.vbo = None self.vaos = [] def add_texture(self, name: str, texture: moderngl.Texture): max_units = self.ctx.info['GL_MAX_TEXTURE_IMAGE_UNITS'] if len(self.textures) >= max_units: raise ValueError(f"Unable to use more than {max_units} textures for a program") # The position in the list determines its id self.texture_names_to_ids[name] = len(self.textures) self.textures.append(texture) def bind_to_mobject_uniforms(self, mobject_uniforms: UniformDict): self.mobject_uniforms = mobject_uniforms def get_id(self) -> int: return self.id def refresh_id(self) -> None: self.id = hash("".join(map(str, [ "".join(map(str, self.program_code.values())), self.mobject_uniforms, self.depth_test, self.render_primitive, self.texture_paths, ]))) def replace_code(self, old: str, new: str) -> None: code_map = self.program_code for name in code_map: if code_map[name] is None: continue code_map[name] = re.sub(old, new, code_map[name]) self.init_program() self.refresh_id() # Changing context def use_clip_plane(self): if "clip_plane" not in self.mobject_uniforms: return False return any(self.mobject_uniforms["clip_plane"]) def set_ctx_depth_test(self, enable: bool = True) -> None: if enable: self.ctx.enable(moderngl.DEPTH_TEST) else: self.ctx.disable(moderngl.DEPTH_TEST) def set_ctx_clip_plane(self, enable: bool = True) -> None: if enable: gl.glEnable(gl.GL_CLIP_DISTANCE0) # Adding data def read_in(self, data_list: Iterable[np.ndarray]): total_len = sum(map(len, data_list)) if total_len == 0: if self.vbo is not None: self.vbo.clear() return # If possible, read concatenated data into existing list if len(self.vert_data) != total_len: self.vert_data = np.concatenate(data_list) else: np.concatenate(data_list, out=self.vert_data) # Either create new vbo, or read data into it total_size = self.vert_data.itemsize * total_len if self.vbo is not None and self.vbo.size != total_size: self.release() # This sets vbo to be None if self.vbo is None: self.vbo = self.ctx.buffer(self.vert_data) self.generate_vaos() else: self.vbo.write(self.vert_data) def generate_vaos(self): # Vertex array object self.vaos = [ self.ctx.vertex_array( program=program, content=[(self.vbo, self.vert_format, *self.vert_attributes)], mode=self.render_primitive, ) for program in self.programs ] # Related to data and rendering def pre_render(self): self.set_ctx_depth_test(self.depth_test) self.set_ctx_clip_plane(self.use_clip_plane()) for tid, texture in enumerate(self.textures): texture.use(tid) def render(self): for vao in self.vaos: vao.render() def update_program_uniforms(self, camera_uniforms: UniformDict): for program in self.programs: if program is None: continue for uniforms in [self.mobject_uniforms, camera_uniforms, self.texture_names_to_ids]: for name, value in uniforms.items(): set_program_uniform(program, name, value) def release(self): for obj in (self.vbo, *self.vaos): if obj is not None: obj.release() self.init_vertex_objects() def release_textures(self): for texture in self.textures: texture.release() del texture self.textures = [] self.texture_names_to_ids = dict() class VShaderWrapper(ShaderWrapper): def __init__( self, ctx: moderngl.context.Context, vert_data: np.ndarray, shader_folder: Optional[str] = None, mobject_uniforms: Optional[UniformDict] = None, # A dictionary mapping names of uniform variables texture_paths: Optional[dict[str, str]] = None, # A dictionary mapping names to filepaths for textures. depth_test: bool = False, render_primitive: int = moderngl.TRIANGLES, code_replacements: dict[str, str] = dict(), stroke_behind: bool = False, ): self.stroke_behind = stroke_behind super().__init__( ctx=ctx, vert_data=vert_data, shader_folder=shader_folder, mobject_uniforms=mobject_uniforms, texture_paths=texture_paths, depth_test=depth_test, render_primitive=render_primitive, code_replacements=code_replacements, ) self.fill_canvas = VShaderWrapper.get_fill_canvas(self.ctx) self.add_texture('Texture', self.fill_canvas[0].color_attachments[0]) self.add_texture('DepthTexture', self.fill_canvas[2].color_attachments[0]) def init_program_code(self) -> None: self.program_code = { f"{vtype}_{name}": get_shader_code_from_file( os.path.join("quadratic_bezier", f"{vtype}", f"{name}.glsl") ) for vtype in ["stroke", "fill", "depth"] for name in ["vert", "geom", "frag"] } def init_program(self): self.stroke_program = get_shader_program( self.ctx, vertex_shader=self.program_code["stroke_vert"], geometry_shader=self.program_code["stroke_geom"], fragment_shader=self.program_code["stroke_frag"], ) self.fill_program = get_shader_program( self.ctx, vertex_shader=self.program_code["fill_vert"], geometry_shader=self.program_code["fill_geom"], fragment_shader=self.program_code["fill_frag"], ) self.fill_border_program = get_shader_program( self.ctx, vertex_shader=self.program_code["stroke_vert"], geometry_shader=self.program_code["stroke_geom"], fragment_shader=self.program_code["stroke_frag"].replace( "// MODIFY FRAG COLOR", "frag_color.a *= 0.95; frag_color.rgb *= frag_color.a;", ) ) self.fill_depth_program = get_shader_program( self.ctx, vertex_shader=self.program_code["depth_vert"], geometry_shader=self.program_code["depth_geom"], fragment_shader=self.program_code["depth_frag"], ) self.programs = [self.stroke_program, self.fill_program, self.fill_border_program, self.fill_depth_program] # Full vert format looks like this (total of 4x23 = 92 bytes): # point 3 # stroke_rgba 4 # stroke_width 1 # joint_angle 1 # fill_rgba 4 # base_normal 3 # fill_border_width 1 self.stroke_vert_format = '3f 4f 1f 1f 16x 3f 4x' self.stroke_vert_attributes = ['point', 'stroke_rgba', 'stroke_width', 'joint_angle', 'unit_normal'] self.fill_vert_format = '3f 24x 4f 3f 4x' self.fill_vert_attributes = ['point', 'fill_rgba', 'base_normal'] self.fill_border_vert_format = '3f 20x 1f 4f 3f 1f' self.fill_border_vert_attributes = ['point', 'joint_angle', 'stroke_rgba', 'unit_normal', 'stroke_width'] self.fill_depth_vert_format = '3f 40x 3f 4x' self.fill_depth_vert_attributes = ['point', 'base_normal'] def init_vertex_objects(self): self.vbo = None self.stroke_vao = None self.fill_vao = None self.fill_border_vao = None self.vaos = [] def generate_vaos(self): self.stroke_vao = self.ctx.vertex_array( program=self.stroke_program, content=[(self.vbo, self.stroke_vert_format, *self.stroke_vert_attributes)], mode=self.render_primitive, ) self.fill_vao = self.ctx.vertex_array( program=self.fill_program, content=[(self.vbo, self.fill_vert_format, *self.fill_vert_attributes)], mode=self.render_primitive, ) self.fill_border_vao = self.ctx.vertex_array( program=self.fill_border_program, content=[(self.vbo, self.fill_border_vert_format, *self.fill_border_vert_attributes)], mode=self.render_primitive, ) self.fill_depth_vao = self.ctx.vertex_array( program=self.fill_depth_program, content=[(self.vbo, self.fill_depth_vert_format, *self.fill_depth_vert_attributes)], mode=self.render_primitive, ) self.vaos = [self.stroke_vao, self.fill_vao, self.fill_border_vao, self.fill_depth_vao] def set_backstroke(self, value: bool = True): self.stroke_behind = value def refresh_id(self): super().refresh_id() self.id = hash(str(self.id) + str(self.stroke_behind)) # Rendering def render_stroke(self): if self.stroke_vao is None: return self.stroke_vao.render() def render_fill(self): if self.fill_vao is None: return original_fbo = self.ctx.fbo fill_tx_fbo, fill_tx_vao, depth_tx_fbo = self.fill_canvas # Render to a separate texture, due to strange alpha compositing # for the blended winding calculation fill_tx_fbo.clear() fill_tx_fbo.use() # Be sure not to apply depth test while rendering fill # but set it back to where it was after apply_depth_test = bool(gl.glGetBooleanv(gl.GL_DEPTH_TEST)) self.ctx.disable(moderngl.DEPTH_TEST) # With this blend function, the effect of blending alpha a with # -a / (1 - a) cancels out, so we can cancel positively and negatively # oriented triangles gl.glBlendFuncSeparate( gl.GL_SRC_ALPHA, gl.GL_ONE_MINUS_SRC_ALPHA, gl.GL_ONE_MINUS_DST_ALPHA, gl.GL_ONE ) self.fill_vao.render() if apply_depth_test: self.ctx.enable(moderngl.DEPTH_TEST) depth_tx_fbo.clear(1.0) depth_tx_fbo.use() gl.glBlendFunc(gl.GL_ONE, gl.GL_ONE) gl.glBlendEquation(gl.GL_MIN) self.fill_depth_vao.render() # Now add border, just taking the max alpha gl.glBlendFunc(gl.GL_ONE, gl.GL_ONE) gl.glBlendEquation(gl.GL_MAX) self.fill_border_vao.render() # Take the texture we were just drawing to, and render it to # the main scene. Account for how alphas have been premultiplied original_fbo.use() gl.glBlendFunc(gl.GL_ONE, gl.GL_ONE_MINUS_SRC_ALPHA) gl.glBlendEquation(gl.GL_FUNC_ADD) fill_tx_vao.render() # Return to original blending state gl.glBlendFunc(gl.GL_SRC_ALPHA, gl.GL_ONE_MINUS_SRC_ALPHA) # Static method returning one shared value across all VShaderWrappers @lru_cache @staticmethod def get_fill_canvas(ctx: moderngl.Context) -> Tuple[Framebuffer, VertexArray, Framebuffer]: """ Because VMobjects with fill are rendered in a funny way, using alpha blending to effectively compute the winding number around each pixel, they need to be rendered to a separate texture, which is then composited onto the ordinary frame buffer. This returns a texture, loaded into a frame buffer, and a vao which can display that texture as a simple quad onto a screen, along with the rgb value which is meant to be discarded. """ size = manim_config.camera.resolution double_size = (2 * size[0], 2 * size[1]) # Important to make sure dtype is floating point (not fixed point) # so that alpha values can be negative and are not clipped fill_texture = ctx.texture(size=double_size, components=4, dtype='f2') # Use another one to keep track of depth depth_texture = ctx.texture(size=size, components=1, dtype='f4') fill_texture_fbo = ctx.framebuffer(fill_texture) depth_texture_fbo = ctx.framebuffer(depth_texture) simple_vert = ''' #version 330 in vec2 texcoord; out vec2 uv; void main() { gl_Position = vec4((2.0 * texcoord - 1.0), 0.0, 1.0); uv = texcoord; } ''' alpha_adjust_frag = ''' #version 330 uniform sampler2D Texture; uniform sampler2D DepthTexture; in vec2 uv; out vec4 color; void main() { color = texture(Texture, uv); if(color.a == 0) discard; if(color.a < 0){ color.a = -color.a / (1.0 - color.a); color.rgb *= (color.a - 1); } // Counteract scaling in fill frag color *= 1.06; gl_FragDepth = texture(DepthTexture, uv)[0]; } ''' fill_program = ctx.program( vertex_shader=simple_vert, fragment_shader=alpha_adjust_frag, ) verts = np.array([[0, 0], [0, 1], [1, 0], [1, 1]]) simple_vbo = ctx.buffer(verts.astype('f4').tobytes()) fill_texture_vao = ctx.simple_vertex_array( fill_program, simple_vbo, 'texcoord', mode=moderngl.TRIANGLE_STRIP ) return (fill_texture_fbo, fill_texture_vao, depth_texture_fbo) def render(self): if self.stroke_behind: self.render_stroke() self.render_fill() else: self.render_fill() self.render_stroke()
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/window.py
manimlib/window.py
from __future__ import annotations import numpy as np import moderngl_window as mglw from moderngl_window.context.pyglet.window import Window as PygletWindow from moderngl_window.timers.clock import Timer from functools import wraps import screeninfo from manimlib.constants import ASPECT_RATIO from manimlib.constants import FRAME_SHAPE from typing import TYPE_CHECKING if TYPE_CHECKING: from typing import Callable, TypeVar, Optional from manimlib.scene.scene import Scene T = TypeVar("T") class Window(PygletWindow): fullscreen: bool = False resizable: bool = True gl_version: tuple[int, int] = (3, 3) vsync: bool = True cursor: bool = True def __init__( self, scene: Optional[Scene] = None, position_string: str = "UR", monitor_index: int = 1, full_screen: bool = False, size: Optional[tuple[int, int]] = None, position: Optional[tuple[int, int]] = None, samples: int = 0 ): self.scene = scene self.monitor = self.get_monitor(monitor_index) self.default_size = size or self.get_default_size(full_screen) self.default_position = position or self.position_from_string(position_string) self.pressed_keys = set() super().__init__(samples=samples) self.to_default_position() if self.scene: self.init_for_scene(scene) def init_for_scene(self, scene: Scene): """ Resets the state and updates the scene associated to this window. This is necessary when we want to reuse an *existing* window after a `scene.reload()` was requested, which will create new scene instances. """ self.pressed_keys.clear() self._has_undrawn_event = True self.scene = scene self.title = str(scene) self.init_mgl_context() self.timer = Timer() self.config = mglw.WindowConfig(ctx=self.ctx, wnd=self, timer=self.timer) mglw.activate_context(window=self, ctx=self.ctx) self.timer.start() # This line seems to resync the viewport self.on_resize(*self.size) def get_monitor(self, index): try: monitors = screeninfo.get_monitors() return monitors[min(index, len(monitors) - 1)] except screeninfo.ScreenInfoError: # Default fallback return screeninfo.Monitor(width=1920, height=1080) def get_default_size(self, full_screen=False): width = self.monitor.width // (1 if full_screen else 2) height = int(width // ASPECT_RATIO) return (width, height) def position_from_string(self, position_string): # Alternatively, it might be specified with a string like # UR, OO, DL, etc. specifying what corner it should go to char_to_n = {"L": 0, "U": 0, "O": 1, "R": 2, "D": 2} size = self.default_size width_diff = self.monitor.width - size[0] height_diff = self.monitor.height - size[1] x_step = char_to_n[position_string[1]] * width_diff // 2 y_step = char_to_n[position_string[0]] * height_diff // 2 return (self.monitor.x + x_step, -self.monitor.y + y_step) def focus(self): """ Puts focus on this window by hiding and showing it again. Note that the pyglet `activate()` method didn't work as expected here, so that's why we have to use this workaround. This will produce a small flicker on the window but at least reliably focuses it. It may also offset the window position slightly. """ self._window.set_visible(False) self._window.set_visible(True) def to_default_position(self): self.position = self.default_position # Hack. Sometimes, namely when configured to open in a separate window, # the window needs to be resized to display correctly. w, h = self.default_size self.size = (w - 1, h - 1) self.size = (w, h) # Delegate event handling to scene def pixel_coords_to_space_coords( self, px: int, py: int, relative: bool = False ) -> np.ndarray: if self.scene is None or not hasattr(self.scene, "frame"): return np.zeros(3) pixel_shape = np.array(self.size) fixed_frame_shape = np.array(FRAME_SHAPE) frame = self.scene.frame coords = np.zeros(3) coords[:2] = (fixed_frame_shape / pixel_shape) * np.array([px, py]) if not relative: coords[:2] -= 0.5 * fixed_frame_shape return frame.from_fixed_frame_point(coords, relative) def has_undrawn_event(self) -> bool: return self._has_undrawn_event def swap_buffers(self): super().swap_buffers() self._has_undrawn_event = False @staticmethod def note_undrawn_event(func: Callable[..., T]) -> Callable[..., T]: @wraps(func) def wrapper(self, *args, **kwargs): func(self, *args, **kwargs) self._has_undrawn_event = True return wrapper @note_undrawn_event def on_mouse_motion(self, x: int, y: int, dx: int, dy: int) -> None: super().on_mouse_motion(x, y, dx, dy) if not self.scene: return point = self.pixel_coords_to_space_coords(x, y) d_point = self.pixel_coords_to_space_coords(dx, dy, relative=True) self.scene.on_mouse_motion(point, d_point) @note_undrawn_event def on_mouse_drag(self, x: int, y: int, dx: int, dy: int, buttons: int, modifiers: int) -> None: super().on_mouse_drag(x, y, dx, dy, buttons, modifiers) if not self.scene: return point = self.pixel_coords_to_space_coords(x, y) d_point = self.pixel_coords_to_space_coords(dx, dy, relative=True) self.scene.on_mouse_drag(point, d_point, buttons, modifiers) @note_undrawn_event def on_mouse_press(self, x: int, y: int, button: int, mods: int) -> None: super().on_mouse_press(x, y, button, mods) if not self.scene: return point = self.pixel_coords_to_space_coords(x, y) self.scene.on_mouse_press(point, button, mods) @note_undrawn_event def on_mouse_release(self, x: int, y: int, button: int, mods: int) -> None: super().on_mouse_release(x, y, button, mods) if not self.scene: return point = self.pixel_coords_to_space_coords(x, y) self.scene.on_mouse_release(point, button, mods) @note_undrawn_event def on_mouse_scroll(self, x: int, y: int, x_offset: float, y_offset: float) -> None: super().on_mouse_scroll(x, y, x_offset, y_offset) if not self.scene: return point = self.pixel_coords_to_space_coords(x, y) offset = self.pixel_coords_to_space_coords(x_offset, y_offset, relative=True) self.scene.on_mouse_scroll(point, offset, x_offset, y_offset) @note_undrawn_event def on_key_press(self, symbol: int, modifiers: int) -> None: self.pressed_keys.add(symbol) # Modifiers? super().on_key_press(symbol, modifiers) if not self.scene: return self.scene.on_key_press(symbol, modifiers) @note_undrawn_event def on_key_release(self, symbol: int, modifiers: int) -> None: self.pressed_keys.difference_update({symbol}) # Modifiers? super().on_key_release(symbol, modifiers) if not self.scene: return self.scene.on_key_release(symbol, modifiers) @note_undrawn_event def on_resize(self, width: int, height: int) -> None: super().on_resize(width, height) if not self.scene: return self.scene.on_resize(width, height) @note_undrawn_event def on_show(self) -> None: super().on_show() if not self.scene: return self.scene.on_show() @note_undrawn_event def on_hide(self) -> None: super().on_hide() if not self.scene: return self.scene.on_hide() @note_undrawn_event def on_close(self) -> None: super().on_close() if not self.scene: return self.scene.on_close() def is_key_pressed(self, symbol: int) -> bool: return (symbol in self.pressed_keys)
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/event_handler/event_listner.py
manimlib/event_handler/event_listner.py
from __future__ import annotations from typing import TYPE_CHECKING if TYPE_CHECKING: from typing import Callable from manimlib.event_handler.event_type import EventType from manimlib.mobject.mobject import Mobject class EventListener(object): def __init__( self, mobject: Mobject, event_type: EventType, event_callback: Callable[[Mobject, dict[str]]] ): self.mobject = mobject self.event_type = event_type self.callback = event_callback def __eq__(self, o: object) -> bool: return_val = False try: return_val = self.callback == o.callback \ and self.mobject == o.mobject \ and self.event_type == o.event_type except: pass return return_val
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/event_handler/event_dispatcher.py
manimlib/event_handler/event_dispatcher.py
from __future__ import annotations import numpy as np from manimlib.event_handler.event_listner import EventListener from manimlib.event_handler.event_type import EventType class EventDispatcher(object): def __init__(self): self.event_listners: dict[ EventType, list[EventListener] ] = { event_type: [] for event_type in EventType } self.mouse_point = np.array((0., 0., 0.)) self.mouse_drag_point = np.array((0., 0., 0.)) self.pressed_keys: set[int] = set() self.draggable_object_listners: list[EventListener] = [] def add_listner(self, event_listner: EventListener): assert isinstance(event_listner, EventListener) self.event_listners[event_listner.event_type].append(event_listner) return self def remove_listner(self, event_listner: EventListener): assert isinstance(event_listner, EventListener) try: while event_listner in self.event_listners[event_listner.event_type]: self.event_listners[event_listner.event_type].remove(event_listner) except: # raise ValueError("Handler is not handling this event, so cannot remove it.") pass return self def dispatch(self, event_type: EventType, **event_data): if event_type == EventType.MouseMotionEvent: self.mouse_point = event_data["point"] elif event_type == EventType.MouseDragEvent: self.mouse_drag_point = event_data["point"] elif event_type == EventType.KeyPressEvent: self.pressed_keys.add(event_data["symbol"]) # Modifiers? elif event_type == EventType.KeyReleaseEvent: self.pressed_keys.difference_update({event_data["symbol"]}) # Modifiers? elif event_type == EventType.MousePressEvent: self.draggable_object_listners = [ listner for listner in self.event_listners[EventType.MouseDragEvent] if listner.mobject.is_point_touching(self.mouse_point) ] elif event_type == EventType.MouseReleaseEvent: self.draggable_object_listners = [] propagate_event = None if event_type == EventType.MouseDragEvent: for listner in self.draggable_object_listners: assert isinstance(listner, EventListener) propagate_event = listner.callback(listner.mobject, event_data) if propagate_event is not None and propagate_event is False: return propagate_event elif event_type.value.startswith('mouse'): for listner in self.event_listners[event_type]: if listner.mobject.is_point_touching(self.mouse_point): propagate_event = listner.callback( listner.mobject, event_data) if propagate_event is not None and propagate_event is False: return propagate_event elif event_type.value.startswith('key'): for listner in self.event_listners[event_type]: propagate_event = listner.callback(listner.mobject, event_data) if propagate_event is not None and propagate_event is False: return propagate_event return propagate_event def get_listners_count(self) -> int: return sum([len(value) for key, value in self.event_listners.items()]) def get_mouse_point(self) -> np.ndarray: return self.mouse_point def get_mouse_drag_point(self) -> np.ndarray: return self.mouse_drag_point def is_key_pressed(self, symbol: int) -> bool: return (symbol in self.pressed_keys) __iadd__ = add_listner __isub__ = remove_listner __call__ = dispatch __len__ = get_listners_count
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/event_handler/__init__.py
manimlib/event_handler/__init__.py
from manimlib.event_handler.event_dispatcher import EventDispatcher # This is supposed to be a Singleton # i.e., during runtime there should be only one object of Event Dispatcher EVENT_DISPATCHER = EventDispatcher()
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/event_handler/event_type.py
manimlib/event_handler/event_type.py
from enum import Enum class EventType(Enum): MouseMotionEvent = 'mouse_motion_event' MousePressEvent = 'mouse_press_event' MouseReleaseEvent = 'mouse_release_event' MouseDragEvent = 'mouse_drag_event' MouseScrollEvent = 'mouse_scroll_event' KeyPressEvent = 'key_press_event' KeyReleaseEvent = 'key_release_event'
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/camera/camera_frame.py
manimlib/camera/camera_frame.py
from __future__ import annotations import math import warnings import numpy as np from scipy.spatial.transform import Rotation from manimlib.constants import DEG, RADIANS from manimlib.constants import FRAME_SHAPE from manimlib.constants import DOWN, LEFT, ORIGIN, OUT, RIGHT, UP from manimlib.constants import PI from manimlib.mobject.mobject import Mobject from manimlib.utils.space_ops import normalize from manimlib.utils.simple_functions import clip from typing import TYPE_CHECKING if TYPE_CHECKING: from manimlib.typing import Vect3 class CameraFrame(Mobject): def __init__( self, frame_shape: tuple[float, float] = FRAME_SHAPE, center_point: Vect3 = ORIGIN, # Field of view in the y direction fovy: float = 45 * DEG, euler_axes: str = "zxz", # This keeps it ordered first in a scene z_index=-1, **kwargs, ): super().__init__(z_index=z_index, **kwargs) self.uniforms["orientation"] = Rotation.identity().as_quat() self.uniforms["fovy"] = fovy self.default_orientation = Rotation.identity() self.view_matrix = np.identity(4) self.id4x4 = np.identity(4) self.camera_location = OUT # This will be updated by set_points self.euler_axes = euler_axes self.set_points(np.array([ORIGIN, LEFT, RIGHT, DOWN, UP])) self.set_width(frame_shape[0], stretch=True) self.set_height(frame_shape[1], stretch=True) self.move_to(center_point) def set_orientation(self, rotation: Rotation): self.uniforms["orientation"][:] = rotation.as_quat() return self def get_orientation(self): return Rotation.from_quat(self.uniforms["orientation"]) def make_orientation_default(self): self.default_orientation = self.get_orientation() return self def to_default_state(self): self.set_shape(*FRAME_SHAPE) self.center() self.set_orientation(self.default_orientation) return self def get_euler_angles(self) -> np.ndarray: orientation = self.get_orientation() if np.isclose(orientation.as_quat(), [0, 0, 0, 1]).all(): return np.zeros(3) with warnings.catch_warnings(): warnings.simplefilter('ignore', UserWarning) # Ignore UserWarnings angles = orientation.as_euler(self.euler_axes)[::-1] # Handle Gimble lock case if self.euler_axes == "zxz": if np.isclose(angles[1], 0, atol=1e-2): angles[0] = angles[0] + angles[2] angles[2] = 0 if np.isclose(angles[1], PI, atol=1e-2): angles[0] = angles[0] - angles[2] angles[2] = 0 return angles def get_theta(self): return self.get_euler_angles()[0] def get_phi(self): return self.get_euler_angles()[1] def get_gamma(self): return self.get_euler_angles()[2] def get_scale(self): return self.get_height() / FRAME_SHAPE[1] def get_inverse_camera_rotation_matrix(self): return self.get_orientation().as_matrix().T def get_view_matrix(self, refresh=False): """ Returns a 4x4 for the affine transformation mapping a point into the camera's internal coordinate system """ if self._data_has_changed: shift = self.id4x4.copy() rotation = self.id4x4.copy() scale = self.get_scale() shift[:3, 3] = -self.get_center() rotation[:3, :3] = self.get_inverse_camera_rotation_matrix() np.dot(rotation, shift, out=self.view_matrix) if scale > 0: self.view_matrix[:3, :4] /= scale return self.view_matrix def get_inv_view_matrix(self): return np.linalg.inv(self.get_view_matrix()) @Mobject.affects_data def interpolate(self, *args, **kwargs): super().interpolate(*args, **kwargs) @Mobject.affects_data def rotate(self, angle: float, axis: np.ndarray = OUT, **kwargs): rot = Rotation.from_rotvec(angle * normalize(axis)) self.set_orientation(rot * self.get_orientation()) return self def set_euler_angles( self, theta: float | None = None, phi: float | None = None, gamma: float | None = None, units: float = RADIANS ): eulers = self.get_euler_angles() # theta, phi, gamma for i, var in enumerate([theta, phi, gamma]): if var is not None: eulers[i] = var * units if all(eulers == 0): rot = Rotation.identity() else: rot = Rotation.from_euler(self.euler_axes, eulers[::-1]) self.set_orientation(rot) return self def increment_euler_angles( self, dtheta: float = 0, dphi: float = 0, dgamma: float = 0, units: float = RADIANS ): angles = self.get_euler_angles() new_angles = angles + np.array([dtheta, dphi, dgamma]) * units # Limit range for phi if self.euler_axes == "zxz": new_angles[1] = clip(new_angles[1], 0, PI) elif self.euler_axes == "zxy": new_angles[1] = clip(new_angles[1], -PI / 2, PI / 2) new_rot = Rotation.from_euler(self.euler_axes, new_angles[::-1]) self.set_orientation(new_rot) return self def set_euler_axes(self, seq: str): self.euler_axes = seq def reorient( self, theta_degrees: float | None = None, phi_degrees: float | None = None, gamma_degrees: float | None = None, center: Vect3 | tuple[float, float, float] | None = None, height: float | None = None ): """ Shortcut for set_euler_angles, defaulting to taking in angles in degrees """ self.set_euler_angles(theta_degrees, phi_degrees, gamma_degrees, units=DEG) if center is not None: self.move_to(np.array(center)) if height is not None: self.set_height(height) return self def set_theta(self, theta: float): return self.set_euler_angles(theta=theta) def set_phi(self, phi: float): return self.set_euler_angles(phi=phi) def set_gamma(self, gamma: float): return self.set_euler_angles(gamma=gamma) def increment_theta(self, dtheta: float, units=RADIANS): self.increment_euler_angles(dtheta=dtheta, units=units) return self def increment_phi(self, dphi: float, units=RADIANS): self.increment_euler_angles(dphi=dphi, units=units) return self def increment_gamma(self, dgamma: float, units=RADIANS): self.increment_euler_angles(dgamma=dgamma, units=units) return self def add_ambient_rotation(self, angular_speed=1 * DEG): self.add_updater(lambda m, dt: m.increment_theta(angular_speed * dt)) return self @Mobject.affects_data def set_focal_distance(self, focal_distance: float): self.uniforms["fovy"] = 2 * math.atan(0.5 * self.get_height() / focal_distance) return self @Mobject.affects_data def set_field_of_view(self, field_of_view: float): self.uniforms["fovy"] = field_of_view return self def get_shape(self): return (self.get_width(), self.get_height()) def get_aspect_ratio(self): width, height = self.get_shape() return width / height def get_center(self) -> np.ndarray: # Assumes first point is at the center return self.get_points()[0] def get_width(self) -> float: points = self.get_points() return points[2, 0] - points[1, 0] def get_height(self) -> float: points = self.get_points() return points[4, 1] - points[3, 1] def get_focal_distance(self) -> float: return 0.5 * self.get_height() / math.tan(0.5 * self.uniforms["fovy"]) def get_field_of_view(self) -> float: return self.uniforms["fovy"] def get_implied_camera_location(self) -> np.ndarray: if self._data_has_changed: to_camera = self.get_inverse_camera_rotation_matrix()[2] dist = self.get_focal_distance() self.camera_location = self.get_center() + dist * to_camera return self.camera_location def to_fixed_frame_point(self, point: Vect3, relative: bool = False): view = self.get_view_matrix() point4d = [*point, 0 if relative else 1] return np.dot(point4d, view.T)[:3] def from_fixed_frame_point(self, point: Vect3, relative: bool = False): inv_view = self.get_inv_view_matrix() point4d = [*point, 0 if relative else 1] return np.dot(point4d, inv_view.T)[:3]
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/camera/camera.py
manimlib/camera/camera.py
from __future__ import annotations import moderngl import numpy as np import OpenGL.GL as gl from PIL import Image from manimlib.camera.camera_frame import CameraFrame from manimlib.constants import BLACK from manimlib.constants import DEFAULT_RESOLUTION from manimlib.constants import FRAME_HEIGHT from manimlib.constants import FRAME_WIDTH from manimlib.mobject.mobject import Mobject from manimlib.mobject.mobject import Point from manimlib.utils.color import color_to_rgba from typing import TYPE_CHECKING if TYPE_CHECKING: from typing import Optional from manimlib.typing import ManimColor, Vect3 from manimlib.window import Window class Camera(object): def __init__( self, window: Optional[Window] = None, background_image: Optional[str] = None, frame_config: dict = dict(), # Note: frame height and width will be resized to match this resolution aspect ratio resolution=DEFAULT_RESOLUTION, fps: int = 30, background_color: ManimColor = BLACK, background_opacity: float = 1.0, # Points in vectorized mobjects with norm greater # than this value will be rescaled. max_allowable_norm: float = FRAME_WIDTH, image_mode: str = "RGBA", n_channels: int = 4, pixel_array_dtype: type = np.uint8, light_source_position: Vect3 = np.array([-10, 10, 10]), # Although vector graphics handle antialiasing fine # without multisampling, for 3d scenes one might want # to set samples to be greater than 0. samples: int = 0, ): self.window = window self.background_image = background_image self.default_pixel_shape = resolution # Rename? self.fps = fps self.max_allowable_norm = max_allowable_norm self.image_mode = image_mode self.n_channels = n_channels self.pixel_array_dtype = pixel_array_dtype self.light_source_position = light_source_position self.samples = samples self.rgb_max_val: float = np.iinfo(self.pixel_array_dtype).max self.background_rgba: list[float] = list(color_to_rgba( background_color, background_opacity )) self.uniforms = dict() self.init_frame(**frame_config) self.init_context() self.init_fbo() self.init_light_source() def init_frame(self, **config) -> None: self.frame = CameraFrame(**config) def init_context(self) -> None: if self.window is None: self.ctx: moderngl.Context = moderngl.create_standalone_context() else: self.ctx: moderngl.Context = self.window.ctx self.ctx.enable(moderngl.PROGRAM_POINT_SIZE) self.ctx.enable(moderngl.BLEND) def init_fbo(self) -> None: # This is the buffer used when writing to a video/image file self.fbo_for_files = self.get_fbo(self.samples) # This is the frame buffer we'll draw into when emitting frames self.draw_fbo = self.get_fbo(samples=0) if self.window is None: self.window_fbo = None self.fbo = self.fbo_for_files else: self.window_fbo = self.ctx.detect_framebuffer() self.fbo = self.window_fbo self.fbo.use() def init_light_source(self) -> None: self.light_source = Point(self.light_source_position) def use_window_fbo(self, use: bool = True): assert self.window is not None if use: self.fbo = self.window_fbo else: self.fbo = self.fbo_for_files # Methods associated with the frame buffer def get_fbo( self, samples: int = 0 ) -> moderngl.Framebuffer: return self.ctx.framebuffer( color_attachments=self.ctx.texture( self.default_pixel_shape, components=self.n_channels, samples=samples, ), depth_attachment=self.ctx.depth_renderbuffer( self.default_pixel_shape, samples=samples ) ) def clear(self) -> None: self.fbo.clear(*self.background_rgba) if self.window: self.window.clear(*self.background_rgba) def blit(self, src_fbo, dst_fbo): """ Copy blocks between fbo's using Blit """ gl.glBindFramebuffer(gl.GL_READ_FRAMEBUFFER, src_fbo.glo) gl.glBindFramebuffer(gl.GL_DRAW_FRAMEBUFFER, dst_fbo.glo) gl.glBlitFramebuffer( *src_fbo.viewport, *dst_fbo.viewport, gl.GL_COLOR_BUFFER_BIT, gl.GL_LINEAR ) def get_raw_fbo_data(self, dtype: str = 'f1') -> bytes: self.blit(self.fbo, self.draw_fbo) return self.draw_fbo.read( viewport=self.draw_fbo.viewport, components=self.n_channels, dtype=dtype, ) def get_image(self) -> Image.Image: return Image.frombytes( 'RGBA', self.get_pixel_shape(), self.get_raw_fbo_data(), 'raw', 'RGBA', 0, -1 ) def get_pixel_array(self) -> np.ndarray: raw = self.get_raw_fbo_data(dtype='f4') flat_arr = np.frombuffer(raw, dtype='f4') arr = flat_arr.reshape([*reversed(self.draw_fbo.size), self.n_channels]) arr = arr[::-1] # Convert from float return (self.rgb_max_val * arr).astype(self.pixel_array_dtype) # Needed? def get_texture(self) -> moderngl.Texture: texture = self.ctx.texture( size=self.fbo.size, components=4, data=self.get_raw_fbo_data(), dtype='f4' ) return texture # Getting camera attributes def get_pixel_size(self) -> float: return self.frame.get_width() / self.get_pixel_shape()[0] def get_pixel_shape(self) -> tuple[int, int]: return self.fbo.size def get_pixel_width(self) -> int: return self.get_pixel_shape()[0] def get_pixel_height(self) -> int: return self.get_pixel_shape()[1] def get_aspect_ratio(self): pw, ph = self.get_pixel_shape() return pw / ph def get_frame_height(self) -> float: return self.frame.get_height() def get_frame_width(self) -> float: return self.frame.get_width() def get_frame_shape(self) -> tuple[float, float]: return (self.get_frame_width(), self.get_frame_height()) def get_frame_center(self) -> np.ndarray: return self.frame.get_center() def get_location(self) -> tuple[float, float, float]: return self.frame.get_implied_camera_location() def resize_frame_shape(self, fixed_dimension: bool = False) -> None: """ Changes frame_shape to match the aspect ratio of the pixels, where fixed_dimension determines whether frame_height or frame_width remains fixed while the other changes accordingly. """ frame_height = self.get_frame_height() frame_width = self.get_frame_width() aspect_ratio = self.get_aspect_ratio() if not fixed_dimension: frame_height = frame_width / aspect_ratio else: frame_width = aspect_ratio * frame_height self.frame.set_height(frame_height, stretch=True) self.frame.set_width(frame_width, stretch=True) # Rendering def capture(self, *mobjects: Mobject) -> None: self.clear() self.refresh_uniforms() self.fbo.use() for mobject in mobjects: mobject.render(self.ctx, self.uniforms) if self.window: self.window.swap_buffers() if self.fbo is not self.window_fbo: self.blit(self.fbo, self.window_fbo) self.window.swap_buffers() def refresh_uniforms(self) -> None: frame = self.frame view_matrix = frame.get_view_matrix() light_pos = self.light_source.get_location() cam_pos = self.frame.get_implied_camera_location() self.uniforms.update( view=tuple(view_matrix.T.flatten()), frame_scale=frame.get_scale(), frame_rescale_factors=( 2.0 / FRAME_WIDTH, 2.0 / FRAME_HEIGHT, frame.get_scale() / frame.get_focal_distance(), ), pixel_size=self.get_pixel_size(), camera_position=tuple(cam_pos), light_position=tuple(light_pos), ) # Mostly just defined so old scenes don't break class ThreeDCamera(Camera): def __init__(self, samples: int = 4, **kwargs): super().__init__(samples=samples, **kwargs)
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/camera/__init__.py
manimlib/camera/__init__.py
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/scene/scene_embed.py
manimlib/scene/scene_embed.py
from __future__ import annotations import inspect import pyperclip import traceback from IPython.terminal import pt_inputhooks from IPython.terminal.embed import InteractiveShellEmbed from manimlib.animation.fading import VFadeInThenOut from manimlib.config import manim_config from manimlib.constants import RED from manimlib.mobject.mobject import Mobject from manimlib.mobject.frame import FullScreenRectangle from manimlib.module_loader import ModuleLoader from typing import TYPE_CHECKING if TYPE_CHECKING: from manimlib.scene.scene import Scene class InteractiveSceneEmbed: def __init__(self, scene: Scene): self.scene = scene self.checkpoint_manager = CheckpointManager() self.shell = self.get_ipython_shell_for_embedded_scene() self.enable_gui() self.ensure_frame_update_post_cell() self.ensure_flash_on_error() if manim_config.embed.autoreload: self.auto_reload() def launch(self): self.shell() def get_ipython_shell_for_embedded_scene(self) -> InteractiveShellEmbed: """ Create embedded IPython terminal configured to have access to the local namespace of the caller """ # Triple back should take us to the context in a user's scene definition # which is calling "self.embed" caller_frame = inspect.currentframe().f_back.f_back.f_back # Update the module's namespace to include local variables module = ModuleLoader.get_module(caller_frame.f_globals["__file__"]) module.__dict__.update(caller_frame.f_locals) module.__dict__.update(self.get_shortcuts()) exception_mode = manim_config.embed.exception_mode return InteractiveShellEmbed( user_module=module, display_banner=False, xmode=exception_mode ) def get_shortcuts(self): """ A few custom shortcuts useful to have in the interactive shell namespace """ scene = self.scene return dict( play=scene.play, wait=scene.wait, add=scene.add, remove=scene.remove, clear=scene.clear, focus=scene.focus, save_state=scene.save_state, undo=scene.undo, redo=scene.redo, i2g=scene.i2g, i2m=scene.i2m, checkpoint_paste=self.checkpoint_paste, clear_checkpoints=self.checkpoint_manager.clear_checkpoints, reload=self.reload_scene # Defined below ) def enable_gui(self): """Enables gui interactions during the embed""" def inputhook(context): while not context.input_is_ready(): if not self.scene.is_window_closing(): self.scene.update_frame(dt=0) if self.scene.is_window_closing(): self.shell.ask_exit() pt_inputhooks.register("manim", inputhook) self.shell.enable_gui("manim") def ensure_frame_update_post_cell(self): """Ensure the scene updates its frame after each ipython cell""" def post_cell_func(*args, **kwargs): if not self.scene.is_window_closing(): self.scene.update_frame(dt=0, force_draw=True) self.shell.events.register("post_run_cell", post_cell_func) def ensure_flash_on_error(self): """Flash border, and potentially play sound, on exceptions""" def custom_exc(shell, etype, evalue, tb, tb_offset=None): # Show the error don't just swallow it shell.showtraceback((etype, evalue, tb), tb_offset=tb_offset) rect = FullScreenRectangle().set_stroke(RED, 30).set_fill(opacity=0) rect.fix_in_frame() self.scene.play(VFadeInThenOut(rect, run_time=0.5)) self.shell.set_custom_exc((Exception,), custom_exc) def validate_syntax(self, file_path: str) -> bool: """ Validates the syntax of a Python file without executing it. Returns True if syntax is valid, False otherwise. Prints syntax errors to the console if found. """ try: with open(file_path, 'r', encoding='utf-8') as f: source_code = f.read() # Use compile() to check for syntax errors without executing compile(source_code, file_path, 'exec') return True except SyntaxError as e: print(f"\nSyntax Error in {file_path}:") print(f" Line {e.lineno}: {e.text.strip() if e.text else ''}") print(f" {' ' * (e.offset - 1 if e.offset else 0)}^") print(f" {e.msg}") return False except Exception as e: print(f"\nError reading {file_path}: {e}") return False def reload_scene(self, embed_line: int | None = None) -> None: """ Reloads the scene just like the `manimgl` command would do with the same arguments that were provided for the initial startup. This allows for quick iteration during scene development since we don't have to exit the IPython kernel and re-run the `manimgl` command again. The GUI stays open during the reload. If `embed_line` is provided, the scene will be reloaded at that line number. This corresponds to the `linemarker` param of the `extract_scene.insert_embed_line_to_module()` method. Before reload, the scene is cleared and the entire state is reset, such that we can start from a clean slate. This is taken care of by the run_scenes function in __main__.py, which will catch the error raised by the `exit_raise` magic command that we invoke here. Note that we cannot define a custom exception class for this error, since the IPython kernel will swallow any exception. While we can catch such an exception in our custom exception handler registered with the `set_custom_exc` method, we cannot break out of the IPython shell by this means. """ # Get the current file path for syntax validation current_file = self.shell.user_module.__file__ # Validate syntax before attempting reload if not self.validate_syntax(current_file): print("[ERROR] Reload cancelled due to syntax errors. Fix the errors and try again.") return # Update the global run configuration. run_config = manim_config.run run_config.is_reload = True if embed_line: run_config.embed_line = embed_line print("Reloading...") self.shell.run_line_magic("exit_raise", "") def auto_reload(self): """Enables reload the shell's module before all calls""" def pre_cell_func(*args, **kwargs): new_mod = ModuleLoader.get_module(self.shell.user_module.__file__, is_during_reload=True) self.shell.user_ns.update(vars(new_mod)) self.shell.events.register("pre_run_cell", pre_cell_func) def checkpoint_paste( self, skip: bool = False, record: bool = False, progress_bar: bool = True ): with self.scene.temp_config_change(skip, record, progress_bar): self.checkpoint_manager.checkpoint_paste(self.shell, self.scene) class CheckpointManager: def __init__(self): self.checkpoint_states: dict[str, list[tuple[Mobject, Mobject]]] = dict() def checkpoint_paste(self, shell, scene): """ Used during interactive development to run (or re-run) a block of scene code. If the copied selection starts with a comment, this will revert to the state of the scene the first time this function was called on a block of code starting with that comment. """ code_string = pyperclip.paste() checkpoint_key = self.get_leading_comment(code_string) self.handle_checkpoint_key(scene, checkpoint_key) shell.run_cell(code_string) @staticmethod def get_leading_comment(code_string: str) -> str: leading_line = code_string.partition("\n")[0].lstrip() if leading_line.startswith("#"): return leading_line return "" def handle_checkpoint_key(self, scene, key: str): if not key: return elif key in self.checkpoint_states: # Revert to checkpoint scene.restore_state(self.checkpoint_states[key]) # Clear out any saved states that show up later all_keys = list(self.checkpoint_states.keys()) index = all_keys.index(key) for later_key in all_keys[index + 1:]: self.checkpoint_states.pop(later_key) else: self.checkpoint_states[key] = scene.get_state() def clear_checkpoints(self): self.checkpoint_states = dict()
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/scene/interactive_scene.py
manimlib/scene/interactive_scene.py
from __future__ import annotations import itertools as it import numpy as np import pyperclip from IPython.core.getipython import get_ipython from pyglet.window import key as PygletWindowKeys from manimlib.animation.fading import FadeIn from manimlib.config import manim_config from manimlib.constants import DL, DOWN, DR, LEFT, ORIGIN, RIGHT, UL, UP, UR from manimlib.constants import FRAME_WIDTH, FRAME_HEIGHT, SMALL_BUFF from manimlib.constants import PI from manimlib.constants import DEG from manimlib.constants import MANIM_COLORS, WHITE, GREY_A, GREY_C from manimlib.mobject.geometry import Line from manimlib.mobject.geometry import Rectangle from manimlib.mobject.geometry import Square from manimlib.mobject.mobject import Group from manimlib.mobject.mobject import Mobject from manimlib.mobject.numbers import DecimalNumber from manimlib.mobject.svg.tex_mobject import Tex from manimlib.mobject.svg.text_mobject import Text from manimlib.mobject.types.dot_cloud import DotCloud from manimlib.mobject.types.vectorized_mobject import VGroup from manimlib.mobject.types.vectorized_mobject import VHighlight from manimlib.mobject.types.vectorized_mobject import VMobject from manimlib.scene.scene import Scene from manimlib.scene.scene import SceneState from manimlib.utils.family_ops import extract_mobject_family_members from manimlib.utils.space_ops import get_norm from manimlib.utils.tex_file_writing import LatexError from typing import TYPE_CHECKING if TYPE_CHECKING: from manimlib.typing import Vect3 SELECT_KEY = manim_config.key_bindings.select UNSELECT_KEY = manim_config.key_bindings.unselect GRAB_KEY = manim_config.key_bindings.grab X_GRAB_KEY = manim_config.key_bindings.x_grab Y_GRAB_KEY = manim_config.key_bindings.y_grab GRAB_KEYS = [GRAB_KEY, X_GRAB_KEY, Y_GRAB_KEY] RESIZE_KEY = manim_config.key_bindings.resize # TODO COLOR_KEY = manim_config.key_bindings.color INFORMATION_KEY = manim_config.key_bindings.information CURSOR_KEY = manim_config.key_bindings.cursor # For keyboard interactions ARROW_SYMBOLS: list[int] = [ PygletWindowKeys.LEFT, PygletWindowKeys.UP, PygletWindowKeys.RIGHT, PygletWindowKeys.DOWN, ] ALL_MODIFIERS = PygletWindowKeys.MOD_CTRL | PygletWindowKeys.MOD_COMMAND | PygletWindowKeys.MOD_SHIFT # Note, a lot of the functionality here is still buggy and very much a work in progress. class InteractiveScene(Scene): """ To select mobjects on screen, hold ctrl and move the mouse to highlight a region, or just tap ctrl to select the mobject under the cursor. Pressing command + t will toggle between modes where you either select top level mobjects part of the scene, or low level pieces. Hold 'g' to grab the selection and move it around Hold 'h' to drag it constrained in the horizontal direction Hold 'v' to drag it constrained in the vertical direction Hold 't' to resize selection, adding 'shift' to resize with respect to a corner Command + 'c' copies the ids of selections to clipboard Command + 'v' will paste either: - The copied mobject - A Tex mobject based on copied LaTeX - A Text mobject based on copied Text Command + 'z' restores selection back to its original state Command + 's' saves the selected mobjects to file """ corner_dot_config = dict( color=WHITE, radius=0.05, glow_factor=2.0, ) selection_rectangle_stroke_color = WHITE selection_rectangle_stroke_width = 1.0 palette_colors = MANIM_COLORS selection_nudge_size = 0.05 cursor_location_config = dict( font_size=24, fill_color=GREY_C, num_decimal_places=3, ) time_label_config = dict( font_size=24, fill_color=GREY_C, num_decimal_places=1, ) crosshair_width = 0.2 crosshair_style = dict( stroke_color=GREY_A, stroke_width=[3, 0, 3], ) def setup(self): self.selection = Group() self.selection_highlight = self.get_selection_highlight() self.selection_rectangle = self.get_selection_rectangle() self.crosshair = self.get_crosshair() self.information_label = self.get_information_label() self.color_palette = self.get_color_palette() self.unselectables = [ self.selection, self.selection_highlight, self.selection_rectangle, self.crosshair, self.information_label, self.camera.frame ] self.select_top_level_mobs = True self.regenerate_selection_search_set() self.is_selecting = False self.is_grabbing = False self.add(self.selection_highlight) def get_selection_rectangle(self): rect = Rectangle( stroke_color=self.selection_rectangle_stroke_color, stroke_width=self.selection_rectangle_stroke_width, ) rect.fix_in_frame() rect.fixed_corner = ORIGIN rect.add_updater(self.update_selection_rectangle) return rect def update_selection_rectangle(self, rect: Rectangle): p1 = rect.fixed_corner p2 = self.frame.to_fixed_frame_point(self.mouse_point.get_center()) rect.set_points_as_corners([ p1, np.array([p2[0], p1[1], 0]), p2, np.array([p1[0], p2[1], 0]), p1, ]) return rect def get_selection_highlight(self): result = Group() result.tracked_mobjects = [] result.add_updater(self.update_selection_highlight) return result def update_selection_highlight(self, highlight: Mobject): if set(highlight.tracked_mobjects) == set(self.selection): return # Otherwise, refresh contents of highlight highlight.tracked_mobjects = list(self.selection) highlight.set_submobjects([ self.get_highlight(mob) for mob in self.selection ]) try: index = min(( i for i, mob in enumerate(self.mobjects) for sm in self.selection if sm in mob.get_family() )) self.mobjects.remove(highlight) self.mobjects.insert(index - 1, highlight) except ValueError: pass def get_crosshair(self): lines = VMobject().replicate(2) lines[0].set_points([LEFT, ORIGIN, RIGHT]) lines[1].set_points([UP, ORIGIN, DOWN]) crosshair = VGroup(*lines) crosshair.set_width(self.crosshair_width) crosshair.set_style(**self.crosshair_style) crosshair.set_animating_status(True) crosshair.fix_in_frame() return crosshair def get_color_palette(self): palette = VGroup(*( Square(fill_color=color, fill_opacity=1, side_length=1) for color in self.palette_colors )) palette.set_stroke(width=0) palette.arrange(RIGHT, buff=0.5) palette.set_width(FRAME_WIDTH - 0.5) palette.to_edge(DOWN, buff=SMALL_BUFF) palette.fix_in_frame() return palette def get_information_label(self): loc_label = VGroup(*( DecimalNumber(**self.cursor_location_config) for n in range(3) )) def update_coords(loc_label): for mob, coord in zip(loc_label, self.mouse_point.get_location()): mob.set_value(coord) loc_label.arrange(RIGHT, buff=loc_label.get_height()) loc_label.to_corner(DR, buff=SMALL_BUFF) loc_label.fix_in_frame() return loc_label loc_label.add_updater(update_coords) time_label = DecimalNumber(0, **self.time_label_config) time_label.to_corner(DL, buff=SMALL_BUFF) time_label.fix_in_frame() time_label.add_updater(lambda m, dt: m.increment_value(dt)) return VGroup(loc_label, time_label) # Overrides def get_state(self): return SceneState(self, ignore=[ self.selection_highlight, self.selection_rectangle, self.crosshair, ]) def restore_state(self, scene_state: SceneState): super().restore_state(scene_state) self.mobjects.insert(0, self.selection_highlight) def add(self, *mobjects: Mobject): super().add(*mobjects) self.regenerate_selection_search_set() def remove(self, *mobjects: Mobject): super().remove(*mobjects) self.regenerate_selection_search_set() def remove_all_except(self, *mobjects_to_keep : Mobject): super().remove_all_except(*mobjects_to_keep) self.regenerate_selection_search_set() # Related to selection def toggle_selection_mode(self): self.select_top_level_mobs = not self.select_top_level_mobs self.refresh_selection_scope() self.regenerate_selection_search_set() def get_selection_search_set(self) -> list[Mobject]: return self.selection_search_set def regenerate_selection_search_set(self): selectable = list(filter( lambda m: m not in self.unselectables, self.mobjects )) if self.select_top_level_mobs: self.selection_search_set = selectable else: self.selection_search_set = [ submob for mob in selectable for submob in mob.family_members_with_points() ] def refresh_selection_scope(self): curr = list(self.selection) if self.select_top_level_mobs: self.selection.set_submobjects([ mob for mob in self.mobjects if any(sm in mob.get_family() for sm in curr) ]) self.selection.refresh_bounding_box(recurse_down=True) else: self.selection.set_submobjects( extract_mobject_family_members( curr, exclude_pointless=True, ) ) def get_corner_dots(self, mobject: Mobject) -> Mobject: dots = DotCloud(**self.corner_dot_config) radius = float(self.corner_dot_config["radius"]) if mobject.get_depth() < 1e-2: vects = [DL, UL, UR, DR] else: vects = np.array(list(it.product(*3 * [[-1, 1]]))) dots.add_updater(lambda d: d.set_points([ mobject.get_corner(v) + v * radius for v in vects ])) return dots def get_highlight(self, mobject: Mobject) -> Mobject: if isinstance(mobject, VMobject) and mobject.has_points() and not self.select_top_level_mobs: length = max([mobject.get_height(), mobject.get_width()]) result = VHighlight( mobject, max_stroke_addition=min([50 * length, 10]), ) result.add_updater(lambda m: m.replace(mobject, stretch=True)) return result elif isinstance(mobject, DotCloud): return Mobject() else: return self.get_corner_dots(mobject) def add_to_selection(self, *mobjects: Mobject): mobs = list(filter( lambda m: m not in self.unselectables and m not in self.selection, mobjects )) if len(mobs) == 0: return self.selection.add(*mobs) for mob in mobs: mob.set_animating_status(True) def toggle_from_selection(self, *mobjects: Mobject): for mob in mobjects: if mob in self.selection: self.selection.remove(mob) mob.set_animating_status(False) mob.refresh_bounding_box() else: self.add_to_selection(mob) def clear_selection(self): for mob in self.selection: mob.set_animating_status(False) mob.refresh_bounding_box() self.selection.set_submobjects([]) def disable_interaction(self, *mobjects: Mobject): for mob in mobjects: for sm in mob.get_family(): self.unselectables.append(sm) self.regenerate_selection_search_set() def enable_interaction(self, *mobjects: Mobject): for mob in mobjects: for sm in mob.get_family(): if sm in self.unselectables: self.unselectables.remove(sm) # Functions for keyboard actions def copy_selection(self): names = [] shell = get_ipython() for mob in self.selection: name = str(id(mob)) if shell is None: continue for key, value in shell.user_ns.items(): if mob is value: name = key names.append(name) pyperclip.copy(", ".join(names)) def paste_selection(self): clipboard_str = pyperclip.paste() # Try pasting a mobject try: ids = map(int, clipboard_str.split(",")) mobs = map(self.id_to_mobject, ids) mob_copies = [m.copy() for m in mobs if m is not None] self.clear_selection() self.play(*( FadeIn(mc, run_time=0.5, scale=1.5) for mc in mob_copies )) self.add_to_selection(*mob_copies) return except ValueError: pass # Otherwise, treat as tex or text if set("\\^=+").intersection(clipboard_str): # Proxy to text for LaTeX try: new_mob = Tex(clipboard_str) except LatexError: return else: new_mob = Text(clipboard_str) self.clear_selection() self.add(new_mob) self.add_to_selection(new_mob) def delete_selection(self): self.remove(*self.selection) self.clear_selection() def enable_selection(self): self.is_selecting = True self.add(self.selection_rectangle) self.selection_rectangle.fixed_corner = self.frame.to_fixed_frame_point( self.mouse_point.get_center() ) def gather_new_selection(self): self.is_selecting = False if self.selection_rectangle in self.mobjects: self.remove(self.selection_rectangle) additions = [] for mob in reversed(self.get_selection_search_set()): if self.selection_rectangle.is_touching(mob): additions.append(mob) if self.selection_rectangle.get_arc_length() < 1e-2: break self.toggle_from_selection(*additions) def prepare_grab(self): mp = self.mouse_point.get_center() self.mouse_to_selection = mp - self.selection.get_center() self.is_grabbing = True def prepare_resizing(self, about_corner=False): center = self.selection.get_center() mp = self.mouse_point.get_center() if about_corner: self.scale_about_point = self.selection.get_corner(center - mp) else: self.scale_about_point = center self.scale_ref_vect = mp - self.scale_about_point self.scale_ref_width = self.selection.get_width() self.scale_ref_height = self.selection.get_height() def toggle_color_palette(self): if len(self.selection) == 0: return if self.color_palette not in self.mobjects: self.save_state() self.add(self.color_palette) else: self.remove(self.color_palette) def display_information(self, show=True): if show: self.add(self.information_label) else: self.remove(self.information_label) def group_selection(self): group = self.get_group(*self.selection) self.add(group) self.clear_selection() self.add_to_selection(group) def ungroup_selection(self): pieces = [] for mob in list(self.selection): self.remove(mob) pieces.extend(list(mob)) self.clear_selection() self.add(*pieces) self.add_to_selection(*pieces) def nudge_selection(self, vect: np.ndarray, large: bool = False): nudge = self.selection_nudge_size if large: nudge *= 10 self.selection.shift(nudge * vect) # Key actions def on_key_press(self, symbol: int, modifiers: int) -> None: super().on_key_press(symbol, modifiers) char = chr(symbol) if char == SELECT_KEY and (modifiers & ALL_MODIFIERS) == 0: self.enable_selection() if char == UNSELECT_KEY: self.clear_selection() elif char in GRAB_KEYS and (modifiers & ALL_MODIFIERS) == 0: self.prepare_grab() elif char == RESIZE_KEY and (modifiers & PygletWindowKeys.MOD_SHIFT): self.prepare_resizing(about_corner=((modifiers & PygletWindowKeys.MOD_SHIFT) > 0)) elif symbol == PygletWindowKeys.LSHIFT: if self.window.is_key_pressed(ord("t")): self.prepare_resizing(about_corner=True) elif char == COLOR_KEY and (modifiers & ALL_MODIFIERS) == 0: self.toggle_color_palette() elif char == INFORMATION_KEY and (modifiers & ALL_MODIFIERS) == 0: self.display_information() elif char == "c" and (modifiers & (PygletWindowKeys.MOD_COMMAND | PygletWindowKeys.MOD_CTRL)): self.copy_selection() elif char == "v" and (modifiers & (PygletWindowKeys.MOD_COMMAND | PygletWindowKeys.MOD_CTRL)): self.paste_selection() elif char == "x" and (modifiers & (PygletWindowKeys.MOD_COMMAND | PygletWindowKeys.MOD_CTRL)): self.copy_selection() self.delete_selection() elif symbol == PygletWindowKeys.BACKSPACE: self.delete_selection() elif char == "a" and (modifiers & (PygletWindowKeys.MOD_COMMAND | PygletWindowKeys.MOD_CTRL)): self.clear_selection() self.add_to_selection(*self.mobjects) elif char == "g" and (modifiers & (PygletWindowKeys.MOD_COMMAND | PygletWindowKeys.MOD_CTRL)): self.group_selection() elif char == "g" and (modifiers & (PygletWindowKeys.MOD_COMMAND | PygletWindowKeys.MOD_CTRL | PygletWindowKeys.MOD_SHIFT)): self.ungroup_selection() elif char == "t" and (modifiers & (PygletWindowKeys.MOD_COMMAND | PygletWindowKeys.MOD_CTRL)): self.toggle_selection_mode() elif char == "d" and (modifiers & PygletWindowKeys.MOD_SHIFT): self.copy_frame_positioning() elif char == "c" and (modifiers & PygletWindowKeys.MOD_SHIFT): self.copy_cursor_position() elif symbol in ARROW_SYMBOLS: self.nudge_selection( vect=[LEFT, UP, RIGHT, DOWN][ARROW_SYMBOLS.index(symbol)], large=(modifiers & PygletWindowKeys.MOD_SHIFT), ) # Adding crosshair if char == CURSOR_KEY: if self.crosshair in self.mobjects: self.remove(self.crosshair) else: self.add(self.crosshair) if char == SELECT_KEY: self.add(self.crosshair) # Conditions for saving state if char in [GRAB_KEY, X_GRAB_KEY, Y_GRAB_KEY, RESIZE_KEY]: self.save_state() def on_key_release(self, symbol: int, modifiers: int) -> None: super().on_key_release(symbol, modifiers) if chr(symbol) == SELECT_KEY: self.gather_new_selection() if chr(symbol) in GRAB_KEYS: self.is_grabbing = False elif chr(symbol) == INFORMATION_KEY: self.display_information(False) elif symbol == PygletWindowKeys.LSHIFT and self.window.is_key_pressed(ord(RESIZE_KEY)): self.prepare_resizing(about_corner=False) # Mouse actions def handle_grabbing(self, point: Vect3): diff = point - self.mouse_to_selection if self.window.is_key_pressed(ord(GRAB_KEY)): self.selection.move_to(diff) elif self.window.is_key_pressed(ord(X_GRAB_KEY)): self.selection.set_x(diff[0]) elif self.window.is_key_pressed(ord(Y_GRAB_KEY)): self.selection.set_y(diff[1]) def handle_resizing(self, point: Vect3): if not hasattr(self, "scale_about_point"): return vect = point - self.scale_about_point if self.window.is_key_pressed(PygletWindowKeys.LCTRL): for i in (0, 1): scalar = vect[i] / self.scale_ref_vect[i] self.selection.rescale_to_fit( scalar * [self.scale_ref_width, self.scale_ref_height][i], dim=i, about_point=self.scale_about_point, stretch=True, ) else: scalar = get_norm(vect) / get_norm(self.scale_ref_vect) self.selection.set_width( scalar * self.scale_ref_width, about_point=self.scale_about_point ) def handle_sweeping_selection(self, point: Vect3): mob = self.point_to_mobject( point, search_set=self.get_selection_search_set(), buff=SMALL_BUFF ) if mob is not None: self.add_to_selection(mob) def choose_color(self, point: Vect3): # Search through all mobject on the screen, not just the palette to_search = [ sm for mobject in self.mobjects for sm in mobject.family_members_with_points() if mobject not in self.unselectables ] mob = self.point_to_mobject(point, to_search) if mob is not None: self.selection.set_color(mob.get_color()) self.remove(self.color_palette) def on_mouse_motion(self, point: Vect3, d_point: Vect3) -> None: super().on_mouse_motion(point, d_point) self.crosshair.move_to(self.frame.to_fixed_frame_point(point)) if self.is_grabbing: self.handle_grabbing(point) elif self.window.is_key_pressed(ord(RESIZE_KEY)): self.handle_resizing(point) elif self.window.is_key_pressed(ord(SELECT_KEY)) and self.window.is_key_pressed(PygletWindowKeys.LSHIFT): self.handle_sweeping_selection(point) def on_mouse_drag( self, point: Vect3, d_point: Vect3, buttons: int, modifiers: int ) -> None: super().on_mouse_drag(point, d_point, buttons, modifiers) self.crosshair.move_to(self.frame.to_fixed_frame_point(point)) def on_mouse_release(self, point: Vect3, button: int, mods: int) -> None: super().on_mouse_release(point, button, mods) if self.color_palette in self.mobjects: self.choose_color(point) else: self.clear_selection() # Copying code to recreate state def copy_frame_positioning(self): frame = self.frame center = frame.get_center() height = frame.get_height() angles = frame.get_euler_angles() call = f"reorient(" theta, phi, gamma = (angles / DEG).astype(int) call += f"{theta}, {phi}, {gamma}" if any(center != 0): call += f", {tuple(np.round(center, 2))}" if height != FRAME_HEIGHT: call += ", {:.2f}".format(height) call += ")" pyperclip.copy(call) def copy_cursor_position(self): pyperclip.copy(str(tuple(self.mouse_point.get_center().round(2))))
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/scene/__init__.py
manimlib/scene/__init__.py
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/scene/scene.py
manimlib/scene/scene.py
from __future__ import annotations from collections import OrderedDict import platform import random import time from functools import wraps from contextlib import contextmanager from contextlib import ExitStack import numpy as np from tqdm.auto import tqdm as ProgressDisplay from pyglet.window import key as PygletWindowKeys from manimlib.animation.animation import prepare_animation from manimlib.camera.camera import Camera from manimlib.camera.camera_frame import CameraFrame from manimlib.config import manim_config from manimlib.event_handler import EVENT_DISPATCHER from manimlib.event_handler.event_type import EventType from manimlib.logger import log from manimlib.mobject.mobject import _AnimationBuilder from manimlib.mobject.mobject import Group from manimlib.mobject.mobject import Mobject from manimlib.mobject.mobject import Point from manimlib.mobject.types.vectorized_mobject import VGroup from manimlib.mobject.types.vectorized_mobject import VMobject from manimlib.scene.scene_embed import InteractiveSceneEmbed from manimlib.scene.scene_embed import CheckpointManager from manimlib.scene.scene_file_writer import SceneFileWriter from manimlib.utils.dict_ops import merge_dicts_recursively from manimlib.utils.family_ops import extract_mobject_family_members from manimlib.utils.family_ops import recursive_mobject_remove from manimlib.utils.iterables import batch_by_property from manimlib.utils.sounds import play_sound from manimlib.utils.color import color_to_rgba from manimlib.window import Window from typing import TYPE_CHECKING if TYPE_CHECKING: from typing import Callable, Iterable, TypeVar, Optional from manimlib.typing import Vect3 T = TypeVar('T') from PIL.Image import Image from manimlib.animation.animation import Animation class Scene(object): random_seed: int = 0 pan_sensitivity: float = 0.5 scroll_sensitivity: float = 20 drag_to_pan: bool = True max_num_saved_states: int = 50 default_camera_config: dict = dict() default_file_writer_config: dict = dict() samples = 0 # Euler angles, in degrees default_frame_orientation = (0, 0) def __init__( self, window: Optional[Window] = None, camera_config: dict = dict(), file_writer_config: dict = dict(), skip_animations: bool = False, always_update_mobjects: bool = False, start_at_animation_number: int | None = None, end_at_animation_number: int | None = None, show_animation_progress: bool = False, leave_progress_bars: bool = False, preview_while_skipping: bool = True, presenter_mode: bool = False, default_wait_time: float = 1.0, ): self.skip_animations = skip_animations self.always_update_mobjects = always_update_mobjects self.start_at_animation_number = start_at_animation_number self.end_at_animation_number = end_at_animation_number self.show_animation_progress = show_animation_progress self.leave_progress_bars = leave_progress_bars self.preview_while_skipping = preview_while_skipping self.presenter_mode = presenter_mode self.default_wait_time = default_wait_time self.camera_config = merge_dicts_recursively( manim_config.camera, # Global default self.default_camera_config, # Updated configuration that subclasses may specify camera_config, # Updated configuration from instantiation ) self.file_writer_config = merge_dicts_recursively( manim_config.file_writer, self.default_file_writer_config, file_writer_config, ) self.window = window if self.window: self.window.init_for_scene(self) # Make sure camera and Pyglet window sync self.camera_config["fps"] = 30 # Core state of the scene self.camera: Camera = Camera( window=self.window, samples=self.samples, **self.camera_config ) self.frame: CameraFrame = self.camera.frame self.frame.reorient(*self.default_frame_orientation) self.frame.make_orientation_default() self.file_writer = SceneFileWriter(self, **self.file_writer_config) self.mobjects: list[Mobject] = [self.camera.frame] self.render_groups: list[Mobject] = [] self.id_to_mobject_map: dict[int, Mobject] = dict() self.num_plays: int = 0 self.time: float = 0 self.skip_time: float = 0 self.original_skipping_status: bool = self.skip_animations self.undo_stack = [] self.redo_stack = [] if self.start_at_animation_number is not None: self.skip_animations = True if self.file_writer.has_progress_display(): self.show_animation_progress = False # Items associated with interaction self.mouse_point = Point() self.mouse_drag_point = Point() self.hold_on_wait = self.presenter_mode self.quit_interaction = False # Much nicer to work with deterministic scenes if self.random_seed is not None: random.seed(self.random_seed) np.random.seed(self.random_seed) def __str__(self) -> str: return self.__class__.__name__ def get_window(self) -> Window | None: return self.window def run(self) -> None: self.virtual_animation_start_time: float = 0 self.real_animation_start_time: float = time.time() self.file_writer.begin() self.setup() try: self.construct() self.interact() except EndScene: pass except KeyboardInterrupt: # Get rid keyboard interupt symbols print("", end="\r") self.file_writer.ended_with_interrupt = True self.tear_down() def setup(self) -> None: """ This is meant to be implement by any scenes which are comonly subclassed, and have some common setup involved before the construct method is called. """ pass def construct(self) -> None: # Where all the animation happens # To be implemented in subclasses pass def tear_down(self) -> None: self.stop_skipping() self.file_writer.finish() if self.window: self.window.destroy() self.window = None def interact(self) -> None: """ If there is a window, enter a loop which updates the frame while under the hood calling the pyglet event loop """ if self.window is None: return log.info( "\nTips: Using the keys `d`, `f`, or `z` " + "you can interact with the scene. " + "Press `command + q` or `esc` to quit" ) self.skip_animations = False while not self.is_window_closing(): self.update_frame(1 / self.camera.fps) def embed( self, close_scene_on_exit: bool = True, show_animation_progress: bool = False, ) -> None: if not self.window: # Embed is only relevant for interactive development with a Window return self.show_animation_progress = show_animation_progress self.stop_skipping() self.update_frame(force_draw=True) InteractiveSceneEmbed(self).launch() # End scene when exiting an embed if close_scene_on_exit: raise EndScene() # Only these methods should touch the camera def get_image(self) -> Image: if self.window is not None: self.camera.use_window_fbo(False) self.camera.capture(*self.render_groups) image = self.camera.get_image() if self.window is not None: self.camera.use_window_fbo(True) return image def show(self) -> None: self.update_frame(force_draw=True) self.get_image().show() def update_frame(self, dt: float = 0, force_draw: bool = False) -> None: self.increment_time(dt) self.update_mobjects(dt) if self.skip_animations and not force_draw: return if self.is_window_closing(): raise EndScene() if self.window and dt == 0 and not self.window.has_undrawn_event() and not force_draw: # In this case, there's no need for new rendering, but we # shoudl still listen for new events self.window._window.dispatch_events() return self.camera.capture(*self.render_groups) if self.window and not self.skip_animations: vt = self.time - self.virtual_animation_start_time rt = time.time() - self.real_animation_start_time time.sleep(max(vt - rt, 0)) def emit_frame(self) -> None: if not self.skip_animations: self.file_writer.write_frame(self.camera) # Related to updating def update_mobjects(self, dt: float) -> None: for mobject in self.mobjects: mobject.update(dt) def should_update_mobjects(self) -> bool: return self.always_update_mobjects or any( mob.has_updaters() for mob in self.mobjects ) # Related to time def get_time(self) -> float: return self.time def increment_time(self, dt: float) -> None: self.time += dt # Related to internal mobject organization def get_top_level_mobjects(self) -> list[Mobject]: # Return only those which are not in the family # of another mobject from the scene mobjects = self.get_mobjects() families = [m.get_family() for m in mobjects] def is_top_level(mobject): num_families = sum([ (mobject in family) for family in families ]) return num_families == 1 return list(filter(is_top_level, mobjects)) def get_mobject_family_members(self) -> list[Mobject]: return extract_mobject_family_members(self.mobjects) def assemble_render_groups(self): """ Rendering can be more efficient when mobjects of the same type are grouped together, so this function creates Groups of all clusters of adjacent Mobjects in the scene """ batches = batch_by_property( self.mobjects, lambda m: str(type(m)) + str(m.get_shader_wrapper(self.camera.ctx).get_id()) + str(m.z_index) ) for group in self.render_groups: group.clear() self.render_groups = [ batch[0].get_group_class()(*batch) for batch, key in batches ] @staticmethod def affects_mobject_list(func: Callable[..., T]) -> Callable[..., T]: @wraps(func) def wrapper(self, *args, **kwargs): func(self, *args, **kwargs) self.assemble_render_groups() return self return wrapper @affects_mobject_list def add(self, *new_mobjects: Mobject): """ Mobjects will be displayed, from background to foreground in the order with which they are added. """ self.remove(*new_mobjects) self.mobjects += new_mobjects # Reorder based on z_index id_to_scene_order = {id(m): idx for idx, m in enumerate(self.mobjects)} self.mobjects.sort(key=lambda m: (m.z_index, id_to_scene_order[id(m)])) self.id_to_mobject_map.update({ id(sm): sm for m in new_mobjects for sm in m.get_family() }) return self def add_mobjects_among(self, values: Iterable): """ This is meant mostly for quick prototyping, e.g. to add all mobjects defined up to a point, call self.add_mobjects_among(locals().values()) """ self.add(*filter( lambda m: isinstance(m, Mobject), values )) return self @affects_mobject_list def replace(self, mobject: Mobject, *replacements: Mobject): if mobject in self.mobjects: index = self.mobjects.index(mobject) self.mobjects = [ *self.mobjects[:index], *replacements, *self.mobjects[index + 1:] ] return self @affects_mobject_list def remove(self, *mobjects_to_remove: Mobject): """ Removes anything in mobjects from scenes mobject list, but in the event that one of the items to be removed is a member of the family of an item in mobject_list, the other family members are added back into the list. For example, if the scene includes Group(m1, m2, m3), and we call scene.remove(m1), the desired behavior is for the scene to then include m2 and m3 (ungrouped). """ to_remove = set(extract_mobject_family_members(mobjects_to_remove)) new_mobjects, _ = recursive_mobject_remove(self.mobjects, to_remove) self.mobjects = new_mobjects @affects_mobject_list def remove_all_except(self, *mobjects_to_keep : Mobject): self.clear() self.add(*mobjects_to_keep) def bring_to_front(self, *mobjects: Mobject): self.add(*mobjects) return self @affects_mobject_list def bring_to_back(self, *mobjects: Mobject): self.remove(*mobjects) self.mobjects = list(mobjects) + self.mobjects return self @affects_mobject_list def clear(self): self.mobjects = [] return self def get_mobjects(self) -> list[Mobject]: return list(self.mobjects) def get_mobject_copies(self) -> list[Mobject]: return [m.copy() for m in self.mobjects] def point_to_mobject( self, point: np.ndarray, search_set: Iterable[Mobject] | None = None, buff: float = 0 ) -> Mobject | None: """ E.g. if clicking on the scene, this returns the top layer mobject under a given point """ if search_set is None: search_set = self.mobjects for mobject in reversed(search_set): if mobject.is_point_touching(point, buff=buff): return mobject return None def get_group(self, *mobjects): if all(isinstance(m, VMobject) for m in mobjects): return VGroup(*mobjects) else: return Group(*mobjects) def id_to_mobject(self, id_value): return self.id_to_mobject_map[id_value] def ids_to_group(self, *id_values): return self.get_group(*filter( lambda x: x is not None, map(self.id_to_mobject, id_values) )) def i2g(self, *id_values): return self.ids_to_group(*id_values) def i2m(self, id_value): return self.id_to_mobject(id_value) # Related to skipping def update_skipping_status(self) -> None: if self.start_at_animation_number is not None: if self.num_plays == self.start_at_animation_number: self.skip_time = self.time if not self.original_skipping_status: self.stop_skipping() if self.end_at_animation_number is not None: if self.num_plays >= self.end_at_animation_number: raise EndScene() def stop_skipping(self) -> None: self.virtual_animation_start_time = self.time self.real_animation_start_time = time.time() self.skip_animations = False # Methods associated with running animations def get_time_progression( self, run_time: float, n_iterations: int | None = None, desc: str = "", override_skip_animations: bool = False ) -> list[float] | np.ndarray | ProgressDisplay: if self.skip_animations and not override_skip_animations: return [run_time] times = np.arange(0, run_time, 1 / self.camera.fps) + 1 / self.camera.fps self.file_writer.set_progress_display_description(sub_desc=desc) if self.show_animation_progress: return ProgressDisplay( times, total=n_iterations, leave=self.leave_progress_bars, ascii=True if platform.system() == 'Windows' else None, desc=desc, bar_format="{l_bar} {n_fmt:3}/{total_fmt:3} {rate_fmt}{postfix}", ) else: return times def get_run_time(self, animations: Iterable[Animation]) -> float: return np.max([animation.get_run_time() for animation in animations]) def get_animation_time_progression( self, animations: Iterable[Animation] ) -> list[float] | np.ndarray | ProgressDisplay: animations = list(animations) run_time = self.get_run_time(animations) description = f"{self.num_plays} {animations[0]}" if len(animations) > 1: description += ", etc." time_progression = self.get_time_progression(run_time, desc=description) return time_progression def get_wait_time_progression( self, duration: float, stop_condition: Callable[[], bool] | None = None ) -> list[float] | np.ndarray | ProgressDisplay: kw = {"desc": f"{self.num_plays} Waiting"} if stop_condition is not None: kw["n_iterations"] = -1 # So it doesn't show % progress kw["override_skip_animations"] = True return self.get_time_progression(duration, **kw) def pre_play(self): if self.presenter_mode and self.num_plays == 0: self.hold_loop() self.update_skipping_status() if not self.skip_animations: self.file_writer.begin_animation() if self.window: self.virtual_animation_start_time = self.time self.real_animation_start_time = time.time() def post_play(self): if not self.skip_animations: self.file_writer.end_animation() if self.preview_while_skipping and self.skip_animations and self.window is not None: # Show some quick frames along the way self.update_frame(dt=0, force_draw=True) self.num_plays += 1 def begin_animations(self, animations: Iterable[Animation]) -> None: all_mobjects = set(self.get_mobject_family_members()) for animation in animations: animation.begin() # Anything animated that's not already in the # scene gets added to the scene. Note, for # animated mobjects that are in the family of # those on screen, this can result in a restructuring # of the scene.mobjects list, which is usually desired. if animation.mobject not in all_mobjects: self.add(animation.mobject) all_mobjects = all_mobjects.union(animation.mobject.get_family()) def progress_through_animations(self, animations: Iterable[Animation]) -> None: last_t = 0 for t in self.get_animation_time_progression(animations): dt = t - last_t last_t = t for animation in animations: animation.update_mobjects(dt) alpha = t / animation.run_time animation.interpolate(alpha) self.update_frame(dt) self.emit_frame() def finish_animations(self, animations: Iterable[Animation]) -> None: for animation in animations: animation.finish() animation.clean_up_from_scene(self) if self.skip_animations: self.update_mobjects(self.get_run_time(animations)) else: self.update_mobjects(0) @affects_mobject_list def play( self, *proto_animations: Animation | _AnimationBuilder, run_time: float | None = None, rate_func: Callable[[float], float] | None = None, lag_ratio: float | None = None, ) -> None: if len(proto_animations) == 0: log.warning("Called Scene.play with no animations") return animations = list(map(prepare_animation, proto_animations)) for anim in animations: anim.update_rate_info(run_time, rate_func, lag_ratio) self.pre_play() self.begin_animations(animations) self.progress_through_animations(animations) self.finish_animations(animations) self.post_play() def wait( self, duration: Optional[float] = None, stop_condition: Callable[[], bool] = None, note: str = None, ignore_presenter_mode: bool = False ): if duration is None: duration = self.default_wait_time self.pre_play() self.update_mobjects(dt=0) # Any problems with this? if self.presenter_mode and not self.skip_animations and not ignore_presenter_mode: if note: log.info(note) self.hold_loop() else: time_progression = self.get_wait_time_progression(duration, stop_condition) last_t = 0 for t in time_progression: dt = t - last_t last_t = t self.update_frame(dt) self.emit_frame() if stop_condition is not None and stop_condition(): break self.post_play() def hold_loop(self): while self.hold_on_wait: self.update_frame(dt=1 / self.camera.fps) self.hold_on_wait = True def wait_until( self, stop_condition: Callable[[], bool], max_time: float = 60 ): self.wait(max_time, stop_condition=stop_condition) def force_skipping(self): self.original_skipping_status = self.skip_animations self.skip_animations = True return self def revert_to_original_skipping_status(self): if hasattr(self, "original_skipping_status"): self.skip_animations = self.original_skipping_status return self def add_sound( self, sound_file: str, time_offset: float = 0, gain: float | None = None, gain_to_background: float | None = None ): if self.skip_animations: return time = self.get_time() + time_offset self.file_writer.add_sound(sound_file, time, gain, gain_to_background) # Helpers for interactive development def get_state(self) -> SceneState: return SceneState(self) @affects_mobject_list def restore_state(self, scene_state: SceneState): scene_state.restore_scene(self) def save_state(self) -> None: state = self.get_state() if self.undo_stack and state.mobjects_match(self.undo_stack[-1]): return self.redo_stack = [] self.undo_stack.append(state) if len(self.undo_stack) > self.max_num_saved_states: self.undo_stack.pop(0) def undo(self): if self.undo_stack: self.redo_stack.append(self.get_state()) self.restore_state(self.undo_stack.pop()) def redo(self): if self.redo_stack: self.undo_stack.append(self.get_state()) self.restore_state(self.redo_stack.pop()) @contextmanager def temp_skip(self): prev_status = self.skip_animations self.skip_animations = True try: yield finally: if not prev_status: self.stop_skipping() @contextmanager def temp_progress_bar(self): prev_progress = self.show_animation_progress self.show_animation_progress = True try: yield finally: self.show_animation_progress = prev_progress @contextmanager def temp_record(self): self.camera.use_window_fbo(False) self.file_writer.begin_insert() try: yield finally: self.file_writer.end_insert() self.camera.use_window_fbo(True) def temp_config_change(self, skip=False, record=False, progress_bar=False): stack = ExitStack() if skip: stack.enter_context(self.temp_skip()) if record: stack.enter_context(self.temp_record()) if progress_bar: stack.enter_context(self.temp_progress_bar()) return stack def is_window_closing(self): return self.window and (self.window.is_closing or self.quit_interaction) # Event handling def set_floor_plane(self, plane: str = "xy"): if plane == "xy": self.frame.set_euler_axes("zxz") elif plane == "xz": self.frame.set_euler_axes("zxy") else: raise Exception("Only `xz` and `xy` are valid floor planes") def on_mouse_motion( self, point: Vect3, d_point: Vect3 ) -> None: assert self.window is not None self.mouse_point.move_to(point) event_data = {"point": point, "d_point": d_point} propagate_event = EVENT_DISPATCHER.dispatch(EventType.MouseMotionEvent, **event_data) if propagate_event is not None and propagate_event is False: return frame = self.camera.frame # Handle perspective changes if self.window.is_key_pressed(ord(manim_config.key_bindings.pan_3d)): ff_d_point = frame.to_fixed_frame_point(d_point, relative=True) ff_d_point *= self.pan_sensitivity frame.increment_theta(-ff_d_point[0]) frame.increment_phi(ff_d_point[1]) # Handle frame movements elif self.window.is_key_pressed(ord(manim_config.key_bindings.pan)): frame.shift(-d_point) def on_mouse_drag( self, point: Vect3, d_point: Vect3, buttons: int, modifiers: int ) -> None: self.mouse_drag_point.move_to(point) if self.drag_to_pan: self.frame.shift(-d_point) event_data = {"point": point, "d_point": d_point, "buttons": buttons, "modifiers": modifiers} propagate_event = EVENT_DISPATCHER.dispatch(EventType.MouseDragEvent, **event_data) if propagate_event is not None and propagate_event is False: return def on_mouse_press( self, point: Vect3, button: int, mods: int ) -> None: self.mouse_drag_point.move_to(point) event_data = {"point": point, "button": button, "mods": mods} propagate_event = EVENT_DISPATCHER.dispatch(EventType.MousePressEvent, **event_data) if propagate_event is not None and propagate_event is False: return def on_mouse_release( self, point: Vect3, button: int, mods: int ) -> None: event_data = {"point": point, "button": button, "mods": mods} propagate_event = EVENT_DISPATCHER.dispatch(EventType.MouseReleaseEvent, **event_data) if propagate_event is not None and propagate_event is False: return def on_mouse_scroll( self, point: Vect3, offset: Vect3, x_pixel_offset: float, y_pixel_offset: float ) -> None: event_data = {"point": point, "offset": offset} propagate_event = EVENT_DISPATCHER.dispatch(EventType.MouseScrollEvent, **event_data) if propagate_event is not None and propagate_event is False: return rel_offset = y_pixel_offset / self.camera.get_pixel_height() self.frame.scale( 1 - self.scroll_sensitivity * rel_offset, about_point=point ) def on_key_release( self, symbol: int, modifiers: int ) -> None: event_data = {"symbol": symbol, "modifiers": modifiers} propagate_event = EVENT_DISPATCHER.dispatch(EventType.KeyReleaseEvent, **event_data) if propagate_event is not None and propagate_event is False: return def on_key_press( self, symbol: int, modifiers: int ) -> None: try: char = chr(symbol) except OverflowError: log.warning("The value of the pressed key is too large.") return event_data = {"symbol": symbol, "modifiers": modifiers} propagate_event = EVENT_DISPATCHER.dispatch(EventType.KeyPressEvent, **event_data) if propagate_event is not None and propagate_event is False: return if char == manim_config.key_bindings.reset: self.play(self.camera.frame.animate.to_default_state()) elif char == "z" and (modifiers & (PygletWindowKeys.MOD_COMMAND | PygletWindowKeys.MOD_CTRL)): self.undo() elif char == "z" and (modifiers & (PygletWindowKeys.MOD_COMMAND | PygletWindowKeys.MOD_CTRL | PygletWindowKeys.MOD_SHIFT)): self.redo() # command + q elif char == manim_config.key_bindings.quit and (modifiers & (PygletWindowKeys.MOD_COMMAND | PygletWindowKeys.MOD_CTRL)): self.quit_interaction = True # Space or right arrow elif char == " " or symbol == PygletWindowKeys.RIGHT: self.hold_on_wait = False def on_resize(self, width: int, height: int) -> None: pass def on_show(self) -> None: pass def on_hide(self) -> None: pass def on_close(self) -> None: pass def focus(self) -> None: """ Puts focus on the ManimGL window. """ if not self.window: return self.window.focus() def set_background_color(self, background_color, background_opacity=1) -> None: self.camera.background_rgba = list(color_to_rgba( background_color, background_opacity )) class SceneState(): def __init__(self, scene: Scene, ignore: list[Mobject] | None = None): self.time = scene.time self.num_plays = scene.num_plays self.mobjects_to_copies = OrderedDict.fromkeys(scene.mobjects) if ignore: for mob in ignore: self.mobjects_to_copies.pop(mob, None) last_m2c = scene.undo_stack[-1].mobjects_to_copies if scene.undo_stack else dict() for mob in self.mobjects_to_copies: # If it hasn't changed since the last state, just point to the # same copy as before if mob in last_m2c and last_m2c[mob].looks_identical(mob): self.mobjects_to_copies[mob] = last_m2c[mob] else: self.mobjects_to_copies[mob] = mob.copy() def __eq__(self, state: SceneState): return all(( self.time == state.time, self.num_plays == state.num_plays, self.mobjects_to_copies == state.mobjects_to_copies )) def mobjects_match(self, state: SceneState): return self.mobjects_to_copies == state.mobjects_to_copies def n_changes(self, state: SceneState): m2c = state.mobjects_to_copies return sum( 1 - int(mob in m2c and mob.looks_identical(m2c[mob])) for mob in self.mobjects_to_copies ) def restore_scene(self, scene: Scene): scene.time = self.time scene.num_plays = self.num_plays scene.mobjects = [ mob.become(mob_copy) for mob, mob_copy in self.mobjects_to_copies.items() ] class EndScene(Exception): pass class ThreeDScene(Scene): samples = 4 default_frame_orientation = (-30, 70) always_depth_test = True def add(self, *mobjects: Mobject, set_depth_test: bool = True, perp_stroke: bool = True): for mob in mobjects: if set_depth_test and not mob.is_fixed_in_frame() and self.always_depth_test: mob.apply_depth_test() if isinstance(mob, VMobject) and mob.has_stroke() and perp_stroke: mob.set_flat_stroke(False) super().add(*mobjects)
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/scene/scene_file_writer.py
manimlib/scene/scene_file_writer.py
from __future__ import annotations import os import platform import shutil import subprocess as sp import sys import numpy as np from pydub import AudioSegment from tqdm.auto import tqdm as ProgressDisplay from pathlib import Path from manimlib.logger import log from manimlib.mobject.mobject import Mobject from manimlib.utils.file_ops import guarantee_existence from manimlib.utils.sounds import get_full_sound_file_path from typing import TYPE_CHECKING if TYPE_CHECKING: from PIL.Image import Image from manimlib.camera.camera import Camera from manimlib.scene.scene import Scene class SceneFileWriter(object): def __init__( self, scene: Scene, write_to_movie: bool = False, subdivide_output: bool = False, png_mode: str = "RGBA", save_last_frame: bool = False, movie_file_extension: str = ".mp4", # Where should this be written output_directory: str = ".", file_name: str | None = None, open_file_upon_completion: bool = False, show_file_location_upon_completion: bool = False, quiet: bool = False, total_frames: int = 0, progress_description_len: int = 40, # Name of the binary used for ffmpeg ffmpeg_bin: str = "ffmpeg", video_codec: str = "libx264", pixel_format: str = "yuv420p", saturation: float = 1.0, gamma: float = 1.0, ): self.scene: Scene = scene self.write_to_movie = write_to_movie self.subdivide_output = subdivide_output self.png_mode = png_mode self.save_last_frame = save_last_frame self.movie_file_extension = movie_file_extension self.output_directory = output_directory self.file_name = file_name self.open_file_upon_completion = open_file_upon_completion self.show_file_location_upon_completion = show_file_location_upon_completion self.quiet = quiet self.total_frames = total_frames self.progress_description_len = progress_description_len self.ffmpeg_bin = ffmpeg_bin self.video_codec = video_codec self.pixel_format = pixel_format self.saturation = saturation self.gamma = gamma # State during file writing self.writing_process: sp.Popen | None = None self.progress_display: ProgressDisplay | None = None self.ended_with_interrupt: bool = False self.init_output_directories() self.init_audio() # Output directories and files def init_output_directories(self) -> None: if self.save_last_frame: self.image_file_path = self.init_image_file_path() if self.write_to_movie: self.movie_file_path = self.init_movie_file_path() if self.subdivide_output: self.partial_movie_directory = self.init_partial_movie_directory() def init_image_file_path(self) -> Path: return self.get_output_file_rootname().with_suffix(".png") def init_movie_file_path(self) -> Path: return self.get_output_file_rootname().with_suffix(self.movie_file_extension) def init_partial_movie_directory(self): return guarantee_existence(self.get_output_file_rootname()) def get_output_file_rootname(self) -> Path: return Path( guarantee_existence(self.output_directory), self.get_output_file_name() ) def get_output_file_name(self) -> str: if self.file_name: return self.file_name # Otherwise, use the name of the scene, potentially # appending animation numbers name = str(self.scene) saan = self.scene.start_at_animation_number eaan = self.scene.end_at_animation_number if saan is not None: name += f"_{saan}" if eaan is not None: name += f"_{eaan}" return name # Directory getters def get_image_file_path(self) -> str: return self.image_file_path def get_next_partial_movie_path(self) -> str: result = Path(self.partial_movie_directory, f"{self.scene.num_plays:05}") return result.with_suffix(self.movie_file_extension) def get_movie_file_path(self) -> str: return self.movie_file_path # Sound def init_audio(self) -> None: self.includes_sound: bool = False def create_audio_segment(self) -> None: self.audio_segment = AudioSegment.silent() def add_audio_segment( self, new_segment: AudioSegment, time: float | None = None, gain_to_background: float | None = None ) -> None: if not self.includes_sound: self.includes_sound = True self.create_audio_segment() segment = self.audio_segment curr_end = segment.duration_seconds if time is None: time = curr_end if time < 0: raise Exception("Adding sound at timestamp < 0") new_end = time + new_segment.duration_seconds diff = new_end - curr_end if diff > 0: segment = segment.append( AudioSegment.silent(int(np.ceil(diff * 1000))), crossfade=0, ) self.audio_segment = segment.overlay( new_segment, position=int(1000 * time), gain_during_overlay=gain_to_background, ) def add_sound( self, sound_file: str, time: float | None = None, gain: float | None = None, gain_to_background: float | None = None ) -> None: file_path = get_full_sound_file_path(sound_file) new_segment = AudioSegment.from_file(file_path) if gain: new_segment = new_segment.apply_gain(gain) self.add_audio_segment(new_segment, time, gain_to_background) # Writers def begin(self) -> None: if not self.subdivide_output and self.write_to_movie: self.open_movie_pipe(self.get_movie_file_path()) def begin_animation(self) -> None: if self.subdivide_output and self.write_to_movie: self.open_movie_pipe(self.get_next_partial_movie_path()) def end_animation(self) -> None: if self.subdivide_output and self.write_to_movie: self.close_movie_pipe() def finish(self) -> None: if not self.subdivide_output and self.write_to_movie: self.close_movie_pipe() if self.includes_sound: self.add_sound_to_video() self.print_file_ready_message(self.get_movie_file_path()) if self.save_last_frame: self.scene.update_frame(force_draw=True) self.save_final_image(self.scene.get_image()) if self.should_open_file(): self.open_file() def open_movie_pipe(self, file_path: str) -> None: stem, ext = os.path.splitext(file_path) self.final_file_path = file_path self.temp_file_path = stem + "_temp" + ext fps = self.scene.camera.fps width, height = self.scene.camera.get_pixel_shape() vf_arg = 'vflip' vf_arg += f',eq=saturation={self.saturation}:gamma={self.gamma}' command = [ self.ffmpeg_bin, '-y', # overwrite output file if it exists '-f', 'rawvideo', '-s', f'{width}x{height}', # size of one frame '-pix_fmt', 'rgba', '-r', str(fps), # frames per second '-i', '-', # The input comes from a pipe '-vf', vf_arg, '-an', # Tells ffmpeg not to expect any audio '-loglevel', 'error', ] if self.video_codec: command += ['-vcodec', self.video_codec] if self.pixel_format: command += ['-pix_fmt', self.pixel_format] command += [self.temp_file_path] self.writing_process = sp.Popen(command, stdin=sp.PIPE) if not self.quiet: self.progress_display = ProgressDisplay( range(self.total_frames), leave=False, ascii=True if platform.system() == 'Windows' else None, dynamic_ncols=True, ) self.set_progress_display_description() def use_fast_encoding(self): self.video_codec = "libx264rgb" self.pixel_format = "rgb32" def get_insert_file_path(self, index: int) -> Path: movie_path = Path(self.get_movie_file_path()) scene_name = movie_path.stem insert_dir = Path(movie_path.parent, "inserts") guarantee_existence(insert_dir) return Path(insert_dir, f"{scene_name}_{index}").with_suffix(self.movie_file_extension) def begin_insert(self): # Begin writing process self.write_to_movie = True self.init_output_directories() index = 0 while (insert_path := self.get_insert_file_path(index)).exists(): index += 1 self.inserted_file_path = insert_path self.open_movie_pipe(self.inserted_file_path) def end_insert(self): self.close_movie_pipe() self.write_to_movie = False self.print_file_ready_message(self.inserted_file_path) def has_progress_display(self): return self.progress_display is not None def set_progress_display_description(self, file: str = "", sub_desc: str = "") -> None: if self.progress_display is None: return desc_len = self.progress_description_len if not file: file = os.path.split(self.get_movie_file_path())[1] full_desc = f"{file} {sub_desc}" if len(full_desc) > desc_len: full_desc = full_desc[:desc_len - 3] + "..." else: full_desc += " " * (desc_len - len(full_desc)) self.progress_display.set_description(full_desc) def write_frame(self, camera: Camera) -> None: if self.write_to_movie: raw_bytes = camera.get_raw_fbo_data() self.writing_process.stdin.write(raw_bytes) if self.progress_display is not None: self.progress_display.update() def close_movie_pipe(self) -> None: self.writing_process.stdin.close() self.writing_process.wait() self.writing_process.terminate() if self.progress_display is not None: self.progress_display.close() if not self.ended_with_interrupt: shutil.move(self.temp_file_path, self.final_file_path) else: self.movie_file_path = self.temp_file_path def add_sound_to_video(self) -> None: movie_file_path = self.get_movie_file_path() stem, ext = os.path.splitext(movie_file_path) sound_file_path = stem + ".wav" # Makes sure sound file length will match video file self.add_audio_segment(AudioSegment.silent(0)) self.audio_segment.export( sound_file_path, bitrate='312k', ) temp_file_path = stem + "_temp" + ext commands = [ self.ffmpeg_bin, "-i", movie_file_path, "-i", sound_file_path, '-y', # overwrite output file if it exists "-c:v", "copy", "-c:a", "aac", "-b:a", "320k", # select video stream from first file "-map", "0:v:0", # select audio stream from second file "-map", "1:a:0", '-loglevel', 'error', # "-shortest", temp_file_path, ] sp.call(commands) shutil.move(temp_file_path, movie_file_path) os.remove(sound_file_path) def save_final_image(self, image: Image) -> None: file_path = self.get_image_file_path() image.save(file_path) self.print_file_ready_message(file_path) def print_file_ready_message(self, file_path: str) -> None: if not self.quiet: log.info(f"File ready at {file_path}") def should_open_file(self) -> bool: return any([ self.show_file_location_upon_completion, self.open_file_upon_completion, ]) def open_file(self) -> None: if self.quiet: curr_stdout = sys.stdout sys.stdout = open(os.devnull, "w") current_os = platform.system() file_paths = [] if self.save_last_frame: file_paths.append(self.get_image_file_path()) if self.write_to_movie: file_paths.append(self.get_movie_file_path()) for file_path in file_paths: if current_os == "Windows": os.startfile(file_path) else: commands = [] if current_os == "Linux": commands.append("xdg-open") elif current_os.startswith("CYGWIN"): commands.append("cygstart") else: # Assume macOS commands.append("open") if self.show_file_location_upon_completion: commands.append("-R") commands.append(file_path) FNULL = open(os.devnull, 'w') sp.call(commands, stdout=FNULL, stderr=sp.STDOUT) FNULL.close() if self.quiet: sys.stdout.close() sys.stdout = curr_stdout
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/utils/file_ops.py
manimlib/utils/file_ops.py
from __future__ import annotations import os from pathlib import Path import hashlib import numpy as np import validators import urllib.request import manimlib.utils.directories from manimlib.utils.simple_functions import hash_string from typing import TYPE_CHECKING if TYPE_CHECKING: from typing import Iterable def guarantee_existence(path: str | Path) -> Path: path = Path(path) path.mkdir(parents=True, exist_ok=True) return path.absolute() def find_file( file_name: str, directories: Iterable[str] | None = None, extensions: Iterable[str] | None = None ) -> Path: # Check if this is a file online first, and if so, download # it to a temporary directory if validators.url(file_name): suffix = Path(file_name).suffix file_hash = hash_string(file_name) folder = manimlib.utils.directories.get_downloads_dir() path = Path(folder, file_hash).with_suffix(suffix) urllib.request.urlretrieve(file_name, path) return path # Check if what was passed in is already a valid path to a file if os.path.exists(file_name): return Path(file_name) # Otherwise look in local file system directories = directories or [""] extensions = extensions or [""] possible_paths = ( Path(directory, file_name + extension) for directory in directories for extension in extensions ) for path in possible_paths: if path.exists(): return path raise IOError(f"{file_name} not Found")
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/utils/tex_file_writing.py
manimlib/utils/tex_file_writing.py
from __future__ import annotations import os import re import yaml import subprocess from functools import lru_cache from pathlib import Path import tempfile from manimlib.utils.cache import cache_on_disk from manimlib.config import manim_config from manimlib.config import get_manim_dir from manimlib.logger import log from manimlib.utils.simple_functions import hash_string def get_tex_template_config(template_name: str) -> dict[str, str]: name = template_name.replace(" ", "_").lower() template_path = os.path.join(get_manim_dir(), "manimlib", "tex_templates.yml") with open(template_path, encoding="utf-8") as tex_templates_file: templates_dict = yaml.safe_load(tex_templates_file) if name not in templates_dict: log.warning(f"Cannot recognize template {name}, falling back to 'default'.") name = "default" return templates_dict[name] @lru_cache def get_tex_config(template: str = "") -> tuple[str, str]: """ Returns a compiler and preamble to use for rendering LaTeX """ template = template or manim_config.tex.template config = get_tex_template_config(template) return config["compiler"], config["preamble"] def get_full_tex(content: str, preamble: str = ""): return "\n\n".join(( "\\documentclass[preview]{standalone}", preamble, "\\begin{document}", content, "\\end{document}" )) + "\n" @lru_cache(maxsize=128) def latex_to_svg( latex: str, template: str = "", additional_preamble: str = "", short_tex: str = "", show_message_during_execution: bool = True, ) -> str: """Convert LaTeX string to SVG string. Args: latex: LaTeX source code template: Path to a template LaTeX file additional_preamble: String including any added "\\usepackage{...}" style imports Returns: str: SVG source code Raises: LatexError: If LaTeX compilation fails NotImplementedError: If compiler is not supported """ if show_message_during_execution: message = f"Writing {(short_tex or latex)[:70]}..." else: message = "" compiler, preamble = get_tex_config(template) preamble = "\n".join([preamble, additional_preamble]) full_tex = get_full_tex(latex, preamble) return full_tex_to_svg(full_tex, compiler, message) @cache_on_disk def full_tex_to_svg(full_tex: str, compiler: str = "latex", message: str = ""): if message: print(message, end="\r") if compiler == "latex": dvi_ext = ".dvi" elif compiler == "xelatex": dvi_ext = ".xdv" else: raise NotImplementedError(f"Compiler '{compiler}' is not implemented") # Write intermediate files to a temporary directory with tempfile.TemporaryDirectory() as temp_dir: tex_path = Path(temp_dir, "working").with_suffix(".tex") dvi_path = tex_path.with_suffix(dvi_ext) # Write tex file tex_path.write_text(full_tex) # Run latex compiler process = subprocess.run( [ compiler, *(['-no-pdf'] if compiler == "xelatex" else []), "-interaction=batchmode", "-halt-on-error", f"-output-directory={temp_dir}", tex_path ], capture_output=True, text=True ) if process.returncode != 0: # Handle error error_str = "" log_path = tex_path.with_suffix(".log") if log_path.exists(): content = log_path.read_text() error_match = re.search(r"(?<=\n! ).*\n.*\n", content) if error_match: error_str = error_match.group() raise LatexError(error_str or "LaTeX compilation failed") # Run dvisvgm and capture output directly process = subprocess.run( [ "dvisvgm", dvi_path, "-n", # no fonts "-v", "0", # quiet "--stdout", # output to stdout instead of file ], capture_output=True ) # Return SVG string result = process.stdout.decode('utf-8') if message: print(" " * len(message), end="\r") return result class LatexError(Exception): pass
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/utils/paths.py
manimlib/utils/paths.py
from __future__ import annotations import math import numpy as np from manimlib.constants import OUT from manimlib.utils.bezier import interpolate from manimlib.utils.space_ops import get_norm from manimlib.utils.space_ops import rotation_matrix_transpose from typing import TYPE_CHECKING if TYPE_CHECKING: from typing import Callable from manimlib.typing import Vect3, Vect3Array STRAIGHT_PATH_THRESHOLD = 0.01 def straight_path( start_points: np.ndarray, end_points: np.ndarray, alpha: float ) -> np.ndarray: """ Same function as interpolate, but renamed to reflect intent of being used to determine how a set of points move to another set. For instance, it should be a specific case of path_along_arc """ return interpolate(start_points, end_points, alpha) def path_along_arc( arc_angle: float, axis: Vect3 = OUT ) -> Callable[[Vect3Array, Vect3Array, float], Vect3Array]: """ If vect is vector from start to end, [vect[:,1], -vect[:,0]] is perpendicular to vect in the left direction. """ if abs(arc_angle) < STRAIGHT_PATH_THRESHOLD: return straight_path if get_norm(axis) == 0: axis = OUT unit_axis = axis / get_norm(axis) def path(start_points, end_points, alpha): vects = end_points - start_points centers = start_points + 0.5 * vects if arc_angle != np.pi: centers += np.cross(unit_axis, vects / 2.0) / math.tan(arc_angle / 2) rot_matrix_T = rotation_matrix_transpose(alpha * arc_angle, unit_axis) return centers + np.dot(start_points - centers, rot_matrix_T) return path def clockwise_path() -> Callable[[Vect3Array, Vect3Array, float], Vect3Array]: return path_along_arc(-np.pi) def counterclockwise_path() -> Callable[[Vect3Array, Vect3Array, float], Vect3Array]: return path_along_arc(np.pi)
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/utils/tex.py
manimlib/utils/tex.py
from __future__ import annotations import re from functools import lru_cache from manimlib.utils.tex_to_symbol_count import TEX_TO_SYMBOL_COUNT @lru_cache def num_tex_symbols(tex: str) -> int: tex = remove_tex_environments(tex) commands_pattern = r""" (?P<sqrt>\\sqrt\[[0-9]+\])| # Special sqrt with number (?P<escaped_brace>\\[{}])| # Escaped braces (?P<cmd>\\[a-zA-Z!,-/:;<>]+) # Regular commands """ total = 0 pos = 0 for match in re.finditer(commands_pattern, tex, re.VERBOSE): # Count normal characters up to this command total += sum(1 for c in tex[pos:match.start()] if c not in "^{} \n\t_$\\&") if match.group("sqrt"): total += len(match.group()) - 5 elif match.group("escaped_brace"): total += 1 # Count escaped brace as one symbol else: total += TEX_TO_SYMBOL_COUNT.get(match.group(), 1) pos = match.end() # Count remaining characters total += sum(1 for c in tex[pos:] if c not in "^{} \n\t_$\\&") return total def remove_tex_environments(tex: str) -> str: # Handle \phantom{...} with any content tex = re.sub(r"\\phantom\{[^}]*\}", "", tex) # Handle other environment commands tex = re.sub(r"\\(begin|end)(\{\w+\})?(\{\w+\})?(\[\w+\])?", "", tex) return tex
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/utils/debug.py
manimlib/utils/debug.py
from __future__ import annotations from manimlib.constants import BLACK from manimlib.logger import log from manimlib.mobject.numbers import Integer from manimlib.mobject.types.vectorized_mobject import VGroup from typing import TYPE_CHECKING if TYPE_CHECKING: from manimlib.mobject.mobject import Mobject def print_family(mobject: Mobject, n_tabs: int = 0) -> None: """For debugging purposes""" log.debug("\t" * n_tabs + str(mobject) + " " + str(id(mobject))) for submob in mobject.submobjects: print_family(submob, n_tabs + 1) def index_labels( mobject: Mobject, label_height: float = 0.15 ) -> VGroup: labels = VGroup() for n, submob in enumerate(mobject): label = Integer(n) label.set_height(label_height) label.move_to(submob) label.set_backstroke(BLACK, 5) labels.add(label) return labels
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/utils/iterables.py
manimlib/utils/iterables.py
from __future__ import annotations from colour import Color import numpy as np import random from typing import TYPE_CHECKING if TYPE_CHECKING: from typing import Callable, Iterable, Sequence, TypeVar T = TypeVar("T") S = TypeVar("S") def remove_list_redundancies(lst: Sequence[T]) -> list[T]: """ Remove duplicate elements while preserving order. Keeps the last occurrence of each element """ return list(reversed(dict.fromkeys(reversed(lst)))) def list_update(l1: Iterable[T], l2: Iterable[T]) -> list[T]: """ Used instead of list(set(l1).update(l2)) to maintain order, making sure duplicates are removed from l1, not l2. """ return remove_list_redundancies([*l1, *l2]) def list_difference_update(l1: Iterable[T], l2: Iterable[T]) -> list[T]: return [e for e in l1 if e not in l2] def adjacent_n_tuples(objects: Sequence[T], n: int) -> zip[tuple[T, ...]]: return zip(*[ [*objects[k:], *objects[:k]] for k in range(n) ]) def adjacent_pairs(objects: Sequence[T]) -> zip[tuple[T, T]]: return adjacent_n_tuples(objects, 2) def batch_by_property( items: Iterable[T], property_func: Callable[[T], S] ) -> list[tuple[T, S]]: """ Takes in a list, and returns a list of tuples, (batch, prop) such that all items in a batch have the same output when put into property_func, and such that chaining all these batches together would give the original list (i.e. order is preserved) """ batch_prop_pairs = [] curr_batch = [] curr_prop = None for item in items: prop = property_func(item) if prop != curr_prop: # Add current batch if len(curr_batch) > 0: batch_prop_pairs.append((curr_batch, curr_prop)) # Redefine curr curr_prop = prop curr_batch = [item] else: curr_batch.append(item) if len(curr_batch) > 0: batch_prop_pairs.append((curr_batch, curr_prop)) return batch_prop_pairs def listify(obj: object) -> list: if isinstance(obj, str): return [obj] try: return list(obj) except TypeError: return [obj] def shuffled(iterable: Iterable) -> list: as_list = list(iterable) random.shuffle(as_list) return as_list def resize_array(nparray: np.ndarray, length: int) -> np.ndarray: if len(nparray) == length: return nparray return np.resize(nparray, (length, *nparray.shape[1:])) def resize_preserving_order(nparray: np.ndarray, length: int) -> np.ndarray: if len(nparray) == 0: return np.resize(nparray, length) if len(nparray) == length: return nparray indices = np.arange(length) * len(nparray) // length return nparray[indices] def resize_with_interpolation(nparray: np.ndarray, length: int) -> np.ndarray: if len(nparray) == length: return nparray if len(nparray) == 1 or array_is_constant(nparray): return nparray[:1].repeat(length, axis=0) if length == 0: return np.zeros((0, *nparray.shape[1:])) cont_indices = np.linspace(0, len(nparray) - 1, length) return np.array([ (1 - a) * nparray[lh] + a * nparray[rh] for ci in cont_indices for lh, rh, a in [(int(ci), int(np.ceil(ci)), ci % 1)] ]) def make_even( iterable_1: Sequence[T], iterable_2: Sequence[S] ) -> tuple[Sequence[T], Sequence[S]]: len1 = len(iterable_1) len2 = len(iterable_2) if len1 == len2: return iterable_1, iterable_2 new_len = max(len1, len2) return ( [iterable_1[(n * len1) // new_len] for n in range(new_len)], [iterable_2[(n * len2) // new_len] for n in range(new_len)] ) def arrays_match(arr1: np.ndarray, arr2: np.ndarray) -> bool: return arr1.shape == arr2.shape and (arr1 == arr2).all() def array_is_constant(arr: np.ndarray) -> bool: return len(arr) > 0 and (arr == arr[0]).all() def cartesian_product(*arrays: np.ndarray): """ Copied from https://stackoverflow.com/a/11146645 """ la = len(arrays) dtype = np.result_type(*arrays) arr = np.empty([len(a) for a in arrays] + [la], dtype=dtype) for i, a in enumerate(np.ix_(*arrays)): arr[..., i] = a return arr.reshape(-1, la) def hash_obj(obj: object) -> int: if isinstance(obj, dict): return hash(tuple(sorted([ (hash_obj(k), hash_obj(v)) for k, v in obj.items() ]))) if isinstance(obj, set): return hash(tuple(sorted(hash_obj(e) for e in obj))) if isinstance(obj, (tuple, list)): return hash(tuple(hash_obj(e) for e in obj)) if isinstance(obj, Color): return hash(obj.get_rgb()) return hash(obj)
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/utils/rate_functions.py
manimlib/utils/rate_functions.py
from __future__ import annotations import numpy as np from manimlib.utils.bezier import bezier from typing import TYPE_CHECKING if TYPE_CHECKING: from typing import Callable def linear(t: float) -> float: return t def smooth(t: float) -> float: # Zero first and second derivatives at t=0 and t=1. # Equivalent to bezier([0, 0, 0, 1, 1, 1]) s = 1 - t return (t**3) * (10 * s * s + 5 * s * t + t * t) def rush_into(t: float) -> float: return 2 * smooth(0.5 * t) def rush_from(t: float) -> float: return 2 * smooth(0.5 * (t + 1)) - 1 def slow_into(t: float) -> float: return np.sqrt(1 - (1 - t) * (1 - t)) def double_smooth(t: float) -> float: if t < 0.5: return 0.5 * smooth(2 * t) else: return 0.5 * (1 + smooth(2 * t - 1)) def there_and_back(t: float) -> float: new_t = 2 * t if t < 0.5 else 2 * (1 - t) return smooth(new_t) def there_and_back_with_pause(t: float, pause_ratio: float = 1. / 3) -> float: a = 2. / (1. - pause_ratio) if t < 0.5 - pause_ratio / 2: return smooth(a * t) elif t < 0.5 + pause_ratio / 2: return 1 else: return smooth(a - a * t) def running_start(t: float, pull_factor: float = -0.5) -> float: return bezier([0, 0, pull_factor, pull_factor, 1, 1, 1])(t) def overshoot(t: float, pull_factor: float = 1.5) -> float: return bezier([0, 0, pull_factor, pull_factor, 1, 1])(t) def not_quite_there( func: Callable[[float], float] = smooth, proportion: float = 0.7 ) -> Callable[[float], float]: def result(t): return proportion * func(t) return result def wiggle(t: float, wiggles: float = 2) -> float: return there_and_back(t) * np.sin(wiggles * np.pi * t) def squish_rate_func( func: Callable[[float], float], a: float = 0.4, b: float = 0.6 ) -> Callable[[float], float]: def result(t): if a == b: return a elif t < a: return func(0) elif t > b: return func(1) else: return func((t - a) / (b - a)) return result # Stylistically, should this take parameters (with default values)? # Ultimately, the functionality is entirely subsumed by squish_rate_func, # but it may be useful to have a nice name for with nice default params for # "lingering", different from squish_rate_func's default params def lingering(t: float) -> float: return squish_rate_func(lambda t: t, 0, 0.8)(t) def exponential_decay(t: float, half_life: float = 0.1) -> float: # The half-life should be rather small to minimize # the cut-off error at the end return 1 - np.exp(-t / half_life)
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/utils/simple_functions.py
manimlib/utils/simple_functions.py
from __future__ import annotations from functools import lru_cache import hashlib import inspect import math import numpy as np from typing import TYPE_CHECKING if TYPE_CHECKING: from typing import Callable, TypeVar, Iterable from manimlib.typing import FloatArray Scalable = TypeVar("Scalable", float, FloatArray) def sigmoid(x: float | FloatArray): return 1.0 / (1 + np.exp(-x)) @lru_cache(maxsize=10) def choose(n: int, k: int) -> int: return math.comb(n, k) def gen_choose(n: int, r: int) -> int: return int(np.prod(range(n, n - r, -1)) / math.factorial(r)) def get_num_args(function: Callable) -> int: return function.__code__.co_argcount def get_parameters(function: Callable) -> Iterable[str]: return inspect.signature(function).parameters.keys() def clip(a: float, min_a: float, max_a: float) -> float: if a < min_a: return min_a elif a > max_a: return max_a return a def arr_clip(arr: np.ndarray, min_a: float, max_a: float) -> np.ndarray: arr[arr < min_a] = min_a arr[arr > max_a] = max_a return arr def fdiv(a: Scalable, b: Scalable, zero_over_zero_value: Scalable | None = None) -> Scalable: """ Less heavyweight name for np.true_divide, enabling default behavior for 0/0 """ if zero_over_zero_value is not None: out = np.full_like(a, zero_over_zero_value) where = np.logical_or(a != 0, b != 0) else: out = None where = True return np.true_divide(a, b, out=out, where=where) def binary_search( function: Callable[[float], float], target: float, lower_bound: float, upper_bound: float, tolerance:float = 1e-4 ) -> float | None: lh = lower_bound rh = upper_bound mh = (lh + rh) / 2 while abs(rh - lh) > tolerance: lx, mx, rx = [function(h) for h in (lh, mh, rh)] if lx == target: return lx if rx == target: return rx if lx <= target and rx >= target: if mx > target: rh = mh else: lh = mh elif lx > target and rx < target: lh, rh = rh, lh else: return None mh = (lh + rh) / 2 return mh def hash_string(string: str, n_bytes=16) -> str: hasher = hashlib.sha256(string.encode()) return hasher.hexdigest()[:n_bytes]
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false
3b1b/manim
https://github.com/3b1b/manim/blob/e5298385edabb34a09cf63743c1a4ec983f1ab11/manimlib/utils/dict_ops.py
manimlib/utils/dict_ops.py
import itertools as it import numpy as np def merge_dicts_recursively(*dicts): """ Creates a dict whose keyset is the union of all the input dictionaries. The value for each key is based on the first dict in the list with that key. dicts later in the list have higher priority When values are dictionaries, it is applied recursively """ result = dict() all_items = it.chain(*[d.items() for d in dicts]) for key, value in all_items: if key in result and isinstance(result[key], dict) and isinstance(value, dict): result[key] = merge_dicts_recursively(result[key], value) else: result[key] = value return result
python
MIT
e5298385edabb34a09cf63743c1a4ec983f1ab11
2026-01-04T14:38:15.512738Z
false