from dataclasses import dataclass import sys import os import hashlib import shutil import subprocess from ..Utils import safe_makedirs, cached_function import zipfile from .. import __version__ try: import zlib zipfile_compression_mode = zipfile.ZIP_DEFLATED except ImportError: zipfile_compression_mode = zipfile.ZIP_STORED try: import gzip gzip_open = gzip.open gzip_ext = ".gz" except ImportError: gzip_open = open gzip_ext = "" zip_ext = ".zip" MAX_CACHE_SIZE = 1024 * 1024 * 100 join_path = cached_function(os.path.join) @cached_function def file_hash(filename): path = os.path.normpath(filename) prefix = ("%d:%s" % (len(path), path)).encode("UTF-8") m = hashlib.sha256(prefix) with open(path, "rb") as f: data = f.read(65000) while data: m.update(data) data = f.read(65000) return m.hexdigest() @cached_function def get_cython_cache_dir(): r""" Return the base directory containing Cython's caches. Priority: 1. CYTHON_CACHE_DIR 2. (OS X): ~/Library/Caches/Cython (posix not OS X): XDG_CACHE_HOME/cython if XDG_CACHE_HOME defined 3. ~/.cython """ if "CYTHON_CACHE_DIR" in os.environ: return os.environ["CYTHON_CACHE_DIR"] parent = None if os.name == "posix": if sys.platform == "darwin": parent = os.path.expanduser("~/Library/Caches") else: # this could fallback on ~/.cache parent = os.environ.get("XDG_CACHE_HOME") if parent and os.path.isdir(parent): return join_path(parent, "cython") # last fallback: ~/.cython return os.path.expanduser(join_path("~", ".cython")) @dataclass class FingerprintFlags: language: str = "c" py_limited_api: bool = False np_pythran: bool = False def get_fingerprint(self): return str((self.language, self.py_limited_api, self.np_pythran)) class Cache: def __init__(self, path, cache_size=None): if path is None: self.path = join_path(get_cython_cache_dir(), "compiler") else: self.path = path self.cache_size = cache_size if cache_size is not None else MAX_CACHE_SIZE if not os.path.exists(self.path): os.makedirs(self.path) def transitive_fingerprint( self, filename, dependencies, compilation_options, flags=FingerprintFlags() ): r""" Return a fingerprint of a cython file that is about to be cythonized. Fingerprints are looked up in future compilations. If the fingerprint is found, the cythonization can be skipped. The fingerprint must incorporate everything that has an influence on the generated code. """ try: m = hashlib.sha256(__version__.encode("UTF-8")) m.update(file_hash(filename).encode("UTF-8")) for x in sorted(dependencies): if os.path.splitext(x)[1] not in (".c", ".cpp", ".h"): m.update(file_hash(x).encode("UTF-8")) # Include the module attributes that change the compilation result # in the fingerprint. We do not iterate over module.__dict__ and # include almost everything here as users might extend Extension # with arbitrary (random) attributes that would lead to cache # misses. m.update(flags.get_fingerprint().encode("UTF-8")) m.update(compilation_options.get_fingerprint().encode("UTF-8")) return m.hexdigest() except OSError: return None def fingerprint_file(self, cfile, fingerprint, ext): return ( join_path(self.path, "%s-%s" % (os.path.basename(cfile), fingerprint)) + ext ) def lookup_cache(self, c_file, fingerprint): # Cython-generated c files are highly compressible. # (E.g. a compression ratio of about 10 for Sage). if not os.path.exists(self.path): safe_makedirs(self.path) gz_fingerprint_file = self.fingerprint_file(c_file, fingerprint, gzip_ext) if os.path.exists(gz_fingerprint_file): return gz_fingerprint_file zip_fingerprint_file = self.fingerprint_file(c_file, fingerprint, zip_ext) if os.path.exists(zip_fingerprint_file): return zip_fingerprint_file return None def load_from_cache(self, c_file, cached): ext = os.path.splitext(cached)[1] if ext == gzip_ext: os.utime(cached, None) with gzip_open(cached, "rb") as g: with open(c_file, "wb") as f: shutil.copyfileobj(g, f) elif ext == zip_ext: os.utime(cached, None) dirname = os.path.dirname(c_file) with zipfile.ZipFile(cached) as z: for artifact in z.namelist(): z.extract(artifact, join_path(dirname, artifact)) else: raise ValueError(f"Unsupported cache file extension: {ext}") def store_to_cache(self, c_file, fingerprint, compilation_result): artifacts = compilation_result.get_generated_source_files() if len(artifacts) == 1: fingerprint_file = self.fingerprint_file(c_file, fingerprint, gzip_ext) with open(c_file, "rb") as f: with gzip_open(fingerprint_file + ".tmp", "wb") as g: shutil.copyfileobj(f, g) else: fingerprint_file = self.fingerprint_file(c_file, fingerprint, zip_ext) with zipfile.ZipFile( fingerprint_file + ".tmp", "w", zipfile_compression_mode ) as zip: for artifact in artifacts: zip.write(artifact, os.path.basename(artifact)) os.rename(fingerprint_file + ".tmp", fingerprint_file) def cleanup_cache(self, ratio=0.85): try: completed_process = subprocess.run( ["du", "-s", "-k", os.path.abspath(self.path)], stdout=subprocess.PIPE ) stdout = completed_process.stdout if completed_process.returncode == 0: total_size = 1024 * int(stdout.strip().split()[0]) if total_size < self.cache_size: return except (OSError, ValueError): pass total_size = 0 all = [] for file in os.listdir(self.path): path = join_path(self.path, file) s = os.stat(path) total_size += s.st_size all.append((s.st_atime, s.st_size, path)) if total_size > self.cache_size: for time, size, file in reversed(sorted(all)): os.unlink(file) total_size -= size if total_size < self.cache_size * ratio: break