DeepSolanaCoder
/
DeepSeek-Coder-main
/finetune
/venv
/lib
/python3.12
/site-packages
/aiohttp
/web_fileresponse.py
| import asyncio | |
| import io | |
| import os | |
| import pathlib | |
| import sys | |
| from contextlib import suppress | |
| from enum import Enum, auto | |
| from mimetypes import MimeTypes | |
| from stat import S_ISREG | |
| from types import MappingProxyType | |
| from typing import ( # noqa | |
| IO, | |
| TYPE_CHECKING, | |
| Any, | |
| Awaitable, | |
| Callable, | |
| Final, | |
| Iterator, | |
| List, | |
| Optional, | |
| Set, | |
| Tuple, | |
| Union, | |
| cast, | |
| ) | |
| from . import hdrs | |
| from .abc import AbstractStreamWriter | |
| from .helpers import ETAG_ANY, ETag, must_be_empty_body | |
| from .typedefs import LooseHeaders, PathLike | |
| from .web_exceptions import ( | |
| HTTPForbidden, | |
| HTTPNotFound, | |
| HTTPNotModified, | |
| HTTPPartialContent, | |
| HTTPPreconditionFailed, | |
| HTTPRequestRangeNotSatisfiable, | |
| ) | |
| from .web_response import StreamResponse | |
| __all__ = ("FileResponse",) | |
| if TYPE_CHECKING: | |
| from .web_request import BaseRequest | |
| _T_OnChunkSent = Optional[Callable[[bytes], Awaitable[None]]] | |
| NOSENDFILE: Final[bool] = bool(os.environ.get("AIOHTTP_NOSENDFILE")) | |
| CONTENT_TYPES: Final[MimeTypes] = MimeTypes() | |
| # File extension to IANA encodings map that will be checked in the order defined. | |
| ENCODING_EXTENSIONS = MappingProxyType( | |
| {ext: CONTENT_TYPES.encodings_map[ext] for ext in (".br", ".gz")} | |
| ) | |
| FALLBACK_CONTENT_TYPE = "application/octet-stream" | |
| # Provide additional MIME type/extension pairs to be recognized. | |
| # https://en.wikipedia.org/wiki/List_of_archive_formats#Compression_only | |
| ADDITIONAL_CONTENT_TYPES = MappingProxyType( | |
| { | |
| "application/gzip": ".gz", | |
| "application/x-brotli": ".br", | |
| "application/x-bzip2": ".bz2", | |
| "application/x-compress": ".Z", | |
| "application/x-xz": ".xz", | |
| } | |
| ) | |
| class _FileResponseResult(Enum): | |
| """The result of the file response.""" | |
| SEND_FILE = auto() # Ie a regular file to send | |
| NOT_ACCEPTABLE = auto() # Ie a socket, or non-regular file | |
| PRE_CONDITION_FAILED = auto() # Ie If-Match or If-None-Match failed | |
| NOT_MODIFIED = auto() # 304 Not Modified | |
| # Add custom pairs and clear the encodings map so guess_type ignores them. | |
| CONTENT_TYPES.encodings_map.clear() | |
| for content_type, extension in ADDITIONAL_CONTENT_TYPES.items(): | |
| CONTENT_TYPES.add_type(content_type, extension) # type: ignore[attr-defined] | |
| _CLOSE_FUTURES: Set[asyncio.Future[None]] = set() | |
| class FileResponse(StreamResponse): | |
| """A response object can be used to send files.""" | |
| def __init__( | |
| self, | |
| path: PathLike, | |
| chunk_size: int = 256 * 1024, | |
| status: int = 200, | |
| reason: Optional[str] = None, | |
| headers: Optional[LooseHeaders] = None, | |
| ) -> None: | |
| super().__init__(status=status, reason=reason, headers=headers) | |
| self._path = pathlib.Path(path) | |
| self._chunk_size = chunk_size | |
| def _seek_and_read(self, fobj: IO[Any], offset: int, chunk_size: int) -> bytes: | |
| fobj.seek(offset) | |
| return fobj.read(chunk_size) # type: ignore[no-any-return] | |
| async def _sendfile_fallback( | |
| self, writer: AbstractStreamWriter, fobj: IO[Any], offset: int, count: int | |
| ) -> AbstractStreamWriter: | |
| # To keep memory usage low,fobj is transferred in chunks | |
| # controlled by the constructor's chunk_size argument. | |
| chunk_size = self._chunk_size | |
| loop = asyncio.get_event_loop() | |
| chunk = await loop.run_in_executor( | |
| None, self._seek_and_read, fobj, offset, chunk_size | |
| ) | |
| while chunk: | |
| await writer.write(chunk) | |
| count = count - chunk_size | |
| if count <= 0: | |
| break | |
| chunk = await loop.run_in_executor(None, fobj.read, min(chunk_size, count)) | |
| await writer.drain() | |
| return writer | |
| async def _sendfile( | |
| self, request: "BaseRequest", fobj: IO[Any], offset: int, count: int | |
| ) -> AbstractStreamWriter: | |
| writer = await super().prepare(request) | |
| assert writer is not None | |
| if NOSENDFILE or self.compression: | |
| return await self._sendfile_fallback(writer, fobj, offset, count) | |
| loop = request._loop | |
| transport = request.transport | |
| assert transport is not None | |
| try: | |
| await loop.sendfile(transport, fobj, offset, count) | |
| except NotImplementedError: | |
| return await self._sendfile_fallback(writer, fobj, offset, count) | |
| await super().write_eof() | |
| return writer | |
| def _etag_match(etag_value: str, etags: Tuple[ETag, ...], *, weak: bool) -> bool: | |
| if len(etags) == 1 and etags[0].value == ETAG_ANY: | |
| return True | |
| return any( | |
| etag.value == etag_value for etag in etags if weak or not etag.is_weak | |
| ) | |
| async def _not_modified( | |
| self, request: "BaseRequest", etag_value: str, last_modified: float | |
| ) -> Optional[AbstractStreamWriter]: | |
| self.set_status(HTTPNotModified.status_code) | |
| self._length_check = False | |
| self.etag = etag_value # type: ignore[assignment] | |
| self.last_modified = last_modified # type: ignore[assignment] | |
| # Delete any Content-Length headers provided by user. HTTP 304 | |
| # should always have empty response body | |
| return await super().prepare(request) | |
| async def _precondition_failed( | |
| self, request: "BaseRequest" | |
| ) -> Optional[AbstractStreamWriter]: | |
| self.set_status(HTTPPreconditionFailed.status_code) | |
| self.content_length = 0 | |
| return await super().prepare(request) | |
| def _make_response( | |
| self, request: "BaseRequest", accept_encoding: str | |
| ) -> Tuple[ | |
| _FileResponseResult, Optional[io.BufferedReader], os.stat_result, Optional[str] | |
| ]: | |
| """Return the response result, io object, stat result, and encoding. | |
| If an uncompressed file is returned, the encoding is set to | |
| :py:data:`None`. | |
| This method should be called from a thread executor | |
| since it calls os.stat which may block. | |
| """ | |
| file_path, st, file_encoding = self._get_file_path_stat_encoding( | |
| accept_encoding | |
| ) | |
| if not file_path: | |
| return _FileResponseResult.NOT_ACCEPTABLE, None, st, None | |
| etag_value = f"{st.st_mtime_ns:x}-{st.st_size:x}" | |
| # https://www.rfc-editor.org/rfc/rfc9110#section-13.1.1-2 | |
| if (ifmatch := request.if_match) is not None and not self._etag_match( | |
| etag_value, ifmatch, weak=False | |
| ): | |
| return _FileResponseResult.PRE_CONDITION_FAILED, None, st, file_encoding | |
| if ( | |
| (unmodsince := request.if_unmodified_since) is not None | |
| and ifmatch is None | |
| and st.st_mtime > unmodsince.timestamp() | |
| ): | |
| return _FileResponseResult.PRE_CONDITION_FAILED, None, st, file_encoding | |
| # https://www.rfc-editor.org/rfc/rfc9110#section-13.1.2-2 | |
| if (ifnonematch := request.if_none_match) is not None and self._etag_match( | |
| etag_value, ifnonematch, weak=True | |
| ): | |
| return _FileResponseResult.NOT_MODIFIED, None, st, file_encoding | |
| if ( | |
| (modsince := request.if_modified_since) is not None | |
| and ifnonematch is None | |
| and st.st_mtime <= modsince.timestamp() | |
| ): | |
| return _FileResponseResult.NOT_MODIFIED, None, st, file_encoding | |
| fobj = file_path.open("rb") | |
| with suppress(OSError): | |
| # fstat() may not be available on all platforms | |
| # Once we open the file, we want the fstat() to ensure | |
| # the file has not changed between the first stat() | |
| # and the open(). | |
| st = os.stat(fobj.fileno()) | |
| return _FileResponseResult.SEND_FILE, fobj, st, file_encoding | |
| def _get_file_path_stat_encoding( | |
| self, accept_encoding: str | |
| ) -> Tuple[Optional[pathlib.Path], os.stat_result, Optional[str]]: | |
| file_path = self._path | |
| for file_extension, file_encoding in ENCODING_EXTENSIONS.items(): | |
| if file_encoding not in accept_encoding: | |
| continue | |
| compressed_path = file_path.with_suffix(file_path.suffix + file_extension) | |
| with suppress(OSError): | |
| # Do not follow symlinks and ignore any non-regular files. | |
| st = compressed_path.lstat() | |
| if S_ISREG(st.st_mode): | |
| return compressed_path, st, file_encoding | |
| # Fallback to the uncompressed file | |
| st = file_path.stat() | |
| return file_path if S_ISREG(st.st_mode) else None, st, None | |
| async def prepare(self, request: "BaseRequest") -> Optional[AbstractStreamWriter]: | |
| loop = asyncio.get_running_loop() | |
| # Encoding comparisons should be case-insensitive | |
| # https://www.rfc-editor.org/rfc/rfc9110#section-8.4.1 | |
| accept_encoding = request.headers.get(hdrs.ACCEPT_ENCODING, "").lower() | |
| try: | |
| response_result, fobj, st, file_encoding = await loop.run_in_executor( | |
| None, self._make_response, request, accept_encoding | |
| ) | |
| except PermissionError: | |
| self.set_status(HTTPForbidden.status_code) | |
| return await super().prepare(request) | |
| except OSError: | |
| # Most likely to be FileNotFoundError or OSError for circular | |
| # symlinks in python >= 3.13, so respond with 404. | |
| self.set_status(HTTPNotFound.status_code) | |
| return await super().prepare(request) | |
| # Forbid special files like sockets, pipes, devices, etc. | |
| if response_result is _FileResponseResult.NOT_ACCEPTABLE: | |
| self.set_status(HTTPForbidden.status_code) | |
| return await super().prepare(request) | |
| if response_result is _FileResponseResult.PRE_CONDITION_FAILED: | |
| return await self._precondition_failed(request) | |
| if response_result is _FileResponseResult.NOT_MODIFIED: | |
| etag_value = f"{st.st_mtime_ns:x}-{st.st_size:x}" | |
| last_modified = st.st_mtime | |
| return await self._not_modified(request, etag_value, last_modified) | |
| assert fobj is not None | |
| try: | |
| return await self._prepare_open_file(request, fobj, st, file_encoding) | |
| finally: | |
| # We do not await here because we do not want to wait | |
| # for the executor to finish before returning the response | |
| # so the connection can begin servicing another request | |
| # as soon as possible. | |
| close_future = loop.run_in_executor(None, fobj.close) | |
| # Hold a strong reference to the future to prevent it from being | |
| # garbage collected before it completes. | |
| _CLOSE_FUTURES.add(close_future) | |
| close_future.add_done_callback(_CLOSE_FUTURES.remove) | |
| async def _prepare_open_file( | |
| self, | |
| request: "BaseRequest", | |
| fobj: io.BufferedReader, | |
| st: os.stat_result, | |
| file_encoding: Optional[str], | |
| ) -> Optional[AbstractStreamWriter]: | |
| status = self._status | |
| file_size: int = st.st_size | |
| file_mtime: float = st.st_mtime | |
| count: int = file_size | |
| start: Optional[int] = None | |
| if (ifrange := request.if_range) is None or file_mtime <= ifrange.timestamp(): | |
| # If-Range header check: | |
| # condition = cached date >= last modification date | |
| # return 206 if True else 200. | |
| # if False: | |
| # Range header would not be processed, return 200 | |
| # if True but Range header missing | |
| # return 200 | |
| try: | |
| rng = request.http_range | |
| start = rng.start | |
| end: Optional[int] = rng.stop | |
| except ValueError: | |
| # https://tools.ietf.org/html/rfc7233: | |
| # A server generating a 416 (Range Not Satisfiable) response to | |
| # a byte-range request SHOULD send a Content-Range header field | |
| # with an unsatisfied-range value. | |
| # The complete-length in a 416 response indicates the current | |
| # length of the selected representation. | |
| # | |
| # Will do the same below. Many servers ignore this and do not | |
| # send a Content-Range header with HTTP 416 | |
| self._headers[hdrs.CONTENT_RANGE] = f"bytes */{file_size}" | |
| self.set_status(HTTPRequestRangeNotSatisfiable.status_code) | |
| return await super().prepare(request) | |
| # If a range request has been made, convert start, end slice | |
| # notation into file pointer offset and count | |
| if start is not None: | |
| if start < 0 and end is None: # return tail of file | |
| start += file_size | |
| if start < 0: | |
| # if Range:bytes=-1000 in request header but file size | |
| # is only 200, there would be trouble without this | |
| start = 0 | |
| count = file_size - start | |
| else: | |
| # rfc7233:If the last-byte-pos value is | |
| # absent, or if the value is greater than or equal to | |
| # the current length of the representation data, | |
| # the byte range is interpreted as the remainder | |
| # of the representation (i.e., the server replaces the | |
| # value of last-byte-pos with a value that is one less than | |
| # the current length of the selected representation). | |
| count = ( | |
| min(end if end is not None else file_size, file_size) - start | |
| ) | |
| if start >= file_size: | |
| # HTTP 416 should be returned in this case. | |
| # | |
| # According to https://tools.ietf.org/html/rfc7233: | |
| # If a valid byte-range-set includes at least one | |
| # byte-range-spec with a first-byte-pos that is less than | |
| # the current length of the representation, or at least one | |
| # suffix-byte-range-spec with a non-zero suffix-length, | |
| # then the byte-range-set is satisfiable. Otherwise, the | |
| # byte-range-set is unsatisfiable. | |
| self._headers[hdrs.CONTENT_RANGE] = f"bytes */{file_size}" | |
| self.set_status(HTTPRequestRangeNotSatisfiable.status_code) | |
| return await super().prepare(request) | |
| status = HTTPPartialContent.status_code | |
| # Even though you are sending the whole file, you should still | |
| # return a HTTP 206 for a Range request. | |
| self.set_status(status) | |
| # If the Content-Type header is not already set, guess it based on the | |
| # extension of the request path. The encoding returned by guess_type | |
| # can be ignored since the map was cleared above. | |
| if hdrs.CONTENT_TYPE not in self._headers: | |
| if sys.version_info >= (3, 13): | |
| guesser = CONTENT_TYPES.guess_file_type | |
| else: | |
| guesser = CONTENT_TYPES.guess_type | |
| self.content_type = guesser(self._path)[0] or FALLBACK_CONTENT_TYPE | |
| if file_encoding: | |
| self._headers[hdrs.CONTENT_ENCODING] = file_encoding | |
| self._headers[hdrs.VARY] = hdrs.ACCEPT_ENCODING | |
| # Disable compression if we are already sending | |
| # a compressed file since we don't want to double | |
| # compress. | |
| self._compression = False | |
| self.etag = f"{st.st_mtime_ns:x}-{st.st_size:x}" # type: ignore[assignment] | |
| self.last_modified = file_mtime # type: ignore[assignment] | |
| self.content_length = count | |
| self._headers[hdrs.ACCEPT_RANGES] = "bytes" | |
| if status == HTTPPartialContent.status_code: | |
| real_start = start | |
| assert real_start is not None | |
| self._headers[hdrs.CONTENT_RANGE] = "bytes {}-{}/{}".format( | |
| real_start, real_start + count - 1, file_size | |
| ) | |
| # If we are sending 0 bytes calling sendfile() will throw a ValueError | |
| if count == 0 or must_be_empty_body(request.method, status): | |
| return await super().prepare(request) | |
| # be aware that start could be None or int=0 here. | |
| offset = start or 0 | |
| return await self._sendfile(request, fobj, offset, count) | |