ZTWHHH commited on
Commit
a7f22cb
·
verified ·
1 Parent(s): 7749173

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. janus/lib/python3.10/site-packages/certifi/__init__.py +4 -0
  2. janus/lib/python3.10/site-packages/certifi/__main__.py +12 -0
  3. janus/lib/python3.10/site-packages/certifi/__pycache__/__init__.cpython-310.pyc +0 -0
  4. janus/lib/python3.10/site-packages/certifi/__pycache__/__main__.cpython-310.pyc +0 -0
  5. janus/lib/python3.10/site-packages/certifi/__pycache__/core.cpython-310.pyc +0 -0
  6. janus/lib/python3.10/site-packages/certifi/cacert.pem +0 -0
  7. janus/lib/python3.10/site-packages/certifi/core.py +114 -0
  8. janus/lib/python3.10/site-packages/certifi/py.typed +0 -0
  9. janus/lib/python3.10/site-packages/jinja2/__init__.py +38 -0
  10. janus/lib/python3.10/site-packages/jinja2/__pycache__/bccache.cpython-310.pyc +0 -0
  11. janus/lib/python3.10/site-packages/jinja2/__pycache__/defaults.cpython-310.pyc +0 -0
  12. janus/lib/python3.10/site-packages/jinja2/__pycache__/exceptions.cpython-310.pyc +0 -0
  13. janus/lib/python3.10/site-packages/jinja2/_identifier.py +6 -0
  14. janus/lib/python3.10/site-packages/jinja2/bccache.py +408 -0
  15. janus/lib/python3.10/site-packages/jinja2/compiler.py +1998 -0
  16. janus/lib/python3.10/site-packages/jinja2/constants.py +20 -0
  17. janus/lib/python3.10/site-packages/jinja2/debug.py +191 -0
  18. janus/lib/python3.10/site-packages/jinja2/defaults.py +48 -0
  19. janus/lib/python3.10/site-packages/jinja2/environment.py +1672 -0
  20. janus/lib/python3.10/site-packages/jinja2/exceptions.py +166 -0
  21. janus/lib/python3.10/site-packages/jinja2/ext.py +870 -0
  22. janus/lib/python3.10/site-packages/jinja2/idtracking.py +318 -0
  23. janus/lib/python3.10/site-packages/jinja2/lexer.py +868 -0
  24. janus/lib/python3.10/site-packages/jinja2/meta.py +112 -0
  25. janus/lib/python3.10/site-packages/jinja2/nativetypes.py +130 -0
  26. janus/lib/python3.10/site-packages/jinja2/nodes.py +1206 -0
  27. janus/lib/python3.10/site-packages/jinja2/optimizer.py +48 -0
  28. janus/lib/python3.10/site-packages/jinja2/parser.py +1049 -0
  29. janus/lib/python3.10/site-packages/jinja2/py.typed +0 -0
  30. janus/lib/python3.10/site-packages/jinja2/runtime.py +1062 -0
  31. janus/lib/python3.10/site-packages/jinja2/tests.py +256 -0
  32. janus/lib/python3.10/site-packages/jinja2/utils.py +766 -0
  33. janus/lib/python3.10/site-packages/jinja2/visitor.py +92 -0
  34. janus/lib/python3.10/site-packages/numpy/_core/__pycache__/__init__.cpython-310.pyc +0 -0
  35. janus/lib/python3.10/site-packages/numpy/_core/__pycache__/_add_newdocs_scalars.cpython-310.pyc +0 -0
  36. janus/lib/python3.10/site-packages/numpy/_core/__pycache__/_asarray.cpython-310.pyc +0 -0
  37. janus/lib/python3.10/site-packages/numpy/_core/__pycache__/_string_helpers.cpython-310.pyc +0 -0
  38. janus/lib/python3.10/site-packages/numpy/_core/__pycache__/cversions.cpython-310.pyc +0 -0
  39. janus/lib/python3.10/site-packages/numpy/_core/__pycache__/defchararray.cpython-310.pyc +0 -0
  40. janus/lib/python3.10/site-packages/numpy/_core/__pycache__/getlimits.cpython-310.pyc +0 -0
  41. janus/lib/python3.10/site-packages/numpy/_core/__pycache__/memmap.cpython-310.pyc +0 -0
  42. janus/lib/python3.10/site-packages/numpy/_core/__pycache__/records.cpython-310.pyc +0 -0
  43. janus/lib/python3.10/site-packages/numpy/_core/__pycache__/shape_base.cpython-310.pyc +0 -0
  44. janus/lib/python3.10/site-packages/numpy/_core/include/numpy/arrayobject.h +7 -0
  45. janus/lib/python3.10/site-packages/numpy/_core/include/numpy/npy_common.h +1070 -0
  46. janus/lib/python3.10/site-packages/numpy/_core/lib/npy-pkg-config/npymath.ini +20 -0
  47. janus/lib/python3.10/site-packages/numpy/_core/lib/pkgconfig/numpy.pc +7 -0
  48. janus/lib/python3.10/site-packages/numpy/random/__pycache__/__init__.cpython-310.pyc +0 -0
  49. janus/lib/python3.10/site-packages/numpy/random/__pycache__/_pickle.cpython-310.pyc +0 -0
  50. janus/lib/python3.10/site-packages/numpy/random/_examples/numba/__pycache__/extending_distributions.cpython-310.pyc +0 -0
janus/lib/python3.10/site-packages/certifi/__init__.py ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ from .core import contents, where
2
+
3
+ __all__ = ["contents", "where"]
4
+ __version__ = "2024.12.14"
janus/lib/python3.10/site-packages/certifi/__main__.py ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import argparse
2
+
3
+ from certifi import contents, where
4
+
5
+ parser = argparse.ArgumentParser()
6
+ parser.add_argument("-c", "--contents", action="store_true")
7
+ args = parser.parse_args()
8
+
9
+ if args.contents:
10
+ print(contents())
11
+ else:
12
+ print(where())
janus/lib/python3.10/site-packages/certifi/__pycache__/__init__.cpython-310.pyc ADDED
Binary file (271 Bytes). View file
 
janus/lib/python3.10/site-packages/certifi/__pycache__/__main__.cpython-310.pyc ADDED
Binary file (409 Bytes). View file
 
janus/lib/python3.10/site-packages/certifi/__pycache__/core.cpython-310.pyc ADDED
Binary file (2.1 kB). View file
 
janus/lib/python3.10/site-packages/certifi/cacert.pem ADDED
The diff for this file is too large to render. See raw diff
 
janus/lib/python3.10/site-packages/certifi/core.py ADDED
@@ -0,0 +1,114 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """
2
+ certifi.py
3
+ ~~~~~~~~~~
4
+
5
+ This module returns the installation location of cacert.pem or its contents.
6
+ """
7
+ import sys
8
+ import atexit
9
+
10
+ def exit_cacert_ctx() -> None:
11
+ _CACERT_CTX.__exit__(None, None, None) # type: ignore[union-attr]
12
+
13
+
14
+ if sys.version_info >= (3, 11):
15
+
16
+ from importlib.resources import as_file, files
17
+
18
+ _CACERT_CTX = None
19
+ _CACERT_PATH = None
20
+
21
+ def where() -> str:
22
+ # This is slightly terrible, but we want to delay extracting the file
23
+ # in cases where we're inside of a zipimport situation until someone
24
+ # actually calls where(), but we don't want to re-extract the file
25
+ # on every call of where(), so we'll do it once then store it in a
26
+ # global variable.
27
+ global _CACERT_CTX
28
+ global _CACERT_PATH
29
+ if _CACERT_PATH is None:
30
+ # This is slightly janky, the importlib.resources API wants you to
31
+ # manage the cleanup of this file, so it doesn't actually return a
32
+ # path, it returns a context manager that will give you the path
33
+ # when you enter it and will do any cleanup when you leave it. In
34
+ # the common case of not needing a temporary file, it will just
35
+ # return the file system location and the __exit__() is a no-op.
36
+ #
37
+ # We also have to hold onto the actual context manager, because
38
+ # it will do the cleanup whenever it gets garbage collected, so
39
+ # we will also store that at the global level as well.
40
+ _CACERT_CTX = as_file(files("certifi").joinpath("cacert.pem"))
41
+ _CACERT_PATH = str(_CACERT_CTX.__enter__())
42
+ atexit.register(exit_cacert_ctx)
43
+
44
+ return _CACERT_PATH
45
+
46
+ def contents() -> str:
47
+ return files("certifi").joinpath("cacert.pem").read_text(encoding="ascii")
48
+
49
+ elif sys.version_info >= (3, 7):
50
+
51
+ from importlib.resources import path as get_path, read_text
52
+
53
+ _CACERT_CTX = None
54
+ _CACERT_PATH = None
55
+
56
+ def where() -> str:
57
+ # This is slightly terrible, but we want to delay extracting the
58
+ # file in cases where we're inside of a zipimport situation until
59
+ # someone actually calls where(), but we don't want to re-extract
60
+ # the file on every call of where(), so we'll do it once then store
61
+ # it in a global variable.
62
+ global _CACERT_CTX
63
+ global _CACERT_PATH
64
+ if _CACERT_PATH is None:
65
+ # This is slightly janky, the importlib.resources API wants you
66
+ # to manage the cleanup of this file, so it doesn't actually
67
+ # return a path, it returns a context manager that will give
68
+ # you the path when you enter it and will do any cleanup when
69
+ # you leave it. In the common case of not needing a temporary
70
+ # file, it will just return the file system location and the
71
+ # __exit__() is a no-op.
72
+ #
73
+ # We also have to hold onto the actual context manager, because
74
+ # it will do the cleanup whenever it gets garbage collected, so
75
+ # we will also store that at the global level as well.
76
+ _CACERT_CTX = get_path("certifi", "cacert.pem")
77
+ _CACERT_PATH = str(_CACERT_CTX.__enter__())
78
+ atexit.register(exit_cacert_ctx)
79
+
80
+ return _CACERT_PATH
81
+
82
+ def contents() -> str:
83
+ return read_text("certifi", "cacert.pem", encoding="ascii")
84
+
85
+ else:
86
+ import os
87
+ import types
88
+ from typing import Union
89
+
90
+ Package = Union[types.ModuleType, str]
91
+ Resource = Union[str, "os.PathLike"]
92
+
93
+ # This fallback will work for Python versions prior to 3.7 that lack the
94
+ # importlib.resources module but relies on the existing `where` function
95
+ # so won't address issues with environments like PyOxidizer that don't set
96
+ # __file__ on modules.
97
+ def read_text(
98
+ package: Package,
99
+ resource: Resource,
100
+ encoding: str = 'utf-8',
101
+ errors: str = 'strict'
102
+ ) -> str:
103
+ with open(where(), encoding=encoding) as data:
104
+ return data.read()
105
+
106
+ # If we don't have importlib.resources, then we will just do the old logic
107
+ # of assuming we're on the filesystem and munge the path directly.
108
+ def where() -> str:
109
+ f = os.path.dirname(__file__)
110
+
111
+ return os.path.join(f, "cacert.pem")
112
+
113
+ def contents() -> str:
114
+ return read_text("certifi", "cacert.pem", encoding="ascii")
janus/lib/python3.10/site-packages/certifi/py.typed ADDED
File without changes
janus/lib/python3.10/site-packages/jinja2/__init__.py ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """Jinja is a template engine written in pure Python. It provides a
2
+ non-XML syntax that supports inline expressions and an optional
3
+ sandboxed environment.
4
+ """
5
+
6
+ from .bccache import BytecodeCache as BytecodeCache
7
+ from .bccache import FileSystemBytecodeCache as FileSystemBytecodeCache
8
+ from .bccache import MemcachedBytecodeCache as MemcachedBytecodeCache
9
+ from .environment import Environment as Environment
10
+ from .environment import Template as Template
11
+ from .exceptions import TemplateAssertionError as TemplateAssertionError
12
+ from .exceptions import TemplateError as TemplateError
13
+ from .exceptions import TemplateNotFound as TemplateNotFound
14
+ from .exceptions import TemplateRuntimeError as TemplateRuntimeError
15
+ from .exceptions import TemplatesNotFound as TemplatesNotFound
16
+ from .exceptions import TemplateSyntaxError as TemplateSyntaxError
17
+ from .exceptions import UndefinedError as UndefinedError
18
+ from .loaders import BaseLoader as BaseLoader
19
+ from .loaders import ChoiceLoader as ChoiceLoader
20
+ from .loaders import DictLoader as DictLoader
21
+ from .loaders import FileSystemLoader as FileSystemLoader
22
+ from .loaders import FunctionLoader as FunctionLoader
23
+ from .loaders import ModuleLoader as ModuleLoader
24
+ from .loaders import PackageLoader as PackageLoader
25
+ from .loaders import PrefixLoader as PrefixLoader
26
+ from .runtime import ChainableUndefined as ChainableUndefined
27
+ from .runtime import DebugUndefined as DebugUndefined
28
+ from .runtime import make_logging_undefined as make_logging_undefined
29
+ from .runtime import StrictUndefined as StrictUndefined
30
+ from .runtime import Undefined as Undefined
31
+ from .utils import clear_caches as clear_caches
32
+ from .utils import is_undefined as is_undefined
33
+ from .utils import pass_context as pass_context
34
+ from .utils import pass_environment as pass_environment
35
+ from .utils import pass_eval_context as pass_eval_context
36
+ from .utils import select_autoescape as select_autoescape
37
+
38
+ __version__ = "3.1.5"
janus/lib/python3.10/site-packages/jinja2/__pycache__/bccache.cpython-310.pyc ADDED
Binary file (14 kB). View file
 
janus/lib/python3.10/site-packages/jinja2/__pycache__/defaults.cpython-310.pyc ADDED
Binary file (1.32 kB). View file
 
janus/lib/python3.10/site-packages/jinja2/__pycache__/exceptions.cpython-310.pyc ADDED
Binary file (5.52 kB). View file
 
janus/lib/python3.10/site-packages/jinja2/_identifier.py ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ import re
2
+
3
+ # generated by scripts/generate_identifier_pattern.py
4
+ pattern = re.compile(
5
+ r"[\w·̀-ͯ·҃-֑҇-ׇֽֿׁׂׅׄؐ-ًؚ-ٰٟۖ-ۜ۟-۪ۤۧۨ-ܑۭܰ-݊ަ-ް߫-߽߳ࠖ-࠙ࠛ-ࠣࠥ-ࠧࠩ-࡙࠭-࡛࣓-ࣣ࣡-ःऺ-़ा-ॏ॑-ॗॢॣঁ-ঃ়া-ৄেৈো-্ৗৢৣ৾ਁ-ਃ਼ਾ-ੂੇੈੋ-੍ੑੰੱੵઁ-ઃ઼ા-ૅે-ૉો-્ૢૣૺ-૿ଁ-ଃ଼ା-ୄେୈୋ-୍ୖୗୢୣஂா-ூெ-ைொ-்ௗఀ-ఄా-ౄె-ైొ-్ౕౖౢౣಁ-ಃ಼ಾ-ೄೆ-ೈೊ-್ೕೖೢೣഀ-ഃ഻഼ാ-ൄെ-ൈൊ-്ൗൢൣංඃ්ා-ුූෘ-ෟෲෳัิ-ฺ็-๎ັິ-ູົຼ່-ໍ༹༘༙༵༷༾༿ཱ-྄྆྇ྍ-ྗྙ-ྼ࿆ါ-ှၖ-ၙၞ-ၠၢ-ၤၧ-ၭၱ-ၴႂ-ႍႏႚ-ႝ፝-፟ᜒ-᜔ᜲ-᜴ᝒᝓᝲᝳ឴-៓៝᠋-᠍ᢅᢆᢩᤠ-ᤫᤰ-᤻ᨗ-ᨛᩕ-ᩞ᩠-᩿᩼᪰-᪽ᬀ-ᬄ᬴-᭄᭫-᭳ᮀ-ᮂᮡ-ᮭ᯦-᯳ᰤ-᰷᳐-᳔᳒-᳨᳭ᳲ-᳴᳷-᳹᷀-᷹᷻-᷿‿⁀⁔⃐-⃥⃜⃡-⃰℘℮⳯-⵿⳱ⷠ-〪ⷿ-゙゚〯꙯ꙴ-꙽ꚞꚟ꛰꛱ꠂ꠆ꠋꠣ-ꠧꢀꢁꢴ-ꣅ꣠-꣱ꣿꤦ-꤭ꥇ-꥓ꦀ-ꦃ꦳-꧀ꧥꨩ-ꨶꩃꩌꩍꩻ-ꩽꪰꪲ-ꪴꪷꪸꪾ꪿꫁ꫫ-ꫯꫵ꫶ꯣ-ꯪ꯬꯭ﬞ︀-️︠-︯︳︴﹍-﹏_𐇽𐋠𐍶-𐍺𐨁-𐨃𐨅𐨆𐨌-𐨏𐨸-𐨿𐨺𐫦𐫥𐴤-𐽆𐴧-𐽐𑀀-𑀂𑀸-𑁆𑁿-𑂂𑂰-𑂺𑄀-𑄂𑄧-𑄴𑅅𑅆𑅳𑆀-𑆂𑆳-𑇀𑇉-𑇌𑈬-𑈷𑈾𑋟-𑋪𑌀-𑌃𑌻𑌼𑌾-𑍄𑍇𑍈𑍋-𑍍𑍗𑍢𑍣𑍦-𑍬𑍰-𑍴𑐵-𑑆𑑞𑒰-𑓃𑖯-𑖵𑖸-𑗀𑗜𑗝𑘰-𑙀𑚫-𑚷𑜝-𑜫𑠬-𑠺𑨁-𑨊𑨳-𑨹𑨻-𑨾𑩇𑩑-𑩛𑪊-𑪙𑰯-𑰶𑰸-𑰿𑲒-𑲧𑲩-𑲶𑴱-𑴶𑴺𑴼𑴽𑴿-𑵅𑵇𑶊-𑶎𑶐𑶑𑶓-𑶗𑻳-𑻶𖫰-𖫴𖬰-𖬶𖽑-𖽾𖾏-𖾒𛲝𛲞𝅥-𝅩𝅭-𝅲𝅻-𝆂𝆅-𝆋𝆪-𝆭𝉂-𝉄𝨀-𝨶𝨻-𝩬𝩵𝪄𝪛-𝪟𝪡-𝪯𞀀-𞀆𞀈-𞀘𞀛-𞀡𞀣𞀤𞀦-𞣐𞀪-𞣖𞥄-𞥊󠄀-󠇯]+" # noqa: B950
6
+ )
janus/lib/python3.10/site-packages/jinja2/bccache.py ADDED
@@ -0,0 +1,408 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """The optional bytecode cache system. This is useful if you have very
2
+ complex template situations and the compilation of all those templates
3
+ slows down your application too much.
4
+
5
+ Situations where this is useful are often forking web applications that
6
+ are initialized on the first request.
7
+ """
8
+
9
+ import errno
10
+ import fnmatch
11
+ import marshal
12
+ import os
13
+ import pickle
14
+ import stat
15
+ import sys
16
+ import tempfile
17
+ import typing as t
18
+ from hashlib import sha1
19
+ from io import BytesIO
20
+ from types import CodeType
21
+
22
+ if t.TYPE_CHECKING:
23
+ import typing_extensions as te
24
+
25
+ from .environment import Environment
26
+
27
+ class _MemcachedClient(te.Protocol):
28
+ def get(self, key: str) -> bytes: ...
29
+
30
+ def set(
31
+ self, key: str, value: bytes, timeout: t.Optional[int] = None
32
+ ) -> None: ...
33
+
34
+
35
+ bc_version = 5
36
+ # Magic bytes to identify Jinja bytecode cache files. Contains the
37
+ # Python major and minor version to avoid loading incompatible bytecode
38
+ # if a project upgrades its Python version.
39
+ bc_magic = (
40
+ b"j2"
41
+ + pickle.dumps(bc_version, 2)
42
+ + pickle.dumps((sys.version_info[0] << 24) | sys.version_info[1], 2)
43
+ )
44
+
45
+
46
+ class Bucket:
47
+ """Buckets are used to store the bytecode for one template. It's created
48
+ and initialized by the bytecode cache and passed to the loading functions.
49
+
50
+ The buckets get an internal checksum from the cache assigned and use this
51
+ to automatically reject outdated cache material. Individual bytecode
52
+ cache subclasses don't have to care about cache invalidation.
53
+ """
54
+
55
+ def __init__(self, environment: "Environment", key: str, checksum: str) -> None:
56
+ self.environment = environment
57
+ self.key = key
58
+ self.checksum = checksum
59
+ self.reset()
60
+
61
+ def reset(self) -> None:
62
+ """Resets the bucket (unloads the bytecode)."""
63
+ self.code: t.Optional[CodeType] = None
64
+
65
+ def load_bytecode(self, f: t.BinaryIO) -> None:
66
+ """Loads bytecode from a file or file like object."""
67
+ # make sure the magic header is correct
68
+ magic = f.read(len(bc_magic))
69
+ if magic != bc_magic:
70
+ self.reset()
71
+ return
72
+ # the source code of the file changed, we need to reload
73
+ checksum = pickle.load(f)
74
+ if self.checksum != checksum:
75
+ self.reset()
76
+ return
77
+ # if marshal_load fails then we need to reload
78
+ try:
79
+ self.code = marshal.load(f)
80
+ except (EOFError, ValueError, TypeError):
81
+ self.reset()
82
+ return
83
+
84
+ def write_bytecode(self, f: t.IO[bytes]) -> None:
85
+ """Dump the bytecode into the file or file like object passed."""
86
+ if self.code is None:
87
+ raise TypeError("can't write empty bucket")
88
+ f.write(bc_magic)
89
+ pickle.dump(self.checksum, f, 2)
90
+ marshal.dump(self.code, f)
91
+
92
+ def bytecode_from_string(self, string: bytes) -> None:
93
+ """Load bytecode from bytes."""
94
+ self.load_bytecode(BytesIO(string))
95
+
96
+ def bytecode_to_string(self) -> bytes:
97
+ """Return the bytecode as bytes."""
98
+ out = BytesIO()
99
+ self.write_bytecode(out)
100
+ return out.getvalue()
101
+
102
+
103
+ class BytecodeCache:
104
+ """To implement your own bytecode cache you have to subclass this class
105
+ and override :meth:`load_bytecode` and :meth:`dump_bytecode`. Both of
106
+ these methods are passed a :class:`~jinja2.bccache.Bucket`.
107
+
108
+ A very basic bytecode cache that saves the bytecode on the file system::
109
+
110
+ from os import path
111
+
112
+ class MyCache(BytecodeCache):
113
+
114
+ def __init__(self, directory):
115
+ self.directory = directory
116
+
117
+ def load_bytecode(self, bucket):
118
+ filename = path.join(self.directory, bucket.key)
119
+ if path.exists(filename):
120
+ with open(filename, 'rb') as f:
121
+ bucket.load_bytecode(f)
122
+
123
+ def dump_bytecode(self, bucket):
124
+ filename = path.join(self.directory, bucket.key)
125
+ with open(filename, 'wb') as f:
126
+ bucket.write_bytecode(f)
127
+
128
+ A more advanced version of a filesystem based bytecode cache is part of
129
+ Jinja.
130
+ """
131
+
132
+ def load_bytecode(self, bucket: Bucket) -> None:
133
+ """Subclasses have to override this method to load bytecode into a
134
+ bucket. If they are not able to find code in the cache for the
135
+ bucket, it must not do anything.
136
+ """
137
+ raise NotImplementedError()
138
+
139
+ def dump_bytecode(self, bucket: Bucket) -> None:
140
+ """Subclasses have to override this method to write the bytecode
141
+ from a bucket back to the cache. If it unable to do so it must not
142
+ fail silently but raise an exception.
143
+ """
144
+ raise NotImplementedError()
145
+
146
+ def clear(self) -> None:
147
+ """Clears the cache. This method is not used by Jinja but should be
148
+ implemented to allow applications to clear the bytecode cache used
149
+ by a particular environment.
150
+ """
151
+
152
+ def get_cache_key(
153
+ self, name: str, filename: t.Optional[t.Union[str]] = None
154
+ ) -> str:
155
+ """Returns the unique hash key for this template name."""
156
+ hash = sha1(name.encode("utf-8"))
157
+
158
+ if filename is not None:
159
+ hash.update(f"|{filename}".encode())
160
+
161
+ return hash.hexdigest()
162
+
163
+ def get_source_checksum(self, source: str) -> str:
164
+ """Returns a checksum for the source."""
165
+ return sha1(source.encode("utf-8")).hexdigest()
166
+
167
+ def get_bucket(
168
+ self,
169
+ environment: "Environment",
170
+ name: str,
171
+ filename: t.Optional[str],
172
+ source: str,
173
+ ) -> Bucket:
174
+ """Return a cache bucket for the given template. All arguments are
175
+ mandatory but filename may be `None`.
176
+ """
177
+ key = self.get_cache_key(name, filename)
178
+ checksum = self.get_source_checksum(source)
179
+ bucket = Bucket(environment, key, checksum)
180
+ self.load_bytecode(bucket)
181
+ return bucket
182
+
183
+ def set_bucket(self, bucket: Bucket) -> None:
184
+ """Put the bucket into the cache."""
185
+ self.dump_bytecode(bucket)
186
+
187
+
188
+ class FileSystemBytecodeCache(BytecodeCache):
189
+ """A bytecode cache that stores bytecode on the filesystem. It accepts
190
+ two arguments: The directory where the cache items are stored and a
191
+ pattern string that is used to build the filename.
192
+
193
+ If no directory is specified a default cache directory is selected. On
194
+ Windows the user's temp directory is used, on UNIX systems a directory
195
+ is created for the user in the system temp directory.
196
+
197
+ The pattern can be used to have multiple separate caches operate on the
198
+ same directory. The default pattern is ``'__jinja2_%s.cache'``. ``%s``
199
+ is replaced with the cache key.
200
+
201
+ >>> bcc = FileSystemBytecodeCache('/tmp/jinja_cache', '%s.cache')
202
+
203
+ This bytecode cache supports clearing of the cache using the clear method.
204
+ """
205
+
206
+ def __init__(
207
+ self, directory: t.Optional[str] = None, pattern: str = "__jinja2_%s.cache"
208
+ ) -> None:
209
+ if directory is None:
210
+ directory = self._get_default_cache_dir()
211
+ self.directory = directory
212
+ self.pattern = pattern
213
+
214
+ def _get_default_cache_dir(self) -> str:
215
+ def _unsafe_dir() -> "te.NoReturn":
216
+ raise RuntimeError(
217
+ "Cannot determine safe temp directory. You "
218
+ "need to explicitly provide one."
219
+ )
220
+
221
+ tmpdir = tempfile.gettempdir()
222
+
223
+ # On windows the temporary directory is used specific unless
224
+ # explicitly forced otherwise. We can just use that.
225
+ if os.name == "nt":
226
+ return tmpdir
227
+ if not hasattr(os, "getuid"):
228
+ _unsafe_dir()
229
+
230
+ dirname = f"_jinja2-cache-{os.getuid()}"
231
+ actual_dir = os.path.join(tmpdir, dirname)
232
+
233
+ try:
234
+ os.mkdir(actual_dir, stat.S_IRWXU)
235
+ except OSError as e:
236
+ if e.errno != errno.EEXIST:
237
+ raise
238
+ try:
239
+ os.chmod(actual_dir, stat.S_IRWXU)
240
+ actual_dir_stat = os.lstat(actual_dir)
241
+ if (
242
+ actual_dir_stat.st_uid != os.getuid()
243
+ or not stat.S_ISDIR(actual_dir_stat.st_mode)
244
+ or stat.S_IMODE(actual_dir_stat.st_mode) != stat.S_IRWXU
245
+ ):
246
+ _unsafe_dir()
247
+ except OSError as e:
248
+ if e.errno != errno.EEXIST:
249
+ raise
250
+
251
+ actual_dir_stat = os.lstat(actual_dir)
252
+ if (
253
+ actual_dir_stat.st_uid != os.getuid()
254
+ or not stat.S_ISDIR(actual_dir_stat.st_mode)
255
+ or stat.S_IMODE(actual_dir_stat.st_mode) != stat.S_IRWXU
256
+ ):
257
+ _unsafe_dir()
258
+
259
+ return actual_dir
260
+
261
+ def _get_cache_filename(self, bucket: Bucket) -> str:
262
+ return os.path.join(self.directory, self.pattern % (bucket.key,))
263
+
264
+ def load_bytecode(self, bucket: Bucket) -> None:
265
+ filename = self._get_cache_filename(bucket)
266
+
267
+ # Don't test for existence before opening the file, since the
268
+ # file could disappear after the test before the open.
269
+ try:
270
+ f = open(filename, "rb")
271
+ except (FileNotFoundError, IsADirectoryError, PermissionError):
272
+ # PermissionError can occur on Windows when an operation is
273
+ # in progress, such as calling clear().
274
+ return
275
+
276
+ with f:
277
+ bucket.load_bytecode(f)
278
+
279
+ def dump_bytecode(self, bucket: Bucket) -> None:
280
+ # Write to a temporary file, then rename to the real name after
281
+ # writing. This avoids another process reading the file before
282
+ # it is fully written.
283
+ name = self._get_cache_filename(bucket)
284
+ f = tempfile.NamedTemporaryFile(
285
+ mode="wb",
286
+ dir=os.path.dirname(name),
287
+ prefix=os.path.basename(name),
288
+ suffix=".tmp",
289
+ delete=False,
290
+ )
291
+
292
+ def remove_silent() -> None:
293
+ try:
294
+ os.remove(f.name)
295
+ except OSError:
296
+ # Another process may have called clear(). On Windows,
297
+ # another program may be holding the file open.
298
+ pass
299
+
300
+ try:
301
+ with f:
302
+ bucket.write_bytecode(f)
303
+ except BaseException:
304
+ remove_silent()
305
+ raise
306
+
307
+ try:
308
+ os.replace(f.name, name)
309
+ except OSError:
310
+ # Another process may have called clear(). On Windows,
311
+ # another program may be holding the file open.
312
+ remove_silent()
313
+ except BaseException:
314
+ remove_silent()
315
+ raise
316
+
317
+ def clear(self) -> None:
318
+ # imported lazily here because google app-engine doesn't support
319
+ # write access on the file system and the function does not exist
320
+ # normally.
321
+ from os import remove
322
+
323
+ files = fnmatch.filter(os.listdir(self.directory), self.pattern % ("*",))
324
+ for filename in files:
325
+ try:
326
+ remove(os.path.join(self.directory, filename))
327
+ except OSError:
328
+ pass
329
+
330
+
331
+ class MemcachedBytecodeCache(BytecodeCache):
332
+ """This class implements a bytecode cache that uses a memcache cache for
333
+ storing the information. It does not enforce a specific memcache library
334
+ (tummy's memcache or cmemcache) but will accept any class that provides
335
+ the minimal interface required.
336
+
337
+ Libraries compatible with this class:
338
+
339
+ - `cachelib <https://github.com/pallets/cachelib>`_
340
+ - `python-memcached <https://pypi.org/project/python-memcached/>`_
341
+
342
+ (Unfortunately the django cache interface is not compatible because it
343
+ does not support storing binary data, only text. You can however pass
344
+ the underlying cache client to the bytecode cache which is available
345
+ as `django.core.cache.cache._client`.)
346
+
347
+ The minimal interface for the client passed to the constructor is this:
348
+
349
+ .. class:: MinimalClientInterface
350
+
351
+ .. method:: set(key, value[, timeout])
352
+
353
+ Stores the bytecode in the cache. `value` is a string and
354
+ `timeout` the timeout of the key. If timeout is not provided
355
+ a default timeout or no timeout should be assumed, if it's
356
+ provided it's an integer with the number of seconds the cache
357
+ item should exist.
358
+
359
+ .. method:: get(key)
360
+
361
+ Returns the value for the cache key. If the item does not
362
+ exist in the cache the return value must be `None`.
363
+
364
+ The other arguments to the constructor are the prefix for all keys that
365
+ is added before the actual cache key and the timeout for the bytecode in
366
+ the cache system. We recommend a high (or no) timeout.
367
+
368
+ This bytecode cache does not support clearing of used items in the cache.
369
+ The clear method is a no-operation function.
370
+
371
+ .. versionadded:: 2.7
372
+ Added support for ignoring memcache errors through the
373
+ `ignore_memcache_errors` parameter.
374
+ """
375
+
376
+ def __init__(
377
+ self,
378
+ client: "_MemcachedClient",
379
+ prefix: str = "jinja2/bytecode/",
380
+ timeout: t.Optional[int] = None,
381
+ ignore_memcache_errors: bool = True,
382
+ ):
383
+ self.client = client
384
+ self.prefix = prefix
385
+ self.timeout = timeout
386
+ self.ignore_memcache_errors = ignore_memcache_errors
387
+
388
+ def load_bytecode(self, bucket: Bucket) -> None:
389
+ try:
390
+ code = self.client.get(self.prefix + bucket.key)
391
+ except Exception:
392
+ if not self.ignore_memcache_errors:
393
+ raise
394
+ else:
395
+ bucket.bytecode_from_string(code)
396
+
397
+ def dump_bytecode(self, bucket: Bucket) -> None:
398
+ key = self.prefix + bucket.key
399
+ value = bucket.bytecode_to_string()
400
+
401
+ try:
402
+ if self.timeout is not None:
403
+ self.client.set(key, value, self.timeout)
404
+ else:
405
+ self.client.set(key, value)
406
+ except Exception:
407
+ if not self.ignore_memcache_errors:
408
+ raise
janus/lib/python3.10/site-packages/jinja2/compiler.py ADDED
@@ -0,0 +1,1998 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """Compiles nodes from the parser into Python code."""
2
+
3
+ import typing as t
4
+ from contextlib import contextmanager
5
+ from functools import update_wrapper
6
+ from io import StringIO
7
+ from itertools import chain
8
+ from keyword import iskeyword as is_python_keyword
9
+
10
+ from markupsafe import escape
11
+ from markupsafe import Markup
12
+
13
+ from . import nodes
14
+ from .exceptions import TemplateAssertionError
15
+ from .idtracking import Symbols
16
+ from .idtracking import VAR_LOAD_ALIAS
17
+ from .idtracking import VAR_LOAD_PARAMETER
18
+ from .idtracking import VAR_LOAD_RESOLVE
19
+ from .idtracking import VAR_LOAD_UNDEFINED
20
+ from .nodes import EvalContext
21
+ from .optimizer import Optimizer
22
+ from .utils import _PassArg
23
+ from .utils import concat
24
+ from .visitor import NodeVisitor
25
+
26
+ if t.TYPE_CHECKING:
27
+ import typing_extensions as te
28
+
29
+ from .environment import Environment
30
+
31
+ F = t.TypeVar("F", bound=t.Callable[..., t.Any])
32
+
33
+ operators = {
34
+ "eq": "==",
35
+ "ne": "!=",
36
+ "gt": ">",
37
+ "gteq": ">=",
38
+ "lt": "<",
39
+ "lteq": "<=",
40
+ "in": "in",
41
+ "notin": "not in",
42
+ }
43
+
44
+
45
+ def optimizeconst(f: F) -> F:
46
+ def new_func(
47
+ self: "CodeGenerator", node: nodes.Expr, frame: "Frame", **kwargs: t.Any
48
+ ) -> t.Any:
49
+ # Only optimize if the frame is not volatile
50
+ if self.optimizer is not None and not frame.eval_ctx.volatile:
51
+ new_node = self.optimizer.visit(node, frame.eval_ctx)
52
+
53
+ if new_node != node:
54
+ return self.visit(new_node, frame)
55
+
56
+ return f(self, node, frame, **kwargs)
57
+
58
+ return update_wrapper(new_func, f) # type: ignore[return-value]
59
+
60
+
61
+ def _make_binop(op: str) -> t.Callable[["CodeGenerator", nodes.BinExpr, "Frame"], None]:
62
+ @optimizeconst
63
+ def visitor(self: "CodeGenerator", node: nodes.BinExpr, frame: Frame) -> None:
64
+ if (
65
+ self.environment.sandboxed and op in self.environment.intercepted_binops # type: ignore
66
+ ):
67
+ self.write(f"environment.call_binop(context, {op!r}, ")
68
+ self.visit(node.left, frame)
69
+ self.write(", ")
70
+ self.visit(node.right, frame)
71
+ else:
72
+ self.write("(")
73
+ self.visit(node.left, frame)
74
+ self.write(f" {op} ")
75
+ self.visit(node.right, frame)
76
+
77
+ self.write(")")
78
+
79
+ return visitor
80
+
81
+
82
+ def _make_unop(
83
+ op: str,
84
+ ) -> t.Callable[["CodeGenerator", nodes.UnaryExpr, "Frame"], None]:
85
+ @optimizeconst
86
+ def visitor(self: "CodeGenerator", node: nodes.UnaryExpr, frame: Frame) -> None:
87
+ if (
88
+ self.environment.sandboxed and op in self.environment.intercepted_unops # type: ignore
89
+ ):
90
+ self.write(f"environment.call_unop(context, {op!r}, ")
91
+ self.visit(node.node, frame)
92
+ else:
93
+ self.write("(" + op)
94
+ self.visit(node.node, frame)
95
+
96
+ self.write(")")
97
+
98
+ return visitor
99
+
100
+
101
+ def generate(
102
+ node: nodes.Template,
103
+ environment: "Environment",
104
+ name: t.Optional[str],
105
+ filename: t.Optional[str],
106
+ stream: t.Optional[t.TextIO] = None,
107
+ defer_init: bool = False,
108
+ optimized: bool = True,
109
+ ) -> t.Optional[str]:
110
+ """Generate the python source for a node tree."""
111
+ if not isinstance(node, nodes.Template):
112
+ raise TypeError("Can't compile non template nodes")
113
+
114
+ generator = environment.code_generator_class(
115
+ environment, name, filename, stream, defer_init, optimized
116
+ )
117
+ generator.visit(node)
118
+
119
+ if stream is None:
120
+ return generator.stream.getvalue() # type: ignore
121
+
122
+ return None
123
+
124
+
125
+ def has_safe_repr(value: t.Any) -> bool:
126
+ """Does the node have a safe representation?"""
127
+ if value is None or value is NotImplemented or value is Ellipsis:
128
+ return True
129
+
130
+ if type(value) in {bool, int, float, complex, range, str, Markup}:
131
+ return True
132
+
133
+ if type(value) in {tuple, list, set, frozenset}:
134
+ return all(has_safe_repr(v) for v in value)
135
+
136
+ if type(value) is dict: # noqa E721
137
+ return all(has_safe_repr(k) and has_safe_repr(v) for k, v in value.items())
138
+
139
+ return False
140
+
141
+
142
+ def find_undeclared(
143
+ nodes: t.Iterable[nodes.Node], names: t.Iterable[str]
144
+ ) -> t.Set[str]:
145
+ """Check if the names passed are accessed undeclared. The return value
146
+ is a set of all the undeclared names from the sequence of names found.
147
+ """
148
+ visitor = UndeclaredNameVisitor(names)
149
+ try:
150
+ for node in nodes:
151
+ visitor.visit(node)
152
+ except VisitorExit:
153
+ pass
154
+ return visitor.undeclared
155
+
156
+
157
+ class MacroRef:
158
+ def __init__(self, node: t.Union[nodes.Macro, nodes.CallBlock]) -> None:
159
+ self.node = node
160
+ self.accesses_caller = False
161
+ self.accesses_kwargs = False
162
+ self.accesses_varargs = False
163
+
164
+
165
+ class Frame:
166
+ """Holds compile time information for us."""
167
+
168
+ def __init__(
169
+ self,
170
+ eval_ctx: EvalContext,
171
+ parent: t.Optional["Frame"] = None,
172
+ level: t.Optional[int] = None,
173
+ ) -> None:
174
+ self.eval_ctx = eval_ctx
175
+
176
+ # the parent of this frame
177
+ self.parent = parent
178
+
179
+ if parent is None:
180
+ self.symbols = Symbols(level=level)
181
+
182
+ # in some dynamic inheritance situations the compiler needs to add
183
+ # write tests around output statements.
184
+ self.require_output_check = False
185
+
186
+ # inside some tags we are using a buffer rather than yield statements.
187
+ # this for example affects {% filter %} or {% macro %}. If a frame
188
+ # is buffered this variable points to the name of the list used as
189
+ # buffer.
190
+ self.buffer: t.Optional[str] = None
191
+
192
+ # the name of the block we're in, otherwise None.
193
+ self.block: t.Optional[str] = None
194
+
195
+ else:
196
+ self.symbols = Symbols(parent.symbols, level=level)
197
+ self.require_output_check = parent.require_output_check
198
+ self.buffer = parent.buffer
199
+ self.block = parent.block
200
+
201
+ # a toplevel frame is the root + soft frames such as if conditions.
202
+ self.toplevel = False
203
+
204
+ # the root frame is basically just the outermost frame, so no if
205
+ # conditions. This information is used to optimize inheritance
206
+ # situations.
207
+ self.rootlevel = False
208
+
209
+ # variables set inside of loops and blocks should not affect outer frames,
210
+ # but they still needs to be kept track of as part of the active context.
211
+ self.loop_frame = False
212
+ self.block_frame = False
213
+
214
+ # track whether the frame is being used in an if-statement or conditional
215
+ # expression as it determines which errors should be raised during runtime
216
+ # or compile time.
217
+ self.soft_frame = False
218
+
219
+ def copy(self) -> "te.Self":
220
+ """Create a copy of the current one."""
221
+ rv = object.__new__(self.__class__)
222
+ rv.__dict__.update(self.__dict__)
223
+ rv.symbols = self.symbols.copy()
224
+ return rv
225
+
226
+ def inner(self, isolated: bool = False) -> "Frame":
227
+ """Return an inner frame."""
228
+ if isolated:
229
+ return Frame(self.eval_ctx, level=self.symbols.level + 1)
230
+ return Frame(self.eval_ctx, self)
231
+
232
+ def soft(self) -> "te.Self":
233
+ """Return a soft frame. A soft frame may not be modified as
234
+ standalone thing as it shares the resources with the frame it
235
+ was created of, but it's not a rootlevel frame any longer.
236
+
237
+ This is only used to implement if-statements and conditional
238
+ expressions.
239
+ """
240
+ rv = self.copy()
241
+ rv.rootlevel = False
242
+ rv.soft_frame = True
243
+ return rv
244
+
245
+ __copy__ = copy
246
+
247
+
248
+ class VisitorExit(RuntimeError):
249
+ """Exception used by the `UndeclaredNameVisitor` to signal a stop."""
250
+
251
+
252
+ class DependencyFinderVisitor(NodeVisitor):
253
+ """A visitor that collects filter and test calls."""
254
+
255
+ def __init__(self) -> None:
256
+ self.filters: t.Set[str] = set()
257
+ self.tests: t.Set[str] = set()
258
+
259
+ def visit_Filter(self, node: nodes.Filter) -> None:
260
+ self.generic_visit(node)
261
+ self.filters.add(node.name)
262
+
263
+ def visit_Test(self, node: nodes.Test) -> None:
264
+ self.generic_visit(node)
265
+ self.tests.add(node.name)
266
+
267
+ def visit_Block(self, node: nodes.Block) -> None:
268
+ """Stop visiting at blocks."""
269
+
270
+
271
+ class UndeclaredNameVisitor(NodeVisitor):
272
+ """A visitor that checks if a name is accessed without being
273
+ declared. This is different from the frame visitor as it will
274
+ not stop at closure frames.
275
+ """
276
+
277
+ def __init__(self, names: t.Iterable[str]) -> None:
278
+ self.names = set(names)
279
+ self.undeclared: t.Set[str] = set()
280
+
281
+ def visit_Name(self, node: nodes.Name) -> None:
282
+ if node.ctx == "load" and node.name in self.names:
283
+ self.undeclared.add(node.name)
284
+ if self.undeclared == self.names:
285
+ raise VisitorExit()
286
+ else:
287
+ self.names.discard(node.name)
288
+
289
+ def visit_Block(self, node: nodes.Block) -> None:
290
+ """Stop visiting a blocks."""
291
+
292
+
293
+ class CompilerExit(Exception):
294
+ """Raised if the compiler encountered a situation where it just
295
+ doesn't make sense to further process the code. Any block that
296
+ raises such an exception is not further processed.
297
+ """
298
+
299
+
300
+ class CodeGenerator(NodeVisitor):
301
+ def __init__(
302
+ self,
303
+ environment: "Environment",
304
+ name: t.Optional[str],
305
+ filename: t.Optional[str],
306
+ stream: t.Optional[t.TextIO] = None,
307
+ defer_init: bool = False,
308
+ optimized: bool = True,
309
+ ) -> None:
310
+ if stream is None:
311
+ stream = StringIO()
312
+ self.environment = environment
313
+ self.name = name
314
+ self.filename = filename
315
+ self.stream = stream
316
+ self.created_block_context = False
317
+ self.defer_init = defer_init
318
+ self.optimizer: t.Optional[Optimizer] = None
319
+
320
+ if optimized:
321
+ self.optimizer = Optimizer(environment)
322
+
323
+ # aliases for imports
324
+ self.import_aliases: t.Dict[str, str] = {}
325
+
326
+ # a registry for all blocks. Because blocks are moved out
327
+ # into the global python scope they are registered here
328
+ self.blocks: t.Dict[str, nodes.Block] = {}
329
+
330
+ # the number of extends statements so far
331
+ self.extends_so_far = 0
332
+
333
+ # some templates have a rootlevel extends. In this case we
334
+ # can safely assume that we're a child template and do some
335
+ # more optimizations.
336
+ self.has_known_extends = False
337
+
338
+ # the current line number
339
+ self.code_lineno = 1
340
+
341
+ # registry of all filters and tests (global, not block local)
342
+ self.tests: t.Dict[str, str] = {}
343
+ self.filters: t.Dict[str, str] = {}
344
+
345
+ # the debug information
346
+ self.debug_info: t.List[t.Tuple[int, int]] = []
347
+ self._write_debug_info: t.Optional[int] = None
348
+
349
+ # the number of new lines before the next write()
350
+ self._new_lines = 0
351
+
352
+ # the line number of the last written statement
353
+ self._last_line = 0
354
+
355
+ # true if nothing was written so far.
356
+ self._first_write = True
357
+
358
+ # used by the `temporary_identifier` method to get new
359
+ # unique, temporary identifier
360
+ self._last_identifier = 0
361
+
362
+ # the current indentation
363
+ self._indentation = 0
364
+
365
+ # Tracks toplevel assignments
366
+ self._assign_stack: t.List[t.Set[str]] = []
367
+
368
+ # Tracks parameter definition blocks
369
+ self._param_def_block: t.List[t.Set[str]] = []
370
+
371
+ # Tracks the current context.
372
+ self._context_reference_stack = ["context"]
373
+
374
+ @property
375
+ def optimized(self) -> bool:
376
+ return self.optimizer is not None
377
+
378
+ # -- Various compilation helpers
379
+
380
+ def fail(self, msg: str, lineno: int) -> "te.NoReturn":
381
+ """Fail with a :exc:`TemplateAssertionError`."""
382
+ raise TemplateAssertionError(msg, lineno, self.name, self.filename)
383
+
384
+ def temporary_identifier(self) -> str:
385
+ """Get a new unique identifier."""
386
+ self._last_identifier += 1
387
+ return f"t_{self._last_identifier}"
388
+
389
+ def buffer(self, frame: Frame) -> None:
390
+ """Enable buffering for the frame from that point onwards."""
391
+ frame.buffer = self.temporary_identifier()
392
+ self.writeline(f"{frame.buffer} = []")
393
+
394
+ def return_buffer_contents(
395
+ self, frame: Frame, force_unescaped: bool = False
396
+ ) -> None:
397
+ """Return the buffer contents of the frame."""
398
+ if not force_unescaped:
399
+ if frame.eval_ctx.volatile:
400
+ self.writeline("if context.eval_ctx.autoescape:")
401
+ self.indent()
402
+ self.writeline(f"return Markup(concat({frame.buffer}))")
403
+ self.outdent()
404
+ self.writeline("else:")
405
+ self.indent()
406
+ self.writeline(f"return concat({frame.buffer})")
407
+ self.outdent()
408
+ return
409
+ elif frame.eval_ctx.autoescape:
410
+ self.writeline(f"return Markup(concat({frame.buffer}))")
411
+ return
412
+ self.writeline(f"return concat({frame.buffer})")
413
+
414
+ def indent(self) -> None:
415
+ """Indent by one."""
416
+ self._indentation += 1
417
+
418
+ def outdent(self, step: int = 1) -> None:
419
+ """Outdent by step."""
420
+ self._indentation -= step
421
+
422
+ def start_write(self, frame: Frame, node: t.Optional[nodes.Node] = None) -> None:
423
+ """Yield or write into the frame buffer."""
424
+ if frame.buffer is None:
425
+ self.writeline("yield ", node)
426
+ else:
427
+ self.writeline(f"{frame.buffer}.append(", node)
428
+
429
+ def end_write(self, frame: Frame) -> None:
430
+ """End the writing process started by `start_write`."""
431
+ if frame.buffer is not None:
432
+ self.write(")")
433
+
434
+ def simple_write(
435
+ self, s: str, frame: Frame, node: t.Optional[nodes.Node] = None
436
+ ) -> None:
437
+ """Simple shortcut for start_write + write + end_write."""
438
+ self.start_write(frame, node)
439
+ self.write(s)
440
+ self.end_write(frame)
441
+
442
+ def blockvisit(self, nodes: t.Iterable[nodes.Node], frame: Frame) -> None:
443
+ """Visit a list of nodes as block in a frame. If the current frame
444
+ is no buffer a dummy ``if 0: yield None`` is written automatically.
445
+ """
446
+ try:
447
+ self.writeline("pass")
448
+ for node in nodes:
449
+ self.visit(node, frame)
450
+ except CompilerExit:
451
+ pass
452
+
453
+ def write(self, x: str) -> None:
454
+ """Write a string into the output stream."""
455
+ if self._new_lines:
456
+ if not self._first_write:
457
+ self.stream.write("\n" * self._new_lines)
458
+ self.code_lineno += self._new_lines
459
+ if self._write_debug_info is not None:
460
+ self.debug_info.append((self._write_debug_info, self.code_lineno))
461
+ self._write_debug_info = None
462
+ self._first_write = False
463
+ self.stream.write(" " * self._indentation)
464
+ self._new_lines = 0
465
+ self.stream.write(x)
466
+
467
+ def writeline(
468
+ self, x: str, node: t.Optional[nodes.Node] = None, extra: int = 0
469
+ ) -> None:
470
+ """Combination of newline and write."""
471
+ self.newline(node, extra)
472
+ self.write(x)
473
+
474
+ def newline(self, node: t.Optional[nodes.Node] = None, extra: int = 0) -> None:
475
+ """Add one or more newlines before the next write."""
476
+ self._new_lines = max(self._new_lines, 1 + extra)
477
+ if node is not None and node.lineno != self._last_line:
478
+ self._write_debug_info = node.lineno
479
+ self._last_line = node.lineno
480
+
481
+ def signature(
482
+ self,
483
+ node: t.Union[nodes.Call, nodes.Filter, nodes.Test],
484
+ frame: Frame,
485
+ extra_kwargs: t.Optional[t.Mapping[str, t.Any]] = None,
486
+ ) -> None:
487
+ """Writes a function call to the stream for the current node.
488
+ A leading comma is added automatically. The extra keyword
489
+ arguments may not include python keywords otherwise a syntax
490
+ error could occur. The extra keyword arguments should be given
491
+ as python dict.
492
+ """
493
+ # if any of the given keyword arguments is a python keyword
494
+ # we have to make sure that no invalid call is created.
495
+ kwarg_workaround = any(
496
+ is_python_keyword(t.cast(str, k))
497
+ for k in chain((x.key for x in node.kwargs), extra_kwargs or ())
498
+ )
499
+
500
+ for arg in node.args:
501
+ self.write(", ")
502
+ self.visit(arg, frame)
503
+
504
+ if not kwarg_workaround:
505
+ for kwarg in node.kwargs:
506
+ self.write(", ")
507
+ self.visit(kwarg, frame)
508
+ if extra_kwargs is not None:
509
+ for key, value in extra_kwargs.items():
510
+ self.write(f", {key}={value}")
511
+ if node.dyn_args:
512
+ self.write(", *")
513
+ self.visit(node.dyn_args, frame)
514
+
515
+ if kwarg_workaround:
516
+ if node.dyn_kwargs is not None:
517
+ self.write(", **dict({")
518
+ else:
519
+ self.write(", **{")
520
+ for kwarg in node.kwargs:
521
+ self.write(f"{kwarg.key!r}: ")
522
+ self.visit(kwarg.value, frame)
523
+ self.write(", ")
524
+ if extra_kwargs is not None:
525
+ for key, value in extra_kwargs.items():
526
+ self.write(f"{key!r}: {value}, ")
527
+ if node.dyn_kwargs is not None:
528
+ self.write("}, **")
529
+ self.visit(node.dyn_kwargs, frame)
530
+ self.write(")")
531
+ else:
532
+ self.write("}")
533
+
534
+ elif node.dyn_kwargs is not None:
535
+ self.write(", **")
536
+ self.visit(node.dyn_kwargs, frame)
537
+
538
+ def pull_dependencies(self, nodes: t.Iterable[nodes.Node]) -> None:
539
+ """Find all filter and test names used in the template and
540
+ assign them to variables in the compiled namespace. Checking
541
+ that the names are registered with the environment is done when
542
+ compiling the Filter and Test nodes. If the node is in an If or
543
+ CondExpr node, the check is done at runtime instead.
544
+
545
+ .. versionchanged:: 3.0
546
+ Filters and tests in If and CondExpr nodes are checked at
547
+ runtime instead of compile time.
548
+ """
549
+ visitor = DependencyFinderVisitor()
550
+
551
+ for node in nodes:
552
+ visitor.visit(node)
553
+
554
+ for id_map, names, dependency in (
555
+ (self.filters, visitor.filters, "filters"),
556
+ (
557
+ self.tests,
558
+ visitor.tests,
559
+ "tests",
560
+ ),
561
+ ):
562
+ for name in sorted(names):
563
+ if name not in id_map:
564
+ id_map[name] = self.temporary_identifier()
565
+
566
+ # add check during runtime that dependencies used inside of executed
567
+ # blocks are defined, as this step may be skipped during compile time
568
+ self.writeline("try:")
569
+ self.indent()
570
+ self.writeline(f"{id_map[name]} = environment.{dependency}[{name!r}]")
571
+ self.outdent()
572
+ self.writeline("except KeyError:")
573
+ self.indent()
574
+ self.writeline("@internalcode")
575
+ self.writeline(f"def {id_map[name]}(*unused):")
576
+ self.indent()
577
+ self.writeline(
578
+ f'raise TemplateRuntimeError("No {dependency[:-1]}'
579
+ f' named {name!r} found.")'
580
+ )
581
+ self.outdent()
582
+ self.outdent()
583
+
584
+ def enter_frame(self, frame: Frame) -> None:
585
+ undefs = []
586
+ for target, (action, param) in frame.symbols.loads.items():
587
+ if action == VAR_LOAD_PARAMETER:
588
+ pass
589
+ elif action == VAR_LOAD_RESOLVE:
590
+ self.writeline(f"{target} = {self.get_resolve_func()}({param!r})")
591
+ elif action == VAR_LOAD_ALIAS:
592
+ self.writeline(f"{target} = {param}")
593
+ elif action == VAR_LOAD_UNDEFINED:
594
+ undefs.append(target)
595
+ else:
596
+ raise NotImplementedError("unknown load instruction")
597
+ if undefs:
598
+ self.writeline(f"{' = '.join(undefs)} = missing")
599
+
600
+ def leave_frame(self, frame: Frame, with_python_scope: bool = False) -> None:
601
+ if not with_python_scope:
602
+ undefs = []
603
+ for target in frame.symbols.loads:
604
+ undefs.append(target)
605
+ if undefs:
606
+ self.writeline(f"{' = '.join(undefs)} = missing")
607
+
608
+ def choose_async(self, async_value: str = "async ", sync_value: str = "") -> str:
609
+ return async_value if self.environment.is_async else sync_value
610
+
611
+ def func(self, name: str) -> str:
612
+ return f"{self.choose_async()}def {name}"
613
+
614
+ def macro_body(
615
+ self, node: t.Union[nodes.Macro, nodes.CallBlock], frame: Frame
616
+ ) -> t.Tuple[Frame, MacroRef]:
617
+ """Dump the function def of a macro or call block."""
618
+ frame = frame.inner()
619
+ frame.symbols.analyze_node(node)
620
+ macro_ref = MacroRef(node)
621
+
622
+ explicit_caller = None
623
+ skip_special_params = set()
624
+ args = []
625
+
626
+ for idx, arg in enumerate(node.args):
627
+ if arg.name == "caller":
628
+ explicit_caller = idx
629
+ if arg.name in ("kwargs", "varargs"):
630
+ skip_special_params.add(arg.name)
631
+ args.append(frame.symbols.ref(arg.name))
632
+
633
+ undeclared = find_undeclared(node.body, ("caller", "kwargs", "varargs"))
634
+
635
+ if "caller" in undeclared:
636
+ # In older Jinja versions there was a bug that allowed caller
637
+ # to retain the special behavior even if it was mentioned in
638
+ # the argument list. However thankfully this was only really
639
+ # working if it was the last argument. So we are explicitly
640
+ # checking this now and error out if it is anywhere else in
641
+ # the argument list.
642
+ if explicit_caller is not None:
643
+ try:
644
+ node.defaults[explicit_caller - len(node.args)]
645
+ except IndexError:
646
+ self.fail(
647
+ "When defining macros or call blocks the "
648
+ 'special "caller" argument must be omitted '
649
+ "or be given a default.",
650
+ node.lineno,
651
+ )
652
+ else:
653
+ args.append(frame.symbols.declare_parameter("caller"))
654
+ macro_ref.accesses_caller = True
655
+ if "kwargs" in undeclared and "kwargs" not in skip_special_params:
656
+ args.append(frame.symbols.declare_parameter("kwargs"))
657
+ macro_ref.accesses_kwargs = True
658
+ if "varargs" in undeclared and "varargs" not in skip_special_params:
659
+ args.append(frame.symbols.declare_parameter("varargs"))
660
+ macro_ref.accesses_varargs = True
661
+
662
+ # macros are delayed, they never require output checks
663
+ frame.require_output_check = False
664
+ frame.symbols.analyze_node(node)
665
+ self.writeline(f"{self.func('macro')}({', '.join(args)}):", node)
666
+ self.indent()
667
+
668
+ self.buffer(frame)
669
+ self.enter_frame(frame)
670
+
671
+ self.push_parameter_definitions(frame)
672
+ for idx, arg in enumerate(node.args):
673
+ ref = frame.symbols.ref(arg.name)
674
+ self.writeline(f"if {ref} is missing:")
675
+ self.indent()
676
+ try:
677
+ default = node.defaults[idx - len(node.args)]
678
+ except IndexError:
679
+ self.writeline(
680
+ f'{ref} = undefined("parameter {arg.name!r} was not provided",'
681
+ f" name={arg.name!r})"
682
+ )
683
+ else:
684
+ self.writeline(f"{ref} = ")
685
+ self.visit(default, frame)
686
+ self.mark_parameter_stored(ref)
687
+ self.outdent()
688
+ self.pop_parameter_definitions()
689
+
690
+ self.blockvisit(node.body, frame)
691
+ self.return_buffer_contents(frame, force_unescaped=True)
692
+ self.leave_frame(frame, with_python_scope=True)
693
+ self.outdent()
694
+
695
+ return frame, macro_ref
696
+
697
+ def macro_def(self, macro_ref: MacroRef, frame: Frame) -> None:
698
+ """Dump the macro definition for the def created by macro_body."""
699
+ arg_tuple = ", ".join(repr(x.name) for x in macro_ref.node.args)
700
+ name = getattr(macro_ref.node, "name", None)
701
+ if len(macro_ref.node.args) == 1:
702
+ arg_tuple += ","
703
+ self.write(
704
+ f"Macro(environment, macro, {name!r}, ({arg_tuple}),"
705
+ f" {macro_ref.accesses_kwargs!r}, {macro_ref.accesses_varargs!r},"
706
+ f" {macro_ref.accesses_caller!r}, context.eval_ctx.autoescape)"
707
+ )
708
+
709
+ def position(self, node: nodes.Node) -> str:
710
+ """Return a human readable position for the node."""
711
+ rv = f"line {node.lineno}"
712
+ if self.name is not None:
713
+ rv = f"{rv} in {self.name!r}"
714
+ return rv
715
+
716
+ def dump_local_context(self, frame: Frame) -> str:
717
+ items_kv = ", ".join(
718
+ f"{name!r}: {target}"
719
+ for name, target in frame.symbols.dump_stores().items()
720
+ )
721
+ return f"{{{items_kv}}}"
722
+
723
+ def write_commons(self) -> None:
724
+ """Writes a common preamble that is used by root and block functions.
725
+ Primarily this sets up common local helpers and enforces a generator
726
+ through a dead branch.
727
+ """
728
+ self.writeline("resolve = context.resolve_or_missing")
729
+ self.writeline("undefined = environment.undefined")
730
+ self.writeline("concat = environment.concat")
731
+ # always use the standard Undefined class for the implicit else of
732
+ # conditional expressions
733
+ self.writeline("cond_expr_undefined = Undefined")
734
+ self.writeline("if 0: yield None")
735
+
736
+ def push_parameter_definitions(self, frame: Frame) -> None:
737
+ """Pushes all parameter targets from the given frame into a local
738
+ stack that permits tracking of yet to be assigned parameters. In
739
+ particular this enables the optimization from `visit_Name` to skip
740
+ undefined expressions for parameters in macros as macros can reference
741
+ otherwise unbound parameters.
742
+ """
743
+ self._param_def_block.append(frame.symbols.dump_param_targets())
744
+
745
+ def pop_parameter_definitions(self) -> None:
746
+ """Pops the current parameter definitions set."""
747
+ self._param_def_block.pop()
748
+
749
+ def mark_parameter_stored(self, target: str) -> None:
750
+ """Marks a parameter in the current parameter definitions as stored.
751
+ This will skip the enforced undefined checks.
752
+ """
753
+ if self._param_def_block:
754
+ self._param_def_block[-1].discard(target)
755
+
756
+ def push_context_reference(self, target: str) -> None:
757
+ self._context_reference_stack.append(target)
758
+
759
+ def pop_context_reference(self) -> None:
760
+ self._context_reference_stack.pop()
761
+
762
+ def get_context_ref(self) -> str:
763
+ return self._context_reference_stack[-1]
764
+
765
+ def get_resolve_func(self) -> str:
766
+ target = self._context_reference_stack[-1]
767
+ if target == "context":
768
+ return "resolve"
769
+ return f"{target}.resolve"
770
+
771
+ def derive_context(self, frame: Frame) -> str:
772
+ return f"{self.get_context_ref()}.derived({self.dump_local_context(frame)})"
773
+
774
+ def parameter_is_undeclared(self, target: str) -> bool:
775
+ """Checks if a given target is an undeclared parameter."""
776
+ if not self._param_def_block:
777
+ return False
778
+ return target in self._param_def_block[-1]
779
+
780
+ def push_assign_tracking(self) -> None:
781
+ """Pushes a new layer for assignment tracking."""
782
+ self._assign_stack.append(set())
783
+
784
+ def pop_assign_tracking(self, frame: Frame) -> None:
785
+ """Pops the topmost level for assignment tracking and updates the
786
+ context variables if necessary.
787
+ """
788
+ vars = self._assign_stack.pop()
789
+ if (
790
+ not frame.block_frame
791
+ and not frame.loop_frame
792
+ and not frame.toplevel
793
+ or not vars
794
+ ):
795
+ return
796
+ public_names = [x for x in vars if x[:1] != "_"]
797
+ if len(vars) == 1:
798
+ name = next(iter(vars))
799
+ ref = frame.symbols.ref(name)
800
+ if frame.loop_frame:
801
+ self.writeline(f"_loop_vars[{name!r}] = {ref}")
802
+ return
803
+ if frame.block_frame:
804
+ self.writeline(f"_block_vars[{name!r}] = {ref}")
805
+ return
806
+ self.writeline(f"context.vars[{name!r}] = {ref}")
807
+ else:
808
+ if frame.loop_frame:
809
+ self.writeline("_loop_vars.update({")
810
+ elif frame.block_frame:
811
+ self.writeline("_block_vars.update({")
812
+ else:
813
+ self.writeline("context.vars.update({")
814
+ for idx, name in enumerate(sorted(vars)):
815
+ if idx:
816
+ self.write(", ")
817
+ ref = frame.symbols.ref(name)
818
+ self.write(f"{name!r}: {ref}")
819
+ self.write("})")
820
+ if not frame.block_frame and not frame.loop_frame and public_names:
821
+ if len(public_names) == 1:
822
+ self.writeline(f"context.exported_vars.add({public_names[0]!r})")
823
+ else:
824
+ names_str = ", ".join(map(repr, sorted(public_names)))
825
+ self.writeline(f"context.exported_vars.update(({names_str}))")
826
+
827
+ # -- Statement Visitors
828
+
829
+ def visit_Template(
830
+ self, node: nodes.Template, frame: t.Optional[Frame] = None
831
+ ) -> None:
832
+ assert frame is None, "no root frame allowed"
833
+ eval_ctx = EvalContext(self.environment, self.name)
834
+
835
+ from .runtime import async_exported
836
+ from .runtime import exported
837
+
838
+ if self.environment.is_async:
839
+ exported_names = sorted(exported + async_exported)
840
+ else:
841
+ exported_names = sorted(exported)
842
+
843
+ self.writeline("from jinja2.runtime import " + ", ".join(exported_names))
844
+
845
+ # if we want a deferred initialization we cannot move the
846
+ # environment into a local name
847
+ envenv = "" if self.defer_init else ", environment=environment"
848
+
849
+ # do we have an extends tag at all? If not, we can save some
850
+ # overhead by just not processing any inheritance code.
851
+ have_extends = node.find(nodes.Extends) is not None
852
+
853
+ # find all blocks
854
+ for block in node.find_all(nodes.Block):
855
+ if block.name in self.blocks:
856
+ self.fail(f"block {block.name!r} defined twice", block.lineno)
857
+ self.blocks[block.name] = block
858
+
859
+ # find all imports and import them
860
+ for import_ in node.find_all(nodes.ImportedName):
861
+ if import_.importname not in self.import_aliases:
862
+ imp = import_.importname
863
+ self.import_aliases[imp] = alias = self.temporary_identifier()
864
+ if "." in imp:
865
+ module, obj = imp.rsplit(".", 1)
866
+ self.writeline(f"from {module} import {obj} as {alias}")
867
+ else:
868
+ self.writeline(f"import {imp} as {alias}")
869
+
870
+ # add the load name
871
+ self.writeline(f"name = {self.name!r}")
872
+
873
+ # generate the root render function.
874
+ self.writeline(
875
+ f"{self.func('root')}(context, missing=missing{envenv}):", extra=1
876
+ )
877
+ self.indent()
878
+ self.write_commons()
879
+
880
+ # process the root
881
+ frame = Frame(eval_ctx)
882
+ if "self" in find_undeclared(node.body, ("self",)):
883
+ ref = frame.symbols.declare_parameter("self")
884
+ self.writeline(f"{ref} = TemplateReference(context)")
885
+ frame.symbols.analyze_node(node)
886
+ frame.toplevel = frame.rootlevel = True
887
+ frame.require_output_check = have_extends and not self.has_known_extends
888
+ if have_extends:
889
+ self.writeline("parent_template = None")
890
+ self.enter_frame(frame)
891
+ self.pull_dependencies(node.body)
892
+ self.blockvisit(node.body, frame)
893
+ self.leave_frame(frame, with_python_scope=True)
894
+ self.outdent()
895
+
896
+ # make sure that the parent root is called.
897
+ if have_extends:
898
+ if not self.has_known_extends:
899
+ self.indent()
900
+ self.writeline("if parent_template is not None:")
901
+ self.indent()
902
+ if not self.environment.is_async:
903
+ self.writeline("yield from parent_template.root_render_func(context)")
904
+ else:
905
+ self.writeline("agen = parent_template.root_render_func(context)")
906
+ self.writeline("try:")
907
+ self.indent()
908
+ self.writeline("async for event in agen:")
909
+ self.indent()
910
+ self.writeline("yield event")
911
+ self.outdent()
912
+ self.outdent()
913
+ self.writeline("finally: await agen.aclose()")
914
+ self.outdent(1 + (not self.has_known_extends))
915
+
916
+ # at this point we now have the blocks collected and can visit them too.
917
+ for name, block in self.blocks.items():
918
+ self.writeline(
919
+ f"{self.func('block_' + name)}(context, missing=missing{envenv}):",
920
+ block,
921
+ 1,
922
+ )
923
+ self.indent()
924
+ self.write_commons()
925
+ # It's important that we do not make this frame a child of the
926
+ # toplevel template. This would cause a variety of
927
+ # interesting issues with identifier tracking.
928
+ block_frame = Frame(eval_ctx)
929
+ block_frame.block_frame = True
930
+ undeclared = find_undeclared(block.body, ("self", "super"))
931
+ if "self" in undeclared:
932
+ ref = block_frame.symbols.declare_parameter("self")
933
+ self.writeline(f"{ref} = TemplateReference(context)")
934
+ if "super" in undeclared:
935
+ ref = block_frame.symbols.declare_parameter("super")
936
+ self.writeline(f"{ref} = context.super({name!r}, block_{name})")
937
+ block_frame.symbols.analyze_node(block)
938
+ block_frame.block = name
939
+ self.writeline("_block_vars = {}")
940
+ self.enter_frame(block_frame)
941
+ self.pull_dependencies(block.body)
942
+ self.blockvisit(block.body, block_frame)
943
+ self.leave_frame(block_frame, with_python_scope=True)
944
+ self.outdent()
945
+
946
+ blocks_kv_str = ", ".join(f"{x!r}: block_{x}" for x in self.blocks)
947
+ self.writeline(f"blocks = {{{blocks_kv_str}}}", extra=1)
948
+ debug_kv_str = "&".join(f"{k}={v}" for k, v in self.debug_info)
949
+ self.writeline(f"debug_info = {debug_kv_str!r}")
950
+
951
+ def visit_Block(self, node: nodes.Block, frame: Frame) -> None:
952
+ """Call a block and register it for the template."""
953
+ level = 0
954
+ if frame.toplevel:
955
+ # if we know that we are a child template, there is no need to
956
+ # check if we are one
957
+ if self.has_known_extends:
958
+ return
959
+ if self.extends_so_far > 0:
960
+ self.writeline("if parent_template is None:")
961
+ self.indent()
962
+ level += 1
963
+
964
+ if node.scoped:
965
+ context = self.derive_context(frame)
966
+ else:
967
+ context = self.get_context_ref()
968
+
969
+ if node.required:
970
+ self.writeline(f"if len(context.blocks[{node.name!r}]) <= 1:", node)
971
+ self.indent()
972
+ self.writeline(
973
+ f'raise TemplateRuntimeError("Required block {node.name!r} not found")',
974
+ node,
975
+ )
976
+ self.outdent()
977
+
978
+ if not self.environment.is_async and frame.buffer is None:
979
+ self.writeline(
980
+ f"yield from context.blocks[{node.name!r}][0]({context})", node
981
+ )
982
+ else:
983
+ self.writeline(f"gen = context.blocks[{node.name!r}][0]({context})")
984
+ self.writeline("try:")
985
+ self.indent()
986
+ self.writeline(
987
+ f"{self.choose_async()}for event in gen:",
988
+ node,
989
+ )
990
+ self.indent()
991
+ self.simple_write("event", frame)
992
+ self.outdent()
993
+ self.outdent()
994
+ self.writeline(
995
+ f"finally: {self.choose_async('await gen.aclose()', 'gen.close()')}"
996
+ )
997
+
998
+ self.outdent(level)
999
+
1000
+ def visit_Extends(self, node: nodes.Extends, frame: Frame) -> None:
1001
+ """Calls the extender."""
1002
+ if not frame.toplevel:
1003
+ self.fail("cannot use extend from a non top-level scope", node.lineno)
1004
+
1005
+ # if the number of extends statements in general is zero so
1006
+ # far, we don't have to add a check if something extended
1007
+ # the template before this one.
1008
+ if self.extends_so_far > 0:
1009
+ # if we have a known extends we just add a template runtime
1010
+ # error into the generated code. We could catch that at compile
1011
+ # time too, but i welcome it not to confuse users by throwing the
1012
+ # same error at different times just "because we can".
1013
+ if not self.has_known_extends:
1014
+ self.writeline("if parent_template is not None:")
1015
+ self.indent()
1016
+ self.writeline('raise TemplateRuntimeError("extended multiple times")')
1017
+
1018
+ # if we have a known extends already we don't need that code here
1019
+ # as we know that the template execution will end here.
1020
+ if self.has_known_extends:
1021
+ raise CompilerExit()
1022
+ else:
1023
+ self.outdent()
1024
+
1025
+ self.writeline("parent_template = environment.get_template(", node)
1026
+ self.visit(node.template, frame)
1027
+ self.write(f", {self.name!r})")
1028
+ self.writeline("for name, parent_block in parent_template.blocks.items():")
1029
+ self.indent()
1030
+ self.writeline("context.blocks.setdefault(name, []).append(parent_block)")
1031
+ self.outdent()
1032
+
1033
+ # if this extends statement was in the root level we can take
1034
+ # advantage of that information and simplify the generated code
1035
+ # in the top level from this point onwards
1036
+ if frame.rootlevel:
1037
+ self.has_known_extends = True
1038
+
1039
+ # and now we have one more
1040
+ self.extends_so_far += 1
1041
+
1042
+ def visit_Include(self, node: nodes.Include, frame: Frame) -> None:
1043
+ """Handles includes."""
1044
+ if node.ignore_missing:
1045
+ self.writeline("try:")
1046
+ self.indent()
1047
+
1048
+ func_name = "get_or_select_template"
1049
+ if isinstance(node.template, nodes.Const):
1050
+ if isinstance(node.template.value, str):
1051
+ func_name = "get_template"
1052
+ elif isinstance(node.template.value, (tuple, list)):
1053
+ func_name = "select_template"
1054
+ elif isinstance(node.template, (nodes.Tuple, nodes.List)):
1055
+ func_name = "select_template"
1056
+
1057
+ self.writeline(f"template = environment.{func_name}(", node)
1058
+ self.visit(node.template, frame)
1059
+ self.write(f", {self.name!r})")
1060
+ if node.ignore_missing:
1061
+ self.outdent()
1062
+ self.writeline("except TemplateNotFound:")
1063
+ self.indent()
1064
+ self.writeline("pass")
1065
+ self.outdent()
1066
+ self.writeline("else:")
1067
+ self.indent()
1068
+
1069
+ def loop_body() -> None:
1070
+ self.indent()
1071
+ self.simple_write("event", frame)
1072
+ self.outdent()
1073
+
1074
+ if node.with_context:
1075
+ self.writeline(
1076
+ f"gen = template.root_render_func("
1077
+ "template.new_context(context.get_all(), True,"
1078
+ f" {self.dump_local_context(frame)}))"
1079
+ )
1080
+ self.writeline("try:")
1081
+ self.indent()
1082
+ self.writeline(f"{self.choose_async()}for event in gen:")
1083
+ loop_body()
1084
+ self.outdent()
1085
+ self.writeline(
1086
+ f"finally: {self.choose_async('await gen.aclose()', 'gen.close()')}"
1087
+ )
1088
+ elif self.environment.is_async:
1089
+ self.writeline(
1090
+ "for event in (await template._get_default_module_async())"
1091
+ "._body_stream:"
1092
+ )
1093
+ loop_body()
1094
+ else:
1095
+ self.writeline("yield from template._get_default_module()._body_stream")
1096
+
1097
+ if node.ignore_missing:
1098
+ self.outdent()
1099
+
1100
+ def _import_common(
1101
+ self, node: t.Union[nodes.Import, nodes.FromImport], frame: Frame
1102
+ ) -> None:
1103
+ self.write(f"{self.choose_async('await ')}environment.get_template(")
1104
+ self.visit(node.template, frame)
1105
+ self.write(f", {self.name!r}).")
1106
+
1107
+ if node.with_context:
1108
+ f_name = f"make_module{self.choose_async('_async')}"
1109
+ self.write(
1110
+ f"{f_name}(context.get_all(), True, {self.dump_local_context(frame)})"
1111
+ )
1112
+ else:
1113
+ self.write(f"_get_default_module{self.choose_async('_async')}(context)")
1114
+
1115
+ def visit_Import(self, node: nodes.Import, frame: Frame) -> None:
1116
+ """Visit regular imports."""
1117
+ self.writeline(f"{frame.symbols.ref(node.target)} = ", node)
1118
+ if frame.toplevel:
1119
+ self.write(f"context.vars[{node.target!r}] = ")
1120
+
1121
+ self._import_common(node, frame)
1122
+
1123
+ if frame.toplevel and not node.target.startswith("_"):
1124
+ self.writeline(f"context.exported_vars.discard({node.target!r})")
1125
+
1126
+ def visit_FromImport(self, node: nodes.FromImport, frame: Frame) -> None:
1127
+ """Visit named imports."""
1128
+ self.newline(node)
1129
+ self.write("included_template = ")
1130
+ self._import_common(node, frame)
1131
+ var_names = []
1132
+ discarded_names = []
1133
+ for name in node.names:
1134
+ if isinstance(name, tuple):
1135
+ name, alias = name
1136
+ else:
1137
+ alias = name
1138
+ self.writeline(
1139
+ f"{frame.symbols.ref(alias)} ="
1140
+ f" getattr(included_template, {name!r}, missing)"
1141
+ )
1142
+ self.writeline(f"if {frame.symbols.ref(alias)} is missing:")
1143
+ self.indent()
1144
+ # The position will contain the template name, and will be formatted
1145
+ # into a string that will be compiled into an f-string. Curly braces
1146
+ # in the name must be replaced with escapes so that they will not be
1147
+ # executed as part of the f-string.
1148
+ position = self.position(node).replace("{", "{{").replace("}", "}}")
1149
+ message = (
1150
+ "the template {included_template.__name__!r}"
1151
+ f" (imported on {position})"
1152
+ f" does not export the requested name {name!r}"
1153
+ )
1154
+ self.writeline(
1155
+ f"{frame.symbols.ref(alias)} = undefined(f{message!r}, name={name!r})"
1156
+ )
1157
+ self.outdent()
1158
+ if frame.toplevel:
1159
+ var_names.append(alias)
1160
+ if not alias.startswith("_"):
1161
+ discarded_names.append(alias)
1162
+
1163
+ if var_names:
1164
+ if len(var_names) == 1:
1165
+ name = var_names[0]
1166
+ self.writeline(f"context.vars[{name!r}] = {frame.symbols.ref(name)}")
1167
+ else:
1168
+ names_kv = ", ".join(
1169
+ f"{name!r}: {frame.symbols.ref(name)}" for name in var_names
1170
+ )
1171
+ self.writeline(f"context.vars.update({{{names_kv}}})")
1172
+ if discarded_names:
1173
+ if len(discarded_names) == 1:
1174
+ self.writeline(f"context.exported_vars.discard({discarded_names[0]!r})")
1175
+ else:
1176
+ names_str = ", ".join(map(repr, discarded_names))
1177
+ self.writeline(
1178
+ f"context.exported_vars.difference_update(({names_str}))"
1179
+ )
1180
+
1181
+ def visit_For(self, node: nodes.For, frame: Frame) -> None:
1182
+ loop_frame = frame.inner()
1183
+ loop_frame.loop_frame = True
1184
+ test_frame = frame.inner()
1185
+ else_frame = frame.inner()
1186
+
1187
+ # try to figure out if we have an extended loop. An extended loop
1188
+ # is necessary if the loop is in recursive mode if the special loop
1189
+ # variable is accessed in the body if the body is a scoped block.
1190
+ extended_loop = (
1191
+ node.recursive
1192
+ or "loop"
1193
+ in find_undeclared(node.iter_child_nodes(only=("body",)), ("loop",))
1194
+ or any(block.scoped for block in node.find_all(nodes.Block))
1195
+ )
1196
+
1197
+ loop_ref = None
1198
+ if extended_loop:
1199
+ loop_ref = loop_frame.symbols.declare_parameter("loop")
1200
+
1201
+ loop_frame.symbols.analyze_node(node, for_branch="body")
1202
+ if node.else_:
1203
+ else_frame.symbols.analyze_node(node, for_branch="else")
1204
+
1205
+ if node.test:
1206
+ loop_filter_func = self.temporary_identifier()
1207
+ test_frame.symbols.analyze_node(node, for_branch="test")
1208
+ self.writeline(f"{self.func(loop_filter_func)}(fiter):", node.test)
1209
+ self.indent()
1210
+ self.enter_frame(test_frame)
1211
+ self.writeline(self.choose_async("async for ", "for "))
1212
+ self.visit(node.target, loop_frame)
1213
+ self.write(" in ")
1214
+ self.write(self.choose_async("auto_aiter(fiter)", "fiter"))
1215
+ self.write(":")
1216
+ self.indent()
1217
+ self.writeline("if ", node.test)
1218
+ self.visit(node.test, test_frame)
1219
+ self.write(":")
1220
+ self.indent()
1221
+ self.writeline("yield ")
1222
+ self.visit(node.target, loop_frame)
1223
+ self.outdent(3)
1224
+ self.leave_frame(test_frame, with_python_scope=True)
1225
+
1226
+ # if we don't have an recursive loop we have to find the shadowed
1227
+ # variables at that point. Because loops can be nested but the loop
1228
+ # variable is a special one we have to enforce aliasing for it.
1229
+ if node.recursive:
1230
+ self.writeline(
1231
+ f"{self.func('loop')}(reciter, loop_render_func, depth=0):", node
1232
+ )
1233
+ self.indent()
1234
+ self.buffer(loop_frame)
1235
+
1236
+ # Use the same buffer for the else frame
1237
+ else_frame.buffer = loop_frame.buffer
1238
+
1239
+ # make sure the loop variable is a special one and raise a template
1240
+ # assertion error if a loop tries to write to loop
1241
+ if extended_loop:
1242
+ self.writeline(f"{loop_ref} = missing")
1243
+
1244
+ for name in node.find_all(nodes.Name):
1245
+ if name.ctx == "store" and name.name == "loop":
1246
+ self.fail(
1247
+ "Can't assign to special loop variable in for-loop target",
1248
+ name.lineno,
1249
+ )
1250
+
1251
+ if node.else_:
1252
+ iteration_indicator = self.temporary_identifier()
1253
+ self.writeline(f"{iteration_indicator} = 1")
1254
+
1255
+ self.writeline(self.choose_async("async for ", "for "), node)
1256
+ self.visit(node.target, loop_frame)
1257
+ if extended_loop:
1258
+ self.write(f", {loop_ref} in {self.choose_async('Async')}LoopContext(")
1259
+ else:
1260
+ self.write(" in ")
1261
+
1262
+ if node.test:
1263
+ self.write(f"{loop_filter_func}(")
1264
+ if node.recursive:
1265
+ self.write("reciter")
1266
+ else:
1267
+ if self.environment.is_async and not extended_loop:
1268
+ self.write("auto_aiter(")
1269
+ self.visit(node.iter, frame)
1270
+ if self.environment.is_async and not extended_loop:
1271
+ self.write(")")
1272
+ if node.test:
1273
+ self.write(")")
1274
+
1275
+ if node.recursive:
1276
+ self.write(", undefined, loop_render_func, depth):")
1277
+ else:
1278
+ self.write(", undefined):" if extended_loop else ":")
1279
+
1280
+ self.indent()
1281
+ self.enter_frame(loop_frame)
1282
+
1283
+ self.writeline("_loop_vars = {}")
1284
+ self.blockvisit(node.body, loop_frame)
1285
+ if node.else_:
1286
+ self.writeline(f"{iteration_indicator} = 0")
1287
+ self.outdent()
1288
+ self.leave_frame(
1289
+ loop_frame, with_python_scope=node.recursive and not node.else_
1290
+ )
1291
+
1292
+ if node.else_:
1293
+ self.writeline(f"if {iteration_indicator}:")
1294
+ self.indent()
1295
+ self.enter_frame(else_frame)
1296
+ self.blockvisit(node.else_, else_frame)
1297
+ self.leave_frame(else_frame)
1298
+ self.outdent()
1299
+
1300
+ # if the node was recursive we have to return the buffer contents
1301
+ # and start the iteration code
1302
+ if node.recursive:
1303
+ self.return_buffer_contents(loop_frame)
1304
+ self.outdent()
1305
+ self.start_write(frame, node)
1306
+ self.write(f"{self.choose_async('await ')}loop(")
1307
+ if self.environment.is_async:
1308
+ self.write("auto_aiter(")
1309
+ self.visit(node.iter, frame)
1310
+ if self.environment.is_async:
1311
+ self.write(")")
1312
+ self.write(", loop)")
1313
+ self.end_write(frame)
1314
+
1315
+ # at the end of the iteration, clear any assignments made in the
1316
+ # loop from the top level
1317
+ if self._assign_stack:
1318
+ self._assign_stack[-1].difference_update(loop_frame.symbols.stores)
1319
+
1320
+ def visit_If(self, node: nodes.If, frame: Frame) -> None:
1321
+ if_frame = frame.soft()
1322
+ self.writeline("if ", node)
1323
+ self.visit(node.test, if_frame)
1324
+ self.write(":")
1325
+ self.indent()
1326
+ self.blockvisit(node.body, if_frame)
1327
+ self.outdent()
1328
+ for elif_ in node.elif_:
1329
+ self.writeline("elif ", elif_)
1330
+ self.visit(elif_.test, if_frame)
1331
+ self.write(":")
1332
+ self.indent()
1333
+ self.blockvisit(elif_.body, if_frame)
1334
+ self.outdent()
1335
+ if node.else_:
1336
+ self.writeline("else:")
1337
+ self.indent()
1338
+ self.blockvisit(node.else_, if_frame)
1339
+ self.outdent()
1340
+
1341
+ def visit_Macro(self, node: nodes.Macro, frame: Frame) -> None:
1342
+ macro_frame, macro_ref = self.macro_body(node, frame)
1343
+ self.newline()
1344
+ if frame.toplevel:
1345
+ if not node.name.startswith("_"):
1346
+ self.write(f"context.exported_vars.add({node.name!r})")
1347
+ self.writeline(f"context.vars[{node.name!r}] = ")
1348
+ self.write(f"{frame.symbols.ref(node.name)} = ")
1349
+ self.macro_def(macro_ref, macro_frame)
1350
+
1351
+ def visit_CallBlock(self, node: nodes.CallBlock, frame: Frame) -> None:
1352
+ call_frame, macro_ref = self.macro_body(node, frame)
1353
+ self.writeline("caller = ")
1354
+ self.macro_def(macro_ref, call_frame)
1355
+ self.start_write(frame, node)
1356
+ self.visit_Call(node.call, frame, forward_caller=True)
1357
+ self.end_write(frame)
1358
+
1359
+ def visit_FilterBlock(self, node: nodes.FilterBlock, frame: Frame) -> None:
1360
+ filter_frame = frame.inner()
1361
+ filter_frame.symbols.analyze_node(node)
1362
+ self.enter_frame(filter_frame)
1363
+ self.buffer(filter_frame)
1364
+ self.blockvisit(node.body, filter_frame)
1365
+ self.start_write(frame, node)
1366
+ self.visit_Filter(node.filter, filter_frame)
1367
+ self.end_write(frame)
1368
+ self.leave_frame(filter_frame)
1369
+
1370
+ def visit_With(self, node: nodes.With, frame: Frame) -> None:
1371
+ with_frame = frame.inner()
1372
+ with_frame.symbols.analyze_node(node)
1373
+ self.enter_frame(with_frame)
1374
+ for target, expr in zip(node.targets, node.values):
1375
+ self.newline()
1376
+ self.visit(target, with_frame)
1377
+ self.write(" = ")
1378
+ self.visit(expr, frame)
1379
+ self.blockvisit(node.body, with_frame)
1380
+ self.leave_frame(with_frame)
1381
+
1382
+ def visit_ExprStmt(self, node: nodes.ExprStmt, frame: Frame) -> None:
1383
+ self.newline(node)
1384
+ self.visit(node.node, frame)
1385
+
1386
+ class _FinalizeInfo(t.NamedTuple):
1387
+ const: t.Optional[t.Callable[..., str]]
1388
+ src: t.Optional[str]
1389
+
1390
+ @staticmethod
1391
+ def _default_finalize(value: t.Any) -> t.Any:
1392
+ """The default finalize function if the environment isn't
1393
+ configured with one. Or, if the environment has one, this is
1394
+ called on that function's output for constants.
1395
+ """
1396
+ return str(value)
1397
+
1398
+ _finalize: t.Optional[_FinalizeInfo] = None
1399
+
1400
+ def _make_finalize(self) -> _FinalizeInfo:
1401
+ """Build the finalize function to be used on constants and at
1402
+ runtime. Cached so it's only created once for all output nodes.
1403
+
1404
+ Returns a ``namedtuple`` with the following attributes:
1405
+
1406
+ ``const``
1407
+ A function to finalize constant data at compile time.
1408
+
1409
+ ``src``
1410
+ Source code to output around nodes to be evaluated at
1411
+ runtime.
1412
+ """
1413
+ if self._finalize is not None:
1414
+ return self._finalize
1415
+
1416
+ finalize: t.Optional[t.Callable[..., t.Any]]
1417
+ finalize = default = self._default_finalize
1418
+ src = None
1419
+
1420
+ if self.environment.finalize:
1421
+ src = "environment.finalize("
1422
+ env_finalize = self.environment.finalize
1423
+ pass_arg = {
1424
+ _PassArg.context: "context",
1425
+ _PassArg.eval_context: "context.eval_ctx",
1426
+ _PassArg.environment: "environment",
1427
+ }.get(
1428
+ _PassArg.from_obj(env_finalize) # type: ignore
1429
+ )
1430
+ finalize = None
1431
+
1432
+ if pass_arg is None:
1433
+
1434
+ def finalize(value: t.Any) -> t.Any: # noqa: F811
1435
+ return default(env_finalize(value))
1436
+
1437
+ else:
1438
+ src = f"{src}{pass_arg}, "
1439
+
1440
+ if pass_arg == "environment":
1441
+
1442
+ def finalize(value: t.Any) -> t.Any: # noqa: F811
1443
+ return default(env_finalize(self.environment, value))
1444
+
1445
+ self._finalize = self._FinalizeInfo(finalize, src)
1446
+ return self._finalize
1447
+
1448
+ def _output_const_repr(self, group: t.Iterable[t.Any]) -> str:
1449
+ """Given a group of constant values converted from ``Output``
1450
+ child nodes, produce a string to write to the template module
1451
+ source.
1452
+ """
1453
+ return repr(concat(group))
1454
+
1455
+ def _output_child_to_const(
1456
+ self, node: nodes.Expr, frame: Frame, finalize: _FinalizeInfo
1457
+ ) -> str:
1458
+ """Try to optimize a child of an ``Output`` node by trying to
1459
+ convert it to constant, finalized data at compile time.
1460
+
1461
+ If :exc:`Impossible` is raised, the node is not constant and
1462
+ will be evaluated at runtime. Any other exception will also be
1463
+ evaluated at runtime for easier debugging.
1464
+ """
1465
+ const = node.as_const(frame.eval_ctx)
1466
+
1467
+ if frame.eval_ctx.autoescape:
1468
+ const = escape(const)
1469
+
1470
+ # Template data doesn't go through finalize.
1471
+ if isinstance(node, nodes.TemplateData):
1472
+ return str(const)
1473
+
1474
+ return finalize.const(const) # type: ignore
1475
+
1476
+ def _output_child_pre(
1477
+ self, node: nodes.Expr, frame: Frame, finalize: _FinalizeInfo
1478
+ ) -> None:
1479
+ """Output extra source code before visiting a child of an
1480
+ ``Output`` node.
1481
+ """
1482
+ if frame.eval_ctx.volatile:
1483
+ self.write("(escape if context.eval_ctx.autoescape else str)(")
1484
+ elif frame.eval_ctx.autoescape:
1485
+ self.write("escape(")
1486
+ else:
1487
+ self.write("str(")
1488
+
1489
+ if finalize.src is not None:
1490
+ self.write(finalize.src)
1491
+
1492
+ def _output_child_post(
1493
+ self, node: nodes.Expr, frame: Frame, finalize: _FinalizeInfo
1494
+ ) -> None:
1495
+ """Output extra source code after visiting a child of an
1496
+ ``Output`` node.
1497
+ """
1498
+ self.write(")")
1499
+
1500
+ if finalize.src is not None:
1501
+ self.write(")")
1502
+
1503
+ def visit_Output(self, node: nodes.Output, frame: Frame) -> None:
1504
+ # If an extends is active, don't render outside a block.
1505
+ if frame.require_output_check:
1506
+ # A top-level extends is known to exist at compile time.
1507
+ if self.has_known_extends:
1508
+ return
1509
+
1510
+ self.writeline("if parent_template is None:")
1511
+ self.indent()
1512
+
1513
+ finalize = self._make_finalize()
1514
+ body: t.List[t.Union[t.List[t.Any], nodes.Expr]] = []
1515
+
1516
+ # Evaluate constants at compile time if possible. Each item in
1517
+ # body will be either a list of static data or a node to be
1518
+ # evaluated at runtime.
1519
+ for child in node.nodes:
1520
+ try:
1521
+ if not (
1522
+ # If the finalize function requires runtime context,
1523
+ # constants can't be evaluated at compile time.
1524
+ finalize.const
1525
+ # Unless it's basic template data that won't be
1526
+ # finalized anyway.
1527
+ or isinstance(child, nodes.TemplateData)
1528
+ ):
1529
+ raise nodes.Impossible()
1530
+
1531
+ const = self._output_child_to_const(child, frame, finalize)
1532
+ except (nodes.Impossible, Exception):
1533
+ # The node was not constant and needs to be evaluated at
1534
+ # runtime. Or another error was raised, which is easier
1535
+ # to debug at runtime.
1536
+ body.append(child)
1537
+ continue
1538
+
1539
+ if body and isinstance(body[-1], list):
1540
+ body[-1].append(const)
1541
+ else:
1542
+ body.append([const])
1543
+
1544
+ if frame.buffer is not None:
1545
+ if len(body) == 1:
1546
+ self.writeline(f"{frame.buffer}.append(")
1547
+ else:
1548
+ self.writeline(f"{frame.buffer}.extend((")
1549
+
1550
+ self.indent()
1551
+
1552
+ for item in body:
1553
+ if isinstance(item, list):
1554
+ # A group of constant data to join and output.
1555
+ val = self._output_const_repr(item)
1556
+
1557
+ if frame.buffer is None:
1558
+ self.writeline("yield " + val)
1559
+ else:
1560
+ self.writeline(val + ",")
1561
+ else:
1562
+ if frame.buffer is None:
1563
+ self.writeline("yield ", item)
1564
+ else:
1565
+ self.newline(item)
1566
+
1567
+ # A node to be evaluated at runtime.
1568
+ self._output_child_pre(item, frame, finalize)
1569
+ self.visit(item, frame)
1570
+ self._output_child_post(item, frame, finalize)
1571
+
1572
+ if frame.buffer is not None:
1573
+ self.write(",")
1574
+
1575
+ if frame.buffer is not None:
1576
+ self.outdent()
1577
+ self.writeline(")" if len(body) == 1 else "))")
1578
+
1579
+ if frame.require_output_check:
1580
+ self.outdent()
1581
+
1582
+ def visit_Assign(self, node: nodes.Assign, frame: Frame) -> None:
1583
+ self.push_assign_tracking()
1584
+
1585
+ # ``a.b`` is allowed for assignment, and is parsed as an NSRef. However,
1586
+ # it is only valid if it references a Namespace object. Emit a check for
1587
+ # that for each ref here, before assignment code is emitted. This can't
1588
+ # be done in visit_NSRef as the ref could be in the middle of a tuple.
1589
+ seen_refs: t.Set[str] = set()
1590
+
1591
+ for nsref in node.find_all(nodes.NSRef):
1592
+ if nsref.name in seen_refs:
1593
+ # Only emit the check for each reference once, in case the same
1594
+ # ref is used multiple times in a tuple, `ns.a, ns.b = c, d`.
1595
+ continue
1596
+
1597
+ seen_refs.add(nsref.name)
1598
+ ref = frame.symbols.ref(nsref.name)
1599
+ self.writeline(f"if not isinstance({ref}, Namespace):")
1600
+ self.indent()
1601
+ self.writeline(
1602
+ "raise TemplateRuntimeError"
1603
+ '("cannot assign attribute on non-namespace object")'
1604
+ )
1605
+ self.outdent()
1606
+
1607
+ self.newline(node)
1608
+ self.visit(node.target, frame)
1609
+ self.write(" = ")
1610
+ self.visit(node.node, frame)
1611
+ self.pop_assign_tracking(frame)
1612
+
1613
+ def visit_AssignBlock(self, node: nodes.AssignBlock, frame: Frame) -> None:
1614
+ self.push_assign_tracking()
1615
+ block_frame = frame.inner()
1616
+ # This is a special case. Since a set block always captures we
1617
+ # will disable output checks. This way one can use set blocks
1618
+ # toplevel even in extended templates.
1619
+ block_frame.require_output_check = False
1620
+ block_frame.symbols.analyze_node(node)
1621
+ self.enter_frame(block_frame)
1622
+ self.buffer(block_frame)
1623
+ self.blockvisit(node.body, block_frame)
1624
+ self.newline(node)
1625
+ self.visit(node.target, frame)
1626
+ self.write(" = (Markup if context.eval_ctx.autoescape else identity)(")
1627
+ if node.filter is not None:
1628
+ self.visit_Filter(node.filter, block_frame)
1629
+ else:
1630
+ self.write(f"concat({block_frame.buffer})")
1631
+ self.write(")")
1632
+ self.pop_assign_tracking(frame)
1633
+ self.leave_frame(block_frame)
1634
+
1635
+ # -- Expression Visitors
1636
+
1637
+ def visit_Name(self, node: nodes.Name, frame: Frame) -> None:
1638
+ if node.ctx == "store" and (
1639
+ frame.toplevel or frame.loop_frame or frame.block_frame
1640
+ ):
1641
+ if self._assign_stack:
1642
+ self._assign_stack[-1].add(node.name)
1643
+ ref = frame.symbols.ref(node.name)
1644
+
1645
+ # If we are looking up a variable we might have to deal with the
1646
+ # case where it's undefined. We can skip that case if the load
1647
+ # instruction indicates a parameter which are always defined.
1648
+ if node.ctx == "load":
1649
+ load = frame.symbols.find_load(ref)
1650
+ if not (
1651
+ load is not None
1652
+ and load[0] == VAR_LOAD_PARAMETER
1653
+ and not self.parameter_is_undeclared(ref)
1654
+ ):
1655
+ self.write(
1656
+ f"(undefined(name={node.name!r}) if {ref} is missing else {ref})"
1657
+ )
1658
+ return
1659
+
1660
+ self.write(ref)
1661
+
1662
+ def visit_NSRef(self, node: nodes.NSRef, frame: Frame) -> None:
1663
+ # NSRef is a dotted assignment target a.b=c, but uses a[b]=c internally.
1664
+ # visit_Assign emits code to validate that each ref is to a Namespace
1665
+ # object only. That can't be emitted here as the ref could be in the
1666
+ # middle of a tuple assignment.
1667
+ ref = frame.symbols.ref(node.name)
1668
+ self.writeline(f"{ref}[{node.attr!r}]")
1669
+
1670
+ def visit_Const(self, node: nodes.Const, frame: Frame) -> None:
1671
+ val = node.as_const(frame.eval_ctx)
1672
+ if isinstance(val, float):
1673
+ self.write(str(val))
1674
+ else:
1675
+ self.write(repr(val))
1676
+
1677
+ def visit_TemplateData(self, node: nodes.TemplateData, frame: Frame) -> None:
1678
+ try:
1679
+ self.write(repr(node.as_const(frame.eval_ctx)))
1680
+ except nodes.Impossible:
1681
+ self.write(
1682
+ f"(Markup if context.eval_ctx.autoescape else identity)({node.data!r})"
1683
+ )
1684
+
1685
+ def visit_Tuple(self, node: nodes.Tuple, frame: Frame) -> None:
1686
+ self.write("(")
1687
+ idx = -1
1688
+ for idx, item in enumerate(node.items):
1689
+ if idx:
1690
+ self.write(", ")
1691
+ self.visit(item, frame)
1692
+ self.write(",)" if idx == 0 else ")")
1693
+
1694
+ def visit_List(self, node: nodes.List, frame: Frame) -> None:
1695
+ self.write("[")
1696
+ for idx, item in enumerate(node.items):
1697
+ if idx:
1698
+ self.write(", ")
1699
+ self.visit(item, frame)
1700
+ self.write("]")
1701
+
1702
+ def visit_Dict(self, node: nodes.Dict, frame: Frame) -> None:
1703
+ self.write("{")
1704
+ for idx, item in enumerate(node.items):
1705
+ if idx:
1706
+ self.write(", ")
1707
+ self.visit(item.key, frame)
1708
+ self.write(": ")
1709
+ self.visit(item.value, frame)
1710
+ self.write("}")
1711
+
1712
+ visit_Add = _make_binop("+")
1713
+ visit_Sub = _make_binop("-")
1714
+ visit_Mul = _make_binop("*")
1715
+ visit_Div = _make_binop("/")
1716
+ visit_FloorDiv = _make_binop("//")
1717
+ visit_Pow = _make_binop("**")
1718
+ visit_Mod = _make_binop("%")
1719
+ visit_And = _make_binop("and")
1720
+ visit_Or = _make_binop("or")
1721
+ visit_Pos = _make_unop("+")
1722
+ visit_Neg = _make_unop("-")
1723
+ visit_Not = _make_unop("not ")
1724
+
1725
+ @optimizeconst
1726
+ def visit_Concat(self, node: nodes.Concat, frame: Frame) -> None:
1727
+ if frame.eval_ctx.volatile:
1728
+ func_name = "(markup_join if context.eval_ctx.volatile else str_join)"
1729
+ elif frame.eval_ctx.autoescape:
1730
+ func_name = "markup_join"
1731
+ else:
1732
+ func_name = "str_join"
1733
+ self.write(f"{func_name}((")
1734
+ for arg in node.nodes:
1735
+ self.visit(arg, frame)
1736
+ self.write(", ")
1737
+ self.write("))")
1738
+
1739
+ @optimizeconst
1740
+ def visit_Compare(self, node: nodes.Compare, frame: Frame) -> None:
1741
+ self.write("(")
1742
+ self.visit(node.expr, frame)
1743
+ for op in node.ops:
1744
+ self.visit(op, frame)
1745
+ self.write(")")
1746
+
1747
+ def visit_Operand(self, node: nodes.Operand, frame: Frame) -> None:
1748
+ self.write(f" {operators[node.op]} ")
1749
+ self.visit(node.expr, frame)
1750
+
1751
+ @optimizeconst
1752
+ def visit_Getattr(self, node: nodes.Getattr, frame: Frame) -> None:
1753
+ if self.environment.is_async:
1754
+ self.write("(await auto_await(")
1755
+
1756
+ self.write("environment.getattr(")
1757
+ self.visit(node.node, frame)
1758
+ self.write(f", {node.attr!r})")
1759
+
1760
+ if self.environment.is_async:
1761
+ self.write("))")
1762
+
1763
+ @optimizeconst
1764
+ def visit_Getitem(self, node: nodes.Getitem, frame: Frame) -> None:
1765
+ # slices bypass the environment getitem method.
1766
+ if isinstance(node.arg, nodes.Slice):
1767
+ self.visit(node.node, frame)
1768
+ self.write("[")
1769
+ self.visit(node.arg, frame)
1770
+ self.write("]")
1771
+ else:
1772
+ if self.environment.is_async:
1773
+ self.write("(await auto_await(")
1774
+
1775
+ self.write("environment.getitem(")
1776
+ self.visit(node.node, frame)
1777
+ self.write(", ")
1778
+ self.visit(node.arg, frame)
1779
+ self.write(")")
1780
+
1781
+ if self.environment.is_async:
1782
+ self.write("))")
1783
+
1784
+ def visit_Slice(self, node: nodes.Slice, frame: Frame) -> None:
1785
+ if node.start is not None:
1786
+ self.visit(node.start, frame)
1787
+ self.write(":")
1788
+ if node.stop is not None:
1789
+ self.visit(node.stop, frame)
1790
+ if node.step is not None:
1791
+ self.write(":")
1792
+ self.visit(node.step, frame)
1793
+
1794
+ @contextmanager
1795
+ def _filter_test_common(
1796
+ self, node: t.Union[nodes.Filter, nodes.Test], frame: Frame, is_filter: bool
1797
+ ) -> t.Iterator[None]:
1798
+ if self.environment.is_async:
1799
+ self.write("(await auto_await(")
1800
+
1801
+ if is_filter:
1802
+ self.write(f"{self.filters[node.name]}(")
1803
+ func = self.environment.filters.get(node.name)
1804
+ else:
1805
+ self.write(f"{self.tests[node.name]}(")
1806
+ func = self.environment.tests.get(node.name)
1807
+
1808
+ # When inside an If or CondExpr frame, allow the filter to be
1809
+ # undefined at compile time and only raise an error if it's
1810
+ # actually called at runtime. See pull_dependencies.
1811
+ if func is None and not frame.soft_frame:
1812
+ type_name = "filter" if is_filter else "test"
1813
+ self.fail(f"No {type_name} named {node.name!r}.", node.lineno)
1814
+
1815
+ pass_arg = {
1816
+ _PassArg.context: "context",
1817
+ _PassArg.eval_context: "context.eval_ctx",
1818
+ _PassArg.environment: "environment",
1819
+ }.get(
1820
+ _PassArg.from_obj(func) # type: ignore
1821
+ )
1822
+
1823
+ if pass_arg is not None:
1824
+ self.write(f"{pass_arg}, ")
1825
+
1826
+ # Back to the visitor function to handle visiting the target of
1827
+ # the filter or test.
1828
+ yield
1829
+
1830
+ self.signature(node, frame)
1831
+ self.write(")")
1832
+
1833
+ if self.environment.is_async:
1834
+ self.write("))")
1835
+
1836
+ @optimizeconst
1837
+ def visit_Filter(self, node: nodes.Filter, frame: Frame) -> None:
1838
+ with self._filter_test_common(node, frame, True):
1839
+ # if the filter node is None we are inside a filter block
1840
+ # and want to write to the current buffer
1841
+ if node.node is not None:
1842
+ self.visit(node.node, frame)
1843
+ elif frame.eval_ctx.volatile:
1844
+ self.write(
1845
+ f"(Markup(concat({frame.buffer}))"
1846
+ f" if context.eval_ctx.autoescape else concat({frame.buffer}))"
1847
+ )
1848
+ elif frame.eval_ctx.autoescape:
1849
+ self.write(f"Markup(concat({frame.buffer}))")
1850
+ else:
1851
+ self.write(f"concat({frame.buffer})")
1852
+
1853
+ @optimizeconst
1854
+ def visit_Test(self, node: nodes.Test, frame: Frame) -> None:
1855
+ with self._filter_test_common(node, frame, False):
1856
+ self.visit(node.node, frame)
1857
+
1858
+ @optimizeconst
1859
+ def visit_CondExpr(self, node: nodes.CondExpr, frame: Frame) -> None:
1860
+ frame = frame.soft()
1861
+
1862
+ def write_expr2() -> None:
1863
+ if node.expr2 is not None:
1864
+ self.visit(node.expr2, frame)
1865
+ return
1866
+
1867
+ self.write(
1868
+ f'cond_expr_undefined("the inline if-expression on'
1869
+ f" {self.position(node)} evaluated to false and no else"
1870
+ f' section was defined.")'
1871
+ )
1872
+
1873
+ self.write("(")
1874
+ self.visit(node.expr1, frame)
1875
+ self.write(" if ")
1876
+ self.visit(node.test, frame)
1877
+ self.write(" else ")
1878
+ write_expr2()
1879
+ self.write(")")
1880
+
1881
+ @optimizeconst
1882
+ def visit_Call(
1883
+ self, node: nodes.Call, frame: Frame, forward_caller: bool = False
1884
+ ) -> None:
1885
+ if self.environment.is_async:
1886
+ self.write("(await auto_await(")
1887
+ if self.environment.sandboxed:
1888
+ self.write("environment.call(context, ")
1889
+ else:
1890
+ self.write("context.call(")
1891
+ self.visit(node.node, frame)
1892
+ extra_kwargs = {"caller": "caller"} if forward_caller else None
1893
+ loop_kwargs = {"_loop_vars": "_loop_vars"} if frame.loop_frame else {}
1894
+ block_kwargs = {"_block_vars": "_block_vars"} if frame.block_frame else {}
1895
+ if extra_kwargs:
1896
+ extra_kwargs.update(loop_kwargs, **block_kwargs)
1897
+ elif loop_kwargs or block_kwargs:
1898
+ extra_kwargs = dict(loop_kwargs, **block_kwargs)
1899
+ self.signature(node, frame, extra_kwargs)
1900
+ self.write(")")
1901
+ if self.environment.is_async:
1902
+ self.write("))")
1903
+
1904
+ def visit_Keyword(self, node: nodes.Keyword, frame: Frame) -> None:
1905
+ self.write(node.key + "=")
1906
+ self.visit(node.value, frame)
1907
+
1908
+ # -- Unused nodes for extensions
1909
+
1910
+ def visit_MarkSafe(self, node: nodes.MarkSafe, frame: Frame) -> None:
1911
+ self.write("Markup(")
1912
+ self.visit(node.expr, frame)
1913
+ self.write(")")
1914
+
1915
+ def visit_MarkSafeIfAutoescape(
1916
+ self, node: nodes.MarkSafeIfAutoescape, frame: Frame
1917
+ ) -> None:
1918
+ self.write("(Markup if context.eval_ctx.autoescape else identity)(")
1919
+ self.visit(node.expr, frame)
1920
+ self.write(")")
1921
+
1922
+ def visit_EnvironmentAttribute(
1923
+ self, node: nodes.EnvironmentAttribute, frame: Frame
1924
+ ) -> None:
1925
+ self.write("environment." + node.name)
1926
+
1927
+ def visit_ExtensionAttribute(
1928
+ self, node: nodes.ExtensionAttribute, frame: Frame
1929
+ ) -> None:
1930
+ self.write(f"environment.extensions[{node.identifier!r}].{node.name}")
1931
+
1932
+ def visit_ImportedName(self, node: nodes.ImportedName, frame: Frame) -> None:
1933
+ self.write(self.import_aliases[node.importname])
1934
+
1935
+ def visit_InternalName(self, node: nodes.InternalName, frame: Frame) -> None:
1936
+ self.write(node.name)
1937
+
1938
+ def visit_ContextReference(
1939
+ self, node: nodes.ContextReference, frame: Frame
1940
+ ) -> None:
1941
+ self.write("context")
1942
+
1943
+ def visit_DerivedContextReference(
1944
+ self, node: nodes.DerivedContextReference, frame: Frame
1945
+ ) -> None:
1946
+ self.write(self.derive_context(frame))
1947
+
1948
+ def visit_Continue(self, node: nodes.Continue, frame: Frame) -> None:
1949
+ self.writeline("continue", node)
1950
+
1951
+ def visit_Break(self, node: nodes.Break, frame: Frame) -> None:
1952
+ self.writeline("break", node)
1953
+
1954
+ def visit_Scope(self, node: nodes.Scope, frame: Frame) -> None:
1955
+ scope_frame = frame.inner()
1956
+ scope_frame.symbols.analyze_node(node)
1957
+ self.enter_frame(scope_frame)
1958
+ self.blockvisit(node.body, scope_frame)
1959
+ self.leave_frame(scope_frame)
1960
+
1961
+ def visit_OverlayScope(self, node: nodes.OverlayScope, frame: Frame) -> None:
1962
+ ctx = self.temporary_identifier()
1963
+ self.writeline(f"{ctx} = {self.derive_context(frame)}")
1964
+ self.writeline(f"{ctx}.vars = ")
1965
+ self.visit(node.context, frame)
1966
+ self.push_context_reference(ctx)
1967
+
1968
+ scope_frame = frame.inner(isolated=True)
1969
+ scope_frame.symbols.analyze_node(node)
1970
+ self.enter_frame(scope_frame)
1971
+ self.blockvisit(node.body, scope_frame)
1972
+ self.leave_frame(scope_frame)
1973
+ self.pop_context_reference()
1974
+
1975
+ def visit_EvalContextModifier(
1976
+ self, node: nodes.EvalContextModifier, frame: Frame
1977
+ ) -> None:
1978
+ for keyword in node.options:
1979
+ self.writeline(f"context.eval_ctx.{keyword.key} = ")
1980
+ self.visit(keyword.value, frame)
1981
+ try:
1982
+ val = keyword.value.as_const(frame.eval_ctx)
1983
+ except nodes.Impossible:
1984
+ frame.eval_ctx.volatile = True
1985
+ else:
1986
+ setattr(frame.eval_ctx, keyword.key, val)
1987
+
1988
+ def visit_ScopedEvalContextModifier(
1989
+ self, node: nodes.ScopedEvalContextModifier, frame: Frame
1990
+ ) -> None:
1991
+ old_ctx_name = self.temporary_identifier()
1992
+ saved_ctx = frame.eval_ctx.save()
1993
+ self.writeline(f"{old_ctx_name} = context.eval_ctx.save()")
1994
+ self.visit_EvalContextModifier(node, frame)
1995
+ for child in node.body:
1996
+ self.visit(child, frame)
1997
+ frame.eval_ctx.revert(saved_ctx)
1998
+ self.writeline(f"context.eval_ctx.revert({old_ctx_name})")
janus/lib/python3.10/site-packages/jinja2/constants.py ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #: list of lorem ipsum words used by the lipsum() helper function
2
+ LOREM_IPSUM_WORDS = """\
3
+ a ac accumsan ad adipiscing aenean aliquam aliquet amet ante aptent arcu at
4
+ auctor augue bibendum blandit class commodo condimentum congue consectetuer
5
+ consequat conubia convallis cras cubilia cum curabitur curae cursus dapibus
6
+ diam dictum dictumst dignissim dis dolor donec dui duis egestas eget eleifend
7
+ elementum elit enim erat eros est et etiam eu euismod facilisi facilisis fames
8
+ faucibus felis fermentum feugiat fringilla fusce gravida habitant habitasse hac
9
+ hendrerit hymenaeos iaculis id imperdiet in inceptos integer interdum ipsum
10
+ justo lacinia lacus laoreet lectus leo libero ligula litora lobortis lorem
11
+ luctus maecenas magna magnis malesuada massa mattis mauris metus mi molestie
12
+ mollis montes morbi mus nam nascetur natoque nec neque netus nibh nisi nisl non
13
+ nonummy nostra nulla nullam nunc odio orci ornare parturient pede pellentesque
14
+ penatibus per pharetra phasellus placerat platea porta porttitor posuere
15
+ potenti praesent pretium primis proin pulvinar purus quam quis quisque rhoncus
16
+ ridiculus risus rutrum sagittis sapien scelerisque sed sem semper senectus sit
17
+ sociis sociosqu sodales sollicitudin suscipit suspendisse taciti tellus tempor
18
+ tempus tincidunt torquent tortor tristique turpis ullamcorper ultrices
19
+ ultricies urna ut varius vehicula vel velit venenatis vestibulum vitae vivamus
20
+ viverra volutpat vulputate"""
janus/lib/python3.10/site-packages/jinja2/debug.py ADDED
@@ -0,0 +1,191 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import sys
2
+ import typing as t
3
+ from types import CodeType
4
+ from types import TracebackType
5
+
6
+ from .exceptions import TemplateSyntaxError
7
+ from .utils import internal_code
8
+ from .utils import missing
9
+
10
+ if t.TYPE_CHECKING:
11
+ from .runtime import Context
12
+
13
+
14
+ def rewrite_traceback_stack(source: t.Optional[str] = None) -> BaseException:
15
+ """Rewrite the current exception to replace any tracebacks from
16
+ within compiled template code with tracebacks that look like they
17
+ came from the template source.
18
+
19
+ This must be called within an ``except`` block.
20
+
21
+ :param source: For ``TemplateSyntaxError``, the original source if
22
+ known.
23
+ :return: The original exception with the rewritten traceback.
24
+ """
25
+ _, exc_value, tb = sys.exc_info()
26
+ exc_value = t.cast(BaseException, exc_value)
27
+ tb = t.cast(TracebackType, tb)
28
+
29
+ if isinstance(exc_value, TemplateSyntaxError) and not exc_value.translated:
30
+ exc_value.translated = True
31
+ exc_value.source = source
32
+ # Remove the old traceback, otherwise the frames from the
33
+ # compiler still show up.
34
+ exc_value.with_traceback(None)
35
+ # Outside of runtime, so the frame isn't executing template
36
+ # code, but it still needs to point at the template.
37
+ tb = fake_traceback(
38
+ exc_value, None, exc_value.filename or "<unknown>", exc_value.lineno
39
+ )
40
+ else:
41
+ # Skip the frame for the render function.
42
+ tb = tb.tb_next
43
+
44
+ stack = []
45
+
46
+ # Build the stack of traceback object, replacing any in template
47
+ # code with the source file and line information.
48
+ while tb is not None:
49
+ # Skip frames decorated with @internalcode. These are internal
50
+ # calls that aren't useful in template debugging output.
51
+ if tb.tb_frame.f_code in internal_code:
52
+ tb = tb.tb_next
53
+ continue
54
+
55
+ template = tb.tb_frame.f_globals.get("__jinja_template__")
56
+
57
+ if template is not None:
58
+ lineno = template.get_corresponding_lineno(tb.tb_lineno)
59
+ fake_tb = fake_traceback(exc_value, tb, template.filename, lineno)
60
+ stack.append(fake_tb)
61
+ else:
62
+ stack.append(tb)
63
+
64
+ tb = tb.tb_next
65
+
66
+ tb_next = None
67
+
68
+ # Assign tb_next in reverse to avoid circular references.
69
+ for tb in reversed(stack):
70
+ tb.tb_next = tb_next
71
+ tb_next = tb
72
+
73
+ return exc_value.with_traceback(tb_next)
74
+
75
+
76
+ def fake_traceback( # type: ignore
77
+ exc_value: BaseException, tb: t.Optional[TracebackType], filename: str, lineno: int
78
+ ) -> TracebackType:
79
+ """Produce a new traceback object that looks like it came from the
80
+ template source instead of the compiled code. The filename, line
81
+ number, and location name will point to the template, and the local
82
+ variables will be the current template context.
83
+
84
+ :param exc_value: The original exception to be re-raised to create
85
+ the new traceback.
86
+ :param tb: The original traceback to get the local variables and
87
+ code info from.
88
+ :param filename: The template filename.
89
+ :param lineno: The line number in the template source.
90
+ """
91
+ if tb is not None:
92
+ # Replace the real locals with the context that would be
93
+ # available at that point in the template.
94
+ locals = get_template_locals(tb.tb_frame.f_locals)
95
+ locals.pop("__jinja_exception__", None)
96
+ else:
97
+ locals = {}
98
+
99
+ globals = {
100
+ "__name__": filename,
101
+ "__file__": filename,
102
+ "__jinja_exception__": exc_value,
103
+ }
104
+ # Raise an exception at the correct line number.
105
+ code: CodeType = compile(
106
+ "\n" * (lineno - 1) + "raise __jinja_exception__", filename, "exec"
107
+ )
108
+
109
+ # Build a new code object that points to the template file and
110
+ # replaces the location with a block name.
111
+ location = "template"
112
+
113
+ if tb is not None:
114
+ function = tb.tb_frame.f_code.co_name
115
+
116
+ if function == "root":
117
+ location = "top-level template code"
118
+ elif function.startswith("block_"):
119
+ location = f"block {function[6:]!r}"
120
+
121
+ if sys.version_info >= (3, 8):
122
+ code = code.replace(co_name=location)
123
+ else:
124
+ code = CodeType(
125
+ code.co_argcount,
126
+ code.co_kwonlyargcount,
127
+ code.co_nlocals,
128
+ code.co_stacksize,
129
+ code.co_flags,
130
+ code.co_code,
131
+ code.co_consts,
132
+ code.co_names,
133
+ code.co_varnames,
134
+ code.co_filename,
135
+ location,
136
+ code.co_firstlineno,
137
+ code.co_lnotab,
138
+ code.co_freevars,
139
+ code.co_cellvars,
140
+ )
141
+
142
+ # Execute the new code, which is guaranteed to raise, and return
143
+ # the new traceback without this frame.
144
+ try:
145
+ exec(code, globals, locals)
146
+ except BaseException:
147
+ return sys.exc_info()[2].tb_next # type: ignore
148
+
149
+
150
+ def get_template_locals(real_locals: t.Mapping[str, t.Any]) -> t.Dict[str, t.Any]:
151
+ """Based on the runtime locals, get the context that would be
152
+ available at that point in the template.
153
+ """
154
+ # Start with the current template context.
155
+ ctx: t.Optional[Context] = real_locals.get("context")
156
+
157
+ if ctx is not None:
158
+ data: t.Dict[str, t.Any] = ctx.get_all().copy()
159
+ else:
160
+ data = {}
161
+
162
+ # Might be in a derived context that only sets local variables
163
+ # rather than pushing a context. Local variables follow the scheme
164
+ # l_depth_name. Find the highest-depth local that has a value for
165
+ # each name.
166
+ local_overrides: t.Dict[str, t.Tuple[int, t.Any]] = {}
167
+
168
+ for name, value in real_locals.items():
169
+ if not name.startswith("l_") or value is missing:
170
+ # Not a template variable, or no longer relevant.
171
+ continue
172
+
173
+ try:
174
+ _, depth_str, name = name.split("_", 2)
175
+ depth = int(depth_str)
176
+ except ValueError:
177
+ continue
178
+
179
+ cur_depth = local_overrides.get(name, (-1,))[0]
180
+
181
+ if cur_depth < depth:
182
+ local_overrides[name] = (depth, value)
183
+
184
+ # Modify the context with any derived context.
185
+ for name, (_, value) in local_overrides.items():
186
+ if value is missing:
187
+ data.pop(name, None)
188
+ else:
189
+ data[name] = value
190
+
191
+ return data
janus/lib/python3.10/site-packages/jinja2/defaults.py ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import typing as t
2
+
3
+ from .filters import FILTERS as DEFAULT_FILTERS # noqa: F401
4
+ from .tests import TESTS as DEFAULT_TESTS # noqa: F401
5
+ from .utils import Cycler
6
+ from .utils import generate_lorem_ipsum
7
+ from .utils import Joiner
8
+ from .utils import Namespace
9
+
10
+ if t.TYPE_CHECKING:
11
+ import typing_extensions as te
12
+
13
+ # defaults for the parser / lexer
14
+ BLOCK_START_STRING = "{%"
15
+ BLOCK_END_STRING = "%}"
16
+ VARIABLE_START_STRING = "{{"
17
+ VARIABLE_END_STRING = "}}"
18
+ COMMENT_START_STRING = "{#"
19
+ COMMENT_END_STRING = "#}"
20
+ LINE_STATEMENT_PREFIX: t.Optional[str] = None
21
+ LINE_COMMENT_PREFIX: t.Optional[str] = None
22
+ TRIM_BLOCKS = False
23
+ LSTRIP_BLOCKS = False
24
+ NEWLINE_SEQUENCE: "te.Literal['\\n', '\\r\\n', '\\r']" = "\n"
25
+ KEEP_TRAILING_NEWLINE = False
26
+
27
+ # default filters, tests and namespace
28
+
29
+ DEFAULT_NAMESPACE = {
30
+ "range": range,
31
+ "dict": dict,
32
+ "lipsum": generate_lorem_ipsum,
33
+ "cycler": Cycler,
34
+ "joiner": Joiner,
35
+ "namespace": Namespace,
36
+ }
37
+
38
+ # default policies
39
+ DEFAULT_POLICIES: t.Dict[str, t.Any] = {
40
+ "compiler.ascii_str": True,
41
+ "urlize.rel": "noopener",
42
+ "urlize.target": None,
43
+ "urlize.extra_schemes": None,
44
+ "truncate.leeway": 5,
45
+ "json.dumps_function": None,
46
+ "json.dumps_kwargs": {"sort_keys": True},
47
+ "ext.i18n.trimmed": False,
48
+ }
janus/lib/python3.10/site-packages/jinja2/environment.py ADDED
@@ -0,0 +1,1672 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """Classes for managing templates and their runtime and compile time
2
+ options.
3
+ """
4
+
5
+ import os
6
+ import typing
7
+ import typing as t
8
+ import weakref
9
+ from collections import ChainMap
10
+ from functools import lru_cache
11
+ from functools import partial
12
+ from functools import reduce
13
+ from types import CodeType
14
+
15
+ from markupsafe import Markup
16
+
17
+ from . import nodes
18
+ from .compiler import CodeGenerator
19
+ from .compiler import generate
20
+ from .defaults import BLOCK_END_STRING
21
+ from .defaults import BLOCK_START_STRING
22
+ from .defaults import COMMENT_END_STRING
23
+ from .defaults import COMMENT_START_STRING
24
+ from .defaults import DEFAULT_FILTERS # type: ignore[attr-defined]
25
+ from .defaults import DEFAULT_NAMESPACE
26
+ from .defaults import DEFAULT_POLICIES
27
+ from .defaults import DEFAULT_TESTS # type: ignore[attr-defined]
28
+ from .defaults import KEEP_TRAILING_NEWLINE
29
+ from .defaults import LINE_COMMENT_PREFIX
30
+ from .defaults import LINE_STATEMENT_PREFIX
31
+ from .defaults import LSTRIP_BLOCKS
32
+ from .defaults import NEWLINE_SEQUENCE
33
+ from .defaults import TRIM_BLOCKS
34
+ from .defaults import VARIABLE_END_STRING
35
+ from .defaults import VARIABLE_START_STRING
36
+ from .exceptions import TemplateNotFound
37
+ from .exceptions import TemplateRuntimeError
38
+ from .exceptions import TemplatesNotFound
39
+ from .exceptions import TemplateSyntaxError
40
+ from .exceptions import UndefinedError
41
+ from .lexer import get_lexer
42
+ from .lexer import Lexer
43
+ from .lexer import TokenStream
44
+ from .nodes import EvalContext
45
+ from .parser import Parser
46
+ from .runtime import Context
47
+ from .runtime import new_context
48
+ from .runtime import Undefined
49
+ from .utils import _PassArg
50
+ from .utils import concat
51
+ from .utils import consume
52
+ from .utils import import_string
53
+ from .utils import internalcode
54
+ from .utils import LRUCache
55
+ from .utils import missing
56
+
57
+ if t.TYPE_CHECKING:
58
+ import typing_extensions as te
59
+
60
+ from .bccache import BytecodeCache
61
+ from .ext import Extension
62
+ from .loaders import BaseLoader
63
+
64
+ _env_bound = t.TypeVar("_env_bound", bound="Environment")
65
+
66
+
67
+ # for direct template usage we have up to ten living environments
68
+ @lru_cache(maxsize=10)
69
+ def get_spontaneous_environment(cls: t.Type[_env_bound], *args: t.Any) -> _env_bound:
70
+ """Return a new spontaneous environment. A spontaneous environment
71
+ is used for templates created directly rather than through an
72
+ existing environment.
73
+
74
+ :param cls: Environment class to create.
75
+ :param args: Positional arguments passed to environment.
76
+ """
77
+ env = cls(*args)
78
+ env.shared = True
79
+ return env
80
+
81
+
82
+ def create_cache(
83
+ size: int,
84
+ ) -> t.Optional[t.MutableMapping[t.Tuple["weakref.ref[t.Any]", str], "Template"]]:
85
+ """Return the cache class for the given size."""
86
+ if size == 0:
87
+ return None
88
+
89
+ if size < 0:
90
+ return {}
91
+
92
+ return LRUCache(size) # type: ignore
93
+
94
+
95
+ def copy_cache(
96
+ cache: t.Optional[t.MutableMapping[t.Any, t.Any]],
97
+ ) -> t.Optional[t.MutableMapping[t.Tuple["weakref.ref[t.Any]", str], "Template"]]:
98
+ """Create an empty copy of the given cache."""
99
+ if cache is None:
100
+ return None
101
+
102
+ if type(cache) is dict: # noqa E721
103
+ return {}
104
+
105
+ return LRUCache(cache.capacity) # type: ignore
106
+
107
+
108
+ def load_extensions(
109
+ environment: "Environment",
110
+ extensions: t.Sequence[t.Union[str, t.Type["Extension"]]],
111
+ ) -> t.Dict[str, "Extension"]:
112
+ """Load the extensions from the list and bind it to the environment.
113
+ Returns a dict of instantiated extensions.
114
+ """
115
+ result = {}
116
+
117
+ for extension in extensions:
118
+ if isinstance(extension, str):
119
+ extension = t.cast(t.Type["Extension"], import_string(extension))
120
+
121
+ result[extension.identifier] = extension(environment)
122
+
123
+ return result
124
+
125
+
126
+ def _environment_config_check(environment: _env_bound) -> _env_bound:
127
+ """Perform a sanity check on the environment."""
128
+ assert issubclass(
129
+ environment.undefined, Undefined
130
+ ), "'undefined' must be a subclass of 'jinja2.Undefined'."
131
+ assert (
132
+ environment.block_start_string
133
+ != environment.variable_start_string
134
+ != environment.comment_start_string
135
+ ), "block, variable and comment start strings must be different."
136
+ assert environment.newline_sequence in {
137
+ "\r",
138
+ "\r\n",
139
+ "\n",
140
+ }, "'newline_sequence' must be one of '\\n', '\\r\\n', or '\\r'."
141
+ return environment
142
+
143
+
144
+ class Environment:
145
+ r"""The core component of Jinja is the `Environment`. It contains
146
+ important shared variables like configuration, filters, tests,
147
+ globals and others. Instances of this class may be modified if
148
+ they are not shared and if no template was loaded so far.
149
+ Modifications on environments after the first template was loaded
150
+ will lead to surprising effects and undefined behavior.
151
+
152
+ Here are the possible initialization parameters:
153
+
154
+ `block_start_string`
155
+ The string marking the beginning of a block. Defaults to ``'{%'``.
156
+
157
+ `block_end_string`
158
+ The string marking the end of a block. Defaults to ``'%}'``.
159
+
160
+ `variable_start_string`
161
+ The string marking the beginning of a print statement.
162
+ Defaults to ``'{{'``.
163
+
164
+ `variable_end_string`
165
+ The string marking the end of a print statement. Defaults to
166
+ ``'}}'``.
167
+
168
+ `comment_start_string`
169
+ The string marking the beginning of a comment. Defaults to ``'{#'``.
170
+
171
+ `comment_end_string`
172
+ The string marking the end of a comment. Defaults to ``'#}'``.
173
+
174
+ `line_statement_prefix`
175
+ If given and a string, this will be used as prefix for line based
176
+ statements. See also :ref:`line-statements`.
177
+
178
+ `line_comment_prefix`
179
+ If given and a string, this will be used as prefix for line based
180
+ comments. See also :ref:`line-statements`.
181
+
182
+ .. versionadded:: 2.2
183
+
184
+ `trim_blocks`
185
+ If this is set to ``True`` the first newline after a block is
186
+ removed (block, not variable tag!). Defaults to `False`.
187
+
188
+ `lstrip_blocks`
189
+ If this is set to ``True`` leading spaces and tabs are stripped
190
+ from the start of a line to a block. Defaults to `False`.
191
+
192
+ `newline_sequence`
193
+ The sequence that starts a newline. Must be one of ``'\r'``,
194
+ ``'\n'`` or ``'\r\n'``. The default is ``'\n'`` which is a
195
+ useful default for Linux and OS X systems as well as web
196
+ applications.
197
+
198
+ `keep_trailing_newline`
199
+ Preserve the trailing newline when rendering templates.
200
+ The default is ``False``, which causes a single newline,
201
+ if present, to be stripped from the end of the template.
202
+
203
+ .. versionadded:: 2.7
204
+
205
+ `extensions`
206
+ List of Jinja extensions to use. This can either be import paths
207
+ as strings or extension classes. For more information have a
208
+ look at :ref:`the extensions documentation <jinja-extensions>`.
209
+
210
+ `optimized`
211
+ should the optimizer be enabled? Default is ``True``.
212
+
213
+ `undefined`
214
+ :class:`Undefined` or a subclass of it that is used to represent
215
+ undefined values in the template.
216
+
217
+ `finalize`
218
+ A callable that can be used to process the result of a variable
219
+ expression before it is output. For example one can convert
220
+ ``None`` implicitly into an empty string here.
221
+
222
+ `autoescape`
223
+ If set to ``True`` the XML/HTML autoescaping feature is enabled by
224
+ default. For more details about autoescaping see
225
+ :class:`~markupsafe.Markup`. As of Jinja 2.4 this can also
226
+ be a callable that is passed the template name and has to
227
+ return ``True`` or ``False`` depending on autoescape should be
228
+ enabled by default.
229
+
230
+ .. versionchanged:: 2.4
231
+ `autoescape` can now be a function
232
+
233
+ `loader`
234
+ The template loader for this environment.
235
+
236
+ `cache_size`
237
+ The size of the cache. Per default this is ``400`` which means
238
+ that if more than 400 templates are loaded the loader will clean
239
+ out the least recently used template. If the cache size is set to
240
+ ``0`` templates are recompiled all the time, if the cache size is
241
+ ``-1`` the cache will not be cleaned.
242
+
243
+ .. versionchanged:: 2.8
244
+ The cache size was increased to 400 from a low 50.
245
+
246
+ `auto_reload`
247
+ Some loaders load templates from locations where the template
248
+ sources may change (ie: file system or database). If
249
+ ``auto_reload`` is set to ``True`` (default) every time a template is
250
+ requested the loader checks if the source changed and if yes, it
251
+ will reload the template. For higher performance it's possible to
252
+ disable that.
253
+
254
+ `bytecode_cache`
255
+ If set to a bytecode cache object, this object will provide a
256
+ cache for the internal Jinja bytecode so that templates don't
257
+ have to be parsed if they were not changed.
258
+
259
+ See :ref:`bytecode-cache` for more information.
260
+
261
+ `enable_async`
262
+ If set to true this enables async template execution which
263
+ allows using async functions and generators.
264
+ """
265
+
266
+ #: if this environment is sandboxed. Modifying this variable won't make
267
+ #: the environment sandboxed though. For a real sandboxed environment
268
+ #: have a look at jinja2.sandbox. This flag alone controls the code
269
+ #: generation by the compiler.
270
+ sandboxed = False
271
+
272
+ #: True if the environment is just an overlay
273
+ overlayed = False
274
+
275
+ #: the environment this environment is linked to if it is an overlay
276
+ linked_to: t.Optional["Environment"] = None
277
+
278
+ #: shared environments have this set to `True`. A shared environment
279
+ #: must not be modified
280
+ shared = False
281
+
282
+ #: the class that is used for code generation. See
283
+ #: :class:`~jinja2.compiler.CodeGenerator` for more information.
284
+ code_generator_class: t.Type["CodeGenerator"] = CodeGenerator
285
+
286
+ concat = "".join
287
+
288
+ #: the context class that is used for templates. See
289
+ #: :class:`~jinja2.runtime.Context` for more information.
290
+ context_class: t.Type[Context] = Context
291
+
292
+ template_class: t.Type["Template"]
293
+
294
+ def __init__(
295
+ self,
296
+ block_start_string: str = BLOCK_START_STRING,
297
+ block_end_string: str = BLOCK_END_STRING,
298
+ variable_start_string: str = VARIABLE_START_STRING,
299
+ variable_end_string: str = VARIABLE_END_STRING,
300
+ comment_start_string: str = COMMENT_START_STRING,
301
+ comment_end_string: str = COMMENT_END_STRING,
302
+ line_statement_prefix: t.Optional[str] = LINE_STATEMENT_PREFIX,
303
+ line_comment_prefix: t.Optional[str] = LINE_COMMENT_PREFIX,
304
+ trim_blocks: bool = TRIM_BLOCKS,
305
+ lstrip_blocks: bool = LSTRIP_BLOCKS,
306
+ newline_sequence: "te.Literal['\\n', '\\r\\n', '\\r']" = NEWLINE_SEQUENCE,
307
+ keep_trailing_newline: bool = KEEP_TRAILING_NEWLINE,
308
+ extensions: t.Sequence[t.Union[str, t.Type["Extension"]]] = (),
309
+ optimized: bool = True,
310
+ undefined: t.Type[Undefined] = Undefined,
311
+ finalize: t.Optional[t.Callable[..., t.Any]] = None,
312
+ autoescape: t.Union[bool, t.Callable[[t.Optional[str]], bool]] = False,
313
+ loader: t.Optional["BaseLoader"] = None,
314
+ cache_size: int = 400,
315
+ auto_reload: bool = True,
316
+ bytecode_cache: t.Optional["BytecodeCache"] = None,
317
+ enable_async: bool = False,
318
+ ):
319
+ # !!Important notice!!
320
+ # The constructor accepts quite a few arguments that should be
321
+ # passed by keyword rather than position. However it's important to
322
+ # not change the order of arguments because it's used at least
323
+ # internally in those cases:
324
+ # - spontaneous environments (i18n extension and Template)
325
+ # - unittests
326
+ # If parameter changes are required only add parameters at the end
327
+ # and don't change the arguments (or the defaults!) of the arguments
328
+ # existing already.
329
+
330
+ # lexer / parser information
331
+ self.block_start_string = block_start_string
332
+ self.block_end_string = block_end_string
333
+ self.variable_start_string = variable_start_string
334
+ self.variable_end_string = variable_end_string
335
+ self.comment_start_string = comment_start_string
336
+ self.comment_end_string = comment_end_string
337
+ self.line_statement_prefix = line_statement_prefix
338
+ self.line_comment_prefix = line_comment_prefix
339
+ self.trim_blocks = trim_blocks
340
+ self.lstrip_blocks = lstrip_blocks
341
+ self.newline_sequence = newline_sequence
342
+ self.keep_trailing_newline = keep_trailing_newline
343
+
344
+ # runtime information
345
+ self.undefined: t.Type[Undefined] = undefined
346
+ self.optimized = optimized
347
+ self.finalize = finalize
348
+ self.autoescape = autoescape
349
+
350
+ # defaults
351
+ self.filters = DEFAULT_FILTERS.copy()
352
+ self.tests = DEFAULT_TESTS.copy()
353
+ self.globals = DEFAULT_NAMESPACE.copy()
354
+
355
+ # set the loader provided
356
+ self.loader = loader
357
+ self.cache = create_cache(cache_size)
358
+ self.bytecode_cache = bytecode_cache
359
+ self.auto_reload = auto_reload
360
+
361
+ # configurable policies
362
+ self.policies = DEFAULT_POLICIES.copy()
363
+
364
+ # load extensions
365
+ self.extensions = load_extensions(self, extensions)
366
+
367
+ self.is_async = enable_async
368
+ _environment_config_check(self)
369
+
370
+ def add_extension(self, extension: t.Union[str, t.Type["Extension"]]) -> None:
371
+ """Adds an extension after the environment was created.
372
+
373
+ .. versionadded:: 2.5
374
+ """
375
+ self.extensions.update(load_extensions(self, [extension]))
376
+
377
+ def extend(self, **attributes: t.Any) -> None:
378
+ """Add the items to the instance of the environment if they do not exist
379
+ yet. This is used by :ref:`extensions <writing-extensions>` to register
380
+ callbacks and configuration values without breaking inheritance.
381
+ """
382
+ for key, value in attributes.items():
383
+ if not hasattr(self, key):
384
+ setattr(self, key, value)
385
+
386
+ def overlay(
387
+ self,
388
+ block_start_string: str = missing,
389
+ block_end_string: str = missing,
390
+ variable_start_string: str = missing,
391
+ variable_end_string: str = missing,
392
+ comment_start_string: str = missing,
393
+ comment_end_string: str = missing,
394
+ line_statement_prefix: t.Optional[str] = missing,
395
+ line_comment_prefix: t.Optional[str] = missing,
396
+ trim_blocks: bool = missing,
397
+ lstrip_blocks: bool = missing,
398
+ newline_sequence: "te.Literal['\\n', '\\r\\n', '\\r']" = missing,
399
+ keep_trailing_newline: bool = missing,
400
+ extensions: t.Sequence[t.Union[str, t.Type["Extension"]]] = missing,
401
+ optimized: bool = missing,
402
+ undefined: t.Type[Undefined] = missing,
403
+ finalize: t.Optional[t.Callable[..., t.Any]] = missing,
404
+ autoescape: t.Union[bool, t.Callable[[t.Optional[str]], bool]] = missing,
405
+ loader: t.Optional["BaseLoader"] = missing,
406
+ cache_size: int = missing,
407
+ auto_reload: bool = missing,
408
+ bytecode_cache: t.Optional["BytecodeCache"] = missing,
409
+ enable_async: bool = missing,
410
+ ) -> "te.Self":
411
+ """Create a new overlay environment that shares all the data with the
412
+ current environment except for cache and the overridden attributes.
413
+ Extensions cannot be removed for an overlayed environment. An overlayed
414
+ environment automatically gets all the extensions of the environment it
415
+ is linked to plus optional extra extensions.
416
+
417
+ Creating overlays should happen after the initial environment was set
418
+ up completely. Not all attributes are truly linked, some are just
419
+ copied over so modifications on the original environment may not shine
420
+ through.
421
+
422
+ .. versionchanged:: 3.1.5
423
+ ``enable_async`` is applied correctly.
424
+
425
+ .. versionchanged:: 3.1.2
426
+ Added the ``newline_sequence``, ``keep_trailing_newline``,
427
+ and ``enable_async`` parameters to match ``__init__``.
428
+ """
429
+ args = dict(locals())
430
+ del args["self"], args["cache_size"], args["extensions"], args["enable_async"]
431
+
432
+ rv = object.__new__(self.__class__)
433
+ rv.__dict__.update(self.__dict__)
434
+ rv.overlayed = True
435
+ rv.linked_to = self
436
+
437
+ for key, value in args.items():
438
+ if value is not missing:
439
+ setattr(rv, key, value)
440
+
441
+ if cache_size is not missing:
442
+ rv.cache = create_cache(cache_size)
443
+ else:
444
+ rv.cache = copy_cache(self.cache)
445
+
446
+ rv.extensions = {}
447
+ for key, value in self.extensions.items():
448
+ rv.extensions[key] = value.bind(rv)
449
+ if extensions is not missing:
450
+ rv.extensions.update(load_extensions(rv, extensions))
451
+
452
+ if enable_async is not missing:
453
+ rv.is_async = enable_async
454
+
455
+ return _environment_config_check(rv)
456
+
457
+ @property
458
+ def lexer(self) -> Lexer:
459
+ """The lexer for this environment."""
460
+ return get_lexer(self)
461
+
462
+ def iter_extensions(self) -> t.Iterator["Extension"]:
463
+ """Iterates over the extensions by priority."""
464
+ return iter(sorted(self.extensions.values(), key=lambda x: x.priority))
465
+
466
+ def getitem(
467
+ self, obj: t.Any, argument: t.Union[str, t.Any]
468
+ ) -> t.Union[t.Any, Undefined]:
469
+ """Get an item or attribute of an object but prefer the item."""
470
+ try:
471
+ return obj[argument]
472
+ except (AttributeError, TypeError, LookupError):
473
+ if isinstance(argument, str):
474
+ try:
475
+ attr = str(argument)
476
+ except Exception:
477
+ pass
478
+ else:
479
+ try:
480
+ return getattr(obj, attr)
481
+ except AttributeError:
482
+ pass
483
+ return self.undefined(obj=obj, name=argument)
484
+
485
+ def getattr(self, obj: t.Any, attribute: str) -> t.Any:
486
+ """Get an item or attribute of an object but prefer the attribute.
487
+ Unlike :meth:`getitem` the attribute *must* be a string.
488
+ """
489
+ try:
490
+ return getattr(obj, attribute)
491
+ except AttributeError:
492
+ pass
493
+ try:
494
+ return obj[attribute]
495
+ except (TypeError, LookupError, AttributeError):
496
+ return self.undefined(obj=obj, name=attribute)
497
+
498
+ def _filter_test_common(
499
+ self,
500
+ name: t.Union[str, Undefined],
501
+ value: t.Any,
502
+ args: t.Optional[t.Sequence[t.Any]],
503
+ kwargs: t.Optional[t.Mapping[str, t.Any]],
504
+ context: t.Optional[Context],
505
+ eval_ctx: t.Optional[EvalContext],
506
+ is_filter: bool,
507
+ ) -> t.Any:
508
+ if is_filter:
509
+ env_map = self.filters
510
+ type_name = "filter"
511
+ else:
512
+ env_map = self.tests
513
+ type_name = "test"
514
+
515
+ func = env_map.get(name) # type: ignore
516
+
517
+ if func is None:
518
+ msg = f"No {type_name} named {name!r}."
519
+
520
+ if isinstance(name, Undefined):
521
+ try:
522
+ name._fail_with_undefined_error()
523
+ except Exception as e:
524
+ msg = f"{msg} ({e}; did you forget to quote the callable name?)"
525
+
526
+ raise TemplateRuntimeError(msg)
527
+
528
+ args = [value, *(args if args is not None else ())]
529
+ kwargs = kwargs if kwargs is not None else {}
530
+ pass_arg = _PassArg.from_obj(func)
531
+
532
+ if pass_arg is _PassArg.context:
533
+ if context is None:
534
+ raise TemplateRuntimeError(
535
+ f"Attempted to invoke a context {type_name} without context."
536
+ )
537
+
538
+ args.insert(0, context)
539
+ elif pass_arg is _PassArg.eval_context:
540
+ if eval_ctx is None:
541
+ if context is not None:
542
+ eval_ctx = context.eval_ctx
543
+ else:
544
+ eval_ctx = EvalContext(self)
545
+
546
+ args.insert(0, eval_ctx)
547
+ elif pass_arg is _PassArg.environment:
548
+ args.insert(0, self)
549
+
550
+ return func(*args, **kwargs)
551
+
552
+ def call_filter(
553
+ self,
554
+ name: str,
555
+ value: t.Any,
556
+ args: t.Optional[t.Sequence[t.Any]] = None,
557
+ kwargs: t.Optional[t.Mapping[str, t.Any]] = None,
558
+ context: t.Optional[Context] = None,
559
+ eval_ctx: t.Optional[EvalContext] = None,
560
+ ) -> t.Any:
561
+ """Invoke a filter on a value the same way the compiler does.
562
+
563
+ This might return a coroutine if the filter is running from an
564
+ environment in async mode and the filter supports async
565
+ execution. It's your responsibility to await this if needed.
566
+
567
+ .. versionadded:: 2.7
568
+ """
569
+ return self._filter_test_common(
570
+ name, value, args, kwargs, context, eval_ctx, True
571
+ )
572
+
573
+ def call_test(
574
+ self,
575
+ name: str,
576
+ value: t.Any,
577
+ args: t.Optional[t.Sequence[t.Any]] = None,
578
+ kwargs: t.Optional[t.Mapping[str, t.Any]] = None,
579
+ context: t.Optional[Context] = None,
580
+ eval_ctx: t.Optional[EvalContext] = None,
581
+ ) -> t.Any:
582
+ """Invoke a test on a value the same way the compiler does.
583
+
584
+ This might return a coroutine if the test is running from an
585
+ environment in async mode and the test supports async execution.
586
+ It's your responsibility to await this if needed.
587
+
588
+ .. versionchanged:: 3.0
589
+ Tests support ``@pass_context``, etc. decorators. Added
590
+ the ``context`` and ``eval_ctx`` parameters.
591
+
592
+ .. versionadded:: 2.7
593
+ """
594
+ return self._filter_test_common(
595
+ name, value, args, kwargs, context, eval_ctx, False
596
+ )
597
+
598
+ @internalcode
599
+ def parse(
600
+ self,
601
+ source: str,
602
+ name: t.Optional[str] = None,
603
+ filename: t.Optional[str] = None,
604
+ ) -> nodes.Template:
605
+ """Parse the sourcecode and return the abstract syntax tree. This
606
+ tree of nodes is used by the compiler to convert the template into
607
+ executable source- or bytecode. This is useful for debugging or to
608
+ extract information from templates.
609
+
610
+ If you are :ref:`developing Jinja extensions <writing-extensions>`
611
+ this gives you a good overview of the node tree generated.
612
+ """
613
+ try:
614
+ return self._parse(source, name, filename)
615
+ except TemplateSyntaxError:
616
+ self.handle_exception(source=source)
617
+
618
+ def _parse(
619
+ self, source: str, name: t.Optional[str], filename: t.Optional[str]
620
+ ) -> nodes.Template:
621
+ """Internal parsing function used by `parse` and `compile`."""
622
+ return Parser(self, source, name, filename).parse()
623
+
624
+ def lex(
625
+ self,
626
+ source: str,
627
+ name: t.Optional[str] = None,
628
+ filename: t.Optional[str] = None,
629
+ ) -> t.Iterator[t.Tuple[int, str, str]]:
630
+ """Lex the given sourcecode and return a generator that yields
631
+ tokens as tuples in the form ``(lineno, token_type, value)``.
632
+ This can be useful for :ref:`extension development <writing-extensions>`
633
+ and debugging templates.
634
+
635
+ This does not perform preprocessing. If you want the preprocessing
636
+ of the extensions to be applied you have to filter source through
637
+ the :meth:`preprocess` method.
638
+ """
639
+ source = str(source)
640
+ try:
641
+ return self.lexer.tokeniter(source, name, filename)
642
+ except TemplateSyntaxError:
643
+ self.handle_exception(source=source)
644
+
645
+ def preprocess(
646
+ self,
647
+ source: str,
648
+ name: t.Optional[str] = None,
649
+ filename: t.Optional[str] = None,
650
+ ) -> str:
651
+ """Preprocesses the source with all extensions. This is automatically
652
+ called for all parsing and compiling methods but *not* for :meth:`lex`
653
+ because there you usually only want the actual source tokenized.
654
+ """
655
+ return reduce(
656
+ lambda s, e: e.preprocess(s, name, filename),
657
+ self.iter_extensions(),
658
+ str(source),
659
+ )
660
+
661
+ def _tokenize(
662
+ self,
663
+ source: str,
664
+ name: t.Optional[str],
665
+ filename: t.Optional[str] = None,
666
+ state: t.Optional[str] = None,
667
+ ) -> TokenStream:
668
+ """Called by the parser to do the preprocessing and filtering
669
+ for all the extensions. Returns a :class:`~jinja2.lexer.TokenStream`.
670
+ """
671
+ source = self.preprocess(source, name, filename)
672
+ stream = self.lexer.tokenize(source, name, filename, state)
673
+
674
+ for ext in self.iter_extensions():
675
+ stream = ext.filter_stream(stream) # type: ignore
676
+
677
+ if not isinstance(stream, TokenStream):
678
+ stream = TokenStream(stream, name, filename)
679
+
680
+ return stream
681
+
682
+ def _generate(
683
+ self,
684
+ source: nodes.Template,
685
+ name: t.Optional[str],
686
+ filename: t.Optional[str],
687
+ defer_init: bool = False,
688
+ ) -> str:
689
+ """Internal hook that can be overridden to hook a different generate
690
+ method in.
691
+
692
+ .. versionadded:: 2.5
693
+ """
694
+ return generate( # type: ignore
695
+ source,
696
+ self,
697
+ name,
698
+ filename,
699
+ defer_init=defer_init,
700
+ optimized=self.optimized,
701
+ )
702
+
703
+ def _compile(self, source: str, filename: str) -> CodeType:
704
+ """Internal hook that can be overridden to hook a different compile
705
+ method in.
706
+
707
+ .. versionadded:: 2.5
708
+ """
709
+ return compile(source, filename, "exec")
710
+
711
+ @typing.overload
712
+ def compile(
713
+ self,
714
+ source: t.Union[str, nodes.Template],
715
+ name: t.Optional[str] = None,
716
+ filename: t.Optional[str] = None,
717
+ raw: "te.Literal[False]" = False,
718
+ defer_init: bool = False,
719
+ ) -> CodeType: ...
720
+
721
+ @typing.overload
722
+ def compile(
723
+ self,
724
+ source: t.Union[str, nodes.Template],
725
+ name: t.Optional[str] = None,
726
+ filename: t.Optional[str] = None,
727
+ raw: "te.Literal[True]" = ...,
728
+ defer_init: bool = False,
729
+ ) -> str: ...
730
+
731
+ @internalcode
732
+ def compile(
733
+ self,
734
+ source: t.Union[str, nodes.Template],
735
+ name: t.Optional[str] = None,
736
+ filename: t.Optional[str] = None,
737
+ raw: bool = False,
738
+ defer_init: bool = False,
739
+ ) -> t.Union[str, CodeType]:
740
+ """Compile a node or template source code. The `name` parameter is
741
+ the load name of the template after it was joined using
742
+ :meth:`join_path` if necessary, not the filename on the file system.
743
+ the `filename` parameter is the estimated filename of the template on
744
+ the file system. If the template came from a database or memory this
745
+ can be omitted.
746
+
747
+ The return value of this method is a python code object. If the `raw`
748
+ parameter is `True` the return value will be a string with python
749
+ code equivalent to the bytecode returned otherwise. This method is
750
+ mainly used internally.
751
+
752
+ `defer_init` is use internally to aid the module code generator. This
753
+ causes the generated code to be able to import without the global
754
+ environment variable to be set.
755
+
756
+ .. versionadded:: 2.4
757
+ `defer_init` parameter added.
758
+ """
759
+ source_hint = None
760
+ try:
761
+ if isinstance(source, str):
762
+ source_hint = source
763
+ source = self._parse(source, name, filename)
764
+ source = self._generate(source, name, filename, defer_init=defer_init)
765
+ if raw:
766
+ return source
767
+ if filename is None:
768
+ filename = "<template>"
769
+ return self._compile(source, filename)
770
+ except TemplateSyntaxError:
771
+ self.handle_exception(source=source_hint)
772
+
773
+ def compile_expression(
774
+ self, source: str, undefined_to_none: bool = True
775
+ ) -> "TemplateExpression":
776
+ """A handy helper method that returns a callable that accepts keyword
777
+ arguments that appear as variables in the expression. If called it
778
+ returns the result of the expression.
779
+
780
+ This is useful if applications want to use the same rules as Jinja
781
+ in template "configuration files" or similar situations.
782
+
783
+ Example usage:
784
+
785
+ >>> env = Environment()
786
+ >>> expr = env.compile_expression('foo == 42')
787
+ >>> expr(foo=23)
788
+ False
789
+ >>> expr(foo=42)
790
+ True
791
+
792
+ Per default the return value is converted to `None` if the
793
+ expression returns an undefined value. This can be changed
794
+ by setting `undefined_to_none` to `False`.
795
+
796
+ >>> env.compile_expression('var')() is None
797
+ True
798
+ >>> env.compile_expression('var', undefined_to_none=False)()
799
+ Undefined
800
+
801
+ .. versionadded:: 2.1
802
+ """
803
+ parser = Parser(self, source, state="variable")
804
+ try:
805
+ expr = parser.parse_expression()
806
+ if not parser.stream.eos:
807
+ raise TemplateSyntaxError(
808
+ "chunk after expression", parser.stream.current.lineno, None, None
809
+ )
810
+ expr.set_environment(self)
811
+ except TemplateSyntaxError:
812
+ self.handle_exception(source=source)
813
+
814
+ body = [nodes.Assign(nodes.Name("result", "store"), expr, lineno=1)]
815
+ template = self.from_string(nodes.Template(body, lineno=1))
816
+ return TemplateExpression(template, undefined_to_none)
817
+
818
+ def compile_templates(
819
+ self,
820
+ target: t.Union[str, "os.PathLike[str]"],
821
+ extensions: t.Optional[t.Collection[str]] = None,
822
+ filter_func: t.Optional[t.Callable[[str], bool]] = None,
823
+ zip: t.Optional[str] = "deflated",
824
+ log_function: t.Optional[t.Callable[[str], None]] = None,
825
+ ignore_errors: bool = True,
826
+ ) -> None:
827
+ """Finds all the templates the loader can find, compiles them
828
+ and stores them in `target`. If `zip` is `None`, instead of in a
829
+ zipfile, the templates will be stored in a directory.
830
+ By default a deflate zip algorithm is used. To switch to
831
+ the stored algorithm, `zip` can be set to ``'stored'``.
832
+
833
+ `extensions` and `filter_func` are passed to :meth:`list_templates`.
834
+ Each template returned will be compiled to the target folder or
835
+ zipfile.
836
+
837
+ By default template compilation errors are ignored. In case a
838
+ log function is provided, errors are logged. If you want template
839
+ syntax errors to abort the compilation you can set `ignore_errors`
840
+ to `False` and you will get an exception on syntax errors.
841
+
842
+ .. versionadded:: 2.4
843
+ """
844
+ from .loaders import ModuleLoader
845
+
846
+ if log_function is None:
847
+
848
+ def log_function(x: str) -> None:
849
+ pass
850
+
851
+ assert log_function is not None
852
+ assert self.loader is not None, "No loader configured."
853
+
854
+ def write_file(filename: str, data: str) -> None:
855
+ if zip:
856
+ info = ZipInfo(filename)
857
+ info.external_attr = 0o755 << 16
858
+ zip_file.writestr(info, data)
859
+ else:
860
+ with open(os.path.join(target, filename), "wb") as f:
861
+ f.write(data.encode("utf8"))
862
+
863
+ if zip is not None:
864
+ from zipfile import ZIP_DEFLATED
865
+ from zipfile import ZIP_STORED
866
+ from zipfile import ZipFile
867
+ from zipfile import ZipInfo
868
+
869
+ zip_file = ZipFile(
870
+ target, "w", dict(deflated=ZIP_DEFLATED, stored=ZIP_STORED)[zip]
871
+ )
872
+ log_function(f"Compiling into Zip archive {target!r}")
873
+ else:
874
+ if not os.path.isdir(target):
875
+ os.makedirs(target)
876
+ log_function(f"Compiling into folder {target!r}")
877
+
878
+ try:
879
+ for name in self.list_templates(extensions, filter_func):
880
+ source, filename, _ = self.loader.get_source(self, name)
881
+ try:
882
+ code = self.compile(source, name, filename, True, True)
883
+ except TemplateSyntaxError as e:
884
+ if not ignore_errors:
885
+ raise
886
+ log_function(f'Could not compile "{name}": {e}')
887
+ continue
888
+
889
+ filename = ModuleLoader.get_module_filename(name)
890
+
891
+ write_file(filename, code)
892
+ log_function(f'Compiled "{name}" as {filename}')
893
+ finally:
894
+ if zip:
895
+ zip_file.close()
896
+
897
+ log_function("Finished compiling templates")
898
+
899
+ def list_templates(
900
+ self,
901
+ extensions: t.Optional[t.Collection[str]] = None,
902
+ filter_func: t.Optional[t.Callable[[str], bool]] = None,
903
+ ) -> t.List[str]:
904
+ """Returns a list of templates for this environment. This requires
905
+ that the loader supports the loader's
906
+ :meth:`~BaseLoader.list_templates` method.
907
+
908
+ If there are other files in the template folder besides the
909
+ actual templates, the returned list can be filtered. There are two
910
+ ways: either `extensions` is set to a list of file extensions for
911
+ templates, or a `filter_func` can be provided which is a callable that
912
+ is passed a template name and should return `True` if it should end up
913
+ in the result list.
914
+
915
+ If the loader does not support that, a :exc:`TypeError` is raised.
916
+
917
+ .. versionadded:: 2.4
918
+ """
919
+ assert self.loader is not None, "No loader configured."
920
+ names = self.loader.list_templates()
921
+
922
+ if extensions is not None:
923
+ if filter_func is not None:
924
+ raise TypeError(
925
+ "either extensions or filter_func can be passed, but not both"
926
+ )
927
+
928
+ def filter_func(x: str) -> bool:
929
+ return "." in x and x.rsplit(".", 1)[1] in extensions
930
+
931
+ if filter_func is not None:
932
+ names = [name for name in names if filter_func(name)]
933
+
934
+ return names
935
+
936
+ def handle_exception(self, source: t.Optional[str] = None) -> "te.NoReturn":
937
+ """Exception handling helper. This is used internally to either raise
938
+ rewritten exceptions or return a rendered traceback for the template.
939
+ """
940
+ from .debug import rewrite_traceback_stack
941
+
942
+ raise rewrite_traceback_stack(source=source)
943
+
944
+ def join_path(self, template: str, parent: str) -> str:
945
+ """Join a template with the parent. By default all the lookups are
946
+ relative to the loader root so this method returns the `template`
947
+ parameter unchanged, but if the paths should be relative to the
948
+ parent template, this function can be used to calculate the real
949
+ template name.
950
+
951
+ Subclasses may override this method and implement template path
952
+ joining here.
953
+ """
954
+ return template
955
+
956
+ @internalcode
957
+ def _load_template(
958
+ self, name: str, globals: t.Optional[t.MutableMapping[str, t.Any]]
959
+ ) -> "Template":
960
+ if self.loader is None:
961
+ raise TypeError("no loader for this environment specified")
962
+ cache_key = (weakref.ref(self.loader), name)
963
+ if self.cache is not None:
964
+ template = self.cache.get(cache_key)
965
+ if template is not None and (
966
+ not self.auto_reload or template.is_up_to_date
967
+ ):
968
+ # template.globals is a ChainMap, modifying it will only
969
+ # affect the template, not the environment globals.
970
+ if globals:
971
+ template.globals.update(globals)
972
+
973
+ return template
974
+
975
+ template = self.loader.load(self, name, self.make_globals(globals))
976
+
977
+ if self.cache is not None:
978
+ self.cache[cache_key] = template
979
+ return template
980
+
981
+ @internalcode
982
+ def get_template(
983
+ self,
984
+ name: t.Union[str, "Template"],
985
+ parent: t.Optional[str] = None,
986
+ globals: t.Optional[t.MutableMapping[str, t.Any]] = None,
987
+ ) -> "Template":
988
+ """Load a template by name with :attr:`loader` and return a
989
+ :class:`Template`. If the template does not exist a
990
+ :exc:`TemplateNotFound` exception is raised.
991
+
992
+ :param name: Name of the template to load. When loading
993
+ templates from the filesystem, "/" is used as the path
994
+ separator, even on Windows.
995
+ :param parent: The name of the parent template importing this
996
+ template. :meth:`join_path` can be used to implement name
997
+ transformations with this.
998
+ :param globals: Extend the environment :attr:`globals` with
999
+ these extra variables available for all renders of this
1000
+ template. If the template has already been loaded and
1001
+ cached, its globals are updated with any new items.
1002
+
1003
+ .. versionchanged:: 3.0
1004
+ If a template is loaded from cache, ``globals`` will update
1005
+ the template's globals instead of ignoring the new values.
1006
+
1007
+ .. versionchanged:: 2.4
1008
+ If ``name`` is a :class:`Template` object it is returned
1009
+ unchanged.
1010
+ """
1011
+ if isinstance(name, Template):
1012
+ return name
1013
+ if parent is not None:
1014
+ name = self.join_path(name, parent)
1015
+
1016
+ return self._load_template(name, globals)
1017
+
1018
+ @internalcode
1019
+ def select_template(
1020
+ self,
1021
+ names: t.Iterable[t.Union[str, "Template"]],
1022
+ parent: t.Optional[str] = None,
1023
+ globals: t.Optional[t.MutableMapping[str, t.Any]] = None,
1024
+ ) -> "Template":
1025
+ """Like :meth:`get_template`, but tries loading multiple names.
1026
+ If none of the names can be loaded a :exc:`TemplatesNotFound`
1027
+ exception is raised.
1028
+
1029
+ :param names: List of template names to try loading in order.
1030
+ :param parent: The name of the parent template importing this
1031
+ template. :meth:`join_path` can be used to implement name
1032
+ transformations with this.
1033
+ :param globals: Extend the environment :attr:`globals` with
1034
+ these extra variables available for all renders of this
1035
+ template. If the template has already been loaded and
1036
+ cached, its globals are updated with any new items.
1037
+
1038
+ .. versionchanged:: 3.0
1039
+ If a template is loaded from cache, ``globals`` will update
1040
+ the template's globals instead of ignoring the new values.
1041
+
1042
+ .. versionchanged:: 2.11
1043
+ If ``names`` is :class:`Undefined`, an :exc:`UndefinedError`
1044
+ is raised instead. If no templates were found and ``names``
1045
+ contains :class:`Undefined`, the message is more helpful.
1046
+
1047
+ .. versionchanged:: 2.4
1048
+ If ``names`` contains a :class:`Template` object it is
1049
+ returned unchanged.
1050
+
1051
+ .. versionadded:: 2.3
1052
+ """
1053
+ if isinstance(names, Undefined):
1054
+ names._fail_with_undefined_error()
1055
+
1056
+ if not names:
1057
+ raise TemplatesNotFound(
1058
+ message="Tried to select from an empty list of templates."
1059
+ )
1060
+
1061
+ for name in names:
1062
+ if isinstance(name, Template):
1063
+ return name
1064
+ if parent is not None:
1065
+ name = self.join_path(name, parent)
1066
+ try:
1067
+ return self._load_template(name, globals)
1068
+ except (TemplateNotFound, UndefinedError):
1069
+ pass
1070
+ raise TemplatesNotFound(names) # type: ignore
1071
+
1072
+ @internalcode
1073
+ def get_or_select_template(
1074
+ self,
1075
+ template_name_or_list: t.Union[
1076
+ str, "Template", t.List[t.Union[str, "Template"]]
1077
+ ],
1078
+ parent: t.Optional[str] = None,
1079
+ globals: t.Optional[t.MutableMapping[str, t.Any]] = None,
1080
+ ) -> "Template":
1081
+ """Use :meth:`select_template` if an iterable of template names
1082
+ is given, or :meth:`get_template` if one name is given.
1083
+
1084
+ .. versionadded:: 2.3
1085
+ """
1086
+ if isinstance(template_name_or_list, (str, Undefined)):
1087
+ return self.get_template(template_name_or_list, parent, globals)
1088
+ elif isinstance(template_name_or_list, Template):
1089
+ return template_name_or_list
1090
+ return self.select_template(template_name_or_list, parent, globals)
1091
+
1092
+ def from_string(
1093
+ self,
1094
+ source: t.Union[str, nodes.Template],
1095
+ globals: t.Optional[t.MutableMapping[str, t.Any]] = None,
1096
+ template_class: t.Optional[t.Type["Template"]] = None,
1097
+ ) -> "Template":
1098
+ """Load a template from a source string without using
1099
+ :attr:`loader`.
1100
+
1101
+ :param source: Jinja source to compile into a template.
1102
+ :param globals: Extend the environment :attr:`globals` with
1103
+ these extra variables available for all renders of this
1104
+ template. If the template has already been loaded and
1105
+ cached, its globals are updated with any new items.
1106
+ :param template_class: Return an instance of this
1107
+ :class:`Template` class.
1108
+ """
1109
+ gs = self.make_globals(globals)
1110
+ cls = template_class or self.template_class
1111
+ return cls.from_code(self, self.compile(source), gs, None)
1112
+
1113
+ def make_globals(
1114
+ self, d: t.Optional[t.MutableMapping[str, t.Any]]
1115
+ ) -> t.MutableMapping[str, t.Any]:
1116
+ """Make the globals map for a template. Any given template
1117
+ globals overlay the environment :attr:`globals`.
1118
+
1119
+ Returns a :class:`collections.ChainMap`. This allows any changes
1120
+ to a template's globals to only affect that template, while
1121
+ changes to the environment's globals are still reflected.
1122
+ However, avoid modifying any globals after a template is loaded.
1123
+
1124
+ :param d: Dict of template-specific globals.
1125
+
1126
+ .. versionchanged:: 3.0
1127
+ Use :class:`collections.ChainMap` to always prevent mutating
1128
+ environment globals.
1129
+ """
1130
+ if d is None:
1131
+ d = {}
1132
+
1133
+ return ChainMap(d, self.globals)
1134
+
1135
+
1136
+ class Template:
1137
+ """A compiled template that can be rendered.
1138
+
1139
+ Use the methods on :class:`Environment` to create or load templates.
1140
+ The environment is used to configure how templates are compiled and
1141
+ behave.
1142
+
1143
+ It is also possible to create a template object directly. This is
1144
+ not usually recommended. The constructor takes most of the same
1145
+ arguments as :class:`Environment`. All templates created with the
1146
+ same environment arguments share the same ephemeral ``Environment``
1147
+ instance behind the scenes.
1148
+
1149
+ A template object should be considered immutable. Modifications on
1150
+ the object are not supported.
1151
+ """
1152
+
1153
+ #: Type of environment to create when creating a template directly
1154
+ #: rather than through an existing environment.
1155
+ environment_class: t.Type[Environment] = Environment
1156
+
1157
+ environment: Environment
1158
+ globals: t.MutableMapping[str, t.Any]
1159
+ name: t.Optional[str]
1160
+ filename: t.Optional[str]
1161
+ blocks: t.Dict[str, t.Callable[[Context], t.Iterator[str]]]
1162
+ root_render_func: t.Callable[[Context], t.Iterator[str]]
1163
+ _module: t.Optional["TemplateModule"]
1164
+ _debug_info: str
1165
+ _uptodate: t.Optional[t.Callable[[], bool]]
1166
+
1167
+ def __new__(
1168
+ cls,
1169
+ source: t.Union[str, nodes.Template],
1170
+ block_start_string: str = BLOCK_START_STRING,
1171
+ block_end_string: str = BLOCK_END_STRING,
1172
+ variable_start_string: str = VARIABLE_START_STRING,
1173
+ variable_end_string: str = VARIABLE_END_STRING,
1174
+ comment_start_string: str = COMMENT_START_STRING,
1175
+ comment_end_string: str = COMMENT_END_STRING,
1176
+ line_statement_prefix: t.Optional[str] = LINE_STATEMENT_PREFIX,
1177
+ line_comment_prefix: t.Optional[str] = LINE_COMMENT_PREFIX,
1178
+ trim_blocks: bool = TRIM_BLOCKS,
1179
+ lstrip_blocks: bool = LSTRIP_BLOCKS,
1180
+ newline_sequence: "te.Literal['\\n', '\\r\\n', '\\r']" = NEWLINE_SEQUENCE,
1181
+ keep_trailing_newline: bool = KEEP_TRAILING_NEWLINE,
1182
+ extensions: t.Sequence[t.Union[str, t.Type["Extension"]]] = (),
1183
+ optimized: bool = True,
1184
+ undefined: t.Type[Undefined] = Undefined,
1185
+ finalize: t.Optional[t.Callable[..., t.Any]] = None,
1186
+ autoescape: t.Union[bool, t.Callable[[t.Optional[str]], bool]] = False,
1187
+ enable_async: bool = False,
1188
+ ) -> t.Any: # it returns a `Template`, but this breaks the sphinx build...
1189
+ env = get_spontaneous_environment(
1190
+ cls.environment_class, # type: ignore
1191
+ block_start_string,
1192
+ block_end_string,
1193
+ variable_start_string,
1194
+ variable_end_string,
1195
+ comment_start_string,
1196
+ comment_end_string,
1197
+ line_statement_prefix,
1198
+ line_comment_prefix,
1199
+ trim_blocks,
1200
+ lstrip_blocks,
1201
+ newline_sequence,
1202
+ keep_trailing_newline,
1203
+ frozenset(extensions),
1204
+ optimized,
1205
+ undefined, # type: ignore
1206
+ finalize,
1207
+ autoescape,
1208
+ None,
1209
+ 0,
1210
+ False,
1211
+ None,
1212
+ enable_async,
1213
+ )
1214
+ return env.from_string(source, template_class=cls)
1215
+
1216
+ @classmethod
1217
+ def from_code(
1218
+ cls,
1219
+ environment: Environment,
1220
+ code: CodeType,
1221
+ globals: t.MutableMapping[str, t.Any],
1222
+ uptodate: t.Optional[t.Callable[[], bool]] = None,
1223
+ ) -> "Template":
1224
+ """Creates a template object from compiled code and the globals. This
1225
+ is used by the loaders and environment to create a template object.
1226
+ """
1227
+ namespace = {"environment": environment, "__file__": code.co_filename}
1228
+ exec(code, namespace)
1229
+ rv = cls._from_namespace(environment, namespace, globals)
1230
+ rv._uptodate = uptodate
1231
+ return rv
1232
+
1233
+ @classmethod
1234
+ def from_module_dict(
1235
+ cls,
1236
+ environment: Environment,
1237
+ module_dict: t.MutableMapping[str, t.Any],
1238
+ globals: t.MutableMapping[str, t.Any],
1239
+ ) -> "Template":
1240
+ """Creates a template object from a module. This is used by the
1241
+ module loader to create a template object.
1242
+
1243
+ .. versionadded:: 2.4
1244
+ """
1245
+ return cls._from_namespace(environment, module_dict, globals)
1246
+
1247
+ @classmethod
1248
+ def _from_namespace(
1249
+ cls,
1250
+ environment: Environment,
1251
+ namespace: t.MutableMapping[str, t.Any],
1252
+ globals: t.MutableMapping[str, t.Any],
1253
+ ) -> "Template":
1254
+ t: Template = object.__new__(cls)
1255
+ t.environment = environment
1256
+ t.globals = globals
1257
+ t.name = namespace["name"]
1258
+ t.filename = namespace["__file__"]
1259
+ t.blocks = namespace["blocks"]
1260
+
1261
+ # render function and module
1262
+ t.root_render_func = namespace["root"]
1263
+ t._module = None
1264
+
1265
+ # debug and loader helpers
1266
+ t._debug_info = namespace["debug_info"]
1267
+ t._uptodate = None
1268
+
1269
+ # store the reference
1270
+ namespace["environment"] = environment
1271
+ namespace["__jinja_template__"] = t
1272
+
1273
+ return t
1274
+
1275
+ def render(self, *args: t.Any, **kwargs: t.Any) -> str:
1276
+ """This method accepts the same arguments as the `dict` constructor:
1277
+ A dict, a dict subclass or some keyword arguments. If no arguments
1278
+ are given the context will be empty. These two calls do the same::
1279
+
1280
+ template.render(knights='that say nih')
1281
+ template.render({'knights': 'that say nih'})
1282
+
1283
+ This will return the rendered template as a string.
1284
+ """
1285
+ if self.environment.is_async:
1286
+ import asyncio
1287
+
1288
+ return asyncio.run(self.render_async(*args, **kwargs))
1289
+
1290
+ ctx = self.new_context(dict(*args, **kwargs))
1291
+
1292
+ try:
1293
+ return self.environment.concat(self.root_render_func(ctx)) # type: ignore
1294
+ except Exception:
1295
+ self.environment.handle_exception()
1296
+
1297
+ async def render_async(self, *args: t.Any, **kwargs: t.Any) -> str:
1298
+ """This works similar to :meth:`render` but returns a coroutine
1299
+ that when awaited returns the entire rendered template string. This
1300
+ requires the async feature to be enabled.
1301
+
1302
+ Example usage::
1303
+
1304
+ await template.render_async(knights='that say nih; asynchronously')
1305
+ """
1306
+ if not self.environment.is_async:
1307
+ raise RuntimeError(
1308
+ "The environment was not created with async mode enabled."
1309
+ )
1310
+
1311
+ ctx = self.new_context(dict(*args, **kwargs))
1312
+
1313
+ try:
1314
+ return self.environment.concat( # type: ignore
1315
+ [n async for n in self.root_render_func(ctx)] # type: ignore
1316
+ )
1317
+ except Exception:
1318
+ return self.environment.handle_exception()
1319
+
1320
+ def stream(self, *args: t.Any, **kwargs: t.Any) -> "TemplateStream":
1321
+ """Works exactly like :meth:`generate` but returns a
1322
+ :class:`TemplateStream`.
1323
+ """
1324
+ return TemplateStream(self.generate(*args, **kwargs))
1325
+
1326
+ def generate(self, *args: t.Any, **kwargs: t.Any) -> t.Iterator[str]:
1327
+ """For very large templates it can be useful to not render the whole
1328
+ template at once but evaluate each statement after another and yield
1329
+ piece for piece. This method basically does exactly that and returns
1330
+ a generator that yields one item after another as strings.
1331
+
1332
+ It accepts the same arguments as :meth:`render`.
1333
+ """
1334
+ if self.environment.is_async:
1335
+ import asyncio
1336
+
1337
+ async def to_list() -> t.List[str]:
1338
+ return [x async for x in self.generate_async(*args, **kwargs)]
1339
+
1340
+ yield from asyncio.run(to_list())
1341
+ return
1342
+
1343
+ ctx = self.new_context(dict(*args, **kwargs))
1344
+
1345
+ try:
1346
+ yield from self.root_render_func(ctx)
1347
+ except Exception:
1348
+ yield self.environment.handle_exception()
1349
+
1350
+ async def generate_async(
1351
+ self, *args: t.Any, **kwargs: t.Any
1352
+ ) -> t.AsyncGenerator[str, object]:
1353
+ """An async version of :meth:`generate`. Works very similarly but
1354
+ returns an async iterator instead.
1355
+ """
1356
+ if not self.environment.is_async:
1357
+ raise RuntimeError(
1358
+ "The environment was not created with async mode enabled."
1359
+ )
1360
+
1361
+ ctx = self.new_context(dict(*args, **kwargs))
1362
+
1363
+ try:
1364
+ agen = self.root_render_func(ctx)
1365
+ try:
1366
+ async for event in agen: # type: ignore
1367
+ yield event
1368
+ finally:
1369
+ # we can't use async with aclosing(...) because that's only
1370
+ # in 3.10+
1371
+ await agen.aclose() # type: ignore
1372
+ except Exception:
1373
+ yield self.environment.handle_exception()
1374
+
1375
+ def new_context(
1376
+ self,
1377
+ vars: t.Optional[t.Dict[str, t.Any]] = None,
1378
+ shared: bool = False,
1379
+ locals: t.Optional[t.Mapping[str, t.Any]] = None,
1380
+ ) -> Context:
1381
+ """Create a new :class:`Context` for this template. The vars
1382
+ provided will be passed to the template. Per default the globals
1383
+ are added to the context. If shared is set to `True` the data
1384
+ is passed as is to the context without adding the globals.
1385
+
1386
+ `locals` can be a dict of local variables for internal usage.
1387
+ """
1388
+ return new_context(
1389
+ self.environment, self.name, self.blocks, vars, shared, self.globals, locals
1390
+ )
1391
+
1392
+ def make_module(
1393
+ self,
1394
+ vars: t.Optional[t.Dict[str, t.Any]] = None,
1395
+ shared: bool = False,
1396
+ locals: t.Optional[t.Mapping[str, t.Any]] = None,
1397
+ ) -> "TemplateModule":
1398
+ """This method works like the :attr:`module` attribute when called
1399
+ without arguments but it will evaluate the template on every call
1400
+ rather than caching it. It's also possible to provide
1401
+ a dict which is then used as context. The arguments are the same
1402
+ as for the :meth:`new_context` method.
1403
+ """
1404
+ ctx = self.new_context(vars, shared, locals)
1405
+ return TemplateModule(self, ctx)
1406
+
1407
+ async def make_module_async(
1408
+ self,
1409
+ vars: t.Optional[t.Dict[str, t.Any]] = None,
1410
+ shared: bool = False,
1411
+ locals: t.Optional[t.Mapping[str, t.Any]] = None,
1412
+ ) -> "TemplateModule":
1413
+ """As template module creation can invoke template code for
1414
+ asynchronous executions this method must be used instead of the
1415
+ normal :meth:`make_module` one. Likewise the module attribute
1416
+ becomes unavailable in async mode.
1417
+ """
1418
+ ctx = self.new_context(vars, shared, locals)
1419
+ return TemplateModule(
1420
+ self,
1421
+ ctx,
1422
+ [x async for x in self.root_render_func(ctx)], # type: ignore
1423
+ )
1424
+
1425
+ @internalcode
1426
+ def _get_default_module(self, ctx: t.Optional[Context] = None) -> "TemplateModule":
1427
+ """If a context is passed in, this means that the template was
1428
+ imported. Imported templates have access to the current
1429
+ template's globals by default, but they can only be accessed via
1430
+ the context during runtime.
1431
+
1432
+ If there are new globals, we need to create a new module because
1433
+ the cached module is already rendered and will not have access
1434
+ to globals from the current context. This new module is not
1435
+ cached because the template can be imported elsewhere, and it
1436
+ should have access to only the current template's globals.
1437
+ """
1438
+ if self.environment.is_async:
1439
+ raise RuntimeError("Module is not available in async mode.")
1440
+
1441
+ if ctx is not None:
1442
+ keys = ctx.globals_keys - self.globals.keys()
1443
+
1444
+ if keys:
1445
+ return self.make_module({k: ctx.parent[k] for k in keys})
1446
+
1447
+ if self._module is None:
1448
+ self._module = self.make_module()
1449
+
1450
+ return self._module
1451
+
1452
+ async def _get_default_module_async(
1453
+ self, ctx: t.Optional[Context] = None
1454
+ ) -> "TemplateModule":
1455
+ if ctx is not None:
1456
+ keys = ctx.globals_keys - self.globals.keys()
1457
+
1458
+ if keys:
1459
+ return await self.make_module_async({k: ctx.parent[k] for k in keys})
1460
+
1461
+ if self._module is None:
1462
+ self._module = await self.make_module_async()
1463
+
1464
+ return self._module
1465
+
1466
+ @property
1467
+ def module(self) -> "TemplateModule":
1468
+ """The template as module. This is used for imports in the
1469
+ template runtime but is also useful if one wants to access
1470
+ exported template variables from the Python layer:
1471
+
1472
+ >>> t = Template('{% macro foo() %}42{% endmacro %}23')
1473
+ >>> str(t.module)
1474
+ '23'
1475
+ >>> t.module.foo() == u'42'
1476
+ True
1477
+
1478
+ This attribute is not available if async mode is enabled.
1479
+ """
1480
+ return self._get_default_module()
1481
+
1482
+ def get_corresponding_lineno(self, lineno: int) -> int:
1483
+ """Return the source line number of a line number in the
1484
+ generated bytecode as they are not in sync.
1485
+ """
1486
+ for template_line, code_line in reversed(self.debug_info):
1487
+ if code_line <= lineno:
1488
+ return template_line
1489
+ return 1
1490
+
1491
+ @property
1492
+ def is_up_to_date(self) -> bool:
1493
+ """If this variable is `False` there is a newer version available."""
1494
+ if self._uptodate is None:
1495
+ return True
1496
+ return self._uptodate()
1497
+
1498
+ @property
1499
+ def debug_info(self) -> t.List[t.Tuple[int, int]]:
1500
+ """The debug info mapping."""
1501
+ if self._debug_info:
1502
+ return [
1503
+ tuple(map(int, x.split("="))) # type: ignore
1504
+ for x in self._debug_info.split("&")
1505
+ ]
1506
+
1507
+ return []
1508
+
1509
+ def __repr__(self) -> str:
1510
+ if self.name is None:
1511
+ name = f"memory:{id(self):x}"
1512
+ else:
1513
+ name = repr(self.name)
1514
+ return f"<{type(self).__name__} {name}>"
1515
+
1516
+
1517
+ class TemplateModule:
1518
+ """Represents an imported template. All the exported names of the
1519
+ template are available as attributes on this object. Additionally
1520
+ converting it into a string renders the contents.
1521
+ """
1522
+
1523
+ def __init__(
1524
+ self,
1525
+ template: Template,
1526
+ context: Context,
1527
+ body_stream: t.Optional[t.Iterable[str]] = None,
1528
+ ) -> None:
1529
+ if body_stream is None:
1530
+ if context.environment.is_async:
1531
+ raise RuntimeError(
1532
+ "Async mode requires a body stream to be passed to"
1533
+ " a template module. Use the async methods of the"
1534
+ " API you are using."
1535
+ )
1536
+
1537
+ body_stream = list(template.root_render_func(context))
1538
+
1539
+ self._body_stream = body_stream
1540
+ self.__dict__.update(context.get_exported())
1541
+ self.__name__ = template.name
1542
+
1543
+ def __html__(self) -> Markup:
1544
+ return Markup(concat(self._body_stream))
1545
+
1546
+ def __str__(self) -> str:
1547
+ return concat(self._body_stream)
1548
+
1549
+ def __repr__(self) -> str:
1550
+ if self.__name__ is None:
1551
+ name = f"memory:{id(self):x}"
1552
+ else:
1553
+ name = repr(self.__name__)
1554
+ return f"<{type(self).__name__} {name}>"
1555
+
1556
+
1557
+ class TemplateExpression:
1558
+ """The :meth:`jinja2.Environment.compile_expression` method returns an
1559
+ instance of this object. It encapsulates the expression-like access
1560
+ to the template with an expression it wraps.
1561
+ """
1562
+
1563
+ def __init__(self, template: Template, undefined_to_none: bool) -> None:
1564
+ self._template = template
1565
+ self._undefined_to_none = undefined_to_none
1566
+
1567
+ def __call__(self, *args: t.Any, **kwargs: t.Any) -> t.Optional[t.Any]:
1568
+ context = self._template.new_context(dict(*args, **kwargs))
1569
+ consume(self._template.root_render_func(context))
1570
+ rv = context.vars["result"]
1571
+ if self._undefined_to_none and isinstance(rv, Undefined):
1572
+ rv = None
1573
+ return rv
1574
+
1575
+
1576
+ class TemplateStream:
1577
+ """A template stream works pretty much like an ordinary python generator
1578
+ but it can buffer multiple items to reduce the number of total iterations.
1579
+ Per default the output is unbuffered which means that for every unbuffered
1580
+ instruction in the template one string is yielded.
1581
+
1582
+ If buffering is enabled with a buffer size of 5, five items are combined
1583
+ into a new string. This is mainly useful if you are streaming
1584
+ big templates to a client via WSGI which flushes after each iteration.
1585
+ """
1586
+
1587
+ def __init__(self, gen: t.Iterator[str]) -> None:
1588
+ self._gen = gen
1589
+ self.disable_buffering()
1590
+
1591
+ def dump(
1592
+ self,
1593
+ fp: t.Union[str, t.IO[bytes]],
1594
+ encoding: t.Optional[str] = None,
1595
+ errors: t.Optional[str] = "strict",
1596
+ ) -> None:
1597
+ """Dump the complete stream into a file or file-like object.
1598
+ Per default strings are written, if you want to encode
1599
+ before writing specify an `encoding`.
1600
+
1601
+ Example usage::
1602
+
1603
+ Template('Hello {{ name }}!').stream(name='foo').dump('hello.html')
1604
+ """
1605
+ close = False
1606
+
1607
+ if isinstance(fp, str):
1608
+ if encoding is None:
1609
+ encoding = "utf-8"
1610
+
1611
+ real_fp: t.IO[bytes] = open(fp, "wb")
1612
+ close = True
1613
+ else:
1614
+ real_fp = fp
1615
+
1616
+ try:
1617
+ if encoding is not None:
1618
+ iterable = (x.encode(encoding, errors) for x in self) # type: ignore
1619
+ else:
1620
+ iterable = self # type: ignore
1621
+
1622
+ if hasattr(real_fp, "writelines"):
1623
+ real_fp.writelines(iterable)
1624
+ else:
1625
+ for item in iterable:
1626
+ real_fp.write(item)
1627
+ finally:
1628
+ if close:
1629
+ real_fp.close()
1630
+
1631
+ def disable_buffering(self) -> None:
1632
+ """Disable the output buffering."""
1633
+ self._next = partial(next, self._gen)
1634
+ self.buffered = False
1635
+
1636
+ def _buffered_generator(self, size: int) -> t.Iterator[str]:
1637
+ buf: t.List[str] = []
1638
+ c_size = 0
1639
+ push = buf.append
1640
+
1641
+ while True:
1642
+ try:
1643
+ while c_size < size:
1644
+ c = next(self._gen)
1645
+ push(c)
1646
+ if c:
1647
+ c_size += 1
1648
+ except StopIteration:
1649
+ if not c_size:
1650
+ return
1651
+ yield concat(buf)
1652
+ del buf[:]
1653
+ c_size = 0
1654
+
1655
+ def enable_buffering(self, size: int = 5) -> None:
1656
+ """Enable buffering. Buffer `size` items before yielding them."""
1657
+ if size <= 1:
1658
+ raise ValueError("buffer size too small")
1659
+
1660
+ self.buffered = True
1661
+ self._next = partial(next, self._buffered_generator(size))
1662
+
1663
+ def __iter__(self) -> "TemplateStream":
1664
+ return self
1665
+
1666
+ def __next__(self) -> str:
1667
+ return self._next() # type: ignore
1668
+
1669
+
1670
+ # hook in default template class. if anyone reads this comment: ignore that
1671
+ # it's possible to use custom templates ;-)
1672
+ Environment.template_class = Template
janus/lib/python3.10/site-packages/jinja2/exceptions.py ADDED
@@ -0,0 +1,166 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import typing as t
2
+
3
+ if t.TYPE_CHECKING:
4
+ from .runtime import Undefined
5
+
6
+
7
+ class TemplateError(Exception):
8
+ """Baseclass for all template errors."""
9
+
10
+ def __init__(self, message: t.Optional[str] = None) -> None:
11
+ super().__init__(message)
12
+
13
+ @property
14
+ def message(self) -> t.Optional[str]:
15
+ return self.args[0] if self.args else None
16
+
17
+
18
+ class TemplateNotFound(IOError, LookupError, TemplateError):
19
+ """Raised if a template does not exist.
20
+
21
+ .. versionchanged:: 2.11
22
+ If the given name is :class:`Undefined` and no message was
23
+ provided, an :exc:`UndefinedError` is raised.
24
+ """
25
+
26
+ # Silence the Python warning about message being deprecated since
27
+ # it's not valid here.
28
+ message: t.Optional[str] = None
29
+
30
+ def __init__(
31
+ self,
32
+ name: t.Optional[t.Union[str, "Undefined"]],
33
+ message: t.Optional[str] = None,
34
+ ) -> None:
35
+ IOError.__init__(self, name)
36
+
37
+ if message is None:
38
+ from .runtime import Undefined
39
+
40
+ if isinstance(name, Undefined):
41
+ name._fail_with_undefined_error()
42
+
43
+ message = name
44
+
45
+ self.message = message
46
+ self.name = name
47
+ self.templates = [name]
48
+
49
+ def __str__(self) -> str:
50
+ return str(self.message)
51
+
52
+
53
+ class TemplatesNotFound(TemplateNotFound):
54
+ """Like :class:`TemplateNotFound` but raised if multiple templates
55
+ are selected. This is a subclass of :class:`TemplateNotFound`
56
+ exception, so just catching the base exception will catch both.
57
+
58
+ .. versionchanged:: 2.11
59
+ If a name in the list of names is :class:`Undefined`, a message
60
+ about it being undefined is shown rather than the empty string.
61
+
62
+ .. versionadded:: 2.2
63
+ """
64
+
65
+ def __init__(
66
+ self,
67
+ names: t.Sequence[t.Union[str, "Undefined"]] = (),
68
+ message: t.Optional[str] = None,
69
+ ) -> None:
70
+ if message is None:
71
+ from .runtime import Undefined
72
+
73
+ parts = []
74
+
75
+ for name in names:
76
+ if isinstance(name, Undefined):
77
+ parts.append(name._undefined_message)
78
+ else:
79
+ parts.append(name)
80
+
81
+ parts_str = ", ".join(map(str, parts))
82
+ message = f"none of the templates given were found: {parts_str}"
83
+
84
+ super().__init__(names[-1] if names else None, message)
85
+ self.templates = list(names)
86
+
87
+
88
+ class TemplateSyntaxError(TemplateError):
89
+ """Raised to tell the user that there is a problem with the template."""
90
+
91
+ def __init__(
92
+ self,
93
+ message: str,
94
+ lineno: int,
95
+ name: t.Optional[str] = None,
96
+ filename: t.Optional[str] = None,
97
+ ) -> None:
98
+ super().__init__(message)
99
+ self.lineno = lineno
100
+ self.name = name
101
+ self.filename = filename
102
+ self.source: t.Optional[str] = None
103
+
104
+ # this is set to True if the debug.translate_syntax_error
105
+ # function translated the syntax error into a new traceback
106
+ self.translated = False
107
+
108
+ def __str__(self) -> str:
109
+ # for translated errors we only return the message
110
+ if self.translated:
111
+ return t.cast(str, self.message)
112
+
113
+ # otherwise attach some stuff
114
+ location = f"line {self.lineno}"
115
+ name = self.filename or self.name
116
+ if name:
117
+ location = f'File "{name}", {location}'
118
+ lines = [t.cast(str, self.message), " " + location]
119
+
120
+ # if the source is set, add the line to the output
121
+ if self.source is not None:
122
+ try:
123
+ line = self.source.splitlines()[self.lineno - 1]
124
+ except IndexError:
125
+ pass
126
+ else:
127
+ lines.append(" " + line.strip())
128
+
129
+ return "\n".join(lines)
130
+
131
+ def __reduce__(self): # type: ignore
132
+ # https://bugs.python.org/issue1692335 Exceptions that take
133
+ # multiple required arguments have problems with pickling.
134
+ # Without this, raises TypeError: __init__() missing 1 required
135
+ # positional argument: 'lineno'
136
+ return self.__class__, (self.message, self.lineno, self.name, self.filename)
137
+
138
+
139
+ class TemplateAssertionError(TemplateSyntaxError):
140
+ """Like a template syntax error, but covers cases where something in the
141
+ template caused an error at compile time that wasn't necessarily caused
142
+ by a syntax error. However it's a direct subclass of
143
+ :exc:`TemplateSyntaxError` and has the same attributes.
144
+ """
145
+
146
+
147
+ class TemplateRuntimeError(TemplateError):
148
+ """A generic runtime error in the template engine. Under some situations
149
+ Jinja may raise this exception.
150
+ """
151
+
152
+
153
+ class UndefinedError(TemplateRuntimeError):
154
+ """Raised if a template tries to operate on :class:`Undefined`."""
155
+
156
+
157
+ class SecurityError(TemplateRuntimeError):
158
+ """Raised if a template tries to do something insecure if the
159
+ sandbox is enabled.
160
+ """
161
+
162
+
163
+ class FilterArgumentError(TemplateRuntimeError):
164
+ """This error is raised if a filter was called with inappropriate
165
+ arguments
166
+ """
janus/lib/python3.10/site-packages/jinja2/ext.py ADDED
@@ -0,0 +1,870 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """Extension API for adding custom tags and behavior."""
2
+
3
+ import pprint
4
+ import re
5
+ import typing as t
6
+
7
+ from markupsafe import Markup
8
+
9
+ from . import defaults
10
+ from . import nodes
11
+ from .environment import Environment
12
+ from .exceptions import TemplateAssertionError
13
+ from .exceptions import TemplateSyntaxError
14
+ from .runtime import concat # type: ignore
15
+ from .runtime import Context
16
+ from .runtime import Undefined
17
+ from .utils import import_string
18
+ from .utils import pass_context
19
+
20
+ if t.TYPE_CHECKING:
21
+ import typing_extensions as te
22
+
23
+ from .lexer import Token
24
+ from .lexer import TokenStream
25
+ from .parser import Parser
26
+
27
+ class _TranslationsBasic(te.Protocol):
28
+ def gettext(self, message: str) -> str: ...
29
+
30
+ def ngettext(self, singular: str, plural: str, n: int) -> str:
31
+ pass
32
+
33
+ class _TranslationsContext(_TranslationsBasic):
34
+ def pgettext(self, context: str, message: str) -> str: ...
35
+
36
+ def npgettext(
37
+ self, context: str, singular: str, plural: str, n: int
38
+ ) -> str: ...
39
+
40
+ _SupportedTranslations = t.Union[_TranslationsBasic, _TranslationsContext]
41
+
42
+
43
+ # I18N functions available in Jinja templates. If the I18N library
44
+ # provides ugettext, it will be assigned to gettext.
45
+ GETTEXT_FUNCTIONS: t.Tuple[str, ...] = (
46
+ "_",
47
+ "gettext",
48
+ "ngettext",
49
+ "pgettext",
50
+ "npgettext",
51
+ )
52
+ _ws_re = re.compile(r"\s*\n\s*")
53
+
54
+
55
+ class Extension:
56
+ """Extensions can be used to add extra functionality to the Jinja template
57
+ system at the parser level. Custom extensions are bound to an environment
58
+ but may not store environment specific data on `self`. The reason for
59
+ this is that an extension can be bound to another environment (for
60
+ overlays) by creating a copy and reassigning the `environment` attribute.
61
+
62
+ As extensions are created by the environment they cannot accept any
63
+ arguments for configuration. One may want to work around that by using
64
+ a factory function, but that is not possible as extensions are identified
65
+ by their import name. The correct way to configure the extension is
66
+ storing the configuration values on the environment. Because this way the
67
+ environment ends up acting as central configuration storage the
68
+ attributes may clash which is why extensions have to ensure that the names
69
+ they choose for configuration are not too generic. ``prefix`` for example
70
+ is a terrible name, ``fragment_cache_prefix`` on the other hand is a good
71
+ name as includes the name of the extension (fragment cache).
72
+ """
73
+
74
+ identifier: t.ClassVar[str]
75
+
76
+ def __init_subclass__(cls) -> None:
77
+ cls.identifier = f"{cls.__module__}.{cls.__name__}"
78
+
79
+ #: if this extension parses this is the list of tags it's listening to.
80
+ tags: t.Set[str] = set()
81
+
82
+ #: the priority of that extension. This is especially useful for
83
+ #: extensions that preprocess values. A lower value means higher
84
+ #: priority.
85
+ #:
86
+ #: .. versionadded:: 2.4
87
+ priority = 100
88
+
89
+ def __init__(self, environment: Environment) -> None:
90
+ self.environment = environment
91
+
92
+ def bind(self, environment: Environment) -> "te.Self":
93
+ """Create a copy of this extension bound to another environment."""
94
+ rv = object.__new__(self.__class__)
95
+ rv.__dict__.update(self.__dict__)
96
+ rv.environment = environment
97
+ return rv
98
+
99
+ def preprocess(
100
+ self, source: str, name: t.Optional[str], filename: t.Optional[str] = None
101
+ ) -> str:
102
+ """This method is called before the actual lexing and can be used to
103
+ preprocess the source. The `filename` is optional. The return value
104
+ must be the preprocessed source.
105
+ """
106
+ return source
107
+
108
+ def filter_stream(
109
+ self, stream: "TokenStream"
110
+ ) -> t.Union["TokenStream", t.Iterable["Token"]]:
111
+ """It's passed a :class:`~jinja2.lexer.TokenStream` that can be used
112
+ to filter tokens returned. This method has to return an iterable of
113
+ :class:`~jinja2.lexer.Token`\\s, but it doesn't have to return a
114
+ :class:`~jinja2.lexer.TokenStream`.
115
+ """
116
+ return stream
117
+
118
+ def parse(self, parser: "Parser") -> t.Union[nodes.Node, t.List[nodes.Node]]:
119
+ """If any of the :attr:`tags` matched this method is called with the
120
+ parser as first argument. The token the parser stream is pointing at
121
+ is the name token that matched. This method has to return one or a
122
+ list of multiple nodes.
123
+ """
124
+ raise NotImplementedError()
125
+
126
+ def attr(
127
+ self, name: str, lineno: t.Optional[int] = None
128
+ ) -> nodes.ExtensionAttribute:
129
+ """Return an attribute node for the current extension. This is useful
130
+ to pass constants on extensions to generated template code.
131
+
132
+ ::
133
+
134
+ self.attr('_my_attribute', lineno=lineno)
135
+ """
136
+ return nodes.ExtensionAttribute(self.identifier, name, lineno=lineno)
137
+
138
+ def call_method(
139
+ self,
140
+ name: str,
141
+ args: t.Optional[t.List[nodes.Expr]] = None,
142
+ kwargs: t.Optional[t.List[nodes.Keyword]] = None,
143
+ dyn_args: t.Optional[nodes.Expr] = None,
144
+ dyn_kwargs: t.Optional[nodes.Expr] = None,
145
+ lineno: t.Optional[int] = None,
146
+ ) -> nodes.Call:
147
+ """Call a method of the extension. This is a shortcut for
148
+ :meth:`attr` + :class:`jinja2.nodes.Call`.
149
+ """
150
+ if args is None:
151
+ args = []
152
+ if kwargs is None:
153
+ kwargs = []
154
+ return nodes.Call(
155
+ self.attr(name, lineno=lineno),
156
+ args,
157
+ kwargs,
158
+ dyn_args,
159
+ dyn_kwargs,
160
+ lineno=lineno,
161
+ )
162
+
163
+
164
+ @pass_context
165
+ def _gettext_alias(
166
+ __context: Context, *args: t.Any, **kwargs: t.Any
167
+ ) -> t.Union[t.Any, Undefined]:
168
+ return __context.call(__context.resolve("gettext"), *args, **kwargs)
169
+
170
+
171
+ def _make_new_gettext(func: t.Callable[[str], str]) -> t.Callable[..., str]:
172
+ @pass_context
173
+ def gettext(__context: Context, __string: str, **variables: t.Any) -> str:
174
+ rv = __context.call(func, __string)
175
+ if __context.eval_ctx.autoescape:
176
+ rv = Markup(rv)
177
+ # Always treat as a format string, even if there are no
178
+ # variables. This makes translation strings more consistent
179
+ # and predictable. This requires escaping
180
+ return rv % variables # type: ignore
181
+
182
+ return gettext
183
+
184
+
185
+ def _make_new_ngettext(func: t.Callable[[str, str, int], str]) -> t.Callable[..., str]:
186
+ @pass_context
187
+ def ngettext(
188
+ __context: Context,
189
+ __singular: str,
190
+ __plural: str,
191
+ __num: int,
192
+ **variables: t.Any,
193
+ ) -> str:
194
+ variables.setdefault("num", __num)
195
+ rv = __context.call(func, __singular, __plural, __num)
196
+ if __context.eval_ctx.autoescape:
197
+ rv = Markup(rv)
198
+ # Always treat as a format string, see gettext comment above.
199
+ return rv % variables # type: ignore
200
+
201
+ return ngettext
202
+
203
+
204
+ def _make_new_pgettext(func: t.Callable[[str, str], str]) -> t.Callable[..., str]:
205
+ @pass_context
206
+ def pgettext(
207
+ __context: Context, __string_ctx: str, __string: str, **variables: t.Any
208
+ ) -> str:
209
+ variables.setdefault("context", __string_ctx)
210
+ rv = __context.call(func, __string_ctx, __string)
211
+
212
+ if __context.eval_ctx.autoescape:
213
+ rv = Markup(rv)
214
+
215
+ # Always treat as a format string, see gettext comment above.
216
+ return rv % variables # type: ignore
217
+
218
+ return pgettext
219
+
220
+
221
+ def _make_new_npgettext(
222
+ func: t.Callable[[str, str, str, int], str],
223
+ ) -> t.Callable[..., str]:
224
+ @pass_context
225
+ def npgettext(
226
+ __context: Context,
227
+ __string_ctx: str,
228
+ __singular: str,
229
+ __plural: str,
230
+ __num: int,
231
+ **variables: t.Any,
232
+ ) -> str:
233
+ variables.setdefault("context", __string_ctx)
234
+ variables.setdefault("num", __num)
235
+ rv = __context.call(func, __string_ctx, __singular, __plural, __num)
236
+
237
+ if __context.eval_ctx.autoescape:
238
+ rv = Markup(rv)
239
+
240
+ # Always treat as a format string, see gettext comment above.
241
+ return rv % variables # type: ignore
242
+
243
+ return npgettext
244
+
245
+
246
+ class InternationalizationExtension(Extension):
247
+ """This extension adds gettext support to Jinja."""
248
+
249
+ tags = {"trans"}
250
+
251
+ # TODO: the i18n extension is currently reevaluating values in a few
252
+ # situations. Take this example:
253
+ # {% trans count=something() %}{{ count }} foo{% pluralize
254
+ # %}{{ count }} fooss{% endtrans %}
255
+ # something is called twice here. One time for the gettext value and
256
+ # the other time for the n-parameter of the ngettext function.
257
+
258
+ def __init__(self, environment: Environment) -> None:
259
+ super().__init__(environment)
260
+ environment.globals["_"] = _gettext_alias
261
+ environment.extend(
262
+ install_gettext_translations=self._install,
263
+ install_null_translations=self._install_null,
264
+ install_gettext_callables=self._install_callables,
265
+ uninstall_gettext_translations=self._uninstall,
266
+ extract_translations=self._extract,
267
+ newstyle_gettext=False,
268
+ )
269
+
270
+ def _install(
271
+ self, translations: "_SupportedTranslations", newstyle: t.Optional[bool] = None
272
+ ) -> None:
273
+ # ugettext and ungettext are preferred in case the I18N library
274
+ # is providing compatibility with older Python versions.
275
+ gettext = getattr(translations, "ugettext", None)
276
+ if gettext is None:
277
+ gettext = translations.gettext
278
+ ngettext = getattr(translations, "ungettext", None)
279
+ if ngettext is None:
280
+ ngettext = translations.ngettext
281
+
282
+ pgettext = getattr(translations, "pgettext", None)
283
+ npgettext = getattr(translations, "npgettext", None)
284
+ self._install_callables(
285
+ gettext, ngettext, newstyle=newstyle, pgettext=pgettext, npgettext=npgettext
286
+ )
287
+
288
+ def _install_null(self, newstyle: t.Optional[bool] = None) -> None:
289
+ import gettext
290
+
291
+ translations = gettext.NullTranslations()
292
+
293
+ if hasattr(translations, "pgettext"):
294
+ # Python < 3.8
295
+ pgettext = translations.pgettext
296
+ else:
297
+
298
+ def pgettext(c: str, s: str) -> str: # type: ignore[misc]
299
+ return s
300
+
301
+ if hasattr(translations, "npgettext"):
302
+ npgettext = translations.npgettext
303
+ else:
304
+
305
+ def npgettext(c: str, s: str, p: str, n: int) -> str: # type: ignore[misc]
306
+ return s if n == 1 else p
307
+
308
+ self._install_callables(
309
+ gettext=translations.gettext,
310
+ ngettext=translations.ngettext,
311
+ newstyle=newstyle,
312
+ pgettext=pgettext,
313
+ npgettext=npgettext,
314
+ )
315
+
316
+ def _install_callables(
317
+ self,
318
+ gettext: t.Callable[[str], str],
319
+ ngettext: t.Callable[[str, str, int], str],
320
+ newstyle: t.Optional[bool] = None,
321
+ pgettext: t.Optional[t.Callable[[str, str], str]] = None,
322
+ npgettext: t.Optional[t.Callable[[str, str, str, int], str]] = None,
323
+ ) -> None:
324
+ if newstyle is not None:
325
+ self.environment.newstyle_gettext = newstyle # type: ignore
326
+ if self.environment.newstyle_gettext: # type: ignore
327
+ gettext = _make_new_gettext(gettext)
328
+ ngettext = _make_new_ngettext(ngettext)
329
+
330
+ if pgettext is not None:
331
+ pgettext = _make_new_pgettext(pgettext)
332
+
333
+ if npgettext is not None:
334
+ npgettext = _make_new_npgettext(npgettext)
335
+
336
+ self.environment.globals.update(
337
+ gettext=gettext, ngettext=ngettext, pgettext=pgettext, npgettext=npgettext
338
+ )
339
+
340
+ def _uninstall(self, translations: "_SupportedTranslations") -> None:
341
+ for key in ("gettext", "ngettext", "pgettext", "npgettext"):
342
+ self.environment.globals.pop(key, None)
343
+
344
+ def _extract(
345
+ self,
346
+ source: t.Union[str, nodes.Template],
347
+ gettext_functions: t.Sequence[str] = GETTEXT_FUNCTIONS,
348
+ ) -> t.Iterator[
349
+ t.Tuple[int, str, t.Union[t.Optional[str], t.Tuple[t.Optional[str], ...]]]
350
+ ]:
351
+ if isinstance(source, str):
352
+ source = self.environment.parse(source)
353
+ return extract_from_ast(source, gettext_functions)
354
+
355
+ def parse(self, parser: "Parser") -> t.Union[nodes.Node, t.List[nodes.Node]]:
356
+ """Parse a translatable tag."""
357
+ lineno = next(parser.stream).lineno
358
+
359
+ context = None
360
+ context_token = parser.stream.next_if("string")
361
+
362
+ if context_token is not None:
363
+ context = context_token.value
364
+
365
+ # find all the variables referenced. Additionally a variable can be
366
+ # defined in the body of the trans block too, but this is checked at
367
+ # a later state.
368
+ plural_expr: t.Optional[nodes.Expr] = None
369
+ plural_expr_assignment: t.Optional[nodes.Assign] = None
370
+ num_called_num = False
371
+ variables: t.Dict[str, nodes.Expr] = {}
372
+ trimmed = None
373
+ while parser.stream.current.type != "block_end":
374
+ if variables:
375
+ parser.stream.expect("comma")
376
+
377
+ # skip colon for python compatibility
378
+ if parser.stream.skip_if("colon"):
379
+ break
380
+
381
+ token = parser.stream.expect("name")
382
+ if token.value in variables:
383
+ parser.fail(
384
+ f"translatable variable {token.value!r} defined twice.",
385
+ token.lineno,
386
+ exc=TemplateAssertionError,
387
+ )
388
+
389
+ # expressions
390
+ if parser.stream.current.type == "assign":
391
+ next(parser.stream)
392
+ variables[token.value] = var = parser.parse_expression()
393
+ elif trimmed is None and token.value in ("trimmed", "notrimmed"):
394
+ trimmed = token.value == "trimmed"
395
+ continue
396
+ else:
397
+ variables[token.value] = var = nodes.Name(token.value, "load")
398
+
399
+ if plural_expr is None:
400
+ if isinstance(var, nodes.Call):
401
+ plural_expr = nodes.Name("_trans", "load")
402
+ variables[token.value] = plural_expr
403
+ plural_expr_assignment = nodes.Assign(
404
+ nodes.Name("_trans", "store"), var
405
+ )
406
+ else:
407
+ plural_expr = var
408
+ num_called_num = token.value == "num"
409
+
410
+ parser.stream.expect("block_end")
411
+
412
+ plural = None
413
+ have_plural = False
414
+ referenced = set()
415
+
416
+ # now parse until endtrans or pluralize
417
+ singular_names, singular = self._parse_block(parser, True)
418
+ if singular_names:
419
+ referenced.update(singular_names)
420
+ if plural_expr is None:
421
+ plural_expr = nodes.Name(singular_names[0], "load")
422
+ num_called_num = singular_names[0] == "num"
423
+
424
+ # if we have a pluralize block, we parse that too
425
+ if parser.stream.current.test("name:pluralize"):
426
+ have_plural = True
427
+ next(parser.stream)
428
+ if parser.stream.current.type != "block_end":
429
+ token = parser.stream.expect("name")
430
+ if token.value not in variables:
431
+ parser.fail(
432
+ f"unknown variable {token.value!r} for pluralization",
433
+ token.lineno,
434
+ exc=TemplateAssertionError,
435
+ )
436
+ plural_expr = variables[token.value]
437
+ num_called_num = token.value == "num"
438
+ parser.stream.expect("block_end")
439
+ plural_names, plural = self._parse_block(parser, False)
440
+ next(parser.stream)
441
+ referenced.update(plural_names)
442
+ else:
443
+ next(parser.stream)
444
+
445
+ # register free names as simple name expressions
446
+ for name in referenced:
447
+ if name not in variables:
448
+ variables[name] = nodes.Name(name, "load")
449
+
450
+ if not have_plural:
451
+ plural_expr = None
452
+ elif plural_expr is None:
453
+ parser.fail("pluralize without variables", lineno)
454
+
455
+ if trimmed is None:
456
+ trimmed = self.environment.policies["ext.i18n.trimmed"]
457
+ if trimmed:
458
+ singular = self._trim_whitespace(singular)
459
+ if plural:
460
+ plural = self._trim_whitespace(plural)
461
+
462
+ node = self._make_node(
463
+ singular,
464
+ plural,
465
+ context,
466
+ variables,
467
+ plural_expr,
468
+ bool(referenced),
469
+ num_called_num and have_plural,
470
+ )
471
+ node.set_lineno(lineno)
472
+ if plural_expr_assignment is not None:
473
+ return [plural_expr_assignment, node]
474
+ else:
475
+ return node
476
+
477
+ def _trim_whitespace(self, string: str, _ws_re: t.Pattern[str] = _ws_re) -> str:
478
+ return _ws_re.sub(" ", string.strip())
479
+
480
+ def _parse_block(
481
+ self, parser: "Parser", allow_pluralize: bool
482
+ ) -> t.Tuple[t.List[str], str]:
483
+ """Parse until the next block tag with a given name."""
484
+ referenced = []
485
+ buf = []
486
+
487
+ while True:
488
+ if parser.stream.current.type == "data":
489
+ buf.append(parser.stream.current.value.replace("%", "%%"))
490
+ next(parser.stream)
491
+ elif parser.stream.current.type == "variable_begin":
492
+ next(parser.stream)
493
+ name = parser.stream.expect("name").value
494
+ referenced.append(name)
495
+ buf.append(f"%({name})s")
496
+ parser.stream.expect("variable_end")
497
+ elif parser.stream.current.type == "block_begin":
498
+ next(parser.stream)
499
+ block_name = (
500
+ parser.stream.current.value
501
+ if parser.stream.current.type == "name"
502
+ else None
503
+ )
504
+ if block_name == "endtrans":
505
+ break
506
+ elif block_name == "pluralize":
507
+ if allow_pluralize:
508
+ break
509
+ parser.fail(
510
+ "a translatable section can have only one pluralize section"
511
+ )
512
+ elif block_name == "trans":
513
+ parser.fail(
514
+ "trans blocks can't be nested; did you mean `endtrans`?"
515
+ )
516
+ parser.fail(
517
+ f"control structures in translatable sections are not allowed; "
518
+ f"saw `{block_name}`"
519
+ )
520
+ elif parser.stream.eos:
521
+ parser.fail("unclosed translation block")
522
+ else:
523
+ raise RuntimeError("internal parser error")
524
+
525
+ return referenced, concat(buf)
526
+
527
+ def _make_node(
528
+ self,
529
+ singular: str,
530
+ plural: t.Optional[str],
531
+ context: t.Optional[str],
532
+ variables: t.Dict[str, nodes.Expr],
533
+ plural_expr: t.Optional[nodes.Expr],
534
+ vars_referenced: bool,
535
+ num_called_num: bool,
536
+ ) -> nodes.Output:
537
+ """Generates a useful node from the data provided."""
538
+ newstyle = self.environment.newstyle_gettext # type: ignore
539
+ node: nodes.Expr
540
+
541
+ # no variables referenced? no need to escape for old style
542
+ # gettext invocations only if there are vars.
543
+ if not vars_referenced and not newstyle:
544
+ singular = singular.replace("%%", "%")
545
+ if plural:
546
+ plural = plural.replace("%%", "%")
547
+
548
+ func_name = "gettext"
549
+ func_args: t.List[nodes.Expr] = [nodes.Const(singular)]
550
+
551
+ if context is not None:
552
+ func_args.insert(0, nodes.Const(context))
553
+ func_name = f"p{func_name}"
554
+
555
+ if plural_expr is not None:
556
+ func_name = f"n{func_name}"
557
+ func_args.extend((nodes.Const(plural), plural_expr))
558
+
559
+ node = nodes.Call(nodes.Name(func_name, "load"), func_args, [], None, None)
560
+
561
+ # in case newstyle gettext is used, the method is powerful
562
+ # enough to handle the variable expansion and autoescape
563
+ # handling itself
564
+ if newstyle:
565
+ for key, value in variables.items():
566
+ # the function adds that later anyways in case num was
567
+ # called num, so just skip it.
568
+ if num_called_num and key == "num":
569
+ continue
570
+ node.kwargs.append(nodes.Keyword(key, value))
571
+
572
+ # otherwise do that here
573
+ else:
574
+ # mark the return value as safe if we are in an
575
+ # environment with autoescaping turned on
576
+ node = nodes.MarkSafeIfAutoescape(node)
577
+ if variables:
578
+ node = nodes.Mod(
579
+ node,
580
+ nodes.Dict(
581
+ [
582
+ nodes.Pair(nodes.Const(key), value)
583
+ for key, value in variables.items()
584
+ ]
585
+ ),
586
+ )
587
+ return nodes.Output([node])
588
+
589
+
590
+ class ExprStmtExtension(Extension):
591
+ """Adds a `do` tag to Jinja that works like the print statement just
592
+ that it doesn't print the return value.
593
+ """
594
+
595
+ tags = {"do"}
596
+
597
+ def parse(self, parser: "Parser") -> nodes.ExprStmt:
598
+ node = nodes.ExprStmt(lineno=next(parser.stream).lineno)
599
+ node.node = parser.parse_tuple()
600
+ return node
601
+
602
+
603
+ class LoopControlExtension(Extension):
604
+ """Adds break and continue to the template engine."""
605
+
606
+ tags = {"break", "continue"}
607
+
608
+ def parse(self, parser: "Parser") -> t.Union[nodes.Break, nodes.Continue]:
609
+ token = next(parser.stream)
610
+ if token.value == "break":
611
+ return nodes.Break(lineno=token.lineno)
612
+ return nodes.Continue(lineno=token.lineno)
613
+
614
+
615
+ class DebugExtension(Extension):
616
+ """A ``{% debug %}`` tag that dumps the available variables,
617
+ filters, and tests.
618
+
619
+ .. code-block:: html+jinja
620
+
621
+ <pre>{% debug %}</pre>
622
+
623
+ .. code-block:: text
624
+
625
+ {'context': {'cycler': <class 'jinja2.utils.Cycler'>,
626
+ ...,
627
+ 'namespace': <class 'jinja2.utils.Namespace'>},
628
+ 'filters': ['abs', 'attr', 'batch', 'capitalize', 'center', 'count', 'd',
629
+ ..., 'urlencode', 'urlize', 'wordcount', 'wordwrap', 'xmlattr'],
630
+ 'tests': ['!=', '<', '<=', '==', '>', '>=', 'callable', 'defined',
631
+ ..., 'odd', 'sameas', 'sequence', 'string', 'undefined', 'upper']}
632
+
633
+ .. versionadded:: 2.11.0
634
+ """
635
+
636
+ tags = {"debug"}
637
+
638
+ def parse(self, parser: "Parser") -> nodes.Output:
639
+ lineno = parser.stream.expect("name:debug").lineno
640
+ context = nodes.ContextReference()
641
+ result = self.call_method("_render", [context], lineno=lineno)
642
+ return nodes.Output([result], lineno=lineno)
643
+
644
+ def _render(self, context: Context) -> str:
645
+ result = {
646
+ "context": context.get_all(),
647
+ "filters": sorted(self.environment.filters.keys()),
648
+ "tests": sorted(self.environment.tests.keys()),
649
+ }
650
+
651
+ # Set the depth since the intent is to show the top few names.
652
+ return pprint.pformat(result, depth=3, compact=True)
653
+
654
+
655
+ def extract_from_ast(
656
+ ast: nodes.Template,
657
+ gettext_functions: t.Sequence[str] = GETTEXT_FUNCTIONS,
658
+ babel_style: bool = True,
659
+ ) -> t.Iterator[
660
+ t.Tuple[int, str, t.Union[t.Optional[str], t.Tuple[t.Optional[str], ...]]]
661
+ ]:
662
+ """Extract localizable strings from the given template node. Per
663
+ default this function returns matches in babel style that means non string
664
+ parameters as well as keyword arguments are returned as `None`. This
665
+ allows Babel to figure out what you really meant if you are using
666
+ gettext functions that allow keyword arguments for placeholder expansion.
667
+ If you don't want that behavior set the `babel_style` parameter to `False`
668
+ which causes only strings to be returned and parameters are always stored
669
+ in tuples. As a consequence invalid gettext calls (calls without a single
670
+ string parameter or string parameters after non-string parameters) are
671
+ skipped.
672
+
673
+ This example explains the behavior:
674
+
675
+ >>> from jinja2 import Environment
676
+ >>> env = Environment()
677
+ >>> node = env.parse('{{ (_("foo"), _(), ngettext("foo", "bar", 42)) }}')
678
+ >>> list(extract_from_ast(node))
679
+ [(1, '_', 'foo'), (1, '_', ()), (1, 'ngettext', ('foo', 'bar', None))]
680
+ >>> list(extract_from_ast(node, babel_style=False))
681
+ [(1, '_', ('foo',)), (1, 'ngettext', ('foo', 'bar'))]
682
+
683
+ For every string found this function yields a ``(lineno, function,
684
+ message)`` tuple, where:
685
+
686
+ * ``lineno`` is the number of the line on which the string was found,
687
+ * ``function`` is the name of the ``gettext`` function used (if the
688
+ string was extracted from embedded Python code), and
689
+ * ``message`` is the string, or a tuple of strings for functions
690
+ with multiple string arguments.
691
+
692
+ This extraction function operates on the AST and is because of that unable
693
+ to extract any comments. For comment support you have to use the babel
694
+ extraction interface or extract comments yourself.
695
+ """
696
+ out: t.Union[t.Optional[str], t.Tuple[t.Optional[str], ...]]
697
+
698
+ for node in ast.find_all(nodes.Call):
699
+ if (
700
+ not isinstance(node.node, nodes.Name)
701
+ or node.node.name not in gettext_functions
702
+ ):
703
+ continue
704
+
705
+ strings: t.List[t.Optional[str]] = []
706
+
707
+ for arg in node.args:
708
+ if isinstance(arg, nodes.Const) and isinstance(arg.value, str):
709
+ strings.append(arg.value)
710
+ else:
711
+ strings.append(None)
712
+
713
+ for _ in node.kwargs:
714
+ strings.append(None)
715
+ if node.dyn_args is not None:
716
+ strings.append(None)
717
+ if node.dyn_kwargs is not None:
718
+ strings.append(None)
719
+
720
+ if not babel_style:
721
+ out = tuple(x for x in strings if x is not None)
722
+
723
+ if not out:
724
+ continue
725
+ else:
726
+ if len(strings) == 1:
727
+ out = strings[0]
728
+ else:
729
+ out = tuple(strings)
730
+
731
+ yield node.lineno, node.node.name, out
732
+
733
+
734
+ class _CommentFinder:
735
+ """Helper class to find comments in a token stream. Can only
736
+ find comments for gettext calls forwards. Once the comment
737
+ from line 4 is found, a comment for line 1 will not return a
738
+ usable value.
739
+ """
740
+
741
+ def __init__(
742
+ self, tokens: t.Sequence[t.Tuple[int, str, str]], comment_tags: t.Sequence[str]
743
+ ) -> None:
744
+ self.tokens = tokens
745
+ self.comment_tags = comment_tags
746
+ self.offset = 0
747
+ self.last_lineno = 0
748
+
749
+ def find_backwards(self, offset: int) -> t.List[str]:
750
+ try:
751
+ for _, token_type, token_value in reversed(
752
+ self.tokens[self.offset : offset]
753
+ ):
754
+ if token_type in ("comment", "linecomment"):
755
+ try:
756
+ prefix, comment = token_value.split(None, 1)
757
+ except ValueError:
758
+ continue
759
+ if prefix in self.comment_tags:
760
+ return [comment.rstrip()]
761
+ return []
762
+ finally:
763
+ self.offset = offset
764
+
765
+ def find_comments(self, lineno: int) -> t.List[str]:
766
+ if not self.comment_tags or self.last_lineno > lineno:
767
+ return []
768
+ for idx, (token_lineno, _, _) in enumerate(self.tokens[self.offset :]):
769
+ if token_lineno > lineno:
770
+ return self.find_backwards(self.offset + idx)
771
+ return self.find_backwards(len(self.tokens))
772
+
773
+
774
+ def babel_extract(
775
+ fileobj: t.BinaryIO,
776
+ keywords: t.Sequence[str],
777
+ comment_tags: t.Sequence[str],
778
+ options: t.Dict[str, t.Any],
779
+ ) -> t.Iterator[
780
+ t.Tuple[
781
+ int, str, t.Union[t.Optional[str], t.Tuple[t.Optional[str], ...]], t.List[str]
782
+ ]
783
+ ]:
784
+ """Babel extraction method for Jinja templates.
785
+
786
+ .. versionchanged:: 2.3
787
+ Basic support for translation comments was added. If `comment_tags`
788
+ is now set to a list of keywords for extraction, the extractor will
789
+ try to find the best preceding comment that begins with one of the
790
+ keywords. For best results, make sure to not have more than one
791
+ gettext call in one line of code and the matching comment in the
792
+ same line or the line before.
793
+
794
+ .. versionchanged:: 2.5.1
795
+ The `newstyle_gettext` flag can be set to `True` to enable newstyle
796
+ gettext calls.
797
+
798
+ .. versionchanged:: 2.7
799
+ A `silent` option can now be provided. If set to `False` template
800
+ syntax errors are propagated instead of being ignored.
801
+
802
+ :param fileobj: the file-like object the messages should be extracted from
803
+ :param keywords: a list of keywords (i.e. function names) that should be
804
+ recognized as translation functions
805
+ :param comment_tags: a list of translator tags to search for and include
806
+ in the results.
807
+ :param options: a dictionary of additional options (optional)
808
+ :return: an iterator over ``(lineno, funcname, message, comments)`` tuples.
809
+ (comments will be empty currently)
810
+ """
811
+ extensions: t.Dict[t.Type[Extension], None] = {}
812
+
813
+ for extension_name in options.get("extensions", "").split(","):
814
+ extension_name = extension_name.strip()
815
+
816
+ if not extension_name:
817
+ continue
818
+
819
+ extensions[import_string(extension_name)] = None
820
+
821
+ if InternationalizationExtension not in extensions:
822
+ extensions[InternationalizationExtension] = None
823
+
824
+ def getbool(options: t.Mapping[str, str], key: str, default: bool = False) -> bool:
825
+ return options.get(key, str(default)).lower() in {"1", "on", "yes", "true"}
826
+
827
+ silent = getbool(options, "silent", True)
828
+ environment = Environment(
829
+ options.get("block_start_string", defaults.BLOCK_START_STRING),
830
+ options.get("block_end_string", defaults.BLOCK_END_STRING),
831
+ options.get("variable_start_string", defaults.VARIABLE_START_STRING),
832
+ options.get("variable_end_string", defaults.VARIABLE_END_STRING),
833
+ options.get("comment_start_string", defaults.COMMENT_START_STRING),
834
+ options.get("comment_end_string", defaults.COMMENT_END_STRING),
835
+ options.get("line_statement_prefix") or defaults.LINE_STATEMENT_PREFIX,
836
+ options.get("line_comment_prefix") or defaults.LINE_COMMENT_PREFIX,
837
+ getbool(options, "trim_blocks", defaults.TRIM_BLOCKS),
838
+ getbool(options, "lstrip_blocks", defaults.LSTRIP_BLOCKS),
839
+ defaults.NEWLINE_SEQUENCE,
840
+ getbool(options, "keep_trailing_newline", defaults.KEEP_TRAILING_NEWLINE),
841
+ tuple(extensions),
842
+ cache_size=0,
843
+ auto_reload=False,
844
+ )
845
+
846
+ if getbool(options, "trimmed"):
847
+ environment.policies["ext.i18n.trimmed"] = True
848
+ if getbool(options, "newstyle_gettext"):
849
+ environment.newstyle_gettext = True # type: ignore
850
+
851
+ source = fileobj.read().decode(options.get("encoding", "utf-8"))
852
+ try:
853
+ node = environment.parse(source)
854
+ tokens = list(environment.lex(environment.preprocess(source)))
855
+ except TemplateSyntaxError:
856
+ if not silent:
857
+ raise
858
+ # skip templates with syntax errors
859
+ return
860
+
861
+ finder = _CommentFinder(tokens, comment_tags)
862
+ for lineno, func, message in extract_from_ast(node, keywords):
863
+ yield lineno, func, message, finder.find_comments(lineno)
864
+
865
+
866
+ #: nicer import names
867
+ i18n = InternationalizationExtension
868
+ do = ExprStmtExtension
869
+ loopcontrols = LoopControlExtension
870
+ debug = DebugExtension
janus/lib/python3.10/site-packages/jinja2/idtracking.py ADDED
@@ -0,0 +1,318 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import typing as t
2
+
3
+ from . import nodes
4
+ from .visitor import NodeVisitor
5
+
6
+ if t.TYPE_CHECKING:
7
+ import typing_extensions as te
8
+
9
+ VAR_LOAD_PARAMETER = "param"
10
+ VAR_LOAD_RESOLVE = "resolve"
11
+ VAR_LOAD_ALIAS = "alias"
12
+ VAR_LOAD_UNDEFINED = "undefined"
13
+
14
+
15
+ def find_symbols(
16
+ nodes: t.Iterable[nodes.Node], parent_symbols: t.Optional["Symbols"] = None
17
+ ) -> "Symbols":
18
+ sym = Symbols(parent=parent_symbols)
19
+ visitor = FrameSymbolVisitor(sym)
20
+ for node in nodes:
21
+ visitor.visit(node)
22
+ return sym
23
+
24
+
25
+ def symbols_for_node(
26
+ node: nodes.Node, parent_symbols: t.Optional["Symbols"] = None
27
+ ) -> "Symbols":
28
+ sym = Symbols(parent=parent_symbols)
29
+ sym.analyze_node(node)
30
+ return sym
31
+
32
+
33
+ class Symbols:
34
+ def __init__(
35
+ self, parent: t.Optional["Symbols"] = None, level: t.Optional[int] = None
36
+ ) -> None:
37
+ if level is None:
38
+ if parent is None:
39
+ level = 0
40
+ else:
41
+ level = parent.level + 1
42
+
43
+ self.level: int = level
44
+ self.parent = parent
45
+ self.refs: t.Dict[str, str] = {}
46
+ self.loads: t.Dict[str, t.Any] = {}
47
+ self.stores: t.Set[str] = set()
48
+
49
+ def analyze_node(self, node: nodes.Node, **kwargs: t.Any) -> None:
50
+ visitor = RootVisitor(self)
51
+ visitor.visit(node, **kwargs)
52
+
53
+ def _define_ref(
54
+ self, name: str, load: t.Optional[t.Tuple[str, t.Optional[str]]] = None
55
+ ) -> str:
56
+ ident = f"l_{self.level}_{name}"
57
+ self.refs[name] = ident
58
+ if load is not None:
59
+ self.loads[ident] = load
60
+ return ident
61
+
62
+ def find_load(self, target: str) -> t.Optional[t.Any]:
63
+ if target in self.loads:
64
+ return self.loads[target]
65
+
66
+ if self.parent is not None:
67
+ return self.parent.find_load(target)
68
+
69
+ return None
70
+
71
+ def find_ref(self, name: str) -> t.Optional[str]:
72
+ if name in self.refs:
73
+ return self.refs[name]
74
+
75
+ if self.parent is not None:
76
+ return self.parent.find_ref(name)
77
+
78
+ return None
79
+
80
+ def ref(self, name: str) -> str:
81
+ rv = self.find_ref(name)
82
+ if rv is None:
83
+ raise AssertionError(
84
+ "Tried to resolve a name to a reference that was"
85
+ f" unknown to the frame ({name!r})"
86
+ )
87
+ return rv
88
+
89
+ def copy(self) -> "te.Self":
90
+ rv = object.__new__(self.__class__)
91
+ rv.__dict__.update(self.__dict__)
92
+ rv.refs = self.refs.copy()
93
+ rv.loads = self.loads.copy()
94
+ rv.stores = self.stores.copy()
95
+ return rv
96
+
97
+ def store(self, name: str) -> None:
98
+ self.stores.add(name)
99
+
100
+ # If we have not see the name referenced yet, we need to figure
101
+ # out what to set it to.
102
+ if name not in self.refs:
103
+ # If there is a parent scope we check if the name has a
104
+ # reference there. If it does it means we might have to alias
105
+ # to a variable there.
106
+ if self.parent is not None:
107
+ outer_ref = self.parent.find_ref(name)
108
+ if outer_ref is not None:
109
+ self._define_ref(name, load=(VAR_LOAD_ALIAS, outer_ref))
110
+ return
111
+
112
+ # Otherwise we can just set it to undefined.
113
+ self._define_ref(name, load=(VAR_LOAD_UNDEFINED, None))
114
+
115
+ def declare_parameter(self, name: str) -> str:
116
+ self.stores.add(name)
117
+ return self._define_ref(name, load=(VAR_LOAD_PARAMETER, None))
118
+
119
+ def load(self, name: str) -> None:
120
+ if self.find_ref(name) is None:
121
+ self._define_ref(name, load=(VAR_LOAD_RESOLVE, name))
122
+
123
+ def branch_update(self, branch_symbols: t.Sequence["Symbols"]) -> None:
124
+ stores: t.Set[str] = set()
125
+
126
+ for branch in branch_symbols:
127
+ stores.update(branch.stores)
128
+
129
+ stores.difference_update(self.stores)
130
+
131
+ for sym in branch_symbols:
132
+ self.refs.update(sym.refs)
133
+ self.loads.update(sym.loads)
134
+ self.stores.update(sym.stores)
135
+
136
+ for name in stores:
137
+ target = self.find_ref(name)
138
+ assert target is not None, "should not happen"
139
+
140
+ if self.parent is not None:
141
+ outer_target = self.parent.find_ref(name)
142
+ if outer_target is not None:
143
+ self.loads[target] = (VAR_LOAD_ALIAS, outer_target)
144
+ continue
145
+ self.loads[target] = (VAR_LOAD_RESOLVE, name)
146
+
147
+ def dump_stores(self) -> t.Dict[str, str]:
148
+ rv: t.Dict[str, str] = {}
149
+ node: t.Optional[Symbols] = self
150
+
151
+ while node is not None:
152
+ for name in sorted(node.stores):
153
+ if name not in rv:
154
+ rv[name] = self.find_ref(name) # type: ignore
155
+
156
+ node = node.parent
157
+
158
+ return rv
159
+
160
+ def dump_param_targets(self) -> t.Set[str]:
161
+ rv = set()
162
+ node: t.Optional[Symbols] = self
163
+
164
+ while node is not None:
165
+ for target, (instr, _) in self.loads.items():
166
+ if instr == VAR_LOAD_PARAMETER:
167
+ rv.add(target)
168
+
169
+ node = node.parent
170
+
171
+ return rv
172
+
173
+
174
+ class RootVisitor(NodeVisitor):
175
+ def __init__(self, symbols: "Symbols") -> None:
176
+ self.sym_visitor = FrameSymbolVisitor(symbols)
177
+
178
+ def _simple_visit(self, node: nodes.Node, **kwargs: t.Any) -> None:
179
+ for child in node.iter_child_nodes():
180
+ self.sym_visitor.visit(child)
181
+
182
+ visit_Template = _simple_visit
183
+ visit_Block = _simple_visit
184
+ visit_Macro = _simple_visit
185
+ visit_FilterBlock = _simple_visit
186
+ visit_Scope = _simple_visit
187
+ visit_If = _simple_visit
188
+ visit_ScopedEvalContextModifier = _simple_visit
189
+
190
+ def visit_AssignBlock(self, node: nodes.AssignBlock, **kwargs: t.Any) -> None:
191
+ for child in node.body:
192
+ self.sym_visitor.visit(child)
193
+
194
+ def visit_CallBlock(self, node: nodes.CallBlock, **kwargs: t.Any) -> None:
195
+ for child in node.iter_child_nodes(exclude=("call",)):
196
+ self.sym_visitor.visit(child)
197
+
198
+ def visit_OverlayScope(self, node: nodes.OverlayScope, **kwargs: t.Any) -> None:
199
+ for child in node.body:
200
+ self.sym_visitor.visit(child)
201
+
202
+ def visit_For(
203
+ self, node: nodes.For, for_branch: str = "body", **kwargs: t.Any
204
+ ) -> None:
205
+ if for_branch == "body":
206
+ self.sym_visitor.visit(node.target, store_as_param=True)
207
+ branch = node.body
208
+ elif for_branch == "else":
209
+ branch = node.else_
210
+ elif for_branch == "test":
211
+ self.sym_visitor.visit(node.target, store_as_param=True)
212
+ if node.test is not None:
213
+ self.sym_visitor.visit(node.test)
214
+ return
215
+ else:
216
+ raise RuntimeError("Unknown for branch")
217
+
218
+ if branch:
219
+ for item in branch:
220
+ self.sym_visitor.visit(item)
221
+
222
+ def visit_With(self, node: nodes.With, **kwargs: t.Any) -> None:
223
+ for target in node.targets:
224
+ self.sym_visitor.visit(target)
225
+ for child in node.body:
226
+ self.sym_visitor.visit(child)
227
+
228
+ def generic_visit(self, node: nodes.Node, *args: t.Any, **kwargs: t.Any) -> None:
229
+ raise NotImplementedError(f"Cannot find symbols for {type(node).__name__!r}")
230
+
231
+
232
+ class FrameSymbolVisitor(NodeVisitor):
233
+ """A visitor for `Frame.inspect`."""
234
+
235
+ def __init__(self, symbols: "Symbols") -> None:
236
+ self.symbols = symbols
237
+
238
+ def visit_Name(
239
+ self, node: nodes.Name, store_as_param: bool = False, **kwargs: t.Any
240
+ ) -> None:
241
+ """All assignments to names go through this function."""
242
+ if store_as_param or node.ctx == "param":
243
+ self.symbols.declare_parameter(node.name)
244
+ elif node.ctx == "store":
245
+ self.symbols.store(node.name)
246
+ elif node.ctx == "load":
247
+ self.symbols.load(node.name)
248
+
249
+ def visit_NSRef(self, node: nodes.NSRef, **kwargs: t.Any) -> None:
250
+ self.symbols.load(node.name)
251
+
252
+ def visit_If(self, node: nodes.If, **kwargs: t.Any) -> None:
253
+ self.visit(node.test, **kwargs)
254
+ original_symbols = self.symbols
255
+
256
+ def inner_visit(nodes: t.Iterable[nodes.Node]) -> "Symbols":
257
+ self.symbols = rv = original_symbols.copy()
258
+
259
+ for subnode in nodes:
260
+ self.visit(subnode, **kwargs)
261
+
262
+ self.symbols = original_symbols
263
+ return rv
264
+
265
+ body_symbols = inner_visit(node.body)
266
+ elif_symbols = inner_visit(node.elif_)
267
+ else_symbols = inner_visit(node.else_ or ())
268
+ self.symbols.branch_update([body_symbols, elif_symbols, else_symbols])
269
+
270
+ def visit_Macro(self, node: nodes.Macro, **kwargs: t.Any) -> None:
271
+ self.symbols.store(node.name)
272
+
273
+ def visit_Import(self, node: nodes.Import, **kwargs: t.Any) -> None:
274
+ self.generic_visit(node, **kwargs)
275
+ self.symbols.store(node.target)
276
+
277
+ def visit_FromImport(self, node: nodes.FromImport, **kwargs: t.Any) -> None:
278
+ self.generic_visit(node, **kwargs)
279
+
280
+ for name in node.names:
281
+ if isinstance(name, tuple):
282
+ self.symbols.store(name[1])
283
+ else:
284
+ self.symbols.store(name)
285
+
286
+ def visit_Assign(self, node: nodes.Assign, **kwargs: t.Any) -> None:
287
+ """Visit assignments in the correct order."""
288
+ self.visit(node.node, **kwargs)
289
+ self.visit(node.target, **kwargs)
290
+
291
+ def visit_For(self, node: nodes.For, **kwargs: t.Any) -> None:
292
+ """Visiting stops at for blocks. However the block sequence
293
+ is visited as part of the outer scope.
294
+ """
295
+ self.visit(node.iter, **kwargs)
296
+
297
+ def visit_CallBlock(self, node: nodes.CallBlock, **kwargs: t.Any) -> None:
298
+ self.visit(node.call, **kwargs)
299
+
300
+ def visit_FilterBlock(self, node: nodes.FilterBlock, **kwargs: t.Any) -> None:
301
+ self.visit(node.filter, **kwargs)
302
+
303
+ def visit_With(self, node: nodes.With, **kwargs: t.Any) -> None:
304
+ for target in node.values:
305
+ self.visit(target)
306
+
307
+ def visit_AssignBlock(self, node: nodes.AssignBlock, **kwargs: t.Any) -> None:
308
+ """Stop visiting at block assigns."""
309
+ self.visit(node.target, **kwargs)
310
+
311
+ def visit_Scope(self, node: nodes.Scope, **kwargs: t.Any) -> None:
312
+ """Stop visiting at scopes."""
313
+
314
+ def visit_Block(self, node: nodes.Block, **kwargs: t.Any) -> None:
315
+ """Stop visiting at blocks."""
316
+
317
+ def visit_OverlayScope(self, node: nodes.OverlayScope, **kwargs: t.Any) -> None:
318
+ """Do not visit into overlay scopes."""
janus/lib/python3.10/site-packages/jinja2/lexer.py ADDED
@@ -0,0 +1,868 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """Implements a Jinja / Python combination lexer. The ``Lexer`` class
2
+ is used to do some preprocessing. It filters out invalid operators like
3
+ the bitshift operators we don't allow in templates. It separates
4
+ template code and python code in expressions.
5
+ """
6
+
7
+ import re
8
+ import typing as t
9
+ from ast import literal_eval
10
+ from collections import deque
11
+ from sys import intern
12
+
13
+ from ._identifier import pattern as name_re
14
+ from .exceptions import TemplateSyntaxError
15
+ from .utils import LRUCache
16
+
17
+ if t.TYPE_CHECKING:
18
+ import typing_extensions as te
19
+
20
+ from .environment import Environment
21
+
22
+ # cache for the lexers. Exists in order to be able to have multiple
23
+ # environments with the same lexer
24
+ _lexer_cache: t.MutableMapping[t.Tuple, "Lexer"] = LRUCache(50) # type: ignore
25
+
26
+ # static regular expressions
27
+ whitespace_re = re.compile(r"\s+")
28
+ newline_re = re.compile(r"(\r\n|\r|\n)")
29
+ string_re = re.compile(
30
+ r"('([^'\\]*(?:\\.[^'\\]*)*)'" r'|"([^"\\]*(?:\\.[^"\\]*)*)")', re.S
31
+ )
32
+ integer_re = re.compile(
33
+ r"""
34
+ (
35
+ 0b(_?[0-1])+ # binary
36
+ |
37
+ 0o(_?[0-7])+ # octal
38
+ |
39
+ 0x(_?[\da-f])+ # hex
40
+ |
41
+ [1-9](_?\d)* # decimal
42
+ |
43
+ 0(_?0)* # decimal zero
44
+ )
45
+ """,
46
+ re.IGNORECASE | re.VERBOSE,
47
+ )
48
+ float_re = re.compile(
49
+ r"""
50
+ (?<!\.) # doesn't start with a .
51
+ (\d+_)*\d+ # digits, possibly _ separated
52
+ (
53
+ (\.(\d+_)*\d+)? # optional fractional part
54
+ e[+\-]?(\d+_)*\d+ # exponent part
55
+ |
56
+ \.(\d+_)*\d+ # required fractional part
57
+ )
58
+ """,
59
+ re.IGNORECASE | re.VERBOSE,
60
+ )
61
+
62
+ # internal the tokens and keep references to them
63
+ TOKEN_ADD = intern("add")
64
+ TOKEN_ASSIGN = intern("assign")
65
+ TOKEN_COLON = intern("colon")
66
+ TOKEN_COMMA = intern("comma")
67
+ TOKEN_DIV = intern("div")
68
+ TOKEN_DOT = intern("dot")
69
+ TOKEN_EQ = intern("eq")
70
+ TOKEN_FLOORDIV = intern("floordiv")
71
+ TOKEN_GT = intern("gt")
72
+ TOKEN_GTEQ = intern("gteq")
73
+ TOKEN_LBRACE = intern("lbrace")
74
+ TOKEN_LBRACKET = intern("lbracket")
75
+ TOKEN_LPAREN = intern("lparen")
76
+ TOKEN_LT = intern("lt")
77
+ TOKEN_LTEQ = intern("lteq")
78
+ TOKEN_MOD = intern("mod")
79
+ TOKEN_MUL = intern("mul")
80
+ TOKEN_NE = intern("ne")
81
+ TOKEN_PIPE = intern("pipe")
82
+ TOKEN_POW = intern("pow")
83
+ TOKEN_RBRACE = intern("rbrace")
84
+ TOKEN_RBRACKET = intern("rbracket")
85
+ TOKEN_RPAREN = intern("rparen")
86
+ TOKEN_SEMICOLON = intern("semicolon")
87
+ TOKEN_SUB = intern("sub")
88
+ TOKEN_TILDE = intern("tilde")
89
+ TOKEN_WHITESPACE = intern("whitespace")
90
+ TOKEN_FLOAT = intern("float")
91
+ TOKEN_INTEGER = intern("integer")
92
+ TOKEN_NAME = intern("name")
93
+ TOKEN_STRING = intern("string")
94
+ TOKEN_OPERATOR = intern("operator")
95
+ TOKEN_BLOCK_BEGIN = intern("block_begin")
96
+ TOKEN_BLOCK_END = intern("block_end")
97
+ TOKEN_VARIABLE_BEGIN = intern("variable_begin")
98
+ TOKEN_VARIABLE_END = intern("variable_end")
99
+ TOKEN_RAW_BEGIN = intern("raw_begin")
100
+ TOKEN_RAW_END = intern("raw_end")
101
+ TOKEN_COMMENT_BEGIN = intern("comment_begin")
102
+ TOKEN_COMMENT_END = intern("comment_end")
103
+ TOKEN_COMMENT = intern("comment")
104
+ TOKEN_LINESTATEMENT_BEGIN = intern("linestatement_begin")
105
+ TOKEN_LINESTATEMENT_END = intern("linestatement_end")
106
+ TOKEN_LINECOMMENT_BEGIN = intern("linecomment_begin")
107
+ TOKEN_LINECOMMENT_END = intern("linecomment_end")
108
+ TOKEN_LINECOMMENT = intern("linecomment")
109
+ TOKEN_DATA = intern("data")
110
+ TOKEN_INITIAL = intern("initial")
111
+ TOKEN_EOF = intern("eof")
112
+
113
+ # bind operators to token types
114
+ operators = {
115
+ "+": TOKEN_ADD,
116
+ "-": TOKEN_SUB,
117
+ "/": TOKEN_DIV,
118
+ "//": TOKEN_FLOORDIV,
119
+ "*": TOKEN_MUL,
120
+ "%": TOKEN_MOD,
121
+ "**": TOKEN_POW,
122
+ "~": TOKEN_TILDE,
123
+ "[": TOKEN_LBRACKET,
124
+ "]": TOKEN_RBRACKET,
125
+ "(": TOKEN_LPAREN,
126
+ ")": TOKEN_RPAREN,
127
+ "{": TOKEN_LBRACE,
128
+ "}": TOKEN_RBRACE,
129
+ "==": TOKEN_EQ,
130
+ "!=": TOKEN_NE,
131
+ ">": TOKEN_GT,
132
+ ">=": TOKEN_GTEQ,
133
+ "<": TOKEN_LT,
134
+ "<=": TOKEN_LTEQ,
135
+ "=": TOKEN_ASSIGN,
136
+ ".": TOKEN_DOT,
137
+ ":": TOKEN_COLON,
138
+ "|": TOKEN_PIPE,
139
+ ",": TOKEN_COMMA,
140
+ ";": TOKEN_SEMICOLON,
141
+ }
142
+
143
+ reverse_operators = {v: k for k, v in operators.items()}
144
+ assert len(operators) == len(reverse_operators), "operators dropped"
145
+ operator_re = re.compile(
146
+ f"({'|'.join(re.escape(x) for x in sorted(operators, key=lambda x: -len(x)))})"
147
+ )
148
+
149
+ ignored_tokens = frozenset(
150
+ [
151
+ TOKEN_COMMENT_BEGIN,
152
+ TOKEN_COMMENT,
153
+ TOKEN_COMMENT_END,
154
+ TOKEN_WHITESPACE,
155
+ TOKEN_LINECOMMENT_BEGIN,
156
+ TOKEN_LINECOMMENT_END,
157
+ TOKEN_LINECOMMENT,
158
+ ]
159
+ )
160
+ ignore_if_empty = frozenset(
161
+ [TOKEN_WHITESPACE, TOKEN_DATA, TOKEN_COMMENT, TOKEN_LINECOMMENT]
162
+ )
163
+
164
+
165
+ def _describe_token_type(token_type: str) -> str:
166
+ if token_type in reverse_operators:
167
+ return reverse_operators[token_type]
168
+
169
+ return {
170
+ TOKEN_COMMENT_BEGIN: "begin of comment",
171
+ TOKEN_COMMENT_END: "end of comment",
172
+ TOKEN_COMMENT: "comment",
173
+ TOKEN_LINECOMMENT: "comment",
174
+ TOKEN_BLOCK_BEGIN: "begin of statement block",
175
+ TOKEN_BLOCK_END: "end of statement block",
176
+ TOKEN_VARIABLE_BEGIN: "begin of print statement",
177
+ TOKEN_VARIABLE_END: "end of print statement",
178
+ TOKEN_LINESTATEMENT_BEGIN: "begin of line statement",
179
+ TOKEN_LINESTATEMENT_END: "end of line statement",
180
+ TOKEN_DATA: "template data / text",
181
+ TOKEN_EOF: "end of template",
182
+ }.get(token_type, token_type)
183
+
184
+
185
+ def describe_token(token: "Token") -> str:
186
+ """Returns a description of the token."""
187
+ if token.type == TOKEN_NAME:
188
+ return token.value
189
+
190
+ return _describe_token_type(token.type)
191
+
192
+
193
+ def describe_token_expr(expr: str) -> str:
194
+ """Like `describe_token` but for token expressions."""
195
+ if ":" in expr:
196
+ type, value = expr.split(":", 1)
197
+
198
+ if type == TOKEN_NAME:
199
+ return value
200
+ else:
201
+ type = expr
202
+
203
+ return _describe_token_type(type)
204
+
205
+
206
+ def count_newlines(value: str) -> int:
207
+ """Count the number of newline characters in the string. This is
208
+ useful for extensions that filter a stream.
209
+ """
210
+ return len(newline_re.findall(value))
211
+
212
+
213
+ def compile_rules(environment: "Environment") -> t.List[t.Tuple[str, str]]:
214
+ """Compiles all the rules from the environment into a list of rules."""
215
+ e = re.escape
216
+ rules = [
217
+ (
218
+ len(environment.comment_start_string),
219
+ TOKEN_COMMENT_BEGIN,
220
+ e(environment.comment_start_string),
221
+ ),
222
+ (
223
+ len(environment.block_start_string),
224
+ TOKEN_BLOCK_BEGIN,
225
+ e(environment.block_start_string),
226
+ ),
227
+ (
228
+ len(environment.variable_start_string),
229
+ TOKEN_VARIABLE_BEGIN,
230
+ e(environment.variable_start_string),
231
+ ),
232
+ ]
233
+
234
+ if environment.line_statement_prefix is not None:
235
+ rules.append(
236
+ (
237
+ len(environment.line_statement_prefix),
238
+ TOKEN_LINESTATEMENT_BEGIN,
239
+ r"^[ \t\v]*" + e(environment.line_statement_prefix),
240
+ )
241
+ )
242
+ if environment.line_comment_prefix is not None:
243
+ rules.append(
244
+ (
245
+ len(environment.line_comment_prefix),
246
+ TOKEN_LINECOMMENT_BEGIN,
247
+ r"(?:^|(?<=\S))[^\S\r\n]*" + e(environment.line_comment_prefix),
248
+ )
249
+ )
250
+
251
+ return [x[1:] for x in sorted(rules, reverse=True)]
252
+
253
+
254
+ class Failure:
255
+ """Class that raises a `TemplateSyntaxError` if called.
256
+ Used by the `Lexer` to specify known errors.
257
+ """
258
+
259
+ def __init__(
260
+ self, message: str, cls: t.Type[TemplateSyntaxError] = TemplateSyntaxError
261
+ ) -> None:
262
+ self.message = message
263
+ self.error_class = cls
264
+
265
+ def __call__(self, lineno: int, filename: t.Optional[str]) -> "te.NoReturn":
266
+ raise self.error_class(self.message, lineno, filename)
267
+
268
+
269
+ class Token(t.NamedTuple):
270
+ lineno: int
271
+ type: str
272
+ value: str
273
+
274
+ def __str__(self) -> str:
275
+ return describe_token(self)
276
+
277
+ def test(self, expr: str) -> bool:
278
+ """Test a token against a token expression. This can either be a
279
+ token type or ``'token_type:token_value'``. This can only test
280
+ against string values and types.
281
+ """
282
+ # here we do a regular string equality check as test_any is usually
283
+ # passed an iterable of not interned strings.
284
+ if self.type == expr:
285
+ return True
286
+
287
+ if ":" in expr:
288
+ return expr.split(":", 1) == [self.type, self.value]
289
+
290
+ return False
291
+
292
+ def test_any(self, *iterable: str) -> bool:
293
+ """Test against multiple token expressions."""
294
+ return any(self.test(expr) for expr in iterable)
295
+
296
+
297
+ class TokenStreamIterator:
298
+ """The iterator for tokenstreams. Iterate over the stream
299
+ until the eof token is reached.
300
+ """
301
+
302
+ def __init__(self, stream: "TokenStream") -> None:
303
+ self.stream = stream
304
+
305
+ def __iter__(self) -> "TokenStreamIterator":
306
+ return self
307
+
308
+ def __next__(self) -> Token:
309
+ token = self.stream.current
310
+
311
+ if token.type is TOKEN_EOF:
312
+ self.stream.close()
313
+ raise StopIteration
314
+
315
+ next(self.stream)
316
+ return token
317
+
318
+
319
+ class TokenStream:
320
+ """A token stream is an iterable that yields :class:`Token`\\s. The
321
+ parser however does not iterate over it but calls :meth:`next` to go
322
+ one token ahead. The current active token is stored as :attr:`current`.
323
+ """
324
+
325
+ def __init__(
326
+ self,
327
+ generator: t.Iterable[Token],
328
+ name: t.Optional[str],
329
+ filename: t.Optional[str],
330
+ ):
331
+ self._iter = iter(generator)
332
+ self._pushed: te.Deque[Token] = deque()
333
+ self.name = name
334
+ self.filename = filename
335
+ self.closed = False
336
+ self.current = Token(1, TOKEN_INITIAL, "")
337
+ next(self)
338
+
339
+ def __iter__(self) -> TokenStreamIterator:
340
+ return TokenStreamIterator(self)
341
+
342
+ def __bool__(self) -> bool:
343
+ return bool(self._pushed) or self.current.type is not TOKEN_EOF
344
+
345
+ @property
346
+ def eos(self) -> bool:
347
+ """Are we at the end of the stream?"""
348
+ return not self
349
+
350
+ def push(self, token: Token) -> None:
351
+ """Push a token back to the stream."""
352
+ self._pushed.append(token)
353
+
354
+ def look(self) -> Token:
355
+ """Look at the next token."""
356
+ old_token = next(self)
357
+ result = self.current
358
+ self.push(result)
359
+ self.current = old_token
360
+ return result
361
+
362
+ def skip(self, n: int = 1) -> None:
363
+ """Got n tokens ahead."""
364
+ for _ in range(n):
365
+ next(self)
366
+
367
+ def next_if(self, expr: str) -> t.Optional[Token]:
368
+ """Perform the token test and return the token if it matched.
369
+ Otherwise the return value is `None`.
370
+ """
371
+ if self.current.test(expr):
372
+ return next(self)
373
+
374
+ return None
375
+
376
+ def skip_if(self, expr: str) -> bool:
377
+ """Like :meth:`next_if` but only returns `True` or `False`."""
378
+ return self.next_if(expr) is not None
379
+
380
+ def __next__(self) -> Token:
381
+ """Go one token ahead and return the old one.
382
+
383
+ Use the built-in :func:`next` instead of calling this directly.
384
+ """
385
+ rv = self.current
386
+
387
+ if self._pushed:
388
+ self.current = self._pushed.popleft()
389
+ elif self.current.type is not TOKEN_EOF:
390
+ try:
391
+ self.current = next(self._iter)
392
+ except StopIteration:
393
+ self.close()
394
+
395
+ return rv
396
+
397
+ def close(self) -> None:
398
+ """Close the stream."""
399
+ self.current = Token(self.current.lineno, TOKEN_EOF, "")
400
+ self._iter = iter(())
401
+ self.closed = True
402
+
403
+ def expect(self, expr: str) -> Token:
404
+ """Expect a given token type and return it. This accepts the same
405
+ argument as :meth:`jinja2.lexer.Token.test`.
406
+ """
407
+ if not self.current.test(expr):
408
+ expr = describe_token_expr(expr)
409
+
410
+ if self.current.type is TOKEN_EOF:
411
+ raise TemplateSyntaxError(
412
+ f"unexpected end of template, expected {expr!r}.",
413
+ self.current.lineno,
414
+ self.name,
415
+ self.filename,
416
+ )
417
+
418
+ raise TemplateSyntaxError(
419
+ f"expected token {expr!r}, got {describe_token(self.current)!r}",
420
+ self.current.lineno,
421
+ self.name,
422
+ self.filename,
423
+ )
424
+
425
+ return next(self)
426
+
427
+
428
+ def get_lexer(environment: "Environment") -> "Lexer":
429
+ """Return a lexer which is probably cached."""
430
+ key = (
431
+ environment.block_start_string,
432
+ environment.block_end_string,
433
+ environment.variable_start_string,
434
+ environment.variable_end_string,
435
+ environment.comment_start_string,
436
+ environment.comment_end_string,
437
+ environment.line_statement_prefix,
438
+ environment.line_comment_prefix,
439
+ environment.trim_blocks,
440
+ environment.lstrip_blocks,
441
+ environment.newline_sequence,
442
+ environment.keep_trailing_newline,
443
+ )
444
+ lexer = _lexer_cache.get(key)
445
+
446
+ if lexer is None:
447
+ _lexer_cache[key] = lexer = Lexer(environment)
448
+
449
+ return lexer
450
+
451
+
452
+ class OptionalLStrip(tuple): # type: ignore[type-arg]
453
+ """A special tuple for marking a point in the state that can have
454
+ lstrip applied.
455
+ """
456
+
457
+ __slots__ = ()
458
+
459
+ # Even though it looks like a no-op, creating instances fails
460
+ # without this.
461
+ def __new__(cls, *members, **kwargs): # type: ignore
462
+ return super().__new__(cls, members)
463
+
464
+
465
+ class _Rule(t.NamedTuple):
466
+ pattern: t.Pattern[str]
467
+ tokens: t.Union[str, t.Tuple[str, ...], t.Tuple[Failure]]
468
+ command: t.Optional[str]
469
+
470
+
471
+ class Lexer:
472
+ """Class that implements a lexer for a given environment. Automatically
473
+ created by the environment class, usually you don't have to do that.
474
+
475
+ Note that the lexer is not automatically bound to an environment.
476
+ Multiple environments can share the same lexer.
477
+ """
478
+
479
+ def __init__(self, environment: "Environment") -> None:
480
+ # shortcuts
481
+ e = re.escape
482
+
483
+ def c(x: str) -> t.Pattern[str]:
484
+ return re.compile(x, re.M | re.S)
485
+
486
+ # lexing rules for tags
487
+ tag_rules: t.List[_Rule] = [
488
+ _Rule(whitespace_re, TOKEN_WHITESPACE, None),
489
+ _Rule(float_re, TOKEN_FLOAT, None),
490
+ _Rule(integer_re, TOKEN_INTEGER, None),
491
+ _Rule(name_re, TOKEN_NAME, None),
492
+ _Rule(string_re, TOKEN_STRING, None),
493
+ _Rule(operator_re, TOKEN_OPERATOR, None),
494
+ ]
495
+
496
+ # assemble the root lexing rule. because "|" is ungreedy
497
+ # we have to sort by length so that the lexer continues working
498
+ # as expected when we have parsing rules like <% for block and
499
+ # <%= for variables. (if someone wants asp like syntax)
500
+ # variables are just part of the rules if variable processing
501
+ # is required.
502
+ root_tag_rules = compile_rules(environment)
503
+
504
+ block_start_re = e(environment.block_start_string)
505
+ block_end_re = e(environment.block_end_string)
506
+ comment_end_re = e(environment.comment_end_string)
507
+ variable_end_re = e(environment.variable_end_string)
508
+
509
+ # block suffix if trimming is enabled
510
+ block_suffix_re = "\\n?" if environment.trim_blocks else ""
511
+
512
+ self.lstrip_blocks = environment.lstrip_blocks
513
+
514
+ self.newline_sequence = environment.newline_sequence
515
+ self.keep_trailing_newline = environment.keep_trailing_newline
516
+
517
+ root_raw_re = (
518
+ rf"(?P<raw_begin>{block_start_re}(\-|\+|)\s*raw\s*"
519
+ rf"(?:\-{block_end_re}\s*|{block_end_re}))"
520
+ )
521
+ root_parts_re = "|".join(
522
+ [root_raw_re] + [rf"(?P<{n}>{r}(\-|\+|))" for n, r in root_tag_rules]
523
+ )
524
+
525
+ # global lexing rules
526
+ self.rules: t.Dict[str, t.List[_Rule]] = {
527
+ "root": [
528
+ # directives
529
+ _Rule(
530
+ c(rf"(.*?)(?:{root_parts_re})"),
531
+ OptionalLStrip(TOKEN_DATA, "#bygroup"), # type: ignore
532
+ "#bygroup",
533
+ ),
534
+ # data
535
+ _Rule(c(".+"), TOKEN_DATA, None),
536
+ ],
537
+ # comments
538
+ TOKEN_COMMENT_BEGIN: [
539
+ _Rule(
540
+ c(
541
+ rf"(.*?)((?:\+{comment_end_re}|\-{comment_end_re}\s*"
542
+ rf"|{comment_end_re}{block_suffix_re}))"
543
+ ),
544
+ (TOKEN_COMMENT, TOKEN_COMMENT_END),
545
+ "#pop",
546
+ ),
547
+ _Rule(c(r"(.)"), (Failure("Missing end of comment tag"),), None),
548
+ ],
549
+ # blocks
550
+ TOKEN_BLOCK_BEGIN: [
551
+ _Rule(
552
+ c(
553
+ rf"(?:\+{block_end_re}|\-{block_end_re}\s*"
554
+ rf"|{block_end_re}{block_suffix_re})"
555
+ ),
556
+ TOKEN_BLOCK_END,
557
+ "#pop",
558
+ ),
559
+ ]
560
+ + tag_rules,
561
+ # variables
562
+ TOKEN_VARIABLE_BEGIN: [
563
+ _Rule(
564
+ c(rf"\-{variable_end_re}\s*|{variable_end_re}"),
565
+ TOKEN_VARIABLE_END,
566
+ "#pop",
567
+ )
568
+ ]
569
+ + tag_rules,
570
+ # raw block
571
+ TOKEN_RAW_BEGIN: [
572
+ _Rule(
573
+ c(
574
+ rf"(.*?)((?:{block_start_re}(\-|\+|))\s*endraw\s*"
575
+ rf"(?:\+{block_end_re}|\-{block_end_re}\s*"
576
+ rf"|{block_end_re}{block_suffix_re}))"
577
+ ),
578
+ OptionalLStrip(TOKEN_DATA, TOKEN_RAW_END), # type: ignore
579
+ "#pop",
580
+ ),
581
+ _Rule(c(r"(.)"), (Failure("Missing end of raw directive"),), None),
582
+ ],
583
+ # line statements
584
+ TOKEN_LINESTATEMENT_BEGIN: [
585
+ _Rule(c(r"\s*(\n|$)"), TOKEN_LINESTATEMENT_END, "#pop")
586
+ ]
587
+ + tag_rules,
588
+ # line comments
589
+ TOKEN_LINECOMMENT_BEGIN: [
590
+ _Rule(
591
+ c(r"(.*?)()(?=\n|$)"),
592
+ (TOKEN_LINECOMMENT, TOKEN_LINECOMMENT_END),
593
+ "#pop",
594
+ )
595
+ ],
596
+ }
597
+
598
+ def _normalize_newlines(self, value: str) -> str:
599
+ """Replace all newlines with the configured sequence in strings
600
+ and template data.
601
+ """
602
+ return newline_re.sub(self.newline_sequence, value)
603
+
604
+ def tokenize(
605
+ self,
606
+ source: str,
607
+ name: t.Optional[str] = None,
608
+ filename: t.Optional[str] = None,
609
+ state: t.Optional[str] = None,
610
+ ) -> TokenStream:
611
+ """Calls tokeniter + tokenize and wraps it in a token stream."""
612
+ stream = self.tokeniter(source, name, filename, state)
613
+ return TokenStream(self.wrap(stream, name, filename), name, filename)
614
+
615
+ def wrap(
616
+ self,
617
+ stream: t.Iterable[t.Tuple[int, str, str]],
618
+ name: t.Optional[str] = None,
619
+ filename: t.Optional[str] = None,
620
+ ) -> t.Iterator[Token]:
621
+ """This is called with the stream as returned by `tokenize` and wraps
622
+ every token in a :class:`Token` and converts the value.
623
+ """
624
+ for lineno, token, value_str in stream:
625
+ if token in ignored_tokens:
626
+ continue
627
+
628
+ value: t.Any = value_str
629
+
630
+ if token == TOKEN_LINESTATEMENT_BEGIN:
631
+ token = TOKEN_BLOCK_BEGIN
632
+ elif token == TOKEN_LINESTATEMENT_END:
633
+ token = TOKEN_BLOCK_END
634
+ # we are not interested in those tokens in the parser
635
+ elif token in (TOKEN_RAW_BEGIN, TOKEN_RAW_END):
636
+ continue
637
+ elif token == TOKEN_DATA:
638
+ value = self._normalize_newlines(value_str)
639
+ elif token == "keyword":
640
+ token = value_str
641
+ elif token == TOKEN_NAME:
642
+ value = value_str
643
+
644
+ if not value.isidentifier():
645
+ raise TemplateSyntaxError(
646
+ "Invalid character in identifier", lineno, name, filename
647
+ )
648
+ elif token == TOKEN_STRING:
649
+ # try to unescape string
650
+ try:
651
+ value = (
652
+ self._normalize_newlines(value_str[1:-1])
653
+ .encode("ascii", "backslashreplace")
654
+ .decode("unicode-escape")
655
+ )
656
+ except Exception as e:
657
+ msg = str(e).split(":")[-1].strip()
658
+ raise TemplateSyntaxError(msg, lineno, name, filename) from e
659
+ elif token == TOKEN_INTEGER:
660
+ value = int(value_str.replace("_", ""), 0)
661
+ elif token == TOKEN_FLOAT:
662
+ # remove all "_" first to support more Python versions
663
+ value = literal_eval(value_str.replace("_", ""))
664
+ elif token == TOKEN_OPERATOR:
665
+ token = operators[value_str]
666
+
667
+ yield Token(lineno, token, value)
668
+
669
+ def tokeniter(
670
+ self,
671
+ source: str,
672
+ name: t.Optional[str],
673
+ filename: t.Optional[str] = None,
674
+ state: t.Optional[str] = None,
675
+ ) -> t.Iterator[t.Tuple[int, str, str]]:
676
+ """This method tokenizes the text and returns the tokens in a
677
+ generator. Use this method if you just want to tokenize a template.
678
+
679
+ .. versionchanged:: 3.0
680
+ Only ``\\n``, ``\\r\\n`` and ``\\r`` are treated as line
681
+ breaks.
682
+ """
683
+ lines = newline_re.split(source)[::2]
684
+
685
+ if not self.keep_trailing_newline and lines[-1] == "":
686
+ del lines[-1]
687
+
688
+ source = "\n".join(lines)
689
+ pos = 0
690
+ lineno = 1
691
+ stack = ["root"]
692
+
693
+ if state is not None and state != "root":
694
+ assert state in ("variable", "block"), "invalid state"
695
+ stack.append(state + "_begin")
696
+
697
+ statetokens = self.rules[stack[-1]]
698
+ source_length = len(source)
699
+ balancing_stack: t.List[str] = []
700
+ newlines_stripped = 0
701
+ line_starting = True
702
+
703
+ while True:
704
+ # tokenizer loop
705
+ for regex, tokens, new_state in statetokens:
706
+ m = regex.match(source, pos)
707
+
708
+ # if no match we try again with the next rule
709
+ if m is None:
710
+ continue
711
+
712
+ # we only match blocks and variables if braces / parentheses
713
+ # are balanced. continue parsing with the lower rule which
714
+ # is the operator rule. do this only if the end tags look
715
+ # like operators
716
+ if balancing_stack and tokens in (
717
+ TOKEN_VARIABLE_END,
718
+ TOKEN_BLOCK_END,
719
+ TOKEN_LINESTATEMENT_END,
720
+ ):
721
+ continue
722
+
723
+ # tuples support more options
724
+ if isinstance(tokens, tuple):
725
+ groups: t.Sequence[str] = m.groups()
726
+
727
+ if isinstance(tokens, OptionalLStrip):
728
+ # Rule supports lstrip. Match will look like
729
+ # text, block type, whitespace control, type, control, ...
730
+ text = groups[0]
731
+ # Skipping the text and first type, every other group is the
732
+ # whitespace control for each type. One of the groups will be
733
+ # -, +, or empty string instead of None.
734
+ strip_sign = next(g for g in groups[2::2] if g is not None)
735
+
736
+ if strip_sign == "-":
737
+ # Strip all whitespace between the text and the tag.
738
+ stripped = text.rstrip()
739
+ newlines_stripped = text[len(stripped) :].count("\n")
740
+ groups = [stripped, *groups[1:]]
741
+ elif (
742
+ # Not marked for preserving whitespace.
743
+ strip_sign != "+"
744
+ # lstrip is enabled.
745
+ and self.lstrip_blocks
746
+ # Not a variable expression.
747
+ and not m.groupdict().get(TOKEN_VARIABLE_BEGIN)
748
+ ):
749
+ # The start of text between the last newline and the tag.
750
+ l_pos = text.rfind("\n") + 1
751
+
752
+ if l_pos > 0 or line_starting:
753
+ # If there's only whitespace between the newline and the
754
+ # tag, strip it.
755
+ if whitespace_re.fullmatch(text, l_pos):
756
+ groups = [text[:l_pos], *groups[1:]]
757
+
758
+ for idx, token in enumerate(tokens):
759
+ # failure group
760
+ if isinstance(token, Failure):
761
+ raise token(lineno, filename)
762
+ # bygroup is a bit more complex, in that case we
763
+ # yield for the current token the first named
764
+ # group that matched
765
+ elif token == "#bygroup":
766
+ for key, value in m.groupdict().items():
767
+ if value is not None:
768
+ yield lineno, key, value
769
+ lineno += value.count("\n")
770
+ break
771
+ else:
772
+ raise RuntimeError(
773
+ f"{regex!r} wanted to resolve the token dynamically"
774
+ " but no group matched"
775
+ )
776
+ # normal group
777
+ else:
778
+ data = groups[idx]
779
+
780
+ if data or token not in ignore_if_empty:
781
+ yield lineno, token, data # type: ignore[misc]
782
+
783
+ lineno += data.count("\n") + newlines_stripped
784
+ newlines_stripped = 0
785
+
786
+ # strings as token just are yielded as it.
787
+ else:
788
+ data = m.group()
789
+
790
+ # update brace/parentheses balance
791
+ if tokens == TOKEN_OPERATOR:
792
+ if data == "{":
793
+ balancing_stack.append("}")
794
+ elif data == "(":
795
+ balancing_stack.append(")")
796
+ elif data == "[":
797
+ balancing_stack.append("]")
798
+ elif data in ("}", ")", "]"):
799
+ if not balancing_stack:
800
+ raise TemplateSyntaxError(
801
+ f"unexpected '{data}'", lineno, name, filename
802
+ )
803
+
804
+ expected_op = balancing_stack.pop()
805
+
806
+ if expected_op != data:
807
+ raise TemplateSyntaxError(
808
+ f"unexpected '{data}', expected '{expected_op}'",
809
+ lineno,
810
+ name,
811
+ filename,
812
+ )
813
+
814
+ # yield items
815
+ if data or tokens not in ignore_if_empty:
816
+ yield lineno, tokens, data
817
+
818
+ lineno += data.count("\n")
819
+
820
+ line_starting = m.group()[-1:] == "\n"
821
+ # fetch new position into new variable so that we can check
822
+ # if there is a internal parsing error which would result
823
+ # in an infinite loop
824
+ pos2 = m.end()
825
+
826
+ # handle state changes
827
+ if new_state is not None:
828
+ # remove the uppermost state
829
+ if new_state == "#pop":
830
+ stack.pop()
831
+ # resolve the new state by group checking
832
+ elif new_state == "#bygroup":
833
+ for key, value in m.groupdict().items():
834
+ if value is not None:
835
+ stack.append(key)
836
+ break
837
+ else:
838
+ raise RuntimeError(
839
+ f"{regex!r} wanted to resolve the new state dynamically"
840
+ f" but no group matched"
841
+ )
842
+ # direct state name given
843
+ else:
844
+ stack.append(new_state)
845
+
846
+ statetokens = self.rules[stack[-1]]
847
+ # we are still at the same position and no stack change.
848
+ # this means a loop without break condition, avoid that and
849
+ # raise error
850
+ elif pos2 == pos:
851
+ raise RuntimeError(
852
+ f"{regex!r} yielded empty string without stack change"
853
+ )
854
+
855
+ # publish new function and start again
856
+ pos = pos2
857
+ break
858
+ # if loop terminated without break we haven't found a single match
859
+ # either we are at the end of the file or we have a problem
860
+ else:
861
+ # end of text
862
+ if pos >= source_length:
863
+ return
864
+
865
+ # something went wrong
866
+ raise TemplateSyntaxError(
867
+ f"unexpected char {source[pos]!r} at {pos}", lineno, name, filename
868
+ )
janus/lib/python3.10/site-packages/jinja2/meta.py ADDED
@@ -0,0 +1,112 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """Functions that expose information about templates that might be
2
+ interesting for introspection.
3
+ """
4
+
5
+ import typing as t
6
+
7
+ from . import nodes
8
+ from .compiler import CodeGenerator
9
+ from .compiler import Frame
10
+
11
+ if t.TYPE_CHECKING:
12
+ from .environment import Environment
13
+
14
+
15
+ class TrackingCodeGenerator(CodeGenerator):
16
+ """We abuse the code generator for introspection."""
17
+
18
+ def __init__(self, environment: "Environment") -> None:
19
+ super().__init__(environment, "<introspection>", "<introspection>")
20
+ self.undeclared_identifiers: t.Set[str] = set()
21
+
22
+ def write(self, x: str) -> None:
23
+ """Don't write."""
24
+
25
+ def enter_frame(self, frame: Frame) -> None:
26
+ """Remember all undeclared identifiers."""
27
+ super().enter_frame(frame)
28
+
29
+ for _, (action, param) in frame.symbols.loads.items():
30
+ if action == "resolve" and param not in self.environment.globals:
31
+ self.undeclared_identifiers.add(param)
32
+
33
+
34
+ def find_undeclared_variables(ast: nodes.Template) -> t.Set[str]:
35
+ """Returns a set of all variables in the AST that will be looked up from
36
+ the context at runtime. Because at compile time it's not known which
37
+ variables will be used depending on the path the execution takes at
38
+ runtime, all variables are returned.
39
+
40
+ >>> from jinja2 import Environment, meta
41
+ >>> env = Environment()
42
+ >>> ast = env.parse('{% set foo = 42 %}{{ bar + foo }}')
43
+ >>> meta.find_undeclared_variables(ast) == {'bar'}
44
+ True
45
+
46
+ .. admonition:: Implementation
47
+
48
+ Internally the code generator is used for finding undeclared variables.
49
+ This is good to know because the code generator might raise a
50
+ :exc:`TemplateAssertionError` during compilation and as a matter of
51
+ fact this function can currently raise that exception as well.
52
+ """
53
+ codegen = TrackingCodeGenerator(ast.environment) # type: ignore
54
+ codegen.visit(ast)
55
+ return codegen.undeclared_identifiers
56
+
57
+
58
+ _ref_types = (nodes.Extends, nodes.FromImport, nodes.Import, nodes.Include)
59
+ _RefType = t.Union[nodes.Extends, nodes.FromImport, nodes.Import, nodes.Include]
60
+
61
+
62
+ def find_referenced_templates(ast: nodes.Template) -> t.Iterator[t.Optional[str]]:
63
+ """Finds all the referenced templates from the AST. This will return an
64
+ iterator over all the hardcoded template extensions, inclusions and
65
+ imports. If dynamic inheritance or inclusion is used, `None` will be
66
+ yielded.
67
+
68
+ >>> from jinja2 import Environment, meta
69
+ >>> env = Environment()
70
+ >>> ast = env.parse('{% extends "layout.html" %}{% include helper %}')
71
+ >>> list(meta.find_referenced_templates(ast))
72
+ ['layout.html', None]
73
+
74
+ This function is useful for dependency tracking. For example if you want
75
+ to rebuild parts of the website after a layout template has changed.
76
+ """
77
+ template_name: t.Any
78
+
79
+ for node in ast.find_all(_ref_types):
80
+ template: nodes.Expr = node.template # type: ignore
81
+
82
+ if not isinstance(template, nodes.Const):
83
+ # a tuple with some non consts in there
84
+ if isinstance(template, (nodes.Tuple, nodes.List)):
85
+ for template_name in template.items:
86
+ # something const, only yield the strings and ignore
87
+ # non-string consts that really just make no sense
88
+ if isinstance(template_name, nodes.Const):
89
+ if isinstance(template_name.value, str):
90
+ yield template_name.value
91
+ # something dynamic in there
92
+ else:
93
+ yield None
94
+ # something dynamic we don't know about here
95
+ else:
96
+ yield None
97
+ continue
98
+ # constant is a basestring, direct template name
99
+ if isinstance(template.value, str):
100
+ yield template.value
101
+ # a tuple or list (latter *should* not happen) made of consts,
102
+ # yield the consts that are strings. We could warn here for
103
+ # non string values
104
+ elif isinstance(node, nodes.Include) and isinstance(
105
+ template.value, (tuple, list)
106
+ ):
107
+ for template_name in template.value:
108
+ if isinstance(template_name, str):
109
+ yield template_name
110
+ # something else we don't care about, we could warn here
111
+ else:
112
+ yield None
janus/lib/python3.10/site-packages/jinja2/nativetypes.py ADDED
@@ -0,0 +1,130 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import typing as t
2
+ from ast import literal_eval
3
+ from ast import parse
4
+ from itertools import chain
5
+ from itertools import islice
6
+ from types import GeneratorType
7
+
8
+ from . import nodes
9
+ from .compiler import CodeGenerator
10
+ from .compiler import Frame
11
+ from .compiler import has_safe_repr
12
+ from .environment import Environment
13
+ from .environment import Template
14
+
15
+
16
+ def native_concat(values: t.Iterable[t.Any]) -> t.Optional[t.Any]:
17
+ """Return a native Python type from the list of compiled nodes. If
18
+ the result is a single node, its value is returned. Otherwise, the
19
+ nodes are concatenated as strings. If the result can be parsed with
20
+ :func:`ast.literal_eval`, the parsed value is returned. Otherwise,
21
+ the string is returned.
22
+
23
+ :param values: Iterable of outputs to concatenate.
24
+ """
25
+ head = list(islice(values, 2))
26
+
27
+ if not head:
28
+ return None
29
+
30
+ if len(head) == 1:
31
+ raw = head[0]
32
+ if not isinstance(raw, str):
33
+ return raw
34
+ else:
35
+ if isinstance(values, GeneratorType):
36
+ values = chain(head, values)
37
+ raw = "".join([str(v) for v in values])
38
+
39
+ try:
40
+ return literal_eval(
41
+ # In Python 3.10+ ast.literal_eval removes leading spaces/tabs
42
+ # from the given string. For backwards compatibility we need to
43
+ # parse the string ourselves without removing leading spaces/tabs.
44
+ parse(raw, mode="eval")
45
+ )
46
+ except (ValueError, SyntaxError, MemoryError):
47
+ return raw
48
+
49
+
50
+ class NativeCodeGenerator(CodeGenerator):
51
+ """A code generator which renders Python types by not adding
52
+ ``str()`` around output nodes.
53
+ """
54
+
55
+ @staticmethod
56
+ def _default_finalize(value: t.Any) -> t.Any:
57
+ return value
58
+
59
+ def _output_const_repr(self, group: t.Iterable[t.Any]) -> str:
60
+ return repr("".join([str(v) for v in group]))
61
+
62
+ def _output_child_to_const(
63
+ self, node: nodes.Expr, frame: Frame, finalize: CodeGenerator._FinalizeInfo
64
+ ) -> t.Any:
65
+ const = node.as_const(frame.eval_ctx)
66
+
67
+ if not has_safe_repr(const):
68
+ raise nodes.Impossible()
69
+
70
+ if isinstance(node, nodes.TemplateData):
71
+ return const
72
+
73
+ return finalize.const(const) # type: ignore
74
+
75
+ def _output_child_pre(
76
+ self, node: nodes.Expr, frame: Frame, finalize: CodeGenerator._FinalizeInfo
77
+ ) -> None:
78
+ if finalize.src is not None:
79
+ self.write(finalize.src)
80
+
81
+ def _output_child_post(
82
+ self, node: nodes.Expr, frame: Frame, finalize: CodeGenerator._FinalizeInfo
83
+ ) -> None:
84
+ if finalize.src is not None:
85
+ self.write(")")
86
+
87
+
88
+ class NativeEnvironment(Environment):
89
+ """An environment that renders templates to native Python types."""
90
+
91
+ code_generator_class = NativeCodeGenerator
92
+ concat = staticmethod(native_concat) # type: ignore
93
+
94
+
95
+ class NativeTemplate(Template):
96
+ environment_class = NativeEnvironment
97
+
98
+ def render(self, *args: t.Any, **kwargs: t.Any) -> t.Any:
99
+ """Render the template to produce a native Python type. If the
100
+ result is a single node, its value is returned. Otherwise, the
101
+ nodes are concatenated as strings. If the result can be parsed
102
+ with :func:`ast.literal_eval`, the parsed value is returned.
103
+ Otherwise, the string is returned.
104
+ """
105
+ ctx = self.new_context(dict(*args, **kwargs))
106
+
107
+ try:
108
+ return self.environment_class.concat( # type: ignore
109
+ self.root_render_func(ctx)
110
+ )
111
+ except Exception:
112
+ return self.environment.handle_exception()
113
+
114
+ async def render_async(self, *args: t.Any, **kwargs: t.Any) -> t.Any:
115
+ if not self.environment.is_async:
116
+ raise RuntimeError(
117
+ "The environment was not created with async mode enabled."
118
+ )
119
+
120
+ ctx = self.new_context(dict(*args, **kwargs))
121
+
122
+ try:
123
+ return self.environment_class.concat( # type: ignore
124
+ [n async for n in self.root_render_func(ctx)] # type: ignore
125
+ )
126
+ except Exception:
127
+ return self.environment.handle_exception()
128
+
129
+
130
+ NativeEnvironment.template_class = NativeTemplate
janus/lib/python3.10/site-packages/jinja2/nodes.py ADDED
@@ -0,0 +1,1206 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """AST nodes generated by the parser for the compiler. Also provides
2
+ some node tree helper functions used by the parser and compiler in order
3
+ to normalize nodes.
4
+ """
5
+
6
+ import inspect
7
+ import operator
8
+ import typing as t
9
+ from collections import deque
10
+
11
+ from markupsafe import Markup
12
+
13
+ from .utils import _PassArg
14
+
15
+ if t.TYPE_CHECKING:
16
+ import typing_extensions as te
17
+
18
+ from .environment import Environment
19
+
20
+ _NodeBound = t.TypeVar("_NodeBound", bound="Node")
21
+
22
+ _binop_to_func: t.Dict[str, t.Callable[[t.Any, t.Any], t.Any]] = {
23
+ "*": operator.mul,
24
+ "/": operator.truediv,
25
+ "//": operator.floordiv,
26
+ "**": operator.pow,
27
+ "%": operator.mod,
28
+ "+": operator.add,
29
+ "-": operator.sub,
30
+ }
31
+
32
+ _uaop_to_func: t.Dict[str, t.Callable[[t.Any], t.Any]] = {
33
+ "not": operator.not_,
34
+ "+": operator.pos,
35
+ "-": operator.neg,
36
+ }
37
+
38
+ _cmpop_to_func: t.Dict[str, t.Callable[[t.Any, t.Any], t.Any]] = {
39
+ "eq": operator.eq,
40
+ "ne": operator.ne,
41
+ "gt": operator.gt,
42
+ "gteq": operator.ge,
43
+ "lt": operator.lt,
44
+ "lteq": operator.le,
45
+ "in": lambda a, b: a in b,
46
+ "notin": lambda a, b: a not in b,
47
+ }
48
+
49
+
50
+ class Impossible(Exception):
51
+ """Raised if the node could not perform a requested action."""
52
+
53
+
54
+ class NodeType(type):
55
+ """A metaclass for nodes that handles the field and attribute
56
+ inheritance. fields and attributes from the parent class are
57
+ automatically forwarded to the child."""
58
+
59
+ def __new__(mcs, name, bases, d): # type: ignore
60
+ for attr in "fields", "attributes":
61
+ storage: t.List[t.Tuple[str, ...]] = []
62
+ storage.extend(getattr(bases[0] if bases else object, attr, ()))
63
+ storage.extend(d.get(attr, ()))
64
+ assert len(bases) <= 1, "multiple inheritance not allowed"
65
+ assert len(storage) == len(set(storage)), "layout conflict"
66
+ d[attr] = tuple(storage)
67
+ d.setdefault("abstract", False)
68
+ return type.__new__(mcs, name, bases, d)
69
+
70
+
71
+ class EvalContext:
72
+ """Holds evaluation time information. Custom attributes can be attached
73
+ to it in extensions.
74
+ """
75
+
76
+ def __init__(
77
+ self, environment: "Environment", template_name: t.Optional[str] = None
78
+ ) -> None:
79
+ self.environment = environment
80
+ if callable(environment.autoescape):
81
+ self.autoescape = environment.autoescape(template_name)
82
+ else:
83
+ self.autoescape = environment.autoescape
84
+ self.volatile = False
85
+
86
+ def save(self) -> t.Mapping[str, t.Any]:
87
+ return self.__dict__.copy()
88
+
89
+ def revert(self, old: t.Mapping[str, t.Any]) -> None:
90
+ self.__dict__.clear()
91
+ self.__dict__.update(old)
92
+
93
+
94
+ def get_eval_context(node: "Node", ctx: t.Optional[EvalContext]) -> EvalContext:
95
+ if ctx is None:
96
+ if node.environment is None:
97
+ raise RuntimeError(
98
+ "if no eval context is passed, the node must have an"
99
+ " attached environment."
100
+ )
101
+ return EvalContext(node.environment)
102
+ return ctx
103
+
104
+
105
+ class Node(metaclass=NodeType):
106
+ """Baseclass for all Jinja nodes. There are a number of nodes available
107
+ of different types. There are four major types:
108
+
109
+ - :class:`Stmt`: statements
110
+ - :class:`Expr`: expressions
111
+ - :class:`Helper`: helper nodes
112
+ - :class:`Template`: the outermost wrapper node
113
+
114
+ All nodes have fields and attributes. Fields may be other nodes, lists,
115
+ or arbitrary values. Fields are passed to the constructor as regular
116
+ positional arguments, attributes as keyword arguments. Each node has
117
+ two attributes: `lineno` (the line number of the node) and `environment`.
118
+ The `environment` attribute is set at the end of the parsing process for
119
+ all nodes automatically.
120
+ """
121
+
122
+ fields: t.Tuple[str, ...] = ()
123
+ attributes: t.Tuple[str, ...] = ("lineno", "environment")
124
+ abstract = True
125
+
126
+ lineno: int
127
+ environment: t.Optional["Environment"]
128
+
129
+ def __init__(self, *fields: t.Any, **attributes: t.Any) -> None:
130
+ if self.abstract:
131
+ raise TypeError("abstract nodes are not instantiable")
132
+ if fields:
133
+ if len(fields) != len(self.fields):
134
+ if not self.fields:
135
+ raise TypeError(f"{type(self).__name__!r} takes 0 arguments")
136
+ raise TypeError(
137
+ f"{type(self).__name__!r} takes 0 or {len(self.fields)}"
138
+ f" argument{'s' if len(self.fields) != 1 else ''}"
139
+ )
140
+ for name, arg in zip(self.fields, fields):
141
+ setattr(self, name, arg)
142
+ for attr in self.attributes:
143
+ setattr(self, attr, attributes.pop(attr, None))
144
+ if attributes:
145
+ raise TypeError(f"unknown attribute {next(iter(attributes))!r}")
146
+
147
+ def iter_fields(
148
+ self,
149
+ exclude: t.Optional[t.Container[str]] = None,
150
+ only: t.Optional[t.Container[str]] = None,
151
+ ) -> t.Iterator[t.Tuple[str, t.Any]]:
152
+ """This method iterates over all fields that are defined and yields
153
+ ``(key, value)`` tuples. Per default all fields are returned, but
154
+ it's possible to limit that to some fields by providing the `only`
155
+ parameter or to exclude some using the `exclude` parameter. Both
156
+ should be sets or tuples of field names.
157
+ """
158
+ for name in self.fields:
159
+ if (
160
+ (exclude is None and only is None)
161
+ or (exclude is not None and name not in exclude)
162
+ or (only is not None and name in only)
163
+ ):
164
+ try:
165
+ yield name, getattr(self, name)
166
+ except AttributeError:
167
+ pass
168
+
169
+ def iter_child_nodes(
170
+ self,
171
+ exclude: t.Optional[t.Container[str]] = None,
172
+ only: t.Optional[t.Container[str]] = None,
173
+ ) -> t.Iterator["Node"]:
174
+ """Iterates over all direct child nodes of the node. This iterates
175
+ over all fields and yields the values of they are nodes. If the value
176
+ of a field is a list all the nodes in that list are returned.
177
+ """
178
+ for _, item in self.iter_fields(exclude, only):
179
+ if isinstance(item, list):
180
+ for n in item:
181
+ if isinstance(n, Node):
182
+ yield n
183
+ elif isinstance(item, Node):
184
+ yield item
185
+
186
+ def find(self, node_type: t.Type[_NodeBound]) -> t.Optional[_NodeBound]:
187
+ """Find the first node of a given type. If no such node exists the
188
+ return value is `None`.
189
+ """
190
+ for result in self.find_all(node_type):
191
+ return result
192
+
193
+ return None
194
+
195
+ def find_all(
196
+ self, node_type: t.Union[t.Type[_NodeBound], t.Tuple[t.Type[_NodeBound], ...]]
197
+ ) -> t.Iterator[_NodeBound]:
198
+ """Find all the nodes of a given type. If the type is a tuple,
199
+ the check is performed for any of the tuple items.
200
+ """
201
+ for child in self.iter_child_nodes():
202
+ if isinstance(child, node_type):
203
+ yield child # type: ignore
204
+ yield from child.find_all(node_type)
205
+
206
+ def set_ctx(self, ctx: str) -> "Node":
207
+ """Reset the context of a node and all child nodes. Per default the
208
+ parser will all generate nodes that have a 'load' context as it's the
209
+ most common one. This method is used in the parser to set assignment
210
+ targets and other nodes to a store context.
211
+ """
212
+ todo = deque([self])
213
+ while todo:
214
+ node = todo.popleft()
215
+ if "ctx" in node.fields:
216
+ node.ctx = ctx # type: ignore
217
+ todo.extend(node.iter_child_nodes())
218
+ return self
219
+
220
+ def set_lineno(self, lineno: int, override: bool = False) -> "Node":
221
+ """Set the line numbers of the node and children."""
222
+ todo = deque([self])
223
+ while todo:
224
+ node = todo.popleft()
225
+ if "lineno" in node.attributes:
226
+ if node.lineno is None or override:
227
+ node.lineno = lineno
228
+ todo.extend(node.iter_child_nodes())
229
+ return self
230
+
231
+ def set_environment(self, environment: "Environment") -> "Node":
232
+ """Set the environment for all nodes."""
233
+ todo = deque([self])
234
+ while todo:
235
+ node = todo.popleft()
236
+ node.environment = environment
237
+ todo.extend(node.iter_child_nodes())
238
+ return self
239
+
240
+ def __eq__(self, other: t.Any) -> bool:
241
+ if type(self) is not type(other):
242
+ return NotImplemented
243
+
244
+ return tuple(self.iter_fields()) == tuple(other.iter_fields())
245
+
246
+ __hash__ = object.__hash__
247
+
248
+ def __repr__(self) -> str:
249
+ args_str = ", ".join(f"{a}={getattr(self, a, None)!r}" for a in self.fields)
250
+ return f"{type(self).__name__}({args_str})"
251
+
252
+ def dump(self) -> str:
253
+ def _dump(node: t.Union[Node, t.Any]) -> None:
254
+ if not isinstance(node, Node):
255
+ buf.append(repr(node))
256
+ return
257
+
258
+ buf.append(f"nodes.{type(node).__name__}(")
259
+ if not node.fields:
260
+ buf.append(")")
261
+ return
262
+ for idx, field in enumerate(node.fields):
263
+ if idx:
264
+ buf.append(", ")
265
+ value = getattr(node, field)
266
+ if isinstance(value, list):
267
+ buf.append("[")
268
+ for idx, item in enumerate(value):
269
+ if idx:
270
+ buf.append(", ")
271
+ _dump(item)
272
+ buf.append("]")
273
+ else:
274
+ _dump(value)
275
+ buf.append(")")
276
+
277
+ buf: t.List[str] = []
278
+ _dump(self)
279
+ return "".join(buf)
280
+
281
+
282
+ class Stmt(Node):
283
+ """Base node for all statements."""
284
+
285
+ abstract = True
286
+
287
+
288
+ class Helper(Node):
289
+ """Nodes that exist in a specific context only."""
290
+
291
+ abstract = True
292
+
293
+
294
+ class Template(Node):
295
+ """Node that represents a template. This must be the outermost node that
296
+ is passed to the compiler.
297
+ """
298
+
299
+ fields = ("body",)
300
+ body: t.List[Node]
301
+
302
+
303
+ class Output(Stmt):
304
+ """A node that holds multiple expressions which are then printed out.
305
+ This is used both for the `print` statement and the regular template data.
306
+ """
307
+
308
+ fields = ("nodes",)
309
+ nodes: t.List["Expr"]
310
+
311
+
312
+ class Extends(Stmt):
313
+ """Represents an extends statement."""
314
+
315
+ fields = ("template",)
316
+ template: "Expr"
317
+
318
+
319
+ class For(Stmt):
320
+ """The for loop. `target` is the target for the iteration (usually a
321
+ :class:`Name` or :class:`Tuple`), `iter` the iterable. `body` is a list
322
+ of nodes that are used as loop-body, and `else_` a list of nodes for the
323
+ `else` block. If no else node exists it has to be an empty list.
324
+
325
+ For filtered nodes an expression can be stored as `test`, otherwise `None`.
326
+ """
327
+
328
+ fields = ("target", "iter", "body", "else_", "test", "recursive")
329
+ target: Node
330
+ iter: Node
331
+ body: t.List[Node]
332
+ else_: t.List[Node]
333
+ test: t.Optional[Node]
334
+ recursive: bool
335
+
336
+
337
+ class If(Stmt):
338
+ """If `test` is true, `body` is rendered, else `else_`."""
339
+
340
+ fields = ("test", "body", "elif_", "else_")
341
+ test: Node
342
+ body: t.List[Node]
343
+ elif_: t.List["If"]
344
+ else_: t.List[Node]
345
+
346
+
347
+ class Macro(Stmt):
348
+ """A macro definition. `name` is the name of the macro, `args` a list of
349
+ arguments and `defaults` a list of defaults if there are any. `body` is
350
+ a list of nodes for the macro body.
351
+ """
352
+
353
+ fields = ("name", "args", "defaults", "body")
354
+ name: str
355
+ args: t.List["Name"]
356
+ defaults: t.List["Expr"]
357
+ body: t.List[Node]
358
+
359
+
360
+ class CallBlock(Stmt):
361
+ """Like a macro without a name but a call instead. `call` is called with
362
+ the unnamed macro as `caller` argument this node holds.
363
+ """
364
+
365
+ fields = ("call", "args", "defaults", "body")
366
+ call: "Call"
367
+ args: t.List["Name"]
368
+ defaults: t.List["Expr"]
369
+ body: t.List[Node]
370
+
371
+
372
+ class FilterBlock(Stmt):
373
+ """Node for filter sections."""
374
+
375
+ fields = ("body", "filter")
376
+ body: t.List[Node]
377
+ filter: "Filter"
378
+
379
+
380
+ class With(Stmt):
381
+ """Specific node for with statements. In older versions of Jinja the
382
+ with statement was implemented on the base of the `Scope` node instead.
383
+
384
+ .. versionadded:: 2.9.3
385
+ """
386
+
387
+ fields = ("targets", "values", "body")
388
+ targets: t.List["Expr"]
389
+ values: t.List["Expr"]
390
+ body: t.List[Node]
391
+
392
+
393
+ class Block(Stmt):
394
+ """A node that represents a block.
395
+
396
+ .. versionchanged:: 3.0.0
397
+ the `required` field was added.
398
+ """
399
+
400
+ fields = ("name", "body", "scoped", "required")
401
+ name: str
402
+ body: t.List[Node]
403
+ scoped: bool
404
+ required: bool
405
+
406
+
407
+ class Include(Stmt):
408
+ """A node that represents the include tag."""
409
+
410
+ fields = ("template", "with_context", "ignore_missing")
411
+ template: "Expr"
412
+ with_context: bool
413
+ ignore_missing: bool
414
+
415
+
416
+ class Import(Stmt):
417
+ """A node that represents the import tag."""
418
+
419
+ fields = ("template", "target", "with_context")
420
+ template: "Expr"
421
+ target: str
422
+ with_context: bool
423
+
424
+
425
+ class FromImport(Stmt):
426
+ """A node that represents the from import tag. It's important to not
427
+ pass unsafe names to the name attribute. The compiler translates the
428
+ attribute lookups directly into getattr calls and does *not* use the
429
+ subscript callback of the interface. As exported variables may not
430
+ start with double underscores (which the parser asserts) this is not a
431
+ problem for regular Jinja code, but if this node is used in an extension
432
+ extra care must be taken.
433
+
434
+ The list of names may contain tuples if aliases are wanted.
435
+ """
436
+
437
+ fields = ("template", "names", "with_context")
438
+ template: "Expr"
439
+ names: t.List[t.Union[str, t.Tuple[str, str]]]
440
+ with_context: bool
441
+
442
+
443
+ class ExprStmt(Stmt):
444
+ """A statement that evaluates an expression and discards the result."""
445
+
446
+ fields = ("node",)
447
+ node: Node
448
+
449
+
450
+ class Assign(Stmt):
451
+ """Assigns an expression to a target."""
452
+
453
+ fields = ("target", "node")
454
+ target: "Expr"
455
+ node: Node
456
+
457
+
458
+ class AssignBlock(Stmt):
459
+ """Assigns a block to a target."""
460
+
461
+ fields = ("target", "filter", "body")
462
+ target: "Expr"
463
+ filter: t.Optional["Filter"]
464
+ body: t.List[Node]
465
+
466
+
467
+ class Expr(Node):
468
+ """Baseclass for all expressions."""
469
+
470
+ abstract = True
471
+
472
+ def as_const(self, eval_ctx: t.Optional[EvalContext] = None) -> t.Any:
473
+ """Return the value of the expression as constant or raise
474
+ :exc:`Impossible` if this was not possible.
475
+
476
+ An :class:`EvalContext` can be provided, if none is given
477
+ a default context is created which requires the nodes to have
478
+ an attached environment.
479
+
480
+ .. versionchanged:: 2.4
481
+ the `eval_ctx` parameter was added.
482
+ """
483
+ raise Impossible()
484
+
485
+ def can_assign(self) -> bool:
486
+ """Check if it's possible to assign something to this node."""
487
+ return False
488
+
489
+
490
+ class BinExpr(Expr):
491
+ """Baseclass for all binary expressions."""
492
+
493
+ fields = ("left", "right")
494
+ left: Expr
495
+ right: Expr
496
+ operator: str
497
+ abstract = True
498
+
499
+ def as_const(self, eval_ctx: t.Optional[EvalContext] = None) -> t.Any:
500
+ eval_ctx = get_eval_context(self, eval_ctx)
501
+
502
+ # intercepted operators cannot be folded at compile time
503
+ if (
504
+ eval_ctx.environment.sandboxed
505
+ and self.operator in eval_ctx.environment.intercepted_binops # type: ignore
506
+ ):
507
+ raise Impossible()
508
+ f = _binop_to_func[self.operator]
509
+ try:
510
+ return f(self.left.as_const(eval_ctx), self.right.as_const(eval_ctx))
511
+ except Exception as e:
512
+ raise Impossible() from e
513
+
514
+
515
+ class UnaryExpr(Expr):
516
+ """Baseclass for all unary expressions."""
517
+
518
+ fields = ("node",)
519
+ node: Expr
520
+ operator: str
521
+ abstract = True
522
+
523
+ def as_const(self, eval_ctx: t.Optional[EvalContext] = None) -> t.Any:
524
+ eval_ctx = get_eval_context(self, eval_ctx)
525
+
526
+ # intercepted operators cannot be folded at compile time
527
+ if (
528
+ eval_ctx.environment.sandboxed
529
+ and self.operator in eval_ctx.environment.intercepted_unops # type: ignore
530
+ ):
531
+ raise Impossible()
532
+ f = _uaop_to_func[self.operator]
533
+ try:
534
+ return f(self.node.as_const(eval_ctx))
535
+ except Exception as e:
536
+ raise Impossible() from e
537
+
538
+
539
+ class Name(Expr):
540
+ """Looks up a name or stores a value in a name.
541
+ The `ctx` of the node can be one of the following values:
542
+
543
+ - `store`: store a value in the name
544
+ - `load`: load that name
545
+ - `param`: like `store` but if the name was defined as function parameter.
546
+ """
547
+
548
+ fields = ("name", "ctx")
549
+ name: str
550
+ ctx: str
551
+
552
+ def can_assign(self) -> bool:
553
+ return self.name not in {"true", "false", "none", "True", "False", "None"}
554
+
555
+
556
+ class NSRef(Expr):
557
+ """Reference to a namespace value assignment"""
558
+
559
+ fields = ("name", "attr")
560
+ name: str
561
+ attr: str
562
+
563
+ def can_assign(self) -> bool:
564
+ # We don't need any special checks here; NSRef assignments have a
565
+ # runtime check to ensure the target is a namespace object which will
566
+ # have been checked already as it is created using a normal assignment
567
+ # which goes through a `Name` node.
568
+ return True
569
+
570
+
571
+ class Literal(Expr):
572
+ """Baseclass for literals."""
573
+
574
+ abstract = True
575
+
576
+
577
+ class Const(Literal):
578
+ """All constant values. The parser will return this node for simple
579
+ constants such as ``42`` or ``"foo"`` but it can be used to store more
580
+ complex values such as lists too. Only constants with a safe
581
+ representation (objects where ``eval(repr(x)) == x`` is true).
582
+ """
583
+
584
+ fields = ("value",)
585
+ value: t.Any
586
+
587
+ def as_const(self, eval_ctx: t.Optional[EvalContext] = None) -> t.Any:
588
+ return self.value
589
+
590
+ @classmethod
591
+ def from_untrusted(
592
+ cls,
593
+ value: t.Any,
594
+ lineno: t.Optional[int] = None,
595
+ environment: "t.Optional[Environment]" = None,
596
+ ) -> "Const":
597
+ """Return a const object if the value is representable as
598
+ constant value in the generated code, otherwise it will raise
599
+ an `Impossible` exception.
600
+ """
601
+ from .compiler import has_safe_repr
602
+
603
+ if not has_safe_repr(value):
604
+ raise Impossible()
605
+ return cls(value, lineno=lineno, environment=environment)
606
+
607
+
608
+ class TemplateData(Literal):
609
+ """A constant template string."""
610
+
611
+ fields = ("data",)
612
+ data: str
613
+
614
+ def as_const(self, eval_ctx: t.Optional[EvalContext] = None) -> str:
615
+ eval_ctx = get_eval_context(self, eval_ctx)
616
+ if eval_ctx.volatile:
617
+ raise Impossible()
618
+ if eval_ctx.autoescape:
619
+ return Markup(self.data)
620
+ return self.data
621
+
622
+
623
+ class Tuple(Literal):
624
+ """For loop unpacking and some other things like multiple arguments
625
+ for subscripts. Like for :class:`Name` `ctx` specifies if the tuple
626
+ is used for loading the names or storing.
627
+ """
628
+
629
+ fields = ("items", "ctx")
630
+ items: t.List[Expr]
631
+ ctx: str
632
+
633
+ def as_const(self, eval_ctx: t.Optional[EvalContext] = None) -> t.Tuple[t.Any, ...]:
634
+ eval_ctx = get_eval_context(self, eval_ctx)
635
+ return tuple(x.as_const(eval_ctx) for x in self.items)
636
+
637
+ def can_assign(self) -> bool:
638
+ for item in self.items:
639
+ if not item.can_assign():
640
+ return False
641
+ return True
642
+
643
+
644
+ class List(Literal):
645
+ """Any list literal such as ``[1, 2, 3]``"""
646
+
647
+ fields = ("items",)
648
+ items: t.List[Expr]
649
+
650
+ def as_const(self, eval_ctx: t.Optional[EvalContext] = None) -> t.List[t.Any]:
651
+ eval_ctx = get_eval_context(self, eval_ctx)
652
+ return [x.as_const(eval_ctx) for x in self.items]
653
+
654
+
655
+ class Dict(Literal):
656
+ """Any dict literal such as ``{1: 2, 3: 4}``. The items must be a list of
657
+ :class:`Pair` nodes.
658
+ """
659
+
660
+ fields = ("items",)
661
+ items: t.List["Pair"]
662
+
663
+ def as_const(
664
+ self, eval_ctx: t.Optional[EvalContext] = None
665
+ ) -> t.Dict[t.Any, t.Any]:
666
+ eval_ctx = get_eval_context(self, eval_ctx)
667
+ return dict(x.as_const(eval_ctx) for x in self.items)
668
+
669
+
670
+ class Pair(Helper):
671
+ """A key, value pair for dicts."""
672
+
673
+ fields = ("key", "value")
674
+ key: Expr
675
+ value: Expr
676
+
677
+ def as_const(
678
+ self, eval_ctx: t.Optional[EvalContext] = None
679
+ ) -> t.Tuple[t.Any, t.Any]:
680
+ eval_ctx = get_eval_context(self, eval_ctx)
681
+ return self.key.as_const(eval_ctx), self.value.as_const(eval_ctx)
682
+
683
+
684
+ class Keyword(Helper):
685
+ """A key, value pair for keyword arguments where key is a string."""
686
+
687
+ fields = ("key", "value")
688
+ key: str
689
+ value: Expr
690
+
691
+ def as_const(self, eval_ctx: t.Optional[EvalContext] = None) -> t.Tuple[str, t.Any]:
692
+ eval_ctx = get_eval_context(self, eval_ctx)
693
+ return self.key, self.value.as_const(eval_ctx)
694
+
695
+
696
+ class CondExpr(Expr):
697
+ """A conditional expression (inline if expression). (``{{
698
+ foo if bar else baz }}``)
699
+ """
700
+
701
+ fields = ("test", "expr1", "expr2")
702
+ test: Expr
703
+ expr1: Expr
704
+ expr2: t.Optional[Expr]
705
+
706
+ def as_const(self, eval_ctx: t.Optional[EvalContext] = None) -> t.Any:
707
+ eval_ctx = get_eval_context(self, eval_ctx)
708
+ if self.test.as_const(eval_ctx):
709
+ return self.expr1.as_const(eval_ctx)
710
+
711
+ # if we evaluate to an undefined object, we better do that at runtime
712
+ if self.expr2 is None:
713
+ raise Impossible()
714
+
715
+ return self.expr2.as_const(eval_ctx)
716
+
717
+
718
+ def args_as_const(
719
+ node: t.Union["_FilterTestCommon", "Call"], eval_ctx: t.Optional[EvalContext]
720
+ ) -> t.Tuple[t.List[t.Any], t.Dict[t.Any, t.Any]]:
721
+ args = [x.as_const(eval_ctx) for x in node.args]
722
+ kwargs = dict(x.as_const(eval_ctx) for x in node.kwargs)
723
+
724
+ if node.dyn_args is not None:
725
+ try:
726
+ args.extend(node.dyn_args.as_const(eval_ctx))
727
+ except Exception as e:
728
+ raise Impossible() from e
729
+
730
+ if node.dyn_kwargs is not None:
731
+ try:
732
+ kwargs.update(node.dyn_kwargs.as_const(eval_ctx))
733
+ except Exception as e:
734
+ raise Impossible() from e
735
+
736
+ return args, kwargs
737
+
738
+
739
+ class _FilterTestCommon(Expr):
740
+ fields = ("node", "name", "args", "kwargs", "dyn_args", "dyn_kwargs")
741
+ node: Expr
742
+ name: str
743
+ args: t.List[Expr]
744
+ kwargs: t.List[Pair]
745
+ dyn_args: t.Optional[Expr]
746
+ dyn_kwargs: t.Optional[Expr]
747
+ abstract = True
748
+ _is_filter = True
749
+
750
+ def as_const(self, eval_ctx: t.Optional[EvalContext] = None) -> t.Any:
751
+ eval_ctx = get_eval_context(self, eval_ctx)
752
+
753
+ if eval_ctx.volatile:
754
+ raise Impossible()
755
+
756
+ if self._is_filter:
757
+ env_map = eval_ctx.environment.filters
758
+ else:
759
+ env_map = eval_ctx.environment.tests
760
+
761
+ func = env_map.get(self.name)
762
+ pass_arg = _PassArg.from_obj(func) # type: ignore
763
+
764
+ if func is None or pass_arg is _PassArg.context:
765
+ raise Impossible()
766
+
767
+ if eval_ctx.environment.is_async and (
768
+ getattr(func, "jinja_async_variant", False) is True
769
+ or inspect.iscoroutinefunction(func)
770
+ ):
771
+ raise Impossible()
772
+
773
+ args, kwargs = args_as_const(self, eval_ctx)
774
+ args.insert(0, self.node.as_const(eval_ctx))
775
+
776
+ if pass_arg is _PassArg.eval_context:
777
+ args.insert(0, eval_ctx)
778
+ elif pass_arg is _PassArg.environment:
779
+ args.insert(0, eval_ctx.environment)
780
+
781
+ try:
782
+ return func(*args, **kwargs)
783
+ except Exception as e:
784
+ raise Impossible() from e
785
+
786
+
787
+ class Filter(_FilterTestCommon):
788
+ """Apply a filter to an expression. ``name`` is the name of the
789
+ filter, the other fields are the same as :class:`Call`.
790
+
791
+ If ``node`` is ``None``, the filter is being used in a filter block
792
+ and is applied to the content of the block.
793
+ """
794
+
795
+ node: t.Optional[Expr] # type: ignore
796
+
797
+ def as_const(self, eval_ctx: t.Optional[EvalContext] = None) -> t.Any:
798
+ if self.node is None:
799
+ raise Impossible()
800
+
801
+ return super().as_const(eval_ctx=eval_ctx)
802
+
803
+
804
+ class Test(_FilterTestCommon):
805
+ """Apply a test to an expression. ``name`` is the name of the test,
806
+ the other field are the same as :class:`Call`.
807
+
808
+ .. versionchanged:: 3.0
809
+ ``as_const`` shares the same logic for filters and tests. Tests
810
+ check for volatile, async, and ``@pass_context`` etc.
811
+ decorators.
812
+ """
813
+
814
+ _is_filter = False
815
+
816
+
817
+ class Call(Expr):
818
+ """Calls an expression. `args` is a list of arguments, `kwargs` a list
819
+ of keyword arguments (list of :class:`Keyword` nodes), and `dyn_args`
820
+ and `dyn_kwargs` has to be either `None` or a node that is used as
821
+ node for dynamic positional (``*args``) or keyword (``**kwargs``)
822
+ arguments.
823
+ """
824
+
825
+ fields = ("node", "args", "kwargs", "dyn_args", "dyn_kwargs")
826
+ node: Expr
827
+ args: t.List[Expr]
828
+ kwargs: t.List[Keyword]
829
+ dyn_args: t.Optional[Expr]
830
+ dyn_kwargs: t.Optional[Expr]
831
+
832
+
833
+ class Getitem(Expr):
834
+ """Get an attribute or item from an expression and prefer the item."""
835
+
836
+ fields = ("node", "arg", "ctx")
837
+ node: Expr
838
+ arg: Expr
839
+ ctx: str
840
+
841
+ def as_const(self, eval_ctx: t.Optional[EvalContext] = None) -> t.Any:
842
+ if self.ctx != "load":
843
+ raise Impossible()
844
+
845
+ eval_ctx = get_eval_context(self, eval_ctx)
846
+
847
+ try:
848
+ return eval_ctx.environment.getitem(
849
+ self.node.as_const(eval_ctx), self.arg.as_const(eval_ctx)
850
+ )
851
+ except Exception as e:
852
+ raise Impossible() from e
853
+
854
+
855
+ class Getattr(Expr):
856
+ """Get an attribute or item from an expression that is a ascii-only
857
+ bytestring and prefer the attribute.
858
+ """
859
+
860
+ fields = ("node", "attr", "ctx")
861
+ node: Expr
862
+ attr: str
863
+ ctx: str
864
+
865
+ def as_const(self, eval_ctx: t.Optional[EvalContext] = None) -> t.Any:
866
+ if self.ctx != "load":
867
+ raise Impossible()
868
+
869
+ eval_ctx = get_eval_context(self, eval_ctx)
870
+
871
+ try:
872
+ return eval_ctx.environment.getattr(self.node.as_const(eval_ctx), self.attr)
873
+ except Exception as e:
874
+ raise Impossible() from e
875
+
876
+
877
+ class Slice(Expr):
878
+ """Represents a slice object. This must only be used as argument for
879
+ :class:`Subscript`.
880
+ """
881
+
882
+ fields = ("start", "stop", "step")
883
+ start: t.Optional[Expr]
884
+ stop: t.Optional[Expr]
885
+ step: t.Optional[Expr]
886
+
887
+ def as_const(self, eval_ctx: t.Optional[EvalContext] = None) -> slice:
888
+ eval_ctx = get_eval_context(self, eval_ctx)
889
+
890
+ def const(obj: t.Optional[Expr]) -> t.Optional[t.Any]:
891
+ if obj is None:
892
+ return None
893
+ return obj.as_const(eval_ctx)
894
+
895
+ return slice(const(self.start), const(self.stop), const(self.step))
896
+
897
+
898
+ class Concat(Expr):
899
+ """Concatenates the list of expressions provided after converting
900
+ them to strings.
901
+ """
902
+
903
+ fields = ("nodes",)
904
+ nodes: t.List[Expr]
905
+
906
+ def as_const(self, eval_ctx: t.Optional[EvalContext] = None) -> str:
907
+ eval_ctx = get_eval_context(self, eval_ctx)
908
+ return "".join(str(x.as_const(eval_ctx)) for x in self.nodes)
909
+
910
+
911
+ class Compare(Expr):
912
+ """Compares an expression with some other expressions. `ops` must be a
913
+ list of :class:`Operand`\\s.
914
+ """
915
+
916
+ fields = ("expr", "ops")
917
+ expr: Expr
918
+ ops: t.List["Operand"]
919
+
920
+ def as_const(self, eval_ctx: t.Optional[EvalContext] = None) -> t.Any:
921
+ eval_ctx = get_eval_context(self, eval_ctx)
922
+ result = value = self.expr.as_const(eval_ctx)
923
+
924
+ try:
925
+ for op in self.ops:
926
+ new_value = op.expr.as_const(eval_ctx)
927
+ result = _cmpop_to_func[op.op](value, new_value)
928
+
929
+ if not result:
930
+ return False
931
+
932
+ value = new_value
933
+ except Exception as e:
934
+ raise Impossible() from e
935
+
936
+ return result
937
+
938
+
939
+ class Operand(Helper):
940
+ """Holds an operator and an expression."""
941
+
942
+ fields = ("op", "expr")
943
+ op: str
944
+ expr: Expr
945
+
946
+
947
+ class Mul(BinExpr):
948
+ """Multiplies the left with the right node."""
949
+
950
+ operator = "*"
951
+
952
+
953
+ class Div(BinExpr):
954
+ """Divides the left by the right node."""
955
+
956
+ operator = "/"
957
+
958
+
959
+ class FloorDiv(BinExpr):
960
+ """Divides the left by the right node and converts the
961
+ result into an integer by truncating.
962
+ """
963
+
964
+ operator = "//"
965
+
966
+
967
+ class Add(BinExpr):
968
+ """Add the left to the right node."""
969
+
970
+ operator = "+"
971
+
972
+
973
+ class Sub(BinExpr):
974
+ """Subtract the right from the left node."""
975
+
976
+ operator = "-"
977
+
978
+
979
+ class Mod(BinExpr):
980
+ """Left modulo right."""
981
+
982
+ operator = "%"
983
+
984
+
985
+ class Pow(BinExpr):
986
+ """Left to the power of right."""
987
+
988
+ operator = "**"
989
+
990
+
991
+ class And(BinExpr):
992
+ """Short circuited AND."""
993
+
994
+ operator = "and"
995
+
996
+ def as_const(self, eval_ctx: t.Optional[EvalContext] = None) -> t.Any:
997
+ eval_ctx = get_eval_context(self, eval_ctx)
998
+ return self.left.as_const(eval_ctx) and self.right.as_const(eval_ctx)
999
+
1000
+
1001
+ class Or(BinExpr):
1002
+ """Short circuited OR."""
1003
+
1004
+ operator = "or"
1005
+
1006
+ def as_const(self, eval_ctx: t.Optional[EvalContext] = None) -> t.Any:
1007
+ eval_ctx = get_eval_context(self, eval_ctx)
1008
+ return self.left.as_const(eval_ctx) or self.right.as_const(eval_ctx)
1009
+
1010
+
1011
+ class Not(UnaryExpr):
1012
+ """Negate the expression."""
1013
+
1014
+ operator = "not"
1015
+
1016
+
1017
+ class Neg(UnaryExpr):
1018
+ """Make the expression negative."""
1019
+
1020
+ operator = "-"
1021
+
1022
+
1023
+ class Pos(UnaryExpr):
1024
+ """Make the expression positive (noop for most expressions)"""
1025
+
1026
+ operator = "+"
1027
+
1028
+
1029
+ # Helpers for extensions
1030
+
1031
+
1032
+ class EnvironmentAttribute(Expr):
1033
+ """Loads an attribute from the environment object. This is useful for
1034
+ extensions that want to call a callback stored on the environment.
1035
+ """
1036
+
1037
+ fields = ("name",)
1038
+ name: str
1039
+
1040
+
1041
+ class ExtensionAttribute(Expr):
1042
+ """Returns the attribute of an extension bound to the environment.
1043
+ The identifier is the identifier of the :class:`Extension`.
1044
+
1045
+ This node is usually constructed by calling the
1046
+ :meth:`~jinja2.ext.Extension.attr` method on an extension.
1047
+ """
1048
+
1049
+ fields = ("identifier", "name")
1050
+ identifier: str
1051
+ name: str
1052
+
1053
+
1054
+ class ImportedName(Expr):
1055
+ """If created with an import name the import name is returned on node
1056
+ access. For example ``ImportedName('cgi.escape')`` returns the `escape`
1057
+ function from the cgi module on evaluation. Imports are optimized by the
1058
+ compiler so there is no need to assign them to local variables.
1059
+ """
1060
+
1061
+ fields = ("importname",)
1062
+ importname: str
1063
+
1064
+
1065
+ class InternalName(Expr):
1066
+ """An internal name in the compiler. You cannot create these nodes
1067
+ yourself but the parser provides a
1068
+ :meth:`~jinja2.parser.Parser.free_identifier` method that creates
1069
+ a new identifier for you. This identifier is not available from the
1070
+ template and is not treated specially by the compiler.
1071
+ """
1072
+
1073
+ fields = ("name",)
1074
+ name: str
1075
+
1076
+ def __init__(self) -> None:
1077
+ raise TypeError(
1078
+ "Can't create internal names. Use the "
1079
+ "`free_identifier` method on a parser."
1080
+ )
1081
+
1082
+
1083
+ class MarkSafe(Expr):
1084
+ """Mark the wrapped expression as safe (wrap it as `Markup`)."""
1085
+
1086
+ fields = ("expr",)
1087
+ expr: Expr
1088
+
1089
+ def as_const(self, eval_ctx: t.Optional[EvalContext] = None) -> Markup:
1090
+ eval_ctx = get_eval_context(self, eval_ctx)
1091
+ return Markup(self.expr.as_const(eval_ctx))
1092
+
1093
+
1094
+ class MarkSafeIfAutoescape(Expr):
1095
+ """Mark the wrapped expression as safe (wrap it as `Markup`) but
1096
+ only if autoescaping is active.
1097
+
1098
+ .. versionadded:: 2.5
1099
+ """
1100
+
1101
+ fields = ("expr",)
1102
+ expr: Expr
1103
+
1104
+ def as_const(
1105
+ self, eval_ctx: t.Optional[EvalContext] = None
1106
+ ) -> t.Union[Markup, t.Any]:
1107
+ eval_ctx = get_eval_context(self, eval_ctx)
1108
+ if eval_ctx.volatile:
1109
+ raise Impossible()
1110
+ expr = self.expr.as_const(eval_ctx)
1111
+ if eval_ctx.autoescape:
1112
+ return Markup(expr)
1113
+ return expr
1114
+
1115
+
1116
+ class ContextReference(Expr):
1117
+ """Returns the current template context. It can be used like a
1118
+ :class:`Name` node, with a ``'load'`` ctx and will return the
1119
+ current :class:`~jinja2.runtime.Context` object.
1120
+
1121
+ Here an example that assigns the current template name to a
1122
+ variable named `foo`::
1123
+
1124
+ Assign(Name('foo', ctx='store'),
1125
+ Getattr(ContextReference(), 'name'))
1126
+
1127
+ This is basically equivalent to using the
1128
+ :func:`~jinja2.pass_context` decorator when using the high-level
1129
+ API, which causes a reference to the context to be passed as the
1130
+ first argument to a function.
1131
+ """
1132
+
1133
+
1134
+ class DerivedContextReference(Expr):
1135
+ """Return the current template context including locals. Behaves
1136
+ exactly like :class:`ContextReference`, but includes local
1137
+ variables, such as from a ``for`` loop.
1138
+
1139
+ .. versionadded:: 2.11
1140
+ """
1141
+
1142
+
1143
+ class Continue(Stmt):
1144
+ """Continue a loop."""
1145
+
1146
+
1147
+ class Break(Stmt):
1148
+ """Break a loop."""
1149
+
1150
+
1151
+ class Scope(Stmt):
1152
+ """An artificial scope."""
1153
+
1154
+ fields = ("body",)
1155
+ body: t.List[Node]
1156
+
1157
+
1158
+ class OverlayScope(Stmt):
1159
+ """An overlay scope for extensions. This is a largely unoptimized scope
1160
+ that however can be used to introduce completely arbitrary variables into
1161
+ a sub scope from a dictionary or dictionary like object. The `context`
1162
+ field has to evaluate to a dictionary object.
1163
+
1164
+ Example usage::
1165
+
1166
+ OverlayScope(context=self.call_method('get_context'),
1167
+ body=[...])
1168
+
1169
+ .. versionadded:: 2.10
1170
+ """
1171
+
1172
+ fields = ("context", "body")
1173
+ context: Expr
1174
+ body: t.List[Node]
1175
+
1176
+
1177
+ class EvalContextModifier(Stmt):
1178
+ """Modifies the eval context. For each option that should be modified,
1179
+ a :class:`Keyword` has to be added to the :attr:`options` list.
1180
+
1181
+ Example to change the `autoescape` setting::
1182
+
1183
+ EvalContextModifier(options=[Keyword('autoescape', Const(True))])
1184
+ """
1185
+
1186
+ fields = ("options",)
1187
+ options: t.List[Keyword]
1188
+
1189
+
1190
+ class ScopedEvalContextModifier(EvalContextModifier):
1191
+ """Modifies the eval context and reverts it later. Works exactly like
1192
+ :class:`EvalContextModifier` but will only modify the
1193
+ :class:`~jinja2.nodes.EvalContext` for nodes in the :attr:`body`.
1194
+ """
1195
+
1196
+ fields = ("body",)
1197
+ body: t.List[Node]
1198
+
1199
+
1200
+ # make sure nobody creates custom nodes
1201
+ def _failing_new(*args: t.Any, **kwargs: t.Any) -> "te.NoReturn":
1202
+ raise TypeError("can't create custom node types")
1203
+
1204
+
1205
+ NodeType.__new__ = staticmethod(_failing_new) # type: ignore
1206
+ del _failing_new
janus/lib/python3.10/site-packages/jinja2/optimizer.py ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """The optimizer tries to constant fold expressions and modify the AST
2
+ in place so that it should be faster to evaluate.
3
+
4
+ Because the AST does not contain all the scoping information and the
5
+ compiler has to find that out, we cannot do all the optimizations we
6
+ want. For example, loop unrolling doesn't work because unrolled loops
7
+ would have a different scope. The solution would be a second syntax tree
8
+ that stored the scoping rules.
9
+ """
10
+
11
+ import typing as t
12
+
13
+ from . import nodes
14
+ from .visitor import NodeTransformer
15
+
16
+ if t.TYPE_CHECKING:
17
+ from .environment import Environment
18
+
19
+
20
+ def optimize(node: nodes.Node, environment: "Environment") -> nodes.Node:
21
+ """The context hint can be used to perform an static optimization
22
+ based on the context given."""
23
+ optimizer = Optimizer(environment)
24
+ return t.cast(nodes.Node, optimizer.visit(node))
25
+
26
+
27
+ class Optimizer(NodeTransformer):
28
+ def __init__(self, environment: "t.Optional[Environment]") -> None:
29
+ self.environment = environment
30
+
31
+ def generic_visit(
32
+ self, node: nodes.Node, *args: t.Any, **kwargs: t.Any
33
+ ) -> nodes.Node:
34
+ node = super().generic_visit(node, *args, **kwargs)
35
+
36
+ # Do constant folding. Some other nodes besides Expr have
37
+ # as_const, but folding them causes errors later on.
38
+ if isinstance(node, nodes.Expr):
39
+ try:
40
+ return nodes.Const.from_untrusted(
41
+ node.as_const(args[0] if args else None),
42
+ lineno=node.lineno,
43
+ environment=self.environment,
44
+ )
45
+ except nodes.Impossible:
46
+ pass
47
+
48
+ return node
janus/lib/python3.10/site-packages/jinja2/parser.py ADDED
@@ -0,0 +1,1049 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """Parse tokens from the lexer into nodes for the compiler."""
2
+
3
+ import typing
4
+ import typing as t
5
+
6
+ from . import nodes
7
+ from .exceptions import TemplateAssertionError
8
+ from .exceptions import TemplateSyntaxError
9
+ from .lexer import describe_token
10
+ from .lexer import describe_token_expr
11
+
12
+ if t.TYPE_CHECKING:
13
+ import typing_extensions as te
14
+
15
+ from .environment import Environment
16
+
17
+ _ImportInclude = t.TypeVar("_ImportInclude", nodes.Import, nodes.Include)
18
+ _MacroCall = t.TypeVar("_MacroCall", nodes.Macro, nodes.CallBlock)
19
+
20
+ _statement_keywords = frozenset(
21
+ [
22
+ "for",
23
+ "if",
24
+ "block",
25
+ "extends",
26
+ "print",
27
+ "macro",
28
+ "include",
29
+ "from",
30
+ "import",
31
+ "set",
32
+ "with",
33
+ "autoescape",
34
+ ]
35
+ )
36
+ _compare_operators = frozenset(["eq", "ne", "lt", "lteq", "gt", "gteq"])
37
+
38
+ _math_nodes: t.Dict[str, t.Type[nodes.Expr]] = {
39
+ "add": nodes.Add,
40
+ "sub": nodes.Sub,
41
+ "mul": nodes.Mul,
42
+ "div": nodes.Div,
43
+ "floordiv": nodes.FloorDiv,
44
+ "mod": nodes.Mod,
45
+ }
46
+
47
+
48
+ class Parser:
49
+ """This is the central parsing class Jinja uses. It's passed to
50
+ extensions and can be used to parse expressions or statements.
51
+ """
52
+
53
+ def __init__(
54
+ self,
55
+ environment: "Environment",
56
+ source: str,
57
+ name: t.Optional[str] = None,
58
+ filename: t.Optional[str] = None,
59
+ state: t.Optional[str] = None,
60
+ ) -> None:
61
+ self.environment = environment
62
+ self.stream = environment._tokenize(source, name, filename, state)
63
+ self.name = name
64
+ self.filename = filename
65
+ self.closed = False
66
+ self.extensions: t.Dict[
67
+ str, t.Callable[[Parser], t.Union[nodes.Node, t.List[nodes.Node]]]
68
+ ] = {}
69
+ for extension in environment.iter_extensions():
70
+ for tag in extension.tags:
71
+ self.extensions[tag] = extension.parse
72
+ self._last_identifier = 0
73
+ self._tag_stack: t.List[str] = []
74
+ self._end_token_stack: t.List[t.Tuple[str, ...]] = []
75
+
76
+ def fail(
77
+ self,
78
+ msg: str,
79
+ lineno: t.Optional[int] = None,
80
+ exc: t.Type[TemplateSyntaxError] = TemplateSyntaxError,
81
+ ) -> "te.NoReturn":
82
+ """Convenience method that raises `exc` with the message, passed
83
+ line number or last line number as well as the current name and
84
+ filename.
85
+ """
86
+ if lineno is None:
87
+ lineno = self.stream.current.lineno
88
+ raise exc(msg, lineno, self.name, self.filename)
89
+
90
+ def _fail_ut_eof(
91
+ self,
92
+ name: t.Optional[str],
93
+ end_token_stack: t.List[t.Tuple[str, ...]],
94
+ lineno: t.Optional[int],
95
+ ) -> "te.NoReturn":
96
+ expected: t.Set[str] = set()
97
+ for exprs in end_token_stack:
98
+ expected.update(map(describe_token_expr, exprs))
99
+ if end_token_stack:
100
+ currently_looking: t.Optional[str] = " or ".join(
101
+ map(repr, map(describe_token_expr, end_token_stack[-1]))
102
+ )
103
+ else:
104
+ currently_looking = None
105
+
106
+ if name is None:
107
+ message = ["Unexpected end of template."]
108
+ else:
109
+ message = [f"Encountered unknown tag {name!r}."]
110
+
111
+ if currently_looking:
112
+ if name is not None and name in expected:
113
+ message.append(
114
+ "You probably made a nesting mistake. Jinja is expecting this tag,"
115
+ f" but currently looking for {currently_looking}."
116
+ )
117
+ else:
118
+ message.append(
119
+ f"Jinja was looking for the following tags: {currently_looking}."
120
+ )
121
+
122
+ if self._tag_stack:
123
+ message.append(
124
+ "The innermost block that needs to be closed is"
125
+ f" {self._tag_stack[-1]!r}."
126
+ )
127
+
128
+ self.fail(" ".join(message), lineno)
129
+
130
+ def fail_unknown_tag(
131
+ self, name: str, lineno: t.Optional[int] = None
132
+ ) -> "te.NoReturn":
133
+ """Called if the parser encounters an unknown tag. Tries to fail
134
+ with a human readable error message that could help to identify
135
+ the problem.
136
+ """
137
+ self._fail_ut_eof(name, self._end_token_stack, lineno)
138
+
139
+ def fail_eof(
140
+ self,
141
+ end_tokens: t.Optional[t.Tuple[str, ...]] = None,
142
+ lineno: t.Optional[int] = None,
143
+ ) -> "te.NoReturn":
144
+ """Like fail_unknown_tag but for end of template situations."""
145
+ stack = list(self._end_token_stack)
146
+ if end_tokens is not None:
147
+ stack.append(end_tokens)
148
+ self._fail_ut_eof(None, stack, lineno)
149
+
150
+ def is_tuple_end(
151
+ self, extra_end_rules: t.Optional[t.Tuple[str, ...]] = None
152
+ ) -> bool:
153
+ """Are we at the end of a tuple?"""
154
+ if self.stream.current.type in ("variable_end", "block_end", "rparen"):
155
+ return True
156
+ elif extra_end_rules is not None:
157
+ return self.stream.current.test_any(extra_end_rules) # type: ignore
158
+ return False
159
+
160
+ def free_identifier(self, lineno: t.Optional[int] = None) -> nodes.InternalName:
161
+ """Return a new free identifier as :class:`~jinja2.nodes.InternalName`."""
162
+ self._last_identifier += 1
163
+ rv = object.__new__(nodes.InternalName)
164
+ nodes.Node.__init__(rv, f"fi{self._last_identifier}", lineno=lineno)
165
+ return rv
166
+
167
+ def parse_statement(self) -> t.Union[nodes.Node, t.List[nodes.Node]]:
168
+ """Parse a single statement."""
169
+ token = self.stream.current
170
+ if token.type != "name":
171
+ self.fail("tag name expected", token.lineno)
172
+ self._tag_stack.append(token.value)
173
+ pop_tag = True
174
+ try:
175
+ if token.value in _statement_keywords:
176
+ f = getattr(self, f"parse_{self.stream.current.value}")
177
+ return f() # type: ignore
178
+ if token.value == "call":
179
+ return self.parse_call_block()
180
+ if token.value == "filter":
181
+ return self.parse_filter_block()
182
+ ext = self.extensions.get(token.value)
183
+ if ext is not None:
184
+ return ext(self)
185
+
186
+ # did not work out, remove the token we pushed by accident
187
+ # from the stack so that the unknown tag fail function can
188
+ # produce a proper error message.
189
+ self._tag_stack.pop()
190
+ pop_tag = False
191
+ self.fail_unknown_tag(token.value, token.lineno)
192
+ finally:
193
+ if pop_tag:
194
+ self._tag_stack.pop()
195
+
196
+ def parse_statements(
197
+ self, end_tokens: t.Tuple[str, ...], drop_needle: bool = False
198
+ ) -> t.List[nodes.Node]:
199
+ """Parse multiple statements into a list until one of the end tokens
200
+ is reached. This is used to parse the body of statements as it also
201
+ parses template data if appropriate. The parser checks first if the
202
+ current token is a colon and skips it if there is one. Then it checks
203
+ for the block end and parses until if one of the `end_tokens` is
204
+ reached. Per default the active token in the stream at the end of
205
+ the call is the matched end token. If this is not wanted `drop_needle`
206
+ can be set to `True` and the end token is removed.
207
+ """
208
+ # the first token may be a colon for python compatibility
209
+ self.stream.skip_if("colon")
210
+
211
+ # in the future it would be possible to add whole code sections
212
+ # by adding some sort of end of statement token and parsing those here.
213
+ self.stream.expect("block_end")
214
+ result = self.subparse(end_tokens)
215
+
216
+ # we reached the end of the template too early, the subparser
217
+ # does not check for this, so we do that now
218
+ if self.stream.current.type == "eof":
219
+ self.fail_eof(end_tokens)
220
+
221
+ if drop_needle:
222
+ next(self.stream)
223
+ return result
224
+
225
+ def parse_set(self) -> t.Union[nodes.Assign, nodes.AssignBlock]:
226
+ """Parse an assign statement."""
227
+ lineno = next(self.stream).lineno
228
+ target = self.parse_assign_target(with_namespace=True)
229
+ if self.stream.skip_if("assign"):
230
+ expr = self.parse_tuple()
231
+ return nodes.Assign(target, expr, lineno=lineno)
232
+ filter_node = self.parse_filter(None)
233
+ body = self.parse_statements(("name:endset",), drop_needle=True)
234
+ return nodes.AssignBlock(target, filter_node, body, lineno=lineno)
235
+
236
+ def parse_for(self) -> nodes.For:
237
+ """Parse a for loop."""
238
+ lineno = self.stream.expect("name:for").lineno
239
+ target = self.parse_assign_target(extra_end_rules=("name:in",))
240
+ self.stream.expect("name:in")
241
+ iter = self.parse_tuple(
242
+ with_condexpr=False, extra_end_rules=("name:recursive",)
243
+ )
244
+ test = None
245
+ if self.stream.skip_if("name:if"):
246
+ test = self.parse_expression()
247
+ recursive = self.stream.skip_if("name:recursive")
248
+ body = self.parse_statements(("name:endfor", "name:else"))
249
+ if next(self.stream).value == "endfor":
250
+ else_ = []
251
+ else:
252
+ else_ = self.parse_statements(("name:endfor",), drop_needle=True)
253
+ return nodes.For(target, iter, body, else_, test, recursive, lineno=lineno)
254
+
255
+ def parse_if(self) -> nodes.If:
256
+ """Parse an if construct."""
257
+ node = result = nodes.If(lineno=self.stream.expect("name:if").lineno)
258
+ while True:
259
+ node.test = self.parse_tuple(with_condexpr=False)
260
+ node.body = self.parse_statements(("name:elif", "name:else", "name:endif"))
261
+ node.elif_ = []
262
+ node.else_ = []
263
+ token = next(self.stream)
264
+ if token.test("name:elif"):
265
+ node = nodes.If(lineno=self.stream.current.lineno)
266
+ result.elif_.append(node)
267
+ continue
268
+ elif token.test("name:else"):
269
+ result.else_ = self.parse_statements(("name:endif",), drop_needle=True)
270
+ break
271
+ return result
272
+
273
+ def parse_with(self) -> nodes.With:
274
+ node = nodes.With(lineno=next(self.stream).lineno)
275
+ targets: t.List[nodes.Expr] = []
276
+ values: t.List[nodes.Expr] = []
277
+ while self.stream.current.type != "block_end":
278
+ if targets:
279
+ self.stream.expect("comma")
280
+ target = self.parse_assign_target()
281
+ target.set_ctx("param")
282
+ targets.append(target)
283
+ self.stream.expect("assign")
284
+ values.append(self.parse_expression())
285
+ node.targets = targets
286
+ node.values = values
287
+ node.body = self.parse_statements(("name:endwith",), drop_needle=True)
288
+ return node
289
+
290
+ def parse_autoescape(self) -> nodes.Scope:
291
+ node = nodes.ScopedEvalContextModifier(lineno=next(self.stream).lineno)
292
+ node.options = [nodes.Keyword("autoescape", self.parse_expression())]
293
+ node.body = self.parse_statements(("name:endautoescape",), drop_needle=True)
294
+ return nodes.Scope([node])
295
+
296
+ def parse_block(self) -> nodes.Block:
297
+ node = nodes.Block(lineno=next(self.stream).lineno)
298
+ node.name = self.stream.expect("name").value
299
+ node.scoped = self.stream.skip_if("name:scoped")
300
+ node.required = self.stream.skip_if("name:required")
301
+
302
+ # common problem people encounter when switching from django
303
+ # to jinja. we do not support hyphens in block names, so let's
304
+ # raise a nicer error message in that case.
305
+ if self.stream.current.type == "sub":
306
+ self.fail(
307
+ "Block names in Jinja have to be valid Python identifiers and may not"
308
+ " contain hyphens, use an underscore instead."
309
+ )
310
+
311
+ node.body = self.parse_statements(("name:endblock",), drop_needle=True)
312
+
313
+ # enforce that required blocks only contain whitespace or comments
314
+ # by asserting that the body, if not empty, is just TemplateData nodes
315
+ # with whitespace data
316
+ if node.required:
317
+ for body_node in node.body:
318
+ if not isinstance(body_node, nodes.Output) or any(
319
+ not isinstance(output_node, nodes.TemplateData)
320
+ or not output_node.data.isspace()
321
+ for output_node in body_node.nodes
322
+ ):
323
+ self.fail("Required blocks can only contain comments or whitespace")
324
+
325
+ self.stream.skip_if("name:" + node.name)
326
+ return node
327
+
328
+ def parse_extends(self) -> nodes.Extends:
329
+ node = nodes.Extends(lineno=next(self.stream).lineno)
330
+ node.template = self.parse_expression()
331
+ return node
332
+
333
+ def parse_import_context(
334
+ self, node: _ImportInclude, default: bool
335
+ ) -> _ImportInclude:
336
+ if self.stream.current.test_any(
337
+ "name:with", "name:without"
338
+ ) and self.stream.look().test("name:context"):
339
+ node.with_context = next(self.stream).value == "with"
340
+ self.stream.skip()
341
+ else:
342
+ node.with_context = default
343
+ return node
344
+
345
+ def parse_include(self) -> nodes.Include:
346
+ node = nodes.Include(lineno=next(self.stream).lineno)
347
+ node.template = self.parse_expression()
348
+ if self.stream.current.test("name:ignore") and self.stream.look().test(
349
+ "name:missing"
350
+ ):
351
+ node.ignore_missing = True
352
+ self.stream.skip(2)
353
+ else:
354
+ node.ignore_missing = False
355
+ return self.parse_import_context(node, True)
356
+
357
+ def parse_import(self) -> nodes.Import:
358
+ node = nodes.Import(lineno=next(self.stream).lineno)
359
+ node.template = self.parse_expression()
360
+ self.stream.expect("name:as")
361
+ node.target = self.parse_assign_target(name_only=True).name
362
+ return self.parse_import_context(node, False)
363
+
364
+ def parse_from(self) -> nodes.FromImport:
365
+ node = nodes.FromImport(lineno=next(self.stream).lineno)
366
+ node.template = self.parse_expression()
367
+ self.stream.expect("name:import")
368
+ node.names = []
369
+
370
+ def parse_context() -> bool:
371
+ if self.stream.current.value in {
372
+ "with",
373
+ "without",
374
+ } and self.stream.look().test("name:context"):
375
+ node.with_context = next(self.stream).value == "with"
376
+ self.stream.skip()
377
+ return True
378
+ return False
379
+
380
+ while True:
381
+ if node.names:
382
+ self.stream.expect("comma")
383
+ if self.stream.current.type == "name":
384
+ if parse_context():
385
+ break
386
+ target = self.parse_assign_target(name_only=True)
387
+ if target.name.startswith("_"):
388
+ self.fail(
389
+ "names starting with an underline can not be imported",
390
+ target.lineno,
391
+ exc=TemplateAssertionError,
392
+ )
393
+ if self.stream.skip_if("name:as"):
394
+ alias = self.parse_assign_target(name_only=True)
395
+ node.names.append((target.name, alias.name))
396
+ else:
397
+ node.names.append(target.name)
398
+ if parse_context() or self.stream.current.type != "comma":
399
+ break
400
+ else:
401
+ self.stream.expect("name")
402
+ if not hasattr(node, "with_context"):
403
+ node.with_context = False
404
+ return node
405
+
406
+ def parse_signature(self, node: _MacroCall) -> None:
407
+ args = node.args = []
408
+ defaults = node.defaults = []
409
+ self.stream.expect("lparen")
410
+ while self.stream.current.type != "rparen":
411
+ if args:
412
+ self.stream.expect("comma")
413
+ arg = self.parse_assign_target(name_only=True)
414
+ arg.set_ctx("param")
415
+ if self.stream.skip_if("assign"):
416
+ defaults.append(self.parse_expression())
417
+ elif defaults:
418
+ self.fail("non-default argument follows default argument")
419
+ args.append(arg)
420
+ self.stream.expect("rparen")
421
+
422
+ def parse_call_block(self) -> nodes.CallBlock:
423
+ node = nodes.CallBlock(lineno=next(self.stream).lineno)
424
+ if self.stream.current.type == "lparen":
425
+ self.parse_signature(node)
426
+ else:
427
+ node.args = []
428
+ node.defaults = []
429
+
430
+ call_node = self.parse_expression()
431
+ if not isinstance(call_node, nodes.Call):
432
+ self.fail("expected call", node.lineno)
433
+ node.call = call_node
434
+ node.body = self.parse_statements(("name:endcall",), drop_needle=True)
435
+ return node
436
+
437
+ def parse_filter_block(self) -> nodes.FilterBlock:
438
+ node = nodes.FilterBlock(lineno=next(self.stream).lineno)
439
+ node.filter = self.parse_filter(None, start_inline=True) # type: ignore
440
+ node.body = self.parse_statements(("name:endfilter",), drop_needle=True)
441
+ return node
442
+
443
+ def parse_macro(self) -> nodes.Macro:
444
+ node = nodes.Macro(lineno=next(self.stream).lineno)
445
+ node.name = self.parse_assign_target(name_only=True).name
446
+ self.parse_signature(node)
447
+ node.body = self.parse_statements(("name:endmacro",), drop_needle=True)
448
+ return node
449
+
450
+ def parse_print(self) -> nodes.Output:
451
+ node = nodes.Output(lineno=next(self.stream).lineno)
452
+ node.nodes = []
453
+ while self.stream.current.type != "block_end":
454
+ if node.nodes:
455
+ self.stream.expect("comma")
456
+ node.nodes.append(self.parse_expression())
457
+ return node
458
+
459
+ @typing.overload
460
+ def parse_assign_target(
461
+ self, with_tuple: bool = ..., name_only: "te.Literal[True]" = ...
462
+ ) -> nodes.Name: ...
463
+
464
+ @typing.overload
465
+ def parse_assign_target(
466
+ self,
467
+ with_tuple: bool = True,
468
+ name_only: bool = False,
469
+ extra_end_rules: t.Optional[t.Tuple[str, ...]] = None,
470
+ with_namespace: bool = False,
471
+ ) -> t.Union[nodes.NSRef, nodes.Name, nodes.Tuple]: ...
472
+
473
+ def parse_assign_target(
474
+ self,
475
+ with_tuple: bool = True,
476
+ name_only: bool = False,
477
+ extra_end_rules: t.Optional[t.Tuple[str, ...]] = None,
478
+ with_namespace: bool = False,
479
+ ) -> t.Union[nodes.NSRef, nodes.Name, nodes.Tuple]:
480
+ """Parse an assignment target. As Jinja allows assignments to
481
+ tuples, this function can parse all allowed assignment targets. Per
482
+ default assignments to tuples are parsed, that can be disable however
483
+ by setting `with_tuple` to `False`. If only assignments to names are
484
+ wanted `name_only` can be set to `True`. The `extra_end_rules`
485
+ parameter is forwarded to the tuple parsing function. If
486
+ `with_namespace` is enabled, a namespace assignment may be parsed.
487
+ """
488
+ target: nodes.Expr
489
+
490
+ if name_only:
491
+ token = self.stream.expect("name")
492
+ target = nodes.Name(token.value, "store", lineno=token.lineno)
493
+ else:
494
+ if with_tuple:
495
+ target = self.parse_tuple(
496
+ simplified=True,
497
+ extra_end_rules=extra_end_rules,
498
+ with_namespace=with_namespace,
499
+ )
500
+ else:
501
+ target = self.parse_primary(with_namespace=with_namespace)
502
+
503
+ target.set_ctx("store")
504
+
505
+ if not target.can_assign():
506
+ self.fail(
507
+ f"can't assign to {type(target).__name__.lower()!r}", target.lineno
508
+ )
509
+
510
+ return target # type: ignore
511
+
512
+ def parse_expression(self, with_condexpr: bool = True) -> nodes.Expr:
513
+ """Parse an expression. Per default all expressions are parsed, if
514
+ the optional `with_condexpr` parameter is set to `False` conditional
515
+ expressions are not parsed.
516
+ """
517
+ if with_condexpr:
518
+ return self.parse_condexpr()
519
+ return self.parse_or()
520
+
521
+ def parse_condexpr(self) -> nodes.Expr:
522
+ lineno = self.stream.current.lineno
523
+ expr1 = self.parse_or()
524
+ expr3: t.Optional[nodes.Expr]
525
+
526
+ while self.stream.skip_if("name:if"):
527
+ expr2 = self.parse_or()
528
+ if self.stream.skip_if("name:else"):
529
+ expr3 = self.parse_condexpr()
530
+ else:
531
+ expr3 = None
532
+ expr1 = nodes.CondExpr(expr2, expr1, expr3, lineno=lineno)
533
+ lineno = self.stream.current.lineno
534
+ return expr1
535
+
536
+ def parse_or(self) -> nodes.Expr:
537
+ lineno = self.stream.current.lineno
538
+ left = self.parse_and()
539
+ while self.stream.skip_if("name:or"):
540
+ right = self.parse_and()
541
+ left = nodes.Or(left, right, lineno=lineno)
542
+ lineno = self.stream.current.lineno
543
+ return left
544
+
545
+ def parse_and(self) -> nodes.Expr:
546
+ lineno = self.stream.current.lineno
547
+ left = self.parse_not()
548
+ while self.stream.skip_if("name:and"):
549
+ right = self.parse_not()
550
+ left = nodes.And(left, right, lineno=lineno)
551
+ lineno = self.stream.current.lineno
552
+ return left
553
+
554
+ def parse_not(self) -> nodes.Expr:
555
+ if self.stream.current.test("name:not"):
556
+ lineno = next(self.stream).lineno
557
+ return nodes.Not(self.parse_not(), lineno=lineno)
558
+ return self.parse_compare()
559
+
560
+ def parse_compare(self) -> nodes.Expr:
561
+ lineno = self.stream.current.lineno
562
+ expr = self.parse_math1()
563
+ ops = []
564
+ while True:
565
+ token_type = self.stream.current.type
566
+ if token_type in _compare_operators:
567
+ next(self.stream)
568
+ ops.append(nodes.Operand(token_type, self.parse_math1()))
569
+ elif self.stream.skip_if("name:in"):
570
+ ops.append(nodes.Operand("in", self.parse_math1()))
571
+ elif self.stream.current.test("name:not") and self.stream.look().test(
572
+ "name:in"
573
+ ):
574
+ self.stream.skip(2)
575
+ ops.append(nodes.Operand("notin", self.parse_math1()))
576
+ else:
577
+ break
578
+ lineno = self.stream.current.lineno
579
+ if not ops:
580
+ return expr
581
+ return nodes.Compare(expr, ops, lineno=lineno)
582
+
583
+ def parse_math1(self) -> nodes.Expr:
584
+ lineno = self.stream.current.lineno
585
+ left = self.parse_concat()
586
+ while self.stream.current.type in ("add", "sub"):
587
+ cls = _math_nodes[self.stream.current.type]
588
+ next(self.stream)
589
+ right = self.parse_concat()
590
+ left = cls(left, right, lineno=lineno)
591
+ lineno = self.stream.current.lineno
592
+ return left
593
+
594
+ def parse_concat(self) -> nodes.Expr:
595
+ lineno = self.stream.current.lineno
596
+ args = [self.parse_math2()]
597
+ while self.stream.current.type == "tilde":
598
+ next(self.stream)
599
+ args.append(self.parse_math2())
600
+ if len(args) == 1:
601
+ return args[0]
602
+ return nodes.Concat(args, lineno=lineno)
603
+
604
+ def parse_math2(self) -> nodes.Expr:
605
+ lineno = self.stream.current.lineno
606
+ left = self.parse_pow()
607
+ while self.stream.current.type in ("mul", "div", "floordiv", "mod"):
608
+ cls = _math_nodes[self.stream.current.type]
609
+ next(self.stream)
610
+ right = self.parse_pow()
611
+ left = cls(left, right, lineno=lineno)
612
+ lineno = self.stream.current.lineno
613
+ return left
614
+
615
+ def parse_pow(self) -> nodes.Expr:
616
+ lineno = self.stream.current.lineno
617
+ left = self.parse_unary()
618
+ while self.stream.current.type == "pow":
619
+ next(self.stream)
620
+ right = self.parse_unary()
621
+ left = nodes.Pow(left, right, lineno=lineno)
622
+ lineno = self.stream.current.lineno
623
+ return left
624
+
625
+ def parse_unary(self, with_filter: bool = True) -> nodes.Expr:
626
+ token_type = self.stream.current.type
627
+ lineno = self.stream.current.lineno
628
+ node: nodes.Expr
629
+
630
+ if token_type == "sub":
631
+ next(self.stream)
632
+ node = nodes.Neg(self.parse_unary(False), lineno=lineno)
633
+ elif token_type == "add":
634
+ next(self.stream)
635
+ node = nodes.Pos(self.parse_unary(False), lineno=lineno)
636
+ else:
637
+ node = self.parse_primary()
638
+ node = self.parse_postfix(node)
639
+ if with_filter:
640
+ node = self.parse_filter_expr(node)
641
+ return node
642
+
643
+ def parse_primary(self, with_namespace: bool = False) -> nodes.Expr:
644
+ """Parse a name or literal value. If ``with_namespace`` is enabled, also
645
+ parse namespace attr refs, for use in assignments."""
646
+ token = self.stream.current
647
+ node: nodes.Expr
648
+ if token.type == "name":
649
+ next(self.stream)
650
+ if token.value in ("true", "false", "True", "False"):
651
+ node = nodes.Const(token.value in ("true", "True"), lineno=token.lineno)
652
+ elif token.value in ("none", "None"):
653
+ node = nodes.Const(None, lineno=token.lineno)
654
+ elif with_namespace and self.stream.current.type == "dot":
655
+ # If namespace attributes are allowed at this point, and the next
656
+ # token is a dot, produce a namespace reference.
657
+ next(self.stream)
658
+ attr = self.stream.expect("name")
659
+ node = nodes.NSRef(token.value, attr.value, lineno=token.lineno)
660
+ else:
661
+ node = nodes.Name(token.value, "load", lineno=token.lineno)
662
+ elif token.type == "string":
663
+ next(self.stream)
664
+ buf = [token.value]
665
+ lineno = token.lineno
666
+ while self.stream.current.type == "string":
667
+ buf.append(self.stream.current.value)
668
+ next(self.stream)
669
+ node = nodes.Const("".join(buf), lineno=lineno)
670
+ elif token.type in ("integer", "float"):
671
+ next(self.stream)
672
+ node = nodes.Const(token.value, lineno=token.lineno)
673
+ elif token.type == "lparen":
674
+ next(self.stream)
675
+ node = self.parse_tuple(explicit_parentheses=True)
676
+ self.stream.expect("rparen")
677
+ elif token.type == "lbracket":
678
+ node = self.parse_list()
679
+ elif token.type == "lbrace":
680
+ node = self.parse_dict()
681
+ else:
682
+ self.fail(f"unexpected {describe_token(token)!r}", token.lineno)
683
+ return node
684
+
685
+ def parse_tuple(
686
+ self,
687
+ simplified: bool = False,
688
+ with_condexpr: bool = True,
689
+ extra_end_rules: t.Optional[t.Tuple[str, ...]] = None,
690
+ explicit_parentheses: bool = False,
691
+ with_namespace: bool = False,
692
+ ) -> t.Union[nodes.Tuple, nodes.Expr]:
693
+ """Works like `parse_expression` but if multiple expressions are
694
+ delimited by a comma a :class:`~jinja2.nodes.Tuple` node is created.
695
+ This method could also return a regular expression instead of a tuple
696
+ if no commas where found.
697
+
698
+ The default parsing mode is a full tuple. If `simplified` is `True`
699
+ only names and literals are parsed; ``with_namespace`` allows namespace
700
+ attr refs as well. The `no_condexpr` parameter is forwarded to
701
+ :meth:`parse_expression`.
702
+
703
+ Because tuples do not require delimiters and may end in a bogus comma
704
+ an extra hint is needed that marks the end of a tuple. For example
705
+ for loops support tuples between `for` and `in`. In that case the
706
+ `extra_end_rules` is set to ``['name:in']``.
707
+
708
+ `explicit_parentheses` is true if the parsing was triggered by an
709
+ expression in parentheses. This is used to figure out if an empty
710
+ tuple is a valid expression or not.
711
+ """
712
+ lineno = self.stream.current.lineno
713
+ if simplified:
714
+
715
+ def parse() -> nodes.Expr:
716
+ return self.parse_primary(with_namespace=with_namespace)
717
+
718
+ else:
719
+
720
+ def parse() -> nodes.Expr:
721
+ return self.parse_expression(with_condexpr=with_condexpr)
722
+
723
+ args: t.List[nodes.Expr] = []
724
+ is_tuple = False
725
+
726
+ while True:
727
+ if args:
728
+ self.stream.expect("comma")
729
+ if self.is_tuple_end(extra_end_rules):
730
+ break
731
+ args.append(parse())
732
+ if self.stream.current.type == "comma":
733
+ is_tuple = True
734
+ else:
735
+ break
736
+ lineno = self.stream.current.lineno
737
+
738
+ if not is_tuple:
739
+ if args:
740
+ return args[0]
741
+
742
+ # if we don't have explicit parentheses, an empty tuple is
743
+ # not a valid expression. This would mean nothing (literally
744
+ # nothing) in the spot of an expression would be an empty
745
+ # tuple.
746
+ if not explicit_parentheses:
747
+ self.fail(
748
+ "Expected an expression,"
749
+ f" got {describe_token(self.stream.current)!r}"
750
+ )
751
+
752
+ return nodes.Tuple(args, "load", lineno=lineno)
753
+
754
+ def parse_list(self) -> nodes.List:
755
+ token = self.stream.expect("lbracket")
756
+ items: t.List[nodes.Expr] = []
757
+ while self.stream.current.type != "rbracket":
758
+ if items:
759
+ self.stream.expect("comma")
760
+ if self.stream.current.type == "rbracket":
761
+ break
762
+ items.append(self.parse_expression())
763
+ self.stream.expect("rbracket")
764
+ return nodes.List(items, lineno=token.lineno)
765
+
766
+ def parse_dict(self) -> nodes.Dict:
767
+ token = self.stream.expect("lbrace")
768
+ items: t.List[nodes.Pair] = []
769
+ while self.stream.current.type != "rbrace":
770
+ if items:
771
+ self.stream.expect("comma")
772
+ if self.stream.current.type == "rbrace":
773
+ break
774
+ key = self.parse_expression()
775
+ self.stream.expect("colon")
776
+ value = self.parse_expression()
777
+ items.append(nodes.Pair(key, value, lineno=key.lineno))
778
+ self.stream.expect("rbrace")
779
+ return nodes.Dict(items, lineno=token.lineno)
780
+
781
+ def parse_postfix(self, node: nodes.Expr) -> nodes.Expr:
782
+ while True:
783
+ token_type = self.stream.current.type
784
+ if token_type == "dot" or token_type == "lbracket":
785
+ node = self.parse_subscript(node)
786
+ # calls are valid both after postfix expressions (getattr
787
+ # and getitem) as well as filters and tests
788
+ elif token_type == "lparen":
789
+ node = self.parse_call(node)
790
+ else:
791
+ break
792
+ return node
793
+
794
+ def parse_filter_expr(self, node: nodes.Expr) -> nodes.Expr:
795
+ while True:
796
+ token_type = self.stream.current.type
797
+ if token_type == "pipe":
798
+ node = self.parse_filter(node) # type: ignore
799
+ elif token_type == "name" and self.stream.current.value == "is":
800
+ node = self.parse_test(node)
801
+ # calls are valid both after postfix expressions (getattr
802
+ # and getitem) as well as filters and tests
803
+ elif token_type == "lparen":
804
+ node = self.parse_call(node)
805
+ else:
806
+ break
807
+ return node
808
+
809
+ def parse_subscript(
810
+ self, node: nodes.Expr
811
+ ) -> t.Union[nodes.Getattr, nodes.Getitem]:
812
+ token = next(self.stream)
813
+ arg: nodes.Expr
814
+
815
+ if token.type == "dot":
816
+ attr_token = self.stream.current
817
+ next(self.stream)
818
+ if attr_token.type == "name":
819
+ return nodes.Getattr(
820
+ node, attr_token.value, "load", lineno=token.lineno
821
+ )
822
+ elif attr_token.type != "integer":
823
+ self.fail("expected name or number", attr_token.lineno)
824
+ arg = nodes.Const(attr_token.value, lineno=attr_token.lineno)
825
+ return nodes.Getitem(node, arg, "load", lineno=token.lineno)
826
+ if token.type == "lbracket":
827
+ args: t.List[nodes.Expr] = []
828
+ while self.stream.current.type != "rbracket":
829
+ if args:
830
+ self.stream.expect("comma")
831
+ args.append(self.parse_subscribed())
832
+ self.stream.expect("rbracket")
833
+ if len(args) == 1:
834
+ arg = args[0]
835
+ else:
836
+ arg = nodes.Tuple(args, "load", lineno=token.lineno)
837
+ return nodes.Getitem(node, arg, "load", lineno=token.lineno)
838
+ self.fail("expected subscript expression", token.lineno)
839
+
840
+ def parse_subscribed(self) -> nodes.Expr:
841
+ lineno = self.stream.current.lineno
842
+ args: t.List[t.Optional[nodes.Expr]]
843
+
844
+ if self.stream.current.type == "colon":
845
+ next(self.stream)
846
+ args = [None]
847
+ else:
848
+ node = self.parse_expression()
849
+ if self.stream.current.type != "colon":
850
+ return node
851
+ next(self.stream)
852
+ args = [node]
853
+
854
+ if self.stream.current.type == "colon":
855
+ args.append(None)
856
+ elif self.stream.current.type not in ("rbracket", "comma"):
857
+ args.append(self.parse_expression())
858
+ else:
859
+ args.append(None)
860
+
861
+ if self.stream.current.type == "colon":
862
+ next(self.stream)
863
+ if self.stream.current.type not in ("rbracket", "comma"):
864
+ args.append(self.parse_expression())
865
+ else:
866
+ args.append(None)
867
+ else:
868
+ args.append(None)
869
+
870
+ return nodes.Slice(lineno=lineno, *args) # noqa: B026
871
+
872
+ def parse_call_args(
873
+ self,
874
+ ) -> t.Tuple[
875
+ t.List[nodes.Expr],
876
+ t.List[nodes.Keyword],
877
+ t.Optional[nodes.Expr],
878
+ t.Optional[nodes.Expr],
879
+ ]:
880
+ token = self.stream.expect("lparen")
881
+ args = []
882
+ kwargs = []
883
+ dyn_args = None
884
+ dyn_kwargs = None
885
+ require_comma = False
886
+
887
+ def ensure(expr: bool) -> None:
888
+ if not expr:
889
+ self.fail("invalid syntax for function call expression", token.lineno)
890
+
891
+ while self.stream.current.type != "rparen":
892
+ if require_comma:
893
+ self.stream.expect("comma")
894
+
895
+ # support for trailing comma
896
+ if self.stream.current.type == "rparen":
897
+ break
898
+
899
+ if self.stream.current.type == "mul":
900
+ ensure(dyn_args is None and dyn_kwargs is None)
901
+ next(self.stream)
902
+ dyn_args = self.parse_expression()
903
+ elif self.stream.current.type == "pow":
904
+ ensure(dyn_kwargs is None)
905
+ next(self.stream)
906
+ dyn_kwargs = self.parse_expression()
907
+ else:
908
+ if (
909
+ self.stream.current.type == "name"
910
+ and self.stream.look().type == "assign"
911
+ ):
912
+ # Parsing a kwarg
913
+ ensure(dyn_kwargs is None)
914
+ key = self.stream.current.value
915
+ self.stream.skip(2)
916
+ value = self.parse_expression()
917
+ kwargs.append(nodes.Keyword(key, value, lineno=value.lineno))
918
+ else:
919
+ # Parsing an arg
920
+ ensure(dyn_args is None and dyn_kwargs is None and not kwargs)
921
+ args.append(self.parse_expression())
922
+
923
+ require_comma = True
924
+
925
+ self.stream.expect("rparen")
926
+ return args, kwargs, dyn_args, dyn_kwargs
927
+
928
+ def parse_call(self, node: nodes.Expr) -> nodes.Call:
929
+ # The lparen will be expected in parse_call_args, but the lineno
930
+ # needs to be recorded before the stream is advanced.
931
+ token = self.stream.current
932
+ args, kwargs, dyn_args, dyn_kwargs = self.parse_call_args()
933
+ return nodes.Call(node, args, kwargs, dyn_args, dyn_kwargs, lineno=token.lineno)
934
+
935
+ def parse_filter(
936
+ self, node: t.Optional[nodes.Expr], start_inline: bool = False
937
+ ) -> t.Optional[nodes.Expr]:
938
+ while self.stream.current.type == "pipe" or start_inline:
939
+ if not start_inline:
940
+ next(self.stream)
941
+ token = self.stream.expect("name")
942
+ name = token.value
943
+ while self.stream.current.type == "dot":
944
+ next(self.stream)
945
+ name += "." + self.stream.expect("name").value
946
+ if self.stream.current.type == "lparen":
947
+ args, kwargs, dyn_args, dyn_kwargs = self.parse_call_args()
948
+ else:
949
+ args = []
950
+ kwargs = []
951
+ dyn_args = dyn_kwargs = None
952
+ node = nodes.Filter(
953
+ node, name, args, kwargs, dyn_args, dyn_kwargs, lineno=token.lineno
954
+ )
955
+ start_inline = False
956
+ return node
957
+
958
+ def parse_test(self, node: nodes.Expr) -> nodes.Expr:
959
+ token = next(self.stream)
960
+ if self.stream.current.test("name:not"):
961
+ next(self.stream)
962
+ negated = True
963
+ else:
964
+ negated = False
965
+ name = self.stream.expect("name").value
966
+ while self.stream.current.type == "dot":
967
+ next(self.stream)
968
+ name += "." + self.stream.expect("name").value
969
+ dyn_args = dyn_kwargs = None
970
+ kwargs: t.List[nodes.Keyword] = []
971
+ if self.stream.current.type == "lparen":
972
+ args, kwargs, dyn_args, dyn_kwargs = self.parse_call_args()
973
+ elif self.stream.current.type in {
974
+ "name",
975
+ "string",
976
+ "integer",
977
+ "float",
978
+ "lparen",
979
+ "lbracket",
980
+ "lbrace",
981
+ } and not self.stream.current.test_any("name:else", "name:or", "name:and"):
982
+ if self.stream.current.test("name:is"):
983
+ self.fail("You cannot chain multiple tests with is")
984
+ arg_node = self.parse_primary()
985
+ arg_node = self.parse_postfix(arg_node)
986
+ args = [arg_node]
987
+ else:
988
+ args = []
989
+ node = nodes.Test(
990
+ node, name, args, kwargs, dyn_args, dyn_kwargs, lineno=token.lineno
991
+ )
992
+ if negated:
993
+ node = nodes.Not(node, lineno=token.lineno)
994
+ return node
995
+
996
+ def subparse(
997
+ self, end_tokens: t.Optional[t.Tuple[str, ...]] = None
998
+ ) -> t.List[nodes.Node]:
999
+ body: t.List[nodes.Node] = []
1000
+ data_buffer: t.List[nodes.Node] = []
1001
+ add_data = data_buffer.append
1002
+
1003
+ if end_tokens is not None:
1004
+ self._end_token_stack.append(end_tokens)
1005
+
1006
+ def flush_data() -> None:
1007
+ if data_buffer:
1008
+ lineno = data_buffer[0].lineno
1009
+ body.append(nodes.Output(data_buffer[:], lineno=lineno))
1010
+ del data_buffer[:]
1011
+
1012
+ try:
1013
+ while self.stream:
1014
+ token = self.stream.current
1015
+ if token.type == "data":
1016
+ if token.value:
1017
+ add_data(nodes.TemplateData(token.value, lineno=token.lineno))
1018
+ next(self.stream)
1019
+ elif token.type == "variable_begin":
1020
+ next(self.stream)
1021
+ add_data(self.parse_tuple(with_condexpr=True))
1022
+ self.stream.expect("variable_end")
1023
+ elif token.type == "block_begin":
1024
+ flush_data()
1025
+ next(self.stream)
1026
+ if end_tokens is not None and self.stream.current.test_any(
1027
+ *end_tokens
1028
+ ):
1029
+ return body
1030
+ rv = self.parse_statement()
1031
+ if isinstance(rv, list):
1032
+ body.extend(rv)
1033
+ else:
1034
+ body.append(rv)
1035
+ self.stream.expect("block_end")
1036
+ else:
1037
+ raise AssertionError("internal parsing error")
1038
+
1039
+ flush_data()
1040
+ finally:
1041
+ if end_tokens is not None:
1042
+ self._end_token_stack.pop()
1043
+ return body
1044
+
1045
+ def parse(self) -> nodes.Template:
1046
+ """Parse the whole template into a `Template` node."""
1047
+ result = nodes.Template(self.subparse(), lineno=1)
1048
+ result.set_environment(self.environment)
1049
+ return result
janus/lib/python3.10/site-packages/jinja2/py.typed ADDED
File without changes
janus/lib/python3.10/site-packages/jinja2/runtime.py ADDED
@@ -0,0 +1,1062 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """The runtime functions and state used by compiled templates."""
2
+
3
+ import functools
4
+ import sys
5
+ import typing as t
6
+ from collections import abc
7
+ from itertools import chain
8
+
9
+ from markupsafe import escape # noqa: F401
10
+ from markupsafe import Markup
11
+ from markupsafe import soft_str
12
+
13
+ from .async_utils import auto_aiter
14
+ from .async_utils import auto_await # noqa: F401
15
+ from .exceptions import TemplateNotFound # noqa: F401
16
+ from .exceptions import TemplateRuntimeError # noqa: F401
17
+ from .exceptions import UndefinedError
18
+ from .nodes import EvalContext
19
+ from .utils import _PassArg
20
+ from .utils import concat
21
+ from .utils import internalcode
22
+ from .utils import missing
23
+ from .utils import Namespace # noqa: F401
24
+ from .utils import object_type_repr
25
+ from .utils import pass_eval_context
26
+
27
+ V = t.TypeVar("V")
28
+ F = t.TypeVar("F", bound=t.Callable[..., t.Any])
29
+
30
+ if t.TYPE_CHECKING:
31
+ import logging
32
+
33
+ import typing_extensions as te
34
+
35
+ from .environment import Environment
36
+
37
+ class LoopRenderFunc(te.Protocol):
38
+ def __call__(
39
+ self,
40
+ reciter: t.Iterable[V],
41
+ loop_render_func: "LoopRenderFunc",
42
+ depth: int = 0,
43
+ ) -> str: ...
44
+
45
+
46
+ # these variables are exported to the template runtime
47
+ exported = [
48
+ "LoopContext",
49
+ "TemplateReference",
50
+ "Macro",
51
+ "Markup",
52
+ "TemplateRuntimeError",
53
+ "missing",
54
+ "escape",
55
+ "markup_join",
56
+ "str_join",
57
+ "identity",
58
+ "TemplateNotFound",
59
+ "Namespace",
60
+ "Undefined",
61
+ "internalcode",
62
+ ]
63
+ async_exported = [
64
+ "AsyncLoopContext",
65
+ "auto_aiter",
66
+ "auto_await",
67
+ ]
68
+
69
+
70
+ def identity(x: V) -> V:
71
+ """Returns its argument. Useful for certain things in the
72
+ environment.
73
+ """
74
+ return x
75
+
76
+
77
+ def markup_join(seq: t.Iterable[t.Any]) -> str:
78
+ """Concatenation that escapes if necessary and converts to string."""
79
+ buf = []
80
+ iterator = map(soft_str, seq)
81
+ for arg in iterator:
82
+ buf.append(arg)
83
+ if hasattr(arg, "__html__"):
84
+ return Markup("").join(chain(buf, iterator))
85
+ return concat(buf)
86
+
87
+
88
+ def str_join(seq: t.Iterable[t.Any]) -> str:
89
+ """Simple args to string conversion and concatenation."""
90
+ return concat(map(str, seq))
91
+
92
+
93
+ def new_context(
94
+ environment: "Environment",
95
+ template_name: t.Optional[str],
96
+ blocks: t.Dict[str, t.Callable[["Context"], t.Iterator[str]]],
97
+ vars: t.Optional[t.Dict[str, t.Any]] = None,
98
+ shared: bool = False,
99
+ globals: t.Optional[t.MutableMapping[str, t.Any]] = None,
100
+ locals: t.Optional[t.Mapping[str, t.Any]] = None,
101
+ ) -> "Context":
102
+ """Internal helper for context creation."""
103
+ if vars is None:
104
+ vars = {}
105
+ if shared:
106
+ parent = vars
107
+ else:
108
+ parent = dict(globals or (), **vars)
109
+ if locals:
110
+ # if the parent is shared a copy should be created because
111
+ # we don't want to modify the dict passed
112
+ if shared:
113
+ parent = dict(parent)
114
+ for key, value in locals.items():
115
+ if value is not missing:
116
+ parent[key] = value
117
+ return environment.context_class(
118
+ environment, parent, template_name, blocks, globals=globals
119
+ )
120
+
121
+
122
+ class TemplateReference:
123
+ """The `self` in templates."""
124
+
125
+ def __init__(self, context: "Context") -> None:
126
+ self.__context = context
127
+
128
+ def __getitem__(self, name: str) -> t.Any:
129
+ blocks = self.__context.blocks[name]
130
+ return BlockReference(name, self.__context, blocks, 0)
131
+
132
+ def __repr__(self) -> str:
133
+ return f"<{type(self).__name__} {self.__context.name!r}>"
134
+
135
+
136
+ def _dict_method_all(dict_method: F) -> F:
137
+ @functools.wraps(dict_method)
138
+ def f_all(self: "Context") -> t.Any:
139
+ return dict_method(self.get_all())
140
+
141
+ return t.cast(F, f_all)
142
+
143
+
144
+ @abc.Mapping.register
145
+ class Context:
146
+ """The template context holds the variables of a template. It stores the
147
+ values passed to the template and also the names the template exports.
148
+ Creating instances is neither supported nor useful as it's created
149
+ automatically at various stages of the template evaluation and should not
150
+ be created by hand.
151
+
152
+ The context is immutable. Modifications on :attr:`parent` **must not**
153
+ happen and modifications on :attr:`vars` are allowed from generated
154
+ template code only. Template filters and global functions marked as
155
+ :func:`pass_context` get the active context passed as first argument
156
+ and are allowed to access the context read-only.
157
+
158
+ The template context supports read only dict operations (`get`,
159
+ `keys`, `values`, `items`, `iterkeys`, `itervalues`, `iteritems`,
160
+ `__getitem__`, `__contains__`). Additionally there is a :meth:`resolve`
161
+ method that doesn't fail with a `KeyError` but returns an
162
+ :class:`Undefined` object for missing variables.
163
+ """
164
+
165
+ def __init__(
166
+ self,
167
+ environment: "Environment",
168
+ parent: t.Dict[str, t.Any],
169
+ name: t.Optional[str],
170
+ blocks: t.Dict[str, t.Callable[["Context"], t.Iterator[str]]],
171
+ globals: t.Optional[t.MutableMapping[str, t.Any]] = None,
172
+ ):
173
+ self.parent = parent
174
+ self.vars: t.Dict[str, t.Any] = {}
175
+ self.environment: Environment = environment
176
+ self.eval_ctx = EvalContext(self.environment, name)
177
+ self.exported_vars: t.Set[str] = set()
178
+ self.name = name
179
+ self.globals_keys = set() if globals is None else set(globals)
180
+
181
+ # create the initial mapping of blocks. Whenever template inheritance
182
+ # takes place the runtime will update this mapping with the new blocks
183
+ # from the template.
184
+ self.blocks = {k: [v] for k, v in blocks.items()}
185
+
186
+ def super(
187
+ self, name: str, current: t.Callable[["Context"], t.Iterator[str]]
188
+ ) -> t.Union["BlockReference", "Undefined"]:
189
+ """Render a parent block."""
190
+ try:
191
+ blocks = self.blocks[name]
192
+ index = blocks.index(current) + 1
193
+ blocks[index]
194
+ except LookupError:
195
+ return self.environment.undefined(
196
+ f"there is no parent block called {name!r}.", name="super"
197
+ )
198
+ return BlockReference(name, self, blocks, index)
199
+
200
+ def get(self, key: str, default: t.Any = None) -> t.Any:
201
+ """Look up a variable by name, or return a default if the key is
202
+ not found.
203
+
204
+ :param key: The variable name to look up.
205
+ :param default: The value to return if the key is not found.
206
+ """
207
+ try:
208
+ return self[key]
209
+ except KeyError:
210
+ return default
211
+
212
+ def resolve(self, key: str) -> t.Union[t.Any, "Undefined"]:
213
+ """Look up a variable by name, or return an :class:`Undefined`
214
+ object if the key is not found.
215
+
216
+ If you need to add custom behavior, override
217
+ :meth:`resolve_or_missing`, not this method. The various lookup
218
+ functions use that method, not this one.
219
+
220
+ :param key: The variable name to look up.
221
+ """
222
+ rv = self.resolve_or_missing(key)
223
+
224
+ if rv is missing:
225
+ return self.environment.undefined(name=key)
226
+
227
+ return rv
228
+
229
+ def resolve_or_missing(self, key: str) -> t.Any:
230
+ """Look up a variable by name, or return a ``missing`` sentinel
231
+ if the key is not found.
232
+
233
+ Override this method to add custom lookup behavior.
234
+ :meth:`resolve`, :meth:`get`, and :meth:`__getitem__` use this
235
+ method. Don't call this method directly.
236
+
237
+ :param key: The variable name to look up.
238
+ """
239
+ if key in self.vars:
240
+ return self.vars[key]
241
+
242
+ if key in self.parent:
243
+ return self.parent[key]
244
+
245
+ return missing
246
+
247
+ def get_exported(self) -> t.Dict[str, t.Any]:
248
+ """Get a new dict with the exported variables."""
249
+ return {k: self.vars[k] for k in self.exported_vars}
250
+
251
+ def get_all(self) -> t.Dict[str, t.Any]:
252
+ """Return the complete context as dict including the exported
253
+ variables. For optimizations reasons this might not return an
254
+ actual copy so be careful with using it.
255
+ """
256
+ if not self.vars:
257
+ return self.parent
258
+ if not self.parent:
259
+ return self.vars
260
+ return dict(self.parent, **self.vars)
261
+
262
+ @internalcode
263
+ def call(
264
+ __self,
265
+ __obj: t.Callable[..., t.Any],
266
+ *args: t.Any,
267
+ **kwargs: t.Any, # noqa: B902
268
+ ) -> t.Union[t.Any, "Undefined"]:
269
+ """Call the callable with the arguments and keyword arguments
270
+ provided but inject the active context or environment as first
271
+ argument if the callable has :func:`pass_context` or
272
+ :func:`pass_environment`.
273
+ """
274
+ if __debug__:
275
+ __traceback_hide__ = True # noqa
276
+
277
+ # Allow callable classes to take a context
278
+ if (
279
+ hasattr(__obj, "__call__") # noqa: B004
280
+ and _PassArg.from_obj(__obj.__call__) is not None
281
+ ):
282
+ __obj = __obj.__call__
283
+
284
+ pass_arg = _PassArg.from_obj(__obj)
285
+
286
+ if pass_arg is _PassArg.context:
287
+ # the active context should have access to variables set in
288
+ # loops and blocks without mutating the context itself
289
+ if kwargs.get("_loop_vars"):
290
+ __self = __self.derived(kwargs["_loop_vars"])
291
+ if kwargs.get("_block_vars"):
292
+ __self = __self.derived(kwargs["_block_vars"])
293
+ args = (__self,) + args
294
+ elif pass_arg is _PassArg.eval_context:
295
+ args = (__self.eval_ctx,) + args
296
+ elif pass_arg is _PassArg.environment:
297
+ args = (__self.environment,) + args
298
+
299
+ kwargs.pop("_block_vars", None)
300
+ kwargs.pop("_loop_vars", None)
301
+
302
+ try:
303
+ return __obj(*args, **kwargs)
304
+ except StopIteration:
305
+ return __self.environment.undefined(
306
+ "value was undefined because a callable raised a"
307
+ " StopIteration exception"
308
+ )
309
+
310
+ def derived(self, locals: t.Optional[t.Dict[str, t.Any]] = None) -> "Context":
311
+ """Internal helper function to create a derived context. This is
312
+ used in situations where the system needs a new context in the same
313
+ template that is independent.
314
+ """
315
+ context = new_context(
316
+ self.environment, self.name, {}, self.get_all(), True, None, locals
317
+ )
318
+ context.eval_ctx = self.eval_ctx
319
+ context.blocks.update((k, list(v)) for k, v in self.blocks.items())
320
+ return context
321
+
322
+ keys = _dict_method_all(dict.keys)
323
+ values = _dict_method_all(dict.values)
324
+ items = _dict_method_all(dict.items)
325
+
326
+ def __contains__(self, name: str) -> bool:
327
+ return name in self.vars or name in self.parent
328
+
329
+ def __getitem__(self, key: str) -> t.Any:
330
+ """Look up a variable by name with ``[]`` syntax, or raise a
331
+ ``KeyError`` if the key is not found.
332
+ """
333
+ item = self.resolve_or_missing(key)
334
+
335
+ if item is missing:
336
+ raise KeyError(key)
337
+
338
+ return item
339
+
340
+ def __repr__(self) -> str:
341
+ return f"<{type(self).__name__} {self.get_all()!r} of {self.name!r}>"
342
+
343
+
344
+ class BlockReference:
345
+ """One block on a template reference."""
346
+
347
+ def __init__(
348
+ self,
349
+ name: str,
350
+ context: "Context",
351
+ stack: t.List[t.Callable[["Context"], t.Iterator[str]]],
352
+ depth: int,
353
+ ) -> None:
354
+ self.name = name
355
+ self._context = context
356
+ self._stack = stack
357
+ self._depth = depth
358
+
359
+ @property
360
+ def super(self) -> t.Union["BlockReference", "Undefined"]:
361
+ """Super the block."""
362
+ if self._depth + 1 >= len(self._stack):
363
+ return self._context.environment.undefined(
364
+ f"there is no parent block called {self.name!r}.", name="super"
365
+ )
366
+ return BlockReference(self.name, self._context, self._stack, self._depth + 1)
367
+
368
+ @internalcode
369
+ async def _async_call(self) -> str:
370
+ rv = self._context.environment.concat( # type: ignore
371
+ [x async for x in self._stack[self._depth](self._context)] # type: ignore
372
+ )
373
+
374
+ if self._context.eval_ctx.autoescape:
375
+ return Markup(rv)
376
+
377
+ return rv
378
+
379
+ @internalcode
380
+ def __call__(self) -> str:
381
+ if self._context.environment.is_async:
382
+ return self._async_call() # type: ignore
383
+
384
+ rv = self._context.environment.concat( # type: ignore
385
+ self._stack[self._depth](self._context)
386
+ )
387
+
388
+ if self._context.eval_ctx.autoescape:
389
+ return Markup(rv)
390
+
391
+ return rv
392
+
393
+
394
+ class LoopContext:
395
+ """A wrapper iterable for dynamic ``for`` loops, with information
396
+ about the loop and iteration.
397
+ """
398
+
399
+ #: Current iteration of the loop, starting at 0.
400
+ index0 = -1
401
+
402
+ _length: t.Optional[int] = None
403
+ _after: t.Any = missing
404
+ _current: t.Any = missing
405
+ _before: t.Any = missing
406
+ _last_changed_value: t.Any = missing
407
+
408
+ def __init__(
409
+ self,
410
+ iterable: t.Iterable[V],
411
+ undefined: t.Type["Undefined"],
412
+ recurse: t.Optional["LoopRenderFunc"] = None,
413
+ depth0: int = 0,
414
+ ) -> None:
415
+ """
416
+ :param iterable: Iterable to wrap.
417
+ :param undefined: :class:`Undefined` class to use for next and
418
+ previous items.
419
+ :param recurse: The function to render the loop body when the
420
+ loop is marked recursive.
421
+ :param depth0: Incremented when looping recursively.
422
+ """
423
+ self._iterable = iterable
424
+ self._iterator = self._to_iterator(iterable)
425
+ self._undefined = undefined
426
+ self._recurse = recurse
427
+ #: How many levels deep a recursive loop currently is, starting at 0.
428
+ self.depth0 = depth0
429
+
430
+ @staticmethod
431
+ def _to_iterator(iterable: t.Iterable[V]) -> t.Iterator[V]:
432
+ return iter(iterable)
433
+
434
+ @property
435
+ def length(self) -> int:
436
+ """Length of the iterable.
437
+
438
+ If the iterable is a generator or otherwise does not have a
439
+ size, it is eagerly evaluated to get a size.
440
+ """
441
+ if self._length is not None:
442
+ return self._length
443
+
444
+ try:
445
+ self._length = len(self._iterable) # type: ignore
446
+ except TypeError:
447
+ iterable = list(self._iterator)
448
+ self._iterator = self._to_iterator(iterable)
449
+ self._length = len(iterable) + self.index + (self._after is not missing)
450
+
451
+ return self._length
452
+
453
+ def __len__(self) -> int:
454
+ return self.length
455
+
456
+ @property
457
+ def depth(self) -> int:
458
+ """How many levels deep a recursive loop currently is, starting at 1."""
459
+ return self.depth0 + 1
460
+
461
+ @property
462
+ def index(self) -> int:
463
+ """Current iteration of the loop, starting at 1."""
464
+ return self.index0 + 1
465
+
466
+ @property
467
+ def revindex0(self) -> int:
468
+ """Number of iterations from the end of the loop, ending at 0.
469
+
470
+ Requires calculating :attr:`length`.
471
+ """
472
+ return self.length - self.index
473
+
474
+ @property
475
+ def revindex(self) -> int:
476
+ """Number of iterations from the end of the loop, ending at 1.
477
+
478
+ Requires calculating :attr:`length`.
479
+ """
480
+ return self.length - self.index0
481
+
482
+ @property
483
+ def first(self) -> bool:
484
+ """Whether this is the first iteration of the loop."""
485
+ return self.index0 == 0
486
+
487
+ def _peek_next(self) -> t.Any:
488
+ """Return the next element in the iterable, or :data:`missing`
489
+ if the iterable is exhausted. Only peeks one item ahead, caching
490
+ the result in :attr:`_last` for use in subsequent checks. The
491
+ cache is reset when :meth:`__next__` is called.
492
+ """
493
+ if self._after is not missing:
494
+ return self._after
495
+
496
+ self._after = next(self._iterator, missing)
497
+ return self._after
498
+
499
+ @property
500
+ def last(self) -> bool:
501
+ """Whether this is the last iteration of the loop.
502
+
503
+ Causes the iterable to advance early. See
504
+ :func:`itertools.groupby` for issues this can cause.
505
+ The :func:`groupby` filter avoids that issue.
506
+ """
507
+ return self._peek_next() is missing
508
+
509
+ @property
510
+ def previtem(self) -> t.Union[t.Any, "Undefined"]:
511
+ """The item in the previous iteration. Undefined during the
512
+ first iteration.
513
+ """
514
+ if self.first:
515
+ return self._undefined("there is no previous item")
516
+
517
+ return self._before
518
+
519
+ @property
520
+ def nextitem(self) -> t.Union[t.Any, "Undefined"]:
521
+ """The item in the next iteration. Undefined during the last
522
+ iteration.
523
+
524
+ Causes the iterable to advance early. See
525
+ :func:`itertools.groupby` for issues this can cause.
526
+ The :func:`jinja-filters.groupby` filter avoids that issue.
527
+ """
528
+ rv = self._peek_next()
529
+
530
+ if rv is missing:
531
+ return self._undefined("there is no next item")
532
+
533
+ return rv
534
+
535
+ def cycle(self, *args: V) -> V:
536
+ """Return a value from the given args, cycling through based on
537
+ the current :attr:`index0`.
538
+
539
+ :param args: One or more values to cycle through.
540
+ """
541
+ if not args:
542
+ raise TypeError("no items for cycling given")
543
+
544
+ return args[self.index0 % len(args)]
545
+
546
+ def changed(self, *value: t.Any) -> bool:
547
+ """Return ``True`` if previously called with a different value
548
+ (including when called for the first time).
549
+
550
+ :param value: One or more values to compare to the last call.
551
+ """
552
+ if self._last_changed_value != value:
553
+ self._last_changed_value = value
554
+ return True
555
+
556
+ return False
557
+
558
+ def __iter__(self) -> "LoopContext":
559
+ return self
560
+
561
+ def __next__(self) -> t.Tuple[t.Any, "LoopContext"]:
562
+ if self._after is not missing:
563
+ rv = self._after
564
+ self._after = missing
565
+ else:
566
+ rv = next(self._iterator)
567
+
568
+ self.index0 += 1
569
+ self._before = self._current
570
+ self._current = rv
571
+ return rv, self
572
+
573
+ @internalcode
574
+ def __call__(self, iterable: t.Iterable[V]) -> str:
575
+ """When iterating over nested data, render the body of the loop
576
+ recursively with the given inner iterable data.
577
+
578
+ The loop must have the ``recursive`` marker for this to work.
579
+ """
580
+ if self._recurse is None:
581
+ raise TypeError(
582
+ "The loop must have the 'recursive' marker to be called recursively."
583
+ )
584
+
585
+ return self._recurse(iterable, self._recurse, depth=self.depth)
586
+
587
+ def __repr__(self) -> str:
588
+ return f"<{type(self).__name__} {self.index}/{self.length}>"
589
+
590
+
591
+ class AsyncLoopContext(LoopContext):
592
+ _iterator: t.AsyncIterator[t.Any] # type: ignore
593
+
594
+ @staticmethod
595
+ def _to_iterator( # type: ignore
596
+ iterable: t.Union[t.Iterable[V], t.AsyncIterable[V]],
597
+ ) -> t.AsyncIterator[V]:
598
+ return auto_aiter(iterable)
599
+
600
+ @property
601
+ async def length(self) -> int: # type: ignore
602
+ if self._length is not None:
603
+ return self._length
604
+
605
+ try:
606
+ self._length = len(self._iterable) # type: ignore
607
+ except TypeError:
608
+ iterable = [x async for x in self._iterator]
609
+ self._iterator = self._to_iterator(iterable)
610
+ self._length = len(iterable) + self.index + (self._after is not missing)
611
+
612
+ return self._length
613
+
614
+ @property
615
+ async def revindex0(self) -> int: # type: ignore
616
+ return await self.length - self.index
617
+
618
+ @property
619
+ async def revindex(self) -> int: # type: ignore
620
+ return await self.length - self.index0
621
+
622
+ async def _peek_next(self) -> t.Any:
623
+ if self._after is not missing:
624
+ return self._after
625
+
626
+ try:
627
+ self._after = await self._iterator.__anext__()
628
+ except StopAsyncIteration:
629
+ self._after = missing
630
+
631
+ return self._after
632
+
633
+ @property
634
+ async def last(self) -> bool: # type: ignore
635
+ return await self._peek_next() is missing
636
+
637
+ @property
638
+ async def nextitem(self) -> t.Union[t.Any, "Undefined"]:
639
+ rv = await self._peek_next()
640
+
641
+ if rv is missing:
642
+ return self._undefined("there is no next item")
643
+
644
+ return rv
645
+
646
+ def __aiter__(self) -> "AsyncLoopContext":
647
+ return self
648
+
649
+ async def __anext__(self) -> t.Tuple[t.Any, "AsyncLoopContext"]:
650
+ if self._after is not missing:
651
+ rv = self._after
652
+ self._after = missing
653
+ else:
654
+ rv = await self._iterator.__anext__()
655
+
656
+ self.index0 += 1
657
+ self._before = self._current
658
+ self._current = rv
659
+ return rv, self
660
+
661
+
662
+ class Macro:
663
+ """Wraps a macro function."""
664
+
665
+ def __init__(
666
+ self,
667
+ environment: "Environment",
668
+ func: t.Callable[..., str],
669
+ name: str,
670
+ arguments: t.List[str],
671
+ catch_kwargs: bool,
672
+ catch_varargs: bool,
673
+ caller: bool,
674
+ default_autoescape: t.Optional[bool] = None,
675
+ ):
676
+ self._environment = environment
677
+ self._func = func
678
+ self._argument_count = len(arguments)
679
+ self.name = name
680
+ self.arguments = arguments
681
+ self.catch_kwargs = catch_kwargs
682
+ self.catch_varargs = catch_varargs
683
+ self.caller = caller
684
+ self.explicit_caller = "caller" in arguments
685
+
686
+ if default_autoescape is None:
687
+ if callable(environment.autoescape):
688
+ default_autoescape = environment.autoescape(None)
689
+ else:
690
+ default_autoescape = environment.autoescape
691
+
692
+ self._default_autoescape = default_autoescape
693
+
694
+ @internalcode
695
+ @pass_eval_context
696
+ def __call__(self, *args: t.Any, **kwargs: t.Any) -> str:
697
+ # This requires a bit of explanation, In the past we used to
698
+ # decide largely based on compile-time information if a macro is
699
+ # safe or unsafe. While there was a volatile mode it was largely
700
+ # unused for deciding on escaping. This turns out to be
701
+ # problematic for macros because whether a macro is safe depends not
702
+ # on the escape mode when it was defined, but rather when it was used.
703
+ #
704
+ # Because however we export macros from the module system and
705
+ # there are historic callers that do not pass an eval context (and
706
+ # will continue to not pass one), we need to perform an instance
707
+ # check here.
708
+ #
709
+ # This is considered safe because an eval context is not a valid
710
+ # argument to callables otherwise anyway. Worst case here is
711
+ # that if no eval context is passed we fall back to the compile
712
+ # time autoescape flag.
713
+ if args and isinstance(args[0], EvalContext):
714
+ autoescape = args[0].autoescape
715
+ args = args[1:]
716
+ else:
717
+ autoescape = self._default_autoescape
718
+
719
+ # try to consume the positional arguments
720
+ arguments = list(args[: self._argument_count])
721
+ off = len(arguments)
722
+
723
+ # For information why this is necessary refer to the handling
724
+ # of caller in the `macro_body` handler in the compiler.
725
+ found_caller = False
726
+
727
+ # if the number of arguments consumed is not the number of
728
+ # arguments expected we start filling in keyword arguments
729
+ # and defaults.
730
+ if off != self._argument_count:
731
+ for name in self.arguments[len(arguments) :]:
732
+ try:
733
+ value = kwargs.pop(name)
734
+ except KeyError:
735
+ value = missing
736
+ if name == "caller":
737
+ found_caller = True
738
+ arguments.append(value)
739
+ else:
740
+ found_caller = self.explicit_caller
741
+
742
+ # it's important that the order of these arguments does not change
743
+ # if not also changed in the compiler's `function_scoping` method.
744
+ # the order is caller, keyword arguments, positional arguments!
745
+ if self.caller and not found_caller:
746
+ caller = kwargs.pop("caller", None)
747
+ if caller is None:
748
+ caller = self._environment.undefined("No caller defined", name="caller")
749
+ arguments.append(caller)
750
+
751
+ if self.catch_kwargs:
752
+ arguments.append(kwargs)
753
+ elif kwargs:
754
+ if "caller" in kwargs:
755
+ raise TypeError(
756
+ f"macro {self.name!r} was invoked with two values for the special"
757
+ " caller argument. This is most likely a bug."
758
+ )
759
+ raise TypeError(
760
+ f"macro {self.name!r} takes no keyword argument {next(iter(kwargs))!r}"
761
+ )
762
+ if self.catch_varargs:
763
+ arguments.append(args[self._argument_count :])
764
+ elif len(args) > self._argument_count:
765
+ raise TypeError(
766
+ f"macro {self.name!r} takes not more than"
767
+ f" {len(self.arguments)} argument(s)"
768
+ )
769
+
770
+ return self._invoke(arguments, autoescape)
771
+
772
+ async def _async_invoke(self, arguments: t.List[t.Any], autoescape: bool) -> str:
773
+ rv = await self._func(*arguments) # type: ignore
774
+
775
+ if autoescape:
776
+ return Markup(rv)
777
+
778
+ return rv # type: ignore
779
+
780
+ def _invoke(self, arguments: t.List[t.Any], autoescape: bool) -> str:
781
+ if self._environment.is_async:
782
+ return self._async_invoke(arguments, autoescape) # type: ignore
783
+
784
+ rv = self._func(*arguments)
785
+
786
+ if autoescape:
787
+ rv = Markup(rv)
788
+
789
+ return rv
790
+
791
+ def __repr__(self) -> str:
792
+ name = "anonymous" if self.name is None else repr(self.name)
793
+ return f"<{type(self).__name__} {name}>"
794
+
795
+
796
+ class Undefined:
797
+ """The default undefined type. This can be printed, iterated, and treated as
798
+ a boolean. Any other operation will raise an :exc:`UndefinedError`.
799
+
800
+ >>> foo = Undefined(name='foo')
801
+ >>> str(foo)
802
+ ''
803
+ >>> not foo
804
+ True
805
+ >>> foo + 42
806
+ Traceback (most recent call last):
807
+ ...
808
+ jinja2.exceptions.UndefinedError: 'foo' is undefined
809
+ """
810
+
811
+ __slots__ = (
812
+ "_undefined_hint",
813
+ "_undefined_obj",
814
+ "_undefined_name",
815
+ "_undefined_exception",
816
+ )
817
+
818
+ def __init__(
819
+ self,
820
+ hint: t.Optional[str] = None,
821
+ obj: t.Any = missing,
822
+ name: t.Optional[str] = None,
823
+ exc: t.Type[TemplateRuntimeError] = UndefinedError,
824
+ ) -> None:
825
+ self._undefined_hint = hint
826
+ self._undefined_obj = obj
827
+ self._undefined_name = name
828
+ self._undefined_exception = exc
829
+
830
+ @property
831
+ def _undefined_message(self) -> str:
832
+ """Build a message about the undefined value based on how it was
833
+ accessed.
834
+ """
835
+ if self._undefined_hint:
836
+ return self._undefined_hint
837
+
838
+ if self._undefined_obj is missing:
839
+ return f"{self._undefined_name!r} is undefined"
840
+
841
+ if not isinstance(self._undefined_name, str):
842
+ return (
843
+ f"{object_type_repr(self._undefined_obj)} has no"
844
+ f" element {self._undefined_name!r}"
845
+ )
846
+
847
+ return (
848
+ f"{object_type_repr(self._undefined_obj)!r} has no"
849
+ f" attribute {self._undefined_name!r}"
850
+ )
851
+
852
+ @internalcode
853
+ def _fail_with_undefined_error(
854
+ self, *args: t.Any, **kwargs: t.Any
855
+ ) -> "te.NoReturn":
856
+ """Raise an :exc:`UndefinedError` when operations are performed
857
+ on the undefined value.
858
+ """
859
+ raise self._undefined_exception(self._undefined_message)
860
+
861
+ @internalcode
862
+ def __getattr__(self, name: str) -> t.Any:
863
+ # Raise AttributeError on requests for names that appear to be unimplemented
864
+ # dunder methods to keep Python's internal protocol probing behaviors working
865
+ # properly in cases where another exception type could cause unexpected or
866
+ # difficult-to-diagnose failures.
867
+ if name[:2] == "__" and name[-2:] == "__":
868
+ raise AttributeError(name)
869
+
870
+ return self._fail_with_undefined_error()
871
+
872
+ __add__ = __radd__ = __sub__ = __rsub__ = _fail_with_undefined_error
873
+ __mul__ = __rmul__ = __div__ = __rdiv__ = _fail_with_undefined_error
874
+ __truediv__ = __rtruediv__ = _fail_with_undefined_error
875
+ __floordiv__ = __rfloordiv__ = _fail_with_undefined_error
876
+ __mod__ = __rmod__ = _fail_with_undefined_error
877
+ __pos__ = __neg__ = _fail_with_undefined_error
878
+ __call__ = __getitem__ = _fail_with_undefined_error
879
+ __lt__ = __le__ = __gt__ = __ge__ = _fail_with_undefined_error
880
+ __int__ = __float__ = __complex__ = _fail_with_undefined_error
881
+ __pow__ = __rpow__ = _fail_with_undefined_error
882
+
883
+ def __eq__(self, other: t.Any) -> bool:
884
+ return type(self) is type(other)
885
+
886
+ def __ne__(self, other: t.Any) -> bool:
887
+ return not self.__eq__(other)
888
+
889
+ def __hash__(self) -> int:
890
+ return id(type(self))
891
+
892
+ def __str__(self) -> str:
893
+ return ""
894
+
895
+ def __len__(self) -> int:
896
+ return 0
897
+
898
+ def __iter__(self) -> t.Iterator[t.Any]:
899
+ yield from ()
900
+
901
+ async def __aiter__(self) -> t.AsyncIterator[t.Any]:
902
+ for _ in ():
903
+ yield
904
+
905
+ def __bool__(self) -> bool:
906
+ return False
907
+
908
+ def __repr__(self) -> str:
909
+ return "Undefined"
910
+
911
+
912
+ def make_logging_undefined(
913
+ logger: t.Optional["logging.Logger"] = None, base: t.Type[Undefined] = Undefined
914
+ ) -> t.Type[Undefined]:
915
+ """Given a logger object this returns a new undefined class that will
916
+ log certain failures. It will log iterations and printing. If no
917
+ logger is given a default logger is created.
918
+
919
+ Example::
920
+
921
+ logger = logging.getLogger(__name__)
922
+ LoggingUndefined = make_logging_undefined(
923
+ logger=logger,
924
+ base=Undefined
925
+ )
926
+
927
+ .. versionadded:: 2.8
928
+
929
+ :param logger: the logger to use. If not provided, a default logger
930
+ is created.
931
+ :param base: the base class to add logging functionality to. This
932
+ defaults to :class:`Undefined`.
933
+ """
934
+ if logger is None:
935
+ import logging
936
+
937
+ logger = logging.getLogger(__name__)
938
+ logger.addHandler(logging.StreamHandler(sys.stderr))
939
+
940
+ def _log_message(undef: Undefined) -> None:
941
+ logger.warning("Template variable warning: %s", undef._undefined_message)
942
+
943
+ class LoggingUndefined(base): # type: ignore
944
+ __slots__ = ()
945
+
946
+ def _fail_with_undefined_error( # type: ignore
947
+ self, *args: t.Any, **kwargs: t.Any
948
+ ) -> "te.NoReturn":
949
+ try:
950
+ super()._fail_with_undefined_error(*args, **kwargs)
951
+ except self._undefined_exception as e:
952
+ logger.error("Template variable error: %s", e) # type: ignore
953
+ raise e
954
+
955
+ def __str__(self) -> str:
956
+ _log_message(self)
957
+ return super().__str__() # type: ignore
958
+
959
+ def __iter__(self) -> t.Iterator[t.Any]:
960
+ _log_message(self)
961
+ return super().__iter__() # type: ignore
962
+
963
+ def __bool__(self) -> bool:
964
+ _log_message(self)
965
+ return super().__bool__() # type: ignore
966
+
967
+ return LoggingUndefined
968
+
969
+
970
+ class ChainableUndefined(Undefined):
971
+ """An undefined that is chainable, where both ``__getattr__`` and
972
+ ``__getitem__`` return itself rather than raising an
973
+ :exc:`UndefinedError`.
974
+
975
+ >>> foo = ChainableUndefined(name='foo')
976
+ >>> str(foo.bar['baz'])
977
+ ''
978
+ >>> foo.bar['baz'] + 42
979
+ Traceback (most recent call last):
980
+ ...
981
+ jinja2.exceptions.UndefinedError: 'foo' is undefined
982
+
983
+ .. versionadded:: 2.11.0
984
+ """
985
+
986
+ __slots__ = ()
987
+
988
+ def __html__(self) -> str:
989
+ return str(self)
990
+
991
+ def __getattr__(self, name: str) -> "ChainableUndefined":
992
+ # Raise AttributeError on requests for names that appear to be unimplemented
993
+ # dunder methods to avoid confusing Python with truthy non-method objects that
994
+ # do not implement the protocol being probed for. e.g., copy.copy(Undefined())
995
+ # fails spectacularly if getattr(Undefined(), '__setstate__') returns an
996
+ # Undefined object instead of raising AttributeError to signal that it does not
997
+ # support that style of object initialization.
998
+ if name[:2] == "__" and name[-2:] == "__":
999
+ raise AttributeError(name)
1000
+
1001
+ return self
1002
+
1003
+ def __getitem__(self, _name: str) -> "ChainableUndefined": # type: ignore[override]
1004
+ return self
1005
+
1006
+
1007
+ class DebugUndefined(Undefined):
1008
+ """An undefined that returns the debug info when printed.
1009
+
1010
+ >>> foo = DebugUndefined(name='foo')
1011
+ >>> str(foo)
1012
+ '{{ foo }}'
1013
+ >>> not foo
1014
+ True
1015
+ >>> foo + 42
1016
+ Traceback (most recent call last):
1017
+ ...
1018
+ jinja2.exceptions.UndefinedError: 'foo' is undefined
1019
+ """
1020
+
1021
+ __slots__ = ()
1022
+
1023
+ def __str__(self) -> str:
1024
+ if self._undefined_hint:
1025
+ message = f"undefined value printed: {self._undefined_hint}"
1026
+
1027
+ elif self._undefined_obj is missing:
1028
+ message = self._undefined_name # type: ignore
1029
+
1030
+ else:
1031
+ message = (
1032
+ f"no such element: {object_type_repr(self._undefined_obj)}"
1033
+ f"[{self._undefined_name!r}]"
1034
+ )
1035
+
1036
+ return f"{{{{ {message} }}}}"
1037
+
1038
+
1039
+ class StrictUndefined(Undefined):
1040
+ """An undefined that barks on print and iteration as well as boolean
1041
+ tests and all kinds of comparisons. In other words: you can do nothing
1042
+ with it except checking if it's defined using the `defined` test.
1043
+
1044
+ >>> foo = StrictUndefined(name='foo')
1045
+ >>> str(foo)
1046
+ Traceback (most recent call last):
1047
+ ...
1048
+ jinja2.exceptions.UndefinedError: 'foo' is undefined
1049
+ >>> not foo
1050
+ Traceback (most recent call last):
1051
+ ...
1052
+ jinja2.exceptions.UndefinedError: 'foo' is undefined
1053
+ >>> foo + 42
1054
+ Traceback (most recent call last):
1055
+ ...
1056
+ jinja2.exceptions.UndefinedError: 'foo' is undefined
1057
+ """
1058
+
1059
+ __slots__ = ()
1060
+ __iter__ = __str__ = __len__ = Undefined._fail_with_undefined_error
1061
+ __eq__ = __ne__ = __bool__ = __hash__ = Undefined._fail_with_undefined_error
1062
+ __contains__ = Undefined._fail_with_undefined_error
janus/lib/python3.10/site-packages/jinja2/tests.py ADDED
@@ -0,0 +1,256 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """Built-in template tests used with the ``is`` operator."""
2
+
3
+ import operator
4
+ import typing as t
5
+ from collections import abc
6
+ from numbers import Number
7
+
8
+ from .runtime import Undefined
9
+ from .utils import pass_environment
10
+
11
+ if t.TYPE_CHECKING:
12
+ from .environment import Environment
13
+
14
+
15
+ def test_odd(value: int) -> bool:
16
+ """Return true if the variable is odd."""
17
+ return value % 2 == 1
18
+
19
+
20
+ def test_even(value: int) -> bool:
21
+ """Return true if the variable is even."""
22
+ return value % 2 == 0
23
+
24
+
25
+ def test_divisibleby(value: int, num: int) -> bool:
26
+ """Check if a variable is divisible by a number."""
27
+ return value % num == 0
28
+
29
+
30
+ def test_defined(value: t.Any) -> bool:
31
+ """Return true if the variable is defined:
32
+
33
+ .. sourcecode:: jinja
34
+
35
+ {% if variable is defined %}
36
+ value of variable: {{ variable }}
37
+ {% else %}
38
+ variable is not defined
39
+ {% endif %}
40
+
41
+ See the :func:`default` filter for a simple way to set undefined
42
+ variables.
43
+ """
44
+ return not isinstance(value, Undefined)
45
+
46
+
47
+ def test_undefined(value: t.Any) -> bool:
48
+ """Like :func:`defined` but the other way round."""
49
+ return isinstance(value, Undefined)
50
+
51
+
52
+ @pass_environment
53
+ def test_filter(env: "Environment", value: str) -> bool:
54
+ """Check if a filter exists by name. Useful if a filter may be
55
+ optionally available.
56
+
57
+ .. code-block:: jinja
58
+
59
+ {% if 'markdown' is filter %}
60
+ {{ value | markdown }}
61
+ {% else %}
62
+ {{ value }}
63
+ {% endif %}
64
+
65
+ .. versionadded:: 3.0
66
+ """
67
+ return value in env.filters
68
+
69
+
70
+ @pass_environment
71
+ def test_test(env: "Environment", value: str) -> bool:
72
+ """Check if a test exists by name. Useful if a test may be
73
+ optionally available.
74
+
75
+ .. code-block:: jinja
76
+
77
+ {% if 'loud' is test %}
78
+ {% if value is loud %}
79
+ {{ value|upper }}
80
+ {% else %}
81
+ {{ value|lower }}
82
+ {% endif %}
83
+ {% else %}
84
+ {{ value }}
85
+ {% endif %}
86
+
87
+ .. versionadded:: 3.0
88
+ """
89
+ return value in env.tests
90
+
91
+
92
+ def test_none(value: t.Any) -> bool:
93
+ """Return true if the variable is none."""
94
+ return value is None
95
+
96
+
97
+ def test_boolean(value: t.Any) -> bool:
98
+ """Return true if the object is a boolean value.
99
+
100
+ .. versionadded:: 2.11
101
+ """
102
+ return value is True or value is False
103
+
104
+
105
+ def test_false(value: t.Any) -> bool:
106
+ """Return true if the object is False.
107
+
108
+ .. versionadded:: 2.11
109
+ """
110
+ return value is False
111
+
112
+
113
+ def test_true(value: t.Any) -> bool:
114
+ """Return true if the object is True.
115
+
116
+ .. versionadded:: 2.11
117
+ """
118
+ return value is True
119
+
120
+
121
+ # NOTE: The existing 'number' test matches booleans and floats
122
+ def test_integer(value: t.Any) -> bool:
123
+ """Return true if the object is an integer.
124
+
125
+ .. versionadded:: 2.11
126
+ """
127
+ return isinstance(value, int) and value is not True and value is not False
128
+
129
+
130
+ # NOTE: The existing 'number' test matches booleans and integers
131
+ def test_float(value: t.Any) -> bool:
132
+ """Return true if the object is a float.
133
+
134
+ .. versionadded:: 2.11
135
+ """
136
+ return isinstance(value, float)
137
+
138
+
139
+ def test_lower(value: str) -> bool:
140
+ """Return true if the variable is lowercased."""
141
+ return str(value).islower()
142
+
143
+
144
+ def test_upper(value: str) -> bool:
145
+ """Return true if the variable is uppercased."""
146
+ return str(value).isupper()
147
+
148
+
149
+ def test_string(value: t.Any) -> bool:
150
+ """Return true if the object is a string."""
151
+ return isinstance(value, str)
152
+
153
+
154
+ def test_mapping(value: t.Any) -> bool:
155
+ """Return true if the object is a mapping (dict etc.).
156
+
157
+ .. versionadded:: 2.6
158
+ """
159
+ return isinstance(value, abc.Mapping)
160
+
161
+
162
+ def test_number(value: t.Any) -> bool:
163
+ """Return true if the variable is a number."""
164
+ return isinstance(value, Number)
165
+
166
+
167
+ def test_sequence(value: t.Any) -> bool:
168
+ """Return true if the variable is a sequence. Sequences are variables
169
+ that are iterable.
170
+ """
171
+ try:
172
+ len(value)
173
+ value.__getitem__ # noqa B018
174
+ except Exception:
175
+ return False
176
+
177
+ return True
178
+
179
+
180
+ def test_sameas(value: t.Any, other: t.Any) -> bool:
181
+ """Check if an object points to the same memory address than another
182
+ object:
183
+
184
+ .. sourcecode:: jinja
185
+
186
+ {% if foo.attribute is sameas false %}
187
+ the foo attribute really is the `False` singleton
188
+ {% endif %}
189
+ """
190
+ return value is other
191
+
192
+
193
+ def test_iterable(value: t.Any) -> bool:
194
+ """Check if it's possible to iterate over an object."""
195
+ try:
196
+ iter(value)
197
+ except TypeError:
198
+ return False
199
+
200
+ return True
201
+
202
+
203
+ def test_escaped(value: t.Any) -> bool:
204
+ """Check if the value is escaped."""
205
+ return hasattr(value, "__html__")
206
+
207
+
208
+ def test_in(value: t.Any, seq: t.Container[t.Any]) -> bool:
209
+ """Check if value is in seq.
210
+
211
+ .. versionadded:: 2.10
212
+ """
213
+ return value in seq
214
+
215
+
216
+ TESTS = {
217
+ "odd": test_odd,
218
+ "even": test_even,
219
+ "divisibleby": test_divisibleby,
220
+ "defined": test_defined,
221
+ "undefined": test_undefined,
222
+ "filter": test_filter,
223
+ "test": test_test,
224
+ "none": test_none,
225
+ "boolean": test_boolean,
226
+ "false": test_false,
227
+ "true": test_true,
228
+ "integer": test_integer,
229
+ "float": test_float,
230
+ "lower": test_lower,
231
+ "upper": test_upper,
232
+ "string": test_string,
233
+ "mapping": test_mapping,
234
+ "number": test_number,
235
+ "sequence": test_sequence,
236
+ "iterable": test_iterable,
237
+ "callable": callable,
238
+ "sameas": test_sameas,
239
+ "escaped": test_escaped,
240
+ "in": test_in,
241
+ "==": operator.eq,
242
+ "eq": operator.eq,
243
+ "equalto": operator.eq,
244
+ "!=": operator.ne,
245
+ "ne": operator.ne,
246
+ ">": operator.gt,
247
+ "gt": operator.gt,
248
+ "greaterthan": operator.gt,
249
+ "ge": operator.ge,
250
+ ">=": operator.ge,
251
+ "<": operator.lt,
252
+ "lt": operator.lt,
253
+ "lessthan": operator.lt,
254
+ "<=": operator.le,
255
+ "le": operator.le,
256
+ }
janus/lib/python3.10/site-packages/jinja2/utils.py ADDED
@@ -0,0 +1,766 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import enum
2
+ import json
3
+ import os
4
+ import re
5
+ import typing as t
6
+ from collections import abc
7
+ from collections import deque
8
+ from random import choice
9
+ from random import randrange
10
+ from threading import Lock
11
+ from types import CodeType
12
+ from urllib.parse import quote_from_bytes
13
+
14
+ import markupsafe
15
+
16
+ if t.TYPE_CHECKING:
17
+ import typing_extensions as te
18
+
19
+ F = t.TypeVar("F", bound=t.Callable[..., t.Any])
20
+
21
+
22
+ class _MissingType:
23
+ def __repr__(self) -> str:
24
+ return "missing"
25
+
26
+ def __reduce__(self) -> str:
27
+ return "missing"
28
+
29
+
30
+ missing: t.Any = _MissingType()
31
+ """Special singleton representing missing values for the runtime."""
32
+
33
+ internal_code: t.MutableSet[CodeType] = set()
34
+
35
+ concat = "".join
36
+
37
+
38
+ def pass_context(f: F) -> F:
39
+ """Pass the :class:`~jinja2.runtime.Context` as the first argument
40
+ to the decorated function when called while rendering a template.
41
+
42
+ Can be used on functions, filters, and tests.
43
+
44
+ If only ``Context.eval_context`` is needed, use
45
+ :func:`pass_eval_context`. If only ``Context.environment`` is
46
+ needed, use :func:`pass_environment`.
47
+
48
+ .. versionadded:: 3.0.0
49
+ Replaces ``contextfunction`` and ``contextfilter``.
50
+ """
51
+ f.jinja_pass_arg = _PassArg.context # type: ignore
52
+ return f
53
+
54
+
55
+ def pass_eval_context(f: F) -> F:
56
+ """Pass the :class:`~jinja2.nodes.EvalContext` as the first argument
57
+ to the decorated function when called while rendering a template.
58
+ See :ref:`eval-context`.
59
+
60
+ Can be used on functions, filters, and tests.
61
+
62
+ If only ``EvalContext.environment`` is needed, use
63
+ :func:`pass_environment`.
64
+
65
+ .. versionadded:: 3.0.0
66
+ Replaces ``evalcontextfunction`` and ``evalcontextfilter``.
67
+ """
68
+ f.jinja_pass_arg = _PassArg.eval_context # type: ignore
69
+ return f
70
+
71
+
72
+ def pass_environment(f: F) -> F:
73
+ """Pass the :class:`~jinja2.Environment` as the first argument to
74
+ the decorated function when called while rendering a template.
75
+
76
+ Can be used on functions, filters, and tests.
77
+
78
+ .. versionadded:: 3.0.0
79
+ Replaces ``environmentfunction`` and ``environmentfilter``.
80
+ """
81
+ f.jinja_pass_arg = _PassArg.environment # type: ignore
82
+ return f
83
+
84
+
85
+ class _PassArg(enum.Enum):
86
+ context = enum.auto()
87
+ eval_context = enum.auto()
88
+ environment = enum.auto()
89
+
90
+ @classmethod
91
+ def from_obj(cls, obj: F) -> t.Optional["_PassArg"]:
92
+ if hasattr(obj, "jinja_pass_arg"):
93
+ return obj.jinja_pass_arg # type: ignore
94
+
95
+ return None
96
+
97
+
98
+ def internalcode(f: F) -> F:
99
+ """Marks the function as internally used"""
100
+ internal_code.add(f.__code__)
101
+ return f
102
+
103
+
104
+ def is_undefined(obj: t.Any) -> bool:
105
+ """Check if the object passed is undefined. This does nothing more than
106
+ performing an instance check against :class:`Undefined` but looks nicer.
107
+ This can be used for custom filters or tests that want to react to
108
+ undefined variables. For example a custom default filter can look like
109
+ this::
110
+
111
+ def default(var, default=''):
112
+ if is_undefined(var):
113
+ return default
114
+ return var
115
+ """
116
+ from .runtime import Undefined
117
+
118
+ return isinstance(obj, Undefined)
119
+
120
+
121
+ def consume(iterable: t.Iterable[t.Any]) -> None:
122
+ """Consumes an iterable without doing anything with it."""
123
+ for _ in iterable:
124
+ pass
125
+
126
+
127
+ def clear_caches() -> None:
128
+ """Jinja keeps internal caches for environments and lexers. These are
129
+ used so that Jinja doesn't have to recreate environments and lexers all
130
+ the time. Normally you don't have to care about that but if you are
131
+ measuring memory consumption you may want to clean the caches.
132
+ """
133
+ from .environment import get_spontaneous_environment
134
+ from .lexer import _lexer_cache
135
+
136
+ get_spontaneous_environment.cache_clear()
137
+ _lexer_cache.clear()
138
+
139
+
140
+ def import_string(import_name: str, silent: bool = False) -> t.Any:
141
+ """Imports an object based on a string. This is useful if you want to
142
+ use import paths as endpoints or something similar. An import path can
143
+ be specified either in dotted notation (``xml.sax.saxutils.escape``)
144
+ or with a colon as object delimiter (``xml.sax.saxutils:escape``).
145
+
146
+ If the `silent` is True the return value will be `None` if the import
147
+ fails.
148
+
149
+ :return: imported object
150
+ """
151
+ try:
152
+ if ":" in import_name:
153
+ module, obj = import_name.split(":", 1)
154
+ elif "." in import_name:
155
+ module, _, obj = import_name.rpartition(".")
156
+ else:
157
+ return __import__(import_name)
158
+ return getattr(__import__(module, None, None, [obj]), obj)
159
+ except (ImportError, AttributeError):
160
+ if not silent:
161
+ raise
162
+
163
+
164
+ def open_if_exists(filename: str, mode: str = "rb") -> t.Optional[t.IO[t.Any]]:
165
+ """Returns a file descriptor for the filename if that file exists,
166
+ otherwise ``None``.
167
+ """
168
+ if not os.path.isfile(filename):
169
+ return None
170
+
171
+ return open(filename, mode)
172
+
173
+
174
+ def object_type_repr(obj: t.Any) -> str:
175
+ """Returns the name of the object's type. For some recognized
176
+ singletons the name of the object is returned instead. (For
177
+ example for `None` and `Ellipsis`).
178
+ """
179
+ if obj is None:
180
+ return "None"
181
+ elif obj is Ellipsis:
182
+ return "Ellipsis"
183
+
184
+ cls = type(obj)
185
+
186
+ if cls.__module__ == "builtins":
187
+ return f"{cls.__name__} object"
188
+
189
+ return f"{cls.__module__}.{cls.__name__} object"
190
+
191
+
192
+ def pformat(obj: t.Any) -> str:
193
+ """Format an object using :func:`pprint.pformat`."""
194
+ from pprint import pformat
195
+
196
+ return pformat(obj)
197
+
198
+
199
+ _http_re = re.compile(
200
+ r"""
201
+ ^
202
+ (
203
+ (https?://|www\.) # scheme or www
204
+ (([\w%-]+\.)+)? # subdomain
205
+ (
206
+ [a-z]{2,63} # basic tld
207
+ |
208
+ xn--[\w%]{2,59} # idna tld
209
+ )
210
+ |
211
+ ([\w%-]{2,63}\.)+ # basic domain
212
+ (com|net|int|edu|gov|org|info|mil) # basic tld
213
+ |
214
+ (https?://) # scheme
215
+ (
216
+ (([\d]{1,3})(\.[\d]{1,3}){3}) # IPv4
217
+ |
218
+ (\[([\da-f]{0,4}:){2}([\da-f]{0,4}:?){1,6}]) # IPv6
219
+ )
220
+ )
221
+ (?::[\d]{1,5})? # port
222
+ (?:[/?#]\S*)? # path, query, and fragment
223
+ $
224
+ """,
225
+ re.IGNORECASE | re.VERBOSE,
226
+ )
227
+ _email_re = re.compile(r"^\S+@\w[\w.-]*\.\w+$")
228
+
229
+
230
+ def urlize(
231
+ text: str,
232
+ trim_url_limit: t.Optional[int] = None,
233
+ rel: t.Optional[str] = None,
234
+ target: t.Optional[str] = None,
235
+ extra_schemes: t.Optional[t.Iterable[str]] = None,
236
+ ) -> str:
237
+ """Convert URLs in text into clickable links.
238
+
239
+ This may not recognize links in some situations. Usually, a more
240
+ comprehensive formatter, such as a Markdown library, is a better
241
+ choice.
242
+
243
+ Works on ``http://``, ``https://``, ``www.``, ``mailto:``, and email
244
+ addresses. Links with trailing punctuation (periods, commas, closing
245
+ parentheses) and leading punctuation (opening parentheses) are
246
+ recognized excluding the punctuation. Email addresses that include
247
+ header fields are not recognized (for example,
248
+ ``mailto:address@example.com?cc=copy@example.com``).
249
+
250
+ :param text: Original text containing URLs to link.
251
+ :param trim_url_limit: Shorten displayed URL values to this length.
252
+ :param target: Add the ``target`` attribute to links.
253
+ :param rel: Add the ``rel`` attribute to links.
254
+ :param extra_schemes: Recognize URLs that start with these schemes
255
+ in addition to the default behavior.
256
+
257
+ .. versionchanged:: 3.0
258
+ The ``extra_schemes`` parameter was added.
259
+
260
+ .. versionchanged:: 3.0
261
+ Generate ``https://`` links for URLs without a scheme.
262
+
263
+ .. versionchanged:: 3.0
264
+ The parsing rules were updated. Recognize email addresses with
265
+ or without the ``mailto:`` scheme. Validate IP addresses. Ignore
266
+ parentheses and brackets in more cases.
267
+ """
268
+ if trim_url_limit is not None:
269
+
270
+ def trim_url(x: str) -> str:
271
+ if len(x) > trim_url_limit:
272
+ return f"{x[:trim_url_limit]}..."
273
+
274
+ return x
275
+
276
+ else:
277
+
278
+ def trim_url(x: str) -> str:
279
+ return x
280
+
281
+ words = re.split(r"(\s+)", str(markupsafe.escape(text)))
282
+ rel_attr = f' rel="{markupsafe.escape(rel)}"' if rel else ""
283
+ target_attr = f' target="{markupsafe.escape(target)}"' if target else ""
284
+
285
+ for i, word in enumerate(words):
286
+ head, middle, tail = "", word, ""
287
+ match = re.match(r"^([(<]|&lt;)+", middle)
288
+
289
+ if match:
290
+ head = match.group()
291
+ middle = middle[match.end() :]
292
+
293
+ # Unlike lead, which is anchored to the start of the string,
294
+ # need to check that the string ends with any of the characters
295
+ # before trying to match all of them, to avoid backtracking.
296
+ if middle.endswith((")", ">", ".", ",", "\n", "&gt;")):
297
+ match = re.search(r"([)>.,\n]|&gt;)+$", middle)
298
+
299
+ if match:
300
+ tail = match.group()
301
+ middle = middle[: match.start()]
302
+
303
+ # Prefer balancing parentheses in URLs instead of ignoring a
304
+ # trailing character.
305
+ for start_char, end_char in ("(", ")"), ("<", ">"), ("&lt;", "&gt;"):
306
+ start_count = middle.count(start_char)
307
+
308
+ if start_count <= middle.count(end_char):
309
+ # Balanced, or lighter on the left
310
+ continue
311
+
312
+ # Move as many as possible from the tail to balance
313
+ for _ in range(min(start_count, tail.count(end_char))):
314
+ end_index = tail.index(end_char) + len(end_char)
315
+ # Move anything in the tail before the end char too
316
+ middle += tail[:end_index]
317
+ tail = tail[end_index:]
318
+
319
+ if _http_re.match(middle):
320
+ if middle.startswith("https://") or middle.startswith("http://"):
321
+ middle = (
322
+ f'<a href="{middle}"{rel_attr}{target_attr}>{trim_url(middle)}</a>'
323
+ )
324
+ else:
325
+ middle = (
326
+ f'<a href="https://{middle}"{rel_attr}{target_attr}>'
327
+ f"{trim_url(middle)}</a>"
328
+ )
329
+
330
+ elif middle.startswith("mailto:") and _email_re.match(middle[7:]):
331
+ middle = f'<a href="{middle}">{middle[7:]}</a>'
332
+
333
+ elif (
334
+ "@" in middle
335
+ and not middle.startswith("www.")
336
+ # ignore values like `@a@b`
337
+ and not middle.startswith("@")
338
+ and ":" not in middle
339
+ and _email_re.match(middle)
340
+ ):
341
+ middle = f'<a href="mailto:{middle}">{middle}</a>'
342
+
343
+ elif extra_schemes is not None:
344
+ for scheme in extra_schemes:
345
+ if middle != scheme and middle.startswith(scheme):
346
+ middle = f'<a href="{middle}"{rel_attr}{target_attr}>{middle}</a>'
347
+
348
+ words[i] = f"{head}{middle}{tail}"
349
+
350
+ return "".join(words)
351
+
352
+
353
+ def generate_lorem_ipsum(
354
+ n: int = 5, html: bool = True, min: int = 20, max: int = 100
355
+ ) -> str:
356
+ """Generate some lorem ipsum for the template."""
357
+ from .constants import LOREM_IPSUM_WORDS
358
+
359
+ words = LOREM_IPSUM_WORDS.split()
360
+ result = []
361
+
362
+ for _ in range(n):
363
+ next_capitalized = True
364
+ last_comma = last_fullstop = 0
365
+ word = None
366
+ last = None
367
+ p = []
368
+
369
+ # each paragraph contains out of 20 to 100 words.
370
+ for idx, _ in enumerate(range(randrange(min, max))):
371
+ while True:
372
+ word = choice(words)
373
+ if word != last:
374
+ last = word
375
+ break
376
+ if next_capitalized:
377
+ word = word.capitalize()
378
+ next_capitalized = False
379
+ # add commas
380
+ if idx - randrange(3, 8) > last_comma:
381
+ last_comma = idx
382
+ last_fullstop += 2
383
+ word += ","
384
+ # add end of sentences
385
+ if idx - randrange(10, 20) > last_fullstop:
386
+ last_comma = last_fullstop = idx
387
+ word += "."
388
+ next_capitalized = True
389
+ p.append(word)
390
+
391
+ # ensure that the paragraph ends with a dot.
392
+ p_str = " ".join(p)
393
+
394
+ if p_str.endswith(","):
395
+ p_str = p_str[:-1] + "."
396
+ elif not p_str.endswith("."):
397
+ p_str += "."
398
+
399
+ result.append(p_str)
400
+
401
+ if not html:
402
+ return "\n\n".join(result)
403
+ return markupsafe.Markup(
404
+ "\n".join(f"<p>{markupsafe.escape(x)}</p>" for x in result)
405
+ )
406
+
407
+
408
+ def url_quote(obj: t.Any, charset: str = "utf-8", for_qs: bool = False) -> str:
409
+ """Quote a string for use in a URL using the given charset.
410
+
411
+ :param obj: String or bytes to quote. Other types are converted to
412
+ string then encoded to bytes using the given charset.
413
+ :param charset: Encode text to bytes using this charset.
414
+ :param for_qs: Quote "/" and use "+" for spaces.
415
+ """
416
+ if not isinstance(obj, bytes):
417
+ if not isinstance(obj, str):
418
+ obj = str(obj)
419
+
420
+ obj = obj.encode(charset)
421
+
422
+ safe = b"" if for_qs else b"/"
423
+ rv = quote_from_bytes(obj, safe)
424
+
425
+ if for_qs:
426
+ rv = rv.replace("%20", "+")
427
+
428
+ return rv
429
+
430
+
431
+ @abc.MutableMapping.register
432
+ class LRUCache:
433
+ """A simple LRU Cache implementation."""
434
+
435
+ # this is fast for small capacities (something below 1000) but doesn't
436
+ # scale. But as long as it's only used as storage for templates this
437
+ # won't do any harm.
438
+
439
+ def __init__(self, capacity: int) -> None:
440
+ self.capacity = capacity
441
+ self._mapping: t.Dict[t.Any, t.Any] = {}
442
+ self._queue: te.Deque[t.Any] = deque()
443
+ self._postinit()
444
+
445
+ def _postinit(self) -> None:
446
+ # alias all queue methods for faster lookup
447
+ self._popleft = self._queue.popleft
448
+ self._pop = self._queue.pop
449
+ self._remove = self._queue.remove
450
+ self._wlock = Lock()
451
+ self._append = self._queue.append
452
+
453
+ def __getstate__(self) -> t.Mapping[str, t.Any]:
454
+ return {
455
+ "capacity": self.capacity,
456
+ "_mapping": self._mapping,
457
+ "_queue": self._queue,
458
+ }
459
+
460
+ def __setstate__(self, d: t.Mapping[str, t.Any]) -> None:
461
+ self.__dict__.update(d)
462
+ self._postinit()
463
+
464
+ def __getnewargs__(self) -> t.Tuple[t.Any, ...]:
465
+ return (self.capacity,)
466
+
467
+ def copy(self) -> "te.Self":
468
+ """Return a shallow copy of the instance."""
469
+ rv = self.__class__(self.capacity)
470
+ rv._mapping.update(self._mapping)
471
+ rv._queue.extend(self._queue)
472
+ return rv
473
+
474
+ def get(self, key: t.Any, default: t.Any = None) -> t.Any:
475
+ """Return an item from the cache dict or `default`"""
476
+ try:
477
+ return self[key]
478
+ except KeyError:
479
+ return default
480
+
481
+ def setdefault(self, key: t.Any, default: t.Any = None) -> t.Any:
482
+ """Set `default` if the key is not in the cache otherwise
483
+ leave unchanged. Return the value of this key.
484
+ """
485
+ try:
486
+ return self[key]
487
+ except KeyError:
488
+ self[key] = default
489
+ return default
490
+
491
+ def clear(self) -> None:
492
+ """Clear the cache."""
493
+ with self._wlock:
494
+ self._mapping.clear()
495
+ self._queue.clear()
496
+
497
+ def __contains__(self, key: t.Any) -> bool:
498
+ """Check if a key exists in this cache."""
499
+ return key in self._mapping
500
+
501
+ def __len__(self) -> int:
502
+ """Return the current size of the cache."""
503
+ return len(self._mapping)
504
+
505
+ def __repr__(self) -> str:
506
+ return f"<{type(self).__name__} {self._mapping!r}>"
507
+
508
+ def __getitem__(self, key: t.Any) -> t.Any:
509
+ """Get an item from the cache. Moves the item up so that it has the
510
+ highest priority then.
511
+
512
+ Raise a `KeyError` if it does not exist.
513
+ """
514
+ with self._wlock:
515
+ rv = self._mapping[key]
516
+
517
+ if self._queue[-1] != key:
518
+ try:
519
+ self._remove(key)
520
+ except ValueError:
521
+ # if something removed the key from the container
522
+ # when we read, ignore the ValueError that we would
523
+ # get otherwise.
524
+ pass
525
+
526
+ self._append(key)
527
+
528
+ return rv
529
+
530
+ def __setitem__(self, key: t.Any, value: t.Any) -> None:
531
+ """Sets the value for an item. Moves the item up so that it
532
+ has the highest priority then.
533
+ """
534
+ with self._wlock:
535
+ if key in self._mapping:
536
+ self._remove(key)
537
+ elif len(self._mapping) == self.capacity:
538
+ del self._mapping[self._popleft()]
539
+
540
+ self._append(key)
541
+ self._mapping[key] = value
542
+
543
+ def __delitem__(self, key: t.Any) -> None:
544
+ """Remove an item from the cache dict.
545
+ Raise a `KeyError` if it does not exist.
546
+ """
547
+ with self._wlock:
548
+ del self._mapping[key]
549
+
550
+ try:
551
+ self._remove(key)
552
+ except ValueError:
553
+ pass
554
+
555
+ def items(self) -> t.Iterable[t.Tuple[t.Any, t.Any]]:
556
+ """Return a list of items."""
557
+ result = [(key, self._mapping[key]) for key in list(self._queue)]
558
+ result.reverse()
559
+ return result
560
+
561
+ def values(self) -> t.Iterable[t.Any]:
562
+ """Return a list of all values."""
563
+ return [x[1] for x in self.items()]
564
+
565
+ def keys(self) -> t.Iterable[t.Any]:
566
+ """Return a list of all keys ordered by most recent usage."""
567
+ return list(self)
568
+
569
+ def __iter__(self) -> t.Iterator[t.Any]:
570
+ return reversed(tuple(self._queue))
571
+
572
+ def __reversed__(self) -> t.Iterator[t.Any]:
573
+ """Iterate over the keys in the cache dict, oldest items
574
+ coming first.
575
+ """
576
+ return iter(tuple(self._queue))
577
+
578
+ __copy__ = copy
579
+
580
+
581
+ def select_autoescape(
582
+ enabled_extensions: t.Collection[str] = ("html", "htm", "xml"),
583
+ disabled_extensions: t.Collection[str] = (),
584
+ default_for_string: bool = True,
585
+ default: bool = False,
586
+ ) -> t.Callable[[t.Optional[str]], bool]:
587
+ """Intelligently sets the initial value of autoescaping based on the
588
+ filename of the template. This is the recommended way to configure
589
+ autoescaping if you do not want to write a custom function yourself.
590
+
591
+ If you want to enable it for all templates created from strings or
592
+ for all templates with `.html` and `.xml` extensions::
593
+
594
+ from jinja2 import Environment, select_autoescape
595
+ env = Environment(autoescape=select_autoescape(
596
+ enabled_extensions=('html', 'xml'),
597
+ default_for_string=True,
598
+ ))
599
+
600
+ Example configuration to turn it on at all times except if the template
601
+ ends with `.txt`::
602
+
603
+ from jinja2 import Environment, select_autoescape
604
+ env = Environment(autoescape=select_autoescape(
605
+ disabled_extensions=('txt',),
606
+ default_for_string=True,
607
+ default=True,
608
+ ))
609
+
610
+ The `enabled_extensions` is an iterable of all the extensions that
611
+ autoescaping should be enabled for. Likewise `disabled_extensions` is
612
+ a list of all templates it should be disabled for. If a template is
613
+ loaded from a string then the default from `default_for_string` is used.
614
+ If nothing matches then the initial value of autoescaping is set to the
615
+ value of `default`.
616
+
617
+ For security reasons this function operates case insensitive.
618
+
619
+ .. versionadded:: 2.9
620
+ """
621
+ enabled_patterns = tuple(f".{x.lstrip('.').lower()}" for x in enabled_extensions)
622
+ disabled_patterns = tuple(f".{x.lstrip('.').lower()}" for x in disabled_extensions)
623
+
624
+ def autoescape(template_name: t.Optional[str]) -> bool:
625
+ if template_name is None:
626
+ return default_for_string
627
+ template_name = template_name.lower()
628
+ if template_name.endswith(enabled_patterns):
629
+ return True
630
+ if template_name.endswith(disabled_patterns):
631
+ return False
632
+ return default
633
+
634
+ return autoescape
635
+
636
+
637
+ def htmlsafe_json_dumps(
638
+ obj: t.Any, dumps: t.Optional[t.Callable[..., str]] = None, **kwargs: t.Any
639
+ ) -> markupsafe.Markup:
640
+ """Serialize an object to a string of JSON with :func:`json.dumps`,
641
+ then replace HTML-unsafe characters with Unicode escapes and mark
642
+ the result safe with :class:`~markupsafe.Markup`.
643
+
644
+ This is available in templates as the ``|tojson`` filter.
645
+
646
+ The following characters are escaped: ``<``, ``>``, ``&``, ``'``.
647
+
648
+ The returned string is safe to render in HTML documents and
649
+ ``<script>`` tags. The exception is in HTML attributes that are
650
+ double quoted; either use single quotes or the ``|forceescape``
651
+ filter.
652
+
653
+ :param obj: The object to serialize to JSON.
654
+ :param dumps: The ``dumps`` function to use. Defaults to
655
+ ``env.policies["json.dumps_function"]``, which defaults to
656
+ :func:`json.dumps`.
657
+ :param kwargs: Extra arguments to pass to ``dumps``. Merged onto
658
+ ``env.policies["json.dumps_kwargs"]``.
659
+
660
+ .. versionchanged:: 3.0
661
+ The ``dumper`` parameter is renamed to ``dumps``.
662
+
663
+ .. versionadded:: 2.9
664
+ """
665
+ if dumps is None:
666
+ dumps = json.dumps
667
+
668
+ return markupsafe.Markup(
669
+ dumps(obj, **kwargs)
670
+ .replace("<", "\\u003c")
671
+ .replace(">", "\\u003e")
672
+ .replace("&", "\\u0026")
673
+ .replace("'", "\\u0027")
674
+ )
675
+
676
+
677
+ class Cycler:
678
+ """Cycle through values by yield them one at a time, then restarting
679
+ once the end is reached. Available as ``cycler`` in templates.
680
+
681
+ Similar to ``loop.cycle``, but can be used outside loops or across
682
+ multiple loops. For example, render a list of folders and files in a
683
+ list, alternating giving them "odd" and "even" classes.
684
+
685
+ .. code-block:: html+jinja
686
+
687
+ {% set row_class = cycler("odd", "even") %}
688
+ <ul class="browser">
689
+ {% for folder in folders %}
690
+ <li class="folder {{ row_class.next() }}">{{ folder }}
691
+ {% endfor %}
692
+ {% for file in files %}
693
+ <li class="file {{ row_class.next() }}">{{ file }}
694
+ {% endfor %}
695
+ </ul>
696
+
697
+ :param items: Each positional argument will be yielded in the order
698
+ given for each cycle.
699
+
700
+ .. versionadded:: 2.1
701
+ """
702
+
703
+ def __init__(self, *items: t.Any) -> None:
704
+ if not items:
705
+ raise RuntimeError("at least one item has to be provided")
706
+ self.items = items
707
+ self.pos = 0
708
+
709
+ def reset(self) -> None:
710
+ """Resets the current item to the first item."""
711
+ self.pos = 0
712
+
713
+ @property
714
+ def current(self) -> t.Any:
715
+ """Return the current item. Equivalent to the item that will be
716
+ returned next time :meth:`next` is called.
717
+ """
718
+ return self.items[self.pos]
719
+
720
+ def next(self) -> t.Any:
721
+ """Return the current item, then advance :attr:`current` to the
722
+ next item.
723
+ """
724
+ rv = self.current
725
+ self.pos = (self.pos + 1) % len(self.items)
726
+ return rv
727
+
728
+ __next__ = next
729
+
730
+
731
+ class Joiner:
732
+ """A joining helper for templates."""
733
+
734
+ def __init__(self, sep: str = ", ") -> None:
735
+ self.sep = sep
736
+ self.used = False
737
+
738
+ def __call__(self) -> str:
739
+ if not self.used:
740
+ self.used = True
741
+ return ""
742
+ return self.sep
743
+
744
+
745
+ class Namespace:
746
+ """A namespace object that can hold arbitrary attributes. It may be
747
+ initialized from a dictionary or with keyword arguments."""
748
+
749
+ def __init__(*args: t.Any, **kwargs: t.Any) -> None: # noqa: B902
750
+ self, args = args[0], args[1:]
751
+ self.__attrs = dict(*args, **kwargs)
752
+
753
+ def __getattribute__(self, name: str) -> t.Any:
754
+ # __class__ is needed for the awaitable check in async mode
755
+ if name in {"_Namespace__attrs", "__class__"}:
756
+ return object.__getattribute__(self, name)
757
+ try:
758
+ return self.__attrs[name]
759
+ except KeyError:
760
+ raise AttributeError(name) from None
761
+
762
+ def __setitem__(self, name: str, value: t.Any) -> None:
763
+ self.__attrs[name] = value
764
+
765
+ def __repr__(self) -> str:
766
+ return f"<Namespace {self.__attrs!r}>"
janus/lib/python3.10/site-packages/jinja2/visitor.py ADDED
@@ -0,0 +1,92 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """API for traversing the AST nodes. Implemented by the compiler and
2
+ meta introspection.
3
+ """
4
+
5
+ import typing as t
6
+
7
+ from .nodes import Node
8
+
9
+ if t.TYPE_CHECKING:
10
+ import typing_extensions as te
11
+
12
+ class VisitCallable(te.Protocol):
13
+ def __call__(self, node: Node, *args: t.Any, **kwargs: t.Any) -> t.Any: ...
14
+
15
+
16
+ class NodeVisitor:
17
+ """Walks the abstract syntax tree and call visitor functions for every
18
+ node found. The visitor functions may return values which will be
19
+ forwarded by the `visit` method.
20
+
21
+ Per default the visitor functions for the nodes are ``'visit_'`` +
22
+ class name of the node. So a `TryFinally` node visit function would
23
+ be `visit_TryFinally`. This behavior can be changed by overriding
24
+ the `get_visitor` function. If no visitor function exists for a node
25
+ (return value `None`) the `generic_visit` visitor is used instead.
26
+ """
27
+
28
+ def get_visitor(self, node: Node) -> "t.Optional[VisitCallable]":
29
+ """Return the visitor function for this node or `None` if no visitor
30
+ exists for this node. In that case the generic visit function is
31
+ used instead.
32
+ """
33
+ return getattr(self, f"visit_{type(node).__name__}", None)
34
+
35
+ def visit(self, node: Node, *args: t.Any, **kwargs: t.Any) -> t.Any:
36
+ """Visit a node."""
37
+ f = self.get_visitor(node)
38
+
39
+ if f is not None:
40
+ return f(node, *args, **kwargs)
41
+
42
+ return self.generic_visit(node, *args, **kwargs)
43
+
44
+ def generic_visit(self, node: Node, *args: t.Any, **kwargs: t.Any) -> t.Any:
45
+ """Called if no explicit visitor function exists for a node."""
46
+ for child_node in node.iter_child_nodes():
47
+ self.visit(child_node, *args, **kwargs)
48
+
49
+
50
+ class NodeTransformer(NodeVisitor):
51
+ """Walks the abstract syntax tree and allows modifications of nodes.
52
+
53
+ The `NodeTransformer` will walk the AST and use the return value of the
54
+ visitor functions to replace or remove the old node. If the return
55
+ value of the visitor function is `None` the node will be removed
56
+ from the previous location otherwise it's replaced with the return
57
+ value. The return value may be the original node in which case no
58
+ replacement takes place.
59
+ """
60
+
61
+ def generic_visit(self, node: Node, *args: t.Any, **kwargs: t.Any) -> Node:
62
+ for field, old_value in node.iter_fields():
63
+ if isinstance(old_value, list):
64
+ new_values = []
65
+ for value in old_value:
66
+ if isinstance(value, Node):
67
+ value = self.visit(value, *args, **kwargs)
68
+ if value is None:
69
+ continue
70
+ elif not isinstance(value, Node):
71
+ new_values.extend(value)
72
+ continue
73
+ new_values.append(value)
74
+ old_value[:] = new_values
75
+ elif isinstance(old_value, Node):
76
+ new_node = self.visit(old_value, *args, **kwargs)
77
+ if new_node is None:
78
+ delattr(node, field)
79
+ else:
80
+ setattr(node, field, new_node)
81
+ return node
82
+
83
+ def visit_list(self, node: Node, *args: t.Any, **kwargs: t.Any) -> t.List[Node]:
84
+ """As transformers may return lists in some places this method
85
+ can be used to enforce a list as return value.
86
+ """
87
+ rv = self.visit(node, *args, **kwargs)
88
+
89
+ if not isinstance(rv, list):
90
+ return [rv]
91
+
92
+ return rv
janus/lib/python3.10/site-packages/numpy/_core/__pycache__/__init__.cpython-310.pyc ADDED
Binary file (4.12 kB). View file
 
janus/lib/python3.10/site-packages/numpy/_core/__pycache__/_add_newdocs_scalars.cpython-310.pyc ADDED
Binary file (11.3 kB). View file
 
janus/lib/python3.10/site-packages/numpy/_core/__pycache__/_asarray.cpython-310.pyc ADDED
Binary file (3.78 kB). View file
 
janus/lib/python3.10/site-packages/numpy/_core/__pycache__/_string_helpers.cpython-310.pyc ADDED
Binary file (3 kB). View file
 
janus/lib/python3.10/site-packages/numpy/_core/__pycache__/cversions.cpython-310.pyc ADDED
Binary file (544 Bytes). View file
 
janus/lib/python3.10/site-packages/numpy/_core/__pycache__/defchararray.cpython-310.pyc ADDED
Binary file (37 kB). View file
 
janus/lib/python3.10/site-packages/numpy/_core/__pycache__/getlimits.cpython-310.pyc ADDED
Binary file (19.3 kB). View file
 
janus/lib/python3.10/site-packages/numpy/_core/__pycache__/memmap.cpython-310.pyc ADDED
Binary file (10.8 kB). View file
 
janus/lib/python3.10/site-packages/numpy/_core/__pycache__/records.cpython-310.pyc ADDED
Binary file (29.2 kB). View file
 
janus/lib/python3.10/site-packages/numpy/_core/__pycache__/shape_base.cpython-310.pyc ADDED
Binary file (29.8 kB). View file
 
janus/lib/python3.10/site-packages/numpy/_core/include/numpy/arrayobject.h ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ #ifndef NUMPY_CORE_INCLUDE_NUMPY_ARRAYOBJECT_H_
2
+ #define NUMPY_CORE_INCLUDE_NUMPY_ARRAYOBJECT_H_
3
+ #define Py_ARRAYOBJECT_H
4
+
5
+ #include "ndarrayobject.h"
6
+
7
+ #endif /* NUMPY_CORE_INCLUDE_NUMPY_ARRAYOBJECT_H_ */
janus/lib/python3.10/site-packages/numpy/_core/include/numpy/npy_common.h ADDED
@@ -0,0 +1,1070 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #ifndef NUMPY_CORE_INCLUDE_NUMPY_NPY_COMMON_H_
2
+ #define NUMPY_CORE_INCLUDE_NUMPY_NPY_COMMON_H_
3
+
4
+ /* need Python.h for npy_intp, npy_uintp */
5
+ #include <Python.h>
6
+
7
+ /* numpconfig.h is auto-generated */
8
+ #include "numpyconfig.h"
9
+ #ifdef HAVE_NPY_CONFIG_H
10
+ #include <npy_config.h>
11
+ #endif
12
+
13
+ /*
14
+ * using static inline modifiers when defining npy_math functions
15
+ * allows the compiler to make optimizations when possible
16
+ */
17
+ #ifndef NPY_INLINE_MATH
18
+ #if defined(NPY_INTERNAL_BUILD) && NPY_INTERNAL_BUILD
19
+ #define NPY_INLINE_MATH 1
20
+ #else
21
+ #define NPY_INLINE_MATH 0
22
+ #endif
23
+ #endif
24
+
25
+ /*
26
+ * gcc does not unroll even with -O3
27
+ * use with care, unrolling on modern cpus rarely speeds things up
28
+ */
29
+ #ifdef HAVE_ATTRIBUTE_OPTIMIZE_UNROLL_LOOPS
30
+ #define NPY_GCC_UNROLL_LOOPS \
31
+ __attribute__((optimize("unroll-loops")))
32
+ #else
33
+ #define NPY_GCC_UNROLL_LOOPS
34
+ #endif
35
+
36
+ /* highest gcc optimization level, enabled autovectorizer */
37
+ #ifdef HAVE_ATTRIBUTE_OPTIMIZE_OPT_3
38
+ #define NPY_GCC_OPT_3 __attribute__((optimize("O3")))
39
+ #else
40
+ #define NPY_GCC_OPT_3
41
+ #endif
42
+
43
+ /*
44
+ * mark an argument (starting from 1) that must not be NULL and is not checked
45
+ * DO NOT USE IF FUNCTION CHECKS FOR NULL!! the compiler will remove the check
46
+ */
47
+ #ifdef HAVE_ATTRIBUTE_NONNULL
48
+ #define NPY_GCC_NONNULL(n) __attribute__((nonnull(n)))
49
+ #else
50
+ #define NPY_GCC_NONNULL(n)
51
+ #endif
52
+
53
+ /*
54
+ * give a hint to the compiler which branch is more likely or unlikely
55
+ * to occur, e.g. rare error cases:
56
+ *
57
+ * if (NPY_UNLIKELY(failure == 0))
58
+ * return NULL;
59
+ *
60
+ * the double !! is to cast the expression (e.g. NULL) to a boolean required by
61
+ * the intrinsic
62
+ */
63
+ #ifdef HAVE___BUILTIN_EXPECT
64
+ #define NPY_LIKELY(x) __builtin_expect(!!(x), 1)
65
+ #define NPY_UNLIKELY(x) __builtin_expect(!!(x), 0)
66
+ #else
67
+ #define NPY_LIKELY(x) (x)
68
+ #define NPY_UNLIKELY(x) (x)
69
+ #endif
70
+
71
+ #ifdef HAVE___BUILTIN_PREFETCH
72
+ /* unlike _mm_prefetch also works on non-x86 */
73
+ #define NPY_PREFETCH(x, rw, loc) __builtin_prefetch((x), (rw), (loc))
74
+ #else
75
+ #ifdef NPY_HAVE_SSE
76
+ /* _MM_HINT_ET[01] (rw = 1) unsupported, only available in gcc >= 4.9 */
77
+ #define NPY_PREFETCH(x, rw, loc) _mm_prefetch((x), loc == 0 ? _MM_HINT_NTA : \
78
+ (loc == 1 ? _MM_HINT_T2 : \
79
+ (loc == 2 ? _MM_HINT_T1 : \
80
+ (loc == 3 ? _MM_HINT_T0 : -1))))
81
+ #else
82
+ #define NPY_PREFETCH(x, rw,loc)
83
+ #endif
84
+ #endif
85
+
86
+ /* `NPY_INLINE` kept for backwards compatibility; use `inline` instead */
87
+ #if defined(_MSC_VER) && !defined(__clang__)
88
+ #define NPY_INLINE __inline
89
+ /* clang included here to handle clang-cl on Windows */
90
+ #elif defined(__GNUC__) || defined(__clang__)
91
+ #if defined(__STRICT_ANSI__)
92
+ #define NPY_INLINE __inline__
93
+ #else
94
+ #define NPY_INLINE inline
95
+ #endif
96
+ #else
97
+ #define NPY_INLINE
98
+ #endif
99
+
100
+ #ifdef _MSC_VER
101
+ #define NPY_FINLINE static __forceinline
102
+ #elif defined(__GNUC__)
103
+ #define NPY_FINLINE static inline __attribute__((always_inline))
104
+ #else
105
+ #define NPY_FINLINE static
106
+ #endif
107
+
108
+ #if defined(_MSC_VER)
109
+ #define NPY_NOINLINE static __declspec(noinline)
110
+ #elif defined(__GNUC__) || defined(__clang__)
111
+ #define NPY_NOINLINE static __attribute__((noinline))
112
+ #else
113
+ #define NPY_NOINLINE static
114
+ #endif
115
+
116
+ #ifdef __cplusplus
117
+ #define NPY_TLS thread_local
118
+ #elif defined(HAVE_THREAD_LOCAL)
119
+ #define NPY_TLS thread_local
120
+ #elif defined(HAVE__THREAD_LOCAL)
121
+ #define NPY_TLS _Thread_local
122
+ #elif defined(HAVE___THREAD)
123
+ #define NPY_TLS __thread
124
+ #elif defined(HAVE___DECLSPEC_THREAD_)
125
+ #define NPY_TLS __declspec(thread)
126
+ #else
127
+ #define NPY_TLS
128
+ #endif
129
+
130
+ #ifdef WITH_CPYCHECKER_RETURNS_BORROWED_REF_ATTRIBUTE
131
+ #define NPY_RETURNS_BORROWED_REF \
132
+ __attribute__((cpychecker_returns_borrowed_ref))
133
+ #else
134
+ #define NPY_RETURNS_BORROWED_REF
135
+ #endif
136
+
137
+ #ifdef WITH_CPYCHECKER_STEALS_REFERENCE_TO_ARG_ATTRIBUTE
138
+ #define NPY_STEALS_REF_TO_ARG(n) \
139
+ __attribute__((cpychecker_steals_reference_to_arg(n)))
140
+ #else
141
+ #define NPY_STEALS_REF_TO_ARG(n)
142
+ #endif
143
+
144
+ /* 64 bit file position support, also on win-amd64. Issue gh-2256 */
145
+ #if defined(_MSC_VER) && defined(_WIN64) && (_MSC_VER > 1400) || \
146
+ defined(__MINGW32__) || defined(__MINGW64__)
147
+ #include <io.h>
148
+
149
+ #define npy_fseek _fseeki64
150
+ #define npy_ftell _ftelli64
151
+ #define npy_lseek _lseeki64
152
+ #define npy_off_t npy_int64
153
+
154
+ #if NPY_SIZEOF_INT == 8
155
+ #define NPY_OFF_T_PYFMT "i"
156
+ #elif NPY_SIZEOF_LONG == 8
157
+ #define NPY_OFF_T_PYFMT "l"
158
+ #elif NPY_SIZEOF_LONGLONG == 8
159
+ #define NPY_OFF_T_PYFMT "L"
160
+ #else
161
+ #error Unsupported size for type off_t
162
+ #endif
163
+ #else
164
+ #ifdef HAVE_FSEEKO
165
+ #define npy_fseek fseeko
166
+ #else
167
+ #define npy_fseek fseek
168
+ #endif
169
+ #ifdef HAVE_FTELLO
170
+ #define npy_ftell ftello
171
+ #else
172
+ #define npy_ftell ftell
173
+ #endif
174
+ #include <sys/types.h>
175
+ #ifndef _WIN32
176
+ #include <unistd.h>
177
+ #endif
178
+ #define npy_lseek lseek
179
+ #define npy_off_t off_t
180
+
181
+ #if NPY_SIZEOF_OFF_T == NPY_SIZEOF_SHORT
182
+ #define NPY_OFF_T_PYFMT "h"
183
+ #elif NPY_SIZEOF_OFF_T == NPY_SIZEOF_INT
184
+ #define NPY_OFF_T_PYFMT "i"
185
+ #elif NPY_SIZEOF_OFF_T == NPY_SIZEOF_LONG
186
+ #define NPY_OFF_T_PYFMT "l"
187
+ #elif NPY_SIZEOF_OFF_T == NPY_SIZEOF_LONGLONG
188
+ #define NPY_OFF_T_PYFMT "L"
189
+ #else
190
+ #error Unsupported size for type off_t
191
+ #endif
192
+ #endif
193
+
194
+ /* enums for detected endianness */
195
+ enum {
196
+ NPY_CPU_UNKNOWN_ENDIAN,
197
+ NPY_CPU_LITTLE,
198
+ NPY_CPU_BIG
199
+ };
200
+
201
+ /*
202
+ * This is to typedef npy_intp to the appropriate size for Py_ssize_t.
203
+ * (Before NumPy 2.0 we used Py_intptr_t and Py_uintptr_t from `pyport.h`.)
204
+ */
205
+ typedef Py_ssize_t npy_intp;
206
+ typedef size_t npy_uintp;
207
+
208
+ /*
209
+ * Define sizes that were not defined in numpyconfig.h.
210
+ */
211
+ #define NPY_SIZEOF_CHAR 1
212
+ #define NPY_SIZEOF_BYTE 1
213
+ #define NPY_SIZEOF_DATETIME 8
214
+ #define NPY_SIZEOF_TIMEDELTA 8
215
+ #define NPY_SIZEOF_HALF 2
216
+ #define NPY_SIZEOF_CFLOAT NPY_SIZEOF_COMPLEX_FLOAT
217
+ #define NPY_SIZEOF_CDOUBLE NPY_SIZEOF_COMPLEX_DOUBLE
218
+ #define NPY_SIZEOF_CLONGDOUBLE NPY_SIZEOF_COMPLEX_LONGDOUBLE
219
+
220
+ #ifdef constchar
221
+ #undef constchar
222
+ #endif
223
+
224
+ #define NPY_SSIZE_T_PYFMT "n"
225
+ #define constchar char
226
+
227
+ /* NPY_INTP_FMT Note:
228
+ * Unlike the other NPY_*_FMT macros, which are used with PyOS_snprintf,
229
+ * NPY_INTP_FMT is used with PyErr_Format and PyUnicode_FromFormat. Those
230
+ * functions use different formatting codes that are portably specified
231
+ * according to the Python documentation. See issue gh-2388.
232
+ */
233
+ #if NPY_SIZEOF_INTP == NPY_SIZEOF_LONG
234
+ #define NPY_INTP NPY_LONG
235
+ #define NPY_UINTP NPY_ULONG
236
+ #define PyIntpArrType_Type PyLongArrType_Type
237
+ #define PyUIntpArrType_Type PyULongArrType_Type
238
+ #define NPY_MAX_INTP NPY_MAX_LONG
239
+ #define NPY_MIN_INTP NPY_MIN_LONG
240
+ #define NPY_MAX_UINTP NPY_MAX_ULONG
241
+ #define NPY_INTP_FMT "ld"
242
+ #elif NPY_SIZEOF_INTP == NPY_SIZEOF_INT
243
+ #define NPY_INTP NPY_INT
244
+ #define NPY_UINTP NPY_UINT
245
+ #define PyIntpArrType_Type PyIntArrType_Type
246
+ #define PyUIntpArrType_Type PyUIntArrType_Type
247
+ #define NPY_MAX_INTP NPY_MAX_INT
248
+ #define NPY_MIN_INTP NPY_MIN_INT
249
+ #define NPY_MAX_UINTP NPY_MAX_UINT
250
+ #define NPY_INTP_FMT "d"
251
+ #elif defined(PY_LONG_LONG) && (NPY_SIZEOF_INTP == NPY_SIZEOF_LONGLONG)
252
+ #define NPY_INTP NPY_LONGLONG
253
+ #define NPY_UINTP NPY_ULONGLONG
254
+ #define PyIntpArrType_Type PyLongLongArrType_Type
255
+ #define PyUIntpArrType_Type PyULongLongArrType_Type
256
+ #define NPY_MAX_INTP NPY_MAX_LONGLONG
257
+ #define NPY_MIN_INTP NPY_MIN_LONGLONG
258
+ #define NPY_MAX_UINTP NPY_MAX_ULONGLONG
259
+ #define NPY_INTP_FMT "lld"
260
+ #else
261
+ #error "Failed to correctly define NPY_INTP and NPY_UINTP"
262
+ #endif
263
+
264
+
265
+ /*
266
+ * Some platforms don't define bool, long long, or long double.
267
+ * Handle that here.
268
+ */
269
+ #define NPY_BYTE_FMT "hhd"
270
+ #define NPY_UBYTE_FMT "hhu"
271
+ #define NPY_SHORT_FMT "hd"
272
+ #define NPY_USHORT_FMT "hu"
273
+ #define NPY_INT_FMT "d"
274
+ #define NPY_UINT_FMT "u"
275
+ #define NPY_LONG_FMT "ld"
276
+ #define NPY_ULONG_FMT "lu"
277
+ #define NPY_HALF_FMT "g"
278
+ #define NPY_FLOAT_FMT "g"
279
+ #define NPY_DOUBLE_FMT "g"
280
+
281
+
282
+ #ifdef PY_LONG_LONG
283
+ typedef PY_LONG_LONG npy_longlong;
284
+ typedef unsigned PY_LONG_LONG npy_ulonglong;
285
+ # ifdef _MSC_VER
286
+ # define NPY_LONGLONG_FMT "I64d"
287
+ # define NPY_ULONGLONG_FMT "I64u"
288
+ # else
289
+ # define NPY_LONGLONG_FMT "lld"
290
+ # define NPY_ULONGLONG_FMT "llu"
291
+ # endif
292
+ # ifdef _MSC_VER
293
+ # define NPY_LONGLONG_SUFFIX(x) (x##i64)
294
+ # define NPY_ULONGLONG_SUFFIX(x) (x##Ui64)
295
+ # else
296
+ # define NPY_LONGLONG_SUFFIX(x) (x##LL)
297
+ # define NPY_ULONGLONG_SUFFIX(x) (x##ULL)
298
+ # endif
299
+ #else
300
+ typedef long npy_longlong;
301
+ typedef unsigned long npy_ulonglong;
302
+ # define NPY_LONGLONG_SUFFIX(x) (x##L)
303
+ # define NPY_ULONGLONG_SUFFIX(x) (x##UL)
304
+ #endif
305
+
306
+
307
+ typedef unsigned char npy_bool;
308
+ #define NPY_FALSE 0
309
+ #define NPY_TRUE 1
310
+ /*
311
+ * `NPY_SIZEOF_LONGDOUBLE` isn't usually equal to sizeof(long double).
312
+ * In some certain cases, it may forced to be equal to sizeof(double)
313
+ * even against the compiler implementation and the same goes for
314
+ * `complex long double`.
315
+ *
316
+ * Therefore, avoid `long double`, use `npy_longdouble` instead,
317
+ * and when it comes to standard math functions make sure of using
318
+ * the double version when `NPY_SIZEOF_LONGDOUBLE` == `NPY_SIZEOF_DOUBLE`.
319
+ * For example:
320
+ * npy_longdouble *ptr, x;
321
+ * #if NPY_SIZEOF_LONGDOUBLE == NPY_SIZEOF_DOUBLE
322
+ * npy_longdouble r = modf(x, ptr);
323
+ * #else
324
+ * npy_longdouble r = modfl(x, ptr);
325
+ * #endif
326
+ *
327
+ * See https://github.com/numpy/numpy/issues/20348
328
+ */
329
+ #if NPY_SIZEOF_LONGDOUBLE == NPY_SIZEOF_DOUBLE
330
+ #define NPY_LONGDOUBLE_FMT "g"
331
+ #define longdouble_t double
332
+ typedef double npy_longdouble;
333
+ #else
334
+ #define NPY_LONGDOUBLE_FMT "Lg"
335
+ #define longdouble_t long double
336
+ typedef long double npy_longdouble;
337
+ #endif
338
+
339
+ #ifndef Py_USING_UNICODE
340
+ #error Must use Python with unicode enabled.
341
+ #endif
342
+
343
+
344
+ typedef signed char npy_byte;
345
+ typedef unsigned char npy_ubyte;
346
+ typedef unsigned short npy_ushort;
347
+ typedef unsigned int npy_uint;
348
+ typedef unsigned long npy_ulong;
349
+
350
+ /* These are for completeness */
351
+ typedef char npy_char;
352
+ typedef short npy_short;
353
+ typedef int npy_int;
354
+ typedef long npy_long;
355
+ typedef float npy_float;
356
+ typedef double npy_double;
357
+
358
+ typedef Py_hash_t npy_hash_t;
359
+ #define NPY_SIZEOF_HASH_T NPY_SIZEOF_INTP
360
+
361
+ #if defined(__cplusplus)
362
+
363
+ typedef struct
364
+ {
365
+ double _Val[2];
366
+ } npy_cdouble;
367
+
368
+ typedef struct
369
+ {
370
+ float _Val[2];
371
+ } npy_cfloat;
372
+
373
+ typedef struct
374
+ {
375
+ long double _Val[2];
376
+ } npy_clongdouble;
377
+
378
+ #else
379
+
380
+ #include <complex.h>
381
+
382
+
383
+ #if defined(_MSC_VER) && !defined(__INTEL_COMPILER)
384
+ typedef _Dcomplex npy_cdouble;
385
+ typedef _Fcomplex npy_cfloat;
386
+ typedef _Lcomplex npy_clongdouble;
387
+ #else /* !defined(_MSC_VER) || defined(__INTEL_COMPILER) */
388
+ typedef double _Complex npy_cdouble;
389
+ typedef float _Complex npy_cfloat;
390
+ typedef longdouble_t _Complex npy_clongdouble;
391
+ #endif
392
+
393
+ #endif
394
+
395
+ /*
396
+ * numarray-style bit-width typedefs
397
+ */
398
+ #define NPY_MAX_INT8 127
399
+ #define NPY_MIN_INT8 -128
400
+ #define NPY_MAX_UINT8 255
401
+ #define NPY_MAX_INT16 32767
402
+ #define NPY_MIN_INT16 -32768
403
+ #define NPY_MAX_UINT16 65535
404
+ #define NPY_MAX_INT32 2147483647
405
+ #define NPY_MIN_INT32 (-NPY_MAX_INT32 - 1)
406
+ #define NPY_MAX_UINT32 4294967295U
407
+ #define NPY_MAX_INT64 NPY_LONGLONG_SUFFIX(9223372036854775807)
408
+ #define NPY_MIN_INT64 (-NPY_MAX_INT64 - NPY_LONGLONG_SUFFIX(1))
409
+ #define NPY_MAX_UINT64 NPY_ULONGLONG_SUFFIX(18446744073709551615)
410
+ #define NPY_MAX_INT128 NPY_LONGLONG_SUFFIX(85070591730234615865843651857942052864)
411
+ #define NPY_MIN_INT128 (-NPY_MAX_INT128 - NPY_LONGLONG_SUFFIX(1))
412
+ #define NPY_MAX_UINT128 NPY_ULONGLONG_SUFFIX(170141183460469231731687303715884105728)
413
+ #define NPY_MAX_INT256 NPY_LONGLONG_SUFFIX(57896044618658097711785492504343953926634992332820282019728792003956564819967)
414
+ #define NPY_MIN_INT256 (-NPY_MAX_INT256 - NPY_LONGLONG_SUFFIX(1))
415
+ #define NPY_MAX_UINT256 NPY_ULONGLONG_SUFFIX(115792089237316195423570985008687907853269984665640564039457584007913129639935)
416
+ #define NPY_MIN_DATETIME NPY_MIN_INT64
417
+ #define NPY_MAX_DATETIME NPY_MAX_INT64
418
+ #define NPY_MIN_TIMEDELTA NPY_MIN_INT64
419
+ #define NPY_MAX_TIMEDELTA NPY_MAX_INT64
420
+
421
+ /* Need to find the number of bits for each type and
422
+ make definitions accordingly.
423
+
424
+ C states that sizeof(char) == 1 by definition
425
+
426
+ So, just using the sizeof keyword won't help.
427
+
428
+ It also looks like Python itself uses sizeof(char) quite a
429
+ bit, which by definition should be 1 all the time.
430
+
431
+ Idea: Make Use of CHAR_BIT which should tell us how many
432
+ BITS per CHARACTER
433
+ */
434
+
435
+ /* Include platform definitions -- These are in the C89/90 standard */
436
+ #include <limits.h>
437
+ #define NPY_MAX_BYTE SCHAR_MAX
438
+ #define NPY_MIN_BYTE SCHAR_MIN
439
+ #define NPY_MAX_UBYTE UCHAR_MAX
440
+ #define NPY_MAX_SHORT SHRT_MAX
441
+ #define NPY_MIN_SHORT SHRT_MIN
442
+ #define NPY_MAX_USHORT USHRT_MAX
443
+ #define NPY_MAX_INT INT_MAX
444
+ #ifndef INT_MIN
445
+ #define INT_MIN (-INT_MAX - 1)
446
+ #endif
447
+ #define NPY_MIN_INT INT_MIN
448
+ #define NPY_MAX_UINT UINT_MAX
449
+ #define NPY_MAX_LONG LONG_MAX
450
+ #define NPY_MIN_LONG LONG_MIN
451
+ #define NPY_MAX_ULONG ULONG_MAX
452
+
453
+ #define NPY_BITSOF_BOOL (sizeof(npy_bool) * CHAR_BIT)
454
+ #define NPY_BITSOF_CHAR CHAR_BIT
455
+ #define NPY_BITSOF_BYTE (NPY_SIZEOF_BYTE * CHAR_BIT)
456
+ #define NPY_BITSOF_SHORT (NPY_SIZEOF_SHORT * CHAR_BIT)
457
+ #define NPY_BITSOF_INT (NPY_SIZEOF_INT * CHAR_BIT)
458
+ #define NPY_BITSOF_LONG (NPY_SIZEOF_LONG * CHAR_BIT)
459
+ #define NPY_BITSOF_LONGLONG (NPY_SIZEOF_LONGLONG * CHAR_BIT)
460
+ #define NPY_BITSOF_INTP (NPY_SIZEOF_INTP * CHAR_BIT)
461
+ #define NPY_BITSOF_HALF (NPY_SIZEOF_HALF * CHAR_BIT)
462
+ #define NPY_BITSOF_FLOAT (NPY_SIZEOF_FLOAT * CHAR_BIT)
463
+ #define NPY_BITSOF_DOUBLE (NPY_SIZEOF_DOUBLE * CHAR_BIT)
464
+ #define NPY_BITSOF_LONGDOUBLE (NPY_SIZEOF_LONGDOUBLE * CHAR_BIT)
465
+ #define NPY_BITSOF_CFLOAT (NPY_SIZEOF_CFLOAT * CHAR_BIT)
466
+ #define NPY_BITSOF_CDOUBLE (NPY_SIZEOF_CDOUBLE * CHAR_BIT)
467
+ #define NPY_BITSOF_CLONGDOUBLE (NPY_SIZEOF_CLONGDOUBLE * CHAR_BIT)
468
+ #define NPY_BITSOF_DATETIME (NPY_SIZEOF_DATETIME * CHAR_BIT)
469
+ #define NPY_BITSOF_TIMEDELTA (NPY_SIZEOF_TIMEDELTA * CHAR_BIT)
470
+
471
+ #if NPY_BITSOF_LONG == 8
472
+ #define NPY_INT8 NPY_LONG
473
+ #define NPY_UINT8 NPY_ULONG
474
+ typedef long npy_int8;
475
+ typedef unsigned long npy_uint8;
476
+ #define PyInt8ScalarObject PyLongScalarObject
477
+ #define PyInt8ArrType_Type PyLongArrType_Type
478
+ #define PyUInt8ScalarObject PyULongScalarObject
479
+ #define PyUInt8ArrType_Type PyULongArrType_Type
480
+ #define NPY_INT8_FMT NPY_LONG_FMT
481
+ #define NPY_UINT8_FMT NPY_ULONG_FMT
482
+ #elif NPY_BITSOF_LONG == 16
483
+ #define NPY_INT16 NPY_LONG
484
+ #define NPY_UINT16 NPY_ULONG
485
+ typedef long npy_int16;
486
+ typedef unsigned long npy_uint16;
487
+ #define PyInt16ScalarObject PyLongScalarObject
488
+ #define PyInt16ArrType_Type PyLongArrType_Type
489
+ #define PyUInt16ScalarObject PyULongScalarObject
490
+ #define PyUInt16ArrType_Type PyULongArrType_Type
491
+ #define NPY_INT16_FMT NPY_LONG_FMT
492
+ #define NPY_UINT16_FMT NPY_ULONG_FMT
493
+ #elif NPY_BITSOF_LONG == 32
494
+ #define NPY_INT32 NPY_LONG
495
+ #define NPY_UINT32 NPY_ULONG
496
+ typedef long npy_int32;
497
+ typedef unsigned long npy_uint32;
498
+ typedef unsigned long npy_ucs4;
499
+ #define PyInt32ScalarObject PyLongScalarObject
500
+ #define PyInt32ArrType_Type PyLongArrType_Type
501
+ #define PyUInt32ScalarObject PyULongScalarObject
502
+ #define PyUInt32ArrType_Type PyULongArrType_Type
503
+ #define NPY_INT32_FMT NPY_LONG_FMT
504
+ #define NPY_UINT32_FMT NPY_ULONG_FMT
505
+ #elif NPY_BITSOF_LONG == 64
506
+ #define NPY_INT64 NPY_LONG
507
+ #define NPY_UINT64 NPY_ULONG
508
+ typedef long npy_int64;
509
+ typedef unsigned long npy_uint64;
510
+ #define PyInt64ScalarObject PyLongScalarObject
511
+ #define PyInt64ArrType_Type PyLongArrType_Type
512
+ #define PyUInt64ScalarObject PyULongScalarObject
513
+ #define PyUInt64ArrType_Type PyULongArrType_Type
514
+ #define NPY_INT64_FMT NPY_LONG_FMT
515
+ #define NPY_UINT64_FMT NPY_ULONG_FMT
516
+ #define MyPyLong_FromInt64 PyLong_FromLong
517
+ #define MyPyLong_AsInt64 PyLong_AsLong
518
+ #elif NPY_BITSOF_LONG == 128
519
+ #define NPY_INT128 NPY_LONG
520
+ #define NPY_UINT128 NPY_ULONG
521
+ typedef long npy_int128;
522
+ typedef unsigned long npy_uint128;
523
+ #define PyInt128ScalarObject PyLongScalarObject
524
+ #define PyInt128ArrType_Type PyLongArrType_Type
525
+ #define PyUInt128ScalarObject PyULongScalarObject
526
+ #define PyUInt128ArrType_Type PyULongArrType_Type
527
+ #define NPY_INT128_FMT NPY_LONG_FMT
528
+ #define NPY_UINT128_FMT NPY_ULONG_FMT
529
+ #endif
530
+
531
+ #if NPY_BITSOF_LONGLONG == 8
532
+ # ifndef NPY_INT8
533
+ # define NPY_INT8 NPY_LONGLONG
534
+ # define NPY_UINT8 NPY_ULONGLONG
535
+ typedef npy_longlong npy_int8;
536
+ typedef npy_ulonglong npy_uint8;
537
+ # define PyInt8ScalarObject PyLongLongScalarObject
538
+ # define PyInt8ArrType_Type PyLongLongArrType_Type
539
+ # define PyUInt8ScalarObject PyULongLongScalarObject
540
+ # define PyUInt8ArrType_Type PyULongLongArrType_Type
541
+ #define NPY_INT8_FMT NPY_LONGLONG_FMT
542
+ #define NPY_UINT8_FMT NPY_ULONGLONG_FMT
543
+ # endif
544
+ # define NPY_MAX_LONGLONG NPY_MAX_INT8
545
+ # define NPY_MIN_LONGLONG NPY_MIN_INT8
546
+ # define NPY_MAX_ULONGLONG NPY_MAX_UINT8
547
+ #elif NPY_BITSOF_LONGLONG == 16
548
+ # ifndef NPY_INT16
549
+ # define NPY_INT16 NPY_LONGLONG
550
+ # define NPY_UINT16 NPY_ULONGLONG
551
+ typedef npy_longlong npy_int16;
552
+ typedef npy_ulonglong npy_uint16;
553
+ # define PyInt16ScalarObject PyLongLongScalarObject
554
+ # define PyInt16ArrType_Type PyLongLongArrType_Type
555
+ # define PyUInt16ScalarObject PyULongLongScalarObject
556
+ # define PyUInt16ArrType_Type PyULongLongArrType_Type
557
+ #define NPY_INT16_FMT NPY_LONGLONG_FMT
558
+ #define NPY_UINT16_FMT NPY_ULONGLONG_FMT
559
+ # endif
560
+ # define NPY_MAX_LONGLONG NPY_MAX_INT16
561
+ # define NPY_MIN_LONGLONG NPY_MIN_INT16
562
+ # define NPY_MAX_ULONGLONG NPY_MAX_UINT16
563
+ #elif NPY_BITSOF_LONGLONG == 32
564
+ # ifndef NPY_INT32
565
+ # define NPY_INT32 NPY_LONGLONG
566
+ # define NPY_UINT32 NPY_ULONGLONG
567
+ typedef npy_longlong npy_int32;
568
+ typedef npy_ulonglong npy_uint32;
569
+ typedef npy_ulonglong npy_ucs4;
570
+ # define PyInt32ScalarObject PyLongLongScalarObject
571
+ # define PyInt32ArrType_Type PyLongLongArrType_Type
572
+ # define PyUInt32ScalarObject PyULongLongScalarObject
573
+ # define PyUInt32ArrType_Type PyULongLongArrType_Type
574
+ #define NPY_INT32_FMT NPY_LONGLONG_FMT
575
+ #define NPY_UINT32_FMT NPY_ULONGLONG_FMT
576
+ # endif
577
+ # define NPY_MAX_LONGLONG NPY_MAX_INT32
578
+ # define NPY_MIN_LONGLONG NPY_MIN_INT32
579
+ # define NPY_MAX_ULONGLONG NPY_MAX_UINT32
580
+ #elif NPY_BITSOF_LONGLONG == 64
581
+ # ifndef NPY_INT64
582
+ # define NPY_INT64 NPY_LONGLONG
583
+ # define NPY_UINT64 NPY_ULONGLONG
584
+ typedef npy_longlong npy_int64;
585
+ typedef npy_ulonglong npy_uint64;
586
+ # define PyInt64ScalarObject PyLongLongScalarObject
587
+ # define PyInt64ArrType_Type PyLongLongArrType_Type
588
+ # define PyUInt64ScalarObject PyULongLongScalarObject
589
+ # define PyUInt64ArrType_Type PyULongLongArrType_Type
590
+ #define NPY_INT64_FMT NPY_LONGLONG_FMT
591
+ #define NPY_UINT64_FMT NPY_ULONGLONG_FMT
592
+ # define MyPyLong_FromInt64 PyLong_FromLongLong
593
+ # define MyPyLong_AsInt64 PyLong_AsLongLong
594
+ # endif
595
+ # define NPY_MAX_LONGLONG NPY_MAX_INT64
596
+ # define NPY_MIN_LONGLONG NPY_MIN_INT64
597
+ # define NPY_MAX_ULONGLONG NPY_MAX_UINT64
598
+ #elif NPY_BITSOF_LONGLONG == 128
599
+ # ifndef NPY_INT128
600
+ # define NPY_INT128 NPY_LONGLONG
601
+ # define NPY_UINT128 NPY_ULONGLONG
602
+ typedef npy_longlong npy_int128;
603
+ typedef npy_ulonglong npy_uint128;
604
+ # define PyInt128ScalarObject PyLongLongScalarObject
605
+ # define PyInt128ArrType_Type PyLongLongArrType_Type
606
+ # define PyUInt128ScalarObject PyULongLongScalarObject
607
+ # define PyUInt128ArrType_Type PyULongLongArrType_Type
608
+ #define NPY_INT128_FMT NPY_LONGLONG_FMT
609
+ #define NPY_UINT128_FMT NPY_ULONGLONG_FMT
610
+ # endif
611
+ # define NPY_MAX_LONGLONG NPY_MAX_INT128
612
+ # define NPY_MIN_LONGLONG NPY_MIN_INT128
613
+ # define NPY_MAX_ULONGLONG NPY_MAX_UINT128
614
+ #elif NPY_BITSOF_LONGLONG == 256
615
+ # define NPY_INT256 NPY_LONGLONG
616
+ # define NPY_UINT256 NPY_ULONGLONG
617
+ typedef npy_longlong npy_int256;
618
+ typedef npy_ulonglong npy_uint256;
619
+ # define PyInt256ScalarObject PyLongLongScalarObject
620
+ # define PyInt256ArrType_Type PyLongLongArrType_Type
621
+ # define PyUInt256ScalarObject PyULongLongScalarObject
622
+ # define PyUInt256ArrType_Type PyULongLongArrType_Type
623
+ #define NPY_INT256_FMT NPY_LONGLONG_FMT
624
+ #define NPY_UINT256_FMT NPY_ULONGLONG_FMT
625
+ # define NPY_MAX_LONGLONG NPY_MAX_INT256
626
+ # define NPY_MIN_LONGLONG NPY_MIN_INT256
627
+ # define NPY_MAX_ULONGLONG NPY_MAX_UINT256
628
+ #endif
629
+
630
+ #if NPY_BITSOF_INT == 8
631
+ #ifndef NPY_INT8
632
+ #define NPY_INT8 NPY_INT
633
+ #define NPY_UINT8 NPY_UINT
634
+ typedef int npy_int8;
635
+ typedef unsigned int npy_uint8;
636
+ # define PyInt8ScalarObject PyIntScalarObject
637
+ # define PyInt8ArrType_Type PyIntArrType_Type
638
+ # define PyUInt8ScalarObject PyUIntScalarObject
639
+ # define PyUInt8ArrType_Type PyUIntArrType_Type
640
+ #define NPY_INT8_FMT NPY_INT_FMT
641
+ #define NPY_UINT8_FMT NPY_UINT_FMT
642
+ #endif
643
+ #elif NPY_BITSOF_INT == 16
644
+ #ifndef NPY_INT16
645
+ #define NPY_INT16 NPY_INT
646
+ #define NPY_UINT16 NPY_UINT
647
+ typedef int npy_int16;
648
+ typedef unsigned int npy_uint16;
649
+ # define PyInt16ScalarObject PyIntScalarObject
650
+ # define PyInt16ArrType_Type PyIntArrType_Type
651
+ # define PyUInt16ScalarObject PyIntUScalarObject
652
+ # define PyUInt16ArrType_Type PyIntUArrType_Type
653
+ #define NPY_INT16_FMT NPY_INT_FMT
654
+ #define NPY_UINT16_FMT NPY_UINT_FMT
655
+ #endif
656
+ #elif NPY_BITSOF_INT == 32
657
+ #ifndef NPY_INT32
658
+ #define NPY_INT32 NPY_INT
659
+ #define NPY_UINT32 NPY_UINT
660
+ typedef int npy_int32;
661
+ typedef unsigned int npy_uint32;
662
+ typedef unsigned int npy_ucs4;
663
+ # define PyInt32ScalarObject PyIntScalarObject
664
+ # define PyInt32ArrType_Type PyIntArrType_Type
665
+ # define PyUInt32ScalarObject PyUIntScalarObject
666
+ # define PyUInt32ArrType_Type PyUIntArrType_Type
667
+ #define NPY_INT32_FMT NPY_INT_FMT
668
+ #define NPY_UINT32_FMT NPY_UINT_FMT
669
+ #endif
670
+ #elif NPY_BITSOF_INT == 64
671
+ #ifndef NPY_INT64
672
+ #define NPY_INT64 NPY_INT
673
+ #define NPY_UINT64 NPY_UINT
674
+ typedef int npy_int64;
675
+ typedef unsigned int npy_uint64;
676
+ # define PyInt64ScalarObject PyIntScalarObject
677
+ # define PyInt64ArrType_Type PyIntArrType_Type
678
+ # define PyUInt64ScalarObject PyUIntScalarObject
679
+ # define PyUInt64ArrType_Type PyUIntArrType_Type
680
+ #define NPY_INT64_FMT NPY_INT_FMT
681
+ #define NPY_UINT64_FMT NPY_UINT_FMT
682
+ # define MyPyLong_FromInt64 PyLong_FromLong
683
+ # define MyPyLong_AsInt64 PyLong_AsLong
684
+ #endif
685
+ #elif NPY_BITSOF_INT == 128
686
+ #ifndef NPY_INT128
687
+ #define NPY_INT128 NPY_INT
688
+ #define NPY_UINT128 NPY_UINT
689
+ typedef int npy_int128;
690
+ typedef unsigned int npy_uint128;
691
+ # define PyInt128ScalarObject PyIntScalarObject
692
+ # define PyInt128ArrType_Type PyIntArrType_Type
693
+ # define PyUInt128ScalarObject PyUIntScalarObject
694
+ # define PyUInt128ArrType_Type PyUIntArrType_Type
695
+ #define NPY_INT128_FMT NPY_INT_FMT
696
+ #define NPY_UINT128_FMT NPY_UINT_FMT
697
+ #endif
698
+ #endif
699
+
700
+ #if NPY_BITSOF_SHORT == 8
701
+ #ifndef NPY_INT8
702
+ #define NPY_INT8 NPY_SHORT
703
+ #define NPY_UINT8 NPY_USHORT
704
+ typedef short npy_int8;
705
+ typedef unsigned short npy_uint8;
706
+ # define PyInt8ScalarObject PyShortScalarObject
707
+ # define PyInt8ArrType_Type PyShortArrType_Type
708
+ # define PyUInt8ScalarObject PyUShortScalarObject
709
+ # define PyUInt8ArrType_Type PyUShortArrType_Type
710
+ #define NPY_INT8_FMT NPY_SHORT_FMT
711
+ #define NPY_UINT8_FMT NPY_USHORT_FMT
712
+ #endif
713
+ #elif NPY_BITSOF_SHORT == 16
714
+ #ifndef NPY_INT16
715
+ #define NPY_INT16 NPY_SHORT
716
+ #define NPY_UINT16 NPY_USHORT
717
+ typedef short npy_int16;
718
+ typedef unsigned short npy_uint16;
719
+ # define PyInt16ScalarObject PyShortScalarObject
720
+ # define PyInt16ArrType_Type PyShortArrType_Type
721
+ # define PyUInt16ScalarObject PyUShortScalarObject
722
+ # define PyUInt16ArrType_Type PyUShortArrType_Type
723
+ #define NPY_INT16_FMT NPY_SHORT_FMT
724
+ #define NPY_UINT16_FMT NPY_USHORT_FMT
725
+ #endif
726
+ #elif NPY_BITSOF_SHORT == 32
727
+ #ifndef NPY_INT32
728
+ #define NPY_INT32 NPY_SHORT
729
+ #define NPY_UINT32 NPY_USHORT
730
+ typedef short npy_int32;
731
+ typedef unsigned short npy_uint32;
732
+ typedef unsigned short npy_ucs4;
733
+ # define PyInt32ScalarObject PyShortScalarObject
734
+ # define PyInt32ArrType_Type PyShortArrType_Type
735
+ # define PyUInt32ScalarObject PyUShortScalarObject
736
+ # define PyUInt32ArrType_Type PyUShortArrType_Type
737
+ #define NPY_INT32_FMT NPY_SHORT_FMT
738
+ #define NPY_UINT32_FMT NPY_USHORT_FMT
739
+ #endif
740
+ #elif NPY_BITSOF_SHORT == 64
741
+ #ifndef NPY_INT64
742
+ #define NPY_INT64 NPY_SHORT
743
+ #define NPY_UINT64 NPY_USHORT
744
+ typedef short npy_int64;
745
+ typedef unsigned short npy_uint64;
746
+ # define PyInt64ScalarObject PyShortScalarObject
747
+ # define PyInt64ArrType_Type PyShortArrType_Type
748
+ # define PyUInt64ScalarObject PyUShortScalarObject
749
+ # define PyUInt64ArrType_Type PyUShortArrType_Type
750
+ #define NPY_INT64_FMT NPY_SHORT_FMT
751
+ #define NPY_UINT64_FMT NPY_USHORT_FMT
752
+ # define MyPyLong_FromInt64 PyLong_FromLong
753
+ # define MyPyLong_AsInt64 PyLong_AsLong
754
+ #endif
755
+ #elif NPY_BITSOF_SHORT == 128
756
+ #ifndef NPY_INT128
757
+ #define NPY_INT128 NPY_SHORT
758
+ #define NPY_UINT128 NPY_USHORT
759
+ typedef short npy_int128;
760
+ typedef unsigned short npy_uint128;
761
+ # define PyInt128ScalarObject PyShortScalarObject
762
+ # define PyInt128ArrType_Type PyShortArrType_Type
763
+ # define PyUInt128ScalarObject PyUShortScalarObject
764
+ # define PyUInt128ArrType_Type PyUShortArrType_Type
765
+ #define NPY_INT128_FMT NPY_SHORT_FMT
766
+ #define NPY_UINT128_FMT NPY_USHORT_FMT
767
+ #endif
768
+ #endif
769
+
770
+
771
+ #if NPY_BITSOF_CHAR == 8
772
+ #ifndef NPY_INT8
773
+ #define NPY_INT8 NPY_BYTE
774
+ #define NPY_UINT8 NPY_UBYTE
775
+ typedef signed char npy_int8;
776
+ typedef unsigned char npy_uint8;
777
+ # define PyInt8ScalarObject PyByteScalarObject
778
+ # define PyInt8ArrType_Type PyByteArrType_Type
779
+ # define PyUInt8ScalarObject PyUByteScalarObject
780
+ # define PyUInt8ArrType_Type PyUByteArrType_Type
781
+ #define NPY_INT8_FMT NPY_BYTE_FMT
782
+ #define NPY_UINT8_FMT NPY_UBYTE_FMT
783
+ #endif
784
+ #elif NPY_BITSOF_CHAR == 16
785
+ #ifndef NPY_INT16
786
+ #define NPY_INT16 NPY_BYTE
787
+ #define NPY_UINT16 NPY_UBYTE
788
+ typedef signed char npy_int16;
789
+ typedef unsigned char npy_uint16;
790
+ # define PyInt16ScalarObject PyByteScalarObject
791
+ # define PyInt16ArrType_Type PyByteArrType_Type
792
+ # define PyUInt16ScalarObject PyUByteScalarObject
793
+ # define PyUInt16ArrType_Type PyUByteArrType_Type
794
+ #define NPY_INT16_FMT NPY_BYTE_FMT
795
+ #define NPY_UINT16_FMT NPY_UBYTE_FMT
796
+ #endif
797
+ #elif NPY_BITSOF_CHAR == 32
798
+ #ifndef NPY_INT32
799
+ #define NPY_INT32 NPY_BYTE
800
+ #define NPY_UINT32 NPY_UBYTE
801
+ typedef signed char npy_int32;
802
+ typedef unsigned char npy_uint32;
803
+ typedef unsigned char npy_ucs4;
804
+ # define PyInt32ScalarObject PyByteScalarObject
805
+ # define PyInt32ArrType_Type PyByteArrType_Type
806
+ # define PyUInt32ScalarObject PyUByteScalarObject
807
+ # define PyUInt32ArrType_Type PyUByteArrType_Type
808
+ #define NPY_INT32_FMT NPY_BYTE_FMT
809
+ #define NPY_UINT32_FMT NPY_UBYTE_FMT
810
+ #endif
811
+ #elif NPY_BITSOF_CHAR == 64
812
+ #ifndef NPY_INT64
813
+ #define NPY_INT64 NPY_BYTE
814
+ #define NPY_UINT64 NPY_UBYTE
815
+ typedef signed char npy_int64;
816
+ typedef unsigned char npy_uint64;
817
+ # define PyInt64ScalarObject PyByteScalarObject
818
+ # define PyInt64ArrType_Type PyByteArrType_Type
819
+ # define PyUInt64ScalarObject PyUByteScalarObject
820
+ # define PyUInt64ArrType_Type PyUByteArrType_Type
821
+ #define NPY_INT64_FMT NPY_BYTE_FMT
822
+ #define NPY_UINT64_FMT NPY_UBYTE_FMT
823
+ # define MyPyLong_FromInt64 PyLong_FromLong
824
+ # define MyPyLong_AsInt64 PyLong_AsLong
825
+ #endif
826
+ #elif NPY_BITSOF_CHAR == 128
827
+ #ifndef NPY_INT128
828
+ #define NPY_INT128 NPY_BYTE
829
+ #define NPY_UINT128 NPY_UBYTE
830
+ typedef signed char npy_int128;
831
+ typedef unsigned char npy_uint128;
832
+ # define PyInt128ScalarObject PyByteScalarObject
833
+ # define PyInt128ArrType_Type PyByteArrType_Type
834
+ # define PyUInt128ScalarObject PyUByteScalarObject
835
+ # define PyUInt128ArrType_Type PyUByteArrType_Type
836
+ #define NPY_INT128_FMT NPY_BYTE_FMT
837
+ #define NPY_UINT128_FMT NPY_UBYTE_FMT
838
+ #endif
839
+ #endif
840
+
841
+
842
+
843
+ #if NPY_BITSOF_DOUBLE == 32
844
+ #ifndef NPY_FLOAT32
845
+ #define NPY_FLOAT32 NPY_DOUBLE
846
+ #define NPY_COMPLEX64 NPY_CDOUBLE
847
+ typedef double npy_float32;
848
+ typedef npy_cdouble npy_complex64;
849
+ # define PyFloat32ScalarObject PyDoubleScalarObject
850
+ # define PyComplex64ScalarObject PyCDoubleScalarObject
851
+ # define PyFloat32ArrType_Type PyDoubleArrType_Type
852
+ # define PyComplex64ArrType_Type PyCDoubleArrType_Type
853
+ #define NPY_FLOAT32_FMT NPY_DOUBLE_FMT
854
+ #define NPY_COMPLEX64_FMT NPY_CDOUBLE_FMT
855
+ #endif
856
+ #elif NPY_BITSOF_DOUBLE == 64
857
+ #ifndef NPY_FLOAT64
858
+ #define NPY_FLOAT64 NPY_DOUBLE
859
+ #define NPY_COMPLEX128 NPY_CDOUBLE
860
+ typedef double npy_float64;
861
+ typedef npy_cdouble npy_complex128;
862
+ # define PyFloat64ScalarObject PyDoubleScalarObject
863
+ # define PyComplex128ScalarObject PyCDoubleScalarObject
864
+ # define PyFloat64ArrType_Type PyDoubleArrType_Type
865
+ # define PyComplex128ArrType_Type PyCDoubleArrType_Type
866
+ #define NPY_FLOAT64_FMT NPY_DOUBLE_FMT
867
+ #define NPY_COMPLEX128_FMT NPY_CDOUBLE_FMT
868
+ #endif
869
+ #elif NPY_BITSOF_DOUBLE == 80
870
+ #ifndef NPY_FLOAT80
871
+ #define NPY_FLOAT80 NPY_DOUBLE
872
+ #define NPY_COMPLEX160 NPY_CDOUBLE
873
+ typedef double npy_float80;
874
+ typedef npy_cdouble npy_complex160;
875
+ # define PyFloat80ScalarObject PyDoubleScalarObject
876
+ # define PyComplex160ScalarObject PyCDoubleScalarObject
877
+ # define PyFloat80ArrType_Type PyDoubleArrType_Type
878
+ # define PyComplex160ArrType_Type PyCDoubleArrType_Type
879
+ #define NPY_FLOAT80_FMT NPY_DOUBLE_FMT
880
+ #define NPY_COMPLEX160_FMT NPY_CDOUBLE_FMT
881
+ #endif
882
+ #elif NPY_BITSOF_DOUBLE == 96
883
+ #ifndef NPY_FLOAT96
884
+ #define NPY_FLOAT96 NPY_DOUBLE
885
+ #define NPY_COMPLEX192 NPY_CDOUBLE
886
+ typedef double npy_float96;
887
+ typedef npy_cdouble npy_complex192;
888
+ # define PyFloat96ScalarObject PyDoubleScalarObject
889
+ # define PyComplex192ScalarObject PyCDoubleScalarObject
890
+ # define PyFloat96ArrType_Type PyDoubleArrType_Type
891
+ # define PyComplex192ArrType_Type PyCDoubleArrType_Type
892
+ #define NPY_FLOAT96_FMT NPY_DOUBLE_FMT
893
+ #define NPY_COMPLEX192_FMT NPY_CDOUBLE_FMT
894
+ #endif
895
+ #elif NPY_BITSOF_DOUBLE == 128
896
+ #ifndef NPY_FLOAT128
897
+ #define NPY_FLOAT128 NPY_DOUBLE
898
+ #define NPY_COMPLEX256 NPY_CDOUBLE
899
+ typedef double npy_float128;
900
+ typedef npy_cdouble npy_complex256;
901
+ # define PyFloat128ScalarObject PyDoubleScalarObject
902
+ # define PyComplex256ScalarObject PyCDoubleScalarObject
903
+ # define PyFloat128ArrType_Type PyDoubleArrType_Type
904
+ # define PyComplex256ArrType_Type PyCDoubleArrType_Type
905
+ #define NPY_FLOAT128_FMT NPY_DOUBLE_FMT
906
+ #define NPY_COMPLEX256_FMT NPY_CDOUBLE_FMT
907
+ #endif
908
+ #endif
909
+
910
+
911
+
912
+ #if NPY_BITSOF_FLOAT == 32
913
+ #ifndef NPY_FLOAT32
914
+ #define NPY_FLOAT32 NPY_FLOAT
915
+ #define NPY_COMPLEX64 NPY_CFLOAT
916
+ typedef float npy_float32;
917
+ typedef npy_cfloat npy_complex64;
918
+ # define PyFloat32ScalarObject PyFloatScalarObject
919
+ # define PyComplex64ScalarObject PyCFloatScalarObject
920
+ # define PyFloat32ArrType_Type PyFloatArrType_Type
921
+ # define PyComplex64ArrType_Type PyCFloatArrType_Type
922
+ #define NPY_FLOAT32_FMT NPY_FLOAT_FMT
923
+ #define NPY_COMPLEX64_FMT NPY_CFLOAT_FMT
924
+ #endif
925
+ #elif NPY_BITSOF_FLOAT == 64
926
+ #ifndef NPY_FLOAT64
927
+ #define NPY_FLOAT64 NPY_FLOAT
928
+ #define NPY_COMPLEX128 NPY_CFLOAT
929
+ typedef float npy_float64;
930
+ typedef npy_cfloat npy_complex128;
931
+ # define PyFloat64ScalarObject PyFloatScalarObject
932
+ # define PyComplex128ScalarObject PyCFloatScalarObject
933
+ # define PyFloat64ArrType_Type PyFloatArrType_Type
934
+ # define PyComplex128ArrType_Type PyCFloatArrType_Type
935
+ #define NPY_FLOAT64_FMT NPY_FLOAT_FMT
936
+ #define NPY_COMPLEX128_FMT NPY_CFLOAT_FMT
937
+ #endif
938
+ #elif NPY_BITSOF_FLOAT == 80
939
+ #ifndef NPY_FLOAT80
940
+ #define NPY_FLOAT80 NPY_FLOAT
941
+ #define NPY_COMPLEX160 NPY_CFLOAT
942
+ typedef float npy_float80;
943
+ typedef npy_cfloat npy_complex160;
944
+ # define PyFloat80ScalarObject PyFloatScalarObject
945
+ # define PyComplex160ScalarObject PyCFloatScalarObject
946
+ # define PyFloat80ArrType_Type PyFloatArrType_Type
947
+ # define PyComplex160ArrType_Type PyCFloatArrType_Type
948
+ #define NPY_FLOAT80_FMT NPY_FLOAT_FMT
949
+ #define NPY_COMPLEX160_FMT NPY_CFLOAT_FMT
950
+ #endif
951
+ #elif NPY_BITSOF_FLOAT == 96
952
+ #ifndef NPY_FLOAT96
953
+ #define NPY_FLOAT96 NPY_FLOAT
954
+ #define NPY_COMPLEX192 NPY_CFLOAT
955
+ typedef float npy_float96;
956
+ typedef npy_cfloat npy_complex192;
957
+ # define PyFloat96ScalarObject PyFloatScalarObject
958
+ # define PyComplex192ScalarObject PyCFloatScalarObject
959
+ # define PyFloat96ArrType_Type PyFloatArrType_Type
960
+ # define PyComplex192ArrType_Type PyCFloatArrType_Type
961
+ #define NPY_FLOAT96_FMT NPY_FLOAT_FMT
962
+ #define NPY_COMPLEX192_FMT NPY_CFLOAT_FMT
963
+ #endif
964
+ #elif NPY_BITSOF_FLOAT == 128
965
+ #ifndef NPY_FLOAT128
966
+ #define NPY_FLOAT128 NPY_FLOAT
967
+ #define NPY_COMPLEX256 NPY_CFLOAT
968
+ typedef float npy_float128;
969
+ typedef npy_cfloat npy_complex256;
970
+ # define PyFloat128ScalarObject PyFloatScalarObject
971
+ # define PyComplex256ScalarObject PyCFloatScalarObject
972
+ # define PyFloat128ArrType_Type PyFloatArrType_Type
973
+ # define PyComplex256ArrType_Type PyCFloatArrType_Type
974
+ #define NPY_FLOAT128_FMT NPY_FLOAT_FMT
975
+ #define NPY_COMPLEX256_FMT NPY_CFLOAT_FMT
976
+ #endif
977
+ #endif
978
+
979
+ /* half/float16 isn't a floating-point type in C */
980
+ #define NPY_FLOAT16 NPY_HALF
981
+ typedef npy_uint16 npy_half;
982
+ typedef npy_half npy_float16;
983
+
984
+ #if NPY_BITSOF_LONGDOUBLE == 32
985
+ #ifndef NPY_FLOAT32
986
+ #define NPY_FLOAT32 NPY_LONGDOUBLE
987
+ #define NPY_COMPLEX64 NPY_CLONGDOUBLE
988
+ typedef npy_longdouble npy_float32;
989
+ typedef npy_clongdouble npy_complex64;
990
+ # define PyFloat32ScalarObject PyLongDoubleScalarObject
991
+ # define PyComplex64ScalarObject PyCLongDoubleScalarObject
992
+ # define PyFloat32ArrType_Type PyLongDoubleArrType_Type
993
+ # define PyComplex64ArrType_Type PyCLongDoubleArrType_Type
994
+ #define NPY_FLOAT32_FMT NPY_LONGDOUBLE_FMT
995
+ #define NPY_COMPLEX64_FMT NPY_CLONGDOUBLE_FMT
996
+ #endif
997
+ #elif NPY_BITSOF_LONGDOUBLE == 64
998
+ #ifndef NPY_FLOAT64
999
+ #define NPY_FLOAT64 NPY_LONGDOUBLE
1000
+ #define NPY_COMPLEX128 NPY_CLONGDOUBLE
1001
+ typedef npy_longdouble npy_float64;
1002
+ typedef npy_clongdouble npy_complex128;
1003
+ # define PyFloat64ScalarObject PyLongDoubleScalarObject
1004
+ # define PyComplex128ScalarObject PyCLongDoubleScalarObject
1005
+ # define PyFloat64ArrType_Type PyLongDoubleArrType_Type
1006
+ # define PyComplex128ArrType_Type PyCLongDoubleArrType_Type
1007
+ #define NPY_FLOAT64_FMT NPY_LONGDOUBLE_FMT
1008
+ #define NPY_COMPLEX128_FMT NPY_CLONGDOUBLE_FMT
1009
+ #endif
1010
+ #elif NPY_BITSOF_LONGDOUBLE == 80
1011
+ #ifndef NPY_FLOAT80
1012
+ #define NPY_FLOAT80 NPY_LONGDOUBLE
1013
+ #define NPY_COMPLEX160 NPY_CLONGDOUBLE
1014
+ typedef npy_longdouble npy_float80;
1015
+ typedef npy_clongdouble npy_complex160;
1016
+ # define PyFloat80ScalarObject PyLongDoubleScalarObject
1017
+ # define PyComplex160ScalarObject PyCLongDoubleScalarObject
1018
+ # define PyFloat80ArrType_Type PyLongDoubleArrType_Type
1019
+ # define PyComplex160ArrType_Type PyCLongDoubleArrType_Type
1020
+ #define NPY_FLOAT80_FMT NPY_LONGDOUBLE_FMT
1021
+ #define NPY_COMPLEX160_FMT NPY_CLONGDOUBLE_FMT
1022
+ #endif
1023
+ #elif NPY_BITSOF_LONGDOUBLE == 96
1024
+ #ifndef NPY_FLOAT96
1025
+ #define NPY_FLOAT96 NPY_LONGDOUBLE
1026
+ #define NPY_COMPLEX192 NPY_CLONGDOUBLE
1027
+ typedef npy_longdouble npy_float96;
1028
+ typedef npy_clongdouble npy_complex192;
1029
+ # define PyFloat96ScalarObject PyLongDoubleScalarObject
1030
+ # define PyComplex192ScalarObject PyCLongDoubleScalarObject
1031
+ # define PyFloat96ArrType_Type PyLongDoubleArrType_Type
1032
+ # define PyComplex192ArrType_Type PyCLongDoubleArrType_Type
1033
+ #define NPY_FLOAT96_FMT NPY_LONGDOUBLE_FMT
1034
+ #define NPY_COMPLEX192_FMT NPY_CLONGDOUBLE_FMT
1035
+ #endif
1036
+ #elif NPY_BITSOF_LONGDOUBLE == 128
1037
+ #ifndef NPY_FLOAT128
1038
+ #define NPY_FLOAT128 NPY_LONGDOUBLE
1039
+ #define NPY_COMPLEX256 NPY_CLONGDOUBLE
1040
+ typedef npy_longdouble npy_float128;
1041
+ typedef npy_clongdouble npy_complex256;
1042
+ # define PyFloat128ScalarObject PyLongDoubleScalarObject
1043
+ # define PyComplex256ScalarObject PyCLongDoubleScalarObject
1044
+ # define PyFloat128ArrType_Type PyLongDoubleArrType_Type
1045
+ # define PyComplex256ArrType_Type PyCLongDoubleArrType_Type
1046
+ #define NPY_FLOAT128_FMT NPY_LONGDOUBLE_FMT
1047
+ #define NPY_COMPLEX256_FMT NPY_CLONGDOUBLE_FMT
1048
+ #endif
1049
+ #elif NPY_BITSOF_LONGDOUBLE == 256
1050
+ #define NPY_FLOAT256 NPY_LONGDOUBLE
1051
+ #define NPY_COMPLEX512 NPY_CLONGDOUBLE
1052
+ typedef npy_longdouble npy_float256;
1053
+ typedef npy_clongdouble npy_complex512;
1054
+ # define PyFloat256ScalarObject PyLongDoubleScalarObject
1055
+ # define PyComplex512ScalarObject PyCLongDoubleScalarObject
1056
+ # define PyFloat256ArrType_Type PyLongDoubleArrType_Type
1057
+ # define PyComplex512ArrType_Type PyCLongDoubleArrType_Type
1058
+ #define NPY_FLOAT256_FMT NPY_LONGDOUBLE_FMT
1059
+ #define NPY_COMPLEX512_FMT NPY_CLONGDOUBLE_FMT
1060
+ #endif
1061
+
1062
+ /* datetime typedefs */
1063
+ typedef npy_int64 npy_timedelta;
1064
+ typedef npy_int64 npy_datetime;
1065
+ #define NPY_DATETIME_FMT NPY_INT64_FMT
1066
+ #define NPY_TIMEDELTA_FMT NPY_INT64_FMT
1067
+
1068
+ /* End of typedefs for numarray style bit-width names */
1069
+
1070
+ #endif /* NUMPY_CORE_INCLUDE_NUMPY_NPY_COMMON_H_ */
janus/lib/python3.10/site-packages/numpy/_core/lib/npy-pkg-config/npymath.ini ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [meta]
2
+ Name=npymath
3
+ Description=Portable, core math library implementing C99 standard
4
+ Version=0.1
5
+
6
+ [variables]
7
+ pkgname=numpy._core
8
+ prefix=${pkgdir}
9
+ libdir=${prefix}/lib
10
+ includedir=${prefix}/include
11
+
12
+ [default]
13
+ Libs=-L${libdir} -lnpymath
14
+ Cflags=-I${includedir}
15
+ Requires=mlib
16
+
17
+ [msvc]
18
+ Libs=/LIBPATH:${libdir} npymath.lib
19
+ Cflags=/INCLUDE:${includedir}
20
+ Requires=mlib
janus/lib/python3.10/site-packages/numpy/_core/lib/pkgconfig/numpy.pc ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ prefix=${pcfiledir}/../..
2
+ includedir=${prefix}/include
3
+
4
+ Name: numpy
5
+ Description: NumPy is the fundamental package for scientific computing with Python.
6
+ Version: 2.2.2
7
+ Cflags: -I${includedir}
janus/lib/python3.10/site-packages/numpy/random/__pycache__/__init__.cpython-310.pyc ADDED
Binary file (7.42 kB). View file
 
janus/lib/python3.10/site-packages/numpy/random/__pycache__/_pickle.cpython-310.pyc ADDED
Binary file (2.47 kB). View file
 
janus/lib/python3.10/site-packages/numpy/random/_examples/numba/__pycache__/extending_distributions.cpython-310.pyc ADDED
Binary file (2.09 kB). View file