| | """Extension API for adding custom tags and behavior.""" |
| |
|
| | import pprint |
| | import re |
| | import typing as t |
| |
|
| | from markupsafe import Markup |
| |
|
| | from . import defaults |
| | from . import nodes |
| | from .environment import Environment |
| | from .exceptions import TemplateAssertionError |
| | from .exceptions import TemplateSyntaxError |
| | from .runtime import concat |
| | from .runtime import Context |
| | from .runtime import Undefined |
| | from .utils import import_string |
| | from .utils import pass_context |
| |
|
| | if t.TYPE_CHECKING: |
| | import typing_extensions as te |
| |
|
| | from .lexer import Token |
| | from .lexer import TokenStream |
| | from .parser import Parser |
| |
|
| | class _TranslationsBasic(te.Protocol): |
| | def gettext(self, message: str) -> str: ... |
| |
|
| | def ngettext(self, singular: str, plural: str, n: int) -> str: |
| | pass |
| |
|
| | class _TranslationsContext(_TranslationsBasic): |
| | def pgettext(self, context: str, message: str) -> str: ... |
| |
|
| | def npgettext( |
| | self, context: str, singular: str, plural: str, n: int |
| | ) -> str: ... |
| |
|
| | _SupportedTranslations = t.Union[_TranslationsBasic, _TranslationsContext] |
| |
|
| |
|
| | |
| | |
| | GETTEXT_FUNCTIONS: t.Tuple[str, ...] = ( |
| | "_", |
| | "gettext", |
| | "ngettext", |
| | "pgettext", |
| | "npgettext", |
| | ) |
| | _ws_re = re.compile(r"\s*\n\s*") |
| |
|
| |
|
| | class Extension: |
| | """Extensions can be used to add extra functionality to the Jinja template |
| | system at the parser level. Custom extensions are bound to an environment |
| | but may not store environment specific data on `self`. The reason for |
| | this is that an extension can be bound to another environment (for |
| | overlays) by creating a copy and reassigning the `environment` attribute. |
| | |
| | As extensions are created by the environment they cannot accept any |
| | arguments for configuration. One may want to work around that by using |
| | a factory function, but that is not possible as extensions are identified |
| | by their import name. The correct way to configure the extension is |
| | storing the configuration values on the environment. Because this way the |
| | environment ends up acting as central configuration storage the |
| | attributes may clash which is why extensions have to ensure that the names |
| | they choose for configuration are not too generic. ``prefix`` for example |
| | is a terrible name, ``fragment_cache_prefix`` on the other hand is a good |
| | name as includes the name of the extension (fragment cache). |
| | """ |
| |
|
| | identifier: t.ClassVar[str] |
| |
|
| | def __init_subclass__(cls) -> None: |
| | cls.identifier = f"{cls.__module__}.{cls.__name__}" |
| |
|
| | |
| | tags: t.Set[str] = set() |
| |
|
| | |
| | |
| | |
| | |
| | |
| | priority = 100 |
| |
|
| | def __init__(self, environment: Environment) -> None: |
| | self.environment = environment |
| |
|
| | def bind(self, environment: Environment) -> "te.Self": |
| | """Create a copy of this extension bound to another environment.""" |
| | rv = object.__new__(self.__class__) |
| | rv.__dict__.update(self.__dict__) |
| | rv.environment = environment |
| | return rv |
| |
|
| | def preprocess( |
| | self, source: str, name: t.Optional[str], filename: t.Optional[str] = None |
| | ) -> str: |
| | """This method is called before the actual lexing and can be used to |
| | preprocess the source. The `filename` is optional. The return value |
| | must be the preprocessed source. |
| | """ |
| | return source |
| |
|
| | def filter_stream( |
| | self, stream: "TokenStream" |
| | ) -> t.Union["TokenStream", t.Iterable["Token"]]: |
| | """It's passed a :class:`~jinja2.lexer.TokenStream` that can be used |
| | to filter tokens returned. This method has to return an iterable of |
| | :class:`~jinja2.lexer.Token`\\s, but it doesn't have to return a |
| | :class:`~jinja2.lexer.TokenStream`. |
| | """ |
| | return stream |
| |
|
| | def parse(self, parser: "Parser") -> t.Union[nodes.Node, t.List[nodes.Node]]: |
| | """If any of the :attr:`tags` matched this method is called with the |
| | parser as first argument. The token the parser stream is pointing at |
| | is the name token that matched. This method has to return one or a |
| | list of multiple nodes. |
| | """ |
| | raise NotImplementedError() |
| |
|
| | def attr( |
| | self, name: str, lineno: t.Optional[int] = None |
| | ) -> nodes.ExtensionAttribute: |
| | """Return an attribute node for the current extension. This is useful |
| | to pass constants on extensions to generated template code. |
| | |
| | :: |
| | |
| | self.attr('_my_attribute', lineno=lineno) |
| | """ |
| | return nodes.ExtensionAttribute(self.identifier, name, lineno=lineno) |
| |
|
| | def call_method( |
| | self, |
| | name: str, |
| | args: t.Optional[t.List[nodes.Expr]] = None, |
| | kwargs: t.Optional[t.List[nodes.Keyword]] = None, |
| | dyn_args: t.Optional[nodes.Expr] = None, |
| | dyn_kwargs: t.Optional[nodes.Expr] = None, |
| | lineno: t.Optional[int] = None, |
| | ) -> nodes.Call: |
| | """Call a method of the extension. This is a shortcut for |
| | :meth:`attr` + :class:`jinja2.nodes.Call`. |
| | """ |
| | if args is None: |
| | args = [] |
| | if kwargs is None: |
| | kwargs = [] |
| | return nodes.Call( |
| | self.attr(name, lineno=lineno), |
| | args, |
| | kwargs, |
| | dyn_args, |
| | dyn_kwargs, |
| | lineno=lineno, |
| | ) |
| |
|
| |
|
| | @pass_context |
| | def _gettext_alias( |
| | __context: Context, *args: t.Any, **kwargs: t.Any |
| | ) -> t.Union[t.Any, Undefined]: |
| | return __context.call(__context.resolve("gettext"), *args, **kwargs) |
| |
|
| |
|
| | def _make_new_gettext(func: t.Callable[[str], str]) -> t.Callable[..., str]: |
| | @pass_context |
| | def gettext(__context: Context, __string: str, **variables: t.Any) -> str: |
| | rv = __context.call(func, __string) |
| | if __context.eval_ctx.autoescape: |
| | rv = Markup(rv) |
| | |
| | |
| | |
| | return rv % variables |
| |
|
| | return gettext |
| |
|
| |
|
| | def _make_new_ngettext(func: t.Callable[[str, str, int], str]) -> t.Callable[..., str]: |
| | @pass_context |
| | def ngettext( |
| | __context: Context, |
| | __singular: str, |
| | __plural: str, |
| | __num: int, |
| | **variables: t.Any, |
| | ) -> str: |
| | variables.setdefault("num", __num) |
| | rv = __context.call(func, __singular, __plural, __num) |
| | if __context.eval_ctx.autoescape: |
| | rv = Markup(rv) |
| | |
| | return rv % variables |
| |
|
| | return ngettext |
| |
|
| |
|
| | def _make_new_pgettext(func: t.Callable[[str, str], str]) -> t.Callable[..., str]: |
| | @pass_context |
| | def pgettext( |
| | __context: Context, __string_ctx: str, __string: str, **variables: t.Any |
| | ) -> str: |
| | variables.setdefault("context", __string_ctx) |
| | rv = __context.call(func, __string_ctx, __string) |
| |
|
| | if __context.eval_ctx.autoescape: |
| | rv = Markup(rv) |
| |
|
| | |
| | return rv % variables |
| |
|
| | return pgettext |
| |
|
| |
|
| | def _make_new_npgettext( |
| | func: t.Callable[[str, str, str, int], str], |
| | ) -> t.Callable[..., str]: |
| | @pass_context |
| | def npgettext( |
| | __context: Context, |
| | __string_ctx: str, |
| | __singular: str, |
| | __plural: str, |
| | __num: int, |
| | **variables: t.Any, |
| | ) -> str: |
| | variables.setdefault("context", __string_ctx) |
| | variables.setdefault("num", __num) |
| | rv = __context.call(func, __string_ctx, __singular, __plural, __num) |
| |
|
| | if __context.eval_ctx.autoescape: |
| | rv = Markup(rv) |
| |
|
| | |
| | return rv % variables |
| |
|
| | return npgettext |
| |
|
| |
|
| | class InternationalizationExtension(Extension): |
| | """This extension adds gettext support to Jinja.""" |
| |
|
| | tags = {"trans"} |
| |
|
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | def __init__(self, environment: Environment) -> None: |
| | super().__init__(environment) |
| | environment.globals["_"] = _gettext_alias |
| | environment.extend( |
| | install_gettext_translations=self._install, |
| | install_null_translations=self._install_null, |
| | install_gettext_callables=self._install_callables, |
| | uninstall_gettext_translations=self._uninstall, |
| | extract_translations=self._extract, |
| | newstyle_gettext=False, |
| | ) |
| |
|
| | def _install( |
| | self, translations: "_SupportedTranslations", newstyle: t.Optional[bool] = None |
| | ) -> None: |
| | |
| | |
| | gettext = getattr(translations, "ugettext", None) |
| | if gettext is None: |
| | gettext = translations.gettext |
| | ngettext = getattr(translations, "ungettext", None) |
| | if ngettext is None: |
| | ngettext = translations.ngettext |
| |
|
| | pgettext = getattr(translations, "pgettext", None) |
| | npgettext = getattr(translations, "npgettext", None) |
| | self._install_callables( |
| | gettext, ngettext, newstyle=newstyle, pgettext=pgettext, npgettext=npgettext |
| | ) |
| |
|
| | def _install_null(self, newstyle: t.Optional[bool] = None) -> None: |
| | import gettext |
| |
|
| | translations = gettext.NullTranslations() |
| |
|
| | if hasattr(translations, "pgettext"): |
| | |
| | pgettext = translations.pgettext |
| | else: |
| |
|
| | def pgettext(c: str, s: str) -> str: |
| | return s |
| |
|
| | if hasattr(translations, "npgettext"): |
| | npgettext = translations.npgettext |
| | else: |
| |
|
| | def npgettext(c: str, s: str, p: str, n: int) -> str: |
| | return s if n == 1 else p |
| |
|
| | self._install_callables( |
| | gettext=translations.gettext, |
| | ngettext=translations.ngettext, |
| | newstyle=newstyle, |
| | pgettext=pgettext, |
| | npgettext=npgettext, |
| | ) |
| |
|
| | def _install_callables( |
| | self, |
| | gettext: t.Callable[[str], str], |
| | ngettext: t.Callable[[str, str, int], str], |
| | newstyle: t.Optional[bool] = None, |
| | pgettext: t.Optional[t.Callable[[str, str], str]] = None, |
| | npgettext: t.Optional[t.Callable[[str, str, str, int], str]] = None, |
| | ) -> None: |
| | if newstyle is not None: |
| | self.environment.newstyle_gettext = newstyle |
| | if self.environment.newstyle_gettext: |
| | gettext = _make_new_gettext(gettext) |
| | ngettext = _make_new_ngettext(ngettext) |
| |
|
| | if pgettext is not None: |
| | pgettext = _make_new_pgettext(pgettext) |
| |
|
| | if npgettext is not None: |
| | npgettext = _make_new_npgettext(npgettext) |
| |
|
| | self.environment.globals.update( |
| | gettext=gettext, ngettext=ngettext, pgettext=pgettext, npgettext=npgettext |
| | ) |
| |
|
| | def _uninstall(self, translations: "_SupportedTranslations") -> None: |
| | for key in ("gettext", "ngettext", "pgettext", "npgettext"): |
| | self.environment.globals.pop(key, None) |
| |
|
| | def _extract( |
| | self, |
| | source: t.Union[str, nodes.Template], |
| | gettext_functions: t.Sequence[str] = GETTEXT_FUNCTIONS, |
| | ) -> t.Iterator[ |
| | t.Tuple[int, str, t.Union[t.Optional[str], t.Tuple[t.Optional[str], ...]]] |
| | ]: |
| | if isinstance(source, str): |
| | source = self.environment.parse(source) |
| | return extract_from_ast(source, gettext_functions) |
| |
|
| | def parse(self, parser: "Parser") -> t.Union[nodes.Node, t.List[nodes.Node]]: |
| | """Parse a translatable tag.""" |
| | lineno = next(parser.stream).lineno |
| |
|
| | context = None |
| | context_token = parser.stream.next_if("string") |
| |
|
| | if context_token is not None: |
| | context = context_token.value |
| |
|
| | |
| | |
| | |
| | plural_expr: t.Optional[nodes.Expr] = None |
| | plural_expr_assignment: t.Optional[nodes.Assign] = None |
| | num_called_num = False |
| | variables: t.Dict[str, nodes.Expr] = {} |
| | trimmed = None |
| | while parser.stream.current.type != "block_end": |
| | if variables: |
| | parser.stream.expect("comma") |
| |
|
| | |
| | if parser.stream.skip_if("colon"): |
| | break |
| |
|
| | token = parser.stream.expect("name") |
| | if token.value in variables: |
| | parser.fail( |
| | f"translatable variable {token.value!r} defined twice.", |
| | token.lineno, |
| | exc=TemplateAssertionError, |
| | ) |
| |
|
| | |
| | if parser.stream.current.type == "assign": |
| | next(parser.stream) |
| | variables[token.value] = var = parser.parse_expression() |
| | elif trimmed is None and token.value in ("trimmed", "notrimmed"): |
| | trimmed = token.value == "trimmed" |
| | continue |
| | else: |
| | variables[token.value] = var = nodes.Name(token.value, "load") |
| |
|
| | if plural_expr is None: |
| | if isinstance(var, nodes.Call): |
| | plural_expr = nodes.Name("_trans", "load") |
| | variables[token.value] = plural_expr |
| | plural_expr_assignment = nodes.Assign( |
| | nodes.Name("_trans", "store"), var |
| | ) |
| | else: |
| | plural_expr = var |
| | num_called_num = token.value == "num" |
| |
|
| | parser.stream.expect("block_end") |
| |
|
| | plural = None |
| | have_plural = False |
| | referenced = set() |
| |
|
| | |
| | singular_names, singular = self._parse_block(parser, True) |
| | if singular_names: |
| | referenced.update(singular_names) |
| | if plural_expr is None: |
| | plural_expr = nodes.Name(singular_names[0], "load") |
| | num_called_num = singular_names[0] == "num" |
| |
|
| | |
| | if parser.stream.current.test("name:pluralize"): |
| | have_plural = True |
| | next(parser.stream) |
| | if parser.stream.current.type != "block_end": |
| | token = parser.stream.expect("name") |
| | if token.value not in variables: |
| | parser.fail( |
| | f"unknown variable {token.value!r} for pluralization", |
| | token.lineno, |
| | exc=TemplateAssertionError, |
| | ) |
| | plural_expr = variables[token.value] |
| | num_called_num = token.value == "num" |
| | parser.stream.expect("block_end") |
| | plural_names, plural = self._parse_block(parser, False) |
| | next(parser.stream) |
| | referenced.update(plural_names) |
| | else: |
| | next(parser.stream) |
| |
|
| | |
| | for name in referenced: |
| | if name not in variables: |
| | variables[name] = nodes.Name(name, "load") |
| |
|
| | if not have_plural: |
| | plural_expr = None |
| | elif plural_expr is None: |
| | parser.fail("pluralize without variables", lineno) |
| |
|
| | if trimmed is None: |
| | trimmed = self.environment.policies["ext.i18n.trimmed"] |
| | if trimmed: |
| | singular = self._trim_whitespace(singular) |
| | if plural: |
| | plural = self._trim_whitespace(plural) |
| |
|
| | node = self._make_node( |
| | singular, |
| | plural, |
| | context, |
| | variables, |
| | plural_expr, |
| | bool(referenced), |
| | num_called_num and have_plural, |
| | ) |
| | node.set_lineno(lineno) |
| | if plural_expr_assignment is not None: |
| | return [plural_expr_assignment, node] |
| | else: |
| | return node |
| |
|
| | def _trim_whitespace(self, string: str, _ws_re: t.Pattern[str] = _ws_re) -> str: |
| | return _ws_re.sub(" ", string.strip()) |
| |
|
| | def _parse_block( |
| | self, parser: "Parser", allow_pluralize: bool |
| | ) -> t.Tuple[t.List[str], str]: |
| | """Parse until the next block tag with a given name.""" |
| | referenced = [] |
| | buf = [] |
| |
|
| | while True: |
| | if parser.stream.current.type == "data": |
| | buf.append(parser.stream.current.value.replace("%", "%%")) |
| | next(parser.stream) |
| | elif parser.stream.current.type == "variable_begin": |
| | next(parser.stream) |
| | name = parser.stream.expect("name").value |
| | referenced.append(name) |
| | buf.append(f"%({name})s") |
| | parser.stream.expect("variable_end") |
| | elif parser.stream.current.type == "block_begin": |
| | next(parser.stream) |
| | block_name = ( |
| | parser.stream.current.value |
| | if parser.stream.current.type == "name" |
| | else None |
| | ) |
| | if block_name == "endtrans": |
| | break |
| | elif block_name == "pluralize": |
| | if allow_pluralize: |
| | break |
| | parser.fail( |
| | "a translatable section can have only one pluralize section" |
| | ) |
| | elif block_name == "trans": |
| | parser.fail( |
| | "trans blocks can't be nested; did you mean `endtrans`?" |
| | ) |
| | parser.fail( |
| | f"control structures in translatable sections are not allowed; " |
| | f"saw `{block_name}`" |
| | ) |
| | elif parser.stream.eos: |
| | parser.fail("unclosed translation block") |
| | else: |
| | raise RuntimeError("internal parser error") |
| |
|
| | return referenced, concat(buf) |
| |
|
| | def _make_node( |
| | self, |
| | singular: str, |
| | plural: t.Optional[str], |
| | context: t.Optional[str], |
| | variables: t.Dict[str, nodes.Expr], |
| | plural_expr: t.Optional[nodes.Expr], |
| | vars_referenced: bool, |
| | num_called_num: bool, |
| | ) -> nodes.Output: |
| | """Generates a useful node from the data provided.""" |
| | newstyle = self.environment.newstyle_gettext |
| | node: nodes.Expr |
| |
|
| | |
| | |
| | if not vars_referenced and not newstyle: |
| | singular = singular.replace("%%", "%") |
| | if plural: |
| | plural = plural.replace("%%", "%") |
| |
|
| | func_name = "gettext" |
| | func_args: t.List[nodes.Expr] = [nodes.Const(singular)] |
| |
|
| | if context is not None: |
| | func_args.insert(0, nodes.Const(context)) |
| | func_name = f"p{func_name}" |
| |
|
| | if plural_expr is not None: |
| | func_name = f"n{func_name}" |
| | func_args.extend((nodes.Const(plural), plural_expr)) |
| |
|
| | node = nodes.Call(nodes.Name(func_name, "load"), func_args, [], None, None) |
| |
|
| | |
| | |
| | |
| | if newstyle: |
| | for key, value in variables.items(): |
| | |
| | |
| | if num_called_num and key == "num": |
| | continue |
| | node.kwargs.append(nodes.Keyword(key, value)) |
| |
|
| | |
| | else: |
| | |
| | |
| | node = nodes.MarkSafeIfAutoescape(node) |
| | if variables: |
| | node = nodes.Mod( |
| | node, |
| | nodes.Dict( |
| | [ |
| | nodes.Pair(nodes.Const(key), value) |
| | for key, value in variables.items() |
| | ] |
| | ), |
| | ) |
| | return nodes.Output([node]) |
| |
|
| |
|
| | class ExprStmtExtension(Extension): |
| | """Adds a `do` tag to Jinja that works like the print statement just |
| | that it doesn't print the return value. |
| | """ |
| |
|
| | tags = {"do"} |
| |
|
| | def parse(self, parser: "Parser") -> nodes.ExprStmt: |
| | node = nodes.ExprStmt(lineno=next(parser.stream).lineno) |
| | node.node = parser.parse_tuple() |
| | return node |
| |
|
| |
|
| | class LoopControlExtension(Extension): |
| | """Adds break and continue to the template engine.""" |
| |
|
| | tags = {"break", "continue"} |
| |
|
| | def parse(self, parser: "Parser") -> t.Union[nodes.Break, nodes.Continue]: |
| | token = next(parser.stream) |
| | if token.value == "break": |
| | return nodes.Break(lineno=token.lineno) |
| | return nodes.Continue(lineno=token.lineno) |
| |
|
| |
|
| | class DebugExtension(Extension): |
| | """A ``{% debug %}`` tag that dumps the available variables, |
| | filters, and tests. |
| | |
| | .. code-block:: html+jinja |
| | |
| | <pre>{% debug %}</pre> |
| | |
| | .. code-block:: text |
| | |
| | {'context': {'cycler': <class 'jinja2.utils.Cycler'>, |
| | ..., |
| | 'namespace': <class 'jinja2.utils.Namespace'>}, |
| | 'filters': ['abs', 'attr', 'batch', 'capitalize', 'center', 'count', 'd', |
| | ..., 'urlencode', 'urlize', 'wordcount', 'wordwrap', 'xmlattr'], |
| | 'tests': ['!=', '<', '<=', '==', '>', '>=', 'callable', 'defined', |
| | ..., 'odd', 'sameas', 'sequence', 'string', 'undefined', 'upper']} |
| | |
| | .. versionadded:: 2.11.0 |
| | """ |
| |
|
| | tags = {"debug"} |
| |
|
| | def parse(self, parser: "Parser") -> nodes.Output: |
| | lineno = parser.stream.expect("name:debug").lineno |
| | context = nodes.ContextReference() |
| | result = self.call_method("_render", [context], lineno=lineno) |
| | return nodes.Output([result], lineno=lineno) |
| |
|
| | def _render(self, context: Context) -> str: |
| | result = { |
| | "context": context.get_all(), |
| | "filters": sorted(self.environment.filters.keys()), |
| | "tests": sorted(self.environment.tests.keys()), |
| | } |
| |
|
| | |
| | return pprint.pformat(result, depth=3, compact=True) |
| |
|
| |
|
| | def extract_from_ast( |
| | ast: nodes.Template, |
| | gettext_functions: t.Sequence[str] = GETTEXT_FUNCTIONS, |
| | babel_style: bool = True, |
| | ) -> t.Iterator[ |
| | t.Tuple[int, str, t.Union[t.Optional[str], t.Tuple[t.Optional[str], ...]]] |
| | ]: |
| | """Extract localizable strings from the given template node. Per |
| | default this function returns matches in babel style that means non string |
| | parameters as well as keyword arguments are returned as `None`. This |
| | allows Babel to figure out what you really meant if you are using |
| | gettext functions that allow keyword arguments for placeholder expansion. |
| | If you don't want that behavior set the `babel_style` parameter to `False` |
| | which causes only strings to be returned and parameters are always stored |
| | in tuples. As a consequence invalid gettext calls (calls without a single |
| | string parameter or string parameters after non-string parameters) are |
| | skipped. |
| | |
| | This example explains the behavior: |
| | |
| | >>> from jinja2 import Environment |
| | >>> env = Environment() |
| | >>> node = env.parse('{{ (_("foo"), _(), ngettext("foo", "bar", 42)) }}') |
| | >>> list(extract_from_ast(node)) |
| | [(1, '_', 'foo'), (1, '_', ()), (1, 'ngettext', ('foo', 'bar', None))] |
| | >>> list(extract_from_ast(node, babel_style=False)) |
| | [(1, '_', ('foo',)), (1, 'ngettext', ('foo', 'bar'))] |
| | |
| | For every string found this function yields a ``(lineno, function, |
| | message)`` tuple, where: |
| | |
| | * ``lineno`` is the number of the line on which the string was found, |
| | * ``function`` is the name of the ``gettext`` function used (if the |
| | string was extracted from embedded Python code), and |
| | * ``message`` is the string, or a tuple of strings for functions |
| | with multiple string arguments. |
| | |
| | This extraction function operates on the AST and is because of that unable |
| | to extract any comments. For comment support you have to use the babel |
| | extraction interface or extract comments yourself. |
| | """ |
| | out: t.Union[t.Optional[str], t.Tuple[t.Optional[str], ...]] |
| |
|
| | for node in ast.find_all(nodes.Call): |
| | if ( |
| | not isinstance(node.node, nodes.Name) |
| | or node.node.name not in gettext_functions |
| | ): |
| | continue |
| |
|
| | strings: t.List[t.Optional[str]] = [] |
| |
|
| | for arg in node.args: |
| | if isinstance(arg, nodes.Const) and isinstance(arg.value, str): |
| | strings.append(arg.value) |
| | else: |
| | strings.append(None) |
| |
|
| | for _ in node.kwargs: |
| | strings.append(None) |
| | if node.dyn_args is not None: |
| | strings.append(None) |
| | if node.dyn_kwargs is not None: |
| | strings.append(None) |
| |
|
| | if not babel_style: |
| | out = tuple(x for x in strings if x is not None) |
| |
|
| | if not out: |
| | continue |
| | else: |
| | if len(strings) == 1: |
| | out = strings[0] |
| | else: |
| | out = tuple(strings) |
| |
|
| | yield node.lineno, node.node.name, out |
| |
|
| |
|
| | class _CommentFinder: |
| | """Helper class to find comments in a token stream. Can only |
| | find comments for gettext calls forwards. Once the comment |
| | from line 4 is found, a comment for line 1 will not return a |
| | usable value. |
| | """ |
| |
|
| | def __init__( |
| | self, tokens: t.Sequence[t.Tuple[int, str, str]], comment_tags: t.Sequence[str] |
| | ) -> None: |
| | self.tokens = tokens |
| | self.comment_tags = comment_tags |
| | self.offset = 0 |
| | self.last_lineno = 0 |
| |
|
| | def find_backwards(self, offset: int) -> t.List[str]: |
| | try: |
| | for _, token_type, token_value in reversed( |
| | self.tokens[self.offset : offset] |
| | ): |
| | if token_type in ("comment", "linecomment"): |
| | try: |
| | prefix, comment = token_value.split(None, 1) |
| | except ValueError: |
| | continue |
| | if prefix in self.comment_tags: |
| | return [comment.rstrip()] |
| | return [] |
| | finally: |
| | self.offset = offset |
| |
|
| | def find_comments(self, lineno: int) -> t.List[str]: |
| | if not self.comment_tags or self.last_lineno > lineno: |
| | return [] |
| | for idx, (token_lineno, _, _) in enumerate(self.tokens[self.offset :]): |
| | if token_lineno > lineno: |
| | return self.find_backwards(self.offset + idx) |
| | return self.find_backwards(len(self.tokens)) |
| |
|
| |
|
| | def babel_extract( |
| | fileobj: t.BinaryIO, |
| | keywords: t.Sequence[str], |
| | comment_tags: t.Sequence[str], |
| | options: t.Dict[str, t.Any], |
| | ) -> t.Iterator[ |
| | t.Tuple[ |
| | int, str, t.Union[t.Optional[str], t.Tuple[t.Optional[str], ...]], t.List[str] |
| | ] |
| | ]: |
| | """Babel extraction method for Jinja templates. |
| | |
| | .. versionchanged:: 2.3 |
| | Basic support for translation comments was added. If `comment_tags` |
| | is now set to a list of keywords for extraction, the extractor will |
| | try to find the best preceding comment that begins with one of the |
| | keywords. For best results, make sure to not have more than one |
| | gettext call in one line of code and the matching comment in the |
| | same line or the line before. |
| | |
| | .. versionchanged:: 2.5.1 |
| | The `newstyle_gettext` flag can be set to `True` to enable newstyle |
| | gettext calls. |
| | |
| | .. versionchanged:: 2.7 |
| | A `silent` option can now be provided. If set to `False` template |
| | syntax errors are propagated instead of being ignored. |
| | |
| | :param fileobj: the file-like object the messages should be extracted from |
| | :param keywords: a list of keywords (i.e. function names) that should be |
| | recognized as translation functions |
| | :param comment_tags: a list of translator tags to search for and include |
| | in the results. |
| | :param options: a dictionary of additional options (optional) |
| | :return: an iterator over ``(lineno, funcname, message, comments)`` tuples. |
| | (comments will be empty currently) |
| | """ |
| | extensions: t.Dict[t.Type[Extension], None] = {} |
| |
|
| | for extension_name in options.get("extensions", "").split(","): |
| | extension_name = extension_name.strip() |
| |
|
| | if not extension_name: |
| | continue |
| |
|
| | extensions[import_string(extension_name)] = None |
| |
|
| | if InternationalizationExtension not in extensions: |
| | extensions[InternationalizationExtension] = None |
| |
|
| | def getbool(options: t.Mapping[str, str], key: str, default: bool = False) -> bool: |
| | return options.get(key, str(default)).lower() in {"1", "on", "yes", "true"} |
| |
|
| | silent = getbool(options, "silent", True) |
| | environment = Environment( |
| | options.get("block_start_string", defaults.BLOCK_START_STRING), |
| | options.get("block_end_string", defaults.BLOCK_END_STRING), |
| | options.get("variable_start_string", defaults.VARIABLE_START_STRING), |
| | options.get("variable_end_string", defaults.VARIABLE_END_STRING), |
| | options.get("comment_start_string", defaults.COMMENT_START_STRING), |
| | options.get("comment_end_string", defaults.COMMENT_END_STRING), |
| | options.get("line_statement_prefix") or defaults.LINE_STATEMENT_PREFIX, |
| | options.get("line_comment_prefix") or defaults.LINE_COMMENT_PREFIX, |
| | getbool(options, "trim_blocks", defaults.TRIM_BLOCKS), |
| | getbool(options, "lstrip_blocks", defaults.LSTRIP_BLOCKS), |
| | defaults.NEWLINE_SEQUENCE, |
| | getbool(options, "keep_trailing_newline", defaults.KEEP_TRAILING_NEWLINE), |
| | tuple(extensions), |
| | cache_size=0, |
| | auto_reload=False, |
| | ) |
| |
|
| | if getbool(options, "trimmed"): |
| | environment.policies["ext.i18n.trimmed"] = True |
| | if getbool(options, "newstyle_gettext"): |
| | environment.newstyle_gettext = True |
| |
|
| | source = fileobj.read().decode(options.get("encoding", "utf-8")) |
| | try: |
| | node = environment.parse(source) |
| | tokens = list(environment.lex(environment.preprocess(source))) |
| | except TemplateSyntaxError: |
| | if not silent: |
| | raise |
| | |
| | return |
| |
|
| | finder = _CommentFinder(tokens, comment_tags) |
| | for lineno, func, message in extract_from_ast(node, keywords): |
| | yield lineno, func, message, finder.find_comments(lineno) |
| |
|
| |
|
| | |
| | i18n = InternationalizationExtension |
| | do = ExprStmtExtension |
| | loopcontrols = LoopControlExtension |
| | debug = DebugExtension |
| |
|