Spaces:
Runtime error
Runtime error
| # cython: auto_cpdef=True, infer_types=True, language_level=3, py2_import=True | |
| # | |
| # Parser | |
| # | |
| from __future__ import absolute_import | |
| # This should be done automatically | |
| import cython | |
| cython.declare(Nodes=object, ExprNodes=object, EncodedString=object, | |
| bytes_literal=object, StringEncoding=object, | |
| FileSourceDescriptor=object, lookup_unicodechar=object, unicode_category=object, | |
| Future=object, Options=object, error=object, warning=object, | |
| Builtin=object, ModuleNode=object, Utils=object, _unicode=object, _bytes=object, | |
| re=object, sys=object, _parse_escape_sequences=object, _parse_escape_sequences_raw=object, | |
| partial=object, reduce=object, _IS_PY3=cython.bint, _IS_2BYTE_UNICODE=cython.bint, | |
| _CDEF_MODIFIERS=tuple) | |
| from io import StringIO | |
| import re | |
| import sys | |
| from unicodedata import lookup as lookup_unicodechar, category as unicode_category | |
| from functools import partial, reduce | |
| from .Scanning import PyrexScanner, FileSourceDescriptor, StringSourceDescriptor | |
| from . import Nodes | |
| from . import ExprNodes | |
| from . import Builtin | |
| from . import StringEncoding | |
| from .StringEncoding import EncodedString, bytes_literal, _unicode, _bytes | |
| from .ModuleNode import ModuleNode | |
| from .Errors import error, warning | |
| from .. import Utils | |
| from . import Future | |
| from . import Options | |
| _IS_PY3 = sys.version_info[0] >= 3 | |
| _IS_2BYTE_UNICODE = sys.maxunicode == 0xffff | |
| _CDEF_MODIFIERS = ('inline', 'nogil', 'api') | |
| class Ctx(object): | |
| # Parsing context | |
| level = 'other' | |
| visibility = 'private' | |
| cdef_flag = 0 | |
| typedef_flag = 0 | |
| api = 0 | |
| overridable = 0 | |
| nogil = 0 | |
| namespace = None | |
| templates = None | |
| allow_struct_enum_decorator = False | |
| def __init__(self, **kwds): | |
| self.__dict__.update(kwds) | |
| def __call__(self, **kwds): | |
| ctx = Ctx() | |
| d = ctx.__dict__ | |
| d.update(self.__dict__) | |
| d.update(kwds) | |
| return ctx | |
| def p_ident(s, message="Expected an identifier"): | |
| if s.sy == 'IDENT': | |
| name = s.systring | |
| s.next() | |
| return name | |
| else: | |
| s.error(message) | |
| def p_ident_list(s): | |
| names = [] | |
| while s.sy == 'IDENT': | |
| names.append(s.systring) | |
| s.next() | |
| if s.sy != ',': | |
| break | |
| s.next() | |
| return names | |
| #------------------------------------------ | |
| # | |
| # Expressions | |
| # | |
| #------------------------------------------ | |
| def p_binop_operator(s): | |
| pos = s.position() | |
| op = s.sy | |
| s.next() | |
| return op, pos | |
| def p_binop_expr(s, ops, p_sub_expr): | |
| n1 = p_sub_expr(s) | |
| while s.sy in ops: | |
| op, pos = p_binop_operator(s) | |
| n2 = p_sub_expr(s) | |
| n1 = ExprNodes.binop_node(pos, op, n1, n2) | |
| if op == '/': | |
| if Future.division in s.context.future_directives: | |
| n1.truedivision = True | |
| else: | |
| n1.truedivision = None # unknown | |
| return n1 | |
| #lambdef: 'lambda' [varargslist] ':' test | |
| def p_lambdef(s, allow_conditional=True): | |
| # s.sy == 'lambda' | |
| pos = s.position() | |
| s.next() | |
| if s.sy == ':': | |
| args = [] | |
| star_arg = starstar_arg = None | |
| else: | |
| args, star_arg, starstar_arg = p_varargslist( | |
| s, terminator=':', annotated=False) | |
| s.expect(':') | |
| if allow_conditional: | |
| expr = p_test(s) | |
| else: | |
| expr = p_test_nocond(s) | |
| return ExprNodes.LambdaNode( | |
| pos, args = args, | |
| star_arg = star_arg, starstar_arg = starstar_arg, | |
| result_expr = expr) | |
| #lambdef_nocond: 'lambda' [varargslist] ':' test_nocond | |
| def p_lambdef_nocond(s): | |
| return p_lambdef(s, allow_conditional=False) | |
| #test: or_test ['if' or_test 'else' test] | lambdef | |
| def p_test(s): | |
| if s.sy == 'lambda': | |
| return p_lambdef(s) | |
| pos = s.position() | |
| expr = p_or_test(s) | |
| if s.sy == 'if': | |
| s.next() | |
| test = p_or_test(s) | |
| s.expect('else') | |
| other = p_test(s) | |
| return ExprNodes.CondExprNode(pos, test=test, true_val=expr, false_val=other) | |
| else: | |
| return expr | |
| #test_nocond: or_test | lambdef_nocond | |
| def p_test_nocond(s): | |
| if s.sy == 'lambda': | |
| return p_lambdef_nocond(s) | |
| else: | |
| return p_or_test(s) | |
| #or_test: and_test ('or' and_test)* | |
| def p_or_test(s): | |
| return p_rassoc_binop_expr(s, ('or',), p_and_test) | |
| def p_rassoc_binop_expr(s, ops, p_subexpr): | |
| n1 = p_subexpr(s) | |
| if s.sy in ops: | |
| pos = s.position() | |
| op = s.sy | |
| s.next() | |
| n2 = p_rassoc_binop_expr(s, ops, p_subexpr) | |
| n1 = ExprNodes.binop_node(pos, op, n1, n2) | |
| return n1 | |
| #and_test: not_test ('and' not_test)* | |
| def p_and_test(s): | |
| #return p_binop_expr(s, ('and',), p_not_test) | |
| return p_rassoc_binop_expr(s, ('and',), p_not_test) | |
| #not_test: 'not' not_test | comparison | |
| def p_not_test(s): | |
| if s.sy == 'not': | |
| pos = s.position() | |
| s.next() | |
| return ExprNodes.NotNode(pos, operand = p_not_test(s)) | |
| else: | |
| return p_comparison(s) | |
| #comparison: expr (comp_op expr)* | |
| #comp_op: '<'|'>'|'=='|'>='|'<='|'<>'|'!='|'in'|'not' 'in'|'is'|'is' 'not' | |
| def p_comparison(s): | |
| n1 = p_starred_expr(s) | |
| if s.sy in comparison_ops: | |
| pos = s.position() | |
| op = p_cmp_op(s) | |
| n2 = p_starred_expr(s) | |
| n1 = ExprNodes.PrimaryCmpNode(pos, | |
| operator = op, operand1 = n1, operand2 = n2) | |
| if s.sy in comparison_ops: | |
| n1.cascade = p_cascaded_cmp(s) | |
| return n1 | |
| def p_test_or_starred_expr(s): | |
| if s.sy == '*': | |
| return p_starred_expr(s) | |
| else: | |
| return p_test(s) | |
| def p_starred_expr(s): | |
| pos = s.position() | |
| if s.sy == '*': | |
| starred = True | |
| s.next() | |
| else: | |
| starred = False | |
| expr = p_bit_expr(s) | |
| if starred: | |
| expr = ExprNodes.StarredUnpackingNode(pos, expr) | |
| return expr | |
| def p_cascaded_cmp(s): | |
| pos = s.position() | |
| op = p_cmp_op(s) | |
| n2 = p_starred_expr(s) | |
| result = ExprNodes.CascadedCmpNode(pos, | |
| operator = op, operand2 = n2) | |
| if s.sy in comparison_ops: | |
| result.cascade = p_cascaded_cmp(s) | |
| return result | |
| def p_cmp_op(s): | |
| if s.sy == 'not': | |
| s.next() | |
| s.expect('in') | |
| op = 'not_in' | |
| elif s.sy == 'is': | |
| s.next() | |
| if s.sy == 'not': | |
| s.next() | |
| op = 'is_not' | |
| else: | |
| op = 'is' | |
| else: | |
| op = s.sy | |
| s.next() | |
| if op == '<>': | |
| op = '!=' | |
| return op | |
| comparison_ops = cython.declare(set, set([ | |
| '<', '>', '==', '>=', '<=', '<>', '!=', | |
| 'in', 'is', 'not' | |
| ])) | |
| #expr: xor_expr ('|' xor_expr)* | |
| def p_bit_expr(s): | |
| return p_binop_expr(s, ('|',), p_xor_expr) | |
| #xor_expr: and_expr ('^' and_expr)* | |
| def p_xor_expr(s): | |
| return p_binop_expr(s, ('^',), p_and_expr) | |
| #and_expr: shift_expr ('&' shift_expr)* | |
| def p_and_expr(s): | |
| return p_binop_expr(s, ('&',), p_shift_expr) | |
| #shift_expr: arith_expr (('<<'|'>>') arith_expr)* | |
| def p_shift_expr(s): | |
| return p_binop_expr(s, ('<<', '>>'), p_arith_expr) | |
| #arith_expr: term (('+'|'-') term)* | |
| def p_arith_expr(s): | |
| return p_binop_expr(s, ('+', '-'), p_term) | |
| #term: factor (('*'|'@'|'/'|'%'|'//') factor)* | |
| def p_term(s): | |
| return p_binop_expr(s, ('*', '@', '/', '%', '//'), p_factor) | |
| #factor: ('+'|'-'|'~'|'&'|typecast|sizeof) factor | power | |
| def p_factor(s): | |
| # little indirection for C-ification purposes | |
| return _p_factor(s) | |
| def _p_factor(s): | |
| sy = s.sy | |
| if sy in ('+', '-', '~'): | |
| op = s.sy | |
| pos = s.position() | |
| s.next() | |
| return ExprNodes.unop_node(pos, op, p_factor(s)) | |
| elif not s.in_python_file: | |
| if sy == '&': | |
| pos = s.position() | |
| s.next() | |
| arg = p_factor(s) | |
| return ExprNodes.AmpersandNode(pos, operand = arg) | |
| elif sy == "<": | |
| return p_typecast(s) | |
| elif sy == 'IDENT' and s.systring == "sizeof": | |
| return p_sizeof(s) | |
| return p_power(s) | |
| def p_typecast(s): | |
| # s.sy == "<" | |
| pos = s.position() | |
| s.next() | |
| base_type = p_c_base_type(s) | |
| is_memslice = isinstance(base_type, Nodes.MemoryViewSliceTypeNode) | |
| is_template = isinstance(base_type, Nodes.TemplatedTypeNode) | |
| is_const = isinstance(base_type, Nodes.CConstTypeNode) | |
| if (not is_memslice and not is_template and not is_const | |
| and base_type.name is None): | |
| s.error("Unknown type") | |
| declarator = p_c_declarator(s, empty = 1) | |
| if s.sy == '?': | |
| s.next() | |
| typecheck = 1 | |
| else: | |
| typecheck = 0 | |
| s.expect(">") | |
| operand = p_factor(s) | |
| if is_memslice: | |
| return ExprNodes.CythonArrayNode(pos, base_type_node=base_type, | |
| operand=operand) | |
| return ExprNodes.TypecastNode(pos, | |
| base_type = base_type, | |
| declarator = declarator, | |
| operand = operand, | |
| typecheck = typecheck) | |
| def p_sizeof(s): | |
| # s.sy == ident "sizeof" | |
| pos = s.position() | |
| s.next() | |
| s.expect('(') | |
| # Here we decide if we are looking at an expression or type | |
| # If it is actually a type, but parsable as an expression, | |
| # we treat it as an expression here. | |
| if looking_at_expr(s): | |
| operand = p_test(s) | |
| node = ExprNodes.SizeofVarNode(pos, operand = operand) | |
| else: | |
| base_type = p_c_base_type(s) | |
| declarator = p_c_declarator(s, empty = 1) | |
| node = ExprNodes.SizeofTypeNode(pos, | |
| base_type = base_type, declarator = declarator) | |
| s.expect(')') | |
| return node | |
| def p_yield_expression(s): | |
| # s.sy == "yield" | |
| pos = s.position() | |
| s.next() | |
| is_yield_from = False | |
| if s.sy == 'from': | |
| is_yield_from = True | |
| s.next() | |
| if s.sy != ')' and s.sy not in statement_terminators: | |
| # "yield from" does not support implicit tuples, but "yield" does ("yield 1,2") | |
| arg = p_test(s) if is_yield_from else p_testlist(s) | |
| else: | |
| if is_yield_from: | |
| s.error("'yield from' requires a source argument", | |
| pos=pos, fatal=False) | |
| arg = None | |
| if is_yield_from: | |
| return ExprNodes.YieldFromExprNode(pos, arg=arg) | |
| else: | |
| return ExprNodes.YieldExprNode(pos, arg=arg) | |
| def p_yield_statement(s): | |
| # s.sy == "yield" | |
| yield_expr = p_yield_expression(s) | |
| return Nodes.ExprStatNode(yield_expr.pos, expr=yield_expr) | |
| def p_async_statement(s, ctx, decorators): | |
| # s.sy >> 'async' ... | |
| if s.sy == 'def': | |
| # 'async def' statements aren't allowed in pxd files | |
| if 'pxd' in ctx.level: | |
| s.error('def statement not allowed here') | |
| s.level = ctx.level | |
| return p_def_statement(s, decorators, is_async_def=True) | |
| elif decorators: | |
| s.error("Decorators can only be followed by functions or classes") | |
| elif s.sy == 'for': | |
| return p_for_statement(s, is_async=True) | |
| elif s.sy == 'with': | |
| s.next() | |
| return p_with_items(s, is_async=True) | |
| else: | |
| s.error("expected one of 'def', 'for', 'with' after 'async'") | |
| #power: atom_expr ('**' factor)* | |
| #atom_expr: ['await'] atom trailer* | |
| def p_power(s): | |
| if s.systring == 'new' and s.peek()[0] == 'IDENT': | |
| return p_new_expr(s) | |
| await_pos = None | |
| if s.sy == 'await': | |
| await_pos = s.position() | |
| s.next() | |
| n1 = p_atom(s) | |
| while s.sy in ('(', '[', '.'): | |
| n1 = p_trailer(s, n1) | |
| if await_pos: | |
| n1 = ExprNodes.AwaitExprNode(await_pos, arg=n1) | |
| if s.sy == '**': | |
| pos = s.position() | |
| s.next() | |
| n2 = p_factor(s) | |
| n1 = ExprNodes.binop_node(pos, '**', n1, n2) | |
| return n1 | |
| def p_new_expr(s): | |
| # s.systring == 'new'. | |
| pos = s.position() | |
| s.next() | |
| cppclass = p_c_base_type(s) | |
| return p_call(s, ExprNodes.NewExprNode(pos, cppclass = cppclass)) | |
| #trailer: '(' [arglist] ')' | '[' subscriptlist ']' | '.' NAME | |
| def p_trailer(s, node1): | |
| pos = s.position() | |
| if s.sy == '(': | |
| return p_call(s, node1) | |
| elif s.sy == '[': | |
| return p_index(s, node1) | |
| else: # s.sy == '.' | |
| s.next() | |
| name = p_ident(s) | |
| return ExprNodes.AttributeNode(pos, | |
| obj=node1, attribute=name) | |
| # arglist: argument (',' argument)* [','] | |
| # argument: [test '='] test # Really [keyword '='] test | |
| # since PEP 448: | |
| # argument: ( test [comp_for] | | |
| # test '=' test | | |
| # '**' expr | | |
| # star_expr ) | |
| def p_call_parse_args(s, allow_genexp=True): | |
| # s.sy == '(' | |
| pos = s.position() | |
| s.next() | |
| positional_args = [] | |
| keyword_args = [] | |
| starstar_seen = False | |
| last_was_tuple_unpack = False | |
| while s.sy != ')': | |
| if s.sy == '*': | |
| if starstar_seen: | |
| s.error("Non-keyword arg following keyword arg", pos=s.position()) | |
| s.next() | |
| positional_args.append(p_test(s)) | |
| last_was_tuple_unpack = True | |
| elif s.sy == '**': | |
| s.next() | |
| keyword_args.append(p_test(s)) | |
| starstar_seen = True | |
| else: | |
| arg = p_test(s) | |
| if s.sy == '=': | |
| s.next() | |
| if not arg.is_name: | |
| s.error("Expected an identifier before '='", | |
| pos=arg.pos) | |
| encoded_name = s.context.intern_ustring(arg.name) | |
| keyword = ExprNodes.IdentifierStringNode( | |
| arg.pos, value=encoded_name) | |
| arg = p_test(s) | |
| keyword_args.append((keyword, arg)) | |
| else: | |
| if keyword_args: | |
| s.error("Non-keyword arg following keyword arg", pos=arg.pos) | |
| if positional_args and not last_was_tuple_unpack: | |
| positional_args[-1].append(arg) | |
| else: | |
| positional_args.append([arg]) | |
| last_was_tuple_unpack = False | |
| if s.sy != ',': | |
| break | |
| s.next() | |
| if s.sy in ('for', 'async'): | |
| if not keyword_args and not last_was_tuple_unpack: | |
| if len(positional_args) == 1 and len(positional_args[0]) == 1: | |
| positional_args = [[p_genexp(s, positional_args[0][0])]] | |
| s.expect(')') | |
| return positional_args or [[]], keyword_args | |
| def p_call_build_packed_args(pos, positional_args, keyword_args): | |
| keyword_dict = None | |
| subtuples = [ | |
| ExprNodes.TupleNode(pos, args=arg) if isinstance(arg, list) else ExprNodes.AsTupleNode(pos, arg=arg) | |
| for arg in positional_args | |
| ] | |
| # TODO: implement a faster way to join tuples than creating each one and adding them | |
| arg_tuple = reduce(partial(ExprNodes.binop_node, pos, '+'), subtuples) | |
| if keyword_args: | |
| kwargs = [] | |
| dict_items = [] | |
| for item in keyword_args: | |
| if isinstance(item, tuple): | |
| key, value = item | |
| dict_items.append(ExprNodes.DictItemNode(pos=key.pos, key=key, value=value)) | |
| elif item.is_dict_literal: | |
| # unpack "**{a:b}" directly | |
| dict_items.extend(item.key_value_pairs) | |
| else: | |
| if dict_items: | |
| kwargs.append(ExprNodes.DictNode( | |
| dict_items[0].pos, key_value_pairs=dict_items, reject_duplicates=True)) | |
| dict_items = [] | |
| kwargs.append(item) | |
| if dict_items: | |
| kwargs.append(ExprNodes.DictNode( | |
| dict_items[0].pos, key_value_pairs=dict_items, reject_duplicates=True)) | |
| if kwargs: | |
| if len(kwargs) == 1 and kwargs[0].is_dict_literal: | |
| # only simple keyword arguments found -> one dict | |
| keyword_dict = kwargs[0] | |
| else: | |
| # at least one **kwargs | |
| keyword_dict = ExprNodes.MergedDictNode(pos, keyword_args=kwargs) | |
| return arg_tuple, keyword_dict | |
| def p_call(s, function): | |
| # s.sy == '(' | |
| pos = s.position() | |
| positional_args, keyword_args = p_call_parse_args(s) | |
| if not keyword_args and len(positional_args) == 1 and isinstance(positional_args[0], list): | |
| return ExprNodes.SimpleCallNode(pos, function=function, args=positional_args[0]) | |
| else: | |
| arg_tuple, keyword_dict = p_call_build_packed_args(pos, positional_args, keyword_args) | |
| return ExprNodes.GeneralCallNode( | |
| pos, function=function, positional_args=arg_tuple, keyword_args=keyword_dict) | |
| #lambdef: 'lambda' [varargslist] ':' test | |
| #subscriptlist: subscript (',' subscript)* [','] | |
| def p_index(s, base): | |
| # s.sy == '[' | |
| pos = s.position() | |
| s.next() | |
| subscripts, is_single_value = p_subscript_list(s) | |
| if is_single_value and len(subscripts[0]) == 2: | |
| start, stop = subscripts[0] | |
| result = ExprNodes.SliceIndexNode(pos, | |
| base = base, start = start, stop = stop) | |
| else: | |
| indexes = make_slice_nodes(pos, subscripts) | |
| if is_single_value: | |
| index = indexes[0] | |
| else: | |
| index = ExprNodes.TupleNode(pos, args = indexes) | |
| result = ExprNodes.IndexNode(pos, | |
| base = base, index = index) | |
| s.expect(']') | |
| return result | |
| def p_subscript_list(s): | |
| is_single_value = True | |
| items = [p_subscript(s)] | |
| while s.sy == ',': | |
| is_single_value = False | |
| s.next() | |
| if s.sy == ']': | |
| break | |
| items.append(p_subscript(s)) | |
| return items, is_single_value | |
| #subscript: '.' '.' '.' | test | [test] ':' [test] [':' [test]] | |
| def p_subscript(s): | |
| # Parse a subscript and return a list of | |
| # 1, 2 or 3 ExprNodes, depending on how | |
| # many slice elements were encountered. | |
| pos = s.position() | |
| start = p_slice_element(s, (':',)) | |
| if s.sy != ':': | |
| return [start] | |
| s.next() | |
| stop = p_slice_element(s, (':', ',', ']')) | |
| if s.sy != ':': | |
| return [start, stop] | |
| s.next() | |
| step = p_slice_element(s, (':', ',', ']')) | |
| return [start, stop, step] | |
| def p_slice_element(s, follow_set): | |
| # Simple expression which may be missing iff | |
| # it is followed by something in follow_set. | |
| if s.sy not in follow_set: | |
| return p_test(s) | |
| else: | |
| return None | |
| def expect_ellipsis(s): | |
| s.expect('.') | |
| s.expect('.') | |
| s.expect('.') | |
| def make_slice_nodes(pos, subscripts): | |
| # Convert a list of subscripts as returned | |
| # by p_subscript_list into a list of ExprNodes, | |
| # creating SliceNodes for elements with 2 or | |
| # more components. | |
| result = [] | |
| for subscript in subscripts: | |
| if len(subscript) == 1: | |
| result.append(subscript[0]) | |
| else: | |
| result.append(make_slice_node(pos, *subscript)) | |
| return result | |
| def make_slice_node(pos, start, stop = None, step = None): | |
| if not start: | |
| start = ExprNodes.NoneNode(pos) | |
| if not stop: | |
| stop = ExprNodes.NoneNode(pos) | |
| if not step: | |
| step = ExprNodes.NoneNode(pos) | |
| return ExprNodes.SliceNode(pos, | |
| start = start, stop = stop, step = step) | |
| #atom: '(' [yield_expr|testlist_comp] ')' | '[' [listmaker] ']' | '{' [dict_or_set_maker] '}' | '`' testlist '`' | NAME | NUMBER | STRING+ | |
| def p_atom(s): | |
| pos = s.position() | |
| sy = s.sy | |
| if sy == '(': | |
| s.next() | |
| if s.sy == ')': | |
| result = ExprNodes.TupleNode(pos, args = []) | |
| elif s.sy == 'yield': | |
| result = p_yield_expression(s) | |
| else: | |
| result = p_testlist_comp(s) | |
| s.expect(')') | |
| return result | |
| elif sy == '[': | |
| return p_list_maker(s) | |
| elif sy == '{': | |
| return p_dict_or_set_maker(s) | |
| elif sy == '`': | |
| return p_backquote_expr(s) | |
| elif sy == '.': | |
| expect_ellipsis(s) | |
| return ExprNodes.EllipsisNode(pos) | |
| elif sy == 'INT': | |
| return p_int_literal(s) | |
| elif sy == 'FLOAT': | |
| value = s.systring | |
| s.next() | |
| return ExprNodes.FloatNode(pos, value = value) | |
| elif sy == 'IMAG': | |
| value = s.systring[:-1] | |
| s.next() | |
| return ExprNodes.ImagNode(pos, value = value) | |
| elif sy == 'BEGIN_STRING': | |
| kind, bytes_value, unicode_value = p_cat_string_literal(s) | |
| if kind == 'c': | |
| return ExprNodes.CharNode(pos, value = bytes_value) | |
| elif kind == 'u': | |
| return ExprNodes.UnicodeNode(pos, value = unicode_value, bytes_value = bytes_value) | |
| elif kind == 'b': | |
| return ExprNodes.BytesNode(pos, value = bytes_value) | |
| elif kind == 'f': | |
| return ExprNodes.JoinedStrNode(pos, values = unicode_value) | |
| elif kind == '': | |
| return ExprNodes.StringNode(pos, value = bytes_value, unicode_value = unicode_value) | |
| else: | |
| s.error("invalid string kind '%s'" % kind) | |
| elif sy == 'IDENT': | |
| name = s.systring | |
| if name == "None": | |
| result = ExprNodes.NoneNode(pos) | |
| elif name == "True": | |
| result = ExprNodes.BoolNode(pos, value=True) | |
| elif name == "False": | |
| result = ExprNodes.BoolNode(pos, value=False) | |
| elif name == "NULL" and not s.in_python_file: | |
| result = ExprNodes.NullNode(pos) | |
| else: | |
| result = p_name(s, name) | |
| s.next() | |
| return result | |
| else: | |
| s.error("Expected an identifier or literal") | |
| def p_int_literal(s): | |
| pos = s.position() | |
| value = s.systring | |
| s.next() | |
| unsigned = "" | |
| longness = "" | |
| while value[-1] in u"UuLl": | |
| if value[-1] in u"Ll": | |
| longness += "L" | |
| else: | |
| unsigned += "U" | |
| value = value[:-1] | |
| # '3L' is ambiguous in Py2 but not in Py3. '3U' and '3LL' are | |
| # illegal in Py2 Python files. All suffixes are illegal in Py3 | |
| # Python files. | |
| is_c_literal = None | |
| if unsigned: | |
| is_c_literal = True | |
| elif longness: | |
| if longness == 'LL' or s.context.language_level >= 3: | |
| is_c_literal = True | |
| if s.in_python_file: | |
| if is_c_literal: | |
| error(pos, "illegal integer literal syntax in Python source file") | |
| is_c_literal = False | |
| return ExprNodes.IntNode(pos, | |
| is_c_literal = is_c_literal, | |
| value = value, | |
| unsigned = unsigned, | |
| longness = longness) | |
| def p_name(s, name): | |
| pos = s.position() | |
| if not s.compile_time_expr and name in s.compile_time_env: | |
| value = s.compile_time_env.lookup_here(name) | |
| node = wrap_compile_time_constant(pos, value) | |
| if node is not None: | |
| return node | |
| return ExprNodes.NameNode(pos, name=name) | |
| def wrap_compile_time_constant(pos, value): | |
| rep = repr(value) | |
| if value is None: | |
| return ExprNodes.NoneNode(pos) | |
| elif value is Ellipsis: | |
| return ExprNodes.EllipsisNode(pos) | |
| elif isinstance(value, bool): | |
| return ExprNodes.BoolNode(pos, value=value) | |
| elif isinstance(value, int): | |
| return ExprNodes.IntNode(pos, value=rep, constant_result=value) | |
| elif isinstance(value, float): | |
| return ExprNodes.FloatNode(pos, value=rep, constant_result=value) | |
| elif isinstance(value, complex): | |
| node = ExprNodes.ImagNode(pos, value=repr(value.imag), constant_result=complex(0.0, value.imag)) | |
| if value.real: | |
| # FIXME: should we care about -0.0 ? | |
| # probably not worth using the '-' operator for negative imag values | |
| node = ExprNodes.binop_node( | |
| pos, '+', ExprNodes.FloatNode(pos, value=repr(value.real), constant_result=value.real), node, | |
| constant_result=value) | |
| return node | |
| elif isinstance(value, _unicode): | |
| return ExprNodes.UnicodeNode(pos, value=EncodedString(value)) | |
| elif isinstance(value, _bytes): | |
| bvalue = bytes_literal(value, 'ascii') # actually: unknown encoding, but BytesLiteral requires one | |
| return ExprNodes.BytesNode(pos, value=bvalue, constant_result=value) | |
| elif isinstance(value, tuple): | |
| args = [wrap_compile_time_constant(pos, arg) | |
| for arg in value] | |
| if None not in args: | |
| return ExprNodes.TupleNode(pos, args=args) | |
| else: | |
| # error already reported | |
| return None | |
| elif not _IS_PY3 and isinstance(value, long): | |
| return ExprNodes.IntNode(pos, value=rep.rstrip('L'), constant_result=value) | |
| error(pos, "Invalid type for compile-time constant: %r (type %s)" | |
| % (value, value.__class__.__name__)) | |
| return None | |
| def p_cat_string_literal(s): | |
| # A sequence of one or more adjacent string literals. | |
| # Returns (kind, bytes_value, unicode_value) | |
| # where kind in ('b', 'c', 'u', 'f', '') | |
| pos = s.position() | |
| kind, bytes_value, unicode_value = p_string_literal(s) | |
| if kind == 'c' or s.sy != 'BEGIN_STRING': | |
| return kind, bytes_value, unicode_value | |
| bstrings, ustrings, positions = [bytes_value], [unicode_value], [pos] | |
| bytes_value = unicode_value = None | |
| while s.sy == 'BEGIN_STRING': | |
| pos = s.position() | |
| next_kind, next_bytes_value, next_unicode_value = p_string_literal(s) | |
| if next_kind == 'c': | |
| error(pos, "Cannot concatenate char literal with another string or char literal") | |
| continue | |
| elif next_kind != kind: | |
| # concatenating f strings and normal strings is allowed and leads to an f string | |
| if set([kind, next_kind]) in (set(['f', 'u']), set(['f', ''])): | |
| kind = 'f' | |
| else: | |
| error(pos, "Cannot mix string literals of different types, expected %s'', got %s''" % ( | |
| kind, next_kind)) | |
| continue | |
| bstrings.append(next_bytes_value) | |
| ustrings.append(next_unicode_value) | |
| positions.append(pos) | |
| # join and rewrap the partial literals | |
| if kind in ('b', 'c', '') or kind == 'u' and None not in bstrings: | |
| # Py3 enforced unicode literals are parsed as bytes/unicode combination | |
| bytes_value = bytes_literal(StringEncoding.join_bytes(bstrings), s.source_encoding) | |
| if kind in ('u', ''): | |
| unicode_value = EncodedString(u''.join([u for u in ustrings if u is not None])) | |
| if kind == 'f': | |
| unicode_value = [] | |
| for u, pos in zip(ustrings, positions): | |
| if isinstance(u, list): | |
| unicode_value += u | |
| else: | |
| # non-f-string concatenated into the f-string | |
| unicode_value.append(ExprNodes.UnicodeNode(pos, value=EncodedString(u))) | |
| return kind, bytes_value, unicode_value | |
| def p_opt_string_literal(s, required_type='u'): | |
| if s.sy != 'BEGIN_STRING': | |
| return None | |
| pos = s.position() | |
| kind, bytes_value, unicode_value = p_string_literal(s, required_type) | |
| if required_type == 'u': | |
| if kind == 'f': | |
| s.error("f-string not allowed here", pos) | |
| return unicode_value | |
| elif required_type == 'b': | |
| return bytes_value | |
| else: | |
| s.error("internal parser configuration error") | |
| def check_for_non_ascii_characters(string): | |
| for c in string: | |
| if c >= u'\x80': | |
| return True | |
| return False | |
| def p_string_literal(s, kind_override=None): | |
| # A single string or char literal. Returns (kind, bvalue, uvalue) | |
| # where kind in ('b', 'c', 'u', 'f', ''). The 'bvalue' is the source | |
| # code byte sequence of the string literal, 'uvalue' is the | |
| # decoded Unicode string. Either of the two may be None depending | |
| # on the 'kind' of string, only unprefixed strings have both | |
| # representations. In f-strings, the uvalue is a list of the Unicode | |
| # strings and f-string expressions that make up the f-string. | |
| # s.sy == 'BEGIN_STRING' | |
| pos = s.position() | |
| is_python3_source = s.context.language_level >= 3 | |
| has_non_ascii_literal_characters = False | |
| string_start_pos = (pos[0], pos[1], pos[2] + len(s.systring)) | |
| kind_string = s.systring.rstrip('"\'').lower() | |
| if len(kind_string) > 1: | |
| if len(set(kind_string)) != len(kind_string): | |
| error(pos, 'Duplicate string prefix character') | |
| if 'b' in kind_string and 'u' in kind_string: | |
| error(pos, 'String prefixes b and u cannot be combined') | |
| if 'b' in kind_string and 'f' in kind_string: | |
| error(pos, 'String prefixes b and f cannot be combined') | |
| if 'u' in kind_string and 'f' in kind_string: | |
| error(pos, 'String prefixes u and f cannot be combined') | |
| is_raw = 'r' in kind_string | |
| if 'c' in kind_string: | |
| # this should never happen, since the lexer does not allow combining c | |
| # with other prefix characters | |
| if len(kind_string) != 1: | |
| error(pos, 'Invalid string prefix for character literal') | |
| kind = 'c' | |
| elif 'f' in kind_string: | |
| kind = 'f' # u is ignored | |
| is_raw = True # postpone the escape resolution | |
| elif 'b' in kind_string: | |
| kind = 'b' | |
| elif 'u' in kind_string: | |
| kind = 'u' | |
| else: | |
| kind = '' | |
| if kind == '' and kind_override is None and Future.unicode_literals in s.context.future_directives: | |
| chars = StringEncoding.StrLiteralBuilder(s.source_encoding) | |
| kind = 'u' | |
| else: | |
| if kind_override is not None and kind_override in 'ub': | |
| kind = kind_override | |
| if kind in ('u', 'f'): # f-strings are scanned exactly like Unicode literals, but are parsed further later | |
| chars = StringEncoding.UnicodeLiteralBuilder() | |
| elif kind == '': | |
| chars = StringEncoding.StrLiteralBuilder(s.source_encoding) | |
| else: | |
| chars = StringEncoding.BytesLiteralBuilder(s.source_encoding) | |
| while 1: | |
| s.next() | |
| sy = s.sy | |
| systr = s.systring | |
| # print "p_string_literal: sy =", sy, repr(s.systring) ### | |
| if sy == 'CHARS': | |
| chars.append(systr) | |
| if is_python3_source and not has_non_ascii_literal_characters and check_for_non_ascii_characters(systr): | |
| has_non_ascii_literal_characters = True | |
| elif sy == 'ESCAPE': | |
| # in Py2, 'ur' raw unicode strings resolve unicode escapes but nothing else | |
| if is_raw and (is_python3_source or kind != 'u' or systr[1] not in u'Uu'): | |
| chars.append(systr) | |
| if is_python3_source and not has_non_ascii_literal_characters and check_for_non_ascii_characters(systr): | |
| has_non_ascii_literal_characters = True | |
| else: | |
| _append_escape_sequence(kind, chars, systr, s) | |
| elif sy == 'NEWLINE': | |
| chars.append(u'\n') | |
| elif sy == 'END_STRING': | |
| break | |
| elif sy == 'EOF': | |
| s.error("Unclosed string literal", pos=pos) | |
| else: | |
| s.error("Unexpected token %r:%r in string literal" % ( | |
| sy, s.systring)) | |
| if kind == 'c': | |
| unicode_value = None | |
| bytes_value = chars.getchar() | |
| if len(bytes_value) != 1: | |
| error(pos, u"invalid character literal: %r" % bytes_value) | |
| else: | |
| bytes_value, unicode_value = chars.getstrings() | |
| if (has_non_ascii_literal_characters | |
| and is_python3_source and Future.unicode_literals in s.context.future_directives): | |
| # Python 3 forbids literal non-ASCII characters in byte strings | |
| if kind == 'b': | |
| s.error("bytes can only contain ASCII literal characters.", pos=pos) | |
| bytes_value = None | |
| if kind == 'f': | |
| unicode_value = p_f_string(s, unicode_value, string_start_pos, is_raw='r' in kind_string) | |
| s.next() | |
| return (kind, bytes_value, unicode_value) | |
| def _append_escape_sequence(kind, builder, escape_sequence, s): | |
| c = escape_sequence[1] | |
| if c in u"01234567": | |
| builder.append_charval(int(escape_sequence[1:], 8)) | |
| elif c in u"'\"\\": | |
| builder.append(c) | |
| elif c in u"abfnrtv": | |
| builder.append(StringEncoding.char_from_escape_sequence(escape_sequence)) | |
| elif c == u'\n': | |
| pass # line continuation | |
| elif c == u'x': # \xXX | |
| if len(escape_sequence) == 4: | |
| builder.append_charval(int(escape_sequence[2:], 16)) | |
| else: | |
| s.error("Invalid hex escape '%s'" % escape_sequence, fatal=False) | |
| elif c in u'NUu' and kind in ('u', 'f', ''): # \uxxxx, \Uxxxxxxxx, \N{...} | |
| chrval = -1 | |
| if c == u'N': | |
| uchar = None | |
| try: | |
| uchar = lookup_unicodechar(escape_sequence[3:-1]) | |
| chrval = ord(uchar) | |
| except KeyError: | |
| s.error("Unknown Unicode character name %s" % | |
| repr(escape_sequence[3:-1]).lstrip('u'), fatal=False) | |
| except TypeError: | |
| # 2-byte unicode build of CPython? | |
| if (uchar is not None and _IS_2BYTE_UNICODE and len(uchar) == 2 and | |
| unicode_category(uchar[0]) == 'Cs' and unicode_category(uchar[1]) == 'Cs'): | |
| # surrogate pair instead of single character | |
| chrval = 0x10000 + (ord(uchar[0]) - 0xd800) >> 10 + (ord(uchar[1]) - 0xdc00) | |
| else: | |
| raise | |
| elif len(escape_sequence) in (6, 10): | |
| chrval = int(escape_sequence[2:], 16) | |
| if chrval > 1114111: # sys.maxunicode: | |
| s.error("Invalid unicode escape '%s'" % escape_sequence) | |
| chrval = -1 | |
| else: | |
| s.error("Invalid unicode escape '%s'" % escape_sequence, fatal=False) | |
| if chrval >= 0: | |
| builder.append_uescape(chrval, escape_sequence) | |
| else: | |
| builder.append(escape_sequence) | |
| _parse_escape_sequences_raw, _parse_escape_sequences = [re.compile(( | |
| # escape sequences: | |
| br'(\\(?:' + | |
| (br'\\?' if is_raw else ( | |
| br'[\\abfnrtv"\'{]|' | |
| br'[0-7]{2,3}|' | |
| br'N\{[^}]*\}|' | |
| br'x[0-9a-fA-F]{2}|' | |
| br'u[0-9a-fA-F]{4}|' | |
| br'U[0-9a-fA-F]{8}|' | |
| br'[NxuU]|' # detect invalid escape sequences that do not match above | |
| )) + | |
| br')?|' | |
| # non-escape sequences: | |
| br'\{\{?|' | |
| br'\}\}?|' | |
| br'[^\\{}]+)' | |
| ).decode('us-ascii')).match | |
| for is_raw in (True, False)] | |
| def _f_string_error_pos(pos, string, i): | |
| return (pos[0], pos[1], pos[2] + i + 1) # FIXME: handle newlines in string | |
| def p_f_string(s, unicode_value, pos, is_raw): | |
| # Parses a PEP 498 f-string literal into a list of nodes. Nodes are either UnicodeNodes | |
| # or FormattedValueNodes. | |
| values = [] | |
| next_start = 0 | |
| size = len(unicode_value) | |
| builder = StringEncoding.UnicodeLiteralBuilder() | |
| _parse_seq = _parse_escape_sequences_raw if is_raw else _parse_escape_sequences | |
| while next_start < size: | |
| end = next_start | |
| match = _parse_seq(unicode_value, next_start) | |
| if match is None: | |
| error(_f_string_error_pos(pos, unicode_value, next_start), "Invalid escape sequence") | |
| next_start = match.end() | |
| part = match.group() | |
| c = part[0] | |
| if c == '\\': | |
| if not is_raw and len(part) > 1: | |
| _append_escape_sequence('f', builder, part, s) | |
| else: | |
| builder.append(part) | |
| elif c == '{': | |
| if part == '{{': | |
| builder.append('{') | |
| else: | |
| # start of an expression | |
| if builder.chars: | |
| values.append(ExprNodes.UnicodeNode(pos, value=builder.getstring())) | |
| builder = StringEncoding.UnicodeLiteralBuilder() | |
| next_start, expr_node = p_f_string_expr(s, unicode_value, pos, next_start, is_raw) | |
| values.append(expr_node) | |
| elif c == '}': | |
| if part == '}}': | |
| builder.append('}') | |
| else: | |
| error(_f_string_error_pos(pos, unicode_value, end), | |
| "f-string: single '}' is not allowed") | |
| else: | |
| builder.append(part) | |
| if builder.chars: | |
| values.append(ExprNodes.UnicodeNode(pos, value=builder.getstring())) | |
| return values | |
| def p_f_string_expr(s, unicode_value, pos, starting_index, is_raw): | |
| # Parses a {}-delimited expression inside an f-string. Returns a FormattedValueNode | |
| # and the index in the string that follows the expression. | |
| i = starting_index | |
| size = len(unicode_value) | |
| conversion_char = terminal_char = format_spec = None | |
| format_spec_str = None | |
| NO_CHAR = 2**30 | |
| nested_depth = 0 | |
| quote_char = NO_CHAR | |
| in_triple_quotes = False | |
| backslash_reported = False | |
| while True: | |
| if i >= size: | |
| break # error will be reported below | |
| c = unicode_value[i] | |
| if quote_char != NO_CHAR: | |
| if c == '\\': | |
| # avoid redundant error reports along '\' sequences | |
| if not backslash_reported: | |
| error(_f_string_error_pos(pos, unicode_value, i), | |
| "backslashes not allowed in f-strings") | |
| backslash_reported = True | |
| elif c == quote_char: | |
| if in_triple_quotes: | |
| if i + 2 < size and unicode_value[i + 1] == c and unicode_value[i + 2] == c: | |
| in_triple_quotes = False | |
| quote_char = NO_CHAR | |
| i += 2 | |
| else: | |
| quote_char = NO_CHAR | |
| elif c in '\'"': | |
| quote_char = c | |
| if i + 2 < size and unicode_value[i + 1] == c and unicode_value[i + 2] == c: | |
| in_triple_quotes = True | |
| i += 2 | |
| elif c in '{[(': | |
| nested_depth += 1 | |
| elif nested_depth != 0 and c in '}])': | |
| nested_depth -= 1 | |
| elif c == '#': | |
| error(_f_string_error_pos(pos, unicode_value, i), | |
| "format string cannot include #") | |
| elif nested_depth == 0 and c in '!:}': | |
| # allow != as a special case | |
| if c == '!' and i + 1 < size and unicode_value[i + 1] == '=': | |
| i += 1 | |
| continue | |
| terminal_char = c | |
| break | |
| i += 1 | |
| # normalise line endings as the parser expects that | |
| expr_str = unicode_value[starting_index:i].replace('\r\n', '\n').replace('\r', '\n') | |
| expr_pos = (pos[0], pos[1], pos[2] + starting_index + 2) # TODO: find exact code position (concat, multi-line, ...) | |
| if not expr_str.strip(): | |
| error(_f_string_error_pos(pos, unicode_value, starting_index), | |
| "empty expression not allowed in f-string") | |
| if terminal_char == '!': | |
| i += 1 | |
| if i + 2 > size: | |
| pass # error will be reported below | |
| else: | |
| conversion_char = unicode_value[i] | |
| i += 1 | |
| terminal_char = unicode_value[i] | |
| if terminal_char == ':': | |
| in_triple_quotes = False | |
| in_string = False | |
| nested_depth = 0 | |
| start_format_spec = i + 1 | |
| while True: | |
| if i >= size: | |
| break # error will be reported below | |
| c = unicode_value[i] | |
| if not in_triple_quotes and not in_string: | |
| if c == '{': | |
| nested_depth += 1 | |
| elif c == '}': | |
| if nested_depth > 0: | |
| nested_depth -= 1 | |
| else: | |
| terminal_char = c | |
| break | |
| if c in '\'"': | |
| if not in_string and i + 2 < size and unicode_value[i + 1] == c and unicode_value[i + 2] == c: | |
| in_triple_quotes = not in_triple_quotes | |
| i += 2 | |
| elif not in_triple_quotes: | |
| in_string = not in_string | |
| i += 1 | |
| format_spec_str = unicode_value[start_format_spec:i] | |
| if terminal_char != '}': | |
| error(_f_string_error_pos(pos, unicode_value, i), | |
| "missing '}' in format string expression" + ( | |
| ", found '%s'" % terminal_char if terminal_char else "")) | |
| # parse the expression as if it was surrounded by parentheses | |
| buf = StringIO('(%s)' % expr_str) | |
| scanner = PyrexScanner(buf, expr_pos[0], parent_scanner=s, source_encoding=s.source_encoding, initial_pos=expr_pos) | |
| expr = p_testlist(scanner) # TODO is testlist right here? | |
| # validate the conversion char | |
| if conversion_char is not None and not ExprNodes.FormattedValueNode.find_conversion_func(conversion_char): | |
| error(expr_pos, "invalid conversion character '%s'" % conversion_char) | |
| # the format spec is itself treated like an f-string | |
| if format_spec_str: | |
| format_spec = ExprNodes.JoinedStrNode(pos, values=p_f_string(s, format_spec_str, pos, is_raw)) | |
| return i + 1, ExprNodes.FormattedValueNode( | |
| pos, value=expr, conversion_char=conversion_char, format_spec=format_spec) | |
| # since PEP 448: | |
| # list_display ::= "[" [listmaker] "]" | |
| # listmaker ::= (test|star_expr) ( comp_for | (',' (test|star_expr))* [','] ) | |
| # comp_iter ::= comp_for | comp_if | |
| # comp_for ::= ["async"] "for" expression_list "in" testlist [comp_iter] | |
| # comp_if ::= "if" test [comp_iter] | |
| def p_list_maker(s): | |
| # s.sy == '[' | |
| pos = s.position() | |
| s.next() | |
| if s.sy == ']': | |
| s.expect(']') | |
| return ExprNodes.ListNode(pos, args=[]) | |
| expr = p_test_or_starred_expr(s) | |
| if s.sy in ('for', 'async'): | |
| if expr.is_starred: | |
| s.error("iterable unpacking cannot be used in comprehension") | |
| append = ExprNodes.ComprehensionAppendNode(pos, expr=expr) | |
| loop = p_comp_for(s, append) | |
| s.expect(']') | |
| return ExprNodes.ComprehensionNode( | |
| pos, loop=loop, append=append, type=Builtin.list_type, | |
| # list comprehensions leak their loop variable in Py2 | |
| has_local_scope=s.context.language_level >= 3) | |
| # (merged) list literal | |
| if s.sy == ',': | |
| s.next() | |
| exprs = p_test_or_starred_expr_list(s, expr) | |
| else: | |
| exprs = [expr] | |
| s.expect(']') | |
| return ExprNodes.ListNode(pos, args=exprs) | |
| def p_comp_iter(s, body): | |
| if s.sy in ('for', 'async'): | |
| return p_comp_for(s, body) | |
| elif s.sy == 'if': | |
| return p_comp_if(s, body) | |
| else: | |
| # insert the 'append' operation into the loop | |
| return body | |
| def p_comp_for(s, body): | |
| pos = s.position() | |
| # [async] for ... | |
| is_async = False | |
| if s.sy == 'async': | |
| is_async = True | |
| s.next() | |
| # s.sy == 'for' | |
| s.expect('for') | |
| kw = p_for_bounds(s, allow_testlist=False, is_async=is_async) | |
| kw.update(else_clause=None, body=p_comp_iter(s, body), is_async=is_async) | |
| return Nodes.ForStatNode(pos, **kw) | |
| def p_comp_if(s, body): | |
| # s.sy == 'if' | |
| pos = s.position() | |
| s.next() | |
| test = p_test_nocond(s) | |
| return Nodes.IfStatNode(pos, | |
| if_clauses = [Nodes.IfClauseNode(pos, condition = test, | |
| body = p_comp_iter(s, body))], | |
| else_clause = None ) | |
| # since PEP 448: | |
| #dictorsetmaker: ( ((test ':' test | '**' expr) | |
| # (comp_for | (',' (test ':' test | '**' expr))* [','])) | | |
| # ((test | star_expr) | |
| # (comp_for | (',' (test | star_expr))* [','])) ) | |
| def p_dict_or_set_maker(s): | |
| # s.sy == '{' | |
| pos = s.position() | |
| s.next() | |
| if s.sy == '}': | |
| s.next() | |
| return ExprNodes.DictNode(pos, key_value_pairs=[]) | |
| parts = [] | |
| target_type = 0 | |
| last_was_simple_item = False | |
| while True: | |
| if s.sy in ('*', '**'): | |
| # merged set/dict literal | |
| if target_type == 0: | |
| target_type = 1 if s.sy == '*' else 2 # 'stars' | |
| elif target_type != len(s.sy): | |
| s.error("unexpected %sitem found in %s literal" % ( | |
| s.sy, 'set' if target_type == 1 else 'dict')) | |
| s.next() | |
| if s.sy == '*': | |
| s.error("expected expression, found '*'") | |
| item = p_starred_expr(s) | |
| parts.append(item) | |
| last_was_simple_item = False | |
| else: | |
| item = p_test(s) | |
| if target_type == 0: | |
| target_type = 2 if s.sy == ':' else 1 # dict vs. set | |
| if target_type == 2: | |
| # dict literal | |
| s.expect(':') | |
| key = item | |
| value = p_test(s) | |
| item = ExprNodes.DictItemNode(key.pos, key=key, value=value) | |
| if last_was_simple_item: | |
| parts[-1].append(item) | |
| else: | |
| parts.append([item]) | |
| last_was_simple_item = True | |
| if s.sy == ',': | |
| s.next() | |
| if s.sy == '}': | |
| break | |
| else: | |
| break | |
| if s.sy in ('for', 'async'): | |
| # dict/set comprehension | |
| if len(parts) == 1 and isinstance(parts[0], list) and len(parts[0]) == 1: | |
| item = parts[0][0] | |
| if target_type == 2: | |
| assert isinstance(item, ExprNodes.DictItemNode), type(item) | |
| comprehension_type = Builtin.dict_type | |
| append = ExprNodes.DictComprehensionAppendNode( | |
| item.pos, key_expr=item.key, value_expr=item.value) | |
| else: | |
| comprehension_type = Builtin.set_type | |
| append = ExprNodes.ComprehensionAppendNode(item.pos, expr=item) | |
| loop = p_comp_for(s, append) | |
| s.expect('}') | |
| return ExprNodes.ComprehensionNode(pos, loop=loop, append=append, type=comprehension_type) | |
| else: | |
| # syntax error, try to find a good error message | |
| if len(parts) == 1 and not isinstance(parts[0], list): | |
| s.error("iterable unpacking cannot be used in comprehension") | |
| else: | |
| # e.g. "{1,2,3 for ..." | |
| s.expect('}') | |
| return ExprNodes.DictNode(pos, key_value_pairs=[]) | |
| s.expect('}') | |
| if target_type == 1: | |
| # (merged) set literal | |
| items = [] | |
| set_items = [] | |
| for part in parts: | |
| if isinstance(part, list): | |
| set_items.extend(part) | |
| else: | |
| if set_items: | |
| items.append(ExprNodes.SetNode(set_items[0].pos, args=set_items)) | |
| set_items = [] | |
| items.append(part) | |
| if set_items: | |
| items.append(ExprNodes.SetNode(set_items[0].pos, args=set_items)) | |
| if len(items) == 1 and items[0].is_set_literal: | |
| return items[0] | |
| return ExprNodes.MergedSequenceNode(pos, args=items, type=Builtin.set_type) | |
| else: | |
| # (merged) dict literal | |
| items = [] | |
| dict_items = [] | |
| for part in parts: | |
| if isinstance(part, list): | |
| dict_items.extend(part) | |
| else: | |
| if dict_items: | |
| items.append(ExprNodes.DictNode(dict_items[0].pos, key_value_pairs=dict_items)) | |
| dict_items = [] | |
| items.append(part) | |
| if dict_items: | |
| items.append(ExprNodes.DictNode(dict_items[0].pos, key_value_pairs=dict_items)) | |
| if len(items) == 1 and items[0].is_dict_literal: | |
| return items[0] | |
| return ExprNodes.MergedDictNode(pos, keyword_args=items, reject_duplicates=False) | |
| # NOTE: no longer in Py3 :) | |
| def p_backquote_expr(s): | |
| # s.sy == '`' | |
| pos = s.position() | |
| s.next() | |
| args = [p_test(s)] | |
| while s.sy == ',': | |
| s.next() | |
| args.append(p_test(s)) | |
| s.expect('`') | |
| if len(args) == 1: | |
| arg = args[0] | |
| else: | |
| arg = ExprNodes.TupleNode(pos, args = args) | |
| return ExprNodes.BackquoteNode(pos, arg = arg) | |
| def p_simple_expr_list(s, expr=None): | |
| exprs = expr is not None and [expr] or [] | |
| while s.sy not in expr_terminators: | |
| exprs.append( p_test(s) ) | |
| if s.sy != ',': | |
| break | |
| s.next() | |
| return exprs | |
| def p_test_or_starred_expr_list(s, expr=None): | |
| exprs = expr is not None and [expr] or [] | |
| while s.sy not in expr_terminators: | |
| exprs.append(p_test_or_starred_expr(s)) | |
| if s.sy != ',': | |
| break | |
| s.next() | |
| return exprs | |
| #testlist: test (',' test)* [','] | |
| def p_testlist(s): | |
| pos = s.position() | |
| expr = p_test(s) | |
| if s.sy == ',': | |
| s.next() | |
| exprs = p_simple_expr_list(s, expr) | |
| return ExprNodes.TupleNode(pos, args = exprs) | |
| else: | |
| return expr | |
| # testlist_star_expr: (test|star_expr) ( comp_for | (',' (test|star_expr))* [','] ) | |
| def p_testlist_star_expr(s): | |
| pos = s.position() | |
| expr = p_test_or_starred_expr(s) | |
| if s.sy == ',': | |
| s.next() | |
| exprs = p_test_or_starred_expr_list(s, expr) | |
| return ExprNodes.TupleNode(pos, args = exprs) | |
| else: | |
| return expr | |
| # testlist_comp: (test|star_expr) ( comp_for | (',' (test|star_expr))* [','] ) | |
| def p_testlist_comp(s): | |
| pos = s.position() | |
| expr = p_test_or_starred_expr(s) | |
| if s.sy == ',': | |
| s.next() | |
| exprs = p_test_or_starred_expr_list(s, expr) | |
| return ExprNodes.TupleNode(pos, args = exprs) | |
| elif s.sy in ('for', 'async'): | |
| return p_genexp(s, expr) | |
| else: | |
| return expr | |
| def p_genexp(s, expr): | |
| # s.sy == 'async' | 'for' | |
| loop = p_comp_for(s, Nodes.ExprStatNode( | |
| expr.pos, expr = ExprNodes.YieldExprNode(expr.pos, arg=expr))) | |
| return ExprNodes.GeneratorExpressionNode(expr.pos, loop=loop) | |
| expr_terminators = cython.declare(set, set([ | |
| ')', ']', '}', ':', '=', 'NEWLINE'])) | |
| #------------------------------------------------------- | |
| # | |
| # Statements | |
| # | |
| #------------------------------------------------------- | |
| def p_global_statement(s): | |
| # assume s.sy == 'global' | |
| pos = s.position() | |
| s.next() | |
| names = p_ident_list(s) | |
| return Nodes.GlobalNode(pos, names = names) | |
| def p_nonlocal_statement(s): | |
| pos = s.position() | |
| s.next() | |
| names = p_ident_list(s) | |
| return Nodes.NonlocalNode(pos, names = names) | |
| def p_expression_or_assignment(s): | |
| expr = p_testlist_star_expr(s) | |
| if s.sy == ':' and (expr.is_name or expr.is_subscript or expr.is_attribute): | |
| s.next() | |
| expr.annotation = p_test(s) | |
| if s.sy == '=' and expr.is_starred: | |
| # This is a common enough error to make when learning Cython to let | |
| # it fail as early as possible and give a very clear error message. | |
| s.error("a starred assignment target must be in a list or tuple" | |
| " - maybe you meant to use an index assignment: var[0] = ...", | |
| pos=expr.pos) | |
| expr_list = [expr] | |
| while s.sy == '=': | |
| s.next() | |
| if s.sy == 'yield': | |
| expr = p_yield_expression(s) | |
| else: | |
| expr = p_testlist_star_expr(s) | |
| expr_list.append(expr) | |
| if len(expr_list) == 1: | |
| if re.match(r"([-+*/%^&|]|<<|>>|\*\*|//|@)=", s.sy): | |
| lhs = expr_list[0] | |
| if isinstance(lhs, ExprNodes.SliceIndexNode): | |
| # implementation requires IndexNode | |
| lhs = ExprNodes.IndexNode( | |
| lhs.pos, | |
| base=lhs.base, | |
| index=make_slice_node(lhs.pos, lhs.start, lhs.stop)) | |
| elif not isinstance(lhs, (ExprNodes.AttributeNode, ExprNodes.IndexNode, ExprNodes.NameNode)): | |
| error(lhs.pos, "Illegal operand for inplace operation.") | |
| operator = s.sy[:-1] | |
| s.next() | |
| if s.sy == 'yield': | |
| rhs = p_yield_expression(s) | |
| else: | |
| rhs = p_testlist(s) | |
| return Nodes.InPlaceAssignmentNode(lhs.pos, operator=operator, lhs=lhs, rhs=rhs) | |
| expr = expr_list[0] | |
| return Nodes.ExprStatNode(expr.pos, expr=expr) | |
| rhs = expr_list[-1] | |
| if len(expr_list) == 2: | |
| return Nodes.SingleAssignmentNode(rhs.pos, lhs=expr_list[0], rhs=rhs) | |
| else: | |
| return Nodes.CascadedAssignmentNode(rhs.pos, lhs_list=expr_list[:-1], rhs=rhs) | |
| def p_print_statement(s): | |
| # s.sy == 'print' | |
| pos = s.position() | |
| ends_with_comma = 0 | |
| s.next() | |
| if s.sy == '>>': | |
| s.next() | |
| stream = p_test(s) | |
| if s.sy == ',': | |
| s.next() | |
| ends_with_comma = s.sy in ('NEWLINE', 'EOF') | |
| else: | |
| stream = None | |
| args = [] | |
| if s.sy not in ('NEWLINE', 'EOF'): | |
| args.append(p_test(s)) | |
| while s.sy == ',': | |
| s.next() | |
| if s.sy in ('NEWLINE', 'EOF'): | |
| ends_with_comma = 1 | |
| break | |
| args.append(p_test(s)) | |
| arg_tuple = ExprNodes.TupleNode(pos, args=args) | |
| return Nodes.PrintStatNode(pos, | |
| arg_tuple=arg_tuple, stream=stream, | |
| append_newline=not ends_with_comma) | |
| def p_exec_statement(s): | |
| # s.sy == 'exec' | |
| pos = s.position() | |
| s.next() | |
| code = p_bit_expr(s) | |
| if isinstance(code, ExprNodes.TupleNode): | |
| # Py3 compatibility syntax | |
| tuple_variant = True | |
| args = code.args | |
| if len(args) not in (2, 3): | |
| s.error("expected tuple of length 2 or 3, got length %d" % len(args), | |
| pos=pos, fatal=False) | |
| args = [code] | |
| else: | |
| tuple_variant = False | |
| args = [code] | |
| if s.sy == 'in': | |
| if tuple_variant: | |
| s.error("tuple variant of exec does not support additional 'in' arguments", | |
| fatal=False) | |
| s.next() | |
| args.append(p_test(s)) | |
| if s.sy == ',': | |
| s.next() | |
| args.append(p_test(s)) | |
| return Nodes.ExecStatNode(pos, args=args) | |
| def p_del_statement(s): | |
| # s.sy == 'del' | |
| pos = s.position() | |
| s.next() | |
| # FIXME: 'exprlist' in Python | |
| args = p_simple_expr_list(s) | |
| return Nodes.DelStatNode(pos, args = args) | |
| def p_pass_statement(s, with_newline = 0): | |
| pos = s.position() | |
| s.expect('pass') | |
| if with_newline: | |
| s.expect_newline("Expected a newline", ignore_semicolon=True) | |
| return Nodes.PassStatNode(pos) | |
| def p_break_statement(s): | |
| # s.sy == 'break' | |
| pos = s.position() | |
| s.next() | |
| return Nodes.BreakStatNode(pos) | |
| def p_continue_statement(s): | |
| # s.sy == 'continue' | |
| pos = s.position() | |
| s.next() | |
| return Nodes.ContinueStatNode(pos) | |
| def p_return_statement(s): | |
| # s.sy == 'return' | |
| pos = s.position() | |
| s.next() | |
| if s.sy not in statement_terminators: | |
| value = p_testlist(s) | |
| else: | |
| value = None | |
| return Nodes.ReturnStatNode(pos, value = value) | |
| def p_raise_statement(s): | |
| # s.sy == 'raise' | |
| pos = s.position() | |
| s.next() | |
| exc_type = None | |
| exc_value = None | |
| exc_tb = None | |
| cause = None | |
| if s.sy not in statement_terminators: | |
| exc_type = p_test(s) | |
| if s.sy == ',': | |
| s.next() | |
| exc_value = p_test(s) | |
| if s.sy == ',': | |
| s.next() | |
| exc_tb = p_test(s) | |
| elif s.sy == 'from': | |
| s.next() | |
| cause = p_test(s) | |
| if exc_type or exc_value or exc_tb: | |
| return Nodes.RaiseStatNode(pos, | |
| exc_type = exc_type, | |
| exc_value = exc_value, | |
| exc_tb = exc_tb, | |
| cause = cause) | |
| else: | |
| return Nodes.ReraiseStatNode(pos) | |
| def p_import_statement(s): | |
| # s.sy in ('import', 'cimport') | |
| pos = s.position() | |
| kind = s.sy | |
| s.next() | |
| items = [p_dotted_name(s, as_allowed=1)] | |
| while s.sy == ',': | |
| s.next() | |
| items.append(p_dotted_name(s, as_allowed=1)) | |
| stats = [] | |
| is_absolute = Future.absolute_import in s.context.future_directives | |
| for pos, target_name, dotted_name, as_name in items: | |
| if kind == 'cimport': | |
| stat = Nodes.CImportStatNode( | |
| pos, | |
| module_name=dotted_name, | |
| as_name=as_name, | |
| is_absolute=is_absolute) | |
| else: | |
| if as_name and "." in dotted_name: | |
| name_list = ExprNodes.ListNode(pos, args=[ | |
| ExprNodes.IdentifierStringNode(pos, value=s.context.intern_ustring("*"))]) | |
| else: | |
| name_list = None | |
| stat = Nodes.SingleAssignmentNode( | |
| pos, | |
| lhs=ExprNodes.NameNode(pos, name=as_name or target_name), | |
| rhs=ExprNodes.ImportNode( | |
| pos, | |
| module_name=ExprNodes.IdentifierStringNode(pos, value=dotted_name), | |
| level=0 if is_absolute else None, | |
| name_list=name_list)) | |
| stats.append(stat) | |
| return Nodes.StatListNode(pos, stats=stats) | |
| def p_from_import_statement(s, first_statement = 0): | |
| # s.sy == 'from' | |
| pos = s.position() | |
| s.next() | |
| if s.sy == '.': | |
| # count relative import level | |
| level = 0 | |
| while s.sy == '.': | |
| level += 1 | |
| s.next() | |
| else: | |
| level = None | |
| if level is not None and s.sy in ('import', 'cimport'): | |
| # we are dealing with "from .. import foo, bar" | |
| dotted_name_pos, dotted_name = s.position(), s.context.intern_ustring('') | |
| else: | |
| if level is None and Future.absolute_import in s.context.future_directives: | |
| level = 0 | |
| (dotted_name_pos, _, dotted_name, _) = p_dotted_name(s, as_allowed=False) | |
| if s.sy not in ('import', 'cimport'): | |
| s.error("Expected 'import' or 'cimport'") | |
| kind = s.sy | |
| s.next() | |
| is_cimport = kind == 'cimport' | |
| is_parenthesized = False | |
| if s.sy == '*': | |
| imported_names = [(s.position(), s.context.intern_ustring("*"), None, None)] | |
| s.next() | |
| else: | |
| if s.sy == '(': | |
| is_parenthesized = True | |
| s.next() | |
| imported_names = [p_imported_name(s, is_cimport)] | |
| while s.sy == ',': | |
| s.next() | |
| if is_parenthesized and s.sy == ')': | |
| break | |
| imported_names.append(p_imported_name(s, is_cimport)) | |
| if is_parenthesized: | |
| s.expect(')') | |
| if dotted_name == '__future__': | |
| if not first_statement: | |
| s.error("from __future__ imports must occur at the beginning of the file") | |
| elif level: | |
| s.error("invalid syntax") | |
| else: | |
| for (name_pos, name, as_name, kind) in imported_names: | |
| if name == "braces": | |
| s.error("not a chance", name_pos) | |
| break | |
| try: | |
| directive = getattr(Future, name) | |
| except AttributeError: | |
| s.error("future feature %s is not defined" % name, name_pos) | |
| break | |
| s.context.future_directives.add(directive) | |
| return Nodes.PassStatNode(pos) | |
| elif kind == 'cimport': | |
| return Nodes.FromCImportStatNode( | |
| pos, module_name=dotted_name, | |
| relative_level=level, | |
| imported_names=imported_names) | |
| else: | |
| imported_name_strings = [] | |
| items = [] | |
| for (name_pos, name, as_name, kind) in imported_names: | |
| imported_name_strings.append( | |
| ExprNodes.IdentifierStringNode(name_pos, value=name)) | |
| items.append( | |
| (name, ExprNodes.NameNode(name_pos, name=as_name or name))) | |
| import_list = ExprNodes.ListNode( | |
| imported_names[0][0], args=imported_name_strings) | |
| return Nodes.FromImportStatNode(pos, | |
| module = ExprNodes.ImportNode(dotted_name_pos, | |
| module_name = ExprNodes.IdentifierStringNode(pos, value = dotted_name), | |
| level = level, | |
| name_list = import_list), | |
| items = items) | |
| imported_name_kinds = cython.declare(set, set(['class', 'struct', 'union'])) | |
| def p_imported_name(s, is_cimport): | |
| pos = s.position() | |
| kind = None | |
| if is_cimport and s.systring in imported_name_kinds: | |
| kind = s.systring | |
| s.next() | |
| name = p_ident(s) | |
| as_name = p_as_name(s) | |
| return (pos, name, as_name, kind) | |
| def p_dotted_name(s, as_allowed): | |
| pos = s.position() | |
| target_name = p_ident(s) | |
| as_name = None | |
| names = [target_name] | |
| while s.sy == '.': | |
| s.next() | |
| names.append(p_ident(s)) | |
| if as_allowed: | |
| as_name = p_as_name(s) | |
| return (pos, target_name, s.context.intern_ustring(u'.'.join(names)), as_name) | |
| def p_as_name(s): | |
| if s.sy == 'IDENT' and s.systring == 'as': | |
| s.next() | |
| return p_ident(s) | |
| else: | |
| return None | |
| def p_assert_statement(s): | |
| # s.sy == 'assert' | |
| pos = s.position() | |
| s.next() | |
| cond = p_test(s) | |
| if s.sy == ',': | |
| s.next() | |
| value = p_test(s) | |
| else: | |
| value = None | |
| return Nodes.AssertStatNode(pos, cond = cond, value = value) | |
| statement_terminators = cython.declare(set, set([';', 'NEWLINE', 'EOF'])) | |
| def p_if_statement(s): | |
| # s.sy == 'if' | |
| pos = s.position() | |
| s.next() | |
| if_clauses = [p_if_clause(s)] | |
| while s.sy == 'elif': | |
| s.next() | |
| if_clauses.append(p_if_clause(s)) | |
| else_clause = p_else_clause(s) | |
| return Nodes.IfStatNode(pos, | |
| if_clauses = if_clauses, else_clause = else_clause) | |
| def p_if_clause(s): | |
| pos = s.position() | |
| test = p_test(s) | |
| body = p_suite(s) | |
| return Nodes.IfClauseNode(pos, | |
| condition = test, body = body) | |
| def p_else_clause(s): | |
| if s.sy == 'else': | |
| s.next() | |
| return p_suite(s) | |
| else: | |
| return None | |
| def p_while_statement(s): | |
| # s.sy == 'while' | |
| pos = s.position() | |
| s.next() | |
| test = p_test(s) | |
| body = p_suite(s) | |
| else_clause = p_else_clause(s) | |
| return Nodes.WhileStatNode(pos, | |
| condition = test, body = body, | |
| else_clause = else_clause) | |
| def p_for_statement(s, is_async=False): | |
| # s.sy == 'for' | |
| pos = s.position() | |
| s.next() | |
| kw = p_for_bounds(s, allow_testlist=True, is_async=is_async) | |
| body = p_suite(s) | |
| else_clause = p_else_clause(s) | |
| kw.update(body=body, else_clause=else_clause, is_async=is_async) | |
| return Nodes.ForStatNode(pos, **kw) | |
| def p_for_bounds(s, allow_testlist=True, is_async=False): | |
| target = p_for_target(s) | |
| if s.sy == 'in': | |
| s.next() | |
| iterator = p_for_iterator(s, allow_testlist, is_async=is_async) | |
| return dict(target=target, iterator=iterator) | |
| elif not s.in_python_file and not is_async: | |
| if s.sy == 'from': | |
| s.next() | |
| bound1 = p_bit_expr(s) | |
| else: | |
| # Support shorter "for a <= x < b" syntax | |
| bound1, target = target, None | |
| rel1 = p_for_from_relation(s) | |
| name2_pos = s.position() | |
| name2 = p_ident(s) | |
| rel2_pos = s.position() | |
| rel2 = p_for_from_relation(s) | |
| bound2 = p_bit_expr(s) | |
| step = p_for_from_step(s) | |
| if target is None: | |
| target = ExprNodes.NameNode(name2_pos, name = name2) | |
| else: | |
| if not target.is_name: | |
| error(target.pos, | |
| "Target of for-from statement must be a variable name") | |
| elif name2 != target.name: | |
| error(name2_pos, | |
| "Variable name in for-from range does not match target") | |
| if rel1[0] != rel2[0]: | |
| error(rel2_pos, | |
| "Relation directions in for-from do not match") | |
| return dict(target = target, | |
| bound1 = bound1, | |
| relation1 = rel1, | |
| relation2 = rel2, | |
| bound2 = bound2, | |
| step = step, | |
| ) | |
| else: | |
| s.expect('in') | |
| return {} | |
| def p_for_from_relation(s): | |
| if s.sy in inequality_relations: | |
| op = s.sy | |
| s.next() | |
| return op | |
| else: | |
| s.error("Expected one of '<', '<=', '>' '>='") | |
| def p_for_from_step(s): | |
| if s.sy == 'IDENT' and s.systring == 'by': | |
| s.next() | |
| step = p_bit_expr(s) | |
| return step | |
| else: | |
| return None | |
| inequality_relations = cython.declare(set, set(['<', '<=', '>', '>='])) | |
| def p_target(s, terminator): | |
| pos = s.position() | |
| expr = p_starred_expr(s) | |
| if s.sy == ',': | |
| s.next() | |
| exprs = [expr] | |
| while s.sy != terminator: | |
| exprs.append(p_starred_expr(s)) | |
| if s.sy != ',': | |
| break | |
| s.next() | |
| return ExprNodes.TupleNode(pos, args = exprs) | |
| else: | |
| return expr | |
| def p_for_target(s): | |
| return p_target(s, 'in') | |
| def p_for_iterator(s, allow_testlist=True, is_async=False): | |
| pos = s.position() | |
| if allow_testlist: | |
| expr = p_testlist(s) | |
| else: | |
| expr = p_or_test(s) | |
| return (ExprNodes.AsyncIteratorNode if is_async else ExprNodes.IteratorNode)(pos, sequence=expr) | |
| def p_try_statement(s): | |
| # s.sy == 'try' | |
| pos = s.position() | |
| s.next() | |
| body = p_suite(s) | |
| except_clauses = [] | |
| else_clause = None | |
| if s.sy in ('except', 'else'): | |
| while s.sy == 'except': | |
| except_clauses.append(p_except_clause(s)) | |
| if s.sy == 'else': | |
| s.next() | |
| else_clause = p_suite(s) | |
| body = Nodes.TryExceptStatNode(pos, | |
| body = body, except_clauses = except_clauses, | |
| else_clause = else_clause) | |
| if s.sy != 'finally': | |
| return body | |
| # try-except-finally is equivalent to nested try-except/try-finally | |
| if s.sy == 'finally': | |
| s.next() | |
| finally_clause = p_suite(s) | |
| return Nodes.TryFinallyStatNode(pos, | |
| body = body, finally_clause = finally_clause) | |
| else: | |
| s.error("Expected 'except' or 'finally'") | |
| def p_except_clause(s): | |
| # s.sy == 'except' | |
| pos = s.position() | |
| s.next() | |
| exc_type = None | |
| exc_value = None | |
| is_except_as = False | |
| if s.sy != ':': | |
| exc_type = p_test(s) | |
| # normalise into list of single exception tests | |
| if isinstance(exc_type, ExprNodes.TupleNode): | |
| exc_type = exc_type.args | |
| else: | |
| exc_type = [exc_type] | |
| if s.sy == ',' or (s.sy == 'IDENT' and s.systring == 'as' | |
| and s.context.language_level == 2): | |
| s.next() | |
| exc_value = p_test(s) | |
| elif s.sy == 'IDENT' and s.systring == 'as': | |
| # Py3 syntax requires a name here | |
| s.next() | |
| pos2 = s.position() | |
| name = p_ident(s) | |
| exc_value = ExprNodes.NameNode(pos2, name = name) | |
| is_except_as = True | |
| body = p_suite(s) | |
| return Nodes.ExceptClauseNode(pos, | |
| pattern = exc_type, target = exc_value, | |
| body = body, is_except_as=is_except_as) | |
| def p_include_statement(s, ctx): | |
| pos = s.position() | |
| s.next() # 'include' | |
| unicode_include_file_name = p_string_literal(s, 'u')[2] | |
| s.expect_newline("Syntax error in include statement") | |
| if s.compile_time_eval: | |
| include_file_name = unicode_include_file_name | |
| include_file_path = s.context.find_include_file(include_file_name, pos) | |
| if include_file_path: | |
| s.included_files.append(include_file_name) | |
| with Utils.open_source_file(include_file_path) as f: | |
| source_desc = FileSourceDescriptor(include_file_path) | |
| s2 = PyrexScanner(f, source_desc, s, source_encoding=f.encoding, parse_comments=s.parse_comments) | |
| tree = p_statement_list(s2, ctx) | |
| return tree | |
| else: | |
| return None | |
| else: | |
| return Nodes.PassStatNode(pos) | |
| def p_with_statement(s): | |
| s.next() # 'with' | |
| if s.systring == 'template' and not s.in_python_file: | |
| node = p_with_template(s) | |
| else: | |
| node = p_with_items(s) | |
| return node | |
| def p_with_items(s, is_async=False): | |
| pos = s.position() | |
| if not s.in_python_file and s.sy == 'IDENT' and s.systring in ('nogil', 'gil'): | |
| if is_async: | |
| s.error("with gil/nogil cannot be async") | |
| state = s.systring | |
| s.next() | |
| if s.sy == ',': | |
| s.next() | |
| body = p_with_items(s) | |
| else: | |
| body = p_suite(s) | |
| return Nodes.GILStatNode(pos, state=state, body=body) | |
| else: | |
| manager = p_test(s) | |
| target = None | |
| if s.sy == 'IDENT' and s.systring == 'as': | |
| s.next() | |
| target = p_starred_expr(s) | |
| if s.sy == ',': | |
| s.next() | |
| body = p_with_items(s, is_async=is_async) | |
| else: | |
| body = p_suite(s) | |
| return Nodes.WithStatNode(pos, manager=manager, target=target, body=body, is_async=is_async) | |
| def p_with_template(s): | |
| pos = s.position() | |
| templates = [] | |
| s.next() | |
| s.expect('[') | |
| templates.append(s.systring) | |
| s.next() | |
| while s.systring == ',': | |
| s.next() | |
| templates.append(s.systring) | |
| s.next() | |
| s.expect(']') | |
| if s.sy == ':': | |
| s.next() | |
| s.expect_newline("Syntax error in template function declaration") | |
| s.expect_indent() | |
| body_ctx = Ctx() | |
| body_ctx.templates = templates | |
| func_or_var = p_c_func_or_var_declaration(s, pos, body_ctx) | |
| s.expect_dedent() | |
| return func_or_var | |
| else: | |
| error(pos, "Syntax error in template function declaration") | |
| def p_simple_statement(s, first_statement = 0): | |
| #print "p_simple_statement:", s.sy, s.systring ### | |
| if s.sy == 'global': | |
| node = p_global_statement(s) | |
| elif s.sy == 'nonlocal': | |
| node = p_nonlocal_statement(s) | |
| elif s.sy == 'print': | |
| node = p_print_statement(s) | |
| elif s.sy == 'exec': | |
| node = p_exec_statement(s) | |
| elif s.sy == 'del': | |
| node = p_del_statement(s) | |
| elif s.sy == 'break': | |
| node = p_break_statement(s) | |
| elif s.sy == 'continue': | |
| node = p_continue_statement(s) | |
| elif s.sy == 'return': | |
| node = p_return_statement(s) | |
| elif s.sy == 'raise': | |
| node = p_raise_statement(s) | |
| elif s.sy in ('import', 'cimport'): | |
| node = p_import_statement(s) | |
| elif s.sy == 'from': | |
| node = p_from_import_statement(s, first_statement = first_statement) | |
| elif s.sy == 'yield': | |
| node = p_yield_statement(s) | |
| elif s.sy == 'assert': | |
| node = p_assert_statement(s) | |
| elif s.sy == 'pass': | |
| node = p_pass_statement(s) | |
| else: | |
| node = p_expression_or_assignment(s) | |
| return node | |
| def p_simple_statement_list(s, ctx, first_statement = 0): | |
| # Parse a series of simple statements on one line | |
| # separated by semicolons. | |
| stat = p_simple_statement(s, first_statement = first_statement) | |
| pos = stat.pos | |
| stats = [] | |
| if not isinstance(stat, Nodes.PassStatNode): | |
| stats.append(stat) | |
| while s.sy == ';': | |
| #print "p_simple_statement_list: maybe more to follow" ### | |
| s.next() | |
| if s.sy in ('NEWLINE', 'EOF'): | |
| break | |
| stat = p_simple_statement(s, first_statement = first_statement) | |
| if isinstance(stat, Nodes.PassStatNode): | |
| continue | |
| stats.append(stat) | |
| first_statement = False | |
| if not stats: | |
| stat = Nodes.PassStatNode(pos) | |
| elif len(stats) == 1: | |
| stat = stats[0] | |
| else: | |
| stat = Nodes.StatListNode(pos, stats = stats) | |
| if s.sy not in ('NEWLINE', 'EOF'): | |
| # provide a better error message for users who accidentally write Cython code in .py files | |
| if isinstance(stat, Nodes.ExprStatNode): | |
| if stat.expr.is_name and stat.expr.name == 'cdef': | |
| s.error("The 'cdef' keyword is only allowed in Cython files (pyx/pxi/pxd)", pos) | |
| s.expect_newline("Syntax error in simple statement list") | |
| return stat | |
| def p_compile_time_expr(s): | |
| old = s.compile_time_expr | |
| s.compile_time_expr = 1 | |
| expr = p_testlist(s) | |
| s.compile_time_expr = old | |
| return expr | |
| def p_DEF_statement(s): | |
| pos = s.position() | |
| denv = s.compile_time_env | |
| s.next() # 'DEF' | |
| name = p_ident(s) | |
| s.expect('=') | |
| expr = p_compile_time_expr(s) | |
| if s.compile_time_eval: | |
| value = expr.compile_time_value(denv) | |
| #print "p_DEF_statement: %s = %r" % (name, value) ### | |
| denv.declare(name, value) | |
| s.expect_newline("Expected a newline", ignore_semicolon=True) | |
| return Nodes.PassStatNode(pos) | |
| def p_IF_statement(s, ctx): | |
| pos = s.position() | |
| saved_eval = s.compile_time_eval | |
| current_eval = saved_eval | |
| denv = s.compile_time_env | |
| result = None | |
| while 1: | |
| s.next() # 'IF' or 'ELIF' | |
| expr = p_compile_time_expr(s) | |
| s.compile_time_eval = current_eval and bool(expr.compile_time_value(denv)) | |
| body = p_suite(s, ctx) | |
| if s.compile_time_eval: | |
| result = body | |
| current_eval = 0 | |
| if s.sy != 'ELIF': | |
| break | |
| if s.sy == 'ELSE': | |
| s.next() | |
| s.compile_time_eval = current_eval | |
| body = p_suite(s, ctx) | |
| if current_eval: | |
| result = body | |
| if not result: | |
| result = Nodes.PassStatNode(pos) | |
| s.compile_time_eval = saved_eval | |
| return result | |
| def p_statement(s, ctx, first_statement = 0): | |
| cdef_flag = ctx.cdef_flag | |
| decorators = None | |
| if s.sy == 'ctypedef': | |
| if ctx.level not in ('module', 'module_pxd'): | |
| s.error("ctypedef statement not allowed here") | |
| #if ctx.api: | |
| # error(s.position(), "'api' not allowed with 'ctypedef'") | |
| return p_ctypedef_statement(s, ctx) | |
| elif s.sy == 'DEF': | |
| return p_DEF_statement(s) | |
| elif s.sy == 'IF': | |
| return p_IF_statement(s, ctx) | |
| elif s.sy == '@': | |
| if ctx.level not in ('module', 'class', 'c_class', 'function', 'property', 'module_pxd', 'c_class_pxd', 'other'): | |
| s.error('decorator not allowed here') | |
| s.level = ctx.level | |
| decorators = p_decorators(s) | |
| if not ctx.allow_struct_enum_decorator and s.sy not in ('def', 'cdef', 'cpdef', 'class', 'async'): | |
| if s.sy == 'IDENT' and s.systring == 'async': | |
| pass # handled below | |
| else: | |
| s.error("Decorators can only be followed by functions or classes") | |
| elif s.sy == 'pass' and cdef_flag: | |
| # empty cdef block | |
| return p_pass_statement(s, with_newline=1) | |
| overridable = 0 | |
| if s.sy == 'cdef': | |
| cdef_flag = 1 | |
| s.next() | |
| elif s.sy == 'cpdef': | |
| cdef_flag = 1 | |
| overridable = 1 | |
| s.next() | |
| if cdef_flag: | |
| if ctx.level not in ('module', 'module_pxd', 'function', 'c_class', 'c_class_pxd'): | |
| s.error('cdef statement not allowed here') | |
| s.level = ctx.level | |
| node = p_cdef_statement(s, ctx(overridable=overridable)) | |
| if decorators is not None: | |
| tup = (Nodes.CFuncDefNode, Nodes.CVarDefNode, Nodes.CClassDefNode) | |
| if ctx.allow_struct_enum_decorator: | |
| tup += (Nodes.CStructOrUnionDefNode, Nodes.CEnumDefNode) | |
| if not isinstance(node, tup): | |
| s.error("Decorators can only be followed by functions or classes") | |
| node.decorators = decorators | |
| return node | |
| else: | |
| if ctx.api: | |
| s.error("'api' not allowed with this statement", fatal=False) | |
| elif s.sy == 'def': | |
| # def statements aren't allowed in pxd files, except | |
| # as part of a cdef class | |
| if ('pxd' in ctx.level) and (ctx.level != 'c_class_pxd'): | |
| s.error('def statement not allowed here') | |
| s.level = ctx.level | |
| return p_def_statement(s, decorators) | |
| elif s.sy == 'class': | |
| if ctx.level not in ('module', 'function', 'class', 'other'): | |
| s.error("class definition not allowed here") | |
| return p_class_statement(s, decorators) | |
| elif s.sy == 'include': | |
| if ctx.level not in ('module', 'module_pxd'): | |
| s.error("include statement not allowed here") | |
| return p_include_statement(s, ctx) | |
| elif ctx.level == 'c_class' and s.sy == 'IDENT' and s.systring == 'property': | |
| return p_property_decl(s) | |
| elif s.sy == 'pass' and ctx.level != 'property': | |
| return p_pass_statement(s, with_newline=True) | |
| else: | |
| if ctx.level in ('c_class_pxd', 'property'): | |
| node = p_ignorable_statement(s) | |
| if node is not None: | |
| return node | |
| s.error("Executable statement not allowed here") | |
| if s.sy == 'if': | |
| return p_if_statement(s) | |
| elif s.sy == 'while': | |
| return p_while_statement(s) | |
| elif s.sy == 'for': | |
| return p_for_statement(s) | |
| elif s.sy == 'try': | |
| return p_try_statement(s) | |
| elif s.sy == 'with': | |
| return p_with_statement(s) | |
| elif s.sy == 'async': | |
| s.next() | |
| return p_async_statement(s, ctx, decorators) | |
| else: | |
| if s.sy == 'IDENT' and s.systring == 'async': | |
| ident_name = s.systring | |
| # PEP 492 enables the async/await keywords when it spots "async def ..." | |
| s.next() | |
| if s.sy == 'def': | |
| return p_async_statement(s, ctx, decorators) | |
| elif decorators: | |
| s.error("Decorators can only be followed by functions or classes") | |
| s.put_back('IDENT', ident_name) # re-insert original token | |
| return p_simple_statement_list(s, ctx, first_statement=first_statement) | |
| def p_statement_list(s, ctx, first_statement = 0): | |
| # Parse a series of statements separated by newlines. | |
| pos = s.position() | |
| stats = [] | |
| while s.sy not in ('DEDENT', 'EOF'): | |
| stat = p_statement(s, ctx, first_statement = first_statement) | |
| if isinstance(stat, Nodes.PassStatNode): | |
| continue | |
| stats.append(stat) | |
| first_statement = False | |
| if not stats: | |
| return Nodes.PassStatNode(pos) | |
| elif len(stats) == 1: | |
| return stats[0] | |
| else: | |
| return Nodes.StatListNode(pos, stats = stats) | |
| def p_suite(s, ctx=Ctx()): | |
| return p_suite_with_docstring(s, ctx, with_doc_only=False)[1] | |
| def p_suite_with_docstring(s, ctx, with_doc_only=False): | |
| s.expect(':') | |
| doc = None | |
| if s.sy == 'NEWLINE': | |
| s.next() | |
| s.expect_indent() | |
| if with_doc_only: | |
| doc = p_doc_string(s) | |
| body = p_statement_list(s, ctx) | |
| s.expect_dedent() | |
| else: | |
| if ctx.api: | |
| s.error("'api' not allowed with this statement", fatal=False) | |
| if ctx.level in ('module', 'class', 'function', 'other'): | |
| body = p_simple_statement_list(s, ctx) | |
| else: | |
| body = p_pass_statement(s) | |
| s.expect_newline("Syntax error in declarations", ignore_semicolon=True) | |
| if not with_doc_only: | |
| doc, body = _extract_docstring(body) | |
| return doc, body | |
| def p_positional_and_keyword_args(s, end_sy_set, templates = None): | |
| """ | |
| Parses positional and keyword arguments. end_sy_set | |
| should contain any s.sy that terminate the argument list. | |
| Argument expansion (* and **) are not allowed. | |
| Returns: (positional_args, keyword_args) | |
| """ | |
| positional_args = [] | |
| keyword_args = [] | |
| pos_idx = 0 | |
| while s.sy not in end_sy_set: | |
| if s.sy == '*' or s.sy == '**': | |
| s.error('Argument expansion not allowed here.', fatal=False) | |
| parsed_type = False | |
| if s.sy == 'IDENT' and s.peek()[0] == '=': | |
| ident = s.systring | |
| s.next() # s.sy is '=' | |
| s.next() | |
| if looking_at_expr(s): | |
| arg = p_test(s) | |
| else: | |
| base_type = p_c_base_type(s, templates = templates) | |
| declarator = p_c_declarator(s, empty = 1) | |
| arg = Nodes.CComplexBaseTypeNode(base_type.pos, | |
| base_type = base_type, declarator = declarator) | |
| parsed_type = True | |
| keyword_node = ExprNodes.IdentifierStringNode(arg.pos, value=ident) | |
| keyword_args.append((keyword_node, arg)) | |
| was_keyword = True | |
| else: | |
| if looking_at_expr(s): | |
| arg = p_test(s) | |
| else: | |
| base_type = p_c_base_type(s, templates = templates) | |
| declarator = p_c_declarator(s, empty = 1) | |
| arg = Nodes.CComplexBaseTypeNode(base_type.pos, | |
| base_type = base_type, declarator = declarator) | |
| parsed_type = True | |
| positional_args.append(arg) | |
| pos_idx += 1 | |
| if len(keyword_args) > 0: | |
| s.error("Non-keyword arg following keyword arg", | |
| pos=arg.pos) | |
| if s.sy != ',': | |
| if s.sy not in end_sy_set: | |
| if parsed_type: | |
| s.error("Unmatched %s" % " or ".join(end_sy_set)) | |
| break | |
| s.next() | |
| return positional_args, keyword_args | |
| def p_c_base_type(s, self_flag = 0, nonempty = 0, templates = None): | |
| # If self_flag is true, this is the base type for the | |
| # self argument of a C method of an extension type. | |
| if s.sy == '(': | |
| return p_c_complex_base_type(s, templates = templates) | |
| else: | |
| return p_c_simple_base_type(s, self_flag, nonempty = nonempty, templates = templates) | |
| def p_calling_convention(s): | |
| if s.sy == 'IDENT' and s.systring in calling_convention_words: | |
| result = s.systring | |
| s.next() | |
| return result | |
| else: | |
| return "" | |
| calling_convention_words = cython.declare( | |
| set, set(["__stdcall", "__cdecl", "__fastcall"])) | |
| def p_c_complex_base_type(s, templates = None): | |
| # s.sy == '(' | |
| pos = s.position() | |
| s.next() | |
| base_type = p_c_base_type(s, templates=templates) | |
| declarator = p_c_declarator(s, empty=True) | |
| type_node = Nodes.CComplexBaseTypeNode( | |
| pos, base_type=base_type, declarator=declarator) | |
| if s.sy == ',': | |
| components = [type_node] | |
| while s.sy == ',': | |
| s.next() | |
| if s.sy == ')': | |
| break | |
| base_type = p_c_base_type(s, templates=templates) | |
| declarator = p_c_declarator(s, empty=True) | |
| components.append(Nodes.CComplexBaseTypeNode( | |
| pos, base_type=base_type, declarator=declarator)) | |
| type_node = Nodes.CTupleBaseTypeNode(pos, components = components) | |
| s.expect(')') | |
| if s.sy == '[': | |
| if is_memoryviewslice_access(s): | |
| type_node = p_memoryviewslice_access(s, type_node) | |
| else: | |
| type_node = p_buffer_or_template(s, type_node, templates) | |
| return type_node | |
| def p_c_simple_base_type(s, self_flag, nonempty, templates = None): | |
| #print "p_c_simple_base_type: self_flag =", self_flag, nonempty | |
| is_basic = 0 | |
| signed = 1 | |
| longness = 0 | |
| complex = 0 | |
| module_path = [] | |
| pos = s.position() | |
| if not s.sy == 'IDENT': | |
| error(pos, "Expected an identifier, found '%s'" % s.sy) | |
| if s.systring == 'const': | |
| s.next() | |
| base_type = p_c_base_type(s, self_flag=self_flag, nonempty=nonempty, templates=templates) | |
| if isinstance(base_type, Nodes.MemoryViewSliceTypeNode): | |
| # reverse order to avoid having to write "(const int)[:]" | |
| base_type.base_type_node = Nodes.CConstTypeNode(pos, base_type=base_type.base_type_node) | |
| return base_type | |
| return Nodes.CConstTypeNode(pos, base_type=base_type) | |
| if looking_at_base_type(s): | |
| #print "p_c_simple_base_type: looking_at_base_type at", s.position() | |
| is_basic = 1 | |
| if s.sy == 'IDENT' and s.systring in special_basic_c_types: | |
| signed, longness = special_basic_c_types[s.systring] | |
| name = s.systring | |
| s.next() | |
| else: | |
| signed, longness = p_sign_and_longness(s) | |
| if s.sy == 'IDENT' and s.systring in basic_c_type_names: | |
| name = s.systring | |
| s.next() | |
| else: | |
| name = 'int' # long [int], short [int], long [int] complex, etc. | |
| if s.sy == 'IDENT' and s.systring == 'complex': | |
| complex = 1 | |
| s.next() | |
| elif looking_at_dotted_name(s): | |
| #print "p_c_simple_base_type: looking_at_type_name at", s.position() | |
| name = s.systring | |
| s.next() | |
| while s.sy == '.': | |
| module_path.append(name) | |
| s.next() | |
| name = p_ident(s) | |
| else: | |
| name = s.systring | |
| s.next() | |
| if nonempty and s.sy != 'IDENT': | |
| # Make sure this is not a declaration of a variable or function. | |
| if s.sy == '(': | |
| s.next() | |
| if (s.sy == '*' or s.sy == '**' or s.sy == '&' | |
| or (s.sy == 'IDENT' and s.systring in calling_convention_words)): | |
| s.put_back('(', '(') | |
| else: | |
| s.put_back('(', '(') | |
| s.put_back('IDENT', name) | |
| name = None | |
| elif s.sy not in ('*', '**', '[', '&'): | |
| s.put_back('IDENT', name) | |
| name = None | |
| type_node = Nodes.CSimpleBaseTypeNode(pos, | |
| name = name, module_path = module_path, | |
| is_basic_c_type = is_basic, signed = signed, | |
| complex = complex, longness = longness, | |
| is_self_arg = self_flag, templates = templates) | |
| # declarations here. | |
| if s.sy == '[': | |
| if is_memoryviewslice_access(s): | |
| type_node = p_memoryviewslice_access(s, type_node) | |
| else: | |
| type_node = p_buffer_or_template(s, type_node, templates) | |
| if s.sy == '.': | |
| s.next() | |
| name = p_ident(s) | |
| type_node = Nodes.CNestedBaseTypeNode(pos, base_type = type_node, name = name) | |
| return type_node | |
| def p_buffer_or_template(s, base_type_node, templates): | |
| # s.sy == '[' | |
| pos = s.position() | |
| s.next() | |
| # Note that buffer_positional_options_count=1, so the only positional argument is dtype. | |
| # For templated types, all parameters are types. | |
| positional_args, keyword_args = ( | |
| p_positional_and_keyword_args(s, (']',), templates) | |
| ) | |
| s.expect(']') | |
| if s.sy == '[': | |
| base_type_node = p_buffer_or_template(s, base_type_node, templates) | |
| keyword_dict = ExprNodes.DictNode(pos, | |
| key_value_pairs = [ | |
| ExprNodes.DictItemNode(pos=key.pos, key=key, value=value) | |
| for key, value in keyword_args | |
| ]) | |
| result = Nodes.TemplatedTypeNode(pos, | |
| positional_args = positional_args, | |
| keyword_args = keyword_dict, | |
| base_type_node = base_type_node) | |
| return result | |
| def p_bracketed_base_type(s, base_type_node, nonempty, empty): | |
| # s.sy == '[' | |
| if empty and not nonempty: | |
| # sizeof-like thing. Only anonymous C arrays allowed (int[SIZE]). | |
| return base_type_node | |
| elif not empty and nonempty: | |
| # declaration of either memoryview slice or buffer. | |
| if is_memoryviewslice_access(s): | |
| return p_memoryviewslice_access(s, base_type_node) | |
| else: | |
| return p_buffer_or_template(s, base_type_node, None) | |
| # return p_buffer_access(s, base_type_node) | |
| elif not empty and not nonempty: | |
| # only anonymous C arrays and memoryview slice arrays here. We | |
| # disallow buffer declarations for now, due to ambiguity with anonymous | |
| # C arrays. | |
| if is_memoryviewslice_access(s): | |
| return p_memoryviewslice_access(s, base_type_node) | |
| else: | |
| return base_type_node | |
| def is_memoryviewslice_access(s): | |
| # s.sy == '[' | |
| # a memoryview slice declaration is distinguishable from a buffer access | |
| # declaration by the first entry in the bracketed list. The buffer will | |
| # not have an unnested colon in the first entry; the memoryview slice will. | |
| saved = [(s.sy, s.systring)] | |
| s.next() | |
| retval = False | |
| if s.systring == ':': | |
| retval = True | |
| elif s.sy == 'INT': | |
| saved.append((s.sy, s.systring)) | |
| s.next() | |
| if s.sy == ':': | |
| retval = True | |
| for sv in saved[::-1]: | |
| s.put_back(*sv) | |
| return retval | |
| def p_memoryviewslice_access(s, base_type_node): | |
| # s.sy == '[' | |
| pos = s.position() | |
| s.next() | |
| subscripts, _ = p_subscript_list(s) | |
| # make sure each entry in subscripts is a slice | |
| for subscript in subscripts: | |
| if len(subscript) < 2: | |
| s.error("An axis specification in memoryview declaration does not have a ':'.") | |
| s.expect(']') | |
| indexes = make_slice_nodes(pos, subscripts) | |
| result = Nodes.MemoryViewSliceTypeNode(pos, | |
| base_type_node = base_type_node, | |
| axes = indexes) | |
| return result | |
| def looking_at_name(s): | |
| return s.sy == 'IDENT' and not s.systring in calling_convention_words | |
| def looking_at_expr(s): | |
| if s.systring in base_type_start_words: | |
| return False | |
| elif s.sy == 'IDENT': | |
| is_type = False | |
| name = s.systring | |
| dotted_path = [] | |
| s.next() | |
| while s.sy == '.': | |
| s.next() | |
| dotted_path.append(s.systring) | |
| s.expect('IDENT') | |
| saved = s.sy, s.systring | |
| if s.sy == 'IDENT': | |
| is_type = True | |
| elif s.sy == '*' or s.sy == '**': | |
| s.next() | |
| is_type = s.sy in (')', ']') | |
| s.put_back(*saved) | |
| elif s.sy == '(': | |
| s.next() | |
| is_type = s.sy == '*' | |
| s.put_back(*saved) | |
| elif s.sy == '[': | |
| s.next() | |
| is_type = s.sy == ']' or not looking_at_expr(s) # could be a nested template type | |
| s.put_back(*saved) | |
| dotted_path.reverse() | |
| for p in dotted_path: | |
| s.put_back('IDENT', p) | |
| s.put_back('.', '.') | |
| s.put_back('IDENT', name) | |
| return not is_type and saved[0] | |
| else: | |
| return True | |
| def looking_at_base_type(s): | |
| #print "looking_at_base_type?", s.sy, s.systring, s.position() | |
| return s.sy == 'IDENT' and s.systring in base_type_start_words | |
| def looking_at_dotted_name(s): | |
| if s.sy == 'IDENT': | |
| name = s.systring | |
| s.next() | |
| result = s.sy == '.' | |
| s.put_back('IDENT', name) | |
| return result | |
| else: | |
| return 0 | |
| def looking_at_call(s): | |
| "See if we're looking at a.b.c(" | |
| # Don't mess up the original position, so save and restore it. | |
| # Unfortunately there's no good way to handle this, as a subsequent call | |
| # to next() will not advance the position until it reads a new token. | |
| position = s.start_line, s.start_col | |
| result = looking_at_expr(s) == u'(' | |
| if not result: | |
| s.start_line, s.start_col = position | |
| return result | |
| basic_c_type_names = cython.declare( | |
| set, set(["void", "char", "int", "float", "double", "bint"])) | |
| special_basic_c_types = cython.declare(dict, { | |
| # name : (signed, longness) | |
| "Py_UNICODE" : (0, 0), | |
| "Py_UCS4" : (0, 0), | |
| "Py_hash_t" : (2, 0), | |
| "Py_ssize_t" : (2, 0), | |
| "ssize_t" : (2, 0), | |
| "size_t" : (0, 0), | |
| "ptrdiff_t" : (2, 0), | |
| "Py_tss_t" : (1, 0), | |
| }) | |
| sign_and_longness_words = cython.declare( | |
| set, set(["short", "long", "signed", "unsigned"])) | |
| base_type_start_words = cython.declare( | |
| set, | |
| basic_c_type_names | |
| | sign_and_longness_words | |
| | set(special_basic_c_types)) | |
| struct_enum_union = cython.declare( | |
| set, set(["struct", "union", "enum", "packed"])) | |
| def p_sign_and_longness(s): | |
| signed = 1 | |
| longness = 0 | |
| while s.sy == 'IDENT' and s.systring in sign_and_longness_words: | |
| if s.systring == 'unsigned': | |
| signed = 0 | |
| elif s.systring == 'signed': | |
| signed = 2 | |
| elif s.systring == 'short': | |
| longness = -1 | |
| elif s.systring == 'long': | |
| longness += 1 | |
| s.next() | |
| return signed, longness | |
| def p_opt_cname(s): | |
| literal = p_opt_string_literal(s, 'u') | |
| if literal is not None: | |
| cname = EncodedString(literal) | |
| cname.encoding = s.source_encoding | |
| else: | |
| cname = None | |
| return cname | |
| def p_c_declarator(s, ctx = Ctx(), empty = 0, is_type = 0, cmethod_flag = 0, | |
| assignable = 0, nonempty = 0, | |
| calling_convention_allowed = 0): | |
| # If empty is true, the declarator must be empty. If nonempty is true, | |
| # the declarator must be nonempty. Otherwise we don't care. | |
| # If cmethod_flag is true, then if this declarator declares | |
| # a function, it's a C method of an extension type. | |
| pos = s.position() | |
| if s.sy == '(': | |
| s.next() | |
| if s.sy == ')' or looking_at_name(s): | |
| base = Nodes.CNameDeclaratorNode(pos, name=s.context.intern_ustring(u""), cname=None) | |
| result = p_c_func_declarator(s, pos, ctx, base, cmethod_flag) | |
| else: | |
| result = p_c_declarator(s, ctx, empty = empty, is_type = is_type, | |
| cmethod_flag = cmethod_flag, | |
| nonempty = nonempty, | |
| calling_convention_allowed = 1) | |
| s.expect(')') | |
| else: | |
| result = p_c_simple_declarator(s, ctx, empty, is_type, cmethod_flag, | |
| assignable, nonempty) | |
| if not calling_convention_allowed and result.calling_convention and s.sy != '(': | |
| error(s.position(), "%s on something that is not a function" | |
| % result.calling_convention) | |
| while s.sy in ('[', '('): | |
| pos = s.position() | |
| if s.sy == '[': | |
| result = p_c_array_declarator(s, result) | |
| else: # sy == '(' | |
| s.next() | |
| result = p_c_func_declarator(s, pos, ctx, result, cmethod_flag) | |
| cmethod_flag = 0 | |
| return result | |
| def p_c_array_declarator(s, base): | |
| pos = s.position() | |
| s.next() # '[' | |
| if s.sy != ']': | |
| dim = p_testlist(s) | |
| else: | |
| dim = None | |
| s.expect(']') | |
| return Nodes.CArrayDeclaratorNode(pos, base = base, dimension = dim) | |
| def p_c_func_declarator(s, pos, ctx, base, cmethod_flag): | |
| # Opening paren has already been skipped | |
| args = p_c_arg_list(s, ctx, cmethod_flag = cmethod_flag, | |
| nonempty_declarators = 0) | |
| ellipsis = p_optional_ellipsis(s) | |
| s.expect(')') | |
| nogil = p_nogil(s) | |
| exc_val, exc_check = p_exception_value_clause(s) | |
| with_gil = p_with_gil(s) | |
| return Nodes.CFuncDeclaratorNode(pos, | |
| base = base, args = args, has_varargs = ellipsis, | |
| exception_value = exc_val, exception_check = exc_check, | |
| nogil = nogil or ctx.nogil or with_gil, with_gil = with_gil) | |
| supported_overloaded_operators = cython.declare(set, set([ | |
| '+', '-', '*', '/', '%', | |
| '++', '--', '~', '|', '&', '^', '<<', '>>', ',', | |
| '==', '!=', '>=', '>', '<=', '<', | |
| '[]', '()', '!', '=', | |
| 'bool', | |
| ])) | |
| def p_c_simple_declarator(s, ctx, empty, is_type, cmethod_flag, | |
| assignable, nonempty): | |
| pos = s.position() | |
| calling_convention = p_calling_convention(s) | |
| if s.sy == '*': | |
| s.next() | |
| if s.systring == 'const': | |
| const_pos = s.position() | |
| s.next() | |
| const_base = p_c_declarator(s, ctx, empty = empty, | |
| is_type = is_type, | |
| cmethod_flag = cmethod_flag, | |
| assignable = assignable, | |
| nonempty = nonempty) | |
| base = Nodes.CConstDeclaratorNode(const_pos, base = const_base) | |
| else: | |
| base = p_c_declarator(s, ctx, empty = empty, is_type = is_type, | |
| cmethod_flag = cmethod_flag, | |
| assignable = assignable, nonempty = nonempty) | |
| result = Nodes.CPtrDeclaratorNode(pos, | |
| base = base) | |
| elif s.sy == '**': # scanner returns this as a single token | |
| s.next() | |
| base = p_c_declarator(s, ctx, empty = empty, is_type = is_type, | |
| cmethod_flag = cmethod_flag, | |
| assignable = assignable, nonempty = nonempty) | |
| result = Nodes.CPtrDeclaratorNode(pos, | |
| base = Nodes.CPtrDeclaratorNode(pos, | |
| base = base)) | |
| elif s.sy == '&': | |
| s.next() | |
| base = p_c_declarator(s, ctx, empty = empty, is_type = is_type, | |
| cmethod_flag = cmethod_flag, | |
| assignable = assignable, nonempty = nonempty) | |
| result = Nodes.CReferenceDeclaratorNode(pos, base = base) | |
| else: | |
| rhs = None | |
| if s.sy == 'IDENT': | |
| name = s.systring | |
| if empty: | |
| error(s.position(), "Declarator should be empty") | |
| s.next() | |
| cname = p_opt_cname(s) | |
| if name != 'operator' and s.sy == '=' and assignable: | |
| s.next() | |
| rhs = p_test(s) | |
| else: | |
| if nonempty: | |
| error(s.position(), "Empty declarator") | |
| name = "" | |
| cname = None | |
| if cname is None and ctx.namespace is not None and nonempty: | |
| cname = ctx.namespace + "::" + name | |
| if name == 'operator' and ctx.visibility == 'extern' and nonempty: | |
| op = s.sy | |
| if [1 for c in op if c in '+-*/<=>!%&|([^~,']: | |
| s.next() | |
| # Handle diphthong operators. | |
| if op == '(': | |
| s.expect(')') | |
| op = '()' | |
| elif op == '[': | |
| s.expect(']') | |
| op = '[]' | |
| elif op in ('-', '+', '|', '&') and s.sy == op: | |
| op *= 2 # ++, --, ... | |
| s.next() | |
| elif s.sy == '=': | |
| op += s.sy # +=, -=, ... | |
| s.next() | |
| if op not in supported_overloaded_operators: | |
| s.error("Overloading operator '%s' not yet supported." % op, | |
| fatal=False) | |
| name += op | |
| elif op == 'IDENT': | |
| op = s.systring; | |
| if op not in supported_overloaded_operators: | |
| s.error("Overloading operator '%s' not yet supported." % op, | |
| fatal=False) | |
| name = name + ' ' + op | |
| s.next() | |
| result = Nodes.CNameDeclaratorNode(pos, | |
| name = name, cname = cname, default = rhs) | |
| result.calling_convention = calling_convention | |
| return result | |
| def p_nogil(s): | |
| if s.sy == 'IDENT' and s.systring == 'nogil': | |
| s.next() | |
| return 1 | |
| else: | |
| return 0 | |
| def p_with_gil(s): | |
| if s.sy == 'with': | |
| s.next() | |
| s.expect_keyword('gil') | |
| return 1 | |
| else: | |
| return 0 | |
| def p_exception_value_clause(s): | |
| exc_val = None | |
| exc_check = 0 | |
| if s.sy == 'except': | |
| s.next() | |
| if s.sy == '*': | |
| exc_check = 1 | |
| s.next() | |
| elif s.sy == '+': | |
| exc_check = '+' | |
| s.next() | |
| if s.sy == 'IDENT': | |
| name = s.systring | |
| s.next() | |
| exc_val = p_name(s, name) | |
| elif s.sy == '*': | |
| exc_val = ExprNodes.CharNode(s.position(), value=u'*') | |
| s.next() | |
| else: | |
| if s.sy == '?': | |
| exc_check = 1 | |
| s.next() | |
| exc_val = p_test(s) | |
| return exc_val, exc_check | |
| c_arg_list_terminators = cython.declare(set, set(['*', '**', '.', ')', ':'])) | |
| def p_c_arg_list(s, ctx = Ctx(), in_pyfunc = 0, cmethod_flag = 0, | |
| nonempty_declarators = 0, kw_only = 0, annotated = 1): | |
| # Comma-separated list of C argument declarations, possibly empty. | |
| # May have a trailing comma. | |
| args = [] | |
| is_self_arg = cmethod_flag | |
| while s.sy not in c_arg_list_terminators: | |
| args.append(p_c_arg_decl(s, ctx, in_pyfunc, is_self_arg, | |
| nonempty = nonempty_declarators, kw_only = kw_only, | |
| annotated = annotated)) | |
| if s.sy != ',': | |
| break | |
| s.next() | |
| is_self_arg = 0 | |
| return args | |
| def p_optional_ellipsis(s): | |
| if s.sy == '.': | |
| expect_ellipsis(s) | |
| return 1 | |
| else: | |
| return 0 | |
| def p_c_arg_decl(s, ctx, in_pyfunc, cmethod_flag = 0, nonempty = 0, | |
| kw_only = 0, annotated = 1): | |
| pos = s.position() | |
| not_none = or_none = 0 | |
| default = None | |
| annotation = None | |
| if s.in_python_file: | |
| # empty type declaration | |
| base_type = Nodes.CSimpleBaseTypeNode(pos, | |
| name = None, module_path = [], | |
| is_basic_c_type = 0, signed = 0, | |
| complex = 0, longness = 0, | |
| is_self_arg = cmethod_flag, templates = None) | |
| else: | |
| base_type = p_c_base_type(s, cmethod_flag, nonempty = nonempty) | |
| declarator = p_c_declarator(s, ctx, nonempty = nonempty) | |
| if s.sy in ('not', 'or') and not s.in_python_file: | |
| kind = s.sy | |
| s.next() | |
| if s.sy == 'IDENT' and s.systring == 'None': | |
| s.next() | |
| else: | |
| s.error("Expected 'None'") | |
| if not in_pyfunc: | |
| error(pos, "'%s None' only allowed in Python functions" % kind) | |
| or_none = kind == 'or' | |
| not_none = kind == 'not' | |
| if annotated and s.sy == ':': | |
| s.next() | |
| annotation = p_test(s) | |
| if s.sy == '=': | |
| s.next() | |
| if 'pxd' in ctx.level: | |
| if s.sy in ['*', '?']: | |
| # TODO(github/1736): Make this an error for inline declarations. | |
| default = ExprNodes.NoneNode(pos) | |
| s.next() | |
| elif 'inline' in ctx.modifiers: | |
| default = p_test(s) | |
| else: | |
| error(pos, "default values cannot be specified in pxd files, use ? or *") | |
| else: | |
| default = p_test(s) | |
| return Nodes.CArgDeclNode(pos, | |
| base_type = base_type, | |
| declarator = declarator, | |
| not_none = not_none, | |
| or_none = or_none, | |
| default = default, | |
| annotation = annotation, | |
| kw_only = kw_only) | |
| def p_api(s): | |
| if s.sy == 'IDENT' and s.systring == 'api': | |
| s.next() | |
| return 1 | |
| else: | |
| return 0 | |
| def p_cdef_statement(s, ctx): | |
| pos = s.position() | |
| ctx.visibility = p_visibility(s, ctx.visibility) | |
| ctx.api = ctx.api or p_api(s) | |
| if ctx.api: | |
| if ctx.visibility not in ('private', 'public'): | |
| error(pos, "Cannot combine 'api' with '%s'" % ctx.visibility) | |
| if (ctx.visibility == 'extern') and s.sy == 'from': | |
| return p_cdef_extern_block(s, pos, ctx) | |
| elif s.sy == 'import': | |
| s.next() | |
| return p_cdef_extern_block(s, pos, ctx) | |
| elif p_nogil(s): | |
| ctx.nogil = 1 | |
| if ctx.overridable: | |
| error(pos, "cdef blocks cannot be declared cpdef") | |
| return p_cdef_block(s, ctx) | |
| elif s.sy == ':': | |
| if ctx.overridable: | |
| error(pos, "cdef blocks cannot be declared cpdef") | |
| return p_cdef_block(s, ctx) | |
| elif s.sy == 'class': | |
| if ctx.level not in ('module', 'module_pxd'): | |
| error(pos, "Extension type definition not allowed here") | |
| if ctx.overridable: | |
| error(pos, "Extension types cannot be declared cpdef") | |
| return p_c_class_definition(s, pos, ctx) | |
| elif s.sy == 'IDENT' and s.systring == 'cppclass': | |
| return p_cpp_class_definition(s, pos, ctx) | |
| elif s.sy == 'IDENT' and s.systring in struct_enum_union: | |
| if ctx.level not in ('module', 'module_pxd'): | |
| error(pos, "C struct/union/enum definition not allowed here") | |
| if ctx.overridable: | |
| if s.systring != 'enum': | |
| error(pos, "C struct/union cannot be declared cpdef") | |
| return p_struct_enum(s, pos, ctx) | |
| elif s.sy == 'IDENT' and s.systring == 'fused': | |
| return p_fused_definition(s, pos, ctx) | |
| else: | |
| return p_c_func_or_var_declaration(s, pos, ctx) | |
| def p_cdef_block(s, ctx): | |
| return p_suite(s, ctx(cdef_flag = 1)) | |
| def p_cdef_extern_block(s, pos, ctx): | |
| if ctx.overridable: | |
| error(pos, "cdef extern blocks cannot be declared cpdef") | |
| include_file = None | |
| s.expect('from') | |
| if s.sy == '*': | |
| s.next() | |
| else: | |
| include_file = p_string_literal(s, 'u')[2] | |
| ctx = ctx(cdef_flag = 1, visibility = 'extern') | |
| if s.systring == "namespace": | |
| s.next() | |
| ctx.namespace = p_string_literal(s, 'u')[2] | |
| if p_nogil(s): | |
| ctx.nogil = 1 | |
| # Use "docstring" as verbatim string to include | |
| verbatim_include, body = p_suite_with_docstring(s, ctx, True) | |
| return Nodes.CDefExternNode(pos, | |
| include_file = include_file, | |
| verbatim_include = verbatim_include, | |
| body = body, | |
| namespace = ctx.namespace) | |
| def p_c_enum_definition(s, pos, ctx): | |
| # s.sy == ident 'enum' | |
| s.next() | |
| if s.sy == 'IDENT': | |
| name = s.systring | |
| s.next() | |
| cname = p_opt_cname(s) | |
| if cname is None and ctx.namespace is not None: | |
| cname = ctx.namespace + "::" + name | |
| else: | |
| name = None | |
| cname = None | |
| items = None | |
| s.expect(':') | |
| items = [] | |
| if s.sy != 'NEWLINE': | |
| p_c_enum_line(s, ctx, items) | |
| else: | |
| s.next() # 'NEWLINE' | |
| s.expect_indent() | |
| while s.sy not in ('DEDENT', 'EOF'): | |
| p_c_enum_line(s, ctx, items) | |
| s.expect_dedent() | |
| return Nodes.CEnumDefNode( | |
| pos, name = name, cname = cname, items = items, | |
| typedef_flag = ctx.typedef_flag, visibility = ctx.visibility, | |
| create_wrapper = ctx.overridable, | |
| api = ctx.api, in_pxd = ctx.level == 'module_pxd') | |
| def p_c_enum_line(s, ctx, items): | |
| if s.sy != 'pass': | |
| p_c_enum_item(s, ctx, items) | |
| while s.sy == ',': | |
| s.next() | |
| if s.sy in ('NEWLINE', 'EOF'): | |
| break | |
| p_c_enum_item(s, ctx, items) | |
| else: | |
| s.next() | |
| s.expect_newline("Syntax error in enum item list") | |
| def p_c_enum_item(s, ctx, items): | |
| pos = s.position() | |
| name = p_ident(s) | |
| cname = p_opt_cname(s) | |
| if cname is None and ctx.namespace is not None: | |
| cname = ctx.namespace + "::" + name | |
| value = None | |
| if s.sy == '=': | |
| s.next() | |
| value = p_test(s) | |
| items.append(Nodes.CEnumDefItemNode(pos, | |
| name = name, cname = cname, value = value)) | |
| def p_c_struct_or_union_definition(s, pos, ctx): | |
| packed = False | |
| if s.systring == 'packed': | |
| packed = True | |
| s.next() | |
| if s.sy != 'IDENT' or s.systring != 'struct': | |
| s.expected('struct') | |
| # s.sy == ident 'struct' or 'union' | |
| kind = s.systring | |
| s.next() | |
| name = p_ident(s) | |
| cname = p_opt_cname(s) | |
| if cname is None and ctx.namespace is not None: | |
| cname = ctx.namespace + "::" + name | |
| attributes = None | |
| if s.sy == ':': | |
| s.next() | |
| s.expect('NEWLINE') | |
| s.expect_indent() | |
| attributes = [] | |
| body_ctx = Ctx() | |
| while s.sy != 'DEDENT': | |
| if s.sy != 'pass': | |
| attributes.append( | |
| p_c_func_or_var_declaration(s, s.position(), body_ctx)) | |
| else: | |
| s.next() | |
| s.expect_newline("Expected a newline") | |
| s.expect_dedent() | |
| else: | |
| s.expect_newline("Syntax error in struct or union definition") | |
| return Nodes.CStructOrUnionDefNode(pos, | |
| name = name, cname = cname, kind = kind, attributes = attributes, | |
| typedef_flag = ctx.typedef_flag, visibility = ctx.visibility, | |
| api = ctx.api, in_pxd = ctx.level == 'module_pxd', packed = packed) | |
| def p_fused_definition(s, pos, ctx): | |
| """ | |
| c(type)def fused my_fused_type: | |
| ... | |
| """ | |
| # s.systring == 'fused' | |
| if ctx.level not in ('module', 'module_pxd'): | |
| error(pos, "Fused type definition not allowed here") | |
| s.next() | |
| name = p_ident(s) | |
| s.expect(":") | |
| s.expect_newline() | |
| s.expect_indent() | |
| types = [] | |
| while s.sy != 'DEDENT': | |
| if s.sy != 'pass': | |
| #types.append(p_c_declarator(s)) | |
| types.append(p_c_base_type(s)) #, nonempty=1)) | |
| else: | |
| s.next() | |
| s.expect_newline() | |
| s.expect_dedent() | |
| if not types: | |
| error(pos, "Need at least one type") | |
| return Nodes.FusedTypeNode(pos, name=name, types=types) | |
| def p_struct_enum(s, pos, ctx): | |
| if s.systring == 'enum': | |
| return p_c_enum_definition(s, pos, ctx) | |
| else: | |
| return p_c_struct_or_union_definition(s, pos, ctx) | |
| def p_visibility(s, prev_visibility): | |
| pos = s.position() | |
| visibility = prev_visibility | |
| if s.sy == 'IDENT' and s.systring in ('extern', 'public', 'readonly'): | |
| visibility = s.systring | |
| if prev_visibility != 'private' and visibility != prev_visibility: | |
| s.error("Conflicting visibility options '%s' and '%s'" | |
| % (prev_visibility, visibility), fatal=False) | |
| s.next() | |
| return visibility | |
| def p_c_modifiers(s): | |
| if s.sy == 'IDENT' and s.systring in ('inline',): | |
| modifier = s.systring | |
| s.next() | |
| return [modifier] + p_c_modifiers(s) | |
| return [] | |
| def p_c_func_or_var_declaration(s, pos, ctx): | |
| cmethod_flag = ctx.level in ('c_class', 'c_class_pxd') | |
| modifiers = p_c_modifiers(s) | |
| base_type = p_c_base_type(s, nonempty = 1, templates = ctx.templates) | |
| declarator = p_c_declarator(s, ctx(modifiers=modifiers), cmethod_flag = cmethod_flag, | |
| assignable = 1, nonempty = 1) | |
| declarator.overridable = ctx.overridable | |
| if s.sy == 'IDENT' and s.systring == 'const' and ctx.level == 'cpp_class': | |
| s.next() | |
| is_const_method = 1 | |
| else: | |
| is_const_method = 0 | |
| if s.sy == '->': | |
| # Special enough to give a better error message and keep going. | |
| s.error( | |
| "Return type annotation is not allowed in cdef/cpdef signatures. " | |
| "Please define it before the function name, as in C signatures.", | |
| fatal=False) | |
| s.next() | |
| p_test(s) # Keep going, but ignore result. | |
| if s.sy == ':': | |
| if ctx.level not in ('module', 'c_class', 'module_pxd', 'c_class_pxd', 'cpp_class') and not ctx.templates: | |
| s.error("C function definition not allowed here") | |
| doc, suite = p_suite_with_docstring(s, Ctx(level='function')) | |
| result = Nodes.CFuncDefNode(pos, | |
| visibility = ctx.visibility, | |
| base_type = base_type, | |
| declarator = declarator, | |
| body = suite, | |
| doc = doc, | |
| modifiers = modifiers, | |
| api = ctx.api, | |
| overridable = ctx.overridable, | |
| is_const_method = is_const_method) | |
| else: | |
| #if api: | |
| # s.error("'api' not allowed with variable declaration") | |
| if is_const_method: | |
| declarator.is_const_method = is_const_method | |
| declarators = [declarator] | |
| while s.sy == ',': | |
| s.next() | |
| if s.sy == 'NEWLINE': | |
| break | |
| declarator = p_c_declarator(s, ctx, cmethod_flag = cmethod_flag, | |
| assignable = 1, nonempty = 1) | |
| declarators.append(declarator) | |
| doc_line = s.start_line + 1 | |
| s.expect_newline("Syntax error in C variable declaration", ignore_semicolon=True) | |
| if ctx.level in ('c_class', 'c_class_pxd') and s.start_line == doc_line: | |
| doc = p_doc_string(s) | |
| else: | |
| doc = None | |
| result = Nodes.CVarDefNode(pos, | |
| visibility = ctx.visibility, | |
| base_type = base_type, | |
| declarators = declarators, | |
| in_pxd = ctx.level in ('module_pxd', 'c_class_pxd'), | |
| doc = doc, | |
| api = ctx.api, | |
| modifiers = modifiers, | |
| overridable = ctx.overridable) | |
| return result | |
| def p_ctypedef_statement(s, ctx): | |
| # s.sy == 'ctypedef' | |
| pos = s.position() | |
| s.next() | |
| visibility = p_visibility(s, ctx.visibility) | |
| api = p_api(s) | |
| ctx = ctx(typedef_flag = 1, visibility = visibility) | |
| if api: | |
| ctx.api = 1 | |
| if s.sy == 'class': | |
| return p_c_class_definition(s, pos, ctx) | |
| elif s.sy == 'IDENT' and s.systring in struct_enum_union: | |
| return p_struct_enum(s, pos, ctx) | |
| elif s.sy == 'IDENT' and s.systring == 'fused': | |
| return p_fused_definition(s, pos, ctx) | |
| else: | |
| base_type = p_c_base_type(s, nonempty = 1) | |
| declarator = p_c_declarator(s, ctx, is_type = 1, nonempty = 1) | |
| s.expect_newline("Syntax error in ctypedef statement", ignore_semicolon=True) | |
| return Nodes.CTypeDefNode( | |
| pos, base_type = base_type, | |
| declarator = declarator, | |
| visibility = visibility, api = api, | |
| in_pxd = ctx.level == 'module_pxd') | |
| def p_decorators(s): | |
| decorators = [] | |
| while s.sy == '@': | |
| pos = s.position() | |
| s.next() | |
| decstring = p_dotted_name(s, as_allowed=0)[2] | |
| names = decstring.split('.') | |
| decorator = ExprNodes.NameNode(pos, name=s.context.intern_ustring(names[0])) | |
| for name in names[1:]: | |
| decorator = ExprNodes.AttributeNode( | |
| pos, attribute=s.context.intern_ustring(name), obj=decorator) | |
| if s.sy == '(': | |
| decorator = p_call(s, decorator) | |
| decorators.append(Nodes.DecoratorNode(pos, decorator=decorator)) | |
| s.expect_newline("Expected a newline after decorator") | |
| return decorators | |
| def _reject_cdef_modifier_in_py(s, name): | |
| """Step over incorrectly placed cdef modifiers (@see _CDEF_MODIFIERS) to provide a good error message for them. | |
| """ | |
| if s.sy == 'IDENT' and name in _CDEF_MODIFIERS: | |
| # Special enough to provide a good error message. | |
| s.error("Cannot use cdef modifier '%s' in Python function signature. Use a decorator instead." % name, fatal=False) | |
| return p_ident(s) # Keep going, in case there are other errors. | |
| return name | |
| def p_def_statement(s, decorators=None, is_async_def=False): | |
| # s.sy == 'def' | |
| pos = s.position() | |
| # PEP 492 switches the async/await keywords on in "async def" functions | |
| if is_async_def: | |
| s.enter_async() | |
| s.next() | |
| name = _reject_cdef_modifier_in_py(s, p_ident(s)) | |
| s.expect( | |
| '(', | |
| "Expected '(', found '%s'. Did you use cdef syntax in a Python declaration? " | |
| "Use decorators and Python type annotations instead." % ( | |
| s.systring if s.sy == 'IDENT' else s.sy)) | |
| args, star_arg, starstar_arg = p_varargslist(s, terminator=')') | |
| s.expect(')') | |
| _reject_cdef_modifier_in_py(s, s.systring) | |
| return_type_annotation = None | |
| if s.sy == '->': | |
| s.next() | |
| return_type_annotation = p_test(s) | |
| _reject_cdef_modifier_in_py(s, s.systring) | |
| doc, body = p_suite_with_docstring(s, Ctx(level='function')) | |
| if is_async_def: | |
| s.exit_async() | |
| return Nodes.DefNode( | |
| pos, name=name, args=args, star_arg=star_arg, starstar_arg=starstar_arg, | |
| doc=doc, body=body, decorators=decorators, is_async_def=is_async_def, | |
| return_type_annotation=return_type_annotation) | |
| def p_varargslist(s, terminator=')', annotated=1): | |
| args = p_c_arg_list(s, in_pyfunc = 1, nonempty_declarators = 1, | |
| annotated = annotated) | |
| star_arg = None | |
| starstar_arg = None | |
| if s.sy == '*': | |
| s.next() | |
| if s.sy == 'IDENT': | |
| star_arg = p_py_arg_decl(s, annotated=annotated) | |
| if s.sy == ',': | |
| s.next() | |
| args.extend(p_c_arg_list(s, in_pyfunc = 1, | |
| nonempty_declarators = 1, kw_only = 1, annotated = annotated)) | |
| elif s.sy != terminator: | |
| s.error("Syntax error in Python function argument list") | |
| if s.sy == '**': | |
| s.next() | |
| starstar_arg = p_py_arg_decl(s, annotated=annotated) | |
| if s.sy == ',': | |
| s.next() | |
| return (args, star_arg, starstar_arg) | |
| def p_py_arg_decl(s, annotated = 1): | |
| pos = s.position() | |
| name = p_ident(s) | |
| annotation = None | |
| if annotated and s.sy == ':': | |
| s.next() | |
| annotation = p_test(s) | |
| return Nodes.PyArgDeclNode(pos, name = name, annotation = annotation) | |
| def p_class_statement(s, decorators): | |
| # s.sy == 'class' | |
| pos = s.position() | |
| s.next() | |
| class_name = EncodedString(p_ident(s)) | |
| class_name.encoding = s.source_encoding # FIXME: why is this needed? | |
| arg_tuple = None | |
| keyword_dict = None | |
| if s.sy == '(': | |
| positional_args, keyword_args = p_call_parse_args(s, allow_genexp=False) | |
| arg_tuple, keyword_dict = p_call_build_packed_args(pos, positional_args, keyword_args) | |
| if arg_tuple is None: | |
| # XXX: empty arg_tuple | |
| arg_tuple = ExprNodes.TupleNode(pos, args=[]) | |
| doc, body = p_suite_with_docstring(s, Ctx(level='class')) | |
| return Nodes.PyClassDefNode( | |
| pos, name=class_name, | |
| bases=arg_tuple, | |
| keyword_args=keyword_dict, | |
| doc=doc, body=body, decorators=decorators, | |
| force_py3_semantics=s.context.language_level >= 3) | |
| def p_c_class_definition(s, pos, ctx): | |
| # s.sy == 'class' | |
| s.next() | |
| module_path = [] | |
| class_name = p_ident(s) | |
| while s.sy == '.': | |
| s.next() | |
| module_path.append(class_name) | |
| class_name = p_ident(s) | |
| if module_path and ctx.visibility != 'extern': | |
| error(pos, "Qualified class name only allowed for 'extern' C class") | |
| if module_path and s.sy == 'IDENT' and s.systring == 'as': | |
| s.next() | |
| as_name = p_ident(s) | |
| else: | |
| as_name = class_name | |
| objstruct_name = None | |
| typeobj_name = None | |
| bases = None | |
| check_size = None | |
| if s.sy == '(': | |
| positional_args, keyword_args = p_call_parse_args(s, allow_genexp=False) | |
| if keyword_args: | |
| s.error("C classes cannot take keyword bases.") | |
| bases, _ = p_call_build_packed_args(pos, positional_args, keyword_args) | |
| if bases is None: | |
| bases = ExprNodes.TupleNode(pos, args=[]) | |
| if s.sy == '[': | |
| if ctx.visibility not in ('public', 'extern') and not ctx.api: | |
| error(s.position(), "Name options only allowed for 'public', 'api', or 'extern' C class") | |
| objstruct_name, typeobj_name, check_size = p_c_class_options(s) | |
| if s.sy == ':': | |
| if ctx.level == 'module_pxd': | |
| body_level = 'c_class_pxd' | |
| else: | |
| body_level = 'c_class' | |
| doc, body = p_suite_with_docstring(s, Ctx(level=body_level)) | |
| else: | |
| s.expect_newline("Syntax error in C class definition") | |
| doc = None | |
| body = None | |
| if ctx.visibility == 'extern': | |
| if not module_path: | |
| error(pos, "Module name required for 'extern' C class") | |
| if typeobj_name: | |
| error(pos, "Type object name specification not allowed for 'extern' C class") | |
| elif ctx.visibility == 'public': | |
| if not objstruct_name: | |
| error(pos, "Object struct name specification required for 'public' C class") | |
| if not typeobj_name: | |
| error(pos, "Type object name specification required for 'public' C class") | |
| elif ctx.visibility == 'private': | |
| if ctx.api: | |
| if not objstruct_name: | |
| error(pos, "Object struct name specification required for 'api' C class") | |
| if not typeobj_name: | |
| error(pos, "Type object name specification required for 'api' C class") | |
| else: | |
| error(pos, "Invalid class visibility '%s'" % ctx.visibility) | |
| return Nodes.CClassDefNode(pos, | |
| visibility = ctx.visibility, | |
| typedef_flag = ctx.typedef_flag, | |
| api = ctx.api, | |
| module_name = ".".join(module_path), | |
| class_name = class_name, | |
| as_name = as_name, | |
| bases = bases, | |
| objstruct_name = objstruct_name, | |
| typeobj_name = typeobj_name, | |
| check_size = check_size, | |
| in_pxd = ctx.level == 'module_pxd', | |
| doc = doc, | |
| body = body) | |
| def p_c_class_options(s): | |
| objstruct_name = None | |
| typeobj_name = None | |
| check_size = None | |
| s.expect('[') | |
| while 1: | |
| if s.sy != 'IDENT': | |
| break | |
| if s.systring == 'object': | |
| s.next() | |
| objstruct_name = p_ident(s) | |
| elif s.systring == 'type': | |
| s.next() | |
| typeobj_name = p_ident(s) | |
| elif s.systring == 'check_size': | |
| s.next() | |
| check_size = p_ident(s) | |
| if check_size not in ('ignore', 'warn', 'error'): | |
| s.error("Expected one of ignore, warn or error, found %r" % check_size) | |
| if s.sy != ',': | |
| break | |
| s.next() | |
| s.expect(']', "Expected 'object', 'type' or 'check_size'") | |
| return objstruct_name, typeobj_name, check_size | |
| def p_property_decl(s): | |
| pos = s.position() | |
| s.next() # 'property' | |
| name = p_ident(s) | |
| doc, body = p_suite_with_docstring( | |
| s, Ctx(level='property'), with_doc_only=True) | |
| return Nodes.PropertyNode(pos, name=name, doc=doc, body=body) | |
| def p_ignorable_statement(s): | |
| """ | |
| Parses any kind of ignorable statement that is allowed in .pxd files. | |
| """ | |
| if s.sy == 'BEGIN_STRING': | |
| pos = s.position() | |
| string_node = p_atom(s) | |
| s.expect_newline("Syntax error in string", ignore_semicolon=True) | |
| return Nodes.ExprStatNode(pos, expr=string_node) | |
| return None | |
| def p_doc_string(s): | |
| if s.sy == 'BEGIN_STRING': | |
| pos = s.position() | |
| kind, bytes_result, unicode_result = p_cat_string_literal(s) | |
| s.expect_newline("Syntax error in doc string", ignore_semicolon=True) | |
| if kind in ('u', ''): | |
| return unicode_result | |
| warning(pos, "Python 3 requires docstrings to be unicode strings") | |
| return bytes_result | |
| else: | |
| return None | |
| def _extract_docstring(node): | |
| """ | |
| Extract a docstring from a statement or from the first statement | |
| in a list. Remove the statement if found. Return a tuple | |
| (plain-docstring or None, node). | |
| """ | |
| doc_node = None | |
| if node is None: | |
| pass | |
| elif isinstance(node, Nodes.ExprStatNode): | |
| if node.expr.is_string_literal: | |
| doc_node = node.expr | |
| node = Nodes.StatListNode(node.pos, stats=[]) | |
| elif isinstance(node, Nodes.StatListNode) and node.stats: | |
| stats = node.stats | |
| if isinstance(stats[0], Nodes.ExprStatNode): | |
| if stats[0].expr.is_string_literal: | |
| doc_node = stats[0].expr | |
| del stats[0] | |
| if doc_node is None: | |
| doc = None | |
| elif isinstance(doc_node, ExprNodes.BytesNode): | |
| warning(node.pos, | |
| "Python 3 requires docstrings to be unicode strings") | |
| doc = doc_node.value | |
| elif isinstance(doc_node, ExprNodes.StringNode): | |
| doc = doc_node.unicode_value | |
| if doc is None: | |
| doc = doc_node.value | |
| else: | |
| doc = doc_node.value | |
| return doc, node | |
| def p_code(s, level=None, ctx=Ctx): | |
| body = p_statement_list(s, ctx(level = level), first_statement = 1) | |
| if s.sy != 'EOF': | |
| s.error("Syntax error in statement [%s,%s]" % ( | |
| repr(s.sy), repr(s.systring))) | |
| return body | |
| _match_compiler_directive_comment = cython.declare(object, re.compile( | |
| r"^#\s*cython\s*:\s*((\w|[.])+\s*=.*)$").match) | |
| def p_compiler_directive_comments(s): | |
| result = {} | |
| while s.sy == 'commentline': | |
| pos = s.position() | |
| m = _match_compiler_directive_comment(s.systring) | |
| if m: | |
| directives_string = m.group(1).strip() | |
| try: | |
| new_directives = Options.parse_directive_list(directives_string, ignore_unknown=True) | |
| except ValueError as e: | |
| s.error(e.args[0], fatal=False) | |
| s.next() | |
| continue | |
| for name in new_directives: | |
| if name not in result: | |
| pass | |
| elif new_directives[name] == result[name]: | |
| warning(pos, "Duplicate directive found: %s" % (name,)) | |
| else: | |
| s.error("Conflicting settings found for top-level directive %s: %r and %r" % ( | |
| name, result[name], new_directives[name]), pos=pos) | |
| if 'language_level' in new_directives: | |
| # Make sure we apply the language level already to the first token that follows the comments. | |
| s.context.set_language_level(new_directives['language_level']) | |
| result.update(new_directives) | |
| s.next() | |
| return result | |
| def p_module(s, pxd, full_module_name, ctx=Ctx): | |
| pos = s.position() | |
| directive_comments = p_compiler_directive_comments(s) | |
| s.parse_comments = False | |
| if s.context.language_level is None: | |
| s.context.set_language_level(2) | |
| if pos[0].filename: | |
| import warnings | |
| warnings.warn( | |
| "Cython directive 'language_level' not set, using 2 for now (Py2). " | |
| "This will change in a later release! File: %s" % pos[0].filename, | |
| FutureWarning, | |
| stacklevel=1 if cython.compiled else 2, | |
| ) | |
| doc = p_doc_string(s) | |
| if pxd: | |
| level = 'module_pxd' | |
| else: | |
| level = 'module' | |
| body = p_statement_list(s, ctx(level=level), first_statement = 1) | |
| if s.sy != 'EOF': | |
| s.error("Syntax error in statement [%s,%s]" % ( | |
| repr(s.sy), repr(s.systring))) | |
| return ModuleNode(pos, doc = doc, body = body, | |
| full_module_name = full_module_name, | |
| directive_comments = directive_comments) | |
| def p_template_definition(s): | |
| name = p_ident(s) | |
| if s.sy == '=': | |
| s.expect('=') | |
| s.expect('*') | |
| required = False | |
| else: | |
| required = True | |
| return name, required | |
| def p_cpp_class_definition(s, pos, ctx): | |
| # s.sy == 'cppclass' | |
| s.next() | |
| module_path = [] | |
| class_name = p_ident(s) | |
| cname = p_opt_cname(s) | |
| if cname is None and ctx.namespace is not None: | |
| cname = ctx.namespace + "::" + class_name | |
| if s.sy == '.': | |
| error(pos, "Qualified class name not allowed C++ class") | |
| if s.sy == '[': | |
| s.next() | |
| templates = [p_template_definition(s)] | |
| while s.sy == ',': | |
| s.next() | |
| templates.append(p_template_definition(s)) | |
| s.expect(']') | |
| template_names = [name for name, required in templates] | |
| else: | |
| templates = None | |
| template_names = None | |
| if s.sy == '(': | |
| s.next() | |
| base_classes = [p_c_base_type(s, templates = template_names)] | |
| while s.sy == ',': | |
| s.next() | |
| base_classes.append(p_c_base_type(s, templates = template_names)) | |
| s.expect(')') | |
| else: | |
| base_classes = [] | |
| if s.sy == '[': | |
| error(s.position(), "Name options not allowed for C++ class") | |
| nogil = p_nogil(s) | |
| if s.sy == ':': | |
| s.next() | |
| s.expect('NEWLINE') | |
| s.expect_indent() | |
| attributes = [] | |
| body_ctx = Ctx(visibility = ctx.visibility, level='cpp_class', nogil=nogil or ctx.nogil) | |
| body_ctx.templates = template_names | |
| while s.sy != 'DEDENT': | |
| if s.sy != 'pass': | |
| attributes.append(p_cpp_class_attribute(s, body_ctx)) | |
| else: | |
| s.next() | |
| s.expect_newline("Expected a newline") | |
| s.expect_dedent() | |
| else: | |
| attributes = None | |
| s.expect_newline("Syntax error in C++ class definition") | |
| return Nodes.CppClassNode(pos, | |
| name = class_name, | |
| cname = cname, | |
| base_classes = base_classes, | |
| visibility = ctx.visibility, | |
| in_pxd = ctx.level == 'module_pxd', | |
| attributes = attributes, | |
| templates = templates) | |
| def p_cpp_class_attribute(s, ctx): | |
| decorators = None | |
| if s.sy == '@': | |
| decorators = p_decorators(s) | |
| if s.systring == 'cppclass': | |
| return p_cpp_class_definition(s, s.position(), ctx) | |
| elif s.systring == 'ctypedef': | |
| return p_ctypedef_statement(s, ctx) | |
| elif s.sy == 'IDENT' and s.systring in struct_enum_union: | |
| if s.systring != 'enum': | |
| return p_cpp_class_definition(s, s.position(), ctx) | |
| else: | |
| return p_struct_enum(s, s.position(), ctx) | |
| else: | |
| node = p_c_func_or_var_declaration(s, s.position(), ctx) | |
| if decorators is not None: | |
| tup = Nodes.CFuncDefNode, Nodes.CVarDefNode, Nodes.CClassDefNode | |
| if ctx.allow_struct_enum_decorator: | |
| tup += Nodes.CStructOrUnionDefNode, Nodes.CEnumDefNode | |
| if not isinstance(node, tup): | |
| s.error("Decorators can only be followed by functions or classes") | |
| node.decorators = decorators | |
| return node | |
| #---------------------------------------------- | |
| # | |
| # Debugging | |
| # | |
| #---------------------------------------------- | |
| def print_parse_tree(f, node, level, key = None): | |
| ind = " " * level | |
| if node: | |
| f.write(ind) | |
| if key: | |
| f.write("%s: " % key) | |
| t = type(node) | |
| if t is tuple: | |
| f.write("(%s @ %s\n" % (node[0], node[1])) | |
| for i in range(2, len(node)): | |
| print_parse_tree(f, node[i], level+1) | |
| f.write("%s)\n" % ind) | |
| return | |
| elif isinstance(node, Nodes.Node): | |
| try: | |
| tag = node.tag | |
| except AttributeError: | |
| tag = node.__class__.__name__ | |
| f.write("%s @ %s\n" % (tag, node.pos)) | |
| for name, value in node.__dict__.items(): | |
| if name != 'tag' and name != 'pos': | |
| print_parse_tree(f, value, level+1, name) | |
| return | |
| elif t is list: | |
| f.write("[\n") | |
| for i in range(len(node)): | |
| print_parse_tree(f, node[i], level+1) | |
| f.write("%s]\n" % ind) | |
| return | |
| f.write("%s%s\n" % (ind, node)) | |