|
|
# iterparse -- event-driven parsing |
|
|
|
|
|
DEF __ITERPARSE_CHUNK_SIZE = 32768 |
|
|
|
|
|
cdef class iterparse: |
|
|
"""iterparse(self, source, events=("end |
|
|
attribute_defaults=False, dtd_validation=False, \ |
|
|
load_dtd=False, no_network=True, remove_blank_text=False, \ |
|
|
remove_comments=False, remove_pis=False, encoding=None, \ |
|
|
html=False, recover=None, huge_tree=False, schema=None) |
|
|
|
|
|
Incremental parser. |
|
|
|
|
|
Parses XML into a tree and generates tuples (event, element) in a |
|
|
SAX-like fashion. ``event`` is any of 'start', 'end', 'start-ns', |
|
|
'end-ns'. |
|
|
|
|
|
For 'start' and 'end', ``element`` is the Element that the parser just |
|
|
found opening or closing. For 'start-ns', it is a tuple (prefix, URI) of |
|
|
a new namespace declaration. For 'end-ns', it is simply None. Note that |
|
|
all start and end events are guaranteed to be properly nested. |
|
|
|
|
|
The keyword argument ``events`` specifies a sequence of event type names |
|
|
that should be generated. By default, only 'end' events will be |
|
|
generated. |
|
|
|
|
|
The additional ``tag`` argument restricts the 'start' and 'end' events to |
|
|
those elements that match the given tag. The ``tag`` argument can also be |
|
|
a sequence of tags to allow matching more than one tag. By default, |
|
|
events are generated for all elements. Note that the 'start-ns' and |
|
|
'end-ns' events are not impacted by this restriction. |
|
|
|
|
|
The other keyword arguments in the constructor are mainly based on the |
|
|
libxml2 parser configuration. A DTD will also be loaded if validation or |
|
|
attribute default values are requested. |
|
|
|
|
|
Available boolean keyword arguments: |
|
|
- attribute_defaults: read default attributes from DTD |
|
|
- dtd_validation: validate (if DTD is available) |
|
|
- load_dtd: use DTD for parsing |
|
|
- no_network: prevent network access for related files |
|
|
- remove_blank_text: discard blank text nodes |
|
|
- remove_comments: discard comments |
|
|
- remove_pis: discard processing instructions |
|
|
- strip_cdata: replace CDATA sections by normal text content (default: True) |
|
|
- compact: safe memory for short text content (default: True) |
|
|
- resolve_entities: replace entities by their text value (default: True) |
|
|
- huge_tree: disable security restrictions and support very deep trees |
|
|
and very long text content (only affects libxml2 2.7+) |
|
|
- html: parse input as HTML (default: XML) |
|
|
- recover: try hard to parse through broken input (default: True for HTML, |
|
|
False otherwise) |
|
|
|
|
|
Other keyword arguments: |
|
|
- encoding: override the document encoding |
|
|
- schema: an XMLSchema to validate against |
|
|
"" |
|
|
cdef _FeedParser _parser |
|
|
cdef object _tag |
|
|
cdef object _events |
|
|
cdef readonly object root |
|
|
cdef object _source |
|
|
cdef object _filename |
|
|
cdef object _error |
|
|
cdef bint _close_source_after_read |
|
|
|
|
|
def __init__(self, source, events=("end",), *, tag=None, |
|
|
attribute_defaults=False, dtd_validation=False, |
|
|
load_dtd=False, no_network=True, remove_blank_text=False, |
|
|
compact=True, resolve_entities=True, remove_comments=False, |
|
|
remove_pis=False, strip_cdata=True, encoding=None, |
|
|
html=False, recover=None, huge_tree=False, collect_ids=True, |
|
|
XMLSchema schema=None): |
|
|
if not hasattr(source, 'read'): |
|
|
source = _getFSPathOrObject(source) |
|
|
self._filename = source |
|
|
self._source = open(source, 'rb') |
|
|
self._close_source_after_read = True |
|
|
else: |
|
|
self._filename = _getFilenameForFile(source) |
|
|
self._source = source |
|
|
self._close_source_after_read = False |
|
|
|
|
|
if recover is None: |
|
|
recover = html |
|
|
|
|
|
if html: |
|
|
# make sure we're not looking for namespaces |
|
|
events = [event for event in events |
|
|
if event not in ('start-ns', 'end-ns')] |
|
|
parser = HTMLPullParser( |
|
|
events, |
|
|
tag=tag, |
|
|
recover=recover, |
|
|
base_url=self._filename, |
|
|
encoding=encoding, |
|
|
remove_blank_text=remove_blank_text, |
|
|
remove_comments=remove_comments, |
|
|
remove_pis=remove_pis, |
|
|
strip_cdata=strip_cdata, |
|
|
no_network=no_network, |
|
|
target=None, # TODO |
|
|
schema=schema, |
|
|
compact=compact) |
|
|
else: |
|
|
parser = XMLPullParser( |
|
|
events, |
|
|
tag=tag, |
|
|
recover=recover, |
|
|
base_url=self._filename, |
|
|
encoding=encoding, |
|
|
attribute_defaults=attribute_defaults, |
|
|
dtd_validation=dtd_validation, |
|
|
load_dtd=load_dtd, |
|
|
no_network=no_network, |
|
|
schema=schema, |
|
|
huge_tree=huge_tree, |
|
|
remove_blank_text=remove_blank_text, |
|
|
resolve_entities=resolve_entities, |
|
|
remove_comments=remove_comments, |
|
|
remove_pis=remove_pis, |
|
|
strip_cdata=strip_cdata, |
|
|
collect_ids=True, |
|
|
target=None, # TODO |
|
|
compact=compact) |
|
|
|
|
|
self._events = parser.read_events() |
|
|
self._parser = parser |
|
|
|
|
|
@property |
|
|
def error_log(self): |
|
|
"""The error log of the last (or current) parser run. |
|
|
""" |
|
|
return self._parser.feed_error_log |
|
|
|
|
|
@property |
|
|
def resolvers(self): |
|
|
"""The custom resolver registry of the last (or current) parser run. |
|
|
""" |
|
|
return self._parser.resolvers |
|
|
|
|
|
@property |
|
|
def version(self): |
|
|
"""The version of the underlying XML parser.""" |
|
|
return self._parser.version |
|
|
|
|
|
def set_element_class_lookup(self, ElementClassLookup lookup = None): |
|
|
"""set_element_class_lookup(self, lookup = None) |
|
|
|
|
|
Set a lookup scheme for element classes generated from this parser. |
|
|
|
|
|
Reset it by passing None or nothing. |
|
|
""" |
|
|
self._parser.set_element_class_lookup(lookup) |
|
|
|
|
|
def makeelement(self, _tag, attrib=None, nsmap=None, **_extra): |
|
|
"""makeelement(self, _tag, attrib=None, nsmap=None, **_extra) |
|
|
|
|
|
Creates a new element associated with this parser. |
|
|
""" |
|
|
self._parser.makeelement( |
|
|
_tag, attrib=None, nsmap=None, **_extra) |
|
|
|
|
|
@cython.final |
|
|
cdef _close_source(self): |
|
|
if self._source is None: |
|
|
return |
|
|
if not self._close_source_after_read: |
|
|
self._source = None |
|
|
return |
|
|
try: |
|
|
close = self._source.close |
|
|
except AttributeError: |
|
|
close = None |
|
|
finally: |
|
|
self._source = None |
|
|
if close is not None: |
|
|
close() |
|
|
|
|
|
def __iter__(self): |
|
|
return self |
|
|
|
|
|
def __next__(self): |
|
|
try: |
|
|
return next(self._events) |
|
|
except StopIteration: |
|
|
pass |
|
|
context = <_SaxParserContext>self._parser._getPushParserContext() |
|
|
if self._source is not None: |
|
|
done = False |
|
|
while not done: |
|
|
try: |
|
|
done = self._read_more_events(context) |
|
|
return next(self._events) |
|
|
except StopIteration: |
|
|
pass # no events yet |
|
|
except Exception as e: |
|
|
self._error = e |
|
|
self._close_source() |
|
|
try: |
|
|
return next(self._events) |
|
|
except StopIteration: |
|
|
break |
|
|
# nothing left to read or return |
|
|
if self._error is not None: |
|
|
error = self._error |
|
|
self._error = None |
|
|
raise error |
|
|
if (context._validator is not None |
|
|
and not context._validator.isvalid()): |
|
|
_raiseParseError(context._c_ctxt, self._filename, |
|
|
context._error_log) |
|
|
# no errors => all done |
|
|
raise StopIteration |
|
|
|
|
|
@cython.final |
|
|
cdef bint _read_more_events(self, _SaxParserContext context) except -123: |
|
|
data = self._source.read(__ITERPARSE_CHUNK_SIZE) |
|
|
if not isinstance(data, bytes): |
|
|
self._close_source() |
|
|
raise TypeError("reading file objects must return bytes objects") |
|
|
if not data: |
|
|
try: |
|
|
self.root = self._parser.close() |
|
|
finally: |
|
|
self._close_source() |
|
|
return True |
|
|
self._parser.feed(data) |
|
|
return False |
|
|
|
|
|
|
|
|
cdef enum _IterwalkSkipStates: |
|
|
IWSKIP_NEXT_IS_START |
|
|
IWSKIP_SKIP_NEXT |
|
|
IWSKIP_CAN_SKIP |
|
|
IWSKIP_CANNOT_SKIP |
|
|
|
|
|
|
|
|
cdef class iterwalk: |
|
|
"""iterwalk(self, element_or_tree, events=("end",), tag=None) |
|
|
|
|
|
A tree walker that generates events from an existing tree as if it |
|
|
was parsing XML data with ``iterparse()``. |
|
|
|
|
|
Just as for ``iterparse()``, the ``tag`` argument can be a single tag or a |
|
|
sequence of tags. |
|
|
|
|
|
After receiving a 'start' or 'start-ns' event, the children and |
|
|
descendants of the current element can be excluded from iteration |
|
|
by calling the ``skip_subtree()`` method. |
|
|
""" |
|
|
cdef _MultiTagMatcher _matcher |
|
|
cdef list _node_stack |
|
|
cdef list _events |
|
|
cdef object _pop_event |
|
|
cdef object _include_siblings |
|
|
cdef int _index |
|
|
cdef int _event_filter |
|
|
cdef _IterwalkSkipStates _skip_state |
|
|
|
|
|
def __init__(self, element_or_tree, events=("end",), tag=None): |
|
|
cdef _Element root |
|
|
cdef int ns_count |
|
|
root = _rootNodeOrRaise(element_or_tree) |
|
|
self._event_filter = _buildParseEventFilter(events) |
|
|
if tag is None or tag == '*': |
|
|
self._matcher = None |
|
|
else: |
|
|
self._matcher = _MultiTagMatcher.__new__(_MultiTagMatcher, tag) |
|
|
self._node_stack = [] |
|
|
self._events = [] |
|
|
self._pop_event = self._events.pop |
|
|
self._skip_state = IWSKIP_CANNOT_SKIP # ignore all skip requests by default |
|
|
|
|
|
if self._event_filter: |
|
|
self._index = 0 |
|
|
if self._matcher is not None and self._event_filter & PARSE_EVENT_FILTER_START: |
|
|
self._matcher.cacheTags(root._doc) |
|
|
|
|
|
# When processing an ElementTree, add events for the preceding comments/PIs. |
|
|
if self._event_filter & (PARSE_EVENT_FILTER_COMMENT | PARSE_EVENT_FILTER_PI): |
|
|
if isinstance(element_or_tree, _ElementTree): |
|
|
self._include_siblings = root |
|
|
for elem in list(root.itersiblings(preceding=True))[::-1]: |
|
|
if self._event_filter & PARSE_EVENT_FILTER_COMMENT and elem.tag is Comment: |
|
|
self._events.append(('comment', elem)) |
|
|
elif self._event_filter & PARSE_EVENT_FILTER_PI and elem.tag is PI: |
|
|
self._events.append(('pi', elem)) |
|
|
|
|
|
ns_count = self._start_node(root) |
|
|
self._node_stack.append( (root, ns_count) ) |
|
|
else: |
|
|
self._index = -1 |
|
|
|
|
|
def __iter__(self): |
|
|
return self |
|
|
|
|
|
def __next__(self): |
|
|
cdef xmlNode* c_child |
|
|
cdef _Element node |
|
|
cdef _Element next_node |
|
|
cdef int ns_count = 0 |
|
|
if self._events: |
|
|
return self._next_event() |
|
|
if self._matcher is not None and self._index >= 0: |
|
|
node = self._node_stack[self._index][0] |
|
|
self._matcher.cacheTags(node._doc) |
|
|
|
|
|
# find next node |
|
|
while self._index >= 0: |
|
|
node = self._node_stack[self._index][0] |
|
|
|
|
|
if self._skip_state == IWSKIP_SKIP_NEXT: |
|
|
c_child = NULL |
|
|
else: |
|
|
c_child = self._process_non_elements( |
|
|
node._doc, _findChildForwards(node._c_node, 0)) |
|
|
self._skip_state = IWSKIP_CANNOT_SKIP |
|
|
|
|
|
while c_child is NULL: |
|
|
# back off through parents |
|
|
self._index -= 1 |
|
|
node = self._end_node() |
|
|
if self._index < 0: |
|
|
break |
|
|
c_child = self._process_non_elements( |
|
|
node._doc, _nextElement(node._c_node)) |
|
|
|
|
|
if c_child is not NULL: |
|
|
next_node = _elementFactory(node._doc, c_child) |
|
|
if self._event_filter & (PARSE_EVENT_FILTER_START | |
|
|
PARSE_EVENT_FILTER_START_NS): |
|
|
ns_count = self._start_node(next_node) |
|
|
elif self._event_filter & PARSE_EVENT_FILTER_END_NS: |
|
|
ns_count = _countNsDefs(next_node._c_node) |
|
|
self._node_stack.append( (next_node, ns_count) ) |
|
|
self._index += 1 |
|
|
if self._events: |
|
|
return self._next_event() |
|
|
|
|
|
if self._include_siblings is not None: |
|
|
node, self._include_siblings = self._include_siblings, None |
|
|
self._process_non_elements(node._doc, _nextElement(node._c_node)) |
|
|
if self._events: |
|
|
return self._next_event() |
|
|
|
|
|
raise StopIteration |
|
|
|
|
|
@cython.final |
|
|
cdef xmlNode* _process_non_elements(self, _Document doc, xmlNode* c_node): |
|
|
while c_node is not NULL and c_node.type != tree.XML_ELEMENT_NODE: |
|
|
if c_node.type == tree.XML_COMMENT_NODE: |
|
|
if self._event_filter & PARSE_EVENT_FILTER_COMMENT: |
|
|
self._events.append( |
|
|
("comment", _elementFactory(doc, c_node))) |
|
|
c_node = _nextElement(c_node) |
|
|
elif c_node.type == tree.XML_PI_NODE: |
|
|
if self._event_filter & PARSE_EVENT_FILTER_PI: |
|
|
self._events.append( |
|
|
("pi", _elementFactory(doc, c_node))) |
|
|
c_node = _nextElement(c_node) |
|
|
else: |
|
|
break |
|
|
return c_node |
|
|
|
|
|
@cython.final |
|
|
cdef _next_event(self): |
|
|
if self._skip_state == IWSKIP_NEXT_IS_START: |
|
|
if self._events[0][0] in ('start', 'start-ns'): |
|
|
self._skip_state = IWSKIP_CAN_SKIP |
|
|
return self._pop_event(0) |
|
|
|
|
|
def skip_subtree(self): |
|
|
"""Prevent descending into the current subtree. |
|
|
Instead, the next returned event will be the 'end' event of the current element |
|
|
(if included), ignoring any children or descendants. |
|
|
|
|
|
This has no effect right after an 'end' or 'end-ns' event. |
|
|
""" |
|
|
if self._skip_state == IWSKIP_CAN_SKIP: |
|
|
self._skip_state = IWSKIP_SKIP_NEXT |
|
|
|
|
|
@cython.final |
|
|
cdef int _start_node(self, _Element node) except -1: |
|
|
cdef int ns_count |
|
|
if self._event_filter & PARSE_EVENT_FILTER_START_NS: |
|
|
ns_count = _appendStartNsEvents(node._c_node, self._events) |
|
|
if self._events: |
|
|
self._skip_state = IWSKIP_NEXT_IS_START |
|
|
elif self._event_filter & PARSE_EVENT_FILTER_END_NS: |
|
|
ns_count = _countNsDefs(node._c_node) |
|
|
else: |
|
|
ns_count = 0 |
|
|
if self._event_filter & PARSE_EVENT_FILTER_START: |
|
|
if self._matcher is None or self._matcher.matches(node._c_node): |
|
|
self._events.append( ("start", node) ) |
|
|
self._skip_state = IWSKIP_NEXT_IS_START |
|
|
return ns_count |
|
|
|
|
|
@cython.final |
|
|
cdef _Element _end_node(self): |
|
|
cdef _Element node |
|
|
cdef int i, ns_count |
|
|
node, ns_count = self._node_stack.pop() |
|
|
if self._event_filter & PARSE_EVENT_FILTER_END: |
|
|
if self._matcher is None or self._matcher.matches(node._c_node): |
|
|
self._events.append( ("end", node) ) |
|
|
if self._event_filter & PARSE_EVENT_FILTER_END_NS and ns_count: |
|
|
event = ("end-ns", None) |
|
|
for i in range(ns_count): |
|
|
self._events.append(event) |
|
|
return node |
|
|
|
|
|
|
|
|
cdef int _countNsDefs(xmlNode* c_node) noexcept: |
|
|
cdef xmlNs* c_ns |
|
|
cdef int count |
|
|
count = 0 |
|
|
c_ns = c_node.nsDef |
|
|
while c_ns is not NULL: |
|
|
count += (c_ns.href is not NULL) |
|
|
c_ns = c_ns.next |
|
|
return count |
|
|
|
|
|
|
|
|
cdef int _appendStartNsEvents(xmlNode* c_node, list event_list) except -1: |
|
|
cdef xmlNs* c_ns |
|
|
cdef int count |
|
|
count = 0 |
|
|
c_ns = c_node.nsDef |
|
|
while c_ns is not NULL: |
|
|
if c_ns.href: |
|
|
ns_tuple = (funicodeOrEmpty(c_ns.prefix), |
|
|
funicode(c_ns.href)) |
|
|
event_list.append( ("start-ns", ns_tuple) ) |
|
|
count += 1 |
|
|
c_ns = c_ns.next |
|
|
return count |
|
|
|