| | |
| |
|
| |
|
| | import difflib |
| | from lxml import etree |
| | from lxml.html import fragment_fromstring |
| | import re |
| |
|
| | __all__ = ['html_annotate', 'htmldiff'] |
| |
|
| | try: |
| | from html import escape as html_escape |
| | except ImportError: |
| | from cgi import escape as html_escape |
| | try: |
| | _unicode = unicode |
| | except NameError: |
| | |
| | _unicode = str |
| | try: |
| | basestring |
| | except NameError: |
| | |
| | basestring = str |
| |
|
| | |
| | |
| | |
| |
|
| | def default_markup(text, version): |
| | return '<span title="%s">%s</span>' % ( |
| | html_escape(_unicode(version), 1), text) |
| |
|
| | def html_annotate(doclist, markup=default_markup): |
| | """ |
| | doclist should be ordered from oldest to newest, like:: |
| | |
| | >>> version1 = 'Hello World' |
| | >>> version2 = 'Goodbye World' |
| | >>> print(html_annotate([(version1, 'version 1'), |
| | ... (version2, 'version 2')])) |
| | <span title="version 2">Goodbye</span> <span title="version 1">World</span> |
| | |
| | The documents must be *fragments* (str/UTF8 or unicode), not |
| | complete documents |
| | |
| | The markup argument is a function to markup the spans of words. |
| | This function is called like markup('Hello', 'version 2'), and |
| | returns HTML. The first argument is text and never includes any |
| | markup. The default uses a span with a title: |
| | |
| | >>> print(default_markup('Some Text', 'by Joe')) |
| | <span title="by Joe">Some Text</span> |
| | """ |
| | |
| | |
| | |
| | |
| | |
| | tokenlist = [tokenize_annotated(doc, version) |
| | for doc, version in doclist] |
| | cur_tokens = tokenlist[0] |
| | for tokens in tokenlist[1:]: |
| | html_annotate_merge_annotations(cur_tokens, tokens) |
| | cur_tokens = tokens |
| |
|
| | |
| | |
| | cur_tokens = compress_tokens(cur_tokens) |
| | |
| | result = markup_serialize_tokens(cur_tokens, markup) |
| | return ''.join(result).strip() |
| |
|
| | def tokenize_annotated(doc, annotation): |
| | """Tokenize a document and add an annotation attribute to each token |
| | """ |
| | tokens = tokenize(doc, include_hrefs=False) |
| | for tok in tokens: |
| | tok.annotation = annotation |
| | return tokens |
| |
|
| | def html_annotate_merge_annotations(tokens_old, tokens_new): |
| | """Merge the annotations from tokens_old into tokens_new, when the |
| | tokens in the new document already existed in the old document. |
| | """ |
| | s = InsensitiveSequenceMatcher(a=tokens_old, b=tokens_new) |
| | commands = s.get_opcodes() |
| |
|
| | for command, i1, i2, j1, j2 in commands: |
| | if command == 'equal': |
| | eq_old = tokens_old[i1:i2] |
| | eq_new = tokens_new[j1:j2] |
| | copy_annotations(eq_old, eq_new) |
| |
|
| | def copy_annotations(src, dest): |
| | """ |
| | Copy annotations from the tokens listed in src to the tokens in dest |
| | """ |
| | assert len(src) == len(dest) |
| | for src_tok, dest_tok in zip(src, dest): |
| | dest_tok.annotation = src_tok.annotation |
| |
|
| | def compress_tokens(tokens): |
| | """ |
| | Combine adjacent tokens when there is no HTML between the tokens, |
| | and they share an annotation |
| | """ |
| | result = [tokens[0]] |
| | for tok in tokens[1:]: |
| | if (not result[-1].post_tags and |
| | not tok.pre_tags and |
| | result[-1].annotation == tok.annotation): |
| | compress_merge_back(result, tok) |
| | else: |
| | result.append(tok) |
| | return result |
| |
|
| | def compress_merge_back(tokens, tok): |
| | """ Merge tok into the last element of tokens (modifying the list of |
| | tokens in-place). """ |
| | last = tokens[-1] |
| | if type(last) is not token or type(tok) is not token: |
| | tokens.append(tok) |
| | else: |
| | text = _unicode(last) |
| | if last.trailing_whitespace: |
| | text += last.trailing_whitespace |
| | text += tok |
| | merged = token(text, |
| | pre_tags=last.pre_tags, |
| | post_tags=tok.post_tags, |
| | trailing_whitespace=tok.trailing_whitespace) |
| | merged.annotation = last.annotation |
| | tokens[-1] = merged |
| | |
| | def markup_serialize_tokens(tokens, markup_func): |
| | """ |
| | Serialize the list of tokens into a list of text chunks, calling |
| | markup_func around text to add annotations. |
| | """ |
| | for token in tokens: |
| | yield from token.pre_tags |
| | html = token.html() |
| | html = markup_func(html, token.annotation) |
| | if token.trailing_whitespace: |
| | html += token.trailing_whitespace |
| | yield html |
| | yield from token.post_tags |
| |
|
| |
|
| | |
| | |
| | |
| |
|
| | def htmldiff(old_html, new_html): |
| | |
| | |
| | """ Do a diff of the old and new document. The documents are HTML |
| | *fragments* (str/UTF8 or unicode), they are not complete documents |
| | (i.e., no <html> tag). |
| | |
| | Returns HTML with <ins> and <del> tags added around the |
| | appropriate text. |
| | |
| | Markup is generally ignored, with the markup from new_html |
| | preserved, and possibly some markup from old_html (though it is |
| | considered acceptable to lose some of the old markup). Only the |
| | words in the HTML are diffed. The exception is <img> tags, which |
| | are treated like words, and the href attribute of <a> tags, which |
| | are noted inside the tag itself when there are changes. |
| | """ |
| | old_html_tokens = tokenize(old_html) |
| | new_html_tokens = tokenize(new_html) |
| | result = htmldiff_tokens(old_html_tokens, new_html_tokens) |
| | result = ''.join(result).strip() |
| | return fixup_ins_del_tags(result) |
| |
|
| | def htmldiff_tokens(html1_tokens, html2_tokens): |
| | """ Does a diff on the tokens themselves, returning a list of text |
| | chunks (not tokens). |
| | """ |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | s = InsensitiveSequenceMatcher(a=html1_tokens, b=html2_tokens) |
| | commands = s.get_opcodes() |
| | result = [] |
| | for command, i1, i2, j1, j2 in commands: |
| | if command == 'equal': |
| | result.extend(expand_tokens(html2_tokens[j1:j2], equal=True)) |
| | continue |
| | if command == 'insert' or command == 'replace': |
| | ins_tokens = expand_tokens(html2_tokens[j1:j2]) |
| | merge_insert(ins_tokens, result) |
| | if command == 'delete' or command == 'replace': |
| | del_tokens = expand_tokens(html1_tokens[i1:i2]) |
| | merge_delete(del_tokens, result) |
| | |
| | |
| | |
| | |
| | result = cleanup_delete(result) |
| |
|
| | return result |
| |
|
| | def expand_tokens(tokens, equal=False): |
| | """Given a list of tokens, return a generator of the chunks of |
| | text for the data in the tokens. |
| | """ |
| | for token in tokens: |
| | yield from token.pre_tags |
| | if not equal or not token.hide_when_equal: |
| | if token.trailing_whitespace: |
| | yield token.html() + token.trailing_whitespace |
| | else: |
| | yield token.html() |
| | yield from token.post_tags |
| |
|
| | def merge_insert(ins_chunks, doc): |
| | """ doc is the already-handled document (as a list of text chunks); |
| | here we add <ins>ins_chunks</ins> to the end of that. """ |
| | |
| | |
| | |
| | unbalanced_start, balanced, unbalanced_end = split_unbalanced(ins_chunks) |
| | doc.extend(unbalanced_start) |
| | if doc and not doc[-1].endswith(' '): |
| | |
| | |
| | doc[-1] += ' ' |
| | doc.append('<ins>') |
| | if balanced and balanced[-1].endswith(' '): |
| | |
| | balanced[-1] = balanced[-1][:-1] |
| | doc.extend(balanced) |
| | doc.append('</ins> ') |
| | doc.extend(unbalanced_end) |
| |
|
| | |
| | |
| | |
| | class DEL_START: |
| | pass |
| | class DEL_END: |
| | pass |
| |
|
| | class NoDeletes(Exception): |
| | """ Raised when the document no longer contains any pending deletes |
| | (DEL_START/DEL_END) """ |
| |
|
| | def merge_delete(del_chunks, doc): |
| | """ Adds the text chunks in del_chunks to the document doc (another |
| | list of text chunks) with marker to show it is a delete. |
| | cleanup_delete later resolves these markers into <del> tags.""" |
| | doc.append(DEL_START) |
| | doc.extend(del_chunks) |
| | doc.append(DEL_END) |
| |
|
| | def cleanup_delete(chunks): |
| | """ Cleans up any DEL_START/DEL_END markers in the document, replacing |
| | them with <del></del>. To do this while keeping the document |
| | valid, it may need to drop some tags (either start or end tags). |
| | |
| | It may also move the del into adjacent tags to try to move it to a |
| | similar location where it was originally located (e.g., moving a |
| | delete into preceding <div> tag, if the del looks like (DEL_START, |
| | 'Text</div>', DEL_END)""" |
| | while 1: |
| | |
| | |
| | |
| | try: |
| | pre_delete, delete, post_delete = split_delete(chunks) |
| | except NoDeletes: |
| | |
| | break |
| | |
| | |
| | unbalanced_start, balanced, unbalanced_end = split_unbalanced(delete) |
| | |
| | |
| | locate_unbalanced_start(unbalanced_start, pre_delete, post_delete) |
| | locate_unbalanced_end(unbalanced_end, pre_delete, post_delete) |
| | doc = pre_delete |
| | if doc and not doc[-1].endswith(' '): |
| | |
| | doc[-1] += ' ' |
| | doc.append('<del>') |
| | if balanced and balanced[-1].endswith(' '): |
| | |
| | balanced[-1] = balanced[-1][:-1] |
| | doc.extend(balanced) |
| | doc.append('</del> ') |
| | doc.extend(post_delete) |
| | chunks = doc |
| | return chunks |
| |
|
| | def split_unbalanced(chunks): |
| | """Return (unbalanced_start, balanced, unbalanced_end), where each is |
| | a list of text and tag chunks. |
| | |
| | unbalanced_start is a list of all the tags that are opened, but |
| | not closed in this span. Similarly, unbalanced_end is a list of |
| | tags that are closed but were not opened. Extracting these might |
| | mean some reordering of the chunks.""" |
| | start = [] |
| | end = [] |
| | tag_stack = [] |
| | balanced = [] |
| | for chunk in chunks: |
| | if not chunk.startswith('<'): |
| | balanced.append(chunk) |
| | continue |
| | endtag = chunk[1] == '/' |
| | name = chunk.split()[0].strip('<>/') |
| | if name in empty_tags: |
| | balanced.append(chunk) |
| | continue |
| | if endtag: |
| | if tag_stack and tag_stack[-1][0] == name: |
| | balanced.append(chunk) |
| | name, pos, tag = tag_stack.pop() |
| | balanced[pos] = tag |
| | elif tag_stack: |
| | start.extend([tag for name, pos, tag in tag_stack]) |
| | tag_stack = [] |
| | end.append(chunk) |
| | else: |
| | end.append(chunk) |
| | else: |
| | tag_stack.append((name, len(balanced), chunk)) |
| | balanced.append(None) |
| | start.extend( |
| | [chunk for name, pos, chunk in tag_stack]) |
| | balanced = [chunk for chunk in balanced if chunk is not None] |
| | return start, balanced, end |
| |
|
| | def split_delete(chunks): |
| | """ Returns (stuff_before_DEL_START, stuff_inside_DEL_START_END, |
| | stuff_after_DEL_END). Returns the first case found (there may be |
| | more DEL_STARTs in stuff_after_DEL_END). Raises NoDeletes if |
| | there's no DEL_START found. """ |
| | try: |
| | pos = chunks.index(DEL_START) |
| | except ValueError: |
| | raise NoDeletes |
| | pos2 = chunks.index(DEL_END) |
| | return chunks[:pos], chunks[pos+1:pos2], chunks[pos2+1:] |
| |
|
| | def locate_unbalanced_start(unbalanced_start, pre_delete, post_delete): |
| | """ pre_delete and post_delete implicitly point to a place in the |
| | document (where the two were split). This moves that point (by |
| | popping items from one and pushing them onto the other). It moves |
| | the point to try to find a place where unbalanced_start applies. |
| | |
| | As an example:: |
| | |
| | >>> unbalanced_start = ['<div>'] |
| | >>> doc = ['<p>', 'Text', '</p>', '<div>', 'More Text', '</div>'] |
| | >>> pre, post = doc[:3], doc[3:] |
| | >>> pre, post |
| | (['<p>', 'Text', '</p>'], ['<div>', 'More Text', '</div>']) |
| | >>> locate_unbalanced_start(unbalanced_start, pre, post) |
| | >>> pre, post |
| | (['<p>', 'Text', '</p>', '<div>'], ['More Text', '</div>']) |
| | |
| | As you can see, we moved the point so that the dangling <div> that |
| | we found will be effectively replaced by the div in the original |
| | document. If this doesn't work out, we just throw away |
| | unbalanced_start without doing anything. |
| | """ |
| | while 1: |
| | if not unbalanced_start: |
| | |
| | break |
| | finding = unbalanced_start[0] |
| | finding_name = finding.split()[0].strip('<>') |
| | if not post_delete: |
| | break |
| | next = post_delete[0] |
| | if next is DEL_START or not next.startswith('<'): |
| | |
| | break |
| | if next[1] == '/': |
| | |
| | break |
| | name = next.split()[0].strip('<>') |
| | if name == 'ins': |
| | |
| | break |
| | assert name != 'del', ( |
| | "Unexpected delete tag: %r" % next) |
| | if name == finding_name: |
| | unbalanced_start.pop(0) |
| | pre_delete.append(post_delete.pop(0)) |
| | else: |
| | |
| | break |
| |
|
| | def locate_unbalanced_end(unbalanced_end, pre_delete, post_delete): |
| | """ like locate_unbalanced_start, except handling end tags and |
| | possibly moving the point earlier in the document. """ |
| | while 1: |
| | if not unbalanced_end: |
| | |
| | break |
| | finding = unbalanced_end[-1] |
| | finding_name = finding.split()[0].strip('<>/') |
| | if not pre_delete: |
| | break |
| | next = pre_delete[-1] |
| | if next is DEL_END or not next.startswith('</'): |
| | |
| | break |
| | name = next.split()[0].strip('<>/') |
| | if name == 'ins' or name == 'del': |
| | |
| | break |
| | if name == finding_name: |
| | unbalanced_end.pop() |
| | post_delete.insert(0, pre_delete.pop()) |
| | else: |
| | |
| | break |
| |
|
| | class token(_unicode): |
| | """ Represents a diffable token, generally a word that is displayed to |
| | the user. Opening tags are attached to this token when they are |
| | adjacent (pre_tags) and closing tags that follow the word |
| | (post_tags). Some exceptions occur when there are empty tags |
| | adjacent to a word, so there may be close tags in pre_tags, or |
| | open tags in post_tags. |
| | |
| | We also keep track of whether the word was originally followed by |
| | whitespace, even though we do not want to treat the word as |
| | equivalent to a similar word that does not have a trailing |
| | space.""" |
| |
|
| | |
| | |
| | hide_when_equal = False |
| |
|
| | def __new__(cls, text, pre_tags=None, post_tags=None, trailing_whitespace=""): |
| | obj = _unicode.__new__(cls, text) |
| |
|
| | if pre_tags is not None: |
| | obj.pre_tags = pre_tags |
| | else: |
| | obj.pre_tags = [] |
| |
|
| | if post_tags is not None: |
| | obj.post_tags = post_tags |
| | else: |
| | obj.post_tags = [] |
| |
|
| | obj.trailing_whitespace = trailing_whitespace |
| |
|
| | return obj |
| |
|
| | def __repr__(self): |
| | return 'token(%s, %r, %r, %r)' % (_unicode.__repr__(self), self.pre_tags, |
| | self.post_tags, self.trailing_whitespace) |
| |
|
| | def html(self): |
| | return _unicode(self) |
| |
|
| | class tag_token(token): |
| |
|
| | """ Represents a token that is actually a tag. Currently this is just |
| | the <img> tag, which takes up visible space just like a word but |
| | is only represented in a document by a tag. """ |
| |
|
| | def __new__(cls, tag, data, html_repr, pre_tags=None, |
| | post_tags=None, trailing_whitespace=""): |
| | obj = token.__new__(cls, "%s: %s" % (type, data), |
| | pre_tags=pre_tags, |
| | post_tags=post_tags, |
| | trailing_whitespace=trailing_whitespace) |
| | obj.tag = tag |
| | obj.data = data |
| | obj.html_repr = html_repr |
| | return obj |
| |
|
| | def __repr__(self): |
| | return 'tag_token(%s, %s, html_repr=%s, post_tags=%r, pre_tags=%r, trailing_whitespace=%r)' % ( |
| | self.tag, |
| | self.data, |
| | self.html_repr, |
| | self.pre_tags, |
| | self.post_tags, |
| | self.trailing_whitespace) |
| | def html(self): |
| | return self.html_repr |
| |
|
| | class href_token(token): |
| |
|
| | """ Represents the href in an anchor tag. Unlike other words, we only |
| | show the href when it changes. """ |
| |
|
| | hide_when_equal = True |
| |
|
| | def html(self): |
| | return ' Link: %s' % self |
| |
|
| | def tokenize(html, include_hrefs=True): |
| | """ |
| | Parse the given HTML and returns token objects (words with attached tags). |
| | |
| | This parses only the content of a page; anything in the head is |
| | ignored, and the <head> and <body> elements are themselves |
| | optional. The content is then parsed by lxml, which ensures the |
| | validity of the resulting parsed document (though lxml may make |
| | incorrect guesses when the markup is particular bad). |
| | |
| | <ins> and <del> tags are also eliminated from the document, as |
| | that gets confusing. |
| | |
| | If include_hrefs is true, then the href attribute of <a> tags is |
| | included as a special kind of diffable token.""" |
| | if etree.iselement(html): |
| | body_el = html |
| | else: |
| | body_el = parse_html(html, cleanup=True) |
| | |
| | chunks = flatten_el(body_el, skip_tag=True, include_hrefs=include_hrefs) |
| | |
| | return fixup_chunks(chunks) |
| |
|
| | def parse_html(html, cleanup=True): |
| | """ |
| | Parses an HTML fragment, returning an lxml element. Note that the HTML will be |
| | wrapped in a <div> tag that was not in the original document. |
| | |
| | If cleanup is true, make sure there's no <head> or <body>, and get |
| | rid of any <ins> and <del> tags. |
| | """ |
| | if cleanup: |
| | |
| | html = cleanup_html(html) |
| | return fragment_fromstring(html, create_parent=True) |
| |
|
| | _body_re = re.compile(r'<body.*?>', re.I|re.S) |
| | _end_body_re = re.compile(r'</body.*?>', re.I|re.S) |
| | _ins_del_re = re.compile(r'</?(ins|del).*?>', re.I|re.S) |
| |
|
| | def cleanup_html(html): |
| | """ This 'cleans' the HTML, meaning that any page structure is removed |
| | (only the contents of <body> are used, if there is any <body). |
| | Also <ins> and <del> tags are removed. """ |
| | match = _body_re.search(html) |
| | if match: |
| | html = html[match.end():] |
| | match = _end_body_re.search(html) |
| | if match: |
| | html = html[:match.start()] |
| | html = _ins_del_re.sub('', html) |
| | return html |
| | |
| |
|
| | end_whitespace_re = re.compile(r'[ \t\n\r]$') |
| |
|
| | def split_trailing_whitespace(word): |
| | """ |
| | This function takes a word, such as 'test\n\n' and returns ('test','\n\n') |
| | """ |
| | stripped_length = len(word.rstrip()) |
| | return word[0:stripped_length], word[stripped_length:] |
| |
|
| |
|
| | def fixup_chunks(chunks): |
| | """ |
| | This function takes a list of chunks and produces a list of tokens. |
| | """ |
| | tag_accum = [] |
| | cur_word = None |
| | result = [] |
| | for chunk in chunks: |
| | if isinstance(chunk, tuple): |
| | if chunk[0] == 'img': |
| | src = chunk[1] |
| | tag, trailing_whitespace = split_trailing_whitespace(chunk[2]) |
| | cur_word = tag_token('img', src, html_repr=tag, |
| | pre_tags=tag_accum, |
| | trailing_whitespace=trailing_whitespace) |
| | tag_accum = [] |
| | result.append(cur_word) |
| |
|
| | elif chunk[0] == 'href': |
| | href = chunk[1] |
| | cur_word = href_token(href, pre_tags=tag_accum, trailing_whitespace=" ") |
| | tag_accum = [] |
| | result.append(cur_word) |
| | continue |
| |
|
| | if is_word(chunk): |
| | chunk, trailing_whitespace = split_trailing_whitespace(chunk) |
| | cur_word = token(chunk, pre_tags=tag_accum, trailing_whitespace=trailing_whitespace) |
| | tag_accum = [] |
| | result.append(cur_word) |
| |
|
| | elif is_start_tag(chunk): |
| | tag_accum.append(chunk) |
| |
|
| | elif is_end_tag(chunk): |
| | if tag_accum: |
| | tag_accum.append(chunk) |
| | else: |
| | assert cur_word, ( |
| | "Weird state, cur_word=%r, result=%r, chunks=%r of %r" |
| | % (cur_word, result, chunk, chunks)) |
| | cur_word.post_tags.append(chunk) |
| | else: |
| | assert False |
| |
|
| | if not result: |
| | return [token('', pre_tags=tag_accum)] |
| | else: |
| | result[-1].post_tags.extend(tag_accum) |
| |
|
| | return result |
| |
|
| |
|
| | |
| | empty_tags = ( |
| | 'param', 'img', 'area', 'br', 'basefont', 'input', |
| | 'base', 'meta', 'link', 'col') |
| |
|
| | block_level_tags = ( |
| | 'address', |
| | 'blockquote', |
| | 'center', |
| | 'dir', |
| | 'div', |
| | 'dl', |
| | 'fieldset', |
| | 'form', |
| | 'h1', |
| | 'h2', |
| | 'h3', |
| | 'h4', |
| | 'h5', |
| | 'h6', |
| | 'hr', |
| | 'isindex', |
| | 'menu', |
| | 'noframes', |
| | 'noscript', |
| | 'ol', |
| | 'p', |
| | 'pre', |
| | 'table', |
| | 'ul', |
| | ) |
| |
|
| | block_level_container_tags = ( |
| | 'dd', |
| | 'dt', |
| | 'frameset', |
| | 'li', |
| | 'tbody', |
| | 'td', |
| | 'tfoot', |
| | 'th', |
| | 'thead', |
| | 'tr', |
| | ) |
| |
|
| |
|
| | def flatten_el(el, include_hrefs, skip_tag=False): |
| | """ Takes an lxml element el, and generates all the text chunks for |
| | that tag. Each start tag is a chunk, each word is a chunk, and each |
| | end tag is a chunk. |
| | |
| | If skip_tag is true, then the outermost container tag is |
| | not returned (just its contents).""" |
| | if not skip_tag: |
| | if el.tag == 'img': |
| | yield ('img', el.get('src'), start_tag(el)) |
| | else: |
| | yield start_tag(el) |
| | if el.tag in empty_tags and not el.text and not len(el) and not el.tail: |
| | return |
| | start_words = split_words(el.text) |
| | for word in start_words: |
| | yield html_escape(word) |
| | for child in el: |
| | yield from flatten_el(child, include_hrefs=include_hrefs) |
| | if el.tag == 'a' and el.get('href') and include_hrefs: |
| | yield ('href', el.get('href')) |
| | if not skip_tag: |
| | yield end_tag(el) |
| | end_words = split_words(el.tail) |
| | for word in end_words: |
| | yield html_escape(word) |
| |
|
| | split_words_re = re.compile(r'\S+(?:\s+|$)', re.U) |
| |
|
| | def split_words(text): |
| | """ Splits some text into words. Includes trailing whitespace |
| | on each word when appropriate. """ |
| | if not text or not text.strip(): |
| | return [] |
| |
|
| | words = split_words_re.findall(text) |
| | return words |
| |
|
| | start_whitespace_re = re.compile(r'^[ \t\n\r]') |
| |
|
| | def start_tag(el): |
| | """ |
| | The text representation of the start tag for a tag. |
| | """ |
| | return '<%s%s>' % ( |
| | el.tag, ''.join([' %s="%s"' % (name, html_escape(value, True)) |
| | for name, value in el.attrib.items()])) |
| |
|
| | def end_tag(el): |
| | """ The text representation of an end tag for a tag. Includes |
| | trailing whitespace when appropriate. """ |
| | if el.tail and start_whitespace_re.search(el.tail): |
| | extra = ' ' |
| | else: |
| | extra = '' |
| | return '</%s>%s' % (el.tag, extra) |
| |
|
| | def is_word(tok): |
| | return not tok.startswith('<') |
| |
|
| | def is_end_tag(tok): |
| | return tok.startswith('</') |
| |
|
| | def is_start_tag(tok): |
| | return tok.startswith('<') and not tok.startswith('</') |
| |
|
| | def fixup_ins_del_tags(html): |
| | """ Given an html string, move any <ins> or <del> tags inside of any |
| | block-level elements, e.g. transform <ins><p>word</p></ins> to |
| | <p><ins>word</ins></p> """ |
| | doc = parse_html(html, cleanup=False) |
| | _fixup_ins_del_tags(doc) |
| | html = serialize_html_fragment(doc, skip_outer=True) |
| | return html |
| |
|
| | def serialize_html_fragment(el, skip_outer=False): |
| | """ Serialize a single lxml element as HTML. The serialized form |
| | includes the elements tail. |
| | |
| | If skip_outer is true, then don't serialize the outermost tag |
| | """ |
| | assert not isinstance(el, basestring), ( |
| | "You should pass in an element, not a string like %r" % el) |
| | html = etree.tostring(el, method="html", encoding=_unicode) |
| | if skip_outer: |
| | |
| | html = html[html.find('>')+1:] |
| | |
| | html = html[:html.rfind('<')] |
| | return html.strip() |
| | else: |
| | return html |
| |
|
| | def _fixup_ins_del_tags(doc): |
| | """fixup_ins_del_tags that works on an lxml document in-place |
| | """ |
| | for tag in ['ins', 'del']: |
| | for el in doc.xpath('descendant-or-self::%s' % tag): |
| | if not _contains_block_level_tag(el): |
| | continue |
| | _move_el_inside_block(el, tag=tag) |
| | el.drop_tag() |
| | |
| |
|
| | def _contains_block_level_tag(el): |
| | """True if the element contains any block-level elements, like <p>, <td>, etc. |
| | """ |
| | if el.tag in block_level_tags or el.tag in block_level_container_tags: |
| | return True |
| | for child in el: |
| | if _contains_block_level_tag(child): |
| | return True |
| | return False |
| |
|
| | def _move_el_inside_block(el, tag): |
| | """ helper for _fixup_ins_del_tags; actually takes the <ins> etc tags |
| | and moves them inside any block-level tags. """ |
| | for child in el: |
| | if _contains_block_level_tag(child): |
| | break |
| | else: |
| | |
| | children_tag = etree.Element(tag) |
| | children_tag.text = el.text |
| | el.text = None |
| | children_tag.extend(list(el)) |
| | el[:] = [children_tag] |
| | return |
| | for child in list(el): |
| | if _contains_block_level_tag(child): |
| | _move_el_inside_block(child, tag) |
| | if child.tail: |
| | tail_tag = etree.Element(tag) |
| | tail_tag.text = child.tail |
| | child.tail = None |
| | el.insert(el.index(child)+1, tail_tag) |
| | else: |
| | child_tag = etree.Element(tag) |
| | el.replace(child, child_tag) |
| | child_tag.append(child) |
| | if el.text: |
| | text_tag = etree.Element(tag) |
| | text_tag.text = el.text |
| | el.text = None |
| | el.insert(0, text_tag) |
| | |
| | def _merge_element_contents(el): |
| | """ |
| | Removes an element, but merges its contents into its place, e.g., |
| | given <p>Hi <i>there!</i></p>, if you remove the <i> element you get |
| | <p>Hi there!</p> |
| | """ |
| | parent = el.getparent() |
| | text = el.text or '' |
| | if el.tail: |
| | if not len(el): |
| | text += el.tail |
| | else: |
| | if el[-1].tail: |
| | el[-1].tail += el.tail |
| | else: |
| | el[-1].tail = el.tail |
| | index = parent.index(el) |
| | if text: |
| | if index == 0: |
| | previous = None |
| | else: |
| | previous = parent[index-1] |
| | if previous is None: |
| | if parent.text: |
| | parent.text += text |
| | else: |
| | parent.text = text |
| | else: |
| | if previous.tail: |
| | previous.tail += text |
| | else: |
| | previous.tail = text |
| | parent[index:index+1] = el.getchildren() |
| |
|
| | class InsensitiveSequenceMatcher(difflib.SequenceMatcher): |
| | """ |
| | Acts like SequenceMatcher, but tries not to find very small equal |
| | blocks amidst large spans of changes |
| | """ |
| |
|
| | threshold = 2 |
| | |
| | def get_matching_blocks(self): |
| | size = min(len(self.b), len(self.b)) |
| | threshold = min(self.threshold, size / 4) |
| | actual = difflib.SequenceMatcher.get_matching_blocks(self) |
| | return [item for item in actual |
| | if item[2] > threshold |
| | or not item[2]] |
| |
|
| | if __name__ == '__main__': |
| | from lxml.html import _diffcommand |
| | _diffcommand.main() |
| | |
| |
|