peacock-data-public-datasets-idc-cronscript
/
venv
/lib
/python3.10
/site-packages
/lxml
/html
/diff.py
| # cython: language_level=3 | |
| import difflib | |
| from lxml import etree | |
| from lxml.html import fragment_fromstring | |
| import re | |
| __all__ = ['html_annotate', 'htmldiff'] | |
| try: | |
| from html import escape as html_escape | |
| except ImportError: | |
| from cgi import escape as html_escape | |
| try: | |
| _unicode = unicode | |
| except NameError: | |
| # Python 3 | |
| _unicode = str | |
| try: | |
| basestring | |
| except NameError: | |
| # Python 3 | |
| basestring = str | |
| ############################################################ | |
| ## Annotation | |
| ############################################################ | |
| def default_markup(text, version): | |
| return '<span title="%s">%s</span>' % ( | |
| html_escape(_unicode(version), 1), text) | |
| def html_annotate(doclist, markup=default_markup): | |
| """ | |
| doclist should be ordered from oldest to newest, like:: | |
| >>> version1 = 'Hello World' | |
| >>> version2 = 'Goodbye World' | |
| >>> print(html_annotate([(version1, 'version 1'), | |
| ... (version2, 'version 2')])) | |
| <span title="version 2">Goodbye</span> <span title="version 1">World</span> | |
| The documents must be *fragments* (str/UTF8 or unicode), not | |
| complete documents | |
| The markup argument is a function to markup the spans of words. | |
| This function is called like markup('Hello', 'version 2'), and | |
| returns HTML. The first argument is text and never includes any | |
| markup. The default uses a span with a title: | |
| >>> print(default_markup('Some Text', 'by Joe')) | |
| <span title="by Joe">Some Text</span> | |
| """ | |
| # The basic strategy we have is to split the documents up into | |
| # logical tokens (which are words with attached markup). We then | |
| # do diffs of each of the versions to track when a token first | |
| # appeared in the document; the annotation attached to the token | |
| # is the version where it first appeared. | |
| tokenlist = [tokenize_annotated(doc, version) | |
| for doc, version in doclist] | |
| cur_tokens = tokenlist[0] | |
| for tokens in tokenlist[1:]: | |
| html_annotate_merge_annotations(cur_tokens, tokens) | |
| cur_tokens = tokens | |
| # After we've tracked all the tokens, we can combine spans of text | |
| # that are adjacent and have the same annotation | |
| cur_tokens = compress_tokens(cur_tokens) | |
| # And finally add markup | |
| result = markup_serialize_tokens(cur_tokens, markup) | |
| return ''.join(result).strip() | |
| def tokenize_annotated(doc, annotation): | |
| """Tokenize a document and add an annotation attribute to each token | |
| """ | |
| tokens = tokenize(doc, include_hrefs=False) | |
| for tok in tokens: | |
| tok.annotation = annotation | |
| return tokens | |
| def html_annotate_merge_annotations(tokens_old, tokens_new): | |
| """Merge the annotations from tokens_old into tokens_new, when the | |
| tokens in the new document already existed in the old document. | |
| """ | |
| s = InsensitiveSequenceMatcher(a=tokens_old, b=tokens_new) | |
| commands = s.get_opcodes() | |
| for command, i1, i2, j1, j2 in commands: | |
| if command == 'equal': | |
| eq_old = tokens_old[i1:i2] | |
| eq_new = tokens_new[j1:j2] | |
| copy_annotations(eq_old, eq_new) | |
| def copy_annotations(src, dest): | |
| """ | |
| Copy annotations from the tokens listed in src to the tokens in dest | |
| """ | |
| assert len(src) == len(dest) | |
| for src_tok, dest_tok in zip(src, dest): | |
| dest_tok.annotation = src_tok.annotation | |
| def compress_tokens(tokens): | |
| """ | |
| Combine adjacent tokens when there is no HTML between the tokens, | |
| and they share an annotation | |
| """ | |
| result = [tokens[0]] | |
| for tok in tokens[1:]: | |
| if (not result[-1].post_tags and | |
| not tok.pre_tags and | |
| result[-1].annotation == tok.annotation): | |
| compress_merge_back(result, tok) | |
| else: | |
| result.append(tok) | |
| return result | |
| def compress_merge_back(tokens, tok): | |
| """ Merge tok into the last element of tokens (modifying the list of | |
| tokens in-place). """ | |
| last = tokens[-1] | |
| if type(last) is not token or type(tok) is not token: | |
| tokens.append(tok) | |
| else: | |
| text = _unicode(last) | |
| if last.trailing_whitespace: | |
| text += last.trailing_whitespace | |
| text += tok | |
| merged = token(text, | |
| pre_tags=last.pre_tags, | |
| post_tags=tok.post_tags, | |
| trailing_whitespace=tok.trailing_whitespace) | |
| merged.annotation = last.annotation | |
| tokens[-1] = merged | |
| def markup_serialize_tokens(tokens, markup_func): | |
| """ | |
| Serialize the list of tokens into a list of text chunks, calling | |
| markup_func around text to add annotations. | |
| """ | |
| for token in tokens: | |
| yield from token.pre_tags | |
| html = token.html() | |
| html = markup_func(html, token.annotation) | |
| if token.trailing_whitespace: | |
| html += token.trailing_whitespace | |
| yield html | |
| yield from token.post_tags | |
| ############################################################ | |
| ## HTML Diffs | |
| ############################################################ | |
| def htmldiff(old_html, new_html): | |
| ## FIXME: this should take parsed documents too, and use their body | |
| ## or other content. | |
| """ Do a diff of the old and new document. The documents are HTML | |
| *fragments* (str/UTF8 or unicode), they are not complete documents | |
| (i.e., no <html> tag). | |
| Returns HTML with <ins> and <del> tags added around the | |
| appropriate text. | |
| Markup is generally ignored, with the markup from new_html | |
| preserved, and possibly some markup from old_html (though it is | |
| considered acceptable to lose some of the old markup). Only the | |
| words in the HTML are diffed. The exception is <img> tags, which | |
| are treated like words, and the href attribute of <a> tags, which | |
| are noted inside the tag itself when there are changes. | |
| """ | |
| old_html_tokens = tokenize(old_html) | |
| new_html_tokens = tokenize(new_html) | |
| result = htmldiff_tokens(old_html_tokens, new_html_tokens) | |
| result = ''.join(result).strip() | |
| return fixup_ins_del_tags(result) | |
| def htmldiff_tokens(html1_tokens, html2_tokens): | |
| """ Does a diff on the tokens themselves, returning a list of text | |
| chunks (not tokens). | |
| """ | |
| # There are several passes as we do the differences. The tokens | |
| # isolate the portion of the content we care to diff; difflib does | |
| # all the actual hard work at that point. | |
| # | |
| # Then we must create a valid document from pieces of both the old | |
| # document and the new document. We generally prefer to take | |
| # markup from the new document, and only do a best effort attempt | |
| # to keep markup from the old document; anything that we can't | |
| # resolve we throw away. Also we try to put the deletes as close | |
| # to the location where we think they would have been -- because | |
| # we are only keeping the markup from the new document, it can be | |
| # fuzzy where in the new document the old text would have gone. | |
| # Again we just do a best effort attempt. | |
| s = InsensitiveSequenceMatcher(a=html1_tokens, b=html2_tokens) | |
| commands = s.get_opcodes() | |
| result = [] | |
| for command, i1, i2, j1, j2 in commands: | |
| if command == 'equal': | |
| result.extend(expand_tokens(html2_tokens[j1:j2], equal=True)) | |
| continue | |
| if command == 'insert' or command == 'replace': | |
| ins_tokens = expand_tokens(html2_tokens[j1:j2]) | |
| merge_insert(ins_tokens, result) | |
| if command == 'delete' or command == 'replace': | |
| del_tokens = expand_tokens(html1_tokens[i1:i2]) | |
| merge_delete(del_tokens, result) | |
| # If deletes were inserted directly as <del> then we'd have an | |
| # invalid document at this point. Instead we put in special | |
| # markers, and when the complete diffed document has been created | |
| # we try to move the deletes around and resolve any problems. | |
| result = cleanup_delete(result) | |
| return result | |
| def expand_tokens(tokens, equal=False): | |
| """Given a list of tokens, return a generator of the chunks of | |
| text for the data in the tokens. | |
| """ | |
| for token in tokens: | |
| yield from token.pre_tags | |
| if not equal or not token.hide_when_equal: | |
| if token.trailing_whitespace: | |
| yield token.html() + token.trailing_whitespace | |
| else: | |
| yield token.html() | |
| yield from token.post_tags | |
| def merge_insert(ins_chunks, doc): | |
| """ doc is the already-handled document (as a list of text chunks); | |
| here we add <ins>ins_chunks</ins> to the end of that. """ | |
| # Though we don't throw away unbalanced_start or unbalanced_end | |
| # (we assume there is accompanying markup later or earlier in the | |
| # document), we only put <ins> around the balanced portion. | |
| unbalanced_start, balanced, unbalanced_end = split_unbalanced(ins_chunks) | |
| doc.extend(unbalanced_start) | |
| if doc and not doc[-1].endswith(' '): | |
| # Fix up the case where the word before the insert didn't end with | |
| # a space | |
| doc[-1] += ' ' | |
| doc.append('<ins>') | |
| if balanced and balanced[-1].endswith(' '): | |
| # We move space outside of </ins> | |
| balanced[-1] = balanced[-1][:-1] | |
| doc.extend(balanced) | |
| doc.append('</ins> ') | |
| doc.extend(unbalanced_end) | |
| # These are sentinels to represent the start and end of a <del> | |
| # segment, until we do the cleanup phase to turn them into proper | |
| # markup: | |
| class DEL_START: | |
| pass | |
| class DEL_END: | |
| pass | |
| class NoDeletes(Exception): | |
| """ Raised when the document no longer contains any pending deletes | |
| (DEL_START/DEL_END) """ | |
| def merge_delete(del_chunks, doc): | |
| """ Adds the text chunks in del_chunks to the document doc (another | |
| list of text chunks) with marker to show it is a delete. | |
| cleanup_delete later resolves these markers into <del> tags.""" | |
| doc.append(DEL_START) | |
| doc.extend(del_chunks) | |
| doc.append(DEL_END) | |
| def cleanup_delete(chunks): | |
| """ Cleans up any DEL_START/DEL_END markers in the document, replacing | |
| them with <del></del>. To do this while keeping the document | |
| valid, it may need to drop some tags (either start or end tags). | |
| It may also move the del into adjacent tags to try to move it to a | |
| similar location where it was originally located (e.g., moving a | |
| delete into preceding <div> tag, if the del looks like (DEL_START, | |
| 'Text</div>', DEL_END)""" | |
| while 1: | |
| # Find a pending DEL_START/DEL_END, splitting the document | |
| # into stuff-preceding-DEL_START, stuff-inside, and | |
| # stuff-following-DEL_END | |
| try: | |
| pre_delete, delete, post_delete = split_delete(chunks) | |
| except NoDeletes: | |
| # Nothing found, we've cleaned up the entire doc | |
| break | |
| # The stuff-inside-DEL_START/END may not be well balanced | |
| # markup. First we figure out what unbalanced portions there are: | |
| unbalanced_start, balanced, unbalanced_end = split_unbalanced(delete) | |
| # Then we move the span forward and/or backward based on these | |
| # unbalanced portions: | |
| locate_unbalanced_start(unbalanced_start, pre_delete, post_delete) | |
| locate_unbalanced_end(unbalanced_end, pre_delete, post_delete) | |
| doc = pre_delete | |
| if doc and not doc[-1].endswith(' '): | |
| # Fix up case where the word before us didn't have a trailing space | |
| doc[-1] += ' ' | |
| doc.append('<del>') | |
| if balanced and balanced[-1].endswith(' '): | |
| # We move space outside of </del> | |
| balanced[-1] = balanced[-1][:-1] | |
| doc.extend(balanced) | |
| doc.append('</del> ') | |
| doc.extend(post_delete) | |
| chunks = doc | |
| return chunks | |
| def split_unbalanced(chunks): | |
| """Return (unbalanced_start, balanced, unbalanced_end), where each is | |
| a list of text and tag chunks. | |
| unbalanced_start is a list of all the tags that are opened, but | |
| not closed in this span. Similarly, unbalanced_end is a list of | |
| tags that are closed but were not opened. Extracting these might | |
| mean some reordering of the chunks.""" | |
| start = [] | |
| end = [] | |
| tag_stack = [] | |
| balanced = [] | |
| for chunk in chunks: | |
| if not chunk.startswith('<'): | |
| balanced.append(chunk) | |
| continue | |
| endtag = chunk[1] == '/' | |
| name = chunk.split()[0].strip('<>/') | |
| if name in empty_tags: | |
| balanced.append(chunk) | |
| continue | |
| if endtag: | |
| if tag_stack and tag_stack[-1][0] == name: | |
| balanced.append(chunk) | |
| name, pos, tag = tag_stack.pop() | |
| balanced[pos] = tag | |
| elif tag_stack: | |
| start.extend([tag for name, pos, tag in tag_stack]) | |
| tag_stack = [] | |
| end.append(chunk) | |
| else: | |
| end.append(chunk) | |
| else: | |
| tag_stack.append((name, len(balanced), chunk)) | |
| balanced.append(None) | |
| start.extend( | |
| [chunk for name, pos, chunk in tag_stack]) | |
| balanced = [chunk for chunk in balanced if chunk is not None] | |
| return start, balanced, end | |
| def split_delete(chunks): | |
| """ Returns (stuff_before_DEL_START, stuff_inside_DEL_START_END, | |
| stuff_after_DEL_END). Returns the first case found (there may be | |
| more DEL_STARTs in stuff_after_DEL_END). Raises NoDeletes if | |
| there's no DEL_START found. """ | |
| try: | |
| pos = chunks.index(DEL_START) | |
| except ValueError: | |
| raise NoDeletes | |
| pos2 = chunks.index(DEL_END) | |
| return chunks[:pos], chunks[pos+1:pos2], chunks[pos2+1:] | |
| def locate_unbalanced_start(unbalanced_start, pre_delete, post_delete): | |
| """ pre_delete and post_delete implicitly point to a place in the | |
| document (where the two were split). This moves that point (by | |
| popping items from one and pushing them onto the other). It moves | |
| the point to try to find a place where unbalanced_start applies. | |
| As an example:: | |
| >>> unbalanced_start = ['<div>'] | |
| >>> doc = ['<p>', 'Text', '</p>', '<div>', 'More Text', '</div>'] | |
| >>> pre, post = doc[:3], doc[3:] | |
| >>> pre, post | |
| (['<p>', 'Text', '</p>'], ['<div>', 'More Text', '</div>']) | |
| >>> locate_unbalanced_start(unbalanced_start, pre, post) | |
| >>> pre, post | |
| (['<p>', 'Text', '</p>', '<div>'], ['More Text', '</div>']) | |
| As you can see, we moved the point so that the dangling <div> that | |
| we found will be effectively replaced by the div in the original | |
| document. If this doesn't work out, we just throw away | |
| unbalanced_start without doing anything. | |
| """ | |
| while 1: | |
| if not unbalanced_start: | |
| # We have totally succeeded in finding the position | |
| break | |
| finding = unbalanced_start[0] | |
| finding_name = finding.split()[0].strip('<>') | |
| if not post_delete: | |
| break | |
| next = post_delete[0] | |
| if next is DEL_START or not next.startswith('<'): | |
| # Reached a word, we can't move the delete text forward | |
| break | |
| if next[1] == '/': | |
| # Reached a closing tag, can we go further? Maybe not... | |
| break | |
| name = next.split()[0].strip('<>') | |
| if name == 'ins': | |
| # Can't move into an insert | |
| break | |
| assert name != 'del', ( | |
| "Unexpected delete tag: %r" % next) | |
| if name == finding_name: | |
| unbalanced_start.pop(0) | |
| pre_delete.append(post_delete.pop(0)) | |
| else: | |
| # Found a tag that doesn't match | |
| break | |
| def locate_unbalanced_end(unbalanced_end, pre_delete, post_delete): | |
| """ like locate_unbalanced_start, except handling end tags and | |
| possibly moving the point earlier in the document. """ | |
| while 1: | |
| if not unbalanced_end: | |
| # Success | |
| break | |
| finding = unbalanced_end[-1] | |
| finding_name = finding.split()[0].strip('<>/') | |
| if not pre_delete: | |
| break | |
| next = pre_delete[-1] | |
| if next is DEL_END or not next.startswith('</'): | |
| # A word or a start tag | |
| break | |
| name = next.split()[0].strip('<>/') | |
| if name == 'ins' or name == 'del': | |
| # Can't move into an insert or delete | |
| break | |
| if name == finding_name: | |
| unbalanced_end.pop() | |
| post_delete.insert(0, pre_delete.pop()) | |
| else: | |
| # Found a tag that doesn't match | |
| break | |
| class token(_unicode): | |
| """ Represents a diffable token, generally a word that is displayed to | |
| the user. Opening tags are attached to this token when they are | |
| adjacent (pre_tags) and closing tags that follow the word | |
| (post_tags). Some exceptions occur when there are empty tags | |
| adjacent to a word, so there may be close tags in pre_tags, or | |
| open tags in post_tags. | |
| We also keep track of whether the word was originally followed by | |
| whitespace, even though we do not want to treat the word as | |
| equivalent to a similar word that does not have a trailing | |
| space.""" | |
| # When this is true, the token will be eliminated from the | |
| # displayed diff if no change has occurred: | |
| hide_when_equal = False | |
| def __new__(cls, text, pre_tags=None, post_tags=None, trailing_whitespace=""): | |
| obj = _unicode.__new__(cls, text) | |
| if pre_tags is not None: | |
| obj.pre_tags = pre_tags | |
| else: | |
| obj.pre_tags = [] | |
| if post_tags is not None: | |
| obj.post_tags = post_tags | |
| else: | |
| obj.post_tags = [] | |
| obj.trailing_whitespace = trailing_whitespace | |
| return obj | |
| def __repr__(self): | |
| return 'token(%s, %r, %r, %r)' % (_unicode.__repr__(self), self.pre_tags, | |
| self.post_tags, self.trailing_whitespace) | |
| def html(self): | |
| return _unicode(self) | |
| class tag_token(token): | |
| """ Represents a token that is actually a tag. Currently this is just | |
| the <img> tag, which takes up visible space just like a word but | |
| is only represented in a document by a tag. """ | |
| def __new__(cls, tag, data, html_repr, pre_tags=None, | |
| post_tags=None, trailing_whitespace=""): | |
| obj = token.__new__(cls, "%s: %s" % (type, data), | |
| pre_tags=pre_tags, | |
| post_tags=post_tags, | |
| trailing_whitespace=trailing_whitespace) | |
| obj.tag = tag | |
| obj.data = data | |
| obj.html_repr = html_repr | |
| return obj | |
| def __repr__(self): | |
| return 'tag_token(%s, %s, html_repr=%s, post_tags=%r, pre_tags=%r, trailing_whitespace=%r)' % ( | |
| self.tag, | |
| self.data, | |
| self.html_repr, | |
| self.pre_tags, | |
| self.post_tags, | |
| self.trailing_whitespace) | |
| def html(self): | |
| return self.html_repr | |
| class href_token(token): | |
| """ Represents the href in an anchor tag. Unlike other words, we only | |
| show the href when it changes. """ | |
| hide_when_equal = True | |
| def html(self): | |
| return ' Link: %s' % self | |
| def tokenize(html, include_hrefs=True): | |
| """ | |
| Parse the given HTML and returns token objects (words with attached tags). | |
| This parses only the content of a page; anything in the head is | |
| ignored, and the <head> and <body> elements are themselves | |
| optional. The content is then parsed by lxml, which ensures the | |
| validity of the resulting parsed document (though lxml may make | |
| incorrect guesses when the markup is particular bad). | |
| <ins> and <del> tags are also eliminated from the document, as | |
| that gets confusing. | |
| If include_hrefs is true, then the href attribute of <a> tags is | |
| included as a special kind of diffable token.""" | |
| if etree.iselement(html): | |
| body_el = html | |
| else: | |
| body_el = parse_html(html, cleanup=True) | |
| # Then we split the document into text chunks for each tag, word, and end tag: | |
| chunks = flatten_el(body_el, skip_tag=True, include_hrefs=include_hrefs) | |
| # Finally re-joining them into token objects: | |
| return fixup_chunks(chunks) | |
| def parse_html(html, cleanup=True): | |
| """ | |
| Parses an HTML fragment, returning an lxml element. Note that the HTML will be | |
| wrapped in a <div> tag that was not in the original document. | |
| If cleanup is true, make sure there's no <head> or <body>, and get | |
| rid of any <ins> and <del> tags. | |
| """ | |
| if cleanup: | |
| # This removes any extra markup or structure like <head>: | |
| html = cleanup_html(html) | |
| return fragment_fromstring(html, create_parent=True) | |
| _body_re = re.compile(r'<body.*?>', re.I|re.S) | |
| _end_body_re = re.compile(r'</body.*?>', re.I|re.S) | |
| _ins_del_re = re.compile(r'</?(ins|del).*?>', re.I|re.S) | |
| def cleanup_html(html): | |
| """ This 'cleans' the HTML, meaning that any page structure is removed | |
| (only the contents of <body> are used, if there is any <body). | |
| Also <ins> and <del> tags are removed. """ | |
| match = _body_re.search(html) | |
| if match: | |
| html = html[match.end():] | |
| match = _end_body_re.search(html) | |
| if match: | |
| html = html[:match.start()] | |
| html = _ins_del_re.sub('', html) | |
| return html | |
| end_whitespace_re = re.compile(r'[ \t\n\r]$') | |
| def split_trailing_whitespace(word): | |
| """ | |
| This function takes a word, such as 'test\n\n' and returns ('test','\n\n') | |
| """ | |
| stripped_length = len(word.rstrip()) | |
| return word[0:stripped_length], word[stripped_length:] | |
| def fixup_chunks(chunks): | |
| """ | |
| This function takes a list of chunks and produces a list of tokens. | |
| """ | |
| tag_accum = [] | |
| cur_word = None | |
| result = [] | |
| for chunk in chunks: | |
| if isinstance(chunk, tuple): | |
| if chunk[0] == 'img': | |
| src = chunk[1] | |
| tag, trailing_whitespace = split_trailing_whitespace(chunk[2]) | |
| cur_word = tag_token('img', src, html_repr=tag, | |
| pre_tags=tag_accum, | |
| trailing_whitespace=trailing_whitespace) | |
| tag_accum = [] | |
| result.append(cur_word) | |
| elif chunk[0] == 'href': | |
| href = chunk[1] | |
| cur_word = href_token(href, pre_tags=tag_accum, trailing_whitespace=" ") | |
| tag_accum = [] | |
| result.append(cur_word) | |
| continue | |
| if is_word(chunk): | |
| chunk, trailing_whitespace = split_trailing_whitespace(chunk) | |
| cur_word = token(chunk, pre_tags=tag_accum, trailing_whitespace=trailing_whitespace) | |
| tag_accum = [] | |
| result.append(cur_word) | |
| elif is_start_tag(chunk): | |
| tag_accum.append(chunk) | |
| elif is_end_tag(chunk): | |
| if tag_accum: | |
| tag_accum.append(chunk) | |
| else: | |
| assert cur_word, ( | |
| "Weird state, cur_word=%r, result=%r, chunks=%r of %r" | |
| % (cur_word, result, chunk, chunks)) | |
| cur_word.post_tags.append(chunk) | |
| else: | |
| assert False | |
| if not result: | |
| return [token('', pre_tags=tag_accum)] | |
| else: | |
| result[-1].post_tags.extend(tag_accum) | |
| return result | |
| # All the tags in HTML that don't require end tags: | |
| empty_tags = ( | |
| 'param', 'img', 'area', 'br', 'basefont', 'input', | |
| 'base', 'meta', 'link', 'col') | |
| block_level_tags = ( | |
| 'address', | |
| 'blockquote', | |
| 'center', | |
| 'dir', | |
| 'div', | |
| 'dl', | |
| 'fieldset', | |
| 'form', | |
| 'h1', | |
| 'h2', | |
| 'h3', | |
| 'h4', | |
| 'h5', | |
| 'h6', | |
| 'hr', | |
| 'isindex', | |
| 'menu', | |
| 'noframes', | |
| 'noscript', | |
| 'ol', | |
| 'p', | |
| 'pre', | |
| 'table', | |
| 'ul', | |
| ) | |
| block_level_container_tags = ( | |
| 'dd', | |
| 'dt', | |
| 'frameset', | |
| 'li', | |
| 'tbody', | |
| 'td', | |
| 'tfoot', | |
| 'th', | |
| 'thead', | |
| 'tr', | |
| ) | |
| def flatten_el(el, include_hrefs, skip_tag=False): | |
| """ Takes an lxml element el, and generates all the text chunks for | |
| that tag. Each start tag is a chunk, each word is a chunk, and each | |
| end tag is a chunk. | |
| If skip_tag is true, then the outermost container tag is | |
| not returned (just its contents).""" | |
| if not skip_tag: | |
| if el.tag == 'img': | |
| yield ('img', el.get('src'), start_tag(el)) | |
| else: | |
| yield start_tag(el) | |
| if el.tag in empty_tags and not el.text and not len(el) and not el.tail: | |
| return | |
| start_words = split_words(el.text) | |
| for word in start_words: | |
| yield html_escape(word) | |
| for child in el: | |
| yield from flatten_el(child, include_hrefs=include_hrefs) | |
| if el.tag == 'a' and el.get('href') and include_hrefs: | |
| yield ('href', el.get('href')) | |
| if not skip_tag: | |
| yield end_tag(el) | |
| end_words = split_words(el.tail) | |
| for word in end_words: | |
| yield html_escape(word) | |
| split_words_re = re.compile(r'\S+(?:\s+|$)', re.U) | |
| def split_words(text): | |
| """ Splits some text into words. Includes trailing whitespace | |
| on each word when appropriate. """ | |
| if not text or not text.strip(): | |
| return [] | |
| words = split_words_re.findall(text) | |
| return words | |
| start_whitespace_re = re.compile(r'^[ \t\n\r]') | |
| def start_tag(el): | |
| """ | |
| The text representation of the start tag for a tag. | |
| """ | |
| return '<%s%s>' % ( | |
| el.tag, ''.join([' %s="%s"' % (name, html_escape(value, True)) | |
| for name, value in el.attrib.items()])) | |
| def end_tag(el): | |
| """ The text representation of an end tag for a tag. Includes | |
| trailing whitespace when appropriate. """ | |
| if el.tail and start_whitespace_re.search(el.tail): | |
| extra = ' ' | |
| else: | |
| extra = '' | |
| return '</%s>%s' % (el.tag, extra) | |
| def is_word(tok): | |
| return not tok.startswith('<') | |
| def is_end_tag(tok): | |
| return tok.startswith('</') | |
| def is_start_tag(tok): | |
| return tok.startswith('<') and not tok.startswith('</') | |
| def fixup_ins_del_tags(html): | |
| """ Given an html string, move any <ins> or <del> tags inside of any | |
| block-level elements, e.g. transform <ins><p>word</p></ins> to | |
| <p><ins>word</ins></p> """ | |
| doc = parse_html(html, cleanup=False) | |
| _fixup_ins_del_tags(doc) | |
| html = serialize_html_fragment(doc, skip_outer=True) | |
| return html | |
| def serialize_html_fragment(el, skip_outer=False): | |
| """ Serialize a single lxml element as HTML. The serialized form | |
| includes the elements tail. | |
| If skip_outer is true, then don't serialize the outermost tag | |
| """ | |
| assert not isinstance(el, basestring), ( | |
| "You should pass in an element, not a string like %r" % el) | |
| html = etree.tostring(el, method="html", encoding=_unicode) | |
| if skip_outer: | |
| # Get rid of the extra starting tag: | |
| html = html[html.find('>')+1:] | |
| # Get rid of the extra end tag: | |
| html = html[:html.rfind('<')] | |
| return html.strip() | |
| else: | |
| return html | |
| def _fixup_ins_del_tags(doc): | |
| """fixup_ins_del_tags that works on an lxml document in-place | |
| """ | |
| for tag in ['ins', 'del']: | |
| for el in doc.xpath('descendant-or-self::%s' % tag): | |
| if not _contains_block_level_tag(el): | |
| continue | |
| _move_el_inside_block(el, tag=tag) | |
| el.drop_tag() | |
| #_merge_element_contents(el) | |
| def _contains_block_level_tag(el): | |
| """True if the element contains any block-level elements, like <p>, <td>, etc. | |
| """ | |
| if el.tag in block_level_tags or el.tag in block_level_container_tags: | |
| return True | |
| for child in el: | |
| if _contains_block_level_tag(child): | |
| return True | |
| return False | |
| def _move_el_inside_block(el, tag): | |
| """ helper for _fixup_ins_del_tags; actually takes the <ins> etc tags | |
| and moves them inside any block-level tags. """ | |
| for child in el: | |
| if _contains_block_level_tag(child): | |
| break | |
| else: | |
| # No block-level tags in any child | |
| children_tag = etree.Element(tag) | |
| children_tag.text = el.text | |
| el.text = None | |
| children_tag.extend(list(el)) | |
| el[:] = [children_tag] | |
| return | |
| for child in list(el): | |
| if _contains_block_level_tag(child): | |
| _move_el_inside_block(child, tag) | |
| if child.tail: | |
| tail_tag = etree.Element(tag) | |
| tail_tag.text = child.tail | |
| child.tail = None | |
| el.insert(el.index(child)+1, tail_tag) | |
| else: | |
| child_tag = etree.Element(tag) | |
| el.replace(child, child_tag) | |
| child_tag.append(child) | |
| if el.text: | |
| text_tag = etree.Element(tag) | |
| text_tag.text = el.text | |
| el.text = None | |
| el.insert(0, text_tag) | |
| def _merge_element_contents(el): | |
| """ | |
| Removes an element, but merges its contents into its place, e.g., | |
| given <p>Hi <i>there!</i></p>, if you remove the <i> element you get | |
| <p>Hi there!</p> | |
| """ | |
| parent = el.getparent() | |
| text = el.text or '' | |
| if el.tail: | |
| if not len(el): | |
| text += el.tail | |
| else: | |
| if el[-1].tail: | |
| el[-1].tail += el.tail | |
| else: | |
| el[-1].tail = el.tail | |
| index = parent.index(el) | |
| if text: | |
| if index == 0: | |
| previous = None | |
| else: | |
| previous = parent[index-1] | |
| if previous is None: | |
| if parent.text: | |
| parent.text += text | |
| else: | |
| parent.text = text | |
| else: | |
| if previous.tail: | |
| previous.tail += text | |
| else: | |
| previous.tail = text | |
| parent[index:index+1] = el.getchildren() | |
| class InsensitiveSequenceMatcher(difflib.SequenceMatcher): | |
| """ | |
| Acts like SequenceMatcher, but tries not to find very small equal | |
| blocks amidst large spans of changes | |
| """ | |
| threshold = 2 | |
| def get_matching_blocks(self): | |
| size = min(len(self.b), len(self.b)) | |
| threshold = min(self.threshold, size / 4) | |
| actual = difflib.SequenceMatcher.get_matching_blocks(self) | |
| return [item for item in actual | |
| if item[2] > threshold | |
| or not item[2]] | |
| if __name__ == '__main__': | |
| from lxml.html import _diffcommand | |
| _diffcommand.main() | |