[arch-commits] Commit in qt5-webengine/trunk (3 files)
Antonio Rojas
arojas at gemini.archlinux.org
Tue Jan 4 17:51:14 UTC 2022
Date: Tuesday, January 4, 2022 @ 17:51:13
Author: arojas
Revision: 433415
Update to 5.15.8, fix build with python 3.10
Modified:
qt5-webengine/trunk/PKGBUILD
qt5-webengine/trunk/qt5-webengine-chromium-python3.patch
qt5-webengine/trunk/qt5-webengine-python3.patch
--------------------------------------+
PKGBUILD | 8
qt5-webengine-chromium-python3.patch | 1073 ++++++++++++++++++++++++++++++++-
qt5-webengine-python3.patch | 27
3 files changed, 1083 insertions(+), 25 deletions(-)
Modified: PKGBUILD
===================================================================
--- PKGBUILD 2022-01-04 17:29:50 UTC (rev 433414)
+++ PKGBUILD 2022-01-04 17:51:13 UTC (rev 433415)
@@ -2,9 +2,9 @@
# Contributor: Andrea Scarpino <andrea at archlinux.org>
pkgname=qt5-webengine
-_qtver=5.15.7
+_qtver=5.15.8
pkgver=${_qtver/-/}
-pkgrel=2
+pkgrel=1
arch=('x86_64')
url='https://www.qt.io'
license=('LGPL3' 'LGPL2.1' 'BSD')
@@ -23,8 +23,8 @@
sha256sums=('SKIP'
'SKIP'
'SKIP'
- '45ccf4ff981a26656e5105259f71a3ce386d4a7112cbf09e41cb61615b355435'
- '8d0327615c6b6a340122640a9ddeb56be185621297a4ead8da4211e37feeaa75')
+ '398c996cb5b606695ac93645143df39e23fa67e768b09e0da6dbd37342a43f32'
+ 'cf7be86f1d301001f91b89455b7edbbd269041e2e196355940c38afc84400826')
prepare() {
mkdir -p build
Modified: qt5-webengine-chromium-python3.patch
===================================================================
--- qt5-webengine-chromium-python3.patch 2022-01-04 17:29:50 UTC (rev 433414)
+++ qt5-webengine-chromium-python3.patch 2022-01-04 17:51:13 UTC (rev 433415)
@@ -1,3 +1,31 @@
+diff --git a/chromium/build/print_python_deps.py b/chromium/build/print_python_deps.py
+index fd29c0972c9..69af247094b 100755
+--- a/chromium/build/print_python_deps.py
++++ b/chromium/build/print_python_deps.py
+@@ -1,4 +1,4 @@
+-#!/usr/bin/python2.7
++#!/usr/bin/python
+ # Copyright 2016 The Chromium Authors. All rights reserved.
+ # Use of this source code is governed by a BSD-style license that can be
+ # found in the LICENSE file.
+@@ -80,7 +80,7 @@ def _GetTargetPythonVersion(module):
+ if shebang.startswith('#!'):
+ # Examples:
+ # '#!/usr/bin/python'
+- # '#!/usr/bin/python2.7'
++ # '#!/usr/bin/python'
+ # '#!/usr/bin/python3'
+ # '#!/usr/bin/env python3'
+ # '#!/usr/bin/env vpython'
+@@ -152,7 +152,7 @@ def main():
+
+ # Trybots run with vpython as default Python, but with a different config
+ # from //.vpython. To make the is_vpython test work, and to match the behavior
+- # of dev machines, the shebang line must be run with python2.7.
++ # of dev machines, the shebang line must be run with python.
+ #
+ # E.g. $HOME/.vpython-root/dd50d3/bin/python
+ # E.g. /b/s/w/ir/cache/vpython/ab5c79/bin/python
diff --git a/chromium/components/resources/protobufs/binary_proto_generator.py b/chromium/components/resources/protobufs/binary_proto_generator.py
index 7422ead9697..16365515f26 100755
--- a/chromium/components/resources/protobufs/binary_proto_generator.py
@@ -1513,21 +1541,1027 @@
{{symbol_export}}extern const blink::QualifiedName& {{attr|symbol}}Attr;
{% endfor %}
-diff --git a/chromium/third_party/dawn/generator/generator_lib.py b/chromium/third_party/dawn/generator/generator_lib.py
-index 5e3734d7833..e3d46bd194f 100644
---- a/chromium/third_party/dawn/generator/generator_lib.py
-+++ b/chromium/third_party/dawn/generator/generator_lib.py
-@@ -201,6 +201,10 @@ def _compute_python_dependencies(root_dir=None):
+diff --git a/chromium/third_party/catapult/common/py_vulcanize/py_vulcanize/parse_html_deps.py b/chromium/third_party/catapult/common/py_vulcanize/py_vulcanize/parse_html_deps.py
+index 441edc5ed0d..4574d8b4951 100644
+--- a/chromium/third_party/catapult/common/py_vulcanize/py_vulcanize/parse_html_deps.py
++++ b/chromium/third_party/catapult/common/py_vulcanize/py_vulcanize/parse_html_deps.py
+@@ -24,10 +24,16 @@ def _InitBeautifulSoup():
+ catapult_path = os.path.abspath(
+ os.path.join(os.path.dirname(__file__),
+ os.path.pardir, os.path.pardir, os.path.pardir))
+- bs_path = os.path.join(catapult_path, 'third_party', 'beautifulsoup4')
++ if six.PY3:
++ bs_path = os.path.join(catapult_path, 'third_party', 'beautifulsoup4-4.9.3', 'py3k')
++ else:
++ bs_path = os.path.join(catapult_path, 'third_party', 'beautifulsoup4')
+ _AddToPathIfNeeded(bs_path)
- paths = set()
- for path in module_paths:
-+ # Builtin/namespaced modules may return None for the file path.
-+ if not path:
-+ continue
-+
- path = os.path.abspath(path)
+- html5lib_path = os.path.join(catapult_path, 'third_party', 'html5lib-python')
++ if six.PY3:
++ html5lib_path = os.path.join(catapult_path, 'third_party', 'html5lib-1.1')
++ else:
++ html5lib_path = os.path.join(catapult_path, 'third_party', 'html5lib-python')
+ _AddToPathIfNeeded(html5lib_path)
- if not path.startswith(root_dir):
+ six_path = os.path.join(catapult_path, 'third_party', 'six')
+diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/__init__.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/__init__.py
+index 7ba34269af7..85d9135f39d 100644
+--- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/__init__.py
++++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/__init__.py
+@@ -16,6 +16,10 @@ documentation:
+ http://www.crummy.com/software/BeautifulSoup/bs4/doc/
+ """
+
++from __future__ import absolute_import
++from __future__ import print_function
++import six
++from six.moves import range
+ __author__ = "Leonard Richardson (leonardr at segfault.org)"
+ __version__ = "4.3.2"
+ __copyright__ = "Copyright (c) 2004-2013 Leonard Richardson"
+@@ -135,12 +139,12 @@ class BeautifulSoup(Tag):
+ "fromEncoding", "from_encoding")
+
+ if len(kwargs) > 0:
+- arg = kwargs.keys().pop()
++ arg = list(kwargs.keys()).pop()
+ raise TypeError(
+ "__init__() got an unexpected keyword argument '%s'" % arg)
+
+ if builder is None:
+- if isinstance(features, basestring):
++ if isinstance(features, six.string_types):
+ features = [features]
+ if features is None or len(features) == 0:
+ features = self.DEFAULT_BUILDER_FEATURES
+@@ -164,7 +168,7 @@ class BeautifulSoup(Tag):
+ # involving passing non-markup to Beautiful Soup.
+ # Beautiful Soup will still parse the input as markup,
+ # just in case that's what the user really wants.
+- if (isinstance(markup, unicode)
++ if (isinstance(markup, six.text_type)
+ and not os.path.supports_unicode_filenames):
+ possible_filename = markup.encode("utf8")
+ else:
+@@ -172,7 +176,7 @@ class BeautifulSoup(Tag):
+ is_file = False
+ try:
+ is_file = os.path.exists(possible_filename)
+- except Exception, e:
++ except Exception as e:
+ # This is almost certainly a problem involving
+ # characters not valid in filenames on this
+ # system. Just let it go.
+@@ -184,7 +188,7 @@ class BeautifulSoup(Tag):
+ # TODO: This is ugly but I couldn't get it to work in
+ # Python 3 otherwise.
+ if ((isinstance(markup, bytes) and not b' ' in markup)
+- or (isinstance(markup, unicode) and not u' ' in markup)):
++ or (isinstance(markup, six.text_type) and not u' ' in markup)):
+ warnings.warn(
+ '"%s" looks like a URL. Beautiful Soup is not an HTTP client. You should probably use an HTTP client to get the document behind the URL, and feed that document to Beautiful Soup.' % markup)
+
+@@ -403,4 +407,4 @@ class FeatureNotFound(ValueError):
+ if __name__ == '__main__':
+ import sys
+ soup = BeautifulSoup(sys.stdin)
+- print soup.prettify()
++ print(soup.prettify())
+diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/__init__.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/__init__.py
+index 740f5f29cd7..819012dc22c 100644
+--- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/__init__.py
++++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/__init__.py
+@@ -1,3 +1,4 @@
++from __future__ import absolute_import
+ from collections import defaultdict
+ import itertools
+ import sys
+@@ -6,6 +7,7 @@ from bs4.element import (
+ ContentMetaAttributeValue,
+ whitespace_re
+ )
++import six
+
+ __all__ = [
+ 'HTMLTreeBuilder',
+@@ -159,7 +161,7 @@ class TreeBuilder(object):
+ # value is a whitespace-separated list of
+ # values. Split it into a list.
+ value = attrs[attr]
+- if isinstance(value, basestring):
++ if isinstance(value, six.string_types):
+ values = whitespace_re.split(value)
+ else:
+ # html5lib sometimes calls setAttributes twice
+diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/_html5lib.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/_html5lib.py
+index 7de36ae75e4..93d53812b0f 100644
+--- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/_html5lib.py
++++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/_html5lib.py
+@@ -1,3 +1,5 @@
++from __future__ import absolute_import
++import six
+ __all__ = [
+ 'HTML5TreeBuilder',
+ ]
+@@ -37,7 +39,7 @@ class HTML5TreeBuilder(HTMLTreeBuilder):
+ doc = parser.parse(markup, encoding=self.user_specified_encoding)
+
+ # Set the character encoding detected by the tokenizer.
+- if isinstance(markup, unicode):
++ if isinstance(markup, six.text_type):
+ # We need to special-case this because html5lib sets
+ # charEncoding to UTF-8 if it gets Unicode input.
+ doc.original_encoding = None
+@@ -124,7 +126,7 @@ class Element(html5lib.treebuilders._base.Node):
+
+ def appendChild(self, node):
+ string_child = child = None
+- if isinstance(node, basestring):
++ if isinstance(node, six.string_types):
+ # Some other piece of code decided to pass in a string
+ # instead of creating a TextElement object to contain the
+ # string.
+@@ -139,7 +141,7 @@ class Element(html5lib.treebuilders._base.Node):
+ else:
+ child = node.element
+
+- if not isinstance(child, basestring) and child.parent is not None:
++ if not isinstance(child, six.string_types) and child.parent is not None:
+ node.element.extract()
+
+ if (string_child and self.element.contents
+@@ -152,7 +154,7 @@ class Element(html5lib.treebuilders._base.Node):
+ old_element.replace_with(new_element)
+ self.soup._most_recent_element = new_element
+ else:
+- if isinstance(node, basestring):
++ if isinstance(node, six.string_types):
+ # Create a brand new NavigableString from this string.
+ child = self.soup.new_string(node)
+
+diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/_htmlparser.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/_htmlparser.py
+index ca8d8b892bf..504a901c737 100644
+--- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/_htmlparser.py
++++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/_htmlparser.py
+@@ -1,13 +1,18 @@
+ """Use the HTMLParser library to parse HTML files that aren't too bad."""
+
++from __future__ import absolute_import
++from six import unichr
++import six
+ __all__ = [
+ 'HTMLParserTreeBuilder',
+ ]
+
+-from HTMLParser import (
+- HTMLParser,
+- HTMLParseError,
+- )
++from six.moves.html_parser import HTMLParser
++if six.PY2:
++ from six.moves.html_parser import HTMLParseError
++else:
++ class HTMLParseError(Exception):
++ pass
+ import sys
+ import warnings
+
+@@ -73,7 +78,7 @@ class BeautifulSoupHTMLParser(HTMLParser):
+
+ try:
+ data = unichr(real_name)
+- except (ValueError, OverflowError), e:
++ except (ValueError, OverflowError) as e:
+ data = u"\N{REPLACEMENT CHARACTER}"
+
+ self.handle_data(data)
+@@ -142,7 +147,7 @@ class HTMLParserTreeBuilder(HTMLTreeBuilder):
+ declared within markup, whether any characters had to be
+ replaced with REPLACEMENT CHARACTER).
+ """
+- if isinstance(markup, unicode):
++ if isinstance(markup, six.text_type):
+ yield (markup, None, None, False)
+ return
+
+@@ -158,7 +163,7 @@ class HTMLParserTreeBuilder(HTMLTreeBuilder):
+ parser.soup = self.soup
+ try:
+ parser.feed(markup)
+- except HTMLParseError, e:
++ except HTMLParseError as e:
+ warnings.warn(RuntimeWarning(
+ "Python's built-in HTMLParser cannot parse the given document. This is not a bug in Beautiful Soup. The best solution is to install an external parser (lxml or html5lib), and use Beautiful Soup with that parser. See http://www.crummy.com/software/BeautifulSoup/bs4/doc/#installing-a-parser for help."))
+ raise e
+diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/_lxml.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/_lxml.py
+index fa5d49875ea..7a5e612a641 100644
+--- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/_lxml.py
++++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/_lxml.py
+@@ -1,3 +1,5 @@
++from __future__ import absolute_import
++import six
+ __all__ = [
+ 'LXMLTreeBuilderForXML',
+ 'LXMLTreeBuilder',
+@@ -78,12 +80,12 @@ class LXMLTreeBuilderForXML(TreeBuilder):
+
+ Each 4-tuple represents a strategy for parsing the document.
+ """
+- if isinstance(markup, unicode):
++ if isinstance(markup, six.text_type):
+ # We were given Unicode. Maybe lxml can parse Unicode on
+ # this system?
+ yield markup, None, document_declared_encoding, False
+
+- if isinstance(markup, unicode):
++ if isinstance(markup, six.text_type):
+ # No, apparently not. Convert the Unicode to UTF-8 and
+ # tell lxml to parse it as UTF-8.
+ yield (markup.encode("utf8"), "utf8",
+@@ -102,7 +104,7 @@ class LXMLTreeBuilderForXML(TreeBuilder):
+ def feed(self, markup):
+ if isinstance(markup, bytes):
+ markup = BytesIO(markup)
+- elif isinstance(markup, unicode):
++ elif isinstance(markup, six.text_type):
+ markup = StringIO(markup)
+
+ # Call feed() at least once, even if the markup is empty,
+@@ -117,7 +119,7 @@ class LXMLTreeBuilderForXML(TreeBuilder):
+ if len(data) != 0:
+ self.parser.feed(data)
+ self.parser.close()
+- except (UnicodeDecodeError, LookupError, etree.ParserError), e:
++ except (UnicodeDecodeError, LookupError, etree.ParserError) as e:
+ raise ParserRejectedMarkup(str(e))
+
+ def close(self):
+@@ -224,7 +226,7 @@ class LXMLTreeBuilder(HTMLTreeBuilder, LXMLTreeBuilderForXML):
+ self.parser = self.parser_for(encoding)
+ self.parser.feed(markup)
+ self.parser.close()
+- except (UnicodeDecodeError, LookupError, etree.ParserError), e:
++ except (UnicodeDecodeError, LookupError, etree.ParserError) as e:
+ raise ParserRejectedMarkup(str(e))
+
+
+diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/dammit.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/dammit.py
+index 59640b7ce3a..5fad7274b8f 100644
+--- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/dammit.py
++++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/dammit.py
+@@ -7,11 +7,14 @@ Feed Parser. It works best on XML and XML, but it does not rewrite the
+ XML or HTML to reflect a new encoding; that's the tree builder's job.
+ """
+
++from __future__ import absolute_import
+ import codecs
+-from htmlentitydefs import codepoint2name
++from six.moves.html_entities import codepoint2name
+ import re
+ import logging
+ import string
++from six import unichr
++import six
+
+ # Import a library to autodetect character encodings.
+ chardet_type = None
+@@ -340,9 +343,9 @@ class UnicodeDammit:
+ self.detector = EncodingDetector(markup, override_encodings, is_html)
+
+ # Short-circuit if the data is in Unicode to begin with.
+- if isinstance(markup, unicode) or markup == '':
++ if isinstance(markup, six.text_type) or markup == '':
+ self.markup = markup
+- self.unicode_markup = unicode(markup)
++ self.unicode_markup = six.text_type(markup)
+ self.original_encoding = None
+ return
+
+@@ -425,7 +428,7 @@ class UnicodeDammit:
+ def _to_unicode(self, data, encoding, errors="strict"):
+ '''Given a string and its encoding, decodes the string into Unicode.
+ %encoding is a string recognized by encodings.aliases'''
+- return unicode(data, encoding, errors)
++ return six.text_type(data, encoding, errors)
+
+ @property
+ def declared_html_encoding(self):
+diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/diagnose.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/diagnose.py
+index 4d0b00afadd..ec417abba3a 100644
+--- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/diagnose.py
++++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/diagnose.py
+@@ -1,7 +1,9 @@
+ """Diagnostic functions, mainly for use when doing tech support."""
++from __future__ import absolute_import
++from __future__ import print_function
+ import cProfile
+ from StringIO import StringIO
+-from HTMLParser import HTMLParser
++from six.moves.html_parser import HTMLParser
+ import bs4
+ from bs4 import BeautifulSoup, __version__
+ from bs4.builder import builder_registry
+@@ -14,11 +16,13 @@ import time
+ import traceback
+ import sys
+ import cProfile
++from six.moves import map
++from six.moves import range
+
+ def diagnose(data):
+ """Diagnostic suite for isolating common problems."""
+- print "Diagnostic running on Beautiful Soup %s" % __version__
+- print "Python version %s" % sys.version
++ print("Diagnostic running on Beautiful Soup %s" % __version__)
++ print("Python version %s" % sys.version)
+
+ basic_parsers = ["html.parser", "html5lib", "lxml"]
+ for name in basic_parsers:
+@@ -27,44 +31,44 @@ def diagnose(data):
+ break
+ else:
+ basic_parsers.remove(name)
+- print (
++ print((
+ "I noticed that %s is not installed. Installing it may help." %
+- name)
++ name))
+
+ if 'lxml' in basic_parsers:
+ basic_parsers.append(["lxml", "xml"])
+ from lxml import etree
+- print "Found lxml version %s" % ".".join(map(str,etree.LXML_VERSION))
++ print("Found lxml version %s" % ".".join(map(str,etree.LXML_VERSION)))
+
+ if 'html5lib' in basic_parsers:
+ import html5lib
+- print "Found html5lib version %s" % html5lib.__version__
++ print("Found html5lib version %s" % html5lib.__version__)
+
+ if hasattr(data, 'read'):
+ data = data.read()
+ elif os.path.exists(data):
+- print '"%s" looks like a filename. Reading data from the file.' % data
++ print('"%s" looks like a filename. Reading data from the file.' % data)
+ data = open(data).read()
+ elif data.startswith("http:") or data.startswith("https:"):
+- print '"%s" looks like a URL. Beautiful Soup is not an HTTP client.' % data
+- print "You need to use some other library to get the document behind the URL, and feed that document to Beautiful Soup."
++ print('"%s" looks like a URL. Beautiful Soup is not an HTTP client.' % data)
++ print("You need to use some other library to get the document behind the URL, and feed that document to Beautiful Soup.")
+ return
+- print
++ print()
+
+ for parser in basic_parsers:
+- print "Trying to parse your markup with %s" % parser
++ print("Trying to parse your markup with %s" % parser)
+ success = False
+ try:
+ soup = BeautifulSoup(data, parser)
+ success = True
+- except Exception, e:
+- print "%s could not parse the markup." % parser
++ except Exception as e:
++ print("%s could not parse the markup." % parser)
+ traceback.print_exc()
+ if success:
+- print "Here's what %s did with the markup:" % parser
+- print soup.prettify()
++ print("Here's what %s did with the markup:" % parser)
++ print(soup.prettify())
+
+- print "-" * 80
++ print("-" * 80)
+
+ def lxml_trace(data, html=True, **kwargs):
+ """Print out the lxml events that occur during parsing.
+@@ -74,7 +78,7 @@ def lxml_trace(data, html=True, **kwargs):
+ """
+ from lxml import etree
+ for event, element in etree.iterparse(StringIO(data), html=html, **kwargs):
+- print("%s, %4s, %s" % (event, element.tag, element.text))
++ print(("%s, %4s, %s" % (event, element.tag, element.text)))
+
+ class AnnouncingParser(HTMLParser):
+ """Announces HTMLParser parse events, without doing anything else."""
+@@ -156,9 +160,9 @@ def rdoc(num_elements=1000):
+
+ def benchmark_parsers(num_elements=100000):
+ """Very basic head-to-head performance benchmark."""
+- print "Comparative parser benchmark on Beautiful Soup %s" % __version__
++ print("Comparative parser benchmark on Beautiful Soup %s" % __version__)
+ data = rdoc(num_elements)
+- print "Generated a large invalid HTML document (%d bytes)." % len(data)
++ print("Generated a large invalid HTML document (%d bytes)." % len(data))
+
+ for parser in ["lxml", ["lxml", "html"], "html5lib", "html.parser"]:
+ success = False
+@@ -167,24 +171,24 @@ def benchmark_parsers(num_elements=100000):
+ soup = BeautifulSoup(data, parser)
+ b = time.time()
+ success = True
+- except Exception, e:
+- print "%s could not parse the markup." % parser
++ except Exception as e:
++ print("%s could not parse the markup." % parser)
+ traceback.print_exc()
+ if success:
+- print "BS4+%s parsed the markup in %.2fs." % (parser, b-a)
++ print("BS4+%s parsed the markup in %.2fs." % (parser, b-a))
+
+ from lxml import etree
+ a = time.time()
+ etree.HTML(data)
+ b = time.time()
+- print "Raw lxml parsed the markup in %.2fs." % (b-a)
++ print("Raw lxml parsed the markup in %.2fs." % (b-a))
+
+ import html5lib
+ parser = html5lib.HTMLParser()
+ a = time.time()
+ parser.parse(data)
+ b = time.time()
+- print "Raw html5lib parsed the markup in %.2fs." % (b-a)
++ print("Raw html5lib parsed the markup in %.2fs." % (b-a))
+
+ def profile(num_elements=100000, parser="lxml"):
+
+diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/element.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/element.py
+index da9afdf48ec..bda27c3e9c9 100644
+--- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/element.py
++++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/element.py
+@@ -1,8 +1,11 @@
++from __future__ import absolute_import
++from __future__ import print_function
+ import collections
+ import re
+ import sys
+ import warnings
+ from bs4.dammit import EntitySubstitution
++import six
+
+ DEFAULT_OUTPUT_ENCODING = "utf-8"
+ PY3K = (sys.version_info[0] > 2)
+@@ -21,22 +24,22 @@ def _alias(attr):
+ return alias
+
+
+-class NamespacedAttribute(unicode):
++class NamespacedAttribute(six.text_type):
+
+ def __new__(cls, prefix, name, namespace=None):
+ if name is None:
+- obj = unicode.__new__(cls, prefix)
++ obj = six.text_type.__new__(cls, prefix)
+ elif prefix is None:
+ # Not really namespaced.
+- obj = unicode.__new__(cls, name)
++ obj = six.text_type.__new__(cls, name)
+ else:
+- obj = unicode.__new__(cls, prefix + ":" + name)
++ obj = six.text_type.__new__(cls, prefix + ":" + name)
+ obj.prefix = prefix
+ obj.name = name
+ obj.namespace = namespace
+ return obj
+
+-class AttributeValueWithCharsetSubstitution(unicode):
++class AttributeValueWithCharsetSubstitution(six.text_type):
+ """A stand-in object for a character encoding specified in HTML."""
+
+ class CharsetMetaAttributeValue(AttributeValueWithCharsetSubstitution):
+@@ -47,7 +50,7 @@ class CharsetMetaAttributeValue(AttributeValueWithCharsetSubstitution):
+ """
+
+ def __new__(cls, original_value):
+- obj = unicode.__new__(cls, original_value)
++ obj = six.text_type.__new__(cls, original_value)
+ obj.original_value = original_value
+ return obj
+
+@@ -70,9 +73,9 @@ class ContentMetaAttributeValue(AttributeValueWithCharsetSubstitution):
+ match = cls.CHARSET_RE.search(original_value)
+ if match is None:
+ # No substitution necessary.
+- return unicode.__new__(unicode, original_value)
++ return six.text_type.__new__(six.text_type, original_value)
+
+- obj = unicode.__new__(cls, original_value)
++ obj = six.text_type.__new__(cls, original_value)
+ obj.original_value = original_value
+ return obj
+
+@@ -272,7 +275,7 @@ class PageElement(object):
+ def insert(self, position, new_child):
+ if new_child is self:
+ raise ValueError("Cannot insert a tag into itself.")
+- if (isinstance(new_child, basestring)
++ if (isinstance(new_child, six.string_types)
+ and not isinstance(new_child, NavigableString)):
+ new_child = NavigableString(new_child)
+
+@@ -489,7 +492,7 @@ class PageElement(object):
+ result = (element for element in generator
+ if isinstance(element, Tag))
+ return ResultSet(strainer, result)
+- elif isinstance(name, basestring):
++ elif isinstance(name, six.string_types):
+ # Optimization to find all tags with a given name.
+ result = (element for element in generator
+ if isinstance(element, Tag)
+@@ -640,7 +643,7 @@ class PageElement(object):
+ return self.parents
+
+
+-class NavigableString(unicode, PageElement):
++class NavigableString(six.text_type, PageElement):
+
+ PREFIX = ''
+ SUFFIX = ''
+@@ -653,15 +656,15 @@ class NavigableString(unicode, PageElement):
+ passed in to the superclass's __new__ or the superclass won't know
+ how to handle non-ASCII characters.
+ """
+- if isinstance(value, unicode):
+- return unicode.__new__(cls, value)
+- return unicode.__new__(cls, value, DEFAULT_OUTPUT_ENCODING)
++ if isinstance(value, six.text_type):
++ return six.text_type.__new__(cls, value)
++ return six.text_type.__new__(cls, value, DEFAULT_OUTPUT_ENCODING)
+
+ def __copy__(self):
+ return self
+
+ def __getnewargs__(self):
+- return (unicode(self),)
++ return (six.text_type(self),)
+
+ def __getattr__(self, attr):
+ """text.string gives you text. This is for backwards
+@@ -1025,8 +1028,8 @@ class Tag(PageElement):
+ else:
+ if isinstance(val, list) or isinstance(val, tuple):
+ val = ' '.join(val)
+- elif not isinstance(val, basestring):
+- val = unicode(val)
++ elif not isinstance(val, six.string_types):
++ val = six.text_type(val)
+ elif (
+ isinstance(val, AttributeValueWithCharsetSubstitution)
+ and eventual_encoding is not None):
+@@ -1034,7 +1037,7 @@ class Tag(PageElement):
+
+ text = self.format_string(val, formatter)
+ decoded = (
+- unicode(key) + '='
++ six.text_type(key) + '='
+ + EntitySubstitution.quoted_attribute_value(text))
+ attrs.append(decoded)
+ close = ''
+@@ -1210,16 +1213,16 @@ class Tag(PageElement):
+ raise ValueError(
+ 'Final combinator "%s" is missing an argument.' % tokens[-1])
+ if self._select_debug:
+- print 'Running CSS selector "%s"' % selector
++ print('Running CSS selector "%s"' % selector)
+ for index, token in enumerate(tokens):
+ if self._select_debug:
+- print ' Considering token "%s"' % token
++ print(' Considering token "%s"' % token)
+ recursive_candidate_generator = None
+ tag_name = None
+ if tokens[index-1] in self._selector_combinators:
+ # This token was consumed by the previous combinator. Skip it.
+ if self._select_debug:
+- print ' Token was consumed by the previous combinator.'
++ print(' Token was consumed by the previous combinator.')
+ continue
+ # Each operation corresponds to a checker function, a rule
+ # for determining whether a candidate matches the
+@@ -1325,14 +1328,14 @@ class Tag(PageElement):
+ next_token = tokens[index+1]
+ def recursive_select(tag):
+ if self._select_debug:
+- print ' Calling select("%s") recursively on %s %s' % (next_token, tag.name, tag.attrs)
+- print '-' * 40
++ print(' Calling select("%s") recursively on %s %s' % (next_token, tag.name, tag.attrs))
++ print('-' * 40)
+ for i in tag.select(next_token, recursive_candidate_generator):
+ if self._select_debug:
+- print '(Recursive select picked up candidate %s %s)' % (i.name, i.attrs)
++ print('(Recursive select picked up candidate %s %s)' % (i.name, i.attrs))
+ yield i
+ if self._select_debug:
+- print '-' * 40
++ print('-' * 40)
+ _use_candidate_generator = recursive_select
+ elif _candidate_generator is None:
+ # By default, a tag's candidates are all of its
+@@ -1343,7 +1346,7 @@ class Tag(PageElement):
+ check = "[any]"
+ else:
+ check = tag_name
+- print ' Default candidate generator, tag name="%s"' % check
++ print(' Default candidate generator, tag name="%s"' % check)
+ if self._select_debug:
+ # This is redundant with later code, but it stops
+ # a bunch of bogus tags from cluttering up the
+@@ -1365,8 +1368,8 @@ class Tag(PageElement):
+ new_context_ids = set([])
+ for tag in current_context:
+ if self._select_debug:
+- print " Running candidate generator on %s %s" % (
+- tag.name, repr(tag.attrs))
++ print(" Running candidate generator on %s %s" % (
++ tag.name, repr(tag.attrs)))
+ for candidate in _use_candidate_generator(tag):
+ if not isinstance(candidate, Tag):
+ continue
+@@ -1381,21 +1384,21 @@ class Tag(PageElement):
+ break
+ if checker is None or result:
+ if self._select_debug:
+- print " SUCCESS %s %s" % (candidate.name, repr(candidate.attrs))
++ print(" SUCCESS %s %s" % (candidate.name, repr(candidate.attrs)))
+ if id(candidate) not in new_context_ids:
+ # If a tag matches a selector more than once,
+ # don't include it in the context more than once.
+ new_context.append(candidate)
+ new_context_ids.add(id(candidate))
+ elif self._select_debug:
+- print " FAILURE %s %s" % (candidate.name, repr(candidate.attrs))
++ print(" FAILURE %s %s" % (candidate.name, repr(candidate.attrs)))
+
+ current_context = new_context
+
+ if self._select_debug:
+- print "Final verdict:"
++ print("Final verdict:")
+ for i in current_context:
+- print " %s %s" % (i.name, i.attrs)
++ print(" %s %s" % (i.name, i.attrs))
+ return current_context
+
+ # Old names for backwards compatibility
+@@ -1448,7 +1451,7 @@ class SoupStrainer(object):
+ def _normalize_search_value(self, value):
+ # Leave it alone if it's a Unicode string, a callable, a
+ # regular expression, a boolean, or None.
+- if (isinstance(value, unicode) or callable(value) or hasattr(value, 'match')
++ if (isinstance(value, six.text_type) or callable(value) or hasattr(value, 'match')
+ or isinstance(value, bool) or value is None):
+ return value
+
+@@ -1461,7 +1464,7 @@ class SoupStrainer(object):
+ new_value = []
+ for v in value:
+ if (hasattr(v, '__iter__') and not isinstance(v, bytes)
+- and not isinstance(v, unicode)):
++ and not isinstance(v, six.text_type)):
+ # This is almost certainly the user's mistake. In the
+ # interests of avoiding infinite loops, we'll let
+ # it through as-is rather than doing a recursive call.
+@@ -1473,7 +1476,7 @@ class SoupStrainer(object):
+ # Otherwise, convert it into a Unicode string.
+ # The unicode(str()) thing is so this will do the same thing on Python 2
+ # and Python 3.
+- return unicode(str(value))
++ return six.text_type(str(value))
+
+ def __str__(self):
+ if self.text:
+@@ -1527,7 +1530,7 @@ class SoupStrainer(object):
+ found = None
+ # If given a list of items, scan it for a text element that
+ # matches.
+- if hasattr(markup, '__iter__') and not isinstance(markup, (Tag, basestring)):
++ if hasattr(markup, '__iter__') and not isinstance(markup, (Tag, six.string_types)):
+ for element in markup:
+ if isinstance(element, NavigableString) \
+ and self.search(element):
+@@ -1540,7 +1543,7 @@ class SoupStrainer(object):
+ found = self.search_tag(markup)
+ # If it's text, make sure the text matches.
+ elif isinstance(markup, NavigableString) or \
+- isinstance(markup, basestring):
++ isinstance(markup, six.string_types):
+ if not self.name and not self.attrs and self._matches(markup, self.text):
+ found = markup
+ else:
+@@ -1554,7 +1557,7 @@ class SoupStrainer(object):
+ if isinstance(markup, list) or isinstance(markup, tuple):
+ # This should only happen when searching a multi-valued attribute
+ # like 'class'.
+- if (isinstance(match_against, unicode)
++ if (isinstance(match_against, six.text_type)
+ and ' ' in match_against):
+ # A bit of a special case. If they try to match "foo
+ # bar" on a multivalue attribute's value, only accept
+@@ -1589,7 +1592,7 @@ class SoupStrainer(object):
+ # None matches None, False, an empty string, an empty list, and so on.
+ return not match_against
+
+- if isinstance(match_against, unicode):
++ if isinstance(match_against, six.text_type):
+ # Exact string match
+ return markup == match_against
+
+diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/testing.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/testing.py
+index fd4495ac58c..4d94c392373 100644
+--- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/testing.py
++++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/testing.py
+@@ -1,5 +1,6 @@
+ """Helper classes for tests."""
+
++from __future__ import absolute_import
+ import copy
+ import functools
+ import unittest
+@@ -14,6 +15,7 @@ from bs4.element import (
+ )
+
+ from bs4.builder import HTMLParserTreeBuilder
++import six
+ default_builder = HTMLParserTreeBuilder
+
+
+@@ -501,7 +503,7 @@ class XMLTreeBuilderSmokeTest(object):
+ markup = '<rss xmlns:dc="foo"><dc:creator>b</dc:creator><dc:date>2012-07-02T20:33:42Z</dc:date><dc:rights>c</dc:rights><image>d</image></rss>'
+ soup = self.soup(markup)
+ self.assertEqual(
+- unicode(soup.rss), markup)
++ six.text_type(soup.rss), markup)
+
+ def test_docstring_includes_correct_encoding(self):
+ soup = self.soup("<root/>")
+@@ -532,17 +534,17 @@ class XMLTreeBuilderSmokeTest(object):
+ def test_closing_namespaced_tag(self):
+ markup = '<p xmlns:dc="http://purl.org/dc/elements/1.1/"><dc:date>20010504</dc:date></p>'
+ soup = self.soup(markup)
+- self.assertEqual(unicode(soup.p), markup)
++ self.assertEqual(six.text_type(soup.p), markup)
+
+ def test_namespaced_attributes(self):
+ markup = '<foo xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"><bar xsi:schemaLocation="http://www.example.com"/></foo>'
+ soup = self.soup(markup)
+- self.assertEqual(unicode(soup.foo), markup)
++ self.assertEqual(six.text_type(soup.foo), markup)
+
+ def test_namespaced_attributes_xml_namespace(self):
+ markup = '<foo xml:lang="fr">bar</foo>'
+ soup = self.soup(markup)
+- self.assertEqual(unicode(soup.foo), markup)
++ self.assertEqual(six.text_type(soup.foo), markup)
+
+ class HTML5TreeBuilderSmokeTest(HTMLTreeBuilderSmokeTest):
+ """Smoke test for a tree builder that supports HTML5."""
+diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_builder_registry.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_builder_registry.py
+index 92ad10fb044..7c84fa3eda2 100644
+--- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_builder_registry.py
++++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_builder_registry.py
+@@ -1,5 +1,6 @@
+ """Tests of the builder registry."""
+
++from __future__ import absolute_import
+ import unittest
+
+ from bs4 import BeautifulSoup
+diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_docs.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_docs.py
+index 5b9f677093a..01eb94ef495 100644
+--- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_docs.py
++++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_docs.py
+@@ -2,6 +2,7 @@
+
+ # pylint: disable-msg=E0611,W0142
+
++from __future__ import absolute_import
+ __metaclass__ = type
+ __all__ = [
+ 'additional_tests',
+diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_html5lib.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_html5lib.py
+index 594c3e1f260..31a06598e9c 100644
+--- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_html5lib.py
++++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_html5lib.py
+@@ -1,11 +1,12 @@
+ """Tests to ensure that the html5lib tree builder generates good trees."""
+
++from __future__ import absolute_import
+ import warnings
+
+ try:
+ from bs4.builder import HTML5TreeBuilder
+ HTML5LIB_PRESENT = True
+-except ImportError, e:
++except ImportError as e:
+ HTML5LIB_PRESENT = False
+ from bs4.element import SoupStrainer
+ from bs4.testing import (
+diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_htmlparser.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_htmlparser.py
+index bcb5ed232f9..9eb4bed6f6c 100644
+--- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_htmlparser.py
++++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_htmlparser.py
+@@ -1,6 +1,7 @@
+ """Tests to ensure that the html.parser tree builder generates good
+ trees."""
+
++from __future__ import absolute_import
+ from bs4.testing import SoupTest, HTMLTreeBuilderSmokeTest
+ from bs4.builder import HTMLParserTreeBuilder
+
+diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_lxml.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_lxml.py
+index 2b2e9b7e780..20ba5ee4f10 100644
+--- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_lxml.py
++++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_lxml.py
+@@ -1,13 +1,15 @@
+ """Tests to ensure that the lxml tree builder generates good trees."""
+
++from __future__ import absolute_import
+ import re
+ import warnings
++import six
+
+ try:
+ import lxml.etree
+ LXML_PRESENT = True
+ LXML_VERSION = lxml.etree.LXML_VERSION
+-except ImportError, e:
++except ImportError as e:
+ LXML_PRESENT = False
+ LXML_VERSION = (0,)
+
+@@ -62,7 +64,7 @@ class LXMLTreeBuilderSmokeTest(SoupTest, HTMLTreeBuilderSmokeTest):
+ # if one is installed.
+ with warnings.catch_warnings(record=True) as w:
+ soup = BeautifulStoneSoup("<b />")
+- self.assertEqual(u"<b/>", unicode(soup.b))
++ self.assertEqual(u"<b/>", six.text_type(soup.b))
+ self.assertTrue("BeautifulStoneSoup class is deprecated" in str(w[0].message))
+
+ def test_real_xhtml_document(self):
+diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_soup.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_soup.py
+index 47ac245f99a..0b42e4f975a 100644
+--- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_soup.py
++++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_soup.py
+@@ -1,6 +1,7 @@
+ # -*- coding: utf-8 -*-
+ """Tests of Beautiful Soup as a whole."""
+
++from __future__ import absolute_import
+ import logging
+ import unittest
+ import sys
+@@ -26,11 +27,12 @@ from bs4.testing import (
+ skipIf,
+ )
+ import warnings
++import six
+
+ try:
+ from bs4.builder import LXMLTreeBuilder, LXMLTreeBuilderForXML
+ LXML_PRESENT = True
+-except ImportError, e:
++except ImportError as e:
+ LXML_PRESENT = False
+
+ PYTHON_2_PRE_2_7 = (sys.version_info < (2,7))
+@@ -204,7 +206,7 @@ class TestEncodingConversion(SoupTest):
+ ascii = b"<foo>a</foo>"
+ soup_from_ascii = self.soup(ascii)
+ unicode_output = soup_from_ascii.decode()
+- self.assertTrue(isinstance(unicode_output, unicode))
++ self.assertTrue(isinstance(unicode_output, six.text_type))
+ self.assertEqual(unicode_output, self.document_for(ascii.decode()))
+ self.assertEqual(soup_from_ascii.original_encoding.lower(), "utf-8")
+ finally:
+diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_tree.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_tree.py
+index f8515c0ea1c..edb2f511628 100644
+--- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_tree.py
++++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_tree.py
+@@ -9,6 +9,7 @@ same markup, but all Beautiful Soup trees can be traversed with the
+ methods tested here.
+ """
+
++from __future__ import absolute_import
+ import copy
+ import pickle
+ import re
+@@ -30,6 +31,7 @@ from bs4.testing import (
+ SoupTest,
+ skipIf,
+ )
++import six
+
+ XML_BUILDER_PRESENT = (builder_registry.lookup("xml") is not None)
+ LXML_PRESENT = (builder_registry.lookup("lxml") is not None)
+@@ -1393,7 +1395,7 @@ class TestSubstitutions(SoupTest):
+
+ def test_prettify_outputs_unicode_by_default(self):
+ soup = self.soup("<a></a>")
+- self.assertEqual(unicode, type(soup.prettify()))
++ self.assertEqual(six.text_type, type(soup.prettify()))
+
+ def test_prettify_can_encode_data(self):
+ soup = self.soup("<a></a>")
+diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/doc/source/conf.py b/chromium/third_party/catapult/third_party/beautifulsoup4/doc/source/conf.py
+index 102c3cf972a..4bad2e9f0f9 100644
+--- a/chromium/third_party/catapult/third_party/beautifulsoup4/doc/source/conf.py
++++ b/chromium/third_party/catapult/third_party/beautifulsoup4/doc/source/conf.py
+@@ -11,6 +11,7 @@
+ # All configuration values have a default; values that are commented out
+ # serve to show the default.
+
++from __future__ import absolute_import
+ import sys, os
+
+ # If extensions (or modules to document with autodoc) are in another directory,
+diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/scripts/demonstrate_parser_differences.py b/chromium/third_party/catapult/third_party/beautifulsoup4/scripts/demonstrate_parser_differences.py
+index d84670a53a6..c62f06d3ef8 100644
+--- a/chromium/third_party/catapult/third_party/beautifulsoup4/scripts/demonstrate_parser_differences.py
++++ b/chromium/third_party/catapult/third_party/beautifulsoup4/scripts/demonstrate_parser_differences.py
+@@ -14,6 +14,8 @@ help you choose a parser, or understand why Beautiful Soup presents
+ your document the way it does.
+ """
+
++from __future__ import absolute_import
++from __future__ import print_function
+ import os
+ import sys
+ from bs4 import BeautifulSoup
+@@ -22,13 +24,13 @@ parsers = ['html.parser']
+ try:
+ from bs4.builder import _lxml
+ parsers.append('lxml')
+-except ImportError, e:
++except ImportError as e:
+ pass
+
+ try:
+ from bs4.builder import _html5lib
+ parsers.append('html5lib')
+-except ImportError, e:
++except ImportError as e:
+ pass
+
+ class Demonstration(object):
+@@ -47,7 +49,7 @@ class Demonstration(object):
+ output = soup.div
+ else:
+ output = soup
+- except Exception, e:
++ except Exception as e:
+ output = "[EXCEPTION] %s" % str(e)
+ self.results[parser] = output
+ if previous_output is None:
+@@ -57,15 +59,15 @@ class Demonstration(object):
+ return uniform_results
+
+ def dump(self):
+- print "%s: %s" % ("Markup".rjust(13), self.markup.encode("utf8"))
++ print("%s: %s" % ("Markup".rjust(13), self.markup.encode("utf8")))
+ for parser, output in self.results.items():
+- print "%s: %s" % (parser.rjust(13), output.encode("utf8"))
++ print("%s: %s" % (parser.rjust(13), output.encode("utf8")))
+
+ different_results = []
+ uniform_results = []
+
+-print "= Testing the following parsers: %s =" % ", ".join(parsers)
+-print
++print("= Testing the following parsers: %s =" % ", ".join(parsers))
++print()
+
+ input_file = sys.stdin
+ if sys.stdin.isatty():
+@@ -83,13 +85,13 @@ for markup in input_file:
+ else:
+ different_results.append(demo)
+
+-print "== Markup that's handled the same in every parser =="
+-print
++print("== Markup that's handled the same in every parser ==")
++print()
+ for demo in uniform_results:
+ demo.dump()
+- print
+-print "== Markup that's not handled the same in every parser =="
+-print
++ print()
++print("== Markup that's not handled the same in every parser ==")
++print()
+ for demo in different_results:
+ demo.dump()
+- print
++ print()
+diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/setup.py b/chromium/third_party/catapult/third_party/beautifulsoup4/setup.py
+index 0142ea00e86..c895096fd62 100644
+--- a/chromium/third_party/catapult/third_party/beautifulsoup4/setup.py
++++ b/chromium/third_party/catapult/third_party/beautifulsoup4/setup.py
+@@ -1,3 +1,4 @@
++from __future__ import absolute_import
+ from distutils.core import setup
+
+ try:
diff --git a/chromium/third_party/devtools-frontend/src/BUILD.gn b/chromium/third_party/devtools-frontend/src/BUILD.gn
index cd488e88b60..ea1dc3d9a79 100644
--- a/chromium/third_party/devtools-frontend/src/BUILD.gn
@@ -1642,6 +2676,19 @@
return json.dumps(result)
def all_compiled_files(self):
+diff --git a/chromium/third_party/jinja2/tests.py b/chromium/third_party/jinja2/tests.py
+index 0adc3d4dbcb..b14f85ff148 100644
+--- a/chromium/third_party/jinja2/tests.py
++++ b/chromium/third_party/jinja2/tests.py
+@@ -10,7 +10,7 @@
+ """
+ import operator
+ import re
+-from collections import Mapping
++from collections.abc import Mapping
+ from jinja2.runtime import Undefined
+ from jinja2._compat import text_type, string_types, integer_types
+ import decimal
diff --git a/chromium/tools/metrics/ukm/gen_builders.py b/chromium/tools/metrics/ukm/gen_builders.py
index f9f61d90a56..44e46fae8cc 100755
--- a/chromium/tools/metrics/ukm/gen_builders.py
Modified: qt5-webengine-python3.patch
===================================================================
--- qt5-webengine-python3.patch 2022-01-04 17:29:50 UTC (rev 433414)
+++ qt5-webengine-python3.patch 2022-01-04 17:51:13 UTC (rev 433415)
@@ -1,18 +1,29 @@
diff --git a/configure.pri b/configure.pri
-index e072961f..5efe30bc 100644
+index 8705ad93..6723bffe 100644
--- a/configure.pri
+++ b/configure.pri
-@@ -12,10 +12,6 @@ defineTest(isPythonVersionSupported) {
- python_version ~= s/[()]//g
- python_version = $$split(python_version, ',')
- python_major_version = $$first(python_version)
+@@ -7,20 +7,7 @@ QTWEBENGINE_SOURCE_TREE = $$PWD
+ equals(QMAKE_HOST.os, Windows): EXE_SUFFIX = .exe
+
+ defineTest(isPythonVersionSupported) {
+- python = $$system_quote($$system_path($$1))
+- python_version = $$system('$$python -c "import sys; print(sys.version_info[0:3])"')
+- python_version ~= s/[()]//g
+- python_version = $$split(python_version, ',')
+- python_major_version = $$first(python_version)
- greaterThan(python_major_version, 2) {
- qtLog("Python version 3 is not supported by Chromium.")
- return(false)
- }
- python_minor_version = $$member(python_version, 1)
- python_patch_version = $$member(python_version, 2)
- greaterThan(python_major_version, 1): greaterThan(python_minor_version, 6): greaterThan(python_patch_version, 4): return(true)
+- python_minor_version = $$member(python_version, 1)
+- python_patch_version = $$member(python_version, 2)
+- greaterThan(python_major_version, 1): greaterThan(python_minor_version, 6): greaterThan(python_patch_version, 4): return(true)
+- qtLog("Unsupported python version: $${python_major_version}.$${python_minor_version}.$${python_patch_version}.")
+- return(false)
++ return(true)
+ }
+
+ defineTest(qtConfTest_detectJumboBuild) {
@@ -52,10 +48,10 @@ defineTest(qtConfReport_jumboBuild) {
qtConfReportPadded($${1}, $$mergeLimit)
}
More information about the arch-commits
mailing list