From 5a3e328cfc34d443db7371cd8955c992c39fb286 Mon Sep 17 00:00:00 2001 From: Kevin Mihelich Date: Sun, 30 Jan 2022 00:52:13 +0000 Subject: [PATCH] extra/qt5-webengine to 5.15.8-3 --- extra/qt5-webengine/PKGBUILD | 9 +- .../qt5-webengine-chromium-python3.patch | 1036 +---------------- .../qt5-webengine/qt5-webengine-ffmpeg5.patch | 150 +++ 3 files changed, 172 insertions(+), 1023 deletions(-) create mode 100644 extra/qt5-webengine/qt5-webengine-ffmpeg5.patch diff --git a/extra/qt5-webengine/PKGBUILD b/extra/qt5-webengine/PKGBUILD index 522a73301..b19e8b1cc 100644 --- a/extra/qt5-webengine/PKGBUILD +++ b/extra/qt5-webengine/PKGBUILD @@ -12,7 +12,7 @@ highmem=1 pkgname=qt5-webengine _qtver=5.15.8 pkgver=${_qtver/-/} -pkgrel=1 +pkgrel=3 arch=('x86_64') url='https://www.qt.io' license=('LGPL3' 'LGPL2.1' 'BSD') @@ -28,6 +28,7 @@ source=(git+https://code.qt.io/qt/qtwebengine.git#tag=v${pkgver}-lts git+https://chromium.googlesource.com/catapult#commit=5eedfe23148a234211ba477f76fc2ea2e8529189 qt5-webengine-python3.patch qt5-webengine-chromium-python3.patch + qt5-webengine-ffmpeg5.patch 0001-ARM-toolchain-fixes.patch 0002-Fix-ARM-skia-ICE.patch 0003-bind-gen-Support-single_process-flag-in-generate_bin.patch @@ -36,11 +37,13 @@ sha256sums=('SKIP' 'SKIP' 'SKIP' '398c996cb5b606695ac93645143df39e23fa67e768b09e0da6dbd37342a43f32' - 'cf7be86f1d301001f91b89455b7edbbd269041e2e196355940c38afc84400826' + 'fda4ff16790799fb285847918a677f4f3f7c0f513d4751f846ffc5aa5d873932' + 'c50d3019626183e753c53a997dc8a55938847543aa3178d4c51f377be741c693' '81b7a62368f40d757b165b88df4813413f30db797e4f93a84dd75e12e1ebd679' '298037fcbc132f1539616cdf6149ad5da104f8e2345a9c1af1e7bf8b0dd52c70' 'b878770648437c9bc24023b1d5a47bcd51382d7142b695864c3379826511bcd9' 'e447f5d2635f8f32914c912d7f99b38726c1541334f3a2c1a8ca2dbde565a7de') +options=(debug) prepare() { mkdir -p build @@ -54,6 +57,8 @@ prepare() { patch -p1 -i "$srcdir"/qt5-webengine-python3.patch # Fix build with Python 3 patch -p1 -d src/3rdparty -i "$srcdir"/qt5-webengine-chromium-python3.patch +# patch -p1 -d src/3rdparty -i "$srcdir"/qt5-webengine-ffmpeg5.patch # Fix build with ffmpeg 5 + # Update catapult for python3 compatibility rm -r src/3rdparty/chromium/third_party/catapult mv "$srcdir"/catapult src/3rdparty/chromium/third_party diff --git a/extra/qt5-webengine/qt5-webengine-chromium-python3.patch b/extra/qt5-webengine/qt5-webengine-chromium-python3.patch index 73999dfd8..c55225b68 100644 --- a/extra/qt5-webengine/qt5-webengine-chromium-python3.patch +++ b/extra/qt5-webengine/qt5-webengine-chromium-python3.patch @@ -1541,1027 +1541,21 @@ index cb05c6c4315..bd5566b03e7 100644 {{symbol_export}}extern const blink::QualifiedName& {{attr|symbol}}Attr; {% endfor %} -diff --git a/chromium/third_party/catapult/common/py_vulcanize/py_vulcanize/parse_html_deps.py b/chromium/third_party/catapult/common/py_vulcanize/py_vulcanize/parse_html_deps.py -index 441edc5ed0d..4574d8b4951 100644 ---- a/chromium/third_party/catapult/common/py_vulcanize/py_vulcanize/parse_html_deps.py -+++ b/chromium/third_party/catapult/common/py_vulcanize/py_vulcanize/parse_html_deps.py -@@ -24,10 +24,16 @@ def _InitBeautifulSoup(): - catapult_path = os.path.abspath( - os.path.join(os.path.dirname(__file__), - os.path.pardir, os.path.pardir, os.path.pardir)) -- bs_path = os.path.join(catapult_path, 'third_party', 'beautifulsoup4') -+ if six.PY3: -+ bs_path = os.path.join(catapult_path, 'third_party', 'beautifulsoup4-4.9.3', 'py3k') -+ else: -+ bs_path = os.path.join(catapult_path, 'third_party', 'beautifulsoup4') - _AddToPathIfNeeded(bs_path) - -- html5lib_path = os.path.join(catapult_path, 'third_party', 'html5lib-python') -+ if six.PY3: -+ html5lib_path = os.path.join(catapult_path, 'third_party', 'html5lib-1.1') -+ else: -+ html5lib_path = os.path.join(catapult_path, 'third_party', 'html5lib-python') - _AddToPathIfNeeded(html5lib_path) - - six_path = os.path.join(catapult_path, 'third_party', 'six') -diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/__init__.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/__init__.py -index 7ba34269af7..85d9135f39d 100644 ---- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/__init__.py -+++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/__init__.py -@@ -16,6 +16,10 @@ documentation: - http://www.crummy.com/software/BeautifulSoup/bs4/doc/ - """ - -+from __future__ import absolute_import -+from __future__ import print_function -+import six -+from six.moves import range - __author__ = "Leonard Richardson (leonardr@segfault.org)" - __version__ = "4.3.2" - __copyright__ = "Copyright (c) 2004-2013 Leonard Richardson" -@@ -135,12 +139,12 @@ class BeautifulSoup(Tag): - "fromEncoding", "from_encoding") - - if len(kwargs) > 0: -- arg = kwargs.keys().pop() -+ arg = list(kwargs.keys()).pop() - raise TypeError( - "__init__() got an unexpected keyword argument '%s'" % arg) - - if builder is None: -- if isinstance(features, basestring): -+ if isinstance(features, six.string_types): - features = [features] - if features is None or len(features) == 0: - features = self.DEFAULT_BUILDER_FEATURES -@@ -164,7 +168,7 @@ class BeautifulSoup(Tag): - # involving passing non-markup to Beautiful Soup. - # Beautiful Soup will still parse the input as markup, - # just in case that's what the user really wants. -- if (isinstance(markup, unicode) -+ if (isinstance(markup, six.text_type) - and not os.path.supports_unicode_filenames): - possible_filename = markup.encode("utf8") - else: -@@ -172,7 +176,7 @@ class BeautifulSoup(Tag): - is_file = False - try: - is_file = os.path.exists(possible_filename) -- except Exception, e: -+ except Exception as e: - # This is almost certainly a problem involving - # characters not valid in filenames on this - # system. Just let it go. -@@ -184,7 +188,7 @@ class BeautifulSoup(Tag): - # TODO: This is ugly but I couldn't get it to work in - # Python 3 otherwise. - if ((isinstance(markup, bytes) and not b' ' in markup) -- or (isinstance(markup, unicode) and not u' ' in markup)): -+ or (isinstance(markup, six.text_type) and not u' ' in markup)): - warnings.warn( - '"%s" looks like a URL. Beautiful Soup is not an HTTP client. You should probably use an HTTP client to get the document behind the URL, and feed that document to Beautiful Soup.' % markup) - -@@ -403,4 +407,4 @@ class FeatureNotFound(ValueError): - if __name__ == '__main__': - import sys - soup = BeautifulSoup(sys.stdin) -- print soup.prettify() -+ print(soup.prettify()) -diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/__init__.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/__init__.py -index 740f5f29cd7..819012dc22c 100644 ---- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/__init__.py -+++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/__init__.py -@@ -1,3 +1,4 @@ -+from __future__ import absolute_import - from collections import defaultdict - import itertools - import sys -@@ -6,6 +7,7 @@ from bs4.element import ( - ContentMetaAttributeValue, - whitespace_re - ) -+import six - - __all__ = [ - 'HTMLTreeBuilder', -@@ -159,7 +161,7 @@ class TreeBuilder(object): - # value is a whitespace-separated list of - # values. Split it into a list. - value = attrs[attr] -- if isinstance(value, basestring): -+ if isinstance(value, six.string_types): - values = whitespace_re.split(value) - else: - # html5lib sometimes calls setAttributes twice -diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/_html5lib.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/_html5lib.py -index 7de36ae75e4..93d53812b0f 100644 ---- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/_html5lib.py -+++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/_html5lib.py -@@ -1,3 +1,5 @@ -+from __future__ import absolute_import -+import six - __all__ = [ - 'HTML5TreeBuilder', - ] -@@ -37,7 +39,7 @@ class HTML5TreeBuilder(HTMLTreeBuilder): - doc = parser.parse(markup, encoding=self.user_specified_encoding) - - # Set the character encoding detected by the tokenizer. -- if isinstance(markup, unicode): -+ if isinstance(markup, six.text_type): - # We need to special-case this because html5lib sets - # charEncoding to UTF-8 if it gets Unicode input. - doc.original_encoding = None -@@ -124,7 +126,7 @@ class Element(html5lib.treebuilders._base.Node): - - def appendChild(self, node): - string_child = child = None -- if isinstance(node, basestring): -+ if isinstance(node, six.string_types): - # Some other piece of code decided to pass in a string - # instead of creating a TextElement object to contain the - # string. -@@ -139,7 +141,7 @@ class Element(html5lib.treebuilders._base.Node): - else: - child = node.element - -- if not isinstance(child, basestring) and child.parent is not None: -+ if not isinstance(child, six.string_types) and child.parent is not None: - node.element.extract() - - if (string_child and self.element.contents -@@ -152,7 +154,7 @@ class Element(html5lib.treebuilders._base.Node): - old_element.replace_with(new_element) - self.soup._most_recent_element = new_element - else: -- if isinstance(node, basestring): -+ if isinstance(node, six.string_types): - # Create a brand new NavigableString from this string. - child = self.soup.new_string(node) - -diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/_htmlparser.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/_htmlparser.py -index ca8d8b892bf..504a901c737 100644 ---- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/_htmlparser.py -+++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/_htmlparser.py -@@ -1,13 +1,18 @@ - """Use the HTMLParser library to parse HTML files that aren't too bad.""" - -+from __future__ import absolute_import -+from six import unichr -+import six - __all__ = [ - 'HTMLParserTreeBuilder', - ] - --from HTMLParser import ( -- HTMLParser, -- HTMLParseError, -- ) -+from six.moves.html_parser import HTMLParser -+if six.PY2: -+ from six.moves.html_parser import HTMLParseError -+else: -+ class HTMLParseError(Exception): -+ pass - import sys - import warnings - -@@ -73,7 +78,7 @@ class BeautifulSoupHTMLParser(HTMLParser): - - try: - data = unichr(real_name) -- except (ValueError, OverflowError), e: -+ except (ValueError, OverflowError) as e: - data = u"\N{REPLACEMENT CHARACTER}" - - self.handle_data(data) -@@ -142,7 +147,7 @@ class HTMLParserTreeBuilder(HTMLTreeBuilder): - declared within markup, whether any characters had to be - replaced with REPLACEMENT CHARACTER). - """ -- if isinstance(markup, unicode): -+ if isinstance(markup, six.text_type): - yield (markup, None, None, False) - return - -@@ -158,7 +163,7 @@ class HTMLParserTreeBuilder(HTMLTreeBuilder): - parser.soup = self.soup - try: - parser.feed(markup) -- except HTMLParseError, e: -+ except HTMLParseError as e: - warnings.warn(RuntimeWarning( - "Python's built-in HTMLParser cannot parse the given document. This is not a bug in Beautiful Soup. The best solution is to install an external parser (lxml or html5lib), and use Beautiful Soup with that parser. See http://www.crummy.com/software/BeautifulSoup/bs4/doc/#installing-a-parser for help.")) - raise e -diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/_lxml.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/_lxml.py -index fa5d49875ea..7a5e612a641 100644 ---- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/_lxml.py -+++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/builder/_lxml.py -@@ -1,3 +1,5 @@ -+from __future__ import absolute_import -+import six - __all__ = [ - 'LXMLTreeBuilderForXML', - 'LXMLTreeBuilder', -@@ -78,12 +80,12 @@ class LXMLTreeBuilderForXML(TreeBuilder): - - Each 4-tuple represents a strategy for parsing the document. - """ -- if isinstance(markup, unicode): -+ if isinstance(markup, six.text_type): - # We were given Unicode. Maybe lxml can parse Unicode on - # this system? - yield markup, None, document_declared_encoding, False - -- if isinstance(markup, unicode): -+ if isinstance(markup, six.text_type): - # No, apparently not. Convert the Unicode to UTF-8 and - # tell lxml to parse it as UTF-8. - yield (markup.encode("utf8"), "utf8", -@@ -102,7 +104,7 @@ class LXMLTreeBuilderForXML(TreeBuilder): - def feed(self, markup): - if isinstance(markup, bytes): - markup = BytesIO(markup) -- elif isinstance(markup, unicode): -+ elif isinstance(markup, six.text_type): - markup = StringIO(markup) - - # Call feed() at least once, even if the markup is empty, -@@ -117,7 +119,7 @@ class LXMLTreeBuilderForXML(TreeBuilder): - if len(data) != 0: - self.parser.feed(data) - self.parser.close() -- except (UnicodeDecodeError, LookupError, etree.ParserError), e: -+ except (UnicodeDecodeError, LookupError, etree.ParserError) as e: - raise ParserRejectedMarkup(str(e)) - - def close(self): -@@ -224,7 +226,7 @@ class LXMLTreeBuilder(HTMLTreeBuilder, LXMLTreeBuilderForXML): - self.parser = self.parser_for(encoding) - self.parser.feed(markup) - self.parser.close() -- except (UnicodeDecodeError, LookupError, etree.ParserError), e: -+ except (UnicodeDecodeError, LookupError, etree.ParserError) as e: - raise ParserRejectedMarkup(str(e)) - - -diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/dammit.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/dammit.py -index 59640b7ce3a..5fad7274b8f 100644 ---- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/dammit.py -+++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/dammit.py -@@ -7,11 +7,14 @@ Feed Parser. It works best on XML and XML, but it does not rewrite the - XML or HTML to reflect a new encoding; that's the tree builder's job. - """ - -+from __future__ import absolute_import - import codecs --from htmlentitydefs import codepoint2name -+from six.moves.html_entities import codepoint2name - import re - import logging - import string -+from six import unichr -+import six - - # Import a library to autodetect character encodings. - chardet_type = None -@@ -340,9 +343,9 @@ class UnicodeDammit: - self.detector = EncodingDetector(markup, override_encodings, is_html) - - # Short-circuit if the data is in Unicode to begin with. -- if isinstance(markup, unicode) or markup == '': -+ if isinstance(markup, six.text_type) or markup == '': - self.markup = markup -- self.unicode_markup = unicode(markup) -+ self.unicode_markup = six.text_type(markup) - self.original_encoding = None - return - -@@ -425,7 +428,7 @@ class UnicodeDammit: - def _to_unicode(self, data, encoding, errors="strict"): - '''Given a string and its encoding, decodes the string into Unicode. - %encoding is a string recognized by encodings.aliases''' -- return unicode(data, encoding, errors) -+ return six.text_type(data, encoding, errors) - - @property - def declared_html_encoding(self): -diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/diagnose.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/diagnose.py -index 4d0b00afadd..ec417abba3a 100644 ---- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/diagnose.py -+++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/diagnose.py -@@ -1,7 +1,9 @@ - """Diagnostic functions, mainly for use when doing tech support.""" -+from __future__ import absolute_import -+from __future__ import print_function - import cProfile - from StringIO import StringIO --from HTMLParser import HTMLParser -+from six.moves.html_parser import HTMLParser - import bs4 - from bs4 import BeautifulSoup, __version__ - from bs4.builder import builder_registry -@@ -14,11 +16,13 @@ import time - import traceback - import sys - import cProfile -+from six.moves import map -+from six.moves import range - - def diagnose(data): - """Diagnostic suite for isolating common problems.""" -- print "Diagnostic running on Beautiful Soup %s" % __version__ -- print "Python version %s" % sys.version -+ print("Diagnostic running on Beautiful Soup %s" % __version__) -+ print("Python version %s" % sys.version) - - basic_parsers = ["html.parser", "html5lib", "lxml"] - for name in basic_parsers: -@@ -27,44 +31,44 @@ def diagnose(data): - break - else: - basic_parsers.remove(name) -- print ( -+ print(( - "I noticed that %s is not installed. Installing it may help." % -- name) -+ name)) - - if 'lxml' in basic_parsers: - basic_parsers.append(["lxml", "xml"]) - from lxml import etree -- print "Found lxml version %s" % ".".join(map(str,etree.LXML_VERSION)) -+ print("Found lxml version %s" % ".".join(map(str,etree.LXML_VERSION))) - - if 'html5lib' in basic_parsers: - import html5lib -- print "Found html5lib version %s" % html5lib.__version__ -+ print("Found html5lib version %s" % html5lib.__version__) - - if hasattr(data, 'read'): - data = data.read() - elif os.path.exists(data): -- print '"%s" looks like a filename. Reading data from the file.' % data -+ print('"%s" looks like a filename. Reading data from the file.' % data) - data = open(data).read() - elif data.startswith("http:") or data.startswith("https:"): -- print '"%s" looks like a URL. Beautiful Soup is not an HTTP client.' % data -- print "You need to use some other library to get the document behind the URL, and feed that document to Beautiful Soup." -+ print('"%s" looks like a URL. Beautiful Soup is not an HTTP client.' % data) -+ print("You need to use some other library to get the document behind the URL, and feed that document to Beautiful Soup.") - return -- print -+ print() - - for parser in basic_parsers: -- print "Trying to parse your markup with %s" % parser -+ print("Trying to parse your markup with %s" % parser) - success = False - try: - soup = BeautifulSoup(data, parser) - success = True -- except Exception, e: -- print "%s could not parse the markup." % parser -+ except Exception as e: -+ print("%s could not parse the markup." % parser) - traceback.print_exc() - if success: -- print "Here's what %s did with the markup:" % parser -- print soup.prettify() -+ print("Here's what %s did with the markup:" % parser) -+ print(soup.prettify()) - -- print "-" * 80 -+ print("-" * 80) - - def lxml_trace(data, html=True, **kwargs): - """Print out the lxml events that occur during parsing. -@@ -74,7 +78,7 @@ def lxml_trace(data, html=True, **kwargs): - """ - from lxml import etree - for event, element in etree.iterparse(StringIO(data), html=html, **kwargs): -- print("%s, %4s, %s" % (event, element.tag, element.text)) -+ print(("%s, %4s, %s" % (event, element.tag, element.text))) - - class AnnouncingParser(HTMLParser): - """Announces HTMLParser parse events, without doing anything else.""" -@@ -156,9 +160,9 @@ def rdoc(num_elements=1000): - - def benchmark_parsers(num_elements=100000): - """Very basic head-to-head performance benchmark.""" -- print "Comparative parser benchmark on Beautiful Soup %s" % __version__ -+ print("Comparative parser benchmark on Beautiful Soup %s" % __version__) - data = rdoc(num_elements) -- print "Generated a large invalid HTML document (%d bytes)." % len(data) -+ print("Generated a large invalid HTML document (%d bytes)." % len(data)) - - for parser in ["lxml", ["lxml", "html"], "html5lib", "html.parser"]: - success = False -@@ -167,24 +171,24 @@ def benchmark_parsers(num_elements=100000): - soup = BeautifulSoup(data, parser) - b = time.time() - success = True -- except Exception, e: -- print "%s could not parse the markup." % parser -+ except Exception as e: -+ print("%s could not parse the markup." % parser) - traceback.print_exc() - if success: -- print "BS4+%s parsed the markup in %.2fs." % (parser, b-a) -+ print("BS4+%s parsed the markup in %.2fs." % (parser, b-a)) - - from lxml import etree - a = time.time() - etree.HTML(data) - b = time.time() -- print "Raw lxml parsed the markup in %.2fs." % (b-a) -+ print("Raw lxml parsed the markup in %.2fs." % (b-a)) - - import html5lib - parser = html5lib.HTMLParser() - a = time.time() - parser.parse(data) - b = time.time() -- print "Raw html5lib parsed the markup in %.2fs." % (b-a) -+ print("Raw html5lib parsed the markup in %.2fs." % (b-a)) - - def profile(num_elements=100000, parser="lxml"): - -diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/element.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/element.py -index da9afdf48ec..bda27c3e9c9 100644 ---- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/element.py -+++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/element.py -@@ -1,8 +1,11 @@ -+from __future__ import absolute_import -+from __future__ import print_function - import collections - import re - import sys - import warnings - from bs4.dammit import EntitySubstitution -+import six - - DEFAULT_OUTPUT_ENCODING = "utf-8" - PY3K = (sys.version_info[0] > 2) -@@ -21,22 +24,22 @@ def _alias(attr): - return alias - - --class NamespacedAttribute(unicode): -+class NamespacedAttribute(six.text_type): - - def __new__(cls, prefix, name, namespace=None): - if name is None: -- obj = unicode.__new__(cls, prefix) -+ obj = six.text_type.__new__(cls, prefix) - elif prefix is None: - # Not really namespaced. -- obj = unicode.__new__(cls, name) -+ obj = six.text_type.__new__(cls, name) - else: -- obj = unicode.__new__(cls, prefix + ":" + name) -+ obj = six.text_type.__new__(cls, prefix + ":" + name) - obj.prefix = prefix - obj.name = name - obj.namespace = namespace - return obj - --class AttributeValueWithCharsetSubstitution(unicode): -+class AttributeValueWithCharsetSubstitution(six.text_type): - """A stand-in object for a character encoding specified in HTML.""" - - class CharsetMetaAttributeValue(AttributeValueWithCharsetSubstitution): -@@ -47,7 +50,7 @@ class CharsetMetaAttributeValue(AttributeValueWithCharsetSubstitution): - """ - - def __new__(cls, original_value): -- obj = unicode.__new__(cls, original_value) -+ obj = six.text_type.__new__(cls, original_value) - obj.original_value = original_value - return obj - -@@ -70,9 +73,9 @@ class ContentMetaAttributeValue(AttributeValueWithCharsetSubstitution): - match = cls.CHARSET_RE.search(original_value) - if match is None: - # No substitution necessary. -- return unicode.__new__(unicode, original_value) -+ return six.text_type.__new__(six.text_type, original_value) - -- obj = unicode.__new__(cls, original_value) -+ obj = six.text_type.__new__(cls, original_value) - obj.original_value = original_value - return obj - -@@ -272,7 +275,7 @@ class PageElement(object): - def insert(self, position, new_child): - if new_child is self: - raise ValueError("Cannot insert a tag into itself.") -- if (isinstance(new_child, basestring) -+ if (isinstance(new_child, six.string_types) - and not isinstance(new_child, NavigableString)): - new_child = NavigableString(new_child) - -@@ -489,7 +492,7 @@ class PageElement(object): - result = (element for element in generator - if isinstance(element, Tag)) - return ResultSet(strainer, result) -- elif isinstance(name, basestring): -+ elif isinstance(name, six.string_types): - # Optimization to find all tags with a given name. - result = (element for element in generator - if isinstance(element, Tag) -@@ -640,7 +643,7 @@ class PageElement(object): - return self.parents - - --class NavigableString(unicode, PageElement): -+class NavigableString(six.text_type, PageElement): - - PREFIX = '' - SUFFIX = '' -@@ -653,15 +656,15 @@ class NavigableString(unicode, PageElement): - passed in to the superclass's __new__ or the superclass won't know - how to handle non-ASCII characters. - """ -- if isinstance(value, unicode): -- return unicode.__new__(cls, value) -- return unicode.__new__(cls, value, DEFAULT_OUTPUT_ENCODING) -+ if isinstance(value, six.text_type): -+ return six.text_type.__new__(cls, value) -+ return six.text_type.__new__(cls, value, DEFAULT_OUTPUT_ENCODING) - - def __copy__(self): - return self - - def __getnewargs__(self): -- return (unicode(self),) -+ return (six.text_type(self),) - - def __getattr__(self, attr): - """text.string gives you text. This is for backwards -@@ -1025,8 +1028,8 @@ class Tag(PageElement): - else: - if isinstance(val, list) or isinstance(val, tuple): - val = ' '.join(val) -- elif not isinstance(val, basestring): -- val = unicode(val) -+ elif not isinstance(val, six.string_types): -+ val = six.text_type(val) - elif ( - isinstance(val, AttributeValueWithCharsetSubstitution) - and eventual_encoding is not None): -@@ -1034,7 +1037,7 @@ class Tag(PageElement): - - text = self.format_string(val, formatter) - decoded = ( -- unicode(key) + '=' -+ six.text_type(key) + '=' - + EntitySubstitution.quoted_attribute_value(text)) - attrs.append(decoded) - close = '' -@@ -1210,16 +1213,16 @@ class Tag(PageElement): - raise ValueError( - 'Final combinator "%s" is missing an argument.' % tokens[-1]) - if self._select_debug: -- print 'Running CSS selector "%s"' % selector -+ print('Running CSS selector "%s"' % selector) - for index, token in enumerate(tokens): - if self._select_debug: -- print ' Considering token "%s"' % token -+ print(' Considering token "%s"' % token) - recursive_candidate_generator = None - tag_name = None - if tokens[index-1] in self._selector_combinators: - # This token was consumed by the previous combinator. Skip it. - if self._select_debug: -- print ' Token was consumed by the previous combinator.' -+ print(' Token was consumed by the previous combinator.') - continue - # Each operation corresponds to a checker function, a rule - # for determining whether a candidate matches the -@@ -1325,14 +1328,14 @@ class Tag(PageElement): - next_token = tokens[index+1] - def recursive_select(tag): - if self._select_debug: -- print ' Calling select("%s") recursively on %s %s' % (next_token, tag.name, tag.attrs) -- print '-' * 40 -+ print(' Calling select("%s") recursively on %s %s' % (next_token, tag.name, tag.attrs)) -+ print('-' * 40) - for i in tag.select(next_token, recursive_candidate_generator): - if self._select_debug: -- print '(Recursive select picked up candidate %s %s)' % (i.name, i.attrs) -+ print('(Recursive select picked up candidate %s %s)' % (i.name, i.attrs)) - yield i - if self._select_debug: -- print '-' * 40 -+ print('-' * 40) - _use_candidate_generator = recursive_select - elif _candidate_generator is None: - # By default, a tag's candidates are all of its -@@ -1343,7 +1346,7 @@ class Tag(PageElement): - check = "[any]" - else: - check = tag_name -- print ' Default candidate generator, tag name="%s"' % check -+ print(' Default candidate generator, tag name="%s"' % check) - if self._select_debug: - # This is redundant with later code, but it stops - # a bunch of bogus tags from cluttering up the -@@ -1365,8 +1368,8 @@ class Tag(PageElement): - new_context_ids = set([]) - for tag in current_context: - if self._select_debug: -- print " Running candidate generator on %s %s" % ( -- tag.name, repr(tag.attrs)) -+ print(" Running candidate generator on %s %s" % ( -+ tag.name, repr(tag.attrs))) - for candidate in _use_candidate_generator(tag): - if not isinstance(candidate, Tag): - continue -@@ -1381,21 +1384,21 @@ class Tag(PageElement): - break - if checker is None or result: - if self._select_debug: -- print " SUCCESS %s %s" % (candidate.name, repr(candidate.attrs)) -+ print(" SUCCESS %s %s" % (candidate.name, repr(candidate.attrs))) - if id(candidate) not in new_context_ids: - # If a tag matches a selector more than once, - # don't include it in the context more than once. - new_context.append(candidate) - new_context_ids.add(id(candidate)) - elif self._select_debug: -- print " FAILURE %s %s" % (candidate.name, repr(candidate.attrs)) -+ print(" FAILURE %s %s" % (candidate.name, repr(candidate.attrs))) - - current_context = new_context - - if self._select_debug: -- print "Final verdict:" -+ print("Final verdict:") - for i in current_context: -- print " %s %s" % (i.name, i.attrs) -+ print(" %s %s" % (i.name, i.attrs)) - return current_context - - # Old names for backwards compatibility -@@ -1448,7 +1451,7 @@ class SoupStrainer(object): - def _normalize_search_value(self, value): - # Leave it alone if it's a Unicode string, a callable, a - # regular expression, a boolean, or None. -- if (isinstance(value, unicode) or callable(value) or hasattr(value, 'match') -+ if (isinstance(value, six.text_type) or callable(value) or hasattr(value, 'match') - or isinstance(value, bool) or value is None): - return value - -@@ -1461,7 +1464,7 @@ class SoupStrainer(object): - new_value = [] - for v in value: - if (hasattr(v, '__iter__') and not isinstance(v, bytes) -- and not isinstance(v, unicode)): -+ and not isinstance(v, six.text_type)): - # This is almost certainly the user's mistake. In the - # interests of avoiding infinite loops, we'll let - # it through as-is rather than doing a recursive call. -@@ -1473,7 +1476,7 @@ class SoupStrainer(object): - # Otherwise, convert it into a Unicode string. - # The unicode(str()) thing is so this will do the same thing on Python 2 - # and Python 3. -- return unicode(str(value)) -+ return six.text_type(str(value)) - - def __str__(self): - if self.text: -@@ -1527,7 +1530,7 @@ class SoupStrainer(object): - found = None - # If given a list of items, scan it for a text element that - # matches. -- if hasattr(markup, '__iter__') and not isinstance(markup, (Tag, basestring)): -+ if hasattr(markup, '__iter__') and not isinstance(markup, (Tag, six.string_types)): - for element in markup: - if isinstance(element, NavigableString) \ - and self.search(element): -@@ -1540,7 +1543,7 @@ class SoupStrainer(object): - found = self.search_tag(markup) - # If it's text, make sure the text matches. - elif isinstance(markup, NavigableString) or \ -- isinstance(markup, basestring): -+ isinstance(markup, six.string_types): - if not self.name and not self.attrs and self._matches(markup, self.text): - found = markup - else: -@@ -1554,7 +1557,7 @@ class SoupStrainer(object): - if isinstance(markup, list) or isinstance(markup, tuple): - # This should only happen when searching a multi-valued attribute - # like 'class'. -- if (isinstance(match_against, unicode) -+ if (isinstance(match_against, six.text_type) - and ' ' in match_against): - # A bit of a special case. If they try to match "foo - # bar" on a multivalue attribute's value, only accept -@@ -1589,7 +1592,7 @@ class SoupStrainer(object): - # None matches None, False, an empty string, an empty list, and so on. - return not match_against - -- if isinstance(match_against, unicode): -+ if isinstance(match_against, six.text_type): - # Exact string match - return markup == match_against - -diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/testing.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/testing.py -index fd4495ac58c..4d94c392373 100644 ---- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/testing.py -+++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/testing.py -@@ -1,5 +1,6 @@ - """Helper classes for tests.""" - -+from __future__ import absolute_import - import copy - import functools - import unittest -@@ -14,6 +15,7 @@ from bs4.element import ( - ) - - from bs4.builder import HTMLParserTreeBuilder -+import six - default_builder = HTMLParserTreeBuilder - - -@@ -501,7 +503,7 @@ class XMLTreeBuilderSmokeTest(object): - markup = 'b2012-07-02T20:33:42Zcd' - soup = self.soup(markup) - self.assertEqual( -- unicode(soup.rss), markup) -+ six.text_type(soup.rss), markup) - - def test_docstring_includes_correct_encoding(self): - soup = self.soup("") -@@ -532,17 +534,17 @@ class XMLTreeBuilderSmokeTest(object): - def test_closing_namespaced_tag(self): - markup = '

20010504

' - soup = self.soup(markup) -- self.assertEqual(unicode(soup.p), markup) -+ self.assertEqual(six.text_type(soup.p), markup) - - def test_namespaced_attributes(self): - markup = '' - soup = self.soup(markup) -- self.assertEqual(unicode(soup.foo), markup) -+ self.assertEqual(six.text_type(soup.foo), markup) - - def test_namespaced_attributes_xml_namespace(self): - markup = 'bar' - soup = self.soup(markup) -- self.assertEqual(unicode(soup.foo), markup) -+ self.assertEqual(six.text_type(soup.foo), markup) - - class HTML5TreeBuilderSmokeTest(HTMLTreeBuilderSmokeTest): - """Smoke test for a tree builder that supports HTML5.""" -diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_builder_registry.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_builder_registry.py -index 92ad10fb044..7c84fa3eda2 100644 ---- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_builder_registry.py -+++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_builder_registry.py -@@ -1,5 +1,6 @@ - """Tests of the builder registry.""" - -+from __future__ import absolute_import - import unittest - - from bs4 import BeautifulSoup -diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_docs.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_docs.py -index 5b9f677093a..01eb94ef495 100644 ---- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_docs.py -+++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_docs.py -@@ -2,6 +2,7 @@ - - # pylint: disable-msg=E0611,W0142 - -+from __future__ import absolute_import - __metaclass__ = type - __all__ = [ - 'additional_tests', -diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_html5lib.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_html5lib.py -index 594c3e1f260..31a06598e9c 100644 ---- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_html5lib.py -+++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_html5lib.py -@@ -1,11 +1,12 @@ - """Tests to ensure that the html5lib tree builder generates good trees.""" - -+from __future__ import absolute_import - import warnings - - try: - from bs4.builder import HTML5TreeBuilder - HTML5LIB_PRESENT = True --except ImportError, e: -+except ImportError as e: - HTML5LIB_PRESENT = False - from bs4.element import SoupStrainer - from bs4.testing import ( -diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_htmlparser.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_htmlparser.py -index bcb5ed232f9..9eb4bed6f6c 100644 ---- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_htmlparser.py -+++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_htmlparser.py -@@ -1,6 +1,7 @@ - """Tests to ensure that the html.parser tree builder generates good - trees.""" - -+from __future__ import absolute_import - from bs4.testing import SoupTest, HTMLTreeBuilderSmokeTest - from bs4.builder import HTMLParserTreeBuilder - -diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_lxml.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_lxml.py -index 2b2e9b7e780..20ba5ee4f10 100644 ---- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_lxml.py -+++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_lxml.py -@@ -1,13 +1,15 @@ - """Tests to ensure that the lxml tree builder generates good trees.""" - -+from __future__ import absolute_import - import re - import warnings -+import six - - try: - import lxml.etree - LXML_PRESENT = True - LXML_VERSION = lxml.etree.LXML_VERSION --except ImportError, e: -+except ImportError as e: - LXML_PRESENT = False - LXML_VERSION = (0,) - -@@ -62,7 +64,7 @@ class LXMLTreeBuilderSmokeTest(SoupTest, HTMLTreeBuilderSmokeTest): - # if one is installed. - with warnings.catch_warnings(record=True) as w: - soup = BeautifulStoneSoup("") -- self.assertEqual(u"", unicode(soup.b)) -+ self.assertEqual(u"", six.text_type(soup.b)) - self.assertTrue("BeautifulStoneSoup class is deprecated" in str(w[0].message)) - - def test_real_xhtml_document(self): -diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_soup.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_soup.py -index 47ac245f99a..0b42e4f975a 100644 ---- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_soup.py -+++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_soup.py -@@ -1,6 +1,7 @@ - # -*- coding: utf-8 -*- - """Tests of Beautiful Soup as a whole.""" - -+from __future__ import absolute_import - import logging - import unittest - import sys -@@ -26,11 +27,12 @@ from bs4.testing import ( - skipIf, - ) - import warnings -+import six - - try: - from bs4.builder import LXMLTreeBuilder, LXMLTreeBuilderForXML - LXML_PRESENT = True --except ImportError, e: -+except ImportError as e: - LXML_PRESENT = False - - PYTHON_2_PRE_2_7 = (sys.version_info < (2,7)) -@@ -204,7 +206,7 @@ class TestEncodingConversion(SoupTest): - ascii = b"a" - soup_from_ascii = self.soup(ascii) - unicode_output = soup_from_ascii.decode() -- self.assertTrue(isinstance(unicode_output, unicode)) -+ self.assertTrue(isinstance(unicode_output, six.text_type)) - self.assertEqual(unicode_output, self.document_for(ascii.decode())) - self.assertEqual(soup_from_ascii.original_encoding.lower(), "utf-8") - finally: -diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_tree.py b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_tree.py -index f8515c0ea1c..edb2f511628 100644 ---- a/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_tree.py -+++ b/chromium/third_party/catapult/third_party/beautifulsoup4/bs4/tests/test_tree.py -@@ -9,6 +9,7 @@ same markup, but all Beautiful Soup trees can be traversed with the - methods tested here. - """ - -+from __future__ import absolute_import - import copy - import pickle - import re -@@ -30,6 +31,7 @@ from bs4.testing import ( - SoupTest, - skipIf, - ) -+import six - - XML_BUILDER_PRESENT = (builder_registry.lookup("xml") is not None) - LXML_PRESENT = (builder_registry.lookup("lxml") is not None) -@@ -1393,7 +1395,7 @@ class TestSubstitutions(SoupTest): - - def test_prettify_outputs_unicode_by_default(self): - soup = self.soup("") -- self.assertEqual(unicode, type(soup.prettify())) -+ self.assertEqual(six.text_type, type(soup.prettify())) - - def test_prettify_can_encode_data(self): - soup = self.soup("") -diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/doc/source/conf.py b/chromium/third_party/catapult/third_party/beautifulsoup4/doc/source/conf.py -index 102c3cf972a..4bad2e9f0f9 100644 ---- a/chromium/third_party/catapult/third_party/beautifulsoup4/doc/source/conf.py -+++ b/chromium/third_party/catapult/third_party/beautifulsoup4/doc/source/conf.py -@@ -11,6 +11,7 @@ - # All configuration values have a default; values that are commented out - # serve to show the default. - -+from __future__ import absolute_import - import sys, os - - # If extensions (or modules to document with autodoc) are in another directory, -diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/scripts/demonstrate_parser_differences.py b/chromium/third_party/catapult/third_party/beautifulsoup4/scripts/demonstrate_parser_differences.py -index d84670a53a6..c62f06d3ef8 100644 ---- a/chromium/third_party/catapult/third_party/beautifulsoup4/scripts/demonstrate_parser_differences.py -+++ b/chromium/third_party/catapult/third_party/beautifulsoup4/scripts/demonstrate_parser_differences.py -@@ -14,6 +14,8 @@ help you choose a parser, or understand why Beautiful Soup presents - your document the way it does. - """ - -+from __future__ import absolute_import -+from __future__ import print_function - import os - import sys - from bs4 import BeautifulSoup -@@ -22,13 +24,13 @@ parsers = ['html.parser'] - try: - from bs4.builder import _lxml - parsers.append('lxml') --except ImportError, e: -+except ImportError as e: - pass - - try: - from bs4.builder import _html5lib - parsers.append('html5lib') --except ImportError, e: -+except ImportError as e: - pass - - class Demonstration(object): -@@ -47,7 +49,7 @@ class Demonstration(object): - output = soup.div - else: - output = soup -- except Exception, e: -+ except Exception as e: - output = "[EXCEPTION] %s" % str(e) - self.results[parser] = output - if previous_output is None: -@@ -57,15 +59,15 @@ class Demonstration(object): - return uniform_results - - def dump(self): -- print "%s: %s" % ("Markup".rjust(13), self.markup.encode("utf8")) -+ print("%s: %s" % ("Markup".rjust(13), self.markup.encode("utf8"))) - for parser, output in self.results.items(): -- print "%s: %s" % (parser.rjust(13), output.encode("utf8")) -+ print("%s: %s" % (parser.rjust(13), output.encode("utf8"))) - - different_results = [] - uniform_results = [] - --print "= Testing the following parsers: %s =" % ", ".join(parsers) --print -+print("= Testing the following parsers: %s =" % ", ".join(parsers)) -+print() - - input_file = sys.stdin - if sys.stdin.isatty(): -@@ -83,13 +85,13 @@ for markup in input_file: - else: - different_results.append(demo) - --print "== Markup that's handled the same in every parser ==" --print -+print("== Markup that's handled the same in every parser ==") -+print() - for demo in uniform_results: - demo.dump() -- print --print "== Markup that's not handled the same in every parser ==" --print -+ print() -+print("== Markup that's not handled the same in every parser ==") -+print() - for demo in different_results: - demo.dump() -- print -+ print() -diff --git a/chromium/third_party/catapult/third_party/beautifulsoup4/setup.py b/chromium/third_party/catapult/third_party/beautifulsoup4/setup.py -index 0142ea00e86..c895096fd62 100644 ---- a/chromium/third_party/catapult/third_party/beautifulsoup4/setup.py -+++ b/chromium/third_party/catapult/third_party/beautifulsoup4/setup.py -@@ -1,3 +1,4 @@ -+from __future__ import absolute_import - from distutils.core import setup - - try: +diff --git a/chromium/third_party/dawn/generator/generator_lib.py b/chromium/third_party/dawn/generator/generator_lib.py +index 5e3734d7833..e3d46bd194f 100644 +--- a/chromium/third_party/dawn/generator/generator_lib.py ++++ b/chromium/third_party/dawn/generator/generator_lib.py +@@ -201,6 +201,10 @@ def _compute_python_dependencies(root_dir=None): + + paths = set() + for path in module_paths: ++ # Builtin/namespaced modules may return None for the file path. ++ if not path: ++ continue ++ + path = os.path.abspath(path) + + if not path.startswith(root_dir): diff --git a/chromium/third_party/devtools-frontend/src/BUILD.gn b/chromium/third_party/devtools-frontend/src/BUILD.gn index cd488e88b60..ea1dc3d9a79 100644 --- a/chromium/third_party/devtools-frontend/src/BUILD.gn diff --git a/extra/qt5-webengine/qt5-webengine-ffmpeg5.patch b/extra/qt5-webengine/qt5-webengine-ffmpeg5.patch new file mode 100644 index 000000000..58e51d0f4 --- /dev/null +++ b/extra/qt5-webengine/qt5-webengine-ffmpeg5.patch @@ -0,0 +1,150 @@ +diff --git a/chromium/media/ffmpeg/ffmpeg_common.h b/chromium/media/ffmpeg/ffmpeg_common.h +index 2734a485cbd..70b1877a43c 100644 +--- a/chromium/media/ffmpeg/ffmpeg_common.h ++++ b/chromium/media/ffmpeg/ffmpeg_common.h +@@ -29,6 +29,7 @@ extern "C" { + #include + #include + #include ++#include + #include + #include + #include +diff --git a/chromium/media/filters/audio_file_reader.cc b/chromium/media/filters/audio_file_reader.cc +index cb81d920def..bd73908d0ca 100644 +--- a/chromium/media/filters/audio_file_reader.cc ++++ b/chromium/media/filters/audio_file_reader.cc +@@ -85,7 +85,7 @@ bool AudioFileReader::OpenDemuxer() { + } + + bool AudioFileReader::OpenDecoder() { +- AVCodec* codec = avcodec_find_decoder(codec_context_->codec_id); ++ const AVCodec* codec = avcodec_find_decoder(codec_context_->codec_id); + if (codec) { + // MP3 decodes to S16P which we don't support, tell it to use S16 instead. + if (codec_context_->sample_fmt == AV_SAMPLE_FMT_S16P) +diff --git a/chromium/media/filters/ffmpeg_audio_decoder.cc b/chromium/media/filters/ffmpeg_audio_decoder.cc +index 0d825ed791b..72fac6167ef 100644 +--- a/chromium/media/filters/ffmpeg_audio_decoder.cc ++++ b/chromium/media/filters/ffmpeg_audio_decoder.cc +@@ -329,7 +329,7 @@ bool FFmpegAudioDecoder::ConfigureDecoder(const AudioDecoderConfig& config) { + } + } + +- AVCodec* codec = avcodec_find_decoder(codec_context_->codec_id); ++ const AVCodec* codec = avcodec_find_decoder(codec_context_->codec_id); + if (!codec || + avcodec_open2(codec_context_.get(), codec, &codec_options) < 0) { + DLOG(ERROR) << "Could not initialize audio decoder: " +diff --git a/chromium/media/filters/ffmpeg_demuxer.cc b/chromium/media/filters/ffmpeg_demuxer.cc +index d34db63f3ef..427565b00c1 100644 +--- a/chromium/media/filters/ffmpeg_demuxer.cc ++++ b/chromium/media/filters/ffmpeg_demuxer.cc +@@ -98,12 +98,12 @@ static base::TimeDelta ExtractStartTime(AVStream* stream) { + + // Next try to use the first DTS value, for codecs where we know PTS == DTS + // (excludes all H26x codecs). The start time must be returned in PTS. +- if (stream->first_dts != kNoFFmpegTimestamp && ++ if (av_stream_get_first_dts(stream) != kNoFFmpegTimestamp && + stream->codecpar->codec_id != AV_CODEC_ID_HEVC && + stream->codecpar->codec_id != AV_CODEC_ID_H264 && + stream->codecpar->codec_id != AV_CODEC_ID_MPEG4) { + const base::TimeDelta first_pts = +- ConvertFromTimeBase(stream->time_base, stream->first_dts); ++ ConvertFromTimeBase(stream->time_base, av_stream_get_first_dts(stream)); + if (first_pts < start_time) + start_time = first_pts; + } +@@ -408,11 +408,11 @@ void FFmpegDemuxerStream::EnqueuePacket(ScopedAVPacket packet) { + scoped_refptr buffer; + + if (type() == DemuxerStream::TEXT) { +- int id_size = 0; ++ size_t id_size = 0; + uint8_t* id_data = av_packet_get_side_data( + packet.get(), AV_PKT_DATA_WEBVTT_IDENTIFIER, &id_size); + +- int settings_size = 0; ++ size_t settings_size = 0; + uint8_t* settings_data = av_packet_get_side_data( + packet.get(), AV_PKT_DATA_WEBVTT_SETTINGS, &settings_size); + +@@ -424,7 +424,7 @@ void FFmpegDemuxerStream::EnqueuePacket(ScopedAVPacket packet) { + buffer = DecoderBuffer::CopyFrom(packet->data, packet->size, + side_data.data(), side_data.size()); + } else { +- int side_data_size = 0; ++ size_t side_data_size = 0; + uint8_t* side_data = av_packet_get_side_data( + packet.get(), AV_PKT_DATA_MATROSKA_BLOCKADDITIONAL, &side_data_size); + +@@ -485,7 +485,7 @@ void FFmpegDemuxerStream::EnqueuePacket(ScopedAVPacket packet) { + packet->size - data_offset); + } + +- int skip_samples_size = 0; ++ size_t skip_samples_size = 0; + const uint32_t* skip_samples_ptr = + reinterpret_cast(av_packet_get_side_data( + packet.get(), AV_PKT_DATA_SKIP_SAMPLES, &skip_samples_size)); +diff --git a/chromium/media/filters/ffmpeg_glue.cc b/chromium/media/filters/ffmpeg_glue.cc +index 0ef3521473d..8483ecc348f 100644 +--- a/chromium/media/filters/ffmpeg_glue.cc ++++ b/chromium/media/filters/ffmpeg_glue.cc +@@ -59,7 +59,6 @@ static int64_t AVIOSeekOperation(void* opaque, int64_t offset, int whence) { + } + + void FFmpegGlue::InitializeFFmpeg() { +- av_register_all(); + } + + static void LogContainer(bool is_local_file, +@@ -95,9 +94,6 @@ FFmpegGlue::FFmpegGlue(FFmpegURLProtocol* protocol) { + // Enable fast, but inaccurate seeks for MP3. + format_context_->flags |= AVFMT_FLAG_FAST_SEEK; + +- // Ensures we can read out various metadata bits like vp8 alpha. +- format_context_->flags |= AVFMT_FLAG_KEEP_SIDE_DATA; +- + // Ensures format parsing errors will bail out. From an audit on 11/2017, all + // instances were real failures. Solves bugs like http://crbug.com/710791. + format_context_->error_recognition |= AV_EF_EXPLODE; +diff --git a/chromium/media/filters/ffmpeg_video_decoder.cc b/chromium/media/filters/ffmpeg_video_decoder.cc +index ef12477ee89..7996606f5f9 100644 +--- a/chromium/media/filters/ffmpeg_video_decoder.cc ++++ b/chromium/media/filters/ffmpeg_video_decoder.cc +@@ -391,7 +391,7 @@ bool FFmpegVideoDecoder::ConfigureDecoder(const VideoDecoderConfig& config, + if (decode_nalus_) + codec_context_->flags2 |= AV_CODEC_FLAG2_CHUNKS; + +- AVCodec* codec = avcodec_find_decoder(codec_context_->codec_id); ++ const AVCodec* codec = avcodec_find_decoder(codec_context_->codec_id); + if (!codec || avcodec_open2(codec_context_.get(), codec, NULL) < 0) { + ReleaseFFmpegResources(); + return false; +diff --git a/chromium/media/filters/media_file_checker.cc b/chromium/media/filters/media_file_checker.cc +index 59c2a2fc618..1a9872c7acb 100644 +--- a/chromium/media/filters/media_file_checker.cc ++++ b/chromium/media/filters/media_file_checker.cc +@@ -68,7 +68,7 @@ bool MediaFileChecker::Start(base::TimeDelta check_time) { + auto context = AVStreamToAVCodecContext(format_context->streams[i]); + if (!context) + continue; +- AVCodec* codec = avcodec_find_decoder(cp->codec_id); ++ const AVCodec* codec = avcodec_find_decoder(cp->codec_id); + if (codec && avcodec_open2(context.get(), codec, nullptr) >= 0) { + auto loop = std::make_unique(context.get()); + stream_contexts[i] = {std::move(context), std::move(loop)}; +diff --git a/chromium/third_party/webrtc/modules/video_coding/codecs/h264/h264_decoder_impl.cc b/chromium/third_party/webrtc/modules/video_coding/codecs/h264/h264_decoder_impl.cc +index 9002b874611..d12fade8b63 100644 +--- a/chromium/third_party/webrtc/modules/video_coding/codecs/h264/h264_decoder_impl.cc ++++ b/chromium/third_party/webrtc/modules/video_coding/codecs/h264/h264_decoder_impl.cc +@@ -203,7 +203,7 @@ int32_t H264DecoderImpl::InitDecode(const VideoCodec* codec_settings, + // a pointer |this|. + av_context_->opaque = this; + +- AVCodec* codec = avcodec_find_decoder(av_context_->codec_id); ++ const AVCodec* codec = avcodec_find_decoder(av_context_->codec_id); + if (!codec) { + // This is an indication that FFmpeg has not been initialized or it has not + // been compiled/initialized with the correct set of codecs.