| # Copyright 2013 The ChromiumOS Authors |
| # Use of this source code is governed by a BSD-style license that can be |
| # found in the LICENSE file. |
| |
| """Verifying the integrity of a Chrome OS update payload. |
| |
| This module is used internally by the main Payload class for verifying the |
| integrity of an update payload. The interface for invoking the checks is as |
| follows: |
| |
| checker = PayloadChecker(payload) |
| checker.Run(...) |
| """ |
| |
| import array |
| import base64 |
| import collections |
| import hashlib |
| import itertools |
| import os |
| import subprocess |
| |
| from update_payload import common |
| from update_payload import error |
| from update_payload import format_utils |
| from update_payload import histogram |
| from update_payload import update_metadata_pb2 |
| |
| |
| # |
| # Constants. |
| # |
| |
| _CHECK_MOVE_SAME_SRC_DST_BLOCK = "move-same-src-dst-block" |
| _CHECK_PAYLOAD_SIG = "payload-sig" |
| CHECKS_TO_DISABLE = ( |
| _CHECK_MOVE_SAME_SRC_DST_BLOCK, |
| _CHECK_PAYLOAD_SIG, |
| ) |
| |
| _TYPE_FULL = "full" |
| _TYPE_DELTA = "delta" |
| |
| _DEFAULT_BLOCK_SIZE = 4096 |
| |
| _DEFAULT_PUBKEY_BASE_NAME = "update-payload-key.pub.pem" |
| _DEFAULT_PUBKEY_FILE_NAME = os.path.join( |
| os.path.dirname(__file__), _DEFAULT_PUBKEY_BASE_NAME |
| ) |
| |
| # Supported minor version map to payload types allowed to be using them. |
| _SUPPORTED_MINOR_VERSIONS = { |
| 0: (_TYPE_FULL,), |
| 2: (_TYPE_DELTA,), |
| 3: (_TYPE_DELTA,), |
| 4: (_TYPE_DELTA,), |
| 5: (_TYPE_DELTA,), |
| 6: (_TYPE_DELTA,), |
| 7: (_TYPE_DELTA,), |
| 8: (_TYPE_DELTA,), |
| 9: (_TYPE_DELTA,), |
| } |
| |
| |
| # |
| # Helper functions. |
| # |
| |
| |
| def _IsPowerOfTwo(val): |
| """Returns True iff val is a power of two.""" |
| return val > 0 and (val & (val - 1)) == 0 |
| |
| |
| def _AddFormat(format_func, value): |
| """Adds a custom formatted representation to ordinary string representation. |
| |
| Args: |
| format_func: A value formatter. |
| value: Value to be formatted and returned. |
| |
| Returns: |
| A string 'x (y)' where x = str(value) and y = format_func(value). |
| """ |
| ret = str(value) |
| formatted_str = format_func(value) |
| if formatted_str: |
| ret += " (%s)" % formatted_str |
| return ret |
| |
| |
| def _AddHumanReadableSize(size): |
| """Adds a human readable representation to a byte size value.""" |
| return _AddFormat(format_utils.BytesToHumanReadable, size) |
| |
| |
| # |
| # Payload report generator. |
| # |
| |
| |
| class _PayloadReport(object): |
| """A payload report generator. |
| |
| A report is essentially a sequence of nodes, which represent data points. It |
| is initialized to have a "global", untitled section. A node may be a |
| sub-report itself. |
| """ |
| |
| # Report nodes: Field, sub-report, section. |
| class Node(object): |
| """A report node interface.""" |
| |
| @staticmethod |
| def _Indent(indent, line): |
| """Indents a line by a given indentation amount. |
| |
| Args: |
| indent: The indentation amount. |
| line: The line content (string). |
| |
| Returns: |
| The properly indented line (string). |
| """ |
| return "%*s%s" % (indent, "", line) |
| |
| def GenerateLines(self, base_indent, sub_indent, curr_section): |
| """Generates the report lines for this node. |
| |
| Args: |
| base_indent: Base indentation for each line. |
| sub_indent: Additional indentation for sub-nodes. |
| curr_section: The current report section object. |
| |
| Returns: |
| A pair consisting of a list of properly indented report lines and a new |
| current section object. |
| """ |
| raise NotImplementedError |
| |
| class FieldNode(Node): |
| """A field report node, representing a (name, value) pair.""" |
| |
| def __init__(self, name, value, linebreak, indent): |
| super(_PayloadReport.FieldNode, self).__init__() |
| self.name = name |
| self.value = value |
| self.linebreak = linebreak |
| self.indent = indent |
| |
| def GenerateLines(self, base_indent, sub_indent, curr_section): |
| """Generates a properly formatted 'name : value' entry.""" |
| report_output = "" |
| if self.name: |
| report_output += ( |
| self.name.ljust(curr_section.max_field_name_len) + " :" |
| ) |
| value_lines = str(self.value).splitlines() |
| if self.linebreak and self.name: |
| report_output += "\n" + "\n".join( |
| ["%*s%s" % (self.indent, "", line) for line in value_lines] |
| ) |
| else: |
| if self.name: |
| report_output += " " |
| report_output += "%*s" % (self.indent, "") |
| cont_line_indent = len(report_output) |
| indented_value_lines = [value_lines[0]] |
| indented_value_lines.extend( |
| [ |
| "%*s%s" % (cont_line_indent, "", line) |
| for line in value_lines[1:] |
| ] |
| ) |
| report_output += "\n".join(indented_value_lines) |
| |
| report_lines = [ |
| self._Indent(base_indent, line + "\n") |
| for line in report_output.split("\n") |
| ] |
| return report_lines, curr_section |
| |
| class SubReportNode(Node): |
| """A sub-report node, representing a nested report.""" |
| |
| def __init__(self, title, report): |
| super(_PayloadReport.SubReportNode, self).__init__() |
| self.title = title |
| self.report = report |
| |
| def GenerateLines(self, base_indent, sub_indent, curr_section): |
| """Recurse with indentation.""" |
| report_lines = [self._Indent(base_indent, self.title + " =>\n")] |
| report_lines.extend( |
| self.report.GenerateLines(base_indent + sub_indent, sub_indent) |
| ) |
| return report_lines, curr_section |
| |
| class SectionNode(Node): |
| """A section header node.""" |
| |
| def __init__(self, title=None): |
| super(_PayloadReport.SectionNode, self).__init__() |
| self.title = title |
| self.max_field_name_len = 0 |
| |
| def GenerateLines(self, base_indent, sub_indent, curr_section): |
| """Dump a title line, return self as the (new) current section.""" |
| report_lines = [] |
| if self.title: |
| report_lines.append( |
| self._Indent(base_indent, "=== %s ===\n" % self.title) |
| ) |
| return report_lines, self |
| |
| def __init__(self): |
| self.report = [] |
| self.last_section = self.global_section = self.SectionNode() |
| self.is_finalized = False |
| |
| def GenerateLines(self, base_indent, sub_indent): |
| """Generates the lines in the report, properly indented. |
| |
| Args: |
| base_indent: The indentation used for root-level report lines. |
| sub_indent: The indentation offset used for sub-reports. |
| |
| Returns: |
| A list of indented report lines. |
| """ |
| report_lines = [] |
| curr_section = self.global_section |
| for node in self.report: |
| node_report_lines, curr_section = node.GenerateLines( |
| base_indent, sub_indent, curr_section |
| ) |
| report_lines.extend(node_report_lines) |
| |
| return report_lines |
| |
| def Dump(self, out_file, base_indent=0, sub_indent=2): |
| """Dumps the report to a file. |
| |
| Args: |
| out_file: File object to output the content to. |
| base_indent: Base indentation for report lines. |
| sub_indent: Added indentation for sub-reports. |
| """ |
| report_lines = self.GenerateLines(base_indent, sub_indent) |
| if report_lines and not self.is_finalized: |
| report_lines.append("(incomplete report)\n") |
| |
| for line in report_lines: |
| out_file.write(line) |
| |
| def AddField(self, name, value, linebreak=False, indent=0): |
| """Adds a field/value pair to the payload report. |
| |
| Args: |
| name: The field's name. |
| value: The field's value. |
| linebreak: Whether the value should be printed on a new line. |
| indent: Amount of extra indent for each line of the value. |
| """ |
| assert not self.is_finalized |
| if name and self.last_section.max_field_name_len < len(name): |
| self.last_section.max_field_name_len = len(name) |
| self.report.append(self.FieldNode(name, value, linebreak, indent)) |
| |
| def AddSubReport(self, title): |
| """Adds and returns a sub-report with a title.""" |
| assert not self.is_finalized |
| sub_report = self.SubReportNode(title, type(self)()) |
| self.report.append(sub_report) |
| return sub_report.report |
| |
| def AddSection(self, title): |
| """Adds a new section title.""" |
| assert not self.is_finalized |
| self.last_section = self.SectionNode(title) |
| self.report.append(self.last_section) |
| |
| def Finalize(self): |
| """Seals the report, marking it as complete.""" |
| self.is_finalized = True |
| |
| |
| # |
| # Payload verification. |
| # |
| |
| |
| class PayloadChecker(object): |
| """Checking the integrity of an update payload. |
| |
| This is a short-lived object whose purpose is to isolate the logic used for |
| verifying the integrity of an update payload. |
| """ |
| |
| def __init__( |
| self, |
| payload, |
| assert_type=None, |
| block_size=0, |
| allow_unhashed=False, |
| disabled_tests=(), |
| ): |
| """Initialize the checker. |
| |
| Args: |
| payload: The payload object to check. |
| assert_type: Assert that payload is either 'full' or 'delta' (optional). |
| block_size: Expected filesystem / payload block size (optional). |
| allow_unhashed: Allow operations with unhashed data blobs. |
| disabled_tests: Sequence of tests to disable. |
| """ |
| if not payload.is_init: |
| raise ValueError("Uninitialized update payload.") |
| |
| # Set checker configuration. |
| self.payload = payload |
| self.block_size = block_size if block_size else _DEFAULT_BLOCK_SIZE |
| if not _IsPowerOfTwo(self.block_size): |
| raise error.PayloadError( |
| "Expected block (%d) size is not a power of two." |
| % self.block_size |
| ) |
| if assert_type not in (None, _TYPE_FULL, _TYPE_DELTA): |
| raise error.PayloadError( |
| "Invalid assert_type value (%r)." % assert_type |
| ) |
| self.payload_type = assert_type |
| self.allow_unhashed = allow_unhashed |
| |
| # Disable specific tests. |
| self.check_move_same_src_dst_block = ( |
| _CHECK_MOVE_SAME_SRC_DST_BLOCK not in disabled_tests |
| ) |
| self.check_payload_sig = _CHECK_PAYLOAD_SIG not in disabled_tests |
| |
| # Reset state; these will be assigned when the manifest is checked. |
| self.sigs_offset = 0 |
| self.sigs_size = 0 |
| self.old_part_info = {} |
| self.new_part_info = {} |
| self.new_fs_sizes = collections.defaultdict(int) |
| self.old_fs_sizes = collections.defaultdict(int) |
| self.minor_version = None |
| self.major_version = None |
| |
| @staticmethod |
| def _CheckElem( |
| msg, |
| name, |
| report, |
| is_mandatory, |
| is_submsg, |
| convert=str, |
| msg_name=None, |
| linebreak=False, |
| indent=0, |
| ): |
| """Adds an element from a protobuf message to the payload report. |
| |
| Checks to see whether a message contains a given element, and if so adds |
| the element value to the provided report. A missing mandatory element |
| causes an exception to be raised. |
| |
| Args: |
| msg: The message containing the element. |
| name: The name of the element. |
| report: A report object to add the element name/value to. |
| is_mandatory: Whether or not this element must be present. |
| is_submsg: Whether this element is itself a message. |
| convert: A function for converting the element value for reporting. |
| msg_name: The name of the message object (for error reporting). |
| linebreak: Whether the value report should induce a line break. |
| indent: Amount of indent used for reporting the value. |
| |
| Returns: |
| A pair consisting of the element value and the generated sub-report for |
| it (if the element is a sub-message, None otherwise). If the element is |
| missing, returns (None, None). |
| |
| Raises: |
| error.PayloadError if a mandatory element is missing. |
| """ |
| element_result = collections.namedtuple( |
| "element_result", ["msg", "report"] |
| ) |
| |
| if not msg.HasField(name): |
| if is_mandatory: |
| raise error.PayloadError( |
| "%smissing mandatory %s %r." |
| % ( |
| msg_name + " " if msg_name else "", |
| "sub-message" if is_submsg else "field", |
| name, |
| ) |
| ) |
| return element_result(None, None) |
| |
| value = getattr(msg, name) |
| if is_submsg: |
| return element_result(value, report and report.AddSubReport(name)) |
| else: |
| if report: |
| report.AddField( |
| name, convert(value), linebreak=linebreak, indent=indent |
| ) |
| return element_result(value, None) |
| |
| @staticmethod |
| def _CheckRepeatedElemNotPresent(msg, field_name, msg_name): |
| """Checks that a repeated element is not specified in the message. |
| |
| Args: |
| msg: The message containing the element. |
| field_name: The name of the element. |
| msg_name: The name of the message object (for error reporting). |
| |
| Raises: |
| error.PayloadError if the repeated element is present or non-empty. |
| """ |
| if getattr(msg, field_name, None): |
| raise error.PayloadError( |
| "%sfield %r not empty." |
| % (msg_name + " " if msg_name else "", field_name) |
| ) |
| |
| @staticmethod |
| def _CheckElemNotPresent(msg, field_name, msg_name): |
| """Checks that an element is not specified in the message. |
| |
| Args: |
| msg: The message containing the element. |
| field_name: The name of the element. |
| msg_name: The name of the message object (for error reporting). |
| |
| Raises: |
| error.PayloadError if the repeated element is present. |
| """ |
| if msg.HasField(field_name): |
| raise error.PayloadError( |
| "%sfield %r exists." |
| % (msg_name + " " if msg_name else "", field_name) |
| ) |
| |
| @staticmethod |
| def _CheckMandatoryField( |
| msg, |
| field_name, |
| report, |
| msg_name, |
| convert=str, |
| linebreak=False, |
| indent=0, |
| ): |
| """Adds a mandatory field; returning first component from _CheckElem.""" |
| return PayloadChecker._CheckElem( |
| msg, |
| field_name, |
| report, |
| True, |
| False, |
| convert=convert, |
| msg_name=msg_name, |
| linebreak=linebreak, |
| indent=indent, |
| )[0] |
| |
| @staticmethod |
| def _CheckOptionalField( |
| msg, field_name, report, convert=str, linebreak=False, indent=0 |
| ): |
| """Adds an optional field; returning first component from _CheckElem.""" |
| return PayloadChecker._CheckElem( |
| msg, |
| field_name, |
| report, |
| False, |
| False, |
| convert=convert, |
| linebreak=linebreak, |
| indent=indent, |
| )[0] |
| |
| @staticmethod |
| def _CheckMandatorySubMsg(msg, submsg_name, report, msg_name): |
| """Adds a mandatory sub-message; wrapper for _CheckElem.""" |
| return PayloadChecker._CheckElem( |
| msg, submsg_name, report, True, True, msg_name |
| ) |
| |
| @staticmethod |
| def _CheckOptionalSubMsg(msg, submsg_name, report): |
| """Adds an optional sub-message; wrapper for _CheckElem.""" |
| return PayloadChecker._CheckElem(msg, submsg_name, report, False, True) |
| |
| @staticmethod |
| def _CheckPresentIff(val1, val2, name1, name2, obj_name): |
| """Checks that val1 is None iff val2 is None. |
| |
| Args: |
| val1: first value to be compared. |
| val2: second value to be compared. |
| name1: name of object holding the first value. |
| name2: name of object holding the second value. |
| obj_name: Name of the object containing these values. |
| |
| Raises: |
| error.PayloadError if assertion does not hold. |
| """ |
| if None in (val1, val2) and val1 is not val2: |
| present, missing = ( |
| (name1, name2) if val2 is None else (name2, name1) |
| ) |
| raise error.PayloadError( |
| "%r present without %r%s." |
| % (present, missing, " in " + obj_name if obj_name else "") |
| ) |
| |
| @staticmethod |
| def _CheckPresentIffMany(vals, name, obj_name): |
| """Checks that a set of vals and names imply every other element. |
| |
| Args: |
| vals: The set of values to be compared. |
| name: The name of the objects holding the corresponding value. |
| obj_name: Name of the object containing these values. |
| |
| Raises: |
| error.PayloadError if assertion does not hold. |
| """ |
| if any(vals) and not all(vals): |
| raise error.PayloadError( |
| "%r is not present in all values%s." |
| % (name, " in " + obj_name if obj_name else "") |
| ) |
| |
| @staticmethod |
| def _Run(cmd, send_data=None): |
| """Runs a subprocess, returns its output. |
| |
| Args: |
| cmd: Sequence of command-line argument for invoking the subprocess. |
| send_data: Data to feed to the process via its stdin. |
| |
| Returns: |
| A tuple containing the stdout and stderr output of the process. |
| """ |
| run_process = subprocess.Popen( |
| cmd, stdin=subprocess.PIPE, stdout=subprocess.PIPE |
| ) |
| try: |
| result = run_process.communicate(input=send_data) |
| finally: |
| exit_code = run_process.wait() |
| |
| if exit_code: |
| raise RuntimeError( |
| "Subprocess %r failed with code %r." % (cmd, exit_code) |
| ) |
| |
| return result |
| |
| @staticmethod |
| def _CheckSha256Signature( |
| sig_data, pubkey_file_name, actual_hash, sig_name |
| ): |
| """Verifies an actual hash against a signed one. |
| |
| Args: |
| sig_data: The raw signature data. |
| pubkey_file_name: Public key used for verifying signature. |
| actual_hash: The actual hash digest. |
| sig_name: Signature name for error reporting. |
| |
| Raises: |
| error.PayloadError if signature could not be verified. |
| """ |
| if len(sig_data) != 256: |
| raise error.PayloadError( |
| "%s: signature size (%d) not as expected (256)." |
| % (sig_name, len(sig_data)) |
| ) |
| signed_data, _ = PayloadChecker._Run( |
| [ |
| "openssl", |
| "rsautl", |
| "-verify", |
| "-pubin", |
| "-inkey", |
| pubkey_file_name, |
| ], |
| send_data=sig_data, |
| ) |
| |
| if len(signed_data) != len(common.SIG_ASN1_HEADER) + 32: |
| raise error.PayloadError( |
| "%s: unexpected signed data length (%d)." |
| % (sig_name, len(signed_data)) |
| ) |
| |
| if not signed_data.startswith(common.SIG_ASN1_HEADER): |
| raise error.PayloadError( |
| "%s: not containing standard ASN.1 prefix." % sig_name |
| ) |
| |
| signed_hash = signed_data[len(common.SIG_ASN1_HEADER) :] |
| if signed_hash != actual_hash: |
| raise error.PayloadError( |
| "%s: signed hash (%s) different from actual (%s)." |
| % ( |
| sig_name, |
| common.FormatSha256(signed_hash), |
| common.FormatSha256(actual_hash), |
| ) |
| ) |
| |
| @staticmethod |
| def _CheckBlocksFitLength( |
| length, num_blocks, block_size, length_name, block_name=None |
| ): |
| """Checks that a given length fits given block space. |
| |
| This ensures that the number of blocks allocated is appropriate for the |
| length of the data residing in these blocks. |
| |
| Args: |
| length: The actual length of the data. |
| num_blocks: The number of blocks allocated for it. |
| block_size: The size of each block in bytes. |
| length_name: Name of length (used for error reporting). |
| block_name: Name of block (used for error reporting). |
| |
| Raises: |
| error.PayloadError if the aforementioned invariant is not satisfied. |
| """ |
| # Check: length <= num_blocks * block_size. |
| if length > num_blocks * block_size: |
| raise error.PayloadError( |
| "%s (%d) > num %sblocks (%d) * block_size (%d)." |
| % ( |
| length_name, |
| length, |
| block_name or "", |
| num_blocks, |
| block_size, |
| ) |
| ) |
| |
| # Check: length > (num_blocks - 1) * block_size. |
| if length <= (num_blocks - 1) * block_size: |
| raise error.PayloadError( |
| "%s (%d) <= (num %sblocks - 1 (%d)) * block_size (%d)." |
| % ( |
| length_name, |
| length, |
| block_name or "", |
| num_blocks - 1, |
| block_size, |
| ) |
| ) |
| |
| def _CheckManifestMinorVersion(self, report): |
| """Checks the payload manifest minor_version field. |
| |
| Args: |
| report: The report object to add to. |
| |
| Raises: |
| error.PayloadError if any of the checks fail. |
| """ |
| self.minor_version = self._CheckOptionalField( |
| self.payload.manifest, "minor_version", report |
| ) |
| if self.minor_version in _SUPPORTED_MINOR_VERSIONS: |
| if ( |
| self.payload_type |
| not in _SUPPORTED_MINOR_VERSIONS[self.minor_version] |
| ): |
| raise error.PayloadError( |
| "Minor version %d not compatible with payload type %s." |
| % (self.minor_version, self.payload_type) |
| ) |
| elif self.minor_version is None: |
| raise error.PayloadError("Minor version is not set.") |
| else: |
| raise error.PayloadError( |
| "Unsupported minor version: %d" % self.minor_version |
| ) |
| |
| def _CheckManifest(self, report, part_sizes=None): |
| """Checks the payload manifest. |
| |
| Args: |
| report: A report object to add to. |
| part_sizes: Map of partition label to partition size in bytes. |
| |
| Returns: |
| A tuple consisting of the partition block size used during the update |
| (integer), the signatures block offset and size. |
| |
| Raises: |
| error.PayloadError if any of the checks fail. |
| """ |
| self.major_version = self.payload.header.version |
| |
| part_sizes = part_sizes or collections.defaultdict(int) |
| manifest = self.payload.manifest |
| report.AddSection("manifest") |
| |
| # Check: block_size must exist and match the expected value. |
| actual_block_size = self._CheckMandatoryField( |
| manifest, "block_size", report, "manifest" |
| ) |
| if actual_block_size != self.block_size: |
| raise error.PayloadError( |
| "Block_size (%d) not as expected (%d)." |
| % (actual_block_size, self.block_size) |
| ) |
| |
| # Check: signatures_offset <==> signatures_size. |
| self.sigs_offset = self._CheckOptionalField( |
| manifest, "signatures_offset", report |
| ) |
| self.sigs_size = self._CheckOptionalField( |
| manifest, "signatures_size", report |
| ) |
| self._CheckPresentIff( |
| self.sigs_offset, |
| self.sigs_size, |
| "signatures_offset", |
| "signatures_size", |
| "manifest", |
| ) |
| |
| for part in manifest.partitions: |
| name = part.partition_name |
| self.old_part_info[name] = self._CheckOptionalSubMsg( |
| part, "old_partition_info", report |
| ) |
| self.new_part_info[name] = self._CheckMandatorySubMsg( |
| part, "new_partition_info", report, "manifest.partitions" |
| ) |
| |
| # Check: Old-style partition infos should not be specified. |
| for _, part in common.CROS_PARTITIONS: |
| self._CheckElemNotPresent( |
| manifest, "old_%s_info" % part, "manifest" |
| ) |
| self._CheckElemNotPresent( |
| manifest, "new_%s_info" % part, "manifest" |
| ) |
| |
| # Check: If old_partition_info is specified anywhere, it must be |
| # specified everywhere. |
| old_part_msgs = [ |
| part.msg for part in self.old_part_info.values() if part |
| ] |
| self._CheckPresentIffMany( |
| old_part_msgs, "old_partition_info", "manifest.partitions" |
| ) |
| |
| is_delta = any( |
| part and part.msg for part in self.old_part_info.values() |
| ) |
| if is_delta: |
| # Assert/mark delta payload. |
| if self.payload_type == _TYPE_FULL: |
| raise error.PayloadError( |
| "Apparent full payload contains old_{kernel,rootfs}_info." |
| ) |
| self.payload_type = _TYPE_DELTA |
| |
| for part, (msg, part_report) in self.old_part_info.items(): |
| # Check: {size, hash} present in old_{kernel,rootfs}_info. |
| field = "old_%s_info" % part |
| self.old_fs_sizes[part] = self._CheckMandatoryField( |
| msg, "size", part_report, field |
| ) |
| self._CheckMandatoryField( |
| msg, "hash", part_report, field, convert=common.FormatSha256 |
| ) |
| |
| # Check: old_{kernel,rootfs} size must fit in respective partition. |
| if self.old_fs_sizes[part] > part_sizes[part] > 0: |
| raise error.PayloadError( |
| "Old %s content (%d) exceed partition size (%d)." |
| % (part, self.old_fs_sizes[part], part_sizes[part]) |
| ) |
| else: |
| # Assert/mark full payload. |
| if self.payload_type == _TYPE_DELTA: |
| raise error.PayloadError( |
| "Apparent delta payload missing old_{kernel,rootfs}_info." |
| ) |
| self.payload_type = _TYPE_FULL |
| |
| # Check: new_{kernel,rootfs}_info present; contains {size, hash}. |
| for part, (msg, part_report) in self.new_part_info.items(): |
| field = "new_%s_info" % part |
| self.new_fs_sizes[part] = self._CheckMandatoryField( |
| msg, "size", part_report, field |
| ) |
| self._CheckMandatoryField( |
| msg, "hash", part_report, field, convert=common.FormatSha256 |
| ) |
| |
| # Check: new_{kernel,rootfs} size must fit in respective partition. |
| if self.new_fs_sizes[part] > part_sizes[part] > 0: |
| raise error.PayloadError( |
| "New %s content (%d) exceed partition size (%d)." |
| % (part, self.new_fs_sizes[part], part_sizes[part]) |
| ) |
| |
| # Check: minor_version makes sense for the payload type. This check should |
| # run after the payload type has been set. |
| self._CheckManifestMinorVersion(report) |
| |
| def _CheckLength(self, length, total_blocks, op_name, length_name): |
| """Checks whether a length matches the space designated in extents. |
| |
| Args: |
| length: The total length of the data. |
| total_blocks: The total number of blocks in extents. |
| op_name: Operation name (for error reporting). |
| length_name: Length name (for error reporting). |
| |
| Raises: |
| error.PayloadError is there a problem with the length. |
| """ |
| # Check: length is non-zero. |
| if length == 0: |
| raise error.PayloadError("%s: %s is zero." % (op_name, length_name)) |
| |
| # Check that length matches number of blocks. |
| self._CheckBlocksFitLength( |
| length, |
| total_blocks, |
| self.block_size, |
| "%s: %s" % (op_name, length_name), |
| ) |
| |
| def _CheckExtents(self, extents, usable_size, block_counters, name): |
| """Checks a sequence of extents. |
| |
| Args: |
| extents: The sequence of extents to check. |
| usable_size: The usable size of the partition to which the extents apply. |
| block_counters: Array of counters corresponding to the number of blocks. |
| name: The name of the extent block. |
| |
| Returns: |
| The total number of blocks in the extents. |
| |
| Raises: |
| error.PayloadError if any of the entailed checks fails. |
| """ |
| total_num_blocks = 0 |
| for ex, ex_name in common.ExtentIter(extents, name): |
| # Check: Mandatory fields. |
| start_block = PayloadChecker._CheckMandatoryField( |
| ex, "start_block", None, ex_name |
| ) |
| num_blocks = PayloadChecker._CheckMandatoryField( |
| ex, "num_blocks", None, ex_name |
| ) |
| end_block = start_block + num_blocks |
| |
| # Check: num_blocks > 0. |
| if num_blocks == 0: |
| raise error.PayloadError("%s: extent length is zero." % ex_name) |
| |
| # Check: Make sure we're within the partition limit. |
| if usable_size and end_block * self.block_size > usable_size: |
| raise error.PayloadError( |
| "%s: extent (%s) exceeds usable partition size (%d)." |
| % ( |
| ex_name, |
| common.FormatExtent(ex, self.block_size), |
| usable_size, |
| ) |
| ) |
| |
| # Record block usage. |
| for i in range(start_block, end_block): |
| block_counters[i] += 1 |
| |
| total_num_blocks += num_blocks |
| |
| return total_num_blocks |
| |
| def _CheckReplaceOperation( |
| self, op, data_length, total_dst_blocks, op_name |
| ): |
| """Specific checks for REPLACE{_BZ,_XZ,_ZSTD*} operations. |
| |
| Args: |
| op: The operation object from the manifest. |
| data_length: The length of the data blob associated with the operation. |
| total_dst_blocks: Total number of blocks in dst_extents. |
| op_name: Operation name for error reporting. |
| |
| Raises: |
| error.PayloadError if any check fails. |
| """ |
| # Check: total_dst_blocks is not a floating point. |
| if isinstance(total_dst_blocks, float): |
| raise error.PayloadError( |
| "%s: contains invalid data type of " |
| "total_dst_blocks." % op_name |
| ) |
| |
| # Check: Does not contain src extents. |
| if op.src_extents: |
| raise error.PayloadError("%s: contains src_extents." % op_name) |
| |
| # Check: Contains data. |
| if data_length is None: |
| raise error.PayloadError( |
| "%s: missing data_{offset,length}." % op_name |
| ) |
| |
| if op.type == common.OpType.REPLACE: |
| PayloadChecker._CheckBlocksFitLength( |
| data_length, |
| total_dst_blocks, |
| self.block_size, |
| op_name + ".data_length", |
| "dst", |
| ) |
| else: |
| # Check: data_length must be smaller than the allotted dst blocks. |
| if data_length >= total_dst_blocks * self.block_size: |
| raise error.PayloadError( |
| "%s: data_length (%d) must be less than allotted dst block " |
| "space (%d * %d)." |
| % (op_name, data_length, total_dst_blocks, self.block_size) |
| ) |
| |
| def _CheckZeroOperation(self, op, op_name): |
| """Specific checks for ZERO operations. |
| |
| Args: |
| op: The operation object from the manifest. |
| op_name: Operation name for error reporting. |
| |
| Raises: |
| error.PayloadError if any check fails. |
| """ |
| # Check: Does not contain src extents, data_length and data_offset. |
| if op.src_extents: |
| raise error.PayloadError("%s: contains src_extents." % op_name) |
| if op.data_length: |
| raise error.PayloadError("%s: contains data_length." % op_name) |
| if op.data_offset: |
| raise error.PayloadError("%s: contains data_offset." % op_name) |
| |
| def _CheckAnyDiffOperation( |
| self, op, data_length, total_dst_blocks, op_name |
| ): |
| """Specific checks for SOURCE_BSDIFF, PUFFDIFF and BROTLI_BSDIFF |
| operations. |
| |
| Args: |
| op: The operation. |
| data_length: The length of the data blob associated with the operation. |
| total_dst_blocks: Total number of blocks in dst_extents. |
| op_name: Operation name for error reporting. |
| |
| Raises: |
| error.PayloadError if any check fails. |
| """ |
| # Check: data_{offset,length} present. |
| if data_length is None: |
| raise error.PayloadError( |
| "%s: missing data_{offset,length}." % op_name |
| ) |
| |
| # Check: data_length is strictly smaller than the allotted dst blocks. |
| if data_length >= total_dst_blocks * self.block_size: |
| raise error.PayloadError( |
| "%s: data_length (%d) must be smaller than allotted dst space " |
| "(%d * %d = %d)." |
| % ( |
| op_name, |
| data_length, |
| total_dst_blocks, |
| self.block_size, |
| total_dst_blocks * self.block_size, |
| ) |
| ) |
| |
| # Check the existence of src_length and dst_length for legacy bsdiffs. |
| if op.type == common.OpType.SOURCE_BSDIFF and self.minor_version <= 3: |
| if not op.HasField("src_length") or not op.HasField("dst_length"): |
| raise error.PayloadError( |
| "%s: require {src,dst}_length." % op_name |
| ) |
| else: |
| if op.HasField("src_length") or op.HasField("dst_length"): |
| raise error.PayloadError( |
| "%s: unneeded {src,dst}_length." % op_name |
| ) |
| |
| def _CheckSourceCopyOperation( |
| self, data_offset, total_src_blocks, total_dst_blocks, op_name |
| ): |
| """Specific checks for SOURCE_COPY. |
| |
| Args: |
| data_offset: The offset of a data blob for the operation. |
| total_src_blocks: Total number of blocks in src_extents. |
| total_dst_blocks: Total number of blocks in dst_extents. |
| op_name: Operation name for error reporting. |
| |
| Raises: |
| error.PayloadError if any check fails. |
| """ |
| # Check: No data_{offset,length}. |
| if data_offset is not None: |
| raise error.PayloadError( |
| "%s: contains data_{offset,length}." % op_name |
| ) |
| |
| # Check: total_src_blocks == total_dst_blocks. |
| if total_src_blocks != total_dst_blocks: |
| raise error.PayloadError( |
| "%s: total src blocks (%d) != total dst blocks (%d)." |
| % (op_name, total_src_blocks, total_dst_blocks) |
| ) |
| |
| def _CheckAnySourceOperation(self, op, total_src_blocks, op_name): |
| """Specific checks for SOURCE_* operations. |
| |
| Args: |
| op: The operation object from the manifest. |
| total_src_blocks: Total number of blocks in src_extents. |
| op_name: Operation name for error reporting. |
| |
| Raises: |
| error.PayloadError if any check fails. |
| """ |
| # Check: total_src_blocks != 0. |
| if total_src_blocks == 0: |
| raise error.PayloadError( |
| "%s: no src blocks in a source op." % op_name |
| ) |
| |
| # Check: src_sha256_hash present in minor version >= 3. |
| if self.minor_version >= 3 and op.src_sha256_hash is None: |
| raise error.PayloadError("%s: source hash missing." % op_name) |
| |
| def _CheckOperation( |
| self, |
| op, |
| op_name, |
| old_block_counters, |
| new_block_counters, |
| old_usable_size, |
| new_usable_size, |
| prev_data_offset, |
| blob_hash_counts, |
| ): |
| """Checks a single update operation. |
| |
| Args: |
| op: The operation object. |
| op_name: Operation name string for error reporting. |
| old_block_counters: Arrays of block read counters. |
| new_block_counters: Arrays of block write counters. |
| old_usable_size: The overall usable size for src data in bytes. |
| new_usable_size: The overall usable size for dst data in bytes. |
| prev_data_offset: Offset of last used data bytes. |
| blob_hash_counts: Counters for hashed/unhashed blobs. |
| |
| Returns: |
| The amount of data blob associated with the operation. |
| |
| Raises: |
| error.PayloadError if any check has failed. |
| """ |
| # Check extents. |
| total_src_blocks = self._CheckExtents( |
| op.src_extents, |
| old_usable_size, |
| old_block_counters, |
| op_name + ".src_extents", |
| ) |
| total_dst_blocks = self._CheckExtents( |
| op.dst_extents, |
| new_usable_size, |
| new_block_counters, |
| op_name + ".dst_extents", |
| ) |
| |
| # Check: data_offset present <==> data_length present. |
| data_offset = self._CheckOptionalField(op, "data_offset", None) |
| data_length = self._CheckOptionalField(op, "data_length", None) |
| self._CheckPresentIff( |
| data_offset, data_length, "data_offset", "data_length", op_name |
| ) |
| |
| # Check: At least one dst_extent. |
| if not op.dst_extents: |
| raise error.PayloadError("%s: dst_extents is empty." % op_name) |
| |
| # Check {src,dst}_length, if present. |
| if op.HasField("src_length"): |
| self._CheckLength( |
| op.src_length, total_src_blocks, op_name, "src_length" |
| ) |
| if op.HasField("dst_length"): |
| self._CheckLength( |
| op.dst_length, total_dst_blocks, op_name, "dst_length" |
| ) |
| |
| if op.HasField("data_sha256_hash"): |
| blob_hash_counts["hashed"] += 1 |
| |
| # Check: Operation carries data. |
| if data_offset is None: |
| raise error.PayloadError( |
| "%s: data_sha256_hash present but no data_{offset,length}." |
| % op_name |
| ) |
| |
| # Check: Hash verifies correctly. |
| actual_hash = hashlib.sha256( |
| self.payload.ReadDataBlob(data_offset, data_length) |
| ) |
| if op.data_sha256_hash != actual_hash.digest(): |
| raise error.PayloadError( |
| "%s: data_sha256_hash (%s) does not match actual hash (%s)." |
| % ( |
| op_name, |
| common.FormatSha256(op.data_sha256_hash), |
| common.FormatSha256(actual_hash.digest()), |
| ) |
| ) |
| elif data_offset is not None: |
| if self.allow_unhashed: |
| blob_hash_counts["unhashed"] += 1 |
| else: |
| raise error.PayloadError( |
| "%s: unhashed operation not allowed." % op_name |
| ) |
| |
| if data_offset is not None: |
| # Check: Contiguous use of data section. |
| if data_offset != prev_data_offset: |
| raise error.PayloadError( |
| "%s: data offset (%d) not matching amount used so far (%d)." |
| % (op_name, data_offset, prev_data_offset) |
| ) |
| |
| # Type-specific checks. |
| if op.type in ( |
| common.OpType.REPLACE, |
| common.OpType.REPLACE_BZ, |
| common.OpType.REPLACE_XZ, |
| common.OpType.REPLACE_ZSTD, |
| common.OpType.REPLACE_ZSTD_INCREASED_WINDOW, |
| ): |
| self._CheckReplaceOperation( |
| op, data_length, total_dst_blocks, op_name |
| ) |
| elif op.type == common.OpType.ZERO and self.minor_version >= 4: |
| self._CheckZeroOperation(op, op_name) |
| elif op.type == common.OpType.SOURCE_COPY and self.minor_version >= 2: |
| self._CheckSourceCopyOperation( |
| data_offset, total_src_blocks, total_dst_blocks, op_name |
| ) |
| self._CheckAnySourceOperation(op, total_src_blocks, op_name) |
| elif op.type == common.OpType.SOURCE_BSDIFF and self.minor_version >= 2: |
| self._CheckAnyDiffOperation( |
| op, data_length, total_dst_blocks, op_name |
| ) |
| self._CheckAnySourceOperation(op, total_src_blocks, op_name) |
| elif op.type == common.OpType.BROTLI_BSDIFF and self.minor_version >= 4: |
| self._CheckAnyDiffOperation( |
| op, data_length, total_dst_blocks, op_name |
| ) |
| self._CheckAnySourceOperation(op, total_src_blocks, op_name) |
| elif op.type == common.OpType.PUFFDIFF and self.minor_version >= 5: |
| self._CheckAnyDiffOperation( |
| op, data_length, total_dst_blocks, op_name |
| ) |
| self._CheckAnySourceOperation(op, total_src_blocks, op_name) |
| else: |
| raise error.PayloadError( |
| "Operation %s (type %d) not allowed in minor version %d" |
| % (op_name, op.type, self.minor_version) |
| ) |
| return data_length if data_length is not None else 0 |
| |
| def _SizeToNumBlocks(self, size): |
| """Returns the number of blocks needed to contain a given byte size.""" |
| return (size + self.block_size - 1) // self.block_size |
| |
| def _AllocBlockCounters(self, total_size): |
| """Returns a freshly initialized array of block counters. |
| |
| Note that the generated array is not portable as is due to byte-ordering |
| issues, hence it should not be serialized. |
| |
| Args: |
| total_size: The total block size in bytes. |
| |
| Returns: |
| An array of unsigned short elements initialized to zero, one for each of |
| the blocks necessary for containing the partition. |
| """ |
| return array.array( |
| "H", itertools.repeat(0, self._SizeToNumBlocks(total_size)) |
| ) |
| |
| def _CheckOperations( |
| self, |
| operations, |
| report, |
| base_name, |
| old_fs_size, |
| new_fs_size, |
| old_usable_size, |
| new_usable_size, |
| prev_data_offset, |
| ): |
| """Checks a sequence of update operations. |
| |
| Args: |
| operations: The sequence of operations to check. |
| report: The report object to add to. |
| base_name: The name of the operation block. |
| old_fs_size: The old filesystem size in bytes. |
| new_fs_size: The new filesystem size in bytes. |
| old_usable_size: The overall usable size of the old partition in bytes. |
| new_usable_size: The overall usable size of the new partition in bytes. |
| prev_data_offset: Offset of last used data bytes. |
| |
| Returns: |
| The total data blob size used. |
| |
| Raises: |
| error.PayloadError if any of the checks fails. |
| """ |
| # The total size of data blobs used by operations scanned thus far. |
| total_data_used = 0 |
| # Counts of specific operation types. |
| op_counts = { |
| common.OpType.REPLACE: 0, |
| common.OpType.REPLACE_BZ: 0, |
| common.OpType.REPLACE_XZ: 0, |
| common.OpType.REPLACE_ZSTD: 0, |
| common.OpType.REPLACE_ZSTD_INCREASED_WINDOW: 0, |
| common.OpType.ZERO: 0, |
| common.OpType.SOURCE_COPY: 0, |
| common.OpType.SOURCE_BSDIFF: 0, |
| common.OpType.PUFFDIFF: 0, |
| common.OpType.BROTLI_BSDIFF: 0, |
| } |
| # Total blob sizes for each operation type. |
| op_blob_totals = { |
| common.OpType.REPLACE: 0, |
| common.OpType.REPLACE_BZ: 0, |
| common.OpType.REPLACE_XZ: 0, |
| common.OpType.REPLACE_ZSTD: 0, |
| common.OpType.REPLACE_ZSTD_INCREASED_WINDOW: 0, |
| # SOURCE_COPY operations don't have blobs. |
| common.OpType.SOURCE_BSDIFF: 0, |
| common.OpType.PUFFDIFF: 0, |
| common.OpType.BROTLI_BSDIFF: 0, |
| } |
| # Counts of hashed vs unhashed operations. |
| blob_hash_counts = { |
| "hashed": 0, |
| "unhashed": 0, |
| } |
| |
| # Allocate old and new block counters. |
| old_block_counters = ( |
| self._AllocBlockCounters(old_usable_size) if old_fs_size else None |
| ) |
| new_block_counters = self._AllocBlockCounters(new_usable_size) |
| |
| # Process and verify each operation. |
| op_num = 0 |
| for op, op_name in common.OperationIter(operations, base_name): |
| op_num += 1 |
| |
| # Check: Type is valid. |
| if op.type not in op_counts: |
| raise error.PayloadError( |
| "%s: invalid type (%d)." % (op_name, op.type) |
| ) |
| op_counts[op.type] += 1 |
| |
| curr_data_used = self._CheckOperation( |
| op, |
| op_name, |
| old_block_counters, |
| new_block_counters, |
| old_usable_size, |
| new_usable_size, |
| prev_data_offset + total_data_used, |
| blob_hash_counts, |
| ) |
| if curr_data_used: |
| op_blob_totals[op.type] += curr_data_used |
| total_data_used += curr_data_used |
| |
| # Report totals and breakdown statistics. |
| report.AddField("total operations", op_num) |
| report.AddField( |
| None, |
| histogram.Histogram.FromCountDict( |
| op_counts, key_names=common.OpType.NAMES |
| ), |
| indent=1, |
| ) |
| report.AddField("total blobs", sum(blob_hash_counts.values())) |
| report.AddField( |
| None, histogram.Histogram.FromCountDict(blob_hash_counts), indent=1 |
| ) |
| report.AddField( |
| "total blob size", _AddHumanReadableSize(total_data_used) |
| ) |
| report.AddField( |
| None, |
| histogram.Histogram.FromCountDict( |
| op_blob_totals, |
| formatter=_AddHumanReadableSize, |
| key_names=common.OpType.NAMES, |
| ), |
| indent=1, |
| ) |
| |
| # Report read/write histograms. |
| if old_block_counters: |
| report.AddField( |
| "block read hist", |
| histogram.Histogram.FromKeyList(old_block_counters), |
| linebreak=True, |
| indent=1, |
| ) |
| |
| new_write_hist = histogram.Histogram.FromKeyList( |
| new_block_counters[: self._SizeToNumBlocks(new_fs_size)] |
| ) |
| report.AddField( |
| "block write hist", new_write_hist, linebreak=True, indent=1 |
| ) |
| |
| # Check: Full update must write each dst block once. |
| if self.payload_type == _TYPE_FULL and new_write_hist.GetKeys() != [1]: |
| raise error.PayloadError( |
| "%s: not all blocks written exactly once during full update." |
| % base_name |
| ) |
| |
| return total_data_used |
| |
| def _CheckSignatures(self, report, pubkey_file_name): |
| """Checks a payload's signature block.""" |
| sigs_raw = self.payload.ReadDataBlob(self.sigs_offset, self.sigs_size) |
| sigs = update_metadata_pb2.Signatures() |
| sigs.ParseFromString(sigs_raw) |
| report.AddSection("signatures") |
| |
| # Check: At least one signature present. |
| if not sigs.signatures: |
| raise error.PayloadError("Signature block is empty.") |
| |
| # Check that we don't have the signature operation blob at the end (used to |
| # be for major version 1). |
| last_partition = self.payload.manifest.partitions[-1] |
| if last_partition.operations: |
| last_op = last_partition.operations[-1] |
| # Check: signatures_{offset,size} must match the last (fake) operation. |
| if ( |
| last_op.type == common.OpType.REPLACE |
| and last_op.data_offset == self.sigs_offset |
| and last_op.data_length == self.sigs_size |
| ): |
| raise error.PayloadError( |
| "It seems like the last operation is the " |
| "signature blob. This is an invalid payload." |
| ) |
| |
| # Compute the checksum of all data up to signature blob. |
| # TODO(garnold) we're re-reading the whole data section into a string |
| # just to compute the checksum; instead, we could do it incrementally as |
| # we read the blobs one-by-one, under the assumption that we're reading |
| # them in order (which currently holds). This should be reconsidered. |
| payload_hasher = self.payload.manifest_hasher.copy() |
| common.Read( |
| self.payload.payload_file, |
| self.sigs_offset, |
| offset=self.payload.data_offset, |
| hasher=payload_hasher, |
| ) |
| |
| for sig, sig_name in common.SignatureIter( |
| sigs.signatures, "signatures" |
| ): |
| sig_report = report.AddSubReport(sig_name) |
| |
| # Check: Signature contains mandatory fields. |
| self._CheckMandatoryField(sig, "data", None, sig_name) |
| sig_report.AddField("data len", len(sig.data)) |
| |
| # Check: Signatures pertains to actual payload hash. |
| if sig.data: |
| self._CheckSha256Signature( |
| sig.data, |
| pubkey_file_name, |
| payload_hasher.digest(), |
| sig_name, |
| ) |
| |
| def Run( |
| self, |
| pubkey_file_name=None, |
| metadata_sig_file=None, |
| metadata_size=0, |
| part_sizes=None, |
| report_out_file=None, |
| ): |
| """Checker entry point, invoking all checks. |
| |
| Args: |
| pubkey_file_name: Public key used for signature verification. |
| metadata_sig_file: Metadata signature, if verification is desired. |
| metadata_size: Metadata size, if verification is desired. |
| part_sizes: Mapping of partition label to size in bytes (default: infer |
| based on payload type and version or filesystem). |
| report_out_file: File object to dump the report to. |
| |
| Raises: |
| error.PayloadError if payload verification failed. |
| """ |
| if not pubkey_file_name: |
| pubkey_file_name = _DEFAULT_PUBKEY_FILE_NAME |
| |
| report = _PayloadReport() |
| |
| # Get payload file size. |
| self.payload.payload_file.seek(0, 2) |
| payload_file_size = self.payload.payload_file.tell() |
| self.payload.ResetFile() |
| |
| try: |
| # Check metadata_size (if provided). |
| if metadata_size and self.payload.metadata_size != metadata_size: |
| raise error.PayloadError( |
| "Invalid payload metadata size in payload(%d) " |
| "vs given(%d)" % (self.payload.metadata_size, metadata_size) |
| ) |
| |
| # Check metadata signature (if provided). |
| if metadata_sig_file: |
| metadata_sig = base64.b64decode(metadata_sig_file.read()) |
| self._CheckSha256Signature( |
| metadata_sig, |
| pubkey_file_name, |
| self.payload.manifest_hasher.digest(), |
| "metadata signature", |
| ) |
| |
| # Part 1: Check the file header. |
| report.AddSection("header") |
| # Check: Payload version is valid. |
| if self.payload.header.version not in (1, 2): |
| raise error.PayloadError( |
| "Unknown payload version (%d)." |
| % self.payload.header.version |
| ) |
| report.AddField("version", self.payload.header.version) |
| report.AddField("manifest len", self.payload.header.manifest_len) |
| |
| # Part 2: Check the manifest. |
| self._CheckManifest(report, part_sizes) |
| assert self.payload_type, "payload type should be known by now" |
| |
| # Make sure deprecated values are not present in the payload. |
| for field in ("install_operations", "kernel_install_operations"): |
| self._CheckRepeatedElemNotPresent( |
| self.payload.manifest, field, "manifest" |
| ) |
| for field in ( |
| "old_kernel_info", |
| "old_rootfs_info", |
| "new_kernel_info", |
| "new_rootfs_info", |
| ): |
| self._CheckElemNotPresent( |
| self.payload.manifest, field, "manifest" |
| ) |
| |
| total_blob_size = 0 |
| for part, operations in ( |
| (p.partition_name, p.operations) |
| for p in self.payload.manifest.partitions |
| ): |
| report.AddSection("%s operations" % part) |
| |
| new_fs_usable_size = self.new_fs_sizes[part] |
| old_fs_usable_size = self.old_fs_sizes[part] |
| |
| if part_sizes is not None and part_sizes.get(part, None): |
| new_fs_usable_size = old_fs_usable_size = part_sizes[part] |
| |
| # TODO(chromium:243559) only default to the filesystem size if no |
| # explicit size provided *and* the partition size is not embedded in the |
| # payload; see issue for more details. |
| total_blob_size += self._CheckOperations( |
| operations, |
| report, |
| "%s_install_operations" % part, |
| self.old_fs_sizes[part], |
| self.new_fs_sizes[part], |
| old_fs_usable_size, |
| new_fs_usable_size, |
| total_blob_size, |
| ) |
| |
| # Check: Operations data reach the end of the payload file. |
| used_payload_size = self.payload.data_offset + total_blob_size |
| # Major versions 2 and higher have a signature at the end, so it should be |
| # considered in the total size of the image. |
| if self.sigs_size: |
| used_payload_size += self.sigs_size |
| |
| if used_payload_size != payload_file_size: |
| raise error.PayloadError( |
| "Used payload size (%d) different from actual file size (%d)." |
| % (used_payload_size, payload_file_size) |
| ) |
| |
| # Part 4: Handle payload signatures message. |
| if self.check_payload_sig and self.sigs_size: |
| self._CheckSignatures(report, pubkey_file_name) |
| |
| # Part 5: Summary. |
| report.AddSection("summary") |
| report.AddField("update type", self.payload_type) |
| |
| report.Finalize() |
| finally: |
| if report_out_file: |
| report.Dump(report_out_file) |