|  | #!/usr/bin/python2 | 
|  | # | 
|  | # Copyright (C) 2013 The Android Open Source Project | 
|  | # | 
|  | # Licensed under the Apache License, Version 2.0 (the "License"); | 
|  | # you may not use this file except in compliance with the License. | 
|  | # You may obtain a copy of the License at | 
|  | # | 
|  | #      http://www.apache.org/licenses/LICENSE-2.0 | 
|  | # | 
|  | # Unless required by applicable law or agreed to in writing, software | 
|  | # distributed under the License is distributed on an "AS IS" BASIS, | 
|  | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | 
|  | # See the License for the specific language governing permissions and | 
|  | # limitations under the License. | 
|  | # | 
|  |  | 
|  | """Command-line tool for checking and applying Chrome OS update payloads.""" | 
|  |  | 
|  | from __future__ import print_function | 
|  |  | 
|  | # pylint: disable=import-error | 
|  | import argparse | 
|  | import filecmp | 
|  | import os | 
|  | import sys | 
|  | import tempfile | 
|  |  | 
|  | from update_payload import common | 
|  | from update_payload import error | 
|  |  | 
|  | lib_dir = os.path.join(os.path.dirname(__file__), 'lib') | 
|  | if os.path.exists(lib_dir) and os.path.isdir(lib_dir): | 
|  | sys.path.insert(1, lib_dir) | 
|  | import update_payload | 
|  |  | 
|  |  | 
|  | _TYPE_FULL = 'full' | 
|  | _TYPE_DELTA = 'delta' | 
|  |  | 
|  |  | 
|  | def ParseArguments(argv): | 
|  | """Parse and validate command-line arguments. | 
|  |  | 
|  | Args: | 
|  | argv: command-line arguments to parse (excluding the program name) | 
|  |  | 
|  | Returns: | 
|  | Returns the arguments returned by the argument parser. | 
|  | """ | 
|  | parser = argparse.ArgumentParser( | 
|  | description=('Applies a Chrome OS update PAYLOAD to src_kern and ' | 
|  | 'src_root emitting dst_kern and dst_root, respectively. ' | 
|  | 'src_kern and src_root are only needed for delta payloads. ' | 
|  | 'When no partitions are provided, verifies the payload ' | 
|  | 'integrity.'), | 
|  | epilog=('Note: a payload may verify correctly but fail to apply, and ' | 
|  | 'vice versa; this is by design and can be thought of as static ' | 
|  | 'vs dynamic correctness. A payload that both verifies and ' | 
|  | 'applies correctly should be safe for use by the Chrome OS ' | 
|  | 'Update Engine. Use --check to verify a payload prior to ' | 
|  | 'applying it.'), | 
|  | formatter_class=argparse.RawDescriptionHelpFormatter | 
|  | ) | 
|  |  | 
|  | check_args = parser.add_argument_group('Checking payload integrity') | 
|  | check_args.add_argument('-c', '--check', action='store_true', default=False, | 
|  | help=('force payload integrity check (e.g. before ' | 
|  | 'applying)')) | 
|  | check_args.add_argument('-D', '--describe', action='store_true', | 
|  | default=False, | 
|  | help='Print a friendly description of the payload.') | 
|  | check_args.add_argument('-r', '--report', metavar='FILE', | 
|  | help="dump payload report (`-' for stdout)") | 
|  | check_args.add_argument('-t', '--type', dest='assert_type', | 
|  | help='assert the payload type', | 
|  | choices=[_TYPE_FULL, _TYPE_DELTA]) | 
|  | check_args.add_argument('-z', '--block-size', metavar='NUM', default=0, | 
|  | type=int, | 
|  | help='assert a non-default (4096) payload block size') | 
|  | check_args.add_argument('-u', '--allow-unhashed', action='store_true', | 
|  | default=False, help='allow unhashed operations') | 
|  | check_args.add_argument('-d', '--disabled_tests', default=(), metavar='', | 
|  | help=('space separated list of tests to disable. ' | 
|  | 'allowed options include: ' + | 
|  | ', '.join(update_payload.CHECKS_TO_DISABLE)), | 
|  | choices=update_payload.CHECKS_TO_DISABLE) | 
|  | check_args.add_argument('-k', '--key', metavar='FILE', | 
|  | help=('override standard key used for signature ' | 
|  | 'validation')) | 
|  | check_args.add_argument('-m', '--meta-sig', metavar='FILE', | 
|  | help='verify metadata against its signature') | 
|  | check_args.add_argument('-s', '--metadata-size', metavar='NUM', default=0, | 
|  | help='the metadata size to verify with the one in' | 
|  | ' payload') | 
|  | # TODO(tbrindus): deprecated in favour of --part_sizes | 
|  | check_args.add_argument('-p', '--root-part-size', metavar='NUM', | 
|  | default=0, type=int, | 
|  | help='override rootfs partition size auto-inference') | 
|  | check_args.add_argument('-P', '--kern-part-size', metavar='NUM', | 
|  | default=0, type=int, | 
|  | help='override kernel partition size auto-inference') | 
|  | check_args.add_argument('--part_sizes', metavar='NUM', nargs='+', type=int, | 
|  | help='override partition size auto-inference') | 
|  |  | 
|  | apply_args = parser.add_argument_group('Applying payload') | 
|  | # TODO(ahassani): Extent extract-bsdiff to puffdiff too. | 
|  | apply_args.add_argument('-x', '--extract-bsdiff', action='store_true', | 
|  | default=False, | 
|  | help=('use temp input/output files with BSDIFF ' | 
|  | 'operations (not in-place)')) | 
|  | apply_args.add_argument('--bspatch-path', metavar='FILE', | 
|  | help='use the specified bspatch binary') | 
|  | apply_args.add_argument('--puffpatch-path', metavar='FILE', | 
|  | help='use the specified puffpatch binary') | 
|  | # TODO(tbrindus): deprecated in favour of --dst_part_paths | 
|  | apply_args.add_argument('--dst_kern', metavar='FILE', | 
|  | help='destination kernel partition file') | 
|  | apply_args.add_argument('--dst_root', metavar='FILE', | 
|  | help='destination root partition file') | 
|  | # TODO(tbrindus): deprecated in favour of --src_part_paths | 
|  | apply_args.add_argument('--src_kern', metavar='FILE', | 
|  | help='source kernel partition file') | 
|  | apply_args.add_argument('--src_root', metavar='FILE', | 
|  | help='source root partition file') | 
|  | # TODO(tbrindus): deprecated in favour of --out_dst_part_paths | 
|  | apply_args.add_argument('--out_dst_kern', metavar='FILE', | 
|  | help='created destination kernel partition file') | 
|  | apply_args.add_argument('--out_dst_root', metavar='FILE', | 
|  | help='created destination root partition file') | 
|  |  | 
|  | apply_args.add_argument('--src_part_paths', metavar='FILE', nargs='+', | 
|  | help='source partitition files') | 
|  | apply_args.add_argument('--dst_part_paths', metavar='FILE', nargs='+', | 
|  | help='destination partition files') | 
|  | apply_args.add_argument('--out_dst_part_paths', metavar='FILE', nargs='+', | 
|  | help='created destination partition files') | 
|  |  | 
|  | parser.add_argument('payload', metavar='PAYLOAD', help='the payload file') | 
|  | parser.add_argument('--part_names', metavar='NAME', nargs='+', | 
|  | help='names of partitions') | 
|  |  | 
|  | # Parse command-line arguments. | 
|  | args = parser.parse_args(argv) | 
|  |  | 
|  | # TODO(tbrindus): temporary workaround to keep old-style flags from breaking | 
|  | # without having to handle both types in our code. Remove after flag usage is | 
|  | # removed from calling scripts. | 
|  | args.part_names = args.part_names or [common.KERNEL, common.ROOTFS] | 
|  | args.part_sizes = args.part_sizes or [args.kern_part_size, | 
|  | args.root_part_size] | 
|  | args.src_part_paths = args.src_part_paths or [args.src_kern, args.src_root] | 
|  | args.dst_part_paths = args.dst_part_paths or [args.dst_kern, args.dst_root] | 
|  | args.out_dst_part_paths = args.out_dst_part_paths or [args.out_dst_kern, | 
|  | args.out_dst_root] | 
|  |  | 
|  | # Make sure we don't have new dependencies on old flags by deleting them from | 
|  | # the namespace here. | 
|  | for old in ['kern_part_size', 'root_part_size', 'src_kern', 'src_root', | 
|  | 'dst_kern', 'dst_root', 'out_dst_kern', 'out_dst_root']: | 
|  | delattr(args, old) | 
|  |  | 
|  | # There are several options that imply --check. | 
|  | args.check = (args.check or args.report or args.assert_type or | 
|  | args.block_size or args.allow_unhashed or | 
|  | args.disabled_tests or args.meta_sig or args.key or | 
|  | any(args.part_sizes) or args.metadata_size) | 
|  |  | 
|  | for arg in ['part_sizes', 'src_part_paths', 'dst_part_paths', | 
|  | 'out_dst_part_paths']: | 
|  | if len(args.part_names) != len(getattr(args, arg, [])): | 
|  | parser.error('partitions in --%s do not match --part_names' % arg) | 
|  |  | 
|  | if all(args.dst_part_paths) or all(args.out_dst_part_paths): | 
|  | if all(args.src_part_paths): | 
|  | if args.assert_type == _TYPE_FULL: | 
|  | parser.error('%s payload does not accept source partition arguments' | 
|  | % _TYPE_FULL) | 
|  | else: | 
|  | args.assert_type = _TYPE_DELTA | 
|  | else: | 
|  | if args.assert_type == _TYPE_DELTA: | 
|  | parser.error('%s payload requires source partitions arguments' | 
|  | % _TYPE_DELTA) | 
|  | else: | 
|  | args.assert_type = _TYPE_FULL | 
|  | else: | 
|  | # Not applying payload. | 
|  | if args.extract_bsdiff: | 
|  | parser.error('--extract-bsdiff can only be used when applying payloads') | 
|  | if args.bspatch_path: | 
|  | parser.error('--bspatch-path can only be used when applying payloads') | 
|  | if args.puffpatch_path: | 
|  | parser.error('--puffpatch-path can only be used when applying payloads') | 
|  |  | 
|  | # By default, look for a metadata-signature file with a name based on the name | 
|  | # of the payload we are checking. We only do it if check was triggered. | 
|  | if args.check and not args.meta_sig: | 
|  | default_meta_sig = args.payload + '.metadata-signature' | 
|  | if os.path.isfile(default_meta_sig): | 
|  | args.meta_sig = default_meta_sig | 
|  | print('Using default metadata signature', args.meta_sig, file=sys.stderr) | 
|  |  | 
|  | return args | 
|  |  | 
|  |  | 
|  | def main(argv): | 
|  | # Parse and validate arguments. | 
|  | args = ParseArguments(argv[1:]) | 
|  |  | 
|  | with open(args.payload) as payload_file: | 
|  | payload = update_payload.Payload(payload_file) | 
|  | try: | 
|  | # Initialize payload. | 
|  | payload.Init() | 
|  |  | 
|  | if args.describe: | 
|  | payload.Describe() | 
|  |  | 
|  | # Perform payload integrity checks. | 
|  | if args.check: | 
|  | report_file = None | 
|  | do_close_report_file = False | 
|  | metadata_sig_file = None | 
|  | try: | 
|  | if args.report: | 
|  | if args.report == '-': | 
|  | report_file = sys.stdout | 
|  | else: | 
|  | report_file = open(args.report, 'w') | 
|  | do_close_report_file = True | 
|  |  | 
|  | part_sizes = dict(zip(args.part_names, args.part_sizes)) | 
|  | metadata_sig_file = args.meta_sig and open(args.meta_sig) | 
|  | payload.Check( | 
|  | pubkey_file_name=args.key, | 
|  | metadata_sig_file=metadata_sig_file, | 
|  | metadata_size=int(args.metadata_size), | 
|  | report_out_file=report_file, | 
|  | assert_type=args.assert_type, | 
|  | block_size=int(args.block_size), | 
|  | part_sizes=part_sizes, | 
|  | allow_unhashed=args.allow_unhashed, | 
|  | disabled_tests=args.disabled_tests) | 
|  | finally: | 
|  | if metadata_sig_file: | 
|  | metadata_sig_file.close() | 
|  | if do_close_report_file: | 
|  | report_file.close() | 
|  |  | 
|  | # Apply payload. | 
|  | if all(args.dst_part_paths) or all(args.out_dst_part_paths): | 
|  | dargs = {'bsdiff_in_place': not args.extract_bsdiff} | 
|  | if args.bspatch_path: | 
|  | dargs['bspatch_path'] = args.bspatch_path | 
|  | if args.puffpatch_path: | 
|  | dargs['puffpatch_path'] = args.puffpatch_path | 
|  | if args.assert_type == _TYPE_DELTA: | 
|  | dargs['old_parts'] = dict(zip(args.part_names, args.src_part_paths)) | 
|  |  | 
|  | out_dst_parts = {} | 
|  | file_handles = [] | 
|  | if all(args.out_dst_part_paths): | 
|  | for name, path in zip(args.part_names, args.out_dst_part_paths): | 
|  | handle = open(path, 'w+') | 
|  | file_handles.append(handle) | 
|  | out_dst_parts[name] = handle.name | 
|  | else: | 
|  | for name in args.part_names: | 
|  | handle = tempfile.NamedTemporaryFile() | 
|  | file_handles.append(handle) | 
|  | out_dst_parts[name] = handle.name | 
|  |  | 
|  | payload.Apply(out_dst_parts, **dargs) | 
|  |  | 
|  | # If destination kernel and rootfs partitions are not given, then this | 
|  | # just becomes an apply operation with no check. | 
|  | if all(args.dst_part_paths): | 
|  | # Prior to comparing, add the unused space past the filesystem | 
|  | # boundary in the new target partitions to become the same size as | 
|  | # the given partitions. This will truncate to larger size. | 
|  | for part_name, out_dst_part, dst_part in zip(args.part_names, | 
|  | file_handles, | 
|  | args.dst_part_paths): | 
|  | out_dst_part.truncate(os.path.getsize(dst_part)) | 
|  |  | 
|  | # Compare resulting partitions with the ones from the target image. | 
|  | if not filecmp.cmp(out_dst_part.name, dst_part): | 
|  | raise error.PayloadError( | 
|  | 'Resulting %s partition corrupted.' % part_name) | 
|  |  | 
|  | # Close the output files. If args.out_dst_* was not given, then these | 
|  | # files are created as temp files and will be deleted upon close(). | 
|  | for handle in file_handles: | 
|  | handle.close() | 
|  | except error.PayloadError, e: | 
|  | sys.stderr.write('Error: %s\n' % e) | 
|  | return 1 | 
|  |  | 
|  | return 0 | 
|  |  | 
|  |  | 
|  | if __name__ == '__main__': | 
|  | sys.exit(main(sys.argv)) |