blob: 55fe6d27b09ff10db5e0a8624986ab0f9b0cad3c [file] [log] [blame]
#!/usr/bin/env python
#
# Copyright (C) 2021 The Android Open Source Project
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""sign_virt_apex is a command line tool for sign the Virt APEX file.
Typical usage:
sign_virt_apex payload_key payload_dir
-v, --verbose
--verify
--avbtool path_to_avbtool
--signing_args args
sign_virt_apex uses external tools which are assumed to be available via PATH.
- avbtool (--avbtool can override the tool)
- lpmake, lpunpack, simg2img, img2simg, initrd_bootconfig
"""
import argparse
import binascii
import builtins
import hashlib
import os
import re
import shlex
import subprocess
import sys
import tempfile
import traceback
from concurrent import futures
# pylint: disable=line-too-long,consider-using-with
# Use executor to parallelize the invocation of external tools
# If a task depends on another, pass the future object of the previous task as wait list.
# Every future object created by a task should be consumed with AwaitAll()
# so that exceptions are propagated .
executor = futures.ThreadPoolExecutor()
# Temporary directory for unpacked super.img.
# We could put its creation/deletion into the task graph as well, but
# having it as a global setup is much simpler.
unpack_dir = tempfile.TemporaryDirectory()
# tasks created with Async() are kept in a list so that they are awaited
# before exit.
tasks = []
# create an async task and return a future value of it.
def Async(fn, *args, wait=None, **kwargs):
# wrap a function with AwaitAll()
def wrapped():
AwaitAll(wait)
fn(*args, **kwargs)
task = executor.submit(wrapped)
tasks.append(task)
return task
# waits for task (captured in fs as future values) with future.result()
# so that any exception raised during task can be raised upward.
def AwaitAll(fs):
if fs:
for f in fs:
f.result()
def ParseArgs(argv):
parser = argparse.ArgumentParser(description='Sign the Virt APEX')
parser.add_argument('--verify', action='store_true',
help='Verify the Virt APEX')
parser.add_argument(
'-v', '--verbose',
action='store_true',
help='verbose execution')
parser.add_argument(
'--avbtool',
default='avbtool',
help='Optional flag that specifies the AVB tool to use. Defaults to `avbtool`.')
parser.add_argument(
'--signing_args',
help='the extra signing arguments passed to avbtool.'
)
parser.add_argument(
'--key_override',
metavar="filename=key",
action='append',
help='Overrides a signing key for a file e.g. microdroid_bootloader=mykey (for testing)')
parser.add_argument(
'key',
help='path to the private key file.')
parser.add_argument(
'input_dir',
help='the directory having files to be packaged')
parser.add_argument(
'--do_not_update_bootconfigs',
action='store_true',
help='This will NOT update the vbmeta related bootconfigs while signing the apex.\
Used for testing only!!')
parser.add_argument('--do_not_validate_avb_version', action='store_true', help='Do not validate the avb_version when updating vbmeta bootconfig. Only use in tests!')
args = parser.parse_args(argv)
# preprocess --key_override into a map
args.key_overrides = {}
if args.key_override:
for pair in args.key_override:
name, key = pair.split('=')
args.key_overrides[name] = key
return args
def RunCommand(args, cmd, env=None, expected_return_values=None):
expected_return_values = expected_return_values or {0}
env = env or {}
env.update(os.environ.copy())
# TODO(b/193504286): we need a way to find other tool (cmd[0]) in various contexts
# e.g. sign_apex.py, sign_target_files_apk.py
if cmd[0] == 'avbtool':
cmd[0] = args.avbtool
if args.verbose:
print('Running: ' + ' '.join(cmd))
p = subprocess.Popen(
cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT, env=env, universal_newlines=True)
output, _ = p.communicate()
if args.verbose or p.returncode not in expected_return_values:
print(output.rstrip())
assert p.returncode in expected_return_values, (
'%d Failed to execute: ' + ' '.join(cmd)) % p.returncode
return (output, p.returncode)
def ReadBytesSize(value):
return int(value.removesuffix(' bytes'))
def ExtractAvbPubkey(args, key, output):
RunCommand(args, ['avbtool', 'extract_public_key',
'--key', key, '--output', output])
def is_lz4(args, path):
# error 44: Unrecognized header
result = RunCommand(args, ['lz4', '-t', path], expected_return_values={0, 44})
return result[1] == 0
def AvbInfo(args, image_path):
"""Parses avbtool --info image output
Args:
args: program arguments.
image_path: The path to the image, either raw or lz4 compressed
descriptor_name: Descriptor name of interest.
Returns:
A pair of
- a dict that contains VBMeta info. None if there's no VBMeta info.
- a list of descriptors.
"""
if not os.path.exists(image_path):
raise ValueError(f'Failed to find image: {image_path}')
if is_lz4(args, image_path):
with tempfile.NamedTemporaryFile() as decompressed_image:
RunCommand(args, ['lz4', '-d', '-f', image_path, decompressed_image.name])
return AvbInfo(args, decompressed_image.name)
output, ret_code = RunCommand(
args, ['avbtool', 'info_image', '--image', image_path], expected_return_values={0, 1})
if ret_code == 1:
return None, None
info, descriptors = {}, []
# Read `avbtool info_image` output as "key:value" lines
matcher = re.compile(r'^(\s*)([^:]+):\s*(.*)$')
def IterateLine(output):
for line in output.split('\n'):
line_info = matcher.match(line)
if not line_info:
continue
yield line_info.group(1), line_info.group(2), line_info.group(3)
gen = IterateLine(output)
def ReadDescriptors(cur_indent, cur_name, cur_value):
descriptor = cur_value if cur_name == 'Prop' else {}
descriptors.append((cur_name, descriptor))
for indent, key, value in gen:
if indent <= cur_indent:
# read descriptors recursively to pass the read key as descriptor name
ReadDescriptors(indent, key, value)
break
descriptor[key] = value
# Read VBMeta info
for _, key, value in gen:
if key == 'Descriptors':
ReadDescriptors(*next(gen))
break
info[key] = value
return info, descriptors
def find_all_values_by_key(pairs, key):
"""Find all the values of the key in the pairs."""
return [v for (k, v) in pairs if k == key]
# Extract properties from the descriptors of original vbmeta image,
# append to command as parameter.
def AppendPropArgument(cmd, descriptors):
for prop in find_all_values_by_key(descriptors, 'Prop'):
cmd.append('--prop')
result = re.match(r"(.+) -> '(.+)'", prop)
cmd.append(result.group(1) + ":" + result.group(2))
def check_resigned_image_avb_info(image_path, original_info, original_descriptors, args):
updated_info, updated_descriptors = AvbInfo(args, image_path)
assert original_info is not None, f'no avbinfo on original image: {image_path}'
assert updated_info is not None, f'no avbinfo on resigned image: {image_path}'
assert_different_value(original_info, updated_info, "Public key (sha1)", image_path)
updated_public_key = updated_info.pop("Public key (sha1)")
if not hasattr(check_resigned_image_avb_info, "new_public_key"):
check_resigned_image_avb_info.new_public_key = updated_public_key
else:
assert check_resigned_image_avb_info.new_public_key == updated_public_key, \
"All images should be resigned with the same public key. Expected public key (sha1):" \
f" {check_resigned_image_avb_info.new_public_key}, actual public key (sha1): " \
f"{updated_public_key}, Path: {image_path}"
original_info.pop("Public key (sha1)")
assert original_info == updated_info, \
f"Original info and updated info should be the same for {image_path}. " \
f"Original info: {original_info}, updated info: {updated_info}"
# Verify the descriptors of the original and updated images.
assert len(original_descriptors) == len(updated_descriptors), \
f"Number of descriptors should be the same for {image_path}. " \
f"Original descriptors: {original_descriptors}, updated descriptors: {updated_descriptors}"
original_prop_descriptors = sorted(find_all_values_by_key(original_descriptors, "Prop"))
updated_prop_descriptors = sorted(find_all_values_by_key(updated_descriptors, "Prop"))
assert original_prop_descriptors == updated_prop_descriptors, \
f"Prop descriptors should be the same for {image_path}. " \
f"Original prop descriptors: {original_prop_descriptors}, " \
f"updated prop descriptors: {updated_prop_descriptors}"
# Remove digest from hash descriptors before comparing, since some digests should change.
original_hash_descriptors = extract_hash_descriptors(original_descriptors, drop_digest)
updated_hash_descriptors = extract_hash_descriptors(updated_descriptors, drop_digest)
assert original_hash_descriptors == updated_hash_descriptors, \
f"Hash descriptors' parameters should be the same for {image_path}. " \
f"Original hash descriptors: {original_hash_descriptors}, " \
f"updated hash descriptors: {updated_hash_descriptors}"
def drop_digest(descriptor):
return {k: v for k, v in descriptor.items() if k != "Digest"}
def AddHashFooter(args, key, image_path, additional_images=()):
if os.path.basename(image_path) in args.key_overrides:
key = args.key_overrides[os.path.basename(image_path)]
info, descriptors = AvbInfo(args, image_path)
assert info is not None, f'no avbinfo: {image_path}'
# Extract hash descriptor of original image.
hash_descriptors_original = extract_hash_descriptors(descriptors, drop_digest)
for additional_image in additional_images:
_, additional_desc = AvbInfo(args, additional_image)
hash_descriptors = extract_hash_descriptors(additional_desc, drop_digest)
for k, v in hash_descriptors.items():
assert v == hash_descriptors_original[k], \
f"Hash descriptor of {k} in {additional_image} and {image_path} should be " \
f"the same. {additional_image}: {v}, {image_path}: {hash_descriptors_original[k]}"
del hash_descriptors_original[k]
assert len(hash_descriptors_original) == 1, \
f"Only one hash descriptor is expected for {image_path} after removing " \
f"additional images. Hash descriptors: {hash_descriptors_original}"
[(original_image_partition_name, original_image_descriptor)] = hash_descriptors_original.items()
assert info["Original image size"] == original_image_descriptor["Image Size"], \
f"Original image size should be the same as the image size in the hash descriptor " \
f"for {image_path}. Original image size: {info['Original image size']}, " \
f"image size in the hash descriptor: {original_image_descriptor['Image Size']}"
partition_size = str(ReadBytesSize(info['Image size']))
algorithm = info['Algorithm']
original_image_salt = original_image_descriptor['Salt']
cmd = ['avbtool', 'add_hash_footer',
'--key', key,
'--algorithm', algorithm,
'--partition_name', original_image_partition_name,
'--salt', original_image_salt,
'--partition_size', partition_size,
'--image', image_path]
AppendPropArgument(cmd, descriptors)
if args.signing_args:
cmd.extend(shlex.split(args.signing_args))
for additional_image in additional_images:
cmd.extend(['--include_descriptors_from_image', additional_image])
cmd.extend(['--rollback_index', info['Rollback Index']])
RunCommand(args, cmd)
check_resigned_image_avb_info(image_path, info, descriptors, args)
def AddHashTreeFooter(args, key, image_path):
if os.path.basename(image_path) in args.key_overrides:
key = args.key_overrides[os.path.basename(image_path)]
info, descriptors = AvbInfo(args, image_path)
if info:
descriptor = find_all_values_by_key(descriptors, 'Hashtree descriptor')[0]
image_size = ReadBytesSize(info['Image size'])
algorithm = info['Algorithm']
partition_name = descriptor['Partition Name']
hash_algorithm = descriptor['Hash Algorithm']
salt = descriptor['Salt']
partition_size = str(image_size)
cmd = ['avbtool', 'add_hashtree_footer',
'--key', key,
'--algorithm', algorithm,
'--partition_name', partition_name,
'--partition_size', partition_size,
'--do_not_generate_fec',
'--hash_algorithm', hash_algorithm,
'--salt', salt,
'--image', image_path]
AppendPropArgument(cmd, descriptors)
if args.signing_args:
cmd.extend(shlex.split(args.signing_args))
RunCommand(args, cmd)
check_resigned_image_avb_info(image_path, info, descriptors, args)
def UpdateVbmetaBootconfig(args, initrds, vbmeta_img):
# Update the bootconfigs in ramdisk
def detach_bootconfigs(initrd_bc, initrd, bc):
cmd = ['initrd_bootconfig', 'detach', initrd_bc, initrd, bc]
RunCommand(args, cmd)
def attach_bootconfigs(initrd_bc, initrd, bc):
cmd = ['initrd_bootconfig', 'attach',
initrd, bc, '--output', initrd_bc]
RunCommand(args, cmd)
# Validate that avb version used while signing the apex is the same as used by build server
def validate_avb_version(bootconfigs):
cmd = ['avbtool', 'version']
stdout, _ = RunCommand(args, cmd)
avb_version_curr = stdout.split(" ")[1].strip()
avb_version_curr = avb_version_curr[0:avb_version_curr.rfind('.')]
avb_version_bc = re.search(
r"androidboot.vbmeta.avb_version = \"([^\"]*)\"", bootconfigs).group(1)
if avb_version_curr != avb_version_bc:
raise builtins.Exception(f'AVB version mismatch between current & one & \
used to build bootconfigs:{avb_version_curr}&{avb_version_bc}')
def calc_vbmeta_digest():
cmd = ['avbtool', 'calculate_vbmeta_digest', '--image',
vbmeta_img, '--hash_algorithm', 'sha256']
stdout, _ = RunCommand(args, cmd)
return stdout.strip()
def calc_vbmeta_size():
cmd = ['avbtool', 'info_image', '--image', vbmeta_img]
stdout, _ = RunCommand(args, cmd)
size = 0
for line in stdout.split("\n"):
line = line.split(":")
if line[0] in ['Header Block', 'Authentication Block', 'Auxiliary Block']:
size += int(line[1].strip()[0:-6])
return size
def update_vbmeta_digest(bootconfigs):
# Update androidboot.vbmeta.digest in bootconfigs
result = re.search(
r"androidboot.vbmeta.digest = \"[^\"]*\"", bootconfigs)
if not result:
raise ValueError("Failed to find androidboot.vbmeta.digest")
return bootconfigs.replace(result.group(),
f'androidboot.vbmeta.digest = "{calc_vbmeta_digest()}"')
def update_vbmeta_size(bootconfigs):
# Update androidboot.vbmeta.size in bootconfigs
result = re.search(r"androidboot.vbmeta.size = [0-9]+", bootconfigs)
if not result:
raise ValueError("Failed to find androidboot.vbmeta.size")
return bootconfigs.replace(result.group(),
f'androidboot.vbmeta.size = {calc_vbmeta_size()}')
with tempfile.TemporaryDirectory() as work_dir:
tmp_initrd = os.path.join(work_dir, 'initrd')
tmp_bc = os.path.join(work_dir, 'bc')
for initrd in initrds:
detach_bootconfigs(initrd, tmp_initrd, tmp_bc)
bc_file = open(tmp_bc, "rt", encoding="utf-8")
bc_data = bc_file.read()
if not args.do_not_validate_avb_version:
validate_avb_version(bc_data)
bc_data = update_vbmeta_digest(bc_data)
bc_data = update_vbmeta_size(bc_data)
bc_file.close()
bc_file = open(tmp_bc, "wt", encoding="utf-8")
bc_file.write(bc_data)
bc_file.flush()
attach_bootconfigs(initrd, tmp_initrd, tmp_bc)
def MakeVbmetaImage(args, key, vbmeta_img, images=None, chained_partitions=None):
if os.path.basename(vbmeta_img) in args.key_overrides:
key = args.key_overrides[os.path.basename(vbmeta_img)]
info, descriptors = AvbInfo(args, vbmeta_img)
if info is None:
return
with tempfile.TemporaryDirectory() as work_dir:
algorithm = info['Algorithm']
rollback_index = info['Rollback Index']
rollback_index_location = info['Rollback Index Location']
cmd = ['avbtool', 'make_vbmeta_image',
'--key', key,
'--algorithm', algorithm,
'--rollback_index', rollback_index,
'--rollback_index_location', rollback_index_location,
'--output', vbmeta_img]
if images:
for img in images:
cmd.extend(['--include_descriptors_from_image', img])
# replace pubkeys of chained_partitions as well
for name, descriptor in descriptors:
if name == 'Chain Partition descriptor':
part_name = descriptor['Partition Name']
ril = descriptor['Rollback Index Location']
part_key = chained_partitions[part_name]
avbpubkey = os.path.join(work_dir, part_name + '.avbpubkey')
ExtractAvbPubkey(args, part_key, avbpubkey)
cmd.extend(['--chain_partition', f'{part_name}:{ril}:{avbpubkey}'])
if args.signing_args:
cmd.extend(shlex.split(args.signing_args))
RunCommand(args, cmd)
check_resigned_image_avb_info(vbmeta_img, info, descriptors, args)
# libavb expects to be able to read the maximum vbmeta size, so we must provide a partition
# which matches this or the read will fail.
with open(vbmeta_img, 'a', encoding='utf8') as f:
f.truncate(65536)
def UnpackSuperImg(args, super_img, work_dir):
tmp_super_img = os.path.join(work_dir, 'super.img')
RunCommand(args, ['simg2img', super_img, tmp_super_img])
RunCommand(args, ['lpunpack', tmp_super_img, work_dir])
def MakeSuperImage(args, partitions, output):
with tempfile.TemporaryDirectory() as work_dir:
cmd = ['lpmake', '--device-size=auto', '--metadata-slots=2', # A/B
'--metadata-size=65536', '--sparse', '--output=' + output]
for part, img in partitions.items():
tmp_img = os.path.join(work_dir, part)
RunCommand(args, ['img2simg', img, tmp_img])
image_arg = f'--image={part}={img}'
partition_arg = f'--partition={part}:readonly:{os.path.getsize(img)}:default'
cmd.extend([image_arg, partition_arg])
RunCommand(args, cmd)
def GenVbmetaImage(args, image, output, partition_name, salt):
cmd = ['avbtool', 'add_hash_footer', '--dynamic_partition_size',
'--do_not_append_vbmeta_image',
'--partition_name', partition_name,
'--salt', salt,
'--image', image,
'--output_vbmeta_image', output]
RunCommand(args, cmd)
gki_versions = ['android14-6.1-pkvm_experimental']
# dict of (key, file) for re-sign/verification. keys are un-versioned for readability.
virt_apex_non_gki_files = {
'kernel': 'etc/fs/microdroid_kernel',
'vbmeta.img': 'etc/fs/microdroid_vbmeta.img',
'super.img': 'etc/fs/microdroid_super.img',
'initrd_normal.img': 'etc/microdroid_initrd_normal.img',
'initrd_debuggable.img': 'etc/microdroid_initrd_debuggable.img',
'rialto': 'etc/rialto.bin',
}
def TargetFiles(input_dir):
ret = {k: os.path.join(input_dir, v) for k, v in virt_apex_non_gki_files.items()}
for ver in gki_versions:
kernel = os.path.join(input_dir, f'etc/fs/microdroid_gki-{ver}_kernel')
initrd_normal = os.path.join(input_dir, f'etc/microdroid_gki-{ver}_initrd_normal.img')
initrd_debug = os.path.join(input_dir, f'etc/microdroid_gki-{ver}_initrd_debuggable.img')
if os.path.isfile(kernel):
ret[f'gki-{ver}_kernel'] = kernel
ret[f'gki-{ver}_initrd_normal.img'] = initrd_normal
ret[f'gki-{ver}_initrd_debuggable.img'] = initrd_debug
return ret
def IsInitrdImage(path):
return path.endswith('initrd_normal.img') or path.endswith('initrd_debuggable.img')
def SignVirtApex(args):
key = args.key
input_dir = args.input_dir
files = TargetFiles(input_dir)
# unpacked files (will be unpacked from super.img below)
system_a_img = os.path.join(unpack_dir.name, 'system_a.img')
vendor_a_img = os.path.join(unpack_dir.name, 'vendor_a.img')
# re-sign super.img
# 1. unpack super.img
# 2. resign system and vendor (if exists)
# 3. repack super.img out of resigned system and vendor (if exists)
UnpackSuperImg(args, files['super.img'], unpack_dir.name)
system_a_f = Async(AddHashTreeFooter, args, key, system_a_img)
partitions = {"system_a": system_a_img}
images = [system_a_img]
images_f = [system_a_f]
# if vendor_a.img exists, resign it
if os.path.exists(vendor_a_img):
partitions.update({'vendor_a': vendor_a_img})
images.append(vendor_a_img)
vendor_a_f = Async(AddHashTreeFooter, args, key, vendor_a_img)
images_f.append(vendor_a_f)
Async(MakeSuperImage, args, partitions,
files['super.img'], wait=images_f)
# re-generate vbmeta from re-signed system_a.img
vbmeta_f = Async(MakeVbmetaImage, args, key, files['vbmeta.img'],
images=images,
wait=images_f)
vbmeta_bc_f = None
if not args.do_not_update_bootconfigs:
initrd_files = [v for k, v in files.items() if IsInitrdImage(k)]
vbmeta_bc_f = Async(UpdateVbmetaBootconfig, args, initrd_files,
files['vbmeta.img'],
wait=[vbmeta_f])
# Re-sign kernel. Note kernel's vbmeta contain addition descriptor from ramdisk(s)
def resign_decompressed_kernel(kernel_file, initrd_normal_file, initrd_debug_file):
_, kernel_image_descriptors = AvbInfo(args, kernel_file)
salts = extract_hash_descriptors(
kernel_image_descriptors, lambda descriptor: descriptor['Salt'])
initrd_normal_hashdesc = tempfile.NamedTemporaryFile(delete=False).name
initrd_debug_hashdesc = tempfile.NamedTemporaryFile(delete=False).name
initrd_n_f = Async(GenVbmetaImage, args, initrd_normal_file,
initrd_normal_hashdesc, "initrd_normal", salts["initrd_normal"],
wait=[vbmeta_bc_f] if vbmeta_bc_f is not None else [])
initrd_d_f = Async(GenVbmetaImage, args, initrd_debug_file,
initrd_debug_hashdesc, "initrd_debug", salts["initrd_debug"],
wait=[vbmeta_bc_f] if vbmeta_bc_f is not None else [])
return Async(AddHashFooter, args, key, kernel_file,
additional_images=[initrd_normal_hashdesc, initrd_debug_hashdesc],
wait=[initrd_n_f, initrd_d_f])
def resign_compressed_kernel(kernel_file, initrd_normal_file, initrd_debug_file):
# decompress, re-sign, compress again
with tempfile.TemporaryDirectory() as work_dir:
decompressed_kernel_file = os.path.join(work_dir, os.path.basename(kernel_file))
RunCommand(args, ['lz4', '-d', kernel_file, decompressed_kernel_file])
resign_decompressed_kernel(decompressed_kernel_file, initrd_normal_file,
initrd_debug_file).result()
RunCommand(args, ['lz4', '-9', '-f', decompressed_kernel_file, kernel_file])
def resign_kernel(kernel, initrd_normal, initrd_debug):
kernel_file = files[kernel]
initrd_normal_file = files[initrd_normal]
initrd_debug_file = files[initrd_debug]
# kernel may be compressed with lz4.
if is_lz4(args, kernel_file):
return Async(resign_compressed_kernel, kernel_file, initrd_normal_file,
initrd_debug_file)
else:
return resign_decompressed_kernel(kernel_file, initrd_normal_file, initrd_debug_file)
_, original_kernel_descriptors = AvbInfo(args, files['kernel'])
resign_kernel_tasks = [resign_kernel('kernel', 'initrd_normal.img', 'initrd_debuggable.img')]
original_kernels = {"kernel" : original_kernel_descriptors}
for ver in gki_versions:
if f'gki-{ver}_kernel' in files:
kernel_name = f'gki-{ver}_kernel'
_, original_kernel_descriptors = AvbInfo(args, files[kernel_name])
task = resign_kernel(
kernel_name,
f'gki-{ver}_initrd_normal.img',
f'gki-{ver}_initrd_debuggable.img')
resign_kernel_tasks.append(task)
original_kernels[kernel_name] = original_kernel_descriptors
# Re-sign rialto if it exists. Rialto only exists in arm64 environment.
if os.path.exists(files['rialto']):
update_initrd_digests_task = Async(
update_initrd_digests_of_kernels_in_rialto, original_kernels, args, files,
wait=resign_kernel_tasks)
Async(resign_rialto, args, key, files['rialto'], wait=[update_initrd_digests_task])
def resign_rialto(args, key, rialto_path):
_, original_descriptors = AvbInfo(args, rialto_path)
AddHashFooter(args, key, rialto_path)
# Verify the new AVB footer.
updated_info, updated_descriptors = AvbInfo(args, rialto_path)
assert len(updated_descriptors) == 2, \
f"There should be two descriptors for rialto. Updated descriptors: {updated_descriptors}"
updated_prop = find_all_values_by_key(updated_descriptors, "Prop")
assert len(updated_prop) == 1, "There should be only one Prop descriptor for rialto. " \
f"Updated descriptors: {updated_descriptors}"
assert updated_info["Rollback Index"] != "0", "Rollback index should not be zero for rialto."
# Verify the only hash descriptor of rialto.
updated_hash_descriptors = extract_hash_descriptors(updated_descriptors)
assert len(updated_hash_descriptors) == 1, \
f"There should be only one hash descriptor for rialto. " \
f"Updated hash descriptors: {updated_hash_descriptors}"
# Since salt is not updated, the change of digest reflects the change of content of rialto
# kernel.
if not args.do_not_update_bootconfigs:
[(_, original_descriptor)] = extract_hash_descriptors(original_descriptors).items()
[(_, updated_descriptor)] = updated_hash_descriptors.items()
assert_different_value(original_descriptor, updated_descriptor, "Digest",
"rialto_hash_descriptor")
def assert_different_value(original, updated, key, context):
assert original[key] != updated[key], \
f"Value of '{key}' should change for '{context}'" \
f"Original value: {original[key]}, updated value: {updated[key]}"
def update_initrd_digests_of_kernels_in_rialto(original_kernels, args, files):
# Update the hashes of initrd_normal and initrd_debug in rialto if the
# bootconfigs in them are updated.
if args.do_not_update_bootconfigs:
return
with open(files['rialto'], "rb") as file:
content = file.read()
for kernel_name, descriptors in original_kernels.items():
content = update_initrd_digests_in_rialto(
descriptors, args, files, kernel_name, content)
with open(files['rialto'], "wb") as file:
file.write(content)
def update_initrd_digests_in_rialto(
original_descriptors, args, files, kernel_name, content):
_, updated_descriptors = AvbInfo(args, files[kernel_name])
original_digests = extract_hash_descriptors(
original_descriptors, lambda x: binascii.unhexlify(x['Digest']))
updated_digests = extract_hash_descriptors(
updated_descriptors, lambda x: binascii.unhexlify(x['Digest']))
assert original_digests.pop("boot") == updated_digests.pop("boot"), \
"Hash descriptor of boot should not change for " + kernel_name + \
f"\nOriginal descriptors: {original_descriptors}, " \
f"\nUpdated descriptors: {updated_descriptors}"
# Check that the original and updated digests are different before updating rialto.
partition_names = {'initrd_normal', 'initrd_debug'}
assert set(original_digests.keys()) == set(updated_digests.keys()) == partition_names, \
f"Original digests' partitions should be {partition_names}. " \
f"Original digests: {original_digests}. Updated digests: {updated_digests}"
assert set(original_digests.values()).isdisjoint(updated_digests.values()), \
"Digests of initrd_normal and initrd_debug should change. " \
f"Original descriptors: {original_descriptors}, " \
f"updated descriptors: {updated_descriptors}"
for partition_name, original_digest in original_digests.items():
updated_digest = updated_digests[partition_name]
assert len(original_digest) == len(updated_digest), \
f"Length of original_digest and updated_digest must be the same for {partition_name}." \
f" Original digest: {original_digest}, updated digest: {updated_digest}"
new_content = content.replace(original_digest, updated_digest)
assert len(new_content) == len(content), \
"Length of new_content and content must be the same."
assert new_content != content, \
f"original digest of the partition {partition_name} not found."
content = new_content
return content
def extract_hash_descriptors(descriptors, f=lambda x: x):
return {desc["Partition Name"]: f(desc) for desc in
find_all_values_by_key(descriptors, "Hash descriptor")}
def VerifyVirtApex(args):
key = args.key
input_dir = args.input_dir
files = TargetFiles(input_dir)
# unpacked files
UnpackSuperImg(args, files['super.img'], unpack_dir.name)
system_a_img = os.path.join(unpack_dir.name, 'system_a.img')
# Read pubkey digest from the input key
with tempfile.NamedTemporaryFile() as pubkey_file:
ExtractAvbPubkey(args, key, pubkey_file.name)
with open(pubkey_file.name, 'rb') as f:
pubkey = f.read()
pubkey_digest = hashlib.sha1(pubkey).hexdigest()
def check_avb_pubkey(file):
info, _ = AvbInfo(args, file)
assert info is not None, f'no avbinfo: {file}'
assert info['Public key (sha1)'] == pubkey_digest, f'pubkey mismatch: {file}'
for k, f in files.items():
if IsInitrdImage(k):
# TODO(b/245277660): Verify that ramdisks contain the correct vbmeta digest
continue
if k == 'rialto' and not os.path.exists(f):
# Rialto only exists in arm64 environment.
continue
if k == 'super.img':
Async(check_avb_pubkey, system_a_img)
else:
# Check pubkey for other files using avbtool
Async(check_avb_pubkey, f)
def main(argv):
try:
args = ParseArgs(argv)
if args.verify:
VerifyVirtApex(args)
else:
SignVirtApex(args)
# ensure all tasks are completed without exceptions
AwaitAll(tasks)
except: # pylint: disable=bare-except
traceback.print_exc()
sys.exit(1)
if __name__ == '__main__':
main(sys.argv[1:])