|  | # Copyright (C) 2008 The Android Open Source Project | 
|  | # | 
|  | # Licensed under the Apache License, Version 2.0 (the "License"); | 
|  | # you may not use this file except in compliance with the License. | 
|  | # You may obtain a copy of the License at | 
|  | # | 
|  | #      http://www.apache.org/licenses/LICENSE-2.0 | 
|  | # | 
|  | # Unless required by applicable law or agreed to in writing, software | 
|  | # distributed under the License is distributed on an "AS IS" BASIS, | 
|  | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | 
|  | # See the License for the specific language governing permissions and | 
|  | # limitations under the License. | 
|  |  | 
|  | import copy | 
|  | import errno | 
|  | import getopt | 
|  | import getpass | 
|  | import imp | 
|  | import os | 
|  | import platform | 
|  | import re | 
|  | import shlex | 
|  | import shutil | 
|  | import subprocess | 
|  | import sys | 
|  | import tempfile | 
|  | import threading | 
|  | import time | 
|  | import zipfile | 
|  |  | 
|  | try: | 
|  | from hashlib import sha1 as sha1 | 
|  | except ImportError: | 
|  | from sha import sha as sha1 | 
|  |  | 
|  | # missing in Python 2.4 and before | 
|  | if not hasattr(os, "SEEK_SET"): | 
|  | os.SEEK_SET = 0 | 
|  |  | 
|  | class Options(object): pass | 
|  | OPTIONS = Options() | 
|  | OPTIONS.search_path = "out/host/linux-x86" | 
|  | OPTIONS.signapk_path = "framework/signapk.jar"  # Relative to search_path | 
|  | OPTIONS.extra_signapk_args = [] | 
|  | OPTIONS.java_path = "java"  # Use the one on the path by default. | 
|  | OPTIONS.public_key_suffix = ".x509.pem" | 
|  | OPTIONS.private_key_suffix = ".pk8" | 
|  | OPTIONS.verbose = False | 
|  | OPTIONS.tempfiles = [] | 
|  | OPTIONS.device_specific = None | 
|  | OPTIONS.extras = {} | 
|  | OPTIONS.info_dict = None | 
|  |  | 
|  |  | 
|  | # Values for "certificate" in apkcerts that mean special things. | 
|  | SPECIAL_CERT_STRINGS = ("PRESIGNED", "EXTERNAL") | 
|  |  | 
|  |  | 
|  | class ExternalError(RuntimeError): pass | 
|  |  | 
|  |  | 
|  | def Run(args, **kwargs): | 
|  | """Create and return a subprocess.Popen object, printing the command | 
|  | line on the terminal if -v was specified.""" | 
|  | if OPTIONS.verbose: | 
|  | print "  running: ", " ".join(args) | 
|  | return subprocess.Popen(args, **kwargs) | 
|  |  | 
|  |  | 
|  | def CloseInheritedPipes(): | 
|  | """ Gmake in MAC OS has file descriptor (PIPE) leak. We close those fds | 
|  | before doing other work.""" | 
|  | if platform.system() != "Darwin": | 
|  | return | 
|  | for d in range(3, 1025): | 
|  | try: | 
|  | stat = os.fstat(d) | 
|  | if stat is not None: | 
|  | pipebit = stat[0] & 0x1000 | 
|  | if pipebit != 0: | 
|  | os.close(d) | 
|  | except OSError: | 
|  | pass | 
|  |  | 
|  |  | 
|  | def LoadInfoDict(zip): | 
|  | """Read and parse the META/misc_info.txt key/value pairs from the | 
|  | input target files and return a dict.""" | 
|  |  | 
|  | d = {} | 
|  | try: | 
|  | for line in zip.read("META/misc_info.txt").split("\n"): | 
|  | line = line.strip() | 
|  | if not line or line.startswith("#"): continue | 
|  | k, v = line.split("=", 1) | 
|  | d[k] = v | 
|  | except KeyError: | 
|  | # ok if misc_info.txt doesn't exist | 
|  | pass | 
|  |  | 
|  | # backwards compatibility: These values used to be in their own | 
|  | # files.  Look for them, in case we're processing an old | 
|  | # target_files zip. | 
|  |  | 
|  | if "mkyaffs2_extra_flags" not in d: | 
|  | try: | 
|  | d["mkyaffs2_extra_flags"] = zip.read("META/mkyaffs2-extra-flags.txt").strip() | 
|  | except KeyError: | 
|  | # ok if flags don't exist | 
|  | pass | 
|  |  | 
|  | if "recovery_api_version" not in d: | 
|  | try: | 
|  | d["recovery_api_version"] = zip.read("META/recovery-api-version.txt").strip() | 
|  | except KeyError: | 
|  | raise ValueError("can't find recovery API version in input target-files") | 
|  |  | 
|  | if "tool_extensions" not in d: | 
|  | try: | 
|  | d["tool_extensions"] = zip.read("META/tool-extensions.txt").strip() | 
|  | except KeyError: | 
|  | # ok if extensions don't exist | 
|  | pass | 
|  |  | 
|  | if "fstab_version" not in d: | 
|  | d["fstab_version"] = "1" | 
|  |  | 
|  | try: | 
|  | data = zip.read("META/imagesizes.txt") | 
|  | for line in data.split("\n"): | 
|  | if not line: continue | 
|  | name, value = line.split(" ", 1) | 
|  | if not value: continue | 
|  | if name == "blocksize": | 
|  | d[name] = value | 
|  | else: | 
|  | d[name + "_size"] = value | 
|  | except KeyError: | 
|  | pass | 
|  |  | 
|  | def makeint(key): | 
|  | if key in d: | 
|  | d[key] = int(d[key], 0) | 
|  |  | 
|  | makeint("recovery_api_version") | 
|  | makeint("blocksize") | 
|  | makeint("system_size") | 
|  | makeint("userdata_size") | 
|  | makeint("cache_size") | 
|  | makeint("recovery_size") | 
|  | makeint("boot_size") | 
|  | makeint("fstab_version") | 
|  |  | 
|  | d["fstab"] = LoadRecoveryFSTab(zip, d["fstab_version"]) | 
|  | d["build.prop"] = LoadBuildProp(zip) | 
|  | return d | 
|  |  | 
|  | def LoadBuildProp(zip): | 
|  | try: | 
|  | data = zip.read("SYSTEM/build.prop") | 
|  | except KeyError: | 
|  | print "Warning: could not find SYSTEM/build.prop in %s" % zip | 
|  | data = "" | 
|  |  | 
|  | d = {} | 
|  | for line in data.split("\n"): | 
|  | line = line.strip() | 
|  | if not line or line.startswith("#"): continue | 
|  | name, value = line.split("=", 1) | 
|  | d[name] = value | 
|  | return d | 
|  |  | 
|  | def LoadRecoveryFSTab(zip, fstab_version): | 
|  | class Partition(object): | 
|  | pass | 
|  |  | 
|  | try: | 
|  | data = zip.read("RECOVERY/RAMDISK/etc/recovery.fstab") | 
|  | except KeyError: | 
|  | print "Warning: could not find RECOVERY/RAMDISK/etc/recovery.fstab in %s." % zip | 
|  | data = "" | 
|  |  | 
|  | if fstab_version == 1: | 
|  | d = {} | 
|  | for line in data.split("\n"): | 
|  | line = line.strip() | 
|  | if not line or line.startswith("#"): continue | 
|  | pieces = line.split() | 
|  | if not (3 <= len(pieces) <= 4): | 
|  | raise ValueError("malformed recovery.fstab line: \"%s\"" % (line,)) | 
|  |  | 
|  | p = Partition() | 
|  | p.mount_point = pieces[0] | 
|  | p.fs_type = pieces[1] | 
|  | p.device = pieces[2] | 
|  | p.length = 0 | 
|  | options = None | 
|  | if len(pieces) >= 4: | 
|  | if pieces[3].startswith("/"): | 
|  | p.device2 = pieces[3] | 
|  | if len(pieces) >= 5: | 
|  | options = pieces[4] | 
|  | else: | 
|  | p.device2 = None | 
|  | options = pieces[3] | 
|  | else: | 
|  | p.device2 = None | 
|  |  | 
|  | if options: | 
|  | options = options.split(",") | 
|  | for i in options: | 
|  | if i.startswith("length="): | 
|  | p.length = int(i[7:]) | 
|  | else: | 
|  | print "%s: unknown option \"%s\"" % (p.mount_point, i) | 
|  |  | 
|  | d[p.mount_point] = p | 
|  |  | 
|  | elif fstab_version == 2: | 
|  | d = {} | 
|  | for line in data.split("\n"): | 
|  | line = line.strip() | 
|  | if not line or line.startswith("#"): continue | 
|  | pieces = line.split() | 
|  | if len(pieces) != 5: | 
|  | raise ValueError("malformed recovery.fstab line: \"%s\"" % (line,)) | 
|  |  | 
|  | # Ignore entries that are managed by vold | 
|  | options = pieces[4] | 
|  | if "voldmanaged=" in options: continue | 
|  |  | 
|  | # It's a good line, parse it | 
|  | p = Partition() | 
|  | p.device = pieces[0] | 
|  | p.mount_point = pieces[1] | 
|  | p.fs_type = pieces[2] | 
|  | p.device2 = None | 
|  | p.length = 0 | 
|  |  | 
|  | options = options.split(",") | 
|  | for i in options: | 
|  | if i.startswith("length="): | 
|  | p.length = int(i[7:]) | 
|  | else: | 
|  | # Ignore all unknown options in the unified fstab | 
|  | continue | 
|  |  | 
|  | d[p.mount_point] = p | 
|  |  | 
|  | else: | 
|  | raise ValueError("Unknown fstab_version: \"%d\"" % (fstab_version,)) | 
|  |  | 
|  | return d | 
|  |  | 
|  |  | 
|  | def DumpInfoDict(d): | 
|  | for k, v in sorted(d.items()): | 
|  | print "%-25s = (%s) %s" % (k, type(v).__name__, v) | 
|  |  | 
|  | def BuildBootableImage(sourcedir, fs_config_file, info_dict=None): | 
|  | """Take a kernel, cmdline, and ramdisk directory from the input (in | 
|  | 'sourcedir'), and turn them into a boot image.  Return the image | 
|  | data, or None if sourcedir does not appear to contains files for | 
|  | building the requested image.""" | 
|  |  | 
|  | if (not os.access(os.path.join(sourcedir, "RAMDISK"), os.F_OK) or | 
|  | not os.access(os.path.join(sourcedir, "kernel"), os.F_OK)): | 
|  | return None | 
|  |  | 
|  | if info_dict is None: | 
|  | info_dict = OPTIONS.info_dict | 
|  |  | 
|  | ramdisk_img = tempfile.NamedTemporaryFile() | 
|  | img = tempfile.NamedTemporaryFile() | 
|  |  | 
|  | if os.access(fs_config_file, os.F_OK): | 
|  | cmd = ["mkbootfs", "-f", fs_config_file, os.path.join(sourcedir, "RAMDISK")] | 
|  | else: | 
|  | cmd = ["mkbootfs", os.path.join(sourcedir, "RAMDISK")] | 
|  | p1 = Run(cmd, stdout=subprocess.PIPE) | 
|  | p2 = Run(["minigzip"], | 
|  | stdin=p1.stdout, stdout=ramdisk_img.file.fileno()) | 
|  |  | 
|  | p2.wait() | 
|  | p1.wait() | 
|  | assert p1.returncode == 0, "mkbootfs of %s ramdisk failed" % (targetname,) | 
|  | assert p2.returncode == 0, "minigzip of %s ramdisk failed" % (targetname,) | 
|  |  | 
|  | cmd = ["mkbootimg", "--kernel", os.path.join(sourcedir, "kernel")] | 
|  |  | 
|  | fn = os.path.join(sourcedir, "cmdline") | 
|  | if os.access(fn, os.F_OK): | 
|  | cmd.append("--cmdline") | 
|  | cmd.append(open(fn).read().rstrip("\n")) | 
|  |  | 
|  | fn = os.path.join(sourcedir, "base") | 
|  | if os.access(fn, os.F_OK): | 
|  | cmd.append("--base") | 
|  | cmd.append(open(fn).read().rstrip("\n")) | 
|  |  | 
|  | fn = os.path.join(sourcedir, "pagesize") | 
|  | if os.access(fn, os.F_OK): | 
|  | cmd.append("--pagesize") | 
|  | cmd.append(open(fn).read().rstrip("\n")) | 
|  |  | 
|  | args = info_dict.get("mkbootimg_args", None) | 
|  | if args and args.strip(): | 
|  | cmd.extend(args.split()) | 
|  |  | 
|  | cmd.extend(["--ramdisk", ramdisk_img.name, | 
|  | "--output", img.name]) | 
|  |  | 
|  | p = Run(cmd, stdout=subprocess.PIPE) | 
|  | p.communicate() | 
|  | assert p.returncode == 0, "mkbootimg of %s image failed" % ( | 
|  | os.path.basename(sourcedir),) | 
|  |  | 
|  | img.seek(os.SEEK_SET, 0) | 
|  | data = img.read() | 
|  |  | 
|  | ramdisk_img.close() | 
|  | img.close() | 
|  |  | 
|  | return data | 
|  |  | 
|  |  | 
|  | def GetBootableImage(name, prebuilt_name, unpack_dir, tree_subdir, | 
|  | info_dict=None): | 
|  | """Return a File object (with name 'name') with the desired bootable | 
|  | image.  Look for it in 'unpack_dir'/BOOTABLE_IMAGES under the name | 
|  | 'prebuilt_name', otherwise construct it from the source files in | 
|  | 'unpack_dir'/'tree_subdir'.""" | 
|  |  | 
|  | prebuilt_path = os.path.join(unpack_dir, "BOOTABLE_IMAGES", prebuilt_name) | 
|  | if os.path.exists(prebuilt_path): | 
|  | print "using prebuilt %s..." % (prebuilt_name,) | 
|  | return File.FromLocalFile(name, prebuilt_path) | 
|  | else: | 
|  | print "building image from target_files %s..." % (tree_subdir,) | 
|  | fs_config = "META/" + tree_subdir.lower() + "_filesystem_config.txt" | 
|  | return File(name, BuildBootableImage(os.path.join(unpack_dir, tree_subdir), | 
|  | os.path.join(unpack_dir, fs_config), | 
|  | info_dict)) | 
|  |  | 
|  |  | 
|  | def UnzipTemp(filename, pattern=None): | 
|  | """Unzip the given archive into a temporary directory and return the name. | 
|  |  | 
|  | If filename is of the form "foo.zip+bar.zip", unzip foo.zip into a | 
|  | temp dir, then unzip bar.zip into that_dir/BOOTABLE_IMAGES. | 
|  |  | 
|  | Returns (tempdir, zipobj) where zipobj is a zipfile.ZipFile (of the | 
|  | main file), open for reading. | 
|  | """ | 
|  |  | 
|  | tmp = tempfile.mkdtemp(prefix="targetfiles-") | 
|  | OPTIONS.tempfiles.append(tmp) | 
|  |  | 
|  | def unzip_to_dir(filename, dirname): | 
|  | cmd = ["unzip", "-o", "-q", filename, "-d", dirname] | 
|  | if pattern is not None: | 
|  | cmd.append(pattern) | 
|  | p = Run(cmd, stdout=subprocess.PIPE) | 
|  | p.communicate() | 
|  | if p.returncode != 0: | 
|  | raise ExternalError("failed to unzip input target-files \"%s\"" % | 
|  | (filename,)) | 
|  |  | 
|  | m = re.match(r"^(.*[.]zip)\+(.*[.]zip)$", filename, re.IGNORECASE) | 
|  | if m: | 
|  | unzip_to_dir(m.group(1), tmp) | 
|  | unzip_to_dir(m.group(2), os.path.join(tmp, "BOOTABLE_IMAGES")) | 
|  | filename = m.group(1) | 
|  | else: | 
|  | unzip_to_dir(filename, tmp) | 
|  |  | 
|  | return tmp, zipfile.ZipFile(filename, "r") | 
|  |  | 
|  |  | 
|  | def GetKeyPasswords(keylist): | 
|  | """Given a list of keys, prompt the user to enter passwords for | 
|  | those which require them.  Return a {key: password} dict.  password | 
|  | will be None if the key has no password.""" | 
|  |  | 
|  | no_passwords = [] | 
|  | need_passwords = [] | 
|  | key_passwords = {} | 
|  | devnull = open("/dev/null", "w+b") | 
|  | for k in sorted(keylist): | 
|  | # We don't need a password for things that aren't really keys. | 
|  | if k in SPECIAL_CERT_STRINGS: | 
|  | no_passwords.append(k) | 
|  | continue | 
|  |  | 
|  | p = Run(["openssl", "pkcs8", "-in", k+OPTIONS.private_key_suffix, | 
|  | "-inform", "DER", "-nocrypt"], | 
|  | stdin=devnull.fileno(), | 
|  | stdout=devnull.fileno(), | 
|  | stderr=subprocess.STDOUT) | 
|  | p.communicate() | 
|  | if p.returncode == 0: | 
|  | # Definitely an unencrypted key. | 
|  | no_passwords.append(k) | 
|  | else: | 
|  | p = Run(["openssl", "pkcs8", "-in", k+OPTIONS.private_key_suffix, | 
|  | "-inform", "DER", "-passin", "pass:"], | 
|  | stdin=devnull.fileno(), | 
|  | stdout=devnull.fileno(), | 
|  | stderr=subprocess.PIPE) | 
|  | stdout, stderr = p.communicate() | 
|  | if p.returncode == 0: | 
|  | # Encrypted key with empty string as password. | 
|  | key_passwords[k] = '' | 
|  | elif stderr.startswith('Error decrypting key'): | 
|  | # Definitely encrypted key. | 
|  | # It would have said "Error reading key" if it didn't parse correctly. | 
|  | need_passwords.append(k) | 
|  | else: | 
|  | # Potentially, a type of key that openssl doesn't understand. | 
|  | # We'll let the routines in signapk.jar handle it. | 
|  | no_passwords.append(k) | 
|  | devnull.close() | 
|  |  | 
|  | key_passwords.update(PasswordManager().GetPasswords(need_passwords)) | 
|  | key_passwords.update(dict.fromkeys(no_passwords, None)) | 
|  | return key_passwords | 
|  |  | 
|  |  | 
|  | def SignFile(input_name, output_name, key, password, align=None, | 
|  | whole_file=False): | 
|  | """Sign the input_name zip/jar/apk, producing output_name.  Use the | 
|  | given key and password (the latter may be None if the key does not | 
|  | have a password. | 
|  |  | 
|  | If align is an integer > 1, zipalign is run to align stored files in | 
|  | the output zip on 'align'-byte boundaries. | 
|  |  | 
|  | If whole_file is true, use the "-w" option to SignApk to embed a | 
|  | signature that covers the whole file in the archive comment of the | 
|  | zip file. | 
|  | """ | 
|  |  | 
|  | if align == 0 or align == 1: | 
|  | align = None | 
|  |  | 
|  | if align: | 
|  | temp = tempfile.NamedTemporaryFile() | 
|  | sign_name = temp.name | 
|  | else: | 
|  | sign_name = output_name | 
|  |  | 
|  | cmd = [OPTIONS.java_path, "-Xmx2048m", "-jar", | 
|  | os.path.join(OPTIONS.search_path, OPTIONS.signapk_path)] | 
|  | cmd.extend(OPTIONS.extra_signapk_args) | 
|  | if whole_file: | 
|  | cmd.append("-w") | 
|  | cmd.extend([key + OPTIONS.public_key_suffix, | 
|  | key + OPTIONS.private_key_suffix, | 
|  | input_name, sign_name]) | 
|  |  | 
|  | p = Run(cmd, stdin=subprocess.PIPE, stdout=subprocess.PIPE) | 
|  | if password is not None: | 
|  | password += "\n" | 
|  | p.communicate(password) | 
|  | if p.returncode != 0: | 
|  | raise ExternalError("signapk.jar failed: return code %s" % (p.returncode,)) | 
|  |  | 
|  | if align: | 
|  | p = Run(["zipalign", "-f", str(align), sign_name, output_name]) | 
|  | p.communicate() | 
|  | if p.returncode != 0: | 
|  | raise ExternalError("zipalign failed: return code %s" % (p.returncode,)) | 
|  | temp.close() | 
|  |  | 
|  |  | 
|  | def CheckSize(data, target, info_dict): | 
|  | """Check the data string passed against the max size limit, if | 
|  | any, for the given target.  Raise exception if the data is too big. | 
|  | Print a warning if the data is nearing the maximum size.""" | 
|  |  | 
|  | if target.endswith(".img"): target = target[:-4] | 
|  | mount_point = "/" + target | 
|  |  | 
|  | fs_type = None | 
|  | limit = None | 
|  | if info_dict["fstab"]: | 
|  | if mount_point == "/userdata": mount_point = "/data" | 
|  | p = info_dict["fstab"][mount_point] | 
|  | fs_type = p.fs_type | 
|  | device = p.device | 
|  | if "/" in device: | 
|  | device = device[device.rfind("/")+1:] | 
|  | limit = info_dict.get(device + "_size", None) | 
|  | if not fs_type or not limit: return | 
|  |  | 
|  | if fs_type == "yaffs2": | 
|  | # image size should be increased by 1/64th to account for the | 
|  | # spare area (64 bytes per 2k page) | 
|  | limit = limit / 2048 * (2048+64) | 
|  | size = len(data) | 
|  | pct = float(size) * 100.0 / limit | 
|  | msg = "%s size (%d) is %.2f%% of limit (%d)" % (target, size, pct, limit) | 
|  | if pct >= 99.0: | 
|  | raise ExternalError(msg) | 
|  | elif pct >= 95.0: | 
|  | print | 
|  | print "  WARNING: ", msg | 
|  | print | 
|  | elif OPTIONS.verbose: | 
|  | print "  ", msg | 
|  |  | 
|  |  | 
|  | def ReadApkCerts(tf_zip): | 
|  | """Given a target_files ZipFile, parse the META/apkcerts.txt file | 
|  | and return a {package: cert} dict.""" | 
|  | certmap = {} | 
|  | for line in tf_zip.read("META/apkcerts.txt").split("\n"): | 
|  | line = line.strip() | 
|  | if not line: continue | 
|  | m = re.match(r'^name="(.*)"\s+certificate="(.*)"\s+' | 
|  | r'private_key="(.*)"$', line) | 
|  | if m: | 
|  | name, cert, privkey = m.groups() | 
|  | public_key_suffix_len = len(OPTIONS.public_key_suffix) | 
|  | private_key_suffix_len = len(OPTIONS.private_key_suffix) | 
|  | if cert in SPECIAL_CERT_STRINGS and not privkey: | 
|  | certmap[name] = cert | 
|  | elif (cert.endswith(OPTIONS.public_key_suffix) and | 
|  | privkey.endswith(OPTIONS.private_key_suffix) and | 
|  | cert[:-public_key_suffix_len] == privkey[:-private_key_suffix_len]): | 
|  | certmap[name] = cert[:-public_key_suffix_len] | 
|  | else: | 
|  | raise ValueError("failed to parse line from apkcerts.txt:\n" + line) | 
|  | return certmap | 
|  |  | 
|  |  | 
|  | COMMON_DOCSTRING = """ | 
|  | -p  (--path)  <dir> | 
|  | Prepend <dir>/bin to the list of places to search for binaries | 
|  | run by this script, and expect to find jars in <dir>/framework. | 
|  |  | 
|  | -s  (--device_specific) <file> | 
|  | Path to the python module containing device-specific | 
|  | releasetools code. | 
|  |  | 
|  | -x  (--extra)  <key=value> | 
|  | Add a key/value pair to the 'extras' dict, which device-specific | 
|  | extension code may look at. | 
|  |  | 
|  | -v  (--verbose) | 
|  | Show command lines being executed. | 
|  |  | 
|  | -h  (--help) | 
|  | Display this usage message and exit. | 
|  | """ | 
|  |  | 
|  | def Usage(docstring): | 
|  | print docstring.rstrip("\n") | 
|  | print COMMON_DOCSTRING | 
|  |  | 
|  |  | 
|  | def ParseOptions(argv, | 
|  | docstring, | 
|  | extra_opts="", extra_long_opts=(), | 
|  | extra_option_handler=None): | 
|  | """Parse the options in argv and return any arguments that aren't | 
|  | flags.  docstring is the calling module's docstring, to be displayed | 
|  | for errors and -h.  extra_opts and extra_long_opts are for flags | 
|  | defined by the caller, which are processed by passing them to | 
|  | extra_option_handler.""" | 
|  |  | 
|  | try: | 
|  | opts, args = getopt.getopt( | 
|  | argv, "hvp:s:x:" + extra_opts, | 
|  | ["help", "verbose", "path=", "signapk_path=", "extra_signapk_args=", | 
|  | "java_path=", "public_key_suffix=", "private_key_suffix=", | 
|  | "device_specific=", "extra="] + | 
|  | list(extra_long_opts)) | 
|  | except getopt.GetoptError, err: | 
|  | Usage(docstring) | 
|  | print "**", str(err), "**" | 
|  | sys.exit(2) | 
|  |  | 
|  | path_specified = False | 
|  |  | 
|  | for o, a in opts: | 
|  | if o in ("-h", "--help"): | 
|  | Usage(docstring) | 
|  | sys.exit() | 
|  | elif o in ("-v", "--verbose"): | 
|  | OPTIONS.verbose = True | 
|  | elif o in ("-p", "--path"): | 
|  | OPTIONS.search_path = a | 
|  | elif o in ("--signapk_path",): | 
|  | OPTIONS.signapk_path = a | 
|  | elif o in ("--extra_signapk_args",): | 
|  | OPTIONS.extra_signapk_args = shlex.split(a) | 
|  | elif o in ("--java_path",): | 
|  | OPTIONS.java_path = a | 
|  | elif o in ("--public_key_suffix",): | 
|  | OPTIONS.public_key_suffix = a | 
|  | elif o in ("--private_key_suffix",): | 
|  | OPTIONS.private_key_suffix = a | 
|  | elif o in ("-s", "--device_specific"): | 
|  | OPTIONS.device_specific = a | 
|  | elif o in ("-x", "--extra"): | 
|  | key, value = a.split("=", 1) | 
|  | OPTIONS.extras[key] = value | 
|  | else: | 
|  | if extra_option_handler is None or not extra_option_handler(o, a): | 
|  | assert False, "unknown option \"%s\"" % (o,) | 
|  |  | 
|  | os.environ["PATH"] = (os.path.join(OPTIONS.search_path, "bin") + | 
|  | os.pathsep + os.environ["PATH"]) | 
|  |  | 
|  | return args | 
|  |  | 
|  |  | 
|  | def Cleanup(): | 
|  | for i in OPTIONS.tempfiles: | 
|  | if os.path.isdir(i): | 
|  | shutil.rmtree(i) | 
|  | else: | 
|  | os.remove(i) | 
|  |  | 
|  |  | 
|  | class PasswordManager(object): | 
|  | def __init__(self): | 
|  | self.editor = os.getenv("EDITOR", None) | 
|  | self.pwfile = os.getenv("ANDROID_PW_FILE", None) | 
|  |  | 
|  | def GetPasswords(self, items): | 
|  | """Get passwords corresponding to each string in 'items', | 
|  | returning a dict.  (The dict may have keys in addition to the | 
|  | values in 'items'.) | 
|  |  | 
|  | Uses the passwords in $ANDROID_PW_FILE if available, letting the | 
|  | user edit that file to add more needed passwords.  If no editor is | 
|  | available, or $ANDROID_PW_FILE isn't define, prompts the user | 
|  | interactively in the ordinary way. | 
|  | """ | 
|  |  | 
|  | current = self.ReadFile() | 
|  |  | 
|  | first = True | 
|  | while True: | 
|  | missing = [] | 
|  | for i in items: | 
|  | if i not in current or not current[i]: | 
|  | missing.append(i) | 
|  | # Are all the passwords already in the file? | 
|  | if not missing: return current | 
|  |  | 
|  | for i in missing: | 
|  | current[i] = "" | 
|  |  | 
|  | if not first: | 
|  | print "key file %s still missing some passwords." % (self.pwfile,) | 
|  | answer = raw_input("try to edit again? [y]> ").strip() | 
|  | if answer and answer[0] not in 'yY': | 
|  | raise RuntimeError("key passwords unavailable") | 
|  | first = False | 
|  |  | 
|  | current = self.UpdateAndReadFile(current) | 
|  |  | 
|  | def PromptResult(self, current): | 
|  | """Prompt the user to enter a value (password) for each key in | 
|  | 'current' whose value is fales.  Returns a new dict with all the | 
|  | values. | 
|  | """ | 
|  | result = {} | 
|  | for k, v in sorted(current.iteritems()): | 
|  | if v: | 
|  | result[k] = v | 
|  | else: | 
|  | while True: | 
|  | result[k] = getpass.getpass("Enter password for %s key> " | 
|  | % (k,)).strip() | 
|  | if result[k]: break | 
|  | return result | 
|  |  | 
|  | def UpdateAndReadFile(self, current): | 
|  | if not self.editor or not self.pwfile: | 
|  | return self.PromptResult(current) | 
|  |  | 
|  | f = open(self.pwfile, "w") | 
|  | os.chmod(self.pwfile, 0600) | 
|  | f.write("# Enter key passwords between the [[[ ]]] brackets.\n") | 
|  | f.write("# (Additional spaces are harmless.)\n\n") | 
|  |  | 
|  | first_line = None | 
|  | sorted = [(not v, k, v) for (k, v) in current.iteritems()] | 
|  | sorted.sort() | 
|  | for i, (_, k, v) in enumerate(sorted): | 
|  | f.write("[[[  %s  ]]] %s\n" % (v, k)) | 
|  | if not v and first_line is None: | 
|  | # position cursor on first line with no password. | 
|  | first_line = i + 4 | 
|  | f.close() | 
|  |  | 
|  | p = Run([self.editor, "+%d" % (first_line,), self.pwfile]) | 
|  | _, _ = p.communicate() | 
|  |  | 
|  | return self.ReadFile() | 
|  |  | 
|  | def ReadFile(self): | 
|  | result = {} | 
|  | if self.pwfile is None: return result | 
|  | try: | 
|  | f = open(self.pwfile, "r") | 
|  | for line in f: | 
|  | line = line.strip() | 
|  | if not line or line[0] == '#': continue | 
|  | m = re.match(r"^\[\[\[\s*(.*?)\s*\]\]\]\s*(\S+)$", line) | 
|  | if not m: | 
|  | print "failed to parse password file: ", line | 
|  | else: | 
|  | result[m.group(2)] = m.group(1) | 
|  | f.close() | 
|  | except IOError, e: | 
|  | if e.errno != errno.ENOENT: | 
|  | print "error reading password file: ", str(e) | 
|  | return result | 
|  |  | 
|  |  | 
|  | def ZipWriteStr(zip, filename, data, perms=0644): | 
|  | # use a fixed timestamp so the output is repeatable. | 
|  | zinfo = zipfile.ZipInfo(filename=filename, | 
|  | date_time=(2009, 1, 1, 0, 0, 0)) | 
|  | zinfo.compress_type = zip.compression | 
|  | zinfo.external_attr = perms << 16 | 
|  | zip.writestr(zinfo, data) | 
|  |  | 
|  |  | 
|  | class DeviceSpecificParams(object): | 
|  | module = None | 
|  | def __init__(self, **kwargs): | 
|  | """Keyword arguments to the constructor become attributes of this | 
|  | object, which is passed to all functions in the device-specific | 
|  | module.""" | 
|  | for k, v in kwargs.iteritems(): | 
|  | setattr(self, k, v) | 
|  | self.extras = OPTIONS.extras | 
|  |  | 
|  | if self.module is None: | 
|  | path = OPTIONS.device_specific | 
|  | if not path: return | 
|  | try: | 
|  | if os.path.isdir(path): | 
|  | info = imp.find_module("releasetools", [path]) | 
|  | else: | 
|  | d, f = os.path.split(path) | 
|  | b, x = os.path.splitext(f) | 
|  | if x == ".py": | 
|  | f = b | 
|  | info = imp.find_module(f, [d]) | 
|  | self.module = imp.load_module("device_specific", *info) | 
|  | except ImportError: | 
|  | print "unable to load device-specific module; assuming none" | 
|  |  | 
|  | def _DoCall(self, function_name, *args, **kwargs): | 
|  | """Call the named function in the device-specific module, passing | 
|  | the given args and kwargs.  The first argument to the call will be | 
|  | the DeviceSpecific object itself.  If there is no module, or the | 
|  | module does not define the function, return the value of the | 
|  | 'default' kwarg (which itself defaults to None).""" | 
|  | if self.module is None or not hasattr(self.module, function_name): | 
|  | return kwargs.get("default", None) | 
|  | return getattr(self.module, function_name)(*((self,) + args), **kwargs) | 
|  |  | 
|  | def FullOTA_Assertions(self): | 
|  | """Called after emitting the block of assertions at the top of a | 
|  | full OTA package.  Implementations can add whatever additional | 
|  | assertions they like.""" | 
|  | return self._DoCall("FullOTA_Assertions") | 
|  |  | 
|  | def FullOTA_InstallBegin(self): | 
|  | """Called at the start of full OTA installation.""" | 
|  | return self._DoCall("FullOTA_InstallBegin") | 
|  |  | 
|  | def FullOTA_InstallEnd(self): | 
|  | """Called at the end of full OTA installation; typically this is | 
|  | used to install the image for the device's baseband processor.""" | 
|  | return self._DoCall("FullOTA_InstallEnd") | 
|  |  | 
|  | def IncrementalOTA_Assertions(self): | 
|  | """Called after emitting the block of assertions at the top of an | 
|  | incremental OTA package.  Implementations can add whatever | 
|  | additional assertions they like.""" | 
|  | return self._DoCall("IncrementalOTA_Assertions") | 
|  |  | 
|  | def IncrementalOTA_VerifyBegin(self): | 
|  | """Called at the start of the verification phase of incremental | 
|  | OTA installation; additional checks can be placed here to abort | 
|  | the script before any changes are made.""" | 
|  | return self._DoCall("IncrementalOTA_VerifyBegin") | 
|  |  | 
|  | def IncrementalOTA_VerifyEnd(self): | 
|  | """Called at the end of the verification phase of incremental OTA | 
|  | installation; additional checks can be placed here to abort the | 
|  | script before any changes are made.""" | 
|  | return self._DoCall("IncrementalOTA_VerifyEnd") | 
|  |  | 
|  | def IncrementalOTA_InstallBegin(self): | 
|  | """Called at the start of incremental OTA installation (after | 
|  | verification is complete).""" | 
|  | return self._DoCall("IncrementalOTA_InstallBegin") | 
|  |  | 
|  | def IncrementalOTA_InstallEnd(self): | 
|  | """Called at the end of incremental OTA installation; typically | 
|  | this is used to install the image for the device's baseband | 
|  | processor.""" | 
|  | return self._DoCall("IncrementalOTA_InstallEnd") | 
|  |  | 
|  | class File(object): | 
|  | def __init__(self, name, data): | 
|  | self.name = name | 
|  | self.data = data | 
|  | self.size = len(data) | 
|  | self.sha1 = sha1(data).hexdigest() | 
|  |  | 
|  | @classmethod | 
|  | def FromLocalFile(cls, name, diskname): | 
|  | f = open(diskname, "rb") | 
|  | data = f.read() | 
|  | f.close() | 
|  | return File(name, data) | 
|  |  | 
|  | def WriteToTemp(self): | 
|  | t = tempfile.NamedTemporaryFile() | 
|  | t.write(self.data) | 
|  | t.flush() | 
|  | return t | 
|  |  | 
|  | def AddToZip(self, z): | 
|  | ZipWriteStr(z, self.name, self.data) | 
|  |  | 
|  | DIFF_PROGRAM_BY_EXT = { | 
|  | ".gz" : "imgdiff", | 
|  | ".zip" : ["imgdiff", "-z"], | 
|  | ".jar" : ["imgdiff", "-z"], | 
|  | ".apk" : ["imgdiff", "-z"], | 
|  | ".img" : "imgdiff", | 
|  | } | 
|  |  | 
|  | class Difference(object): | 
|  | def __init__(self, tf, sf, diff_program=None): | 
|  | self.tf = tf | 
|  | self.sf = sf | 
|  | self.patch = None | 
|  | self.diff_program = diff_program | 
|  |  | 
|  | def ComputePatch(self): | 
|  | """Compute the patch (as a string of data) needed to turn sf into | 
|  | tf.  Returns the same tuple as GetPatch().""" | 
|  |  | 
|  | tf = self.tf | 
|  | sf = self.sf | 
|  |  | 
|  | if self.diff_program: | 
|  | diff_program = self.diff_program | 
|  | else: | 
|  | ext = os.path.splitext(tf.name)[1] | 
|  | diff_program = DIFF_PROGRAM_BY_EXT.get(ext, "bsdiff") | 
|  |  | 
|  | ttemp = tf.WriteToTemp() | 
|  | stemp = sf.WriteToTemp() | 
|  |  | 
|  | ext = os.path.splitext(tf.name)[1] | 
|  |  | 
|  | try: | 
|  | ptemp = tempfile.NamedTemporaryFile() | 
|  | if isinstance(diff_program, list): | 
|  | cmd = copy.copy(diff_program) | 
|  | else: | 
|  | cmd = [diff_program] | 
|  | cmd.append(stemp.name) | 
|  | cmd.append(ttemp.name) | 
|  | cmd.append(ptemp.name) | 
|  | p = Run(cmd, stdout=subprocess.PIPE, stderr=subprocess.PIPE) | 
|  | _, err = p.communicate() | 
|  | if err or p.returncode != 0: | 
|  | print "WARNING: failure running %s:\n%s\n" % (diff_program, err) | 
|  | return None | 
|  | diff = ptemp.read() | 
|  | finally: | 
|  | ptemp.close() | 
|  | stemp.close() | 
|  | ttemp.close() | 
|  |  | 
|  | self.patch = diff | 
|  | return self.tf, self.sf, self.patch | 
|  |  | 
|  |  | 
|  | def GetPatch(self): | 
|  | """Return a tuple (target_file, source_file, patch_data). | 
|  | patch_data may be None if ComputePatch hasn't been called, or if | 
|  | computing the patch failed.""" | 
|  | return self.tf, self.sf, self.patch | 
|  |  | 
|  |  | 
|  | def ComputeDifferences(diffs): | 
|  | """Call ComputePatch on all the Difference objects in 'diffs'.""" | 
|  | print len(diffs), "diffs to compute" | 
|  |  | 
|  | # Do the largest files first, to try and reduce the long-pole effect. | 
|  | by_size = [(i.tf.size, i) for i in diffs] | 
|  | by_size.sort(reverse=True) | 
|  | by_size = [i[1] for i in by_size] | 
|  |  | 
|  | lock = threading.Lock() | 
|  | diff_iter = iter(by_size)   # accessed under lock | 
|  |  | 
|  | def worker(): | 
|  | try: | 
|  | lock.acquire() | 
|  | for d in diff_iter: | 
|  | lock.release() | 
|  | start = time.time() | 
|  | d.ComputePatch() | 
|  | dur = time.time() - start | 
|  | lock.acquire() | 
|  |  | 
|  | tf, sf, patch = d.GetPatch() | 
|  | if sf.name == tf.name: | 
|  | name = tf.name | 
|  | else: | 
|  | name = "%s (%s)" % (tf.name, sf.name) | 
|  | if patch is None: | 
|  | print "patching failed!                                  %s" % (name,) | 
|  | else: | 
|  | print "%8.2f sec %8d / %8d bytes (%6.2f%%) %s" % ( | 
|  | dur, len(patch), tf.size, 100.0 * len(patch) / tf.size, name) | 
|  | lock.release() | 
|  | except Exception, e: | 
|  | print e | 
|  | raise | 
|  |  | 
|  | # start worker threads; wait for them all to finish. | 
|  | threads = [threading.Thread(target=worker) | 
|  | for i in range(OPTIONS.worker_threads)] | 
|  | for th in threads: | 
|  | th.start() | 
|  | while threads: | 
|  | threads.pop().join() | 
|  |  | 
|  |  | 
|  | # map recovery.fstab's fs_types to mount/format "partition types" | 
|  | PARTITION_TYPES = { "yaffs2": "MTD", "mtd": "MTD", | 
|  | "ext4": "EMMC", "emmc": "EMMC" } | 
|  |  | 
|  | def GetTypeAndDevice(mount_point, info): | 
|  | fstab = info["fstab"] | 
|  | if fstab: | 
|  | return PARTITION_TYPES[fstab[mount_point].fs_type], fstab[mount_point].device | 
|  | else: | 
|  | return None | 
|  |  | 
|  |  | 
|  | def ParseCertificate(data): | 
|  | """Parse a PEM-format certificate.""" | 
|  | cert = [] | 
|  | save = False | 
|  | for line in data.split("\n"): | 
|  | if "--END CERTIFICATE--" in line: | 
|  | break | 
|  | if save: | 
|  | cert.append(line) | 
|  | if "--BEGIN CERTIFICATE--" in line: | 
|  | save = True | 
|  | cert = "".join(cert).decode('base64') | 
|  | return cert |