| #!/usr/bin/env python |
| # |
| # Copyright (C) 2014 The Android Open Source Project |
| # |
| # Licensed under the Apache License, Version 2.0 (the "License"); |
| # you may not use this file except in compliance with the License. |
| # You may obtain a copy of the License at |
| # |
| # http://www.apache.org/licenses/LICENSE-2.0 |
| # |
| # Unless required by applicable law or agreed to in writing, software |
| # distributed under the License is distributed on an "AS IS" BASIS, |
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| # See the License for the specific language governing permissions and |
| # limitations under the License. |
| |
| """ |
| Given a target-files zipfile that does not contain images (ie, does |
| not have an IMAGES/ top-level subdirectory), produce the images and |
| add them to the zipfile. |
| |
| Usage: add_img_to_target_files [flag] target_files |
| |
| -a (--add_missing) |
| Build and add missing images to "IMAGES/". If this option is |
| not specified, this script will simply exit when "IMAGES/" |
| directory exists in the target file. |
| |
| -r (--rebuild_recovery) |
| Rebuild the recovery patch and write it to the system image. Only |
| meaningful when system image needs to be rebuilt. |
| |
| --replace_verity_private_key |
| Replace the private key used for verity signing. (same as the option |
| in sign_target_files_apks) |
| |
| --replace_verity_public_key |
| Replace the certificate (public key) used for verity verification. (same |
| as the option in sign_target_files_apks) |
| |
| --is_signing |
| Skip building & adding the images for "userdata" and "cache" if we |
| are signing the target files. |
| """ |
| |
| from __future__ import print_function |
| |
| import sys |
| |
| if sys.hexversion < 0x02070000: |
| print("Python 2.7 or newer is required.", file=sys.stderr) |
| sys.exit(1) |
| |
| import datetime |
| import errno |
| import hashlib |
| import os |
| import shlex |
| import shutil |
| import subprocess |
| import tempfile |
| import uuid |
| import zipfile |
| |
| import build_image |
| import common |
| import rangelib |
| import sparse_img |
| |
| OPTIONS = common.OPTIONS |
| |
| OPTIONS.add_missing = False |
| OPTIONS.rebuild_recovery = False |
| OPTIONS.replace_updated_files_list = [] |
| OPTIONS.replace_verity_public_key = False |
| OPTIONS.replace_verity_private_key = False |
| OPTIONS.is_signing = False |
| |
| |
| class OutputFile(object): |
| def __init__(self, output_zip, input_dir, prefix, name): |
| self._output_zip = output_zip |
| self.input_name = os.path.join(input_dir, prefix, name) |
| |
| if self._output_zip: |
| self._zip_name = os.path.join(prefix, name) |
| |
| root, suffix = os.path.splitext(name) |
| self.name = common.MakeTempFile(prefix=root + '-', suffix=suffix) |
| else: |
| self.name = self.input_name |
| |
| def Write(self): |
| if self._output_zip: |
| common.ZipWrite(self._output_zip, self.name, self._zip_name) |
| |
| |
| def GetCareMap(which, imgname): |
| """Generate care_map of system (or vendor) partition""" |
| |
| assert which in ("system", "vendor") |
| |
| simg = sparse_img.SparseImage(imgname) |
| care_map_list = [which] |
| |
| care_map_ranges = simg.care_map |
| key = which + "_adjusted_partition_size" |
| adjusted_blocks = OPTIONS.info_dict.get(key) |
| if adjusted_blocks: |
| assert adjusted_blocks > 0, "blocks should be positive for " + which |
| care_map_ranges = care_map_ranges.intersect(rangelib.RangeSet( |
| "0-%d" % (adjusted_blocks,))) |
| |
| care_map_list.append(care_map_ranges.to_string_raw()) |
| return care_map_list |
| |
| |
| def AddSystem(output_zip, prefix="IMAGES/", recovery_img=None, boot_img=None): |
| """Turn the contents of SYSTEM into a system image and store it in |
| output_zip. Returns the name of the system image file.""" |
| |
| img = OutputFile(output_zip, OPTIONS.input_tmp, prefix, "system.img") |
| if os.path.exists(img.input_name): |
| print("system.img already exists in %s, no need to rebuild..." % (prefix,)) |
| return img.input_name |
| |
| def output_sink(fn, data): |
| ofile = open(os.path.join(OPTIONS.input_tmp, "SYSTEM", fn), "w") |
| ofile.write(data) |
| ofile.close() |
| |
| arc_name = "SYSTEM/" + fn |
| if arc_name in output_zip.namelist(): |
| OPTIONS.replace_updated_files_list.append(arc_name) |
| else: |
| common.ZipWrite(output_zip, ofile.name, arc_name) |
| |
| if OPTIONS.rebuild_recovery: |
| print("Building new recovery patch") |
| common.MakeRecoveryPatch(OPTIONS.input_tmp, output_sink, recovery_img, |
| boot_img, info_dict=OPTIONS.info_dict) |
| |
| block_list = OutputFile(output_zip, OPTIONS.input_tmp, prefix, "system.map") |
| CreateImage(OPTIONS.input_tmp, OPTIONS.info_dict, "system", img, |
| block_list=block_list) |
| |
| return img.name |
| |
| |
| def AddSystemOther(output_zip, prefix="IMAGES/"): |
| """Turn the contents of SYSTEM_OTHER into a system_other image |
| and store it in output_zip.""" |
| |
| img = OutputFile(output_zip, OPTIONS.input_tmp, prefix, "system_other.img") |
| if os.path.exists(img.input_name): |
| print("system_other.img already exists in %s, no need to rebuild..." % ( |
| prefix,)) |
| return |
| |
| CreateImage(OPTIONS.input_tmp, OPTIONS.info_dict, "system_other", img) |
| |
| |
| def AddVendor(output_zip, prefix="IMAGES/"): |
| """Turn the contents of VENDOR into a vendor image and store in it |
| output_zip.""" |
| |
| img = OutputFile(output_zip, OPTIONS.input_tmp, prefix, "vendor.img") |
| if os.path.exists(img.input_name): |
| print("vendor.img already exists in %s, no need to rebuild..." % (prefix,)) |
| return img.input_name |
| |
| block_list = OutputFile(output_zip, OPTIONS.input_tmp, prefix, "vendor.map") |
| CreateImage(OPTIONS.input_tmp, OPTIONS.info_dict, "vendor", img, |
| block_list=block_list) |
| return img.name |
| |
| |
| def AddDtbo(output_zip, prefix="IMAGES/"): |
| """Adds the DTBO image. |
| |
| Uses the image under prefix if it already exists. Otherwise looks for the |
| image under PREBUILT_IMAGES/, signs it as needed, and returns the image name. |
| """ |
| |
| img = OutputFile(output_zip, OPTIONS.input_tmp, prefix, "dtbo.img") |
| if os.path.exists(img.input_name): |
| print("dtbo.img already exists in %s, no need to rebuild..." % (prefix,)) |
| return img.input_name |
| |
| dtbo_prebuilt_path = os.path.join( |
| OPTIONS.input_tmp, "PREBUILT_IMAGES", "dtbo.img") |
| assert os.path.exists(dtbo_prebuilt_path) |
| shutil.copy(dtbo_prebuilt_path, img.name) |
| |
| # AVB-sign the image as needed. |
| if OPTIONS.info_dict.get("avb_enable") == "true": |
| avbtool = os.getenv('AVBTOOL') or OPTIONS.info_dict["avb_avbtool"] |
| part_size = OPTIONS.info_dict["dtbo_size"] |
| # The AVB hash footer will be replaced if already present. |
| cmd = [avbtool, "add_hash_footer", "--image", img.name, |
| "--partition_size", str(part_size), "--partition_name", "dtbo"] |
| common.AppendAVBSigningArgs(cmd, "dtbo") |
| args = OPTIONS.info_dict.get("avb_dtbo_add_hash_footer_args") |
| if args and args.strip(): |
| cmd.extend(shlex.split(args)) |
| p = common.Run(cmd, stdout=subprocess.PIPE) |
| p.communicate() |
| assert p.returncode == 0, \ |
| "avbtool add_hash_footer of %s failed" % (img.name,) |
| |
| img.Write() |
| return img.name |
| |
| |
| def CreateImage(input_dir, info_dict, what, output_file, block_list=None): |
| print("creating " + what + ".img...") |
| |
| # The name of the directory it is making an image out of matters to |
| # mkyaffs2image. It wants "system" but we have a directory named |
| # "SYSTEM", so create a symlink. |
| temp_dir = tempfile.mkdtemp() |
| OPTIONS.tempfiles.append(temp_dir) |
| try: |
| os.symlink(os.path.join(input_dir, what.upper()), |
| os.path.join(temp_dir, what)) |
| except OSError as e: |
| # bogus error on my mac version? |
| # File "./build/tools/releasetools/img_from_target_files" |
| # os.path.join(OPTIONS.input_tmp, "system")) |
| # OSError: [Errno 17] File exists |
| if e.errno == errno.EEXIST: |
| pass |
| |
| image_props = build_image.ImagePropFromGlobalDict(info_dict, what) |
| fstab = info_dict["fstab"] |
| mount_point = "/" + what |
| if fstab and mount_point in fstab: |
| image_props["fs_type"] = fstab[mount_point].fs_type |
| |
| # Use a fixed timestamp (01/01/2009) when packaging the image. |
| # Bug: 24377993 |
| epoch = datetime.datetime.fromtimestamp(0) |
| timestamp = (datetime.datetime(2009, 1, 1) - epoch).total_seconds() |
| image_props["timestamp"] = int(timestamp) |
| |
| if what == "system": |
| fs_config_prefix = "" |
| else: |
| fs_config_prefix = what + "_" |
| |
| fs_config = os.path.join( |
| input_dir, "META/" + fs_config_prefix + "filesystem_config.txt") |
| if not os.path.exists(fs_config): |
| fs_config = None |
| |
| # Override values loaded from info_dict. |
| if fs_config: |
| image_props["fs_config"] = fs_config |
| if block_list: |
| image_props["block_list"] = block_list.name |
| |
| # Use repeatable ext4 FS UUID and hash_seed UUID (based on partition name and |
| # build fingerprint). |
| uuid_seed = what + "-" |
| if "build.prop" in info_dict: |
| build_prop = info_dict["build.prop"] |
| if "ro.build.fingerprint" in build_prop: |
| uuid_seed += build_prop["ro.build.fingerprint"] |
| elif "ro.build.thumbprint" in build_prop: |
| uuid_seed += build_prop["ro.build.thumbprint"] |
| image_props["uuid"] = str(uuid.uuid5(uuid.NAMESPACE_URL, uuid_seed)) |
| hash_seed = "hash_seed-" + uuid_seed |
| image_props["hash_seed"] = str(uuid.uuid5(uuid.NAMESPACE_URL, hash_seed)) |
| |
| succ = build_image.BuildImage(os.path.join(temp_dir, what), |
| image_props, output_file.name) |
| assert succ, "build " + what + ".img image failed" |
| |
| output_file.Write() |
| if block_list: |
| block_list.Write() |
| |
| # Set the 'adjusted_partition_size' that excludes the verity blocks of the |
| # given image. When avb is enabled, this size is the max image size returned |
| # by the avb tool. |
| is_verity_partition = "verity_block_device" in image_props |
| verity_supported = (image_props.get("verity") == "true" or |
| image_props.get("avb_enable") == "true") |
| is_avb_enable = image_props.get("avb_hashtree_enable") == "true" |
| if verity_supported and (is_verity_partition or is_avb_enable): |
| adjusted_blocks_value = image_props.get("partition_size") |
| if adjusted_blocks_value: |
| adjusted_blocks_key = what + "_adjusted_partition_size" |
| info_dict[adjusted_blocks_key] = int(adjusted_blocks_value)/4096 - 1 |
| |
| |
| def AddUserdata(output_zip, prefix="IMAGES/"): |
| """Create a userdata image and store it in output_zip. |
| |
| In most case we just create and store an empty userdata.img; |
| But the invoker can also request to create userdata.img with real |
| data from the target files, by setting "userdata_img_with_data=true" |
| in OPTIONS.info_dict. |
| """ |
| |
| img = OutputFile(output_zip, OPTIONS.input_tmp, prefix, "userdata.img") |
| if os.path.exists(img.input_name): |
| print("userdata.img already exists in %s, no need to rebuild..." % ( |
| prefix,)) |
| return |
| |
| # Skip userdata.img if no size. |
| image_props = build_image.ImagePropFromGlobalDict(OPTIONS.info_dict, "data") |
| if not image_props.get("partition_size"): |
| return |
| |
| print("creating userdata.img...") |
| |
| # Use a fixed timestamp (01/01/2009) when packaging the image. |
| # Bug: 24377993 |
| epoch = datetime.datetime.fromtimestamp(0) |
| timestamp = (datetime.datetime(2009, 1, 1) - epoch).total_seconds() |
| image_props["timestamp"] = int(timestamp) |
| |
| # The name of the directory it is making an image out of matters to |
| # mkyaffs2image. So we create a temp dir, and within it we create an |
| # empty dir named "data", or a symlink to the DATA dir, |
| # and build the image from that. |
| temp_dir = tempfile.mkdtemp() |
| OPTIONS.tempfiles.append(temp_dir) |
| user_dir = os.path.join(temp_dir, "data") |
| empty = (OPTIONS.info_dict.get("userdata_img_with_data") != "true") |
| if empty: |
| # Create an empty dir. |
| os.mkdir(user_dir) |
| else: |
| # Symlink to the DATA dir. |
| os.symlink(os.path.join(OPTIONS.input_tmp, "DATA"), |
| user_dir) |
| |
| fstab = OPTIONS.info_dict["fstab"] |
| if fstab: |
| image_props["fs_type"] = fstab["/data"].fs_type |
| succ = build_image.BuildImage(user_dir, image_props, img.name) |
| assert succ, "build userdata.img image failed" |
| |
| common.CheckSize(img.name, "userdata.img", OPTIONS.info_dict) |
| img.Write() |
| |
| |
| def AppendVBMetaArgsForPartition(cmd, partition, img_path, public_key_dir): |
| if not img_path: |
| return |
| |
| # Check if chain partition is used. |
| key_path = OPTIONS.info_dict.get("avb_" + partition + "_key_path") |
| if key_path: |
| # extract public key in AVB format to be included in vbmeta.img |
| avbtool = os.getenv('AVBTOOL') or OPTIONS.info_dict["avb_avbtool"] |
| public_key_path = os.path.join(public_key_dir, "%s.avbpubkey" % partition) |
| p = common.Run([avbtool, "extract_public_key", "--key", key_path, |
| "--output", public_key_path], |
| stdout=subprocess.PIPE, stderr=subprocess.PIPE) |
| p.communicate() |
| assert p.returncode == 0, \ |
| "avbtool extract_public_key fail for partition: %r" % partition |
| |
| rollback_index_location = OPTIONS.info_dict[ |
| "avb_" + partition + "_rollback_index_location"] |
| cmd.extend(["--chain_partition", "%s:%s:%s" % ( |
| partition, rollback_index_location, public_key_path)]) |
| else: |
| cmd.extend(["--include_descriptors_from_image", img_path]) |
| |
| |
| def AddVBMeta(output_zip, boot_img_path, system_img_path, vendor_img_path, |
| dtbo_img_path, prefix="IMAGES/"): |
| """Create a VBMeta image and store it in output_zip.""" |
| img = OutputFile(output_zip, OPTIONS.input_tmp, prefix, "vbmeta.img") |
| avbtool = os.getenv('AVBTOOL') or OPTIONS.info_dict["avb_avbtool"] |
| cmd = [avbtool, "make_vbmeta_image", "--output", img.name] |
| common.AppendAVBSigningArgs(cmd, "vbmeta") |
| |
| public_key_dir = tempfile.mkdtemp(prefix="avbpubkey-") |
| OPTIONS.tempfiles.append(public_key_dir) |
| |
| AppendVBMetaArgsForPartition(cmd, "boot", boot_img_path, public_key_dir) |
| AppendVBMetaArgsForPartition(cmd, "system", system_img_path, public_key_dir) |
| AppendVBMetaArgsForPartition(cmd, "vendor", vendor_img_path, public_key_dir) |
| AppendVBMetaArgsForPartition(cmd, "dtbo", dtbo_img_path, public_key_dir) |
| |
| args = OPTIONS.info_dict.get("avb_vbmeta_args") |
| if args and args.strip(): |
| split_args = shlex.split(args) |
| for index, arg in enumerate(split_args[:-1]): |
| # Sanity check that the image file exists. Some images might be defined |
| # as a path relative to source tree, which may not be available at the |
| # same location when running this script (we have the input target_files |
| # zip only). For such cases, we additionally scan other locations (e.g. |
| # IMAGES/, RADIO/, etc) before bailing out. |
| if arg == '--include_descriptors_from_image': |
| image_path = split_args[index + 1] |
| if os.path.exists(image_path): |
| continue |
| found = False |
| for dir in ['IMAGES', 'RADIO', 'VENDOR_IMAGES', 'PREBUILT_IMAGES']: |
| alt_path = os.path.join( |
| OPTIONS.input_tmp, dir, os.path.basename(image_path)) |
| if os.path.exists(alt_path): |
| split_args[index + 1] = alt_path |
| found = True |
| break |
| assert found, 'failed to find %s' % (image_path,) |
| cmd.extend(split_args) |
| |
| p = common.Run(cmd, stdout=subprocess.PIPE, stderr=subprocess.PIPE) |
| p.communicate() |
| assert p.returncode == 0, "avbtool make_vbmeta_image failed" |
| img.Write() |
| |
| |
| def AddPartitionTable(output_zip, prefix="IMAGES/"): |
| """Create a partition table image and store it in output_zip.""" |
| |
| img = OutputFile(output_zip, OPTIONS.input_tmp, prefix, "partition-table.img") |
| bpt = OutputFile(output_zip, OPTIONS.input_tmp, prefix, "partition-table.bpt") |
| |
| # use BPTTOOL from environ, or "bpttool" if empty or not set. |
| bpttool = os.getenv("BPTTOOL") or "bpttool" |
| cmd = [bpttool, "make_table", "--output_json", bpt.name, |
| "--output_gpt", img.name] |
| input_files_str = OPTIONS.info_dict["board_bpt_input_files"] |
| input_files = input_files_str.split(" ") |
| for i in input_files: |
| cmd.extend(["--input", i]) |
| disk_size = OPTIONS.info_dict.get("board_bpt_disk_size") |
| if disk_size: |
| cmd.extend(["--disk_size", disk_size]) |
| args = OPTIONS.info_dict.get("board_bpt_make_table_args") |
| if args: |
| cmd.extend(shlex.split(args)) |
| |
| p = common.Run(cmd, stdout=subprocess.PIPE) |
| p.communicate() |
| assert p.returncode == 0, "bpttool make_table failed" |
| |
| img.Write() |
| bpt.Write() |
| |
| |
| def AddCache(output_zip, prefix="IMAGES/"): |
| """Create an empty cache image and store it in output_zip.""" |
| |
| img = OutputFile(output_zip, OPTIONS.input_tmp, prefix, "cache.img") |
| if os.path.exists(img.input_name): |
| print("cache.img already exists in %s, no need to rebuild..." % (prefix,)) |
| return |
| |
| image_props = build_image.ImagePropFromGlobalDict(OPTIONS.info_dict, "cache") |
| # The build system has to explicitly request for cache.img. |
| if "fs_type" not in image_props: |
| return |
| |
| print("creating cache.img...") |
| |
| # Use a fixed timestamp (01/01/2009) when packaging the image. |
| # Bug: 24377993 |
| epoch = datetime.datetime.fromtimestamp(0) |
| timestamp = (datetime.datetime(2009, 1, 1) - epoch).total_seconds() |
| image_props["timestamp"] = int(timestamp) |
| |
| # The name of the directory it is making an image out of matters to |
| # mkyaffs2image. So we create a temp dir, and within it we create an |
| # empty dir named "cache", and build the image from that. |
| temp_dir = tempfile.mkdtemp() |
| OPTIONS.tempfiles.append(temp_dir) |
| user_dir = os.path.join(temp_dir, "cache") |
| os.mkdir(user_dir) |
| |
| fstab = OPTIONS.info_dict["fstab"] |
| if fstab: |
| image_props["fs_type"] = fstab["/cache"].fs_type |
| succ = build_image.BuildImage(user_dir, image_props, img.name) |
| assert succ, "build cache.img image failed" |
| |
| common.CheckSize(img.name, "cache.img", OPTIONS.info_dict) |
| img.Write() |
| |
| |
| def ReplaceUpdatedFiles(zip_filename, files_list): |
| """Update all the zip entries listed in the files_list. |
| |
| For now the list includes META/care_map.txt, and the related files under |
| SYSTEM/ after rebuilding recovery. |
| """ |
| |
| cmd = ["zip", "-d", zip_filename] + files_list |
| p = common.Run(cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) |
| p.communicate() |
| |
| output_zip = zipfile.ZipFile(zip_filename, "a", |
| compression=zipfile.ZIP_DEFLATED, |
| allowZip64=True) |
| for item in files_list: |
| file_path = os.path.join(OPTIONS.input_tmp, item) |
| assert os.path.exists(file_path) |
| common.ZipWrite(output_zip, file_path, arcname=item) |
| common.ZipClose(output_zip) |
| |
| |
| def AddImagesToTargetFiles(filename): |
| if os.path.isdir(filename): |
| OPTIONS.input_tmp = os.path.abspath(filename) |
| input_zip = None |
| else: |
| OPTIONS.input_tmp, input_zip = common.UnzipTemp(filename) |
| |
| if not OPTIONS.add_missing: |
| if os.path.isdir(os.path.join(OPTIONS.input_tmp, "IMAGES")): |
| print("target_files appears to already contain images.") |
| sys.exit(1) |
| |
| # vendor.img is unlike system.img or system_other.img. Because it could be |
| # built from source, or dropped into target_files.zip as a prebuilt blob. We |
| # consider either of them as vendor.img being available, which could be used |
| # when generating vbmeta.img for AVB. |
| has_vendor = (os.path.isdir(os.path.join(OPTIONS.input_tmp, "VENDOR")) or |
| os.path.exists(os.path.join(OPTIONS.input_tmp, "IMAGES", |
| "vendor.img"))) |
| has_system_other = os.path.isdir(os.path.join(OPTIONS.input_tmp, |
| "SYSTEM_OTHER")) |
| |
| if input_zip: |
| OPTIONS.info_dict = common.LoadInfoDict(input_zip, OPTIONS.input_tmp) |
| |
| common.ZipClose(input_zip) |
| output_zip = zipfile.ZipFile(filename, "a", |
| compression=zipfile.ZIP_DEFLATED, |
| allowZip64=True) |
| else: |
| OPTIONS.info_dict = common.LoadInfoDict(filename, filename) |
| output_zip = None |
| images_dir = os.path.join(OPTIONS.input_tmp, "IMAGES") |
| if not os.path.isdir(images_dir): |
| os.makedirs(images_dir) |
| images_dir = None |
| |
| has_recovery = (OPTIONS.info_dict.get("no_recovery") != "true") |
| |
| if OPTIONS.info_dict.get("avb_enable") == "true": |
| fp = None |
| if "build.prop" in OPTIONS.info_dict: |
| build_prop = OPTIONS.info_dict["build.prop"] |
| if "ro.build.fingerprint" in build_prop: |
| fp = build_prop["ro.build.fingerprint"] |
| elif "ro.build.thumbprint" in build_prop: |
| fp = build_prop["ro.build.thumbprint"] |
| if fp: |
| OPTIONS.info_dict["avb_salt"] = hashlib.sha256(fp).hexdigest() |
| |
| def banner(s): |
| print("\n\n++++ " + s + " ++++\n\n") |
| |
| prebuilt_path = os.path.join(OPTIONS.input_tmp, "IMAGES", "boot.img") |
| boot_image = None |
| if os.path.exists(prebuilt_path): |
| banner("boot") |
| print("boot.img already exists in IMAGES/, no need to rebuild...") |
| if OPTIONS.rebuild_recovery: |
| boot_image = common.GetBootableImage( |
| "IMAGES/boot.img", "boot.img", OPTIONS.input_tmp, "BOOT") |
| else: |
| banner("boot") |
| boot_image = common.GetBootableImage( |
| "IMAGES/boot.img", "boot.img", OPTIONS.input_tmp, "BOOT") |
| if boot_image: |
| if output_zip: |
| boot_image.AddToZip(output_zip) |
| else: |
| boot_image.WriteToDir(OPTIONS.input_tmp) |
| |
| recovery_image = None |
| if has_recovery: |
| banner("recovery") |
| prebuilt_path = os.path.join(OPTIONS.input_tmp, "IMAGES", "recovery.img") |
| if os.path.exists(prebuilt_path): |
| print("recovery.img already exists in IMAGES/, no need to rebuild...") |
| if OPTIONS.rebuild_recovery: |
| recovery_image = common.GetBootableImage( |
| "IMAGES/recovery.img", "recovery.img", OPTIONS.input_tmp, |
| "RECOVERY") |
| else: |
| recovery_image = common.GetBootableImage( |
| "IMAGES/recovery.img", "recovery.img", OPTIONS.input_tmp, "RECOVERY") |
| if recovery_image: |
| if output_zip: |
| recovery_image.AddToZip(output_zip) |
| else: |
| recovery_image.WriteToDir(OPTIONS.input_tmp) |
| |
| banner("recovery (two-step image)") |
| # The special recovery.img for two-step package use. |
| recovery_two_step_image = common.GetBootableImage( |
| "IMAGES/recovery-two-step.img", "recovery-two-step.img", |
| OPTIONS.input_tmp, "RECOVERY", two_step_image=True) |
| if recovery_two_step_image: |
| if output_zip: |
| recovery_two_step_image.AddToZip(output_zip) |
| else: |
| recovery_two_step_image.WriteToDir(OPTIONS.input_tmp) |
| |
| banner("system") |
| system_img_path = AddSystem( |
| output_zip, recovery_img=recovery_image, boot_img=boot_image) |
| vendor_img_path = None |
| if has_vendor: |
| banner("vendor") |
| vendor_img_path = AddVendor(output_zip) |
| if has_system_other: |
| banner("system_other") |
| AddSystemOther(output_zip) |
| if not OPTIONS.is_signing: |
| banner("userdata") |
| AddUserdata(output_zip) |
| banner("cache") |
| AddCache(output_zip) |
| |
| if OPTIONS.info_dict.get("board_bpt_enable") == "true": |
| banner("partition-table") |
| AddPartitionTable(output_zip) |
| |
| dtbo_img_path = None |
| if OPTIONS.info_dict.get("has_dtbo") == "true": |
| banner("dtbo") |
| dtbo_img_path = AddDtbo(output_zip) |
| |
| if OPTIONS.info_dict.get("avb_enable") == "true": |
| banner("vbmeta") |
| boot_contents = boot_image.WriteToTemp() |
| AddVBMeta(output_zip, boot_contents.name, system_img_path, |
| vendor_img_path, dtbo_img_path) |
| |
| # For devices using A/B update, copy over images from RADIO/ and/or |
| # VENDOR_IMAGES/ to IMAGES/ and make sure we have all the needed |
| # images ready under IMAGES/. All images should have '.img' as extension. |
| banner("radio") |
| ab_partitions = os.path.join(OPTIONS.input_tmp, "META", "ab_partitions.txt") |
| if os.path.exists(ab_partitions): |
| with open(ab_partitions, 'r') as f: |
| lines = f.readlines() |
| # For devices using A/B update, generate care_map for system and vendor |
| # partitions (if present), then write this file to target_files package. |
| care_map_list = [] |
| for line in lines: |
| if line.strip() == "system" and ( |
| "system_verity_block_device" in OPTIONS.info_dict or |
| OPTIONS.info_dict.get("avb_system_hashtree_enable") == "true"): |
| assert os.path.exists(system_img_path) |
| care_map_list += GetCareMap("system", system_img_path) |
| if line.strip() == "vendor" and ( |
| "vendor_verity_block_device" in OPTIONS.info_dict or |
| OPTIONS.info_dict.get("avb_vendor_hashtree_enable") == "true"): |
| assert os.path.exists(vendor_img_path) |
| care_map_list += GetCareMap("vendor", vendor_img_path) |
| |
| img_name = line.strip() + ".img" |
| prebuilt_path = os.path.join(OPTIONS.input_tmp, "IMAGES", img_name) |
| if os.path.exists(prebuilt_path): |
| print("%s already exists, no need to overwrite..." % (img_name,)) |
| continue |
| |
| img_radio_path = os.path.join(OPTIONS.input_tmp, "RADIO", img_name) |
| img_vendor_dir = os.path.join( |
| OPTIONS.input_tmp, "VENDOR_IMAGES") |
| if os.path.exists(img_radio_path): |
| if output_zip: |
| common.ZipWrite(output_zip, img_radio_path, |
| os.path.join("IMAGES", img_name)) |
| else: |
| shutil.copy(img_radio_path, prebuilt_path) |
| else: |
| for root, _, files in os.walk(img_vendor_dir): |
| if img_name in files: |
| if output_zip: |
| common.ZipWrite(output_zip, os.path.join(root, img_name), |
| os.path.join("IMAGES", img_name)) |
| else: |
| shutil.copy(os.path.join(root, img_name), prebuilt_path) |
| break |
| |
| if output_zip: |
| # Zip spec says: All slashes MUST be forward slashes. |
| img_path = 'IMAGES/' + img_name |
| assert img_path in output_zip.namelist(), "cannot find " + img_name |
| else: |
| img_path = os.path.join(OPTIONS.input_tmp, "IMAGES", img_name) |
| assert os.path.exists(img_path), "cannot find " + img_name |
| |
| if care_map_list: |
| care_map_path = "META/care_map.txt" |
| if output_zip and care_map_path not in output_zip.namelist(): |
| common.ZipWriteStr(output_zip, care_map_path, '\n'.join(care_map_list)) |
| else: |
| with open(os.path.join(OPTIONS.input_tmp, care_map_path), 'w') as fp: |
| fp.write('\n'.join(care_map_list)) |
| if output_zip: |
| OPTIONS.replace_updated_files_list.append(care_map_path) |
| |
| # Radio images that need to be packed into IMAGES/, and product-img.zip. |
| pack_radioimages = os.path.join( |
| OPTIONS.input_tmp, "META", "pack_radioimages.txt") |
| if os.path.exists(pack_radioimages): |
| with open(pack_radioimages, 'r') as f: |
| lines = f.readlines() |
| for line in lines: |
| img_name = line.strip() |
| _, ext = os.path.splitext(img_name) |
| if not ext: |
| img_name += ".img" |
| prebuilt_path = os.path.join(OPTIONS.input_tmp, "IMAGES", img_name) |
| if os.path.exists(prebuilt_path): |
| print("%s already exists, no need to overwrite..." % (img_name,)) |
| continue |
| |
| img_radio_path = os.path.join(OPTIONS.input_tmp, "RADIO", img_name) |
| assert os.path.exists(img_radio_path), \ |
| "Failed to find %s at %s" % (img_name, img_radio_path) |
| if output_zip: |
| common.ZipWrite(output_zip, img_radio_path, |
| os.path.join("IMAGES", img_name)) |
| else: |
| shutil.copy(img_radio_path, prebuilt_path) |
| |
| if output_zip: |
| common.ZipClose(output_zip) |
| if OPTIONS.replace_updated_files_list: |
| ReplaceUpdatedFiles(output_zip.filename, |
| OPTIONS.replace_updated_files_list) |
| |
| |
| def main(argv): |
| def option_handler(o, a): |
| if o in ("-a", "--add_missing"): |
| OPTIONS.add_missing = True |
| elif o in ("-r", "--rebuild_recovery",): |
| OPTIONS.rebuild_recovery = True |
| elif o == "--replace_verity_private_key": |
| OPTIONS.replace_verity_private_key = (True, a) |
| elif o == "--replace_verity_public_key": |
| OPTIONS.replace_verity_public_key = (True, a) |
| elif o == "--is_signing": |
| OPTIONS.is_signing = True |
| else: |
| return False |
| return True |
| |
| args = common.ParseOptions( |
| argv, __doc__, extra_opts="ar", |
| extra_long_opts=["add_missing", "rebuild_recovery", |
| "replace_verity_public_key=", |
| "replace_verity_private_key=", |
| "is_signing"], |
| extra_option_handler=option_handler) |
| |
| |
| if len(args) != 1: |
| common.Usage(__doc__) |
| sys.exit(1) |
| |
| AddImagesToTargetFiles(args[0]) |
| print("done.") |
| |
| if __name__ == '__main__': |
| try: |
| common.CloseInheritedPipes() |
| main(sys.argv[1:]) |
| except common.ExternalError as e: |
| print("\n ERROR: %s\n" % (e,)) |
| sys.exit(1) |
| finally: |
| common.Cleanup() |