| // |
| // Copyright (C) 2010 The Android Open Source Project |
| // |
| // Licensed under the Apache License, Version 2.0 (the "License"); |
| // you may not use this file except in compliance with the License. |
| // You may obtain a copy of the License at |
| // |
| // http://www.apache.org/licenses/LICENSE-2.0 |
| // |
| // Unless required by applicable law or agreed to in writing, software |
| // distributed under the License is distributed on an "AS IS" BASIS, |
| // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| // See the License for the specific language governing permissions and |
| // limitations under the License. |
| // |
| |
| #include <map> |
| #include <string> |
| #include <vector> |
| |
| #include <base/bind.h> |
| #include <base/files/file_path.h> |
| #include <base/files/file_util.h> |
| #include <base/logging.h> |
| #include <base/strings/string_number_conversions.h> |
| #include <base/strings/string_split.h> |
| #include <base/strings/string_util.h> |
| #include <brillo/flag_helper.h> |
| #include <brillo/key_value_store.h> |
| #include <brillo/message_loops/base_message_loop.h> |
| #include <xz.h> |
| |
| #include "update_engine/common/download_action.h" |
| #include "update_engine/common/fake_boot_control.h" |
| #include "update_engine/common/fake_hardware.h" |
| #include "update_engine/common/file_fetcher.h" |
| #include "update_engine/common/prefs.h" |
| #include "update_engine/common/terminator.h" |
| #include "update_engine/common/utils.h" |
| #include "update_engine/payload_consumer/filesystem_verifier_action.h" |
| #include "update_engine/payload_consumer/payload_constants.h" |
| #include "update_engine/payload_generator/delta_diff_generator.h" |
| #include "update_engine/payload_generator/payload_generation_config.h" |
| #include "update_engine/payload_generator/payload_properties.h" |
| #include "update_engine/payload_generator/payload_signer.h" |
| #include "update_engine/payload_generator/xz.h" |
| #include "update_engine/update_metadata.pb.h" |
| |
| // This file contains a simple program that takes an old path, a new path, |
| // and an output file as arguments and the path to an output file and |
| // generates a delta that can be sent to Chrome OS clients. |
| |
| using std::map; |
| using std::string; |
| using std::vector; |
| |
| namespace chromeos_update_engine { |
| |
| namespace { |
| |
| constexpr char kPayloadPropertiesFormatKeyValue[] = "key-value"; |
| constexpr char kPayloadPropertiesFormatJson[] = "json"; |
| |
| void ParseSignatureSizes(const string& signature_sizes_flag, |
| vector<size_t>* signature_sizes) { |
| signature_sizes->clear(); |
| vector<string> split_strings = base::SplitString( |
| signature_sizes_flag, ":", base::TRIM_WHITESPACE, base::SPLIT_WANT_ALL); |
| for (const string& str : split_strings) { |
| size_t size = 0; |
| bool parsing_successful = base::StringToSizeT(str, &size); |
| LOG_IF(FATAL, !parsing_successful) << "Invalid signature size: " << str; |
| |
| signature_sizes->push_back(size); |
| } |
| } |
| |
| void CalculateHashForSigning(const vector<size_t>& sizes, |
| const string& out_hash_file, |
| const string& out_metadata_hash_file, |
| const string& in_file) { |
| LOG(INFO) << "Calculating hash for signing."; |
| LOG_IF(FATAL, in_file.empty()) |
| << "Must pass --in_file to calculate hash for signing."; |
| LOG_IF(FATAL, out_hash_file.empty()) |
| << "Must pass --out_hash_file to calculate hash for signing."; |
| |
| brillo::Blob payload_hash, metadata_hash; |
| CHECK(PayloadSigner::HashPayloadForSigning( |
| in_file, sizes, &payload_hash, &metadata_hash)); |
| CHECK(utils::WriteFile( |
| out_hash_file.c_str(), payload_hash.data(), payload_hash.size())); |
| if (!out_metadata_hash_file.empty()) |
| CHECK(utils::WriteFile(out_metadata_hash_file.c_str(), |
| metadata_hash.data(), |
| metadata_hash.size())); |
| |
| LOG(INFO) << "Done calculating hash for signing."; |
| } |
| |
| void SignatureFileFlagToBlobs(const string& signature_file_flag, |
| vector<brillo::Blob>* signatures) { |
| vector<string> signature_files = base::SplitString( |
| signature_file_flag, ":", base::TRIM_WHITESPACE, base::SPLIT_WANT_ALL); |
| for (const string& signature_file : signature_files) { |
| brillo::Blob signature; |
| CHECK(utils::ReadFile(signature_file, &signature)); |
| signatures->push_back(signature); |
| } |
| } |
| |
| void SignPayload(const string& in_file, |
| const string& out_file, |
| const vector<size_t>& signature_sizes, |
| const string& payload_signature_file, |
| const string& metadata_signature_file, |
| const string& out_metadata_size_file) { |
| LOG(INFO) << "Signing payload."; |
| LOG_IF(FATAL, in_file.empty()) << "Must pass --in_file to sign payload."; |
| LOG_IF(FATAL, out_file.empty()) << "Must pass --out_file to sign payload."; |
| LOG_IF(FATAL, payload_signature_file.empty()) |
| << "Must pass --payload_signature_file to sign payload."; |
| vector<brillo::Blob> payload_signatures, metadata_signatures; |
| SignatureFileFlagToBlobs(payload_signature_file, &payload_signatures); |
| SignatureFileFlagToBlobs(metadata_signature_file, &metadata_signatures); |
| uint64_t final_metadata_size; |
| CHECK(PayloadSigner::AddSignatureToPayload(in_file, |
| signature_sizes, |
| payload_signatures, |
| metadata_signatures, |
| out_file, |
| &final_metadata_size)); |
| LOG(INFO) << "Done signing payload. Final metadata size = " |
| << final_metadata_size; |
| if (!out_metadata_size_file.empty()) { |
| string metadata_size_string = std::to_string(final_metadata_size); |
| CHECK(utils::WriteFile(out_metadata_size_file.c_str(), |
| metadata_size_string.data(), |
| metadata_size_string.size())); |
| } |
| } |
| |
| int VerifySignedPayload(const string& in_file, const string& public_key) { |
| LOG(INFO) << "Verifying signed payload."; |
| LOG_IF(FATAL, in_file.empty()) |
| << "Must pass --in_file to verify signed payload."; |
| LOG_IF(FATAL, public_key.empty()) |
| << "Must pass --public_key to verify signed payload."; |
| if (!PayloadSigner::VerifySignedPayload(in_file, public_key)) { |
| LOG(INFO) << "VerifySignedPayload failed"; |
| return 1; |
| } |
| |
| LOG(INFO) << "Done verifying signed payload."; |
| return 0; |
| } |
| |
| class ApplyPayloadProcessorDelegate : public ActionProcessorDelegate { |
| public: |
| void ProcessingDone(const ActionProcessor* processor, |
| ErrorCode code) override { |
| brillo::MessageLoop::current()->BreakLoop(); |
| code_ = code; |
| } |
| void ProcessingStopped(const ActionProcessor* processor) override { |
| brillo::MessageLoop::current()->BreakLoop(); |
| } |
| ErrorCode code_; |
| }; |
| |
| // TODO(deymo): Move this function to a new file and make the delta_performer |
| // integration tests use this instead. |
| bool ApplyPayload(const string& payload_file, |
| // Simply reuses the payload config used for payload |
| // generation. |
| const PayloadGenerationConfig& config) { |
| LOG(INFO) << "Applying delta."; |
| FakeBootControl fake_boot_control; |
| FakeHardware fake_hardware; |
| MemoryPrefs prefs; |
| InstallPlan install_plan; |
| InstallPlan::Payload payload; |
| install_plan.source_slot = |
| config.is_delta ? 0 : BootControlInterface::kInvalidSlot; |
| install_plan.target_slot = 1; |
| // For partial updates, we always write kDelta to the payload. Make it |
| // consistent for host simulation. |
| payload.type = config.is_delta || config.is_partial_update |
| ? InstallPayloadType::kDelta |
| : InstallPayloadType::kFull; |
| payload.size = utils::FileSize(payload_file); |
| // TODO(senj): This hash is only correct for unsigned payload, need to support |
| // signed payload using PayloadSigner. |
| HashCalculator::RawHashOfFile(payload_file, payload.size, &payload.hash); |
| install_plan.payloads = {payload}; |
| install_plan.download_url = |
| "file://" + |
| base::MakeAbsoluteFilePath(base::FilePath(payload_file)).value(); |
| |
| for (size_t i = 0; i < config.target.partitions.size(); i++) { |
| const string& part_name = config.target.partitions[i].name; |
| const string& target_path = config.target.partitions[i].path; |
| fake_boot_control.SetPartitionDevice( |
| part_name, install_plan.target_slot, target_path); |
| |
| string source_path; |
| if (config.is_delta) { |
| TEST_AND_RETURN_FALSE(config.target.partitions.size() == |
| config.source.partitions.size()); |
| source_path = config.source.partitions[i].path; |
| fake_boot_control.SetPartitionDevice( |
| part_name, install_plan.source_slot, source_path); |
| } |
| |
| LOG(INFO) << "Install partition:" |
| << " source: " << source_path << "\ttarget: " << target_path; |
| } |
| |
| xz_crc32_init(); |
| brillo::BaseMessageLoop loop; |
| loop.SetAsCurrent(); |
| auto install_plan_action = std::make_unique<InstallPlanAction>(install_plan); |
| auto download_action = |
| std::make_unique<DownloadAction>(&prefs, |
| &fake_boot_control, |
| &fake_hardware, |
| new FileFetcher(), |
| true /* interactive */); |
| auto filesystem_verifier_action = std::make_unique<FilesystemVerifierAction>( |
| fake_boot_control.GetDynamicPartitionControl()); |
| |
| BondActions(install_plan_action.get(), download_action.get()); |
| BondActions(download_action.get(), filesystem_verifier_action.get()); |
| ActionProcessor processor; |
| ApplyPayloadProcessorDelegate delegate; |
| processor.set_delegate(&delegate); |
| processor.EnqueueAction(std::move(install_plan_action)); |
| processor.EnqueueAction(std::move(download_action)); |
| processor.EnqueueAction(std::move(filesystem_verifier_action)); |
| loop.PostTask(FROM_HERE, |
| base::Bind(&ActionProcessor::StartProcessing, |
| base::Unretained(&processor))); |
| loop.Run(); |
| CHECK_EQ(delegate.code_, ErrorCode::kSuccess); |
| LOG(INFO) << "Completed applying " << (config.is_delta ? "delta" : "full") |
| << " payload."; |
| return true; |
| } |
| |
| bool ExtractProperties(const string& payload_path, |
| const string& props_file, |
| const string& props_format) { |
| string properties; |
| PayloadProperties payload_props(payload_path); |
| if (props_format == kPayloadPropertiesFormatKeyValue) { |
| TEST_AND_RETURN_FALSE(payload_props.GetPropertiesAsKeyValue(&properties)); |
| } else if (props_format == kPayloadPropertiesFormatJson) { |
| TEST_AND_RETURN_FALSE(payload_props.GetPropertiesAsJson(&properties)); |
| } else { |
| LOG(FATAL) << "Invalid option " << props_format |
| << " for --properties_format flag."; |
| } |
| if (props_file == "-") { |
| printf("%s", properties.c_str()); |
| } else { |
| utils::WriteFile( |
| props_file.c_str(), properties.c_str(), properties.length()); |
| LOG(INFO) << "Generated properties file at " << props_file; |
| } |
| return true; |
| } |
| |
| template <typename Key, typename Val> |
| string ToString(const map<Key, Val>& map) { |
| vector<string> result; |
| result.reserve(map.size()); |
| for (const auto& it : map) { |
| result.emplace_back(it.first + ": " + it.second); |
| } |
| return "{" + base::JoinString(result, ",") + "}"; |
| } |
| |
| bool ParsePerPartitionTimestamps(const string& partition_timestamps, |
| PayloadGenerationConfig* config) { |
| base::StringPairs pairs; |
| CHECK(base::SplitStringIntoKeyValuePairs( |
| partition_timestamps, ':', ',', &pairs)) |
| << "--partition_timestamps accepts commad " |
| "separated pairs. e.x. system:1234,vendor:5678"; |
| map<string, string> partition_timestamps_map{ |
| std::move_iterator(pairs.begin()), std::move_iterator(pairs.end())}; |
| for (auto&& partition : config->target.partitions) { |
| auto&& it = partition_timestamps_map.find(partition.name); |
| if (it != partition_timestamps_map.end()) { |
| partition.version = std::move(it->second); |
| partition_timestamps_map.erase(it); |
| } |
| } |
| if (!partition_timestamps_map.empty()) { |
| LOG(ERROR) << "Unused timestamps: " << ToString(partition_timestamps_map); |
| return false; |
| } |
| return true; |
| } |
| |
| int Main(int argc, char** argv) { |
| DEFINE_string(old_image, "", "Path to the old rootfs"); |
| DEFINE_string(new_image, "", "Path to the new rootfs"); |
| DEFINE_string(old_kernel, "", "Path to the old kernel partition image"); |
| DEFINE_string(new_kernel, "", "Path to the new kernel partition image"); |
| DEFINE_string(old_partitions, |
| "", |
| "Path to the old partitions. To pass multiple partitions, use " |
| "a single argument with a colon between paths, e.g. " |
| "/path/to/part:/path/to/part2::/path/to/last_part . Path can " |
| "be empty, but it has to match the order of partition_names."); |
| DEFINE_string(new_partitions, |
| "", |
| "Path to the new partitions. To pass multiple partitions, use " |
| "a single argument with a colon between paths, e.g. " |
| "/path/to/part:/path/to/part2:/path/to/last_part . Path has " |
| "to match the order of partition_names."); |
| DEFINE_string(old_mapfiles, |
| "", |
| "Path to the .map files associated with the partition files " |
| "in the old partition. The .map file is normally generated " |
| "when creating the image in Android builds. Only recommended " |
| "for unsupported filesystem. Pass multiple files separated by " |
| "a colon as with -old_partitions."); |
| DEFINE_string(new_mapfiles, |
| "", |
| "Path to the .map files associated with the partition files " |
| "in the new partition, similar to the -old_mapfiles flag."); |
| DEFINE_string(partition_names, |
| string(kPartitionNameRoot) + ":" + kPartitionNameKernel, |
| "Names of the partitions. To pass multiple names, use a single " |
| "argument with a colon between names, e.g. " |
| "name:name2:name3:last_name . Name can not be empty, and it " |
| "has to match the order of partitions."); |
| DEFINE_string(in_file, |
| "", |
| "Path to input delta payload file used to hash/sign payloads " |
| "and apply delta over old_image (for debugging)"); |
| DEFINE_string(out_file, "", "Path to output delta payload file"); |
| DEFINE_string(out_hash_file, "", "Path to output hash file"); |
| DEFINE_string( |
| out_metadata_hash_file, "", "Path to output metadata hash file"); |
| DEFINE_string( |
| out_metadata_size_file, "", "Path to output metadata size file"); |
| DEFINE_string(private_key, "", "Path to private key in .pem format"); |
| DEFINE_string(public_key, "", "Path to public key in .pem format"); |
| DEFINE_int32( |
| public_key_version, -1, "DEPRECATED. Key-check version # of client"); |
| DEFINE_string(signature_size, |
| "", |
| "Raw signature size used for hash calculation. " |
| "You may pass in multiple sizes by colon separating them. E.g. " |
| "2048:2048:4096 will assume 3 signatures, the first two with " |
| "2048 size and the last 4096."); |
| DEFINE_string(payload_signature_file, |
| "", |
| "Raw signature file to sign payload with. To pass multiple " |
| "signatures, use a single argument with a colon between paths, " |
| "e.g. /path/to/sig:/path/to/next:/path/to/last_sig . Each " |
| "signature will be assigned a client version, starting from " |
| "kSignatureOriginalVersion."); |
| DEFINE_string(metadata_signature_file, |
| "", |
| "Raw signature file with the signature of the metadata hash. " |
| "To pass multiple signatures, use a single argument with a " |
| "colon between paths, " |
| "e.g. /path/to/sig:/path/to/next:/path/to/last_sig ."); |
| DEFINE_int32( |
| chunk_size, 200 * 1024 * 1024, "Payload chunk size (-1 for whole files)"); |
| DEFINE_uint64(rootfs_partition_size, |
| chromeos_update_engine::kRootFSPartitionSize, |
| "RootFS partition size for the image once installed"); |
| DEFINE_uint64( |
| major_version, 2, "The major version of the payload being generated."); |
| DEFINE_int32(minor_version, |
| -1, |
| "The minor version of the payload being generated " |
| "(-1 means autodetect)."); |
| DEFINE_string(properties_file, |
| "", |
| "If passed, dumps the payload properties of the payload passed " |
| "in --in_file and exits. Look at --properties_format."); |
| DEFINE_string(properties_format, |
| kPayloadPropertiesFormatKeyValue, |
| "Defines the format of the --properties_file. The acceptable " |
| "values are: key-value (default) and json"); |
| DEFINE_int64(max_timestamp, |
| 0, |
| "The maximum timestamp of the OS allowed to apply this " |
| "payload."); |
| DEFINE_string( |
| partition_timestamps, |
| "", |
| "The per-partition maximum timestamps which the OS allowed to apply this " |
| "payload. Passed in comma separated pairs, e.x. system:1234,vendor:5678"); |
| |
| DEFINE_string(new_postinstall_config_file, |
| "", |
| "A config file specifying postinstall related metadata. " |
| "Only allowed in major version 2 or newer."); |
| DEFINE_string(dynamic_partition_info_file, |
| "", |
| "An info file specifying dynamic partition metadata. " |
| "Only allowed in major version 2 or newer."); |
| DEFINE_bool(disable_fec_computation, |
| false, |
| "Disables the fec data computation on device."); |
| DEFINE_bool(disable_verity_computation, |
| false, |
| "Disables the verity data computation on device."); |
| DEFINE_string( |
| out_maximum_signature_size_file, |
| "", |
| "Path to the output maximum signature size given a private key."); |
| DEFINE_bool(is_partial_update, |
| false, |
| "The payload only targets a subset of partitions on the device," |
| "e.g. generic kernel image update."); |
| DEFINE_bool( |
| disable_vabc, |
| false, |
| "Whether to disable Virtual AB Compression when installing the OTA"); |
| DEFINE_bool(enable_vabc_xor, |
| false, |
| "Whether to use Virtual AB Compression XOR feature"); |
| DEFINE_string( |
| apex_info_file, "", "Path to META/apex_info.pb found in target build"); |
| DEFINE_string(compressor_types, |
| "bz2:brotli", |
| "Colon ':' separated list of compressors. Allowed valures are " |
| "bz2 and brotli."); |
| DEFINE_bool( |
| enable_lz4diff, |
| false, |
| "Whether to enable LZ4diff feature when processing EROFS images."); |
| |
| DEFINE_bool( |
| enable_zucchini, |
| true, |
| "Whether to enable zucchini feature when processing executable files."); |
| |
| DEFINE_string(erofs_compression_param, |
| "", |
| "Compression parameter passed to mkfs.erofs's -z option. " |
| "Example: lz4 lz4hc,9"); |
| |
| brillo::FlagHelper::Init( |
| argc, |
| argv, |
| "Generates a payload to provide to ChromeOS' update_engine.\n\n" |
| "This tool can create full payloads and also delta payloads if the src\n" |
| "image is provided. It also provides debugging options to apply, sign\n" |
| "and verify payloads."); |
| Terminator::Init(); |
| |
| logging::LoggingSettings log_settings; |
| #if BASE_VER < 780000 |
| log_settings.log_file = "delta_generator.log"; |
| #else |
| log_settings.log_file_path = "delta_generator.log"; |
| #endif |
| log_settings.logging_dest = logging::LOG_TO_SYSTEM_DEBUG_LOG; |
| log_settings.lock_log = logging::LOCK_LOG_FILE; |
| log_settings.delete_old = logging::APPEND_TO_OLD_LOG_FILE; |
| |
| logging::InitLogging(log_settings); |
| |
| // Initialize the Xz compressor. |
| XzCompressInit(); |
| |
| if (!FLAGS_out_maximum_signature_size_file.empty()) { |
| LOG_IF(FATAL, FLAGS_private_key.empty()) |
| << "Private key is not provided when calculating the maximum signature " |
| "size."; |
| |
| size_t maximum_signature_size; |
| if (!PayloadSigner::GetMaximumSignatureSize(FLAGS_private_key, |
| &maximum_signature_size)) { |
| LOG(ERROR) << "Failed to get the maximum signature size of private key: " |
| << FLAGS_private_key; |
| return 1; |
| } |
| // Write the size string to output file. |
| string signature_size_string = std::to_string(maximum_signature_size); |
| if (!utils::WriteFile(FLAGS_out_maximum_signature_size_file.c_str(), |
| signature_size_string.c_str(), |
| signature_size_string.size())) { |
| PLOG(ERROR) << "Failed to write the maximum signature size to " |
| << FLAGS_out_maximum_signature_size_file << "."; |
| return 1; |
| } |
| return 0; |
| } |
| |
| vector<size_t> signature_sizes; |
| if (!FLAGS_signature_size.empty()) { |
| ParseSignatureSizes(FLAGS_signature_size, &signature_sizes); |
| } |
| |
| if (!FLAGS_out_hash_file.empty() || !FLAGS_out_metadata_hash_file.empty()) { |
| CHECK(FLAGS_out_metadata_size_file.empty()); |
| CalculateHashForSigning(signature_sizes, |
| FLAGS_out_hash_file, |
| FLAGS_out_metadata_hash_file, |
| FLAGS_in_file); |
| return 0; |
| } |
| if (!FLAGS_payload_signature_file.empty()) { |
| SignPayload(FLAGS_in_file, |
| FLAGS_out_file, |
| signature_sizes, |
| FLAGS_payload_signature_file, |
| FLAGS_metadata_signature_file, |
| FLAGS_out_metadata_size_file); |
| return 0; |
| } |
| if (!FLAGS_public_key.empty()) { |
| LOG_IF(WARNING, FLAGS_public_key_version != -1) |
| << "--public_key_version is deprecated and ignored."; |
| return VerifySignedPayload(FLAGS_in_file, FLAGS_public_key); |
| } |
| if (!FLAGS_properties_file.empty()) { |
| return ExtractProperties( |
| FLAGS_in_file, FLAGS_properties_file, FLAGS_properties_format) |
| ? 0 |
| : 1; |
| } |
| |
| // A payload generation was requested. Convert the flags to a |
| // PayloadGenerationConfig. |
| PayloadGenerationConfig payload_config; |
| vector<string> partition_names, old_partitions, new_partitions; |
| vector<string> old_mapfiles, new_mapfiles; |
| |
| if (!FLAGS_old_mapfiles.empty()) { |
| old_mapfiles = base::SplitString( |
| FLAGS_old_mapfiles, ":", base::TRIM_WHITESPACE, base::SPLIT_WANT_ALL); |
| } |
| if (!FLAGS_new_mapfiles.empty()) { |
| new_mapfiles = base::SplitString( |
| FLAGS_new_mapfiles, ":", base::TRIM_WHITESPACE, base::SPLIT_WANT_ALL); |
| } |
| |
| partition_names = base::SplitString( |
| FLAGS_partition_names, ":", base::TRIM_WHITESPACE, base::SPLIT_WANT_ALL); |
| CHECK(!partition_names.empty()); |
| if (FLAGS_major_version < kMinSupportedMajorPayloadVersion || |
| FLAGS_major_version > kMaxSupportedMajorPayloadVersion) { |
| LOG(FATAL) << "Unsupported major version " << FLAGS_major_version; |
| return 1; |
| } |
| |
| if (!FLAGS_apex_info_file.empty()) { |
| // apex_info_file should point to a regular file(or symlink to a regular |
| // file) |
| CHECK(utils::FileExists(FLAGS_apex_info_file.c_str())); |
| CHECK(utils::IsRegFile(FLAGS_apex_info_file.c_str()) || |
| utils::IsSymlink(FLAGS_apex_info_file.c_str())); |
| payload_config.apex_info_file = FLAGS_apex_info_file; |
| } |
| |
| payload_config.enable_vabc_xor = FLAGS_enable_vabc_xor; |
| payload_config.enable_lz4diff = FLAGS_enable_lz4diff; |
| payload_config.enable_zucchini = FLAGS_enable_zucchini; |
| |
| payload_config.ParseCompressorTypes(FLAGS_compressor_types); |
| |
| if (!FLAGS_new_partitions.empty()) { |
| LOG_IF(FATAL, !FLAGS_new_image.empty() || !FLAGS_new_kernel.empty()) |
| << "--new_image and --new_kernel are deprecated, please use " |
| << "--new_partitions for all partitions."; |
| new_partitions = base::SplitString( |
| FLAGS_new_partitions, ":", base::TRIM_WHITESPACE, base::SPLIT_WANT_ALL); |
| CHECK(partition_names.size() == new_partitions.size()); |
| |
| payload_config.is_delta = !FLAGS_old_partitions.empty(); |
| LOG_IF(FATAL, !FLAGS_old_image.empty() || !FLAGS_old_kernel.empty()) |
| << "--old_image and --old_kernel are deprecated, please use " |
| << "--old_partitions if you are using --new_partitions."; |
| } else { |
| new_partitions = {FLAGS_new_image, FLAGS_new_kernel}; |
| LOG(WARNING) << "--new_partitions is empty, using deprecated --new_image " |
| << "and --new_kernel flags."; |
| |
| payload_config.is_delta = |
| !FLAGS_old_image.empty() || !FLAGS_old_kernel.empty(); |
| LOG_IF(FATAL, !FLAGS_old_partitions.empty()) |
| << "Please use --new_partitions if you are using --old_partitions."; |
| } |
| for (size_t i = 0; i < partition_names.size(); i++) { |
| LOG_IF(FATAL, partition_names[i].empty()) |
| << "Partition name can't be empty, see --partition_names."; |
| payload_config.target.partitions.emplace_back(partition_names[i]); |
| payload_config.target.partitions.back().path = new_partitions[i]; |
| payload_config.target.partitions.back().disable_fec_computation = |
| FLAGS_disable_fec_computation; |
| if (!FLAGS_erofs_compression_param.empty()) { |
| payload_config.target.partitions.back().erofs_compression_param = |
| PartitionConfig::ParseCompressionParam(FLAGS_erofs_compression_param); |
| } |
| if (i < new_mapfiles.size()) |
| payload_config.target.partitions.back().mapfile_path = new_mapfiles[i]; |
| } |
| |
| if (payload_config.is_delta) { |
| if (!FLAGS_old_partitions.empty()) { |
| old_partitions = base::SplitString(FLAGS_old_partitions, |
| ":", |
| base::TRIM_WHITESPACE, |
| base::SPLIT_WANT_ALL); |
| CHECK(old_partitions.size() == new_partitions.size()); |
| } else { |
| old_partitions = {FLAGS_old_image, FLAGS_old_kernel}; |
| LOG(WARNING) << "--old_partitions is empty, using deprecated --old_image " |
| << "and --old_kernel flags."; |
| } |
| for (size_t i = 0; i < partition_names.size(); i++) { |
| payload_config.source.partitions.emplace_back(partition_names[i]); |
| payload_config.source.partitions.back().path = old_partitions[i]; |
| if (i < old_mapfiles.size()) |
| payload_config.source.partitions.back().mapfile_path = old_mapfiles[i]; |
| } |
| } |
| |
| if (FLAGS_is_partial_update) { |
| payload_config.is_partial_update = true; |
| } |
| |
| if (!FLAGS_in_file.empty()) { |
| return ApplyPayload(FLAGS_in_file, payload_config) ? 0 : 1; |
| } |
| |
| if (!FLAGS_new_postinstall_config_file.empty()) { |
| brillo::KeyValueStore store; |
| CHECK(store.Load(base::FilePath(FLAGS_new_postinstall_config_file))); |
| CHECK(payload_config.target.LoadPostInstallConfig(store)); |
| } |
| |
| // Use the default soft_chunk_size defined in the config. |
| payload_config.hard_chunk_size = FLAGS_chunk_size; |
| payload_config.block_size = kBlockSize; |
| |
| // The partition size is never passed to the delta_generator, so we |
| // need to detect those from the provided files. |
| if (payload_config.is_delta) { |
| CHECK(payload_config.source.LoadImageSize()); |
| } |
| CHECK(payload_config.target.LoadImageSize()); |
| |
| if (!FLAGS_dynamic_partition_info_file.empty()) { |
| brillo::KeyValueStore store; |
| CHECK(store.Load(base::FilePath(FLAGS_dynamic_partition_info_file))); |
| CHECK(payload_config.target.LoadDynamicPartitionMetadata(store)); |
| CHECK(payload_config.target.ValidateDynamicPartitionMetadata()); |
| if (FLAGS_disable_vabc) { |
| LOG(INFO) << "Disabling VABC"; |
| payload_config.target.dynamic_partition_metadata->set_vabc_enabled(false); |
| payload_config.target.dynamic_partition_metadata |
| ->set_vabc_compression_param(""); |
| } |
| } |
| |
| CHECK(!FLAGS_out_file.empty()); |
| |
| payload_config.rootfs_partition_size = FLAGS_rootfs_partition_size; |
| |
| if (payload_config.is_delta) { |
| // Avoid opening the filesystem interface for full payloads. |
| for (PartitionConfig& part : payload_config.target.partitions) |
| CHECK(part.OpenFilesystem()); |
| for (PartitionConfig& part : payload_config.source.partitions) |
| CHECK(part.OpenFilesystem()); |
| } |
| |
| payload_config.version.major = FLAGS_major_version; |
| LOG(INFO) << "Using provided major_version=" << FLAGS_major_version; |
| |
| if (FLAGS_minor_version == -1) { |
| // Autodetect minor_version by looking at the update_engine.conf in the old |
| // image. |
| if (payload_config.is_delta) { |
| brillo::KeyValueStore store; |
| uint32_t minor_version; |
| bool minor_version_found = false; |
| for (const PartitionConfig& part : payload_config.source.partitions) { |
| if (part.fs_interface && part.fs_interface->LoadSettings(&store) && |
| utils::GetMinorVersion(store, &minor_version)) { |
| payload_config.version.minor = minor_version; |
| minor_version_found = true; |
| LOG(INFO) << "Auto-detected minor_version=" |
| << payload_config.version.minor; |
| break; |
| } |
| } |
| if (!minor_version_found) { |
| LOG(FATAL) << "Failed to detect the minor version."; |
| return 1; |
| } |
| } else { |
| payload_config.version.minor = kFullPayloadMinorVersion; |
| LOG(INFO) << "Using non-delta minor_version=" |
| << payload_config.version.minor; |
| } |
| } else { |
| payload_config.version.minor = FLAGS_minor_version; |
| LOG(INFO) << "Using provided minor_version=" << FLAGS_minor_version; |
| } |
| |
| if (payload_config.version.minor != kFullPayloadMinorVersion && |
| (payload_config.version.minor < kMinSupportedMinorPayloadVersion || |
| payload_config.version.minor > kMaxSupportedMinorPayloadVersion)) { |
| LOG(FATAL) << "Unsupported minor version " << payload_config.version.minor; |
| return 1; |
| } |
| |
| payload_config.max_timestamp = FLAGS_max_timestamp; |
| if (!FLAGS_partition_timestamps.empty()) { |
| CHECK(ParsePerPartitionTimestamps(FLAGS_partition_timestamps, |
| &payload_config)); |
| } |
| |
| if (payload_config.is_delta && |
| payload_config.version.minor >= kVerityMinorPayloadVersion && |
| !FLAGS_disable_verity_computation) |
| CHECK(payload_config.target.LoadVerityConfig()); |
| |
| LOG(INFO) << "Generating " << (payload_config.is_delta ? "delta" : "full") |
| << " update"; |
| |
| // From this point, all the options have been parsed. |
| if (!payload_config.Validate()) { |
| LOG(ERROR) << "Invalid options passed. See errors above."; |
| return 1; |
| } |
| |
| uint64_t metadata_size; |
| if (!GenerateUpdatePayloadFile( |
| payload_config, FLAGS_out_file, FLAGS_private_key, &metadata_size)) { |
| return 1; |
| } |
| if (!FLAGS_out_metadata_size_file.empty()) { |
| string metadata_size_string = std::to_string(metadata_size); |
| CHECK(utils::WriteFile(FLAGS_out_metadata_size_file.c_str(), |
| metadata_size_string.data(), |
| metadata_size_string.size())); |
| } |
| return 0; |
| } |
| |
| } // namespace |
| |
| } // namespace chromeos_update_engine |
| |
| int main(int argc, char** argv) { |
| return chromeos_update_engine::Main(argc, argv); |
| } |