| // |
| // Copyright (C) 2015 The Android Open Source Project |
| // |
| // Licensed under the Apache License, Version 2.0 (the "License"); |
| // you may not use this file except in compliance with the License. |
| // You may obtain a copy of the License at |
| // |
| // http://www.apache.org/licenses/LICENSE-2.0 |
| // |
| // Unless required by applicable law or agreed to in writing, software |
| // distributed under the License is distributed on an "AS IS" BASIS, |
| // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| // See the License for the specific language governing permissions and |
| // limitations under the License. |
| // |
| |
| #include "update_engine/payload_generator/ab_generator.h" |
| |
| #include <algorithm> |
| |
| #include <base/strings/stringprintf.h> |
| |
| #include "update_engine/omaha_hash_calculator.h" |
| #include "update_engine/payload_constants.h" |
| #include "update_engine/payload_generator/annotated_operation.h" |
| #include "update_engine/payload_generator/bzip.h" |
| #include "update_engine/payload_generator/delta_diff_generator.h" |
| #include "update_engine/payload_generator/delta_diff_utils.h" |
| #include "update_engine/utils.h" |
| |
| using std::string; |
| using std::vector; |
| |
| namespace chromeos_update_engine { |
| |
| bool ABGenerator::GenerateOperations( |
| const PayloadGenerationConfig& config, |
| const PartitionConfig& old_part, |
| const PartitionConfig& new_part, |
| BlobFileWriter* blob_file, |
| vector<AnnotatedOperation>* aops) { |
| TEST_AND_RETURN_FALSE(old_part.name == new_part.name); |
| |
| ssize_t hard_chunk_blocks = (config.hard_chunk_size == -1 ? -1 : |
| config.hard_chunk_size / config.block_size); |
| size_t soft_chunk_blocks = config.soft_chunk_size / config.block_size; |
| |
| aops->clear(); |
| TEST_AND_RETURN_FALSE(diff_utils::DeltaReadPartition( |
| aops, |
| old_part, |
| new_part, |
| hard_chunk_blocks, |
| soft_chunk_blocks, |
| blob_file, |
| true)); // src_ops_allowed |
| LOG(INFO) << "done reading " << new_part.name; |
| |
| TEST_AND_RETURN_FALSE(FragmentOperations(aops, |
| new_part.path, |
| blob_file)); |
| SortOperationsByDestination(aops); |
| |
| // Use the soft_chunk_size when merging operations to prevent merging all |
| // the operations into a huge one if there's no hard limit. |
| size_t merge_chunk_blocks = soft_chunk_blocks; |
| if (hard_chunk_blocks != -1 && |
| static_cast<size_t>(hard_chunk_blocks) < soft_chunk_blocks) { |
| merge_chunk_blocks = hard_chunk_blocks; |
| } |
| |
| TEST_AND_RETURN_FALSE(MergeOperations(aops, |
| merge_chunk_blocks, |
| new_part.path, |
| blob_file)); |
| |
| if (config.minor_version == kOpSrcHashMinorPayloadVersion) |
| TEST_AND_RETURN_FALSE(AddSourceHash(aops, old_part.path)); |
| |
| return true; |
| } |
| |
| void ABGenerator::SortOperationsByDestination( |
| vector<AnnotatedOperation>* aops) { |
| sort(aops->begin(), aops->end(), diff_utils::CompareAopsByDestination); |
| } |
| |
| bool ABGenerator::FragmentOperations( |
| vector<AnnotatedOperation>* aops, |
| const string& target_part_path, |
| BlobFileWriter* blob_file) { |
| vector<AnnotatedOperation> fragmented_aops; |
| for (const AnnotatedOperation& aop : *aops) { |
| if (aop.op.type() == InstallOperation::SOURCE_COPY) { |
| TEST_AND_RETURN_FALSE(SplitSourceCopy(aop, &fragmented_aops)); |
| } else if ((aop.op.type() == InstallOperation::REPLACE) || |
| (aop.op.type() == InstallOperation::REPLACE_BZ)) { |
| TEST_AND_RETURN_FALSE(SplitReplaceOrReplaceBz(aop, &fragmented_aops, |
| target_part_path, |
| blob_file)); |
| } else { |
| fragmented_aops.push_back(aop); |
| } |
| } |
| *aops = fragmented_aops; |
| return true; |
| } |
| |
| bool ABGenerator::SplitSourceCopy( |
| const AnnotatedOperation& original_aop, |
| vector<AnnotatedOperation>* result_aops) { |
| InstallOperation original_op = original_aop.op; |
| TEST_AND_RETURN_FALSE(original_op.type() == InstallOperation::SOURCE_COPY); |
| // Keeps track of the index of curr_src_ext. |
| int curr_src_ext_index = 0; |
| Extent curr_src_ext = original_op.src_extents(curr_src_ext_index); |
| for (int i = 0; i < original_op.dst_extents_size(); i++) { |
| Extent dst_ext = original_op.dst_extents(i); |
| // The new operation which will have only one dst extent. |
| InstallOperation new_op; |
| uint64_t blocks_left = dst_ext.num_blocks(); |
| while (blocks_left > 0) { |
| if (curr_src_ext.num_blocks() <= blocks_left) { |
| // If the curr_src_ext is smaller than dst_ext, add it. |
| blocks_left -= curr_src_ext.num_blocks(); |
| *(new_op.add_src_extents()) = curr_src_ext; |
| if (curr_src_ext_index + 1 < original_op.src_extents().size()) { |
| curr_src_ext = original_op.src_extents(++curr_src_ext_index); |
| } else { |
| break; |
| } |
| } else { |
| // Split src_exts that are bigger than the dst_ext we're dealing with. |
| Extent first_ext; |
| first_ext.set_num_blocks(blocks_left); |
| first_ext.set_start_block(curr_src_ext.start_block()); |
| *(new_op.add_src_extents()) = first_ext; |
| // Keep the second half of the split op. |
| curr_src_ext.set_num_blocks(curr_src_ext.num_blocks() - blocks_left); |
| curr_src_ext.set_start_block(curr_src_ext.start_block() + blocks_left); |
| blocks_left -= first_ext.num_blocks(); |
| } |
| } |
| // Fix up our new operation and add it to the results. |
| new_op.set_type(InstallOperation::SOURCE_COPY); |
| *(new_op.add_dst_extents()) = dst_ext; |
| new_op.set_src_length(dst_ext.num_blocks() * kBlockSize); |
| new_op.set_dst_length(dst_ext.num_blocks() * kBlockSize); |
| |
| AnnotatedOperation new_aop; |
| new_aop.op = new_op; |
| new_aop.name = base::StringPrintf("%s:%d", original_aop.name.c_str(), i); |
| result_aops->push_back(new_aop); |
| } |
| if (curr_src_ext_index != original_op.src_extents().size() - 1) { |
| LOG(FATAL) << "Incorrectly split SOURCE_COPY operation. Did not use all " |
| << "source extents."; |
| } |
| return true; |
| } |
| |
| bool ABGenerator::SplitReplaceOrReplaceBz( |
| const AnnotatedOperation& original_aop, |
| vector<AnnotatedOperation>* result_aops, |
| const string& target_part_path, |
| BlobFileWriter* blob_file) { |
| InstallOperation original_op = original_aop.op; |
| const bool is_replace = original_op.type() == InstallOperation::REPLACE; |
| TEST_AND_RETURN_FALSE(is_replace || |
| original_op.type() == InstallOperation::REPLACE_BZ); |
| |
| uint32_t data_offset = original_op.data_offset(); |
| for (int i = 0; i < original_op.dst_extents_size(); i++) { |
| Extent dst_ext = original_op.dst_extents(i); |
| // Make a new operation with only one dst extent. |
| InstallOperation new_op; |
| *(new_op.add_dst_extents()) = dst_ext; |
| uint32_t data_size = dst_ext.num_blocks() * kBlockSize; |
| new_op.set_dst_length(data_size); |
| // If this is a REPLACE, attempt to reuse portions of the existing blob. |
| if (is_replace) { |
| new_op.set_type(InstallOperation::REPLACE); |
| new_op.set_data_length(data_size); |
| new_op.set_data_offset(data_offset); |
| data_offset += data_size; |
| } |
| |
| AnnotatedOperation new_aop; |
| new_aop.op = new_op; |
| new_aop.name = base::StringPrintf("%s:%d", original_aop.name.c_str(), i); |
| TEST_AND_RETURN_FALSE(AddDataAndSetType(&new_aop, target_part_path, |
| blob_file)); |
| |
| result_aops->push_back(new_aop); |
| } |
| return true; |
| } |
| |
| bool ABGenerator::MergeOperations(vector<AnnotatedOperation>* aops, |
| size_t chunk_blocks, |
| const string& target_part_path, |
| BlobFileWriter* blob_file) { |
| vector<AnnotatedOperation> new_aops; |
| for (const AnnotatedOperation& curr_aop : *aops) { |
| if (new_aops.empty()) { |
| new_aops.push_back(curr_aop); |
| continue; |
| } |
| AnnotatedOperation& last_aop = new_aops.back(); |
| |
| if (last_aop.op.dst_extents_size() <= 0 || |
| curr_aop.op.dst_extents_size() <= 0) { |
| new_aops.push_back(curr_aop); |
| continue; |
| } |
| uint32_t last_dst_idx = last_aop.op.dst_extents_size() - 1; |
| uint32_t last_end_block = |
| last_aop.op.dst_extents(last_dst_idx).start_block() + |
| last_aop.op.dst_extents(last_dst_idx).num_blocks(); |
| uint32_t curr_start_block = curr_aop.op.dst_extents(0).start_block(); |
| uint32_t combined_block_count = |
| last_aop.op.dst_extents(last_dst_idx).num_blocks() + |
| curr_aop.op.dst_extents(0).num_blocks(); |
| bool good_op_type = curr_aop.op.type() == InstallOperation::SOURCE_COPY || |
| curr_aop.op.type() == InstallOperation::REPLACE || |
| curr_aop.op.type() == InstallOperation::REPLACE_BZ; |
| if (good_op_type && |
| last_aop.op.type() == curr_aop.op.type() && |
| last_end_block == curr_start_block && |
| combined_block_count <= chunk_blocks) { |
| // If the operations have the same type (which is a type that we can |
| // merge), are contiguous, are fragmented to have one destination extent, |
| // and their combined block count would be less than chunk size, merge |
| // them. |
| last_aop.name = base::StringPrintf("%s,%s", |
| last_aop.name.c_str(), |
| curr_aop.name.c_str()); |
| |
| ExtendExtents(last_aop.op.mutable_src_extents(), |
| curr_aop.op.src_extents()); |
| if (curr_aop.op.src_length() > 0) |
| last_aop.op.set_src_length(last_aop.op.src_length() + |
| curr_aop.op.src_length()); |
| ExtendExtents(last_aop.op.mutable_dst_extents(), |
| curr_aop.op.dst_extents()); |
| if (curr_aop.op.dst_length() > 0) |
| last_aop.op.set_dst_length(last_aop.op.dst_length() + |
| curr_aop.op.dst_length()); |
| // Set the data length to zero so we know to add the blob later. |
| if (curr_aop.op.type() == InstallOperation::REPLACE || |
| curr_aop.op.type() == InstallOperation::REPLACE_BZ) { |
| last_aop.op.set_data_length(0); |
| } |
| } else { |
| // Otherwise just include the extent as is. |
| new_aops.push_back(curr_aop); |
| } |
| } |
| |
| // Set the blobs for REPLACE/REPLACE_BZ operations that have been merged. |
| for (AnnotatedOperation& curr_aop : new_aops) { |
| if (curr_aop.op.data_length() == 0 && |
| (curr_aop.op.type() == InstallOperation::REPLACE || |
| curr_aop.op.type() == InstallOperation::REPLACE_BZ)) { |
| TEST_AND_RETURN_FALSE(AddDataAndSetType(&curr_aop, target_part_path, |
| blob_file)); |
| } |
| } |
| |
| *aops = new_aops; |
| return true; |
| } |
| |
| bool ABGenerator::AddDataAndSetType(AnnotatedOperation* aop, |
| const string& target_part_path, |
| BlobFileWriter* blob_file) { |
| TEST_AND_RETURN_FALSE(aop->op.type() == InstallOperation::REPLACE || |
| aop->op.type() == InstallOperation::REPLACE_BZ); |
| |
| brillo::Blob data(aop->op.dst_length()); |
| vector<Extent> dst_extents; |
| ExtentsToVector(aop->op.dst_extents(), &dst_extents); |
| TEST_AND_RETURN_FALSE(utils::ReadExtents(target_part_path, |
| dst_extents, |
| &data, |
| data.size(), |
| kBlockSize)); |
| |
| brillo::Blob data_bz; |
| TEST_AND_RETURN_FALSE(BzipCompress(data, &data_bz)); |
| CHECK(!data_bz.empty()); |
| |
| brillo::Blob* data_p = nullptr; |
| InstallOperation_Type new_op_type; |
| if (data_bz.size() < data.size()) { |
| new_op_type = InstallOperation::REPLACE_BZ; |
| data_p = &data_bz; |
| } else { |
| new_op_type = InstallOperation::REPLACE; |
| data_p = &data; |
| } |
| |
| // If the operation doesn't point to a data blob, then we add it. |
| if (aop->op.type() != new_op_type || |
| aop->op.data_length() != data_p->size()) { |
| aop->op.set_type(new_op_type); |
| aop->SetOperationBlob(data_p, blob_file); |
| } |
| |
| return true; |
| } |
| |
| bool ABGenerator::AddSourceHash(vector<AnnotatedOperation>* aops, |
| const string& source_part_path) { |
| for (AnnotatedOperation& aop : *aops) { |
| if (aop.op.src_extents_size() == 0) |
| continue; |
| |
| vector<Extent> src_extents; |
| ExtentsToVector(aop.op.src_extents(), &src_extents); |
| brillo::Blob src_data, src_hash; |
| uint64_t src_length = |
| aop.op.has_src_length() |
| ? aop.op.src_length() |
| : BlocksInExtents(aop.op.src_extents()) * kBlockSize; |
| TEST_AND_RETURN_FALSE(utils::ReadExtents( |
| source_part_path, src_extents, &src_data, src_length, kBlockSize)); |
| TEST_AND_RETURN_FALSE( |
| OmahaHashCalculator::RawHashOfData(src_data, &src_hash)); |
| aop.op.set_src_sha256_hash(src_hash.data(), src_hash.size()); |
| } |
| return true; |
| } |
| |
| } // namespace chromeos_update_engine |