| /* |
| * Copyright (C) 2015 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #include <ctype.h> |
| #include <stdlib.h> |
| |
| #include <algorithm> |
| #include <string> |
| #include <vector> |
| |
| #include <android-base/strings.h> |
| #include <openssl/evp.h> |
| |
| #include "fec_private.h" |
| |
| /* converts a hex nibble into an int */ |
| static inline int hextobin(char c) |
| { |
| if (c >= '0' && c <= '9') { |
| return c - '0'; |
| } else if (c >= 'a' && c <= 'f') { |
| return c - 'a' + 10; |
| } else { |
| errno = EINVAL; |
| return -1; |
| } |
| } |
| |
| /* converts a hex string `src' of `size' characters to binary and copies the |
| the result into `dst' */ |
| static int parse_hex(uint8_t *dst, uint32_t size, const char *src) |
| { |
| int l, h; |
| |
| check(dst); |
| check(src); |
| check(2 * size == strlen(src)); |
| |
| while (size) { |
| h = hextobin(tolower(*src++)); |
| l = hextobin(tolower(*src++)); |
| |
| check(l >= 0); |
| check(h >= 0); |
| |
| *dst++ = (h << 4) | l; |
| --size; |
| } |
| |
| return 0; |
| } |
| |
| /* parses a 64-bit unsigned integer from string `src' into `dst' and if |
| `maxval' is >0, checks that `dst' <= `maxval' */ |
| static int parse_uint64(const char *src, uint64_t maxval, uint64_t *dst) |
| { |
| char *end; |
| unsigned long long int value; |
| |
| check(src); |
| check(dst); |
| |
| errno = 0; |
| value = strtoull(src, &end, 0); |
| |
| if (*src == '\0' || *end != '\0' || |
| (errno == ERANGE && value == ULLONG_MAX)) { |
| errno = EINVAL; |
| return -1; |
| } |
| |
| if (maxval && value > maxval) { |
| errno = EINVAL; |
| return -1; |
| } |
| |
| *dst = (uint64_t)value; |
| return 0; |
| } |
| |
| /* computes the size of verity hash tree for `file_size' bytes and returns the |
| number of hash tree levels in `verity_levels,' and the number of hashes per |
| level in `level_hashes', if the parameters are non-NULL */ |
| uint64_t verity_get_size(uint64_t file_size, uint32_t *verity_levels, |
| uint32_t *level_hashes, uint32_t padded_digest_size) { |
| // we assume a known metadata size, 4 KiB block size, and SHA-256 or SHA1 to |
| // avoid relying on disk content. |
| |
| uint32_t level = 0; |
| uint64_t total = 0; |
| uint64_t hashes = file_size / FEC_BLOCKSIZE; |
| |
| do { |
| if (level_hashes) { |
| level_hashes[level] = hashes; |
| } |
| |
| hashes = fec_div_round_up(hashes * padded_digest_size, FEC_BLOCKSIZE); |
| total += hashes; |
| |
| ++level; |
| } while (hashes > 1); |
| |
| if (verity_levels) { |
| *verity_levels = level; |
| } |
| |
| return total * FEC_BLOCKSIZE; |
| } |
| |
| int hashtree_info::get_hash(const uint8_t *block, uint8_t *hash) { |
| auto md = EVP_get_digestbynid(nid_); |
| check(md); |
| auto mdctx = EVP_MD_CTX_new(); |
| check(mdctx); |
| |
| EVP_DigestInit_ex(mdctx, md, nullptr); |
| EVP_DigestUpdate(mdctx, salt.data(), salt.size()); |
| EVP_DigestUpdate(mdctx, block, FEC_BLOCKSIZE); |
| unsigned int hash_size; |
| EVP_DigestFinal_ex(mdctx, hash, &hash_size); |
| EVP_MD_CTX_free(mdctx); |
| |
| check(hash_size == digest_length_); |
| return 0; |
| } |
| |
| int hashtree_info::initialize(uint64_t hash_start, uint64_t data_blocks, |
| const std::vector<uint8_t> &salt, int nid) { |
| check(nid == NID_sha256 || nid == NID_sha1); |
| |
| this->hash_start = hash_start; |
| this->data_blocks = data_blocks; |
| this->salt = salt; |
| this->nid_ = nid; |
| |
| digest_length_ = nid == NID_sha1 ? SHA_DIGEST_LENGTH : SHA256_DIGEST_LENGTH; |
| // The padded digest size for both sha256 and sha1 are 256 bytes. |
| padded_digest_length_ = SHA256_DIGEST_LENGTH; |
| |
| return 0; |
| } |
| |
| bool hashtree_info::check_block_hash(const uint8_t *expected, |
| const uint8_t *block) { |
| check(block); |
| std::vector<uint8_t> hash(digest_length_, 0); |
| |
| if (unlikely(get_hash(block, hash.data()) == -1)) { |
| error("failed to hash"); |
| return false; |
| } |
| |
| check(expected); |
| return !memcmp(expected, hash.data(), digest_length_); |
| } |
| |
| bool hashtree_info::check_block_hash_with_index(uint64_t index, |
| const uint8_t *block) { |
| check(index < data_blocks); |
| |
| const uint8_t *expected = &hash_data[index * padded_digest_length_]; |
| return check_block_hash(expected, block); |
| } |
| |
| // Reads the hash and the corresponding data block using error correction, if |
| // available. |
| bool hashtree_info::ecc_read_hashes(fec_handle *f, uint64_t hash_offset, |
| uint8_t *hash, uint64_t data_offset, |
| uint8_t *data) { |
| check(f); |
| |
| if (hash && |
| fec_pread(f, hash, digest_length_, hash_offset) != digest_length_) { |
| error("failed to read hash tree: offset %" PRIu64 ": %s", hash_offset, |
| strerror(errno)); |
| return false; |
| } |
| |
| check(data); |
| |
| if (fec_pread(f, data, FEC_BLOCKSIZE, data_offset) != FEC_BLOCKSIZE) { |
| error("failed to read hash tree: data_offset %" PRIu64 ": %s", |
| data_offset, strerror(errno)); |
| return false; |
| } |
| |
| return true; |
| } |
| |
| int hashtree_info::verify_tree(const fec_handle *f, const uint8_t *root) { |
| check(f); |
| check(root); |
| |
| uint8_t data[FEC_BLOCKSIZE]; |
| |
| uint32_t levels = 0; |
| |
| /* calculate the size and the number of levels in the hash tree */ |
| uint64_t hash_size = verity_get_size(data_blocks * FEC_BLOCKSIZE, &levels, |
| NULL, padded_digest_length_); |
| |
| check(hash_start < UINT64_MAX - hash_size); |
| check(hash_start + hash_size <= f->data_size); |
| |
| uint64_t hash_offset = hash_start; |
| uint64_t data_offset = hash_offset + FEC_BLOCKSIZE; |
| |
| /* validate the root hash */ |
| if (!raw_pread(f->fd, data, FEC_BLOCKSIZE, hash_offset) || |
| !check_block_hash(root, data)) { |
| /* try to correct */ |
| if (!ecc_read_hashes(const_cast<fec_handle *>(f), 0, nullptr, |
| hash_offset, data) || |
| !check_block_hash(root, data)) { |
| error("root hash invalid"); |
| return -1; |
| } else if (f->mode & O_RDWR && |
| !raw_pwrite(f->fd, data, FEC_BLOCKSIZE, hash_offset)) { |
| error("failed to rewrite the root block: %s", strerror(errno)); |
| return -1; |
| } |
| } |
| |
| debug("root hash valid"); |
| |
| /* calculate the number of hashes on each level */ |
| uint32_t hashes[levels]; |
| |
| verity_get_size(data_blocks * FEC_BLOCKSIZE, NULL, hashes, |
| padded_digest_length_); |
| |
| uint64_t hash_data_offset = data_offset; |
| uint32_t hash_data_blocks = 0; |
| /* calculate the size and offset for the data hashes */ |
| for (uint32_t i = 1; i < levels; ++i) { |
| uint32_t blocks = hashes[levels - i]; |
| debug("%u hash blocks on level %u", blocks, levels - i); |
| |
| hash_data_offset = data_offset; |
| hash_data_blocks = blocks; |
| |
| data_offset += blocks * FEC_BLOCKSIZE; |
| } |
| |
| check(hash_data_blocks); |
| check(hash_data_blocks <= hash_size / FEC_BLOCKSIZE); |
| |
| check(hash_data_offset); |
| check(hash_data_offset <= UINT64_MAX - (hash_data_blocks * FEC_BLOCKSIZE)); |
| check(hash_data_offset < f->data_size); |
| check(hash_data_offset + hash_data_blocks * FEC_BLOCKSIZE <= f->data_size); |
| |
| /* copy data hashes to memory in case they are corrupted, so we don't |
| have to correct them every time they are needed */ |
| std::vector<uint8_t> data_hashes(hash_data_blocks * FEC_BLOCKSIZE, 0); |
| |
| /* validate the rest of the hash tree */ |
| data_offset = hash_offset + FEC_BLOCKSIZE; |
| |
| std::vector<uint8_t> buffer(padded_digest_length_, 0); |
| for (uint32_t i = 1; i < levels; ++i) { |
| uint32_t blocks = hashes[levels - i]; |
| |
| for (uint32_t j = 0; j < blocks; ++j) { |
| /* ecc reads are very I/O intensive, so read raw hash tree and do |
| error correcting only if it doesn't validate */ |
| if (!raw_pread(f->fd, buffer.data(), padded_digest_length_, |
| hash_offset + j * padded_digest_length_) || |
| !raw_pread(f->fd, data, FEC_BLOCKSIZE, |
| data_offset + j * FEC_BLOCKSIZE)) { |
| error("failed to read hashes: %s", strerror(errno)); |
| return -1; |
| } |
| |
| if (!check_block_hash(buffer.data(), data)) { |
| /* try to correct */ |
| if (!ecc_read_hashes(const_cast<fec_handle *>(f), |
| hash_offset + j * padded_digest_length_, |
| buffer.data(), |
| data_offset + j * FEC_BLOCKSIZE, data) || |
| !check_block_hash(buffer.data(), data)) { |
| error("invalid hash tree: hash_offset %" PRIu64 |
| ", " |
| "data_offset %" PRIu64 ", block %u", |
| hash_offset, data_offset, j); |
| return -1; |
| } |
| |
| /* update the corrected blocks to the file if we are in r/w |
| mode */ |
| if (f->mode & O_RDWR) { |
| if (!raw_pwrite(f->fd, buffer.data(), padded_digest_length_, |
| hash_offset + j * padded_digest_length_) || |
| !raw_pwrite(f->fd, data, FEC_BLOCKSIZE, |
| data_offset + j * FEC_BLOCKSIZE)) { |
| error("failed to write hashes: %s", strerror(errno)); |
| return -1; |
| } |
| } |
| } |
| |
| if (blocks == hash_data_blocks) { |
| std::copy(data, data + FEC_BLOCKSIZE, |
| data_hashes.begin() + j * FEC_BLOCKSIZE); |
| } |
| } |
| |
| hash_offset = data_offset; |
| data_offset += blocks * FEC_BLOCKSIZE; |
| } |
| |
| debug("valid"); |
| |
| this->hash_data = std::move(data_hashes); |
| |
| std::vector<uint8_t> zero_block(FEC_BLOCKSIZE, 0); |
| zero_hash.resize(padded_digest_length_, 0); |
| if (get_hash(zero_block.data(), zero_hash.data()) == -1) { |
| error("failed to hash"); |
| return -1; |
| } |
| return 0; |
| } |
| |
| /* reads, corrects and parses the verity table, validates parameters, and if |
| `f->flags' does not have `FEC_VERITY_DISABLE' set, calls `verify_tree' to |
| load and validate the hash tree */ |
| static int parse_table(fec_handle *f, uint64_t offset, uint32_t size, bool useecc) |
| { |
| check(f); |
| check(size >= VERITY_MIN_TABLE_SIZE); |
| check(size <= VERITY_MAX_TABLE_SIZE); |
| |
| debug("offset = %" PRIu64 ", size = %u", offset, size); |
| |
| std::string table(size, 0); |
| |
| if (!useecc) { |
| if (!raw_pread(f->fd, const_cast<char *>(table.data()), size, offset)) { |
| error("failed to read verity table: %s", strerror(errno)); |
| return -1; |
| } |
| } else if (fec_pread(f, const_cast<char *>(table.data()), size, offset) != |
| (ssize_t)size) { |
| error("failed to ecc read verity table: %s", strerror(errno)); |
| return -1; |
| } |
| |
| debug("verity table: '%s'", table.c_str()); |
| |
| int i = 0; |
| std::vector<uint8_t> salt; |
| uint8_t root[SHA256_DIGEST_LENGTH]; |
| uint64_t hash_start = 0; |
| uint64_t data_blocks = 0; |
| |
| auto tokens = android::base::Split(table, " "); |
| |
| for (const auto& token : tokens) { |
| switch (i++) { |
| case 0: /* version */ |
| if (token != stringify(VERITY_TABLE_VERSION)) { |
| error("unsupported verity table version: %s", token.c_str()); |
| return -1; |
| } |
| break; |
| case 3: /* data_block_size */ |
| case 4: /* hash_block_size */ |
| /* assume 4 KiB block sizes for everything */ |
| if (token != stringify(FEC_BLOCKSIZE)) { |
| error("unsupported verity block size: %s", token.c_str()); |
| return -1; |
| } |
| break; |
| case 5: /* num_data_blocks */ |
| if (parse_uint64(token.c_str(), f->data_size / FEC_BLOCKSIZE, |
| &data_blocks) == -1) { |
| error("invalid number of verity data blocks: %s", |
| token.c_str()); |
| return -1; |
| } |
| break; |
| case 6: /* hash_start_block */ |
| if (parse_uint64(token.c_str(), f->data_size / FEC_BLOCKSIZE, |
| &hash_start) == -1) { |
| error("invalid verity hash start block: %s", token.c_str()); |
| return -1; |
| } |
| |
| hash_start *= FEC_BLOCKSIZE; |
| break; |
| case 7: /* algorithm */ |
| if (token != "sha256") { |
| error("unsupported verity hash algorithm: %s", token.c_str()); |
| return -1; |
| } |
| break; |
| case 8: /* digest */ |
| if (parse_hex(root, sizeof(root), token.c_str()) == -1) { |
| error("invalid verity root hash: %s", token.c_str()); |
| return -1; |
| } |
| break; |
| case 9: /* salt */ |
| { |
| uint32_t salt_size = token.size(); |
| check(salt_size % 2 == 0); |
| salt_size /= 2; |
| |
| salt.resize(salt_size, 0); |
| |
| if (parse_hex(salt.data(), salt_size, token.c_str()) == -1) { |
| error("invalid verity salt: %s", token.c_str()); |
| return -1; |
| } |
| break; |
| } |
| default: |
| break; |
| } |
| } |
| |
| if (i < VERITY_TABLE_ARGS) { |
| error("not enough arguments in verity table: %d; expected at least " |
| stringify(VERITY_TABLE_ARGS), i); |
| return -1; |
| } |
| |
| check(hash_start < f->data_size); |
| |
| verity_info *v = &f->verity; |
| if (v->metadata_start < hash_start) { |
| check(data_blocks == v->metadata_start / FEC_BLOCKSIZE); |
| } else { |
| check(data_blocks == hash_start / FEC_BLOCKSIZE); |
| } |
| |
| v->table = std::move(table); |
| |
| v->hashtree.initialize(hash_start, data_blocks, salt, NID_sha256); |
| if (!(f->flags & FEC_VERITY_DISABLE)) { |
| if (v->hashtree.verify_tree(f, root) == -1) { |
| return -1; |
| } |
| |
| check(!v->hashtree.hash_data.empty()); |
| check(!v->hashtree.zero_hash.empty()); |
| } |
| |
| return 0; |
| } |
| |
| /* rewrites verity metadata block using error corrected data in `f->verity' */ |
| static int rewrite_metadata(fec_handle *f, uint64_t offset) |
| { |
| check(f); |
| check(f->data_size > VERITY_METADATA_SIZE); |
| check(offset <= f->data_size - VERITY_METADATA_SIZE); |
| |
| std::unique_ptr<uint8_t[]> metadata( |
| new (std::nothrow) uint8_t[VERITY_METADATA_SIZE]); |
| |
| if (!metadata) { |
| errno = ENOMEM; |
| return -1; |
| } |
| |
| memset(metadata.get(), 0, VERITY_METADATA_SIZE); |
| |
| verity_info *v = &f->verity; |
| memcpy(metadata.get(), &v->header, sizeof(v->header)); |
| |
| check(!v->table.empty()); |
| size_t len = v->table.size(); |
| |
| check(sizeof(v->header) + len <= VERITY_METADATA_SIZE); |
| memcpy(metadata.get() + sizeof(v->header), v->table.data(), len); |
| |
| return raw_pwrite(f->fd, metadata.get(), VERITY_METADATA_SIZE, offset); |
| } |
| |
| static int validate_header(const fec_handle *f, const verity_header *header, |
| uint64_t offset) |
| { |
| check(f); |
| check(header); |
| |
| if (header->magic != VERITY_MAGIC && |
| header->magic != VERITY_MAGIC_DISABLE) { |
| return -1; |
| } |
| |
| if (header->version != VERITY_VERSION) { |
| error("unsupported verity version %u", header->version); |
| return -1; |
| } |
| |
| if (header->length < VERITY_MIN_TABLE_SIZE || |
| header->length > VERITY_MAX_TABLE_SIZE) { |
| error("invalid verity table size: %u; expected [" |
| stringify(VERITY_MIN_TABLE_SIZE) ", " |
| stringify(VERITY_MAX_TABLE_SIZE) ")", header->length); |
| return -1; |
| } |
| |
| /* signature is skipped, because for our purposes it won't matter from |
| where the data originates; the caller of the library is responsible |
| for signature verification */ |
| |
| if (offset > UINT64_MAX - header->length) { |
| error("invalid verity table length: %u", header->length); |
| return -1; |
| } else if (offset + header->length >= f->data_size) { |
| error("invalid verity table length: %u", header->length); |
| return -1; |
| } |
| |
| return 0; |
| } |
| |
| /* attempts to read verity metadata from `f->fd' position `offset'; if in r/w |
| mode, rewrites the metadata if it had errors */ |
| int verity_parse_header(fec_handle *f, uint64_t offset) |
| { |
| check(f); |
| check(f->data_size > VERITY_METADATA_SIZE); |
| |
| if (offset > f->data_size - VERITY_METADATA_SIZE) { |
| debug("failed to read verity header: offset %" PRIu64 " is too far", |
| offset); |
| return -1; |
| } |
| |
| verity_info *v = &f->verity; |
| uint64_t errors = f->errors; |
| |
| if (!raw_pread(f->fd, &v->header, sizeof(v->header), offset)) { |
| error("failed to read verity header: %s", strerror(errno)); |
| return -1; |
| } |
| |
| /* use raw data to check for the alternative magic, because it will |
| be error corrected to VERITY_MAGIC otherwise */ |
| if (v->header.magic == VERITY_MAGIC_DISABLE) { |
| /* this value is not used by us, but can be used by a caller to |
| decide whether dm-verity should be enabled */ |
| v->disabled = true; |
| } |
| |
| if (fec_pread(f, &v->ecc_header, sizeof(v->ecc_header), offset) != |
| sizeof(v->ecc_header)) { |
| warn("failed to read verity header: %s", strerror(errno)); |
| return -1; |
| } |
| |
| if (validate_header(f, &v->header, offset)) { |
| /* raw verity header is invalid; this could be due to corruption, or |
| due to missing verity metadata */ |
| |
| if (validate_header(f, &v->ecc_header, offset)) { |
| return -1; /* either way, we cannot recover */ |
| } |
| |
| /* report mismatching fields */ |
| if (!v->disabled && v->header.magic != v->ecc_header.magic) { |
| warn("corrected verity header magic"); |
| v->header.magic = v->ecc_header.magic; |
| } |
| |
| if (v->header.version != v->ecc_header.version) { |
| warn("corrected verity header version"); |
| v->header.version = v->ecc_header.version; |
| } |
| |
| if (v->header.length != v->ecc_header.length) { |
| warn("corrected verity header length"); |
| v->header.length = v->ecc_header.length; |
| } |
| |
| if (memcmp(v->header.signature, v->ecc_header.signature, |
| sizeof(v->header.signature))) { |
| warn("corrected verity header signature"); |
| /* we have no way of knowing which signature is correct, if either |
| of them is */ |
| } |
| } |
| |
| v->metadata_start = offset; |
| |
| if (parse_table(f, offset + sizeof(v->header), v->header.length, |
| false) == -1 && |
| parse_table(f, offset + sizeof(v->header), v->header.length, |
| true) == -1) { |
| return -1; |
| } |
| |
| /* if we corrected something while parsing metadata and we are in r/w |
| mode, rewrite the corrected metadata */ |
| if (f->mode & O_RDWR && f->errors > errors && |
| rewrite_metadata(f, offset) < 0) { |
| warn("failed to rewrite verity metadata: %s", strerror(errno)); |
| } |
| |
| if (v->metadata_start < v->hashtree.hash_start) { |
| f->data_size = v->metadata_start; |
| } else { |
| f->data_size = v->hashtree.hash_start; |
| } |
| |
| return 0; |
| } |
| |
| int fec_verity_set_status(struct fec_handle *f, bool enabled) |
| { |
| check(f); |
| |
| if (!(f->mode & O_RDWR)) { |
| error("cannot update verity magic: read-only handle"); |
| errno = EBADF; |
| return -1; |
| } |
| |
| verity_info *v = &f->verity; |
| |
| if (!v->metadata_start) { |
| error("cannot update verity magic: no metadata found"); |
| errno = EINVAL; |
| return -1; |
| } |
| |
| if (v->disabled == !enabled) { |
| return 0; /* nothing to do */ |
| } |
| |
| uint32_t magic = enabled ? VERITY_MAGIC : VERITY_MAGIC_DISABLE; |
| |
| if (!raw_pwrite(f->fd, &magic, sizeof(magic), v->metadata_start)) { |
| error("failed to update verity magic to %08x: %s", magic, |
| strerror(errno)); |
| return -1; |
| } |
| |
| warn("updated verity magic to %08x (%s)", magic, |
| enabled ? "enabled" : "disabled"); |
| v->disabled = !enabled; |
| |
| return 0; |
| } |