/* * Copyright (C) 2015 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include #include #include #include #include #include #include #include "fec_private.h" /* converts a hex nibble into an int */ static inline int hextobin(char c) { if (c >= '0' && c <= '9') { return c - '0'; } else if (c >= 'a' && c <= 'f') { return c - 'a' + 10; } else { errno = EINVAL; return -1; } } /* converts a hex string `src' of `size' characters to binary and copies the the result into `dst' */ static int parse_hex(uint8_t *dst, uint32_t size, const char *src) { int l, h; check(dst); check(src); check(2 * size == strlen(src)); while (size) { h = hextobin(tolower(*src++)); l = hextobin(tolower(*src++)); check(l >= 0); check(h >= 0); *dst++ = (h << 4) | l; --size; } return 0; } /* parses a 64-bit unsigned integer from string `src' into `dst' and if `maxval' is >0, checks that `dst' <= `maxval' */ static int parse_uint64(const char *src, uint64_t maxval, uint64_t *dst) { char *end; unsigned long long int value; check(src); check(dst); errno = 0; value = strtoull(src, &end, 0); if (*src == '\0' || *end != '\0' || (errno == ERANGE && value == ULLONG_MAX)) { errno = EINVAL; return -1; } if (maxval && value > maxval) { errno = EINVAL; return -1; } *dst = (uint64_t)value; return 0; } /* computes the size of verity hash tree for `file_size' bytes and returns the number of hash tree levels in `verity_levels,' and the number of hashes per level in `level_hashes', if the parameters are non-NULL */ uint64_t verity_get_size(uint64_t file_size, uint32_t *verity_levels, uint32_t *level_hashes, uint32_t padded_digest_size) { // we assume a known metadata size, 4 KiB block size, and SHA-256 or SHA1 to // avoid relying on disk content. uint32_t level = 0; uint64_t total = 0; uint64_t hashes = file_size / FEC_BLOCKSIZE; do { if (level_hashes) { level_hashes[level] = hashes; } hashes = fec_div_round_up(hashes * padded_digest_size, FEC_BLOCKSIZE); total += hashes; ++level; } while (hashes > 1); if (verity_levels) { *verity_levels = level; } return total * FEC_BLOCKSIZE; } int hashtree_info::get_hash(const uint8_t *block, uint8_t *hash) { auto md = EVP_get_digestbynid(nid_); check(md); auto mdctx = EVP_MD_CTX_new(); check(mdctx); EVP_DigestInit_ex(mdctx, md, nullptr); EVP_DigestUpdate(mdctx, salt.data(), salt.size()); EVP_DigestUpdate(mdctx, block, FEC_BLOCKSIZE); unsigned int hash_size; EVP_DigestFinal_ex(mdctx, hash, &hash_size); EVP_MD_CTX_free(mdctx); check(hash_size == digest_length_); return 0; } int hashtree_info::initialize(uint64_t hash_start, uint64_t data_blocks, const std::vector &salt, int nid) { check(nid == NID_sha256 || nid == NID_sha1); this->hash_start = hash_start; this->data_blocks = data_blocks; this->salt = salt; this->nid_ = nid; digest_length_ = nid == NID_sha1 ? SHA_DIGEST_LENGTH : SHA256_DIGEST_LENGTH; // The padded digest size for both sha256 and sha1 are 256 bytes. padded_digest_length_ = SHA256_DIGEST_LENGTH; return 0; } bool hashtree_info::check_block_hash(const uint8_t *expected, const uint8_t *block) { check(block); std::vector hash(digest_length_, 0); if (unlikely(get_hash(block, hash.data()) == -1)) { error("failed to hash"); return false; } check(expected); return !memcmp(expected, hash.data(), digest_length_); } bool hashtree_info::check_block_hash_with_index(uint64_t index, const uint8_t *block) { check(index < data_blocks); const uint8_t *expected = &hash_data[index * padded_digest_length_]; return check_block_hash(expected, block); } // Reads the hash and the corresponding data block using error correction, if // available. bool hashtree_info::ecc_read_hashes(fec_handle *f, uint64_t hash_offset, uint8_t *hash, uint64_t data_offset, uint8_t *data) { check(f); if (hash && fec_pread(f, hash, digest_length_, hash_offset) != digest_length_) { error("failed to read hash tree: offset %" PRIu64 ": %s", hash_offset, strerror(errno)); return false; } check(data); if (fec_pread(f, data, FEC_BLOCKSIZE, data_offset) != FEC_BLOCKSIZE) { error("failed to read hash tree: data_offset %" PRIu64 ": %s", data_offset, strerror(errno)); return false; } return true; } int hashtree_info::verify_tree(const fec_handle *f, const uint8_t *root) { check(f); check(root); uint8_t data[FEC_BLOCKSIZE]; uint32_t levels = 0; /* calculate the size and the number of levels in the hash tree */ uint64_t hash_size = verity_get_size(data_blocks * FEC_BLOCKSIZE, &levels, NULL, padded_digest_length_); check(hash_start < UINT64_MAX - hash_size); check(hash_start + hash_size <= f->data_size); uint64_t hash_offset = hash_start; uint64_t data_offset = hash_offset + FEC_BLOCKSIZE; /* validate the root hash */ if (!raw_pread(f->fd, data, FEC_BLOCKSIZE, hash_offset) || !check_block_hash(root, data)) { /* try to correct */ if (!ecc_read_hashes(const_cast(f), 0, nullptr, hash_offset, data) || !check_block_hash(root, data)) { error("root hash invalid"); return -1; } else if (f->mode & O_RDWR && !raw_pwrite(f->fd, data, FEC_BLOCKSIZE, hash_offset)) { error("failed to rewrite the root block: %s", strerror(errno)); return -1; } } debug("root hash valid"); /* calculate the number of hashes on each level */ uint32_t hashes[levels]; verity_get_size(data_blocks * FEC_BLOCKSIZE, NULL, hashes, padded_digest_length_); uint64_t hash_data_offset = data_offset; uint32_t hash_data_blocks = 0; /* calculate the size and offset for the data hashes */ for (uint32_t i = 1; i < levels; ++i) { uint32_t blocks = hashes[levels - i]; debug("%u hash blocks on level %u", blocks, levels - i); hash_data_offset = data_offset; hash_data_blocks = blocks; data_offset += blocks * FEC_BLOCKSIZE; } check(hash_data_blocks); check(hash_data_blocks <= hash_size / FEC_BLOCKSIZE); check(hash_data_offset); check(hash_data_offset <= UINT64_MAX - (hash_data_blocks * FEC_BLOCKSIZE)); check(hash_data_offset < f->data_size); check(hash_data_offset + hash_data_blocks * FEC_BLOCKSIZE <= f->data_size); /* copy data hashes to memory in case they are corrupted, so we don't have to correct them every time they are needed */ std::vector data_hashes(hash_data_blocks * FEC_BLOCKSIZE, 0); /* validate the rest of the hash tree */ data_offset = hash_offset + FEC_BLOCKSIZE; std::vector buffer(padded_digest_length_, 0); for (uint32_t i = 1; i < levels; ++i) { uint32_t blocks = hashes[levels - i]; for (uint32_t j = 0; j < blocks; ++j) { /* ecc reads are very I/O intensive, so read raw hash tree and do error correcting only if it doesn't validate */ if (!raw_pread(f->fd, buffer.data(), padded_digest_length_, hash_offset + j * padded_digest_length_) || !raw_pread(f->fd, data, FEC_BLOCKSIZE, data_offset + j * FEC_BLOCKSIZE)) { error("failed to read hashes: %s", strerror(errno)); return -1; } if (!check_block_hash(buffer.data(), data)) { /* try to correct */ if (!ecc_read_hashes(const_cast(f), hash_offset + j * padded_digest_length_, buffer.data(), data_offset + j * FEC_BLOCKSIZE, data) || !check_block_hash(buffer.data(), data)) { error("invalid hash tree: hash_offset %" PRIu64 ", " "data_offset %" PRIu64 ", block %u", hash_offset, data_offset, j); return -1; } /* update the corrected blocks to the file if we are in r/w mode */ if (f->mode & O_RDWR) { if (!raw_pwrite(f->fd, buffer.data(), padded_digest_length_, hash_offset + j * padded_digest_length_) || !raw_pwrite(f->fd, data, FEC_BLOCKSIZE, data_offset + j * FEC_BLOCKSIZE)) { error("failed to write hashes: %s", strerror(errno)); return -1; } } } if (blocks == hash_data_blocks) { std::copy(data, data + FEC_BLOCKSIZE, data_hashes.begin() + j * FEC_BLOCKSIZE); } } hash_offset = data_offset; data_offset += blocks * FEC_BLOCKSIZE; } debug("valid"); this->hash_data = std::move(data_hashes); std::vector zero_block(FEC_BLOCKSIZE, 0); zero_hash.resize(padded_digest_length_, 0); if (get_hash(zero_block.data(), zero_hash.data()) == -1) { error("failed to hash"); return -1; } return 0; } /* reads, corrects and parses the verity table, validates parameters, and if `f->flags' does not have `FEC_VERITY_DISABLE' set, calls `verify_tree' to load and validate the hash tree */ static int parse_table(fec_handle *f, uint64_t offset, uint32_t size, bool useecc) { check(f); check(size >= VERITY_MIN_TABLE_SIZE); check(size <= VERITY_MAX_TABLE_SIZE); debug("offset = %" PRIu64 ", size = %u", offset, size); std::string table(size, 0); if (!useecc) { if (!raw_pread(f->fd, const_cast(table.data()), size, offset)) { error("failed to read verity table: %s", strerror(errno)); return -1; } } else if (fec_pread(f, const_cast(table.data()), size, offset) != (ssize_t)size) { error("failed to ecc read verity table: %s", strerror(errno)); return -1; } debug("verity table: '%s'", table.c_str()); int i = 0; std::vector salt; uint8_t root[SHA256_DIGEST_LENGTH]; uint64_t hash_start = 0; uint64_t data_blocks = 0; auto tokens = android::base::Split(table, " "); for (const auto& token : tokens) { switch (i++) { case 0: /* version */ if (token != stringify(VERITY_TABLE_VERSION)) { error("unsupported verity table version: %s", token.c_str()); return -1; } break; case 3: /* data_block_size */ case 4: /* hash_block_size */ /* assume 4 KiB block sizes for everything */ if (token != stringify(FEC_BLOCKSIZE)) { error("unsupported verity block size: %s", token.c_str()); return -1; } break; case 5: /* num_data_blocks */ if (parse_uint64(token.c_str(), f->data_size / FEC_BLOCKSIZE, &data_blocks) == -1) { error("invalid number of verity data blocks: %s", token.c_str()); return -1; } break; case 6: /* hash_start_block */ if (parse_uint64(token.c_str(), f->data_size / FEC_BLOCKSIZE, &hash_start) == -1) { error("invalid verity hash start block: %s", token.c_str()); return -1; } hash_start *= FEC_BLOCKSIZE; break; case 7: /* algorithm */ if (token != "sha256") { error("unsupported verity hash algorithm: %s", token.c_str()); return -1; } break; case 8: /* digest */ if (parse_hex(root, sizeof(root), token.c_str()) == -1) { error("invalid verity root hash: %s", token.c_str()); return -1; } break; case 9: /* salt */ { uint32_t salt_size = token.size(); check(salt_size % 2 == 0); salt_size /= 2; salt.resize(salt_size, 0); if (parse_hex(salt.data(), salt_size, token.c_str()) == -1) { error("invalid verity salt: %s", token.c_str()); return -1; } break; } default: break; } } if (i < VERITY_TABLE_ARGS) { error("not enough arguments in verity table: %d; expected at least " stringify(VERITY_TABLE_ARGS), i); return -1; } check(hash_start < f->data_size); verity_info *v = &f->verity; if (v->metadata_start < hash_start) { check(data_blocks == v->metadata_start / FEC_BLOCKSIZE); } else { check(data_blocks == hash_start / FEC_BLOCKSIZE); } v->table = std::move(table); v->hashtree.initialize(hash_start, data_blocks, salt, NID_sha256); if (!(f->flags & FEC_VERITY_DISABLE)) { if (v->hashtree.verify_tree(f, root) == -1) { return -1; } check(!v->hashtree.hash_data.empty()); check(!v->hashtree.zero_hash.empty()); } return 0; } /* rewrites verity metadata block using error corrected data in `f->verity' */ static int rewrite_metadata(fec_handle *f, uint64_t offset) { check(f); check(f->data_size > VERITY_METADATA_SIZE); check(offset <= f->data_size - VERITY_METADATA_SIZE); std::unique_ptr metadata( new (std::nothrow) uint8_t[VERITY_METADATA_SIZE]); if (!metadata) { errno = ENOMEM; return -1; } memset(metadata.get(), 0, VERITY_METADATA_SIZE); verity_info *v = &f->verity; memcpy(metadata.get(), &v->header, sizeof(v->header)); check(!v->table.empty()); size_t len = v->table.size(); check(sizeof(v->header) + len <= VERITY_METADATA_SIZE); memcpy(metadata.get() + sizeof(v->header), v->table.data(), len); return raw_pwrite(f->fd, metadata.get(), VERITY_METADATA_SIZE, offset); } static int validate_header(const fec_handle *f, const verity_header *header, uint64_t offset) { check(f); check(header); if (header->magic != VERITY_MAGIC && header->magic != VERITY_MAGIC_DISABLE) { return -1; } if (header->version != VERITY_VERSION) { error("unsupported verity version %u", header->version); return -1; } if (header->length < VERITY_MIN_TABLE_SIZE || header->length > VERITY_MAX_TABLE_SIZE) { error("invalid verity table size: %u; expected [" stringify(VERITY_MIN_TABLE_SIZE) ", " stringify(VERITY_MAX_TABLE_SIZE) ")", header->length); return -1; } /* signature is skipped, because for our purposes it won't matter from where the data originates; the caller of the library is responsible for signature verification */ if (offset > UINT64_MAX - header->length) { error("invalid verity table length: %u", header->length); return -1; } else if (offset + header->length >= f->data_size) { error("invalid verity table length: %u", header->length); return -1; } return 0; } /* attempts to read verity metadata from `f->fd' position `offset'; if in r/w mode, rewrites the metadata if it had errors */ int verity_parse_header(fec_handle *f, uint64_t offset) { check(f); check(f->data_size > VERITY_METADATA_SIZE); if (offset > f->data_size - VERITY_METADATA_SIZE) { debug("failed to read verity header: offset %" PRIu64 " is too far", offset); return -1; } verity_info *v = &f->verity; uint64_t errors = f->errors; if (!raw_pread(f->fd, &v->header, sizeof(v->header), offset)) { error("failed to read verity header: %s", strerror(errno)); return -1; } /* use raw data to check for the alternative magic, because it will be error corrected to VERITY_MAGIC otherwise */ if (v->header.magic == VERITY_MAGIC_DISABLE) { /* this value is not used by us, but can be used by a caller to decide whether dm-verity should be enabled */ v->disabled = true; } if (fec_pread(f, &v->ecc_header, sizeof(v->ecc_header), offset) != sizeof(v->ecc_header)) { warn("failed to read verity header: %s", strerror(errno)); return -1; } if (validate_header(f, &v->header, offset)) { /* raw verity header is invalid; this could be due to corruption, or due to missing verity metadata */ if (validate_header(f, &v->ecc_header, offset)) { return -1; /* either way, we cannot recover */ } /* report mismatching fields */ if (!v->disabled && v->header.magic != v->ecc_header.magic) { warn("corrected verity header magic"); v->header.magic = v->ecc_header.magic; } if (v->header.version != v->ecc_header.version) { warn("corrected verity header version"); v->header.version = v->ecc_header.version; } if (v->header.length != v->ecc_header.length) { warn("corrected verity header length"); v->header.length = v->ecc_header.length; } if (memcmp(v->header.signature, v->ecc_header.signature, sizeof(v->header.signature))) { warn("corrected verity header signature"); /* we have no way of knowing which signature is correct, if either of them is */ } } v->metadata_start = offset; if (parse_table(f, offset + sizeof(v->header), v->header.length, false) == -1 && parse_table(f, offset + sizeof(v->header), v->header.length, true) == -1) { return -1; } /* if we corrected something while parsing metadata and we are in r/w mode, rewrite the corrected metadata */ if (f->mode & O_RDWR && f->errors > errors && rewrite_metadata(f, offset) < 0) { warn("failed to rewrite verity metadata: %s", strerror(errno)); } if (v->metadata_start < v->hashtree.hash_start) { f->data_size = v->metadata_start; } else { f->data_size = v->hashtree.hash_start; } return 0; } int fec_verity_set_status(struct fec_handle *f, bool enabled) { check(f); if (!(f->mode & O_RDWR)) { error("cannot update verity magic: read-only handle"); errno = EBADF; return -1; } verity_info *v = &f->verity; if (!v->metadata_start) { error("cannot update verity magic: no metadata found"); errno = EINVAL; return -1; } if (v->disabled == !enabled) { return 0; /* nothing to do */ } uint32_t magic = enabled ? VERITY_MAGIC : VERITY_MAGIC_DISABLE; if (!raw_pwrite(f->fd, &magic, sizeof(magic), v->metadata_start)) { error("failed to update verity magic to %08x: %s", magic, strerror(errno)); return -1; } warn("updated verity magic to %08x (%s)", magic, enabled ? "enabled" : "disabled"); v->disabled = !enabled; return 0; }