1516 lines
41 KiB
C
1516 lines
41 KiB
C
/* Copyright 2016 The Chromium OS Authors. All rights reserved.
|
|
* Use of this source code is governed by a BSD-style license that can be
|
|
* found in the LICENSE file.
|
|
*
|
|
* Test Cr-50 Non-Voltatile memory module
|
|
*/
|
|
|
|
#include "nvmem_test.h"
|
|
|
|
#include "common.h"
|
|
#include "console.h"
|
|
#include "crc.h"
|
|
#include "flash.h"
|
|
#include "flash_log.h"
|
|
#include "new_nvmem.h"
|
|
#include "nvmem.h"
|
|
#include "printf.h"
|
|
#include "shared_mem.h"
|
|
#include "task.h"
|
|
#include "test_util.h"
|
|
#include "timer.h"
|
|
#include "util.h"
|
|
|
|
#define WRITE_SEGMENT_LEN 200
|
|
#define WRITE_READ_SEGMENTS 4
|
|
|
|
enum test_failure_mode failure_mode;
|
|
|
|
static const uint8_t legacy_nvmem_image[] = {
|
|
#include "legacy_nvmem_dump.h"
|
|
};
|
|
|
|
BUILD_ASSERT(sizeof(legacy_nvmem_image) == NVMEM_PARTITION_SIZE);
|
|
|
|
static uint8_t write_buffer[NVMEM_PARTITION_SIZE];
|
|
static int flash_write_fail;
|
|
|
|
struct nvmem_test_result {
|
|
int var_count;
|
|
int reserved_obj_count;
|
|
int evictable_obj_count;
|
|
int deleted_obj_count;
|
|
int delimiter_count;
|
|
int unexpected_count;
|
|
size_t valid_data_size;
|
|
size_t erased_data_size;
|
|
};
|
|
|
|
static struct nvmem_test_result test_result;
|
|
|
|
int app_cipher(const void *salt_p, void *out_p, const void *in_p, size_t size)
|
|
{
|
|
|
|
const uint8_t *in = in_p;
|
|
uint8_t *out = out_p;
|
|
const uint8_t *salt = salt_p;
|
|
size_t i;
|
|
|
|
for (i = 0; i < size; i++)
|
|
out[i] = in[i] ^ salt[i % CIPHER_SALT_SIZE];
|
|
|
|
return 1;
|
|
}
|
|
|
|
void app_compute_hash(uint8_t *p_buf, size_t num_bytes,
|
|
uint8_t *p_hash, size_t hash_bytes)
|
|
{
|
|
uint32_t crc;
|
|
uint32_t *p_data;
|
|
int n;
|
|
size_t tail_size;
|
|
|
|
crc32_init();
|
|
/* Assuming here that buffer is 4 byte aligned. */
|
|
p_data = (uint32_t *)p_buf;
|
|
for (n = 0; n < num_bytes / 4; n++)
|
|
crc32_hash32(*p_data++);
|
|
|
|
tail_size = num_bytes % 4;
|
|
if (tail_size) {
|
|
uint32_t tail;
|
|
|
|
tail = 0;
|
|
memcpy(&tail, p_data, tail_size);
|
|
crc32_hash32(tail);
|
|
}
|
|
|
|
/*
|
|
* Crc32 of 0xffffffff is 0xffffffff. Let's spike the results to avoid
|
|
* this unfortunate Crc32 property.
|
|
*/
|
|
crc = crc32_result() ^ 0x55555555;
|
|
|
|
for (n = 0; n < hash_bytes; n += sizeof(crc)) {
|
|
size_t copy_bytes = MIN(sizeof(crc), hash_bytes - n);
|
|
|
|
memcpy(p_hash + n, &crc, copy_bytes);
|
|
}
|
|
}
|
|
|
|
int crypto_enabled(void)
|
|
{
|
|
return 1;
|
|
}
|
|
|
|
/* Used to allow/prevent Flash erase/write operations */
|
|
int flash_pre_op(void)
|
|
{
|
|
return flash_write_fail ? EC_ERROR_UNKNOWN : EC_SUCCESS;
|
|
}
|
|
|
|
static void dump_nvmem_state(const char *title,
|
|
const struct nvmem_test_result *tr)
|
|
{
|
|
ccprintf("\n%s:\n", title);
|
|
ccprintf("var_count: %d\n", tr->var_count);
|
|
ccprintf("reserved_obj_count: %d\n", tr->reserved_obj_count);
|
|
ccprintf("evictable_obj_count: %d\n", tr->evictable_obj_count);
|
|
ccprintf("deleted_obj_count: %d\n", tr->deleted_obj_count);
|
|
ccprintf("deimiter_count: %d\n", tr->delimiter_count);
|
|
ccprintf("unexpected_count: %d\n", tr->unexpected_count);
|
|
ccprintf("valid_data_size: %d\n", tr->valid_data_size);
|
|
ccprintf("erased_data_size: %d\n\n", tr->erased_data_size);
|
|
}
|
|
|
|
static void wipe_out_nvmem_cache(void)
|
|
{
|
|
memset(nvmem_cache_base(NVMEM_TPM), 0, nvmem_user_sizes[NVMEM_TPM]);
|
|
}
|
|
|
|
static int prepare_nvmem_contents(void)
|
|
{
|
|
struct nvmem_tag *tag;
|
|
|
|
memcpy(write_buffer, legacy_nvmem_image, sizeof(write_buffer));
|
|
tag = (struct nvmem_tag *)write_buffer;
|
|
|
|
app_compute_hash(tag->padding, NVMEM_PARTITION_SIZE - NVMEM_SHA_SIZE,
|
|
tag->sha, sizeof(tag->sha));
|
|
app_cipher(tag->sha, tag + 1, tag + 1,
|
|
NVMEM_PARTITION_SIZE - sizeof(struct nvmem_tag));
|
|
|
|
return flash_physical_write(CONFIG_FLASH_NVMEM_BASE_A -
|
|
CONFIG_PROGRAM_MEMORY_BASE,
|
|
sizeof(write_buffer), write_buffer);
|
|
}
|
|
|
|
static int iterate_over_flash(void)
|
|
{
|
|
enum ec_error_list rv;
|
|
struct nn_container *ch;
|
|
struct access_tracker at = {};
|
|
uint8_t buf[CONFIG_FLASH_BANK_SIZE];
|
|
|
|
memset(&test_result, 0, sizeof(test_result));
|
|
ch = (struct nn_container *)buf;
|
|
|
|
while ((rv = get_next_object(&at, ch, 1)) == EC_SUCCESS)
|
|
switch (ch->container_type) {
|
|
case NN_OBJ_OLD_COPY:
|
|
if (ch->container_type_copy == NN_OBJ_TRANSACTION_DEL) {
|
|
test_result.delimiter_count++;
|
|
} else {
|
|
test_result.deleted_obj_count++;
|
|
test_result.erased_data_size += ch->size;
|
|
}
|
|
break;
|
|
|
|
case NN_OBJ_TUPLE:
|
|
test_result.var_count++;
|
|
test_result.valid_data_size += ch->size;
|
|
break;
|
|
|
|
case NN_OBJ_TPM_RESERVED:
|
|
test_result.reserved_obj_count++;
|
|
test_result.valid_data_size += ch->size;
|
|
break;
|
|
|
|
case NN_OBJ_TPM_EVICTABLE:
|
|
test_result.evictable_obj_count++;
|
|
test_result.valid_data_size += ch->size;
|
|
break;
|
|
|
|
case NN_OBJ_TRANSACTION_DEL:
|
|
test_result.delimiter_count++;
|
|
break;
|
|
default:
|
|
test_result.unexpected_count++;
|
|
break;
|
|
}
|
|
|
|
if (rv != EC_ERROR_MEMORY_ALLOCATION) {
|
|
ccprintf("\n%s:%d - unexpected return value %d\n", __func__,
|
|
__LINE__, rv);
|
|
return rv;
|
|
}
|
|
|
|
/* Verify that there is a delimiter at the top of the flash. */
|
|
if (at.mt.data_offset > sizeof(*at.mt.ph)) {
|
|
if ((at.mt.ph == at.dt.ph) &&
|
|
(((at.mt.data_offset - sizeof(struct nn_container))) ==
|
|
at.dt.data_offset)) {
|
|
return EC_SUCCESS;
|
|
}
|
|
} else {
|
|
if ((at.dt.ph == list_element_to_ph(at.list_index)) &&
|
|
(at.dt.data_offset ==
|
|
(CONFIG_FLASH_BANK_SIZE - sizeof(struct nn_container)))) {
|
|
ccprintf("%s:%d edge delimiter case OK\n", __func__,
|
|
__LINE__);
|
|
return EC_SUCCESS;
|
|
}
|
|
}
|
|
ccprintf("%s:%d bad delimiter location: ph %p, "
|
|
"dt.ph %p, offset %d, delim offset %d\n",
|
|
__func__, __LINE__, at.mt.ph, at.dt.ph, at.mt.data_offset,
|
|
at.dt.data_offset);
|
|
|
|
return EC_ERROR_INVAL;
|
|
}
|
|
|
|
static void *page_to_flash_addr(int page_num)
|
|
{
|
|
uint32_t base_offset = CONFIG_FLASH_NEW_NVMEM_BASE_A;
|
|
|
|
if (page_num > NEW_NVMEM_TOTAL_PAGES)
|
|
return NULL;
|
|
|
|
if (page_num >= (NEW_NVMEM_TOTAL_PAGES / 2)) {
|
|
page_num -= (NEW_NVMEM_TOTAL_PAGES / 2);
|
|
base_offset = CONFIG_FLASH_NEW_NVMEM_BASE_B;
|
|
}
|
|
|
|
return (void *)((uintptr_t)base_offset +
|
|
page_num * CONFIG_FLASH_BANK_SIZE);
|
|
}
|
|
|
|
static int post_init_from_scratch(uint8_t flash_value)
|
|
{
|
|
int i;
|
|
void *flash_p;
|
|
|
|
memset(write_buffer, flash_value, sizeof(write_buffer));
|
|
|
|
/* Overwrite nvmem flash space with junk value. */
|
|
flash_physical_write(
|
|
CONFIG_FLASH_NEW_NVMEM_BASE_A - CONFIG_PROGRAM_MEMORY_BASE,
|
|
NEW_FLASH_HALF_NVMEM_SIZE, (const char *)write_buffer);
|
|
flash_physical_write(
|
|
CONFIG_FLASH_NEW_NVMEM_BASE_B - CONFIG_PROGRAM_MEMORY_BASE,
|
|
NEW_FLASH_HALF_NVMEM_SIZE, (const char *)write_buffer);
|
|
|
|
TEST_ASSERT(nvmem_init() == EC_SUCCESS);
|
|
TEST_ASSERT(iterate_over_flash() == EC_SUCCESS);
|
|
TEST_ASSERT(test_result.var_count == 0);
|
|
TEST_ASSERT(test_result.reserved_obj_count == 38);
|
|
TEST_ASSERT(test_result.evictable_obj_count == 0);
|
|
TEST_ASSERT(test_result.deleted_obj_count == 0);
|
|
TEST_ASSERT(test_result.unexpected_count == 0);
|
|
TEST_ASSERT(test_result.valid_data_size == 1088);
|
|
TEST_ASSERT(total_var_space == 0);
|
|
|
|
for (i = 0; i < (NEW_NVMEM_TOTAL_PAGES - 1); i++) {
|
|
flash_p = page_to_flash_addr(i);
|
|
|
|
TEST_ASSERT(!!flash_p);
|
|
TEST_ASSERT(is_uninitialized(flash_p, CONFIG_FLASH_BANK_SIZE));
|
|
}
|
|
|
|
flash_p = page_to_flash_addr(i);
|
|
TEST_ASSERT(!is_uninitialized(flash_p, CONFIG_FLASH_BANK_SIZE));
|
|
|
|
return EC_SUCCESS;
|
|
}
|
|
|
|
/*
|
|
* The purpose of this test is to check NvMem initialization when NvMem is
|
|
* completely erased (i.e. following SpiFlash write of program). In this case,
|
|
* nvmem_init() is expected to create initial flash storage containing
|
|
* reserved objects only.
|
|
*/
|
|
static int test_fully_erased_nvmem(void)
|
|
{
|
|
|
|
return post_init_from_scratch(0xff);
|
|
}
|
|
|
|
/*
|
|
* The purpose of this test is to check nvmem_init() in the case when no valid
|
|
* pages exist but flash space is garbled as opposed to be fully erased. In
|
|
* this case, the initialization is expected to create one new valid page and
|
|
* erase the rest of the pages.
|
|
*/
|
|
static int test_corrupt_nvmem(void)
|
|
{
|
|
return post_init_from_scratch(0x55);
|
|
}
|
|
|
|
static int prepare_new_flash(void)
|
|
{
|
|
TEST_ASSERT(test_fully_erased_nvmem() == EC_SUCCESS);
|
|
|
|
/* Now copy sensible information into the nvmem cache. */
|
|
memcpy(nvmem_cache_base(NVMEM_TPM),
|
|
legacy_nvmem_image + sizeof(struct nvmem_tag),
|
|
nvmem_user_sizes[NVMEM_TPM]);
|
|
|
|
dump_nvmem_state("after first save", &test_result);
|
|
TEST_ASSERT(new_nvmem_save() == EC_SUCCESS);
|
|
TEST_ASSERT(iterate_over_flash() == EC_SUCCESS);
|
|
|
|
TEST_ASSERT(test_result.deleted_obj_count == 24);
|
|
TEST_ASSERT(test_result.var_count == 0);
|
|
TEST_ASSERT(test_result.reserved_obj_count == 40);
|
|
TEST_ASSERT(test_result.evictable_obj_count == 9);
|
|
TEST_ASSERT(test_result.unexpected_count == 0);
|
|
TEST_ASSERT(test_result.valid_data_size == 5128);
|
|
TEST_ASSERT(test_result.erased_data_size == 698);
|
|
|
|
return EC_SUCCESS;
|
|
}
|
|
|
|
static int test_nvmem_save(void)
|
|
{
|
|
const char *key = "var1";
|
|
const char *value = "value of var 1";
|
|
size_t total_var_size;
|
|
struct nvmem_test_result old_result;
|
|
|
|
TEST_ASSERT(prepare_new_flash() == EC_SUCCESS);
|
|
|
|
/*
|
|
* Verify that saving without changing the cache does not affect flash
|
|
* contents.
|
|
*/
|
|
old_result = test_result;
|
|
TEST_ASSERT(new_nvmem_save() == EC_SUCCESS);
|
|
|
|
/*
|
|
* Save of unmodified cache does not modify the flash contents and
|
|
* does not set the delimiter.
|
|
*/
|
|
|
|
TEST_ASSERT(iterate_over_flash() == EC_SUCCESS);
|
|
TEST_ASSERT(!memcmp(&test_result, &old_result, sizeof(test_result)));
|
|
|
|
wipe_out_nvmem_cache();
|
|
TEST_ASSERT(nvmem_init() == EC_SUCCESS);
|
|
TEST_ASSERT(new_nvmem_save() == EC_SUCCESS);
|
|
TEST_ASSERT(iterate_over_flash() == EC_SUCCESS);
|
|
TEST_ASSERT(!memcmp(&test_result, &old_result, sizeof(test_result)));
|
|
|
|
/*
|
|
* Total size test variable storage takes in flash (container header
|
|
* size not included).
|
|
*/
|
|
total_var_size = strlen(key) + strlen(value) + sizeof(struct tuple);
|
|
|
|
/* Verify that we can add a variable to nvmem. */
|
|
TEST_ASSERT(setvar(key, strlen(key), value, strlen(value)) ==
|
|
EC_SUCCESS);
|
|
TEST_ASSERT(iterate_over_flash() == EC_SUCCESS);
|
|
|
|
/* Remove changes caused by the new var addition. */
|
|
test_result.var_count -= 1;
|
|
test_result.delimiter_count -= 1;
|
|
test_result.valid_data_size -= total_var_size;
|
|
|
|
TEST_ASSERT(memcmp(&test_result, &old_result, sizeof(test_result)) ==
|
|
0);
|
|
|
|
/* Verify that we can delete a variable from nvmem. */
|
|
TEST_ASSERT(setvar(key, strlen(key), NULL, 0) == EC_SUCCESS);
|
|
TEST_ASSERT(iterate_over_flash() == EC_SUCCESS);
|
|
test_result.deleted_obj_count -= 1;
|
|
test_result.erased_data_size -= total_var_size;
|
|
test_result.delimiter_count -= 1;
|
|
TEST_ASSERT(memcmp(&test_result, &old_result, sizeof(test_result)) ==
|
|
0);
|
|
|
|
return EC_SUCCESS;
|
|
}
|
|
|
|
static size_t get_free_nvmem_room(void)
|
|
{
|
|
size_t free_room;
|
|
size_t free_pages;
|
|
/* Compaction kicks in when 3 pages or less are left. */
|
|
const size_t max_pages = NEW_NVMEM_TOTAL_PAGES - 3;
|
|
|
|
ccprintf("list index %d, data offset 0x%x\n", master_at.list_index,
|
|
master_at.mt.data_offset);
|
|
|
|
if (master_at.list_index >= max_pages)
|
|
return 0;
|
|
|
|
free_pages = max_pages - master_at.list_index;
|
|
free_room = (free_pages - 1) * (CONFIG_FLASH_BANK_SIZE -
|
|
sizeof(struct nn_page_header)) +
|
|
CONFIG_FLASH_BANK_SIZE - master_at.mt.data_offset;
|
|
ccprintf("free pages %d, data offset 0x%x\n", free_pages,
|
|
master_at.mt.data_offset);
|
|
return free_room;
|
|
}
|
|
|
|
static int test_nvmem_compaction(void)
|
|
{
|
|
char value[100]; /* Definitely more than enough. */
|
|
const char *key = "var 1";
|
|
int i;
|
|
size_t key_len;
|
|
size_t val_len;
|
|
size_t free_room;
|
|
size_t real_var_size;
|
|
size_t var_space;
|
|
int max_vars;
|
|
int erased_data_size;
|
|
const size_t alignment_mask = CONFIG_FLASH_WRITE_SIZE - 1;
|
|
|
|
key_len = strlen(key);
|
|
val_len = snprintf(value, sizeof(value), "variable value is %04d", 0);
|
|
|
|
TEST_ASSERT(prepare_new_flash() == EC_SUCCESS);
|
|
|
|
/*
|
|
* Remember how much room was erased before flooding nvmem with erased
|
|
* values.
|
|
*/
|
|
erased_data_size = test_result.erased_data_size;
|
|
|
|
/* Let's see how much free room there is. */
|
|
free_room = get_free_nvmem_room();
|
|
TEST_ASSERT(free_room);
|
|
|
|
/* How much room (key, value) pair takes in a container. */
|
|
real_var_size = val_len + key_len + sizeof(struct tuple);
|
|
/*
|
|
* See how many vars including containers should be able to fit there.
|
|
*
|
|
* First calculate rounded up space a var will take. Apart from the
|
|
* var itself there will be a container header and a delimiter.
|
|
*/
|
|
var_space = (real_var_size + 2 * sizeof(struct nn_container) +
|
|
alignment_mask) & ~alignment_mask;
|
|
|
|
max_vars = free_room / var_space;
|
|
|
|
/*
|
|
* And now flood the NVMEM with erased values (each new setvar()
|
|
* invocation erases the previous instance.
|
|
*/
|
|
for (i = 0; i <= max_vars; i++) {
|
|
snprintf(value, sizeof(value), "variable value is %04d", i);
|
|
TEST_ASSERT(setvar(key, key_len, value, val_len) == EC_SUCCESS);
|
|
}
|
|
|
|
TEST_ASSERT(iterate_over_flash() == EC_SUCCESS);
|
|
/* Make sure there was no compaction yet. */
|
|
TEST_ASSERT(test_result.erased_data_size > erased_data_size);
|
|
|
|
/* This is how much the erased space grew as a result of flooding. */
|
|
erased_data_size = test_result.erased_data_size - erased_data_size;
|
|
TEST_ASSERT(erased_data_size == max_vars * real_var_size);
|
|
|
|
/* This will take it over the compaction limit. */
|
|
val_len = snprintf(value, sizeof(value), "variable value is %03d", i);
|
|
TEST_ASSERT(setvar(key, key_len, value, val_len) == EC_SUCCESS);
|
|
TEST_ASSERT(iterate_over_flash() == EC_SUCCESS);
|
|
TEST_ASSERT(test_result.erased_data_size < var_space);
|
|
|
|
return EC_SUCCESS;
|
|
}
|
|
|
|
static int test_configured_nvmem(void)
|
|
{
|
|
/*
|
|
* The purpose of this test is to check how nvmem_init() initializes
|
|
* from previously saved flash contents.
|
|
*/
|
|
TEST_ASSERT(prepare_nvmem_contents() == EC_SUCCESS);
|
|
|
|
/*
|
|
* This is initialization from legacy flash contents which replaces
|
|
* legacy flash image with the new format flash image
|
|
*/
|
|
TEST_ASSERT(nvmem_init() == EC_SUCCESS);
|
|
|
|
/* And this is initialization from the new flash layout. */
|
|
return nvmem_init();
|
|
}
|
|
|
|
static uint8_t find_lb(const void *data)
|
|
{
|
|
return (const uint8_t *)memchr(data, '#', 256) - (const uint8_t *)data;
|
|
}
|
|
|
|
/*
|
|
* Helper function, depending on the argument value either writes variables
|
|
* into nvmem and verifies their presence, or deletes them and verifies that
|
|
* they indeed disappear.
|
|
*/
|
|
static int var_read_write_delete_helper(int do_write)
|
|
{
|
|
size_t i;
|
|
uint16_t saved_total_var_space;
|
|
uint32_t coverage_map;
|
|
|
|
const struct {
|
|
uint8_t *key;
|
|
uint8_t *value;
|
|
} kv_pairs[] = {
|
|
/* Use # as the delimiter to allow \0 in keys/values. */
|
|
{"\0key\00#", "value of key2#"}, {"key1#", "value of key1#"},
|
|
{"key2#", "value of key2#"}, {"key3#", "value of\0 key3#"},
|
|
{"ke\04#", "value\0 of\0 key4#"},
|
|
};
|
|
|
|
coverage_map = 0;
|
|
saved_total_var_space = total_var_space;
|
|
|
|
/*
|
|
* Read all vars, one at a time, verifying that they shows up in
|
|
* getvar results when appropriate but not before.
|
|
*/
|
|
for (i = 0; i <= ARRAY_SIZE(kv_pairs); i++) {
|
|
size_t j;
|
|
uint8_t key_len;
|
|
uint8_t val_len;
|
|
const void *value;
|
|
|
|
for (j = 0; j < ARRAY_SIZE(kv_pairs); j++) {
|
|
const struct tuple *t;
|
|
|
|
coverage_map |= 1;
|
|
|
|
key_len = find_lb(kv_pairs[j].key);
|
|
t = getvar(kv_pairs[j].key, key_len);
|
|
|
|
if ((j >= i) ^ !do_write) {
|
|
TEST_ASSERT(t == NULL);
|
|
continue;
|
|
}
|
|
|
|
coverage_map |= 2;
|
|
|
|
TEST_ASSERT(saved_total_var_space == total_var_space);
|
|
|
|
/* Confirm that what we found is the right variable. */
|
|
val_len = find_lb(kv_pairs[j].value);
|
|
|
|
TEST_ASSERT(t->key_len == key_len);
|
|
TEST_ASSERT(t->val_len == val_len);
|
|
TEST_ASSERT(
|
|
!memcmp(kv_pairs[j].key, t->data_, key_len));
|
|
TEST_ASSERT(!memcmp(kv_pairs[j].value,
|
|
t->data_ + key_len, val_len));
|
|
freevar(t);
|
|
}
|
|
|
|
if (i == ARRAY_SIZE(kv_pairs)) {
|
|
coverage_map |= 4;
|
|
/* All four variables have been processed. */
|
|
break;
|
|
}
|
|
|
|
val_len = find_lb(kv_pairs[i].value);
|
|
key_len = find_lb(kv_pairs[i].key);
|
|
value = kv_pairs[i].value;
|
|
if (!do_write) {
|
|
|
|
coverage_map |= 8;
|
|
|
|
saved_total_var_space -= val_len + key_len;
|
|
/*
|
|
* Make sure all val_len == 0 and val == NULL
|
|
* combinations are exercised.
|
|
*/
|
|
switch (i) {
|
|
case 0:
|
|
val_len = 0;
|
|
coverage_map |= 0x10;
|
|
break;
|
|
|
|
case 1:
|
|
coverage_map |= 0x20;
|
|
value = NULL;
|
|
break;
|
|
default:
|
|
coverage_map |= 0x40;
|
|
val_len = 0;
|
|
value = NULL;
|
|
break;
|
|
}
|
|
} else {
|
|
coverage_map |= 0x80;
|
|
saved_total_var_space += val_len + key_len;
|
|
}
|
|
key_len = find_lb(kv_pairs[i].key);
|
|
TEST_ASSERT(setvar(kv_pairs[i].key, key_len, value, val_len) ==
|
|
EC_SUCCESS);
|
|
|
|
TEST_ASSERT(saved_total_var_space == total_var_space);
|
|
}
|
|
|
|
if (do_write)
|
|
TEST_ASSERT(coverage_map == 0x87);
|
|
else
|
|
TEST_ASSERT(coverage_map == 0x7f);
|
|
|
|
return EC_SUCCESS;
|
|
}
|
|
|
|
static int test_var_read_write_delete(void)
|
|
{
|
|
TEST_ASSERT(post_init_from_scratch(0xff) == EC_SUCCESS);
|
|
|
|
ccprintf("\n%s: starting write cycle\n", __func__);
|
|
TEST_ASSERT(var_read_write_delete_helper(1) == EC_SUCCESS);
|
|
|
|
ccprintf("%s: starting delete cycle\n", __func__);
|
|
TEST_ASSERT(var_read_write_delete_helper(0) == EC_SUCCESS);
|
|
|
|
return EC_SUCCESS;
|
|
}
|
|
/* Verify that nvmem_erase_user_data only erases the given user's data. */
|
|
static int test_nvmem_erase_tpm_data(void)
|
|
{
|
|
TEST_ASSERT(prepare_nvmem_contents() == EC_SUCCESS);
|
|
TEST_ASSERT(nvmem_init() == EC_SUCCESS);
|
|
browse_flash_contents(1);
|
|
TEST_ASSERT(nvmem_erase_tpm_data() == EC_SUCCESS);
|
|
browse_flash_contents(1);
|
|
TEST_ASSERT(iterate_over_flash() == EC_SUCCESS);
|
|
TEST_ASSERT(test_result.deleted_obj_count == 0);
|
|
TEST_ASSERT(test_result.var_count == 3);
|
|
TEST_ASSERT(test_result.reserved_obj_count == 38);
|
|
TEST_ASSERT(test_result.evictable_obj_count == 0);
|
|
TEST_ASSERT(test_result.unexpected_count == 0);
|
|
TEST_ASSERT(test_result.valid_data_size == 1174);
|
|
TEST_ASSERT(test_result.erased_data_size == 0);
|
|
|
|
return EC_SUCCESS;
|
|
}
|
|
|
|
static size_t fill_obj_offsets(uint16_t *offsets, size_t max_objects)
|
|
{
|
|
size_t i;
|
|
size_t obj_count;
|
|
|
|
obj_count = init_object_offsets(offsets, max_objects);
|
|
|
|
ccprintf("%d objects\n", obj_count);
|
|
for (i = 0; i < obj_count; i++) {
|
|
uint32_t *op;
|
|
|
|
op = evictable_offs_to_addr(offsets[i]);
|
|
ccprintf("offs %04x:%08x:%08x:%08x addr %p size %d\n",
|
|
offsets[i], op[-1], op[0], op[1], op,
|
|
(uintptr_t)nvmem_cache_base(NVMEM_TPM) + op[-1] -
|
|
(uintptr_t)op);
|
|
}
|
|
|
|
return obj_count;
|
|
}
|
|
|
|
static size_t fill_cache_offsets(const void *cache, uint16_t *offsets,
|
|
size_t max_objects)
|
|
{
|
|
uint8_t buf[nvmem_user_sizes[NVMEM_TPM]];
|
|
void *real_cache;
|
|
size_t num_offsets;
|
|
|
|
real_cache = nvmem_cache_base(NVMEM_TPM);
|
|
memcpy(buf, real_cache, sizeof(buf));
|
|
|
|
memcpy(real_cache, cache, sizeof(buf));
|
|
memset(offsets, 0, sizeof(*offsets) * max_objects);
|
|
num_offsets = fill_obj_offsets(offsets, max_objects);
|
|
|
|
/* Restore the real cache. */
|
|
memcpy(real_cache, buf, sizeof(buf));
|
|
|
|
return num_offsets;
|
|
}
|
|
#define MAX_OFFSETS 20
|
|
|
|
static uint32_t get_evict_size(const uint8_t *cache, uint16_t offset)
|
|
{
|
|
uint32_t next_addr;
|
|
uint32_t cache_offset;
|
|
|
|
cache_offset = s_evictNvStart + offset;
|
|
memcpy(&next_addr, cache + cache_offset - sizeof(next_addr),
|
|
sizeof(next_addr));
|
|
|
|
return next_addr - cache_offset;
|
|
}
|
|
|
|
/* Returns zero if the two objects are identical. */
|
|
static int compare_objects(const uint8_t *cache1, uint16_t offset1,
|
|
const uint8_t *cache2, uint16_t offset2)
|
|
{
|
|
uint32_t size1;
|
|
uint32_t size2;
|
|
|
|
size1 = get_evict_size(cache1, offset1);
|
|
size2 = get_evict_size(cache2, offset2);
|
|
|
|
if (size1 == size2)
|
|
return memcmp(cache1 + s_evictNvStart + offset1,
|
|
cache2 + s_evictNvStart + offset2, size1);
|
|
|
|
return 1;
|
|
}
|
|
/*
|
|
* Compare two instances of NVMEM caches. Reserved spaces should be exactly
|
|
* the same for the match, but evictable objects could be rearranged due to
|
|
* compaction, updating, etc.
|
|
*
|
|
* For the two cache instances to be considered the same the sets and contents
|
|
* of the evictable object spaces must also match object to object.
|
|
*/
|
|
static int caches_match(const uint8_t *cache1, const uint8_t *cache2)
|
|
{
|
|
int failed_count;
|
|
size_t cache1_offs_count;
|
|
size_t cache2_offs_count;
|
|
size_t i;
|
|
uint16_t cache1_offsets[MAX_OFFSETS];
|
|
uint16_t cache2_offsets[MAX_OFFSETS];
|
|
|
|
for (failed_count = i = 0; i < NV_PSEUDO_RESERVE_LAST; i++) {
|
|
NV_RESERVED_ITEM ri;
|
|
struct {
|
|
uint32_t offset;
|
|
uint32_t size;
|
|
} ranges[3];
|
|
size_t j;
|
|
|
|
NvGetReserved(i, &ri);
|
|
|
|
ranges[0].offset = ri.offset;
|
|
|
|
if (i != NV_STATE_CLEAR) {
|
|
ranges[0].size = ri.size;
|
|
ranges[1].size = 0;
|
|
} else {
|
|
ranges[0].size = offsetof(STATE_CLEAR_DATA, pcrSave);
|
|
ranges[1].offset = ranges[0].offset + ranges[0].size;
|
|
ranges[1].size = sizeof(PCR_SAVE);
|
|
ranges[2].offset = ranges[1].offset + ranges[1].size;
|
|
ranges[2].size = sizeof(PCR_AUTHVALUE);
|
|
}
|
|
|
|
for (j = 0; j < ARRAY_SIZE(ranges); j++) {
|
|
|
|
uint32_t offset;
|
|
uint32_t size;
|
|
uint32_t k;
|
|
|
|
size = ranges[j].size;
|
|
if (!size)
|
|
break;
|
|
|
|
offset = ranges[j].offset;
|
|
|
|
if (!memcmp(cache1 + offset, cache2 + offset, size))
|
|
continue;
|
|
|
|
ccprintf("%s:%d failed comparing %d:%d:\n", __func__,
|
|
__LINE__, i, j);
|
|
for (k = offset; k < (offset + size); k++)
|
|
if (cache1[k] != cache2[k])
|
|
ccprintf(" %3d:%02x", k - offset,
|
|
cache1[k]);
|
|
ccprintf("\n");
|
|
for (k = offset; k < (offset + size); k++)
|
|
if (cache1[k] != cache2[k])
|
|
ccprintf(" %3d:%02x", k - offset,
|
|
cache2[k]);
|
|
ccprintf("\n");
|
|
|
|
failed_count++;
|
|
}
|
|
}
|
|
|
|
TEST_ASSERT(!failed_count);
|
|
|
|
cache1_offs_count = fill_cache_offsets(cache1, cache1_offsets,
|
|
ARRAY_SIZE(cache1_offsets));
|
|
cache2_offs_count = fill_cache_offsets(cache2, cache2_offsets,
|
|
ARRAY_SIZE(cache2_offsets));
|
|
|
|
TEST_ASSERT(cache1_offs_count == cache2_offs_count);
|
|
|
|
for (i = 0; (i < ARRAY_SIZE(cache1_offsets)) && cache2_offs_count;
|
|
i++) {
|
|
size_t j;
|
|
|
|
for (j = 0; j < cache2_offs_count; j++) {
|
|
if (compare_objects(cache1, cache1_offsets[i], cache2,
|
|
cache2_offsets[j]))
|
|
continue;
|
|
/* Remove object from the cache2 offsets. */
|
|
cache2_offsets[j] = cache2_offsets[--cache2_offs_count];
|
|
break;
|
|
}
|
|
}
|
|
|
|
TEST_ASSERT(cache2_offs_count == 0);
|
|
return EC_SUCCESS;
|
|
}
|
|
|
|
static int prepare_post_migration_nvmem(void)
|
|
{
|
|
TEST_ASSERT(prepare_nvmem_contents() == EC_SUCCESS);
|
|
TEST_ASSERT(nvmem_init() == EC_SUCCESS);
|
|
TEST_ASSERT(new_nvmem_save() == EC_SUCCESS);
|
|
TEST_ASSERT(nvmem_init() == EC_SUCCESS);
|
|
|
|
return EC_SUCCESS;
|
|
}
|
|
/*
|
|
* This test creates various failure conditions related to interrupted nvmem
|
|
* save operations and verifies that transaction integrity is maintained -
|
|
* i.e. either all variables get updated,
|
|
*/
|
|
static int test_nvmem_incomplete_transaction(void)
|
|
{
|
|
/*
|
|
* Will be more than enough, we can't store more than 15 objects or so
|
|
* anyways.
|
|
*/
|
|
uint16_t offsets[MAX_OFFSETS];
|
|
size_t num_objects;
|
|
uint8_t buf[nvmem_user_sizes[NVMEM_TPM]];
|
|
uint8_t *p;
|
|
size_t object_size;
|
|
union entry_u e;
|
|
|
|
TEST_ASSERT(prepare_post_migration_nvmem() == EC_SUCCESS);
|
|
num_objects = fill_obj_offsets(offsets, ARRAY_SIZE(offsets));
|
|
TEST_ASSERT(num_objects == 9);
|
|
|
|
/* Save cache state before deleting objects. */
|
|
memcpy(buf, nvmem_cache_base(NVMEM_TPM), sizeof(buf));
|
|
|
|
drop_evictable_obj(evictable_offs_to_addr(offsets[4]));
|
|
drop_evictable_obj(evictable_offs_to_addr(offsets[3]));
|
|
|
|
failure_mode = TEST_FAIL_WHEN_SAVING;
|
|
TEST_ASSERT(new_nvmem_save() == EC_SUCCESS);
|
|
wipe_out_nvmem_cache();
|
|
TEST_ASSERT(nvmem_init() == EC_SUCCESS);
|
|
|
|
TEST_ASSERT(caches_match(buf, nvmem_cache_base(NVMEM_TPM)) ==
|
|
EC_SUCCESS);
|
|
drop_evictable_obj(evictable_offs_to_addr(offsets[4]));
|
|
drop_evictable_obj(evictable_offs_to_addr(offsets[3]));
|
|
|
|
/* Check if failure when invalidating is recovered after restart. */
|
|
failure_mode = TEST_FAIL_WHEN_INVALIDATING;
|
|
TEST_ASSERT(new_nvmem_save() == EC_SUCCESS);
|
|
ccprintf("%s:%d\n", __func__, __LINE__);
|
|
wipe_out_nvmem_cache();
|
|
TEST_ASSERT(nvmem_init() == EC_SUCCESS);
|
|
ccprintf("%s:%d\n", __func__, __LINE__);
|
|
num_objects = fill_obj_offsets(offsets, ARRAY_SIZE(offsets));
|
|
TEST_ASSERT(num_objects == 7);
|
|
|
|
/*
|
|
* Now, let's modify an object and introduce corruption when saving
|
|
* it.
|
|
*/
|
|
p = evictable_offs_to_addr(offsets[4]);
|
|
p[10] ^= 0x55;
|
|
failure_mode = TEST_FAILED_HASH;
|
|
new_nvmem_save();
|
|
failure_mode = TEST_NO_FAILURE;
|
|
|
|
/* And verify that nvmem can still successfully initialize. */
|
|
TEST_ASSERT(nvmem_init() == EC_SUCCESS);
|
|
|
|
/*
|
|
* Now let's interrupt saving an object spanning two pages.
|
|
*
|
|
* First, fill up the current page to get close to the limit such that
|
|
* the next save will have to span two flash pages.
|
|
*/
|
|
object_size = offsets[4] - offsets[3];
|
|
p = (uint8_t *)evictable_offs_to_addr(offsets[3]) + object_size - 10;
|
|
while ((master_at.mt.data_offset + object_size +
|
|
sizeof(struct nn_container)) <= CONFIG_FLASH_BANK_SIZE) {
|
|
(*p)++;
|
|
new_nvmem_save();
|
|
}
|
|
|
|
/* This will trigger spilling over the page boundary. */
|
|
(*p)++;
|
|
failure_mode = TEST_SPANNING_PAGES;
|
|
new_nvmem_save();
|
|
failure_mode = TEST_NO_FAILURE;
|
|
|
|
/* Drain the event log. */
|
|
e.r.timestamp = 0;
|
|
while (flash_log_dequeue_event(e.r.timestamp, e.entry, sizeof(e)) > 0)
|
|
;
|
|
|
|
TEST_ASSERT(nvmem_init() == EC_SUCCESS);
|
|
|
|
/* Let's verify that a container mismatch event has been added. */
|
|
TEST_ASSERT(flash_log_dequeue_event(e.r.timestamp, e.entry, sizeof(e))
|
|
> 0);
|
|
TEST_ASSERT(e.r.type == FE_LOG_NVMEM);
|
|
TEST_ASSERT(e.r.payload[0] == NVMEMF_CONTAINER_HASH_MISMATCH);
|
|
return EC_SUCCESS;
|
|
}
|
|
|
|
/*
|
|
* Verify that interrupted compaction results in a consistent state of the
|
|
* NVMEM cache.
|
|
*/
|
|
static int test_nvmem_interrupted_compaction(void)
|
|
{
|
|
uint8_t buf[nvmem_user_sizes[NVMEM_TPM]];
|
|
uint8_t target_list_index;
|
|
uint8_t filler = 1;
|
|
|
|
TEST_ASSERT(prepare_post_migration_nvmem() == EC_SUCCESS);
|
|
|
|
/* Let's fill up a couple of pages with erased objects. */
|
|
target_list_index = master_at.list_index + 2;
|
|
|
|
do {
|
|
/*
|
|
* A few randomly picked reserved objects to modify to create
|
|
* need for compaction.
|
|
*/
|
|
const uint8_t objs_to_modify[] = {1, 3, 19, 42};
|
|
size_t i;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(objs_to_modify); i++) {
|
|
NV_RESERVED_ITEM ri;
|
|
|
|
NvGetReserved(i, &ri);
|
|
|
|
/* Direct access to the object. */
|
|
memset((uint8_t *)nvmem_cache_base(NVMEM_TPM) +
|
|
ri.offset,
|
|
filler++, ri.size);
|
|
}
|
|
TEST_ASSERT(new_nvmem_save() == EC_SUCCESS);
|
|
} while (master_at.list_index != target_list_index);
|
|
|
|
/* Save the state of NVMEM cache. */
|
|
memcpy(buf, nvmem_cache_base(NVMEM_TPM), sizeof(buf));
|
|
failure_mode = TEST_FAIL_WHEN_COMPACTING;
|
|
compact_nvmem();
|
|
wipe_out_nvmem_cache();
|
|
ccprintf("%s:%d\n", __func__, __LINE__);
|
|
TEST_ASSERT(nvmem_init() == EC_SUCCESS);
|
|
TEST_ASSERT(caches_match(buf, nvmem_cache_base(NVMEM_TPM)) ==
|
|
EC_SUCCESS);
|
|
return EC_SUCCESS;
|
|
}
|
|
|
|
int nvmem_first_task(void *unused)
|
|
{
|
|
return EC_SUCCESS;
|
|
}
|
|
|
|
int nvmem_second_task(void *unused)
|
|
{
|
|
return EC_SUCCESS;
|
|
}
|
|
|
|
static void run_test_setup(void)
|
|
{
|
|
/* Allow Flash erase/writes */
|
|
flash_write_fail = 0;
|
|
test_reset();
|
|
}
|
|
|
|
void nvmem_wipe_cache(void)
|
|
{
|
|
}
|
|
|
|
int DCRYPTO_ladder_is_enabled(void)
|
|
{
|
|
return 1;
|
|
}
|
|
|
|
static int test_migration(void)
|
|
{
|
|
/*
|
|
* This purpose of this test is to verify migration of the 'legacy'
|
|
* TPM NVMEM format to the new scheme where each element is stored in
|
|
* flash in its own container.
|
|
*/
|
|
TEST_ASSERT(prepare_nvmem_contents() == EC_SUCCESS);
|
|
TEST_ASSERT(nvmem_init() == EC_SUCCESS);
|
|
TEST_ASSERT(iterate_over_flash() == EC_SUCCESS);
|
|
TEST_ASSERT(test_result.var_count == 3);
|
|
TEST_ASSERT(test_result.reserved_obj_count == 40);
|
|
TEST_ASSERT(test_result.evictable_obj_count == 9);
|
|
TEST_ASSERT(test_result.delimiter_count == 1);
|
|
TEST_ASSERT(test_result.deleted_obj_count == 0);
|
|
TEST_ASSERT(test_result.unexpected_count == 0);
|
|
TEST_ASSERT(test_result.valid_data_size == 5214);
|
|
TEST_ASSERT(total_var_space == 77);
|
|
/* Container pointer not yet set. */
|
|
TEST_ASSERT(!master_at.ct.data_offset && !master_at.ct.ph);
|
|
return EC_SUCCESS;
|
|
}
|
|
|
|
/*
|
|
* The purpose of this test is to verify variable storage limits, both per
|
|
* object and total.
|
|
*/
|
|
static int test_var_boundaries(void)
|
|
{
|
|
const size_t max_size = 255; /* Key and value must fit in a byte. */
|
|
const uint8_t *key;
|
|
const uint8_t *val;
|
|
size_t key_len;
|
|
size_t val_len;
|
|
uint16_t saved_total_var_space;
|
|
uint32_t coverage_map;
|
|
uint8_t var_key[10];
|
|
|
|
TEST_ASSERT(prepare_new_flash() == EC_SUCCESS);
|
|
saved_total_var_space = total_var_space;
|
|
coverage_map = 0;
|
|
|
|
/*
|
|
* Let's use the legacy NVMEM image as a source of fairly random but
|
|
* reproducible data.
|
|
*/
|
|
key = legacy_nvmem_image;
|
|
val = legacy_nvmem_image;
|
|
|
|
/*
|
|
* Test limit of max variable body space, use keys and values of
|
|
* different sizes, below and above the limit.
|
|
*/
|
|
for (key_len = 1; key_len < max_size; key_len += 20) {
|
|
|
|
coverage_map |= 1;
|
|
|
|
val_len = MIN(max_size, MAX_VAR_BODY_SPACE - key_len);
|
|
TEST_ASSERT(setvar(key, key_len, val, val_len) == EC_SUCCESS);
|
|
TEST_ASSERT(total_var_space ==
|
|
saved_total_var_space + key_len + val_len);
|
|
|
|
/* Now drop the variable from the storage. */
|
|
TEST_ASSERT(setvar(key, key_len, NULL, 0) == EC_SUCCESS);
|
|
TEST_ASSERT(total_var_space == saved_total_var_space);
|
|
|
|
/* And if key length allows it, try to write too much. */
|
|
if (val_len == max_size)
|
|
continue;
|
|
|
|
coverage_map |= 2;
|
|
/*
|
|
* Yes, let's try writing one byte too many and see that the
|
|
* attempt is rejected.
|
|
*/
|
|
val_len++;
|
|
TEST_ASSERT(setvar(key, key_len, val, val_len) ==
|
|
EC_ERROR_INVAL);
|
|
TEST_ASSERT(total_var_space == saved_total_var_space);
|
|
}
|
|
|
|
/*
|
|
* Test limit of max total variable space, use keys and values of
|
|
* different sizes, below and above the limit.
|
|
*/
|
|
key_len = sizeof(var_key);
|
|
val_len = 20; /* Anything below 256 would work. */
|
|
memset(var_key, 'x', key_len);
|
|
|
|
while (1) {
|
|
int rv;
|
|
|
|
/*
|
|
* Change the key so that a new variable is added to the
|
|
* storage.
|
|
*/
|
|
rv = setvar(var_key, key_len, val, val_len);
|
|
|
|
if (rv == EC_ERROR_OVERFLOW)
|
|
break;
|
|
|
|
coverage_map |= 4;
|
|
TEST_ASSERT(rv == EC_SUCCESS);
|
|
var_key[0]++;
|
|
saved_total_var_space += key_len + val_len;
|
|
}
|
|
|
|
TEST_ASSERT(saved_total_var_space == total_var_space);
|
|
TEST_ASSERT(saved_total_var_space <= MAX_VAR_TOTAL_SPACE);
|
|
TEST_ASSERT((saved_total_var_space + key_len + val_len) >
|
|
MAX_VAR_TOTAL_SPACE);
|
|
|
|
TEST_ASSERT(coverage_map == 7);
|
|
return EC_SUCCESS;
|
|
}
|
|
|
|
static int verify_ram_index_space(size_t verify_size)
|
|
{
|
|
NV_RESERVED_ITEM ri;
|
|
size_t i;
|
|
uint32_t casted_size;
|
|
uint8_t byte;
|
|
uint8_t fill_byte = 0x55;
|
|
|
|
if (verify_size > RAM_INDEX_SPACE)
|
|
return EC_ERROR_INVAL;
|
|
|
|
NvGetReserved(NV_RAM_INDEX_SPACE, &ri);
|
|
|
|
/*
|
|
* Save the size of the index space, needed on machines where size_t
|
|
* is a 64 bit value.
|
|
*/
|
|
casted_size = verify_size;
|
|
|
|
/*
|
|
* Now write index space in the cache, we write the complete space,
|
|
* but on read back only verify_size bytes are expected to be set.
|
|
*/
|
|
nvmem_write(ri.offset, sizeof(casted_size), &casted_size, NVMEM_TPM);
|
|
|
|
for (i = 0; i < RAM_INDEX_SPACE; i++)
|
|
nvmem_write(ri.offset + sizeof(casted_size) + i,
|
|
sizeof(fill_byte), &fill_byte, NVMEM_TPM);
|
|
|
|
TEST_ASSERT(new_nvmem_save() == EC_SUCCESS);
|
|
wipe_out_nvmem_cache();
|
|
TEST_ASSERT(nvmem_init() == EC_SUCCESS);
|
|
|
|
/* Make sure read back size matches. */
|
|
nvmem_read(ri.offset, sizeof(casted_size), &casted_size, NVMEM_TPM);
|
|
TEST_ASSERT(casted_size == verify_size);
|
|
|
|
/*
|
|
* Now check spaces which were supposed to be written (up to
|
|
* verify_size) and left intact.
|
|
*/
|
|
for (i = 0; i < RAM_INDEX_SPACE; i++) {
|
|
nvmem_read(ri.offset + sizeof(casted_size) + i, sizeof(byte),
|
|
&byte, NVMEM_TPM);
|
|
if (i < verify_size)
|
|
TEST_ASSERT(byte == fill_byte);
|
|
else
|
|
TEST_ASSERT(byte == 0);
|
|
}
|
|
|
|
return EC_SUCCESS;
|
|
}
|
|
|
|
static int test_tpm_nvmem_modify_reserved_objects(void)
|
|
{
|
|
NV_RESERVED_ITEM ri;
|
|
/* Some random reserved objects' indices. */
|
|
const uint8_t res_obj_ids[] = {1, 4, 9, 20};
|
|
size_t i;
|
|
static uint8_t cache_copy[12 * 1024];
|
|
struct nvmem_test_result old_result;
|
|
uint64_t new_values[ARRAY_SIZE(res_obj_ids)];
|
|
size_t erased_size;
|
|
|
|
TEST_ASSERT(sizeof(cache_copy) >= nvmem_user_sizes[NVMEM_TPM]);
|
|
TEST_ASSERT(prepare_new_flash() == EC_SUCCESS);
|
|
TEST_ASSERT(new_nvmem_save() == EC_SUCCESS);
|
|
TEST_ASSERT(nvmem_init() == EC_SUCCESS);
|
|
iterate_over_flash();
|
|
old_result = test_result;
|
|
|
|
/* Preserve NVMEM cache for future comparison. */
|
|
memcpy(cache_copy, nvmem_cache_base(NVMEM_TPM),
|
|
nvmem_user_sizes[NVMEM_TPM]);
|
|
|
|
erased_size = 0;
|
|
/* Modify several reserved objects in the cache. */
|
|
for (i = 0; i < ARRAY_SIZE(res_obj_ids); i++) {
|
|
size_t copy_size;
|
|
uint8_t *addr_in_cache;
|
|
size_t k;
|
|
|
|
NvGetReserved(res_obj_ids[i], &ri);
|
|
copy_size = MIN(sizeof(new_values[0]), ri.size);
|
|
addr_in_cache =
|
|
(uint8_t *)nvmem_cache_base(NVMEM_TPM) + ri.offset;
|
|
|
|
/* Prepare a new value for the variable. */
|
|
memcpy(new_values + i, addr_in_cache, copy_size);
|
|
for (k = 0; k < copy_size; k++)
|
|
((uint8_t *)(new_values + i))[k] ^= 0x55;
|
|
|
|
/* Update value in the cache. */
|
|
memcpy(addr_in_cache, new_values + i, copy_size);
|
|
|
|
/* And in the cache copy. */
|
|
memcpy(cache_copy + ri.offset, new_values + i, copy_size);
|
|
|
|
/*
|
|
* This much will be added to the erased space, object size
|
|
* plus index size.
|
|
*/
|
|
erased_size += ri.size + 1;
|
|
}
|
|
|
|
/* Save it into flash. */
|
|
TEST_ASSERT(new_nvmem_save() == EC_SUCCESS);
|
|
|
|
/* Wipe out the cache to be sure. */
|
|
wipe_out_nvmem_cache();
|
|
|
|
/* Read NVMEM contents from flash. */
|
|
TEST_ASSERT(nvmem_init() == EC_SUCCESS);
|
|
|
|
/* Verify that the cache matches expectations. */
|
|
TEST_ASSERT(!memcmp(cache_copy, nvmem_cache_base(NVMEM_TPM),
|
|
nvmem_user_sizes[NVMEM_TPM]));
|
|
|
|
iterate_over_flash();
|
|
|
|
/* Update previous results with our expectations. */
|
|
old_result.deleted_obj_count += ARRAY_SIZE(res_obj_ids);
|
|
old_result.erased_data_size += erased_size;
|
|
old_result.delimiter_count++;
|
|
|
|
TEST_ASSERT(!memcmp(&test_result, &old_result, sizeof(test_result)));
|
|
|
|
/* Verify several index space cases. */
|
|
for (i = 0; i <= RAM_INDEX_SPACE; i += (RAM_INDEX_SPACE / 2))
|
|
TEST_ASSERT(verify_ram_index_space(i) == EC_SUCCESS);
|
|
|
|
return EC_SUCCESS;
|
|
}
|
|
|
|
static int compare_object(uint16_t obj_offset, size_t obj_size, const void *obj)
|
|
{
|
|
uint32_t next_addr;
|
|
|
|
memcpy(&next_addr,
|
|
evictable_offs_to_addr(obj_offset - sizeof(next_addr)),
|
|
sizeof(next_addr));
|
|
|
|
ccprintf("next_addr %x, sum %x size %d\n", next_addr,
|
|
(s_evictNvStart + obj_offset + obj_size), obj_size);
|
|
TEST_ASSERT(next_addr == (s_evictNvStart + obj_offset + obj_size));
|
|
|
|
if (!memcmp(evictable_offs_to_addr(obj_offset), obj, obj_size))
|
|
return EC_SUCCESS;
|
|
|
|
return EC_ERROR_INVAL;
|
|
}
|
|
|
|
static int test_tpm_nvmem_modify_evictable_objects(void)
|
|
{
|
|
size_t num_objects;
|
|
uint16_t offsets[MAX_OFFSETS];
|
|
uint32_t handles[ARRAY_SIZE(offsets)];
|
|
uint32_t new_evictable_object[30];
|
|
size_t i;
|
|
const uint32_t new_obj_handle = 0x100;
|
|
static uint8_t modified_obj[CONFIG_FLASH_BANK_SIZE];
|
|
size_t modified_obj_size;
|
|
uint32_t modified_obj_handle;
|
|
uint32_t deleted_obj_handle;
|
|
uint8_t *obj_cache_addr;
|
|
size_t num_handles;
|
|
int new_obj_index;
|
|
int modified_obj_index;
|
|
|
|
TEST_ASSERT(prepare_new_flash() == EC_SUCCESS);
|
|
TEST_ASSERT(new_nvmem_save() == EC_SUCCESS);
|
|
TEST_ASSERT(nvmem_init() == EC_SUCCESS);
|
|
iterate_over_flash();
|
|
|
|
/* Verify that all evictable objects are there. */
|
|
num_objects = fill_obj_offsets(offsets, ARRAY_SIZE(offsets));
|
|
TEST_ASSERT(num_objects == 9);
|
|
num_handles = num_objects;
|
|
|
|
/* Save handles of all objects there are. */
|
|
for (i = 0; i < num_objects; i++) {
|
|
memcpy(handles + i, evictable_offs_to_addr(offsets[i]),
|
|
sizeof(handles[i]));
|
|
ccprintf("obj %d handle %08x\n", i, handles[i]);
|
|
}
|
|
/*
|
|
* Let's modify the object which currently is stored second in the
|
|
* stack.
|
|
*/
|
|
modified_obj_size = offsets[3] - offsets[2] - sizeof(uint32_t);
|
|
|
|
/* Modify the object and copy modified value into local buffer. */
|
|
obj_cache_addr = evictable_offs_to_addr(offsets[2]);
|
|
memcpy(&modified_obj_handle, obj_cache_addr,
|
|
sizeof(modified_obj_handle));
|
|
|
|
for (i = 0; i < modified_obj_size; i++) {
|
|
uint8_t c;
|
|
|
|
c = obj_cache_addr[i];
|
|
|
|
if (i >= sizeof(uint32_t)) { /* Preserve the 4 byte handle. */
|
|
c ^= 0x55;
|
|
obj_cache_addr[i] = c;
|
|
}
|
|
modified_obj[i] = c;
|
|
}
|
|
|
|
/* Save its handle and then drop the object at offset 5. */
|
|
memcpy(&deleted_obj_handle, evictable_offs_to_addr(offsets[5]),
|
|
sizeof(deleted_obj_handle));
|
|
drop_evictable_obj(evictable_offs_to_addr(offsets[5]));
|
|
|
|
/* Prepare the new evictable object, first four bytes are the handle. */
|
|
for (i = 0; i < ARRAY_SIZE(new_evictable_object); i++)
|
|
new_evictable_object[i] = new_obj_handle + i;
|
|
|
|
/* Add it to the cache. */
|
|
add_evictable_obj(new_evictable_object, sizeof(new_evictable_object));
|
|
|
|
/* Save the new cache state in the flash. */
|
|
TEST_ASSERT(new_nvmem_save() == EC_SUCCESS);
|
|
|
|
/* Wipe out NVMEM cache just in case. */
|
|
wipe_out_nvmem_cache();
|
|
|
|
/* Read back from flash into cache. */
|
|
TEST_ASSERT(nvmem_init() == EC_SUCCESS);
|
|
|
|
/* One object removed, one added, the number should have not changed. */
|
|
TEST_ASSERT(num_objects ==
|
|
fill_obj_offsets(offsets, ARRAY_SIZE(offsets)));
|
|
|
|
new_obj_index = 0;
|
|
modified_obj_index = 0;
|
|
for (i = 0; i < num_objects; i++) {
|
|
uint32_t handle;
|
|
size_t j;
|
|
|
|
memcpy(&handle, evictable_offs_to_addr(offsets[i]),
|
|
sizeof(handles[i]));
|
|
ASSERT(handle != deleted_obj_handle);
|
|
|
|
if (handle == new_obj_handle)
|
|
new_obj_index = i;
|
|
else if (handle == modified_obj_handle)
|
|
modified_obj_index = i;
|
|
/*
|
|
* Remove the found handle from the set of handles which were
|
|
* there originally.
|
|
*/
|
|
for (j = 0; j < num_handles; j++)
|
|
if (handles[j] == handle) {
|
|
num_handles--;
|
|
handles[j] = handles[num_handles];
|
|
break;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Removed object's handle is still in the array, and it should be the
|
|
* only remaining element.
|
|
*/
|
|
TEST_ASSERT(num_handles == 1);
|
|
TEST_ASSERT(handles[0] == deleted_obj_handle);
|
|
TEST_ASSERT(new_obj_index >= 0); /* New handle was seen in the cache. */
|
|
TEST_ASSERT(modified_obj_index >=
|
|
0); /* Modified object was seen in the cache. */
|
|
|
|
TEST_ASSERT(compare_object(offsets[new_obj_index],
|
|
sizeof(new_evictable_object),
|
|
new_evictable_object) == EC_SUCCESS);
|
|
TEST_ASSERT(compare_object(offsets[modified_obj_index],
|
|
modified_obj_size,
|
|
modified_obj) == EC_SUCCESS);
|
|
return EC_SUCCESS;
|
|
}
|
|
|
|
static int test_nvmem_tuple_updates(void)
|
|
{
|
|
size_t i;
|
|
const char *modified_var1 = "var one after";
|
|
const struct tuple *t;
|
|
|
|
const struct {
|
|
uint8_t *key;
|
|
uint8_t *value;
|
|
} kv_pairs[] = {
|
|
/* Use # as the delimiter to allow \0 in keys/values. */
|
|
{"key0", "var zero before"},
|
|
{"key1", "var one before"}
|
|
};
|
|
|
|
TEST_ASSERT(post_init_from_scratch(0xff) == EC_SUCCESS);
|
|
|
|
/* Save vars in the nvmem. */
|
|
for (i = 0; i < ARRAY_SIZE(kv_pairs); i++)
|
|
TEST_ASSERT(setvar(kv_pairs[i].key, strlen(kv_pairs[i].key),
|
|
kv_pairs[i].value,
|
|
strlen(kv_pairs[i].value)) == EC_SUCCESS);
|
|
|
|
TEST_ASSERT(nvmem_init() == EC_SUCCESS);
|
|
/* Verify the vars are still there. */
|
|
for (i = 0; i < ARRAY_SIZE(kv_pairs); i++) {
|
|
const struct tuple *t;
|
|
|
|
t = getvar(kv_pairs[i].key, strlen(kv_pairs[i].key));
|
|
TEST_ASSERT(t);
|
|
TEST_ASSERT(t->val_len == strlen(kv_pairs[i].value));
|
|
TEST_ASSERT(!memcmp(t->data_ + strlen(kv_pairs[i].key),
|
|
kv_pairs[i].value, t->val_len));
|
|
freevar(t);
|
|
}
|
|
|
|
/*
|
|
* Now, let's try updating variable 'key1' introducing various failure
|
|
* modes.
|
|
*/
|
|
failure_mode = TEST_FAIL_SAVING_VAR;
|
|
TEST_ASSERT(setvar(kv_pairs[1].key, strlen(kv_pairs[1].key),
|
|
modified_var1, strlen(modified_var1)) == EC_SUCCESS);
|
|
TEST_ASSERT(nvmem_init() == EC_SUCCESS);
|
|
/* No change should be seen. */
|
|
for (i = 0; i < ARRAY_SIZE(kv_pairs); i++) {
|
|
t = getvar(kv_pairs[i].key, strlen(kv_pairs[i].key));
|
|
TEST_ASSERT(t);
|
|
TEST_ASSERT(t->val_len == strlen(kv_pairs[i].value));
|
|
TEST_ASSERT(!memcmp(t->data_ + strlen(kv_pairs[i].key),
|
|
kv_pairs[i].value, t->val_len));
|
|
freevar(t);
|
|
}
|
|
failure_mode = TEST_FAIL_FINALIZING_VAR;
|
|
TEST_ASSERT(setvar(kv_pairs[1].key, strlen(kv_pairs[1].key),
|
|
modified_var1, strlen(modified_var1)) == EC_SUCCESS);
|
|
failure_mode = TEST_NO_FAILURE;
|
|
TEST_ASSERT(nvmem_init() == EC_SUCCESS);
|
|
|
|
/* First variable should be still unchanged. */
|
|
t = getvar(kv_pairs[0].key, strlen(kv_pairs[0].key));
|
|
TEST_ASSERT(t);
|
|
TEST_ASSERT(t->val_len == strlen(kv_pairs[0].value));
|
|
TEST_ASSERT(!memcmp(t->data_ + strlen(kv_pairs[0].key),
|
|
kv_pairs[0].value, t->val_len));
|
|
freevar(t);
|
|
|
|
/* Second variable should be updated. */
|
|
t = getvar(kv_pairs[1].key, strlen(kv_pairs[1].key));
|
|
TEST_ASSERT(t);
|
|
TEST_ASSERT(t->val_len == strlen(modified_var1));
|
|
TEST_ASSERT(!memcmp(t->data_ + strlen(kv_pairs[1].key), modified_var1,
|
|
t->val_len));
|
|
freevar(t);
|
|
|
|
/* A corrupted attempt to update second variable. */
|
|
failure_mode = TEST_FAIL_FINALIZING_VAR;
|
|
TEST_ASSERT(setvar(kv_pairs[1].key, strlen(kv_pairs[1].key),
|
|
kv_pairs[1].value, strlen(kv_pairs[1].value))
|
|
== EC_SUCCESS);
|
|
failure_mode = TEST_NO_FAILURE;
|
|
TEST_ASSERT(nvmem_init() == EC_SUCCESS);
|
|
|
|
/* Is there an instance of the second variable still in the flash. */
|
|
t = getvar(kv_pairs[1].key, strlen(kv_pairs[1].key));
|
|
TEST_ASSERT(t);
|
|
freevar(t);
|
|
|
|
/* Delete the remaining instance of the variable. */
|
|
TEST_ASSERT(setvar(kv_pairs[1].key, strlen(kv_pairs[1].key),
|
|
NULL, 0) == EC_SUCCESS);
|
|
|
|
/* Verify that it is indeed deleted before and after re-init. */
|
|
TEST_ASSERT(!getvar(kv_pairs[1].key, strlen(kv_pairs[1].key)));
|
|
TEST_ASSERT(nvmem_init() == EC_SUCCESS);
|
|
TEST_ASSERT(!getvar(kv_pairs[1].key, strlen(kv_pairs[1].key)));
|
|
|
|
return EC_SUCCESS;
|
|
}
|
|
|
|
void run_test(void)
|
|
{
|
|
run_test_setup();
|
|
|
|
RUN_TEST(test_migration);
|
|
RUN_TEST(test_corrupt_nvmem);
|
|
RUN_TEST(test_fully_erased_nvmem);
|
|
RUN_TEST(test_configured_nvmem);
|
|
RUN_TEST(test_nvmem_save);
|
|
RUN_TEST(test_var_read_write_delete);
|
|
RUN_TEST(test_nvmem_compaction);
|
|
RUN_TEST(test_var_boundaries);
|
|
RUN_TEST(test_nvmem_erase_tpm_data);
|
|
RUN_TEST(test_tpm_nvmem_modify_reserved_objects);
|
|
RUN_TEST(test_tpm_nvmem_modify_evictable_objects);
|
|
RUN_TEST(test_nvmem_incomplete_transaction);
|
|
RUN_TEST(test_nvmem_tuple_updates);
|
|
failure_mode = TEST_NO_FAILURE; /* In case the above test failed. */
|
|
RUN_TEST(test_nvmem_interrupted_compaction);
|
|
failure_mode = TEST_NO_FAILURE; /* In case the above test failed. */
|
|
|
|
/*
|
|
* more tests to come
|
|
* RUN_TEST(test_lock);
|
|
* RUN_TEST(test_malloc_blocking);
|
|
*/
|
|
|
|
test_print_result();
|
|
}
|