diff options
Diffstat (limited to 'src/lib/imd_cbmem.c')
-rw-r--r-- | src/lib/imd_cbmem.c | 294 |
1 files changed, 294 insertions, 0 deletions
diff --git a/src/lib/imd_cbmem.c b/src/lib/imd_cbmem.c new file mode 100644 index 0000000000..0649bf3b17 --- /dev/null +++ b/src/lib/imd_cbmem.c @@ -0,0 +1,294 @@ +/* + * This file is part of the coreboot project. + * + * Copyright (C) 2013 Google, Inc. + * + * This program is free software; you can redistribute it and/or modify + * it under the terms of the GNU General Public License as published by + * the Free Software Foundation; version 2 of the License. + * + * This program is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the + * GNU General Public License for more details. + * + * You should have received a copy of the GNU General Public License + * along with this program; if not, write to the Free Software + * Foundation, Inc. + */ + +#include <bootstate.h> +#include <bootmem.h> +#include <console/console.h> +#include <cbmem.h> +#include <imd.h> +#include <rules.h> +#include <string.h> +#include <stdlib.h> +#include <arch/early_variables.h> +#if IS_ENABLED(CONFIG_ARCH_X86) && !IS_ENABLED(CONFIG_EARLY_CBMEM_INIT) +#include <arch/acpi.h> +#endif + +/* The root region is at least DYN_CBMEM_ALIGN_SIZE . */ +#define ROOT_MIN_SIZE DYN_CBMEM_ALIGN_SIZE +#define LG_ALIGN ROOT_MIN_SIZE +/* Small allocation parameters. */ +#define SM_ROOT_SIZE 1024 +#define SM_ALIGN 32 + +static inline struct imd *cbmem_get_imd(void) +{ + /* Only supply a backing store for imd in ramstage. */ + if (ENV_RAMSTAGE) { + static struct imd imd_cbmem; + return &imd_cbmem; + } + return NULL; +} + +/* + * x86 !CONFIG_EARLY_CBMEM_INIT platforms need to do the following in ramstage: + * 1. Call set_top_of_ram() which in turn calls cbmem_set_top(). + * 2. Provide a get_top_of_ram() implementation. + * + * CONFIG_EARLY_CBMEM_INIT platforms just need to provide cbmem_top(). + */ +void cbmem_set_top(void *ramtop) +{ + struct imd *imd = cbmem_get_imd(); + + imd_handle_init(imd, ramtop); +} + +static inline const struct cbmem_entry *imd_to_cbmem(const struct imd_entry *e) +{ + return (const struct cbmem_entry *)e; +} + +static inline const struct imd_entry *cbmem_to_imd(const struct cbmem_entry *e) +{ + return (const struct imd_entry *)e; +} + +/* These are the different situations to handle: + * CONFIG_EARLY_CBMEM_INIT: + * In ramstage cbmem_initialize() attempts a recovery of the + * cbmem region set up by romstage. It uses cbmem_top() as the + * starting point of recovery. + * + * In romstage, similar to ramstage, cbmem_initialize() needs to + * attempt recovery of the cbmem area using cbmem_top() as the limit. + * cbmem_initialize_empty() initializes an empty cbmem area from + * cbmem_top(); + * + */ +static struct imd *imd_init_backing(struct imd *backing) +{ + struct imd *imd; + + imd = cbmem_get_imd(); + + if (imd != NULL) + return imd; + + imd = backing; + + return imd; +} + +static struct imd *imd_init_backing_with_recover(struct imd *backing) +{ + struct imd *imd; + + imd = imd_init_backing(backing); + if (!ENV_RAMSTAGE) { + /* Early cbmem init platforms need to always use cbmem_top(). */ + if (IS_ENABLED(CONFIG_EARLY_CBMEM_INIT)) + imd_handle_init(imd, cbmem_top()); + /* Need to partially recover all the time outside of ramstage + * because there's object storage outside of the stack. */ + imd_handle_init_partial_recovery(imd); + } + + return imd; +} + +void cbmem_initialize_empty(void) +{ + struct imd *imd; + struct imd imd_backing; + + imd = imd_init_backing(&imd_backing); + + /* Early cbmem init platforms need to always use cbmem_top(). */ + if (IS_ENABLED(CONFIG_EARLY_CBMEM_INIT)) + imd_handle_init(imd, cbmem_top()); + + printk(BIOS_DEBUG, "CBMEM:\n"); + + if (imd_create_tiered_empty(imd, ROOT_MIN_SIZE, LG_ALIGN, + SM_ROOT_SIZE, SM_ALIGN)) { + printk(BIOS_DEBUG, "failed.\n"); + return; + } + + /* Complete migration to CBMEM. */ + cbmem_run_init_hooks(); +} + +static inline int cbmem_fail_recovery(void) +{ + cbmem_initialize_empty(); + cbmem_fail_resume(); + return 1; +} + +int cbmem_initialize(void) +{ + struct imd *imd; + struct imd imd_backing; + + imd = imd_init_backing(&imd_backing); + + /* Early cbmem init platforms need to always use cbmem_top(). */ + if (IS_ENABLED(CONFIG_EARLY_CBMEM_INIT)) + imd_handle_init(imd, cbmem_top()); + + if (imd_recover(imd)) + return 1; + +#if defined(__PRE_RAM__) + /* + * Lock the imd in romstage on a recovery. The assumption is that + * if the imd area was recovered in romstage then S3 resume path + * is being taken. + */ + imd_lockdown(imd); +#endif + + /* Complete migration to CBMEM. */ + cbmem_run_init_hooks(); + + /* Recovery successful. */ + return 0; +} + +int cbmem_recovery(int is_wakeup) +{ + int rv = 0; + if (!is_wakeup) + cbmem_initialize_empty(); + else + rv = cbmem_initialize(); + return rv; +} + +const struct cbmem_entry *cbmem_entry_add(u32 id, u64 size64) +{ + struct imd *imd; + struct imd imd_backing; + const struct imd_entry *e; + + imd = imd_init_backing_with_recover(&imd_backing); + + e = imd_entry_find_or_add(imd, id, size64); + + return imd_to_cbmem(e); +} + +void *cbmem_add(u32 id, u64 size) +{ + struct imd *imd; + struct imd imd_backing; + const struct imd_entry *e; + + imd = imd_init_backing_with_recover(&imd_backing); + + e = imd_entry_find_or_add(imd, id, size); + + if (e == NULL) + return NULL; + + return imd_entry_at(imd, e); +} + +/* Retrieve a region provided a given id. */ +const struct cbmem_entry *cbmem_entry_find(u32 id) +{ + struct imd *imd; + struct imd imd_backing; + const struct imd_entry *e; + + imd = imd_init_backing_with_recover(&imd_backing); + + e = imd_entry_find(imd, id); + + return imd_to_cbmem(e); +} + +void *cbmem_find(u32 id) +{ + struct imd *imd; + struct imd imd_backing; + const struct imd_entry *e; + + imd = imd_init_backing_with_recover(&imd_backing); + + e = imd_entry_find(imd, id); + + if (e == NULL) + return NULL; + + return imd_entry_at(imd, e); +} + +/* Remove a reserved region. Returns 0 on success, < 0 on error. Note: A region + * cannot be removed unless it was the last one added. */ +int cbmem_entry_remove(const struct cbmem_entry *entry) +{ + struct imd *imd; + struct imd imd_backing; + + imd = imd_init_backing_with_recover(&imd_backing); + + return imd_entry_remove(imd, cbmem_to_imd(entry)); +} + +u64 cbmem_entry_size(const struct cbmem_entry *entry) +{ + struct imd *imd; + struct imd imd_backing; + + imd = imd_init_backing_with_recover(&imd_backing); + + return imd_entry_size(imd, cbmem_to_imd(entry)); +} + +void *cbmem_entry_start(const struct cbmem_entry *entry) +{ + struct imd *imd; + struct imd imd_backing; + + imd = imd_init_backing_with_recover(&imd_backing); + + return imd_entry_at(imd, cbmem_to_imd(entry)); +} + +#if ENV_RAMSTAGE +void cbmem_add_bootmem(void) +{ + void *base = NULL; + size_t size = 0; + + imd_region_used(cbmem_get_imd(), &base, &size); + bootmem_add_range((uintptr_t)base, size, LB_MEM_TABLE); +} + +void cbmem_list(void) +{ + static const struct imd_lookup lookup[] = { CBMEM_ID_TO_NAME_TABLE }; + + imd_print_entries(cbmem_get_imd(), lookup, ARRAY_SIZE(lookup)); +} +#endif /* __PRE_RAM__ */ |