From a05a8522ce1dd90c6d667b70bafb24757a27c656 Mon Sep 17 00:00:00 2001 From: Aaron Durbin Date: Fri, 22 Mar 2013 20:44:46 -0500 Subject: lib: add memrange infrastructure The memrange infrastructure allows for keeping track of the machine's physical address space. Each memory_range entry in a memory_ranges structure can be tagged with an arbitrary value. It supports merging and deleting ranges as well as filling in holes in the address space with a particular tag. The memrange infrastructure will serve as a shared implementation for address tracking by the MTRR and coreboot mem table code. Change-Id: Id5bea9d2a419114fca55c59af0fdca063551110e Signed-off-by: Aaron Durbin Reviewed-on: http://review.coreboot.org/2888 Tested-by: build bot (Jenkins) Reviewed-by: Stefan Reinauer --- src/include/memrange.h | 107 ++++++++++++++++ src/lib/Makefile.inc | 1 + src/lib/memrange.c | 323 +++++++++++++++++++++++++++++++++++++++++++++++++ 3 files changed, 431 insertions(+) create mode 100644 src/include/memrange.h create mode 100644 src/lib/memrange.c diff --git a/src/include/memrange.h b/src/include/memrange.h new file mode 100644 index 0000000000..23dae6a30a --- /dev/null +++ b/src/include/memrange.h @@ -0,0 +1,107 @@ +/* + * This file is part of the coreboot project. + * + * Copyright (C) 2013 Google, Inc. + * + * This program is free software; you can redistribute it and/or modify + * it under the terms of the GNU General Public License as published by + * the Free Software Foundation; version 2 of the License. + * + * This program is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the + * GNU General Public License for more details. + * + * You should have received a copy of the GNU General Public License + * along with this program; if not, write to the Free Software + * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA + */ +#ifndef MEMRANGE_H_ +#define MEMRANGE_H_ + +#include + +/* A memranges structure consists of a list of range_entry(s). The structure + * is exposed so that a memranges can be used on the stack if needed. */ +struct memranges { + struct range_entry *entries; +}; + +/* Each region within a memranges structure is represented by a + * range_entry structure. Use the associated range_entry_(base|end|size|tag) + * functions to interrogate its properties. i.e. don't rely on one's own + * interpretation of the fields. */ +struct range_entry { + resource_t begin; + resource_t end; + unsigned long tag; + struct range_entry *next; +}; + +/* Return inclusive base address of memory range. */ +static inline resource_t range_entry_base(const struct range_entry *r) +{ + return r->begin; +} + +/* Return exclusive end address of memory range. */ +static inline resource_t range_entry_end(const struct range_entry *r) +{ + return r->end + 1; +} + +/* Return size of of memory range. */ +static inline resource_t range_entry_size(const struct range_entry *r) +{ + return r->end - r->begin + 1; +} + +static inline unsigned long range_entry_tag(const struct range_entry *r) +{ + return r->tag; +} + +/* Iterate over each entry in a memranges structure. Ranges cannot + * be deleted while processing each entry as the list cannot be safely + * traversed after such an operation. + * r - range_entry pointer. + * ranges - memranges pointer */ +#define memranges_each_entry(r, ranges) \ + for (r = (ranges)->entries; r != NULL; r = r->next) + +/* Initialize and fill a memranges structure according to the + * mask and match type for all memory resources. Tag each entry with the + * specified type. */ +void memranges_init(struct memranges *ranges, + unsigned long mask, unsigned long match, + unsigned long tag); + +/* Remove and free all entries within the memranges structure. */ +void memranges_teardown(struct memranges *ranges); + +/* Add memory resources that match with the corresponding mask and match. + * Each entry will be tagged with the provided tag. e.g. To populate + * all cacheable memory resources in the range: + * memranges_add_resources(range, IORESOURCE_CACHEABLE, + * IORESROUCE_CACHEABLE, my_cacheable_tag); */ +void memranges_add_resources(struct memranges *ranges, + unsigned long mask, unsigned long match, + unsigned long tag); + +/* Fill all address ranges up to limit (exclusive) not covered by an entry by + * inserting new entries with the provided tag. */ +void memranges_fill_holes_up_to(struct memranges *ranges, + resource_t limit, unsigned long tag); + +/* Create a hole in the range by deleting/modifying entries that overlap with + * the region specified by base and size. */ +void memranges_create_hole(struct memranges *ranges, + resource_t base, resource_t size); + +/* Insert a resource to the given memranges. All existing ranges + * covered by range specified by base and size will be removed before a + * new one is added. */ +void memranges_insert(struct memranges *ranges, + resource_t base, resource_t size, unsigned long tag); + +#endif /* MEMRANGE_H_ */ diff --git a/src/lib/Makefile.inc b/src/lib/Makefile.inc index 02c42a8626..9d3588ea60 100644 --- a/src/lib/Makefile.inc +++ b/src/lib/Makefile.inc @@ -86,6 +86,7 @@ ramstage-$(CONFIG_TRACE) += trace.c ramstage-$(CONFIG_COLLECT_TIMESTAMPS) += timestamp.c ramstage-$(CONFIG_COVERAGE) += libgcov.c ramstage-$(CONFIG_MAINBOARD_DO_NATIVE_VGA_INIT) += edid.c +ramstage-y += memrange.c # The CBMEM implementations are chosen based on CONFIG_DYNAMIC_CBMEM. ifeq ($(CONFIG_DYNAMIC_CBMEM),y) diff --git a/src/lib/memrange.c b/src/lib/memrange.c new file mode 100644 index 0000000000..c376d52816 --- /dev/null +++ b/src/lib/memrange.c @@ -0,0 +1,323 @@ +/* + * This file is part of the coreboot project. + * + * Copyright (C) 2013 Google, Inc. + * + * This program is free software; you can redistribute it and/or modify + * it under the terms of the GNU General Public License as published by + * the Free Software Foundation; version 2 of the License. + * + * This program is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the + * GNU General Public License for more details. + * + * You should have received a copy of the GNU General Public License + * along with this program; if not, write to the Free Software + * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA + */ +#include +#include +#include + +/* Coreboot doesn't have a free() function. Therefore, keep a cache of + * free'd entries. */ +static struct range_entry *free_list; + +static inline void range_entry_link(struct range_entry **prev_ptr, + struct range_entry *r) +{ + r->next = *prev_ptr; + *prev_ptr = r; +} + +static inline void range_entry_unlink(struct range_entry **prev_ptr, + struct range_entry *r) +{ + *prev_ptr = r->next; + r->next = NULL; +} + +static inline void range_entry_unlink_and_free(struct range_entry **prev_ptr, + struct range_entry *r) +{ + range_entry_unlink(prev_ptr, r); + range_entry_link(&free_list, r); +} + +static struct range_entry *alloc_range(void) +{ + if (free_list != NULL) { + struct range_entry *r; + + r = free_list; + range_entry_unlink(&free_list, r); + return r; + } + return malloc(sizeof(struct range_entry)); +} + +static inline struct range_entry * +range_list_add(struct range_entry **prev_ptr, resource_t begin, resource_t end, + unsigned long tag) +{ + struct range_entry *new_entry; + + new_entry = alloc_range(); + if (new_entry == NULL) { + printk(BIOS_ERR, "Could not allocate range_entry!\n"); + return NULL; + } + new_entry->begin = begin; + new_entry->end = end; + new_entry->tag = tag; + range_entry_link(prev_ptr, new_entry); + + return new_entry; +} + +static void merge_neighbor_entries(struct memranges *ranges) +{ + struct range_entry *cur; + struct range_entry *prev; + + prev = NULL; + /* Merge all neighbors and delete/free the leftover entry. */ + for (cur = ranges->entries; cur != NULL; cur = cur->next) { + /* First entry. Just set prev. */ + if (prev == NULL) { + prev = cur; + continue; + } + + /* If the previous entry merges with the current update the + * previous entry to cover full range and delete current from + * the list. */ + if (prev->end + 1 >= cur->begin && prev->tag == cur->tag) { + prev->end = cur->end; + range_entry_unlink_and_free(&prev->next, cur); + /* Set cur to prev so cur->next is valid since cur + * was just unlinked and free. */ + cur = prev; + continue; + } + + prev = cur; + } +} + +static void remove_memranges(struct memranges *ranges, + resource_t begin, resource_t end, + unsigned long unused) +{ + struct range_entry *cur; + struct range_entry *next; + struct range_entry **prev_ptr; + + prev_ptr = &ranges->entries; + for (cur = ranges->entries; cur != NULL; cur = next) { + resource_t tmp_end; + + /* Cache the next value to handle unlinks. */ + next = cur->next; + + /* No other ranges are affected. */ + if (end < cur->begin) + break; + + /* The removal range starts after this one. */ + if (begin > cur->end) { + prev_ptr = &cur->next; + continue; + } + + /* The removal range overlaps with the current entry either + * partially or fully. However, we need to adjust the removal + * range for any holes. */ + if (begin <= cur->begin) { + begin = cur->begin; + + /* Full removal. */ + if (end >= cur->end) { + begin = cur->end + 1; + range_entry_unlink_and_free(prev_ptr, cur); + continue; + } + } + + /* prev_ptr can be set now that the unlink path wasn't taken. */ + prev_ptr = &cur->next; + + /* Clip the end fragment to do proper splitting. */ + tmp_end = end; + if (end > cur->end) + tmp_end = cur->end; + + /* Hole punched in middle of entry. */ + if (begin > cur->begin && tmp_end < cur->end) { + range_list_add(&cur->next, end + 1, cur->end, cur->tag); + cur->end = begin - 1; + break; + } + + /* Removal at beginning. */ + if (begin == cur->begin) + cur->begin = tmp_end + 1; + + /* Removal at end. */ + if (tmp_end == cur->end) + cur->end = begin - 1; + } +} + +static void merge_add_memranges(struct memranges *ranges, + resource_t begin, resource_t end, + unsigned long tag) +{ + struct range_entry *cur; + struct range_entry **prev_ptr; + + prev_ptr = &ranges->entries; + + /* Remove all existing entries covered by the range. */ + remove_memranges(ranges, begin, end, -1); + + /* Find the entry to place the new entry after. Since + * remove_memranges() was called above there is a guranteed + * spot for this new entry. */ + for (cur = ranges->entries; cur != NULL; cur = cur->next) { + /* Found insertion spot before current entry. */ + if (end < cur->begin) + break; + + /* Keep track of previous entry to insert new entry after it. */ + prev_ptr = &cur->next; + + /* The new entry starts after this one. */ + if (begin > cur->end) + continue; + + } + + /* Add new entry and merge with neighbors. */ + range_list_add(prev_ptr, begin, end, tag); + merge_neighbor_entries(ranges); +} + +typedef void (*range_action_t)(struct memranges *ranges, + resource_t begin, resource_t end, + unsigned long tag); + +static void do_action(struct memranges *ranges, + resource_t base, resource_t size, unsigned long tag, + range_action_t action) +{ + resource_t end; + resource_t begin; + + /* The addresses are aligned to 4096 bytes: the begin address is + * aligned down while the end address is aligned up to be conservative + * about the full range covered. */ + begin = ALIGN_DOWN(base, 4096); + end = begin + size + (base - begin); + end = ALIGN_UP(end, 4096) - 1; + action(ranges, begin, end, tag); +} + +void memranges_create_hole(struct memranges *ranges, + resource_t base, resource_t size) +{ + do_action(ranges, base, size, -1, remove_memranges); +} + +void memranges_insert(struct memranges *ranges, + resource_t base, resource_t size, unsigned long tag) +{ + do_action(ranges, base, size, tag, merge_add_memranges); +} + +struct collect_context { + struct memranges *ranges; + unsigned long tag; +}; + +static void collect_ranges(void *gp, struct device *dev, struct resource *res) +{ + struct collect_context *ctx = gp; + + memranges_insert(ctx->ranges, res->base, res->size, ctx->tag); +} + +void memranges_add_resources(struct memranges *ranges, + unsigned long mask, unsigned long match, + unsigned long tag) +{ + struct collect_context context; + + /* Only deal with MEM resources. */ + mask |= IORESOURCE_MEM; + match |= IORESOURCE_MEM; + + context.ranges = ranges; + context.tag = tag; + search_global_resources(mask, match, collect_ranges, &context); +} + +void memranges_init(struct memranges *ranges, + unsigned long mask, unsigned long match, + unsigned long tag) +{ + ranges->entries = NULL; + memranges_add_resources(ranges, mask, match, tag); +} + +void memranges_teardown(struct memranges *ranges) +{ + while (ranges->entries != NULL) { + range_entry_unlink_and_free(&ranges->entries, ranges->entries); + } +} + +void memranges_fill_holes_up_to(struct memranges *ranges, + resource_t limit, unsigned long tag) +{ + struct range_entry *cur; + struct range_entry *prev; + + prev = NULL; + for (cur = ranges->entries; cur != NULL; cur = cur->next) { + /* First entry. Just set prev. */ + if (prev == NULL) { + prev = cur; + continue; + } + + /* If the previous entry does not directly preceed the current + * entry then add a new entry just after the previous one. */ + if (range_entry_end(prev) != cur->begin) { + resource_t end; + + end = cur->begin - 1; + if (end >= limit) + end = limit - 1; + range_list_add(&prev->next, range_entry_end(prev), + end, tag); + } + + prev = cur; + + /* Hit the requested range limit. No other entries after this + * are affected. */ + if (cur->begin >= limit) + break; + } + + /* Handle the case where the limit was never reached. A new entry needs + * to be added to cover the range up to the limit. */ + if (prev != NULL && range_entry_end(prev) < limit) + range_list_add(&prev->next, range_entry_end(prev), + limit - 1, tag); + + /* Merge all entries that were newly added. */ + merge_neighbor_entries(ranges); +} -- cgit v1.2.3