summaryrefslogtreecommitdiff
path: root/src/soc/amd
diff options
context:
space:
mode:
authorJason Glenesk <jason.glenesk@amd.corp-partner.google.com>2020-12-10 04:22:56 -0800
committerPatrick Georgi <pgeorgi@google.com>2021-01-15 11:27:23 +0000
commit276e865e7fc4c6c7a7fa30e1786fff373b0ae7a1 (patch)
treef7fd3bc1f0812324bf541802409f75b6187423a7 /src/soc/amd
parentfd104e1c5d238c4d6d28b59bf3c3b0be2f3b420d (diff)
soc/amd/picasso: Generate ACPI CRAT objects in cb
Add code to collect all required information and generate ACPI CRAT table entries. Publish tables generated from cb, rather than use the tables created by FSP binary. BUG=b:155307433 TEST=Boot trembyle and compare coreboot generated tables with tables that FSP published previously. BRANCH=Zork Change-Id: If64fd624597b2ced014ba7f0332a6a48143c0e8c Signed-off-by: Jason Glenesk <jason.glenesk@amd.corp-partner.google.com> Reviewed-on: https://review.coreboot.org/c/coreboot/+/47727 Tested-by: build bot (Jenkins) <no-reply@coreboot.org> Reviewed-by: Marshall Dawson <marshalldawson3rd@gmail.com> Reviewed-by: Angel Pons <th3fanbus@gmail.com>
Diffstat (limited to 'src/soc/amd')
-rw-r--r--src/soc/amd/picasso/agesa_acpi.c534
-rw-r--r--src/soc/amd/picasso/include/soc/data_fabric.h32
2 files changed, 564 insertions, 2 deletions
diff --git a/src/soc/amd/picasso/agesa_acpi.c b/src/soc/amd/picasso/agesa_acpi.c
index ec9924bd26..23d10ff039 100644
--- a/src/soc/amd/picasso/agesa_acpi.c
+++ b/src/soc/amd/picasso/agesa_acpi.c
@@ -1,14 +1,19 @@
/* SPDX-License-Identifier: GPL-2.0-only */
#include <acpi/acpi.h>
+#include <acpi/acpi_crat.h>
#include <acpi/acpi_ivrs.h>
#include <console/console.h>
+#include <cpu/amd/cpuid.h>
+#include <cpu/amd/msr.h>
#include <fsp/util.h>
#include <FspGuids.h>
#include <soc/acpi.h>
#include <stdint.h>
#include <device/pci_def.h>
#include <device/pci_ops.h>
+#include <soc/cpu.h>
+#include <soc/data_fabric.h>
#include <soc/pci_devs.h>
#include <stdlib.h>
#include <arch/mmio.h>
@@ -493,14 +498,539 @@ static unsigned long acpi_fill_ivrs(acpi_ivrs_t *ivrs, unsigned long current)
return acpi_fill_ivrs11(current, ivrs);
}
+static unsigned long gen_crat_hsa_entry(struct acpi_crat_header *crat, unsigned long current)
+{
+ struct crat_hsa_processing_unit *hsa_entry = (struct crat_hsa_processing_unit *)current;
+ memset(hsa_entry, 0, sizeof(struct crat_hsa_processing_unit));
+
+ hsa_entry->flags = CRAT_HSA_PR_FLAG_EN | CRAT_HSA_PR_FLAG_CPU_PRES;
+ hsa_entry->wave_front_size = 4;
+ hsa_entry->num_cpu_cores = get_cpu_count();
+ hsa_entry->length = sizeof(struct crat_hsa_processing_unit);
+ crat->total_entries++;
+
+ current += hsa_entry->length;
+ return current;
+}
+
+static unsigned long create_crat_memory_entry(uint32_t domain, uint64_t region_base,
+ uint64_t region_size, unsigned long current)
+{
+ struct crat_memory *mem_affinity = (struct crat_memory *)current;
+ memset(mem_affinity, 0, sizeof(struct crat_memory));
+
+ mem_affinity->type = CRAT_MEMORY_TYPE;
+ mem_affinity->length = sizeof(struct crat_memory);
+ mem_affinity->proximity_domain = 0;
+ mem_affinity->base_address_low = region_base & 0xffffffff;
+ mem_affinity->base_address_high = (region_base >> 32) & 0xffffffff;
+ mem_affinity->length_low = region_size & 0xffffffff;
+ mem_affinity->length_high = (region_size >> 32) & 0xffffffff;
+ mem_affinity->flags = CRAT_MEM_FLAG_EN;
+ mem_affinity->width = 64;
+
+ current += mem_affinity->length;
+ return current;
+}
+
+static unsigned long gen_crat_memory_entries(struct acpi_crat_header *crat,
+ unsigned long current)
+{
+ uint32_t dram_base_reg, dram_limit_reg, dram_hole_ctl;
+ uint64_t memory_length, memory_base, hole_base, size_below_hole;
+ size_t new_entries = 0;
+
+ for (size_t dram_map_idx = 0; dram_map_idx < PICASSO_NUM_DRAM_REG;
+ dram_map_idx++) {
+ dram_base_reg =
+ data_fabric_read_reg32(0, DF_DRAM_BASE(dram_map_idx), IOMS0_FABRIC_ID);
+
+ if (dram_base_reg & DRAM_BASE_REG_VALID) {
+ dram_limit_reg = data_fabric_read_reg32(0, DF_DRAM_LIMIT(dram_map_idx),
+ IOMS0_FABRIC_ID);
+ memory_length =
+ ((dram_limit_reg & DRAM_LIMIT_ADDR) >> DRAM_LIMIT_ADDR_SHFT) + 1
+ - ((dram_base_reg & DRAM_BASE_ADDR) >> DRAM_BASE_ADDR_SHFT);
+ memory_length = memory_length << 28;
+ memory_base = (dram_base_reg & DRAM_BASE_ADDR)
+ << (28 - DRAM_BASE_ADDR_SHFT);
+
+ if (memory_base == 0) {
+ current =
+ create_crat_memory_entry(0, 0ull, 0xa0000ull, current);
+ memory_base = (1 * 1024 * 1024);
+ memory_length = memory_base;
+ new_entries++;
+ }
+
+ if (dram_base_reg & DRAM_BASE_HOLE_EN) {
+ dram_hole_ctl = data_fabric_read_reg32(0, D18F0_DRAM_HOLE_CTL,
+ IOMS0_FABRIC_ID);
+ hole_base = (dram_hole_ctl & DRAM_HOLE_CTL_BASE);
+ size_below_hole = hole_base - memory_base;
+ current = create_crat_memory_entry(0, memory_base,
+ size_below_hole, current);
+ memory_length = (((dram_limit_reg & DRAM_LIMIT_ADDR)
+ >> DRAM_LIMIT_ADDR_SHFT)
+ + 1 - 0x10)
+ << 28;
+ memory_base = 0x100000000;
+ new_entries++;
+ }
+
+ current = create_crat_memory_entry(0, memory_base, memory_length,
+ current);
+ new_entries++;
+ }
+ }
+ crat->total_entries += new_entries;
+ return current;
+}
+
+static unsigned long add_crat_cache_entry(struct crat_cache **cache_affinity,
+ unsigned long current)
+{
+ *cache_affinity = (struct crat_cache *)current;
+ memset(*cache_affinity, 0, sizeof(struct crat_cache));
+
+ (*cache_affinity)->type = CRAT_CACHE_TYPE;
+ (*cache_affinity)->length = sizeof(struct crat_cache);
+ (*cache_affinity)->flags = CRAT_CACHE_FLAG_EN | CRAT_CACHE_FLAG_CPU_CACHE;
+
+ current += sizeof(struct crat_cache);
+ return current;
+}
+
+static uint8_t get_associativity(uint32_t encoded_associativity)
+{
+ uint8_t associativity = 0;
+
+ switch (encoded_associativity) {
+ case 0:
+ case 1:
+ case 2:
+ case 3:
+ case 4:
+ return encoded_associativity;
+ case 5:
+ associativity = 6;
+ break;
+ case 6:
+ associativity = 8;
+ break;
+ case 8:
+ associativity = 16;
+ break;
+ case 0xA:
+ associativity = 32;
+ break;
+ case 0xB:
+ associativity = 48;
+ break;
+ case 0xC:
+ associativity = 64;
+ break;
+ case 0xD:
+ associativity = 96;
+ break;
+ case 0xE:
+ associativity = 128;
+ break;
+ case 0xF:
+ associativity = 0xFF;
+ break;
+ default:
+ return 0;
+ }
+
+ return associativity;
+}
+
+static unsigned long gen_crat_cache_entry(struct acpi_crat_header *crat, unsigned long current)
+{
+ size_t total_num_threads, num_threads_sharing0, num_threads_sharing1,
+ num_threads_sharing2, num_threads_sharing3, thread, new_entries;
+ struct cpuid_result cache_props0, cache_props1, cache_props2, cache_props3;
+ uint8_t sibling_mask = 0;
+ uint32_t l1_data_cache_ids, l1_inst_cache_ids, l2_cache_ids, l3_cache_ids;
+ struct crat_cache *cache_affinity = NULL;
+
+ total_num_threads = get_cpu_count();
+
+ cache_props0 = cpuid_ext(CPUID_CACHE_PROPS, CACHE_PROPS_0);
+ cache_props1 = cpuid_ext(CPUID_CACHE_PROPS, CACHE_PROPS_1);
+ cache_props2 = cpuid_ext(CPUID_CACHE_PROPS, CACHE_PROPS_2);
+ cache_props3 = cpuid_ext(CPUID_CACHE_PROPS, CACHE_PROPS_3);
+
+ l1_data_cache_ids = cpuid_ecx(CPUID_L1_TLB_CACHE_IDS);
+ l1_inst_cache_ids = cpuid_edx(CPUID_L1_TLB_CACHE_IDS);
+ l2_cache_ids = cpuid_ecx(CPUID_L2_L3_CACHE_L2_TLB_IDS);
+ l3_cache_ids = cpuid_edx(CPUID_L2_L3_CACHE_L2_TLB_IDS);
+
+ num_threads_sharing0 =
+ ((cache_props0.eax & NUM_SHARE_CACHE_MASK) >> NUM_SHARE_CACHE_SHFT) + 1;
+ num_threads_sharing1 =
+ ((cache_props1.eax & NUM_SHARE_CACHE_MASK) >> NUM_SHARE_CACHE_SHFT) + 1;
+ num_threads_sharing2 =
+ ((cache_props2.eax & NUM_SHARE_CACHE_MASK) >> NUM_SHARE_CACHE_SHFT) + 1;
+ num_threads_sharing3 =
+ ((cache_props3.eax & NUM_SHARE_CACHE_MASK) >> NUM_SHARE_CACHE_SHFT) + 1;
+
+ new_entries = 0;
+ for (thread = 0; thread < total_num_threads; thread++) {
+ /* L1 data cache */
+ if (thread % num_threads_sharing0 == 0) {
+ current = add_crat_cache_entry(&cache_affinity, current);
+ new_entries++;
+
+ cache_affinity->flags |= CRAT_CACHE_FLAG_DATA_CACHE;
+ cache_affinity->proc_id_low = thread;
+ sibling_mask = 1;
+ for (size_t sibling = 1; sibling < num_threads_sharing0; sibling++)
+ sibling_mask = (sibling_mask << 1) + 1;
+ cache_affinity->sibling_map[thread / 8] = sibling_mask << (thread % 8);
+ cache_affinity->cache_properties =
+ (cache_props0.edx & CACHE_INCLUSIVE_MASK) ? 2 : 0;
+ cache_affinity->cache_size =
+ (l1_data_cache_ids & L1_DC_SIZE_MASK) >> L1_DC_SIZE_SHFT;
+ cache_affinity->cache_level = CRAT_L1_CACHE;
+ cache_affinity->lines_per_tag =
+ (l1_data_cache_ids & L1_DC_LINE_TAG_MASK)
+ >> L1_DC_LINE_TAG_SHFT;
+ cache_affinity->cache_line_size =
+ (l1_data_cache_ids & L1_DC_LINE_SIZE_MASK)
+ >> L1_DC_LINE_SIZE_SHFT;
+ cache_affinity->associativity =
+ (l1_data_cache_ids & L1_DC_ASSOC_MASK) >> L1_DC_ASSOC_SHFT;
+ cache_affinity->cache_latency = 1;
+ }
+
+ /* L1 instruction cache */
+ if (thread % num_threads_sharing1 == 0) {
+ current = add_crat_cache_entry(&cache_affinity, current);
+ new_entries++;
+
+ cache_affinity->flags |= CRAT_CACHE_FLAG_INSTR_CACHE;
+ cache_affinity->proc_id_low = thread;
+ sibling_mask = 1;
+ for (size_t sibling = 1; sibling < num_threads_sharing1; sibling++)
+ sibling_mask = (sibling_mask << 1) + 1;
+ cache_affinity->sibling_map[thread / 8] = sibling_mask << (thread % 8);
+ cache_affinity->cache_properties =
+ (cache_props1.edx & CACHE_INCLUSIVE_MASK) ? 2 : 0;
+ cache_affinity->cache_size =
+ (l1_inst_cache_ids & L1_IC_SIZE_MASK) >> L1_IC_SIZE_SHFT;
+ cache_affinity->cache_level = CRAT_L1_CACHE;
+ cache_affinity->lines_per_tag =
+ (l1_inst_cache_ids & L1_IC_LINE_TAG_MASK)
+ >> L1_IC_LINE_TAG_SHFT;
+ cache_affinity->cache_line_size =
+ (l1_inst_cache_ids & L1_IC_LINE_SIZE_MASK)
+ >> L1_IC_LINE_SIZE_SHFT;
+ cache_affinity->associativity =
+ (l1_inst_cache_ids & L1_IC_ASSOC_MASK) >> L1_IC_ASSOC_SHFT;
+ cache_affinity->cache_latency = 1;
+ }
+
+ /* L2 cache */
+ if (thread % num_threads_sharing2 == 0) {
+ current = add_crat_cache_entry(&cache_affinity, current);
+ new_entries++;
+
+ cache_affinity->flags |=
+ CRAT_CACHE_FLAG_DATA_CACHE | CRAT_CACHE_FLAG_INSTR_CACHE;
+ cache_affinity->proc_id_low = thread;
+ sibling_mask = 1;
+ for (size_t sibling = 1; sibling < num_threads_sharing2; sibling++)
+ sibling_mask = (sibling_mask << 1) + 1;
+ cache_affinity->sibling_map[thread / 8] = sibling_mask << (thread % 8);
+ cache_affinity->cache_properties =
+ (cache_props2.edx & CACHE_INCLUSIVE_MASK) ? 2 : 0;
+ cache_affinity->cache_size =
+ (l2_cache_ids & L2_DC_SIZE_MASK) >> L2_DC_SIZE_SHFT;
+ cache_affinity->cache_level = CRAT_L2_CACHE;
+ cache_affinity->lines_per_tag =
+ (l2_cache_ids & L2_DC_LINE_TAG_MASK) >> L2_DC_LINE_TAG_SHFT;
+ cache_affinity->cache_line_size =
+ (l2_cache_ids & L2_DC_LINE_SIZE_MASK) >> L2_DC_LINE_SIZE_SHFT;
+ cache_affinity->associativity = get_associativity(
+ (l2_cache_ids & L2_DC_ASSOC_MASK) >> L2_DC_ASSOC_SHFT);
+ cache_affinity->cache_latency = 1;
+ }
+
+ /* L3 cache */
+ if (thread % num_threads_sharing3 == 0) {
+ current = add_crat_cache_entry(&cache_affinity, current);
+ new_entries++;
+
+ cache_affinity->flags |=
+ CRAT_CACHE_FLAG_DATA_CACHE | CRAT_CACHE_FLAG_INSTR_CACHE;
+ cache_affinity->proc_id_low = thread;
+ sibling_mask = 1;
+ for (size_t sibling = 1; sibling < num_threads_sharing3; sibling++)
+ sibling_mask = (sibling_mask << 1) + 1;
+ cache_affinity->sibling_map[thread / 8] = sibling_mask << (thread % 8);
+ cache_affinity->cache_properties =
+ (cache_props0.edx & CACHE_INCLUSIVE_MASK) ? 2 : 0;
+ cache_affinity->cache_size =
+ ((l3_cache_ids & L3_DC_SIZE_MASK) >> L3_DC_SIZE_SHFT) * 512;
+ cache_affinity->cache_level = CRAT_L3_CACHE;
+ cache_affinity->lines_per_tag =
+ (l3_cache_ids & L3_DC_LINE_TAG_MASK) >> L3_DC_LINE_TAG_SHFT;
+ cache_affinity->cache_line_size =
+ (l3_cache_ids & L3_DC_LINE_SIZE_MASK) >> L3_DC_LINE_SIZE_SHFT;
+ cache_affinity->associativity = get_associativity(
+ (l3_cache_ids & L3_DC_ASSOC_MASK) >> L3_DC_ASSOC_SHFT);
+ cache_affinity->cache_latency = 1;
+ }
+ }
+ crat->total_entries += new_entries;
+ return current;
+}
+
+static uint8_t get_tlb_size(enum tlb_type type, struct crat_tlb *crat_tlb_entry,
+ uint16_t raw_assoc_size)
+{
+ uint8_t tlbsize;
+
+ if (raw_assoc_size >= 256) {
+ tlbsize = (uint8_t)(raw_assoc_size / 256);
+
+ if (type == tlb_2m)
+ crat_tlb_entry->flags |= CRAT_TLB_FLAG_2MB_BASE_256;
+ else if (type == tlb_4k)
+ crat_tlb_entry->flags |= CRAT_TLB_FLAG_4K_BASE_256;
+ else if (type == tlb_1g)
+ crat_tlb_entry->flags |= CRAT_TLB_FLAG_1GB_BASE_256;
+ } else {
+ tlbsize = (uint8_t)(raw_assoc_size);
+ }
+ return tlbsize;
+}
+
+static unsigned long add_crat_tlb_entry(struct crat_tlb **tlb_affinity, unsigned long current)
+{
+ *tlb_affinity = (struct crat_tlb *)current;
+ memset(*tlb_affinity, 0, sizeof(struct crat_tlb));
+
+ (*tlb_affinity)->type = CRAT_TLB_TYPE;
+ (*tlb_affinity)->length = sizeof(struct crat_tlb);
+ (*tlb_affinity)->flags = CRAT_TLB_FLAG_EN | CRAT_TLB_FLAG_CPU_TLB;
+
+ current += sizeof(struct crat_tlb);
+ return current;
+}
+
+static unsigned long gen_crat_tlb_entry(struct acpi_crat_header *crat, unsigned long current)
+{
+ size_t total_num_threads, num_threads_sharing0, num_threads_sharing1,
+ num_threads_sharing2, thread, new_entries;
+ struct cpuid_result cache_props0, cache_props1, cache_props2;
+ uint8_t sibling_mask = 0;
+ uint32_t l1_tlb_2M4M_ids, l1_tlb_4K_ids, l2_tlb_2M4M_ids, l2_tlb_4K_ids, l1_tlb_1G_ids,
+ l2_tlb_1G_ids;
+ struct crat_tlb *tlb_affinity = NULL;
+
+ total_num_threads = get_cpu_count();
+ cache_props0 = cpuid_ext(CPUID_CACHE_PROPS, CACHE_PROPS_0);
+ cache_props1 = cpuid_ext(CPUID_CACHE_PROPS, CACHE_PROPS_1);
+ cache_props2 = cpuid_ext(CPUID_CACHE_PROPS, CACHE_PROPS_2);
+
+ l1_tlb_2M4M_ids = cpuid_eax(CPUID_L1_TLB_CACHE_IDS);
+ l2_tlb_2M4M_ids = cpuid_eax(CPUID_L2_L3_CACHE_L2_TLB_IDS);
+ l1_tlb_4K_ids = cpuid_ebx(CPUID_L1_TLB_CACHE_IDS);
+ l2_tlb_4K_ids = cpuid_ebx(CPUID_L2_L3_CACHE_L2_TLB_IDS);
+ l1_tlb_1G_ids = cpuid_eax(CPUID_TLB_L1L2_1G_IDS);
+ l2_tlb_1G_ids = cpuid_ebx(CPUID_TLB_L1L2_1G_IDS);
+
+ num_threads_sharing0 =
+ ((cache_props0.eax & NUM_SHARE_CACHE_MASK) >> NUM_SHARE_CACHE_SHFT) + 1;
+ num_threads_sharing1 =
+ ((cache_props1.eax & NUM_SHARE_CACHE_MASK) >> NUM_SHARE_CACHE_SHFT) + 1;
+ num_threads_sharing2 =
+ ((cache_props2.eax & NUM_SHARE_CACHE_MASK) >> NUM_SHARE_CACHE_SHFT) + 1;
+
+ new_entries = 0;
+ for (thread = 0; thread < total_num_threads; thread++) {
+
+ /* L1 data TLB */
+ if (thread % num_threads_sharing0 == 0) {
+ current = add_crat_tlb_entry(&tlb_affinity, current);
+ new_entries++;
+
+ tlb_affinity->flags |= CRAT_TLB_FLAG_DATA_TLB;
+ tlb_affinity->proc_id_low = thread;
+ sibling_mask = 1;
+ for (size_t sibling = 1; sibling < num_threads_sharing0; sibling++)
+ sibling_mask = (sibling_mask << 1) + 1;
+ tlb_affinity->sibling_map[thread / 8] = sibling_mask << (thread % 8);
+ tlb_affinity->tlb_level = CRAT_L1_CACHE;
+
+ tlb_affinity->data_tlb_2mb_assoc =
+ (l1_tlb_2M4M_ids & L1_DAT_TLB_2M4M_ASSOC_MASK)
+ >> L1_DAT_TLB_2M4M_ASSOC_SHFT;
+ tlb_affinity->data_tlb_2mb_size =
+ get_tlb_size(tlb_2m, tlb_affinity,
+ (l1_tlb_2M4M_ids & L1_DAT_TLB_2M4M_SIZE_MASK)
+ >> L1_DAT_TLB_2M4M_SIZE_SHFT);
+
+ tlb_affinity->data_tlb_4k_assoc =
+ (l1_tlb_4K_ids & L1_DAT_TLB_4K_ASSOC_MASK)
+ >> L1_DAT_TLB_4K_ASSOC_SHFT;
+ tlb_affinity->data_tlb_4k_size =
+ get_tlb_size(tlb_4k, tlb_affinity,
+ (l1_tlb_4K_ids & L1_DAT_TLB_4K_SIZE_MASK)
+ >> L1_DAT_TLB_4K_SIZE_SHFT);
+
+ tlb_affinity->data_tlb_1g_assoc =
+ (l1_tlb_1G_ids & L1_DAT_TLB_1G_ASSOC_MASK)
+ >> L1_DAT_TLB_1G_ASSOC_SHFT;
+ tlb_affinity->data_tlb_1g_size =
+ get_tlb_size(tlb_1g, tlb_affinity,
+ (l1_tlb_1G_ids & L1_DAT_TLB_1G_SIZE_MASK)
+ >> L1_DAT_TLB_1G_SIZE_SHFT);
+ }
+
+ /* L1 instruction TLB */
+ if (thread % num_threads_sharing1 == 0) {
+ current = add_crat_tlb_entry(&tlb_affinity, current);
+ new_entries++;
+
+ tlb_affinity->flags |= CRAT_TLB_FLAG_INSTR_TLB;
+ tlb_affinity->proc_id_low = thread;
+ sibling_mask = 1;
+ for (size_t sibling = 1; sibling < num_threads_sharing1; sibling++)
+ sibling_mask = (sibling_mask << 1) + 1;
+ tlb_affinity->sibling_map[thread / 8] = sibling_mask << (thread % 8);
+ tlb_affinity->tlb_level = CRAT_L1_CACHE;
+ tlb_affinity->instr_tlb_2mb_assoc =
+ (l1_tlb_2M4M_ids & L1_INST_TLB_2M4M_ASSOC_MASK)
+ >> L1_INST_TLB_2M4M_ASSOC_SHFT;
+ tlb_affinity->instr_tlb_2mb_size =
+ get_tlb_size(tlb_2m, tlb_affinity,
+ (l1_tlb_2M4M_ids & L1_INST_TLB_2M4M_SIZE_MASK)
+ >> L1_INST_TLB_2M4M_SIZE_SHFT);
+
+ tlb_affinity->instr_tlb_4k_assoc =
+ (l1_tlb_4K_ids & L1_INST_TLB_4K_ASSOC_MASK)
+ >> L1_INST_TLB_4K_ASSOC_SHFT;
+ tlb_affinity->instr_tlb_4k_size =
+ get_tlb_size(tlb_4k, tlb_affinity,
+ (l1_tlb_4K_ids & L1_INST_TLB_4K_SIZE_MASK)
+ >> L1_INST_TLB_4K_SIZE_SHFT);
+
+ tlb_affinity->instr_tlb_1g_assoc =
+ (l1_tlb_1G_ids & L1_INST_TLB_1G_ASSOC_MASK)
+ >> L1_INST_TLB_1G_ASSOC_SHFT;
+ tlb_affinity->instr_tlb_1g_size =
+ get_tlb_size(tlb_1g, tlb_affinity,
+ (l1_tlb_1G_ids & L1_INST_TLB_1G_SIZE_MASK)
+ >> L1_INST_TLB_1G_SIZE_SHFT);
+ }
+
+ /* L2 Data TLB */
+ if (thread % num_threads_sharing2 == 0) {
+ current = add_crat_tlb_entry(&tlb_affinity, current);
+ new_entries++;
+
+ tlb_affinity->flags |= CRAT_TLB_FLAG_DATA_TLB;
+ tlb_affinity->proc_id_low = thread;
+ sibling_mask = 1;
+ for (size_t sibling = 1; sibling < num_threads_sharing2; sibling++)
+ sibling_mask = (sibling_mask << 1) + 1;
+ tlb_affinity->sibling_map[thread / 8] = sibling_mask << (thread % 8);
+ tlb_affinity->tlb_level = CRAT_L2_CACHE;
+ tlb_affinity->data_tlb_2mb_assoc =
+ (l2_tlb_2M4M_ids & L2_DAT_TLB_2M4M_ASSOC_MASK)
+ >> L2_DAT_TLB_2M4M_ASSOC_SHFT;
+ tlb_affinity->data_tlb_2mb_size =
+ get_tlb_size(tlb_2m, tlb_affinity,
+ (l2_tlb_2M4M_ids & L2_DAT_TLB_2M4M_SIZE_MASK)
+ >> L2_DAT_TLB_2M4M_SIZE_SHFT);
+
+ tlb_affinity->data_tlb_4k_assoc =
+ get_associativity((l2_tlb_4K_ids & L2_DAT_TLB_2M4M_ASSOC_MASK)
+ >> L2_DAT_TLB_4K_ASSOC_SHFT);
+ tlb_affinity->data_tlb_4k_size =
+ get_tlb_size(tlb_4k, tlb_affinity,
+ (l2_tlb_2M4M_ids & L2_DAT_TLB_4K_SIZE_MASK)
+ >> L2_DAT_TLB_4K_SIZE_SHFT);
+
+ tlb_affinity->data_tlb_1g_assoc =
+ get_associativity((l2_tlb_1G_ids & L2_DAT_TLB_1G_ASSOC_MASK)
+ >> L2_DAT_TLB_1G_ASSOC_SHFT);
+ tlb_affinity->data_tlb_1g_size =
+ get_tlb_size(tlb_1g, tlb_affinity,
+ (l2_tlb_1G_ids & L2_DAT_TLB_1G_SIZE_MASK)
+ >> L2_DAT_TLB_1G_SIZE_SHFT);
+ }
+
+ /* L2 Instruction TLB */
+ if (thread % num_threads_sharing2 == 0) {
+ current = add_crat_tlb_entry(&tlb_affinity, current);
+ new_entries++;
+
+ tlb_affinity->flags |= CRAT_TLB_FLAG_INSTR_TLB;
+ tlb_affinity->proc_id_low = thread;
+ sibling_mask = 1;
+ for (size_t sibling = 1; sibling < num_threads_sharing2; sibling++)
+ sibling_mask = (sibling_mask << 1) + 1;
+ tlb_affinity->sibling_map[thread / 8] = sibling_mask << (thread % 8);
+ tlb_affinity->tlb_level = CRAT_L2_CACHE;
+ tlb_affinity->instr_tlb_2mb_assoc = get_associativity(
+ (l2_tlb_2M4M_ids & L2_INST_TLB_2M4M_ASSOC_MASK)
+ >> L2_INST_TLB_2M4M_ASSOC_SHFT);
+ tlb_affinity->instr_tlb_2mb_size =
+ get_tlb_size(tlb_2m, tlb_affinity,
+ (l2_tlb_2M4M_ids & L2_INST_TLB_2M4M_SIZE_MASK)
+ >> L2_INST_TLB_2M4M_SIZE_SHFT);
+
+ tlb_affinity->instr_tlb_4k_assoc =
+ get_associativity((l2_tlb_4K_ids & L2_INST_TLB_4K_ASSOC_MASK)
+ >> L2_INST_TLB_4K_ASSOC_SHFT);
+ tlb_affinity->instr_tlb_4k_size =
+ get_tlb_size(tlb_4k, tlb_affinity,
+ (l2_tlb_4K_ids & L2_INST_TLB_4K_SIZE_MASK)
+ >> L2_INST_TLB_4K_SIZE_SHFT);
+
+ tlb_affinity->instr_tlb_1g_assoc =
+ get_associativity((l2_tlb_1G_ids & L2_INST_TLB_1G_ASSOC_MASK)
+ >> L2_INST_TLB_1G_ASSOC_SHFT);
+ tlb_affinity->instr_tlb_1g_size =
+ get_tlb_size(tlb_1g, tlb_affinity,
+ (l2_tlb_1G_ids & L2_INST_TLB_1G_SIZE_MASK)
+ >> L2_INST_TLB_1G_SIZE_SHFT);
+ }
+ }
+
+ crat->total_entries += new_entries;
+ return current;
+}
+
+static unsigned long acpi_fill_crat(struct acpi_crat_header *crat, unsigned long current)
+{
+ current = gen_crat_hsa_entry(crat, current);
+ current = gen_crat_memory_entries(crat, current);
+ current = gen_crat_cache_entry(crat, current);
+ current = gen_crat_tlb_entry(crat, current);
+ crat->num_nodes++;
+
+ return current;
+}
+
uintptr_t agesa_write_acpi_tables(const struct device *device, uintptr_t current,
acpi_rsdp_t *rsdp)
{
acpi_ivrs_t *ivrs;
+ struct acpi_crat_header *crat;
- printk(BIOS_DEBUG, "Searching for AGESA FSP ACPI Tables\n");
+ /* CRAT */
+ current = ALIGN(current, 8);
+ crat = (struct acpi_crat_header *)current;
+ acpi_create_crat(crat, acpi_fill_crat);
+ current += crat->header.length;
+ acpi_add_table(rsdp, crat);
- current = add_agesa_acpi_table(AMD_FSP_ACPI_CRAT_HOB_GUID, "CRAT", rsdp, current);
current = add_agesa_acpi_table(AMD_FSP_ACPI_ALIB_HOB_GUID, "ALIB", rsdp, current);
/* IVRS */
diff --git a/src/soc/amd/picasso/include/soc/data_fabric.h b/src/soc/amd/picasso/include/soc/data_fabric.h
index 01fdc73388..d64c67a2bd 100644
--- a/src/soc/amd/picasso/include/soc/data_fabric.h
+++ b/src/soc/amd/picasso/include/soc/data_fabric.h
@@ -25,6 +25,38 @@
#define NB_MMIO_LIMIT(reg) ((reg) * 4 * sizeof(uint32_t) + D18F0_MMIO_LIMIT0)
#define NB_MMIO_CONTROL(reg) ((reg) * 4 * sizeof(uint32_t) + D18F0_MMIO_CTRL0)
+#define D18F0_DRAM_HOLE_CTL 0x104
+#define DRAM_HOLE_CTL_VALID BIT(0)
+#define DRAM_HOLE_CTL_BASE_SHFT 24
+#define DRAM_HOLE_CTL_BASE (0xff << DRAM_HOLE_CTL_BASE_SHFT)
+
+#define D18F0_DRAM_BASE0 0x110
+#define DRAM_BASE_REG_VALID BIT(0)
+#define DRAM_BASE_HOLE_EN BIT(1)
+#define DRAM_BASE_INTLV_CH_SHFT 4
+#define DRAM_BASE_INTLV_CH (0xf << DRAM_BASE_INTLV_CH_SHFT)
+#define DRAM_BASE_INTLV_SEL_SHFT 8
+#define DRAM_BASE_INTLV_SEL (0x7 << DRAM_BASE_INTLV_SEL_SHFT)
+#define DRAM_BASE_ADDR_SHFT 12
+#define DRAM_BASE_ADDR (0xfffff << DRAM_BASE_ADDR_SHFT)
+
+#define D18F0_DRAM_LIMIT0 0x114
+#define DRAM_LIMIT_DST_ID_SHFT 0
+#define DRAM_LIMIT_DST_ID (0xff << DRAM_LIMIT_DST_ID_SHFT)
+#define DRAM_LIMIT_INTLV_NUM_SOCK_SHFT 8
+#define DRAM_LIMIT_INTLV_NUM_SOCK (0x1 << DRAM_LIMIT_INTLV_NUM_SOCK_SHFT)
+#define DRAM_LIMIT_INTLV_NUM_DIE_SHFT 10
+#define DRAM_LIMIT_INTLV_NUM_DIE (0x3 << DRAM_LIMIT_INTLV_NUM_DIE_SHFT)
+#define DRAM_LIMIT_ADDR_SHFT 12
+#define DRAM_LIMIT_ADDR (0xfffff << DRAM_LIMIT_ADDR_SHFT)
+
+#define PICASSO_NUM_DRAM_REG 2
+
+#define DF_DRAM_BASE(dram_map_pair) ((dram_map_pair) * 2 * sizeof(uint32_t) \
+ + D18F0_DRAM_BASE0)
+#define DF_DRAM_LIMIT(dram_map_pair) ((dram_map_pair) * 2 * sizeof(uint32_t) \
+ + D18F0_DRAM_LIMIT0)
+
#define DF_FICAA_BIOS 0x5C
#define DF_FICAD_LO 0x98
#define DF_FICAD_HI 0x9C