summaryrefslogtreecommitdiff
path: root/src/cpu
diff options
context:
space:
mode:
authorMartin Roth <martin.roth@se-eng.com>2013-07-08 16:23:54 -0600
committerStefan Reinauer <stefan.reinauer@coreboot.org>2013-07-11 22:36:59 +0200
commit4c3ab7376ebb2e3e18919f1ab663d317dfec9b9c (patch)
tree6bd8440a05f6ea1184c0a5500d43cc92ab683f01 /src/cpu
parent0cb07e3476d9408d0935253f9f26c0a8ddc28401 (diff)
cpu: Fix spelling
Change-Id: I69c46648de0689e9bed84c7726906024ad65e769 Signed-off-by: Martin Roth <martin.roth@se-eng.com> Reviewed-on: http://review.coreboot.org/3729 Tested-by: build bot (Jenkins) Reviewed-by: Stefan Reinauer <stefan.reinauer@coreboot.org>
Diffstat (limited to 'src/cpu')
-rw-r--r--src/cpu/amd/car/post_cache_as_ram.c4
-rw-r--r--src/cpu/amd/dualcore/amd_sibling.c4
-rw-r--r--src/cpu/amd/geode_gx2/cpubug.c6
-rw-r--r--src/cpu/amd/geode_gx2/cpureginit.c4
-rw-r--r--src/cpu/amd/geode_lx/cpubug.c2
-rw-r--r--src/cpu/amd/model_10xxx/defaults.h2
-rw-r--r--src/cpu/amd/model_10xxx/fidvid.c6
-rw-r--r--src/cpu/amd/model_10xxx/init_cpus.c2
-rw-r--r--src/cpu/amd/model_fxx/fidvid.c4
-rw-r--r--src/cpu/amd/model_fxx/model_fxx_init.c2
-rw-r--r--src/cpu/amd/model_fxx/powernow_acpi.c6
-rw-r--r--src/cpu/amd/mtrr/amd_mtrr.c4
-rw-r--r--src/cpu/amd/quadcore/amd_sibling.c4
-rw-r--r--src/cpu/intel/haswell/haswell.h2
-rw-r--r--src/cpu/intel/haswell/haswell_init.c10
-rw-r--r--src/cpu/intel/haswell/mp_init.c8
-rw-r--r--src/cpu/intel/haswell/smmrelocate.c2
-rw-r--r--src/cpu/intel/model_2065x/model_2065x_init.c4
-rw-r--r--src/cpu/intel/model_206ax/model_206ax_init.c4
-rw-r--r--src/cpu/intel/slot_1/l2_cache.c6
-rw-r--r--src/cpu/samsung/exynos5250/clk.h4
-rw-r--r--src/cpu/samsung/exynos5250/clock.c8
-rw-r--r--src/cpu/samsung/exynos5250/cpu.c2
-rw-r--r--src/cpu/samsung/exynos5250/dmc.h2
-rw-r--r--src/cpu/samsung/exynos5250/dmc_common.c6
-rw-r--r--src/cpu/samsung/exynos5250/dmc_init_ddr3.c6
-rw-r--r--src/cpu/samsung/exynos5250/dp-core.h8
-rw-r--r--src/cpu/samsung/exynos5250/dp-reg.c4
-rw-r--r--src/cpu/samsung/exynos5250/fimd.h4
-rw-r--r--src/cpu/samsung/exynos5250/pinmux.c2
-rw-r--r--src/cpu/samsung/exynos5250/setup.h4
-rw-r--r--src/cpu/samsung/exynos5250/spi.c4
-rw-r--r--src/cpu/samsung/exynos5250/tmu.c4
-rw-r--r--src/cpu/samsung/exynos5250/tmu.h2
-rw-r--r--src/cpu/samsung/exynos5250/uart.c4
-rw-r--r--src/cpu/ti/am335x/uart.c2
-rw-r--r--src/cpu/ti/am335x/uart.h4
-rw-r--r--src/cpu/via/nano/update_ucode.c4
-rw-r--r--src/cpu/via/nano/update_ucode.h4
-rw-r--r--src/cpu/x86/car.c2
-rw-r--r--src/cpu/x86/lapic/lapic_cpu_init.c2
-rw-r--r--src/cpu/x86/mtrr/earlymtrr.c4
-rw-r--r--src/cpu/x86/mtrr/mtrr.c18
-rw-r--r--src/cpu/x86/smm/smm_module_loader.c14
44 files changed, 102 insertions, 102 deletions
diff --git a/src/cpu/amd/car/post_cache_as_ram.c b/src/cpu/amd/car/post_cache_as_ram.c
index 68e7c09ad3..18c278e41a 100644
--- a/src/cpu/amd/car/post_cache_as_ram.c
+++ b/src/cpu/amd/car/post_cache_as_ram.c
@@ -88,7 +88,7 @@ static void post_cache_as_ram(void)
#endif
#if 1
{
- /* Check value of esp to verify if we have enough rom for stack in Cache as RAM */
+ /* Check value of esp to verify if we have enough room for stack in Cache as RAM */
unsigned v_esp;
__asm__ volatile (
"movl %%esp, %0\n\t"
@@ -123,7 +123,7 @@ static void post_cache_as_ram(void)
/* set new esp */ /* before CONFIG_RAMBASE */
"subl %0, %%esp\n\t"
::"a"( (CONFIG_DCACHE_RAM_BASE + CONFIG_DCACHE_RAM_SIZE)- (CONFIG_RAMTOP) )
- /* discard all registers (eax is used for %0), so gcc redo everything
+ /* discard all registers (eax is used for %0), so gcc redoes everything
after the stack is moved */
: "cc", "memory", "%ebx", "%ecx", "%edx", "%esi", "%edi", "%ebp"
);
diff --git a/src/cpu/amd/dualcore/amd_sibling.c b/src/cpu/amd/dualcore/amd_sibling.c
index 13b9e378f3..b3df0a36e7 100644
--- a/src/cpu/amd/dualcore/amd_sibling.c
+++ b/src/cpu/amd/dualcore/amd_sibling.c
@@ -91,10 +91,10 @@ unsigned get_apicid_base(unsigned ioapic_num)
}
#endif
- //contruct apicid_base
+ //Construct apicid_base
if((!disable_siblings) && (siblings>0) ) {
- /* for 8 way dual core, we will used up apicid 16:16, actualy 16 is not allowed by current kernel
+ /* for 8 way dual core, we will used up apicid 16:16, actually 16 is not allowed by current kernel
and the kernel will try to get one that is small than 16 to make io apic work.
I don't know when the kernel can support 256 apic id. (APIC_EXT_ID is enabled) */
diff --git a/src/cpu/amd/geode_gx2/cpubug.c b/src/cpu/amd/geode_gx2/cpubug.c
index 54d7e0e8d2..eddc92408e 100644
--- a/src/cpu/amd/geode_gx2/cpubug.c
+++ b/src/cpu/amd/geode_gx2/cpubug.c
@@ -163,7 +163,7 @@ static void eng2900(void)
* clocks when CPU is snooped. Because setting XSTATE to 0
* overrides any other XSTATE action, the code will always
* stall for 4 GeodeLink clocks after a snoop request goes
- * away even if it occured a clock or two later than a
+ * away even if it occurred a clock or two later than a
* different snoop; the stall signal will never 'glitch high'
* for only one or two CPU clocks with this code.
*/
@@ -201,7 +201,7 @@ static void eng2900(void)
msr.lo = 0x30000;
wrmsr(MSR_GLCP + 0x0073, msr);
- /* Writing action number 5: STALL_CPU_PIPE when exitting idle
+ /* Writing action number 5: STALL_CPU_PIPE when exiting idle
state or not in idle state */
msr.hi = 0;
msr.lo = 0x00430000;
@@ -293,7 +293,7 @@ static void bug118339(void)
*
* PBZ 3659:
* The MC reordered transactions incorrectly and breaks coherency.
- * Disable reording and take a potential performance hit.
+ * Disable reordering and take a potential performance hit.
* This is safe to do here and not in MC init since there is nothing
* to maintain coherency with and the cache is not enabled yet.
*/
diff --git a/src/cpu/amd/geode_gx2/cpureginit.c b/src/cpu/amd/geode_gx2/cpureginit.c
index 0fc852d531..e0ecd62722 100644
--- a/src/cpu/amd/geode_gx2/cpureginit.c
+++ b/src/cpu/amd/geode_gx2/cpureginit.c
@@ -15,7 +15,7 @@ void cpuRegInit (void)
/* Set up GLCP to grab BTM data. */
msrnum = GLCP_DBGOUT; /* GLCP_DBGOUT MSR */
msr.hi = 0x0;
- msr.lo = 0x08; /* reset value (SCOPE_SEL = 0) causes FIFO toshift out, */
+ msr.lo = 0x08; /* reset value (SCOPE_SEL = 0) causes FIFO to shift out, */
wrmsr(msrnum, msr); /* exchange it to anything else to prevent this */
/* Turn off debug clock */
@@ -119,7 +119,7 @@ void cpuRegInit (void)
wrmsr(msrnum, msr);
}
-/* FPU impercise exceptions bit */
+/* FPU imprecise exceptions bit */
{
msrnum = CPU_FPU_MSR_MODE;
msr = rdmsr(msrnum);
diff --git a/src/cpu/amd/geode_lx/cpubug.c b/src/cpu/amd/geode_lx/cpubug.c
index 4110454566..ebadec7f7a 100644
--- a/src/cpu/amd/geode_lx/cpubug.c
+++ b/src/cpu/amd/geode_lx/cpubug.c
@@ -67,7 +67,7 @@ static void pcideadlock(void)
/***/
/** PBZ 3659:*/
/** The MC reordered transactions incorrectly and breaks coherency.*/
-/** Disable reording and take a potential performance hit.*/
+/** Disable reordering and take a potential performance hit.*/
/** This is safe to do here and not in MC init since there is nothing*/
/** to maintain coherency with and the cache is not enabled yet.*/
/***/
diff --git a/src/cpu/amd/model_10xxx/defaults.h b/src/cpu/amd/model_10xxx/defaults.h
index f7b8b2f9e1..3d33dda029 100644
--- a/src/cpu/amd/model_10xxx/defaults.h
+++ b/src/cpu/amd/model_10xxx/defaults.h
@@ -54,7 +54,7 @@ static const struct {
{ DC_CFG, AMD_DR_Bx, AMD_PTYPE_SVR,
0x00000000, 0x00000000,
- 0x00000000, 0x00000C00 }, /* Errata 326 */
+ 0x00000000, 0x00000C00 }, /* Erratum 326 */
{ NB_CFG, AMD_FAM10_ALL, AMD_PTYPE_DC | AMD_PTYPE_MC,
0x00000000, 1 << 22,
diff --git a/src/cpu/amd/model_10xxx/fidvid.c b/src/cpu/amd/model_10xxx/fidvid.c
index ebe44dd76a..4297c90460 100644
--- a/src/cpu/amd/model_10xxx/fidvid.c
+++ b/src/cpu/amd/model_10xxx/fidvid.c
@@ -68,7 +68,7 @@ Fam10 Bios and Kernel Development Guide #31116, rev 3.48, April 22, 2010
9.- TODO Requires information on current delivery capability
(depends on mainboard and maybe power supply ?). One might use a config
- option with the maximum number of Ampers that the board can deliver to CPU.
+ option with the maximum number of Amperes that the board can deliver to CPU.
10.- [Multiprocessor] TODO 2.4.2.12
[Uniprocessor] FIXME ? We call setPStateMaxVal() in init_fidvid_stage2,
@@ -79,7 +79,7 @@ Fam10 Bios and Kernel Development Guide #31116, rev 3.48, April 22, 2010
11.- finalPstateChange() from init_fidvid_Stage2 (BKDG says just "may", anyway)
12.- generate ACPI for p-states. FIXME
- Needs more assesment. There's some kind of fixed support that
+ Needs more assessment. There's some kind of fixed support that
does not seem to depend on CPU revision or actual MSRC001_00[68:64]
as BKDG apparently requires.
http://www.coreboot.org/ACPI#CPU_Power_Management
@@ -935,7 +935,7 @@ static void fixPsNbVidAfterWR(u32 newNbVid, u8 NbVidUpdatedAll,u8 pviMode)
static void finalPstateChange(void)
{
- /* Enble P0 on all cores for best performance.
+ /* Enable P0 on all cores for best performance.
* Linux can slow them down later if need be.
* It is safe since they will be in C1 halt
* most of the time anyway.
diff --git a/src/cpu/amd/model_10xxx/init_cpus.c b/src/cpu/amd/model_10xxx/init_cpus.c
index 3d5ddcfb35..eb047b8db2 100644
--- a/src/cpu/amd/model_10xxx/init_cpus.c
+++ b/src/cpu/amd/model_10xxx/init_cpus.c
@@ -424,7 +424,7 @@ static void start_node(u8 node)
/**
* static void setup_remote_node(u32 node)
*
- * Copy the BSP Adress Map to each AP.
+ * Copy the BSP Address Map to each AP.
*/
static void setup_remote_node(u8 node)
{
diff --git a/src/cpu/amd/model_fxx/fidvid.c b/src/cpu/amd/model_fxx/fidvid.c
index 06ff636659..e68611bca8 100644
--- a/src/cpu/amd/model_fxx/fidvid.c
+++ b/src/cpu/amd/model_fxx/fidvid.c
@@ -496,7 +496,7 @@ static void init_fidvid_bsp_stage2(unsigned ap_apicid, void *gp)
continue;
if ((readback & 0xff) == 2) {
timeout = 0;
- break; /* target ap is stage 2, it's FID has beed set */
+ break; /* target ap is stage 2, its FID has been set */
}
}
@@ -603,7 +603,7 @@ static void init_fidvid_bsp(unsigned bsp_apicid)
/* For all APs ( We know the APIC ID of all AP even the APIC ID is lifted)
* send signal to the AP it could change it's fid/vid */
- /* remote read singnal from AP that AP is done */
+ /* remote read signal from AP that AP is done */
fv.common_fidvid &= 0xffff00;
diff --git a/src/cpu/amd/model_fxx/model_fxx_init.c b/src/cpu/amd/model_fxx/model_fxx_init.c
index a18b982bdb..42c6f95119 100644
--- a/src/cpu/amd/model_fxx/model_fxx_init.c
+++ b/src/cpu/amd/model_fxx/model_fxx_init.c
@@ -509,7 +509,7 @@ static void model_fxx_init(device_t dev)
id = get_node_core_id(read_nb_cfg_54()); // pre e0 nb_cfg_54 can not be set
- /* Is this a bad location? In particular can another node prefecth
+ /* Is this a bad location? In particular can another node prefetch
* data from this node before we have initialized it?
*/
if (id.coreid == 0)
diff --git a/src/cpu/amd/model_fxx/powernow_acpi.c b/src/cpu/amd/model_fxx/powernow_acpi.c
index 820d4262e6..af1e24b094 100644
--- a/src/cpu/amd/model_fxx/powernow_acpi.c
+++ b/src/cpu/amd/model_fxx/powernow_acpi.c
@@ -75,7 +75,7 @@ static int write_pstates_for_core(u8 pstate_num, u16 *pstate_feq, u8 *pstate_vid
#if CONFIG_K8_REV_F_SUPPORT
/*
-* Details about this algorithm , refert to BDKG 10.5.1
+* Details about this algorithm , refer to BDKG 10.5.1
* Two parts are included, the another is the DSDT reconstruction process
*/
@@ -202,7 +202,7 @@ static int pstates_algorithm(u32 pcontrol_blk, u8 plen, u8 onlyBSP)
goto write_pstates;
}
- /* Get the multipier of the fid frequency */
+ /* Get the multiplier of the fid frequency */
/*
* Fid multiplier is always 100 revF and revG.
*/
@@ -316,7 +316,7 @@ static int pstates_algorithm(u32 pcontrol_blk, u8 plen, u8 onlyBSP)
Pstate_num++;
}
- /* Constuct P[Min] State */
+ /* Construct P[Min] State */
if (Max_fid == 0x2A && Max_vid != 0x0) {
Pstate_fid[Pstate_num] = 0x2;
Pstate_feq[Pstate_num] =
diff --git a/src/cpu/amd/mtrr/amd_mtrr.c b/src/cpu/amd/mtrr/amd_mtrr.c
index 2f3e6e340f..033ec53d3a 100644
--- a/src/cpu/amd/mtrr/amd_mtrr.c
+++ b/src/cpu/amd/mtrr/amd_mtrr.c
@@ -128,10 +128,10 @@ void amd_setup_mtrrs(void)
}
/* Now that I have mapped what is memory and what is not
- * Setup the mtrrs so we can cache the memory.
+ * Set up the mtrrs so we can cache the memory.
*/
- // Rev. F K8 supports has SYSCFG_MSR_TOM2WB and dont need
+ // Rev. F K8 supports has SYSCFG_MSR_TOM2WB and doesn't need
// variable MTRR to span memory above 4GB
// Lower revisions K8 need variable MTRR over 4GB
x86_setup_var_mtrrs(address_bits, has_tom2wb ? 0 : 1);
diff --git a/src/cpu/amd/quadcore/amd_sibling.c b/src/cpu/amd/quadcore/amd_sibling.c
index e0d173dd79..d653a85d5a 100644
--- a/src/cpu/amd/quadcore/amd_sibling.c
+++ b/src/cpu/amd/quadcore/amd_sibling.c
@@ -95,10 +95,10 @@ u32 get_apicid_base(u32 ioapic_num)
nb_cfg_54 = read_nb_cfg_54();
- //contruct apicid_base
+ //Construct apicid_base
if((!disable_siblings) && (siblings>0) ) {
- /* for 8 way dual core, we will used up apicid 16:16, actualy
+ /* for 8 way dual core, we will used up apicid 16:16, actually
16 is not allowed by current kernel and the kernel will try
to get one that is small than 16 to make io apic work. I don't
know when the kernel can support 256 apic id.
diff --git a/src/cpu/intel/haswell/haswell.h b/src/cpu/intel/haswell/haswell.h
index 4a739a9e30..ac082b86eb 100644
--- a/src/cpu/intel/haswell/haswell.h
+++ b/src/cpu/intel/haswell/haswell.h
@@ -169,7 +169,7 @@ void smm_initiate_relocation(void);
void smm_initiate_relocation_parallel(void);
struct bus;
void bsp_init_and_start_aps(struct bus *cpu_bus);
-/* Returns 0 on succes. < 0 on failure. */
+/* Returns 0 on success. < 0 on failure. */
int setup_ap_init(struct bus *cpu_bus, int *max_cpus,
const void *microcode_patch);
/* Returns 0 on success, < 0 on failure. */
diff --git a/src/cpu/intel/haswell/haswell_init.c b/src/cpu/intel/haswell/haswell_init.c
index 056b289bb2..85f6afd031 100644
--- a/src/cpu/intel/haswell/haswell_init.c
+++ b/src/cpu/intel/haswell/haswell_init.c
@@ -39,7 +39,7 @@
#include "chip.h"
/*
- * List of suported C-states in this processor
+ * List of supported C-states in this processor
*
* Latencies are typical worst-case package exit time in uS
* taken from the SandyBridge BIOS specification.
@@ -324,7 +324,7 @@ static void configure_thermal_target(void)
return;
conf = lapic->chip_info;
- /* Set TCC activaiton offset if supported */
+ /* Set TCC activation offset if supported */
msr = rdmsr(MSR_PLATFORM_INFO);
if ((msr.lo & (1 << 30)) && conf->tcc_offset) {
msr = rdmsr(MSR_TEMPERATURE_TARGET);
@@ -508,8 +508,8 @@ void bsp_init_and_start_aps(struct bus *cpu_bus)
int num_aps;
const void *microcode_patch;
- /* Perform any necesarry BSP initialization before APs are brought up.
- * This call alos allows the BSP to prepare for any secondary effects
+ /* Perform any necessary BSP initialization before APs are brought up.
+ * This call also allows the BSP to prepare for any secondary effects
* from calling cpu_initialize() such as smm_init(). */
bsp_init_before_ap_bringup(cpu_bus);
@@ -529,7 +529,7 @@ void bsp_init_and_start_aps(struct bus *cpu_bus)
}
if (smm_initialize()) {
- printk(BIOS_CRIT, "SMM Initialiazation failed...\n");
+ printk(BIOS_CRIT, "SMM Initialization failed...\n");
return;
}
diff --git a/src/cpu/intel/haswell/mp_init.c b/src/cpu/intel/haswell/mp_init.c
index 1358418557..51130a56a3 100644
--- a/src/cpu/intel/haswell/mp_init.c
+++ b/src/cpu/intel/haswell/mp_init.c
@@ -78,7 +78,7 @@ static device_t cpu_devs[CONFIG_MAX_CPUS];
static atomic_t num_aps;
/* Number of APs that have relocated their SMM handler. */
static atomic_t num_aps_relocated_smm;
-/* Barrier to stop APs from performing SMM relcoation. */
+/* Barrier to stop APs from performing SMM relocation. */
static int smm_relocation_barrier_begin __attribute__ ((aligned (64)));
/* Determine if hyperthreading is disabled. */
int ht_disabled;
@@ -145,7 +145,7 @@ void release_aps_for_smm_relocation(int do_parallel)
/* The mtrr code sets up ROM caching on the BSP, but not the others. However,
* the boot loader payload disables this. In order for Linux not to complain
- * ensure the caching is disabled for tha APs before going to sleep. */
+ * ensure the caching is disabled for the APs before going to sleep. */
static void cleanup_rom_caching(void)
{
x86_mtrr_disable_rom_caching();
@@ -178,7 +178,7 @@ static void asmlinkage ap_init(unsigned int cpu, void *microcode_ptr)
ap_initiate_smm_relocation();
- /* Indicate that SMM relocation has occured on this thread. */
+ /* Indicate that SMM relocation has occurred on this thread. */
atomic_inc(&num_aps_relocated_smm);
/* After SMM relocation a 2nd microcode load is required. */
@@ -401,7 +401,7 @@ static int allocate_cpu_devices(struct bus *cpu_bus, int *total_hw_threads)
/* Allocate the new cpu device structure */
new = alloc_find_dev(cpu_bus, &cpu_path);
if (new == NULL) {
- printk(BIOS_CRIT, "Could not allocte cpu device\n");
+ printk(BIOS_CRIT, "Could not allocate cpu device\n");
max_cpus--;
}
cpu_devs[i] = new;
diff --git a/src/cpu/intel/haswell/smmrelocate.c b/src/cpu/intel/haswell/smmrelocate.c
index 4fe6489be2..3f4f45ab4f 100644
--- a/src/cpu/intel/haswell/smmrelocate.c
+++ b/src/cpu/intel/haswell/smmrelocate.c
@@ -348,7 +348,7 @@ static void setup_ied_area(struct smm_relocation_params *params)
memset(ied_base + (1 << 20), 0, (32 << 10));
/* According to the BWG MP init section 2MiB of memory at IEDBASE +
- * 2MiB should be zeroed as well. However, I suspect what is inteneded
+ * 2MiB should be zeroed as well. However, I suspect what is intended
* is to clear the memory covered by EMRR. TODO(adurbin): figure out if * this is really required. */
//memset(ied_base + (2 << 20), 0, (2 << 20));
}
diff --git a/src/cpu/intel/model_2065x/model_2065x_init.c b/src/cpu/intel/model_2065x/model_2065x_init.c
index ab4db59e6d..4c88e4412f 100644
--- a/src/cpu/intel/model_2065x/model_2065x_init.c
+++ b/src/cpu/intel/model_2065x/model_2065x_init.c
@@ -39,7 +39,7 @@
#include "chip.h"
/*
- * List of suported C-states in this processor
+ * List of supported C-states in this processor
*
* Latencies are typical worst-case package exit time in uS
* taken from the SandyBridge BIOS specification.
@@ -249,7 +249,7 @@ static void configure_thermal_target(void)
return;
conf = lapic->chip_info;
- /* Set TCC activaiton offset if supported */
+ /* Set TCC activation offset if supported */
msr = rdmsr(MSR_PLATFORM_INFO);
if ((msr.lo & (1 << 30)) && conf->tcc_offset) {
msr = rdmsr(MSR_TEMPERATURE_TARGET);
diff --git a/src/cpu/intel/model_206ax/model_206ax_init.c b/src/cpu/intel/model_206ax/model_206ax_init.c
index fa974e53cd..60288010da 100644
--- a/src/cpu/intel/model_206ax/model_206ax_init.c
+++ b/src/cpu/intel/model_206ax/model_206ax_init.c
@@ -39,7 +39,7 @@
#include "chip.h"
/*
- * List of suported C-states in this processor
+ * List of supported C-states in this processor
*
* Latencies are typical worst-case package exit time in uS
* taken from the SandyBridge BIOS specification.
@@ -374,7 +374,7 @@ static void configure_thermal_target(void)
return;
conf = lapic->chip_info;
- /* Set TCC activaiton offset if supported */
+ /* Set TCC activation offset if supported */
msr = rdmsr(MSR_PLATFORM_INFO);
if ((msr.lo & (1 << 30)) && conf->tcc_offset) {
msr = rdmsr(MSR_TEMPERATURE_TARGET);
diff --git a/src/cpu/intel/slot_1/l2_cache.c b/src/cpu/intel/slot_1/l2_cache.c
index eca0c89e3b..a974d24c3d 100644
--- a/src/cpu/intel/slot_1/l2_cache.c
+++ b/src/cpu/intel/slot_1/l2_cache.c
@@ -34,7 +34,7 @@
*/
/* This code is ported from coreboot v1.
- * The L2 cache initalization sequence here only apply to SECC/SECC2 P6 family
+ * The L2 cache initialization sequence here only apply to SECC/SECC2 P6 family
* CPUs with Klamath (63x), Deschutes (65x) and Katmai (67x) cores.
* It is not required for Coppermine (68x) and Tualatin (6bx) cores.
* It is currently not known if Celerons with Mendocino (66x) core require the
@@ -295,7 +295,7 @@ int write_l2(u32 address, u32 data)
// data1 = ffffffff
// data2 = 000000dc
// address = 00aaaaaa
- // Final address signalled:
+ // Final address signaled:
// 000fffff fff000c0 000dcaaa aaa00000
data1 = data & 0xff;
data1 = data1 << 21;
@@ -343,7 +343,7 @@ int test_l2_address_alias(u32 address1, u32 address2,
/* Calculates the L2 cache size.
*
- * Reference: Intel(R) 64 and IA-32 Architectures Software Developer’s Manual
+ * Reference: Intel(R) 64 and IA-32 Architectures Software Developer�s Manual
* Volume 3B: System Programming Guide, Part 2, Intel pub. 253669, pg. B-172.
*
*/
diff --git a/src/cpu/samsung/exynos5250/clk.h b/src/cpu/samsung/exynos5250/clk.h
index ba8d960ca4..565cf2b255 100644
--- a/src/cpu/samsung/exynos5250/clk.h
+++ b/src/cpu/samsung/exynos5250/clk.h
@@ -72,7 +72,7 @@ unsigned long clock_get_periph_rate(enum periph_id peripheral);
#define MCT_HZ 24000000
/*
- * Set mshci controller instances clock drivder
+ * Set mshci controller instances clock divider
*
* @param enum periph_id instance of the mshci controller
*
@@ -521,7 +521,7 @@ struct exynos5_mct_regs {
};
#define EXYNOS5_EPLLCON0_LOCKED_SHIFT 29 /* EPLL Locked bit position*/
-#define EPLL_SRC_CLOCK 24000000 /*24 MHz Cristal Input */
+#define EPLL_SRC_CLOCK 24000000 /*24 MHz Crystal Input */
#define TIMEOUT_EPLL_LOCK 1000
#define AUDIO_0_RATIO_MASK 0x0f
diff --git a/src/cpu/samsung/exynos5250/clock.c b/src/cpu/samsung/exynos5250/clock.c
index 78f0bd434d..390fae30f6 100644
--- a/src/cpu/samsung/exynos5250/clock.c
+++ b/src/cpu/samsung/exynos5250/clock.c
@@ -160,7 +160,7 @@ static struct clk_bit_info clk_bit_info[PERIPH_ID_COUNT] = {
{24, 1, 20, -1}, /* PERIPH_ID_SATA */
};
-/* Epll Clock division values to achive different frequency output */
+/* Epll Clock division values to achieve different frequency output */
static struct st_epll_con_val epll_div[] = {
{ 192000000, 0, 48, 3, 1, 0 },
{ 180000000, 0, 45, 3, 1, 0 },
@@ -405,7 +405,7 @@ void clock_ll_set_pre_ratio(enum periph_id periph_id, unsigned divisor)
u32 *reg;
/*
- * For now we only handle a very small subset of peipherals here.
+ * For now we only handle a very small subset of peripherals here.
* Others will need to (and do) mangle the clock registers
* themselves, At some point it is hoped that this function can work
* from a table or calculated register offset / mask. For now this
@@ -636,7 +636,7 @@ int clock_epll_set_rate(unsigned long rate)
epll_con |= epll_div[i].s_div << EPLL_CON0_SDIV_SHIFT;
/*
- * Required period ( in cycles) to genarate a stable clock output.
+ * Required period ( in cycles) to generate a stable clock output.
* The maximum clock time can be up to 3000 * PDIV cycles of PLLs
* frequency input (as per spec)
*/
@@ -675,7 +675,7 @@ int clock_set_i2s_clk_prescaler(unsigned int src_frq, unsigned int dst_frq)
unsigned int div ;
if ((dst_frq == 0) || (src_frq == 0)) {
- printk(BIOS_DEBUG, "%s: Invalid requency input for prescaler\n", __func__);
+ printk(BIOS_DEBUG, "%s: Invalid frequency input for prescaler\n", __func__);
printk(BIOS_DEBUG, "src frq = %d des frq = %d ", src_frq, dst_frq);
return -1;
}
diff --git a/src/cpu/samsung/exynos5250/cpu.c b/src/cpu/samsung/exynos5250/cpu.c
index 96e7dc361a..61f937b378 100644
--- a/src/cpu/samsung/exynos5250/cpu.c
+++ b/src/cpu/samsung/exynos5250/cpu.c
@@ -107,7 +107,7 @@ static void exynos_displayport_init(device_t dev)
*
* Note: We may want to do something clever to ensure the framebuffer
* region is aligned such that we don't change dcache policy for other
- * stuff inadvertantly.
+ * stuff inadvertently.
*/
uint32_t lower = ALIGN_DOWN(lcdbase, MiB);
uint32_t upper = ALIGN_UP(lcdbase + fb_size, MiB);
diff --git a/src/cpu/samsung/exynos5250/dmc.h b/src/cpu/samsung/exynos5250/dmc.h
index d676aa7463..acd0abbcf1 100644
--- a/src/cpu/samsung/exynos5250/dmc.h
+++ b/src/cpu/samsung/exynos5250/dmc.h
@@ -328,7 +328,7 @@ struct mem_timings {
uint8_t chips_per_channel; /* number of chips per channel */
uint8_t chips_to_configure; /* number of chips to configure */
uint8_t send_zq_init; /* 1 to send this command */
- unsigned int impedance; /* drive strength impedeance */
+ unsigned int impedance; /* drive strength impedance */
uint8_t gate_leveling_enable; /* check gate leveling is enabled */
};
diff --git a/src/cpu/samsung/exynos5250/dmc_common.c b/src/cpu/samsung/exynos5250/dmc_common.c
index 55388ac162..b506853df8 100644
--- a/src/cpu/samsung/exynos5250/dmc_common.c
+++ b/src/cpu/samsung/exynos5250/dmc_common.c
@@ -64,7 +64,7 @@ int dmc_config_zq(struct mem_timings *mem,
val &= ~ZQ_MANUAL_STR;
/*
- * Since we are manaully calibrating the ZQ values,
+ * Since we are manually calibrating the ZQ values,
* we are looping for the ZQ_init to complete.
*/
i = ZQ_INIT_TIMEOUT;
@@ -97,12 +97,12 @@ void update_reset_dll(struct exynos5_dmc *dmc, enum ddr_mode mode)
writel(val, &dmc->phycontrol0);
}
- /* Update DLL Information: Force DLL Resyncronization */
+ /* Update DLL Information: Force DLL Resynchronization */
val = readl(&dmc->phycontrol0);
val |= FP_RSYNC;
writel(val, &dmc->phycontrol0);
- /* Reset Force DLL Resyncronization */
+ /* Reset Force DLL Resynchronization */
val = readl(&dmc->phycontrol0);
val &= ~FP_RSYNC;
writel(val, &dmc->phycontrol0);
diff --git a/src/cpu/samsung/exynos5250/dmc_init_ddr3.c b/src/cpu/samsung/exynos5250/dmc_init_ddr3.c
index cb5c61389f..554f4c2dc4 100644
--- a/src/cpu/samsung/exynos5250/dmc_init_ddr3.c
+++ b/src/cpu/samsung/exynos5250/dmc_init_ddr3.c
@@ -139,9 +139,9 @@ int ddr3_mem_ctrl_init(struct mem_timings *mem, unsigned long mem_iv_size,
writel(mem->concontrol | (mem->rd_fetch << CONCONTROL_RD_FETCH_SHIFT),
&dmc->concontrol);
- /* Memory Channel Inteleaving Size */
+ /* Memory Channel Interleaving Size */
printk(BIOS_SPEW, "ddr3_mem_ctrl_init: "
- "Memory Channel Inteleaving Size\n");
+ "Memory Channel Interleaving Size\n");
writel(mem->iv_size, &dmc->ivcontrol);
/* Set DMC MEMCONTROL register */
@@ -161,7 +161,7 @@ int ddr3_mem_ctrl_init(struct mem_timings *mem, unsigned long mem_iv_size,
/* Power Down mode Configuration */
printk(BIOS_SPEW, "ddr3_mem_ctrl_init: "
- "Power Down mode Configuraation\n");
+ "Power Down mode Configuration\n");
writel(mem->dpwrdn_cyc << PWRDNCONFIG_DPWRDN_CYC_SHIFT |
mem->dsref_cyc << PWRDNCONFIG_DSREF_CYC_SHIFT,
&dmc->pwrdnconfig);
diff --git a/src/cpu/samsung/exynos5250/dp-core.h b/src/cpu/samsung/exynos5250/dp-core.h
index 73f4638599..ab7e7e4b32 100644
--- a/src/cpu/samsung/exynos5250/dp-core.h
+++ b/src/cpu/samsung/exynos5250/dp-core.h
@@ -29,7 +29,7 @@
#define MAX_CR_LOOP 5
#define MAX_EQ_LOOP 4
-/* Link tare type */
+/* Link rate type */
enum link_rate {
LINK_RATE_1_62GBPS = 0x06,
LINK_RATE_2_70GBPS = 0x0a
@@ -126,7 +126,7 @@ struct s5p_dp_device {
*/
void s5p_dp_reset(struct s5p_dp_device *dp);
/*
- * Initialize DP to recieve video stream
+ * Initialize DP to receive video stream
*
* param dp pointer to main s5p-dp structure
*/
@@ -214,8 +214,8 @@ int s5p_dp_is_slave_video_stream_clock_on(struct s5p_dp_device *dp);
*
* param dp pointer to main s5p-dp structure
* param type clock_recovery_m_value_type
- * param m_value to caluculate m_vid value
- * param n_value to caluculate n_vid value
+ * param m_value to calculate m_vid value
+ * param n_value to calculate n_vid value
*/
void s5p_dp_set_video_cr_mn(struct s5p_dp_device *dp,
enum clock_recovery_m_value_type type,
diff --git a/src/cpu/samsung/exynos5250/dp-reg.c b/src/cpu/samsung/exynos5250/dp-reg.c
index eb53356449..ac4208d159 100644
--- a/src/cpu/samsung/exynos5250/dp-reg.c
+++ b/src/cpu/samsung/exynos5250/dp-reg.c
@@ -91,7 +91,7 @@ void s5p_dp_reset(struct s5p_dp_device *dp)
/* Set interrupt pin assertion polarity as high */
writel(INT_POL0 | INT_POL1, &base->int_ctl);
- /* Clear pending regisers */
+ /* Clear pending registers */
writel(0xff, &base->common_int_sta_1);
writel(0x4f, &base->common_int_sta_2);
writel(0xe0, &base->common_int_sta_3);
@@ -156,7 +156,7 @@ void s5p_dp_init_aux(struct s5p_dp_device *dp)
u32 reg;
struct exynos5_dp *base = dp->base;
- /* Clear inerrupts related to AUX channel */
+ /* Clear interrupts related to AUX channel */
reg = RPLY_RECEIV | AUX_ERR;
writel(reg, &base->dp_int_sta);
diff --git a/src/cpu/samsung/exynos5250/fimd.h b/src/cpu/samsung/exynos5250/fimd.h
index 71d1785568..f3dea37a1b 100644
--- a/src/cpu/samsung/exynos5250/fimd.h
+++ b/src/cpu/samsung/exynos5250/fimd.h
@@ -86,11 +86,11 @@ struct exynos5_fimd_panel {
unsigned int upper_margin; /* Vertical Backporch */
unsigned int lower_margin; /* Vertical frontporch */
unsigned int vsync; /* Vertical Sync Pulse Width */
- unsigned int left_margin; /* Horizantal Backporch */
+ unsigned int left_margin; /* Horizontal Backporch */
unsigned int right_margin; /* Horizontal Frontporch */
unsigned int hsync; /* Horizontal Sync Pulse Width */
unsigned int xres; /* X Resolution */
- unsigned int yres; /* Y Resopultion */
+ unsigned int yres; /* Y Resolution */
};
/* LCDIF Register Map */
diff --git a/src/cpu/samsung/exynos5250/pinmux.c b/src/cpu/samsung/exynos5250/pinmux.c
index 1c72353aa1..ab1b25ec3b 100644
--- a/src/cpu/samsung/exynos5250/pinmux.c
+++ b/src/cpu/samsung/exynos5250/pinmux.c
@@ -94,7 +94,7 @@ void exynos_pinmux_sdmmc2(void)
void exynos_pinmux_sdmmc3(void)
{
/*
- * TODO: Need to add defintions for GPC4 before
+ * TODO: Need to add definitions for GPC4 before
* enabling this.
*/
printk(BIOS_DEBUG, "SDMMC3 not supported yet");
diff --git a/src/cpu/samsung/exynos5250/setup.h b/src/cpu/samsung/exynos5250/setup.h
index 942a3f1ab5..9f10786ada 100644
--- a/src/cpu/samsung/exynos5250/setup.h
+++ b/src/cpu/samsung/exynos5250/setup.h
@@ -661,7 +661,7 @@ struct exynos5_phy_control;
#define MEM_TERM_EN (1 << 31) /* Termination enable for memory */
#define PHY_TERM_EN (1 << 30) /* Termination enable for PHY */
#define DMC_CTRL_SHGATE (1 << 29) /* Duration of DQS gating signal */
-#define FP_RSYNC (1 << 3) /* Force DLL resyncronization */
+#define FP_RSYNC (1 << 3) /* Force DLL resynchronization */
/* Driver strength for CK, CKE, CS & CA */
#define IMP_OUTPUT_DRV_40_OHM 0x5
@@ -676,7 +676,7 @@ struct exynos5_phy_control;
struct mem_timings;
-/* Errors that we can encourter in low-level setup */
+/* Errors that we can encounter in low-level setup */
enum {
SETUP_ERR_OK,
SETUP_ERR_RDLV_COMPLETE_TIMEOUT = -1,
diff --git a/src/cpu/samsung/exynos5250/spi.c b/src/cpu/samsung/exynos5250/spi.c
index 642ae23bb6..1c365dc5ca 100644
--- a/src/cpu/samsung/exynos5250/spi.c
+++ b/src/cpu/samsung/exynos5250/spi.c
@@ -42,7 +42,7 @@ static void exynos_spi_rx_tx(struct exynos_spi *regs, int todo,
unsigned int *rxp = (unsigned int *)(dinp + (i * (32 * 1024)));
unsigned int out_bytes, in_bytes;
- // TODO In currrent implementation, every read/write must be aligned to
+ // TODO In current implementation, every read/write must be aligned to
// 4 bytes, otherwise you may get timeout or other unexpected results.
ASSERT(todo % 4 == 0);
@@ -90,7 +90,7 @@ int exynos_spi_open(struct exynos_spi *regs)
SPI_MODE_CH_WIDTH_WORD | SPI_MODE_BUS_WIDTH_WORD);
clrbits_le32(&regs->ch_cfg, SPI_CH_CPOL_L); /* CPOL: active high */
- /* clear rx and tx channel if set priveously */
+ /* clear rx and tx channel if set previously */
clrbits_le32(&regs->ch_cfg, SPI_RX_CH_ON | SPI_TX_CH_ON);
setbits_le32(&regs->swap_cfg,
diff --git a/src/cpu/samsung/exynos5250/tmu.c b/src/cpu/samsung/exynos5250/tmu.c
index 5a871fd2ba..1b5e9c2b64 100644
--- a/src/cpu/samsung/exynos5250/tmu.c
+++ b/src/cpu/samsung/exynos5250/tmu.c
@@ -69,8 +69,8 @@ struct tmu_info exynos5250_tmu_info = {
/*
* After reading temperature code from register, compensating
- * its value and calculating celsius temperatue,
- * get current temperatue.
+ * its value and calculating celsius temperature,
+ * get current temperature.
*
* @return current temperature of the chip as sensed by TMU
*/
diff --git a/src/cpu/samsung/exynos5250/tmu.h b/src/cpu/samsung/exynos5250/tmu.h
index 03eacd2595..bda4bd7dd8 100644
--- a/src/cpu/samsung/exynos5250/tmu.h
+++ b/src/cpu/samsung/exynos5250/tmu.h
@@ -65,7 +65,7 @@ enum tmu_status_t {
TMU_STATUS_TRIPPED,
};
-/* Tmeperature threshold values for various thermal events */
+/* Temperature threshold values for various thermal events */
struct temperature_params {
/* minimum value in temperature code range */
unsigned int min_val;
diff --git a/src/cpu/samsung/exynos5250/uart.c b/src/cpu/samsung/exynos5250/uart.c
index 41dc709c97..dbf7202d35 100644
--- a/src/cpu/samsung/exynos5250/uart.c
+++ b/src/cpu/samsung/exynos5250/uart.c
@@ -37,7 +37,7 @@ static uint32_t base_port = CONFIG_CONSOLE_SERIAL_UART_ADDRESS;
* The coefficient, used to calculate the baudrate on S5P UARTs is
* calculated as
* C = UBRDIV * 16 + number_of_set_bits_in_UDIVSLOT
- * however, section 31.6.11 of the datasheet doesn't recomment using 1 for 1,
+ * however, section 31.6.11 of the datasheet doesn't recommend using 1 for 1,
* 3 for 2, ... (2^n - 1) for n, instead, they suggest using these constants:
*/
static const int udivslot[] = {
@@ -129,7 +129,7 @@ static int exynos5_uart_err_check(int op)
/*
* Read a single byte from the serial port. Returns 1 on success, 0
- * otherwise. When the function is succesfull, the character read is
+ * otherwise. When the function is successful, the character read is
* written into its argument c.
*/
static unsigned char exynos5_uart_rx_byte(void)
diff --git a/src/cpu/ti/am335x/uart.c b/src/cpu/ti/am335x/uart.c
index 395ed866d9..bd2ff44044 100644
--- a/src/cpu/ti/am335x/uart.c
+++ b/src/cpu/ti/am335x/uart.c
@@ -147,7 +147,7 @@ static void am335x_uart_init_dev(void)
/*
* Read a single byte from the serial port. Returns 1 on success, 0
- * otherwise. When the function is succesfull, the character read is
+ * otherwise. When the function is successful, the character read is
* written into its argument c.
*/
static unsigned char am335x_uart_rx_byte(void)
diff --git a/src/cpu/ti/am335x/uart.h b/src/cpu/ti/am335x/uart.h
index 56d4c51230..645cd43743 100644
--- a/src/cpu/ti/am335x/uart.h
+++ b/src/cpu/ti/am335x/uart.h
@@ -130,7 +130,7 @@ struct am335x_uart {
uint8_t rsvd_0x36[2];
uint16_t blr; /* BOF control */
uint8_t rsvd_0x3a[2];
- uint16_t acreg; /* auxilliary control */
+ uint16_t acreg; /* auxiliary control */
uint8_t rsvd_0x3e[2];
/* 0x40 */
@@ -139,7 +139,7 @@ struct am335x_uart {
uint16_t ssr; /* supplementary status */
uint8_t rsvd_0x46[2];
- uint16_t eblr; /* BOF length (operatoinal mode only) */
+ uint16_t eblr; /* BOF length (operational mode only) */
uint8_t rsvd_0x4a[6];
/* 0x50 */
diff --git a/src/cpu/via/nano/update_ucode.c b/src/cpu/via/nano/update_ucode.c
index d5757f87a7..7471928641 100644
--- a/src/cpu/via/nano/update_ucode.c
+++ b/src/cpu/via/nano/update_ucode.c
@@ -37,7 +37,7 @@ static ucode_update_status nano_apply_ucode(const nano_ucode_header *ucode)
msr.hi = 0;
wrmsr(MSR_IA32_BIOS_UPDT_TRIG, msr);
- /* Let's see if we updated succesfully */
+ /* Let's see if we updated successfully */
msr = rdmsr(MSR_UCODE_UPDATE_STATUS);
return msr.lo & 0x07;
@@ -80,7 +80,7 @@ static void nano_print_ucode_status(ucode_update_status stat)
switch(stat)
{
case UCODE_UPDATE_SUCCESS:
- printk(BIOS_INFO, "Microcode update succesful.\n");
+ printk(BIOS_INFO, "Microcode update successful.\n");
break;
case UCODE_UPDATE_FAIL:
printk(BIOS_ALERT, "Microcode update failed, bad environment."
diff --git a/src/cpu/via/nano/update_ucode.h b/src/cpu/via/nano/update_ucode.h
index 6a22d183a7..ae58fb228d 100644
--- a/src/cpu/via/nano/update_ucode.h
+++ b/src/cpu/via/nano/update_ucode.h
@@ -56,11 +56,11 @@ typedef struct {
u32 applicable_fms; /* Fam/model/stepping to which ucode applies */
u32 checksum; /* Two's complement checksum of ucode+header */
u32 loader_revision; /* Revision of hardware ucode update loader*/
- u32 rfu_1; /* Reservod for future use */
+ u32 rfu_1; /* Reserved for future use */
u32 payload_size; /* Size of the ucode payload only */
u32 total_size; /* Size of the ucode, including header */
char name[8]; /* ASCII string of ucode filename */
- u32 rfu_2; /* Reservod for future use */
+ u32 rfu_2; /* Reserved for future use */
/* First double-word of the ucode payload
* Its address represents the beginning of the ucode update we need to
* send to the CPU */
diff --git a/src/cpu/x86/car.c b/src/cpu/x86/car.c
index 88f280796b..87fa98b488 100644
--- a/src/cpu/x86/car.c
+++ b/src/cpu/x86/car.c
@@ -33,7 +33,7 @@ extern char _car_data_end[];
/*
* The car_migrated global variable determines if the cache-as-ram space has
- * been migrated to real RAM. It does this by asumming the following things:
+ * been migrated to real RAM. It does this by assuming the following things:
* 1. cache-as-ram space is zero'd out once it is set up.
* 2. Either the cache-as-ram space is memory-backed after getting torn down
* or the space returns 0xff's for each byte read.
diff --git a/src/cpu/x86/lapic/lapic_cpu_init.c b/src/cpu/x86/lapic/lapic_cpu_init.c
index fbc8aa4e45..a3bab4658a 100644
--- a/src/cpu/x86/lapic/lapic_cpu_init.c
+++ b/src/cpu/x86/lapic/lapic_cpu_init.c
@@ -40,7 +40,7 @@
* will return 0, meaning no CPU.
*
* We actually handling that case by noting which cpus startup
- * and not telling anyone about the ones that dont.
+ * and not telling anyone about the ones that don't.
*/
/* Start-UP IPI vector must be 4kB aligned and below 1MB. */
diff --git a/src/cpu/x86/mtrr/earlymtrr.c b/src/cpu/x86/mtrr/earlymtrr.c
index 55dbd2f4a4..f16da279c6 100644
--- a/src/cpu/x86/mtrr/earlymtrr.c
+++ b/src/cpu/x86/mtrr/earlymtrr.c
@@ -32,7 +32,7 @@ static void cache_ramstage(void)
const int addr_det = 0;
/* the fixed and variable MTTRs are power-up with random values,
- * clear them to MTRR_TYPE_UNCACHEABLE for safty.
+ * clear them to MTRR_TYPE_UNCACHEABLE for safety.
*/
static void do_early_mtrr_init(const unsigned long *mtrr_msrs)
{
@@ -43,7 +43,7 @@ static void do_early_mtrr_init(const unsigned long *mtrr_msrs)
msr_t msr;
const unsigned long *msr_addr;
- /* Inialize all of the relevant msrs to 0 */
+ /* Initialize all of the relevant msrs to 0 */
msr.lo = 0;
msr.hi = 0;
unsigned long msr_nr;
diff --git a/src/cpu/x86/mtrr/mtrr.c b/src/cpu/x86/mtrr/mtrr.c
index b69787bf4a..8f1c35ec6e 100644
--- a/src/cpu/x86/mtrr/mtrr.c
+++ b/src/cpu/x86/mtrr/mtrr.c
@@ -94,7 +94,7 @@ static inline unsigned int fms(unsigned int x)
return r;
}
-/* fls: find least sigificant bit set */
+/* fls: find least significant bit set */
static inline unsigned int fls(unsigned int x)
{
int r;
@@ -160,8 +160,8 @@ static struct memranges *get_physical_address_space(void)
static struct memranges addr_space_storage;
/* In order to handle some chipsets not being able to pre-determine
- * uncacheable ranges, such as graphics memory, at resource inseration
- * time remove unacheable regions from the cacheable ones. */
+ * uncacheable ranges, such as graphics memory, at resource insertion
+ * time remove uncacheable regions from the cacheable ones. */
if (addr_space == NULL) {
struct range_entry *r;
unsigned long mask;
@@ -216,7 +216,7 @@ static struct memranges *get_physical_address_space(void)
}
/* Fixed MTRR descriptor. This structure defines the step size and begin
- * and end (exclusive) address covered by a set of fixe MTRR MSRs.
+ * and end (exclusive) address covered by a set of fixed MTRR MSRs.
* It also describes the offset in byte intervals to store the calculated MTRR
* type in an array. */
struct fixed_mtrr_desc {
@@ -533,7 +533,7 @@ static void calc_var_mtrrs_with_hole(struct var_mtrr_state *var_state,
struct range_entry *next;
/*
- * Determine MTRRs based on the following algoirthm for the given entry:
+ * Determine MTRRs based on the following algorithm for the given entry:
* +------------------+ b2 = ALIGN_UP(end)
* | 0 or more bytes | <-- hole is carved out between b1 and b2
* +------------------+ a2 = b1 = end
@@ -571,7 +571,7 @@ static void calc_var_mtrrs_with_hole(struct var_mtrr_state *var_state,
b1 = a2;
- /* First check if a1 is >= 4GiB and the current etnry is the last
+ /* First check if a1 is >= 4GiB and the current entry is the last
* entry. If so perform an optimization of covering a larger range
* defined by the base address' alignment. */
if (a1 >= RANGE_4GB && next == NULL) {
@@ -686,10 +686,10 @@ static int calc_var_mtrrs(struct memranges *addr_space,
* 1. UC as default type with no holes at top of range.
* 2. UC as default using holes at top of range.
* 3. WB as default.
- * The lowest count is then used as default after totalling all
- * MTRRs. Note that the optimal algoirthm for UC default is marked in
+ * The lowest count is then used as default after totaling all
+ * MTRRs. Note that the optimal algorithm for UC default is marked in
* the tag of each range regardless of final decision. UC takes
- * precedence in the MTRR archiecture. Therefore, only holes can be
+ * precedence in the MTRR architecture. Therefore, only holes can be
* used when the type of the region is MTRR_TYPE_WRBACK with
* MTRR_TYPE_UNCACHEABLE as the default type.
*/
diff --git a/src/cpu/x86/smm/smm_module_loader.c b/src/cpu/x86/smm/smm_module_loader.c
index 5eb4c5a0a2..478ae8c10a 100644
--- a/src/cpu/x86/smm/smm_module_loader.c
+++ b/src/cpu/x86/smm/smm_module_loader.c
@@ -24,16 +24,16 @@
#include <console/console.h>
/*
- * Compoments that make up the SMRAM:
+ * Components that make up the SMRAM:
* 1. Save state - the total save state memory used
* 2. Stack - stacks for the CPUs in the SMM handler
* 3. Stub - SMM stub code for calling into handler
* 4. Handler - C-based SMM handler.
*
- * The compoents are assumed to consist of one consecutive region.
+ * The components are assumed to consist of one consecutive region.
*/
-/* These paramters are used by the SMM stub code. A pointer to the params
+/* These parameters are used by the SMM stub code. A pointer to the params
* is also passed to the C-base handler. */
struct smm_stub_params {
u32 stack_size;
@@ -80,7 +80,7 @@ static void smm_place_jmp_instructions(void *entry_start, int stride, int num,
/* Each entry point has an IP value of 0x8000. The SMBASE for each
* cpu is different so the effective address of the entry instruction
- * is different. Therefore, the relative displacment for each entry
+ * is different. Therefore, the relative displacement for each entry
* instruction needs to be updated to reflect the current effective
* IP. Additionally, the IP result from the jmp instruction is
* calculated using the next instruction's address so the size of
@@ -140,7 +140,7 @@ static void smm_stub_place_staggered_entry_points(char *base,
stub_entry_offset = rmodule_entry_offset(smm_stub);
/* If there are staggered entry points or the stub is not located
- * at the SMM entry point then jmp instructionss need to be placed. */
+ * at the SMM entry point then jmp instructions need to be placed. */
if (params->num_concurrent_save_states > 1 || stub_entry_offset != 0) {
int num_entries;
@@ -297,7 +297,7 @@ int smm_setup_relocation_handler(struct smm_loader_params *params)
return -1;
/* Since the relocation handler always uses stack, adjust the number
- * of conccurent stack users to be CONFIG_MAX_CPUS. */
+ * of concurrent stack users to be CONFIG_MAX_CPUS. */
if (params->num_concurrent_stacks == 0)
params->num_concurrent_stacks = CONFIG_MAX_CPUS;
@@ -318,7 +318,7 @@ int smm_setup_relocation_handler(struct smm_loader_params *params)
*
* It should be noted that this algorithm will not work for
* SMM_DEFAULT_SIZE SMRAM regions such as the A segment. This algorithm
- * expectes a region large enough to encompass the handler and stacks
+ * expects a region large enough to encompass the handler and stacks
* as well as the SMM_DEFAULT_SIZE.
*/
int smm_load_module(void *smram, int size, struct smm_loader_params *params)