summaryrefslogtreecommitdiff
path: root/src/cpu
diff options
context:
space:
mode:
authorArthur Heymans <arthur@aheymans.xyz>2021-07-05 21:18:50 +0200
committerFelix Held <felix-coreboot@felixheld.de>2024-02-08 13:11:59 +0000
commit3edf840ad15154d38769c0115811906284762b11 (patch)
tree02b55264c698b3c0fb1422f41ceecf6424a0ae20 /src/cpu
parentcdea508a024376c79c8cf29648ba174867eb2750 (diff)
cpu/x86/64bit: Turn jumping to long mode into a macro
This makes it easier to reuse, e.g. if you want to do it twice in one assembly file. Change-Id: Ida861338004187e4e714be41e17c8447fa4cf935 Signed-off-by: Arthur Heymans <arthur@aheymans.xyz> Reviewed-on: https://review.coreboot.org/c/coreboot/+/79261 Tested-by: build bot (Jenkins) <no-reply@coreboot.org> Reviewed-by: Felix Held <felix-coreboot@felixheld.de>
Diffstat (limited to 'src/cpu')
-rw-r--r--src/cpu/intel/car/core2/cache_as_ram.S4
-rw-r--r--src/cpu/intel/car/non-evict/cache_as_ram.S4
-rw-r--r--src/cpu/intel/car/p4-netburst/cache_as_ram.S3
-rw-r--r--src/cpu/qemu-x86/cache_as_ram_bootblock.S3
-rw-r--r--src/cpu/x86/64bit/entry64.inc12
-rw-r--r--src/cpu/x86/64bit/mode_switch.S5
-rw-r--r--src/cpu/x86/64bit/mode_switch2.S4
-rw-r--r--src/cpu/x86/sipi_vector.S4
-rw-r--r--src/cpu/x86/smm/smm_stub.S3
9 files changed, 25 insertions, 17 deletions
diff --git a/src/cpu/intel/car/core2/cache_as_ram.S b/src/cpu/intel/car/core2/cache_as_ram.S
index e134717b40..2e4d9c8074 100644
--- a/src/cpu/intel/car/core2/cache_as_ram.S
+++ b/src/cpu/intel/car/core2/cache_as_ram.S
@@ -4,6 +4,7 @@
#include <cpu/x86/mtrr.h>
#include <cpu/x86/cache.h>
#include <cpu/x86/post_code.h>
+#include <cpu/x86/64bit/entry64.inc>
.section .init
.global bootblock_pre_c_entry
@@ -162,8 +163,7 @@ addrsize_set_high:
subl $4, %esp
#if ENV_X86_64
-
- #include <cpu/x86/64bit/entry64.inc>
+ setup_longmode $(CONFIG_ARCH_X86_64_PGTBL_LOC)
movd %mm2, %rdi
shlq $32, %rdi
diff --git a/src/cpu/intel/car/non-evict/cache_as_ram.S b/src/cpu/intel/car/non-evict/cache_as_ram.S
index 76986ff68e..578bf03afd 100644
--- a/src/cpu/intel/car/non-evict/cache_as_ram.S
+++ b/src/cpu/intel/car/non-evict/cache_as_ram.S
@@ -4,6 +4,7 @@
#include <cpu/x86/mtrr.h>
#include <cpu/x86/cache.h>
#include <cpu/x86/post_code.h>
+#include <cpu/x86/64bit/entry64.inc>
#define NoEvictMod_MSR 0x2e0
#define BBL_CR_CTL3_MSR 0x11e
@@ -213,8 +214,7 @@ end_microcode_update:
andl $0xfffffff0, %esp
#if ENV_X86_64
-
- #include <cpu/x86/64bit/entry64.inc>
+ setup_longmode $(CONFIG_ARCH_X86_64_PGTBL_LOC)
movd %mm2, %rdi
shlq $32, %rdi
diff --git a/src/cpu/intel/car/p4-netburst/cache_as_ram.S b/src/cpu/intel/car/p4-netburst/cache_as_ram.S
index f7c023b402..32fddd6810 100644
--- a/src/cpu/intel/car/p4-netburst/cache_as_ram.S
+++ b/src/cpu/intel/car/p4-netburst/cache_as_ram.S
@@ -13,6 +13,7 @@
.global bootblock_pre_c_entry
#include <cpu/intel/car/cache_as_ram_symbols.inc>
+#include <cpu/x86/64bit/entry64.inc>
.code32
_cache_as_ram_setup:
@@ -362,7 +363,7 @@ fill_cache:
subl $4, %esp
#if ENV_X86_64
- #include <cpu/x86/64bit/entry64.inc>
+ setup_longmode $(CONFIG_ARCH_X86_64_PGTBL_LOC)
movd %mm2, %rdi
shlq $32, %rdi /* BIST */
diff --git a/src/cpu/qemu-x86/cache_as_ram_bootblock.S b/src/cpu/qemu-x86/cache_as_ram_bootblock.S
index 859b760175..e2662c2945 100644
--- a/src/cpu/qemu-x86/cache_as_ram_bootblock.S
+++ b/src/cpu/qemu-x86/cache_as_ram_bootblock.S
@@ -1,6 +1,7 @@
/* SPDX-License-Identifier: GPL-2.0-only */
#include <cpu/x86/post_code.h>
+#include <cpu/x86/64bit/entry64.inc>
#define CBFS_FILE_MAGIC 0
#define CBFS_FILE_LEN (CBFS_FILE_MAGIC + 8)
@@ -79,7 +80,7 @@ pages_done:
#if ENV_X86_64
/* entry64.inc preserves ebx. */
-#include <cpu/x86/64bit/entry64.inc>
+ setup_longmode $(CONFIG_ARCH_X86_64_PGTBL_LOC)
/* Restore the BIST result and timestamps. */
movd %mm2, %rdi
diff --git a/src/cpu/x86/64bit/entry64.inc b/src/cpu/x86/64bit/entry64.inc
index 7da68b47f9..878f310843 100644
--- a/src/cpu/x86/64bit/entry64.inc
+++ b/src/cpu/x86/64bit/entry64.inc
@@ -22,10 +22,9 @@
#include <arch/rom_segs.h>
#endif
-
-setup_longmode:
+.macro setup_longmode page_table
/* Get page table address */
- movl $(CONFIG_ARCH_X86_64_PGTBL_LOC), %eax
+ movl \page_table, %eax
/* load identity mapped page tables */
movl %eax, %cr3
@@ -48,12 +47,13 @@ setup_longmode:
/* use long jump to switch to 64-bit code segment */
#if defined(__RAMSTAGE__)
- ljmp $RAM_CODE_SEG64, $__longmode_start
+ ljmp $RAM_CODE_SEG64, $jmp_addr\@
#else
- ljmp $ROM_CODE_SEG64, $__longmode_start
+ ljmp $ROM_CODE_SEG64, $jmp_addr\@
#endif
.code64
-__longmode_start:
+jmp_addr\@:
+.endm
#endif
diff --git a/src/cpu/x86/64bit/mode_switch.S b/src/cpu/x86/64bit/mode_switch.S
index c4198f3dad..01fe003cb1 100644
--- a/src/cpu/x86/64bit/mode_switch.S
+++ b/src/cpu/x86/64bit/mode_switch.S
@@ -1,4 +1,5 @@
/* SPDX-License-Identifier: GPL-2.0-only */
+#include <cpu/x86/64bit/entry64.inc>
/* Calls a x86_32 function from x86_64 context */
.text
@@ -42,8 +43,8 @@ protected_mode_call_wrapper:
call *%ebx
movl %eax, %ebx
- /* Jump to long mode. Preserves ebx */
- #include <cpu/x86/64bit/entry64.inc>
+ /* Preserves ebx */
+ setup_longmode $(CONFIG_ARCH_X86_64_PGTBL_LOC)
/* Place return value in rax */
movl %ebx, %eax
diff --git a/src/cpu/x86/64bit/mode_switch2.S b/src/cpu/x86/64bit/mode_switch2.S
index 65e9d94a98..1807d2e404 100644
--- a/src/cpu/x86/64bit/mode_switch2.S
+++ b/src/cpu/x86/64bit/mode_switch2.S
@@ -4,6 +4,8 @@
* Must not be directly invoked from C code!
*/
+#include <cpu/x86/64bit/entry64.inc>
+
.text
.code32
.section ".text.long_mode_call_3arg", "ax", @progbits
@@ -19,7 +21,7 @@ long_mode_call_3arg:
mov %esp, %ebp
/* Enter long mode, preserves ebx */
- #include <cpu/x86/64bit/entry64.inc>
+ setup_longmode $(CONFIG_ARCH_X86_64_PGTBL_LOC)
/* Align stack */
movabs $0xfffffffffffffff0, %rax
diff --git a/src/cpu/x86/sipi_vector.S b/src/cpu/x86/sipi_vector.S
index 39973dbc8b..923e3987d0 100644
--- a/src/cpu/x86/sipi_vector.S
+++ b/src/cpu/x86/sipi_vector.S
@@ -6,6 +6,7 @@
#include <arch/ram_segs.h>
#define __RAMSTAGE__
+#include <cpu/x86/64bit/entry64.inc>
/* The SIPI vector is responsible for initializing the APs in the system. It
* loads microcode, sets up MSRs, and enables caching before calling into
@@ -223,7 +224,8 @@ load_msr:
#if ENV_X86_64
/* entry64.inc preserves ebx, esi, edi, ebp */
-#include <cpu/x86/64bit/entry64.inc>
+ setup_longmode $(CONFIG_ARCH_X86_64_PGTBL_LOC)
+
movabs c_handler, %eax
call *%rax
#else
diff --git a/src/cpu/x86/smm/smm_stub.S b/src/cpu/x86/smm/smm_stub.S
index f97ab59cd9..9f1f21d974 100644
--- a/src/cpu/x86/smm/smm_stub.S
+++ b/src/cpu/x86/smm/smm_stub.S
@@ -12,6 +12,7 @@
#include <cpu/x86/cr.h>
#include <cpu/x86/msr.h>
#include <cpu/x86/lapic_def.h>
+#include <cpu/x86/64bit/entry64.inc>
.code32
.section ".module_parameters", "aw", @progbits
@@ -195,7 +196,7 @@ align_stack:
#if ENV_X86_64
mov %ecx, %edi
/* entry64.inc preserves ebx, esi, edi, ebp */
-#include <cpu/x86/64bit/entry64.inc>
+ setup_longmode $(CONFIG_ARCH_X86_64_PGTBL_LOC)
mov %edi, %ecx