summaryrefslogtreecommitdiff
path: root/src/arch/arm64
diff options
context:
space:
mode:
authorFurquan Shaikh <furquan@google.com>2015-10-15 12:15:31 -0700
committerJulius Werner <jwerner@chromium.org>2015-11-07 03:29:14 +0100
commitb3f6ad35221984419ee0998f47b778d669d1636e (patch)
tree306fc113435cb78c63d2950db33af8f577754d51 /src/arch/arm64
parent1148786c05d97b4c646c11e770b275809b562953 (diff)
arm64: Remove SMP support
As ARM Trusted Firmware is the only first class citizen for booting arm64 multi-processor in coreboot remove SMP support. If SoCs want to bring up MP then ATF needs to be ported and integrated. Change-Id: Ife24d53eed9b7a5a5d8c69a64d7a20a55a4163db Signed-off-by: Furquan Shaikh <furquan@google.com> Reviewed-on: http://review.coreboot.org/11909 Tested-by: build bot (Jenkins) Reviewed-by: Julius Werner <jwerner@chromium.org>
Diffstat (limited to 'src/arch/arm64')
-rw-r--r--src/arch/arm64/Makefile.inc2
-rw-r--r--src/arch/arm64/c_entry.c18
-rw-r--r--src/arch/arm64/cpu-internal.h20
-rw-r--r--src/arch/arm64/cpu.c224
-rw-r--r--src/arch/arm64/cpu_ramstage.c92
-rw-r--r--src/arch/arm64/include/arch/startup.h44
-rw-r--r--src/arch/arm64/include/armv8/arch/cpu.h134
-rw-r--r--src/arch/arm64/include/armv8/arch/smp/spinlock.h29
-rw-r--r--src/arch/arm64/stage_entry.S122
-rw-r--r--src/arch/arm64/startup.c53
10 files changed, 34 insertions, 704 deletions
diff --git a/src/arch/arm64/Makefile.inc b/src/arch/arm64/Makefile.inc
index fabe7b3241..b4ee3a5101 100644
--- a/src/arch/arm64/Makefile.inc
+++ b/src/arch/arm64/Makefile.inc
@@ -130,9 +130,7 @@ ifeq ($(CONFIG_ARCH_RAMSTAGE_ARM64),y)
ramstage-y += c_entry.c
ramstage-y += stages.c
-ramstage-y += startup.c
ramstage-y += div0.c
-ramstage-y += cpu.c
ramstage-y += cpu_ramstage.c
ramstage-y += eabi_compat.c
ramstage-y += boot.c
diff --git a/src/arch/arm64/c_entry.c b/src/arch/arm64/c_entry.c
index 0a5541c4ab..deef83df7e 100644
--- a/src/arch/arm64/c_entry.c
+++ b/src/arch/arm64/c_entry.c
@@ -17,7 +17,6 @@
#include <arch/cpu.h>
#include <arch/mmu.h>
#include <arch/stages.h>
-#include <arch/startup.h>
#include "cpu-internal.h"
void __attribute__((weak)) arm64_soc_init(void)
@@ -32,7 +31,7 @@ static void seed_stack(void)
int i;
int size;
- stack_begin = cpu_get_stack(smp_processor_id());
+ stack_begin = cpu_get_stack();
stack_begin -= CONFIG_STACK_SIZE;
slot = (void *)stack_begin;
@@ -45,21 +44,10 @@ static void seed_stack(void)
static void arm64_init(void)
{
- cpu_set_bsp();
seed_stack();
arm64_soc_init();
main();
}
-/*
- * This variable holds entry point for CPUs starting up. The first
- * element is the BSP path, and the second is the non-BSP path.
- */
-void (*c_entry[2])(void) = { &arm64_init, &arch_secondary_cpu_init };
-
-void *prepare_secondary_cpu_startup(void)
-{
- startup_save_cpu_data();
-
- return secondary_entry_point(&arm64_cpu_startup_resume);
-}
+/* This variable holds entry point for CPU starting up. */
+void (*c_entry)(void) = &arm64_init;
diff --git a/src/arch/arm64/cpu-internal.h b/src/arch/arm64/cpu-internal.h
index fb5d13fe75..75d2d4deb6 100644
--- a/src/arch/arm64/cpu-internal.h
+++ b/src/arch/arm64/cpu-internal.h
@@ -16,22 +16,10 @@
#ifndef ARCH_CPU_INTERNAL_H
#define ARCH_CPU_INTERNAL_H
-/*
- * Do the necessary work to prepare for secondary CPUs coming up. The
- * SoC will call this function before bringing up the other CPUs. The
- * entry point for the seoncdary CPUs is returned.
- */
-void *prepare_secondary_cpu_startup(void);
-
-/*
- * Code path for the non-BSP CPUs. This is an internal function used.
- */
-void arch_secondary_cpu_init(void);
-
-/* Return the top of the stack for the specified cpu. */
-void *cpu_get_stack(unsigned int cpu);
+/* Return the top of the stack for the cpu. */
+void *cpu_get_stack(void);
-/* Return the top of the exception stack for the specified cpu. */
-void *cpu_get_exception_stack(unsigned int cpu);
+/* Return the top of the exception stack for the cpu. */
+void *cpu_get_exception_stack(void);
#endif /* ARCH_CPU_INTERNAL_H */
diff --git a/src/arch/arm64/cpu.c b/src/arch/arm64/cpu.c
deleted file mode 100644
index 1501181364..0000000000
--- a/src/arch/arm64/cpu.c
+++ /dev/null
@@ -1,224 +0,0 @@
-/*
- * This file is part of the coreboot project.
- *
- * Copyright 2013 Google Inc.
- *
- * This program is free software; you can redistribute it and/or modify
- * it under the terms of the GNU General Public License as published by
- * the Free Software Foundation; version 2 of the License.
- *
- * This program is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
- * GNU General Public License for more details.
- */
-
-#include <stdint.h>
-#include <stdlib.h>
-#include <arch/lib_helpers.h>
-#include <cpu/cpu.h>
-#include <console/console.h>
-#include <smp/node.h>
-#include "cpu-internal.h"
-
-struct cpu_info cpu_infos[CONFIG_MAX_CPUS];
-struct cpu_info *bsp_cpu_info;
-
-struct cpu_info *cpu_info(void)
-{
- return cpu_info_for_cpu(smp_processor_id());
-}
-
-size_t cpus_online(void)
-{
- int i;
- size_t num = 0;
-
- for (i = 0; i < ARRAY_SIZE(cpu_infos); i++) {
- if (cpu_online(cpu_info_for_cpu(i)))
- num++;
- }
-
- return num;
-}
-
-static inline int action_queue_empty(struct cpu_action_queue *q)
-{
- return load_acquire_exclusive(&q->todo) == NULL;
-}
-
-static inline int action_completed(struct cpu_action_queue *q,
- struct cpu_action *action)
-{
- return load_acquire(&q->completed) == action;
-}
-
-static inline void wait_for_action_queue_slot(struct cpu_action_queue *q)
-{
- while (!action_queue_empty(q))
- wfe();
-}
-
-static void wait_for_action_complete(struct cpu_action_queue *q,
- struct cpu_action *a)
-{
- while (!action_completed(q, a))
- wfe();
-}
-
-static struct cpu_action *wait_for_action(struct cpu_action_queue *q,
- struct cpu_action *local)
-{
- struct cpu_action *action;
-
- while (action_queue_empty(q))
- wfe();
-
- /*
- * Keep original address, but use a local copy for async processing.
- */
- do {
- action = load_acquire_exclusive(&q->todo);
- *local = *action;
- } while (!store_release_exclusive(&q->todo, NULL));
-
- return action;
-}
-
-static void queue_action(struct cpu_action_queue *q, struct cpu_action *action)
-{
- do {
- wait_for_action_queue_slot(q);
- if (load_acquire_exclusive(&q->todo) != NULL)
- continue;
- } while (!store_release_exclusive(&q->todo, action));
-}
-
-static void action_queue_complete(struct cpu_action_queue *q,
- struct cpu_action *action)
-{
- /* Mark completion and send events to waiters. */
- store_release(&q->completed, action);
- sev();
-}
-
-static void action_run(struct cpu_action *action)
-{
- action->run(action->arg);
-}
-
-static void action_run_on_cpu(struct cpu_info *ci, struct cpu_action *action,
- int sync)
-{
- struct cpu_action_queue *q = &ci->action_queue;
-
- /* Don't run actions on non-online cpus. */
- if (!cpu_online(ci))
- return;
-
- if (ci->id == smp_processor_id()) {
- action->run(action->arg);
- return;
- }
-
- queue_action(q, action);
- /* Wait for CPU to pick it up. Empty slot means it was picked up. */
- wait_for_action_queue_slot(q);
- /* Wait for completion if requested. */
- if (sync)
- wait_for_action_complete(q, action);
-}
-
-static int __arch_run_on_cpu(unsigned int cpu, struct cpu_action *action,
- int sync)
-{
- struct cpu_info *ci;
-
- if (cpu >= CONFIG_MAX_CPUS)
- return -1;
-
- ci = cpu_info_for_cpu(cpu);
-
- action_run_on_cpu(ci, action, sync);
-
- return 0;
-}
-
-int arch_run_on_cpu(unsigned int cpu, struct cpu_action *action)
-{
- return __arch_run_on_cpu(cpu, action, 1);
-}
-
-int arch_run_on_cpu_async(unsigned int cpu, struct cpu_action *action)
-{
- return __arch_run_on_cpu(cpu, action, 0);
-}
-
-static int __arch_run_on_all_cpus(struct cpu_action *action, int sync)
-{
- int i;
-
- for (i = 0; i < CONFIG_MAX_CPUS; i++)
- action_run_on_cpu(cpu_info_for_cpu(i), action, sync);
-
- return 0;
-}
-
-static int __arch_run_on_all_cpus_but_self(struct cpu_action *action, int sync)
-{
- int i;
- struct cpu_info *me = cpu_info();
-
- for (i = 0; i < CONFIG_MAX_CPUS; i++) {
- struct cpu_info *ci = cpu_info_for_cpu(i);
- if (ci == me)
- continue;
- action_run_on_cpu(ci, action, sync);
- }
-
- return 0;
-}
-
-int arch_run_on_all_cpus(struct cpu_action *action)
-{
- return __arch_run_on_all_cpus(action, 1);
-}
-
-int arch_run_on_all_cpus_async(struct cpu_action *action)
-{
- return __arch_run_on_all_cpus(action, 0);
-}
-
-int arch_run_on_all_cpus_but_self(struct cpu_action *action)
-{
- return __arch_run_on_all_cpus_but_self(action, 1);
-}
-
-int arch_run_on_all_cpus_but_self_async(struct cpu_action *action)
-{
- return __arch_run_on_all_cpus_but_self(action, 0);
-}
-
-
-void arch_cpu_wait_for_action(void)
-{
- struct cpu_info *ci = cpu_info();
- struct cpu_action_queue *q = &ci->action_queue;
-
- while (1) {
- struct cpu_action *orig;
- struct cpu_action action;
-
- orig = wait_for_action(q, &action);
-
- action_run(&action);
- action_queue_complete(q, orig);
- }
-}
-
-#if IS_ENABLED(CONFIG_SMP)
-int boot_cpu(void)
-{
- return cpu_is_bsp();
-}
-#endif
diff --git a/src/arch/arm64/cpu_ramstage.c b/src/arch/arm64/cpu_ramstage.c
index 81df98cfe1..c94c26fad1 100644
--- a/src/arch/arm64/cpu_ramstage.c
+++ b/src/arch/arm64/cpu_ramstage.c
@@ -24,6 +24,8 @@
#include <timer.h>
#include "cpu-internal.h"
+static struct cpu_info cpu_info;
+
void __attribute__((weak)) arm64_arch_timer_init(void)
{
/* Default weak implementation does nothing. */
@@ -102,9 +104,9 @@ static void el3_init(void)
isb();
}
-static void init_this_cpu(void *arg)
+static void init_this_cpu(void)
{
- struct cpu_info *ci = arg;
+ struct cpu_info *ci = &cpu_info;
device_t dev = ci->cpu;
cpu_set_device_operations(dev);
@@ -156,14 +158,14 @@ static void init_cpu_info(struct bus *bus)
continue;
/* IDs are currently mapped 1:1 with logical CPU numbers. */
- if (id >= CONFIG_MAX_CPUS) {
+ if (id != 0) {
printk(BIOS_WARNING,
"CPU id %x too large. Disabling.\n", id);
cpu_disable_dev(cur);
continue;
}
- ci = cpu_info_for_cpu(id);
+ ci = &cpu_info;
if (ci->cpu != NULL) {
printk(BIOS_WARNING,
"Duplicate ID %x in device tree.\n", id);
@@ -175,12 +177,8 @@ static void init_cpu_info(struct bus *bus)
}
}
-void arch_initialize_cpus(device_t cluster, struct cpu_control_ops *cntrl_ops)
+void arch_initialize_cpu(device_t cluster)
{
- size_t max_cpus;
- size_t i;
- struct cpu_info *ci;
- void (*entry)(void);
struct bus *bus;
if (cluster->path.type != DEVICE_PATH_CPU_CLUSTER) {
@@ -196,83 +194,11 @@ void arch_initialize_cpus(device_t cluster, struct cpu_control_ops *cntrl_ops)
if (bus == NULL)
return;
- /*
- * el3_init must be performed prior to prepare_secondary_cpu_startup.
- * This is important since el3_init initializes SCR values on BSP CPU
- * and then prepare_secondary_cpu_startup reads the initialized SCR
- * value and saves it for use by non-BSP CPUs.
- */
el3_init();
- /* Mark current cpu online. */
- cpu_mark_online(cpu_info());
- entry = prepare_secondary_cpu_startup();
/* Initialize the cpu_info structures. */
init_cpu_info(bus);
- max_cpus = cntrl_ops->total_cpus();
-
- if (max_cpus > CONFIG_MAX_CPUS) {
- printk(BIOS_WARNING,
- "max_cpus (%zu) exceeds CONFIG_MAX_CPUS (%zu).\n",
- max_cpus, (size_t)CONFIG_MAX_CPUS);
- max_cpus = CONFIG_MAX_CPUS;
- }
-
- for (i = 0; i < max_cpus; i++) {
- device_t dev;
- struct cpu_action action;
- struct stopwatch sw;
-
- ci = cpu_info_for_cpu(i);
- dev = ci->cpu;
-
- /* Disregard CPUs not in device tree. */
- if (dev == NULL)
- continue;
-
- /* Skip disabled CPUs. */
- if (!dev->enabled)
- continue;
-
- if (!cpu_online(ci)) {
- /* Start the CPU. */
- printk(BIOS_DEBUG, "Starting CPU%x\n", ci->id);
-
- if (cntrl_ops->start_cpu(ci->id, entry)) {
- printk(BIOS_ERR,
- "Failed to start CPU%x\n", ci->id);
- continue;
- }
- stopwatch_init_msecs_expire(&sw, 1000);
- /* Wait for CPU to come online. */
- while (!stopwatch_expired(&sw)) {
- if (!cpu_online(ci))
- continue;
- printk(BIOS_DEBUG,
- "CPU%x online in %ld usecs.\n",
- ci->id, stopwatch_duration_usecs(&sw));
- break;
- }
- }
-
- if (!cpu_online(ci)) {
- printk(BIOS_DEBUG,
- "CPU%x failed to come online in %ld usecs.\n",
- ci->id, stopwatch_duration_usecs(&sw));
- continue;
- }
-
- /* Send it the init action. */
- action.run = init_this_cpu;
- action.arg = ci;
- arch_run_on_cpu(ci->id, &action);
- }
-}
-
-void arch_secondary_cpu_init(void)
-{
- /* Mark this CPU online. */
- cpu_mark_online(cpu_info());
- arch_cpu_wait_for_action();
+ /* Send it the init action. */
+ init_this_cpu();
}
diff --git a/src/arch/arm64/include/arch/startup.h b/src/arch/arm64/include/arch/startup.h
deleted file mode 100644
index def0b378fd..0000000000
--- a/src/arch/arm64/include/arch/startup.h
+++ /dev/null
@@ -1,44 +0,0 @@
-/*
- * This file is part of the coreboot project.
- *
- * Copyright (C) 2014 Google Inc
- *
- * This program is free software; you can redistribute it and/or
- * modify it under the terms of the GNU General Public License as
- * published by the Free Software Foundation; version 2 of
- * the License.
- *
- * This program is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
- * GNU General Public License for more details.
- */
-
-#ifndef __ARCH_ARM64_INCLUDE_ARCH_STARTUP_H__
-#define __ARCH_ARM64_INCLUDE_ARCH_STARTUP_H__
-
-/* Every element occupies 8 bytes (64-bit entries) */
-#define PER_ELEMENT_SIZE_BYTES 8
-#define MAIR_INDEX 0
-#define TCR_INDEX 1
-#define TTBR0_INDEX 2
-#define SCR_INDEX 3
-#define VBAR_INDEX 4
-#define CNTFRQ_INDEX 5
-#define CPTR_INDEX 6
-#define CPACR_INDEX 7
-/* IMPORTANT!!! If any new element is added please update NUM_ELEMENTS */
-#define NUM_ELEMENTS 8
-
-#ifndef __ASSEMBLY__
-
-/*
- * startup_save_cpu_data is used to save register values that need to be setup
- * when a CPU starts booting. This is used by secondary CPUs as well as resume
- * path to directly setup MMU and other related registers.
- */
-void startup_save_cpu_data(void);
-
-#endif
-
-#endif /* __ARCH_ARM64_INCLUDE_ARCH_STARTUP_H__ */
diff --git a/src/arch/arm64/include/armv8/arch/cpu.h b/src/arch/arm64/include/armv8/arch/cpu.h
index d1056c5d55..778909d94c 100644
--- a/src/arch/arm64/include/armv8/arch/cpu.h
+++ b/src/arch/arm64/include/armv8/arch/cpu.h
@@ -37,148 +37,28 @@ struct cpu_driver {
const struct cpu_device_id *id_table;
};
-/* Action to run. */
-struct cpu_action {
- void (*run)(void *arg);
- void *arg;
-};
-
-/*
- * Actions are queued to 'todo'. When picked up 'todo' is cleared. The
- * 'completed' field is set to the original 'todo' value when the action
- * is complete.
- */
-struct cpu_action_queue {
- struct cpu_action *todo;
- struct cpu_action *completed;
-};
-
struct cpu_info {
device_t cpu;
- struct cpu_action_queue action_queue;
- unsigned int online;
- /* Current assumption is that id matches smp_processor_id(). */
unsigned int id;
uint64_t mpidr;
};
-/* Obtain cpu_info for current executing CPU. */
-struct cpu_info *cpu_info(void);
-
-extern struct cpu_info *bsp_cpu_info;
-extern struct cpu_info cpu_infos[CONFIG_MAX_CPUS];
-
-static inline struct cpu_info *cpu_info_for_cpu(unsigned int id)
-{
- return &cpu_infos[id];
-}
-
-/* Ran only by BSP at initial boot strapping. */
-static inline void cpu_set_bsp(void)
-{
- bsp_cpu_info = cpu_info();
-}
-
-static inline int cpu_is_bsp(void)
-{
- return cpu_info() == bsp_cpu_info;
-}
-
-static inline int cpu_online(struct cpu_info *ci)
-{
- return load_acquire(&ci->online) != 0;
-}
-
-static inline void cpu_mark_online(struct cpu_info *ci)
-{
- ci->mpidr = read_affinity_mpidr();
- store_release(&ci->online, 1);
-}
-
-/* Provide number of CPUs online. */
-size_t cpus_online(void);
-
-/* Control routines for starting CPUs. */
-struct cpu_control_ops {
- /* Return the maximum number of CPUs supported. */
- size_t (*total_cpus)(void);
- /*
- * Start the requested CPU and have it start running entry().
- * Returns 0 on success, < 0 on error.
- */
- int (*start_cpu)(unsigned int id, void (*entry)(void));
-};
-
-/*
- * Initialize all DEVICE_PATH_CPUS under the DEVICE_PATH_CPU_CLUSTER cluster.
- * type DEVICE_PATH_CPUS. Start up is controlled by cntrl_ops.
- */
-void arch_initialize_cpus(device_t cluster, struct cpu_control_ops *cntrl_ops);
-
-/*
- * Run cpu_action returning < 0 on error, 0 on success. There are synchronous
- * and asynchronous methods. Both cases ensure the action has been picked up
- * by the target cpu. The synchronous variants will wait for the action to
- * be completed before returning.
- *
- * Though the current implementation allows queuing actions on the main cpu,
- * the main cpu doesn't process its own queue.
- */
-int arch_run_on_cpu(unsigned int cpu, struct cpu_action *action);
-int arch_run_on_all_cpus(struct cpu_action *action);
-int arch_run_on_all_cpus_but_self(struct cpu_action *action);
-int arch_run_on_cpu_async(unsigned int cpu, struct cpu_action *action);
-int arch_run_on_all_cpus_async(struct cpu_action *action);
-int arch_run_on_all_cpus_but_self_async(struct cpu_action *action);
-
-/* Wait for actions to be perfomed. */
-void arch_cpu_wait_for_action(void);
+/* Initialize CPU0 under the DEVICE_PATH_CPU_CLUSTER cluster. */
+void arch_initialize_cpu(device_t cluster);
#endif /* !__PRE_RAM__ */
-/*
- * Returns logical cpu in range [0:MAX_CPUS). SoC should define this.
- * Additionally, this is needed early in arm64 init so it should not
- * rely on a stack. Standard clobber list is fair game: x0-x7 and x0
- * returns the logical cpu number.
- */
-unsigned int smp_processor_id(void);
-
-/*
- * Stages and rmodules have 2 entry points: BSP and non-BSP. Provided
- * a pointer the correct non-BSP entry point will be returned. The
- * first instruction is for BSP and the 2nd is for non-BSP. Instructions
- * are all 32-bit on arm64.
- */
-static inline void *secondary_entry_point(void *e)
-{
- uintptr_t nonbsp = (uintptr_t)e;
-
- return (void *)(nonbsp + sizeof(uint32_t));
-}
+static inline unsigned int smp_processor_id(void) { return 0; }
/*
- * The arm64_cpu_startup() initializes a CPU's exception stack and regular
- * stack as well initializing the C environment for the processor. It
- * calls into the array of function pointers at symbol c_entry depending
- * on BSP state. Note that arm64_cpu_startup contains secondary entry
- * point which can be obtained by secondary_entry_point().
+ * The arm64_cpu_startup() initializes CPU's exception stack and regular
+ * stack as well initializing the C environment for the processor. Finally it
+ * calls into c_entry.
*/
void arm64_cpu_startup(void);
/*
- * The arm64_cpu_startup_resume() initializes a CPU's exception stack and
- * regular stack as well initializing the C environment for the processor. It
- * calls into the array of function pointers at symbol c_entry depending
- * on BSP state. Note that arm64_cpu_startup contains secondary entry
- * point which can be obtained by secondary_entry_point().
- * Additionally, it also restores saved register data and enables MMU, caches
- * and exceptions before jumping to C environment for both BSP and non-BSP CPUs.
- */
-void arm64_cpu_startup_resume(void);
-
-/*
- * The arm64_arch_timer_init() initializes the per CPU's cntfrq register of
+ * The arm64_arch_timer_init() initializes the CPU's cntfrq register of
* ARM arch timer.
*/
void arm64_arch_timer_init(void);
diff --git a/src/arch/arm64/include/armv8/arch/smp/spinlock.h b/src/arch/arm64/include/armv8/arch/smp/spinlock.h
deleted file mode 100644
index 10278e72ae..0000000000
--- a/src/arch/arm64/include/armv8/arch/smp/spinlock.h
+++ /dev/null
@@ -1,29 +0,0 @@
-#ifndef ARCH_SMP_SPINLOCK_H
-#define ARCH_SMP_SPINLOCK_H
-
-#include <arch/barrier.h>
-#include <stdint.h>
-
-typedef struct {
- volatile uint32_t lock;
-} spinlock_t;
-
-#define SPIN_LOCK_UNLOCKED (spinlock_t) { 0 }
-#define DECLARE_SPIN_LOCK(x) static spinlock_t x = SPIN_LOCK_UNLOCKED;
-
-static inline void spin_lock(spinlock_t *spin)
-{
- while (1) {
- if (load_acquire_exclusive(&spin->lock) != 0)
- continue;
- if (store_release_exclusive(&spin->lock, 1))
- break;
- }
-}
-
-static inline void spin_unlock(spinlock_t *spin)
-{
- store_release(&spin->lock, 0);
-}
-
-#endif
diff --git a/src/arch/arm64/stage_entry.S b/src/arch/arm64/stage_entry.S
index 79db9ddc1b..7f113fb4fc 100644
--- a/src/arch/arm64/stage_entry.S
+++ b/src/arch/arm64/stage_entry.S
@@ -15,51 +15,41 @@
/*
* ======================== stage_entry.S =====================================
- * This file acts as an entry point to the different stages of arm64 as well as
- * for the secure monitor. They share the same process of setting up stacks and
- * jumping to c code. It is important to save x25 from corruption as it contains
- * the argument for secure monitor.
+ * This file acts as an entry point to the different stages of arm64. They share
+ * the same process of setting up stacks and jumping to c code. It is important
+ * to save x25 from corruption as it contains the argument for rmodule.
* =============================================================================
*/
#include <arch/asm.h>
#define __ASSEMBLY__
#include <arch/lib_helpers.h>
-#include <arch/startup.h>
#define STACK_SZ CONFIG_STACK_SIZE
#define EXCEPTION_STACK_SZ CONFIG_STACK_SIZE
/*
- * The stacks for each of the armv8 cores grows down from _estack. It is sized
- * according to MAX_CPUS. Additionally provide exception stacks for each CPU.
+ * Stack for armv8 CPU grows down from _estack. Additionally, provide exception
+ * stack for the CPU.
*/
.section .bss, "aw", @nobits
-.global _arm64_startup_data
-.balign 8
-_arm64_startup_data:
-.space NUM_ELEMENTS*PER_ELEMENT_SIZE_BYTES
-
.global _stack
.global _estack
.balign STACK_SZ
_stack:
-.space CONFIG_MAX_CPUS*STACK_SZ
+.space STACK_SZ
_estack:
.global _stack_exceptions
.global _estack_exceptions
.balign EXCEPTION_STACK_SZ
_stack_exceptions:
-.space CONFIG_MAX_CPUS*EXCEPTION_STACK_SZ
+.space EXCEPTION_STACK_SZ
_estack_exceptions:
ENTRY(cpu_get_stack)
- mov x1, #STACK_SZ
- mul x0, x0, x1
- ldr x1, 1f
- sub x0, x1, x0
+ ldr x0, 1f
ret
.align 3
1:
@@ -67,10 +57,7 @@ ENTRY(cpu_get_stack)
ENDPROC(cpu_get_stack)
ENTRY(cpu_get_exception_stack)
- mov x1, #EXCEPTION_STACK_SZ
- mul x0, x0, x1
- ldr x1, 1f
- sub x0, x1, x0
+ ldr x0, 1f
ret
.align 3
1:
@@ -87,11 +74,7 @@ ENDPROC(cpu_get_exception_stack)
* any rmodules.
*/
ENTRY(arm64_c_environment)
- bl smp_processor_id /* x0 = cpu */
- mov x24, x0
-
-
- /* Set the exception stack for this cpu. */
+ /* Set the exception stack for the cpu. */
bl cpu_get_exception_stack
msr SPSel, #1
isb
@@ -101,16 +84,12 @@ ENTRY(arm64_c_environment)
msr SPSel, #0
isb
- /* Set stack for this cpu. */
- mov x0, x24 /* x0 = cpu */
+ /* Set the non-exception stack for the cpu. */
bl cpu_get_stack
mov sp, x0
/* Get entry point by dereferencing c_entry. */
ldr x1, 1f
- /* Retrieve entry in c_entry array using x26 as the index. */
- adds x1, x1, x26, lsl #3
- ldr x1, [x1]
/* Move back the arguments from x25 to x0 */
mov x0, x25
br x1
@@ -119,21 +98,7 @@ ENTRY(arm64_c_environment)
.quad c_entry
ENDPROC(arm64_c_environment)
-/* The first 2 instructions are for BSP and secondary CPUs,
- * respectively. x26 holds the index into c_entry array. */
-.macro split_bsp_path
- b 2000f
- b 2001f
- 2000:
- mov x26, #0
- b 2002f
- 2001:
- mov x26, #1
- 2002:
-.endm
-
ENTRY(_start)
- split_bsp_path
/* Save any arguments to current rmodule in x25 */
mov x25, x0
b arm64_c_environment
@@ -153,77 +118,12 @@ ENDPROC(_start)
write_current sctlr, x0, x1
.endm
-/*
- * This macro assumes x2 has base address and returns value read in x0
- * x1 is used as temporary register.
- */
-.macro get_element_addr index
- add x1, x2, #(\index * PER_ELEMENT_SIZE_BYTES)
- ldr x0, [x1]
-.endm
-
-/*
- * Uses following registers:
- * x0 = reading stored value
- * x1 = temp reg
- * x2 = base address of saved data region
- */
-.macro startup_restore
- adr x2, _arm64_startup_data
-
- get_element_addr MAIR_INDEX
- write_current mair, x0, x1
-
- get_element_addr TCR_INDEX
- write_current tcr, x0, x1
-
- get_element_addr TTBR0_INDEX
- write_current ttbr0, x0, x1
-
- get_element_addr SCR_INDEX
- write_el3 scr, x0, x1
-
- get_element_addr VBAR_INDEX
- write_current vbar, x0, x1
-
- get_element_addr CNTFRQ_INDEX
- write_el0 cntfrq, x0, x1
-
- get_element_addr CPTR_INDEX
- write_el3 cptr, x0, x1
-
- get_element_addr CPACR_INDEX
- write_el1 cpacr, x0, x1
-
- dsb sy
- isb
-
- tlbiall_current x1
- read_current x0, sctlr
- orr x0, x0, #(1 << 12) /* Enable Instruction Cache */
- orr x0, x0, #(1 << 2) /* Enable Data/Unified Cache */
- orr x0, x0, #(1 << 0) /* Enable MMU */
- write_current sctlr, x0, x1
-
- dsb sy
- isb
-.endm
-
CPU_RESET_ENTRY(arm64_cpu_startup)
- split_bsp_path
bl arm64_cpu_early_setup
setup_sctlr
b arm64_c_environment
ENDPROC(arm64_cpu_startup)
-CPU_RESET_ENTRY(arm64_cpu_startup_resume)
- split_bsp_path
- bl arm64_cpu_early_setup
- setup_sctlr
- startup_restore
- b arm64_c_environment
-ENDPROC(arm64_cpu_startup_resume)
-
/*
* stage_entry is defined as a weak symbol to allow SoCs/CPUs to define a custom
* entry point to perform any fixups that need to be done immediately after
diff --git a/src/arch/arm64/startup.c b/src/arch/arm64/startup.c
deleted file mode 100644
index 6ffe3f9b29..0000000000
--- a/src/arch/arm64/startup.c
+++ /dev/null
@@ -1,53 +0,0 @@
-/*
- * This file is part of the coreboot project.
- *
- * Copyright (C) 2014 Google Inc
- *
- * This program is free software; you can redistribute it and/or
- * modify it under the terms of the GNU General Public License as
- * published by the Free Software Foundation; version 2 of
- * the License.
- *
- * This program is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
- * GNU General Public License for more details.
- */
-
-#include <arch/cache.h>
-#include <arch/lib_helpers.h>
-#include <arch/startup.h>
-#include <console/console.h>
-
-/* This space is defined in stage_entry.S. */
-extern u8 _arm64_startup_data[];
-
-static inline void save_element(size_t index, uint64_t val)
-{
- uint64_t *ptr = (uint64_t *)_arm64_startup_data;
-
- ptr[index] = val;
-}
-
-/*
- * startup_save_cpu_data is used to save register values that need to be setup
- * when a CPU starts booting. This is used by secondary CPUs as well as resume
- * path to directly setup MMU and other related registers.
- */
-void startup_save_cpu_data(void)
-{
- save_element(MAIR_INDEX, raw_read_mair_current());
- save_element(TCR_INDEX, raw_read_tcr_current());
- save_element(TTBR0_INDEX, raw_read_ttbr0_current());
- save_element(VBAR_INDEX, raw_read_vbar_current());
- save_element(CNTFRQ_INDEX, raw_read_cntfrq_el0());
- save_element(CPACR_INDEX, raw_read_cpacr_el1());
-
- if (get_current_el() == EL3) {
- save_element(SCR_INDEX, raw_read_scr_el3());
- save_element(CPTR_INDEX, raw_read_cptr_el3());
- }
-
- dcache_clean_by_mva(_arm64_startup_data,
- NUM_ELEMENTS * PER_ELEMENT_SIZE_BYTES);
-}