aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorDavid S. Miller <davem@davemloft.net>2009-04-01 16:15:20 -0700
committerDavid S. Miller <davem@davemloft.net>2009-06-16 04:56:23 -0700
commit73fffc037e2383a0ed126d57bdcda9b369769ae8 (patch)
tree9099954565aac8e2ea1b0e040662d6ca14844027
parentb696fdc259f0d94348a9327bed352fac44d4883d (diff)
sparc64: Get rid of real_setup_per_cpu_areas().
Now that we defer the cpu_data() initializations to the end of per-cpu setup, we can get rid of this local hack we had to setup the per-cpu areas eary. This is a necessary step in order to support HAVE_DYNAMIC_PER_CPU_AREA since the per-cpu setup must run when page structs are available. Signed-off-by: David S. Miller <davem@davemloft.net>
-rw-r--r--arch/sparc/include/asm/percpu_64.h4
-rw-r--r--arch/sparc/kernel/smp_64.c11
-rw-r--r--arch/sparc/mm/init_64.c7
3 files changed, 5 insertions, 17 deletions
diff --git a/arch/sparc/include/asm/percpu_64.h b/arch/sparc/include/asm/percpu_64.h
index c0ab102d11f..007aafb4ae9 100644
--- a/arch/sparc/include/asm/percpu_64.h
+++ b/arch/sparc/include/asm/percpu_64.h
@@ -9,8 +9,6 @@ register unsigned long __local_per_cpu_offset asm("g5");
#include <asm/trap_block.h>
-extern void real_setup_per_cpu_areas(void);
-
#define __per_cpu_offset(__cpu) \
(trap_block[(__cpu)].__per_cpu_base)
#define per_cpu_offset(x) (__per_cpu_offset(x))
@@ -19,8 +17,6 @@ extern void real_setup_per_cpu_areas(void);
#else /* ! SMP */
-#define real_setup_per_cpu_areas() do { } while (0)
-
#endif /* SMP */
#include <asm-generic/percpu.h>
diff --git a/arch/sparc/kernel/smp_64.c b/arch/sparc/kernel/smp_64.c
index 045fbb554a9..5302344e3cb 100644
--- a/arch/sparc/kernel/smp_64.c
+++ b/arch/sparc/kernel/smp_64.c
@@ -20,7 +20,7 @@
#include <linux/cache.h>
#include <linux/jiffies.h>
#include <linux/profile.h>
-#include <linux/lmb.h>
+#include <linux/bootmem.h>
#include <linux/cpu.h>
#include <asm/head.h>
@@ -1371,9 +1371,9 @@ void smp_send_stop(void)
{
}
-void __init real_setup_per_cpu_areas(void)
+void __init setup_per_cpu_areas(void)
{
- unsigned long base, shift, paddr, goal, size, i;
+ unsigned long base, shift, goal, size, i;
char *ptr;
/* Copy section for each CPU (we discard the original) */
@@ -1383,13 +1383,12 @@ void __init real_setup_per_cpu_areas(void)
for (size = PAGE_SIZE; size < goal; size <<= 1UL)
shift++;
- paddr = lmb_alloc(size * NR_CPUS, PAGE_SIZE);
- if (!paddr) {
+ ptr = __alloc_bootmem(size * NR_CPUS, PAGE_SIZE, 0);
+ if (!ptr) {
prom_printf("Cannot allocate per-cpu memory.\n");
prom_halt();
}
- ptr = __va(paddr);
base = ptr - __per_cpu_start;
for (i = 0; i < NR_CPUS; i++, ptr += size) {
diff --git a/arch/sparc/mm/init_64.c b/arch/sparc/mm/init_64.c
index 785f0a24fcb..b5a5932def7 100644
--- a/arch/sparc/mm/init_64.c
+++ b/arch/sparc/mm/init_64.c
@@ -1679,11 +1679,6 @@ pgd_t swapper_pg_dir[2048];
static void sun4u_pgprot_init(void);
static void sun4v_pgprot_init(void);
-/* Dummy function */
-void __init setup_per_cpu_areas(void)
-{
-}
-
void __init paging_init(void)
{
unsigned long end_pfn, shift, phys_base;
@@ -1807,8 +1802,6 @@ void __init paging_init(void)
mdesc_populate_present_mask(CPU_MASK_ALL_PTR);
}
- real_setup_per_cpu_areas();
-
/* Once the OF device tree and MDESC have been setup, we know
* the list of possible cpus. Therefore we can allocate the
* IRQ stacks.