diff options
author | Ralf Baechle <ralf@linux-mips.org> | 2000-04-28 01:09:25 +0000 |
---|---|---|
committer | Ralf Baechle <ralf@linux-mips.org> | 2000-04-28 01:09:25 +0000 |
commit | b9ba7aeb165cffecdffb60aec8c3fa8d590d9ca9 (patch) | |
tree | 42d07b0c7246ae2536a702e7c5de9e2732341116 /arch/ppc | |
parent | 7406b0a326f2d70ade2671c37d1beef62249db97 (diff) |
Merge with 2.3.99-pre6.
Diffstat (limited to 'arch/ppc')
-rw-r--r-- | arch/ppc/8xx_io/uart.c | 2 | ||||
-rw-r--r-- | arch/ppc/defconfig | 4 | ||||
-rw-r--r-- | arch/ppc/kernel/apus_setup.c | 12 | ||||
-rw-r--r-- | arch/ppc/kernel/chrp_setup.c | 4 | ||||
-rw-r--r-- | arch/ppc/kernel/gemini_prom.S | 7 | ||||
-rw-r--r-- | arch/ppc/kernel/gemini_setup.c | 6 | ||||
-rw-r--r-- | arch/ppc/kernel/hashtable.S | 32 | ||||
-rw-r--r-- | arch/ppc/kernel/head.S | 30 | ||||
-rw-r--r-- | arch/ppc/kernel/head_8xx.S | 14 | ||||
-rw-r--r-- | arch/ppc/kernel/idle.c | 8 | ||||
-rw-r--r-- | arch/ppc/kernel/irq.c | 14 | ||||
-rw-r--r-- | arch/ppc/kernel/misc.S | 4 | ||||
-rw-r--r-- | arch/ppc/kernel/open_pic.c | 4 | ||||
-rw-r--r-- | arch/ppc/kernel/pmac_pic.c | 4 | ||||
-rw-r--r-- | arch/ppc/kernel/ppc_ksyms.c | 10 | ||||
-rw-r--r-- | arch/ppc/kernel/prep_setup.c | 6 | ||||
-rw-r--r-- | arch/ppc/kernel/process.c | 30 | ||||
-rw-r--r-- | arch/ppc/kernel/ptrace.c | 18 | ||||
-rw-r--r-- | arch/ppc/kernel/setup.c | 6 | ||||
-rw-r--r-- | arch/ppc/kernel/time.c | 8 | ||||
-rw-r--r-- | arch/ppc/mm/fault.c | 2 | ||||
-rw-r--r-- | arch/ppc/mm/init.c | 22 |
22 files changed, 118 insertions, 129 deletions
diff --git a/arch/ppc/8xx_io/uart.c b/arch/ppc/8xx_io/uart.c index 575f68a28..f5088b9e4 100644 --- a/arch/ppc/8xx_io/uart.c +++ b/arch/ppc/8xx_io/uart.c @@ -2063,7 +2063,7 @@ int rs_8xx_read_proc(char *page, char **start, off_t off, int count, done: if (off >= len+begin) return 0; - *start = page + (begin-off); + *start = page + (off-begin); return ((count < begin+len-off) ? count : begin+len-off); } diff --git a/arch/ppc/defconfig b/arch/ppc/defconfig index 8717bc055..229768b9b 100644 --- a/arch/ppc/defconfig +++ b/arch/ppc/defconfig @@ -184,8 +184,8 @@ CONFIG_BLK_DEV_IDEPCI=y CONFIG_IDEDMA_PCI_EXPERIMENTAL=y # CONFIG_IDEDMA_PCI_WIP is not set # CONFIG_IDEDMA_NEW_DRIVE_LISTINGS is not set -# CONFIG_BLK_DEV_AEC6210 is not set -# CONFIG_AEC6210_TUNING is not set +# CONFIG_BLK_DEV_AEC62XX is not set +# CONFIG_AEC62XX_TUNING is not set # CONFIG_BLK_DEV_ALI15X3 is not set # CONFIG_WDC_ALI15X3 is not set # CONFIG_BLK_DEV_AMD7409 is not set diff --git a/arch/ppc/kernel/apus_setup.c b/arch/ppc/kernel/apus_setup.c index a54efc6fd..b93bd8193 100644 --- a/arch/ppc/kernel/apus_setup.c +++ b/arch/ppc/kernel/apus_setup.c @@ -108,18 +108,6 @@ void (*mach_floppy_eject) (void) __apusdata = NULL; #ifdef CONFIG_HEARTBEAT void (*mach_heartbeat) (int) __apusdata = NULL; extern void apus_heartbeat (void); -static int heartbeat_enabled = 1; - -void enable_heartbeat(void) -{ - heartbeat_enabled = 1; -} - -void disable_heartbeat(void) -{ - heartbeat_enabled = 0; - mach_heartbeat(0); -} #endif extern unsigned long amiga_model; diff --git a/arch/ppc/kernel/chrp_setup.c b/arch/ppc/kernel/chrp_setup.c index 21abfc84f..3b541c26f 100644 --- a/arch/ppc/kernel/chrp_setup.c +++ b/arch/ppc/kernel/chrp_setup.c @@ -422,10 +422,10 @@ void __init chrp_init_IRQ(void) request_irq(openpic_to_irq(HYDRA_INT_ADB_NMI), xmon_irq, 0, "NMI", 0); #endif /* CONFIG_XMON */ -#ifdef __SMP__ +#ifdef CONFIG_SMP request_irq(openpic_to_irq(OPENPIC_VEC_IPI), openpic_ipi_action, 0, "IPI0", 0); -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ } void __init diff --git a/arch/ppc/kernel/gemini_prom.S b/arch/ppc/kernel/gemini_prom.S index f91e992f4..0904bb0eb 100644 --- a/arch/ppc/kernel/gemini_prom.S +++ b/arch/ppc/kernel/gemini_prom.S @@ -10,6 +10,7 @@ #include "ppc_asm.tmpl" #include "ppc_defs.h" +#include <linux/config.h> #include <asm/processor.h> #include <asm/page.h> #include <asm/gemini.h> @@ -25,7 +26,7 @@ _GLOBAL(prom_init) _GLOBAL(gemini_prom_init) -#ifdef __SMP__ +#ifdef CONFIG_SMP /* Since the MMU's on, get stuff in rom space that we'll need */ lis r4,GEMINI_CPUSTAT@h ori r4,r4,GEMINI_CPUSTAT@l @@ -74,14 +75,14 @@ prom_no_mmu: addi r3,r3,1 bdnz 3b -#ifdef __SMP__ +#ifdef CONFIG_SMP /* The 750 book (and Mot/IBM support) says that this will "assist" snooping when in SMP. Not sure yet whether this should stay or leave... */ mfspr r4,HID0 ori r4,r4,HID0_ABE mtspr HID0,r4 sync -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ blr /* apparently, SMon doesn't pay attention to HID0[SRST]. Disable the MMU and diff --git a/arch/ppc/kernel/gemini_setup.c b/arch/ppc/kernel/gemini_setup.c index fcf3a701c..ae9a0bd5c 100644 --- a/arch/ppc/kernel/gemini_setup.c +++ b/arch/ppc/kernel/gemini_setup.c @@ -303,7 +303,7 @@ printk("3\n"); /* standard stuff */ cache |= ((1<<reg)<<25); -#ifdef __SMP__ +#ifdef CONFIG_SMP /* A couple errata for the 750's (both IBM and Motorola silicon) note that you can get missed cache lines on MP implementations. The workaround - if you call it that - is to make the L2 @@ -351,7 +351,7 @@ void __init gemini_init_IRQ(void) for( i=0; i < NR_IRQS; i++ ) irq_desc[i].handler = &open_pic; openpic_init(1); -#ifdef __SMP__ +#ifdef CONFIG_SMP request_irq(OPENPIC_VEC_IPI, openpic_ipi_action, 0, "IPI0", 0); request_irq(OPENPIC_VEC_IPI+1, openpic_ipi_action, @@ -360,7 +360,7 @@ void __init gemini_init_IRQ(void) 0, "IPI2 (stop CPU)", 0); request_irq(OPENPIC_VEC_IPI+3, openpic_ipi_action, 0, "IPI3 (reschedule)", 0); -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ } #define gemini_rtc_read(x) (readb(GEMINI_RTC+(x))) diff --git a/arch/ppc/kernel/hashtable.S b/arch/ppc/kernel/hashtable.S index 5593ebe18..58045be1d 100644 --- a/arch/ppc/kernel/hashtable.S +++ b/arch/ppc/kernel/hashtable.S @@ -52,7 +52,7 @@ Hash_msk = (((1 << Hash_bits) - 1) * 64) .globl hash_page hash_page: -#ifdef __SMP__ +#ifdef CONFIG_SMP SAVE_2GPRS(7,r21) eieio lis r2,hash_table_lock@h @@ -88,7 +88,7 @@ hash_page: rlwimi r5,r3,12,20,29 /* insert top 10 bits of address */ lwz r5,0(r5) /* get pmd entry */ rlwinm. r5,r5,0,0,19 /* extract address of pte page */ -#ifdef __SMP__ +#ifdef CONFIG_SMP beq- hash_page_out /* return if no mapping */ #else /* XXX it seems like the 601 will give a machine fault on the @@ -102,7 +102,7 @@ hash_page: lwz r6,0(r2) /* get linux-style pte */ ori r4,r4,1 /* set _PAGE_PRESENT bit in access */ andc. r0,r4,r6 /* check access & ~permission */ -#ifdef __SMP__ +#ifdef CONFIG_SMP bne- hash_page_out /* return if access not permitted */ #else bnelr- @@ -128,7 +128,7 @@ hash_page: rlwinm r5,r5,7,1,24 /* put VSID in 0x7fffff80 bits */ #endif /* CONFIG_PPC64 */ -#ifndef __SMP__ /* do this later for SMP */ +#ifndef CONFIG_SMP /* do this later for SMP */ #ifdef CONFIG_PPC64 ori r5,r5,1 /* set V (valid) bit */ #else /* CONFIG_PPC64 */ @@ -280,7 +280,7 @@ hash_page_patch_C: addi r4,r4,1 stw r4,htab_evicts@l(r2) -#ifndef __SMP__ +#ifndef CONFIG_SMP /* Store PTE in PTEG */ found_empty: #ifdef CONFIG_PPC64 @@ -296,7 +296,7 @@ found_slot: #endif sync -#else /* __SMP__ */ +#else /* CONFIG_SMP */ /* * Between the tlbie above and updating the hash table entry below, * another CPU could read the hash table entry and put it in its TLB. @@ -324,7 +324,7 @@ found_slot: sync oris r5,r5,0x8000 stw r5,0(r3) /* finally set V bit in PTE */ -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ /* * Update the hash table miss count. We only want misses here @@ -342,7 +342,7 @@ found_slot: addi r3,r3,1 stw r3,htab_reloads@l(r2) -#ifdef __SMP__ +#ifdef CONFIG_SMP lis r2,hash_table_lock@ha tophys(r2,r2) li r0,0 @@ -373,7 +373,7 @@ found_slot: lwz r21,GPR21(r21) rfi -#ifdef __SMP__ +#ifdef CONFIG_SMP hash_page_out: lis r2,hash_table_lock@ha tophys(r2,r2) @@ -386,7 +386,7 @@ hash_page_out: .globl hash_table_lock hash_table_lock: .long 0 -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ .data next_slot: @@ -404,13 +404,13 @@ _GLOBAL(flush_hash_segments) bne+ 99f tlbia sync -#ifdef __SMP__ +#ifdef CONFIG_SMP tlbsync sync #endif blr 99: -#ifdef __SMP__ +#ifdef CONFIG_SMP /* Note - we had better not do anything which could generate a hash table miss while we have the hash table locked, or we'll get a deadlock. -paulus */ @@ -451,7 +451,7 @@ _GLOBAL(flush_hash_segments) sync tlbia sync -#ifdef __SMP__ +#ifdef CONFIG_SMP tlbsync sync lis r3,hash_table_lock@ha @@ -473,13 +473,13 @@ _GLOBAL(flush_hash_page) bne+ 99f tlbie r4 /* in hw tlb too */ sync -#ifdef __SMP__ +#ifdef CONFIG_SMP tlbsync sync #endif blr 99: -#ifdef __SMP__ +#ifdef CONFIG_SMP /* Note - we had better not do anything which could generate a hash table miss while we have the hash table locked, or we'll get a deadlock. -paulus */ @@ -531,7 +531,7 @@ _GLOBAL(flush_hash_page) 4: sync tlbie r4 /* in hw tlb too */ sync -#ifdef __SMP__ +#ifdef CONFIG_SMP tlbsync sync li r0,0 diff --git a/arch/ppc/kernel/head.S b/arch/ppc/kernel/head.S index dab413c15..218bb4ebb 100644 --- a/arch/ppc/kernel/head.S +++ b/arch/ppc/kernel/head.S @@ -189,11 +189,11 @@ __after_prom_start: #endif /* CONFIG_PPC64 */ 4: tophys(r8,r11) -#ifdef __SMP__ +#ifdef CONFIG_SMP ori r8,r8,0x12 /* R/W access, M=1 */ #else ori r8,r8,2 /* R/W access */ -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ #ifdef CONFIG_APUS ori r11,r11,BL_8M<<2|0x2 /* set up 8MB BAT registers for 604 */ #else @@ -782,7 +782,7 @@ load_up_fpu: * horrendously complex, especially when a task switches from one CPU * to another. Instead we call giveup_fpu in switch_to. */ -#ifndef __SMP__ +#ifndef CONFIG_SMP lis r6,0 /* get __pa constant */ tophys(r6,r6) addis r3,r6,last_task_used_math@ha @@ -801,18 +801,18 @@ load_up_fpu: andc r4,r4,r20 /* disable FP for previous task */ stw r4,_MSR-STACK_FRAME_OVERHEAD(r5) 1: -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ /* enable use of FP after return */ ori r23,r23,MSR_FP|MSR_FE0|MSR_FE1 mfspr r5,SPRG3 /* current task's THREAD (phys) */ lfd fr0,THREAD_FPSCR-4(r5) mtfsf 0xff,fr0 REST_32FPRS(0, r5) -#ifndef __SMP__ +#ifndef CONFIG_SMP subi r4,r5,THREAD sub r4,r4,r6 stw r4,last_task_used_math@l(r3) -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ /* restore registers and return */ lwz r3,_CCR(r21) lwz r4,_LINK(r21) @@ -868,7 +868,7 @@ load_up_altivec: * horrendously complex, especially when a task switches from one CPU * to another. Instead we call giveup_altivec in switch_to. */ -#ifndef __SMP__ +#ifndef CONFIG_SMP #ifndef CONFIG_APUS lis r6,-KERNELBASE@h #else @@ -892,7 +892,7 @@ load_up_altivec: andc r4,r4,r20 /* disable altivec for previous task */ stw r4,_MSR-STACK_FRAME_OVERHEAD(r5) 1: -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ /* enable use of AltiVec after return */ oris r23,r23,MSR_VEC@h mfspr r5,SPRG3 /* current task's THREAD (phys) */ @@ -900,11 +900,11 @@ load_up_altivec: LVX(vr0,r20,r5) MTVSCR(vr0) REST_32VR(0,r20,r5) -#ifndef __SMP__ +#ifndef CONFIG_SMP subi r4,r5,THREAD sub r4,r4,r6 stw r4,last_task_used_altivec@l(r3) -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ /* restore registers and return */ lwz r3,_CCR(r21) lwz r4,_LINK(r21) @@ -967,11 +967,11 @@ giveup_altivec: andc r4,r4,r3 /* disable AltiVec for previous task */ stw r4,_MSR-STACK_FRAME_OVERHEAD(r5) 1: -#ifndef __SMP__ +#ifndef CONFIG_SMP li r5,0 lis r4,last_task_used_altivec@ha stw r5,last_task_used_altivec@l(r4) -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ blr #endif /* CONFIG_ALTIVEC */ @@ -1002,11 +1002,11 @@ giveup_fpu: andc r4,r4,r3 /* disable FP for previous task */ stw r4,_MSR-STACK_FRAME_OVERHEAD(r5) 1: -#ifndef __SMP__ +#ifndef CONFIG_SMP li r5,0 lis r4,last_task_used_math@ha stw r5,last_task_used_math@l(r4) -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ blr /* @@ -1432,7 +1432,7 @@ start_here: SYNC /* Force all PTE updates to finish */ tlbia /* Clear all TLB entries */ sync /* wait for tlbia/tlbie to finish */ -#ifdef __SMP__ +#ifdef CONFIG_SMP tlbsync /* ... on all CPUs */ sync #endif diff --git a/arch/ppc/kernel/head_8xx.S b/arch/ppc/kernel/head_8xx.S index 94cd67d3b..83f276585 100644 --- a/arch/ppc/kernel/head_8xx.S +++ b/arch/ppc/kernel/head_8xx.S @@ -698,7 +698,7 @@ __secondary_hold: * This is where the main kernel code starts. */ start_here: -#ifdef __SMP__ +#ifdef CONFIG_SMP /* if we're the second cpu stack and r2 are different * and we want to not clear the bss -- Cort */ lis r5,first_cpu_booted@h @@ -715,7 +715,7 @@ start_here: lwz r2,0(r2) b 10f 99: -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ /* ptr to current */ lis r2,init_task_union@h ori r2,r2,init_task_union@l @@ -734,9 +734,9 @@ start_here: 3: stwu r0,4(r8) bdnz 3b 2: -#ifdef __SMP__ +#ifdef CONFIG_SMP 10: -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ /* stack */ addi r1,r2,TASK_UNION_SIZE li r0,0 @@ -779,7 +779,7 @@ start_here: SYNC /* Force all PTE updates to finish */ tlbia /* Clear all TLB entries */ sync /* wait for tlbia/tlbie to finish */ -#ifdef __SMP__ +#ifdef CONFIG_SMP tlbsync /* ... on all CPUs */ sync #endif @@ -794,7 +794,7 @@ start_here: li r4,MSR_KERNEL lis r3,start_kernel@h ori r3,r3,start_kernel@l -#ifdef __SMP__ +#ifdef CONFIG_SMP /* the second time through here we go to * start_secondary(). -- Cort */ @@ -807,7 +807,7 @@ start_here: lis r3,start_secondary@h ori r3,r3,start_secondary@l 10: -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ mtspr SRR0,r3 mtspr SRR1,r4 rfi /* enable MMU and jump to start_kernel */ diff --git a/arch/ppc/kernel/idle.c b/arch/ppc/kernel/idle.c index 7cf97b873..1d3dfd299 100644 --- a/arch/ppc/kernel/idle.c +++ b/arch/ppc/kernel/idle.c @@ -61,7 +61,7 @@ int idled(void) if ( !current->need_resched && htab_reclaim_on ) htab_reclaim(); if ( !current->need_resched ) power_save(); -#ifdef __SMP__ +#ifdef CONFIG_SMP if (current->need_resched) #endif schedule(); @@ -162,11 +162,11 @@ unsigned long get_zero_page_fast(void) : "=&r" (tmp), "=&r" (page), "+m" (zero_cache) : "r" (&zero_quicklist) : "cc" ); -#ifdef __SMP__ +#ifdef CONFIG_SMP /* if another cpu beat us above this can happen -- Cort */ if ( page == 0 ) return 0; -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ /* we can update zerocount after the fact since it is not * used for anything but control of a loop which doesn't * matter since it won't affect anything if it zeros one @@ -253,7 +253,7 @@ void zero_paged(void) /* atomically add this page to the list */ asm ( "101:lwarx %0,0,%2\n" /* reserve zero_cache */ " stw %0,0(%3)\n" /* update *pageptr */ -#ifdef __SMP__ +#ifdef CONFIG_SMP " sync\n" /* let store settle */ #endif " stwcx. %3,0,%2\n" /* update zero_cache in mem */ diff --git a/arch/ppc/kernel/irq.c b/arch/ppc/kernel/irq.c index e4b279032..c0cc492ad 100644 --- a/arch/ppc/kernel/irq.c +++ b/arch/ppc/kernel/irq.c @@ -237,13 +237,13 @@ int get_irq_list(char *buf) if ( !action || !action->handler ) continue; len += sprintf(buf+len, "%3d: ", i); -#ifdef __SMP__ +#ifdef CONFIG_SMP for (j = 0; j < smp_num_cpus; j++) len += sprintf(buf+len, "%10u ", kstat.irqs[cpu_logical_map(j)][i]); #else len += sprintf(buf+len, "%10u ", kstat_irqs(i)); -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ if ( irq_desc[i].handler ) len += sprintf(buf+len, " %s ", irq_desc[i].handler->typename ); else @@ -254,7 +254,7 @@ int get_irq_list(char *buf) } len += sprintf(buf+len, "\n"); } -#ifdef __SMP__ +#ifdef CONFIG_SMP /* should this be per processor send/receive? */ len += sprintf(buf+len, "IPI: %10lu\n", ipi_count); #endif @@ -343,7 +343,7 @@ void __init init_IRQ(void) ppc_md.init_IRQ(); } -#ifdef __SMP__ +#ifdef CONFIG_SMP unsigned char global_irq_holder = NO_PROC_ID; unsigned volatile int global_irq_lock; atomic_t global_irq_count; @@ -611,7 +611,7 @@ void __global_restore_flags(unsigned long flags) } } } -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ static struct proc_dir_entry * root_irq_dir; static struct proc_dir_entry * irq_dir [NR_IRQS]; @@ -733,7 +733,7 @@ static void register_irq_proc (unsigned int irq) irq_dir[irq] = proc_mkdir(name, root_irq_dir); /* create /proc/irq/1234/smp_affinity */ - entry = create_proc_entry("smp_affinity", 0700, irq_dir[irq]); + entry = create_proc_entry("smp_affinity", 0600, irq_dir[irq]); entry->nlink = 1; entry->data = (void *)irq; @@ -754,7 +754,7 @@ void init_irq_proc (void) root_irq_dir = proc_mkdir("irq", 0); /* create /proc/irq/prof_cpu_mask */ - entry = create_proc_entry("prof_cpu_mask", 0700, root_irq_dir); + entry = create_proc_entry("prof_cpu_mask", 0600, root_irq_dir); entry->nlink = 1; entry->data = (void *)&prof_cpu_mask; diff --git a/arch/ppc/kernel/misc.S b/arch/ppc/kernel/misc.S index 8444bb4a0..e4fbefbc3 100644 --- a/arch/ppc/kernel/misc.S +++ b/arch/ppc/kernel/misc.S @@ -142,7 +142,7 @@ _GLOBAL(_tlbia) sync tlbia sync -#ifdef __SMP__ +#ifdef CONFIG_SMP tlbsync sync #endif @@ -154,7 +154,7 @@ _GLOBAL(_tlbia) _GLOBAL(_tlbie) tlbie r3 sync -#ifdef __SMP__ +#ifdef CONFIG_SMP tlbsync sync #endif diff --git a/arch/ppc/kernel/open_pic.c b/arch/ppc/kernel/open_pic.c index 9438c57ee..a3c6cc4dd 100644 --- a/arch/ppc/kernel/open_pic.c +++ b/arch/ppc/kernel/open_pic.c @@ -101,12 +101,12 @@ void no_action(int ir1, void *dev, struct pt_regs *regs) { } -#ifdef __SMP__ +#ifdef CONFIG_SMP void openpic_ipi_action(int cpl, void *dev_id, struct pt_regs *regs) { smp_message_recv(cpl-OPENPIC_VEC_IPI); } -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ #ifdef __i386__ static inline u_int in_le32(volatile u_int *addr) diff --git a/arch/ppc/kernel/pmac_pic.c b/arch/ppc/kernel/pmac_pic.c index f2794ae5a..b58d69fc9 100644 --- a/arch/ppc/kernel/pmac_pic.c +++ b/arch/ppc/kernel/pmac_pic.c @@ -183,7 +183,7 @@ pmac_get_irq(struct pt_regs *regs) int irq; unsigned long bits = 0; -#ifdef __SMP__ +#ifdef CONFIG_SMP void pmac_smp_message_recv(void); /* IPI's are a hack on the powersurge -- Cort */ @@ -197,7 +197,7 @@ pmac_get_irq(struct pt_regs *regs) pmac_smp_message_recv(); return -2; /* ignore, already handled */ } -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ /* Yeah, I know, this could be a separate get_irq function */ if (has_openpic) diff --git a/arch/ppc/kernel/ppc_ksyms.c b/arch/ppc/kernel/ppc_ksyms.c index 87c8d4082..7d4c038ec 100644 --- a/arch/ppc/kernel/ppc_ksyms.c +++ b/arch/ppc/kernel/ppc_ksyms.c @@ -35,9 +35,9 @@ #include <asm/dma.h> #include <asm/machdep.h> #include <asm/hw_irq.h> -#ifdef __SMP__ +#ifdef CONFIG_SMP #include <asm/smplock.h> -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ #include "time.h" /* Tell string.h we don't want memcpy etc. as cpp defines */ @@ -76,9 +76,9 @@ EXPORT_SYMBOL(disable_irq); EXPORT_SYMBOL(disable_irq_nosync); EXPORT_SYMBOL(local_irq_count); EXPORT_SYMBOL(local_bh_count); -#ifdef __SMP__ +#ifdef CONFIG_SMP EXPORT_SYMBOL(kernel_flag); -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ #if !defined(CONFIG_4xx) && !defined(CONFIG_8xx) EXPORT_SYMBOL(isa_io_base); @@ -189,7 +189,7 @@ EXPORT_SYMBOL(giveup_fpu); EXPORT_SYMBOL(enable_kernel_fp); EXPORT_SYMBOL(flush_icache_range); EXPORT_SYMBOL(xchg_u32); -#ifdef __SMP__ +#ifdef CONFIG_SMP EXPORT_SYMBOL(__global_cli); EXPORT_SYMBOL(__global_sti); EXPORT_SYMBOL(__global_save_flags); diff --git a/arch/ppc/kernel/prep_setup.c b/arch/ppc/kernel/prep_setup.c index 19ae61450..cb1110eca 100644 --- a/arch/ppc/kernel/prep_setup.c +++ b/arch/ppc/kernel/prep_setup.c @@ -118,7 +118,7 @@ prep_get_cpuinfo(char *buffer) extern char *Motherboard_map_name; int len, i; -#ifdef __SMP__ +#ifdef CONFIG_SMP #define CD(X) (cpu_data[n].X) #else #define CD(X) (X) @@ -624,10 +624,10 @@ prep_init_IRQ(void) for ( i = 0 ; i < 16 ; i++ ) irq_desc[i].handler = &i8259_pic; i8259_init(); -#ifdef __SMP__ +#ifdef CONFIG_SMP request_irq(openpic_to_irq(OPENPIC_VEC_SPURIOUS), openpic_ipi_action, 0, "IPI0", 0); -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ } #if defined(CONFIG_BLK_DEV_IDE) || defined(CONFIG_BLK_DEV_IDE_MODULE) diff --git a/arch/ppc/kernel/process.c b/arch/ppc/kernel/process.c index 5c01d3c72..ce8e039c0 100644 --- a/arch/ppc/kernel/process.c +++ b/arch/ppc/kernel/process.c @@ -154,7 +154,7 @@ dump_altivec(struct pt_regs *regs, elf_vrregset_t *vrregs) void enable_kernel_altivec(void) { -#ifdef __SMP__ +#ifdef CONFIG_SMP if (current->thread.regs && (current->thread.regs->msr & MSR_VEC)) giveup_altivec(current); else @@ -169,14 +169,14 @@ enable_kernel_altivec(void) void enable_kernel_fp(void) { -#ifdef __SMP__ +#ifdef CONFIG_SMP if (current->thread.regs && (current->thread.regs->msr & MSR_FP)) giveup_fpu(current); else giveup_fpu(NULL); /* just enables FP for kernel */ #else giveup_fpu(last_task_used_math); -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ } int @@ -208,7 +208,7 @@ _switch_to(struct task_struct *prev, struct task_struct *new, new->comm,new->pid,new->thread.regs->nip,new->processor, new->fs->root,prev->fs->root); #endif -#ifdef __SMP__ +#ifdef CONFIG_SMP /* avoid complexity of lazy save/restore of fpu * by just saving it every time we switch out if * this task used the fpu during the last quantum. @@ -236,7 +236,7 @@ _switch_to(struct task_struct *prev, struct task_struct *new, #endif /* CONFIG_ALTIVEC */ prev->last_processor = prev->processor; current_set[smp_processor_id()] = new; -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ /* Avoid the trap. On smp this this never happens since * we don't set last_task_used_altivec -- Cort */ @@ -265,9 +265,9 @@ void show_regs(struct pt_regs * regs) printk("\nlast math %p last altivec %p", last_task_used_math, last_task_used_altivec); -#ifdef __SMP__ +#ifdef CONFIG_SMP printk(" CPU: %d last CPU: %d", current->processor,current->last_processor); -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ printk("\n"); for (i = 0; i < 32; i++) @@ -319,7 +319,7 @@ copy_thread(int nr, unsigned long clone_flags, unsigned long usp, { unsigned long msr; struct pt_regs * childregs, *kregs; -#ifdef __SMP__ +#ifdef CONFIG_SMP extern void ret_from_smpfork(void); #else extern void ret_from_except(void); @@ -336,7 +336,7 @@ copy_thread(int nr, unsigned long clone_flags, unsigned long usp, p->thread.ksp = (unsigned long) childregs - STACK_FRAME_OVERHEAD; p->thread.ksp -= sizeof(struct pt_regs ) + STACK_FRAME_OVERHEAD; kregs = (struct pt_regs *)(p->thread.ksp + STACK_FRAME_OVERHEAD); -#ifdef __SMP__ +#ifdef CONFIG_SMP kregs->nip = (unsigned long)ret_from_smpfork; #else kregs->nip = (unsigned long)ret_from_except; @@ -378,9 +378,9 @@ copy_thread(int nr, unsigned long clone_flags, unsigned long usp, childregs->msr &= ~MSR_VEC; #endif /* CONFIG_ALTIVEC */ -#ifdef __SMP__ +#ifdef CONFIG_SMP p->last_processor = NO_PROC_ID; -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ return 0; } @@ -447,14 +447,14 @@ asmlinkage int sys_clone(int p1, int p2, int p3, int p4, int p5, int p6, int res; lock_kernel(); res = do_fork(clone_flags, regs->gpr[1], regs); -#ifdef __SMP__ +#ifdef CONFIG_SMP /* When we clone the idle task we keep the same pid but * the return value of 0 for both causes problems. * -- Cort */ if ((current->pid == 0) && (current == &init_task)) res = 1; -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ unlock_kernel(); return res; } @@ -466,14 +466,14 @@ asmlinkage int sys_fork(int p1, int p2, int p3, int p4, int p5, int p6, int res; res = do_fork(SIGCHLD, regs->gpr[1], regs); -#ifdef __SMP__ +#ifdef CONFIG_SMP /* When we clone the idle task we keep the same pid but * the return value of 0 for both causes problems. * -- Cort */ if ((current->pid == 0) && (current == &init_task)) res = 1; -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ return res; } diff --git a/arch/ppc/kernel/ptrace.c b/arch/ppc/kernel/ptrace.c index 3a4e3b797..fe0bfcea2 100644 --- a/arch/ppc/kernel/ptrace.c +++ b/arch/ppc/kernel/ptrace.c @@ -100,7 +100,7 @@ static unsigned long get_long(struct task_struct * tsk, repeat: pgdir = pgd_offset(vma->vm_mm, addr); if (pgd_none(*pgdir)) { - handle_mm_fault(tsk, vma, addr, 0); + handle_mm_fault(tsk->mm, vma, addr, 0); goto repeat; } if (pgd_bad(*pgdir)) { @@ -110,7 +110,7 @@ repeat: } pgmiddle = pmd_offset(pgdir,addr); if (pmd_none(*pgmiddle)) { - handle_mm_fault(tsk, vma, addr, 0); + handle_mm_fault(tsk->mm, vma, addr, 0); goto repeat; } if (pmd_bad(*pgmiddle)) { @@ -120,7 +120,7 @@ repeat: } pgtable = pte_offset(pgmiddle, addr); if (!pte_present(*pgtable)) { - handle_mm_fault(tsk, vma, addr, 0); + handle_mm_fault(tsk->mm, vma, addr, 0); goto repeat; } page = pte_page(*pgtable); @@ -151,7 +151,7 @@ static void put_long(struct task_struct * tsk, struct vm_area_struct * vma, repeat: pgdir = pgd_offset(vma->vm_mm, addr); if (!pgd_present(*pgdir)) { - handle_mm_fault(tsk, vma, addr, 1); + handle_mm_fault(tsk->mm, vma, addr, 1); goto repeat; } if (pgd_bad(*pgdir)) { @@ -161,7 +161,7 @@ repeat: } pgmiddle = pmd_offset(pgdir,addr); if (pmd_none(*pgmiddle)) { - handle_mm_fault(tsk, vma, addr, 1); + handle_mm_fault(tsk->mm, vma, addr, 1); goto repeat; } if (pmd_bad(*pgmiddle)) { @@ -171,12 +171,12 @@ repeat: } pgtable = pte_offset(pgmiddle, addr); if (!pte_present(*pgtable)) { - handle_mm_fault(tsk, vma, addr, 1); + handle_mm_fault(tsk->mm, vma, addr, 1); goto repeat; } page = pte_page(*pgtable); if (!pte_write(*pgtable)) { - handle_mm_fault(tsk, vma, addr, 1); + handle_mm_fault(tsk->mm, vma, addr, 1); goto repeat; } /* this is a hack for non-kernel-mapped video buffers and similar */ @@ -198,7 +198,7 @@ repeat: static int read_long(struct task_struct * tsk, unsigned long addr, unsigned long * result) { - struct vm_area_struct * vma = find_extend_vma(tsk, addr); + struct vm_area_struct * vma = find_extend_vma(tsk->mm, addr); if (!vma) return -EIO; @@ -240,7 +240,7 @@ static int read_long(struct task_struct * tsk, unsigned long addr, static int write_long(struct task_struct * tsk, unsigned long addr, unsigned long data) { - struct vm_area_struct * vma = find_extend_vma(tsk, addr); + struct vm_area_struct * vma = find_extend_vma(tsk->mm, addr); if (!vma) return -EIO; diff --git a/arch/ppc/kernel/setup.c b/arch/ppc/kernel/setup.c index 30595e694..2b771ef14 100644 --- a/arch/ppc/kernel/setup.c +++ b/arch/ppc/kernel/setup.c @@ -204,7 +204,7 @@ int get_cpuinfo(char *buffer) unsigned int pvr; unsigned short maj, min; -#ifdef __SMP__ +#ifdef CONFIG_SMP #define CPU_PRESENT(x) (cpu_callin_map[(x)]) #define GET_PVR ((long int)(cpu_data[i].pvr)) #define CD(x) (cpu_data[i].x) @@ -355,13 +355,13 @@ int get_cpuinfo(char *buffer) bogosum += CD(loops_per_sec); } -#ifdef __SMP__ +#ifdef CONFIG_SMP if ( i ) len += sprintf(buffer+len, "\n"); len += sprintf(buffer+len,"total bogomips\t: %lu.%02lu\n", (bogosum+2500)/500000, (bogosum+2500)/5000 % 100); -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ /* * Ooh's and aah's info about zero'd pages in idle task diff --git a/arch/ppc/kernel/time.c b/arch/ppc/kernel/time.c index b695da797..b821249d9 100644 --- a/arch/ppc/kernel/time.c +++ b/arch/ppc/kernel/time.c @@ -74,7 +74,7 @@ int timer_interrupt(struct pt_regs * regs) unsigned long cpu = smp_processor_id(); hardirq_enter(cpu); -#ifdef __SMP__ +#ifdef CONFIG_SMP { unsigned int loops = 100000000; while (test_bit(0, &global_irq_lock)) { @@ -93,7 +93,7 @@ int timer_interrupt(struct pt_regs * regs) } } } -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ dval = get_dec(); /* @@ -133,7 +133,7 @@ int timer_interrupt(struct pt_regs * regs) last_rtc_update = xtime.tv_sec; } } -#ifdef __SMP__ +#ifdef CONFIG_SMP smp_local_timer_interrupt(regs); #endif @@ -155,7 +155,7 @@ void do_gettimeofday(struct timeval *tv) cli(); *tv = xtime; /* XXX we don't seem to have the decrementers synced properly yet */ -#ifndef __SMP__ +#ifndef CONFIG_SMP tv->tv_usec += (decrementer_count - get_dec()) * count_period_num / count_period_den; if (tv->tv_usec >= 1000000) { diff --git a/arch/ppc/mm/fault.c b/arch/ppc/mm/fault.c index 5fee4463b..562ce0295 100644 --- a/arch/ppc/mm/fault.c +++ b/arch/ppc/mm/fault.c @@ -135,7 +135,7 @@ good_area: if (!(vma->vm_flags & (VM_READ | VM_EXEC))) goto bad_area; } - if (!handle_mm_fault(current, vma, address, is_write)) + if (!handle_mm_fault(mm, vma, address, is_write)) goto bad_area; up(&mm->mmap_sem); /* diff --git a/arch/ppc/mm/init.c b/arch/ppc/mm/init.c index eaa5924a2..c92abef90 100644 --- a/arch/ppc/mm/init.c +++ b/arch/ppc/mm/init.c @@ -87,7 +87,7 @@ unsigned long avail_start; extern int num_memory; extern struct mem_info memory[]; extern boot_infos_t *boot_infos; -#ifndef __SMP__ +#ifndef CONFIG_SMP struct pgtable_cache_struct quicklists; #endif @@ -278,9 +278,9 @@ void show_mem(void) show_buffers(); printk("%-8s %3s %8s %8s %8s %9s %8s", "Process", "Pid", "Ctx", "Ctx<<4", "Last Sys", "pc", "task"); -#ifdef __SMP__ +#ifdef CONFIG_SMP printk(" %3s", "CPU"); -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ printk("\n"); for_each_task(p) { @@ -294,7 +294,7 @@ void show_mem(void) (ulong)p); { int iscur = 0; -#ifdef __SMP__ +#ifdef CONFIG_SMP printk("%3d ", p->processor); if ( (p->processor != NO_PROC_ID) && (p == current_set[p->processor]) ) @@ -315,7 +315,7 @@ void show_mem(void) printk(","); printk("last math"); } -#endif /* __SMP__ */ +#endif /* CONFIG_SMP */ printk("\n"); } } @@ -495,7 +495,7 @@ local_flush_tlb_all(void) { __clear_user(Hash, Hash_size); _tlbia(); -#ifdef __SMP__ +#ifdef CONFIG_SMP smp_send_tlb_invalidate(0); #endif } @@ -511,7 +511,7 @@ local_flush_tlb_mm(struct mm_struct *mm) mm->context = NO_CONTEXT; if (mm == current->mm) activate_mm(mm, mm); -#ifdef __SMP__ +#ifdef CONFIG_SMP smp_send_tlb_invalidate(0); #endif } @@ -523,7 +523,7 @@ local_flush_tlb_page(struct vm_area_struct *vma, unsigned long vmaddr) flush_hash_page(vma->vm_mm->context, vmaddr); else flush_hash_page(0, vmaddr); -#ifdef __SMP__ +#ifdef CONFIG_SMP smp_send_tlb_invalidate(0); #endif } @@ -551,7 +551,7 @@ local_flush_tlb_range(struct mm_struct *mm, unsigned long start, unsigned long e { flush_hash_page(mm->context, start); } -#ifdef __SMP__ +#ifdef CONFIG_SMP smp_send_tlb_invalidate(0); #endif } @@ -575,7 +575,7 @@ mmu_context_overflow(void) } read_unlock(&tasklist_lock); flush_hash_segments(0x10, 0xffffff); -#ifdef __SMP__ +#ifdef CONFIG_SMP smp_send_tlb_invalidate(0); #endif atomic_set(&next_mmu_context, 0); @@ -663,7 +663,7 @@ void __init setbat(int index, unsigned long virt, unsigned long phys, } #define IO_PAGE (_PAGE_NO_CACHE | _PAGE_GUARDED | _PAGE_RW) -#ifdef __SMP__ +#ifdef CONFIG_SMP #define RAM_PAGE (_PAGE_RW|_PAGE_COHERENT) #else #define RAM_PAGE (_PAGE_RW) |