diff options
author | Ralf Baechle <ralf@linux-mips.org> | 2000-03-27 23:54:12 +0000 |
---|---|---|
committer | Ralf Baechle <ralf@linux-mips.org> | 2000-03-27 23:54:12 +0000 |
commit | d3e71cb08747743fce908122bab08b479eb403a5 (patch) | |
tree | cbec6948fdbdee9af81cf3ecfb504070d2745d7b /include/asm-i386 | |
parent | fe7ff1706e323d0e5ed83972960a1ecc1ee538b3 (diff) |
Merge with Linux 2.3.99-pre3.
Diffstat (limited to 'include/asm-i386')
-rw-r--r-- | include/asm-i386/cache.h | 8 | ||||
-rw-r--r-- | include/asm-i386/smplock.h | 2 | ||||
-rw-r--r-- | include/asm-i386/string.h | 7 | ||||
-rw-r--r-- | include/asm-i386/system.h | 3 | ||||
-rw-r--r-- | include/asm-i386/uaccess.h | 2 |
5 files changed, 11 insertions, 11 deletions
diff --git a/include/asm-i386/cache.h b/include/asm-i386/cache.h index 7260b50e4..de4a7a87c 100644 --- a/include/asm-i386/cache.h +++ b/include/asm-i386/cache.h @@ -4,11 +4,9 @@ #ifndef __ARCH_I386_CACHE_H #define __ARCH_I386_CACHE_H +#include <linux/config.h> + /* bytes per L1 cache line */ -#if CPU==586 || CPU==686 -#define L1_CACHE_BYTES 32 -#else -#define L1_CACHE_BYTES 16 -#endif +#define L1_CACHE_BYTES CONFIG_X86_L1_CACHE_BYTES #endif diff --git a/include/asm-i386/smplock.h b/include/asm-i386/smplock.h index 2c4b30791..f6f29383b 100644 --- a/include/asm-i386/smplock.h +++ b/include/asm-i386/smplock.h @@ -10,6 +10,8 @@ extern spinlock_t kernel_flag; +#define kernel_locked() spin_is_locked(&kernel_flag) + /* * Release global kernel lock and global interrupt lock */ diff --git a/include/asm-i386/string.h b/include/asm-i386/string.h index 515ffa7d5..ef0dd75d8 100644 --- a/include/asm-i386/string.h +++ b/include/asm-i386/string.h @@ -2,6 +2,7 @@ #define _I386_STRING_H_ #ifdef __KERNEL__ +#include <linux/config.h> /* * On a 486 or Pentium, we are better off not using the * byte string operations. But on a 386 or a PPro the @@ -11,7 +12,7 @@ * Also, the byte strings actually work correctly. Forget * the i486 routines for now as they may be broken.. */ -#if FIXED_486_STRING && (CPU == 486 || CPU == 586) +#if FIXED_486_STRING && defined(CONFIG_X86_USE_STRING_486) #include <asm/string-486.h> #else @@ -284,8 +285,6 @@ __asm__ __volatile__( \ #define __HAVE_ARCH_MEMCPY -#include <linux/config.h> - #ifdef CONFIG_X86_USE_3DNOW /* All this just for in_interrupt() ... */ @@ -529,7 +528,7 @@ extern inline void * memscan(void * addr, int c, size_t size) return addr; } +#endif /* CONFIG_X86_USE_STRING_486 */ #endif /* __KERNEL__ */ #endif -#endif diff --git a/include/asm-i386/system.h b/include/asm-i386/system.h index 36f2c8855..75e04e938 100644 --- a/include/asm-i386/system.h +++ b/include/asm-i386/system.h @@ -1,6 +1,7 @@ #ifndef __ASM_SYSTEM_H #define __ASM_SYSTEM_H +#include <linux/config.h> #include <linux/kernel.h> #include <asm/segment.h> #include <linux/bitops.h> /* for LOCK_PREFIX */ @@ -215,7 +216,7 @@ static inline unsigned long __xchg(unsigned long x, volatile void * ptr, int siz * indicated by comparing RETURN with OLD. */ -#if CPU != 386 +#ifdef CONFIG_X86_CMPXCHG #define __HAVE_ARCH_CMPXCHG 1 static inline unsigned long __cmpxchg(volatile void *ptr, unsigned long old, diff --git a/include/asm-i386/uaccess.h b/include/asm-i386/uaccess.h index 7546159f6..2e3081a5c 100644 --- a/include/asm-i386/uaccess.h +++ b/include/asm-i386/uaccess.h @@ -56,7 +56,7 @@ extern int __verify_write(const void *, unsigned long); segment_eq(get_fs(),KERNEL_DS) || \ __verify_write((void *)(addr),(size)))) -#endif /* CPU */ +#endif extern inline int verify_area(int type, const void * addr, unsigned long size) { |