diff options
author | Ralf Baechle <ralf@linux-mips.org> | 1997-04-29 21:13:14 +0000 |
---|---|---|
committer | <ralf@linux-mips.org> | 1997-04-29 21:13:14 +0000 |
commit | 19c9bba94152148523ba0f7ef7cffe3d45656b11 (patch) | |
tree | 40b1cb534496a7f1ca0f5c314a523c69f1fee464 /arch/sparc/lib/locks.S | |
parent | 7206675c40394c78a90e74812bbdbf8cf3cca1be (diff) |
Import of Linux/MIPS 2.1.36
Diffstat (limited to 'arch/sparc/lib/locks.S')
-rw-r--r-- | arch/sparc/lib/locks.S | 162 |
1 files changed, 162 insertions, 0 deletions
diff --git a/arch/sparc/lib/locks.S b/arch/sparc/lib/locks.S new file mode 100644 index 000000000..e91efdc09 --- /dev/null +++ b/arch/sparc/lib/locks.S @@ -0,0 +1,162 @@ +/* $Id: locks.S,v 1.12 1997/04/22 18:48:07 davem Exp $ + * locks.S: SMP low-level lock primitives on Sparc. + * + * Copyright (C) 1996 David S. Miller (davem@caip.rutgers.edu) + */ + +#include <asm/cprefix.h> +#include <asm/ptrace.h> +#include <asm/psr.h> +#include <asm/smp.h> +#include <asm/spinlock.h> + + .text + .align 4 + + /* This is called when the initial acquisition attempt of a spin + * lock fails. The calling convention is weird, return address + * is in %o7 as usual but we agree with the caller to only touch + * and use %g2 as a temporary. We are passed a ptr to the lock + * itself in %g1, %g4 must be restored into %o7 when we return, + * and the caller wants us to return to him at three instructions + * previous to the call instruction which got us here. See how + * this is used in asm/smp_lock.h and asm/spinlock.h if what I + * just said confuses you to no end. + */ + .globl ___spinlock_waitfor +___spinlock_waitfor: +1: orcc %g2, 0x0, %g0 + bne,a 1b + ldub [%g1], %g2 + ldstub [%g1], %g2 + jmpl %o7 - 12, %g0 + mov %g4, %o7 + +___lk_busy_spin: + orcc %g2, 0, %g0 + bne,a ___lk_busy_spin + ldub [%g1 + 0], %g2 + b 1f + ldstub [%g1 + 0], %g2 + + .globl ___lock_kernel +___lock_kernel: + addcc %g2, -1, %g2 + bcs,a 9f + st %g2, [%g6 + AOFF_task_lock_depth] + rd %psr, %g3 + or %g3, PSR_PIL, %g2 + wr %g2, 0x0, %psr + nop; nop; nop + ldstub [%g1 + 0], %g2 +1: orcc %g2, 0, %g0 + bne,a ___lk_busy_spin + ldub [%g1 + 0], %g2 + stb %g5, [%g1 + 1] + mov -1, %g2 + st %g2, [%g6 + AOFF_task_lock_depth] + wr %g3, 0x0, %psr + nop; nop; nop +9: jmpl %o7 + 0x8, %g0 + mov %g4, %o7 + + .globl ___lock_reacquire_kernel +___lock_reacquire_kernel: + rd %psr, %g3 + or %g3, PSR_PIL, %g7 + wr %g7, 0x0, %psr + nop; nop; nop + st %g2, [%g6 + AOFF_task_lock_depth] + ldstub [%g1 + 0], %g2 +1: orcc %g2, 0, %g0 + be 3f + ldub [%g1 + 0], %g2 +2: orcc %g2, 0, %g0 + bne,a 2b + ldub [%g1 + 0], %g2 + b 1b + ldstub [%g1 + 0], %g2 +3: stb %g5, [%g1 + 1] + wr %g3, 0x0, %psr + nop; nop; nop + jmpl %o7 + 0x8, %g0 + mov %g4, %o7 + + .globl ___unlock_kernel +___unlock_kernel: + addcc %g2, 1, %g2 + bne,a 1f + st %g2, [%g6 + AOFF_task_lock_depth] + rd %psr, %g3 + or %g3, PSR_PIL, %g2 + wr %g2, 0x0, %psr + nop; nop; nop + mov NO_PROC_ID, %g2 + stb %g2, [%g1 + 1] + stb %g0, [%g1 + 0] + st %g0, [%g6 + AOFF_task_lock_depth] + wr %g3, 0x0, %psr + nop; nop; nop; +1: jmpl %o7 + 0x8, %g0 + mov %g4, %o7 + + /* Read/writer locks, as usual this is overly clever to make it + * as fast as possible. + */ + + /* caches... */ +___rw_read_enter_spin_on_wlock: + orcc %g2, 0x0, %g0 + be,a ___rw_read_enter + ldstub [%g1 + 3], %g2 + b ___rw_read_enter_spin_on_wlock + ldub [%g1 + 3], %g2 +___rw_write_enter_spin_on_wlock: + orcc %g2, 0x0, %g0 + be,a ___rw_write_enter + ldstub [%g1 + 3], %g2 + b ___rw_write_enter_spin_on_wlock + ldub [%g1 + 3], %g2 + + .globl ___rw_read_enter +___rw_read_enter: + orcc %g2, 0x0, %g0 + bne,a ___rw_read_enter_spin_on_wlock + ldub [%g1 + 3], %g2 +1: + ldstub [%g1 + 2], %g7 + orcc %g7, 0x0, %g0 + bne 1b + ldsh [%g1], %g2 + add %g2, 1, %g2 + sth %g2, [%g1] + sth %g0, [%g1 + 2] + retl + mov %g4, %o7 + + /* We must be careful here to not blow away wlock. */ + .globl ___rw_read_exit +___rw_read_exit_spin: + ldstub [%g1 + 2], %g2 +___rw_read_exit: + orcc %g2, 0x0, %g0 + bne ___rw_read_exit_spin + ldsh [%g1], %g7 + sub %g7, 1, %g7 + sth %g7, [%g1] + stb %g0, [%g1 + 2] + retl + mov %g4, %o7 + + .globl ___rw_write_enter +___rw_write_enter: + orcc %g2, 0x0, %g0 + bne,a ___rw_write_enter_spin_on_wlock + ldub [%g1 + 3], %g2 + ld [%g1], %g2 +1: + andncc %g2, 0xff, %g0 + bne,a 1b + ld [%g1], %g2 + retl + mov %g4, %o7 |