forked from luck/tmp_suning_uos_patched
1da177e4c3
Initial git repository build. I'm not bothering with the full history, even though we have it. We can create a separate "historical" git archive of that later if we want to, and in the meantime it's about 3.2GB when imported into git - space that would just make the early git days unnecessarily complicated, when we don't have a lot of good infrastructure for it. Let it rip!
101 lines
2.6 KiB
ArmAsm
101 lines
2.6 KiB
ArmAsm
/* atomic.S: Move this stuff here for better ICACHE hit rates.
|
|
*
|
|
* Copyright (C) 1996 David S. Miller (davem@caipfs.rutgers.edu)
|
|
*/
|
|
|
|
#include <linux/config.h>
|
|
#include <asm/ptrace.h>
|
|
#include <asm/psr.h>
|
|
|
|
.text
|
|
.align 4
|
|
|
|
.globl __atomic_begin
|
|
__atomic_begin:
|
|
|
|
#ifndef CONFIG_SMP
|
|
.globl ___xchg32_sun4c
|
|
___xchg32_sun4c:
|
|
rd %psr, %g3
|
|
andcc %g3, PSR_PIL, %g0
|
|
bne 1f
|
|
nop
|
|
wr %g3, PSR_PIL, %psr
|
|
nop; nop; nop
|
|
1:
|
|
andcc %g3, PSR_PIL, %g0
|
|
ld [%g1], %g7
|
|
bne 1f
|
|
st %g2, [%g1]
|
|
wr %g3, 0x0, %psr
|
|
nop; nop; nop
|
|
1:
|
|
mov %g7, %g2
|
|
jmpl %o7 + 8, %g0
|
|
mov %g4, %o7
|
|
|
|
.globl ___xchg32_sun4md
|
|
___xchg32_sun4md:
|
|
swap [%g1], %g2
|
|
jmpl %o7 + 8, %g0
|
|
mov %g4, %o7
|
|
#endif
|
|
|
|
/* Read asm-sparc/atomic.h carefully to understand how this works for SMP.
|
|
* Really, some things here for SMP are overly clever, go read the header.
|
|
*/
|
|
.globl ___atomic24_add
|
|
___atomic24_add:
|
|
rd %psr, %g3 ! Keep the code small, old way was stupid
|
|
nop; nop; nop; ! Let the bits set
|
|
or %g3, PSR_PIL, %g7 ! Disable interrupts
|
|
wr %g7, 0x0, %psr ! Set %psr
|
|
nop; nop; nop; ! Let the bits set
|
|
#ifdef CONFIG_SMP
|
|
1: ldstub [%g1 + 3], %g7 ! Spin on the byte lock for SMP.
|
|
orcc %g7, 0x0, %g0 ! Did we get it?
|
|
bne 1b ! Nope...
|
|
ld [%g1], %g7 ! Load locked atomic24_t
|
|
sra %g7, 8, %g7 ! Get signed 24-bit integer
|
|
add %g7, %g2, %g2 ! Add in argument
|
|
sll %g2, 8, %g7 ! Transpose back to atomic24_t
|
|
st %g7, [%g1] ! Clever: This releases the lock as well.
|
|
#else
|
|
ld [%g1], %g7 ! Load locked atomic24_t
|
|
add %g7, %g2, %g2 ! Add in argument
|
|
st %g2, [%g1] ! Store it back
|
|
#endif
|
|
wr %g3, 0x0, %psr ! Restore original PSR_PIL
|
|
nop; nop; nop; ! Let the bits set
|
|
jmpl %o7, %g0 ! NOTE: not + 8, see callers in atomic.h
|
|
mov %g4, %o7 ! Restore %o7
|
|
|
|
.globl ___atomic24_sub
|
|
___atomic24_sub:
|
|
rd %psr, %g3 ! Keep the code small, old way was stupid
|
|
nop; nop; nop; ! Let the bits set
|
|
or %g3, PSR_PIL, %g7 ! Disable interrupts
|
|
wr %g7, 0x0, %psr ! Set %psr
|
|
nop; nop; nop; ! Let the bits set
|
|
#ifdef CONFIG_SMP
|
|
1: ldstub [%g1 + 3], %g7 ! Spin on the byte lock for SMP.
|
|
orcc %g7, 0x0, %g0 ! Did we get it?
|
|
bne 1b ! Nope...
|
|
ld [%g1], %g7 ! Load locked atomic24_t
|
|
sra %g7, 8, %g7 ! Get signed 24-bit integer
|
|
sub %g7, %g2, %g2 ! Subtract argument
|
|
sll %g2, 8, %g7 ! Transpose back to atomic24_t
|
|
st %g7, [%g1] ! Clever: This releases the lock as well
|
|
#else
|
|
ld [%g1], %g7 ! Load locked atomic24_t
|
|
sub %g7, %g2, %g2 ! Subtract argument
|
|
st %g2, [%g1] ! Store it back
|
|
#endif
|
|
wr %g3, 0x0, %psr ! Restore original PSR_PIL
|
|
nop; nop; nop; ! Let the bits set
|
|
jmpl %o7, %g0 ! NOTE: not + 8, see callers in atomic.h
|
|
mov %g4, %o7 ! Restore %o7
|
|
|
|
.globl __atomic_end
|
|
__atomic_end:
|