| /* $Id: bitops.h,v 1.67 2001/11/19 18:36:34 davem Exp $ |
| * bitops.h: Bit string operations on the Sparc. |
| * |
| * Copyright 1995 David S. Miller (davem@caip.rutgers.edu) |
| * Copyright 1996 Eddie C. Dost (ecd@skynet.be) |
| * Copyright 2001 Anton Blanchard (anton@samba.org) |
| */ |
| |
| #ifndef _SPARC_BITOPS_H |
| #define _SPARC_BITOPS_H |
| |
| #include <linux/compiler.h> |
| #include <asm/byteorder.h> |
| |
| #ifdef __KERNEL__ |
| |
| /* |
| * Set bit 'nr' in 32-bit quantity at address 'addr' where bit '0' |
| * is in the highest of the four bytes and bit '31' is the high bit |
| * within the first byte. Sparc is BIG-Endian. Unless noted otherwise |
| * all bit-ops return 0 if bit was previously clear and != 0 otherwise. |
| */ |
| static inline int test_and_set_bit(unsigned long nr, volatile unsigned long *addr) |
| { |
| register unsigned long mask asm("g2"); |
| register unsigned long *ADDR asm("g1"); |
| register int tmp1 asm("g3"); |
| register int tmp2 asm("g4"); |
| register int tmp3 asm("g5"); |
| register int tmp4 asm("g7"); |
| |
| ADDR = ((unsigned long *) addr) + (nr >> 5); |
| mask = 1 << (nr & 31); |
| |
| __asm__ __volatile__( |
| "mov %%o7, %%g4\n\t" |
| "call ___set_bit\n\t" |
| " add %%o7, 8, %%o7\n" |
| : "=&r" (mask), "=r" (tmp1), "=r" (tmp2), "=r" (tmp3), "=r" (tmp4) |
| : "0" (mask), "r" (ADDR) |
| : "memory", "cc"); |
| |
| return mask != 0; |
| } |
| |
| static inline void set_bit(unsigned long nr, volatile unsigned long *addr) |
| { |
| register unsigned long mask asm("g2"); |
| register unsigned long *ADDR asm("g1"); |
| register int tmp1 asm("g3"); |
| register int tmp2 asm("g4"); |
| register int tmp3 asm("g5"); |
| register int tmp4 asm("g7"); |
| |
| ADDR = ((unsigned long *) addr) + (nr >> 5); |
| mask = 1 << (nr & 31); |
| |
| __asm__ __volatile__( |
| "mov %%o7, %%g4\n\t" |
| "call ___set_bit\n\t" |
| " add %%o7, 8, %%o7\n" |
| : "=&r" (mask), "=r" (tmp1), "=r" (tmp2), "=r" (tmp3), "=r" (tmp4) |
| : "0" (mask), "r" (ADDR) |
| : "memory", "cc"); |
| } |
| |
| static inline int test_and_clear_bit(unsigned long nr, volatile unsigned long *addr) |
| { |
| register unsigned long mask asm("g2"); |
| register unsigned long *ADDR asm("g1"); |
| register int tmp1 asm("g3"); |
| register int tmp2 asm("g4"); |
| register int tmp3 asm("g5"); |
| register int tmp4 asm("g7"); |
| |
| ADDR = ((unsigned long *) addr) + (nr >> 5); |
| mask = 1 << (nr & 31); |
| |
| __asm__ __volatile__( |
| "mov %%o7, %%g4\n\t" |
| "call ___clear_bit\n\t" |
| " add %%o7, 8, %%o7\n" |
| : "=&r" (mask), "=r" (tmp1), "=r" (tmp2), "=r" (tmp3), "=r" (tmp4) |
| : "0" (mask), "r" (ADDR) |
| : "memory", "cc"); |
| |
| return mask != 0; |
| } |
| |
| static inline void clear_bit(unsigned long nr, volatile unsigned long *addr) |
| { |
| register unsigned long mask asm("g2"); |
| register unsigned long *ADDR asm("g1"); |
| register int tmp1 asm("g3"); |
| register int tmp2 asm("g4"); |
| register int tmp3 asm("g5"); |
| register int tmp4 asm("g7"); |
| |
| ADDR = ((unsigned long *) addr) + (nr >> 5); |
| mask = 1 << (nr & 31); |
| |
| __asm__ __volatile__( |
| "mov %%o7, %%g4\n\t" |
| "call ___clear_bit\n\t" |
| " add %%o7, 8, %%o7\n" |
| : "=&r" (mask), "=r" (tmp1), "=r" (tmp2), "=r" (tmp3), "=r" (tmp4) |
| : "0" (mask), "r" (ADDR) |
| : "memory", "cc"); |
| } |
| |
| static inline int test_and_change_bit(unsigned long nr, volatile unsigned long *addr) |
| { |
| register unsigned long mask asm("g2"); |
| register unsigned long *ADDR asm("g1"); |
| register int tmp1 asm("g3"); |
| register int tmp2 asm("g4"); |
| register int tmp3 asm("g5"); |
| register int tmp4 asm("g7"); |
| |
| ADDR = ((unsigned long *) addr) + (nr >> 5); |
| mask = 1 << (nr & 31); |
| |
| __asm__ __volatile__( |
| "mov %%o7, %%g4\n\t" |
| "call ___change_bit\n\t" |
| " add %%o7, 8, %%o7\n" |
| : "=&r" (mask), "=r" (tmp1), "=r" (tmp2), "=r" (tmp3), "=r" (tmp4) |
| : "0" (mask), "r" (ADDR) |
| : "memory", "cc"); |
| |
| return mask != 0; |
| } |
| |
| static inline void change_bit(unsigned long nr, volatile unsigned long *addr) |
| { |
| register unsigned long mask asm("g2"); |
| register unsigned long *ADDR asm("g1"); |
| register int tmp1 asm("g3"); |
| register int tmp2 asm("g4"); |
| register int tmp3 asm("g5"); |
| register int tmp4 asm("g7"); |
| |
| ADDR = ((unsigned long *) addr) + (nr >> 5); |
| mask = 1 << (nr & 31); |
| |
| __asm__ __volatile__( |
| "mov %%o7, %%g4\n\t" |
| "call ___change_bit\n\t" |
| " add %%o7, 8, %%o7\n" |
| : "=&r" (mask), "=r" (tmp1), "=r" (tmp2), "=r" (tmp3), "=r" (tmp4) |
| : "0" (mask), "r" (ADDR) |
| : "memory", "cc"); |
| } |
| |
| #include <asm-generic/bitops/non-atomic.h> |
| |
| #define smp_mb__before_clear_bit() do { } while(0) |
| #define smp_mb__after_clear_bit() do { } while(0) |
| |
| #include <asm-generic/bitops/ffz.h> |
| #include <asm-generic/bitops/__ffs.h> |
| #include <asm-generic/bitops/sched.h> |
| #include <asm-generic/bitops/ffs.h> |
| #include <asm-generic/bitops/fls.h> |
| #include <asm-generic/bitops/fls64.h> |
| #include <asm-generic/bitops/hweight.h> |
| #include <asm-generic/bitops/find.h> |
| #include <asm-generic/bitops/ext2-non-atomic.h> |
| #include <asm-generic/bitops/ext2-atomic.h> |
| #include <asm-generic/bitops/minix.h> |
| |
| #endif /* __KERNEL__ */ |
| |
| #endif /* defined(_SPARC_BITOPS_H) */ |