From: Jaroslav Kysela Date: Thu, 20 Feb 2003 19:24:01 +0000 (+0000) Subject: Renamed ATOMIC->ATOMIC_SMP_LOCK X-Git-Tag: v1.0.3~258 X-Git-Url: https://git.alsa-project.org/?a=commitdiff_plain;h=8d2eecb54720d1a44c01f565933b22674aa6cb1c;p=alsa-lib.git Renamed ATOMIC->ATOMIC_SMP_LOCK --- diff --git a/include/iatomic.h b/include/iatomic.h index b7dd1f69..c30b1e3f 100644 --- a/include/iatomic.h +++ b/include/iatomic.h @@ -8,7 +8,7 @@ * resource counting etc.. */ -#define LOCK "lock ; " +#define ATOMIC_SMP_LOCK "lock ; " /* * Make sure gcc doesn't try to be clever and move things around @@ -49,7 +49,7 @@ typedef struct { volatile int counter; } atomic_t; static __inline__ void atomic_add(int i, atomic_t *v) { __asm__ __volatile__( - LOCK "addl %1,%0" + ATOMIC_SMP_LOCK "addl %1,%0" :"=m" (v->counter) :"ir" (i), "m" (v->counter)); } @@ -65,7 +65,7 @@ static __inline__ void atomic_add(int i, atomic_t *v) static __inline__ void atomic_sub(int i, atomic_t *v) { __asm__ __volatile__( - LOCK "subl %1,%0" + ATOMIC_SMP_LOCK "subl %1,%0" :"=m" (v->counter) :"ir" (i), "m" (v->counter)); } @@ -85,7 +85,7 @@ static __inline__ int atomic_sub_and_test(int i, atomic_t *v) unsigned char c; __asm__ __volatile__( - LOCK "subl %2,%0; sete %1" + ATOMIC_SMP_LOCK "subl %2,%0; sete %1" :"=m" (v->counter), "=qm" (c) :"ir" (i), "m" (v->counter) : "memory"); return c; @@ -101,7 +101,7 @@ static __inline__ int atomic_sub_and_test(int i, atomic_t *v) static __inline__ void atomic_inc(atomic_t *v) { __asm__ __volatile__( - LOCK "incl %0" + ATOMIC_SMP_LOCK "incl %0" :"=m" (v->counter) :"m" (v->counter)); } @@ -116,7 +116,7 @@ static __inline__ void atomic_inc(atomic_t *v) static __inline__ void atomic_dec(atomic_t *v) { __asm__ __volatile__( - LOCK "decl %0" + ATOMIC_SMP_LOCK "decl %0" :"=m" (v->counter) :"m" (v->counter)); } @@ -135,7 +135,7 @@ static __inline__ int atomic_dec_and_test(atomic_t *v) unsigned char c; __asm__ __volatile__( - LOCK "decl %0; sete %1" + ATOMIC_SMP_LOCK "decl %0; sete %1" :"=m" (v->counter), "=qm" (c) :"m" (v->counter) : "memory"); return c != 0; @@ -155,7 +155,7 @@ static __inline__ int atomic_inc_and_test(atomic_t *v) unsigned char c; __asm__ __volatile__( - LOCK "incl %0; sete %1" + ATOMIC_SMP_LOCK "incl %0; sete %1" :"=m" (v->counter), "=qm" (c) :"m" (v->counter) : "memory"); return c != 0; @@ -176,7 +176,7 @@ static __inline__ int atomic_add_negative(int i, atomic_t *v) unsigned char c; __asm__ __volatile__( - LOCK "addl %2,%0; sets %1" + ATOMIC_SMP_LOCK "addl %2,%0; sets %1" :"=m" (v->counter), "=qm" (c) :"ir" (i), "m" (v->counter) : "memory"); return c; @@ -184,11 +184,11 @@ static __inline__ int atomic_add_negative(int i, atomic_t *v) /* These are x86-specific, used by some header files */ #define atomic_clear_mask(mask, addr) \ -__asm__ __volatile__(LOCK "andl %0,%1" \ +__asm__ __volatile__(ATOMIC_SMP_LOCK "andl %0,%1" \ : : "r" (~(mask)),"m" (*addr) : "memory") #define atomic_set_mask(mask, addr) \ -__asm__ __volatile__(LOCK "orl %0,%1" \ +__asm__ __volatile__(ATOMIC_SMP_LOCK "orl %0,%1" \ : : "r" (mask),"m" (*addr) : "memory") /* @@ -216,6 +216,8 @@ __asm__ __volatile__(LOCK "orl %0,%1" \ #define wmb() asm volatile("sfence":::"memory") #endif +#undef ATOMIC_SMP_LOCK + #define IATOMIC_DEFINED 1 #endif /* __i386__ */