blob: 838e4a8d27f1a23f4e027dea2ece4398bd5030c5 [file] [log] [blame]
Adopted from
http://git.yoctoproject.org/cgit/cgit.cgi/poky/commit/?id=e3a384578f677c05d812d99c2c92aa13670bd06a
Upstream-Status: Pending
Remove the `q' suffix on x86-64 atomic instructions.
We don't need the `q' suffix on x86_64 atomic instructions for AO_t,
which is defined as "unsigned long". "unsigned long" is 32bit for x32
and 64bit for x86-64. The register operand in x86-64 atomic instructions
is sufficient to properly determine the register size.
Received this patch from H.J. Lu <hjl.tools@gmail.com>
Signed-Off-By: Nitin A Kamble <nitin.a.kamble@intel.com> 2011/12/02
diff --git a/libatomic_ops/src/atomic_ops/sysdeps/gcc/x86_64.h b/libatomic_ops/src/atomic_ops/sysdeps/gcc/x86_64.h
index 0f68c1e..3bcde88 100644
--- a/libatomic_ops/src/atomic_ops/sysdeps/gcc/x86_64.h
+++ b/libatomic_ops/src/atomic_ops/sysdeps/gcc/x86_64.h
@@ -47,7 +47,7 @@ AO_fetch_and_add_full (volatile AO_t *p, AO_t incr)
{
AO_t result;
- __asm__ __volatile__ ("lock; xaddq %0, %1" :
+ __asm__ __volatile__ ("lock; xadd %0, %1" :
"=r" (result), "=m" (*p) : "0" (incr), "m" (*p)
: "memory");
return result;
@@ -93,7 +93,7 @@ AO_int_fetch_and_add_full (volatile unsigned int *p, unsigned int incr)
AO_INLINE void
AO_or_full (volatile AO_t *p, AO_t incr)
{
- __asm__ __volatile__ ("lock; orq %1, %0" :
+ __asm__ __volatile__ ("lock; or %1, %0" :
"=m" (*p) : "r" (incr), "m" (*p) : "memory");
}
#define AO_HAVE_or_full
@@ -123,7 +123,7 @@ AO_compare_and_swap_full(volatile AO_t *addr, AO_t old, AO_t new_val)
return (int)__sync_bool_compare_and_swap(addr, old, new_val);
# else
char result;
- __asm__ __volatile__("lock; cmpxchgq %3, %0; setz %1"
+ __asm__ __volatile__("lock; cmpxchg %3, %0; setz %1"
: "=m" (*addr), "=a" (result)
: "m" (*addr), "r" (new_val), "a" (old) : "memory");
return (int) result;