mirror of
git://git.musl-libc.org/musl
synced 2025-02-28 08:40:25 +00:00
improve aarch64 atomics
aarch64 provides ll/sc variants with acquire/release memory order, freeing us from the need to have full barriers both before and after the ll/sc operation. previously they were not used because the a_cas can fail without performing a_sc, in which case half of the barrier would be omitted. instead, define a custom version of a_cas for aarch64 which uses a_barrier explicitly when aborting the cas operation. aside from cas, other operations built on top of ll/sc are not affected since they never abort but rather loop until they succeed. a split ll/sc version of the pointer-sized a_cas_p is also introduced using the same technique. patch by Szabolcs Nagy.
This commit is contained in:
parent
0f5eb3de29
commit
b17fbd3520
@ -2,7 +2,7 @@
|
|||||||
static inline int a_ll(volatile int *p)
|
static inline int a_ll(volatile int *p)
|
||||||
{
|
{
|
||||||
int v;
|
int v;
|
||||||
__asm__ __volatile__ ("ldxr %0, %1" : "=r"(v) : "Q"(*p));
|
__asm__ __volatile__ ("ldaxr %0, %1" : "=r"(v) : "Q"(*p));
|
||||||
return v;
|
return v;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -10,7 +10,7 @@ static inline int a_ll(volatile int *p)
|
|||||||
static inline int a_sc(volatile int *p, int v)
|
static inline int a_sc(volatile int *p, int v)
|
||||||
{
|
{
|
||||||
int r;
|
int r;
|
||||||
__asm__ __volatile__ ("stxr %w0,%1,%2" : "=&r"(r) : "r"(v), "Q"(*p) : "memory");
|
__asm__ __volatile__ ("stlxr %w0,%1,%2" : "=&r"(r) : "r"(v), "Q"(*p) : "memory");
|
||||||
return !r;
|
return !r;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -20,25 +20,45 @@ static inline void a_barrier()
|
|||||||
__asm__ __volatile__ ("dmb ish" : : : "memory");
|
__asm__ __volatile__ ("dmb ish" : : : "memory");
|
||||||
}
|
}
|
||||||
|
|
||||||
#define a_pre_llsc a_barrier
|
#define a_cas a_cas
|
||||||
#define a_post_llsc a_barrier
|
static inline int a_cas(volatile int *p, int t, int s)
|
||||||
|
{
|
||||||
|
int old;
|
||||||
|
do {
|
||||||
|
old = a_ll(p);
|
||||||
|
if (old != t) {
|
||||||
|
a_barrier();
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
} while (!a_sc(p, s));
|
||||||
|
return old;
|
||||||
|
}
|
||||||
|
|
||||||
|
static inline void *a_ll_p(volatile void *p)
|
||||||
|
{
|
||||||
|
void *v;
|
||||||
|
__asm__ __volatile__ ("ldaxr %0, %1" : "=r"(v) : "Q"(*(void *volatile *)p));
|
||||||
|
return v;
|
||||||
|
}
|
||||||
|
|
||||||
|
static inline int a_sc_p(volatile int *p, void *v)
|
||||||
|
{
|
||||||
|
int r;
|
||||||
|
__asm__ __volatile__ ("stlxr %w0,%1,%2" : "=&r"(r) : "r"(v), "Q"(*(void *volatile *)p) : "memory");
|
||||||
|
return !r;
|
||||||
|
}
|
||||||
|
|
||||||
#define a_cas_p a_cas_p
|
#define a_cas_p a_cas_p
|
||||||
static inline void *a_cas_p(volatile void *p, void *t, void *s)
|
static inline void *a_cas_p(volatile void *p, void *t, void *s)
|
||||||
{
|
{
|
||||||
void *old;
|
void *old;
|
||||||
__asm__ __volatile__(
|
do {
|
||||||
" dmb ish\n"
|
old = a_ll_p(p);
|
||||||
"1: ldxr %0,%3\n"
|
if (old != t) {
|
||||||
" cmp %0,%1\n"
|
a_barrier();
|
||||||
" b.ne 1f\n"
|
break;
|
||||||
" stxr %w0,%2,%3\n"
|
}
|
||||||
" cbnz %w0,1b\n"
|
} while (!a_sc_p(p, s));
|
||||||
" mov %0,%1\n"
|
|
||||||
"1: dmb ish\n"
|
|
||||||
: "=&r"(old)
|
|
||||||
: "r"(t), "r"(s), "Q"(*(void *volatile *)p)
|
|
||||||
: "memory", "cc");
|
|
||||||
return old;
|
return old;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user