mirror of
https://git.musl-libc.org/git/musl
synced 2025-03-28 05:23:03 +03:00
mitigate performance regression in libc-internal locks on x86_64
commit 3c43c0761e1725fd5f89a9c028cbf43250abb913 fixed missing synchronization in the atomic store operation for i386 and x86_64, but opted to use mfence for the barrier on x86_64 where it's always available. however, in practice mfence is significantly slower than the barrier approach used on i386 (a nop-like lock orl operation). this commit changes x86_64 (and x32) to use the faster barrier.
This commit is contained in:
parent
c13f2af1fe
commit
5a9c8c05a5
@ -83,7 +83,7 @@ static inline void a_dec(volatile int *x)
|
||||
|
||||
static inline void a_store(volatile int *p, int x)
|
||||
{
|
||||
__asm__( "mov %1, %0 ; mfence" : "=m"(*p) : "r"(x) : "memory" );
|
||||
__asm__( "mov %1, %0 ; lock ; orl $0,(%%rsp)" : "=m"(*p) : "r"(x) : "memory" );
|
||||
}
|
||||
|
||||
static inline void a_spin()
|
||||
|
@ -83,7 +83,7 @@ static inline void a_dec(volatile int *x)
|
||||
|
||||
static inline void a_store(volatile int *p, int x)
|
||||
{
|
||||
__asm__( "mov %1, %0 ; mfence" : "=m"(*p) : "r"(x) : "memory" );
|
||||
__asm__( "mov %1, %0 ; lock ; orl $0,(%%rsp)" : "=m"(*p) : "r"(x) : "memory" );
|
||||
}
|
||||
|
||||
static inline void a_spin()
|
||||
|
Loading…
x
Reference in New Issue
Block a user