2008-06-08 00:50:51 +04:00
|
|
|
/*
|
|
|
|
* Copyright (c) 2003 Fabrice Bellard
|
|
|
|
*
|
|
|
|
* This library is free software; you can redistribute it and/or
|
|
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
|
|
* License as published by the Free Software Foundation; either
|
|
|
|
* version 2 of the License, or (at your option) any later version.
|
|
|
|
*
|
|
|
|
* This library is distributed in the hope that it will be useful,
|
|
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
|
|
* Lesser General Public License for more details.
|
|
|
|
*
|
|
|
|
* You should have received a copy of the GNU Lesser General Public
|
2009-07-17 00:47:01 +04:00
|
|
|
* License along with this library; if not, see <http://www.gnu.org/licenses/>
|
2008-06-08 00:50:51 +04:00
|
|
|
*/
|
|
|
|
|
|
|
|
/* Locking primitives. Most of this code should be redundant -
|
|
|
|
system emulation doesn't need/use locking, NPTL userspace uses
|
|
|
|
pthread mutexes, and non-NPTL userspace isn't threadsafe anyway.
|
|
|
|
In either case a spinlock is probably the wrong kind of lock.
|
|
|
|
Spinlocks are only good if you know annother CPU has the lock and is
|
|
|
|
likely to release it soon. In environments where you have more threads
|
|
|
|
than physical CPUs (the extreme case being a single CPU host) a spinlock
|
|
|
|
simply wastes CPU until the OS decides to preempt it. */
|
2009-07-27 18:13:24 +04:00
|
|
|
#if defined(CONFIG_USE_NPTL)
|
2008-06-08 00:50:51 +04:00
|
|
|
|
|
|
|
#include <pthread.h>
|
|
|
|
#define spin_lock pthread_mutex_lock
|
|
|
|
#define spin_unlock pthread_mutex_unlock
|
2009-10-02 01:12:16 +04:00
|
|
|
#define spinlock_t pthread_mutex_t
|
2008-06-08 00:50:51 +04:00
|
|
|
#define SPIN_LOCK_UNLOCKED PTHREAD_MUTEX_INITIALIZER
|
|
|
|
|
|
|
|
#else
|
|
|
|
|
|
|
|
#if defined(__hppa__)
|
|
|
|
|
2009-10-02 01:12:16 +04:00
|
|
|
typedef int spinlock_t[4];
|
2008-06-08 00:50:51 +04:00
|
|
|
|
|
|
|
#define SPIN_LOCK_UNLOCKED { 1, 1, 1, 1 }
|
|
|
|
|
2009-10-02 01:12:16 +04:00
|
|
|
static inline void resetlock (spinlock_t *p)
|
2008-06-08 00:50:51 +04:00
|
|
|
{
|
|
|
|
(*p)[0] = (*p)[1] = (*p)[2] = (*p)[3] = 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
#else
|
|
|
|
|
2009-10-02 01:12:16 +04:00
|
|
|
typedef int spinlock_t;
|
2008-06-08 00:50:51 +04:00
|
|
|
|
|
|
|
#define SPIN_LOCK_UNLOCKED 0
|
|
|
|
|
2009-10-02 01:12:16 +04:00
|
|
|
static inline void resetlock (spinlock_t *p)
|
2008-06-08 00:50:51 +04:00
|
|
|
{
|
|
|
|
*p = SPIN_LOCK_UNLOCKED;
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif
|
|
|
|
|
2009-01-14 21:39:49 +03:00
|
|
|
#if defined(_ARCH_PPC)
|
2008-06-08 00:50:51 +04:00
|
|
|
static inline int testandset (int *p)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
__asm__ __volatile__ (
|
2008-11-18 04:42:17 +03:00
|
|
|
" lwarx %0,0,%1\n"
|
2008-06-08 00:50:51 +04:00
|
|
|
" xor. %0,%3,%0\n"
|
2008-11-18 04:42:17 +03:00
|
|
|
" bne $+12\n"
|
2008-06-08 00:50:51 +04:00
|
|
|
" stwcx. %2,0,%1\n"
|
2008-11-18 04:42:17 +03:00
|
|
|
" bne- $-16\n"
|
2008-06-08 00:50:51 +04:00
|
|
|
: "=&r" (ret)
|
|
|
|
: "r" (p), "r" (1), "r" (0)
|
|
|
|
: "cr0", "memory");
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
#elif defined(__i386__)
|
|
|
|
static inline int testandset (int *p)
|
|
|
|
{
|
|
|
|
long int readval = 0;
|
|
|
|
|
|
|
|
__asm__ __volatile__ ("lock; cmpxchgl %2, %0"
|
|
|
|
: "+m" (*p), "+a" (readval)
|
|
|
|
: "r" (1)
|
|
|
|
: "cc");
|
|
|
|
return readval;
|
|
|
|
}
|
|
|
|
#elif defined(__x86_64__)
|
|
|
|
static inline int testandset (int *p)
|
|
|
|
{
|
|
|
|
long int readval = 0;
|
|
|
|
|
|
|
|
__asm__ __volatile__ ("lock; cmpxchgl %2, %0"
|
|
|
|
: "+m" (*p), "+a" (readval)
|
|
|
|
: "r" (1)
|
|
|
|
: "cc");
|
|
|
|
return readval;
|
|
|
|
}
|
|
|
|
#elif defined(__s390__)
|
|
|
|
static inline int testandset (int *p)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
__asm__ __volatile__ ("0: cs %0,%1,0(%2)\n"
|
|
|
|
" jl 0b"
|
|
|
|
: "=&d" (ret)
|
|
|
|
: "r" (1), "a" (p), "0" (*p)
|
|
|
|
: "cc", "memory" );
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
#elif defined(__alpha__)
|
|
|
|
static inline int testandset (int *p)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
unsigned long one;
|
|
|
|
|
|
|
|
__asm__ __volatile__ ("0: mov 1,%2\n"
|
|
|
|
" ldl_l %0,%1\n"
|
|
|
|
" stl_c %2,%1\n"
|
|
|
|
" beq %2,1f\n"
|
|
|
|
".subsection 2\n"
|
|
|
|
"1: br 0b\n"
|
|
|
|
".previous"
|
|
|
|
: "=r" (ret), "=m" (*p), "=r" (one)
|
|
|
|
: "m" (*p));
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
#elif defined(__sparc__)
|
|
|
|
static inline int testandset (int *p)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
__asm__ __volatile__("ldstub [%1], %0"
|
|
|
|
: "=r" (ret)
|
|
|
|
: "r" (p)
|
|
|
|
: "memory");
|
|
|
|
|
|
|
|
return (ret ? 1 : 0);
|
|
|
|
}
|
|
|
|
#elif defined(__arm__)
|
|
|
|
static inline int testandset (int *spinlock)
|
|
|
|
{
|
|
|
|
register unsigned int ret;
|
|
|
|
__asm__ __volatile__("swp %0, %1, [%2]"
|
|
|
|
: "=r"(ret)
|
|
|
|
: "0"(1), "r"(spinlock));
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
#elif defined(__mc68000)
|
|
|
|
static inline int testandset (int *p)
|
|
|
|
{
|
|
|
|
char ret;
|
|
|
|
__asm__ __volatile__("tas %1; sne %0"
|
|
|
|
: "=r" (ret)
|
|
|
|
: "m" (p)
|
|
|
|
: "cc","memory");
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
#elif defined(__hppa__)
|
|
|
|
|
|
|
|
/* Because malloc only guarantees 8-byte alignment for malloc'd data,
|
|
|
|
and GCC only guarantees 8-byte alignment for stack locals, we can't
|
|
|
|
be assured of 16-byte alignment for atomic lock data even if we
|
|
|
|
specify "__attribute ((aligned(16)))" in the type declaration. So,
|
|
|
|
we use a struct containing an array of four ints for the atomic lock
|
|
|
|
type and dynamically select the 16-byte aligned int from the array
|
|
|
|
for the semaphore. */
|
|
|
|
#define __PA_LDCW_ALIGNMENT 16
|
|
|
|
static inline void *ldcw_align (void *p) {
|
|
|
|
unsigned long a = (unsigned long)p;
|
|
|
|
a = (a + __PA_LDCW_ALIGNMENT - 1) & ~(__PA_LDCW_ALIGNMENT - 1);
|
|
|
|
return (void *)a;
|
|
|
|
}
|
|
|
|
|
2009-10-02 01:12:16 +04:00
|
|
|
static inline int testandset (spinlock_t *p)
|
2008-06-08 00:50:51 +04:00
|
|
|
{
|
|
|
|
unsigned int ret;
|
|
|
|
p = ldcw_align(p);
|
|
|
|
__asm__ __volatile__("ldcw 0(%1),%0"
|
|
|
|
: "=r" (ret)
|
|
|
|
: "r" (p)
|
|
|
|
: "memory" );
|
|
|
|
return !ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
#elif defined(__ia64)
|
|
|
|
|
|
|
|
#include <ia64intrin.h>
|
|
|
|
|
|
|
|
static inline int testandset (int *p)
|
|
|
|
{
|
|
|
|
return __sync_lock_test_and_set (p, 1);
|
|
|
|
}
|
|
|
|
#elif defined(__mips__)
|
|
|
|
static inline int testandset (int *p)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
__asm__ __volatile__ (
|
|
|
|
" .set push \n"
|
|
|
|
" .set noat \n"
|
|
|
|
" .set mips2 \n"
|
|
|
|
"1: li $1, 1 \n"
|
|
|
|
" ll %0, %1 \n"
|
|
|
|
" sc $1, %1 \n"
|
|
|
|
" beqz $1, 1b \n"
|
|
|
|
" .set pop "
|
|
|
|
: "=r" (ret), "+R" (*p)
|
|
|
|
:
|
|
|
|
: "memory");
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
#else
|
|
|
|
#error unimplemented CPU support
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#if defined(CONFIG_USER_ONLY)
|
2009-10-02 01:12:16 +04:00
|
|
|
static inline void spin_lock(spinlock_t *lock)
|
2008-06-08 00:50:51 +04:00
|
|
|
{
|
|
|
|
while (testandset(lock));
|
|
|
|
}
|
|
|
|
|
2009-10-02 01:12:16 +04:00
|
|
|
static inline void spin_unlock(spinlock_t *lock)
|
2008-06-08 00:50:51 +04:00
|
|
|
{
|
|
|
|
resetlock(lock);
|
|
|
|
}
|
|
|
|
|
2009-10-02 01:12:16 +04:00
|
|
|
static inline int spin_trylock(spinlock_t *lock)
|
2008-06-08 00:50:51 +04:00
|
|
|
{
|
|
|
|
return !testandset(lock);
|
|
|
|
}
|
|
|
|
#else
|
2009-10-02 01:12:16 +04:00
|
|
|
static inline void spin_lock(spinlock_t *lock)
|
2008-06-08 00:50:51 +04:00
|
|
|
{
|
|
|
|
}
|
|
|
|
|
2009-10-02 01:12:16 +04:00
|
|
|
static inline void spin_unlock(spinlock_t *lock)
|
2008-06-08 00:50:51 +04:00
|
|
|
{
|
|
|
|
}
|
|
|
|
|
2009-10-02 01:12:16 +04:00
|
|
|
static inline int spin_trylock(spinlock_t *lock)
|
2008-06-08 00:50:51 +04:00
|
|
|
{
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#endif
|