/* ----------------------------------------------------------------------------- * * (c) The GHC Team, 2011 * * Inefficient but necessary atomic locks used for implementing atomic * operations on ARM architectures pre-ARMv6. * * -------------------------------------------------------------------------- */ #include "PosixSource.h" #include "Rts.h" #if defined(HAVE_SCHED_H) #include #endif #if defined(THREADED_RTS) #if arm_HOST_ARCH && defined(arm_HOST_ARCH_PRE_ARMv6) static volatile int atomic_spin = 0; static int arm_atomic_spin_trylock (void) { int result; asm volatile ( "swp %0, %1, [%2]\n" : "=&r,&r" (result) : "r,0" (1), "r,r" (&atomic_spin) : "memory"); if (result == 0) return 0; else return -1; } void arm_atomic_spin_lock() { while (arm_atomic_spin_trylock()) #if defined(HAVE_SCHED_H) sched_yield(); #else ; // inefficient on non-POSIX. #endif } void arm_atomic_spin_unlock() { atomic_spin = 0; } #endif /* arm_HOST_ARCH && defined(arm_HOST_ARCH_PRE_ARMv6) */ #endif /* defined(THREADED_RTS) */