#include #define FUTEX_WAIT 0 #define FUTEX_WAKE 1 /* * int __futex_wait(volatile void *ftx, int val, const struct timespec *timeout) */ .text .globl __futex_wait .type __futex_wait, @function .align 4 __futex_wait: pushl %ebx pushl %esi mov 12(%esp), %ebx /* ftx */ movl $FUTEX_WAIT, %ecx mov 16(%esp), %edx /* val */ mov 20(%esp), %esi /* timeout */ movl $__NR_futex, %eax int $0x80 popl %esi popl %ebx ret /* int __futex_wake(volatile void *ftx, int count) */ .text .globl __futex_wake .type __futex_wake, @function .align 4 __futex_wake: pushl %ebx mov 8(%esp), %ebx /* ftx */ movl $FUTEX_WAKE, %ecx mov 12(%esp), %edx /* count */ movl $__NR_futex, %eax int $0x80 popl %ebx ret /* int __atomic_cmpxchg(int old, int new, volatile int* addr) */ .text .globl __atomic_cmpxchg .type __atomic_cmpxchg, @function .align 4 __atomic_cmpxchg: mov 4(%esp), %eax /* old */ mov 8(%esp), %ecx /* new */ mov 12(%esp), %edx /* addr */ lock cmpxchg %ecx, (%edx) jnz 1f xor %eax, %eax jmp 2f 1: movl $1, %eax 2: ret /* 0 == success, 1 == failure */ /* int __atomic_swap(int new, volatile int* addr) */ .text .globl __atomic_swap .type __atomic_swap, @function .align 4 __atomic_swap: mov 4(%esp), %ecx /* new */ mov 8(%esp), %edx /* addr */ lock xchg %ecx, (%edx) mov %ecx, %eax ret /* * int __atomic_dec(volatile int* addr) * * My x86 asm is really rusty.. this is probably suboptimal */ .text .globl __atomic_dec .type __atomic_dec, @function .align 4 __atomic_dec: pushl %ebx pushl %esi movl 12(%esp), %ebx /* addr */ 1: movl (%ebx), %esi /* old = *addr */ movl %esi, %edx subl $1, %edx /* new = old - 1 */ pushl %ebx pushl %edx pushl %esi call __atomic_cmpxchg addl $12, %esp test %eax, %eax jnz 1b movl %esi, %eax /* return old */ popl %esi popl %ebx ret .text /* int __atomic_inc(volatile int* addr) */ .globl __atomic_inc .type __atomic_inc, @function .align 4 __atomic_inc: pushl %ebx pushl %esi movl 12(%esp), %ebx /* addr */ 1: movl (%ebx), %esi /* old = *addr */ movl %esi, %edx addl $1, %edx /* new = old + 1 */ pushl %ebx pushl %edx pushl %esi call __atomic_cmpxchg addl $12, %esp test %eax, %eax jnz 1b movl %esi, %eax /* return old */ popl %esi popl %ebx ret