mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-24 05:04:00 +08:00
8d7718aa08
Change futex_atomic_op_inuser and futex_atomic_cmpxchg_inatomic prototypes to use u32 types for the futex as this is the data type the futex core code uses all over the place. Signed-off-by: Michel Lespinasse <walken@google.com> Cc: Darren Hart <darren@dvhart.com> Cc: Peter Zijlstra <peterz@infradead.org> Cc: Matt Turner <mattst88@gmail.com> Cc: Russell King <linux@arm.linux.org.uk> Cc: David Howells <dhowells@redhat.com> Cc: Tony Luck <tony.luck@intel.com> Cc: Michal Simek <monstr@monstr.eu> Cc: Ralf Baechle <ralf@linux-mips.org> Cc: "James E.J. Bottomley" <jejb@parisc-linux.org> Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org> Cc: Martin Schwidefsky <schwidefsky@de.ibm.com> Cc: Paul Mundt <lethal@linux-sh.org> Cc: "David S. Miller" <davem@davemloft.net> Cc: Chris Metcalf <cmetcalf@tilera.com> Cc: Linus Torvalds <torvalds@linux-foundation.org> LKML-Reference: <20110311025058.GD26122@google.com> Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
128 lines
3.2 KiB
C
128 lines
3.2 KiB
C
#ifndef _ASM_FUTEX_H
|
|
#define _ASM_FUTEX_H
|
|
|
|
#include <linux/futex.h>
|
|
#include <linux/uaccess.h>
|
|
#include <asm/errno.h>
|
|
#include <asm/system.h>
|
|
|
|
#define __futex_atomic_op1(insn, ret, oldval, uaddr, oparg) \
|
|
do { \
|
|
register unsigned long r8 __asm ("r8") = 0; \
|
|
__asm__ __volatile__( \
|
|
" mf;; \n" \
|
|
"[1:] " insn ";; \n" \
|
|
" .xdata4 \"__ex_table\", 1b-., 2f-. \n" \
|
|
"[2:]" \
|
|
: "+r" (r8), "=r" (oldval) \
|
|
: "r" (uaddr), "r" (oparg) \
|
|
: "memory"); \
|
|
ret = r8; \
|
|
} while (0)
|
|
|
|
#define __futex_atomic_op2(insn, ret, oldval, uaddr, oparg) \
|
|
do { \
|
|
register unsigned long r8 __asm ("r8") = 0; \
|
|
int val, newval; \
|
|
do { \
|
|
__asm__ __volatile__( \
|
|
" mf;; \n" \
|
|
"[1:] ld4 %3=[%4];; \n" \
|
|
" mov %2=%3 \n" \
|
|
insn ";; \n" \
|
|
" mov ar.ccv=%2;; \n" \
|
|
"[2:] cmpxchg4.acq %1=[%4],%3,ar.ccv;; \n" \
|
|
" .xdata4 \"__ex_table\", 1b-., 3f-.\n" \
|
|
" .xdata4 \"__ex_table\", 2b-., 3f-.\n" \
|
|
"[3:]" \
|
|
: "+r" (r8), "=r" (val), "=&r" (oldval), \
|
|
"=&r" (newval) \
|
|
: "r" (uaddr), "r" (oparg) \
|
|
: "memory"); \
|
|
if (unlikely (r8)) \
|
|
break; \
|
|
} while (unlikely (val != oldval)); \
|
|
ret = r8; \
|
|
} while (0)
|
|
|
|
static inline int
|
|
futex_atomic_op_inuser (int encoded_op, u32 __user *uaddr)
|
|
{
|
|
int op = (encoded_op >> 28) & 7;
|
|
int cmp = (encoded_op >> 24) & 15;
|
|
int oparg = (encoded_op << 8) >> 20;
|
|
int cmparg = (encoded_op << 20) >> 20;
|
|
int oldval = 0, ret;
|
|
if (encoded_op & (FUTEX_OP_OPARG_SHIFT << 28))
|
|
oparg = 1 << oparg;
|
|
|
|
if (! access_ok (VERIFY_WRITE, uaddr, sizeof(u32)))
|
|
return -EFAULT;
|
|
|
|
pagefault_disable();
|
|
|
|
switch (op) {
|
|
case FUTEX_OP_SET:
|
|
__futex_atomic_op1("xchg4 %1=[%2],%3", ret, oldval, uaddr,
|
|
oparg);
|
|
break;
|
|
case FUTEX_OP_ADD:
|
|
__futex_atomic_op2("add %3=%3,%5", ret, oldval, uaddr, oparg);
|
|
break;
|
|
case FUTEX_OP_OR:
|
|
__futex_atomic_op2("or %3=%3,%5", ret, oldval, uaddr, oparg);
|
|
break;
|
|
case FUTEX_OP_ANDN:
|
|
__futex_atomic_op2("and %3=%3,%5", ret, oldval, uaddr,
|
|
~oparg);
|
|
break;
|
|
case FUTEX_OP_XOR:
|
|
__futex_atomic_op2("xor %3=%3,%5", ret, oldval, uaddr, oparg);
|
|
break;
|
|
default:
|
|
ret = -ENOSYS;
|
|
}
|
|
|
|
pagefault_enable();
|
|
|
|
if (!ret) {
|
|
switch (cmp) {
|
|
case FUTEX_OP_CMP_EQ: ret = (oldval == cmparg); break;
|
|
case FUTEX_OP_CMP_NE: ret = (oldval != cmparg); break;
|
|
case FUTEX_OP_CMP_LT: ret = (oldval < cmparg); break;
|
|
case FUTEX_OP_CMP_GE: ret = (oldval >= cmparg); break;
|
|
case FUTEX_OP_CMP_LE: ret = (oldval <= cmparg); break;
|
|
case FUTEX_OP_CMP_GT: ret = (oldval > cmparg); break;
|
|
default: ret = -ENOSYS;
|
|
}
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static inline int
|
|
futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr,
|
|
u32 oldval, u32 newval)
|
|
{
|
|
if (!access_ok(VERIFY_WRITE, uaddr, sizeof(u32)))
|
|
return -EFAULT;
|
|
|
|
{
|
|
register unsigned long r8 __asm ("r8") = 0;
|
|
unsigned long prev;
|
|
__asm__ __volatile__(
|
|
" mf;; \n"
|
|
" mov ar.ccv=%3;; \n"
|
|
"[1:] cmpxchg4.acq %0=[%1],%2,ar.ccv \n"
|
|
" .xdata4 \"__ex_table\", 1b-., 2f-. \n"
|
|
"[2:]"
|
|
: "=r" (prev)
|
|
: "r" (uaddr), "r" (newval),
|
|
"rO" ((long) (unsigned) oldval)
|
|
: "memory");
|
|
*uval = prev;
|
|
return r8;
|
|
}
|
|
}
|
|
|
|
#endif /* _ASM_FUTEX_H */
|