LOCK_PREFIX         9 arch/x86/include/asm/alternative-asm.h 	.macro LOCK_PREFIX
LOCK_PREFIX        17 arch/x86/include/asm/alternative-asm.h 	.macro LOCK_PREFIX
LOCK_PREFIX        51 arch/x86/include/asm/atomic.h 	asm volatile(LOCK_PREFIX "addl %1,%0"
LOCK_PREFIX        65 arch/x86/include/asm/atomic.h 	asm volatile(LOCK_PREFIX "subl %1,%0"
LOCK_PREFIX        81 arch/x86/include/asm/atomic.h 	GEN_BINARY_RMWcc(LOCK_PREFIX "subl", v->counter, "er", i, "%0", "e");
LOCK_PREFIX        92 arch/x86/include/asm/atomic.h 	asm volatile(LOCK_PREFIX "incl %0"
LOCK_PREFIX       104 arch/x86/include/asm/atomic.h 	asm volatile(LOCK_PREFIX "decl %0"
LOCK_PREFIX       118 arch/x86/include/asm/atomic.h 	GEN_UNARY_RMWcc(LOCK_PREFIX "decl", v->counter, "%0", "e");
LOCK_PREFIX       131 arch/x86/include/asm/atomic.h 	GEN_UNARY_RMWcc(LOCK_PREFIX "incl", v->counter, "%0", "e");
LOCK_PREFIX       145 arch/x86/include/asm/atomic.h 	GEN_BINARY_RMWcc(LOCK_PREFIX "addl", v->counter, "er", i, "%0", "s");
LOCK_PREFIX       218 arch/x86/include/asm/atomic.h 	asm(LOCK_PREFIX "addw $1, %0" : "+m" (*v));
LOCK_PREFIX       224 arch/x86/include/asm/atomic.h 	asm volatile(LOCK_PREFIX "andl %0,%1"			\
LOCK_PREFIX       228 arch/x86/include/asm/atomic.h 	asm volatile(LOCK_PREFIX "orl %0,%1"			\
LOCK_PREFIX        45 arch/x86/include/asm/atomic64_64.h 	asm volatile(LOCK_PREFIX "addq %1,%0"
LOCK_PREFIX        59 arch/x86/include/asm/atomic64_64.h 	asm volatile(LOCK_PREFIX "subq %1,%0"
LOCK_PREFIX        75 arch/x86/include/asm/atomic64_64.h 	GEN_BINARY_RMWcc(LOCK_PREFIX "subq", v->counter, "er", i, "%0", "e");
LOCK_PREFIX        86 arch/x86/include/asm/atomic64_64.h 	asm volatile(LOCK_PREFIX "incq %0"
LOCK_PREFIX        99 arch/x86/include/asm/atomic64_64.h 	asm volatile(LOCK_PREFIX "decq %0"
LOCK_PREFIX       114 arch/x86/include/asm/atomic64_64.h 	GEN_UNARY_RMWcc(LOCK_PREFIX "decq", v->counter, "%0", "e");
LOCK_PREFIX       127 arch/x86/include/asm/atomic64_64.h 	GEN_UNARY_RMWcc(LOCK_PREFIX "incq", v->counter, "%0", "e");
LOCK_PREFIX       141 arch/x86/include/asm/atomic64_64.h 	GEN_BINARY_RMWcc(LOCK_PREFIX "addq", v->counter, "er", i, "%0", "s");
LOCK_PREFIX        75 arch/x86/include/asm/bitops.h 		asm volatile(LOCK_PREFIX "orb %1,%0"
LOCK_PREFIX        80 arch/x86/include/asm/bitops.h 		asm volatile(LOCK_PREFIX "bts %1,%0"
LOCK_PREFIX       113 arch/x86/include/asm/bitops.h 		asm volatile(LOCK_PREFIX "andb %1,%0"
LOCK_PREFIX       117 arch/x86/include/asm/bitops.h 		asm volatile(LOCK_PREFIX "btr %1,%0"
LOCK_PREFIX       186 arch/x86/include/asm/bitops.h 		asm volatile(LOCK_PREFIX "xorb %1,%0"
LOCK_PREFIX       190 arch/x86/include/asm/bitops.h 		asm volatile(LOCK_PREFIX "btc %1,%0"
LOCK_PREFIX       206 arch/x86/include/asm/bitops.h 	GEN_BINARY_RMWcc(LOCK_PREFIX "bts", *addr, "Ir", nr, "%0", "c");
LOCK_PREFIX       252 arch/x86/include/asm/bitops.h 	GEN_BINARY_RMWcc(LOCK_PREFIX "btr", *addr, "Ir", nr, "%0", "c");
LOCK_PREFIX       305 arch/x86/include/asm/bitops.h 	GEN_BINARY_RMWcc(LOCK_PREFIX "btc", *addr, "Ir", nr, "%0", "c");
LOCK_PREFIX       134 arch/x86/include/asm/cmpxchg.h 	__raw_cmpxchg((ptr), (old), (new), (size), LOCK_PREFIX)
LOCK_PREFIX       166 arch/x86/include/asm/cmpxchg.h #define xadd(ptr, inc)		__xadd((ptr), (inc), LOCK_PREFIX)
LOCK_PREFIX       207 arch/x86/include/asm/cmpxchg.h #define add_smp(ptr, inc)	__add((ptr), (inc), LOCK_PREFIX)
LOCK_PREFIX       228 arch/x86/include/asm/cmpxchg.h 	__cmpxchg_double(LOCK_PREFIX, p1, p2, o1, o2, n1, n2)
LOCK_PREFIX        30 arch/x86/include/asm/cmpxchg_32.h 		     LOCK_PREFIX "cmpxchg8b %0\n\t"
LOCK_PREFIX        49 arch/x86/include/asm/cmpxchg_32.h 	asm volatile(LOCK_PREFIX "cmpxchg8b %1"
LOCK_PREFIX        31 arch/x86/include/asm/futex.h 		     "2:\t" LOCK_PREFIX "cmpxchgl %3, %2\n"	\
LOCK_PREFIX        65 arch/x86/include/asm/futex.h 		__futex_atomic_op1(LOCK_PREFIX "xaddl %0, %2", ret, oldval,
LOCK_PREFIX        31 arch/x86/include/asm/mutex_32.h 	asm volatile(LOCK_PREFIX "   decl (%%eax)\n"		\
LOCK_PREFIX        77 arch/x86/include/asm/mutex_32.h 	asm volatile(LOCK_PREFIX "   incl (%%eax)\n"		\
LOCK_PREFIX        23 arch/x86/include/asm/mutex_64.h 	asm_volatile_goto(LOCK_PREFIX "   decl %0\n"
LOCK_PREFIX        40 arch/x86/include/asm/mutex_64.h 	asm volatile(LOCK_PREFIX "   decl (%%rdi)\n"		\
LOCK_PREFIX        78 arch/x86/include/asm/mutex_64.h 	asm_volatile_goto(LOCK_PREFIX "   incl %0\n"
LOCK_PREFIX        95 arch/x86/include/asm/mutex_64.h 	asm volatile(LOCK_PREFIX "   incl (%%rdi)\n"		\
LOCK_PREFIX        66 arch/x86/include/asm/rwsem.h 		     LOCK_PREFIX _ASM_INC "(%1)\n\t"
LOCK_PREFIX        89 arch/x86/include/asm/rwsem.h 		     LOCK_PREFIX "  cmpxchg  %2,%0\n\t"
LOCK_PREFIX       106 arch/x86/include/asm/rwsem.h 		     LOCK_PREFIX "  xadd      %1,(%2)\n\t"
LOCK_PREFIX       138 arch/x86/include/asm/rwsem.h 		     LOCK_PREFIX "  cmpxchg  %2,%0\n\t"
LOCK_PREFIX       157 arch/x86/include/asm/rwsem.h 		     LOCK_PREFIX "  xadd      %1,(%2)\n\t"
LOCK_PREFIX       175 arch/x86/include/asm/rwsem.h 		     LOCK_PREFIX "  xadd      %1,(%2)\n\t"
LOCK_PREFIX       192 arch/x86/include/asm/rwsem.h 		     LOCK_PREFIX _ASM_ADD "%2,(%1)\n\t"
LOCK_PREFIX       211 arch/x86/include/asm/rwsem.h 	asm volatile(LOCK_PREFIX _ASM_ADD "%1,%0"
LOCK_PREFIX        34 arch/x86/include/asm/spinlock.h # define UNLOCK_LOCK_PREFIX LOCK_PREFIX
LOCK_PREFIX       549 arch/x86/include/asm/uaccess.h 			"1:\t" LOCK_PREFIX "cmpxchgb %4, %2\n"		\
LOCK_PREFIX       565 arch/x86/include/asm/uaccess.h 			"1:\t" LOCK_PREFIX "cmpxchgw %4, %2\n"		\
LOCK_PREFIX       581 arch/x86/include/asm/uaccess.h 			"1:\t" LOCK_PREFIX "cmpxchgl %4, %2\n"		\
LOCK_PREFIX       600 arch/x86/include/asm/uaccess.h 			"1:\t" LOCK_PREFIX "cmpxchgq %4, %2\n"		\
LOCK_PREFIX       763 arch/x86/include/asm/uv/uv_bau.h 	asm volatile(LOCK_PREFIX "xaddw %0, %1"