/linux-master/arch/xtensa/lib/ |
H A D | mulsi3.S | 72 xor a5, a2, a3 /* Top bit is 1 if one input is negative. */
|
/linux-master/arch/arm64/include/asm/ |
H A D | atomic_ll_sc.h | 108 ATOMIC_OPS(xor, eor, K) 204 ATOMIC64_OPS(xor, eor, L)
|
/linux-master/arch/alpha/lib/ |
H A D | ev6-memset.S | 65 xor $16,$6,$1 # E : will complete write be within one quadword? 248 xor $16,$6,$1 # E : will complete write be within one quadword? 432 xor $16,$6,$1 # E : will complete write be within one quadword?
|
H A D | ev6-divide.S | 77 #define GETSIGN(x) xor $24,$25,x
|
/linux-master/arch/x86/crypto/ |
H A D | chacha-avx2-x86_64.S | 169 # xor and write second block 199 # xor remaining bytes from partial register into output 420 # xor and write second block 474 # xor and write fourth block 504 # xor remaining bytes from partial register into output 882 # xor/write first four blocks 943 # xor remaining blocks, write to output 998 # xor remaining bytes from partial register into output
|
H A D | chacha-avx512vl-x86_64.S | 142 # xor and write second block 172 # xor remaining bytes from partial register into output 354 # xor and write second block 408 # xor and write fourth block 438 # xor remaining bytes from partial register into output 705 # xor/write first four blocks 764 # xor remaining blocks, write to output 818 # xor remaining bytes from partial register into output
|
H A D | aesni-intel_asm.S | 252 xor %r11d, %r11d 280 xor %r11d, %r11d # initialise the data pointer offset as zero 537 pxor \TMP3, \TMP2 # xor the shifted versions 554 pxor \TMP3,\TMP2 # xor the shifted versions 571 xor %eax, %eax 582 xor %eax, %eax 696 xor %eax, %eax 731 xor %eax, %eax 1154 pxor \TMP3, \TMP2 # xor the shifted versions 1169 pxor \TMP3,\TMP2 # xor th [all...] |
H A D | camellia-x86_64-asm_64.S | 185 * %rcx: bool xor 221 enc_outunpack(xor, RT1); 419 * %rcx: bool xor 456 enc_outunpack2(xor, RT2);
|
/linux-master/arch/arm/include/asm/ |
H A D | atomic.h | 245 ATOMIC_OPS(xor, ^=, eor) 399 ATOMIC64_OPS(xor, eor, eor)
|
H A D | xor.h | 3 * arch/arm/include/asm/xor.h 8 #include <asm-generic/xor.h>
|
/linux-master/arch/riscv/lib/ |
H A D | memmove.S | 87 xor t0, a0, a1
|
/linux-master/arch/powerpc/kernel/ |
H A D | kvm_emul.S | 129 xor r31, r30, r31
|
/linux-master/arch/sparc/kernel/ |
H A D | ktlb.S | 122 xor %g2, %g4, %g5
|
/linux-master/lib/ |
H A D | atomic64.c | 125 ATOMIC64_OPS(xor, ^=)
|
/linux-master/arch/x86/include/asm/ |
H A D | xor_32.h | 550 #include <asm-generic/xor.h> 552 /* We force the use of the SSE xor block because it can write around L2.
|
H A D | nospec-branch.h | 84 xor %eax, %eax; \
|
/linux-master/tools/testing/selftests/bpf/prog_tests/ |
H A D | atomics.c | 107 prog_fd = skel->progs.xor.prog_fd; 189 if (test__start_subtest("xor"))
|
/linux-master/arch/mips/dec/ |
H A D | int-handler.S | 269 2: xor t1,t3
|
/linux-master/arch/xtensa/kernel/ |
H A D | coprocessor.S | 225 xor a10, a10, a2
|
/linux-master/arch/sparc/lib/ |
H A D | atomic32.c | 65 ATOMIC_FETCH_OP(xor, ^=)
|
H A D | memscan_64.S | 50 xor %o4, %o5, %o4
|
/linux-master/include/net/netfilter/ |
H A D | nf_tables_core.h | 34 u32 xor; member in struct:nft_bitwise_fast_expr
|
/linux-master/arch/parisc/kernel/ |
H A D | syscall.S | 166 xor %r1,%r30,%r30 /* ye olde xor trick */ 167 xor %r1,%r30,%r1 168 xor %r1,%r30,%r30 196 xor %r19,%r2,%r2 /* clear bottom bit */ 561 xor %r30,%r1,%r30
|
/linux-master/drivers/gpu/drm/nouveau/nvkm/engine/gr/fuc/ |
H A D | com.fuc | 37 xor $r8 8
|
/linux-master/arch/x86/boot/compressed/ |
H A D | mem_encrypt.S | 49 xor %eax, %eax
|