/linux-master/arch/x86/crypto/ |
H A D | aesni-intel_asm.S | 178 # states of %xmm registers %xmm6:%xmm15 not saved 179 # all %xmm registers are clobbered 194 # clobbers r12, and tmp xmm registers. 644 movups (\PLAIN_CYPH_IN), %xmm1 # If more than 16 bytes, just fill xmm 791 movdqu AadHash(%arg2), %xmm\i # XMM0 = Y0 804 movdqa \XMM0, %xmm\index 806 MOVADQ \XMM0, %xmm\index 808 pshufb %xmm14, %xmm\index # perform a 16 byte swap 809 pxor \TMP2, %xmm\index 819 aesenc \TMP1, %xmm\inde [all...] |
H A D | sha256-ssse3-asm.S | 66 # COPY_XMM_AND_BSWAP xmm, [mem], byte_flip_mask 67 # Load xmm with mem and byte swap each dword
|
H A D | sha256-avx-asm.S | 72 # COPY_XMM_AND_BSWAP xmm, [mem], byte_flip_mask 73 # Load xmm with mem and byte swap each dword
|
H A D | aes_ctrby8_avx-x86_64.S | 27 * Note: the "x" prefix in these aliases means "this is an xmm register". The 93 /* generate a unique variable for xmm register */ 95 var_xdata = %xmm\n
|
H A D | aesni-intel_avx-x86_64.S | 78 ## padded AAD in xmm register = {A1 A0 0 0} 94 ## padded AAD in xmm register = {A2 A1 A0 0} 202 reg_\r = %xmm\n 262 # clobbering all xmm registers 725 vmovdqu (\PLAIN_CYPH_IN), %xmm1 # If more than 16 bytes, just fill xmm
|
/linux-master/tools/perf/util/intel-pt-decoder/ |
H A D | intel-pt-decoder.h | 202 uint64_t xmm[INTEL_PT_BLK_ITEM_ID_CNT]; member in struct:intel_pt_blk_items::__anon4234::__anon4235
|
/linux-master/arch/x86/kvm/ |
H A D | hyperv.c | 1868 sse128_t xmm[HV_HYPERCALL_MAX_XMM_REGISTERS]; member in struct:kvm_hv_hcall 1904 data[i] = sse128_hi(hc->xmm[j / 2]); 1906 data[i] = sse128_lo(hc->xmm[j / 2]); 2048 flush.processor_mask = sse128_lo(hc->xmm[0]); 2078 &hc->xmm[0], sizeof(hc->xmm[0])); 2254 send_ipi_ex.vp_set.valid_bank_mask = sse128_lo(hc->xmm[0]); 2465 _kvm_read_sse_reg(reg, &hc->xmm[reg]);
|
H A D | kvm_emulate.h | 253 unsigned xmm; member in union:operand::__anon124
|
H A D | emulate.c | 1150 op->addr.xmm = reg; 1201 op->addr.xmm = ctxt->modrm_rm; 1809 kvm_write_sse_reg(op->addr.xmm, &op->vec_val);
|
H A D | x86.c | 11998 memcpy(fpu->xmm, fxsave->xmm_space, sizeof(fxsave->xmm_space)); 12022 memcpy(fxsave->xmm_space, fpu->xmm, sizeof(fxsave->xmm_space));
|
/linux-master/arch/x86/include/asm/ |
H A D | perf_event.h | 436 u64 xmm[16*2]; /* two entries for each register */ member in struct:pebs_xmm
|
/linux-master/arch/x86/include/uapi/asm/ |
H A D | kvm.h | 178 __u8 xmm[16][16]; member in struct:kvm_fpu
|
/linux-master/arch/x86/events/intel/ |
H A D | ds.c | 1902 struct pebs_xmm *xmm = next_record; local 1904 next_record = xmm + 1; 1905 perf_regs->xmm_regs = xmm->xmm;
|
/linux-master/tools/arch/x86/include/uapi/asm/ |
H A D | kvm.h | 178 __u8 xmm[16][16]; member in struct:kvm_fpu
|
/linux-master/tools/perf/util/ |
H A D | intel-pt.c | 2169 const u64 *xmm = items->xmm; local 2179 for (; mask; mask >>= 1, xmm++) { 2181 *pos++ = *xmm;
|