/kernel/linux/linux-5.10/arch/x86/crypto/ |
H A D | aesni-intel_asm.S | 182 # states of %xmm registers %xmm6:%xmm15 not saved 183 # all %xmm registers are clobbered 198 # clobbers r12, and tmp xmm registers. 648 movups (\PLAIN_CYPH_IN), %xmm1 # If more than 16 bytes, just fill xmm 795 movdqu AadHash(%arg2), %xmm\i # XMM0 = Y0 808 movdqa \XMM0, %xmm\index 810 MOVADQ \XMM0, %xmm\index 812 pshufb %xmm14, %xmm\index # perform a 16 byte swap 813 pxor \TMP2, %xmm\index 823 aesenc \TMP1, %xmm\inde [all...] |
H A D | aes_ctrby8_avx-x86_64.S | 129 /* generate a unique variable for xmm register */ 131 var_xdata = %xmm\n
|
H A D | sha256-avx-asm.S | 71 # COPY_XMM_AND_BSWAP xmm, [mem], byte_flip_mask 72 # Load xmm with mem and byte swap each dword
|
H A D | sha256-ssse3-asm.S | 65 # COPY_XMM_AND_BSWAP xmm, [mem], byte_flip_mask 66 # Load xmm with mem and byte swap each dword
|
H A D | aesni-intel_avx-x86_64.S | 78 ## padded AAD in xmm register = {A1 A0 0 0} 94 ## padded AAD in xmm register = {A2 A1 A0 0} 230 reg_\r = %xmm\n 296 # clobbering all xmm registers 757 vmovdqu (\PLAIN_CYPH_IN), %xmm1 # If more than 16 bytes, just fill xmm
|
/kernel/linux/linux-6.6/arch/x86/crypto/ |
H A D | aesni-intel_asm.S | 178 # states of %xmm registers %xmm6:%xmm15 not saved 179 # all %xmm registers are clobbered 194 # clobbers r12, and tmp xmm registers. 644 movups (\PLAIN_CYPH_IN), %xmm1 # If more than 16 bytes, just fill xmm 791 movdqu AadHash(%arg2), %xmm\i # XMM0 = Y0 804 movdqa \XMM0, %xmm\index 806 MOVADQ \XMM0, %xmm\index 808 pshufb %xmm14, %xmm\index # perform a 16 byte swap 809 pxor \TMP2, %xmm\index 819 aesenc \TMP1, %xmm\inde [all...] |
H A D | aes_ctrby8_avx-x86_64.S | 27 * Note: the "x" prefix in these aliases means "this is an xmm register". The 93 /* generate a unique variable for xmm register */ 95 var_xdata = %xmm\n
|
H A D | sha256-avx-asm.S | 72 # COPY_XMM_AND_BSWAP xmm, [mem], byte_flip_mask 73 # Load xmm with mem and byte swap each dword
|
H A D | sha256-ssse3-asm.S | 66 # COPY_XMM_AND_BSWAP xmm, [mem], byte_flip_mask 67 # Load xmm with mem and byte swap each dword
|
/kernel/linux/linux-6.6/arch/x86/kvm/ |
H A D | hyperv.c | 1818 sse128_t xmm[HV_HYPERCALL_MAX_XMM_REGISTERS]; member 1854 data[i] = sse128_hi(hc->xmm[j / 2]); in kvm_hv_get_hc_data() 1856 data[i] = sse128_lo(hc->xmm[j / 2]); in kvm_hv_get_hc_data() 1998 flush.processor_mask = sse128_lo(hc->xmm[0]); in kvm_hv_flush_tlb() 2028 &hc->xmm[0], sizeof(hc->xmm[0])); in kvm_hv_flush_tlb() 2204 send_ipi_ex.vp_set.valid_bank_mask = sse128_lo(hc->xmm[0]); in kvm_hv_send_ipi() 2415 _kvm_read_sse_reg(reg, &hc->xmm[reg]); in kvm_hv_hypercall_read_xmm()
|
H A D | kvm_emulate.h | 244 unsigned xmm; member
|
/kernel/linux/linux-5.10/tools/perf/util/intel-pt-decoder/ |
H A D | intel-pt-decoder.h | 195 uint64_t xmm[INTEL_PT_BLK_ITEM_ID_CNT]; member
|
/kernel/linux/linux-6.6/tools/perf/util/intel-pt-decoder/ |
H A D | intel-pt-decoder.h | 202 uint64_t xmm[INTEL_PT_BLK_ITEM_ID_CNT]; member
|
/kernel/linux/linux-5.10/arch/x86/include/asm/ |
H A D | perf_event.h | 355 u64 xmm[16*2]; /* two entries for each register */ member
|
/kernel/linux/linux-5.10/arch/x86/include/uapi/asm/ |
H A D | kvm.h | 170 __u8 xmm[16][16]; member
|
/kernel/linux/linux-5.10/arch/x86/kvm/ |
H A D | kvm_emulate.h | 257 unsigned xmm; member
|
/kernel/linux/linux-5.10/tools/arch/x86/include/uapi/asm/ |
H A D | kvm.h | 170 __u8 xmm[16][16]; member
|
/kernel/linux/linux-6.6/arch/x86/include/asm/ |
H A D | perf_event.h | 423 u64 xmm[16*2]; /* two entries for each register */ member
|
/kernel/linux/patches/linux-6.6/prebuilts/usr/include/asm-x86/asm/ |
H A D | kvm.h | 156 __u8 xmm[16][16]; member
|
/kernel/linux/patches/linux-5.10/prebuilts/usr/include/asm-x86/asm/ |
H A D | kvm.h | 156 __u8 xmm[16][16]; member
|
/kernel/linux/linux-5.10/arch/x86/events/intel/ |
H A D | ds.c | 1596 struct pebs_xmm *xmm = next_record; in setup_pebs_adaptive_sample_data() local 1598 next_record = xmm + 1; in setup_pebs_adaptive_sample_data() 1599 perf_regs->xmm_regs = xmm->xmm; in setup_pebs_adaptive_sample_data()
|
/kernel/linux/linux-6.6/arch/x86/events/intel/ |
H A D | ds.c | 1901 struct pebs_xmm *xmm = next_record; in setup_pebs_adaptive_sample_data() local 1903 next_record = xmm + 1; in setup_pebs_adaptive_sample_data() 1904 perf_regs->xmm_regs = xmm->xmm; in setup_pebs_adaptive_sample_data()
|
/kernel/linux/linux-6.6/arch/x86/include/uapi/asm/ |
H A D | kvm.h | 178 __u8 xmm[16][16]; member
|
/kernel/linux/linux-6.6/tools/arch/x86/include/uapi/asm/ |
H A D | kvm.h | 178 __u8 xmm[16][16]; member
|
/kernel/linux/linux-5.10/tools/perf/util/ |
H A D | intel-pt.c | 1717 const u64 *xmm = items->xmm; in intel_pt_add_xmm() local 1727 for (; mask; mask >>= 1, xmm++) { in intel_pt_add_xmm() 1729 *pos++ = *xmm; in intel_pt_add_xmm()
|