Lines Matching defs:data

731 				  struct vmx_uret_msr *msr, u64 data)
738 ret = kvm_set_user_return_msr(slot, data, msr->mask);
742 msr->data = data;
1152 vmx->guest_uret_msrs[i].data = guest_efer;
1314 vmx->guest_uret_msrs[i].data,
1412 static void vmx_write_guest_kernel_gs_base(struct vcpu_vmx *vmx, u64 data)
1416 wrmsrl(MSR_KERNEL_GS_BASE, data);
1418 vmx->msr_guest_kernel_gs_base = data;
1597 static int vmx_rtit_ctl_check(struct kvm_vcpu *vcpu, u64 data)
1606 if (data & vmx->pt_desc.ctl_bitmask)
1614 ((vmx->pt_desc.guest.ctl ^ data) & ~RTIT_CTL_TRACEEN))
1622 if ((data & RTIT_CTL_TRACEEN) && !(data & RTIT_CTL_TOPA) &&
1623 !(data & RTIT_CTL_FABRIC_EN) &&
1634 !test_bit((data & RTIT_CTL_MTC_RANGE) >>
1640 !test_bit((data & RTIT_CTL_CYC_THRESH) >>
1645 !test_bit((data & RTIT_CTL_PSB_FREQ) >>
1653 value = (data & RTIT_CTL_ADDR0) >> RTIT_CTL_ADDR0_OFFSET;
1656 value = (data & RTIT_CTL_ADDR1) >> RTIT_CTL_ADDR1_OFFSET;
1659 value = (data & RTIT_CTL_ADDR2) >> RTIT_CTL_ADDR2_OFFSET;
1662 value = (data & RTIT_CTL_ADDR3) >> RTIT_CTL_ADDR3_OFFSET;
1970 return !(msr->data & ~valid_bits);
1979 return vmx_get_vmx_msr(&vmcs_config.nested, msr->index, &msr->data);
1986 * Reads an msr value (of 'msr_info->index') into 'msr_info->data'.
1999 msr_info->data = vmcs_readl(GUEST_FS_BASE);
2002 msr_info->data = vmcs_readl(GUEST_GS_BASE);
2005 msr_info->data = vmx_read_guest_kernel_gs_base(vmx);
2019 msr_info->data = vmx->msr_ia32_umwait_control;
2026 msr_info->data = to_vmx(vcpu)->spec_ctrl;
2029 msr_info->data = vmcs_read32(GUEST_SYSENTER_CS);
2032 msr_info->data = vmcs_readl(GUEST_SYSENTER_EIP);
2035 msr_info->data = vmcs_readl(GUEST_SYSENTER_ESP);
2042 msr_info->data = vmcs_read64(GUEST_BNDCFGS);
2049 msr_info->data = vcpu->arch.mcg_ext_ctl;
2052 msr_info->data = vmx->msr_ia32_feature_control;
2058 msr_info->data = to_vmx(vcpu)->msr_ia32_sgxlepubkeyhash
2065 &msr_info->data))
2076 &msr_info->data);
2081 msr_info->data = vmx->pt_desc.guest.ctl;
2086 msr_info->data = vmx->pt_desc.guest.status;
2093 msr_info->data = vmx->pt_desc.guest.cr3_match;
2102 msr_info->data = vmx->pt_desc.guest.output_base;
2111 msr_info->data = vmx->pt_desc.guest.output_mask;
2119 msr_info->data = vmx->pt_desc.guest.addr_b[index / 2];
2121 msr_info->data = vmx->pt_desc.guest.addr_a[index / 2];
2124 msr_info->data = vmcs_read64(GUEST_IA32_DEBUGCTL);
2130 msr_info->data = msr->data;
2140 u64 data)
2144 return (u32)data;
2146 return (unsigned long)data;
2175 u64 data = msr_info->data;
2185 vmcs_writel(GUEST_FS_BASE, data);
2189 vmcs_writel(GUEST_GS_BASE, data);
2192 vmx_write_guest_kernel_gs_base(vmx, data);
2205 if (!ret && data) {
2215 get_vmcs12(vcpu)->guest_sysenter_cs = data;
2216 vmcs_write32(GUEST_SYSENTER_CS, data);
2220 data = nested_vmx_truncate_sysenter_addr(vcpu, data);
2221 get_vmcs12(vcpu)->guest_sysenter_eip = data;
2223 vmcs_writel(GUEST_SYSENTER_EIP, data);
2227 data = nested_vmx_truncate_sysenter_addr(vcpu, data);
2228 get_vmcs12(vcpu)->guest_sysenter_esp = data;
2230 vmcs_writel(GUEST_SYSENTER_ESP, data);
2235 invalid = data & ~vmx_get_supported_debugctl(vcpu, msr_info->host_initiated);
2237 kvm_pr_unimpl_wrmsr(vcpu, msr_index, data);
2238 data &= ~(DEBUGCTLMSR_BTF|DEBUGCTLMSR_LBR);
2247 get_vmcs12(vcpu)->guest_ia32_debugctl = data;
2249 vmcs_write64(GUEST_IA32_DEBUGCTL, data);
2251 (data & DEBUGCTLMSR_LBR))
2260 if (is_noncanonical_address(data & PAGE_MASK, vcpu) ||
2261 (data & MSR_IA32_BNDCFGS_RSVD))
2267 get_vmcs12(vcpu)->guest_bndcfgs = data;
2269 vmcs_write64(GUEST_BNDCFGS, data);
2276 if (data & (BIT_ULL(1) | GENMASK_ULL(63, 32)))
2279 vmx->msr_ia32_umwait_control = data;
2286 if (kvm_spec_ctrl_test_value(data))
2289 vmx->spec_ctrl = data;
2290 if (!data)
2313 if (data & ~(TSX_CTRL_RTM_DISABLE | TSX_CTRL_CPUID_CLEAR))
2323 get_vmcs12(vcpu)->guest_ia32_pat = data;
2326 vmcs_write64(GUEST_IA32_PAT, data);
2332 (data & ~MCG_EXT_CTL_LMCE_EN))
2334 vcpu->arch.mcg_ext_ctl = data;
2340 vmx->msr_ia32_feature_control = data;
2341 if (msr_info->host_initiated && data == 0)
2365 [msr_index - MSR_IA32_SGXLEPUBKEYHASH0] = data;
2372 return vmx_set_vmx_msr(vcpu, msr_index, data);
2375 vmx_rtit_ctl_check(vcpu, data) ||
2378 vmcs_write64(GUEST_IA32_RTIT_CTL, data);
2379 vmx->pt_desc.guest.ctl = data;
2385 if (data & MSR_IA32_RTIT_STATUS_MASK)
2387 vmx->pt_desc.guest.status = data;
2395 vmx->pt_desc.guest.cr3_match = data;
2405 if (!pt_output_base_valid(vcpu, data))
2407 vmx->pt_desc.guest.output_base = data;
2417 vmx->pt_desc.guest.output_mask = data;
2425 if (is_noncanonical_address(data, vcpu))
2428 vmx->pt_desc.guest.addr_b[index / 2] = data;
2430 vmx->pt_desc.guest.addr_a[index / 2] = data;
2433 if (data && !vcpu_to_pmu(vcpu)->version)
2435 if (data & PMU_CAP_LBR_FMT) {
2436 if ((data & PMU_CAP_LBR_FMT) !=
2442 if (data & PERF_CAP_PEBS_FORMAT) {
2443 if ((data & PERF_CAP_PEBS_MASK) !=
2460 ret = vmx_set_guest_uret_msr(vmx, msr, data);
3848 u16 data;
3856 data = TSS_BASE_SIZE + TSS_REDIRECTION_SIZE;
3857 if (__copy_to_user(ua + TSS_IOPB_BASE_OFFSET, &data, sizeof(u16)))
3860 data = ~0;
3861 if (__copy_to_user(ua + RMODE_TSS_SIZE - 1, &data, sizeof(u8)))
5246 vcpu->run->internal.data[0] = vect_info;
5247 vcpu->run->internal.data[1] = intr_info;
5248 vcpu->run->internal.data[2] = error_code;
5249 vcpu->run->internal.data[3] = vcpu->arch.last_vmentry_cpu;
6520 vcpu->run->internal.data[0] = vectoring_info;
6521 vcpu->run->internal.data[1] = exit_reason.full;
6522 vcpu->run->internal.data[2] = vcpu->arch.exit_qualification;
6524 vcpu->run->internal.data[ndata++] =
6527 vcpu->run->internal.data[ndata++] = vcpu->arch.last_vmentry_cpu;
6586 vcpu->run->internal.data[0] = exit_reason.full;
6587 vcpu->run->internal.data[1] = vcpu->arch.last_vmentry_cpu;
7558 #define L1TF_MSG_SMT "L1TF CPU bug present and SMT on, data leak possible. See CVE-2018-3646 and https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/l1tf.html for details.\n"
7559 #define L1TF_MSG_L1D "L1TF CPU bug present and virtualization mitigation disabled, data leak possible. See CVE-2018-3646 and https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/l1tf.html for details.\n"