/kernel/linux/linux-6.6/tools/testing/selftests/sgx/ |
H A D | main.c | 172 struct sgx_enclave_run run; in FIXTURE() local 262 #define ENCL_CALL(op, run, clobbered) \ 267 EENTER, 0, 0, (run)); \ 270 (run)); \ 274 #define EXPECT_EEXIT(run) \ 276 EXPECT_EQ((run)->function, EEXIT); \ 277 if ((run)->function != EEXIT) \ 278 TH_LOG("0x%02x 0x%02x 0x%016llx", (run)->exception_vector, \ 279 (run)->exception_error_code, (run) 532 test_handler(long rdi, long rsi, long rdx, long ursp, long r8, long r9, struct sgx_enclave_run *run) test_handler() argument [all...] |
/kernel/linux/linux-6.6/drivers/media/test-drivers/visl/ |
H A D | visl-dec.c | 37 __kernel_size_t buflen, struct visl_run *run) in visl_get_ref_frames() 55 vb2_buf = vb2_find_buffer(cap_q, run->fwht.params->backward_ref_ts); in visl_get_ref_frames() 58 run->fwht.params->backward_ref_ts, in visl_get_ref_frames() 67 b_ref = vb2_find_buffer(cap_q, run->mpeg2.pic->backward_ref_ts); in visl_get_ref_frames() 68 f_ref = vb2_find_buffer(cap_q, run->mpeg2.pic->forward_ref_ts); in visl_get_ref_frames() 73 run->mpeg2.pic->backward_ref_ts, in visl_get_ref_frames() 75 run->mpeg2.pic->forward_ref_ts, in visl_get_ref_frames() 85 last = vb2_find_buffer(cap_q, run->vp8.frame->last_frame_ts); in visl_get_ref_frames() 86 golden = vb2_find_buffer(cap_q, run->vp8.frame->golden_frame_ts); in visl_get_ref_frames() 87 alt = vb2_find_buffer(cap_q, run in visl_get_ref_frames() 36 visl_get_ref_frames(struct visl_ctx *ctx, u8 *buf, __kernel_size_t buflen, struct visl_run *run) visl_get_ref_frames() argument 195 visl_tpg_fill_sequence(struct visl_ctx *ctx, struct visl_run *run, char buf[], size_t bufsz) visl_tpg_fill_sequence() argument 215 visl_tpg_fill(struct visl_ctx *ctx, struct visl_run *run) visl_tpg_fill() argument 358 visl_trace_ctrls(struct visl_ctx *ctx, struct visl_run *run) visl_trace_ctrls() argument 420 struct visl_run run = {}; visl_device_run() local [all...] |
/kernel/linux/linux-6.6/tools/testing/selftests/kvm/s390x/ |
H A D | sync_regs_test.c | 78 struct kvm_run *run = vcpu->run; in test_read_invalid() local 82 run->kvm_valid_regs = INVALID_SYNC_FIELD; in test_read_invalid() 87 run->kvm_valid_regs = 0; in test_read_invalid() 89 run->kvm_valid_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS; in test_read_invalid() 94 run->kvm_valid_regs = 0; in test_read_invalid() 99 struct kvm_run *run = vcpu->run; in test_set_invalid() local 103 run->kvm_dirty_regs = INVALID_SYNC_FIELD; in test_set_invalid() 108 run in test_set_invalid() 120 struct kvm_run *run = vcpu->run; test_req_and_verify_all_valid_regs() local 146 struct kvm_run *run = vcpu->run; test_set_and_verify_various_reg_values() local 185 struct kvm_run *run = vcpu->run; test_clear_kvm_dirty_regs_bits() local [all...] |
/kernel/linux/linux-6.6/tools/testing/selftests/kvm/x86_64/ |
H A D | sync_regs_test.c | 91 struct kvm_run *run = (struct kvm_run *)arg; in race_events_inj_pen() local 92 struct kvm_vcpu_events *events = &run->s.regs.events; in race_events_inj_pen() 97 WRITE_ONCE(run->kvm_dirty_regs, KVM_SYNC_X86_EVENTS); in race_events_inj_pen() 114 struct kvm_run *run = (struct kvm_run *)arg; in race_events_exc() local 115 struct kvm_vcpu_events *events = &run->s.regs.events; in race_events_exc() 118 WRITE_ONCE(run->kvm_dirty_regs, KVM_SYNC_X86_EVENTS); in race_events_exc() 136 struct kvm_run *run = (struct kvm_run *)arg; in race_sregs_cr4() local 137 __u64 *cr4 = &run->s.regs.sregs.cr4; in race_sregs_cr4() 142 WRITE_ONCE(run->kvm_dirty_regs, KVM_SYNC_X86_SREGS); in race_sregs_cr4() 161 struct kvm_run *run; in race_sync_regs() local 215 struct kvm_run *run; main() local [all...] |
H A D | debug_regs.c | 83 struct kvm_run *run; in main() local 101 run = vcpu->run; in main() 108 TEST_ASSERT(run->exit_reason == KVM_EXIT_DEBUG && in main() 109 run->debug.arch.exception == BP_VECTOR && in main() 110 run->debug.arch.pc == CAST_TO_RIP(sw_bp), in main() 112 run->exit_reason, run->debug.arch.exception, in main() 113 run->debug.arch.pc, CAST_TO_RIP(sw_bp)); in main() 125 TEST_ASSERT(run in main() [all...] |
H A D | userspace_msr_exit_test.c | 275 * Now run the same tests with the instruction emulator. in guest_code_filter_allow() 401 if (vcpu->run->exit_reason == KVM_EXIT_IO && in check_for_guest_assert() 409 struct kvm_run *run = vcpu->run; in process_rdmsr() local 414 TEST_ASSERT(run->msr.index == msr_index, in process_rdmsr() 416 run->msr.index, msr_index); in process_rdmsr() 418 switch (run->msr.index) { in process_rdmsr() 420 run->msr.data = 0; in process_rdmsr() 423 run->msr.error = 1; in process_rdmsr() 426 run in process_rdmsr() 441 struct kvm_run *run = vcpu->run; process_wrmsr() local 612 handle_rdmsr(struct kvm_run *run) handle_rdmsr() argument 629 handle_wrmsr(struct kvm_run *run) handle_wrmsr() argument 653 struct kvm_run *run; test_msr_filter_deny() local [all...] |
H A D | xen_vmcall_test.c | 105 volatile struct kvm_run *run = vcpu->run; in main() local 110 if (run->exit_reason == KVM_EXIT_XEN) { in main() 111 TEST_ASSERT_EQ(run->xen.type, KVM_EXIT_XEN_HCALL); in main() 112 TEST_ASSERT_EQ(run->xen.u.hcall.cpl, 0); in main() 113 TEST_ASSERT_EQ(run->xen.u.hcall.longmode, 1); in main() 114 TEST_ASSERT_EQ(run->xen.u.hcall.input, INPUTVALUE); in main() 115 TEST_ASSERT_EQ(run->xen.u.hcall.params[0], ARGVALUE(1)); in main() 116 TEST_ASSERT_EQ(run->xen.u.hcall.params[1], ARGVALUE(2)); in main() 117 TEST_ASSERT_EQ(run in main() [all...] |
/kernel/linux/linux-5.10/arch/loongarch/kvm/ |
H A D | emulate.c | 37 vcpu->run->exit_reason = KVM_EXIT_IRQ_WINDOW_OPEN; in _kvm_emu_idle() 50 struct kvm_run *run = vcpu->run; in _kvm_emu_mmio_write() local 53 void *data = run->mmio.data; in _kvm_emu_mmio_write() 65 run->mmio.phys_addr = vcpu->arch.badv; in _kvm_emu_mmio_write() 66 if (run->mmio.phys_addr == KVM_INVALID_ADDR) in _kvm_emu_mmio_write() 76 run->mmio.len = 8; in _kvm_emu_mmio_write() 80 run->mmio.len = 4; in _kvm_emu_mmio_write() 94 run->mmio.len = 8; in _kvm_emu_mmio_write() 98 run in _kvm_emu_mmio_write() 175 struct kvm_run *run = vcpu->run; _kvm_emu_mmio_read() local 295 _kvm_complete_mmio_read(struct kvm_vcpu *vcpu, struct kvm_run *run) _kvm_complete_mmio_read() argument [all...] |
H A D | csr.c | 300 int (*get) (struct kvm_run *run, struct kvm_vcpu *vcpu, u32 addr, u64 *res); 301 int (*set) (struct kvm_run *run, struct kvm_vcpu *vcpu, u32 addr, u64 val); 316 static int kvm_iocsr_common_get(struct kvm_run *run, struct kvm_vcpu *vcpu, in kvm_iocsr_common_get() argument 332 static int kvm_iocsr_common_set(struct kvm_run *run, struct kvm_vcpu *vcpu, in kvm_iocsr_common_set() argument 348 static int kvm_misc_set(struct kvm_run *run, struct kvm_vcpu *vcpu, u32 addr, in kvm_misc_set() argument 351 return kvm_iocsr_common_set(run, vcpu, addr, val); in kvm_misc_set() 354 static int kvm_ipi_get(struct kvm_run *run, struct kvm_vcpu *vcpu, u32 addr, in kvm_ipi_get() argument 360 run->mmio.phys_addr = KVM_IPI_REG_ADDRESS(vcpu->vcpu_id, (addr & 0xff)); in kvm_ipi_get() 361 ret = kvm_io_bus_read(vcpu, KVM_MMIO_BUS, run->mmio.phys_addr, in kvm_ipi_get() 362 run in kvm_ipi_get() 372 kvm_extioi_isr_get(struct kvm_run *run, struct kvm_vcpu *vcpu, u32 addr, u64 *res) kvm_extioi_isr_get() argument 390 kvm_ipi_set(struct kvm_run *run, struct kvm_vcpu *vcpu, u32 addr, u64 val) kvm_ipi_set() argument 408 kvm_extioi_set(struct kvm_run *run, struct kvm_vcpu *vcpu, u32 addr, u64 val) kvm_extioi_set() argument 432 kvm_nop_set(struct kvm_run *run, struct kvm_vcpu *vcpu, u32 addr, u64 val) kvm_nop_set() argument 469 _kvm_emu_iocsr_read(struct kvm_run *run, struct kvm_vcpu *vcpu, u32 addr, u64 *res) _kvm_emu_iocsr_read() argument 496 _kvm_emu_iocsr_write(struct kvm_run *run, struct kvm_vcpu *vcpu, u32 addr, u64 val) _kvm_emu_iocsr_write() argument 524 _kvm_emu_iocsr(larch_inst inst, struct kvm_run *run, struct kvm_vcpu *vcpu) _kvm_emu_iocsr() argument 590 _kvm_complete_iocsr_read(struct kvm_vcpu *vcpu, struct kvm_run *run) _kvm_complete_iocsr_read() argument [all...] |
H A D | exit.c | 48 vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR; in _kvm_fault_ni() 97 struct kvm_run *run = vcpu->run; in _kvm_trap_handle_gspr() local 149 er = _kvm_emu_iocsr(inst, run, vcpu); in _kvm_trap_handle_gspr() 237 vcpu->run->exit_reason = KVM_EXIT_MMIO; in _kvm_handle_gspr() 240 vcpu->run->exit_reason = KVM_EXIT_LOONGARCH_IOCSR; in _kvm_handle_gspr() 244 vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR; in _kvm_handle_gspr() 261 vcpu->run->exit_reason = KVM_EXIT_DEBUG; in _kvm_handle_hypcall() 278 vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR; in _kvm_handle_gcm() 294 struct kvm_run *run in _kvm_handle_fpu_disabled() local 321 struct kvm_run *run = vcpu->run; _kvm_handle_lsx_disabled() local 356 struct kvm_run *run = vcpu->run; _kvm_handle_lasx_disabled() local 388 struct kvm_run *run = vcpu->run; _kvm_handle_lbt_disabled() local 407 struct kvm_run *run = vcpu->run; _kvm_handle_read_fault() local 445 struct kvm_run *run = vcpu->run; _kvm_handle_write_fault() local [all...] |
/kernel/linux/linux-5.10/tools/testing/selftests/kvm/x86_64/ |
H A D | debug_regs.c | 69 struct kvm_run *run; in main() local 89 run = vcpu_state(vm, VCPU_ID); in main() 96 TEST_ASSERT(run->exit_reason == KVM_EXIT_DEBUG && in main() 97 run->debug.arch.exception == BP_VECTOR && in main() 98 run->debug.arch.pc == CAST_TO_RIP(sw_bp), in main() 100 run->exit_reason, run->debug.arch.exception, in main() 101 run->debug.arch.pc, CAST_TO_RIP(sw_bp)); in main() 113 TEST_ASSERT(run->exit_reason == KVM_EXIT_DEBUG && in main() 114 run in main() [all...] |
H A D | sync_regs_test.c | 83 struct kvm_run *run; in main() local 105 run = vcpu_state(vm, VCPU_ID); in main() 108 run->kvm_valid_regs = INVALID_SYNC_FIELD; in main() 115 run->kvm_valid_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS; in main() 123 run->kvm_dirty_regs = INVALID_SYNC_FIELD; in main() 130 run->kvm_dirty_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS; in main() 139 run->kvm_valid_regs = TEST_SYNC_FIELDS; in main() 141 TEST_ASSERT(run->exit_reason == KVM_EXIT_IO, in main() 143 run->exit_reason, in main() 144 exit_reason_str(run in main() [all...] |
H A D | user_msr_test.c | 154 static void handle_rdmsr(struct kvm_run *run) in handle_rdmsr() argument 156 run->msr.data = run->msr.index; in handle_rdmsr() 159 if (run->msr.index == MSR_SYSCALL_MASK || in handle_rdmsr() 160 run->msr.index == MSR_GS_BASE) { in handle_rdmsr() 161 TEST_ASSERT(run->msr.reason == KVM_MSR_EXIT_REASON_FILTER, in handle_rdmsr() 165 if (run->msr.index == 0xdeadbeef) { in handle_rdmsr() 166 TEST_ASSERT(run->msr.reason == KVM_MSR_EXIT_REASON_UNKNOWN, in handle_rdmsr() 171 static void handle_wrmsr(struct kvm_run *run) in handle_wrmsr() argument 176 if (run in handle_wrmsr() 200 struct kvm_run *run; main() local [all...] |
/kernel/linux/linux-5.10/tools/testing/selftests/kvm/s390x/ |
H A D | sync_regs_test.c | 79 struct kvm_run *run; in main() local 96 run = vcpu_state(vm, VCPU_ID); in main() 99 run->kvm_valid_regs = INVALID_SYNC_FIELD; in main() 106 run->kvm_valid_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS; in main() 114 run->kvm_dirty_regs = INVALID_SYNC_FIELD; in main() 121 run->kvm_dirty_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS; in main() 129 run->kvm_valid_regs = TEST_SYNC_FIELDS; in main() 132 TEST_ASSERT(run->exit_reason == KVM_EXIT_S390_SIEIC, in main() 134 run->exit_reason, in main() 135 exit_reason_str(run in main() [all...] |
/kernel/linux/linux-6.6/fs/ntfs3/ |
H A D | run.c | 34 static bool run_lookup(const struct runs_tree *run, CLST vcn, size_t *index) in run_lookup() argument 39 if (!run->count) { in run_lookup() 45 max_idx = run->count - 1; in run_lookup() 48 r = run->runs; in run_lookup() 61 *index = run->count; in run_lookup() 72 r = run->runs + mid_idx; in run_lookup() 93 static void run_consolidate(struct runs_tree *run, size_t index) in run_consolidate() argument 96 struct ntfs_run *r = run->runs + index; in run_consolidate() 98 while (index + 1 < run->count) { in run_consolidate() 100 * I should merge current run wit in run_consolidate() 167 run_is_mapped_full(const struct runs_tree *run, CLST svcn, CLST evcn) run_is_mapped_full() argument 192 run_lookup_entry(const struct runs_tree *run, CLST vcn, CLST *lcn, CLST *len, size_t *index) run_lookup_entry() argument 228 run_truncate_head(struct runs_tree *run, CLST vcn) run_truncate_head() argument 263 run_truncate(struct runs_tree *run, CLST vcn) run_truncate() argument 300 run_truncate_around(struct runs_tree *run, CLST vcn) run_truncate_around() argument 316 run_add_entry(struct runs_tree *run, CLST vcn, CLST lcn, CLST len, bool is_mft) run_add_entry() argument 489 run_collapse_range(struct runs_tree *run, CLST vcn, CLST len) run_collapse_range() argument 555 run_insert_range(struct runs_tree *run, CLST vcn, CLST len) run_insert_range() argument 595 run_get_entry(const struct runs_tree *run, size_t index, CLST *vcn, CLST *lcn, CLST *len) run_get_entry() argument 817 run_pack(const struct runs_tree *run, CLST svcn, CLST len, u8 *run_buf, u32 run_buf_size, CLST *packed_vcns) run_pack() argument 920 run_unpack(struct runs_tree *run, struct ntfs_sb_info *sbi, CLST ino, CLST svcn, CLST evcn, CLST vcn, const u8 *run_buf, int run_buf_size) run_unpack() argument 1050 run_unpack_ex(struct runs_tree *run, struct ntfs_sb_info *sbi, CLST ino, CLST svcn, CLST evcn, CLST vcn, const u8 *run_buf, int run_buf_size) run_unpack_ex() argument 1151 run_clone(const struct runs_tree *run, struct runs_tree *new_run) run_clone() argument [all...] |
/kernel/linux/linux-6.6/drivers/staging/media/sunxi/cedrus/ |
H A D | cedrus_dec.c | 29 struct cedrus_run run = {}; in cedrus_device_run() local 33 run.src = v4l2_m2m_next_src_buf(ctx->fh.m2m_ctx); in cedrus_device_run() 34 run.dst = v4l2_m2m_next_dst_buf(ctx->fh.m2m_ctx); in cedrus_device_run() 37 src_req = run.src->vb2_buf.req_obj.req; in cedrus_device_run() 44 run.mpeg2.sequence = cedrus_find_control_data(ctx, in cedrus_device_run() 46 run.mpeg2.picture = cedrus_find_control_data(ctx, in cedrus_device_run() 48 run.mpeg2.quantisation = cedrus_find_control_data(ctx, in cedrus_device_run() 53 run.h264.decode_params = cedrus_find_control_data(ctx, in cedrus_device_run() 55 run.h264.pps = cedrus_find_control_data(ctx, in cedrus_device_run() 57 run in cedrus_device_run() [all...] |
/kernel/linux/linux-5.10/drivers/staging/media/sunxi/cedrus/ |
H A D | cedrus_dec.c | 29 struct cedrus_run run = {}; in cedrus_device_run() local 32 run.src = v4l2_m2m_next_src_buf(ctx->fh.m2m_ctx); in cedrus_device_run() 33 run.dst = v4l2_m2m_next_dst_buf(ctx->fh.m2m_ctx); in cedrus_device_run() 36 src_req = run.src->vb2_buf.req_obj.req; in cedrus_device_run() 43 run.mpeg2.slice_params = cedrus_find_control_data(ctx, in cedrus_device_run() 45 run.mpeg2.quantization = cedrus_find_control_data(ctx, in cedrus_device_run() 50 run.h264.decode_params = cedrus_find_control_data(ctx, in cedrus_device_run() 52 run.h264.pps = cedrus_find_control_data(ctx, in cedrus_device_run() 54 run.h264.scaling_matrix = cedrus_find_control_data(ctx, in cedrus_device_run() 56 run in cedrus_device_run() [all...] |
/kernel/linux/linux-5.10/tools/testing/selftests/arm64/mte/ |
H A D | check_mmap_options.c | 63 int run, result, map_size; in check_anonymous_memory_mapping() local 68 for (run = 0; run < item; run++) { in check_anonymous_memory_mapping() 69 map_size = sizes[run] + OVERFLOW + UNDERFLOW; in check_anonymous_memory_mapping() 75 mte_initialize_current_context(mode, (uintptr_t)ptr, sizes[run]); in check_anonymous_memory_mapping() 77 ptr = mte_insert_tags((void *)ptr, sizes[run]); in check_anonymous_memory_mapping() 83 result = check_mte_memory(ptr, sizes[run], mode, tag_check); in check_anonymous_memory_mapping() 84 mte_clear_tags((void *)ptr, sizes[run]); in check_anonymous_memory_mapping() 95 int run, f in check_file_memory_mapping() local 134 int run, prot_flag, result, fd, map_size; check_clear_prot_mte_flag() local [all...] |
/kernel/linux/linux-6.6/tools/testing/selftests/arm64/mte/ |
H A D | check_mmap_options.c | 63 int run, result, map_size; in check_anonymous_memory_mapping() local 67 for (run = 0; run < item; run++) { in check_anonymous_memory_mapping() 68 map_size = sizes[run] + OVERFLOW + UNDERFLOW; in check_anonymous_memory_mapping() 74 mte_initialize_current_context(mode, (uintptr_t)ptr, sizes[run]); in check_anonymous_memory_mapping() 76 ptr = mte_insert_tags((void *)ptr, sizes[run]); in check_anonymous_memory_mapping() 82 result = check_mte_memory(ptr, sizes[run], mode, tag_check); in check_anonymous_memory_mapping() 83 mte_clear_tags((void *)ptr, sizes[run]); in check_anonymous_memory_mapping() 94 int run, f in check_file_memory_mapping() local 133 int run, prot_flag, result, fd, map_size; check_clear_prot_mte_flag() local [all...] |
/kernel/linux/linux-5.10/arch/s390/kvm/ |
H A D | diag.c | 25 start = vcpu->run->s.regs.gprs[(vcpu->arch.sie_block->ipa & 0xf0) >> 4]; in diag_release_pages() 26 end = vcpu->run->s.regs.gprs[vcpu->arch.sie_block->ipa & 0xf] + PAGE_SIZE; in diag_release_pages() 76 vcpu->run->s.regs.gprs[rx]); in __diag_page_ref_service() 78 if (vcpu->run->s.regs.gprs[rx] & 7) in __diag_page_ref_service() 80 rc = read_guest(vcpu, vcpu->run->s.regs.gprs[rx], rx, &parm, sizeof(parm)); in __diag_page_ref_service() 97 vcpu->run->s.regs.gprs[ry] = 8; in __diag_page_ref_service() 111 vcpu->run->s.regs.gprs[ry] = 0; in __diag_page_ref_service() 125 vcpu->run->s.regs.gprs[ry] = 0; in __diag_page_ref_service() 131 vcpu->run->s.regs.gprs[ry] = 4; in __diag_page_ref_service() 158 tid = vcpu->run in __diag_time_slice_end_directed() [all...] |
/kernel/linux/linux-6.6/arch/s390/kvm/ |
H A D | diag.c | 25 start = vcpu->run->s.regs.gprs[(vcpu->arch.sie_block->ipa & 0xf0) >> 4]; in diag_release_pages() 26 end = vcpu->run->s.regs.gprs[vcpu->arch.sie_block->ipa & 0xf] + PAGE_SIZE; in diag_release_pages() 76 vcpu->run->s.regs.gprs[rx]); in __diag_page_ref_service() 78 if (vcpu->run->s.regs.gprs[rx] & 7) in __diag_page_ref_service() 80 rc = read_guest(vcpu, vcpu->run->s.regs.gprs[rx], rx, &parm, sizeof(parm)); in __diag_page_ref_service() 97 vcpu->run->s.regs.gprs[ry] = 8; in __diag_page_ref_service() 111 vcpu->run->s.regs.gprs[ry] = 0; in __diag_page_ref_service() 125 vcpu->run->s.regs.gprs[ry] = 0; in __diag_page_ref_service() 131 vcpu->run->s.regs.gprs[ry] = 4; in __diag_page_ref_service() 172 tid = vcpu->run in __diag_time_slice_end_directed() [all...] |
/kernel/linux/linux-6.6/arch/mips/kvm/ |
H A D | emulate.c | 962 vcpu->run->exit_reason = KVM_EXIT_IRQ_WINDOW_OPEN; in kvm_mips_emul_wait() 975 struct kvm_run *run = vcpu->run; in kvm_mips_emulate_store() local 976 void *data = run->mmio.data; in kvm_mips_emulate_store() 991 run->mmio.phys_addr = kvm_mips_callbacks->gva_to_gpa( in kvm_mips_emulate_store() 993 if (run->mmio.phys_addr == KVM_INVALID_ADDR) in kvm_mips_emulate_store() 999 run->mmio.len = 8; in kvm_mips_emulate_store() 1009 run->mmio.len = 4; in kvm_mips_emulate_store() 1018 run->mmio.len = 2; in kvm_mips_emulate_store() 1027 run in kvm_mips_emulate_store() 1271 struct kvm_run *run = vcpu->run; kvm_mips_emulate_load() local 1511 struct kvm_run *run = vcpu->run; kvm_mips_complete_mmio_load() local [all...] |
/kernel/linux/linux-6.6/arch/riscv/kvm/ |
H A D | vcpu_insn.c | 148 int (*func)(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn); 151 static int truly_illegal_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, in truly_illegal_insn() argument 167 static int truly_virtual_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, in truly_virtual_insn() argument 197 static int wfi_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn) in wfi_insn() argument 226 * @run: The VCPU run struct containing the CSR data 230 int kvm_riscv_vcpu_csr_return(struct kvm_vcpu *vcpu, struct kvm_run *run) in kvm_riscv_vcpu_csr_return() argument 242 run->riscv_csr.ret_value); in kvm_riscv_vcpu_csr_return() 250 static int csr_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn) in csr_insn() argument 294 run in csr_insn() 370 system_opcode_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn) system_opcode_insn() argument 410 kvm_riscv_vcpu_virtual_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, struct kvm_cpu_trap *trap) kvm_riscv_vcpu_virtual_insn() argument 453 kvm_riscv_vcpu_mmio_load(struct kvm_vcpu *vcpu, struct kvm_run *run, unsigned long fault_addr, unsigned long htinst) kvm_riscv_vcpu_mmio_load() argument 575 kvm_riscv_vcpu_mmio_store(struct kvm_vcpu *vcpu, struct kvm_run *run, unsigned long fault_addr, unsigned long htinst) kvm_riscv_vcpu_mmio_store() argument 703 kvm_riscv_vcpu_mmio_return(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_riscv_vcpu_mmio_return() argument [all...] |
/kernel/linux/linux-5.10/tools/perf/scripts/python/ |
H A D | stat-cpi.py | 23 def store(time, event, cpu, thread, val, ena, run): 24 #print("event %s cpu %d, thread %d, time %d, val %d, ena %d, run %d" % 25 # (event, cpu, thread, time, val, ena, run)) 29 data[key] = [ val, ena, run] 35 def stat__cycles_k(cpu, thread, time, val, ena, run): 36 store(time, "cycles", cpu, thread, val, ena, run); 38 def stat__instructions_k(cpu, thread, time, val, ena, run): 39 store(time, "instructions", cpu, thread, val, ena, run); 41 def stat__cycles_u(cpu, thread, time, val, ena, run): 42 store(time, "cycles", cpu, thread, val, ena, run); [all...] |
/kernel/linux/linux-6.6/tools/perf/scripts/python/ |
H A D | stat-cpi.py | 23 def store(time, event, cpu, thread, val, ena, run): 24 #print("event %s cpu %d, thread %d, time %d, val %d, ena %d, run %d" % 25 # (event, cpu, thread, time, val, ena, run)) 29 data[key] = [ val, ena, run] 35 def stat__cycles_k(cpu, thread, time, val, ena, run): 36 store(time, "cycles", cpu, thread, val, ena, run); 38 def stat__instructions_k(cpu, thread, time, val, ena, run): 39 store(time, "instructions", cpu, thread, val, ena, run); 41 def stat__cycles_u(cpu, thread, time, val, ena, run): 42 store(time, "cycles", cpu, thread, val, ena, run); [all...] |