Home
last modified time | relevance | path

Searched refs:run (Results 1 - 25 of 684) sorted by relevance

12345678910>>...28

/kernel/linux/linux-6.6/tools/testing/selftests/sgx/
H A Dmain.c172 struct sgx_enclave_run run; in FIXTURE() local
262 #define ENCL_CALL(op, run, clobbered) \
267 EENTER, 0, 0, (run)); \
270 (run)); \
274 #define EXPECT_EEXIT(run) \
276 EXPECT_EQ((run)->function, EEXIT); \
277 if ((run)->function != EEXIT) \
278 TH_LOG("0x%02x 0x%02x 0x%016llx", (run)->exception_vector, \
279 (run)->exception_error_code, (run)
532 test_handler(long rdi, long rsi, long rdx, long ursp, long r8, long r9, struct sgx_enclave_run *run) test_handler() argument
[all...]
/kernel/linux/linux-6.6/drivers/media/test-drivers/visl/
H A Dvisl-dec.c37 __kernel_size_t buflen, struct visl_run *run) in visl_get_ref_frames()
55 vb2_buf = vb2_find_buffer(cap_q, run->fwht.params->backward_ref_ts); in visl_get_ref_frames()
58 run->fwht.params->backward_ref_ts, in visl_get_ref_frames()
67 b_ref = vb2_find_buffer(cap_q, run->mpeg2.pic->backward_ref_ts); in visl_get_ref_frames()
68 f_ref = vb2_find_buffer(cap_q, run->mpeg2.pic->forward_ref_ts); in visl_get_ref_frames()
73 run->mpeg2.pic->backward_ref_ts, in visl_get_ref_frames()
75 run->mpeg2.pic->forward_ref_ts, in visl_get_ref_frames()
85 last = vb2_find_buffer(cap_q, run->vp8.frame->last_frame_ts); in visl_get_ref_frames()
86 golden = vb2_find_buffer(cap_q, run->vp8.frame->golden_frame_ts); in visl_get_ref_frames()
87 alt = vb2_find_buffer(cap_q, run in visl_get_ref_frames()
36 visl_get_ref_frames(struct visl_ctx *ctx, u8 *buf, __kernel_size_t buflen, struct visl_run *run) visl_get_ref_frames() argument
195 visl_tpg_fill_sequence(struct visl_ctx *ctx, struct visl_run *run, char buf[], size_t bufsz) visl_tpg_fill_sequence() argument
215 visl_tpg_fill(struct visl_ctx *ctx, struct visl_run *run) visl_tpg_fill() argument
358 visl_trace_ctrls(struct visl_ctx *ctx, struct visl_run *run) visl_trace_ctrls() argument
420 struct visl_run run = {}; visl_device_run() local
[all...]
/kernel/linux/linux-6.6/tools/testing/selftests/kvm/s390x/
H A Dsync_regs_test.c78 struct kvm_run *run = vcpu->run; in test_read_invalid() local
82 run->kvm_valid_regs = INVALID_SYNC_FIELD; in test_read_invalid()
87 run->kvm_valid_regs = 0; in test_read_invalid()
89 run->kvm_valid_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS; in test_read_invalid()
94 run->kvm_valid_regs = 0; in test_read_invalid()
99 struct kvm_run *run = vcpu->run; in test_set_invalid() local
103 run->kvm_dirty_regs = INVALID_SYNC_FIELD; in test_set_invalid()
108 run in test_set_invalid()
120 struct kvm_run *run = vcpu->run; test_req_and_verify_all_valid_regs() local
146 struct kvm_run *run = vcpu->run; test_set_and_verify_various_reg_values() local
185 struct kvm_run *run = vcpu->run; test_clear_kvm_dirty_regs_bits() local
[all...]
/kernel/linux/linux-6.6/tools/testing/selftests/kvm/x86_64/
H A Dsync_regs_test.c91 struct kvm_run *run = (struct kvm_run *)arg; in race_events_inj_pen() local
92 struct kvm_vcpu_events *events = &run->s.regs.events; in race_events_inj_pen()
97 WRITE_ONCE(run->kvm_dirty_regs, KVM_SYNC_X86_EVENTS); in race_events_inj_pen()
114 struct kvm_run *run = (struct kvm_run *)arg; in race_events_exc() local
115 struct kvm_vcpu_events *events = &run->s.regs.events; in race_events_exc()
118 WRITE_ONCE(run->kvm_dirty_regs, KVM_SYNC_X86_EVENTS); in race_events_exc()
136 struct kvm_run *run = (struct kvm_run *)arg; in race_sregs_cr4() local
137 __u64 *cr4 = &run->s.regs.sregs.cr4; in race_sregs_cr4()
142 WRITE_ONCE(run->kvm_dirty_regs, KVM_SYNC_X86_SREGS); in race_sregs_cr4()
161 struct kvm_run *run; in race_sync_regs() local
215 struct kvm_run *run; main() local
[all...]
H A Ddebug_regs.c83 struct kvm_run *run; in main() local
101 run = vcpu->run; in main()
108 TEST_ASSERT(run->exit_reason == KVM_EXIT_DEBUG && in main()
109 run->debug.arch.exception == BP_VECTOR && in main()
110 run->debug.arch.pc == CAST_TO_RIP(sw_bp), in main()
112 run->exit_reason, run->debug.arch.exception, in main()
113 run->debug.arch.pc, CAST_TO_RIP(sw_bp)); in main()
125 TEST_ASSERT(run in main()
[all...]
H A Duserspace_msr_exit_test.c275 * Now run the same tests with the instruction emulator. in guest_code_filter_allow()
401 if (vcpu->run->exit_reason == KVM_EXIT_IO && in check_for_guest_assert()
409 struct kvm_run *run = vcpu->run; in process_rdmsr() local
414 TEST_ASSERT(run->msr.index == msr_index, in process_rdmsr()
416 run->msr.index, msr_index); in process_rdmsr()
418 switch (run->msr.index) { in process_rdmsr()
420 run->msr.data = 0; in process_rdmsr()
423 run->msr.error = 1; in process_rdmsr()
426 run in process_rdmsr()
441 struct kvm_run *run = vcpu->run; process_wrmsr() local
612 handle_rdmsr(struct kvm_run *run) handle_rdmsr() argument
629 handle_wrmsr(struct kvm_run *run) handle_wrmsr() argument
653 struct kvm_run *run; test_msr_filter_deny() local
[all...]
H A Dxen_vmcall_test.c105 volatile struct kvm_run *run = vcpu->run; in main() local
110 if (run->exit_reason == KVM_EXIT_XEN) { in main()
111 TEST_ASSERT_EQ(run->xen.type, KVM_EXIT_XEN_HCALL); in main()
112 TEST_ASSERT_EQ(run->xen.u.hcall.cpl, 0); in main()
113 TEST_ASSERT_EQ(run->xen.u.hcall.longmode, 1); in main()
114 TEST_ASSERT_EQ(run->xen.u.hcall.input, INPUTVALUE); in main()
115 TEST_ASSERT_EQ(run->xen.u.hcall.params[0], ARGVALUE(1)); in main()
116 TEST_ASSERT_EQ(run->xen.u.hcall.params[1], ARGVALUE(2)); in main()
117 TEST_ASSERT_EQ(run in main()
[all...]
/kernel/linux/linux-5.10/arch/loongarch/kvm/
H A Demulate.c37 vcpu->run->exit_reason = KVM_EXIT_IRQ_WINDOW_OPEN; in _kvm_emu_idle()
50 struct kvm_run *run = vcpu->run; in _kvm_emu_mmio_write() local
53 void *data = run->mmio.data; in _kvm_emu_mmio_write()
65 run->mmio.phys_addr = vcpu->arch.badv; in _kvm_emu_mmio_write()
66 if (run->mmio.phys_addr == KVM_INVALID_ADDR) in _kvm_emu_mmio_write()
76 run->mmio.len = 8; in _kvm_emu_mmio_write()
80 run->mmio.len = 4; in _kvm_emu_mmio_write()
94 run->mmio.len = 8; in _kvm_emu_mmio_write()
98 run in _kvm_emu_mmio_write()
175 struct kvm_run *run = vcpu->run; _kvm_emu_mmio_read() local
295 _kvm_complete_mmio_read(struct kvm_vcpu *vcpu, struct kvm_run *run) _kvm_complete_mmio_read() argument
[all...]
H A Dcsr.c300 int (*get) (struct kvm_run *run, struct kvm_vcpu *vcpu, u32 addr, u64 *res);
301 int (*set) (struct kvm_run *run, struct kvm_vcpu *vcpu, u32 addr, u64 val);
316 static int kvm_iocsr_common_get(struct kvm_run *run, struct kvm_vcpu *vcpu, in kvm_iocsr_common_get() argument
332 static int kvm_iocsr_common_set(struct kvm_run *run, struct kvm_vcpu *vcpu, in kvm_iocsr_common_set() argument
348 static int kvm_misc_set(struct kvm_run *run, struct kvm_vcpu *vcpu, u32 addr, in kvm_misc_set() argument
351 return kvm_iocsr_common_set(run, vcpu, addr, val); in kvm_misc_set()
354 static int kvm_ipi_get(struct kvm_run *run, struct kvm_vcpu *vcpu, u32 addr, in kvm_ipi_get() argument
360 run->mmio.phys_addr = KVM_IPI_REG_ADDRESS(vcpu->vcpu_id, (addr & 0xff)); in kvm_ipi_get()
361 ret = kvm_io_bus_read(vcpu, KVM_MMIO_BUS, run->mmio.phys_addr, in kvm_ipi_get()
362 run in kvm_ipi_get()
372 kvm_extioi_isr_get(struct kvm_run *run, struct kvm_vcpu *vcpu, u32 addr, u64 *res) kvm_extioi_isr_get() argument
390 kvm_ipi_set(struct kvm_run *run, struct kvm_vcpu *vcpu, u32 addr, u64 val) kvm_ipi_set() argument
408 kvm_extioi_set(struct kvm_run *run, struct kvm_vcpu *vcpu, u32 addr, u64 val) kvm_extioi_set() argument
432 kvm_nop_set(struct kvm_run *run, struct kvm_vcpu *vcpu, u32 addr, u64 val) kvm_nop_set() argument
469 _kvm_emu_iocsr_read(struct kvm_run *run, struct kvm_vcpu *vcpu, u32 addr, u64 *res) _kvm_emu_iocsr_read() argument
496 _kvm_emu_iocsr_write(struct kvm_run *run, struct kvm_vcpu *vcpu, u32 addr, u64 val) _kvm_emu_iocsr_write() argument
524 _kvm_emu_iocsr(larch_inst inst, struct kvm_run *run, struct kvm_vcpu *vcpu) _kvm_emu_iocsr() argument
590 _kvm_complete_iocsr_read(struct kvm_vcpu *vcpu, struct kvm_run *run) _kvm_complete_iocsr_read() argument
[all...]
H A Dexit.c48 vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR; in _kvm_fault_ni()
97 struct kvm_run *run = vcpu->run; in _kvm_trap_handle_gspr() local
149 er = _kvm_emu_iocsr(inst, run, vcpu); in _kvm_trap_handle_gspr()
237 vcpu->run->exit_reason = KVM_EXIT_MMIO; in _kvm_handle_gspr()
240 vcpu->run->exit_reason = KVM_EXIT_LOONGARCH_IOCSR; in _kvm_handle_gspr()
244 vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR; in _kvm_handle_gspr()
261 vcpu->run->exit_reason = KVM_EXIT_DEBUG; in _kvm_handle_hypcall()
278 vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR; in _kvm_handle_gcm()
294 struct kvm_run *run in _kvm_handle_fpu_disabled() local
321 struct kvm_run *run = vcpu->run; _kvm_handle_lsx_disabled() local
356 struct kvm_run *run = vcpu->run; _kvm_handle_lasx_disabled() local
388 struct kvm_run *run = vcpu->run; _kvm_handle_lbt_disabled() local
407 struct kvm_run *run = vcpu->run; _kvm_handle_read_fault() local
445 struct kvm_run *run = vcpu->run; _kvm_handle_write_fault() local
[all...]
/kernel/linux/linux-5.10/tools/testing/selftests/kvm/x86_64/
H A Ddebug_regs.c69 struct kvm_run *run; in main() local
89 run = vcpu_state(vm, VCPU_ID); in main()
96 TEST_ASSERT(run->exit_reason == KVM_EXIT_DEBUG && in main()
97 run->debug.arch.exception == BP_VECTOR && in main()
98 run->debug.arch.pc == CAST_TO_RIP(sw_bp), in main()
100 run->exit_reason, run->debug.arch.exception, in main()
101 run->debug.arch.pc, CAST_TO_RIP(sw_bp)); in main()
113 TEST_ASSERT(run->exit_reason == KVM_EXIT_DEBUG && in main()
114 run in main()
[all...]
H A Dsync_regs_test.c83 struct kvm_run *run; in main() local
105 run = vcpu_state(vm, VCPU_ID); in main()
108 run->kvm_valid_regs = INVALID_SYNC_FIELD; in main()
115 run->kvm_valid_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS; in main()
123 run->kvm_dirty_regs = INVALID_SYNC_FIELD; in main()
130 run->kvm_dirty_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS; in main()
139 run->kvm_valid_regs = TEST_SYNC_FIELDS; in main()
141 TEST_ASSERT(run->exit_reason == KVM_EXIT_IO, in main()
143 run->exit_reason, in main()
144 exit_reason_str(run in main()
[all...]
H A Duser_msr_test.c154 static void handle_rdmsr(struct kvm_run *run) in handle_rdmsr() argument
156 run->msr.data = run->msr.index; in handle_rdmsr()
159 if (run->msr.index == MSR_SYSCALL_MASK || in handle_rdmsr()
160 run->msr.index == MSR_GS_BASE) { in handle_rdmsr()
161 TEST_ASSERT(run->msr.reason == KVM_MSR_EXIT_REASON_FILTER, in handle_rdmsr()
165 if (run->msr.index == 0xdeadbeef) { in handle_rdmsr()
166 TEST_ASSERT(run->msr.reason == KVM_MSR_EXIT_REASON_UNKNOWN, in handle_rdmsr()
171 static void handle_wrmsr(struct kvm_run *run) in handle_wrmsr() argument
176 if (run in handle_wrmsr()
200 struct kvm_run *run; main() local
[all...]
/kernel/linux/linux-5.10/tools/testing/selftests/kvm/s390x/
H A Dsync_regs_test.c79 struct kvm_run *run; in main() local
96 run = vcpu_state(vm, VCPU_ID); in main()
99 run->kvm_valid_regs = INVALID_SYNC_FIELD; in main()
106 run->kvm_valid_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS; in main()
114 run->kvm_dirty_regs = INVALID_SYNC_FIELD; in main()
121 run->kvm_dirty_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS; in main()
129 run->kvm_valid_regs = TEST_SYNC_FIELDS; in main()
132 TEST_ASSERT(run->exit_reason == KVM_EXIT_S390_SIEIC, in main()
134 run->exit_reason, in main()
135 exit_reason_str(run in main()
[all...]
/kernel/linux/linux-6.6/fs/ntfs3/
H A Drun.c34 static bool run_lookup(const struct runs_tree *run, CLST vcn, size_t *index) in run_lookup() argument
39 if (!run->count) { in run_lookup()
45 max_idx = run->count - 1; in run_lookup()
48 r = run->runs; in run_lookup()
61 *index = run->count; in run_lookup()
72 r = run->runs + mid_idx; in run_lookup()
93 static void run_consolidate(struct runs_tree *run, size_t index) in run_consolidate() argument
96 struct ntfs_run *r = run->runs + index; in run_consolidate()
98 while (index + 1 < run->count) { in run_consolidate()
100 * I should merge current run wit in run_consolidate()
167 run_is_mapped_full(const struct runs_tree *run, CLST svcn, CLST evcn) run_is_mapped_full() argument
192 run_lookup_entry(const struct runs_tree *run, CLST vcn, CLST *lcn, CLST *len, size_t *index) run_lookup_entry() argument
228 run_truncate_head(struct runs_tree *run, CLST vcn) run_truncate_head() argument
263 run_truncate(struct runs_tree *run, CLST vcn) run_truncate() argument
300 run_truncate_around(struct runs_tree *run, CLST vcn) run_truncate_around() argument
316 run_add_entry(struct runs_tree *run, CLST vcn, CLST lcn, CLST len, bool is_mft) run_add_entry() argument
489 run_collapse_range(struct runs_tree *run, CLST vcn, CLST len) run_collapse_range() argument
555 run_insert_range(struct runs_tree *run, CLST vcn, CLST len) run_insert_range() argument
595 run_get_entry(const struct runs_tree *run, size_t index, CLST *vcn, CLST *lcn, CLST *len) run_get_entry() argument
817 run_pack(const struct runs_tree *run, CLST svcn, CLST len, u8 *run_buf, u32 run_buf_size, CLST *packed_vcns) run_pack() argument
920 run_unpack(struct runs_tree *run, struct ntfs_sb_info *sbi, CLST ino, CLST svcn, CLST evcn, CLST vcn, const u8 *run_buf, int run_buf_size) run_unpack() argument
1050 run_unpack_ex(struct runs_tree *run, struct ntfs_sb_info *sbi, CLST ino, CLST svcn, CLST evcn, CLST vcn, const u8 *run_buf, int run_buf_size) run_unpack_ex() argument
1151 run_clone(const struct runs_tree *run, struct runs_tree *new_run) run_clone() argument
[all...]
/kernel/linux/linux-6.6/drivers/staging/media/sunxi/cedrus/
H A Dcedrus_dec.c29 struct cedrus_run run = {}; in cedrus_device_run() local
33 run.src = v4l2_m2m_next_src_buf(ctx->fh.m2m_ctx); in cedrus_device_run()
34 run.dst = v4l2_m2m_next_dst_buf(ctx->fh.m2m_ctx); in cedrus_device_run()
37 src_req = run.src->vb2_buf.req_obj.req; in cedrus_device_run()
44 run.mpeg2.sequence = cedrus_find_control_data(ctx, in cedrus_device_run()
46 run.mpeg2.picture = cedrus_find_control_data(ctx, in cedrus_device_run()
48 run.mpeg2.quantisation = cedrus_find_control_data(ctx, in cedrus_device_run()
53 run.h264.decode_params = cedrus_find_control_data(ctx, in cedrus_device_run()
55 run.h264.pps = cedrus_find_control_data(ctx, in cedrus_device_run()
57 run in cedrus_device_run()
[all...]
/kernel/linux/linux-5.10/drivers/staging/media/sunxi/cedrus/
H A Dcedrus_dec.c29 struct cedrus_run run = {}; in cedrus_device_run() local
32 run.src = v4l2_m2m_next_src_buf(ctx->fh.m2m_ctx); in cedrus_device_run()
33 run.dst = v4l2_m2m_next_dst_buf(ctx->fh.m2m_ctx); in cedrus_device_run()
36 src_req = run.src->vb2_buf.req_obj.req; in cedrus_device_run()
43 run.mpeg2.slice_params = cedrus_find_control_data(ctx, in cedrus_device_run()
45 run.mpeg2.quantization = cedrus_find_control_data(ctx, in cedrus_device_run()
50 run.h264.decode_params = cedrus_find_control_data(ctx, in cedrus_device_run()
52 run.h264.pps = cedrus_find_control_data(ctx, in cedrus_device_run()
54 run.h264.scaling_matrix = cedrus_find_control_data(ctx, in cedrus_device_run()
56 run in cedrus_device_run()
[all...]
/kernel/linux/linux-5.10/tools/testing/selftests/arm64/mte/
H A Dcheck_mmap_options.c63 int run, result, map_size; in check_anonymous_memory_mapping() local
68 for (run = 0; run < item; run++) { in check_anonymous_memory_mapping()
69 map_size = sizes[run] + OVERFLOW + UNDERFLOW; in check_anonymous_memory_mapping()
75 mte_initialize_current_context(mode, (uintptr_t)ptr, sizes[run]); in check_anonymous_memory_mapping()
77 ptr = mte_insert_tags((void *)ptr, sizes[run]); in check_anonymous_memory_mapping()
83 result = check_mte_memory(ptr, sizes[run], mode, tag_check); in check_anonymous_memory_mapping()
84 mte_clear_tags((void *)ptr, sizes[run]); in check_anonymous_memory_mapping()
95 int run, f in check_file_memory_mapping() local
134 int run, prot_flag, result, fd, map_size; check_clear_prot_mte_flag() local
[all...]
/kernel/linux/linux-6.6/tools/testing/selftests/arm64/mte/
H A Dcheck_mmap_options.c63 int run, result, map_size; in check_anonymous_memory_mapping() local
67 for (run = 0; run < item; run++) { in check_anonymous_memory_mapping()
68 map_size = sizes[run] + OVERFLOW + UNDERFLOW; in check_anonymous_memory_mapping()
74 mte_initialize_current_context(mode, (uintptr_t)ptr, sizes[run]); in check_anonymous_memory_mapping()
76 ptr = mte_insert_tags((void *)ptr, sizes[run]); in check_anonymous_memory_mapping()
82 result = check_mte_memory(ptr, sizes[run], mode, tag_check); in check_anonymous_memory_mapping()
83 mte_clear_tags((void *)ptr, sizes[run]); in check_anonymous_memory_mapping()
94 int run, f in check_file_memory_mapping() local
133 int run, prot_flag, result, fd, map_size; check_clear_prot_mte_flag() local
[all...]
/kernel/linux/linux-5.10/arch/s390/kvm/
H A Ddiag.c25 start = vcpu->run->s.regs.gprs[(vcpu->arch.sie_block->ipa & 0xf0) >> 4]; in diag_release_pages()
26 end = vcpu->run->s.regs.gprs[vcpu->arch.sie_block->ipa & 0xf] + PAGE_SIZE; in diag_release_pages()
76 vcpu->run->s.regs.gprs[rx]); in __diag_page_ref_service()
78 if (vcpu->run->s.regs.gprs[rx] & 7) in __diag_page_ref_service()
80 rc = read_guest(vcpu, vcpu->run->s.regs.gprs[rx], rx, &parm, sizeof(parm)); in __diag_page_ref_service()
97 vcpu->run->s.regs.gprs[ry] = 8; in __diag_page_ref_service()
111 vcpu->run->s.regs.gprs[ry] = 0; in __diag_page_ref_service()
125 vcpu->run->s.regs.gprs[ry] = 0; in __diag_page_ref_service()
131 vcpu->run->s.regs.gprs[ry] = 4; in __diag_page_ref_service()
158 tid = vcpu->run in __diag_time_slice_end_directed()
[all...]
/kernel/linux/linux-6.6/arch/s390/kvm/
H A Ddiag.c25 start = vcpu->run->s.regs.gprs[(vcpu->arch.sie_block->ipa & 0xf0) >> 4]; in diag_release_pages()
26 end = vcpu->run->s.regs.gprs[vcpu->arch.sie_block->ipa & 0xf] + PAGE_SIZE; in diag_release_pages()
76 vcpu->run->s.regs.gprs[rx]); in __diag_page_ref_service()
78 if (vcpu->run->s.regs.gprs[rx] & 7) in __diag_page_ref_service()
80 rc = read_guest(vcpu, vcpu->run->s.regs.gprs[rx], rx, &parm, sizeof(parm)); in __diag_page_ref_service()
97 vcpu->run->s.regs.gprs[ry] = 8; in __diag_page_ref_service()
111 vcpu->run->s.regs.gprs[ry] = 0; in __diag_page_ref_service()
125 vcpu->run->s.regs.gprs[ry] = 0; in __diag_page_ref_service()
131 vcpu->run->s.regs.gprs[ry] = 4; in __diag_page_ref_service()
172 tid = vcpu->run in __diag_time_slice_end_directed()
[all...]
/kernel/linux/linux-6.6/arch/mips/kvm/
H A Demulate.c962 vcpu->run->exit_reason = KVM_EXIT_IRQ_WINDOW_OPEN; in kvm_mips_emul_wait()
975 struct kvm_run *run = vcpu->run; in kvm_mips_emulate_store() local
976 void *data = run->mmio.data; in kvm_mips_emulate_store()
991 run->mmio.phys_addr = kvm_mips_callbacks->gva_to_gpa( in kvm_mips_emulate_store()
993 if (run->mmio.phys_addr == KVM_INVALID_ADDR) in kvm_mips_emulate_store()
999 run->mmio.len = 8; in kvm_mips_emulate_store()
1009 run->mmio.len = 4; in kvm_mips_emulate_store()
1018 run->mmio.len = 2; in kvm_mips_emulate_store()
1027 run in kvm_mips_emulate_store()
1271 struct kvm_run *run = vcpu->run; kvm_mips_emulate_load() local
1511 struct kvm_run *run = vcpu->run; kvm_mips_complete_mmio_load() local
[all...]
/kernel/linux/linux-6.6/arch/riscv/kvm/
H A Dvcpu_insn.c148 int (*func)(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn);
151 static int truly_illegal_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, in truly_illegal_insn() argument
167 static int truly_virtual_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, in truly_virtual_insn() argument
197 static int wfi_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn) in wfi_insn() argument
226 * @run: The VCPU run struct containing the CSR data
230 int kvm_riscv_vcpu_csr_return(struct kvm_vcpu *vcpu, struct kvm_run *run) in kvm_riscv_vcpu_csr_return() argument
242 run->riscv_csr.ret_value); in kvm_riscv_vcpu_csr_return()
250 static int csr_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn) in csr_insn() argument
294 run in csr_insn()
370 system_opcode_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn) system_opcode_insn() argument
410 kvm_riscv_vcpu_virtual_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, struct kvm_cpu_trap *trap) kvm_riscv_vcpu_virtual_insn() argument
453 kvm_riscv_vcpu_mmio_load(struct kvm_vcpu *vcpu, struct kvm_run *run, unsigned long fault_addr, unsigned long htinst) kvm_riscv_vcpu_mmio_load() argument
575 kvm_riscv_vcpu_mmio_store(struct kvm_vcpu *vcpu, struct kvm_run *run, unsigned long fault_addr, unsigned long htinst) kvm_riscv_vcpu_mmio_store() argument
703 kvm_riscv_vcpu_mmio_return(struct kvm_vcpu *vcpu, struct kvm_run *run) kvm_riscv_vcpu_mmio_return() argument
[all...]
/kernel/linux/linux-5.10/tools/perf/scripts/python/
H A Dstat-cpi.py23 def store(time, event, cpu, thread, val, ena, run):
24 #print("event %s cpu %d, thread %d, time %d, val %d, ena %d, run %d" %
25 # (event, cpu, thread, time, val, ena, run))
29 data[key] = [ val, ena, run]
35 def stat__cycles_k(cpu, thread, time, val, ena, run):
36 store(time, "cycles", cpu, thread, val, ena, run);
38 def stat__instructions_k(cpu, thread, time, val, ena, run):
39 store(time, "instructions", cpu, thread, val, ena, run);
41 def stat__cycles_u(cpu, thread, time, val, ena, run):
42 store(time, "cycles", cpu, thread, val, ena, run);
[all...]
/kernel/linux/linux-6.6/tools/perf/scripts/python/
H A Dstat-cpi.py23 def store(time, event, cpu, thread, val, ena, run):
24 #print("event %s cpu %d, thread %d, time %d, val %d, ena %d, run %d" %
25 # (event, cpu, thread, time, val, ena, run))
29 data[key] = [ val, ena, run]
35 def stat__cycles_k(cpu, thread, time, val, ena, run):
36 store(time, "cycles", cpu, thread, val, ena, run);
38 def stat__instructions_k(cpu, thread, time, val, ena, run):
39 store(time, "instructions", cpu, thread, val, ena, run);
41 def stat__cycles_u(cpu, thread, time, val, ena, run):
42 store(time, "cycles", cpu, thread, val, ena, run);
[all...]

Completed in 12 milliseconds

12345678910>>...28