Lines Matching defs:data

2440 				 void *data, int offset, int len)
2448 r = __copy_from_user(data, (void __user *)addr + offset, len);
2454 int kvm_read_guest_page(struct kvm *kvm, gfn_t gfn, void *data, int offset,
2459 return __kvm_read_guest_page(slot, gfn, data, offset, len);
2463 int kvm_vcpu_read_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, void *data,
2468 return __kvm_read_guest_page(slot, gfn, data, offset, len);
2472 int kvm_read_guest(struct kvm *kvm, gpa_t gpa, void *data, unsigned long len)
2480 ret = kvm_read_guest_page(kvm, gfn, data, offset, seg);
2485 data += seg;
2492 int kvm_vcpu_read_guest(struct kvm_vcpu *vcpu, gpa_t gpa, void *data, unsigned long len)
2500 ret = kvm_vcpu_read_guest_page(vcpu, gfn, data, offset, seg);
2505 data += seg;
2513 void *data, int offset, unsigned long len)
2522 r = __copy_from_user_inatomic(data, (void __user *)addr + offset, len);
2530 void *data, unsigned long len)
2536 return __kvm_read_guest_atomic(slot, gfn, data, offset, len);
2541 const void *data, int offset, int len)
2549 r = __copy_to_user((void __user *)addr + offset, data, len);
2557 const void *data, int offset, int len)
2561 return __kvm_write_guest_page(slot, gfn, data, offset, len);
2566 const void *data, int offset, int len)
2570 return __kvm_write_guest_page(slot, gfn, data, offset, len);
2574 int kvm_write_guest(struct kvm *kvm, gpa_t gpa, const void *data,
2583 ret = kvm_write_guest_page(kvm, gfn, data, offset, seg);
2588 data += seg;
2595 int kvm_vcpu_write_guest(struct kvm_vcpu *vcpu, gpa_t gpa, const void *data,
2604 ret = kvm_vcpu_write_guest_page(vcpu, gfn, data, offset, seg);
2609 data += seg;
2666 void *data, unsigned int offset,
2685 return kvm_write_guest(kvm, gpa, data, len);
2687 r = __copy_to_user((void __user *)ghc->hva + offset, data, len);
2697 void *data, unsigned long len)
2699 return kvm_write_guest_offset_cached(kvm, ghc, data, 0, len);
2704 void *data, unsigned int offset,
2723 return kvm_read_guest(kvm, gpa, data, len);
2725 r = __copy_from_user(data, (void __user *)ghc->hva + offset, len);
2734 void *data, unsigned long len)
2736 return kvm_read_guest_offset_cached(kvm, ghc, data, 0, len);
3045 * Since algorithm is based on heuristics, accessing another VCPU data without
3889 struct kvm_irqfd data;
3892 if (copy_from_user(&data, argp, sizeof(data)))
3894 r = kvm_irqfd(kvm, &data);
3898 struct kvm_ioeventfd data;
3901 if (copy_from_user(&data, argp, sizeof(data)))
3903 r = kvm_ioeventfd(kvm, &data);
4156 r += PAGE_SIZE; /* pio data page */
4657 static int kvm_stat_data_get(void *data, u64 *val)
4660 struct kvm_stat_data *stat_data = (struct kvm_stat_data *)data;
4676 static int kvm_stat_data_clear(void *data, u64 val)
4679 struct kvm_stat_data *stat_data = (struct kvm_stat_data *)data;
4941 static void check_processor_compat(void *data)
4943 struct kvm_cpu_compat_check *c = data;
5103 uintptr_t data;
5116 uintptr_t data = init_context->data;
5146 err = thread_fn(kvm, data);
5152 uintptr_t data, const char *name,
5162 init_context.data = data;