Lines Matching refs:from

34  * and hide all the ugliness from the user.
227 * This does an atomic 128 byte aligned load from userspace.
365 const void __user *from, unsigned long size);
369 copy_mc_generic(void *to, const void *from, unsigned long size);
372 copy_mc_to_kernel(void *to, const void *from, unsigned long size)
374 return copy_mc_generic(to, from, size);
379 copy_mc_to_user(void __user *to, const void *from, unsigned long n)
381 if (likely(check_copy_size(from, n, true))) {
384 n = copy_mc_generic((void *)to, from, n);
395 raw_copy_in_user(void __user *to, const void __user *from, unsigned long n)
400 allow_read_write_user(to, from, n);
401 ret = __copy_tofrom_user(to, from, n);
402 prevent_read_write_user(to, from, n);
408 const void __user *from, unsigned long n)
417 __get_user_size(*(u8 *)to, from, 1, ret);
421 __get_user_size(*(u16 *)to, from, 2, ret);
425 __get_user_size(*(u32 *)to, from, 4, ret);
429 __get_user_size(*(u64 *)to, from, 8, ret);
437 allow_read_from_user(from, n);
438 ret = __copy_tofrom_user((__force void __user *)to, from, n);
439 prevent_read_from_user(from, n);
444 raw_copy_to_user_allowed(void __user *to, const void *from, unsigned long n)
451 __put_user_size_allowed(*(u8 *)from, (u8 __user *)to, 1, ret);
454 __put_user_size_allowed(*(u16 *)from, (u16 __user *)to, 2, ret);
457 __put_user_size_allowed(*(u32 *)from, (u32 __user *)to, 4, ret);
460 __put_user_size_allowed(*(u64 *)from, (u64 __user *)to, 8, ret);
467 return __copy_tofrom_user(to, (__force const void __user *)from, n);
471 raw_copy_to_user(void __user *to, const void *from, unsigned long n)
476 ret = raw_copy_to_user_allowed(to, from, n);