Lines Matching refs:in

8  * Use GFP_KERNEL for messages as in 1.2
54 unsigned long q_qnum; /* number of messages in queue */
67 * Similar to the optimization used in ipc/mqueue.c, one syscall return path
68 * does not acquire any locks when it sees that a message exists in
307 static inline unsigned long copy_msqid_to_user(void __user *buf, struct msqid64_ds *in, int version)
311 return copy_to_user(buf, in, sizeof(*in));
317 ipc64_perm_to_ipc_perm(&in->msg_perm, &out.msg_perm);
319 out.msg_stime = in->msg_stime;
320 out.msg_rtime = in->msg_rtime;
321 out.msg_ctime = in->msg_ctime;
323 if (in->msg_cbytes > USHRT_MAX) {
326 out.msg_cbytes = in->msg_cbytes;
328 out.msg_lcbytes = in->msg_cbytes;
330 if (in->msg_qnum > USHRT_MAX) {
333 out.msg_qnum = in->msg_qnum;
336 if (in->msg_qbytes > USHRT_MAX) {
339 out.msg_qbytes = in->msg_qbytes;
341 out.msg_lqbytes = in->msg_qbytes;
343 out.msg_lspid = in->msg_lspid;
344 out.msg_lrpid = in->msg_lrpid;
387 * to be held in write mode.
562 * As defined in SUS:
688 static int copy_compat_msqid_to_user(void __user *buf, struct msqid64_ds *in, int version)
693 to_compat_ipc64_perm(&v.msg_perm, &in->msg_perm);
694 v.msg_stime = lower_32_bits(in->msg_stime);
695 v.msg_stime_high = upper_32_bits(in->msg_stime);
696 v.msg_rtime = lower_32_bits(in->msg_rtime);
697 v.msg_rtime_high = upper_32_bits(in->msg_rtime);
698 v.msg_ctime = lower_32_bits(in->msg_ctime);
699 v.msg_ctime_high = upper_32_bits(in->msg_ctime);
700 v.msg_cbytes = in->msg_cbytes;
701 v.msg_qnum = in->msg_qnum;
702 v.msg_qbytes = in->msg_qbytes;
703 v.msg_lspid = in->msg_lspid;
704 v.msg_lrpid = in->msg_lrpid;
709 to_compat_ipc_perm(&v.msg_perm, &in->msg_perm);
710 v.msg_stime = in->msg_stime;
711 v.msg_rtime = in->msg_rtime;
712 v.msg_ctime = in->msg_ctime;
713 v.msg_cbytes = in->msg_cbytes;
714 v.msg_qnum = in->msg_qnum;
715 v.msg_qbytes = in->msg_qbytes;
716 v.msg_lspid = in->msg_lspid;
717 v.msg_lrpid = in->msg_lrpid;
1227 * The work in pipelined_send() and expunge_all():