Lines Matching defs:new_op
22 struct orangefs_kernel_op_s *new_op;
30 new_op = op_alloc(ORANGEFS_VFS_OP_RA_FLUSH);
31 if (!new_op)
33 new_op->upcall.req.ra_cache_flush.refn = orangefs_inode->refn;
35 ret = service_operation(new_op, "orangefs_flush_racache",
41 op_release(new_op);
55 struct orangefs_kernel_op_s *new_op = NULL;
62 new_op = op_alloc(ORANGEFS_VFS_OP_FILE_IO);
63 if (!new_op)
67 new_op->upcall.req.io.readahead_size = readahead_size;
68 new_op->upcall.req.io.io_type = type;
69 new_op->upcall.req.io.refn = orangefs_inode->refn;
85 new_op,
88 new_op->uses_shared_memory = 1;
89 new_op->upcall.req.io.buf_index = buffer_index;
90 new_op->upcall.req.io.count = total_size;
91 new_op->upcall.req.io.offset = *offset;
93 new_op->upcall.uid = from_kuid(&init_user_ns, wr->uid);
94 new_op->upcall.gid = from_kgid(&init_user_ns, wr->gid);
125 new_op->upcall.uid = 0;
127 new_op->upcall.uid = 0;
152 llu(new_op->tag));
155 ret = service_operation(new_op,
170 if (ret == -EAGAIN && op_state_purged(new_op)) {
193 switch (new_op->op_state - OP_VFS_STATE_GIVEN_UP) {
219 new_op->op_state);
226 new_op->op_state,
227 new_op);
235 if (orangefs_cancel_op_in_progress(new_op))
244 if (type == ORANGEFS_IO_READ && new_op->downcall.resp.io.amt_complete) {
262 min(new_op->downcall.resp.io.amt_complete,
265 copy_amount = new_op->downcall.resp.io.amt_complete;
280 (int)new_op->downcall.resp.io.amt_complete);
282 ret = new_op->downcall.resp.io.amt_complete;
300 op_release(new_op);
550 struct orangefs_kernel_op_s *new_op = NULL;
557 new_op = op_alloc(ORANGEFS_VFS_OP_FSYNC);
558 if (!new_op)
560 new_op->upcall.req.fsync.refn = orangefs_inode->refn;
562 ret = service_operation(new_op,
570 op_release(new_op);