Lines Matching refs:rqstp

31 static int svc_deferred_recv(struct svc_rqst *rqstp);
364 * Copy the local and remote xprt addresses to the rqstp structure
366 void svc_xprt_copy_addrs(struct svc_rqst *rqstp, struct svc_xprt *xprt)
368 memcpy(&rqstp->rq_addr, &xprt->xpt_remote, xprt->xpt_remotelen);
369 rqstp->rq_addrlen = xprt->xpt_remotelen;
375 memcpy(&rqstp->rq_daddr, &xprt->xpt_local, xprt->xpt_locallen);
376 rqstp->rq_daddrlen = xprt->xpt_locallen;
382 * @rqstp: svc_rqst struct containing address to print
387 char *svc_print_addr(struct svc_rqst *rqstp, char *buf, size_t len)
389 return __svc_print_addr(svc_addr(rqstp), buf, len);
401 static bool svc_xprt_reserve_slot(struct svc_rqst *rqstp, struct svc_xprt *xprt)
403 if (!test_bit(RQ_DATA, &rqstp->rq_flags)) {
407 set_bit(RQ_DATA, &rqstp->rq_flags);
412 static void svc_xprt_release_slot(struct svc_rqst *rqstp)
414 struct svc_xprt *xprt = rqstp->rq_xprt;
415 if (test_and_clear_bit(RQ_DATA, &rqstp->rq_flags)) {
507 * @rqstp: The request in question
515 void svc_reserve(struct svc_rqst *rqstp, int space)
517 struct svc_xprt *xprt = rqstp->rq_xprt;
519 space += rqstp->rq_res.head[0].iov_len;
521 if (xprt && space < rqstp->rq_reserved) {
522 atomic_sub((rqstp->rq_reserved - space), &xprt->xpt_reserved);
523 rqstp->rq_reserved = space;
539 static void svc_xprt_release(struct svc_rqst *rqstp)
541 struct svc_xprt *xprt = rqstp->rq_xprt;
543 xprt->xpt_ops->xpo_release_ctxt(xprt, rqstp->rq_xprt_ctxt);
544 rqstp->rq_xprt_ctxt = NULL;
546 free_deferred(xprt, rqstp->rq_deferred);
547 rqstp->rq_deferred = NULL;
549 svc_rqst_release_pages(rqstp);
550 rqstp->rq_res.page_len = 0;
551 rqstp->rq_res.page_base = 0;
558 if ((rqstp->rq_res.len) > rqstp->rq_reserved)
560 rqstp->rq_reserved,
561 rqstp->rq_res.len);
563 rqstp->rq_res.head[0].iov_len = 0;
564 svc_reserve(rqstp, 0);
565 svc_xprt_release_slot(rqstp);
566 rqstp->rq_xprt = NULL;
654 static bool svc_alloc_arg(struct svc_rqst *rqstp)
656 struct svc_serv *serv = rqstp->rq_server;
657 struct xdr_buf *arg = &rqstp->rq_arg;
670 rqstp->rq_pages);
683 rqstp->rq_page_end = &rqstp->rq_pages[pages];
684 rqstp->rq_pages[pages] = NULL; /* this might be seen in nfsd_splice_actor() */
687 arg->head[0].iov_base = page_address(rqstp->rq_pages[0]);
689 arg->pages = rqstp->rq_pages + 1;
696 rqstp->rq_xid = xdr_zero;
701 rqst_should_sleep(struct svc_rqst *rqstp)
703 struct svc_pool *pool = rqstp->rq_pool;
724 static struct svc_xprt *svc_get_next_xprt(struct svc_rqst *rqstp)
726 struct svc_pool *pool = rqstp->rq_pool;
729 WARN_ON_ONCE(rqstp->rq_xprt);
731 rqstp->rq_xprt = svc_xprt_dequeue(pool);
732 if (rqstp->rq_xprt)
738 clear_bit(RQ_BUSY, &rqstp->rq_flags);
741 if (likely(rqst_should_sleep(rqstp)))
748 set_bit(RQ_BUSY, &rqstp->rq_flags);
751 rqstp->rq_xprt = svc_xprt_dequeue(pool);
752 if (rqstp->rq_xprt)
764 rqstp->rq_chandle.thread_wait = 5*HZ;
766 rqstp->rq_chandle.thread_wait = 1*HZ;
767 trace_svc_xprt_dequeue(rqstp);
768 return rqstp->rq_xprt;
787 static int svc_handle_xprt(struct svc_rqst *rqstp, struct svc_xprt *xprt)
789 struct svc_serv *serv = rqstp->rq_server;
819 } else if (svc_xprt_reserve_slot(rqstp, xprt)) {
821 rqstp->rq_deferred = svc_deferred_dequeue(xprt);
822 if (rqstp->rq_deferred)
823 len = svc_deferred_recv(rqstp);
825 len = xprt->xpt_ops->xpo_recvfrom(rqstp);
826 rqstp->rq_reserved = serv->sv_max_mesg;
827 atomic_add(rqstp->rq_reserved, &xprt->xpt_reserved);
837 * @rqstp: an idle RPC service thread
843 void svc_recv(struct svc_rqst *rqstp)
846 struct svc_serv *serv = rqstp->rq_server;
849 if (!svc_alloc_arg(rqstp))
857 xprt = svc_get_next_xprt(rqstp);
861 len = svc_handle_xprt(rqstp, xprt);
867 trace_svc_xdr_recvfrom(&rqstp->rq_arg);
871 rqstp->rq_chandle.defer = svc_defer;
875 percpu_counter_inc(&rqstp->rq_pool->sp_messages_arrived);
876 rqstp->rq_stime = ktime_get();
877 svc_process(rqstp);
881 rqstp->rq_res.len = 0;
882 svc_xprt_release(rqstp);
889 void svc_drop(struct svc_rqst *rqstp)
891 trace_svc_drop(rqstp);
892 svc_xprt_release(rqstp);
898 * @rqstp: RPC transaction context
901 void svc_send(struct svc_rqst *rqstp)
907 xprt = rqstp->rq_xprt;
912 xb = &rqstp->rq_res;
916 trace_svc_xdr_sendto(rqstp->rq_xid, xb);
917 trace_svc_stats_latency(rqstp);
919 status = xprt->xpt_ops->xpo_sendto(rqstp);
921 trace_svc_send(rqstp, status);
922 svc_xprt_release(rqstp);
1184 struct svc_rqst *rqstp = container_of(req, struct svc_rqst, rq_chandle);
1187 if (rqstp->rq_arg.page_len || !test_bit(RQ_USEDEFERRAL, &rqstp->rq_flags))
1189 if (rqstp->rq_deferred) {
1190 dr = rqstp->rq_deferred;
1191 rqstp->rq_deferred = NULL;
1196 size = sizeof(struct svc_deferred_req) + rqstp->rq_arg.len;
1201 dr->handle.owner = rqstp->rq_server;
1202 dr->prot = rqstp->rq_prot;
1203 memcpy(&dr->addr, &rqstp->rq_addr, rqstp->rq_addrlen);
1204 dr->addrlen = rqstp->rq_addrlen;
1205 dr->daddr = rqstp->rq_daddr;
1206 dr->argslen = rqstp->rq_arg.len >> 2;
1209 skip = rqstp->rq_arg.len - rqstp->rq_arg.head[0].iov_len;
1210 memcpy(dr->args, rqstp->rq_arg.head[0].iov_base - skip,
1213 dr->xprt_ctxt = rqstp->rq_xprt_ctxt;
1214 rqstp->rq_xprt_ctxt = NULL;
1215 trace_svc_defer(rqstp);
1216 svc_xprt_get(rqstp->rq_xprt);
1217 dr->xprt = rqstp->rq_xprt;
1218 set_bit(RQ_DROPME, &rqstp->rq_flags);
1227 static noinline int svc_deferred_recv(struct svc_rqst *rqstp)
1229 struct svc_deferred_req *dr = rqstp->rq_deferred;
1234 rqstp->rq_arg.head[0].iov_base = dr->args;
1236 rqstp->rq_arg.head[0].iov_len = dr->argslen << 2;
1237 rqstp->rq_arg.page_len = 0;
1239 rqstp->rq_arg.len = dr->argslen << 2;
1240 rqstp->rq_prot = dr->prot;
1241 memcpy(&rqstp->rq_addr, &dr->addr, dr->addrlen);
1242 rqstp->rq_addrlen = dr->addrlen;
1244 rqstp->rq_daddr = dr->daddr;
1245 rqstp->rq_respages = rqstp->rq_pages;
1246 rqstp->rq_xprt_ctxt = dr->xprt_ctxt;
1249 svc_xprt_received(rqstp->rq_xprt);