Lines Matching refs:nreq
896 struct request_sock *nreq;
898 nreq = kmem_cache_alloc(req->rsk_ops->slab, GFP_ATOMIC | __GFP_NOWARN);
899 if (!nreq) {
908 nreq_sk = req_to_sk(nreq);
922 nreq->rsk_listener = sk;
927 if (sk->sk_protocol == IPPROTO_TCP && tcp_rsk(nreq)->tfo_listener)
928 rcu_assign_pointer(tcp_sk(nreq->sk)->fastopen_rsk, nreq);
930 return nreq;
993 struct request_sock *nreq = NULL, *oreq = req;
1007 nreq = inet_reqsk_clone(req, nsk);
1008 if (!nreq)
1016 refcount_set(&nreq->rsk_refcnt, 2 + 1);
1017 timer_setup(&nreq->rsk_timer, reqsk_timer_handler, TIMER_PINNED);
1020 req = nreq;
1068 if (!nreq)
1071 if (!inet_ehash_insert(req_to_sk(nreq), req_to_sk(oreq), NULL)) {
1073 inet_csk_reqsk_queue_drop(sk_listener, nreq);
1082 reqsk_put(nreq);
1087 * SYN+ACKs (nreq->num_timeout > max_syn_ack_retries, etc), or another
1090 if (nreq) {
1093 reqsk_migrate_reset(nreq);
1094 reqsk_queue_removed(queue, nreq);
1095 __reqsk_free(nreq);
1335 struct request_sock *nreq;
1337 /* hold a refcnt for the nreq->rsk_listener
1341 nreq = inet_reqsk_clone(req, sk);
1342 if (!nreq) {
1347 refcount_set(&nreq->rsk_refcnt, 1);
1348 if (inet_csk_reqsk_queue_add(sk, nreq, child)) {
1356 reqsk_migrate_reset(nreq);
1357 __reqsk_free(nreq);
1390 struct request_sock *nreq;
1399 nreq = inet_reqsk_clone(req, nsk);
1400 if (nreq) {
1401 refcount_set(&nreq->rsk_refcnt, 1);
1403 if (inet_csk_reqsk_queue_add(nsk, nreq, child)) {
1410 reqsk_migrate_reset(nreq);
1411 __reqsk_free(nreq);