/kernel/linux/linux-5.10/net/ipv4/ |
H A D | tcp_fastopen.c | 337 __NET_INC_STATS(sock_net(sk), in tcp_fastopen_queue_check() 354 return (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen) & flag) || in tcp_fastopen_no_cookie() 369 int tcp_fastopen = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen); in tcp_try_fastopen() 375 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPFASTOPENCOOKIEREQD); in tcp_try_fastopen() 394 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen() 412 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen() 417 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen() 421 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen() 509 struct net *net = sock_net(sk); in tcp_fastopen_active_disable() 511 if (!READ_ONCE(sock_net(s in tcp_fastopen_active_disable() [all...] |
H A D | tcp_timer.c | 75 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONTIMEOUT); in tcp_write_err() 126 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONMEMORY); in tcp_out_of_resources() 130 if (!check_net(sock_net(sk))) { in tcp_out_of_resources() 146 int retries = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_orphan_retries); /* May be zero. */ in tcp_orphan_retries() 162 const struct net *net = sock_net(sk); in tcp_mtu_probing() 235 struct net *net = sock_net(sk); in tcp_write_timeout() 283 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPTIMEOUTREHASH); in tcp_write_timeout() 319 __NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKS); in tcp_delack_timer_handler() 347 __NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKLOCKED); in tcp_delack_timer() 384 max_probes = READ_ONCE(sock_net(s in tcp_probe_timer() [all...] |
H A D | inet_hashtables.c | 52 return inet6_ehashfn(sock_net(sk), in sk_ehashfn() 64 return inet_ehashfn(sock_net(sk), in sk_ehashfn() 118 const int bhash = inet_bhashfn(sock_net(sk), inet_sk(sk)->inet_num, in __inet_put_port() 144 const int bhash = inet_bhashfn(sock_net(sk), port, in __inet_inherit_port() 165 if (net_eq(ib_net(tb), sock_net(sk)) && in __inet_inherit_port() 171 sock_net(sk), head, port, in __inet_inherit_port() 194 hash = ipv6_portaddr_hash(sock_net(sk), in inet_lhash2_bucket_sk() 199 hash = ipv4_portaddr_hash(sock_net(sk), in inet_lhash2_bucket_sk() 247 if (net_eq(sock_net(sk), net) && sk->sk_num == hnum && in compute_score() 457 struct net *net = sock_net(s in __inet_check_established() [all...] |
H A D | syncookies.c | 338 if (!READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_syncookies) || in cookie_v4_check() 347 __NET_INC_STATS(sock_net(sk), LINUX_MIB_SYNCOOKIESFAILED); in cookie_v4_check() 351 __NET_INC_STATS(sock_net(sk), LINUX_MIB_SYNCOOKIESRECV); in cookie_v4_check() 355 tcp_parse_options(sock_net(sk), skb, &tcp_opt, 0, NULL); in cookie_v4_check() 358 tsoff = secure_tcp_ts_off(sock_net(sk), in cookie_v4_check() 364 if (!cookie_timestamp_decode(sock_net(sk), &tcp_opt)) in cookie_v4_check() 401 RCU_INIT_POINTER(ireq->ireq_opt, tcp_v4_save_options(sock_net(sk), skb)); in cookie_v4_check() 422 rt = ip_route_output_key(sock_net(sk), &fl4); in cookie_v4_check() 442 ireq->ecn_ok = cookie_ecn_ok(&tcp_opt, sock_net(sk), &rt->dst); in cookie_v4_check()
|
H A D | tcp_minisocks.c | 258 struct inet_timewait_death_row *tcp_death_row = &sock_net(sk)->ipv4.tcp_death_row; in tcp_time_wait() 336 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPTIMEWAITOVERFLOW); in tcp_time_wait() 553 __TCP_INC_STATS(sock_net(sk), TCP_MIB_PASSIVEOPENS); in tcp_create_openreq_child() 586 tcp_parse_options(sock_net(sk), skb, &tmp_opt, 0, NULL); in tcp_check_req() 628 if (!tcp_oow_rate_limited(sock_net(sk), skb, in tcp_check_req() 718 !tcp_oow_rate_limited(sock_net(sk), skb, in tcp_check_req() 723 NET_INC_STATS(sock_net(sk), LINUX_MIB_PAWSESTABREJECTED); in tcp_check_req() 745 TCP_INC_STATS(sock_net(sk), TCP_MIB_ATTEMPTFAILS); in tcp_check_req() 768 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPDEFERACCEPTDROP); in tcp_check_req() 795 if (!sock_net(s in tcp_check_req() [all...] |
H A D | tcp_output.c | 85 NET_ADD_STATS(sock_net(sk), LINUX_MIB_TCPORIGDATASENT, in tcp_event_new_data_sent() 187 NET_ADD_STATS(sock_net(sk), LINUX_MIB_TCPACKCOMPRESSED, in tcp_event_ack_sent() 231 if (sock_net(sk)->ipv4.sysctl_tcp_workaround_signed_windows) in tcp_select_initial_window() 242 space = max_t(u32, space, READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_rmem[2])); in tcp_select_initial_window() 275 NET_INC_STATS(sock_net(sk), in tcp_select_window() 286 sock_net(sk)->ipv4.sysctl_tcp_workaround_signed_windows) in tcp_select_window() 298 NET_INC_STATS(sock_net(sk), in tcp_select_window() 301 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPFROMZEROWINDOWADV); in tcp_select_window() 325 bool use_ecn = sock_net(sk)->ipv4.sysctl_tcp_ecn == 1 || in tcp_ecn_send_syn() 347 if (sock_net(s in tcp_ecn_clear_syn() [all...] |
/kernel/linux/linux-6.6/net/ipv4/ |
H A D | tcp_fastopen.c | 321 __NET_INC_STATS(sock_net(sk), in tcp_fastopen_queue_check() 338 return (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen) & flag) || in tcp_fastopen_no_cookie() 353 int tcp_fastopen = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen); in tcp_try_fastopen() 359 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPFASTOPENCOOKIEREQD); in tcp_try_fastopen() 378 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen() 396 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen() 401 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen() 405 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen() 493 struct net *net = sock_net(sk); in tcp_fastopen_active_disable() 495 if (!READ_ONCE(sock_net(s in tcp_fastopen_active_disable() [all...] |
H A D | tcp_plb.c | 28 struct net *net = sock_net(sk); in tcp_plb_update_state() 48 struct net *net = sock_net(sk); in tcp_plb_check_rehash() 83 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPPLBREHASH); in tcp_plb_check_rehash() 93 struct net *net = sock_net(sk); in tcp_plb_update_state_upon_rto()
|
H A D | tcp_timer.c | 77 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONTIMEOUT); in tcp_write_err() 128 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONMEMORY); in tcp_out_of_resources() 132 if (!check_net(sock_net(sk))) { in tcp_out_of_resources() 148 int retries = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_orphan_retries); /* May be zero. */ in tcp_orphan_retries() 164 const struct net *net = sock_net(sk); in tcp_mtu_probing() 237 struct net *net = sock_net(sk); in tcp_write_timeout() 291 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPTIMEOUTREHASH); in tcp_write_timeout() 335 __NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKS); in tcp_delack_timer_handler() 359 __NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKLOCKED); in tcp_delack_timer() 399 max_probes = READ_ONCE(sock_net(s in tcp_probe_timer() [all...] |
H A D | udp_diag.c | 36 struct net *net = sock_net(in_skb->sk); in udp_dump_one() 94 struct net *net = sock_net(skb->sk); in udp_dump() 117 if (!net_eq(sock_net(sk), net)) in udp_dump() 150 udp_dump(sock_net(cb->skb->sk)->ipv4.udp_table, skb, cb, r); in udp_diag_dump() 156 return udp_dump_one(sock_net(cb->skb->sk)->ipv4.udp_table, cb, req); in udp_diag_dump_one() 171 struct net *net = sock_net(in_skb->sk); in __udp_diag_destroy() 228 return __udp_diag_destroy(in_skb, req, sock_net(in_skb->sk)->ipv4.udp_table); in udp_diag_destroy()
|
H A D | syncookies.c | 336 if (!READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_syncookies) || in cookie_v4_check() 345 __NET_INC_STATS(sock_net(sk), LINUX_MIB_SYNCOOKIESFAILED); in cookie_v4_check() 349 __NET_INC_STATS(sock_net(sk), LINUX_MIB_SYNCOOKIESRECV); in cookie_v4_check() 353 tcp_parse_options(sock_net(sk), skb, &tcp_opt, 0, NULL); in cookie_v4_check() 356 tsoff = secure_tcp_ts_off(sock_net(sk), in cookie_v4_check() 362 if (!cookie_timestamp_decode(sock_net(sk), &tcp_opt)) in cookie_v4_check() 399 RCU_INIT_POINTER(ireq->ireq_opt, tcp_v4_save_options(sock_net(sk), skb)); in cookie_v4_check() 420 rt = ip_route_output_key(sock_net(sk), &fl4); in cookie_v4_check() 440 ireq->ecn_ok = cookie_ecn_ok(&tcp_opt, sock_net(sk), &rt->dst); in cookie_v4_check()
|
H A D | tcp_minisocks.c | 284 struct net *net = sock_net(sk); in tcp_time_wait() 579 __TCP_INC_STATS(sock_net(sk), TCP_MIB_PASSIVEOPENS); in tcp_create_openreq_child() 612 tcp_parse_options(sock_net(sk), skb, &tmp_opt, 0, NULL); in tcp_check_req() 654 if (!tcp_oow_rate_limited(sock_net(sk), skb, in tcp_check_req() 743 !tcp_oow_rate_limited(sock_net(sk), skb, in tcp_check_req() 748 NET_INC_STATS(sock_net(sk), LINUX_MIB_PAWSESTABREJECTED); in tcp_check_req() 770 TCP_INC_STATS(sock_net(sk), TCP_MIB_ATTEMPTFAILS); in tcp_check_req() 793 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPDEFERACCEPTDROP); in tcp_check_req() 821 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPMIGRATEREQFAILURE); in tcp_check_req() 823 if (!READ_ONCE(sock_net(s in tcp_check_req() [all...] |
/kernel/linux/linux-5.10/net/ipv6/ |
H A D | syncookies.c | 144 if (!READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_syncookies) || in cookie_v6_check() 153 __NET_INC_STATS(sock_net(sk), LINUX_MIB_SYNCOOKIESFAILED); in cookie_v6_check() 157 __NET_INC_STATS(sock_net(sk), LINUX_MIB_SYNCOOKIESRECV); in cookie_v6_check() 161 tcp_parse_options(sock_net(sk), skb, &tcp_opt, 0, NULL); in cookie_v6_check() 164 tsoff = secure_tcpv6_ts_off(sock_net(sk), in cookie_v6_check() 170 if (!cookie_timestamp_decode(sock_net(sk), &tcp_opt)) in cookie_v6_check() 241 dst = ip6_dst_lookup_flow(sock_net(sk), sk, &fl6, final_p); in cookie_v6_check() 259 ireq->ecn_ok = cookie_ecn_ok(&tcp_opt, sock_net(sk), dst); in cookie_v6_check()
|
/kernel/linux/linux-6.6/net/ipv6/ |
H A D | syncookies.c | 144 if (!READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_syncookies) || in cookie_v6_check() 153 __NET_INC_STATS(sock_net(sk), LINUX_MIB_SYNCOOKIESFAILED); in cookie_v6_check() 157 __NET_INC_STATS(sock_net(sk), LINUX_MIB_SYNCOOKIESRECV); in cookie_v6_check() 161 tcp_parse_options(sock_net(sk), skb, &tcp_opt, 0, NULL); in cookie_v6_check() 164 tsoff = secure_tcpv6_ts_off(sock_net(sk), in cookie_v6_check() 170 if (!cookie_timestamp_decode(sock_net(sk), &tcp_opt)) in cookie_v6_check() 241 dst = ip6_dst_lookup_flow(sock_net(sk), sk, &fl6, final_p); in cookie_v6_check() 259 ireq->ecn_ok = cookie_ecn_ok(&tcp_opt, sock_net(sk), dst); in cookie_v6_check()
|
/kernel/linux/linux-5.10/net/can/ |
H A D | raw.c | 275 if (!net_eq(dev_net(dev), sock_net(sk))) in raw_notify() 390 dev = dev_get_by_index(sock_net(sk), ro->ifindex); in raw_release() 396 raw_disable_allfilters(sock_net(sk), NULL, sk); in raw_release() 439 dev = dev_get_by_index(sock_net(sk), addr->can_ifindex); in raw_bind() 455 err = raw_enable_allfilters(sock_net(sk), dev, sk); in raw_bind() 461 err = raw_enable_allfilters(sock_net(sk), NULL, sk); in raw_bind() 470 dev = dev_get_by_index(sock_net(sk), in raw_bind() 478 raw_disable_allfilters(sock_net(sk), NULL, sk); in raw_bind() 553 dev = dev_get_by_index(sock_net(sk), ro->ifindex); in raw_setsockopt() 565 err = raw_enable_filters(sock_net(s in raw_setsockopt() [all...] |
/kernel/linux/linux-6.6/net/mptcp/ |
H A D | token.c | 179 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); in mptcp_token_new_connect() 198 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); in mptcp_token_accept() 260 !net_eq(sock_net(sk), net)) in mptcp_token_get_sock() 267 !net_eq(sock_net(sk), net)) { in mptcp_token_get_sock() 316 if (!net_eq(sock_net(sk), net)) in mptcp_token_iter_next() 325 if (!net_eq(sock_net(sk), net)) { in mptcp_token_iter_next() 384 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, -1); in mptcp_token_destroy()
|
/kernel/linux/linux-5.10/net/vmw_vsock/ |
H A D | diag.c | 59 net = sock_net(skb->sk); in vsock_diag_dump() 79 if (!net_eq(sock_net(sk), net)) in vsock_diag_dump() 113 if (!net_eq(sock_net(sk), net)) in vsock_diag_dump() 144 struct net *net = sock_net(skb->sk); in vsock_diag_handler_dump()
|
/kernel/linux/linux-6.6/net/vmw_vsock/ |
H A D | diag.c | 59 net = sock_net(skb->sk); in vsock_diag_dump() 79 if (!net_eq(sock_net(sk), net)) in vsock_diag_dump() 113 if (!net_eq(sock_net(sk), net)) in vsock_diag_dump() 144 struct net *net = sock_net(skb->sk); in vsock_diag_handler_dump()
|
/kernel/linux/common_modules/newip/third_party/linux-5.10/net/newip/ |
H A D | tcp_nip_timer.c | 37 int retries = sock_net(sk)->ipv4.sysctl_tcp_orphan_retries; /* May be zero. */ in tcp_nip_orphan_retries() 70 __NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKS); in tcp_nip_delack_timer_handler() 82 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONTIMEOUT); in tcp_nip_write_err() 95 __NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKLOCKED); in tcp_nip_delack_timer() 121 struct net *net = sock_net(sk); in tcp_nip_write_timeout() 154 struct net *net = sock_net(sk); in tcp_nip_retransmit_timer() 233 /* default: sock_net(sk)->ipv4.sysctl_tcp_retries2 */ in tcp_nip_probe_timer()
|
/kernel/linux/linux-5.10/net/ieee802154/ |
H A D | socket.c | 137 dev_load(sock_net(sk), ifr.ifr_name); in ieee802154_dev_ioctl() 138 dev = dev_get_by_name(sock_net(sk), ifr.ifr_name); in ieee802154_dev_ioctl() 178 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); in raw_hash() 188 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, -1); in raw_unhash() 215 dev = ieee802154_get_dev(sock_net(sk), &addr); in raw_bind() 257 dev = dev_getfirstbyhwtype(sock_net(sk), ARPHRD_IEEE802154); in raw_sendmsg() 259 dev = dev_get_by_index(sock_net(sk), sk->sk_bound_dev_if); in raw_sendmsg() 462 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); in dgram_hash() 472 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, -1); in dgram_unhash() 512 dev = ieee802154_get_dev(sock_net(s in dgram_bind() [all...] |
/kernel/linux/linux-6.6/net/ieee802154/ |
H A D | socket.c | 136 dev_load(sock_net(sk), ifr.ifr_name); in ieee802154_dev_ioctl() 137 dev = dev_get_by_name(sock_net(sk), ifr.ifr_name); in ieee802154_dev_ioctl() 178 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); in raw_hash() 187 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, -1); in raw_unhash() 214 dev = ieee802154_get_dev(sock_net(sk), &addr); in raw_bind() 256 dev = dev_getfirstbyhwtype(sock_net(sk), ARPHRD_IEEE802154); in raw_sendmsg() 258 dev = dev_get_by_index(sock_net(sk), sk->sk_bound_dev_if); in raw_sendmsg() 461 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); in dgram_hash() 470 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, -1); in dgram_unhash() 510 dev = ieee802154_get_dev(sock_net(s in dgram_bind() [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/core/ |
H A D | nldev.c | 886 device = ib_device_get_by_index(sock_net(skb->sk), index); in nldev_get_doit() 907 return rdma_nl_unicast(sock_net(skb->sk), msg, NETLINK_CB(skb).portid); in nldev_get_doit() 930 device = ib_device_get_by_index(sock_net(skb->sk), index); in nldev_set_doit() 1024 device = ib_device_get_by_index(sock_net(skb->sk), index); in nldev_port_get_doit() 1044 err = fill_port_info(msg, device, port, sock_net(skb->sk)); in nldev_port_get_doit() 1051 return rdma_nl_unicast(sock_net(skb->sk), msg, NETLINK_CB(skb).portid); in nldev_port_get_doit() 1078 device = ib_device_get_by_index(sock_net(skb->sk), ifindex); in nldev_port_get_dumpit() 1104 if (fill_port_info(skb, device, p, sock_net(skb->sk))) { in nldev_port_get_dumpit() 1133 device = ib_device_get_by_index(sock_net(skb->sk), index); in nldev_res_get_doit() 1153 return rdma_nl_unicast(sock_net(sk in nldev_res_get_doit() [all...] |
/kernel/linux/linux-5.10/net/tls/ |
H A D | tls_main.c | 285 TLS_DEC_STATS(sock_net(sk), LINUX_MIB_TLSCURRTXSW); in tls_sk_proto_cleanup() 288 TLS_DEC_STATS(sock_net(sk), LINUX_MIB_TLSCURRTXDEVICE); in tls_sk_proto_cleanup() 293 TLS_DEC_STATS(sock_net(sk), LINUX_MIB_TLSCURRRXSW); in tls_sk_proto_cleanup() 296 TLS_DEC_STATS(sock_net(sk), LINUX_MIB_TLSCURRRXDEVICE); in tls_sk_proto_cleanup() 550 TLS_INC_STATS(sock_net(sk), LINUX_MIB_TLSTXDEVICE); in do_tls_setsockopt_conf() 551 TLS_INC_STATS(sock_net(sk), LINUX_MIB_TLSCURRTXDEVICE); in do_tls_setsockopt_conf() 556 TLS_INC_STATS(sock_net(sk), LINUX_MIB_TLSTXSW); in do_tls_setsockopt_conf() 557 TLS_INC_STATS(sock_net(sk), LINUX_MIB_TLSCURRTXSW); in do_tls_setsockopt_conf() 564 TLS_INC_STATS(sock_net(sk), LINUX_MIB_TLSRXDEVICE); in do_tls_setsockopt_conf() 565 TLS_INC_STATS(sock_net(s in do_tls_setsockopt_conf() [all...] |
/kernel/linux/linux-6.6/net/can/ |
H A D | raw.c | 279 if (!net_eq(dev_net(dev), sock_net(sk))) in raw_notify() 401 raw_disable_allfilters(sock_net(sk), NULL, sk); in raw_release() 447 dev = dev_get_by_index(sock_net(sk), addr->can_ifindex); in raw_bind() 463 err = raw_enable_allfilters(sock_net(sk), dev, sk); in raw_bind() 471 err = raw_enable_allfilters(sock_net(sk), NULL, sk); in raw_bind() 483 raw_disable_allfilters(sock_net(sk), NULL, sk); in raw_bind() 580 err = raw_enable_filters(sock_net(sk), dev, sk, in raw_setsockopt() 583 err = raw_enable_filters(sock_net(sk), dev, sk, in raw_setsockopt() 592 raw_disable_filters(sock_net(sk), dev, sk, ro->filter, in raw_setsockopt() 638 err = raw_enable_errfilter(sock_net(s in raw_setsockopt() [all...] |
/kernel/linux/linux-5.10/net/phonet/ |
H A D | socket.c | 79 if (!net_eq(sock_net(sknode), net)) in pn_find_sock_by_sa() 116 if (!net_eq(sock_net(sknode), net)) in pn_deliver_sock_broadcast() 175 if (saddr && phonet_address_lookup(sock_net(sk), saddr)) in pn_socket_bind() 372 dev = dev_get_by_index(sock_net(sk), in pn_socket_ioctl() 375 dev = phonet_device_get(sock_net(sk)); in pn_socket_ioctl() 474 struct net *net = sock_net(sk); in pn_sock_get_port() 529 if (!net_eq(net, sock_net(sknode))) in pn_sock_get_idx() 546 while (sk && !net_eq(net, sock_net(sk))); in pn_sock_get_next() 636 if (!net_eq(sock_net(sk), &init_net)) in pn_sock_bind_res()
|