Lines Matching defs:conn
127 * - reassembly is optional and easily done by transports per conn
138 struct rds_connection *conn = cp->cp_conn;
183 * rds_conn_shutdown() sets the conn state and then tests RDS_IN_XMIT,
192 if (conn->c_trans->xmit_path_prepare)
193 conn->c_trans->xmit_path_prepare(cp);
218 if (!rm && test_and_clear_bit(0, &conn->c_map_queued)) {
219 rm = rds_cong_update_alloc(conn);
315 ret = conn->c_trans->xmit_rdma(conn, &rm->rdma);
331 ret = conn->c_trans->xmit_atomic(conn, &rm->atomic);
366 ret = conn->c_trans->xmit(conn, rm,
420 if (conn->c_trans->xmit_path_complete)
421 conn->c_trans->xmit_path_complete(cp);
453 if ((test_bit(0, &conn->c_map_queued) ||
713 void rds_send_drop_acked(struct rds_connection *conn, u64 ack,
716 WARN_ON(conn->c_trans->t_mp_capable);
717 rds_send_path_drop_acked(&conn->c_path[0], ack, is_acked);
724 struct rds_connection *conn;
751 /* Remove the messages from the conn */
754 conn = rm->m_inc.i_conn;
755 if (conn->c_trans->t_mp_capable)
758 cp = &conn->c_path[0];
762 * Maybe someone else beat us to removing rm from the conn.
817 static int rds_send_queue_rm(struct rds_sock *rs, struct rds_connection *conn,
862 rm->m_inc.i_conn = conn;
1045 struct rds_connection *conn, int nonblock)
1049 if (conn->c_npaths == 0)
1052 hash = RDS_MPATH_HASH(rs, conn->c_npaths);
1053 if (conn->c_npaths == 0 && hash != 0) {
1054 rds_send_ping(conn, 0);
1061 if (conn->c_npaths == 0) {
1067 if (wait_event_interruptible(conn->c_hs_waitq,
1068 conn->c_npaths != 0))
1071 if (conn->c_npaths == 1)
1108 struct rds_connection *conn;
1289 * Caching the conn in the socket helps a lot. */
1292 conn = rs->rs_conn;
1294 conn = rds_conn_create_outgoing(sock_net(sock->sk),
1299 if (IS_ERR(conn)) {
1300 ret = PTR_ERR(conn);
1303 rs->rs_conn = conn;
1306 if (conn->c_trans->t_mp_capable)
1307 cpath = &conn->c_path[rds_send_mprds_hash(rs, conn, nonblock)];
1309 cpath = &conn->c_path[0];
1318 if (rm->rdma.op_active && !conn->c_trans->xmit_rdma) {
1319 printk_ratelimited(KERN_NOTICE "rdma_op %p conn xmit_rdma %p\n",
1320 &rm->rdma, conn->c_trans->xmit_rdma);
1325 if (rm->atomic.op_active && !conn->c_trans->xmit_atomic) {
1326 printk_ratelimited(KERN_NOTICE "atomic_op %p conn xmit_atomic %p\n",
1327 &rm->atomic, conn->c_trans->xmit_atomic);
1332 if (rds_destroy_pending(conn)) {
1338 rds_check_all_paths(conn);
1340 ret = rds_cong_wait(conn->c_fcong, dport, nonblock, rs);
1345 while (!rds_send_queue_rm(rs, conn, cpath, rm, rs->rs_bound_port,
1355 rds_send_queue_rm(rs, conn, cpath, rm,
1495 rds_send_ping(struct rds_connection *conn, int cp_index)
1498 struct rds_conn_path *cp = &conn->c_path[cp_index];
1501 if (conn->c_ping_triggered) {
1505 conn->c_ping_triggered = 1;