Lines Matching defs:pool
56 struct rds_ib_mr_pool *pool;
62 pool = rds_ibdev->mr_8k_pool;
64 pool = rds_ibdev->mr_1m_pool;
66 ibmr = rds_ib_try_reuse_ibmr(pool);
79 pool->max_pages);
86 ibmr->pool = pool;
87 if (pool->pool_type == RDS_IB_MR_8K_POOL)
92 if (atomic_read(&pool->item_count) > pool->max_items_soft)
93 pool->max_items_soft = pool->max_items;
102 atomic_dec(&pool->item_count);
108 struct rds_ib_mr_pool *pool = ibmr->pool;
111 llist_add(&ibmr->llnode, &pool->drop_list);
113 llist_add(&ibmr->llnode, &pool->free_list);
114 atomic_add(ibmr->sg_len, &pool->free_pinned);
115 atomic_inc(&pool->dirty_count);
118 if (atomic_read(&pool->free_pinned) >= pool->max_free_pinned ||
119 atomic_read(&pool->dirty_count) >= pool->max_items / 5)
120 queue_delayed_work(rds_ib_mr_wq, &pool->flush_worker, 10);
188 struct rds_ib_mr_pool *pool,
243 if (frmr->dma_npages > ibmr->pool->max_pages) {
252 if (ibmr->pool->pool_type == RDS_IB_MR_8K_POOL)
387 if (ibmr->pool->pool_type == RDS_IB_MR_8K_POOL)
426 ret = rds_ib_map_frmr(rds_ibdev, ibmr->pool, ibmr, sg, nents);
439 struct rds_ib_mr_pool *pool = ibmr->pool;
443 llist_add(&ibmr->llnode, &pool->drop_list);
445 llist_add(&ibmr->llnode, &pool->free_list);