Lines Matching refs:xa
53 const struct xdp_mem_allocator *xa = ptr;
56 return xa->mem.id != mem_id;
73 struct xdp_mem_allocator *xa;
75 xa = container_of(rcu, struct xdp_mem_allocator, rcu);
78 ida_simple_remove(&mem_id_pool, xa->mem.id);
80 kfree(xa);
83 static void mem_xa_remove(struct xdp_mem_allocator *xa)
85 trace_mem_disconnect(xa);
87 if (!rhashtable_remove_fast(mem_id_ht, &xa->node, mem_id_rht_params))
88 call_rcu(&xa->rcu, __xdp_mem_allocator_rcu_free);
93 struct xdp_mem_allocator *xa;
102 while ((xa = rhashtable_walk_next(&iter)) && !IS_ERR(xa)) {
103 if (xa->allocator == allocator)
104 mem_xa_remove(xa);
109 } while (xa == ERR_PTR(-EAGAIN));
117 struct xdp_mem_allocator *xa;
130 xa = rhashtable_lookup(mem_id_ht, &id, mem_id_rht_params);
131 page_pool_destroy(xa->page_pool);
460 struct xdp_mem_allocator *xa = bq->xa;
462 if (unlikely(!xa || !bq->count))
465 page_pool_put_page_bulk(xa->page_pool, bq->q, bq->count);
466 /* bq->xa is not cleared to save lookup, if mem.id same in next bulk */
476 struct xdp_mem_allocator *xa;
483 xa = bq->xa;
484 if (unlikely(!xa)) {
485 xa = rhashtable_lookup(mem_id_ht, &mem->id, mem_id_rht_params);
487 bq->xa = xa;
493 if (unlikely(mem->id != xa->mem.id)) {
495 bq->xa = rhashtable_lookup(mem_id_ht, &mem->id, mem_id_rht_params);