Lines Matching refs:node

3366 	int node = blk_mq_get_hctx_node(set, hctx_idx);
3369 if (node == NUMA_NO_NODE)
3370 node = set->numa_node;
3372 tags = blk_mq_init_tags(nr_tags, reserved_tags, node,
3379 node);
3385 node);
3399 unsigned int hctx_idx, int node)
3404 ret = set->ops->init_request(set, rq, hctx_idx, node);
3418 int node = blk_mq_get_hctx_node(set, hctx_idx);
3421 if (node == NUMA_NO_NODE)
3422 node = set->numa_node;
3444 page = alloc_pages_node(node,
3474 if (blk_mq_init_request(set, rq, hctx_idx, node)) {
3527 static int blk_mq_hctx_notify_offline(unsigned int cpu, struct hlist_node *node)
3529 struct blk_mq_hw_ctx *hctx = hlist_entry_safe(node,
3560 static int blk_mq_hctx_notify_online(unsigned int cpu, struct hlist_node *node)
3562 struct blk_mq_hw_ctx *hctx = hlist_entry_safe(node,
3575 static int blk_mq_hctx_notify_dead(unsigned int cpu, struct hlist_node *node)
3582 hctx = hlist_entry_safe(node, struct blk_mq_hw_ctx, cpuhp_dead);
3725 int node)
3730 hctx = kzalloc_node(sizeof(struct blk_mq_hw_ctx), gfp, node);
3734 if (!zalloc_cpumask_var_node(&hctx->cpumask, gfp, node))
3738 if (node == NUMA_NO_NODE)
3739 node = set->numa_node;
3740 hctx->numa_node = node;
3755 gfp, node);
3760 gfp, node, false, false))
3807 * Set local node, IFF we have more than one hw queue. If
3808 * not, we remain on the home node of the device
4190 int hctx_idx, int node)
4197 if (tmp->numa_node == node) {
4207 hctx = blk_mq_alloc_hctx(q, set, node);
4232 int node = blk_mq_get_hctx_node(set, i);
4240 if (!blk_mq_alloc_and_init_hctx(set, q, i, node)) {
4243 pr_warn("Allocate new hctx on node %d fails, fallback to previous one on node %d\n",
4244 node, old_node);
4680 struct list_head node;
4707 INIT_LIST_HEAD(&qe->node);
4712 list_add(&qe->node, head);
4725 list_for_each_entry(qe, head, node)
4742 list_del(&qe->node);