Lines Matching refs:node

88  * Per-node worker thread pool
94 int node;
471 struct io_wq_work_node *node, *prev;
476 wq_list_for_each(node, prev, &acct->work_list) {
479 work = container_of(node, struct io_wq_work, list);
483 wq_list_del(&acct->work_list, node, prev);
500 node = &tail->list;
764 tsk = create_io_thread(io_wqe_worker, worker, wqe->node);
812 worker = kzalloc_node(sizeof(*worker), GFP_KERNEL, wqe->node);
831 tsk = create_io_thread(io_wqe_worker, worker, wqe->node);
858 /* no task if node is/was offline */
1026 struct io_wq_work_node *node, *prev;
1029 wq_list_for_each(node, prev, &acct->work_list) {
1030 work = container_of(node, struct io_wq_work, list);
1078 int node;
1085 for_each_node(node) {
1086 struct io_wqe *wqe = wq->wqes[node];
1099 for_each_node(node) {
1100 struct io_wqe *wqe = wq->wqes[node];
1135 int ret, node, i;
1156 for_each_node(node) {
1158 int alloc_node = node;
1165 wq->wqes[node] = wqe;
1168 cpumask_copy(wqe->cpu_mask, cpumask_of_node(node));
1169 wqe->node = alloc_node;
1195 for_each_node(node) {
1196 if (!wq->wqes[node])
1198 free_cpumask_var(wq->wqes[node]->cpu_mask);
1199 kfree(wq->wqes[node]);
1241 int node;
1249 for_each_node(node) {
1250 struct io_wqe *wqe = wq->wqes[node];
1258 for_each_node(node) {
1260 list_del_init(&wq->wqes[node]->wait.entry);
1269 int node;
1273 for_each_node(node) {
1274 struct io_wqe *wqe = wq->wqes[node];
1326 static int io_wq_cpu_online(unsigned int cpu, struct hlist_node *node)
1328 struct io_wq *wq = hlist_entry_safe(node, struct io_wq, cpuhp_node);
1333 static int io_wq_cpu_offline(unsigned int cpu, struct hlist_node *node)
1335 struct io_wq *wq = hlist_entry_safe(node, struct io_wq, cpuhp_node);
1365 int i, node;
1380 for_each_node(node) {
1381 struct io_wqe *wqe = wq->wqes[node];