xref: /kernel/linux/linux-6.6/block/blk-ioc.c (revision 62306a36)
162306a36Sopenharmony_ci// SPDX-License-Identifier: GPL-2.0
262306a36Sopenharmony_ci/*
362306a36Sopenharmony_ci * Functions related to io context handling
462306a36Sopenharmony_ci */
562306a36Sopenharmony_ci#include <linux/kernel.h>
662306a36Sopenharmony_ci#include <linux/module.h>
762306a36Sopenharmony_ci#include <linux/init.h>
862306a36Sopenharmony_ci#include <linux/bio.h>
962306a36Sopenharmony_ci#include <linux/blkdev.h>
1062306a36Sopenharmony_ci#include <linux/slab.h>
1162306a36Sopenharmony_ci#include <linux/security.h>
1262306a36Sopenharmony_ci#include <linux/sched/task.h>
1362306a36Sopenharmony_ci
1462306a36Sopenharmony_ci#include "blk.h"
1562306a36Sopenharmony_ci#include "blk-mq-sched.h"
1662306a36Sopenharmony_ci
1762306a36Sopenharmony_ci/*
1862306a36Sopenharmony_ci * For io context allocations
1962306a36Sopenharmony_ci */
2062306a36Sopenharmony_cistatic struct kmem_cache *iocontext_cachep;
2162306a36Sopenharmony_ci
2262306a36Sopenharmony_ci#ifdef CONFIG_BLK_ICQ
2362306a36Sopenharmony_ci/**
2462306a36Sopenharmony_ci * get_io_context - increment reference count to io_context
2562306a36Sopenharmony_ci * @ioc: io_context to get
2662306a36Sopenharmony_ci *
2762306a36Sopenharmony_ci * Increment reference count to @ioc.
2862306a36Sopenharmony_ci */
2962306a36Sopenharmony_cistatic void get_io_context(struct io_context *ioc)
3062306a36Sopenharmony_ci{
3162306a36Sopenharmony_ci	BUG_ON(atomic_long_read(&ioc->refcount) <= 0);
3262306a36Sopenharmony_ci	atomic_long_inc(&ioc->refcount);
3362306a36Sopenharmony_ci}
3462306a36Sopenharmony_ci
3562306a36Sopenharmony_cistatic void icq_free_icq_rcu(struct rcu_head *head)
3662306a36Sopenharmony_ci{
3762306a36Sopenharmony_ci	struct io_cq *icq = container_of(head, struct io_cq, __rcu_head);
3862306a36Sopenharmony_ci
3962306a36Sopenharmony_ci	kmem_cache_free(icq->__rcu_icq_cache, icq);
4062306a36Sopenharmony_ci}
4162306a36Sopenharmony_ci
4262306a36Sopenharmony_ci/*
4362306a36Sopenharmony_ci * Exit an icq. Called with ioc locked for blk-mq, and with both ioc
4462306a36Sopenharmony_ci * and queue locked for legacy.
4562306a36Sopenharmony_ci */
4662306a36Sopenharmony_cistatic void ioc_exit_icq(struct io_cq *icq)
4762306a36Sopenharmony_ci{
4862306a36Sopenharmony_ci	struct elevator_type *et = icq->q->elevator->type;
4962306a36Sopenharmony_ci
5062306a36Sopenharmony_ci	if (icq->flags & ICQ_EXITED)
5162306a36Sopenharmony_ci		return;
5262306a36Sopenharmony_ci
5362306a36Sopenharmony_ci	if (et->ops.exit_icq)
5462306a36Sopenharmony_ci		et->ops.exit_icq(icq);
5562306a36Sopenharmony_ci
5662306a36Sopenharmony_ci	icq->flags |= ICQ_EXITED;
5762306a36Sopenharmony_ci}
5862306a36Sopenharmony_ci
5962306a36Sopenharmony_cistatic void ioc_exit_icqs(struct io_context *ioc)
6062306a36Sopenharmony_ci{
6162306a36Sopenharmony_ci	struct io_cq *icq;
6262306a36Sopenharmony_ci
6362306a36Sopenharmony_ci	spin_lock_irq(&ioc->lock);
6462306a36Sopenharmony_ci	hlist_for_each_entry(icq, &ioc->icq_list, ioc_node)
6562306a36Sopenharmony_ci		ioc_exit_icq(icq);
6662306a36Sopenharmony_ci	spin_unlock_irq(&ioc->lock);
6762306a36Sopenharmony_ci}
6862306a36Sopenharmony_ci
6962306a36Sopenharmony_ci/*
7062306a36Sopenharmony_ci * Release an icq. Called with ioc locked for blk-mq, and with both ioc
7162306a36Sopenharmony_ci * and queue locked for legacy.
7262306a36Sopenharmony_ci */
7362306a36Sopenharmony_cistatic void ioc_destroy_icq(struct io_cq *icq)
7462306a36Sopenharmony_ci{
7562306a36Sopenharmony_ci	struct io_context *ioc = icq->ioc;
7662306a36Sopenharmony_ci	struct request_queue *q = icq->q;
7762306a36Sopenharmony_ci	struct elevator_type *et = q->elevator->type;
7862306a36Sopenharmony_ci
7962306a36Sopenharmony_ci	lockdep_assert_held(&ioc->lock);
8062306a36Sopenharmony_ci	lockdep_assert_held(&q->queue_lock);
8162306a36Sopenharmony_ci
8262306a36Sopenharmony_ci	if (icq->flags & ICQ_DESTROYED)
8362306a36Sopenharmony_ci		return;
8462306a36Sopenharmony_ci
8562306a36Sopenharmony_ci	radix_tree_delete(&ioc->icq_tree, icq->q->id);
8662306a36Sopenharmony_ci	hlist_del_init(&icq->ioc_node);
8762306a36Sopenharmony_ci	list_del_init(&icq->q_node);
8862306a36Sopenharmony_ci
8962306a36Sopenharmony_ci	/*
9062306a36Sopenharmony_ci	 * Both setting lookup hint to and clearing it from @icq are done
9162306a36Sopenharmony_ci	 * under queue_lock.  If it's not pointing to @icq now, it never
9262306a36Sopenharmony_ci	 * will.  Hint assignment itself can race safely.
9362306a36Sopenharmony_ci	 */
9462306a36Sopenharmony_ci	if (rcu_access_pointer(ioc->icq_hint) == icq)
9562306a36Sopenharmony_ci		rcu_assign_pointer(ioc->icq_hint, NULL);
9662306a36Sopenharmony_ci
9762306a36Sopenharmony_ci	ioc_exit_icq(icq);
9862306a36Sopenharmony_ci
9962306a36Sopenharmony_ci	/*
10062306a36Sopenharmony_ci	 * @icq->q might have gone away by the time RCU callback runs
10162306a36Sopenharmony_ci	 * making it impossible to determine icq_cache.  Record it in @icq.
10262306a36Sopenharmony_ci	 */
10362306a36Sopenharmony_ci	icq->__rcu_icq_cache = et->icq_cache;
10462306a36Sopenharmony_ci	icq->flags |= ICQ_DESTROYED;
10562306a36Sopenharmony_ci	call_rcu(&icq->__rcu_head, icq_free_icq_rcu);
10662306a36Sopenharmony_ci}
10762306a36Sopenharmony_ci
10862306a36Sopenharmony_ci/*
10962306a36Sopenharmony_ci * Slow path for ioc release in put_io_context().  Performs double-lock
11062306a36Sopenharmony_ci * dancing to unlink all icq's and then frees ioc.
11162306a36Sopenharmony_ci */
11262306a36Sopenharmony_cistatic void ioc_release_fn(struct work_struct *work)
11362306a36Sopenharmony_ci{
11462306a36Sopenharmony_ci	struct io_context *ioc = container_of(work, struct io_context,
11562306a36Sopenharmony_ci					      release_work);
11662306a36Sopenharmony_ci	spin_lock_irq(&ioc->lock);
11762306a36Sopenharmony_ci
11862306a36Sopenharmony_ci	while (!hlist_empty(&ioc->icq_list)) {
11962306a36Sopenharmony_ci		struct io_cq *icq = hlist_entry(ioc->icq_list.first,
12062306a36Sopenharmony_ci						struct io_cq, ioc_node);
12162306a36Sopenharmony_ci		struct request_queue *q = icq->q;
12262306a36Sopenharmony_ci
12362306a36Sopenharmony_ci		if (spin_trylock(&q->queue_lock)) {
12462306a36Sopenharmony_ci			ioc_destroy_icq(icq);
12562306a36Sopenharmony_ci			spin_unlock(&q->queue_lock);
12662306a36Sopenharmony_ci		} else {
12762306a36Sopenharmony_ci			/* Make sure q and icq cannot be freed. */
12862306a36Sopenharmony_ci			rcu_read_lock();
12962306a36Sopenharmony_ci
13062306a36Sopenharmony_ci			/* Re-acquire the locks in the correct order. */
13162306a36Sopenharmony_ci			spin_unlock(&ioc->lock);
13262306a36Sopenharmony_ci			spin_lock(&q->queue_lock);
13362306a36Sopenharmony_ci			spin_lock(&ioc->lock);
13462306a36Sopenharmony_ci
13562306a36Sopenharmony_ci			ioc_destroy_icq(icq);
13662306a36Sopenharmony_ci
13762306a36Sopenharmony_ci			spin_unlock(&q->queue_lock);
13862306a36Sopenharmony_ci			rcu_read_unlock();
13962306a36Sopenharmony_ci		}
14062306a36Sopenharmony_ci	}
14162306a36Sopenharmony_ci
14262306a36Sopenharmony_ci	spin_unlock_irq(&ioc->lock);
14362306a36Sopenharmony_ci
14462306a36Sopenharmony_ci	kmem_cache_free(iocontext_cachep, ioc);
14562306a36Sopenharmony_ci}
14662306a36Sopenharmony_ci
14762306a36Sopenharmony_ci/*
14862306a36Sopenharmony_ci * Releasing icqs requires reverse order double locking and we may already be
14962306a36Sopenharmony_ci * holding a queue_lock.  Do it asynchronously from a workqueue.
15062306a36Sopenharmony_ci */
15162306a36Sopenharmony_cistatic bool ioc_delay_free(struct io_context *ioc)
15262306a36Sopenharmony_ci{
15362306a36Sopenharmony_ci	unsigned long flags;
15462306a36Sopenharmony_ci
15562306a36Sopenharmony_ci	spin_lock_irqsave(&ioc->lock, flags);
15662306a36Sopenharmony_ci	if (!hlist_empty(&ioc->icq_list)) {
15762306a36Sopenharmony_ci		queue_work(system_power_efficient_wq, &ioc->release_work);
15862306a36Sopenharmony_ci		spin_unlock_irqrestore(&ioc->lock, flags);
15962306a36Sopenharmony_ci		return true;
16062306a36Sopenharmony_ci	}
16162306a36Sopenharmony_ci	spin_unlock_irqrestore(&ioc->lock, flags);
16262306a36Sopenharmony_ci	return false;
16362306a36Sopenharmony_ci}
16462306a36Sopenharmony_ci
16562306a36Sopenharmony_ci/**
16662306a36Sopenharmony_ci * ioc_clear_queue - break any ioc association with the specified queue
16762306a36Sopenharmony_ci * @q: request_queue being cleared
16862306a36Sopenharmony_ci *
16962306a36Sopenharmony_ci * Walk @q->icq_list and exit all io_cq's.
17062306a36Sopenharmony_ci */
17162306a36Sopenharmony_civoid ioc_clear_queue(struct request_queue *q)
17262306a36Sopenharmony_ci{
17362306a36Sopenharmony_ci	spin_lock_irq(&q->queue_lock);
17462306a36Sopenharmony_ci	while (!list_empty(&q->icq_list)) {
17562306a36Sopenharmony_ci		struct io_cq *icq =
17662306a36Sopenharmony_ci			list_first_entry(&q->icq_list, struct io_cq, q_node);
17762306a36Sopenharmony_ci
17862306a36Sopenharmony_ci		/*
17962306a36Sopenharmony_ci		 * Other context won't hold ioc lock to wait for queue_lock, see
18062306a36Sopenharmony_ci		 * details in ioc_release_fn().
18162306a36Sopenharmony_ci		 */
18262306a36Sopenharmony_ci		spin_lock(&icq->ioc->lock);
18362306a36Sopenharmony_ci		ioc_destroy_icq(icq);
18462306a36Sopenharmony_ci		spin_unlock(&icq->ioc->lock);
18562306a36Sopenharmony_ci	}
18662306a36Sopenharmony_ci	spin_unlock_irq(&q->queue_lock);
18762306a36Sopenharmony_ci}
18862306a36Sopenharmony_ci#else /* CONFIG_BLK_ICQ */
18962306a36Sopenharmony_cistatic inline void ioc_exit_icqs(struct io_context *ioc)
19062306a36Sopenharmony_ci{
19162306a36Sopenharmony_ci}
19262306a36Sopenharmony_cistatic inline bool ioc_delay_free(struct io_context *ioc)
19362306a36Sopenharmony_ci{
19462306a36Sopenharmony_ci	return false;
19562306a36Sopenharmony_ci}
19662306a36Sopenharmony_ci#endif /* CONFIG_BLK_ICQ */
19762306a36Sopenharmony_ci
19862306a36Sopenharmony_ci/**
19962306a36Sopenharmony_ci * put_io_context - put a reference of io_context
20062306a36Sopenharmony_ci * @ioc: io_context to put
20162306a36Sopenharmony_ci *
20262306a36Sopenharmony_ci * Decrement reference count of @ioc and release it if the count reaches
20362306a36Sopenharmony_ci * zero.
20462306a36Sopenharmony_ci */
20562306a36Sopenharmony_civoid put_io_context(struct io_context *ioc)
20662306a36Sopenharmony_ci{
20762306a36Sopenharmony_ci	BUG_ON(atomic_long_read(&ioc->refcount) <= 0);
20862306a36Sopenharmony_ci	if (atomic_long_dec_and_test(&ioc->refcount) && !ioc_delay_free(ioc))
20962306a36Sopenharmony_ci		kmem_cache_free(iocontext_cachep, ioc);
21062306a36Sopenharmony_ci}
21162306a36Sopenharmony_ciEXPORT_SYMBOL_GPL(put_io_context);
21262306a36Sopenharmony_ci
21362306a36Sopenharmony_ci/* Called by the exiting task */
21462306a36Sopenharmony_civoid exit_io_context(struct task_struct *task)
21562306a36Sopenharmony_ci{
21662306a36Sopenharmony_ci	struct io_context *ioc;
21762306a36Sopenharmony_ci
21862306a36Sopenharmony_ci	task_lock(task);
21962306a36Sopenharmony_ci	ioc = task->io_context;
22062306a36Sopenharmony_ci	task->io_context = NULL;
22162306a36Sopenharmony_ci	task_unlock(task);
22262306a36Sopenharmony_ci
22362306a36Sopenharmony_ci	if (atomic_dec_and_test(&ioc->active_ref)) {
22462306a36Sopenharmony_ci		ioc_exit_icqs(ioc);
22562306a36Sopenharmony_ci		put_io_context(ioc);
22662306a36Sopenharmony_ci	}
22762306a36Sopenharmony_ci}
22862306a36Sopenharmony_ci
22962306a36Sopenharmony_cistatic struct io_context *alloc_io_context(gfp_t gfp_flags, int node)
23062306a36Sopenharmony_ci{
23162306a36Sopenharmony_ci	struct io_context *ioc;
23262306a36Sopenharmony_ci
23362306a36Sopenharmony_ci	ioc = kmem_cache_alloc_node(iocontext_cachep, gfp_flags | __GFP_ZERO,
23462306a36Sopenharmony_ci				    node);
23562306a36Sopenharmony_ci	if (unlikely(!ioc))
23662306a36Sopenharmony_ci		return NULL;
23762306a36Sopenharmony_ci
23862306a36Sopenharmony_ci	atomic_long_set(&ioc->refcount, 1);
23962306a36Sopenharmony_ci	atomic_set(&ioc->active_ref, 1);
24062306a36Sopenharmony_ci#ifdef CONFIG_BLK_ICQ
24162306a36Sopenharmony_ci	spin_lock_init(&ioc->lock);
24262306a36Sopenharmony_ci	INIT_RADIX_TREE(&ioc->icq_tree, GFP_ATOMIC);
24362306a36Sopenharmony_ci	INIT_HLIST_HEAD(&ioc->icq_list);
24462306a36Sopenharmony_ci	INIT_WORK(&ioc->release_work, ioc_release_fn);
24562306a36Sopenharmony_ci#endif
24662306a36Sopenharmony_ci	ioc->ioprio = IOPRIO_DEFAULT;
24762306a36Sopenharmony_ci
24862306a36Sopenharmony_ci	return ioc;
24962306a36Sopenharmony_ci}
25062306a36Sopenharmony_ci
25162306a36Sopenharmony_ciint set_task_ioprio(struct task_struct *task, int ioprio)
25262306a36Sopenharmony_ci{
25362306a36Sopenharmony_ci	int err;
25462306a36Sopenharmony_ci	const struct cred *cred = current_cred(), *tcred;
25562306a36Sopenharmony_ci
25662306a36Sopenharmony_ci	rcu_read_lock();
25762306a36Sopenharmony_ci	tcred = __task_cred(task);
25862306a36Sopenharmony_ci	if (!uid_eq(tcred->uid, cred->euid) &&
25962306a36Sopenharmony_ci	    !uid_eq(tcred->uid, cred->uid) && !capable(CAP_SYS_NICE)) {
26062306a36Sopenharmony_ci		rcu_read_unlock();
26162306a36Sopenharmony_ci		return -EPERM;
26262306a36Sopenharmony_ci	}
26362306a36Sopenharmony_ci	rcu_read_unlock();
26462306a36Sopenharmony_ci
26562306a36Sopenharmony_ci	err = security_task_setioprio(task, ioprio);
26662306a36Sopenharmony_ci	if (err)
26762306a36Sopenharmony_ci		return err;
26862306a36Sopenharmony_ci
26962306a36Sopenharmony_ci	task_lock(task);
27062306a36Sopenharmony_ci	if (unlikely(!task->io_context)) {
27162306a36Sopenharmony_ci		struct io_context *ioc;
27262306a36Sopenharmony_ci
27362306a36Sopenharmony_ci		task_unlock(task);
27462306a36Sopenharmony_ci
27562306a36Sopenharmony_ci		ioc = alloc_io_context(GFP_ATOMIC, NUMA_NO_NODE);
27662306a36Sopenharmony_ci		if (!ioc)
27762306a36Sopenharmony_ci			return -ENOMEM;
27862306a36Sopenharmony_ci
27962306a36Sopenharmony_ci		task_lock(task);
28062306a36Sopenharmony_ci		if (task->flags & PF_EXITING) {
28162306a36Sopenharmony_ci			kmem_cache_free(iocontext_cachep, ioc);
28262306a36Sopenharmony_ci			goto out;
28362306a36Sopenharmony_ci		}
28462306a36Sopenharmony_ci		if (task->io_context)
28562306a36Sopenharmony_ci			kmem_cache_free(iocontext_cachep, ioc);
28662306a36Sopenharmony_ci		else
28762306a36Sopenharmony_ci			task->io_context = ioc;
28862306a36Sopenharmony_ci	}
28962306a36Sopenharmony_ci	task->io_context->ioprio = ioprio;
29062306a36Sopenharmony_ciout:
29162306a36Sopenharmony_ci	task_unlock(task);
29262306a36Sopenharmony_ci	return 0;
29362306a36Sopenharmony_ci}
29462306a36Sopenharmony_ciEXPORT_SYMBOL_GPL(set_task_ioprio);
29562306a36Sopenharmony_ci
29662306a36Sopenharmony_ciint __copy_io(unsigned long clone_flags, struct task_struct *tsk)
29762306a36Sopenharmony_ci{
29862306a36Sopenharmony_ci	struct io_context *ioc = current->io_context;
29962306a36Sopenharmony_ci
30062306a36Sopenharmony_ci	/*
30162306a36Sopenharmony_ci	 * Share io context with parent, if CLONE_IO is set
30262306a36Sopenharmony_ci	 */
30362306a36Sopenharmony_ci	if (clone_flags & CLONE_IO) {
30462306a36Sopenharmony_ci		atomic_inc(&ioc->active_ref);
30562306a36Sopenharmony_ci		tsk->io_context = ioc;
30662306a36Sopenharmony_ci	} else if (ioprio_valid(ioc->ioprio)) {
30762306a36Sopenharmony_ci		tsk->io_context = alloc_io_context(GFP_KERNEL, NUMA_NO_NODE);
30862306a36Sopenharmony_ci		if (!tsk->io_context)
30962306a36Sopenharmony_ci			return -ENOMEM;
31062306a36Sopenharmony_ci		tsk->io_context->ioprio = ioc->ioprio;
31162306a36Sopenharmony_ci	}
31262306a36Sopenharmony_ci
31362306a36Sopenharmony_ci	return 0;
31462306a36Sopenharmony_ci}
31562306a36Sopenharmony_ci
31662306a36Sopenharmony_ci#ifdef CONFIG_BLK_ICQ
31762306a36Sopenharmony_ci/**
31862306a36Sopenharmony_ci * ioc_lookup_icq - lookup io_cq from ioc
31962306a36Sopenharmony_ci * @q: the associated request_queue
32062306a36Sopenharmony_ci *
32162306a36Sopenharmony_ci * Look up io_cq associated with @ioc - @q pair from @ioc.  Must be called
32262306a36Sopenharmony_ci * with @q->queue_lock held.
32362306a36Sopenharmony_ci */
32462306a36Sopenharmony_cistruct io_cq *ioc_lookup_icq(struct request_queue *q)
32562306a36Sopenharmony_ci{
32662306a36Sopenharmony_ci	struct io_context *ioc = current->io_context;
32762306a36Sopenharmony_ci	struct io_cq *icq;
32862306a36Sopenharmony_ci
32962306a36Sopenharmony_ci	lockdep_assert_held(&q->queue_lock);
33062306a36Sopenharmony_ci
33162306a36Sopenharmony_ci	/*
33262306a36Sopenharmony_ci	 * icq's are indexed from @ioc using radix tree and hint pointer,
33362306a36Sopenharmony_ci	 * both of which are protected with RCU.  All removals are done
33462306a36Sopenharmony_ci	 * holding both q and ioc locks, and we're holding q lock - if we
33562306a36Sopenharmony_ci	 * find a icq which points to us, it's guaranteed to be valid.
33662306a36Sopenharmony_ci	 */
33762306a36Sopenharmony_ci	rcu_read_lock();
33862306a36Sopenharmony_ci	icq = rcu_dereference(ioc->icq_hint);
33962306a36Sopenharmony_ci	if (icq && icq->q == q)
34062306a36Sopenharmony_ci		goto out;
34162306a36Sopenharmony_ci
34262306a36Sopenharmony_ci	icq = radix_tree_lookup(&ioc->icq_tree, q->id);
34362306a36Sopenharmony_ci	if (icq && icq->q == q)
34462306a36Sopenharmony_ci		rcu_assign_pointer(ioc->icq_hint, icq);	/* allowed to race */
34562306a36Sopenharmony_ci	else
34662306a36Sopenharmony_ci		icq = NULL;
34762306a36Sopenharmony_ciout:
34862306a36Sopenharmony_ci	rcu_read_unlock();
34962306a36Sopenharmony_ci	return icq;
35062306a36Sopenharmony_ci}
35162306a36Sopenharmony_ciEXPORT_SYMBOL(ioc_lookup_icq);
35262306a36Sopenharmony_ci
35362306a36Sopenharmony_ci/**
35462306a36Sopenharmony_ci * ioc_create_icq - create and link io_cq
35562306a36Sopenharmony_ci * @q: request_queue of interest
35662306a36Sopenharmony_ci *
35762306a36Sopenharmony_ci * Make sure io_cq linking @ioc and @q exists.  If icq doesn't exist, they
35862306a36Sopenharmony_ci * will be created using @gfp_mask.
35962306a36Sopenharmony_ci *
36062306a36Sopenharmony_ci * The caller is responsible for ensuring @ioc won't go away and @q is
36162306a36Sopenharmony_ci * alive and will stay alive until this function returns.
36262306a36Sopenharmony_ci */
36362306a36Sopenharmony_cistatic struct io_cq *ioc_create_icq(struct request_queue *q)
36462306a36Sopenharmony_ci{
36562306a36Sopenharmony_ci	struct io_context *ioc = current->io_context;
36662306a36Sopenharmony_ci	struct elevator_type *et = q->elevator->type;
36762306a36Sopenharmony_ci	struct io_cq *icq;
36862306a36Sopenharmony_ci
36962306a36Sopenharmony_ci	/* allocate stuff */
37062306a36Sopenharmony_ci	icq = kmem_cache_alloc_node(et->icq_cache, GFP_ATOMIC | __GFP_ZERO,
37162306a36Sopenharmony_ci				    q->node);
37262306a36Sopenharmony_ci	if (!icq)
37362306a36Sopenharmony_ci		return NULL;
37462306a36Sopenharmony_ci
37562306a36Sopenharmony_ci	if (radix_tree_maybe_preload(GFP_ATOMIC) < 0) {
37662306a36Sopenharmony_ci		kmem_cache_free(et->icq_cache, icq);
37762306a36Sopenharmony_ci		return NULL;
37862306a36Sopenharmony_ci	}
37962306a36Sopenharmony_ci
38062306a36Sopenharmony_ci	icq->ioc = ioc;
38162306a36Sopenharmony_ci	icq->q = q;
38262306a36Sopenharmony_ci	INIT_LIST_HEAD(&icq->q_node);
38362306a36Sopenharmony_ci	INIT_HLIST_NODE(&icq->ioc_node);
38462306a36Sopenharmony_ci
38562306a36Sopenharmony_ci	/* lock both q and ioc and try to link @icq */
38662306a36Sopenharmony_ci	spin_lock_irq(&q->queue_lock);
38762306a36Sopenharmony_ci	spin_lock(&ioc->lock);
38862306a36Sopenharmony_ci
38962306a36Sopenharmony_ci	if (likely(!radix_tree_insert(&ioc->icq_tree, q->id, icq))) {
39062306a36Sopenharmony_ci		hlist_add_head(&icq->ioc_node, &ioc->icq_list);
39162306a36Sopenharmony_ci		list_add(&icq->q_node, &q->icq_list);
39262306a36Sopenharmony_ci		if (et->ops.init_icq)
39362306a36Sopenharmony_ci			et->ops.init_icq(icq);
39462306a36Sopenharmony_ci	} else {
39562306a36Sopenharmony_ci		kmem_cache_free(et->icq_cache, icq);
39662306a36Sopenharmony_ci		icq = ioc_lookup_icq(q);
39762306a36Sopenharmony_ci		if (!icq)
39862306a36Sopenharmony_ci			printk(KERN_ERR "cfq: icq link failed!\n");
39962306a36Sopenharmony_ci	}
40062306a36Sopenharmony_ci
40162306a36Sopenharmony_ci	spin_unlock(&ioc->lock);
40262306a36Sopenharmony_ci	spin_unlock_irq(&q->queue_lock);
40362306a36Sopenharmony_ci	radix_tree_preload_end();
40462306a36Sopenharmony_ci	return icq;
40562306a36Sopenharmony_ci}
40662306a36Sopenharmony_ci
40762306a36Sopenharmony_cistruct io_cq *ioc_find_get_icq(struct request_queue *q)
40862306a36Sopenharmony_ci{
40962306a36Sopenharmony_ci	struct io_context *ioc = current->io_context;
41062306a36Sopenharmony_ci	struct io_cq *icq = NULL;
41162306a36Sopenharmony_ci
41262306a36Sopenharmony_ci	if (unlikely(!ioc)) {
41362306a36Sopenharmony_ci		ioc = alloc_io_context(GFP_ATOMIC, q->node);
41462306a36Sopenharmony_ci		if (!ioc)
41562306a36Sopenharmony_ci			return NULL;
41662306a36Sopenharmony_ci
41762306a36Sopenharmony_ci		task_lock(current);
41862306a36Sopenharmony_ci		if (current->io_context) {
41962306a36Sopenharmony_ci			kmem_cache_free(iocontext_cachep, ioc);
42062306a36Sopenharmony_ci			ioc = current->io_context;
42162306a36Sopenharmony_ci		} else {
42262306a36Sopenharmony_ci			current->io_context = ioc;
42362306a36Sopenharmony_ci		}
42462306a36Sopenharmony_ci
42562306a36Sopenharmony_ci		get_io_context(ioc);
42662306a36Sopenharmony_ci		task_unlock(current);
42762306a36Sopenharmony_ci	} else {
42862306a36Sopenharmony_ci		get_io_context(ioc);
42962306a36Sopenharmony_ci
43062306a36Sopenharmony_ci		spin_lock_irq(&q->queue_lock);
43162306a36Sopenharmony_ci		icq = ioc_lookup_icq(q);
43262306a36Sopenharmony_ci		spin_unlock_irq(&q->queue_lock);
43362306a36Sopenharmony_ci	}
43462306a36Sopenharmony_ci
43562306a36Sopenharmony_ci	if (!icq) {
43662306a36Sopenharmony_ci		icq = ioc_create_icq(q);
43762306a36Sopenharmony_ci		if (!icq) {
43862306a36Sopenharmony_ci			put_io_context(ioc);
43962306a36Sopenharmony_ci			return NULL;
44062306a36Sopenharmony_ci		}
44162306a36Sopenharmony_ci	}
44262306a36Sopenharmony_ci	return icq;
44362306a36Sopenharmony_ci}
44462306a36Sopenharmony_ciEXPORT_SYMBOL_GPL(ioc_find_get_icq);
44562306a36Sopenharmony_ci#endif /* CONFIG_BLK_ICQ */
44662306a36Sopenharmony_ci
44762306a36Sopenharmony_cistatic int __init blk_ioc_init(void)
44862306a36Sopenharmony_ci{
44962306a36Sopenharmony_ci	iocontext_cachep = kmem_cache_create("blkdev_ioc",
45062306a36Sopenharmony_ci			sizeof(struct io_context), 0, SLAB_PANIC, NULL);
45162306a36Sopenharmony_ci	return 0;
45262306a36Sopenharmony_ci}
45362306a36Sopenharmony_cisubsys_initcall(blk_ioc_init);
454