162306a36Sopenharmony_ci// SPDX-License-Identifier: GPL-2.0 262306a36Sopenharmony_ci/* 362306a36Sopenharmony_ci * Functions related to io context handling 462306a36Sopenharmony_ci */ 562306a36Sopenharmony_ci#include <linux/kernel.h> 662306a36Sopenharmony_ci#include <linux/module.h> 762306a36Sopenharmony_ci#include <linux/init.h> 862306a36Sopenharmony_ci#include <linux/bio.h> 962306a36Sopenharmony_ci#include <linux/blkdev.h> 1062306a36Sopenharmony_ci#include <linux/slab.h> 1162306a36Sopenharmony_ci#include <linux/security.h> 1262306a36Sopenharmony_ci#include <linux/sched/task.h> 1362306a36Sopenharmony_ci 1462306a36Sopenharmony_ci#include "blk.h" 1562306a36Sopenharmony_ci#include "blk-mq-sched.h" 1662306a36Sopenharmony_ci 1762306a36Sopenharmony_ci/* 1862306a36Sopenharmony_ci * For io context allocations 1962306a36Sopenharmony_ci */ 2062306a36Sopenharmony_cistatic struct kmem_cache *iocontext_cachep; 2162306a36Sopenharmony_ci 2262306a36Sopenharmony_ci#ifdef CONFIG_BLK_ICQ 2362306a36Sopenharmony_ci/** 2462306a36Sopenharmony_ci * get_io_context - increment reference count to io_context 2562306a36Sopenharmony_ci * @ioc: io_context to get 2662306a36Sopenharmony_ci * 2762306a36Sopenharmony_ci * Increment reference count to @ioc. 2862306a36Sopenharmony_ci */ 2962306a36Sopenharmony_cistatic void get_io_context(struct io_context *ioc) 3062306a36Sopenharmony_ci{ 3162306a36Sopenharmony_ci BUG_ON(atomic_long_read(&ioc->refcount) <= 0); 3262306a36Sopenharmony_ci atomic_long_inc(&ioc->refcount); 3362306a36Sopenharmony_ci} 3462306a36Sopenharmony_ci 3562306a36Sopenharmony_cistatic void icq_free_icq_rcu(struct rcu_head *head) 3662306a36Sopenharmony_ci{ 3762306a36Sopenharmony_ci struct io_cq *icq = container_of(head, struct io_cq, __rcu_head); 3862306a36Sopenharmony_ci 3962306a36Sopenharmony_ci kmem_cache_free(icq->__rcu_icq_cache, icq); 4062306a36Sopenharmony_ci} 4162306a36Sopenharmony_ci 4262306a36Sopenharmony_ci/* 4362306a36Sopenharmony_ci * Exit an icq. Called with ioc locked for blk-mq, and with both ioc 4462306a36Sopenharmony_ci * and queue locked for legacy. 4562306a36Sopenharmony_ci */ 4662306a36Sopenharmony_cistatic void ioc_exit_icq(struct io_cq *icq) 4762306a36Sopenharmony_ci{ 4862306a36Sopenharmony_ci struct elevator_type *et = icq->q->elevator->type; 4962306a36Sopenharmony_ci 5062306a36Sopenharmony_ci if (icq->flags & ICQ_EXITED) 5162306a36Sopenharmony_ci return; 5262306a36Sopenharmony_ci 5362306a36Sopenharmony_ci if (et->ops.exit_icq) 5462306a36Sopenharmony_ci et->ops.exit_icq(icq); 5562306a36Sopenharmony_ci 5662306a36Sopenharmony_ci icq->flags |= ICQ_EXITED; 5762306a36Sopenharmony_ci} 5862306a36Sopenharmony_ci 5962306a36Sopenharmony_cistatic void ioc_exit_icqs(struct io_context *ioc) 6062306a36Sopenharmony_ci{ 6162306a36Sopenharmony_ci struct io_cq *icq; 6262306a36Sopenharmony_ci 6362306a36Sopenharmony_ci spin_lock_irq(&ioc->lock); 6462306a36Sopenharmony_ci hlist_for_each_entry(icq, &ioc->icq_list, ioc_node) 6562306a36Sopenharmony_ci ioc_exit_icq(icq); 6662306a36Sopenharmony_ci spin_unlock_irq(&ioc->lock); 6762306a36Sopenharmony_ci} 6862306a36Sopenharmony_ci 6962306a36Sopenharmony_ci/* 7062306a36Sopenharmony_ci * Release an icq. Called with ioc locked for blk-mq, and with both ioc 7162306a36Sopenharmony_ci * and queue locked for legacy. 7262306a36Sopenharmony_ci */ 7362306a36Sopenharmony_cistatic void ioc_destroy_icq(struct io_cq *icq) 7462306a36Sopenharmony_ci{ 7562306a36Sopenharmony_ci struct io_context *ioc = icq->ioc; 7662306a36Sopenharmony_ci struct request_queue *q = icq->q; 7762306a36Sopenharmony_ci struct elevator_type *et = q->elevator->type; 7862306a36Sopenharmony_ci 7962306a36Sopenharmony_ci lockdep_assert_held(&ioc->lock); 8062306a36Sopenharmony_ci lockdep_assert_held(&q->queue_lock); 8162306a36Sopenharmony_ci 8262306a36Sopenharmony_ci if (icq->flags & ICQ_DESTROYED) 8362306a36Sopenharmony_ci return; 8462306a36Sopenharmony_ci 8562306a36Sopenharmony_ci radix_tree_delete(&ioc->icq_tree, icq->q->id); 8662306a36Sopenharmony_ci hlist_del_init(&icq->ioc_node); 8762306a36Sopenharmony_ci list_del_init(&icq->q_node); 8862306a36Sopenharmony_ci 8962306a36Sopenharmony_ci /* 9062306a36Sopenharmony_ci * Both setting lookup hint to and clearing it from @icq are done 9162306a36Sopenharmony_ci * under queue_lock. If it's not pointing to @icq now, it never 9262306a36Sopenharmony_ci * will. Hint assignment itself can race safely. 9362306a36Sopenharmony_ci */ 9462306a36Sopenharmony_ci if (rcu_access_pointer(ioc->icq_hint) == icq) 9562306a36Sopenharmony_ci rcu_assign_pointer(ioc->icq_hint, NULL); 9662306a36Sopenharmony_ci 9762306a36Sopenharmony_ci ioc_exit_icq(icq); 9862306a36Sopenharmony_ci 9962306a36Sopenharmony_ci /* 10062306a36Sopenharmony_ci * @icq->q might have gone away by the time RCU callback runs 10162306a36Sopenharmony_ci * making it impossible to determine icq_cache. Record it in @icq. 10262306a36Sopenharmony_ci */ 10362306a36Sopenharmony_ci icq->__rcu_icq_cache = et->icq_cache; 10462306a36Sopenharmony_ci icq->flags |= ICQ_DESTROYED; 10562306a36Sopenharmony_ci call_rcu(&icq->__rcu_head, icq_free_icq_rcu); 10662306a36Sopenharmony_ci} 10762306a36Sopenharmony_ci 10862306a36Sopenharmony_ci/* 10962306a36Sopenharmony_ci * Slow path for ioc release in put_io_context(). Performs double-lock 11062306a36Sopenharmony_ci * dancing to unlink all icq's and then frees ioc. 11162306a36Sopenharmony_ci */ 11262306a36Sopenharmony_cistatic void ioc_release_fn(struct work_struct *work) 11362306a36Sopenharmony_ci{ 11462306a36Sopenharmony_ci struct io_context *ioc = container_of(work, struct io_context, 11562306a36Sopenharmony_ci release_work); 11662306a36Sopenharmony_ci spin_lock_irq(&ioc->lock); 11762306a36Sopenharmony_ci 11862306a36Sopenharmony_ci while (!hlist_empty(&ioc->icq_list)) { 11962306a36Sopenharmony_ci struct io_cq *icq = hlist_entry(ioc->icq_list.first, 12062306a36Sopenharmony_ci struct io_cq, ioc_node); 12162306a36Sopenharmony_ci struct request_queue *q = icq->q; 12262306a36Sopenharmony_ci 12362306a36Sopenharmony_ci if (spin_trylock(&q->queue_lock)) { 12462306a36Sopenharmony_ci ioc_destroy_icq(icq); 12562306a36Sopenharmony_ci spin_unlock(&q->queue_lock); 12662306a36Sopenharmony_ci } else { 12762306a36Sopenharmony_ci /* Make sure q and icq cannot be freed. */ 12862306a36Sopenharmony_ci rcu_read_lock(); 12962306a36Sopenharmony_ci 13062306a36Sopenharmony_ci /* Re-acquire the locks in the correct order. */ 13162306a36Sopenharmony_ci spin_unlock(&ioc->lock); 13262306a36Sopenharmony_ci spin_lock(&q->queue_lock); 13362306a36Sopenharmony_ci spin_lock(&ioc->lock); 13462306a36Sopenharmony_ci 13562306a36Sopenharmony_ci ioc_destroy_icq(icq); 13662306a36Sopenharmony_ci 13762306a36Sopenharmony_ci spin_unlock(&q->queue_lock); 13862306a36Sopenharmony_ci rcu_read_unlock(); 13962306a36Sopenharmony_ci } 14062306a36Sopenharmony_ci } 14162306a36Sopenharmony_ci 14262306a36Sopenharmony_ci spin_unlock_irq(&ioc->lock); 14362306a36Sopenharmony_ci 14462306a36Sopenharmony_ci kmem_cache_free(iocontext_cachep, ioc); 14562306a36Sopenharmony_ci} 14662306a36Sopenharmony_ci 14762306a36Sopenharmony_ci/* 14862306a36Sopenharmony_ci * Releasing icqs requires reverse order double locking and we may already be 14962306a36Sopenharmony_ci * holding a queue_lock. Do it asynchronously from a workqueue. 15062306a36Sopenharmony_ci */ 15162306a36Sopenharmony_cistatic bool ioc_delay_free(struct io_context *ioc) 15262306a36Sopenharmony_ci{ 15362306a36Sopenharmony_ci unsigned long flags; 15462306a36Sopenharmony_ci 15562306a36Sopenharmony_ci spin_lock_irqsave(&ioc->lock, flags); 15662306a36Sopenharmony_ci if (!hlist_empty(&ioc->icq_list)) { 15762306a36Sopenharmony_ci queue_work(system_power_efficient_wq, &ioc->release_work); 15862306a36Sopenharmony_ci spin_unlock_irqrestore(&ioc->lock, flags); 15962306a36Sopenharmony_ci return true; 16062306a36Sopenharmony_ci } 16162306a36Sopenharmony_ci spin_unlock_irqrestore(&ioc->lock, flags); 16262306a36Sopenharmony_ci return false; 16362306a36Sopenharmony_ci} 16462306a36Sopenharmony_ci 16562306a36Sopenharmony_ci/** 16662306a36Sopenharmony_ci * ioc_clear_queue - break any ioc association with the specified queue 16762306a36Sopenharmony_ci * @q: request_queue being cleared 16862306a36Sopenharmony_ci * 16962306a36Sopenharmony_ci * Walk @q->icq_list and exit all io_cq's. 17062306a36Sopenharmony_ci */ 17162306a36Sopenharmony_civoid ioc_clear_queue(struct request_queue *q) 17262306a36Sopenharmony_ci{ 17362306a36Sopenharmony_ci spin_lock_irq(&q->queue_lock); 17462306a36Sopenharmony_ci while (!list_empty(&q->icq_list)) { 17562306a36Sopenharmony_ci struct io_cq *icq = 17662306a36Sopenharmony_ci list_first_entry(&q->icq_list, struct io_cq, q_node); 17762306a36Sopenharmony_ci 17862306a36Sopenharmony_ci /* 17962306a36Sopenharmony_ci * Other context won't hold ioc lock to wait for queue_lock, see 18062306a36Sopenharmony_ci * details in ioc_release_fn(). 18162306a36Sopenharmony_ci */ 18262306a36Sopenharmony_ci spin_lock(&icq->ioc->lock); 18362306a36Sopenharmony_ci ioc_destroy_icq(icq); 18462306a36Sopenharmony_ci spin_unlock(&icq->ioc->lock); 18562306a36Sopenharmony_ci } 18662306a36Sopenharmony_ci spin_unlock_irq(&q->queue_lock); 18762306a36Sopenharmony_ci} 18862306a36Sopenharmony_ci#else /* CONFIG_BLK_ICQ */ 18962306a36Sopenharmony_cistatic inline void ioc_exit_icqs(struct io_context *ioc) 19062306a36Sopenharmony_ci{ 19162306a36Sopenharmony_ci} 19262306a36Sopenharmony_cistatic inline bool ioc_delay_free(struct io_context *ioc) 19362306a36Sopenharmony_ci{ 19462306a36Sopenharmony_ci return false; 19562306a36Sopenharmony_ci} 19662306a36Sopenharmony_ci#endif /* CONFIG_BLK_ICQ */ 19762306a36Sopenharmony_ci 19862306a36Sopenharmony_ci/** 19962306a36Sopenharmony_ci * put_io_context - put a reference of io_context 20062306a36Sopenharmony_ci * @ioc: io_context to put 20162306a36Sopenharmony_ci * 20262306a36Sopenharmony_ci * Decrement reference count of @ioc and release it if the count reaches 20362306a36Sopenharmony_ci * zero. 20462306a36Sopenharmony_ci */ 20562306a36Sopenharmony_civoid put_io_context(struct io_context *ioc) 20662306a36Sopenharmony_ci{ 20762306a36Sopenharmony_ci BUG_ON(atomic_long_read(&ioc->refcount) <= 0); 20862306a36Sopenharmony_ci if (atomic_long_dec_and_test(&ioc->refcount) && !ioc_delay_free(ioc)) 20962306a36Sopenharmony_ci kmem_cache_free(iocontext_cachep, ioc); 21062306a36Sopenharmony_ci} 21162306a36Sopenharmony_ciEXPORT_SYMBOL_GPL(put_io_context); 21262306a36Sopenharmony_ci 21362306a36Sopenharmony_ci/* Called by the exiting task */ 21462306a36Sopenharmony_civoid exit_io_context(struct task_struct *task) 21562306a36Sopenharmony_ci{ 21662306a36Sopenharmony_ci struct io_context *ioc; 21762306a36Sopenharmony_ci 21862306a36Sopenharmony_ci task_lock(task); 21962306a36Sopenharmony_ci ioc = task->io_context; 22062306a36Sopenharmony_ci task->io_context = NULL; 22162306a36Sopenharmony_ci task_unlock(task); 22262306a36Sopenharmony_ci 22362306a36Sopenharmony_ci if (atomic_dec_and_test(&ioc->active_ref)) { 22462306a36Sopenharmony_ci ioc_exit_icqs(ioc); 22562306a36Sopenharmony_ci put_io_context(ioc); 22662306a36Sopenharmony_ci } 22762306a36Sopenharmony_ci} 22862306a36Sopenharmony_ci 22962306a36Sopenharmony_cistatic struct io_context *alloc_io_context(gfp_t gfp_flags, int node) 23062306a36Sopenharmony_ci{ 23162306a36Sopenharmony_ci struct io_context *ioc; 23262306a36Sopenharmony_ci 23362306a36Sopenharmony_ci ioc = kmem_cache_alloc_node(iocontext_cachep, gfp_flags | __GFP_ZERO, 23462306a36Sopenharmony_ci node); 23562306a36Sopenharmony_ci if (unlikely(!ioc)) 23662306a36Sopenharmony_ci return NULL; 23762306a36Sopenharmony_ci 23862306a36Sopenharmony_ci atomic_long_set(&ioc->refcount, 1); 23962306a36Sopenharmony_ci atomic_set(&ioc->active_ref, 1); 24062306a36Sopenharmony_ci#ifdef CONFIG_BLK_ICQ 24162306a36Sopenharmony_ci spin_lock_init(&ioc->lock); 24262306a36Sopenharmony_ci INIT_RADIX_TREE(&ioc->icq_tree, GFP_ATOMIC); 24362306a36Sopenharmony_ci INIT_HLIST_HEAD(&ioc->icq_list); 24462306a36Sopenharmony_ci INIT_WORK(&ioc->release_work, ioc_release_fn); 24562306a36Sopenharmony_ci#endif 24662306a36Sopenharmony_ci ioc->ioprio = IOPRIO_DEFAULT; 24762306a36Sopenharmony_ci 24862306a36Sopenharmony_ci return ioc; 24962306a36Sopenharmony_ci} 25062306a36Sopenharmony_ci 25162306a36Sopenharmony_ciint set_task_ioprio(struct task_struct *task, int ioprio) 25262306a36Sopenharmony_ci{ 25362306a36Sopenharmony_ci int err; 25462306a36Sopenharmony_ci const struct cred *cred = current_cred(), *tcred; 25562306a36Sopenharmony_ci 25662306a36Sopenharmony_ci rcu_read_lock(); 25762306a36Sopenharmony_ci tcred = __task_cred(task); 25862306a36Sopenharmony_ci if (!uid_eq(tcred->uid, cred->euid) && 25962306a36Sopenharmony_ci !uid_eq(tcred->uid, cred->uid) && !capable(CAP_SYS_NICE)) { 26062306a36Sopenharmony_ci rcu_read_unlock(); 26162306a36Sopenharmony_ci return -EPERM; 26262306a36Sopenharmony_ci } 26362306a36Sopenharmony_ci rcu_read_unlock(); 26462306a36Sopenharmony_ci 26562306a36Sopenharmony_ci err = security_task_setioprio(task, ioprio); 26662306a36Sopenharmony_ci if (err) 26762306a36Sopenharmony_ci return err; 26862306a36Sopenharmony_ci 26962306a36Sopenharmony_ci task_lock(task); 27062306a36Sopenharmony_ci if (unlikely(!task->io_context)) { 27162306a36Sopenharmony_ci struct io_context *ioc; 27262306a36Sopenharmony_ci 27362306a36Sopenharmony_ci task_unlock(task); 27462306a36Sopenharmony_ci 27562306a36Sopenharmony_ci ioc = alloc_io_context(GFP_ATOMIC, NUMA_NO_NODE); 27662306a36Sopenharmony_ci if (!ioc) 27762306a36Sopenharmony_ci return -ENOMEM; 27862306a36Sopenharmony_ci 27962306a36Sopenharmony_ci task_lock(task); 28062306a36Sopenharmony_ci if (task->flags & PF_EXITING) { 28162306a36Sopenharmony_ci kmem_cache_free(iocontext_cachep, ioc); 28262306a36Sopenharmony_ci goto out; 28362306a36Sopenharmony_ci } 28462306a36Sopenharmony_ci if (task->io_context) 28562306a36Sopenharmony_ci kmem_cache_free(iocontext_cachep, ioc); 28662306a36Sopenharmony_ci else 28762306a36Sopenharmony_ci task->io_context = ioc; 28862306a36Sopenharmony_ci } 28962306a36Sopenharmony_ci task->io_context->ioprio = ioprio; 29062306a36Sopenharmony_ciout: 29162306a36Sopenharmony_ci task_unlock(task); 29262306a36Sopenharmony_ci return 0; 29362306a36Sopenharmony_ci} 29462306a36Sopenharmony_ciEXPORT_SYMBOL_GPL(set_task_ioprio); 29562306a36Sopenharmony_ci 29662306a36Sopenharmony_ciint __copy_io(unsigned long clone_flags, struct task_struct *tsk) 29762306a36Sopenharmony_ci{ 29862306a36Sopenharmony_ci struct io_context *ioc = current->io_context; 29962306a36Sopenharmony_ci 30062306a36Sopenharmony_ci /* 30162306a36Sopenharmony_ci * Share io context with parent, if CLONE_IO is set 30262306a36Sopenharmony_ci */ 30362306a36Sopenharmony_ci if (clone_flags & CLONE_IO) { 30462306a36Sopenharmony_ci atomic_inc(&ioc->active_ref); 30562306a36Sopenharmony_ci tsk->io_context = ioc; 30662306a36Sopenharmony_ci } else if (ioprio_valid(ioc->ioprio)) { 30762306a36Sopenharmony_ci tsk->io_context = alloc_io_context(GFP_KERNEL, NUMA_NO_NODE); 30862306a36Sopenharmony_ci if (!tsk->io_context) 30962306a36Sopenharmony_ci return -ENOMEM; 31062306a36Sopenharmony_ci tsk->io_context->ioprio = ioc->ioprio; 31162306a36Sopenharmony_ci } 31262306a36Sopenharmony_ci 31362306a36Sopenharmony_ci return 0; 31462306a36Sopenharmony_ci} 31562306a36Sopenharmony_ci 31662306a36Sopenharmony_ci#ifdef CONFIG_BLK_ICQ 31762306a36Sopenharmony_ci/** 31862306a36Sopenharmony_ci * ioc_lookup_icq - lookup io_cq from ioc 31962306a36Sopenharmony_ci * @q: the associated request_queue 32062306a36Sopenharmony_ci * 32162306a36Sopenharmony_ci * Look up io_cq associated with @ioc - @q pair from @ioc. Must be called 32262306a36Sopenharmony_ci * with @q->queue_lock held. 32362306a36Sopenharmony_ci */ 32462306a36Sopenharmony_cistruct io_cq *ioc_lookup_icq(struct request_queue *q) 32562306a36Sopenharmony_ci{ 32662306a36Sopenharmony_ci struct io_context *ioc = current->io_context; 32762306a36Sopenharmony_ci struct io_cq *icq; 32862306a36Sopenharmony_ci 32962306a36Sopenharmony_ci lockdep_assert_held(&q->queue_lock); 33062306a36Sopenharmony_ci 33162306a36Sopenharmony_ci /* 33262306a36Sopenharmony_ci * icq's are indexed from @ioc using radix tree and hint pointer, 33362306a36Sopenharmony_ci * both of which are protected with RCU. All removals are done 33462306a36Sopenharmony_ci * holding both q and ioc locks, and we're holding q lock - if we 33562306a36Sopenharmony_ci * find a icq which points to us, it's guaranteed to be valid. 33662306a36Sopenharmony_ci */ 33762306a36Sopenharmony_ci rcu_read_lock(); 33862306a36Sopenharmony_ci icq = rcu_dereference(ioc->icq_hint); 33962306a36Sopenharmony_ci if (icq && icq->q == q) 34062306a36Sopenharmony_ci goto out; 34162306a36Sopenharmony_ci 34262306a36Sopenharmony_ci icq = radix_tree_lookup(&ioc->icq_tree, q->id); 34362306a36Sopenharmony_ci if (icq && icq->q == q) 34462306a36Sopenharmony_ci rcu_assign_pointer(ioc->icq_hint, icq); /* allowed to race */ 34562306a36Sopenharmony_ci else 34662306a36Sopenharmony_ci icq = NULL; 34762306a36Sopenharmony_ciout: 34862306a36Sopenharmony_ci rcu_read_unlock(); 34962306a36Sopenharmony_ci return icq; 35062306a36Sopenharmony_ci} 35162306a36Sopenharmony_ciEXPORT_SYMBOL(ioc_lookup_icq); 35262306a36Sopenharmony_ci 35362306a36Sopenharmony_ci/** 35462306a36Sopenharmony_ci * ioc_create_icq - create and link io_cq 35562306a36Sopenharmony_ci * @q: request_queue of interest 35662306a36Sopenharmony_ci * 35762306a36Sopenharmony_ci * Make sure io_cq linking @ioc and @q exists. If icq doesn't exist, they 35862306a36Sopenharmony_ci * will be created using @gfp_mask. 35962306a36Sopenharmony_ci * 36062306a36Sopenharmony_ci * The caller is responsible for ensuring @ioc won't go away and @q is 36162306a36Sopenharmony_ci * alive and will stay alive until this function returns. 36262306a36Sopenharmony_ci */ 36362306a36Sopenharmony_cistatic struct io_cq *ioc_create_icq(struct request_queue *q) 36462306a36Sopenharmony_ci{ 36562306a36Sopenharmony_ci struct io_context *ioc = current->io_context; 36662306a36Sopenharmony_ci struct elevator_type *et = q->elevator->type; 36762306a36Sopenharmony_ci struct io_cq *icq; 36862306a36Sopenharmony_ci 36962306a36Sopenharmony_ci /* allocate stuff */ 37062306a36Sopenharmony_ci icq = kmem_cache_alloc_node(et->icq_cache, GFP_ATOMIC | __GFP_ZERO, 37162306a36Sopenharmony_ci q->node); 37262306a36Sopenharmony_ci if (!icq) 37362306a36Sopenharmony_ci return NULL; 37462306a36Sopenharmony_ci 37562306a36Sopenharmony_ci if (radix_tree_maybe_preload(GFP_ATOMIC) < 0) { 37662306a36Sopenharmony_ci kmem_cache_free(et->icq_cache, icq); 37762306a36Sopenharmony_ci return NULL; 37862306a36Sopenharmony_ci } 37962306a36Sopenharmony_ci 38062306a36Sopenharmony_ci icq->ioc = ioc; 38162306a36Sopenharmony_ci icq->q = q; 38262306a36Sopenharmony_ci INIT_LIST_HEAD(&icq->q_node); 38362306a36Sopenharmony_ci INIT_HLIST_NODE(&icq->ioc_node); 38462306a36Sopenharmony_ci 38562306a36Sopenharmony_ci /* lock both q and ioc and try to link @icq */ 38662306a36Sopenharmony_ci spin_lock_irq(&q->queue_lock); 38762306a36Sopenharmony_ci spin_lock(&ioc->lock); 38862306a36Sopenharmony_ci 38962306a36Sopenharmony_ci if (likely(!radix_tree_insert(&ioc->icq_tree, q->id, icq))) { 39062306a36Sopenharmony_ci hlist_add_head(&icq->ioc_node, &ioc->icq_list); 39162306a36Sopenharmony_ci list_add(&icq->q_node, &q->icq_list); 39262306a36Sopenharmony_ci if (et->ops.init_icq) 39362306a36Sopenharmony_ci et->ops.init_icq(icq); 39462306a36Sopenharmony_ci } else { 39562306a36Sopenharmony_ci kmem_cache_free(et->icq_cache, icq); 39662306a36Sopenharmony_ci icq = ioc_lookup_icq(q); 39762306a36Sopenharmony_ci if (!icq) 39862306a36Sopenharmony_ci printk(KERN_ERR "cfq: icq link failed!\n"); 39962306a36Sopenharmony_ci } 40062306a36Sopenharmony_ci 40162306a36Sopenharmony_ci spin_unlock(&ioc->lock); 40262306a36Sopenharmony_ci spin_unlock_irq(&q->queue_lock); 40362306a36Sopenharmony_ci radix_tree_preload_end(); 40462306a36Sopenharmony_ci return icq; 40562306a36Sopenharmony_ci} 40662306a36Sopenharmony_ci 40762306a36Sopenharmony_cistruct io_cq *ioc_find_get_icq(struct request_queue *q) 40862306a36Sopenharmony_ci{ 40962306a36Sopenharmony_ci struct io_context *ioc = current->io_context; 41062306a36Sopenharmony_ci struct io_cq *icq = NULL; 41162306a36Sopenharmony_ci 41262306a36Sopenharmony_ci if (unlikely(!ioc)) { 41362306a36Sopenharmony_ci ioc = alloc_io_context(GFP_ATOMIC, q->node); 41462306a36Sopenharmony_ci if (!ioc) 41562306a36Sopenharmony_ci return NULL; 41662306a36Sopenharmony_ci 41762306a36Sopenharmony_ci task_lock(current); 41862306a36Sopenharmony_ci if (current->io_context) { 41962306a36Sopenharmony_ci kmem_cache_free(iocontext_cachep, ioc); 42062306a36Sopenharmony_ci ioc = current->io_context; 42162306a36Sopenharmony_ci } else { 42262306a36Sopenharmony_ci current->io_context = ioc; 42362306a36Sopenharmony_ci } 42462306a36Sopenharmony_ci 42562306a36Sopenharmony_ci get_io_context(ioc); 42662306a36Sopenharmony_ci task_unlock(current); 42762306a36Sopenharmony_ci } else { 42862306a36Sopenharmony_ci get_io_context(ioc); 42962306a36Sopenharmony_ci 43062306a36Sopenharmony_ci spin_lock_irq(&q->queue_lock); 43162306a36Sopenharmony_ci icq = ioc_lookup_icq(q); 43262306a36Sopenharmony_ci spin_unlock_irq(&q->queue_lock); 43362306a36Sopenharmony_ci } 43462306a36Sopenharmony_ci 43562306a36Sopenharmony_ci if (!icq) { 43662306a36Sopenharmony_ci icq = ioc_create_icq(q); 43762306a36Sopenharmony_ci if (!icq) { 43862306a36Sopenharmony_ci put_io_context(ioc); 43962306a36Sopenharmony_ci return NULL; 44062306a36Sopenharmony_ci } 44162306a36Sopenharmony_ci } 44262306a36Sopenharmony_ci return icq; 44362306a36Sopenharmony_ci} 44462306a36Sopenharmony_ciEXPORT_SYMBOL_GPL(ioc_find_get_icq); 44562306a36Sopenharmony_ci#endif /* CONFIG_BLK_ICQ */ 44662306a36Sopenharmony_ci 44762306a36Sopenharmony_cistatic int __init blk_ioc_init(void) 44862306a36Sopenharmony_ci{ 44962306a36Sopenharmony_ci iocontext_cachep = kmem_cache_create("blkdev_ioc", 45062306a36Sopenharmony_ci sizeof(struct io_context), 0, SLAB_PANIC, NULL); 45162306a36Sopenharmony_ci return 0; 45262306a36Sopenharmony_ci} 45362306a36Sopenharmony_cisubsys_initcall(blk_ioc_init); 454