blob: 6f59fbad93d9b19fc56dca824b7aa8a4b9241cf3 [file] [log] [blame]
Jens Axboe86db1e22008-01-29 14:53:40 +01001/*
2 * Functions related to io context handling
3 */
4#include <linux/kernel.h>
5#include <linux/module.h>
6#include <linux/init.h>
7#include <linux/bio.h>
8#include <linux/blkdev.h>
9#include <linux/bootmem.h> /* for max_pfn/max_low_pfn */
Tejun Heo5a0e3ad2010-03-24 17:04:11 +090010#include <linux/slab.h>
Jens Axboe86db1e22008-01-29 14:53:40 +010011
12#include "blk.h"
13
14/*
15 * For io context allocations
16 */
17static struct kmem_cache *iocontext_cachep;
18
Tejun Heo6e736be2011-12-14 00:33:38 +010019/**
20 * get_io_context - increment reference count to io_context
21 * @ioc: io_context to get
22 *
23 * Increment reference count to @ioc.
24 */
25void get_io_context(struct io_context *ioc)
26{
27 BUG_ON(atomic_long_read(&ioc->refcount) <= 0);
28 atomic_long_inc(&ioc->refcount);
29}
30EXPORT_SYMBOL(get_io_context);
31
Jens Axboe86db1e22008-01-29 14:53:40 +010032static void cfq_dtor(struct io_context *ioc)
33{
Jens Axboeffc4e752008-02-19 10:02:29 +010034 if (!hlist_empty(&ioc->cic_list)) {
35 struct cfq_io_context *cic;
Jens Axboe86db1e22008-01-29 14:53:40 +010036
Paul Bollee2bd9672011-06-02 13:05:02 +020037 cic = hlist_entry(ioc->cic_list.first, struct cfq_io_context,
Jens Axboeffc4e752008-02-19 10:02:29 +010038 cic_list);
39 cic->dtor(ioc);
40 }
Jens Axboe86db1e22008-01-29 14:53:40 +010041}
42
Tejun Heo42ec57a2011-12-14 00:33:37 +010043/**
44 * put_io_context - put a reference of io_context
45 * @ioc: io_context to put
46 *
47 * Decrement reference count of @ioc and release it if the count reaches
48 * zero.
Jens Axboe86db1e22008-01-29 14:53:40 +010049 */
Tejun Heo42ec57a2011-12-14 00:33:37 +010050void put_io_context(struct io_context *ioc)
Jens Axboe86db1e22008-01-29 14:53:40 +010051{
52 if (ioc == NULL)
Tejun Heo42ec57a2011-12-14 00:33:37 +010053 return;
Jens Axboe86db1e22008-01-29 14:53:40 +010054
Tejun Heo42ec57a2011-12-14 00:33:37 +010055 BUG_ON(atomic_long_read(&ioc->refcount) <= 0);
Jens Axboe86db1e22008-01-29 14:53:40 +010056
Tejun Heo42ec57a2011-12-14 00:33:37 +010057 if (!atomic_long_dec_and_test(&ioc->refcount))
58 return;
Jens Axboe86db1e22008-01-29 14:53:40 +010059
Tejun Heo42ec57a2011-12-14 00:33:37 +010060 rcu_read_lock();
61 cfq_dtor(ioc);
62 rcu_read_unlock();
63
64 kmem_cache_free(iocontext_cachep, ioc);
Jens Axboe86db1e22008-01-29 14:53:40 +010065}
66EXPORT_SYMBOL(put_io_context);
67
68static void cfq_exit(struct io_context *ioc)
69{
Jens Axboe86db1e22008-01-29 14:53:40 +010070 rcu_read_lock();
Jens Axboe86db1e22008-01-29 14:53:40 +010071
Jens Axboeffc4e752008-02-19 10:02:29 +010072 if (!hlist_empty(&ioc->cic_list)) {
73 struct cfq_io_context *cic;
74
Paul Bollee2bd9672011-06-02 13:05:02 +020075 cic = hlist_entry(ioc->cic_list.first, struct cfq_io_context,
Jens Axboeffc4e752008-02-19 10:02:29 +010076 cic_list);
77 cic->exit(ioc);
78 }
79 rcu_read_unlock();
Jens Axboe86db1e22008-01-29 14:53:40 +010080}
81
Bart Van Assche27667c92010-12-21 15:07:45 +010082/* Called by the exiting task */
Louis Rillingb69f2292009-12-04 14:52:42 +010083void exit_io_context(struct task_struct *task)
Jens Axboe86db1e22008-01-29 14:53:40 +010084{
85 struct io_context *ioc;
86
Tejun Heo6e736be2011-12-14 00:33:38 +010087 /* PF_EXITING prevents new io_context from being attached to @task */
88 WARN_ON_ONCE(!(current->flags & PF_EXITING));
89
Louis Rillingb69f2292009-12-04 14:52:42 +010090 task_lock(task);
91 ioc = task->io_context;
92 task->io_context = NULL;
93 task_unlock(task);
Jens Axboe86db1e22008-01-29 14:53:40 +010094
Bart Van Assche27667c92010-12-21 15:07:45 +010095 if (atomic_dec_and_test(&ioc->nr_tasks))
Jens Axboe86db1e22008-01-29 14:53:40 +010096 cfq_exit(ioc);
97
Louis Rilling61cc74f2009-12-04 14:52:41 +010098 put_io_context(ioc);
Jens Axboe86db1e22008-01-29 14:53:40 +010099}
100
Tejun Heo6e736be2011-12-14 00:33:38 +0100101static struct io_context *create_task_io_context(struct task_struct *task,
102 gfp_t gfp_flags, int node,
103 bool take_ref)
Jens Axboe86db1e22008-01-29 14:53:40 +0100104{
Paul Bolledf415652011-06-06 05:11:34 +0200105 struct io_context *ioc;
Jens Axboe86db1e22008-01-29 14:53:40 +0100106
Tejun Heo42ec57a2011-12-14 00:33:37 +0100107 ioc = kmem_cache_alloc_node(iocontext_cachep, gfp_flags | __GFP_ZERO,
108 node);
109 if (unlikely(!ioc))
110 return NULL;
111
112 /* initialize */
113 atomic_long_set(&ioc->refcount, 1);
114 atomic_set(&ioc->nr_tasks, 1);
115 spin_lock_init(&ioc->lock);
116 INIT_RADIX_TREE(&ioc->radix_root, GFP_ATOMIC | __GFP_HIGH);
117 INIT_HLIST_HEAD(&ioc->cic_list);
Jens Axboe86db1e22008-01-29 14:53:40 +0100118
Tejun Heo6e736be2011-12-14 00:33:38 +0100119 /* try to install, somebody might already have beaten us to it */
120 task_lock(task);
121
122 if (!task->io_context && !(task->flags & PF_EXITING)) {
123 task->io_context = ioc;
124 } else {
125 kmem_cache_free(iocontext_cachep, ioc);
126 ioc = task->io_context;
127 }
128
129 if (ioc && take_ref)
130 get_io_context(ioc);
131
132 task_unlock(task);
Paul Bolledf415652011-06-06 05:11:34 +0200133 return ioc;
Jens Axboe86db1e22008-01-29 14:53:40 +0100134}
135
Tejun Heo42ec57a2011-12-14 00:33:37 +0100136/**
137 * current_io_context - get io_context of %current
138 * @gfp_flags: allocation flags, used if allocation is necessary
139 * @node: allocation node, used if allocation is necessary
Jens Axboe86db1e22008-01-29 14:53:40 +0100140 *
Tejun Heo42ec57a2011-12-14 00:33:37 +0100141 * Return io_context of %current. If it doesn't exist, it is created with
142 * @gfp_flags and @node. The returned io_context does NOT have its
143 * reference count incremented. Because io_context is exited only on task
144 * exit, %current can be sure that the returned io_context is valid and
145 * alive as long as it is executing.
Jens Axboe86db1e22008-01-29 14:53:40 +0100146 */
147struct io_context *current_io_context(gfp_t gfp_flags, int node)
148{
Tejun Heo6e736be2011-12-14 00:33:38 +0100149 might_sleep_if(gfp_flags & __GFP_WAIT);
Jens Axboe86db1e22008-01-29 14:53:40 +0100150
Tejun Heo6e736be2011-12-14 00:33:38 +0100151 if (current->io_context)
152 return current->io_context;
Jens Axboe86db1e22008-01-29 14:53:40 +0100153
Tejun Heo6e736be2011-12-14 00:33:38 +0100154 return create_task_io_context(current, gfp_flags, node, false);
Jens Axboe86db1e22008-01-29 14:53:40 +0100155}
Tejun Heo6e736be2011-12-14 00:33:38 +0100156EXPORT_SYMBOL(current_io_context);
Jens Axboe86db1e22008-01-29 14:53:40 +0100157
Tejun Heo6e736be2011-12-14 00:33:38 +0100158/**
159 * get_task_io_context - get io_context of a task
160 * @task: task of interest
161 * @gfp_flags: allocation flags, used if allocation is necessary
162 * @node: allocation node, used if allocation is necessary
Jens Axboe86db1e22008-01-29 14:53:40 +0100163 *
Tejun Heo6e736be2011-12-14 00:33:38 +0100164 * Return io_context of @task. If it doesn't exist, it is created with
165 * @gfp_flags and @node. The returned io_context has its reference count
166 * incremented.
167 *
168 * This function always goes through task_lock() and it's better to use
169 * current_io_context() + get_io_context() for %current.
Jens Axboe86db1e22008-01-29 14:53:40 +0100170 */
Tejun Heo6e736be2011-12-14 00:33:38 +0100171struct io_context *get_task_io_context(struct task_struct *task,
172 gfp_t gfp_flags, int node)
Jens Axboe86db1e22008-01-29 14:53:40 +0100173{
Tejun Heo6e736be2011-12-14 00:33:38 +0100174 struct io_context *ioc;
Jens Axboe86db1e22008-01-29 14:53:40 +0100175
Tejun Heo6e736be2011-12-14 00:33:38 +0100176 might_sleep_if(gfp_flags & __GFP_WAIT);
Jens Axboe86db1e22008-01-29 14:53:40 +0100177
Tejun Heo6e736be2011-12-14 00:33:38 +0100178 task_lock(task);
179 ioc = task->io_context;
180 if (likely(ioc)) {
181 get_io_context(ioc);
182 task_unlock(task);
183 return ioc;
184 }
185 task_unlock(task);
186
187 return create_task_io_context(task, gfp_flags, node, true);
Jens Axboe86db1e22008-01-29 14:53:40 +0100188}
Tejun Heo6e736be2011-12-14 00:33:38 +0100189EXPORT_SYMBOL(get_task_io_context);
Jens Axboe86db1e22008-01-29 14:53:40 +0100190
Tejun Heodc869002011-12-14 00:33:38 +0100191void ioc_set_changed(struct io_context *ioc, int which)
192{
193 struct cfq_io_context *cic;
194 struct hlist_node *n;
195
196 hlist_for_each_entry(cic, n, &ioc->cic_list, cic_list)
197 set_bit(which, &cic->changed);
198}
199
200/**
201 * ioc_ioprio_changed - notify ioprio change
202 * @ioc: io_context of interest
203 * @ioprio: new ioprio
204 *
205 * @ioc's ioprio has changed to @ioprio. Set %CIC_IOPRIO_CHANGED for all
206 * cic's. iosched is responsible for checking the bit and applying it on
207 * request issue path.
208 */
209void ioc_ioprio_changed(struct io_context *ioc, int ioprio)
210{
211 unsigned long flags;
212
213 spin_lock_irqsave(&ioc->lock, flags);
214 ioc->ioprio = ioprio;
215 ioc_set_changed(ioc, CIC_IOPRIO_CHANGED);
216 spin_unlock_irqrestore(&ioc->lock, flags);
217}
218
219/**
220 * ioc_cgroup_changed - notify cgroup change
221 * @ioc: io_context of interest
222 *
223 * @ioc's cgroup has changed. Set %CIC_CGROUP_CHANGED for all cic's.
224 * iosched is responsible for checking the bit and applying it on request
225 * issue path.
226 */
227void ioc_cgroup_changed(struct io_context *ioc)
228{
229 unsigned long flags;
230
231 spin_lock_irqsave(&ioc->lock, flags);
232 ioc_set_changed(ioc, CIC_CGROUP_CHANGED);
233 spin_unlock_irqrestore(&ioc->lock, flags);
234}
235
Adrian Bunk13341592008-02-18 13:45:53 +0100236static int __init blk_ioc_init(void)
Jens Axboe86db1e22008-01-29 14:53:40 +0100237{
238 iocontext_cachep = kmem_cache_create("blkdev_ioc",
239 sizeof(struct io_context), 0, SLAB_PANIC, NULL);
240 return 0;
241}
242subsys_initcall(blk_ioc_init);