| #ifndef IOCONTEXT_H |
| #define IOCONTEXT_H |
| |
| #include <linux/radix-tree.h> |
| #include <linux/rcupdate.h> |
| |
| /* |
| * This is the per-process anticipatory I/O scheduler state. |
| */ |
| struct as_io_context { |
| spinlock_t lock; |
| |
| void (*dtor)(struct as_io_context *aic); /* destructor */ |
| void (*exit)(struct as_io_context *aic); /* called on task exit */ |
| |
| unsigned long state; |
| atomic_t nr_queued; /* queued reads & sync writes */ |
| atomic_t nr_dispatched; /* number of requests gone to the drivers */ |
| |
| /* IO History tracking */ |
| /* Thinktime */ |
| unsigned long last_end_request; |
| unsigned long ttime_total; |
| unsigned long ttime_samples; |
| unsigned long ttime_mean; |
| /* Layout pattern */ |
| unsigned int seek_samples; |
| sector_t last_request_pos; |
| u64 seek_total; |
| sector_t seek_mean; |
| }; |
| |
| struct cfq_queue; |
| struct cfq_io_context { |
| void *key; |
| unsigned long dead_key; |
| |
| struct cfq_queue *cfqq[2]; |
| |
| struct io_context *ioc; |
| |
| unsigned long last_end_request; |
| sector_t last_request_pos; |
| |
| unsigned long ttime_total; |
| unsigned long ttime_samples; |
| unsigned long ttime_mean; |
| |
| unsigned int seek_samples; |
| u64 seek_total; |
| sector_t seek_mean; |
| |
| struct list_head queue_list; |
| struct hlist_node cic_list; |
| |
| void (*dtor)(struct io_context *); /* destructor */ |
| void (*exit)(struct io_context *); /* called on task exit */ |
| |
| struct rcu_head rcu_head; |
| }; |
| |
| /* |
| * I/O subsystem state of the associated processes. It is refcounted |
| * and kmalloc'ed. These could be shared between processes. |
| */ |
| struct io_context { |
| atomic_t refcount; |
| atomic_t nr_tasks; |
| |
| /* all the fields below are protected by this lock */ |
| spinlock_t lock; |
| |
| unsigned short ioprio; |
| unsigned short ioprio_changed; |
| |
| /* |
| * For request batching |
| */ |
| unsigned long last_waited; /* Time last woken after wait for request */ |
| int nr_batch_requests; /* Number of requests left in the batch */ |
| |
| struct as_io_context *aic; |
| struct radix_tree_root radix_root; |
| struct hlist_head cic_list; |
| void *ioc_data; |
| }; |
| |
| static inline struct io_context *ioc_task_link(struct io_context *ioc) |
| { |
| /* |
| * if ref count is zero, don't allow sharing (ioc is going away, it's |
| * a race). |
| */ |
| if (ioc && atomic_inc_not_zero(&ioc->refcount)) { |
| atomic_inc(&ioc->nr_tasks); |
| return ioc; |
| } |
| |
| return NULL; |
| } |
| |
| #ifdef CONFIG_BLOCK |
| int put_io_context(struct io_context *ioc); |
| void exit_io_context(void); |
| struct io_context *get_io_context(gfp_t gfp_flags, int node); |
| struct io_context *alloc_io_context(gfp_t gfp_flags, int node); |
| void copy_io_context(struct io_context **pdst, struct io_context **psrc); |
| #else |
| static inline void exit_io_context(void) |
| { |
| } |
| |
| struct io_context; |
| static inline int put_io_context(struct io_context *ioc) |
| { |
| return 1; |
| } |
| #endif |
| |
| #endif |