Blame view

block/blk.h 10.5 KB
b24413180   Greg Kroah-Hartman   License cleanup: ...
1
  /* SPDX-License-Identifier: GPL-2.0 */
8324aa91d   Jens Axboe   block: split tag ...
2
3
  #ifndef BLK_INTERNAL_H
  #define BLK_INTERNAL_H
a73f730d0   Tejun Heo   block, cfq: move ...
4
  #include <linux/idr.h>
f70ced091   Ming Lei   blk-mq: support p...
5
  #include <linux/blk-mq.h>
c39ae60df   Christoph Hellwig   block: remove ARC...
6
  #include <xen/xen.h>
f70ced091   Ming Lei   blk-mq: support p...
7
  #include "blk-mq.h"
c3e221921   Ming Lei   block: free sched...
8
  #include "blk-mq-sched.h"
a73f730d0   Tejun Heo   block, cfq: move ...
9

0d2602ca3   Jens Axboe   blk-mq: improve s...
10
11
  /* Max future timer expiry for timeouts */
  #define BLK_MAX_TIMEOUT		(5 * HZ)
18fbda91c   Omar Sandoval   block: use same b...
12
13
14
  #ifdef CONFIG_DEBUG_FS
  extern struct dentry *blk_debugfs_root;
  #endif
7c94e1c15   Ming Lei   block: introduce ...
15
16
17
18
19
20
21
22
  struct blk_flush_queue {
  	unsigned int		flush_queue_delayed:1;
  	unsigned int		flush_pending_idx:1;
  	unsigned int		flush_running_idx:1;
  	unsigned long		flush_pending_since;
  	struct list_head	flush_queue[2];
  	struct list_head	flush_data_in_flight;
  	struct request		*flush_rq;
0048b4837   Ming Lei   blk-mq: fix race ...
23
24
25
26
27
28
  
  	/*
  	 * flush_rq shares tag with this rq, both can't be active
  	 * at the same time
  	 */
  	struct request		*orig_rq;
7c94e1c15   Ming Lei   block: introduce ...
29
30
  	spinlock_t		mq_flush_lock;
  };
8324aa91d   Jens Axboe   block: split tag ...
31
32
  extern struct kmem_cache *blk_requestq_cachep;
  extern struct kobj_type blk_queue_ktype;
a73f730d0   Tejun Heo   block, cfq: move ...
33
  extern struct ida blk_queue_ida;
8324aa91d   Jens Axboe   block: split tag ...
34

f9afca4d3   Jens Axboe   blk-mq: pass in r...
35
36
  static inline struct blk_flush_queue *
  blk_get_flush_queue(struct request_queue *q, struct blk_mq_ctx *ctx)
7c94e1c15   Ming Lei   block: introduce ...
37
  {
8ccdf4a37   Jianchao Wang   blk-mq: save queu...
38
  	return blk_mq_map_queue(q, REQ_OP_FLUSH, ctx)->fq;
7c94e1c15   Ming Lei   block: introduce ...
39
  }
09ac46c42   Tejun Heo   block: misc updat...
40
41
42
43
  static inline void __blk_get_queue(struct request_queue *q)
  {
  	kobject_get(&q->kobj);
  }
f70ced091   Ming Lei   blk-mq: support p...
44
  struct blk_flush_queue *blk_alloc_flush_queue(struct request_queue *q,
5b202853f   Jianchao Wang   blk-mq: change gf...
45
  		int node, int cmd_size, gfp_t flags);
f70ced091   Ming Lei   blk-mq: support p...
46
  void blk_free_flush_queue(struct blk_flush_queue *q);
f35526557   Ming Lei   block: introduce ...
47

3ef28e83a   Dan Williams   block: generic re...
48
49
50
51
52
53
54
55
56
57
58
59
  void blk_freeze_queue(struct request_queue *q);
  
  static inline void blk_queue_enter_live(struct request_queue *q)
  {
  	/*
  	 * Given that running in generic_make_request() context
  	 * guarantees that a live reference against q_usage_counter has
  	 * been established, further references under that same context
  	 * need not check that the queue has been frozen (marked dead).
  	 */
  	percpu_ref_get(&q->q_usage_counter);
  }
8324aa91d   Jens Axboe   block: split tag ...
60

3dccdae54   Christoph Hellwig   block: merge BIOV...
61
62
  static inline bool biovec_phys_mergeable(struct request_queue *q,
  		struct bio_vec *vec1, struct bio_vec *vec2)
6a9f5f240   Christoph Hellwig   block: simplify B...
63
  {
3dccdae54   Christoph Hellwig   block: merge BIOV...
64
  	unsigned long mask = queue_segment_boundary(q);
6e768461c   Christoph Hellwig   block: remove bve...
65
66
  	phys_addr_t addr1 = page_to_phys(vec1->bv_page) + vec1->bv_offset;
  	phys_addr_t addr2 = page_to_phys(vec2->bv_page) + vec2->bv_offset;
3dccdae54   Christoph Hellwig   block: merge BIOV...
67
68
  
  	if (addr1 + vec1->bv_len != addr2)
6a9f5f240   Christoph Hellwig   block: simplify B...
69
  		return false;
0383ad437   Ming Lei   block: pass page ...
70
  	if (xen_domain() && !xen_biovec_phys_mergeable(vec1, vec2->bv_page))
6a9f5f240   Christoph Hellwig   block: simplify B...
71
  		return false;
3dccdae54   Christoph Hellwig   block: merge BIOV...
72
73
  	if ((addr1 | mask) != ((addr2 + vec2->bv_len - 1) | mask))
  		return false;
6a9f5f240   Christoph Hellwig   block: simplify B...
74
75
  	return true;
  }
27ca1d4ed   Christoph Hellwig   block: move req_g...
76
77
78
  static inline bool __bvec_gap_to_prev(struct request_queue *q,
  		struct bio_vec *bprv, unsigned int offset)
  {
df376b2ed   Johannes Thumshirn   block: respect vi...
79
  	return (offset & queue_virt_boundary(q)) ||
27ca1d4ed   Christoph Hellwig   block: move req_g...
80
81
82
83
84
85
86
87
88
89
90
91
92
93
  		((bprv->bv_offset + bprv->bv_len) & queue_virt_boundary(q));
  }
  
  /*
   * Check if adding a bio_vec after bprv with offset would create a gap in
   * the SG list. Most drivers don't care about this, but some do.
   */
  static inline bool bvec_gap_to_prev(struct request_queue *q,
  		struct bio_vec *bprv, unsigned int offset)
  {
  	if (!queue_virt_boundary(q))
  		return false;
  	return __bvec_gap_to_prev(q, bprv, offset);
  }
1aa0a133f   Christoph Hellwig   block: mark blk_r...
94
95
96
97
98
99
100
101
102
103
104
  static inline void blk_rq_bio_prep(struct request *rq, struct bio *bio,
  		unsigned int nr_segs)
  {
  	rq->nr_phys_segments = nr_segs;
  	rq->__data_len = bio->bi_iter.bi_size;
  	rq->bio = rq->biotail = bio;
  	rq->ioprio = bio_prio(bio);
  
  	if (bio->bi_disk)
  		rq->rq_disk = bio->bi_disk;
  }
5a48fc147   Dan Williams   block: blk_flush_...
105
106
  #ifdef CONFIG_BLK_DEV_INTEGRITY
  void blk_flush_integrity(void);
7c20f1168   Christoph Hellwig   bio-integrity: st...
107
108
109
110
111
112
113
  bool __bio_integrity_endio(struct bio *);
  static inline bool bio_integrity_endio(struct bio *bio)
  {
  	if (bio_integrity(bio))
  		return __bio_integrity_endio(bio);
  	return true;
  }
43b729bfe   Christoph Hellwig   block: move integ...
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
  
  static inline bool integrity_req_gap_back_merge(struct request *req,
  		struct bio *next)
  {
  	struct bio_integrity_payload *bip = bio_integrity(req->bio);
  	struct bio_integrity_payload *bip_next = bio_integrity(next);
  
  	return bvec_gap_to_prev(req->q, &bip->bip_vec[bip->bip_vcnt - 1],
  				bip_next->bip_vec[0].bv_offset);
  }
  
  static inline bool integrity_req_gap_front_merge(struct request *req,
  		struct bio *bio)
  {
  	struct bio_integrity_payload *bip = bio_integrity(bio);
  	struct bio_integrity_payload *bip_next = bio_integrity(req->bio);
  
  	return bvec_gap_to_prev(req->q, &bip->bip_vec[bip->bip_vcnt - 1],
  				bip_next->bip_vec[0].bv_offset);
  }
  #else /* CONFIG_BLK_DEV_INTEGRITY */
  static inline bool integrity_req_gap_back_merge(struct request *req,
  		struct bio *next)
  {
  	return false;
  }
  static inline bool integrity_req_gap_front_merge(struct request *req,
  		struct bio *bio)
  {
  	return false;
  }
5a48fc147   Dan Williams   block: blk_flush_...
145
146
147
  static inline void blk_flush_integrity(void)
  {
  }
7c20f1168   Christoph Hellwig   bio-integrity: st...
148
149
150
151
  static inline bool bio_integrity_endio(struct bio *bio)
  {
  	return true;
  }
43b729bfe   Christoph Hellwig   block: move integ...
152
  #endif /* CONFIG_BLK_DEV_INTEGRITY */
8324aa91d   Jens Axboe   block: split tag ...
153

0d2602ca3   Jens Axboe   blk-mq: improve s...
154
  unsigned long blk_rq_timeout(unsigned long timeout);
87ee7b112   Jens Axboe   blk-mq: fix race ...
155
  void blk_add_timer(struct request *req);
320ae51fe   Jens Axboe   blk-mq: new multi...
156

14ccb66b3   Christoph Hellwig   block: remove the...
157
158
159
160
  bool bio_attempt_front_merge(struct request *req, struct bio *bio,
  		unsigned int nr_segs);
  bool bio_attempt_back_merge(struct request *req, struct bio *bio,
  		unsigned int nr_segs);
1e739730c   Christoph Hellwig   block: optionally...
161
162
  bool bio_attempt_discard_merge(struct request_queue *q, struct request *req,
  		struct bio *bio);
320ae51fe   Jens Axboe   blk-mq: new multi...
163
  bool blk_attempt_plug_merge(struct request_queue *q, struct bio *bio,
14ccb66b3   Christoph Hellwig   block: remove the...
164
  		unsigned int nr_segs, struct request **same_queue_rq);
320ae51fe   Jens Axboe   blk-mq: new multi...
165
166
167
  
  void blk_account_io_start(struct request *req, bool new_io);
  void blk_account_io_completion(struct request *req, unsigned int bytes);
522a77756   Omar Sandoval   block: consolidat...
168
  void blk_account_io_done(struct request *req, u64 now);
320ae51fe   Jens Axboe   blk-mq: new multi...
169

242f9dcb8   Jens Axboe   block: unify requ...
170
  /*
158dbda00   Tejun Heo   block: reorganize...
171
172
   * Internal elevator interface
   */
e80640213   Christoph Hellwig   block: split out ...
173
  #define ELV_ON_HASH(rq) ((rq)->rq_flags & RQF_HASHED)
158dbda00   Tejun Heo   block: reorganize...
174

ae1b15396   Tejun Heo   block: reimplemen...
175
  void blk_insert_flush(struct request *rq);
dd831006d   Tejun Heo   block: misc clean...
176

131d08e12   Christoph Hellwig   block: split the ...
177
  int elevator_init_mq(struct request_queue *q);
d48ece209   Jianchao Wang   blk-mq: init hctx...
178
179
  int elevator_switch_mq(struct request_queue *q,
  			      struct elevator_type *new_e);
c3e221921   Ming Lei   block: free sched...
180
  void __elevator_exit(struct request_queue *, struct elevator_queue *);
83d016ac8   Bart Van Assche   block: Unexport e...
181
182
  int elv_register_queue(struct request_queue *q);
  void elv_unregister_queue(struct request_queue *q);
c3e221921   Ming Lei   block: free sched...
183
184
185
186
187
188
  static inline void elevator_exit(struct request_queue *q,
  		struct elevator_queue *e)
  {
  	blk_mq_sched_free_requests(q);
  	__elevator_exit(q, e);
  }
807d4af2f   Christoph Hellwig   block: add a __di...
189
  struct hd_struct *__disk_get_part(struct gendisk *disk, int partno);
581d4e28d   Jens Axboe   block: add fault ...
190
191
192
193
194
195
196
197
198
199
200
  #ifdef CONFIG_FAIL_IO_TIMEOUT
  int blk_should_fake_timeout(struct request_queue *);
  ssize_t part_timeout_show(struct device *, struct device_attribute *, char *);
  ssize_t part_timeout_store(struct device *, struct device_attribute *,
  				const char *, size_t);
  #else
  static inline int blk_should_fake_timeout(struct request_queue *q)
  {
  	return 0;
  }
  #endif
14ccb66b3   Christoph Hellwig   block: remove the...
201
202
203
204
205
206
  void __blk_queue_split(struct request_queue *q, struct bio **bio,
  		unsigned int *nr_segs);
  int ll_back_merge_fn(struct request *req, struct bio *bio,
  		unsigned int nr_segs);
  int ll_front_merge_fn(struct request *req,  struct bio *bio,
  		unsigned int nr_segs);
b973cb7e8   Jens Axboe   blk-merge: return...
207
208
  struct request *attempt_back_merge(struct request_queue *q, struct request *rq);
  struct request *attempt_front_merge(struct request_queue *q, struct request *rq);
5e84ea3a9   Jens Axboe   block: attempt to...
209
210
  int blk_attempt_req_merge(struct request_queue *q, struct request *rq,
  				struct request *next);
e9cd19c0c   Christoph Hellwig   block: simplify b...
211
  unsigned int blk_recalc_rq_segments(struct request *rq);
80a761fd3   Tejun Heo   block: implement ...
212
  void blk_rq_set_mixed_merge(struct request *rq);
050c8ea80   Tejun Heo   block: separate o...
213
  bool blk_rq_merge_ok(struct request *rq, struct bio *bio);
34fe7c054   Christoph Hellwig   block: enumify EL...
214
  enum elv_merge blk_try_merge(struct request *rq, struct bio *bio);
d6d481969   Jens Axboe   block: ll_rw_blk....
215

ff88972c8   Adrian Bunk   proper prototype ...
216
  int blk_dev_init(void);
c2553b584   Jens Axboe   block: make blk_d...
217
218
219
220
221
  /*
   * Contribute to IO statistics IFF:
   *
   *	a) it's attached to a gendisk, and
   *	b) the queue had IO stats enabled when this request was started, and
e2a60da74   Martin K. Petersen   block: Clean up s...
222
   *	c) it's a file system request
c2553b584   Jens Axboe   block: make blk_d...
223
   */
599d067dd   Chengguang Xu   block: change ret...
224
  static inline bool blk_do_io_stat(struct request *rq)
fb8ec18c3   Jens Axboe   block: fix oops i...
225
  {
33659ebba   Christoph Hellwig   block: remove wra...
226
  	return rq->rq_disk &&
e80640213   Christoph Hellwig   block: split out ...
227
  	       (rq->rq_flags & RQF_IO_STAT) &&
57292b58d   Christoph Hellwig   block: introduce ...
228
  		!blk_rq_is_passthrough(rq);
fb8ec18c3   Jens Axboe   block: fix oops i...
229
  }
6cf7677f1   Christoph Hellwig   block: move req_s...
230
231
232
233
234
235
  static inline void req_set_nomerge(struct request_queue *q, struct request *req)
  {
  	req->cmd_flags |= REQ_NOMERGE;
  	if (req == q->last_merge)
  		q->last_merge = NULL;
  }
f2dbd76a0   Tejun Heo   block, cfq: repla...
236
  /*
1adfc5e41   Ming Lei   block: make sure ...
237
238
239
240
241
242
243
244
245
246
   * The max size one bio can handle is UINT_MAX becasue bvec_iter.bi_size
   * is defined as 'unsigned int', meantime it has to aligned to with logical
   * block size which is the minimum accepted unit by hardware.
   */
  static inline unsigned int bio_allowed_max_sectors(struct request_queue *q)
  {
  	return round_down(UINT_MAX, queue_logical_block_size(q)) >> 9;
  }
  
  /*
f2dbd76a0   Tejun Heo   block, cfq: repla...
247
248
249
   * Internal io_context interface
   */
  void get_io_context(struct io_context *ioc);
47fdd4ca9   Tejun Heo   block, cfq: move ...
250
  struct io_cq *ioc_lookup_icq(struct io_context *ioc, struct request_queue *q);
24acfc34f   Tejun Heo   block: interface ...
251
252
  struct io_cq *ioc_create_icq(struct io_context *ioc, struct request_queue *q,
  			     gfp_t gfp_mask);
7e5a87944   Tejun Heo   block, cfq: move ...
253
  void ioc_clear_queue(struct request_queue *q);
f2dbd76a0   Tejun Heo   block, cfq: repla...
254

24acfc34f   Tejun Heo   block: interface ...
255
  int create_task_io_context(struct task_struct *task, gfp_t gfp_mask, int node);
f2dbd76a0   Tejun Heo   block, cfq: repla...
256
257
258
  
  /**
   * create_io_context - try to create task->io_context
f2dbd76a0   Tejun Heo   block, cfq: repla...
259
260
261
   * @gfp_mask: allocation mask
   * @node: allocation node
   *
24acfc34f   Tejun Heo   block: interface ...
262
263
264
   * If %current->io_context is %NULL, allocate a new io_context and install
   * it.  Returns the current %current->io_context which may be %NULL if
   * allocation failed.
f2dbd76a0   Tejun Heo   block, cfq: repla...
265
266
   *
   * Note that this function can't be called with IRQ disabled because
24acfc34f   Tejun Heo   block: interface ...
267
   * task_lock which protects %current->io_context is IRQ-unsafe.
f2dbd76a0   Tejun Heo   block, cfq: repla...
268
   */
24acfc34f   Tejun Heo   block: interface ...
269
  static inline struct io_context *create_io_context(gfp_t gfp_mask, int node)
f2dbd76a0   Tejun Heo   block, cfq: repla...
270
271
  {
  	WARN_ON_ONCE(irqs_disabled());
24acfc34f   Tejun Heo   block: interface ...
272
273
274
  	if (unlikely(!current->io_context))
  		create_task_io_context(current, gfp_mask, node);
  	return current->io_context;
f2dbd76a0   Tejun Heo   block, cfq: repla...
275
276
277
278
279
  }
  
  /*
   * Internal throttling interface
   */
bc9fcbf9c   Tejun Heo   block: move blk_t...
280
  #ifdef CONFIG_BLK_DEV_THROTTLING
c9a929dde   Tejun Heo   block: fix reques...
281
  extern void blk_throtl_drain(struct request_queue *q);
bc9fcbf9c   Tejun Heo   block: move blk_t...
282
283
  extern int blk_throtl_init(struct request_queue *q);
  extern void blk_throtl_exit(struct request_queue *q);
d61fcfa4b   Shaohua Li   blk-throttle: cho...
284
  extern void blk_throtl_register_queue(struct request_queue *q);
bc9fcbf9c   Tejun Heo   block: move blk_t...
285
  #else /* CONFIG_BLK_DEV_THROTTLING */
c9a929dde   Tejun Heo   block: fix reques...
286
  static inline void blk_throtl_drain(struct request_queue *q) { }
bc9fcbf9c   Tejun Heo   block: move blk_t...
287
288
  static inline int blk_throtl_init(struct request_queue *q) { return 0; }
  static inline void blk_throtl_exit(struct request_queue *q) { }
d61fcfa4b   Shaohua Li   blk-throttle: cho...
289
  static inline void blk_throtl_register_queue(struct request_queue *q) { }
bc9fcbf9c   Tejun Heo   block: move blk_t...
290
  #endif /* CONFIG_BLK_DEV_THROTTLING */
297e3d854   Shaohua Li   blk-throttle: mak...
291
292
293
294
  #ifdef CONFIG_BLK_DEV_THROTTLING_LOW
  extern ssize_t blk_throtl_sample_time_show(struct request_queue *q, char *page);
  extern ssize_t blk_throtl_sample_time_store(struct request_queue *q,
  	const char *page, size_t count);
9e234eeaf   Shaohua Li   blk-throttle: add...
295
  extern void blk_throtl_bio_endio(struct bio *bio);
b9147dd1b   Shaohua Li   blk-throttle: add...
296
  extern void blk_throtl_stat_add(struct request *rq, u64 time);
9e234eeaf   Shaohua Li   blk-throttle: add...
297
298
  #else
  static inline void blk_throtl_bio_endio(struct bio *bio) { }
b9147dd1b   Shaohua Li   blk-throttle: add...
299
  static inline void blk_throtl_stat_add(struct request *rq, u64 time) { }
297e3d854   Shaohua Li   blk-throttle: mak...
300
  #endif
bc9fcbf9c   Tejun Heo   block: move blk_t...
301

3bce016a4   Christoph Hellwig   block: move bounc...
302
303
304
305
306
307
308
309
310
311
312
313
  #ifdef CONFIG_BOUNCE
  extern int init_emergency_isa_pool(void);
  extern void blk_queue_bounce(struct request_queue *q, struct bio **bio);
  #else
  static inline int init_emergency_isa_pool(void)
  {
  	return 0;
  }
  static inline void blk_queue_bounce(struct request_queue *q, struct bio **bio)
  {
  }
  #endif /* CONFIG_BOUNCE */
d70675121   Josef Bacik   block: introduce ...
314
315
316
317
318
  #ifdef CONFIG_BLK_CGROUP_IOLATENCY
  extern int blk_iolatency_init(struct request_queue *q);
  #else
  static inline int blk_iolatency_init(struct request_queue *q) { return 0; }
  #endif
a2d6b3a2d   Damien Le Moal   block: Improve zo...
319
  struct bio *blk_next_bio(struct bio *bio, unsigned int nr_pages, gfp_t gfp);
bf5054569   Damien Le Moal   block: Introduce ...
320
321
322
323
324
  #ifdef CONFIG_BLK_DEV_ZONED
  void blk_queue_free_zone_bitmaps(struct request_queue *q);
  #else
  static inline void blk_queue_free_zone_bitmaps(struct request_queue *q) {}
  #endif
bc9fcbf9c   Tejun Heo   block: move blk_t...
325
  #endif /* BLK_INTERNAL_H */