Blame view
drivers/md/dm-rq.h
1.62 KB
4cc96131a dm: move request-... |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 |
/* * Internal header file for device mapper * * Copyright (C) 2016 Red Hat, Inc. All rights reserved. * * This file is released under the LGPL. */ #ifndef DM_RQ_INTERNAL_H #define DM_RQ_INTERNAL_H #include <linux/bio.h> #include <linux/kthread.h> #include "dm-stats.h" struct mapped_device; /* * One of these is allocated per request. */ struct dm_rq_target_io { struct mapped_device *md; struct dm_target *ti; struct request *orig, *clone; struct kthread_work work; int error; union map_info info; struct dm_stats_aux stats_aux; unsigned long duration_jiffies; unsigned n_sectors; }; /* * For request-based dm - the bio clones we allocate are embedded in these * structs. * * We allocate these with bio_alloc_bioset, using the front_pad parameter when * the bioset is created - this means the bio has to come at the end of the * struct. */ struct dm_rq_clone_bio_info { struct bio *orig; struct dm_rq_target_io *tio; struct bio clone; }; bool dm_use_blk_mq_default(void); bool dm_use_blk_mq(struct mapped_device *md); int dm_old_init_request_queue(struct mapped_device *md); |
e83068a5f dm mpath: add opt... |
52 |
int dm_mq_init_request_queue(struct mapped_device *md, struct dm_table *t); |
4cc96131a dm: move request-... |
53 54 55 56 |
void dm_mq_cleanup_mapped_device(struct mapped_device *md); void dm_start_queue(struct request_queue *q); void dm_stop_queue(struct request_queue *q); |
e0c107526 dm rq: introduce ... |
57 |
void dm_mq_kick_requeue_list(struct mapped_device *md); |
4cc96131a dm: move request-... |
58 59 60 61 62 63 64 |
unsigned dm_get_reserved_rq_based_ios(void); ssize_t dm_attr_rq_based_seq_io_merge_deadline_show(struct mapped_device *md, char *buf); ssize_t dm_attr_rq_based_seq_io_merge_deadline_store(struct mapped_device *md, const char *buf, size_t count); #endif |