5 #include "blk-mq-tag.h"
7 int blk_mq_sched_init_hctx_data(struct request_queue
*q
, size_t size
,
8 int (*init
)(struct blk_mq_hw_ctx
*),
9 void (*exit
)(struct blk_mq_hw_ctx
*));
11 void blk_mq_sched_free_hctx_data(struct request_queue
*q
,
12 void (*exit
)(struct blk_mq_hw_ctx
*));
14 struct request
*blk_mq_sched_get_request(struct request_queue
*q
, struct bio
*bio
, unsigned int op
, struct blk_mq_alloc_data
*data
);
15 void blk_mq_sched_put_request(struct request
*rq
);
17 void blk_mq_sched_request_inserted(struct request
*rq
);
18 bool blk_mq_sched_try_merge(struct request_queue
*q
, struct bio
*bio
,
19 struct request
**merged_request
);
20 bool __blk_mq_sched_bio_merge(struct request_queue
*q
, struct bio
*bio
);
21 bool blk_mq_sched_try_insert_merge(struct request_queue
*q
, struct request
*rq
);
22 void blk_mq_sched_restart_queues(struct blk_mq_hw_ctx
*hctx
);
24 void blk_mq_sched_insert_request(struct request
*rq
, bool at_head
,
25 bool run_queue
, bool async
, bool can_block
);
26 void blk_mq_sched_insert_requests(struct request_queue
*q
,
27 struct blk_mq_ctx
*ctx
,
28 struct list_head
*list
, bool run_queue_async
);
30 void blk_mq_sched_dispatch_requests(struct blk_mq_hw_ctx
*hctx
);
31 void blk_mq_sched_move_to_dispatch(struct blk_mq_hw_ctx
*hctx
,
32 struct list_head
*rq_list
,
33 struct request
*(*get_rq
)(struct blk_mq_hw_ctx
*));
35 int blk_mq_sched_setup(struct request_queue
*q
);
36 void blk_mq_sched_teardown(struct request_queue
*q
);
38 int blk_mq_sched_init(struct request_queue
*q
);
41 blk_mq_sched_bio_merge(struct request_queue
*q
, struct bio
*bio
)
43 struct elevator_queue
*e
= q
->elevator
;
45 if (!e
|| blk_queue_nomerges(q
) || !bio_mergeable(bio
))
48 return __blk_mq_sched_bio_merge(q
, bio
);
51 static inline int blk_mq_sched_get_rq_priv(struct request_queue
*q
,
55 struct elevator_queue
*e
= q
->elevator
;
57 if (e
&& e
->type
->ops
.mq
.get_rq_priv
)
58 return e
->type
->ops
.mq
.get_rq_priv(q
, rq
, bio
);
63 static inline void blk_mq_sched_put_rq_priv(struct request_queue
*q
,
66 struct elevator_queue
*e
= q
->elevator
;
68 if (e
&& e
->type
->ops
.mq
.put_rq_priv
)
69 e
->type
->ops
.mq
.put_rq_priv(q
, rq
);
73 blk_mq_sched_allow_merge(struct request_queue
*q
, struct request
*rq
,
76 struct elevator_queue
*e
= q
->elevator
;
78 if (e
&& e
->type
->ops
.mq
.allow_merge
)
79 return e
->type
->ops
.mq
.allow_merge(q
, rq
, bio
);
85 blk_mq_sched_completed_request(struct blk_mq_hw_ctx
*hctx
, struct request
*rq
)
87 struct elevator_queue
*e
= hctx
->queue
->elevator
;
89 if (e
&& e
->type
->ops
.mq
.completed_request
)
90 e
->type
->ops
.mq
.completed_request(hctx
, rq
);
92 BUG_ON(rq
->internal_tag
== -1);
94 blk_mq_put_tag(hctx
, hctx
->sched_tags
, rq
->mq_ctx
, rq
->internal_tag
);
97 static inline void blk_mq_sched_started_request(struct request
*rq
)
99 struct request_queue
*q
= rq
->q
;
100 struct elevator_queue
*e
= q
->elevator
;
102 if (e
&& e
->type
->ops
.mq
.started_request
)
103 e
->type
->ops
.mq
.started_request(rq
);
106 static inline void blk_mq_sched_requeue_request(struct request
*rq
)
108 struct request_queue
*q
= rq
->q
;
109 struct elevator_queue
*e
= q
->elevator
;
111 if (e
&& e
->type
->ops
.mq
.requeue_request
)
112 e
->type
->ops
.mq
.requeue_request(rq
);
115 static inline bool blk_mq_sched_has_work(struct blk_mq_hw_ctx
*hctx
)
117 struct elevator_queue
*e
= hctx
->queue
->elevator
;
119 if (e
&& e
->type
->ops
.mq
.has_work
)
120 return e
->type
->ops
.mq
.has_work(hctx
);
126 * Mark a hardware queue as needing a restart.
128 static inline void blk_mq_sched_mark_restart_hctx(struct blk_mq_hw_ctx
*hctx
)
130 if (!test_bit(BLK_MQ_S_SCHED_RESTART
, &hctx
->state
))
131 set_bit(BLK_MQ_S_SCHED_RESTART
, &hctx
->state
);
135 * Mark a hardware queue and the request queue it belongs to as needing a
138 static inline void blk_mq_sched_mark_restart_queue(struct blk_mq_hw_ctx
*hctx
)
140 struct request_queue
*q
= hctx
->queue
;
142 if (!test_bit(BLK_MQ_S_SCHED_RESTART
, &hctx
->state
))
143 set_bit(BLK_MQ_S_SCHED_RESTART
, &hctx
->state
);
144 if (!test_bit(QUEUE_FLAG_RESTART
, &q
->queue_flags
))
145 set_bit(QUEUE_FLAG_RESTART
, &q
->queue_flags
);
148 static inline bool blk_mq_sched_needs_restart(struct blk_mq_hw_ctx
*hctx
)
150 return test_bit(BLK_MQ_S_SCHED_RESTART
, &hctx
->state
);