1 /* SPDX-License-Identifier: GPL-2.0 */
6 #include "blk-mq-tag.h"
8 void blk_mq_sched_free_hctx_data(struct request_queue
*q
,
9 void (*exit
)(struct blk_mq_hw_ctx
*));
11 void blk_mq_sched_assign_ioc(struct request
*rq
, struct bio
*bio
);
13 void blk_mq_sched_request_inserted(struct request
*rq
);
14 bool blk_mq_sched_try_merge(struct request_queue
*q
, struct bio
*bio
,
15 struct request
**merged_request
);
16 bool __blk_mq_sched_bio_merge(struct request_queue
*q
, struct bio
*bio
);
17 bool blk_mq_sched_try_insert_merge(struct request_queue
*q
, struct request
*rq
);
18 void blk_mq_sched_mark_restart_hctx(struct blk_mq_hw_ctx
*hctx
);
19 void blk_mq_sched_restart(struct blk_mq_hw_ctx
*hctx
);
21 void blk_mq_sched_insert_request(struct request
*rq
, bool at_head
,
22 bool run_queue
, bool async
);
23 void blk_mq_sched_insert_requests(struct request_queue
*q
,
24 struct blk_mq_ctx
*ctx
,
25 struct list_head
*list
, bool run_queue_async
);
27 void blk_mq_sched_dispatch_requests(struct blk_mq_hw_ctx
*hctx
);
29 int blk_mq_init_sched(struct request_queue
*q
, struct elevator_type
*e
);
30 void blk_mq_exit_sched(struct request_queue
*q
, struct elevator_queue
*e
);
33 blk_mq_sched_bio_merge(struct request_queue
*q
, struct bio
*bio
)
35 if (blk_queue_nomerges(q
) || !bio_mergeable(bio
))
38 return __blk_mq_sched_bio_merge(q
, bio
);
42 blk_mq_sched_allow_merge(struct request_queue
*q
, struct request
*rq
,
45 struct elevator_queue
*e
= q
->elevator
;
47 if (e
&& e
->type
->ops
.mq
.allow_merge
)
48 return e
->type
->ops
.mq
.allow_merge(q
, rq
, bio
);
53 static inline void blk_mq_sched_completed_request(struct request
*rq
)
55 struct elevator_queue
*e
= rq
->q
->elevator
;
57 if (e
&& e
->type
->ops
.mq
.completed_request
)
58 e
->type
->ops
.mq
.completed_request(rq
);
61 static inline void blk_mq_sched_started_request(struct request
*rq
)
63 struct request_queue
*q
= rq
->q
;
64 struct elevator_queue
*e
= q
->elevator
;
66 if (e
&& e
->type
->ops
.mq
.started_request
)
67 e
->type
->ops
.mq
.started_request(rq
);
70 static inline void blk_mq_sched_requeue_request(struct request
*rq
)
72 struct request_queue
*q
= rq
->q
;
73 struct elevator_queue
*e
= q
->elevator
;
75 if (e
&& e
->type
->ops
.mq
.requeue_request
)
76 e
->type
->ops
.mq
.requeue_request(rq
);
79 static inline bool blk_mq_sched_has_work(struct blk_mq_hw_ctx
*hctx
)
81 struct elevator_queue
*e
= hctx
->queue
->elevator
;
83 if (e
&& e
->type
->ops
.mq
.has_work
)
84 return e
->type
->ops
.mq
.has_work(hctx
);
89 static inline bool blk_mq_sched_needs_restart(struct blk_mq_hw_ctx
*hctx
)
91 return test_bit(BLK_MQ_S_SCHED_RESTART
, &hctx
->state
);