5 #include "blk-mq-tag.h"
7 void blk_mq_sched_free_hctx_data(struct request_queue
*q
,
8 void (*exit
)(struct blk_mq_hw_ctx
*));
10 struct request
*blk_mq_sched_get_request(struct request_queue
*q
, struct bio
*bio
, unsigned int op
, struct blk_mq_alloc_data
*data
);
11 void blk_mq_sched_put_request(struct request
*rq
);
13 void blk_mq_sched_request_inserted(struct request
*rq
);
14 bool blk_mq_sched_try_merge(struct request_queue
*q
, struct bio
*bio
,
15 struct request
**merged_request
);
16 bool __blk_mq_sched_bio_merge(struct request_queue
*q
, struct bio
*bio
);
17 bool blk_mq_sched_try_insert_merge(struct request_queue
*q
, struct request
*rq
);
18 void blk_mq_sched_restart(struct blk_mq_hw_ctx
*hctx
);
20 void blk_mq_sched_insert_request(struct request
*rq
, bool at_head
,
21 bool run_queue
, bool async
, bool can_block
);
22 void blk_mq_sched_insert_requests(struct request_queue
*q
,
23 struct blk_mq_ctx
*ctx
,
24 struct list_head
*list
, bool run_queue_async
);
26 void blk_mq_sched_dispatch_requests(struct blk_mq_hw_ctx
*hctx
);
28 int blk_mq_init_sched(struct request_queue
*q
, struct elevator_type
*e
);
29 void blk_mq_exit_sched(struct request_queue
*q
, struct elevator_queue
*e
);
31 int blk_mq_sched_init_hctx(struct request_queue
*q
, struct blk_mq_hw_ctx
*hctx
,
32 unsigned int hctx_idx
);
33 void blk_mq_sched_exit_hctx(struct request_queue
*q
, struct blk_mq_hw_ctx
*hctx
,
34 unsigned int hctx_idx
);
36 int blk_mq_sched_init(struct request_queue
*q
);
39 blk_mq_sched_bio_merge(struct request_queue
*q
, struct bio
*bio
)
41 struct elevator_queue
*e
= q
->elevator
;
43 if (!e
|| blk_queue_nomerges(q
) || !bio_mergeable(bio
))
46 return __blk_mq_sched_bio_merge(q
, bio
);
49 static inline int blk_mq_sched_get_rq_priv(struct request_queue
*q
,
53 struct elevator_queue
*e
= q
->elevator
;
55 if (e
&& e
->type
->ops
.mq
.get_rq_priv
)
56 return e
->type
->ops
.mq
.get_rq_priv(q
, rq
, bio
);
61 static inline void blk_mq_sched_put_rq_priv(struct request_queue
*q
,
64 struct elevator_queue
*e
= q
->elevator
;
66 if (e
&& e
->type
->ops
.mq
.put_rq_priv
)
67 e
->type
->ops
.mq
.put_rq_priv(q
, rq
);
71 blk_mq_sched_allow_merge(struct request_queue
*q
, struct request
*rq
,
74 struct elevator_queue
*e
= q
->elevator
;
76 if (e
&& e
->type
->ops
.mq
.allow_merge
)
77 return e
->type
->ops
.mq
.allow_merge(q
, rq
, bio
);
82 static inline void blk_mq_sched_completed_request(struct request
*rq
)
84 struct elevator_queue
*e
= rq
->q
->elevator
;
86 if (e
&& e
->type
->ops
.mq
.completed_request
)
87 e
->type
->ops
.mq
.completed_request(rq
);
90 static inline void blk_mq_sched_started_request(struct request
*rq
)
92 struct request_queue
*q
= rq
->q
;
93 struct elevator_queue
*e
= q
->elevator
;
95 if (e
&& e
->type
->ops
.mq
.started_request
)
96 e
->type
->ops
.mq
.started_request(rq
);
99 static inline void blk_mq_sched_requeue_request(struct request
*rq
)
101 struct request_queue
*q
= rq
->q
;
102 struct elevator_queue
*e
= q
->elevator
;
104 if (e
&& e
->type
->ops
.mq
.requeue_request
)
105 e
->type
->ops
.mq
.requeue_request(rq
);
108 static inline bool blk_mq_sched_has_work(struct blk_mq_hw_ctx
*hctx
)
110 struct elevator_queue
*e
= hctx
->queue
->elevator
;
112 if (e
&& e
->type
->ops
.mq
.has_work
)
113 return e
->type
->ops
.mq
.has_work(hctx
);
119 * Mark a hardware queue as needing a restart.
121 static inline void blk_mq_sched_mark_restart_hctx(struct blk_mq_hw_ctx
*hctx
)
123 if (!test_bit(BLK_MQ_S_SCHED_RESTART
, &hctx
->state
))
124 set_bit(BLK_MQ_S_SCHED_RESTART
, &hctx
->state
);
127 static inline bool blk_mq_sched_needs_restart(struct blk_mq_hw_ctx
*hctx
)
129 return test_bit(BLK_MQ_S_SCHED_RESTART
, &hctx
->state
);