Linux 4.19.133
[linux/fpc-iii.git] / drivers / md / dm-cache-background-tracker.c
blob84814e819e4c35130ad6444feb88082c37c35768
1 /*
2 * Copyright (C) 2017 Red Hat. All rights reserved.
4 * This file is released under the GPL.
5 */
7 #include "dm-cache-background-tracker.h"
9 /*----------------------------------------------------------------*/
11 #define DM_MSG_PREFIX "dm-background-tracker"
13 struct bt_work {
14 struct list_head list;
15 struct rb_node node;
16 struct policy_work work;
19 struct background_tracker {
20 unsigned max_work;
21 atomic_t pending_promotes;
22 atomic_t pending_writebacks;
23 atomic_t pending_demotes;
25 struct list_head issued;
26 struct list_head queued;
27 struct rb_root pending;
29 struct kmem_cache *work_cache;
32 struct background_tracker *btracker_create(unsigned max_work)
34 struct background_tracker *b = kmalloc(sizeof(*b), GFP_KERNEL);
36 if (!b) {
37 DMERR("couldn't create background_tracker");
38 return NULL;
41 b->max_work = max_work;
42 atomic_set(&b->pending_promotes, 0);
43 atomic_set(&b->pending_writebacks, 0);
44 atomic_set(&b->pending_demotes, 0);
46 INIT_LIST_HEAD(&b->issued);
47 INIT_LIST_HEAD(&b->queued);
49 b->pending = RB_ROOT;
50 b->work_cache = KMEM_CACHE(bt_work, 0);
51 if (!b->work_cache) {
52 DMERR("couldn't create mempool for background work items");
53 kfree(b);
54 b = NULL;
57 return b;
59 EXPORT_SYMBOL_GPL(btracker_create);
61 void btracker_destroy(struct background_tracker *b)
63 kmem_cache_destroy(b->work_cache);
64 kfree(b);
66 EXPORT_SYMBOL_GPL(btracker_destroy);
68 static int cmp_oblock(dm_oblock_t lhs, dm_oblock_t rhs)
70 if (from_oblock(lhs) < from_oblock(rhs))
71 return -1;
73 if (from_oblock(rhs) < from_oblock(lhs))
74 return 1;
76 return 0;
79 static bool __insert_pending(struct background_tracker *b,
80 struct bt_work *nw)
82 int cmp;
83 struct bt_work *w;
84 struct rb_node **new = &b->pending.rb_node, *parent = NULL;
86 while (*new) {
87 w = container_of(*new, struct bt_work, node);
89 parent = *new;
90 cmp = cmp_oblock(w->work.oblock, nw->work.oblock);
91 if (cmp < 0)
92 new = &((*new)->rb_left);
94 else if (cmp > 0)
95 new = &((*new)->rb_right);
97 else
98 /* already present */
99 return false;
102 rb_link_node(&nw->node, parent, new);
103 rb_insert_color(&nw->node, &b->pending);
105 return true;
108 static struct bt_work *__find_pending(struct background_tracker *b,
109 dm_oblock_t oblock)
111 int cmp;
112 struct bt_work *w;
113 struct rb_node **new = &b->pending.rb_node;
115 while (*new) {
116 w = container_of(*new, struct bt_work, node);
118 cmp = cmp_oblock(w->work.oblock, oblock);
119 if (cmp < 0)
120 new = &((*new)->rb_left);
122 else if (cmp > 0)
123 new = &((*new)->rb_right);
125 else
126 break;
129 return *new ? w : NULL;
133 static void update_stats(struct background_tracker *b, struct policy_work *w, int delta)
135 switch (w->op) {
136 case POLICY_PROMOTE:
137 atomic_add(delta, &b->pending_promotes);
138 break;
140 case POLICY_DEMOTE:
141 atomic_add(delta, &b->pending_demotes);
142 break;
144 case POLICY_WRITEBACK:
145 atomic_add(delta, &b->pending_writebacks);
146 break;
150 unsigned btracker_nr_writebacks_queued(struct background_tracker *b)
152 return atomic_read(&b->pending_writebacks);
154 EXPORT_SYMBOL_GPL(btracker_nr_writebacks_queued);
156 unsigned btracker_nr_demotions_queued(struct background_tracker *b)
158 return atomic_read(&b->pending_demotes);
160 EXPORT_SYMBOL_GPL(btracker_nr_demotions_queued);
162 static bool max_work_reached(struct background_tracker *b)
164 return atomic_read(&b->pending_promotes) +
165 atomic_read(&b->pending_writebacks) +
166 atomic_read(&b->pending_demotes) >= b->max_work;
169 static struct bt_work *alloc_work(struct background_tracker *b)
171 if (max_work_reached(b))
172 return NULL;
174 return kmem_cache_alloc(b->work_cache, GFP_NOWAIT);
177 int btracker_queue(struct background_tracker *b,
178 struct policy_work *work,
179 struct policy_work **pwork)
181 struct bt_work *w;
183 if (pwork)
184 *pwork = NULL;
186 w = alloc_work(b);
187 if (!w)
188 return -ENOMEM;
190 memcpy(&w->work, work, sizeof(*work));
192 if (!__insert_pending(b, w)) {
194 * There was a race, we'll just ignore this second
195 * bit of work for the same oblock.
197 kmem_cache_free(b->work_cache, w);
198 return -EINVAL;
201 if (pwork) {
202 *pwork = &w->work;
203 list_add(&w->list, &b->issued);
204 } else
205 list_add(&w->list, &b->queued);
206 update_stats(b, &w->work, 1);
208 return 0;
210 EXPORT_SYMBOL_GPL(btracker_queue);
213 * Returns -ENODATA if there's no work.
215 int btracker_issue(struct background_tracker *b, struct policy_work **work)
217 struct bt_work *w;
219 if (list_empty(&b->queued))
220 return -ENODATA;
222 w = list_first_entry(&b->queued, struct bt_work, list);
223 list_move(&w->list, &b->issued);
224 *work = &w->work;
226 return 0;
228 EXPORT_SYMBOL_GPL(btracker_issue);
230 void btracker_complete(struct background_tracker *b,
231 struct policy_work *op)
233 struct bt_work *w = container_of(op, struct bt_work, work);
235 update_stats(b, &w->work, -1);
236 rb_erase(&w->node, &b->pending);
237 list_del(&w->list);
238 kmem_cache_free(b->work_cache, w);
240 EXPORT_SYMBOL_GPL(btracker_complete);
242 bool btracker_promotion_already_present(struct background_tracker *b,
243 dm_oblock_t oblock)
245 return __find_pending(b, oblock) != NULL;
247 EXPORT_SYMBOL_GPL(btracker_promotion_already_present);
249 /*----------------------------------------------------------------*/