2 * Copyright (c) 2013 Red Hat, Inc. and Parallels Inc. All rights reserved.
3 * Authors: David Chinner and Glauber Costa
5 * Generic LRU infrastructure
7 #include <linux/kernel.h>
8 #include <linux/module.h>
10 #include <linux/list_lru.h>
11 #include <linux/slab.h>
12 #include <linux/mutex.h>
13 #include <linux/memcontrol.h>
15 #ifdef CONFIG_MEMCG_KMEM
16 static LIST_HEAD(list_lrus
);
17 static DEFINE_MUTEX(list_lrus_mutex
);
19 static void list_lru_register(struct list_lru
*lru
)
21 mutex_lock(&list_lrus_mutex
);
22 list_add(&lru
->list
, &list_lrus
);
23 mutex_unlock(&list_lrus_mutex
);
26 static void list_lru_unregister(struct list_lru
*lru
)
28 mutex_lock(&list_lrus_mutex
);
30 mutex_unlock(&list_lrus_mutex
);
33 static void list_lru_register(struct list_lru
*lru
)
37 static void list_lru_unregister(struct list_lru
*lru
)
40 #endif /* CONFIG_MEMCG_KMEM */
42 #ifdef CONFIG_MEMCG_KMEM
43 static inline bool list_lru_memcg_aware(struct list_lru
*lru
)
45 return !!lru
->node
[0].memcg_lrus
;
48 static inline struct list_lru_one
*
49 list_lru_from_memcg_idx(struct list_lru_node
*nlru
, int idx
)
52 * The lock protects the array of per cgroup lists from relocation
53 * (see memcg_update_list_lru_node).
55 lockdep_assert_held(&nlru
->lock
);
56 if (nlru
->memcg_lrus
&& idx
>= 0)
57 return nlru
->memcg_lrus
->lru
[idx
];
62 static inline struct list_lru_one
*
63 list_lru_from_kmem(struct list_lru_node
*nlru
, void *ptr
)
65 struct mem_cgroup
*memcg
;
67 if (!nlru
->memcg_lrus
)
70 memcg
= mem_cgroup_from_kmem(ptr
);
74 return list_lru_from_memcg_idx(nlru
, memcg_cache_id(memcg
));
77 static inline bool list_lru_memcg_aware(struct list_lru
*lru
)
82 static inline struct list_lru_one
*
83 list_lru_from_memcg_idx(struct list_lru_node
*nlru
, int idx
)
88 static inline struct list_lru_one
*
89 list_lru_from_kmem(struct list_lru_node
*nlru
, void *ptr
)
93 #endif /* CONFIG_MEMCG_KMEM */
95 bool list_lru_add(struct list_lru
*lru
, struct list_head
*item
)
97 int nid
= page_to_nid(virt_to_page(item
));
98 struct list_lru_node
*nlru
= &lru
->node
[nid
];
99 struct list_lru_one
*l
;
101 spin_lock(&nlru
->lock
);
102 if (list_empty(item
)) {
103 l
= list_lru_from_kmem(nlru
, item
);
104 list_add_tail(item
, &l
->list
);
106 spin_unlock(&nlru
->lock
);
109 spin_unlock(&nlru
->lock
);
112 EXPORT_SYMBOL_GPL(list_lru_add
);
114 bool list_lru_del(struct list_lru
*lru
, struct list_head
*item
)
116 int nid
= page_to_nid(virt_to_page(item
));
117 struct list_lru_node
*nlru
= &lru
->node
[nid
];
118 struct list_lru_one
*l
;
120 spin_lock(&nlru
->lock
);
121 if (!list_empty(item
)) {
122 l
= list_lru_from_kmem(nlru
, item
);
125 spin_unlock(&nlru
->lock
);
128 spin_unlock(&nlru
->lock
);
131 EXPORT_SYMBOL_GPL(list_lru_del
);
133 void list_lru_isolate(struct list_lru_one
*list
, struct list_head
*item
)
138 EXPORT_SYMBOL_GPL(list_lru_isolate
);
140 void list_lru_isolate_move(struct list_lru_one
*list
, struct list_head
*item
,
141 struct list_head
*head
)
143 list_move(item
, head
);
146 EXPORT_SYMBOL_GPL(list_lru_isolate_move
);
148 static unsigned long __list_lru_count_one(struct list_lru
*lru
,
149 int nid
, int memcg_idx
)
151 struct list_lru_node
*nlru
= &lru
->node
[nid
];
152 struct list_lru_one
*l
;
155 spin_lock(&nlru
->lock
);
156 l
= list_lru_from_memcg_idx(nlru
, memcg_idx
);
158 spin_unlock(&nlru
->lock
);
163 unsigned long list_lru_count_one(struct list_lru
*lru
,
164 int nid
, struct mem_cgroup
*memcg
)
166 return __list_lru_count_one(lru
, nid
, memcg_cache_id(memcg
));
168 EXPORT_SYMBOL_GPL(list_lru_count_one
);
170 unsigned long list_lru_count_node(struct list_lru
*lru
, int nid
)
175 count
+= __list_lru_count_one(lru
, nid
, -1);
176 if (list_lru_memcg_aware(lru
)) {
177 for_each_memcg_cache_index(memcg_idx
)
178 count
+= __list_lru_count_one(lru
, nid
, memcg_idx
);
182 EXPORT_SYMBOL_GPL(list_lru_count_node
);
185 __list_lru_walk_one(struct list_lru
*lru
, int nid
, int memcg_idx
,
186 list_lru_walk_cb isolate
, void *cb_arg
,
187 unsigned long *nr_to_walk
)
190 struct list_lru_node
*nlru
= &lru
->node
[nid
];
191 struct list_lru_one
*l
;
192 struct list_head
*item
, *n
;
193 unsigned long isolated
= 0;
195 spin_lock(&nlru
->lock
);
196 l
= list_lru_from_memcg_idx(nlru
, memcg_idx
);
198 list_for_each_safe(item
, n
, &l
->list
) {
202 * decrement nr_to_walk first so that we don't livelock if we
203 * get stuck on large numbesr of LRU_RETRY items
209 ret
= isolate(item
, l
, &nlru
->lock
, cb_arg
);
211 case LRU_REMOVED_RETRY
:
212 assert_spin_locked(&nlru
->lock
);
216 * If the lru lock has been dropped, our list
217 * traversal is now invalid and so we have to
218 * restart from scratch.
220 if (ret
== LRU_REMOVED_RETRY
)
224 list_move_tail(item
, &l
->list
);
230 * The lru lock has been dropped, our list traversal is
231 * now invalid and so we have to restart from scratch.
233 assert_spin_locked(&nlru
->lock
);
240 spin_unlock(&nlru
->lock
);
245 list_lru_walk_one(struct list_lru
*lru
, int nid
, struct mem_cgroup
*memcg
,
246 list_lru_walk_cb isolate
, void *cb_arg
,
247 unsigned long *nr_to_walk
)
249 return __list_lru_walk_one(lru
, nid
, memcg_cache_id(memcg
),
250 isolate
, cb_arg
, nr_to_walk
);
252 EXPORT_SYMBOL_GPL(list_lru_walk_one
);
254 unsigned long list_lru_walk_node(struct list_lru
*lru
, int nid
,
255 list_lru_walk_cb isolate
, void *cb_arg
,
256 unsigned long *nr_to_walk
)
261 isolated
+= __list_lru_walk_one(lru
, nid
, -1, isolate
, cb_arg
,
263 if (*nr_to_walk
> 0 && list_lru_memcg_aware(lru
)) {
264 for_each_memcg_cache_index(memcg_idx
) {
265 isolated
+= __list_lru_walk_one(lru
, nid
, memcg_idx
,
266 isolate
, cb_arg
, nr_to_walk
);
267 if (*nr_to_walk
<= 0)
273 EXPORT_SYMBOL_GPL(list_lru_walk_node
);
275 static void init_one_lru(struct list_lru_one
*l
)
277 INIT_LIST_HEAD(&l
->list
);
281 #ifdef CONFIG_MEMCG_KMEM
282 static void __memcg_destroy_list_lru_node(struct list_lru_memcg
*memcg_lrus
,
287 for (i
= begin
; i
< end
; i
++)
288 kfree(memcg_lrus
->lru
[i
]);
291 static int __memcg_init_list_lru_node(struct list_lru_memcg
*memcg_lrus
,
296 for (i
= begin
; i
< end
; i
++) {
297 struct list_lru_one
*l
;
299 l
= kmalloc(sizeof(struct list_lru_one
), GFP_KERNEL
);
304 memcg_lrus
->lru
[i
] = l
;
308 __memcg_destroy_list_lru_node(memcg_lrus
, begin
, i
- 1);
312 static int memcg_init_list_lru_node(struct list_lru_node
*nlru
)
314 int size
= memcg_nr_cache_ids
;
316 nlru
->memcg_lrus
= kmalloc(size
* sizeof(void *), GFP_KERNEL
);
317 if (!nlru
->memcg_lrus
)
320 if (__memcg_init_list_lru_node(nlru
->memcg_lrus
, 0, size
)) {
321 kfree(nlru
->memcg_lrus
);
328 static void memcg_destroy_list_lru_node(struct list_lru_node
*nlru
)
330 __memcg_destroy_list_lru_node(nlru
->memcg_lrus
, 0, memcg_nr_cache_ids
);
331 kfree(nlru
->memcg_lrus
);
334 static int memcg_update_list_lru_node(struct list_lru_node
*nlru
,
335 int old_size
, int new_size
)
337 struct list_lru_memcg
*old
, *new;
339 BUG_ON(old_size
> new_size
);
341 old
= nlru
->memcg_lrus
;
342 new = kmalloc(new_size
* sizeof(void *), GFP_KERNEL
);
346 if (__memcg_init_list_lru_node(new, old_size
, new_size
)) {
351 memcpy(new, old
, old_size
* sizeof(void *));
354 * The lock guarantees that we won't race with a reader
355 * (see list_lru_from_memcg_idx).
357 * Since list_lru_{add,del} may be called under an IRQ-safe lock,
358 * we have to use IRQ-safe primitives here to avoid deadlock.
360 spin_lock_irq(&nlru
->lock
);
361 nlru
->memcg_lrus
= new;
362 spin_unlock_irq(&nlru
->lock
);
368 static void memcg_cancel_update_list_lru_node(struct list_lru_node
*nlru
,
369 int old_size
, int new_size
)
371 /* do not bother shrinking the array back to the old size, because we
372 * cannot handle allocation failures here */
373 __memcg_destroy_list_lru_node(nlru
->memcg_lrus
, old_size
, new_size
);
376 static int memcg_init_list_lru(struct list_lru
*lru
, bool memcg_aware
)
380 for (i
= 0; i
< nr_node_ids
; i
++) {
382 lru
->node
[i
].memcg_lrus
= NULL
;
383 else if (memcg_init_list_lru_node(&lru
->node
[i
]))
388 for (i
= i
- 1; i
>= 0; i
--)
389 memcg_destroy_list_lru_node(&lru
->node
[i
]);
393 static void memcg_destroy_list_lru(struct list_lru
*lru
)
397 if (!list_lru_memcg_aware(lru
))
400 for (i
= 0; i
< nr_node_ids
; i
++)
401 memcg_destroy_list_lru_node(&lru
->node
[i
]);
404 static int memcg_update_list_lru(struct list_lru
*lru
,
405 int old_size
, int new_size
)
409 if (!list_lru_memcg_aware(lru
))
412 for (i
= 0; i
< nr_node_ids
; i
++) {
413 if (memcg_update_list_lru_node(&lru
->node
[i
],
419 for (i
= i
- 1; i
>= 0; i
--)
420 memcg_cancel_update_list_lru_node(&lru
->node
[i
],
425 static void memcg_cancel_update_list_lru(struct list_lru
*lru
,
426 int old_size
, int new_size
)
430 if (!list_lru_memcg_aware(lru
))
433 for (i
= 0; i
< nr_node_ids
; i
++)
434 memcg_cancel_update_list_lru_node(&lru
->node
[i
],
438 int memcg_update_all_list_lrus(int new_size
)
441 struct list_lru
*lru
;
442 int old_size
= memcg_nr_cache_ids
;
444 mutex_lock(&list_lrus_mutex
);
445 list_for_each_entry(lru
, &list_lrus
, list
) {
446 ret
= memcg_update_list_lru(lru
, old_size
, new_size
);
451 mutex_unlock(&list_lrus_mutex
);
454 list_for_each_entry_continue_reverse(lru
, &list_lrus
, list
)
455 memcg_cancel_update_list_lru(lru
, old_size
, new_size
);
459 static void memcg_drain_list_lru_node(struct list_lru_node
*nlru
,
460 int src_idx
, int dst_idx
)
462 struct list_lru_one
*src
, *dst
;
465 * Since list_lru_{add,del} may be called under an IRQ-safe lock,
466 * we have to use IRQ-safe primitives here to avoid deadlock.
468 spin_lock_irq(&nlru
->lock
);
470 src
= list_lru_from_memcg_idx(nlru
, src_idx
);
471 dst
= list_lru_from_memcg_idx(nlru
, dst_idx
);
473 list_splice_init(&src
->list
, &dst
->list
);
474 dst
->nr_items
+= src
->nr_items
;
477 spin_unlock_irq(&nlru
->lock
);
480 static void memcg_drain_list_lru(struct list_lru
*lru
,
481 int src_idx
, int dst_idx
)
485 if (!list_lru_memcg_aware(lru
))
488 for (i
= 0; i
< nr_node_ids
; i
++)
489 memcg_drain_list_lru_node(&lru
->node
[i
], src_idx
, dst_idx
);
492 void memcg_drain_all_list_lrus(int src_idx
, int dst_idx
)
494 struct list_lru
*lru
;
496 mutex_lock(&list_lrus_mutex
);
497 list_for_each_entry(lru
, &list_lrus
, list
)
498 memcg_drain_list_lru(lru
, src_idx
, dst_idx
);
499 mutex_unlock(&list_lrus_mutex
);
502 static int memcg_init_list_lru(struct list_lru
*lru
, bool memcg_aware
)
507 static void memcg_destroy_list_lru(struct list_lru
*lru
)
510 #endif /* CONFIG_MEMCG_KMEM */
512 int __list_lru_init(struct list_lru
*lru
, bool memcg_aware
,
513 struct lock_class_key
*key
)
516 size_t size
= sizeof(*lru
->node
) * nr_node_ids
;
519 memcg_get_cache_ids();
521 lru
->node
= kzalloc(size
, GFP_KERNEL
);
525 for (i
= 0; i
< nr_node_ids
; i
++) {
526 spin_lock_init(&lru
->node
[i
].lock
);
528 lockdep_set_class(&lru
->node
[i
].lock
, key
);
529 init_one_lru(&lru
->node
[i
].lru
);
532 err
= memcg_init_list_lru(lru
, memcg_aware
);
538 list_lru_register(lru
);
540 memcg_put_cache_ids();
543 EXPORT_SYMBOL_GPL(__list_lru_init
);
545 void list_lru_destroy(struct list_lru
*lru
)
547 /* Already destroyed or not yet initialized? */
551 memcg_get_cache_ids();
553 list_lru_unregister(lru
);
555 memcg_destroy_list_lru(lru
);
559 memcg_put_cache_ids();
561 EXPORT_SYMBOL_GPL(list_lru_destroy
);