1 // SPDX-License-Identifier: GPL-2.0+
3 * test_xarray.c: Test the XArray API
4 * Copyright (c) 2017-2018 Microsoft Corporation
5 * Author: Matthew Wilcox <willy@infradead.org>
8 #include <linux/xarray.h>
9 #include <linux/module.h>
11 static unsigned int tests_run
;
12 static unsigned int tests_passed
;
16 void xa_dump(const struct xarray
*xa
) { }
19 #define XA_BUG_ON(xa, x) do { \
22 printk("BUG at %s:%d\n", __func__, __LINE__); \
31 static void *xa_mk_index(unsigned long index
)
33 return xa_mk_value(index
& LONG_MAX
);
36 static void *xa_store_index(struct xarray
*xa
, unsigned long index
, gfp_t gfp
)
38 return xa_store(xa
, index
, xa_mk_index(index
), gfp
);
41 static void xa_insert_index(struct xarray
*xa
, unsigned long index
)
43 XA_BUG_ON(xa
, xa_insert(xa
, index
, xa_mk_index(index
),
47 static void xa_alloc_index(struct xarray
*xa
, unsigned long index
, gfp_t gfp
)
51 XA_BUG_ON(xa
, xa_alloc(xa
, &id
, xa_mk_index(index
), xa_limit_32b
,
53 XA_BUG_ON(xa
, id
!= index
);
56 static void xa_erase_index(struct xarray
*xa
, unsigned long index
)
58 XA_BUG_ON(xa
, xa_erase(xa
, index
) != xa_mk_index(index
));
59 XA_BUG_ON(xa
, xa_load(xa
, index
) != NULL
);
63 * If anyone needs this, please move it to xarray.c. We have no current
64 * users outside the test suite because all current multislot users want
65 * to use the advanced API.
67 static void *xa_store_order(struct xarray
*xa
, unsigned long index
,
68 unsigned order
, void *entry
, gfp_t gfp
)
70 XA_STATE_ORDER(xas
, xa
, index
, order
);
75 curr
= xas_store(&xas
, entry
);
77 } while (xas_nomem(&xas
, gfp
));
82 static noinline
void check_xa_err(struct xarray
*xa
)
84 XA_BUG_ON(xa
, xa_err(xa_store_index(xa
, 0, GFP_NOWAIT
)) != 0);
85 XA_BUG_ON(xa
, xa_err(xa_erase(xa
, 0)) != 0);
87 /* The kernel does not fail GFP_NOWAIT allocations */
88 XA_BUG_ON(xa
, xa_err(xa_store_index(xa
, 1, GFP_NOWAIT
)) != -ENOMEM
);
89 XA_BUG_ON(xa
, xa_err(xa_store_index(xa
, 1, GFP_NOWAIT
)) != -ENOMEM
);
91 XA_BUG_ON(xa
, xa_err(xa_store_index(xa
, 1, GFP_KERNEL
)) != 0);
92 XA_BUG_ON(xa
, xa_err(xa_store(xa
, 1, xa_mk_value(0), GFP_KERNEL
)) != 0);
93 XA_BUG_ON(xa
, xa_err(xa_erase(xa
, 1)) != 0);
94 // kills the test-suite :-(
95 // XA_BUG_ON(xa, xa_err(xa_store(xa, 0, xa_mk_internal(0), 0)) != -EINVAL);
98 static noinline
void check_xas_retry(struct xarray
*xa
)
100 XA_STATE(xas
, xa
, 0);
103 xa_store_index(xa
, 0, GFP_KERNEL
);
104 xa_store_index(xa
, 1, GFP_KERNEL
);
107 XA_BUG_ON(xa
, xas_find(&xas
, ULONG_MAX
) != xa_mk_value(0));
108 xa_erase_index(xa
, 1);
109 XA_BUG_ON(xa
, !xa_is_retry(xas_reload(&xas
)));
110 XA_BUG_ON(xa
, xas_retry(&xas
, NULL
));
111 XA_BUG_ON(xa
, xas_retry(&xas
, xa_mk_value(0)));
113 XA_BUG_ON(xa
, xas
.xa_node
!= XAS_RESTART
);
114 XA_BUG_ON(xa
, xas_next_entry(&xas
, ULONG_MAX
) != xa_mk_value(0));
115 XA_BUG_ON(xa
, xas
.xa_node
!= NULL
);
118 XA_BUG_ON(xa
, xa_store_index(xa
, 1, GFP_KERNEL
) != NULL
);
121 XA_BUG_ON(xa
, !xa_is_internal(xas_reload(&xas
)));
122 xas
.xa_node
= XAS_RESTART
;
123 XA_BUG_ON(xa
, xas_next_entry(&xas
, ULONG_MAX
) != xa_mk_value(0));
126 /* Make sure we can iterate through retry entries */
129 xas_store(&xas
, XA_RETRY_ENTRY
);
131 xas_store(&xas
, XA_RETRY_ENTRY
);
134 xas_for_each(&xas
, entry
, ULONG_MAX
) {
135 xas_store(&xas
, xa_mk_index(xas
.xa_index
));
139 xa_erase_index(xa
, 0);
140 xa_erase_index(xa
, 1);
143 static noinline
void check_xa_load(struct xarray
*xa
)
147 for (i
= 0; i
< 1024; i
++) {
148 for (j
= 0; j
< 1024; j
++) {
149 void *entry
= xa_load(xa
, j
);
151 XA_BUG_ON(xa
, xa_to_value(entry
) != j
);
153 XA_BUG_ON(xa
, entry
);
155 XA_BUG_ON(xa
, xa_store_index(xa
, i
, GFP_KERNEL
) != NULL
);
158 for (i
= 0; i
< 1024; i
++) {
159 for (j
= 0; j
< 1024; j
++) {
160 void *entry
= xa_load(xa
, j
);
162 XA_BUG_ON(xa
, xa_to_value(entry
) != j
);
164 XA_BUG_ON(xa
, entry
);
166 xa_erase_index(xa
, i
);
168 XA_BUG_ON(xa
, !xa_empty(xa
));
171 static noinline
void check_xa_mark_1(struct xarray
*xa
, unsigned long index
)
174 unsigned int max_order
= IS_ENABLED(CONFIG_XARRAY_MULTI
) ? 8 : 1;
176 /* NULL elements have no marks set */
177 XA_BUG_ON(xa
, xa_get_mark(xa
, index
, XA_MARK_0
));
178 xa_set_mark(xa
, index
, XA_MARK_0
);
179 XA_BUG_ON(xa
, xa_get_mark(xa
, index
, XA_MARK_0
));
181 /* Storing a pointer will not make a mark appear */
182 XA_BUG_ON(xa
, xa_store_index(xa
, index
, GFP_KERNEL
) != NULL
);
183 XA_BUG_ON(xa
, xa_get_mark(xa
, index
, XA_MARK_0
));
184 xa_set_mark(xa
, index
, XA_MARK_0
);
185 XA_BUG_ON(xa
, !xa_get_mark(xa
, index
, XA_MARK_0
));
187 /* Setting one mark will not set another mark */
188 XA_BUG_ON(xa
, xa_get_mark(xa
, index
+ 1, XA_MARK_0
));
189 XA_BUG_ON(xa
, xa_get_mark(xa
, index
, XA_MARK_1
));
191 /* Storing NULL clears marks, and they can't be set again */
192 xa_erase_index(xa
, index
);
193 XA_BUG_ON(xa
, !xa_empty(xa
));
194 XA_BUG_ON(xa
, xa_get_mark(xa
, index
, XA_MARK_0
));
195 xa_set_mark(xa
, index
, XA_MARK_0
);
196 XA_BUG_ON(xa
, xa_get_mark(xa
, index
, XA_MARK_0
));
199 * Storing a multi-index entry over entries with marks gives the
200 * entire entry the union of the marks
202 BUG_ON((index
% 4) != 0);
203 for (order
= 2; order
< max_order
; order
++) {
204 unsigned long base
= round_down(index
, 1UL << order
);
205 unsigned long next
= base
+ (1UL << order
);
208 XA_BUG_ON(xa
, xa_store_index(xa
, index
+ 1, GFP_KERNEL
));
209 xa_set_mark(xa
, index
+ 1, XA_MARK_0
);
210 XA_BUG_ON(xa
, xa_store_index(xa
, index
+ 2, GFP_KERNEL
));
211 xa_set_mark(xa
, index
+ 2, XA_MARK_2
);
212 XA_BUG_ON(xa
, xa_store_index(xa
, next
, GFP_KERNEL
));
213 xa_store_order(xa
, index
, order
, xa_mk_index(index
),
215 for (i
= base
; i
< next
; i
++) {
216 XA_STATE(xas
, xa
, i
);
217 unsigned int seen
= 0;
220 XA_BUG_ON(xa
, !xa_get_mark(xa
, i
, XA_MARK_0
));
221 XA_BUG_ON(xa
, xa_get_mark(xa
, i
, XA_MARK_1
));
222 XA_BUG_ON(xa
, !xa_get_mark(xa
, i
, XA_MARK_2
));
224 /* We should see two elements in the array */
226 xas_for_each(&xas
, entry
, ULONG_MAX
)
229 XA_BUG_ON(xa
, seen
!= 2);
231 /* One of which is marked */
235 xas_for_each_marked(&xas
, entry
, ULONG_MAX
, XA_MARK_0
)
238 XA_BUG_ON(xa
, seen
!= 1);
240 XA_BUG_ON(xa
, xa_get_mark(xa
, next
, XA_MARK_0
));
241 XA_BUG_ON(xa
, xa_get_mark(xa
, next
, XA_MARK_1
));
242 XA_BUG_ON(xa
, xa_get_mark(xa
, next
, XA_MARK_2
));
243 xa_erase_index(xa
, index
);
244 xa_erase_index(xa
, next
);
245 XA_BUG_ON(xa
, !xa_empty(xa
));
247 XA_BUG_ON(xa
, !xa_empty(xa
));
250 static noinline
void check_xa_mark_2(struct xarray
*xa
)
252 XA_STATE(xas
, xa
, 0);
254 unsigned int count
= 0;
257 xa_store_index(xa
, 0, GFP_KERNEL
);
258 xa_set_mark(xa
, 0, XA_MARK_0
);
261 xas_init_marks(&xas
);
263 XA_BUG_ON(xa
, !xa_get_mark(xa
, 0, XA_MARK_0
) == 0);
265 for (index
= 3500; index
< 4500; index
++) {
266 xa_store_index(xa
, index
, GFP_KERNEL
);
267 xa_set_mark(xa
, index
, XA_MARK_0
);
272 xas_for_each_marked(&xas
, entry
, ULONG_MAX
, XA_MARK_0
)
275 XA_BUG_ON(xa
, count
!= 1000);
278 xas_for_each(&xas
, entry
, ULONG_MAX
) {
279 xas_init_marks(&xas
);
280 XA_BUG_ON(xa
, !xa_get_mark(xa
, xas
.xa_index
, XA_MARK_0
));
281 XA_BUG_ON(xa
, !xas_get_mark(&xas
, XA_MARK_0
));
288 static noinline
void check_xa_mark(struct xarray
*xa
)
292 for (index
= 0; index
< 16384; index
+= 4)
293 check_xa_mark_1(xa
, index
);
298 static noinline
void check_xa_shrink(struct xarray
*xa
)
300 XA_STATE(xas
, xa
, 1);
301 struct xa_node
*node
;
303 unsigned int max_order
= IS_ENABLED(CONFIG_XARRAY_MULTI
) ? 15 : 1;
305 XA_BUG_ON(xa
, !xa_empty(xa
));
306 XA_BUG_ON(xa
, xa_store_index(xa
, 0, GFP_KERNEL
) != NULL
);
307 XA_BUG_ON(xa
, xa_store_index(xa
, 1, GFP_KERNEL
) != NULL
);
310 * Check that erasing the entry at 1 shrinks the tree and properly
311 * marks the node as being deleted.
314 XA_BUG_ON(xa
, xas_load(&xas
) != xa_mk_value(1));
316 XA_BUG_ON(xa
, xa_entry_locked(xa
, node
, 0) != xa_mk_value(0));
317 XA_BUG_ON(xa
, xas_store(&xas
, NULL
) != xa_mk_value(1));
318 XA_BUG_ON(xa
, xa_load(xa
, 1) != NULL
);
319 XA_BUG_ON(xa
, xas
.xa_node
!= XAS_BOUNDS
);
320 XA_BUG_ON(xa
, xa_entry_locked(xa
, node
, 0) != XA_RETRY_ENTRY
);
321 XA_BUG_ON(xa
, xas_load(&xas
) != NULL
);
323 XA_BUG_ON(xa
, xa_load(xa
, 0) != xa_mk_value(0));
324 xa_erase_index(xa
, 0);
325 XA_BUG_ON(xa
, !xa_empty(xa
));
327 for (order
= 0; order
< max_order
; order
++) {
328 unsigned long max
= (1UL << order
) - 1;
329 xa_store_order(xa
, 0, order
, xa_mk_value(0), GFP_KERNEL
);
330 XA_BUG_ON(xa
, xa_load(xa
, max
) != xa_mk_value(0));
331 XA_BUG_ON(xa
, xa_load(xa
, max
+ 1) != NULL
);
335 XA_BUG_ON(xa
, xa_store_index(xa
, ULONG_MAX
, GFP_KERNEL
) !=
338 XA_BUG_ON(xa
, xa_head(xa
) == node
);
340 XA_BUG_ON(xa
, xa_load(xa
, max
+ 1) != NULL
);
341 xa_erase_index(xa
, ULONG_MAX
);
342 XA_BUG_ON(xa
, xa
->xa_head
!= node
);
343 xa_erase_index(xa
, 0);
347 static noinline
void check_insert(struct xarray
*xa
)
351 for (i
= 0; i
< 1024; i
++) {
352 xa_insert_index(xa
, i
);
353 XA_BUG_ON(xa
, xa_load(xa
, i
- 1) != NULL
);
354 XA_BUG_ON(xa
, xa_load(xa
, i
+ 1) != NULL
);
355 xa_erase_index(xa
, i
);
358 for (i
= 10; i
< BITS_PER_LONG
; i
++) {
359 xa_insert_index(xa
, 1UL << i
);
360 XA_BUG_ON(xa
, xa_load(xa
, (1UL << i
) - 1) != NULL
);
361 XA_BUG_ON(xa
, xa_load(xa
, (1UL << i
) + 1) != NULL
);
362 xa_erase_index(xa
, 1UL << i
);
364 xa_insert_index(xa
, (1UL << i
) - 1);
365 XA_BUG_ON(xa
, xa_load(xa
, (1UL << i
) - 2) != NULL
);
366 XA_BUG_ON(xa
, xa_load(xa
, 1UL << i
) != NULL
);
367 xa_erase_index(xa
, (1UL << i
) - 1);
370 xa_insert_index(xa
, ~0UL);
371 XA_BUG_ON(xa
, xa_load(xa
, 0UL) != NULL
);
372 XA_BUG_ON(xa
, xa_load(xa
, ~1UL) != NULL
);
373 xa_erase_index(xa
, ~0UL);
375 XA_BUG_ON(xa
, !xa_empty(xa
));
378 static noinline
void check_cmpxchg(struct xarray
*xa
)
380 void *FIVE
= xa_mk_value(5);
381 void *SIX
= xa_mk_value(6);
382 void *LOTS
= xa_mk_value(12345678);
384 XA_BUG_ON(xa
, !xa_empty(xa
));
385 XA_BUG_ON(xa
, xa_store_index(xa
, 12345678, GFP_KERNEL
) != NULL
);
386 XA_BUG_ON(xa
, xa_insert(xa
, 12345678, xa
, GFP_KERNEL
) != -EBUSY
);
387 XA_BUG_ON(xa
, xa_cmpxchg(xa
, 12345678, SIX
, FIVE
, GFP_KERNEL
) != LOTS
);
388 XA_BUG_ON(xa
, xa_cmpxchg(xa
, 12345678, LOTS
, FIVE
, GFP_KERNEL
) != LOTS
);
389 XA_BUG_ON(xa
, xa_cmpxchg(xa
, 12345678, FIVE
, LOTS
, GFP_KERNEL
) != FIVE
);
390 XA_BUG_ON(xa
, xa_cmpxchg(xa
, 5, FIVE
, NULL
, GFP_KERNEL
) != NULL
);
391 XA_BUG_ON(xa
, xa_cmpxchg(xa
, 5, NULL
, FIVE
, GFP_KERNEL
) != NULL
);
392 xa_erase_index(xa
, 12345678);
393 xa_erase_index(xa
, 5);
394 XA_BUG_ON(xa
, !xa_empty(xa
));
397 static noinline
void check_reserve(struct xarray
*xa
)
403 /* An array with a reserved entry is not empty */
404 XA_BUG_ON(xa
, !xa_empty(xa
));
405 XA_BUG_ON(xa
, xa_reserve(xa
, 12345678, GFP_KERNEL
) != 0);
406 XA_BUG_ON(xa
, xa_empty(xa
));
407 XA_BUG_ON(xa
, xa_load(xa
, 12345678));
408 xa_release(xa
, 12345678);
409 XA_BUG_ON(xa
, !xa_empty(xa
));
411 /* Releasing a used entry does nothing */
412 XA_BUG_ON(xa
, xa_reserve(xa
, 12345678, GFP_KERNEL
) != 0);
413 XA_BUG_ON(xa
, xa_store_index(xa
, 12345678, GFP_NOWAIT
) != NULL
);
414 xa_release(xa
, 12345678);
415 xa_erase_index(xa
, 12345678);
416 XA_BUG_ON(xa
, !xa_empty(xa
));
418 /* cmpxchg sees a reserved entry as ZERO */
419 XA_BUG_ON(xa
, xa_reserve(xa
, 12345678, GFP_KERNEL
) != 0);
420 XA_BUG_ON(xa
, xa_cmpxchg(xa
, 12345678, XA_ZERO_ENTRY
,
421 xa_mk_value(12345678), GFP_NOWAIT
) != NULL
);
422 xa_release(xa
, 12345678);
423 xa_erase_index(xa
, 12345678);
424 XA_BUG_ON(xa
, !xa_empty(xa
));
426 /* xa_insert treats it as busy */
427 XA_BUG_ON(xa
, xa_reserve(xa
, 12345678, GFP_KERNEL
) != 0);
428 XA_BUG_ON(xa
, xa_insert(xa
, 12345678, xa_mk_value(12345678), 0) !=
430 XA_BUG_ON(xa
, xa_empty(xa
));
431 XA_BUG_ON(xa
, xa_erase(xa
, 12345678) != NULL
);
432 XA_BUG_ON(xa
, !xa_empty(xa
));
434 /* Can iterate through a reserved entry */
435 xa_store_index(xa
, 5, GFP_KERNEL
);
436 XA_BUG_ON(xa
, xa_reserve(xa
, 6, GFP_KERNEL
) != 0);
437 xa_store_index(xa
, 7, GFP_KERNEL
);
440 xa_for_each(xa
, index
, entry
) {
441 XA_BUG_ON(xa
, index
!= 5 && index
!= 7);
444 XA_BUG_ON(xa
, count
!= 2);
446 /* If we free a reserved entry, we should be able to allocate it */
447 if (xa
->xa_flags
& XA_FLAGS_ALLOC
) {
450 XA_BUG_ON(xa
, xa_alloc(xa
, &id
, xa_mk_value(8),
451 XA_LIMIT(5, 10), GFP_KERNEL
) != 0);
452 XA_BUG_ON(xa
, id
!= 8);
455 XA_BUG_ON(xa
, xa_alloc(xa
, &id
, xa_mk_value(6),
456 XA_LIMIT(5, 10), GFP_KERNEL
) != 0);
457 XA_BUG_ON(xa
, id
!= 6);
463 static noinline
void check_xas_erase(struct xarray
*xa
)
465 XA_STATE(xas
, xa
, 0);
469 for (i
= 0; i
< 200; i
++) {
470 for (j
= i
; j
< 2 * i
+ 17; j
++) {
474 xas_store(&xas
, xa_mk_index(j
));
476 } while (xas_nomem(&xas
, GFP_KERNEL
));
479 xas_set(&xas
, ULONG_MAX
);
482 xas_store(&xas
, xa_mk_value(0));
484 } while (xas_nomem(&xas
, GFP_KERNEL
));
487 xas_store(&xas
, NULL
);
491 xas_for_each(&xas
, entry
, ULONG_MAX
) {
492 XA_BUG_ON(xa
, entry
!= xa_mk_index(j
));
493 xas_store(&xas
, NULL
);
497 XA_BUG_ON(xa
, !xa_empty(xa
));
501 #ifdef CONFIG_XARRAY_MULTI
502 static noinline
void check_multi_store_1(struct xarray
*xa
, unsigned long index
,
505 XA_STATE(xas
, xa
, index
);
506 unsigned long min
= index
& ~((1UL << order
) - 1);
507 unsigned long max
= min
+ (1UL << order
);
509 xa_store_order(xa
, index
, order
, xa_mk_index(index
), GFP_KERNEL
);
510 XA_BUG_ON(xa
, xa_load(xa
, min
) != xa_mk_index(index
));
511 XA_BUG_ON(xa
, xa_load(xa
, max
- 1) != xa_mk_index(index
));
512 XA_BUG_ON(xa
, xa_load(xa
, max
) != NULL
);
513 XA_BUG_ON(xa
, xa_load(xa
, min
- 1) != NULL
);
516 XA_BUG_ON(xa
, xas_store(&xas
, xa_mk_index(min
)) != xa_mk_index(index
));
518 XA_BUG_ON(xa
, xa_load(xa
, min
) != xa_mk_index(min
));
519 XA_BUG_ON(xa
, xa_load(xa
, max
- 1) != xa_mk_index(min
));
520 XA_BUG_ON(xa
, xa_load(xa
, max
) != NULL
);
521 XA_BUG_ON(xa
, xa_load(xa
, min
- 1) != NULL
);
523 xa_erase_index(xa
, min
);
524 XA_BUG_ON(xa
, !xa_empty(xa
));
527 static noinline
void check_multi_store_2(struct xarray
*xa
, unsigned long index
,
530 XA_STATE(xas
, xa
, index
);
531 xa_store_order(xa
, index
, order
, xa_mk_value(0), GFP_KERNEL
);
534 XA_BUG_ON(xa
, xas_store(&xas
, xa_mk_value(1)) != xa_mk_value(0));
535 XA_BUG_ON(xa
, xas
.xa_index
!= index
);
536 XA_BUG_ON(xa
, xas_store(&xas
, NULL
) != xa_mk_value(1));
538 XA_BUG_ON(xa
, !xa_empty(xa
));
541 static noinline
void check_multi_store_3(struct xarray
*xa
, unsigned long index
,
544 XA_STATE(xas
, xa
, 0);
548 xa_store_order(xa
, index
, order
, xa_mk_index(index
), GFP_KERNEL
);
551 xas_for_each(&xas
, entry
, ULONG_MAX
) {
552 XA_BUG_ON(xa
, entry
!= xa_mk_index(index
));
555 XA_BUG_ON(xa
, n
!= 1);
556 xas_set(&xas
, index
+ 1);
557 xas_for_each(&xas
, entry
, ULONG_MAX
) {
558 XA_BUG_ON(xa
, entry
!= xa_mk_index(index
));
561 XA_BUG_ON(xa
, n
!= 2);
568 static noinline
void check_multi_store(struct xarray
*xa
)
570 #ifdef CONFIG_XARRAY_MULTI
571 unsigned long i
, j
, k
;
572 unsigned int max_order
= (sizeof(long) == 4) ? 30 : 60;
574 /* Loading from any position returns the same value */
575 xa_store_order(xa
, 0, 1, xa_mk_value(0), GFP_KERNEL
);
576 XA_BUG_ON(xa
, xa_load(xa
, 0) != xa_mk_value(0));
577 XA_BUG_ON(xa
, xa_load(xa
, 1) != xa_mk_value(0));
578 XA_BUG_ON(xa
, xa_load(xa
, 2) != NULL
);
580 XA_BUG_ON(xa
, xa_to_node(xa_head(xa
))->count
!= 2);
581 XA_BUG_ON(xa
, xa_to_node(xa_head(xa
))->nr_values
!= 2);
584 /* Storing adjacent to the value does not alter the value */
585 xa_store(xa
, 3, xa
, GFP_KERNEL
);
586 XA_BUG_ON(xa
, xa_load(xa
, 0) != xa_mk_value(0));
587 XA_BUG_ON(xa
, xa_load(xa
, 1) != xa_mk_value(0));
588 XA_BUG_ON(xa
, xa_load(xa
, 2) != NULL
);
590 XA_BUG_ON(xa
, xa_to_node(xa_head(xa
))->count
!= 3);
591 XA_BUG_ON(xa
, xa_to_node(xa_head(xa
))->nr_values
!= 2);
594 /* Overwriting multiple indexes works */
595 xa_store_order(xa
, 0, 2, xa_mk_value(1), GFP_KERNEL
);
596 XA_BUG_ON(xa
, xa_load(xa
, 0) != xa_mk_value(1));
597 XA_BUG_ON(xa
, xa_load(xa
, 1) != xa_mk_value(1));
598 XA_BUG_ON(xa
, xa_load(xa
, 2) != xa_mk_value(1));
599 XA_BUG_ON(xa
, xa_load(xa
, 3) != xa_mk_value(1));
600 XA_BUG_ON(xa
, xa_load(xa
, 4) != NULL
);
602 XA_BUG_ON(xa
, xa_to_node(xa_head(xa
))->count
!= 4);
603 XA_BUG_ON(xa
, xa_to_node(xa_head(xa
))->nr_values
!= 4);
606 /* We can erase multiple values with a single store */
607 xa_store_order(xa
, 0, BITS_PER_LONG
- 1, NULL
, GFP_KERNEL
);
608 XA_BUG_ON(xa
, !xa_empty(xa
));
610 /* Even when the first slot is empty but the others aren't */
611 xa_store_index(xa
, 1, GFP_KERNEL
);
612 xa_store_index(xa
, 2, GFP_KERNEL
);
613 xa_store_order(xa
, 0, 2, NULL
, GFP_KERNEL
);
614 XA_BUG_ON(xa
, !xa_empty(xa
));
616 for (i
= 0; i
< max_order
; i
++) {
617 for (j
= 0; j
< max_order
; j
++) {
618 xa_store_order(xa
, 0, i
, xa_mk_index(i
), GFP_KERNEL
);
619 xa_store_order(xa
, 0, j
, xa_mk_index(j
), GFP_KERNEL
);
621 for (k
= 0; k
< max_order
; k
++) {
622 void *entry
= xa_load(xa
, (1UL << k
) - 1);
623 if ((i
< k
) && (j
< k
))
624 XA_BUG_ON(xa
, entry
!= NULL
);
626 XA_BUG_ON(xa
, entry
!= xa_mk_index(j
));
630 XA_BUG_ON(xa
, !xa_empty(xa
));
634 for (i
= 0; i
< 20; i
++) {
635 check_multi_store_1(xa
, 200, i
);
636 check_multi_store_1(xa
, 0, i
);
637 check_multi_store_1(xa
, (1UL << i
) + 1, i
);
639 check_multi_store_2(xa
, 4095, 9);
641 for (i
= 1; i
< 20; i
++) {
642 check_multi_store_3(xa
, 0, i
);
643 check_multi_store_3(xa
, 1UL << i
, i
);
648 static noinline
void check_xa_alloc_1(struct xarray
*xa
, unsigned int base
)
653 XA_BUG_ON(xa
, !xa_empty(xa
));
654 /* An empty array should assign %base to the first alloc */
655 xa_alloc_index(xa
, base
, GFP_KERNEL
);
657 /* Erasing it should make the array empty again */
658 xa_erase_index(xa
, base
);
659 XA_BUG_ON(xa
, !xa_empty(xa
));
661 /* And it should assign %base again */
662 xa_alloc_index(xa
, base
, GFP_KERNEL
);
664 /* Allocating and then erasing a lot should not lose base */
665 for (i
= base
+ 1; i
< 2 * XA_CHUNK_SIZE
; i
++)
666 xa_alloc_index(xa
, i
, GFP_KERNEL
);
667 for (i
= base
; i
< 2 * XA_CHUNK_SIZE
; i
++)
668 xa_erase_index(xa
, i
);
669 xa_alloc_index(xa
, base
, GFP_KERNEL
);
671 /* Destroying the array should do the same as erasing */
674 /* And it should assign %base again */
675 xa_alloc_index(xa
, base
, GFP_KERNEL
);
677 /* The next assigned ID should be base+1 */
678 xa_alloc_index(xa
, base
+ 1, GFP_KERNEL
);
679 xa_erase_index(xa
, base
+ 1);
681 /* Storing a value should mark it used */
682 xa_store_index(xa
, base
+ 1, GFP_KERNEL
);
683 xa_alloc_index(xa
, base
+ 2, GFP_KERNEL
);
685 /* If we then erase base, it should be free */
686 xa_erase_index(xa
, base
);
687 xa_alloc_index(xa
, base
, GFP_KERNEL
);
689 xa_erase_index(xa
, base
+ 1);
690 xa_erase_index(xa
, base
+ 2);
692 for (i
= 1; i
< 5000; i
++) {
693 xa_alloc_index(xa
, base
+ i
, GFP_KERNEL
);
698 /* Check that we fail properly at the limit of allocation */
699 XA_BUG_ON(xa
, xa_alloc(xa
, &id
, xa_mk_index(UINT_MAX
- 1),
700 XA_LIMIT(UINT_MAX
- 1, UINT_MAX
),
702 XA_BUG_ON(xa
, id
!= 0xfffffffeU
);
703 XA_BUG_ON(xa
, xa_alloc(xa
, &id
, xa_mk_index(UINT_MAX
),
704 XA_LIMIT(UINT_MAX
- 1, UINT_MAX
),
706 XA_BUG_ON(xa
, id
!= 0xffffffffU
);
708 XA_BUG_ON(xa
, xa_alloc(xa
, &id
, xa_mk_index(0),
709 XA_LIMIT(UINT_MAX
- 1, UINT_MAX
),
710 GFP_KERNEL
) != -EBUSY
);
711 XA_BUG_ON(xa
, id
!= 3);
714 XA_BUG_ON(xa
, xa_alloc(xa
, &id
, xa_mk_index(10), XA_LIMIT(10, 5),
715 GFP_KERNEL
) != -EBUSY
);
716 XA_BUG_ON(xa
, xa_store_index(xa
, 3, GFP_KERNEL
) != 0);
717 XA_BUG_ON(xa
, xa_alloc(xa
, &id
, xa_mk_index(10), XA_LIMIT(10, 5),
718 GFP_KERNEL
) != -EBUSY
);
719 xa_erase_index(xa
, 3);
720 XA_BUG_ON(xa
, !xa_empty(xa
));
723 static noinline
void check_xa_alloc_2(struct xarray
*xa
, unsigned int base
)
729 /* Allocate and free a NULL and check xa_empty() behaves */
730 XA_BUG_ON(xa
, !xa_empty(xa
));
731 XA_BUG_ON(xa
, xa_alloc(xa
, &id
, NULL
, xa_limit_32b
, GFP_KERNEL
) != 0);
732 XA_BUG_ON(xa
, id
!= base
);
733 XA_BUG_ON(xa
, xa_empty(xa
));
734 XA_BUG_ON(xa
, xa_erase(xa
, id
) != NULL
);
735 XA_BUG_ON(xa
, !xa_empty(xa
));
737 /* Ditto, but check destroy instead of erase */
738 XA_BUG_ON(xa
, !xa_empty(xa
));
739 XA_BUG_ON(xa
, xa_alloc(xa
, &id
, NULL
, xa_limit_32b
, GFP_KERNEL
) != 0);
740 XA_BUG_ON(xa
, id
!= base
);
741 XA_BUG_ON(xa
, xa_empty(xa
));
743 XA_BUG_ON(xa
, !xa_empty(xa
));
745 for (i
= base
; i
< base
+ 10; i
++) {
746 XA_BUG_ON(xa
, xa_alloc(xa
, &id
, NULL
, xa_limit_32b
,
748 XA_BUG_ON(xa
, id
!= i
);
751 XA_BUG_ON(xa
, xa_store(xa
, 3, xa_mk_index(3), GFP_KERNEL
) != NULL
);
752 XA_BUG_ON(xa
, xa_store(xa
, 4, xa_mk_index(4), GFP_KERNEL
) != NULL
);
753 XA_BUG_ON(xa
, xa_store(xa
, 4, NULL
, GFP_KERNEL
) != xa_mk_index(4));
754 XA_BUG_ON(xa
, xa_erase(xa
, 5) != NULL
);
755 XA_BUG_ON(xa
, xa_alloc(xa
, &id
, NULL
, xa_limit_32b
, GFP_KERNEL
) != 0);
756 XA_BUG_ON(xa
, id
!= 5);
758 xa_for_each(xa
, index
, entry
) {
759 xa_erase_index(xa
, index
);
762 for (i
= base
; i
< base
+ 9; i
++) {
763 XA_BUG_ON(xa
, xa_erase(xa
, i
) != NULL
);
764 XA_BUG_ON(xa
, xa_empty(xa
));
766 XA_BUG_ON(xa
, xa_erase(xa
, 8) != NULL
);
767 XA_BUG_ON(xa
, xa_empty(xa
));
768 XA_BUG_ON(xa
, xa_erase(xa
, base
+ 9) != NULL
);
769 XA_BUG_ON(xa
, !xa_empty(xa
));
774 static noinline
void check_xa_alloc_3(struct xarray
*xa
, unsigned int base
)
776 struct xa_limit limit
= XA_LIMIT(1, 0x3fff);
782 XA_BUG_ON(xa
, xa_alloc_cyclic(xa
, &id
, xa_mk_index(1), limit
,
783 &next
, GFP_KERNEL
) != 0);
784 XA_BUG_ON(xa
, id
!= 1);
787 XA_BUG_ON(xa
, xa_alloc_cyclic(xa
, &id
, xa_mk_index(0x3ffd), limit
,
788 &next
, GFP_KERNEL
) != 0);
789 XA_BUG_ON(xa
, id
!= 0x3ffd);
790 xa_erase_index(xa
, 0x3ffd);
791 xa_erase_index(xa
, 1);
792 XA_BUG_ON(xa
, !xa_empty(xa
));
794 for (i
= 0x3ffe; i
< 0x4003; i
++) {
796 entry
= xa_mk_index(i
);
798 entry
= xa_mk_index(i
- 0x3fff);
799 XA_BUG_ON(xa
, xa_alloc_cyclic(xa
, &id
, entry
, limit
,
800 &next
, GFP_KERNEL
) != (id
== 1));
801 XA_BUG_ON(xa
, xa_mk_index(id
) != entry
);
804 /* Check wrap-around is handled correctly */
806 xa_erase_index(xa
, base
);
807 xa_erase_index(xa
, base
+ 1);
809 XA_BUG_ON(xa
, xa_alloc_cyclic(xa
, &id
, xa_mk_index(UINT_MAX
),
810 xa_limit_32b
, &next
, GFP_KERNEL
) != 0);
811 XA_BUG_ON(xa
, id
!= UINT_MAX
);
812 XA_BUG_ON(xa
, xa_alloc_cyclic(xa
, &id
, xa_mk_index(base
),
813 xa_limit_32b
, &next
, GFP_KERNEL
) != 1);
814 XA_BUG_ON(xa
, id
!= base
);
815 XA_BUG_ON(xa
, xa_alloc_cyclic(xa
, &id
, xa_mk_index(base
+ 1),
816 xa_limit_32b
, &next
, GFP_KERNEL
) != 0);
817 XA_BUG_ON(xa
, id
!= base
+ 1);
819 xa_for_each(xa
, index
, entry
)
820 xa_erase_index(xa
, index
);
822 XA_BUG_ON(xa
, !xa_empty(xa
));
825 static DEFINE_XARRAY_ALLOC(xa0
);
826 static DEFINE_XARRAY_ALLOC1(xa1
);
828 static noinline
void check_xa_alloc(void)
830 check_xa_alloc_1(&xa0
, 0);
831 check_xa_alloc_1(&xa1
, 1);
832 check_xa_alloc_2(&xa0
, 0);
833 check_xa_alloc_2(&xa1
, 1);
834 check_xa_alloc_3(&xa0
, 0);
835 check_xa_alloc_3(&xa1
, 1);
838 static noinline
void __check_store_iter(struct xarray
*xa
, unsigned long start
,
839 unsigned int order
, unsigned int present
)
841 XA_STATE_ORDER(xas
, xa
, start
, order
);
843 unsigned int count
= 0;
847 xas_for_each_conflict(&xas
, entry
) {
848 XA_BUG_ON(xa
, !xa_is_value(entry
));
849 XA_BUG_ON(xa
, entry
< xa_mk_index(start
));
850 XA_BUG_ON(xa
, entry
> xa_mk_index(start
+ (1UL << order
) - 1));
853 xas_store(&xas
, xa_mk_index(start
));
855 if (xas_nomem(&xas
, GFP_KERNEL
)) {
859 XA_BUG_ON(xa
, xas_error(&xas
));
860 XA_BUG_ON(xa
, count
!= present
);
861 XA_BUG_ON(xa
, xa_load(xa
, start
) != xa_mk_index(start
));
862 XA_BUG_ON(xa
, xa_load(xa
, start
+ (1UL << order
) - 1) !=
864 xa_erase_index(xa
, start
);
867 static noinline
void check_store_iter(struct xarray
*xa
)
870 unsigned int max_order
= IS_ENABLED(CONFIG_XARRAY_MULTI
) ? 20 : 1;
872 for (i
= 0; i
< max_order
; i
++) {
873 unsigned int min
= 1 << i
;
874 unsigned int max
= (2 << i
) - 1;
875 __check_store_iter(xa
, 0, i
, 0);
876 XA_BUG_ON(xa
, !xa_empty(xa
));
877 __check_store_iter(xa
, min
, i
, 0);
878 XA_BUG_ON(xa
, !xa_empty(xa
));
880 xa_store_index(xa
, min
, GFP_KERNEL
);
881 __check_store_iter(xa
, min
, i
, 1);
882 XA_BUG_ON(xa
, !xa_empty(xa
));
883 xa_store_index(xa
, max
, GFP_KERNEL
);
884 __check_store_iter(xa
, min
, i
, 1);
885 XA_BUG_ON(xa
, !xa_empty(xa
));
887 for (j
= 0; j
< min
; j
++)
888 xa_store_index(xa
, j
, GFP_KERNEL
);
889 __check_store_iter(xa
, 0, i
, min
);
890 XA_BUG_ON(xa
, !xa_empty(xa
));
891 for (j
= 0; j
< min
; j
++)
892 xa_store_index(xa
, min
+ j
, GFP_KERNEL
);
893 __check_store_iter(xa
, min
, i
, min
);
894 XA_BUG_ON(xa
, !xa_empty(xa
));
896 #ifdef CONFIG_XARRAY_MULTI
897 xa_store_index(xa
, 63, GFP_KERNEL
);
898 xa_store_index(xa
, 65, GFP_KERNEL
);
899 __check_store_iter(xa
, 64, 2, 1);
900 xa_erase_index(xa
, 63);
902 XA_BUG_ON(xa
, !xa_empty(xa
));
905 static noinline
void check_multi_find(struct xarray
*xa
)
907 #ifdef CONFIG_XARRAY_MULTI
910 xa_store_order(xa
, 12, 2, xa_mk_value(12), GFP_KERNEL
);
911 XA_BUG_ON(xa
, xa_store_index(xa
, 16, GFP_KERNEL
) != NULL
);
914 XA_BUG_ON(xa
, xa_find(xa
, &index
, ULONG_MAX
, XA_PRESENT
) !=
916 XA_BUG_ON(xa
, index
!= 12);
918 XA_BUG_ON(xa
, xa_find(xa
, &index
, ULONG_MAX
, XA_PRESENT
) !=
920 XA_BUG_ON(xa
, (index
< 12) || (index
>= 16));
921 XA_BUG_ON(xa
, xa_find_after(xa
, &index
, ULONG_MAX
, XA_PRESENT
) !=
923 XA_BUG_ON(xa
, index
!= 16);
925 xa_erase_index(xa
, 12);
926 xa_erase_index(xa
, 16);
927 XA_BUG_ON(xa
, !xa_empty(xa
));
931 static noinline
void check_multi_find_2(struct xarray
*xa
)
933 unsigned int max_order
= IS_ENABLED(CONFIG_XARRAY_MULTI
) ? 10 : 1;
937 for (i
= 0; i
< max_order
; i
++) {
938 unsigned long index
= 1UL << i
;
939 for (j
= 0; j
< index
; j
++) {
940 XA_STATE(xas
, xa
, j
+ index
);
941 xa_store_index(xa
, index
- 1, GFP_KERNEL
);
942 xa_store_order(xa
, index
, i
, xa_mk_index(index
),
945 xas_for_each(&xas
, entry
, ULONG_MAX
) {
946 xa_erase_index(xa
, index
);
949 xa_erase_index(xa
, index
- 1);
950 XA_BUG_ON(xa
, !xa_empty(xa
));
955 static noinline
void check_find_1(struct xarray
*xa
)
957 unsigned long i
, j
, k
;
959 XA_BUG_ON(xa
, !xa_empty(xa
));
962 * Check xa_find with all pairs between 0 and 99 inclusive,
963 * starting at every index between 0 and 99
965 for (i
= 0; i
< 100; i
++) {
966 XA_BUG_ON(xa
, xa_store_index(xa
, i
, GFP_KERNEL
) != NULL
);
967 xa_set_mark(xa
, i
, XA_MARK_0
);
968 for (j
= 0; j
< i
; j
++) {
969 XA_BUG_ON(xa
, xa_store_index(xa
, j
, GFP_KERNEL
) !=
971 xa_set_mark(xa
, j
, XA_MARK_0
);
972 for (k
= 0; k
< 100; k
++) {
973 unsigned long index
= k
;
974 void *entry
= xa_find(xa
, &index
, ULONG_MAX
,
977 XA_BUG_ON(xa
, index
!= j
);
979 XA_BUG_ON(xa
, index
!= i
);
981 XA_BUG_ON(xa
, entry
!= NULL
);
984 entry
= xa_find(xa
, &index
, ULONG_MAX
,
987 XA_BUG_ON(xa
, index
!= j
);
989 XA_BUG_ON(xa
, index
!= i
);
991 XA_BUG_ON(xa
, entry
!= NULL
);
993 xa_erase_index(xa
, j
);
994 XA_BUG_ON(xa
, xa_get_mark(xa
, j
, XA_MARK_0
));
995 XA_BUG_ON(xa
, !xa_get_mark(xa
, i
, XA_MARK_0
));
997 xa_erase_index(xa
, i
);
998 XA_BUG_ON(xa
, xa_get_mark(xa
, i
, XA_MARK_0
));
1000 XA_BUG_ON(xa
, !xa_empty(xa
));
1003 static noinline
void check_find_2(struct xarray
*xa
)
1006 unsigned long i
, j
, index
;
1008 xa_for_each(xa
, index
, entry
) {
1009 XA_BUG_ON(xa
, true);
1012 for (i
= 0; i
< 1024; i
++) {
1013 xa_store_index(xa
, index
, GFP_KERNEL
);
1015 xa_for_each(xa
, index
, entry
) {
1016 XA_BUG_ON(xa
, xa_mk_index(index
) != entry
);
1017 XA_BUG_ON(xa
, index
!= j
++);
1024 static noinline
void check_find_3(struct xarray
*xa
)
1026 XA_STATE(xas
, xa
, 0);
1027 unsigned long i
, j
, k
;
1030 for (i
= 0; i
< 100; i
++) {
1031 for (j
= 0; j
< 100; j
++) {
1033 for (k
= 0; k
< 100; k
++) {
1035 xas_for_each_marked(&xas
, entry
, k
, XA_MARK_0
)
1039 xas
.xa_node
!= XAS_RESTART
);
1043 xa_store_index(xa
, i
, GFP_KERNEL
);
1044 xa_set_mark(xa
, i
, XA_MARK_0
);
1049 static noinline
void check_find(struct xarray
*xa
)
1054 check_multi_find(xa
);
1055 check_multi_find_2(xa
);
1058 /* See find_swap_entry() in mm/shmem.c */
1059 static noinline
unsigned long xa_find_entry(struct xarray
*xa
, void *item
)
1061 XA_STATE(xas
, xa
, 0);
1062 unsigned int checked
= 0;
1066 xas_for_each(&xas
, entry
, ULONG_MAX
) {
1067 if (xas_retry(&xas
, entry
))
1072 if ((checked
% 4) != 0)
1078 return entry
? xas
.xa_index
: -1;
1081 static noinline
void check_find_entry(struct xarray
*xa
)
1083 #ifdef CONFIG_XARRAY_MULTI
1085 unsigned long offset
, index
;
1087 for (order
= 0; order
< 20; order
++) {
1088 for (offset
= 0; offset
< (1UL << (order
+ 3));
1089 offset
+= (1UL << order
)) {
1090 for (index
= 0; index
< (1UL << (order
+ 5));
1091 index
+= (1UL << order
)) {
1092 xa_store_order(xa
, index
, order
,
1093 xa_mk_index(index
), GFP_KERNEL
);
1094 XA_BUG_ON(xa
, xa_load(xa
, index
) !=
1095 xa_mk_index(index
));
1096 XA_BUG_ON(xa
, xa_find_entry(xa
,
1097 xa_mk_index(index
)) != index
);
1099 XA_BUG_ON(xa
, xa_find_entry(xa
, xa
) != -1);
1105 XA_BUG_ON(xa
, xa_find_entry(xa
, xa
) != -1);
1106 xa_store_index(xa
, ULONG_MAX
, GFP_KERNEL
);
1107 XA_BUG_ON(xa
, xa_find_entry(xa
, xa
) != -1);
1108 XA_BUG_ON(xa
, xa_find_entry(xa
, xa_mk_index(ULONG_MAX
)) != -1);
1109 xa_erase_index(xa
, ULONG_MAX
);
1110 XA_BUG_ON(xa
, !xa_empty(xa
));
1113 static noinline
void check_move_tiny(struct xarray
*xa
)
1115 XA_STATE(xas
, xa
, 0);
1117 XA_BUG_ON(xa
, !xa_empty(xa
));
1119 XA_BUG_ON(xa
, xas_next(&xas
) != NULL
);
1120 XA_BUG_ON(xa
, xas_next(&xas
) != NULL
);
1122 xa_store_index(xa
, 0, GFP_KERNEL
);
1125 XA_BUG_ON(xa
, xas_next(&xas
) != xa_mk_index(0));
1126 XA_BUG_ON(xa
, xas_next(&xas
) != NULL
);
1128 XA_BUG_ON(xa
, xas_prev(&xas
) != xa_mk_index(0));
1129 XA_BUG_ON(xa
, xas_prev(&xas
) != NULL
);
1131 xa_erase_index(xa
, 0);
1132 XA_BUG_ON(xa
, !xa_empty(xa
));
1135 static noinline
void check_move_small(struct xarray
*xa
, unsigned long idx
)
1137 XA_STATE(xas
, xa
, 0);
1140 xa_store_index(xa
, 0, GFP_KERNEL
);
1141 xa_store_index(xa
, idx
, GFP_KERNEL
);
1144 for (i
= 0; i
< idx
* 4; i
++) {
1145 void *entry
= xas_next(&xas
);
1147 XA_BUG_ON(xa
, xas
.xa_node
== XAS_RESTART
);
1148 XA_BUG_ON(xa
, xas
.xa_index
!= i
);
1149 if (i
== 0 || i
== idx
)
1150 XA_BUG_ON(xa
, entry
!= xa_mk_index(i
));
1152 XA_BUG_ON(xa
, entry
!= NULL
);
1155 XA_BUG_ON(xa
, xas
.xa_index
!= i
);
1158 void *entry
= xas_prev(&xas
);
1161 XA_BUG_ON(xa
, xas
.xa_node
== XAS_RESTART
);
1162 XA_BUG_ON(xa
, xas
.xa_index
!= i
);
1163 if (i
== 0 || i
== idx
)
1164 XA_BUG_ON(xa
, entry
!= xa_mk_index(i
));
1166 XA_BUG_ON(xa
, entry
!= NULL
);
1169 xas_set(&xas
, ULONG_MAX
);
1170 XA_BUG_ON(xa
, xas_next(&xas
) != NULL
);
1171 XA_BUG_ON(xa
, xas
.xa_index
!= ULONG_MAX
);
1172 XA_BUG_ON(xa
, xas_next(&xas
) != xa_mk_value(0));
1173 XA_BUG_ON(xa
, xas
.xa_index
!= 0);
1174 XA_BUG_ON(xa
, xas_prev(&xas
) != NULL
);
1175 XA_BUG_ON(xa
, xas
.xa_index
!= ULONG_MAX
);
1178 xa_erase_index(xa
, 0);
1179 xa_erase_index(xa
, idx
);
1180 XA_BUG_ON(xa
, !xa_empty(xa
));
1183 static noinline
void check_move(struct xarray
*xa
)
1185 XA_STATE(xas
, xa
, (1 << 16) - 1);
1188 for (i
= 0; i
< (1 << 16); i
++)
1189 XA_BUG_ON(xa
, xa_store_index(xa
, i
, GFP_KERNEL
) != NULL
);
1193 void *entry
= xas_prev(&xas
);
1195 XA_BUG_ON(xa
, entry
!= xa_mk_index(i
));
1196 XA_BUG_ON(xa
, i
!= xas
.xa_index
);
1199 XA_BUG_ON(xa
, xas_prev(&xas
) != NULL
);
1200 XA_BUG_ON(xa
, xas
.xa_index
!= ULONG_MAX
);
1203 void *entry
= xas_next(&xas
);
1204 XA_BUG_ON(xa
, entry
!= xa_mk_index(i
));
1205 XA_BUG_ON(xa
, i
!= xas
.xa_index
);
1207 } while (i
< (1 << 16));
1210 for (i
= (1 << 8); i
< (1 << 15); i
++)
1211 xa_erase_index(xa
, i
);
1217 void *entry
= xas_prev(&xas
);
1219 if ((i
< (1 << 8)) || (i
>= (1 << 15)))
1220 XA_BUG_ON(xa
, entry
!= xa_mk_index(i
));
1222 XA_BUG_ON(xa
, entry
!= NULL
);
1223 XA_BUG_ON(xa
, i
!= xas
.xa_index
);
1226 XA_BUG_ON(xa
, xas_prev(&xas
) != NULL
);
1227 XA_BUG_ON(xa
, xas
.xa_index
!= ULONG_MAX
);
1230 void *entry
= xas_next(&xas
);
1231 if ((i
< (1 << 8)) || (i
>= (1 << 15)))
1232 XA_BUG_ON(xa
, entry
!= xa_mk_index(i
));
1234 XA_BUG_ON(xa
, entry
!= NULL
);
1235 XA_BUG_ON(xa
, i
!= xas
.xa_index
);
1237 } while (i
< (1 << 16));
1242 check_move_tiny(xa
);
1244 for (i
= 0; i
< 16; i
++)
1245 check_move_small(xa
, 1UL << i
);
1247 for (i
= 2; i
< 16; i
++)
1248 check_move_small(xa
, (1UL << i
) - 1);
1251 static noinline
void xa_store_many_order(struct xarray
*xa
,
1252 unsigned long index
, unsigned order
)
1254 XA_STATE_ORDER(xas
, xa
, index
, order
);
1259 XA_BUG_ON(xa
, xas_find_conflict(&xas
));
1260 xas_create_range(&xas
);
1261 if (xas_error(&xas
))
1263 for (i
= 0; i
< (1U << order
); i
++) {
1264 XA_BUG_ON(xa
, xas_store(&xas
, xa_mk_index(index
+ i
)));
1269 } while (xas_nomem(&xas
, GFP_KERNEL
));
1271 XA_BUG_ON(xa
, xas_error(&xas
));
1274 static noinline
void check_create_range_1(struct xarray
*xa
,
1275 unsigned long index
, unsigned order
)
1279 xa_store_many_order(xa
, index
, order
);
1280 for (i
= index
; i
< index
+ (1UL << order
); i
++)
1281 xa_erase_index(xa
, i
);
1282 XA_BUG_ON(xa
, !xa_empty(xa
));
1285 static noinline
void check_create_range_2(struct xarray
*xa
, unsigned order
)
1288 unsigned long nr
= 1UL << order
;
1290 for (i
= 0; i
< nr
* nr
; i
+= nr
)
1291 xa_store_many_order(xa
, i
, order
);
1292 for (i
= 0; i
< nr
* nr
; i
++)
1293 xa_erase_index(xa
, i
);
1294 XA_BUG_ON(xa
, !xa_empty(xa
));
1297 static noinline
void check_create_range_3(void)
1299 XA_STATE(xas
, NULL
, 0);
1300 xas_set_err(&xas
, -EEXIST
);
1301 xas_create_range(&xas
);
1302 XA_BUG_ON(NULL
, xas_error(&xas
) != -EEXIST
);
1305 static noinline
void check_create_range_4(struct xarray
*xa
,
1306 unsigned long index
, unsigned order
)
1308 XA_STATE_ORDER(xas
, xa
, index
, order
);
1309 unsigned long base
= xas
.xa_index
;
1310 unsigned long i
= 0;
1312 xa_store_index(xa
, index
, GFP_KERNEL
);
1315 xas_create_range(&xas
);
1316 if (xas_error(&xas
))
1318 for (i
= 0; i
< (1UL << order
); i
++) {
1319 void *old
= xas_store(&xas
, xa_mk_index(base
+ i
));
1320 if (xas
.xa_index
== index
)
1321 XA_BUG_ON(xa
, old
!= xa_mk_index(base
+ i
));
1323 XA_BUG_ON(xa
, old
!= NULL
);
1328 } while (xas_nomem(&xas
, GFP_KERNEL
));
1330 XA_BUG_ON(xa
, xas_error(&xas
));
1332 for (i
= base
; i
< base
+ (1UL << order
); i
++)
1333 xa_erase_index(xa
, i
);
1334 XA_BUG_ON(xa
, !xa_empty(xa
));
1337 static noinline
void check_create_range(struct xarray
*xa
)
1340 unsigned int max_order
= IS_ENABLED(CONFIG_XARRAY_MULTI
) ? 12 : 1;
1342 for (order
= 0; order
< max_order
; order
++) {
1343 check_create_range_1(xa
, 0, order
);
1344 check_create_range_1(xa
, 1U << order
, order
);
1345 check_create_range_1(xa
, 2U << order
, order
);
1346 check_create_range_1(xa
, 3U << order
, order
);
1347 check_create_range_1(xa
, 1U << 24, order
);
1349 check_create_range_2(xa
, order
);
1351 check_create_range_4(xa
, 0, order
);
1352 check_create_range_4(xa
, 1U << order
, order
);
1353 check_create_range_4(xa
, 2U << order
, order
);
1354 check_create_range_4(xa
, 3U << order
, order
);
1355 check_create_range_4(xa
, 1U << 24, order
);
1357 check_create_range_4(xa
, 1, order
);
1358 check_create_range_4(xa
, (1U << order
) + 1, order
);
1359 check_create_range_4(xa
, (2U << order
) + 1, order
);
1360 check_create_range_4(xa
, (2U << order
) - 1, order
);
1361 check_create_range_4(xa
, (3U << order
) + 1, order
);
1362 check_create_range_4(xa
, (3U << order
) - 1, order
);
1363 check_create_range_4(xa
, (1U << 24) + 1, order
);
1366 check_create_range_3();
1369 static noinline
void __check_store_range(struct xarray
*xa
, unsigned long first
,
1372 #ifdef CONFIG_XARRAY_MULTI
1373 xa_store_range(xa
, first
, last
, xa_mk_index(first
), GFP_KERNEL
);
1375 XA_BUG_ON(xa
, xa_load(xa
, first
) != xa_mk_index(first
));
1376 XA_BUG_ON(xa
, xa_load(xa
, last
) != xa_mk_index(first
));
1377 XA_BUG_ON(xa
, xa_load(xa
, first
- 1) != NULL
);
1378 XA_BUG_ON(xa
, xa_load(xa
, last
+ 1) != NULL
);
1380 xa_store_range(xa
, first
, last
, NULL
, GFP_KERNEL
);
1383 XA_BUG_ON(xa
, !xa_empty(xa
));
1386 static noinline
void check_store_range(struct xarray
*xa
)
1390 for (i
= 0; i
< 128; i
++) {
1391 for (j
= i
; j
< 128; j
++) {
1392 __check_store_range(xa
, i
, j
);
1393 __check_store_range(xa
, 128 + i
, 128 + j
);
1394 __check_store_range(xa
, 4095 + i
, 4095 + j
);
1395 __check_store_range(xa
, 4096 + i
, 4096 + j
);
1396 __check_store_range(xa
, 123456 + i
, 123456 + j
);
1397 __check_store_range(xa
, (1 << 24) + i
, (1 << 24) + j
);
1402 static void check_align_1(struct xarray
*xa
, char *name
)
1406 unsigned long index
;
1409 for (i
= 0; i
< 8; i
++) {
1410 XA_BUG_ON(xa
, xa_alloc(xa
, &id
, name
+ i
, xa_limit_32b
,
1412 XA_BUG_ON(xa
, id
!= i
);
1414 xa_for_each(xa
, index
, entry
)
1415 XA_BUG_ON(xa
, xa_is_err(entry
));
1420 * We should always be able to store without allocating memory after
1423 static void check_align_2(struct xarray
*xa
, char *name
)
1427 XA_BUG_ON(xa
, !xa_empty(xa
));
1429 for (i
= 0; i
< 8; i
++) {
1430 XA_BUG_ON(xa
, xa_store(xa
, 0, name
+ i
, GFP_KERNEL
) != NULL
);
1434 for (i
= 0; i
< 8; i
++) {
1435 XA_BUG_ON(xa
, xa_reserve(xa
, 0, GFP_KERNEL
) != 0);
1436 XA_BUG_ON(xa
, xa_store(xa
, 0, name
+ i
, 0) != NULL
);
1440 XA_BUG_ON(xa
, !xa_empty(xa
));
1443 static noinline
void check_align(struct xarray
*xa
)
1445 char name
[] = "Motorola 68000";
1447 check_align_1(xa
, name
);
1448 check_align_1(xa
, name
+ 1);
1449 check_align_1(xa
, name
+ 2);
1450 check_align_1(xa
, name
+ 3);
1451 check_align_2(xa
, name
);
1454 static LIST_HEAD(shadow_nodes
);
1456 static void test_update_node(struct xa_node
*node
)
1458 if (node
->count
&& node
->count
== node
->nr_values
) {
1459 if (list_empty(&node
->private_list
))
1460 list_add(&shadow_nodes
, &node
->private_list
);
1462 if (!list_empty(&node
->private_list
))
1463 list_del_init(&node
->private_list
);
1467 static noinline
void shadow_remove(struct xarray
*xa
)
1469 struct xa_node
*node
;
1472 while ((node
= list_first_entry_or_null(&shadow_nodes
,
1473 struct xa_node
, private_list
))) {
1474 XA_STATE(xas
, node
->array
, 0);
1475 XA_BUG_ON(xa
, node
->array
!= xa
);
1476 list_del_init(&node
->private_list
);
1477 xas
.xa_node
= xa_parent_locked(node
->array
, node
);
1478 xas
.xa_offset
= node
->offset
;
1479 xas
.xa_shift
= node
->shift
+ XA_CHUNK_SHIFT
;
1480 xas_set_update(&xas
, test_update_node
);
1481 xas_store(&xas
, NULL
);
1486 static noinline
void check_workingset(struct xarray
*xa
, unsigned long index
)
1488 XA_STATE(xas
, xa
, index
);
1489 xas_set_update(&xas
, test_update_node
);
1493 xas_store(&xas
, xa_mk_value(0));
1495 xas_store(&xas
, xa_mk_value(1));
1497 } while (xas_nomem(&xas
, GFP_KERNEL
));
1499 XA_BUG_ON(xa
, list_empty(&shadow_nodes
));
1503 xas_store(&xas
, &xas
);
1504 XA_BUG_ON(xa
, !list_empty(&shadow_nodes
));
1506 xas_store(&xas
, xa_mk_value(2));
1508 XA_BUG_ON(xa
, list_empty(&shadow_nodes
));
1511 XA_BUG_ON(xa
, !list_empty(&shadow_nodes
));
1512 XA_BUG_ON(xa
, !xa_empty(xa
));
1516 * Check that the pointer / value / sibling entries are accounted the
1517 * way we expect them to be.
1519 static noinline
void check_account(struct xarray
*xa
)
1521 #ifdef CONFIG_XARRAY_MULTI
1524 for (order
= 1; order
< 12; order
++) {
1525 XA_STATE(xas
, xa
, 1 << order
);
1527 xa_store_order(xa
, 0, order
, xa
, GFP_KERNEL
);
1530 XA_BUG_ON(xa
, xas
.xa_node
->count
== 0);
1531 XA_BUG_ON(xa
, xas
.xa_node
->count
> (1 << order
));
1532 XA_BUG_ON(xa
, xas
.xa_node
->nr_values
!= 0);
1535 xa_store_order(xa
, 1 << order
, order
, xa_mk_index(1UL << order
),
1537 XA_BUG_ON(xa
, xas
.xa_node
->count
!= xas
.xa_node
->nr_values
* 2);
1539 xa_erase(xa
, 1 << order
);
1540 XA_BUG_ON(xa
, xas
.xa_node
->nr_values
!= 0);
1543 XA_BUG_ON(xa
, !xa_empty(xa
));
1548 static noinline
void check_destroy(struct xarray
*xa
)
1550 unsigned long index
;
1552 XA_BUG_ON(xa
, !xa_empty(xa
));
1554 /* Destroying an empty array is a no-op */
1556 XA_BUG_ON(xa
, !xa_empty(xa
));
1558 /* Destroying an array with a single entry */
1559 for (index
= 0; index
< 1000; index
++) {
1560 xa_store_index(xa
, index
, GFP_KERNEL
);
1561 XA_BUG_ON(xa
, xa_empty(xa
));
1563 XA_BUG_ON(xa
, !xa_empty(xa
));
1566 /* Destroying an array with a single entry at ULONG_MAX */
1567 xa_store(xa
, ULONG_MAX
, xa
, GFP_KERNEL
);
1568 XA_BUG_ON(xa
, xa_empty(xa
));
1570 XA_BUG_ON(xa
, !xa_empty(xa
));
1572 #ifdef CONFIG_XARRAY_MULTI
1573 /* Destroying an array with a multi-index entry */
1574 xa_store_order(xa
, 1 << 11, 11, xa
, GFP_KERNEL
);
1575 XA_BUG_ON(xa
, xa_empty(xa
));
1577 XA_BUG_ON(xa
, !xa_empty(xa
));
1581 static DEFINE_XARRAY(array
);
1583 static int xarray_checks(void)
1585 check_xa_err(&array
);
1586 check_xas_retry(&array
);
1587 check_xa_load(&array
);
1588 check_xa_mark(&array
);
1589 check_xa_shrink(&array
);
1590 check_xas_erase(&array
);
1591 check_insert(&array
);
1592 check_cmpxchg(&array
);
1593 check_reserve(&array
);
1594 check_reserve(&xa0
);
1595 check_multi_store(&array
);
1598 check_find_entry(&array
);
1599 check_account(&array
);
1600 check_destroy(&array
);
1602 check_create_range(&array
);
1603 check_store_range(&array
);
1604 check_store_iter(&array
);
1607 check_workingset(&array
, 0);
1608 check_workingset(&array
, 64);
1609 check_workingset(&array
, 4096);
1611 printk("XArray: %u of %u tests passed\n", tests_passed
, tests_run
);
1612 return (tests_run
== tests_passed
) ? 0 : -EINVAL
;
1615 static void xarray_exit(void)
1619 module_init(xarray_checks
);
1620 module_exit(xarray_exit
);
1621 MODULE_AUTHOR("Matthew Wilcox <willy@infradead.org>");
1622 MODULE_LICENSE("GPL");