2 * btree.c - NILFS B-tree.
4 * Copyright (C) 2005-2008 Nippon Telegraph and Telephone Corporation.
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License as published by
8 * the Free Software Foundation; either version 2 of the License, or
9 * (at your option) any later version.
11 * This program is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 * GNU General Public License for more details.
16 * You should have received a copy of the GNU General Public License
17 * along with this program; if not, write to the Free Software
18 * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
20 * Written by Koji Sato <koji@osrg.net>.
23 #include <linux/slab.h>
24 #include <linux/string.h>
25 #include <linux/errno.h>
26 #include <linux/pagevec.h>
35 * struct nilfs_btree_path - A path on which B-tree operations are executed
36 * @bp_bh: buffer head of node block
37 * @bp_sib_bh: buffer head of sibling node block
38 * @bp_index: index of child node
39 * @bp_oldreq: ptr end request for old ptr
40 * @bp_newreq: ptr alloc request for new ptr
41 * @bp_op: rebalance operation
43 struct nilfs_btree_path
{
44 struct buffer_head
*bp_bh
;
45 struct buffer_head
*bp_sib_bh
;
47 union nilfs_bmap_ptr_req bp_oldreq
;
48 union nilfs_bmap_ptr_req bp_newreq
;
49 struct nilfs_btnode_chkey_ctxt bp_ctxt
;
50 void (*bp_op
)(struct nilfs_btree
*, struct nilfs_btree_path
*,
51 int, __u64
*, __u64
*);
55 * B-tree path operations
58 static struct kmem_cache
*nilfs_btree_path_cache
;
60 int __init
nilfs_btree_path_cache_init(void)
62 nilfs_btree_path_cache
=
63 kmem_cache_create("nilfs2_btree_path_cache",
64 sizeof(struct nilfs_btree_path
) *
65 NILFS_BTREE_LEVEL_MAX
, 0, 0, NULL
);
66 return (nilfs_btree_path_cache
!= NULL
) ? 0 : -ENOMEM
;
69 void nilfs_btree_path_cache_destroy(void)
71 kmem_cache_destroy(nilfs_btree_path_cache
);
74 static inline struct nilfs_btree_path
*nilfs_btree_alloc_path(void)
76 return kmem_cache_alloc(nilfs_btree_path_cache
, GFP_NOFS
);
79 static inline void nilfs_btree_free_path(struct nilfs_btree_path
*path
)
81 kmem_cache_free(nilfs_btree_path_cache
, path
);
84 static void nilfs_btree_init_path(struct nilfs_btree_path
*path
)
88 for (level
= NILFS_BTREE_LEVEL_DATA
;
89 level
< NILFS_BTREE_LEVEL_MAX
;
91 path
[level
].bp_bh
= NULL
;
92 path
[level
].bp_sib_bh
= NULL
;
93 path
[level
].bp_index
= 0;
94 path
[level
].bp_oldreq
.bpr_ptr
= NILFS_BMAP_INVALID_PTR
;
95 path
[level
].bp_newreq
.bpr_ptr
= NILFS_BMAP_INVALID_PTR
;
96 path
[level
].bp_op
= NULL
;
100 static void nilfs_btree_release_path(struct nilfs_btree_path
*path
)
104 for (level
= NILFS_BTREE_LEVEL_DATA
; level
< NILFS_BTREE_LEVEL_MAX
;
106 brelse(path
[level
].bp_bh
);
110 * B-tree node operations
112 static int nilfs_btree_get_block(const struct nilfs_btree
*btree
, __u64 ptr
,
113 struct buffer_head
**bhp
)
115 struct address_space
*btnc
=
116 &NILFS_BMAP_I((struct nilfs_bmap
*)btree
)->i_btnode_cache
;
119 err
= nilfs_btnode_submit_block(btnc
, ptr
, 0, bhp
);
121 return err
== -EEXIST
? 0 : err
;
123 wait_on_buffer(*bhp
);
124 if (!buffer_uptodate(*bhp
)) {
131 static int nilfs_btree_get_new_block(const struct nilfs_btree
*btree
,
132 __u64 ptr
, struct buffer_head
**bhp
)
134 struct address_space
*btnc
=
135 &NILFS_BMAP_I((struct nilfs_bmap
*)btree
)->i_btnode_cache
;
136 struct buffer_head
*bh
;
138 bh
= nilfs_btnode_create_block(btnc
, ptr
);
142 set_buffer_nilfs_volatile(bh
);
148 nilfs_btree_node_get_flags(const struct nilfs_btree_node
*node
)
150 return node
->bn_flags
;
154 nilfs_btree_node_set_flags(struct nilfs_btree_node
*node
, int flags
)
156 node
->bn_flags
= flags
;
159 static inline int nilfs_btree_node_root(const struct nilfs_btree_node
*node
)
161 return nilfs_btree_node_get_flags(node
) & NILFS_BTREE_NODE_ROOT
;
165 nilfs_btree_node_get_level(const struct nilfs_btree_node
*node
)
167 return node
->bn_level
;
171 nilfs_btree_node_set_level(struct nilfs_btree_node
*node
, int level
)
173 node
->bn_level
= level
;
177 nilfs_btree_node_get_nchildren(const struct nilfs_btree_node
*node
)
179 return le16_to_cpu(node
->bn_nchildren
);
183 nilfs_btree_node_set_nchildren(struct nilfs_btree_node
*node
, int nchildren
)
185 node
->bn_nchildren
= cpu_to_le16(nchildren
);
188 static inline int nilfs_btree_node_size(const struct nilfs_btree
*btree
)
190 return 1 << btree
->bt_bmap
.b_inode
->i_blkbits
;
194 nilfs_btree_node_nchildren_min(const struct nilfs_btree_node
*node
,
195 const struct nilfs_btree
*btree
)
197 return nilfs_btree_node_root(node
) ?
198 NILFS_BTREE_ROOT_NCHILDREN_MIN
:
199 NILFS_BTREE_NODE_NCHILDREN_MIN(nilfs_btree_node_size(btree
));
203 nilfs_btree_node_nchildren_max(const struct nilfs_btree_node
*node
,
204 const struct nilfs_btree
*btree
)
206 return nilfs_btree_node_root(node
) ?
207 NILFS_BTREE_ROOT_NCHILDREN_MAX
:
208 NILFS_BTREE_NODE_NCHILDREN_MAX(nilfs_btree_node_size(btree
));
211 static inline __le64
*
212 nilfs_btree_node_dkeys(const struct nilfs_btree_node
*node
)
214 return (__le64
*)((char *)(node
+ 1) +
215 (nilfs_btree_node_root(node
) ?
216 0 : NILFS_BTREE_NODE_EXTRA_PAD_SIZE
));
219 static inline __le64
*
220 nilfs_btree_node_dptrs(const struct nilfs_btree_node
*node
,
221 const struct nilfs_btree
*btree
)
223 return (__le64
*)(nilfs_btree_node_dkeys(node
) +
224 nilfs_btree_node_nchildren_max(node
, btree
));
228 nilfs_btree_node_get_key(const struct nilfs_btree_node
*node
, int index
)
230 return nilfs_bmap_dkey_to_key(*(nilfs_btree_node_dkeys(node
) + index
));
234 nilfs_btree_node_set_key(struct nilfs_btree_node
*node
, int index
, __u64 key
)
236 *(nilfs_btree_node_dkeys(node
) + index
) = nilfs_bmap_key_to_dkey(key
);
240 nilfs_btree_node_get_ptr(const struct nilfs_btree
*btree
,
241 const struct nilfs_btree_node
*node
, int index
)
243 return nilfs_bmap_dptr_to_ptr(*(nilfs_btree_node_dptrs(node
, btree
) +
248 nilfs_btree_node_set_ptr(struct nilfs_btree
*btree
,
249 struct nilfs_btree_node
*node
, int index
, __u64 ptr
)
251 *(nilfs_btree_node_dptrs(node
, btree
) + index
) =
252 nilfs_bmap_ptr_to_dptr(ptr
);
255 static void nilfs_btree_node_init(struct nilfs_btree
*btree
,
256 struct nilfs_btree_node
*node
,
257 int flags
, int level
, int nchildren
,
258 const __u64
*keys
, const __u64
*ptrs
)
264 nilfs_btree_node_set_flags(node
, flags
);
265 nilfs_btree_node_set_level(node
, level
);
266 nilfs_btree_node_set_nchildren(node
, nchildren
);
268 dkeys
= nilfs_btree_node_dkeys(node
);
269 dptrs
= nilfs_btree_node_dptrs(node
, btree
);
270 for (i
= 0; i
< nchildren
; i
++) {
271 dkeys
[i
] = nilfs_bmap_key_to_dkey(keys
[i
]);
272 dptrs
[i
] = nilfs_bmap_ptr_to_dptr(ptrs
[i
]);
276 /* Assume the buffer heads corresponding to left and right are locked. */
277 static void nilfs_btree_node_move_left(struct nilfs_btree
*btree
,
278 struct nilfs_btree_node
*left
,
279 struct nilfs_btree_node
*right
,
282 __le64
*ldkeys
, *rdkeys
;
283 __le64
*ldptrs
, *rdptrs
;
284 int lnchildren
, rnchildren
;
286 ldkeys
= nilfs_btree_node_dkeys(left
);
287 ldptrs
= nilfs_btree_node_dptrs(left
, btree
);
288 lnchildren
= nilfs_btree_node_get_nchildren(left
);
290 rdkeys
= nilfs_btree_node_dkeys(right
);
291 rdptrs
= nilfs_btree_node_dptrs(right
, btree
);
292 rnchildren
= nilfs_btree_node_get_nchildren(right
);
294 memcpy(ldkeys
+ lnchildren
, rdkeys
, n
* sizeof(*rdkeys
));
295 memcpy(ldptrs
+ lnchildren
, rdptrs
, n
* sizeof(*rdptrs
));
296 memmove(rdkeys
, rdkeys
+ n
, (rnchildren
- n
) * sizeof(*rdkeys
));
297 memmove(rdptrs
, rdptrs
+ n
, (rnchildren
- n
) * sizeof(*rdptrs
));
301 nilfs_btree_node_set_nchildren(left
, lnchildren
);
302 nilfs_btree_node_set_nchildren(right
, rnchildren
);
305 /* Assume that the buffer heads corresponding to left and right are locked. */
306 static void nilfs_btree_node_move_right(struct nilfs_btree
*btree
,
307 struct nilfs_btree_node
*left
,
308 struct nilfs_btree_node
*right
,
311 __le64
*ldkeys
, *rdkeys
;
312 __le64
*ldptrs
, *rdptrs
;
313 int lnchildren
, rnchildren
;
315 ldkeys
= nilfs_btree_node_dkeys(left
);
316 ldptrs
= nilfs_btree_node_dptrs(left
, btree
);
317 lnchildren
= nilfs_btree_node_get_nchildren(left
);
319 rdkeys
= nilfs_btree_node_dkeys(right
);
320 rdptrs
= nilfs_btree_node_dptrs(right
, btree
);
321 rnchildren
= nilfs_btree_node_get_nchildren(right
);
323 memmove(rdkeys
+ n
, rdkeys
, rnchildren
* sizeof(*rdkeys
));
324 memmove(rdptrs
+ n
, rdptrs
, rnchildren
* sizeof(*rdptrs
));
325 memcpy(rdkeys
, ldkeys
+ lnchildren
- n
, n
* sizeof(*rdkeys
));
326 memcpy(rdptrs
, ldptrs
+ lnchildren
- n
, n
* sizeof(*rdptrs
));
330 nilfs_btree_node_set_nchildren(left
, lnchildren
);
331 nilfs_btree_node_set_nchildren(right
, rnchildren
);
334 /* Assume that the buffer head corresponding to node is locked. */
335 static void nilfs_btree_node_insert(struct nilfs_btree
*btree
,
336 struct nilfs_btree_node
*node
,
337 __u64 key
, __u64 ptr
, int index
)
343 dkeys
= nilfs_btree_node_dkeys(node
);
344 dptrs
= nilfs_btree_node_dptrs(node
, btree
);
345 nchildren
= nilfs_btree_node_get_nchildren(node
);
346 if (index
< nchildren
) {
347 memmove(dkeys
+ index
+ 1, dkeys
+ index
,
348 (nchildren
- index
) * sizeof(*dkeys
));
349 memmove(dptrs
+ index
+ 1, dptrs
+ index
,
350 (nchildren
- index
) * sizeof(*dptrs
));
352 dkeys
[index
] = nilfs_bmap_key_to_dkey(key
);
353 dptrs
[index
] = nilfs_bmap_ptr_to_dptr(ptr
);
355 nilfs_btree_node_set_nchildren(node
, nchildren
);
358 /* Assume that the buffer head corresponding to node is locked. */
359 static void nilfs_btree_node_delete(struct nilfs_btree
*btree
,
360 struct nilfs_btree_node
*node
,
361 __u64
*keyp
, __u64
*ptrp
, int index
)
369 dkeys
= nilfs_btree_node_dkeys(node
);
370 dptrs
= nilfs_btree_node_dptrs(node
, btree
);
371 key
= nilfs_bmap_dkey_to_key(dkeys
[index
]);
372 ptr
= nilfs_bmap_dptr_to_ptr(dptrs
[index
]);
373 nchildren
= nilfs_btree_node_get_nchildren(node
);
379 if (index
< nchildren
- 1) {
380 memmove(dkeys
+ index
, dkeys
+ index
+ 1,
381 (nchildren
- index
- 1) * sizeof(*dkeys
));
382 memmove(dptrs
+ index
, dptrs
+ index
+ 1,
383 (nchildren
- index
- 1) * sizeof(*dptrs
));
386 nilfs_btree_node_set_nchildren(node
, nchildren
);
389 static int nilfs_btree_node_lookup(const struct nilfs_btree_node
*node
,
390 __u64 key
, int *indexp
)
393 int index
, low
, high
, s
;
397 high
= nilfs_btree_node_get_nchildren(node
) - 1;
400 while (low
<= high
) {
401 index
= (low
+ high
) / 2;
402 nkey
= nilfs_btree_node_get_key(node
, index
);
406 } else if (nkey
< key
) {
416 if (nilfs_btree_node_get_level(node
) > NILFS_BTREE_LEVEL_NODE_MIN
) {
417 if (s
> 0 && index
> 0)
428 static inline struct nilfs_btree_node
*
429 nilfs_btree_get_root(const struct nilfs_btree
*btree
)
431 return (struct nilfs_btree_node
*)btree
->bt_bmap
.b_u
.u_data
;
434 static inline struct nilfs_btree_node
*
435 nilfs_btree_get_nonroot_node(const struct nilfs_btree_path
*path
, int level
)
437 return (struct nilfs_btree_node
*)path
[level
].bp_bh
->b_data
;
440 static inline struct nilfs_btree_node
*
441 nilfs_btree_get_sib_node(const struct nilfs_btree_path
*path
, int level
)
443 return (struct nilfs_btree_node
*)path
[level
].bp_sib_bh
->b_data
;
446 static inline int nilfs_btree_height(const struct nilfs_btree
*btree
)
448 return nilfs_btree_node_get_level(nilfs_btree_get_root(btree
)) + 1;
451 static inline struct nilfs_btree_node
*
452 nilfs_btree_get_node(const struct nilfs_btree
*btree
,
453 const struct nilfs_btree_path
*path
,
456 return (level
== nilfs_btree_height(btree
) - 1) ?
457 nilfs_btree_get_root(btree
) :
458 nilfs_btree_get_nonroot_node(path
, level
);
462 nilfs_btree_bad_node(struct nilfs_btree_node
*node
, int level
)
464 if (unlikely(nilfs_btree_node_get_level(node
) != level
)) {
466 printk(KERN_CRIT
"NILFS: btree level mismatch: %d != %d\n",
467 nilfs_btree_node_get_level(node
), level
);
473 static int nilfs_btree_do_lookup(const struct nilfs_btree
*btree
,
474 struct nilfs_btree_path
*path
,
475 __u64 key
, __u64
*ptrp
, int minlevel
)
477 struct nilfs_btree_node
*node
;
479 int level
, index
, found
, ret
;
481 node
= nilfs_btree_get_root(btree
);
482 level
= nilfs_btree_node_get_level(node
);
483 if (level
< minlevel
|| nilfs_btree_node_get_nchildren(node
) <= 0)
486 found
= nilfs_btree_node_lookup(node
, key
, &index
);
487 ptr
= nilfs_btree_node_get_ptr(btree
, node
, index
);
488 path
[level
].bp_bh
= NULL
;
489 path
[level
].bp_index
= index
;
491 for (level
--; level
>= minlevel
; level
--) {
492 ret
= nilfs_btree_get_block(btree
, ptr
, &path
[level
].bp_bh
);
495 node
= nilfs_btree_get_nonroot_node(path
, level
);
496 if (nilfs_btree_bad_node(node
, level
))
499 found
= nilfs_btree_node_lookup(node
, key
, &index
);
502 if (index
< nilfs_btree_node_nchildren_max(node
, btree
))
503 ptr
= nilfs_btree_node_get_ptr(btree
, node
, index
);
505 WARN_ON(found
|| level
!= NILFS_BTREE_LEVEL_NODE_MIN
);
507 ptr
= NILFS_BMAP_INVALID_PTR
;
509 path
[level
].bp_index
= index
;
520 static int nilfs_btree_do_lookup_last(const struct nilfs_btree
*btree
,
521 struct nilfs_btree_path
*path
,
522 __u64
*keyp
, __u64
*ptrp
)
524 struct nilfs_btree_node
*node
;
526 int index
, level
, ret
;
528 node
= nilfs_btree_get_root(btree
);
529 index
= nilfs_btree_node_get_nchildren(node
) - 1;
532 level
= nilfs_btree_node_get_level(node
);
533 ptr
= nilfs_btree_node_get_ptr(btree
, node
, index
);
534 path
[level
].bp_bh
= NULL
;
535 path
[level
].bp_index
= index
;
537 for (level
--; level
> 0; level
--) {
538 ret
= nilfs_btree_get_block(btree
, ptr
, &path
[level
].bp_bh
);
541 node
= nilfs_btree_get_nonroot_node(path
, level
);
542 if (nilfs_btree_bad_node(node
, level
))
544 index
= nilfs_btree_node_get_nchildren(node
) - 1;
545 ptr
= nilfs_btree_node_get_ptr(btree
, node
, index
);
546 path
[level
].bp_index
= index
;
550 *keyp
= nilfs_btree_node_get_key(node
, index
);
557 static int nilfs_btree_lookup(const struct nilfs_bmap
*bmap
,
558 __u64 key
, int level
, __u64
*ptrp
)
560 struct nilfs_btree
*btree
;
561 struct nilfs_btree_path
*path
;
565 btree
= (struct nilfs_btree
*)bmap
;
566 path
= nilfs_btree_alloc_path();
569 nilfs_btree_init_path(path
);
571 ret
= nilfs_btree_do_lookup(btree
, path
, key
, &ptr
, level
);
576 nilfs_btree_release_path(path
);
577 nilfs_btree_free_path(path
);
582 static int nilfs_btree_lookup_contig(const struct nilfs_bmap
*bmap
,
583 __u64 key
, __u64
*ptrp
, unsigned maxblocks
)
585 struct nilfs_btree
*btree
= (struct nilfs_btree
*)bmap
;
586 struct nilfs_btree_path
*path
;
587 struct nilfs_btree_node
*node
;
588 struct inode
*dat
= NULL
;
591 int level
= NILFS_BTREE_LEVEL_NODE_MIN
;
592 int ret
, cnt
, index
, maxlevel
;
594 path
= nilfs_btree_alloc_path();
597 nilfs_btree_init_path(path
);
598 ret
= nilfs_btree_do_lookup(btree
, path
, key
, &ptr
, level
);
602 if (NILFS_BMAP_USE_VBN(bmap
)) {
603 dat
= nilfs_bmap_get_dat(bmap
);
604 ret
= nilfs_dat_translate(dat
, ptr
, &blocknr
);
610 if (cnt
== maxblocks
)
613 maxlevel
= nilfs_btree_height(btree
) - 1;
614 node
= nilfs_btree_get_node(btree
, path
, level
);
615 index
= path
[level
].bp_index
+ 1;
617 while (index
< nilfs_btree_node_get_nchildren(node
)) {
618 if (nilfs_btree_node_get_key(node
, index
) !=
621 ptr2
= nilfs_btree_node_get_ptr(btree
, node
, index
);
623 ret
= nilfs_dat_translate(dat
, ptr2
, &blocknr
);
628 if (ptr2
!= ptr
+ cnt
|| ++cnt
== maxblocks
)
633 if (level
== maxlevel
)
636 /* look-up right sibling node */
637 node
= nilfs_btree_get_node(btree
, path
, level
+ 1);
638 index
= path
[level
+ 1].bp_index
+ 1;
639 if (index
>= nilfs_btree_node_get_nchildren(node
) ||
640 nilfs_btree_node_get_key(node
, index
) != key
+ cnt
)
642 ptr2
= nilfs_btree_node_get_ptr(btree
, node
, index
);
643 path
[level
+ 1].bp_index
= index
;
645 brelse(path
[level
].bp_bh
);
646 path
[level
].bp_bh
= NULL
;
647 ret
= nilfs_btree_get_block(btree
, ptr2
, &path
[level
].bp_bh
);
650 node
= nilfs_btree_get_nonroot_node(path
, level
);
652 path
[level
].bp_index
= index
;
658 nilfs_btree_release_path(path
);
659 nilfs_btree_free_path(path
);
663 static void nilfs_btree_promote_key(struct nilfs_btree
*btree
,
664 struct nilfs_btree_path
*path
,
665 int level
, __u64 key
)
667 if (level
< nilfs_btree_height(btree
) - 1) {
669 nilfs_btree_node_set_key(
670 nilfs_btree_get_nonroot_node(path
, level
),
671 path
[level
].bp_index
, key
);
672 if (!buffer_dirty(path
[level
].bp_bh
))
673 nilfs_btnode_mark_dirty(path
[level
].bp_bh
);
674 } while ((path
[level
].bp_index
== 0) &&
675 (++level
< nilfs_btree_height(btree
) - 1));
679 if (level
== nilfs_btree_height(btree
) - 1) {
680 nilfs_btree_node_set_key(nilfs_btree_get_root(btree
),
681 path
[level
].bp_index
, key
);
685 static void nilfs_btree_do_insert(struct nilfs_btree
*btree
,
686 struct nilfs_btree_path
*path
,
687 int level
, __u64
*keyp
, __u64
*ptrp
)
689 struct nilfs_btree_node
*node
;
691 if (level
< nilfs_btree_height(btree
) - 1) {
692 node
= nilfs_btree_get_nonroot_node(path
, level
);
693 nilfs_btree_node_insert(btree
, node
, *keyp
, *ptrp
,
694 path
[level
].bp_index
);
695 if (!buffer_dirty(path
[level
].bp_bh
))
696 nilfs_btnode_mark_dirty(path
[level
].bp_bh
);
698 if (path
[level
].bp_index
== 0)
699 nilfs_btree_promote_key(btree
, path
, level
+ 1,
700 nilfs_btree_node_get_key(node
,
703 node
= nilfs_btree_get_root(btree
);
704 nilfs_btree_node_insert(btree
, node
, *keyp
, *ptrp
,
705 path
[level
].bp_index
);
709 static void nilfs_btree_carry_left(struct nilfs_btree
*btree
,
710 struct nilfs_btree_path
*path
,
711 int level
, __u64
*keyp
, __u64
*ptrp
)
713 struct nilfs_btree_node
*node
, *left
;
714 int nchildren
, lnchildren
, n
, move
;
716 node
= nilfs_btree_get_nonroot_node(path
, level
);
717 left
= nilfs_btree_get_sib_node(path
, level
);
718 nchildren
= nilfs_btree_node_get_nchildren(node
);
719 lnchildren
= nilfs_btree_node_get_nchildren(left
);
722 n
= (nchildren
+ lnchildren
+ 1) / 2 - lnchildren
;
723 if (n
> path
[level
].bp_index
) {
724 /* move insert point */
729 nilfs_btree_node_move_left(btree
, left
, node
, n
);
731 if (!buffer_dirty(path
[level
].bp_bh
))
732 nilfs_btnode_mark_dirty(path
[level
].bp_bh
);
733 if (!buffer_dirty(path
[level
].bp_sib_bh
))
734 nilfs_btnode_mark_dirty(path
[level
].bp_sib_bh
);
736 nilfs_btree_promote_key(btree
, path
, level
+ 1,
737 nilfs_btree_node_get_key(node
, 0));
740 brelse(path
[level
].bp_bh
);
741 path
[level
].bp_bh
= path
[level
].bp_sib_bh
;
742 path
[level
].bp_sib_bh
= NULL
;
743 path
[level
].bp_index
+= lnchildren
;
744 path
[level
+ 1].bp_index
--;
746 brelse(path
[level
].bp_sib_bh
);
747 path
[level
].bp_sib_bh
= NULL
;
748 path
[level
].bp_index
-= n
;
751 nilfs_btree_do_insert(btree
, path
, level
, keyp
, ptrp
);
754 static void nilfs_btree_carry_right(struct nilfs_btree
*btree
,
755 struct nilfs_btree_path
*path
,
756 int level
, __u64
*keyp
, __u64
*ptrp
)
758 struct nilfs_btree_node
*node
, *right
;
759 int nchildren
, rnchildren
, n
, move
;
761 node
= nilfs_btree_get_nonroot_node(path
, level
);
762 right
= nilfs_btree_get_sib_node(path
, level
);
763 nchildren
= nilfs_btree_node_get_nchildren(node
);
764 rnchildren
= nilfs_btree_node_get_nchildren(right
);
767 n
= (nchildren
+ rnchildren
+ 1) / 2 - rnchildren
;
768 if (n
> nchildren
- path
[level
].bp_index
) {
769 /* move insert point */
774 nilfs_btree_node_move_right(btree
, node
, right
, n
);
776 if (!buffer_dirty(path
[level
].bp_bh
))
777 nilfs_btnode_mark_dirty(path
[level
].bp_bh
);
778 if (!buffer_dirty(path
[level
].bp_sib_bh
))
779 nilfs_btnode_mark_dirty(path
[level
].bp_sib_bh
);
781 path
[level
+ 1].bp_index
++;
782 nilfs_btree_promote_key(btree
, path
, level
+ 1,
783 nilfs_btree_node_get_key(right
, 0));
784 path
[level
+ 1].bp_index
--;
787 brelse(path
[level
].bp_bh
);
788 path
[level
].bp_bh
= path
[level
].bp_sib_bh
;
789 path
[level
].bp_sib_bh
= NULL
;
790 path
[level
].bp_index
-= nilfs_btree_node_get_nchildren(node
);
791 path
[level
+ 1].bp_index
++;
793 brelse(path
[level
].bp_sib_bh
);
794 path
[level
].bp_sib_bh
= NULL
;
797 nilfs_btree_do_insert(btree
, path
, level
, keyp
, ptrp
);
800 static void nilfs_btree_split(struct nilfs_btree
*btree
,
801 struct nilfs_btree_path
*path
,
802 int level
, __u64
*keyp
, __u64
*ptrp
)
804 struct nilfs_btree_node
*node
, *right
;
807 int nchildren
, n
, move
;
809 node
= nilfs_btree_get_nonroot_node(path
, level
);
810 right
= nilfs_btree_get_sib_node(path
, level
);
811 nchildren
= nilfs_btree_node_get_nchildren(node
);
814 n
= (nchildren
+ 1) / 2;
815 if (n
> nchildren
- path
[level
].bp_index
) {
820 nilfs_btree_node_move_right(btree
, node
, right
, n
);
822 if (!buffer_dirty(path
[level
].bp_bh
))
823 nilfs_btnode_mark_dirty(path
[level
].bp_bh
);
824 if (!buffer_dirty(path
[level
].bp_sib_bh
))
825 nilfs_btnode_mark_dirty(path
[level
].bp_sib_bh
);
827 newkey
= nilfs_btree_node_get_key(right
, 0);
828 newptr
= path
[level
].bp_newreq
.bpr_ptr
;
831 path
[level
].bp_index
-= nilfs_btree_node_get_nchildren(node
);
832 nilfs_btree_node_insert(btree
, right
, *keyp
, *ptrp
,
833 path
[level
].bp_index
);
835 *keyp
= nilfs_btree_node_get_key(right
, 0);
836 *ptrp
= path
[level
].bp_newreq
.bpr_ptr
;
838 brelse(path
[level
].bp_bh
);
839 path
[level
].bp_bh
= path
[level
].bp_sib_bh
;
840 path
[level
].bp_sib_bh
= NULL
;
842 nilfs_btree_do_insert(btree
, path
, level
, keyp
, ptrp
);
844 *keyp
= nilfs_btree_node_get_key(right
, 0);
845 *ptrp
= path
[level
].bp_newreq
.bpr_ptr
;
847 brelse(path
[level
].bp_sib_bh
);
848 path
[level
].bp_sib_bh
= NULL
;
851 path
[level
+ 1].bp_index
++;
854 static void nilfs_btree_grow(struct nilfs_btree
*btree
,
855 struct nilfs_btree_path
*path
,
856 int level
, __u64
*keyp
, __u64
*ptrp
)
858 struct nilfs_btree_node
*root
, *child
;
861 root
= nilfs_btree_get_root(btree
);
862 child
= nilfs_btree_get_sib_node(path
, level
);
864 n
= nilfs_btree_node_get_nchildren(root
);
866 nilfs_btree_node_move_right(btree
, root
, child
, n
);
867 nilfs_btree_node_set_level(root
, level
+ 1);
869 if (!buffer_dirty(path
[level
].bp_sib_bh
))
870 nilfs_btnode_mark_dirty(path
[level
].bp_sib_bh
);
872 path
[level
].bp_bh
= path
[level
].bp_sib_bh
;
873 path
[level
].bp_sib_bh
= NULL
;
875 nilfs_btree_do_insert(btree
, path
, level
, keyp
, ptrp
);
877 *keyp
= nilfs_btree_node_get_key(child
, 0);
878 *ptrp
= path
[level
].bp_newreq
.bpr_ptr
;
881 static __u64
nilfs_btree_find_near(const struct nilfs_btree
*btree
,
882 const struct nilfs_btree_path
*path
)
884 struct nilfs_btree_node
*node
;
888 return NILFS_BMAP_INVALID_PTR
;
891 level
= NILFS_BTREE_LEVEL_NODE_MIN
;
892 if (path
[level
].bp_index
> 0) {
893 node
= nilfs_btree_get_node(btree
, path
, level
);
894 return nilfs_btree_node_get_ptr(btree
, node
,
895 path
[level
].bp_index
- 1);
899 level
= NILFS_BTREE_LEVEL_NODE_MIN
+ 1;
900 if (level
<= nilfs_btree_height(btree
) - 1) {
901 node
= nilfs_btree_get_node(btree
, path
, level
);
902 return nilfs_btree_node_get_ptr(btree
, node
,
903 path
[level
].bp_index
);
906 return NILFS_BMAP_INVALID_PTR
;
909 static __u64
nilfs_btree_find_target_v(const struct nilfs_btree
*btree
,
910 const struct nilfs_btree_path
*path
,
915 ptr
= nilfs_bmap_find_target_seq(&btree
->bt_bmap
, key
);
916 if (ptr
!= NILFS_BMAP_INVALID_PTR
)
917 /* sequential access */
920 ptr
= nilfs_btree_find_near(btree
, path
);
921 if (ptr
!= NILFS_BMAP_INVALID_PTR
)
926 return nilfs_bmap_find_target_in_group(&btree
->bt_bmap
);
929 static void nilfs_btree_set_target_v(struct nilfs_btree
*btree
, __u64 key
,
932 btree
->bt_bmap
.b_last_allocated_key
= key
;
933 btree
->bt_bmap
.b_last_allocated_ptr
= ptr
;
936 static int nilfs_btree_prepare_insert(struct nilfs_btree
*btree
,
937 struct nilfs_btree_path
*path
,
938 int *levelp
, __u64 key
, __u64 ptr
,
939 struct nilfs_bmap_stats
*stats
)
941 struct buffer_head
*bh
;
942 struct nilfs_btree_node
*node
, *parent
, *sib
;
944 int pindex
, level
, ret
;
945 struct inode
*dat
= NULL
;
947 stats
->bs_nblocks
= 0;
948 level
= NILFS_BTREE_LEVEL_DATA
;
950 /* allocate a new ptr for data block */
951 if (NILFS_BMAP_USE_VBN(&btree
->bt_bmap
)) {
952 path
[level
].bp_newreq
.bpr_ptr
=
953 nilfs_btree_find_target_v(btree
, path
, key
);
954 dat
= nilfs_bmap_get_dat(&btree
->bt_bmap
);
957 ret
= nilfs_bmap_prepare_alloc_ptr(&btree
->bt_bmap
,
958 &path
[level
].bp_newreq
, dat
);
962 for (level
= NILFS_BTREE_LEVEL_NODE_MIN
;
963 level
< nilfs_btree_height(btree
) - 1;
965 node
= nilfs_btree_get_nonroot_node(path
, level
);
966 if (nilfs_btree_node_get_nchildren(node
) <
967 nilfs_btree_node_nchildren_max(node
, btree
)) {
968 path
[level
].bp_op
= nilfs_btree_do_insert
;
973 parent
= nilfs_btree_get_node(btree
, path
, level
+ 1);
974 pindex
= path
[level
+ 1].bp_index
;
978 sibptr
= nilfs_btree_node_get_ptr(btree
, parent
,
980 ret
= nilfs_btree_get_block(btree
, sibptr
, &bh
);
982 goto err_out_child_node
;
983 sib
= (struct nilfs_btree_node
*)bh
->b_data
;
984 if (nilfs_btree_node_get_nchildren(sib
) <
985 nilfs_btree_node_nchildren_max(sib
, btree
)) {
986 path
[level
].bp_sib_bh
= bh
;
987 path
[level
].bp_op
= nilfs_btree_carry_left
;
996 nilfs_btree_node_get_nchildren(parent
) - 1) {
997 sibptr
= nilfs_btree_node_get_ptr(btree
, parent
,
999 ret
= nilfs_btree_get_block(btree
, sibptr
, &bh
);
1001 goto err_out_child_node
;
1002 sib
= (struct nilfs_btree_node
*)bh
->b_data
;
1003 if (nilfs_btree_node_get_nchildren(sib
) <
1004 nilfs_btree_node_nchildren_max(sib
, btree
)) {
1005 path
[level
].bp_sib_bh
= bh
;
1006 path
[level
].bp_op
= nilfs_btree_carry_right
;
1007 stats
->bs_nblocks
++;
1014 path
[level
].bp_newreq
.bpr_ptr
=
1015 path
[level
- 1].bp_newreq
.bpr_ptr
+ 1;
1016 ret
= nilfs_bmap_prepare_alloc_ptr(&btree
->bt_bmap
,
1017 &path
[level
].bp_newreq
, dat
);
1019 goto err_out_child_node
;
1020 ret
= nilfs_btree_get_new_block(btree
,
1021 path
[level
].bp_newreq
.bpr_ptr
,
1024 goto err_out_curr_node
;
1026 stats
->bs_nblocks
++;
1028 nilfs_btree_node_init(btree
,
1029 (struct nilfs_btree_node
*)bh
->b_data
,
1030 0, level
, 0, NULL
, NULL
);
1031 path
[level
].bp_sib_bh
= bh
;
1032 path
[level
].bp_op
= nilfs_btree_split
;
1036 node
= nilfs_btree_get_root(btree
);
1037 if (nilfs_btree_node_get_nchildren(node
) <
1038 nilfs_btree_node_nchildren_max(node
, btree
)) {
1039 path
[level
].bp_op
= nilfs_btree_do_insert
;
1040 stats
->bs_nblocks
++;
1045 path
[level
].bp_newreq
.bpr_ptr
= path
[level
- 1].bp_newreq
.bpr_ptr
+ 1;
1046 ret
= nilfs_bmap_prepare_alloc_ptr(&btree
->bt_bmap
,
1047 &path
[level
].bp_newreq
, dat
);
1049 goto err_out_child_node
;
1050 ret
= nilfs_btree_get_new_block(btree
, path
[level
].bp_newreq
.bpr_ptr
,
1053 goto err_out_curr_node
;
1055 nilfs_btree_node_init(btree
, (struct nilfs_btree_node
*)bh
->b_data
,
1056 0, level
, 0, NULL
, NULL
);
1057 path
[level
].bp_sib_bh
= bh
;
1058 path
[level
].bp_op
= nilfs_btree_grow
;
1061 path
[level
].bp_op
= nilfs_btree_do_insert
;
1063 /* a newly-created node block and a data block are added */
1064 stats
->bs_nblocks
+= 2;
1073 nilfs_bmap_abort_alloc_ptr(&btree
->bt_bmap
, &path
[level
].bp_newreq
,
1076 for (level
--; level
> NILFS_BTREE_LEVEL_DATA
; level
--) {
1077 nilfs_btnode_delete(path
[level
].bp_sib_bh
);
1078 nilfs_bmap_abort_alloc_ptr(&btree
->bt_bmap
,
1079 &path
[level
].bp_newreq
, dat
);
1083 nilfs_bmap_abort_alloc_ptr(&btree
->bt_bmap
, &path
[level
].bp_newreq
,
1087 stats
->bs_nblocks
= 0;
1091 static void nilfs_btree_commit_insert(struct nilfs_btree
*btree
,
1092 struct nilfs_btree_path
*path
,
1093 int maxlevel
, __u64 key
, __u64 ptr
)
1095 struct inode
*dat
= NULL
;
1098 set_buffer_nilfs_volatile((struct buffer_head
*)((unsigned long)ptr
));
1099 ptr
= path
[NILFS_BTREE_LEVEL_DATA
].bp_newreq
.bpr_ptr
;
1100 if (NILFS_BMAP_USE_VBN(&btree
->bt_bmap
)) {
1101 nilfs_btree_set_target_v(btree
, key
, ptr
);
1102 dat
= nilfs_bmap_get_dat(&btree
->bt_bmap
);
1105 for (level
= NILFS_BTREE_LEVEL_NODE_MIN
; level
<= maxlevel
; level
++) {
1106 nilfs_bmap_commit_alloc_ptr(&btree
->bt_bmap
,
1107 &path
[level
- 1].bp_newreq
, dat
);
1108 path
[level
].bp_op(btree
, path
, level
, &key
, &ptr
);
1111 if (!nilfs_bmap_dirty(&btree
->bt_bmap
))
1112 nilfs_bmap_set_dirty(&btree
->bt_bmap
);
1115 static int nilfs_btree_insert(struct nilfs_bmap
*bmap
, __u64 key
, __u64 ptr
)
1117 struct nilfs_btree
*btree
;
1118 struct nilfs_btree_path
*path
;
1119 struct nilfs_bmap_stats stats
;
1122 btree
= (struct nilfs_btree
*)bmap
;
1123 path
= nilfs_btree_alloc_path();
1126 nilfs_btree_init_path(path
);
1128 ret
= nilfs_btree_do_lookup(btree
, path
, key
, NULL
,
1129 NILFS_BTREE_LEVEL_NODE_MIN
);
1130 if (ret
!= -ENOENT
) {
1136 ret
= nilfs_btree_prepare_insert(btree
, path
, &level
, key
, ptr
, &stats
);
1139 nilfs_btree_commit_insert(btree
, path
, level
, key
, ptr
);
1140 nilfs_bmap_add_blocks(bmap
, stats
.bs_nblocks
);
1143 nilfs_btree_release_path(path
);
1144 nilfs_btree_free_path(path
);
1148 static void nilfs_btree_do_delete(struct nilfs_btree
*btree
,
1149 struct nilfs_btree_path
*path
,
1150 int level
, __u64
*keyp
, __u64
*ptrp
)
1152 struct nilfs_btree_node
*node
;
1154 if (level
< nilfs_btree_height(btree
) - 1) {
1155 node
= nilfs_btree_get_nonroot_node(path
, level
);
1156 nilfs_btree_node_delete(btree
, node
, keyp
, ptrp
,
1157 path
[level
].bp_index
);
1158 if (!buffer_dirty(path
[level
].bp_bh
))
1159 nilfs_btnode_mark_dirty(path
[level
].bp_bh
);
1160 if (path
[level
].bp_index
== 0)
1161 nilfs_btree_promote_key(btree
, path
, level
+ 1,
1162 nilfs_btree_node_get_key(node
, 0));
1164 node
= nilfs_btree_get_root(btree
);
1165 nilfs_btree_node_delete(btree
, node
, keyp
, ptrp
,
1166 path
[level
].bp_index
);
1170 static void nilfs_btree_borrow_left(struct nilfs_btree
*btree
,
1171 struct nilfs_btree_path
*path
,
1172 int level
, __u64
*keyp
, __u64
*ptrp
)
1174 struct nilfs_btree_node
*node
, *left
;
1175 int nchildren
, lnchildren
, n
;
1177 nilfs_btree_do_delete(btree
, path
, level
, keyp
, ptrp
);
1179 node
= nilfs_btree_get_nonroot_node(path
, level
);
1180 left
= nilfs_btree_get_sib_node(path
, level
);
1181 nchildren
= nilfs_btree_node_get_nchildren(node
);
1182 lnchildren
= nilfs_btree_node_get_nchildren(left
);
1184 n
= (nchildren
+ lnchildren
) / 2 - nchildren
;
1186 nilfs_btree_node_move_right(btree
, left
, node
, n
);
1188 if (!buffer_dirty(path
[level
].bp_bh
))
1189 nilfs_btnode_mark_dirty(path
[level
].bp_bh
);
1190 if (!buffer_dirty(path
[level
].bp_sib_bh
))
1191 nilfs_btnode_mark_dirty(path
[level
].bp_sib_bh
);
1193 nilfs_btree_promote_key(btree
, path
, level
+ 1,
1194 nilfs_btree_node_get_key(node
, 0));
1196 brelse(path
[level
].bp_sib_bh
);
1197 path
[level
].bp_sib_bh
= NULL
;
1198 path
[level
].bp_index
+= n
;
1201 static void nilfs_btree_borrow_right(struct nilfs_btree
*btree
,
1202 struct nilfs_btree_path
*path
,
1203 int level
, __u64
*keyp
, __u64
*ptrp
)
1205 struct nilfs_btree_node
*node
, *right
;
1206 int nchildren
, rnchildren
, n
;
1208 nilfs_btree_do_delete(btree
, path
, level
, keyp
, ptrp
);
1210 node
= nilfs_btree_get_nonroot_node(path
, level
);
1211 right
= nilfs_btree_get_sib_node(path
, level
);
1212 nchildren
= nilfs_btree_node_get_nchildren(node
);
1213 rnchildren
= nilfs_btree_node_get_nchildren(right
);
1215 n
= (nchildren
+ rnchildren
) / 2 - nchildren
;
1217 nilfs_btree_node_move_left(btree
, node
, right
, n
);
1219 if (!buffer_dirty(path
[level
].bp_bh
))
1220 nilfs_btnode_mark_dirty(path
[level
].bp_bh
);
1221 if (!buffer_dirty(path
[level
].bp_sib_bh
))
1222 nilfs_btnode_mark_dirty(path
[level
].bp_sib_bh
);
1224 path
[level
+ 1].bp_index
++;
1225 nilfs_btree_promote_key(btree
, path
, level
+ 1,
1226 nilfs_btree_node_get_key(right
, 0));
1227 path
[level
+ 1].bp_index
--;
1229 brelse(path
[level
].bp_sib_bh
);
1230 path
[level
].bp_sib_bh
= NULL
;
1233 static void nilfs_btree_concat_left(struct nilfs_btree
*btree
,
1234 struct nilfs_btree_path
*path
,
1235 int level
, __u64
*keyp
, __u64
*ptrp
)
1237 struct nilfs_btree_node
*node
, *left
;
1240 nilfs_btree_do_delete(btree
, path
, level
, keyp
, ptrp
);
1242 node
= nilfs_btree_get_nonroot_node(path
, level
);
1243 left
= nilfs_btree_get_sib_node(path
, level
);
1245 n
= nilfs_btree_node_get_nchildren(node
);
1247 nilfs_btree_node_move_left(btree
, left
, node
, n
);
1249 if (!buffer_dirty(path
[level
].bp_sib_bh
))
1250 nilfs_btnode_mark_dirty(path
[level
].bp_sib_bh
);
1252 nilfs_btnode_delete(path
[level
].bp_bh
);
1253 path
[level
].bp_bh
= path
[level
].bp_sib_bh
;
1254 path
[level
].bp_sib_bh
= NULL
;
1255 path
[level
].bp_index
+= nilfs_btree_node_get_nchildren(left
);
1258 static void nilfs_btree_concat_right(struct nilfs_btree
*btree
,
1259 struct nilfs_btree_path
*path
,
1260 int level
, __u64
*keyp
, __u64
*ptrp
)
1262 struct nilfs_btree_node
*node
, *right
;
1265 nilfs_btree_do_delete(btree
, path
, level
, keyp
, ptrp
);
1267 node
= nilfs_btree_get_nonroot_node(path
, level
);
1268 right
= nilfs_btree_get_sib_node(path
, level
);
1270 n
= nilfs_btree_node_get_nchildren(right
);
1272 nilfs_btree_node_move_left(btree
, node
, right
, n
);
1274 if (!buffer_dirty(path
[level
].bp_bh
))
1275 nilfs_btnode_mark_dirty(path
[level
].bp_bh
);
1277 nilfs_btnode_delete(path
[level
].bp_sib_bh
);
1278 path
[level
].bp_sib_bh
= NULL
;
1279 path
[level
+ 1].bp_index
++;
1282 static void nilfs_btree_shrink(struct nilfs_btree
*btree
,
1283 struct nilfs_btree_path
*path
,
1284 int level
, __u64
*keyp
, __u64
*ptrp
)
1286 struct nilfs_btree_node
*root
, *child
;
1289 nilfs_btree_do_delete(btree
, path
, level
, keyp
, ptrp
);
1291 root
= nilfs_btree_get_root(btree
);
1292 child
= nilfs_btree_get_nonroot_node(path
, level
);
1294 nilfs_btree_node_delete(btree
, root
, NULL
, NULL
, 0);
1295 nilfs_btree_node_set_level(root
, level
);
1296 n
= nilfs_btree_node_get_nchildren(child
);
1297 nilfs_btree_node_move_left(btree
, root
, child
, n
);
1299 nilfs_btnode_delete(path
[level
].bp_bh
);
1300 path
[level
].bp_bh
= NULL
;
1304 static int nilfs_btree_prepare_delete(struct nilfs_btree
*btree
,
1305 struct nilfs_btree_path
*path
,
1307 struct nilfs_bmap_stats
*stats
,
1310 struct buffer_head
*bh
;
1311 struct nilfs_btree_node
*node
, *parent
, *sib
;
1313 int pindex
, level
, ret
;
1316 stats
->bs_nblocks
= 0;
1317 for (level
= NILFS_BTREE_LEVEL_NODE_MIN
;
1318 level
< nilfs_btree_height(btree
) - 1;
1320 node
= nilfs_btree_get_nonroot_node(path
, level
);
1321 path
[level
].bp_oldreq
.bpr_ptr
=
1322 nilfs_btree_node_get_ptr(btree
, node
,
1323 path
[level
].bp_index
);
1324 ret
= nilfs_bmap_prepare_end_ptr(&btree
->bt_bmap
,
1325 &path
[level
].bp_oldreq
, dat
);
1327 goto err_out_child_node
;
1329 if (nilfs_btree_node_get_nchildren(node
) >
1330 nilfs_btree_node_nchildren_min(node
, btree
)) {
1331 path
[level
].bp_op
= nilfs_btree_do_delete
;
1332 stats
->bs_nblocks
++;
1336 parent
= nilfs_btree_get_node(btree
, path
, level
+ 1);
1337 pindex
= path
[level
+ 1].bp_index
;
1341 sibptr
= nilfs_btree_node_get_ptr(btree
, parent
,
1343 ret
= nilfs_btree_get_block(btree
, sibptr
, &bh
);
1345 goto err_out_curr_node
;
1346 sib
= (struct nilfs_btree_node
*)bh
->b_data
;
1347 if (nilfs_btree_node_get_nchildren(sib
) >
1348 nilfs_btree_node_nchildren_min(sib
, btree
)) {
1349 path
[level
].bp_sib_bh
= bh
;
1350 path
[level
].bp_op
= nilfs_btree_borrow_left
;
1351 stats
->bs_nblocks
++;
1354 path
[level
].bp_sib_bh
= bh
;
1355 path
[level
].bp_op
= nilfs_btree_concat_left
;
1356 stats
->bs_nblocks
++;
1360 nilfs_btree_node_get_nchildren(parent
) - 1) {
1362 sibptr
= nilfs_btree_node_get_ptr(btree
, parent
,
1364 ret
= nilfs_btree_get_block(btree
, sibptr
, &bh
);
1366 goto err_out_curr_node
;
1367 sib
= (struct nilfs_btree_node
*)bh
->b_data
;
1368 if (nilfs_btree_node_get_nchildren(sib
) >
1369 nilfs_btree_node_nchildren_min(sib
, btree
)) {
1370 path
[level
].bp_sib_bh
= bh
;
1371 path
[level
].bp_op
= nilfs_btree_borrow_right
;
1372 stats
->bs_nblocks
++;
1375 path
[level
].bp_sib_bh
= bh
;
1376 path
[level
].bp_op
= nilfs_btree_concat_right
;
1377 stats
->bs_nblocks
++;
1382 /* the only child of the root node */
1383 WARN_ON(level
!= nilfs_btree_height(btree
) - 2);
1384 if (nilfs_btree_node_get_nchildren(node
) - 1 <=
1385 NILFS_BTREE_ROOT_NCHILDREN_MAX
) {
1386 path
[level
].bp_op
= nilfs_btree_shrink
;
1387 stats
->bs_nblocks
+= 2;
1389 path
[level
].bp_op
= nilfs_btree_do_delete
;
1390 stats
->bs_nblocks
++;
1398 node
= nilfs_btree_get_root(btree
);
1399 path
[level
].bp_oldreq
.bpr_ptr
=
1400 nilfs_btree_node_get_ptr(btree
, node
, path
[level
].bp_index
);
1402 ret
= nilfs_bmap_prepare_end_ptr(&btree
->bt_bmap
,
1403 &path
[level
].bp_oldreq
, dat
);
1405 goto err_out_child_node
;
1407 /* child of the root node is deleted */
1408 path
[level
].bp_op
= nilfs_btree_do_delete
;
1409 stats
->bs_nblocks
++;
1418 nilfs_bmap_abort_end_ptr(&btree
->bt_bmap
, &path
[level
].bp_oldreq
, dat
);
1420 for (level
--; level
>= NILFS_BTREE_LEVEL_NODE_MIN
; level
--) {
1421 brelse(path
[level
].bp_sib_bh
);
1422 nilfs_bmap_abort_end_ptr(&btree
->bt_bmap
,
1423 &path
[level
].bp_oldreq
, dat
);
1426 stats
->bs_nblocks
= 0;
1430 static void nilfs_btree_commit_delete(struct nilfs_btree
*btree
,
1431 struct nilfs_btree_path
*path
,
1432 int maxlevel
, struct inode
*dat
)
1436 for (level
= NILFS_BTREE_LEVEL_NODE_MIN
; level
<= maxlevel
; level
++) {
1437 nilfs_bmap_commit_end_ptr(&btree
->bt_bmap
,
1438 &path
[level
].bp_oldreq
, dat
);
1439 path
[level
].bp_op(btree
, path
, level
, NULL
, NULL
);
1442 if (!nilfs_bmap_dirty(&btree
->bt_bmap
))
1443 nilfs_bmap_set_dirty(&btree
->bt_bmap
);
1446 static int nilfs_btree_delete(struct nilfs_bmap
*bmap
, __u64 key
)
1449 struct nilfs_btree
*btree
;
1450 struct nilfs_btree_path
*path
;
1451 struct nilfs_bmap_stats stats
;
1455 btree
= (struct nilfs_btree
*)bmap
;
1456 path
= nilfs_btree_alloc_path();
1459 nilfs_btree_init_path(path
);
1460 ret
= nilfs_btree_do_lookup(btree
, path
, key
, NULL
,
1461 NILFS_BTREE_LEVEL_NODE_MIN
);
1466 dat
= NILFS_BMAP_USE_VBN(&btree
->bt_bmap
) ?
1467 nilfs_bmap_get_dat(&btree
->bt_bmap
) : NULL
;
1469 ret
= nilfs_btree_prepare_delete(btree
, path
, &level
, &stats
, dat
);
1472 nilfs_btree_commit_delete(btree
, path
, level
, dat
);
1473 nilfs_bmap_sub_blocks(bmap
, stats
.bs_nblocks
);
1476 nilfs_btree_release_path(path
);
1477 nilfs_btree_free_path(path
);
1481 static int nilfs_btree_last_key(const struct nilfs_bmap
*bmap
, __u64
*keyp
)
1483 struct nilfs_btree
*btree
;
1484 struct nilfs_btree_path
*path
;
1487 btree
= (struct nilfs_btree
*)bmap
;
1488 path
= nilfs_btree_alloc_path();
1491 nilfs_btree_init_path(path
);
1493 ret
= nilfs_btree_do_lookup_last(btree
, path
, keyp
, NULL
);
1495 nilfs_btree_release_path(path
);
1496 nilfs_btree_free_path(path
);
1501 static int nilfs_btree_check_delete(struct nilfs_bmap
*bmap
, __u64 key
)
1503 struct buffer_head
*bh
;
1504 struct nilfs_btree
*btree
;
1505 struct nilfs_btree_node
*root
, *node
;
1506 __u64 maxkey
, nextmaxkey
;
1510 btree
= (struct nilfs_btree
*)bmap
;
1511 root
= nilfs_btree_get_root(btree
);
1512 switch (nilfs_btree_height(btree
)) {
1518 nchildren
= nilfs_btree_node_get_nchildren(root
);
1521 ptr
= nilfs_btree_node_get_ptr(btree
, root
, nchildren
- 1);
1522 ret
= nilfs_btree_get_block(btree
, ptr
, &bh
);
1525 node
= (struct nilfs_btree_node
*)bh
->b_data
;
1531 nchildren
= nilfs_btree_node_get_nchildren(node
);
1532 maxkey
= nilfs_btree_node_get_key(node
, nchildren
- 1);
1533 nextmaxkey
= (nchildren
> 1) ?
1534 nilfs_btree_node_get_key(node
, nchildren
- 2) : 0;
1538 return (maxkey
== key
) && (nextmaxkey
< NILFS_BMAP_LARGE_LOW
);
1541 static int nilfs_btree_gather_data(struct nilfs_bmap
*bmap
,
1542 __u64
*keys
, __u64
*ptrs
, int nitems
)
1544 struct buffer_head
*bh
;
1545 struct nilfs_btree
*btree
;
1546 struct nilfs_btree_node
*node
, *root
;
1550 int nchildren
, i
, ret
;
1552 btree
= (struct nilfs_btree
*)bmap
;
1553 root
= nilfs_btree_get_root(btree
);
1554 switch (nilfs_btree_height(btree
)) {
1560 nchildren
= nilfs_btree_node_get_nchildren(root
);
1561 WARN_ON(nchildren
> 1);
1562 ptr
= nilfs_btree_node_get_ptr(btree
, root
, nchildren
- 1);
1563 ret
= nilfs_btree_get_block(btree
, ptr
, &bh
);
1566 node
= (struct nilfs_btree_node
*)bh
->b_data
;
1573 nchildren
= nilfs_btree_node_get_nchildren(node
);
1574 if (nchildren
< nitems
)
1576 dkeys
= nilfs_btree_node_dkeys(node
);
1577 dptrs
= nilfs_btree_node_dptrs(node
, btree
);
1578 for (i
= 0; i
< nitems
; i
++) {
1579 keys
[i
] = nilfs_bmap_dkey_to_key(dkeys
[i
]);
1580 ptrs
[i
] = nilfs_bmap_dptr_to_ptr(dptrs
[i
]);
1590 nilfs_btree_prepare_convert_and_insert(struct nilfs_bmap
*bmap
, __u64 key
,
1591 union nilfs_bmap_ptr_req
*dreq
,
1592 union nilfs_bmap_ptr_req
*nreq
,
1593 struct buffer_head
**bhp
,
1594 struct nilfs_bmap_stats
*stats
)
1596 struct buffer_head
*bh
;
1597 struct nilfs_btree
*btree
= (struct nilfs_btree
*)bmap
;
1598 struct inode
*dat
= NULL
;
1601 stats
->bs_nblocks
= 0;
1604 /* cannot find near ptr */
1605 if (NILFS_BMAP_USE_VBN(bmap
)) {
1606 dreq
->bpr_ptr
= nilfs_btree_find_target_v(btree
, NULL
, key
);
1607 dat
= nilfs_bmap_get_dat(bmap
);
1610 ret
= nilfs_bmap_prepare_alloc_ptr(bmap
, dreq
, dat
);
1615 stats
->bs_nblocks
++;
1617 nreq
->bpr_ptr
= dreq
->bpr_ptr
+ 1;
1618 ret
= nilfs_bmap_prepare_alloc_ptr(bmap
, nreq
, dat
);
1622 ret
= nilfs_btree_get_new_block(btree
, nreq
->bpr_ptr
, &bh
);
1627 stats
->bs_nblocks
++;
1635 nilfs_bmap_abort_alloc_ptr(bmap
, nreq
, dat
);
1637 nilfs_bmap_abort_alloc_ptr(bmap
, dreq
, dat
);
1638 stats
->bs_nblocks
= 0;
1644 nilfs_btree_commit_convert_and_insert(struct nilfs_bmap
*bmap
,
1645 __u64 key
, __u64 ptr
,
1646 const __u64
*keys
, const __u64
*ptrs
,
1648 union nilfs_bmap_ptr_req
*dreq
,
1649 union nilfs_bmap_ptr_req
*nreq
,
1650 struct buffer_head
*bh
)
1652 struct nilfs_btree
*btree
= (struct nilfs_btree
*)bmap
;
1653 struct nilfs_btree_node
*node
;
1657 /* free resources */
1658 if (bmap
->b_ops
->bop_clear
!= NULL
)
1659 bmap
->b_ops
->bop_clear(bmap
);
1661 /* ptr must be a pointer to a buffer head. */
1662 set_buffer_nilfs_volatile((struct buffer_head
*)((unsigned long)ptr
));
1664 /* convert and insert */
1665 dat
= NILFS_BMAP_USE_VBN(bmap
) ? nilfs_bmap_get_dat(bmap
) : NULL
;
1666 nilfs_btree_init(bmap
);
1668 nilfs_bmap_commit_alloc_ptr(bmap
, dreq
, dat
);
1669 nilfs_bmap_commit_alloc_ptr(bmap
, nreq
, dat
);
1671 /* create child node at level 1 */
1672 node
= (struct nilfs_btree_node
*)bh
->b_data
;
1673 nilfs_btree_node_init(btree
, node
, 0, 1, n
, keys
, ptrs
);
1674 nilfs_btree_node_insert(btree
, node
,
1675 key
, dreq
->bpr_ptr
, n
);
1676 if (!buffer_dirty(bh
))
1677 nilfs_btnode_mark_dirty(bh
);
1678 if (!nilfs_bmap_dirty(bmap
))
1679 nilfs_bmap_set_dirty(bmap
);
1683 /* create root node at level 2 */
1684 node
= nilfs_btree_get_root(btree
);
1685 tmpptr
= nreq
->bpr_ptr
;
1686 nilfs_btree_node_init(btree
, node
, NILFS_BTREE_NODE_ROOT
,
1687 2, 1, &keys
[0], &tmpptr
);
1689 nilfs_bmap_commit_alloc_ptr(bmap
, dreq
, dat
);
1691 /* create root node at level 1 */
1692 node
= nilfs_btree_get_root(btree
);
1693 nilfs_btree_node_init(btree
, node
, NILFS_BTREE_NODE_ROOT
,
1695 nilfs_btree_node_insert(btree
, node
,
1696 key
, dreq
->bpr_ptr
, n
);
1697 if (!nilfs_bmap_dirty(bmap
))
1698 nilfs_bmap_set_dirty(bmap
);
1701 if (NILFS_BMAP_USE_VBN(bmap
))
1702 nilfs_btree_set_target_v(btree
, key
, dreq
->bpr_ptr
);
1706 * nilfs_btree_convert_and_insert -
1714 int nilfs_btree_convert_and_insert(struct nilfs_bmap
*bmap
,
1715 __u64 key
, __u64 ptr
,
1716 const __u64
*keys
, const __u64
*ptrs
, int n
)
1718 struct buffer_head
*bh
;
1719 union nilfs_bmap_ptr_req dreq
, nreq
, *di
, *ni
;
1720 struct nilfs_bmap_stats stats
;
1723 if (n
+ 1 <= NILFS_BTREE_ROOT_NCHILDREN_MAX
) {
1726 } else if ((n
+ 1) <= NILFS_BTREE_NODE_NCHILDREN_MAX(
1727 1 << bmap
->b_inode
->i_blkbits
)) {
1736 ret
= nilfs_btree_prepare_convert_and_insert(bmap
, key
, di
, ni
, &bh
,
1740 nilfs_btree_commit_convert_and_insert(bmap
, key
, ptr
, keys
, ptrs
, n
,
1742 nilfs_bmap_add_blocks(bmap
, stats
.bs_nblocks
);
1746 static int nilfs_btree_propagate_p(struct nilfs_btree
*btree
,
1747 struct nilfs_btree_path
*path
,
1749 struct buffer_head
*bh
)
1751 while ((++level
< nilfs_btree_height(btree
) - 1) &&
1752 !buffer_dirty(path
[level
].bp_bh
))
1753 nilfs_btnode_mark_dirty(path
[level
].bp_bh
);
1758 static int nilfs_btree_prepare_update_v(struct nilfs_btree
*btree
,
1759 struct nilfs_btree_path
*path
,
1760 int level
, struct inode
*dat
)
1762 struct nilfs_btree_node
*parent
;
1765 parent
= nilfs_btree_get_node(btree
, path
, level
+ 1);
1766 path
[level
].bp_oldreq
.bpr_ptr
=
1767 nilfs_btree_node_get_ptr(btree
, parent
,
1768 path
[level
+ 1].bp_index
);
1769 path
[level
].bp_newreq
.bpr_ptr
= path
[level
].bp_oldreq
.bpr_ptr
+ 1;
1770 ret
= nilfs_dat_prepare_update(dat
, &path
[level
].bp_oldreq
.bpr_req
,
1771 &path
[level
].bp_newreq
.bpr_req
);
1775 if (buffer_nilfs_node(path
[level
].bp_bh
)) {
1776 path
[level
].bp_ctxt
.oldkey
= path
[level
].bp_oldreq
.bpr_ptr
;
1777 path
[level
].bp_ctxt
.newkey
= path
[level
].bp_newreq
.bpr_ptr
;
1778 path
[level
].bp_ctxt
.bh
= path
[level
].bp_bh
;
1779 ret
= nilfs_btnode_prepare_change_key(
1780 &NILFS_BMAP_I(&btree
->bt_bmap
)->i_btnode_cache
,
1781 &path
[level
].bp_ctxt
);
1783 nilfs_dat_abort_update(dat
,
1784 &path
[level
].bp_oldreq
.bpr_req
,
1785 &path
[level
].bp_newreq
.bpr_req
);
1793 static void nilfs_btree_commit_update_v(struct nilfs_btree
*btree
,
1794 struct nilfs_btree_path
*path
,
1795 int level
, struct inode
*dat
)
1797 struct nilfs_btree_node
*parent
;
1799 nilfs_dat_commit_update(dat
, &path
[level
].bp_oldreq
.bpr_req
,
1800 &path
[level
].bp_newreq
.bpr_req
,
1801 btree
->bt_bmap
.b_ptr_type
== NILFS_BMAP_PTR_VS
);
1803 if (buffer_nilfs_node(path
[level
].bp_bh
)) {
1804 nilfs_btnode_commit_change_key(
1805 &NILFS_BMAP_I(&btree
->bt_bmap
)->i_btnode_cache
,
1806 &path
[level
].bp_ctxt
);
1807 path
[level
].bp_bh
= path
[level
].bp_ctxt
.bh
;
1809 set_buffer_nilfs_volatile(path
[level
].bp_bh
);
1811 parent
= nilfs_btree_get_node(btree
, path
, level
+ 1);
1812 nilfs_btree_node_set_ptr(btree
, parent
, path
[level
+ 1].bp_index
,
1813 path
[level
].bp_newreq
.bpr_ptr
);
1816 static void nilfs_btree_abort_update_v(struct nilfs_btree
*btree
,
1817 struct nilfs_btree_path
*path
,
1818 int level
, struct inode
*dat
)
1820 nilfs_dat_abort_update(dat
, &path
[level
].bp_oldreq
.bpr_req
,
1821 &path
[level
].bp_newreq
.bpr_req
);
1822 if (buffer_nilfs_node(path
[level
].bp_bh
))
1823 nilfs_btnode_abort_change_key(
1824 &NILFS_BMAP_I(&btree
->bt_bmap
)->i_btnode_cache
,
1825 &path
[level
].bp_ctxt
);
1828 static int nilfs_btree_prepare_propagate_v(struct nilfs_btree
*btree
,
1829 struct nilfs_btree_path
*path
,
1830 int minlevel
, int *maxlevelp
,
1836 if (!buffer_nilfs_volatile(path
[level
].bp_bh
)) {
1837 ret
= nilfs_btree_prepare_update_v(btree
, path
, level
, dat
);
1841 while ((++level
< nilfs_btree_height(btree
) - 1) &&
1842 !buffer_dirty(path
[level
].bp_bh
)) {
1844 WARN_ON(buffer_nilfs_volatile(path
[level
].bp_bh
));
1845 ret
= nilfs_btree_prepare_update_v(btree
, path
, level
, dat
);
1851 *maxlevelp
= level
- 1;
1856 while (--level
> minlevel
)
1857 nilfs_btree_abort_update_v(btree
, path
, level
, dat
);
1858 if (!buffer_nilfs_volatile(path
[level
].bp_bh
))
1859 nilfs_btree_abort_update_v(btree
, path
, level
, dat
);
1863 static void nilfs_btree_commit_propagate_v(struct nilfs_btree
*btree
,
1864 struct nilfs_btree_path
*path
,
1865 int minlevel
, int maxlevel
,
1866 struct buffer_head
*bh
,
1871 if (!buffer_nilfs_volatile(path
[minlevel
].bp_bh
))
1872 nilfs_btree_commit_update_v(btree
, path
, minlevel
, dat
);
1874 for (level
= minlevel
+ 1; level
<= maxlevel
; level
++)
1875 nilfs_btree_commit_update_v(btree
, path
, level
, dat
);
1878 static int nilfs_btree_propagate_v(struct nilfs_btree
*btree
,
1879 struct nilfs_btree_path
*path
,
1880 int level
, struct buffer_head
*bh
)
1882 int maxlevel
= 0, ret
;
1883 struct nilfs_btree_node
*parent
;
1884 struct inode
*dat
= nilfs_bmap_get_dat(&btree
->bt_bmap
);
1888 path
[level
].bp_bh
= bh
;
1889 ret
= nilfs_btree_prepare_propagate_v(btree
, path
, level
, &maxlevel
,
1894 if (buffer_nilfs_volatile(path
[level
].bp_bh
)) {
1895 parent
= nilfs_btree_get_node(btree
, path
, level
+ 1);
1896 ptr
= nilfs_btree_node_get_ptr(btree
, parent
,
1897 path
[level
+ 1].bp_index
);
1898 ret
= nilfs_dat_mark_dirty(dat
, ptr
);
1903 nilfs_btree_commit_propagate_v(btree
, path
, level
, maxlevel
, bh
, dat
);
1906 brelse(path
[level
].bp_bh
);
1907 path
[level
].bp_bh
= NULL
;
1911 static int nilfs_btree_propagate(const struct nilfs_bmap
*bmap
,
1912 struct buffer_head
*bh
)
1914 struct nilfs_btree
*btree
;
1915 struct nilfs_btree_path
*path
;
1916 struct nilfs_btree_node
*node
;
1920 WARN_ON(!buffer_dirty(bh
));
1922 btree
= (struct nilfs_btree
*)bmap
;
1923 path
= nilfs_btree_alloc_path();
1926 nilfs_btree_init_path(path
);
1928 if (buffer_nilfs_node(bh
)) {
1929 node
= (struct nilfs_btree_node
*)bh
->b_data
;
1930 key
= nilfs_btree_node_get_key(node
, 0);
1931 level
= nilfs_btree_node_get_level(node
);
1933 key
= nilfs_bmap_data_get_key(bmap
, bh
);
1934 level
= NILFS_BTREE_LEVEL_DATA
;
1937 ret
= nilfs_btree_do_lookup(btree
, path
, key
, NULL
, level
+ 1);
1939 if (unlikely(ret
== -ENOENT
))
1940 printk(KERN_CRIT
"%s: key = %llu, level == %d\n",
1941 __func__
, (unsigned long long)key
, level
);
1945 ret
= NILFS_BMAP_USE_VBN(bmap
) ?
1946 nilfs_btree_propagate_v(btree
, path
, level
, bh
) :
1947 nilfs_btree_propagate_p(btree
, path
, level
, bh
);
1950 nilfs_btree_release_path(path
);
1951 nilfs_btree_free_path(path
);
1956 static int nilfs_btree_propagate_gc(const struct nilfs_bmap
*bmap
,
1957 struct buffer_head
*bh
)
1959 return nilfs_dat_mark_dirty(nilfs_bmap_get_dat(bmap
), bh
->b_blocknr
);
1962 static void nilfs_btree_add_dirty_buffer(struct nilfs_btree
*btree
,
1963 struct list_head
*lists
,
1964 struct buffer_head
*bh
)
1966 struct list_head
*head
;
1967 struct buffer_head
*cbh
;
1968 struct nilfs_btree_node
*node
, *cnode
;
1973 node
= (struct nilfs_btree_node
*)bh
->b_data
;
1974 key
= nilfs_btree_node_get_key(node
, 0);
1975 level
= nilfs_btree_node_get_level(node
);
1976 list_for_each(head
, &lists
[level
]) {
1977 cbh
= list_entry(head
, struct buffer_head
, b_assoc_buffers
);
1978 cnode
= (struct nilfs_btree_node
*)cbh
->b_data
;
1979 ckey
= nilfs_btree_node_get_key(cnode
, 0);
1983 list_add_tail(&bh
->b_assoc_buffers
, head
);
1986 static void nilfs_btree_lookup_dirty_buffers(struct nilfs_bmap
*bmap
,
1987 struct list_head
*listp
)
1989 struct nilfs_btree
*btree
= (struct nilfs_btree
*)bmap
;
1990 struct address_space
*btcache
= &NILFS_BMAP_I(bmap
)->i_btnode_cache
;
1991 struct list_head lists
[NILFS_BTREE_LEVEL_MAX
];
1992 struct pagevec pvec
;
1993 struct buffer_head
*bh
, *head
;
1997 for (level
= NILFS_BTREE_LEVEL_NODE_MIN
;
1998 level
< NILFS_BTREE_LEVEL_MAX
;
2000 INIT_LIST_HEAD(&lists
[level
]);
2002 pagevec_init(&pvec
, 0);
2004 while (pagevec_lookup_tag(&pvec
, btcache
, &index
, PAGECACHE_TAG_DIRTY
,
2006 for (i
= 0; i
< pagevec_count(&pvec
); i
++) {
2007 bh
= head
= page_buffers(pvec
.pages
[i
]);
2009 if (buffer_dirty(bh
))
2010 nilfs_btree_add_dirty_buffer(btree
,
2012 } while ((bh
= bh
->b_this_page
) != head
);
2014 pagevec_release(&pvec
);
2018 for (level
= NILFS_BTREE_LEVEL_NODE_MIN
;
2019 level
< NILFS_BTREE_LEVEL_MAX
;
2021 list_splice_tail(&lists
[level
], listp
);
2024 static int nilfs_btree_assign_p(struct nilfs_btree
*btree
,
2025 struct nilfs_btree_path
*path
,
2027 struct buffer_head
**bh
,
2029 union nilfs_binfo
*binfo
)
2031 struct nilfs_btree_node
*parent
;
2036 parent
= nilfs_btree_get_node(btree
, path
, level
+ 1);
2037 ptr
= nilfs_btree_node_get_ptr(btree
, parent
,
2038 path
[level
+ 1].bp_index
);
2039 if (buffer_nilfs_node(*bh
)) {
2040 path
[level
].bp_ctxt
.oldkey
= ptr
;
2041 path
[level
].bp_ctxt
.newkey
= blocknr
;
2042 path
[level
].bp_ctxt
.bh
= *bh
;
2043 ret
= nilfs_btnode_prepare_change_key(
2044 &NILFS_BMAP_I(&btree
->bt_bmap
)->i_btnode_cache
,
2045 &path
[level
].bp_ctxt
);
2048 nilfs_btnode_commit_change_key(
2049 &NILFS_BMAP_I(&btree
->bt_bmap
)->i_btnode_cache
,
2050 &path
[level
].bp_ctxt
);
2051 *bh
= path
[level
].bp_ctxt
.bh
;
2054 nilfs_btree_node_set_ptr(btree
, parent
,
2055 path
[level
+ 1].bp_index
, blocknr
);
2057 key
= nilfs_btree_node_get_key(parent
, path
[level
+ 1].bp_index
);
2058 /* on-disk format */
2059 binfo
->bi_dat
.bi_blkoff
= nilfs_bmap_key_to_dkey(key
);
2060 binfo
->bi_dat
.bi_level
= level
;
2065 static int nilfs_btree_assign_v(struct nilfs_btree
*btree
,
2066 struct nilfs_btree_path
*path
,
2068 struct buffer_head
**bh
,
2070 union nilfs_binfo
*binfo
)
2072 struct nilfs_btree_node
*parent
;
2073 struct inode
*dat
= nilfs_bmap_get_dat(&btree
->bt_bmap
);
2076 union nilfs_bmap_ptr_req req
;
2079 parent
= nilfs_btree_get_node(btree
, path
, level
+ 1);
2080 ptr
= nilfs_btree_node_get_ptr(btree
, parent
,
2081 path
[level
+ 1].bp_index
);
2083 ret
= nilfs_dat_prepare_start(dat
, &req
.bpr_req
);
2086 nilfs_dat_commit_start(dat
, &req
.bpr_req
, blocknr
);
2088 key
= nilfs_btree_node_get_key(parent
, path
[level
+ 1].bp_index
);
2089 /* on-disk format */
2090 binfo
->bi_v
.bi_vblocknr
= nilfs_bmap_ptr_to_dptr(ptr
);
2091 binfo
->bi_v
.bi_blkoff
= nilfs_bmap_key_to_dkey(key
);
2096 static int nilfs_btree_assign(struct nilfs_bmap
*bmap
,
2097 struct buffer_head
**bh
,
2099 union nilfs_binfo
*binfo
)
2101 struct nilfs_btree
*btree
;
2102 struct nilfs_btree_path
*path
;
2103 struct nilfs_btree_node
*node
;
2107 btree
= (struct nilfs_btree
*)bmap
;
2108 path
= nilfs_btree_alloc_path();
2111 nilfs_btree_init_path(path
);
2113 if (buffer_nilfs_node(*bh
)) {
2114 node
= (struct nilfs_btree_node
*)(*bh
)->b_data
;
2115 key
= nilfs_btree_node_get_key(node
, 0);
2116 level
= nilfs_btree_node_get_level(node
);
2118 key
= nilfs_bmap_data_get_key(bmap
, *bh
);
2119 level
= NILFS_BTREE_LEVEL_DATA
;
2122 ret
= nilfs_btree_do_lookup(btree
, path
, key
, NULL
, level
+ 1);
2124 WARN_ON(ret
== -ENOENT
);
2128 ret
= NILFS_BMAP_USE_VBN(bmap
) ?
2129 nilfs_btree_assign_v(btree
, path
, level
, bh
, blocknr
, binfo
) :
2130 nilfs_btree_assign_p(btree
, path
, level
, bh
, blocknr
, binfo
);
2133 nilfs_btree_release_path(path
);
2134 nilfs_btree_free_path(path
);
2139 static int nilfs_btree_assign_gc(struct nilfs_bmap
*bmap
,
2140 struct buffer_head
**bh
,
2142 union nilfs_binfo
*binfo
)
2144 struct nilfs_btree_node
*node
;
2148 ret
= nilfs_dat_move(nilfs_bmap_get_dat(bmap
), (*bh
)->b_blocknr
,
2153 if (buffer_nilfs_node(*bh
)) {
2154 node
= (struct nilfs_btree_node
*)(*bh
)->b_data
;
2155 key
= nilfs_btree_node_get_key(node
, 0);
2157 key
= nilfs_bmap_data_get_key(bmap
, *bh
);
2159 /* on-disk format */
2160 binfo
->bi_v
.bi_vblocknr
= cpu_to_le64((*bh
)->b_blocknr
);
2161 binfo
->bi_v
.bi_blkoff
= nilfs_bmap_key_to_dkey(key
);
2166 static int nilfs_btree_mark(struct nilfs_bmap
*bmap
, __u64 key
, int level
)
2168 struct buffer_head
*bh
;
2169 struct nilfs_btree
*btree
;
2170 struct nilfs_btree_path
*path
;
2174 btree
= (struct nilfs_btree
*)bmap
;
2175 path
= nilfs_btree_alloc_path();
2178 nilfs_btree_init_path(path
);
2180 ret
= nilfs_btree_do_lookup(btree
, path
, key
, &ptr
, level
+ 1);
2182 WARN_ON(ret
== -ENOENT
);
2185 ret
= nilfs_btree_get_block(btree
, ptr
, &bh
);
2187 WARN_ON(ret
== -ENOENT
);
2191 if (!buffer_dirty(bh
))
2192 nilfs_btnode_mark_dirty(bh
);
2194 if (!nilfs_bmap_dirty(&btree
->bt_bmap
))
2195 nilfs_bmap_set_dirty(&btree
->bt_bmap
);
2198 nilfs_btree_release_path(path
);
2199 nilfs_btree_free_path(path
);
2203 static const struct nilfs_bmap_operations nilfs_btree_ops
= {
2204 .bop_lookup
= nilfs_btree_lookup
,
2205 .bop_lookup_contig
= nilfs_btree_lookup_contig
,
2206 .bop_insert
= nilfs_btree_insert
,
2207 .bop_delete
= nilfs_btree_delete
,
2210 .bop_propagate
= nilfs_btree_propagate
,
2212 .bop_lookup_dirty_buffers
= nilfs_btree_lookup_dirty_buffers
,
2214 .bop_assign
= nilfs_btree_assign
,
2215 .bop_mark
= nilfs_btree_mark
,
2217 .bop_last_key
= nilfs_btree_last_key
,
2218 .bop_check_insert
= NULL
,
2219 .bop_check_delete
= nilfs_btree_check_delete
,
2220 .bop_gather_data
= nilfs_btree_gather_data
,
2223 static const struct nilfs_bmap_operations nilfs_btree_ops_gc
= {
2225 .bop_lookup_contig
= NULL
,
2230 .bop_propagate
= nilfs_btree_propagate_gc
,
2232 .bop_lookup_dirty_buffers
= nilfs_btree_lookup_dirty_buffers
,
2234 .bop_assign
= nilfs_btree_assign_gc
,
2237 .bop_last_key
= NULL
,
2238 .bop_check_insert
= NULL
,
2239 .bop_check_delete
= NULL
,
2240 .bop_gather_data
= NULL
,
2243 int nilfs_btree_init(struct nilfs_bmap
*bmap
)
2245 bmap
->b_ops
= &nilfs_btree_ops
;
2249 void nilfs_btree_init_gc(struct nilfs_bmap
*bmap
)
2251 bmap
->b_ops
= &nilfs_btree_ops_gc
;