2 * Copyright (c) 2000-2003,2005 Silicon Graphics, Inc.
5 * This program is free software; you can redistribute it and/or
6 * modify it under the terms of the GNU General Public License as
7 * published by the Free Software Foundation.
9 * This program is distributed in the hope that it would be useful,
10 * but WITHOUT ANY WARRANTY; without even the implied warranty of
11 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
12 * GNU General Public License for more details.
14 * You should have received a copy of the GNU General Public License
15 * along with this program; if not, write the Free Software Foundation,
16 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
20 #include "xfs_types.h"
23 #include "xfs_trans.h"
27 #include "xfs_dmapi.h"
28 #include "xfs_mount.h"
29 #include "xfs_da_btree.h"
30 #include "xfs_bmap_btree.h"
31 #include "xfs_alloc_btree.h"
32 #include "xfs_ialloc_btree.h"
33 #include "xfs_dir_sf.h"
34 #include "xfs_dir2_sf.h"
35 #include "xfs_attr_sf.h"
36 #include "xfs_dinode.h"
37 #include "xfs_inode.h"
38 #include "xfs_inode_item.h"
39 #include "xfs_alloc.h"
40 #include "xfs_btree.h"
42 #include "xfs_dir_leaf.h"
43 #include "xfs_error.h"
48 * Routines to implement leaf blocks of directories as Btrees of hashed names.
51 /*========================================================================
52 * Function prototypes for the kernel.
53 *========================================================================*/
56 * Routines used for growing the Btree.
58 STATIC
void xfs_dir_leaf_add_work(xfs_dabuf_t
*leaf_buffer
, xfs_da_args_t
*args
,
61 STATIC
int xfs_dir_leaf_compact(xfs_trans_t
*trans
, xfs_dabuf_t
*leaf_buffer
,
62 int musthave
, int justcheck
);
63 STATIC
void xfs_dir_leaf_rebalance(xfs_da_state_t
*state
,
64 xfs_da_state_blk_t
*blk1
,
65 xfs_da_state_blk_t
*blk2
);
66 STATIC
int xfs_dir_leaf_figure_balance(xfs_da_state_t
*state
,
67 xfs_da_state_blk_t
*leaf_blk_1
,
68 xfs_da_state_blk_t
*leaf_blk_2
,
69 int *number_entries_in_blk1
,
70 int *number_namebytes_in_blk1
);
72 STATIC
int xfs_dir_leaf_create(struct xfs_da_args
*args
,
73 xfs_dablk_t which_block
,
74 struct xfs_dabuf
**bpp
);
79 STATIC
void xfs_dir_leaf_moveents(xfs_dir_leafblock_t
*src_leaf
,
81 xfs_dir_leafblock_t
*dst_leaf
,
82 int dst_start
, int move_count
,
86 /*========================================================================
87 * External routines when dirsize < XFS_IFORK_DSIZE(dp).
88 *========================================================================*/
92 * Validate a given inode number.
95 xfs_dir_ino_validate(xfs_mount_t
*mp
, xfs_ino_t ino
)
97 xfs_agblock_t agblkno
;
103 agno
= XFS_INO_TO_AGNO(mp
, ino
);
104 agblkno
= XFS_INO_TO_AGBNO(mp
, ino
);
105 ioff
= XFS_INO_TO_OFFSET(mp
, ino
);
106 agino
= XFS_OFFBNO_TO_AGINO(mp
, agblkno
, ioff
);
108 agno
< mp
->m_sb
.sb_agcount
&&
109 agblkno
< mp
->m_sb
.sb_agblocks
&&
111 ioff
< (1 << mp
->m_sb
.sb_inopblog
) &&
112 XFS_AGINO_TO_INO(mp
, agno
, agino
) == ino
;
113 if (unlikely(XFS_TEST_ERROR(!ino_ok
, mp
, XFS_ERRTAG_DIR_INO_VALIDATE
,
114 XFS_RANDOM_DIR_INO_VALIDATE
))) {
115 xfs_fs_cmn_err(CE_WARN
, mp
, "Invalid inode number 0x%Lx",
116 (unsigned long long) ino
);
117 XFS_ERROR_REPORT("xfs_dir_ino_validate", XFS_ERRLEVEL_LOW
, mp
);
118 return XFS_ERROR(EFSCORRUPTED
);
124 * Create the initial contents of a shortform directory.
127 xfs_dir_shortform_create(xfs_da_args_t
*args
, xfs_ino_t parent
)
129 xfs_dir_sf_hdr_t
*hdr
;
134 ASSERT(dp
->i_d
.di_size
== 0);
135 if (dp
->i_d
.di_format
== XFS_DINODE_FMT_EXTENTS
) {
136 dp
->i_df
.if_flags
&= ~XFS_IFEXTENTS
; /* just in case */
137 dp
->i_d
.di_format
= XFS_DINODE_FMT_LOCAL
;
138 xfs_trans_log_inode(args
->trans
, dp
, XFS_ILOG_CORE
);
139 dp
->i_df
.if_flags
|= XFS_IFINLINE
;
141 ASSERT(dp
->i_df
.if_flags
& XFS_IFINLINE
);
142 ASSERT(dp
->i_df
.if_bytes
== 0);
143 xfs_idata_realloc(dp
, sizeof(*hdr
), XFS_DATA_FORK
);
144 hdr
= (xfs_dir_sf_hdr_t
*)dp
->i_df
.if_u1
.if_data
;
145 XFS_DIR_SF_PUT_DIRINO(&parent
, &hdr
->parent
);
148 dp
->i_d
.di_size
= sizeof(*hdr
);
149 xfs_trans_log_inode(args
->trans
, dp
, XFS_ILOG_CORE
| XFS_ILOG_DDATA
);
154 * Add a name to the shortform directory structure.
155 * Overflow from the inode has already been checked for.
158 xfs_dir_shortform_addname(xfs_da_args_t
*args
)
160 xfs_dir_shortform_t
*sf
;
161 xfs_dir_sf_entry_t
*sfe
;
166 ASSERT(dp
->i_df
.if_flags
& XFS_IFINLINE
);
168 * Catch the case where the conversion from shortform to leaf
169 * failed part way through.
171 if (dp
->i_d
.di_size
< sizeof(xfs_dir_sf_hdr_t
)) {
172 ASSERT(XFS_FORCED_SHUTDOWN(dp
->i_mount
));
173 return XFS_ERROR(EIO
);
175 ASSERT(dp
->i_df
.if_bytes
== dp
->i_d
.di_size
);
176 ASSERT(dp
->i_df
.if_u1
.if_data
!= NULL
);
177 sf
= (xfs_dir_shortform_t
*)dp
->i_df
.if_u1
.if_data
;
179 for (i
= sf
->hdr
.count
-1; i
>= 0; i
--) {
180 if (sfe
->namelen
== args
->namelen
&&
181 args
->name
[0] == sfe
->name
[0] &&
182 memcmp(args
->name
, sfe
->name
, args
->namelen
) == 0)
183 return XFS_ERROR(EEXIST
);
184 sfe
= XFS_DIR_SF_NEXTENTRY(sfe
);
187 offset
= (int)((char *)sfe
- (char *)sf
);
188 size
= XFS_DIR_SF_ENTSIZE_BYNAME(args
->namelen
);
189 xfs_idata_realloc(dp
, size
, XFS_DATA_FORK
);
190 sf
= (xfs_dir_shortform_t
*)dp
->i_df
.if_u1
.if_data
;
191 sfe
= (xfs_dir_sf_entry_t
*)((char *)sf
+ offset
);
193 XFS_DIR_SF_PUT_DIRINO(&args
->inumber
, &sfe
->inumber
);
194 sfe
->namelen
= args
->namelen
;
195 memcpy(sfe
->name
, args
->name
, sfe
->namelen
);
198 dp
->i_d
.di_size
+= size
;
199 xfs_trans_log_inode(args
->trans
, dp
, XFS_ILOG_CORE
| XFS_ILOG_DDATA
);
205 * Remove a name from the shortform directory structure.
208 xfs_dir_shortform_removename(xfs_da_args_t
*args
)
210 xfs_dir_shortform_t
*sf
;
211 xfs_dir_sf_entry_t
*sfe
;
212 int base
, size
= 0, i
;
216 ASSERT(dp
->i_df
.if_flags
& XFS_IFINLINE
);
218 * Catch the case where the conversion from shortform to leaf
219 * failed part way through.
221 if (dp
->i_d
.di_size
< sizeof(xfs_dir_sf_hdr_t
)) {
222 ASSERT(XFS_FORCED_SHUTDOWN(dp
->i_mount
));
223 return XFS_ERROR(EIO
);
225 ASSERT(dp
->i_df
.if_bytes
== dp
->i_d
.di_size
);
226 ASSERT(dp
->i_df
.if_u1
.if_data
!= NULL
);
227 base
= sizeof(xfs_dir_sf_hdr_t
);
228 sf
= (xfs_dir_shortform_t
*)dp
->i_df
.if_u1
.if_data
;
230 for (i
= sf
->hdr
.count
-1; i
>= 0; i
--) {
231 size
= XFS_DIR_SF_ENTSIZE_BYENTRY(sfe
);
232 if (sfe
->namelen
== args
->namelen
&&
233 sfe
->name
[0] == args
->name
[0] &&
234 memcmp(sfe
->name
, args
->name
, args
->namelen
) == 0)
237 sfe
= XFS_DIR_SF_NEXTENTRY(sfe
);
240 ASSERT(args
->oknoent
);
241 return XFS_ERROR(ENOENT
);
244 if ((base
+ size
) != dp
->i_d
.di_size
) {
245 memmove(&((char *)sf
)[base
], &((char *)sf
)[base
+size
],
246 dp
->i_d
.di_size
- (base
+size
));
250 xfs_idata_realloc(dp
, -size
, XFS_DATA_FORK
);
251 dp
->i_d
.di_size
-= size
;
252 xfs_trans_log_inode(args
->trans
, dp
, XFS_ILOG_CORE
| XFS_ILOG_DDATA
);
258 * Look up a name in a shortform directory structure.
261 xfs_dir_shortform_lookup(xfs_da_args_t
*args
)
263 xfs_dir_shortform_t
*sf
;
264 xfs_dir_sf_entry_t
*sfe
;
269 ASSERT(dp
->i_df
.if_flags
& XFS_IFINLINE
);
271 * Catch the case where the conversion from shortform to leaf
272 * failed part way through.
274 if (dp
->i_d
.di_size
< sizeof(xfs_dir_sf_hdr_t
)) {
275 ASSERT(XFS_FORCED_SHUTDOWN(dp
->i_mount
));
276 return XFS_ERROR(EIO
);
278 ASSERT(dp
->i_df
.if_bytes
== dp
->i_d
.di_size
);
279 ASSERT(dp
->i_df
.if_u1
.if_data
!= NULL
);
280 sf
= (xfs_dir_shortform_t
*)dp
->i_df
.if_u1
.if_data
;
281 if (args
->namelen
== 2 &&
282 args
->name
[0] == '.' && args
->name
[1] == '.') {
283 XFS_DIR_SF_GET_DIRINO(&sf
->hdr
.parent
, &args
->inumber
);
284 return(XFS_ERROR(EEXIST
));
286 if (args
->namelen
== 1 && args
->name
[0] == '.') {
287 args
->inumber
= dp
->i_ino
;
288 return(XFS_ERROR(EEXIST
));
291 for (i
= sf
->hdr
.count
-1; i
>= 0; i
--) {
292 if (sfe
->namelen
== args
->namelen
&&
293 sfe
->name
[0] == args
->name
[0] &&
294 memcmp(args
->name
, sfe
->name
, args
->namelen
) == 0) {
295 XFS_DIR_SF_GET_DIRINO(&sfe
->inumber
, &args
->inumber
);
296 return(XFS_ERROR(EEXIST
));
298 sfe
= XFS_DIR_SF_NEXTENTRY(sfe
);
300 ASSERT(args
->oknoent
);
301 return(XFS_ERROR(ENOENT
));
305 * Convert from using the shortform to the leaf.
308 xfs_dir_shortform_to_leaf(xfs_da_args_t
*iargs
)
311 xfs_dir_shortform_t
*sf
;
312 xfs_dir_sf_entry_t
*sfe
;
322 * Catch the case where the conversion from shortform to leaf
323 * failed part way through.
325 if (dp
->i_d
.di_size
< sizeof(xfs_dir_sf_hdr_t
)) {
326 ASSERT(XFS_FORCED_SHUTDOWN(dp
->i_mount
));
327 return XFS_ERROR(EIO
);
329 ASSERT(dp
->i_df
.if_bytes
== dp
->i_d
.di_size
);
330 ASSERT(dp
->i_df
.if_u1
.if_data
!= NULL
);
331 size
= dp
->i_df
.if_bytes
;
332 tmpbuffer
= kmem_alloc(size
, KM_SLEEP
);
333 ASSERT(tmpbuffer
!= NULL
);
335 memcpy(tmpbuffer
, dp
->i_df
.if_u1
.if_data
, size
);
337 sf
= (xfs_dir_shortform_t
*)tmpbuffer
;
338 XFS_DIR_SF_GET_DIRINO(&sf
->hdr
.parent
, &inumber
);
340 xfs_idata_realloc(dp
, -size
, XFS_DATA_FORK
);
342 xfs_trans_log_inode(iargs
->trans
, dp
, XFS_ILOG_CORE
);
343 retval
= xfs_da_grow_inode(iargs
, &blkno
);
348 retval
= xfs_dir_leaf_create(iargs
, blkno
, &bp
);
355 args
.hashval
= xfs_dir_hash_dot
;
356 args
.inumber
= dp
->i_ino
;
358 args
.firstblock
= iargs
->firstblock
;
359 args
.flist
= iargs
->flist
;
360 args
.total
= iargs
->total
;
361 args
.whichfork
= XFS_DATA_FORK
;
362 args
.trans
= iargs
->trans
;
364 args
.addname
= args
.oknoent
= 1;
365 retval
= xfs_dir_leaf_addname(&args
);
371 args
.hashval
= xfs_dir_hash_dotdot
;
372 args
.inumber
= inumber
;
373 retval
= xfs_dir_leaf_addname(&args
);
378 for (i
= 0; i
< sf
->hdr
.count
; i
++) {
379 args
.name
= (char *)(sfe
->name
);
380 args
.namelen
= sfe
->namelen
;
381 args
.hashval
= xfs_da_hashname((char *)(sfe
->name
),
383 XFS_DIR_SF_GET_DIRINO(&sfe
->inumber
, &args
.inumber
);
384 retval
= xfs_dir_leaf_addname(&args
);
387 sfe
= XFS_DIR_SF_NEXTENTRY(sfe
);
392 kmem_free(tmpbuffer
, size
);
397 xfs_dir_shortform_compare(const void *a
, const void *b
)
399 xfs_dir_sf_sort_t
*sa
, *sb
;
401 sa
= (xfs_dir_sf_sort_t
*)a
;
402 sb
= (xfs_dir_sf_sort_t
*)b
;
403 if (sa
->hash
< sb
->hash
)
405 else if (sa
->hash
> sb
->hash
)
408 return sa
->entno
- sb
->entno
;
412 * Copy out directory entries for getdents(), for shortform directories.
416 xfs_dir_shortform_getdents(xfs_inode_t
*dp
, uio_t
*uio
, int *eofp
,
417 xfs_dirent_t
*dbp
, xfs_dir_put_t put
)
419 xfs_dir_shortform_t
*sf
;
420 xfs_dir_sf_entry_t
*sfe
;
421 int retval
, i
, sbsize
, nsbuf
, lastresid
=0, want_entno
;
423 xfs_dahash_t cookhash
, hash
;
424 xfs_dir_put_args_t p
;
425 xfs_dir_sf_sort_t
*sbuf
, *sbp
;
428 sf
= (xfs_dir_shortform_t
*)dp
->i_df
.if_u1
.if_data
;
429 cookhash
= XFS_DA_COOKIE_HASH(mp
, uio
->uio_offset
);
430 want_entno
= XFS_DA_COOKIE_ENTRY(mp
, uio
->uio_offset
);
431 nsbuf
= sf
->hdr
.count
+ 2;
432 sbsize
= (nsbuf
+ 1) * sizeof(*sbuf
);
433 sbp
= sbuf
= kmem_alloc(sbsize
, KM_SLEEP
);
435 xfs_dir_trace_g_du("sf: start", dp
, uio
);
438 * Collect all the entries into the buffer.
443 sbp
->hash
= xfs_dir_hash_dot
;
444 sbp
->ino
= dp
->i_ino
;
454 sbp
->hash
= xfs_dir_hash_dotdot
;
455 sbp
->ino
= XFS_GET_DIR_INO8(sf
->hdr
.parent
);
461 * Scan the directory data for the rest of the entries.
463 for (i
= 0, sfe
= &sf
->list
[0]; i
< sf
->hdr
.count
; i
++) {
466 ((char *)sfe
< (char *)sf
) ||
467 ((char *)sfe
>= ((char *)sf
+ dp
->i_df
.if_bytes
)))) {
468 xfs_dir_trace_g_du("sf: corrupted", dp
, uio
);
469 XFS_CORRUPTION_ERROR("xfs_dir_shortform_getdents",
470 XFS_ERRLEVEL_LOW
, mp
, sfe
);
471 kmem_free(sbuf
, sbsize
);
472 return XFS_ERROR(EFSCORRUPTED
);
477 sbp
->hash
= xfs_da_hashname((char *)sfe
->name
, sfe
->namelen
);
478 sbp
->ino
= XFS_GET_DIR_INO8(sfe
->inumber
);
479 sbp
->name
= (char *)sfe
->name
;
480 sbp
->namelen
= sfe
->namelen
;
481 sfe
= XFS_DIR_SF_NEXTENTRY(sfe
);
486 * Sort the entries on hash then entno.
488 xfs_sort(sbuf
, nsbuf
, sizeof(*sbuf
), xfs_dir_shortform_compare
);
490 * Stuff in last entry.
493 sbp
->hash
= XFS_DA_MAXHASH
;
496 * Figure out the sequence numbers in case there's a hash duplicate.
498 for (hash
= sbuf
->hash
, sbp
= sbuf
+ 1;
499 sbp
< &sbuf
[nsbuf
+ 1]; sbp
++) {
500 if (sbp
->hash
== hash
)
501 sbp
->seqno
= sbp
[-1].seqno
+ 1;
507 * Set up put routine.
516 for (sbp
= sbuf
; sbp
< &sbuf
[nsbuf
+ 1]; sbp
++) {
517 if (sbp
->hash
> cookhash
||
518 (sbp
->hash
== cookhash
&& sbp
->seqno
>= want_entno
))
523 * Did we fail to find anything? We stop at the last entry,
524 * the one we put maxhash into.
526 if (sbp
== &sbuf
[nsbuf
]) {
527 kmem_free(sbuf
, sbsize
);
528 xfs_dir_trace_g_du("sf: hash beyond end", dp
, uio
);
529 uio
->uio_offset
= XFS_DA_MAKE_COOKIE(mp
, 0, 0, XFS_DA_MAXHASH
);
535 * Loop putting entries into the user buffer.
537 while (sbp
< &sbuf
[nsbuf
]) {
539 * Save the first resid in a run of equal-hashval entries
540 * so that we can back them out if they don't all fit.
542 if (sbp
->seqno
== 0 || sbp
== sbuf
)
543 lastresid
= uio
->uio_resid
;
544 XFS_PUT_COOKIE(p
.cook
, mp
, 0, sbp
[1].seqno
, sbp
[1].hash
);
547 p
.ino
+= mp
->m_inoadd
;
550 p
.namelen
= sbp
->namelen
;
554 XFS_DA_MAKE_COOKIE(mp
, 0, 0, sbp
->hash
);
555 kmem_free(sbuf
, sbsize
);
556 uio
->uio_resid
= lastresid
;
557 xfs_dir_trace_g_du("sf: E-O-B", dp
, uio
);
562 kmem_free(sbuf
, sbsize
);
563 uio
->uio_offset
= p
.cook
.o
;
565 xfs_dir_trace_g_du("sf: E-O-F", dp
, uio
);
570 * Look up a name in a shortform directory structure, replace the inode number.
573 xfs_dir_shortform_replace(xfs_da_args_t
*args
)
575 xfs_dir_shortform_t
*sf
;
576 xfs_dir_sf_entry_t
*sfe
;
581 ASSERT(dp
->i_df
.if_flags
& XFS_IFINLINE
);
583 * Catch the case where the conversion from shortform to leaf
584 * failed part way through.
586 if (dp
->i_d
.di_size
< sizeof(xfs_dir_sf_hdr_t
)) {
587 ASSERT(XFS_FORCED_SHUTDOWN(dp
->i_mount
));
588 return XFS_ERROR(EIO
);
590 ASSERT(dp
->i_df
.if_bytes
== dp
->i_d
.di_size
);
591 ASSERT(dp
->i_df
.if_u1
.if_data
!= NULL
);
592 sf
= (xfs_dir_shortform_t
*)dp
->i_df
.if_u1
.if_data
;
593 if (args
->namelen
== 2 &&
594 args
->name
[0] == '.' && args
->name
[1] == '.') {
595 /* XXX - replace assert? */
596 XFS_DIR_SF_PUT_DIRINO(&args
->inumber
, &sf
->hdr
.parent
);
597 xfs_trans_log_inode(args
->trans
, dp
, XFS_ILOG_DDATA
);
600 ASSERT(args
->namelen
!= 1 || args
->name
[0] != '.');
602 for (i
= sf
->hdr
.count
-1; i
>= 0; i
--) {
603 if (sfe
->namelen
== args
->namelen
&&
604 sfe
->name
[0] == args
->name
[0] &&
605 memcmp(args
->name
, sfe
->name
, args
->namelen
) == 0) {
606 ASSERT(memcmp((char *)&args
->inumber
,
607 (char *)&sfe
->inumber
, sizeof(xfs_ino_t
)));
608 XFS_DIR_SF_PUT_DIRINO(&args
->inumber
, &sfe
->inumber
);
609 xfs_trans_log_inode(args
->trans
, dp
, XFS_ILOG_DDATA
);
612 sfe
= XFS_DIR_SF_NEXTENTRY(sfe
);
614 ASSERT(args
->oknoent
);
615 return XFS_ERROR(ENOENT
);
619 * Convert a leaf directory to shortform structure
622 xfs_dir_leaf_to_shortform(xfs_da_args_t
*iargs
)
624 xfs_dir_leafblock_t
*leaf
;
625 xfs_dir_leaf_hdr_t
*hdr
;
626 xfs_dir_leaf_entry_t
*entry
;
627 xfs_dir_leaf_name_t
*namest
;
630 xfs_ino_t parent
= 0;
636 tmpbuffer
= kmem_alloc(XFS_LBSIZE(dp
->i_mount
), KM_SLEEP
);
637 ASSERT(tmpbuffer
!= NULL
);
639 retval
= xfs_da_read_buf(iargs
->trans
, iargs
->dp
, 0, -1, &bp
,
644 memcpy(tmpbuffer
, bp
->data
, XFS_LBSIZE(dp
->i_mount
));
645 leaf
= (xfs_dir_leafblock_t
*)tmpbuffer
;
646 ASSERT(be16_to_cpu(leaf
->hdr
.info
.magic
) == XFS_DIR_LEAF_MAGIC
);
647 memset(bp
->data
, 0, XFS_LBSIZE(dp
->i_mount
));
650 * Find and special case the parent inode number
653 entry
= &leaf
->entries
[0];
654 for (i
= INT_GET(hdr
->count
, ARCH_CONVERT
)-1; i
>= 0; entry
++, i
--) {
655 namest
= XFS_DIR_LEAF_NAMESTRUCT(leaf
, INT_GET(entry
->nameidx
, ARCH_CONVERT
));
656 if ((entry
->namelen
== 2) &&
657 (namest
->name
[0] == '.') &&
658 (namest
->name
[1] == '.')) {
659 XFS_DIR_SF_GET_DIRINO(&namest
->inumber
, &parent
);
661 } else if ((entry
->namelen
== 1) && (namest
->name
[0] == '.')) {
665 retval
= xfs_da_shrink_inode(iargs
, 0, bp
);
668 retval
= xfs_dir_shortform_create(iargs
, parent
);
673 * Copy the rest of the filenames
675 entry
= &leaf
->entries
[0];
677 args
.firstblock
= iargs
->firstblock
;
678 args
.flist
= iargs
->flist
;
679 args
.total
= iargs
->total
;
680 args
.whichfork
= XFS_DATA_FORK
;
681 args
.trans
= iargs
->trans
;
683 args
.addname
= args
.oknoent
= 1;
684 for (i
= 0; i
< INT_GET(hdr
->count
, ARCH_CONVERT
); entry
++, i
++) {
687 namest
= XFS_DIR_LEAF_NAMESTRUCT(leaf
, INT_GET(entry
->nameidx
, ARCH_CONVERT
));
688 args
.name
= (char *)(namest
->name
);
689 args
.namelen
= entry
->namelen
;
690 args
.hashval
= INT_GET(entry
->hashval
, ARCH_CONVERT
);
691 XFS_DIR_SF_GET_DIRINO(&namest
->inumber
, &args
.inumber
);
692 xfs_dir_shortform_addname(&args
);
696 kmem_free(tmpbuffer
, XFS_LBSIZE(dp
->i_mount
));
701 * Convert from using a single leaf to a root node and a leaf.
704 xfs_dir_leaf_to_node(xfs_da_args_t
*args
)
706 xfs_dir_leafblock_t
*leaf
;
707 xfs_da_intnode_t
*node
;
709 xfs_dabuf_t
*bp1
, *bp2
;
714 retval
= xfs_da_grow_inode(args
, &blkno
);
718 retval
= xfs_da_read_buf(args
->trans
, args
->dp
, 0, -1, &bp1
,
723 retval
= xfs_da_get_buf(args
->trans
, args
->dp
, 1, -1, &bp2
,
726 xfs_da_buf_done(bp1
);
730 memcpy(bp2
->data
, bp1
->data
, XFS_LBSIZE(dp
->i_mount
));
731 xfs_da_buf_done(bp1
);
732 xfs_da_log_buf(args
->trans
, bp2
, 0, XFS_LBSIZE(dp
->i_mount
) - 1);
735 * Set up the new root node.
737 retval
= xfs_da_node_create(args
, 0, 1, &bp1
, XFS_DATA_FORK
);
739 xfs_da_buf_done(bp2
);
744 ASSERT(be16_to_cpu(leaf
->hdr
.info
.magic
) == XFS_DIR_LEAF_MAGIC
);
745 node
->btree
[0].hashval
= cpu_to_be32(
746 INT_GET(leaf
->entries
[
747 INT_GET(leaf
->hdr
.count
, ARCH_CONVERT
)-1].hashval
, ARCH_CONVERT
));
748 xfs_da_buf_done(bp2
);
749 node
->btree
[0].before
= cpu_to_be32(blkno
);
750 node
->hdr
.count
= cpu_to_be16(1);
751 xfs_da_log_buf(args
->trans
, bp1
,
752 XFS_DA_LOGRANGE(node
, &node
->btree
[0], sizeof(node
->btree
[0])));
753 xfs_da_buf_done(bp1
);
759 /*========================================================================
760 * Routines used for growing the Btree.
761 *========================================================================*/
764 * Create the initial contents of a leaf directory
765 * or a leaf in a node directory.
768 xfs_dir_leaf_create(xfs_da_args_t
*args
, xfs_dablk_t blkno
, xfs_dabuf_t
**bpp
)
770 xfs_dir_leafblock_t
*leaf
;
771 xfs_dir_leaf_hdr_t
*hdr
;
778 retval
= xfs_da_get_buf(args
->trans
, dp
, blkno
, -1, &bp
, XFS_DATA_FORK
);
783 memset((char *)leaf
, 0, XFS_LBSIZE(dp
->i_mount
));
785 hdr
->info
.magic
= cpu_to_be16(XFS_DIR_LEAF_MAGIC
);
786 INT_SET(hdr
->firstused
, ARCH_CONVERT
, XFS_LBSIZE(dp
->i_mount
));
788 INT_SET(hdr
->firstused
, ARCH_CONVERT
, XFS_LBSIZE(dp
->i_mount
) - 1);
789 INT_SET(hdr
->freemap
[0].base
, ARCH_CONVERT
, sizeof(xfs_dir_leaf_hdr_t
));
790 INT_SET(hdr
->freemap
[0].size
, ARCH_CONVERT
, INT_GET(hdr
->firstused
, ARCH_CONVERT
) - INT_GET(hdr
->freemap
[0].base
, ARCH_CONVERT
));
792 xfs_da_log_buf(args
->trans
, bp
, 0, XFS_LBSIZE(dp
->i_mount
) - 1);
799 * Split the leaf node, rebalance, then add the new entry.
802 xfs_dir_leaf_split(xfs_da_state_t
*state
, xfs_da_state_blk_t
*oldblk
,
803 xfs_da_state_blk_t
*newblk
)
810 * Allocate space for a new leaf node.
813 ASSERT(args
!= NULL
);
814 ASSERT(oldblk
->magic
== XFS_DIR_LEAF_MAGIC
);
815 error
= xfs_da_grow_inode(args
, &blkno
);
818 error
= xfs_dir_leaf_create(args
, blkno
, &newblk
->bp
);
821 newblk
->blkno
= blkno
;
822 newblk
->magic
= XFS_DIR_LEAF_MAGIC
;
825 * Rebalance the entries across the two leaves.
827 xfs_dir_leaf_rebalance(state
, oldblk
, newblk
);
828 error
= xfs_da_blk_link(state
, oldblk
, newblk
);
833 * Insert the new entry in the correct block.
836 error
= xfs_dir_leaf_add(oldblk
->bp
, args
, oldblk
->index
);
838 error
= xfs_dir_leaf_add(newblk
->bp
, args
, newblk
->index
);
842 * Update last hashval in each block since we added the name.
844 oldblk
->hashval
= xfs_dir_leaf_lasthash(oldblk
->bp
, NULL
);
845 newblk
->hashval
= xfs_dir_leaf_lasthash(newblk
->bp
, NULL
);
850 * Add a name to the leaf directory structure.
852 * Must take into account fragmented leaves and leaves where spacemap has
853 * lost some freespace information (ie: holes).
856 xfs_dir_leaf_add(xfs_dabuf_t
*bp
, xfs_da_args_t
*args
, int index
)
858 xfs_dir_leafblock_t
*leaf
;
859 xfs_dir_leaf_hdr_t
*hdr
;
860 xfs_dir_leaf_map_t
*map
;
861 int tablesize
, entsize
, sum
, i
, tmp
, error
;
864 ASSERT(be16_to_cpu(leaf
->hdr
.info
.magic
) == XFS_DIR_LEAF_MAGIC
);
865 ASSERT((index
>= 0) && (index
<= INT_GET(leaf
->hdr
.count
, ARCH_CONVERT
)));
867 entsize
= XFS_DIR_LEAF_ENTSIZE_BYNAME(args
->namelen
);
870 * Search through freemap for first-fit on new name length.
871 * (may need to figure in size of entry struct too)
873 tablesize
= (INT_GET(hdr
->count
, ARCH_CONVERT
) + 1) * (uint
)sizeof(xfs_dir_leaf_entry_t
)
874 + (uint
)sizeof(xfs_dir_leaf_hdr_t
);
875 map
= &hdr
->freemap
[XFS_DIR_LEAF_MAPSIZE
-1];
876 for (sum
= 0, i
= XFS_DIR_LEAF_MAPSIZE
-1; i
>= 0; map
--, i
--) {
877 if (tablesize
> INT_GET(hdr
->firstused
, ARCH_CONVERT
)) {
878 sum
+= INT_GET(map
->size
, ARCH_CONVERT
);
882 continue; /* no space in this map */
884 if (INT_GET(map
->base
, ARCH_CONVERT
) < INT_GET(hdr
->firstused
, ARCH_CONVERT
))
885 tmp
+= (uint
)sizeof(xfs_dir_leaf_entry_t
);
886 if (INT_GET(map
->size
, ARCH_CONVERT
) >= tmp
) {
887 if (!args
->justcheck
)
888 xfs_dir_leaf_add_work(bp
, args
, index
, i
);
891 sum
+= INT_GET(map
->size
, ARCH_CONVERT
);
895 * If there are no holes in the address space of the block,
896 * and we don't have enough freespace, then compaction will do us
897 * no good and we should just give up.
899 if (!hdr
->holes
&& (sum
< entsize
))
900 return XFS_ERROR(ENOSPC
);
903 * Compact the entries to coalesce free space.
904 * Pass the justcheck flag so the checking pass can return
905 * an error, without changing anything, if it won't fit.
907 error
= xfs_dir_leaf_compact(args
->trans
, bp
,
910 (uint
)sizeof(xfs_dir_leaf_entry_t
) : 0,
915 * After compaction, the block is guaranteed to have only one
916 * free region, in freemap[0]. If it is not big enough, give up.
918 if (INT_GET(hdr
->freemap
[0].size
, ARCH_CONVERT
) <
919 (entsize
+ (uint
)sizeof(xfs_dir_leaf_entry_t
)))
920 return XFS_ERROR(ENOSPC
);
922 if (!args
->justcheck
)
923 xfs_dir_leaf_add_work(bp
, args
, index
, 0);
928 * Add a name to a leaf directory structure.
931 xfs_dir_leaf_add_work(xfs_dabuf_t
*bp
, xfs_da_args_t
*args
, int index
,
934 xfs_dir_leafblock_t
*leaf
;
935 xfs_dir_leaf_hdr_t
*hdr
;
936 xfs_dir_leaf_entry_t
*entry
;
937 xfs_dir_leaf_name_t
*namest
;
938 xfs_dir_leaf_map_t
*map
;
944 ASSERT(be16_to_cpu(leaf
->hdr
.info
.magic
) == XFS_DIR_LEAF_MAGIC
);
946 ASSERT((mapindex
>= 0) && (mapindex
< XFS_DIR_LEAF_MAPSIZE
));
947 ASSERT((index
>= 0) && (index
<= INT_GET(hdr
->count
, ARCH_CONVERT
)));
950 * Force open some space in the entry array and fill it in.
952 entry
= &leaf
->entries
[index
];
953 if (index
< INT_GET(hdr
->count
, ARCH_CONVERT
)) {
954 tmp
= INT_GET(hdr
->count
, ARCH_CONVERT
) - index
;
955 tmp
*= (uint
)sizeof(xfs_dir_leaf_entry_t
);
956 memmove(entry
+ 1, entry
, tmp
);
957 xfs_da_log_buf(args
->trans
, bp
,
958 XFS_DA_LOGRANGE(leaf
, entry
, tmp
+ (uint
)sizeof(*entry
)));
960 INT_MOD(hdr
->count
, ARCH_CONVERT
, +1);
963 * Allocate space for the new string (at the end of the run).
965 map
= &hdr
->freemap
[mapindex
];
966 mp
= args
->trans
->t_mountp
;
967 ASSERT(INT_GET(map
->base
, ARCH_CONVERT
) < XFS_LBSIZE(mp
));
968 ASSERT(INT_GET(map
->size
, ARCH_CONVERT
) >= XFS_DIR_LEAF_ENTSIZE_BYNAME(args
->namelen
));
969 ASSERT(INT_GET(map
->size
, ARCH_CONVERT
) < XFS_LBSIZE(mp
));
970 INT_MOD(map
->size
, ARCH_CONVERT
, -(XFS_DIR_LEAF_ENTSIZE_BYNAME(args
->namelen
)));
971 INT_SET(entry
->nameidx
, ARCH_CONVERT
, INT_GET(map
->base
, ARCH_CONVERT
) + INT_GET(map
->size
, ARCH_CONVERT
));
972 INT_SET(entry
->hashval
, ARCH_CONVERT
, args
->hashval
);
973 entry
->namelen
= args
->namelen
;
974 xfs_da_log_buf(args
->trans
, bp
,
975 XFS_DA_LOGRANGE(leaf
, entry
, sizeof(*entry
)));
978 * Copy the string and inode number into the new space.
980 namest
= XFS_DIR_LEAF_NAMESTRUCT(leaf
, INT_GET(entry
->nameidx
, ARCH_CONVERT
));
981 XFS_DIR_SF_PUT_DIRINO(&args
->inumber
, &namest
->inumber
);
982 memcpy(namest
->name
, args
->name
, args
->namelen
);
983 xfs_da_log_buf(args
->trans
, bp
,
984 XFS_DA_LOGRANGE(leaf
, namest
, XFS_DIR_LEAF_ENTSIZE_BYENTRY(entry
)));
987 * Update the control info for this leaf node
989 if (INT_GET(entry
->nameidx
, ARCH_CONVERT
) < INT_GET(hdr
->firstused
, ARCH_CONVERT
))
990 INT_COPY(hdr
->firstused
, entry
->nameidx
, ARCH_CONVERT
);
991 ASSERT(INT_GET(hdr
->firstused
, ARCH_CONVERT
) >= ((INT_GET(hdr
->count
, ARCH_CONVERT
)*sizeof(*entry
))+sizeof(*hdr
)));
992 tmp
= (INT_GET(hdr
->count
, ARCH_CONVERT
)-1) * (uint
)sizeof(xfs_dir_leaf_entry_t
)
993 + (uint
)sizeof(xfs_dir_leaf_hdr_t
);
994 map
= &hdr
->freemap
[0];
995 for (i
= 0; i
< XFS_DIR_LEAF_MAPSIZE
; map
++, i
++) {
996 if (INT_GET(map
->base
, ARCH_CONVERT
) == tmp
) {
997 INT_MOD(map
->base
, ARCH_CONVERT
, (uint
)sizeof(xfs_dir_leaf_entry_t
));
998 INT_MOD(map
->size
, ARCH_CONVERT
, -((uint
)sizeof(xfs_dir_leaf_entry_t
)));
1001 INT_MOD(hdr
->namebytes
, ARCH_CONVERT
, args
->namelen
);
1002 xfs_da_log_buf(args
->trans
, bp
,
1003 XFS_DA_LOGRANGE(leaf
, hdr
, sizeof(*hdr
)));
1007 * Garbage collect a leaf directory block by copying it to a new buffer.
1010 xfs_dir_leaf_compact(xfs_trans_t
*trans
, xfs_dabuf_t
*bp
, int musthave
,
1013 xfs_dir_leafblock_t
*leaf_s
, *leaf_d
;
1014 xfs_dir_leaf_hdr_t
*hdr_s
, *hdr_d
;
1017 char *tmpbuffer2
=NULL
;
1021 mp
= trans
->t_mountp
;
1022 lbsize
= XFS_LBSIZE(mp
);
1023 tmpbuffer
= kmem_alloc(lbsize
, KM_SLEEP
);
1024 ASSERT(tmpbuffer
!= NULL
);
1025 memcpy(tmpbuffer
, bp
->data
, lbsize
);
1028 * Make a second copy in case xfs_dir_leaf_moveents()
1029 * below destroys the original.
1031 if (musthave
|| justcheck
) {
1032 tmpbuffer2
= kmem_alloc(lbsize
, KM_SLEEP
);
1033 memcpy(tmpbuffer2
, bp
->data
, lbsize
);
1035 memset(bp
->data
, 0, lbsize
);
1038 * Copy basic information
1040 leaf_s
= (xfs_dir_leafblock_t
*)tmpbuffer
;
1042 hdr_s
= &leaf_s
->hdr
;
1043 hdr_d
= &leaf_d
->hdr
;
1044 hdr_d
->info
= hdr_s
->info
; /* struct copy */
1045 INT_SET(hdr_d
->firstused
, ARCH_CONVERT
, lbsize
);
1046 if (!hdr_d
->firstused
)
1047 INT_SET(hdr_d
->firstused
, ARCH_CONVERT
, lbsize
- 1);
1048 hdr_d
->namebytes
= 0;
1051 INT_SET(hdr_d
->freemap
[0].base
, ARCH_CONVERT
, sizeof(xfs_dir_leaf_hdr_t
));
1052 INT_SET(hdr_d
->freemap
[0].size
, ARCH_CONVERT
, INT_GET(hdr_d
->firstused
, ARCH_CONVERT
) - INT_GET(hdr_d
->freemap
[0].base
, ARCH_CONVERT
));
1055 * Copy all entry's in the same (sorted) order,
1056 * but allocate filenames packed and in sequence.
1057 * This changes the source (leaf_s) as well.
1059 xfs_dir_leaf_moveents(leaf_s
, 0, leaf_d
, 0, (int)INT_GET(hdr_s
->count
, ARCH_CONVERT
), mp
);
1061 if (musthave
&& INT_GET(hdr_d
->freemap
[0].size
, ARCH_CONVERT
) < musthave
)
1062 rval
= XFS_ERROR(ENOSPC
);
1066 if (justcheck
|| rval
== ENOSPC
) {
1068 memcpy(bp
->data
, tmpbuffer2
, lbsize
);
1070 xfs_da_log_buf(trans
, bp
, 0, lbsize
- 1);
1073 kmem_free(tmpbuffer
, lbsize
);
1074 if (musthave
|| justcheck
)
1075 kmem_free(tmpbuffer2
, lbsize
);
1080 * Redistribute the directory entries between two leaf nodes,
1081 * taking into account the size of the new entry.
1083 * NOTE: if new block is empty, then it will get the upper half of old block.
1086 xfs_dir_leaf_rebalance(xfs_da_state_t
*state
, xfs_da_state_blk_t
*blk1
,
1087 xfs_da_state_blk_t
*blk2
)
1089 xfs_da_state_blk_t
*tmp_blk
;
1090 xfs_dir_leafblock_t
*leaf1
, *leaf2
;
1091 xfs_dir_leaf_hdr_t
*hdr1
, *hdr2
;
1092 int count
, totallen
, max
, space
, swap
;
1095 * Set up environment.
1097 ASSERT(blk1
->magic
== XFS_DIR_LEAF_MAGIC
);
1098 ASSERT(blk2
->magic
== XFS_DIR_LEAF_MAGIC
);
1099 leaf1
= blk1
->bp
->data
;
1100 leaf2
= blk2
->bp
->data
;
1101 ASSERT(be16_to_cpu(leaf1
->hdr
.info
.magic
) == XFS_DIR_LEAF_MAGIC
);
1102 ASSERT(be16_to_cpu(leaf2
->hdr
.info
.magic
) == XFS_DIR_LEAF_MAGIC
);
1105 * Check ordering of blocks, reverse if it makes things simpler.
1108 if (xfs_dir_leaf_order(blk1
->bp
, blk2
->bp
)) {
1112 leaf1
= blk1
->bp
->data
;
1113 leaf2
= blk2
->bp
->data
;
1120 * Examine entries until we reduce the absolute difference in
1121 * byte usage between the two blocks to a minimum. Then get
1122 * the direction to copy and the number of elements to move.
1124 state
->inleaf
= xfs_dir_leaf_figure_balance(state
, blk1
, blk2
,
1127 state
->inleaf
= !state
->inleaf
;
1130 * Move any entries required from leaf to leaf:
1132 if (count
< INT_GET(hdr1
->count
, ARCH_CONVERT
)) {
1134 * Figure the total bytes to be added to the destination leaf.
1136 count
= INT_GET(hdr1
->count
, ARCH_CONVERT
) - count
; /* number entries being moved */
1137 space
= INT_GET(hdr1
->namebytes
, ARCH_CONVERT
) - totallen
;
1138 space
+= count
* ((uint
)sizeof(xfs_dir_leaf_name_t
)-1);
1139 space
+= count
* (uint
)sizeof(xfs_dir_leaf_entry_t
);
1142 * leaf2 is the destination, compact it if it looks tight.
1144 max
= INT_GET(hdr2
->firstused
, ARCH_CONVERT
) - (uint
)sizeof(xfs_dir_leaf_hdr_t
);
1145 max
-= INT_GET(hdr2
->count
, ARCH_CONVERT
) * (uint
)sizeof(xfs_dir_leaf_entry_t
);
1147 xfs_dir_leaf_compact(state
->args
->trans
, blk2
->bp
,
1152 * Move high entries from leaf1 to low end of leaf2.
1154 xfs_dir_leaf_moveents(leaf1
, INT_GET(hdr1
->count
, ARCH_CONVERT
) - count
,
1155 leaf2
, 0, count
, state
->mp
);
1157 xfs_da_log_buf(state
->args
->trans
, blk1
->bp
, 0,
1158 state
->blocksize
-1);
1159 xfs_da_log_buf(state
->args
->trans
, blk2
->bp
, 0,
1160 state
->blocksize
-1);
1162 } else if (count
> INT_GET(hdr1
->count
, ARCH_CONVERT
)) {
1164 * Figure the total bytes to be added to the destination leaf.
1166 count
-= INT_GET(hdr1
->count
, ARCH_CONVERT
); /* number entries being moved */
1167 space
= totallen
- INT_GET(hdr1
->namebytes
, ARCH_CONVERT
);
1168 space
+= count
* ((uint
)sizeof(xfs_dir_leaf_name_t
)-1);
1169 space
+= count
* (uint
)sizeof(xfs_dir_leaf_entry_t
);
1172 * leaf1 is the destination, compact it if it looks tight.
1174 max
= INT_GET(hdr1
->firstused
, ARCH_CONVERT
) - (uint
)sizeof(xfs_dir_leaf_hdr_t
);
1175 max
-= INT_GET(hdr1
->count
, ARCH_CONVERT
) * (uint
)sizeof(xfs_dir_leaf_entry_t
);
1177 xfs_dir_leaf_compact(state
->args
->trans
, blk1
->bp
,
1182 * Move low entries from leaf2 to high end of leaf1.
1184 xfs_dir_leaf_moveents(leaf2
, 0, leaf1
, (int)INT_GET(hdr1
->count
, ARCH_CONVERT
),
1187 xfs_da_log_buf(state
->args
->trans
, blk1
->bp
, 0,
1188 state
->blocksize
-1);
1189 xfs_da_log_buf(state
->args
->trans
, blk2
->bp
, 0,
1190 state
->blocksize
-1);
1194 * Copy out last hashval in each block for B-tree code.
1196 blk1
->hashval
= INT_GET(leaf1
->entries
[ INT_GET(leaf1
->hdr
.count
, ARCH_CONVERT
)-1 ].hashval
, ARCH_CONVERT
);
1197 blk2
->hashval
= INT_GET(leaf2
->entries
[ INT_GET(leaf2
->hdr
.count
, ARCH_CONVERT
)-1 ].hashval
, ARCH_CONVERT
);
1200 * Adjust the expected index for insertion.
1201 * GROT: this doesn't work unless blk2 was originally empty.
1203 if (!state
->inleaf
) {
1204 blk2
->index
= blk1
->index
- INT_GET(leaf1
->hdr
.count
, ARCH_CONVERT
);
1209 * Examine entries until we reduce the absolute difference in
1210 * byte usage between the two blocks to a minimum.
1211 * GROT: Is this really necessary? With other than a 512 byte blocksize,
1212 * GROT: there will always be enough room in either block for a new entry.
1213 * GROT: Do a double-split for this case?
1216 xfs_dir_leaf_figure_balance(xfs_da_state_t
*state
,
1217 xfs_da_state_blk_t
*blk1
,
1218 xfs_da_state_blk_t
*blk2
,
1219 int *countarg
, int *namebytesarg
)
1221 xfs_dir_leafblock_t
*leaf1
, *leaf2
;
1222 xfs_dir_leaf_hdr_t
*hdr1
, *hdr2
;
1223 xfs_dir_leaf_entry_t
*entry
;
1224 int count
, max
, totallen
, half
;
1225 int lastdelta
, foundit
, tmp
;
1228 * Set up environment.
1230 leaf1
= blk1
->bp
->data
;
1231 leaf2
= blk2
->bp
->data
;
1238 * Examine entries until we reduce the absolute difference in
1239 * byte usage between the two blocks to a minimum.
1241 max
= INT_GET(hdr1
->count
, ARCH_CONVERT
) + INT_GET(hdr2
->count
, ARCH_CONVERT
);
1242 half
= (max
+1) * (uint
)(sizeof(*entry
)+sizeof(xfs_dir_leaf_entry_t
)-1);
1243 half
+= INT_GET(hdr1
->namebytes
, ARCH_CONVERT
) + INT_GET(hdr2
->namebytes
, ARCH_CONVERT
) + state
->args
->namelen
;
1245 lastdelta
= state
->blocksize
;
1246 entry
= &leaf1
->entries
[0];
1247 for (count
= 0; count
< max
; entry
++, count
++) {
1249 #define XFS_DIR_ABS(A) (((A) < 0) ? -(A) : (A))
1251 * The new entry is in the first block, account for it.
1253 if (count
== blk1
->index
) {
1254 tmp
= totallen
+ (uint
)sizeof(*entry
)
1255 + XFS_DIR_LEAF_ENTSIZE_BYNAME(state
->args
->namelen
);
1256 if (XFS_DIR_ABS(half
- tmp
) > lastdelta
)
1258 lastdelta
= XFS_DIR_ABS(half
- tmp
);
1264 * Wrap around into the second block if necessary.
1266 if (count
== INT_GET(hdr1
->count
, ARCH_CONVERT
)) {
1268 entry
= &leaf1
->entries
[0];
1272 * Figure out if next leaf entry would be too much.
1274 tmp
= totallen
+ (uint
)sizeof(*entry
)
1275 + XFS_DIR_LEAF_ENTSIZE_BYENTRY(entry
);
1276 if (XFS_DIR_ABS(half
- tmp
) > lastdelta
)
1278 lastdelta
= XFS_DIR_ABS(half
- tmp
);
1284 * Calculate the number of namebytes that will end up in lower block.
1285 * If new entry not in lower block, fix up the count.
1288 count
* (uint
)(sizeof(*entry
)+sizeof(xfs_dir_leaf_entry_t
)-1);
1290 totallen
-= (sizeof(*entry
)+sizeof(xfs_dir_leaf_entry_t
)-1) +
1291 state
->args
->namelen
;
1295 *namebytesarg
= totallen
;
1299 /*========================================================================
1300 * Routines used for shrinking the Btree.
1301 *========================================================================*/
1304 * Check a leaf block and its neighbors to see if the block should be
1305 * collapsed into one or the other neighbor. Always keep the block
1306 * with the smaller block number.
1307 * If the current block is over 50% full, don't try to join it, return 0.
1308 * If the block is empty, fill in the state structure and return 2.
1309 * If it can be collapsed, fill in the state structure and return 1.
1310 * If nothing can be done, return 0.
1313 xfs_dir_leaf_toosmall(xfs_da_state_t
*state
, int *action
)
1315 xfs_dir_leafblock_t
*leaf
;
1316 xfs_da_state_blk_t
*blk
;
1317 xfs_da_blkinfo_t
*info
;
1318 int count
, bytes
, forward
, error
, retval
, i
;
1323 * Check for the degenerate case of the block being over 50% full.
1324 * If so, it's not worth even looking to see if we might be able
1325 * to coalesce with a sibling.
1327 blk
= &state
->path
.blk
[ state
->path
.active
-1 ];
1328 info
= blk
->bp
->data
;
1329 ASSERT(be16_to_cpu(info
->magic
) == XFS_DIR_LEAF_MAGIC
);
1330 leaf
= (xfs_dir_leafblock_t
*)info
;
1331 count
= INT_GET(leaf
->hdr
.count
, ARCH_CONVERT
);
1332 bytes
= (uint
)sizeof(xfs_dir_leaf_hdr_t
) +
1333 count
* (uint
)sizeof(xfs_dir_leaf_entry_t
) +
1334 count
* ((uint
)sizeof(xfs_dir_leaf_name_t
)-1) +
1335 INT_GET(leaf
->hdr
.namebytes
, ARCH_CONVERT
);
1336 if (bytes
> (state
->blocksize
>> 1)) {
1337 *action
= 0; /* blk over 50%, don't try to join */
1342 * Check for the degenerate case of the block being empty.
1343 * If the block is empty, we'll simply delete it, no need to
1344 * coalesce it with a sibling block. We choose (arbitrarily)
1345 * to merge with the forward block unless it is NULL.
1349 * Make altpath point to the block we want to keep and
1350 * path point to the block we want to drop (this one).
1352 forward
= (info
->forw
!= 0);
1353 memcpy(&state
->altpath
, &state
->path
, sizeof(state
->path
));
1354 error
= xfs_da_path_shift(state
, &state
->altpath
, forward
,
1367 * Examine each sibling block to see if we can coalesce with
1368 * at least 25% free space to spare. We need to figure out
1369 * whether to merge with the forward or the backward block.
1370 * We prefer coalescing with the lower numbered sibling so as
1371 * to shrink a directory over time.
1373 forward
= (be32_to_cpu(info
->forw
) < be32_to_cpu(info
->back
)); /* start with smaller blk num */
1374 for (i
= 0; i
< 2; forward
= !forward
, i
++) {
1376 blkno
= be32_to_cpu(info
->forw
);
1378 blkno
= be32_to_cpu(info
->back
);
1381 error
= xfs_da_read_buf(state
->args
->trans
, state
->args
->dp
,
1388 leaf
= (xfs_dir_leafblock_t
*)info
;
1389 count
= INT_GET(leaf
->hdr
.count
, ARCH_CONVERT
);
1390 bytes
= state
->blocksize
- (state
->blocksize
>>2);
1391 bytes
-= INT_GET(leaf
->hdr
.namebytes
, ARCH_CONVERT
);
1393 ASSERT(be16_to_cpu(leaf
->hdr
.info
.magic
) == XFS_DIR_LEAF_MAGIC
);
1394 count
+= INT_GET(leaf
->hdr
.count
, ARCH_CONVERT
);
1395 bytes
-= INT_GET(leaf
->hdr
.namebytes
, ARCH_CONVERT
);
1396 bytes
-= count
* ((uint
)sizeof(xfs_dir_leaf_name_t
) - 1);
1397 bytes
-= count
* (uint
)sizeof(xfs_dir_leaf_entry_t
);
1398 bytes
-= (uint
)sizeof(xfs_dir_leaf_hdr_t
);
1400 break; /* fits with at least 25% to spare */
1402 xfs_da_brelse(state
->args
->trans
, bp
);
1408 xfs_da_buf_done(bp
);
1411 * Make altpath point to the block we want to keep (the lower
1412 * numbered block) and path point to the block we want to drop.
1414 memcpy(&state
->altpath
, &state
->path
, sizeof(state
->path
));
1415 if (blkno
< blk
->blkno
) {
1416 error
= xfs_da_path_shift(state
, &state
->altpath
, forward
,
1419 error
= xfs_da_path_shift(state
, &state
->path
, forward
,
1433 * Remove a name from the leaf directory structure.
1435 * Return 1 if leaf is less than 37% full, 0 if >= 37% full.
1436 * If two leaves are 37% full, when combined they will leave 25% free.
1439 xfs_dir_leaf_remove(xfs_trans_t
*trans
, xfs_dabuf_t
*bp
, int index
)
1441 xfs_dir_leafblock_t
*leaf
;
1442 xfs_dir_leaf_hdr_t
*hdr
;
1443 xfs_dir_leaf_map_t
*map
;
1444 xfs_dir_leaf_entry_t
*entry
;
1445 xfs_dir_leaf_name_t
*namest
;
1446 int before
, after
, smallest
, entsize
;
1447 int tablesize
, tmp
, i
;
1451 ASSERT(be16_to_cpu(leaf
->hdr
.info
.magic
) == XFS_DIR_LEAF_MAGIC
);
1453 mp
= trans
->t_mountp
;
1454 ASSERT((INT_GET(hdr
->count
, ARCH_CONVERT
) > 0) && (INT_GET(hdr
->count
, ARCH_CONVERT
) < (XFS_LBSIZE(mp
)/8)));
1455 ASSERT((index
>= 0) && (index
< INT_GET(hdr
->count
, ARCH_CONVERT
)));
1456 ASSERT(INT_GET(hdr
->firstused
, ARCH_CONVERT
) >= ((INT_GET(hdr
->count
, ARCH_CONVERT
)*sizeof(*entry
))+sizeof(*hdr
)));
1457 entry
= &leaf
->entries
[index
];
1458 ASSERT(INT_GET(entry
->nameidx
, ARCH_CONVERT
) >= INT_GET(hdr
->firstused
, ARCH_CONVERT
));
1459 ASSERT(INT_GET(entry
->nameidx
, ARCH_CONVERT
) < XFS_LBSIZE(mp
));
1462 * Scan through free region table:
1463 * check for adjacency of free'd entry with an existing one,
1464 * find smallest free region in case we need to replace it,
1465 * adjust any map that borders the entry table,
1467 tablesize
= INT_GET(hdr
->count
, ARCH_CONVERT
) * (uint
)sizeof(xfs_dir_leaf_entry_t
)
1468 + (uint
)sizeof(xfs_dir_leaf_hdr_t
);
1469 map
= &hdr
->freemap
[0];
1470 tmp
= INT_GET(map
->size
, ARCH_CONVERT
);
1471 before
= after
= -1;
1472 smallest
= XFS_DIR_LEAF_MAPSIZE
- 1;
1473 entsize
= XFS_DIR_LEAF_ENTSIZE_BYENTRY(entry
);
1474 for (i
= 0; i
< XFS_DIR_LEAF_MAPSIZE
; map
++, i
++) {
1475 ASSERT(INT_GET(map
->base
, ARCH_CONVERT
) < XFS_LBSIZE(mp
));
1476 ASSERT(INT_GET(map
->size
, ARCH_CONVERT
) < XFS_LBSIZE(mp
));
1477 if (INT_GET(map
->base
, ARCH_CONVERT
) == tablesize
) {
1478 INT_MOD(map
->base
, ARCH_CONVERT
, -((uint
)sizeof(xfs_dir_leaf_entry_t
)));
1479 INT_MOD(map
->size
, ARCH_CONVERT
, (uint
)sizeof(xfs_dir_leaf_entry_t
));
1482 if ((INT_GET(map
->base
, ARCH_CONVERT
) + INT_GET(map
->size
, ARCH_CONVERT
)) == INT_GET(entry
->nameidx
, ARCH_CONVERT
)) {
1484 } else if (INT_GET(map
->base
, ARCH_CONVERT
) == (INT_GET(entry
->nameidx
, ARCH_CONVERT
) + entsize
)) {
1486 } else if (INT_GET(map
->size
, ARCH_CONVERT
) < tmp
) {
1487 tmp
= INT_GET(map
->size
, ARCH_CONVERT
);
1493 * Coalesce adjacent freemap regions,
1494 * or replace the smallest region.
1496 if ((before
>= 0) || (after
>= 0)) {
1497 if ((before
>= 0) && (after
>= 0)) {
1498 map
= &hdr
->freemap
[before
];
1499 INT_MOD(map
->size
, ARCH_CONVERT
, entsize
);
1500 INT_MOD(map
->size
, ARCH_CONVERT
, INT_GET(hdr
->freemap
[after
].size
, ARCH_CONVERT
));
1501 hdr
->freemap
[after
].base
= 0;
1502 hdr
->freemap
[after
].size
= 0;
1503 } else if (before
>= 0) {
1504 map
= &hdr
->freemap
[before
];
1505 INT_MOD(map
->size
, ARCH_CONVERT
, entsize
);
1507 map
= &hdr
->freemap
[after
];
1508 INT_COPY(map
->base
, entry
->nameidx
, ARCH_CONVERT
);
1509 INT_MOD(map
->size
, ARCH_CONVERT
, entsize
);
1513 * Replace smallest region (if it is smaller than free'd entry)
1515 map
= &hdr
->freemap
[smallest
];
1516 if (INT_GET(map
->size
, ARCH_CONVERT
) < entsize
) {
1517 INT_COPY(map
->base
, entry
->nameidx
, ARCH_CONVERT
);
1518 INT_SET(map
->size
, ARCH_CONVERT
, entsize
);
1523 * Did we remove the first entry?
1525 if (INT_GET(entry
->nameidx
, ARCH_CONVERT
) == INT_GET(hdr
->firstused
, ARCH_CONVERT
))
1531 * Compress the remaining entries and zero out the removed stuff.
1533 namest
= XFS_DIR_LEAF_NAMESTRUCT(leaf
, INT_GET(entry
->nameidx
, ARCH_CONVERT
));
1534 memset((char *)namest
, 0, entsize
);
1535 xfs_da_log_buf(trans
, bp
, XFS_DA_LOGRANGE(leaf
, namest
, entsize
));
1537 INT_MOD(hdr
->namebytes
, ARCH_CONVERT
, -(entry
->namelen
));
1538 tmp
= (INT_GET(hdr
->count
, ARCH_CONVERT
) - index
) * (uint
)sizeof(xfs_dir_leaf_entry_t
);
1539 memmove(entry
, entry
+ 1, tmp
);
1540 INT_MOD(hdr
->count
, ARCH_CONVERT
, -1);
1541 xfs_da_log_buf(trans
, bp
,
1542 XFS_DA_LOGRANGE(leaf
, entry
, tmp
+ (uint
)sizeof(*entry
)));
1543 entry
= &leaf
->entries
[INT_GET(hdr
->count
, ARCH_CONVERT
)];
1544 memset((char *)entry
, 0, sizeof(xfs_dir_leaf_entry_t
));
1547 * If we removed the first entry, re-find the first used byte
1548 * in the name area. Note that if the entry was the "firstused",
1549 * then we don't have a "hole" in our block resulting from
1550 * removing the name.
1553 tmp
= XFS_LBSIZE(mp
);
1554 entry
= &leaf
->entries
[0];
1555 for (i
= INT_GET(hdr
->count
, ARCH_CONVERT
)-1; i
>= 0; entry
++, i
--) {
1556 ASSERT(INT_GET(entry
->nameidx
, ARCH_CONVERT
) >= INT_GET(hdr
->firstused
, ARCH_CONVERT
));
1557 ASSERT(INT_GET(entry
->nameidx
, ARCH_CONVERT
) < XFS_LBSIZE(mp
));
1558 if (INT_GET(entry
->nameidx
, ARCH_CONVERT
) < tmp
)
1559 tmp
= INT_GET(entry
->nameidx
, ARCH_CONVERT
);
1561 INT_SET(hdr
->firstused
, ARCH_CONVERT
, tmp
);
1562 if (!hdr
->firstused
)
1563 INT_SET(hdr
->firstused
, ARCH_CONVERT
, tmp
- 1);
1565 hdr
->holes
= 1; /* mark as needing compaction */
1568 xfs_da_log_buf(trans
, bp
, XFS_DA_LOGRANGE(leaf
, hdr
, sizeof(*hdr
)));
1571 * Check if leaf is less than 50% full, caller may want to
1572 * "join" the leaf with a sibling if so.
1574 tmp
= (uint
)sizeof(xfs_dir_leaf_hdr_t
);
1575 tmp
+= INT_GET(leaf
->hdr
.count
, ARCH_CONVERT
) * (uint
)sizeof(xfs_dir_leaf_entry_t
);
1576 tmp
+= INT_GET(leaf
->hdr
.count
, ARCH_CONVERT
) * ((uint
)sizeof(xfs_dir_leaf_name_t
) - 1);
1577 tmp
+= INT_GET(leaf
->hdr
.namebytes
, ARCH_CONVERT
);
1578 if (tmp
< mp
->m_dir_magicpct
)
1579 return 1; /* leaf is < 37% full */
1584 * Move all the directory entries from drop_leaf into save_leaf.
1587 xfs_dir_leaf_unbalance(xfs_da_state_t
*state
, xfs_da_state_blk_t
*drop_blk
,
1588 xfs_da_state_blk_t
*save_blk
)
1590 xfs_dir_leafblock_t
*drop_leaf
, *save_leaf
, *tmp_leaf
;
1591 xfs_dir_leaf_hdr_t
*drop_hdr
, *save_hdr
, *tmp_hdr
;
1596 * Set up environment.
1599 ASSERT(drop_blk
->magic
== XFS_DIR_LEAF_MAGIC
);
1600 ASSERT(save_blk
->magic
== XFS_DIR_LEAF_MAGIC
);
1601 drop_leaf
= drop_blk
->bp
->data
;
1602 save_leaf
= save_blk
->bp
->data
;
1603 ASSERT(be16_to_cpu(drop_leaf
->hdr
.info
.magic
) == XFS_DIR_LEAF_MAGIC
);
1604 ASSERT(be16_to_cpu(save_leaf
->hdr
.info
.magic
) == XFS_DIR_LEAF_MAGIC
);
1605 drop_hdr
= &drop_leaf
->hdr
;
1606 save_hdr
= &save_leaf
->hdr
;
1609 * Save last hashval from dying block for later Btree fixup.
1611 drop_blk
->hashval
= INT_GET(drop_leaf
->entries
[ drop_leaf
->hdr
.count
-1 ].hashval
, ARCH_CONVERT
);
1614 * Check if we need a temp buffer, or can we do it in place.
1615 * Note that we don't check "leaf" for holes because we will
1616 * always be dropping it, toosmall() decided that for us already.
1618 if (save_hdr
->holes
== 0) {
1620 * dest leaf has no holes, so we add there. May need
1621 * to make some room in the entry array.
1623 if (xfs_dir_leaf_order(save_blk
->bp
, drop_blk
->bp
)) {
1624 xfs_dir_leaf_moveents(drop_leaf
, 0, save_leaf
, 0,
1625 (int)INT_GET(drop_hdr
->count
, ARCH_CONVERT
), mp
);
1627 xfs_dir_leaf_moveents(drop_leaf
, 0,
1628 save_leaf
, INT_GET(save_hdr
->count
, ARCH_CONVERT
),
1629 (int)INT_GET(drop_hdr
->count
, ARCH_CONVERT
), mp
);
1633 * Destination has holes, so we make a temporary copy
1634 * of the leaf and add them both to that.
1636 tmpbuffer
= kmem_alloc(state
->blocksize
, KM_SLEEP
);
1637 ASSERT(tmpbuffer
!= NULL
);
1638 memset(tmpbuffer
, 0, state
->blocksize
);
1639 tmp_leaf
= (xfs_dir_leafblock_t
*)tmpbuffer
;
1640 tmp_hdr
= &tmp_leaf
->hdr
;
1641 tmp_hdr
->info
= save_hdr
->info
; /* struct copy */
1643 INT_SET(tmp_hdr
->firstused
, ARCH_CONVERT
, state
->blocksize
);
1644 if (!tmp_hdr
->firstused
)
1645 INT_SET(tmp_hdr
->firstused
, ARCH_CONVERT
, state
->blocksize
- 1);
1646 tmp_hdr
->namebytes
= 0;
1647 if (xfs_dir_leaf_order(save_blk
->bp
, drop_blk
->bp
)) {
1648 xfs_dir_leaf_moveents(drop_leaf
, 0, tmp_leaf
, 0,
1649 (int)INT_GET(drop_hdr
->count
, ARCH_CONVERT
), mp
);
1650 xfs_dir_leaf_moveents(save_leaf
, 0,
1651 tmp_leaf
, INT_GET(tmp_leaf
->hdr
.count
, ARCH_CONVERT
),
1652 (int)INT_GET(save_hdr
->count
, ARCH_CONVERT
), mp
);
1654 xfs_dir_leaf_moveents(save_leaf
, 0, tmp_leaf
, 0,
1655 (int)INT_GET(save_hdr
->count
, ARCH_CONVERT
), mp
);
1656 xfs_dir_leaf_moveents(drop_leaf
, 0,
1657 tmp_leaf
, INT_GET(tmp_leaf
->hdr
.count
, ARCH_CONVERT
),
1658 (int)INT_GET(drop_hdr
->count
, ARCH_CONVERT
), mp
);
1660 memcpy(save_leaf
, tmp_leaf
, state
->blocksize
);
1661 kmem_free(tmpbuffer
, state
->blocksize
);
1664 xfs_da_log_buf(state
->args
->trans
, save_blk
->bp
, 0,
1665 state
->blocksize
- 1);
1668 * Copy out last hashval in each block for B-tree code.
1670 save_blk
->hashval
= INT_GET(save_leaf
->entries
[ INT_GET(save_leaf
->hdr
.count
, ARCH_CONVERT
)-1 ].hashval
, ARCH_CONVERT
);
1673 /*========================================================================
1674 * Routines used for finding things in the Btree.
1675 *========================================================================*/
1678 * Look up a name in a leaf directory structure.
1679 * This is the internal routine, it uses the caller's buffer.
1681 * Note that duplicate keys are allowed, but only check within the
1682 * current leaf node. The Btree code must check in adjacent leaf nodes.
1684 * Return in *index the index into the entry[] array of either the found
1685 * entry, or where the entry should have been (insert before that entry).
1687 * Don't change the args->inumber unless we find the filename.
1690 xfs_dir_leaf_lookup_int(xfs_dabuf_t
*bp
, xfs_da_args_t
*args
, int *index
)
1692 xfs_dir_leafblock_t
*leaf
;
1693 xfs_dir_leaf_entry_t
*entry
;
1694 xfs_dir_leaf_name_t
*namest
;
1696 xfs_dahash_t hashval
;
1699 ASSERT(be16_to_cpu(leaf
->hdr
.info
.magic
) == XFS_DIR_LEAF_MAGIC
);
1700 ASSERT(INT_GET(leaf
->hdr
.count
, ARCH_CONVERT
) < (XFS_LBSIZE(args
->dp
->i_mount
)/8));
1703 * Binary search. (note: small blocks will skip this loop)
1705 hashval
= args
->hashval
;
1706 probe
= span
= INT_GET(leaf
->hdr
.count
, ARCH_CONVERT
) / 2;
1707 for (entry
= &leaf
->entries
[probe
]; span
> 4;
1708 entry
= &leaf
->entries
[probe
]) {
1710 if (INT_GET(entry
->hashval
, ARCH_CONVERT
) < hashval
)
1712 else if (INT_GET(entry
->hashval
, ARCH_CONVERT
) > hashval
)
1717 ASSERT((probe
>= 0) && \
1718 ((!leaf
->hdr
.count
) || (probe
< INT_GET(leaf
->hdr
.count
, ARCH_CONVERT
))));
1719 ASSERT((span
<= 4) || (INT_GET(entry
->hashval
, ARCH_CONVERT
) == hashval
));
1722 * Since we may have duplicate hashval's, find the first matching
1723 * hashval in the leaf.
1725 while ((probe
> 0) && (INT_GET(entry
->hashval
, ARCH_CONVERT
) >= hashval
)) {
1729 while ((probe
< INT_GET(leaf
->hdr
.count
, ARCH_CONVERT
)) && (INT_GET(entry
->hashval
, ARCH_CONVERT
) < hashval
)) {
1733 if ((probe
== INT_GET(leaf
->hdr
.count
, ARCH_CONVERT
)) || (INT_GET(entry
->hashval
, ARCH_CONVERT
) != hashval
)) {
1735 ASSERT(args
->oknoent
);
1736 return XFS_ERROR(ENOENT
);
1740 * Duplicate keys may be present, so search all of them for a match.
1742 while ((probe
< INT_GET(leaf
->hdr
.count
, ARCH_CONVERT
)) && (INT_GET(entry
->hashval
, ARCH_CONVERT
) == hashval
)) {
1743 namest
= XFS_DIR_LEAF_NAMESTRUCT(leaf
, INT_GET(entry
->nameidx
, ARCH_CONVERT
));
1744 if (entry
->namelen
== args
->namelen
&&
1745 namest
->name
[0] == args
->name
[0] &&
1746 memcmp(args
->name
, namest
->name
, args
->namelen
) == 0) {
1747 XFS_DIR_SF_GET_DIRINO(&namest
->inumber
, &args
->inumber
);
1749 return XFS_ERROR(EEXIST
);
1755 ASSERT(probe
== INT_GET(leaf
->hdr
.count
, ARCH_CONVERT
) || args
->oknoent
);
1756 return XFS_ERROR(ENOENT
);
1759 /*========================================================================
1761 *========================================================================*/
1764 * Move the indicated entries from one leaf to another.
1765 * NOTE: this routine modifies both source and destination leaves.
1769 xfs_dir_leaf_moveents(xfs_dir_leafblock_t
*leaf_s
, int start_s
,
1770 xfs_dir_leafblock_t
*leaf_d
, int start_d
,
1771 int count
, xfs_mount_t
*mp
)
1773 xfs_dir_leaf_hdr_t
*hdr_s
, *hdr_d
;
1774 xfs_dir_leaf_entry_t
*entry_s
, *entry_d
;
1778 * Check for nothing to do.
1784 * Set up environment.
1786 ASSERT(be16_to_cpu(leaf_s
->hdr
.info
.magic
) == XFS_DIR_LEAF_MAGIC
);
1787 ASSERT(be16_to_cpu(leaf_d
->hdr
.info
.magic
) == XFS_DIR_LEAF_MAGIC
);
1788 hdr_s
= &leaf_s
->hdr
;
1789 hdr_d
= &leaf_d
->hdr
;
1790 ASSERT((INT_GET(hdr_s
->count
, ARCH_CONVERT
) > 0) && (INT_GET(hdr_s
->count
, ARCH_CONVERT
) < (XFS_LBSIZE(mp
)/8)));
1791 ASSERT(INT_GET(hdr_s
->firstused
, ARCH_CONVERT
) >=
1792 ((INT_GET(hdr_s
->count
, ARCH_CONVERT
)*sizeof(*entry_s
))+sizeof(*hdr_s
)));
1793 ASSERT(INT_GET(hdr_d
->count
, ARCH_CONVERT
) < (XFS_LBSIZE(mp
)/8));
1794 ASSERT(INT_GET(hdr_d
->firstused
, ARCH_CONVERT
) >=
1795 ((INT_GET(hdr_d
->count
, ARCH_CONVERT
)*sizeof(*entry_d
))+sizeof(*hdr_d
)));
1797 ASSERT(start_s
< INT_GET(hdr_s
->count
, ARCH_CONVERT
));
1798 ASSERT(start_d
<= INT_GET(hdr_d
->count
, ARCH_CONVERT
));
1799 ASSERT(count
<= INT_GET(hdr_s
->count
, ARCH_CONVERT
));
1802 * Move the entries in the destination leaf up to make a hole?
1804 if (start_d
< INT_GET(hdr_d
->count
, ARCH_CONVERT
)) {
1805 tmp
= INT_GET(hdr_d
->count
, ARCH_CONVERT
) - start_d
;
1806 tmp
*= (uint
)sizeof(xfs_dir_leaf_entry_t
);
1807 entry_s
= &leaf_d
->entries
[start_d
];
1808 entry_d
= &leaf_d
->entries
[start_d
+ count
];
1809 memcpy(entry_d
, entry_s
, tmp
);
1813 * Copy all entry's in the same (sorted) order,
1814 * but allocate filenames packed and in sequence.
1816 entry_s
= &leaf_s
->entries
[start_s
];
1817 entry_d
= &leaf_d
->entries
[start_d
];
1818 for (i
= 0; i
< count
; entry_s
++, entry_d
++, i
++) {
1819 ASSERT(INT_GET(entry_s
->nameidx
, ARCH_CONVERT
) >= INT_GET(hdr_s
->firstused
, ARCH_CONVERT
));
1820 tmp
= XFS_DIR_LEAF_ENTSIZE_BYENTRY(entry_s
);
1821 INT_MOD(hdr_d
->firstused
, ARCH_CONVERT
, -(tmp
));
1822 entry_d
->hashval
= entry_s
->hashval
; /* INT_: direct copy */
1823 INT_COPY(entry_d
->nameidx
, hdr_d
->firstused
, ARCH_CONVERT
);
1824 entry_d
->namelen
= entry_s
->namelen
;
1825 ASSERT(INT_GET(entry_d
->nameidx
, ARCH_CONVERT
) + tmp
<= XFS_LBSIZE(mp
));
1826 memcpy(XFS_DIR_LEAF_NAMESTRUCT(leaf_d
, INT_GET(entry_d
->nameidx
, ARCH_CONVERT
)),
1827 XFS_DIR_LEAF_NAMESTRUCT(leaf_s
, INT_GET(entry_s
->nameidx
, ARCH_CONVERT
)), tmp
);
1828 ASSERT(INT_GET(entry_s
->nameidx
, ARCH_CONVERT
) + tmp
<= XFS_LBSIZE(mp
));
1829 memset((char *)XFS_DIR_LEAF_NAMESTRUCT(leaf_s
, INT_GET(entry_s
->nameidx
, ARCH_CONVERT
)),
1831 INT_MOD(hdr_s
->namebytes
, ARCH_CONVERT
, -(entry_d
->namelen
));
1832 INT_MOD(hdr_d
->namebytes
, ARCH_CONVERT
, entry_d
->namelen
);
1833 INT_MOD(hdr_s
->count
, ARCH_CONVERT
, -1);
1834 INT_MOD(hdr_d
->count
, ARCH_CONVERT
, +1);
1835 tmp
= INT_GET(hdr_d
->count
, ARCH_CONVERT
) * (uint
)sizeof(xfs_dir_leaf_entry_t
)
1836 + (uint
)sizeof(xfs_dir_leaf_hdr_t
);
1837 ASSERT(INT_GET(hdr_d
->firstused
, ARCH_CONVERT
) >= tmp
);
1842 * Zero out the entries we just copied.
1844 if (start_s
== INT_GET(hdr_s
->count
, ARCH_CONVERT
)) {
1845 tmp
= count
* (uint
)sizeof(xfs_dir_leaf_entry_t
);
1846 entry_s
= &leaf_s
->entries
[start_s
];
1847 ASSERT((char *)entry_s
+ tmp
<= (char *)leaf_s
+ XFS_LBSIZE(mp
));
1848 memset((char *)entry_s
, 0, tmp
);
1851 * Move the remaining entries down to fill the hole,
1852 * then zero the entries at the top.
1854 tmp
= INT_GET(hdr_s
->count
, ARCH_CONVERT
) - count
;
1855 tmp
*= (uint
)sizeof(xfs_dir_leaf_entry_t
);
1856 entry_s
= &leaf_s
->entries
[start_s
+ count
];
1857 entry_d
= &leaf_s
->entries
[start_s
];
1858 memcpy(entry_d
, entry_s
, tmp
);
1860 tmp
= count
* (uint
)sizeof(xfs_dir_leaf_entry_t
);
1861 entry_s
= &leaf_s
->entries
[INT_GET(hdr_s
->count
, ARCH_CONVERT
)];
1862 ASSERT((char *)entry_s
+ tmp
<= (char *)leaf_s
+ XFS_LBSIZE(mp
));
1863 memset((char *)entry_s
, 0, tmp
);
1867 * Fill in the freemap information
1869 INT_SET(hdr_d
->freemap
[0].base
, ARCH_CONVERT
, (uint
)sizeof(xfs_dir_leaf_hdr_t
));
1870 INT_MOD(hdr_d
->freemap
[0].base
, ARCH_CONVERT
, INT_GET(hdr_d
->count
, ARCH_CONVERT
) * (uint
)sizeof(xfs_dir_leaf_entry_t
));
1871 INT_SET(hdr_d
->freemap
[0].size
, ARCH_CONVERT
, INT_GET(hdr_d
->firstused
, ARCH_CONVERT
) - INT_GET(hdr_d
->freemap
[0].base
, ARCH_CONVERT
));
1872 INT_SET(hdr_d
->freemap
[1].base
, ARCH_CONVERT
, (hdr_d
->freemap
[2].base
= 0));
1873 INT_SET(hdr_d
->freemap
[1].size
, ARCH_CONVERT
, (hdr_d
->freemap
[2].size
= 0));
1874 hdr_s
->holes
= 1; /* leaf may not be compact */
1878 * Compare two leaf blocks "order".
1881 xfs_dir_leaf_order(xfs_dabuf_t
*leaf1_bp
, xfs_dabuf_t
*leaf2_bp
)
1883 xfs_dir_leafblock_t
*leaf1
, *leaf2
;
1885 leaf1
= leaf1_bp
->data
;
1886 leaf2
= leaf2_bp
->data
;
1887 ASSERT((be16_to_cpu(leaf1
->hdr
.info
.magic
) == XFS_DIR_LEAF_MAGIC
) &&
1888 (be16_to_cpu(leaf2
->hdr
.info
.magic
) == XFS_DIR_LEAF_MAGIC
));
1889 if ((INT_GET(leaf1
->hdr
.count
, ARCH_CONVERT
) > 0) && (INT_GET(leaf2
->hdr
.count
, ARCH_CONVERT
) > 0) &&
1890 ((INT_GET(leaf2
->entries
[ 0 ].hashval
, ARCH_CONVERT
) <
1891 INT_GET(leaf1
->entries
[ 0 ].hashval
, ARCH_CONVERT
)) ||
1892 (INT_GET(leaf2
->entries
[ INT_GET(leaf2
->hdr
.count
, ARCH_CONVERT
)-1 ].hashval
, ARCH_CONVERT
) <
1893 INT_GET(leaf1
->entries
[ INT_GET(leaf1
->hdr
.count
, ARCH_CONVERT
)-1 ].hashval
, ARCH_CONVERT
)))) {
1900 * Pick up the last hashvalue from a leaf block.
1903 xfs_dir_leaf_lasthash(xfs_dabuf_t
*bp
, int *count
)
1905 xfs_dir_leafblock_t
*leaf
;
1908 ASSERT(be16_to_cpu(leaf
->hdr
.info
.magic
) == XFS_DIR_LEAF_MAGIC
);
1910 *count
= INT_GET(leaf
->hdr
.count
, ARCH_CONVERT
);
1911 if (!leaf
->hdr
.count
)
1913 return(INT_GET(leaf
->entries
[ INT_GET(leaf
->hdr
.count
, ARCH_CONVERT
)-1 ].hashval
, ARCH_CONVERT
));
1917 * Copy out directory entries for getdents(), for leaf directories.
1920 xfs_dir_leaf_getdents_int(
1930 xfs_dir_leafblock_t
*leaf
;
1931 xfs_dir_leaf_entry_t
*entry
;
1932 xfs_dir_leaf_name_t
*namest
;
1933 int entno
, want_entno
, i
, nextentno
;
1935 xfs_dahash_t cookhash
;
1936 xfs_dahash_t nexthash
= 0;
1937 #if (BITS_PER_LONG == 32)
1938 xfs_dahash_t lasthash
= XFS_DA_MAXHASH
;
1940 xfs_dir_put_args_t p
;
1944 if (be16_to_cpu(leaf
->hdr
.info
.magic
) != XFS_DIR_LEAF_MAGIC
) {
1946 return XFS_ERROR(ENOENT
); /* XXX wrong code */
1949 want_entno
= XFS_DA_COOKIE_ENTRY(mp
, uio
->uio_offset
);
1951 cookhash
= XFS_DA_COOKIE_HASH(mp
, uio
->uio_offset
);
1953 xfs_dir_trace_g_dul("leaf: start", dp
, uio
, leaf
);
1956 * Re-find our place.
1958 for (i
= entno
= 0, entry
= &leaf
->entries
[0];
1959 i
< INT_GET(leaf
->hdr
.count
, ARCH_CONVERT
);
1962 namest
= XFS_DIR_LEAF_NAMESTRUCT(leaf
,
1963 INT_GET(entry
->nameidx
, ARCH_CONVERT
));
1966 ((char *)namest
< (char *)leaf
) ||
1967 ((char *)namest
>= (char *)leaf
+ XFS_LBSIZE(mp
)))) {
1968 XFS_CORRUPTION_ERROR("xfs_dir_leaf_getdents_int(1)",
1969 XFS_ERRLEVEL_LOW
, mp
, leaf
);
1970 xfs_dir_trace_g_du("leaf: corrupted", dp
, uio
);
1971 return XFS_ERROR(EFSCORRUPTED
);
1973 if (INT_GET(entry
->hashval
, ARCH_CONVERT
) >= cookhash
) {
1974 if ( entno
< want_entno
1975 && INT_GET(entry
->hashval
, ARCH_CONVERT
)
1978 * Trying to get to a particular offset in a
1979 * run of equal-hashval entries.
1982 } else if ( want_entno
> 0
1983 && entno
== want_entno
1984 && INT_GET(entry
->hashval
, ARCH_CONVERT
)
1994 if (i
== INT_GET(leaf
->hdr
.count
, ARCH_CONVERT
)) {
1995 xfs_dir_trace_g_du("leaf: hash not found", dp
, uio
);
1996 if (!leaf
->hdr
.info
.forw
)
1998 XFS_DA_MAKE_COOKIE(mp
, 0, 0, XFS_DA_MAXHASH
);
2000 * Don't set uio_offset if there's another block:
2001 * the node code will be setting uio_offset anyway.
2006 xfs_dir_trace_g_due("leaf: hash found", dp
, uio
, entry
);
2013 * We're synchronized, start copying entries out to the user.
2015 for (; entno
>= 0 && i
< INT_GET(leaf
->hdr
.count
, ARCH_CONVERT
);
2016 entry
++, i
++, (entno
= nextentno
)) {
2017 int lastresid
=0, retval
;
2018 xfs_dircook_t lastoffset
;
2019 xfs_dahash_t thishash
;
2022 * Check for a damaged directory leaf block and pick up
2023 * the inode number from this entry.
2025 namest
= XFS_DIR_LEAF_NAMESTRUCT(leaf
,
2026 INT_GET(entry
->nameidx
, ARCH_CONVERT
));
2029 ((char *)namest
< (char *)leaf
) ||
2030 ((char *)namest
>= (char *)leaf
+ XFS_LBSIZE(mp
)))) {
2031 XFS_CORRUPTION_ERROR("xfs_dir_leaf_getdents_int(2)",
2032 XFS_ERRLEVEL_LOW
, mp
, leaf
);
2033 xfs_dir_trace_g_du("leaf: corrupted", dp
, uio
);
2034 return XFS_ERROR(EFSCORRUPTED
);
2037 xfs_dir_trace_g_duc("leaf: middle cookie ",
2040 if (i
< (INT_GET(leaf
->hdr
.count
, ARCH_CONVERT
) - 1)) {
2041 nexthash
= INT_GET(entry
[1].hashval
, ARCH_CONVERT
);
2043 if (nexthash
== INT_GET(entry
->hashval
, ARCH_CONVERT
))
2044 nextentno
= entno
+ 1;
2047 XFS_PUT_COOKIE(p
.cook
, mp
, bno
, nextentno
, nexthash
);
2048 xfs_dir_trace_g_duc("leaf: middle cookie ",
2051 } else if ((thishash
= be32_to_cpu(leaf
->hdr
.info
.forw
))) {
2053 xfs_dir_leafblock_t
*leaf2
;
2055 ASSERT(nextda
!= -1);
2057 retval
= xfs_da_read_buf(dp
->i_transp
, dp
, thishash
,
2058 nextda
, &bp2
, XFS_DATA_FORK
);
2062 ASSERT(bp2
!= NULL
);
2067 (be16_to_cpu(leaf2
->hdr
.info
.magic
)
2068 != XFS_DIR_LEAF_MAGIC
)
2069 || (be32_to_cpu(leaf2
->hdr
.info
.back
)
2070 != bno
))) { /* GROT */
2071 XFS_CORRUPTION_ERROR("xfs_dir_leaf_getdents_int(3)",
2072 XFS_ERRLEVEL_LOW
, mp
,
2074 xfs_da_brelse(dp
->i_transp
, bp2
);
2076 return XFS_ERROR(EFSCORRUPTED
);
2079 nexthash
= INT_GET(leaf2
->entries
[0].hashval
,
2082 XFS_PUT_COOKIE(p
.cook
, mp
, thishash
, 0, nexthash
);
2083 xfs_da_brelse(dp
->i_transp
, bp2
);
2084 xfs_dir_trace_g_duc("leaf: next blk cookie",
2088 XFS_PUT_COOKIE(p
.cook
, mp
, 0, 0, XFS_DA_MAXHASH
);
2092 * Save off the cookie so we can fall back should the
2093 * 'put' into the outgoing buffer fails. To handle a run
2094 * of equal-hashvals, the off_t structure on 64bit
2095 * builds has entno built into the cookie to ID the
2096 * entry. On 32bit builds, we only have space for the
2097 * hashval so we can't ID specific entries within a group
2098 * of same hashval entries. For this, lastoffset is set
2099 * to the first in the run of equal hashvals so we don't
2100 * include any entries unless we can include all entries
2101 * that share the same hashval. Hopefully the buffer
2102 * provided is big enough to handle it (see pv763517).
2104 #if (BITS_PER_LONG == 32)
2105 if ((thishash
= INT_GET(entry
->hashval
, ARCH_CONVERT
))
2107 XFS_PUT_COOKIE(lastoffset
, mp
, bno
, entno
, thishash
);
2108 lastresid
= uio
->uio_resid
;
2109 lasthash
= thishash
;
2111 xfs_dir_trace_g_duc("leaf: DUP COOKIES, skipped",
2115 thishash
= INT_GET(entry
->hashval
, ARCH_CONVERT
);
2116 XFS_PUT_COOKIE(lastoffset
, mp
, bno
, entno
, thishash
);
2117 lastresid
= uio
->uio_resid
;
2118 #endif /* BITS_PER_LONG == 32 */
2121 * Put the current entry into the outgoing buffer. If we fail
2122 * then restore the UIO to the first entry in the current
2123 * run of equal-hashval entries (probably one 1 entry long).
2125 p
.ino
= XFS_GET_DIR_INO8(namest
->inumber
);
2127 p
.ino
+= mp
->m_inoadd
;
2129 p
.name
= (char *)namest
->name
;
2130 p
.namelen
= entry
->namelen
;
2135 uio
->uio_offset
= lastoffset
.o
;
2136 uio
->uio_resid
= lastresid
;
2140 xfs_dir_trace_g_du("leaf: E-O-B", dp
, uio
);
2146 uio
->uio_offset
= p
.cook
.o
;
2150 xfs_dir_trace_g_du("leaf: E-O-F", dp
, uio
);
2156 * Format a dirent64 structure and copy it out the the user's buffer.
2159 xfs_dir_put_dirent64_direct(xfs_dir_put_args_t
*pa
)
2162 int reclen
, namelen
;
2166 namelen
= pa
->namelen
;
2167 reclen
= DIRENTSIZE(namelen
);
2169 if (reclen
> uio
->uio_resid
) {
2173 iovp
= uio
->uio_iov
;
2174 idbp
= (xfs_dirent_t
*)iovp
->iov_base
;
2175 iovp
->iov_base
= (char *)idbp
+ reclen
;
2176 iovp
->iov_len
-= reclen
;
2177 uio
->uio_resid
-= reclen
;
2178 idbp
->d_reclen
= reclen
;
2179 idbp
->d_ino
= pa
->ino
;
2180 idbp
->d_off
= pa
->cook
.o
;
2181 idbp
->d_name
[namelen
] = '\0';
2183 memcpy(idbp
->d_name
, pa
->name
, namelen
);
2188 * Format a dirent64 structure and copy it out the the user's buffer.
2191 xfs_dir_put_dirent64_uio(xfs_dir_put_args_t
*pa
)
2193 int retval
, reclen
, namelen
;
2197 namelen
= pa
->namelen
;
2198 reclen
= DIRENTSIZE(namelen
);
2200 if (reclen
> uio
->uio_resid
) {
2205 idbp
->d_reclen
= reclen
;
2206 idbp
->d_ino
= pa
->ino
;
2207 idbp
->d_off
= pa
->cook
.o
;
2208 idbp
->d_name
[namelen
] = '\0';
2209 memcpy(idbp
->d_name
, pa
->name
, namelen
);
2210 retval
= uio_read((caddr_t
)idbp
, reclen
, uio
);
2211 pa
->done
= (retval
== 0);