2 * linux/fs/hfsplus/super.c
5 * Brad Boyer (flar@allandria.com)
6 * (C) 2003 Ardis Technologies <roman@ardistech.com>
10 #include <linux/module.h>
11 #include <linux/init.h>
12 #include <linux/pagemap.h>
13 #include <linux/blkdev.h>
15 #include <linux/slab.h>
16 #include <linux/vfs.h>
17 #include <linux/nls.h>
19 static struct inode
*hfsplus_alloc_inode(struct super_block
*sb
);
20 static void hfsplus_destroy_inode(struct inode
*inode
);
22 #include "hfsplus_fs.h"
25 static int hfsplus_system_read_inode(struct inode
*inode
)
27 struct hfsplus_vh
*vhdr
= HFSPLUS_SB(inode
->i_sb
)->s_vhdr
;
29 switch (inode
->i_ino
) {
30 case HFSPLUS_EXT_CNID
:
31 hfsplus_inode_read_fork(inode
, &vhdr
->ext_file
);
32 inode
->i_mapping
->a_ops
= &hfsplus_btree_aops
;
34 case HFSPLUS_CAT_CNID
:
35 hfsplus_inode_read_fork(inode
, &vhdr
->cat_file
);
36 inode
->i_mapping
->a_ops
= &hfsplus_btree_aops
;
38 case HFSPLUS_ALLOC_CNID
:
39 hfsplus_inode_read_fork(inode
, &vhdr
->alloc_file
);
40 inode
->i_mapping
->a_ops
= &hfsplus_aops
;
42 case HFSPLUS_START_CNID
:
43 hfsplus_inode_read_fork(inode
, &vhdr
->start_file
);
45 case HFSPLUS_ATTR_CNID
:
46 hfsplus_inode_read_fork(inode
, &vhdr
->attr_file
);
47 inode
->i_mapping
->a_ops
= &hfsplus_btree_aops
;
56 struct inode
*hfsplus_iget(struct super_block
*sb
, unsigned long ino
)
58 struct hfs_find_data fd
;
62 inode
= iget_locked(sb
, ino
);
64 return ERR_PTR(-ENOMEM
);
65 if (!(inode
->i_state
& I_NEW
))
68 INIT_LIST_HEAD(&HFSPLUS_I(inode
)->open_dir_list
);
69 mutex_init(&HFSPLUS_I(inode
)->extents_lock
);
70 HFSPLUS_I(inode
)->flags
= 0;
71 HFSPLUS_I(inode
)->extent_state
= 0;
72 HFSPLUS_I(inode
)->rsrc_inode
= NULL
;
73 atomic_set(&HFSPLUS_I(inode
)->opencnt
, 0);
75 if (inode
->i_ino
>= HFSPLUS_FIRSTUSER_CNID
||
76 inode
->i_ino
== HFSPLUS_ROOT_CNID
) {
77 err
= hfs_find_init(HFSPLUS_SB(inode
->i_sb
)->cat_tree
, &fd
);
79 err
= hfsplus_find_cat(inode
->i_sb
, inode
->i_ino
, &fd
);
81 err
= hfsplus_cat_read_inode(inode
, &fd
);
85 err
= hfsplus_system_read_inode(inode
);
93 unlock_new_inode(inode
);
97 static int hfsplus_system_write_inode(struct inode
*inode
)
99 struct hfsplus_sb_info
*sbi
= HFSPLUS_SB(inode
->i_sb
);
100 struct hfsplus_vh
*vhdr
= sbi
->s_vhdr
;
101 struct hfsplus_fork_raw
*fork
;
102 struct hfs_btree
*tree
= NULL
;
104 switch (inode
->i_ino
) {
105 case HFSPLUS_EXT_CNID
:
106 fork
= &vhdr
->ext_file
;
107 tree
= sbi
->ext_tree
;
109 case HFSPLUS_CAT_CNID
:
110 fork
= &vhdr
->cat_file
;
111 tree
= sbi
->cat_tree
;
113 case HFSPLUS_ALLOC_CNID
:
114 fork
= &vhdr
->alloc_file
;
116 case HFSPLUS_START_CNID
:
117 fork
= &vhdr
->start_file
;
119 case HFSPLUS_ATTR_CNID
:
120 fork
= &vhdr
->attr_file
;
121 tree
= sbi
->attr_tree
;
127 if (fork
->total_size
!= cpu_to_be64(inode
->i_size
)) {
128 set_bit(HFSPLUS_SB_WRITEBACKUP
, &sbi
->flags
);
129 hfsplus_mark_mdb_dirty(inode
->i_sb
);
131 hfsplus_inode_write_fork(inode
, fork
);
133 int err
= hfs_btree_write(tree
);
135 pr_err("b-tree write err: %d, ino %lu\n",
143 static int hfsplus_write_inode(struct inode
*inode
,
144 struct writeback_control
*wbc
)
148 hfs_dbg(INODE
, "hfsplus_write_inode: %lu\n", inode
->i_ino
);
150 err
= hfsplus_ext_write_extent(inode
);
154 if (inode
->i_ino
>= HFSPLUS_FIRSTUSER_CNID
||
155 inode
->i_ino
== HFSPLUS_ROOT_CNID
)
156 return hfsplus_cat_write_inode(inode
);
158 return hfsplus_system_write_inode(inode
);
161 static void hfsplus_evict_inode(struct inode
*inode
)
163 hfs_dbg(INODE
, "hfsplus_evict_inode: %lu\n", inode
->i_ino
);
164 truncate_inode_pages(&inode
->i_data
, 0);
166 if (HFSPLUS_IS_RSRC(inode
)) {
167 HFSPLUS_I(HFSPLUS_I(inode
)->rsrc_inode
)->rsrc_inode
= NULL
;
168 iput(HFSPLUS_I(inode
)->rsrc_inode
);
172 static int hfsplus_sync_fs(struct super_block
*sb
, int wait
)
174 struct hfsplus_sb_info
*sbi
= HFSPLUS_SB(sb
);
175 struct hfsplus_vh
*vhdr
= sbi
->s_vhdr
;
176 int write_backup
= 0;
182 hfs_dbg(SUPER
, "hfsplus_sync_fs\n");
185 * Explicitly write out the special metadata inodes.
187 * While these special inodes are marked as hashed and written
188 * out peridocically by the flusher threads we redirty them
189 * during writeout of normal inodes, and thus the life lock
190 * prevents us from getting the latest state to disk.
192 error
= filemap_write_and_wait(sbi
->cat_tree
->inode
->i_mapping
);
193 error2
= filemap_write_and_wait(sbi
->ext_tree
->inode
->i_mapping
);
196 if (sbi
->attr_tree
) {
198 filemap_write_and_wait(sbi
->attr_tree
->inode
->i_mapping
);
202 error2
= filemap_write_and_wait(sbi
->alloc_file
->i_mapping
);
206 mutex_lock(&sbi
->vh_mutex
);
207 mutex_lock(&sbi
->alloc_mutex
);
208 vhdr
->free_blocks
= cpu_to_be32(sbi
->free_blocks
);
209 vhdr
->next_cnid
= cpu_to_be32(sbi
->next_cnid
);
210 vhdr
->folder_count
= cpu_to_be32(sbi
->folder_count
);
211 vhdr
->file_count
= cpu_to_be32(sbi
->file_count
);
213 if (test_and_clear_bit(HFSPLUS_SB_WRITEBACKUP
, &sbi
->flags
)) {
214 memcpy(sbi
->s_backup_vhdr
, sbi
->s_vhdr
, sizeof(*sbi
->s_vhdr
));
218 error2
= hfsplus_submit_bio(sb
,
219 sbi
->part_start
+ HFSPLUS_VOLHEAD_SECTOR
,
220 sbi
->s_vhdr_buf
, NULL
, WRITE_SYNC
);
226 error2
= hfsplus_submit_bio(sb
,
227 sbi
->part_start
+ sbi
->sect_count
- 2,
228 sbi
->s_backup_vhdr_buf
, NULL
, WRITE_SYNC
);
232 mutex_unlock(&sbi
->alloc_mutex
);
233 mutex_unlock(&sbi
->vh_mutex
);
235 if (!test_bit(HFSPLUS_SB_NOBARRIER
, &sbi
->flags
))
236 blkdev_issue_flush(sb
->s_bdev
, GFP_KERNEL
, NULL
);
241 static void delayed_sync_fs(struct work_struct
*work
)
244 struct hfsplus_sb_info
*sbi
;
246 sbi
= container_of(work
, struct hfsplus_sb_info
, sync_work
.work
);
248 spin_lock(&sbi
->work_lock
);
249 sbi
->work_queued
= 0;
250 spin_unlock(&sbi
->work_lock
);
252 err
= hfsplus_sync_fs(sbi
->alloc_file
->i_sb
, 1);
254 pr_err("delayed sync fs err %d\n", err
);
257 void hfsplus_mark_mdb_dirty(struct super_block
*sb
)
259 struct hfsplus_sb_info
*sbi
= HFSPLUS_SB(sb
);
262 if (sb
->s_flags
& MS_RDONLY
)
265 spin_lock(&sbi
->work_lock
);
266 if (!sbi
->work_queued
) {
267 delay
= msecs_to_jiffies(dirty_writeback_interval
* 10);
268 queue_delayed_work(system_long_wq
, &sbi
->sync_work
, delay
);
269 sbi
->work_queued
= 1;
271 spin_unlock(&sbi
->work_lock
);
274 static void hfsplus_put_super(struct super_block
*sb
)
276 struct hfsplus_sb_info
*sbi
= HFSPLUS_SB(sb
);
278 hfs_dbg(SUPER
, "hfsplus_put_super\n");
280 cancel_delayed_work_sync(&sbi
->sync_work
);
282 if (!(sb
->s_flags
& MS_RDONLY
) && sbi
->s_vhdr
) {
283 struct hfsplus_vh
*vhdr
= sbi
->s_vhdr
;
285 vhdr
->modify_date
= hfsp_now2mt();
286 vhdr
->attributes
|= cpu_to_be32(HFSPLUS_VOL_UNMNT
);
287 vhdr
->attributes
&= cpu_to_be32(~HFSPLUS_VOL_INCNSTNT
);
289 hfsplus_sync_fs(sb
, 1);
292 hfs_btree_close(sbi
->attr_tree
);
293 hfs_btree_close(sbi
->cat_tree
);
294 hfs_btree_close(sbi
->ext_tree
);
295 iput(sbi
->alloc_file
);
296 iput(sbi
->hidden_dir
);
297 kfree(sbi
->s_vhdr_buf
);
298 kfree(sbi
->s_backup_vhdr_buf
);
299 unload_nls(sbi
->nls
);
300 kfree(sb
->s_fs_info
);
301 sb
->s_fs_info
= NULL
;
304 static int hfsplus_statfs(struct dentry
*dentry
, struct kstatfs
*buf
)
306 struct super_block
*sb
= dentry
->d_sb
;
307 struct hfsplus_sb_info
*sbi
= HFSPLUS_SB(sb
);
308 u64 id
= huge_encode_dev(sb
->s_bdev
->bd_dev
);
310 buf
->f_type
= HFSPLUS_SUPER_MAGIC
;
311 buf
->f_bsize
= sb
->s_blocksize
;
312 buf
->f_blocks
= sbi
->total_blocks
<< sbi
->fs_shift
;
313 buf
->f_bfree
= sbi
->free_blocks
<< sbi
->fs_shift
;
314 buf
->f_bavail
= buf
->f_bfree
;
315 buf
->f_files
= 0xFFFFFFFF;
316 buf
->f_ffree
= 0xFFFFFFFF - sbi
->next_cnid
;
317 buf
->f_fsid
.val
[0] = (u32
)id
;
318 buf
->f_fsid
.val
[1] = (u32
)(id
>> 32);
319 buf
->f_namelen
= HFSPLUS_MAX_STRLEN
;
324 static int hfsplus_remount(struct super_block
*sb
, int *flags
, char *data
)
326 if ((*flags
& MS_RDONLY
) == (sb
->s_flags
& MS_RDONLY
))
328 if (!(*flags
& MS_RDONLY
)) {
329 struct hfsplus_vh
*vhdr
= HFSPLUS_SB(sb
)->s_vhdr
;
332 if (!hfsplus_parse_options_remount(data
, &force
))
335 if (!(vhdr
->attributes
& cpu_to_be32(HFSPLUS_VOL_UNMNT
))) {
336 pr_warn("filesystem was not cleanly unmounted, running fsck.hfsplus is recommended. leaving read-only.\n");
337 sb
->s_flags
|= MS_RDONLY
;
341 } else if (vhdr
->attributes
&
342 cpu_to_be32(HFSPLUS_VOL_SOFTLOCK
)) {
343 pr_warn("filesystem is marked locked, leaving read-only.\n");
344 sb
->s_flags
|= MS_RDONLY
;
346 } else if (vhdr
->attributes
&
347 cpu_to_be32(HFSPLUS_VOL_JOURNALED
)) {
348 pr_warn("filesystem is marked journaled, leaving read-only.\n");
349 sb
->s_flags
|= MS_RDONLY
;
356 static const struct super_operations hfsplus_sops
= {
357 .alloc_inode
= hfsplus_alloc_inode
,
358 .destroy_inode
= hfsplus_destroy_inode
,
359 .write_inode
= hfsplus_write_inode
,
360 .evict_inode
= hfsplus_evict_inode
,
361 .put_super
= hfsplus_put_super
,
362 .sync_fs
= hfsplus_sync_fs
,
363 .statfs
= hfsplus_statfs
,
364 .remount_fs
= hfsplus_remount
,
365 .show_options
= hfsplus_show_options
,
368 static int hfsplus_fill_super(struct super_block
*sb
, void *data
, int silent
)
370 struct hfsplus_vh
*vhdr
;
371 struct hfsplus_sb_info
*sbi
;
372 hfsplus_cat_entry entry
;
373 struct hfs_find_data fd
;
374 struct inode
*root
, *inode
;
376 struct nls_table
*nls
= NULL
;
377 u64 last_fs_block
, last_fs_page
;
381 sbi
= kzalloc(sizeof(*sbi
), GFP_KERNEL
);
386 mutex_init(&sbi
->alloc_mutex
);
387 mutex_init(&sbi
->vh_mutex
);
388 spin_lock_init(&sbi
->work_lock
);
389 INIT_DELAYED_WORK(&sbi
->sync_work
, delayed_sync_fs
);
390 hfsplus_fill_defaults(sbi
);
393 if (!hfsplus_parse_options(data
, sbi
)) {
394 pr_err("unable to parse mount options\n");
398 /* temporarily use utf8 to correctly find the hidden dir below */
400 sbi
->nls
= load_nls("utf8");
402 pr_err("unable to load nls for utf8\n");
406 /* Grab the volume header */
407 if (hfsplus_read_wrapper(sb
)) {
409 pr_warn("unable to find HFS+ superblock\n");
414 /* Copy parts of the volume header into the superblock */
415 sb
->s_magic
= HFSPLUS_VOLHEAD_SIG
;
416 if (be16_to_cpu(vhdr
->version
) < HFSPLUS_MIN_VERSION
||
417 be16_to_cpu(vhdr
->version
) > HFSPLUS_CURRENT_VERSION
) {
418 pr_err("wrong filesystem version\n");
421 sbi
->total_blocks
= be32_to_cpu(vhdr
->total_blocks
);
422 sbi
->free_blocks
= be32_to_cpu(vhdr
->free_blocks
);
423 sbi
->next_cnid
= be32_to_cpu(vhdr
->next_cnid
);
424 sbi
->file_count
= be32_to_cpu(vhdr
->file_count
);
425 sbi
->folder_count
= be32_to_cpu(vhdr
->folder_count
);
426 sbi
->data_clump_blocks
=
427 be32_to_cpu(vhdr
->data_clump_sz
) >> sbi
->alloc_blksz_shift
;
428 if (!sbi
->data_clump_blocks
)
429 sbi
->data_clump_blocks
= 1;
430 sbi
->rsrc_clump_blocks
=
431 be32_to_cpu(vhdr
->rsrc_clump_sz
) >> sbi
->alloc_blksz_shift
;
432 if (!sbi
->rsrc_clump_blocks
)
433 sbi
->rsrc_clump_blocks
= 1;
436 last_fs_block
= sbi
->total_blocks
- 1;
437 last_fs_page
= (last_fs_block
<< sbi
->alloc_blksz_shift
) >>
440 if ((last_fs_block
> (sector_t
)(~0ULL) >> (sbi
->alloc_blksz_shift
- 9)) ||
441 (last_fs_page
> (pgoff_t
)(~0ULL))) {
442 pr_err("filesystem size too large\n");
446 /* Set up operations so we can load metadata */
447 sb
->s_op
= &hfsplus_sops
;
448 sb
->s_maxbytes
= MAX_LFS_FILESIZE
;
450 if (!(vhdr
->attributes
& cpu_to_be32(HFSPLUS_VOL_UNMNT
))) {
451 pr_warn("Filesystem was not cleanly unmounted, running fsck.hfsplus is recommended. mounting read-only.\n");
452 sb
->s_flags
|= MS_RDONLY
;
453 } else if (test_and_clear_bit(HFSPLUS_SB_FORCE
, &sbi
->flags
)) {
455 } else if (vhdr
->attributes
& cpu_to_be32(HFSPLUS_VOL_SOFTLOCK
)) {
456 pr_warn("Filesystem is marked locked, mounting read-only.\n");
457 sb
->s_flags
|= MS_RDONLY
;
458 } else if ((vhdr
->attributes
& cpu_to_be32(HFSPLUS_VOL_JOURNALED
)) &&
459 !(sb
->s_flags
& MS_RDONLY
)) {
460 pr_warn("write access to a journaled filesystem is not supported, use the force option at your own risk, mounting read-only.\n");
461 sb
->s_flags
|= MS_RDONLY
;
466 /* Load metadata objects (B*Trees) */
467 sbi
->ext_tree
= hfs_btree_open(sb
, HFSPLUS_EXT_CNID
);
468 if (!sbi
->ext_tree
) {
469 pr_err("failed to load extents file\n");
472 sbi
->cat_tree
= hfs_btree_open(sb
, HFSPLUS_CAT_CNID
);
473 if (!sbi
->cat_tree
) {
474 pr_err("failed to load catalog file\n");
475 goto out_close_ext_tree
;
477 atomic_set(&sbi
->attr_tree_state
, HFSPLUS_EMPTY_ATTR_TREE
);
478 if (vhdr
->attr_file
.total_blocks
!= 0) {
479 sbi
->attr_tree
= hfs_btree_open(sb
, HFSPLUS_ATTR_CNID
);
480 if (!sbi
->attr_tree
) {
481 pr_err("failed to load attributes file\n");
482 goto out_close_cat_tree
;
484 atomic_set(&sbi
->attr_tree_state
, HFSPLUS_VALID_ATTR_TREE
);
486 sb
->s_xattr
= hfsplus_xattr_handlers
;
488 inode
= hfsplus_iget(sb
, HFSPLUS_ALLOC_CNID
);
490 pr_err("failed to load allocation file\n");
491 err
= PTR_ERR(inode
);
492 goto out_close_attr_tree
;
494 sbi
->alloc_file
= inode
;
496 /* Load the root directory */
497 root
= hfsplus_iget(sb
, HFSPLUS_ROOT_CNID
);
499 pr_err("failed to load root directory\n");
501 goto out_put_alloc_file
;
504 sb
->s_d_op
= &hfsplus_dentry_operations
;
505 sb
->s_root
= d_make_root(root
);
508 goto out_put_alloc_file
;
511 str
.len
= sizeof(HFSP_HIDDENDIR_NAME
) - 1;
512 str
.name
= HFSP_HIDDENDIR_NAME
;
513 err
= hfs_find_init(sbi
->cat_tree
, &fd
);
516 hfsplus_cat_build_key(sb
, fd
.search_key
, HFSPLUS_ROOT_CNID
, &str
);
517 if (!hfs_brec_read(&fd
, &entry
, sizeof(entry
))) {
519 if (entry
.type
!= cpu_to_be16(HFSPLUS_FOLDER
))
521 inode
= hfsplus_iget(sb
, be32_to_cpu(entry
.folder
.id
));
523 err
= PTR_ERR(inode
);
526 sbi
->hidden_dir
= inode
;
530 if (!(sb
->s_flags
& MS_RDONLY
)) {
532 * H+LX == hfsplusutils, H+Lx == this driver, H+lx is unused
533 * all three are registered with Apple for our use
535 vhdr
->last_mount_vers
= cpu_to_be32(HFSP_MOUNT_VERSION
);
536 vhdr
->modify_date
= hfsp_now2mt();
537 be32_add_cpu(&vhdr
->write_count
, 1);
538 vhdr
->attributes
&= cpu_to_be32(~HFSPLUS_VOL_UNMNT
);
539 vhdr
->attributes
|= cpu_to_be32(HFSPLUS_VOL_INCNSTNT
);
540 hfsplus_sync_fs(sb
, 1);
542 if (!sbi
->hidden_dir
) {
543 mutex_lock(&sbi
->vh_mutex
);
544 sbi
->hidden_dir
= hfsplus_new_inode(sb
, S_IFDIR
);
545 if (!sbi
->hidden_dir
) {
546 mutex_unlock(&sbi
->vh_mutex
);
550 err
= hfsplus_create_cat(sbi
->hidden_dir
->i_ino
, root
,
551 &str
, sbi
->hidden_dir
);
553 mutex_unlock(&sbi
->vh_mutex
);
554 goto out_put_hidden_dir
;
557 err
= hfsplus_init_inode_security(sbi
->hidden_dir
,
559 if (err
== -EOPNOTSUPP
)
560 err
= 0; /* Operation is not supported. */
563 * Try to delete anyway without
566 hfsplus_delete_cat(sbi
->hidden_dir
->i_ino
,
568 mutex_unlock(&sbi
->vh_mutex
);
569 goto out_put_hidden_dir
;
572 mutex_unlock(&sbi
->vh_mutex
);
573 hfsplus_mark_inode_dirty(sbi
->hidden_dir
,
574 HFSPLUS_I_CAT_DIRTY
);
578 unload_nls(sbi
->nls
);
583 iput(sbi
->hidden_dir
);
588 iput(sbi
->alloc_file
);
590 hfs_btree_close(sbi
->attr_tree
);
592 hfs_btree_close(sbi
->cat_tree
);
594 hfs_btree_close(sbi
->ext_tree
);
596 kfree(sbi
->s_vhdr_buf
);
597 kfree(sbi
->s_backup_vhdr_buf
);
599 unload_nls(sbi
->nls
);
606 MODULE_AUTHOR("Brad Boyer");
607 MODULE_DESCRIPTION("Extended Macintosh Filesystem");
608 MODULE_LICENSE("GPL");
610 static struct kmem_cache
*hfsplus_inode_cachep
;
612 static struct inode
*hfsplus_alloc_inode(struct super_block
*sb
)
614 struct hfsplus_inode_info
*i
;
616 i
= kmem_cache_alloc(hfsplus_inode_cachep
, GFP_KERNEL
);
617 return i
? &i
->vfs_inode
: NULL
;
620 static void hfsplus_i_callback(struct rcu_head
*head
)
622 struct inode
*inode
= container_of(head
, struct inode
, i_rcu
);
624 kmem_cache_free(hfsplus_inode_cachep
, HFSPLUS_I(inode
));
627 static void hfsplus_destroy_inode(struct inode
*inode
)
629 call_rcu(&inode
->i_rcu
, hfsplus_i_callback
);
632 #define HFSPLUS_INODE_SIZE sizeof(struct hfsplus_inode_info)
634 static struct dentry
*hfsplus_mount(struct file_system_type
*fs_type
,
635 int flags
, const char *dev_name
, void *data
)
637 return mount_bdev(fs_type
, flags
, dev_name
, data
, hfsplus_fill_super
);
640 static struct file_system_type hfsplus_fs_type
= {
641 .owner
= THIS_MODULE
,
643 .mount
= hfsplus_mount
,
644 .kill_sb
= kill_block_super
,
645 .fs_flags
= FS_REQUIRES_DEV
,
647 MODULE_ALIAS_FS("hfsplus");
649 static void hfsplus_init_once(void *p
)
651 struct hfsplus_inode_info
*i
= p
;
653 inode_init_once(&i
->vfs_inode
);
656 static int __init
init_hfsplus_fs(void)
660 hfsplus_inode_cachep
= kmem_cache_create("hfsplus_icache",
661 HFSPLUS_INODE_SIZE
, 0, SLAB_HWCACHE_ALIGN
,
663 if (!hfsplus_inode_cachep
)
665 err
= hfsplus_create_attr_tree_cache();
667 goto destroy_inode_cache
;
668 err
= register_filesystem(&hfsplus_fs_type
);
670 goto destroy_attr_tree_cache
;
673 destroy_attr_tree_cache
:
674 hfsplus_destroy_attr_tree_cache();
677 kmem_cache_destroy(hfsplus_inode_cachep
);
682 static void __exit
exit_hfsplus_fs(void)
684 unregister_filesystem(&hfsplus_fs_type
);
687 * Make sure all delayed rcu free inodes are flushed before we
691 hfsplus_destroy_attr_tree_cache();
692 kmem_cache_destroy(hfsplus_inode_cachep
);
695 module_init(init_hfsplus_fs
)
696 module_exit(exit_hfsplus_fs
)