2 * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved.
3 * Copyright (C) 2004-2007 Red Hat, Inc. All rights reserved.
5 * This copyrighted material is made available to anyone wishing to use,
6 * modify, copy, or redistribute it subject to the terms and conditions
7 * of the GNU General Public License version 2.
10 #include <linux/sched.h>
11 #include <linux/slab.h>
12 #include <linux/spinlock.h>
13 #include <linux/completion.h>
14 #include <linux/buffer_head.h>
15 #include <linux/crc32.h>
16 #include <linux/gfs2_ondisk.h>
17 #include <linux/bio.h>
18 #include <linux/lm_interface.h>
36 static const u32 gfs2_old_fs_formats
[] = {
40 static const u32 gfs2_old_multihost_formats
[] = {
45 * gfs2_tune_init - Fill a gfs2_tune structure with default values
50 void gfs2_tune_init(struct gfs2_tune
*gt
)
52 spin_lock_init(>
->gt_spin
);
54 gt
->gt_demote_secs
= 300;
55 gt
->gt_incore_log_blocks
= 1024;
56 gt
->gt_log_flush_secs
= 60;
57 gt
->gt_recoverd_secs
= 60;
59 gt
->gt_quotad_secs
= 5;
60 gt
->gt_quota_simul_sync
= 64;
61 gt
->gt_quota_warn_period
= 10;
62 gt
->gt_quota_scale_num
= 1;
63 gt
->gt_quota_scale_den
= 1;
64 gt
->gt_quota_cache_secs
= 300;
65 gt
->gt_quota_quantum
= 60;
66 gt
->gt_atime_quantum
= 3600;
67 gt
->gt_new_files_jdata
= 0;
68 gt
->gt_new_files_directio
= 0;
69 gt
->gt_max_readahead
= 1 << 18;
70 gt
->gt_stall_secs
= 600;
71 gt
->gt_complain_secs
= 10;
72 gt
->gt_statfs_quantum
= 30;
73 gt
->gt_statfs_slow
= 0;
77 * gfs2_check_sb - Check superblock
78 * @sdp: the filesystem
80 * @silent: Don't print a message if the check fails
82 * Checks the version code of the FS is one that we understand how to
83 * read and that the sizes of the various on-disk structures have not
87 int gfs2_check_sb(struct gfs2_sbd
*sdp
, struct gfs2_sb_host
*sb
, int silent
)
91 if (sb
->sb_magic
!= GFS2_MAGIC
||
92 sb
->sb_type
!= GFS2_METATYPE_SB
) {
94 printk(KERN_WARNING
"GFS2: not a GFS2 filesystem\n");
98 /* If format numbers match exactly, we're done. */
100 if (sb
->sb_fs_format
== GFS2_FORMAT_FS
&&
101 sb
->sb_multihost_format
== GFS2_FORMAT_MULTI
)
104 if (sb
->sb_fs_format
!= GFS2_FORMAT_FS
) {
105 for (x
= 0; gfs2_old_fs_formats
[x
]; x
++)
106 if (gfs2_old_fs_formats
[x
] == sb
->sb_fs_format
)
109 if (!gfs2_old_fs_formats
[x
]) {
111 "GFS2: code version (%u, %u) is incompatible "
112 "with ondisk format (%u, %u)\n",
113 GFS2_FORMAT_FS
, GFS2_FORMAT_MULTI
,
114 sb
->sb_fs_format
, sb
->sb_multihost_format
);
116 "GFS2: I don't know how to upgrade this FS\n");
121 if (sb
->sb_multihost_format
!= GFS2_FORMAT_MULTI
) {
122 for (x
= 0; gfs2_old_multihost_formats
[x
]; x
++)
123 if (gfs2_old_multihost_formats
[x
] ==
124 sb
->sb_multihost_format
)
127 if (!gfs2_old_multihost_formats
[x
]) {
129 "GFS2: code version (%u, %u) is incompatible "
130 "with ondisk format (%u, %u)\n",
131 GFS2_FORMAT_FS
, GFS2_FORMAT_MULTI
,
132 sb
->sb_fs_format
, sb
->sb_multihost_format
);
134 "GFS2: I don't know how to upgrade this FS\n");
139 if (!sdp
->sd_args
.ar_upgrade
) {
141 "GFS2: code version (%u, %u) is incompatible "
142 "with ondisk format (%u, %u)\n",
143 GFS2_FORMAT_FS
, GFS2_FORMAT_MULTI
,
144 sb
->sb_fs_format
, sb
->sb_multihost_format
);
146 "GFS2: Use the \"upgrade\" mount option to upgrade "
148 printk(KERN_INFO
"GFS2: See the manual for more details\n");
156 static void end_bio_io_page(struct bio
*bio
, int error
)
158 struct page
*page
= bio
->bi_private
;
161 SetPageUptodate(page
);
163 printk(KERN_WARNING
"gfs2: error %d reading superblock\n", error
);
167 static void gfs2_sb_in(struct gfs2_sb_host
*sb
, const void *buf
)
169 const struct gfs2_sb
*str
= buf
;
171 sb
->sb_magic
= be32_to_cpu(str
->sb_header
.mh_magic
);
172 sb
->sb_type
= be32_to_cpu(str
->sb_header
.mh_type
);
173 sb
->sb_format
= be32_to_cpu(str
->sb_header
.mh_format
);
174 sb
->sb_fs_format
= be32_to_cpu(str
->sb_fs_format
);
175 sb
->sb_multihost_format
= be32_to_cpu(str
->sb_multihost_format
);
176 sb
->sb_bsize
= be32_to_cpu(str
->sb_bsize
);
177 sb
->sb_bsize_shift
= be32_to_cpu(str
->sb_bsize_shift
);
178 sb
->sb_master_dir
.no_addr
= be64_to_cpu(str
->sb_master_dir
.no_addr
);
179 sb
->sb_master_dir
.no_formal_ino
= be64_to_cpu(str
->sb_master_dir
.no_formal_ino
);
180 sb
->sb_root_dir
.no_addr
= be64_to_cpu(str
->sb_root_dir
.no_addr
);
181 sb
->sb_root_dir
.no_formal_ino
= be64_to_cpu(str
->sb_root_dir
.no_formal_ino
);
183 memcpy(sb
->sb_lockproto
, str
->sb_lockproto
, GFS2_LOCKNAME_LEN
);
184 memcpy(sb
->sb_locktable
, str
->sb_locktable
, GFS2_LOCKNAME_LEN
);
188 * gfs2_read_super - Read the gfs2 super block from disk
189 * @sdp: The GFS2 super block
190 * @sector: The location of the super block
191 * @error: The error code to return
193 * This uses the bio functions to read the super block from disk
194 * because we want to be 100% sure that we never read cached data.
195 * A super block is read twice only during each GFS2 mount and is
196 * never written to by the filesystem. The first time its read no
197 * locks are held, and the only details which are looked at are those
198 * relating to the locking protocol. Once locking is up and working,
199 * the sb is read again under the lock to establish the location of
200 * the master directory (contains pointers to journals etc) and the
203 * Returns: 0 on success or error
206 int gfs2_read_super(struct gfs2_sbd
*sdp
, sector_t sector
)
208 struct super_block
*sb
= sdp
->sd_vfs
;
213 page
= alloc_page(GFP_NOFS
);
217 ClearPageUptodate(page
);
218 ClearPageDirty(page
);
221 bio
= bio_alloc(GFP_NOFS
, 1);
222 if (unlikely(!bio
)) {
227 bio
->bi_sector
= sector
* (sb
->s_blocksize
>> 9);
228 bio
->bi_bdev
= sb
->s_bdev
;
229 bio_add_page(bio
, page
, PAGE_SIZE
, 0);
231 bio
->bi_end_io
= end_bio_io_page
;
232 bio
->bi_private
= page
;
233 submit_bio(READ_SYNC
| (1 << BIO_RW_META
), bio
);
234 wait_on_page_locked(page
);
236 if (!PageUptodate(page
)) {
241 gfs2_sb_in(&sdp
->sd_sb
, p
);
248 * gfs2_read_sb - Read super block
249 * @sdp: The GFS2 superblock
250 * @gl: the glock for the superblock (assumed to be held)
251 * @silent: Don't print message if mount fails
255 int gfs2_read_sb(struct gfs2_sbd
*sdp
, struct gfs2_glock
*gl
, int silent
)
257 u32 hash_blocks
, ind_blocks
, leaf_blocks
;
262 error
= gfs2_read_super(sdp
, GFS2_SB_ADDR
>> sdp
->sd_fsb2bb_shift
);
265 fs_err(sdp
, "can't read superblock\n");
269 error
= gfs2_check_sb(sdp
, &sdp
->sd_sb
, silent
);
273 sdp
->sd_fsb2bb_shift
= sdp
->sd_sb
.sb_bsize_shift
-
274 GFS2_BASIC_BLOCK_SHIFT
;
275 sdp
->sd_fsb2bb
= 1 << sdp
->sd_fsb2bb_shift
;
276 sdp
->sd_diptrs
= (sdp
->sd_sb
.sb_bsize
-
277 sizeof(struct gfs2_dinode
)) / sizeof(u64
);
278 sdp
->sd_inptrs
= (sdp
->sd_sb
.sb_bsize
-
279 sizeof(struct gfs2_meta_header
)) / sizeof(u64
);
280 sdp
->sd_jbsize
= sdp
->sd_sb
.sb_bsize
- sizeof(struct gfs2_meta_header
);
281 sdp
->sd_hash_bsize
= sdp
->sd_sb
.sb_bsize
/ 2;
282 sdp
->sd_hash_bsize_shift
= sdp
->sd_sb
.sb_bsize_shift
- 1;
283 sdp
->sd_hash_ptrs
= sdp
->sd_hash_bsize
/ sizeof(u64
);
284 sdp
->sd_qc_per_block
= (sdp
->sd_sb
.sb_bsize
-
285 sizeof(struct gfs2_meta_header
)) /
286 sizeof(struct gfs2_quota_change
);
288 /* Compute maximum reservation required to add a entry to a directory */
290 hash_blocks
= DIV_ROUND_UP(sizeof(u64
) * (1 << GFS2_DIR_MAX_DEPTH
),
294 for (tmp_blocks
= hash_blocks
; tmp_blocks
> sdp
->sd_diptrs
;) {
295 tmp_blocks
= DIV_ROUND_UP(tmp_blocks
, sdp
->sd_inptrs
);
296 ind_blocks
+= tmp_blocks
;
299 leaf_blocks
= 2 + GFS2_DIR_MAX_DEPTH
;
301 sdp
->sd_max_dirres
= hash_blocks
+ ind_blocks
+ leaf_blocks
;
303 sdp
->sd_heightsize
[0] = sdp
->sd_sb
.sb_bsize
-
304 sizeof(struct gfs2_dinode
);
305 sdp
->sd_heightsize
[1] = sdp
->sd_sb
.sb_bsize
* sdp
->sd_diptrs
;
310 space
= sdp
->sd_heightsize
[x
- 1] * sdp
->sd_inptrs
;
312 m
= do_div(d
, sdp
->sd_inptrs
);
314 if (d
!= sdp
->sd_heightsize
[x
- 1] || m
)
316 sdp
->sd_heightsize
[x
] = space
;
318 sdp
->sd_max_height
= x
;
319 sdp
->sd_heightsize
[x
] = ~0;
320 gfs2_assert(sdp
, sdp
->sd_max_height
<= GFS2_MAX_META_HEIGHT
);
322 sdp
->sd_jheightsize
[0] = sdp
->sd_sb
.sb_bsize
-
323 sizeof(struct gfs2_dinode
);
324 sdp
->sd_jheightsize
[1] = sdp
->sd_jbsize
* sdp
->sd_diptrs
;
329 space
= sdp
->sd_jheightsize
[x
- 1] * sdp
->sd_inptrs
;
331 m
= do_div(d
, sdp
->sd_inptrs
);
333 if (d
!= sdp
->sd_jheightsize
[x
- 1] || m
)
335 sdp
->sd_jheightsize
[x
] = space
;
337 sdp
->sd_max_jheight
= x
;
338 sdp
->sd_jheightsize
[x
] = ~0;
339 gfs2_assert(sdp
, sdp
->sd_max_jheight
<= GFS2_MAX_META_HEIGHT
);
345 * gfs2_jindex_hold - Grab a lock on the jindex
346 * @sdp: The GFS2 superblock
347 * @ji_gh: the holder for the jindex glock
349 * This is very similar to the gfs2_rindex_hold() function, except that
350 * in general we hold the jindex lock for longer periods of time and
351 * we grab it far less frequently (in general) then the rgrp lock.
356 int gfs2_jindex_hold(struct gfs2_sbd
*sdp
, struct gfs2_holder
*ji_gh
)
358 struct gfs2_inode
*dip
= GFS2_I(sdp
->sd_jindex
);
361 struct gfs2_jdesc
*jd
;
366 mutex_lock(&sdp
->sd_jindex_mutex
);
369 error
= gfs2_glock_nq_init(dip
->i_gl
, LM_ST_SHARED
, 0, ji_gh
);
373 name
.len
= sprintf(buf
, "journal%u", sdp
->sd_journals
);
374 name
.hash
= gfs2_disk_hash(name
.name
, name
.len
);
376 error
= gfs2_dir_check(sdp
->sd_jindex
, &name
, NULL
);
377 if (error
== -ENOENT
) {
382 gfs2_glock_dq_uninit(ji_gh
);
388 jd
= kzalloc(sizeof(struct gfs2_jdesc
), GFP_KERNEL
);
392 INIT_LIST_HEAD(&jd
->extent_list
);
393 jd
->jd_inode
= gfs2_lookupi(sdp
->sd_jindex
, &name
, 1, NULL
);
394 if (!jd
->jd_inode
|| IS_ERR(jd
->jd_inode
)) {
398 error
= PTR_ERR(jd
->jd_inode
);
403 spin_lock(&sdp
->sd_jindex_spin
);
404 jd
->jd_jid
= sdp
->sd_journals
++;
405 list_add_tail(&jd
->jd_list
, &sdp
->sd_jindex_list
);
406 spin_unlock(&sdp
->sd_jindex_spin
);
409 mutex_unlock(&sdp
->sd_jindex_mutex
);
415 * gfs2_jindex_free - Clear all the journal index information
416 * @sdp: The GFS2 superblock
420 void gfs2_jindex_free(struct gfs2_sbd
*sdp
)
422 struct list_head list
, *head
;
423 struct gfs2_jdesc
*jd
;
424 struct gfs2_journal_extent
*jext
;
426 spin_lock(&sdp
->sd_jindex_spin
);
427 list_add(&list
, &sdp
->sd_jindex_list
);
428 list_del_init(&sdp
->sd_jindex_list
);
429 sdp
->sd_journals
= 0;
430 spin_unlock(&sdp
->sd_jindex_spin
);
432 while (!list_empty(&list
)) {
433 jd
= list_entry(list
.next
, struct gfs2_jdesc
, jd_list
);
434 head
= &jd
->extent_list
;
435 while (!list_empty(head
)) {
436 jext
= list_entry(head
->next
,
437 struct gfs2_journal_extent
,
439 list_del(&jext
->extent_list
);
442 list_del(&jd
->jd_list
);
448 static struct gfs2_jdesc
*jdesc_find_i(struct list_head
*head
, unsigned int jid
)
450 struct gfs2_jdesc
*jd
;
453 list_for_each_entry(jd
, head
, jd_list
) {
454 if (jd
->jd_jid
== jid
) {
466 struct gfs2_jdesc
*gfs2_jdesc_find(struct gfs2_sbd
*sdp
, unsigned int jid
)
468 struct gfs2_jdesc
*jd
;
470 spin_lock(&sdp
->sd_jindex_spin
);
471 jd
= jdesc_find_i(&sdp
->sd_jindex_list
, jid
);
472 spin_unlock(&sdp
->sd_jindex_spin
);
477 void gfs2_jdesc_make_dirty(struct gfs2_sbd
*sdp
, unsigned int jid
)
479 struct gfs2_jdesc
*jd
;
481 spin_lock(&sdp
->sd_jindex_spin
);
482 jd
= jdesc_find_i(&sdp
->sd_jindex_list
, jid
);
485 spin_unlock(&sdp
->sd_jindex_spin
);
488 struct gfs2_jdesc
*gfs2_jdesc_find_dirty(struct gfs2_sbd
*sdp
)
490 struct gfs2_jdesc
*jd
;
493 spin_lock(&sdp
->sd_jindex_spin
);
495 list_for_each_entry(jd
, &sdp
->sd_jindex_list
, jd_list
) {
502 spin_unlock(&sdp
->sd_jindex_spin
);
510 int gfs2_jdesc_check(struct gfs2_jdesc
*jd
)
512 struct gfs2_inode
*ip
= GFS2_I(jd
->jd_inode
);
513 struct gfs2_sbd
*sdp
= GFS2_SB(jd
->jd_inode
);
517 if (ip
->i_di
.di_size
< (8 << 20) || ip
->i_di
.di_size
> (1 << 30) ||
518 (ip
->i_di
.di_size
& (sdp
->sd_sb
.sb_bsize
- 1))) {
519 gfs2_consist_inode(ip
);
522 jd
->jd_blocks
= ip
->i_di
.di_size
>> sdp
->sd_sb
.sb_bsize_shift
;
524 error
= gfs2_write_alloc_required(ip
, 0, ip
->i_di
.di_size
, &ar
);
526 gfs2_consist_inode(ip
);
534 * gfs2_make_fs_rw - Turn a Read-Only FS into a Read-Write one
535 * @sdp: the filesystem
540 int gfs2_make_fs_rw(struct gfs2_sbd
*sdp
)
542 struct gfs2_inode
*ip
= GFS2_I(sdp
->sd_jdesc
->jd_inode
);
543 struct gfs2_glock
*j_gl
= ip
->i_gl
;
544 struct gfs2_holder t_gh
;
545 struct gfs2_log_header_host head
;
548 error
= gfs2_glock_nq_init(sdp
->sd_trans_gl
, LM_ST_SHARED
, 0, &t_gh
);
552 j_gl
->gl_ops
->go_inval(j_gl
, DIO_METADATA
);
554 error
= gfs2_find_jhead(sdp
->sd_jdesc
, &head
);
558 if (!(head
.lh_flags
& GFS2_LOG_HEAD_UNMOUNT
)) {
564 /* Initialize some head of the log stuff */
565 sdp
->sd_log_sequence
= head
.lh_sequence
+ 1;
566 gfs2_log_pointers_init(sdp
, head
.lh_blkno
);
568 error
= gfs2_quota_init(sdp
);
572 set_bit(SDF_JOURNAL_LIVE
, &sdp
->sd_flags
);
574 gfs2_glock_dq_uninit(&t_gh
);
579 t_gh
.gh_flags
|= GL_NOCACHE
;
580 gfs2_glock_dq_uninit(&t_gh
);
586 * gfs2_make_fs_ro - Turn a Read-Write FS into a Read-Only one
587 * @sdp: the filesystem
592 int gfs2_make_fs_ro(struct gfs2_sbd
*sdp
)
594 struct gfs2_holder t_gh
;
597 gfs2_quota_sync(sdp
);
598 gfs2_statfs_sync(sdp
);
600 error
= gfs2_glock_nq_init(sdp
->sd_trans_gl
, LM_ST_SHARED
, GL_NOCACHE
,
602 if (error
&& !test_bit(SDF_SHUTDOWN
, &sdp
->sd_flags
))
605 gfs2_meta_syncfs(sdp
);
606 gfs2_log_shutdown(sdp
);
608 clear_bit(SDF_JOURNAL_LIVE
, &sdp
->sd_flags
);
611 gfs2_glock_dq_uninit(&t_gh
);
613 gfs2_quota_cleanup(sdp
);
618 static void gfs2_statfs_change_in(struct gfs2_statfs_change_host
*sc
, const void *buf
)
620 const struct gfs2_statfs_change
*str
= buf
;
622 sc
->sc_total
= be64_to_cpu(str
->sc_total
);
623 sc
->sc_free
= be64_to_cpu(str
->sc_free
);
624 sc
->sc_dinodes
= be64_to_cpu(str
->sc_dinodes
);
627 static void gfs2_statfs_change_out(const struct gfs2_statfs_change_host
*sc
, void *buf
)
629 struct gfs2_statfs_change
*str
= buf
;
631 str
->sc_total
= cpu_to_be64(sc
->sc_total
);
632 str
->sc_free
= cpu_to_be64(sc
->sc_free
);
633 str
->sc_dinodes
= cpu_to_be64(sc
->sc_dinodes
);
636 int gfs2_statfs_init(struct gfs2_sbd
*sdp
)
638 struct gfs2_inode
*m_ip
= GFS2_I(sdp
->sd_statfs_inode
);
639 struct gfs2_statfs_change_host
*m_sc
= &sdp
->sd_statfs_master
;
640 struct gfs2_inode
*l_ip
= GFS2_I(sdp
->sd_sc_inode
);
641 struct gfs2_statfs_change_host
*l_sc
= &sdp
->sd_statfs_local
;
642 struct buffer_head
*m_bh
, *l_bh
;
643 struct gfs2_holder gh
;
646 error
= gfs2_glock_nq_init(m_ip
->i_gl
, LM_ST_EXCLUSIVE
, GL_NOCACHE
,
651 error
= gfs2_meta_inode_buffer(m_ip
, &m_bh
);
655 if (sdp
->sd_args
.ar_spectator
) {
656 spin_lock(&sdp
->sd_statfs_spin
);
657 gfs2_statfs_change_in(m_sc
, m_bh
->b_data
+
658 sizeof(struct gfs2_dinode
));
659 spin_unlock(&sdp
->sd_statfs_spin
);
661 error
= gfs2_meta_inode_buffer(l_ip
, &l_bh
);
665 spin_lock(&sdp
->sd_statfs_spin
);
666 gfs2_statfs_change_in(m_sc
, m_bh
->b_data
+
667 sizeof(struct gfs2_dinode
));
668 gfs2_statfs_change_in(l_sc
, l_bh
->b_data
+
669 sizeof(struct gfs2_dinode
));
670 spin_unlock(&sdp
->sd_statfs_spin
);
678 gfs2_glock_dq_uninit(&gh
);
682 void gfs2_statfs_change(struct gfs2_sbd
*sdp
, s64 total
, s64 free
,
685 struct gfs2_inode
*l_ip
= GFS2_I(sdp
->sd_sc_inode
);
686 struct gfs2_statfs_change_host
*l_sc
= &sdp
->sd_statfs_local
;
687 struct buffer_head
*l_bh
;
690 error
= gfs2_meta_inode_buffer(l_ip
, &l_bh
);
694 gfs2_trans_add_bh(l_ip
->i_gl
, l_bh
, 1);
696 spin_lock(&sdp
->sd_statfs_spin
);
697 l_sc
->sc_total
+= total
;
698 l_sc
->sc_free
+= free
;
699 l_sc
->sc_dinodes
+= dinodes
;
700 gfs2_statfs_change_out(l_sc
, l_bh
->b_data
+ sizeof(struct gfs2_dinode
));
701 spin_unlock(&sdp
->sd_statfs_spin
);
706 int gfs2_statfs_sync(struct gfs2_sbd
*sdp
)
708 struct gfs2_inode
*m_ip
= GFS2_I(sdp
->sd_statfs_inode
);
709 struct gfs2_inode
*l_ip
= GFS2_I(sdp
->sd_sc_inode
);
710 struct gfs2_statfs_change_host
*m_sc
= &sdp
->sd_statfs_master
;
711 struct gfs2_statfs_change_host
*l_sc
= &sdp
->sd_statfs_local
;
712 struct gfs2_holder gh
;
713 struct buffer_head
*m_bh
, *l_bh
;
716 error
= gfs2_glock_nq_init(m_ip
->i_gl
, LM_ST_EXCLUSIVE
, GL_NOCACHE
,
721 error
= gfs2_meta_inode_buffer(m_ip
, &m_bh
);
725 spin_lock(&sdp
->sd_statfs_spin
);
726 gfs2_statfs_change_in(m_sc
, m_bh
->b_data
+
727 sizeof(struct gfs2_dinode
));
728 if (!l_sc
->sc_total
&& !l_sc
->sc_free
&& !l_sc
->sc_dinodes
) {
729 spin_unlock(&sdp
->sd_statfs_spin
);
732 spin_unlock(&sdp
->sd_statfs_spin
);
734 error
= gfs2_meta_inode_buffer(l_ip
, &l_bh
);
738 error
= gfs2_trans_begin(sdp
, 2 * RES_DINODE
, 0);
742 gfs2_trans_add_bh(l_ip
->i_gl
, l_bh
, 1);
744 spin_lock(&sdp
->sd_statfs_spin
);
745 m_sc
->sc_total
+= l_sc
->sc_total
;
746 m_sc
->sc_free
+= l_sc
->sc_free
;
747 m_sc
->sc_dinodes
+= l_sc
->sc_dinodes
;
748 memset(l_sc
, 0, sizeof(struct gfs2_statfs_change
));
749 memset(l_bh
->b_data
+ sizeof(struct gfs2_dinode
),
750 0, sizeof(struct gfs2_statfs_change
));
751 spin_unlock(&sdp
->sd_statfs_spin
);
753 gfs2_trans_add_bh(m_ip
->i_gl
, m_bh
, 1);
754 gfs2_statfs_change_out(m_sc
, m_bh
->b_data
+ sizeof(struct gfs2_dinode
));
763 gfs2_glock_dq_uninit(&gh
);
768 * gfs2_statfs_i - Do a statfs
769 * @sdp: the filesystem
770 * @sg: the sg structure
775 int gfs2_statfs_i(struct gfs2_sbd
*sdp
, struct gfs2_statfs_change_host
*sc
)
777 struct gfs2_statfs_change_host
*m_sc
= &sdp
->sd_statfs_master
;
778 struct gfs2_statfs_change_host
*l_sc
= &sdp
->sd_statfs_local
;
780 spin_lock(&sdp
->sd_statfs_spin
);
783 sc
->sc_total
+= l_sc
->sc_total
;
784 sc
->sc_free
+= l_sc
->sc_free
;
785 sc
->sc_dinodes
+= l_sc
->sc_dinodes
;
787 spin_unlock(&sdp
->sd_statfs_spin
);
791 if (sc
->sc_free
> sc
->sc_total
)
792 sc
->sc_free
= sc
->sc_total
;
793 if (sc
->sc_dinodes
< 0)
800 * statfs_fill - fill in the sg for a given RG
802 * @sc: the sc structure
804 * Returns: 0 on success, -ESTALE if the LVB is invalid
807 static int statfs_slow_fill(struct gfs2_rgrpd
*rgd
,
808 struct gfs2_statfs_change_host
*sc
)
810 gfs2_rgrp_verify(rgd
);
811 sc
->sc_total
+= rgd
->rd_data
;
812 sc
->sc_free
+= rgd
->rd_rg
.rg_free
;
813 sc
->sc_dinodes
+= rgd
->rd_rg
.rg_dinodes
;
818 * gfs2_statfs_slow - Stat a filesystem using asynchronous locking
819 * @sdp: the filesystem
820 * @sc: the sc info that will be returned
822 * Any error (other than a signal) will cause this routine to fall back
823 * to the synchronous version.
825 * FIXME: This really shouldn't busy wait like this.
830 int gfs2_statfs_slow(struct gfs2_sbd
*sdp
, struct gfs2_statfs_change_host
*sc
)
832 struct gfs2_holder ri_gh
;
833 struct gfs2_rgrpd
*rgd_next
;
834 struct gfs2_holder
*gha
, *gh
;
835 unsigned int slots
= 64;
840 memset(sc
, 0, sizeof(struct gfs2_statfs_change_host
));
841 gha
= kcalloc(slots
, sizeof(struct gfs2_holder
), GFP_KERNEL
);
845 error
= gfs2_rindex_hold(sdp
, &ri_gh
);
849 rgd_next
= gfs2_rgrpd_get_first(sdp
);
854 for (x
= 0; x
< slots
; x
++) {
857 if (gh
->gh_gl
&& gfs2_glock_poll(gh
)) {
858 err
= gfs2_glock_wait(gh
);
860 gfs2_holder_uninit(gh
);
864 error
= statfs_slow_fill(
865 gh
->gh_gl
->gl_object
, sc
);
866 gfs2_glock_dq_uninit(gh
);
872 else if (rgd_next
&& !error
) {
873 error
= gfs2_glock_nq_init(rgd_next
->rd_gl
,
877 rgd_next
= gfs2_rgrpd_get_next(rgd_next
);
881 if (signal_pending(current
))
882 error
= -ERESTARTSYS
;
891 gfs2_glock_dq_uninit(&ri_gh
);
899 struct list_head list
;
900 struct gfs2_holder gh
;
904 * gfs2_lock_fs_check_clean - Stop all writes to the FS and check that all
906 * @sdp: the file system
907 * @state: the state to put the transaction lock into
908 * @t_gh: the hold on the transaction lock
913 static int gfs2_lock_fs_check_clean(struct gfs2_sbd
*sdp
,
914 struct gfs2_holder
*t_gh
)
916 struct gfs2_inode
*ip
;
917 struct gfs2_holder ji_gh
;
918 struct gfs2_jdesc
*jd
;
921 struct gfs2_log_header_host lh
;
924 error
= gfs2_jindex_hold(sdp
, &ji_gh
);
928 list_for_each_entry(jd
, &sdp
->sd_jindex_list
, jd_list
) {
929 lfcc
= kmalloc(sizeof(struct lfcc
), GFP_KERNEL
);
934 ip
= GFS2_I(jd
->jd_inode
);
935 error
= gfs2_glock_nq_init(ip
->i_gl
, LM_ST_SHARED
, 0, &lfcc
->gh
);
940 list_add(&lfcc
->list
, &list
);
943 error
= gfs2_glock_nq_init(sdp
->sd_trans_gl
, LM_ST_DEFERRED
,
944 LM_FLAG_PRIORITY
| GL_NOCACHE
,
947 list_for_each_entry(jd
, &sdp
->sd_jindex_list
, jd_list
) {
948 error
= gfs2_jdesc_check(jd
);
951 error
= gfs2_find_jhead(jd
, &lh
);
954 if (!(lh
.lh_flags
& GFS2_LOG_HEAD_UNMOUNT
)) {
961 gfs2_glock_dq_uninit(t_gh
);
964 while (!list_empty(&list
)) {
965 lfcc
= list_entry(list
.next
, struct lfcc
, list
);
966 list_del(&lfcc
->list
);
967 gfs2_glock_dq_uninit(&lfcc
->gh
);
970 gfs2_glock_dq_uninit(&ji_gh
);
975 * gfs2_freeze_fs - freezes the file system
976 * @sdp: the file system
978 * This function flushes data and meta data for all machines by
979 * aquiring the transaction log exclusively. All journals are
980 * ensured to be in a clean state as well.
985 int gfs2_freeze_fs(struct gfs2_sbd
*sdp
)
989 mutex_lock(&sdp
->sd_freeze_lock
);
991 if (!sdp
->sd_freeze_count
++) {
992 error
= gfs2_lock_fs_check_clean(sdp
, &sdp
->sd_freeze_gh
);
994 sdp
->sd_freeze_count
--;
997 mutex_unlock(&sdp
->sd_freeze_lock
);
1003 * gfs2_unfreeze_fs - unfreezes the file system
1004 * @sdp: the file system
1006 * This function allows the file system to proceed by unlocking
1007 * the exclusively held transaction lock. Other GFS2 nodes are
1008 * now free to acquire the lock shared and go on with their lives.
1012 void gfs2_unfreeze_fs(struct gfs2_sbd
*sdp
)
1014 mutex_lock(&sdp
->sd_freeze_lock
);
1016 if (sdp
->sd_freeze_count
&& !--sdp
->sd_freeze_count
)
1017 gfs2_glock_dq_uninit(&sdp
->sd_freeze_gh
);
1019 mutex_unlock(&sdp
->sd_freeze_lock
);