2 * mdt.c - meta data file for NILFS
4 * Copyright (C) 2005-2008 Nippon Telegraph and Telephone Corporation.
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License as published by
8 * the Free Software Foundation; either version 2 of the License, or
9 * (at your option) any later version.
11 * This program is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 * GNU General Public License for more details.
16 * You should have received a copy of the GNU General Public License
17 * along with this program; if not, write to the Free Software
18 * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
20 * Written by Ryusuke Konishi <ryusuke@osrg.net>
23 #include <linux/buffer_head.h>
24 #include <linux/mpage.h>
26 #include <linux/writeback.h>
27 #include <linux/backing-dev.h>
28 #include <linux/swap.h>
29 #include <linux/slab.h>
37 #define NILFS_MDT_MAX_RA_BLOCKS (16 - 1)
41 nilfs_mdt_insert_new_block(struct inode
*inode
, unsigned long block
,
42 struct buffer_head
*bh
,
43 void (*init_block
)(struct inode
*,
44 struct buffer_head
*, void *))
46 struct nilfs_inode_info
*ii
= NILFS_I(inode
);
50 /* Caller exclude read accesses using page lock */
52 /* set_buffer_new(bh); */
55 ret
= nilfs_bmap_insert(ii
->i_bmap
, block
, (unsigned long)bh
);
59 set_buffer_mapped(bh
);
61 kaddr
= kmap_atomic(bh
->b_page
, KM_USER0
);
62 memset(kaddr
+ bh_offset(bh
), 0, 1 << inode
->i_blkbits
);
64 init_block(inode
, bh
, kaddr
);
65 flush_dcache_page(bh
->b_page
);
66 kunmap_atomic(kaddr
, KM_USER0
);
68 set_buffer_uptodate(bh
);
69 nilfs_mark_buffer_dirty(bh
);
70 nilfs_mdt_mark_dirty(inode
);
74 static int nilfs_mdt_create_block(struct inode
*inode
, unsigned long block
,
75 struct buffer_head
**out_bh
,
76 void (*init_block
)(struct inode
*,
80 struct super_block
*sb
= inode
->i_sb
;
81 struct nilfs_transaction_info ti
;
82 struct buffer_head
*bh
;
85 nilfs_transaction_begin(sb
, &ti
, 0);
88 bh
= nilfs_grab_buffer(inode
, inode
->i_mapping
, block
, 0);
93 if (buffer_uptodate(bh
))
97 if (buffer_uptodate(bh
))
100 bh
->b_bdev
= sb
->s_bdev
;
101 err
= nilfs_mdt_insert_new_block(inode
, block
, bh
, init_block
);
108 unlock_page(bh
->b_page
);
109 page_cache_release(bh
->b_page
);
114 err
= nilfs_transaction_commit(sb
);
116 nilfs_transaction_abort(sb
);
122 nilfs_mdt_submit_block(struct inode
*inode
, unsigned long blkoff
,
123 int mode
, struct buffer_head
**out_bh
)
125 struct buffer_head
*bh
;
129 bh
= nilfs_grab_buffer(inode
, inode
->i_mapping
, blkoff
, 0);
133 ret
= -EEXIST
; /* internal code */
134 if (buffer_uptodate(bh
))
138 if (!trylock_buffer(bh
)) {
142 } else /* mode == READ */
145 if (buffer_uptodate(bh
)) {
150 ret
= nilfs_bmap_lookup(NILFS_I(inode
)->i_bmap
, blkoff
, &blknum
);
155 map_bh(bh
, inode
->i_sb
, (sector_t
)blknum
);
157 bh
->b_end_io
= end_buffer_read_sync
;
166 unlock_page(bh
->b_page
);
167 page_cache_release(bh
->b_page
);
173 static int nilfs_mdt_read_block(struct inode
*inode
, unsigned long block
,
174 int readahead
, struct buffer_head
**out_bh
)
176 struct buffer_head
*first_bh
, *bh
;
177 unsigned long blkoff
;
178 int i
, nr_ra_blocks
= NILFS_MDT_MAX_RA_BLOCKS
;
181 err
= nilfs_mdt_submit_block(inode
, block
, READ
, &first_bh
);
182 if (err
== -EEXIST
) /* internal code */
190 for (i
= 0; i
< nr_ra_blocks
; i
++, blkoff
++) {
191 err
= nilfs_mdt_submit_block(inode
, blkoff
, READA
, &bh
);
192 if (likely(!err
|| err
== -EEXIST
))
194 else if (err
!= -EBUSY
)
196 /* abort readahead if bmap lookup failed */
197 if (!buffer_locked(first_bh
))
202 wait_on_buffer(first_bh
);
206 if (!buffer_uptodate(first_bh
))
219 * nilfs_mdt_get_block - read or create a buffer on meta data file.
220 * @inode: inode of the meta data file
221 * @blkoff: block offset
222 * @create: create flag
223 * @init_block: initializer used for newly allocated block
224 * @out_bh: output of a pointer to the buffer_head
226 * nilfs_mdt_get_block() looks up the specified buffer and tries to create
227 * a new buffer if @create is not zero. On success, the returned buffer is
228 * assured to be either existing or formatted using a buffer lock on success.
229 * @out_bh is substituted only when zero is returned.
231 * Return Value: On success, it returns 0. On error, the following negative
232 * error code is returned.
234 * %-ENOMEM - Insufficient memory available.
238 * %-ENOENT - the specified block does not exist (hole block)
240 * %-EINVAL - bmap is broken. (the caller should call nilfs_error())
242 * %-EROFS - Read only filesystem (for create mode)
244 int nilfs_mdt_get_block(struct inode
*inode
, unsigned long blkoff
, int create
,
245 void (*init_block
)(struct inode
*,
246 struct buffer_head
*, void *),
247 struct buffer_head
**out_bh
)
251 /* Should be rewritten with merging nilfs_mdt_read_block() */
253 ret
= nilfs_mdt_read_block(inode
, blkoff
, !create
, out_bh
);
254 if (!create
|| ret
!= -ENOENT
)
257 ret
= nilfs_mdt_create_block(inode
, blkoff
, out_bh
, init_block
);
258 if (unlikely(ret
== -EEXIST
)) {
259 /* create = 0; */ /* limit read-create loop retries */
266 * nilfs_mdt_delete_block - make a hole on the meta data file.
267 * @inode: inode of the meta data file
268 * @block: block offset
270 * Return Value: On success, zero is returned.
271 * On error, one of the following negative error code is returned.
273 * %-ENOMEM - Insufficient memory available.
277 * %-EINVAL - bmap is broken. (the caller should call nilfs_error())
279 int nilfs_mdt_delete_block(struct inode
*inode
, unsigned long block
)
281 struct nilfs_inode_info
*ii
= NILFS_I(inode
);
284 err
= nilfs_bmap_delete(ii
->i_bmap
, block
);
285 if (!err
|| err
== -ENOENT
) {
286 nilfs_mdt_mark_dirty(inode
);
287 nilfs_mdt_forget_block(inode
, block
);
293 * nilfs_mdt_forget_block - discard dirty state and try to remove the page
294 * @inode: inode of the meta data file
295 * @block: block offset
297 * nilfs_mdt_forget_block() clears a dirty flag of the specified buffer, and
298 * tries to release the page including the buffer from a page cache.
300 * Return Value: On success, 0 is returned. On error, one of the following
301 * negative error code is returned.
303 * %-EBUSY - page has an active buffer.
305 * %-ENOENT - page cache has no page addressed by the offset.
307 int nilfs_mdt_forget_block(struct inode
*inode
, unsigned long block
)
309 pgoff_t index
= (pgoff_t
)block
>>
310 (PAGE_CACHE_SHIFT
- inode
->i_blkbits
);
312 unsigned long first_block
;
316 page
= find_lock_page(inode
->i_mapping
, index
);
320 wait_on_page_writeback(page
);
322 first_block
= (unsigned long)index
<<
323 (PAGE_CACHE_SHIFT
- inode
->i_blkbits
);
324 if (page_has_buffers(page
)) {
325 struct buffer_head
*bh
;
327 bh
= nilfs_page_get_nth_block(page
, block
- first_block
);
328 nilfs_forget_buffer(bh
);
330 still_dirty
= PageDirty(page
);
332 page_cache_release(page
);
335 invalidate_inode_pages2_range(inode
->i_mapping
, index
, index
) != 0)
341 * nilfs_mdt_mark_block_dirty - mark a block on the meta data file dirty.
342 * @inode: inode of the meta data file
343 * @block: block offset
345 * Return Value: On success, it returns 0. On error, the following negative
346 * error code is returned.
348 * %-ENOMEM - Insufficient memory available.
352 * %-ENOENT - the specified block does not exist (hole block)
354 * %-EINVAL - bmap is broken. (the caller should call nilfs_error())
356 int nilfs_mdt_mark_block_dirty(struct inode
*inode
, unsigned long block
)
358 struct buffer_head
*bh
;
361 err
= nilfs_mdt_read_block(inode
, block
, 0, &bh
);
364 nilfs_mark_buffer_dirty(bh
);
365 nilfs_mdt_mark_dirty(inode
);
370 int nilfs_mdt_fetch_dirty(struct inode
*inode
)
372 struct nilfs_inode_info
*ii
= NILFS_I(inode
);
374 if (nilfs_bmap_test_and_clear_dirty(ii
->i_bmap
)) {
375 set_bit(NILFS_I_DIRTY
, &ii
->i_state
);
378 return test_bit(NILFS_I_DIRTY
, &ii
->i_state
);
382 nilfs_mdt_write_page(struct page
*page
, struct writeback_control
*wbc
)
385 struct super_block
*sb
;
388 redirty_page_for_writepage(wbc
, page
);
391 inode
= page
->mapping
->host
;
397 if (wbc
->sync_mode
== WB_SYNC_ALL
)
398 err
= nilfs_construct_segment(sb
);
399 else if (wbc
->for_reclaim
)
400 nilfs_flush_segment(sb
, inode
->i_ino
);
406 static const struct address_space_operations def_mdt_aops
= {
407 .writepage
= nilfs_mdt_write_page
,
408 .sync_page
= block_sync_page
,
411 static const struct inode_operations def_mdt_iops
;
412 static const struct file_operations def_mdt_fops
;
415 int nilfs_mdt_init(struct inode
*inode
, gfp_t gfp_mask
, size_t objsz
)
417 struct nilfs_mdt_info
*mi
;
419 mi
= kzalloc(max(sizeof(*mi
), objsz
), GFP_NOFS
);
423 init_rwsem(&mi
->mi_sem
);
424 inode
->i_private
= mi
;
426 inode
->i_mode
= S_IFREG
;
427 mapping_set_gfp_mask(inode
->i_mapping
, gfp_mask
);
428 inode
->i_mapping
->backing_dev_info
= inode
->i_sb
->s_bdi
;
430 inode
->i_op
= &def_mdt_iops
;
431 inode
->i_fop
= &def_mdt_fops
;
432 inode
->i_mapping
->a_ops
= &def_mdt_aops
;
437 void nilfs_mdt_set_entry_size(struct inode
*inode
, unsigned entry_size
,
438 unsigned header_size
)
440 struct nilfs_mdt_info
*mi
= NILFS_MDT(inode
);
442 mi
->mi_entry_size
= entry_size
;
443 mi
->mi_entries_per_block
= (1 << inode
->i_blkbits
) / entry_size
;
444 mi
->mi_first_entry_offset
= DIV_ROUND_UP(header_size
, entry_size
);
447 static const struct address_space_operations shadow_map_aops
= {
448 .sync_page
= block_sync_page
,
452 * nilfs_mdt_setup_shadow_map - setup shadow map and bind it to metadata file
453 * @inode: inode of the metadata file
454 * @shadow: shadow mapping
456 int nilfs_mdt_setup_shadow_map(struct inode
*inode
,
457 struct nilfs_shadow_map
*shadow
)
459 struct nilfs_mdt_info
*mi
= NILFS_MDT(inode
);
460 struct backing_dev_info
*bdi
= inode
->i_sb
->s_bdi
;
462 INIT_LIST_HEAD(&shadow
->frozen_buffers
);
463 nilfs_mapping_init_once(&shadow
->frozen_data
);
464 nilfs_mapping_init(&shadow
->frozen_data
, bdi
, &shadow_map_aops
);
465 nilfs_mapping_init_once(&shadow
->frozen_btnodes
);
466 nilfs_mapping_init(&shadow
->frozen_btnodes
, bdi
, &shadow_map_aops
);
467 mi
->mi_shadow
= shadow
;
472 * nilfs_mdt_save_to_shadow_map - copy bmap and dirty pages to shadow map
473 * @inode: inode of the metadata file
475 int nilfs_mdt_save_to_shadow_map(struct inode
*inode
)
477 struct nilfs_mdt_info
*mi
= NILFS_MDT(inode
);
478 struct nilfs_inode_info
*ii
= NILFS_I(inode
);
479 struct nilfs_shadow_map
*shadow
= mi
->mi_shadow
;
482 ret
= nilfs_copy_dirty_pages(&shadow
->frozen_data
, inode
->i_mapping
);
486 ret
= nilfs_copy_dirty_pages(&shadow
->frozen_btnodes
,
487 &ii
->i_btnode_cache
);
491 nilfs_bmap_save(ii
->i_bmap
, &shadow
->bmap_store
);
496 int nilfs_mdt_freeze_buffer(struct inode
*inode
, struct buffer_head
*bh
)
498 struct nilfs_shadow_map
*shadow
= NILFS_MDT(inode
)->mi_shadow
;
499 struct buffer_head
*bh_frozen
;
501 int blkbits
= inode
->i_blkbits
;
504 page
= grab_cache_page(&shadow
->frozen_data
, bh
->b_page
->index
);
508 if (!page_has_buffers(page
))
509 create_empty_buffers(page
, 1 << blkbits
, 0);
511 bh_frozen
= nilfs_page_get_nth_block(page
, bh_offset(bh
) >> blkbits
);
513 if (!buffer_uptodate(bh_frozen
))
514 nilfs_copy_buffer(bh_frozen
, bh
);
515 if (list_empty(&bh_frozen
->b_assoc_buffers
)) {
516 list_add_tail(&bh_frozen
->b_assoc_buffers
,
517 &shadow
->frozen_buffers
);
518 set_buffer_nilfs_redirected(bh
);
520 brelse(bh_frozen
); /* already frozen */
525 page_cache_release(page
);
530 nilfs_mdt_get_frozen_buffer(struct inode
*inode
, struct buffer_head
*bh
)
532 struct nilfs_shadow_map
*shadow
= NILFS_MDT(inode
)->mi_shadow
;
533 struct buffer_head
*bh_frozen
= NULL
;
537 page
= find_lock_page(&shadow
->frozen_data
, bh
->b_page
->index
);
539 if (page_has_buffers(page
)) {
540 n
= bh_offset(bh
) >> inode
->i_blkbits
;
541 bh_frozen
= nilfs_page_get_nth_block(page
, n
);
544 page_cache_release(page
);
549 static void nilfs_release_frozen_buffers(struct nilfs_shadow_map
*shadow
)
551 struct list_head
*head
= &shadow
->frozen_buffers
;
552 struct buffer_head
*bh
;
554 while (!list_empty(head
)) {
555 bh
= list_first_entry(head
, struct buffer_head
,
557 list_del_init(&bh
->b_assoc_buffers
);
558 brelse(bh
); /* drop ref-count to make it releasable */
563 * nilfs_mdt_restore_from_shadow_map - restore dirty pages and bmap state
564 * @inode: inode of the metadata file
566 void nilfs_mdt_restore_from_shadow_map(struct inode
*inode
)
568 struct nilfs_mdt_info
*mi
= NILFS_MDT(inode
);
569 struct nilfs_inode_info
*ii
= NILFS_I(inode
);
570 struct nilfs_shadow_map
*shadow
= mi
->mi_shadow
;
572 down_write(&mi
->mi_sem
);
574 if (mi
->mi_palloc_cache
)
575 nilfs_palloc_clear_cache(inode
);
577 nilfs_clear_dirty_pages(inode
->i_mapping
);
578 nilfs_copy_back_pages(inode
->i_mapping
, &shadow
->frozen_data
);
580 nilfs_clear_dirty_pages(&ii
->i_btnode_cache
);
581 nilfs_copy_back_pages(&ii
->i_btnode_cache
, &shadow
->frozen_btnodes
);
583 nilfs_bmap_restore(ii
->i_bmap
, &shadow
->bmap_store
);
585 up_write(&mi
->mi_sem
);
589 * nilfs_mdt_clear_shadow_map - truncate pages in shadow map caches
590 * @inode: inode of the metadata file
592 void nilfs_mdt_clear_shadow_map(struct inode
*inode
)
594 struct nilfs_mdt_info
*mi
= NILFS_MDT(inode
);
595 struct nilfs_shadow_map
*shadow
= mi
->mi_shadow
;
597 down_write(&mi
->mi_sem
);
598 nilfs_release_frozen_buffers(shadow
);
599 truncate_inode_pages(&shadow
->frozen_data
, 0);
600 truncate_inode_pages(&shadow
->frozen_btnodes
, 0);
601 up_write(&mi
->mi_sem
);