2 * segbuf.c - NILFS segment buffer
4 * Copyright (C) 2005-2008 Nippon Telegraph and Telephone Corporation.
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License as published by
8 * the Free Software Foundation; either version 2 of the License, or
9 * (at your option) any later version.
11 * This program is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 * GNU General Public License for more details.
16 * You should have received a copy of the GNU General Public License
17 * along with this program; if not, write to the Free Software
18 * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
20 * Written by Ryusuke Konishi <ryusuke@osrg.net>
24 #include <linux/buffer_head.h>
25 #include <linux/writeback.h>
26 #include <linux/crc32.h>
31 static struct kmem_cache
*nilfs_segbuf_cachep
;
33 static void nilfs_segbuf_init_once(void *obj
)
35 memset(obj
, 0, sizeof(struct nilfs_segment_buffer
));
38 int __init
nilfs_init_segbuf_cache(void)
41 kmem_cache_create("nilfs2_segbuf_cache",
42 sizeof(struct nilfs_segment_buffer
),
43 0, SLAB_RECLAIM_ACCOUNT
,
44 nilfs_segbuf_init_once
);
46 return (nilfs_segbuf_cachep
== NULL
) ? -ENOMEM
: 0;
49 void nilfs_destroy_segbuf_cache(void)
51 kmem_cache_destroy(nilfs_segbuf_cachep
);
54 struct nilfs_segment_buffer
*nilfs_segbuf_new(struct super_block
*sb
)
56 struct nilfs_segment_buffer
*segbuf
;
58 segbuf
= kmem_cache_alloc(nilfs_segbuf_cachep
, GFP_NOFS
);
59 if (unlikely(!segbuf
))
62 segbuf
->sb_super
= sb
;
63 INIT_LIST_HEAD(&segbuf
->sb_list
);
64 INIT_LIST_HEAD(&segbuf
->sb_segsum_buffers
);
65 INIT_LIST_HEAD(&segbuf
->sb_payload_buffers
);
69 void nilfs_segbuf_free(struct nilfs_segment_buffer
*segbuf
)
71 kmem_cache_free(nilfs_segbuf_cachep
, segbuf
);
74 void nilfs_segbuf_map(struct nilfs_segment_buffer
*segbuf
, __u64 segnum
,
75 unsigned long offset
, struct the_nilfs
*nilfs
)
77 segbuf
->sb_segnum
= segnum
;
78 nilfs_get_segment_range(nilfs
, segnum
, &segbuf
->sb_fseg_start
,
79 &segbuf
->sb_fseg_end
);
81 segbuf
->sb_pseg_start
= segbuf
->sb_fseg_start
+ offset
;
82 segbuf
->sb_rest_blocks
=
83 segbuf
->sb_fseg_end
- segbuf
->sb_pseg_start
+ 1;
86 void nilfs_segbuf_set_next_segnum(struct nilfs_segment_buffer
*segbuf
,
87 __u64 nextnum
, struct the_nilfs
*nilfs
)
89 segbuf
->sb_nextnum
= nextnum
;
90 segbuf
->sb_sum
.next
= nilfs_get_segment_start_blocknr(nilfs
, nextnum
);
93 int nilfs_segbuf_extend_segsum(struct nilfs_segment_buffer
*segbuf
)
95 struct buffer_head
*bh
;
97 bh
= sb_getblk(segbuf
->sb_super
,
98 segbuf
->sb_pseg_start
+ segbuf
->sb_sum
.nsumblk
);
102 nilfs_segbuf_add_segsum_buffer(segbuf
, bh
);
106 int nilfs_segbuf_extend_payload(struct nilfs_segment_buffer
*segbuf
,
107 struct buffer_head
**bhp
)
109 struct buffer_head
*bh
;
111 bh
= sb_getblk(segbuf
->sb_super
,
112 segbuf
->sb_pseg_start
+ segbuf
->sb_sum
.nblocks
);
116 nilfs_segbuf_add_payload_buffer(segbuf
, bh
);
121 int nilfs_segbuf_reset(struct nilfs_segment_buffer
*segbuf
, unsigned flags
,
126 segbuf
->sb_sum
.nblocks
= segbuf
->sb_sum
.nsumblk
= 0;
127 err
= nilfs_segbuf_extend_segsum(segbuf
);
131 segbuf
->sb_sum
.flags
= flags
;
132 segbuf
->sb_sum
.sumbytes
= sizeof(struct nilfs_segment_summary
);
133 segbuf
->sb_sum
.nfinfo
= segbuf
->sb_sum
.nfileblk
= 0;
134 segbuf
->sb_sum
.ctime
= ctime
;
136 segbuf
->sb_io_error
= 0;
141 * Setup segument summary
143 void nilfs_segbuf_fill_in_segsum(struct nilfs_segment_buffer
*segbuf
)
145 struct nilfs_segment_summary
*raw_sum
;
146 struct buffer_head
*bh_sum
;
148 bh_sum
= list_entry(segbuf
->sb_segsum_buffers
.next
,
149 struct buffer_head
, b_assoc_buffers
);
150 raw_sum
= (struct nilfs_segment_summary
*)bh_sum
->b_data
;
152 raw_sum
->ss_magic
= cpu_to_le32(NILFS_SEGSUM_MAGIC
);
153 raw_sum
->ss_bytes
= cpu_to_le16(sizeof(*raw_sum
));
154 raw_sum
->ss_flags
= cpu_to_le16(segbuf
->sb_sum
.flags
);
155 raw_sum
->ss_seq
= cpu_to_le64(segbuf
->sb_sum
.seg_seq
);
156 raw_sum
->ss_create
= cpu_to_le64(segbuf
->sb_sum
.ctime
);
157 raw_sum
->ss_next
= cpu_to_le64(segbuf
->sb_sum
.next
);
158 raw_sum
->ss_nblocks
= cpu_to_le32(segbuf
->sb_sum
.nblocks
);
159 raw_sum
->ss_nfinfo
= cpu_to_le32(segbuf
->sb_sum
.nfinfo
);
160 raw_sum
->ss_sumbytes
= cpu_to_le32(segbuf
->sb_sum
.sumbytes
);
165 * CRC calculation routines
167 void nilfs_segbuf_fill_in_segsum_crc(struct nilfs_segment_buffer
*segbuf
,
170 struct buffer_head
*bh
;
171 struct nilfs_segment_summary
*raw_sum
;
172 unsigned long size
, bytes
= segbuf
->sb_sum
.sumbytes
;
175 bh
= list_entry(segbuf
->sb_segsum_buffers
.next
, struct buffer_head
,
178 raw_sum
= (struct nilfs_segment_summary
*)bh
->b_data
;
179 size
= min_t(unsigned long, bytes
, bh
->b_size
);
181 (unsigned char *)raw_sum
+
182 sizeof(raw_sum
->ss_datasum
) + sizeof(raw_sum
->ss_sumsum
),
183 size
- (sizeof(raw_sum
->ss_datasum
) +
184 sizeof(raw_sum
->ss_sumsum
)));
186 list_for_each_entry_continue(bh
, &segbuf
->sb_segsum_buffers
,
189 size
= min_t(unsigned long, bytes
, bh
->b_size
);
190 crc
= crc32_le(crc
, bh
->b_data
, size
);
192 raw_sum
->ss_sumsum
= cpu_to_le32(crc
);
195 void nilfs_segbuf_fill_in_data_crc(struct nilfs_segment_buffer
*segbuf
,
198 struct buffer_head
*bh
;
199 struct nilfs_segment_summary
*raw_sum
;
203 bh
= list_entry(segbuf
->sb_segsum_buffers
.next
, struct buffer_head
,
205 raw_sum
= (struct nilfs_segment_summary
*)bh
->b_data
;
207 (unsigned char *)raw_sum
+ sizeof(raw_sum
->ss_datasum
),
208 bh
->b_size
- sizeof(raw_sum
->ss_datasum
));
210 list_for_each_entry_continue(bh
, &segbuf
->sb_segsum_buffers
,
212 crc
= crc32_le(crc
, bh
->b_data
, bh
->b_size
);
214 list_for_each_entry(bh
, &segbuf
->sb_payload_buffers
, b_assoc_buffers
) {
215 kaddr
= kmap_atomic(bh
->b_page
, KM_USER0
);
216 crc
= crc32_le(crc
, kaddr
+ bh_offset(bh
), bh
->b_size
);
217 kunmap_atomic(kaddr
, KM_USER0
);
219 raw_sum
->ss_datasum
= cpu_to_le32(crc
);
222 void nilfs_release_buffers(struct list_head
*list
)
224 struct buffer_head
*bh
, *n
;
226 list_for_each_entry_safe(bh
, n
, list
, b_assoc_buffers
) {
227 list_del_init(&bh
->b_assoc_buffers
);
228 if (buffer_nilfs_allocated(bh
)) {
229 struct page
*clone_page
= bh
->b_page
;
231 /* remove clone page */
233 page_cache_release(clone_page
); /* for each bh */
234 if (page_count(clone_page
) <= 2) {
235 lock_page(clone_page
);
236 nilfs_free_private_page(clone_page
);
247 static void nilfs_end_bio_write(struct bio
*bio
, int err
)
249 const int uptodate
= test_bit(BIO_UPTODATE
, &bio
->bi_flags
);
250 struct nilfs_write_info
*wi
= bio
->bi_private
;
252 if (err
== -EOPNOTSUPP
) {
253 set_bit(BIO_EOPNOTSUPP
, &bio
->bi_flags
);
255 /* to be detected by submit_seg_bio() */
259 atomic_inc(&wi
->err
);
262 complete(&wi
->bio_event
);
265 static int nilfs_submit_seg_bio(struct nilfs_write_info
*wi
, int mode
)
267 struct bio
*bio
= wi
->bio
;
270 if (wi
->nbio
> 0 && bdi_write_congested(wi
->bdi
)) {
271 wait_for_completion(&wi
->bio_event
);
273 if (unlikely(atomic_read(&wi
->err
))) {
280 bio
->bi_end_io
= nilfs_end_bio_write
;
281 bio
->bi_private
= wi
;
283 submit_bio(mode
, bio
);
284 if (bio_flagged(bio
, BIO_EOPNOTSUPP
)) {
293 wi
->rest_blocks
-= wi
->end
- wi
->start
;
294 wi
->nr_vecs
= min(wi
->max_pages
, wi
->rest_blocks
);
304 * nilfs_alloc_seg_bio - allocate a bio for writing segment.
306 * @start: beginning disk block number of this BIO.
307 * @nr_vecs: request size of page vector.
309 * alloc_seg_bio() allocates a new BIO structure and initialize it.
311 * Return Value: On success, pointer to the struct bio is returned.
312 * On error, NULL is returned.
314 static struct bio
*nilfs_alloc_seg_bio(struct super_block
*sb
, sector_t start
,
319 bio
= bio_alloc(GFP_NOWAIT
, nr_vecs
);
321 while (!bio
&& (nr_vecs
>>= 1))
322 bio
= bio_alloc(GFP_NOWAIT
, nr_vecs
);
325 bio
->bi_bdev
= sb
->s_bdev
;
326 bio
->bi_sector
= (sector_t
)start
<< (sb
->s_blocksize_bits
- 9);
331 void nilfs_segbuf_prepare_write(struct nilfs_segment_buffer
*segbuf
,
332 struct nilfs_write_info
*wi
)
335 wi
->rest_blocks
= segbuf
->sb_sum
.nblocks
;
336 wi
->max_pages
= bio_get_nr_vecs(wi
->sb
->s_bdev
);
337 wi
->nr_vecs
= min(wi
->max_pages
, wi
->rest_blocks
);
338 wi
->start
= wi
->end
= 0;
340 wi
->blocknr
= segbuf
->sb_pseg_start
;
342 atomic_set(&wi
->err
, 0);
343 init_completion(&wi
->bio_event
);
346 static int nilfs_submit_bh(struct nilfs_write_info
*wi
, struct buffer_head
*bh
,
351 BUG_ON(wi
->nr_vecs
<= 0);
354 wi
->bio
= nilfs_alloc_seg_bio(wi
->sb
, wi
->blocknr
+ wi
->end
,
356 if (unlikely(!wi
->bio
))
360 len
= bio_add_page(wi
->bio
, bh
->b_page
, bh
->b_size
, bh_offset(bh
));
361 if (len
== bh
->b_size
) {
366 err
= nilfs_submit_seg_bio(wi
, mode
);
367 /* never submit current bh */
373 int nilfs_segbuf_write(struct nilfs_segment_buffer
*segbuf
,
374 struct nilfs_write_info
*wi
)
376 struct buffer_head
*bh
;
379 list_for_each_entry(bh
, &segbuf
->sb_segsum_buffers
, b_assoc_buffers
) {
380 res
= nilfs_submit_bh(wi
, bh
, rw
);
385 list_for_each_entry(bh
, &segbuf
->sb_payload_buffers
, b_assoc_buffers
) {
386 res
= nilfs_submit_bh(wi
, bh
, rw
);
393 * Last BIO is always sent through the following
396 rw
|= (1 << BIO_RW_SYNCIO
) | (1 << BIO_RW_UNPLUG
);
397 res
= nilfs_submit_seg_bio(wi
, rw
);
407 atomic_inc(&wi
->err
);
412 * nilfs_segbuf_wait - wait for completion of requested BIOs
413 * @wi: nilfs_write_info
415 * Return Value: On Success, 0 is returned. On Error, one of the following
416 * negative error code is returned.
420 int nilfs_segbuf_wait(struct nilfs_segment_buffer
*segbuf
,
421 struct nilfs_write_info
*wi
)
429 wait_for_completion(&wi
->bio_event
);
430 } while (--wi
->nbio
> 0);
432 if (unlikely(atomic_read(&wi
->err
) > 0)) {
433 printk(KERN_ERR
"NILFS: IO error writing segment\n");
435 segbuf
->sb_io_error
= 1;