2 linear.c : Multiple Devices driver for Linux
3 Copyright (C) 1994-96 Marc ZYNGIER
4 <zyngier@ufr-info-p7.ibp.fr> or
7 Linear mode management functions.
9 This program is free software; you can redistribute it and/or modify
10 it under the terms of the GNU General Public License as published by
11 the Free Software Foundation; either version 2, or (at your option)
14 You should have received a copy of the GNU General Public License
15 (for example /usr/src/linux/COPYING); if not, write to the Free
16 Software Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
19 #include <linux/blkdev.h>
20 #include <linux/raid/md_u.h>
21 #include <linux/seq_file.h>
22 #include <linux/module.h>
23 #include <linux/slab.h>
28 * find which device holds a particular offset
30 static inline struct dev_info
*which_dev(struct mddev
*mddev
, sector_t sector
)
33 struct linear_conf
*conf
;
36 hi
= mddev
->raid_disks
- 1;
37 conf
= rcu_dereference(mddev
->private);
46 if (sector
< conf
->disks
[mid
].end_sector
)
52 return conf
->disks
+ lo
;
56 * linear_mergeable_bvec -- tell bio layer if two requests can be merged
58 * @bvm: properties of new bio
59 * @biovec: the request that could be merged to it.
61 * Return amount of bytes we can take at this offset
63 static int linear_mergeable_bvec(struct request_queue
*q
,
64 struct bvec_merge_data
*bvm
,
65 struct bio_vec
*biovec
)
67 struct mddev
*mddev
= q
->queuedata
;
68 struct dev_info
*dev0
;
69 unsigned long maxsectors
, bio_sectors
= bvm
->bi_size
>> 9;
70 sector_t sector
= bvm
->bi_sector
+ get_start_sect(bvm
->bi_bdev
);
73 dev0
= which_dev(mddev
, sector
);
74 maxsectors
= dev0
->end_sector
- sector
;
77 if (maxsectors
< bio_sectors
)
80 maxsectors
-= bio_sectors
;
82 if (maxsectors
<= (PAGE_SIZE
>> 9 ) && bio_sectors
== 0)
83 return biovec
->bv_len
;
84 /* The bytes available at this offset could be really big,
85 * so we cap at 2^31 to avoid overflow */
86 if (maxsectors
> (1 << (31-9)))
88 return maxsectors
<< 9;
91 static int linear_congested(void *data
, int bits
)
93 struct mddev
*mddev
= data
;
94 struct linear_conf
*conf
;
97 if (mddev_congested(mddev
, bits
))
101 conf
= rcu_dereference(mddev
->private);
103 for (i
= 0; i
< mddev
->raid_disks
&& !ret
; i
++) {
104 struct request_queue
*q
= bdev_get_queue(conf
->disks
[i
].rdev
->bdev
);
105 ret
|= bdi_congested(&q
->backing_dev_info
, bits
);
112 static sector_t
linear_size(struct mddev
*mddev
, sector_t sectors
, int raid_disks
)
114 struct linear_conf
*conf
;
115 sector_t array_sectors
;
118 conf
= rcu_dereference(mddev
->private);
119 WARN_ONCE(sectors
|| raid_disks
,
120 "%s does not support generic reshape\n", __func__
);
121 array_sectors
= conf
->array_sectors
;
124 return array_sectors
;
127 static struct linear_conf
*linear_conf(struct mddev
*mddev
, int raid_disks
)
129 struct linear_conf
*conf
;
130 struct md_rdev
*rdev
;
133 conf
= kzalloc (sizeof (*conf
) + raid_disks
*sizeof(struct dev_info
),
139 conf
->array_sectors
= 0;
141 list_for_each_entry(rdev
, &mddev
->disks
, same_set
) {
142 int j
= rdev
->raid_disk
;
143 struct dev_info
*disk
= conf
->disks
+ j
;
146 if (j
< 0 || j
>= raid_disks
|| disk
->rdev
) {
147 printk(KERN_ERR
"md/linear:%s: disk numbering problem. Aborting!\n",
153 if (mddev
->chunk_sectors
) {
154 sectors
= rdev
->sectors
;
155 sector_div(sectors
, mddev
->chunk_sectors
);
156 rdev
->sectors
= sectors
* mddev
->chunk_sectors
;
159 disk_stack_limits(mddev
->gendisk
, rdev
->bdev
,
160 rdev
->data_offset
<< 9);
161 /* as we don't honour merge_bvec_fn, we must never risk
162 * violating it, so limit max_segments to 1 lying within
165 if (rdev
->bdev
->bd_disk
->queue
->merge_bvec_fn
) {
166 blk_queue_max_segments(mddev
->queue
, 1);
167 blk_queue_segment_boundary(mddev
->queue
,
168 PAGE_CACHE_SIZE
- 1);
171 conf
->array_sectors
+= rdev
->sectors
;
175 if (cnt
!= raid_disks
) {
176 printk(KERN_ERR
"md/linear:%s: not enough drives present. Aborting!\n",
182 * Here we calculate the device offsets.
184 conf
->disks
[0].end_sector
= conf
->disks
[0].rdev
->sectors
;
186 for (i
= 1; i
< raid_disks
; i
++)
187 conf
->disks
[i
].end_sector
=
188 conf
->disks
[i
-1].end_sector
+
189 conf
->disks
[i
].rdev
->sectors
;
198 static int linear_run (struct mddev
*mddev
)
200 struct linear_conf
*conf
;
202 if (md_check_no_bitmap(mddev
))
204 conf
= linear_conf(mddev
, mddev
->raid_disks
);
208 mddev
->private = conf
;
209 md_set_array_sectors(mddev
, linear_size(mddev
, 0, 0));
211 blk_queue_merge_bvec(mddev
->queue
, linear_mergeable_bvec
);
212 mddev
->queue
->backing_dev_info
.congested_fn
= linear_congested
;
213 mddev
->queue
->backing_dev_info
.congested_data
= mddev
;
214 return md_integrity_register(mddev
);
217 static int linear_add(struct mddev
*mddev
, struct md_rdev
*rdev
)
219 /* Adding a drive to a linear array allows the array to grow.
220 * It is permitted if the new drive has a matching superblock
221 * already on it, with raid_disk equal to raid_disks.
222 * It is achieved by creating a new linear_private_data structure
223 * and swapping it in in-place of the current one.
224 * The current one is never freed until the array is stopped.
227 struct linear_conf
*newconf
, *oldconf
;
229 if (rdev
->saved_raid_disk
!= mddev
->raid_disks
)
232 rdev
->raid_disk
= rdev
->saved_raid_disk
;
234 newconf
= linear_conf(mddev
,mddev
->raid_disks
+1);
239 oldconf
= rcu_dereference(mddev
->private);
241 rcu_assign_pointer(mddev
->private, newconf
);
242 md_set_array_sectors(mddev
, linear_size(mddev
, 0, 0));
243 set_capacity(mddev
->gendisk
, mddev
->array_sectors
);
244 revalidate_disk(mddev
->gendisk
);
245 kfree_rcu(oldconf
, rcu
);
249 static int linear_stop (struct mddev
*mddev
)
251 struct linear_conf
*conf
= mddev
->private;
254 * We do not require rcu protection here since
255 * we hold reconfig_mutex for both linear_add and
256 * linear_stop, so they cannot race.
257 * We should make sure any old 'conf's are properly
261 blk_sync_queue(mddev
->queue
); /* the unplug fn references 'conf'*/
263 mddev
->private = NULL
;
268 static void linear_make_request(struct mddev
*mddev
, struct bio
*bio
)
270 struct dev_info
*tmp_dev
;
271 sector_t start_sector
;
273 if (unlikely(bio
->bi_rw
& REQ_FLUSH
)) {
274 md_flush_request(mddev
, bio
);
279 tmp_dev
= which_dev(mddev
, bio
->bi_sector
);
280 start_sector
= tmp_dev
->end_sector
- tmp_dev
->rdev
->sectors
;
283 if (unlikely(bio
->bi_sector
>= (tmp_dev
->end_sector
)
284 || (bio
->bi_sector
< start_sector
))) {
285 char b
[BDEVNAME_SIZE
];
288 "md/linear:%s: make_request: Sector %llu out of bounds on "
289 "dev %s: %llu sectors, offset %llu\n",
291 (unsigned long long)bio
->bi_sector
,
292 bdevname(tmp_dev
->rdev
->bdev
, b
),
293 (unsigned long long)tmp_dev
->rdev
->sectors
,
294 (unsigned long long)start_sector
);
299 if (unlikely(bio
->bi_sector
+ (bio
->bi_size
>> 9) >
300 tmp_dev
->end_sector
)) {
301 /* This bio crosses a device boundary, so we have to
305 sector_t end_sector
= tmp_dev
->end_sector
;
309 bp
= bio_split(bio
, end_sector
- bio
->bi_sector
);
311 linear_make_request(mddev
, &bp
->bio1
);
312 linear_make_request(mddev
, &bp
->bio2
);
313 bio_pair_release(bp
);
317 bio
->bi_bdev
= tmp_dev
->rdev
->bdev
;
318 bio
->bi_sector
= bio
->bi_sector
- start_sector
319 + tmp_dev
->rdev
->data_offset
;
321 generic_make_request(bio
);
324 static void linear_status (struct seq_file
*seq
, struct mddev
*mddev
)
327 seq_printf(seq
, " %dk rounding", mddev
->chunk_sectors
/ 2);
331 static struct md_personality linear_personality
=
334 .level
= LEVEL_LINEAR
,
335 .owner
= THIS_MODULE
,
336 .make_request
= linear_make_request
,
339 .status
= linear_status
,
340 .hot_add_disk
= linear_add
,
344 static int __init
linear_init (void)
346 return register_md_personality (&linear_personality
);
349 static void linear_exit (void)
351 unregister_md_personality (&linear_personality
);
355 module_init(linear_init
);
356 module_exit(linear_exit
);
357 MODULE_LICENSE("GPL");
358 MODULE_DESCRIPTION("Linear device concatenation personality for MD");
359 MODULE_ALIAS("md-personality-1"); /* LINEAR - deprecated*/
360 MODULE_ALIAS("md-linear");
361 MODULE_ALIAS("md-level--1");