2 * Copyright (C) 2001-2003 Sistina Software (UK) Limited.
4 * This file is released under the GPL.
8 #include <linux/device-mapper.h>
10 #include <linux/module.h>
11 #include <linux/init.h>
12 #include <linux/blkdev.h>
13 #include <linux/bio.h>
14 #include <linux/slab.h>
15 #include <linux/log2.h>
17 #define DM_MSG_PREFIX "striped"
18 #define DM_IO_ERROR_THRESHOLD 15
22 sector_t physical_start
;
31 /* The size of this target / num. stripes */
32 sector_t stripe_width
;
37 /* Needed for handling events */
40 /* Work struct used for triggering events*/
41 struct work_struct trigger_event
;
43 struct stripe stripe
[0];
47 * An event is triggered whenever a drive
48 * drops out of a stripe volume.
50 static void trigger_event(struct work_struct
*work
)
52 struct stripe_c
*sc
= container_of(work
, struct stripe_c
,
54 dm_table_event(sc
->ti
->table
);
57 static inline struct stripe_c
*alloc_context(unsigned int stripes
)
61 if (dm_array_too_big(sizeof(struct stripe_c
), sizeof(struct stripe
),
65 len
= sizeof(struct stripe_c
) + (sizeof(struct stripe
) * stripes
);
67 return kmalloc(len
, GFP_KERNEL
);
71 * Parse a single <dev> <sector> pair
73 static int get_stripe(struct dm_target
*ti
, struct stripe_c
*sc
,
74 unsigned int stripe
, char **argv
)
76 unsigned long long start
;
80 if (sscanf(argv
[1], "%llu%c", &start
, &dummy
) != 1)
83 ret
= dm_get_device(ti
, argv
[0], dm_table_get_mode(ti
->table
),
84 &sc
->stripe
[stripe
].dev
);
88 sc
->stripe
[stripe
].physical_start
= start
;
94 * Construct a striped mapping.
95 * <number of stripes> <chunk size> [<dev_path> <offset>]+
97 static int stripe_ctr(struct dm_target
*ti
, unsigned int argc
, char **argv
)
100 sector_t width
, tmp_len
;
107 ti
->error
= "Not enough arguments";
111 if (kstrtouint(argv
[0], 10, &stripes
) || !stripes
) {
112 ti
->error
= "Invalid stripe count";
116 if (kstrtouint(argv
[1], 10, &chunk_size
) || !chunk_size
) {
117 ti
->error
= "Invalid chunk_size";
122 if (sector_div(width
, stripes
)) {
123 ti
->error
= "Target length not divisible by "
129 if (sector_div(tmp_len
, chunk_size
)) {
130 ti
->error
= "Target length not divisible by "
136 * Do we have enough arguments for that many stripes ?
138 if (argc
!= (2 + 2 * stripes
)) {
139 ti
->error
= "Not enough destinations "
144 sc
= alloc_context(stripes
);
146 ti
->error
= "Memory allocation for striped context "
151 INIT_WORK(&sc
->trigger_event
, trigger_event
);
153 /* Set pointer to dm target; used in trigger_event */
155 sc
->stripes
= stripes
;
156 sc
->stripe_width
= width
;
158 if (stripes
& (stripes
- 1))
159 sc
->stripes_shift
= -1;
161 sc
->stripes_shift
= __ffs(stripes
);
163 r
= dm_set_target_max_io_len(ti
, chunk_size
);
169 ti
->num_flush_bios
= stripes
;
170 ti
->num_discard_bios
= stripes
;
171 ti
->num_write_same_bios
= stripes
;
173 sc
->chunk_size
= chunk_size
;
174 if (chunk_size
& (chunk_size
- 1))
175 sc
->chunk_size_shift
= -1;
177 sc
->chunk_size_shift
= __ffs(chunk_size
);
180 * Get the stripe destinations.
182 for (i
= 0; i
< stripes
; i
++) {
185 r
= get_stripe(ti
, sc
, i
, argv
);
187 ti
->error
= "Couldn't parse stripe destination";
189 dm_put_device(ti
, sc
->stripe
[i
].dev
);
193 atomic_set(&(sc
->stripe
[i
].error_count
), 0);
201 static void stripe_dtr(struct dm_target
*ti
)
204 struct stripe_c
*sc
= (struct stripe_c
*) ti
->private;
206 for (i
= 0; i
< sc
->stripes
; i
++)
207 dm_put_device(ti
, sc
->stripe
[i
].dev
);
209 flush_work(&sc
->trigger_event
);
213 static void stripe_map_sector(struct stripe_c
*sc
, sector_t sector
,
214 uint32_t *stripe
, sector_t
*result
)
216 sector_t chunk
= dm_target_offset(sc
->ti
, sector
);
217 sector_t chunk_offset
;
219 if (sc
->chunk_size_shift
< 0)
220 chunk_offset
= sector_div(chunk
, sc
->chunk_size
);
222 chunk_offset
= chunk
& (sc
->chunk_size
- 1);
223 chunk
>>= sc
->chunk_size_shift
;
226 if (sc
->stripes_shift
< 0)
227 *stripe
= sector_div(chunk
, sc
->stripes
);
229 *stripe
= chunk
& (sc
->stripes
- 1);
230 chunk
>>= sc
->stripes_shift
;
233 if (sc
->chunk_size_shift
< 0)
234 chunk
*= sc
->chunk_size
;
236 chunk
<<= sc
->chunk_size_shift
;
238 *result
= chunk
+ chunk_offset
;
241 static void stripe_map_range_sector(struct stripe_c
*sc
, sector_t sector
,
242 uint32_t target_stripe
, sector_t
*result
)
246 stripe_map_sector(sc
, sector
, &stripe
, result
);
247 if (stripe
== target_stripe
)
252 if (sc
->chunk_size_shift
< 0)
253 *result
-= sector_div(sector
, sc
->chunk_size
);
255 *result
= sector
& ~(sector_t
)(sc
->chunk_size
- 1);
257 if (target_stripe
< stripe
)
258 *result
+= sc
->chunk_size
; /* next chunk */
261 static int stripe_map_range(struct stripe_c
*sc
, struct bio
*bio
,
262 uint32_t target_stripe
)
266 stripe_map_range_sector(sc
, bio
->bi_iter
.bi_sector
,
267 target_stripe
, &begin
);
268 stripe_map_range_sector(sc
, bio_end_sector(bio
),
269 target_stripe
, &end
);
271 bio
->bi_bdev
= sc
->stripe
[target_stripe
].dev
->bdev
;
272 bio
->bi_iter
.bi_sector
= begin
+
273 sc
->stripe
[target_stripe
].physical_start
;
274 bio
->bi_iter
.bi_size
= to_bytes(end
- begin
);
275 return DM_MAPIO_REMAPPED
;
277 /* The range doesn't map to the target stripe */
279 return DM_MAPIO_SUBMITTED
;
283 static int stripe_map(struct dm_target
*ti
, struct bio
*bio
)
285 struct stripe_c
*sc
= ti
->private;
287 unsigned target_bio_nr
;
289 if (bio
->bi_opf
& REQ_PREFLUSH
) {
290 target_bio_nr
= dm_bio_get_target_bio_nr(bio
);
291 BUG_ON(target_bio_nr
>= sc
->stripes
);
292 bio
->bi_bdev
= sc
->stripe
[target_bio_nr
].dev
->bdev
;
293 return DM_MAPIO_REMAPPED
;
295 if (unlikely(bio_op(bio
) == REQ_OP_DISCARD
) ||
296 unlikely(bio_op(bio
) == REQ_OP_WRITE_SAME
)) {
297 target_bio_nr
= dm_bio_get_target_bio_nr(bio
);
298 BUG_ON(target_bio_nr
>= sc
->stripes
);
299 return stripe_map_range(sc
, bio
, target_bio_nr
);
302 stripe_map_sector(sc
, bio
->bi_iter
.bi_sector
,
303 &stripe
, &bio
->bi_iter
.bi_sector
);
305 bio
->bi_iter
.bi_sector
+= sc
->stripe
[stripe
].physical_start
;
306 bio
->bi_bdev
= sc
->stripe
[stripe
].dev
->bdev
;
308 return DM_MAPIO_REMAPPED
;
311 static long stripe_direct_access(struct dm_target
*ti
, sector_t sector
,
312 void **kaddr
, pfn_t
*pfn
, long size
)
314 struct stripe_c
*sc
= ti
->private;
316 struct block_device
*bdev
;
317 struct blk_dax_ctl dax
= {
322 stripe_map_sector(sc
, sector
, &stripe
, &dax
.sector
);
324 dax
.sector
+= sc
->stripe
[stripe
].physical_start
;
325 bdev
= sc
->stripe
[stripe
].dev
->bdev
;
327 ret
= bdev_direct_access(bdev
, &dax
);
338 * #stripes [stripe_name <stripe_name>] [group word count]
339 * [error count 'A|D' <error count 'A|D'>]
342 * #stripes [stripe chunk size]
343 * [stripe_name physical_start <stripe_name physical_start>]
347 static void stripe_status(struct dm_target
*ti
, status_type_t type
,
348 unsigned status_flags
, char *result
, unsigned maxlen
)
350 struct stripe_c
*sc
= (struct stripe_c
*) ti
->private;
351 char buffer
[sc
->stripes
+ 1];
356 case STATUSTYPE_INFO
:
357 DMEMIT("%d ", sc
->stripes
);
358 for (i
= 0; i
< sc
->stripes
; i
++) {
359 DMEMIT("%s ", sc
->stripe
[i
].dev
->name
);
360 buffer
[i
] = atomic_read(&(sc
->stripe
[i
].error_count
)) ?
364 DMEMIT("1 %s", buffer
);
367 case STATUSTYPE_TABLE
:
368 DMEMIT("%d %llu", sc
->stripes
,
369 (unsigned long long)sc
->chunk_size
);
370 for (i
= 0; i
< sc
->stripes
; i
++)
371 DMEMIT(" %s %llu", sc
->stripe
[i
].dev
->name
,
372 (unsigned long long)sc
->stripe
[i
].physical_start
);
377 static int stripe_end_io(struct dm_target
*ti
, struct bio
*bio
, int error
)
380 char major_minor
[16];
381 struct stripe_c
*sc
= ti
->private;
384 return 0; /* I/O complete */
386 if ((error
== -EWOULDBLOCK
) && (bio
->bi_opf
& REQ_RAHEAD
))
389 if (error
== -EOPNOTSUPP
)
392 memset(major_minor
, 0, sizeof(major_minor
));
393 sprintf(major_minor
, "%d:%d",
394 MAJOR(disk_devt(bio
->bi_bdev
->bd_disk
)),
395 MINOR(disk_devt(bio
->bi_bdev
->bd_disk
)));
398 * Test to see which stripe drive triggered the event
399 * and increment error count for all stripes on that device.
400 * If the error count for a given device exceeds the threshold
401 * value we will no longer trigger any further events.
403 for (i
= 0; i
< sc
->stripes
; i
++)
404 if (!strcmp(sc
->stripe
[i
].dev
->name
, major_minor
)) {
405 atomic_inc(&(sc
->stripe
[i
].error_count
));
406 if (atomic_read(&(sc
->stripe
[i
].error_count
)) <
407 DM_IO_ERROR_THRESHOLD
)
408 schedule_work(&sc
->trigger_event
);
414 static int stripe_iterate_devices(struct dm_target
*ti
,
415 iterate_devices_callout_fn fn
, void *data
)
417 struct stripe_c
*sc
= ti
->private;
422 ret
= fn(ti
, sc
->stripe
[i
].dev
,
423 sc
->stripe
[i
].physical_start
,
424 sc
->stripe_width
, data
);
425 } while (!ret
&& ++i
< sc
->stripes
);
430 static void stripe_io_hints(struct dm_target
*ti
,
431 struct queue_limits
*limits
)
433 struct stripe_c
*sc
= ti
->private;
434 unsigned chunk_size
= sc
->chunk_size
<< SECTOR_SHIFT
;
436 blk_limits_io_min(limits
, chunk_size
);
437 blk_limits_io_opt(limits
, chunk_size
* sc
->stripes
);
440 static struct target_type stripe_target
= {
442 .version
= {1, 6, 0},
443 .module
= THIS_MODULE
,
447 .end_io
= stripe_end_io
,
448 .status
= stripe_status
,
449 .iterate_devices
= stripe_iterate_devices
,
450 .io_hints
= stripe_io_hints
,
451 .direct_access
= stripe_direct_access
,
454 int __init
dm_stripe_init(void)
458 r
= dm_register_target(&stripe_target
);
460 DMWARN("target registration failed");
465 void dm_stripe_exit(void)
467 dm_unregister_target(&stripe_target
);