1 // SPDX-License-Identifier: GPL-2.0-only
3 * Copyright (C) 2001-2003 Sistina Software (UK) Limited.
5 * This file is released under the GPL.
9 #include <linux/device-mapper.h>
11 #include <linux/module.h>
12 #include <linux/init.h>
13 #include <linux/blkdev.h>
14 #include <linux/bio.h>
15 #include <linux/dax.h>
16 #include <linux/slab.h>
17 #include <linux/log2.h>
19 static struct workqueue_struct
*dm_stripe_wq
;
21 #define DM_MSG_PREFIX "striped"
22 #define DM_IO_ERROR_THRESHOLD 15
26 sector_t physical_start
;
35 /* The size of this target / num. stripes */
36 sector_t stripe_width
;
41 /* Needed for handling events */
44 /* Work struct used for triggering events*/
45 struct work_struct trigger_event
;
47 struct stripe stripe
[] __counted_by(stripes
);
51 * An event is triggered whenever a drive
52 * drops out of a stripe volume.
54 static void trigger_event(struct work_struct
*work
)
56 struct stripe_c
*sc
= container_of(work
, struct stripe_c
,
58 dm_table_event(sc
->ti
->table
);
62 * Parse a single <dev> <sector> pair
64 static int get_stripe(struct dm_target
*ti
, struct stripe_c
*sc
,
65 unsigned int stripe
, char **argv
)
67 unsigned long long start
;
71 if (sscanf(argv
[1], "%llu%c", &start
, &dummy
) != 1)
74 ret
= dm_get_device(ti
, argv
[0], dm_table_get_mode(ti
->table
),
75 &sc
->stripe
[stripe
].dev
);
79 sc
->stripe
[stripe
].physical_start
= start
;
85 * Construct a striped mapping.
86 * <number of stripes> <chunk size> [<dev_path> <offset>]+
88 static int stripe_ctr(struct dm_target
*ti
, unsigned int argc
, char **argv
)
91 sector_t width
, tmp_len
;
98 ti
->error
= "Not enough arguments";
102 if (kstrtouint(argv
[0], 10, &stripes
) || !stripes
) {
103 ti
->error
= "Invalid stripe count";
107 if (kstrtouint(argv
[1], 10, &chunk_size
) || !chunk_size
) {
108 ti
->error
= "Invalid chunk_size";
113 if (sector_div(width
, stripes
)) {
114 ti
->error
= "Target length not divisible by number of stripes";
119 if (sector_div(tmp_len
, chunk_size
)) {
120 ti
->error
= "Target length not divisible by chunk size";
125 * Do we have enough arguments for that many stripes ?
127 if (argc
!= (2 + 2 * stripes
)) {
128 ti
->error
= "Not enough destinations specified";
132 sc
= kmalloc(struct_size(sc
, stripe
, stripes
), GFP_KERNEL
);
134 ti
->error
= "Memory allocation for striped context failed";
138 INIT_WORK(&sc
->trigger_event
, trigger_event
);
140 /* Set pointer to dm target; used in trigger_event */
142 sc
->stripes
= stripes
;
143 sc
->stripe_width
= width
;
145 if (stripes
& (stripes
- 1))
146 sc
->stripes_shift
= -1;
148 sc
->stripes_shift
= __ffs(stripes
);
150 r
= dm_set_target_max_io_len(ti
, chunk_size
);
156 ti
->num_flush_bios
= stripes
;
157 ti
->num_discard_bios
= stripes
;
158 ti
->num_secure_erase_bios
= stripes
;
159 ti
->num_write_zeroes_bios
= stripes
;
160 ti
->flush_bypasses_map
= true;
162 sc
->chunk_size
= chunk_size
;
163 if (chunk_size
& (chunk_size
- 1))
164 sc
->chunk_size_shift
= -1;
166 sc
->chunk_size_shift
= __ffs(chunk_size
);
169 * Get the stripe destinations.
171 for (i
= 0; i
< stripes
; i
++) {
174 r
= get_stripe(ti
, sc
, i
, argv
);
176 ti
->error
= "Couldn't parse stripe destination";
178 dm_put_device(ti
, sc
->stripe
[i
].dev
);
182 atomic_set(&(sc
->stripe
[i
].error_count
), 0);
190 static void stripe_dtr(struct dm_target
*ti
)
193 struct stripe_c
*sc
= ti
->private;
195 for (i
= 0; i
< sc
->stripes
; i
++)
196 dm_put_device(ti
, sc
->stripe
[i
].dev
);
198 flush_work(&sc
->trigger_event
);
202 static void stripe_map_sector(struct stripe_c
*sc
, sector_t sector
,
203 uint32_t *stripe
, sector_t
*result
)
205 sector_t chunk
= dm_target_offset(sc
->ti
, sector
);
206 sector_t chunk_offset
;
208 if (sc
->chunk_size_shift
< 0)
209 chunk_offset
= sector_div(chunk
, sc
->chunk_size
);
211 chunk_offset
= chunk
& (sc
->chunk_size
- 1);
212 chunk
>>= sc
->chunk_size_shift
;
215 if (sc
->stripes_shift
< 0)
216 *stripe
= sector_div(chunk
, sc
->stripes
);
218 *stripe
= chunk
& (sc
->stripes
- 1);
219 chunk
>>= sc
->stripes_shift
;
222 if (sc
->chunk_size_shift
< 0)
223 chunk
*= sc
->chunk_size
;
225 chunk
<<= sc
->chunk_size_shift
;
227 *result
= chunk
+ chunk_offset
;
230 static void stripe_map_range_sector(struct stripe_c
*sc
, sector_t sector
,
231 uint32_t target_stripe
, sector_t
*result
)
235 stripe_map_sector(sc
, sector
, &stripe
, result
);
236 if (stripe
== target_stripe
)
241 if (sc
->chunk_size_shift
< 0)
242 *result
-= sector_div(sector
, sc
->chunk_size
);
244 *result
= sector
& ~(sector_t
)(sc
->chunk_size
- 1);
246 if (target_stripe
< stripe
)
247 *result
+= sc
->chunk_size
; /* next chunk */
250 static int stripe_map_range(struct stripe_c
*sc
, struct bio
*bio
,
251 uint32_t target_stripe
)
255 stripe_map_range_sector(sc
, bio
->bi_iter
.bi_sector
,
256 target_stripe
, &begin
);
257 stripe_map_range_sector(sc
, bio_end_sector(bio
),
258 target_stripe
, &end
);
260 bio_set_dev(bio
, sc
->stripe
[target_stripe
].dev
->bdev
);
261 bio
->bi_iter
.bi_sector
= begin
+
262 sc
->stripe
[target_stripe
].physical_start
;
263 bio
->bi_iter
.bi_size
= to_bytes(end
- begin
);
264 return DM_MAPIO_REMAPPED
;
267 /* The range doesn't map to the target stripe */
269 return DM_MAPIO_SUBMITTED
;
272 int stripe_map(struct dm_target
*ti
, struct bio
*bio
)
274 struct stripe_c
*sc
= ti
->private;
276 unsigned int target_bio_nr
;
278 if (bio
->bi_opf
& REQ_PREFLUSH
) {
279 target_bio_nr
= dm_bio_get_target_bio_nr(bio
);
280 BUG_ON(target_bio_nr
>= sc
->stripes
);
281 bio_set_dev(bio
, sc
->stripe
[target_bio_nr
].dev
->bdev
);
282 return DM_MAPIO_REMAPPED
;
284 if (unlikely(bio_op(bio
) == REQ_OP_DISCARD
) ||
285 unlikely(bio_op(bio
) == REQ_OP_SECURE_ERASE
) ||
286 unlikely(bio_op(bio
) == REQ_OP_WRITE_ZEROES
)) {
287 target_bio_nr
= dm_bio_get_target_bio_nr(bio
);
288 BUG_ON(target_bio_nr
>= sc
->stripes
);
289 return stripe_map_range(sc
, bio
, target_bio_nr
);
292 stripe_map_sector(sc
, bio
->bi_iter
.bi_sector
,
293 &stripe
, &bio
->bi_iter
.bi_sector
);
295 bio
->bi_iter
.bi_sector
+= sc
->stripe
[stripe
].physical_start
;
296 bio_set_dev(bio
, sc
->stripe
[stripe
].dev
->bdev
);
298 return DM_MAPIO_REMAPPED
;
301 #if IS_ENABLED(CONFIG_FS_DAX)
302 static struct dax_device
*stripe_dax_pgoff(struct dm_target
*ti
, pgoff_t
*pgoff
)
304 struct stripe_c
*sc
= ti
->private;
305 struct block_device
*bdev
;
309 stripe_map_sector(sc
, *pgoff
* PAGE_SECTORS
, &stripe
, &dev_sector
);
310 dev_sector
+= sc
->stripe
[stripe
].physical_start
;
311 bdev
= sc
->stripe
[stripe
].dev
->bdev
;
313 *pgoff
= (get_start_sect(bdev
) + dev_sector
) >> PAGE_SECTORS_SHIFT
;
314 return sc
->stripe
[stripe
].dev
->dax_dev
;
317 static long stripe_dax_direct_access(struct dm_target
*ti
, pgoff_t pgoff
,
318 long nr_pages
, enum dax_access_mode mode
, void **kaddr
,
321 struct dax_device
*dax_dev
= stripe_dax_pgoff(ti
, &pgoff
);
323 return dax_direct_access(dax_dev
, pgoff
, nr_pages
, mode
, kaddr
, pfn
);
326 static int stripe_dax_zero_page_range(struct dm_target
*ti
, pgoff_t pgoff
,
329 struct dax_device
*dax_dev
= stripe_dax_pgoff(ti
, &pgoff
);
331 return dax_zero_page_range(dax_dev
, pgoff
, nr_pages
);
334 static size_t stripe_dax_recovery_write(struct dm_target
*ti
, pgoff_t pgoff
,
335 void *addr
, size_t bytes
, struct iov_iter
*i
)
337 struct dax_device
*dax_dev
= stripe_dax_pgoff(ti
, &pgoff
);
339 return dax_recovery_write(dax_dev
, pgoff
, addr
, bytes
, i
);
343 #define stripe_dax_direct_access NULL
344 #define stripe_dax_zero_page_range NULL
345 #define stripe_dax_recovery_write NULL
352 * #stripes [stripe_name <stripe_name>] [group word count]
353 * [error count 'A|D' <error count 'A|D'>]
356 * #stripes [stripe chunk size]
357 * [stripe_name physical_start <stripe_name physical_start>]
361 static void stripe_status(struct dm_target
*ti
, status_type_t type
,
362 unsigned int status_flags
, char *result
, unsigned int maxlen
)
364 struct stripe_c
*sc
= ti
->private;
369 case STATUSTYPE_INFO
:
370 DMEMIT("%d ", sc
->stripes
);
371 for (i
= 0; i
< sc
->stripes
; i
++)
372 DMEMIT("%s ", sc
->stripe
[i
].dev
->name
);
375 for (i
= 0; i
< sc
->stripes
; i
++)
376 DMEMIT("%c", atomic_read(&(sc
->stripe
[i
].error_count
)) ? 'D' : 'A');
379 case STATUSTYPE_TABLE
:
380 DMEMIT("%d %llu", sc
->stripes
,
381 (unsigned long long)sc
->chunk_size
);
382 for (i
= 0; i
< sc
->stripes
; i
++)
383 DMEMIT(" %s %llu", sc
->stripe
[i
].dev
->name
,
384 (unsigned long long)sc
->stripe
[i
].physical_start
);
388 DMEMIT_TARGET_NAME_VERSION(ti
->type
);
389 DMEMIT(",stripes=%d,chunk_size=%llu", sc
->stripes
,
390 (unsigned long long)sc
->chunk_size
);
392 for (i
= 0; i
< sc
->stripes
; i
++) {
393 DMEMIT(",stripe_%d_device_name=%s", i
, sc
->stripe
[i
].dev
->name
);
394 DMEMIT(",stripe_%d_physical_start=%llu", i
,
395 (unsigned long long)sc
->stripe
[i
].physical_start
);
396 DMEMIT(",stripe_%d_status=%c", i
,
397 atomic_read(&(sc
->stripe
[i
].error_count
)) ? 'D' : 'A');
404 static int stripe_end_io(struct dm_target
*ti
, struct bio
*bio
,
408 char major_minor
[16];
409 struct stripe_c
*sc
= ti
->private;
412 return DM_ENDIO_DONE
; /* I/O complete */
414 if (bio
->bi_opf
& REQ_RAHEAD
)
415 return DM_ENDIO_DONE
;
417 if (*error
== BLK_STS_NOTSUPP
)
418 return DM_ENDIO_DONE
;
420 memset(major_minor
, 0, sizeof(major_minor
));
421 sprintf(major_minor
, "%d:%d", MAJOR(bio_dev(bio
)), MINOR(bio_dev(bio
)));
424 * Test to see which stripe drive triggered the event
425 * and increment error count for all stripes on that device.
426 * If the error count for a given device exceeds the threshold
427 * value we will no longer trigger any further events.
429 for (i
= 0; i
< sc
->stripes
; i
++)
430 if (!strcmp(sc
->stripe
[i
].dev
->name
, major_minor
)) {
431 atomic_inc(&(sc
->stripe
[i
].error_count
));
432 if (atomic_read(&(sc
->stripe
[i
].error_count
)) <
433 DM_IO_ERROR_THRESHOLD
)
434 queue_work(dm_stripe_wq
, &sc
->trigger_event
);
437 return DM_ENDIO_DONE
;
440 static int stripe_iterate_devices(struct dm_target
*ti
,
441 iterate_devices_callout_fn fn
, void *data
)
443 struct stripe_c
*sc
= ti
->private;
448 ret
= fn(ti
, sc
->stripe
[i
].dev
,
449 sc
->stripe
[i
].physical_start
,
450 sc
->stripe_width
, data
);
451 } while (!ret
&& ++i
< sc
->stripes
);
456 static void stripe_io_hints(struct dm_target
*ti
,
457 struct queue_limits
*limits
)
459 struct stripe_c
*sc
= ti
->private;
460 unsigned int chunk_size
= sc
->chunk_size
<< SECTOR_SHIFT
;
462 limits
->io_min
= chunk_size
;
463 limits
->io_opt
= chunk_size
* sc
->stripes
;
466 static struct target_type stripe_target
= {
468 .version
= {1, 6, 0},
469 .features
= DM_TARGET_PASSES_INTEGRITY
| DM_TARGET_NOWAIT
,
470 .module
= THIS_MODULE
,
474 .end_io
= stripe_end_io
,
475 .status
= stripe_status
,
476 .iterate_devices
= stripe_iterate_devices
,
477 .io_hints
= stripe_io_hints
,
478 .direct_access
= stripe_dax_direct_access
,
479 .dax_zero_page_range
= stripe_dax_zero_page_range
,
480 .dax_recovery_write
= stripe_dax_recovery_write
,
483 int __init
dm_stripe_init(void)
487 dm_stripe_wq
= alloc_workqueue("dm_stripe_wq", 0, 0);
490 r
= dm_register_target(&stripe_target
);
492 destroy_workqueue(dm_stripe_wq
);
493 DMWARN("target registration failed");
499 void dm_stripe_exit(void)
501 dm_unregister_target(&stripe_target
);
502 destroy_workqueue(dm_stripe_wq
);