1 // SPDX-License-Identifier: GPL-2.0-or-later
3 * faulty.c : Multiple Devices driver for Linux
5 * Copyright (C) 2004 Neil Brown
7 * fautly-device-simulator personality for md
12 * The "faulty" personality causes some requests to fail.
14 * Possible failure modes are:
15 * reads fail "randomly" but succeed on retry
16 * writes fail "randomly" but succeed on retry
17 * reads for some address fail and then persist until a write
18 * reads for some address fail and then persist irrespective of write
19 * writes for some address fail and persist
22 * Different modes can be active at a time, but only
23 * one can be set at array creation. Others can be added later.
24 * A mode can be one-shot or recurrent with the recurrence being
25 * once in every N requests.
26 * The bottom 5 bits of the "layout" indicate the mode. The
27 * remainder indicate a period, or 0 for one-shot.
29 * There is an implementation limit on the number of concurrently
30 * persisting-faulty blocks. When a new fault is requested that would
31 * exceed the limit, it is ignored.
32 * All current faults can be clear using a layout of "0".
34 * Requests are always sent to the device. If they are to fail,
35 * we clone the bio and insert a new b_end_io into the chain.
38 #define WriteTransient 0
39 #define ReadTransient 1
40 #define WritePersistent 2
41 #define ReadPersistent 3
42 #define WriteAll 4 /* doesn't go to device */
46 #define ClearErrors 31
47 #define ClearFaults 30
49 #define AllPersist 100 /* internal use only */
56 #include <linux/blkdev.h>
57 #include <linux/module.h>
58 #include <linux/raid/md_u.h>
59 #include <linux/slab.h>
61 #include <linux/seq_file.h>
64 static void faulty_fail(struct bio
*bio
)
66 struct bio
*b
= bio
->bi_private
;
68 b
->bi_iter
.bi_size
= bio
->bi_iter
.bi_size
;
69 b
->bi_iter
.bi_sector
= bio
->bi_iter
.bi_sector
;
78 atomic_t counters
[Modes
];
79 sector_t faults
[MaxFault
];
85 static int check_mode(struct faulty_conf
*conf
, int mode
)
87 if (conf
->period
[mode
] == 0 &&
88 atomic_read(&conf
->counters
[mode
]) <= 0)
89 return 0; /* no failure, no decrement */
92 if (atomic_dec_and_test(&conf
->counters
[mode
])) {
93 if (conf
->period
[mode
])
94 atomic_set(&conf
->counters
[mode
], conf
->period
[mode
]);
100 static int check_sector(struct faulty_conf
*conf
, sector_t start
, sector_t end
, int dir
)
102 /* If we find a ReadFixable sector, we fix it ... */
104 for (i
=0; i
<conf
->nfaults
; i
++)
105 if (conf
->faults
[i
] >= start
&&
106 conf
->faults
[i
] < end
) {
108 switch (conf
->modes
[i
] * 2 + dir
) {
109 case WritePersistent
*2+WRITE
: return 1;
110 case ReadPersistent
*2+READ
: return 1;
111 case ReadFixable
*2+READ
: return 1;
112 case ReadFixable
*2+WRITE
:
113 conf
->modes
[i
] = NoPersist
;
115 case AllPersist
*2+READ
:
116 case AllPersist
*2+WRITE
: return 1;
124 static void add_sector(struct faulty_conf
*conf
, sector_t start
, int mode
)
127 int n
= conf
->nfaults
;
128 for (i
=0; i
<conf
->nfaults
; i
++)
129 if (conf
->faults
[i
] == start
) {
131 case NoPersist
: conf
->modes
[i
] = mode
; return;
132 case WritePersistent
:
133 if (conf
->modes
[i
] == ReadPersistent
||
134 conf
->modes
[i
] == ReadFixable
)
135 conf
->modes
[i
] = AllPersist
;
137 conf
->modes
[i
] = WritePersistent
;
140 if (conf
->modes
[i
] == WritePersistent
)
141 conf
->modes
[i
] = AllPersist
;
143 conf
->modes
[i
] = ReadPersistent
;
146 if (conf
->modes
[i
] == WritePersistent
||
147 conf
->modes
[i
] == ReadPersistent
)
148 conf
->modes
[i
] = AllPersist
;
150 conf
->modes
[i
] = ReadFixable
;
153 } else if (conf
->modes
[i
] == NoPersist
)
158 conf
->faults
[n
] = start
;
159 conf
->modes
[n
] = mode
;
160 if (conf
->nfaults
== n
)
164 static bool faulty_make_request(struct mddev
*mddev
, struct bio
*bio
)
166 struct faulty_conf
*conf
= mddev
->private;
169 if (bio_data_dir(bio
) == WRITE
) {
171 if (atomic_read(&conf
->counters
[WriteAll
])) {
172 /* special case - don't decrement, don't submit_bio_noacct,
173 * just fail immediately
179 if (check_sector(conf
, bio
->bi_iter
.bi_sector
,
180 bio_end_sector(bio
), WRITE
))
182 if (check_mode(conf
, WritePersistent
)) {
183 add_sector(conf
, bio
->bi_iter
.bi_sector
,
187 if (check_mode(conf
, WriteTransient
))
191 if (check_sector(conf
, bio
->bi_iter
.bi_sector
,
192 bio_end_sector(bio
), READ
))
194 if (check_mode(conf
, ReadTransient
))
196 if (check_mode(conf
, ReadPersistent
)) {
197 add_sector(conf
, bio
->bi_iter
.bi_sector
,
201 if (check_mode(conf
, ReadFixable
)) {
202 add_sector(conf
, bio
->bi_iter
.bi_sector
,
208 struct bio
*b
= bio_clone_fast(bio
, GFP_NOIO
, &mddev
->bio_set
);
210 bio_set_dev(b
, conf
->rdev
->bdev
);
212 b
->bi_end_io
= faulty_fail
;
215 bio_set_dev(bio
, conf
->rdev
->bdev
);
217 submit_bio_noacct(bio
);
221 static void faulty_status(struct seq_file
*seq
, struct mddev
*mddev
)
223 struct faulty_conf
*conf
= mddev
->private;
226 if ((n
=atomic_read(&conf
->counters
[WriteTransient
])) != 0)
227 seq_printf(seq
, " WriteTransient=%d(%d)",
228 n
, conf
->period
[WriteTransient
]);
230 if ((n
=atomic_read(&conf
->counters
[ReadTransient
])) != 0)
231 seq_printf(seq
, " ReadTransient=%d(%d)",
232 n
, conf
->period
[ReadTransient
]);
234 if ((n
=atomic_read(&conf
->counters
[WritePersistent
])) != 0)
235 seq_printf(seq
, " WritePersistent=%d(%d)",
236 n
, conf
->period
[WritePersistent
]);
238 if ((n
=atomic_read(&conf
->counters
[ReadPersistent
])) != 0)
239 seq_printf(seq
, " ReadPersistent=%d(%d)",
240 n
, conf
->period
[ReadPersistent
]);
243 if ((n
=atomic_read(&conf
->counters
[ReadFixable
])) != 0)
244 seq_printf(seq
, " ReadFixable=%d(%d)",
245 n
, conf
->period
[ReadFixable
]);
247 if ((n
=atomic_read(&conf
->counters
[WriteAll
])) != 0)
248 seq_printf(seq
, " WriteAll");
250 seq_printf(seq
, " nfaults=%d", conf
->nfaults
);
254 static int faulty_reshape(struct mddev
*mddev
)
256 int mode
= mddev
->new_layout
& ModeMask
;
257 int count
= mddev
->new_layout
>> ModeShift
;
258 struct faulty_conf
*conf
= mddev
->private;
260 if (mddev
->new_layout
< 0)
264 if (mode
== ClearFaults
)
266 else if (mode
== ClearErrors
) {
268 for (i
=0 ; i
< Modes
; i
++) {
270 atomic_set(&conf
->counters
[i
], 0);
272 } else if (mode
< Modes
) {
273 conf
->period
[mode
] = count
;
275 atomic_set(&conf
->counters
[mode
], count
);
278 mddev
->new_layout
= -1;
279 mddev
->layout
= -1; /* makes sure further changes come through */
283 static sector_t
faulty_size(struct mddev
*mddev
, sector_t sectors
, int raid_disks
)
285 WARN_ONCE(raid_disks
,
286 "%s does not support generic reshape\n", __func__
);
289 return mddev
->dev_sectors
;
294 static int faulty_run(struct mddev
*mddev
)
296 struct md_rdev
*rdev
;
298 struct faulty_conf
*conf
;
300 if (md_check_no_bitmap(mddev
))
303 conf
= kmalloc(sizeof(*conf
), GFP_KERNEL
);
307 for (i
=0; i
<Modes
; i
++) {
308 atomic_set(&conf
->counters
[i
], 0);
313 rdev_for_each(rdev
, mddev
) {
315 disk_stack_limits(mddev
->gendisk
, rdev
->bdev
,
316 rdev
->data_offset
<< 9);
319 md_set_array_sectors(mddev
, faulty_size(mddev
, 0, 0));
320 mddev
->private = conf
;
322 faulty_reshape(mddev
);
327 static void faulty_free(struct mddev
*mddev
, void *priv
)
329 struct faulty_conf
*conf
= priv
;
334 static struct md_personality faulty_personality
=
337 .level
= LEVEL_FAULTY
,
338 .owner
= THIS_MODULE
,
339 .make_request
= faulty_make_request
,
342 .status
= faulty_status
,
343 .check_reshape
= faulty_reshape
,
347 static int __init
raid_init(void)
349 return register_md_personality(&faulty_personality
);
352 static void raid_exit(void)
354 unregister_md_personality(&faulty_personality
);
357 module_init(raid_init
);
358 module_exit(raid_exit
);
359 MODULE_LICENSE("GPL");
360 MODULE_DESCRIPTION("Fault injection personality for MD");
361 MODULE_ALIAS("md-personality-10"); /* faulty */
362 MODULE_ALIAS("md-faulty");
363 MODULE_ALIAS("md-level--5");