|  | // SPDX-License-Identifier: GPL-2.0-only | 
|  | /* | 
|  | * Copyright (C) 2003 Sistina Software (UK) Limited. | 
|  | * Copyright (C) 2004, 2010-2011 Red Hat, Inc. All rights reserved. | 
|  | * | 
|  | * This file is released under the GPL. | 
|  | */ | 
|  |  | 
|  | #include <linux/device-mapper.h> | 
|  |  | 
|  | #include <linux/module.h> | 
|  | #include <linux/init.h> | 
|  | #include <linux/blkdev.h> | 
|  | #include <linux/bio.h> | 
|  | #include <linux/slab.h> | 
|  |  | 
|  | #define DM_MSG_PREFIX "flakey" | 
|  |  | 
|  | #define PROBABILITY_BASE	1000000000 | 
|  |  | 
|  | #define all_corrupt_bio_flags_match(bio, fc)	\ | 
|  | (((bio)->bi_opf & (fc)->corrupt_bio_flags) == (fc)->corrupt_bio_flags) | 
|  |  | 
|  | /* | 
|  | * Flakey: Used for testing only, simulates intermittent, | 
|  | * catastrophic device failure. | 
|  | */ | 
|  | struct flakey_c { | 
|  | struct dm_dev *dev; | 
|  | unsigned long start_time; | 
|  | sector_t start; | 
|  | unsigned int up_interval; | 
|  | unsigned int down_interval; | 
|  | unsigned long flags; | 
|  | unsigned int corrupt_bio_byte; | 
|  | unsigned int corrupt_bio_rw; | 
|  | unsigned int corrupt_bio_value; | 
|  | blk_opf_t corrupt_bio_flags; | 
|  | unsigned int random_read_corrupt; | 
|  | unsigned int random_write_corrupt; | 
|  | }; | 
|  |  | 
|  | enum feature_flag_bits { | 
|  | ERROR_READS, | 
|  | DROP_WRITES, | 
|  | ERROR_WRITES | 
|  | }; | 
|  |  | 
|  | struct per_bio_data { | 
|  | bool bio_submitted; | 
|  | }; | 
|  |  | 
|  | static int parse_features(struct dm_arg_set *as, struct flakey_c *fc, | 
|  | struct dm_target *ti) | 
|  | { | 
|  | int r; | 
|  | unsigned int argc; | 
|  | const char *arg_name; | 
|  |  | 
|  | static const struct dm_arg _args[] = { | 
|  | {0, 11, "Invalid number of feature args"}, | 
|  | {1, UINT_MAX, "Invalid corrupt bio byte"}, | 
|  | {0, 255, "Invalid corrupt value to write into bio byte (0-255)"}, | 
|  | {0, UINT_MAX, "Invalid corrupt bio flags mask"}, | 
|  | {0, PROBABILITY_BASE, "Invalid random corrupt argument"}, | 
|  | }; | 
|  |  | 
|  | /* No feature arguments supplied. */ | 
|  | if (!as->argc) | 
|  | return 0; | 
|  |  | 
|  | r = dm_read_arg_group(_args, as, &argc, &ti->error); | 
|  | if (r) | 
|  | return r; | 
|  |  | 
|  | while (argc) { | 
|  | arg_name = dm_shift_arg(as); | 
|  | argc--; | 
|  |  | 
|  | if (!arg_name) { | 
|  | ti->error = "Insufficient feature arguments"; | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * error_reads | 
|  | */ | 
|  | if (!strcasecmp(arg_name, "error_reads")) { | 
|  | if (test_and_set_bit(ERROR_READS, &fc->flags)) { | 
|  | ti->error = "Feature error_reads duplicated"; | 
|  | return -EINVAL; | 
|  | } | 
|  | continue; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * drop_writes | 
|  | */ | 
|  | if (!strcasecmp(arg_name, "drop_writes")) { | 
|  | if (test_and_set_bit(DROP_WRITES, &fc->flags)) { | 
|  | ti->error = "Feature drop_writes duplicated"; | 
|  | return -EINVAL; | 
|  | } else if (test_bit(ERROR_WRITES, &fc->flags)) { | 
|  | ti->error = "Feature drop_writes conflicts with feature error_writes"; | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | continue; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * error_writes | 
|  | */ | 
|  | if (!strcasecmp(arg_name, "error_writes")) { | 
|  | if (test_and_set_bit(ERROR_WRITES, &fc->flags)) { | 
|  | ti->error = "Feature error_writes duplicated"; | 
|  | return -EINVAL; | 
|  |  | 
|  | } else if (test_bit(DROP_WRITES, &fc->flags)) { | 
|  | ti->error = "Feature error_writes conflicts with feature drop_writes"; | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | continue; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * corrupt_bio_byte <Nth_byte> <direction> <value> <bio_flags> | 
|  | */ | 
|  | if (!strcasecmp(arg_name, "corrupt_bio_byte")) { | 
|  | if (!argc) { | 
|  | ti->error = "Feature corrupt_bio_byte requires parameters"; | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | r = dm_read_arg(_args + 1, as, &fc->corrupt_bio_byte, &ti->error); | 
|  | if (r) | 
|  | return r; | 
|  | argc--; | 
|  |  | 
|  | /* | 
|  | * Direction r or w? | 
|  | */ | 
|  | arg_name = dm_shift_arg(as); | 
|  | if (arg_name && !strcasecmp(arg_name, "w")) | 
|  | fc->corrupt_bio_rw = WRITE; | 
|  | else if (arg_name && !strcasecmp(arg_name, "r")) | 
|  | fc->corrupt_bio_rw = READ; | 
|  | else { | 
|  | ti->error = "Invalid corrupt bio direction (r or w)"; | 
|  | return -EINVAL; | 
|  | } | 
|  | argc--; | 
|  |  | 
|  | /* | 
|  | * Value of byte (0-255) to write in place of correct one. | 
|  | */ | 
|  | r = dm_read_arg(_args + 2, as, &fc->corrupt_bio_value, &ti->error); | 
|  | if (r) | 
|  | return r; | 
|  | argc--; | 
|  |  | 
|  | /* | 
|  | * Only corrupt bios with these flags set. | 
|  | */ | 
|  | BUILD_BUG_ON(sizeof(fc->corrupt_bio_flags) != | 
|  | sizeof(unsigned int)); | 
|  | r = dm_read_arg(_args + 3, as, | 
|  | (__force unsigned int *)&fc->corrupt_bio_flags, | 
|  | &ti->error); | 
|  | if (r) | 
|  | return r; | 
|  | argc--; | 
|  |  | 
|  | continue; | 
|  | } | 
|  |  | 
|  | if (!strcasecmp(arg_name, "random_read_corrupt")) { | 
|  | if (!argc) { | 
|  | ti->error = "Feature random_read_corrupt requires a parameter"; | 
|  | return -EINVAL; | 
|  | } | 
|  | r = dm_read_arg(_args + 4, as, &fc->random_read_corrupt, &ti->error); | 
|  | if (r) | 
|  | return r; | 
|  | argc--; | 
|  |  | 
|  | continue; | 
|  | } | 
|  |  | 
|  | if (!strcasecmp(arg_name, "random_write_corrupt")) { | 
|  | if (!argc) { | 
|  | ti->error = "Feature random_write_corrupt requires a parameter"; | 
|  | return -EINVAL; | 
|  | } | 
|  | r = dm_read_arg(_args + 4, as, &fc->random_write_corrupt, &ti->error); | 
|  | if (r) | 
|  | return r; | 
|  | argc--; | 
|  |  | 
|  | continue; | 
|  | } | 
|  |  | 
|  | ti->error = "Unrecognised flakey feature requested"; | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | if (test_bit(DROP_WRITES, &fc->flags) && (fc->corrupt_bio_rw == WRITE)) { | 
|  | ti->error = "drop_writes is incompatible with corrupt_bio_byte with the WRITE flag set"; | 
|  | return -EINVAL; | 
|  |  | 
|  | } else if (test_bit(ERROR_WRITES, &fc->flags) && (fc->corrupt_bio_rw == WRITE)) { | 
|  | ti->error = "error_writes is incompatible with corrupt_bio_byte with the WRITE flag set"; | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | if (!fc->corrupt_bio_byte && !test_bit(ERROR_READS, &fc->flags) && | 
|  | !test_bit(DROP_WRITES, &fc->flags) && !test_bit(ERROR_WRITES, &fc->flags) && | 
|  | !fc->random_read_corrupt && !fc->random_write_corrupt) { | 
|  | set_bit(ERROR_WRITES, &fc->flags); | 
|  | set_bit(ERROR_READS, &fc->flags); | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Construct a flakey mapping: | 
|  | * <dev_path> <offset> <up interval> <down interval> [<#feature args> [<arg>]*] | 
|  | * | 
|  | *   Feature args: | 
|  | *     [drop_writes] | 
|  | *     [corrupt_bio_byte <Nth_byte> <direction> <value> <bio_flags>] | 
|  | * | 
|  | *   Nth_byte starts from 1 for the first byte. | 
|  | *   Direction is r for READ or w for WRITE. | 
|  | *   bio_flags is ignored if 0. | 
|  | */ | 
|  | static int flakey_ctr(struct dm_target *ti, unsigned int argc, char **argv) | 
|  | { | 
|  | static const struct dm_arg _args[] = { | 
|  | {0, UINT_MAX, "Invalid up interval"}, | 
|  | {0, UINT_MAX, "Invalid down interval"}, | 
|  | }; | 
|  |  | 
|  | int r; | 
|  | struct flakey_c *fc; | 
|  | unsigned long long tmpll; | 
|  | struct dm_arg_set as; | 
|  | const char *devname; | 
|  | char dummy; | 
|  |  | 
|  | as.argc = argc; | 
|  | as.argv = argv; | 
|  |  | 
|  | if (argc < 4) { | 
|  | ti->error = "Invalid argument count"; | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | fc = kzalloc(sizeof(*fc), GFP_KERNEL); | 
|  | if (!fc) { | 
|  | ti->error = "Cannot allocate context"; | 
|  | return -ENOMEM; | 
|  | } | 
|  | fc->start_time = jiffies; | 
|  |  | 
|  | devname = dm_shift_arg(&as); | 
|  |  | 
|  | r = -EINVAL; | 
|  | if (sscanf(dm_shift_arg(&as), "%llu%c", &tmpll, &dummy) != 1 || tmpll != (sector_t)tmpll) { | 
|  | ti->error = "Invalid device sector"; | 
|  | goto bad; | 
|  | } | 
|  | fc->start = tmpll; | 
|  |  | 
|  | r = dm_read_arg(_args, &as, &fc->up_interval, &ti->error); | 
|  | if (r) | 
|  | goto bad; | 
|  |  | 
|  | r = dm_read_arg(_args, &as, &fc->down_interval, &ti->error); | 
|  | if (r) | 
|  | goto bad; | 
|  |  | 
|  | if (!(fc->up_interval + fc->down_interval)) { | 
|  | ti->error = "Total (up + down) interval is zero"; | 
|  | r = -EINVAL; | 
|  | goto bad; | 
|  | } | 
|  |  | 
|  | if (fc->up_interval + fc->down_interval < fc->up_interval) { | 
|  | ti->error = "Interval overflow"; | 
|  | r = -EINVAL; | 
|  | goto bad; | 
|  | } | 
|  |  | 
|  | r = parse_features(&as, fc, ti); | 
|  | if (r) | 
|  | goto bad; | 
|  |  | 
|  | r = dm_get_device(ti, devname, dm_table_get_mode(ti->table), &fc->dev); | 
|  | if (r) { | 
|  | ti->error = "Device lookup failed"; | 
|  | goto bad; | 
|  | } | 
|  |  | 
|  | ti->num_flush_bios = 1; | 
|  | ti->num_discard_bios = 1; | 
|  | ti->per_io_data_size = sizeof(struct per_bio_data); | 
|  | ti->private = fc; | 
|  | return 0; | 
|  |  | 
|  | bad: | 
|  | kfree(fc); | 
|  | return r; | 
|  | } | 
|  |  | 
|  | static void flakey_dtr(struct dm_target *ti) | 
|  | { | 
|  | struct flakey_c *fc = ti->private; | 
|  |  | 
|  | dm_put_device(ti, fc->dev); | 
|  | kfree(fc); | 
|  | } | 
|  |  | 
|  | static sector_t flakey_map_sector(struct dm_target *ti, sector_t bi_sector) | 
|  | { | 
|  | struct flakey_c *fc = ti->private; | 
|  |  | 
|  | return fc->start + dm_target_offset(ti, bi_sector); | 
|  | } | 
|  |  | 
|  | static void flakey_map_bio(struct dm_target *ti, struct bio *bio) | 
|  | { | 
|  | struct flakey_c *fc = ti->private; | 
|  |  | 
|  | bio_set_dev(bio, fc->dev->bdev); | 
|  | bio->bi_iter.bi_sector = flakey_map_sector(ti, bio->bi_iter.bi_sector); | 
|  | } | 
|  |  | 
|  | static void corrupt_bio_common(struct bio *bio, unsigned int corrupt_bio_byte, | 
|  | unsigned char corrupt_bio_value) | 
|  | { | 
|  | struct bvec_iter iter; | 
|  | struct bio_vec bvec; | 
|  |  | 
|  | /* | 
|  | * Overwrite the Nth byte of the bio's data, on whichever page | 
|  | * it falls. | 
|  | */ | 
|  | bio_for_each_segment(bvec, bio, iter) { | 
|  | if (bio_iter_len(bio, iter) > corrupt_bio_byte) { | 
|  | unsigned char *segment = bvec_kmap_local(&bvec); | 
|  | segment[corrupt_bio_byte] = corrupt_bio_value; | 
|  | kunmap_local(segment); | 
|  | DMDEBUG("Corrupting data bio=%p by writing %u to byte %u " | 
|  | "(rw=%c bi_opf=%u bi_sector=%llu size=%u)\n", | 
|  | bio, corrupt_bio_value, corrupt_bio_byte, | 
|  | (bio_data_dir(bio) == WRITE) ? 'w' : 'r', bio->bi_opf, | 
|  | (unsigned long long)bio->bi_iter.bi_sector, | 
|  | bio->bi_iter.bi_size); | 
|  | break; | 
|  | } | 
|  | corrupt_bio_byte -= bio_iter_len(bio, iter); | 
|  | } | 
|  | } | 
|  |  | 
|  | static void corrupt_bio_data(struct bio *bio, struct flakey_c *fc) | 
|  | { | 
|  | unsigned int corrupt_bio_byte = fc->corrupt_bio_byte - 1; | 
|  |  | 
|  | if (!bio_has_data(bio)) | 
|  | return; | 
|  |  | 
|  | corrupt_bio_common(bio, corrupt_bio_byte, fc->corrupt_bio_value); | 
|  | } | 
|  |  | 
|  | static void corrupt_bio_random(struct bio *bio) | 
|  | { | 
|  | unsigned int corrupt_byte; | 
|  | unsigned char corrupt_value; | 
|  |  | 
|  | if (!bio_has_data(bio)) | 
|  | return; | 
|  |  | 
|  | corrupt_byte = get_random_u32() % bio->bi_iter.bi_size; | 
|  | corrupt_value = get_random_u8(); | 
|  |  | 
|  | corrupt_bio_common(bio, corrupt_byte, corrupt_value); | 
|  | } | 
|  |  | 
|  | static void clone_free(struct bio *clone) | 
|  | { | 
|  | struct folio_iter fi; | 
|  |  | 
|  | if (clone->bi_vcnt > 0) { /* bio_for_each_folio_all crashes with an empty bio */ | 
|  | bio_for_each_folio_all(fi, clone) | 
|  | folio_put(fi.folio); | 
|  | } | 
|  |  | 
|  | bio_uninit(clone); | 
|  | kfree(clone); | 
|  | } | 
|  |  | 
|  | static void clone_endio(struct bio *clone) | 
|  | { | 
|  | struct bio *bio = clone->bi_private; | 
|  | bio->bi_status = clone->bi_status; | 
|  | clone_free(clone); | 
|  | bio_endio(bio); | 
|  | } | 
|  |  | 
|  | static struct bio *clone_bio(struct dm_target *ti, struct flakey_c *fc, struct bio *bio) | 
|  | { | 
|  | struct bio *clone; | 
|  | unsigned size, remaining_size, nr_iovecs, order; | 
|  | struct bvec_iter iter = bio->bi_iter; | 
|  |  | 
|  | if (unlikely(bio->bi_iter.bi_size > UIO_MAXIOV << PAGE_SHIFT)) | 
|  | dm_accept_partial_bio(bio, UIO_MAXIOV << PAGE_SHIFT >> SECTOR_SHIFT); | 
|  |  | 
|  | size = bio->bi_iter.bi_size; | 
|  | nr_iovecs = (size + PAGE_SIZE - 1) >> PAGE_SHIFT; | 
|  |  | 
|  | clone = bio_kmalloc(nr_iovecs, GFP_NOIO | __GFP_NORETRY | __GFP_NOWARN); | 
|  | if (!clone) | 
|  | return NULL; | 
|  |  | 
|  | bio_init(clone, fc->dev->bdev, bio->bi_inline_vecs, nr_iovecs, bio->bi_opf); | 
|  |  | 
|  | clone->bi_iter.bi_sector = flakey_map_sector(ti, bio->bi_iter.bi_sector); | 
|  | clone->bi_private = bio; | 
|  | clone->bi_end_io = clone_endio; | 
|  |  | 
|  | remaining_size = size; | 
|  |  | 
|  | order = MAX_PAGE_ORDER; | 
|  | while (remaining_size) { | 
|  | struct page *pages; | 
|  | unsigned size_to_add, to_copy; | 
|  | unsigned char *virt; | 
|  | unsigned remaining_order = __fls((remaining_size + PAGE_SIZE - 1) >> PAGE_SHIFT); | 
|  | order = min(order, remaining_order); | 
|  |  | 
|  | retry_alloc_pages: | 
|  | pages = alloc_pages(GFP_NOIO | __GFP_NORETRY | __GFP_NOWARN | __GFP_COMP, order); | 
|  | if (unlikely(!pages)) { | 
|  | if (order) { | 
|  | order--; | 
|  | goto retry_alloc_pages; | 
|  | } | 
|  | clone_free(clone); | 
|  | return NULL; | 
|  | } | 
|  | size_to_add = min((unsigned)PAGE_SIZE << order, remaining_size); | 
|  |  | 
|  | virt = page_to_virt(pages); | 
|  | to_copy = size_to_add; | 
|  | do { | 
|  | struct bio_vec bvec = bvec_iter_bvec(bio->bi_io_vec, iter); | 
|  | unsigned this_step = min(bvec.bv_len, to_copy); | 
|  | void *map = bvec_kmap_local(&bvec); | 
|  | memcpy(virt, map, this_step); | 
|  | kunmap_local(map); | 
|  |  | 
|  | bvec_iter_advance(bio->bi_io_vec, &iter, this_step); | 
|  | to_copy -= this_step; | 
|  | virt += this_step; | 
|  | } while (to_copy); | 
|  |  | 
|  | __bio_add_page(clone, pages, size_to_add, 0); | 
|  | remaining_size -= size_to_add; | 
|  | } | 
|  |  | 
|  | return clone; | 
|  | } | 
|  |  | 
|  | static int flakey_map(struct dm_target *ti, struct bio *bio) | 
|  | { | 
|  | struct flakey_c *fc = ti->private; | 
|  | unsigned int elapsed; | 
|  | struct per_bio_data *pb = dm_per_bio_data(bio, sizeof(struct per_bio_data)); | 
|  |  | 
|  | pb->bio_submitted = false; | 
|  |  | 
|  | if (op_is_zone_mgmt(bio_op(bio))) | 
|  | goto map_bio; | 
|  |  | 
|  | /* Are we alive ? */ | 
|  | elapsed = (jiffies - fc->start_time) / HZ; | 
|  | if (elapsed % (fc->up_interval + fc->down_interval) >= fc->up_interval) { | 
|  | bool corrupt_fixed, corrupt_random; | 
|  | /* | 
|  | * Flag this bio as submitted while down. | 
|  | */ | 
|  | pb->bio_submitted = true; | 
|  |  | 
|  | /* | 
|  | * Error reads if neither corrupt_bio_byte or drop_writes or error_writes are set. | 
|  | * Otherwise, flakey_end_io() will decide if the reads should be modified. | 
|  | */ | 
|  | if (bio_data_dir(bio) == READ) { | 
|  | if (test_bit(ERROR_READS, &fc->flags)) | 
|  | return DM_MAPIO_KILL; | 
|  | goto map_bio; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Drop or error writes? | 
|  | */ | 
|  | if (test_bit(DROP_WRITES, &fc->flags)) { | 
|  | bio_endio(bio); | 
|  | return DM_MAPIO_SUBMITTED; | 
|  | } else if (test_bit(ERROR_WRITES, &fc->flags)) { | 
|  | bio_io_error(bio); | 
|  | return DM_MAPIO_SUBMITTED; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Corrupt matching writes. | 
|  | */ | 
|  | corrupt_fixed = false; | 
|  | corrupt_random = false; | 
|  | if (fc->corrupt_bio_byte && fc->corrupt_bio_rw == WRITE) { | 
|  | if (all_corrupt_bio_flags_match(bio, fc)) | 
|  | corrupt_fixed = true; | 
|  | } | 
|  | if (fc->random_write_corrupt) { | 
|  | u64 rnd = get_random_u64(); | 
|  | u32 rem = do_div(rnd, PROBABILITY_BASE); | 
|  | if (rem < fc->random_write_corrupt) | 
|  | corrupt_random = true; | 
|  | } | 
|  | if (corrupt_fixed || corrupt_random) { | 
|  | struct bio *clone = clone_bio(ti, fc, bio); | 
|  | if (clone) { | 
|  | if (corrupt_fixed) | 
|  | corrupt_bio_data(clone, fc); | 
|  | if (corrupt_random) | 
|  | corrupt_bio_random(clone); | 
|  | submit_bio(clone); | 
|  | return DM_MAPIO_SUBMITTED; | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | map_bio: | 
|  | flakey_map_bio(ti, bio); | 
|  |  | 
|  | return DM_MAPIO_REMAPPED; | 
|  | } | 
|  |  | 
|  | static int flakey_end_io(struct dm_target *ti, struct bio *bio, | 
|  | blk_status_t *error) | 
|  | { | 
|  | struct flakey_c *fc = ti->private; | 
|  | struct per_bio_data *pb = dm_per_bio_data(bio, sizeof(struct per_bio_data)); | 
|  |  | 
|  | if (op_is_zone_mgmt(bio_op(bio))) | 
|  | return DM_ENDIO_DONE; | 
|  |  | 
|  | if (!*error && pb->bio_submitted && (bio_data_dir(bio) == READ)) { | 
|  | if (fc->corrupt_bio_byte) { | 
|  | if ((fc->corrupt_bio_rw == READ) && | 
|  | all_corrupt_bio_flags_match(bio, fc)) { | 
|  | /* | 
|  | * Corrupt successful matching READs while in down state. | 
|  | */ | 
|  | corrupt_bio_data(bio, fc); | 
|  | } | 
|  | } | 
|  | if (fc->random_read_corrupt) { | 
|  | u64 rnd = get_random_u64(); | 
|  | u32 rem = do_div(rnd, PROBABILITY_BASE); | 
|  | if (rem < fc->random_read_corrupt) | 
|  | corrupt_bio_random(bio); | 
|  | } | 
|  | if (test_bit(ERROR_READS, &fc->flags)) { | 
|  | /* | 
|  | * Error read during the down_interval if drop_writes | 
|  | * and error_writes were not configured. | 
|  | */ | 
|  | *error = BLK_STS_IOERR; | 
|  | } | 
|  | } | 
|  |  | 
|  | return DM_ENDIO_DONE; | 
|  | } | 
|  |  | 
|  | static void flakey_status(struct dm_target *ti, status_type_t type, | 
|  | unsigned int status_flags, char *result, unsigned int maxlen) | 
|  | { | 
|  | unsigned int sz = 0; | 
|  | struct flakey_c *fc = ti->private; | 
|  | unsigned int error_reads, drop_writes, error_writes; | 
|  |  | 
|  | switch (type) { | 
|  | case STATUSTYPE_INFO: | 
|  | result[0] = '\0'; | 
|  | break; | 
|  |  | 
|  | case STATUSTYPE_TABLE: | 
|  | DMEMIT("%s %llu %u %u", fc->dev->name, | 
|  | (unsigned long long)fc->start, fc->up_interval, | 
|  | fc->down_interval); | 
|  |  | 
|  | error_reads = test_bit(ERROR_READS, &fc->flags); | 
|  | drop_writes = test_bit(DROP_WRITES, &fc->flags); | 
|  | error_writes = test_bit(ERROR_WRITES, &fc->flags); | 
|  | DMEMIT(" %u", error_reads + drop_writes + error_writes + | 
|  | (fc->corrupt_bio_byte > 0) * 5 + | 
|  | (fc->random_read_corrupt > 0) * 2 + | 
|  | (fc->random_write_corrupt > 0) * 2); | 
|  |  | 
|  | if (error_reads) | 
|  | DMEMIT(" error_reads"); | 
|  | if (drop_writes) | 
|  | DMEMIT(" drop_writes"); | 
|  | else if (error_writes) | 
|  | DMEMIT(" error_writes"); | 
|  |  | 
|  | if (fc->corrupt_bio_byte) | 
|  | DMEMIT(" corrupt_bio_byte %u %c %u %u", | 
|  | fc->corrupt_bio_byte, | 
|  | (fc->corrupt_bio_rw == WRITE) ? 'w' : 'r', | 
|  | fc->corrupt_bio_value, fc->corrupt_bio_flags); | 
|  |  | 
|  | if (fc->random_read_corrupt > 0) | 
|  | DMEMIT(" random_read_corrupt %u", fc->random_read_corrupt); | 
|  | if (fc->random_write_corrupt > 0) | 
|  | DMEMIT(" random_write_corrupt %u", fc->random_write_corrupt); | 
|  |  | 
|  | break; | 
|  |  | 
|  | case STATUSTYPE_IMA: | 
|  | result[0] = '\0'; | 
|  | break; | 
|  | } | 
|  | } | 
|  |  | 
|  | static int flakey_prepare_ioctl(struct dm_target *ti, struct block_device **bdev) | 
|  | { | 
|  | struct flakey_c *fc = ti->private; | 
|  |  | 
|  | *bdev = fc->dev->bdev; | 
|  |  | 
|  | /* | 
|  | * Only pass ioctls through if the device sizes match exactly. | 
|  | */ | 
|  | if (fc->start || ti->len != bdev_nr_sectors((*bdev))) | 
|  | return 1; | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | #ifdef CONFIG_BLK_DEV_ZONED | 
|  | static int flakey_report_zones(struct dm_target *ti, | 
|  | struct dm_report_zones_args *args, unsigned int nr_zones) | 
|  | { | 
|  | struct flakey_c *fc = ti->private; | 
|  |  | 
|  | return dm_report_zones(fc->dev->bdev, fc->start, | 
|  | flakey_map_sector(ti, args->next_sector), | 
|  | args, nr_zones); | 
|  | } | 
|  | #else | 
|  | #define flakey_report_zones NULL | 
|  | #endif | 
|  |  | 
|  | static int flakey_iterate_devices(struct dm_target *ti, iterate_devices_callout_fn fn, void *data) | 
|  | { | 
|  | struct flakey_c *fc = ti->private; | 
|  |  | 
|  | return fn(ti, fc->dev, fc->start, ti->len, data); | 
|  | } | 
|  |  | 
|  | static struct target_type flakey_target = { | 
|  | .name   = "flakey", | 
|  | .version = {1, 5, 0}, | 
|  | .features = DM_TARGET_ZONED_HM | DM_TARGET_PASSES_CRYPTO, | 
|  | .report_zones = flakey_report_zones, | 
|  | .module = THIS_MODULE, | 
|  | .ctr    = flakey_ctr, | 
|  | .dtr    = flakey_dtr, | 
|  | .map    = flakey_map, | 
|  | .end_io = flakey_end_io, | 
|  | .status = flakey_status, | 
|  | .prepare_ioctl = flakey_prepare_ioctl, | 
|  | .iterate_devices = flakey_iterate_devices, | 
|  | }; | 
|  | module_dm(flakey); | 
|  |  | 
|  | MODULE_DESCRIPTION(DM_NAME " flakey target"); | 
|  | MODULE_AUTHOR("Joe Thornber <dm-devel@redhat.com>"); | 
|  | MODULE_LICENSE("GPL"); |