|  | // SPDX-License-Identifier: GPL-2.0-only | 
|  | /* | 
|  | * Copyright (C) 2001 Sistina Software (UK) Limited | 
|  | * | 
|  | * This file is released under the GPL. | 
|  | */ | 
|  |  | 
|  | #include "dm-core.h" | 
|  |  | 
|  | #include <linux/module.h> | 
|  | #include <linux/init.h> | 
|  | #include <linux/kmod.h> | 
|  | #include <linux/bio.h> | 
|  | #include <linux/dax.h> | 
|  |  | 
|  | #define DM_MSG_PREFIX "target" | 
|  |  | 
|  | static LIST_HEAD(_targets); | 
|  | static DECLARE_RWSEM(_lock); | 
|  |  | 
|  | static inline struct target_type *__find_target_type(const char *name) | 
|  | { | 
|  | struct target_type *tt; | 
|  |  | 
|  | list_for_each_entry(tt, &_targets, list) | 
|  | if (!strcmp(name, tt->name)) | 
|  | return tt; | 
|  |  | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | static struct target_type *get_target_type(const char *name) | 
|  | { | 
|  | struct target_type *tt; | 
|  |  | 
|  | down_read(&_lock); | 
|  |  | 
|  | tt = __find_target_type(name); | 
|  | if (tt && !try_module_get(tt->module)) | 
|  | tt = NULL; | 
|  |  | 
|  | up_read(&_lock); | 
|  | return tt; | 
|  | } | 
|  |  | 
|  | static void load_module(const char *name) | 
|  | { | 
|  | request_module("dm-%s", name); | 
|  | } | 
|  |  | 
|  | struct target_type *dm_get_target_type(const char *name) | 
|  | { | 
|  | struct target_type *tt = get_target_type(name); | 
|  |  | 
|  | if (!tt) { | 
|  | load_module(name); | 
|  | tt = get_target_type(name); | 
|  | } | 
|  |  | 
|  | return tt; | 
|  | } | 
|  |  | 
|  | void dm_put_target_type(struct target_type *tt) | 
|  | { | 
|  | down_read(&_lock); | 
|  | module_put(tt->module); | 
|  | up_read(&_lock); | 
|  | } | 
|  |  | 
|  | int dm_target_iterate(void (*iter_func)(struct target_type *tt, | 
|  | void *param), void *param) | 
|  | { | 
|  | struct target_type *tt; | 
|  |  | 
|  | down_read(&_lock); | 
|  | list_for_each_entry(tt, &_targets, list) | 
|  | iter_func(tt, param); | 
|  | up_read(&_lock); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | int dm_register_target(struct target_type *tt) | 
|  | { | 
|  | int rv = 0; | 
|  |  | 
|  | down_write(&_lock); | 
|  | if (__find_target_type(tt->name)) { | 
|  | DMERR("%s: '%s' target already registered", | 
|  | __func__, tt->name); | 
|  | rv = -EEXIST; | 
|  | } else { | 
|  | list_add(&tt->list, &_targets); | 
|  | } | 
|  | up_write(&_lock); | 
|  |  | 
|  | return rv; | 
|  | } | 
|  | EXPORT_SYMBOL(dm_register_target); | 
|  |  | 
|  | void dm_unregister_target(struct target_type *tt) | 
|  | { | 
|  | down_write(&_lock); | 
|  | if (!__find_target_type(tt->name)) { | 
|  | DMCRIT("Unregistering unrecognised target: %s", tt->name); | 
|  | BUG(); | 
|  | } | 
|  |  | 
|  | list_del(&tt->list); | 
|  |  | 
|  | up_write(&_lock); | 
|  | } | 
|  | EXPORT_SYMBOL(dm_unregister_target); | 
|  |  | 
|  | /* | 
|  | * io-err: always fails an io, useful for bringing | 
|  | * up LVs that have holes in them. | 
|  | */ | 
|  | struct io_err_c { | 
|  | struct dm_dev *dev; | 
|  | sector_t start; | 
|  | }; | 
|  |  | 
|  | static int io_err_get_args(struct dm_target *tt, unsigned int argc, char **args) | 
|  | { | 
|  | unsigned long long start; | 
|  | struct io_err_c *ioec; | 
|  | char dummy; | 
|  | int ret; | 
|  |  | 
|  | ioec = kmalloc(sizeof(*ioec), GFP_KERNEL); | 
|  | if (!ioec) { | 
|  | tt->error = "Cannot allocate io_err context"; | 
|  | return -ENOMEM; | 
|  | } | 
|  |  | 
|  | ret = -EINVAL; | 
|  | if (sscanf(args[1], "%llu%c", &start, &dummy) != 1 || | 
|  | start != (sector_t)start) { | 
|  | tt->error = "Invalid device sector"; | 
|  | goto bad; | 
|  | } | 
|  | ioec->start = start; | 
|  |  | 
|  | ret = dm_get_device(tt, args[0], dm_table_get_mode(tt->table), &ioec->dev); | 
|  | if (ret) { | 
|  | tt->error = "Device lookup failed"; | 
|  | goto bad; | 
|  | } | 
|  |  | 
|  | tt->private = ioec; | 
|  |  | 
|  | return 0; | 
|  |  | 
|  | bad: | 
|  | kfree(ioec); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int io_err_ctr(struct dm_target *tt, unsigned int argc, char **args) | 
|  | { | 
|  | /* | 
|  | * If we have arguments, assume it is the path to the backing | 
|  | * block device and its mapping start sector (same as dm-linear). | 
|  | * In this case, get the device so that we can get its limits. | 
|  | */ | 
|  | if (argc == 2) { | 
|  | int ret = io_err_get_args(tt, argc, args); | 
|  |  | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Return error for discards instead of -EOPNOTSUPP | 
|  | */ | 
|  | tt->num_discard_bios = 1; | 
|  | tt->discards_supported = true; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void io_err_dtr(struct dm_target *tt) | 
|  | { | 
|  | struct io_err_c *ioec = tt->private; | 
|  |  | 
|  | if (ioec) { | 
|  | dm_put_device(tt, ioec->dev); | 
|  | kfree(ioec); | 
|  | } | 
|  | } | 
|  |  | 
|  | static int io_err_map(struct dm_target *tt, struct bio *bio) | 
|  | { | 
|  | return DM_MAPIO_KILL; | 
|  | } | 
|  |  | 
|  | static int io_err_clone_and_map_rq(struct dm_target *ti, struct request *rq, | 
|  | union map_info *map_context, | 
|  | struct request **clone) | 
|  | { | 
|  | return DM_MAPIO_KILL; | 
|  | } | 
|  |  | 
|  | static void io_err_release_clone_rq(struct request *clone, | 
|  | union map_info *map_context) | 
|  | { | 
|  | } | 
|  |  | 
|  | #ifdef CONFIG_BLK_DEV_ZONED | 
|  | static sector_t io_err_map_sector(struct dm_target *ti, sector_t bi_sector) | 
|  | { | 
|  | struct io_err_c *ioec = ti->private; | 
|  |  | 
|  | return ioec->start + dm_target_offset(ti, bi_sector); | 
|  | } | 
|  |  | 
|  | static int io_err_report_zones(struct dm_target *ti, | 
|  | struct dm_report_zones_args *args, unsigned int nr_zones) | 
|  | { | 
|  | struct io_err_c *ioec = ti->private; | 
|  |  | 
|  | /* | 
|  | * This should never be called when we do not have a backing device | 
|  | * as that mean the target is not a zoned one. | 
|  | */ | 
|  | if (WARN_ON_ONCE(!ioec)) | 
|  | return -EIO; | 
|  |  | 
|  | return dm_report_zones(ioec->dev->bdev, ioec->start, | 
|  | io_err_map_sector(ti, args->next_sector), | 
|  | args, nr_zones); | 
|  | } | 
|  | #else | 
|  | #define io_err_report_zones NULL | 
|  | #endif | 
|  |  | 
|  | static int io_err_iterate_devices(struct dm_target *ti, | 
|  | iterate_devices_callout_fn fn, void *data) | 
|  | { | 
|  | struct io_err_c *ioec = ti->private; | 
|  |  | 
|  | if (!ioec) | 
|  | return 0; | 
|  |  | 
|  | return fn(ti, ioec->dev, ioec->start, ti->len, data); | 
|  | } | 
|  |  | 
|  | static void io_err_io_hints(struct dm_target *ti, struct queue_limits *limits) | 
|  | { | 
|  | limits->max_hw_discard_sectors = UINT_MAX; | 
|  | limits->discard_granularity = 512; | 
|  | } | 
|  |  | 
|  | static long io_err_dax_direct_access(struct dm_target *ti, pgoff_t pgoff, | 
|  | long nr_pages, enum dax_access_mode mode, void **kaddr, | 
|  | pfn_t *pfn) | 
|  | { | 
|  | return -EIO; | 
|  | } | 
|  |  | 
|  | static struct target_type error_target = { | 
|  | .name = "error", | 
|  | .version = {1, 7, 0}, | 
|  | .features = DM_TARGET_WILDCARD | DM_TARGET_ZONED_HM, | 
|  | .ctr  = io_err_ctr, | 
|  | .dtr  = io_err_dtr, | 
|  | .map  = io_err_map, | 
|  | .clone_and_map_rq = io_err_clone_and_map_rq, | 
|  | .release_clone_rq = io_err_release_clone_rq, | 
|  | .iterate_devices = io_err_iterate_devices, | 
|  | .io_hints = io_err_io_hints, | 
|  | .direct_access = io_err_dax_direct_access, | 
|  | .report_zones = io_err_report_zones, | 
|  | }; | 
|  |  | 
|  | int __init dm_target_init(void) | 
|  | { | 
|  | return dm_register_target(&error_target); | 
|  | } | 
|  |  | 
|  | void dm_target_exit(void) | 
|  | { | 
|  | dm_unregister_target(&error_target); | 
|  | } |