mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-12-13 14:04:05 +08:00
81e7706345
This commit implements processing of the REQ_OP_ZONE_RESET_ALL operation for zoned mapped devices. Given that this operation always has a BIO sector of 0 and a 0 size, processing through the regular BIO __split_and_process_bio() function does not work because this function would always select the first target. Instead, handling of this operation is implemented using the function __send_zone_reset_all(). Similarly to the __send_empty_flush() function, the new __send_zone_reset_all() function manually goes through all targets of a mapped device table doing the following: 1) If the target can natively support REQ_OP_ZONE_RESET_ALL, __send_duplicate_bios() is used to forward the reset all operation to the target. This case is handled with the __send_zone_reset_all_native() function. 2) For other targets, the function __send_zone_reset_all_emulated() is executed to emulate the execution of REQ_OP_ZONE_RESET_ALL using regular REQ_OP_ZONE_RESET operations. Targets that can natively support REQ_OP_ZONE_RESET_ALL are identified using the new target field zone_reset_all_supported. This boolean is set to true in for targets that have reliable zone limits, that is, targets that map all sequential write required zones of their zoned device(s). Setting this field is handled in dm_set_zones_restrictions() and device_get_zone_resource_limits(). For targets with unreliable zone limits, REQ_OP_ZONE_RESET_ALL must be emulated (case 2 above). This is implemented with __send_zone_reset_all_emulated() and is similar to the block layer function blkdev_zone_reset_all_emulated(): first a report zones is done for the zones of the target to identify zones that need reset, that is, any sequential write required zone that is not already empty. This is done using a bitmap and the function dm_zone_get_reset_bitmap() which sets to 1 the bit corresponding to a zone that needs reset. Next, this zone bitmap is inspected and a clone BIO modified to use the REQ_OP_ZONE_RESET operation issued for any zone with its bit set in the zone bitmap. This implementation is more efficient than what the block layer does with blkdev_zone_reset_all_emulated(), which is always used for DM zoned devices currently: as we can natively use REQ_OP_ZONE_RESET_ALL on targets mapping all sequential write required zones, resetting all zones of a zoned mapped device can be much faster compared to always emulating this operation using regular per-zone reset. In the worst case, this implementation is as-efficient as the block layer emulation. This reduction in the time it takes to reset all zones of a zoned mapped device depends directly on the mapped device targets mapping (reliable zone limits or not). Signed-off-by: Damien Le Moal <dlemoal@kernel.org> Reviewed-by: Christoph Hellwig <hch@lst.de> Reviewed-by: Johannes Thumshirn <johannes.thumshirn@wdc.com> Reviewed-by: Martin K. Petersen <martin.petersen@oracle.com> Link: https://lore.kernel.org/r/20240704052816.623865-4-dlemoal@kernel.org Signed-off-by: Jens Axboe <axboe@kernel.dk>
256 lines
7.4 KiB
C
256 lines
7.4 KiB
C
/*
|
|
* Internal header file for device mapper
|
|
*
|
|
* Copyright (C) 2001, 2002 Sistina Software
|
|
* Copyright (C) 2004-2006 Red Hat, Inc. All rights reserved.
|
|
*
|
|
* This file is released under the LGPL.
|
|
*/
|
|
|
|
#ifndef DM_INTERNAL_H
|
|
#define DM_INTERNAL_H
|
|
|
|
#include <linux/fs.h>
|
|
#include <linux/device-mapper.h>
|
|
#include <linux/list.h>
|
|
#include <linux/moduleparam.h>
|
|
#include <linux/blkdev.h>
|
|
#include <linux/backing-dev.h>
|
|
#include <linux/hdreg.h>
|
|
#include <linux/completion.h>
|
|
#include <linux/kobject.h>
|
|
#include <linux/refcount.h>
|
|
#include <linux/log2.h>
|
|
|
|
#include "dm-stats.h"
|
|
|
|
/*
|
|
* Suspend feature flags
|
|
*/
|
|
#define DM_SUSPEND_LOCKFS_FLAG (1 << 0)
|
|
#define DM_SUSPEND_NOFLUSH_FLAG (1 << 1)
|
|
|
|
/*
|
|
* Status feature flags
|
|
*/
|
|
#define DM_STATUS_NOFLUSH_FLAG (1 << 0)
|
|
|
|
/*
|
|
* List of devices that a metadevice uses and should open/close.
|
|
*/
|
|
struct dm_dev_internal {
|
|
struct list_head list;
|
|
refcount_t count;
|
|
struct dm_dev *dm_dev;
|
|
};
|
|
|
|
struct dm_table;
|
|
struct dm_md_mempools;
|
|
struct dm_target_io;
|
|
struct dm_io;
|
|
|
|
/*
|
|
*---------------------------------------------------------------
|
|
* Internal table functions.
|
|
*---------------------------------------------------------------
|
|
*/
|
|
void dm_table_event_callback(struct dm_table *t,
|
|
void (*fn)(void *), void *context);
|
|
struct dm_target *dm_table_find_target(struct dm_table *t, sector_t sector);
|
|
bool dm_table_has_no_data_devices(struct dm_table *table);
|
|
int dm_calculate_queue_limits(struct dm_table *table,
|
|
struct queue_limits *limits);
|
|
int dm_table_set_restrictions(struct dm_table *t, struct request_queue *q,
|
|
struct queue_limits *limits);
|
|
struct list_head *dm_table_get_devices(struct dm_table *t);
|
|
void dm_table_presuspend_targets(struct dm_table *t);
|
|
void dm_table_presuspend_undo_targets(struct dm_table *t);
|
|
void dm_table_postsuspend_targets(struct dm_table *t);
|
|
int dm_table_resume_targets(struct dm_table *t);
|
|
enum dm_queue_mode dm_table_get_type(struct dm_table *t);
|
|
struct target_type *dm_table_get_immutable_target_type(struct dm_table *t);
|
|
struct dm_target *dm_table_get_immutable_target(struct dm_table *t);
|
|
struct dm_target *dm_table_get_wildcard_target(struct dm_table *t);
|
|
bool dm_table_bio_based(struct dm_table *t);
|
|
bool dm_table_request_based(struct dm_table *t);
|
|
|
|
void dm_lock_md_type(struct mapped_device *md);
|
|
void dm_unlock_md_type(struct mapped_device *md);
|
|
void dm_set_md_type(struct mapped_device *md, enum dm_queue_mode type);
|
|
enum dm_queue_mode dm_get_md_type(struct mapped_device *md);
|
|
struct target_type *dm_get_immutable_target_type(struct mapped_device *md);
|
|
|
|
int dm_setup_md_queue(struct mapped_device *md, struct dm_table *t);
|
|
|
|
/*
|
|
* To check whether the target type is bio-based or not (request-based).
|
|
*/
|
|
#define dm_target_bio_based(t) ((t)->type->map != NULL)
|
|
|
|
/*
|
|
* To check whether the target type is request-based or not (bio-based).
|
|
*/
|
|
#define dm_target_request_based(t) ((t)->type->clone_and_map_rq != NULL)
|
|
|
|
/*
|
|
* To check whether the target type is a hybrid (capable of being
|
|
* either request-based or bio-based).
|
|
*/
|
|
#define dm_target_hybrid(t) (dm_target_bio_based(t) && dm_target_request_based(t))
|
|
|
|
/*
|
|
* Zoned targets related functions.
|
|
*/
|
|
int dm_set_zones_restrictions(struct dm_table *t, struct request_queue *q,
|
|
struct queue_limits *lim);
|
|
int dm_revalidate_zones(struct dm_table *t, struct request_queue *q);
|
|
void dm_zone_endio(struct dm_io *io, struct bio *clone);
|
|
#ifdef CONFIG_BLK_DEV_ZONED
|
|
int dm_blk_report_zones(struct gendisk *disk, sector_t sector,
|
|
unsigned int nr_zones, report_zones_cb cb, void *data);
|
|
bool dm_is_zone_write(struct mapped_device *md, struct bio *bio);
|
|
int dm_zone_map_bio(struct dm_target_io *io);
|
|
int dm_zone_get_reset_bitmap(struct mapped_device *md, struct dm_table *t,
|
|
sector_t sector, unsigned int nr_zones,
|
|
unsigned long *need_reset);
|
|
#else
|
|
#define dm_blk_report_zones NULL
|
|
static inline bool dm_is_zone_write(struct mapped_device *md, struct bio *bio)
|
|
{
|
|
return false;
|
|
}
|
|
static inline int dm_zone_map_bio(struct dm_target_io *tio)
|
|
{
|
|
return DM_MAPIO_KILL;
|
|
}
|
|
#endif
|
|
|
|
/*
|
|
*---------------------------------------------------------------
|
|
* A registry of target types.
|
|
*---------------------------------------------------------------
|
|
*/
|
|
int dm_target_init(void);
|
|
void dm_target_exit(void);
|
|
struct target_type *dm_get_target_type(const char *name);
|
|
void dm_put_target_type(struct target_type *tt);
|
|
int dm_target_iterate(void (*iter_func)(struct target_type *tt,
|
|
void *param), void *param);
|
|
|
|
int dm_split_args(int *argc, char ***argvp, char *input);
|
|
|
|
/*
|
|
* Is this mapped_device being deleted?
|
|
*/
|
|
int dm_deleting_md(struct mapped_device *md);
|
|
|
|
/*
|
|
* Is this mapped_device suspended?
|
|
*/
|
|
int dm_suspended_md(struct mapped_device *md);
|
|
|
|
/*
|
|
* Internal suspend and resume methods.
|
|
*/
|
|
int dm_suspended_internally_md(struct mapped_device *md);
|
|
void dm_internal_suspend_fast(struct mapped_device *md);
|
|
void dm_internal_resume_fast(struct mapped_device *md);
|
|
void dm_internal_suspend_noflush(struct mapped_device *md);
|
|
void dm_internal_resume(struct mapped_device *md);
|
|
|
|
/*
|
|
* Test if the device is scheduled for deferred remove.
|
|
*/
|
|
int dm_test_deferred_remove_flag(struct mapped_device *md);
|
|
|
|
/*
|
|
* Try to remove devices marked for deferred removal.
|
|
*/
|
|
void dm_deferred_remove(void);
|
|
|
|
/*
|
|
* The device-mapper can be driven through one of two interfaces;
|
|
* ioctl or filesystem, depending which patch you have applied.
|
|
*/
|
|
int dm_interface_init(void);
|
|
void dm_interface_exit(void);
|
|
|
|
/*
|
|
* sysfs interface
|
|
*/
|
|
int dm_sysfs_init(struct mapped_device *md);
|
|
void dm_sysfs_exit(struct mapped_device *md);
|
|
struct kobject *dm_kobject(struct mapped_device *md);
|
|
struct mapped_device *dm_get_from_kobject(struct kobject *kobj);
|
|
|
|
/*
|
|
* The kobject helper
|
|
*/
|
|
void dm_kobject_release(struct kobject *kobj);
|
|
|
|
/*
|
|
* Targets for linear and striped mappings
|
|
*/
|
|
int linear_map(struct dm_target *ti, struct bio *bio);
|
|
int dm_linear_init(void);
|
|
void dm_linear_exit(void);
|
|
|
|
int stripe_map(struct dm_target *ti, struct bio *bio);
|
|
int dm_stripe_init(void);
|
|
void dm_stripe_exit(void);
|
|
|
|
/*
|
|
* mapped_device operations
|
|
*/
|
|
void dm_destroy(struct mapped_device *md);
|
|
void dm_destroy_immediate(struct mapped_device *md);
|
|
int dm_open_count(struct mapped_device *md);
|
|
int dm_lock_for_deletion(struct mapped_device *md, bool mark_deferred, bool only_deferred);
|
|
int dm_cancel_deferred_remove(struct mapped_device *md);
|
|
int dm_request_based(struct mapped_device *md);
|
|
int dm_get_table_device(struct mapped_device *md, dev_t dev, blk_mode_t mode,
|
|
struct dm_dev **result);
|
|
void dm_put_table_device(struct mapped_device *md, struct dm_dev *d);
|
|
|
|
int dm_kobject_uevent(struct mapped_device *md, enum kobject_action action,
|
|
unsigned int cookie, bool need_resize_uevent);
|
|
|
|
int dm_io_init(void);
|
|
void dm_io_exit(void);
|
|
|
|
int dm_kcopyd_init(void);
|
|
void dm_kcopyd_exit(void);
|
|
|
|
/*
|
|
* Mempool operations
|
|
*/
|
|
void dm_free_md_mempools(struct dm_md_mempools *pools);
|
|
|
|
/*
|
|
* Various helpers
|
|
*/
|
|
unsigned int dm_get_reserved_bio_based_ios(void);
|
|
|
|
#define DM_HASH_LOCKS_MAX 64
|
|
|
|
static inline unsigned int dm_num_hash_locks(void)
|
|
{
|
|
unsigned int num_locks = roundup_pow_of_two(num_online_cpus()) << 1;
|
|
|
|
return min_t(unsigned int, num_locks, DM_HASH_LOCKS_MAX);
|
|
}
|
|
|
|
#define DM_HASH_LOCKS_MULT 4294967291ULL
|
|
#define DM_HASH_LOCKS_SHIFT 6
|
|
|
|
static inline unsigned int dm_hash_locks_index(sector_t block,
|
|
unsigned int num_locks)
|
|
{
|
|
sector_t h1 = (block * DM_HASH_LOCKS_MULT) >> DM_HASH_LOCKS_SHIFT;
|
|
sector_t h2 = h1 >> DM_HASH_LOCKS_SHIFT;
|
|
|
|
return (h1 ^ h2) & (num_locks - 1);
|
|
}
|
|
|
|
#endif
|