mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-16 08:44:21 +08:00
8ec456629d
For device mapper targets to take advantage of IMA's measurement capabilities, the status functions for the individual targets need to be updated to handle the status_type_t case for value STATUSTYPE_IMA. Update status functions for the following target types, to log their respective attributes to be measured using IMA. 01. cache 02. crypt 03. integrity 04. linear 05. mirror 06. multipath 07. raid 08. snapshot 09. striped 10. verity For rest of the targets, handle the STATUSTYPE_IMA case by setting the measurement buffer to NULL. For IMA to measure the data on a given system, the IMA policy on the system needs to be updated to have the following line, and the system needs to be restarted for the measurements to take effect. /etc/ima/ima-policy measure func=CRITICAL_DATA label=device-mapper template=ima-buf The measurements will be reflected in the IMA logs, which are located at: /sys/kernel/security/integrity/ima/ascii_runtime_measurements /sys/kernel/security/integrity/ima/binary_runtime_measurements These IMA logs can later be consumed by various attestation clients running on the system, and send them to external services for attesting the system. The DM target data measured by IMA subsystem can alternatively be queried from userspace by setting DM_IMA_MEASUREMENT_FLAG with DM_TABLE_STATUS_CMD. Signed-off-by: Tushar Sugandhi <tusharsu@linux.microsoft.com> Signed-off-by: Mike Snitzer <snitzer@redhat.com>
544 lines
13 KiB
C
544 lines
13 KiB
C
/*
|
|
* Copyright (C) 2001-2003 Sistina Software (UK) Limited.
|
|
*
|
|
* This file is released under the GPL.
|
|
*/
|
|
|
|
#include "dm.h"
|
|
#include <linux/device-mapper.h>
|
|
|
|
#include <linux/module.h>
|
|
#include <linux/init.h>
|
|
#include <linux/blkdev.h>
|
|
#include <linux/bio.h>
|
|
#include <linux/dax.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/log2.h>
|
|
|
|
#define DM_MSG_PREFIX "striped"
|
|
#define DM_IO_ERROR_THRESHOLD 15
|
|
|
|
struct stripe {
|
|
struct dm_dev *dev;
|
|
sector_t physical_start;
|
|
|
|
atomic_t error_count;
|
|
};
|
|
|
|
struct stripe_c {
|
|
uint32_t stripes;
|
|
int stripes_shift;
|
|
|
|
/* The size of this target / num. stripes */
|
|
sector_t stripe_width;
|
|
|
|
uint32_t chunk_size;
|
|
int chunk_size_shift;
|
|
|
|
/* Needed for handling events */
|
|
struct dm_target *ti;
|
|
|
|
/* Work struct used for triggering events*/
|
|
struct work_struct trigger_event;
|
|
|
|
struct stripe stripe[];
|
|
};
|
|
|
|
/*
|
|
* An event is triggered whenever a drive
|
|
* drops out of a stripe volume.
|
|
*/
|
|
static void trigger_event(struct work_struct *work)
|
|
{
|
|
struct stripe_c *sc = container_of(work, struct stripe_c,
|
|
trigger_event);
|
|
dm_table_event(sc->ti->table);
|
|
}
|
|
|
|
/*
|
|
* Parse a single <dev> <sector> pair
|
|
*/
|
|
static int get_stripe(struct dm_target *ti, struct stripe_c *sc,
|
|
unsigned int stripe, char **argv)
|
|
{
|
|
unsigned long long start;
|
|
char dummy;
|
|
int ret;
|
|
|
|
if (sscanf(argv[1], "%llu%c", &start, &dummy) != 1)
|
|
return -EINVAL;
|
|
|
|
ret = dm_get_device(ti, argv[0], dm_table_get_mode(ti->table),
|
|
&sc->stripe[stripe].dev);
|
|
if (ret)
|
|
return ret;
|
|
|
|
sc->stripe[stripe].physical_start = start;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Construct a striped mapping.
|
|
* <number of stripes> <chunk size> [<dev_path> <offset>]+
|
|
*/
|
|
static int stripe_ctr(struct dm_target *ti, unsigned int argc, char **argv)
|
|
{
|
|
struct stripe_c *sc;
|
|
sector_t width, tmp_len;
|
|
uint32_t stripes;
|
|
uint32_t chunk_size;
|
|
int r;
|
|
unsigned int i;
|
|
|
|
if (argc < 2) {
|
|
ti->error = "Not enough arguments";
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (kstrtouint(argv[0], 10, &stripes) || !stripes) {
|
|
ti->error = "Invalid stripe count";
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (kstrtouint(argv[1], 10, &chunk_size) || !chunk_size) {
|
|
ti->error = "Invalid chunk_size";
|
|
return -EINVAL;
|
|
}
|
|
|
|
width = ti->len;
|
|
if (sector_div(width, stripes)) {
|
|
ti->error = "Target length not divisible by "
|
|
"number of stripes";
|
|
return -EINVAL;
|
|
}
|
|
|
|
tmp_len = width;
|
|
if (sector_div(tmp_len, chunk_size)) {
|
|
ti->error = "Target length not divisible by "
|
|
"chunk size";
|
|
return -EINVAL;
|
|
}
|
|
|
|
/*
|
|
* Do we have enough arguments for that many stripes ?
|
|
*/
|
|
if (argc != (2 + 2 * stripes)) {
|
|
ti->error = "Not enough destinations "
|
|
"specified";
|
|
return -EINVAL;
|
|
}
|
|
|
|
sc = kmalloc(struct_size(sc, stripe, stripes), GFP_KERNEL);
|
|
if (!sc) {
|
|
ti->error = "Memory allocation for striped context "
|
|
"failed";
|
|
return -ENOMEM;
|
|
}
|
|
|
|
INIT_WORK(&sc->trigger_event, trigger_event);
|
|
|
|
/* Set pointer to dm target; used in trigger_event */
|
|
sc->ti = ti;
|
|
sc->stripes = stripes;
|
|
sc->stripe_width = width;
|
|
|
|
if (stripes & (stripes - 1))
|
|
sc->stripes_shift = -1;
|
|
else
|
|
sc->stripes_shift = __ffs(stripes);
|
|
|
|
r = dm_set_target_max_io_len(ti, chunk_size);
|
|
if (r) {
|
|
kfree(sc);
|
|
return r;
|
|
}
|
|
|
|
ti->num_flush_bios = stripes;
|
|
ti->num_discard_bios = stripes;
|
|
ti->num_secure_erase_bios = stripes;
|
|
ti->num_write_same_bios = stripes;
|
|
ti->num_write_zeroes_bios = stripes;
|
|
|
|
sc->chunk_size = chunk_size;
|
|
if (chunk_size & (chunk_size - 1))
|
|
sc->chunk_size_shift = -1;
|
|
else
|
|
sc->chunk_size_shift = __ffs(chunk_size);
|
|
|
|
/*
|
|
* Get the stripe destinations.
|
|
*/
|
|
for (i = 0; i < stripes; i++) {
|
|
argv += 2;
|
|
|
|
r = get_stripe(ti, sc, i, argv);
|
|
if (r < 0) {
|
|
ti->error = "Couldn't parse stripe destination";
|
|
while (i--)
|
|
dm_put_device(ti, sc->stripe[i].dev);
|
|
kfree(sc);
|
|
return r;
|
|
}
|
|
atomic_set(&(sc->stripe[i].error_count), 0);
|
|
}
|
|
|
|
ti->private = sc;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void stripe_dtr(struct dm_target *ti)
|
|
{
|
|
unsigned int i;
|
|
struct stripe_c *sc = (struct stripe_c *) ti->private;
|
|
|
|
for (i = 0; i < sc->stripes; i++)
|
|
dm_put_device(ti, sc->stripe[i].dev);
|
|
|
|
flush_work(&sc->trigger_event);
|
|
kfree(sc);
|
|
}
|
|
|
|
static void stripe_map_sector(struct stripe_c *sc, sector_t sector,
|
|
uint32_t *stripe, sector_t *result)
|
|
{
|
|
sector_t chunk = dm_target_offset(sc->ti, sector);
|
|
sector_t chunk_offset;
|
|
|
|
if (sc->chunk_size_shift < 0)
|
|
chunk_offset = sector_div(chunk, sc->chunk_size);
|
|
else {
|
|
chunk_offset = chunk & (sc->chunk_size - 1);
|
|
chunk >>= sc->chunk_size_shift;
|
|
}
|
|
|
|
if (sc->stripes_shift < 0)
|
|
*stripe = sector_div(chunk, sc->stripes);
|
|
else {
|
|
*stripe = chunk & (sc->stripes - 1);
|
|
chunk >>= sc->stripes_shift;
|
|
}
|
|
|
|
if (sc->chunk_size_shift < 0)
|
|
chunk *= sc->chunk_size;
|
|
else
|
|
chunk <<= sc->chunk_size_shift;
|
|
|
|
*result = chunk + chunk_offset;
|
|
}
|
|
|
|
static void stripe_map_range_sector(struct stripe_c *sc, sector_t sector,
|
|
uint32_t target_stripe, sector_t *result)
|
|
{
|
|
uint32_t stripe;
|
|
|
|
stripe_map_sector(sc, sector, &stripe, result);
|
|
if (stripe == target_stripe)
|
|
return;
|
|
|
|
/* round down */
|
|
sector = *result;
|
|
if (sc->chunk_size_shift < 0)
|
|
*result -= sector_div(sector, sc->chunk_size);
|
|
else
|
|
*result = sector & ~(sector_t)(sc->chunk_size - 1);
|
|
|
|
if (target_stripe < stripe)
|
|
*result += sc->chunk_size; /* next chunk */
|
|
}
|
|
|
|
static int stripe_map_range(struct stripe_c *sc, struct bio *bio,
|
|
uint32_t target_stripe)
|
|
{
|
|
sector_t begin, end;
|
|
|
|
stripe_map_range_sector(sc, bio->bi_iter.bi_sector,
|
|
target_stripe, &begin);
|
|
stripe_map_range_sector(sc, bio_end_sector(bio),
|
|
target_stripe, &end);
|
|
if (begin < end) {
|
|
bio_set_dev(bio, sc->stripe[target_stripe].dev->bdev);
|
|
bio->bi_iter.bi_sector = begin +
|
|
sc->stripe[target_stripe].physical_start;
|
|
bio->bi_iter.bi_size = to_bytes(end - begin);
|
|
return DM_MAPIO_REMAPPED;
|
|
} else {
|
|
/* The range doesn't map to the target stripe */
|
|
bio_endio(bio);
|
|
return DM_MAPIO_SUBMITTED;
|
|
}
|
|
}
|
|
|
|
static int stripe_map(struct dm_target *ti, struct bio *bio)
|
|
{
|
|
struct stripe_c *sc = ti->private;
|
|
uint32_t stripe;
|
|
unsigned target_bio_nr;
|
|
|
|
if (bio->bi_opf & REQ_PREFLUSH) {
|
|
target_bio_nr = dm_bio_get_target_bio_nr(bio);
|
|
BUG_ON(target_bio_nr >= sc->stripes);
|
|
bio_set_dev(bio, sc->stripe[target_bio_nr].dev->bdev);
|
|
return DM_MAPIO_REMAPPED;
|
|
}
|
|
if (unlikely(bio_op(bio) == REQ_OP_DISCARD) ||
|
|
unlikely(bio_op(bio) == REQ_OP_SECURE_ERASE) ||
|
|
unlikely(bio_op(bio) == REQ_OP_WRITE_ZEROES) ||
|
|
unlikely(bio_op(bio) == REQ_OP_WRITE_SAME)) {
|
|
target_bio_nr = dm_bio_get_target_bio_nr(bio);
|
|
BUG_ON(target_bio_nr >= sc->stripes);
|
|
return stripe_map_range(sc, bio, target_bio_nr);
|
|
}
|
|
|
|
stripe_map_sector(sc, bio->bi_iter.bi_sector,
|
|
&stripe, &bio->bi_iter.bi_sector);
|
|
|
|
bio->bi_iter.bi_sector += sc->stripe[stripe].physical_start;
|
|
bio_set_dev(bio, sc->stripe[stripe].dev->bdev);
|
|
|
|
return DM_MAPIO_REMAPPED;
|
|
}
|
|
|
|
#if IS_ENABLED(CONFIG_DAX_DRIVER)
|
|
static long stripe_dax_direct_access(struct dm_target *ti, pgoff_t pgoff,
|
|
long nr_pages, void **kaddr, pfn_t *pfn)
|
|
{
|
|
sector_t dev_sector, sector = pgoff * PAGE_SECTORS;
|
|
struct stripe_c *sc = ti->private;
|
|
struct dax_device *dax_dev;
|
|
struct block_device *bdev;
|
|
uint32_t stripe;
|
|
long ret;
|
|
|
|
stripe_map_sector(sc, sector, &stripe, &dev_sector);
|
|
dev_sector += sc->stripe[stripe].physical_start;
|
|
dax_dev = sc->stripe[stripe].dev->dax_dev;
|
|
bdev = sc->stripe[stripe].dev->bdev;
|
|
|
|
ret = bdev_dax_pgoff(bdev, dev_sector, nr_pages * PAGE_SIZE, &pgoff);
|
|
if (ret)
|
|
return ret;
|
|
return dax_direct_access(dax_dev, pgoff, nr_pages, kaddr, pfn);
|
|
}
|
|
|
|
static size_t stripe_dax_copy_from_iter(struct dm_target *ti, pgoff_t pgoff,
|
|
void *addr, size_t bytes, struct iov_iter *i)
|
|
{
|
|
sector_t dev_sector, sector = pgoff * PAGE_SECTORS;
|
|
struct stripe_c *sc = ti->private;
|
|
struct dax_device *dax_dev;
|
|
struct block_device *bdev;
|
|
uint32_t stripe;
|
|
|
|
stripe_map_sector(sc, sector, &stripe, &dev_sector);
|
|
dev_sector += sc->stripe[stripe].physical_start;
|
|
dax_dev = sc->stripe[stripe].dev->dax_dev;
|
|
bdev = sc->stripe[stripe].dev->bdev;
|
|
|
|
if (bdev_dax_pgoff(bdev, dev_sector, ALIGN(bytes, PAGE_SIZE), &pgoff))
|
|
return 0;
|
|
return dax_copy_from_iter(dax_dev, pgoff, addr, bytes, i);
|
|
}
|
|
|
|
static size_t stripe_dax_copy_to_iter(struct dm_target *ti, pgoff_t pgoff,
|
|
void *addr, size_t bytes, struct iov_iter *i)
|
|
{
|
|
sector_t dev_sector, sector = pgoff * PAGE_SECTORS;
|
|
struct stripe_c *sc = ti->private;
|
|
struct dax_device *dax_dev;
|
|
struct block_device *bdev;
|
|
uint32_t stripe;
|
|
|
|
stripe_map_sector(sc, sector, &stripe, &dev_sector);
|
|
dev_sector += sc->stripe[stripe].physical_start;
|
|
dax_dev = sc->stripe[stripe].dev->dax_dev;
|
|
bdev = sc->stripe[stripe].dev->bdev;
|
|
|
|
if (bdev_dax_pgoff(bdev, dev_sector, ALIGN(bytes, PAGE_SIZE), &pgoff))
|
|
return 0;
|
|
return dax_copy_to_iter(dax_dev, pgoff, addr, bytes, i);
|
|
}
|
|
|
|
static int stripe_dax_zero_page_range(struct dm_target *ti, pgoff_t pgoff,
|
|
size_t nr_pages)
|
|
{
|
|
int ret;
|
|
sector_t dev_sector, sector = pgoff * PAGE_SECTORS;
|
|
struct stripe_c *sc = ti->private;
|
|
struct dax_device *dax_dev;
|
|
struct block_device *bdev;
|
|
uint32_t stripe;
|
|
|
|
stripe_map_sector(sc, sector, &stripe, &dev_sector);
|
|
dev_sector += sc->stripe[stripe].physical_start;
|
|
dax_dev = sc->stripe[stripe].dev->dax_dev;
|
|
bdev = sc->stripe[stripe].dev->bdev;
|
|
|
|
ret = bdev_dax_pgoff(bdev, dev_sector, nr_pages << PAGE_SHIFT, &pgoff);
|
|
if (ret)
|
|
return ret;
|
|
return dax_zero_page_range(dax_dev, pgoff, nr_pages);
|
|
}
|
|
|
|
#else
|
|
#define stripe_dax_direct_access NULL
|
|
#define stripe_dax_copy_from_iter NULL
|
|
#define stripe_dax_copy_to_iter NULL
|
|
#define stripe_dax_zero_page_range NULL
|
|
#endif
|
|
|
|
/*
|
|
* Stripe status:
|
|
*
|
|
* INFO
|
|
* #stripes [stripe_name <stripe_name>] [group word count]
|
|
* [error count 'A|D' <error count 'A|D'>]
|
|
*
|
|
* TABLE
|
|
* #stripes [stripe chunk size]
|
|
* [stripe_name physical_start <stripe_name physical_start>]
|
|
*
|
|
*/
|
|
|
|
static void stripe_status(struct dm_target *ti, status_type_t type,
|
|
unsigned status_flags, char *result, unsigned maxlen)
|
|
{
|
|
struct stripe_c *sc = (struct stripe_c *) ti->private;
|
|
unsigned int sz = 0;
|
|
unsigned int i;
|
|
|
|
switch (type) {
|
|
case STATUSTYPE_INFO:
|
|
DMEMIT("%d ", sc->stripes);
|
|
for (i = 0; i < sc->stripes; i++) {
|
|
DMEMIT("%s ", sc->stripe[i].dev->name);
|
|
}
|
|
DMEMIT("1 ");
|
|
for (i = 0; i < sc->stripes; i++) {
|
|
DMEMIT("%c", atomic_read(&(sc->stripe[i].error_count)) ?
|
|
'D' : 'A');
|
|
}
|
|
break;
|
|
|
|
case STATUSTYPE_TABLE:
|
|
DMEMIT("%d %llu", sc->stripes,
|
|
(unsigned long long)sc->chunk_size);
|
|
for (i = 0; i < sc->stripes; i++)
|
|
DMEMIT(" %s %llu", sc->stripe[i].dev->name,
|
|
(unsigned long long)sc->stripe[i].physical_start);
|
|
break;
|
|
|
|
case STATUSTYPE_IMA:
|
|
DMEMIT_TARGET_NAME_VERSION(ti->type);
|
|
DMEMIT(",stripes=%d,chunk_size=%llu", sc->stripes,
|
|
(unsigned long long)sc->chunk_size);
|
|
|
|
for (i = 0; i < sc->stripes; i++) {
|
|
DMEMIT(",stripe_%d_device_name=%s", i, sc->stripe[i].dev->name);
|
|
DMEMIT(",stripe_%d_physical_start=%llu", i,
|
|
(unsigned long long)sc->stripe[i].physical_start);
|
|
DMEMIT(",stripe_%d_status=%c", i,
|
|
atomic_read(&(sc->stripe[i].error_count)) ? 'D' : 'A');
|
|
}
|
|
DMEMIT(";");
|
|
break;
|
|
}
|
|
}
|
|
|
|
static int stripe_end_io(struct dm_target *ti, struct bio *bio,
|
|
blk_status_t *error)
|
|
{
|
|
unsigned i;
|
|
char major_minor[16];
|
|
struct stripe_c *sc = ti->private;
|
|
|
|
if (!*error)
|
|
return DM_ENDIO_DONE; /* I/O complete */
|
|
|
|
if (bio->bi_opf & REQ_RAHEAD)
|
|
return DM_ENDIO_DONE;
|
|
|
|
if (*error == BLK_STS_NOTSUPP)
|
|
return DM_ENDIO_DONE;
|
|
|
|
memset(major_minor, 0, sizeof(major_minor));
|
|
sprintf(major_minor, "%d:%d", MAJOR(bio_dev(bio)), MINOR(bio_dev(bio)));
|
|
|
|
/*
|
|
* Test to see which stripe drive triggered the event
|
|
* and increment error count for all stripes on that device.
|
|
* If the error count for a given device exceeds the threshold
|
|
* value we will no longer trigger any further events.
|
|
*/
|
|
for (i = 0; i < sc->stripes; i++)
|
|
if (!strcmp(sc->stripe[i].dev->name, major_minor)) {
|
|
atomic_inc(&(sc->stripe[i].error_count));
|
|
if (atomic_read(&(sc->stripe[i].error_count)) <
|
|
DM_IO_ERROR_THRESHOLD)
|
|
schedule_work(&sc->trigger_event);
|
|
}
|
|
|
|
return DM_ENDIO_DONE;
|
|
}
|
|
|
|
static int stripe_iterate_devices(struct dm_target *ti,
|
|
iterate_devices_callout_fn fn, void *data)
|
|
{
|
|
struct stripe_c *sc = ti->private;
|
|
int ret = 0;
|
|
unsigned i = 0;
|
|
|
|
do {
|
|
ret = fn(ti, sc->stripe[i].dev,
|
|
sc->stripe[i].physical_start,
|
|
sc->stripe_width, data);
|
|
} while (!ret && ++i < sc->stripes);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void stripe_io_hints(struct dm_target *ti,
|
|
struct queue_limits *limits)
|
|
{
|
|
struct stripe_c *sc = ti->private;
|
|
unsigned chunk_size = sc->chunk_size << SECTOR_SHIFT;
|
|
|
|
blk_limits_io_min(limits, chunk_size);
|
|
blk_limits_io_opt(limits, chunk_size * sc->stripes);
|
|
}
|
|
|
|
static struct target_type stripe_target = {
|
|
.name = "striped",
|
|
.version = {1, 6, 0},
|
|
.features = DM_TARGET_PASSES_INTEGRITY | DM_TARGET_NOWAIT,
|
|
.module = THIS_MODULE,
|
|
.ctr = stripe_ctr,
|
|
.dtr = stripe_dtr,
|
|
.map = stripe_map,
|
|
.end_io = stripe_end_io,
|
|
.status = stripe_status,
|
|
.iterate_devices = stripe_iterate_devices,
|
|
.io_hints = stripe_io_hints,
|
|
.direct_access = stripe_dax_direct_access,
|
|
.dax_copy_from_iter = stripe_dax_copy_from_iter,
|
|
.dax_copy_to_iter = stripe_dax_copy_to_iter,
|
|
.dax_zero_page_range = stripe_dax_zero_page_range,
|
|
};
|
|
|
|
int __init dm_stripe_init(void)
|
|
{
|
|
int r;
|
|
|
|
r = dm_register_target(&stripe_target);
|
|
if (r < 0)
|
|
DMWARN("target registration failed");
|
|
|
|
return r;
|
|
}
|
|
|
|
void dm_stripe_exit(void)
|
|
{
|
|
dm_unregister_target(&stripe_target);
|
|
}
|