mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-28 06:34:12 +08:00
0760267809
- Make an informative message less ominous (Keith) - Enhanced trace decoding (Guixin) - TCP updates (Hannes, Li) - Fabrics connect deadlock fix (Chunguang) - Platform API migration update (Uwe) - A new device quirk (Jiawei) -----BEGIN PGP SIGNATURE----- iQIzBAABCAAdFiEE3Fbyvv+648XNRdHTPe3zGtjzRgkFAmX8eFoACgkQPe3zGtjz RglB1A//Tw/U2aXAMSe6sX01vL6u3uCY6moXz/C7taa3tkn6TeFDg905jBDbX89E HCaQKBpppG9SfQg/CS5pkTmUboW0kDjDKUUwK5IOUrVbaPC9JQ4FhJ4dqoTvd8I+ CgsUpuF6HWQaSWr+2atfJLDSklGCVJrvs2YxycCpYaaHaBeBkc8dwRk+ec1RlePW m2Kq4volZ8mVQ4FXu21YfH281Hw9gndqWMXPFMiv7U2zrVB9OYP9Z3BqLTq2cUa5 ce/4CY7nXv/kEddnuLYTBKUz/ymnXvD+FE8Ibzt1YH8laC96Juj9WUZmRy/okzAZ V/BpGriYzzCANVNyQxNmp9okskBdGXuhpefwOXZ17jbRUrfmNo5yU/TuxM8nvG9P ikR5qZSaOq3cELDAwtVOHkq6dyhimxW4HNVr2fIY0KEVgqL1ph4zSVckMGiW4/Dq FbIB2fTc4sVjZmhHuuevYoelAulHnlMxX8RNrk5us6AT4f+1hYcqdxCCR3IsX/Jw GdUYPz/EB8HnSmxCGEWPyI2Ldf+Id9xFUOMpsoEWz+637TRGdNv1ADfBeUUnBoWz DBEBo8YMcYeKgv0ohZql5B5g3Bl9GfyRCcDua/uubwobrm/tzGfHamE4LYp5NeQY fqt0F8WmAG0Dey/6whSo+29dhwYnCpHl7xQLNhz6wMhqlXUdQm0= =Cntu -----END PGP SIGNATURE----- Merge tag 'nvme-6.9-2024-03-21' of git://git.infradead.org/nvme into block-6.9 Pull NVMe fixes from Keith: "nvme updates for Linux 6.9 - Make an informative message less ominous (Keith) - Enhanced trace decoding (Guixin) - TCP updates (Hannes, Li) - Fabrics connect deadlock fix (Chunguang) - Platform API migration update (Uwe) - A new device quirk (Jiawei)" * tag 'nvme-6.9-2024-03-21' of git://git.infradead.org/nvme: nvmet-rdma: remove NVMET_RDMA_REQ_INVALIDATE_RKEY flag nvme: remove redundant BUILD_BUG_ON check nvme/tcp: Add wq_unbound modparam for nvme_tcp_wq nvme-tcp: Export the nvme_tcp_wq to sysfs drivers/nvme: Add quirks for device 126f:2262 nvme: parse format command's lbafu when tracing nvme: add tracing of reservation commands nvme: parse zns command's zsa and zrasf to string nvme: use nvme_disk_is_ns_head helper nvme: fix reconnection fail due to reserved tag allocation nvmet: add tracing of zns commands nvmet: add tracing of authentication commands nvme-apple: Convert to platform remove callback returning void nvmet-tcp: do not continue for invalid icreq nvme: change shutdown timeout setting message
829 lines
22 KiB
C
829 lines
22 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Sysfs interface for the NVMe core driver.
|
|
*
|
|
* Copyright (c) 2011-2014, Intel Corporation.
|
|
*/
|
|
|
|
#include <linux/nvme-auth.h>
|
|
|
|
#include "nvme.h"
|
|
#include "fabrics.h"
|
|
|
|
static ssize_t nvme_sysfs_reset(struct device *dev,
|
|
struct device_attribute *attr, const char *buf,
|
|
size_t count)
|
|
{
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev);
|
|
int ret;
|
|
|
|
ret = nvme_reset_ctrl_sync(ctrl);
|
|
if (ret < 0)
|
|
return ret;
|
|
return count;
|
|
}
|
|
static DEVICE_ATTR(reset_controller, S_IWUSR, NULL, nvme_sysfs_reset);
|
|
|
|
static ssize_t nvme_sysfs_rescan(struct device *dev,
|
|
struct device_attribute *attr, const char *buf,
|
|
size_t count)
|
|
{
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev);
|
|
|
|
nvme_queue_scan(ctrl);
|
|
return count;
|
|
}
|
|
static DEVICE_ATTR(rescan_controller, S_IWUSR, NULL, nvme_sysfs_rescan);
|
|
|
|
static ssize_t nvme_adm_passthru_err_log_enabled_show(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev);
|
|
|
|
return sysfs_emit(buf,
|
|
ctrl->passthru_err_log_enabled ? "on\n" : "off\n");
|
|
}
|
|
|
|
static ssize_t nvme_adm_passthru_err_log_enabled_store(struct device *dev,
|
|
struct device_attribute *attr, const char *buf, size_t count)
|
|
{
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev);
|
|
bool passthru_err_log_enabled;
|
|
int err;
|
|
|
|
err = kstrtobool(buf, &passthru_err_log_enabled);
|
|
if (err)
|
|
return -EINVAL;
|
|
|
|
ctrl->passthru_err_log_enabled = passthru_err_log_enabled;
|
|
|
|
return count;
|
|
}
|
|
|
|
static inline struct nvme_ns_head *dev_to_ns_head(struct device *dev)
|
|
{
|
|
struct gendisk *disk = dev_to_disk(dev);
|
|
|
|
if (nvme_disk_is_ns_head(disk))
|
|
return disk->private_data;
|
|
return nvme_get_ns_from_dev(dev)->head;
|
|
}
|
|
|
|
static ssize_t nvme_io_passthru_err_log_enabled_show(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
struct nvme_ns_head *head = dev_to_ns_head(dev);
|
|
|
|
return sysfs_emit(buf, head->passthru_err_log_enabled ? "on\n" : "off\n");
|
|
}
|
|
|
|
static ssize_t nvme_io_passthru_err_log_enabled_store(struct device *dev,
|
|
struct device_attribute *attr, const char *buf, size_t count)
|
|
{
|
|
struct nvme_ns_head *head = dev_to_ns_head(dev);
|
|
bool passthru_err_log_enabled;
|
|
int err;
|
|
|
|
err = kstrtobool(buf, &passthru_err_log_enabled);
|
|
if (err)
|
|
return -EINVAL;
|
|
head->passthru_err_log_enabled = passthru_err_log_enabled;
|
|
|
|
return count;
|
|
}
|
|
|
|
static struct device_attribute dev_attr_adm_passthru_err_log_enabled = \
|
|
__ATTR(passthru_err_log_enabled, S_IRUGO | S_IWUSR, \
|
|
nvme_adm_passthru_err_log_enabled_show, nvme_adm_passthru_err_log_enabled_store);
|
|
|
|
static struct device_attribute dev_attr_io_passthru_err_log_enabled = \
|
|
__ATTR(passthru_err_log_enabled, S_IRUGO | S_IWUSR, \
|
|
nvme_io_passthru_err_log_enabled_show, nvme_io_passthru_err_log_enabled_store);
|
|
|
|
static ssize_t wwid_show(struct device *dev, struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct nvme_ns_head *head = dev_to_ns_head(dev);
|
|
struct nvme_ns_ids *ids = &head->ids;
|
|
struct nvme_subsystem *subsys = head->subsys;
|
|
int serial_len = sizeof(subsys->serial);
|
|
int model_len = sizeof(subsys->model);
|
|
|
|
if (!uuid_is_null(&ids->uuid))
|
|
return sysfs_emit(buf, "uuid.%pU\n", &ids->uuid);
|
|
|
|
if (memchr_inv(ids->nguid, 0, sizeof(ids->nguid)))
|
|
return sysfs_emit(buf, "eui.%16phN\n", ids->nguid);
|
|
|
|
if (memchr_inv(ids->eui64, 0, sizeof(ids->eui64)))
|
|
return sysfs_emit(buf, "eui.%8phN\n", ids->eui64);
|
|
|
|
while (serial_len > 0 && (subsys->serial[serial_len - 1] == ' ' ||
|
|
subsys->serial[serial_len - 1] == '\0'))
|
|
serial_len--;
|
|
while (model_len > 0 && (subsys->model[model_len - 1] == ' ' ||
|
|
subsys->model[model_len - 1] == '\0'))
|
|
model_len--;
|
|
|
|
return sysfs_emit(buf, "nvme.%04x-%*phN-%*phN-%08x\n", subsys->vendor_id,
|
|
serial_len, subsys->serial, model_len, subsys->model,
|
|
head->ns_id);
|
|
}
|
|
static DEVICE_ATTR_RO(wwid);
|
|
|
|
static ssize_t nguid_show(struct device *dev, struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
return sysfs_emit(buf, "%pU\n", dev_to_ns_head(dev)->ids.nguid);
|
|
}
|
|
static DEVICE_ATTR_RO(nguid);
|
|
|
|
static ssize_t uuid_show(struct device *dev, struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct nvme_ns_ids *ids = &dev_to_ns_head(dev)->ids;
|
|
|
|
/* For backward compatibility expose the NGUID to userspace if
|
|
* we have no UUID set
|
|
*/
|
|
if (uuid_is_null(&ids->uuid)) {
|
|
dev_warn_once(dev,
|
|
"No UUID available providing old NGUID\n");
|
|
return sysfs_emit(buf, "%pU\n", ids->nguid);
|
|
}
|
|
return sysfs_emit(buf, "%pU\n", &ids->uuid);
|
|
}
|
|
static DEVICE_ATTR_RO(uuid);
|
|
|
|
static ssize_t eui_show(struct device *dev, struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
return sysfs_emit(buf, "%8ph\n", dev_to_ns_head(dev)->ids.eui64);
|
|
}
|
|
static DEVICE_ATTR_RO(eui);
|
|
|
|
static ssize_t nsid_show(struct device *dev, struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
return sysfs_emit(buf, "%d\n", dev_to_ns_head(dev)->ns_id);
|
|
}
|
|
static DEVICE_ATTR_RO(nsid);
|
|
|
|
static ssize_t csi_show(struct device *dev, struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
return sysfs_emit(buf, "%u\n", dev_to_ns_head(dev)->ids.csi);
|
|
}
|
|
static DEVICE_ATTR_RO(csi);
|
|
|
|
static ssize_t metadata_bytes_show(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
return sysfs_emit(buf, "%u\n", dev_to_ns_head(dev)->ms);
|
|
}
|
|
static DEVICE_ATTR_RO(metadata_bytes);
|
|
|
|
static int ns_head_update_nuse(struct nvme_ns_head *head)
|
|
{
|
|
struct nvme_id_ns *id;
|
|
struct nvme_ns *ns;
|
|
int srcu_idx, ret = -EWOULDBLOCK;
|
|
|
|
/* Avoid issuing commands too often by rate limiting the update */
|
|
if (!__ratelimit(&head->rs_nuse))
|
|
return 0;
|
|
|
|
srcu_idx = srcu_read_lock(&head->srcu);
|
|
ns = nvme_find_path(head);
|
|
if (!ns)
|
|
goto out_unlock;
|
|
|
|
ret = nvme_identify_ns(ns->ctrl, head->ns_id, &id);
|
|
if (ret)
|
|
goto out_unlock;
|
|
|
|
head->nuse = le64_to_cpu(id->nuse);
|
|
kfree(id);
|
|
|
|
out_unlock:
|
|
srcu_read_unlock(&head->srcu, srcu_idx);
|
|
return ret;
|
|
}
|
|
|
|
static int ns_update_nuse(struct nvme_ns *ns)
|
|
{
|
|
struct nvme_id_ns *id;
|
|
int ret;
|
|
|
|
/* Avoid issuing commands too often by rate limiting the update. */
|
|
if (!__ratelimit(&ns->head->rs_nuse))
|
|
return 0;
|
|
|
|
ret = nvme_identify_ns(ns->ctrl, ns->head->ns_id, &id);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ns->head->nuse = le64_to_cpu(id->nuse);
|
|
kfree(id);
|
|
return 0;
|
|
}
|
|
|
|
static ssize_t nuse_show(struct device *dev, struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct nvme_ns_head *head = dev_to_ns_head(dev);
|
|
struct gendisk *disk = dev_to_disk(dev);
|
|
struct block_device *bdev = disk->part0;
|
|
int ret;
|
|
|
|
if (nvme_disk_is_ns_head(bdev->bd_disk))
|
|
ret = ns_head_update_nuse(head);
|
|
else
|
|
ret = ns_update_nuse(bdev->bd_disk->private_data);
|
|
if (ret)
|
|
return ret;
|
|
|
|
return sysfs_emit(buf, "%llu\n", head->nuse);
|
|
}
|
|
static DEVICE_ATTR_RO(nuse);
|
|
|
|
static struct attribute *nvme_ns_attrs[] = {
|
|
&dev_attr_wwid.attr,
|
|
&dev_attr_uuid.attr,
|
|
&dev_attr_nguid.attr,
|
|
&dev_attr_eui.attr,
|
|
&dev_attr_csi.attr,
|
|
&dev_attr_nsid.attr,
|
|
&dev_attr_metadata_bytes.attr,
|
|
&dev_attr_nuse.attr,
|
|
#ifdef CONFIG_NVME_MULTIPATH
|
|
&dev_attr_ana_grpid.attr,
|
|
&dev_attr_ana_state.attr,
|
|
#endif
|
|
&dev_attr_io_passthru_err_log_enabled.attr,
|
|
NULL,
|
|
};
|
|
|
|
static umode_t nvme_ns_attrs_are_visible(struct kobject *kobj,
|
|
struct attribute *a, int n)
|
|
{
|
|
struct device *dev = container_of(kobj, struct device, kobj);
|
|
struct nvme_ns_ids *ids = &dev_to_ns_head(dev)->ids;
|
|
|
|
if (a == &dev_attr_uuid.attr) {
|
|
if (uuid_is_null(&ids->uuid) &&
|
|
!memchr_inv(ids->nguid, 0, sizeof(ids->nguid)))
|
|
return 0;
|
|
}
|
|
if (a == &dev_attr_nguid.attr) {
|
|
if (!memchr_inv(ids->nguid, 0, sizeof(ids->nguid)))
|
|
return 0;
|
|
}
|
|
if (a == &dev_attr_eui.attr) {
|
|
if (!memchr_inv(ids->eui64, 0, sizeof(ids->eui64)))
|
|
return 0;
|
|
}
|
|
#ifdef CONFIG_NVME_MULTIPATH
|
|
if (a == &dev_attr_ana_grpid.attr || a == &dev_attr_ana_state.attr) {
|
|
/* per-path attr */
|
|
if (nvme_disk_is_ns_head(dev_to_disk(dev)))
|
|
return 0;
|
|
if (!nvme_ctrl_use_ana(nvme_get_ns_from_dev(dev)->ctrl))
|
|
return 0;
|
|
}
|
|
#endif
|
|
return a->mode;
|
|
}
|
|
|
|
static const struct attribute_group nvme_ns_attr_group = {
|
|
.attrs = nvme_ns_attrs,
|
|
.is_visible = nvme_ns_attrs_are_visible,
|
|
};
|
|
|
|
const struct attribute_group *nvme_ns_attr_groups[] = {
|
|
&nvme_ns_attr_group,
|
|
NULL,
|
|
};
|
|
|
|
#define nvme_show_str_function(field) \
|
|
static ssize_t field##_show(struct device *dev, \
|
|
struct device_attribute *attr, char *buf) \
|
|
{ \
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev); \
|
|
return sysfs_emit(buf, "%.*s\n", \
|
|
(int)sizeof(ctrl->subsys->field), ctrl->subsys->field); \
|
|
} \
|
|
static DEVICE_ATTR(field, S_IRUGO, field##_show, NULL);
|
|
|
|
nvme_show_str_function(model);
|
|
nvme_show_str_function(serial);
|
|
nvme_show_str_function(firmware_rev);
|
|
|
|
#define nvme_show_int_function(field) \
|
|
static ssize_t field##_show(struct device *dev, \
|
|
struct device_attribute *attr, char *buf) \
|
|
{ \
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev); \
|
|
return sysfs_emit(buf, "%d\n", ctrl->field); \
|
|
} \
|
|
static DEVICE_ATTR(field, S_IRUGO, field##_show, NULL);
|
|
|
|
nvme_show_int_function(cntlid);
|
|
nvme_show_int_function(numa_node);
|
|
nvme_show_int_function(queue_count);
|
|
nvme_show_int_function(sqsize);
|
|
nvme_show_int_function(kato);
|
|
|
|
static ssize_t nvme_sysfs_delete(struct device *dev,
|
|
struct device_attribute *attr, const char *buf,
|
|
size_t count)
|
|
{
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev);
|
|
|
|
if (!test_bit(NVME_CTRL_STARTED_ONCE, &ctrl->flags))
|
|
return -EBUSY;
|
|
|
|
if (device_remove_file_self(dev, attr))
|
|
nvme_delete_ctrl_sync(ctrl);
|
|
return count;
|
|
}
|
|
static DEVICE_ATTR(delete_controller, S_IWUSR, NULL, nvme_sysfs_delete);
|
|
|
|
static ssize_t nvme_sysfs_show_transport(struct device *dev,
|
|
struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev);
|
|
|
|
return sysfs_emit(buf, "%s\n", ctrl->ops->name);
|
|
}
|
|
static DEVICE_ATTR(transport, S_IRUGO, nvme_sysfs_show_transport, NULL);
|
|
|
|
static ssize_t nvme_sysfs_show_state(struct device *dev,
|
|
struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev);
|
|
unsigned state = (unsigned)nvme_ctrl_state(ctrl);
|
|
static const char *const state_name[] = {
|
|
[NVME_CTRL_NEW] = "new",
|
|
[NVME_CTRL_LIVE] = "live",
|
|
[NVME_CTRL_RESETTING] = "resetting",
|
|
[NVME_CTRL_CONNECTING] = "connecting",
|
|
[NVME_CTRL_DELETING] = "deleting",
|
|
[NVME_CTRL_DELETING_NOIO]= "deleting (no IO)",
|
|
[NVME_CTRL_DEAD] = "dead",
|
|
};
|
|
|
|
if (state < ARRAY_SIZE(state_name) && state_name[state])
|
|
return sysfs_emit(buf, "%s\n", state_name[state]);
|
|
|
|
return sysfs_emit(buf, "unknown state\n");
|
|
}
|
|
|
|
static DEVICE_ATTR(state, S_IRUGO, nvme_sysfs_show_state, NULL);
|
|
|
|
static ssize_t nvme_sysfs_show_subsysnqn(struct device *dev,
|
|
struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev);
|
|
|
|
return sysfs_emit(buf, "%s\n", ctrl->subsys->subnqn);
|
|
}
|
|
static DEVICE_ATTR(subsysnqn, S_IRUGO, nvme_sysfs_show_subsysnqn, NULL);
|
|
|
|
static ssize_t nvme_sysfs_show_hostnqn(struct device *dev,
|
|
struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev);
|
|
|
|
return sysfs_emit(buf, "%s\n", ctrl->opts->host->nqn);
|
|
}
|
|
static DEVICE_ATTR(hostnqn, S_IRUGO, nvme_sysfs_show_hostnqn, NULL);
|
|
|
|
static ssize_t nvme_sysfs_show_hostid(struct device *dev,
|
|
struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev);
|
|
|
|
return sysfs_emit(buf, "%pU\n", &ctrl->opts->host->id);
|
|
}
|
|
static DEVICE_ATTR(hostid, S_IRUGO, nvme_sysfs_show_hostid, NULL);
|
|
|
|
static ssize_t nvme_sysfs_show_address(struct device *dev,
|
|
struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev);
|
|
|
|
return ctrl->ops->get_address(ctrl, buf, PAGE_SIZE);
|
|
}
|
|
static DEVICE_ATTR(address, S_IRUGO, nvme_sysfs_show_address, NULL);
|
|
|
|
static ssize_t nvme_ctrl_loss_tmo_show(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev);
|
|
struct nvmf_ctrl_options *opts = ctrl->opts;
|
|
|
|
if (ctrl->opts->max_reconnects == -1)
|
|
return sysfs_emit(buf, "off\n");
|
|
return sysfs_emit(buf, "%d\n",
|
|
opts->max_reconnects * opts->reconnect_delay);
|
|
}
|
|
|
|
static ssize_t nvme_ctrl_loss_tmo_store(struct device *dev,
|
|
struct device_attribute *attr, const char *buf, size_t count)
|
|
{
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev);
|
|
struct nvmf_ctrl_options *opts = ctrl->opts;
|
|
int ctrl_loss_tmo, err;
|
|
|
|
err = kstrtoint(buf, 10, &ctrl_loss_tmo);
|
|
if (err)
|
|
return -EINVAL;
|
|
|
|
if (ctrl_loss_tmo < 0)
|
|
opts->max_reconnects = -1;
|
|
else
|
|
opts->max_reconnects = DIV_ROUND_UP(ctrl_loss_tmo,
|
|
opts->reconnect_delay);
|
|
return count;
|
|
}
|
|
static DEVICE_ATTR(ctrl_loss_tmo, S_IRUGO | S_IWUSR,
|
|
nvme_ctrl_loss_tmo_show, nvme_ctrl_loss_tmo_store);
|
|
|
|
static ssize_t nvme_ctrl_reconnect_delay_show(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev);
|
|
|
|
if (ctrl->opts->reconnect_delay == -1)
|
|
return sysfs_emit(buf, "off\n");
|
|
return sysfs_emit(buf, "%d\n", ctrl->opts->reconnect_delay);
|
|
}
|
|
|
|
static ssize_t nvme_ctrl_reconnect_delay_store(struct device *dev,
|
|
struct device_attribute *attr, const char *buf, size_t count)
|
|
{
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev);
|
|
unsigned int v;
|
|
int err;
|
|
|
|
err = kstrtou32(buf, 10, &v);
|
|
if (err)
|
|
return err;
|
|
|
|
ctrl->opts->reconnect_delay = v;
|
|
return count;
|
|
}
|
|
static DEVICE_ATTR(reconnect_delay, S_IRUGO | S_IWUSR,
|
|
nvme_ctrl_reconnect_delay_show, nvme_ctrl_reconnect_delay_store);
|
|
|
|
static ssize_t nvme_ctrl_fast_io_fail_tmo_show(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev);
|
|
|
|
if (ctrl->opts->fast_io_fail_tmo == -1)
|
|
return sysfs_emit(buf, "off\n");
|
|
return sysfs_emit(buf, "%d\n", ctrl->opts->fast_io_fail_tmo);
|
|
}
|
|
|
|
static ssize_t nvme_ctrl_fast_io_fail_tmo_store(struct device *dev,
|
|
struct device_attribute *attr, const char *buf, size_t count)
|
|
{
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev);
|
|
struct nvmf_ctrl_options *opts = ctrl->opts;
|
|
int fast_io_fail_tmo, err;
|
|
|
|
err = kstrtoint(buf, 10, &fast_io_fail_tmo);
|
|
if (err)
|
|
return -EINVAL;
|
|
|
|
if (fast_io_fail_tmo < 0)
|
|
opts->fast_io_fail_tmo = -1;
|
|
else
|
|
opts->fast_io_fail_tmo = fast_io_fail_tmo;
|
|
return count;
|
|
}
|
|
static DEVICE_ATTR(fast_io_fail_tmo, S_IRUGO | S_IWUSR,
|
|
nvme_ctrl_fast_io_fail_tmo_show, nvme_ctrl_fast_io_fail_tmo_store);
|
|
|
|
static ssize_t cntrltype_show(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
static const char * const type[] = {
|
|
[NVME_CTRL_IO] = "io\n",
|
|
[NVME_CTRL_DISC] = "discovery\n",
|
|
[NVME_CTRL_ADMIN] = "admin\n",
|
|
};
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev);
|
|
|
|
if (ctrl->cntrltype > NVME_CTRL_ADMIN || !type[ctrl->cntrltype])
|
|
return sysfs_emit(buf, "reserved\n");
|
|
|
|
return sysfs_emit(buf, type[ctrl->cntrltype]);
|
|
}
|
|
static DEVICE_ATTR_RO(cntrltype);
|
|
|
|
static ssize_t dctype_show(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
static const char * const type[] = {
|
|
[NVME_DCTYPE_NOT_REPORTED] = "none\n",
|
|
[NVME_DCTYPE_DDC] = "ddc\n",
|
|
[NVME_DCTYPE_CDC] = "cdc\n",
|
|
};
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev);
|
|
|
|
if (ctrl->dctype > NVME_DCTYPE_CDC || !type[ctrl->dctype])
|
|
return sysfs_emit(buf, "reserved\n");
|
|
|
|
return sysfs_emit(buf, type[ctrl->dctype]);
|
|
}
|
|
static DEVICE_ATTR_RO(dctype);
|
|
|
|
#ifdef CONFIG_NVME_HOST_AUTH
|
|
static ssize_t nvme_ctrl_dhchap_secret_show(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev);
|
|
struct nvmf_ctrl_options *opts = ctrl->opts;
|
|
|
|
if (!opts->dhchap_secret)
|
|
return sysfs_emit(buf, "none\n");
|
|
return sysfs_emit(buf, "%s\n", opts->dhchap_secret);
|
|
}
|
|
|
|
static ssize_t nvme_ctrl_dhchap_secret_store(struct device *dev,
|
|
struct device_attribute *attr, const char *buf, size_t count)
|
|
{
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev);
|
|
struct nvmf_ctrl_options *opts = ctrl->opts;
|
|
char *dhchap_secret;
|
|
|
|
if (!ctrl->opts->dhchap_secret)
|
|
return -EINVAL;
|
|
if (count < 7)
|
|
return -EINVAL;
|
|
if (memcmp(buf, "DHHC-1:", 7))
|
|
return -EINVAL;
|
|
|
|
dhchap_secret = kzalloc(count + 1, GFP_KERNEL);
|
|
if (!dhchap_secret)
|
|
return -ENOMEM;
|
|
memcpy(dhchap_secret, buf, count);
|
|
nvme_auth_stop(ctrl);
|
|
if (strcmp(dhchap_secret, opts->dhchap_secret)) {
|
|
struct nvme_dhchap_key *key, *host_key;
|
|
int ret;
|
|
|
|
ret = nvme_auth_generate_key(dhchap_secret, &key);
|
|
if (ret) {
|
|
kfree(dhchap_secret);
|
|
return ret;
|
|
}
|
|
kfree(opts->dhchap_secret);
|
|
opts->dhchap_secret = dhchap_secret;
|
|
host_key = ctrl->host_key;
|
|
mutex_lock(&ctrl->dhchap_auth_mutex);
|
|
ctrl->host_key = key;
|
|
mutex_unlock(&ctrl->dhchap_auth_mutex);
|
|
nvme_auth_free_key(host_key);
|
|
} else
|
|
kfree(dhchap_secret);
|
|
/* Start re-authentication */
|
|
dev_info(ctrl->device, "re-authenticating controller\n");
|
|
queue_work(nvme_wq, &ctrl->dhchap_auth_work);
|
|
|
|
return count;
|
|
}
|
|
|
|
static DEVICE_ATTR(dhchap_secret, S_IRUGO | S_IWUSR,
|
|
nvme_ctrl_dhchap_secret_show, nvme_ctrl_dhchap_secret_store);
|
|
|
|
static ssize_t nvme_ctrl_dhchap_ctrl_secret_show(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev);
|
|
struct nvmf_ctrl_options *opts = ctrl->opts;
|
|
|
|
if (!opts->dhchap_ctrl_secret)
|
|
return sysfs_emit(buf, "none\n");
|
|
return sysfs_emit(buf, "%s\n", opts->dhchap_ctrl_secret);
|
|
}
|
|
|
|
static ssize_t nvme_ctrl_dhchap_ctrl_secret_store(struct device *dev,
|
|
struct device_attribute *attr, const char *buf, size_t count)
|
|
{
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev);
|
|
struct nvmf_ctrl_options *opts = ctrl->opts;
|
|
char *dhchap_secret;
|
|
|
|
if (!ctrl->opts->dhchap_ctrl_secret)
|
|
return -EINVAL;
|
|
if (count < 7)
|
|
return -EINVAL;
|
|
if (memcmp(buf, "DHHC-1:", 7))
|
|
return -EINVAL;
|
|
|
|
dhchap_secret = kzalloc(count + 1, GFP_KERNEL);
|
|
if (!dhchap_secret)
|
|
return -ENOMEM;
|
|
memcpy(dhchap_secret, buf, count);
|
|
nvme_auth_stop(ctrl);
|
|
if (strcmp(dhchap_secret, opts->dhchap_ctrl_secret)) {
|
|
struct nvme_dhchap_key *key, *ctrl_key;
|
|
int ret;
|
|
|
|
ret = nvme_auth_generate_key(dhchap_secret, &key);
|
|
if (ret) {
|
|
kfree(dhchap_secret);
|
|
return ret;
|
|
}
|
|
kfree(opts->dhchap_ctrl_secret);
|
|
opts->dhchap_ctrl_secret = dhchap_secret;
|
|
ctrl_key = ctrl->ctrl_key;
|
|
mutex_lock(&ctrl->dhchap_auth_mutex);
|
|
ctrl->ctrl_key = key;
|
|
mutex_unlock(&ctrl->dhchap_auth_mutex);
|
|
nvme_auth_free_key(ctrl_key);
|
|
} else
|
|
kfree(dhchap_secret);
|
|
/* Start re-authentication */
|
|
dev_info(ctrl->device, "re-authenticating controller\n");
|
|
queue_work(nvme_wq, &ctrl->dhchap_auth_work);
|
|
|
|
return count;
|
|
}
|
|
|
|
static DEVICE_ATTR(dhchap_ctrl_secret, S_IRUGO | S_IWUSR,
|
|
nvme_ctrl_dhchap_ctrl_secret_show, nvme_ctrl_dhchap_ctrl_secret_store);
|
|
#endif
|
|
|
|
#ifdef CONFIG_NVME_TCP_TLS
|
|
static ssize_t tls_key_show(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev);
|
|
|
|
if (!ctrl->tls_key)
|
|
return 0;
|
|
return sysfs_emit(buf, "%08x", key_serial(ctrl->tls_key));
|
|
}
|
|
static DEVICE_ATTR_RO(tls_key);
|
|
#endif
|
|
|
|
static struct attribute *nvme_dev_attrs[] = {
|
|
&dev_attr_reset_controller.attr,
|
|
&dev_attr_rescan_controller.attr,
|
|
&dev_attr_model.attr,
|
|
&dev_attr_serial.attr,
|
|
&dev_attr_firmware_rev.attr,
|
|
&dev_attr_cntlid.attr,
|
|
&dev_attr_delete_controller.attr,
|
|
&dev_attr_transport.attr,
|
|
&dev_attr_subsysnqn.attr,
|
|
&dev_attr_address.attr,
|
|
&dev_attr_state.attr,
|
|
&dev_attr_numa_node.attr,
|
|
&dev_attr_queue_count.attr,
|
|
&dev_attr_sqsize.attr,
|
|
&dev_attr_hostnqn.attr,
|
|
&dev_attr_hostid.attr,
|
|
&dev_attr_ctrl_loss_tmo.attr,
|
|
&dev_attr_reconnect_delay.attr,
|
|
&dev_attr_fast_io_fail_tmo.attr,
|
|
&dev_attr_kato.attr,
|
|
&dev_attr_cntrltype.attr,
|
|
&dev_attr_dctype.attr,
|
|
#ifdef CONFIG_NVME_HOST_AUTH
|
|
&dev_attr_dhchap_secret.attr,
|
|
&dev_attr_dhchap_ctrl_secret.attr,
|
|
#endif
|
|
#ifdef CONFIG_NVME_TCP_TLS
|
|
&dev_attr_tls_key.attr,
|
|
#endif
|
|
&dev_attr_adm_passthru_err_log_enabled.attr,
|
|
NULL
|
|
};
|
|
|
|
static umode_t nvme_dev_attrs_are_visible(struct kobject *kobj,
|
|
struct attribute *a, int n)
|
|
{
|
|
struct device *dev = container_of(kobj, struct device, kobj);
|
|
struct nvme_ctrl *ctrl = dev_get_drvdata(dev);
|
|
|
|
if (a == &dev_attr_delete_controller.attr && !ctrl->ops->delete_ctrl)
|
|
return 0;
|
|
if (a == &dev_attr_address.attr && !ctrl->ops->get_address)
|
|
return 0;
|
|
if (a == &dev_attr_hostnqn.attr && !ctrl->opts)
|
|
return 0;
|
|
if (a == &dev_attr_hostid.attr && !ctrl->opts)
|
|
return 0;
|
|
if (a == &dev_attr_ctrl_loss_tmo.attr && !ctrl->opts)
|
|
return 0;
|
|
if (a == &dev_attr_reconnect_delay.attr && !ctrl->opts)
|
|
return 0;
|
|
if (a == &dev_attr_fast_io_fail_tmo.attr && !ctrl->opts)
|
|
return 0;
|
|
#ifdef CONFIG_NVME_HOST_AUTH
|
|
if (a == &dev_attr_dhchap_secret.attr && !ctrl->opts)
|
|
return 0;
|
|
if (a == &dev_attr_dhchap_ctrl_secret.attr && !ctrl->opts)
|
|
return 0;
|
|
#endif
|
|
#ifdef CONFIG_NVME_TCP_TLS
|
|
if (a == &dev_attr_tls_key.attr &&
|
|
(!ctrl->opts || strcmp(ctrl->opts->transport, "tcp")))
|
|
return 0;
|
|
#endif
|
|
|
|
return a->mode;
|
|
}
|
|
|
|
const struct attribute_group nvme_dev_attrs_group = {
|
|
.attrs = nvme_dev_attrs,
|
|
.is_visible = nvme_dev_attrs_are_visible,
|
|
};
|
|
EXPORT_SYMBOL_GPL(nvme_dev_attrs_group);
|
|
|
|
const struct attribute_group *nvme_dev_attr_groups[] = {
|
|
&nvme_dev_attrs_group,
|
|
NULL,
|
|
};
|
|
|
|
#define SUBSYS_ATTR_RO(_name, _mode, _show) \
|
|
struct device_attribute subsys_attr_##_name = \
|
|
__ATTR(_name, _mode, _show, NULL)
|
|
|
|
static ssize_t nvme_subsys_show_nqn(struct device *dev,
|
|
struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct nvme_subsystem *subsys =
|
|
container_of(dev, struct nvme_subsystem, dev);
|
|
|
|
return sysfs_emit(buf, "%s\n", subsys->subnqn);
|
|
}
|
|
static SUBSYS_ATTR_RO(subsysnqn, S_IRUGO, nvme_subsys_show_nqn);
|
|
|
|
static ssize_t nvme_subsys_show_type(struct device *dev,
|
|
struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct nvme_subsystem *subsys =
|
|
container_of(dev, struct nvme_subsystem, dev);
|
|
|
|
switch (subsys->subtype) {
|
|
case NVME_NQN_DISC:
|
|
return sysfs_emit(buf, "discovery\n");
|
|
case NVME_NQN_NVME:
|
|
return sysfs_emit(buf, "nvm\n");
|
|
default:
|
|
return sysfs_emit(buf, "reserved\n");
|
|
}
|
|
}
|
|
static SUBSYS_ATTR_RO(subsystype, S_IRUGO, nvme_subsys_show_type);
|
|
|
|
#define nvme_subsys_show_str_function(field) \
|
|
static ssize_t subsys_##field##_show(struct device *dev, \
|
|
struct device_attribute *attr, char *buf) \
|
|
{ \
|
|
struct nvme_subsystem *subsys = \
|
|
container_of(dev, struct nvme_subsystem, dev); \
|
|
return sysfs_emit(buf, "%.*s\n", \
|
|
(int)sizeof(subsys->field), subsys->field); \
|
|
} \
|
|
static SUBSYS_ATTR_RO(field, S_IRUGO, subsys_##field##_show);
|
|
|
|
nvme_subsys_show_str_function(model);
|
|
nvme_subsys_show_str_function(serial);
|
|
nvme_subsys_show_str_function(firmware_rev);
|
|
|
|
static struct attribute *nvme_subsys_attrs[] = {
|
|
&subsys_attr_model.attr,
|
|
&subsys_attr_serial.attr,
|
|
&subsys_attr_firmware_rev.attr,
|
|
&subsys_attr_subsysnqn.attr,
|
|
&subsys_attr_subsystype.attr,
|
|
#ifdef CONFIG_NVME_MULTIPATH
|
|
&subsys_attr_iopolicy.attr,
|
|
#endif
|
|
NULL,
|
|
};
|
|
|
|
static const struct attribute_group nvme_subsys_attrs_group = {
|
|
.attrs = nvme_subsys_attrs,
|
|
};
|
|
|
|
const struct attribute_group *nvme_subsys_attrs_groups[] = {
|
|
&nvme_subsys_attrs_group,
|
|
NULL,
|
|
};
|