mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-12-25 03:55:09 +08:00
19398821b2
The cxl_pmem.ko module houses the driver for both cxl_nvdimm_bridge
objects and cxl_nvdimm objects. When the core creates a cxl_nvdimm it
arranges for it to be autoremoved when the bridge goes down. However, if
the bridge never initialized because the cxl_pmem.ko module never
loaded, it sets up a the following crash scenario:
BUG: kernel NULL pointer dereference, address: 0000000000000478
[..]
RIP: 0010:cxl_nvdimm_probe+0x99/0x140 [cxl_pmem]
[..]
Call Trace:
<TASK>
cxl_bus_probe+0x17/0x50 [cxl_core]
really_probe+0xde/0x380
__driver_probe_device+0x78/0x170
driver_probe_device+0x1f/0x90
__driver_attach+0xd2/0x1c0
bus_for_each_dev+0x79/0xc0
bus_add_driver+0x1b1/0x200
driver_register+0x89/0xe0
cxl_pmem_init+0x50/0xff0 [cxl_pmem]
It turns out the recent rework to simplify nvdimm probing obviated the
need to unregister cxl_nvdimm objects at cxl_nvdimm_bridge ->remove()
time. Leave the cxl_nvdimm device registered until the hosting
cxl_memdev departs. The alternative is that the cxl_memdev needs to be
reattached whenever the cxl_nvdimm_bridge attach state cycles, which is
awkward and unnecessary.
The only requirement is to make sure that when the cxl_nvdimm_bridge
goes away any dependent cxl_nvdimm objects are shutdown. Handle that in
unregister_nvdimm_bus().
With these registration entanglements removed there is no longer a need
to pre-load the cxl_pmem module in cxl_acpi.
Fixes: cb9cfff82f
("cxl/acpi: Simplify cxl_nvdimm_bridge probing")
Reported-by: Gregory Price <gregory.price@memverge.com>
Debugged-by: Jonathan Cameron <Jonathan.Cameron@huawei.com>
Tested-by: Jonathan Cameron <Jonathan.Cameron@huawei.com>
Reviewed-by: Jonathan Cameron <Jonathan.Cameron@huawei.com>
Reviewed-by: Dave Jiang <dave.jiang@intel.com>
Link: https://lore.kernel.org/r/167426077263.3955046.9695309346988027311.stgit@dwillia2-xfh.jf.intel.com
Signed-off-by: Dan Williams <dan.j.williams@intel.com>
291 lines
6.9 KiB
C
291 lines
6.9 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/* Copyright(c) 2020 Intel Corporation. */
|
|
#include <linux/device.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/idr.h>
|
|
#include <cxlmem.h>
|
|
#include <cxl.h>
|
|
#include "core.h"
|
|
|
|
/**
|
|
* DOC: cxl pmem
|
|
*
|
|
* The core CXL PMEM infrastructure supports persistent memory
|
|
* provisioning and serves as a bridge to the LIBNVDIMM subsystem. A CXL
|
|
* 'bridge' device is added at the root of a CXL device topology if
|
|
* platform firmware advertises at least one persistent memory capable
|
|
* CXL window. That root-level bridge corresponds to a LIBNVDIMM 'bus'
|
|
* device. Then for each cxl_memdev in the CXL device topology a bridge
|
|
* device is added to host a LIBNVDIMM dimm object. When these bridges
|
|
* are registered native LIBNVDIMM uapis are translated to CXL
|
|
* operations, for example, namespace label access commands.
|
|
*/
|
|
|
|
static DEFINE_IDA(cxl_nvdimm_bridge_ida);
|
|
|
|
static void cxl_nvdimm_bridge_release(struct device *dev)
|
|
{
|
|
struct cxl_nvdimm_bridge *cxl_nvb = to_cxl_nvdimm_bridge(dev);
|
|
|
|
ida_free(&cxl_nvdimm_bridge_ida, cxl_nvb->id);
|
|
kfree(cxl_nvb);
|
|
}
|
|
|
|
static const struct attribute_group *cxl_nvdimm_bridge_attribute_groups[] = {
|
|
&cxl_base_attribute_group,
|
|
NULL,
|
|
};
|
|
|
|
const struct device_type cxl_nvdimm_bridge_type = {
|
|
.name = "cxl_nvdimm_bridge",
|
|
.release = cxl_nvdimm_bridge_release,
|
|
.groups = cxl_nvdimm_bridge_attribute_groups,
|
|
};
|
|
|
|
struct cxl_nvdimm_bridge *to_cxl_nvdimm_bridge(struct device *dev)
|
|
{
|
|
if (dev_WARN_ONCE(dev, dev->type != &cxl_nvdimm_bridge_type,
|
|
"not a cxl_nvdimm_bridge device\n"))
|
|
return NULL;
|
|
return container_of(dev, struct cxl_nvdimm_bridge, dev);
|
|
}
|
|
EXPORT_SYMBOL_NS_GPL(to_cxl_nvdimm_bridge, CXL);
|
|
|
|
bool is_cxl_nvdimm_bridge(struct device *dev)
|
|
{
|
|
return dev->type == &cxl_nvdimm_bridge_type;
|
|
}
|
|
EXPORT_SYMBOL_NS_GPL(is_cxl_nvdimm_bridge, CXL);
|
|
|
|
static int match_nvdimm_bridge(struct device *dev, void *data)
|
|
{
|
|
return is_cxl_nvdimm_bridge(dev);
|
|
}
|
|
|
|
struct cxl_nvdimm_bridge *cxl_find_nvdimm_bridge(struct device *start)
|
|
{
|
|
struct cxl_port *port = find_cxl_root(start);
|
|
struct device *dev;
|
|
|
|
if (!port)
|
|
return NULL;
|
|
|
|
dev = device_find_child(&port->dev, NULL, match_nvdimm_bridge);
|
|
put_device(&port->dev);
|
|
|
|
if (!dev)
|
|
return NULL;
|
|
|
|
return to_cxl_nvdimm_bridge(dev);
|
|
}
|
|
EXPORT_SYMBOL_NS_GPL(cxl_find_nvdimm_bridge, CXL);
|
|
|
|
static struct lock_class_key cxl_nvdimm_bridge_key;
|
|
|
|
static struct cxl_nvdimm_bridge *cxl_nvdimm_bridge_alloc(struct cxl_port *port)
|
|
{
|
|
struct cxl_nvdimm_bridge *cxl_nvb;
|
|
struct device *dev;
|
|
int rc;
|
|
|
|
cxl_nvb = kzalloc(sizeof(*cxl_nvb), GFP_KERNEL);
|
|
if (!cxl_nvb)
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
rc = ida_alloc(&cxl_nvdimm_bridge_ida, GFP_KERNEL);
|
|
if (rc < 0)
|
|
goto err;
|
|
cxl_nvb->id = rc;
|
|
|
|
dev = &cxl_nvb->dev;
|
|
cxl_nvb->port = port;
|
|
device_initialize(dev);
|
|
lockdep_set_class(&dev->mutex, &cxl_nvdimm_bridge_key);
|
|
device_set_pm_not_required(dev);
|
|
dev->parent = &port->dev;
|
|
dev->bus = &cxl_bus_type;
|
|
dev->type = &cxl_nvdimm_bridge_type;
|
|
|
|
return cxl_nvb;
|
|
|
|
err:
|
|
kfree(cxl_nvb);
|
|
return ERR_PTR(rc);
|
|
}
|
|
|
|
static void unregister_nvb(void *_cxl_nvb)
|
|
{
|
|
struct cxl_nvdimm_bridge *cxl_nvb = _cxl_nvb;
|
|
|
|
device_unregister(&cxl_nvb->dev);
|
|
}
|
|
|
|
/**
|
|
* devm_cxl_add_nvdimm_bridge() - add the root of a LIBNVDIMM topology
|
|
* @host: platform firmware root device
|
|
* @port: CXL port at the root of a CXL topology
|
|
*
|
|
* Return: bridge device that can host cxl_nvdimm objects
|
|
*/
|
|
struct cxl_nvdimm_bridge *devm_cxl_add_nvdimm_bridge(struct device *host,
|
|
struct cxl_port *port)
|
|
{
|
|
struct cxl_nvdimm_bridge *cxl_nvb;
|
|
struct device *dev;
|
|
int rc;
|
|
|
|
if (!IS_ENABLED(CONFIG_CXL_PMEM))
|
|
return ERR_PTR(-ENXIO);
|
|
|
|
cxl_nvb = cxl_nvdimm_bridge_alloc(port);
|
|
if (IS_ERR(cxl_nvb))
|
|
return cxl_nvb;
|
|
|
|
dev = &cxl_nvb->dev;
|
|
rc = dev_set_name(dev, "nvdimm-bridge%d", cxl_nvb->id);
|
|
if (rc)
|
|
goto err;
|
|
|
|
rc = device_add(dev);
|
|
if (rc)
|
|
goto err;
|
|
|
|
rc = devm_add_action_or_reset(host, unregister_nvb, cxl_nvb);
|
|
if (rc)
|
|
return ERR_PTR(rc);
|
|
|
|
return cxl_nvb;
|
|
|
|
err:
|
|
put_device(dev);
|
|
return ERR_PTR(rc);
|
|
}
|
|
EXPORT_SYMBOL_NS_GPL(devm_cxl_add_nvdimm_bridge, CXL);
|
|
|
|
static void cxl_nvdimm_release(struct device *dev)
|
|
{
|
|
struct cxl_nvdimm *cxl_nvd = to_cxl_nvdimm(dev);
|
|
|
|
kfree(cxl_nvd);
|
|
}
|
|
|
|
static const struct attribute_group *cxl_nvdimm_attribute_groups[] = {
|
|
&cxl_base_attribute_group,
|
|
NULL,
|
|
};
|
|
|
|
const struct device_type cxl_nvdimm_type = {
|
|
.name = "cxl_nvdimm",
|
|
.release = cxl_nvdimm_release,
|
|
.groups = cxl_nvdimm_attribute_groups,
|
|
};
|
|
|
|
bool is_cxl_nvdimm(struct device *dev)
|
|
{
|
|
return dev->type == &cxl_nvdimm_type;
|
|
}
|
|
EXPORT_SYMBOL_NS_GPL(is_cxl_nvdimm, CXL);
|
|
|
|
struct cxl_nvdimm *to_cxl_nvdimm(struct device *dev)
|
|
{
|
|
if (dev_WARN_ONCE(dev, !is_cxl_nvdimm(dev),
|
|
"not a cxl_nvdimm device\n"))
|
|
return NULL;
|
|
return container_of(dev, struct cxl_nvdimm, dev);
|
|
}
|
|
EXPORT_SYMBOL_NS_GPL(to_cxl_nvdimm, CXL);
|
|
|
|
static struct lock_class_key cxl_nvdimm_key;
|
|
|
|
static struct cxl_nvdimm *cxl_nvdimm_alloc(struct cxl_nvdimm_bridge *cxl_nvb,
|
|
struct cxl_memdev *cxlmd)
|
|
{
|
|
struct cxl_nvdimm *cxl_nvd;
|
|
struct device *dev;
|
|
|
|
cxl_nvd = kzalloc(sizeof(*cxl_nvd), GFP_KERNEL);
|
|
if (!cxl_nvd)
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
dev = &cxl_nvd->dev;
|
|
cxl_nvd->cxlmd = cxlmd;
|
|
cxlmd->cxl_nvd = cxl_nvd;
|
|
device_initialize(dev);
|
|
lockdep_set_class(&dev->mutex, &cxl_nvdimm_key);
|
|
device_set_pm_not_required(dev);
|
|
dev->parent = &cxlmd->dev;
|
|
dev->bus = &cxl_bus_type;
|
|
dev->type = &cxl_nvdimm_type;
|
|
/*
|
|
* A "%llx" string is 17-bytes vs dimm_id that is max
|
|
* NVDIMM_KEY_DESC_LEN
|
|
*/
|
|
BUILD_BUG_ON(sizeof(cxl_nvd->dev_id) < 17 ||
|
|
sizeof(cxl_nvd->dev_id) > NVDIMM_KEY_DESC_LEN);
|
|
sprintf(cxl_nvd->dev_id, "%llx", cxlmd->cxlds->serial);
|
|
|
|
return cxl_nvd;
|
|
}
|
|
|
|
static void cxlmd_release_nvdimm(void *_cxlmd)
|
|
{
|
|
struct cxl_memdev *cxlmd = _cxlmd;
|
|
struct cxl_nvdimm *cxl_nvd = cxlmd->cxl_nvd;
|
|
struct cxl_nvdimm_bridge *cxl_nvb = cxlmd->cxl_nvb;
|
|
|
|
cxl_nvd->cxlmd = NULL;
|
|
cxlmd->cxl_nvd = NULL;
|
|
cxlmd->cxl_nvb = NULL;
|
|
device_unregister(&cxl_nvd->dev);
|
|
put_device(&cxl_nvb->dev);
|
|
}
|
|
|
|
/**
|
|
* devm_cxl_add_nvdimm() - add a bridge between a cxl_memdev and an nvdimm
|
|
* @cxlmd: cxl_memdev instance that will perform LIBNVDIMM operations
|
|
*
|
|
* Return: 0 on success negative error code on failure.
|
|
*/
|
|
int devm_cxl_add_nvdimm(struct cxl_memdev *cxlmd)
|
|
{
|
|
struct cxl_nvdimm_bridge *cxl_nvb;
|
|
struct cxl_nvdimm *cxl_nvd;
|
|
struct device *dev;
|
|
int rc;
|
|
|
|
cxl_nvb = cxl_find_nvdimm_bridge(&cxlmd->dev);
|
|
if (!cxl_nvb)
|
|
return -ENODEV;
|
|
|
|
cxl_nvd = cxl_nvdimm_alloc(cxl_nvb, cxlmd);
|
|
if (IS_ERR(cxl_nvd)) {
|
|
rc = PTR_ERR(cxl_nvd);
|
|
goto err_alloc;
|
|
}
|
|
cxlmd->cxl_nvb = cxl_nvb;
|
|
|
|
dev = &cxl_nvd->dev;
|
|
rc = dev_set_name(dev, "pmem%d", cxlmd->id);
|
|
if (rc)
|
|
goto err;
|
|
|
|
rc = device_add(dev);
|
|
if (rc)
|
|
goto err;
|
|
|
|
dev_dbg(&cxlmd->dev, "register %s\n", dev_name(dev));
|
|
|
|
/* @cxlmd carries a reference on @cxl_nvb until cxlmd_release_nvdimm */
|
|
return devm_add_action_or_reset(&cxlmd->dev, cxlmd_release_nvdimm, cxlmd);
|
|
|
|
err:
|
|
put_device(dev);
|
|
err_alloc:
|
|
cxlmd->cxl_nvb = NULL;
|
|
cxlmd->cxl_nvd = NULL;
|
|
put_device(&cxl_nvb->dev);
|
|
|
|
return rc;
|
|
}
|
|
EXPORT_SYMBOL_NS_GPL(devm_cxl_add_nvdimm, CXL);
|