mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-17 17:24:17 +08:00
004f1afbe1
The expectation is that the legacy / non-standard pmem discovery method (e820 type-12) will only ever be used to describe small quantities of persistent memory. Larger capacities will be described via the ACPI NFIT. When "allocate struct page from pmem" support is added this default policy can be overridden by assigning a legacy pmem namespace to a pfn device, however this would be only be necessary if a platform used the legacy mechanism to define a very large range. Cc: Christoph Hellwig <hch@lst.de> Signed-off-by: Dan Williams <dan.j.williams@intel.com>
282 lines
7.7 KiB
C
282 lines
7.7 KiB
C
/*
|
|
* Copyright(c) 2013-2015 Intel Corporation. All rights reserved.
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of version 2 of the GNU General Public License as
|
|
* published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope that it will be useful, but
|
|
* WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* General Public License for more details.
|
|
*/
|
|
#ifndef __ND_H__
|
|
#define __ND_H__
|
|
#include <linux/libnvdimm.h>
|
|
#include <linux/blkdev.h>
|
|
#include <linux/device.h>
|
|
#include <linux/mutex.h>
|
|
#include <linux/ndctl.h>
|
|
#include <linux/types.h>
|
|
#include "label.h"
|
|
|
|
enum {
|
|
/*
|
|
* Limits the maximum number of block apertures a dimm can
|
|
* support and is an input to the geometry/on-disk-format of a
|
|
* BTT instance
|
|
*/
|
|
ND_MAX_LANES = 256,
|
|
SECTOR_SHIFT = 9,
|
|
INT_LBASIZE_ALIGNMENT = 64,
|
|
#if IS_ENABLED(CONFIG_NVDIMM_PFN)
|
|
ND_PFN_ALIGN = PAGES_PER_SECTION * PAGE_SIZE,
|
|
ND_PFN_MASK = ND_PFN_ALIGN - 1,
|
|
#else
|
|
ND_PFN_ALIGN = 0,
|
|
ND_PFN_MASK = 0,
|
|
#endif
|
|
};
|
|
|
|
struct nvdimm_drvdata {
|
|
struct device *dev;
|
|
int nsindex_size;
|
|
struct nd_cmd_get_config_size nsarea;
|
|
void *data;
|
|
int ns_current, ns_next;
|
|
struct resource dpa;
|
|
struct kref kref;
|
|
};
|
|
|
|
struct nd_region_namespaces {
|
|
int count;
|
|
int active;
|
|
};
|
|
|
|
static inline struct nd_namespace_index *to_namespace_index(
|
|
struct nvdimm_drvdata *ndd, int i)
|
|
{
|
|
if (i < 0)
|
|
return NULL;
|
|
|
|
return ndd->data + sizeof_namespace_index(ndd) * i;
|
|
}
|
|
|
|
static inline struct nd_namespace_index *to_current_namespace_index(
|
|
struct nvdimm_drvdata *ndd)
|
|
{
|
|
return to_namespace_index(ndd, ndd->ns_current);
|
|
}
|
|
|
|
static inline struct nd_namespace_index *to_next_namespace_index(
|
|
struct nvdimm_drvdata *ndd)
|
|
{
|
|
return to_namespace_index(ndd, ndd->ns_next);
|
|
}
|
|
|
|
#define nd_dbg_dpa(r, d, res, fmt, arg...) \
|
|
dev_dbg((r) ? &(r)->dev : (d)->dev, "%s: %.13s: %#llx @ %#llx " fmt, \
|
|
(r) ? dev_name((d)->dev) : "", res ? res->name : "null", \
|
|
(unsigned long long) (res ? resource_size(res) : 0), \
|
|
(unsigned long long) (res ? res->start : 0), ##arg)
|
|
|
|
#define for_each_label(l, label, labels) \
|
|
for (l = 0; (label = labels ? labels[l] : NULL); l++)
|
|
|
|
#define for_each_dpa_resource(ndd, res) \
|
|
for (res = (ndd)->dpa.child; res; res = res->sibling)
|
|
|
|
#define for_each_dpa_resource_safe(ndd, res, next) \
|
|
for (res = (ndd)->dpa.child, next = res ? res->sibling : NULL; \
|
|
res; res = next, next = next ? next->sibling : NULL)
|
|
|
|
struct nd_percpu_lane {
|
|
int count;
|
|
spinlock_t lock;
|
|
};
|
|
|
|
struct nd_region {
|
|
struct device dev;
|
|
struct ida ns_ida;
|
|
struct ida btt_ida;
|
|
struct ida pfn_ida;
|
|
unsigned long flags;
|
|
struct device *ns_seed;
|
|
struct device *btt_seed;
|
|
struct device *pfn_seed;
|
|
u16 ndr_mappings;
|
|
u64 ndr_size;
|
|
u64 ndr_start;
|
|
int id, num_lanes, ro, numa_node;
|
|
void *provider_data;
|
|
struct nd_interleave_set *nd_set;
|
|
struct nd_percpu_lane __percpu *lane;
|
|
struct nd_mapping mapping[0];
|
|
};
|
|
|
|
struct nd_blk_region {
|
|
int (*enable)(struct nvdimm_bus *nvdimm_bus, struct device *dev);
|
|
void (*disable)(struct nvdimm_bus *nvdimm_bus, struct device *dev);
|
|
int (*do_io)(struct nd_blk_region *ndbr, resource_size_t dpa,
|
|
void *iobuf, u64 len, int rw);
|
|
void *blk_provider_data;
|
|
struct nd_region nd_region;
|
|
};
|
|
|
|
/*
|
|
* Lookup next in the repeating sequence of 01, 10, and 11.
|
|
*/
|
|
static inline unsigned nd_inc_seq(unsigned seq)
|
|
{
|
|
static const unsigned next[] = { 0, 2, 3, 1 };
|
|
|
|
return next[seq & 3];
|
|
}
|
|
|
|
struct btt;
|
|
struct nd_btt {
|
|
struct device dev;
|
|
struct nd_namespace_common *ndns;
|
|
struct btt *btt;
|
|
unsigned long lbasize;
|
|
u8 *uuid;
|
|
int id;
|
|
};
|
|
|
|
enum nd_pfn_mode {
|
|
PFN_MODE_NONE,
|
|
PFN_MODE_RAM,
|
|
PFN_MODE_PMEM,
|
|
};
|
|
|
|
struct nd_pfn {
|
|
int id;
|
|
u8 *uuid;
|
|
struct device dev;
|
|
unsigned long npfns;
|
|
enum nd_pfn_mode mode;
|
|
struct nd_pfn_sb *pfn_sb;
|
|
struct nd_namespace_common *ndns;
|
|
};
|
|
|
|
enum nd_async_mode {
|
|
ND_SYNC,
|
|
ND_ASYNC,
|
|
};
|
|
|
|
int nd_integrity_init(struct gendisk *disk, unsigned long meta_size);
|
|
void wait_nvdimm_bus_probe_idle(struct device *dev);
|
|
void nd_device_register(struct device *dev);
|
|
void nd_device_unregister(struct device *dev, enum nd_async_mode mode);
|
|
int nd_uuid_store(struct device *dev, u8 **uuid_out, const char *buf,
|
|
size_t len);
|
|
ssize_t nd_sector_size_show(unsigned long current_lbasize,
|
|
const unsigned long *supported, char *buf);
|
|
ssize_t nd_sector_size_store(struct device *dev, const char *buf,
|
|
unsigned long *current_lbasize, const unsigned long *supported);
|
|
int __init nvdimm_init(void);
|
|
int __init nd_region_init(void);
|
|
void nvdimm_exit(void);
|
|
void nd_region_exit(void);
|
|
struct nvdimm;
|
|
struct nvdimm_drvdata *to_ndd(struct nd_mapping *nd_mapping);
|
|
int nvdimm_init_nsarea(struct nvdimm_drvdata *ndd);
|
|
int nvdimm_init_config_data(struct nvdimm_drvdata *ndd);
|
|
int nvdimm_set_config_data(struct nvdimm_drvdata *ndd, size_t offset,
|
|
void *buf, size_t len);
|
|
struct nd_btt *to_nd_btt(struct device *dev);
|
|
|
|
struct nd_gen_sb {
|
|
char reserved[SZ_4K - 8];
|
|
__le64 checksum;
|
|
};
|
|
|
|
u64 nd_sb_checksum(struct nd_gen_sb *sb);
|
|
#if IS_ENABLED(CONFIG_BTT)
|
|
int nd_btt_probe(struct nd_namespace_common *ndns, void *drvdata);
|
|
bool is_nd_btt(struct device *dev);
|
|
struct device *nd_btt_create(struct nd_region *nd_region);
|
|
#else
|
|
static inline int nd_btt_probe(struct nd_namespace_common *ndns, void *drvdata)
|
|
{
|
|
return -ENODEV;
|
|
}
|
|
|
|
static inline bool is_nd_btt(struct device *dev)
|
|
{
|
|
return false;
|
|
}
|
|
|
|
static inline struct device *nd_btt_create(struct nd_region *nd_region)
|
|
{
|
|
return NULL;
|
|
}
|
|
#endif
|
|
|
|
struct nd_pfn *to_nd_pfn(struct device *dev);
|
|
#if IS_ENABLED(CONFIG_NVDIMM_PFN)
|
|
int nd_pfn_probe(struct nd_namespace_common *ndns, void *drvdata);
|
|
bool is_nd_pfn(struct device *dev);
|
|
struct device *nd_pfn_create(struct nd_region *nd_region);
|
|
int nd_pfn_validate(struct nd_pfn *nd_pfn);
|
|
#else
|
|
static inline int nd_pfn_probe(struct nd_namespace_common *ndns, void *drvdata)
|
|
{
|
|
return -ENODEV;
|
|
}
|
|
|
|
static inline bool is_nd_pfn(struct device *dev)
|
|
{
|
|
return false;
|
|
}
|
|
|
|
static inline struct device *nd_pfn_create(struct nd_region *nd_region)
|
|
{
|
|
return NULL;
|
|
}
|
|
|
|
static inline int nd_pfn_validate(struct nd_pfn *nd_pfn)
|
|
{
|
|
return -ENODEV;
|
|
}
|
|
#endif
|
|
|
|
struct nd_region *to_nd_region(struct device *dev);
|
|
int nd_region_to_nstype(struct nd_region *nd_region);
|
|
int nd_region_register_namespaces(struct nd_region *nd_region, int *err);
|
|
u64 nd_region_interleave_set_cookie(struct nd_region *nd_region);
|
|
void nvdimm_bus_lock(struct device *dev);
|
|
void nvdimm_bus_unlock(struct device *dev);
|
|
bool is_nvdimm_bus_locked(struct device *dev);
|
|
int nvdimm_revalidate_disk(struct gendisk *disk);
|
|
void nvdimm_drvdata_release(struct kref *kref);
|
|
void put_ndd(struct nvdimm_drvdata *ndd);
|
|
int nd_label_reserve_dpa(struct nvdimm_drvdata *ndd);
|
|
void nvdimm_free_dpa(struct nvdimm_drvdata *ndd, struct resource *res);
|
|
struct resource *nvdimm_allocate_dpa(struct nvdimm_drvdata *ndd,
|
|
struct nd_label_id *label_id, resource_size_t start,
|
|
resource_size_t n);
|
|
resource_size_t nvdimm_namespace_capacity(struct nd_namespace_common *ndns);
|
|
struct nd_namespace_common *nvdimm_namespace_common_probe(struct device *dev);
|
|
int nvdimm_namespace_attach_btt(struct nd_namespace_common *ndns);
|
|
int nvdimm_namespace_detach_btt(struct nd_namespace_common *ndns);
|
|
const char *nvdimm_namespace_disk_name(struct nd_namespace_common *ndns,
|
|
char *name);
|
|
int nd_blk_region_init(struct nd_region *nd_region);
|
|
void __nd_iostat_start(struct bio *bio, unsigned long *start);
|
|
static inline bool nd_iostat_start(struct bio *bio, unsigned long *start)
|
|
{
|
|
struct gendisk *disk = bio->bi_bdev->bd_disk;
|
|
|
|
if (!blk_queue_io_stat(disk->queue))
|
|
return false;
|
|
|
|
__nd_iostat_start(bio, start);
|
|
return true;
|
|
}
|
|
void nd_iostat_end(struct bio *bio, unsigned long start);
|
|
resource_size_t nd_namespace_blk_validate(struct nd_namespace_blk *nsblk);
|
|
const u8 *nd_dev_to_uuid(struct device *dev);
|
|
bool pmem_should_map_pages(struct device *dev);
|
|
#endif /* __ND_H__ */
|