mirror of
git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-09-04 20:19:47 +08:00

Add CXL RAS Features support. Features include "patrol scrub control", "error check scrub", "perform maintenance", and "memory sparing". This support connects the RAS Featurs to EDAC.
140 lines
4.6 KiB
C
140 lines
4.6 KiB
C
/* SPDX-License-Identifier: GPL-2.0-only */
|
|
/* Copyright(c) 2020 Intel Corporation. */
|
|
|
|
#ifndef __CXL_CORE_H__
|
|
#define __CXL_CORE_H__
|
|
|
|
#include <cxl/mailbox.h>
|
|
|
|
extern const struct device_type cxl_nvdimm_bridge_type;
|
|
extern const struct device_type cxl_nvdimm_type;
|
|
extern const struct device_type cxl_pmu_type;
|
|
|
|
extern struct attribute_group cxl_base_attribute_group;
|
|
|
|
#ifdef CONFIG_CXL_REGION
|
|
extern struct device_attribute dev_attr_create_pmem_region;
|
|
extern struct device_attribute dev_attr_create_ram_region;
|
|
extern struct device_attribute dev_attr_delete_region;
|
|
extern struct device_attribute dev_attr_region;
|
|
extern const struct device_type cxl_pmem_region_type;
|
|
extern const struct device_type cxl_dax_region_type;
|
|
extern const struct device_type cxl_region_type;
|
|
void cxl_decoder_kill_region(struct cxl_endpoint_decoder *cxled);
|
|
#define CXL_REGION_ATTR(x) (&dev_attr_##x.attr)
|
|
#define CXL_REGION_TYPE(x) (&cxl_region_type)
|
|
#define SET_CXL_REGION_ATTR(x) (&dev_attr_##x.attr),
|
|
#define CXL_PMEM_REGION_TYPE(x) (&cxl_pmem_region_type)
|
|
#define CXL_DAX_REGION_TYPE(x) (&cxl_dax_region_type)
|
|
int cxl_region_init(void);
|
|
void cxl_region_exit(void);
|
|
int cxl_get_poison_by_endpoint(struct cxl_port *port);
|
|
struct cxl_region *cxl_dpa_to_region(const struct cxl_memdev *cxlmd, u64 dpa);
|
|
u64 cxl_dpa_to_hpa(struct cxl_region *cxlr, const struct cxl_memdev *cxlmd,
|
|
u64 dpa);
|
|
|
|
#else
|
|
static inline u64 cxl_dpa_to_hpa(struct cxl_region *cxlr,
|
|
const struct cxl_memdev *cxlmd, u64 dpa)
|
|
{
|
|
return ULLONG_MAX;
|
|
}
|
|
static inline
|
|
struct cxl_region *cxl_dpa_to_region(const struct cxl_memdev *cxlmd, u64 dpa)
|
|
{
|
|
return NULL;
|
|
}
|
|
static inline int cxl_get_poison_by_endpoint(struct cxl_port *port)
|
|
{
|
|
return 0;
|
|
}
|
|
static inline void cxl_decoder_kill_region(struct cxl_endpoint_decoder *cxled)
|
|
{
|
|
}
|
|
static inline int cxl_region_init(void)
|
|
{
|
|
return 0;
|
|
}
|
|
static inline void cxl_region_exit(void)
|
|
{
|
|
}
|
|
#define CXL_REGION_ATTR(x) NULL
|
|
#define CXL_REGION_TYPE(x) NULL
|
|
#define SET_CXL_REGION_ATTR(x)
|
|
#define CXL_PMEM_REGION_TYPE(x) NULL
|
|
#define CXL_DAX_REGION_TYPE(x) NULL
|
|
#endif
|
|
|
|
struct cxl_send_command;
|
|
struct cxl_mem_query_commands;
|
|
int cxl_query_cmd(struct cxl_mailbox *cxl_mbox,
|
|
struct cxl_mem_query_commands __user *q);
|
|
int cxl_send_cmd(struct cxl_mailbox *cxl_mbox, struct cxl_send_command __user *s);
|
|
void __iomem *devm_cxl_iomap_block(struct device *dev, resource_size_t addr,
|
|
resource_size_t length);
|
|
|
|
struct dentry *cxl_debugfs_create_dir(const char *dir);
|
|
int cxl_dpa_set_part(struct cxl_endpoint_decoder *cxled,
|
|
enum cxl_partition_mode mode);
|
|
int cxl_dpa_alloc(struct cxl_endpoint_decoder *cxled, u64 size);
|
|
int cxl_dpa_free(struct cxl_endpoint_decoder *cxled);
|
|
resource_size_t cxl_dpa_size(struct cxl_endpoint_decoder *cxled);
|
|
resource_size_t cxl_dpa_resource_start(struct cxl_endpoint_decoder *cxled);
|
|
|
|
enum cxl_rcrb {
|
|
CXL_RCRB_DOWNSTREAM,
|
|
CXL_RCRB_UPSTREAM,
|
|
};
|
|
struct cxl_rcrb_info;
|
|
resource_size_t __rcrb_to_component(struct device *dev,
|
|
struct cxl_rcrb_info *ri,
|
|
enum cxl_rcrb which);
|
|
u16 cxl_rcrb_to_aer(struct device *dev, resource_size_t rcrb);
|
|
|
|
#define PCI_RCRB_CAP_LIST_ID_MASK GENMASK(7, 0)
|
|
#define PCI_RCRB_CAP_HDR_ID_MASK GENMASK(7, 0)
|
|
#define PCI_RCRB_CAP_HDR_NEXT_MASK GENMASK(15, 8)
|
|
#define PCI_CAP_EXP_SIZEOF 0x3c
|
|
|
|
extern struct rw_semaphore cxl_dpa_rwsem;
|
|
extern struct rw_semaphore cxl_region_rwsem;
|
|
|
|
int cxl_memdev_init(void);
|
|
void cxl_memdev_exit(void);
|
|
void cxl_mbox_init(void);
|
|
|
|
enum cxl_poison_trace_type {
|
|
CXL_POISON_TRACE_LIST,
|
|
CXL_POISON_TRACE_INJECT,
|
|
CXL_POISON_TRACE_CLEAR,
|
|
};
|
|
|
|
long cxl_pci_get_latency(struct pci_dev *pdev);
|
|
int cxl_pci_get_bandwidth(struct pci_dev *pdev, struct access_coordinate *c);
|
|
int cxl_update_hmat_access_coordinates(int nid, struct cxl_region *cxlr,
|
|
enum access_coordinate_class access);
|
|
bool cxl_need_node_perf_attrs_update(int nid);
|
|
int cxl_port_get_switch_dport_bandwidth(struct cxl_port *port,
|
|
struct access_coordinate *c);
|
|
|
|
int cxl_ras_init(void);
|
|
void cxl_ras_exit(void);
|
|
int cxl_gpf_port_setup(struct cxl_dport *dport);
|
|
int cxl_acpi_get_extended_linear_cache_size(struct resource *backing_res,
|
|
int nid, resource_size_t *size);
|
|
|
|
#ifdef CONFIG_CXL_FEATURES
|
|
struct cxl_feat_entry *
|
|
cxl_feature_info(struct cxl_features_state *cxlfs, const uuid_t *uuid);
|
|
size_t cxl_get_feature(struct cxl_mailbox *cxl_mbox, const uuid_t *feat_uuid,
|
|
enum cxl_get_feat_selection selection,
|
|
void *feat_out, size_t feat_out_size, u16 offset,
|
|
u16 *return_code);
|
|
int cxl_set_feature(struct cxl_mailbox *cxl_mbox, const uuid_t *feat_uuid,
|
|
u8 feat_version, const void *feat_data,
|
|
size_t feat_data_size, u32 feat_flag, u16 offset,
|
|
u16 *return_code);
|
|
#endif
|
|
|
|
#endif /* __CXL_CORE_H__ */
|