| /* SPDX-License-Identifier: GPL-2.0-only */ |
| /* |
| * Copyright(c) 2013-2015 Intel Corporation. All rights reserved. |
| */ |
| #ifndef __ND_CORE_H__ |
| #define __ND_CORE_H__ |
| #include <linux/libnvdimm.h> |
| #include <linux/device.h> |
| #include <linux/sizes.h> |
| #include <linux/mutex.h> |
| #include <linux/nd.h> |
| |
| extern struct list_head nvdimm_bus_list; |
| extern struct mutex nvdimm_bus_list_mutex; |
| extern int nvdimm_major; |
| extern struct workqueue_struct *nvdimm_wq; |
| |
| struct nvdimm_bus { |
| struct nvdimm_bus_descriptor *nd_desc; |
| wait_queue_head_t probe_wait; |
| struct list_head list; |
| struct device dev; |
| int id, probe_active; |
| struct list_head mapping_list; |
| struct mutex reconfig_mutex; |
| struct badrange badrange; |
| }; |
| |
| struct nvdimm { |
| unsigned long flags; |
| void *provider_data; |
| unsigned long cmd_mask; |
| struct device dev; |
| atomic_t busy; |
| int id, num_flush; |
| struct resource *flush_wpq; |
| const char *dimm_id; |
| struct { |
| const struct nvdimm_security_ops *ops; |
| enum nvdimm_security_state state; |
| enum nvdimm_security_state ext_state; |
| unsigned int overwrite_tmo; |
| struct kernfs_node *overwrite_state; |
| } sec; |
| struct delayed_work dwork; |
| }; |
| |
| static inline enum nvdimm_security_state nvdimm_security_state( |
| struct nvdimm *nvdimm, enum nvdimm_passphrase_type ptype) |
| { |
| if (!nvdimm->sec.ops) |
| return -ENXIO; |
| |
| return nvdimm->sec.ops->state(nvdimm, ptype); |
| } |
| int nvdimm_security_freeze(struct nvdimm *nvdimm); |
| #if IS_ENABLED(CONFIG_NVDIMM_KEYS) |
| int nvdimm_security_disable(struct nvdimm *nvdimm, unsigned int keyid); |
| int nvdimm_security_update(struct nvdimm *nvdimm, unsigned int keyid, |
| unsigned int new_keyid, |
| enum nvdimm_passphrase_type pass_type); |
| int nvdimm_security_erase(struct nvdimm *nvdimm, unsigned int keyid, |
| enum nvdimm_passphrase_type pass_type); |
| int nvdimm_security_overwrite(struct nvdimm *nvdimm, unsigned int keyid); |
| void nvdimm_security_overwrite_query(struct work_struct *work); |
| #else |
| static inline int nvdimm_security_disable(struct nvdimm *nvdimm, |
| unsigned int keyid) |
| { |
| return -EOPNOTSUPP; |
| } |
| static inline int nvdimm_security_update(struct nvdimm *nvdimm, |
| unsigned int keyid, |
| unsigned int new_keyid, |
| enum nvdimm_passphrase_type pass_type) |
| { |
| return -EOPNOTSUPP; |
| } |
| static inline int nvdimm_security_erase(struct nvdimm *nvdimm, |
| unsigned int keyid, |
| enum nvdimm_passphrase_type pass_type) |
| { |
| return -EOPNOTSUPP; |
| } |
| static inline int nvdimm_security_overwrite(struct nvdimm *nvdimm, |
| unsigned int keyid) |
| { |
| return -EOPNOTSUPP; |
| } |
| static inline void nvdimm_security_overwrite_query(struct work_struct *work) |
| { |
| } |
| #endif |
| |
| /** |
| * struct blk_alloc_info - tracking info for BLK dpa scanning |
| * @nd_mapping: blk region mapping boundaries |
| * @available: decremented in alias_dpa_busy as aliased PMEM is scanned |
| * @busy: decremented in blk_dpa_busy to account for ranges already |
| * handled by alias_dpa_busy |
| * @res: alias_dpa_busy interprets this a free space range that needs to |
| * be truncated to the valid BLK allocation starting DPA, blk_dpa_busy |
| * treats it as a busy range that needs the aliased PMEM ranges |
| * truncated. |
| */ |
| struct blk_alloc_info { |
| struct nd_mapping *nd_mapping; |
| resource_size_t available, busy; |
| struct resource *res; |
| }; |
| |
| bool is_nvdimm(struct device *dev); |
| bool is_nd_pmem(struct device *dev); |
| bool is_nd_volatile(struct device *dev); |
| bool is_nd_blk(struct device *dev); |
| static inline bool is_nd_region(struct device *dev) |
| { |
| return is_nd_pmem(dev) || is_nd_blk(dev) || is_nd_volatile(dev); |
| } |
| static inline bool is_memory(struct device *dev) |
| { |
| return is_nd_pmem(dev) || is_nd_volatile(dev); |
| } |
| struct nvdimm_bus *walk_to_nvdimm_bus(struct device *nd_dev); |
| int __init nvdimm_bus_init(void); |
| void nvdimm_bus_exit(void); |
| void nvdimm_devs_exit(void); |
| void nd_region_devs_exit(void); |
| void nd_region_probe_success(struct nvdimm_bus *nvdimm_bus, struct device *dev); |
| struct nd_region; |
| void nd_region_create_ns_seed(struct nd_region *nd_region); |
| void nd_region_create_btt_seed(struct nd_region *nd_region); |
| void nd_region_create_pfn_seed(struct nd_region *nd_region); |
| void nd_region_create_dax_seed(struct nd_region *nd_region); |
| void nd_region_disable(struct nvdimm_bus *nvdimm_bus, struct device *dev); |
| int nvdimm_bus_create_ndctl(struct nvdimm_bus *nvdimm_bus); |
| void nvdimm_bus_destroy_ndctl(struct nvdimm_bus *nvdimm_bus); |
| void nd_synchronize(void); |
| int nvdimm_bus_register_dimms(struct nvdimm_bus *nvdimm_bus); |
| int nvdimm_bus_register_regions(struct nvdimm_bus *nvdimm_bus); |
| int nvdimm_bus_init_interleave_sets(struct nvdimm_bus *nvdimm_bus); |
| void __nd_device_register(struct device *dev); |
| int nd_match_dimm(struct device *dev, void *data); |
| struct nd_label_id; |
| char *nd_label_gen_id(struct nd_label_id *label_id, u8 *uuid, u32 flags); |
| bool nd_is_uuid_unique(struct device *dev, u8 *uuid); |
| struct nd_region; |
| struct nvdimm_drvdata; |
| struct nd_mapping; |
| void nd_mapping_free_labels(struct nd_mapping *nd_mapping); |
| |
| int __reserve_free_pmem(struct device *dev, void *data); |
| void release_free_pmem(struct nvdimm_bus *nvdimm_bus, |
| struct nd_mapping *nd_mapping); |
| |
| resource_size_t nd_pmem_max_contiguous_dpa(struct nd_region *nd_region, |
| struct nd_mapping *nd_mapping); |
| resource_size_t nd_region_allocatable_dpa(struct nd_region *nd_region); |
| resource_size_t nd_pmem_available_dpa(struct nd_region *nd_region, |
| struct nd_mapping *nd_mapping, resource_size_t *overlap); |
| resource_size_t nd_blk_available_dpa(struct nd_region *nd_region); |
| resource_size_t nd_region_available_dpa(struct nd_region *nd_region); |
| int nd_region_conflict(struct nd_region *nd_region, resource_size_t start, |
| resource_size_t size); |
| resource_size_t nvdimm_allocated_dpa(struct nvdimm_drvdata *ndd, |
| struct nd_label_id *label_id); |
| int alias_dpa_busy(struct device *dev, void *data); |
| struct resource *nsblk_add_resource(struct nd_region *nd_region, |
| struct nvdimm_drvdata *ndd, struct nd_namespace_blk *nsblk, |
| resource_size_t start); |
| int nvdimm_num_label_slots(struct nvdimm_drvdata *ndd); |
| void get_ndd(struct nvdimm_drvdata *ndd); |
| resource_size_t __nvdimm_namespace_capacity(struct nd_namespace_common *ndns); |
| void nd_detach_ndns(struct device *dev, struct nd_namespace_common **_ndns); |
| void __nd_detach_ndns(struct device *dev, struct nd_namespace_common **_ndns); |
| bool nd_attach_ndns(struct device *dev, struct nd_namespace_common *attach, |
| struct nd_namespace_common **_ndns); |
| bool __nd_attach_ndns(struct device *dev, struct nd_namespace_common *attach, |
| struct nd_namespace_common **_ndns); |
| ssize_t nd_namespace_store(struct device *dev, |
| struct nd_namespace_common **_ndns, const char *buf, |
| size_t len); |
| struct nd_pfn *to_nd_pfn_safe(struct device *dev); |
| #endif /* __ND_CORE_H__ */ |