| // SPDX-License-Identifier: GPL-2.0-or-later |
| /* |
| * CXL Flash Device Driver |
| * |
| * Written by: Matthew R. Ochs <mrochs@linux.vnet.ibm.com>, IBM Corporation |
| * Uma Krishnan <ukrishn@linux.vnet.ibm.com>, IBM Corporation |
| * |
| * Copyright (C) 2018 IBM Corporation |
| */ |
| |
| #include <linux/file.h> |
| #include <linux/idr.h> |
| #include <linux/module.h> |
| #include <linux/mount.h> |
| #include <linux/pseudo_fs.h> |
| #include <linux/poll.h> |
| #include <linux/sched/signal.h> |
| #include <linux/interrupt.h> |
| #include <linux/irqdomain.h> |
| #include <asm/xive.h> |
| #include <misc/ocxl.h> |
| |
| #include <uapi/misc/cxl.h> |
| |
| #include "backend.h" |
| #include "ocxl_hw.h" |
| |
| /* |
| * Pseudo-filesystem to allocate inodes. |
| */ |
| |
| #define OCXLFLASH_FS_MAGIC 0x1697698f |
| |
| static int ocxlflash_fs_cnt; |
| static struct vfsmount *ocxlflash_vfs_mount; |
| |
| static int ocxlflash_fs_init_fs_context(struct fs_context *fc) |
| { |
| return init_pseudo(fc, OCXLFLASH_FS_MAGIC) ? 0 : -ENOMEM; |
| } |
| |
| static struct file_system_type ocxlflash_fs_type = { |
| .name = "ocxlflash", |
| .owner = THIS_MODULE, |
| .init_fs_context = ocxlflash_fs_init_fs_context, |
| .kill_sb = kill_anon_super, |
| }; |
| |
| /* |
| * ocxlflash_release_mapping() - release the memory mapping |
| * @ctx: Context whose mapping is to be released. |
| */ |
| static void ocxlflash_release_mapping(struct ocxlflash_context *ctx) |
| { |
| if (ctx->mapping) |
| simple_release_fs(&ocxlflash_vfs_mount, &ocxlflash_fs_cnt); |
| ctx->mapping = NULL; |
| } |
| |
| /* |
| * ocxlflash_getfile() - allocate pseudo filesystem, inode, and the file |
| * @dev: Generic device of the host. |
| * @name: Name of the pseudo filesystem. |
| * @fops: File operations. |
| * @priv: Private data. |
| * @flags: Flags for the file. |
| * |
| * Return: pointer to the file on success, ERR_PTR on failure |
| */ |
| static struct file *ocxlflash_getfile(struct device *dev, const char *name, |
| const struct file_operations *fops, |
| void *priv, int flags) |
| { |
| struct file *file; |
| struct inode *inode; |
| int rc; |
| |
| if (fops->owner && !try_module_get(fops->owner)) { |
| dev_err(dev, "%s: Owner does not exist\n", __func__); |
| rc = -ENOENT; |
| goto err1; |
| } |
| |
| rc = simple_pin_fs(&ocxlflash_fs_type, &ocxlflash_vfs_mount, |
| &ocxlflash_fs_cnt); |
| if (unlikely(rc < 0)) { |
| dev_err(dev, "%s: Cannot mount ocxlflash pseudofs rc=%d\n", |
| __func__, rc); |
| goto err2; |
| } |
| |
| inode = alloc_anon_inode(ocxlflash_vfs_mount->mnt_sb); |
| if (IS_ERR(inode)) { |
| rc = PTR_ERR(inode); |
| dev_err(dev, "%s: alloc_anon_inode failed rc=%d\n", |
| __func__, rc); |
| goto err3; |
| } |
| |
| file = alloc_file_pseudo(inode, ocxlflash_vfs_mount, name, |
| flags & (O_ACCMODE | O_NONBLOCK), fops); |
| if (IS_ERR(file)) { |
| rc = PTR_ERR(file); |
| dev_err(dev, "%s: alloc_file failed rc=%d\n", |
| __func__, rc); |
| goto err4; |
| } |
| |
| file->private_data = priv; |
| out: |
| return file; |
| err4: |
| iput(inode); |
| err3: |
| simple_release_fs(&ocxlflash_vfs_mount, &ocxlflash_fs_cnt); |
| err2: |
| module_put(fops->owner); |
| err1: |
| file = ERR_PTR(rc); |
| goto out; |
| } |
| |
| /** |
| * ocxlflash_psa_map() - map the process specific MMIO space |
| * @ctx_cookie: Adapter context for which the mapping needs to be done. |
| * |
| * Return: MMIO pointer of the mapped region |
| */ |
| static void __iomem *ocxlflash_psa_map(void *ctx_cookie) |
| { |
| struct ocxlflash_context *ctx = ctx_cookie; |
| struct device *dev = ctx->hw_afu->dev; |
| |
| mutex_lock(&ctx->state_mutex); |
| if (ctx->state != STARTED) { |
| dev_err(dev, "%s: Context not started, state=%d\n", __func__, |
| ctx->state); |
| mutex_unlock(&ctx->state_mutex); |
| return NULL; |
| } |
| mutex_unlock(&ctx->state_mutex); |
| |
| return ioremap(ctx->psn_phys, ctx->psn_size); |
| } |
| |
| /** |
| * ocxlflash_psa_unmap() - unmap the process specific MMIO space |
| * @addr: MMIO pointer to unmap. |
| */ |
| static void ocxlflash_psa_unmap(void __iomem *addr) |
| { |
| iounmap(addr); |
| } |
| |
| /** |
| * ocxlflash_process_element() - get process element of the adapter context |
| * @ctx_cookie: Adapter context associated with the process element. |
| * |
| * Return: process element of the adapter context |
| */ |
| static int ocxlflash_process_element(void *ctx_cookie) |
| { |
| struct ocxlflash_context *ctx = ctx_cookie; |
| |
| return ctx->pe; |
| } |
| |
| /** |
| * afu_map_irq() - map the interrupt of the adapter context |
| * @flags: Flags. |
| * @ctx: Adapter context. |
| * @num: Per-context AFU interrupt number. |
| * @handler: Interrupt handler to register. |
| * @cookie: Interrupt handler private data. |
| * @name: Name of the interrupt. |
| * |
| * Return: 0 on success, -errno on failure |
| */ |
| static int afu_map_irq(u64 flags, struct ocxlflash_context *ctx, int num, |
| irq_handler_t handler, void *cookie, char *name) |
| { |
| struct ocxl_hw_afu *afu = ctx->hw_afu; |
| struct device *dev = afu->dev; |
| struct ocxlflash_irqs *irq; |
| struct xive_irq_data *xd; |
| u32 virq; |
| int rc = 0; |
| |
| if (num < 0 || num >= ctx->num_irqs) { |
| dev_err(dev, "%s: Interrupt %d not allocated\n", __func__, num); |
| rc = -ENOENT; |
| goto out; |
| } |
| |
| irq = &ctx->irqs[num]; |
| virq = irq_create_mapping(NULL, irq->hwirq); |
| if (unlikely(!virq)) { |
| dev_err(dev, "%s: irq_create_mapping failed\n", __func__); |
| rc = -ENOMEM; |
| goto out; |
| } |
| |
| rc = request_irq(virq, handler, 0, name, cookie); |
| if (unlikely(rc)) { |
| dev_err(dev, "%s: request_irq failed rc=%d\n", __func__, rc); |
| goto err1; |
| } |
| |
| xd = irq_get_handler_data(virq); |
| if (unlikely(!xd)) { |
| dev_err(dev, "%s: Can't get interrupt data\n", __func__); |
| rc = -ENXIO; |
| goto err2; |
| } |
| |
| irq->virq = virq; |
| irq->vtrig = xd->trig_mmio; |
| out: |
| return rc; |
| err2: |
| free_irq(virq, cookie); |
| err1: |
| irq_dispose_mapping(virq); |
| goto out; |
| } |
| |
| /** |
| * ocxlflash_map_afu_irq() - map the interrupt of the adapter context |
| * @ctx_cookie: Adapter context. |
| * @num: Per-context AFU interrupt number. |
| * @handler: Interrupt handler to register. |
| * @cookie: Interrupt handler private data. |
| * @name: Name of the interrupt. |
| * |
| * Return: 0 on success, -errno on failure |
| */ |
| static int ocxlflash_map_afu_irq(void *ctx_cookie, int num, |
| irq_handler_t handler, void *cookie, |
| char *name) |
| { |
| return afu_map_irq(0, ctx_cookie, num, handler, cookie, name); |
| } |
| |
| /** |
| * afu_unmap_irq() - unmap the interrupt |
| * @flags: Flags. |
| * @ctx: Adapter context. |
| * @num: Per-context AFU interrupt number. |
| * @cookie: Interrupt handler private data. |
| */ |
| static void afu_unmap_irq(u64 flags, struct ocxlflash_context *ctx, int num, |
| void *cookie) |
| { |
| struct ocxl_hw_afu *afu = ctx->hw_afu; |
| struct device *dev = afu->dev; |
| struct ocxlflash_irqs *irq; |
| |
| if (num < 0 || num >= ctx->num_irqs) { |
| dev_err(dev, "%s: Interrupt %d not allocated\n", __func__, num); |
| return; |
| } |
| |
| irq = &ctx->irqs[num]; |
| |
| if (irq_find_mapping(NULL, irq->hwirq)) { |
| free_irq(irq->virq, cookie); |
| irq_dispose_mapping(irq->virq); |
| } |
| |
| memset(irq, 0, sizeof(*irq)); |
| } |
| |
| /** |
| * ocxlflash_unmap_afu_irq() - unmap the interrupt |
| * @ctx_cookie: Adapter context. |
| * @num: Per-context AFU interrupt number. |
| * @cookie: Interrupt handler private data. |
| */ |
| static void ocxlflash_unmap_afu_irq(void *ctx_cookie, int num, void *cookie) |
| { |
| return afu_unmap_irq(0, ctx_cookie, num, cookie); |
| } |
| |
| /** |
| * ocxlflash_get_irq_objhndl() - get the object handle for an interrupt |
| * @ctx_cookie: Context associated with the interrupt. |
| * @irq: Interrupt number. |
| * |
| * Return: effective address of the mapped region |
| */ |
| static u64 ocxlflash_get_irq_objhndl(void *ctx_cookie, int irq) |
| { |
| struct ocxlflash_context *ctx = ctx_cookie; |
| |
| if (irq < 0 || irq >= ctx->num_irqs) |
| return 0; |
| |
| return (__force u64)ctx->irqs[irq].vtrig; |
| } |
| |
| /** |
| * ocxlflash_xsl_fault() - callback when translation error is triggered |
| * @data: Private data provided at callback registration, the context. |
| * @addr: Address that triggered the error. |
| * @dsisr: Value of dsisr register. |
| */ |
| static void ocxlflash_xsl_fault(void *data, u64 addr, u64 dsisr) |
| { |
| struct ocxlflash_context *ctx = data; |
| |
| spin_lock(&ctx->slock); |
| ctx->fault_addr = addr; |
| ctx->fault_dsisr = dsisr; |
| ctx->pending_fault = true; |
| spin_unlock(&ctx->slock); |
| |
| wake_up_all(&ctx->wq); |
| } |
| |
| /** |
| * start_context() - local routine to start a context |
| * @ctx: Adapter context to be started. |
| * |
| * Assign the context specific MMIO space, add and enable the PE. |
| * |
| * Return: 0 on success, -errno on failure |
| */ |
| static int start_context(struct ocxlflash_context *ctx) |
| { |
| struct ocxl_hw_afu *afu = ctx->hw_afu; |
| struct ocxl_afu_config *acfg = &afu->acfg; |
| void *link_token = afu->link_token; |
| struct pci_dev *pdev = afu->pdev; |
| struct device *dev = afu->dev; |
| bool master = ctx->master; |
| struct mm_struct *mm; |
| int rc = 0; |
| u32 pid; |
| |
| mutex_lock(&ctx->state_mutex); |
| if (ctx->state != OPENED) { |
| dev_err(dev, "%s: Context state invalid, state=%d\n", |
| __func__, ctx->state); |
| rc = -EINVAL; |
| goto out; |
| } |
| |
| if (master) { |
| ctx->psn_size = acfg->global_mmio_size; |
| ctx->psn_phys = afu->gmmio_phys; |
| } else { |
| ctx->psn_size = acfg->pp_mmio_stride; |
| ctx->psn_phys = afu->ppmmio_phys + (ctx->pe * ctx->psn_size); |
| } |
| |
| /* pid and mm not set for master contexts */ |
| if (master) { |
| pid = 0; |
| mm = NULL; |
| } else { |
| pid = current->mm->context.id; |
| mm = current->mm; |
| } |
| |
| rc = ocxl_link_add_pe(link_token, ctx->pe, pid, 0, 0, |
| pci_dev_id(pdev), mm, ocxlflash_xsl_fault, |
| ctx); |
| if (unlikely(rc)) { |
| dev_err(dev, "%s: ocxl_link_add_pe failed rc=%d\n", |
| __func__, rc); |
| goto out; |
| } |
| |
| ctx->state = STARTED; |
| out: |
| mutex_unlock(&ctx->state_mutex); |
| return rc; |
| } |
| |
| /** |
| * ocxlflash_start_context() - start a kernel context |
| * @ctx_cookie: Adapter context to be started. |
| * |
| * Return: 0 on success, -errno on failure |
| */ |
| static int ocxlflash_start_context(void *ctx_cookie) |
| { |
| struct ocxlflash_context *ctx = ctx_cookie; |
| |
| return start_context(ctx); |
| } |
| |
| /** |
| * ocxlflash_stop_context() - stop a context |
| * @ctx_cookie: Adapter context to be stopped. |
| * |
| * Return: 0 on success, -errno on failure |
| */ |
| static int ocxlflash_stop_context(void *ctx_cookie) |
| { |
| struct ocxlflash_context *ctx = ctx_cookie; |
| struct ocxl_hw_afu *afu = ctx->hw_afu; |
| struct ocxl_afu_config *acfg = &afu->acfg; |
| struct pci_dev *pdev = afu->pdev; |
| struct device *dev = afu->dev; |
| enum ocxlflash_ctx_state state; |
| int rc = 0; |
| |
| mutex_lock(&ctx->state_mutex); |
| state = ctx->state; |
| ctx->state = CLOSED; |
| mutex_unlock(&ctx->state_mutex); |
| if (state != STARTED) |
| goto out; |
| |
| rc = ocxl_config_terminate_pasid(pdev, acfg->dvsec_afu_control_pos, |
| ctx->pe); |
| if (unlikely(rc)) { |
| dev_err(dev, "%s: ocxl_config_terminate_pasid failed rc=%d\n", |
| __func__, rc); |
| /* If EBUSY, PE could be referenced in future by the AFU */ |
| if (rc == -EBUSY) |
| goto out; |
| } |
| |
| rc = ocxl_link_remove_pe(afu->link_token, ctx->pe); |
| if (unlikely(rc)) { |
| dev_err(dev, "%s: ocxl_link_remove_pe failed rc=%d\n", |
| __func__, rc); |
| goto out; |
| } |
| out: |
| return rc; |
| } |
| |
| /** |
| * ocxlflash_afu_reset() - reset the AFU |
| * @ctx_cookie: Adapter context. |
| */ |
| static int ocxlflash_afu_reset(void *ctx_cookie) |
| { |
| struct ocxlflash_context *ctx = ctx_cookie; |
| struct device *dev = ctx->hw_afu->dev; |
| |
| /* Pending implementation from OCXL transport services */ |
| dev_err_once(dev, "%s: afu_reset() fop not supported\n", __func__); |
| |
| /* Silently return success until it is implemented */ |
| return 0; |
| } |
| |
| /** |
| * ocxlflash_set_master() - sets the context as master |
| * @ctx_cookie: Adapter context to set as master. |
| */ |
| static void ocxlflash_set_master(void *ctx_cookie) |
| { |
| struct ocxlflash_context *ctx = ctx_cookie; |
| |
| ctx->master = true; |
| } |
| |
| /** |
| * ocxlflash_get_context() - obtains the context associated with the host |
| * @pdev: PCI device associated with the host. |
| * @afu_cookie: Hardware AFU associated with the host. |
| * |
| * Return: returns the pointer to host adapter context |
| */ |
| static void *ocxlflash_get_context(struct pci_dev *pdev, void *afu_cookie) |
| { |
| struct ocxl_hw_afu *afu = afu_cookie; |
| |
| return afu->ocxl_ctx; |
| } |
| |
| /** |
| * ocxlflash_dev_context_init() - allocate and initialize an adapter context |
| * @pdev: PCI device associated with the host. |
| * @afu_cookie: Hardware AFU associated with the host. |
| * |
| * Return: returns the adapter context on success, ERR_PTR on failure |
| */ |
| static void *ocxlflash_dev_context_init(struct pci_dev *pdev, void *afu_cookie) |
| { |
| struct ocxl_hw_afu *afu = afu_cookie; |
| struct device *dev = afu->dev; |
| struct ocxlflash_context *ctx; |
| int rc; |
| |
| ctx = kzalloc(sizeof(*ctx), GFP_KERNEL); |
| if (unlikely(!ctx)) { |
| dev_err(dev, "%s: Context allocation failed\n", __func__); |
| rc = -ENOMEM; |
| goto err1; |
| } |
| |
| idr_preload(GFP_KERNEL); |
| rc = idr_alloc(&afu->idr, ctx, 0, afu->max_pasid, GFP_NOWAIT); |
| idr_preload_end(); |
| if (unlikely(rc < 0)) { |
| dev_err(dev, "%s: idr_alloc failed rc=%d\n", __func__, rc); |
| goto err2; |
| } |
| |
| spin_lock_init(&ctx->slock); |
| init_waitqueue_head(&ctx->wq); |
| mutex_init(&ctx->state_mutex); |
| |
| ctx->state = OPENED; |
| ctx->pe = rc; |
| ctx->master = false; |
| ctx->mapping = NULL; |
| ctx->hw_afu = afu; |
| ctx->irq_bitmap = 0; |
| ctx->pending_irq = false; |
| ctx->pending_fault = false; |
| out: |
| return ctx; |
| err2: |
| kfree(ctx); |
| err1: |
| ctx = ERR_PTR(rc); |
| goto out; |
| } |
| |
| /** |
| * ocxlflash_release_context() - releases an adapter context |
| * @ctx_cookie: Adapter context to be released. |
| * |
| * Return: 0 on success, -errno on failure |
| */ |
| static int ocxlflash_release_context(void *ctx_cookie) |
| { |
| struct ocxlflash_context *ctx = ctx_cookie; |
| struct device *dev; |
| int rc = 0; |
| |
| if (!ctx) |
| goto out; |
| |
| dev = ctx->hw_afu->dev; |
| mutex_lock(&ctx->state_mutex); |
| if (ctx->state >= STARTED) { |
| dev_err(dev, "%s: Context in use, state=%d\n", __func__, |
| ctx->state); |
| mutex_unlock(&ctx->state_mutex); |
| rc = -EBUSY; |
| goto out; |
| } |
| mutex_unlock(&ctx->state_mutex); |
| |
| idr_remove(&ctx->hw_afu->idr, ctx->pe); |
| ocxlflash_release_mapping(ctx); |
| kfree(ctx); |
| out: |
| return rc; |
| } |
| |
| /** |
| * ocxlflash_perst_reloads_same_image() - sets the image reload policy |
| * @afu_cookie: Hardware AFU associated with the host. |
| * @image: Whether to load the same image on PERST. |
| */ |
| static void ocxlflash_perst_reloads_same_image(void *afu_cookie, bool image) |
| { |
| struct ocxl_hw_afu *afu = afu_cookie; |
| |
| afu->perst_same_image = image; |
| } |
| |
| /** |
| * ocxlflash_read_adapter_vpd() - reads the adapter VPD |
| * @pdev: PCI device associated with the host. |
| * @buf: Buffer to get the VPD data. |
| * @count: Size of buffer (maximum bytes that can be read). |
| * |
| * Return: size of VPD on success, -errno on failure |
| */ |
| static ssize_t ocxlflash_read_adapter_vpd(struct pci_dev *pdev, void *buf, |
| size_t count) |
| { |
| return pci_read_vpd(pdev, 0, count, buf); |
| } |
| |
| /** |
| * free_afu_irqs() - internal service to free interrupts |
| * @ctx: Adapter context. |
| */ |
| static void free_afu_irqs(struct ocxlflash_context *ctx) |
| { |
| struct ocxl_hw_afu *afu = ctx->hw_afu; |
| struct device *dev = afu->dev; |
| int i; |
| |
| if (!ctx->irqs) { |
| dev_err(dev, "%s: Interrupts not allocated\n", __func__); |
| return; |
| } |
| |
| for (i = ctx->num_irqs; i >= 0; i--) |
| ocxl_link_free_irq(afu->link_token, ctx->irqs[i].hwirq); |
| |
| kfree(ctx->irqs); |
| ctx->irqs = NULL; |
| } |
| |
| /** |
| * alloc_afu_irqs() - internal service to allocate interrupts |
| * @ctx: Context associated with the request. |
| * @num: Number of interrupts requested. |
| * |
| * Return: 0 on success, -errno on failure |
| */ |
| static int alloc_afu_irqs(struct ocxlflash_context *ctx, int num) |
| { |
| struct ocxl_hw_afu *afu = ctx->hw_afu; |
| struct device *dev = afu->dev; |
| struct ocxlflash_irqs *irqs; |
| int rc = 0; |
| int hwirq; |
| int i; |
| |
| if (ctx->irqs) { |
| dev_err(dev, "%s: Interrupts already allocated\n", __func__); |
| rc = -EEXIST; |
| goto out; |
| } |
| |
| if (num > OCXL_MAX_IRQS) { |
| dev_err(dev, "%s: Too many interrupts num=%d\n", __func__, num); |
| rc = -EINVAL; |
| goto out; |
| } |
| |
| irqs = kcalloc(num, sizeof(*irqs), GFP_KERNEL); |
| if (unlikely(!irqs)) { |
| dev_err(dev, "%s: Context irqs allocation failed\n", __func__); |
| rc = -ENOMEM; |
| goto out; |
| } |
| |
| for (i = 0; i < num; i++) { |
| rc = ocxl_link_irq_alloc(afu->link_token, &hwirq); |
| if (unlikely(rc)) { |
| dev_err(dev, "%s: ocxl_link_irq_alloc failed rc=%d\n", |
| __func__, rc); |
| goto err; |
| } |
| |
| irqs[i].hwirq = hwirq; |
| } |
| |
| ctx->irqs = irqs; |
| ctx->num_irqs = num; |
| out: |
| return rc; |
| err: |
| for (i = i-1; i >= 0; i--) |
| ocxl_link_free_irq(afu->link_token, irqs[i].hwirq); |
| kfree(irqs); |
| goto out; |
| } |
| |
| /** |
| * ocxlflash_allocate_afu_irqs() - allocates the requested number of interrupts |
| * @ctx_cookie: Context associated with the request. |
| * @num: Number of interrupts requested. |
| * |
| * Return: 0 on success, -errno on failure |
| */ |
| static int ocxlflash_allocate_afu_irqs(void *ctx_cookie, int num) |
| { |
| return alloc_afu_irqs(ctx_cookie, num); |
| } |
| |
| /** |
| * ocxlflash_free_afu_irqs() - frees the interrupts of an adapter context |
| * @ctx_cookie: Adapter context. |
| */ |
| static void ocxlflash_free_afu_irqs(void *ctx_cookie) |
| { |
| free_afu_irqs(ctx_cookie); |
| } |
| |
| /** |
| * ocxlflash_unconfig_afu() - unconfigure the AFU |
| * @afu: AFU associated with the host. |
| */ |
| static void ocxlflash_unconfig_afu(struct ocxl_hw_afu *afu) |
| { |
| if (afu->gmmio_virt) { |
| iounmap(afu->gmmio_virt); |
| afu->gmmio_virt = NULL; |
| } |
| } |
| |
| /** |
| * ocxlflash_destroy_afu() - destroy the AFU structure |
| * @afu_cookie: AFU to be freed. |
| */ |
| static void ocxlflash_destroy_afu(void *afu_cookie) |
| { |
| struct ocxl_hw_afu *afu = afu_cookie; |
| int pos; |
| |
| if (!afu) |
| return; |
| |
| ocxlflash_release_context(afu->ocxl_ctx); |
| idr_destroy(&afu->idr); |
| |
| /* Disable the AFU */ |
| pos = afu->acfg.dvsec_afu_control_pos; |
| ocxl_config_set_afu_state(afu->pdev, pos, 0); |
| |
| ocxlflash_unconfig_afu(afu); |
| kfree(afu); |
| } |
| |
| /** |
| * ocxlflash_config_fn() - configure the host function |
| * @pdev: PCI device associated with the host. |
| * @afu: AFU associated with the host. |
| * |
| * Return: 0 on success, -errno on failure |
| */ |
| static int ocxlflash_config_fn(struct pci_dev *pdev, struct ocxl_hw_afu *afu) |
| { |
| struct ocxl_fn_config *fcfg = &afu->fcfg; |
| struct device *dev = &pdev->dev; |
| u16 base, enabled, supported; |
| int rc = 0; |
| |
| /* Read DVSEC config of the function */ |
| rc = ocxl_config_read_function(pdev, fcfg); |
| if (unlikely(rc)) { |
| dev_err(dev, "%s: ocxl_config_read_function failed rc=%d\n", |
| __func__, rc); |
| goto out; |
| } |
| |
| /* Check if function has AFUs defined, only 1 per function supported */ |
| if (fcfg->max_afu_index >= 0) { |
| afu->is_present = true; |
| if (fcfg->max_afu_index != 0) |
| dev_warn(dev, "%s: Unexpected AFU index value %d\n", |
| __func__, fcfg->max_afu_index); |
| } |
| |
| rc = ocxl_config_get_actag_info(pdev, &base, &enabled, &supported); |
| if (unlikely(rc)) { |
| dev_err(dev, "%s: ocxl_config_get_actag_info failed rc=%d\n", |
| __func__, rc); |
| goto out; |
| } |
| |
| afu->fn_actag_base = base; |
| afu->fn_actag_enabled = enabled; |
| |
| ocxl_config_set_actag(pdev, fcfg->dvsec_function_pos, base, enabled); |
| dev_dbg(dev, "%s: Function acTag range base=%u enabled=%u\n", |
| __func__, base, enabled); |
| |
| rc = ocxl_link_setup(pdev, 0, &afu->link_token); |
| if (unlikely(rc)) { |
| dev_err(dev, "%s: ocxl_link_setup failed rc=%d\n", |
| __func__, rc); |
| goto out; |
| } |
| |
| rc = ocxl_config_set_TL(pdev, fcfg->dvsec_tl_pos); |
| if (unlikely(rc)) { |
| dev_err(dev, "%s: ocxl_config_set_TL failed rc=%d\n", |
| __func__, rc); |
| goto err; |
| } |
| out: |
| return rc; |
| err: |
| ocxl_link_release(pdev, afu->link_token); |
| goto out; |
| } |
| |
| /** |
| * ocxlflash_unconfig_fn() - unconfigure the host function |
| * @pdev: PCI device associated with the host. |
| * @afu: AFU associated with the host. |
| */ |
| static void ocxlflash_unconfig_fn(struct pci_dev *pdev, struct ocxl_hw_afu *afu) |
| { |
| ocxl_link_release(pdev, afu->link_token); |
| } |
| |
| /** |
| * ocxlflash_map_mmio() - map the AFU MMIO space |
| * @afu: AFU associated with the host. |
| * |
| * Return: 0 on success, -errno on failure |
| */ |
| static int ocxlflash_map_mmio(struct ocxl_hw_afu *afu) |
| { |
| struct ocxl_afu_config *acfg = &afu->acfg; |
| struct pci_dev *pdev = afu->pdev; |
| struct device *dev = afu->dev; |
| phys_addr_t gmmio, ppmmio; |
| int rc = 0; |
| |
| rc = pci_request_region(pdev, acfg->global_mmio_bar, "ocxlflash"); |
| if (unlikely(rc)) { |
| dev_err(dev, "%s: pci_request_region for global failed rc=%d\n", |
| __func__, rc); |
| goto out; |
| } |
| gmmio = pci_resource_start(pdev, acfg->global_mmio_bar); |
| gmmio += acfg->global_mmio_offset; |
| |
| rc = pci_request_region(pdev, acfg->pp_mmio_bar, "ocxlflash"); |
| if (unlikely(rc)) { |
| dev_err(dev, "%s: pci_request_region for pp bar failed rc=%d\n", |
| __func__, rc); |
| goto err1; |
| } |
| ppmmio = pci_resource_start(pdev, acfg->pp_mmio_bar); |
| ppmmio += acfg->pp_mmio_offset; |
| |
| afu->gmmio_virt = ioremap(gmmio, acfg->global_mmio_size); |
| if (unlikely(!afu->gmmio_virt)) { |
| dev_err(dev, "%s: MMIO mapping failed\n", __func__); |
| rc = -ENOMEM; |
| goto err2; |
| } |
| |
| afu->gmmio_phys = gmmio; |
| afu->ppmmio_phys = ppmmio; |
| out: |
| return rc; |
| err2: |
| pci_release_region(pdev, acfg->pp_mmio_bar); |
| err1: |
| pci_release_region(pdev, acfg->global_mmio_bar); |
| goto out; |
| } |
| |
| /** |
| * ocxlflash_config_afu() - configure the host AFU |
| * @pdev: PCI device associated with the host. |
| * @afu: AFU associated with the host. |
| * |
| * Must be called _after_ host function configuration. |
| * |
| * Return: 0 on success, -errno on failure |
| */ |
| static int ocxlflash_config_afu(struct pci_dev *pdev, struct ocxl_hw_afu *afu) |
| { |
| struct ocxl_afu_config *acfg = &afu->acfg; |
| struct ocxl_fn_config *fcfg = &afu->fcfg; |
| struct device *dev = &pdev->dev; |
| int count; |
| int base; |
| int pos; |
| int rc = 0; |
| |
| /* This HW AFU function does not have any AFUs defined */ |
| if (!afu->is_present) |
| goto out; |
| |
| /* Read AFU config at index 0 */ |
| rc = ocxl_config_read_afu(pdev, fcfg, acfg, 0); |
| if (unlikely(rc)) { |
| dev_err(dev, "%s: ocxl_config_read_afu failed rc=%d\n", |
| __func__, rc); |
| goto out; |
| } |
| |
| /* Only one AFU per function is supported, so actag_base is same */ |
| base = afu->fn_actag_base; |
| count = min_t(int, acfg->actag_supported, afu->fn_actag_enabled); |
| pos = acfg->dvsec_afu_control_pos; |
| |
| ocxl_config_set_afu_actag(pdev, pos, base, count); |
| dev_dbg(dev, "%s: acTag base=%d enabled=%d\n", __func__, base, count); |
| afu->afu_actag_base = base; |
| afu->afu_actag_enabled = count; |
| afu->max_pasid = 1 << acfg->pasid_supported_log; |
| |
| ocxl_config_set_afu_pasid(pdev, pos, 0, acfg->pasid_supported_log); |
| |
| rc = ocxlflash_map_mmio(afu); |
| if (unlikely(rc)) { |
| dev_err(dev, "%s: ocxlflash_map_mmio failed rc=%d\n", |
| __func__, rc); |
| goto out; |
| } |
| |
| /* Enable the AFU */ |
| ocxl_config_set_afu_state(pdev, acfg->dvsec_afu_control_pos, 1); |
| out: |
| return rc; |
| } |
| |
| /** |
| * ocxlflash_create_afu() - create the AFU for OCXL |
| * @pdev: PCI device associated with the host. |
| * |
| * Return: AFU on success, NULL on failure |
| */ |
| static void *ocxlflash_create_afu(struct pci_dev *pdev) |
| { |
| struct device *dev = &pdev->dev; |
| struct ocxlflash_context *ctx; |
| struct ocxl_hw_afu *afu; |
| int rc; |
| |
| afu = kzalloc(sizeof(*afu), GFP_KERNEL); |
| if (unlikely(!afu)) { |
| dev_err(dev, "%s: HW AFU allocation failed\n", __func__); |
| goto out; |
| } |
| |
| afu->pdev = pdev; |
| afu->dev = dev; |
| idr_init(&afu->idr); |
| |
| rc = ocxlflash_config_fn(pdev, afu); |
| if (unlikely(rc)) { |
| dev_err(dev, "%s: Function configuration failed rc=%d\n", |
| __func__, rc); |
| goto err1; |
| } |
| |
| rc = ocxlflash_config_afu(pdev, afu); |
| if (unlikely(rc)) { |
| dev_err(dev, "%s: AFU configuration failed rc=%d\n", |
| __func__, rc); |
| goto err2; |
| } |
| |
| ctx = ocxlflash_dev_context_init(pdev, afu); |
| if (IS_ERR(ctx)) { |
| rc = PTR_ERR(ctx); |
| dev_err(dev, "%s: ocxlflash_dev_context_init failed rc=%d\n", |
| __func__, rc); |
| goto err3; |
| } |
| |
| afu->ocxl_ctx = ctx; |
| out: |
| return afu; |
| err3: |
| ocxlflash_unconfig_afu(afu); |
| err2: |
| ocxlflash_unconfig_fn(pdev, afu); |
| err1: |
| idr_destroy(&afu->idr); |
| kfree(afu); |
| afu = NULL; |
| goto out; |
| } |
| |
| /** |
| * ctx_event_pending() - check for any event pending on the context |
| * @ctx: Context to be checked. |
| * |
| * Return: true if there is an event pending, false if none pending |
| */ |
| static inline bool ctx_event_pending(struct ocxlflash_context *ctx) |
| { |
| if (ctx->pending_irq || ctx->pending_fault) |
| return true; |
| |
| return false; |
| } |
| |
| /** |
| * afu_poll() - poll the AFU for events on the context |
| * @file: File associated with the adapter context. |
| * @poll: Poll structure from the user. |
| * |
| * Return: poll mask |
| */ |
| static unsigned int afu_poll(struct file *file, struct poll_table_struct *poll) |
| { |
| struct ocxlflash_context *ctx = file->private_data; |
| struct device *dev = ctx->hw_afu->dev; |
| ulong lock_flags; |
| int mask = 0; |
| |
| poll_wait(file, &ctx->wq, poll); |
| |
| spin_lock_irqsave(&ctx->slock, lock_flags); |
| if (ctx_event_pending(ctx)) |
| mask |= POLLIN | POLLRDNORM; |
| else if (ctx->state == CLOSED) |
| mask |= POLLERR; |
| spin_unlock_irqrestore(&ctx->slock, lock_flags); |
| |
| dev_dbg(dev, "%s: Poll wait completed for pe %i mask %i\n", |
| __func__, ctx->pe, mask); |
| |
| return mask; |
| } |
| |
| /** |
| * afu_read() - perform a read on the context for any event |
| * @file: File associated with the adapter context. |
| * @buf: Buffer to receive the data. |
| * @count: Size of buffer (maximum bytes that can be read). |
| * @off: Offset. |
| * |
| * Return: size of the data read on success, -errno on failure |
| */ |
| static ssize_t afu_read(struct file *file, char __user *buf, size_t count, |
| loff_t *off) |
| { |
| struct ocxlflash_context *ctx = file->private_data; |
| struct device *dev = ctx->hw_afu->dev; |
| struct cxl_event event; |
| ulong lock_flags; |
| ssize_t esize; |
| ssize_t rc; |
| int bit; |
| DEFINE_WAIT(event_wait); |
| |
| if (*off != 0) { |
| dev_err(dev, "%s: Non-zero offset not supported, off=%lld\n", |
| __func__, *off); |
| rc = -EINVAL; |
| goto out; |
| } |
| |
| spin_lock_irqsave(&ctx->slock, lock_flags); |
| |
| for (;;) { |
| prepare_to_wait(&ctx->wq, &event_wait, TASK_INTERRUPTIBLE); |
| |
| if (ctx_event_pending(ctx) || (ctx->state == CLOSED)) |
| break; |
| |
| if (file->f_flags & O_NONBLOCK) { |
| dev_err(dev, "%s: File cannot be blocked on I/O\n", |
| __func__); |
| rc = -EAGAIN; |
| goto err; |
| } |
| |
| if (signal_pending(current)) { |
| dev_err(dev, "%s: Signal pending on the process\n", |
| __func__); |
| rc = -ERESTARTSYS; |
| goto err; |
| } |
| |
| spin_unlock_irqrestore(&ctx->slock, lock_flags); |
| schedule(); |
| spin_lock_irqsave(&ctx->slock, lock_flags); |
| } |
| |
| finish_wait(&ctx->wq, &event_wait); |
| |
| memset(&event, 0, sizeof(event)); |
| event.header.process_element = ctx->pe; |
| event.header.size = sizeof(struct cxl_event_header); |
| if (ctx->pending_irq) { |
| esize = sizeof(struct cxl_event_afu_interrupt); |
| event.header.size += esize; |
| event.header.type = CXL_EVENT_AFU_INTERRUPT; |
| |
| bit = find_first_bit(&ctx->irq_bitmap, ctx->num_irqs); |
| clear_bit(bit, &ctx->irq_bitmap); |
| event.irq.irq = bit + 1; |
| if (bitmap_empty(&ctx->irq_bitmap, ctx->num_irqs)) |
| ctx->pending_irq = false; |
| } else if (ctx->pending_fault) { |
| event.header.size += sizeof(struct cxl_event_data_storage); |
| event.header.type = CXL_EVENT_DATA_STORAGE; |
| event.fault.addr = ctx->fault_addr; |
| event.fault.dsisr = ctx->fault_dsisr; |
| ctx->pending_fault = false; |
| } |
| |
| spin_unlock_irqrestore(&ctx->slock, lock_flags); |
| |
| if (copy_to_user(buf, &event, event.header.size)) { |
| dev_err(dev, "%s: copy_to_user failed\n", __func__); |
| rc = -EFAULT; |
| goto out; |
| } |
| |
| rc = event.header.size; |
| out: |
| return rc; |
| err: |
| finish_wait(&ctx->wq, &event_wait); |
| spin_unlock_irqrestore(&ctx->slock, lock_flags); |
| goto out; |
| } |
| |
| /** |
| * afu_release() - release and free the context |
| * @inode: File inode pointer. |
| * @file: File associated with the context. |
| * |
| * Return: 0 on success, -errno on failure |
| */ |
| static int afu_release(struct inode *inode, struct file *file) |
| { |
| struct ocxlflash_context *ctx = file->private_data; |
| int i; |
| |
| /* Unmap and free the interrupts associated with the context */ |
| for (i = ctx->num_irqs; i >= 0; i--) |
| afu_unmap_irq(0, ctx, i, ctx); |
| free_afu_irqs(ctx); |
| |
| return ocxlflash_release_context(ctx); |
| } |
| |
| /** |
| * ocxlflash_mmap_fault() - mmap fault handler |
| * @vmf: VM fault associated with current fault. |
| * |
| * Return: 0 on success, -errno on failure |
| */ |
| static vm_fault_t ocxlflash_mmap_fault(struct vm_fault *vmf) |
| { |
| struct vm_area_struct *vma = vmf->vma; |
| struct ocxlflash_context *ctx = vma->vm_file->private_data; |
| struct device *dev = ctx->hw_afu->dev; |
| u64 mmio_area, offset; |
| |
| offset = vmf->pgoff << PAGE_SHIFT; |
| if (offset >= ctx->psn_size) |
| return VM_FAULT_SIGBUS; |
| |
| mutex_lock(&ctx->state_mutex); |
| if (ctx->state != STARTED) { |
| dev_err(dev, "%s: Context not started, state=%d\n", |
| __func__, ctx->state); |
| mutex_unlock(&ctx->state_mutex); |
| return VM_FAULT_SIGBUS; |
| } |
| mutex_unlock(&ctx->state_mutex); |
| |
| mmio_area = ctx->psn_phys; |
| mmio_area += offset; |
| |
| return vmf_insert_pfn(vma, vmf->address, mmio_area >> PAGE_SHIFT); |
| } |
| |
| static const struct vm_operations_struct ocxlflash_vmops = { |
| .fault = ocxlflash_mmap_fault, |
| }; |
| |
| /** |
| * afu_mmap() - map the fault handler operations |
| * @file: File associated with the context. |
| * @vma: VM area associated with mapping. |
| * |
| * Return: 0 on success, -errno on failure |
| */ |
| static int afu_mmap(struct file *file, struct vm_area_struct *vma) |
| { |
| struct ocxlflash_context *ctx = file->private_data; |
| |
| if ((vma_pages(vma) + vma->vm_pgoff) > |
| (ctx->psn_size >> PAGE_SHIFT)) |
| return -EINVAL; |
| |
| vma->vm_flags |= VM_IO | VM_PFNMAP; |
| vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); |
| vma->vm_ops = &ocxlflash_vmops; |
| return 0; |
| } |
| |
| static const struct file_operations ocxl_afu_fops = { |
| .owner = THIS_MODULE, |
| .poll = afu_poll, |
| .read = afu_read, |
| .release = afu_release, |
| .mmap = afu_mmap, |
| }; |
| |
| #define PATCH_FOPS(NAME) \ |
| do { if (!fops->NAME) fops->NAME = ocxl_afu_fops.NAME; } while (0) |
| |
| /** |
| * ocxlflash_get_fd() - get file descriptor for an adapter context |
| * @ctx_cookie: Adapter context. |
| * @fops: File operations to be associated. |
| * @fd: File descriptor to be returned back. |
| * |
| * Return: pointer to the file on success, ERR_PTR on failure |
| */ |
| static struct file *ocxlflash_get_fd(void *ctx_cookie, |
| struct file_operations *fops, int *fd) |
| { |
| struct ocxlflash_context *ctx = ctx_cookie; |
| struct device *dev = ctx->hw_afu->dev; |
| struct file *file; |
| int flags, fdtmp; |
| int rc = 0; |
| char *name = NULL; |
| |
| /* Only allow one fd per context */ |
| if (ctx->mapping) { |
| dev_err(dev, "%s: Context is already mapped to an fd\n", |
| __func__); |
| rc = -EEXIST; |
| goto err1; |
| } |
| |
| flags = O_RDWR | O_CLOEXEC; |
| |
| /* This code is similar to anon_inode_getfd() */ |
| rc = get_unused_fd_flags(flags); |
| if (unlikely(rc < 0)) { |
| dev_err(dev, "%s: get_unused_fd_flags failed rc=%d\n", |
| __func__, rc); |
| goto err1; |
| } |
| fdtmp = rc; |
| |
| /* Patch the file ops that are not defined */ |
| if (fops) { |
| PATCH_FOPS(poll); |
| PATCH_FOPS(read); |
| PATCH_FOPS(release); |
| PATCH_FOPS(mmap); |
| } else /* Use default ops */ |
| fops = (struct file_operations *)&ocxl_afu_fops; |
| |
| name = kasprintf(GFP_KERNEL, "ocxlflash:%d", ctx->pe); |
| file = ocxlflash_getfile(dev, name, fops, ctx, flags); |
| kfree(name); |
| if (IS_ERR(file)) { |
| rc = PTR_ERR(file); |
| dev_err(dev, "%s: ocxlflash_getfile failed rc=%d\n", |
| __func__, rc); |
| goto err2; |
| } |
| |
| ctx->mapping = file->f_mapping; |
| *fd = fdtmp; |
| out: |
| return file; |
| err2: |
| put_unused_fd(fdtmp); |
| err1: |
| file = ERR_PTR(rc); |
| goto out; |
| } |
| |
| /** |
| * ocxlflash_fops_get_context() - get the context associated with the file |
| * @file: File associated with the adapter context. |
| * |
| * Return: pointer to the context |
| */ |
| static void *ocxlflash_fops_get_context(struct file *file) |
| { |
| return file->private_data; |
| } |
| |
| /** |
| * ocxlflash_afu_irq() - interrupt handler for user contexts |
| * @irq: Interrupt number. |
| * @data: Private data provided at interrupt registration, the context. |
| * |
| * Return: Always return IRQ_HANDLED. |
| */ |
| static irqreturn_t ocxlflash_afu_irq(int irq, void *data) |
| { |
| struct ocxlflash_context *ctx = data; |
| struct device *dev = ctx->hw_afu->dev; |
| int i; |
| |
| dev_dbg(dev, "%s: Interrupt raised for pe %i virq %i\n", |
| __func__, ctx->pe, irq); |
| |
| for (i = 0; i < ctx->num_irqs; i++) { |
| if (ctx->irqs[i].virq == irq) |
| break; |
| } |
| if (unlikely(i >= ctx->num_irqs)) { |
| dev_err(dev, "%s: Received AFU IRQ out of range\n", __func__); |
| goto out; |
| } |
| |
| spin_lock(&ctx->slock); |
| set_bit(i - 1, &ctx->irq_bitmap); |
| ctx->pending_irq = true; |
| spin_unlock(&ctx->slock); |
| |
| wake_up_all(&ctx->wq); |
| out: |
| return IRQ_HANDLED; |
| } |
| |
| /** |
| * ocxlflash_start_work() - start a user context |
| * @ctx_cookie: Context to be started. |
| * @num_irqs: Number of interrupts requested. |
| * |
| * Return: 0 on success, -errno on failure |
| */ |
| static int ocxlflash_start_work(void *ctx_cookie, u64 num_irqs) |
| { |
| struct ocxlflash_context *ctx = ctx_cookie; |
| struct ocxl_hw_afu *afu = ctx->hw_afu; |
| struct device *dev = afu->dev; |
| char *name; |
| int rc = 0; |
| int i; |
| |
| rc = alloc_afu_irqs(ctx, num_irqs); |
| if (unlikely(rc < 0)) { |
| dev_err(dev, "%s: alloc_afu_irqs failed rc=%d\n", __func__, rc); |
| goto out; |
| } |
| |
| for (i = 0; i < num_irqs; i++) { |
| name = kasprintf(GFP_KERNEL, "ocxlflash-%s-pe%i-%i", |
| dev_name(dev), ctx->pe, i); |
| rc = afu_map_irq(0, ctx, i, ocxlflash_afu_irq, ctx, name); |
| kfree(name); |
| if (unlikely(rc < 0)) { |
| dev_err(dev, "%s: afu_map_irq failed rc=%d\n", |
| __func__, rc); |
| goto err; |
| } |
| } |
| |
| rc = start_context(ctx); |
| if (unlikely(rc)) { |
| dev_err(dev, "%s: start_context failed rc=%d\n", __func__, rc); |
| goto err; |
| } |
| out: |
| return rc; |
| err: |
| for (i = i-1; i >= 0; i--) |
| afu_unmap_irq(0, ctx, i, ctx); |
| free_afu_irqs(ctx); |
| goto out; |
| }; |
| |
| /** |
| * ocxlflash_fd_mmap() - mmap handler for adapter file descriptor |
| * @file: File installed with adapter file descriptor. |
| * @vma: VM area associated with mapping. |
| * |
| * Return: 0 on success, -errno on failure |
| */ |
| static int ocxlflash_fd_mmap(struct file *file, struct vm_area_struct *vma) |
| { |
| return afu_mmap(file, vma); |
| } |
| |
| /** |
| * ocxlflash_fd_release() - release the context associated with the file |
| * @inode: File inode pointer. |
| * @file: File associated with the adapter context. |
| * |
| * Return: 0 on success, -errno on failure |
| */ |
| static int ocxlflash_fd_release(struct inode *inode, struct file *file) |
| { |
| return afu_release(inode, file); |
| } |
| |
| /* Backend ops to ocxlflash services */ |
| const struct cxlflash_backend_ops cxlflash_ocxl_ops = { |
| .module = THIS_MODULE, |
| .psa_map = ocxlflash_psa_map, |
| .psa_unmap = ocxlflash_psa_unmap, |
| .process_element = ocxlflash_process_element, |
| .map_afu_irq = ocxlflash_map_afu_irq, |
| .unmap_afu_irq = ocxlflash_unmap_afu_irq, |
| .get_irq_objhndl = ocxlflash_get_irq_objhndl, |
| .start_context = ocxlflash_start_context, |
| .stop_context = ocxlflash_stop_context, |
| .afu_reset = ocxlflash_afu_reset, |
| .set_master = ocxlflash_set_master, |
| .get_context = ocxlflash_get_context, |
| .dev_context_init = ocxlflash_dev_context_init, |
| .release_context = ocxlflash_release_context, |
| .perst_reloads_same_image = ocxlflash_perst_reloads_same_image, |
| .read_adapter_vpd = ocxlflash_read_adapter_vpd, |
| .allocate_afu_irqs = ocxlflash_allocate_afu_irqs, |
| .free_afu_irqs = ocxlflash_free_afu_irqs, |
| .create_afu = ocxlflash_create_afu, |
| .destroy_afu = ocxlflash_destroy_afu, |
| .get_fd = ocxlflash_get_fd, |
| .fops_get_context = ocxlflash_fops_get_context, |
| .start_work = ocxlflash_start_work, |
| .fd_mmap = ocxlflash_fd_mmap, |
| .fd_release = ocxlflash_fd_release, |
| }; |