| // SPDX-License-Identifier: GPL-2.0 |
| /* |
| * Copyright (C) 2021 Broadcom. All Rights Reserved. The term |
| * “Broadcom” refers to Broadcom Inc. and/or its subsidiaries. |
| */ |
| |
| #include "efclib.h" |
| #include "../libefc_sli/sli4.h" |
| #include "efc_cmds.h" |
| #include "efc_sm.h" |
| |
| static void |
| efc_nport_free_resources(struct efc_nport *nport, int evt, void *data) |
| { |
| struct efc *efc = nport->efc; |
| |
| /* Clear the nport attached flag */ |
| nport->attached = false; |
| |
| /* Free the service parameters buffer */ |
| if (nport->dma.virt) { |
| dma_free_coherent(&efc->pci->dev, nport->dma.size, |
| nport->dma.virt, nport->dma.phys); |
| memset(&nport->dma, 0, sizeof(struct efc_dma)); |
| } |
| |
| /* Free the SLI resources */ |
| sli_resource_free(efc->sli, SLI4_RSRC_VPI, nport->indicator); |
| |
| efc_nport_cb(efc, evt, nport); |
| } |
| |
| static int |
| efc_nport_get_mbox_status(struct efc_nport *nport, u8 *mqe, int status) |
| { |
| struct efc *efc = nport->efc; |
| struct sli4_mbox_command_header *hdr = |
| (struct sli4_mbox_command_header *)mqe; |
| |
| if (status || le16_to_cpu(hdr->status)) { |
| efc_log_debug(efc, "bad status vpi=%#x st=%x hdr=%x\n", |
| nport->indicator, status, le16_to_cpu(hdr->status)); |
| return -EIO; |
| } |
| |
| return 0; |
| } |
| |
| static int |
| efc_nport_free_unreg_vpi_cb(struct efc *efc, int status, u8 *mqe, void *arg) |
| { |
| struct efc_nport *nport = arg; |
| int evt = EFC_EVT_NPORT_FREE_OK; |
| int rc; |
| |
| rc = efc_nport_get_mbox_status(nport, mqe, status); |
| if (rc) |
| evt = EFC_EVT_NPORT_FREE_FAIL; |
| |
| efc_nport_free_resources(nport, evt, mqe); |
| return rc; |
| } |
| |
| static void |
| efc_nport_free_unreg_vpi(struct efc_nport *nport) |
| { |
| struct efc *efc = nport->efc; |
| int rc; |
| u8 data[SLI4_BMBX_SIZE]; |
| |
| rc = sli_cmd_unreg_vpi(efc->sli, data, nport->indicator, |
| SLI4_UNREG_TYPE_PORT); |
| if (rc) { |
| efc_log_err(efc, "UNREG_VPI format failure\n"); |
| efc_nport_free_resources(nport, EFC_EVT_NPORT_FREE_FAIL, data); |
| return; |
| } |
| |
| rc = efc->tt.issue_mbox_rqst(efc->base, data, |
| efc_nport_free_unreg_vpi_cb, nport); |
| if (rc) { |
| efc_log_err(efc, "UNREG_VPI command failure\n"); |
| efc_nport_free_resources(nport, EFC_EVT_NPORT_FREE_FAIL, data); |
| } |
| } |
| |
| static void |
| efc_nport_send_evt(struct efc_nport *nport, int evt, void *data) |
| { |
| struct efc *efc = nport->efc; |
| |
| /* Now inform the registered callbacks */ |
| efc_nport_cb(efc, evt, nport); |
| |
| /* Set the nport attached flag */ |
| if (evt == EFC_EVT_NPORT_ATTACH_OK) |
| nport->attached = true; |
| |
| /* If there is a pending free request, then handle it now */ |
| if (nport->free_req_pending) |
| efc_nport_free_unreg_vpi(nport); |
| } |
| |
| static int |
| efc_nport_alloc_init_vpi_cb(struct efc *efc, int status, u8 *mqe, void *arg) |
| { |
| struct efc_nport *nport = arg; |
| |
| if (efc_nport_get_mbox_status(nport, mqe, status)) { |
| efc_nport_free_resources(nport, EFC_EVT_NPORT_ALLOC_FAIL, mqe); |
| return -EIO; |
| } |
| |
| efc_nport_send_evt(nport, EFC_EVT_NPORT_ALLOC_OK, mqe); |
| return 0; |
| } |
| |
| static void |
| efc_nport_alloc_init_vpi(struct efc_nport *nport) |
| { |
| struct efc *efc = nport->efc; |
| u8 data[SLI4_BMBX_SIZE]; |
| int rc; |
| |
| /* If there is a pending free request, then handle it now */ |
| if (nport->free_req_pending) { |
| efc_nport_free_resources(nport, EFC_EVT_NPORT_FREE_OK, data); |
| return; |
| } |
| |
| rc = sli_cmd_init_vpi(efc->sli, data, |
| nport->indicator, nport->domain->indicator); |
| if (rc) { |
| efc_log_err(efc, "INIT_VPI format failure\n"); |
| efc_nport_free_resources(nport, EFC_EVT_NPORT_ALLOC_FAIL, data); |
| return; |
| } |
| |
| rc = efc->tt.issue_mbox_rqst(efc->base, data, |
| efc_nport_alloc_init_vpi_cb, nport); |
| if (rc) { |
| efc_log_err(efc, "INIT_VPI command failure\n"); |
| efc_nport_free_resources(nport, EFC_EVT_NPORT_ALLOC_FAIL, data); |
| } |
| } |
| |
| static int |
| efc_nport_alloc_read_sparm64_cb(struct efc *efc, int status, u8 *mqe, void *arg) |
| { |
| struct efc_nport *nport = arg; |
| u8 *payload = NULL; |
| |
| if (efc_nport_get_mbox_status(nport, mqe, status)) { |
| efc_nport_free_resources(nport, EFC_EVT_NPORT_ALLOC_FAIL, mqe); |
| return -EIO; |
| } |
| |
| payload = nport->dma.virt; |
| |
| memcpy(&nport->sli_wwpn, payload + SLI4_READ_SPARM64_WWPN_OFFSET, |
| sizeof(nport->sli_wwpn)); |
| memcpy(&nport->sli_wwnn, payload + SLI4_READ_SPARM64_WWNN_OFFSET, |
| sizeof(nport->sli_wwnn)); |
| |
| dma_free_coherent(&efc->pci->dev, nport->dma.size, nport->dma.virt, |
| nport->dma.phys); |
| memset(&nport->dma, 0, sizeof(struct efc_dma)); |
| efc_nport_alloc_init_vpi(nport); |
| return 0; |
| } |
| |
| static void |
| efc_nport_alloc_read_sparm64(struct efc *efc, struct efc_nport *nport) |
| { |
| u8 data[SLI4_BMBX_SIZE]; |
| int rc; |
| |
| /* Allocate memory for the service parameters */ |
| nport->dma.size = EFC_SPARAM_DMA_SZ; |
| nport->dma.virt = dma_alloc_coherent(&efc->pci->dev, |
| nport->dma.size, &nport->dma.phys, |
| GFP_KERNEL); |
| if (!nport->dma.virt) { |
| efc_log_err(efc, "Failed to allocate DMA memory\n"); |
| efc_nport_free_resources(nport, EFC_EVT_NPORT_ALLOC_FAIL, data); |
| return; |
| } |
| |
| rc = sli_cmd_read_sparm64(efc->sli, data, |
| &nport->dma, nport->indicator); |
| if (rc) { |
| efc_log_err(efc, "READ_SPARM64 format failure\n"); |
| efc_nport_free_resources(nport, EFC_EVT_NPORT_ALLOC_FAIL, data); |
| return; |
| } |
| |
| rc = efc->tt.issue_mbox_rqst(efc->base, data, |
| efc_nport_alloc_read_sparm64_cb, nport); |
| if (rc) { |
| efc_log_err(efc, "READ_SPARM64 command failure\n"); |
| efc_nport_free_resources(nport, EFC_EVT_NPORT_ALLOC_FAIL, data); |
| } |
| } |
| |
| int |
| efc_cmd_nport_alloc(struct efc *efc, struct efc_nport *nport, |
| struct efc_domain *domain, u8 *wwpn) |
| { |
| u32 index; |
| |
| nport->indicator = U32_MAX; |
| nport->free_req_pending = false; |
| |
| if (wwpn) |
| memcpy(&nport->sli_wwpn, wwpn, sizeof(nport->sli_wwpn)); |
| |
| /* |
| * allocate a VPI object for the port and stores it in the |
| * indicator field of the port object. |
| */ |
| if (sli_resource_alloc(efc->sli, SLI4_RSRC_VPI, |
| &nport->indicator, &index)) { |
| efc_log_err(efc, "VPI allocation failure\n"); |
| return -EIO; |
| } |
| |
| if (domain) { |
| /* |
| * If the WWPN is NULL, fetch the default |
| * WWPN and WWNN before initializing the VPI |
| */ |
| if (!wwpn) |
| efc_nport_alloc_read_sparm64(efc, nport); |
| else |
| efc_nport_alloc_init_vpi(nport); |
| } else if (!wwpn) { |
| /* domain NULL and wwpn non-NULL */ |
| efc_log_err(efc, "need WWN for physical port\n"); |
| sli_resource_free(efc->sli, SLI4_RSRC_VPI, nport->indicator); |
| return -EIO; |
| } |
| |
| return 0; |
| } |
| |
| static int |
| efc_nport_attach_reg_vpi_cb(struct efc *efc, int status, u8 *mqe, |
| void *arg) |
| { |
| struct efc_nport *nport = arg; |
| |
| nport->attaching = false; |
| if (efc_nport_get_mbox_status(nport, mqe, status)) { |
| efc_nport_free_resources(nport, EFC_EVT_NPORT_ATTACH_FAIL, mqe); |
| return -EIO; |
| } |
| |
| efc_nport_send_evt(nport, EFC_EVT_NPORT_ATTACH_OK, mqe); |
| return 0; |
| } |
| |
| int |
| efc_cmd_nport_attach(struct efc *efc, struct efc_nport *nport, u32 fc_id) |
| { |
| u8 buf[SLI4_BMBX_SIZE]; |
| int rc = 0; |
| |
| if (!nport) { |
| efc_log_err(efc, "bad param(s) nport=%p\n", nport); |
| return -EIO; |
| } |
| |
| nport->fc_id = fc_id; |
| |
| /* register previously-allocated VPI with the device */ |
| rc = sli_cmd_reg_vpi(efc->sli, buf, nport->fc_id, |
| nport->sli_wwpn, nport->indicator, |
| nport->domain->indicator, false); |
| if (rc) { |
| efc_log_err(efc, "REG_VPI format failure\n"); |
| efc_nport_free_resources(nport, EFC_EVT_NPORT_ATTACH_FAIL, buf); |
| return rc; |
| } |
| |
| rc = efc->tt.issue_mbox_rqst(efc->base, buf, |
| efc_nport_attach_reg_vpi_cb, nport); |
| if (rc) { |
| efc_log_err(efc, "REG_VPI command failure\n"); |
| efc_nport_free_resources(nport, EFC_EVT_NPORT_ATTACH_FAIL, buf); |
| } else { |
| nport->attaching = true; |
| } |
| |
| return rc; |
| } |
| |
| int |
| efc_cmd_nport_free(struct efc *efc, struct efc_nport *nport) |
| { |
| if (!nport) { |
| efc_log_err(efc, "bad parameter(s) nport=%p\n", nport); |
| return -EIO; |
| } |
| |
| /* Issue the UNREG_VPI command to free the assigned VPI context */ |
| if (nport->attached) |
| efc_nport_free_unreg_vpi(nport); |
| else if (nport->attaching) |
| nport->free_req_pending = true; |
| else |
| efc_sm_post_event(&nport->sm, EFC_EVT_NPORT_FREE_OK, NULL); |
| |
| return 0; |
| } |
| |
| static int |
| efc_domain_get_mbox_status(struct efc_domain *domain, u8 *mqe, int status) |
| { |
| struct efc *efc = domain->efc; |
| struct sli4_mbox_command_header *hdr = |
| (struct sli4_mbox_command_header *)mqe; |
| |
| if (status || le16_to_cpu(hdr->status)) { |
| efc_log_debug(efc, "bad status vfi=%#x st=%x hdr=%x\n", |
| domain->indicator, status, |
| le16_to_cpu(hdr->status)); |
| return -EIO; |
| } |
| |
| return 0; |
| } |
| |
| static void |
| efc_domain_free_resources(struct efc_domain *domain, int evt, void *data) |
| { |
| struct efc *efc = domain->efc; |
| |
| /* Free the service parameters buffer */ |
| if (domain->dma.virt) { |
| dma_free_coherent(&efc->pci->dev, |
| domain->dma.size, domain->dma.virt, |
| domain->dma.phys); |
| memset(&domain->dma, 0, sizeof(struct efc_dma)); |
| } |
| |
| /* Free the SLI resources */ |
| sli_resource_free(efc->sli, SLI4_RSRC_VFI, domain->indicator); |
| |
| efc_domain_cb(efc, evt, domain); |
| } |
| |
| static void |
| efc_domain_send_nport_evt(struct efc_domain *domain, |
| int port_evt, int domain_evt, void *data) |
| { |
| struct efc *efc = domain->efc; |
| |
| /* Send alloc/attach ok to the physical nport */ |
| efc_nport_send_evt(domain->nport, port_evt, NULL); |
| |
| /* Now inform the registered callbacks */ |
| efc_domain_cb(efc, domain_evt, domain); |
| } |
| |
| static int |
| efc_domain_alloc_read_sparm64_cb(struct efc *efc, int status, u8 *mqe, |
| void *arg) |
| { |
| struct efc_domain *domain = arg; |
| |
| if (efc_domain_get_mbox_status(domain, mqe, status)) { |
| efc_domain_free_resources(domain, |
| EFC_HW_DOMAIN_ALLOC_FAIL, mqe); |
| return -EIO; |
| } |
| |
| efc_domain_send_nport_evt(domain, EFC_EVT_NPORT_ALLOC_OK, |
| EFC_HW_DOMAIN_ALLOC_OK, mqe); |
| return 0; |
| } |
| |
| static void |
| efc_domain_alloc_read_sparm64(struct efc_domain *domain) |
| { |
| struct efc *efc = domain->efc; |
| u8 data[SLI4_BMBX_SIZE]; |
| int rc; |
| |
| rc = sli_cmd_read_sparm64(efc->sli, data, &domain->dma, 0); |
| if (rc) { |
| efc_log_err(efc, "READ_SPARM64 format failure\n"); |
| efc_domain_free_resources(domain, |
| EFC_HW_DOMAIN_ALLOC_FAIL, data); |
| return; |
| } |
| |
| rc = efc->tt.issue_mbox_rqst(efc->base, data, |
| efc_domain_alloc_read_sparm64_cb, domain); |
| if (rc) { |
| efc_log_err(efc, "READ_SPARM64 command failure\n"); |
| efc_domain_free_resources(domain, |
| EFC_HW_DOMAIN_ALLOC_FAIL, data); |
| } |
| } |
| |
| static int |
| efc_domain_alloc_init_vfi_cb(struct efc *efc, int status, u8 *mqe, |
| void *arg) |
| { |
| struct efc_domain *domain = arg; |
| |
| if (efc_domain_get_mbox_status(domain, mqe, status)) { |
| efc_domain_free_resources(domain, |
| EFC_HW_DOMAIN_ALLOC_FAIL, mqe); |
| return -EIO; |
| } |
| |
| efc_domain_alloc_read_sparm64(domain); |
| return 0; |
| } |
| |
| static void |
| efc_domain_alloc_init_vfi(struct efc_domain *domain) |
| { |
| struct efc *efc = domain->efc; |
| struct efc_nport *nport = domain->nport; |
| u8 data[SLI4_BMBX_SIZE]; |
| int rc; |
| |
| /* |
| * For FC, the HW alread registered an FCFI. |
| * Copy FCF information into the domain and jump to INIT_VFI. |
| */ |
| domain->fcf_indicator = efc->fcfi; |
| rc = sli_cmd_init_vfi(efc->sli, data, domain->indicator, |
| domain->fcf_indicator, nport->indicator); |
| if (rc) { |
| efc_log_err(efc, "INIT_VFI format failure\n"); |
| efc_domain_free_resources(domain, |
| EFC_HW_DOMAIN_ALLOC_FAIL, data); |
| return; |
| } |
| |
| efc_log_err(efc, "%s issue mbox\n", __func__); |
| rc = efc->tt.issue_mbox_rqst(efc->base, data, |
| efc_domain_alloc_init_vfi_cb, domain); |
| if (rc) { |
| efc_log_err(efc, "INIT_VFI command failure\n"); |
| efc_domain_free_resources(domain, |
| EFC_HW_DOMAIN_ALLOC_FAIL, data); |
| } |
| } |
| |
| int |
| efc_cmd_domain_alloc(struct efc *efc, struct efc_domain *domain, u32 fcf) |
| { |
| u32 index; |
| |
| if (!domain || !domain->nport) { |
| efc_log_err(efc, "bad parameter(s) domain=%p nport=%p\n", |
| domain, domain ? domain->nport : NULL); |
| return -EIO; |
| } |
| |
| /* allocate memory for the service parameters */ |
| domain->dma.size = EFC_SPARAM_DMA_SZ; |
| domain->dma.virt = dma_alloc_coherent(&efc->pci->dev, |
| domain->dma.size, |
| &domain->dma.phys, GFP_KERNEL); |
| if (!domain->dma.virt) { |
| efc_log_err(efc, "Failed to allocate DMA memory\n"); |
| return -EIO; |
| } |
| |
| domain->fcf = fcf; |
| domain->fcf_indicator = U32_MAX; |
| domain->indicator = U32_MAX; |
| |
| if (sli_resource_alloc(efc->sli, SLI4_RSRC_VFI, &domain->indicator, |
| &index)) { |
| efc_log_err(efc, "VFI allocation failure\n"); |
| |
| dma_free_coherent(&efc->pci->dev, |
| domain->dma.size, domain->dma.virt, |
| domain->dma.phys); |
| memset(&domain->dma, 0, sizeof(struct efc_dma)); |
| |
| return -EIO; |
| } |
| |
| efc_domain_alloc_init_vfi(domain); |
| return 0; |
| } |
| |
| static int |
| efc_domain_attach_reg_vfi_cb(struct efc *efc, int status, u8 *mqe, |
| void *arg) |
| { |
| struct efc_domain *domain = arg; |
| |
| if (efc_domain_get_mbox_status(domain, mqe, status)) { |
| efc_domain_free_resources(domain, |
| EFC_HW_DOMAIN_ATTACH_FAIL, mqe); |
| return -EIO; |
| } |
| |
| efc_domain_send_nport_evt(domain, EFC_EVT_NPORT_ATTACH_OK, |
| EFC_HW_DOMAIN_ATTACH_OK, mqe); |
| return 0; |
| } |
| |
| int |
| efc_cmd_domain_attach(struct efc *efc, struct efc_domain *domain, u32 fc_id) |
| { |
| u8 buf[SLI4_BMBX_SIZE]; |
| int rc = 0; |
| |
| if (!domain) { |
| efc_log_err(efc, "bad param(s) domain=%p\n", domain); |
| return -EIO; |
| } |
| |
| domain->nport->fc_id = fc_id; |
| |
| rc = sli_cmd_reg_vfi(efc->sli, buf, SLI4_BMBX_SIZE, domain->indicator, |
| domain->fcf_indicator, domain->dma, |
| domain->nport->indicator, domain->nport->sli_wwpn, |
| domain->nport->fc_id); |
| if (rc) { |
| efc_log_err(efc, "REG_VFI format failure\n"); |
| goto cleanup; |
| } |
| |
| rc = efc->tt.issue_mbox_rqst(efc->base, buf, |
| efc_domain_attach_reg_vfi_cb, domain); |
| if (rc) { |
| efc_log_err(efc, "REG_VFI command failure\n"); |
| goto cleanup; |
| } |
| |
| return rc; |
| |
| cleanup: |
| efc_domain_free_resources(domain, EFC_HW_DOMAIN_ATTACH_FAIL, buf); |
| |
| return rc; |
| } |
| |
| static int |
| efc_domain_free_unreg_vfi_cb(struct efc *efc, int status, u8 *mqe, void *arg) |
| { |
| struct efc_domain *domain = arg; |
| int evt = EFC_HW_DOMAIN_FREE_OK; |
| int rc; |
| |
| rc = efc_domain_get_mbox_status(domain, mqe, status); |
| if (rc) { |
| evt = EFC_HW_DOMAIN_FREE_FAIL; |
| rc = -EIO; |
| } |
| |
| efc_domain_free_resources(domain, evt, mqe); |
| return rc; |
| } |
| |
| static void |
| efc_domain_free_unreg_vfi(struct efc_domain *domain) |
| { |
| struct efc *efc = domain->efc; |
| int rc; |
| u8 data[SLI4_BMBX_SIZE]; |
| |
| rc = sli_cmd_unreg_vfi(efc->sli, data, domain->indicator, |
| SLI4_UNREG_TYPE_DOMAIN); |
| if (rc) { |
| efc_log_err(efc, "UNREG_VFI format failure\n"); |
| goto cleanup; |
| } |
| |
| rc = efc->tt.issue_mbox_rqst(efc->base, data, |
| efc_domain_free_unreg_vfi_cb, domain); |
| if (rc) { |
| efc_log_err(efc, "UNREG_VFI command failure\n"); |
| goto cleanup; |
| } |
| |
| return; |
| |
| cleanup: |
| efc_domain_free_resources(domain, EFC_HW_DOMAIN_FREE_FAIL, data); |
| } |
| |
| int |
| efc_cmd_domain_free(struct efc *efc, struct efc_domain *domain) |
| { |
| if (!domain) { |
| efc_log_err(efc, "bad parameter(s) domain=%p\n", domain); |
| return -EIO; |
| } |
| |
| efc_domain_free_unreg_vfi(domain); |
| return 0; |
| } |
| |
| int |
| efc_cmd_node_alloc(struct efc *efc, struct efc_remote_node *rnode, u32 fc_addr, |
| struct efc_nport *nport) |
| { |
| /* Check for invalid indicator */ |
| if (rnode->indicator != U32_MAX) { |
| efc_log_err(efc, |
| "RPI allocation failure addr=%#x rpi=%#x\n", |
| fc_addr, rnode->indicator); |
| return -EIO; |
| } |
| |
| /* NULL SLI port indicates an unallocated remote node */ |
| rnode->nport = NULL; |
| |
| if (sli_resource_alloc(efc->sli, SLI4_RSRC_RPI, |
| &rnode->indicator, &rnode->index)) { |
| efc_log_err(efc, "RPI allocation failure addr=%#x\n", |
| fc_addr); |
| return -EIO; |
| } |
| |
| rnode->fc_id = fc_addr; |
| rnode->nport = nport; |
| |
| return 0; |
| } |
| |
| static int |
| efc_cmd_node_attach_cb(struct efc *efc, int status, u8 *mqe, void *arg) |
| { |
| struct efc_remote_node *rnode = arg; |
| struct sli4_mbox_command_header *hdr = |
| (struct sli4_mbox_command_header *)mqe; |
| int evt = 0; |
| |
| if (status || le16_to_cpu(hdr->status)) { |
| efc_log_debug(efc, "bad status cqe=%#x mqe=%#x\n", status, |
| le16_to_cpu(hdr->status)); |
| rnode->attached = false; |
| evt = EFC_EVT_NODE_ATTACH_FAIL; |
| } else { |
| rnode->attached = true; |
| evt = EFC_EVT_NODE_ATTACH_OK; |
| } |
| |
| efc_remote_node_cb(efc, evt, rnode); |
| |
| return 0; |
| } |
| |
| int |
| efc_cmd_node_attach(struct efc *efc, struct efc_remote_node *rnode, |
| struct efc_dma *sparms) |
| { |
| int rc = -EIO; |
| u8 buf[SLI4_BMBX_SIZE]; |
| |
| if (!rnode || !sparms) { |
| efc_log_err(efc, "bad parameter(s) rnode=%p sparms=%p\n", |
| rnode, sparms); |
| return -EIO; |
| } |
| |
| /* |
| * If the attach count is non-zero, this RPI has already been reg'd. |
| * Otherwise, register the RPI |
| */ |
| if (rnode->index == U32_MAX) { |
| efc_log_err(efc, "bad parameter rnode->index invalid\n"); |
| return -EIO; |
| } |
| |
| /* Update a remote node object with the remote port's service params */ |
| if (!sli_cmd_reg_rpi(efc->sli, buf, rnode->indicator, |
| rnode->nport->indicator, rnode->fc_id, sparms, 0, 0)) |
| rc = efc->tt.issue_mbox_rqst(efc->base, buf, |
| efc_cmd_node_attach_cb, rnode); |
| |
| return rc; |
| } |
| |
| int |
| efc_node_free_resources(struct efc *efc, struct efc_remote_node *rnode) |
| { |
| int rc = 0; |
| |
| if (!rnode) { |
| efc_log_err(efc, "bad parameter rnode=%p\n", rnode); |
| return -EIO; |
| } |
| |
| if (rnode->nport) { |
| if (rnode->attached) { |
| efc_log_err(efc, "rnode is still attached\n"); |
| return -EIO; |
| } |
| if (rnode->indicator != U32_MAX) { |
| if (sli_resource_free(efc->sli, SLI4_RSRC_RPI, |
| rnode->indicator)) { |
| efc_log_err(efc, |
| "RPI free fail RPI %d addr=%#x\n", |
| rnode->indicator, rnode->fc_id); |
| rc = -EIO; |
| } else { |
| rnode->indicator = U32_MAX; |
| rnode->index = U32_MAX; |
| } |
| } |
| } |
| |
| return rc; |
| } |
| |
| static int |
| efc_cmd_node_free_cb(struct efc *efc, int status, u8 *mqe, void *arg) |
| { |
| struct efc_remote_node *rnode = arg; |
| struct sli4_mbox_command_header *hdr = |
| (struct sli4_mbox_command_header *)mqe; |
| int evt = EFC_EVT_NODE_FREE_FAIL; |
| int rc = 0; |
| |
| if (status || le16_to_cpu(hdr->status)) { |
| efc_log_debug(efc, "bad status cqe=%#x mqe=%#x\n", status, |
| le16_to_cpu(hdr->status)); |
| |
| /* |
| * In certain cases, a non-zero MQE status is OK (all must be |
| * true): |
| * - node is attached |
| * - status is 0x1400 |
| */ |
| if (!rnode->attached || |
| (le16_to_cpu(hdr->status) != SLI4_MBX_STATUS_RPI_NOT_REG)) |
| rc = -EIO; |
| } |
| |
| if (!rc) { |
| rnode->attached = false; |
| evt = EFC_EVT_NODE_FREE_OK; |
| } |
| |
| efc_remote_node_cb(efc, evt, rnode); |
| |
| return rc; |
| } |
| |
| int |
| efc_cmd_node_detach(struct efc *efc, struct efc_remote_node *rnode) |
| { |
| u8 buf[SLI4_BMBX_SIZE]; |
| int rc = -EIO; |
| |
| if (!rnode) { |
| efc_log_err(efc, "bad parameter rnode=%p\n", rnode); |
| return -EIO; |
| } |
| |
| if (rnode->nport) { |
| if (!rnode->attached) |
| return -EIO; |
| |
| rc = -EIO; |
| |
| if (!sli_cmd_unreg_rpi(efc->sli, buf, rnode->indicator, |
| SLI4_RSRC_RPI, U32_MAX)) |
| rc = efc->tt.issue_mbox_rqst(efc->base, buf, |
| efc_cmd_node_free_cb, rnode); |
| |
| if (rc != 0) { |
| efc_log_err(efc, "UNREG_RPI failed\n"); |
| rc = -EIO; |
| } |
| } |
| |
| return rc; |
| } |