| // SPDX-License-Identifier: GPL-2.0-or-later |
| /* |
| * Copyright 2016-17 IBM Corp. |
| */ |
| |
| #define pr_fmt(fmt) "vas: " fmt |
| |
| #include <linux/types.h> |
| #include <linux/mutex.h> |
| #include <linux/slab.h> |
| #include <linux/io.h> |
| #include <linux/log2.h> |
| #include <linux/rcupdate.h> |
| #include <linux/cred.h> |
| #include <linux/sched/mm.h> |
| #include <linux/mmu_context.h> |
| #include <asm/switch_to.h> |
| #include <asm/ppc-opcode.h> |
| #include <asm/vas.h> |
| #include "vas.h" |
| #include "copy-paste.h" |
| |
| #define CREATE_TRACE_POINTS |
| #include "vas-trace.h" |
| |
| /* |
| * Compute the paste address region for the window @window using the |
| * ->paste_base_addr and ->paste_win_id_shift we got from device tree. |
| */ |
| void vas_win_paste_addr(struct pnv_vas_window *window, u64 *addr, int *len) |
| { |
| int winid; |
| u64 base, shift; |
| |
| base = window->vinst->paste_base_addr; |
| shift = window->vinst->paste_win_id_shift; |
| winid = window->vas_win.winid; |
| |
| *addr = base + (winid << shift); |
| if (len) |
| *len = PAGE_SIZE; |
| |
| pr_debug("Txwin #%d: Paste addr 0x%llx\n", winid, *addr); |
| } |
| |
| static inline void get_hvwc_mmio_bar(struct pnv_vas_window *window, |
| u64 *start, int *len) |
| { |
| u64 pbaddr; |
| |
| pbaddr = window->vinst->hvwc_bar_start; |
| *start = pbaddr + window->vas_win.winid * VAS_HVWC_SIZE; |
| *len = VAS_HVWC_SIZE; |
| } |
| |
| static inline void get_uwc_mmio_bar(struct pnv_vas_window *window, |
| u64 *start, int *len) |
| { |
| u64 pbaddr; |
| |
| pbaddr = window->vinst->uwc_bar_start; |
| *start = pbaddr + window->vas_win.winid * VAS_UWC_SIZE; |
| *len = VAS_UWC_SIZE; |
| } |
| |
| /* |
| * Map the paste bus address of the given send window into kernel address |
| * space. Unlike MMIO regions (map_mmio_region() below), paste region must |
| * be mapped cache-able and is only applicable to send windows. |
| */ |
| static void *map_paste_region(struct pnv_vas_window *txwin) |
| { |
| int len; |
| void *map; |
| char *name; |
| u64 start; |
| |
| name = kasprintf(GFP_KERNEL, "window-v%d-w%d", txwin->vinst->vas_id, |
| txwin->vas_win.winid); |
| if (!name) |
| goto free_name; |
| |
| txwin->paste_addr_name = name; |
| vas_win_paste_addr(txwin, &start, &len); |
| |
| if (!request_mem_region(start, len, name)) { |
| pr_devel("%s(): request_mem_region(0x%llx, %d) failed\n", |
| __func__, start, len); |
| goto free_name; |
| } |
| |
| map = ioremap_cache(start, len); |
| if (!map) { |
| pr_devel("%s(): ioremap_cache(0x%llx, %d) failed\n", __func__, |
| start, len); |
| goto free_name; |
| } |
| |
| pr_devel("Mapped paste addr 0x%llx to kaddr 0x%p\n", start, map); |
| return map; |
| |
| free_name: |
| kfree(name); |
| return ERR_PTR(-ENOMEM); |
| } |
| |
| static void *map_mmio_region(char *name, u64 start, int len) |
| { |
| void *map; |
| |
| if (!request_mem_region(start, len, name)) { |
| pr_devel("%s(): request_mem_region(0x%llx, %d) failed\n", |
| __func__, start, len); |
| return NULL; |
| } |
| |
| map = ioremap(start, len); |
| if (!map) { |
| pr_devel("%s(): ioremap(0x%llx, %d) failed\n", __func__, start, |
| len); |
| return NULL; |
| } |
| |
| return map; |
| } |
| |
| static void unmap_region(void *addr, u64 start, int len) |
| { |
| iounmap(addr); |
| release_mem_region((phys_addr_t)start, len); |
| } |
| |
| /* |
| * Unmap the paste address region for a window. |
| */ |
| static void unmap_paste_region(struct pnv_vas_window *window) |
| { |
| int len; |
| u64 busaddr_start; |
| |
| if (window->paste_kaddr) { |
| vas_win_paste_addr(window, &busaddr_start, &len); |
| unmap_region(window->paste_kaddr, busaddr_start, len); |
| window->paste_kaddr = NULL; |
| kfree(window->paste_addr_name); |
| window->paste_addr_name = NULL; |
| } |
| } |
| |
| /* |
| * Unmap the MMIO regions for a window. Hold the vas_mutex so we don't |
| * unmap when the window's debugfs dir is in use. This serializes close |
| * of a window even on another VAS instance but since its not a critical |
| * path, just minimize the time we hold the mutex for now. We can add |
| * a per-instance mutex later if necessary. |
| */ |
| static void unmap_winctx_mmio_bars(struct pnv_vas_window *window) |
| { |
| int len; |
| void *uwc_map; |
| void *hvwc_map; |
| u64 busaddr_start; |
| |
| mutex_lock(&vas_mutex); |
| |
| hvwc_map = window->hvwc_map; |
| window->hvwc_map = NULL; |
| |
| uwc_map = window->uwc_map; |
| window->uwc_map = NULL; |
| |
| mutex_unlock(&vas_mutex); |
| |
| if (hvwc_map) { |
| get_hvwc_mmio_bar(window, &busaddr_start, &len); |
| unmap_region(hvwc_map, busaddr_start, len); |
| } |
| |
| if (uwc_map) { |
| get_uwc_mmio_bar(window, &busaddr_start, &len); |
| unmap_region(uwc_map, busaddr_start, len); |
| } |
| } |
| |
| /* |
| * Find the Hypervisor Window Context (HVWC) MMIO Base Address Region and the |
| * OS/User Window Context (UWC) MMIO Base Address Region for the given window. |
| * Map these bus addresses and save the mapped kernel addresses in @window. |
| */ |
| static int map_winctx_mmio_bars(struct pnv_vas_window *window) |
| { |
| int len; |
| u64 start; |
| |
| get_hvwc_mmio_bar(window, &start, &len); |
| window->hvwc_map = map_mmio_region("HVWCM_Window", start, len); |
| |
| get_uwc_mmio_bar(window, &start, &len); |
| window->uwc_map = map_mmio_region("UWCM_Window", start, len); |
| |
| if (!window->hvwc_map || !window->uwc_map) { |
| unmap_winctx_mmio_bars(window); |
| return -1; |
| } |
| |
| return 0; |
| } |
| |
| /* |
| * Reset all valid registers in the HV and OS/User Window Contexts for |
| * the window identified by @window. |
| * |
| * NOTE: We cannot really use a for loop to reset window context. Not all |
| * offsets in a window context are valid registers and the valid |
| * registers are not sequential. And, we can only write to offsets |
| * with valid registers. |
| */ |
| static void reset_window_regs(struct pnv_vas_window *window) |
| { |
| write_hvwc_reg(window, VREG(LPID), 0ULL); |
| write_hvwc_reg(window, VREG(PID), 0ULL); |
| write_hvwc_reg(window, VREG(XLATE_MSR), 0ULL); |
| write_hvwc_reg(window, VREG(XLATE_LPCR), 0ULL); |
| write_hvwc_reg(window, VREG(XLATE_CTL), 0ULL); |
| write_hvwc_reg(window, VREG(AMR), 0ULL); |
| write_hvwc_reg(window, VREG(SEIDR), 0ULL); |
| write_hvwc_reg(window, VREG(FAULT_TX_WIN), 0ULL); |
| write_hvwc_reg(window, VREG(OSU_INTR_SRC_RA), 0ULL); |
| write_hvwc_reg(window, VREG(HV_INTR_SRC_RA), 0ULL); |
| write_hvwc_reg(window, VREG(PSWID), 0ULL); |
| write_hvwc_reg(window, VREG(LFIFO_BAR), 0ULL); |
| write_hvwc_reg(window, VREG(LDATA_STAMP_CTL), 0ULL); |
| write_hvwc_reg(window, VREG(LDMA_CACHE_CTL), 0ULL); |
| write_hvwc_reg(window, VREG(LRFIFO_PUSH), 0ULL); |
| write_hvwc_reg(window, VREG(CURR_MSG_COUNT), 0ULL); |
| write_hvwc_reg(window, VREG(LNOTIFY_AFTER_COUNT), 0ULL); |
| write_hvwc_reg(window, VREG(LRX_WCRED), 0ULL); |
| write_hvwc_reg(window, VREG(LRX_WCRED_ADDER), 0ULL); |
| write_hvwc_reg(window, VREG(TX_WCRED), 0ULL); |
| write_hvwc_reg(window, VREG(TX_WCRED_ADDER), 0ULL); |
| write_hvwc_reg(window, VREG(LFIFO_SIZE), 0ULL); |
| write_hvwc_reg(window, VREG(WINCTL), 0ULL); |
| write_hvwc_reg(window, VREG(WIN_STATUS), 0ULL); |
| write_hvwc_reg(window, VREG(WIN_CTX_CACHING_CTL), 0ULL); |
| write_hvwc_reg(window, VREG(TX_RSVD_BUF_COUNT), 0ULL); |
| write_hvwc_reg(window, VREG(LRFIFO_WIN_PTR), 0ULL); |
| write_hvwc_reg(window, VREG(LNOTIFY_CTL), 0ULL); |
| write_hvwc_reg(window, VREG(LNOTIFY_PID), 0ULL); |
| write_hvwc_reg(window, VREG(LNOTIFY_LPID), 0ULL); |
| write_hvwc_reg(window, VREG(LNOTIFY_TID), 0ULL); |
| write_hvwc_reg(window, VREG(LNOTIFY_SCOPE), 0ULL); |
| write_hvwc_reg(window, VREG(NX_UTIL_ADDER), 0ULL); |
| |
| /* Skip read-only registers: NX_UTIL and NX_UTIL_SE */ |
| |
| /* |
| * The send and receive window credit adder registers are also |
| * accessible from HVWC and have been initialized above. We don't |
| * need to initialize from the OS/User Window Context, so skip |
| * following calls: |
| * |
| * write_uwc_reg(window, VREG(TX_WCRED_ADDER), 0ULL); |
| * write_uwc_reg(window, VREG(LRX_WCRED_ADDER), 0ULL); |
| */ |
| } |
| |
| /* |
| * Initialize window context registers related to Address Translation. |
| * These registers are common to send/receive windows although they |
| * differ for user/kernel windows. As we resolve the TODOs we may |
| * want to add fields to vas_winctx and move the initialization to |
| * init_vas_winctx_regs(). |
| */ |
| static void init_xlate_regs(struct pnv_vas_window *window, bool user_win) |
| { |
| u64 lpcr, val; |
| |
| /* |
| * MSR_TA, MSR_US are false for both kernel and user. |
| * MSR_DR and MSR_PR are false for kernel. |
| */ |
| val = 0ULL; |
| val = SET_FIELD(VAS_XLATE_MSR_HV, val, 1); |
| val = SET_FIELD(VAS_XLATE_MSR_SF, val, 1); |
| if (user_win) { |
| val = SET_FIELD(VAS_XLATE_MSR_DR, val, 1); |
| val = SET_FIELD(VAS_XLATE_MSR_PR, val, 1); |
| } |
| write_hvwc_reg(window, VREG(XLATE_MSR), val); |
| |
| lpcr = mfspr(SPRN_LPCR); |
| val = 0ULL; |
| /* |
| * NOTE: From Section 5.7.8.1 Segment Lookaside Buffer of the |
| * Power ISA, v3.0B, Page size encoding is 0 = 4KB, 5 = 64KB. |
| * |
| * NOTE: From Section 1.3.1, Address Translation Context of the |
| * Nest MMU Workbook, LPCR_SC should be 0 for Power9. |
| */ |
| val = SET_FIELD(VAS_XLATE_LPCR_PAGE_SIZE, val, 5); |
| val = SET_FIELD(VAS_XLATE_LPCR_ISL, val, lpcr & LPCR_ISL); |
| val = SET_FIELD(VAS_XLATE_LPCR_TC, val, lpcr & LPCR_TC); |
| val = SET_FIELD(VAS_XLATE_LPCR_SC, val, 0); |
| write_hvwc_reg(window, VREG(XLATE_LPCR), val); |
| |
| /* |
| * Section 1.3.1 (Address translation Context) of NMMU workbook. |
| * 0b00 Hashed Page Table mode |
| * 0b01 Reserved |
| * 0b10 Radix on HPT |
| * 0b11 Radix on Radix |
| */ |
| val = 0ULL; |
| val = SET_FIELD(VAS_XLATE_MODE, val, radix_enabled() ? 3 : 2); |
| write_hvwc_reg(window, VREG(XLATE_CTL), val); |
| |
| /* |
| * TODO: Can we mfspr(AMR) even for user windows? |
| */ |
| val = 0ULL; |
| val = SET_FIELD(VAS_AMR, val, mfspr(SPRN_AMR)); |
| write_hvwc_reg(window, VREG(AMR), val); |
| |
| val = 0ULL; |
| val = SET_FIELD(VAS_SEIDR, val, 0); |
| write_hvwc_reg(window, VREG(SEIDR), val); |
| } |
| |
| /* |
| * Initialize Reserved Send Buffer Count for the send window. It involves |
| * writing to the register, reading it back to confirm that the hardware |
| * has enough buffers to reserve. See section 1.3.1.2.1 of VAS workbook. |
| * |
| * Since we can only make a best-effort attempt to fulfill the request, |
| * we don't return any errors if we cannot. |
| * |
| * TODO: Reserved (aka dedicated) send buffers are not supported yet. |
| */ |
| static void init_rsvd_tx_buf_count(struct pnv_vas_window *txwin, |
| struct vas_winctx *winctx) |
| { |
| write_hvwc_reg(txwin, VREG(TX_RSVD_BUF_COUNT), 0ULL); |
| } |
| |
| /* |
| * init_winctx_regs() |
| * Initialize window context registers for a receive window. |
| * Except for caching control and marking window open, the registers |
| * are initialized in the order listed in Section 3.1.4 (Window Context |
| * Cache Register Details) of the VAS workbook although they don't need |
| * to be. |
| * |
| * Design note: For NX receive windows, NX allocates the FIFO buffer in OPAL |
| * (so that it can get a large contiguous area) and passes that buffer |
| * to kernel via device tree. We now write that buffer address to the |
| * FIFO BAR. Would it make sense to do this all in OPAL? i.e have OPAL |
| * write the per-chip RX FIFO addresses to the windows during boot-up |
| * as a one-time task? That could work for NX but what about other |
| * receivers? Let the receivers tell us the rx-fifo buffers for now. |
| */ |
| static void init_winctx_regs(struct pnv_vas_window *window, |
| struct vas_winctx *winctx) |
| { |
| u64 val; |
| int fifo_size; |
| |
| reset_window_regs(window); |
| |
| val = 0ULL; |
| val = SET_FIELD(VAS_LPID, val, winctx->lpid); |
| write_hvwc_reg(window, VREG(LPID), val); |
| |
| val = 0ULL; |
| val = SET_FIELD(VAS_PID_ID, val, winctx->pidr); |
| write_hvwc_reg(window, VREG(PID), val); |
| |
| init_xlate_regs(window, winctx->user_win); |
| |
| val = 0ULL; |
| val = SET_FIELD(VAS_FAULT_TX_WIN, val, winctx->fault_win_id); |
| write_hvwc_reg(window, VREG(FAULT_TX_WIN), val); |
| |
| /* In PowerNV, interrupts go to HV. */ |
| write_hvwc_reg(window, VREG(OSU_INTR_SRC_RA), 0ULL); |
| |
| val = 0ULL; |
| val = SET_FIELD(VAS_HV_INTR_SRC_RA, val, winctx->irq_port); |
| write_hvwc_reg(window, VREG(HV_INTR_SRC_RA), val); |
| |
| val = 0ULL; |
| val = SET_FIELD(VAS_PSWID_EA_HANDLE, val, winctx->pswid); |
| write_hvwc_reg(window, VREG(PSWID), val); |
| |
| write_hvwc_reg(window, VREG(SPARE1), 0ULL); |
| write_hvwc_reg(window, VREG(SPARE2), 0ULL); |
| write_hvwc_reg(window, VREG(SPARE3), 0ULL); |
| |
| /* |
| * NOTE: VAS expects the FIFO address to be copied into the LFIFO_BAR |
| * register as is - do NOT shift the address into VAS_LFIFO_BAR |
| * bit fields! Ok to set the page migration select fields - |
| * VAS ignores the lower 10+ bits in the address anyway, because |
| * the minimum FIFO size is 1K? |
| * |
| * See also: Design note in function header. |
| */ |
| val = __pa(winctx->rx_fifo); |
| val = SET_FIELD(VAS_PAGE_MIGRATION_SELECT, val, 0); |
| write_hvwc_reg(window, VREG(LFIFO_BAR), val); |
| |
| val = 0ULL; |
| val = SET_FIELD(VAS_LDATA_STAMP, val, winctx->data_stamp); |
| write_hvwc_reg(window, VREG(LDATA_STAMP_CTL), val); |
| |
| val = 0ULL; |
| val = SET_FIELD(VAS_LDMA_TYPE, val, winctx->dma_type); |
| val = SET_FIELD(VAS_LDMA_FIFO_DISABLE, val, winctx->fifo_disable); |
| write_hvwc_reg(window, VREG(LDMA_CACHE_CTL), val); |
| |
| write_hvwc_reg(window, VREG(LRFIFO_PUSH), 0ULL); |
| write_hvwc_reg(window, VREG(CURR_MSG_COUNT), 0ULL); |
| write_hvwc_reg(window, VREG(LNOTIFY_AFTER_COUNT), 0ULL); |
| |
| val = 0ULL; |
| val = SET_FIELD(VAS_LRX_WCRED, val, winctx->wcreds_max); |
| write_hvwc_reg(window, VREG(LRX_WCRED), val); |
| |
| val = 0ULL; |
| val = SET_FIELD(VAS_TX_WCRED, val, winctx->wcreds_max); |
| write_hvwc_reg(window, VREG(TX_WCRED), val); |
| |
| write_hvwc_reg(window, VREG(LRX_WCRED_ADDER), 0ULL); |
| write_hvwc_reg(window, VREG(TX_WCRED_ADDER), 0ULL); |
| |
| fifo_size = winctx->rx_fifo_size / 1024; |
| |
| val = 0ULL; |
| val = SET_FIELD(VAS_LFIFO_SIZE, val, ilog2(fifo_size)); |
| write_hvwc_reg(window, VREG(LFIFO_SIZE), val); |
| |
| /* Update window control and caching control registers last so |
| * we mark the window open only after fully initializing it and |
| * pushing context to cache. |
| */ |
| |
| write_hvwc_reg(window, VREG(WIN_STATUS), 0ULL); |
| |
| init_rsvd_tx_buf_count(window, winctx); |
| |
| /* for a send window, point to the matching receive window */ |
| val = 0ULL; |
| val = SET_FIELD(VAS_LRX_WIN_ID, val, winctx->rx_win_id); |
| write_hvwc_reg(window, VREG(LRFIFO_WIN_PTR), val); |
| |
| write_hvwc_reg(window, VREG(SPARE4), 0ULL); |
| |
| val = 0ULL; |
| val = SET_FIELD(VAS_NOTIFY_DISABLE, val, winctx->notify_disable); |
| val = SET_FIELD(VAS_INTR_DISABLE, val, winctx->intr_disable); |
| val = SET_FIELD(VAS_NOTIFY_EARLY, val, winctx->notify_early); |
| val = SET_FIELD(VAS_NOTIFY_OSU_INTR, val, winctx->notify_os_intr_reg); |
| write_hvwc_reg(window, VREG(LNOTIFY_CTL), val); |
| |
| val = 0ULL; |
| val = SET_FIELD(VAS_LNOTIFY_PID, val, winctx->lnotify_pid); |
| write_hvwc_reg(window, VREG(LNOTIFY_PID), val); |
| |
| val = 0ULL; |
| val = SET_FIELD(VAS_LNOTIFY_LPID, val, winctx->lnotify_lpid); |
| write_hvwc_reg(window, VREG(LNOTIFY_LPID), val); |
| |
| val = 0ULL; |
| val = SET_FIELD(VAS_LNOTIFY_TID, val, winctx->lnotify_tid); |
| write_hvwc_reg(window, VREG(LNOTIFY_TID), val); |
| |
| val = 0ULL; |
| val = SET_FIELD(VAS_LNOTIFY_MIN_SCOPE, val, winctx->min_scope); |
| val = SET_FIELD(VAS_LNOTIFY_MAX_SCOPE, val, winctx->max_scope); |
| write_hvwc_reg(window, VREG(LNOTIFY_SCOPE), val); |
| |
| /* Skip read-only registers NX_UTIL and NX_UTIL_SE */ |
| |
| write_hvwc_reg(window, VREG(SPARE5), 0ULL); |
| write_hvwc_reg(window, VREG(NX_UTIL_ADDER), 0ULL); |
| write_hvwc_reg(window, VREG(SPARE6), 0ULL); |
| |
| /* Finally, push window context to memory and... */ |
| val = 0ULL; |
| val = SET_FIELD(VAS_PUSH_TO_MEM, val, 1); |
| write_hvwc_reg(window, VREG(WIN_CTX_CACHING_CTL), val); |
| |
| /* ... mark the window open for business */ |
| val = 0ULL; |
| val = SET_FIELD(VAS_WINCTL_REJ_NO_CREDIT, val, winctx->rej_no_credit); |
| val = SET_FIELD(VAS_WINCTL_PIN, val, winctx->pin_win); |
| val = SET_FIELD(VAS_WINCTL_TX_WCRED_MODE, val, winctx->tx_wcred_mode); |
| val = SET_FIELD(VAS_WINCTL_RX_WCRED_MODE, val, winctx->rx_wcred_mode); |
| val = SET_FIELD(VAS_WINCTL_TX_WORD_MODE, val, winctx->tx_word_mode); |
| val = SET_FIELD(VAS_WINCTL_RX_WORD_MODE, val, winctx->rx_word_mode); |
| val = SET_FIELD(VAS_WINCTL_FAULT_WIN, val, winctx->fault_win); |
| val = SET_FIELD(VAS_WINCTL_NX_WIN, val, winctx->nx_win); |
| val = SET_FIELD(VAS_WINCTL_OPEN, val, 1); |
| write_hvwc_reg(window, VREG(WINCTL), val); |
| } |
| |
| static void vas_release_window_id(struct ida *ida, int winid) |
| { |
| ida_free(ida, winid); |
| } |
| |
| static int vas_assign_window_id(struct ida *ida) |
| { |
| int winid = ida_alloc_max(ida, VAS_WINDOWS_PER_CHIP - 1, GFP_KERNEL); |
| |
| if (winid == -ENOSPC) { |
| pr_err("Too many (%d) open windows\n", VAS_WINDOWS_PER_CHIP); |
| return -EAGAIN; |
| } |
| |
| return winid; |
| } |
| |
| static void vas_window_free(struct pnv_vas_window *window) |
| { |
| struct vas_instance *vinst = window->vinst; |
| int winid = window->vas_win.winid; |
| |
| unmap_winctx_mmio_bars(window); |
| |
| vas_window_free_dbgdir(window); |
| |
| kfree(window); |
| |
| vas_release_window_id(&vinst->ida, winid); |
| } |
| |
| static struct pnv_vas_window *vas_window_alloc(struct vas_instance *vinst) |
| { |
| int winid; |
| struct pnv_vas_window *window; |
| |
| winid = vas_assign_window_id(&vinst->ida); |
| if (winid < 0) |
| return ERR_PTR(winid); |
| |
| window = kzalloc(sizeof(*window), GFP_KERNEL); |
| if (!window) |
| goto out_free; |
| |
| window->vinst = vinst; |
| window->vas_win.winid = winid; |
| |
| if (map_winctx_mmio_bars(window)) |
| goto out_free; |
| |
| vas_window_init_dbgdir(window); |
| |
| return window; |
| |
| out_free: |
| kfree(window); |
| vas_release_window_id(&vinst->ida, winid); |
| return ERR_PTR(-ENOMEM); |
| } |
| |
| static void put_rx_win(struct pnv_vas_window *rxwin) |
| { |
| /* Better not be a send window! */ |
| WARN_ON_ONCE(rxwin->tx_win); |
| |
| atomic_dec(&rxwin->num_txwins); |
| } |
| |
| /* |
| * Find the user space receive window given the @pswid. |
| * - We must have a valid vasid and it must belong to this instance. |
| * (so both send and receive windows are on the same VAS instance) |
| * - The window must refer to an OPEN, FTW, RECEIVE window. |
| * |
| * NOTE: We access ->windows[] table and assume that vinst->mutex is held. |
| */ |
| static struct pnv_vas_window *get_user_rxwin(struct vas_instance *vinst, |
| u32 pswid) |
| { |
| int vasid, winid; |
| struct pnv_vas_window *rxwin; |
| |
| decode_pswid(pswid, &vasid, &winid); |
| |
| if (vinst->vas_id != vasid) |
| return ERR_PTR(-EINVAL); |
| |
| rxwin = vinst->windows[winid]; |
| |
| if (!rxwin || rxwin->tx_win || rxwin->vas_win.cop != VAS_COP_TYPE_FTW) |
| return ERR_PTR(-EINVAL); |
| |
| return rxwin; |
| } |
| |
| /* |
| * Get the VAS receive window associated with NX engine identified |
| * by @cop and if applicable, @pswid. |
| * |
| * See also function header of set_vinst_win(). |
| */ |
| static struct pnv_vas_window *get_vinst_rxwin(struct vas_instance *vinst, |
| enum vas_cop_type cop, u32 pswid) |
| { |
| struct pnv_vas_window *rxwin; |
| |
| mutex_lock(&vinst->mutex); |
| |
| if (cop == VAS_COP_TYPE_FTW) |
| rxwin = get_user_rxwin(vinst, pswid); |
| else |
| rxwin = vinst->rxwin[cop] ?: ERR_PTR(-EINVAL); |
| |
| if (!IS_ERR(rxwin)) |
| atomic_inc(&rxwin->num_txwins); |
| |
| mutex_unlock(&vinst->mutex); |
| |
| return rxwin; |
| } |
| |
| /* |
| * We have two tables of windows in a VAS instance. The first one, |
| * ->windows[], contains all the windows in the instance and allows |
| * looking up a window by its id. It is used to look up send windows |
| * during fault handling and receive windows when pairing user space |
| * send/receive windows. |
| * |
| * The second table, ->rxwin[], contains receive windows that are |
| * associated with NX engines. This table has VAS_COP_TYPE_MAX |
| * entries and is used to look up a receive window by its |
| * coprocessor type. |
| * |
| * Here, we save @window in the ->windows[] table. If it is a receive |
| * window, we also save the window in the ->rxwin[] table. |
| */ |
| static void set_vinst_win(struct vas_instance *vinst, |
| struct pnv_vas_window *window) |
| { |
| int id = window->vas_win.winid; |
| |
| mutex_lock(&vinst->mutex); |
| |
| /* |
| * There should only be one receive window for a coprocessor type |
| * unless its a user (FTW) window. |
| */ |
| if (!window->user_win && !window->tx_win) { |
| WARN_ON_ONCE(vinst->rxwin[window->vas_win.cop]); |
| vinst->rxwin[window->vas_win.cop] = window; |
| } |
| |
| WARN_ON_ONCE(vinst->windows[id] != NULL); |
| vinst->windows[id] = window; |
| |
| mutex_unlock(&vinst->mutex); |
| } |
| |
| /* |
| * Clear this window from the table(s) of windows for this VAS instance. |
| * See also function header of set_vinst_win(). |
| */ |
| static void clear_vinst_win(struct pnv_vas_window *window) |
| { |
| int id = window->vas_win.winid; |
| struct vas_instance *vinst = window->vinst; |
| |
| mutex_lock(&vinst->mutex); |
| |
| if (!window->user_win && !window->tx_win) { |
| WARN_ON_ONCE(!vinst->rxwin[window->vas_win.cop]); |
| vinst->rxwin[window->vas_win.cop] = NULL; |
| } |
| |
| WARN_ON_ONCE(vinst->windows[id] != window); |
| vinst->windows[id] = NULL; |
| |
| mutex_unlock(&vinst->mutex); |
| } |
| |
| static void init_winctx_for_rxwin(struct pnv_vas_window *rxwin, |
| struct vas_rx_win_attr *rxattr, |
| struct vas_winctx *winctx) |
| { |
| /* |
| * We first zero (memset()) all fields and only set non-zero fields. |
| * Following fields are 0/false but maybe deserve a comment: |
| * |
| * ->notify_os_intr_reg In powerNV, send intrs to HV |
| * ->notify_disable False for NX windows |
| * ->intr_disable False for Fault Windows |
| * ->xtra_write False for NX windows |
| * ->notify_early NA for NX windows |
| * ->rsvd_txbuf_count NA for Rx windows |
| * ->lpid, ->pid, ->tid NA for Rx windows |
| */ |
| |
| memset(winctx, 0, sizeof(struct vas_winctx)); |
| |
| winctx->rx_fifo = rxattr->rx_fifo; |
| winctx->rx_fifo_size = rxattr->rx_fifo_size; |
| winctx->wcreds_max = rxwin->vas_win.wcreds_max; |
| winctx->pin_win = rxattr->pin_win; |
| |
| winctx->nx_win = rxattr->nx_win; |
| winctx->fault_win = rxattr->fault_win; |
| winctx->user_win = rxattr->user_win; |
| winctx->rej_no_credit = rxattr->rej_no_credit; |
| winctx->rx_word_mode = rxattr->rx_win_ord_mode; |
| winctx->tx_word_mode = rxattr->tx_win_ord_mode; |
| winctx->rx_wcred_mode = rxattr->rx_wcred_mode; |
| winctx->tx_wcred_mode = rxattr->tx_wcred_mode; |
| winctx->notify_early = rxattr->notify_early; |
| |
| if (winctx->nx_win) { |
| winctx->data_stamp = true; |
| winctx->intr_disable = true; |
| winctx->pin_win = true; |
| |
| WARN_ON_ONCE(winctx->fault_win); |
| WARN_ON_ONCE(!winctx->rx_word_mode); |
| WARN_ON_ONCE(!winctx->tx_word_mode); |
| WARN_ON_ONCE(winctx->notify_after_count); |
| } else if (winctx->fault_win) { |
| winctx->notify_disable = true; |
| } else if (winctx->user_win) { |
| /* |
| * Section 1.8.1 Low Latency Core-Core Wake up of |
| * the VAS workbook: |
| * |
| * - disable credit checks ([tr]x_wcred_mode = false) |
| * - disable FIFO writes |
| * - enable ASB_Notify, disable interrupt |
| */ |
| winctx->fifo_disable = true; |
| winctx->intr_disable = true; |
| winctx->rx_fifo = NULL; |
| } |
| |
| winctx->lnotify_lpid = rxattr->lnotify_lpid; |
| winctx->lnotify_pid = rxattr->lnotify_pid; |
| winctx->lnotify_tid = rxattr->lnotify_tid; |
| winctx->pswid = rxattr->pswid; |
| winctx->dma_type = VAS_DMA_TYPE_INJECT; |
| winctx->tc_mode = rxattr->tc_mode; |
| |
| winctx->min_scope = VAS_SCOPE_LOCAL; |
| winctx->max_scope = VAS_SCOPE_VECTORED_GROUP; |
| if (rxwin->vinst->virq) |
| winctx->irq_port = rxwin->vinst->irq_port; |
| } |
| |
| static bool rx_win_args_valid(enum vas_cop_type cop, |
| struct vas_rx_win_attr *attr) |
| { |
| pr_debug("Rxattr: fault %d, notify %d, intr %d, early %d, fifo %d\n", |
| attr->fault_win, attr->notify_disable, |
| attr->intr_disable, attr->notify_early, |
| attr->rx_fifo_size); |
| |
| if (cop >= VAS_COP_TYPE_MAX) |
| return false; |
| |
| if (cop != VAS_COP_TYPE_FTW && |
| attr->rx_fifo_size < VAS_RX_FIFO_SIZE_MIN) |
| return false; |
| |
| if (attr->rx_fifo_size > VAS_RX_FIFO_SIZE_MAX) |
| return false; |
| |
| if (!attr->wcreds_max) |
| return false; |
| |
| if (attr->nx_win) { |
| /* cannot be fault or user window if it is nx */ |
| if (attr->fault_win || attr->user_win) |
| return false; |
| /* |
| * Section 3.1.4.32: NX Windows must not disable notification, |
| * and must not enable interrupts or early notification. |
| */ |
| if (attr->notify_disable || !attr->intr_disable || |
| attr->notify_early) |
| return false; |
| } else if (attr->fault_win) { |
| /* cannot be both fault and user window */ |
| if (attr->user_win) |
| return false; |
| |
| /* |
| * Section 3.1.4.32: Fault windows must disable notification |
| * but not interrupts. |
| */ |
| if (!attr->notify_disable || attr->intr_disable) |
| return false; |
| |
| } else if (attr->user_win) { |
| /* |
| * User receive windows are only for fast-thread-wakeup |
| * (FTW). They don't need a FIFO and must disable interrupts |
| */ |
| if (attr->rx_fifo || attr->rx_fifo_size || !attr->intr_disable) |
| return false; |
| } else { |
| /* Rx window must be one of NX or Fault or User window. */ |
| return false; |
| } |
| |
| return true; |
| } |
| |
| void vas_init_rx_win_attr(struct vas_rx_win_attr *rxattr, enum vas_cop_type cop) |
| { |
| memset(rxattr, 0, sizeof(*rxattr)); |
| |
| if (cop == VAS_COP_TYPE_842 || cop == VAS_COP_TYPE_842_HIPRI || |
| cop == VAS_COP_TYPE_GZIP || cop == VAS_COP_TYPE_GZIP_HIPRI) { |
| rxattr->pin_win = true; |
| rxattr->nx_win = true; |
| rxattr->fault_win = false; |
| rxattr->intr_disable = true; |
| rxattr->rx_wcred_mode = true; |
| rxattr->tx_wcred_mode = true; |
| rxattr->rx_win_ord_mode = true; |
| rxattr->tx_win_ord_mode = true; |
| } else if (cop == VAS_COP_TYPE_FAULT) { |
| rxattr->pin_win = true; |
| rxattr->fault_win = true; |
| rxattr->notify_disable = true; |
| rxattr->rx_wcred_mode = true; |
| rxattr->rx_win_ord_mode = true; |
| rxattr->rej_no_credit = true; |
| rxattr->tc_mode = VAS_THRESH_DISABLED; |
| } else if (cop == VAS_COP_TYPE_FTW) { |
| rxattr->user_win = true; |
| rxattr->intr_disable = true; |
| |
| /* |
| * As noted in the VAS Workbook we disable credit checks. |
| * If we enable credit checks in the future, we must also |
| * implement a mechanism to return the user credits or new |
| * paste operations will fail. |
| */ |
| } |
| } |
| EXPORT_SYMBOL_GPL(vas_init_rx_win_attr); |
| |
| struct vas_window *vas_rx_win_open(int vasid, enum vas_cop_type cop, |
| struct vas_rx_win_attr *rxattr) |
| { |
| struct pnv_vas_window *rxwin; |
| struct vas_winctx winctx; |
| struct vas_instance *vinst; |
| |
| trace_vas_rx_win_open(current, vasid, cop, rxattr); |
| |
| if (!rx_win_args_valid(cop, rxattr)) |
| return ERR_PTR(-EINVAL); |
| |
| vinst = find_vas_instance(vasid); |
| if (!vinst) { |
| pr_devel("vasid %d not found!\n", vasid); |
| return ERR_PTR(-EINVAL); |
| } |
| pr_devel("Found instance %d\n", vasid); |
| |
| rxwin = vas_window_alloc(vinst); |
| if (IS_ERR(rxwin)) { |
| pr_devel("Unable to allocate memory for Rx window\n"); |
| return (struct vas_window *)rxwin; |
| } |
| |
| rxwin->tx_win = false; |
| rxwin->nx_win = rxattr->nx_win; |
| rxwin->user_win = rxattr->user_win; |
| rxwin->vas_win.cop = cop; |
| rxwin->vas_win.wcreds_max = rxattr->wcreds_max; |
| |
| init_winctx_for_rxwin(rxwin, rxattr, &winctx); |
| init_winctx_regs(rxwin, &winctx); |
| |
| set_vinst_win(vinst, rxwin); |
| |
| return &rxwin->vas_win; |
| } |
| EXPORT_SYMBOL_GPL(vas_rx_win_open); |
| |
| void vas_init_tx_win_attr(struct vas_tx_win_attr *txattr, enum vas_cop_type cop) |
| { |
| memset(txattr, 0, sizeof(*txattr)); |
| |
| if (cop == VAS_COP_TYPE_842 || cop == VAS_COP_TYPE_842_HIPRI || |
| cop == VAS_COP_TYPE_GZIP || cop == VAS_COP_TYPE_GZIP_HIPRI) { |
| txattr->rej_no_credit = false; |
| txattr->rx_wcred_mode = true; |
| txattr->tx_wcred_mode = true; |
| txattr->rx_win_ord_mode = true; |
| txattr->tx_win_ord_mode = true; |
| } else if (cop == VAS_COP_TYPE_FTW) { |
| txattr->user_win = true; |
| } |
| } |
| EXPORT_SYMBOL_GPL(vas_init_tx_win_attr); |
| |
| static void init_winctx_for_txwin(struct pnv_vas_window *txwin, |
| struct vas_tx_win_attr *txattr, |
| struct vas_winctx *winctx) |
| { |
| /* |
| * We first zero all fields and only set non-zero ones. Following |
| * are some fields set to 0/false for the stated reason: |
| * |
| * ->notify_os_intr_reg In powernv, send intrs to HV |
| * ->rsvd_txbuf_count Not supported yet. |
| * ->notify_disable False for NX windows |
| * ->xtra_write False for NX windows |
| * ->notify_early NA for NX windows |
| * ->lnotify_lpid NA for Tx windows |
| * ->lnotify_pid NA for Tx windows |
| * ->lnotify_tid NA for Tx windows |
| * ->tx_win_cred_mode Ignore for now for NX windows |
| * ->rx_win_cred_mode Ignore for now for NX windows |
| */ |
| memset(winctx, 0, sizeof(struct vas_winctx)); |
| |
| winctx->wcreds_max = txwin->vas_win.wcreds_max; |
| |
| winctx->user_win = txattr->user_win; |
| winctx->nx_win = txwin->rxwin->nx_win; |
| winctx->pin_win = txattr->pin_win; |
| winctx->rej_no_credit = txattr->rej_no_credit; |
| winctx->rsvd_txbuf_enable = txattr->rsvd_txbuf_enable; |
| |
| winctx->rx_wcred_mode = txattr->rx_wcred_mode; |
| winctx->tx_wcred_mode = txattr->tx_wcred_mode; |
| winctx->rx_word_mode = txattr->rx_win_ord_mode; |
| winctx->tx_word_mode = txattr->tx_win_ord_mode; |
| winctx->rsvd_txbuf_count = txattr->rsvd_txbuf_count; |
| |
| winctx->intr_disable = true; |
| if (winctx->nx_win) |
| winctx->data_stamp = true; |
| |
| winctx->lpid = txattr->lpid; |
| winctx->pidr = txattr->pidr; |
| winctx->rx_win_id = txwin->rxwin->vas_win.winid; |
| /* |
| * IRQ and fault window setup is successful. Set fault window |
| * for the send window so that ready to handle faults. |
| */ |
| if (txwin->vinst->virq) |
| winctx->fault_win_id = txwin->vinst->fault_win->vas_win.winid; |
| |
| winctx->dma_type = VAS_DMA_TYPE_INJECT; |
| winctx->tc_mode = txattr->tc_mode; |
| winctx->min_scope = VAS_SCOPE_LOCAL; |
| winctx->max_scope = VAS_SCOPE_VECTORED_GROUP; |
| if (txwin->vinst->virq) |
| winctx->irq_port = txwin->vinst->irq_port; |
| |
| winctx->pswid = txattr->pswid ? txattr->pswid : |
| encode_pswid(txwin->vinst->vas_id, |
| txwin->vas_win.winid); |
| } |
| |
| static bool tx_win_args_valid(enum vas_cop_type cop, |
| struct vas_tx_win_attr *attr) |
| { |
| if (attr->tc_mode != VAS_THRESH_DISABLED) |
| return false; |
| |
| if (cop > VAS_COP_TYPE_MAX) |
| return false; |
| |
| if (attr->wcreds_max > VAS_TX_WCREDS_MAX) |
| return false; |
| |
| if (attr->user_win) { |
| if (attr->rsvd_txbuf_count) |
| return false; |
| |
| if (cop != VAS_COP_TYPE_FTW && cop != VAS_COP_TYPE_GZIP && |
| cop != VAS_COP_TYPE_GZIP_HIPRI) |
| return false; |
| } |
| |
| return true; |
| } |
| |
| struct vas_window *vas_tx_win_open(int vasid, enum vas_cop_type cop, |
| struct vas_tx_win_attr *attr) |
| { |
| int rc; |
| struct pnv_vas_window *txwin; |
| struct pnv_vas_window *rxwin; |
| struct vas_winctx winctx; |
| struct vas_instance *vinst; |
| |
| trace_vas_tx_win_open(current, vasid, cop, attr); |
| |
| if (!tx_win_args_valid(cop, attr)) |
| return ERR_PTR(-EINVAL); |
| |
| /* |
| * If caller did not specify a vasid but specified the PSWID of a |
| * receive window (applicable only to FTW windows), use the vasid |
| * from that receive window. |
| */ |
| if (vasid == -1 && attr->pswid) |
| decode_pswid(attr->pswid, &vasid, NULL); |
| |
| vinst = find_vas_instance(vasid); |
| if (!vinst) { |
| pr_devel("vasid %d not found!\n", vasid); |
| return ERR_PTR(-EINVAL); |
| } |
| |
| rxwin = get_vinst_rxwin(vinst, cop, attr->pswid); |
| if (IS_ERR(rxwin)) { |
| pr_devel("No RxWin for vasid %d, cop %d\n", vasid, cop); |
| return (struct vas_window *)rxwin; |
| } |
| |
| txwin = vas_window_alloc(vinst); |
| if (IS_ERR(txwin)) { |
| rc = PTR_ERR(txwin); |
| goto put_rxwin; |
| } |
| |
| txwin->vas_win.cop = cop; |
| txwin->tx_win = 1; |
| txwin->rxwin = rxwin; |
| txwin->nx_win = txwin->rxwin->nx_win; |
| txwin->user_win = attr->user_win; |
| txwin->vas_win.wcreds_max = attr->wcreds_max ?: VAS_WCREDS_DEFAULT; |
| |
| init_winctx_for_txwin(txwin, attr, &winctx); |
| |
| init_winctx_regs(txwin, &winctx); |
| |
| /* |
| * If its a kernel send window, map the window address into the |
| * kernel's address space. For user windows, user must issue an |
| * mmap() to map the window into their address space. |
| * |
| * NOTE: If kernel ever resubmits a user CRB after handling a page |
| * fault, we will need to map this into kernel as well. |
| */ |
| if (!txwin->user_win) { |
| txwin->paste_kaddr = map_paste_region(txwin); |
| if (IS_ERR(txwin->paste_kaddr)) { |
| rc = PTR_ERR(txwin->paste_kaddr); |
| goto free_window; |
| } |
| } else { |
| /* |
| * Interrupt hanlder or fault window setup failed. Means |
| * NX can not generate fault for page fault. So not |
| * opening for user space tx window. |
| */ |
| if (!vinst->virq) { |
| rc = -ENODEV; |
| goto free_window; |
| } |
| rc = get_vas_user_win_ref(&txwin->vas_win.task_ref); |
| if (rc) |
| goto free_window; |
| |
| vas_user_win_add_mm_context(&txwin->vas_win.task_ref); |
| } |
| |
| set_vinst_win(vinst, txwin); |
| |
| return &txwin->vas_win; |
| |
| free_window: |
| vas_window_free(txwin); |
| |
| put_rxwin: |
| put_rx_win(rxwin); |
| return ERR_PTR(rc); |
| |
| } |
| EXPORT_SYMBOL_GPL(vas_tx_win_open); |
| |
| int vas_copy_crb(void *crb, int offset) |
| { |
| return vas_copy(crb, offset); |
| } |
| EXPORT_SYMBOL_GPL(vas_copy_crb); |
| |
| #define RMA_LSMP_REPORT_ENABLE PPC_BIT(53) |
| int vas_paste_crb(struct vas_window *vwin, int offset, bool re) |
| { |
| struct pnv_vas_window *txwin; |
| int rc; |
| void *addr; |
| uint64_t val; |
| |
| txwin = container_of(vwin, struct pnv_vas_window, vas_win); |
| trace_vas_paste_crb(current, txwin); |
| |
| /* |
| * Only NX windows are supported for now and hardware assumes |
| * report-enable flag is set for NX windows. Ensure software |
| * complies too. |
| */ |
| WARN_ON_ONCE(txwin->nx_win && !re); |
| |
| addr = txwin->paste_kaddr; |
| if (re) { |
| /* |
| * Set the REPORT_ENABLE bit (equivalent to writing |
| * to 1K offset of the paste address) |
| */ |
| val = SET_FIELD(RMA_LSMP_REPORT_ENABLE, 0ULL, 1); |
| addr += val; |
| } |
| |
| /* |
| * Map the raw CR value from vas_paste() to an error code (there |
| * is just pass or fail for now though). |
| */ |
| rc = vas_paste(addr, offset); |
| if (rc == 2) |
| rc = 0; |
| else |
| rc = -EINVAL; |
| |
| pr_debug("Txwin #%d: Msg count %llu\n", txwin->vas_win.winid, |
| read_hvwc_reg(txwin, VREG(LRFIFO_PUSH))); |
| |
| return rc; |
| } |
| EXPORT_SYMBOL_GPL(vas_paste_crb); |
| |
| /* |
| * If credit checking is enabled for this window, poll for the return |
| * of window credits (i.e for NX engines to process any outstanding CRBs). |
| * Since NX-842 waits for the CRBs to be processed before closing the |
| * window, we should not have to wait for too long. |
| * |
| * TODO: We retry in 10ms intervals now. We could/should probably peek at |
| * the VAS_LRFIFO_PUSH_OFFSET register to get an estimate of pending |
| * CRBs on the FIFO and compute the delay dynamically on each retry. |
| * But that is not really needed until we support NX-GZIP access from |
| * user space. (NX-842 driver waits for CSB and Fast thread-wakeup |
| * doesn't use credit checking). |
| */ |
| static void poll_window_credits(struct pnv_vas_window *window) |
| { |
| u64 val; |
| int creds, mode; |
| int count = 0; |
| |
| val = read_hvwc_reg(window, VREG(WINCTL)); |
| if (window->tx_win) |
| mode = GET_FIELD(VAS_WINCTL_TX_WCRED_MODE, val); |
| else |
| mode = GET_FIELD(VAS_WINCTL_RX_WCRED_MODE, val); |
| |
| if (!mode) |
| return; |
| retry: |
| if (window->tx_win) { |
| val = read_hvwc_reg(window, VREG(TX_WCRED)); |
| creds = GET_FIELD(VAS_TX_WCRED, val); |
| } else { |
| val = read_hvwc_reg(window, VREG(LRX_WCRED)); |
| creds = GET_FIELD(VAS_LRX_WCRED, val); |
| } |
| |
| /* |
| * Takes around few milliseconds to complete all pending requests |
| * and return credits. |
| * TODO: Scan fault FIFO and invalidate CRBs points to this window |
| * and issue CRB Kill to stop all pending requests. Need only |
| * if there is a bug in NX or fault handling in kernel. |
| */ |
| if (creds < window->vas_win.wcreds_max) { |
| val = 0; |
| set_current_state(TASK_UNINTERRUPTIBLE); |
| schedule_timeout(msecs_to_jiffies(10)); |
| count++; |
| /* |
| * Process can not close send window until all credits are |
| * returned. |
| */ |
| if (!(count % 1000)) |
| pr_warn_ratelimited("VAS: pid %d stuck. Waiting for credits returned for Window(%d). creds %d, Retries %d\n", |
| vas_window_pid(&window->vas_win), |
| window->vas_win.winid, |
| creds, count); |
| |
| goto retry; |
| } |
| } |
| |
| /* |
| * Wait for the window to go to "not-busy" state. It should only take a |
| * short time to queue a CRB, so window should not be busy for too long. |
| * Trying 5ms intervals. |
| */ |
| static void poll_window_busy_state(struct pnv_vas_window *window) |
| { |
| int busy; |
| u64 val; |
| int count = 0; |
| |
| retry: |
| val = read_hvwc_reg(window, VREG(WIN_STATUS)); |
| busy = GET_FIELD(VAS_WIN_BUSY, val); |
| if (busy) { |
| val = 0; |
| set_current_state(TASK_UNINTERRUPTIBLE); |
| schedule_timeout(msecs_to_jiffies(10)); |
| count++; |
| /* |
| * Takes around few milliseconds to process all pending |
| * requests. |
| */ |
| if (!(count % 1000)) |
| pr_warn_ratelimited("VAS: pid %d stuck. Window (ID=%d) is in busy state. Retries %d\n", |
| vas_window_pid(&window->vas_win), |
| window->vas_win.winid, count); |
| |
| goto retry; |
| } |
| } |
| |
| /* |
| * Have the hardware cast a window out of cache and wait for it to |
| * be completed. |
| * |
| * NOTE: It can take a relatively long time to cast the window context |
| * out of the cache. It is not strictly necessary to cast out if: |
| * |
| * - we clear the "Pin Window" bit (so hardware is free to evict) |
| * |
| * - we re-initialize the window context when it is reassigned. |
| * |
| * We do the former in vas_win_close() and latter in vas_win_open(). |
| * So, ignoring the cast-out for now. We can add it as needed. If |
| * casting out becomes necessary we should consider offloading the |
| * job to a worker thread, so the window close can proceed quickly. |
| */ |
| static void poll_window_castout(struct pnv_vas_window *window) |
| { |
| /* stub for now */ |
| } |
| |
| /* |
| * Unpin and close a window so no new requests are accepted and the |
| * hardware can evict this window from cache if necessary. |
| */ |
| static void unpin_close_window(struct pnv_vas_window *window) |
| { |
| u64 val; |
| |
| val = read_hvwc_reg(window, VREG(WINCTL)); |
| val = SET_FIELD(VAS_WINCTL_PIN, val, 0); |
| val = SET_FIELD(VAS_WINCTL_OPEN, val, 0); |
| write_hvwc_reg(window, VREG(WINCTL), val); |
| } |
| |
| /* |
| * Close a window. |
| * |
| * See Section 1.12.1 of VAS workbook v1.05 for details on closing window: |
| * - Disable new paste operations (unmap paste address) |
| * - Poll for the "Window Busy" bit to be cleared |
| * - Clear the Open/Enable bit for the Window. |
| * - Poll for return of window Credits (implies FIFO empty for Rx win?) |
| * - Unpin and cast window context out of cache |
| * |
| * Besides the hardware, kernel has some bookkeeping of course. |
| */ |
| int vas_win_close(struct vas_window *vwin) |
| { |
| struct pnv_vas_window *window; |
| |
| if (!vwin) |
| return 0; |
| |
| window = container_of(vwin, struct pnv_vas_window, vas_win); |
| |
| if (!window->tx_win && atomic_read(&window->num_txwins) != 0) { |
| pr_devel("Attempting to close an active Rx window!\n"); |
| WARN_ON_ONCE(1); |
| return -EBUSY; |
| } |
| |
| unmap_paste_region(window); |
| |
| poll_window_busy_state(window); |
| |
| unpin_close_window(window); |
| |
| poll_window_credits(window); |
| |
| clear_vinst_win(window); |
| |
| poll_window_castout(window); |
| |
| /* if send window, drop reference to matching receive window */ |
| if (window->tx_win) { |
| if (window->user_win) { |
| put_vas_user_win_ref(&vwin->task_ref); |
| mm_context_remove_vas_window(vwin->task_ref.mm); |
| } |
| put_rx_win(window->rxwin); |
| } |
| |
| vas_window_free(window); |
| |
| return 0; |
| } |
| EXPORT_SYMBOL_GPL(vas_win_close); |
| |
| /* |
| * Return credit for the given window. |
| * Send windows and fault window uses credit mechanism as follows: |
| * |
| * Send windows: |
| * - The default number of credits available for each send window is |
| * 1024. It means 1024 requests can be issued asynchronously at the |
| * same time. If the credit is not available, that request will be |
| * returned with RMA_Busy. |
| * - One credit is taken when NX request is issued. |
| * - This credit is returned after NX processed that request. |
| * - If NX encounters translation error, kernel will return the |
| * credit on the specific send window after processing the fault CRB. |
| * |
| * Fault window: |
| * - The total number credits available is FIFO_SIZE/CRB_SIZE. |
| * Means 4MB/128 in the current implementation. If credit is not |
| * available, RMA_Reject is returned. |
| * - A credit is taken when NX pastes CRB in fault FIFO. |
| * - The kernel with return credit on fault window after reading entry |
| * from fault FIFO. |
| */ |
| void vas_return_credit(struct pnv_vas_window *window, bool tx) |
| { |
| uint64_t val; |
| |
| val = 0ULL; |
| if (tx) { /* send window */ |
| val = SET_FIELD(VAS_TX_WCRED, val, 1); |
| write_hvwc_reg(window, VREG(TX_WCRED_ADDER), val); |
| } else { |
| val = SET_FIELD(VAS_LRX_WCRED, val, 1); |
| write_hvwc_reg(window, VREG(LRX_WCRED_ADDER), val); |
| } |
| } |
| |
| struct pnv_vas_window *vas_pswid_to_window(struct vas_instance *vinst, |
| uint32_t pswid) |
| { |
| struct pnv_vas_window *window; |
| int winid; |
| |
| if (!pswid) { |
| pr_devel("%s: called for pswid 0!\n", __func__); |
| return ERR_PTR(-ESRCH); |
| } |
| |
| decode_pswid(pswid, NULL, &winid); |
| |
| if (winid >= VAS_WINDOWS_PER_CHIP) |
| return ERR_PTR(-ESRCH); |
| |
| /* |
| * If application closes the window before the hardware |
| * returns the fault CRB, we should wait in vas_win_close() |
| * for the pending requests. so the window must be active |
| * and the process alive. |
| * |
| * If its a kernel process, we should not get any faults and |
| * should not get here. |
| */ |
| window = vinst->windows[winid]; |
| |
| if (!window) { |
| pr_err("PSWID decode: Could not find window for winid %d pswid %d vinst 0x%p\n", |
| winid, pswid, vinst); |
| return NULL; |
| } |
| |
| /* |
| * Do some sanity checks on the decoded window. Window should be |
| * NX GZIP user send window. FTW windows should not incur faults |
| * since their CRBs are ignored (not queued on FIFO or processed |
| * by NX). |
| */ |
| if (!window->tx_win || !window->user_win || !window->nx_win || |
| window->vas_win.cop == VAS_COP_TYPE_FAULT || |
| window->vas_win.cop == VAS_COP_TYPE_FTW) { |
| pr_err("PSWID decode: id %d, tx %d, user %d, nx %d, cop %d\n", |
| winid, window->tx_win, window->user_win, |
| window->nx_win, window->vas_win.cop); |
| WARN_ON(1); |
| } |
| |
| return window; |
| } |
| |
| static struct vas_window *vas_user_win_open(int vas_id, u64 flags, |
| enum vas_cop_type cop_type) |
| { |
| struct vas_tx_win_attr txattr = {}; |
| |
| vas_init_tx_win_attr(&txattr, cop_type); |
| |
| txattr.lpid = mfspr(SPRN_LPID); |
| txattr.pidr = mfspr(SPRN_PID); |
| txattr.user_win = true; |
| txattr.rsvd_txbuf_count = false; |
| txattr.pswid = false; |
| |
| pr_devel("Pid %d: Opening txwin, PIDR %ld\n", txattr.pidr, |
| mfspr(SPRN_PID)); |
| |
| return vas_tx_win_open(vas_id, cop_type, &txattr); |
| } |
| |
| static u64 vas_user_win_paste_addr(struct vas_window *txwin) |
| { |
| struct pnv_vas_window *win; |
| u64 paste_addr; |
| |
| win = container_of(txwin, struct pnv_vas_window, vas_win); |
| vas_win_paste_addr(win, &paste_addr, NULL); |
| |
| return paste_addr; |
| } |
| |
| static int vas_user_win_close(struct vas_window *txwin) |
| { |
| vas_win_close(txwin); |
| |
| return 0; |
| } |
| |
| static const struct vas_user_win_ops vops = { |
| .open_win = vas_user_win_open, |
| .paste_addr = vas_user_win_paste_addr, |
| .close_win = vas_user_win_close, |
| }; |
| |
| /* |
| * Supporting only nx-gzip coprocessor type now, but this API code |
| * extended to other coprocessor types later. |
| */ |
| int vas_register_api_powernv(struct module *mod, enum vas_cop_type cop_type, |
| const char *name) |
| { |
| |
| return vas_register_coproc_api(mod, cop_type, name, &vops); |
| } |
| EXPORT_SYMBOL_GPL(vas_register_api_powernv); |
| |
| void vas_unregister_api_powernv(void) |
| { |
| vas_unregister_coproc_api(); |
| } |
| EXPORT_SYMBOL_GPL(vas_unregister_api_powernv); |