|  | /* | 
|  | *  ICSWX and ACOP Management | 
|  | * | 
|  | *  Copyright (C) 2011 Anton Blanchard, IBM Corp. <anton@samba.org> | 
|  | * | 
|  | *  This program is free software; you can redistribute it and/or | 
|  | *  modify it under the terms of the GNU General Public License | 
|  | *  as published by the Free Software Foundation; either version | 
|  | *  2 of the License, or (at your option) any later version. | 
|  | * | 
|  | */ | 
|  |  | 
|  | #include <linux/sched.h> | 
|  | #include <linux/kernel.h> | 
|  | #include <linux/errno.h> | 
|  | #include <linux/types.h> | 
|  | #include <linux/mm.h> | 
|  | #include <linux/spinlock.h> | 
|  | #include <linux/module.h> | 
|  | #include <linux/uaccess.h> | 
|  |  | 
|  | #include "icswx.h" | 
|  |  | 
|  | /* | 
|  | * The processor and its L2 cache cause the icswx instruction to | 
|  | * generate a COP_REQ transaction on PowerBus. The transaction has no | 
|  | * address, and the processor does not perform an MMU access to | 
|  | * authenticate the transaction. The command portion of the PowerBus | 
|  | * COP_REQ transaction includes the LPAR_ID (LPID) and the coprocessor | 
|  | * Process ID (PID), which the coprocessor compares to the authorized | 
|  | * LPID and PID held in the coprocessor, to determine if the process | 
|  | * is authorized to generate the transaction.  The data of the COP_REQ | 
|  | * transaction is 128-byte or less in size and is placed in cacheable | 
|  | * memory on a 128-byte cache line boundary. | 
|  | * | 
|  | * The task to use a coprocessor should use use_cop() to mark the use | 
|  | * of the Coprocessor Type (CT) and context switching. On a server | 
|  | * class processor, the PID register is used only for coprocessor | 
|  | * management + * and so a coprocessor PID is allocated before | 
|  | * executing icswx + * instruction. Drop_cop() is used to free the | 
|  | * coprocessor PID. | 
|  | * | 
|  | * Example: | 
|  | * Host Fabric Interface (HFI) is a PowerPC network coprocessor. | 
|  | * Each HFI have multiple windows. Each HFI window serves as a | 
|  | * network device sending to and receiving from HFI network. | 
|  | * HFI immediate send function uses icswx instruction. The immediate | 
|  | * send function allows small (single cache-line) packets be sent | 
|  | * without using the regular HFI send FIFO and doorbell, which are | 
|  | * much slower than immediate send. | 
|  | * | 
|  | * For each task intending to use HFI immediate send, the HFI driver | 
|  | * calls use_cop() to obtain a coprocessor PID for the task. | 
|  | * The HFI driver then allocate a free HFI window and save the | 
|  | * coprocessor PID to the HFI window to allow the task to use the | 
|  | * HFI window. | 
|  | * | 
|  | * The HFI driver repeatedly creates immediate send packets and | 
|  | * issues icswx instruction to send data through the HFI window. | 
|  | * The HFI compares the coprocessor PID in the CPU PID register | 
|  | * to the PID held in the HFI window to determine if the transaction | 
|  | * is allowed. | 
|  | * | 
|  | * When the task to release the HFI window, the HFI driver calls | 
|  | * drop_cop() to release the coprocessor PID. | 
|  | */ | 
|  |  | 
|  | void switch_cop(struct mm_struct *next) | 
|  | { | 
|  | #ifdef CONFIG_PPC_ICSWX_PID | 
|  | mtspr(SPRN_PID, next->context.cop_pid); | 
|  | #endif | 
|  | mtspr(SPRN_ACOP, next->context.acop); | 
|  | } | 
|  |  | 
|  | /** | 
|  | * Start using a coprocessor. | 
|  | * @acop: mask of coprocessor to be used. | 
|  | * @mm: The mm the coprocessor to associate with. Most likely current mm. | 
|  | * | 
|  | * Return a positive PID if successful. Negative errno otherwise. | 
|  | * The returned PID will be fed to the coprocessor to determine if an | 
|  | * icswx transaction is authenticated. | 
|  | */ | 
|  | int use_cop(unsigned long acop, struct mm_struct *mm) | 
|  | { | 
|  | int ret; | 
|  |  | 
|  | if (!cpu_has_feature(CPU_FTR_ICSWX)) | 
|  | return -ENODEV; | 
|  |  | 
|  | if (!mm || !acop) | 
|  | return -EINVAL; | 
|  |  | 
|  | /* The page_table_lock ensures mm_users won't change under us */ | 
|  | spin_lock(&mm->page_table_lock); | 
|  | spin_lock(mm->context.cop_lockp); | 
|  |  | 
|  | ret = get_cop_pid(mm); | 
|  | if (ret < 0) | 
|  | goto out; | 
|  |  | 
|  | /* update acop */ | 
|  | mm->context.acop |= acop; | 
|  |  | 
|  | sync_cop(mm); | 
|  |  | 
|  | /* | 
|  | * If this is a threaded process then there might be other threads | 
|  | * running. We need to send an IPI to force them to pick up any | 
|  | * change in PID and ACOP. | 
|  | */ | 
|  | if (atomic_read(&mm->mm_users) > 1) | 
|  | smp_call_function(sync_cop, mm, 1); | 
|  |  | 
|  | out: | 
|  | spin_unlock(mm->context.cop_lockp); | 
|  | spin_unlock(&mm->page_table_lock); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  | EXPORT_SYMBOL_GPL(use_cop); | 
|  |  | 
|  | /** | 
|  | * Stop using a coprocessor. | 
|  | * @acop: mask of coprocessor to be stopped. | 
|  | * @mm: The mm the coprocessor associated with. | 
|  | */ | 
|  | void drop_cop(unsigned long acop, struct mm_struct *mm) | 
|  | { | 
|  | int free_pid; | 
|  |  | 
|  | if (!cpu_has_feature(CPU_FTR_ICSWX)) | 
|  | return; | 
|  |  | 
|  | if (WARN_ON_ONCE(!mm)) | 
|  | return; | 
|  |  | 
|  | /* The page_table_lock ensures mm_users won't change under us */ | 
|  | spin_lock(&mm->page_table_lock); | 
|  | spin_lock(mm->context.cop_lockp); | 
|  |  | 
|  | mm->context.acop &= ~acop; | 
|  |  | 
|  | free_pid = disable_cop_pid(mm); | 
|  | sync_cop(mm); | 
|  |  | 
|  | /* | 
|  | * If this is a threaded process then there might be other threads | 
|  | * running. We need to send an IPI to force them to pick up any | 
|  | * change in PID and ACOP. | 
|  | */ | 
|  | if (atomic_read(&mm->mm_users) > 1) | 
|  | smp_call_function(sync_cop, mm, 1); | 
|  |  | 
|  | if (free_pid != COP_PID_NONE) | 
|  | free_cop_pid(free_pid); | 
|  |  | 
|  | spin_unlock(mm->context.cop_lockp); | 
|  | spin_unlock(&mm->page_table_lock); | 
|  | } | 
|  | EXPORT_SYMBOL_GPL(drop_cop); | 
|  |  | 
|  | static int acop_use_cop(int ct) | 
|  | { | 
|  | /* There is no alternate policy, yet */ | 
|  | return -1; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Get the instruction word at the NIP | 
|  | */ | 
|  | static u32 acop_get_inst(struct pt_regs *regs) | 
|  | { | 
|  | u32 inst; | 
|  | u32 __user *p; | 
|  |  | 
|  | p = (u32 __user *)regs->nip; | 
|  | if (!access_ok(VERIFY_READ, p, sizeof(*p))) | 
|  | return 0; | 
|  |  | 
|  | if (__get_user(inst, p)) | 
|  | return 0; | 
|  |  | 
|  | return inst; | 
|  | } | 
|  |  | 
|  | /** | 
|  | * @regs: regsiters at time of interrupt | 
|  | * @address: storage address | 
|  | * @error_code: Fault code, usually the DSISR or ESR depending on | 
|  | *		processor type | 
|  | * | 
|  | * Return 0 if we are able to resolve the data storage fault that | 
|  | * results from a CT miss in the ACOP register. | 
|  | */ | 
|  | int acop_handle_fault(struct pt_regs *regs, unsigned long address, | 
|  | unsigned long error_code) | 
|  | { | 
|  | int ct; | 
|  | u32 inst = 0; | 
|  |  | 
|  | if (!cpu_has_feature(CPU_FTR_ICSWX)) { | 
|  | pr_info("No coprocessors available"); | 
|  | _exception(SIGILL, regs, ILL_ILLOPN, address); | 
|  | } | 
|  |  | 
|  | if (!user_mode(regs)) { | 
|  | /* this could happen if the HV denies the | 
|  | * kernel access, for now we just die */ | 
|  | die("ICSWX from kernel failed", regs, SIGSEGV); | 
|  | } | 
|  |  | 
|  | /* Some implementations leave us a hint for the CT */ | 
|  | ct = ICSWX_GET_CT_HINT(error_code); | 
|  | if (ct < 0) { | 
|  | /* we have to peek at the instruction word to figure out CT */ | 
|  | u32 ccw; | 
|  | u32 rs; | 
|  |  | 
|  | inst = acop_get_inst(regs); | 
|  | if (inst == 0) | 
|  | return -1; | 
|  |  | 
|  | rs = (inst >> (31 - 10)) & 0x1f; | 
|  | ccw = regs->gpr[rs]; | 
|  | ct = (ccw >> 16) & 0x3f; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * We could be here because another thread has enabled acop | 
|  | * but the ACOP register has yet to be updated. | 
|  | * | 
|  | * This should have been taken care of by the IPI to sync all | 
|  | * the threads (see smp_call_function(sync_cop, mm, 1)), but | 
|  | * that could take forever if there are a significant amount | 
|  | * of threads. | 
|  | * | 
|  | * Given the number of threads on some of these systems, | 
|  | * perhaps this is the best way to sync ACOP rather than whack | 
|  | * every thread with an IPI. | 
|  | */ | 
|  | if ((acop_copro_type_bit(ct) & current->active_mm->context.acop) != 0) { | 
|  | sync_cop(current->active_mm); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* check for alternate policy */ | 
|  | if (!acop_use_cop(ct)) | 
|  | return 0; | 
|  |  | 
|  | /* at this point the CT is unknown to the system */ | 
|  | pr_warn("%s[%d]: Coprocessor %d is unavailable\n", | 
|  | current->comm, current->pid, ct); | 
|  |  | 
|  | /* get inst if we don't already have it */ | 
|  | if (inst == 0) { | 
|  | inst = acop_get_inst(regs); | 
|  | if (inst == 0) | 
|  | return -1; | 
|  | } | 
|  |  | 
|  | /* Check if the instruction is the "record form" */ | 
|  | if (inst & 1) { | 
|  | /* | 
|  | * the instruction is "record" form so we can reject | 
|  | * using CR0 | 
|  | */ | 
|  | regs->ccr &= ~(0xful << 28); | 
|  | regs->ccr |= ICSWX_RC_NOT_FOUND << 28; | 
|  |  | 
|  | /* Move on to the next instruction */ | 
|  | regs->nip += 4; | 
|  | } else { | 
|  | /* | 
|  | * There is no architected mechanism to report a bad | 
|  | * CT so we could either SIGILL or report nothing. | 
|  | * Since the non-record version should only bu used | 
|  | * for "hints" or "don't care" we should probably do | 
|  | * nothing.  However, I could see how some people | 
|  | * might want an SIGILL so it here if you want it. | 
|  | */ | 
|  | #ifdef CONFIG_PPC_ICSWX_USE_SIGILL | 
|  | _exception(SIGILL, regs, ILL_ILLOPN, address); | 
|  | #else | 
|  | regs->nip += 4; | 
|  | #endif | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  | EXPORT_SYMBOL_GPL(acop_handle_fault); |