| /* |
| * This file is subject to the terms and conditions of the GNU General Public |
| * License. See the file "COPYING" in the main directory of this archive |
| * for more details. |
| * |
| * KVM/MIPS TLB handling, this file is part of the Linux host kernel so that |
| * TLB handlers run from KSEG0 |
| * |
| * Copyright (C) 2012 MIPS Technologies, Inc. All rights reserved. |
| * Authors: Sanjay Lal <sanjayl@kymasys.com> |
| */ |
| |
| #include <linux/sched.h> |
| #include <linux/smp.h> |
| #include <linux/mm.h> |
| #include <linux/delay.h> |
| #include <linux/export.h> |
| #include <linux/kvm_host.h> |
| #include <linux/srcu.h> |
| |
| #include <asm/cpu.h> |
| #include <asm/bootinfo.h> |
| #include <asm/mipsregs.h> |
| #include <asm/mmu_context.h> |
| #include <asm/cacheflush.h> |
| #include <asm/tlb.h> |
| #include <asm/tlbdebug.h> |
| |
| #undef CONFIG_MIPS_MT |
| #include <asm/r4kcache.h> |
| #define CONFIG_MIPS_MT |
| |
| unsigned long GUESTID_MASK; |
| EXPORT_SYMBOL_GPL(GUESTID_MASK); |
| unsigned long GUESTID_FIRST_VERSION; |
| EXPORT_SYMBOL_GPL(GUESTID_FIRST_VERSION); |
| unsigned long GUESTID_VERSION_MASK; |
| EXPORT_SYMBOL_GPL(GUESTID_VERSION_MASK); |
| |
| static u32 kvm_mips_get_root_asid(struct kvm_vcpu *vcpu) |
| { |
| struct mm_struct *gpa_mm = &vcpu->kvm->arch.gpa_mm; |
| |
| if (cpu_has_guestid) |
| return 0; |
| else |
| return cpu_asid(smp_processor_id(), gpa_mm); |
| } |
| |
| static int _kvm_mips_host_tlb_inv(unsigned long entryhi) |
| { |
| int idx; |
| |
| write_c0_entryhi(entryhi); |
| mtc0_tlbw_hazard(); |
| |
| tlb_probe(); |
| tlb_probe_hazard(); |
| idx = read_c0_index(); |
| |
| BUG_ON(idx >= current_cpu_data.tlbsize); |
| |
| if (idx >= 0) { |
| write_c0_entryhi(UNIQUE_ENTRYHI(idx)); |
| write_c0_entrylo0(0); |
| write_c0_entrylo1(0); |
| mtc0_tlbw_hazard(); |
| |
| tlb_write_indexed(); |
| tlbw_use_hazard(); |
| } |
| |
| return idx; |
| } |
| |
| /* GuestID management */ |
| |
| /** |
| * clear_root_gid() - Set GuestCtl1.RID for normal root operation. |
| */ |
| static inline void clear_root_gid(void) |
| { |
| if (cpu_has_guestid) { |
| clear_c0_guestctl1(MIPS_GCTL1_RID); |
| mtc0_tlbw_hazard(); |
| } |
| } |
| |
| /** |
| * set_root_gid_to_guest_gid() - Set GuestCtl1.RID to match GuestCtl1.ID. |
| * |
| * Sets the root GuestID to match the current guest GuestID, for TLB operation |
| * on the GPA->RPA mappings in the root TLB. |
| * |
| * The caller must be sure to disable HTW while the root GID is set, and |
| * possibly longer if TLB registers are modified. |
| */ |
| static inline void set_root_gid_to_guest_gid(void) |
| { |
| unsigned int guestctl1; |
| |
| if (cpu_has_guestid) { |
| back_to_back_c0_hazard(); |
| guestctl1 = read_c0_guestctl1(); |
| guestctl1 = (guestctl1 & ~MIPS_GCTL1_RID) | |
| ((guestctl1 & MIPS_GCTL1_ID) >> MIPS_GCTL1_ID_SHIFT) |
| << MIPS_GCTL1_RID_SHIFT; |
| write_c0_guestctl1(guestctl1); |
| mtc0_tlbw_hazard(); |
| } |
| } |
| |
| int kvm_vz_host_tlb_inv(struct kvm_vcpu *vcpu, unsigned long va) |
| { |
| int idx; |
| unsigned long flags, old_entryhi; |
| |
| local_irq_save(flags); |
| htw_stop(); |
| |
| /* Set root GuestID for root probe and write of guest TLB entry */ |
| set_root_gid_to_guest_gid(); |
| |
| old_entryhi = read_c0_entryhi(); |
| |
| idx = _kvm_mips_host_tlb_inv((va & VPN2_MASK) | |
| kvm_mips_get_root_asid(vcpu)); |
| |
| write_c0_entryhi(old_entryhi); |
| clear_root_gid(); |
| mtc0_tlbw_hazard(); |
| |
| htw_start(); |
| local_irq_restore(flags); |
| |
| /* |
| * We don't want to get reserved instruction exceptions for missing tlb |
| * entries. |
| */ |
| if (cpu_has_vtag_icache) |
| flush_icache_all(); |
| |
| if (idx > 0) |
| kvm_debug("%s: Invalidated root entryhi %#lx @ idx %d\n", |
| __func__, (va & VPN2_MASK) | |
| kvm_mips_get_root_asid(vcpu), idx); |
| |
| return 0; |
| } |
| EXPORT_SYMBOL_GPL(kvm_vz_host_tlb_inv); |
| |
| /** |
| * kvm_vz_guest_tlb_lookup() - Lookup a guest VZ TLB mapping. |
| * @vcpu: KVM VCPU pointer. |
| * @gpa: Guest virtual address in a TLB mapped guest segment. |
| * @gpa: Pointer to output guest physical address it maps to. |
| * |
| * Converts a guest virtual address in a guest TLB mapped segment to a guest |
| * physical address, by probing the guest TLB. |
| * |
| * Returns: 0 if guest TLB mapping exists for @gva. *@gpa will have been |
| * written. |
| * -EFAULT if no guest TLB mapping exists for @gva. *@gpa may not |
| * have been written. |
| */ |
| int kvm_vz_guest_tlb_lookup(struct kvm_vcpu *vcpu, unsigned long gva, |
| unsigned long *gpa) |
| { |
| unsigned long o_entryhi, o_entrylo[2], o_pagemask; |
| unsigned int o_index; |
| unsigned long entrylo[2], pagemask, pagemaskbit, pa; |
| unsigned long flags; |
| int index; |
| |
| /* Probe the guest TLB for a mapping */ |
| local_irq_save(flags); |
| /* Set root GuestID for root probe of guest TLB entry */ |
| htw_stop(); |
| set_root_gid_to_guest_gid(); |
| |
| o_entryhi = read_gc0_entryhi(); |
| o_index = read_gc0_index(); |
| |
| write_gc0_entryhi((o_entryhi & 0x3ff) | (gva & ~0xfffl)); |
| mtc0_tlbw_hazard(); |
| guest_tlb_probe(); |
| tlb_probe_hazard(); |
| |
| index = read_gc0_index(); |
| if (index < 0) { |
| /* No match, fail */ |
| write_gc0_entryhi(o_entryhi); |
| write_gc0_index(o_index); |
| |
| clear_root_gid(); |
| htw_start(); |
| local_irq_restore(flags); |
| return -EFAULT; |
| } |
| |
| /* Match! read the TLB entry */ |
| o_entrylo[0] = read_gc0_entrylo0(); |
| o_entrylo[1] = read_gc0_entrylo1(); |
| o_pagemask = read_gc0_pagemask(); |
| |
| mtc0_tlbr_hazard(); |
| guest_tlb_read(); |
| tlb_read_hazard(); |
| |
| entrylo[0] = read_gc0_entrylo0(); |
| entrylo[1] = read_gc0_entrylo1(); |
| pagemask = ~read_gc0_pagemask() & ~0x1fffl; |
| |
| write_gc0_entryhi(o_entryhi); |
| write_gc0_index(o_index); |
| write_gc0_entrylo0(o_entrylo[0]); |
| write_gc0_entrylo1(o_entrylo[1]); |
| write_gc0_pagemask(o_pagemask); |
| |
| clear_root_gid(); |
| htw_start(); |
| local_irq_restore(flags); |
| |
| /* Select one of the EntryLo values and interpret the GPA */ |
| pagemaskbit = (pagemask ^ (pagemask & (pagemask - 1))) >> 1; |
| pa = entrylo[!!(gva & pagemaskbit)]; |
| |
| /* |
| * TLB entry may have become invalid since TLB probe if physical FTLB |
| * entries are shared between threads (e.g. I6400). |
| */ |
| if (!(pa & ENTRYLO_V)) |
| return -EFAULT; |
| |
| /* |
| * Note, this doesn't take guest MIPS32 XPA into account, where PFN is |
| * split with XI/RI in the middle. |
| */ |
| pa = (pa << 6) & ~0xfffl; |
| pa |= gva & ~(pagemask | pagemaskbit); |
| |
| *gpa = pa; |
| return 0; |
| } |
| EXPORT_SYMBOL_GPL(kvm_vz_guest_tlb_lookup); |
| |
| /** |
| * kvm_vz_local_flush_roottlb_all_guests() - Flush all root TLB entries for |
| * guests. |
| * |
| * Invalidate all entries in root tlb which are GPA mappings. |
| */ |
| void kvm_vz_local_flush_roottlb_all_guests(void) |
| { |
| unsigned long flags; |
| unsigned long old_entryhi, old_pagemask, old_guestctl1; |
| int entry; |
| |
| if (WARN_ON(!cpu_has_guestid)) |
| return; |
| |
| local_irq_save(flags); |
| htw_stop(); |
| |
| /* TLBR may clobber EntryHi.ASID, PageMask, and GuestCtl1.RID */ |
| old_entryhi = read_c0_entryhi(); |
| old_pagemask = read_c0_pagemask(); |
| old_guestctl1 = read_c0_guestctl1(); |
| |
| /* |
| * Invalidate guest entries in root TLB while leaving root entries |
| * intact when possible. |
| */ |
| for (entry = 0; entry < current_cpu_data.tlbsize; entry++) { |
| write_c0_index(entry); |
| mtc0_tlbw_hazard(); |
| tlb_read(); |
| tlb_read_hazard(); |
| |
| /* Don't invalidate non-guest (RVA) mappings in the root TLB */ |
| if (!(read_c0_guestctl1() & MIPS_GCTL1_RID)) |
| continue; |
| |
| /* Make sure all entries differ. */ |
| write_c0_entryhi(UNIQUE_ENTRYHI(entry)); |
| write_c0_entrylo0(0); |
| write_c0_entrylo1(0); |
| write_c0_guestctl1(0); |
| mtc0_tlbw_hazard(); |
| tlb_write_indexed(); |
| } |
| |
| write_c0_entryhi(old_entryhi); |
| write_c0_pagemask(old_pagemask); |
| write_c0_guestctl1(old_guestctl1); |
| tlbw_use_hazard(); |
| |
| htw_start(); |
| local_irq_restore(flags); |
| } |
| EXPORT_SYMBOL_GPL(kvm_vz_local_flush_roottlb_all_guests); |
| |
| /** |
| * kvm_vz_local_flush_guesttlb_all() - Flush all guest TLB entries. |
| * |
| * Invalidate all entries in guest tlb irrespective of guestid. |
| */ |
| void kvm_vz_local_flush_guesttlb_all(void) |
| { |
| unsigned long flags; |
| unsigned long old_index; |
| unsigned long old_entryhi; |
| unsigned long old_entrylo[2]; |
| unsigned long old_pagemask; |
| int entry; |
| u64 cvmmemctl2 = 0; |
| |
| local_irq_save(flags); |
| |
| /* Preserve all clobbered guest registers */ |
| old_index = read_gc0_index(); |
| old_entryhi = read_gc0_entryhi(); |
| old_entrylo[0] = read_gc0_entrylo0(); |
| old_entrylo[1] = read_gc0_entrylo1(); |
| old_pagemask = read_gc0_pagemask(); |
| |
| switch (current_cpu_type()) { |
| case CPU_CAVIUM_OCTEON3: |
| /* Inhibit machine check due to multiple matching TLB entries */ |
| cvmmemctl2 = read_c0_cvmmemctl2(); |
| cvmmemctl2 |= CVMMEMCTL2_INHIBITTS; |
| write_c0_cvmmemctl2(cvmmemctl2); |
| break; |
| } |
| |
| /* Invalidate guest entries in guest TLB */ |
| write_gc0_entrylo0(0); |
| write_gc0_entrylo1(0); |
| write_gc0_pagemask(0); |
| for (entry = 0; entry < current_cpu_data.guest.tlbsize; entry++) { |
| /* Make sure all entries differ. */ |
| write_gc0_index(entry); |
| write_gc0_entryhi(UNIQUE_GUEST_ENTRYHI(entry)); |
| mtc0_tlbw_hazard(); |
| guest_tlb_write_indexed(); |
| } |
| |
| if (cvmmemctl2) { |
| cvmmemctl2 &= ~CVMMEMCTL2_INHIBITTS; |
| write_c0_cvmmemctl2(cvmmemctl2); |
| } |
| |
| write_gc0_index(old_index); |
| write_gc0_entryhi(old_entryhi); |
| write_gc0_entrylo0(old_entrylo[0]); |
| write_gc0_entrylo1(old_entrylo[1]); |
| write_gc0_pagemask(old_pagemask); |
| tlbw_use_hazard(); |
| |
| local_irq_restore(flags); |
| } |
| EXPORT_SYMBOL_GPL(kvm_vz_local_flush_guesttlb_all); |
| |
| /** |
| * kvm_vz_save_guesttlb() - Save a range of guest TLB entries. |
| * @buf: Buffer to write TLB entries into. |
| * @index: Start index. |
| * @count: Number of entries to save. |
| * |
| * Save a range of guest TLB entries. The caller must ensure interrupts are |
| * disabled. |
| */ |
| void kvm_vz_save_guesttlb(struct kvm_mips_tlb *buf, unsigned int index, |
| unsigned int count) |
| { |
| unsigned int end = index + count; |
| unsigned long old_entryhi, old_entrylo0, old_entrylo1, old_pagemask; |
| unsigned int guestctl1 = 0; |
| int old_index, i; |
| |
| /* Save registers we're about to clobber */ |
| old_index = read_gc0_index(); |
| old_entryhi = read_gc0_entryhi(); |
| old_entrylo0 = read_gc0_entrylo0(); |
| old_entrylo1 = read_gc0_entrylo1(); |
| old_pagemask = read_gc0_pagemask(); |
| |
| /* Set root GuestID for root probe */ |
| htw_stop(); |
| set_root_gid_to_guest_gid(); |
| if (cpu_has_guestid) |
| guestctl1 = read_c0_guestctl1(); |
| |
| /* Read each entry from guest TLB */ |
| for (i = index; i < end; ++i, ++buf) { |
| write_gc0_index(i); |
| |
| mtc0_tlbr_hazard(); |
| guest_tlb_read(); |
| tlb_read_hazard(); |
| |
| if (cpu_has_guestid && |
| (read_c0_guestctl1() ^ guestctl1) & MIPS_GCTL1_RID) { |
| /* Entry invalid or belongs to another guest */ |
| buf->tlb_hi = UNIQUE_GUEST_ENTRYHI(i); |
| buf->tlb_lo[0] = 0; |
| buf->tlb_lo[1] = 0; |
| buf->tlb_mask = 0; |
| } else { |
| /* Entry belongs to the right guest */ |
| buf->tlb_hi = read_gc0_entryhi(); |
| buf->tlb_lo[0] = read_gc0_entrylo0(); |
| buf->tlb_lo[1] = read_gc0_entrylo1(); |
| buf->tlb_mask = read_gc0_pagemask(); |
| } |
| } |
| |
| /* Clear root GuestID again */ |
| clear_root_gid(); |
| htw_start(); |
| |
| /* Restore clobbered registers */ |
| write_gc0_index(old_index); |
| write_gc0_entryhi(old_entryhi); |
| write_gc0_entrylo0(old_entrylo0); |
| write_gc0_entrylo1(old_entrylo1); |
| write_gc0_pagemask(old_pagemask); |
| |
| tlbw_use_hazard(); |
| } |
| EXPORT_SYMBOL_GPL(kvm_vz_save_guesttlb); |
| |
| /** |
| * kvm_vz_load_guesttlb() - Save a range of guest TLB entries. |
| * @buf: Buffer to read TLB entries from. |
| * @index: Start index. |
| * @count: Number of entries to load. |
| * |
| * Load a range of guest TLB entries. The caller must ensure interrupts are |
| * disabled. |
| */ |
| void kvm_vz_load_guesttlb(const struct kvm_mips_tlb *buf, unsigned int index, |
| unsigned int count) |
| { |
| unsigned int end = index + count; |
| unsigned long old_entryhi, old_entrylo0, old_entrylo1, old_pagemask; |
| int old_index, i; |
| |
| /* Save registers we're about to clobber */ |
| old_index = read_gc0_index(); |
| old_entryhi = read_gc0_entryhi(); |
| old_entrylo0 = read_gc0_entrylo0(); |
| old_entrylo1 = read_gc0_entrylo1(); |
| old_pagemask = read_gc0_pagemask(); |
| |
| /* Set root GuestID for root probe */ |
| htw_stop(); |
| set_root_gid_to_guest_gid(); |
| |
| /* Write each entry to guest TLB */ |
| for (i = index; i < end; ++i, ++buf) { |
| write_gc0_index(i); |
| write_gc0_entryhi(buf->tlb_hi); |
| write_gc0_entrylo0(buf->tlb_lo[0]); |
| write_gc0_entrylo1(buf->tlb_lo[1]); |
| write_gc0_pagemask(buf->tlb_mask); |
| |
| mtc0_tlbw_hazard(); |
| guest_tlb_write_indexed(); |
| } |
| |
| /* Clear root GuestID again */ |
| clear_root_gid(); |
| htw_start(); |
| |
| /* Restore clobbered registers */ |
| write_gc0_index(old_index); |
| write_gc0_entryhi(old_entryhi); |
| write_gc0_entrylo0(old_entrylo0); |
| write_gc0_entrylo1(old_entrylo1); |
| write_gc0_pagemask(old_pagemask); |
| |
| tlbw_use_hazard(); |
| } |
| EXPORT_SYMBOL_GPL(kvm_vz_load_guesttlb); |
| |
| #ifdef CONFIG_CPU_LOONGSON64 |
| void kvm_loongson_clear_guest_vtlb(void) |
| { |
| int idx = read_gc0_index(); |
| |
| /* Set root GuestID for root probe and write of guest TLB entry */ |
| set_root_gid_to_guest_gid(); |
| |
| write_gc0_index(0); |
| guest_tlbinvf(); |
| write_gc0_index(idx); |
| |
| clear_root_gid(); |
| set_c0_diag(LOONGSON_DIAG_ITLB | LOONGSON_DIAG_DTLB); |
| } |
| EXPORT_SYMBOL_GPL(kvm_loongson_clear_guest_vtlb); |
| |
| void kvm_loongson_clear_guest_ftlb(void) |
| { |
| int i; |
| int idx = read_gc0_index(); |
| |
| /* Set root GuestID for root probe and write of guest TLB entry */ |
| set_root_gid_to_guest_gid(); |
| |
| for (i = current_cpu_data.tlbsizevtlb; |
| i < (current_cpu_data.tlbsizevtlb + |
| current_cpu_data.tlbsizeftlbsets); |
| i++) { |
| write_gc0_index(i); |
| guest_tlbinvf(); |
| } |
| write_gc0_index(idx); |
| |
| clear_root_gid(); |
| set_c0_diag(LOONGSON_DIAG_ITLB | LOONGSON_DIAG_DTLB); |
| } |
| EXPORT_SYMBOL_GPL(kvm_loongson_clear_guest_ftlb); |
| #endif |