| /* SPDX-License-Identifier: GPL-2.0-only */ |
| /* |
| * s390x MMU |
| * |
| * Copyright (c) 2017 Red Hat Inc |
| * |
| * Authors: |
| * David Hildenbrand <david@redhat.com> |
| */ |
| |
| #include <libcflat.h> |
| #include <asm/pgtable.h> |
| #include <asm/arch_def.h> |
| #include <asm/barrier.h> |
| #include <vmalloc.h> |
| #include "mmu.h" |
| |
| /* |
| * The naming convention used here is the same as used in the Linux kernel; |
| * this is the correspondence between the s390x architectural names and the |
| * Linux ones: |
| * |
| * pgd - region 1 table entry |
| * p4d - region 2 table entry |
| * pud - region 3 table entry |
| * pmd - segment table entry |
| * pte - page table entry |
| */ |
| |
| static pgd_t *table_root; |
| |
| static void mmu_enable(pgd_t *pgtable) |
| { |
| const uint64_t asce = __pa(pgtable) | ASCE_DT_REGION1 | |
| REGION_TABLE_LENGTH; |
| |
| /* set primary asce */ |
| lctlg(1, asce); |
| assert(stctg(1) == asce); |
| |
| /* enable dat (primary == 0 set as default) */ |
| enable_dat(); |
| |
| /* we can now also use DAT unconditionally in our PGM handler */ |
| lowcore.pgm_new_psw.mask |= PSW_MASK_DAT; |
| } |
| |
| /* |
| * Get the pud (region 3) DAT table entry for the given address and root, |
| * allocating it if necessary |
| */ |
| static inline pud_t *get_pud(pgd_t *pgtable, uintptr_t vaddr) |
| { |
| pgd_t *pgd = pgd_offset(pgtable, vaddr); |
| p4d_t *p4d = p4d_alloc(pgd, vaddr); |
| pud_t *pud = pud_alloc(p4d, vaddr); |
| |
| return pud; |
| } |
| |
| /* |
| * Get the pmd (segment) DAT table entry for the given address and pud, |
| * allocating it if necessary. |
| * The pud must not be huge. |
| */ |
| static inline pmd_t *get_pmd(pud_t *pud, uintptr_t vaddr) |
| { |
| pmd_t *pmd; |
| |
| assert(!pud_huge(*pud)); |
| pmd = pmd_alloc(pud, vaddr); |
| return pmd; |
| } |
| |
| /* |
| * Get the pte (page) DAT table entry for the given address and pmd, |
| * allocating it if necessary. |
| * The pmd must not be large. |
| */ |
| static inline pte_t *get_pte(pmd_t *pmd, uintptr_t vaddr) |
| { |
| pte_t *pte; |
| |
| assert(!pmd_large(*pmd)); |
| pte = pte_alloc(pmd, vaddr); |
| return pte; |
| } |
| |
| /* |
| * Splits a large pmd (segment) DAT table entry into equivalent 4kB small |
| * pages. |
| * @pmd The pmd to split, it must be large. |
| * @va the virtual address corresponding to this pmd. |
| */ |
| static void split_pmd(pmd_t *pmd, uintptr_t va) |
| { |
| phys_addr_t pa = pmd_val(*pmd) & SEGMENT_ENTRY_SFAA; |
| unsigned long i, prot; |
| pte_t *pte; |
| |
| assert(pmd_large(*pmd)); |
| pte = alloc_pages(PAGE_TABLE_ORDER); |
| prot = pmd_val(*pmd) & (SEGMENT_ENTRY_IEP | SEGMENT_ENTRY_P); |
| for (i = 0; i < PAGE_TABLE_ENTRIES; i++) |
| pte_val(pte[i]) = pa | PAGE_SIZE * i | prot; |
| idte_pmdp(va, &pmd_val(*pmd)); |
| pmd_val(*pmd) = __pa(pte) | SEGMENT_ENTRY_TT_SEGMENT; |
| |
| } |
| |
| /* |
| * Splits a huge pud (region 3) DAT table entry into equivalent 1MB large |
| * pages. |
| * @pud The pud to split, it must be huge. |
| * @va the virtual address corresponding to this pud. |
| */ |
| static void split_pud(pud_t *pud, uintptr_t va) |
| { |
| phys_addr_t pa = pud_val(*pud) & REGION3_ENTRY_RFAA; |
| unsigned long i, prot; |
| pmd_t *pmd; |
| |
| assert(pud_huge(*pud)); |
| pmd = alloc_pages(SEGMENT_TABLE_ORDER); |
| prot = pud_val(*pud) & (REGION3_ENTRY_IEP | REGION_ENTRY_P); |
| for (i = 0; i < SEGMENT_TABLE_ENTRIES; i++) |
| pmd_val(pmd[i]) = pa | SZ_1M * i | prot | SEGMENT_ENTRY_FC | SEGMENT_ENTRY_TT_SEGMENT; |
| idte_pudp(va, &pud_val(*pud)); |
| pud_val(*pud) = __pa(pmd) | REGION_ENTRY_TT_REGION3 | REGION_TABLE_LENGTH; |
| } |
| |
| void *get_dat_entry(pgd_t *pgtable, void *vaddr, enum pgt_level level) |
| { |
| uintptr_t va = (uintptr_t)vaddr; |
| pgd_t *pgd; |
| p4d_t *p4d; |
| pud_t *pud; |
| pmd_t *pmd; |
| |
| assert(level && (level <= 5)); |
| pgd = pgd_offset(pgtable, va); |
| if (level == pgtable_level_pgd) |
| return pgd; |
| p4d = p4d_alloc(pgd, va); |
| if (level == pgtable_level_p4d) |
| return p4d; |
| pud = pud_alloc(p4d, va); |
| |
| if (level == pgtable_level_pud) |
| return pud; |
| if (!pud_none(*pud) && pud_huge(*pud)) |
| split_pud(pud, va); |
| pmd = get_pmd(pud, va); |
| if (level == pgtable_level_pmd) |
| return pmd; |
| if (!pmd_none(*pmd) && pmd_large(*pmd)) |
| split_pmd(pmd, va); |
| return get_pte(pmd, va); |
| } |
| |
| void *split_page(pgd_t *pgtable, void *vaddr, enum pgt_level level) |
| { |
| assert((level >= 3) && (level <= 5)); |
| return get_dat_entry(pgtable ? pgtable : table_root, vaddr, level); |
| } |
| |
| phys_addr_t virt_to_pte_phys(pgd_t *pgtable, void *vaddr) |
| { |
| uintptr_t va = (uintptr_t)vaddr; |
| pud_t *pud; |
| pmd_t *pmd; |
| pte_t *pte; |
| |
| pud = get_pud(pgtable, va); |
| if (pud_huge(*pud)) |
| return (pud_val(*pud) & REGION3_ENTRY_RFAA) | (va & ~REGION3_ENTRY_RFAA); |
| pmd = get_pmd(pud, va); |
| if (pmd_large(*pmd)) |
| return (pmd_val(*pmd) & SEGMENT_ENTRY_SFAA) | (va & ~SEGMENT_ENTRY_SFAA); |
| pte = get_pte(pmd, va); |
| return (pte_val(*pte) & PAGE_MASK) | (va & ~PAGE_MASK); |
| } |
| |
| /* |
| * Get the DAT table entry of the given level for the given address, |
| * splitting if necessary. If the entry was not invalid, invalidate it, and |
| * return the pointer to the entry and, if requested, its old value. |
| * @pgtable root of the page tables |
| * @vaddr virtual address |
| * @level 3 (for 2GB pud), 4 (for 1MB pmd) or 5 (for 4kB pages) |
| * @old if not NULL, will be written with the old value of the DAT table |
| * entry before invalidation |
| */ |
| static void *dat_get_and_invalidate(pgd_t *pgtable, void *vaddr, enum pgt_level level, unsigned long *old) |
| { |
| unsigned long va = (unsigned long)vaddr; |
| void *ptr; |
| |
| ptr = get_dat_entry(pgtable, vaddr, level); |
| if (old) |
| *old = *(unsigned long *)ptr; |
| if ((level == pgtable_level_pgd) && !pgd_none(*(pgd_t *)ptr)) |
| idte_pgdp(va, ptr); |
| else if ((level == pgtable_level_p4d) && !p4d_none(*(p4d_t *)ptr)) |
| idte_p4dp(va, ptr); |
| else if ((level == pgtable_level_pud) && !pud_none(*(pud_t *)ptr)) |
| idte_pudp(va, ptr); |
| else if ((level == pgtable_level_pmd) && !pmd_none(*(pmd_t *)ptr)) |
| idte_pmdp(va, ptr); |
| else if (!pte_none(*(pte_t *)ptr)) |
| ipte(va, ptr); |
| return ptr; |
| } |
| |
| static void cleanup_pmd(pmd_t *pmd) |
| { |
| /* was invalid or large, nothing to do */ |
| if (pmd_none(*pmd) || pmd_large(*pmd)) |
| return; |
| /* was not large, free the corresponding page table */ |
| free_pages((void *)(pmd_val(*pmd) & PAGE_MASK)); |
| } |
| |
| static void cleanup_pud(pud_t *pud) |
| { |
| unsigned long i; |
| pmd_t *pmd; |
| |
| /* was invalid or large, nothing to do */ |
| if (pud_none(*pud) || pud_huge(*pud)) |
| return; |
| /* recursively clean up all pmds if needed */ |
| pmd = (pmd_t *)(pud_val(*pud) & PAGE_MASK); |
| for (i = 0; i < SEGMENT_TABLE_ENTRIES; i++) |
| cleanup_pmd(pmd + i); |
| /* free the corresponding segment table */ |
| free_pages(pmd); |
| } |
| |
| /* |
| * Set the DAT entry for the given level of the given virtual address. If a |
| * mapping already existed, it is overwritten. If an existing mapping with |
| * smaller pages existed, all the lower tables are freed. |
| * Returns the pointer to the DAT table entry. |
| * @pgtable root of the page tables |
| * @val the new value for the DAT table entry |
| * @vaddr the virtual address |
| * @level 3 for pud (region 3), 4 for pmd (segment) and 5 for pte (pages) |
| */ |
| static void *set_dat_entry(pgd_t *pgtable, unsigned long val, void *vaddr, enum pgt_level level) |
| { |
| unsigned long old, *res; |
| |
| res = dat_get_and_invalidate(pgtable, vaddr, level, &old); |
| if (level == pgtable_level_pmd) |
| cleanup_pmd((pmd_t *)&old); |
| if (level == pgtable_level_pud) |
| cleanup_pud((pud_t *)&old); |
| *res = val; |
| return res; |
| } |
| |
| pteval_t *install_page(pgd_t *pgtable, phys_addr_t phys, void *vaddr) |
| { |
| assert(IS_ALIGNED(phys, PAGE_SIZE)); |
| assert(IS_ALIGNED((uintptr_t)vaddr, PAGE_SIZE)); |
| return set_dat_entry(pgtable, phys, vaddr, pgtable_level_pte); |
| } |
| |
| pmdval_t *install_large_page(pgd_t *pgtable, phys_addr_t phys, void *vaddr) |
| { |
| assert(IS_ALIGNED(phys, SZ_1M)); |
| assert(IS_ALIGNED((uintptr_t)vaddr, SZ_1M)); |
| return set_dat_entry(pgtable, phys | SEGMENT_ENTRY_FC, vaddr, pgtable_level_pmd); |
| } |
| |
| pudval_t *install_huge_page(pgd_t *pgtable, phys_addr_t phys, void *vaddr) |
| { |
| assert(IS_ALIGNED(phys, SZ_2G)); |
| assert(IS_ALIGNED((uintptr_t)vaddr, SZ_2G)); |
| return set_dat_entry(pgtable, phys | REGION3_ENTRY_FC | REGION_ENTRY_TT_REGION3, vaddr, pgtable_level_pud); |
| } |
| |
| void protect_dat_entry(void *vaddr, unsigned long prot, enum pgt_level level) |
| { |
| unsigned long old, *ptr; |
| |
| ptr = dat_get_and_invalidate(table_root, vaddr, level, &old); |
| *ptr = old | prot; |
| } |
| |
| void unprotect_dat_entry(void *vaddr, unsigned long prot, enum pgt_level level) |
| { |
| unsigned long old, *ptr; |
| |
| ptr = dat_get_and_invalidate(table_root, vaddr, level, &old); |
| *ptr = old & ~prot; |
| } |
| |
| void protect_range(void *start, unsigned long len, unsigned long prot) |
| { |
| uintptr_t curr = (uintptr_t)start & PAGE_MASK; |
| |
| len &= PAGE_MASK; |
| for (; len; len -= PAGE_SIZE, curr += PAGE_SIZE) |
| protect_dat_entry((void *)curr, prot, 5); |
| } |
| |
| void unprotect_range(void *start, unsigned long len, unsigned long prot) |
| { |
| uintptr_t curr = (uintptr_t)start & PAGE_MASK; |
| |
| len &= PAGE_MASK; |
| for (; len; len -= PAGE_SIZE, curr += PAGE_SIZE) |
| unprotect_dat_entry((void *)curr, prot, 5); |
| } |
| |
| static void setup_identity(pgd_t *pgtable, phys_addr_t start_addr, |
| phys_addr_t end_addr) |
| { |
| phys_addr_t cur; |
| |
| start_addr &= PAGE_MASK; |
| for (cur = start_addr; true; cur += PAGE_SIZE) { |
| if (start_addr < end_addr && cur >= end_addr) |
| break; |
| if (start_addr > end_addr && cur <= end_addr) |
| break; |
| install_page(pgtable, cur, __va(cur)); |
| } |
| } |
| |
| void *setup_mmu(phys_addr_t phys_end, void *unused) |
| { |
| pgd_t *page_root; |
| |
| /* allocate a region-1 table */ |
| page_root = pgd_alloc_one(); |
| |
| /* map all physical memory 1:1 */ |
| setup_identity(page_root, 0, phys_end); |
| |
| /* generate 128MB of invalid adresses at the end (for testing PGM) */ |
| init_alloc_vpage((void *) -(1UL << 27)); |
| setup_identity(page_root, -(1UL << 27), 0); |
| |
| /* finally enable DAT with the new table */ |
| mmu_enable(page_root); |
| table_root = page_root; |
| return page_root; |
| } |