| // SPDX-License-Identifier: GPL-2.0 |
| // Copyright (C) 2018 Hangzhou C-SKY Microsystems co.,ltd. |
| |
| #include <linux/bug.h> |
| #include <linux/module.h> |
| #include <linux/init.h> |
| #include <linux/signal.h> |
| #include <linux/sched.h> |
| #include <linux/kernel.h> |
| #include <linux/errno.h> |
| #include <linux/string.h> |
| #include <linux/types.h> |
| #include <linux/pagemap.h> |
| #include <linux/ptrace.h> |
| #include <linux/mman.h> |
| #include <linux/mm.h> |
| #include <linux/highmem.h> |
| #include <linux/memblock.h> |
| #include <linux/swap.h> |
| #include <linux/proc_fs.h> |
| #include <linux/pfn.h> |
| #include <linux/initrd.h> |
| |
| #include <asm/setup.h> |
| #include <asm/cachectl.h> |
| #include <asm/dma.h> |
| #include <asm/pgalloc.h> |
| #include <asm/mmu_context.h> |
| #include <asm/sections.h> |
| #include <asm/tlb.h> |
| #include <asm/cacheflush.h> |
| |
| #define PTRS_KERN_TABLE \ |
| ((PTRS_PER_PGD - USER_PTRS_PER_PGD) * PTRS_PER_PTE) |
| |
| pgd_t swapper_pg_dir[PTRS_PER_PGD] __page_aligned_bss; |
| pte_t invalid_pte_table[PTRS_PER_PTE] __page_aligned_bss; |
| pte_t kernel_pte_tables[PTRS_KERN_TABLE] __page_aligned_bss; |
| |
| EXPORT_SYMBOL(invalid_pte_table); |
| unsigned long empty_zero_page[PAGE_SIZE / sizeof(unsigned long)] |
| __page_aligned_bss; |
| EXPORT_SYMBOL(empty_zero_page); |
| |
| #ifdef CONFIG_BLK_DEV_INITRD |
| static void __init setup_initrd(void) |
| { |
| unsigned long size; |
| |
| if (initrd_start >= initrd_end) { |
| pr_err("initrd not found or empty"); |
| goto disable; |
| } |
| |
| if (__pa(initrd_end) > PFN_PHYS(max_low_pfn)) { |
| pr_err("initrd extends beyond end of memory"); |
| goto disable; |
| } |
| |
| size = initrd_end - initrd_start; |
| |
| if (memblock_is_region_reserved(__pa(initrd_start), size)) { |
| pr_err("INITRD: 0x%08lx+0x%08lx overlaps in-use memory region", |
| __pa(initrd_start), size); |
| goto disable; |
| } |
| |
| memblock_reserve(__pa(initrd_start), size); |
| |
| pr_info("Initial ramdisk at: 0x%p (%lu bytes)\n", |
| (void *)(initrd_start), size); |
| |
| initrd_below_start_ok = 1; |
| |
| return; |
| |
| disable: |
| initrd_start = initrd_end = 0; |
| |
| pr_err(" - disabling initrd\n"); |
| } |
| #endif |
| |
| void __init mem_init(void) |
| { |
| #ifdef CONFIG_HIGHMEM |
| unsigned long tmp; |
| |
| set_max_mapnr(highend_pfn - ARCH_PFN_OFFSET); |
| #else |
| set_max_mapnr(max_low_pfn - ARCH_PFN_OFFSET); |
| #endif |
| high_memory = (void *) __va(max_low_pfn << PAGE_SHIFT); |
| |
| #ifdef CONFIG_BLK_DEV_INITRD |
| setup_initrd(); |
| #endif |
| |
| memblock_free_all(); |
| |
| #ifdef CONFIG_HIGHMEM |
| for (tmp = highstart_pfn; tmp < highend_pfn; tmp++) { |
| struct page *page = pfn_to_page(tmp); |
| |
| /* FIXME not sure about */ |
| if (!memblock_is_reserved(tmp << PAGE_SHIFT)) |
| free_highmem_page(page); |
| } |
| #endif |
| } |
| |
| void free_initmem(void) |
| { |
| free_initmem_default(-1); |
| } |
| |
| void pgd_init(unsigned long *p) |
| { |
| int i; |
| |
| for (i = 0; i < PTRS_PER_PGD; i++) |
| p[i] = __pa(invalid_pte_table); |
| |
| flush_tlb_all(); |
| local_icache_inv_all(NULL); |
| } |
| |
| void __init mmu_init(unsigned long min_pfn, unsigned long max_pfn) |
| { |
| int i; |
| |
| for (i = 0; i < USER_PTRS_PER_PGD; i++) |
| swapper_pg_dir[i].pgd = __pa(invalid_pte_table); |
| |
| for (i = USER_PTRS_PER_PGD; i < PTRS_PER_PGD; i++) |
| swapper_pg_dir[i].pgd = |
| __pa(kernel_pte_tables + (PTRS_PER_PTE * (i - USER_PTRS_PER_PGD))); |
| |
| for (i = 0; i < PTRS_KERN_TABLE; i++) |
| set_pte(&kernel_pte_tables[i], __pte(_PAGE_GLOBAL)); |
| |
| for (i = min_pfn; i < max_pfn; i++) |
| set_pte(&kernel_pte_tables[i - PFN_DOWN(va_pa_offset)], pfn_pte(i, PAGE_KERNEL)); |
| |
| flush_tlb_all(); |
| local_icache_inv_all(NULL); |
| |
| /* Setup page mask to 4k */ |
| write_mmu_pagemask(0); |
| |
| setup_pgd(swapper_pg_dir, 0); |
| } |
| |
| void __init fixrange_init(unsigned long start, unsigned long end, |
| pgd_t *pgd_base) |
| { |
| pgd_t *pgd; |
| pud_t *pud; |
| pmd_t *pmd; |
| pte_t *pte; |
| int i, j, k; |
| unsigned long vaddr; |
| |
| vaddr = start; |
| i = pgd_index(vaddr); |
| j = pud_index(vaddr); |
| k = pmd_index(vaddr); |
| pgd = pgd_base + i; |
| |
| for ( ; (i < PTRS_PER_PGD) && (vaddr != end); pgd++, i++) { |
| pud = (pud_t *)pgd; |
| for ( ; (j < PTRS_PER_PUD) && (vaddr != end); pud++, j++) { |
| pmd = (pmd_t *)pud; |
| for (; (k < PTRS_PER_PMD) && (vaddr != end); pmd++, k++) { |
| if (pmd_none(*pmd)) { |
| pte = (pte_t *) memblock_alloc_low(PAGE_SIZE, PAGE_SIZE); |
| if (!pte) |
| panic("%s: Failed to allocate %lu bytes align=%lx\n", |
| __func__, PAGE_SIZE, |
| PAGE_SIZE); |
| |
| set_pmd(pmd, __pmd(__pa(pte))); |
| BUG_ON(pte != pte_offset_kernel(pmd, 0)); |
| } |
| vaddr += PMD_SIZE; |
| } |
| k = 0; |
| } |
| j = 0; |
| } |
| } |
| |
| void __init fixaddr_init(void) |
| { |
| unsigned long vaddr; |
| |
| vaddr = __fix_to_virt(__end_of_fixed_addresses - 1) & PMD_MASK; |
| fixrange_init(vaddr, vaddr + PMD_SIZE, swapper_pg_dir); |
| } |
| |
| static const pgprot_t protection_map[16] = { |
| [VM_NONE] = PAGE_NONE, |
| [VM_READ] = PAGE_READ, |
| [VM_WRITE] = PAGE_READ, |
| [VM_WRITE | VM_READ] = PAGE_READ, |
| [VM_EXEC] = PAGE_READ, |
| [VM_EXEC | VM_READ] = PAGE_READ, |
| [VM_EXEC | VM_WRITE] = PAGE_READ, |
| [VM_EXEC | VM_WRITE | VM_READ] = PAGE_READ, |
| [VM_SHARED] = PAGE_NONE, |
| [VM_SHARED | VM_READ] = PAGE_READ, |
| [VM_SHARED | VM_WRITE] = PAGE_WRITE, |
| [VM_SHARED | VM_WRITE | VM_READ] = PAGE_WRITE, |
| [VM_SHARED | VM_EXEC] = PAGE_READ, |
| [VM_SHARED | VM_EXEC | VM_READ] = PAGE_READ, |
| [VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_WRITE, |
| [VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_WRITE |
| }; |
| DECLARE_VM_GET_PAGE_PROT |