blob: a155180d7a837be3ae49a929ab65a86a90d6743c [file] [log] [blame]
Greg Kroah-Hartmanb2441312017-11-01 15:07:57 +01001// SPDX-License-Identifier: GPL-2.0
Linus Torvalds1da177e2005-04-16 15:20:36 -07002/*
3 * linux/arch/alpha/mm/init.c
4 *
5 * Copyright (C) 1995 Linus Torvalds
6 */
7
8/* 2.3.x zone allocator, 1999 Andrea Arcangeli <andrea@suse.de> */
9
akpm@osdl.orgdf2e71f2006-01-09 20:51:37 -080010#include <linux/pagemap.h>
Linus Torvalds1da177e2005-04-16 15:20:36 -070011#include <linux/signal.h>
12#include <linux/sched.h>
13#include <linux/kernel.h>
14#include <linux/errno.h>
15#include <linux/string.h>
16#include <linux/types.h>
17#include <linux/ptrace.h>
18#include <linux/mman.h>
19#include <linux/mm.h>
20#include <linux/swap.h>
21#include <linux/init.h>
Mike Rapoport57c8a662018-10-30 15:09:49 -070022#include <linux/memblock.h> /* max_low_pfn */
Linus Torvalds1da177e2005-04-16 15:20:36 -070023#include <linux/vmalloc.h>
Tejun Heo5a0e3ad2010-03-24 17:04:11 +090024#include <linux/gfp.h>
Linus Torvalds1da177e2005-04-16 15:20:36 -070025
Linus Torvalds7c0f6ba2016-12-24 11:46:01 -080026#include <linux/uaccess.h>
Linus Torvalds1da177e2005-04-16 15:20:36 -070027#include <asm/pgalloc.h>
28#include <asm/hwrpb.h>
29#include <asm/dma.h>
30#include <asm/mmu_context.h>
31#include <asm/console.h>
32#include <asm/tlb.h>
David Howellsec221202012-03-28 18:11:12 +010033#include <asm/setup.h>
Jiang Liuf3beeb42013-04-29 15:06:25 -070034#include <asm/sections.h>
Linus Torvalds1da177e2005-04-16 15:20:36 -070035
Linus Torvalds1da177e2005-04-16 15:20:36 -070036extern void die_if_kernel(char *,struct pt_regs *,long);
37
38static struct pcb_struct original_pcb;
39
40pgd_t *
41pgd_alloc(struct mm_struct *mm)
42{
43 pgd_t *ret, *init;
44
45 ret = (pgd_t *)__get_free_page(GFP_KERNEL | __GFP_ZERO);
46 init = pgd_offset(&init_mm, 0UL);
47 if (ret) {
48#ifdef CONFIG_ALPHA_LARGE_VMALLOC
49 memcpy (ret + USER_PTRS_PER_PGD, init + USER_PTRS_PER_PGD,
50 (PTRS_PER_PGD - USER_PTRS_PER_PGD - 1)*sizeof(pgd_t));
51#else
52 pgd_val(ret[PTRS_PER_PGD-2]) = pgd_val(init[PTRS_PER_PGD-2]);
53#endif
54
55 /* The last PGD entry is the VPTB self-map. */
56 pgd_val(ret[PTRS_PER_PGD-1])
57 = pte_val(mk_pte(virt_to_page(ret), PAGE_KERNEL));
58 }
59 return ret;
60}
61
Linus Torvalds1da177e2005-04-16 15:20:36 -070062
63/*
64 * BAD_PAGE is the page that is used for page faults when linux
65 * is out-of-memory. Older versions of linux just did a
66 * do_exit(), but using this instead means there is less risk
67 * for a process dying in kernel mode, possibly leaving an inode
68 * unused etc..
69 *
70 * BAD_PAGETABLE is the accompanying page-table: it is initialized
71 * to point to BAD_PAGE entries.
72 *
73 * ZERO_PAGE is a special page that is used for zero-initialized
74 * data and COW.
75 */
76pmd_t *
77__bad_pagetable(void)
78{
Kees Cook545c2722022-02-02 14:13:15 -080079 memset(absolute_pointer(EMPTY_PGT), 0, PAGE_SIZE);
Linus Torvalds1da177e2005-04-16 15:20:36 -070080 return (pmd_t *) EMPTY_PGT;
81}
82
83pte_t
84__bad_page(void)
85{
Kees Cook545c2722022-02-02 14:13:15 -080086 memset(absolute_pointer(EMPTY_PGE), 0, PAGE_SIZE);
Linus Torvalds1da177e2005-04-16 15:20:36 -070087 return pte_mkdirty(mk_pte(virt_to_page(EMPTY_PGE), PAGE_SHARED));
88}
89
Linus Torvalds1da177e2005-04-16 15:20:36 -070090static inline unsigned long
91load_PCB(struct pcb_struct *pcb)
92{
93 register unsigned long sp __asm__("$30");
94 pcb->ksp = sp;
95 return __reload_thread(pcb);
96}
97
98/* Set up initial PCB, VPTB, and other such nicities. */
99
100static inline void
101switch_to_system_map(void)
102{
103 unsigned long newptbr;
104 unsigned long original_pcb_ptr;
105
106 /* Initialize the kernel's page tables. Linux puts the vptb in
107 the last slot of the L1 page table. */
108 memset(swapper_pg_dir, 0, PAGE_SIZE);
109 newptbr = ((unsigned long) swapper_pg_dir - PAGE_OFFSET) >> PAGE_SHIFT;
110 pgd_val(swapper_pg_dir[1023]) =
111 (newptbr << 32) | pgprot_val(PAGE_KERNEL);
112
113 /* Set the vptb. This is often done by the bootloader, but
114 shouldn't be required. */
115 if (hwrpb->vptb != 0xfffffffe00000000UL) {
116 wrvptptr(0xfffffffe00000000UL);
117 hwrpb->vptb = 0xfffffffe00000000UL;
118 hwrpb_update_checksum(hwrpb);
119 }
120
121 /* Also set up the real kernel PCB while we're at it. */
122 init_thread_info.pcb.ptbr = newptbr;
123 init_thread_info.pcb.flags = 1; /* set FEN, clear everything else */
124 original_pcb_ptr = load_PCB(&init_thread_info.pcb);
125 tbia();
126
127 /* Save off the contents of the original PCB so that we can
128 restore the original console's page tables for a clean reboot.
129
130 Note that the PCB is supposed to be a physical address, but
131 since KSEG values also happen to work, folks get confused.
132 Check this here. */
133
134 if (original_pcb_ptr < PAGE_OFFSET) {
135 original_pcb_ptr = (unsigned long)
136 phys_to_virt(original_pcb_ptr);
137 }
138 original_pcb = *(struct pcb_struct *) original_pcb_ptr;
139}
140
141int callback_init_done;
142
143void * __init
144callback_init(void * kernel_end)
145{
146 struct crb_struct * crb;
147 pgd_t *pgd;
Mike Rapoporta73c9482019-12-04 16:53:44 -0800148 p4d_t *p4d;
149 pud_t *pud;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700150 pmd_t *pmd;
151 void *two_pages;
152
153 /* Starting at the HWRPB, locate the CRB. */
154 crb = (struct crb_struct *)((char *)hwrpb + hwrpb->crb_offset);
155
156 if (alpha_using_srm) {
157 /* Tell the console whither it is to be remapped. */
158 if (srm_fixup(VMALLOC_START, (unsigned long)hwrpb))
159 __halt(); /* "We're boned." --Bender */
160
161 /* Edit the procedure descriptors for DISPATCH and FIXUP. */
162 crb->dispatch_va = (struct procdesc_struct *)
163 (VMALLOC_START + (unsigned long)crb->dispatch_va
164 - crb->map[0].va);
165 crb->fixup_va = (struct procdesc_struct *)
166 (VMALLOC_START + (unsigned long)crb->fixup_va
167 - crb->map[0].va);
168 }
169
170 switch_to_system_map();
171
172 /* Allocate one PGD and one PMD. In the case of SRM, we'll need
173 these to actually remap the console. There is an assumption
174 here that only one of each is needed, and this allows for 8MB.
175 On systems with larger consoles, additional pages will be
176 allocated as needed during the mapping process.
177
178 In the case of not SRM, but not CONFIG_ALPHA_LARGE_VMALLOC,
179 we need to allocate the PGD we use for vmalloc before we start
180 forking other tasks. */
181
182 two_pages = (void *)
183 (((unsigned long)kernel_end + ~PAGE_MASK) & PAGE_MASK);
184 kernel_end = two_pages + 2*PAGE_SIZE;
185 memset(two_pages, 0, 2*PAGE_SIZE);
186
187 pgd = pgd_offset_k(VMALLOC_START);
Mike Rapoporta73c9482019-12-04 16:53:44 -0800188 p4d = p4d_offset(pgd, VMALLOC_START);
189 pud = pud_offset(p4d, VMALLOC_START);
190 pud_set(pud, (pmd_t *)two_pages);
191 pmd = pmd_offset(pud, VMALLOC_START);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700192 pmd_set(pmd, (pte_t *)(two_pages + PAGE_SIZE));
193
194 if (alpha_using_srm) {
195 static struct vm_struct console_remap_vm;
Tejun Heof0aa6612009-02-20 16:29:08 +0900196 unsigned long nr_pages = 0;
197 unsigned long vaddr;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700198 unsigned long i, j;
199
Tejun Heof0aa6612009-02-20 16:29:08 +0900200 /* calculate needed size */
201 for (i = 0; i < crb->map_entries; ++i)
202 nr_pages += crb->map[i].count;
203
204 /* register the vm area */
205 console_remap_vm.flags = VM_ALLOC;
206 console_remap_vm.size = nr_pages << PAGE_SHIFT;
Tejun Heoc0c0a292009-02-24 11:57:21 +0900207 vm_area_register_early(&console_remap_vm, PAGE_SIZE);
Tejun Heof0aa6612009-02-20 16:29:08 +0900208
Tejun Heoaf6326d2009-03-01 16:03:16 +0900209 vaddr = (unsigned long)console_remap_vm.addr;
Tejun Heof0aa6612009-02-20 16:29:08 +0900210
Linus Torvalds1da177e2005-04-16 15:20:36 -0700211 /* Set up the third level PTEs and update the virtual
212 addresses of the CRB entries. */
213 for (i = 0; i < crb->map_entries; ++i) {
214 unsigned long pfn = crb->map[i].pa >> PAGE_SHIFT;
215 crb->map[i].va = vaddr;
216 for (j = 0; j < crb->map[i].count; ++j) {
Simon Arlottc3a2dde2007-10-20 01:04:37 +0200217 /* Newer consoles (especially on larger
Linus Torvalds1da177e2005-04-16 15:20:36 -0700218 systems) may require more pages of
219 PTEs. Grab additional pages as needed. */
Mike Rapoporta73c9482019-12-04 16:53:44 -0800220 if (pmd != pmd_offset(pud, vaddr)) {
Linus Torvalds1da177e2005-04-16 15:20:36 -0700221 memset(kernel_end, 0, PAGE_SIZE);
Mike Rapoporta73c9482019-12-04 16:53:44 -0800222 pmd = pmd_offset(pud, vaddr);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700223 pmd_set(pmd, (pte_t *)kernel_end);
224 kernel_end += PAGE_SIZE;
225 }
226 set_pte(pte_offset_kernel(pmd, vaddr),
227 pfn_pte(pfn, PAGE_KERNEL));
228 pfn++;
229 vaddr += PAGE_SIZE;
230 }
231 }
Linus Torvalds1da177e2005-04-16 15:20:36 -0700232 }
233
234 callback_init_done = 1;
235 return kernel_end;
236}
237
Linus Torvalds1da177e2005-04-16 15:20:36 -0700238/*
239 * paging_init() sets up the memory map.
240 */
Sam Ravnborg64d158b2007-07-31 00:37:53 -0700241void __init paging_init(void)
Linus Torvalds1da177e2005-04-16 15:20:36 -0700242{
Mike Rapoportfa3354e2020-06-03 15:57:06 -0700243 unsigned long max_zone_pfn[MAX_NR_ZONES] = {0, };
244 unsigned long dma_pfn;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700245
246 dma_pfn = virt_to_phys((char *)MAX_DMA_ADDRESS) >> PAGE_SHIFT;
Mike Rapoportfa3354e2020-06-03 15:57:06 -0700247 max_pfn = max_low_pfn;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700248
Mike Rapoportfa3354e2020-06-03 15:57:06 -0700249 max_zone_pfn[ZONE_DMA] = dma_pfn;
250 max_zone_pfn[ZONE_NORMAL] = max_pfn;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700251
252 /* Initialize mem_map[]. */
Mike Rapoportfa3354e2020-06-03 15:57:06 -0700253 free_area_init(max_zone_pfn);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700254
255 /* Initialize the kernel's ZERO_PGE. */
Kees Cook545c2722022-02-02 14:13:15 -0800256 memset(absolute_pointer(ZERO_PGE), 0, PAGE_SIZE);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700257}
Linus Torvalds1da177e2005-04-16 15:20:36 -0700258
259#if defined(CONFIG_ALPHA_GENERIC) || defined(CONFIG_ALPHA_SRM)
260void
261srm_paging_stop (void)
262{
263 /* Move the vptb back to where the SRM console expects it. */
264 swapper_pg_dir[1] = swapper_pg_dir[1023];
265 tbia();
266 wrvptptr(0x200000000UL);
267 hwrpb->vptb = 0x200000000UL;
268 hwrpb_update_checksum(hwrpb);
269
270 /* Reload the page tables that the console had in use. */
271 load_PCB(&original_pcb);
272 tbia();
273}
274#endif
275
Linus Torvalds1da177e2005-04-16 15:20:36 -0700276void __init
277mem_init(void)
278{
Jiang Liu2fb1cd52013-07-03 15:04:36 -0700279 set_max_mapnr(max_low_pfn);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700280 high_memory = (void *) __va(max_low_pfn * PAGE_SIZE);
Mike Rapoportc6ffc5c2018-10-30 15:09:30 -0700281 memblock_free_all();
Linus Torvalds1da177e2005-04-16 15:20:36 -0700282}
Anshuman Khandual4975f602022-07-11 12:35:48 +0530283
284static const pgprot_t protection_map[16] = {
285 [VM_NONE] = _PAGE_P(_PAGE_FOE | _PAGE_FOW |
286 _PAGE_FOR),
287 [VM_READ] = _PAGE_P(_PAGE_FOE | _PAGE_FOW),
288 [VM_WRITE] = _PAGE_P(_PAGE_FOE),
289 [VM_WRITE | VM_READ] = _PAGE_P(_PAGE_FOE),
290 [VM_EXEC] = _PAGE_P(_PAGE_FOW | _PAGE_FOR),
291 [VM_EXEC | VM_READ] = _PAGE_P(_PAGE_FOW),
292 [VM_EXEC | VM_WRITE] = _PAGE_P(0),
293 [VM_EXEC | VM_WRITE | VM_READ] = _PAGE_P(0),
294 [VM_SHARED] = _PAGE_S(_PAGE_FOE | _PAGE_FOW |
295 _PAGE_FOR),
296 [VM_SHARED | VM_READ] = _PAGE_S(_PAGE_FOE | _PAGE_FOW),
297 [VM_SHARED | VM_WRITE] = _PAGE_S(_PAGE_FOE),
298 [VM_SHARED | VM_WRITE | VM_READ] = _PAGE_S(_PAGE_FOE),
299 [VM_SHARED | VM_EXEC] = _PAGE_S(_PAGE_FOW | _PAGE_FOR),
300 [VM_SHARED | VM_EXEC | VM_READ] = _PAGE_S(_PAGE_FOW),
301 [VM_SHARED | VM_EXEC | VM_WRITE] = _PAGE_S(0),
302 [VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = _PAGE_S(0)
303};
304DECLARE_VM_GET_PAGE_PROT