| /* SPDX-License-Identifier: GPL-2.0-only */ |
| /* |
| * linux/arch/arm/mm/proc-v7m.S |
| * |
| * Copyright (C) 2008 ARM Ltd. |
| * Copyright (C) 2001 Deep Blue Solutions Ltd. |
| * |
| * This is the "shell" of the ARMv7-M processor support. |
| */ |
| #include <linux/linkage.h> |
| #include <linux/cfi_types.h> |
| #include <asm/assembler.h> |
| #include <asm/page.h> |
| #include <asm/v7m.h> |
| #include "proc-macros.S" |
| |
| SYM_TYPED_FUNC_START(cpu_v7m_proc_init) |
| ret lr |
| SYM_FUNC_END(cpu_v7m_proc_init) |
| |
| SYM_TYPED_FUNC_START(cpu_v7m_proc_fin) |
| ret lr |
| SYM_FUNC_END(cpu_v7m_proc_fin) |
| |
| /* |
| * cpu_v7m_reset(loc) |
| * |
| * Perform a soft reset of the system. Put the CPU into the |
| * same state as it would be if it had been reset, and branch |
| * to what would be the reset vector. |
| * |
| * - loc - location to jump to for soft reset |
| */ |
| .align 5 |
| SYM_TYPED_FUNC_START(cpu_v7m_reset) |
| ret r0 |
| SYM_FUNC_END(cpu_v7m_reset) |
| |
| /* |
| * cpu_v7m_do_idle() |
| * |
| * Idle the processor (eg, wait for interrupt). |
| * |
| * IRQs are already disabled. |
| */ |
| SYM_TYPED_FUNC_START(cpu_v7m_do_idle) |
| wfi |
| ret lr |
| SYM_FUNC_END(cpu_v7m_do_idle) |
| |
| SYM_TYPED_FUNC_START(cpu_v7m_dcache_clean_area) |
| ret lr |
| SYM_FUNC_END(cpu_v7m_dcache_clean_area) |
| |
| /* |
| * There is no MMU, so here is nothing to do. |
| */ |
| SYM_TYPED_FUNC_START(cpu_v7m_switch_mm) |
| ret lr |
| SYM_FUNC_END(cpu_v7m_switch_mm) |
| |
| .globl cpu_v7m_suspend_size |
| .equ cpu_v7m_suspend_size, 0 |
| |
| #ifdef CONFIG_ARM_CPU_SUSPEND |
| SYM_TYPED_FUNC_START(cpu_v7m_do_suspend) |
| ret lr |
| SYM_FUNC_END(cpu_v7m_do_suspend) |
| |
| SYM_TYPED_FUNC_START(cpu_v7m_do_resume) |
| ret lr |
| SYM_FUNC_END(cpu_v7m_do_resume) |
| #endif |
| |
| SYM_TYPED_FUNC_START(cpu_cm7_dcache_clean_area) |
| dcache_line_size r2, r3 |
| movw r3, #:lower16:BASEADDR_V7M_SCB + V7M_SCB_DCCMVAC |
| movt r3, #:upper16:BASEADDR_V7M_SCB + V7M_SCB_DCCMVAC |
| |
| 1: str r0, [r3] @ clean D entry |
| add r0, r0, r2 |
| subs r1, r1, r2 |
| bhi 1b |
| dsb |
| ret lr |
| SYM_FUNC_END(cpu_cm7_dcache_clean_area) |
| |
| SYM_TYPED_FUNC_START(cpu_cm7_proc_fin) |
| movw r2, #:lower16:(BASEADDR_V7M_SCB + V7M_SCB_CCR) |
| movt r2, #:upper16:(BASEADDR_V7M_SCB + V7M_SCB_CCR) |
| ldr r0, [r2] |
| bic r0, r0, #(V7M_SCB_CCR_DC | V7M_SCB_CCR_IC) |
| str r0, [r2] |
| ret lr |
| SYM_FUNC_END(cpu_cm7_proc_fin) |
| |
| .section ".init.text", "ax" |
| |
| __v7m_cm7_setup: |
| mov r8, #(V7M_SCB_CCR_DC | V7M_SCB_CCR_IC| V7M_SCB_CCR_BP) |
| b __v7m_setup_cont |
| /* |
| * __v7m_setup |
| * |
| * This should be able to cover all ARMv7-M cores. |
| */ |
| __v7m_setup: |
| mov r8, 0 |
| |
| __v7m_setup_cont: |
| @ Configure the vector table base address |
| ldr r0, =BASEADDR_V7M_SCB |
| ldr r12, =vector_table |
| str r12, [r0, V7M_SCB_VTOR] |
| |
| @ enable UsageFault, BusFault and MemManage fault. |
| ldr r5, [r0, #V7M_SCB_SHCSR] |
| orr r5, #(V7M_SCB_SHCSR_USGFAULTENA | V7M_SCB_SHCSR_BUSFAULTENA | V7M_SCB_SHCSR_MEMFAULTENA) |
| str r5, [r0, #V7M_SCB_SHCSR] |
| |
| @ Lower the priority of the SVC and PendSV exceptions |
| mov r5, #0x80000000 |
| str r5, [r0, V7M_SCB_SHPR2] @ set SVC priority |
| mov r5, #0x00800000 |
| str r5, [r0, V7M_SCB_SHPR3] @ set PendSV priority |
| |
| @ SVC to switch to handler mode. Notice that this requires sp to |
| @ point to writeable memory because the processor saves |
| @ some registers to the stack. |
| badr r1, 1f |
| ldr r5, [r12, #11 * 4] @ read the SVC vector entry |
| str r1, [r12, #11 * 4] @ write the temporary SVC vector entry |
| dsb |
| mov r6, lr @ save LR |
| ldr sp, =init_thread_union + THREAD_START_SP |
| cpsie i |
| svc #0 |
| 1: cpsid i |
| /* Calculate exc_ret */ |
| orr r10, lr, #EXC_RET_THREADMODE_PROCESSSTACK |
| ldmia sp, {r0-r3, r12} |
| str r5, [r12, #11 * 4] @ restore the original SVC vector entry |
| mov lr, r6 @ restore LR |
| |
| @ Special-purpose control register |
| mov r1, #1 |
| msr control, r1 @ Thread mode has unpriviledged access |
| |
| @ Configure caches (if implemented) |
| teq r8, #0 |
| stmiane sp, {r0-r6, lr} @ v7m_invalidate_l1 touches r0-r6 |
| blne v7m_invalidate_l1 |
| teq r8, #0 @ re-evalutae condition |
| ldmiane sp, {r0-r6, lr} |
| |
| @ Configure the System Control Register to ensure 8-byte stack alignment |
| @ Note the STKALIGN bit is either RW or RAO. |
| ldr r0, [r0, V7M_SCB_CCR] @ system control register |
| orr r0, #V7M_SCB_CCR_STKALIGN |
| orr r0, r0, r8 |
| |
| ret lr |
| ENDPROC(__v7m_setup) |
| |
| /* |
| * Cortex-M7 processor functions |
| */ |
| globl_equ cpu_cm7_proc_init, cpu_v7m_proc_init |
| globl_equ cpu_cm7_reset, cpu_v7m_reset |
| globl_equ cpu_cm7_do_idle, cpu_v7m_do_idle |
| globl_equ cpu_cm7_switch_mm, cpu_v7m_switch_mm |
| |
| define_processor_functions v7m, dabort=nommu_early_abort, pabort=legacy_pabort, nommu=1 |
| define_processor_functions cm7, dabort=nommu_early_abort, pabort=legacy_pabort, nommu=1 |
| |
| .section ".rodata" |
| string cpu_arch_name, "armv7m" |
| string cpu_elf_name "v7m" |
| string cpu_v7m_name "ARMv7-M" |
| |
| .section ".proc.info.init", "a" |
| |
| .macro __v7m_proc name, initfunc, cache_fns = nop_cache_fns, hwcaps = 0, proc_fns = v7m_processor_functions |
| .long 0 /* proc_info_list.__cpu_mm_mmu_flags */ |
| .long 0 /* proc_info_list.__cpu_io_mmu_flags */ |
| initfn \initfunc, \name |
| .long cpu_arch_name |
| .long cpu_elf_name |
| .long HWCAP_HALF | HWCAP_THUMB | HWCAP_FAST_MULT | \hwcaps |
| .long cpu_v7m_name |
| .long \proc_fns |
| .long 0 /* proc_info_list.tlb */ |
| .long 0 /* proc_info_list.user */ |
| .long \cache_fns |
| .endm |
| |
| /* |
| * Match ARM Cortex-M55 processor. |
| */ |
| .type __v7m_cm55_proc_info, #object |
| __v7m_cm55_proc_info: |
| .long 0x410fd220 /* ARM Cortex-M55 0xD22 */ |
| .long 0xff0ffff0 /* Mask off revision, patch release */ |
| __v7m_proc __v7m_cm55_proc_info, __v7m_cm7_setup, hwcaps = HWCAP_EDSP, cache_fns = v7m_cache_fns, proc_fns = cm7_processor_functions |
| .size __v7m_cm55_proc_info, . - __v7m_cm55_proc_info |
| |
| /* |
| * Match ARM Cortex-M33 processor. |
| */ |
| .type __v7m_cm33_proc_info, #object |
| __v7m_cm33_proc_info: |
| .long 0x410fd210 /* ARM Cortex-M33 0xD21 */ |
| .long 0xff0ffff0 /* Mask off revision, patch release */ |
| __v7m_proc __v7m_cm33_proc_info, __v7m_setup, hwcaps = HWCAP_EDSP |
| .size __v7m_cm33_proc_info, . - __v7m_cm33_proc_info |
| |
| /* |
| * Match ARM Cortex-M7 processor. |
| */ |
| .type __v7m_cm7_proc_info, #object |
| __v7m_cm7_proc_info: |
| .long 0x410fc270 /* ARM Cortex-M7 0xC27 */ |
| .long 0xff0ffff0 /* Mask off revision, patch release */ |
| __v7m_proc __v7m_cm7_proc_info, __v7m_cm7_setup, hwcaps = HWCAP_EDSP, cache_fns = v7m_cache_fns, proc_fns = cm7_processor_functions |
| .size __v7m_cm7_proc_info, . - __v7m_cm7_proc_info |
| |
| /* |
| * Match ARM Cortex-M4 processor. |
| */ |
| .type __v7m_cm4_proc_info, #object |
| __v7m_cm4_proc_info: |
| .long 0x410fc240 /* ARM Cortex-M4 0xC24 */ |
| .long 0xff0ffff0 /* Mask off revision, patch release */ |
| __v7m_proc __v7m_cm4_proc_info, __v7m_setup, hwcaps = HWCAP_EDSP |
| .size __v7m_cm4_proc_info, . - __v7m_cm4_proc_info |
| |
| /* |
| * Match ARM Cortex-M3 processor. |
| */ |
| .type __v7m_cm3_proc_info, #object |
| __v7m_cm3_proc_info: |
| .long 0x410fc230 /* ARM Cortex-M3 0xC23 */ |
| .long 0xff0ffff0 /* Mask off revision, patch release */ |
| __v7m_proc __v7m_cm3_proc_info, __v7m_setup |
| .size __v7m_cm3_proc_info, . - __v7m_cm3_proc_info |
| |
| /* |
| * Match any ARMv7-M processor core. |
| */ |
| .type __v7m_proc_info, #object |
| __v7m_proc_info: |
| .long 0x000f0000 @ Required ID value |
| .long 0x000f0000 @ Mask for ID |
| __v7m_proc __v7m_proc_info, __v7m_setup |
| .size __v7m_proc_info, . - __v7m_proc_info |
| |