| /* SPDX-License-Identifier: GPL-2.0-or-later */ |
| /* |
| * This file contains low level CPU setup functions. |
| * Kumar Gala <galak@kernel.crashing.org> |
| * Copyright 2009 Freescale Semiconductor, Inc. |
| * |
| * Based on cpu_setup_6xx code by |
| * Benjamin Herrenschmidt <benh@kernel.crashing.org> |
| */ |
| |
| #include <asm/page.h> |
| #include <asm/processor.h> |
| #include <asm/cputable.h> |
| #include <asm/ppc_asm.h> |
| #include <asm/nohash/mmu-e500.h> |
| #include <asm/asm-offsets.h> |
| #include <asm/mpc85xx.h> |
| |
| _GLOBAL(__e500_icache_setup) |
| mfspr r0, SPRN_L1CSR1 |
| andi. r3, r0, L1CSR1_ICE |
| bnelr /* Already enabled */ |
| oris r0, r0, L1CSR1_CPE@h |
| ori r0, r0, (L1CSR1_ICFI | L1CSR1_ICLFR | L1CSR1_ICE) |
| mtspr SPRN_L1CSR1, r0 /* Enable I-Cache */ |
| isync |
| blr |
| |
| _GLOBAL(__e500_dcache_setup) |
| mfspr r0, SPRN_L1CSR0 |
| andi. r3, r0, L1CSR0_DCE |
| bnelr /* Already enabled */ |
| msync |
| isync |
| li r0, 0 |
| mtspr SPRN_L1CSR0, r0 /* Disable */ |
| msync |
| isync |
| li r0, (L1CSR0_DCFI | L1CSR0_CLFC) |
| mtspr SPRN_L1CSR0, r0 /* Invalidate */ |
| isync |
| 1: mfspr r0, SPRN_L1CSR0 |
| andi. r3, r0, L1CSR0_CLFC |
| bne+ 1b /* Wait for lock bits reset */ |
| oris r0, r0, L1CSR0_CPE@h |
| ori r0, r0, L1CSR0_DCE |
| msync |
| isync |
| mtspr SPRN_L1CSR0, r0 /* Enable */ |
| isync |
| blr |
| |
| /* |
| * FIXME - we haven't yet done testing to determine a reasonable default |
| * value for PW20_WAIT_IDLE_BIT. |
| */ |
| #define PW20_WAIT_IDLE_BIT 50 /* 1ms, TB frequency is 41.66MHZ */ |
| _GLOBAL(setup_pw20_idle) |
| mfspr r3, SPRN_PWRMGTCR0 |
| |
| /* Set PW20_WAIT bit, enable pw20 state*/ |
| ori r3, r3, PWRMGTCR0_PW20_WAIT |
| li r11, PW20_WAIT_IDLE_BIT |
| |
| /* Set Automatic PW20 Core Idle Count */ |
| rlwimi r3, r11, PWRMGTCR0_PW20_ENT_SHIFT, PWRMGTCR0_PW20_ENT |
| |
| mtspr SPRN_PWRMGTCR0, r3 |
| |
| blr |
| |
| /* |
| * FIXME - we haven't yet done testing to determine a reasonable default |
| * value for AV_WAIT_IDLE_BIT. |
| */ |
| #define AV_WAIT_IDLE_BIT 50 /* 1ms, TB frequency is 41.66MHZ */ |
| _GLOBAL(setup_altivec_idle) |
| mfspr r3, SPRN_PWRMGTCR0 |
| |
| /* Enable Altivec Idle */ |
| oris r3, r3, PWRMGTCR0_AV_IDLE_PD_EN@h |
| li r11, AV_WAIT_IDLE_BIT |
| |
| /* Set Automatic AltiVec Idle Count */ |
| rlwimi r3, r11, PWRMGTCR0_AV_IDLE_CNT_SHIFT, PWRMGTCR0_AV_IDLE_CNT |
| |
| mtspr SPRN_PWRMGTCR0, r3 |
| |
| blr |
| |
| #ifdef CONFIG_PPC_E500MC |
| _GLOBAL(__setup_cpu_e6500) |
| mflr r6 |
| #ifdef CONFIG_PPC64 |
| bl setup_altivec_ivors |
| /* Touch IVOR42 only if the CPU supports E.HV category */ |
| mfspr r10,SPRN_MMUCFG |
| rlwinm. r10,r10,0,MMUCFG_LPIDSIZE |
| beq 1f |
| bl setup_lrat_ivor |
| 1: |
| #endif |
| bl setup_pw20_idle |
| bl setup_altivec_idle |
| bl __setup_cpu_e5500 |
| mtlr r6 |
| blr |
| #endif /* CONFIG_PPC_E500MC */ |
| |
| #ifdef CONFIG_PPC32 |
| #ifdef CONFIG_PPC_E500 |
| #ifndef CONFIG_PPC_E500MC |
| _GLOBAL(__setup_cpu_e500v1) |
| _GLOBAL(__setup_cpu_e500v2) |
| mflr r4 |
| bl __e500_icache_setup |
| bl __e500_dcache_setup |
| bl __setup_e500_ivors |
| #if defined(CONFIG_FSL_RIO) || defined(CONFIG_FSL_PCI) |
| /* Ensure that RFXE is set */ |
| mfspr r3,SPRN_HID1 |
| oris r3,r3,HID1_RFXE@h |
| mtspr SPRN_HID1,r3 |
| #endif |
| mtlr r4 |
| blr |
| #else /* CONFIG_PPC_E500MC */ |
| _GLOBAL(__setup_cpu_e500mc) |
| _GLOBAL(__setup_cpu_e5500) |
| mflr r5 |
| bl __e500_icache_setup |
| bl __e500_dcache_setup |
| bl __setup_e500mc_ivors |
| /* |
| * We only want to touch IVOR38-41 if we're running on hardware |
| * that supports category E.HV. The architectural way to determine |
| * this is MMUCFG[LPIDSIZE]. |
| */ |
| mfspr r3, SPRN_MMUCFG |
| rlwinm. r3, r3, 0, MMUCFG_LPIDSIZE |
| beq 1f |
| bl __setup_ehv_ivors |
| b 2f |
| 1: |
| lwz r3, CPU_SPEC_FEATURES(r4) |
| /* We need this check as cpu_setup is also called for |
| * the secondary cores. So, if we have already cleared |
| * the feature on the primary core, avoid doing it on the |
| * secondary core. |
| */ |
| andi. r6, r3, CPU_FTR_EMB_HV |
| beq 2f |
| rlwinm r3, r3, 0, ~CPU_FTR_EMB_HV |
| stw r3, CPU_SPEC_FEATURES(r4) |
| 2: |
| mtlr r5 |
| blr |
| #endif /* CONFIG_PPC_E500MC */ |
| #endif /* CONFIG_PPC_E500 */ |
| #endif /* CONFIG_PPC32 */ |
| |
| #ifdef CONFIG_PPC_BOOK3E_64 |
| _GLOBAL(__restore_cpu_e6500) |
| mflr r5 |
| bl setup_altivec_ivors |
| /* Touch IVOR42 only if the CPU supports E.HV category */ |
| mfspr r10,SPRN_MMUCFG |
| rlwinm. r10,r10,0,MMUCFG_LPIDSIZE |
| beq 1f |
| bl setup_lrat_ivor |
| 1: |
| bl setup_pw20_idle |
| bl setup_altivec_idle |
| bl __restore_cpu_e5500 |
| mtlr r5 |
| blr |
| |
| _GLOBAL(__restore_cpu_e5500) |
| mflr r4 |
| bl __e500_icache_setup |
| bl __e500_dcache_setup |
| bl __setup_base_ivors |
| bl setup_perfmon_ivor |
| bl setup_doorbell_ivors |
| /* |
| * We only want to touch IVOR38-41 if we're running on hardware |
| * that supports category E.HV. The architectural way to determine |
| * this is MMUCFG[LPIDSIZE]. |
| */ |
| mfspr r10,SPRN_MMUCFG |
| rlwinm. r10,r10,0,MMUCFG_LPIDSIZE |
| beq 1f |
| bl setup_ehv_ivors |
| 1: |
| mtlr r4 |
| blr |
| |
| _GLOBAL(__setup_cpu_e5500) |
| mflr r5 |
| bl __e500_icache_setup |
| bl __e500_dcache_setup |
| bl __setup_base_ivors |
| bl setup_perfmon_ivor |
| bl setup_doorbell_ivors |
| /* |
| * We only want to touch IVOR38-41 if we're running on hardware |
| * that supports category E.HV. The architectural way to determine |
| * this is MMUCFG[LPIDSIZE]. |
| */ |
| mfspr r10,SPRN_MMUCFG |
| rlwinm. r10,r10,0,MMUCFG_LPIDSIZE |
| beq 1f |
| bl setup_ehv_ivors |
| b 2f |
| 1: |
| ld r10,CPU_SPEC_FEATURES(r4) |
| LOAD_REG_IMMEDIATE(r9,CPU_FTR_EMB_HV) |
| andc r10,r10,r9 |
| std r10,CPU_SPEC_FEATURES(r4) |
| 2: |
| mtlr r5 |
| blr |
| #endif |
| |
| /* flush L1 data cache, it can apply to e500v2, e500mc and e5500 */ |
| _GLOBAL(flush_dcache_L1) |
| mfmsr r10 |
| wrteei 0 |
| |
| mfspr r3,SPRN_L1CFG0 |
| rlwinm r5,r3,9,3 /* Extract cache block size */ |
| twlgti r5,1 /* Only 32 and 64 byte cache blocks |
| * are currently defined. |
| */ |
| li r4,32 |
| subfic r6,r5,2 /* r6 = log2(1KiB / cache block size) - |
| * log2(number of ways) |
| */ |
| slw r5,r4,r5 /* r5 = cache block size */ |
| |
| rlwinm r7,r3,0,0xff /* Extract number of KiB in the cache */ |
| mulli r7,r7,13 /* An 8-way cache will require 13 |
| * loads per set. |
| */ |
| slw r7,r7,r6 |
| |
| /* save off HID0 and set DCFA */ |
| mfspr r8,SPRN_HID0 |
| ori r9,r8,HID0_DCFA@l |
| mtspr SPRN_HID0,r9 |
| isync |
| |
| LOAD_REG_IMMEDIATE(r6, KERNELBASE) |
| mr r4, r6 |
| mtctr r7 |
| |
| 1: lwz r3,0(r4) /* Load... */ |
| add r4,r4,r5 |
| bdnz 1b |
| |
| msync |
| mr r4, r6 |
| mtctr r7 |
| |
| 1: dcbf 0,r4 /* ...and flush. */ |
| add r4,r4,r5 |
| bdnz 1b |
| |
| /* restore HID0 */ |
| mtspr SPRN_HID0,r8 |
| isync |
| |
| wrtee r10 |
| |
| blr |
| |
| has_L2_cache: |
| /* skip L2 cache on P2040/P2040E as they have no L2 cache */ |
| mfspr r3, SPRN_SVR |
| /* shift right by 8 bits and clear E bit of SVR */ |
| rlwinm r4, r3, 24, ~0x800 |
| |
| lis r3, SVR_P2040@h |
| ori r3, r3, SVR_P2040@l |
| cmpw r4, r3 |
| beq 1f |
| |
| li r3, 1 |
| blr |
| 1: |
| li r3, 0 |
| blr |
| |
| /* flush backside L2 cache */ |
| flush_backside_L2_cache: |
| mflr r10 |
| bl has_L2_cache |
| mtlr r10 |
| cmpwi r3, 0 |
| beq 2f |
| |
| /* Flush the L2 cache */ |
| mfspr r3, SPRN_L2CSR0 |
| ori r3, r3, L2CSR0_L2FL@l |
| msync |
| isync |
| mtspr SPRN_L2CSR0,r3 |
| isync |
| |
| /* check if it is complete */ |
| 1: mfspr r3,SPRN_L2CSR0 |
| andi. r3, r3, L2CSR0_L2FL@l |
| bne 1b |
| 2: |
| blr |
| |
| _GLOBAL(cpu_down_flush_e500v2) |
| mflr r0 |
| bl flush_dcache_L1 |
| mtlr r0 |
| blr |
| |
| _GLOBAL(cpu_down_flush_e500mc) |
| _GLOBAL(cpu_down_flush_e5500) |
| mflr r0 |
| bl flush_dcache_L1 |
| bl flush_backside_L2_cache |
| mtlr r0 |
| blr |
| |
| /* L1 Data Cache of e6500 contains no modified data, no flush is required */ |
| _GLOBAL(cpu_down_flush_e6500) |
| blr |