| /* SPDX-License-Identifier: GPL-2.0 */ |
| /* |
| * Copyright (C) 2020-2022 Loongson Technology Corporation Limited |
| */ |
| #ifndef _ASM_PROCESSOR_H |
| #define _ASM_PROCESSOR_H |
| |
| #include <linux/atomic.h> |
| #include <linux/cpumask.h> |
| #include <linux/sizes.h> |
| |
| #include <asm/cpu.h> |
| #include <asm/cpu-info.h> |
| #include <asm/hw_breakpoint.h> |
| #include <asm/loongarch.h> |
| #include <asm/vdso/processor.h> |
| #include <uapi/asm/ptrace.h> |
| #include <uapi/asm/sigcontext.h> |
| |
| #ifdef CONFIG_32BIT |
| |
| #define TASK_SIZE 0x80000000UL |
| #define TASK_SIZE_MIN TASK_SIZE |
| #define STACK_TOP_MAX TASK_SIZE |
| |
| #define TASK_IS_32BIT_ADDR 1 |
| |
| #endif |
| |
| #ifdef CONFIG_64BIT |
| |
| #define TASK_SIZE32 0x100000000UL |
| #define TASK_SIZE64 (0x1UL << ((cpu_vabits > VA_BITS) ? VA_BITS : cpu_vabits)) |
| |
| #define TASK_SIZE (test_thread_flag(TIF_32BIT_ADDR) ? TASK_SIZE32 : TASK_SIZE64) |
| #define TASK_SIZE_MIN TASK_SIZE32 |
| #define STACK_TOP_MAX TASK_SIZE64 |
| |
| #define TASK_SIZE_OF(tsk) \ |
| (test_tsk_thread_flag(tsk, TIF_32BIT_ADDR) ? TASK_SIZE32 : TASK_SIZE64) |
| |
| #define TASK_IS_32BIT_ADDR test_thread_flag(TIF_32BIT_ADDR) |
| |
| #endif |
| |
| #define VDSO_RANDOMIZE_SIZE (TASK_IS_32BIT_ADDR ? SZ_1M : SZ_64M) |
| |
| unsigned long stack_top(void); |
| #define STACK_TOP stack_top() |
| |
| /* |
| * This decides where the kernel will search for a free chunk of vm |
| * space during mmap's. |
| */ |
| #define TASK_UNMAPPED_BASE PAGE_ALIGN(TASK_SIZE / 3) |
| |
| #define FPU_REG_WIDTH 256 |
| #define FPU_ALIGN __attribute__((aligned(32))) |
| |
| union fpureg { |
| __u32 val32[FPU_REG_WIDTH / 32]; |
| __u64 val64[FPU_REG_WIDTH / 64]; |
| }; |
| |
| #define FPR_IDX(width, idx) (idx) |
| |
| #define BUILD_FPR_ACCESS(width) \ |
| static inline u##width get_fpr##width(union fpureg *fpr, unsigned idx) \ |
| { \ |
| return fpr->val##width[FPR_IDX(width, idx)]; \ |
| } \ |
| \ |
| static inline void set_fpr##width(union fpureg *fpr, unsigned int idx, \ |
| u##width val) \ |
| { \ |
| fpr->val##width[FPR_IDX(width, idx)] = val; \ |
| } |
| |
| BUILD_FPR_ACCESS(32) |
| BUILD_FPR_ACCESS(64) |
| |
| struct loongarch_fpu { |
| uint64_t fcc; /* 8x8 */ |
| uint32_t fcsr; |
| uint32_t ftop; |
| union fpureg fpr[NUM_FPU_REGS]; |
| }; |
| |
| struct loongarch_lbt { |
| /* Scratch registers */ |
| unsigned long scr0; |
| unsigned long scr1; |
| unsigned long scr2; |
| unsigned long scr3; |
| /* Eflags register */ |
| unsigned long eflags; |
| }; |
| |
| #define INIT_CPUMASK { \ |
| {0,} \ |
| } |
| |
| #define ARCH_MIN_TASKALIGN 32 |
| |
| struct loongarch_vdso_info; |
| |
| /* |
| * If you change thread_struct remember to change the #defines below too! |
| */ |
| struct thread_struct { |
| /* Main processor registers. */ |
| unsigned long reg01, reg03, reg22; /* ra sp fp */ |
| unsigned long reg23, reg24, reg25, reg26; /* s0-s3 */ |
| unsigned long reg27, reg28, reg29, reg30, reg31; /* s4-s8 */ |
| |
| /* __schedule() return address / call frame address */ |
| unsigned long sched_ra; |
| unsigned long sched_cfa; |
| |
| /* CSR registers */ |
| unsigned long csr_prmd; |
| unsigned long csr_crmd; |
| unsigned long csr_euen; |
| unsigned long csr_ecfg; |
| unsigned long csr_badvaddr; /* Last user fault */ |
| |
| /* Other stuff associated with the thread. */ |
| unsigned long trap_nr; |
| unsigned long error_code; |
| unsigned long single_step; /* Used by PTRACE_SINGLESTEP */ |
| struct loongarch_vdso_info *vdso; |
| |
| /* |
| * FPU & vector registers, must be at the last of inherited |
| * context because they are conditionally copied at fork(). |
| */ |
| struct loongarch_fpu fpu FPU_ALIGN; |
| struct loongarch_lbt lbt; /* Also conditionally copied */ |
| |
| /* Hardware breakpoints pinned to this task. */ |
| struct perf_event *hbp_break[LOONGARCH_MAX_BRP]; |
| struct perf_event *hbp_watch[LOONGARCH_MAX_WRP]; |
| }; |
| |
| #define thread_saved_ra(tsk) (tsk->thread.sched_ra) |
| #define thread_saved_fp(tsk) (tsk->thread.sched_cfa) |
| |
| #define INIT_THREAD { \ |
| /* \ |
| * Main processor registers \ |
| */ \ |
| .reg01 = 0, \ |
| .reg03 = 0, \ |
| .reg22 = 0, \ |
| .reg23 = 0, \ |
| .reg24 = 0, \ |
| .reg25 = 0, \ |
| .reg26 = 0, \ |
| .reg27 = 0, \ |
| .reg28 = 0, \ |
| .reg29 = 0, \ |
| .reg30 = 0, \ |
| .reg31 = 0, \ |
| .sched_ra = 0, \ |
| .sched_cfa = 0, \ |
| .csr_crmd = 0, \ |
| .csr_prmd = 0, \ |
| .csr_euen = 0, \ |
| .csr_ecfg = 0, \ |
| .csr_badvaddr = 0, \ |
| /* \ |
| * Other stuff associated with the process \ |
| */ \ |
| .trap_nr = 0, \ |
| .error_code = 0, \ |
| /* \ |
| * FPU & vector registers \ |
| */ \ |
| .fpu = { \ |
| .fcc = 0, \ |
| .fcsr = 0, \ |
| .ftop = 0, \ |
| .fpr = {{{0,},},}, \ |
| }, \ |
| .hbp_break = {0}, \ |
| .hbp_watch = {0}, \ |
| } |
| |
| struct task_struct; |
| |
| enum idle_boot_override {IDLE_NO_OVERRIDE = 0, IDLE_HALT, IDLE_NOMWAIT, IDLE_POLL}; |
| |
| extern unsigned long boot_option_idle_override; |
| /* |
| * Do necessary setup to start up a newly executed thread. |
| */ |
| extern void start_thread(struct pt_regs *regs, unsigned long pc, unsigned long sp); |
| |
| unsigned long __get_wchan(struct task_struct *p); |
| |
| #define __KSTK_TOS(tsk) ((unsigned long)task_stack_page(tsk) + \ |
| THREAD_SIZE - sizeof(struct pt_regs)) |
| #define task_pt_regs(tsk) ((struct pt_regs *)__KSTK_TOS(tsk)) |
| #define KSTK_EIP(tsk) (task_pt_regs(tsk)->csr_era) |
| #define KSTK_ESP(tsk) (task_pt_regs(tsk)->regs[3]) |
| #define KSTK_EUEN(tsk) (task_pt_regs(tsk)->csr_euen) |
| #define KSTK_ECFG(tsk) (task_pt_regs(tsk)->csr_ecfg) |
| |
| #define return_address() ({__asm__ __volatile__("":::"$1"); __builtin_return_address(0);}) |
| |
| #ifdef CONFIG_CPU_HAS_PREFETCH |
| |
| #define ARCH_HAS_PREFETCH |
| #define prefetch(x) __builtin_prefetch((x), 0, 1) |
| |
| #define ARCH_HAS_PREFETCHW |
| #define prefetchw(x) __builtin_prefetch((x), 1, 1) |
| |
| #endif |
| |
| #endif /* _ASM_PROCESSOR_H */ |