| /* SPDX-License-Identifier: GPL-2.0-or-later |
| * -*- mode: asm -*- |
| * |
| * linux/arch/m68k/kernel/entry.S |
| * |
| * Copyright (C) 1991, 1992 Linus Torvalds |
| * |
| * Linux/m68k support by Hamish Macdonald |
| * |
| * 68060 fixes by Jesper Skov |
| * |
| */ |
| |
| /* |
| * entry.S contains the system-call and fault low-level handling routines. |
| * This also contains the timer-interrupt handler, as well as all interrupts |
| * and faults that can result in a task-switch. |
| * |
| * NOTE: This code handles signal-recognition, which happens every time |
| * after a timer-interrupt and after each system call. |
| * |
| */ |
| |
| /* |
| * 12/03/96 Jes: Currently we only support m68k single-cpu systems, so |
| * all pointers that used to be 'current' are now entry |
| * number 0 in the 'current_set' list. |
| * |
| * 6/05/00 RZ: addedd writeback completion after return from sighandler |
| * for 68040 |
| */ |
| |
| #include <linux/linkage.h> |
| #include <asm/errno.h> |
| #include <asm/setup.h> |
| #include <asm/traps.h> |
| #include <asm/unistd.h> |
| #include <asm/asm-offsets.h> |
| #include <asm/entry.h> |
| |
| .globl system_call, buserr, trap, resume |
| .globl sys_call_table |
| .globl __sys_fork, __sys_clone, __sys_vfork |
| .globl bad_interrupt |
| .globl auto_irqhandler_fixup |
| .globl user_irqvec_fixup |
| |
| .text |
| ENTRY(__sys_fork) |
| SAVE_SWITCH_STACK |
| jbsr sys_fork |
| lea %sp@(24),%sp |
| rts |
| |
| ENTRY(__sys_clone) |
| SAVE_SWITCH_STACK |
| pea %sp@(SWITCH_STACK_SIZE) |
| jbsr m68k_clone |
| lea %sp@(28),%sp |
| rts |
| |
| ENTRY(__sys_vfork) |
| SAVE_SWITCH_STACK |
| jbsr sys_vfork |
| lea %sp@(24),%sp |
| rts |
| |
| ENTRY(__sys_clone3) |
| SAVE_SWITCH_STACK |
| pea %sp@(SWITCH_STACK_SIZE) |
| jbsr m68k_clone3 |
| lea %sp@(28),%sp |
| rts |
| |
| ENTRY(sys_sigreturn) |
| SAVE_SWITCH_STACK |
| movel %sp,%a1 | switch_stack pointer |
| lea %sp@(SWITCH_STACK_SIZE),%a0 | pt_regs pointer |
| lea %sp@(-84),%sp | leave a gap |
| movel %a1,%sp@- |
| movel %a0,%sp@- |
| jbsr do_sigreturn |
| jra 1f | shared with rt_sigreturn() |
| |
| ENTRY(sys_rt_sigreturn) |
| SAVE_SWITCH_STACK |
| movel %sp,%a1 | switch_stack pointer |
| lea %sp@(SWITCH_STACK_SIZE),%a0 | pt_regs pointer |
| lea %sp@(-84),%sp | leave a gap |
| movel %a1,%sp@- |
| movel %a0,%sp@- |
| | stack contents: |
| | [original pt_regs address] [original switch_stack address] |
| | [gap] [switch_stack] [pt_regs] [exception frame] |
| jbsr do_rt_sigreturn |
| |
| 1: |
| | stack contents now: |
| | [original pt_regs address] [original switch_stack address] |
| | [unused part of the gap] [moved switch_stack] [moved pt_regs] |
| | [replacement exception frame] |
| | return value of do_{rt_,}sigreturn() points to moved switch_stack. |
| |
| movel %d0,%sp | discard the leftover junk |
| RESTORE_SWITCH_STACK |
| | stack contents now is just [syscall return address] [pt_regs] [frame] |
| | return pt_regs.d0 |
| movel %sp@(PT_OFF_D0+4),%d0 |
| rts |
| |
| ENTRY(buserr) |
| SAVE_ALL_INT |
| GET_CURRENT(%d0) |
| movel %sp,%sp@- | stack frame pointer argument |
| jbsr buserr_c |
| addql #4,%sp |
| jra ret_from_exception |
| |
| ENTRY(trap) |
| SAVE_ALL_INT |
| GET_CURRENT(%d0) |
| movel %sp,%sp@- | stack frame pointer argument |
| jbsr trap_c |
| addql #4,%sp |
| jra ret_from_exception |
| |
| | After a fork we jump here directly from resume, |
| | so that %d1 contains the previous task |
| | schedule_tail now used regardless of CONFIG_SMP |
| ENTRY(ret_from_fork) |
| movel %d1,%sp@- |
| jsr schedule_tail |
| addql #4,%sp |
| jra ret_from_exception |
| |
| ENTRY(ret_from_kernel_thread) |
| | a3 contains the kernel thread payload, d7 - its argument |
| movel %d1,%sp@- |
| jsr schedule_tail |
| movel %d7,(%sp) |
| jsr %a3@ |
| addql #4,%sp |
| jra ret_from_exception |
| |
| #if defined(CONFIG_COLDFIRE) || !defined(CONFIG_MMU) |
| |
| #ifdef TRAP_DBG_INTERRUPT |
| |
| .globl dbginterrupt |
| ENTRY(dbginterrupt) |
| SAVE_ALL_INT |
| GET_CURRENT(%d0) |
| movel %sp,%sp@- /* stack frame pointer argument */ |
| jsr dbginterrupt_c |
| addql #4,%sp |
| jra ret_from_exception |
| #endif |
| |
| ENTRY(reschedule) |
| /* save top of frame */ |
| pea %sp@ |
| jbsr set_esp0 |
| addql #4,%sp |
| pea ret_from_exception |
| jmp schedule |
| |
| ENTRY(ret_from_user_signal) |
| moveq #__NR_sigreturn,%d0 |
| trap #0 |
| |
| ENTRY(ret_from_user_rt_signal) |
| movel #__NR_rt_sigreturn,%d0 |
| trap #0 |
| |
| #else |
| |
| do_trace_entry: |
| movel #-ENOSYS,%sp@(PT_OFF_D0)| needed for strace |
| subql #4,%sp |
| SAVE_SWITCH_STACK |
| jbsr syscall_trace_enter |
| RESTORE_SWITCH_STACK |
| addql #4,%sp |
| addql #1,%d0 | optimization for cmpil #-1,%d0 |
| jeq ret_from_syscall |
| movel %sp@(PT_OFF_ORIG_D0),%d0 |
| cmpl #NR_syscalls,%d0 |
| jcs syscall |
| jra ret_from_syscall |
| badsys: |
| movel #-ENOSYS,%sp@(PT_OFF_D0) |
| jra ret_from_syscall |
| |
| do_trace_exit: |
| subql #4,%sp |
| SAVE_SWITCH_STACK |
| jbsr syscall_trace_leave |
| RESTORE_SWITCH_STACK |
| addql #4,%sp |
| jra .Lret_from_exception |
| |
| ENTRY(system_call) |
| SAVE_ALL_SYS |
| |
| GET_CURRENT(%d1) |
| movel %d1,%a1 |
| |
| | save top of frame |
| movel %sp,%curptr@(TASK_THREAD+THREAD_ESP0) |
| |
| | syscall trace? |
| tstb %a1@(TINFO_FLAGS+2) |
| jmi do_trace_entry |
| | seccomp filter active? |
| btst #5,%a1@(TINFO_FLAGS+2) |
| bnes do_trace_entry |
| cmpl #NR_syscalls,%d0 |
| jcc badsys |
| syscall: |
| jbsr @(sys_call_table,%d0:l:4)@(0) |
| movel %d0,%sp@(PT_OFF_D0) | save the return value |
| ret_from_syscall: |
| |oriw #0x0700,%sr |
| movel %curptr@(TASK_STACK),%a1 |
| movew %a1@(TINFO_FLAGS+2),%d0 |
| jne syscall_exit_work |
| 1: RESTORE_ALL |
| |
| syscall_exit_work: |
| btst #5,%sp@(PT_OFF_SR) | check if returning to kernel |
| bnes 1b | if so, skip resched, signals |
| lslw #1,%d0 |
| jcs do_trace_exit |
| jmi do_delayed_trace |
| lslw #8,%d0 |
| jne do_signal_return |
| pea resume_userspace |
| jra schedule |
| |
| |
| ENTRY(ret_from_exception) |
| .Lret_from_exception: |
| btst #5,%sp@(PT_OFF_SR) | check if returning to kernel |
| bnes 1f | if so, skip resched, signals |
| | only allow interrupts when we are really the last one on the |
| | kernel stack, otherwise stack overflow can occur during |
| | heavy interrupt load |
| andw #ALLOWINT,%sr |
| |
| resume_userspace: |
| movel %curptr@(TASK_STACK),%a1 |
| moveb %a1@(TINFO_FLAGS+3),%d0 |
| jne exit_work |
| 1: RESTORE_ALL |
| |
| exit_work: |
| | save top of frame |
| movel %sp,%curptr@(TASK_THREAD+THREAD_ESP0) |
| lslb #1,%d0 |
| jne do_signal_return |
| pea resume_userspace |
| jra schedule |
| |
| |
| do_signal_return: |
| |andw #ALLOWINT,%sr |
| subql #4,%sp | dummy return address |
| SAVE_SWITCH_STACK |
| pea %sp@(SWITCH_STACK_SIZE) |
| bsrl do_notify_resume |
| addql #4,%sp |
| RESTORE_SWITCH_STACK |
| addql #4,%sp |
| jbra resume_userspace |
| |
| do_delayed_trace: |
| bclr #7,%sp@(PT_OFF_SR) | clear trace bit in SR |
| pea 1 | send SIGTRAP |
| movel %curptr,%sp@- |
| pea LSIGTRAP |
| jbsr send_sig |
| addql #8,%sp |
| addql #4,%sp |
| jbra resume_userspace |
| |
| |
| /* This is the main interrupt handler for autovector interrupts */ |
| |
| ENTRY(auto_inthandler) |
| SAVE_ALL_INT |
| GET_CURRENT(%d0) |
| | put exception # in d0 |
| bfextu %sp@(PT_OFF_FORMATVEC){#4,#10},%d0 |
| subw #VEC_SPUR,%d0 |
| |
| movel %sp,%sp@- |
| movel %d0,%sp@- | put vector # on stack |
| auto_irqhandler_fixup = . + 2 |
| jsr do_IRQ | process the IRQ |
| addql #8,%sp | pop parameters off stack |
| jra ret_from_exception |
| |
| /* Handler for user defined interrupt vectors */ |
| |
| ENTRY(user_inthandler) |
| SAVE_ALL_INT |
| GET_CURRENT(%d0) |
| | put exception # in d0 |
| bfextu %sp@(PT_OFF_FORMATVEC){#4,#10},%d0 |
| user_irqvec_fixup = . + 2 |
| subw #VEC_USER,%d0 |
| |
| movel %sp,%sp@- |
| movel %d0,%sp@- | put vector # on stack |
| jsr do_IRQ | process the IRQ |
| addql #8,%sp | pop parameters off stack |
| jra ret_from_exception |
| |
| /* Handler for uninitialized and spurious interrupts */ |
| |
| ENTRY(bad_inthandler) |
| SAVE_ALL_INT |
| GET_CURRENT(%d0) |
| |
| movel %sp,%sp@- |
| jsr handle_badint |
| addql #4,%sp |
| jra ret_from_exception |
| |
| resume: |
| /* |
| * Beware - when entering resume, prev (the current task) is |
| * in a0, next (the new task) is in a1,so don't change these |
| * registers until their contents are no longer needed. |
| */ |
| |
| /* save sr */ |
| movew %sr,%a0@(TASK_THREAD+THREAD_SR) |
| |
| /* save fs (sfc,%dfc) (may be pointing to kernel memory) */ |
| movec %sfc,%d0 |
| movew %d0,%a0@(TASK_THREAD+THREAD_FC) |
| |
| /* save usp */ |
| /* it is better to use a movel here instead of a movew 8*) */ |
| movec %usp,%d0 |
| movel %d0,%a0@(TASK_THREAD+THREAD_USP) |
| |
| /* save non-scratch registers on stack */ |
| SAVE_SWITCH_STACK |
| |
| /* save current kernel stack pointer */ |
| movel %sp,%a0@(TASK_THREAD+THREAD_KSP) |
| |
| /* save floating point context */ |
| #ifndef CONFIG_M68KFPU_EMU_ONLY |
| #ifdef CONFIG_M68KFPU_EMU |
| tstl m68k_fputype |
| jeq 3f |
| #endif |
| fsave %a0@(TASK_THREAD+THREAD_FPSTATE) |
| |
| #if defined(CONFIG_M68060) |
| #if !defined(CPU_M68060_ONLY) |
| btst #3,m68k_cputype+3 |
| beqs 1f |
| #endif |
| /* The 060 FPU keeps status in bits 15-8 of the first longword */ |
| tstb %a0@(TASK_THREAD+THREAD_FPSTATE+2) |
| jeq 3f |
| #if !defined(CPU_M68060_ONLY) |
| jra 2f |
| #endif |
| #endif /* CONFIG_M68060 */ |
| #if !defined(CPU_M68060_ONLY) |
| 1: tstb %a0@(TASK_THREAD+THREAD_FPSTATE) |
| jeq 3f |
| #endif |
| 2: fmovemx %fp0-%fp7,%a0@(TASK_THREAD+THREAD_FPREG) |
| fmoveml %fpcr/%fpsr/%fpiar,%a0@(TASK_THREAD+THREAD_FPCNTL) |
| 3: |
| #endif /* CONFIG_M68KFPU_EMU_ONLY */ |
| /* Return previous task in %d1 */ |
| movel %curptr,%d1 |
| |
| /* switch to new task (a1 contains new task) */ |
| movel %a1,%curptr |
| |
| /* restore floating point context */ |
| #ifndef CONFIG_M68KFPU_EMU_ONLY |
| #ifdef CONFIG_M68KFPU_EMU |
| tstl m68k_fputype |
| jeq 4f |
| #endif |
| #if defined(CONFIG_M68060) |
| #if !defined(CPU_M68060_ONLY) |
| btst #3,m68k_cputype+3 |
| beqs 1f |
| #endif |
| /* The 060 FPU keeps status in bits 15-8 of the first longword */ |
| tstb %a1@(TASK_THREAD+THREAD_FPSTATE+2) |
| jeq 3f |
| #if !defined(CPU_M68060_ONLY) |
| jra 2f |
| #endif |
| #endif /* CONFIG_M68060 */ |
| #if !defined(CPU_M68060_ONLY) |
| 1: tstb %a1@(TASK_THREAD+THREAD_FPSTATE) |
| jeq 3f |
| #endif |
| 2: fmovemx %a1@(TASK_THREAD+THREAD_FPREG),%fp0-%fp7 |
| fmoveml %a1@(TASK_THREAD+THREAD_FPCNTL),%fpcr/%fpsr/%fpiar |
| 3: frestore %a1@(TASK_THREAD+THREAD_FPSTATE) |
| 4: |
| #endif /* CONFIG_M68KFPU_EMU_ONLY */ |
| |
| /* restore the kernel stack pointer */ |
| movel %a1@(TASK_THREAD+THREAD_KSP),%sp |
| |
| /* restore non-scratch registers */ |
| RESTORE_SWITCH_STACK |
| |
| /* restore user stack pointer */ |
| movel %a1@(TASK_THREAD+THREAD_USP),%a0 |
| movel %a0,%usp |
| |
| /* restore fs (sfc,%dfc) */ |
| movew %a1@(TASK_THREAD+THREAD_FC),%a0 |
| movec %a0,%sfc |
| movec %a0,%dfc |
| |
| /* restore status register */ |
| movew %a1@(TASK_THREAD+THREAD_SR),%d0 |
| oriw #0x0700,%d0 |
| movew %d0,%sr |
| |
| rts |
| |
| #endif /* CONFIG_MMU && !CONFIG_COLDFIRE */ |