2007-08-02 15:19:14 +08:00
|
|
|
/* arch/sparc/kernel/entry.S: Sparc trap low-level entry points.
|
2005-04-17 06:20:36 +08:00
|
|
|
*
|
2007-08-02 15:19:14 +08:00
|
|
|
* Copyright (C) 1995, 2007 David S. Miller (davem@davemloft.net)
|
2005-04-17 06:20:36 +08:00
|
|
|
* Copyright (C) 1996 Eddie C. Dost (ecd@skynet.be)
|
|
|
|
* Copyright (C) 1996 Miguel de Icaza (miguel@nuclecu.unam.mx)
|
|
|
|
* Copyright (C) 1996-1999 Jakub Jelinek (jj@sunsite.mff.cuni.cz)
|
|
|
|
* Copyright (C) 1997 Anton Blanchard (anton@progsoc.uts.edu.au)
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/errno.h>
|
|
|
|
|
|
|
|
#include <asm/head.h>
|
|
|
|
#include <asm/asi.h>
|
|
|
|
#include <asm/smp.h>
|
|
|
|
#include <asm/contregs.h>
|
|
|
|
#include <asm/ptrace.h>
|
2005-09-10 02:35:55 +08:00
|
|
|
#include <asm/asm-offsets.h>
|
2005-04-17 06:20:36 +08:00
|
|
|
#include <asm/psr.h>
|
|
|
|
#include <asm/vaddrs.h>
|
|
|
|
#include <asm/memreg.h>
|
|
|
|
#include <asm/page.h>
|
2008-06-14 02:20:54 +08:00
|
|
|
#include <asm/pgtable.h>
|
2005-04-17 06:20:36 +08:00
|
|
|
#include <asm/pgtsun4c.h>
|
|
|
|
#include <asm/winmacro.h>
|
|
|
|
#include <asm/signal.h>
|
|
|
|
#include <asm/obio.h>
|
|
|
|
#include <asm/mxcc.h>
|
|
|
|
#include <asm/thread_info.h>
|
|
|
|
#include <asm/param.h>
|
2006-11-06 08:51:03 +08:00
|
|
|
#include <asm/unistd.h>
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
#include <asm/asmmacro.h>
|
|
|
|
|
|
|
|
#define curptr g6
|
|
|
|
|
|
|
|
/* These are just handy. */
|
|
|
|
#define _SV save %sp, -STACKFRAME_SZ, %sp
|
|
|
|
#define _RS restore
|
|
|
|
|
|
|
|
#define FLUSH_ALL_KERNEL_WINDOWS \
|
|
|
|
_SV; _SV; _SV; _SV; _SV; _SV; _SV; \
|
|
|
|
_RS; _RS; _RS; _RS; _RS; _RS; _RS;
|
|
|
|
|
|
|
|
.text
|
|
|
|
|
2008-04-29 17:38:50 +08:00
|
|
|
#ifdef CONFIG_KGDB
|
|
|
|
.align 4
|
|
|
|
.globl arch_kgdb_breakpoint
|
|
|
|
.type arch_kgdb_breakpoint,#function
|
|
|
|
arch_kgdb_breakpoint:
|
|
|
|
ta 0x7d
|
|
|
|
retl
|
|
|
|
nop
|
|
|
|
.size arch_kgdb_breakpoint,.-arch_kgdb_breakpoint
|
2005-04-17 06:20:36 +08:00
|
|
|
#endif
|
|
|
|
|
2007-08-02 15:19:14 +08:00
|
|
|
#if defined(CONFIG_BLK_DEV_FD) || defined(CONFIG_BLK_DEV_FD_MODULE)
|
2005-04-17 06:20:36 +08:00
|
|
|
.align 4
|
|
|
|
.globl floppy_hardint
|
|
|
|
floppy_hardint:
|
|
|
|
/*
|
|
|
|
* This code cannot touch registers %l0 %l1 and %l2
|
|
|
|
* because SAVE_ALL depends on their values. It depends
|
|
|
|
* on %l3 also, but we regenerate it before a call.
|
|
|
|
* Other registers are:
|
|
|
|
* %l3 -- base address of fdc registers
|
|
|
|
* %l4 -- pdma_vaddr
|
|
|
|
* %l5 -- scratch for ld/st address
|
|
|
|
* %l6 -- pdma_size
|
|
|
|
* %l7 -- scratch [floppy byte, ld/st address, aux. data]
|
|
|
|
*/
|
|
|
|
|
|
|
|
/* Do we have work to do? */
|
|
|
|
sethi %hi(doing_pdma), %l7
|
|
|
|
ld [%l7 + %lo(doing_pdma)], %l7
|
|
|
|
cmp %l7, 0
|
|
|
|
be floppy_dosoftint
|
|
|
|
nop
|
|
|
|
|
|
|
|
/* Load fdc register base */
|
|
|
|
sethi %hi(fdc_status), %l3
|
|
|
|
ld [%l3 + %lo(fdc_status)], %l3
|
|
|
|
|
|
|
|
/* Setup register addresses */
|
|
|
|
sethi %hi(pdma_vaddr), %l5 ! transfer buffer
|
|
|
|
ld [%l5 + %lo(pdma_vaddr)], %l4
|
|
|
|
sethi %hi(pdma_size), %l5 ! bytes to go
|
|
|
|
ld [%l5 + %lo(pdma_size)], %l6
|
|
|
|
next_byte:
|
|
|
|
ldub [%l3], %l7
|
|
|
|
|
|
|
|
andcc %l7, 0x80, %g0 ! Does fifo still have data
|
|
|
|
bz floppy_fifo_emptied ! fifo has been emptied...
|
|
|
|
andcc %l7, 0x20, %g0 ! in non-dma mode still?
|
|
|
|
bz floppy_overrun ! nope, overrun
|
|
|
|
andcc %l7, 0x40, %g0 ! 0=write 1=read
|
|
|
|
bz floppy_write
|
|
|
|
sub %l6, 0x1, %l6
|
|
|
|
|
|
|
|
/* Ok, actually read this byte */
|
|
|
|
ldub [%l3 + 1], %l7
|
|
|
|
orcc %g0, %l6, %g0
|
|
|
|
stb %l7, [%l4]
|
|
|
|
bne next_byte
|
|
|
|
add %l4, 0x1, %l4
|
|
|
|
|
|
|
|
b floppy_tdone
|
|
|
|
nop
|
|
|
|
|
|
|
|
floppy_write:
|
|
|
|
/* Ok, actually write this byte */
|
|
|
|
ldub [%l4], %l7
|
|
|
|
orcc %g0, %l6, %g0
|
|
|
|
stb %l7, [%l3 + 1]
|
|
|
|
bne next_byte
|
|
|
|
add %l4, 0x1, %l4
|
|
|
|
|
|
|
|
/* fall through... */
|
|
|
|
floppy_tdone:
|
|
|
|
sethi %hi(pdma_vaddr), %l5
|
|
|
|
st %l4, [%l5 + %lo(pdma_vaddr)]
|
|
|
|
sethi %hi(pdma_size), %l5
|
|
|
|
st %l6, [%l5 + %lo(pdma_size)]
|
|
|
|
/* Flip terminal count pin */
|
|
|
|
set auxio_register, %l7
|
|
|
|
ld [%l7], %l7
|
|
|
|
|
|
|
|
set sparc_cpu_model, %l5
|
|
|
|
ld [%l5], %l5
|
|
|
|
subcc %l5, 1, %g0 /* enum { sun4c = 1 }; */
|
|
|
|
be 1f
|
|
|
|
ldub [%l7], %l5
|
|
|
|
|
|
|
|
or %l5, 0xc2, %l5
|
|
|
|
stb %l5, [%l7]
|
|
|
|
andn %l5, 0x02, %l5
|
|
|
|
b 2f
|
|
|
|
nop
|
|
|
|
|
|
|
|
1:
|
|
|
|
or %l5, 0xf4, %l5
|
|
|
|
stb %l5, [%l7]
|
|
|
|
andn %l5, 0x04, %l5
|
|
|
|
|
|
|
|
2:
|
|
|
|
/* Kill some time so the bits set */
|
|
|
|
WRITE_PAUSE
|
|
|
|
WRITE_PAUSE
|
|
|
|
|
|
|
|
stb %l5, [%l7]
|
|
|
|
|
|
|
|
/* Prevent recursion */
|
|
|
|
sethi %hi(doing_pdma), %l7
|
|
|
|
b floppy_dosoftint
|
|
|
|
st %g0, [%l7 + %lo(doing_pdma)]
|
|
|
|
|
|
|
|
/* We emptied the FIFO, but we haven't read everything
|
|
|
|
* as of yet. Store the current transfer address and
|
|
|
|
* bytes left to read so we can continue when the next
|
|
|
|
* fast IRQ comes in.
|
|
|
|
*/
|
|
|
|
floppy_fifo_emptied:
|
|
|
|
sethi %hi(pdma_vaddr), %l5
|
|
|
|
st %l4, [%l5 + %lo(pdma_vaddr)]
|
|
|
|
sethi %hi(pdma_size), %l7
|
|
|
|
st %l6, [%l7 + %lo(pdma_size)]
|
|
|
|
|
|
|
|
/* Restore condition codes */
|
|
|
|
wr %l0, 0x0, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
|
|
|
|
jmp %l1
|
|
|
|
rett %l2
|
|
|
|
|
|
|
|
floppy_overrun:
|
|
|
|
sethi %hi(pdma_vaddr), %l5
|
|
|
|
st %l4, [%l5 + %lo(pdma_vaddr)]
|
|
|
|
sethi %hi(pdma_size), %l5
|
|
|
|
st %l6, [%l5 + %lo(pdma_size)]
|
|
|
|
/* Prevent recursion */
|
|
|
|
sethi %hi(doing_pdma), %l7
|
|
|
|
st %g0, [%l7 + %lo(doing_pdma)]
|
|
|
|
|
|
|
|
/* fall through... */
|
|
|
|
floppy_dosoftint:
|
|
|
|
rd %wim, %l3
|
|
|
|
SAVE_ALL
|
|
|
|
|
|
|
|
/* Set all IRQs off. */
|
|
|
|
or %l0, PSR_PIL, %l4
|
|
|
|
wr %l4, 0x0, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
wr %l4, PSR_ET, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
|
|
|
|
mov 11, %o0 ! floppy irq level (unused anyway)
|
|
|
|
mov %g0, %o1 ! devid is not used in fast interrupts
|
|
|
|
call sparc_floppy_irq
|
|
|
|
add %sp, STACKFRAME_SZ, %o2 ! struct pt_regs *regs
|
|
|
|
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
#endif /* (CONFIG_BLK_DEV_FD) */
|
|
|
|
|
|
|
|
/* Bad trap handler */
|
|
|
|
.globl bad_trap_handler
|
|
|
|
bad_trap_handler:
|
|
|
|
SAVE_ALL
|
|
|
|
|
|
|
|
wr %l0, PSR_ET, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
|
|
|
|
add %sp, STACKFRAME_SZ, %o0 ! pt_regs
|
|
|
|
call do_hw_interrupt
|
|
|
|
mov %l7, %o1 ! trap number
|
|
|
|
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
/* For now all IRQ's not registered get sent here. handler_irq() will
|
|
|
|
* see if a routine is registered to handle this interrupt and if not
|
|
|
|
* it will say so on the console.
|
|
|
|
*/
|
|
|
|
|
|
|
|
.align 4
|
|
|
|
.globl real_irq_entry, patch_handler_irq
|
|
|
|
real_irq_entry:
|
|
|
|
SAVE_ALL
|
|
|
|
|
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
.globl patchme_maybe_smp_msg
|
|
|
|
|
|
|
|
cmp %l7, 12
|
|
|
|
patchme_maybe_smp_msg:
|
|
|
|
bgu maybe_smp4m_msg
|
|
|
|
nop
|
|
|
|
#endif
|
|
|
|
|
|
|
|
real_irq_continue:
|
|
|
|
or %l0, PSR_PIL, %g2
|
|
|
|
wr %g2, 0x0, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
wr %g2, PSR_ET, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
mov %l7, %o0 ! irq level
|
|
|
|
patch_handler_irq:
|
|
|
|
call handler_irq
|
|
|
|
add %sp, STACKFRAME_SZ, %o1 ! pt_regs ptr
|
|
|
|
or %l0, PSR_PIL, %g2 ! restore PIL after handler_irq
|
|
|
|
wr %g2, PSR_ET, %psr ! keep ET up
|
|
|
|
WRITE_PAUSE
|
|
|
|
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
/* SMP per-cpu ticker interrupts are handled specially. */
|
|
|
|
smp4m_ticker:
|
|
|
|
bne real_irq_continue+4
|
|
|
|
or %l0, PSR_PIL, %g2
|
|
|
|
wr %g2, 0x0, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
wr %g2, PSR_ET, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
call smp4m_percpu_timer_interrupt
|
|
|
|
add %sp, STACKFRAME_SZ, %o0
|
|
|
|
wr %l0, PSR_ET, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
/* Here is where we check for possible SMP IPI passed to us
|
|
|
|
* on some level other than 15 which is the NMI and only used
|
|
|
|
* for cross calls. That has a separate entry point below.
|
|
|
|
*/
|
|
|
|
maybe_smp4m_msg:
|
|
|
|
GET_PROCESSOR4M_ID(o3)
|
2008-09-20 12:17:43 +08:00
|
|
|
sethi %hi(sun4m_irq_percpu), %l5
|
|
|
|
sll %o3, 2, %o3
|
|
|
|
or %l5, %lo(sun4m_irq_percpu), %o5
|
2005-04-17 06:20:36 +08:00
|
|
|
sethi %hi(0x40000000), %o2
|
|
|
|
ld [%o5 + %o3], %o1
|
2008-09-20 12:17:43 +08:00
|
|
|
ld [%o1 + 0x00], %o3 ! sun4m_irq_percpu[cpu]->pending
|
|
|
|
andcc %o3, %o2, %g0
|
2005-04-17 06:20:36 +08:00
|
|
|
be,a smp4m_ticker
|
|
|
|
cmp %l7, 14
|
2008-09-20 12:17:43 +08:00
|
|
|
st %o2, [%o1 + 0x04] ! sun4m_irq_percpu[cpu]->clear=0x40000000
|
2005-04-17 06:20:36 +08:00
|
|
|
WRITE_PAUSE
|
2008-09-20 12:17:43 +08:00
|
|
|
ld [%o1 + 0x00], %g0 ! sun4m_irq_percpu[cpu]->pending
|
2005-04-17 06:20:36 +08:00
|
|
|
WRITE_PAUSE
|
|
|
|
or %l0, PSR_PIL, %l4
|
|
|
|
wr %l4, 0x0, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
wr %l4, PSR_ET, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
call smp_reschedule_irq
|
|
|
|
nop
|
|
|
|
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
.align 4
|
|
|
|
.globl linux_trap_ipi15_sun4m
|
|
|
|
linux_trap_ipi15_sun4m:
|
|
|
|
SAVE_ALL
|
|
|
|
sethi %hi(0x80000000), %o2
|
|
|
|
GET_PROCESSOR4M_ID(o0)
|
2008-09-20 12:17:43 +08:00
|
|
|
sethi %hi(sun4m_irq_percpu), %l5
|
|
|
|
or %l5, %lo(sun4m_irq_percpu), %o5
|
|
|
|
sll %o0, 2, %o0
|
|
|
|
ld [%o5 + %o0], %o5
|
|
|
|
ld [%o5 + 0x00], %o3 ! sun4m_irq_percpu[cpu]->pending
|
2005-04-17 06:20:36 +08:00
|
|
|
andcc %o3, %o2, %g0
|
|
|
|
be 1f ! Must be an NMI async memory error
|
2008-09-20 12:17:43 +08:00
|
|
|
st %o2, [%o5 + 0x04] ! sun4m_irq_percpu[cpu]->clear=0x80000000
|
2005-04-17 06:20:36 +08:00
|
|
|
WRITE_PAUSE
|
2008-09-20 12:17:43 +08:00
|
|
|
ld [%o5 + 0x00], %g0 ! sun4m_irq_percpu[cpu]->pending
|
2005-04-17 06:20:36 +08:00
|
|
|
WRITE_PAUSE
|
|
|
|
or %l0, PSR_PIL, %l4
|
|
|
|
wr %l4, 0x0, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
wr %l4, PSR_ET, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
call smp4m_cross_call_irq
|
|
|
|
nop
|
|
|
|
b ret_trap_lockless_ipi
|
|
|
|
clr %l6
|
|
|
|
1:
|
|
|
|
/* NMI async memory error handling. */
|
|
|
|
sethi %hi(0x80000000), %l4
|
2008-09-20 12:17:43 +08:00
|
|
|
sethi %hi(sun4m_irq_global), %o5
|
|
|
|
ld [%o5 + %lo(sun4m_irq_global)], %l5
|
|
|
|
st %l4, [%l5 + 0x0c] ! sun4m_irq_global->mask_set=0x80000000
|
2005-04-17 06:20:36 +08:00
|
|
|
WRITE_PAUSE
|
2008-09-20 12:17:43 +08:00
|
|
|
ld [%l5 + 0x00], %g0 ! sun4m_irq_global->pending
|
2005-04-17 06:20:36 +08:00
|
|
|
WRITE_PAUSE
|
|
|
|
or %l0, PSR_PIL, %l4
|
|
|
|
wr %l4, 0x0, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
wr %l4, PSR_ET, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
call sun4m_nmi
|
|
|
|
nop
|
2008-09-20 12:17:43 +08:00
|
|
|
st %l4, [%l5 + 0x08] ! sun4m_irq_global->mask_clear=0x80000000
|
2005-04-17 06:20:36 +08:00
|
|
|
WRITE_PAUSE
|
2008-09-20 12:17:43 +08:00
|
|
|
ld [%l5 + 0x00], %g0 ! sun4m_irq_global->pending
|
2005-04-17 06:20:36 +08:00
|
|
|
WRITE_PAUSE
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
.globl smp4d_ticker
|
|
|
|
/* SMP per-cpu ticker interrupts are handled specially. */
|
|
|
|
smp4d_ticker:
|
|
|
|
SAVE_ALL
|
|
|
|
or %l0, PSR_PIL, %g2
|
|
|
|
sethi %hi(CC_ICLR), %o0
|
|
|
|
sethi %hi(1 << 14), %o1
|
|
|
|
or %o0, %lo(CC_ICLR), %o0
|
|
|
|
stha %o1, [%o0] ASI_M_MXCC /* Clear PIL 14 in MXCC's ICLR */
|
|
|
|
wr %g2, 0x0, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
wr %g2, PSR_ET, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
call smp4d_percpu_timer_interrupt
|
|
|
|
add %sp, STACKFRAME_SZ, %o0
|
|
|
|
wr %l0, PSR_ET, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
.align 4
|
|
|
|
.globl linux_trap_ipi15_sun4d
|
|
|
|
linux_trap_ipi15_sun4d:
|
|
|
|
SAVE_ALL
|
|
|
|
sethi %hi(CC_BASE), %o4
|
|
|
|
sethi %hi(MXCC_ERR_ME|MXCC_ERR_PEW|MXCC_ERR_ASE|MXCC_ERR_PEE), %o2
|
|
|
|
or %o4, (CC_EREG - CC_BASE), %o0
|
|
|
|
ldda [%o0] ASI_M_MXCC, %o0
|
|
|
|
andcc %o0, %o2, %g0
|
|
|
|
bne 1f
|
|
|
|
sethi %hi(BB_STAT2), %o2
|
|
|
|
lduba [%o2] ASI_M_CTL, %o2
|
|
|
|
andcc %o2, BB_STAT2_MASK, %g0
|
|
|
|
bne 2f
|
|
|
|
or %o4, (CC_ICLR - CC_BASE), %o0
|
|
|
|
sethi %hi(1 << 15), %o1
|
|
|
|
stha %o1, [%o0] ASI_M_MXCC /* Clear PIL 15 in MXCC's ICLR */
|
|
|
|
or %l0, PSR_PIL, %l4
|
|
|
|
wr %l4, 0x0, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
wr %l4, PSR_ET, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
call smp4d_cross_call_irq
|
|
|
|
nop
|
|
|
|
b ret_trap_lockless_ipi
|
|
|
|
clr %l6
|
|
|
|
|
|
|
|
1: /* MXCC error */
|
|
|
|
2: /* BB error */
|
|
|
|
/* Disable PIL 15 */
|
|
|
|
set CC_IMSK, %l4
|
|
|
|
lduha [%l4] ASI_M_MXCC, %l5
|
|
|
|
sethi %hi(1 << 15), %l7
|
|
|
|
or %l5, %l7, %l5
|
|
|
|
stha %l5, [%l4] ASI_M_MXCC
|
|
|
|
/* FIXME */
|
|
|
|
1: b,a 1b
|
|
|
|
|
2009-09-01 06:08:13 +08:00
|
|
|
#ifdef CONFIG_SPARC_LEON
|
|
|
|
|
|
|
|
.globl smpleon_ticker
|
|
|
|
/* SMP per-cpu ticker interrupts are handled specially. */
|
|
|
|
smpleon_ticker:
|
|
|
|
SAVE_ALL
|
|
|
|
or %l0, PSR_PIL, %g2
|
|
|
|
wr %g2, 0x0, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
wr %g2, PSR_ET, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
call leon_percpu_timer_interrupt
|
|
|
|
add %sp, STACKFRAME_SZ, %o0
|
|
|
|
wr %l0, PSR_ET, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
.align 4
|
|
|
|
.globl linux_trap_ipi15_leon
|
|
|
|
linux_trap_ipi15_leon:
|
|
|
|
SAVE_ALL
|
|
|
|
or %l0, PSR_PIL, %l4
|
|
|
|
wr %l4, 0x0, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
wr %l4, PSR_ET, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
call leon_cross_call_irq
|
|
|
|
nop
|
|
|
|
b ret_trap_lockless_ipi
|
|
|
|
clr %l6
|
|
|
|
|
|
|
|
#endif /* CONFIG_SPARC_LEON */
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
#endif /* CONFIG_SMP */
|
|
|
|
|
|
|
|
/* This routine handles illegal instructions and privileged
|
|
|
|
* instruction attempts from user code.
|
|
|
|
*/
|
|
|
|
.align 4
|
|
|
|
.globl bad_instruction
|
|
|
|
bad_instruction:
|
|
|
|
sethi %hi(0xc1f80000), %l4
|
|
|
|
ld [%l1], %l5
|
|
|
|
sethi %hi(0x81d80000), %l7
|
|
|
|
and %l5, %l4, %l5
|
|
|
|
cmp %l5, %l7
|
|
|
|
be 1f
|
|
|
|
SAVE_ALL
|
|
|
|
|
|
|
|
wr %l0, PSR_ET, %psr ! re-enable traps
|
|
|
|
WRITE_PAUSE
|
|
|
|
|
|
|
|
add %sp, STACKFRAME_SZ, %o0
|
|
|
|
mov %l1, %o1
|
|
|
|
mov %l2, %o2
|
|
|
|
call do_illegal_instruction
|
|
|
|
mov %l0, %o3
|
|
|
|
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
1: /* unimplemented flush - just skip */
|
|
|
|
jmpl %l2, %g0
|
|
|
|
rett %l2 + 4
|
|
|
|
|
|
|
|
.align 4
|
|
|
|
.globl priv_instruction
|
|
|
|
priv_instruction:
|
|
|
|
SAVE_ALL
|
|
|
|
|
|
|
|
wr %l0, PSR_ET, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
|
|
|
|
add %sp, STACKFRAME_SZ, %o0
|
|
|
|
mov %l1, %o1
|
|
|
|
mov %l2, %o2
|
|
|
|
call do_priv_instruction
|
|
|
|
mov %l0, %o3
|
|
|
|
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
/* This routine handles unaligned data accesses. */
|
|
|
|
.align 4
|
|
|
|
.globl mna_handler
|
|
|
|
mna_handler:
|
|
|
|
andcc %l0, PSR_PS, %g0
|
|
|
|
be mna_fromuser
|
|
|
|
nop
|
|
|
|
|
|
|
|
SAVE_ALL
|
|
|
|
|
|
|
|
wr %l0, PSR_ET, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
|
|
|
|
ld [%l1], %o1
|
|
|
|
call kernel_unaligned_trap
|
|
|
|
add %sp, STACKFRAME_SZ, %o0
|
|
|
|
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
mna_fromuser:
|
|
|
|
SAVE_ALL
|
|
|
|
|
|
|
|
wr %l0, PSR_ET, %psr ! re-enable traps
|
|
|
|
WRITE_PAUSE
|
|
|
|
|
|
|
|
ld [%l1], %o1
|
|
|
|
call user_unaligned_trap
|
|
|
|
add %sp, STACKFRAME_SZ, %o0
|
|
|
|
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
/* This routine handles floating point disabled traps. */
|
|
|
|
.align 4
|
|
|
|
.globl fpd_trap_handler
|
|
|
|
fpd_trap_handler:
|
|
|
|
SAVE_ALL
|
|
|
|
|
|
|
|
wr %l0, PSR_ET, %psr ! re-enable traps
|
|
|
|
WRITE_PAUSE
|
|
|
|
|
|
|
|
add %sp, STACKFRAME_SZ, %o0
|
|
|
|
mov %l1, %o1
|
|
|
|
mov %l2, %o2
|
|
|
|
call do_fpd_trap
|
|
|
|
mov %l0, %o3
|
|
|
|
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
/* This routine handles Floating Point Exceptions. */
|
|
|
|
.align 4
|
|
|
|
.globl fpe_trap_handler
|
|
|
|
fpe_trap_handler:
|
|
|
|
set fpsave_magic, %l5
|
|
|
|
cmp %l1, %l5
|
|
|
|
be 1f
|
|
|
|
sethi %hi(fpsave), %l5
|
|
|
|
or %l5, %lo(fpsave), %l5
|
|
|
|
cmp %l1, %l5
|
|
|
|
bne 2f
|
|
|
|
sethi %hi(fpsave_catch2), %l5
|
|
|
|
or %l5, %lo(fpsave_catch2), %l5
|
|
|
|
wr %l0, 0x0, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
jmp %l5
|
|
|
|
rett %l5 + 4
|
|
|
|
1:
|
|
|
|
sethi %hi(fpsave_catch), %l5
|
|
|
|
or %l5, %lo(fpsave_catch), %l5
|
|
|
|
wr %l0, 0x0, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
jmp %l5
|
|
|
|
rett %l5 + 4
|
|
|
|
|
|
|
|
2:
|
|
|
|
SAVE_ALL
|
|
|
|
|
|
|
|
wr %l0, PSR_ET, %psr ! re-enable traps
|
|
|
|
WRITE_PAUSE
|
|
|
|
|
|
|
|
add %sp, STACKFRAME_SZ, %o0
|
|
|
|
mov %l1, %o1
|
|
|
|
mov %l2, %o2
|
|
|
|
call do_fpe_trap
|
|
|
|
mov %l0, %o3
|
|
|
|
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
/* This routine handles Tag Overflow Exceptions. */
|
|
|
|
.align 4
|
|
|
|
.globl do_tag_overflow
|
|
|
|
do_tag_overflow:
|
|
|
|
SAVE_ALL
|
|
|
|
|
|
|
|
wr %l0, PSR_ET, %psr ! re-enable traps
|
|
|
|
WRITE_PAUSE
|
|
|
|
|
|
|
|
add %sp, STACKFRAME_SZ, %o0
|
|
|
|
mov %l1, %o1
|
|
|
|
mov %l2, %o2
|
|
|
|
call handle_tag_overflow
|
|
|
|
mov %l0, %o3
|
|
|
|
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
/* This routine handles Watchpoint Exceptions. */
|
|
|
|
.align 4
|
|
|
|
.globl do_watchpoint
|
|
|
|
do_watchpoint:
|
|
|
|
SAVE_ALL
|
|
|
|
|
|
|
|
wr %l0, PSR_ET, %psr ! re-enable traps
|
|
|
|
WRITE_PAUSE
|
|
|
|
|
|
|
|
add %sp, STACKFRAME_SZ, %o0
|
|
|
|
mov %l1, %o1
|
|
|
|
mov %l2, %o2
|
|
|
|
call handle_watchpoint
|
|
|
|
mov %l0, %o3
|
|
|
|
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
/* This routine handles Register Access Exceptions. */
|
|
|
|
.align 4
|
|
|
|
.globl do_reg_access
|
|
|
|
do_reg_access:
|
|
|
|
SAVE_ALL
|
|
|
|
|
|
|
|
wr %l0, PSR_ET, %psr ! re-enable traps
|
|
|
|
WRITE_PAUSE
|
|
|
|
|
|
|
|
add %sp, STACKFRAME_SZ, %o0
|
|
|
|
mov %l1, %o1
|
|
|
|
mov %l2, %o2
|
|
|
|
call handle_reg_access
|
|
|
|
mov %l0, %o3
|
|
|
|
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
/* This routine handles Co-Processor Disabled Exceptions. */
|
|
|
|
.align 4
|
|
|
|
.globl do_cp_disabled
|
|
|
|
do_cp_disabled:
|
|
|
|
SAVE_ALL
|
|
|
|
|
|
|
|
wr %l0, PSR_ET, %psr ! re-enable traps
|
|
|
|
WRITE_PAUSE
|
|
|
|
|
|
|
|
add %sp, STACKFRAME_SZ, %o0
|
|
|
|
mov %l1, %o1
|
|
|
|
mov %l2, %o2
|
|
|
|
call handle_cp_disabled
|
|
|
|
mov %l0, %o3
|
|
|
|
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
/* This routine handles Co-Processor Exceptions. */
|
|
|
|
.align 4
|
|
|
|
.globl do_cp_exception
|
|
|
|
do_cp_exception:
|
|
|
|
SAVE_ALL
|
|
|
|
|
|
|
|
wr %l0, PSR_ET, %psr ! re-enable traps
|
|
|
|
WRITE_PAUSE
|
|
|
|
|
|
|
|
add %sp, STACKFRAME_SZ, %o0
|
|
|
|
mov %l1, %o1
|
|
|
|
mov %l2, %o2
|
|
|
|
call handle_cp_exception
|
|
|
|
mov %l0, %o3
|
|
|
|
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
/* This routine handles Hardware Divide By Zero Exceptions. */
|
|
|
|
.align 4
|
|
|
|
.globl do_hw_divzero
|
|
|
|
do_hw_divzero:
|
|
|
|
SAVE_ALL
|
|
|
|
|
|
|
|
wr %l0, PSR_ET, %psr ! re-enable traps
|
|
|
|
WRITE_PAUSE
|
|
|
|
|
|
|
|
add %sp, STACKFRAME_SZ, %o0
|
|
|
|
mov %l1, %o1
|
|
|
|
mov %l2, %o2
|
|
|
|
call handle_hw_divzero
|
|
|
|
mov %l0, %o3
|
|
|
|
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
.align 4
|
|
|
|
.globl do_flush_windows
|
|
|
|
do_flush_windows:
|
|
|
|
SAVE_ALL
|
|
|
|
|
|
|
|
wr %l0, PSR_ET, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
|
|
|
|
andcc %l0, PSR_PS, %g0
|
|
|
|
bne dfw_kernel
|
|
|
|
nop
|
|
|
|
|
|
|
|
call flush_user_windows
|
|
|
|
nop
|
|
|
|
|
|
|
|
/* Advance over the trap instruction. */
|
|
|
|
ld [%sp + STACKFRAME_SZ + PT_NPC], %l1
|
|
|
|
add %l1, 0x4, %l2
|
|
|
|
st %l1, [%sp + STACKFRAME_SZ + PT_PC]
|
|
|
|
st %l2, [%sp + STACKFRAME_SZ + PT_NPC]
|
|
|
|
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
.globl flush_patch_one
|
|
|
|
|
|
|
|
/* We get these for debugging routines using __builtin_return_address() */
|
|
|
|
dfw_kernel:
|
|
|
|
flush_patch_one:
|
|
|
|
FLUSH_ALL_KERNEL_WINDOWS
|
|
|
|
|
|
|
|
/* Advance over the trap instruction. */
|
|
|
|
ld [%sp + STACKFRAME_SZ + PT_NPC], %l1
|
|
|
|
add %l1, 0x4, %l2
|
|
|
|
st %l1, [%sp + STACKFRAME_SZ + PT_PC]
|
|
|
|
st %l2, [%sp + STACKFRAME_SZ + PT_NPC]
|
|
|
|
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
/* The getcc software trap. The user wants the condition codes from
|
|
|
|
* the %psr in register %g1.
|
|
|
|
*/
|
|
|
|
|
|
|
|
.align 4
|
|
|
|
.globl getcc_trap_handler
|
|
|
|
getcc_trap_handler:
|
|
|
|
srl %l0, 20, %g1 ! give user
|
|
|
|
and %g1, 0xf, %g1 ! only ICC bits in %psr
|
|
|
|
jmp %l2 ! advance over trap instruction
|
|
|
|
rett %l2 + 0x4 ! like this...
|
|
|
|
|
|
|
|
/* The setcc software trap. The user has condition codes in %g1
|
|
|
|
* that it would like placed in the %psr. Be careful not to flip
|
|
|
|
* any unintentional bits!
|
|
|
|
*/
|
|
|
|
|
|
|
|
.align 4
|
|
|
|
.globl setcc_trap_handler
|
|
|
|
setcc_trap_handler:
|
|
|
|
sll %g1, 0x14, %l4
|
|
|
|
set PSR_ICC, %l5
|
|
|
|
andn %l0, %l5, %l0 ! clear ICC bits in %psr
|
|
|
|
and %l4, %l5, %l4 ! clear non-ICC bits in user value
|
|
|
|
or %l4, %l0, %l4 ! or them in... mix mix mix
|
|
|
|
|
|
|
|
wr %l4, 0x0, %psr ! set new %psr
|
|
|
|
WRITE_PAUSE ! TI scumbags...
|
|
|
|
|
|
|
|
jmp %l2 ! advance over trap instruction
|
|
|
|
rett %l2 + 0x4 ! like this...
|
|
|
|
|
|
|
|
.align 4
|
|
|
|
.globl linux_trap_nmi_sun4c
|
|
|
|
linux_trap_nmi_sun4c:
|
|
|
|
SAVE_ALL
|
|
|
|
|
|
|
|
/* Ugh, we need to clear the IRQ line. This is now
|
|
|
|
* a very sun4c specific trap handler...
|
|
|
|
*/
|
|
|
|
sethi %hi(interrupt_enable), %l5
|
|
|
|
ld [%l5 + %lo(interrupt_enable)], %l5
|
|
|
|
ldub [%l5], %l6
|
|
|
|
andn %l6, INTS_ENAB, %l6
|
|
|
|
stb %l6, [%l5]
|
|
|
|
|
|
|
|
/* Now it is safe to re-enable traps without recursion. */
|
|
|
|
or %l0, PSR_PIL, %l0
|
|
|
|
wr %l0, PSR_ET, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
|
|
|
|
/* Now call the c-code with the pt_regs frame ptr and the
|
|
|
|
* memory error registers as arguments. The ordering chosen
|
|
|
|
* here is due to unlatching semantics.
|
|
|
|
*/
|
|
|
|
sethi %hi(AC_SYNC_ERR), %o0
|
|
|
|
add %o0, 0x4, %o0
|
|
|
|
lda [%o0] ASI_CONTROL, %o2 ! sync vaddr
|
|
|
|
sub %o0, 0x4, %o0
|
|
|
|
lda [%o0] ASI_CONTROL, %o1 ! sync error
|
|
|
|
add %o0, 0xc, %o0
|
|
|
|
lda [%o0] ASI_CONTROL, %o4 ! async vaddr
|
|
|
|
sub %o0, 0x4, %o0
|
|
|
|
lda [%o0] ASI_CONTROL, %o3 ! async error
|
|
|
|
call sparc_lvl15_nmi
|
|
|
|
add %sp, STACKFRAME_SZ, %o0
|
|
|
|
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
.align 4
|
|
|
|
.globl invalid_segment_patch1_ff
|
|
|
|
.globl invalid_segment_patch2_ff
|
|
|
|
invalid_segment_patch1_ff: cmp %l4, 0xff
|
|
|
|
invalid_segment_patch2_ff: mov 0xff, %l3
|
|
|
|
|
|
|
|
.align 4
|
|
|
|
.globl invalid_segment_patch1_1ff
|
|
|
|
.globl invalid_segment_patch2_1ff
|
|
|
|
invalid_segment_patch1_1ff: cmp %l4, 0x1ff
|
|
|
|
invalid_segment_patch2_1ff: mov 0x1ff, %l3
|
|
|
|
|
|
|
|
.align 4
|
|
|
|
.globl num_context_patch1_16, num_context_patch2_16
|
|
|
|
num_context_patch1_16: mov 0x10, %l7
|
|
|
|
num_context_patch2_16: mov 0x10, %l7
|
|
|
|
|
|
|
|
.align 4
|
|
|
|
.globl vac_linesize_patch_32
|
|
|
|
vac_linesize_patch_32: subcc %l7, 32, %l7
|
|
|
|
|
|
|
|
.align 4
|
|
|
|
.globl vac_hwflush_patch1_on, vac_hwflush_patch2_on
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Ugly, but we cant use hardware flushing on the sun4 and we'd require
|
|
|
|
* two instructions (Anton)
|
|
|
|
*/
|
|
|
|
vac_hwflush_patch1_on: addcc %l7, -PAGE_SIZE, %l7
|
|
|
|
|
|
|
|
vac_hwflush_patch2_on: sta %g0, [%l3 + %l7] ASI_HWFLUSHSEG
|
|
|
|
|
|
|
|
.globl invalid_segment_patch1, invalid_segment_patch2
|
|
|
|
.globl num_context_patch1
|
|
|
|
.globl vac_linesize_patch, vac_hwflush_patch1
|
|
|
|
.globl vac_hwflush_patch2
|
|
|
|
|
|
|
|
.align 4
|
|
|
|
.globl sun4c_fault
|
|
|
|
|
|
|
|
! %l0 = %psr
|
|
|
|
! %l1 = %pc
|
|
|
|
! %l2 = %npc
|
|
|
|
! %l3 = %wim
|
|
|
|
! %l7 = 1 for textfault
|
|
|
|
! We want error in %l5, vaddr in %l6
|
|
|
|
sun4c_fault:
|
|
|
|
sethi %hi(AC_SYNC_ERR), %l4
|
|
|
|
add %l4, 0x4, %l6 ! AC_SYNC_VA in %l6
|
|
|
|
lda [%l6] ASI_CONTROL, %l5 ! Address
|
|
|
|
lda [%l4] ASI_CONTROL, %l6 ! Error, retained for a bit
|
|
|
|
|
|
|
|
andn %l5, 0xfff, %l5 ! Encode all info into l7
|
|
|
|
srl %l6, 14, %l4
|
|
|
|
|
|
|
|
and %l4, 2, %l4
|
|
|
|
or %l5, %l4, %l4
|
|
|
|
|
|
|
|
or %l4, %l7, %l7 ! l7 = [addr,write,txtfault]
|
|
|
|
|
|
|
|
andcc %l0, PSR_PS, %g0
|
|
|
|
be sun4c_fault_fromuser
|
|
|
|
andcc %l7, 1, %g0 ! Text fault?
|
|
|
|
|
|
|
|
be 1f
|
|
|
|
sethi %hi(KERNBASE), %l4
|
|
|
|
|
|
|
|
mov %l1, %l5 ! PC
|
|
|
|
|
|
|
|
1:
|
|
|
|
cmp %l5, %l4
|
|
|
|
blu sun4c_fault_fromuser
|
|
|
|
sethi %hi(~((1 << SUN4C_REAL_PGDIR_SHIFT) - 1)), %l4
|
|
|
|
|
|
|
|
/* If the kernel references a bum kernel pointer, or a pte which
|
|
|
|
* points to a non existant page in ram, we will run this code
|
|
|
|
* _forever_ and lock up the machine!!!!! So we must check for
|
|
|
|
* this condition, the AC_SYNC_ERR bits are what we must examine.
|
|
|
|
* Also a parity error would make this happen as well. So we just
|
|
|
|
* check that we are in fact servicing a tlb miss and not some
|
|
|
|
* other type of fault for the kernel.
|
|
|
|
*/
|
|
|
|
andcc %l6, 0x80, %g0
|
|
|
|
be sun4c_fault_fromuser
|
|
|
|
and %l5, %l4, %l5
|
|
|
|
|
|
|
|
/* Test for NULL pte_t * in vmalloc area. */
|
|
|
|
sethi %hi(VMALLOC_START), %l4
|
|
|
|
cmp %l5, %l4
|
|
|
|
blu,a invalid_segment_patch1
|
|
|
|
lduXa [%l5] ASI_SEGMAP, %l4
|
|
|
|
|
|
|
|
sethi %hi(swapper_pg_dir), %l4
|
|
|
|
srl %l5, SUN4C_PGDIR_SHIFT, %l6
|
|
|
|
or %l4, %lo(swapper_pg_dir), %l4
|
|
|
|
sll %l6, 2, %l6
|
|
|
|
ld [%l4 + %l6], %l4
|
|
|
|
andcc %l4, PAGE_MASK, %g0
|
|
|
|
be sun4c_fault_fromuser
|
|
|
|
lduXa [%l5] ASI_SEGMAP, %l4
|
|
|
|
|
|
|
|
invalid_segment_patch1:
|
|
|
|
cmp %l4, 0x7f
|
|
|
|
bne 1f
|
|
|
|
sethi %hi(sun4c_kfree_ring), %l4
|
|
|
|
or %l4, %lo(sun4c_kfree_ring), %l4
|
|
|
|
ld [%l4 + 0x18], %l3
|
|
|
|
deccc %l3 ! do we have a free entry?
|
|
|
|
bcs,a 2f ! no, unmap one.
|
|
|
|
sethi %hi(sun4c_kernel_ring), %l4
|
|
|
|
|
|
|
|
st %l3, [%l4 + 0x18] ! sun4c_kfree_ring.num_entries--
|
|
|
|
|
|
|
|
ld [%l4 + 0x00], %l6 ! entry = sun4c_kfree_ring.ringhd.next
|
|
|
|
st %l5, [%l6 + 0x08] ! entry->vaddr = address
|
|
|
|
|
|
|
|
ld [%l6 + 0x00], %l3 ! next = entry->next
|
|
|
|
ld [%l6 + 0x04], %l7 ! entry->prev
|
|
|
|
|
|
|
|
st %l7, [%l3 + 0x04] ! next->prev = entry->prev
|
|
|
|
st %l3, [%l7 + 0x00] ! entry->prev->next = next
|
|
|
|
|
|
|
|
sethi %hi(sun4c_kernel_ring), %l4
|
|
|
|
or %l4, %lo(sun4c_kernel_ring), %l4
|
|
|
|
! head = &sun4c_kernel_ring.ringhd
|
|
|
|
|
|
|
|
ld [%l4 + 0x00], %l7 ! head->next
|
|
|
|
|
|
|
|
st %l4, [%l6 + 0x04] ! entry->prev = head
|
|
|
|
st %l7, [%l6 + 0x00] ! entry->next = head->next
|
|
|
|
st %l6, [%l7 + 0x04] ! head->next->prev = entry
|
|
|
|
|
|
|
|
st %l6, [%l4 + 0x00] ! head->next = entry
|
|
|
|
|
|
|
|
ld [%l4 + 0x18], %l3
|
|
|
|
inc %l3 ! sun4c_kernel_ring.num_entries++
|
|
|
|
st %l3, [%l4 + 0x18]
|
|
|
|
b 4f
|
|
|
|
ld [%l6 + 0x08], %l5
|
|
|
|
|
|
|
|
2:
|
|
|
|
or %l4, %lo(sun4c_kernel_ring), %l4
|
|
|
|
! head = &sun4c_kernel_ring.ringhd
|
|
|
|
|
|
|
|
ld [%l4 + 0x04], %l6 ! entry = head->prev
|
|
|
|
|
|
|
|
ld [%l6 + 0x08], %l3 ! tmp = entry->vaddr
|
|
|
|
|
|
|
|
! Flush segment from the cache.
|
|
|
|
sethi %hi((64 * 1024)), %l7
|
|
|
|
9:
|
|
|
|
vac_hwflush_patch1:
|
|
|
|
vac_linesize_patch:
|
|
|
|
subcc %l7, 16, %l7
|
|
|
|
bne 9b
|
|
|
|
vac_hwflush_patch2:
|
|
|
|
sta %g0, [%l3 + %l7] ASI_FLUSHSEG
|
|
|
|
|
|
|
|
st %l5, [%l6 + 0x08] ! entry->vaddr = address
|
|
|
|
|
|
|
|
ld [%l6 + 0x00], %l5 ! next = entry->next
|
|
|
|
ld [%l6 + 0x04], %l7 ! entry->prev
|
|
|
|
|
|
|
|
st %l7, [%l5 + 0x04] ! next->prev = entry->prev
|
|
|
|
st %l5, [%l7 + 0x00] ! entry->prev->next = next
|
|
|
|
st %l4, [%l6 + 0x04] ! entry->prev = head
|
|
|
|
|
|
|
|
ld [%l4 + 0x00], %l7 ! head->next
|
|
|
|
|
|
|
|
st %l7, [%l6 + 0x00] ! entry->next = head->next
|
|
|
|
st %l6, [%l7 + 0x04] ! head->next->prev = entry
|
|
|
|
st %l6, [%l4 + 0x00] ! head->next = entry
|
|
|
|
|
|
|
|
mov %l3, %l5 ! address = tmp
|
|
|
|
|
|
|
|
4:
|
|
|
|
num_context_patch1:
|
|
|
|
mov 0x08, %l7
|
|
|
|
|
|
|
|
ld [%l6 + 0x08], %l4
|
|
|
|
ldub [%l6 + 0x0c], %l3
|
|
|
|
or %l4, %l3, %l4 ! encode new vaddr/pseg into l4
|
|
|
|
|
|
|
|
sethi %hi(AC_CONTEXT), %l3
|
|
|
|
lduba [%l3] ASI_CONTROL, %l6
|
|
|
|
|
|
|
|
/* Invalidate old mapping, instantiate new mapping,
|
|
|
|
* for each context. Registers l6/l7 are live across
|
|
|
|
* this loop.
|
|
|
|
*/
|
|
|
|
3: deccc %l7
|
|
|
|
sethi %hi(AC_CONTEXT), %l3
|
|
|
|
stba %l7, [%l3] ASI_CONTROL
|
|
|
|
invalid_segment_patch2:
|
|
|
|
mov 0x7f, %l3
|
|
|
|
stXa %l3, [%l5] ASI_SEGMAP
|
|
|
|
andn %l4, 0x1ff, %l3
|
|
|
|
bne 3b
|
|
|
|
stXa %l4, [%l3] ASI_SEGMAP
|
|
|
|
|
|
|
|
sethi %hi(AC_CONTEXT), %l3
|
|
|
|
stba %l6, [%l3] ASI_CONTROL
|
|
|
|
|
|
|
|
andn %l4, 0x1ff, %l5
|
|
|
|
|
|
|
|
1:
|
|
|
|
sethi %hi(VMALLOC_START), %l4
|
|
|
|
cmp %l5, %l4
|
|
|
|
|
|
|
|
bgeu 1f
|
|
|
|
mov 1 << (SUN4C_REAL_PGDIR_SHIFT - PAGE_SHIFT), %l7
|
|
|
|
|
|
|
|
sethi %hi(KERNBASE), %l6
|
|
|
|
|
|
|
|
sub %l5, %l6, %l4
|
|
|
|
srl %l4, PAGE_SHIFT, %l4
|
|
|
|
sethi %hi((SUN4C_PAGE_KERNEL & 0xf4000000)), %l3
|
|
|
|
or %l3, %l4, %l3
|
|
|
|
|
|
|
|
sethi %hi(PAGE_SIZE), %l4
|
|
|
|
|
|
|
|
2:
|
|
|
|
sta %l3, [%l5] ASI_PTE
|
|
|
|
deccc %l7
|
|
|
|
inc %l3
|
|
|
|
bne 2b
|
|
|
|
add %l5, %l4, %l5
|
|
|
|
|
|
|
|
b 7f
|
|
|
|
sethi %hi(sun4c_kernel_faults), %l4
|
|
|
|
|
|
|
|
1:
|
|
|
|
srl %l5, SUN4C_PGDIR_SHIFT, %l3
|
|
|
|
sethi %hi(swapper_pg_dir), %l4
|
|
|
|
or %l4, %lo(swapper_pg_dir), %l4
|
|
|
|
sll %l3, 2, %l3
|
|
|
|
ld [%l4 + %l3], %l4
|
|
|
|
and %l4, PAGE_MASK, %l4
|
|
|
|
|
|
|
|
srl %l5, (PAGE_SHIFT - 2), %l6
|
|
|
|
and %l6, ((SUN4C_PTRS_PER_PTE - 1) << 2), %l6
|
|
|
|
add %l6, %l4, %l6
|
|
|
|
|
|
|
|
sethi %hi(PAGE_SIZE), %l4
|
|
|
|
|
|
|
|
2:
|
|
|
|
ld [%l6], %l3
|
|
|
|
deccc %l7
|
|
|
|
sta %l3, [%l5] ASI_PTE
|
|
|
|
add %l6, 0x4, %l6
|
|
|
|
bne 2b
|
|
|
|
add %l5, %l4, %l5
|
|
|
|
|
|
|
|
sethi %hi(sun4c_kernel_faults), %l4
|
|
|
|
7:
|
|
|
|
ld [%l4 + %lo(sun4c_kernel_faults)], %l3
|
|
|
|
inc %l3
|
|
|
|
st %l3, [%l4 + %lo(sun4c_kernel_faults)]
|
|
|
|
|
|
|
|
/* Restore condition codes */
|
|
|
|
wr %l0, 0x0, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
jmp %l1
|
|
|
|
rett %l2
|
|
|
|
|
|
|
|
sun4c_fault_fromuser:
|
|
|
|
SAVE_ALL
|
|
|
|
nop
|
|
|
|
|
|
|
|
mov %l7, %o1 ! Decode the info from %l7
|
|
|
|
mov %l7, %o2
|
|
|
|
and %o1, 1, %o1 ! arg2 = text_faultp
|
|
|
|
mov %l7, %o3
|
|
|
|
and %o2, 2, %o2 ! arg3 = writep
|
|
|
|
andn %o3, 0xfff, %o3 ! arg4 = faulting address
|
|
|
|
|
|
|
|
wr %l0, PSR_ET, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
|
|
|
|
call do_sun4c_fault
|
|
|
|
add %sp, STACKFRAME_SZ, %o0 ! arg1 = pt_regs ptr
|
|
|
|
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
.align 4
|
|
|
|
.globl srmmu_fault
|
|
|
|
srmmu_fault:
|
|
|
|
mov 0x400, %l5
|
|
|
|
mov 0x300, %l4
|
|
|
|
|
|
|
|
lda [%l5] ASI_M_MMUREGS, %l6 ! read sfar first
|
|
|
|
lda [%l4] ASI_M_MMUREGS, %l5 ! read sfsr last
|
|
|
|
|
|
|
|
andn %l6, 0xfff, %l6
|
|
|
|
srl %l5, 6, %l5 ! and encode all info into l7
|
|
|
|
|
|
|
|
and %l5, 2, %l5
|
|
|
|
or %l5, %l6, %l6
|
|
|
|
|
|
|
|
or %l6, %l7, %l7 ! l7 = [addr,write,txtfault]
|
|
|
|
|
|
|
|
SAVE_ALL
|
|
|
|
|
|
|
|
mov %l7, %o1
|
|
|
|
mov %l7, %o2
|
|
|
|
and %o1, 1, %o1 ! arg2 = text_faultp
|
|
|
|
mov %l7, %o3
|
|
|
|
and %o2, 2, %o2 ! arg3 = writep
|
|
|
|
andn %o3, 0xfff, %o3 ! arg4 = faulting address
|
|
|
|
|
|
|
|
wr %l0, PSR_ET, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
|
|
|
|
call do_sparc_fault
|
|
|
|
add %sp, STACKFRAME_SZ, %o0 ! arg1 = pt_regs ptr
|
|
|
|
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
.align 4
|
|
|
|
.globl sys_nis_syscall
|
|
|
|
sys_nis_syscall:
|
|
|
|
mov %o7, %l5
|
|
|
|
add %sp, STACKFRAME_SZ, %o0 ! pt_regs *regs arg
|
|
|
|
call c_sys_nis_syscall
|
|
|
|
mov %l5, %o7
|
|
|
|
|
|
|
|
.align 4
|
|
|
|
.globl sys_execve
|
|
|
|
sys_execve:
|
|
|
|
mov %o7, %l5
|
|
|
|
add %sp, STACKFRAME_SZ, %o0 ! pt_regs *regs arg
|
|
|
|
call sparc_execve
|
|
|
|
mov %l5, %o7
|
|
|
|
|
2008-04-20 17:14:23 +08:00
|
|
|
.globl sunos_execv
|
|
|
|
sunos_execv:
|
|
|
|
st %g0, [%sp + STACKFRAME_SZ + PT_I2]
|
|
|
|
|
|
|
|
call sparc_execve
|
|
|
|
add %sp, STACKFRAME_SZ, %o0
|
|
|
|
|
|
|
|
b ret_sys_call
|
|
|
|
ld [%sp + STACKFRAME_SZ + PT_I0], %o0
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
.align 4
|
2009-01-14 21:13:56 +08:00
|
|
|
.globl sys_sparc_pipe
|
|
|
|
sys_sparc_pipe:
|
2005-04-17 06:20:36 +08:00
|
|
|
mov %o7, %l5
|
|
|
|
add %sp, STACKFRAME_SZ, %o0 ! pt_regs *regs arg
|
|
|
|
call sparc_pipe
|
|
|
|
mov %l5, %o7
|
|
|
|
|
|
|
|
.align 4
|
|
|
|
.globl sys_sigaltstack
|
|
|
|
sys_sigaltstack:
|
|
|
|
mov %o7, %l5
|
|
|
|
mov %fp, %o2
|
|
|
|
call do_sigaltstack
|
|
|
|
mov %l5, %o7
|
|
|
|
|
|
|
|
.align 4
|
|
|
|
.globl sys_sigstack
|
|
|
|
sys_sigstack:
|
|
|
|
mov %o7, %l5
|
|
|
|
mov %fp, %o2
|
|
|
|
call do_sys_sigstack
|
|
|
|
mov %l5, %o7
|
|
|
|
|
|
|
|
.align 4
|
|
|
|
.globl sys_sigreturn
|
|
|
|
sys_sigreturn:
|
|
|
|
call do_sigreturn
|
|
|
|
add %sp, STACKFRAME_SZ, %o0
|
|
|
|
|
|
|
|
ld [%curptr + TI_FLAGS], %l5
|
|
|
|
andcc %l5, _TIF_SYSCALL_TRACE, %g0
|
|
|
|
be 1f
|
|
|
|
nop
|
|
|
|
|
|
|
|
call syscall_trace
|
|
|
|
nop
|
|
|
|
|
|
|
|
1:
|
|
|
|
/* We don't want to muck with user registers like a
|
|
|
|
* normal syscall, just return.
|
|
|
|
*/
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
.align 4
|
|
|
|
.globl sys_rt_sigreturn
|
|
|
|
sys_rt_sigreturn:
|
|
|
|
call do_rt_sigreturn
|
|
|
|
add %sp, STACKFRAME_SZ, %o0
|
|
|
|
|
|
|
|
ld [%curptr + TI_FLAGS], %l5
|
|
|
|
andcc %l5, _TIF_SYSCALL_TRACE, %g0
|
|
|
|
be 1f
|
|
|
|
nop
|
|
|
|
|
2008-07-27 18:13:13 +08:00
|
|
|
add %sp, STACKFRAME_SZ, %o0
|
2005-04-17 06:20:36 +08:00
|
|
|
call syscall_trace
|
2008-07-27 18:13:13 +08:00
|
|
|
mov 1, %o1
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
1:
|
|
|
|
/* We are returning to a signal handler. */
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
/* Now that we have a real sys_clone, sys_fork() is
|
|
|
|
* implemented in terms of it. Our _real_ implementation
|
|
|
|
* of SunOS vfork() will use sys_vfork().
|
|
|
|
*
|
|
|
|
* XXX These three should be consolidated into mostly shared
|
|
|
|
* XXX code just like on sparc64... -DaveM
|
|
|
|
*/
|
|
|
|
.align 4
|
|
|
|
.globl sys_fork, flush_patch_two
|
|
|
|
sys_fork:
|
|
|
|
mov %o7, %l5
|
|
|
|
flush_patch_two:
|
|
|
|
FLUSH_ALL_KERNEL_WINDOWS;
|
|
|
|
ld [%curptr + TI_TASK], %o4
|
|
|
|
rd %psr, %g4
|
|
|
|
WRITE_PAUSE
|
|
|
|
mov SIGCHLD, %o0 ! arg0: clone flags
|
|
|
|
rd %wim, %g5
|
|
|
|
WRITE_PAUSE
|
|
|
|
mov %fp, %o1 ! arg1: usp
|
|
|
|
std %g4, [%o4 + AOFF_task_thread + AOFF_thread_fork_kpsr]
|
|
|
|
add %sp, STACKFRAME_SZ, %o2 ! arg2: pt_regs ptr
|
|
|
|
mov 0, %o3
|
|
|
|
call sparc_do_fork
|
|
|
|
mov %l5, %o7
|
|
|
|
|
|
|
|
/* Whee, kernel threads! */
|
|
|
|
.globl sys_clone, flush_patch_three
|
|
|
|
sys_clone:
|
|
|
|
mov %o7, %l5
|
|
|
|
flush_patch_three:
|
|
|
|
FLUSH_ALL_KERNEL_WINDOWS;
|
|
|
|
ld [%curptr + TI_TASK], %o4
|
|
|
|
rd %psr, %g4
|
|
|
|
WRITE_PAUSE
|
|
|
|
|
|
|
|
/* arg0,1: flags,usp -- loaded already */
|
|
|
|
cmp %o1, 0x0 ! Is new_usp NULL?
|
|
|
|
rd %wim, %g5
|
|
|
|
WRITE_PAUSE
|
|
|
|
be,a 1f
|
|
|
|
mov %fp, %o1 ! yes, use callers usp
|
|
|
|
andn %o1, 7, %o1 ! no, align to 8 bytes
|
|
|
|
1:
|
|
|
|
std %g4, [%o4 + AOFF_task_thread + AOFF_thread_fork_kpsr]
|
|
|
|
add %sp, STACKFRAME_SZ, %o2 ! arg2: pt_regs ptr
|
|
|
|
mov 0, %o3
|
|
|
|
call sparc_do_fork
|
|
|
|
mov %l5, %o7
|
|
|
|
|
|
|
|
/* Whee, real vfork! */
|
|
|
|
.globl sys_vfork, flush_patch_four
|
|
|
|
sys_vfork:
|
|
|
|
flush_patch_four:
|
|
|
|
FLUSH_ALL_KERNEL_WINDOWS;
|
|
|
|
ld [%curptr + TI_TASK], %o4
|
|
|
|
rd %psr, %g4
|
|
|
|
WRITE_PAUSE
|
|
|
|
rd %wim, %g5
|
|
|
|
WRITE_PAUSE
|
|
|
|
std %g4, [%o4 + AOFF_task_thread + AOFF_thread_fork_kpsr]
|
|
|
|
sethi %hi(0x4000 | 0x0100 | SIGCHLD), %o0
|
|
|
|
mov %fp, %o1
|
|
|
|
or %o0, %lo(0x4000 | 0x0100 | SIGCHLD), %o0
|
|
|
|
sethi %hi(sparc_do_fork), %l1
|
|
|
|
mov 0, %o3
|
|
|
|
jmpl %l1 + %lo(sparc_do_fork), %g0
|
|
|
|
add %sp, STACKFRAME_SZ, %o2
|
|
|
|
|
|
|
|
.align 4
|
|
|
|
linux_sparc_ni_syscall:
|
|
|
|
sethi %hi(sys_ni_syscall), %l7
|
|
|
|
b syscall_is_too_hard
|
|
|
|
or %l7, %lo(sys_ni_syscall), %l7
|
|
|
|
|
|
|
|
linux_fast_syscall:
|
|
|
|
andn %l7, 3, %l7
|
|
|
|
mov %i0, %o0
|
|
|
|
mov %i1, %o1
|
|
|
|
mov %i2, %o2
|
|
|
|
jmpl %l7 + %g0, %g0
|
|
|
|
mov %i3, %o3
|
|
|
|
|
|
|
|
linux_syscall_trace:
|
2008-07-27 18:13:13 +08:00
|
|
|
add %sp, STACKFRAME_SZ, %o0
|
2005-04-17 06:20:36 +08:00
|
|
|
call syscall_trace
|
2008-07-27 18:13:13 +08:00
|
|
|
mov 0, %o1
|
|
|
|
cmp %o0, 0
|
|
|
|
bne 3f
|
|
|
|
mov -ENOSYS, %o0
|
2005-04-17 06:20:36 +08:00
|
|
|
mov %i0, %o0
|
|
|
|
mov %i1, %o1
|
|
|
|
mov %i2, %o2
|
|
|
|
mov %i3, %o3
|
|
|
|
b 2f
|
|
|
|
mov %i4, %o4
|
|
|
|
|
|
|
|
.globl ret_from_fork
|
|
|
|
ret_from_fork:
|
|
|
|
call schedule_tail
|
|
|
|
mov %g3, %o0
|
|
|
|
b ret_sys_call
|
|
|
|
ld [%sp + STACKFRAME_SZ + PT_I0], %o0
|
|
|
|
|
2008-04-20 17:14:23 +08:00
|
|
|
/* Linux native system calls enter here... */
|
2005-04-17 06:20:36 +08:00
|
|
|
.align 4
|
|
|
|
.globl linux_sparc_syscall
|
|
|
|
linux_sparc_syscall:
|
sparc: Fix debugger syscall restart interactions.
So, forever, we've had this ptrace_signal_deliver implementation
which tries to handle all of the nasties that can occur when the
debugger looks at a process about to take a signal. It's meant
to address all of these issues inside of the kernel so that the
debugger need not be mindful of such things.
Problem is, this doesn't work.
The idea was that we should do the syscall restart business first, so
that the debugger captures that state. Otherwise, if the debugger for
example saves the child's state, makes the child execute something
else, then restores the saved state, we won't handle the syscall
restart properly because we lose the "we're in a syscall" state.
The code here worked for most cases, but if the debugger actually
passes the signal through to the child unaltered, it's possible that
we would do a syscall restart when we shouldn't have.
In particular this breaks the case of debugging a process under a gdb
which is being debugged by yet another gdb. gdb uses sigsuspend
to wait for SIGCHLD of the inferior, but if gdb itself is being
debugged by a top-level gdb we get a ptrace_stop(). The top-level gdb
does a PTRACE_CONT with SIGCHLD to let the inferior gdb see the
signal. But ptrace_signal_deliver() assumed the debugger would cancel
out the signal and therefore did a syscall restart, because the return
error was ERESTARTNOHAND.
Fix this by simply making ptrace_signal_deliver() a nop, and providing
a way for the debugger to control system call restarting properly:
1) Report a "in syscall" software bit in regs->{tstate,psr}.
It is set early on in trap entry to a system call and is fully
visible to the debugger via ptrace() and regsets.
2) Test this bit right before doing a syscall restart. We have
to do a final recheck right after get_signal_to_deliver() in
case the debugger cleared the bit during ptrace_stop().
3) Clear the bit in trap return so we don't accidently try to set
that bit in the real register.
As a result we also get a ptrace_{is,clear}_syscall() for sparc32 just
like sparc64 has.
M68K has this same exact bug, and is now the only other user of the
ptrace_signal_deliver hook. It needs to be fixed in the same exact
way as sparc.
Signed-off-by: David S. Miller <davem@davemloft.net>
2008-05-11 17:07:19 +08:00
|
|
|
sethi %hi(PSR_SYSCALL), %l4
|
|
|
|
or %l0, %l4, %l0
|
2005-04-17 06:20:36 +08:00
|
|
|
/* Direct access to user regs, must faster. */
|
|
|
|
cmp %g1, NR_SYSCALLS
|
|
|
|
bgeu linux_sparc_ni_syscall
|
|
|
|
sll %g1, 2, %l4
|
|
|
|
ld [%l7 + %l4], %l7
|
|
|
|
andcc %l7, 1, %g0
|
|
|
|
bne linux_fast_syscall
|
|
|
|
/* Just do first insn from SAVE_ALL in the delay slot */
|
|
|
|
|
|
|
|
syscall_is_too_hard:
|
|
|
|
SAVE_ALL_HEAD
|
|
|
|
rd %wim, %l3
|
|
|
|
|
|
|
|
wr %l0, PSR_ET, %psr
|
|
|
|
mov %i0, %o0
|
|
|
|
mov %i1, %o1
|
|
|
|
mov %i2, %o2
|
|
|
|
|
|
|
|
ld [%curptr + TI_FLAGS], %l5
|
|
|
|
mov %i3, %o3
|
|
|
|
andcc %l5, _TIF_SYSCALL_TRACE, %g0
|
|
|
|
mov %i4, %o4
|
|
|
|
bne linux_syscall_trace
|
|
|
|
mov %i0, %l5
|
|
|
|
2:
|
|
|
|
call %l7
|
|
|
|
mov %i5, %o5
|
|
|
|
|
2008-07-27 18:13:13 +08:00
|
|
|
3:
|
2005-04-17 06:20:36 +08:00
|
|
|
st %o0, [%sp + STACKFRAME_SZ + PT_I0]
|
|
|
|
|
|
|
|
ret_sys_call:
|
|
|
|
ld [%curptr + TI_FLAGS], %l6
|
|
|
|
cmp %o0, -ERESTART_RESTARTBLOCK
|
|
|
|
ld [%sp + STACKFRAME_SZ + PT_PSR], %g3
|
|
|
|
set PSR_C, %g2
|
|
|
|
bgeu 1f
|
|
|
|
andcc %l6, _TIF_SYSCALL_TRACE, %g0
|
|
|
|
|
|
|
|
/* System call success, clear Carry condition code. */
|
|
|
|
andn %g3, %g2, %g3
|
|
|
|
clr %l6
|
|
|
|
st %g3, [%sp + STACKFRAME_SZ + PT_PSR]
|
|
|
|
bne linux_syscall_trace2
|
|
|
|
ld [%sp + STACKFRAME_SZ + PT_NPC], %l1 /* pc = npc */
|
|
|
|
add %l1, 0x4, %l2 /* npc = npc+4 */
|
|
|
|
st %l1, [%sp + STACKFRAME_SZ + PT_PC]
|
|
|
|
b ret_trap_entry
|
|
|
|
st %l2, [%sp + STACKFRAME_SZ + PT_NPC]
|
|
|
|
1:
|
|
|
|
/* System call failure, set Carry condition code.
|
|
|
|
* Also, get abs(errno) to return to the process.
|
|
|
|
*/
|
|
|
|
sub %g0, %o0, %o0
|
|
|
|
or %g3, %g2, %g3
|
|
|
|
st %o0, [%sp + STACKFRAME_SZ + PT_I0]
|
|
|
|
mov 1, %l6
|
|
|
|
st %g3, [%sp + STACKFRAME_SZ + PT_PSR]
|
|
|
|
bne linux_syscall_trace2
|
|
|
|
ld [%sp + STACKFRAME_SZ + PT_NPC], %l1 /* pc = npc */
|
|
|
|
add %l1, 0x4, %l2 /* npc = npc+4 */
|
|
|
|
st %l1, [%sp + STACKFRAME_SZ + PT_PC]
|
|
|
|
b ret_trap_entry
|
|
|
|
st %l2, [%sp + STACKFRAME_SZ + PT_NPC]
|
|
|
|
|
|
|
|
linux_syscall_trace2:
|
2008-07-27 18:13:13 +08:00
|
|
|
add %sp, STACKFRAME_SZ, %o0
|
|
|
|
mov 1, %o1
|
2005-04-17 06:20:36 +08:00
|
|
|
call syscall_trace
|
|
|
|
add %l1, 0x4, %l2 /* npc = npc+4 */
|
|
|
|
st %l1, [%sp + STACKFRAME_SZ + PT_PC]
|
|
|
|
b ret_trap_entry
|
|
|
|
st %l2, [%sp + STACKFRAME_SZ + PT_NPC]
|
|
|
|
|
|
|
|
|
|
|
|
/* Saving and restoring the FPU state is best done from lowlevel code.
|
|
|
|
*
|
|
|
|
* void fpsave(unsigned long *fpregs, unsigned long *fsr,
|
|
|
|
* void *fpqueue, unsigned long *fpqdepth)
|
|
|
|
*/
|
|
|
|
|
|
|
|
.globl fpsave
|
|
|
|
fpsave:
|
|
|
|
st %fsr, [%o1] ! this can trap on us if fpu is in bogon state
|
|
|
|
ld [%o1], %g1
|
|
|
|
set 0x2000, %g4
|
|
|
|
andcc %g1, %g4, %g0
|
|
|
|
be 2f
|
|
|
|
mov 0, %g2
|
|
|
|
|
|
|
|
/* We have an fpqueue to save. */
|
|
|
|
1:
|
|
|
|
std %fq, [%o2]
|
|
|
|
fpsave_magic:
|
|
|
|
st %fsr, [%o1]
|
|
|
|
ld [%o1], %g3
|
|
|
|
andcc %g3, %g4, %g0
|
|
|
|
add %g2, 1, %g2
|
|
|
|
bne 1b
|
|
|
|
add %o2, 8, %o2
|
|
|
|
|
|
|
|
2:
|
|
|
|
st %g2, [%o3]
|
|
|
|
|
|
|
|
std %f0, [%o0 + 0x00]
|
|
|
|
std %f2, [%o0 + 0x08]
|
|
|
|
std %f4, [%o0 + 0x10]
|
|
|
|
std %f6, [%o0 + 0x18]
|
|
|
|
std %f8, [%o0 + 0x20]
|
|
|
|
std %f10, [%o0 + 0x28]
|
|
|
|
std %f12, [%o0 + 0x30]
|
|
|
|
std %f14, [%o0 + 0x38]
|
|
|
|
std %f16, [%o0 + 0x40]
|
|
|
|
std %f18, [%o0 + 0x48]
|
|
|
|
std %f20, [%o0 + 0x50]
|
|
|
|
std %f22, [%o0 + 0x58]
|
|
|
|
std %f24, [%o0 + 0x60]
|
|
|
|
std %f26, [%o0 + 0x68]
|
|
|
|
std %f28, [%o0 + 0x70]
|
|
|
|
retl
|
|
|
|
std %f30, [%o0 + 0x78]
|
|
|
|
|
|
|
|
/* Thanks for Theo Deraadt and the authors of the Sprite/netbsd/openbsd
|
|
|
|
* code for pointing out this possible deadlock, while we save state
|
|
|
|
* above we could trap on the fsr store so our low level fpu trap
|
|
|
|
* code has to know how to deal with this.
|
|
|
|
*/
|
|
|
|
fpsave_catch:
|
|
|
|
b fpsave_magic + 4
|
|
|
|
st %fsr, [%o1]
|
|
|
|
|
|
|
|
fpsave_catch2:
|
|
|
|
b fpsave + 4
|
|
|
|
st %fsr, [%o1]
|
|
|
|
|
|
|
|
/* void fpload(unsigned long *fpregs, unsigned long *fsr); */
|
|
|
|
|
|
|
|
.globl fpload
|
|
|
|
fpload:
|
|
|
|
ldd [%o0 + 0x00], %f0
|
|
|
|
ldd [%o0 + 0x08], %f2
|
|
|
|
ldd [%o0 + 0x10], %f4
|
|
|
|
ldd [%o0 + 0x18], %f6
|
|
|
|
ldd [%o0 + 0x20], %f8
|
|
|
|
ldd [%o0 + 0x28], %f10
|
|
|
|
ldd [%o0 + 0x30], %f12
|
|
|
|
ldd [%o0 + 0x38], %f14
|
|
|
|
ldd [%o0 + 0x40], %f16
|
|
|
|
ldd [%o0 + 0x48], %f18
|
|
|
|
ldd [%o0 + 0x50], %f20
|
|
|
|
ldd [%o0 + 0x58], %f22
|
|
|
|
ldd [%o0 + 0x60], %f24
|
|
|
|
ldd [%o0 + 0x68], %f26
|
|
|
|
ldd [%o0 + 0x70], %f28
|
|
|
|
ldd [%o0 + 0x78], %f30
|
|
|
|
ld [%o1], %fsr
|
|
|
|
retl
|
|
|
|
nop
|
|
|
|
|
|
|
|
/* __ndelay and __udelay take two arguments:
|
|
|
|
* 0 - nsecs or usecs to delay
|
|
|
|
* 1 - per_cpu udelay_val (loops per jiffy)
|
|
|
|
*
|
|
|
|
* Note that ndelay gives HZ times higher resolution but has a 10ms
|
|
|
|
* limit. udelay can handle up to 1s.
|
|
|
|
*/
|
|
|
|
.globl __ndelay
|
|
|
|
__ndelay:
|
|
|
|
save %sp, -STACKFRAME_SZ, %sp
|
|
|
|
mov %i0, %o0
|
2007-07-22 10:17:41 +08:00
|
|
|
call .umul ! round multiplier up so large ns ok
|
|
|
|
mov 0x1ae, %o1 ! 2**32 / (1 000 000 000 / HZ)
|
2005-04-17 06:20:36 +08:00
|
|
|
call .umul
|
|
|
|
mov %i1, %o1 ! udelay_val
|
|
|
|
ba delay_continue
|
|
|
|
mov %o1, %o0 ! >>32 later for better resolution
|
|
|
|
|
|
|
|
.globl __udelay
|
|
|
|
__udelay:
|
|
|
|
save %sp, -STACKFRAME_SZ, %sp
|
|
|
|
mov %i0, %o0
|
2007-07-22 10:17:41 +08:00
|
|
|
sethi %hi(0x10c7), %o1 ! round multiplier up so large us ok
|
2005-04-17 06:20:36 +08:00
|
|
|
call .umul
|
2007-07-22 10:17:41 +08:00
|
|
|
or %o1, %lo(0x10c7), %o1 ! 2**32 / 1 000 000
|
2005-04-17 06:20:36 +08:00
|
|
|
call .umul
|
|
|
|
mov %i1, %o1 ! udelay_val
|
2007-07-22 10:17:41 +08:00
|
|
|
sethi %hi(0x028f4b62), %l0 ! Add in rounding constant * 2**32,
|
|
|
|
or %g0, %lo(0x028f4b62), %l0
|
|
|
|
addcc %o0, %l0, %o0 ! 2**32 * 0.009 999
|
|
|
|
bcs,a 3f
|
|
|
|
add %o1, 0x01, %o1
|
|
|
|
3:
|
2005-04-17 06:20:36 +08:00
|
|
|
call .umul
|
|
|
|
mov HZ, %o0 ! >>32 earlier for wider range
|
|
|
|
|
|
|
|
delay_continue:
|
|
|
|
cmp %o0, 0x0
|
|
|
|
1:
|
|
|
|
bne 1b
|
|
|
|
subcc %o0, 1, %o0
|
|
|
|
|
|
|
|
ret
|
|
|
|
restore
|
|
|
|
|
|
|
|
/* Handle a software breakpoint */
|
|
|
|
/* We have to inform parent that child has stopped */
|
|
|
|
.align 4
|
|
|
|
.globl breakpoint_trap
|
|
|
|
breakpoint_trap:
|
|
|
|
rd %wim,%l3
|
|
|
|
SAVE_ALL
|
|
|
|
wr %l0, PSR_ET, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
|
|
|
|
st %i0, [%sp + STACKFRAME_SZ + PT_G0] ! for restarting syscalls
|
|
|
|
call sparc_breakpoint
|
|
|
|
add %sp, STACKFRAME_SZ, %o0
|
|
|
|
|
|
|
|
RESTORE_ALL
|
|
|
|
|
2008-04-29 17:38:50 +08:00
|
|
|
#ifdef CONFIG_KGDB
|
|
|
|
.align 4
|
|
|
|
.globl kgdb_trap_low
|
|
|
|
.type kgdb_trap_low,#function
|
|
|
|
kgdb_trap_low:
|
|
|
|
rd %wim,%l3
|
|
|
|
SAVE_ALL
|
|
|
|
wr %l0, PSR_ET, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
|
|
|
|
call kgdb_trap
|
|
|
|
add %sp, STACKFRAME_SZ, %o0
|
|
|
|
|
|
|
|
RESTORE_ALL
|
|
|
|
.size kgdb_trap_low,.-kgdb_trap_low
|
|
|
|
#endif
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
.align 4
|
2008-06-06 02:40:58 +08:00
|
|
|
.globl flush_patch_exception
|
2005-04-17 06:20:36 +08:00
|
|
|
flush_patch_exception:
|
|
|
|
FLUSH_ALL_KERNEL_WINDOWS;
|
|
|
|
ldd [%o0], %o6
|
|
|
|
jmpl %o7 + 0xc, %g0 ! see asm-sparc/processor.h
|
|
|
|
mov 1, %g1 ! signal EFAULT condition
|
|
|
|
|
|
|
|
.align 4
|
|
|
|
.globl kill_user_windows, kuw_patch1_7win
|
|
|
|
.globl kuw_patch1
|
|
|
|
kuw_patch1_7win: sll %o3, 6, %o3
|
|
|
|
|
|
|
|
/* No matter how much overhead this routine has in the worst
|
|
|
|
* case scenerio, it is several times better than taking the
|
|
|
|
* traps with the old method of just doing flush_user_windows().
|
|
|
|
*/
|
|
|
|
kill_user_windows:
|
|
|
|
ld [%g6 + TI_UWINMASK], %o0 ! get current umask
|
|
|
|
orcc %g0, %o0, %g0 ! if no bits set, we are done
|
|
|
|
be 3f ! nothing to do
|
|
|
|
rd %psr, %o5 ! must clear interrupts
|
|
|
|
or %o5, PSR_PIL, %o4 ! or else that could change
|
|
|
|
wr %o4, 0x0, %psr ! the uwinmask state
|
|
|
|
WRITE_PAUSE ! burn them cycles
|
|
|
|
1:
|
|
|
|
ld [%g6 + TI_UWINMASK], %o0 ! get consistent state
|
|
|
|
orcc %g0, %o0, %g0 ! did an interrupt come in?
|
|
|
|
be 4f ! yep, we are done
|
|
|
|
rd %wim, %o3 ! get current wim
|
|
|
|
srl %o3, 1, %o4 ! simulate a save
|
|
|
|
kuw_patch1:
|
|
|
|
sll %o3, 7, %o3 ! compute next wim
|
|
|
|
or %o4, %o3, %o3 ! result
|
|
|
|
andncc %o0, %o3, %o0 ! clean this bit in umask
|
|
|
|
bne kuw_patch1 ! not done yet
|
|
|
|
srl %o3, 1, %o4 ! begin another save simulation
|
|
|
|
wr %o3, 0x0, %wim ! set the new wim
|
|
|
|
st %g0, [%g6 + TI_UWINMASK] ! clear uwinmask
|
|
|
|
4:
|
|
|
|
wr %o5, 0x0, %psr ! re-enable interrupts
|
|
|
|
WRITE_PAUSE ! burn baby burn
|
|
|
|
3:
|
|
|
|
retl ! return
|
|
|
|
st %g0, [%g6 + TI_W_SAVED] ! no windows saved
|
|
|
|
|
|
|
|
.align 4
|
|
|
|
.globl restore_current
|
|
|
|
restore_current:
|
|
|
|
LOAD_CURRENT(g6, o0)
|
|
|
|
retl
|
|
|
|
nop
|
|
|
|
|
|
|
|
#ifdef CONFIG_PCI
|
|
|
|
#include <asm/pcic.h>
|
|
|
|
|
|
|
|
.align 4
|
|
|
|
.globl linux_trap_ipi15_pcic
|
|
|
|
linux_trap_ipi15_pcic:
|
|
|
|
rd %wim, %l3
|
|
|
|
SAVE_ALL
|
|
|
|
|
|
|
|
/*
|
|
|
|
* First deactivate NMI
|
|
|
|
* or we cannot drop ET, cannot get window spill traps.
|
|
|
|
* The busy loop is necessary because the PIO error
|
|
|
|
* sometimes does not go away quickly and we trap again.
|
|
|
|
*/
|
|
|
|
sethi %hi(pcic_regs), %o1
|
|
|
|
ld [%o1 + %lo(pcic_regs)], %o2
|
|
|
|
|
|
|
|
! Get pending status for printouts later.
|
|
|
|
ld [%o2 + PCI_SYS_INT_PENDING], %o0
|
|
|
|
|
|
|
|
mov PCI_SYS_INT_PENDING_CLEAR_ALL, %o1
|
|
|
|
stb %o1, [%o2 + PCI_SYS_INT_PENDING_CLEAR]
|
|
|
|
1:
|
|
|
|
ld [%o2 + PCI_SYS_INT_PENDING], %o1
|
|
|
|
andcc %o1, ((PCI_SYS_INT_PENDING_PIO|PCI_SYS_INT_PENDING_PCI)>>24), %g0
|
|
|
|
bne 1b
|
|
|
|
nop
|
|
|
|
|
|
|
|
or %l0, PSR_PIL, %l4
|
|
|
|
wr %l4, 0x0, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
wr %l4, PSR_ET, %psr
|
|
|
|
WRITE_PAUSE
|
|
|
|
|
|
|
|
call pcic_nmi
|
|
|
|
add %sp, STACKFRAME_SZ, %o1 ! struct pt_regs *regs
|
|
|
|
RESTORE_ALL
|
|
|
|
|
|
|
|
.globl pcic_nmi_trap_patch
|
|
|
|
pcic_nmi_trap_patch:
|
|
|
|
sethi %hi(linux_trap_ipi15_pcic), %l3
|
|
|
|
jmpl %l3 + %lo(linux_trap_ipi15_pcic), %g0
|
|
|
|
rd %psr, %l0
|
|
|
|
.word 0
|
|
|
|
|
|
|
|
#endif /* CONFIG_PCI */
|
|
|
|
|
2008-04-29 17:38:50 +08:00
|
|
|
.globl flushw_all
|
|
|
|
flushw_all:
|
|
|
|
save %sp, -0x40, %sp
|
|
|
|
save %sp, -0x40, %sp
|
|
|
|
save %sp, -0x40, %sp
|
|
|
|
save %sp, -0x40, %sp
|
|
|
|
save %sp, -0x40, %sp
|
|
|
|
save %sp, -0x40, %sp
|
|
|
|
save %sp, -0x40, %sp
|
|
|
|
restore
|
|
|
|
restore
|
|
|
|
restore
|
|
|
|
restore
|
|
|
|
restore
|
|
|
|
restore
|
|
|
|
ret
|
|
|
|
restore
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
/* End of entry.S */
|