2019-05-27 14:55:01 +08:00
|
|
|
/* SPDX-License-Identifier: GPL-2.0-or-later */
|
2005-09-26 14:04:21 +08:00
|
|
|
/*
|
|
|
|
* FPU support code, moved here from head.S so that it can be used
|
|
|
|
* by chips which use other head-whatever.S files.
|
|
|
|
*
|
2006-08-30 12:45:35 +08:00
|
|
|
* Copyright (C) 1995-1996 Gary Thomas (gdt@linuxppc.org)
|
|
|
|
* Copyright (C) 1996 Cort Dougan <cort@cs.nmt.edu>
|
|
|
|
* Copyright (C) 1996 Paul Mackerras.
|
|
|
|
* Copyright (C) 1997 Dan Malek (dmalek@jlc.net).
|
2005-09-26 14:04:21 +08:00
|
|
|
*/
|
|
|
|
|
2005-10-10 20:20:10 +08:00
|
|
|
#include <asm/reg.h>
|
2005-09-26 14:04:21 +08:00
|
|
|
#include <asm/page.h>
|
|
|
|
#include <asm/mmu.h>
|
|
|
|
#include <asm/cputable.h>
|
|
|
|
#include <asm/cache.h>
|
|
|
|
#include <asm/thread_info.h>
|
|
|
|
#include <asm/ppc_asm.h>
|
|
|
|
#include <asm/asm-offsets.h>
|
2010-11-18 23:06:17 +08:00
|
|
|
#include <asm/ptrace.h>
|
2016-01-14 12:33:46 +08:00
|
|
|
#include <asm/export.h>
|
2018-07-06 00:24:57 +08:00
|
|
|
#include <asm/asm-compat.h>
|
2018-07-06 00:25:01 +08:00
|
|
|
#include <asm/feature-fixups.h>
|
2005-09-26 14:04:21 +08:00
|
|
|
|
2008-06-25 12:07:18 +08:00
|
|
|
#ifdef CONFIG_VSX
|
2012-06-25 21:33:23 +08:00
|
|
|
#define __REST_32FPVSRS(n,c,base) \
|
2008-06-25 12:07:18 +08:00
|
|
|
BEGIN_FTR_SECTION \
|
|
|
|
b 2f; \
|
|
|
|
END_FTR_SECTION_IFSET(CPU_FTR_VSX); \
|
|
|
|
REST_32FPRS(n,base); \
|
|
|
|
b 3f; \
|
|
|
|
2: REST_32VSRS(n,c,base); \
|
|
|
|
3:
|
|
|
|
|
2012-06-25 21:33:23 +08:00
|
|
|
#define __SAVE_32FPVSRS(n,c,base) \
|
2008-06-25 12:07:18 +08:00
|
|
|
BEGIN_FTR_SECTION \
|
|
|
|
b 2f; \
|
|
|
|
END_FTR_SECTION_IFSET(CPU_FTR_VSX); \
|
|
|
|
SAVE_32FPRS(n,base); \
|
|
|
|
b 3f; \
|
|
|
|
2: SAVE_32VSRS(n,c,base); \
|
|
|
|
3:
|
|
|
|
#else
|
2012-06-25 21:33:23 +08:00
|
|
|
#define __REST_32FPVSRS(n,b,base) REST_32FPRS(n, base)
|
|
|
|
#define __SAVE_32FPVSRS(n,b,base) SAVE_32FPRS(n, base)
|
2008-06-25 12:07:18 +08:00
|
|
|
#endif
|
2012-06-25 21:33:23 +08:00
|
|
|
#define REST_32FPVSRS(n,c,base) __REST_32FPVSRS(n,__REG_##c,__REG_##base)
|
|
|
|
#define SAVE_32FPVSRS(n,c,base) __SAVE_32FPVSRS(n,__REG_##c,__REG_##base)
|
2008-06-25 12:07:18 +08:00
|
|
|
|
2013-09-10 18:21:10 +08:00
|
|
|
/*
|
|
|
|
* Load state from memory into FP registers including FPSCR.
|
|
|
|
* Assumes the caller has enabled FP in the MSR.
|
|
|
|
*/
|
|
|
|
_GLOBAL(load_fp_state)
|
|
|
|
lfd fr0,FPSTATE_FPSCR(r3)
|
|
|
|
MTFSF_L(fr0)
|
|
|
|
REST_32FPVSRS(0, R4, R3)
|
|
|
|
blr
|
2016-01-14 12:33:46 +08:00
|
|
|
EXPORT_SYMBOL(load_fp_state)
|
powerpc/64: Don't trace code that runs with the soft irq mask unreconciled
"Reconciling" in terms of interrupt handling, is to bring the soft irq
mask state in to synch with the hardware, after an interrupt causes
MSR[EE] to be cleared (while the soft mask may be enabled, and hard
irqs not marked disabled).
General kernel code should not be called while unreconciled, because
local_irq_disable, etc. manipulations can cause surprising irq traces,
and it's fragile because the soft irq code does not really expect to
be called in this situation.
When exiting from an interrupt, MSR[EE] is cleared to prevent races,
but soft irq state is enabled for the returned-to context, so this is
now an unreconciled state. restore_math is called in this state, and
that can be ftraced, and the ftrace subsystem disables local irqs.
Mark restore_math and its callees as notrace. Restore a sanity check
in the soft irq code that had to be disabled for this case, by commit
4da1f79227ad4 ("powerpc/64: Disable irq restore warning for now").
Signed-off-by: Nicholas Piggin <npiggin@gmail.com>
Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
2019-05-02 13:21:07 +08:00
|
|
|
_ASM_NOKPROBE_SYMBOL(load_fp_state); /* used by restore_math */
|
2013-09-10 18:21:10 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Store FP state into memory, including FPSCR
|
|
|
|
* Assumes the caller has enabled FP in the MSR.
|
|
|
|
*/
|
|
|
|
_GLOBAL(store_fp_state)
|
|
|
|
SAVE_32FPVSRS(0, R4, R3)
|
|
|
|
mffs fr0
|
|
|
|
stfd fr0,FPSTATE_FPSCR(r3)
|
|
|
|
blr
|
2016-01-14 12:33:46 +08:00
|
|
|
EXPORT_SYMBOL(store_fp_state)
|
2013-09-10 18:21:10 +08:00
|
|
|
|
2005-09-26 14:04:21 +08:00
|
|
|
/*
|
|
|
|
* This task wants to use the FPU now.
|
|
|
|
* On UP, disable FP for the task which had the FPU previously,
|
|
|
|
* and save its floating-point registers in its thread_struct.
|
|
|
|
* Load up this task's FP registers from its thread_struct,
|
|
|
|
* enable the FPU for the current task and return to the task.
|
2013-10-23 16:40:02 +08:00
|
|
|
* Note that on 32-bit this can only use registers that will be
|
|
|
|
* restored by fast_exception_return, i.e. r3 - r6, r10 and r11.
|
2005-09-26 14:04:21 +08:00
|
|
|
*/
|
2005-10-06 08:59:19 +08:00
|
|
|
_GLOBAL(load_up_fpu)
|
2005-09-26 14:04:21 +08:00
|
|
|
mfmsr r5
|
|
|
|
ori r5,r5,MSR_FP
|
2008-06-25 12:07:18 +08:00
|
|
|
#ifdef CONFIG_VSX
|
|
|
|
BEGIN_FTR_SECTION
|
|
|
|
oris r5,r5,MSR_VSX@h
|
|
|
|
END_FTR_SECTION_IFSET(CPU_FTR_VSX)
|
|
|
|
#endif
|
2005-09-26 14:04:21 +08:00
|
|
|
SYNC
|
|
|
|
MTMSRD(r5) /* enable use of fpu now */
|
|
|
|
isync
|
|
|
|
/* enable use of FP after return */
|
2005-10-06 08:59:19 +08:00
|
|
|
#ifdef CONFIG_PPC32
|
2013-09-10 18:20:42 +08:00
|
|
|
mfspr r5,SPRN_SPRG_THREAD /* current task's THREAD (phys) */
|
2019-12-21 16:32:38 +08:00
|
|
|
#ifdef CONFIG_VMAP_STACK
|
|
|
|
tovirt(r5, r5)
|
|
|
|
#endif
|
2005-09-26 14:04:21 +08:00
|
|
|
lwz r4,THREAD_FPEXC_MODE(r5)
|
|
|
|
ori r9,r9,MSR_FP /* enable FP for current */
|
|
|
|
or r9,r9,r4
|
2005-10-06 08:59:19 +08:00
|
|
|
#else
|
|
|
|
ld r4,PACACURRENT(r13)
|
|
|
|
addi r5,r4,THREAD /* Get THREAD */
|
2006-02-07 10:55:30 +08:00
|
|
|
lwz r4,THREAD_FPEXC_MODE(r5)
|
2005-10-06 08:59:19 +08:00
|
|
|
ori r12,r12,MSR_FP
|
|
|
|
or r12,r12,r4
|
|
|
|
std r12,_MSR(r1)
|
|
|
|
#endif
|
2020-06-24 07:41:39 +08:00
|
|
|
li r4,1
|
2016-02-29 14:53:47 +08:00
|
|
|
stb r4,THREAD_LOAD_FP(r5)
|
2013-10-23 16:40:02 +08:00
|
|
|
addi r10,r5,THREAD_FPSTATE
|
|
|
|
lfd fr0,FPSTATE_FPSCR(r10)
|
2006-06-10 18:18:39 +08:00
|
|
|
MTFSF_L(fr0)
|
2013-10-23 16:40:02 +08:00
|
|
|
REST_32FPVSRS(0, R4, R10)
|
2005-09-26 14:04:21 +08:00
|
|
|
/* restore registers and return */
|
|
|
|
/* we haven't used ctr or xer or lr */
|
2008-06-25 12:07:18 +08:00
|
|
|
blr
|
2020-04-01 00:03:44 +08:00
|
|
|
_ASM_NOKPROBE_SYMBOL(load_up_fpu)
|
2005-09-26 14:04:21 +08:00
|
|
|
|
|
|
|
/*
|
2016-02-29 14:53:49 +08:00
|
|
|
* save_fpu(tsk)
|
|
|
|
* Save the floating-point registers in its thread_struct.
|
2005-09-26 14:04:21 +08:00
|
|
|
* Enables the FPU for use in the kernel on return.
|
|
|
|
*/
|
2016-02-29 14:53:49 +08:00
|
|
|
_GLOBAL(save_fpu)
|
2005-09-26 14:04:21 +08:00
|
|
|
addi r3,r3,THREAD /* want THREAD of task */
|
2013-09-10 18:21:10 +08:00
|
|
|
PPC_LL r6,THREAD_FPSAVEAREA(r3)
|
[PATCH] powerpc: Consolidate asm compatibility macros
This patch consolidates macros used to generate assembly for
compatibility across different CPUs or configs. A new header,
asm-powerpc/asm-compat.h contains the main compatibility macros. It
uses some preprocessor magic to make the macros suitable both for use
in .S files, and in inline asm in .c files. Headers (bitops.h,
uaccess.h, atomic.h, bug.h) which had their own such compatibility
macros are changed to use asm-compat.h.
ppc_asm.h is now for use in .S files *only*, and a #error enforces
that. As such, we're a lot more careless about namespace pollution
here than in asm-compat.h.
While we're at it, this patch adds a call to the PPC405_ERR77 macro in
futex.h which should have had it already, but didn't.
Built and booted on pSeries, Maple and iSeries (ARCH=powerpc). Built
for 32-bit powermac (ARCH=powerpc) and Walnut (ARCH=ppc).
Signed-off-by: David Gibson <dwg@au1.ibm.com>
Signed-off-by: Paul Mackerras <paulus@samba.org>
2005-11-10 09:56:55 +08:00
|
|
|
PPC_LL r5,PT_REGS(r3)
|
2013-09-10 18:21:10 +08:00
|
|
|
PPC_LCMPI 0,r6,0
|
|
|
|
bne 2f
|
2013-09-10 18:20:42 +08:00
|
|
|
addi r6,r3,THREAD_FPSTATE
|
2016-02-29 14:53:49 +08:00
|
|
|
2: SAVE_32FPVSRS(0, R4, R6)
|
2005-09-26 14:04:21 +08:00
|
|
|
mffs fr0
|
2013-09-10 18:20:42 +08:00
|
|
|
stfd fr0,FPSTATE_FPSCR(r6)
|
2005-09-26 14:04:21 +08:00
|
|
|
blr
|
[PATCH] powerpc: Fix handling of fpscr on 64-bit
The recent merge of fpu.S broken the handling of fpscr for
ARCH=powerpc and CONFIG_PPC64=y. FP registers could be corrupted,
leading to strange random application crashes.
The confusion arises, because the thread_struct has (and requires) a
64-bit area to save the fpscr, because we use load/store double
instructions to get it in to/out of the FPU. However, only the low
32-bits are actually used, so we want to treat it as a 32-bit quantity
when manipulating its bits to avoid extra load/stores on 32-bit. This
patch replaces the current definition with a structure of two 32-bit
quantities (pad and val), to clarify things as much as is possible.
The 'val' field is used when manipulating bits, the structure itself
is used when obtaining the address for loading/unloading the value
from the FPU.
While we're at it, consolidate the 4 (!) almost identical versions of
cvt_fd() and cvt_df() (arch/ppc/kernel/misc.S,
arch/ppc64/kernel/misc.S, arch/powerpc/kernel/misc_32.S,
arch/powerpc/kernel/misc_64.S) into a single version in fpu.S. The
new version takes a pointer to thread_struct and applies the correct
offset itself, rather than a pointer to the fpscr field itself, again
to avoid confusion as to which is the correct field to use.
Finally, this patch makes ARCH=ppc64 also use the consolidated fpu.S
code, which it previously did not.
Built for G5 (ARCH=ppc64 and ARCH=powerpc), 32-bit powermac (ARCH=ppc
and ARCH=powerpc) and Walnut (ARCH=ppc, CONFIG_MATH_EMULATION=y).
Booted on G5 (ARCH=powerpc) and things which previously fell over no
longer do.
Signed-off-by: David Gibson <dwg@au1.ibm.com>
Signed-off-by: Paul Mackerras <paulus@samba.org>
2005-10-27 14:27:25 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* These are used in the alignment trap handler when emulating
|
|
|
|
* single-precision loads and stores.
|
|
|
|
*/
|
|
|
|
|
|
|
|
_GLOBAL(cvt_fd)
|
|
|
|
lfs 0,0(r3)
|
|
|
|
stfd 0,0(r4)
|
|
|
|
blr
|
|
|
|
|
|
|
|
_GLOBAL(cvt_df)
|
|
|
|
lfd 0,0(r3)
|
|
|
|
stfs 0,0(r4)
|
|
|
|
blr
|