2013-10-30 22:35:40 +08:00
|
|
|
/*
|
|
|
|
* Machine check exception handling.
|
|
|
|
*
|
|
|
|
* This program is free software; you can redistribute it and/or modify
|
|
|
|
* it under the terms of the GNU General Public License as published by
|
|
|
|
* the Free Software Foundation; either version 2 of the License, or
|
|
|
|
* (at your option) any later version.
|
|
|
|
*
|
|
|
|
* This program is distributed in the hope that it will be useful,
|
|
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
* GNU General Public License for more details.
|
|
|
|
*
|
|
|
|
* You should have received a copy of the GNU General Public License
|
|
|
|
* along with this program; if not, write to the Free Software
|
|
|
|
* Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
|
|
|
|
*
|
|
|
|
* Copyright 2013 IBM Corporation
|
|
|
|
* Author: Mahesh Salgaonkar <mahesh@linux.vnet.ibm.com>
|
|
|
|
*/
|
|
|
|
|
|
|
|
#undef DEBUG
|
|
|
|
#define pr_fmt(fmt) "mce: " fmt
|
|
|
|
|
2017-08-08 14:39:24 +08:00
|
|
|
#include <linux/hardirq.h>
|
2013-10-30 22:35:40 +08:00
|
|
|
#include <linux/types.h>
|
|
|
|
#include <linux/ptrace.h>
|
|
|
|
#include <linux/percpu.h>
|
|
|
|
#include <linux/export.h>
|
2014-01-14 18:15:09 +08:00
|
|
|
#include <linux/irq_work.h>
|
2017-08-08 14:39:24 +08:00
|
|
|
|
|
|
|
#include <asm/machdep.h>
|
2013-10-30 22:35:40 +08:00
|
|
|
#include <asm/mce.h>
|
|
|
|
|
|
|
|
static DEFINE_PER_CPU(int, mce_nest_count);
|
|
|
|
static DEFINE_PER_CPU(struct machine_check_event[MAX_MC_EVT], mce_event);
|
|
|
|
|
2013-10-30 22:35:49 +08:00
|
|
|
/* Queue for delayed MCE events. */
|
|
|
|
static DEFINE_PER_CPU(int, mce_queue_count);
|
|
|
|
static DEFINE_PER_CPU(struct machine_check_event[MAX_MC_EVT], mce_event_queue);
|
|
|
|
|
2014-01-14 18:15:09 +08:00
|
|
|
static void machine_check_process_queued_event(struct irq_work *work);
|
2016-01-06 08:45:50 +08:00
|
|
|
static struct irq_work mce_event_process_work = {
|
2014-01-14 18:15:09 +08:00
|
|
|
.func = machine_check_process_queued_event,
|
|
|
|
};
|
|
|
|
|
2013-10-30 22:35:40 +08:00
|
|
|
static void mce_set_error_info(struct machine_check_event *mce,
|
|
|
|
struct mce_error_info *mce_err)
|
|
|
|
{
|
|
|
|
mce->error_type = mce_err->error_type;
|
|
|
|
switch (mce_err->error_type) {
|
|
|
|
case MCE_ERROR_TYPE_UE:
|
|
|
|
mce->u.ue_error.ue_error_type = mce_err->u.ue_error_type;
|
|
|
|
break;
|
|
|
|
case MCE_ERROR_TYPE_SLB:
|
|
|
|
mce->u.slb_error.slb_error_type = mce_err->u.slb_error_type;
|
|
|
|
break;
|
|
|
|
case MCE_ERROR_TYPE_ERAT:
|
|
|
|
mce->u.erat_error.erat_error_type = mce_err->u.erat_error_type;
|
|
|
|
break;
|
|
|
|
case MCE_ERROR_TYPE_TLB:
|
|
|
|
mce->u.tlb_error.tlb_error_type = mce_err->u.tlb_error_type;
|
|
|
|
break;
|
2017-02-28 10:00:48 +08:00
|
|
|
case MCE_ERROR_TYPE_USER:
|
|
|
|
mce->u.user_error.user_error_type = mce_err->u.user_error_type;
|
|
|
|
break;
|
|
|
|
case MCE_ERROR_TYPE_RA:
|
|
|
|
mce->u.ra_error.ra_error_type = mce_err->u.ra_error_type;
|
|
|
|
break;
|
|
|
|
case MCE_ERROR_TYPE_LINK:
|
|
|
|
mce->u.link_error.link_error_type = mce_err->u.link_error_type;
|
|
|
|
break;
|
2013-10-30 22:35:40 +08:00
|
|
|
case MCE_ERROR_TYPE_UNKNOWN:
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Decode and save high level MCE information into per cpu buffer which
|
|
|
|
* is an array of machine_check_event structure.
|
|
|
|
*/
|
|
|
|
void save_mce_event(struct pt_regs *regs, long handled,
|
|
|
|
struct mce_error_info *mce_err,
|
2017-09-29 12:26:53 +08:00
|
|
|
uint64_t nip, uint64_t addr, uint64_t phys_addr)
|
2013-10-30 22:35:40 +08:00
|
|
|
{
|
2015-05-12 11:23:59 +08:00
|
|
|
int index = __this_cpu_inc_return(mce_nest_count) - 1;
|
powerpc: Replace __get_cpu_var uses
This still has not been merged and now powerpc is the only arch that does
not have this change. Sorry about missing linuxppc-dev before.
V2->V2
- Fix up to work against 3.18-rc1
__get_cpu_var() is used for multiple purposes in the kernel source. One of
them is address calculation via the form &__get_cpu_var(x). This calculates
the address for the instance of the percpu variable of the current processor
based on an offset.
Other use cases are for storing and retrieving data from the current
processors percpu area. __get_cpu_var() can be used as an lvalue when
writing data or on the right side of an assignment.
__get_cpu_var() is defined as :
__get_cpu_var() always only does an address determination. However, store
and retrieve operations could use a segment prefix (or global register on
other platforms) to avoid the address calculation.
this_cpu_write() and this_cpu_read() can directly take an offset into a
percpu area and use optimized assembly code to read and write per cpu
variables.
This patch converts __get_cpu_var into either an explicit address
calculation using this_cpu_ptr() or into a use of this_cpu operations that
use the offset. Thereby address calculations are avoided and less registers
are used when code is generated.
At the end of the patch set all uses of __get_cpu_var have been removed so
the macro is removed too.
The patch set includes passes over all arches as well. Once these operations
are used throughout then specialized macros can be defined in non -x86
arches as well in order to optimize per cpu access by f.e. using a global
register that may be set to the per cpu base.
Transformations done to __get_cpu_var()
1. Determine the address of the percpu instance of the current processor.
DEFINE_PER_CPU(int, y);
int *x = &__get_cpu_var(y);
Converts to
int *x = this_cpu_ptr(&y);
2. Same as #1 but this time an array structure is involved.
DEFINE_PER_CPU(int, y[20]);
int *x = __get_cpu_var(y);
Converts to
int *x = this_cpu_ptr(y);
3. Retrieve the content of the current processors instance of a per cpu
variable.
DEFINE_PER_CPU(int, y);
int x = __get_cpu_var(y)
Converts to
int x = __this_cpu_read(y);
4. Retrieve the content of a percpu struct
DEFINE_PER_CPU(struct mystruct, y);
struct mystruct x = __get_cpu_var(y);
Converts to
memcpy(&x, this_cpu_ptr(&y), sizeof(x));
5. Assignment to a per cpu variable
DEFINE_PER_CPU(int, y)
__get_cpu_var(y) = x;
Converts to
__this_cpu_write(y, x);
6. Increment/Decrement etc of a per cpu variable
DEFINE_PER_CPU(int, y);
__get_cpu_var(y)++
Converts to
__this_cpu_inc(y)
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
CC: Paul Mackerras <paulus@samba.org>
Signed-off-by: Christoph Lameter <cl@linux.com>
[mpe: Fix build errors caused by set/or_softirq_pending(), and rework
assignment in __set_breakpoint() to use memcpy().]
Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
2014-10-22 04:23:25 +08:00
|
|
|
struct machine_check_event *mce = this_cpu_ptr(&mce_event[index]);
|
2013-10-30 22:35:40 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Return if we don't have enough space to log mce event.
|
|
|
|
* mce_nest_count may go beyond MAX_MC_EVT but that's ok,
|
|
|
|
* the check below will stop buffer overrun.
|
|
|
|
*/
|
|
|
|
if (index >= MAX_MC_EVT)
|
|
|
|
return;
|
|
|
|
|
|
|
|
/* Populate generic machine check info */
|
|
|
|
mce->version = MCE_V1;
|
2013-12-16 13:16:24 +08:00
|
|
|
mce->srr0 = nip;
|
2013-10-30 22:35:40 +08:00
|
|
|
mce->srr1 = regs->msr;
|
|
|
|
mce->gpr3 = regs->gpr[3];
|
|
|
|
mce->in_use = 1;
|
|
|
|
|
2016-08-09 13:09:13 +08:00
|
|
|
/* Mark it recovered if we have handled it and MSR(RI=1). */
|
|
|
|
if (handled && (regs->msr & MSR_RI))
|
2013-10-30 22:35:40 +08:00
|
|
|
mce->disposition = MCE_DISPOSITION_RECOVERED;
|
|
|
|
else
|
|
|
|
mce->disposition = MCE_DISPOSITION_NOT_RECOVERED;
|
2017-02-28 10:00:47 +08:00
|
|
|
|
|
|
|
mce->initiator = mce_err->initiator;
|
|
|
|
mce->severity = mce_err->severity;
|
2013-10-30 22:35:40 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Populate the mce error_type and type-specific error_type.
|
|
|
|
*/
|
|
|
|
mce_set_error_info(mce, mce_err);
|
|
|
|
|
|
|
|
if (!addr)
|
|
|
|
return;
|
|
|
|
|
|
|
|
if (mce->error_type == MCE_ERROR_TYPE_TLB) {
|
|
|
|
mce->u.tlb_error.effective_address_provided = true;
|
|
|
|
mce->u.tlb_error.effective_address = addr;
|
|
|
|
} else if (mce->error_type == MCE_ERROR_TYPE_SLB) {
|
|
|
|
mce->u.slb_error.effective_address_provided = true;
|
|
|
|
mce->u.slb_error.effective_address = addr;
|
|
|
|
} else if (mce->error_type == MCE_ERROR_TYPE_ERAT) {
|
|
|
|
mce->u.erat_error.effective_address_provided = true;
|
|
|
|
mce->u.erat_error.effective_address = addr;
|
2017-02-28 10:00:48 +08:00
|
|
|
} else if (mce->error_type == MCE_ERROR_TYPE_USER) {
|
|
|
|
mce->u.user_error.effective_address_provided = true;
|
|
|
|
mce->u.user_error.effective_address = addr;
|
|
|
|
} else if (mce->error_type == MCE_ERROR_TYPE_RA) {
|
|
|
|
mce->u.ra_error.effective_address_provided = true;
|
|
|
|
mce->u.ra_error.effective_address = addr;
|
|
|
|
} else if (mce->error_type == MCE_ERROR_TYPE_LINK) {
|
|
|
|
mce->u.link_error.effective_address_provided = true;
|
|
|
|
mce->u.link_error.effective_address = addr;
|
2013-10-30 22:35:40 +08:00
|
|
|
} else if (mce->error_type == MCE_ERROR_TYPE_UE) {
|
|
|
|
mce->u.ue_error.effective_address_provided = true;
|
|
|
|
mce->u.ue_error.effective_address = addr;
|
2017-09-29 12:26:53 +08:00
|
|
|
if (phys_addr != ULONG_MAX) {
|
|
|
|
mce->u.ue_error.physical_address_provided = true;
|
|
|
|
mce->u.ue_error.physical_address = phys_addr;
|
|
|
|
}
|
2013-10-30 22:35:40 +08:00
|
|
|
}
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* get_mce_event:
|
|
|
|
* mce Pointer to machine_check_event structure to be filled.
|
|
|
|
* release Flag to indicate whether to free the event slot or not.
|
|
|
|
* 0 <= do not release the mce event. Caller will invoke
|
|
|
|
* release_mce_event() once event has been consumed.
|
|
|
|
* 1 <= release the slot.
|
|
|
|
*
|
|
|
|
* return 1 = success
|
|
|
|
* 0 = failure
|
|
|
|
*
|
|
|
|
* get_mce_event() will be called by platform specific machine check
|
|
|
|
* handle routine and in KVM.
|
|
|
|
* When we call get_mce_event(), we are still in interrupt context and
|
|
|
|
* preemption will not be scheduled until ret_from_expect() routine
|
|
|
|
* is called.
|
|
|
|
*/
|
|
|
|
int get_mce_event(struct machine_check_event *mce, bool release)
|
|
|
|
{
|
powerpc: Replace __get_cpu_var uses
This still has not been merged and now powerpc is the only arch that does
not have this change. Sorry about missing linuxppc-dev before.
V2->V2
- Fix up to work against 3.18-rc1
__get_cpu_var() is used for multiple purposes in the kernel source. One of
them is address calculation via the form &__get_cpu_var(x). This calculates
the address for the instance of the percpu variable of the current processor
based on an offset.
Other use cases are for storing and retrieving data from the current
processors percpu area. __get_cpu_var() can be used as an lvalue when
writing data or on the right side of an assignment.
__get_cpu_var() is defined as :
__get_cpu_var() always only does an address determination. However, store
and retrieve operations could use a segment prefix (or global register on
other platforms) to avoid the address calculation.
this_cpu_write() and this_cpu_read() can directly take an offset into a
percpu area and use optimized assembly code to read and write per cpu
variables.
This patch converts __get_cpu_var into either an explicit address
calculation using this_cpu_ptr() or into a use of this_cpu operations that
use the offset. Thereby address calculations are avoided and less registers
are used when code is generated.
At the end of the patch set all uses of __get_cpu_var have been removed so
the macro is removed too.
The patch set includes passes over all arches as well. Once these operations
are used throughout then specialized macros can be defined in non -x86
arches as well in order to optimize per cpu access by f.e. using a global
register that may be set to the per cpu base.
Transformations done to __get_cpu_var()
1. Determine the address of the percpu instance of the current processor.
DEFINE_PER_CPU(int, y);
int *x = &__get_cpu_var(y);
Converts to
int *x = this_cpu_ptr(&y);
2. Same as #1 but this time an array structure is involved.
DEFINE_PER_CPU(int, y[20]);
int *x = __get_cpu_var(y);
Converts to
int *x = this_cpu_ptr(y);
3. Retrieve the content of the current processors instance of a per cpu
variable.
DEFINE_PER_CPU(int, y);
int x = __get_cpu_var(y)
Converts to
int x = __this_cpu_read(y);
4. Retrieve the content of a percpu struct
DEFINE_PER_CPU(struct mystruct, y);
struct mystruct x = __get_cpu_var(y);
Converts to
memcpy(&x, this_cpu_ptr(&y), sizeof(x));
5. Assignment to a per cpu variable
DEFINE_PER_CPU(int, y)
__get_cpu_var(y) = x;
Converts to
__this_cpu_write(y, x);
6. Increment/Decrement etc of a per cpu variable
DEFINE_PER_CPU(int, y);
__get_cpu_var(y)++
Converts to
__this_cpu_inc(y)
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
CC: Paul Mackerras <paulus@samba.org>
Signed-off-by: Christoph Lameter <cl@linux.com>
[mpe: Fix build errors caused by set/or_softirq_pending(), and rework
assignment in __set_breakpoint() to use memcpy().]
Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
2014-10-22 04:23:25 +08:00
|
|
|
int index = __this_cpu_read(mce_nest_count) - 1;
|
2013-10-30 22:35:40 +08:00
|
|
|
struct machine_check_event *mc_evt;
|
|
|
|
int ret = 0;
|
|
|
|
|
|
|
|
/* Sanity check */
|
|
|
|
if (index < 0)
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
/* Check if we have MCE info to process. */
|
|
|
|
if (index < MAX_MC_EVT) {
|
powerpc: Replace __get_cpu_var uses
This still has not been merged and now powerpc is the only arch that does
not have this change. Sorry about missing linuxppc-dev before.
V2->V2
- Fix up to work against 3.18-rc1
__get_cpu_var() is used for multiple purposes in the kernel source. One of
them is address calculation via the form &__get_cpu_var(x). This calculates
the address for the instance of the percpu variable of the current processor
based on an offset.
Other use cases are for storing and retrieving data from the current
processors percpu area. __get_cpu_var() can be used as an lvalue when
writing data or on the right side of an assignment.
__get_cpu_var() is defined as :
__get_cpu_var() always only does an address determination. However, store
and retrieve operations could use a segment prefix (or global register on
other platforms) to avoid the address calculation.
this_cpu_write() and this_cpu_read() can directly take an offset into a
percpu area and use optimized assembly code to read and write per cpu
variables.
This patch converts __get_cpu_var into either an explicit address
calculation using this_cpu_ptr() or into a use of this_cpu operations that
use the offset. Thereby address calculations are avoided and less registers
are used when code is generated.
At the end of the patch set all uses of __get_cpu_var have been removed so
the macro is removed too.
The patch set includes passes over all arches as well. Once these operations
are used throughout then specialized macros can be defined in non -x86
arches as well in order to optimize per cpu access by f.e. using a global
register that may be set to the per cpu base.
Transformations done to __get_cpu_var()
1. Determine the address of the percpu instance of the current processor.
DEFINE_PER_CPU(int, y);
int *x = &__get_cpu_var(y);
Converts to
int *x = this_cpu_ptr(&y);
2. Same as #1 but this time an array structure is involved.
DEFINE_PER_CPU(int, y[20]);
int *x = __get_cpu_var(y);
Converts to
int *x = this_cpu_ptr(y);
3. Retrieve the content of the current processors instance of a per cpu
variable.
DEFINE_PER_CPU(int, y);
int x = __get_cpu_var(y)
Converts to
int x = __this_cpu_read(y);
4. Retrieve the content of a percpu struct
DEFINE_PER_CPU(struct mystruct, y);
struct mystruct x = __get_cpu_var(y);
Converts to
memcpy(&x, this_cpu_ptr(&y), sizeof(x));
5. Assignment to a per cpu variable
DEFINE_PER_CPU(int, y)
__get_cpu_var(y) = x;
Converts to
__this_cpu_write(y, x);
6. Increment/Decrement etc of a per cpu variable
DEFINE_PER_CPU(int, y);
__get_cpu_var(y)++
Converts to
__this_cpu_inc(y)
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
CC: Paul Mackerras <paulus@samba.org>
Signed-off-by: Christoph Lameter <cl@linux.com>
[mpe: Fix build errors caused by set/or_softirq_pending(), and rework
assignment in __set_breakpoint() to use memcpy().]
Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
2014-10-22 04:23:25 +08:00
|
|
|
mc_evt = this_cpu_ptr(&mce_event[index]);
|
2013-10-30 22:35:40 +08:00
|
|
|
/* Copy the event structure and release the original */
|
|
|
|
if (mce)
|
|
|
|
*mce = *mc_evt;
|
|
|
|
if (release)
|
|
|
|
mc_evt->in_use = 0;
|
|
|
|
ret = 1;
|
|
|
|
}
|
|
|
|
/* Decrement the count to free the slot. */
|
|
|
|
if (release)
|
powerpc: Replace __get_cpu_var uses
This still has not been merged and now powerpc is the only arch that does
not have this change. Sorry about missing linuxppc-dev before.
V2->V2
- Fix up to work against 3.18-rc1
__get_cpu_var() is used for multiple purposes in the kernel source. One of
them is address calculation via the form &__get_cpu_var(x). This calculates
the address for the instance of the percpu variable of the current processor
based on an offset.
Other use cases are for storing and retrieving data from the current
processors percpu area. __get_cpu_var() can be used as an lvalue when
writing data or on the right side of an assignment.
__get_cpu_var() is defined as :
__get_cpu_var() always only does an address determination. However, store
and retrieve operations could use a segment prefix (or global register on
other platforms) to avoid the address calculation.
this_cpu_write() and this_cpu_read() can directly take an offset into a
percpu area and use optimized assembly code to read and write per cpu
variables.
This patch converts __get_cpu_var into either an explicit address
calculation using this_cpu_ptr() or into a use of this_cpu operations that
use the offset. Thereby address calculations are avoided and less registers
are used when code is generated.
At the end of the patch set all uses of __get_cpu_var have been removed so
the macro is removed too.
The patch set includes passes over all arches as well. Once these operations
are used throughout then specialized macros can be defined in non -x86
arches as well in order to optimize per cpu access by f.e. using a global
register that may be set to the per cpu base.
Transformations done to __get_cpu_var()
1. Determine the address of the percpu instance of the current processor.
DEFINE_PER_CPU(int, y);
int *x = &__get_cpu_var(y);
Converts to
int *x = this_cpu_ptr(&y);
2. Same as #1 but this time an array structure is involved.
DEFINE_PER_CPU(int, y[20]);
int *x = __get_cpu_var(y);
Converts to
int *x = this_cpu_ptr(y);
3. Retrieve the content of the current processors instance of a per cpu
variable.
DEFINE_PER_CPU(int, y);
int x = __get_cpu_var(y)
Converts to
int x = __this_cpu_read(y);
4. Retrieve the content of a percpu struct
DEFINE_PER_CPU(struct mystruct, y);
struct mystruct x = __get_cpu_var(y);
Converts to
memcpy(&x, this_cpu_ptr(&y), sizeof(x));
5. Assignment to a per cpu variable
DEFINE_PER_CPU(int, y)
__get_cpu_var(y) = x;
Converts to
__this_cpu_write(y, x);
6. Increment/Decrement etc of a per cpu variable
DEFINE_PER_CPU(int, y);
__get_cpu_var(y)++
Converts to
__this_cpu_inc(y)
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
CC: Paul Mackerras <paulus@samba.org>
Signed-off-by: Christoph Lameter <cl@linux.com>
[mpe: Fix build errors caused by set/or_softirq_pending(), and rework
assignment in __set_breakpoint() to use memcpy().]
Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
2014-10-22 04:23:25 +08:00
|
|
|
__this_cpu_dec(mce_nest_count);
|
2013-10-30 22:35:40 +08:00
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
void release_mce_event(void)
|
|
|
|
{
|
|
|
|
get_mce_event(NULL, true);
|
|
|
|
}
|
2013-10-30 22:35:49 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Queue up the MCE event which then can be handled later.
|
|
|
|
*/
|
|
|
|
void machine_check_queue_event(void)
|
|
|
|
{
|
|
|
|
int index;
|
|
|
|
struct machine_check_event evt;
|
|
|
|
|
|
|
|
if (!get_mce_event(&evt, MCE_EVENT_RELEASE))
|
|
|
|
return;
|
|
|
|
|
2015-05-12 11:23:59 +08:00
|
|
|
index = __this_cpu_inc_return(mce_queue_count) - 1;
|
2013-10-30 22:35:49 +08:00
|
|
|
/* If queue is full, just return for now. */
|
|
|
|
if (index >= MAX_MC_EVT) {
|
powerpc: Replace __get_cpu_var uses
This still has not been merged and now powerpc is the only arch that does
not have this change. Sorry about missing linuxppc-dev before.
V2->V2
- Fix up to work against 3.18-rc1
__get_cpu_var() is used for multiple purposes in the kernel source. One of
them is address calculation via the form &__get_cpu_var(x). This calculates
the address for the instance of the percpu variable of the current processor
based on an offset.
Other use cases are for storing and retrieving data from the current
processors percpu area. __get_cpu_var() can be used as an lvalue when
writing data or on the right side of an assignment.
__get_cpu_var() is defined as :
__get_cpu_var() always only does an address determination. However, store
and retrieve operations could use a segment prefix (or global register on
other platforms) to avoid the address calculation.
this_cpu_write() and this_cpu_read() can directly take an offset into a
percpu area and use optimized assembly code to read and write per cpu
variables.
This patch converts __get_cpu_var into either an explicit address
calculation using this_cpu_ptr() or into a use of this_cpu operations that
use the offset. Thereby address calculations are avoided and less registers
are used when code is generated.
At the end of the patch set all uses of __get_cpu_var have been removed so
the macro is removed too.
The patch set includes passes over all arches as well. Once these operations
are used throughout then specialized macros can be defined in non -x86
arches as well in order to optimize per cpu access by f.e. using a global
register that may be set to the per cpu base.
Transformations done to __get_cpu_var()
1. Determine the address of the percpu instance of the current processor.
DEFINE_PER_CPU(int, y);
int *x = &__get_cpu_var(y);
Converts to
int *x = this_cpu_ptr(&y);
2. Same as #1 but this time an array structure is involved.
DEFINE_PER_CPU(int, y[20]);
int *x = __get_cpu_var(y);
Converts to
int *x = this_cpu_ptr(y);
3. Retrieve the content of the current processors instance of a per cpu
variable.
DEFINE_PER_CPU(int, y);
int x = __get_cpu_var(y)
Converts to
int x = __this_cpu_read(y);
4. Retrieve the content of a percpu struct
DEFINE_PER_CPU(struct mystruct, y);
struct mystruct x = __get_cpu_var(y);
Converts to
memcpy(&x, this_cpu_ptr(&y), sizeof(x));
5. Assignment to a per cpu variable
DEFINE_PER_CPU(int, y)
__get_cpu_var(y) = x;
Converts to
__this_cpu_write(y, x);
6. Increment/Decrement etc of a per cpu variable
DEFINE_PER_CPU(int, y);
__get_cpu_var(y)++
Converts to
__this_cpu_inc(y)
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
CC: Paul Mackerras <paulus@samba.org>
Signed-off-by: Christoph Lameter <cl@linux.com>
[mpe: Fix build errors caused by set/or_softirq_pending(), and rework
assignment in __set_breakpoint() to use memcpy().]
Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
2014-10-22 04:23:25 +08:00
|
|
|
__this_cpu_dec(mce_queue_count);
|
2013-10-30 22:35:49 +08:00
|
|
|
return;
|
|
|
|
}
|
powerpc: Replace __get_cpu_var uses
This still has not been merged and now powerpc is the only arch that does
not have this change. Sorry about missing linuxppc-dev before.
V2->V2
- Fix up to work against 3.18-rc1
__get_cpu_var() is used for multiple purposes in the kernel source. One of
them is address calculation via the form &__get_cpu_var(x). This calculates
the address for the instance of the percpu variable of the current processor
based on an offset.
Other use cases are for storing and retrieving data from the current
processors percpu area. __get_cpu_var() can be used as an lvalue when
writing data or on the right side of an assignment.
__get_cpu_var() is defined as :
__get_cpu_var() always only does an address determination. However, store
and retrieve operations could use a segment prefix (or global register on
other platforms) to avoid the address calculation.
this_cpu_write() and this_cpu_read() can directly take an offset into a
percpu area and use optimized assembly code to read and write per cpu
variables.
This patch converts __get_cpu_var into either an explicit address
calculation using this_cpu_ptr() or into a use of this_cpu operations that
use the offset. Thereby address calculations are avoided and less registers
are used when code is generated.
At the end of the patch set all uses of __get_cpu_var have been removed so
the macro is removed too.
The patch set includes passes over all arches as well. Once these operations
are used throughout then specialized macros can be defined in non -x86
arches as well in order to optimize per cpu access by f.e. using a global
register that may be set to the per cpu base.
Transformations done to __get_cpu_var()
1. Determine the address of the percpu instance of the current processor.
DEFINE_PER_CPU(int, y);
int *x = &__get_cpu_var(y);
Converts to
int *x = this_cpu_ptr(&y);
2. Same as #1 but this time an array structure is involved.
DEFINE_PER_CPU(int, y[20]);
int *x = __get_cpu_var(y);
Converts to
int *x = this_cpu_ptr(y);
3. Retrieve the content of the current processors instance of a per cpu
variable.
DEFINE_PER_CPU(int, y);
int x = __get_cpu_var(y)
Converts to
int x = __this_cpu_read(y);
4. Retrieve the content of a percpu struct
DEFINE_PER_CPU(struct mystruct, y);
struct mystruct x = __get_cpu_var(y);
Converts to
memcpy(&x, this_cpu_ptr(&y), sizeof(x));
5. Assignment to a per cpu variable
DEFINE_PER_CPU(int, y)
__get_cpu_var(y) = x;
Converts to
__this_cpu_write(y, x);
6. Increment/Decrement etc of a per cpu variable
DEFINE_PER_CPU(int, y);
__get_cpu_var(y)++
Converts to
__this_cpu_inc(y)
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
CC: Paul Mackerras <paulus@samba.org>
Signed-off-by: Christoph Lameter <cl@linux.com>
[mpe: Fix build errors caused by set/or_softirq_pending(), and rework
assignment in __set_breakpoint() to use memcpy().]
Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
2014-10-22 04:23:25 +08:00
|
|
|
memcpy(this_cpu_ptr(&mce_event_queue[index]), &evt, sizeof(evt));
|
2014-01-14 18:15:09 +08:00
|
|
|
|
|
|
|
/* Queue irq work to process this event later. */
|
|
|
|
irq_work_queue(&mce_event_process_work);
|
2013-10-30 22:35:49 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* process pending MCE event from the mce event queue. This function will be
|
|
|
|
* called during syscall exit.
|
|
|
|
*/
|
2014-01-14 18:15:09 +08:00
|
|
|
static void machine_check_process_queued_event(struct irq_work *work)
|
2013-10-30 22:35:49 +08:00
|
|
|
{
|
|
|
|
int index;
|
|
|
|
|
2017-04-19 00:38:17 +08:00
|
|
|
add_taint(TAINT_MACHINE_CHECK, LOCKDEP_NOW_UNRELIABLE);
|
|
|
|
|
2013-10-30 22:35:49 +08:00
|
|
|
/*
|
|
|
|
* For now just print it to console.
|
|
|
|
* TODO: log this error event to FSP or nvram.
|
|
|
|
*/
|
powerpc: Replace __get_cpu_var uses
This still has not been merged and now powerpc is the only arch that does
not have this change. Sorry about missing linuxppc-dev before.
V2->V2
- Fix up to work against 3.18-rc1
__get_cpu_var() is used for multiple purposes in the kernel source. One of
them is address calculation via the form &__get_cpu_var(x). This calculates
the address for the instance of the percpu variable of the current processor
based on an offset.
Other use cases are for storing and retrieving data from the current
processors percpu area. __get_cpu_var() can be used as an lvalue when
writing data or on the right side of an assignment.
__get_cpu_var() is defined as :
__get_cpu_var() always only does an address determination. However, store
and retrieve operations could use a segment prefix (or global register on
other platforms) to avoid the address calculation.
this_cpu_write() and this_cpu_read() can directly take an offset into a
percpu area and use optimized assembly code to read and write per cpu
variables.
This patch converts __get_cpu_var into either an explicit address
calculation using this_cpu_ptr() or into a use of this_cpu operations that
use the offset. Thereby address calculations are avoided and less registers
are used when code is generated.
At the end of the patch set all uses of __get_cpu_var have been removed so
the macro is removed too.
The patch set includes passes over all arches as well. Once these operations
are used throughout then specialized macros can be defined in non -x86
arches as well in order to optimize per cpu access by f.e. using a global
register that may be set to the per cpu base.
Transformations done to __get_cpu_var()
1. Determine the address of the percpu instance of the current processor.
DEFINE_PER_CPU(int, y);
int *x = &__get_cpu_var(y);
Converts to
int *x = this_cpu_ptr(&y);
2. Same as #1 but this time an array structure is involved.
DEFINE_PER_CPU(int, y[20]);
int *x = __get_cpu_var(y);
Converts to
int *x = this_cpu_ptr(y);
3. Retrieve the content of the current processors instance of a per cpu
variable.
DEFINE_PER_CPU(int, y);
int x = __get_cpu_var(y)
Converts to
int x = __this_cpu_read(y);
4. Retrieve the content of a percpu struct
DEFINE_PER_CPU(struct mystruct, y);
struct mystruct x = __get_cpu_var(y);
Converts to
memcpy(&x, this_cpu_ptr(&y), sizeof(x));
5. Assignment to a per cpu variable
DEFINE_PER_CPU(int, y)
__get_cpu_var(y) = x;
Converts to
__this_cpu_write(y, x);
6. Increment/Decrement etc of a per cpu variable
DEFINE_PER_CPU(int, y);
__get_cpu_var(y)++
Converts to
__this_cpu_inc(y)
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
CC: Paul Mackerras <paulus@samba.org>
Signed-off-by: Christoph Lameter <cl@linux.com>
[mpe: Fix build errors caused by set/or_softirq_pending(), and rework
assignment in __set_breakpoint() to use memcpy().]
Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
2014-10-22 04:23:25 +08:00
|
|
|
while (__this_cpu_read(mce_queue_count) > 0) {
|
|
|
|
index = __this_cpu_read(mce_queue_count) - 1;
|
2013-10-30 22:35:49 +08:00
|
|
|
machine_check_print_event_info(
|
2017-04-03 13:29:34 +08:00
|
|
|
this_cpu_ptr(&mce_event_queue[index]), false);
|
powerpc: Replace __get_cpu_var uses
This still has not been merged and now powerpc is the only arch that does
not have this change. Sorry about missing linuxppc-dev before.
V2->V2
- Fix up to work against 3.18-rc1
__get_cpu_var() is used for multiple purposes in the kernel source. One of
them is address calculation via the form &__get_cpu_var(x). This calculates
the address for the instance of the percpu variable of the current processor
based on an offset.
Other use cases are for storing and retrieving data from the current
processors percpu area. __get_cpu_var() can be used as an lvalue when
writing data or on the right side of an assignment.
__get_cpu_var() is defined as :
__get_cpu_var() always only does an address determination. However, store
and retrieve operations could use a segment prefix (or global register on
other platforms) to avoid the address calculation.
this_cpu_write() and this_cpu_read() can directly take an offset into a
percpu area and use optimized assembly code to read and write per cpu
variables.
This patch converts __get_cpu_var into either an explicit address
calculation using this_cpu_ptr() or into a use of this_cpu operations that
use the offset. Thereby address calculations are avoided and less registers
are used when code is generated.
At the end of the patch set all uses of __get_cpu_var have been removed so
the macro is removed too.
The patch set includes passes over all arches as well. Once these operations
are used throughout then specialized macros can be defined in non -x86
arches as well in order to optimize per cpu access by f.e. using a global
register that may be set to the per cpu base.
Transformations done to __get_cpu_var()
1. Determine the address of the percpu instance of the current processor.
DEFINE_PER_CPU(int, y);
int *x = &__get_cpu_var(y);
Converts to
int *x = this_cpu_ptr(&y);
2. Same as #1 but this time an array structure is involved.
DEFINE_PER_CPU(int, y[20]);
int *x = __get_cpu_var(y);
Converts to
int *x = this_cpu_ptr(y);
3. Retrieve the content of the current processors instance of a per cpu
variable.
DEFINE_PER_CPU(int, y);
int x = __get_cpu_var(y)
Converts to
int x = __this_cpu_read(y);
4. Retrieve the content of a percpu struct
DEFINE_PER_CPU(struct mystruct, y);
struct mystruct x = __get_cpu_var(y);
Converts to
memcpy(&x, this_cpu_ptr(&y), sizeof(x));
5. Assignment to a per cpu variable
DEFINE_PER_CPU(int, y)
__get_cpu_var(y) = x;
Converts to
__this_cpu_write(y, x);
6. Increment/Decrement etc of a per cpu variable
DEFINE_PER_CPU(int, y);
__get_cpu_var(y)++
Converts to
__this_cpu_inc(y)
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
CC: Paul Mackerras <paulus@samba.org>
Signed-off-by: Christoph Lameter <cl@linux.com>
[mpe: Fix build errors caused by set/or_softirq_pending(), and rework
assignment in __set_breakpoint() to use memcpy().]
Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
2014-10-22 04:23:25 +08:00
|
|
|
__this_cpu_dec(mce_queue_count);
|
2013-10-30 22:35:49 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-04-03 13:29:34 +08:00
|
|
|
void machine_check_print_event_info(struct machine_check_event *evt,
|
|
|
|
bool user_mode)
|
2013-10-30 22:35:49 +08:00
|
|
|
{
|
|
|
|
const char *level, *sevstr, *subtype;
|
|
|
|
static const char *mc_ue_types[] = {
|
|
|
|
"Indeterminate",
|
|
|
|
"Instruction fetch",
|
|
|
|
"Page table walk ifetch",
|
|
|
|
"Load/Store",
|
|
|
|
"Page table walk Load/Store",
|
|
|
|
};
|
|
|
|
static const char *mc_slb_types[] = {
|
|
|
|
"Indeterminate",
|
|
|
|
"Parity",
|
|
|
|
"Multihit",
|
|
|
|
};
|
|
|
|
static const char *mc_erat_types[] = {
|
|
|
|
"Indeterminate",
|
|
|
|
"Parity",
|
|
|
|
"Multihit",
|
|
|
|
};
|
|
|
|
static const char *mc_tlb_types[] = {
|
|
|
|
"Indeterminate",
|
|
|
|
"Parity",
|
|
|
|
"Multihit",
|
|
|
|
};
|
2017-02-28 10:00:48 +08:00
|
|
|
static const char *mc_user_types[] = {
|
|
|
|
"Indeterminate",
|
|
|
|
"tlbie(l) invalid",
|
|
|
|
};
|
|
|
|
static const char *mc_ra_types[] = {
|
|
|
|
"Indeterminate",
|
|
|
|
"Instruction fetch (bad)",
|
2017-05-29 14:26:44 +08:00
|
|
|
"Instruction fetch (foreign)",
|
2017-02-28 10:00:48 +08:00
|
|
|
"Page table walk ifetch (bad)",
|
|
|
|
"Page table walk ifetch (foreign)",
|
|
|
|
"Load (bad)",
|
|
|
|
"Store (bad)",
|
|
|
|
"Page table walk Load/Store (bad)",
|
|
|
|
"Page table walk Load/Store (foreign)",
|
|
|
|
"Load/Store (foreign)",
|
|
|
|
};
|
|
|
|
static const char *mc_link_types[] = {
|
|
|
|
"Indeterminate",
|
|
|
|
"Instruction fetch (timeout)",
|
|
|
|
"Page table walk ifetch (timeout)",
|
|
|
|
"Load (timeout)",
|
|
|
|
"Store (timeout)",
|
|
|
|
"Page table walk Load/Store (timeout)",
|
|
|
|
};
|
2013-10-30 22:35:49 +08:00
|
|
|
|
|
|
|
/* Print things out */
|
|
|
|
if (evt->version != MCE_V1) {
|
|
|
|
pr_err("Machine Check Exception, Unknown event version %d !\n",
|
|
|
|
evt->version);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
switch (evt->severity) {
|
|
|
|
case MCE_SEV_NO_ERROR:
|
|
|
|
level = KERN_INFO;
|
|
|
|
sevstr = "Harmless";
|
|
|
|
break;
|
|
|
|
case MCE_SEV_WARNING:
|
|
|
|
level = KERN_WARNING;
|
|
|
|
sevstr = "";
|
|
|
|
break;
|
|
|
|
case MCE_SEV_ERROR_SYNC:
|
|
|
|
level = KERN_ERR;
|
|
|
|
sevstr = "Severe";
|
|
|
|
break;
|
|
|
|
case MCE_SEV_FATAL:
|
|
|
|
default:
|
|
|
|
level = KERN_ERR;
|
|
|
|
sevstr = "Fatal";
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
printk("%s%s Machine check interrupt [%s]\n", level, sevstr,
|
|
|
|
evt->disposition == MCE_DISPOSITION_RECOVERED ?
|
2017-03-14 20:36:43 +08:00
|
|
|
"Recovered" : "Not recovered");
|
2017-04-03 13:29:34 +08:00
|
|
|
|
|
|
|
if (user_mode) {
|
|
|
|
printk("%s NIP: [%016llx] PID: %d Comm: %s\n", level,
|
|
|
|
evt->srr0, current->pid, current->comm);
|
|
|
|
} else {
|
|
|
|
printk("%s NIP [%016llx]: %pS\n", level, evt->srr0,
|
|
|
|
(void *)evt->srr0);
|
|
|
|
}
|
|
|
|
|
2013-10-30 22:35:49 +08:00
|
|
|
printk("%s Initiator: %s\n", level,
|
|
|
|
evt->initiator == MCE_INITIATOR_CPU ? "CPU" : "Unknown");
|
|
|
|
switch (evt->error_type) {
|
|
|
|
case MCE_ERROR_TYPE_UE:
|
|
|
|
subtype = evt->u.ue_error.ue_error_type <
|
|
|
|
ARRAY_SIZE(mc_ue_types) ?
|
|
|
|
mc_ue_types[evt->u.ue_error.ue_error_type]
|
|
|
|
: "Unknown";
|
|
|
|
printk("%s Error type: UE [%s]\n", level, subtype);
|
|
|
|
if (evt->u.ue_error.effective_address_provided)
|
|
|
|
printk("%s Effective address: %016llx\n",
|
|
|
|
level, evt->u.ue_error.effective_address);
|
|
|
|
if (evt->u.ue_error.physical_address_provided)
|
2017-09-29 12:26:52 +08:00
|
|
|
printk("%s Physical address: %016llx\n",
|
2013-10-30 22:35:49 +08:00
|
|
|
level, evt->u.ue_error.physical_address);
|
|
|
|
break;
|
|
|
|
case MCE_ERROR_TYPE_SLB:
|
|
|
|
subtype = evt->u.slb_error.slb_error_type <
|
|
|
|
ARRAY_SIZE(mc_slb_types) ?
|
|
|
|
mc_slb_types[evt->u.slb_error.slb_error_type]
|
|
|
|
: "Unknown";
|
|
|
|
printk("%s Error type: SLB [%s]\n", level, subtype);
|
|
|
|
if (evt->u.slb_error.effective_address_provided)
|
|
|
|
printk("%s Effective address: %016llx\n",
|
|
|
|
level, evt->u.slb_error.effective_address);
|
|
|
|
break;
|
|
|
|
case MCE_ERROR_TYPE_ERAT:
|
|
|
|
subtype = evt->u.erat_error.erat_error_type <
|
|
|
|
ARRAY_SIZE(mc_erat_types) ?
|
|
|
|
mc_erat_types[evt->u.erat_error.erat_error_type]
|
|
|
|
: "Unknown";
|
|
|
|
printk("%s Error type: ERAT [%s]\n", level, subtype);
|
|
|
|
if (evt->u.erat_error.effective_address_provided)
|
|
|
|
printk("%s Effective address: %016llx\n",
|
|
|
|
level, evt->u.erat_error.effective_address);
|
|
|
|
break;
|
|
|
|
case MCE_ERROR_TYPE_TLB:
|
|
|
|
subtype = evt->u.tlb_error.tlb_error_type <
|
|
|
|
ARRAY_SIZE(mc_tlb_types) ?
|
|
|
|
mc_tlb_types[evt->u.tlb_error.tlb_error_type]
|
|
|
|
: "Unknown";
|
|
|
|
printk("%s Error type: TLB [%s]\n", level, subtype);
|
|
|
|
if (evt->u.tlb_error.effective_address_provided)
|
|
|
|
printk("%s Effective address: %016llx\n",
|
|
|
|
level, evt->u.tlb_error.effective_address);
|
|
|
|
break;
|
2017-02-28 10:00:48 +08:00
|
|
|
case MCE_ERROR_TYPE_USER:
|
|
|
|
subtype = evt->u.user_error.user_error_type <
|
|
|
|
ARRAY_SIZE(mc_user_types) ?
|
|
|
|
mc_user_types[evt->u.user_error.user_error_type]
|
|
|
|
: "Unknown";
|
|
|
|
printk("%s Error type: User [%s]\n", level, subtype);
|
|
|
|
if (evt->u.user_error.effective_address_provided)
|
|
|
|
printk("%s Effective address: %016llx\n",
|
|
|
|
level, evt->u.user_error.effective_address);
|
|
|
|
break;
|
|
|
|
case MCE_ERROR_TYPE_RA:
|
|
|
|
subtype = evt->u.ra_error.ra_error_type <
|
|
|
|
ARRAY_SIZE(mc_ra_types) ?
|
|
|
|
mc_ra_types[evt->u.ra_error.ra_error_type]
|
|
|
|
: "Unknown";
|
|
|
|
printk("%s Error type: Real address [%s]\n", level, subtype);
|
|
|
|
if (evt->u.ra_error.effective_address_provided)
|
|
|
|
printk("%s Effective address: %016llx\n",
|
|
|
|
level, evt->u.ra_error.effective_address);
|
|
|
|
break;
|
|
|
|
case MCE_ERROR_TYPE_LINK:
|
|
|
|
subtype = evt->u.link_error.link_error_type <
|
|
|
|
ARRAY_SIZE(mc_link_types) ?
|
|
|
|
mc_link_types[evt->u.link_error.link_error_type]
|
|
|
|
: "Unknown";
|
|
|
|
printk("%s Error type: Link [%s]\n", level, subtype);
|
|
|
|
if (evt->u.link_error.effective_address_provided)
|
|
|
|
printk("%s Effective address: %016llx\n",
|
|
|
|
level, evt->u.link_error.effective_address);
|
|
|
|
break;
|
2013-10-30 22:35:49 +08:00
|
|
|
default:
|
|
|
|
case MCE_ERROR_TYPE_UNKNOWN:
|
|
|
|
printk("%s Error type: Unknown\n", level);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
2017-05-11 19:03:12 +08:00
|
|
|
EXPORT_SYMBOL_GPL(machine_check_print_event_info);
|
2013-10-30 22:36:13 +08:00
|
|
|
|
2017-08-08 14:39:24 +08:00
|
|
|
/*
|
|
|
|
* This function is called in real mode. Strictly no printk's please.
|
|
|
|
*
|
|
|
|
* regs->nip and regs->msr contains srr0 and ssr1.
|
|
|
|
*/
|
|
|
|
long machine_check_early(struct pt_regs *regs)
|
|
|
|
{
|
|
|
|
long handled = 0;
|
|
|
|
|
|
|
|
__this_cpu_inc(irq_stat.mce_exceptions);
|
|
|
|
|
|
|
|
if (cur_cpu_spec && cur_cpu_spec->machine_check_early)
|
|
|
|
handled = cur_cpu_spec->machine_check_early(regs);
|
|
|
|
return handled;
|
|
|
|
}
|
|
|
|
|
|
|
|
long hmi_exception_realmode(struct pt_regs *regs)
|
|
|
|
{
|
|
|
|
__this_cpu_inc(irq_stat.hmi_exceptions);
|
|
|
|
|
2017-09-15 13:25:48 +08:00
|
|
|
#ifdef CONFIG_PPC_BOOK3S_64
|
|
|
|
/* Workaround for P9 vector CI loads (see p9_hmi_special_emu) */
|
|
|
|
if (pvr_version_is(PVR_POWER9)) {
|
|
|
|
unsigned long hmer = mfspr(SPRN_HMER);
|
|
|
|
|
|
|
|
/* Do we have the debug bit set */
|
|
|
|
if (hmer & PPC_BIT(17)) {
|
|
|
|
hmer &= ~PPC_BIT(17);
|
|
|
|
mtspr(SPRN_HMER, hmer);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Now to avoid problems with soft-disable we
|
|
|
|
* only do the emulation if we are coming from
|
|
|
|
* user space
|
|
|
|
*/
|
|
|
|
if (user_mode(regs))
|
|
|
|
local_paca->hmi_p9_special_emu = 1;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Don't bother going to OPAL if that's the
|
|
|
|
* only relevant bit.
|
|
|
|
*/
|
|
|
|
if (!(hmer & mfspr(SPRN_HMEER)))
|
|
|
|
return local_paca->hmi_p9_special_emu;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
#endif /* CONFIG_PPC_BOOK3S_64 */
|
|
|
|
|
2017-08-08 14:39:24 +08:00
|
|
|
wait_for_subcore_guest_exit();
|
|
|
|
|
|
|
|
if (ppc_md.hmi_exception_early)
|
|
|
|
ppc_md.hmi_exception_early(regs);
|
|
|
|
|
|
|
|
wait_for_tb_resync();
|
|
|
|
|
2017-09-15 13:25:48 +08:00
|
|
|
return 1;
|
2017-08-08 14:39:24 +08:00
|
|
|
}
|