mirror of https://gitee.com/openkylin/linux.git
perf_events, x86: Implement intel core solo/duo support
Implement Intel Core Solo/Duo, aka. Intel Architectural Performance Monitoring Version 1. Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl> Cc: Peter Zijlstra <a.p.zijlstra@chello.nl> Cc: Mike Galbraith <efault@gmx.de> Cc: Paul Mackerras <paulus@samba.org> Cc: Arnaldo Carvalho de Melo <acme@redhat.com> Cc: Frederic Weisbecker <fweisbec@gmail.com> Cc: Arjan van de Ven <arjan@linux.intel.com> LKML-Reference: <new-submission> Signed-off-by: Ingo Molnar <mingo@elte.hu>
This commit is contained in:
parent
9717e6cd3d
commit
8c48e44419
|
@ -227,6 +227,17 @@ static const u64 intel_perfmon_event_map[] =
|
||||||
};
|
};
|
||||||
|
|
||||||
static struct event_constraint intel_core_event_constraints[] =
|
static struct event_constraint intel_core_event_constraints[] =
|
||||||
|
{
|
||||||
|
INTEL_EVENT_CONSTRAINT(0x11, 0x2), /* FP_ASSIST */
|
||||||
|
INTEL_EVENT_CONSTRAINT(0x12, 0x2), /* MUL */
|
||||||
|
INTEL_EVENT_CONSTRAINT(0x13, 0x2), /* DIV */
|
||||||
|
INTEL_EVENT_CONSTRAINT(0x14, 0x1), /* CYCLES_DIV_BUSY */
|
||||||
|
INTEL_EVENT_CONSTRAINT(0x19, 0x2), /* DELAYED_BYPASS */
|
||||||
|
INTEL_EVENT_CONSTRAINT(0xc1, 0x1), /* FP_COMP_INSTR_RET */
|
||||||
|
EVENT_CONSTRAINT_END
|
||||||
|
};
|
||||||
|
|
||||||
|
static struct event_constraint intel_core2_event_constraints[] =
|
||||||
{
|
{
|
||||||
FIXED_EVENT_CONSTRAINT(0xc0, (0x3|(1ULL<<32))), /* INSTRUCTIONS_RETIRED */
|
FIXED_EVENT_CONSTRAINT(0xc0, (0x3|(1ULL<<32))), /* INSTRUCTIONS_RETIRED */
|
||||||
FIXED_EVENT_CONSTRAINT(0x3c, (0x3|(1ULL<<33))), /* UNHALTED_CORE_CYCLES */
|
FIXED_EVENT_CONSTRAINT(0x3c, (0x3|(1ULL<<33))), /* UNHALTED_CORE_CYCLES */
|
||||||
|
@ -1216,7 +1227,7 @@ static void intel_pmu_disable_all(void)
|
||||||
intel_pmu_disable_bts();
|
intel_pmu_disable_bts();
|
||||||
}
|
}
|
||||||
|
|
||||||
static void amd_pmu_disable_all(void)
|
static void x86_pmu_disable_all(void)
|
||||||
{
|
{
|
||||||
struct cpu_hw_events *cpuc = &__get_cpu_var(cpu_hw_events);
|
struct cpu_hw_events *cpuc = &__get_cpu_var(cpu_hw_events);
|
||||||
int idx;
|
int idx;
|
||||||
|
@ -1226,11 +1237,11 @@ static void amd_pmu_disable_all(void)
|
||||||
|
|
||||||
if (!test_bit(idx, cpuc->active_mask))
|
if (!test_bit(idx, cpuc->active_mask))
|
||||||
continue;
|
continue;
|
||||||
rdmsrl(MSR_K7_EVNTSEL0 + idx, val);
|
rdmsrl(x86_pmu.eventsel + idx, val);
|
||||||
if (!(val & ARCH_PERFMON_EVENTSEL0_ENABLE))
|
if (!(val & ARCH_PERFMON_EVENTSEL0_ENABLE))
|
||||||
continue;
|
continue;
|
||||||
val &= ~ARCH_PERFMON_EVENTSEL0_ENABLE;
|
val &= ~ARCH_PERFMON_EVENTSEL0_ENABLE;
|
||||||
wrmsrl(MSR_K7_EVNTSEL0 + idx, val);
|
wrmsrl(x86_pmu.eventsel + idx, val);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1278,7 +1289,7 @@ static void intel_pmu_enable_all(void)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static void amd_pmu_enable_all(void)
|
static void x86_pmu_enable_all(void)
|
||||||
{
|
{
|
||||||
struct cpu_hw_events *cpuc = &__get_cpu_var(cpu_hw_events);
|
struct cpu_hw_events *cpuc = &__get_cpu_var(cpu_hw_events);
|
||||||
int idx;
|
int idx;
|
||||||
|
@ -1292,7 +1303,7 @@ static void amd_pmu_enable_all(void)
|
||||||
|
|
||||||
val = event->hw.config;
|
val = event->hw.config;
|
||||||
val |= ARCH_PERFMON_EVENTSEL0_ENABLE;
|
val |= ARCH_PERFMON_EVENTSEL0_ENABLE;
|
||||||
wrmsrl(MSR_K7_EVNTSEL0 + idx, val);
|
wrmsrl(x86_pmu.eventsel + idx, val);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1546,7 +1557,7 @@ static inline void intel_pmu_ack_status(u64 ack)
|
||||||
wrmsrl(MSR_CORE_PERF_GLOBAL_OVF_CTRL, ack);
|
wrmsrl(MSR_CORE_PERF_GLOBAL_OVF_CTRL, ack);
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline void x86_pmu_enable_event(struct hw_perf_event *hwc, int idx)
|
static inline void __x86_pmu_enable_event(struct hw_perf_event *hwc, int idx)
|
||||||
{
|
{
|
||||||
(void)checking_wrmsrl(hwc->config_base + idx,
|
(void)checking_wrmsrl(hwc->config_base + idx,
|
||||||
hwc->config | ARCH_PERFMON_EVENTSEL0_ENABLE);
|
hwc->config | ARCH_PERFMON_EVENTSEL0_ENABLE);
|
||||||
|
@ -1598,12 +1609,6 @@ intel_pmu_disable_event(struct hw_perf_event *hwc, int idx)
|
||||||
x86_pmu_disable_event(hwc, idx);
|
x86_pmu_disable_event(hwc, idx);
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline void
|
|
||||||
amd_pmu_disable_event(struct hw_perf_event *hwc, int idx)
|
|
||||||
{
|
|
||||||
x86_pmu_disable_event(hwc, idx);
|
|
||||||
}
|
|
||||||
|
|
||||||
static DEFINE_PER_CPU(u64 [X86_PMC_IDX_MAX], pmc_prev_left);
|
static DEFINE_PER_CPU(u64 [X86_PMC_IDX_MAX], pmc_prev_left);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -1723,15 +1728,14 @@ static void intel_pmu_enable_event(struct hw_perf_event *hwc, int idx)
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
x86_pmu_enable_event(hwc, idx);
|
__x86_pmu_enable_event(hwc, idx);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void amd_pmu_enable_event(struct hw_perf_event *hwc, int idx)
|
static void x86_pmu_enable_event(struct hw_perf_event *hwc, int idx)
|
||||||
{
|
{
|
||||||
struct cpu_hw_events *cpuc = &__get_cpu_var(cpu_hw_events);
|
struct cpu_hw_events *cpuc = &__get_cpu_var(cpu_hw_events);
|
||||||
|
|
||||||
if (cpuc->enabled)
|
if (cpuc->enabled)
|
||||||
x86_pmu_enable_event(hwc, idx);
|
__x86_pmu_enable_event(hwc, idx);
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -1988,50 +1992,6 @@ static void intel_pmu_reset(void)
|
||||||
local_irq_restore(flags);
|
local_irq_restore(flags);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int p6_pmu_handle_irq(struct pt_regs *regs)
|
|
||||||
{
|
|
||||||
struct perf_sample_data data;
|
|
||||||
struct cpu_hw_events *cpuc;
|
|
||||||
struct perf_event *event;
|
|
||||||
struct hw_perf_event *hwc;
|
|
||||||
int idx, handled = 0;
|
|
||||||
u64 val;
|
|
||||||
|
|
||||||
data.addr = 0;
|
|
||||||
data.raw = NULL;
|
|
||||||
|
|
||||||
cpuc = &__get_cpu_var(cpu_hw_events);
|
|
||||||
|
|
||||||
for (idx = 0; idx < x86_pmu.num_events; idx++) {
|
|
||||||
if (!test_bit(idx, cpuc->active_mask))
|
|
||||||
continue;
|
|
||||||
|
|
||||||
event = cpuc->events[idx];
|
|
||||||
hwc = &event->hw;
|
|
||||||
|
|
||||||
val = x86_perf_event_update(event, hwc, idx);
|
|
||||||
if (val & (1ULL << (x86_pmu.event_bits - 1)))
|
|
||||||
continue;
|
|
||||||
|
|
||||||
/*
|
|
||||||
* event overflow
|
|
||||||
*/
|
|
||||||
handled = 1;
|
|
||||||
data.period = event->hw.last_period;
|
|
||||||
|
|
||||||
if (!x86_perf_event_set_period(event, hwc, idx))
|
|
||||||
continue;
|
|
||||||
|
|
||||||
if (perf_event_overflow(event, 1, &data, regs))
|
|
||||||
p6_pmu_disable_event(hwc, idx);
|
|
||||||
}
|
|
||||||
|
|
||||||
if (handled)
|
|
||||||
inc_irq_stat(apic_perf_irqs);
|
|
||||||
|
|
||||||
return handled;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* This handler is triggered by the local APIC, so the APIC IRQ handling
|
* This handler is triggered by the local APIC, so the APIC IRQ handling
|
||||||
* rules apply:
|
* rules apply:
|
||||||
|
@ -2098,7 +2058,7 @@ static int intel_pmu_handle_irq(struct pt_regs *regs)
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int amd_pmu_handle_irq(struct pt_regs *regs)
|
static int x86_pmu_handle_irq(struct pt_regs *regs)
|
||||||
{
|
{
|
||||||
struct perf_sample_data data;
|
struct perf_sample_data data;
|
||||||
struct cpu_hw_events *cpuc;
|
struct cpu_hw_events *cpuc;
|
||||||
|
@ -2133,7 +2093,7 @@ static int amd_pmu_handle_irq(struct pt_regs *regs)
|
||||||
continue;
|
continue;
|
||||||
|
|
||||||
if (perf_event_overflow(event, 1, &data, regs))
|
if (perf_event_overflow(event, 1, &data, regs))
|
||||||
amd_pmu_disable_event(hwc, idx);
|
x86_pmu.disable(hwc, idx);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (handled)
|
if (handled)
|
||||||
|
@ -2374,7 +2334,7 @@ static __read_mostly struct notifier_block perf_event_nmi_notifier = {
|
||||||
|
|
||||||
static __initconst struct x86_pmu p6_pmu = {
|
static __initconst struct x86_pmu p6_pmu = {
|
||||||
.name = "p6",
|
.name = "p6",
|
||||||
.handle_irq = p6_pmu_handle_irq,
|
.handle_irq = x86_pmu_handle_irq,
|
||||||
.disable_all = p6_pmu_disable_all,
|
.disable_all = p6_pmu_disable_all,
|
||||||
.enable_all = p6_pmu_enable_all,
|
.enable_all = p6_pmu_enable_all,
|
||||||
.enable = p6_pmu_enable_event,
|
.enable = p6_pmu_enable_event,
|
||||||
|
@ -2401,6 +2361,29 @@ static __initconst struct x86_pmu p6_pmu = {
|
||||||
.event_constraints = intel_p6_event_constraints
|
.event_constraints = intel_p6_event_constraints
|
||||||
};
|
};
|
||||||
|
|
||||||
|
static __initconst struct x86_pmu core_pmu = {
|
||||||
|
.name = "core",
|
||||||
|
.handle_irq = x86_pmu_handle_irq,
|
||||||
|
.disable_all = x86_pmu_disable_all,
|
||||||
|
.enable_all = x86_pmu_enable_all,
|
||||||
|
.enable = x86_pmu_enable_event,
|
||||||
|
.disable = x86_pmu_disable_event,
|
||||||
|
.eventsel = MSR_ARCH_PERFMON_EVENTSEL0,
|
||||||
|
.perfctr = MSR_ARCH_PERFMON_PERFCTR0,
|
||||||
|
.event_map = intel_pmu_event_map,
|
||||||
|
.raw_event = intel_pmu_raw_event,
|
||||||
|
.max_events = ARRAY_SIZE(intel_perfmon_event_map),
|
||||||
|
.apic = 1,
|
||||||
|
/*
|
||||||
|
* Intel PMCs cannot be accessed sanely above 32 bit width,
|
||||||
|
* so we install an artificial 1<<31 period regardless of
|
||||||
|
* the generic event period:
|
||||||
|
*/
|
||||||
|
.max_period = (1ULL << 31) - 1,
|
||||||
|
.get_event_constraints = intel_get_event_constraints,
|
||||||
|
.event_constraints = intel_core_event_constraints,
|
||||||
|
};
|
||||||
|
|
||||||
static __initconst struct x86_pmu intel_pmu = {
|
static __initconst struct x86_pmu intel_pmu = {
|
||||||
.name = "Intel",
|
.name = "Intel",
|
||||||
.handle_irq = intel_pmu_handle_irq,
|
.handle_irq = intel_pmu_handle_irq,
|
||||||
|
@ -2427,11 +2410,11 @@ static __initconst struct x86_pmu intel_pmu = {
|
||||||
|
|
||||||
static __initconst struct x86_pmu amd_pmu = {
|
static __initconst struct x86_pmu amd_pmu = {
|
||||||
.name = "AMD",
|
.name = "AMD",
|
||||||
.handle_irq = amd_pmu_handle_irq,
|
.handle_irq = x86_pmu_handle_irq,
|
||||||
.disable_all = amd_pmu_disable_all,
|
.disable_all = x86_pmu_disable_all,
|
||||||
.enable_all = amd_pmu_enable_all,
|
.enable_all = x86_pmu_enable_all,
|
||||||
.enable = amd_pmu_enable_event,
|
.enable = x86_pmu_enable_event,
|
||||||
.disable = amd_pmu_disable_event,
|
.disable = x86_pmu_disable_event,
|
||||||
.eventsel = MSR_K7_EVNTSEL0,
|
.eventsel = MSR_K7_EVNTSEL0,
|
||||||
.perfctr = MSR_K7_PERFCTR0,
|
.perfctr = MSR_K7_PERFCTR0,
|
||||||
.event_map = amd_pmu_event_map,
|
.event_map = amd_pmu_event_map,
|
||||||
|
@ -2498,9 +2481,10 @@ static __init int intel_pmu_init(void)
|
||||||
|
|
||||||
version = eax.split.version_id;
|
version = eax.split.version_id;
|
||||||
if (version < 2)
|
if (version < 2)
|
||||||
return -ENODEV;
|
x86_pmu = core_pmu;
|
||||||
|
else
|
||||||
|
x86_pmu = intel_pmu;
|
||||||
|
|
||||||
x86_pmu = intel_pmu;
|
|
||||||
x86_pmu.version = version;
|
x86_pmu.version = version;
|
||||||
x86_pmu.num_events = eax.split.num_events;
|
x86_pmu.num_events = eax.split.num_events;
|
||||||
x86_pmu.event_bits = eax.split.bit_width;
|
x86_pmu.event_bits = eax.split.bit_width;
|
||||||
|
@ -2510,12 +2494,17 @@ static __init int intel_pmu_init(void)
|
||||||
* Quirk: v2 perfmon does not report fixed-purpose events, so
|
* Quirk: v2 perfmon does not report fixed-purpose events, so
|
||||||
* assume at least 3 events:
|
* assume at least 3 events:
|
||||||
*/
|
*/
|
||||||
x86_pmu.num_events_fixed = max((int)edx.split.num_events_fixed, 3);
|
if (version > 1)
|
||||||
|
x86_pmu.num_events_fixed = max((int)edx.split.num_events_fixed, 3);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Install the hw-cache-events table:
|
* Install the hw-cache-events table:
|
||||||
*/
|
*/
|
||||||
switch (boot_cpu_data.x86_model) {
|
switch (boot_cpu_data.x86_model) {
|
||||||
|
case 14: /* 65 nm core solo/duo, "Yonah" */
|
||||||
|
pr_cont("Core events, ");
|
||||||
|
break;
|
||||||
|
|
||||||
case 15: /* original 65 nm celeron/pentium/core2/xeon, "Merom"/"Conroe" */
|
case 15: /* original 65 nm celeron/pentium/core2/xeon, "Merom"/"Conroe" */
|
||||||
case 22: /* single-core 65 nm celeron/core2solo "Merom-L"/"Conroe-L" */
|
case 22: /* single-core 65 nm celeron/core2solo "Merom-L"/"Conroe-L" */
|
||||||
case 23: /* current 45 nm celeron/core2/xeon "Penryn"/"Wolfdale" */
|
case 23: /* current 45 nm celeron/core2/xeon "Penryn"/"Wolfdale" */
|
||||||
|
@ -2523,7 +2512,7 @@ static __init int intel_pmu_init(void)
|
||||||
memcpy(hw_cache_event_ids, core2_hw_cache_event_ids,
|
memcpy(hw_cache_event_ids, core2_hw_cache_event_ids,
|
||||||
sizeof(hw_cache_event_ids));
|
sizeof(hw_cache_event_ids));
|
||||||
|
|
||||||
x86_pmu.event_constraints = intel_core_event_constraints;
|
x86_pmu.event_constraints = intel_core2_event_constraints;
|
||||||
pr_cont("Core2 events, ");
|
pr_cont("Core2 events, ");
|
||||||
break;
|
break;
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue