| /* |
| * Xtensa Performance Monitor Module driver |
| * See Tensilica Debug User's Guide for PMU registers documentation. |
| * |
| * Copyright (C) 2015 Cadence Design Systems Inc. |
| * |
| * This program is free software; you can redistribute it and/or modify |
| * it under the terms of the GNU General Public License version 2 as |
| * published by the Free Software Foundation. |
| */ |
| |
| #include <linux/interrupt.h> |
| #include <linux/irqdomain.h> |
| #include <linux/module.h> |
| #include <linux/of.h> |
| #include <linux/perf_event.h> |
| #include <linux/platform_device.h> |
| |
| #include <asm/processor.h> |
| #include <asm/stacktrace.h> |
| |
| /* Global control/status for all perf counters */ |
| #define XTENSA_PMU_PMG 0x1000 |
| /* Perf counter values */ |
| #define XTENSA_PMU_PM(i) (0x1080 + (i) * 4) |
| /* Perf counter control registers */ |
| #define XTENSA_PMU_PMCTRL(i) (0x1100 + (i) * 4) |
| /* Perf counter status registers */ |
| #define XTENSA_PMU_PMSTAT(i) (0x1180 + (i) * 4) |
| |
| #define XTENSA_PMU_PMG_PMEN 0x1 |
| |
| #define XTENSA_PMU_COUNTER_MASK 0xffffffffULL |
| #define XTENSA_PMU_COUNTER_MAX 0x7fffffff |
| |
| #define XTENSA_PMU_PMCTRL_INTEN 0x00000001 |
| #define XTENSA_PMU_PMCTRL_KRNLCNT 0x00000008 |
| #define XTENSA_PMU_PMCTRL_TRACELEVEL 0x000000f0 |
| #define XTENSA_PMU_PMCTRL_SELECT_SHIFT 8 |
| #define XTENSA_PMU_PMCTRL_SELECT 0x00001f00 |
| #define XTENSA_PMU_PMCTRL_MASK_SHIFT 16 |
| #define XTENSA_PMU_PMCTRL_MASK 0xffff0000 |
| |
| #define XTENSA_PMU_MASK(select, mask) \ |
| (((select) << XTENSA_PMU_PMCTRL_SELECT_SHIFT) | \ |
| ((mask) << XTENSA_PMU_PMCTRL_MASK_SHIFT) | \ |
| XTENSA_PMU_PMCTRL_TRACELEVEL | \ |
| XTENSA_PMU_PMCTRL_INTEN) |
| |
| #define XTENSA_PMU_PMSTAT_OVFL 0x00000001 |
| #define XTENSA_PMU_PMSTAT_INTASRT 0x00000010 |
| |
| struct xtensa_pmu_events { |
| /* Array of events currently on this core */ |
| struct perf_event *event[XCHAL_NUM_PERF_COUNTERS]; |
| /* Bitmap of used hardware counters */ |
| unsigned long used_mask[BITS_TO_LONGS(XCHAL_NUM_PERF_COUNTERS)]; |
| }; |
| static DEFINE_PER_CPU(struct xtensa_pmu_events, xtensa_pmu_events); |
| |
| static const u32 xtensa_hw_ctl[] = { |
| [PERF_COUNT_HW_CPU_CYCLES] = XTENSA_PMU_MASK(0, 0x1), |
| [PERF_COUNT_HW_INSTRUCTIONS] = XTENSA_PMU_MASK(2, 0xffff), |
| [PERF_COUNT_HW_CACHE_REFERENCES] = XTENSA_PMU_MASK(10, 0x1), |
| [PERF_COUNT_HW_CACHE_MISSES] = XTENSA_PMU_MASK(12, 0x1), |
| /* Taken and non-taken branches + taken loop ends */ |
| [PERF_COUNT_HW_BRANCH_INSTRUCTIONS] = XTENSA_PMU_MASK(2, 0x490), |
| /* Instruction-related + other global stall cycles */ |
| [PERF_COUNT_HW_STALLED_CYCLES_FRONTEND] = XTENSA_PMU_MASK(4, 0x1ff), |
| /* Data-related global stall cycles */ |
| [PERF_COUNT_HW_STALLED_CYCLES_BACKEND] = XTENSA_PMU_MASK(3, 0x1ff), |
| }; |
| |
| #define C(_x) PERF_COUNT_HW_CACHE_##_x |
| |
| static const u32 xtensa_cache_ctl[][C(OP_MAX)][C(RESULT_MAX)] = { |
| [C(L1D)] = { |
| [C(OP_READ)] = { |
| [C(RESULT_ACCESS)] = XTENSA_PMU_MASK(10, 0x1), |
| [C(RESULT_MISS)] = XTENSA_PMU_MASK(10, 0x2), |
| }, |
| [C(OP_WRITE)] = { |
| [C(RESULT_ACCESS)] = XTENSA_PMU_MASK(11, 0x1), |
| [C(RESULT_MISS)] = XTENSA_PMU_MASK(11, 0x2), |
| }, |
| }, |
| [C(L1I)] = { |
| [C(OP_READ)] = { |
| [C(RESULT_ACCESS)] = XTENSA_PMU_MASK(8, 0x1), |
| [C(RESULT_MISS)] = XTENSA_PMU_MASK(8, 0x2), |
| }, |
| }, |
| [C(DTLB)] = { |
| [C(OP_READ)] = { |
| [C(RESULT_ACCESS)] = XTENSA_PMU_MASK(9, 0x1), |
| [C(RESULT_MISS)] = XTENSA_PMU_MASK(9, 0x8), |
| }, |
| }, |
| [C(ITLB)] = { |
| [C(OP_READ)] = { |
| [C(RESULT_ACCESS)] = XTENSA_PMU_MASK(7, 0x1), |
| [C(RESULT_MISS)] = XTENSA_PMU_MASK(7, 0x8), |
| }, |
| }, |
| }; |
| |
| static int xtensa_pmu_cache_event(u64 config) |
| { |
| unsigned int cache_type, cache_op, cache_result; |
| int ret; |
| |
| cache_type = (config >> 0) & 0xff; |
| cache_op = (config >> 8) & 0xff; |
| cache_result = (config >> 16) & 0xff; |
| |
| if (cache_type >= ARRAY_SIZE(xtensa_cache_ctl) || |
| cache_op >= C(OP_MAX) || |
| cache_result >= C(RESULT_MAX)) |
| return -EINVAL; |
| |
| ret = xtensa_cache_ctl[cache_type][cache_op][cache_result]; |
| |
| if (ret == 0) |
| return -EINVAL; |
| |
| return ret; |
| } |
| |
| static inline uint32_t xtensa_pmu_read_counter(int idx) |
| { |
| return get_er(XTENSA_PMU_PM(idx)); |
| } |
| |
| static inline void xtensa_pmu_write_counter(int idx, uint32_t v) |
| { |
| set_er(v, XTENSA_PMU_PM(idx)); |
| } |
| |
| static void xtensa_perf_event_update(struct perf_event *event, |
| struct hw_perf_event *hwc, int idx) |
| { |
| uint64_t prev_raw_count, new_raw_count; |
| int64_t delta; |
| |
| do { |
| prev_raw_count = local64_read(&hwc->prev_count); |
| new_raw_count = xtensa_pmu_read_counter(event->hw.idx); |
| } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, |
| new_raw_count) != prev_raw_count); |
| |
| delta = (new_raw_count - prev_raw_count) & XTENSA_PMU_COUNTER_MASK; |
| |
| local64_add(delta, &event->count); |
| local64_sub(delta, &hwc->period_left); |
| } |
| |
| static bool xtensa_perf_event_set_period(struct perf_event *event, |
| struct hw_perf_event *hwc, int idx) |
| { |
| bool rc = false; |
| s64 left; |
| |
| if (!is_sampling_event(event)) { |
| left = XTENSA_PMU_COUNTER_MAX; |
| } else { |
| s64 period = hwc->sample_period; |
| |
| left = local64_read(&hwc->period_left); |
| if (left <= -period) { |
| left = period; |
| local64_set(&hwc->period_left, left); |
| hwc->last_period = period; |
| rc = true; |
| } else if (left <= 0) { |
| left += period; |
| local64_set(&hwc->period_left, left); |
| hwc->last_period = period; |
| rc = true; |
| } |
| if (left > XTENSA_PMU_COUNTER_MAX) |
| left = XTENSA_PMU_COUNTER_MAX; |
| } |
| |
| local64_set(&hwc->prev_count, -left); |
| xtensa_pmu_write_counter(idx, -left); |
| perf_event_update_userpage(event); |
| |
| return rc; |
| } |
| |
| static void xtensa_pmu_enable(struct pmu *pmu) |
| { |
| set_er(get_er(XTENSA_PMU_PMG) | XTENSA_PMU_PMG_PMEN, XTENSA_PMU_PMG); |
| } |
| |
| static void xtensa_pmu_disable(struct pmu *pmu) |
| { |
| set_er(get_er(XTENSA_PMU_PMG) & ~XTENSA_PMU_PMG_PMEN, XTENSA_PMU_PMG); |
| } |
| |
| static int xtensa_pmu_event_init(struct perf_event *event) |
| { |
| int ret; |
| |
| switch (event->attr.type) { |
| case PERF_TYPE_HARDWARE: |
| if (event->attr.config >= ARRAY_SIZE(xtensa_hw_ctl) || |
| xtensa_hw_ctl[event->attr.config] == 0) |
| return -EINVAL; |
| event->hw.config = xtensa_hw_ctl[event->attr.config]; |
| return 0; |
| |
| case PERF_TYPE_HW_CACHE: |
| ret = xtensa_pmu_cache_event(event->attr.config); |
| if (ret < 0) |
| return ret; |
| event->hw.config = ret; |
| return 0; |
| |
| case PERF_TYPE_RAW: |
| /* Not 'previous counter' select */ |
| if ((event->attr.config & XTENSA_PMU_PMCTRL_SELECT) == |
| (1 << XTENSA_PMU_PMCTRL_SELECT_SHIFT)) |
| return -EINVAL; |
| event->hw.config = (event->attr.config & |
| (XTENSA_PMU_PMCTRL_KRNLCNT | |
| XTENSA_PMU_PMCTRL_TRACELEVEL | |
| XTENSA_PMU_PMCTRL_SELECT | |
| XTENSA_PMU_PMCTRL_MASK)) | |
| XTENSA_PMU_PMCTRL_INTEN; |
| return 0; |
| |
| default: |
| return -ENOENT; |
| } |
| } |
| |
| /* |
| * Starts/Stops a counter present on the PMU. The PMI handler |
| * should stop the counter when perf_event_overflow() returns |
| * !0. ->start() will be used to continue. |
| */ |
| static void xtensa_pmu_start(struct perf_event *event, int flags) |
| { |
| struct hw_perf_event *hwc = &event->hw; |
| int idx = hwc->idx; |
| |
| if (WARN_ON_ONCE(idx == -1)) |
| return; |
| |
| if (flags & PERF_EF_RELOAD) { |
| WARN_ON_ONCE(!(event->hw.state & PERF_HES_UPTODATE)); |
| xtensa_perf_event_set_period(event, hwc, idx); |
| } |
| |
| hwc->state = 0; |
| |
| set_er(hwc->config, XTENSA_PMU_PMCTRL(idx)); |
| } |
| |
| static void xtensa_pmu_stop(struct perf_event *event, int flags) |
| { |
| struct hw_perf_event *hwc = &event->hw; |
| int idx = hwc->idx; |
| |
| if (!(hwc->state & PERF_HES_STOPPED)) { |
| set_er(0, XTENSA_PMU_PMCTRL(idx)); |
| set_er(get_er(XTENSA_PMU_PMSTAT(idx)), |
| XTENSA_PMU_PMSTAT(idx)); |
| hwc->state |= PERF_HES_STOPPED; |
| } |
| |
| if ((flags & PERF_EF_UPDATE) && |
| !(event->hw.state & PERF_HES_UPTODATE)) { |
| xtensa_perf_event_update(event, &event->hw, idx); |
| event->hw.state |= PERF_HES_UPTODATE; |
| } |
| } |
| |
| /* |
| * Adds/Removes a counter to/from the PMU, can be done inside |
| * a transaction, see the ->*_txn() methods. |
| */ |
| static int xtensa_pmu_add(struct perf_event *event, int flags) |
| { |
| struct xtensa_pmu_events *ev = this_cpu_ptr(&xtensa_pmu_events); |
| struct hw_perf_event *hwc = &event->hw; |
| int idx = hwc->idx; |
| |
| if (__test_and_set_bit(idx, ev->used_mask)) { |
| idx = find_first_zero_bit(ev->used_mask, |
| XCHAL_NUM_PERF_COUNTERS); |
| if (idx == XCHAL_NUM_PERF_COUNTERS) |
| return -EAGAIN; |
| |
| __set_bit(idx, ev->used_mask); |
| hwc->idx = idx; |
| } |
| ev->event[idx] = event; |
| |
| hwc->state = PERF_HES_UPTODATE | PERF_HES_STOPPED; |
| |
| if (flags & PERF_EF_START) |
| xtensa_pmu_start(event, PERF_EF_RELOAD); |
| |
| perf_event_update_userpage(event); |
| return 0; |
| } |
| |
| static void xtensa_pmu_del(struct perf_event *event, int flags) |
| { |
| struct xtensa_pmu_events *ev = this_cpu_ptr(&xtensa_pmu_events); |
| |
| xtensa_pmu_stop(event, PERF_EF_UPDATE); |
| __clear_bit(event->hw.idx, ev->used_mask); |
| perf_event_update_userpage(event); |
| } |
| |
| static void xtensa_pmu_read(struct perf_event *event) |
| { |
| xtensa_perf_event_update(event, &event->hw, event->hw.idx); |
| } |
| |
| static int callchain_trace(struct stackframe *frame, void *data) |
| { |
| struct perf_callchain_entry_ctx *entry = data; |
| |
| perf_callchain_store(entry, frame->pc); |
| return 0; |
| } |
| |
| void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, |
| struct pt_regs *regs) |
| { |
| xtensa_backtrace_kernel(regs, entry->max_stack, |
| callchain_trace, NULL, entry); |
| } |
| |
| void perf_callchain_user(struct perf_callchain_entry_ctx *entry, |
| struct pt_regs *regs) |
| { |
| xtensa_backtrace_user(regs, entry->max_stack, |
| callchain_trace, entry); |
| } |
| |
| void perf_event_print_debug(void) |
| { |
| unsigned long flags; |
| unsigned i; |
| |
| local_irq_save(flags); |
| pr_info("CPU#%d: PMG: 0x%08lx\n", smp_processor_id(), |
| get_er(XTENSA_PMU_PMG)); |
| for (i = 0; i < XCHAL_NUM_PERF_COUNTERS; ++i) |
| pr_info("PM%d: 0x%08lx, PMCTRL%d: 0x%08lx, PMSTAT%d: 0x%08lx\n", |
| i, get_er(XTENSA_PMU_PM(i)), |
| i, get_er(XTENSA_PMU_PMCTRL(i)), |
| i, get_er(XTENSA_PMU_PMSTAT(i))); |
| local_irq_restore(flags); |
| } |
| |
| irqreturn_t xtensa_pmu_irq_handler(int irq, void *dev_id) |
| { |
| irqreturn_t rc = IRQ_NONE; |
| struct xtensa_pmu_events *ev = this_cpu_ptr(&xtensa_pmu_events); |
| unsigned i; |
| |
| for (i = find_first_bit(ev->used_mask, XCHAL_NUM_PERF_COUNTERS); |
| i < XCHAL_NUM_PERF_COUNTERS; |
| i = find_next_bit(ev->used_mask, XCHAL_NUM_PERF_COUNTERS, i + 1)) { |
| uint32_t v = get_er(XTENSA_PMU_PMSTAT(i)); |
| struct perf_event *event = ev->event[i]; |
| struct hw_perf_event *hwc = &event->hw; |
| u64 last_period; |
| |
| if (!(v & XTENSA_PMU_PMSTAT_OVFL)) |
| continue; |
| |
| set_er(v, XTENSA_PMU_PMSTAT(i)); |
| xtensa_perf_event_update(event, hwc, i); |
| last_period = hwc->last_period; |
| if (xtensa_perf_event_set_period(event, hwc, i)) { |
| struct perf_sample_data data; |
| struct pt_regs *regs = get_irq_regs(); |
| |
| perf_sample_data_init(&data, 0, last_period); |
| if (perf_event_overflow(event, &data, regs)) |
| xtensa_pmu_stop(event, 0); |
| } |
| |
| rc = IRQ_HANDLED; |
| } |
| return rc; |
| } |
| |
| static struct pmu xtensa_pmu = { |
| .pmu_enable = xtensa_pmu_enable, |
| .pmu_disable = xtensa_pmu_disable, |
| .event_init = xtensa_pmu_event_init, |
| .add = xtensa_pmu_add, |
| .del = xtensa_pmu_del, |
| .start = xtensa_pmu_start, |
| .stop = xtensa_pmu_stop, |
| .read = xtensa_pmu_read, |
| }; |
| |
| static int xtensa_pmu_setup(unsigned int cpu) |
| { |
| unsigned i; |
| |
| set_er(0, XTENSA_PMU_PMG); |
| for (i = 0; i < XCHAL_NUM_PERF_COUNTERS; ++i) { |
| set_er(0, XTENSA_PMU_PMCTRL(i)); |
| set_er(get_er(XTENSA_PMU_PMSTAT(i)), XTENSA_PMU_PMSTAT(i)); |
| } |
| return 0; |
| } |
| |
| static int __init xtensa_pmu_init(void) |
| { |
| int ret; |
| int irq = irq_create_mapping(NULL, XCHAL_PROFILING_INTERRUPT); |
| |
| ret = cpuhp_setup_state(CPUHP_AP_PERF_XTENSA_STARTING, |
| "perf/xtensa:starting", xtensa_pmu_setup, |
| NULL); |
| if (ret) { |
| pr_err("xtensa_pmu: failed to register CPU-hotplug.\n"); |
| return ret; |
| } |
| #if XTENSA_FAKE_NMI |
| enable_irq(irq); |
| #else |
| ret = request_irq(irq, xtensa_pmu_irq_handler, IRQF_PERCPU, |
| "pmu", NULL); |
| if (ret < 0) |
| return ret; |
| #endif |
| |
| ret = perf_pmu_register(&xtensa_pmu, "cpu", PERF_TYPE_RAW); |
| if (ret) |
| free_irq(irq, NULL); |
| |
| return ret; |
| } |
| early_initcall(xtensa_pmu_init); |