mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/next/linux-next.git
synced 2025-01-10 07:50:04 +00:00
x86, perf_counter, bts: Optimize BTS overflow handling
Draining the BTS buffer on a buffer overflow interrupt takes too long resulting in a kernel lockup when tracing the kernel. Restructure perf_counter sampling into sample creation and sample output. Prepare a single reference sample for BTS sampling and update the from and to address fields when draining the BTS buffer. Drain the entire BTS buffer between a single perf_output_begin() / perf_output_end() pair. Signed-off-by: Markus Metzger <markus.t.metzger@intel.com> Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl> LKML-Reference: <20090915130023.A16204@sedona.ch.intel.com> Signed-off-by: Ingo Molnar <mingo@elte.hu>
This commit is contained in:
parent
4b77a72977
commit
5622f295b5
@ -36,10 +36,10 @@ static u64 perf_counter_mask __read_mostly;
|
|||||||
#define BTS_RECORD_SIZE 24
|
#define BTS_RECORD_SIZE 24
|
||||||
|
|
||||||
/* The size of a per-cpu BTS buffer in bytes: */
|
/* The size of a per-cpu BTS buffer in bytes: */
|
||||||
#define BTS_BUFFER_SIZE (BTS_RECORD_SIZE * 1024)
|
#define BTS_BUFFER_SIZE (BTS_RECORD_SIZE * 2048)
|
||||||
|
|
||||||
/* The BTS overflow threshold in bytes from the end of the buffer: */
|
/* The BTS overflow threshold in bytes from the end of the buffer: */
|
||||||
#define BTS_OVFL_TH (BTS_RECORD_SIZE * 64)
|
#define BTS_OVFL_TH (BTS_RECORD_SIZE * 128)
|
||||||
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@ -1488,8 +1488,7 @@ void perf_counter_print_debug(void)
|
|||||||
local_irq_restore(flags);
|
local_irq_restore(flags);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void intel_pmu_drain_bts_buffer(struct cpu_hw_counters *cpuc,
|
static void intel_pmu_drain_bts_buffer(struct cpu_hw_counters *cpuc)
|
||||||
struct perf_sample_data *data)
|
|
||||||
{
|
{
|
||||||
struct debug_store *ds = cpuc->ds;
|
struct debug_store *ds = cpuc->ds;
|
||||||
struct bts_record {
|
struct bts_record {
|
||||||
@ -1498,8 +1497,11 @@ static void intel_pmu_drain_bts_buffer(struct cpu_hw_counters *cpuc,
|
|||||||
u64 flags;
|
u64 flags;
|
||||||
};
|
};
|
||||||
struct perf_counter *counter = cpuc->counters[X86_PMC_IDX_FIXED_BTS];
|
struct perf_counter *counter = cpuc->counters[X86_PMC_IDX_FIXED_BTS];
|
||||||
unsigned long orig_ip = data->regs->ip;
|
|
||||||
struct bts_record *at, *top;
|
struct bts_record *at, *top;
|
||||||
|
struct perf_output_handle handle;
|
||||||
|
struct perf_event_header header;
|
||||||
|
struct perf_sample_data data;
|
||||||
|
struct pt_regs regs;
|
||||||
|
|
||||||
if (!counter)
|
if (!counter)
|
||||||
return;
|
return;
|
||||||
@ -1510,19 +1512,38 @@ static void intel_pmu_drain_bts_buffer(struct cpu_hw_counters *cpuc,
|
|||||||
at = (struct bts_record *)(unsigned long)ds->bts_buffer_base;
|
at = (struct bts_record *)(unsigned long)ds->bts_buffer_base;
|
||||||
top = (struct bts_record *)(unsigned long)ds->bts_index;
|
top = (struct bts_record *)(unsigned long)ds->bts_index;
|
||||||
|
|
||||||
|
if (top <= at)
|
||||||
|
return;
|
||||||
|
|
||||||
ds->bts_index = ds->bts_buffer_base;
|
ds->bts_index = ds->bts_buffer_base;
|
||||||
|
|
||||||
for (; at < top; at++) {
|
|
||||||
data->regs->ip = at->from;
|
|
||||||
data->addr = at->to;
|
|
||||||
|
|
||||||
perf_counter_output(counter, 1, data);
|
data.period = counter->hw.last_period;
|
||||||
|
data.addr = 0;
|
||||||
|
regs.ip = 0;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Prepare a generic sample, i.e. fill in the invariant fields.
|
||||||
|
* We will overwrite the from and to address before we output
|
||||||
|
* the sample.
|
||||||
|
*/
|
||||||
|
perf_prepare_sample(&header, &data, counter, ®s);
|
||||||
|
|
||||||
|
if (perf_output_begin(&handle, counter,
|
||||||
|
header.size * (top - at), 1, 1))
|
||||||
|
return;
|
||||||
|
|
||||||
|
for (; at < top; at++) {
|
||||||
|
data.ip = at->from;
|
||||||
|
data.addr = at->to;
|
||||||
|
|
||||||
|
perf_output_sample(&handle, &header, &data, counter);
|
||||||
}
|
}
|
||||||
|
|
||||||
data->regs->ip = orig_ip;
|
perf_output_end(&handle);
|
||||||
data->addr = 0;
|
|
||||||
|
|
||||||
/* There's new data available. */
|
/* There's new data available. */
|
||||||
|
counter->hw.interrupts++;
|
||||||
counter->pending_kill = POLL_IN;
|
counter->pending_kill = POLL_IN;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1552,13 +1573,9 @@ static void x86_pmu_disable(struct perf_counter *counter)
|
|||||||
x86_perf_counter_update(counter, hwc, idx);
|
x86_perf_counter_update(counter, hwc, idx);
|
||||||
|
|
||||||
/* Drain the remaining BTS records. */
|
/* Drain the remaining BTS records. */
|
||||||
if (unlikely(idx == X86_PMC_IDX_FIXED_BTS)) {
|
if (unlikely(idx == X86_PMC_IDX_FIXED_BTS))
|
||||||
struct perf_sample_data data;
|
intel_pmu_drain_bts_buffer(cpuc);
|
||||||
struct pt_regs regs;
|
|
||||||
|
|
||||||
data.regs = ®s;
|
|
||||||
intel_pmu_drain_bts_buffer(cpuc, &data);
|
|
||||||
}
|
|
||||||
cpuc->counters[idx] = NULL;
|
cpuc->counters[idx] = NULL;
|
||||||
clear_bit(idx, cpuc->used_mask);
|
clear_bit(idx, cpuc->used_mask);
|
||||||
|
|
||||||
@ -1619,7 +1636,6 @@ static int p6_pmu_handle_irq(struct pt_regs *regs)
|
|||||||
int idx, handled = 0;
|
int idx, handled = 0;
|
||||||
u64 val;
|
u64 val;
|
||||||
|
|
||||||
data.regs = regs;
|
|
||||||
data.addr = 0;
|
data.addr = 0;
|
||||||
|
|
||||||
cpuc = &__get_cpu_var(cpu_hw_counters);
|
cpuc = &__get_cpu_var(cpu_hw_counters);
|
||||||
@ -1644,7 +1660,7 @@ static int p6_pmu_handle_irq(struct pt_regs *regs)
|
|||||||
if (!x86_perf_counter_set_period(counter, hwc, idx))
|
if (!x86_perf_counter_set_period(counter, hwc, idx))
|
||||||
continue;
|
continue;
|
||||||
|
|
||||||
if (perf_counter_overflow(counter, 1, &data))
|
if (perf_counter_overflow(counter, 1, &data, regs))
|
||||||
p6_pmu_disable_counter(hwc, idx);
|
p6_pmu_disable_counter(hwc, idx);
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1665,13 +1681,12 @@ static int intel_pmu_handle_irq(struct pt_regs *regs)
|
|||||||
int bit, loops;
|
int bit, loops;
|
||||||
u64 ack, status;
|
u64 ack, status;
|
||||||
|
|
||||||
data.regs = regs;
|
|
||||||
data.addr = 0;
|
data.addr = 0;
|
||||||
|
|
||||||
cpuc = &__get_cpu_var(cpu_hw_counters);
|
cpuc = &__get_cpu_var(cpu_hw_counters);
|
||||||
|
|
||||||
perf_disable();
|
perf_disable();
|
||||||
intel_pmu_drain_bts_buffer(cpuc, &data);
|
intel_pmu_drain_bts_buffer(cpuc);
|
||||||
status = intel_pmu_get_status();
|
status = intel_pmu_get_status();
|
||||||
if (!status) {
|
if (!status) {
|
||||||
perf_enable();
|
perf_enable();
|
||||||
@ -1702,7 +1717,7 @@ again:
|
|||||||
|
|
||||||
data.period = counter->hw.last_period;
|
data.period = counter->hw.last_period;
|
||||||
|
|
||||||
if (perf_counter_overflow(counter, 1, &data))
|
if (perf_counter_overflow(counter, 1, &data, regs))
|
||||||
intel_pmu_disable_counter(&counter->hw, bit);
|
intel_pmu_disable_counter(&counter->hw, bit);
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1729,7 +1744,6 @@ static int amd_pmu_handle_irq(struct pt_regs *regs)
|
|||||||
int idx, handled = 0;
|
int idx, handled = 0;
|
||||||
u64 val;
|
u64 val;
|
||||||
|
|
||||||
data.regs = regs;
|
|
||||||
data.addr = 0;
|
data.addr = 0;
|
||||||
|
|
||||||
cpuc = &__get_cpu_var(cpu_hw_counters);
|
cpuc = &__get_cpu_var(cpu_hw_counters);
|
||||||
@ -1754,7 +1768,7 @@ static int amd_pmu_handle_irq(struct pt_regs *regs)
|
|||||||
if (!x86_perf_counter_set_period(counter, hwc, idx))
|
if (!x86_perf_counter_set_period(counter, hwc, idx))
|
||||||
continue;
|
continue;
|
||||||
|
|
||||||
if (perf_counter_overflow(counter, 1, &data))
|
if (perf_counter_overflow(counter, 1, &data, regs))
|
||||||
amd_pmu_disable_counter(hwc, idx);
|
amd_pmu_disable_counter(hwc, idx);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -691,6 +691,17 @@ struct perf_cpu_context {
|
|||||||
int recursion[4];
|
int recursion[4];
|
||||||
};
|
};
|
||||||
|
|
||||||
|
struct perf_output_handle {
|
||||||
|
struct perf_counter *counter;
|
||||||
|
struct perf_mmap_data *data;
|
||||||
|
unsigned long head;
|
||||||
|
unsigned long offset;
|
||||||
|
int nmi;
|
||||||
|
int sample;
|
||||||
|
int locked;
|
||||||
|
unsigned long flags;
|
||||||
|
};
|
||||||
|
|
||||||
#ifdef CONFIG_PERF_COUNTERS
|
#ifdef CONFIG_PERF_COUNTERS
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@ -722,16 +733,38 @@ extern int hw_perf_group_sched_in(struct perf_counter *group_leader,
|
|||||||
extern void perf_counter_update_userpage(struct perf_counter *counter);
|
extern void perf_counter_update_userpage(struct perf_counter *counter);
|
||||||
|
|
||||||
struct perf_sample_data {
|
struct perf_sample_data {
|
||||||
struct pt_regs *regs;
|
u64 type;
|
||||||
|
|
||||||
|
u64 ip;
|
||||||
|
struct {
|
||||||
|
u32 pid;
|
||||||
|
u32 tid;
|
||||||
|
} tid_entry;
|
||||||
|
u64 time;
|
||||||
u64 addr;
|
u64 addr;
|
||||||
|
u64 id;
|
||||||
|
u64 stream_id;
|
||||||
|
struct {
|
||||||
|
u32 cpu;
|
||||||
|
u32 reserved;
|
||||||
|
} cpu_entry;
|
||||||
u64 period;
|
u64 period;
|
||||||
|
struct perf_callchain_entry *callchain;
|
||||||
struct perf_raw_record *raw;
|
struct perf_raw_record *raw;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
extern void perf_output_sample(struct perf_output_handle *handle,
|
||||||
|
struct perf_event_header *header,
|
||||||
|
struct perf_sample_data *data,
|
||||||
|
struct perf_counter *counter);
|
||||||
|
extern void perf_prepare_sample(struct perf_event_header *header,
|
||||||
|
struct perf_sample_data *data,
|
||||||
|
struct perf_counter *counter,
|
||||||
|
struct pt_regs *regs);
|
||||||
|
|
||||||
extern int perf_counter_overflow(struct perf_counter *counter, int nmi,
|
extern int perf_counter_overflow(struct perf_counter *counter, int nmi,
|
||||||
struct perf_sample_data *data);
|
struct perf_sample_data *data,
|
||||||
extern void perf_counter_output(struct perf_counter *counter, int nmi,
|
struct pt_regs *regs);
|
||||||
struct perf_sample_data *data);
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Return 1 for a software counter, 0 for a hardware counter
|
* Return 1 for a software counter, 0 for a hardware counter
|
||||||
@ -781,6 +814,12 @@ extern void perf_tpcounter_event(int event_id, u64 addr, u64 count,
|
|||||||
#define perf_instruction_pointer(regs) instruction_pointer(regs)
|
#define perf_instruction_pointer(regs) instruction_pointer(regs)
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
extern int perf_output_begin(struct perf_output_handle *handle,
|
||||||
|
struct perf_counter *counter, unsigned int size,
|
||||||
|
int nmi, int sample);
|
||||||
|
extern void perf_output_end(struct perf_output_handle *handle);
|
||||||
|
extern void perf_output_copy(struct perf_output_handle *handle,
|
||||||
|
const void *buf, unsigned int len);
|
||||||
#else
|
#else
|
||||||
static inline void
|
static inline void
|
||||||
perf_counter_task_sched_in(struct task_struct *task, int cpu) { }
|
perf_counter_task_sched_in(struct task_struct *task, int cpu) { }
|
||||||
@ -807,7 +846,28 @@ static inline void perf_counter_mmap(struct vm_area_struct *vma) { }
|
|||||||
static inline void perf_counter_comm(struct task_struct *tsk) { }
|
static inline void perf_counter_comm(struct task_struct *tsk) { }
|
||||||
static inline void perf_counter_fork(struct task_struct *tsk) { }
|
static inline void perf_counter_fork(struct task_struct *tsk) { }
|
||||||
static inline void perf_counter_init(void) { }
|
static inline void perf_counter_init(void) { }
|
||||||
|
|
||||||
|
static inline int
|
||||||
|
perf_output_begin(struct perf_output_handle *handle, struct perf_counter *c,
|
||||||
|
unsigned int size, int nmi, int sample) { }
|
||||||
|
static inline void perf_output_end(struct perf_output_handle *handle) { }
|
||||||
|
static inline void
|
||||||
|
perf_output_copy(struct perf_output_handle *handle,
|
||||||
|
const void *buf, unsigned int len) { }
|
||||||
|
static inline void
|
||||||
|
perf_output_sample(struct perf_output_handle *handle,
|
||||||
|
struct perf_event_header *header,
|
||||||
|
struct perf_sample_data *data,
|
||||||
|
struct perf_counter *counter) { }
|
||||||
|
static inline void
|
||||||
|
perf_prepare_sample(struct perf_event_header *header,
|
||||||
|
struct perf_sample_data *data,
|
||||||
|
struct perf_counter *counter,
|
||||||
|
struct pt_regs *regs) { }
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
#define perf_output_put(handle, x) \
|
||||||
|
perf_output_copy((handle), &(x), sizeof(x))
|
||||||
|
|
||||||
#endif /* __KERNEL__ */
|
#endif /* __KERNEL__ */
|
||||||
#endif /* _LINUX_PERF_COUNTER_H */
|
#endif /* _LINUX_PERF_COUNTER_H */
|
||||||
|
@ -2512,18 +2512,6 @@ __weak struct perf_callchain_entry *perf_callchain(struct pt_regs *regs)
|
|||||||
/*
|
/*
|
||||||
* Output
|
* Output
|
||||||
*/
|
*/
|
||||||
|
|
||||||
struct perf_output_handle {
|
|
||||||
struct perf_counter *counter;
|
|
||||||
struct perf_mmap_data *data;
|
|
||||||
unsigned long head;
|
|
||||||
unsigned long offset;
|
|
||||||
int nmi;
|
|
||||||
int sample;
|
|
||||||
int locked;
|
|
||||||
unsigned long flags;
|
|
||||||
};
|
|
||||||
|
|
||||||
static bool perf_output_space(struct perf_mmap_data *data, unsigned long tail,
|
static bool perf_output_space(struct perf_mmap_data *data, unsigned long tail,
|
||||||
unsigned long offset, unsigned long head)
|
unsigned long offset, unsigned long head)
|
||||||
{
|
{
|
||||||
@ -2633,8 +2621,8 @@ out:
|
|||||||
local_irq_restore(handle->flags);
|
local_irq_restore(handle->flags);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void perf_output_copy(struct perf_output_handle *handle,
|
void perf_output_copy(struct perf_output_handle *handle,
|
||||||
const void *buf, unsigned int len)
|
const void *buf, unsigned int len)
|
||||||
{
|
{
|
||||||
unsigned int pages_mask;
|
unsigned int pages_mask;
|
||||||
unsigned int offset;
|
unsigned int offset;
|
||||||
@ -2669,12 +2657,9 @@ static void perf_output_copy(struct perf_output_handle *handle,
|
|||||||
WARN_ON_ONCE(((long)(handle->head - handle->offset)) < 0);
|
WARN_ON_ONCE(((long)(handle->head - handle->offset)) < 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
#define perf_output_put(handle, x) \
|
int perf_output_begin(struct perf_output_handle *handle,
|
||||||
perf_output_copy((handle), &(x), sizeof(x))
|
struct perf_counter *counter, unsigned int size,
|
||||||
|
int nmi, int sample)
|
||||||
static int perf_output_begin(struct perf_output_handle *handle,
|
|
||||||
struct perf_counter *counter, unsigned int size,
|
|
||||||
int nmi, int sample)
|
|
||||||
{
|
{
|
||||||
struct perf_counter *output_counter;
|
struct perf_counter *output_counter;
|
||||||
struct perf_mmap_data *data;
|
struct perf_mmap_data *data;
|
||||||
@ -2756,7 +2741,7 @@ out:
|
|||||||
return -ENOSPC;
|
return -ENOSPC;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void perf_output_end(struct perf_output_handle *handle)
|
void perf_output_end(struct perf_output_handle *handle)
|
||||||
{
|
{
|
||||||
struct perf_counter *counter = handle->counter;
|
struct perf_counter *counter = handle->counter;
|
||||||
struct perf_mmap_data *data = handle->data;
|
struct perf_mmap_data *data = handle->data;
|
||||||
@ -2870,82 +2855,151 @@ static void perf_output_read(struct perf_output_handle *handle,
|
|||||||
perf_output_read_one(handle, counter);
|
perf_output_read_one(handle, counter);
|
||||||
}
|
}
|
||||||
|
|
||||||
void perf_counter_output(struct perf_counter *counter, int nmi,
|
void perf_output_sample(struct perf_output_handle *handle,
|
||||||
struct perf_sample_data *data)
|
struct perf_event_header *header,
|
||||||
|
struct perf_sample_data *data,
|
||||||
|
struct perf_counter *counter)
|
||||||
|
{
|
||||||
|
u64 sample_type = data->type;
|
||||||
|
|
||||||
|
perf_output_put(handle, *header);
|
||||||
|
|
||||||
|
if (sample_type & PERF_SAMPLE_IP)
|
||||||
|
perf_output_put(handle, data->ip);
|
||||||
|
|
||||||
|
if (sample_type & PERF_SAMPLE_TID)
|
||||||
|
perf_output_put(handle, data->tid_entry);
|
||||||
|
|
||||||
|
if (sample_type & PERF_SAMPLE_TIME)
|
||||||
|
perf_output_put(handle, data->time);
|
||||||
|
|
||||||
|
if (sample_type & PERF_SAMPLE_ADDR)
|
||||||
|
perf_output_put(handle, data->addr);
|
||||||
|
|
||||||
|
if (sample_type & PERF_SAMPLE_ID)
|
||||||
|
perf_output_put(handle, data->id);
|
||||||
|
|
||||||
|
if (sample_type & PERF_SAMPLE_STREAM_ID)
|
||||||
|
perf_output_put(handle, data->stream_id);
|
||||||
|
|
||||||
|
if (sample_type & PERF_SAMPLE_CPU)
|
||||||
|
perf_output_put(handle, data->cpu_entry);
|
||||||
|
|
||||||
|
if (sample_type & PERF_SAMPLE_PERIOD)
|
||||||
|
perf_output_put(handle, data->period);
|
||||||
|
|
||||||
|
if (sample_type & PERF_SAMPLE_READ)
|
||||||
|
perf_output_read(handle, counter);
|
||||||
|
|
||||||
|
if (sample_type & PERF_SAMPLE_CALLCHAIN) {
|
||||||
|
if (data->callchain) {
|
||||||
|
int size = 1;
|
||||||
|
|
||||||
|
if (data->callchain)
|
||||||
|
size += data->callchain->nr;
|
||||||
|
|
||||||
|
size *= sizeof(u64);
|
||||||
|
|
||||||
|
perf_output_copy(handle, data->callchain, size);
|
||||||
|
} else {
|
||||||
|
u64 nr = 0;
|
||||||
|
perf_output_put(handle, nr);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (sample_type & PERF_SAMPLE_RAW) {
|
||||||
|
if (data->raw) {
|
||||||
|
perf_output_put(handle, data->raw->size);
|
||||||
|
perf_output_copy(handle, data->raw->data,
|
||||||
|
data->raw->size);
|
||||||
|
} else {
|
||||||
|
struct {
|
||||||
|
u32 size;
|
||||||
|
u32 data;
|
||||||
|
} raw = {
|
||||||
|
.size = sizeof(u32),
|
||||||
|
.data = 0,
|
||||||
|
};
|
||||||
|
perf_output_put(handle, raw);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
void perf_prepare_sample(struct perf_event_header *header,
|
||||||
|
struct perf_sample_data *data,
|
||||||
|
struct perf_counter *counter,
|
||||||
|
struct pt_regs *regs)
|
||||||
{
|
{
|
||||||
int ret;
|
|
||||||
u64 sample_type = counter->attr.sample_type;
|
u64 sample_type = counter->attr.sample_type;
|
||||||
struct perf_output_handle handle;
|
|
||||||
struct perf_event_header header;
|
|
||||||
u64 ip;
|
|
||||||
struct {
|
|
||||||
u32 pid, tid;
|
|
||||||
} tid_entry;
|
|
||||||
struct perf_callchain_entry *callchain = NULL;
|
|
||||||
int callchain_size = 0;
|
|
||||||
u64 time;
|
|
||||||
struct {
|
|
||||||
u32 cpu, reserved;
|
|
||||||
} cpu_entry;
|
|
||||||
|
|
||||||
header.type = PERF_EVENT_SAMPLE;
|
data->type = sample_type;
|
||||||
header.size = sizeof(header);
|
|
||||||
|
|
||||||
header.misc = 0;
|
header->type = PERF_EVENT_SAMPLE;
|
||||||
header.misc |= perf_misc_flags(data->regs);
|
header->size = sizeof(*header);
|
||||||
|
|
||||||
|
header->misc = 0;
|
||||||
|
header->misc |= perf_misc_flags(regs);
|
||||||
|
|
||||||
if (sample_type & PERF_SAMPLE_IP) {
|
if (sample_type & PERF_SAMPLE_IP) {
|
||||||
ip = perf_instruction_pointer(data->regs);
|
data->ip = perf_instruction_pointer(regs);
|
||||||
header.size += sizeof(ip);
|
|
||||||
|
header->size += sizeof(data->ip);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (sample_type & PERF_SAMPLE_TID) {
|
if (sample_type & PERF_SAMPLE_TID) {
|
||||||
/* namespace issues */
|
/* namespace issues */
|
||||||
tid_entry.pid = perf_counter_pid(counter, current);
|
data->tid_entry.pid = perf_counter_pid(counter, current);
|
||||||
tid_entry.tid = perf_counter_tid(counter, current);
|
data->tid_entry.tid = perf_counter_tid(counter, current);
|
||||||
|
|
||||||
header.size += sizeof(tid_entry);
|
header->size += sizeof(data->tid_entry);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (sample_type & PERF_SAMPLE_TIME) {
|
if (sample_type & PERF_SAMPLE_TIME) {
|
||||||
/*
|
/*
|
||||||
* Maybe do better on x86 and provide cpu_clock_nmi()
|
* Maybe do better on x86 and provide cpu_clock_nmi()
|
||||||
*/
|
*/
|
||||||
time = sched_clock();
|
data->time = sched_clock();
|
||||||
|
|
||||||
header.size += sizeof(u64);
|
header->size += sizeof(data->time);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (sample_type & PERF_SAMPLE_ADDR)
|
if (sample_type & PERF_SAMPLE_ADDR)
|
||||||
header.size += sizeof(u64);
|
header->size += sizeof(data->addr);
|
||||||
|
|
||||||
if (sample_type & PERF_SAMPLE_ID)
|
if (sample_type & PERF_SAMPLE_ID) {
|
||||||
header.size += sizeof(u64);
|
data->id = primary_counter_id(counter);
|
||||||
|
|
||||||
if (sample_type & PERF_SAMPLE_STREAM_ID)
|
header->size += sizeof(data->id);
|
||||||
header.size += sizeof(u64);
|
}
|
||||||
|
|
||||||
|
if (sample_type & PERF_SAMPLE_STREAM_ID) {
|
||||||
|
data->stream_id = counter->id;
|
||||||
|
|
||||||
|
header->size += sizeof(data->stream_id);
|
||||||
|
}
|
||||||
|
|
||||||
if (sample_type & PERF_SAMPLE_CPU) {
|
if (sample_type & PERF_SAMPLE_CPU) {
|
||||||
header.size += sizeof(cpu_entry);
|
data->cpu_entry.cpu = raw_smp_processor_id();
|
||||||
|
data->cpu_entry.reserved = 0;
|
||||||
|
|
||||||
cpu_entry.cpu = raw_smp_processor_id();
|
header->size += sizeof(data->cpu_entry);
|
||||||
cpu_entry.reserved = 0;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if (sample_type & PERF_SAMPLE_PERIOD)
|
if (sample_type & PERF_SAMPLE_PERIOD)
|
||||||
header.size += sizeof(u64);
|
header->size += sizeof(data->period);
|
||||||
|
|
||||||
if (sample_type & PERF_SAMPLE_READ)
|
if (sample_type & PERF_SAMPLE_READ)
|
||||||
header.size += perf_counter_read_size(counter);
|
header->size += perf_counter_read_size(counter);
|
||||||
|
|
||||||
if (sample_type & PERF_SAMPLE_CALLCHAIN) {
|
if (sample_type & PERF_SAMPLE_CALLCHAIN) {
|
||||||
callchain = perf_callchain(data->regs);
|
int size = 1;
|
||||||
|
|
||||||
if (callchain) {
|
data->callchain = perf_callchain(regs);
|
||||||
callchain_size = (1 + callchain->nr) * sizeof(u64);
|
|
||||||
header.size += callchain_size;
|
if (data->callchain)
|
||||||
} else
|
size += data->callchain->nr;
|
||||||
header.size += sizeof(u64);
|
|
||||||
|
header->size += size * sizeof(u64);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (sample_type & PERF_SAMPLE_RAW) {
|
if (sample_type & PERF_SAMPLE_RAW) {
|
||||||
@ -2957,69 +3011,23 @@ void perf_counter_output(struct perf_counter *counter, int nmi,
|
|||||||
size += sizeof(u32);
|
size += sizeof(u32);
|
||||||
|
|
||||||
WARN_ON_ONCE(size & (sizeof(u64)-1));
|
WARN_ON_ONCE(size & (sizeof(u64)-1));
|
||||||
header.size += size;
|
header->size += size;
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
ret = perf_output_begin(&handle, counter, header.size, nmi, 1);
|
static void perf_counter_output(struct perf_counter *counter, int nmi,
|
||||||
if (ret)
|
struct perf_sample_data *data,
|
||||||
|
struct pt_regs *regs)
|
||||||
|
{
|
||||||
|
struct perf_output_handle handle;
|
||||||
|
struct perf_event_header header;
|
||||||
|
|
||||||
|
perf_prepare_sample(&header, data, counter, regs);
|
||||||
|
|
||||||
|
if (perf_output_begin(&handle, counter, header.size, nmi, 1))
|
||||||
return;
|
return;
|
||||||
|
|
||||||
perf_output_put(&handle, header);
|
perf_output_sample(&handle, &header, data, counter);
|
||||||
|
|
||||||
if (sample_type & PERF_SAMPLE_IP)
|
|
||||||
perf_output_put(&handle, ip);
|
|
||||||
|
|
||||||
if (sample_type & PERF_SAMPLE_TID)
|
|
||||||
perf_output_put(&handle, tid_entry);
|
|
||||||
|
|
||||||
if (sample_type & PERF_SAMPLE_TIME)
|
|
||||||
perf_output_put(&handle, time);
|
|
||||||
|
|
||||||
if (sample_type & PERF_SAMPLE_ADDR)
|
|
||||||
perf_output_put(&handle, data->addr);
|
|
||||||
|
|
||||||
if (sample_type & PERF_SAMPLE_ID) {
|
|
||||||
u64 id = primary_counter_id(counter);
|
|
||||||
|
|
||||||
perf_output_put(&handle, id);
|
|
||||||
}
|
|
||||||
|
|
||||||
if (sample_type & PERF_SAMPLE_STREAM_ID)
|
|
||||||
perf_output_put(&handle, counter->id);
|
|
||||||
|
|
||||||
if (sample_type & PERF_SAMPLE_CPU)
|
|
||||||
perf_output_put(&handle, cpu_entry);
|
|
||||||
|
|
||||||
if (sample_type & PERF_SAMPLE_PERIOD)
|
|
||||||
perf_output_put(&handle, data->period);
|
|
||||||
|
|
||||||
if (sample_type & PERF_SAMPLE_READ)
|
|
||||||
perf_output_read(&handle, counter);
|
|
||||||
|
|
||||||
if (sample_type & PERF_SAMPLE_CALLCHAIN) {
|
|
||||||
if (callchain)
|
|
||||||
perf_output_copy(&handle, callchain, callchain_size);
|
|
||||||
else {
|
|
||||||
u64 nr = 0;
|
|
||||||
perf_output_put(&handle, nr);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if (sample_type & PERF_SAMPLE_RAW) {
|
|
||||||
if (data->raw) {
|
|
||||||
perf_output_put(&handle, data->raw->size);
|
|
||||||
perf_output_copy(&handle, data->raw->data, data->raw->size);
|
|
||||||
} else {
|
|
||||||
struct {
|
|
||||||
u32 size;
|
|
||||||
u32 data;
|
|
||||||
} raw = {
|
|
||||||
.size = sizeof(u32),
|
|
||||||
.data = 0,
|
|
||||||
};
|
|
||||||
perf_output_put(&handle, raw);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
perf_output_end(&handle);
|
perf_output_end(&handle);
|
||||||
}
|
}
|
||||||
@ -3501,7 +3509,8 @@ static void perf_log_throttle(struct perf_counter *counter, int enable)
|
|||||||
*/
|
*/
|
||||||
|
|
||||||
static int __perf_counter_overflow(struct perf_counter *counter, int nmi,
|
static int __perf_counter_overflow(struct perf_counter *counter, int nmi,
|
||||||
int throttle, struct perf_sample_data *data)
|
int throttle, struct perf_sample_data *data,
|
||||||
|
struct pt_regs *regs)
|
||||||
{
|
{
|
||||||
int events = atomic_read(&counter->event_limit);
|
int events = atomic_read(&counter->event_limit);
|
||||||
struct hw_perf_counter *hwc = &counter->hw;
|
struct hw_perf_counter *hwc = &counter->hw;
|
||||||
@ -3557,14 +3566,15 @@ static int __perf_counter_overflow(struct perf_counter *counter, int nmi,
|
|||||||
perf_counter_disable(counter);
|
perf_counter_disable(counter);
|
||||||
}
|
}
|
||||||
|
|
||||||
perf_counter_output(counter, nmi, data);
|
perf_counter_output(counter, nmi, data, regs);
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
int perf_counter_overflow(struct perf_counter *counter, int nmi,
|
int perf_counter_overflow(struct perf_counter *counter, int nmi,
|
||||||
struct perf_sample_data *data)
|
struct perf_sample_data *data,
|
||||||
|
struct pt_regs *regs)
|
||||||
{
|
{
|
||||||
return __perf_counter_overflow(counter, nmi, 1, data);
|
return __perf_counter_overflow(counter, nmi, 1, data, regs);
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@ -3602,7 +3612,8 @@ again:
|
|||||||
}
|
}
|
||||||
|
|
||||||
static void perf_swcounter_overflow(struct perf_counter *counter,
|
static void perf_swcounter_overflow(struct perf_counter *counter,
|
||||||
int nmi, struct perf_sample_data *data)
|
int nmi, struct perf_sample_data *data,
|
||||||
|
struct pt_regs *regs)
|
||||||
{
|
{
|
||||||
struct hw_perf_counter *hwc = &counter->hw;
|
struct hw_perf_counter *hwc = &counter->hw;
|
||||||
int throttle = 0;
|
int throttle = 0;
|
||||||
@ -3615,7 +3626,8 @@ static void perf_swcounter_overflow(struct perf_counter *counter,
|
|||||||
return;
|
return;
|
||||||
|
|
||||||
for (; overflow; overflow--) {
|
for (; overflow; overflow--) {
|
||||||
if (__perf_counter_overflow(counter, nmi, throttle, data)) {
|
if (__perf_counter_overflow(counter, nmi, throttle,
|
||||||
|
data, regs)) {
|
||||||
/*
|
/*
|
||||||
* We inhibit the overflow from happening when
|
* We inhibit the overflow from happening when
|
||||||
* hwc->interrupts == MAX_INTERRUPTS.
|
* hwc->interrupts == MAX_INTERRUPTS.
|
||||||
@ -3634,7 +3646,8 @@ static void perf_swcounter_unthrottle(struct perf_counter *counter)
|
|||||||
}
|
}
|
||||||
|
|
||||||
static void perf_swcounter_add(struct perf_counter *counter, u64 nr,
|
static void perf_swcounter_add(struct perf_counter *counter, u64 nr,
|
||||||
int nmi, struct perf_sample_data *data)
|
int nmi, struct perf_sample_data *data,
|
||||||
|
struct pt_regs *regs)
|
||||||
{
|
{
|
||||||
struct hw_perf_counter *hwc = &counter->hw;
|
struct hw_perf_counter *hwc = &counter->hw;
|
||||||
|
|
||||||
@ -3643,11 +3656,11 @@ static void perf_swcounter_add(struct perf_counter *counter, u64 nr,
|
|||||||
if (!hwc->sample_period)
|
if (!hwc->sample_period)
|
||||||
return;
|
return;
|
||||||
|
|
||||||
if (!data->regs)
|
if (!regs)
|
||||||
return;
|
return;
|
||||||
|
|
||||||
if (!atomic64_add_negative(nr, &hwc->period_left))
|
if (!atomic64_add_negative(nr, &hwc->period_left))
|
||||||
perf_swcounter_overflow(counter, nmi, data);
|
perf_swcounter_overflow(counter, nmi, data, regs);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int perf_swcounter_is_counting(struct perf_counter *counter)
|
static int perf_swcounter_is_counting(struct perf_counter *counter)
|
||||||
@ -3706,7 +3719,8 @@ static int perf_swcounter_match(struct perf_counter *counter,
|
|||||||
static void perf_swcounter_ctx_event(struct perf_counter_context *ctx,
|
static void perf_swcounter_ctx_event(struct perf_counter_context *ctx,
|
||||||
enum perf_type_id type,
|
enum perf_type_id type,
|
||||||
u32 event, u64 nr, int nmi,
|
u32 event, u64 nr, int nmi,
|
||||||
struct perf_sample_data *data)
|
struct perf_sample_data *data,
|
||||||
|
struct pt_regs *regs)
|
||||||
{
|
{
|
||||||
struct perf_counter *counter;
|
struct perf_counter *counter;
|
||||||
|
|
||||||
@ -3715,8 +3729,8 @@ static void perf_swcounter_ctx_event(struct perf_counter_context *ctx,
|
|||||||
|
|
||||||
rcu_read_lock();
|
rcu_read_lock();
|
||||||
list_for_each_entry_rcu(counter, &ctx->event_list, event_entry) {
|
list_for_each_entry_rcu(counter, &ctx->event_list, event_entry) {
|
||||||
if (perf_swcounter_match(counter, type, event, data->regs))
|
if (perf_swcounter_match(counter, type, event, regs))
|
||||||
perf_swcounter_add(counter, nr, nmi, data);
|
perf_swcounter_add(counter, nr, nmi, data, regs);
|
||||||
}
|
}
|
||||||
rcu_read_unlock();
|
rcu_read_unlock();
|
||||||
}
|
}
|
||||||
@ -3737,7 +3751,8 @@ static int *perf_swcounter_recursion_context(struct perf_cpu_context *cpuctx)
|
|||||||
|
|
||||||
static void do_perf_swcounter_event(enum perf_type_id type, u32 event,
|
static void do_perf_swcounter_event(enum perf_type_id type, u32 event,
|
||||||
u64 nr, int nmi,
|
u64 nr, int nmi,
|
||||||
struct perf_sample_data *data)
|
struct perf_sample_data *data,
|
||||||
|
struct pt_regs *regs)
|
||||||
{
|
{
|
||||||
struct perf_cpu_context *cpuctx = &get_cpu_var(perf_cpu_context);
|
struct perf_cpu_context *cpuctx = &get_cpu_var(perf_cpu_context);
|
||||||
int *recursion = perf_swcounter_recursion_context(cpuctx);
|
int *recursion = perf_swcounter_recursion_context(cpuctx);
|
||||||
@ -3750,7 +3765,7 @@ static void do_perf_swcounter_event(enum perf_type_id type, u32 event,
|
|||||||
barrier();
|
barrier();
|
||||||
|
|
||||||
perf_swcounter_ctx_event(&cpuctx->ctx, type, event,
|
perf_swcounter_ctx_event(&cpuctx->ctx, type, event,
|
||||||
nr, nmi, data);
|
nr, nmi, data, regs);
|
||||||
rcu_read_lock();
|
rcu_read_lock();
|
||||||
/*
|
/*
|
||||||
* doesn't really matter which of the child contexts the
|
* doesn't really matter which of the child contexts the
|
||||||
@ -3758,7 +3773,7 @@ static void do_perf_swcounter_event(enum perf_type_id type, u32 event,
|
|||||||
*/
|
*/
|
||||||
ctx = rcu_dereference(current->perf_counter_ctxp);
|
ctx = rcu_dereference(current->perf_counter_ctxp);
|
||||||
if (ctx)
|
if (ctx)
|
||||||
perf_swcounter_ctx_event(ctx, type, event, nr, nmi, data);
|
perf_swcounter_ctx_event(ctx, type, event, nr, nmi, data, regs);
|
||||||
rcu_read_unlock();
|
rcu_read_unlock();
|
||||||
|
|
||||||
barrier();
|
barrier();
|
||||||
@ -3772,11 +3787,11 @@ void __perf_swcounter_event(u32 event, u64 nr, int nmi,
|
|||||||
struct pt_regs *regs, u64 addr)
|
struct pt_regs *regs, u64 addr)
|
||||||
{
|
{
|
||||||
struct perf_sample_data data = {
|
struct perf_sample_data data = {
|
||||||
.regs = regs,
|
|
||||||
.addr = addr,
|
.addr = addr,
|
||||||
};
|
};
|
||||||
|
|
||||||
do_perf_swcounter_event(PERF_TYPE_SOFTWARE, event, nr, nmi, &data);
|
do_perf_swcounter_event(PERF_TYPE_SOFTWARE, event, nr, nmi,
|
||||||
|
&data, regs);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void perf_swcounter_read(struct perf_counter *counter)
|
static void perf_swcounter_read(struct perf_counter *counter)
|
||||||
@ -3813,6 +3828,7 @@ static enum hrtimer_restart perf_swcounter_hrtimer(struct hrtimer *hrtimer)
|
|||||||
{
|
{
|
||||||
enum hrtimer_restart ret = HRTIMER_RESTART;
|
enum hrtimer_restart ret = HRTIMER_RESTART;
|
||||||
struct perf_sample_data data;
|
struct perf_sample_data data;
|
||||||
|
struct pt_regs *regs;
|
||||||
struct perf_counter *counter;
|
struct perf_counter *counter;
|
||||||
u64 period;
|
u64 period;
|
||||||
|
|
||||||
@ -3820,17 +3836,17 @@ static enum hrtimer_restart perf_swcounter_hrtimer(struct hrtimer *hrtimer)
|
|||||||
counter->pmu->read(counter);
|
counter->pmu->read(counter);
|
||||||
|
|
||||||
data.addr = 0;
|
data.addr = 0;
|
||||||
data.regs = get_irq_regs();
|
regs = get_irq_regs();
|
||||||
/*
|
/*
|
||||||
* In case we exclude kernel IPs or are somehow not in interrupt
|
* In case we exclude kernel IPs or are somehow not in interrupt
|
||||||
* context, provide the next best thing, the user IP.
|
* context, provide the next best thing, the user IP.
|
||||||
*/
|
*/
|
||||||
if ((counter->attr.exclude_kernel || !data.regs) &&
|
if ((counter->attr.exclude_kernel || !regs) &&
|
||||||
!counter->attr.exclude_user)
|
!counter->attr.exclude_user)
|
||||||
data.regs = task_pt_regs(current);
|
regs = task_pt_regs(current);
|
||||||
|
|
||||||
if (data.regs) {
|
if (regs) {
|
||||||
if (perf_counter_overflow(counter, 0, &data))
|
if (perf_counter_overflow(counter, 0, &data, regs))
|
||||||
ret = HRTIMER_NORESTART;
|
ret = HRTIMER_NORESTART;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -3966,15 +3982,17 @@ void perf_tpcounter_event(int event_id, u64 addr, u64 count, void *record,
|
|||||||
};
|
};
|
||||||
|
|
||||||
struct perf_sample_data data = {
|
struct perf_sample_data data = {
|
||||||
.regs = get_irq_regs(),
|
|
||||||
.addr = addr,
|
.addr = addr,
|
||||||
.raw = &raw,
|
.raw = &raw,
|
||||||
};
|
};
|
||||||
|
|
||||||
if (!data.regs)
|
struct pt_regs *regs = get_irq_regs();
|
||||||
data.regs = task_pt_regs(current);
|
|
||||||
|
|
||||||
do_perf_swcounter_event(PERF_TYPE_TRACEPOINT, event_id, count, 1, &data);
|
if (!regs)
|
||||||
|
regs = task_pt_regs(current);
|
||||||
|
|
||||||
|
do_perf_swcounter_event(PERF_TYPE_TRACEPOINT, event_id, count, 1,
|
||||||
|
&data, regs);
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(perf_tpcounter_event);
|
EXPORT_SYMBOL_GPL(perf_tpcounter_event);
|
||||||
|
|
||||||
|
Loading…
x
Reference in New Issue
Block a user