mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/next/linux-next.git
synced 2025-01-09 23:39:18 +00:00
5f7fb89a11
All architectures that implement function graph also implements HAVE_FUNCTION_GRAPH_RET_ADDR_PTR. Remove it, as it is no longer a differentiator. Link: https://lore.kernel.org/linux-trace-kernel/20240611031737.982047614@goodmis.org Cc: Masami Hiramatsu <mhiramat@kernel.org> Cc: Mark Rutland <mark.rutland@arm.com> Cc: Mathieu Desnoyers <mathieu.desnoyers@efficios.com> Cc: Jonathan Corbet <corbet@lwn.net> Cc: Catalin Marinas <catalin.marinas@arm.com> Cc: Will Deacon <will@kernel.org> Cc: Guo Ren <guoren@kernel.org> Cc: Huacai Chen <chenhuacai@kernel.org> Cc: WANG Xuerui <kernel@xen0n.name> Cc: Michael Ellerman <mpe@ellerman.id.au> Cc: Nicholas Piggin <npiggin@gmail.com> Cc: Christophe Leroy <christophe.leroy@csgroup.eu> Cc: "Naveen N. Rao" <naveen.n.rao@linux.ibm.com> Cc: Paul Walmsley <paul.walmsley@sifive.com> Cc: Palmer Dabbelt <palmer@dabbelt.com> Cc: Albert Ou <aou@eecs.berkeley.edu> Cc: Heiko Carstens <hca@linux.ibm.com> Cc: Vasily Gorbik <gor@linux.ibm.com> Cc: Alexander Gordeev <agordeev@linux.ibm.com> Cc: Christian Borntraeger <borntraeger@linux.ibm.com> Cc: Sven Schnelle <svens@linux.ibm.com> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: Ingo Molnar <mingo@redhat.com> Cc: Borislav Petkov <bp@alien8.de> Cc: Dave Hansen <dave.hansen@linux.intel.com> Cc: "H. Peter Anvin" <hpa@zytor.com> Signed-off-by: Steven Rostedt (Google) <rostedt@goodmis.org>
229 lines
5.6 KiB
C
229 lines
5.6 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
/* Copyright (C) 2017 Andes Technology Corporation */
|
|
|
|
#ifndef _ASM_RISCV_FTRACE_H
|
|
#define _ASM_RISCV_FTRACE_H
|
|
|
|
/*
|
|
* The graph frame test is not possible if CONFIG_FRAME_POINTER is not enabled.
|
|
* Check arch/riscv/kernel/mcount.S for detail.
|
|
*/
|
|
#if defined(CONFIG_FUNCTION_GRAPH_TRACER) && defined(CONFIG_FRAME_POINTER)
|
|
#define HAVE_FUNCTION_GRAPH_FP_TEST
|
|
#endif
|
|
|
|
#define ARCH_SUPPORTS_FTRACE_OPS 1
|
|
#ifndef __ASSEMBLY__
|
|
|
|
extern void *return_address(unsigned int level);
|
|
|
|
#define ftrace_return_address(n) return_address(n)
|
|
|
|
void _mcount(void);
|
|
static inline unsigned long ftrace_call_adjust(unsigned long addr)
|
|
{
|
|
return addr;
|
|
}
|
|
|
|
/*
|
|
* Let's do like x86/arm64 and ignore the compat syscalls.
|
|
*/
|
|
#define ARCH_TRACE_IGNORE_COMPAT_SYSCALLS
|
|
static inline bool arch_trace_is_compat_syscall(struct pt_regs *regs)
|
|
{
|
|
return is_compat_task();
|
|
}
|
|
|
|
#define ARCH_HAS_SYSCALL_MATCH_SYM_NAME
|
|
static inline bool arch_syscall_match_sym_name(const char *sym,
|
|
const char *name)
|
|
{
|
|
/*
|
|
* Since all syscall functions have __riscv_ prefix, we must skip it.
|
|
* However, as we described above, we decided to ignore compat
|
|
* syscalls, so we don't care about __riscv_compat_ prefix here.
|
|
*/
|
|
return !strcmp(sym + 8, name);
|
|
}
|
|
|
|
struct dyn_arch_ftrace {
|
|
};
|
|
#endif
|
|
|
|
#ifdef CONFIG_DYNAMIC_FTRACE
|
|
/*
|
|
* A general call in RISC-V is a pair of insts:
|
|
* 1) auipc: setting high-20 pc-related bits to ra register
|
|
* 2) jalr: setting low-12 offset to ra, jump to ra, and set ra to
|
|
* return address (original pc + 4)
|
|
*
|
|
*<ftrace enable>:
|
|
* 0: auipc t0/ra, 0x?
|
|
* 4: jalr t0/ra, ?(t0/ra)
|
|
*
|
|
*<ftrace disable>:
|
|
* 0: nop
|
|
* 4: nop
|
|
*
|
|
* Dynamic ftrace generates probes to call sites, so we must deal with
|
|
* both auipc and jalr at the same time.
|
|
*/
|
|
|
|
#define MCOUNT_ADDR ((unsigned long)_mcount)
|
|
#define JALR_SIGN_MASK (0x00000800)
|
|
#define JALR_OFFSET_MASK (0x00000fff)
|
|
#define AUIPC_OFFSET_MASK (0xfffff000)
|
|
#define AUIPC_PAD (0x00001000)
|
|
#define JALR_SHIFT 20
|
|
#define JALR_RA (0x000080e7)
|
|
#define AUIPC_RA (0x00000097)
|
|
#define JALR_T0 (0x000282e7)
|
|
#define AUIPC_T0 (0x00000297)
|
|
#define NOP4 (0x00000013)
|
|
|
|
#define to_jalr_t0(offset) \
|
|
(((offset & JALR_OFFSET_MASK) << JALR_SHIFT) | JALR_T0)
|
|
|
|
#define to_auipc_t0(offset) \
|
|
((offset & JALR_SIGN_MASK) ? \
|
|
(((offset & AUIPC_OFFSET_MASK) + AUIPC_PAD) | AUIPC_T0) : \
|
|
((offset & AUIPC_OFFSET_MASK) | AUIPC_T0))
|
|
|
|
#define make_call_t0(caller, callee, call) \
|
|
do { \
|
|
unsigned int offset = \
|
|
(unsigned long) callee - (unsigned long) caller; \
|
|
call[0] = to_auipc_t0(offset); \
|
|
call[1] = to_jalr_t0(offset); \
|
|
} while (0)
|
|
|
|
#define to_jalr_ra(offset) \
|
|
(((offset & JALR_OFFSET_MASK) << JALR_SHIFT) | JALR_RA)
|
|
|
|
#define to_auipc_ra(offset) \
|
|
((offset & JALR_SIGN_MASK) ? \
|
|
(((offset & AUIPC_OFFSET_MASK) + AUIPC_PAD) | AUIPC_RA) : \
|
|
((offset & AUIPC_OFFSET_MASK) | AUIPC_RA))
|
|
|
|
#define make_call_ra(caller, callee, call) \
|
|
do { \
|
|
unsigned int offset = \
|
|
(unsigned long) callee - (unsigned long) caller; \
|
|
call[0] = to_auipc_ra(offset); \
|
|
call[1] = to_jalr_ra(offset); \
|
|
} while (0)
|
|
|
|
/*
|
|
* Let auipc+jalr be the basic *mcount unit*, so we make it 8 bytes here.
|
|
*/
|
|
#define MCOUNT_INSN_SIZE 8
|
|
|
|
#ifndef __ASSEMBLY__
|
|
struct dyn_ftrace;
|
|
int ftrace_init_nop(struct module *mod, struct dyn_ftrace *rec);
|
|
#define ftrace_init_nop ftrace_init_nop
|
|
|
|
#ifdef CONFIG_DYNAMIC_FTRACE_WITH_ARGS
|
|
#define arch_ftrace_get_regs(regs) NULL
|
|
struct ftrace_ops;
|
|
struct ftrace_regs {
|
|
unsigned long epc;
|
|
unsigned long ra;
|
|
unsigned long sp;
|
|
unsigned long s0;
|
|
unsigned long t1;
|
|
union {
|
|
unsigned long args[8];
|
|
struct {
|
|
unsigned long a0;
|
|
unsigned long a1;
|
|
unsigned long a2;
|
|
unsigned long a3;
|
|
unsigned long a4;
|
|
unsigned long a5;
|
|
unsigned long a6;
|
|
unsigned long a7;
|
|
};
|
|
};
|
|
};
|
|
|
|
static __always_inline unsigned long ftrace_regs_get_instruction_pointer(const struct ftrace_regs
|
|
*fregs)
|
|
{
|
|
return fregs->epc;
|
|
}
|
|
|
|
static __always_inline void ftrace_regs_set_instruction_pointer(struct ftrace_regs *fregs,
|
|
unsigned long pc)
|
|
{
|
|
fregs->epc = pc;
|
|
}
|
|
|
|
static __always_inline unsigned long ftrace_regs_get_stack_pointer(const struct ftrace_regs *fregs)
|
|
{
|
|
return fregs->sp;
|
|
}
|
|
|
|
static __always_inline unsigned long ftrace_regs_get_argument(struct ftrace_regs *fregs,
|
|
unsigned int n)
|
|
{
|
|
if (n < 8)
|
|
return fregs->args[n];
|
|
return 0;
|
|
}
|
|
|
|
static __always_inline unsigned long ftrace_regs_get_return_value(const struct ftrace_regs *fregs)
|
|
{
|
|
return fregs->a0;
|
|
}
|
|
|
|
static __always_inline void ftrace_regs_set_return_value(struct ftrace_regs *fregs,
|
|
unsigned long ret)
|
|
{
|
|
fregs->a0 = ret;
|
|
}
|
|
|
|
static __always_inline void ftrace_override_function_with_return(struct ftrace_regs *fregs)
|
|
{
|
|
fregs->epc = fregs->ra;
|
|
}
|
|
|
|
int ftrace_regs_query_register_offset(const char *name);
|
|
|
|
void ftrace_graph_func(unsigned long ip, unsigned long parent_ip,
|
|
struct ftrace_ops *op, struct ftrace_regs *fregs);
|
|
#define ftrace_graph_func ftrace_graph_func
|
|
|
|
static inline void arch_ftrace_set_direct_caller(struct ftrace_regs *fregs, unsigned long addr)
|
|
{
|
|
fregs->t1 = addr;
|
|
}
|
|
#endif /* CONFIG_DYNAMIC_FTRACE_WITH_ARGS */
|
|
|
|
#endif /* __ASSEMBLY__ */
|
|
|
|
#endif /* CONFIG_DYNAMIC_FTRACE */
|
|
|
|
#ifndef __ASSEMBLY__
|
|
#ifdef CONFIG_FUNCTION_GRAPH_TRACER
|
|
struct fgraph_ret_regs {
|
|
unsigned long a1;
|
|
unsigned long a0;
|
|
unsigned long s0;
|
|
unsigned long ra;
|
|
};
|
|
|
|
static inline unsigned long fgraph_ret_regs_return_value(struct fgraph_ret_regs *ret_regs)
|
|
{
|
|
return ret_regs->a0;
|
|
}
|
|
|
|
static inline unsigned long fgraph_ret_regs_frame_pointer(struct fgraph_ret_regs *ret_regs)
|
|
{
|
|
return ret_regs->s0;
|
|
}
|
|
#endif /* ifdef CONFIG_FUNCTION_GRAPH_TRACER */
|
|
#endif
|
|
|
|
#endif /* _ASM_RISCV_FTRACE_H */
|