linux/lib/test_kprobes.c

372 lines
9.4 KiB
C
Raw Normal View History

// SPDX-License-Identifier: GPL-2.0-or-later
/*
* test_kprobes.c - simple sanity test for *probes
*
* Copyright IBM Corp. 2008
*/
#include <linux/kernel.h>
#include <linux/kprobes.h>
#include <linux/random.h>
#include <kunit/test.h>
#define div_factor 3
static u32 rand1, preh_val, posth_val;
static u32 (*target)(u32 value);
static u32 (*target2)(u32 value);
static struct kunit *current_test;
static unsigned long (*internal_target)(void);
static unsigned long (*stacktrace_target)(void);
static unsigned long (*stacktrace_driver)(void);
static unsigned long target_return_address[2];
static noinline u32 kprobe_target(u32 value)
{
return (value / div_factor);
}
static int kp_pre_handler(struct kprobe *p, struct pt_regs *regs)
{
KUNIT_EXPECT_FALSE(current_test, preemptible());
preh_val = (rand1 / div_factor);
return 0;
}
static void kp_post_handler(struct kprobe *p, struct pt_regs *regs,
unsigned long flags)
{
KUNIT_EXPECT_FALSE(current_test, preemptible());
KUNIT_EXPECT_EQ(current_test, preh_val, (rand1 / div_factor));
posth_val = preh_val + div_factor;
}
static struct kprobe kp = {
.symbol_name = "kprobe_target",
.pre_handler = kp_pre_handler,
.post_handler = kp_post_handler
};
static void test_kprobe(struct kunit *test)
{
current_test = test;
KUNIT_EXPECT_EQ(test, 0, register_kprobe(&kp));
target(rand1);
unregister_kprobe(&kp);
KUNIT_EXPECT_NE(test, 0, preh_val);
KUNIT_EXPECT_NE(test, 0, posth_val);
}
static noinline u32 kprobe_target2(u32 value)
{
return (value / div_factor) + 1;
}
static noinline unsigned long kprobe_stacktrace_internal_target(void)
{
if (!target_return_address[0])
target_return_address[0] = (unsigned long)__builtin_return_address(0);
return target_return_address[0];
}
static noinline unsigned long kprobe_stacktrace_target(void)
{
if (!target_return_address[1])
target_return_address[1] = (unsigned long)__builtin_return_address(0);
if (internal_target)
internal_target();
return target_return_address[1];
}
static noinline unsigned long kprobe_stacktrace_driver(void)
{
if (stacktrace_target)
stacktrace_target();
/* This is for preventing inlining the function */
return (unsigned long)__builtin_return_address(0);
}
static int kp_pre_handler2(struct kprobe *p, struct pt_regs *regs)
{
preh_val = (rand1 / div_factor) + 1;
return 0;
}
static void kp_post_handler2(struct kprobe *p, struct pt_regs *regs,
unsigned long flags)
{
KUNIT_EXPECT_EQ(current_test, preh_val, (rand1 / div_factor) + 1);
posth_val = preh_val + div_factor;
}
static struct kprobe kp2 = {
.symbol_name = "kprobe_target2",
.pre_handler = kp_pre_handler2,
.post_handler = kp_post_handler2
};
static void test_kprobes(struct kunit *test)
{
struct kprobe *kps[2] = {&kp, &kp2};
current_test = test;
/* addr and flags should be cleard for reusing kprobe. */
kp.addr = NULL;
kp.flags = 0;
KUNIT_EXPECT_EQ(test, 0, register_kprobes(kps, 2));
preh_val = 0;
posth_val = 0;
target(rand1);
KUNIT_EXPECT_NE(test, 0, preh_val);
KUNIT_EXPECT_NE(test, 0, posth_val);
preh_val = 0;
posth_val = 0;
target2(rand1);
KUNIT_EXPECT_NE(test, 0, preh_val);
KUNIT_EXPECT_NE(test, 0, posth_val);
unregister_kprobes(kps, 2);
}
#ifdef CONFIG_KRETPROBES
static u32 krph_val;
static int entry_handler(struct kretprobe_instance *ri, struct pt_regs *regs)
{
KUNIT_EXPECT_FALSE(current_test, preemptible());
krph_val = (rand1 / div_factor);
return 0;
}
static int return_handler(struct kretprobe_instance *ri, struct pt_regs *regs)
{
unsigned long ret = regs_return_value(regs);
KUNIT_EXPECT_FALSE(current_test, preemptible());
KUNIT_EXPECT_EQ(current_test, ret, rand1 / div_factor);
KUNIT_EXPECT_NE(current_test, krph_val, 0);
krph_val = rand1;
return 0;
}
static struct kretprobe rp = {
.handler = return_handler,
.entry_handler = entry_handler,
.kp.symbol_name = "kprobe_target"
};
static void test_kretprobe(struct kunit *test)
{
current_test = test;
KUNIT_EXPECT_EQ(test, 0, register_kretprobe(&rp));
target(rand1);
unregister_kretprobe(&rp);
KUNIT_EXPECT_EQ(test, krph_val, rand1);
}
static int return_handler2(struct kretprobe_instance *ri, struct pt_regs *regs)
{
unsigned long ret = regs_return_value(regs);
KUNIT_EXPECT_EQ(current_test, ret, (rand1 / div_factor) + 1);
KUNIT_EXPECT_NE(current_test, krph_val, 0);
krph_val = rand1;
return 0;
}
static struct kretprobe rp2 = {
.handler = return_handler2,
.entry_handler = entry_handler,
.kp.symbol_name = "kprobe_target2"
};
static void test_kretprobes(struct kunit *test)
{
struct kretprobe *rps[2] = {&rp, &rp2};
current_test = test;
/* addr and flags should be cleard for reusing kprobe. */
rp.kp.addr = NULL;
rp.kp.flags = 0;
KUNIT_EXPECT_EQ(test, 0, register_kretprobes(rps, 2));
krph_val = 0;
target(rand1);
KUNIT_EXPECT_EQ(test, krph_val, rand1);
krph_val = 0;
target2(rand1);
KUNIT_EXPECT_EQ(test, krph_val, rand1);
unregister_kretprobes(rps, 2);
}
#ifdef CONFIG_ARCH_CORRECT_STACKTRACE_ON_KRETPROBE
#define STACK_BUF_SIZE 16
static unsigned long stack_buf[STACK_BUF_SIZE];
static int stacktrace_return_handler(struct kretprobe_instance *ri, struct pt_regs *regs)
{
unsigned long retval = regs_return_value(regs);
int i, ret;
KUNIT_EXPECT_FALSE(current_test, preemptible());
KUNIT_EXPECT_EQ(current_test, retval, target_return_address[1]);
/*
* Test stacktrace inside the kretprobe handler, this will involves
* kretprobe trampoline, but must include correct return address
* of the target function.
*/
ret = stack_trace_save(stack_buf, STACK_BUF_SIZE, 0);
KUNIT_EXPECT_NE(current_test, ret, 0);
for (i = 0; i < ret; i++) {
if (stack_buf[i] == target_return_address[1])
break;
}
KUNIT_EXPECT_NE(current_test, i, ret);
#if !IS_MODULE(CONFIG_KPROBES_SANITY_TEST)
/*
* Test stacktrace from pt_regs at the return address. Thus the stack
* trace must start from the target return address.
*/
ret = stack_trace_save_regs(regs, stack_buf, STACK_BUF_SIZE, 0);
KUNIT_EXPECT_NE(current_test, ret, 0);
KUNIT_EXPECT_EQ(current_test, stack_buf[0], target_return_address[1]);
#endif
return 0;
}
static struct kretprobe rp3 = {
.handler = stacktrace_return_handler,
.kp.symbol_name = "kprobe_stacktrace_target"
};
static void test_stacktrace_on_kretprobe(struct kunit *test)
{
unsigned long myretaddr = (unsigned long)__builtin_return_address(0);
current_test = test;
rp3.kp.addr = NULL;
rp3.kp.flags = 0;
/*
* Run the stacktrace_driver() to record correct return address in
* stacktrace_target() and ensure stacktrace_driver() call is not
* inlined by checking the return address of stacktrace_driver()
* and the return address of this function is different.
*/
KUNIT_ASSERT_NE(test, myretaddr, stacktrace_driver());
KUNIT_ASSERT_EQ(test, 0, register_kretprobe(&rp3));
KUNIT_ASSERT_NE(test, myretaddr, stacktrace_driver());
unregister_kretprobe(&rp3);
}
static int stacktrace_internal_return_handler(struct kretprobe_instance *ri, struct pt_regs *regs)
{
unsigned long retval = regs_return_value(regs);
int i, ret;
KUNIT_EXPECT_FALSE(current_test, preemptible());
KUNIT_EXPECT_EQ(current_test, retval, target_return_address[0]);
/*
* Test stacktrace inside the kretprobe handler for nested case.
* The unwinder will find the kretprobe_trampoline address on the
* return address, and kretprobe must solve that.
*/
ret = stack_trace_save(stack_buf, STACK_BUF_SIZE, 0);
KUNIT_EXPECT_NE(current_test, ret, 0);
for (i = 0; i < ret - 1; i++) {
if (stack_buf[i] == target_return_address[0]) {
KUNIT_EXPECT_EQ(current_test, stack_buf[i + 1], target_return_address[1]);
break;
}
}
KUNIT_EXPECT_NE(current_test, i, ret);
#if !IS_MODULE(CONFIG_KPROBES_SANITY_TEST)
/* Ditto for the regs version. */
ret = stack_trace_save_regs(regs, stack_buf, STACK_BUF_SIZE, 0);
KUNIT_EXPECT_NE(current_test, ret, 0);
KUNIT_EXPECT_EQ(current_test, stack_buf[0], target_return_address[0]);
KUNIT_EXPECT_EQ(current_test, stack_buf[1], target_return_address[1]);
#endif
return 0;
}
static struct kretprobe rp4 = {
.handler = stacktrace_internal_return_handler,
.kp.symbol_name = "kprobe_stacktrace_internal_target"
};
static void test_stacktrace_on_nested_kretprobe(struct kunit *test)
{
unsigned long myretaddr = (unsigned long)__builtin_return_address(0);
struct kretprobe *rps[2] = {&rp3, &rp4};
current_test = test;
rp3.kp.addr = NULL;
rp3.kp.flags = 0;
//KUNIT_ASSERT_NE(test, myretaddr, stacktrace_driver());
KUNIT_ASSERT_EQ(test, 0, register_kretprobes(rps, 2));
KUNIT_ASSERT_NE(test, myretaddr, stacktrace_driver());
unregister_kretprobes(rps, 2);
}
#endif /* CONFIG_ARCH_CORRECT_STACKTRACE_ON_KRETPROBE */
#endif /* CONFIG_KRETPROBES */
static int kprobes_test_init(struct kunit *test)
{
target = kprobe_target;
target2 = kprobe_target2;
stacktrace_target = kprobe_stacktrace_target;
internal_target = kprobe_stacktrace_internal_target;
stacktrace_driver = kprobe_stacktrace_driver;
do {
rand1 = prandom_u32();
} while (rand1 <= div_factor);
return 0;
}
static struct kunit_case kprobes_testcases[] = {
KUNIT_CASE(test_kprobe),
KUNIT_CASE(test_kprobes),
#ifdef CONFIG_KRETPROBES
KUNIT_CASE(test_kretprobe),
KUNIT_CASE(test_kretprobes),
#ifdef CONFIG_ARCH_CORRECT_STACKTRACE_ON_KRETPROBE
KUNIT_CASE(test_stacktrace_on_kretprobe),
KUNIT_CASE(test_stacktrace_on_nested_kretprobe),
#endif
#endif
{}
};
static struct kunit_suite kprobes_test_suite = {
.name = "kprobes_test",
.init = kprobes_test_init,
.test_cases = kprobes_testcases,
};
kunit_test_suites(&kprobes_test_suite);
MODULE_LICENSE("GPL");