// SPDX-License-Identifier: GPL-2.0 #include "util/debug.h" #include "util/dso.h" #include "util/event.h" #include "util/map.h" #include "util/symbol.h" #include "util/sort.h" #include "util/evsel.h" #include "util/evlist.h" #include "util/machine.h" #include "util/parse-events.h" #include "util/thread.h" #include "tests/tests.h" #include "tests/hists_common.h" #include <linux/kernel.h> struct sample { u32 pid; u64 ip; struct thread *thread; struct map *map; struct symbol *sym; }; /* For the numbers, see hists_common.c */ static struct sample fake_samples[] = { /* perf [kernel] schedule() */ { .pid = FAKE_PID_PERF1, .ip = FAKE_IP_KERNEL_SCHEDULE, }, /* perf [perf] main() */ { .pid = FAKE_PID_PERF1, .ip = FAKE_IP_PERF_MAIN, }, /* perf [perf] cmd_record() */ { .pid = FAKE_PID_PERF1, .ip = FAKE_IP_PERF_CMD_RECORD, }, /* perf [libc] malloc() */ { .pid = FAKE_PID_PERF1, .ip = FAKE_IP_LIBC_MALLOC, }, /* perf [libc] free() */ { .pid = FAKE_PID_PERF1, .ip = FAKE_IP_LIBC_FREE, }, /* perf [perf] main() */ { .pid = FAKE_PID_PERF2, .ip = FAKE_IP_PERF_MAIN, }, /* perf [kernel] page_fault() */ { .pid = FAKE_PID_PERF2, .ip = FAKE_IP_KERNEL_PAGE_FAULT, }, /* bash [bash] main() */ { .pid = FAKE_PID_BASH, .ip = FAKE_IP_BASH_MAIN, }, /* bash [bash] xmalloc() */ { .pid = FAKE_PID_BASH, .ip = FAKE_IP_BASH_XMALLOC, }, /* bash [kernel] page_fault() */ { .pid = FAKE_PID_BASH, .ip = FAKE_IP_KERNEL_PAGE_FAULT, }, }; /* * Will be cast to struct ip_callchain which has all 64 bit entries * of nr and ips[]. */ static u64 fake_callchains[][10] = { /* schedule => run_command => main */ { 3, FAKE_IP_KERNEL_SCHEDULE, FAKE_IP_PERF_RUN_COMMAND, FAKE_IP_PERF_MAIN, }, /* main */ { 1, FAKE_IP_PERF_MAIN, }, /* cmd_record => run_command => main */ { 3, FAKE_IP_PERF_CMD_RECORD, FAKE_IP_PERF_RUN_COMMAND, FAKE_IP_PERF_MAIN, }, /* malloc => cmd_record => run_command => main */ { 4, FAKE_IP_LIBC_MALLOC, FAKE_IP_PERF_CMD_RECORD, FAKE_IP_PERF_RUN_COMMAND, FAKE_IP_PERF_MAIN, }, /* free => cmd_record => run_command => main */ { 4, FAKE_IP_LIBC_FREE, FAKE_IP_PERF_CMD_RECORD, FAKE_IP_PERF_RUN_COMMAND, FAKE_IP_PERF_MAIN, }, /* main */ { 1, FAKE_IP_PERF_MAIN, }, /* page_fault => sys_perf_event_open => run_command => main */ { 4, FAKE_IP_KERNEL_PAGE_FAULT, FAKE_IP_KERNEL_SYS_PERF_EVENT_OPEN, FAKE_IP_PERF_RUN_COMMAND, FAKE_IP_PERF_MAIN, }, /* main */ { 1, FAKE_IP_BASH_MAIN, }, /* xmalloc => malloc => xmalloc => malloc => xmalloc => main */ { 6, FAKE_IP_BASH_XMALLOC, FAKE_IP_LIBC_MALLOC, FAKE_IP_BASH_XMALLOC, FAKE_IP_LIBC_MALLOC, FAKE_IP_BASH_XMALLOC, FAKE_IP_BASH_MAIN, }, /* page_fault => malloc => main */ { 3, FAKE_IP_KERNEL_PAGE_FAULT, FAKE_IP_LIBC_MALLOC, FAKE_IP_BASH_MAIN, }, }; static int add_hist_entries(struct hists *hists, struct machine *machine) { struct addr_location al; struct evsel *evsel = hists_to_evsel(hists); struct perf_sample sample = { .period = 1000, }; size_t i; addr_location__init(&al); for (i = 0; i < ARRAY_SIZE(fake_samples); i++) { struct hist_entry_iter iter = { .evsel = evsel, .sample = &sample, .hide_unresolved = false, }; if (symbol_conf.cumulate_callchain) iter.ops = &hist_iter_cumulative; else iter.ops = &hist_iter_normal; sample.cpumode = PERF_RECORD_MISC_USER; sample.pid = fake_samples[i].pid; sample.tid = fake_samples[i].pid; sample.ip = fake_samples[i].ip; sample.callchain = (struct ip_callchain *)fake_callchains[i]; if (machine__resolve(machine, &al, &sample) < 0) goto out; if (hist_entry_iter__add(&iter, &al, sysctl_perf_event_max_stack, NULL) < 0) { goto out; } thread__put(fake_samples[i].thread); fake_samples[i].thread = thread__get(al.thread); map__put(fake_samples[i].map); fake_samples[i].map = map__get(al.map); fake_samples[i].sym = al.sym; } addr_location__exit(&al); return TEST_OK; out: pr_debug("Not enough memory for adding a hist entry\n"); addr_location__exit(&al); return TEST_FAIL; } static void del_hist_entries(struct hists *hists) { struct hist_entry *he; struct rb_root_cached *root_in; struct rb_root_cached *root_out; struct rb_node *node; if (hists__has(hists, need_collapse)) root_in = &hists->entries_collapsed; else root_in = hists->entries_in; root_out = &hists->entries; while (!RB_EMPTY_ROOT(&root_out->rb_root)) { node = rb_first_cached(root_out); he = rb_entry(node, struct hist_entry, rb_node); rb_erase_cached(node, root_out); rb_erase_cached(&he->rb_node_in, root_in); hist_entry__delete(he); } } static void put_fake_samples(void) { size_t i; for (i = 0; i < ARRAY_SIZE(fake_samples); i++) { map__zput(fake_samples[i].map); thread__zput(fake_samples[i].thread); } } typedef int (*test_fn_t)(struct evsel *, struct machine *); #define COMM(he) (thread__comm_str(he->thread)) #define DSO(he) (map__dso(he->ms.map)->short_name) #define SYM(he) (he->ms.sym->name) #define CPU(he) (he->cpu) #define DEPTH(he) (he->callchain->max_depth) #define CDSO(cl) (map__dso(cl->ms.map)->short_name) #define CSYM(cl) (cl->ms.sym->name) struct result { u64 children; u64 self; const char *comm; const char *dso; const char *sym; }; struct callchain_result { u64 nr; struct { const char *dso; const char *sym; } node[10]; }; static int do_test(struct hists *hists, struct result *expected, size_t nr_expected, struct callchain_result *expected_callchain, size_t nr_callchain) { char buf[32]; size_t i, c; struct hist_entry *he; struct rb_root *root; struct rb_node *node; struct callchain_node *cnode; struct callchain_list *clist; /* * adding and deleting hist entries must be done outside of this * function since TEST_ASSERT_VAL() returns in case of failure. */ hists__collapse_resort(hists, NULL); evsel__output_resort(hists_to_evsel(hists), NULL); if (verbose > 2) { pr_info("use callchain: %d, cumulate callchain: %d\n", symbol_conf.use_callchain, symbol_conf.cumulate_callchain); print_hists_out(hists); } root = &hists->entries.rb_root; for (node = rb_first(root), i = 0; node && (he = rb_entry(node, struct hist_entry, rb_node)); node = rb_next(node), i++) { scnprintf(buf, sizeof(buf), "Invalid hist entry #%zd", i); TEST_ASSERT_VAL("Incorrect number of hist entry", i < nr_expected); TEST_ASSERT_VAL(buf, he->stat.period == expected[i].self && !strcmp(COMM(he), expected[i].comm) && !strcmp(DSO(he), expected[i].dso) && !strcmp(SYM(he), expected[i].sym)); if (symbol_conf.cumulate_callchain) TEST_ASSERT_VAL(buf, he->stat_acc->period == expected[i].children); if (!symbol_conf.use_callchain) continue; /* check callchain entries */ root = &he->callchain->node.rb_root; TEST_ASSERT_VAL("callchains expected", !RB_EMPTY_ROOT(root)); cnode = rb_entry(rb_first(root), struct callchain_node, rb_node); c = 0; list_for_each_entry(clist, &cnode->val, list) { scnprintf(buf, sizeof(buf), "Invalid callchain entry #%zd/%zd", i, c); TEST_ASSERT_VAL("Incorrect number of callchain entry", c < expected_callchain[i].nr); TEST_ASSERT_VAL(buf, !strcmp(CDSO(clist), expected_callchain[i].node[c].dso) && !strcmp(CSYM(clist), expected_callchain[i].node[c].sym)); c++; } /* TODO: handle multiple child nodes properly */ TEST_ASSERT_VAL("Incorrect number of callchain entry", c <= expected_callchain[i].nr); } TEST_ASSERT_VAL("Incorrect number of hist entry", i == nr_expected); TEST_ASSERT_VAL("Incorrect number of callchain entry", !symbol_conf.use_callchain || nr_expected == nr_callchain); return 0; } /* NO callchain + NO children */ static int test1(struct evsel *evsel, struct machine *machine) { int err; struct hists *hists = evsel__hists(evsel); /* * expected output: * * Overhead Command Shared Object Symbol * ======== ======= ============= ============== * 20.00% perf perf [.] main * 10.00% bash [kernel] [k] page_fault * 10.00% bash bash [.] main * 10.00% bash bash [.] xmalloc * 10.00% perf [kernel] [k] page_fault * 10.00% perf [kernel] [k] schedule * 10.00% perf libc [.] free * 10.00% perf libc [.] malloc * 10.00% perf perf [.] cmd_record */ struct result expected[] = { { 0, 2000, "perf", "perf", "main" }, { 0, 1000, "bash", "[kernel]", "page_fault" }, { 0, 1000, "bash", "bash", "main" }, { 0, 1000, "bash", "bash", "xmalloc" }, { 0, 1000, "perf", "[kernel]", "page_fault" }, { 0, 1000, "perf", "[kernel]", "schedule" }, { 0, 1000, "perf", "libc", "free" }, { 0, 1000, "perf", "libc", "malloc" }, { 0, 1000, "perf", "perf", "cmd_record" }, }; symbol_conf.use_callchain = false; symbol_conf.cumulate_callchain = false; evsel__reset_sample_bit(evsel, CALLCHAIN); setup_sorting(NULL); callchain_register_param(&callchain_param); err = add_hist_entries(hists, machine); if (err < 0) goto out; err = do_test(hists, expected, ARRAY_SIZE(expected), NULL, 0); out: del_hist_entries(hists); reset_output_field(); return err; } /* callchain + NO children */ static int test2(struct evsel *evsel, struct machine *machine) { int err; struct hists *hists = evsel__hists(evsel); /* * expected output: * * Overhead Command Shared Object Symbol * ======== ======= ============= ============== * 20.00% perf perf [.] main * | * --- main * * 10.00% bash [kernel] [k] page_fault * | * --- page_fault * malloc * main * * 10.00% bash bash [.] main * | * --- main * * 10.00% bash bash [.] xmalloc * | * --- xmalloc * malloc * xmalloc <--- NOTE: there's a cycle * malloc * xmalloc * main * * 10.00% perf [kernel] [k] page_fault * | * --- page_fault * sys_perf_event_open * run_command * main * * 10.00% perf [kernel] [k] schedule * | * --- schedule * run_command * main * * 10.00% perf libc [.] free * | * --- free * cmd_record * run_command * main * * 10.00% perf libc [.] malloc * | * --- malloc * cmd_record * run_command * main * * 10.00% perf perf [.] cmd_record * | * --- cmd_record * run_command * main * */ struct result expected[] = { { 0, 2000, "perf", "perf", "main" }, { 0, 1000, "bash", "[kernel]", "page_fault" }, { 0, 1000, "bash", "bash", "main" }, { 0, 1000, "bash", "bash", "xmalloc" }, { 0, 1000, "perf", "[kernel]", "page_fault" }, { 0, 1000, "perf", "[kernel]", "schedule" }, { 0, 1000, "perf", "libc", "free" }, { 0, 1000, "perf", "libc", "malloc" }, { 0, 1000, "perf", "perf", "cmd_record" }, }; struct callchain_result expected_callchain[] = { { 1, { { "perf", "main" }, }, }, { 3, { { "[kernel]", "page_fault" }, { "libc", "malloc" }, { "bash", "main" }, }, }, { 1, { { "bash", "main" }, }, }, { 6, { { "bash", "xmalloc" }, { "libc", "malloc" }, { "bash", "xmalloc" }, { "libc", "malloc" }, { "bash", "xmalloc" }, { "bash", "main" }, }, }, { 4, { { "[kernel]", "page_fault" }, { "[kernel]", "sys_perf_event_open" }, { "perf", "run_command" }, { "perf", "main" }, }, }, { 3, { { "[kernel]", "schedule" }, { "perf", "run_command" }, { "perf", "main" }, }, }, { 4, { { "libc", "free" }, { "perf", "cmd_record" }, { "perf", "run_command" }, { "perf", "main" }, }, }, { 4, { { "libc", "malloc" }, { "perf", "cmd_record" }, { "perf", "run_command" }, { "perf", "main" }, }, }, { 3, { { "perf", "cmd_record" }, { "perf", "run_command" }, { "perf", "main" }, }, }, }; symbol_conf.use_callchain = true; symbol_conf.cumulate_callchain = false; evsel__set_sample_bit(evsel, CALLCHAIN); setup_sorting(NULL); callchain_register_param(&callchain_param); err = add_hist_entries(hists, machine); if (err < 0) goto out; err = do_test(hists, expected, ARRAY_SIZE(expected), expected_callchain, ARRAY_SIZE(expected_callchain)); out: del_hist_entries(hists); reset_output_field(); return err; } /* NO callchain + children */ static int test3(struct evsel *evsel, struct machine *machine) { int err; struct hists *hists = evsel__hists(evsel); /* * expected output: * * Children Self Command Shared Object Symbol * ======== ======== ======= ============= ======================= * 70.00% 20.00% perf perf [.] main * 50.00% 0.00% perf perf [.] run_command * 30.00% 10.00% bash bash [.] main * 30.00% 10.00% perf perf [.] cmd_record * 20.00% 0.00% bash libc [.] malloc * 10.00% 10.00% bash [kernel] [k] page_fault * 10.00% 10.00% bash bash [.] xmalloc * 10.00% 10.00% perf [kernel] [k] page_fault * 10.00% 10.00% perf libc [.] malloc * 10.00% 10.00% perf [kernel] [k] schedule * 10.00% 10.00% perf libc [.] free * 10.00% 0.00% perf [kernel] [k] sys_perf_event_open */ struct result expected[] = { { 7000, 2000, "perf", "perf", "main" }, { 5000, 0, "perf", "perf", "run_command" }, { 3000, 1000, "bash", "bash", "main" }, { 3000, 1000, "perf", "perf", "cmd_record" }, { 2000, 0, "bash", "libc", "malloc" }, { 1000, 1000, "bash", "[kernel]", "page_fault" }, { 1000, 1000, "bash", "bash", "xmalloc" }, { 1000, 1000, "perf", "[kernel]", "page_fault" }, { 1000, 1000, "perf", "[kernel]", "schedule" }, { 1000, 1000, "perf", "libc", "free" }, { 1000, 1000, "perf", "libc", "malloc" }, { 1000, 0, "perf", "[kernel]", "sys_perf_event_open" }, }; symbol_conf.use_callchain = false; symbol_conf.cumulate_callchain = true; evsel__reset_sample_bit(evsel, CALLCHAIN); setup_sorting(NULL); callchain_register_param(&callchain_param); err = add_hist_entries(hists, machine); if (err < 0) goto out; err = do_test(hists, expected, ARRAY_SIZE(expected), NULL, 0); out: del_hist_entries(hists); reset_output_field(); return err; } /* callchain + children */ static int test4(struct evsel *evsel, struct machine *machine) { int err; struct hists *hists = evsel__hists(evsel); /* * expected output: * * Children Self Command Shared Object Symbol * ======== ======== ======= ============= ======================= * 70.00% 20.00% perf perf [.] main * | * --- main * * 50.00% 0.00% perf perf [.] run_command * | * --- run_command * main * * 30.00% 10.00% bash bash [.] main * | * --- main * * 30.00% 10.00% perf perf [.] cmd_record * | * --- cmd_record * run_command * main * * 20.00% 0.00% bash libc [.] malloc * | * --- malloc * | * |--50.00%-- xmalloc * | main * --50.00%-- main * * 10.00% 10.00% bash [kernel] [k] page_fault * | * --- page_fault * malloc * main * * 10.00% 10.00% bash bash [.] xmalloc * | * --- xmalloc * malloc * xmalloc <--- NOTE: there's a cycle * malloc * xmalloc * main * * 10.00% 0.00% perf [kernel] [k] sys_perf_event_open * | * --- sys_perf_event_open * run_command * main * * 10.00% 10.00% perf [kernel] [k] page_fault * | * --- page_fault * sys_perf_event_open * run_command * main * * 10.00% 10.00% perf [kernel] [k] schedule * | * --- schedule * run_command * main * * 10.00% 10.00% perf libc [.] free * | * --- free * cmd_record * run_command * main * * 10.00% 10.00% perf libc [.] malloc * | * --- malloc * cmd_record * run_command * main * */ struct result expected[] = { { 7000, 2000, "perf", "perf", "main" }, { 5000, 0, "perf", "perf", "run_command" }, { 3000, 1000, "bash", "bash", "main" }, { 3000, 1000, "perf", "perf", "cmd_record" }, { 2000, 0, "bash", "libc", "malloc" }, { 1000, 1000, "bash", "[kernel]", "page_fault" }, { 1000, 1000, "bash", "bash", "xmalloc" }, { 1000, 0, "perf", "[kernel]", "sys_perf_event_open" }, { 1000, 1000, "perf", "[kernel]", "page_fault" }, { 1000, 1000, "perf", "[kernel]", "schedule" }, { 1000, 1000, "perf", "libc", "free" }, { 1000, 1000, "perf", "libc", "malloc" }, }; struct callchain_result expected_callchain[] = { { 1, { { "perf", "main" }, }, }, { 2, { { "perf", "run_command" }, { "perf", "main" }, }, }, { 1, { { "bash", "main" }, }, }, { 3, { { "perf", "cmd_record" }, { "perf", "run_command" }, { "perf", "main" }, }, }, { 4, { { "libc", "malloc" }, { "bash", "xmalloc" }, { "bash", "main" }, { "bash", "main" }, }, }, { 3, { { "[kernel]", "page_fault" }, { "libc", "malloc" }, { "bash", "main" }, }, }, { 6, { { "bash", "xmalloc" }, { "libc", "malloc" }, { "bash", "xmalloc" }, { "libc", "malloc" }, { "bash", "xmalloc" }, { "bash", "main" }, }, }, { 3, { { "[kernel]", "sys_perf_event_open" }, { "perf", "run_command" }, { "perf", "main" }, }, }, { 4, { { "[kernel]", "page_fault" }, { "[kernel]", "sys_perf_event_open" }, { "perf", "run_command" }, { "perf", "main" }, }, }, { 3, { { "[kernel]", "schedule" }, { "perf", "run_command" }, { "perf", "main" }, }, }, { 4, { { "libc", "free" }, { "perf", "cmd_record" }, { "perf", "run_command" }, { "perf", "main" }, }, }, { 4, { { "libc", "malloc" }, { "perf", "cmd_record" }, { "perf", "run_command" }, { "perf", "main" }, }, }, }; symbol_conf.use_callchain = true; symbol_conf.cumulate_callchain = true; evsel__set_sample_bit(evsel, CALLCHAIN); setup_sorting(NULL); callchain_param = callchain_param_default; callchain_register_param(&callchain_param); err = add_hist_entries(hists, machine); if (err < 0) goto out; err = do_test(hists, expected, ARRAY_SIZE(expected), expected_callchain, ARRAY_SIZE(expected_callchain)); out: del_hist_entries(hists); reset_output_field(); return err; } static int test__hists_cumulate(struct test_suite *test __maybe_unused, int subtest __maybe_unused) { int err = TEST_FAIL; struct machines machines; struct machine *machine; struct evsel *evsel; struct evlist *evlist = evlist__new(); size_t i; test_fn_t testcases[] = { test1, test2, test3, test4, }; TEST_ASSERT_VAL("No memory", evlist); err = parse_event(evlist, "cpu-clock"); if (err) goto out; err = TEST_FAIL; machines__init(&machines); /* setup threads/dso/map/symbols also */ machine = setup_fake_machine(&machines); if (!machine) goto out; if (verbose > 1) machine__fprintf(machine, stderr); evsel = evlist__first(evlist); for (i = 0; i < ARRAY_SIZE(testcases); i++) { err = testcases[i](evsel, machine); if (err < 0) break; } out: /* tear down everything */ evlist__delete(evlist); machines__exit(&machines); put_fake_samples(); return err; } DEFINE_SUITE("Cumulate child hist entries", hists_cumulate);