|  | // SPDX-License-Identifier: GPL-2.0 | 
|  | #include <linux/delay.h> | 
|  | #include <linux/module.h> | 
|  | #include <linux/kthread.h> | 
|  | #include <linux/trace_clock.h> | 
|  |  | 
|  | #define CREATE_TRACE_POINTS | 
|  | #include "trace_benchmark.h" | 
|  |  | 
|  | static struct task_struct *bm_event_thread; | 
|  |  | 
|  | static char bm_str[BENCHMARK_EVENT_STRLEN] = "START"; | 
|  |  | 
|  | static u64 bm_total; | 
|  | static u64 bm_totalsq; | 
|  | static u64 bm_last; | 
|  | static u64 bm_max; | 
|  | static u64 bm_min; | 
|  | static u64 bm_first; | 
|  | static u64 bm_cnt; | 
|  | static u64 bm_stddev; | 
|  | static unsigned int bm_avg; | 
|  | static unsigned int bm_std; | 
|  |  | 
|  | static bool ok_to_run; | 
|  |  | 
|  | /* | 
|  | * This gets called in a loop recording the time it took to write | 
|  | * the tracepoint. What it writes is the time statistics of the last | 
|  | * tracepoint write. As there is nothing to write the first time | 
|  | * it simply writes "START". As the first write is cold cache and | 
|  | * the rest is hot, we save off that time in bm_first and it is | 
|  | * reported as "first", which is shown in the second write to the | 
|  | * tracepoint. The "first" field is writen within the statics from | 
|  | * then on but never changes. | 
|  | */ | 
|  | static void trace_do_benchmark(void) | 
|  | { | 
|  | u64 start; | 
|  | u64 stop; | 
|  | u64 delta; | 
|  | u64 stddev; | 
|  | u64 seed; | 
|  | u64 last_seed; | 
|  | unsigned int avg; | 
|  | unsigned int std = 0; | 
|  |  | 
|  | /* Only run if the tracepoint is actually active */ | 
|  | if (!trace_benchmark_event_enabled() || !tracing_is_on()) | 
|  | return; | 
|  |  | 
|  | local_irq_disable(); | 
|  | start = trace_clock_local(); | 
|  | trace_benchmark_event(bm_str); | 
|  | stop = trace_clock_local(); | 
|  | local_irq_enable(); | 
|  |  | 
|  | bm_cnt++; | 
|  |  | 
|  | delta = stop - start; | 
|  |  | 
|  | /* | 
|  | * The first read is cold cached, keep it separate from the | 
|  | * other calculations. | 
|  | */ | 
|  | if (bm_cnt == 1) { | 
|  | bm_first = delta; | 
|  | scnprintf(bm_str, BENCHMARK_EVENT_STRLEN, | 
|  | "first=%llu [COLD CACHED]", bm_first); | 
|  | return; | 
|  | } | 
|  |  | 
|  | bm_last = delta; | 
|  |  | 
|  | if (delta > bm_max) | 
|  | bm_max = delta; | 
|  | if (!bm_min || delta < bm_min) | 
|  | bm_min = delta; | 
|  |  | 
|  | /* | 
|  | * When bm_cnt is greater than UINT_MAX, it breaks the statistics | 
|  | * accounting. Freeze the statistics when that happens. | 
|  | * We should have enough data for the avg and stddev anyway. | 
|  | */ | 
|  | if (bm_cnt > UINT_MAX) { | 
|  | scnprintf(bm_str, BENCHMARK_EVENT_STRLEN, | 
|  | "last=%llu first=%llu max=%llu min=%llu ** avg=%u std=%d std^2=%lld", | 
|  | bm_last, bm_first, bm_max, bm_min, bm_avg, bm_std, bm_stddev); | 
|  | return; | 
|  | } | 
|  |  | 
|  | bm_total += delta; | 
|  | bm_totalsq += delta * delta; | 
|  |  | 
|  |  | 
|  | if (bm_cnt > 1) { | 
|  | /* | 
|  | * Apply Welford's method to calculate standard deviation: | 
|  | * s^2 = 1 / (n * (n-1)) * (n * \Sum (x_i)^2 - (\Sum x_i)^2) | 
|  | */ | 
|  | stddev = (u64)bm_cnt * bm_totalsq - bm_total * bm_total; | 
|  | do_div(stddev, (u32)bm_cnt); | 
|  | do_div(stddev, (u32)bm_cnt - 1); | 
|  | } else | 
|  | stddev = 0; | 
|  |  | 
|  | delta = bm_total; | 
|  | do_div(delta, bm_cnt); | 
|  | avg = delta; | 
|  |  | 
|  | if (stddev > 0) { | 
|  | int i = 0; | 
|  | /* | 
|  | * stddev is the square of standard deviation but | 
|  | * we want the actualy number. Use the average | 
|  | * as our seed to find the std. | 
|  | * | 
|  | * The next try is: | 
|  | *  x = (x + N/x) / 2 | 
|  | * | 
|  | * Where N is the squared number to find the square | 
|  | * root of. | 
|  | */ | 
|  | seed = avg; | 
|  | do { | 
|  | last_seed = seed; | 
|  | seed = stddev; | 
|  | if (!last_seed) | 
|  | break; | 
|  | do_div(seed, last_seed); | 
|  | seed += last_seed; | 
|  | do_div(seed, 2); | 
|  | } while (i++ < 10 && last_seed != seed); | 
|  |  | 
|  | std = seed; | 
|  | } | 
|  |  | 
|  | scnprintf(bm_str, BENCHMARK_EVENT_STRLEN, | 
|  | "last=%llu first=%llu max=%llu min=%llu avg=%u std=%d std^2=%lld", | 
|  | bm_last, bm_first, bm_max, bm_min, avg, std, stddev); | 
|  |  | 
|  | bm_std = std; | 
|  | bm_avg = avg; | 
|  | bm_stddev = stddev; | 
|  | } | 
|  |  | 
|  | static int benchmark_event_kthread(void *arg) | 
|  | { | 
|  | /* sleep a bit to make sure the tracepoint gets activated */ | 
|  | msleep(100); | 
|  |  | 
|  | while (!kthread_should_stop()) { | 
|  |  | 
|  | trace_do_benchmark(); | 
|  |  | 
|  | /* | 
|  | * We don't go to sleep, but let others run as well. | 
|  | * This is bascially a "yield()" to let any task that | 
|  | * wants to run, schedule in, but if the CPU is idle, | 
|  | * we'll keep burning cycles. | 
|  | * | 
|  | * Note the tasks_rcu_qs() version of cond_resched() will | 
|  | * notify synchronize_rcu_tasks() that this thread has | 
|  | * passed a quiescent state for rcu_tasks. Otherwise | 
|  | * this thread will never voluntarily schedule which would | 
|  | * block synchronize_rcu_tasks() indefinitely. | 
|  | */ | 
|  | cond_resched_tasks_rcu_qs(); | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * When the benchmark tracepoint is enabled, it calls this | 
|  | * function and the thread that calls the tracepoint is created. | 
|  | */ | 
|  | int trace_benchmark_reg(void) | 
|  | { | 
|  | if (!ok_to_run) { | 
|  | pr_warning("trace benchmark cannot be started via kernel command line\n"); | 
|  | return -EBUSY; | 
|  | } | 
|  |  | 
|  | bm_event_thread = kthread_run(benchmark_event_kthread, | 
|  | NULL, "event_benchmark"); | 
|  | if (IS_ERR(bm_event_thread)) { | 
|  | pr_warning("trace benchmark failed to create kernel thread\n"); | 
|  | return PTR_ERR(bm_event_thread); | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * When the benchmark tracepoint is disabled, it calls this | 
|  | * function and the thread that calls the tracepoint is deleted | 
|  | * and all the numbers are reset. | 
|  | */ | 
|  | void trace_benchmark_unreg(void) | 
|  | { | 
|  | if (!bm_event_thread) | 
|  | return; | 
|  |  | 
|  | kthread_stop(bm_event_thread); | 
|  | bm_event_thread = NULL; | 
|  |  | 
|  | strcpy(bm_str, "START"); | 
|  | bm_total = 0; | 
|  | bm_totalsq = 0; | 
|  | bm_last = 0; | 
|  | bm_max = 0; | 
|  | bm_min = 0; | 
|  | bm_cnt = 0; | 
|  | /* These don't need to be reset but reset them anyway */ | 
|  | bm_first = 0; | 
|  | bm_std = 0; | 
|  | bm_avg = 0; | 
|  | bm_stddev = 0; | 
|  | } | 
|  |  | 
|  | static __init int ok_to_run_trace_benchmark(void) | 
|  | { | 
|  | ok_to_run = true; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | early_initcall(ok_to_run_trace_benchmark); |