| .. SPDX-License-Identifier: GPL-2.0 |
| |
| ========================== |
| RCU Torture Test Operation |
| ========================== |
| |
| |
| CONFIG_RCU_TORTURE_TEST |
| ======================= |
| |
| The CONFIG_RCU_TORTURE_TEST config option is available for all RCU |
| implementations. It creates an rcutorture kernel module that can |
| be loaded to run a torture test. The test periodically outputs |
| status messages via printk(), which can be examined via the dmesg |
| command (perhaps grepping for "torture"). The test is started |
| when the module is loaded, and stops when the module is unloaded. |
| |
| Module parameters are prefixed by "rcutorture." in |
| Documentation/admin-guide/kernel-parameters.txt. |
| |
| Output |
| ====== |
| |
| The statistics output is as follows:: |
| |
| rcu-torture:--- Start of test: nreaders=16 nfakewriters=4 stat_interval=30 verbose=0 test_no_idle_hz=1 shuffle_interval=3 stutter=5 irqreader=1 fqs_duration=0 fqs_holdoff=0 fqs_stutter=3 test_boost=1/0 test_boost_interval=7 test_boost_duration=4 |
| rcu-torture: rtc: (null) ver: 155441 tfle: 0 rta: 155441 rtaf: 8884 rtf: 155440 rtmbe: 0 rtbe: 0 rtbke: 0 rtbre: 0 rtbf: 0 rtb: 0 nt: 3055767 |
| rcu-torture: Reader Pipe: 727860534 34213 0 0 0 0 0 0 0 0 0 |
| rcu-torture: Reader Batch: 727877838 17003 0 0 0 0 0 0 0 0 0 |
| rcu-torture: Free-Block Circulation: 155440 155440 155440 155440 155440 155440 155440 155440 155440 155440 0 |
| rcu-torture:--- End of test: SUCCESS: nreaders=16 nfakewriters=4 stat_interval=30 verbose=0 test_no_idle_hz=1 shuffle_interval=3 stutter=5 irqreader=1 fqs_duration=0 fqs_holdoff=0 fqs_stutter=3 test_boost=1/0 test_boost_interval=7 test_boost_duration=4 |
| |
| The command "dmesg | grep torture:" will extract this information on |
| most systems. On more esoteric configurations, it may be necessary to |
| use other commands to access the output of the printk()s used by |
| the RCU torture test. The printk()s use KERN_ALERT, so they should |
| be evident. ;-) |
| |
| The first and last lines show the rcutorture module parameters, and the |
| last line shows either "SUCCESS" or "FAILURE", based on rcutorture's |
| automatic determination as to whether RCU operated correctly. |
| |
| The entries are as follows: |
| |
| * "rtc": The hexadecimal address of the structure currently visible |
| to readers. |
| |
| * "ver": The number of times since boot that the RCU writer task |
| has changed the structure visible to readers. |
| |
| * "tfle": If non-zero, indicates that the "torture freelist" |
| containing structures to be placed into the "rtc" area is empty. |
| This condition is important, since it can fool you into thinking |
| that RCU is working when it is not. :-/ |
| |
| * "rta": Number of structures allocated from the torture freelist. |
| |
| * "rtaf": Number of allocations from the torture freelist that have |
| failed due to the list being empty. It is not unusual for this |
| to be non-zero, but it is bad for it to be a large fraction of |
| the value indicated by "rta". |
| |
| * "rtf": Number of frees into the torture freelist. |
| |
| * "rtmbe": A non-zero value indicates that rcutorture believes that |
| rcu_assign_pointer() and rcu_dereference() are not working |
| correctly. This value should be zero. |
| |
| * "rtbe": A non-zero value indicates that one of the rcu_barrier() |
| family of functions is not working correctly. |
| |
| * "rtbke": rcutorture was unable to create the real-time kthreads |
| used to force RCU priority inversion. This value should be zero. |
| |
| * "rtbre": Although rcutorture successfully created the kthreads |
| used to force RCU priority inversion, it was unable to set them |
| to the real-time priority level of 1. This value should be zero. |
| |
| * "rtbf": The number of times that RCU priority boosting failed |
| to resolve RCU priority inversion. |
| |
| * "rtb": The number of times that rcutorture attempted to force |
| an RCU priority inversion condition. If you are testing RCU |
| priority boosting via the "test_boost" module parameter, this |
| value should be non-zero. |
| |
| * "nt": The number of times rcutorture ran RCU read-side code from |
| within a timer handler. This value should be non-zero only |
| if you specified the "irqreader" module parameter. |
| |
| * "Reader Pipe": Histogram of "ages" of structures seen by readers. |
| If any entries past the first two are non-zero, RCU is broken. |
| And rcutorture prints the error flag string "!!!" to make sure |
| you notice. The age of a newly allocated structure is zero, |
| it becomes one when removed from reader visibility, and is |
| incremented once per grace period subsequently -- and is freed |
| after passing through (RCU_TORTURE_PIPE_LEN-2) grace periods. |
| |
| The output displayed above was taken from a correctly working |
| RCU. If you want to see what it looks like when broken, break |
| it yourself. ;-) |
| |
| * "Reader Batch": Another histogram of "ages" of structures seen |
| by readers, but in terms of counter flips (or batches) rather |
| than in terms of grace periods. The legal number of non-zero |
| entries is again two. The reason for this separate view is that |
| it is sometimes easier to get the third entry to show up in the |
| "Reader Batch" list than in the "Reader Pipe" list. |
| |
| * "Free-Block Circulation": Shows the number of torture structures |
| that have reached a given point in the pipeline. The first element |
| should closely correspond to the number of structures allocated, |
| the second to the number that have been removed from reader view, |
| and all but the last remaining to the corresponding number of |
| passes through a grace period. The last entry should be zero, |
| as it is only incremented if a torture structure's counter |
| somehow gets incremented farther than it should. |
| |
| Different implementations of RCU can provide implementation-specific |
| additional information. For example, Tree SRCU provides the following |
| additional line:: |
| |
| srcud-torture: Tree SRCU per-CPU(idx=0): 0(35,-21) 1(-4,24) 2(1,1) 3(-26,20) 4(28,-47) 5(-9,4) 6(-10,14) 7(-14,11) T(1,6) |
| |
| This line shows the per-CPU counter state, in this case for Tree SRCU |
| using a dynamically allocated srcu_struct (hence "srcud-" rather than |
| "srcu-"). The numbers in parentheses are the values of the "old" and |
| "current" counters for the corresponding CPU. The "idx" value maps the |
| "old" and "current" values to the underlying array, and is useful for |
| debugging. The final "T" entry contains the totals of the counters. |
| |
| Usage on Specific Kernel Builds |
| =============================== |
| |
| It is sometimes desirable to torture RCU on a specific kernel build, |
| for example, when preparing to put that kernel build into production. |
| In that case, the kernel should be built with CONFIG_RCU_TORTURE_TEST=m |
| so that the test can be started using modprobe and terminated using rmmod. |
| |
| For example, the following script may be used to torture RCU:: |
| |
| #!/bin/sh |
| |
| modprobe rcutorture |
| sleep 3600 |
| rmmod rcutorture |
| dmesg | grep torture: |
| |
| The output can be manually inspected for the error flag of "!!!". |
| One could of course create a more elaborate script that automatically |
| checked for such errors. The "rmmod" command forces a "SUCCESS", |
| "FAILURE", or "RCU_HOTPLUG" indication to be printk()ed. The first |
| two are self-explanatory, while the last indicates that while there |
| were no RCU failures, CPU-hotplug problems were detected. |
| |
| |
| Usage on Mainline Kernels |
| ========================= |
| |
| When using rcutorture to test changes to RCU itself, it is often |
| necessary to build a number of kernels in order to test that change |
| across a broad range of combinations of the relevant Kconfig options |
| and of the relevant kernel boot parameters. In this situation, use |
| of modprobe and rmmod can be quite time-consuming and error-prone. |
| |
| Therefore, the tools/testing/selftests/rcutorture/bin/kvm.sh |
| script is available for mainline testing for x86, arm64, and |
| powerpc. By default, it will run the series of tests specified by |
| tools/testing/selftests/rcutorture/configs/rcu/CFLIST, with each test |
| running for 30 minutes within a guest OS using a minimal userspace |
| supplied by an automatically generated initrd. After the tests are |
| complete, the resulting build products and console output are analyzed |
| for errors and the results of the runs are summarized. |
| |
| On larger systems, rcutorture testing can be accelerated by passing the |
| --cpus argument to kvm.sh. For example, on a 64-CPU system, "--cpus 43" |
| would use up to 43 CPUs to run tests concurrently, which as of v5.4 would |
| complete all the scenarios in two batches, reducing the time to complete |
| from about eight hours to about one hour (not counting the time to build |
| the sixteen kernels). The "--dryrun sched" argument will not run tests, |
| but rather tell you how the tests would be scheduled into batches. This |
| can be useful when working out how many CPUs to specify in the --cpus |
| argument. |
| |
| Not all changes require that all scenarios be run. For example, a change |
| to Tree SRCU might run only the SRCU-N and SRCU-P scenarios using the |
| --configs argument to kvm.sh as follows: "--configs 'SRCU-N SRCU-P'". |
| Large systems can run multiple copies of the full set of scenarios, |
| for example, a system with 448 hardware threads can run five instances |
| of the full set concurrently. To make this happen:: |
| |
| kvm.sh --cpus 448 --configs '5*CFLIST' |
| |
| Alternatively, such a system can run 56 concurrent instances of a single |
| eight-CPU scenario:: |
| |
| kvm.sh --cpus 448 --configs '56*TREE04' |
| |
| Or 28 concurrent instances of each of two eight-CPU scenarios:: |
| |
| kvm.sh --cpus 448 --configs '28*TREE03 28*TREE04' |
| |
| Of course, each concurrent instance will use memory, which can be |
| limited using the --memory argument, which defaults to 512M. Small |
| values for memory may require disabling the callback-flooding tests |
| using the --bootargs parameter discussed below. |
| |
| Sometimes additional debugging is useful, and in such cases the --kconfig |
| parameter to kvm.sh may be used, for example, ``--kconfig 'CONFIG_RCU_EQS_DEBUG=y'``. |
| In addition, there are the --gdb, --kasan, and --kcsan parameters. |
| Note that --gdb limits you to one scenario per kvm.sh run and requires |
| that you have another window open from which to run ``gdb`` as instructed |
| by the script. |
| |
| Kernel boot arguments can also be supplied, for example, to control |
| rcutorture's module parameters. For example, to test a change to RCU's |
| CPU stall-warning code, use "--bootargs 'rcutorture.stall_cpu=30'". |
| This will of course result in the scripting reporting a failure, namely |
| the resulting RCU CPU stall warning. As noted above, reducing memory may |
| require disabling rcutorture's callback-flooding tests:: |
| |
| kvm.sh --cpus 448 --configs '56*TREE04' --memory 128M \ |
| --bootargs 'rcutorture.fwd_progress=0' |
| |
| Sometimes all that is needed is a full set of kernel builds. This is |
| what the --buildonly parameter does. |
| |
| The --duration parameter can override the default run time of 30 minutes. |
| For example, ``--duration 2d`` would run for two days, ``--duration 3h`` |
| would run for three hours, ``--duration 5m`` would run for five minutes, |
| and ``--duration 45s`` would run for 45 seconds. This last can be useful |
| for tracking down rare boot-time failures. |
| |
| Finally, the --trust-make parameter allows each kernel build to reuse what |
| it can from the previous kernel build. Please note that without the |
| --trust-make parameter, your tags files may be demolished. |
| |
| There are additional more arcane arguments that are documented in the |
| source code of the kvm.sh script. |
| |
| If a run contains failures, the number of buildtime and runtime failures |
| is listed at the end of the kvm.sh output, which you really should redirect |
| to a file. The build products and console output of each run is kept in |
| tools/testing/selftests/rcutorture/res in timestamped directories. A |
| given directory can be supplied to kvm-find-errors.sh in order to have |
| it cycle you through summaries of errors and full error logs. For example:: |
| |
| tools/testing/selftests/rcutorture/bin/kvm-find-errors.sh \ |
| tools/testing/selftests/rcutorture/res/2020.01.20-15.54.23 |
| |
| However, it is often more convenient to access the files directly. |
| Files pertaining to all scenarios in a run reside in the top-level |
| directory (2020.01.20-15.54.23 in the example above), while per-scenario |
| files reside in a subdirectory named after the scenario (for example, |
| "TREE04"). If a given scenario ran more than once (as in "--configs |
| '56*TREE04'" above), the directories corresponding to the second and |
| subsequent runs of that scenario include a sequence number, for example, |
| "TREE04.2", "TREE04.3", and so on. |
| |
| The most frequently used file in the top-level directory is testid.txt. |
| If the test ran in a git repository, then this file contains the commit |
| that was tested and any uncommitted changes in diff format. |
| |
| The most frequently used files in each per-scenario-run directory are: |
| |
| .config: |
| This file contains the Kconfig options. |
| |
| Make.out: |
| This contains build output for a specific scenario. |
| |
| console.log: |
| This contains the console output for a specific scenario. |
| This file may be examined once the kernel has booted, but |
| it might not exist if the build failed. |
| |
| vmlinux: |
| This contains the kernel, which can be useful with tools like |
| objdump and gdb. |
| |
| A number of additional files are available, but are less frequently used. |
| Many are intended for debugging of rcutorture itself or of its scripting. |
| |
| As of v5.4, a successful run with the default set of scenarios produces |
| the following summary at the end of the run on a 12-CPU system:: |
| |
| SRCU-N ------- 804233 GPs (148.932/s) [srcu: g10008272 f0x0 ] |
| SRCU-P ------- 202320 GPs (37.4667/s) [srcud: g1809476 f0x0 ] |
| SRCU-t ------- 1122086 GPs (207.794/s) [srcu: g0 f0x0 ] |
| SRCU-u ------- 1111285 GPs (205.794/s) [srcud: g1 f0x0 ] |
| TASKS01 ------- 19666 GPs (3.64185/s) [tasks: g0 f0x0 ] |
| TASKS02 ------- 20541 GPs (3.80389/s) [tasks: g0 f0x0 ] |
| TASKS03 ------- 19416 GPs (3.59556/s) [tasks: g0 f0x0 ] |
| TINY01 ------- 836134 GPs (154.84/s) [rcu: g0 f0x0 ] n_max_cbs: 34198 |
| TINY02 ------- 850371 GPs (157.476/s) [rcu: g0 f0x0 ] n_max_cbs: 2631 |
| TREE01 ------- 162625 GPs (30.1157/s) [rcu: g1124169 f0x0 ] |
| TREE02 ------- 333003 GPs (61.6672/s) [rcu: g2647753 f0x0 ] n_max_cbs: 35844 |
| TREE03 ------- 306623 GPs (56.782/s) [rcu: g2975325 f0x0 ] n_max_cbs: 1496497 |
| CPU count limited from 16 to 12 |
| TREE04 ------- 246149 GPs (45.5831/s) [rcu: g1695737 f0x0 ] n_max_cbs: 434961 |
| TREE05 ------- 314603 GPs (58.2598/s) [rcu: g2257741 f0x2 ] n_max_cbs: 193997 |
| TREE07 ------- 167347 GPs (30.9902/s) [rcu: g1079021 f0x0 ] n_max_cbs: 478732 |
| CPU count limited from 16 to 12 |
| TREE09 ------- 752238 GPs (139.303/s) [rcu: g13075057 f0x0 ] n_max_cbs: 99011 |
| |
| |
| Repeated Runs |
| ============= |
| |
| Suppose that you are chasing down a rare boot-time failure. Although you |
| could use kvm.sh, doing so will rebuild the kernel on each run. If you |
| need (say) 1,000 runs to have confidence that you have fixed the bug, |
| these pointless rebuilds can become extremely annoying. |
| |
| This is why kvm-again.sh exists. |
| |
| Suppose that a previous kvm.sh run left its output in this directory:: |
| |
| tools/testing/selftests/rcutorture/res/2022.11.03-11.26.28 |
| |
| Then this run can be re-run without rebuilding as follow: |
| |
| kvm-again.sh tools/testing/selftests/rcutorture/res/2022.11.03-11.26.28 |
| |
| A few of the original run's kvm.sh parameters may be overridden, perhaps |
| most notably --duration and --bootargs. For example:: |
| |
| kvm-again.sh tools/testing/selftests/rcutorture/res/2022.11.03-11.26.28 \ |
| --duration 45s |
| |
| would re-run the previous test, but for only 45 seconds, thus facilitating |
| tracking down the aforementioned rare boot-time failure. |
| |
| |
| Distributed Runs |
| ================ |
| |
| Although kvm.sh is quite useful, its testing is confined to a single |
| system. It is not all that hard to use your favorite framework to cause |
| (say) 5 instances of kvm.sh to run on your 5 systems, but this will very |
| likely unnecessarily rebuild kernels. In addition, manually distributing |
| the desired rcutorture scenarios across the available systems can be |
| painstaking and error-prone. |
| |
| And this is why the kvm-remote.sh script exists. |
| |
| If you the following command works:: |
| |
| ssh system0 date |
| |
| and if it also works for system1, system2, system3, system4, and system5, |
| and all of these systems have 64 CPUs, you can type:: |
| |
| kvm-remote.sh "system0 system1 system2 system3 system4 system5" \ |
| --cpus 64 --duration 8h --configs "5*CFLIST" |
| |
| This will build each default scenario's kernel on the local system, then |
| spread each of five instances of each scenario over the systems listed, |
| running each scenario for eight hours. At the end of the runs, the |
| results will be gathered, recorded, and printed. Most of the parameters |
| that kvm.sh will accept can be passed to kvm-remote.sh, but the list of |
| systems must come first. |
| |
| The kvm.sh ``--dryrun scenarios`` argument is useful for working out |
| how many scenarios may be run in one batch across a group of systems. |
| |
| You can also re-run a previous remote run in a manner similar to kvm.sh: |
| |
| kvm-remote.sh "system0 system1 system2 system3 system4 system5" \ |
| tools/testing/selftests/rcutorture/res/2022.11.03-11.26.28-remote \ |
| --duration 24h |
| |
| In this case, most of the kvm-again.sh parameters may be supplied following |
| the pathname of the old run-results directory. |