blob: 8625946d8d00651d01c7275c99560dfc19e13195 [file] [log] [blame]
Greg Kroah-Hartmanb2441312017-11-01 15:07:57 +01001/* SPDX-License-Identifier: GPL-2.0 */
Sam Ravnborga00736e2008-06-19 20:26:19 +02002#ifndef _SPARC64_BACKOFF_H
3#define _SPARC64_BACKOFF_H
4
David S. Miller187818c2012-10-28 13:04:47 -07005/* The macros in this file implement an exponential backoff facility
6 * for atomic operations.
7 *
8 * When multiple threads compete on an atomic operation, it is
9 * possible for one thread to be continually denied a successful
10 * completion of the compare-and-swap instruction. Heavily
11 * threaded cpu implementations like Niagara can compound this
12 * problem even further.
13 *
14 * When an atomic operation fails and needs to be retried, we spin a
15 * certain number of times. At each subsequent failure of the same
16 * operation we double the spin count, realizing an exponential
17 * backoff.
18 *
19 * When we spin, we try to use an operation that will cause the
20 * current cpu strand to block, and therefore make the core fully
21 * available to any other other runnable strands. There are two
22 * options, based upon cpu capabilities.
23 *
24 * On all cpus prior to SPARC-T4 we do three dummy reads of the
25 * condition code register. Each read blocks the strand for something
26 * between 40 and 50 cpu cycles.
27 *
28 * For SPARC-T4 and later we have a special "pause" instruction
29 * available. This is implemented using writes to register %asr27.
30 * The cpu will block the number of cycles written into the register,
31 * unless a disrupting trap happens first. SPARC-T4 specifically
32 * implements pause with a granularity of 8 cycles. Each strand has
33 * an internal pause counter which decrements every 8 cycles. So the
34 * chip shifts the %asr27 value down by 3 bits, and writes the result
35 * into the pause counter. If a value smaller than 8 is written, the
36 * chip blocks for 1 cycle.
37 *
38 * To achieve the same amount of backoff as the three %ccr reads give
39 * on earlier chips, we shift the backoff value up by 7 bits. (Three
40 * %ccr reads block for about 128 cycles, 1 << 7 == 128) We write the
41 * whole amount we want to block into the pause register, rather than
42 * loop writing 128 each time.
43 */
44
Sam Ravnborga00736e2008-06-19 20:26:19 +020045#define BACKOFF_LIMIT (4 * 1024)
46
47#ifdef CONFIG_SMP
48
49#define BACKOFF_SETUP(reg) \
50 mov 1, reg
51
David S. Miller0f581892010-08-18 22:53:26 -070052#define BACKOFF_LABEL(spin_label, continue_label) \
53 spin_label
54
David S. Millere9b9eb52012-10-27 23:00:41 -070055#define BACKOFF_SPIN(reg, tmp, label) \
56 mov reg, tmp; \
5788: rd %ccr, %g0; \
58 rd %ccr, %g0; \
59 rd %ccr, %g0; \
David S. Miller187818c2012-10-28 13:04:47 -070060 .section .pause_3insn_patch,"ax";\
David S. Millere9b9eb52012-10-27 23:00:41 -070061 .word 88b; \
62 sllx tmp, 7, tmp; \
63 wr tmp, 0, %asr27; \
64 clr tmp; \
65 .previous; \
66 brnz,pt tmp, 88b; \
67 sub tmp, 1, tmp; \
68 set BACKOFF_LIMIT, tmp; \
69 cmp reg, tmp; \
70 bg,pn %xcc, label; \
71 nop; \
72 ba,pt %xcc, label; \
73 sllx reg, 1, reg;
Sam Ravnborga00736e2008-06-19 20:26:19 +020074
75#else
76
77#define BACKOFF_SETUP(reg)
David S. Miller0f581892010-08-18 22:53:26 -070078
79#define BACKOFF_LABEL(spin_label, continue_label) \
80 continue_label
81
82#define BACKOFF_SPIN(reg, tmp, label)
Sam Ravnborga00736e2008-06-19 20:26:19 +020083
84#endif
85
86#endif /* _SPARC64_BACKOFF_H */