Greg Kroah-Hartman | b244131 | 2017-11-01 15:07:57 +0100 | [diff] [blame] | 1 | /* SPDX-License-Identifier: GPL-2.0 */ |
Sam Ravnborg | a00736e | 2008-06-19 20:26:19 +0200 | [diff] [blame] | 2 | #ifndef _SPARC64_BACKOFF_H |
| 3 | #define _SPARC64_BACKOFF_H |
| 4 | |
David S. Miller | 187818c | 2012-10-28 13:04:47 -0700 | [diff] [blame] | 5 | /* The macros in this file implement an exponential backoff facility |
| 6 | * for atomic operations. |
| 7 | * |
| 8 | * When multiple threads compete on an atomic operation, it is |
| 9 | * possible for one thread to be continually denied a successful |
| 10 | * completion of the compare-and-swap instruction. Heavily |
| 11 | * threaded cpu implementations like Niagara can compound this |
| 12 | * problem even further. |
| 13 | * |
| 14 | * When an atomic operation fails and needs to be retried, we spin a |
| 15 | * certain number of times. At each subsequent failure of the same |
| 16 | * operation we double the spin count, realizing an exponential |
| 17 | * backoff. |
| 18 | * |
| 19 | * When we spin, we try to use an operation that will cause the |
| 20 | * current cpu strand to block, and therefore make the core fully |
| 21 | * available to any other other runnable strands. There are two |
| 22 | * options, based upon cpu capabilities. |
| 23 | * |
| 24 | * On all cpus prior to SPARC-T4 we do three dummy reads of the |
| 25 | * condition code register. Each read blocks the strand for something |
| 26 | * between 40 and 50 cpu cycles. |
| 27 | * |
| 28 | * For SPARC-T4 and later we have a special "pause" instruction |
| 29 | * available. This is implemented using writes to register %asr27. |
| 30 | * The cpu will block the number of cycles written into the register, |
| 31 | * unless a disrupting trap happens first. SPARC-T4 specifically |
| 32 | * implements pause with a granularity of 8 cycles. Each strand has |
| 33 | * an internal pause counter which decrements every 8 cycles. So the |
| 34 | * chip shifts the %asr27 value down by 3 bits, and writes the result |
| 35 | * into the pause counter. If a value smaller than 8 is written, the |
| 36 | * chip blocks for 1 cycle. |
| 37 | * |
| 38 | * To achieve the same amount of backoff as the three %ccr reads give |
| 39 | * on earlier chips, we shift the backoff value up by 7 bits. (Three |
| 40 | * %ccr reads block for about 128 cycles, 1 << 7 == 128) We write the |
| 41 | * whole amount we want to block into the pause register, rather than |
| 42 | * loop writing 128 each time. |
| 43 | */ |
| 44 | |
Sam Ravnborg | a00736e | 2008-06-19 20:26:19 +0200 | [diff] [blame] | 45 | #define BACKOFF_LIMIT (4 * 1024) |
| 46 | |
| 47 | #ifdef CONFIG_SMP |
| 48 | |
| 49 | #define BACKOFF_SETUP(reg) \ |
| 50 | mov 1, reg |
| 51 | |
David S. Miller | 0f58189 | 2010-08-18 22:53:26 -0700 | [diff] [blame] | 52 | #define BACKOFF_LABEL(spin_label, continue_label) \ |
| 53 | spin_label |
| 54 | |
David S. Miller | e9b9eb5 | 2012-10-27 23:00:41 -0700 | [diff] [blame] | 55 | #define BACKOFF_SPIN(reg, tmp, label) \ |
| 56 | mov reg, tmp; \ |
| 57 | 88: rd %ccr, %g0; \ |
| 58 | rd %ccr, %g0; \ |
| 59 | rd %ccr, %g0; \ |
David S. Miller | 187818c | 2012-10-28 13:04:47 -0700 | [diff] [blame] | 60 | .section .pause_3insn_patch,"ax";\ |
David S. Miller | e9b9eb5 | 2012-10-27 23:00:41 -0700 | [diff] [blame] | 61 | .word 88b; \ |
| 62 | sllx tmp, 7, tmp; \ |
| 63 | wr tmp, 0, %asr27; \ |
| 64 | clr tmp; \ |
| 65 | .previous; \ |
| 66 | brnz,pt tmp, 88b; \ |
| 67 | sub tmp, 1, tmp; \ |
| 68 | set BACKOFF_LIMIT, tmp; \ |
| 69 | cmp reg, tmp; \ |
| 70 | bg,pn %xcc, label; \ |
| 71 | nop; \ |
| 72 | ba,pt %xcc, label; \ |
| 73 | sllx reg, 1, reg; |
Sam Ravnborg | a00736e | 2008-06-19 20:26:19 +0200 | [diff] [blame] | 74 | |
| 75 | #else |
| 76 | |
| 77 | #define BACKOFF_SETUP(reg) |
David S. Miller | 0f58189 | 2010-08-18 22:53:26 -0700 | [diff] [blame] | 78 | |
| 79 | #define BACKOFF_LABEL(spin_label, continue_label) \ |
| 80 | continue_label |
| 81 | |
| 82 | #define BACKOFF_SPIN(reg, tmp, label) |
Sam Ravnborg | a00736e | 2008-06-19 20:26:19 +0200 | [diff] [blame] | 83 | |
| 84 | #endif |
| 85 | |
| 86 | #endif /* _SPARC64_BACKOFF_H */ |