Andrii Nakryiko | 929ffa6 | 2019-08-15 22:45:43 -0700 | [diff] [blame] | 1 | /* SPDX-License-Identifier: (LGPL-2.1 OR BSD-2-Clause) */ |
David S. Miller | bc1bafb | 2017-05-01 12:43:49 -0700 | [diff] [blame] | 2 | #ifndef __BPF_ENDIAN__ |
| 3 | #define __BPF_ENDIAN__ |
| 4 | |
Andrii Nakryiko | 30ad688 | 2020-06-30 08:21:24 -0700 | [diff] [blame] | 5 | /* |
| 6 | * Isolate byte #n and put it into byte #m, for __u##b type. |
| 7 | * E.g., moving byte #6 (nnnnnnnn) into byte #1 (mmmmmmmm) for __u64: |
| 8 | * 1) xxxxxxxx nnnnnnnn xxxxxxxx xxxxxxxx xxxxxxxx xxxxxxxx mmmmmmmm xxxxxxxx |
| 9 | * 2) nnnnnnnn xxxxxxxx xxxxxxxx xxxxxxxx xxxxxxxx mmmmmmmm xxxxxxxx 00000000 |
| 10 | * 3) 00000000 00000000 00000000 00000000 00000000 00000000 00000000 nnnnnnnn |
| 11 | * 4) 00000000 00000000 00000000 00000000 00000000 00000000 nnnnnnnn 00000000 |
| 12 | */ |
| 13 | #define ___bpf_mvb(x, b, n, m) ((__u##b)(x) << (b-(n+1)*8) >> (b-8) << (m*8)) |
| 14 | |
| 15 | #define ___bpf_swab16(x) ((__u16)( \ |
| 16 | ___bpf_mvb(x, 16, 0, 1) | \ |
| 17 | ___bpf_mvb(x, 16, 1, 0))) |
| 18 | |
| 19 | #define ___bpf_swab32(x) ((__u32)( \ |
| 20 | ___bpf_mvb(x, 32, 0, 3) | \ |
| 21 | ___bpf_mvb(x, 32, 1, 2) | \ |
| 22 | ___bpf_mvb(x, 32, 2, 1) | \ |
| 23 | ___bpf_mvb(x, 32, 3, 0))) |
| 24 | |
| 25 | #define ___bpf_swab64(x) ((__u64)( \ |
| 26 | ___bpf_mvb(x, 64, 0, 7) | \ |
| 27 | ___bpf_mvb(x, 64, 1, 6) | \ |
| 28 | ___bpf_mvb(x, 64, 2, 5) | \ |
| 29 | ___bpf_mvb(x, 64, 3, 4) | \ |
| 30 | ___bpf_mvb(x, 64, 4, 3) | \ |
| 31 | ___bpf_mvb(x, 64, 5, 2) | \ |
| 32 | ___bpf_mvb(x, 64, 6, 1) | \ |
| 33 | ___bpf_mvb(x, 64, 7, 0))) |
David S. Miller | bc1bafb | 2017-05-01 12:43:49 -0700 | [diff] [blame] | 34 | |
Daniel Borkmann | 78a5a93 | 2017-06-08 19:06:25 +0200 | [diff] [blame] | 35 | /* LLVM's BPF target selects the endianness of the CPU |
| 36 | * it compiles on, or the user specifies (bpfel/bpfeb), |
| 37 | * respectively. The used __BYTE_ORDER__ is defined by |
| 38 | * the compiler, we cannot rely on __BYTE_ORDER from |
| 39 | * libc headers, since it doesn't reflect the actual |
| 40 | * requested byte order. |
| 41 | * |
| 42 | * Note, LLVM's BPF target has different __builtin_bswapX() |
| 43 | * semantics. It does map to BPF_ALU | BPF_END | BPF_TO_BE |
| 44 | * in bpfel and bpfeb case, which means below, that we map |
| 45 | * to cpu_to_be16(). We could use it unconditionally in BPF |
| 46 | * case, but better not rely on it, so that this header here |
| 47 | * can be used from application and BPF program side, which |
| 48 | * use different targets. |
| 49 | */ |
| 50 | #if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__ |
| 51 | # define __bpf_ntohs(x) __builtin_bswap16(x) |
| 52 | # define __bpf_htons(x) __builtin_bswap16(x) |
Andrii Nakryiko | 30ad688 | 2020-06-30 08:21:24 -0700 | [diff] [blame] | 53 | # define __bpf_constant_ntohs(x) ___bpf_swab16(x) |
| 54 | # define __bpf_constant_htons(x) ___bpf_swab16(x) |
Yonghong Song | 5333502 | 2017-07-10 14:04:28 -0700 | [diff] [blame] | 55 | # define __bpf_ntohl(x) __builtin_bswap32(x) |
| 56 | # define __bpf_htonl(x) __builtin_bswap32(x) |
Andrii Nakryiko | 30ad688 | 2020-06-30 08:21:24 -0700 | [diff] [blame] | 57 | # define __bpf_constant_ntohl(x) ___bpf_swab32(x) |
| 58 | # define __bpf_constant_htonl(x) ___bpf_swab32(x) |
Ilya Leoshkevich | b963267 | 2019-08-30 13:07:29 +0200 | [diff] [blame] | 59 | # define __bpf_be64_to_cpu(x) __builtin_bswap64(x) |
| 60 | # define __bpf_cpu_to_be64(x) __builtin_bswap64(x) |
Andrii Nakryiko | 30ad688 | 2020-06-30 08:21:24 -0700 | [diff] [blame] | 61 | # define __bpf_constant_be64_to_cpu(x) ___bpf_swab64(x) |
| 62 | # define __bpf_constant_cpu_to_be64(x) ___bpf_swab64(x) |
Daniel Borkmann | 78a5a93 | 2017-06-08 19:06:25 +0200 | [diff] [blame] | 63 | #elif __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__ |
| 64 | # define __bpf_ntohs(x) (x) |
| 65 | # define __bpf_htons(x) (x) |
| 66 | # define __bpf_constant_ntohs(x) (x) |
| 67 | # define __bpf_constant_htons(x) (x) |
Yonghong Song | 5333502 | 2017-07-10 14:04:28 -0700 | [diff] [blame] | 68 | # define __bpf_ntohl(x) (x) |
| 69 | # define __bpf_htonl(x) (x) |
| 70 | # define __bpf_constant_ntohl(x) (x) |
| 71 | # define __bpf_constant_htonl(x) (x) |
Ilya Leoshkevich | b963267 | 2019-08-30 13:07:29 +0200 | [diff] [blame] | 72 | # define __bpf_be64_to_cpu(x) (x) |
| 73 | # define __bpf_cpu_to_be64(x) (x) |
| 74 | # define __bpf_constant_be64_to_cpu(x) (x) |
| 75 | # define __bpf_constant_cpu_to_be64(x) (x) |
David S. Miller | bc1bafb | 2017-05-01 12:43:49 -0700 | [diff] [blame] | 76 | #else |
Daniel Borkmann | 78a5a93 | 2017-06-08 19:06:25 +0200 | [diff] [blame] | 77 | # error "Fix your compiler's __BYTE_ORDER__?!" |
David S. Miller | bc1bafb | 2017-05-01 12:43:49 -0700 | [diff] [blame] | 78 | #endif |
| 79 | |
| 80 | #define bpf_htons(x) \ |
| 81 | (__builtin_constant_p(x) ? \ |
Daniel Borkmann | 78a5a93 | 2017-06-08 19:06:25 +0200 | [diff] [blame] | 82 | __bpf_constant_htons(x) : __bpf_htons(x)) |
David S. Miller | bc1bafb | 2017-05-01 12:43:49 -0700 | [diff] [blame] | 83 | #define bpf_ntohs(x) \ |
| 84 | (__builtin_constant_p(x) ? \ |
Daniel Borkmann | 78a5a93 | 2017-06-08 19:06:25 +0200 | [diff] [blame] | 85 | __bpf_constant_ntohs(x) : __bpf_ntohs(x)) |
Yonghong Song | 5333502 | 2017-07-10 14:04:28 -0700 | [diff] [blame] | 86 | #define bpf_htonl(x) \ |
| 87 | (__builtin_constant_p(x) ? \ |
| 88 | __bpf_constant_htonl(x) : __bpf_htonl(x)) |
| 89 | #define bpf_ntohl(x) \ |
| 90 | (__builtin_constant_p(x) ? \ |
| 91 | __bpf_constant_ntohl(x) : __bpf_ntohl(x)) |
Ilya Leoshkevich | b963267 | 2019-08-30 13:07:29 +0200 | [diff] [blame] | 92 | #define bpf_cpu_to_be64(x) \ |
| 93 | (__builtin_constant_p(x) ? \ |
| 94 | __bpf_constant_cpu_to_be64(x) : __bpf_cpu_to_be64(x)) |
| 95 | #define bpf_be64_to_cpu(x) \ |
| 96 | (__builtin_constant_p(x) ? \ |
| 97 | __bpf_constant_be64_to_cpu(x) : __bpf_be64_to_cpu(x)) |
David S. Miller | bc1bafb | 2017-05-01 12:43:49 -0700 | [diff] [blame] | 98 | |
Daniel Borkmann | 78a5a93 | 2017-06-08 19:06:25 +0200 | [diff] [blame] | 99 | #endif /* __BPF_ENDIAN__ */ |