intel_pstate: Remove C0 tracking
[linux/fpc-iii.git] / samples / seccomp / bpf-helper.h
blob38ee70f3cd5b970101edd6d2edb52b4df8d73d14
1 /*
2 * Example wrapper around BPF macros.
4 * Copyright (c) 2012 The Chromium OS Authors <chromium-os-dev@chromium.org>
5 * Author: Will Drewry <wad@chromium.org>
7 * The code may be used by anyone for any purpose,
8 * and can serve as a starting point for developing
9 * applications using prctl(PR_SET_SECCOMP, 2, ...).
11 * No guarantees are provided with respect to the correctness
12 * or functionality of this code.
14 #ifndef __BPF_HELPER_H__
15 #define __BPF_HELPER_H__
17 #include <asm/bitsperlong.h> /* for __BITS_PER_LONG */
18 #include <endian.h>
19 #include <linux/filter.h>
20 #include <linux/seccomp.h> /* for seccomp_data */
21 #include <linux/types.h>
22 #include <linux/unistd.h>
23 #include <stddef.h>
25 #define BPF_LABELS_MAX 256
26 struct bpf_labels {
27 int count;
28 struct __bpf_label {
29 const char *label;
30 __u32 location;
31 } labels[BPF_LABELS_MAX];
34 int bpf_resolve_jumps(struct bpf_labels *labels,
35 struct sock_filter *filter, size_t count);
36 __u32 seccomp_bpf_label(struct bpf_labels *labels, const char *label);
37 void seccomp_bpf_print(struct sock_filter *filter, size_t count);
39 #define JUMP_JT 0xff
40 #define JUMP_JF 0xff
41 #define LABEL_JT 0xfe
42 #define LABEL_JF 0xfe
44 #define ALLOW \
45 BPF_STMT(BPF_RET+BPF_K, SECCOMP_RET_ALLOW)
46 #define DENY \
47 BPF_STMT(BPF_RET+BPF_K, SECCOMP_RET_KILL)
48 #define JUMP(labels, label) \
49 BPF_JUMP(BPF_JMP+BPF_JA, FIND_LABEL((labels), (label)), \
50 JUMP_JT, JUMP_JF)
51 #define LABEL(labels, label) \
52 BPF_JUMP(BPF_JMP+BPF_JA, FIND_LABEL((labels), (label)), \
53 LABEL_JT, LABEL_JF)
54 #define SYSCALL(nr, jt) \
55 BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (nr), 0, 1), \
58 /* Lame, but just an example */
59 #define FIND_LABEL(labels, label) seccomp_bpf_label((labels), #label)
61 #define EXPAND(...) __VA_ARGS__
63 /* Ensure that we load the logically correct offset. */
64 #if __BYTE_ORDER == __LITTLE_ENDIAN
65 #define LO_ARG(idx) offsetof(struct seccomp_data, args[(idx)])
66 #elif __BYTE_ORDER == __BIG_ENDIAN
67 #define LO_ARG(idx) offsetof(struct seccomp_data, args[(idx)]) + sizeof(__u32)
68 #else
69 #error "Unknown endianness"
70 #endif
72 /* Map all width-sensitive operations */
73 #if __BITS_PER_LONG == 32
75 #define JEQ(x, jt) JEQ32(x, EXPAND(jt))
76 #define JNE(x, jt) JNE32(x, EXPAND(jt))
77 #define JGT(x, jt) JGT32(x, EXPAND(jt))
78 #define JLT(x, jt) JLT32(x, EXPAND(jt))
79 #define JGE(x, jt) JGE32(x, EXPAND(jt))
80 #define JLE(x, jt) JLE32(x, EXPAND(jt))
81 #define JA(x, jt) JA32(x, EXPAND(jt))
82 #define ARG(i) ARG_32(i)
84 #elif __BITS_PER_LONG == 64
86 /* Ensure that we load the logically correct offset. */
87 #if __BYTE_ORDER == __LITTLE_ENDIAN
88 #define ENDIAN(_lo, _hi) _lo, _hi
89 #define HI_ARG(idx) offsetof(struct seccomp_data, args[(idx)]) + sizeof(__u32)
90 #elif __BYTE_ORDER == __BIG_ENDIAN
91 #define ENDIAN(_lo, _hi) _hi, _lo
92 #define HI_ARG(idx) offsetof(struct seccomp_data, args[(idx)])
93 #endif
95 union arg64 {
96 struct {
97 __u32 ENDIAN(lo32, hi32);
99 __u64 u64;
102 #define JEQ(x, jt) \
103 JEQ64(((union arg64){.u64 = (x)}).lo32, \
104 ((union arg64){.u64 = (x)}).hi32, \
105 EXPAND(jt))
106 #define JGT(x, jt) \
107 JGT64(((union arg64){.u64 = (x)}).lo32, \
108 ((union arg64){.u64 = (x)}).hi32, \
109 EXPAND(jt))
110 #define JGE(x, jt) \
111 JGE64(((union arg64){.u64 = (x)}).lo32, \
112 ((union arg64){.u64 = (x)}).hi32, \
113 EXPAND(jt))
114 #define JNE(x, jt) \
115 JNE64(((union arg64){.u64 = (x)}).lo32, \
116 ((union arg64){.u64 = (x)}).hi32, \
117 EXPAND(jt))
118 #define JLT(x, jt) \
119 JLT64(((union arg64){.u64 = (x)}).lo32, \
120 ((union arg64){.u64 = (x)}).hi32, \
121 EXPAND(jt))
122 #define JLE(x, jt) \
123 JLE64(((union arg64){.u64 = (x)}).lo32, \
124 ((union arg64){.u64 = (x)}).hi32, \
125 EXPAND(jt))
127 #define JA(x, jt) \
128 JA64(((union arg64){.u64 = (x)}).lo32, \
129 ((union arg64){.u64 = (x)}).hi32, \
130 EXPAND(jt))
131 #define ARG(i) ARG_64(i)
133 #else
134 #error __BITS_PER_LONG value unusable.
135 #endif
137 /* Loads the arg into A */
138 #define ARG_32(idx) \
139 BPF_STMT(BPF_LD+BPF_W+BPF_ABS, LO_ARG(idx))
141 /* Loads hi into A and lo in X */
142 #define ARG_64(idx) \
143 BPF_STMT(BPF_LD+BPF_W+BPF_ABS, LO_ARG(idx)), \
144 BPF_STMT(BPF_ST, 0), /* lo -> M[0] */ \
145 BPF_STMT(BPF_LD+BPF_W+BPF_ABS, HI_ARG(idx)), \
146 BPF_STMT(BPF_ST, 1) /* hi -> M[1] */
148 #define JEQ32(value, jt) \
149 BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (value), 0, 1), \
152 #define JNE32(value, jt) \
153 BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (value), 1, 0), \
156 /* Checks the lo, then swaps to check the hi. A=lo,X=hi */
157 #define JEQ64(lo, hi, jt) \
158 BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (hi), 0, 5), \
159 BPF_STMT(BPF_LD+BPF_MEM, 0), /* swap in lo */ \
160 BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (lo), 0, 2), \
161 BPF_STMT(BPF_LD+BPF_MEM, 1), /* passed: swap hi back in */ \
162 jt, \
163 BPF_STMT(BPF_LD+BPF_MEM, 1) /* failed: swap hi back in */
165 #define JNE64(lo, hi, jt) \
166 BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (hi), 5, 0), \
167 BPF_STMT(BPF_LD+BPF_MEM, 0), /* swap in lo */ \
168 BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (lo), 2, 0), \
169 BPF_STMT(BPF_LD+BPF_MEM, 1), /* passed: swap hi back in */ \
170 jt, \
171 BPF_STMT(BPF_LD+BPF_MEM, 1) /* failed: swap hi back in */
173 #define JA32(value, jt) \
174 BPF_JUMP(BPF_JMP+BPF_JSET+BPF_K, (value), 0, 1), \
177 #define JA64(lo, hi, jt) \
178 BPF_JUMP(BPF_JMP+BPF_JSET+BPF_K, (hi), 3, 0), \
179 BPF_STMT(BPF_LD+BPF_MEM, 0), /* swap in lo */ \
180 BPF_JUMP(BPF_JMP+BPF_JSET+BPF_K, (lo), 0, 2), \
181 BPF_STMT(BPF_LD+BPF_MEM, 1), /* passed: swap hi back in */ \
182 jt, \
183 BPF_STMT(BPF_LD+BPF_MEM, 1) /* failed: swap hi back in */
185 #define JGE32(value, jt) \
186 BPF_JUMP(BPF_JMP+BPF_JGE+BPF_K, (value), 0, 1), \
189 #define JLT32(value, jt) \
190 BPF_JUMP(BPF_JMP+BPF_JGE+BPF_K, (value), 1, 0), \
193 /* Shortcut checking if hi > arg.hi. */
194 #define JGE64(lo, hi, jt) \
195 BPF_JUMP(BPF_JMP+BPF_JGT+BPF_K, (hi), 4, 0), \
196 BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (hi), 0, 5), \
197 BPF_STMT(BPF_LD+BPF_MEM, 0), /* swap in lo */ \
198 BPF_JUMP(BPF_JMP+BPF_JGE+BPF_K, (lo), 0, 2), \
199 BPF_STMT(BPF_LD+BPF_MEM, 1), /* passed: swap hi back in */ \
200 jt, \
201 BPF_STMT(BPF_LD+BPF_MEM, 1) /* failed: swap hi back in */
203 #define JLT64(lo, hi, jt) \
204 BPF_JUMP(BPF_JMP+BPF_JGE+BPF_K, (hi), 0, 4), \
205 BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (hi), 0, 5), \
206 BPF_STMT(BPF_LD+BPF_MEM, 0), /* swap in lo */ \
207 BPF_JUMP(BPF_JMP+BPF_JGT+BPF_K, (lo), 2, 0), \
208 BPF_STMT(BPF_LD+BPF_MEM, 1), /* passed: swap hi back in */ \
209 jt, \
210 BPF_STMT(BPF_LD+BPF_MEM, 1) /* failed: swap hi back in */
212 #define JGT32(value, jt) \
213 BPF_JUMP(BPF_JMP+BPF_JGT+BPF_K, (value), 0, 1), \
216 #define JLE32(value, jt) \
217 BPF_JUMP(BPF_JMP+BPF_JGT+BPF_K, (value), 1, 0), \
220 /* Check hi > args.hi first, then do the GE checking */
221 #define JGT64(lo, hi, jt) \
222 BPF_JUMP(BPF_JMP+BPF_JGT+BPF_K, (hi), 4, 0), \
223 BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (hi), 0, 5), \
224 BPF_STMT(BPF_LD+BPF_MEM, 0), /* swap in lo */ \
225 BPF_JUMP(BPF_JMP+BPF_JGT+BPF_K, (lo), 0, 2), \
226 BPF_STMT(BPF_LD+BPF_MEM, 1), /* passed: swap hi back in */ \
227 jt, \
228 BPF_STMT(BPF_LD+BPF_MEM, 1) /* failed: swap hi back in */
230 #define JLE64(lo, hi, jt) \
231 BPF_JUMP(BPF_JMP+BPF_JGT+BPF_K, (hi), 6, 0), \
232 BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (hi), 0, 3), \
233 BPF_STMT(BPF_LD+BPF_MEM, 0), /* swap in lo */ \
234 BPF_JUMP(BPF_JMP+BPF_JGT+BPF_K, (lo), 2, 0), \
235 BPF_STMT(BPF_LD+BPF_MEM, 1), /* passed: swap hi back in */ \
236 jt, \
237 BPF_STMT(BPF_LD+BPF_MEM, 1) /* failed: swap hi back in */
239 #define LOAD_SYSCALL_NR \
240 BPF_STMT(BPF_LD+BPF_W+BPF_ABS, \
241 offsetof(struct seccomp_data, nr))
243 #endif /* __BPF_HELPER_H__ */