1 /* Copyright (c) 2016 Facebook
2  *
3  * This program is free software; you can redistribute it and/or
4  * modify it under the terms of version 2 of the GNU General Public
5  * License as published by the Free Software Foundation.
6  */
7 #include <linux/skbuff.h>
8 #include <linux/netdevice.h>
9 #include <linux/version.h>
10 #include <uapi/linux/bpf.h>
11 #include <bpf/bpf_helpers.h>
12 #include <bpf/bpf_tracing.h>
13 #include <bpf/bpf_core_read.h>
14 #include "trace_common.h"
15 
16 #define MAX_ENTRIES 1000
17 #define MAX_NR_CPUS 1024
18 
19 struct {
20 	__uint(type, BPF_MAP_TYPE_HASH);
21 	__type(key, u32);
22 	__type(value, long);
23 	__uint(max_entries, MAX_ENTRIES);
24 } hash_map SEC(".maps");
25 
26 struct {
27 	__uint(type, BPF_MAP_TYPE_LRU_HASH);
28 	__type(key, u32);
29 	__type(value, long);
30 	__uint(max_entries, 10000);
31 } lru_hash_map SEC(".maps");
32 
33 struct {
34 	__uint(type, BPF_MAP_TYPE_LRU_HASH);
35 	__type(key, u32);
36 	__type(value, long);
37 	__uint(max_entries, 10000);
38 	__uint(map_flags, BPF_F_NO_COMMON_LRU);
39 } nocommon_lru_hash_map SEC(".maps");
40 
41 struct inner_lru {
42 	__uint(type, BPF_MAP_TYPE_LRU_HASH);
43 	__type(key, u32);
44 	__type(value, long);
45 	__uint(max_entries, MAX_ENTRIES);
46 	__uint(map_flags, BPF_F_NUMA_NODE);
47 	__uint(numa_node, 0);
48 } inner_lru_hash_map SEC(".maps");
49 
50 struct {
51 	__uint(type, BPF_MAP_TYPE_ARRAY_OF_MAPS);
52 	__uint(max_entries, MAX_NR_CPUS);
53 	__uint(key_size, sizeof(u32));
54 	__array(values, struct inner_lru); /* use inner_lru as inner map */
55 } array_of_lru_hashs SEC(".maps") = {
56 	/* statically initialize the first element */
57 	.values = { &inner_lru_hash_map },
58 };
59 
60 struct {
61 	__uint(type, BPF_MAP_TYPE_PERCPU_HASH);
62 	__uint(key_size, sizeof(u32));
63 	__uint(value_size, sizeof(long));
64 	__uint(max_entries, MAX_ENTRIES);
65 } percpu_hash_map SEC(".maps");
66 
67 struct {
68 	__uint(type, BPF_MAP_TYPE_HASH);
69 	__type(key, u32);
70 	__type(value, long);
71 	__uint(max_entries, MAX_ENTRIES);
72 	__uint(map_flags, BPF_F_NO_PREALLOC);
73 } hash_map_alloc SEC(".maps");
74 
75 struct {
76 	__uint(type, BPF_MAP_TYPE_PERCPU_HASH);
77 	__uint(key_size, sizeof(u32));
78 	__uint(value_size, sizeof(long));
79 	__uint(max_entries, MAX_ENTRIES);
80 	__uint(map_flags, BPF_F_NO_PREALLOC);
81 } percpu_hash_map_alloc SEC(".maps");
82 
83 struct {
84 	__uint(type, BPF_MAP_TYPE_LPM_TRIE);
85 	__uint(key_size, 8);
86 	__uint(value_size, sizeof(long));
87 	__uint(max_entries, 10000);
88 	__uint(map_flags, BPF_F_NO_PREALLOC);
89 } lpm_trie_map_alloc SEC(".maps");
90 
91 struct {
92 	__uint(type, BPF_MAP_TYPE_ARRAY);
93 	__type(key, u32);
94 	__type(value, long);
95 	__uint(max_entries, MAX_ENTRIES);
96 } array_map SEC(".maps");
97 
98 struct {
99 	__uint(type, BPF_MAP_TYPE_LRU_HASH);
100 	__type(key, u32);
101 	__type(value, long);
102 	__uint(max_entries, MAX_ENTRIES);
103 } lru_hash_lookup_map SEC(".maps");
104 
SYSCALL(sys_getuid)105 SEC("kprobe/" SYSCALL(sys_getuid))
106 int stress_hmap(struct pt_regs *ctx)
107 {
108 	u32 key = bpf_get_current_pid_tgid();
109 	long init_val = 1;
110 	long *value;
111 
112 	bpf_map_update_elem(&hash_map, &key, &init_val, BPF_ANY);
113 	value = bpf_map_lookup_elem(&hash_map, &key);
114 	if (value)
115 		bpf_map_delete_elem(&hash_map, &key);
116 
117 	return 0;
118 }
119 
SYSCALL(sys_geteuid)120 SEC("kprobe/" SYSCALL(sys_geteuid))
121 int stress_percpu_hmap(struct pt_regs *ctx)
122 {
123 	u32 key = bpf_get_current_pid_tgid();
124 	long init_val = 1;
125 	long *value;
126 
127 	bpf_map_update_elem(&percpu_hash_map, &key, &init_val, BPF_ANY);
128 	value = bpf_map_lookup_elem(&percpu_hash_map, &key);
129 	if (value)
130 		bpf_map_delete_elem(&percpu_hash_map, &key);
131 	return 0;
132 }
133 
SYSCALL(sys_getgid)134 SEC("kprobe/" SYSCALL(sys_getgid))
135 int stress_hmap_alloc(struct pt_regs *ctx)
136 {
137 	u32 key = bpf_get_current_pid_tgid();
138 	long init_val = 1;
139 	long *value;
140 
141 	bpf_map_update_elem(&hash_map_alloc, &key, &init_val, BPF_ANY);
142 	value = bpf_map_lookup_elem(&hash_map_alloc, &key);
143 	if (value)
144 		bpf_map_delete_elem(&hash_map_alloc, &key);
145 	return 0;
146 }
147 
SYSCALL(sys_getegid)148 SEC("kprobe/" SYSCALL(sys_getegid))
149 int stress_percpu_hmap_alloc(struct pt_regs *ctx)
150 {
151 	u32 key = bpf_get_current_pid_tgid();
152 	long init_val = 1;
153 	long *value;
154 
155 	bpf_map_update_elem(&percpu_hash_map_alloc, &key, &init_val, BPF_ANY);
156 	value = bpf_map_lookup_elem(&percpu_hash_map_alloc, &key);
157 	if (value)
158 		bpf_map_delete_elem(&percpu_hash_map_alloc, &key);
159 	return 0;
160 }
161 
SYSCALL(sys_connect)162 SEC("kprobe/" SYSCALL(sys_connect))
163 int stress_lru_hmap_alloc(struct pt_regs *ctx)
164 {
165 	struct pt_regs *real_regs = (struct pt_regs *)PT_REGS_PARM1_CORE(ctx);
166 	char fmt[] = "Failed at stress_lru_hmap_alloc. ret:%dn";
167 	union {
168 		u16 dst6[8];
169 		struct {
170 			u16 magic0;
171 			u16 magic1;
172 			u16 tcase;
173 			u16 unused16;
174 			u32 unused32;
175 			u32 key;
176 		};
177 	} test_params;
178 	struct sockaddr_in6 *in6;
179 	u16 test_case;
180 	int addrlen, ret;
181 	long val = 1;
182 	u32 key = 0;
183 
184 	in6 = (struct sockaddr_in6 *)PT_REGS_PARM2_CORE(real_regs);
185 	addrlen = (int)PT_REGS_PARM3_CORE(real_regs);
186 
187 	if (addrlen != sizeof(*in6))
188 		return 0;
189 
190 	ret = bpf_probe_read_user(test_params.dst6, sizeof(test_params.dst6),
191 				  &in6->sin6_addr);
192 	if (ret)
193 		goto done;
194 
195 	if (test_params.magic0 != 0xdead ||
196 	    test_params.magic1 != 0xbeef)
197 		return 0;
198 
199 	test_case = test_params.tcase;
200 	if (test_case != 3)
201 		key = bpf_get_prandom_u32();
202 
203 	if (test_case == 0) {
204 		ret = bpf_map_update_elem(&lru_hash_map, &key, &val, BPF_ANY);
205 	} else if (test_case == 1) {
206 		ret = bpf_map_update_elem(&nocommon_lru_hash_map, &key, &val,
207 					  BPF_ANY);
208 	} else if (test_case == 2) {
209 		void *nolocal_lru_map;
210 		int cpu = bpf_get_smp_processor_id();
211 
212 		nolocal_lru_map = bpf_map_lookup_elem(&array_of_lru_hashs,
213 						      &cpu);
214 		if (!nolocal_lru_map) {
215 			ret = -ENOENT;
216 			goto done;
217 		}
218 
219 		ret = bpf_map_update_elem(nolocal_lru_map, &key, &val,
220 					  BPF_ANY);
221 	} else if (test_case == 3) {
222 		u32 i;
223 
224 		key = test_params.key;
225 
226 #pragma clang loop unroll(full)
227 		for (i = 0; i < 32; i++) {
228 			bpf_map_lookup_elem(&lru_hash_lookup_map, &key);
229 			key++;
230 		}
231 	} else {
232 		ret = -EINVAL;
233 	}
234 
235 done:
236 	if (ret)
237 		bpf_trace_printk(fmt, sizeof(fmt), ret);
238 
239 	return 0;
240 }
241 
SYSCALL(sys_gettid)242 SEC("kprobe/" SYSCALL(sys_gettid))
243 int stress_lpm_trie_map_alloc(struct pt_regs *ctx)
244 {
245 	union {
246 		u32 b32[2];
247 		u8 b8[8];
248 	} key;
249 	unsigned int i;
250 
251 	key.b32[0] = 32;
252 	key.b8[4] = 192;
253 	key.b8[5] = 168;
254 	key.b8[6] = 0;
255 	key.b8[7] = 1;
256 
257 #pragma clang loop unroll(full)
258 	for (i = 0; i < 32; ++i)
259 		bpf_map_lookup_elem(&lpm_trie_map_alloc, &key);
260 
261 	return 0;
262 }
263 
SYSCALL(sys_getpgid)264 SEC("kprobe/" SYSCALL(sys_getpgid))
265 int stress_hash_map_lookup(struct pt_regs *ctx)
266 {
267 	u32 key = 1, i;
268 	long *value;
269 
270 #pragma clang loop unroll(full)
271 	for (i = 0; i < 64; ++i)
272 		value = bpf_map_lookup_elem(&hash_map, &key);
273 
274 	return 0;
275 }
276 
SYSCALL(sys_getppid)277 SEC("kprobe/" SYSCALL(sys_getppid))
278 int stress_array_map_lookup(struct pt_regs *ctx)
279 {
280 	u32 key = 1, i;
281 	long *value;
282 
283 #pragma clang loop unroll(full)
284 	for (i = 0; i < 64; ++i)
285 		value = bpf_map_lookup_elem(&array_map, &key);
286 
287 	return 0;
288 }
289 
290 char _license[] SEC("license") = "GPL";
291 u32 _version SEC("version") = LINUX_VERSION_CODE;
292