/linux-6.1.9/samples/bpf/ |
D | test_map_in_map_user.c | 17 static int map_fd[7]; variable 19 #define PORT_A (map_fd[0]) 20 #define PORT_H (map_fd[1]) 21 #define REG_RESULT_H (map_fd[2]) 22 #define INLINE_RESULT_H (map_fd[3]) 23 #define A_OF_PORT_A (map_fd[4]) /* Test case #0 */ 24 #define H_OF_PORT_A (map_fd[5]) /* Test case #1 */ 25 #define H_OF_PORT_H (map_fd[6]) /* Test case #2 */ 142 map_fd[0] = bpf_object__find_map_fd_by_name(obj, "port_a"); in main() 143 map_fd[1] = bpf_object__find_map_fd_by_name(obj, "port_h"); in main() [all …]
|
D | test_lru_dist.c | 101 int map_fd; member 107 lru->map_fd = bpf_map_create(BPF_MAP_TYPE_HASH, NULL, in pfect_lru_init() 111 assert(lru->map_fd != -1); in pfect_lru_init() 124 close(lru->map_fd); in pfect_lru_destroy() 135 if (!bpf_map_lookup_elem(lru->map_fd, &key, &node)) { in pfect_lru_lookup_or_insert() 152 bpf_map_update_elem(lru->map_fd, &node->key, &null_node, BPF_EXIST); in pfect_lru_lookup_or_insert() 160 assert(!bpf_map_update_elem(lru->map_fd, &key, &node, BPF_EXIST)); in pfect_lru_lookup_or_insert() 163 assert(!bpf_map_update_elem(lru->map_fd, &key, &node, BPF_NOEXIST)); in pfect_lru_lookup_or_insert() 212 int map_fd; in create_map() local 214 map_fd = bpf_map_create(map_type, NULL, sizeof(unsigned long long), in create_map() [all …]
|
D | test_current_task_under_cgroup_user.c | 22 int map_fd[2]; in main() local 43 map_fd[0] = bpf_object__find_map_fd_by_name(obj, "cgroup_map"); in main() 44 map_fd[1] = bpf_object__find_map_fd_by_name(obj, "perf_map"); in main() 45 if (map_fd[0] < 0 || map_fd[1] < 0) { in main() 65 if (bpf_map_update_elem(map_fd[0], &idx, &cg2, BPF_ANY)) { in main() 79 bpf_map_lookup_elem(map_fd[1], &idx, &remote_pid); in main() 93 bpf_map_update_elem(map_fd[1], &idx, &remote_pid, BPF_ANY); in main() 96 bpf_map_lookup_elem(map_fd[1], &idx, &remote_pid); in main()
|
D | tracex6_user.c | 23 static int map_fd[3]; variable 43 assert(bpf_map_update_elem(map_fd[0], &cpu, &pmu_fd, BPF_ANY) == 0); in check_on_cpu() 46 bpf_map_get_next_key(map_fd[1], &cpu, NULL); in check_on_cpu() 48 if (bpf_map_lookup_elem(map_fd[1], &cpu, &value)) { in check_on_cpu() 56 if (bpf_map_lookup_elem(map_fd[2], &cpu, &value2)) { in check_on_cpu() 66 assert(bpf_map_delete_elem(map_fd[0], &cpu) == 0 || error); in check_on_cpu() 69 assert(bpf_map_delete_elem(map_fd[1], &cpu) == 0 || error); in check_on_cpu() 196 map_fd[0] = bpf_object__find_map_fd_by_name(obj, "counters"); in main() 197 map_fd[1] = bpf_object__find_map_fd_by_name(obj, "values"); in main() 198 map_fd[2] = bpf_object__find_map_fd_by_name(obj, "values2"); in main() [all …]
|
D | offwaketime_user.c | 18 static int map_fd[2]; variable 54 if (bpf_map_lookup_elem(map_fd[1], &key->tret, ip) != 0) { in print_stack() 61 if (bpf_map_lookup_elem(map_fd[1], &key->wret, ip) != 0) { in print_stack() 91 print_stacks(map_fd[0]); in int_exit() 122 map_fd[0] = bpf_object__find_map_fd_by_name(obj, "counts"); in main() 123 map_fd[1] = bpf_object__find_map_fd_by_name(obj, "stackmap"); in main() 124 if (map_fd[0] < 0 || map_fd[1] < 0) { in main() 145 print_stacks(map_fd[0]); in main()
|
D | xdp_adjust_tail_user.c | 52 static void poll_stats(unsigned int map_fd, unsigned int kill_after_s) in poll_stats() argument 62 assert(bpf_map_lookup_elem(map_fd, &key, &value) == 0); in poll_stats() 89 int i, prog_fd, map_fd, opt; in main() local 164 map_fd = bpf_object__find_map_fd_by_name(obj, "xdp_adju.data"); in main() 165 if (map_fd < 0) { in main() 169 bpf_map_update_elem(map_fd, &key, &max_pckt_size, BPF_ANY); in main() 173 map_fd = bpf_object__find_map_fd_by_name(obj, "icmpcnt"); in main() 174 if (map_fd < 0) { in main() 194 poll_stats(map_fd, kill_after_s); in main()
|
D | test_cgrp2_attach.c | 43 static int prog_load(int map_fd, int verdict) in prog_load() argument 53 BPF_LD_MAP_FD(BPF_REG_1, map_fd), /* load map fd to r1 */ in prog_load() 64 BPF_LD_MAP_FD(BPF_REG_1, map_fd), in prog_load() 94 int prog_fd, map_fd, ret, key; in attach_filter() local 97 map_fd = bpf_map_create(BPF_MAP_TYPE_ARRAY, NULL, in attach_filter() 100 if (map_fd < 0) { in attach_filter() 105 prog_fd = prog_load(map_fd, verdict); in attach_filter() 121 assert(bpf_map_lookup_elem(map_fd, &key, &pkt_cnt) == 0); in attach_filter() 124 assert(bpf_map_lookup_elem(map_fd, &key, &byte_cnt) == 0); in attach_filter()
|
D | spintest_user.c | 17 int map_fd, i, j = 0; in main() local 40 map_fd = bpf_object__find_map_fd_by_name(obj, "my_map"); in main() 41 if (map_fd < 0) { in main() 66 while (bpf_map_get_next_key(map_fd, &key, &next_key) == 0) { in main() 67 bpf_map_lookup_elem(map_fd, &next_key, &value); in main() 81 while (bpf_map_get_next_key(map_fd, &key, &next_key) == 0) in main() 82 bpf_map_delete_elem(map_fd, &next_key); in main()
|
/linux-6.1.9/tools/testing/selftests/bpf/prog_tests/ |
D | lookup_and_delete.c | 13 static int fill_values(int map_fd) in fill_values() argument 19 err = bpf_map_update_elem(map_fd, &key, &value, BPF_NOEXIST); in fill_values() 27 static int fill_values_percpu(int map_fd) in fill_values_percpu() argument 36 err = bpf_map_update_elem(map_fd, &key, value, BPF_NOEXIST); in fill_values_percpu() 45 int *map_fd) in setup_prog() argument 66 *map_fd = bpf_map__fd(skel->maps.hash_map); in setup_prog() 67 if (!ASSERT_GE(*map_fd, 0, "bpf_map__fd")) in setup_prog() 102 int map_fd, err; in test_lookup_and_delete_hash() local 105 skel = setup_prog(BPF_MAP_TYPE_HASH, &map_fd); in test_lookup_and_delete_hash() 109 err = fill_values(map_fd); in test_lookup_and_delete_hash() [all …]
|
D | tailcalls.c | 11 int err, map_fd, prog_fd, main_fd, i, j; in test_tailcall_1() local 40 map_fd = bpf_map__fd(prog_array); in test_tailcall_1() 41 if (CHECK_FAIL(map_fd < 0)) in test_tailcall_1() 55 err = bpf_map_update_elem(map_fd, &i, &prog_fd, BPF_ANY); in test_tailcall_1() 65 err = bpf_map_delete_elem(map_fd, &i); in test_tailcall_1() 85 err = bpf_map_update_elem(map_fd, &i, &prog_fd, BPF_ANY); in test_tailcall_1() 106 err = bpf_map_update_elem(map_fd, &i, &prog_fd, BPF_ANY); in test_tailcall_1() 118 err = bpf_map_delete_elem(map_fd, &i); in test_tailcall_1() 128 err = bpf_map_delete_elem(map_fd, &i); in test_tailcall_1() 147 int err, map_fd, prog_fd, main_fd, i; in test_tailcall_2() local [all …]
|
D | l4lb_all.c | 26 int err, i, prog_fd, map_fd; in test_l4lb() local 41 map_fd = bpf_find_map(__func__, obj, "vip_map"); in test_l4lb() 42 if (map_fd < 0) in test_l4lb() 44 bpf_map_update_elem(map_fd, &key, &value, 0); in test_l4lb() 46 map_fd = bpf_find_map(__func__, obj, "ch_rings"); in test_l4lb() 47 if (map_fd < 0) in test_l4lb() 49 bpf_map_update_elem(map_fd, &ch_key, &real_num, 0); in test_l4lb() 51 map_fd = bpf_find_map(__func__, obj, "reals"); in test_l4lb() 52 if (map_fd < 0) in test_l4lb() 54 bpf_map_update_elem(map_fd, &real_num, &real_def, 0); in test_l4lb() [all …]
|
D | global_data.c | 7 int i, err, map_fd; in test_global_data_number() local 10 map_fd = bpf_find_map(__func__, obj, "result_number"); in test_global_data_number() 11 if (CHECK_FAIL(map_fd < 0)) in test_global_data_number() 33 err = bpf_map_lookup_elem(map_fd, &tests[i].key, &num); in test_global_data_number() 42 int i, err, map_fd; in test_global_data_string() local 45 map_fd = bpf_find_map(__func__, obj, "result_string"); in test_global_data_string() 46 if (CHECK_FAIL(map_fd < 0)) in test_global_data_string() 62 err = bpf_map_lookup_elem(map_fd, &tests[i].key, str); in test_global_data_string() 77 int i, err, map_fd; in test_global_data_struct() local 80 map_fd = bpf_find_map(__func__, obj, "result_struct"); in test_global_data_struct() [all …]
|
D | map_init.c | 20 static int map_populate(int map_fd, int num) in map_populate() argument 30 err = bpf_map_update_elem(map_fd, &key, value, BPF_NOEXIST); in map_populate() 39 int *map_fd, int populate) in setup() argument 60 *map_fd = bpf_map__fd(skel->maps.hashmap1); in setup() 61 if (CHECK(*map_fd < 0, "bpf_map__fd", "failed\n")) in setup() 64 err = map_populate(*map_fd, populate); in setup() 71 close(*map_fd); in setup() 133 int map_fd, err; in test_pcpu_map_init() local 137 skel = setup(BPF_MAP_TYPE_PERCPU_HASH, 1, &map_fd, 1); in test_pcpu_map_init() 143 err = bpf_map_delete_elem(map_fd, &key); in test_pcpu_map_init() [all …]
|
D | map_lock.c | 23 int err, map_fd = *(u32 *) arg; in parallel_map_access() local 27 err = bpf_map_lookup_elem_flags(map_fd, &key, vars, BPF_F_LOCK); in parallel_map_access() 53 int prog_fd, map_fd[2], vars[17] = {}; in test_map_lock() local 64 map_fd[0] = bpf_find_map(__func__, obj, "hash_map"); in test_map_lock() 65 if (CHECK_FAIL(map_fd[0] < 0)) in test_map_lock() 67 map_fd[1] = bpf_find_map(__func__, obj, "array_map"); in test_map_lock() 68 if (CHECK_FAIL(map_fd[1] < 0)) in test_map_lock() 71 bpf_map_update_elem(map_fd[0], &key, vars, BPF_F_LOCK); in test_map_lock() 80 &map_fd[i - 4]))) in test_map_lock() 88 ret != (void *)&map_fd[i - 4])) in test_map_lock()
|
D | xdp_cpumap_attach.c | 19 int err, prog_fd, map_fd; in test_xdp_with_cpumap_helpers() local 35 map_fd = bpf_map__fd(skel->maps.cpu_map); in test_xdp_with_cpumap_helpers() 41 err = bpf_map_update_elem(map_fd, &idx, &val, 0); in test_xdp_with_cpumap_helpers() 44 err = bpf_map_lookup_elem(map_fd, &idx, &val); in test_xdp_with_cpumap_helpers() 55 err = bpf_map_update_elem(map_fd, &idx, &val, 0); in test_xdp_with_cpumap_helpers() 64 err = bpf_map_update_elem(map_fd, &idx, &val, 0); in test_xdp_with_cpumap_helpers() 79 int err, frags_prog_fd, map_fd; in test_xdp_with_cpumap_frags_helpers() local 87 map_fd = bpf_map__fd(skel->maps.cpu_map); in test_xdp_with_cpumap_frags_helpers() 93 err = bpf_map_update_elem(map_fd, &idx, &val, 0); in test_xdp_with_cpumap_frags_helpers() 96 err = bpf_map_lookup_elem(map_fd, &idx, &val); in test_xdp_with_cpumap_frags_helpers() [all …]
|
D | xdp_devmap_attach.c | 20 int err, dm_fd, map_fd; in test_xdp_with_devmap_helpers() local 37 map_fd = bpf_map__fd(skel->maps.dm_ports); in test_xdp_with_devmap_helpers() 43 err = bpf_map_update_elem(map_fd, &idx, &val, 0); in test_xdp_with_devmap_helpers() 46 err = bpf_map_lookup_elem(map_fd, &idx, &val); in test_xdp_with_devmap_helpers() 57 err = bpf_map_update_elem(map_fd, &idx, &val, 0); in test_xdp_with_devmap_helpers() 66 err = bpf_map_update_elem(map_fd, &idx, &val, 0); in test_xdp_with_devmap_helpers() 92 int err, dm_fd_frags, map_fd; in test_xdp_with_devmap_frags_helpers() local 100 map_fd = bpf_map__fd(skel->maps.dm_ports); in test_xdp_with_devmap_frags_helpers() 106 err = bpf_map_update_elem(map_fd, &idx, &val, 0); in test_xdp_with_devmap_frags_helpers() 109 err = bpf_map_lookup_elem(map_fd, &idx, &val); in test_xdp_with_devmap_frags_helpers() [all …]
|
D | mptcp.c | 23 static int verify_tsk(int map_fd, int client_fd) in verify_tsk() argument 28 err = bpf_map_lookup_elem(map_fd, &cfd, &val); in verify_tsk() 61 static int verify_msk(int map_fd, int client_fd, __u32 token) in verify_msk() argument 72 err = bpf_map_lookup_elem(map_fd, &cfd, &val); in verify_msk() 96 int client_fd, prog_fd, map_fd, err; in run_test() local 113 map_fd = bpf_map__fd(sock_skel->maps.socket_storage_map); in run_test() 114 if (!ASSERT_GE(map_fd, 0, "bpf_map__fd")) { in run_test() 129 err += is_mptcp ? verify_msk(map_fd, client_fd, sock_skel->bss->token) : in run_test() 130 verify_tsk(map_fd, client_fd); in run_test()
|
/linux-6.1.9/tools/testing/selftests/bpf/ |
D | test_lpm_map.c | 428 int map_fd; in test_lpm_delete() local 434 map_fd = bpf_map_create(BPF_MAP_TYPE_LPM_TRIE, NULL, in test_lpm_delete() 437 assert(map_fd >= 0); in test_lpm_delete() 454 assert(bpf_map_update_elem(map_fd, key, &value, 0) == 0); in test_lpm_delete() 459 assert(bpf_map_update_elem(map_fd, key, &value, 0) == 0); in test_lpm_delete() 464 assert(bpf_map_update_elem(map_fd, key, &value, 0) == 0); in test_lpm_delete() 469 assert(bpf_map_update_elem(map_fd, key, &value, 0) == 0); in test_lpm_delete() 474 assert(bpf_map_lookup_elem(map_fd, key, &value) == -ENOENT); in test_lpm_delete() 478 assert(bpf_map_delete_elem(map_fd, key) == -ENOENT); in test_lpm_delete() 482 assert(bpf_map_delete_elem(map_fd, key) == -ENOENT); in test_lpm_delete() [all …]
|
D | test_skb_cgroup_id_user.c | 85 int map_fd = -1; in get_map_fd_by_prog_id() local 106 map_fd = bpf_map_get_fd_by_id(map_ids[0]); in get_map_fd_by_prog_id() 107 if (map_fd < 0) in get_map_fd_by_prog_id() 112 return map_fd; in get_map_fd_by_prog_id() 120 int map_fd; in check_ancestor_cgroup_ids() local 127 map_fd = get_map_fd_by_prog_id(prog_id); in check_ancestor_cgroup_ids() 128 if (map_fd < 0) in check_ancestor_cgroup_ids() 132 if (bpf_map_lookup_elem(map_fd, &level, &actual_ids[level])) { in check_ancestor_cgroup_ids() 147 if (map_fd >= 0) in check_ancestor_cgroup_ids() 148 close(map_fd); in check_ancestor_cgroup_ids()
|
D | test_cgroup_storage.c | 41 int map_fd, percpu_map_fd, prog_fd, cgroup_fd; in main() local 57 map_fd = bpf_map_create(BPF_MAP_TYPE_CGROUP_STORAGE, NULL, sizeof(key), in main() 59 if (map_fd < 0) { in main() 72 prog[7].imm = map_fd; in main() 89 if (bpf_map_get_next_key(map_fd, NULL, &key)) { in main() 94 if (bpf_map_lookup_elem(map_fd, &key, &value)) { in main() 113 if (bpf_map_lookup_elem(map_fd, &key, &value)) { in main() 125 if (bpf_map_update_elem(map_fd, &key, &value, 0)) { in main() 136 if (bpf_map_lookup_elem(map_fd, &key, &value)) { in main()
|
/linux-6.1.9/tools/testing/selftests/bpf/map_tests/ |
D | sk_storage_map.c | 144 int btf_fd, map_fd; in create_sk_storage_map() local 151 map_fd = bpf_map_create(BPF_MAP_TYPE_SK_STORAGE, "sk_storage_map", 4, 8, 0, &map_opts); in create_sk_storage_map() 154 CHECK(map_fd == -1, in create_sk_storage_map() 157 return map_fd; in create_sk_storage_map() 166 int i, map_fd, err, *sk_fds; in insert_close_thread() local 181 map_fd = READ_ONCE(sk_storage_map); in insert_close_thread() 189 err = bpf_map_update_elem(map_fd, &sk_fds[i], &value, in insert_close_thread() 225 int i, map_fd = -1, err = 0, nr_threads_created = 0; in do_sk_storage_map_stress_free() local 246 map_fd = create_sk_storage_map(); in do_sk_storage_map_stress_free() 247 WRITE_ONCE(sk_storage_map, map_fd); in do_sk_storage_map_stress_free() [all …]
|
D | htab_map_batch_ops.c | 14 static void map_batch_update(int map_fd, __u32 max_entries, int *keys, in map_batch_update() argument 37 err = bpf_map_update_batch(map_fd, keys, values, &max_entries, &opts); in map_batch_update() 81 int map_fd, *keys, *visited, key; in __test_map_lookup_and_delete_batch() local 92 map_fd = bpf_map_create(is_pcpu ? BPF_MAP_TYPE_PERCPU_HASH : BPF_MAP_TYPE_HASH, in __test_map_lookup_and_delete_batch() 94 CHECK(map_fd == -1, in __test_map_lookup_and_delete_batch() 109 err = bpf_map_lookup_and_delete_batch(map_fd, NULL, &batch, keys, in __test_map_lookup_and_delete_batch() 115 map_batch_update(map_fd, max_entries, keys, values, is_pcpu); in __test_map_lookup_and_delete_batch() 119 err = bpf_map_lookup_and_delete_batch(map_fd, NULL, &batch, keys, in __test_map_lookup_and_delete_batch() 127 err = bpf_map_lookup_and_delete_batch(map_fd, NULL, &batch, keys, in __test_map_lookup_and_delete_batch() 136 err = bpf_map_get_next_key(map_fd, NULL, &key); in __test_map_lookup_and_delete_batch() [all …]
|
D | lpm_trie_map_batch_ops.c | 22 static void map_batch_update(int map_fd, __u32 max_entries, in map_batch_update() argument 40 err = bpf_map_update_batch(map_fd, keys, values, &max_entries, &opts); in map_batch_update() 70 int map_fd, *values, *visited; in test_lpm_trie_map_batch_ops() local 80 map_fd = bpf_map_create(BPF_MAP_TYPE_LPM_TRIE, "lpm_trie_map", in test_lpm_trie_map_batch_ops() 83 CHECK(map_fd == -1, "bpf_map_create()", "error:%s\n", in test_lpm_trie_map_batch_ops() 94 map_batch_update(map_fd, max_entries, keys, values); in test_lpm_trie_map_batch_ops() 105 err = bpf_map_lookup_batch(map_fd, in test_lpm_trie_map_batch_ops() 127 err = bpf_map_delete_batch(map_fd, keys + total, &count, in test_lpm_trie_map_batch_ops() 139 err = bpf_map_get_next_key(map_fd, NULL, &key); in test_lpm_trie_map_batch_ops() 154 close(map_fd); in test_lpm_trie_map_batch_ops()
|
/linux-6.1.9/tools/lib/bpf/ |
D | ringbuf.c | 32 int map_fd; member 50 int map_fd; member 73 int ring_buffer__add(struct ring_buffer *rb, int map_fd, in ring_buffer__add() argument 86 err = bpf_obj_get_info_by_fd(map_fd, &info, &len); in ring_buffer__add() 90 map_fd, err); in ring_buffer__add() 96 map_fd); in ring_buffer__add() 113 r->map_fd = map_fd; in ring_buffer__add() 119 tmp = mmap(NULL, rb->page_size, PROT_READ | PROT_WRITE, MAP_SHARED, map_fd, 0); in ring_buffer__add() 123 map_fd, err); in ring_buffer__add() 137 tmp = mmap(NULL, (size_t)mmap_sz, PROT_READ, MAP_SHARED, map_fd, rb->page_size); in ring_buffer__add() [all …]
|
/linux-6.1.9/tools/bpf/bpftool/ |
D | iter.c | 19 int err = -1, map_fd = -1; in do_pin() local 37 map_fd = map_parse_fd(&argc, &argv); in do_pin() 38 if (map_fd < 0) in do_pin() 42 linfo.map.map_fd = map_fd; in do_pin() 91 if (map_fd >= 0) in do_pin() 92 close(map_fd); in do_pin()
|