/linux/tools/testing/selftests/bpf/ |
A D | test_lru_map.c | 166 map_flags); in test_lru_sanity0() 170 if (map_flags & BPF_F_NO_COMMON_LRU) in test_lru_sanity0() 269 map_flags); in test_lru_sanity1() 346 map_flags); in test_lru_sanity2() 453 map_flags); in test_lru_sanity3() 512 map_flags); in test_lru_sanity4() 591 map_flags); in test_lru_sanity5() 639 map_flags); in test_lru_sanity6() 702 map_flags); in test_lru_sanity7() 798 map_flags); in test_lru_sanity8() [all …]
|
A D | test_maps.c | 35 static int map_flags; variable 43 2, map_flags); in test_hashmap() 141 2, map_flags); in test_hashmap_sizes() 334 old_flags = map_flags; in test_hashmap_zero_seed() 357 map_flags = old_flags; in test_hashmap_zero_seed() 555 map_flags); in test_queuemap() 559 map_flags); in test_queuemap() 613 map_flags); in test_stackmap() 617 map_flags); in test_stackmap() 1318 MAP_SIZE, map_flags); in test_map_large() [all …]
|
/linux/samples/bpf/ |
A D | test_lru_dist.c | 317 map_flags); in test_parallel_lru_dist() 319 if (map_flags & BPF_F_NO_COMMON_LRU) in test_parallel_lru_dist() 344 map_flags); in test_lru_loss0() 348 if (map_flags & BPF_F_NO_COMMON_LRU) in test_lru_loss0() 396 map_flags); in test_lru_loss1() 400 if (map_flags & BPF_F_NO_COMMON_LRU) in test_lru_loss1() 473 map_flags); in test_parallel_lru_loss() 476 if (map_flags & BPF_F_NO_COMMON_LRU) in test_parallel_lru_loss() 477 map_fd = create_map(map_type, map_flags, in test_parallel_lru_loss() 480 map_fd = create_map(map_type, map_flags, in test_parallel_lru_loss() [all …]
|
A D | map_perf_test_kern.c | 38 __uint(map_flags, BPF_F_NO_COMMON_LRU); 46 __uint(map_flags, BPF_F_NUMA_NODE); 72 __uint(map_flags, BPF_F_NO_PREALLOC); 80 __uint(map_flags, BPF_F_NO_PREALLOC); 88 __uint(map_flags, BPF_F_NO_PREALLOC);
|
A D | tcp_dumpstats_kern.c | 17 __u32 map_flags; member 22 .map_flags = BPF_F_NO_PREALLOC,
|
/linux/kernel/bpf/ |
A D | bpf_local_storage.c | 248 u64 map_flags) in check_flags() argument 254 if (!old_sdata && (map_flags & ~BPF_F_LOCK) == BPF_EXIST) in check_flags() 330 void *value, u64 map_flags) in bpf_local_storage_update() argument 339 if (unlikely((map_flags & ~BPF_F_LOCK) > BPF_EXIST) || in bpf_local_storage_update() 341 unlikely((map_flags & BPF_F_LOCK) && in bpf_local_storage_update() 348 err = check_flags(NULL, map_flags); in bpf_local_storage_update() 366 if ((map_flags & BPF_F_LOCK) && !(map_flags & BPF_NOEXIST)) { in bpf_local_storage_update() 373 err = check_flags(old_sdata, map_flags); in bpf_local_storage_update() 397 err = check_flags(old_sdata, map_flags); in bpf_local_storage_update() 401 if (old_sdata && (map_flags & BPF_F_LOCK)) { in bpf_local_storage_update() [all …]
|
A D | hashtab.c | 1005 u64 map_flags) in check_flags() argument 1020 u64 map_flags) in htab_map_update_elem() argument 1122 u64 map_flags) in htab_lru_map_update_elem() argument 1189 void *value, u64 map_flags, in __htab_percpu_map_update_elem() argument 1274 if (map_flags != BPF_EXIST) { in __htab_lru_percpu_map_update_elem() 1317 void *value, u64 map_flags) in htab_lru_percpu_map_update_elem() argument 1599 u64 elem_map_flags, map_flags; in __htab_map_lookup_and_delete_batch() local 1613 map_flags = attr->batch.flags; in __htab_map_lookup_and_delete_batch() 1614 if (map_flags) in __htab_map_lookup_and_delete_batch() 2209 u64 map_flags) in bpf_percpu_hash_update() argument [all …]
|
A D | arraymap.c | 69 attr->map_flags & BPF_F_PRESERVE_ELEMS) in array_map_alloc_check() 120 if (attr->map_flags & BPF_F_MMAPABLE) { in array_map_alloc() 129 if (attr->map_flags & BPF_F_MMAPABLE) { in array_map_alloc() 210 if (map->map_flags & BPF_F_INNER_MAP) in array_map_gen_lookup() 298 u64 map_flags) in array_map_update_elem() argument 312 if (unlikely(map_flags & BPF_NOEXIST)) in array_map_update_elem() 326 if (map_flags & BPF_F_LOCK) in array_map_update_elem() 336 u64 map_flags) in bpf_percpu_array_update() argument 344 if (unlikely(map_flags > BPF_EXIST)) in bpf_percpu_array_update() 352 if (unlikely(map_flags == BPF_NOEXIST)) in bpf_percpu_array_update() [all …]
|
A D | reuseport_array.c | 201 u32 map_flags) in reuseport_array_update_check() argument 203 if (osk && map_flags == BPF_NOEXIST) in reuseport_array_update_check() 206 if (!osk && map_flags == BPF_EXIST) in reuseport_array_update_check() 242 void *value, u64 map_flags) in bpf_fd_reuseport_array_update_elem() argument 252 if (map_flags > BPF_EXIST) in bpf_fd_reuseport_array_update_elem() 282 map_flags); in bpf_fd_reuseport_array_update_elem() 298 err = reuseport_array_update_check(array, nsk, osk, reuse, map_flags); in bpf_fd_reuseport_array_update_elem()
|
A D | map_in_map.c | 51 inner_map_meta->map_flags = inner_map->map_flags; in bpf_map_meta_alloc() 86 meta0->map_flags == meta1->map_flags; in bpf_map_meta_equal()
|
A D | devmap.c | 121 attr->map_flags & ~DEV_CREATE_FLAG_MASK) in dev_map_init_map() 127 attr->map_flags |= BPF_F_RDONLY_PROG; in dev_map_init_map() 897 void *key, void *value, u64 map_flags) in __dev_map_update_elem() argument 904 if (unlikely(map_flags > BPF_EXIST)) in __dev_map_update_elem() 908 if (unlikely(map_flags == BPF_NOEXIST)) in __dev_map_update_elem() 937 u64 map_flags) in dev_map_update_elem() argument 940 map, key, value, map_flags); in dev_map_update_elem() 944 void *key, void *value, u64 map_flags) in __dev_map_hash_update_elem() argument 962 if (old_dev && (map_flags & BPF_NOEXIST)) in __dev_map_hash_update_elem() 997 u64 map_flags) in dev_map_hash_update_elem() argument [all …]
|
A D | bloom_filter.c | 96 attr->map_flags & ~BLOOM_CREATE_FLAG_MASK || in bloom_map_alloc() 97 !bpf_map_flags_access_ok(attr->map_flags) || in bloom_map_alloc() 154 if (!(attr->map_flags & BPF_F_ZERO_SEED)) in bloom_map_alloc()
|
/linux/tools/lib/bpf/ |
A D | bpf.c | 100 attr.map_flags = create_attr->map_flags; in libbpf__bpf_create_map_xattr() 128 p.map_flags = create_attr->map_flags; in bpf_create_map_xattr() 146 __u32 map_flags, int node) in bpf_create_map_node() argument 152 map_attr.map_flags = map_flags; in bpf_create_map_node() 170 map_attr.map_flags = map_flags; in bpf_create_map() 180 __u32 map_flags) in bpf_create_map_name() argument 186 map_attr.map_flags = map_flags; in bpf_create_map_name() 196 __u32 map_flags, int node) in bpf_create_map_in_map_node() argument 208 attr.map_flags = map_flags; in bpf_create_map_in_map_node() 214 attr.map_flags |= BPF_F_NUMA_NODE; in bpf_create_map_in_map_node() [all …]
|
A D | libbpf_probes.c | 204 int key_size, value_size, max_entries, map_flags; in bpf_probe_map_type() local 212 map_flags = 0; in bpf_probe_map_type() 221 map_flags = BPF_F_NO_PREALLOC; in bpf_probe_map_type() 240 map_flags = BPF_F_NO_PREALLOC; in bpf_probe_map_type() 295 attr.map_flags = map_flags; in bpf_probe_map_type()
|
A D | bpf.h | 40 __u32 map_flags; member 59 int max_entries, __u32 map_flags, int node); 62 int max_entries, __u32 map_flags); 64 int value_size, int max_entries, __u32 map_flags); 68 __u32 map_flags, int node); 72 __u32 map_flags);
|
/linux/tools/testing/selftests/bpf/progs/ |
A D | sockopt_inherit.c | 18 __uint(map_flags, BPF_F_NO_PREALLOC | BPF_F_CLONE); 25 __uint(map_flags, BPF_F_NO_PREALLOC | BPF_F_CLONE); 32 __uint(map_flags, BPF_F_NO_PREALLOC);
|
A D | test_btf_map_in_map.c | 46 __uint(map_flags, BPF_F_INNER_MAP); 55 __uint(map_flags, BPF_F_INNER_MAP); 68 __uint(map_flags, BPF_F_INNER_MAP);
|
A D | test_map_in_map.c | 11 __uint(map_flags, 0); 19 __uint(map_flags, 0);
|
A D | test_mmap.c | 12 __uint(map_flags, BPF_F_MMAPABLE | BPF_F_RDONLY_PROG); 19 __uint(map_flags, BPF_F_MMAPABLE);
|
A D | local_storage.c | 28 __uint(map_flags, BPF_F_NO_PREALLOC); 35 __uint(map_flags, BPF_F_NO_PREALLOC | BPF_F_CLONE); 42 __uint(map_flags, BPF_F_NO_PREALLOC);
|
A D | test_queue_stack_map.h | 14 __uint(map_flags, 0); 22 __uint(map_flags, 0);
|
A D | task_ls_recursion.c | 12 __uint(map_flags, BPF_F_NO_PREALLOC); 19 __uint(map_flags, BPF_F_NO_PREALLOC);
|
A D | netns_cookie_prog.c | 11 __uint(map_flags, BPF_F_NO_PREALLOC); 18 __uint(map_flags, BPF_F_NO_PREALLOC);
|
A D | test_sk_storage_tracing.c | 17 __uint(map_flags, BPF_F_NO_PREALLOC); 25 __uint(map_flags, BPF_F_NO_PREALLOC);
|
/linux/net/xdp/ |
A D | xskmap.c | 70 attr->map_flags & ~(BPF_F_NUMA_NODE | BPF_F_RDONLY | BPF_F_WRONLY)) in xsk_map_alloc() 152 u64 map_flags) in xsk_map_update_elem() argument 162 if (unlikely(map_flags > BPF_EXIST)) in xsk_map_update_elem() 190 } else if (old_xs && map_flags == BPF_NOEXIST) { in xsk_map_update_elem() 193 } else if (!old_xs && map_flags == BPF_EXIST) { in xsk_map_update_elem()
|