map_flags 137 drivers/infiniband/core/iwpm_util.c u8 nl_client, u32 map_flags) map_flags 155 drivers/infiniband/core/iwpm_util.c map_info->map_flags = map_flags; map_flags 731 drivers/infiniband/core/iwpm_util.c &map_info->map_flags, map_flags 81 drivers/infiniband/core/iwpm_util.h u32 map_flags; map_flags 53 drivers/iommu/virtio-iommu.c u32 map_flags; map_flags 68 drivers/iommu/virtio-iommu.c u32 map_flags; map_flags 627 drivers/iommu/virtio-iommu.c vdomain->map_flags = viommu->map_flags; map_flags 729 drivers/iommu/virtio-iommu.c if (flags & ~vdomain->map_flags) map_flags 1042 drivers/iommu/virtio-iommu.c viommu->map_flags = VIRTIO_IOMMU_MAP_F_READ | VIRTIO_IOMMU_MAP_F_WRITE; map_flags 1073 drivers/iommu/virtio-iommu.c viommu->map_flags |= VIRTIO_IOMMU_MAP_F_MMIO; map_flags 199 drivers/misc/mic/scif/scif_epd.h int map_flags, scif_pinned_pages_t *pages); map_flags 114 drivers/misc/mic/scif/scif_rma.c int kernel = SCIF_MAP_KERNEL & pin->map_flags; map_flags 716 drivers/misc/mic/scif/scif_rma.c if ((!!(window->pinned_pages->map_flags & SCIF_MAP_KERNEL)) && map_flags 1308 drivers/misc/mic/scif/scif_rma.c int map_flags, scif_pinned_pages_t *pages) map_flags 1319 drivers/misc/mic/scif/scif_rma.c if (map_flags & ~(SCIF_MAP_KERNEL | SCIF_MAP_ULIMIT)) map_flags 1321 drivers/misc/mic/scif/scif_rma.c ulimit = !!(map_flags & SCIF_MAP_ULIMIT); map_flags 1342 drivers/misc/mic/scif/scif_rma.c if (map_flags & SCIF_MAP_KERNEL) { map_flags 1355 drivers/misc/mic/scif/scif_rma.c pinned_pages->map_flags = SCIF_MAP_KERNEL; map_flags 1398 drivers/misc/mic/scif/scif_rma.c pinned_pages->map_flags = 0; map_flags 1425 drivers/misc/mic/scif/scif_rma.c int map_flags, scif_pinned_pages_t *pages) map_flags 1427 drivers/misc/mic/scif/scif_rma.c return __scif_pin_pages(addr, len, &prot, map_flags, pages); map_flags 1468 drivers/misc/mic/scif/scif_rma.c off_t offset, int map_flags) map_flags 1478 drivers/misc/mic/scif/scif_rma.c if (map_flags & ~SCIF_MAP_FIXED) map_flags 1487 drivers/misc/mic/scif/scif_rma.c if ((map_flags & SCIF_MAP_FIXED) && map_flags 1506 drivers/misc/mic/scif/scif_rma.c err = scif_get_window_offset(ep, map_flags, offset, map_flags 1569 drivers/misc/mic/scif/scif_rma.c int prot, int map_flags) map_flags 1581 drivers/misc/mic/scif/scif_rma.c epd, addr, len, offset, prot, map_flags); map_flags 1583 drivers/misc/mic/scif/scif_rma.c if (map_flags & ~(SCIF_MAP_FIXED | SCIF_MAP_KERNEL)) map_flags 1590 drivers/misc/mic/scif/scif_rma.c if ((map_flags & SCIF_MAP_FIXED) && map_flags 1612 drivers/misc/mic/scif/scif_rma.c err = scif_get_window_offset(ep, map_flags, offset, map_flags 1641 drivers/misc/mic/scif/scif_rma.c if (!(map_flags & SCIF_MAP_KERNEL)) { map_flags 1643 drivers/misc/mic/scif/scif_rma.c map_flags |= SCIF_MAP_ULIMIT; map_flags 1647 drivers/misc/mic/scif/scif_rma.c map_flags & (SCIF_MAP_KERNEL | SCIF_MAP_ULIMIT), map_flags 189 drivers/misc/mic/scif/scif_rma.h int map_flags; map_flags 1439 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c u16 map_func, map_flags; map_flags 1464 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c map_flags = TXSCH_MAP_FLAGS(pfvf_map[schq]); map_flags 1467 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c if (map_flags & NIX_TXSCHQ_TL1_CFG_DONE) map_flags 1481 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c map_flags |= NIX_TXSCHQ_TL1_CFG_DONE; map_flags 1482 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c pfvf_map[schq] = TXSCH_MAP(map_func, map_flags); map_flags 1497 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c u16 map_func, map_flags; map_flags 1555 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c map_flags = TXSCH_MAP_FLAGS(pfvf_map[schq]); map_flags 1557 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c map_flags |= NIX_TXSCHQ_TL1_CFG_DONE; map_flags 1558 drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c pfvf_map[schq] = TXSCH_MAP(map_func, map_flags); map_flags 99 drivers/net/ethernet/netronome/nfp/bpf/cmsg.c req->map_flags = 0; map_flags 84 drivers/net/ethernet/netronome/nfp/bpf/fw.h __be32 map_flags; /* reserved */ map_flags 337 drivers/net/ethernet/netronome/nfp/bpf/offload.c if (offmap->map.map_flags || map_flags 492 drivers/net/netdevsim/bpf.c if (offmap->map.map_flags) map_flags 1423 fs/ceph/file.c u32 map_flags; map_flags 1488 fs/ceph/file.c map_flags = osdc->osdmap->flags; map_flags 1491 fs/ceph/file.c if ((map_flags & CEPH_OSDMAP_FULL) || map_flags 1585 fs/ceph/file.c if ((map_flags & CEPH_OSDMAP_NEARFULL) || map_flags 972 fs/ext4/inode.c ext4_lblk_t block, int map_flags) map_flags 976 fs/ext4/inode.c int create = map_flags & EXT4_GET_BLOCKS_CREATE; map_flags 983 fs/ext4/inode.c err = ext4_map_blocks(handle, inode, &map, map_flags); map_flags 1029 fs/ext4/inode.c ext4_lblk_t block, int map_flags) map_flags 1033 fs/ext4/inode.c bh = ext4_getblk(handle, inode, block, map_flags); map_flags 87 include/linux/bpf.h u32 map_flags; map_flags 457 include/linux/bpf.h u32 access_flags = map->map_flags & (BPF_F_RDONLY_PROG | BPF_F_WRONLY_PROG); map_flags 685 include/linux/bpf.h void *key, void *value, u64 map_flags); map_flags 688 include/linux/bpf.h void *key, void *value, u64 map_flags); map_flags 736 include/linux/bpf.h return (attr->map_flags & BPF_F_NUMA_NODE) ? map_flags 1005 include/linux/bpf.h void *value, u64 map_flags); map_flags 1020 include/linux/bpf.h u64 map_flags) map_flags 549 include/linux/scif.h int prot_flags, int map_flags); map_flags 1093 include/linux/scif.h int scif_pin_pages(void *addr, size_t len, int prot_flags, int map_flags, map_flags 1183 include/linux/scif.h off_t offset, int map_flags); map_flags 86 include/rdma/iw_portmap.h u32 map_flags); map_flags 375 include/uapi/linux/bpf.h __u32 map_flags; /* BPF_MAP_CREATE related map_flags 3263 include/uapi/linux/bpf.h __u32 map_flags; map_flags 57 kernel/bpf/arraymap.c attr->map_flags & ~ARRAY_CREATE_FLAG_MASK || map_flags 58 kernel/bpf/arraymap.c !bpf_map_flags_access_ok(attr->map_flags) || map_flags 273 kernel/bpf/arraymap.c u64 map_flags) map_flags 279 kernel/bpf/arraymap.c if (unlikely((map_flags & ~BPF_F_LOCK) > BPF_EXIST)) map_flags 287 kernel/bpf/arraymap.c if (unlikely(map_flags & BPF_NOEXIST)) map_flags 291 kernel/bpf/arraymap.c if (unlikely((map_flags & BPF_F_LOCK) && map_flags 301 kernel/bpf/arraymap.c if (map_flags & BPF_F_LOCK) map_flags 310 kernel/bpf/arraymap.c u64 map_flags) map_flags 318 kernel/bpf/arraymap.c if (unlikely(map_flags > BPF_EXIST)) map_flags 326 kernel/bpf/arraymap.c if (unlikely(map_flags == BPF_NOEXIST)) map_flags 480 kernel/bpf/arraymap.c if (attr->map_flags & (BPF_F_RDONLY_PROG | BPF_F_WRONLY_PROG)) map_flags 526 kernel/bpf/arraymap.c void *key, void *value, u64 map_flags) map_flags 532 kernel/bpf/arraymap.c if (map_flags != BPF_ANY) map_flags 92 kernel/bpf/cpumap.c attr->value_size != 4 || attr->map_flags & ~BPF_F_NUMA_NODE) map_flags 469 kernel/bpf/cpumap.c u64 map_flags) map_flags 479 kernel/bpf/cpumap.c if (unlikely(map_flags > BPF_EXIST)) map_flags 483 kernel/bpf/cpumap.c if (unlikely(map_flags == BPF_NOEXIST)) map_flags 117 kernel/bpf/devmap.c attr->value_size != 4 || attr->map_flags & ~DEV_CREATE_FLAG_MASK) map_flags 123 kernel/bpf/devmap.c attr->map_flags |= BPF_F_RDONLY_PROG; map_flags 624 kernel/bpf/devmap.c void *key, void *value, u64 map_flags) map_flags 631 kernel/bpf/devmap.c if (unlikely(map_flags > BPF_EXIST)) map_flags 635 kernel/bpf/devmap.c if (unlikely(map_flags == BPF_NOEXIST)) map_flags 658 kernel/bpf/devmap.c u64 map_flags) map_flags 661 kernel/bpf/devmap.c map, key, value, map_flags); map_flags 665 kernel/bpf/devmap.c void *key, void *value, u64 map_flags) map_flags 674 kernel/bpf/devmap.c if (unlikely(map_flags > BPF_EXIST || !ifindex)) map_flags 680 kernel/bpf/devmap.c if (old_dev && (map_flags & BPF_NOEXIST)) map_flags 715 kernel/bpf/devmap.c u64 map_flags) map_flags 718 kernel/bpf/devmap.c map, key, value, map_flags); map_flags 76 kernel/bpf/hashtab.c return !(htab->map.map_flags & BPF_F_NO_PREALLOC); map_flags 165 kernel/bpf/hashtab.c htab->map.map_flags & BPF_F_NO_COMMON_LRU, map_flags 237 kernel/bpf/hashtab.c bool percpu_lru = (attr->map_flags & BPF_F_NO_COMMON_LRU); map_flags 238 kernel/bpf/hashtab.c bool prealloc = !(attr->map_flags & BPF_F_NO_PREALLOC); map_flags 239 kernel/bpf/hashtab.c bool zero_seed = (attr->map_flags & BPF_F_ZERO_SEED); map_flags 257 kernel/bpf/hashtab.c if (attr->map_flags & ~HTAB_CREATE_FLAG_MASK || map_flags 258 kernel/bpf/hashtab.c !bpf_map_flags_access_ok(attr->map_flags)) map_flags 306 kernel/bpf/hashtab.c bool percpu_lru = (attr->map_flags & BPF_F_NO_COMMON_LRU); map_flags 307 kernel/bpf/hashtab.c bool prealloc = !(attr->map_flags & BPF_F_NO_PREALLOC); map_flags 367 kernel/bpf/hashtab.c if (htab->map.map_flags & BPF_F_ZERO_SEED) map_flags 805 kernel/bpf/hashtab.c u64 map_flags) map_flags 807 kernel/bpf/hashtab.c if (l_old && (map_flags & ~BPF_F_LOCK) == BPF_NOEXIST) map_flags 811 kernel/bpf/hashtab.c if (!l_old && (map_flags & ~BPF_F_LOCK) == BPF_EXIST) map_flags 820 kernel/bpf/hashtab.c u64 map_flags) map_flags 830 kernel/bpf/hashtab.c if (unlikely((map_flags & ~BPF_F_LOCK) > BPF_EXIST)) map_flags 843 kernel/bpf/hashtab.c if (unlikely(map_flags & BPF_F_LOCK)) { map_flags 849 kernel/bpf/hashtab.c ret = check_flags(htab, l_old, map_flags); map_flags 870 kernel/bpf/hashtab.c ret = check_flags(htab, l_old, map_flags); map_flags 874 kernel/bpf/hashtab.c if (unlikely(l_old && (map_flags & BPF_F_LOCK))) { map_flags 912 kernel/bpf/hashtab.c u64 map_flags) map_flags 922 kernel/bpf/hashtab.c if (unlikely(map_flags > BPF_EXIST)) map_flags 950 kernel/bpf/hashtab.c ret = check_flags(htab, l_old, map_flags); map_flags 976 kernel/bpf/hashtab.c void *value, u64 map_flags, map_flags 987 kernel/bpf/hashtab.c if (unlikely(map_flags > BPF_EXIST)) map_flags 1005 kernel/bpf/hashtab.c ret = check_flags(htab, l_old, map_flags); map_flags 1029 kernel/bpf/hashtab.c void *value, u64 map_flags, map_flags 1040 kernel/bpf/hashtab.c if (unlikely(map_flags > BPF_EXIST)) map_flags 1058 kernel/bpf/hashtab.c if (map_flags != BPF_EXIST) { map_flags 1069 kernel/bpf/hashtab.c ret = check_flags(htab, l_old, map_flags); map_flags 1094 kernel/bpf/hashtab.c void *value, u64 map_flags) map_flags 1096 kernel/bpf/hashtab.c return __htab_percpu_map_update_elem(map, key, value, map_flags, false); map_flags 1100 kernel/bpf/hashtab.c void *value, u64 map_flags) map_flags 1102 kernel/bpf/hashtab.c return __htab_lru_percpu_map_update_elem(map, key, value, map_flags, map_flags 1316 kernel/bpf/hashtab.c u64 map_flags) map_flags 1324 kernel/bpf/hashtab.c map_flags, true); map_flags 1326 kernel/bpf/hashtab.c ret = __htab_percpu_map_update_elem(map, key, value, map_flags, map_flags 1434 kernel/bpf/hashtab.c void *key, void *value, u64 map_flags) map_flags 1444 kernel/bpf/hashtab.c ret = htab_map_update_elem(map, key, &ptr, map_flags); map_flags 201 kernel/bpf/local_storage.c void *value, u64 map_flags) map_flags 209 kernel/bpf/local_storage.c if (map_flags != BPF_ANY && map_flags != BPF_EXIST) map_flags 287 kernel/bpf/local_storage.c if (attr->map_flags & ~LOCAL_STORAGE_CREATE_FLAG_MASK || map_flags 288 kernel/bpf/local_storage.c !bpf_map_flags_access_ok(attr->map_flags)) map_flags 551 kernel/bpf/lpm_trie.c !(attr->map_flags & BPF_F_NO_PREALLOC) || map_flags 552 kernel/bpf/lpm_trie.c attr->map_flags & ~LPM_CREATE_FLAG_MASK || map_flags 553 kernel/bpf/lpm_trie.c !bpf_map_flags_access_ok(attr->map_flags) || map_flags 56 kernel/bpf/map_in_map.c inner_map_meta->map_flags = inner_map->map_flags; map_flags 84 kernel/bpf/map_in_map.c meta0->map_flags == meta1->map_flags && map_flags 54 kernel/bpf/queue_stack_maps.c attr->map_flags & ~QUEUE_STACK_CREATE_FLAG_MASK || map_flags 55 kernel/bpf/queue_stack_maps.c !bpf_map_flags_access_ok(attr->map_flags)) map_flags 208 kernel/bpf/reuseport_array.c u32 map_flags) map_flags 210 kernel/bpf/reuseport_array.c if (osk && map_flags == BPF_NOEXIST) map_flags 213 kernel/bpf/reuseport_array.c if (!osk && map_flags == BPF_EXIST) map_flags 249 kernel/bpf/reuseport_array.c void *value, u64 map_flags) map_flags 258 kernel/bpf/reuseport_array.c if (map_flags > BPF_EXIST) map_flags 288 kernel/bpf/reuseport_array.c map_flags); map_flags 304 kernel/bpf/reuseport_array.c err = reuseport_array_update_check(array, nsk, osk, reuse, map_flags); map_flags 52 kernel/bpf/stackmap.c return (map->map_flags & BPF_F_STACK_BUILD_ID); map_flags 96 kernel/bpf/stackmap.c if (attr->map_flags & ~STACK_CREATE_FLAG_MASK) map_flags 105 kernel/bpf/stackmap.c if (attr->map_flags & BPF_F_STACK_BUILD_ID) { map_flags 573 kernel/bpf/stackmap.c u64 map_flags) map_flags 182 kernel/bpf/syscall.c map->map_flags = bpf_map_flags_retain_permanent(attr->map_flags); map_flags 398 kernel/bpf/syscall.c map->map_flags, map_flags 526 kernel/bpf/syscall.c if (map->map_flags & BPF_F_RDONLY_PROG) map_flags 562 kernel/bpf/syscall.c f_flags = bpf_get_file_flag(attr->map_flags); map_flags 2621 kernel/bpf/syscall.c info.map_flags = map->map_flags; map_flags 3942 kernel/bpf/verifier.c if ((map->map_flags & BPF_F_RDONLY_PROG) && map_flags 7911 kernel/bpf/verifier.c !(map->map_flags & BPF_F_NO_PREALLOC); map_flags 92 kernel/bpf/xskmap.c attr->map_flags & ~(BPF_F_NUMA_NODE | BPF_F_RDONLY | BPF_F_WRONLY)) map_flags 216 kernel/bpf/xskmap.c u64 map_flags) map_flags 225 kernel/bpf/xskmap.c if (unlikely(map_flags > BPF_EXIST)) map_flags 258 kernel/bpf/xskmap.c } else if (old_xs && map_flags == BPF_NOEXIST) { map_flags 261 kernel/bpf/xskmap.c } else if (!old_xs && map_flags == BPF_EXIST) { map_flags 508 mm/mremap.c unsigned long map_flags; map_flags 554 mm/mremap.c map_flags = MAP_FIXED; map_flags 556 mm/mremap.c map_flags |= MAP_SHARED; map_flags 560 mm/mremap.c map_flags); map_flags 710 mm/mremap.c unsigned long map_flags = 0; map_flags 712 mm/mremap.c map_flags |= MAP_SHARED; map_flags 717 mm/mremap.c map_flags); map_flags 313 net/core/bpf_sk_storage.c u64 map_flags) map_flags 315 net/core/bpf_sk_storage.c if (old_sdata && (map_flags & ~BPF_F_LOCK) == BPF_NOEXIST) map_flags 319 net/core/bpf_sk_storage.c if (!old_sdata && (map_flags & ~BPF_F_LOCK) == BPF_EXIST) map_flags 390 net/core/bpf_sk_storage.c u64 map_flags) map_flags 399 net/core/bpf_sk_storage.c if (unlikely((map_flags & ~BPF_F_LOCK) > BPF_EXIST) || map_flags 401 net/core/bpf_sk_storage.c unlikely((map_flags & BPF_F_LOCK) && !map_value_has_spin_lock(map))) map_flags 408 net/core/bpf_sk_storage.c err = check_flags(NULL, map_flags); map_flags 426 net/core/bpf_sk_storage.c if ((map_flags & BPF_F_LOCK) && !(map_flags & BPF_NOEXIST)) { map_flags 432 net/core/bpf_sk_storage.c err = check_flags(old_sdata, map_flags); map_flags 456 net/core/bpf_sk_storage.c err = check_flags(old_sdata, map_flags); map_flags 460 net/core/bpf_sk_storage.c if (old_sdata && (map_flags & BPF_F_LOCK)) { map_flags 611 net/core/bpf_sk_storage.c if (attr->map_flags & ~SK_STORAGE_CREATE_FLAG_MASK || map_flags 612 net/core/bpf_sk_storage.c !(attr->map_flags & BPF_F_NO_PREALLOC) || map_flags 719 net/core/bpf_sk_storage.c void *value, u64 map_flags) map_flags 728 net/core/bpf_sk_storage.c sdata = sk_storage_update(sock->sk, map, value, map_flags); map_flags 794 net/core/bpf_sk_storage.c if (!(smap->map.map_flags & BPF_F_CLONE)) map_flags 35 net/core/sock_map.c attr->map_flags & ~SOCK_CREATE_FLAG_MASK) map_flags 814 net/core/sock_map.c attr->map_flags & ~SOCK_CREATE_FLAG_MASK) map_flags 280 samples/bpf/bpf_load.c numa_node = maps[i].def.map_flags & BPF_F_NUMA_NODE ? map_flags 292 samples/bpf/bpf_load.c maps[i].def.map_flags, map_flags 300 samples/bpf/bpf_load.c maps[i].def.map_flags, map_flags 15 samples/bpf/bpf_load.h unsigned int map_flags; map_flags 35 samples/bpf/map_perf_test_kern.c .map_flags = BPF_F_NO_COMMON_LRU, map_flags 43 samples/bpf/map_perf_test_kern.c .map_flags = BPF_F_NUMA_NODE, map_flags 65 samples/bpf/map_perf_test_kern.c .map_flags = BPF_F_NO_PREALLOC, map_flags 73 samples/bpf/map_perf_test_kern.c .map_flags = BPF_F_NO_PREALLOC, map_flags 81 samples/bpf/map_perf_test_kern.c .map_flags = BPF_F_NO_PREALLOC, map_flags 17 samples/bpf/tcp_dumpstats_kern.c __u32 map_flags; map_flags 22 samples/bpf/tcp_dumpstats_kern.c .map_flags = BPF_F_NO_PREALLOC, map_flags 208 samples/bpf/test_lru_dist.c static int create_map(int map_type, int map_flags, unsigned int size) map_flags 213 samples/bpf/test_lru_dist.c sizeof(unsigned long long), size, map_flags); map_flags 310 samples/bpf/test_lru_dist.c static void test_parallel_lru_dist(int map_type, int map_flags, map_flags 317 samples/bpf/test_lru_dist.c map_flags); map_flags 319 samples/bpf/test_lru_dist.c if (map_flags & BPF_F_NO_COMMON_LRU) map_flags 320 samples/bpf/test_lru_dist.c lru_map_fd = create_map(map_type, map_flags, map_flags 323 samples/bpf/test_lru_dist.c lru_map_fd = create_map(map_type, map_flags, map_flags 335 samples/bpf/test_lru_dist.c static void test_lru_loss0(int map_type, int map_flags) map_flags 344 samples/bpf/test_lru_dist.c map_flags); map_flags 348 samples/bpf/test_lru_dist.c if (map_flags & BPF_F_NO_COMMON_LRU) map_flags 349 samples/bpf/test_lru_dist.c map_fd = create_map(map_type, map_flags, 900 * nr_cpus); map_flags 351 samples/bpf/test_lru_dist.c map_fd = create_map(map_type, map_flags, 900); map_flags 389 samples/bpf/test_lru_dist.c static void test_lru_loss1(int map_type, int map_flags) map_flags 396 samples/bpf/test_lru_dist.c map_flags); map_flags 400 samples/bpf/test_lru_dist.c if (map_flags & BPF_F_NO_COMMON_LRU) map_flags 401 samples/bpf/test_lru_dist.c map_fd = create_map(map_type, map_flags, 1000 * nr_cpus); map_flags 403 samples/bpf/test_lru_dist.c map_fd = create_map(map_type, map_flags, 1000); map_flags 468 samples/bpf/test_lru_dist.c static void test_parallel_lru_loss(int map_type, int map_flags, int nr_tasks) map_flags 473 samples/bpf/test_lru_dist.c map_flags); map_flags 476 samples/bpf/test_lru_dist.c if (map_flags & BPF_F_NO_COMMON_LRU) map_flags 477 samples/bpf/test_lru_dist.c map_fd = create_map(map_type, map_flags, map_flags 480 samples/bpf/test_lru_dist.c map_fd = create_map(map_type, map_flags, map_flags 493 samples/bpf/test_lru_dist.c int map_flags[] = {0, BPF_F_NO_COMMON_LRU}; map_flags 527 samples/bpf/test_lru_dist.c for (f = 0; f < sizeof(map_flags) / sizeof(*map_flags); f++) { map_flags 528 samples/bpf/test_lru_dist.c test_lru_loss0(BPF_MAP_TYPE_LRU_HASH, map_flags[f]); map_flags 529 samples/bpf/test_lru_dist.c test_lru_loss1(BPF_MAP_TYPE_LRU_HASH, map_flags[f]); map_flags 530 samples/bpf/test_lru_dist.c test_parallel_lru_loss(BPF_MAP_TYPE_LRU_HASH, map_flags[f], map_flags 532 samples/bpf/test_lru_dist.c test_parallel_lru_dist(BPF_MAP_TYPE_LRU_HASH, map_flags[f], map_flags 50 samples/bpf/xdp_router_ipv4_kern.c __uint(map_flags, BPF_F_NO_PREALLOC); map_flags 503 tools/bpf/bpftool/map.c jsonw_printf(json_wtr, "%d", info->map_flags); map_flags 581 tools/bpf/bpftool/map.c printf("flags 0x%x", info->map_flags); map_flags 1175 tools/bpf/bpftool/map.c if (parse_u32_arg(&argc, &argv, &attr.map_flags, map_flags 375 tools/include/uapi/linux/bpf.h __u32 map_flags; /* BPF_MAP_CREATE related map_flags 3263 tools/include/uapi/linux/bpf.h __u32 map_flags; map_flags 89 tools/lib/bpf/bpf.c attr.map_flags = create_attr->map_flags; map_flags 105 tools/lib/bpf/bpf.c __u32 map_flags, int node) map_flags 111 tools/lib/bpf/bpf.c map_attr.map_flags = map_flags; map_flags 117 tools/lib/bpf/bpf.c map_attr.map_flags |= BPF_F_NUMA_NODE; map_flags 124 tools/lib/bpf/bpf.c int value_size, int max_entries, __u32 map_flags) map_flags 129 tools/lib/bpf/bpf.c map_attr.map_flags = map_flags; map_flags 139 tools/lib/bpf/bpf.c __u32 map_flags) map_flags 145 tools/lib/bpf/bpf.c map_attr.map_flags = map_flags; map_flags 155 tools/lib/bpf/bpf.c __u32 map_flags, int node) map_flags 166 tools/lib/bpf/bpf.c attr.map_flags = map_flags; map_flags 172 tools/lib/bpf/bpf.c attr.map_flags |= BPF_F_NUMA_NODE; map_flags 181 tools/lib/bpf/bpf.c __u32 map_flags) map_flags 184 tools/lib/bpf/bpf.c inner_map_fd, max_entries, map_flags, map_flags 42 tools/lib/bpf/bpf.h __u32 map_flags; map_flags 58 tools/lib/bpf/bpf.h int max_entries, __u32 map_flags, int node); map_flags 61 tools/lib/bpf/bpf.h int max_entries, __u32 map_flags); map_flags 63 tools/lib/bpf/bpf.h int value_size, int max_entries, __u32 map_flags); map_flags 67 tools/lib/bpf/bpf.h __u32 map_flags, int node); map_flags 71 tools/lib/bpf/bpf.h __u32 map_flags); map_flags 842 tools/lib/bpf/libbpf.c def->map_flags = type == LIBBPF_MAP_RODATA ? BPF_F_RDONLY_PROG : 0; map_flags 1161 tools/lib/bpf/libbpf.c &map->def.map_flags)) map_flags 1164 tools/lib/bpf/libbpf.c map_name, map->def.map_flags); map_flags 1924 tools/lib/bpf/libbpf.c map->def.map_flags = info.map_flags; map_flags 2162 tools/lib/bpf/libbpf.c create_attr.map_flags = def->map_flags; map_flags 291 tools/lib/bpf/libbpf.h unsigned int map_flags; map_flags 195 tools/lib/bpf/libbpf_probes.c int key_size, value_size, max_entries, map_flags; map_flags 203 tools/lib/bpf/libbpf_probes.c map_flags = 0; map_flags 212 tools/lib/bpf/libbpf_probes.c map_flags = BPF_F_NO_PREALLOC; map_flags 229 tools/lib/bpf/libbpf_probes.c map_flags = BPF_F_NO_PREALLOC; map_flags 278 tools/lib/bpf/libbpf_probes.c attr.map_flags = map_flags; map_flags 367 tools/perf/bench/numa.c static u8 *alloc_data(ssize_t bytes0, int map_flags, map_flags 388 tools/perf/bench/numa.c buf = (void *)mmap(0, bytes, PROT_READ|PROT_WRITE, MAP_ANON|map_flags, -1, 0); map_flags 391 tools/perf/bench/numa.c if (map_flags == MAP_PRIVATE) { map_flags 16 tools/perf/include/bpf/bpf.h unsigned int map_flags; map_flags 256 tools/testing/selftests/bpf/bpf_helpers.h unsigned int map_flags; map_flags 25 tools/testing/selftests/bpf/map_tests/sk_storage_map.c .map_flags = BPF_F_NO_PREALLOC, map_flags 582 tools/testing/selftests/bpf/map_tests/sk_storage_map.c bad_xattr.map_flags = 0; map_flags 73 tools/testing/selftests/bpf/prog_tests/bpf_obj_id.c map_infos[i].map_flags != 0 || map_flags 84 tools/testing/selftests/bpf/prog_tests/bpf_obj_id.c map_infos[i].map_flags, map_flags 58 tools/testing/selftests/bpf/prog_tests/obj_name.c attr.map_flags = 0; map_flags 17 tools/testing/selftests/bpf/progs/socket_cookie_prog.c __uint(map_flags, BPF_F_NO_PREALLOC); map_flags 19 tools/testing/selftests/bpf/progs/sockopt_inherit.c __uint(map_flags, BPF_F_NO_PREALLOC | BPF_F_CLONE); map_flags 26 tools/testing/selftests/bpf/progs/sockopt_inherit.c __uint(map_flags, BPF_F_NO_PREALLOC | BPF_F_CLONE); map_flags 33 tools/testing/selftests/bpf/progs/sockopt_inherit.c __uint(map_flags, BPF_F_NO_PREALLOC); map_flags 21 tools/testing/selftests/bpf/progs/sockopt_sk.c .map_flags = BPF_F_NO_PREALLOC, map_flags 20 tools/testing/selftests/bpf/progs/tcp_rtt.c .map_flags = BPF_F_NO_PREALLOC, map_flags 11 tools/testing/selftests/bpf/progs/test_map_in_map.c __uint(map_flags, 0); map_flags 20 tools/testing/selftests/bpf/progs/test_map_in_map.c __uint(map_flags, 0); map_flags 65 tools/testing/selftests/bpf/progs/test_sock_fields_kern.c __uint(map_flags, BPF_F_NO_PREALLOC); map_flags 72 tools/testing/selftests/bpf/progs/test_sock_fields_kern.c __uint(map_flags, BPF_F_NO_PREALLOC); map_flags 30 tools/testing/selftests/bpf/progs/test_stacktrace_build_id.c __uint(map_flags, BPF_F_STACK_BUILD_ID); map_flags 177 tools/testing/selftests/bpf/progs/test_xdp_noinline.c __uint(map_flags, 1U << 1); map_flags 29 tools/testing/selftests/bpf/test_lru_map.c static int create_map(int map_type, int map_flags, unsigned int size) map_flags 34 tools/testing/selftests/bpf/test_lru_map.c sizeof(unsigned long long), size, map_flags); map_flags 159 tools/testing/selftests/bpf/test_lru_map.c static void test_lru_sanity0(int map_type, int map_flags) map_flags 166 tools/testing/selftests/bpf/test_lru_map.c map_flags); map_flags 170 tools/testing/selftests/bpf/test_lru_map.c if (map_flags & BPF_F_NO_COMMON_LRU) map_flags 171 tools/testing/selftests/bpf/test_lru_map.c lru_map_fd = create_map(map_type, map_flags, 2 * nr_cpus); map_flags 173 tools/testing/selftests/bpf/test_lru_map.c lru_map_fd = create_map(map_type, map_flags, 2); map_flags 248 tools/testing/selftests/bpf/test_lru_map.c static void test_lru_sanity1(int map_type, int map_flags, unsigned int tgt_free) map_flags 256 tools/testing/selftests/bpf/test_lru_map.c if (map_flags & BPF_F_NO_COMMON_LRU) map_flags 261 tools/testing/selftests/bpf/test_lru_map.c map_flags); map_flags 269 tools/testing/selftests/bpf/test_lru_map.c lru_map_fd = create_map(map_type, map_flags, map_size); map_flags 324 tools/testing/selftests/bpf/test_lru_map.c static void test_lru_sanity2(int map_type, int map_flags, unsigned int tgt_free) map_flags 333 tools/testing/selftests/bpf/test_lru_map.c if (map_flags & BPF_F_NO_COMMON_LRU) map_flags 338 tools/testing/selftests/bpf/test_lru_map.c map_flags); map_flags 346 tools/testing/selftests/bpf/test_lru_map.c lru_map_fd = create_map(map_type, map_flags, map_size); map_flags 432 tools/testing/selftests/bpf/test_lru_map.c static void test_lru_sanity3(int map_type, int map_flags, unsigned int tgt_free) map_flags 440 tools/testing/selftests/bpf/test_lru_map.c if (map_flags & BPF_F_NO_COMMON_LRU) map_flags 445 tools/testing/selftests/bpf/test_lru_map.c map_flags); map_flags 453 tools/testing/selftests/bpf/test_lru_map.c lru_map_fd = create_map(map_type, map_flags, map_size); map_flags 496 tools/testing/selftests/bpf/test_lru_map.c static void test_lru_sanity4(int map_type, int map_flags, unsigned int tgt_free) map_flags 504 tools/testing/selftests/bpf/test_lru_map.c map_flags); map_flags 508 tools/testing/selftests/bpf/test_lru_map.c if (map_flags & BPF_F_NO_COMMON_LRU) map_flags 509 tools/testing/selftests/bpf/test_lru_map.c lru_map_fd = create_map(map_type, map_flags, map_flags 512 tools/testing/selftests/bpf/test_lru_map.c lru_map_fd = create_map(map_type, map_flags, 3 * tgt_free); map_flags 573 tools/testing/selftests/bpf/test_lru_map.c static void test_lru_sanity5(int map_type, int map_flags) map_flags 579 tools/testing/selftests/bpf/test_lru_map.c if (map_flags & BPF_F_NO_COMMON_LRU) map_flags 583 tools/testing/selftests/bpf/test_lru_map.c map_flags); map_flags 585 tools/testing/selftests/bpf/test_lru_map.c map_fd = create_map(map_type, map_flags, 1); map_flags 620 tools/testing/selftests/bpf/test_lru_map.c static void test_lru_sanity6(int map_type, int map_flags, int tgt_free) map_flags 627 tools/testing/selftests/bpf/test_lru_map.c if (!(map_flags & BPF_F_NO_COMMON_LRU)) map_flags 631 tools/testing/selftests/bpf/test_lru_map.c map_flags); map_flags 638 tools/testing/selftests/bpf/test_lru_map.c lru_map_fd = create_map(map_type, map_flags, map_size * nr_cpus); map_flags 687 tools/testing/selftests/bpf/test_lru_map.c static void test_lru_sanity7(int map_type, int map_flags) map_flags 694 tools/testing/selftests/bpf/test_lru_map.c map_flags); map_flags 698 tools/testing/selftests/bpf/test_lru_map.c if (map_flags & BPF_F_NO_COMMON_LRU) map_flags 699 tools/testing/selftests/bpf/test_lru_map.c lru_map_fd = create_map(map_type, map_flags, 2 * nr_cpus); map_flags 701 tools/testing/selftests/bpf/test_lru_map.c lru_map_fd = create_map(map_type, map_flags, 2); map_flags 783 tools/testing/selftests/bpf/test_lru_map.c static void test_lru_sanity8(int map_type, int map_flags) map_flags 790 tools/testing/selftests/bpf/test_lru_map.c map_flags); map_flags 794 tools/testing/selftests/bpf/test_lru_map.c if (map_flags & BPF_F_NO_COMMON_LRU) map_flags 795 tools/testing/selftests/bpf/test_lru_map.c lru_map_fd = create_map(map_type, map_flags, 2 * nr_cpus); map_flags 797 tools/testing/selftests/bpf/test_lru_map.c lru_map_fd = create_map(map_type, map_flags, 2); map_flags 874 tools/testing/selftests/bpf/test_lru_map.c int map_flags[] = {0, BPF_F_NO_COMMON_LRU}; map_flags 883 tools/testing/selftests/bpf/test_lru_map.c for (f = 0; f < sizeof(map_flags) / sizeof(*map_flags); f++) { map_flags 884 tools/testing/selftests/bpf/test_lru_map.c unsigned int tgt_free = (map_flags[f] & BPF_F_NO_COMMON_LRU) ? map_flags 888 tools/testing/selftests/bpf/test_lru_map.c test_lru_sanity0(map_types[t], map_flags[f]); map_flags 889 tools/testing/selftests/bpf/test_lru_map.c test_lru_sanity1(map_types[t], map_flags[f], tgt_free); map_flags 890 tools/testing/selftests/bpf/test_lru_map.c test_lru_sanity2(map_types[t], map_flags[f], tgt_free); map_flags 891 tools/testing/selftests/bpf/test_lru_map.c test_lru_sanity3(map_types[t], map_flags[f], tgt_free); map_flags 892 tools/testing/selftests/bpf/test_lru_map.c test_lru_sanity4(map_types[t], map_flags[f], tgt_free); map_flags 893 tools/testing/selftests/bpf/test_lru_map.c test_lru_sanity5(map_types[t], map_flags[f]); map_flags 894 tools/testing/selftests/bpf/test_lru_map.c test_lru_sanity6(map_types[t], map_flags[f], tgt_free); map_flags 895 tools/testing/selftests/bpf/test_lru_map.c test_lru_sanity7(map_types[t], map_flags[f]); map_flags 896 tools/testing/selftests/bpf/test_lru_map.c test_lru_sanity8(map_types[t], map_flags[f]); map_flags 35 tools/testing/selftests/bpf/test_maps.c static int map_flags; map_flags 43 tools/testing/selftests/bpf/test_maps.c 2, map_flags); map_flags 134 tools/testing/selftests/bpf/test_maps.c 2, map_flags); map_flags 156 tools/testing/selftests/bpf/test_maps.c sizeof(bpf_percpu(value, 0)), 2, map_flags); map_flags 258 tools/testing/selftests/bpf/test_maps.c max_entries, map_flags); map_flags 261 tools/testing/selftests/bpf/test_maps.c "err: %s, flags: 0x%x\n", strerror(errno), map_flags); map_flags 317 tools/testing/selftests/bpf/test_maps.c old_flags = map_flags; map_flags 318 tools/testing/selftests/bpf/test_maps.c map_flags |= BPF_F_ZERO_SEED; map_flags 340 tools/testing/selftests/bpf/test_maps.c map_flags = old_flags; map_flags 538 tools/testing/selftests/bpf/test_maps.c map_flags); map_flags 542 tools/testing/selftests/bpf/test_maps.c map_flags); map_flags 544 tools/testing/selftests/bpf/test_maps.c if (map_flags & BPF_F_NO_PREALLOC) { map_flags 596 tools/testing/selftests/bpf/test_maps.c map_flags); map_flags 600 tools/testing/selftests/bpf/test_maps.c map_flags); map_flags 602 tools/testing/selftests/bpf/test_maps.c if (map_flags & BPF_F_NO_PREALLOC) { map_flags 1242 tools/testing/selftests/bpf/test_maps.c MAP_SIZE, map_flags); map_flags 1346 tools/testing/selftests/bpf/test_maps.c MAP_SIZE, map_flags); map_flags 1396 tools/testing/selftests/bpf/test_maps.c MAP_SIZE, map_flags | BPF_F_RDONLY); map_flags 1419 tools/testing/selftests/bpf/test_maps.c MAP_SIZE, map_flags | BPF_F_WRONLY); map_flags 1728 tools/testing/selftests/bpf/test_maps.c map_flags = 0; map_flags 1731 tools/testing/selftests/bpf/test_maps.c map_flags = BPF_F_NO_PREALLOC; map_flags 16 tools/testing/selftests/bpf/test_queue_stack_map.h __uint(map_flags, 0); map_flags 24 tools/testing/selftests/bpf/test_queue_stack_map.h __uint(map_flags, 0); map_flags 598 tools/testing/selftests/bpf/test_verifier.c .map_flags = BPF_F_NO_PREALLOC,