kmem_cache_alloc 27 arch/arm64/mm/pgd.c return kmem_cache_alloc(pgd_cache, gfp); kmem_cache_alloc 10 arch/powerpc/include/asm/book3s/32/pgalloc.h return kmem_cache_alloc(PGT_CACHE(PGD_INDEX_SIZE), kmem_cache_alloc 55 arch/powerpc/include/asm/book3s/64/pgalloc.h pgd = kmem_cache_alloc(PGT_CACHE(PGD_INDEX_SIZE), kmem_cache_alloc 97 arch/powerpc/include/asm/book3s/64/pgalloc.h pud = kmem_cache_alloc(PGT_CACHE(PUD_CACHE_INDEX), kmem_cache_alloc 22 arch/powerpc/include/asm/nohash/64/pgalloc.h return kmem_cache_alloc(PGT_CACHE(PUD_INDEX_SIZE), kmem_cache_alloc 50 arch/powerpc/include/asm/nohash/64/pgalloc.h return kmem_cache_alloc(PGT_CACHE(PMD_CACHE_INDEX), kmem_cache_alloc 22 arch/powerpc/include/asm/nohash/pgalloc.h return kmem_cache_alloc(PGT_CACHE(PGD_INDEX_SIZE), kmem_cache_alloc 356 arch/powerpc/kvm/book3s_64_mmu_radix.c return kmem_cache_alloc(kvm_pte_cache, GFP_KERNEL); kmem_cache_alloc 366 arch/powerpc/kvm/book3s_64_mmu_radix.c return kmem_cache_alloc(kvm_pmd_cache, GFP_KERNEL); kmem_cache_alloc 73 arch/powerpc/mm/hugetlbpage.c new = kmem_cache_alloc(cachep, pgtable_gfp_flags(mm, GFP_KERNEL)); kmem_cache_alloc 696 arch/powerpc/perf/hv-24x7.c kmem_cache_alloc(hv_page_cache, GFP_KERNEL); kmem_cache_alloc 933 arch/powerpc/perf/hv-24x7.c void *page = kmem_cache_alloc(hv_page_cache, GFP_USER); kmem_cache_alloc 1013 arch/powerpc/perf/hv-24x7.c void *page = kmem_cache_alloc(hv_page_cache, GFP_USER); \ kmem_cache_alloc 51 arch/powerpc/platforms/cell/spufs/inode.c ei = kmem_cache_alloc(spufs_inode_cache, GFP_KERNEL); kmem_cache_alloc 91 arch/powerpc/platforms/pseries/lpar.c dtl = kmem_cache_alloc(dtl_cache, GFP_KERNEL); kmem_cache_alloc 644 arch/powerpc/sysdev/xive/native.c p = kmem_cache_alloc(xive_provision_cache, GFP_KERNEL); kmem_cache_alloc 89 arch/s390/kernel/nmi.c origin = (unsigned long) kmem_cache_alloc(mcesa_cache, GFP_KERNEL); kmem_cache_alloc 109 arch/s390/kernel/nmi.c origin = (unsigned long) kmem_cache_alloc(mcesa_cache, GFP_KERNEL); kmem_cache_alloc 340 arch/s390/mm/pgalloc.c table = kmem_cache_alloc(base_pgt_cache, GFP_KERNEL); kmem_cache_alloc 32 arch/s390/pci/pci_dma.c table = kmem_cache_alloc(dma_region_table_cache, GFP_ATOMIC); kmem_cache_alloc 50 arch/s390/pci/pci_dma.c table = kmem_cache_alloc(dma_page_table_cache, GFP_ATOMIC); kmem_cache_alloc 23 arch/sh/kernel/cpu/fpu.c tsk->thread.xstate = kmem_cache_alloc(task_xstate_cachep, kmem_cache_alloc 161 arch/sh/kernel/cpu/sh4/sq.c map = kmem_cache_alloc(sq_cache, GFP_KERNEL); kmem_cache_alloc 32 arch/sh/kernel/process.c dst->thread.xstate = kmem_cache_alloc(task_xstate_cachep, kmem_cache_alloc 35 arch/sh/mm/pgtable.c return kmem_cache_alloc(pgd_cachep, PGALLOC_GFP); kmem_cache_alloc 51 arch/sh/mm/pgtable.c return kmem_cache_alloc(pmd_cachep, PGALLOC_GFP); kmem_cache_alloc 28 arch/sparc/include/asm/pgalloc_64.h return kmem_cache_alloc(pgtable_cache, GFP_KERNEL); kmem_cache_alloc 45 arch/sparc/include/asm/pgalloc_64.h return kmem_cache_alloc(pgtable_cache, GFP_KERNEL); kmem_cache_alloc 55 arch/sparc/include/asm/pgalloc_64.h return kmem_cache_alloc(pgtable_cache, GFP_KERNEL); kmem_cache_alloc 393 arch/x86/mm/pgtable.c return kmem_cache_alloc(pgd_cache, GFP_PGTABLE_USER); kmem_cache_alloc 222 block/bio.c bvl = kmem_cache_alloc(bvs->slab, __gfp_mask); kmem_cache_alloc 168 drivers/block/rsxx/cregs.c cmd = kmem_cache_alloc(creg_cmd_pool, GFP_KERNEL); kmem_cache_alloc 161 drivers/block/rsxx/dev.c bio_meta = kmem_cache_alloc(bio_meta_pool, GFP_KERNEL); kmem_cache_alloc 608 drivers/block/rsxx/dma.c dma = kmem_cache_alloc(rsxx_dma_pool, GFP_KERNEL); kmem_cache_alloc 642 drivers/block/rsxx/dma.c dma = kmem_cache_alloc(rsxx_dma_pool, GFP_KERNEL); kmem_cache_alloc 2609 drivers/block/skd_main.c buf = kmem_cache_alloc(s, gfp); kmem_cache_alloc 1050 drivers/crypto/axis/artpec6_crypto.c common->dma = kmem_cache_alloc(ac->dma_cache, flags); kmem_cache_alloc 463 drivers/crypto/caam/qi.c return kmem_cache_alloc(qi_cache, flags); kmem_cache_alloc 325 drivers/crypto/ccp/ccp-dmaengine.c cmd = kmem_cache_alloc(chan->ccp->dma_cmd_cache, GFP_NOWAIT); kmem_cache_alloc 434 drivers/dax/super.c dax_dev = kmem_cache_alloc(dax_cache, GFP_KERNEL); kmem_cache_alloc 87 drivers/dma-buf/st-dma-fence.c f = kmem_cache_alloc(slab_fences, GFP_KERNEL); kmem_cache_alloc 85 drivers/dma/ioat/prep.c sed = kmem_cache_alloc(ioat_sed_cache, flags); kmem_cache_alloc 1258 drivers/firewire/net.c ptask = kmem_cache_alloc(fwnet_packet_task_cache, GFP_ATOMIC); kmem_cache_alloc 145 drivers/gpu/drm/amd/amdgpu/amdgpu_fence.c fence = kmem_cache_alloc(amdgpu_fence_slab, GFP_KERNEL); kmem_cache_alloc 171 drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c e = kmem_cache_alloc(amdgpu_sync_slab, GFP_KERNEL); kmem_cache_alloc 1306 drivers/gpu/drm/exynos/exynos_drm_g2d.c runqueue_node = kmem_cache_alloc(g2d->runqueue_slab, GFP_KERNEL); kmem_cache_alloc 89 drivers/gpu/drm/i915/gem/i915_gem_context.c return kmem_cache_alloc(global.slab_luts, GFP_KERNEL); kmem_cache_alloc 200 drivers/gpu/drm/i915/i915_active.c prealloc = kmem_cache_alloc(global.slab_cache, GFP_KERNEL); kmem_cache_alloc 607 drivers/gpu/drm/i915/i915_active.c node = kmem_cache_alloc(global.slab_cache, GFP_KERNEL); kmem_cache_alloc 350 drivers/gpu/drm/i915/i915_request.c cb = kmem_cache_alloc(global.slab_execute_cbs, gfp); kmem_cache_alloc 618 drivers/gpu/drm/i915/i915_request.c rq = kmem_cache_alloc(global.slab_requests, kmem_cache_alloc 631 drivers/gpu/drm/i915/i915_request.c return kmem_cache_alloc(global.slab_requests, gfp); kmem_cache_alloc 676 drivers/gpu/drm/i915/i915_request.c rq = kmem_cache_alloc(global.slab_requests, kmem_cache_alloc 109 drivers/gpu/drm/i915/i915_scheduler.c p = kmem_cache_alloc(global.slab_priorities, GFP_ATOMIC); kmem_cache_alloc 401 drivers/gpu/drm/i915/i915_scheduler.c return kmem_cache_alloc(global.slab_dependencies, GFP_KERNEL); kmem_cache_alloc 820 drivers/infiniband/hw/hfi1/user_sdma.c tx = kmem_cache_alloc(pq->txreq_cache, GFP_KERNEL); kmem_cache_alloc 103 drivers/infiniband/hw/hfi1/verbs_txreq.c tx = kmem_cache_alloc(dev->verbs_txreq_cache, VERBS_TXREQ_GFP); kmem_cache_alloc 83 drivers/infiniband/hw/hfi1/verbs_txreq.h tx = kmem_cache_alloc(dev->verbs_txreq_cache, VERBS_TXREQ_GFP); kmem_cache_alloc 182 drivers/infiniband/hw/hfi1/vnic_sdma.c tx = kmem_cache_alloc(dd->vnic.txreq_cache, GFP_ATOMIC); kmem_cache_alloc 959 drivers/infiniband/hw/qib/qib_user_sdma.c pkt = kmem_cache_alloc(pq->pkt_slab, GFP_KERNEL); kmem_cache_alloc 665 drivers/infiniband/ulp/srpt/ib_srpt.c ioctx->buf = kmem_cache_alloc(buf_cache, GFP_KERNEL); kmem_cache_alloc 3718 drivers/iommu/amd_iommu.c table->table = kmem_cache_alloc(amd_iommu_irq_cache, GFP_KERNEL); kmem_cache_alloc 532 drivers/iommu/intel-iommu.c return kmem_cache_alloc(iommu_domain_cache, GFP_ATOMIC); kmem_cache_alloc 542 drivers/iommu/intel-iommu.c return kmem_cache_alloc(iommu_devinfo_cache, GFP_ATOMIC); kmem_cache_alloc 380 drivers/md/dm-bufio.c return kmem_cache_alloc(c->slab_cache, gfp_mask); kmem_cache_alloc 444 drivers/md/dm-bufio.c struct dm_buffer *b = kmem_cache_alloc(c->slab_buffer, gfp_mask); kmem_cache_alloc 174 drivers/md/dm-cache-background-tracker.c return kmem_cache_alloc(b->work_cache, GFP_NOWAIT); kmem_cache_alloc 724 drivers/md/dm-snap.c e = kmem_cache_alloc(exception_cache, gfp); kmem_cache_alloc 726 drivers/md/dm-snap.c e = kmem_cache_alloc(exception_cache, GFP_ATOMIC); kmem_cache_alloc 210 drivers/md/raid5-ppl.c io = kmem_cache_alloc(kc, gfp_mask); kmem_cache_alloc 158 drivers/misc/lkdtm/heap.c val = kmem_cache_alloc(double_free_cache, GFP_KERNEL); kmem_cache_alloc 175 drivers/misc/lkdtm/heap.c val = kmem_cache_alloc(a_cache, GFP_KERNEL); kmem_cache_alloc 204 drivers/misc/lkdtm/usercopy.c buf = kmem_cache_alloc(whitelist_cache, GFP_KERNEL); kmem_cache_alloc 1579 drivers/misc/mic/scif/scif_dma.c temp = kmem_cache_alloc(unaligned_cache, GFP_KERNEL); kmem_cache_alloc 1535 drivers/mtd/nand/raw/nandsim.c mypage->byte = kmem_cache_alloc(ns->nand_pages_slab, GFP_NOFS); kmem_cache_alloc 1053 drivers/mtd/ubi/fastmap.c e = kmem_cache_alloc(ubi_wl_entry_slab, GFP_KERNEL); kmem_cache_alloc 1698 drivers/mtd/ubi/wl.c e = kmem_cache_alloc(ubi_wl_entry_slab, GFP_KERNEL); kmem_cache_alloc 1775 drivers/mtd/ubi/wl.c e = kmem_cache_alloc(ubi_wl_entry_slab, GFP_KERNEL); kmem_cache_alloc 1797 drivers/mtd/ubi/wl.c e = kmem_cache_alloc(ubi_wl_entry_slab, GFP_KERNEL); kmem_cache_alloc 3944 drivers/s390/block/dasd_eckd.c char *copy = kmem_cache_alloc(dasd_page_cache, kmem_cache_alloc 514 drivers/s390/block/dasd_fba.c char *copy = kmem_cache_alloc(dasd_page_cache, kmem_cache_alloc 3700 drivers/s390/net/qeth_core_main.c *hdr = kmem_cache_alloc(qeth_core_header_cache, GFP_ATOMIC); kmem_cache_alloc 566 drivers/s390/net/qeth_l2_main.c hdr = kmem_cache_alloc(qeth_core_header_cache, GFP_ATOMIC); kmem_cache_alloc 697 drivers/s390/scsi/zfcp_fsf.c qtcb = kmem_cache_alloc(zfcp_fsf_qtcb_cache, GFP_ATOMIC); kmem_cache_alloc 233 drivers/scsi/aic94xx/aic94xx_hwi.h return kmem_cache_alloc(asd_dma_token_cache, flags); kmem_cache_alloc 4620 drivers/scsi/pmcraid.c kmem_cache_alloc(pinstance->cmd_cachep, GFP_KERNEL); kmem_cache_alloc 328 drivers/soc/fsl/qbman/qman_test_stash.c handler = kmem_cache_alloc(hp_handler_slab, GFP_KERNEL); kmem_cache_alloc 3395 drivers/staging/exfat/exfat_super.c ei = kmem_cache_alloc(exfat_inode_cachep, GFP_NOFS); kmem_cache_alloc 285 drivers/staging/fieldbus/anybuss/host.c t = kmem_cache_alloc(cache, GFP_KERNEL); kmem_cache_alloc 235 drivers/staging/fwserial/fwserial.c txn = kmem_cache_alloc(fwtty_txn_cache, GFP_ATOMIC); kmem_cache_alloc 727 drivers/staging/fwserial/fwserial.c txn = kmem_cache_alloc(fwtty_txn_cache, GFP_ATOMIC); kmem_cache_alloc 2452 drivers/usb/dwc2/hcd.c qh->dw_align_buf = kmem_cache_alloc(hsotg->unaligned_cache, kmem_cache_alloc 205 drivers/usb/mon/mon_text.c (ep = kmem_cache_alloc(rp->e_slab, GFP_ATOMIC)) == NULL) { kmem_cache_alloc 277 drivers/usb/mon/mon_text.c (ep = kmem_cache_alloc(rp->e_slab, GFP_ATOMIC)) == NULL) { kmem_cache_alloc 226 fs/9p/vfs_inode.c v9inode = (struct v9fs_inode *)kmem_cache_alloc(v9fs_inode_cache, kmem_cache_alloc 274 fs/adfs/super.c ei = kmem_cache_alloc(adfs_inode_cachep, GFP_KERNEL); kmem_cache_alloc 103 fs/affs/super.c i = kmem_cache_alloc(affs_inode_cachep, GFP_KERNEL); kmem_cache_alloc 670 fs/afs/super.c vnode = kmem_cache_alloc(afs_inode_cachep, GFP_KERNEL); kmem_cache_alloc 1031 fs/aio.c req = kmem_cache_alloc(kiocb_cachep, GFP_KERNEL); kmem_cache_alloc 279 fs/befs/linuxvfs.c bi = kmem_cache_alloc(befs_inode_cachep, GFP_KERNEL); kmem_cache_alloc 243 fs/bfs/inode.c bi = kmem_cache_alloc(bfs_inode_cachep, GFP_KERNEL); kmem_cache_alloc 773 fs/block_dev.c struct bdev_inode *ei = kmem_cache_alloc(bdev_cachep, GFP_KERNEL); kmem_cache_alloc 345 fs/btrfs/backref.c ref = kmem_cache_alloc(btrfs_prelim_ref_cache, gfp_mask); kmem_cache_alloc 683 fs/btrfs/backref.c new_ref = kmem_cache_alloc(btrfs_prelim_ref_cache, kmem_cache_alloc 935 fs/btrfs/delayed-ref.c ref = kmem_cache_alloc(btrfs_delayed_tree_ref_cachep, GFP_NOFS); kmem_cache_alloc 939 fs/btrfs/delayed-ref.c head_ref = kmem_cache_alloc(btrfs_delayed_ref_head_cachep, GFP_NOFS); kmem_cache_alloc 1030 fs/btrfs/delayed-ref.c ref = kmem_cache_alloc(btrfs_delayed_data_ref_cachep, GFP_NOFS); kmem_cache_alloc 1046 fs/btrfs/delayed-ref.c head_ref = kmem_cache_alloc(btrfs_delayed_ref_head_cachep, GFP_NOFS); kmem_cache_alloc 1108 fs/btrfs/delayed-ref.c head_ref = kmem_cache_alloc(btrfs_delayed_ref_head_cachep, GFP_NOFS); kmem_cache_alloc 299 fs/btrfs/delayed-ref.h return kmem_cache_alloc(btrfs_delayed_extent_op_cachep, GFP_NOFS); kmem_cache_alloc 742 fs/btrfs/disk-io.c end_io_wq = kmem_cache_alloc(btrfs_end_io_wq_cache, GFP_NOFS); kmem_cache_alloc 301 fs/btrfs/extent_io.c state = kmem_cache_alloc(extent_state_cache, mask); kmem_cache_alloc 9385 fs/btrfs/inode.c ei = kmem_cache_alloc(btrfs_inode_cachep, GFP_KERNEL); kmem_cache_alloc 102 fs/cachefiles/bind.c fsdef = kmem_cache_alloc(cachefiles_object_jar, GFP_KERNEL); kmem_cache_alloc 43 fs/cachefiles/interface.c object = kmem_cache_alloc(cachefiles_object_jar, cachefiles_gfp); kmem_cache_alloc 229 fs/ceph/caps.c cap = kmem_cache_alloc(ceph_cap_cachep, GFP_NOFS); kmem_cache_alloc 333 fs/ceph/caps.c cap = kmem_cache_alloc(ceph_cap_cachep, GFP_NOFS); kmem_cache_alloc 1666 fs/ceph/caps.c return kmem_cache_alloc(ceph_cap_flush_cachep, GFP_KERNEL); kmem_cache_alloc 426 fs/ceph/inode.c ci = kmem_cache_alloc(ceph_inode_cachep, GFP_NOFS); kmem_cache_alloc 326 fs/cifs/cifsfs.c cifs_inode = kmem_cache_alloc(cifs_inode_cachep, GFP_KERNEL); kmem_cache_alloc 46 fs/coda/inode.c ei = kmem_cache_alloc(coda_inode_cachep, GFP_KERNEL); kmem_cache_alloc 1688 fs/dcache.c dentry = kmem_cache_alloc(dentry_cache, GFP_KERNEL); kmem_cache_alloc 96 fs/dcookies.c struct dcookie_struct *dcs = kmem_cache_alloc(dcookie_cache, kmem_cache_alloc 1199 fs/direct-io.c dio = kmem_cache_alloc(dio_cache, GFP_KERNEL); kmem_cache_alloc 1414 fs/ecryptfs/crypto.c page_virt = kmem_cache_alloc(ecryptfs_header_cache, GFP_USER); kmem_cache_alloc 1646 fs/ecryptfs/crypto.c tmp_tfm = kmem_cache_alloc(ecryptfs_key_tfm_cache, GFP_KERNEL); kmem_cache_alloc 327 fs/ecryptfs/inode.c dentry_info = kmem_cache_alloc(ecryptfs_dentry_info_cache, GFP_KERNEL); kmem_cache_alloc 2415 fs/ecryptfs/keystore.c key_rec = kmem_cache_alloc(ecryptfs_key_record_cache, GFP_KERNEL); kmem_cache_alloc 2496 fs/ecryptfs/keystore.c new_key_sig = kmem_cache_alloc(ecryptfs_key_sig_cache, GFP_KERNEL); kmem_cache_alloc 418 fs/ecryptfs/mmap.c xattr_virt = kmem_cache_alloc(ecryptfs_xattr_cache, GFP_KERNEL); kmem_cache_alloc 41 fs/ecryptfs/super.c inode_info = kmem_cache_alloc(ecryptfs_inode_info_cache, GFP_KERNEL); kmem_cache_alloc 71 fs/efs/super.c ei = kmem_cache_alloc(efs_inode_cachep, GFP_KERNEL); kmem_cache_alloc 59 fs/erofs/super.c kmem_cache_alloc(erofs_inode_cachep, GFP_KERNEL); kmem_cache_alloc 409 fs/erofs/zdata.c pcl = kmem_cache_alloc(pcluster_cachep, GFP_NOFS); kmem_cache_alloc 1322 fs/eventpoll.c if (epi->nwait >= 0 && (pwq = kmem_cache_alloc(pwq_cache, GFP_KERNEL))) { kmem_cache_alloc 1510 fs/eventpoll.c if (!(epi = kmem_cache_alloc(epi_cache, GFP_KERNEL))) kmem_cache_alloc 1200 fs/exec.c newsighand = kmem_cache_alloc(sighand_cachep, GFP_KERNEL); kmem_cache_alloc 183 fs/ext2/super.c ei = kmem_cache_alloc(ext2_inode_cachep, GFP_KERNEL); kmem_cache_alloc 95 fs/ext4/block_validity.c new_entry = kmem_cache_alloc(ext4_system_zone_cachep, kmem_cache_alloc 449 fs/ext4/extents_status.c es = kmem_cache_alloc(ext4_es_cachep, GFP_ATOMIC); kmem_cache_alloc 1876 fs/ext4/extents_status.c pr = kmem_cache_alloc(ext4_pending_cachep, GFP_ATOMIC); kmem_cache_alloc 3659 fs/ext4/mballoc.c pa = kmem_cache_alloc(ext4_pspace_cachep, GFP_NOFS); kmem_cache_alloc 3753 fs/ext4/mballoc.c pa = kmem_cache_alloc(ext4_pspace_cachep, GFP_NOFS); kmem_cache_alloc 4898 fs/ext4/mballoc.c new_entry = kmem_cache_alloc(ext4_free_data_cachep, kmem_cache_alloc 1084 fs/ext4/super.c ei = kmem_cache_alloc(ext4_inode_cachep, GFP_NOFS); kmem_cache_alloc 209 fs/f2fs/extent_cache.c en = kmem_cache_alloc(extent_node_slab, GFP_ATOMIC); kmem_cache_alloc 2209 fs/f2fs/f2fs.h entry = kmem_cache_alloc(cachep, flags); kmem_cache_alloc 2211 fs/f2fs/f2fs.h entry = kmem_cache_alloc(cachep, flags | __GFP_NOFAIL); kmem_cache_alloc 151 fs/f2fs/node.c new = kmem_cache_alloc(nat_entry_slab, GFP_F2FS_ZERO); kmem_cache_alloc 887 fs/f2fs/super.c fi = kmem_cache_alloc(f2fs_inode_cachep, GFP_F2FS_ZERO); kmem_cache_alloc 64 fs/fat/cache.c return kmem_cache_alloc(fat_cache_cachep, GFP_NOFS); kmem_cache_alloc 747 fs/fat/inode.c ei = kmem_cache_alloc(fat_inode_cachep, GFP_NOFS); kmem_cache_alloc 895 fs/fcntl.c return kmem_cache_alloc(fasync_cache, GFP_KERNEL); kmem_cache_alloc 280 fs/file.c newf = kmem_cache_alloc(files_cachep, GFP_KERNEL); kmem_cache_alloc 127 fs/freevxfs/vxfs_super.c vi = kmem_cache_alloc(vxfs_inode_cachep, GFP_KERNEL); kmem_cache_alloc 114 fs/fs_struct.c struct fs_struct *fs = kmem_cache_alloc(fs_cachep, GFP_KERNEL); kmem_cache_alloc 76 fs/fuse/inode.c fi = kmem_cache_alloc(fuse_inode_cachep, GFP_KERNEL); kmem_cache_alloc 816 fs/gfs2/glock.c gl = kmem_cache_alloc(cachep, GFP_NOFS); kmem_cache_alloc 1394 fs/gfs2/super.c ip = kmem_cache_alloc(gfs2_inode_cachep, GFP_KERNEL); kmem_cache_alloc 166 fs/hfs/super.c i = kmem_cache_alloc(hfs_inode_cachep, GFP_KERNEL); kmem_cache_alloc 85 fs/hfsplus/attributes.c return kmem_cache_alloc(hfsplus_attr_tree_cachep, GFP_KERNEL); kmem_cache_alloc 628 fs/hfsplus/super.c i = kmem_cache_alloc(hfsplus_inode_cachep, GFP_KERNEL); kmem_cache_alloc 236 fs/hpfs/super.c ei = kmem_cache_alloc(hpfs_inode_cachep, GFP_NOFS); kmem_cache_alloc 1038 fs/hugetlbfs/inode.c p = kmem_cache_alloc(hugetlbfs_inode_cachep, GFP_KERNEL); kmem_cache_alloc 233 fs/inode.c inode = kmem_cache_alloc(inode_cachep, GFP_KERNEL); kmem_cache_alloc 624 fs/io_uring.c req = kmem_cache_alloc(req_cachep, gfp); kmem_cache_alloc 639 fs/io_uring.c state->reqs[0] = kmem_cache_alloc(req_cachep, gfp); kmem_cache_alloc 73 fs/isofs/inode.c ei = kmem_cache_alloc(isofs_inode_cachep, GFP_KERNEL); kmem_cache_alloc 2345 fs/jbd2/journal.c ptr = kmem_cache_alloc(get_slab(size), flags); kmem_cache_alloc 145 fs/jbd2/revoke.c record = kmem_cache_alloc(jbd2_revoke_record_cache, gfp_mask); kmem_cache_alloc 224 fs/jbd2/revoke.c table = kmem_cache_alloc(jbd2_revoke_table_cache, GFP_KERNEL); kmem_cache_alloc 130 fs/jffs2/malloc.c ret = kmem_cache_alloc(full_dnode_slab, GFP_KERNEL); kmem_cache_alloc 144 fs/jffs2/malloc.c ret = kmem_cache_alloc(raw_dirent_slab, GFP_KERNEL); kmem_cache_alloc 158 fs/jffs2/malloc.c ret = kmem_cache_alloc(raw_inode_slab, GFP_KERNEL); kmem_cache_alloc 172 fs/jffs2/malloc.c ret = kmem_cache_alloc(tmp_dnode_info_slab, GFP_KERNEL); kmem_cache_alloc 188 fs/jffs2/malloc.c ret = kmem_cache_alloc(raw_node_ref_slab, GFP_KERNEL); kmem_cache_alloc 251 fs/jffs2/malloc.c ret = kmem_cache_alloc(node_frag_slab, GFP_KERNEL); kmem_cache_alloc 265 fs/jffs2/malloc.c ret = kmem_cache_alloc(inode_cache_slab, GFP_KERNEL); kmem_cache_alloc 42 fs/jffs2/super.c f = kmem_cache_alloc(jffs2_inode_cachep, GFP_KERNEL); kmem_cache_alloc 105 fs/jfs/super.c jfs_inode = kmem_cache_alloc(jfs_inode_cachep, GFP_NOFS); kmem_cache_alloc 258 fs/locks.c ctx = kmem_cache_alloc(flctx_cache, GFP_KERNEL); kmem_cache_alloc 88 fs/mbcache.c entry = kmem_cache_alloc(mb_entry_cache, mask); kmem_cache_alloc 66 fs/minix/inode.c ei = kmem_cache_alloc(minix_inode_cachep, GFP_KERNEL); kmem_cache_alloc 2068 fs/nfs/inode.c nfsi = kmem_cache_alloc(nfs_inode_cachep, GFP_KERNEL); kmem_cache_alloc 87 fs/nfs/write.c p = kmem_cache_alloc(nfs_cdata_cachep, GFP_NOIO | kmem_cache_alloc 143 fs/nfsd/filecache.c new = kmem_cache_alloc(nfsd_file_mark_slab, GFP_KERNEL); kmem_cache_alloc 176 fs/nfsd/filecache.c nf = kmem_cache_alloc(nfsd_file_slab, GFP_KERNEL); kmem_cache_alloc 421 fs/nfsd/nfs4layouts.c new = kmem_cache_alloc(nfs4_layout_cache, GFP_KERNEL); kmem_cache_alloc 3937 fs/nfsd/nfs4state.c return kmem_cache_alloc(file_slab, GFP_KERNEL); kmem_cache_alloc 4058 fs/nfsd/nfs4state.c sop = kmem_cache_alloc(slab, GFP_KERNEL); kmem_cache_alloc 98 fs/nfsd/nfscache.c rp = kmem_cache_alloc(nn->drc_slab, GFP_KERNEL); kmem_cache_alloc 28 fs/nilfs2/btree.c path = kmem_cache_alloc(nilfs_btree_path_cache, GFP_NOFS); kmem_cache_alloc 38 fs/nilfs2/segbuf.c segbuf = kmem_cache_alloc(nilfs_segbuf_cachep, GFP_NOFS); kmem_cache_alloc 165 fs/nilfs2/segment.c ti = kmem_cache_alloc(nilfs_transaction_cachep, GFP_NOFS); kmem_cache_alloc 148 fs/nilfs2/super.c ii = kmem_cache_alloc(nilfs_inode_cachep, GFP_NOFS); kmem_cache_alloc 295 fs/notify/dnotify/dnotify.c dn = kmem_cache_alloc(dnotify_struct_cache, GFP_KERNEL); kmem_cache_alloc 302 fs/notify/dnotify/dnotify.c new_dn_mark = kmem_cache_alloc(dnotify_mark_cache, GFP_KERNEL); kmem_cache_alloc 311 fs/notify/fanotify/fanotify.c pevent = kmem_cache_alloc(fanotify_perm_event_cachep, gfp); kmem_cache_alloc 319 fs/notify/fanotify/fanotify.c event = kmem_cache_alloc(fanotify_event_cachep, gfp); kmem_cache_alloc 688 fs/notify/fanotify/fanotify_user.c mark = kmem_cache_alloc(fanotify_mark_cache, GFP_KERNEL); kmem_cache_alloc 570 fs/notify/inotify/inotify_user.c tmp_i_mark = kmem_cache_alloc(inotify_inode_mark_cachep, GFP_KERNEL); kmem_cache_alloc 479 fs/notify/mark.c conn = kmem_cache_alloc(fsnotify_mark_connector_cachep, GFP_KERNEL); kmem_cache_alloc 1259 fs/ntfs/attrib.c ctx = kmem_cache_alloc(ntfs_attr_ctx_cache, GFP_NOFS); kmem_cache_alloc 29 fs/ntfs/index.c ictx = kmem_cache_alloc(ntfs_index_ctx_cache, GFP_NOFS); kmem_cache_alloc 312 fs/ntfs/inode.c ni = kmem_cache_alloc(ntfs_big_inode_cache, GFP_NOFS); kmem_cache_alloc 331 fs/ntfs/inode.c ni = kmem_cache_alloc(ntfs_inode_cache, GFP_NOFS); kmem_cache_alloc 255 fs/ntfs/unistr.c ucs = kmem_cache_alloc(ntfs_name_cache, GFP_NOFS); kmem_cache_alloc 788 fs/ocfs2/dlm/dlmmaster.c alloc_mle = kmem_cache_alloc(dlm_mle_cache, GFP_NOFS); kmem_cache_alloc 1569 fs/ocfs2/dlm/dlmmaster.c mle = kmem_cache_alloc(dlm_mle_cache, GFP_NOFS); kmem_cache_alloc 2573 fs/ocfs2/dlm/dlmmaster.c mle = kmem_cache_alloc(dlm_mle_cache, GFP_NOFS); kmem_cache_alloc 3123 fs/ocfs2/dlm/dlmmaster.c mle = kmem_cache_alloc(dlm_mle_cache, GFP_NOFS); kmem_cache_alloc 328 fs/ocfs2/dlmfs/dlmfs.c ip = kmem_cache_alloc(dlmfs_inode_cache, GFP_NOFS); kmem_cache_alloc 258 fs/ocfs2/quota_local.c newchunk = kmem_cache_alloc(ocfs2_qf_chunk_cachep, GFP_NOFS); kmem_cache_alloc 990 fs/ocfs2/quota_local.c chunk = kmem_cache_alloc(ocfs2_qf_chunk_cachep, GFP_NOFS); kmem_cache_alloc 552 fs/ocfs2/super.c oi = kmem_cache_alloc(ocfs2_inode_cachep, GFP_NOFS); kmem_cache_alloc 399 fs/ocfs2/uptodate.c new = kmem_cache_alloc(ocfs2_uptodate_cachep, GFP_NOFS); kmem_cache_alloc 410 fs/ocfs2/uptodate.c tree[i] = kmem_cache_alloc(ocfs2_uptodate_cachep, kmem_cache_alloc 334 fs/openpromfs/inode.c oi = kmem_cache_alloc(op_inode_cachep, GFP_KERNEL); kmem_cache_alloc 109 fs/orangefs/super.c orangefs_inode = kmem_cache_alloc(orangefs_inode_cache, GFP_KERNEL); kmem_cache_alloc 173 fs/overlayfs/super.c struct ovl_inode *oi = kmem_cache_alloc(ovl_inode_cachep, GFP_KERNEL); kmem_cache_alloc 62 fs/proc/inode.c ei = kmem_cache_alloc(proc_inode_cachep, GFP_KERNEL); kmem_cache_alloc 356 fs/proc/inode.c pdeo = kmem_cache_alloc(pde_opener_cache, GFP_KERNEL); kmem_cache_alloc 342 fs/qnx4/inode.c ei = kmem_cache_alloc(qnx4_inode_cachep, GFP_KERNEL); kmem_cache_alloc 602 fs/qnx6/inode.c ei = kmem_cache_alloc(qnx6_inode_cachep, GFP_KERNEL); kmem_cache_alloc 642 fs/reiserfs/super.c ei = kmem_cache_alloc(reiserfs_inode_cachep, GFP_KERNEL); kmem_cache_alloc 377 fs/romfs/super.c inode = kmem_cache_alloc(romfs_inode_cachep, GFP_KERNEL); kmem_cache_alloc 470 fs/squashfs/super.c kmem_cache_alloc(squashfs_inode_cachep, GFP_KERNEL); kmem_cache_alloc 310 fs/sysv/inode.c si = kmem_cache_alloc(sysv_inode_cachep, GFP_KERNEL); kmem_cache_alloc 253 fs/ubifs/super.c ui = kmem_cache_alloc(ubifs_inode_slab, GFP_NOFS); kmem_cache_alloc 144 fs/udf/super.c ei = kmem_cache_alloc(udf_inode_cachep, GFP_KERNEL); kmem_cache_alloc 1448 fs/ufs/super.c ei = kmem_cache_alloc(ufs_inode_cachep, GFP_NOFS); kmem_cache_alloc 694 fs/userfaultfd.c ctx = kmem_cache_alloc(userfaultfd_ctx_cachep, GFP_KERNEL); kmem_cache_alloc 1962 fs/userfaultfd.c ctx = kmem_cache_alloc(userfaultfd_ctx_cachep, GFP_KERNEL); kmem_cache_alloc 128 fs/xfs/kmem.c ptr = kmem_cache_alloc(zone, lflags); kmem_cache_alloc 2566 include/linux/fs.h #define __getname() kmem_cache_alloc(names_cachep, GFP_KERNEL) kmem_cache_alloc 1454 include/linux/jbd2.h return kmem_cache_alloc(jbd2_inode_cache, gfp_flags); kmem_cache_alloc 394 include/linux/slab.h void *kmem_cache_alloc(struct kmem_cache *, gfp_t flags) __assume_slab_alignment __malloc; kmem_cache_alloc 427 include/linux/slab.h return kmem_cache_alloc(s, flags); kmem_cache_alloc 452 include/linux/slab.h void *ret = kmem_cache_alloc(s, flags); kmem_cache_alloc 680 include/linux/slab.h return kmem_cache_alloc(k, flags | __GFP_ZERO); kmem_cache_alloc 84 include/net/request_sock.h req = kmem_cache_alloc(ops->slab, GFP_ATOMIC | __GFP_NOWARN); kmem_cache_alloc 54 include/trace/events/kmem.h DEFINE_EVENT(kmem_alloc, kmem_cache_alloc, kmem_cache_alloc 427 ipc/mqueue.c ei = kmem_cache_alloc(mqueue_inode_cachep, GFP_KERNEL); kmem_cache_alloc 1681 kernel/audit.c ab = kmem_cache_alloc(audit_buffer_cache, gfp_mask); kmem_cache_alloc 258 kernel/cred.c new = kmem_cache_alloc(cred_jar, GFP_KERNEL); kmem_cache_alloc 686 kernel/cred.c new = kmem_cache_alloc(cred_jar, GFP_KERNEL); kmem_cache_alloc 348 kernel/fork.c vma = kmem_cache_alloc(vm_area_cachep, GFP_KERNEL); kmem_cache_alloc 356 kernel/fork.c struct vm_area_struct *new = kmem_cache_alloc(vm_area_cachep, GFP_KERNEL); kmem_cache_alloc 677 kernel/fork.c #define allocate_mm() (kmem_cache_alloc(mm_cachep, GFP_KERNEL)) kmem_cache_alloc 1507 kernel/fork.c sig = kmem_cache_alloc(sighand_cachep, GFP_KERNEL); kmem_cache_alloc 49 kernel/nsproxy.c nsproxy = kmem_cache_alloc(nsproxy_cachep, GFP_KERNEL); kmem_cache_alloc 169 kernel/pid.c pid = kmem_cache_alloc(ns->pid_cachep, GFP_KERNEL); kmem_cache_alloc 6941 kernel/sched/core.c tg = kmem_cache_alloc(task_group_cache, GFP_KERNEL | __GFP_ZERO); kmem_cache_alloc 434 kernel/signal.c q = kmem_cache_alloc(sigqueue_cachep, flags); kmem_cache_alloc 111 kernel/trace/trace_events.c field = kmem_cache_alloc(field_cachep, GFP_TRACE); kmem_cache_alloc 2249 kernel/trace/trace_events.c file = kmem_cache_alloc(file_cachep, GFP_TRACE); kmem_cache_alloc 34 kernel/utsname.c uts_ns = kmem_cache_alloc(uts_ns_cache, GFP_KERNEL); kmem_cache_alloc 83 lib/btree.c return kmem_cache_alloc(btree_cachep, gfp_mask); kmem_cache_alloc 134 lib/lru_cache.c void *p = kmem_cache_alloc(cache, GFP_KERNEL); kmem_cache_alloc 262 lib/radix-tree.c ret = kmem_cache_alloc(radix_tree_node_cachep, kmem_cache_alloc 285 lib/radix-tree.c ret = kmem_cache_alloc(radix_tree_node_cachep, gfp_mask); kmem_cache_alloc 347 lib/radix-tree.c node = kmem_cache_alloc(radix_tree_node_cachep, gfp_mask); kmem_cache_alloc 392 lib/test_kasan.c p = kmem_cache_alloc(cache, GFP_KERNEL); kmem_cache_alloc 423 lib/test_kasan.c p = kmem_cache_alloc(cache, GFP_KERNEL); kmem_cache_alloc 552 lib/test_kasan.c p = kmem_cache_alloc(cache, GFP_KERNEL); kmem_cache_alloc 577 lib/test_kasan.c p = kmem_cache_alloc(cache, GFP_KERNEL); kmem_cache_alloc 206 lib/test_meminit.c buf = kmem_cache_alloc(c, alloc_mask); kmem_cache_alloc 263 lib/test_meminit.c buf = kmem_cache_alloc(c, GFP_KERNEL); kmem_cache_alloc 281 lib/test_meminit.c buf = kmem_cache_alloc(c, GFP_KERNEL); kmem_cache_alloc 304 lib/xarray.c xas->xa_alloc = kmem_cache_alloc(radix_tree_node_cachep, gfp); kmem_cache_alloc 335 lib/xarray.c xas->xa_alloc = kmem_cache_alloc(radix_tree_node_cachep, gfp); kmem_cache_alloc 338 lib/xarray.c xas->xa_alloc = kmem_cache_alloc(radix_tree_node_cachep, gfp); kmem_cache_alloc 371 lib/xarray.c node = kmem_cache_alloc(radix_tree_node_cachep, gfp); kmem_cache_alloc 53 mm/kmemleak-test.c kmem_cache_alloc(files_cachep, GFP_KERNEL)); kmem_cache_alloc 55 mm/kmemleak-test.c kmem_cache_alloc(files_cachep, GFP_KERNEL)); kmem_cache_alloc 423 mm/kmemleak.c object = kmem_cache_alloc(object_cache, gfp_kmemleak_mask(gfp)); kmem_cache_alloc 799 mm/kmemleak.c area = kmem_cache_alloc(scan_area_cache, gfp_kmemleak_mask(gfp)); kmem_cache_alloc 402 mm/ksm.c return kmem_cache_alloc(stable_node_cache, GFP_KERNEL | __GFP_HIGH); kmem_cache_alloc 4661 mm/memory.c ptl = kmem_cache_alloc(page_ptl_cachep, GFP_KERNEL); kmem_cache_alloc 276 mm/mempolicy.c policy = kmem_cache_alloc(policy_cache, GFP_KERNEL); kmem_cache_alloc 2224 mm/mempolicy.c struct mempolicy *new = kmem_cache_alloc(policy_cache, GFP_KERNEL); kmem_cache_alloc 2489 mm/mempolicy.c n = kmem_cache_alloc(sn_cache, GFP_KERNEL); kmem_cache_alloc 2561 mm/mempolicy.c n_new = kmem_cache_alloc(sn_cache, GFP_KERNEL); kmem_cache_alloc 2564 mm/mempolicy.c mpol_new = kmem_cache_alloc(policy_cache, GFP_KERNEL); kmem_cache_alloc 513 mm/mempool.c return kmem_cache_alloc(mem, gfp_mask); kmem_cache_alloc 1398 mm/nommu.c region = kmem_cache_alloc(vm_region_jar, GFP_KERNEL); kmem_cache_alloc 83 mm/rmap.c anon_vma = kmem_cache_alloc(anon_vma_cachep, GFP_KERNEL); kmem_cache_alloc 130 mm/rmap.c return kmem_cache_alloc(anon_vma_chain_cachep, gfp); kmem_cache_alloc 3741 mm/shmem.c info = kmem_cache_alloc(shmem_inode_cachep, GFP_KERNEL); kmem_cache_alloc 3490 mm/slab.c EXPORT_SYMBOL(kmem_cache_alloc); kmem_cache_alloc 122 mm/slab_common.c void *x = p[i] = kmem_cache_alloc(s, flags); kmem_cache_alloc 1796 mm/slab_common.c EXPORT_TRACEPOINT_SYMBOL(kmem_cache_alloc); kmem_cache_alloc 626 mm/slob.c EXPORT_SYMBOL(kmem_cache_alloc); kmem_cache_alloc 2786 mm/slub.c EXPORT_SYMBOL(kmem_cache_alloc); kmem_cache_alloc 4156 mm/slub.c n = kmem_cache_alloc(kmem_cache_node, GFP_KERNEL); kmem_cache_alloc 973 mm/vmalloc.c lva = kmem_cache_alloc(vmap_area_cachep, GFP_NOWAIT); kmem_cache_alloc 201 mm/z3fold.c slots = kmem_cache_alloc(pool->c_handle, kmem_cache_alloc 353 mm/zsmalloc.c return (unsigned long)kmem_cache_alloc(pool->handle_cachep, kmem_cache_alloc 364 mm/zsmalloc.c return kmem_cache_alloc(pool->zspage_cachep, kmem_cache_alloc 251 mm/zswap.c entry = kmem_cache_alloc(zswap_entry_cache, gfp); kmem_cache_alloc 229 net/9p/client.c fc->sdata = kmem_cache_alloc(c->fcall_cache, GFP_NOFS); kmem_cache_alloc 270 net/9p/client.c struct p9_req_t *req = kmem_cache_alloc(p9_req_cache, GFP_NOFS); kmem_cache_alloc 475 net/batman-adv/translation-table.c tt_change_node = kmem_cache_alloc(batadv_tt_change_cache, GFP_ATOMIC); kmem_cache_alloc 711 net/batman-adv/translation-table.c tt_local = kmem_cache_alloc(batadv_tl_cache, GFP_ATOMIC); kmem_cache_alloc 2896 net/batman-adv/translation-table.c tt_req_node = kmem_cache_alloc(batadv_tt_req_cache, GFP_ATOMIC); kmem_cache_alloc 3725 net/batman-adv/translation-table.c tt_roam_node = kmem_cache_alloc(batadv_tt_roam_cache, kmem_cache_alloc 492 net/bridge/br_fdb.c fdb = kmem_cache_alloc(br_fdb_cache, GFP_ATOMIC); kmem_cache_alloc 459 net/can/af_can.c rcv = kmem_cache_alloc(rcv_cache, GFP_KERNEL); kmem_cache_alloc 1041 net/can/gw.c gwj = kmem_cache_alloc(cgw_cache, GFP_KERNEL); kmem_cache_alloc 587 net/ceph/osd_client.c req = kmem_cache_alloc(ceph_osd_request_cache, gfp_flags); kmem_cache_alloc 93 net/core/dst.c dst = kmem_cache_alloc(ops->kmem_cachep, GFP_ATOMIC); kmem_cache_alloc 310 net/core/skbuff.c skb = kmem_cache_alloc(skbuff_head_cache, GFP_ATOMIC); kmem_cache_alloc 1448 net/core/skbuff.c n = kmem_cache_alloc(skbuff_head_cache, gfp_mask); kmem_cache_alloc 5981 net/core/skbuff.c struct skb_ext *new = kmem_cache_alloc(skbuff_ext_cache, GFP_ATOMIC); kmem_cache_alloc 5999 net/core/skbuff.c new = kmem_cache_alloc(skbuff_ext_cache, GFP_ATOMIC); kmem_cache_alloc 1599 net/core/sock.c sk = kmem_cache_alloc(slab, priority & ~__GFP_ZERO); kmem_cache_alloc 55 net/dccp/ackvec.c avr = kmem_cache_alloc(dccp_ackvec_record_slab, GFP_ATOMIC); kmem_cache_alloc 148 net/dccp/ccid.c ccid = kmem_cache_alloc(rx ? ccid_ops->ccid_hc_rx_slab : kmem_cache_alloc 40 net/dccp/ccids/lib/loss_interval.c lh->ring[LIH_INDEX(lh->counter)] = kmem_cache_alloc(tfrc_lh_slab, kmem_cache_alloc 51 net/dccp/ccids/lib/packet_history.c struct tfrc_tx_hist_entry *entry = kmem_cache_alloc(tfrc_tx_hist_slab, gfp_any()); kmem_cache_alloc 339 net/dccp/ccids/lib/packet_history.c h->ring[i] = kmem_cache_alloc(tfrc_rx_hist_slab, GFP_ATOMIC); kmem_cache_alloc 360 net/ipv4/fib_trie.c kv = kmem_cache_alloc(trie_leaf_kmem, GFP_KERNEL); kmem_cache_alloc 1204 net/ipv4/fib_trie.c new_fa = kmem_cache_alloc(fn_alias_kmem, GFP_KERNEL); kmem_cache_alloc 1258 net/ipv4/fib_trie.c new_fa = kmem_cache_alloc(fn_alias_kmem, GFP_KERNEL); kmem_cache_alloc 1761 net/ipv4/fib_trie.c new_fa = kmem_cache_alloc(fn_alias_kmem, GFP_KERNEL); kmem_cache_alloc 67 net/ipv4/inet_hashtables.c struct inet_bind_bucket *tb = kmem_cache_alloc(cachep, GFP_ATOMIC); kmem_cache_alloc 164 net/ipv4/inet_timewait_sock.c tw = kmem_cache_alloc(sk->sk_prot_creator->twsk_prot->twsk_slab, kmem_cache_alloc 216 net/ipv4/inetpeer.c p = kmem_cache_alloc(peer_cachep, GFP_ATOMIC); kmem_cache_alloc 147 net/ipv6/xfrm6_tunnel.c x6spi = kmem_cache_alloc(xfrm6_tunnel_spi_kmem, GFP_ATOMIC); kmem_cache_alloc 237 net/mac80211/mesh.c p = kmem_cache_alloc(rm_cache, GFP_ATOMIC); kmem_cache_alloc 910 net/netfilter/ipvs/ip_vs_conn.c cp = kmem_cache_alloc(ip_vs_conn_cachep, GFP_ATOMIC); kmem_cache_alloc 183 net/netfilter/nf_conncount.c conn = kmem_cache_alloc(conncount_conn_cachep, GFP_ATOMIC); kmem_cache_alloc 358 net/netfilter/nf_conncount.c rbconn = kmem_cache_alloc(conncount_rb_cachep, GFP_ATOMIC); kmem_cache_alloc 362 net/netfilter/nf_conncount.c conn = kmem_cache_alloc(conncount_conn_cachep, GFP_ATOMIC); kmem_cache_alloc 1371 net/netfilter/nf_conntrack_core.c ct = kmem_cache_alloc(nf_conntrack_cachep, gfp); kmem_cache_alloc 286 net/netfilter/nf_conntrack_expect.c new = kmem_cache_alloc(nf_ct_expect_cachep, GFP_ATOMIC); kmem_cache_alloc 246 net/netfilter/xt_hashlimit.c ent = kmem_cache_alloc(hashlimit_cachep, GFP_ATOMIC); kmem_cache_alloc 262 net/rds/ib_recv.c ibinc = kmem_cache_alloc(rds_ib_incoming_slab, slab_mask); kmem_cache_alloc 288 net/rds/ib_recv.c frag = kmem_cache_alloc(rds_ib_frag_slab, slab_mask); kmem_cache_alloc 383 net/rds/tcp.c tc = kmem_cache_alloc(rds_tcp_conn_slab, gfp); kmem_cache_alloc 173 net/rds/tcp_recv.c tinc = kmem_cache_alloc(rds_tcp_incoming_slab, kmem_cache_alloc 8538 net/sctp/socket.c pp = kmem_cache_alloc(sctp_bucket_cachep, GFP_ATOMIC); kmem_cache_alloc 239 net/socket.c ei = kmem_cache_alloc(sock_inode_cachep, GFP_KERNEL); kmem_cache_alloc 200 net/sunrpc/rpc_pipe.c rpci = kmem_cache_alloc(rpc_inode_cachep, GFP_KERNEL); kmem_cache_alloc 615 net/xfrm/xfrm_state.c x = kmem_cache_alloc(xfrm_state_cache, GFP_ATOMIC | __GFP_ZERO); kmem_cache_alloc 105 security/integrity/iint.c iint = kmem_cache_alloc(iint_cache, GFP_NOFS); kmem_cache_alloc 20 tools/testing/radix-tree/linux/slab.h void *kmem_cache_alloc(struct kmem_cache *cachep, int flags);