ctor 120 arch/powerpc/mm/init-common.c new = kmem_cache_create(name, table_size, align, 0, ctor(shift)); ctor 306 arch/powerpc/platforms/pseries/setup.c void (*ctor)(void *) = get_dtl_cache_ctor(); ctor 309 arch/powerpc/platforms/pseries/setup.c DISPATCH_LOG_BYTES, 0, ctor); ctor 275 drivers/gpu/drm/nouveau/include/nvkm/core/device.h int (*ctor)(struct nvkm_device *, const struct nvkm_oclass *, ctor 21 drivers/gpu/drm/nouveau/include/nvkm/core/event.h int (*ctor)(struct nvkm_object *, void *data, u32 size, ctor 13 drivers/gpu/drm/nouveau/include/nvkm/core/oclass.h int (*ctor)(const struct nvkm_oclass *, void *data, u32 size, ctor 18 drivers/gpu/drm/nouveau/include/nvkm/core/oclass.h int (*ctor)(const struct nvkm_oclass *, void *data, u32 size, ctor 68 drivers/gpu/drm/nouveau/nvkm/core/client.c .ctor = nvkm_uclient_new, ctor 234 drivers/gpu/drm/nouveau/nvkm/core/client.c return oclass->base.ctor(oclass, data, size, pobject); ctor 250 drivers/gpu/drm/nouveau/nvkm/core/client.c oclass->ctor = nvkm_client_child_new; ctor 126 drivers/gpu/drm/nouveau/nvkm/core/ioctl.c ret = oclass.ctor(&oclass, data, size, &object); ctor 140 drivers/gpu/drm/nouveau/nvkm/core/notify.c ret = event->func->ctor(object, data, size, notify); ctor 2628 drivers/gpu/drm/nouveau/nvkm/engine/device/base.c .ctor = nvkm_device_event_ctor, ctor 211 drivers/gpu/drm/nouveau/nvkm/engine/device/ctrl.c .ctor = nvkm_control_new, ctor 341 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c return sclass->ctor(udev->device, oclass, data, size, pobject); ctor 380 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c oclass->ctor = nvkm_udevice_child_new; ctor 455 drivers/gpu/drm/nouveau/nvkm/engine/device/user.c .ctor = nvkm_udevice_new, ctor 85 drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c .ctor = nvkm_disp_vblank_ctor, ctor 128 drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c .ctor = nvkm_disp_hpd_ctor ctor 186 drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c return sclass->ctor(disp, oclass, data, size, &oproxy->object); ctor 191 drivers/gpu/drm/nouveau/nvkm/engine/disp/base.c .ctor = nvkm_disp_class_new, ctor 46 drivers/gpu/drm/nouveau/nvkm/engine/disp/changf119.c .ctor = nv50_disp_chan_uevent_ctor, ctor 26 drivers/gpu/drm/nouveau/nvkm/engine/disp/changv100.c .ctor = nv50_disp_chan_uevent_ctor, ctor 155 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c .ctor = nv50_disp_chan_uevent_ctor, ctor 260 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c ret = sclass->ctor(device, oclass, argv, argc, &object->oproxy.object); ctor 288 drivers/gpu/drm/nouveau/nvkm/engine/disp/channv50.c sclass->ctor = nv50_disp_chan_child_new, ctor 28 drivers/gpu/drm/nouveau/nvkm/engine/disp/priv.h int (*ctor)(struct nvkm_disp *, const struct nvkm_oclass *, ctor 54 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootg84.c .ctor = g84_disp_root_new, ctor 54 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootg94.c .ctor = g94_disp_root_new, ctor 54 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootgf119.c .ctor = gf119_disp_root_new, ctor 54 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootgk104.c .ctor = gk104_disp_root_new, ctor 54 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootgk110.c .ctor = gk110_disp_root_new, ctor 54 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootgm107.c .ctor = gm107_disp_root_new, ctor 54 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootgm200.c .ctor = gm200_disp_root_new, ctor 54 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootgp100.c .ctor = gp100_disp_root_new, ctor 54 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootgp102.c .ctor = gp102_disp_root_new, ctor 54 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootgt200.c .ctor = gt200_disp_root_new, ctor 54 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootgt215.c .ctor = gt215_disp_root_new, ctor 51 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootgv100.c .ctor = gv100_disp_root_new, ctor 97 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv04.c .ctor = nv04_disp_root_new, ctor 283 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c return user->ctor(oclass, argv, argc, disp, pobject); ctor 292 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c if (root->func->user[index].ctor) { ctor 295 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c sclass->ctor = nv50_disp_root_child_new_; ctor 360 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.c .ctor = nv50_disp_root_new, ctor 18 drivers/gpu/drm/nouveau/nvkm/engine/disp/rootnv50.h int (*ctor)(const struct nvkm_oclass *, void *argv, u32 argc, ctor 51 drivers/gpu/drm/nouveau/nvkm/engine/disp/roottu102.c .ctor = tu102_disp_root_new, ctor 48 drivers/gpu/drm/nouveau/nvkm/engine/dma/base.c .ctor = nvkm_dma_oclass_new, ctor 144 drivers/gpu/drm/nouveau/nvkm/engine/fifo/base.c .ctor = nvkm_fifo_kevent_ctor, ctor 162 drivers/gpu/drm/nouveau/nvkm/engine/fifo/base.c .ctor = nvkm_fifo_cevent_ctor, ctor 205 drivers/gpu/drm/nouveau/nvkm/engine/fifo/base.c .ctor = nvkm_fifo_uevent_ctor, ctor 229 drivers/gpu/drm/nouveau/nvkm/engine/fifo/base.c .ctor = nvkm_fifo_class_new_, ctor 239 drivers/gpu/drm/nouveau/nvkm/engine/fifo/base.c return sclass->ctor(fifo, oclass, data, size, pobject); ctor 244 drivers/gpu/drm/nouveau/nvkm/engine/fifo/base.c .ctor = nvkm_fifo_class_new, ctor 176 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c ret = oclass->base.ctor(&(const struct nvkm_oclass) { ctor 220 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c if (!oclass->base.ctor) ctor 221 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c oclass->base.ctor = nvkm_object_new; ctor 222 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c oclass->ctor = nvkm_fifo_chan_child_new; ctor 233 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c if (!oclass->base.ctor) ctor 234 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c oclass->base.ctor = nvkm_object_new; ctor 235 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.c oclass->ctor = nvkm_fifo_chan_child_new; ctor 29 drivers/gpu/drm/nouveau/nvkm/engine/fifo/chan.h int (*ctor)(struct nvkm_fifo *, const struct nvkm_oclass *, ctor 93 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmag84.c .ctor = g84_fifo_dma_new, ctor 225 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv04.c .ctor = nv04_fifo_dma_new, ctor 96 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv10.c .ctor = nv10_fifo_dma_new, ctor 97 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv17.c .ctor = nv17_fifo_dma_new, ctor 245 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv40.c .ctor = nv40_fifo_dma_new, ctor 91 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv50.c .ctor = nv50_fifo_dma_new, ctor 105 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c return user->ctor(fifo, oclass, argv, argc, pobject); ctor 109 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c return user->ctor(oclass, argv, argc, pobject); ctor 122 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c if (fifo->func->user.ctor && c++ == index) { ctor 128 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c if (fifo->func->chan.ctor && c++ == index) { ctor 78 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.h int (*ctor)(const struct nvkm_oclass *, void *, u32, ctor 84 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.h int (*ctor)(struct gk104_fifo *, const struct nvkm_oclass *, ctor 94 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gpfifog84.c .ctor = g84_fifo_gpfifo_new, ctor 300 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gpfifogf100.c .ctor = gf100_fifo_gpfifo_new, ctor 92 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gpfifonv50.c .ctor = nv50_fifo_gpfifo_new, ctor 307 drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c sclass->ctor = gf100_gr_object_new; ctor 620 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c oclass->ctor = nvkm_perfmon_child_new; ctor 687 drivers/gpu/drm/nouveau/nvkm/engine/pm/base.c .ctor = nvkm_pm_oclass_new, ctor 55 drivers/gpu/drm/nouveau/nvkm/engine/sw/base.c return sclass->ctor(chan, oclass, data, size, pobject); ctor 64 drivers/gpu/drm/nouveau/nvkm/engine/sw/base.c while (sw->func->sclass[c].ctor) { ctor 68 drivers/gpu/drm/nouveau/nvkm/engine/sw/base.c oclass->base.ctor = nvkm_sw_oclass_new; ctor 69 drivers/gpu/drm/nouveau/nvkm/engine/sw/chan.c .ctor = nvkm_sw_chan_event_ctor, ctor 12 drivers/gpu/drm/nouveau/nvkm/engine/sw/priv.h int (*ctor)(struct nvkm_sw_chan *, const struct nvkm_oclass *, ctor 57 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c .ctor = nvkm_fault_ntfy_ctor, ctor 180 drivers/gpu/drm/nouveau/nvkm/subdev/fault/base.c fault->user.ctor = nvkm_ufault_new; ctor 142 drivers/gpu/drm/nouveau/nvkm/subdev/gpio/base.c .ctor = nvkm_gpio_intr_ctor, ctor 123 drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c .ctor = nvkm_i2c_intr_ctor, ctor 423 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c mmu->user.ctor = nvkm_ummu_new; ctor 31 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/priv.h int (*ctor)(struct nvkm_mmu *, bool managed, u64 addr, u64 size, ctor 40 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ummu.c oclass->ctor = nvkm_umem_new; ctor 48 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ummu.c oclass->ctor = nvkm_uvmm_new; ctor 399 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c ret = mmu->func->vmm.ctor(mmu, managed, addr, size, argv, argc, ctor 1868 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c ret = mmu->func->vmm.ctor(mmu, false, addr, size, argv, argc, ctor 650 fs/ecryptfs/main.c void (*ctor)(void *obj); ctor 672 fs/ecryptfs/main.c .ctor = inode_info_init_once, ctor 743 fs/ecryptfs/main.c SLAB_HWCACHE_ALIGN | info->flags, info->ctor); ctor 761 fs/erofs/zdata.c struct z_erofs_pagevec_ctor ctor; ctor 803 fs/erofs/zdata.c z_erofs_pagevec_ctor_init(&ctor, Z_EROFS_NR_INLINE_PAGEVECS, ctor 809 fs/erofs/zdata.c page = z_erofs_pagevec_dequeue(&ctor, &page_type); ctor 837 fs/erofs/zdata.c z_erofs_pagevec_ctor_exit(&ctor, true); ctor 37 fs/erofs/zpvec.h static inline void z_erofs_pagevec_ctor_exit(struct z_erofs_pagevec_ctor *ctor, ctor 40 fs/erofs/zpvec.h if (!ctor->curr) ctor 44 fs/erofs/zpvec.h kunmap_atomic(ctor->pages); ctor 46 fs/erofs/zpvec.h kunmap(ctor->curr); ctor 50 fs/erofs/zpvec.h z_erofs_pagevec_ctor_next_page(struct z_erofs_pagevec_ctor *ctor, ctor 56 fs/erofs/zpvec.h if (ctor->next) ctor 57 fs/erofs/zpvec.h return ctor->next; ctor 60 fs/erofs/zpvec.h const erofs_vtptr_t t = ctor->pages[index]; ctor 66 fs/erofs/zpvec.h DBG_BUGON(nr >= ctor->nr); ctor 71 fs/erofs/zpvec.h z_erofs_pagevec_ctor_pagedown(struct z_erofs_pagevec_ctor *ctor, ctor 74 fs/erofs/zpvec.h struct page *next = z_erofs_pagevec_ctor_next_page(ctor, ctor->nr); ctor 76 fs/erofs/zpvec.h z_erofs_pagevec_ctor_exit(ctor, atomic); ctor 78 fs/erofs/zpvec.h ctor->curr = next; ctor 79 fs/erofs/zpvec.h ctor->next = NULL; ctor 80 fs/erofs/zpvec.h ctor->pages = atomic ? ctor 81 fs/erofs/zpvec.h kmap_atomic(ctor->curr) : kmap(ctor->curr); ctor 83 fs/erofs/zpvec.h ctor->nr = PAGE_SIZE / sizeof(struct page *); ctor 84 fs/erofs/zpvec.h ctor->index = 0; ctor 87 fs/erofs/zpvec.h static inline void z_erofs_pagevec_ctor_init(struct z_erofs_pagevec_ctor *ctor, ctor 92 fs/erofs/zpvec.h ctor->nr = nr; ctor 93 fs/erofs/zpvec.h ctor->curr = ctor->next = NULL; ctor 94 fs/erofs/zpvec.h ctor->pages = pages; ctor 98 fs/erofs/zpvec.h z_erofs_pagevec_ctor_pagedown(ctor, false); ctor 99 fs/erofs/zpvec.h while (i > ctor->nr) { ctor 100 fs/erofs/zpvec.h i -= ctor->nr; ctor 101 fs/erofs/zpvec.h z_erofs_pagevec_ctor_pagedown(ctor, false); ctor 104 fs/erofs/zpvec.h ctor->next = z_erofs_pagevec_ctor_next_page(ctor, i); ctor 105 fs/erofs/zpvec.h ctor->index = i; ctor 108 fs/erofs/zpvec.h static inline bool z_erofs_pagevec_enqueue(struct z_erofs_pagevec_ctor *ctor, ctor 114 fs/erofs/zpvec.h if (!ctor->next && type) ctor 115 fs/erofs/zpvec.h if (ctor->index + 1 == ctor->nr) ctor 118 fs/erofs/zpvec.h if (ctor->index >= ctor->nr) ctor 119 fs/erofs/zpvec.h z_erofs_pagevec_ctor_pagedown(ctor, false); ctor 126 fs/erofs/zpvec.h if (type == (uintptr_t)ctor->next) { ctor 127 fs/erofs/zpvec.h ctor->next = page; ctor 130 fs/erofs/zpvec.h ctor->pages[ctor->index++] = tagptr_fold(erofs_vtptr_t, page, type); ctor 135 fs/erofs/zpvec.h z_erofs_pagevec_dequeue(struct z_erofs_pagevec_ctor *ctor, ctor 140 fs/erofs/zpvec.h if (ctor->index >= ctor->nr) { ctor 141 fs/erofs/zpvec.h DBG_BUGON(!ctor->next); ctor 142 fs/erofs/zpvec.h z_erofs_pagevec_ctor_pagedown(ctor, true); ctor 145 fs/erofs/zpvec.h t = ctor->pages[ctor->index]; ctor 150 fs/erofs/zpvec.h if (*type == (uintptr_t)ctor->next) ctor 151 fs/erofs/zpvec.h ctor->next = tagptr_unfold_ptr(t); ctor 153 fs/erofs/zpvec.h ctor->pages[ctor->index++] = tagptr_fold(erofs_vtptr_t, NULL, 0); ctor 213 include/linux/livepatch.h klp_shadow_ctor_t ctor, void *ctor_data); ctor 216 include/linux/livepatch.h klp_shadow_ctor_t ctor, void *ctor_data); ctor 149 include/linux/slab.h void (*ctor)(void *)); ctor 154 include/linux/slab.h void (*ctor)(void *)); ctor 39 include/linux/slab_def.h void (*ctor)(void *obj); ctor 101 include/linux/slub_def.h void (*ctor)(void *); ctor 106 kernel/livepatch/shadow.c klp_shadow_ctor_t ctor, void *ctor_data, ctor 143 kernel/livepatch/shadow.c if (ctor) { ctor 146 kernel/livepatch/shadow.c err = ctor(obj, new_shadow->data, ctor_data); ctor 198 kernel/livepatch/shadow.c klp_shadow_ctor_t ctor, void *ctor_data) ctor 201 kernel/livepatch/shadow.c ctor, ctor_data, true); ctor 227 kernel/livepatch/shadow.c klp_shadow_ctor_t ctor, void *ctor_data) ctor 230 kernel/livepatch/shadow.c ctor, ctor_data, false); ctor 72 lib/livepatch/test_klp_shadow_vars.c gfp_t gfp_flags, klp_shadow_ctor_t ctor, ctor 75 lib/livepatch/test_klp_shadow_vars.c void *ret = klp_shadow_alloc(obj, id, size, gfp_flags, ctor, ctor 78 lib/livepatch/test_klp_shadow_vars.c __func__, ptr_id(obj), id, size, &gfp_flags, ptr_id(ctor), ctor 84 lib/livepatch/test_klp_shadow_vars.c gfp_t gfp_flags, klp_shadow_ctor_t ctor, ctor 87 lib/livepatch/test_klp_shadow_vars.c void *ret = klp_shadow_get_or_alloc(obj, id, size, gfp_flags, ctor, ctor 90 lib/livepatch/test_klp_shadow_vars.c __func__, ptr_id(obj), id, size, &gfp_flags, ptr_id(ctor), ctor 334 lib/test_meminit.c bool ctor, rcu, zero; ctor 339 lib/test_meminit.c ctor = flags & 1; ctor 342 lib/test_meminit.c if (ctor & zero) ctor 344 lib/test_meminit.c num_tests += do_kmem_cache_size(size, ctor, rcu, zero, ctor 268 mm/kasan/common.c (cache->flags & SLAB_TYPESAFE_BY_RCU || cache->ctor || ctor 387 mm/kasan/common.c if (!cache->ctor && !(cache->flags & SLAB_TYPESAFE_BY_RCU)) ctor 512 mm/mempool.c VM_BUG_ON(mem->ctor); ctor 1783 mm/slab.c void (*ctor)(void *)) ctor 1790 mm/slab.c slab_flags_t flags, void (*ctor)(void *)) ctor 1794 mm/slab.c cachep = find_mergeable(size, align, flags, name, ctor); ctor 1822 mm/slab.c if (cachep->ctor || flags & SLAB_TYPESAFE_BY_RCU) ctor 2351 mm/slab.c if (cachep->ctor && !(cachep->flags & SLAB_POISON)) { ctor 2354 mm/slab.c cachep->ctor(objp + obj_offset(cachep)); ctor 2507 mm/slab.c if (DEBUG == 0 && cachep->ctor) { ctor 2509 mm/slab.c cachep->ctor(objp); ctor 2588 mm/slab.c WARN_ON_ONCE(cachep->ctor && (flags & __GFP_ZERO)); ctor 3000 mm/slab.c WARN_ON_ONCE(cachep->ctor && (flags & __GFP_ZERO)); ctor 3024 mm/slab.c if (cachep->ctor && cachep->flags & SLAB_POISON) ctor 3025 mm/slab.c cachep->ctor(objp); ctor 29 mm/slab.h void (*ctor)(void *); /* Called on object slot creation */ ctor 168 mm/slab.h slab_flags_t flags, const char *name, void (*ctor)(void *)); ctor 172 mm/slab.h slab_flags_t flags, void (*ctor)(void *)); ctor 176 mm/slab.h void (*ctor)(void *)); ctor 180 mm/slab.h slab_flags_t flags, void (*ctor)(void *)) ctor 185 mm/slab.h void (*ctor)(void *)) ctor 677 mm/slab.h if (c->ctor) ctor 689 mm/slab.h return !(c->ctor || ctor 317 mm/slab_common.c if (s->ctor) ctor 333 mm/slab_common.c slab_flags_t flags, const char *name, void (*ctor)(void *)) ctor 340 mm/slab_common.c if (ctor) ctor 382 mm/slab_common.c unsigned int usersize, void (*ctor)(void *), ctor 399 mm/slab_common.c s->ctor = ctor; ctor 458 mm/slab_common.c void (*ctor)(void *)) ctor 495 mm/slab_common.c s = __kmem_cache_alias(name, size, align, flags, ctor); ctor 507 mm/slab_common.c flags, useroffset, usersize, ctor, NULL, NULL); ctor 562 mm/slab_common.c slab_flags_t flags, void (*ctor)(void *)) ctor 565 mm/slab_common.c ctor); ctor 684 mm/slab_common.c root_cache->ctor, memcg, root_cache); ctor 613 mm/slob.c if (b && c->ctor) { ctor 615 mm/slob.c c->ctor(b); ctor 1306 mm/slub.c void (*ctor)(void *)) ctor 1365 mm/slub.c void (*ctor)(void *)) ctor 1477 mm/slub.c if (unlikely(s->ctor)) { ctor 1479 mm/slub.c s->ctor(object); ctor 2448 mm/slub.c WARN_ON_ONCE(s->ctor && (flags & __GFP_ZERO)); ctor 3531 mm/slub.c !s->ctor) ctor 3553 mm/slub.c s->ctor)) { ctor 3634 mm/slub.c s->flags = kmem_cache_flags(s->size, flags, s->name, s->ctor); ctor 4291 mm/slub.c slab_flags_t flags, void (*ctor)(void *)) ctor 4295 mm/slub.c s = find_mergeable(size, align, flags, name, ctor); ctor 5035 mm/slub.c if (!s->ctor) ctor 5037 mm/slub.c return sprintf(buf, "%pS\n", s->ctor); ctor 5039 mm/slub.c SLAB_ATTR_RO(ctor); ctor 25 tools/testing/radix-tree/linux.c void (*ctor)(void *); ctor 45 tools/testing/radix-tree/linux.c if (cachep->ctor) ctor 46 tools/testing/radix-tree/linux.c cachep->ctor(node); ctor 102 tools/testing/radix-tree/linux.c unsigned long flags, void (*ctor)(void *)) ctor 110 tools/testing/radix-tree/linux.c ret->ctor = ctor; ctor 25 tools/testing/radix-tree/linux/slab.h unsigned long flags, void (*ctor)(void *));