new_ext 3101 drivers/iommu/intel-iommu.c bool new_ext, ext; new_ext 3105 drivers/iommu/intel-iommu.c new_ext = !!ecap_ecs(iommu->ecap); new_ext 3113 drivers/iommu/intel-iommu.c if (new_ext != ext) new_ext 606 kernel/power/snapshot.c struct mem_extent *new_ext; new_ext 608 kernel/power/snapshot.c new_ext = kzalloc(sizeof(struct mem_extent), gfp_mask); new_ext 609 kernel/power/snapshot.c if (!new_ext) { new_ext 613 kernel/power/snapshot.c new_ext->start = zone_start; new_ext 614 kernel/power/snapshot.c new_ext->end = zone_end; new_ext 615 kernel/power/snapshot.c list_add_tail(&new_ext->hook, &ext->hook); new_ext 226 mm/page_owner.c struct page_ext *new_ext = lookup_page_ext(newpage); new_ext 229 mm/page_owner.c if (unlikely(!old_ext || !new_ext)) new_ext 233 mm/page_owner.c new_page_owner = get_page_owner(new_ext); new_ext 249 mm/page_owner.c __set_bit(PAGE_EXT_OWNER, &new_ext->flags); new_ext 250 mm/page_owner.c __set_bit(PAGE_EXT_OWNER_ALLOCATED, &new_ext->flags);