new_ext          3101 drivers/iommu/intel-iommu.c 	bool new_ext, ext;
new_ext          3105 drivers/iommu/intel-iommu.c 	new_ext    = !!ecap_ecs(iommu->ecap);
new_ext          3113 drivers/iommu/intel-iommu.c 	if (new_ext != ext)
new_ext           606 kernel/power/snapshot.c 			struct mem_extent *new_ext;
new_ext           608 kernel/power/snapshot.c 			new_ext = kzalloc(sizeof(struct mem_extent), gfp_mask);
new_ext           609 kernel/power/snapshot.c 			if (!new_ext) {
new_ext           613 kernel/power/snapshot.c 			new_ext->start = zone_start;
new_ext           614 kernel/power/snapshot.c 			new_ext->end = zone_end;
new_ext           615 kernel/power/snapshot.c 			list_add_tail(&new_ext->hook, &ext->hook);
new_ext           226 mm/page_owner.c 	struct page_ext *new_ext = lookup_page_ext(newpage);
new_ext           229 mm/page_owner.c 	if (unlikely(!old_ext || !new_ext))
new_ext           233 mm/page_owner.c 	new_page_owner = get_page_owner(new_ext);
new_ext           249 mm/page_owner.c 	__set_bit(PAGE_EXT_OWNER, &new_ext->flags);
new_ext           250 mm/page_owner.c 	__set_bit(PAGE_EXT_OWNER_ALLOCATED, &new_ext->flags);