new_pages 1231 arch/s390/kernel/debug.c int rc, new_pages; new_pages 1245 arch/s390/kernel/debug.c new_pages = debug_get_uint(str); new_pages 1246 arch/s390/kernel/debug.c if (new_pages < 0) { new_pages 1250 arch/s390/kernel/debug.c rc = debug_set_size(id, id->nr_areas, new_pages); new_pages 292 drivers/base/firmware_loader/main.c struct page **new_pages; new_pages 294 drivers/base/firmware_loader/main.c new_pages = kvmalloc_array(new_array_size, sizeof(void *), new_pages 296 drivers/base/firmware_loader/main.c if (!new_pages) new_pages 298 drivers/base/firmware_loader/main.c memcpy(new_pages, fw_priv->pages, new_pages 300 drivers/base/firmware_loader/main.c memset(&new_pages[fw_priv->page_array_size], 0, sizeof(void *) * new_pages 303 drivers/base/firmware_loader/main.c fw_priv->pages = new_pages; new_pages 381 drivers/block/drbd/drbd_bitmap.c struct page **new_pages, *page; new_pages 397 drivers/block/drbd/drbd_bitmap.c new_pages = kzalloc(bytes, GFP_NOIO | __GFP_NOWARN); new_pages 398 drivers/block/drbd/drbd_bitmap.c if (!new_pages) { new_pages 399 drivers/block/drbd/drbd_bitmap.c new_pages = __vmalloc(bytes, new_pages 402 drivers/block/drbd/drbd_bitmap.c if (!new_pages) new_pages 408 drivers/block/drbd/drbd_bitmap.c new_pages[i] = old_pages[i]; new_pages 412 drivers/block/drbd/drbd_bitmap.c bm_free_pages(new_pages + have, i - have); new_pages 413 drivers/block/drbd/drbd_bitmap.c bm_vk_free(new_pages); new_pages 419 drivers/block/drbd/drbd_bitmap.c new_pages[i] = page; new_pages 423 drivers/block/drbd/drbd_bitmap.c new_pages[i] = old_pages[i]; new_pages 429 drivers/block/drbd/drbd_bitmap.c return new_pages; new_pages 592 drivers/gpu/drm/ttm/ttm_page_alloc.c struct list_head new_pages; new_pages 601 drivers/gpu/drm/ttm/ttm_page_alloc.c INIT_LIST_HEAD(&new_pages); new_pages 602 drivers/gpu/drm/ttm/ttm_page_alloc.c r = ttm_alloc_new_pages(&new_pages, pool->gfp_flags, ttm_flags, new_pages 607 drivers/gpu/drm/ttm/ttm_page_alloc.c list_splice(&new_pages, &pool->list); new_pages 613 drivers/gpu/drm/ttm/ttm_page_alloc.c list_for_each_entry(p, &new_pages, lru) { new_pages 616 drivers/gpu/drm/ttm/ttm_page_alloc.c list_splice(&new_pages, &pool->list); new_pages 3319 fs/io_uring.c unsigned long page_limit, cur_pages, new_pages; new_pages 3326 fs/io_uring.c new_pages = cur_pages + nr_pages; new_pages 3327 fs/io_uring.c if (new_pages > page_limit) new_pages 3330 fs/io_uring.c new_pages) != cur_pages); new_pages 478 kernel/trace/ring_buffer.c struct list_head new_pages; /* new pages to add */ new_pages 1324 kernel/trace/ring_buffer.c INIT_LIST_HEAD(&cpu_buffer->new_pages); new_pages 1609 kernel/trace/ring_buffer.c struct list_head *pages = &cpu_buffer->new_pages; new_pages 1674 kernel/trace/ring_buffer.c list_for_each_entry_safe(bpage, tmp, &cpu_buffer->new_pages, new_pages 1768 kernel/trace/ring_buffer.c INIT_LIST_HEAD(&cpu_buffer->new_pages); new_pages 1770 kernel/trace/ring_buffer.c &cpu_buffer->new_pages, cpu)) { new_pages 1823 kernel/trace/ring_buffer.c INIT_LIST_HEAD(&cpu_buffer->new_pages); new_pages 1826 kernel/trace/ring_buffer.c &cpu_buffer->new_pages, cpu_id)) { new_pages 1880 kernel/trace/ring_buffer.c if (list_empty(&cpu_buffer->new_pages)) new_pages 1883 kernel/trace/ring_buffer.c list_for_each_entry_safe(bpage, tmp, &cpu_buffer->new_pages, new_pages 4367 kernel/trace/ring_buffer.c INIT_LIST_HEAD(&cpu_buffer->new_pages);