dpage 1323 drivers/crypto/hifn_795x.c struct page *spage, *dpage; dpage 1344 drivers/crypto/hifn_795x.c dpage = sg_page(t); dpage 1349 drivers/crypto/hifn_795x.c dpage = sg_page(dst); dpage 1355 drivers/crypto/hifn_795x.c hifn_setup_dst_desc(dev, dpage, doff, len, n - len == 0); dpage 686 drivers/dma/nbpfaxi.c struct nbpf_desc_page *dpage = (void *)get_zeroed_page(GFP_KERNEL | GFP_DMA); dpage 695 drivers/dma/nbpfaxi.c if (!dpage) dpage 699 drivers/dma/nbpfaxi.c __func__, NBPF_DESCS_PER_PAGE, NBPF_SEGMENTS_PER_PAGE, sizeof(*dpage)); dpage 701 drivers/dma/nbpfaxi.c for (i = 0, ldesc = dpage->ldesc, hwdesc = dpage->hwdesc; dpage 702 drivers/dma/nbpfaxi.c i < ARRAY_SIZE(dpage->ldesc); dpage 713 drivers/dma/nbpfaxi.c for (i = 0, desc = dpage->desc; dpage 714 drivers/dma/nbpfaxi.c i < ARRAY_SIZE(dpage->desc); dpage 730 drivers/dma/nbpfaxi.c list_add(&dpage->node, &chan->desc_page); dpage 733 drivers/dma/nbpfaxi.c return ARRAY_SIZE(dpage->desc); dpage 1059 drivers/dma/nbpfaxi.c struct nbpf_desc_page *dpage, *tmp; dpage 1068 drivers/dma/nbpfaxi.c list_for_each_entry_safe(dpage, tmp, &chan->desc_page, node) { dpage 1071 drivers/dma/nbpfaxi.c list_del(&dpage->node); dpage 1072 drivers/dma/nbpfaxi.c for (i = 0, ldesc = dpage->ldesc; dpage 1073 drivers/dma/nbpfaxi.c i < ARRAY_SIZE(dpage->ldesc); dpage 1077 drivers/dma/nbpfaxi.c free_page((unsigned long)dpage); dpage 136 drivers/gpu/drm/nouveau/nouveau_dmem.c struct page *dpage, *spage; dpage 142 drivers/gpu/drm/nouveau/nouveau_dmem.c dpage = alloc_page_vma(GFP_HIGHUSER, vmf->vma, vmf->address); dpage 143 drivers/gpu/drm/nouveau/nouveau_dmem.c if (!dpage) dpage 145 drivers/gpu/drm/nouveau/nouveau_dmem.c lock_page(dpage); dpage 147 drivers/gpu/drm/nouveau/nouveau_dmem.c *dma_addr = dma_map_page(dev, dpage, 0, PAGE_SIZE, DMA_BIDIRECTIONAL); dpage 155 drivers/gpu/drm/nouveau/nouveau_dmem.c args->dst[0] = migrate_pfn(page_to_pfn(dpage)) | MIGRATE_PFN_LOCKED; dpage 161 drivers/gpu/drm/nouveau/nouveau_dmem.c __free_page(dpage); dpage 564 drivers/gpu/drm/nouveau/nouveau_dmem.c struct page *dpage, *spage; dpage 570 drivers/gpu/drm/nouveau/nouveau_dmem.c dpage = nouveau_dmem_page_alloc_locked(drm); dpage 571 drivers/gpu/drm/nouveau/nouveau_dmem.c if (!dpage) dpage 579 drivers/gpu/drm/nouveau/nouveau_dmem.c nouveau_dmem_page_addr(dpage), NOUVEAU_APER_HOST, dpage 583 drivers/gpu/drm/nouveau/nouveau_dmem.c return migrate_pfn(page_to_pfn(dpage)) | MIGRATE_PFN_LOCKED; dpage 588 drivers/gpu/drm/nouveau/nouveau_dmem.c nouveau_dmem_page_free_locked(drm, dpage); dpage 1337 drivers/net/ethernet/3com/typhoon.c void *dpage; dpage 1358 drivers/net/ethernet/3com/typhoon.c dpage = pci_alloc_consistent(pdev, PAGE_SIZE, &dpage_dma); dpage 1359 drivers/net/ethernet/3com/typhoon.c if(!dpage) { dpage 1423 drivers/net/ethernet/3com/typhoon.c dpage, len, dpage 1464 drivers/net/ethernet/3com/typhoon.c pci_free_consistent(pdev, PAGE_SIZE, dpage, dpage_dma); dpage 4080 drivers/scsi/st.c struct page *dpage = st_bp->reserved_pages[dst_seg]; dpage 4084 drivers/scsi/st.c memmove(page_address(dpage) + dst_offset, dpage 168 fs/f2fs/acl.c struct page *dpage) dpage 178 fs/f2fs/acl.c retval = f2fs_getxattr(inode, name_index, "", NULL, 0, dpage); dpage 184 fs/f2fs/acl.c retval, dpage); dpage 334 fs/f2fs/acl.c struct page *dpage) dpage 346 fs/f2fs/acl.c p = __f2fs_get_acl(dir, ACL_TYPE_DEFAULT, dpage); dpage 384 fs/f2fs/acl.c struct page *dpage) dpage 389 fs/f2fs/acl.c error = f2fs_acl_create(dir, &inode->i_mode, &default_acl, &acl, dpage); dpage 45 fs/f2fs/acl.h struct page *ipage, struct page *dpage) dpage 471 fs/f2fs/dir.c struct page *dpage) dpage 493 fs/f2fs/dir.c err = f2fs_init_acl(inode, dir, page, dpage); dpage 2980 fs/f2fs/f2fs.h const struct qstr *orig_name, struct page *dpage); dpage 144 fs/f2fs/xattr.h size_t buffer_size, struct page *dpage) dpage 104 fs/nilfs2/page.c struct page *spage = sbh->b_page, *dpage = dbh->b_page; dpage 108 fs/nilfs2/page.c kaddr1 = kmap_atomic(dpage); dpage 125 fs/nilfs2/page.c SetPageUptodate(dpage); dpage 127 fs/nilfs2/page.c ClearPageUptodate(dpage); dpage 129 fs/nilfs2/page.c SetPageMappedToDisk(dpage); dpage 131 fs/nilfs2/page.c ClearPageMappedToDisk(dpage); dpage 255 fs/nilfs2/page.c struct page *page = pvec.pages[i], *dpage; dpage 261 fs/nilfs2/page.c dpage = grab_cache_page(dmap, page->index); dpage 262 fs/nilfs2/page.c if (unlikely(!dpage)) { dpage 272 fs/nilfs2/page.c nilfs_copy_page(dpage, page, 1); dpage 273 fs/nilfs2/page.c __set_page_dirty_nobuffers(dpage); dpage 275 fs/nilfs2/page.c unlock_page(dpage); dpage 276 fs/nilfs2/page.c put_page(dpage); dpage 309 fs/nilfs2/page.c struct page *page = pvec.pages[i], *dpage; dpage 313 fs/nilfs2/page.c dpage = find_lock_page(dmap, offset); dpage 314 fs/nilfs2/page.c if (dpage) { dpage 316 fs/nilfs2/page.c WARN_ON(PageDirty(dpage)); dpage 317 fs/nilfs2/page.c nilfs_copy_page(dpage, page, 0); dpage 318 fs/nilfs2/page.c unlock_page(dpage); dpage 319 fs/nilfs2/page.c put_page(dpage);