pde 356 arch/ia64/hp/common/sba_iommu.c u32 pde = ((u32)((*pptr >> (63)) & 0x1)); pde 357 arch/ia64/hp/common/sba_iommu.c if ((rval & 0x1) ^ pde) pde 61 arch/m68k/kernel/bootinfo_proc.c struct proc_dir_entry *pde; pde 70 arch/m68k/kernel/bootinfo_proc.c pde = proc_create_data("bootinfo", 0400, NULL, &bootinfo_fops, NULL); pde 71 arch/m68k/kernel/bootinfo_proc.c if (!pde) { pde 115 arch/parisc/kernel/pdt.c static void report_mem_err(unsigned long pde) pde 121 arch/parisc/kernel/pdt.c addr = pde & PDT_ADDR_PHYS_MASK; pde 132 arch/parisc/kernel/pdt.c pde & PDT_ADDR_PERM_ERR ? "permanent ":"", pde 133 arch/parisc/kernel/pdt.c pde & PDT_ADDR_SINGLE_ERR ? "single":"multi"); pde 321 arch/parisc/kernel/pdt.c unsigned long pde = *bad_mem_ptr++; pde 323 arch/parisc/kernel/pdt.c report_mem_err(pde); pde 326 arch/parisc/kernel/pdt.c if ((pde & PDT_ADDR_PERM_ERR) || pde 327 arch/parisc/kernel/pdt.c ((pde & PDT_ADDR_SINGLE_ERR) == 0)) pde 328 arch/parisc/kernel/pdt.c memory_failure(pde >> PAGE_SHIFT, 0); pde 331 arch/parisc/kernel/pdt.c pfn_to_page(pde >> PAGE_SHIFT), 0); pde 336 arch/parisc/kernel/pdt.c pde & PDT_ADDR_PHYS_MASK); pde 51 arch/powerpc/kernel/proc_powerpc.c struct proc_dir_entry *pde; pde 53 arch/powerpc/kernel/proc_powerpc.c pde = proc_create_data("powerpc/systemcfg", S_IFREG | 0444, NULL, pde 55 arch/powerpc/kernel/proc_powerpc.c if (!pde) pde 57 arch/powerpc/kernel/proc_powerpc.c proc_set_size(pde, PAGE_SIZE); pde 1104 arch/s390/kernel/debug.c struct dentry *pde; pde 1116 arch/s390/kernel/debug.c pde = debugfs_create_file(view->name, mode, id->debugfs_root_entry, pde 1129 arch/s390/kernel/debug.c id->debugfs_entries[i] = pde; pde 1133 arch/s390/kernel/debug.c debugfs_remove(pde); pde 529 drivers/block/ps3vram.c struct proc_dir_entry *pde; pde 531 drivers/block/ps3vram.c pde = proc_create_single_data(DEVICE_NAME, 0444, NULL, pde 533 drivers/block/ps3vram.c if (!pde) pde 502 drivers/char/toshiba.c struct proc_dir_entry *pde; pde 504 drivers/char/toshiba.c pde = proc_create_single("toshiba", 0, NULL, proc_toshiba_show); pde 505 drivers/char/toshiba.c if (!pde) { pde 1196 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c uint64_t pde, pt, flags; pde 1204 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c pde = (entry - parent->entries) * 8; pde 1205 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c return vm->update_funcs->update(params, bo, pde, pt, 1, 0, flags); pde 494 drivers/gpu/drm/i915/gvt/gtt.c DEFINE_PPGTT_GMA_TO_INDEX(gen8, pde, (gma >> 21 & 0x1ff)); pde 225 drivers/gpu/drm/i915/i915_gem_gtt.c u64 pde = _PAGE_PRESENT | _PAGE_RW; pde 226 drivers/gpu/drm/i915/i915_gem_gtt.c pde |= addr; pde 228 drivers/gpu/drm/i915/i915_gem_gtt.c pde |= PPAT_CACHED_PDE; pde 230 drivers/gpu/drm/i915/i915_gem_gtt.c pde |= PPAT_UNCACHED; pde 231 drivers/gpu/drm/i915/i915_gem_gtt.c return pde; pde 939 drivers/gpu/drm/i915/i915_gem_gtt.c void **pde = pd->entry; pde 942 drivers/gpu/drm/i915/i915_gem_gtt.c if (!*pde) pde 945 drivers/gpu/drm/i915/i915_gem_gtt.c __gen8_ppgtt_cleanup(vm, *pde, GEN8_PDES, lvl - 1); pde 946 drivers/gpu/drm/i915/i915_gem_gtt.c } while (pde++, --count); pde 1416 drivers/gpu/drm/i915/i915_gem_gtt.c struct i915_page_directory *pde; pde 1418 drivers/gpu/drm/i915/i915_gem_gtt.c pde = alloc_pd(vm); pde 1419 drivers/gpu/drm/i915/i915_gem_gtt.c if (IS_ERR(pde)) pde 1420 drivers/gpu/drm/i915/i915_gem_gtt.c return PTR_ERR(pde); pde 1422 drivers/gpu/drm/i915/i915_gem_gtt.c fill_px(pde, vm->scratch[1].encode); pde 1423 drivers/gpu/drm/i915/i915_gem_gtt.c set_pd_entry(pd, idx, pde); pde 1424 drivers/gpu/drm/i915/i915_gem_gtt.c atomic_inc(px_used(pde)); /* keep pinned */ pde 1543 drivers/gpu/drm/i915/i915_gem_gtt.c const unsigned int pde, pde 1548 drivers/gpu/drm/i915/i915_gem_gtt.c ppgtt->pd_addr + pde); pde 1610 drivers/gpu/drm/i915/i915_gem_gtt.c unsigned int pde = first_entry / GEN6_PTES; pde 1616 drivers/gpu/drm/i915/i915_gem_gtt.c i915_pt_entry(ppgtt->base.pd, pde++); pde 1693 drivers/gpu/drm/i915/i915_gem_gtt.c unsigned int pde; pde 1700 drivers/gpu/drm/i915/i915_gem_gtt.c gen6_for_each_pde(pt, pd, start, length, pde) { pde 1717 drivers/gpu/drm/i915/i915_gem_gtt.c if (pd->entry[pde] == &vm->scratch[1]) { pde 1718 drivers/gpu/drm/i915/i915_gem_gtt.c pd->entry[pde] = pt; pde 1721 drivers/gpu/drm/i915/i915_gem_gtt.c gen6_write_pde(ppgtt, pde, pt); pde 1726 drivers/gpu/drm/i915/i915_gem_gtt.c pt = pd->entry[pde]; pde 1781 drivers/gpu/drm/i915/i915_gem_gtt.c u32 pde; pde 1783 drivers/gpu/drm/i915/i915_gem_gtt.c gen6_for_all_pdes(pt, pd, pde) pde 1824 drivers/gpu/drm/i915/i915_gem_gtt.c unsigned int pde; pde 1829 drivers/gpu/drm/i915/i915_gem_gtt.c gen6_for_all_pdes(pt, ppgtt->base.pd, pde) pde 1830 drivers/gpu/drm/i915/i915_gem_gtt.c gen6_write_pde(ppgtt, pde, pt); pde 1845 drivers/gpu/drm/i915/i915_gem_gtt.c unsigned int pde; pde 1851 drivers/gpu/drm/i915/i915_gem_gtt.c gen6_for_all_pdes(pt, ppgtt->base.pd, pde) { pde 1856 drivers/gpu/drm/i915/i915_gem_gtt.c pd->entry[pde] = scratch; pde 32 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c kvfree(pgt->pde); pde 62 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c pgt->pde = kvcalloc(pten, sizeof(*pgt->pde), GFP_KERNEL); pde 63 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (!pgt->pde) { pde 160 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c pgd->pde[pdei] = NVKM_VMM_PDE_SPARSE; pde 163 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c pgd->pde[pdei] = NULL; pde 170 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c func->pde(vmm, pgd, pdei); pde 171 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c pgd->pde[pdei] = NULL; pde 177 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c func->pde(vmm, pgd, pdei); pde 386 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c pgt->pde[ptei++] = NVKM_VMM_PDE_SPARSE; pde 398 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c memset(&pt->pde[ptei], 0x00, sizeof(pt->pde[0]) * ptes); pde 417 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct nvkm_vmm_pt *pgt = pgd->pde[pdei]; pde 478 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c it->desc[it->lvl].func->pde(it->vmm, pgd, pdei); pde 487 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c struct nvkm_vmm_pt *pgt = pgd->pde[pdei]; pde 496 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c pgd->pde[pdei] = pgt; pde 544 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (ref && NVKM_VMM_PDE_INVALID(pgd->pde[pdei])) { pde 548 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c it.pt[it.lvl - 1] = pgt = pgd->pde[pdei]; pde 35 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h #define NVKM_VMM_PDE_INVALID(pde) IS_ERR_OR_NULL(pde) pde 36 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h #define NVKM_VMM_PDE_SPARSED(pde) IS_ERR(pde) pde 38 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h struct nvkm_vmm_pt **pde; pde 65 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h nvkm_vmm_pde_func pde; pde 108 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c struct nvkm_vmm_pt *pgt = pgd->pde[pdei]; pde 149 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c .pde = gf100_vmm_pgd_pde, pde 64 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c .pde = gf100_vmm_pgd_pde, pde 232 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c struct nvkm_vmm_pt *pgt = pgd->pde[pdei]; pde 265 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c .pde = gp100_vmm_pd0_pde, pde 272 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c struct nvkm_vmm_pt *pgt = pgd->pde[pdei]; pde 288 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c .pde = gp100_vmm_pd1_pde, pde 151 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c if (!nv50_vmm_pde(vmm, pgd->pde[pdei], &data)) pde 163 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c .pde = nv50_vmm_pgd_pde, pde 354 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c if (!nv50_vmm_pde(vmm, vmm->pd->pde[pdei], &data)) { pde 670 drivers/gpu/drm/radeon/radeon_vm.c uint64_t pde, pt; pde 680 drivers/gpu/drm/radeon/radeon_vm.c pde = pd_addr + pt_idx * 8; pde 681 drivers/gpu/drm/radeon/radeon_vm.c if (((last_pde + 8 * count) != pde) || pde 691 drivers/gpu/drm/radeon/radeon_vm.c last_pde = pde; pde 896 drivers/iommu/amd_iommu.c size_t size, u16 domid, int pde) pde 922 drivers/iommu/amd_iommu.c if (pde) /* PDE bit - we want to flush everything, not only the PTEs */ pde 1267 drivers/iommu/amd_iommu.c u64 address, size_t size, int pde) pde 1273 drivers/iommu/amd_iommu.c build_inv_iommu_pages(&cmd, address, size, domain->id, pde); pde 57 drivers/iommu/intel-pasid.h static inline bool pasid_pde_is_present(struct pasid_dir_entry *pde) pde 59 drivers/iommu/intel-pasid.h return READ_ONCE(pde->val) & PASID_PTE_PRESENT; pde 64 drivers/iommu/intel-pasid.h get_pasid_table_from_pde(struct pasid_dir_entry *pde) pde 66 drivers/iommu/intel-pasid.h if (!pasid_pde_is_present(pde)) pde 69 drivers/iommu/intel-pasid.h return phys_to_virt(READ_ONCE(pde->val) & PDE_PFN_MASK); pde 162 drivers/iommu/tegra-smmu.c static dma_addr_t smmu_pde_to_dma(struct tegra_smmu *smmu, u32 pde) pde 164 drivers/iommu/tegra-smmu.c return (dma_addr_t)(pde & smmu->pfn_mask) << 12; pde 569 drivers/iommu/tegra-smmu.c unsigned int pde = iova_pd_index(iova); pde 572 drivers/iommu/tegra-smmu.c if (!as->pts[pde]) { pde 594 drivers/iommu/tegra-smmu.c as->pts[pde] = page; pde 603 drivers/iommu/tegra-smmu.c *dmap = smmu_pde_to_dma(smmu, pd[pde]); pde 606 drivers/iommu/tegra-smmu.c return tegra_smmu_pte_offset(as->pts[pde], iova); pde 618 drivers/iommu/tegra-smmu.c unsigned int pde = iova_pd_index(iova); pde 619 drivers/iommu/tegra-smmu.c struct page *page = as->pts[pde]; pde 625 drivers/iommu/tegra-smmu.c if (--as->count[pde] == 0) { pde 628 drivers/iommu/tegra-smmu.c dma_addr_t pte_dma = smmu_pde_to_dma(smmu, pd[pde]); pde 634 drivers/iommu/tegra-smmu.c as->pts[pde] = NULL; pde 1135 drivers/net/ppp/pppoe.c struct proc_dir_entry *pde; pde 1139 drivers/net/ppp/pppoe.c pde = proc_create_net("pppoe", 0444, net->proc_net, pde 1142 drivers/net/ppp/pppoe.c if (!pde) pde 161 drivers/net/wireless/ath/dfs_pri_detector.c static struct pulse_elem *pulse_queue_get_tail(struct pri_detector *pde) pde 163 drivers/net/wireless/ath/dfs_pri_detector.c struct list_head *l = &pde->pulses; pde 169 drivers/net/wireless/ath/dfs_pri_detector.c static bool pulse_queue_dequeue(struct pri_detector *pde) pde 171 drivers/net/wireless/ath/dfs_pri_detector.c struct pulse_elem *p = pulse_queue_get_tail(pde); pde 174 drivers/net/wireless/ath/dfs_pri_detector.c pde->count--; pde 178 drivers/net/wireless/ath/dfs_pri_detector.c return (pde->count > 0); pde 182 drivers/net/wireless/ath/dfs_pri_detector.c static void pulse_queue_check_window(struct pri_detector *pde) pde 188 drivers/net/wireless/ath/dfs_pri_detector.c if (pde->count < 2) pde 191 drivers/net/wireless/ath/dfs_pri_detector.c if (pde->last_ts <= pde->window_size) pde 194 drivers/net/wireless/ath/dfs_pri_detector.c min_valid_ts = pde->last_ts - pde->window_size; pde 195 drivers/net/wireless/ath/dfs_pri_detector.c while ((p = pulse_queue_get_tail(pde)) != NULL) { pde 198 drivers/net/wireless/ath/dfs_pri_detector.c pulse_queue_dequeue(pde); pde 202 drivers/net/wireless/ath/dfs_pri_detector.c static bool pulse_queue_enqueue(struct pri_detector *pde, u64 ts) pde 216 drivers/net/wireless/ath/dfs_pri_detector.c list_add(&p->head, &pde->pulses); pde 217 drivers/net/wireless/ath/dfs_pri_detector.c pde->count++; pde 218 drivers/net/wireless/ath/dfs_pri_detector.c pde->last_ts = ts; pde 219 drivers/net/wireless/ath/dfs_pri_detector.c pulse_queue_check_window(pde); pde 220 drivers/net/wireless/ath/dfs_pri_detector.c if (pde->count >= pde->max_count) pde 221 drivers/net/wireless/ath/dfs_pri_detector.c pulse_queue_dequeue(pde); pde 225 drivers/net/wireless/ath/dfs_pri_detector.c static bool pseq_handler_create_sequences(struct pri_detector *pde, pde 229 drivers/net/wireless/ath/dfs_pri_detector.c list_for_each_entry(p, &pde->pulses, head) { pde 236 drivers/net/wireless/ath/dfs_pri_detector.c if (delta_ts < pde->rs->pri_min) pde 240 drivers/net/wireless/ath/dfs_pri_detector.c if (delta_ts > pde->rs->pri_max) pde 249 drivers/net/wireless/ath/dfs_pri_detector.c ps.pri = GET_PRI_TO_USE(pde->rs->pri_min, pde 250 drivers/net/wireless/ath/dfs_pri_detector.c pde->rs->pri_max, ts - p->ts); pde 251 drivers/net/wireless/ath/dfs_pri_detector.c ps.dur = ps.pri * (pde->rs->ppb - 1) pde 252 drivers/net/wireless/ath/dfs_pri_detector.c + 2 * pde->rs->max_pri_tolerance; pde 258 drivers/net/wireless/ath/dfs_pri_detector.c list_for_each_entry_continue(p2, &pde->pulses, head) { pde 265 drivers/net/wireless/ath/dfs_pri_detector.c pde->rs->max_pri_tolerance); pde 298 drivers/net/wireless/ath/dfs_pri_detector.c list_add(&new_ps->head, &pde->sequences); pde 305 drivers/net/wireless/ath/dfs_pri_detector.c pseq_handler_add_to_existing_seqs(struct pri_detector *pde, u64 ts) pde 309 drivers/net/wireless/ath/dfs_pri_detector.c list_for_each_entry_safe(ps, ps2, &pde->sequences, head) { pde 322 drivers/net/wireless/ath/dfs_pri_detector.c pde->rs->max_pri_tolerance); pde 337 drivers/net/wireless/ath/dfs_pri_detector.c pseq_handler_check_detection(struct pri_detector *pde) pde 341 drivers/net/wireless/ath/dfs_pri_detector.c if (list_empty(&pde->sequences)) pde 344 drivers/net/wireless/ath/dfs_pri_detector.c list_for_each_entry(ps, &pde->sequences, head) { pde 350 drivers/net/wireless/ath/dfs_pri_detector.c if ((ps->count >= pde->rs->ppb_thresh) && pde 351 drivers/net/wireless/ath/dfs_pri_detector.c (ps->count * pde->rs->num_pri >= ps->count_falses)) pde 359 drivers/net/wireless/ath/dfs_pri_detector.c static void pri_detector_reset(struct pri_detector *pde, u64 ts) pde 363 drivers/net/wireless/ath/dfs_pri_detector.c list_for_each_entry_safe(ps, ps0, &pde->sequences, head) { pde 367 drivers/net/wireless/ath/dfs_pri_detector.c list_for_each_entry_safe(p, p0, &pde->pulses, head) { pde 371 drivers/net/wireless/ath/dfs_pri_detector.c pde->count = 0; pde 372 drivers/net/wireless/ath/dfs_pri_detector.c pde->last_ts = ts; pde 242 drivers/parisc/sba_iommu.c u32 pde = ((u32) (((char *)pptr)[7])) << 24; pde 243 drivers/parisc/sba_iommu.c if ((rval ^ pde) & 0x80000000) pde 347 drivers/perf/xgene_pmu.c XGENE_PMU_EVENT_ATTR(pde-cmd-sent, 0x08), pde 243 drivers/s390/block/dasd.c struct dentry *pde; pde 247 drivers/s390/block/dasd.c pde = debugfs_create_dir(name, base_dentry); pde 248 drivers/s390/block/dasd.c if (!pde || IS_ERR(pde)) pde 250 drivers/s390/block/dasd.c return pde; pde 1102 drivers/s390/block/dasd.c struct dentry *pde; pde 1109 drivers/s390/block/dasd.c pde = debugfs_create_file("statistics", mode, base_dentry, pde 1111 drivers/s390/block/dasd.c if (pde && !IS_ERR(pde)) pde 1112 drivers/s390/block/dasd.c profile->dentry = pde; pde 1133 drivers/s390/block/dasd.c struct dentry *pde; pde 1136 drivers/s390/block/dasd.c pde = debugfs_create_dir("dasd", NULL); pde 1137 drivers/s390/block/dasd.c if (!pde || IS_ERR(pde)) pde 1139 drivers/s390/block/dasd.c dasd_debugfs_root_entry = pde; pde 1140 drivers/s390/block/dasd.c pde = debugfs_create_dir("global", dasd_debugfs_root_entry); pde 1141 drivers/s390/block/dasd.c if (!pde || IS_ERR(pde)) pde 1143 drivers/s390/block/dasd.c dasd_debugfs_global_entry = pde; pde 1218 drivers/s390/block/dasd.c struct dentry *pde; pde 1225 drivers/s390/block/dasd.c pde = debugfs_create_file("host_access_list", mode, base_dentry, pde 1227 drivers/s390/block/dasd.c if (pde && !IS_ERR(pde)) pde 1228 drivers/s390/block/dasd.c device->hosts_dentry = pde; pde 727 drivers/scsi/esas2r/esas2r_main.c struct proc_dir_entry *pde; pde 729 drivers/scsi/esas2r/esas2r_main.c pde = proc_create(ATTONODE_NAME, 0, pde 733 drivers/scsi/esas2r/esas2r_main.c if (!pde) { pde 453 drivers/scsi/scsi_proc.c struct proc_dir_entry *pde; pde 459 drivers/scsi/scsi_proc.c pde = proc_create("scsi/scsi", 0, NULL, &proc_scsi_operations); pde 460 drivers/scsi/scsi_proc.c if (!pde) pde 239 drivers/usb/gadget/udc/at91_udc.c udc->pde = proc_create_single_data(debug_filename, 0, NULL, pde 245 drivers/usb/gadget/udc/at91_udc.c if (udc->pde) pde 140 drivers/usb/gadget/udc/at91_udc.h struct proc_dir_entry *pde; pde 130 drivers/usb/gadget/udc/lpc32xx_udc.c struct dentry *pde; pde 542 drivers/usb/gadget/udc/lpc32xx_udc.c udc->pde = debugfs_create_file(debug_filename, 0, NULL, udc, &proc_ops); pde 547 drivers/usb/gadget/udc/lpc32xx_udc.c debugfs_remove(udc->pde); pde 568 fs/ext2/dir.c ext2_dirent * pde = NULL; pde 579 fs/ext2/dir.c pde = de; pde 582 fs/ext2/dir.c if (pde) pde 583 fs/ext2/dir.c from = (char*)pde - (char*)page_address(page); pde 588 fs/ext2/dir.c if (pde) pde 589 fs/ext2/dir.c pde->rec_len = ext2_rec_len_to_disk(to - from); pde 2456 fs/ext4/namei.c struct ext4_dir_entry_2 *de, *pde; pde 2461 fs/ext4/namei.c pde = NULL; pde 2468 fs/ext4/namei.c if (pde) pde 2469 fs/ext4/namei.c pde->rec_len = ext4_rec_len_to_disk( pde 2470 fs/ext4/namei.c ext4_rec_len_from_disk(pde->rec_len, pde 2481 fs/ext4/namei.c pde = de; pde 544 fs/nilfs2/dir.c struct nilfs_dir_entry *de, *pde = NULL; pde 558 fs/nilfs2/dir.c pde = de; pde 561 fs/nilfs2/dir.c if (pde) pde 562 fs/nilfs2/dir.c from = (char *)pde - (char *)page_address(page); pde 566 fs/nilfs2/dir.c if (pde) pde 567 fs/nilfs2/dir.c pde->rec_len = nilfs_rec_len_to_disk(to - from); pde 1132 fs/ocfs2/dir.c struct ocfs2_dir_entry *de, *pde; pde 1140 fs/ocfs2/dir.c pde = NULL; pde 1156 fs/ocfs2/dir.c if (pde) pde 1157 fs/ocfs2/dir.c le16_add_cpu(&pde->rec_len, pde 1165 fs/ocfs2/dir.c pde = de; pde 37 fs/proc/generic.c void pde_free(struct proc_dir_entry *pde) pde 39 fs/proc/generic.c if (S_ISLNK(pde->mode)) pde 40 fs/proc/generic.c kfree(pde->data); pde 41 fs/proc/generic.c if (pde->name != pde->inline_name) pde 42 fs/proc/generic.c kfree(pde->name); pde 43 fs/proc/generic.c kmem_cache_free(proc_dir_entry_cache, pde); pde 644 fs/proc/generic.c void pde_put(struct proc_dir_entry *pde) pde 646 fs/proc/generic.c if (refcount_dec_and_test(&pde->refcnt)) { pde 647 fs/proc/generic.c proc_free_inum(pde->low_ino); pde 648 fs/proc/generic.c pde_free(pde); pde 764 fs/proc/generic.c struct proc_dir_entry *pde = PDE(file_inode(f)); pde 768 fs/proc/generic.c if (!pde->write) pde 775 fs/proc/generic.c ret = pde->write(f, buf, size); pde 68 fs/proc/inode.c ei->pde = NULL; pde 129 fs/proc/inode.c static inline int use_pde(struct proc_dir_entry *pde) pde 131 fs/proc/inode.c return likely(atomic_inc_unless_negative(&pde->in_use)); pde 134 fs/proc/inode.c static void unuse_pde(struct proc_dir_entry *pde) pde 136 fs/proc/inode.c if (unlikely(atomic_dec_return(&pde->in_use) == BIAS)) pde 137 fs/proc/inode.c complete(pde->pde_unload_completion); pde 141 fs/proc/inode.c static void close_pdeo(struct proc_dir_entry *pde, struct pde_opener *pdeo) pde 157 fs/proc/inode.c spin_unlock(&pde->pde_unload_lock); pde 164 fs/proc/inode.c spin_unlock(&pde->pde_unload_lock); pde 166 fs/proc/inode.c pde->proc_fops->release(file_inode(file), file); pde 167 fs/proc/inode.c spin_lock(&pde->pde_unload_lock); pde 171 fs/proc/inode.c spin_unlock(&pde->pde_unload_lock); pde 200 fs/proc/inode.c struct proc_dir_entry *pde = PDE(file_inode(file)); pde 202 fs/proc/inode.c if (use_pde(pde)) { pde 205 fs/proc/inode.c llseek = pde->proc_fops->llseek; pde 209 fs/proc/inode.c unuse_pde(pde); pde 216 fs/proc/inode.c struct proc_dir_entry *pde = PDE(file_inode(file)); pde 218 fs/proc/inode.c if (use_pde(pde)) { pde 221 fs/proc/inode.c read = pde->proc_fops->read; pde 224 fs/proc/inode.c unuse_pde(pde); pde 231 fs/proc/inode.c struct proc_dir_entry *pde = PDE(file_inode(file)); pde 233 fs/proc/inode.c if (use_pde(pde)) { pde 236 fs/proc/inode.c write = pde->proc_fops->write; pde 239 fs/proc/inode.c unuse_pde(pde); pde 246 fs/proc/inode.c struct proc_dir_entry *pde = PDE(file_inode(file)); pde 248 fs/proc/inode.c if (use_pde(pde)) { pde 251 fs/proc/inode.c poll = pde->proc_fops->poll; pde 254 fs/proc/inode.c unuse_pde(pde); pde 261 fs/proc/inode.c struct proc_dir_entry *pde = PDE(file_inode(file)); pde 263 fs/proc/inode.c if (use_pde(pde)) { pde 266 fs/proc/inode.c ioctl = pde->proc_fops->unlocked_ioctl; pde 269 fs/proc/inode.c unuse_pde(pde); pde 277 fs/proc/inode.c struct proc_dir_entry *pde = PDE(file_inode(file)); pde 279 fs/proc/inode.c if (use_pde(pde)) { pde 282 fs/proc/inode.c compat_ioctl = pde->proc_fops->compat_ioctl; pde 285 fs/proc/inode.c unuse_pde(pde); pde 293 fs/proc/inode.c struct proc_dir_entry *pde = PDE(file_inode(file)); pde 295 fs/proc/inode.c if (use_pde(pde)) { pde 298 fs/proc/inode.c mmap = pde->proc_fops->mmap; pde 301 fs/proc/inode.c unuse_pde(pde); pde 311 fs/proc/inode.c struct proc_dir_entry *pde = PDE(file_inode(file)); pde 314 fs/proc/inode.c if (use_pde(pde)) { pde 317 fs/proc/inode.c get_area = pde->proc_fops->get_unmapped_area; pde 327 fs/proc/inode.c unuse_pde(pde); pde 334 fs/proc/inode.c struct proc_dir_entry *pde = PDE(inode); pde 351 fs/proc/inode.c if (!use_pde(pde)) pde 354 fs/proc/inode.c release = pde->proc_fops->release; pde 363 fs/proc/inode.c open = pde->proc_fops->open; pde 373 fs/proc/inode.c spin_lock(&pde->pde_unload_lock); pde 374 fs/proc/inode.c list_add(&pdeo->lh, &pde->pde_openers); pde 375 fs/proc/inode.c spin_unlock(&pde->pde_unload_lock); pde 381 fs/proc/inode.c unuse_pde(pde); pde 387 fs/proc/inode.c struct proc_dir_entry *pde = PDE(inode); pde 389 fs/proc/inode.c spin_lock(&pde->pde_unload_lock); pde 390 fs/proc/inode.c list_for_each_entry(pdeo, &pde->pde_openers, lh) { pde 392 fs/proc/inode.c close_pdeo(pde, pdeo); pde 396 fs/proc/inode.c spin_unlock(&pde->pde_unload_lock); pde 438 fs/proc/inode.c struct proc_dir_entry *pde = PDE(inode); pde 439 fs/proc/inode.c if (!use_pde(pde)) pde 441 fs/proc/inode.c set_delayed_call(done, proc_put_link, pde); pde 442 fs/proc/inode.c return pde->data; pde 456 fs/proc/inode.c PROC_I(inode)->pde = de; pde 74 fs/proc/internal.h void pde_free(struct proc_dir_entry *pde); pde 88 fs/proc/internal.h struct proc_dir_entry *pde; pde 106 fs/proc/internal.h return PROC_I(inode)->pde; pde 183 fs/proc/internal.h static inline struct proc_dir_entry *pde_get(struct proc_dir_entry *pde) pde 185 fs/proc/internal.h refcount_inc(&pde->refcnt); pde 186 fs/proc/internal.h return pde; pde 190 fs/proc/internal.h static inline bool is_empty_pde(const struct proc_dir_entry *pde) pde 192 fs/proc/internal.h return S_ISDIR(pde->mode) && !pde->proc_iops; pde 32 fs/proc/proc_net.c static inline struct net *PDE_NET(struct proc_dir_entry *pde) pde 34 fs/proc/proc_net.c return pde->parent->data; pde 52 fs/proc/proc_net.c static void pde_force_lookup(struct proc_dir_entry *pde) pde 55 fs/proc/proc_net.c pde->proc_dops = &proc_net_dentry_ops; pde 503 fs/ufs/dir.c struct ufs_dir_entry *pde = NULL; pde 521 fs/ufs/dir.c pde = de; pde 524 fs/ufs/dir.c if (pde) pde 525 fs/ufs/dir.c from = (char*)pde - (char*)page_address(page); pde 531 fs/ufs/dir.c if (pde) pde 532 fs/ufs/dir.c pde->d_reclen = cpu_to_fs16(sb, to - from); pde 140 ipc/util.c struct proc_dir_entry *pde; pde 151 ipc/util.c pde = proc_create_data(path, pde 156 ipc/util.c if (!pde) pde 3005 net/ipv4/igmp.c struct proc_dir_entry *pde; pde 3008 net/ipv4/igmp.c pde = proc_create_net("igmp", 0444, net->proc_net, &igmp_mc_seq_ops, pde 3010 net/ipv4/igmp.c if (!pde) pde 3012 net/ipv4/igmp.c pde = proc_create_net("mcfilter", 0444, net->proc_net, pde 3014 net/ipv4/igmp.c if (!pde) pde 51 net/ipv4/netfilter/ipt_CLUSTERIP.c struct proc_dir_entry *pde; /* proc dir entry */ pde 128 net/ipv4/netfilter/ipt_CLUSTERIP.c proc_remove(c->pde); pde 159 net/ipv4/netfilter/ipt_CLUSTERIP.c if (!c->pde) pde 281 net/ipv4/netfilter/ipt_CLUSTERIP.c c->pde = proc_create_data(buffer, 0600, pde 285 net/ipv4/netfilter/ipt_CLUSTERIP.c if (!c->pde) { pde 365 net/ipv4/route.c struct proc_dir_entry *pde; pde 367 net/ipv4/route.c pde = proc_create("rt_cache", 0444, net->proc_net, pde 369 net/ipv4/route.c if (!pde) pde 372 net/ipv4/route.c pde = proc_create("rt_cache", 0444, pde 374 net/ipv4/route.c if (!pde) pde 378 net/ipv4/route.c pde = proc_create_single("rt_acct", 0, net->proc_net, pde 380 net/ipv4/route.c if (!pde) pde 1637 net/l2tp/l2tp_ppp.c struct proc_dir_entry *pde; pde 1640 net/l2tp/l2tp_ppp.c pde = proc_create_net("pppol2tp", 0444, net->proc_net, pde 1642 net/l2tp/l2tp_ppp.c if (!pde) { pde 466 net/netfilter/nf_conntrack_standalone.c struct proc_dir_entry *pde; pde 470 net/netfilter/nf_conntrack_standalone.c pde = proc_create_net("nf_conntrack", 0440, net->proc_net, &ct_seq_ops, pde 472 net/netfilter/nf_conntrack_standalone.c if (!pde) pde 478 net/netfilter/nf_conntrack_standalone.c proc_set_user(pde, root_uid, root_gid); pde 480 net/netfilter/nf_conntrack_standalone.c pde = proc_create_net("nf_conntrack", 0444, net->proc_net_stat, pde 482 net/netfilter/nf_conntrack_standalone.c if (!pde) pde 131 net/netfilter/xt_hashlimit.c struct proc_dir_entry *pde; pde 341 net/netfilter/xt_hashlimit.c hinfo->pde = proc_create_seq_data(name, 0, pde 345 net/netfilter/xt_hashlimit.c if (hinfo->pde == NULL) { pde 330 net/netfilter/xt_recent.c struct proc_dir_entry *pde; pde 407 net/netfilter/xt_recent.c pde = proc_create_data(t->name, ip_list_perms, recent_net->xt_recent, pde 409 net/netfilter/xt_recent.c if (pde == NULL) { pde 414 net/netfilter/xt_recent.c proc_set_user(pde, uid, gid); pde 311 tools/testing/selftests/kvm/lib/x86_64/processor.c struct pageDirectoryEntry *pde; pde 312 tools/testing/selftests/kvm/lib/x86_64/processor.c pde = addr_gpa2hva(vm, pdpe[index[2]].address * vm->page_size); pde 313 tools/testing/selftests/kvm/lib/x86_64/processor.c if (!pde[index[1]].present) { pde 314 tools/testing/selftests/kvm/lib/x86_64/processor.c pde[index[1]].address = vm_phy_page_alloc(vm, pde 317 tools/testing/selftests/kvm/lib/x86_64/processor.c pde[index[1]].writable = true; pde 318 tools/testing/selftests/kvm/lib/x86_64/processor.c pde[index[1]].present = true; pde 323 tools/testing/selftests/kvm/lib/x86_64/processor.c pte = addr_gpa2hva(vm, pde[index[1]].address * vm->page_size); pde 347 tools/testing/selftests/kvm/lib/x86_64/processor.c struct pageDirectoryEntry *pde, *pde_start; pde 388 tools/testing/selftests/kvm/lib/x86_64/processor.c pde = &pde_start[n3]; pde 389 tools/testing/selftests/kvm/lib/x86_64/processor.c if (!pde->present) pde 393 tools/testing/selftests/kvm/lib/x86_64/processor.c indent, "", pde - pde_start, pde, pde 394 tools/testing/selftests/kvm/lib/x86_64/processor.c addr_hva2gpa(vm, pde), pde 395 tools/testing/selftests/kvm/lib/x86_64/processor.c (uint64_t) pde->address, pde->writable, pde 396 tools/testing/selftests/kvm/lib/x86_64/processor.c pde->execute_disable); pde 399 tools/testing/selftests/kvm/lib/x86_64/processor.c pde->address * vm->page_size); pde 547 tools/testing/selftests/kvm/lib/x86_64/processor.c struct pageDirectoryEntry *pde; pde 568 tools/testing/selftests/kvm/lib/x86_64/processor.c pde = addr_gpa2hva(vm, pdpe[index[2]].address * vm->page_size); pde 569 tools/testing/selftests/kvm/lib/x86_64/processor.c if (!pde[index[1]].present) pde 572 tools/testing/selftests/kvm/lib/x86_64/processor.c pte = addr_gpa2hva(vm, pde[index[1]].address * vm->page_size); pde 447 tools/testing/selftests/kvm/lib/x86_64/vmx.c struct eptPageTableEntry *pde; pde 448 tools/testing/selftests/kvm/lib/x86_64/vmx.c pde = addr_gpa2hva(vm, pdpe[index[2]].address * vm->page_size); pde 449 tools/testing/selftests/kvm/lib/x86_64/vmx.c if (!pde[index[1]].readable) { pde 450 tools/testing/selftests/kvm/lib/x86_64/vmx.c pde[index[1]].address = vm_phy_page_alloc(vm, pde 453 tools/testing/selftests/kvm/lib/x86_64/vmx.c pde[index[1]].writable = true; pde 454 tools/testing/selftests/kvm/lib/x86_64/vmx.c pde[index[1]].readable = true; pde 455 tools/testing/selftests/kvm/lib/x86_64/vmx.c pde[index[1]].executable = true; pde 460 tools/testing/selftests/kvm/lib/x86_64/vmx.c pte = addr_gpa2hva(vm, pde[index[1]].address * vm->page_size);