pgsz 314 arch/mips/sgi-ip22/ip28-berr.c unsigned int pgsz = (ctl & 2) ? 14:12; /* 16k:4k */ pgsz 317 arch/mips/sgi-ip22/ip28-berr.c pte += 8*((vaddr >> pgsz) & 0x1ff); pgsz 327 arch/mips/sgi-ip22/ip28-berr.c a += vaddr & ((1 << pgsz) - 1); pgsz 1387 drivers/hwtracing/intel_th/msu.c size_t pgsz = PFN_DOWN(sg->length); pgsz 1389 drivers/hwtracing/intel_th/msu.c if (pgoff < pgsz) pgsz 1392 drivers/hwtracing/intel_th/msu.c pgoff -= pgsz; pgsz 2829 drivers/infiniband/core/verbs.c unsigned long pgsz) pgsz 2836 drivers/infiniband/core/verbs.c biter->__pg_bit = __fls(pgsz); pgsz 238 drivers/iommu/omap-iommu.c attr <<= (((e->pgsz == MMU_CAM_PGSZ_4K) || pgsz 239 drivers/iommu/omap-iommu.c (e->pgsz == MMU_CAM_PGSZ_64K)) ? 0 : 6); pgsz 320 drivers/iommu/omap-iommu.c if (e->da & ~(get_cam_va_mask(e->pgsz))) { pgsz 330 drivers/iommu/omap-iommu.c cr->cam = (e->da & MMU_CAM_VATAG_MASK) | e->prsvd | e->pgsz | e->valid; pgsz 642 drivers/iommu/omap-iommu.c switch (e->pgsz) { pgsz 1326 drivers/iommu/omap-iommu.c static u32 iotlb_init_entry(struct iotlb_entry *e, u32 da, u32 pa, int pgsz) pgsz 1333 drivers/iommu/omap-iommu.c e->pgsz = pgsz; pgsz 1338 drivers/iommu/omap-iommu.c return iopgsz_to_bytes(e->pgsz); pgsz 24 drivers/iommu/omap-iommu.h u32 pgsz, prsvd, valid; pgsz 194 drivers/iommu/omap-iommu.h #define get_cam_va_mask(pgsz) \ pgsz 195 drivers/iommu/omap-iommu.h (((pgsz) == MMU_CAM_PGSZ_16M) ? 0xff000000 : \ pgsz 196 drivers/iommu/omap-iommu.h ((pgsz) == MMU_CAM_PGSZ_1M) ? 0xfff00000 : \ pgsz 197 drivers/iommu/omap-iommu.h ((pgsz) == MMU_CAM_PGSZ_64K) ? 0xffff0000 : \ pgsz 198 drivers/iommu/omap-iommu.h ((pgsz) == MMU_CAM_PGSZ_4K) ? 0xfffff000 : 0) pgsz 203 drivers/mtd/nand/raw/nandsim.c #define NS_RAW_OFFSET_OOB(ns) (NS_RAW_OFFSET(ns) + ns->geom.pgsz) pgsz 317 drivers/mtd/nand/raw/nandsim.c uint pgsz; /* NAND flash page size, bytes */ pgsz 639 drivers/mtd/nand/raw/nandsim.c ns->geom.pgsz = mtd->writesize; pgsz 642 drivers/mtd/nand/raw/nandsim.c ns->geom.pgszoob = ns->geom.pgsz + ns->geom.oobsz; pgsz 643 drivers/mtd/nand/raw/nandsim.c ns->geom.pgnum = div_u64(ns->geom.totsz, ns->geom.pgsz); pgsz 647 drivers/mtd/nand/raw/nandsim.c ns->geom.pgsec = ns->geom.secsz / ns->geom.pgsz; pgsz 651 drivers/mtd/nand/raw/nandsim.c if (ns->geom.pgsz == 512) { pgsz 655 drivers/mtd/nand/raw/nandsim.c } else if (ns->geom.pgsz == 2048) { pgsz 657 drivers/mtd/nand/raw/nandsim.c } else if (ns->geom.pgsz == 4096) { pgsz 660 drivers/mtd/nand/raw/nandsim.c NS_ERR("init_nandsim: unknown page size %u\n", ns->geom.pgsz); pgsz 727 drivers/mtd/nand/raw/nandsim.c printk("page size: %u bytes\n", ns->geom.pgsz); pgsz 1589 drivers/mtd/nand/raw/nandsim.c else if (ns->regs.off < ns->geom.pgsz) pgsz 1595 drivers/mtd/nand/raw/nandsim.c NS_UDELAY(input_cycle * ns->geom.pgsz / 1000 / busdiv); pgsz 1666 drivers/mtd/nand/raw/nandsim.c NS_UDELAY(output_cycle * ns->geom.pgsz / 1000 / busdiv); pgsz 1686 drivers/mtd/nand/raw/nandsim.c NS_DBG("do_state_action: set internal offset to %d\n", ns->geom.pgsz/2); pgsz 1687 drivers/mtd/nand/raw/nandsim.c ns->regs.off = ns->geom.pgsz/2; pgsz 1691 drivers/mtd/nand/raw/nandsim.c NS_DBG("do_state_action: set internal offset to %d\n", ns->geom.pgsz); pgsz 1692 drivers/mtd/nand/raw/nandsim.c ns->regs.off = ns->geom.pgsz; pgsz 1141 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c pci_read_config_dword(dev, pos + PCI_SRIOV_SUP_PGSIZE, &iov->pgsz); pgsz 1171 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c iov->initial, iov->nr_virtfn, iov->offset, iov->stride, iov->pgsz); pgsz 56 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h u32 pgsz; /* page size for BAR alignment */ pgsz 66 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c int cxgbi_ppm_find_page_index(struct cxgbi_ppm *ppm, unsigned long pgsz) pgsz 72 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c if (pgsz == 1UL << (DDP_PGSZ_BASE_SHIFT + pgsz 75 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c __func__, ppm->ndev->name, pgsz, i); pgsz 79 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c pr_info("ippm: ddp page size %lu not supported.\n", pgsz); pgsz 317 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.h int cxgbi_ppm_find_page_index(struct cxgbi_ppm *ppm, unsigned long pgsz); pgsz 393 drivers/net/ethernet/qlogic/qed/qed_sriov.c pos + PCI_SRIOV_SUP_PGSIZE, &iov->pgsz); pgsz 407 drivers/net/ethernet/qlogic/qed/qed_sriov.c iov->nr_virtfn, iov->offset, iov->stride, iov->pgsz); pgsz 125 drivers/net/ethernet/qlogic/qed/qed_sriov.h u32 pgsz; /* page size for BAR alignment */ pgsz 590 drivers/pci/iov.c u32 pgsz; pgsz 618 drivers/pci/iov.c pci_read_config_dword(dev, pos + PCI_SRIOV_SUP_PGSIZE, &pgsz); pgsz 620 drivers/pci/iov.c pgsz &= ~((1 << i) - 1); pgsz 621 drivers/pci/iov.c if (!pgsz) pgsz 624 drivers/pci/iov.c pgsz &= ~(pgsz - 1); pgsz 625 drivers/pci/iov.c pci_write_config_dword(dev, pos + PCI_SRIOV_SYS_PGSIZE, pgsz); pgsz 663 drivers/pci/iov.c iov->pgsz = pgsz; pgsz 729 drivers/pci/iov.c pci_write_config_dword(dev, iov->pos + PCI_SRIOV_SYS_PGSIZE, iov->pgsz); pgsz 329 drivers/pci/pci.h u32 pgsz; /* Page size for BAR alignment */ pgsz 2746 include/rdma/ib_verbs.h unsigned long pgsz); pgsz 2770 include/rdma/ib_verbs.h #define rdma_for_each_block(sglist, biter, nents, pgsz) \ pgsz 2772 include/rdma/ib_verbs.h pgsz); \ pgsz 3289 include/rdma/ib_verbs.h unsigned long pgsz; pgsz 3296 include/rdma/ib_verbs.h pgsz = pgsz_bitmap & ~(-align << 1); pgsz 3297 include/rdma/ib_verbs.h if (!pgsz) pgsz 3300 include/rdma/ib_verbs.h return __fls(pgsz); pgsz 199 tools/perf/jvmti/jvmti_agent.c long pgsz; pgsz 201 tools/perf/jvmti/jvmti_agent.c pgsz = sysconf(_SC_PAGESIZE); pgsz 202 tools/perf/jvmti/jvmti_agent.c if (pgsz == -1) pgsz 216 tools/perf/jvmti/jvmti_agent.c marker_addr = mmap(NULL, pgsz, PROT_READ|PROT_EXEC, MAP_PRIVATE, fd, 0); pgsz 223 tools/perf/jvmti/jvmti_agent.c long pgsz; pgsz 228 tools/perf/jvmti/jvmti_agent.c pgsz = sysconf(_SC_PAGESIZE); pgsz 229 tools/perf/jvmti/jvmti_agent.c if (pgsz == -1) pgsz 232 tools/perf/jvmti/jvmti_agent.c munmap(marker_addr, pgsz);