page_num 4271 arch/x86/kvm/mmu.c int page_num = KVM_PAGES_PER_HPAGE(level); page_num 4273 arch/x86/kvm/mmu.c gfn &= ~(page_num - 1); page_num 4275 arch/x86/kvm/mmu.c return kvm_mtrr_check_gfn_range_consistency(vcpu, gfn, page_num); page_num 696 arch/x86/kvm/mtrr.c int page_num) page_num 704 arch/x86/kvm/mtrr.c end = gfn_to_gpa(gfn + page_num); page_num 2586 arch/x86/kvm/x86.c u32 page_num = data & ~PAGE_MASK; page_num 2592 arch/x86/kvm/x86.c if (page_num >= blob_size) page_num 2595 arch/x86/kvm/x86.c page = memdup_user(blob_addr + (page_num * PAGE_SIZE), PAGE_SIZE); page_num 287 arch/x86/kvm/x86.h int page_num); page_num 329 drivers/crypto/qat/qat_common/icp_qat_uclo.h unsigned int page_num; page_num 340 drivers/crypto/qat/qat_common/icp_qat_uclo.h unsigned int page_num; page_num 997 drivers/dma-buf/dma-buf.c void *dma_buf_kmap(struct dma_buf *dmabuf, unsigned long page_num) page_num 1003 drivers/dma-buf/dma-buf.c return dmabuf->ops->map(dmabuf, page_num); page_num 1015 drivers/dma-buf/dma-buf.c void dma_buf_kunmap(struct dma_buf *dmabuf, unsigned long page_num, page_num 1021 drivers/dma-buf/dma-buf.c dmabuf->ops->unmap(dmabuf, page_num, vaddr); page_num 96 drivers/dma-buf/udmabuf.c static void *kmap_udmabuf(struct dma_buf *buf, unsigned long page_num) page_num 99 drivers/dma-buf/udmabuf.c struct page *page = ubuf->pages[page_num]; page_num 104 drivers/dma-buf/udmabuf.c static void kunmap_udmabuf(struct dma_buf *buf, unsigned long page_num, page_num 106 drivers/gpu/drm/gma500/framebuffer.c int page_num; page_num 114 drivers/gpu/drm/gma500/framebuffer.c page_num = vma_pages(vma); page_num 119 drivers/gpu/drm/gma500/framebuffer.c for (i = 0; i < page_num; i++) { page_num 96 drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c static void *i915_gem_dmabuf_kmap(struct dma_buf *dma_buf, unsigned long page_num) page_num 101 drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c if (page_num >= obj->base.size >> PAGE_SHIFT) page_num 111 drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c page = i915_gem_object_get_page(obj, page_num); page_num 122 drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c static void i915_gem_dmabuf_kunmap(struct dma_buf *dma_buf, unsigned long page_num, void *addr) page_num 603 drivers/gpu/drm/i915/gem/selftests/huge_pages.c unsigned long page_num; page_num 609 drivers/gpu/drm/i915/gem/selftests/huge_pages.c for_each_prime_number_from(page_num, 1, max_pages) { page_num 611 drivers/gpu/drm/i915/gem/selftests/huge_pages.c u64 size = page_num << PAGE_SHIFT; page_num 79 drivers/gpu/drm/i915/gem/selftests/mock_dmabuf.c static void *mock_dmabuf_kmap(struct dma_buf *dma_buf, unsigned long page_num) page_num 83 drivers/gpu/drm/i915/gem/selftests/mock_dmabuf.c return kmap(mock->pages[page_num]); page_num 86 drivers/gpu/drm/i915/gem/selftests/mock_dmabuf.c static void mock_dmabuf_kunmap(struct dma_buf *dma_buf, unsigned long page_num, void *addr) page_num 90 drivers/gpu/drm/i915/gem/selftests/mock_dmabuf.c return kunmap(mock->pages[page_num]); page_num 48 drivers/gpu/drm/i915/gvt/dmabuf.c u32 page_num; page_num 58 drivers/gpu/drm/i915/gvt/dmabuf.c page_num = obj->base.size >> PAGE_SHIFT; page_num 59 drivers/gpu/drm/i915/gvt/dmabuf.c ret = sg_alloc_table(st, page_num, GFP_KERNEL); page_num 66 drivers/gpu/drm/i915/gvt/dmabuf.c for_each_sg(st->sgl, sg, page_num, i) { page_num 616 drivers/gpu/drm/i915/selftests/i915_buddy.c unsigned long page_num; page_num 643 drivers/gpu/drm/i915/selftests/i915_buddy.c for_each_prime_number_from(page_num, 1, ULONG_MAX - 1) { page_num 647 drivers/gpu/drm/i915/selftests/i915_buddy.c size = min(page_num * mm.chunk_size, rem); page_num 89 drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c unsigned long page_num) page_num 94 drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c omap_gem_cpu_sync_page(obj, page_num); page_num 95 drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c return kmap(pages[page_num]); page_num 99 drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c unsigned long page_num, void *addr) page_num 104 drivers/gpu/drm/omapdrm/omap_gem_dmabuf.c kunmap(pages[page_num]); page_num 143 drivers/gpu/drm/udl/udl_dmabuf.c static void *udl_dmabuf_kmap(struct dma_buf *dma_buf, unsigned long page_num) page_num 151 drivers/gpu/drm/udl/udl_dmabuf.c unsigned long page_num, void *addr) page_num 75 drivers/gpu/drm/vmwgfx/vmwgfx_prime.c unsigned long page_num) page_num 81 drivers/gpu/drm/vmwgfx/vmwgfx_prime.c unsigned long page_num, void *addr) page_num 374 drivers/infiniband/hw/efa/efa_admin_cmds_defs.h u32 page_num; page_num 175 drivers/infiniband/hw/efa/efa_com_cmd.h u32 page_num; page_num 1347 drivers/infiniband/hw/efa/efa_verbs.c params->page_num, params->page_shift); page_num 1352 drivers/infiniband/hw/efa/efa_verbs.c "inline_pbl_array - pages[%u]\n", params->page_num); page_num 1364 drivers/infiniband/hw/efa/efa_verbs.c err = pbl_create(dev, pbl, mr->umem, params->page_num, page_num 1450 drivers/infiniband/hw/efa/efa_verbs.c params.page_num = DIV_ROUND_UP(length + (start & (pg_sz - 1)), page_num 1455 drivers/infiniband/hw/efa/efa_verbs.c start, length, params.page_shift, params.page_num); page_num 1458 drivers/infiniband/hw/efa/efa_verbs.c if (params.page_num <= inline_size) { page_num 1789 drivers/infiniband/hw/hns/hns_roce_hw_v2.c u32 page_num; page_num 1805 drivers/infiniband/hw/hns/hns_roce_hw_v2.c page_num = link_tbl->npages; page_num 1839 drivers/infiniband/hw/hns/hns_roce_hw_v2.c cpu_to_le32(entry[page_num - 1].blk_ba0); page_num 1843 drivers/infiniband/hw/hns/hns_roce_hw_v2.c entry[page_num - 1].blk_ba1_nxt_ptr & page_num 1848 drivers/infiniband/hw/hns/hns_roce_hw_v2.c (entry[page_num - 2].blk_ba1_nxt_ptr & page_num 301 drivers/iommu/mtk_iommu_v1.c unsigned int page_num = size >> MT2701_IOMMU_PAGE_SHIFT; page_num 309 drivers/iommu/mtk_iommu_v1.c for (i = 0; i < page_num; i++) { page_num 333 drivers/iommu/mtk_iommu_v1.c unsigned int page_num = size >> MT2701_IOMMU_PAGE_SHIFT; page_num 336 drivers/iommu/mtk_iommu_v1.c memset(pgt_base_iova, 0, page_num * sizeof(u32)); page_num 2288 drivers/net/ethernet/emulex/benet/be_cmds.c u8 page_num, u8 *data) page_num 2295 drivers/net/ethernet/emulex/benet/be_cmds.c if (page_num > TR_PAGE_A2) page_num 2320 drivers/net/ethernet/emulex/benet/be_cmds.c req->page_num = cpu_to_le32(page_num); page_num 1094 drivers/net/ethernet/emulex/benet/be_cmds.h __le32 page_num; page_num 1119 drivers/net/ethernet/emulex/benet/be_cmds.h u32 page_num; page_num 2430 drivers/net/ethernet/emulex/benet/be_cmds.h u8 page_num, u8 *data); page_num 1980 drivers/net/ethernet/mellanox/mlx4/port.c u8 page_num; page_num 2083 drivers/net/ethernet/mellanox/mlx4/port.c cable_info->page_num = 0; page_num 310 drivers/net/ethernet/mellanox/mlx5/core/port.c static int mlx5_eeprom_high_page_offset(int page_num) page_num 312 drivers/net/ethernet/mellanox/mlx5/core/port.c if (!page_num) /* Page 0 always start from low page */ page_num 316 drivers/net/ethernet/mellanox/mlx5/core/port.c return page_num * MLX5_EEPROM_HIGH_PAGE_LENGTH; page_num 322 drivers/net/ethernet/mellanox/mlx5/core/port.c int module_num, page_num, status, err; page_num 336 drivers/net/ethernet/mellanox/mlx5/core/port.c page_num = mlx5_eeprom_page(offset); page_num 341 drivers/net/ethernet/mellanox/mlx5/core/port.c offset -= mlx5_eeprom_high_page_offset(page_num); page_num 352 drivers/net/ethernet/mellanox/mlx5/core/port.c MLX5_SET(mcia_reg, in, page_number, page_num); page_num 625 drivers/net/ethernet/neterion/s2io.c int page_num = TXD_MEM_PAGE_CNT(config->tx_cfg[i].fifo_len, page_num 639 drivers/net/ethernet/neterion/s2io.c for (j = 0; j < page_num; j++) { page_num 875 drivers/net/ethernet/neterion/s2io.c int page_num = 0; page_num 898 drivers/net/ethernet/neterion/s2io.c page_num = TXD_MEM_PAGE_CNT(tx_cfg->fifo_len, lst_per_page); page_num 899 drivers/net/ethernet/neterion/s2io.c for (j = 0; j < page_num; j++) { page_num 13 drivers/net/ethernet/pensando/ionic/ionic_bus.h void __iomem *ionic_bus_map_dbpage(struct ionic *ionic, int page_num); page_num 95 drivers/net/ethernet/pensando/ionic/ionic_bus_pci.c void __iomem *ionic_bus_map_dbpage(struct ionic *ionic, int page_num) page_num 99 drivers/net/ethernet/pensando/ionic/ionic_bus_pci.c (u64)page_num << PAGE_SHIFT, PAGE_SIZE); page_num 1124 drivers/net/wireless/intel/iwlwifi/fw/dbg.c range->page_num = cpu_to_le32(idx); page_num 1148 drivers/net/wireless/intel/iwlwifi/fw/dbg.c range->page_num = cpu_to_le32(idx); page_num 337 drivers/net/wireless/intel/iwlwifi/fw/error-dump.h __le32 page_num; page_num 25 drivers/net/wireless/realtek/rtw88/debug.c u32 page_num; page_num 221 drivers/net/wireless/realtek/rtw88/debug.c u32 buf_size = debugfs_priv->rsvd_page.page_num * page_size; page_num 260 drivers/net/wireless/realtek/rtw88/debug.c u32 offset, page_num; page_num 265 drivers/net/wireless/realtek/rtw88/debug.c num = sscanf(tmp, "%d %d", &offset, &page_num); page_num 273 drivers/net/wireless/realtek/rtw88/debug.c debugfs_priv->rsvd_page.page_num = page_num; page_num 1092 drivers/staging/gasket/gasket_page_table.c uint page_num; page_num 1097 drivers/staging/gasket/gasket_page_table.c page_num = gasket_simple_page_idx(pg_tbl, dev_addr); page_num 1098 drivers/staging/gasket/gasket_page_table.c if (page_num >= pg_tbl->num_simple_entries) page_num 1101 drivers/staging/gasket/gasket_page_table.c pte = pg_tbl->entries + page_num; page_num 1106 drivers/staging/gasket/gasket_page_table.c page_num = gasket_extended_lvl0_page_idx(pg_tbl, dev_addr); page_num 1107 drivers/staging/gasket/gasket_page_table.c if (page_num >= pg_tbl->num_extended_entries) page_num 1110 drivers/staging/gasket/gasket_page_table.c pte = pg_tbl->entries + pg_tbl->num_simple_entries + page_num; page_num 1115 drivers/staging/gasket/gasket_page_table.c page_num = gasket_extended_lvl1_page_idx(pg_tbl, dev_addr); page_num 1116 drivers/staging/gasket/gasket_page_table.c pte = pte->sublevel + page_num; page_num 1252 drivers/staging/rts5208/ms.c u16 block_addr, u8 page_num, u8 *buf, int buf_len) page_num 1274 drivers/staging/rts5208/ms.c data[5] = page_num; page_num 1333 drivers/staging/rts5208/ms.c u8 page_num, u8 *buf, int buf_len) page_num 1356 drivers/staging/rts5208/ms.c data[5] = page_num; page_num 1389 drivers/staging/rts5208/ms.c static int ms_read_page(struct rtsx_chip *chip, u16 block_addr, u8 page_num) page_num 1409 drivers/staging/rts5208/ms.c data[5] = page_num; page_num 245 drivers/tee/optee/rpc.c size_t page_num; page_num 247 drivers/tee/optee/rpc.c pages = tee_shm_get_pages(shm, &page_num); page_num 248 drivers/tee/optee/rpc.c if (!pages || !page_num) { page_num 253 drivers/tee/optee/rpc.c pages_list = optee_allocate_pages_list(page_num); page_num 260 drivers/tee/optee/rpc.c call_ctx->num_entries = page_num; page_num 274 drivers/tee/optee/rpc.c optee_fill_pages_list(pages_list, pages, page_num, page_num 346 drivers/xen/gntdev-dmabuf.c unsigned long page_num) page_num 353 drivers/xen/gntdev-dmabuf.c unsigned long page_num, void *addr) page_num 222 fs/btrfs/scrub.c int page_num, int force_write); page_num 225 fs/btrfs/scrub.c int page_num); page_num 835 fs/btrfs/scrub.c int page_num; page_num 1094 fs/btrfs/scrub.c for (page_num = 0; page_num < sblock_bad->page_count; page_num 1095 fs/btrfs/scrub.c page_num++) { page_num 1096 fs/btrfs/scrub.c struct scrub_page *page_bad = sblock_bad->pagev[page_num]; page_num 1119 fs/btrfs/scrub.c pagev[page_num]->io_error) { page_num 1141 fs/btrfs/scrub.c page_num) != 0) { page_num 1149 fs/btrfs/scrub.c page_num, 0); page_num 1424 fs/btrfs/scrub.c int page_num; page_num 1434 fs/btrfs/scrub.c for (page_num = 0; page_num < sblock->page_count; page_num++) { page_num 1435 fs/btrfs/scrub.c struct scrub_page *page = sblock->pagev[page_num]; page_num 1452 fs/btrfs/scrub.c for (page_num = 0; page_num < sblock->page_count; page_num++) page_num 1453 fs/btrfs/scrub.c sblock->pagev[page_num]->io_error = 1; page_num 1469 fs/btrfs/scrub.c int page_num; page_num 1477 fs/btrfs/scrub.c for (page_num = 0; page_num < sblock->page_count; page_num++) { page_num 1479 fs/btrfs/scrub.c struct scrub_page *page = sblock->pagev[page_num]; page_num 1532 fs/btrfs/scrub.c int page_num; page_num 1535 fs/btrfs/scrub.c for (page_num = 0; page_num < sblock_bad->page_count; page_num++) { page_num 1540 fs/btrfs/scrub.c page_num, 1); page_num 1550 fs/btrfs/scrub.c int page_num, int force_write) page_num 1552 fs/btrfs/scrub.c struct scrub_page *page_bad = sblock_bad->pagev[page_num]; page_num 1553 fs/btrfs/scrub.c struct scrub_page *page_good = sblock_good->pagev[page_num]; page_num 1596 fs/btrfs/scrub.c int page_num; page_num 1605 fs/btrfs/scrub.c for (page_num = 0; page_num < sblock->page_count; page_num++) { page_num 1608 fs/btrfs/scrub.c ret = scrub_write_page_to_dev_replace(sblock, page_num); page_num 1615 fs/btrfs/scrub.c int page_num) page_num 1617 fs/btrfs/scrub.c struct scrub_page *spage = sblock->pagev[page_num]; page_num 894 samples/vfio-mdev/mbochs.c static void *mbochs_kmap_dmabuf(struct dma_buf *buf, unsigned long page_num) page_num 897 samples/vfio-mdev/mbochs.c struct page *page = dmabuf->pages[page_num]; page_num 902 samples/vfio-mdev/mbochs.c static void mbochs_kunmap_dmabuf(struct dma_buf *buf, unsigned long page_num,