/linux-4.4.14/drivers/gpu/drm/ |
H A D | ati_pcigart.c | 104 u32 *pci_gart = NULL, page_base, gart_idx; drm_ati_pcigart_init() local 165 page_base = (u32) entry->busaddr[i]; drm_ati_pcigart_init() 172 val = page_base | 0xc; drm_ati_pcigart_init() 175 val = (page_base >> 8) | 0xc; drm_ati_pcigart_init() 179 val = page_base; drm_ati_pcigart_init() 188 page_base += ATI_PCIGART_PAGE_SIZE; drm_ati_pcigart_init()
|
/linux-4.4.14/arch/ia64/mm/ |
H A D | ioremap.c | 43 unsigned long page_base; ioremap() local 71 page_base = phys_addr & PAGE_MASK; ioremap() 72 size = PAGE_ALIGN(phys_addr + size) - page_base; ioremap() 73 if (efi_mem_attribute(page_base, size) & EFI_MEMORY_WB) { ioremap()
|
/linux-4.4.14/net/sunrpc/xprtrdma/ |
H A D | rpc_rdma.c | 150 int page_base; rpcrdma_convert_iovs() local 161 ppages = xdrbuf->pages + (xdrbuf->page_base >> PAGE_SHIFT); rpcrdma_convert_iovs() 162 page_base = xdrbuf->page_base & ~PAGE_MASK; rpcrdma_convert_iovs() 172 seg[n].mr_offset = (void *)(unsigned long) page_base; rpcrdma_convert_iovs() 173 seg[n].mr_len = min_t(u32, PAGE_SIZE - page_base, len); rpcrdma_convert_iovs() 179 page_base = 0; /* page offset only applies to first page */ rpcrdma_convert_iovs() 371 int page_base; rpcrdma_inline_pullup() local 396 page_base = rqst->rq_snd_buf.page_base; rpcrdma_inline_pullup() 397 ppages = rqst->rq_snd_buf.pages + (page_base >> PAGE_SHIFT); rpcrdma_inline_pullup() 398 page_base &= ~PAGE_MASK; rpcrdma_inline_pullup() 399 npages = PAGE_ALIGN(page_base+copy_len) >> PAGE_SHIFT; rpcrdma_inline_pullup() 401 curlen = PAGE_SIZE - page_base; rpcrdma_inline_pullup() 407 memcpy(destp, srcp+page_base, curlen); rpcrdma_inline_pullup() 412 page_base = 0; rpcrdma_inline_pullup() 623 int page_base; rpcrdma_inline_fixup() local 642 page_base = rqst->rq_rcv_buf.page_base; rpcrdma_inline_fixup() 643 ppages = rqst->rq_rcv_buf.pages + (page_base >> PAGE_SHIFT); rpcrdma_inline_fixup() 644 page_base &= ~PAGE_MASK; rpcrdma_inline_fixup() 647 npages = PAGE_ALIGN(page_base + rpcrdma_inline_fixup() 650 curlen = PAGE_SIZE - page_base; rpcrdma_inline_fixup() 657 memcpy(destp + page_base, srcp, curlen); rpcrdma_inline_fixup() 664 page_base = 0; rpcrdma_inline_fixup()
|
H A D | svc_rdma_recvfrom.c | 84 rqstp->rq_arg.page_base = 0; rdma_build_arg_xdr() 443 head->arg.page_base = 0; rdma_read_chunks() 540 rqstp->rq_arg.page_base = head->arg.page_base; rdma_read_complete()
|
H A D | svc_rdma_sendto.c | 80 page_off = xdr->page_base; map_xdr() 101 "page_base %u page_len %u head_len %zu tail_len %zu\n", map_xdr() 102 sge_no, page_no, xdr->page_base, xdr->page_len, map_xdr() 123 xdr_off += xdr->page_base; dma_map_xdr()
|
/linux-4.4.14/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_gart.c | 230 u64 page_base; amdgpu_gart_unbind() local 244 page_base = adev->gart.pages_addr[p]; amdgpu_gart_unbind() 250 t, page_base, flags); amdgpu_gart_unbind() 251 page_base += AMDGPU_GPU_PAGE_SIZE; amdgpu_gart_unbind() 278 uint64_t page_base; amdgpu_gart_bind() local 293 page_base = adev->gart.pages_addr[p]; amdgpu_gart_bind() 295 amdgpu_gart_set_pte_pde(adev, adev->gart.ptr, t, page_base, flags); amdgpu_gart_bind() 296 page_base += AMDGPU_GPU_PAGE_SIZE; amdgpu_gart_bind()
|
/linux-4.4.14/net/sunrpc/ |
H A D | xdr.c | 126 kaddr[buf->page_base + len] = '\0'; xdr_terminate_string() 143 xdr->page_base = base; xdr_inline_pages() 348 buf->page_base + pglen + offs - len, xdr_shrink_bufhead() 365 buf->page_base + len, xdr_shrink_bufhead() 366 buf->page_base, xdr_shrink_bufhead() 371 _copy_to_pages(buf->pages, buf->page_base, xdr_shrink_bufhead() 420 buf->pages, buf->page_base + pglen - len, xdr_shrink_pagelen() 631 new = buf->page_base + buf->page_len; xdr_truncate_encode() 697 buf->page_base = base; xdr_write_pages() 745 base += xdr->buf->page_base; xdr_set_page_base() 767 newbase -= xdr->buf->page_base; xdr_set_next_page() 1040 base += buf->page_base; xdr_buf_subsegment() 1041 subbuf->page_base = base & ~PAGE_CACHE_MASK; xdr_buf_subsegment() 1118 _copy_from_pages(obj, subbuf->pages, subbuf->page_base, this_len); __read_bytes_from_xdr_buf() 1149 _copy_to_pages(subbuf->pages, subbuf->page_base, obj, this_len); __write_bytes_to_xdr_buf() 1299 base += buf->page_base; xdr_xcode_array2() 1482 page_offset = (offset + buf->page_base) & (PAGE_CACHE_SIZE - 1); xdr_process_buf() 1483 i = (offset + buf->page_base) >> PAGE_CACHE_SHIFT; xdr_process_buf()
|
H A D | socklib.c | 96 if (base || xdr->page_base) { xdr_partial_copy_from_skb() 98 base += xdr->page_base; xdr_partial_copy_from_skb()
|
H A D | svcsock.c | 182 size_t base = xdr->page_base; svc_send_common() 652 rqstp->rq_arg.page_base = 0; svc_udp_recvfrom() 1164 rqstp->rq_arg.page_base = 0; svc_tcp_recvfrom()
|
H A D | svc_xprt.c | 482 rqstp->rq_res.page_base = 0; svc_xprt_release() 634 arg->page_base = 0; svc_alloc_arg()
|
H A D | svc.c | 1311 rqstp->rq_res.page_base = 0; svc_process()
|
H A D | xprtsock.c | 352 base += xdr->page_base; xs_send_pagedata()
|
/linux-4.4.14/drivers/gpu/drm/radeon/ |
H A D | radeon_gart.c | 289 uint64_t page_base, page_entry; radeon_gart_bind() local 301 page_base = dma_addr[i]; radeon_gart_bind() 303 page_entry = radeon_gart_get_page_entry(page_base, flags); radeon_gart_bind() 308 page_base += RADEON_GPU_PAGE_SIZE; radeon_gart_bind()
|
H A D | r600_cp.c | 187 u64 page_base; r600_page_table_init() local 213 page_base = (u64) entry_addr & ATI_PCIGART_PAGE_MASK; r600_page_table_init() 214 page_base |= R600_PTE_VALID | R600_PTE_SYSTEM | R600_PTE_SNOOPED; r600_page_table_init() 215 page_base |= R600_PTE_READABLE | R600_PTE_WRITEABLE; r600_page_table_init() 217 DRM_WRITE64(map, gart_idx * sizeof(u64), page_base); r600_page_table_init() 223 i, (unsigned long long)page_base); r600_page_table_init()
|
/linux-4.4.14/drivers/gpu/drm/qxl/ |
H A D | qxl_image.c | 165 unsigned page_base, page_offset, out_offset; qxl_image_init_helper() local 172 page_base = out_offset & PAGE_MASK; qxl_image_init_helper() 176 ptr = qxl_bo_kmap_atomic_page(qdev, chunk_bo, page_base); qxl_image_init_helper()
|
/linux-4.4.14/arch/x86/mm/ |
H A D | kmmio.c | 229 unsigned long page_base = addr; kmmio_handler() local 234 page_base &= page_level_mask(l); kmmio_handler() 247 faultpage = get_kmmio_fault_page(page_base); kmmio_handler() 259 if (page_base == ctx->addr) { kmmio_handler() 287 ctx->probe = get_kmmio_probe(page_base); kmmio_handler() 289 ctx->addr = page_base; kmmio_handler()
|
/linux-4.4.14/drivers/hid/ |
H A D | hid-rmi.c | 675 u16 page_base = page << 8; rmi_register_function() local 691 f->query_base_addr = page_base | pdt_entry->query_base_addr; rmi_register_function() 692 f->command_base_addr = page_base | pdt_entry->command_base_addr; rmi_register_function() 693 f->control_base_addr = page_base | pdt_entry->control_base_addr; rmi_register_function() 694 f->data_base_addr = page_base | pdt_entry->data_base_addr; rmi_register_function()
|
/linux-4.4.14/fs/nfsd/ |
H A D | nfscache.c | 301 idx = buf->page_base / PAGE_SIZE; nfsd_cache_csum() 302 base = buf->page_base & ~PAGE_MASK; nfsd_cache_csum()
|
H A D | vfs.c | 802 rqstp->rq_res.page_base = buf->offset; nfsd_splice_actor()
|
H A D | nfs4xdr.c | 3376 xdr->page_ptr += (buf->page_base + maxcount + PAGE_SIZE - 1) nfsd4_encode_splice_read()
|
/linux-4.4.14/net/sunrpc/auth_gss/ |
H A D | gss_rpc_xdr.h | 138 unsigned int page_base; /* Start of page data */ member in struct:gssp_in_token
|
H A D | gss_krb5_wrap.c | 81 unsigned int last = (buf->page_base + len - 1) gss_krb5_remove_padding() 83 unsigned int offset = (buf->page_base + len - 1) gss_krb5_remove_padding()
|
H A D | auth_gss.c | 1731 first = snd_buf->page_base >> PAGE_CACHE_SHIFT; alloc_enc_pages() 1732 last = (snd_buf->page_base + snd_buf->page_len - 1) >> PAGE_CACHE_SHIFT; alloc_enc_pages() 1778 first = snd_buf->page_base >> PAGE_CACHE_SHIFT; gss_wrap_req_priv() 1781 snd_buf->page_base -= first << PAGE_CACHE_SHIFT; gss_wrap_req_priv()
|
H A D | gss_rpc_xdr.c | 69 xdr_write_pages(xdr, in->pages, in->page_base, in->page_len); gssx_enc_in_token()
|
H A D | svcauth_gss.c | 1045 * rq_arg.head[0].iov_base will provide us the page_base to pass 1067 in_token->page_base = (ulong)argv->iov_base & ~PAGE_MASK; gss_read_proxy_verf()
|
H A D | gss_krb5_crypto.c | 417 int i = (page_pos + outbuf->page_base) >> PAGE_CACHE_SHIFT; encryptor()
|
/linux-4.4.14/include/linux/sunrpc/ |
H A D | xdr.h | 60 unsigned int page_base, /* Start of page data */ member in struct:xdr_buf
|
/linux-4.4.14/drivers/block/ |
H A D | cpqarray.c | 716 ulong page_base = ((ulong) base) & PAGE_MASK; remap_pci_mem() local 717 ulong page_offs = ((ulong) base) - page_base; remap_pci_mem() 718 void __iomem *page_remapped = ioremap(page_base, page_offs+size); remap_pci_mem()
|
H A D | cciss.c | 3009 ulong page_base = ((ulong) base) & PAGE_MASK; remap_pci_mem() local 3010 ulong page_offs = ((ulong) base) - page_base; remap_pci_mem() 3011 void __iomem *page_remapped = ioremap(page_base, page_offs + size); remap_pci_mem()
|
/linux-4.4.14/drivers/gpu/drm/i915/ |
H A D | i915_gem.c | 741 loff_t page_base, int page_offset, fast_user_write() 749 vaddr_atomic = io_mapping_map_atomic_wc(mapping, page_base); fast_user_write() 770 loff_t offset, page_base; i915_gem_gtt_pwrite_fast() local 796 * page_base = page offset within aperture i915_gem_gtt_pwrite_fast() 800 page_base = offset & PAGE_MASK; i915_gem_gtt_pwrite_fast() 810 if (fast_user_write(dev_priv->gtt.mappable, page_base, i915_gem_gtt_pwrite_fast() 740 fast_user_write(struct io_mapping *mapping, loff_t page_base, int page_offset, char __user *user_data, int length) fast_user_write() argument
|
/linux-4.4.14/drivers/scsi/ |
H A D | hpsa.c | 6739 ulong page_base = ((ulong) base) & PAGE_MASK; remap_pci_mem() local 6740 ulong page_offs = ((ulong) base) - page_base; remap_pci_mem() 6741 void __iomem *page_remapped = ioremap_nocache(page_base, remap_pci_mem()
|