/linux-4.1.27/arch/powerpc/include/asm/ |
H A D | mman.h | 21 static inline unsigned long arch_calc_vm_prot_bits(unsigned long prot) arch_calc_vm_prot_bits() argument 23 return (prot & PROT_SAO) ? VM_SAO : 0; arch_calc_vm_prot_bits() 25 #define arch_calc_vm_prot_bits(prot) arch_calc_vm_prot_bits(prot) 33 static inline int arch_validate_prot(unsigned long prot) arch_validate_prot() argument 35 if (prot & ~(PROT_READ | PROT_WRITE | PROT_EXEC | PROT_SEM | PROT_SAO)) arch_validate_prot() 37 if ((prot & PROT_SAO) && !cpu_has_feature(CPU_FTR_SAO)) arch_validate_prot() 41 #define arch_validate_prot(prot) arch_validate_prot(prot)
|
H A D | syscalls.h | 12 unsigned long prot, unsigned long flags, 15 unsigned long prot, unsigned long flags,
|
H A D | pgtable.h | 187 #define pgprot_noncached(prot) (__pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | \ 190 #define pgprot_noncached_wc(prot) (__pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | \ 193 #define pgprot_cached(prot) (__pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | \ 196 #define pgprot_cached_wthru(prot) (__pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | \ 199 #define pgprot_cached_noncoherent(prot) \ 200 (__pgprot(pgprot_val(prot) & ~_PAGE_CACHE_CTL))
|
H A D | pgtable-ppc64-4k.h | 85 #define remap_4k_pfn(vma, addr, pfn, prot) \ 86 remap_pfn_range((vma), (addr), (pfn), PAGE_SIZE, (prot))
|
H A D | pte-hash64-64k.h | 97 #define remap_4k_pfn(vma, addr, pfn, prot) \ 100 __pgprot(pgprot_val((prot)) | _PAGE_4K_PFN)))
|
H A D | highmem.h | 63 extern void *kmap_atomic_prot(struct page *page, pgprot_t prot);
|
/linux-4.1.27/arch/x86/include/asm/ |
H A D | fb.h | 11 unsigned long prot; fb_pgprotect() local 13 prot = pgprot_val(vma->vm_page_prot) & ~_PAGE_CACHE_MASK; fb_pgprotect() 16 prot | cachemode2protval(_PAGE_CACHE_MODE_UC_MINUS); fb_pgprotect()
|
H A D | iomap.h | 30 iomap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot); 36 iomap_create_wc(resource_size_t base, unsigned long size, pgprot_t *prot);
|
H A D | highmem.h | 66 void *kmap_atomic_prot(struct page *page, pgprot_t prot); 70 void *kmap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot);
|
/linux-4.1.27/net/ipv6/ |
H A D | protocol.c | 32 int inet6_add_protocol(const struct inet6_protocol *prot, unsigned char protocol) inet6_add_protocol() argument 35 NULL, prot) ? 0 : -1; inet6_add_protocol() 39 int inet6_del_protocol(const struct inet6_protocol *prot, unsigned char protocol) inet6_del_protocol() argument 44 prot, NULL) == prot) ? 0 : -1; inet6_del_protocol() 56 int inet6_add_offload(const struct net_offload *prot, unsigned char protocol) inet6_add_offload() argument 59 NULL, prot) ? 0 : -1; inet6_add_offload() 63 int inet6_del_offload(const struct net_offload *prot, unsigned char protocol) inet6_del_offload() argument 68 prot, NULL) == prot) ? 0 : -1; inet6_del_offload()
|
H A D | udplite.c | 65 .prot = &udplitev6_prot,
|
/linux-4.1.27/include/linux/ |
H A D | mman.h | 38 #define arch_calc_vm_prot_bits(prot) 0 50 * Returns true if the prot flags are valid 52 static inline int arch_validate_prot(unsigned long prot) arch_validate_prot() argument 54 return (prot & ~(PROT_READ | PROT_WRITE | PROT_EXEC | PROT_SEM)) == 0; arch_validate_prot() 70 * Combine the mmap "prot" argument into "vm_flags" used internally. 73 calc_vm_prot_bits(unsigned long prot) calc_vm_prot_bits() argument 75 return _calc_vm_trans(prot, PROT_READ, VM_READ ) | calc_vm_prot_bits() 76 _calc_vm_trans(prot, PROT_WRITE, VM_WRITE) | calc_vm_prot_bits() 77 _calc_vm_trans(prot, PROT_EXEC, VM_EXEC) | calc_vm_prot_bits() 78 arch_calc_vm_prot_bits(prot); calc_vm_prot_bits()
|
H A D | io-mapping.h | 41 pgprot_t prot; member in struct:io_mapping 55 pgprot_t prot; io_mapping_create_wc() local 61 if (iomap_create_wc(base, size, &prot)) io_mapping_create_wc() 66 iomap->prot = prot; io_mapping_create_wc() 93 return iomap_atomic_prot_pfn(pfn, mapping->prot); io_mapping_map_atomic_wc()
|
H A D | agpgart.h | 51 * The "prot" down below needs still a "sleep" flag somehow ... 56 int prot; /* prot flags for mmap */ member in struct:agp_segment 62 pgprot_t prot; member in struct:agp_segment_priv
|
H A D | ima.h | 20 extern int ima_file_mmap(struct file *file, unsigned long prot); 40 static inline int ima_file_mmap(struct file *file, unsigned long prot) ima_file_mmap() argument
|
H A D | vmalloc.h | 58 int node, pgprot_t prot); 77 extern void *__vmalloc(unsigned long size, gfp_t gfp_mask, pgprot_t prot); 80 pgprot_t prot, unsigned long vm_flags, int node, 86 unsigned long flags, pgprot_t prot); 123 extern int map_vm_area(struct vm_struct *area, pgprot_t prot, 127 pgprot_t prot, struct page **pages); 133 pgprot_t prot, struct page **pages) map_kernel_range_noflush() 132 map_kernel_range_noflush(unsigned long start, unsigned long size, pgprot_t prot, struct page **pages) map_kernel_range_noflush() argument
|
H A D | iommu.h | 148 phys_addr_t paddr, size_t size, int prot); 152 struct scatterlist *sg, unsigned int nents, int prot); 164 phys_addr_t paddr, u64 size, int prot); 197 phys_addr_t paddr, size_t size, int prot); 202 int prot); 245 int prot); 290 unsigned int nents, int prot) iommu_map_sg() 292 return domain->ops->map_sg(domain, iova, sg, nents, prot); iommu_map_sg() 336 phys_addr_t paddr, int gfp_order, int prot) iommu_map() 349 unsigned int nents, int prot) iommu_map_sg() 356 u64 size, int prot) iommu_domain_window_enable() 288 iommu_map_sg(struct iommu_domain *domain, unsigned long iova, struct scatterlist *sg, unsigned int nents, int prot) iommu_map_sg() argument 335 iommu_map(struct iommu_domain *domain, unsigned long iova, phys_addr_t paddr, int gfp_order, int prot) iommu_map() argument 347 iommu_map_sg(struct iommu_domain *domain, unsigned long iova, struct scatterlist *sg, unsigned int nents, int prot) iommu_map_sg() argument 354 iommu_domain_window_enable(struct iommu_domain *domain, u32 wnd_nr, phys_addr_t paddr, u64 size, int prot) iommu_domain_window_enable() argument
|
H A D | io.h | 32 phys_addr_t phys_addr, pgprot_t prot); 35 phys_addr_t phys_addr, pgprot_t prot) ioremap_page_range() 34 ioremap_page_range(unsigned long addr, unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_page_range() argument
|
H A D | crash_dump.h | 23 unsigned long size, pgprot_t prot);
|
/linux-4.1.27/net/ipv4/ |
H A D | protocol.c | 35 int inet_add_protocol(const struct net_protocol *prot, unsigned char protocol) inet_add_protocol() argument 37 if (!prot->netns_ok) { inet_add_protocol() 44 NULL, prot) ? 0 : -1; inet_add_protocol() 48 int inet_add_offload(const struct net_offload *prot, unsigned char protocol) inet_add_offload() argument 51 NULL, prot) ? 0 : -1; inet_add_offload() 55 int inet_del_protocol(const struct net_protocol *prot, unsigned char protocol) inet_del_protocol() argument 60 prot, NULL) == prot) ? 0 : -1; inet_del_protocol() 68 int inet_del_offload(const struct net_offload *prot, unsigned char protocol) inet_del_offload() argument 73 prot, NULL) == prot) ? 0 : -1; inet_del_offload()
|
H A D | udplite.c | 71 .prot = &udplite_prot,
|
/linux-4.1.27/arch/sh/include/asm/ |
H A D | syscalls.h | 7 unsigned long prot, unsigned long flags, 10 unsigned long prot, unsigned long flags,
|
H A D | mmu.h | 62 unsigned long size, pgprot_t prot); 64 pgprot_t prot, void *caller); 71 unsigned long size, pgprot_t prot) pmb_bolt_mapping() 78 pgprot_t prot, void *caller) pmb_remap_caller() 99 pmb_remap(phys_addr_t phys, unsigned long size, pgprot_t prot) pmb_remap() argument 101 return pmb_remap_caller(phys, size, prot, __builtin_return_address(0)); pmb_remap() 70 pmb_bolt_mapping(unsigned long virt, phys_addr_t phys, unsigned long size, pgprot_t prot) pmb_bolt_mapping() argument 77 pmb_remap_caller(phys_addr_t phys, unsigned long size, pgprot_t prot, void *caller) pmb_remap_caller() argument
|
H A D | io.h | 270 pgprot_t prot, void *caller); 274 __ioremap(phys_addr_t offset, unsigned long size, pgprot_t prot) __ioremap() argument 276 return __ioremap_caller(offset, size, prot, __builtin_return_address(0)); __ioremap() 280 __ioremap_29bit(phys_addr_t offset, unsigned long size, pgprot_t prot) __ioremap_29bit() argument 292 u64 flags = pgprot_val(prot); __ioremap_29bit() 315 __ioremap_mode(phys_addr_t offset, unsigned long size, pgprot_t prot) __ioremap_mode() argument 323 ret = __ioremap_29bit(offset, size, prot); __ioremap_mode() 327 return __ioremap(offset, size, prot); __ioremap_mode() 330 #define __ioremap(offset, size, prot) ((void __iomem *)(offset)) 331 #define __ioremap_mode(offset, size, prot) ((void __iomem *)(offset)) 360 ioremap_fixed(phys_addr_t phys_addr, unsigned long size, pgprot_t prot) ioremap_fixed() argument
|
H A D | pgtable_32.h | 317 #define pfn_pte(pfn, prot) \ 318 __pte(((unsigned long long)(pfn) << PAGE_SHIFT) | pgprot_val(prot)) 319 #define pfn_pmd(pfn, prot) \ 320 __pmd(((unsigned long long)(pfn) << PAGE_SHIFT) | pgprot_val(prot)) 378 #define pgprot_writecombine(prot) \ 379 __pgprot(pgprot_val(prot) & ~_PAGE_CACHABLE)
|
H A D | pgtable_64.h | 207 #define pgprot_writecombine(prot) __pgprot(pgprot_val(prot) & ~_PAGE_CACHABLE) 307 #define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) 308 #define pfn_pmd(pfn, prot) __pmd(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
|
/linux-4.1.27/include/trace/events/ |
H A D | sock.h | 34 TP_PROTO(struct sock *sk, struct proto *prot, long allocated), 36 TP_ARGS(sk, prot, allocated), 47 strncpy(__entry->name, prot->name, 32); 48 __entry->sysctl_mem = prot->sysctl_mem; 50 __entry->sysctl_rmem = prot->sysctl_rmem[0];
|
/linux-4.1.27/arch/um/include/asm/ |
H A D | pgtable-2level.h | 41 #define pfn_pte(pfn, prot) __pte(pfn_to_phys(pfn) | pgprot_val(prot)) 42 #define pfn_pmd(pfn, prot) __pmd(pfn_to_phys(pfn) | pgprot_val(prot))
|
H A D | page.h | 48 #define pte_set_val(pte, phys, prot) \ 50 (pte).pte_low = (phys) | pgprot_val(prot); }) 77 #define pte_set_val(p, phys, prot) (p).pte = (phys | pgprot_val(prot))
|
/linux-4.1.27/arch/avr32/include/asm/ |
H A D | pgtable-2level.h | 42 #define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) 43 #define pfn_pmd(pfn, prot) __pmd(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
|
H A D | pgtable.h | 263 * Mark the prot value as uncacheable and unbufferable 265 #define pgprot_noncached(prot) \ 266 __pgprot(pgprot_val(prot) & ~(_PAGE_BUFFER | _PAGE_CACHABLE)) 269 * Mark the prot value as uncacheable but bufferable 271 #define pgprot_writecombine(prot) \ 272 __pgprot((pgprot_val(prot) & ~_PAGE_CACHABLE) | _PAGE_BUFFER)
|
/linux-4.1.27/arch/um/kernel/ |
H A D | syscall.c | 18 unsigned long prot, unsigned long flags, old_mmap() 25 err = sys_mmap_pgoff(addr, len, prot, flags, fd, offset >> PAGE_SHIFT); old_mmap() 17 old_mmap(unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long fd, unsigned long offset) old_mmap() argument
|
H A D | tlb.c | 24 unsigned int prot; member in struct:host_vm_change::host_vm_op::__anon2946::__anon2947 35 unsigned int prot; member in struct:host_vm_change::host_vm_op::__anon2946::__anon2949 64 op->u.mmap.prot, op->u.mmap.fd, do_ops() 73 op->u.mprotect.len, op->u.mprotect.prot, do_ops() 88 unsigned int prot, struct host_vm_change *hvc) add_mmap() 99 (last->u.mmap.prot == prot) && (last->u.mmap.fd == fd) && add_mmap() 115 .prot = prot, add_mmap() 153 unsigned int prot, struct host_vm_change *hvc) add_mprotect() 162 (last->u.mprotect.prot == prot)) { add_mprotect() 177 .prot = prot } } }); add_mprotect() 188 int r, w, x, prot, ret = 0; update_pte_range() local 204 prot = ((r ? UM_PROT_READ : 0) | (w ? UM_PROT_WRITE : 0) | update_pte_range() 209 PAGE_SIZE, prot, hvc); update_pte_range() 213 ret = add_mprotect(addr, PAGE_SIZE, prot, hvc); update_pte_range() 391 int r, w, x, prot, err = 0; flush_tlb_page() local 420 prot = ((r ? UM_PROT_READ : 0) | (w ? UM_PROT_WRITE : 0) | flush_tlb_page() 428 err = map(mm_id, address, PAGE_SIZE, prot, fd, offset, flush_tlb_page() 434 err = protect(mm_id, address, PAGE_SIZE, prot, 1, &flush); flush_tlb_page() 87 add_mmap(unsigned long virt, unsigned long phys, unsigned long len, unsigned int prot, struct host_vm_change *hvc) add_mmap() argument 152 add_mprotect(unsigned long addr, unsigned long len, unsigned int prot, struct host_vm_change *hvc) add_mprotect() argument
|
/linux-4.1.27/include/asm-generic/ |
H A D | syscalls.h | 14 unsigned long prot, unsigned long flags, 20 unsigned long prot, unsigned long flags,
|
H A D | pgtable.h | 246 #define move_pte(pte, prot, old_addr, new_addr) (pte) 258 #define pgprot_noncached(prot) (prot) 492 static inline int track_pfn_remap(struct vm_area_struct *vma, pgprot_t *prot, track_pfn_remap() argument 503 static inline int track_pfn_insert(struct vm_area_struct *vma, pgprot_t *prot, track_pfn_insert() argument 528 extern int track_pfn_remap(struct vm_area_struct *vma, pgprot_t *prot, 531 extern int track_pfn_insert(struct vm_area_struct *vma, pgprot_t *prot, 701 int pud_set_huge(pud_t *pud, phys_addr_t addr, pgprot_t prot); 702 int pmd_set_huge(pmd_t *pmd, phys_addr_t addr, pgprot_t prot); 706 static inline int pud_set_huge(pud_t *pud, phys_addr_t addr, pgprot_t prot) pud_set_huge() argument 710 static inline int pmd_set_huge(pmd_t *pmd, phys_addr_t addr, pgprot_t prot) pmd_set_huge() argument
|
/linux-4.1.27/arch/arm/mm/ |
H A D | idmap.c | 22 unsigned long prot) idmap_add_pmd() 47 *pmd = __pmd((addr & PMD_MASK) | prot); idmap_add_pmd() 53 unsigned long prot) idmap_add_pmd() 57 addr = (addr & PMD_MASK) | prot; idmap_add_pmd() 66 unsigned long prot) idmap_add_pud() 73 idmap_add_pmd(pud, addr, next, prot); idmap_add_pud() 78 const char *text_end, unsigned long prot) identity_mapping_add() 87 prot |= PMD_TYPE_SECT | PMD_SECT_AP_WRITE | PMD_SECT_AF; identity_mapping_add() 90 prot |= PMD_BIT4; identity_mapping_add() 95 idmap_add_pud(pgd, addr, next, prot); identity_mapping_add() 21 idmap_add_pmd(pud_t *pud, unsigned long addr, unsigned long end, unsigned long prot) idmap_add_pmd() argument 52 idmap_add_pmd(pud_t *pud, unsigned long addr, unsigned long end, unsigned long prot) idmap_add_pmd() argument 65 idmap_add_pud(pgd_t *pgd, unsigned long addr, unsigned long end, unsigned long prot) idmap_add_pud() argument 77 identity_mapping_add(pgd_t *pgd, const char *text_start, const char *text_end, unsigned long prot) identity_mapping_add() argument
|
H A D | dma-mapping.c | 291 pgprot_t prot, struct page **ret_page, 295 pgprot_t prot, struct page **ret_page, 299 __dma_alloc_remap(struct page *page, size_t size, gfp_t gfp, pgprot_t prot, __dma_alloc_remap() argument 308 prot, caller); __dma_alloc_remap() 349 pgprot_t prot = pgprot_dmacoherent(PAGE_KERNEL); atomic_pool_init() local 359 ptr = __alloc_from_contiguous(NULL, atomic_pool_size, prot, atomic_pool_init() 362 ptr = __alloc_remap_buffer(NULL, atomic_pool_size, gfp, prot, atomic_pool_init() 453 pgprot_t prot = *(pgprot_t *)data; __dma_update_pte() local 455 set_pte_ext(pte, mk_pte(page, prot), 0); __dma_update_pte() 459 static void __dma_remap(struct page *page, size_t size, pgprot_t prot) __dma_remap() argument 464 apply_to_page_range(&init_mm, start, size, __dma_update_pte, &prot); __dma_remap() 469 pgprot_t prot, struct page **ret_page, __alloc_remap_buffer() 480 ptr = __dma_alloc_remap(page, size, gfp, prot, caller); __alloc_remap_buffer() 528 pgprot_t prot, struct page **ret_page, __alloc_from_contiguous() 546 ptr = __dma_alloc_remap(page, size, GFP_KERNEL, prot, caller); __alloc_from_contiguous() 552 __dma_remap(page, size, prot); __alloc_from_contiguous() 573 static inline pgprot_t __get_dma_pgprot(struct dma_attrs *attrs, pgprot_t prot) __get_dma_pgprot() argument 575 prot = dma_get_attr(DMA_ATTR_WRITE_COMBINE, attrs) ? __get_dma_pgprot() 576 pgprot_writecombine(prot) : __get_dma_pgprot() 577 pgprot_dmacoherent(prot); __get_dma_pgprot() 578 return prot; __get_dma_pgprot() 587 #define __get_dma_pgprot(attrs, prot) __pgprot(0) 588 #define __alloc_remap_buffer(dev, size, gfp, prot, ret, c, wv) NULL 590 #define __alloc_from_contiguous(dev, size, prot, ret, c, wv) NULL 612 gfp_t gfp, pgprot_t prot, bool is_coherent, __dma_alloc() 653 addr = __alloc_remap_buffer(dev, size, gfp, prot, &page, caller, want_vaddr); __dma_alloc() 655 addr = __alloc_from_contiguous(dev, size, prot, &page, caller, want_vaddr); __dma_alloc() 670 pgprot_t prot = __get_dma_pgprot(attrs, PAGE_KERNEL); arm_dma_alloc() local 676 return __dma_alloc(dev, size, handle, gfp, prot, false, arm_dma_alloc() 683 pgprot_t prot = __get_dma_pgprot(attrs, PAGE_KERNEL); arm_coherent_dma_alloc() local 689 return __dma_alloc(dev, size, handle, gfp, prot, true, arm_coherent_dma_alloc() 1226 __iommu_alloc_remap(struct page **pages, size_t size, gfp_t gfp, pgprot_t prot, __iommu_alloc_remap() argument 1230 VM_ARM_DMA_CONSISTENT | VM_USERMAP, prot, caller); __iommu_alloc_remap() 1347 pgprot_t prot = __get_dma_pgprot(attrs, PAGE_KERNEL); arm_iommu_alloc_attrs() local 1377 addr = __iommu_alloc_remap(pages, size, gfp, prot, arm_iommu_alloc_attrs() 1470 int prot; __dma_direction_to_prot() local 1474 prot = IOMMU_READ | IOMMU_WRITE; __dma_direction_to_prot() 1477 prot = IOMMU_READ; __dma_direction_to_prot() 1480 prot = IOMMU_WRITE; __dma_direction_to_prot() 1483 prot = 0; __dma_direction_to_prot() 1486 return prot; __dma_direction_to_prot() 1502 int prot; __map_sg_chunk() local 1519 prot = __dma_direction_to_prot(dir); __map_sg_chunk() 1521 ret = iommu_map(mapping->domain, iova, phys, len, prot); __map_sg_chunk() 1720 int ret, prot, len = PAGE_ALIGN(size + offset); arm_coherent_iommu_map_page() local 1726 prot = __dma_direction_to_prot(dir); arm_coherent_iommu_map_page() 1728 ret = iommu_map(mapping->domain, dma_addr, page_to_phys(page), len, prot); arm_coherent_iommu_map_page() 468 __alloc_remap_buffer(struct device *dev, size_t size, gfp_t gfp, pgprot_t prot, struct page **ret_page, const void *caller, bool want_vaddr) __alloc_remap_buffer() argument 527 __alloc_from_contiguous(struct device *dev, size_t size, pgprot_t prot, struct page **ret_page, const void *caller, bool want_vaddr) __alloc_from_contiguous() argument 611 __dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, gfp_t gfp, pgprot_t prot, bool is_coherent, struct dma_attrs *attrs, const void *caller) __dma_alloc() argument
|
H A D | init.c | 575 pmdval_t prot; member in struct:section_perm 585 .prot = PMD_SECT_XN, 592 .prot = PMD_SECT_XN, 600 .prot = PMD_SECT_XN, 613 .prot = L_PMD_SECT_RDONLY, 616 .prot = PMD_SECT_APX | PMD_SECT_AP_WRITE, 629 pmdval_t prot) section_update() 638 pmd[0] = __pmd((pmd_val(pmd[0]) & mask) | prot); section_update() 641 pmd[1] = __pmd((pmd_val(pmd[1]) & mask) | prot); section_update() 643 pmd[0] = __pmd((pmd_val(pmd[0]) & mask) | prot); section_update() 684 set_section_perms(nx_perms, prot); fix_kernmem_perms() 690 set_section_perms(ro_perms, prot); mark_rodata_ro() 700 set_section_perms(ro_perms, prot); set_kernel_text_ro() 628 section_update(unsigned long addr, pmdval_t mask, pmdval_t prot) section_update() argument
|
H A D | dump.c | 221 u64 prot = val & pg_level[level].mask; note_page() local 225 st->current_prot = prot; note_page() 227 } else if (prot != st->current_prot || level != st->level || note_page() 252 st->current_prot = prot; note_page()
|
H A D | fault-armv.c | 243 pgprot_t prot = __pgprot_modify(PAGE_KERNEL, check_writebuffer_bugs() local 246 p1 = vmap(&page, 1, VM_IOREMAP, prot); check_writebuffer_bugs() 247 p2 = vmap(&page, 1, VM_IOREMAP, prot); check_writebuffer_bugs()
|
/linux-4.1.27/lib/ |
H A D | ioremap.c | 54 unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_pte_range() 65 set_pte_at(&init_mm, addr, pte, pfn_pte(pfn, prot)); ioremap_pte_range() 72 unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_pmd_range() 87 if (pmd_set_huge(pmd, phys_addr + addr, prot)) ioremap_pmd_range() 91 if (ioremap_pte_range(pmd, addr, next, phys_addr + addr, prot)) ioremap_pmd_range() 98 unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_pud_range() 113 if (pud_set_huge(pud, phys_addr + addr, prot)) ioremap_pud_range() 117 if (ioremap_pmd_range(pud, addr, next, phys_addr + addr, prot)) ioremap_pud_range() 124 unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_page_range() 138 err = ioremap_pud_range(pgd, addr, next, phys_addr+addr, prot); ioremap_page_range() 53 ioremap_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_pte_range() argument 71 ioremap_pmd_range(pud_t *pud, unsigned long addr, unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_pmd_range() argument 97 ioremap_pud_range(pgd_t *pgd, unsigned long addr, unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_pud_range() argument 123 ioremap_page_range(unsigned long addr, unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_page_range() argument
|
/linux-4.1.27/arch/x86/mm/ |
H A D | iomap_32.c | 34 int iomap_create_wc(resource_size_t base, unsigned long size, pgprot_t *prot) iomap_create_wc() argument 46 *prot = __pgprot(__PAGE_KERNEL | cachemode2protval(pcm)); iomap_create_wc() 57 void *kmap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot) kmap_atomic_prot_pfn() argument 67 set_pte(kmap_pte - idx, pfn_pte(pfn, prot)); kmap_atomic_prot_pfn() 74 * Map 'pfn' using protections 'prot' 77 iomap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot) iomap_atomic_prot_pfn() argument 85 if (!pat_enabled && pgprot_val(prot) == iomap_atomic_prot_pfn() 87 prot = __pgprot(__PAGE_KERNEL | iomap_atomic_prot_pfn() 90 return (void __force __iomem *) kmap_atomic_prot_pfn(pfn, prot); iomap_atomic_prot_pfn()
|
H A D | dump_pagetables.c | 124 static void printk_prot(struct seq_file *m, pgprot_t prot, int level, bool dmsg) printk_prot() argument 126 pgprotval_t pr = pgprot_val(prot); printk_prot() 130 if (!pgprot_val(prot)) { printk_prot() 193 pgprotval_t prot, cur; note_page() local 201 prot = pgprot_val(new_prot) & PTE_FLAGS_MASK; note_page() 212 } else if (prot != cur || level != st->level || note_page() 275 pgprot_t prot = pte_pgprot(*start); walk_pte_level() local 278 note_page(m, st, prot, 4); walk_pte_level() 295 pgprotval_t prot = pmd_val(*start) & PTE_FLAGS_MASK; walk_pmd_level() local 298 note_page(m, st, __pgprot(prot), 3); walk_pmd_level() 327 pgprotval_t prot = pud_val(*start) & PTE_FLAGS_MASK; walk_pud_level() local 330 note_page(m, st, __pgprot(prot), 2); walk_pud_level() 365 pgprotval_t prot = pgd_val(*start) & PTE_FLAGS_MASK; ptdump_walk_pgd_level() local 368 note_page(m, &st, __pgprot(prot), 1); ptdump_walk_pgd_level()
|
H A D | ioremap.c | 91 pgprot_t prot; __ioremap_caller() local 160 prot = PAGE_KERNEL_IO; __ioremap_caller() 164 prot = __pgprot(pgprot_val(prot) | __ioremap_caller() 168 prot = __pgprot(pgprot_val(prot) | __ioremap_caller() 172 prot = __pgprot(pgprot_val(prot) | __ioremap_caller() 191 if (ioremap_page_range(vaddr, vaddr + size, phys_addr, prot)) __ioremap_caller()
|
H A D | pat.c | 712 * Internal interface to reserve a range of physical memory with prot. 714 * this func also keeps identity mapping (if any) in sync with this new prot. 800 * If the vma has a linear pfn mapping for the entire range, we get the prot 806 unsigned long prot; track_pfn_copy() local 815 if (follow_phys(vma, vma->vm_start, 0, &prot, &paddr)) { track_pfn_copy() 819 pgprot = __pgprot(prot); track_pfn_copy() 827 * prot is passed in as a parameter for the new mapping. If the vma has a 831 int track_pfn_remap(struct vm_area_struct *vma, pgprot_t *prot, track_pfn_remap() argument 841 ret = reserve_pfn_range(paddr, size, prot, 0); track_pfn_remap() 851 * For anything smaller than the vma size we set prot based on the track_pfn_remap() 864 *prot = __pgprot((pgprot_val(vma->vm_page_prot) & (~_PAGE_CACHE_MASK)) | track_pfn_remap() 870 int track_pfn_insert(struct vm_area_struct *vma, pgprot_t *prot, track_pfn_insert() argument 878 /* Set prot based on lookup */ track_pfn_insert() 880 *prot = __pgprot((pgprot_val(vma->vm_page_prot) & (~_PAGE_CACHE_MASK)) | track_pfn_insert() 895 unsigned long prot; untrack_pfn() local 903 if (follow_phys(vma, vma->vm_start, 0, &prot, &paddr)) { untrack_pfn() 914 pgprot_t pgprot_writecombine(pgprot_t prot) pgprot_writecombine() argument 917 return __pgprot(pgprot_val(prot) | pgprot_writecombine() 920 return pgprot_noncached(prot); pgprot_writecombine()
|
H A D | highmem_32.c | 33 void *kmap_atomic_prot(struct page *page, pgprot_t prot) kmap_atomic_prot() argument 48 set_pte(kmap_pte-idx, mk_pte(page, prot)); kmap_atomic_prot()
|
H A D | pgtable.c | 566 int pud_set_huge(pud_t *pud, phys_addr_t addr, pgprot_t prot) pud_set_huge() argument 578 prot = pgprot_4k_2_large(prot); pud_set_huge() 582 __pgprot(pgprot_val(prot) | _PAGE_PSE))); pud_set_huge() 587 int pmd_set_huge(pmd_t *pmd, phys_addr_t addr, pgprot_t prot) pmd_set_huge() argument 599 prot = pgprot_4k_2_large(prot); pmd_set_huge() 603 __pgprot(pgprot_val(prot) | _PAGE_PSE))); pmd_set_huge()
|
H A D | init_64.c | 331 pgprot_t prot; __init_extra_mapping() local 333 pgprot_val(prot) = pgprot_val(PAGE_KERNEL_LARGE) | __init_extra_mapping() 351 set_pmd(pmd, __pmd(phys | pgprot_val(prot))); __init_extra_mapping() 403 pgprot_t prot) phys_pte_init() 437 set_pte(pte, pfn_pte(addr >> PAGE_SHIFT, prot)); phys_pte_init() 448 unsigned long page_size_mask, pgprot_t prot) phys_pmd_init() 458 pgprot_t new_prot = prot; phys_pmd_init() 474 end, prot); phys_pmd_init() 504 __pgprot(pgprot_val(prot) | _PAGE_PSE))); phys_pmd_init() 532 pgprot_t prot = PAGE_KERNEL; phys_pud_init() local 547 page_size_mask, prot); phys_pud_init() 569 prot = pte_pgprot(pte_clrhuge(*(pte_t *)pud)); phys_pud_init() 585 prot); phys_pud_init() 402 phys_pte_init(pte_t *pte_page, unsigned long addr, unsigned long end, pgprot_t prot) phys_pte_init() argument 447 phys_pmd_init(pmd_t *pmd_page, unsigned long address, unsigned long end, unsigned long page_size_mask, pgprot_t prot) phys_pmd_init() argument
|
/linux-4.1.27/arch/s390/mm/ |
H A D | dump_pagetables.c | 62 unsigned int prot, cur; note_page() local 70 prot = new_prot; note_page() 79 } else if (prot != cur || level != st->level || note_page() 113 unsigned int prot; walk_pte_level() local 120 prot = pte_val(*pte) & (_PAGE_PROTECT | _PAGE_INVALID); walk_pte_level() 121 note_page(m, st, prot, 4); walk_pte_level() 129 unsigned int prot; walk_pmd_level() local 138 prot = pmd_val(*pmd) & _SEGMENT_ENTRY_PROTECT; walk_pmd_level() 139 note_page(m, st, prot, 3); walk_pmd_level() 151 unsigned int prot; walk_pud_level() local 160 prot = pud_val(*pud) & _REGION3_ENTRY_RO; walk_pud_level() 161 note_page(m, st, prot, 2); walk_pud_level()
|
H A D | hugetlbpage.c | 19 * prot-none, clean, old 111.000000.1 -> 00..1...1...00 __pte_to_pmd() 20 * prot-none, clean, young 111.000001.1 -> 01..1...1...00 __pte_to_pmd() 21 * prot-none, dirty, old 111.000010.1 -> 10..1...1...00 __pte_to_pmd() 22 * prot-none, dirty, young 111.000011.1 -> 11..1...1...00 __pte_to_pmd() 56 * prot-none, clean, old 00..1...1...00 -> 111.000000.1 __pmd_to_pte() 57 * prot-none, clean, young 01..1...1...00 -> 111.000001.1 __pmd_to_pte() 58 * prot-none, dirty, old 10..1...1...00 -> 111.000010.1 __pmd_to_pte() 59 * prot-none, dirty, young 11..1...1...00 -> 111.000011.1 __pmd_to_pte()
|
/linux-4.1.27/include/net/ |
H A D | protocol.h | 86 struct proto *prot; member in struct:inet_protosw 103 int inet_add_protocol(const struct net_protocol *prot, unsigned char num); 104 int inet_del_protocol(const struct net_protocol *prot, unsigned char num); 105 int inet_add_offload(const struct net_offload *prot, unsigned char num); 106 int inet_del_offload(const struct net_offload *prot, unsigned char num); 110 int udp_add_offload(struct udp_offload *prot); 111 void udp_del_offload(struct udp_offload *prot); 114 int inet6_add_protocol(const struct inet6_protocol *prot, unsigned char num); 115 int inet6_del_protocol(const struct inet6_protocol *prot, unsigned char num); 119 int inet6_add_offload(const struct net_offload *prot, unsigned char num); 120 int inet6_del_offload(const struct net_offload *prot, unsigned char num);
|
H A D | sock.h | 1083 int proto_register(struct proto *prot, int alloc_slab); 1084 void proto_unregister(struct proto *prot); 1177 struct proto *prot = sk->sk_prot; sk_leave_memory_pressure() local 1179 for (; cg_proto; cg_proto = parent_cg_proto(prot, cg_proto)) sk_leave_memory_pressure() 1192 struct proto *prot = sk->sk_prot; sk_enter_memory_pressure() local 1194 for (; cg_proto; cg_proto = parent_cg_proto(prot, cg_proto)) sk_enter_memory_pressure() 1203 long *prot = sk->sk_prot->sysctl_mem; sk_prot_mem_limits() local 1205 prot = sk->sk_cgrp->sysctl_mem; sk_prot_mem_limits() 1206 return prot[index]; sk_prot_mem_limits() 1209 static inline void memcg_memory_allocated_add(struct cg_proto *prot, memcg_memory_allocated_add() argument 1213 page_counter_charge(&prot->memory_allocated, amt); memcg_memory_allocated_add() 1215 if (page_counter_read(&prot->memory_allocated) > memcg_memory_allocated_add() 1216 prot->memory_allocated.limit) memcg_memory_allocated_add() 1220 static inline void memcg_memory_allocated_sub(struct cg_proto *prot, memcg_memory_allocated_sub() argument 1223 page_counter_uncharge(&prot->memory_allocated, amt); memcg_memory_allocated_sub() 1229 struct proto *prot = sk->sk_prot; sk_memory_allocated() local 1234 return atomic_long_read(prot->memory_allocated); sk_memory_allocated() 1240 struct proto *prot = sk->sk_prot; sk_memory_allocated_add() local 1245 atomic_long_add_return(amt, prot->memory_allocated); sk_memory_allocated_add() 1249 return atomic_long_add_return(amt, prot->memory_allocated); sk_memory_allocated_add() 1255 struct proto *prot = sk->sk_prot; sk_memory_allocated_sub() local 1260 atomic_long_sub(amt, prot->memory_allocated); sk_memory_allocated_sub() 1265 struct proto *prot = sk->sk_prot; sk_sockets_allocated_dec() local 1270 for (; cg_proto; cg_proto = parent_cg_proto(prot, cg_proto)) sk_sockets_allocated_dec() 1274 percpu_counter_dec(prot->sockets_allocated); sk_sockets_allocated_dec() 1279 struct proto *prot = sk->sk_prot; sk_sockets_allocated_inc() local 1284 for (; cg_proto; cg_proto = parent_cg_proto(prot, cg_proto)) sk_sockets_allocated_inc() 1288 percpu_counter_inc(prot->sockets_allocated); sk_sockets_allocated_inc() 1294 struct proto *prot = sk->sk_prot; sk_sockets_allocated_read_positive() local 1299 return percpu_counter_read_positive(prot->sockets_allocated); sk_sockets_allocated_read_positive() 1303 proto_sockets_allocated_sum_positive(struct proto *prot) proto_sockets_allocated_sum_positive() argument 1305 return percpu_counter_sum_positive(prot->sockets_allocated); proto_sockets_allocated_sum_positive() 1309 proto_memory_allocated(struct proto *prot) proto_memory_allocated() argument 1311 return atomic_long_read(prot->memory_allocated); proto_memory_allocated() 1315 proto_memory_pressure(struct proto *prot) proto_memory_pressure() argument 1317 if (!prot->memory_pressure) proto_memory_pressure() 1319 return !!*prot->memory_pressure; proto_memory_pressure() 1325 void sock_prot_inuse_add(struct net *net, struct proto *prot, int inc); 1328 static inline void sock_prot_inuse_add(struct net *net, struct proto *prot, sock_prot_inuse_add() argument 1528 struct proto *prot);
|
/linux-4.1.27/arch/tile/gxio/ |
H A D | kiorpc.c | 33 pgprot_t mmio_base, prot = { 0 }; iorpc_ioremap() local 55 prot = hv_pte_set_lotar(prot, hv_pte_get_lotar(mmio_base)); iorpc_ioremap() 58 return ioremap_prot(PFN_PHYS(pfn), size, prot); iorpc_ioremap()
|
/linux-4.1.27/arch/microblaze/kernel/ |
H A D | sys_microblaze.c | 37 unsigned long, prot, unsigned long, flags, unsigned long, fd, SYSCALL_DEFINE6() 43 return sys_mmap_pgoff(addr, len, prot, flags, fd, pgoff >> PAGE_SHIFT); SYSCALL_DEFINE6() 47 unsigned long, prot, unsigned long, flags, unsigned long, fd, SYSCALL_DEFINE6() 53 return sys_mmap_pgoff(addr, len, prot, flags, fd, SYSCALL_DEFINE6()
|
/linux-4.1.27/arch/score/kernel/ |
H A D | sys_score.c | 37 sys_mmap2(unsigned long addr, unsigned long len, unsigned long prot, sys_mmap2() argument 40 return sys_mmap_pgoff(addr, len, prot, flags, fd, pgoff); sys_mmap2() 44 sys_mmap(unsigned long addr, unsigned long len, unsigned long prot, sys_mmap() argument 49 return sys_mmap_pgoff(addr, len, prot, flags, fd, offset >> PAGE_SHIFT); sys_mmap()
|
/linux-4.1.27/arch/arm/include/asm/ |
H A D | fixmap.h | 21 void __set_fixmap(enum fixed_addresses idx, phys_addr_t phys, pgprot_t prot);
|
H A D | pgtable.h | 113 #define __pgprot_modify(prot,mask,bits) \ 114 __pgprot((pgprot_val(prot) & ~(mask)) | (bits)) 116 #define pgprot_noncached(prot) \ 117 __pgprot_modify(prot, L_PTE_MT_MASK, L_PTE_MT_UNCACHED) 119 #define pgprot_writecombine(prot) \ 120 __pgprot_modify(prot, L_PTE_MT_MASK, L_PTE_MT_BUFFERABLE) 122 #define pgprot_stronglyordered(prot) \ 123 __pgprot_modify(prot, L_PTE_MT_MASK, L_PTE_MT_UNCACHED) 126 #define pgprot_dmacoherent(prot) \ 127 __pgprot_modify(prot, L_PTE_MT_MASK, L_PTE_MT_BUFFERABLE | L_PTE_XN) 133 #define pgprot_dmacoherent(prot) \ 134 __pgprot_modify(prot, L_PTE_MT_MASK, L_PTE_MT_UNCACHED | L_PTE_XN) 210 #define pfn_pte(pfn,prot) __pte(__pfn_to_phys(pfn) | pgprot_val(prot)) 213 #define mk_pte(page,prot) pfn_pte(page_to_pfn(page), prot) 255 static inline pte_t clear_pte_bit(pte_t pte, pgprot_t prot) clear_pte_bit() argument 257 pte_val(pte) &= ~pgprot_val(prot); clear_pte_bit() 261 static inline pte_t set_pte_bit(pte_t pte, pgprot_t prot) set_pte_bit() argument 263 pte_val(pte) |= pgprot_val(prot); set_pte_bit()
|
H A D | pgalloc.h | 132 pmdval_t prot) __pmd_populate() 134 pmdval_t pmdval = (pte + PTE_HWTABLE_OFF) | prot; __pmd_populate() 161 pmdval_t prot; pmd_populate() local 164 prot = user_pmd_table; pmd_populate() 166 prot = _PAGE_USER_TABLE; pmd_populate() 168 __pmd_populate(pmdp, page_to_phys(ptep), prot); pmd_populate() 131 __pmd_populate(pmd_t *pmdp, phys_addr_t pte, pmdval_t prot) __pmd_populate() argument
|
H A D | pgtable-nommu.h | 64 * Mark the prot value as uncacheable and unbufferable. 66 #define pgprot_noncached(prot) __pgprot(0) 67 #define pgprot_writecombine(prot) __pgprot(0) 68 #define pgprot_dmacoherent(prot) __pgprot(0)
|
H A D | pgtable-3level.h | 257 #define pfn_pmd(pfn,prot) (__pmd(((phys_addr_t)(pfn) << PAGE_SHIFT) | pgprot_val(prot))) 258 #define mk_pmd(page,prot) pfn_pmd(page_to_pfn(page),prot)
|
/linux-4.1.27/arch/arm64/include/asm/ |
H A D | pgtable.h | 122 #define pfn_pte(pfn,prot) (__pte(((phys_addr_t)(pfn) << PAGE_SHIFT) | pgprot_val(prot))) 153 static inline pte_t clear_pte_bit(pte_t pte, pgprot_t prot) clear_pte_bit() argument 155 pte_val(pte) &= ~pgprot_val(prot); clear_pte_bit() 159 static inline pte_t set_pte_bit(pte_t pte, pgprot_t prot) set_pte_bit() argument 161 pte_val(pte) |= pgprot_val(prot); set_pte_bit() 268 static inline pgprot_t mk_sect_prot(pgprot_t prot) mk_sect_prot() argument 270 return __pgprot(pgprot_val(prot) & ~PTE_TABLE_BIT); mk_sect_prot() 305 #define pfn_pmd(pfn,prot) (__pmd(((phys_addr_t)(pfn) << PAGE_SHIFT) | pgprot_val(prot))) 306 #define mk_pmd(page,prot) pfn_pmd(page_to_pfn(page),prot) 318 #define __pgprot_modify(prot,mask,bits) \ 319 __pgprot((pgprot_val(prot) & ~(mask)) | (bits)) 322 * Mark the prot value as uncacheable and unbufferable. 324 #define pgprot_noncached(prot) \ 325 __pgprot_modify(prot, PTE_ATTRINDX_MASK, PTE_ATTRINDX(MT_DEVICE_nGnRnE) | PTE_PXN | PTE_UXN) 326 #define pgprot_writecombine(prot) \ 327 __pgprot_modify(prot, PTE_ATTRINDX_MASK, PTE_ATTRINDX(MT_NORMAL_NC) | PTE_PXN | PTE_UXN) 328 #define pgprot_device(prot) \ 329 __pgprot_modify(prot, PTE_ATTRINDX_MASK, PTE_ATTRINDX(MT_DEVICE_nGnRE) | PTE_PXN | PTE_UXN) 377 #define mk_pte(page,prot) pfn_pte(page_to_pfn(page),prot)
|
H A D | mmu.h | 36 pgprot_t prot);
|
H A D | pgalloc.h | 111 pmdval_t prot) __pmd_populate() 113 set_pmd(pmdp, __pmd(pte | prot)); __pmd_populate() 110 __pmd_populate(pmd_t *pmdp, phys_addr_t pte, pmdval_t prot) __pmd_populate() argument
|
H A D | fixmap.h | 68 extern void __set_fixmap(enum fixed_addresses idx, phys_addr_t phys, pgprot_t prot);
|
/linux-4.1.27/arch/mn10300/kernel/ |
H A D | sys_mn10300.c | 27 unsigned long prot, unsigned long flags, old_mmap() 32 return sys_mmap_pgoff(addr, len, prot, flags, fd, offset >> PAGE_SHIFT); old_mmap() 26 old_mmap(unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long fd, unsigned long offset) old_mmap() argument
|
/linux-4.1.27/arch/cris/kernel/ |
H A D | sys_cris.c | 30 sys_mmap2(unsigned long addr, unsigned long len, unsigned long prot, sys_mmap2() argument 34 return sys_mmap_pgoff(addr, len, prot, flags, fd, pgoff); sys_mmap2()
|
/linux-4.1.27/arch/ia64/include/asm/ |
H A D | unistd.h | 40 int prot, int flags, 44 int prot, int flags,
|
H A D | pgtable.h | 351 #define pgprot_cacheable(prot) __pgprot((pgprot_val(prot) & ~_PAGE_MA_MASK) | _PAGE_MA_WB) 352 #define pgprot_noncached(prot) __pgprot((pgprot_val(prot) & ~_PAGE_MA_MASK) | _PAGE_MA_UC) 353 #define pgprot_writecombine(prot) __pgprot((pgprot_val(prot) & ~_PAGE_MA_MASK) | _PAGE_MA_WC)
|
/linux-4.1.27/arch/x86/um/shared/sysdep/ |
H A D | syscalls_64.h | 27 unsigned long prot, unsigned long flags,
|
/linux-4.1.27/arch/sh/kernel/ |
H A D | sys_sh.c | 32 unsigned long prot, unsigned long flags, old_mmap() 37 return sys_mmap_pgoff(addr, len, prot, flags, fd, off>>PAGE_SHIFT); old_mmap() 41 unsigned long prot, unsigned long flags, sys_mmap2() 53 return sys_mmap_pgoff(addr, len, prot, flags, fd, pgoff); sys_mmap2() 31 old_mmap(unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, int fd, unsigned long off) old_mmap() argument 40 sys_mmap2(unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long fd, unsigned long pgoff) sys_mmap2() argument
|
/linux-4.1.27/arch/m32r/include/asm/ |
H A D | pgtable-2level.h | 71 #define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) 72 #define pfn_pmd(pfn, prot) __pmd(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
|
H A D | pgtable.h | 267 unsigned long prot = pgprot_val(_prot); pgprot_noncached() local 269 prot |= _PAGE_NONCACHABLE; pgprot_noncached() 270 return __pgprot(prot); pgprot_noncached() 273 #define pgprot_writecombine(prot) pgprot_noncached(prot)
|
/linux-4.1.27/arch/arc/mm/ |
H A D | ioremap.c | 47 pgprot_t prot = __pgprot(flags); ioremap_prot() local 59 prot = pgprot_noncached(prot); ioremap_prot() 75 (unsigned long)vaddr + size, paddr, prot)) { ioremap_prot()
|
/linux-4.1.27/arch/sh/boards/mach-landisk/ |
H A D | setup.c | 57 pgprot_t prot; landisk_devices_setup() local 63 prot = PAGE_KERNEL_PCC(1, _PAGE_PCC_IO16); landisk_devices_setup() 64 cf_ide_base = ioremap_prot(paddrbase, PAGE_SIZE, pgprot_val(prot)); landisk_devices_setup()
|
/linux-4.1.27/arch/sh/boards/mach-lboxre2/ |
H A D | setup.c | 53 pgprot_t prot; lboxre2_devices_setup() local 59 prot = PAGE_KERNEL_PCC(1, _PAGE_PCC_IO16); lboxre2_devices_setup() 60 cf0_io_base = (u32)ioremap_prot(paddrbase, psize, pgprot_val(prot)); lboxre2_devices_setup()
|
/linux-4.1.27/arch/sh/boards/mach-sh03/ |
H A D | setup.c | 78 pgprot_t prot; sh03_devices_setup() local 84 prot = PAGE_KERNEL_PCC(1, _PAGE_PCC_IO16); sh03_devices_setup() 85 cf_ide_base = ioremap_prot(paddrbase, PAGE_SIZE, pgprot_val(prot)); sh03_devices_setup()
|
/linux-4.1.27/fs/affs/ |
H A D | amigaffs.c | 388 prot_to_mode(u32 prot) prot_to_mode() argument 392 if (!(prot & FIBF_NOWRITE)) prot_to_mode() 394 if (!(prot & FIBF_NOREAD)) prot_to_mode() 396 if (!(prot & FIBF_NOEXECUTE)) prot_to_mode() 398 if (prot & FIBF_GRP_WRITE) prot_to_mode() 400 if (prot & FIBF_GRP_READ) prot_to_mode() 402 if (prot & FIBF_GRP_EXECUTE) prot_to_mode() 404 if (prot & FIBF_OTR_WRITE) prot_to_mode() 406 if (prot & FIBF_OTR_READ) prot_to_mode() 408 if (prot & FIBF_OTR_EXECUTE) prot_to_mode() 417 u32 prot = AFFS_I(inode)->i_protect; mode_to_prot() local 421 prot |= FIBF_NOEXECUTE; mode_to_prot() 423 prot |= FIBF_NOREAD; mode_to_prot() 425 prot |= FIBF_NOWRITE; mode_to_prot() 427 prot |= FIBF_GRP_EXECUTE; mode_to_prot() 429 prot |= FIBF_GRP_READ; mode_to_prot() 431 prot |= FIBF_GRP_WRITE; mode_to_prot() 433 prot |= FIBF_OTR_EXECUTE; mode_to_prot() 435 prot |= FIBF_OTR_READ; mode_to_prot() 437 prot |= FIBF_OTR_WRITE; mode_to_prot() 439 AFFS_I(inode)->i_protect = prot; mode_to_prot()
|
H A D | inode.c | 24 u32 prot; affs_iget() local 49 prot = be32_to_cpu(tail->protect); affs_iget() 56 AFFS_I(inode)->i_protect = prot; affs_iget() 72 inode->i_mode = prot_to_mode(prot); affs_iget()
|
/linux-4.1.27/include/xen/ |
H A D | xen-ops.h | 39 * @prot: page protection mask 52 int *err_ptr, pgprot_t prot, 61 * @prot: page protection mask 71 pgprot_t prot, unsigned domid, 78 int *err_ptr, pgprot_t prot,
|
/linux-4.1.27/arch/parisc/kernel/ |
H A D | entry.S | 50 /* space_to_prot macro creates a prot id from a space id */ 53 .macro space_to_prot spc prot 54 depd,z \spc,62,31,\prot 57 .macro space_to_prot spc prot 58 extrd,u \spc,(64 - (SPACEID_SHIFT)),32,\prot 509 /* Drop prot bits and convert to page addr for iitlbt and idtlbt */ 517 /* Convert the pte and prot to tlb insertion values. How 519 .macro make_insert_tlb spc,pte,prot 520 space_to_prot \spc \prot /* create prot id from space */ 533 depd \pte,8,7,\prot 539 depdi 7,11,3,\prot 544 depd %r0,11,2,\prot /* If Gateway, Set PL2 to 0 */ 553 depdi 1,12,1,\prot 555 /* Drop prot bits and convert to page addr for iitlbt and idtlbt */ 562 .macro make_insert_tlb_11 spc,pte,prot 563 zdep \spc,30,15,\prot 564 dep \pte,8,7,\prot 566 depi 1,12,1,\prot 568 depi 7,11,3,\prot /* Set for user space (1 rsvd for read) */ 570 depi 0,11,2,\prot /* If Gateway, Set PL2 to 0 */ 572 /* Get rid of prot bits and convert to page addr for iitlba */ 596 .macro do_alias spc,tmp,tmp1,va,pte,prot,fault,patype 617 ldi (_PAGE_REFTRAP|_PAGE_READ|_PAGE_WRITE),\prot 620 * instruction) nullify the next load so \prot is only T. 624 ldi (_PAGE_DIRTY|_PAGE_READ|_PAGE_WRITE),\prot 626 depd,z \prot,8,7,\prot 629 depw,z \prot,8,7,\prot 1134 prot = r17 /* prot bits */ define 1150 make_insert_tlb spc,pte,prot 1152 idtlbt pte,prot 1159 do_alias spc,t0,t1,va,pte,prot,dtlb_fault,20 1161 idtlbt pte,prot 1176 make_insert_tlb spc,pte,prot 1178 idtlbt pte,prot 1185 do_alias spc,t0,t1,va,pte,prot,nadtlb_emulate,20 1187 idtlbt pte,prot 1204 make_insert_tlb_11 spc,pte,prot 1210 idtlbp prot,(%sr1,va) 1219 do_alias spc,t0,t1,va,pte,prot,dtlb_fault,11 1222 idtlbp prot,(va) 1237 make_insert_tlb_11 spc,pte,prot 1243 idtlbp prot,(%sr1,va) 1252 do_alias spc,t0,t1,va,pte,prot,nadtlb_emulate,11 1255 idtlbp prot,(va) 1270 make_insert_tlb spc,pte,prot 1274 idtlbt pte,prot 1281 do_alias spc,t0,t1,va,pte,prot,dtlb_fault,20 1283 idtlbt pte,prot 1298 make_insert_tlb spc,pte,prot 1302 idtlbt pte,prot 1309 do_alias spc,t0,t1,va,pte,prot,nadtlb_emulate,20 1311 idtlbt pte,prot 1407 make_insert_tlb spc,pte,prot 1409 iitlbt pte,prot 1431 make_insert_tlb spc,pte,prot 1433 iitlbt pte,prot 1440 do_alias spc,t0,t1,va,pte,prot,naitlb_fault,20 1442 iitlbt pte,prot 1459 make_insert_tlb_11 spc,pte,prot 1465 iitlbp prot,(%sr1,va) 1483 make_insert_tlb_11 spc,pte,prot 1489 iitlbp prot,(%sr1,va) 1498 do_alias spc,t0,t1,va,pte,prot,itlb_fault,11 1501 iitlbp prot,(%sr0, va) 1517 make_insert_tlb spc,pte,prot 1521 iitlbt pte,prot 1537 make_insert_tlb spc,pte,prot 1541 iitlbt pte,prot 1548 do_alias spc,t0,t1,va,pte,prot,naitlb_fault,20 1550 iitlbt pte,prot 1569 make_insert_tlb spc,pte,prot 1571 idtlbt pte,prot 1589 make_insert_tlb_11 spc,pte,prot 1595 idtlbp prot,(%sr1,va) 1613 make_insert_tlb spc,pte,prot 1617 idtlbt pte,prot
|
H A D | pci.c | 226 unsigned long prot; pci_mmap_page_range() local 242 prot = pgprot_val(vma->vm_page_prot); pci_mmap_page_range() 243 prot |= _PAGE_NO_CACHE; pci_mmap_page_range() 244 vma->vm_page_prot = __pgprot(prot); pci_mmap_page_range()
|
H A D | sys_parisc.c | 265 unsigned long prot, unsigned long flags, unsigned long fd, sys_mmap2() 270 return sys_mmap_pgoff(addr, len, prot, flags, fd, sys_mmap2() 275 unsigned long prot, unsigned long flags, unsigned long fd, sys_mmap() 279 return sys_mmap_pgoff(addr, len, prot, flags, fd, sys_mmap() 371 asmlinkage unsigned long sys_alloc_hugepages(int key, unsigned long addr, unsigned long len, int prot, int flag) sys_alloc_hugepages() argument 264 sys_mmap2(unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long fd, unsigned long pgoff) sys_mmap2() argument 274 sys_mmap(unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long fd, unsigned long offset) sys_mmap() argument
|
/linux-4.1.27/arch/powerpc/kernel/ |
H A D | syscalls.c | 45 unsigned long prot, unsigned long flags, do_mmap2() 50 if (!arch_validate_prot(prot)) do_mmap2() 59 ret = sys_mmap_pgoff(addr, len, prot, flags, fd, off); do_mmap2() 65 unsigned long prot, unsigned long flags, sys_mmap2() 68 return do_mmap2(addr, len, prot, flags, fd, pgoff, PAGE_SHIFT-12); sys_mmap2() 72 unsigned long prot, unsigned long flags, sys_mmap() 75 return do_mmap2(addr, len, prot, flags, fd, offset, PAGE_SHIFT); sys_mmap() 44 do_mmap2(unsigned long addr, size_t len, unsigned long prot, unsigned long flags, unsigned long fd, unsigned long off, int shift) do_mmap2() argument 64 sys_mmap2(unsigned long addr, size_t len, unsigned long prot, unsigned long flags, unsigned long fd, unsigned long pgoff) sys_mmap2() argument 71 sys_mmap(unsigned long addr, size_t len, unsigned long prot, unsigned long flags, unsigned long fd, off_t offset) sys_mmap() argument
|
H A D | sys_ppc32.c | 65 unsigned long prot, unsigned long flags, compat_sys_mmap2() 69 return sys_mmap(addr, len, prot, flags, fd, pgoff << 12); compat_sys_mmap2() 64 compat_sys_mmap2(unsigned long addr, size_t len, unsigned long prot, unsigned long flags, unsigned long fd, unsigned long pgoff) compat_sys_mmap2() argument
|
/linux-4.1.27/arch/m68k/include/asm/ |
H A D | pgtable_mm.h | 155 # define pgprot_noncached(prot) (__pgprot(pgprot_val(prot) | CF_PAGE_NOCACHE)) 162 #define pgprot_noncached(prot) \ 164 ? (__pgprot(pgprot_val(prot) | __SUN3_PAGE_NOCACHE)) \ 166 ? (__pgprot(pgprot_val(prot) | _PAGE_NOCACHE030)) \ 168 ? (__pgprot((pgprot_val(prot) & _CACHEMASK040) | _PAGE_NOCACHE_S)) \ 169 : (prot)))
|
/linux-4.1.27/arch/arm64/mm/ |
H A D | mmu.c | 91 pgprot_t prot, alloc_init_pte() 107 set_pte(pte, pfn_pte(pfn, prot)); alloc_init_pte() 115 pgprot_t prot = __pgprot(pud_val(*old_pud) ^ addr); split_pud() local 119 set_pmd(pmd, __pmd(addr | prot)); split_pud() 126 phys_addr_t phys, pgprot_t prot, alloc_init_pmd() 156 pgprot_val(mk_sect_prot(prot)))); alloc_init_pmd() 171 prot, alloc); alloc_init_pmd() 191 phys_addr_t phys, pgprot_t prot, alloc_init_pud() 213 pgprot_val(mk_sect_prot(prot)))); alloc_init_pud() 231 alloc_init_pmd(mm, pud, addr, next, phys, prot, alloc); alloc_init_pud() 243 phys_addr_t size, pgprot_t prot, __create_mapping() 254 alloc_init_pud(mm, pgd, addr, next, phys, prot, alloc); __create_mapping() 270 phys_addr_t size, pgprot_t prot) create_mapping() 278 size, prot, early_alloc); create_mapping() 283 pgprot_t prot) create_pgd_mapping() 285 __create_mapping(mm, pgd_offset(mm, virt), phys, virt, size, prot, create_pgd_mapping() 290 phys_addr_t size, pgprot_t prot) create_mapping_late() 299 phys, virt, size, prot, late_alloc); create_mapping_late() 89 alloc_init_pte(pmd_t *pmd, unsigned long addr, unsigned long end, unsigned long pfn, pgprot_t prot, void *(*alloc)(unsigned long size)) alloc_init_pte() argument 124 alloc_init_pmd(struct mm_struct *mm, pud_t *pud, unsigned long addr, unsigned long end, phys_addr_t phys, pgprot_t prot, void *(*alloc)(unsigned long size)) alloc_init_pmd() argument 189 alloc_init_pud(struct mm_struct *mm, pgd_t *pgd, unsigned long addr, unsigned long end, phys_addr_t phys, pgprot_t prot, void *(*alloc)(unsigned long size)) alloc_init_pud() argument 241 __create_mapping(struct mm_struct *mm, pgd_t *pgd, phys_addr_t phys, unsigned long virt, phys_addr_t size, pgprot_t prot, void *(*alloc)(unsigned long size)) __create_mapping() argument 269 create_mapping(phys_addr_t phys, unsigned long virt, phys_addr_t size, pgprot_t prot) create_mapping() argument 281 create_pgd_mapping(struct mm_struct *mm, phys_addr_t phys, unsigned long virt, phys_addr_t size, pgprot_t prot) create_pgd_mapping() argument 289 create_mapping_late(phys_addr_t phys, unsigned long virt, phys_addr_t size, pgprot_t prot) create_mapping_late() argument
|
H A D | ioremap.c | 33 pgprot_t prot, void *caller) __ioremap_caller() 67 err = ioremap_page_range(addr, addr + size, phys_addr, prot); __ioremap_caller() 76 void __iomem *__ioremap(phys_addr_t phys_addr, size_t size, pgprot_t prot) __ioremap() argument 78 return __ioremap_caller(phys_addr, size, prot, __ioremap() 32 __ioremap_caller(phys_addr_t phys_addr, size_t size, pgprot_t prot, void *caller) __ioremap_caller() argument
|
H A D | dump.c | 188 u64 prot = val & pg_level[level].mask; note_page() local 192 st->current_prot = prot; note_page() 195 } else if (prot != st->current_prot || level != st->level || note_page() 222 st->current_prot = prot; note_page()
|
H A D | dma-mapping.c | 34 static pgprot_t __get_dma_pgprot(struct dma_attrs *attrs, pgprot_t prot, __get_dma_pgprot() argument 38 return pgprot_writecombine(prot); __get_dma_pgprot() 39 return prot; __get_dma_pgprot() 358 pgprot_t prot = __pgprot(PROT_NORMAL_NC); atomic_pool_init() local 382 VM_USERMAP, prot, atomic_pool_init); atomic_pool_init()
|
/linux-4.1.27/drivers/vfio/ |
H A D | vfio_iommu_type1.c | 68 int prot; /* IOMMU_CACHE */ member in struct:vfio_domain 77 int prot; /* IOMMU_READ/WRITE */ member in struct:vfio_dma 219 static int put_pfn(unsigned long pfn, int prot) put_pfn() argument 223 if (prot & IOMMU_WRITE) put_pfn() 231 static int vaddr_get_pfn(unsigned long vaddr, int prot, unsigned long *pfn) vaddr_get_pfn() argument 237 if (get_user_pages_fast(vaddr, 1, !!(prot & IOMMU_WRITE), page) == 1) { vaddr_get_pfn() 263 int prot, unsigned long *pfn_base) vfio_pin_pages() 273 ret = vaddr_get_pfn(vaddr, prot, pfn_base); vfio_pin_pages() 280 put_pfn(*pfn_base, prot); vfio_pin_pages() 296 ret = vaddr_get_pfn(vaddr, prot, &pfn); vfio_pin_pages() 302 put_pfn(pfn, prot); vfio_pin_pages() 308 put_pfn(pfn, prot); vfio_pin_pages() 322 int prot, bool do_accounting) vfio_unpin_pages() 328 unlocked += put_pfn(pfn++, prot); vfio_unpin_pages() 387 dma->prot, false); vfio_unmap_unpin() 502 unsigned long pfn, long npage, int prot) map_try_harder() 510 PAGE_SIZE, prot | domain->prot); map_try_harder() 522 unsigned long pfn, long npage, int prot) vfio_iommu_map() 529 npage << PAGE_SHIFT, prot | d->prot); vfio_iommu_map() 532 map_try_harder(d, iova, pfn, npage, prot)) vfio_iommu_map() 555 int ret = 0, prot = 0; vfio_dma_do_map() local 570 prot |= IOMMU_WRITE; vfio_dma_do_map() 572 prot |= IOMMU_READ; vfio_dma_do_map() 574 if (!prot || !size || (size | iova | vaddr) & mask) vfio_dma_do_map() 596 dma->prot = prot; vfio_dma_do_map() 604 size >> PAGE_SHIFT, prot, &pfn); vfio_dma_do_map() 612 ret = vfio_iommu_map(iommu, iova + dma->size, pfn, npage, prot); vfio_dma_do_map() 614 vfio_unpin_pages(pfn, npage, prot, true); vfio_dma_do_map() 680 size, dma->prot | domain->prot); vfio_iommu_replay() 711 IOMMU_READ | IOMMU_WRITE | domain->prot); vfio_test_domain_fgsp() 790 domain->prot |= IOMMU_CACHE; vfio_iommu_type1_attach_group() 801 d->prot == domain->prot) { vfio_iommu_type1_attach_group() 945 if (!(domain->prot & IOMMU_CACHE)) { vfio_domains_have_iommu_cache() 262 vfio_pin_pages(unsigned long vaddr, long npage, int prot, unsigned long *pfn_base) vfio_pin_pages() argument 321 vfio_unpin_pages(unsigned long pfn, long npage, int prot, bool do_accounting) vfio_unpin_pages() argument 501 map_try_harder(struct vfio_domain *domain, dma_addr_t iova, unsigned long pfn, long npage, int prot) map_try_harder() argument 521 vfio_iommu_map(struct vfio_iommu *iommu, dma_addr_t iova, unsigned long pfn, long npage, int prot) vfio_iommu_map() argument
|
/linux-4.1.27/arch/sparc/kernel/ |
H A D | sys_sparc_32.c | 99 unsigned long prot, unsigned long flags, unsigned long fd, sys_mmap2() 104 return sys_mmap_pgoff(addr, len, prot, flags, fd, sys_mmap2() 109 unsigned long prot, unsigned long flags, unsigned long fd, sys_mmap() 113 return sys_mmap_pgoff(addr, len, prot, flags, fd, off >> PAGE_SHIFT); sys_mmap() 117 unsigned long prot, unsigned long pgoff, sparc_remap_file_pages() 123 return sys_remap_file_pages(start, size, prot, sparc_remap_file_pages() 98 sys_mmap2(unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long fd, unsigned long pgoff) sys_mmap2() argument 108 sys_mmap(unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long fd, unsigned long off) sys_mmap() argument 116 sparc_remap_file_pages(unsigned long start, unsigned long size, unsigned long prot, unsigned long pgoff, unsigned long flags) sparc_remap_file_pages() argument
|
H A D | systbls.h | 17 unsigned long prot, unsigned long flags, 23 unsigned long prot, unsigned long flags, 26 unsigned long prot, unsigned long pgoff,
|
H A D | pci_sun4v.c | 40 unsigned long prot; /* IOMMU page protections */ member in struct:iommu_batch 50 static inline void iommu_batch_start(struct device *dev, unsigned long prot, unsigned long entry) iommu_batch_start() argument 55 p->prot = prot; iommu_batch_start() 65 unsigned long prot = p->prot; iommu_batch_flush() local 74 npages, prot, __pa(pglist)); iommu_batch_flush() 81 npages, prot, __pa(pglist), num); iommu_batch_flush() 244 unsigned long prot; dma_4v_map_page() local 265 prot = HV_PCI_MAP_ATTR_READ; dma_4v_map_page() 267 prot |= HV_PCI_MAP_ATTR_WRITE; dma_4v_map_page() 271 iommu_batch_start(dev, prot, entry); dma_4v_map_page() 328 unsigned long flags, handle, prot; dma_4v_map_sg() local 343 prot = HV_PCI_MAP_ATTR_READ; dma_4v_map_sg() 345 prot |= HV_PCI_MAP_ATTR_WRITE; dma_4v_map_sg() 357 iommu_batch_start(dev, prot, ~0UL); dma_4v_map_sg()
|
/linux-4.1.27/arch/sh/mm/ |
H A D | ioremap_fixed.c | 47 ioremap_fixed(phys_addr_t phys_addr, unsigned long size, pgprot_t prot) ioremap_fixed() argument 88 pgprot_val(prot) |= _PAGE_WIRED; ioremap_fixed() 89 __set_fixmap(idx, phys_addr, prot); ioremap_fixed()
|
H A D | init.c | 72 static void set_pte_phys(unsigned long addr, unsigned long phys, pgprot_t prot) set_pte_phys() argument 82 set_pte(pte, pfn_pte(phys >> PAGE_SHIFT, prot)); set_pte_phys() 85 if (pgprot_val(prot) & _PAGE_WIRED) set_pte_phys() 89 static void clear_pte_phys(unsigned long addr, pgprot_t prot) clear_pte_phys() argument 95 if (pgprot_val(prot) & _PAGE_WIRED) clear_pte_phys() 102 void __set_fixmap(enum fixed_addresses idx, unsigned long phys, pgprot_t prot) __set_fixmap() argument 111 set_pte_phys(address, phys, prot); __set_fixmap() 114 void __clear_fixmap(enum fixed_addresses idx, pgprot_t prot) __clear_fixmap() argument 123 clear_pte_phys(address, prot); __clear_fixmap()
|
H A D | nommu.c | 98 void __set_fixmap(enum fixed_addresses idx, unsigned long phys, pgprot_t prot) __set_fixmap() argument
|
H A D | pmb.c | 117 static inline unsigned long pgprot_to_pmb_flags(pgprot_t prot) pgprot_to_pmb_flags() argument 120 u64 flags = pgprot_val(prot); pgprot_to_pmb_flags() 207 static inline bool pmb_prot_valid(pgprot_t prot) pmb_prot_valid() argument 209 return (pgprot_val(prot) & _PAGE_USER) == 0; pmb_prot_valid() 336 unsigned long size, pgprot_t prot) pmb_bolt_mapping() 355 pmb_flags = pgprot_to_pmb_flags(prot); pmb_bolt_mapping() 411 pgprot_t prot, void *caller) pmb_remap_caller() 428 if (!pmb_prot_valid(prot)) pmb_remap_caller() 454 ret = pmb_bolt_mapping(vaddr, phys, size, prot); pmb_remap_caller() 335 pmb_bolt_mapping(unsigned long vaddr, phys_addr_t phys, unsigned long size, pgprot_t prot) pmb_bolt_mapping() argument 410 pmb_remap_caller(phys_addr_t phys, unsigned long size, pgprot_t prot, void *caller) pmb_remap_caller() argument
|
/linux-4.1.27/arch/metag/mm/ |
H A D | ioremap.c | 34 pgprot_t prot; __ioremap() local 52 prot = __pgprot(_PAGE_PRESENT | _PAGE_WRITE | _PAGE_DIRTY | __ioremap() 64 if (ioremap_page_range(addr, addr + size, phys_addr, prot)) { __ioremap()
|
/linux-4.1.27/arch/avr32/mm/ |
H A D | ioremap.c | 28 pgprot_t prot; __ioremap() local 56 prot = __pgprot(_PAGE_PRESENT | _PAGE_GLOBAL | _PAGE_RW | _PAGE_DIRTY __ioremap() 67 if (ioremap_page_range(addr, addr + size, phys_addr, prot)) { __ioremap()
|
/linux-4.1.27/arch/cris/arch-v32/drivers/pci/ |
H A D | bios.c | 20 unsigned long prot; pci_mmap_page_range() local 25 prot = pgprot_val(vma->vm_page_prot); pci_mmap_page_range() 26 vma->vm_page_prot = __pgprot(prot); pci_mmap_page_range()
|
/linux-4.1.27/arch/powerpc/mm/ |
H A D | Makefile | 35 obj-$(CONFIG_PPC_SUBPAGE_PROT) += subpage-prot.o
|
H A D | highmem.c | 32 void *kmap_atomic_prot(struct page *page, pgprot_t prot) kmap_atomic_prot() argument 48 __set_pte_at(&init_mm, vaddr, kmap_pte-idx, mk_pte(page, prot), 1); kmap_atomic_prot()
|
H A D | pgtable_32.c | 397 static int __change_page_attr(struct page *page, pgprot_t prot) __change_page_attr() argument 410 __set_pte_at(&init_mm, address, kpte, mk_pte(page, prot), 0); __change_page_attr() 423 static int change_page_attr(struct page *page, int numpages, pgprot_t prot) change_page_attr() argument 430 err = __change_page_attr(page, prot); change_page_attr()
|
H A D | hash_utils_64.c | 182 unsigned long pstart, unsigned long prot, htab_bolt_mapping() 192 prot = htab_convert_pte_flags(prot); htab_bolt_mapping() 195 vstart, vend, pstart, prot, psize, ssize); htab_bolt_mapping() 202 unsigned long tprot = prot; htab_bolt_mapping() 675 unsigned long prot; htab_initialize() local 743 prot = pgprot_val(PAGE_KERNEL); htab_initialize() 762 DBG("creating mapping for region: %lx..%lx (prot: %lx)\n", for_each_memblock() 763 base, size, prot); for_each_memblock() 780 __pa(base), prot, for_each_memblock() 787 prot, for_each_memblock() 794 prot, mmu_linear_psize, mmu_kernel_ssize)); for_each_memblock() 813 __pa(tce_alloc_start), prot, 181 htab_bolt_mapping(unsigned long vstart, unsigned long vend, unsigned long pstart, unsigned long prot, int psize, int ssize) htab_bolt_mapping() argument
|
/linux-4.1.27/arch/unicore32/include/asm/ |
H A D | pgtable.h | 153 #define pfn_pte(pfn, prot) (__pte(((pfn) << PAGE_SHIFT) \ 154 | pgprot_val(prot))) 197 * Mark the prot value as uncacheable. pte_mkspecial() 199 #define pgprot_noncached(prot) \ pte_mkspecial() 200 __pgprot(pgprot_val(prot) & ~PTE_CACHEABLE) pte_mkspecial() 201 #define pgprot_writecombine(prot) \ pte_mkspecial() 202 __pgprot(pgprot_val(prot) & ~PTE_CACHEABLE) pte_mkspecial() 203 #define pgprot_dmacoherent(prot) \ pte_mkspecial() 204 __pgprot(pgprot_val(prot) & ~PTE_CACHEABLE) pte_mkspecial() 230 #define mk_pte(page, prot) pfn_pte(page_to_pfn(page), prot)
|
/linux-4.1.27/drivers/net/ethernet/mellanox/mlx4/ |
H A D | mcg.c | 159 u32 prot; new_steering_entry() local 207 prot = be32_to_cpu(mgm->members_count) >> 30; new_steering_entry() 222 mgm->members_count = cpu_to_be32(members_count | (prot << 30)); new_steering_entry() 429 u32 prot; add_promisc_qp() local 475 prot = be32_to_cpu(mgm->members_count) >> 30; add_promisc_qp() 506 (prot << 30)); add_promisc_qp() 696 u8 *gid, enum mlx4_protocol prot, find_entry() 705 u8 op_mod = (prot == MLX4_PROT_ETH) ? find_entry() 740 be32_to_cpu(mgm->members_count) >> 30 == prot) find_entry() 1098 int block_mcast_loopback, enum mlx4_protocol prot, mlx4_qp_attach_common() 1118 err = find_entry(dev, port, gid, prot, mlx4_qp_attach_common() 1164 mgm->members_count = cpu_to_be32(members_count | (u32) prot << 30); mlx4_qp_attach_common() 1184 if (prot == MLX4_PROT_ETH) { mlx4_qp_attach_common() 1207 enum mlx4_protocol prot, enum mlx4_steer_type steer) mlx4_qp_detach_common() 1226 err = find_entry(dev, port, gid, prot, mlx4_qp_detach_common() 1240 if (prot == MLX4_PROT_ETH && mlx4_qp_detach_common() 1261 mgm->members_count = cpu_to_be32(--members_count | (u32) prot << 30); mlx4_qp_detach_common() 1263 if (prot == MLX4_PROT_ETH) mlx4_qp_detach_common() 1266 if (members_count && (prot != MLX4_PROT_ETH || !removed_entry)) { mlx4_qp_detach_common() 1272 mgm->members_count = cpu_to_be32((u32) prot << 30); mlx4_qp_detach_common() 1329 enum mlx4_protocol prot) mlx4_QP_ATTACH() 1344 qpn |= (prot << 28); mlx4_QP_ATTACH() 1362 enum mlx4_protocol prot, u64 *reg_id) mlx4_trans_to_dmfs_attach() 1379 switch (prot) { mlx4_trans_to_dmfs_attach() 1401 enum mlx4_protocol prot, u64 *reg_id) mlx4_multicast_attach() 1405 if (prot == MLX4_PROT_ETH) mlx4_multicast_attach() 1409 if (prot == MLX4_PROT_ETH) mlx4_multicast_attach() 1414 block_mcast_loopback, prot); mlx4_multicast_attach() 1416 block_mcast_loopback, prot, mlx4_multicast_attach() 1422 prot, reg_id); mlx4_multicast_attach() 1430 enum mlx4_protocol prot, u64 reg_id) mlx4_multicast_detach() 1434 if (prot == MLX4_PROT_ETH) mlx4_multicast_detach() 1438 if (prot == MLX4_PROT_ETH) mlx4_multicast_detach() 1442 return mlx4_QP_ATTACH(dev, qp, gid, 0, 0, prot); mlx4_multicast_detach() 1444 return mlx4_qp_detach_common(dev, qp, gid, prot, mlx4_multicast_detach() 1516 int block_mcast_loopback, enum mlx4_protocol prot) mlx4_unicast_attach() 1518 if (prot == MLX4_PROT_ETH) mlx4_unicast_attach() 1523 block_mcast_loopback, prot); mlx4_unicast_attach() 1526 prot, MLX4_UC_STEER); mlx4_unicast_attach() 1531 u8 gid[16], enum mlx4_protocol prot) mlx4_unicast_detach() 1533 if (prot == MLX4_PROT_ETH) mlx4_unicast_detach() 1537 return mlx4_QP_ATTACH(dev, qp, gid, 0, 0, prot); mlx4_unicast_detach() 1539 return mlx4_qp_detach_common(dev, qp, gid, prot, MLX4_UC_STEER); mlx4_unicast_detach() 695 find_entry(struct mlx4_dev *dev, u8 port, u8 *gid, enum mlx4_protocol prot, struct mlx4_cmd_mailbox *mgm_mailbox, int *prev, int *index) find_entry() argument 1097 mlx4_qp_attach_common(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], int block_mcast_loopback, enum mlx4_protocol prot, enum mlx4_steer_type steer) mlx4_qp_attach_common() argument 1206 mlx4_qp_detach_common(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], enum mlx4_protocol prot, enum mlx4_steer_type steer) mlx4_qp_detach_common() argument 1327 mlx4_QP_ATTACH(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], u8 attach, u8 block_loopback, enum mlx4_protocol prot) mlx4_QP_ATTACH() argument 1359 mlx4_trans_to_dmfs_attach(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], u8 port, int block_mcast_loopback, enum mlx4_protocol prot, u64 *reg_id) mlx4_trans_to_dmfs_attach() argument 1399 mlx4_multicast_attach(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], u8 port, int block_mcast_loopback, enum mlx4_protocol prot, u64 *reg_id) mlx4_multicast_attach() argument 1429 mlx4_multicast_detach(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], enum mlx4_protocol prot, u64 reg_id) mlx4_multicast_detach() argument 1514 mlx4_unicast_attach(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], int block_mcast_loopback, enum mlx4_protocol prot) mlx4_unicast_attach() argument 1530 mlx4_unicast_detach(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], enum mlx4_protocol prot) mlx4_unicast_detach() argument
|
/linux-4.1.27/arch/tile/include/asm/ |
H A D | highmem.h | 65 void *kmap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot); 67 void *kmap_atomic_prot(struct page *page, pgprot_t prot);
|
H A D | mmu_context.h | 37 static inline void __install_page_table(pgd_t *pgdir, int asid, pgprot_t prot) __install_page_table() argument 40 int rc = hv_install_context(__pa(pgdir), prot, asid, __install_page_table()
|
H A D | pgtable.h | 275 extern pgprot_t set_remote_cache_cpu(pgprot_t prot, int cpu); 276 extern int get_remote_cache_cpu(pgprot_t prot); 278 static inline pte_t pfn_pte(unsigned long pfn, pgprot_t prot) pfn_pte() argument 280 return hv_pte_set_pa(prot, PFN_PHYS(pfn)); pfn_pte() 433 static inline pmd_t ptfn_pmd(unsigned long ptfn, pgprot_t prot) ptfn_pmd() argument 435 return pte_pmd(hv_pte_set_ptfn(prot, ptfn)); ptfn_pmd()
|
/linux-4.1.27/arch/microblaze/mm/ |
H A D | highmem.c | 34 void *kmap_atomic_prot(struct page *page, pgprot_t prot) kmap_atomic_prot() argument 52 set_pte_at(&init_mm, vaddr, kmap_pte-idx, mk_pte(page, prot)); kmap_atomic_prot()
|
/linux-4.1.27/arch/score/include/asm/ |
H A D | pgtable.h | 67 #define pfn_pte(pfn, prot) \ 68 __pte(((unsigned long long)(pfn) << PAGE_SHIFT) | pgprot_val(prot)) 99 #define mk_pte(page, prot) pfn_pte(page_to_pfn(page), prot) pte_mkspecial() 155 unsigned long prot = pgprot_val(_prot); pgprot_noncached() local 157 prot = (prot & ~_CACHE_MASK); pgprot_noncached() 159 return __pgprot(prot); pgprot_noncached()
|
/linux-4.1.27/arch/mips/mm/ |
H A D | pgtable-64.c | 79 pmd_t mk_pmd(struct page *page, pgprot_t prot) mk_pmd() argument 83 pmd_val(pmd) = (page_to_pfn(page) << _PFN_SHIFT) | pgprot_val(prot); mk_pmd()
|
/linux-4.1.27/arch/arm64/kernel/ |
H A D | sys.c | 30 unsigned long prot, unsigned long flags, sys_mmap() 36 return sys_mmap_pgoff(addr, len, prot, flags, fd, off >> PAGE_SHIFT); sys_mmap() 29 sys_mmap(unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long fd, off_t off) sys_mmap() argument
|
H A D | efi.c | 238 pgprot_t prot; efi_virtmap_init() local 259 prot = __pgprot(PROT_DEVICE_nGnRE); efi_virtmap_init() 262 prot = PAGE_KERNEL_EXEC; efi_virtmap_init() 264 prot = PAGE_KERNEL; efi_virtmap_init() 266 create_pgd_mapping(&efi_mm, paddr, md->virt_addr, size, prot); efi_virtmap_init()
|
/linux-4.1.27/arch/frv/kernel/ |
H A D | sys_frv.c | 31 unsigned long prot, unsigned long flags, sys_mmap2() 42 return sys_mmap_pgoff(addr, len, prot, flags, fd, sys_mmap2() 30 sys_mmap2(unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long fd, unsigned long pgoff) sys_mmap2() argument
|
/linux-4.1.27/arch/hexagon/mm/ |
H A D | ioremap.c | 31 pgprot_t prot = __pgprot(_PAGE_PRESENT|_PAGE_READ|_PAGE_WRITE ioremap_nocache() local 46 if (ioremap_page_range(addr, addr+size, phys_addr, prot)) { ioremap_nocache()
|
/linux-4.1.27/include/linux/can/ |
H A D | core.h | 34 * @prot: pointer to struct proto structure. 40 struct proto *prot; member in struct:can_proto
|
/linux-4.1.27/arch/um/os-Linux/ |
H A D | process.c | 136 int prot; os_map_memory() local 138 prot = (r ? PROT_READ : 0) | (w ? PROT_WRITE : 0) | os_map_memory() 141 loc = mmap64((void *) virt, len, prot, MAP_SHARED | MAP_FIXED, os_map_memory() 150 int prot = ((r ? PROT_READ : 0) | (w ? PROT_WRITE : 0) | os_protect_memory() local 153 if (mprotect(addr, len, prot) < 0) os_protect_memory()
|
/linux-4.1.27/drivers/xen/ |
H A D | xlate_mmu.c | 66 pgprot_t prot; member in struct:remap_data 82 pte_t pte = pte_mkspecial(pfn_pte(pfn, info->prot)); remap_pte_fn() 99 int *err_ptr, pgprot_t prot, xen_xlate_remap_gfn_array() 112 data.prot = prot; xen_xlate_remap_gfn_array() 96 xen_xlate_remap_gfn_array(struct vm_area_struct *vma, unsigned long addr, xen_pfn_t *mfn, int nr, int *err_ptr, pgprot_t prot, unsigned domid, struct page **pages) xen_xlate_remap_gfn_array() argument
|
/linux-4.1.27/arch/tile/mm/ |
H A D | init.c | 210 static pgprot_t __init construct_pgprot(pgprot_t prot, int home) construct_pgprot() argument 212 prot = pte_set_home(prot, home); construct_pgprot() 215 prot = hv_pte_set_mode(prot, HV_PTE_MODE_CACHE_HASH_L3); construct_pgprot() 217 prot = hv_pte_set_mode(prot, HV_PTE_MODE_CACHE_NO_L3); construct_pgprot() 219 return prot; construct_pgprot() 378 static inline pgprot_t ktext_set_nocache(pgprot_t prot) ktext_set_nocache() argument 381 prot = hv_pte_set_nc(prot); ktext_set_nocache() 383 prot = hv_pte_set_no_alloc_l2(prot); ktext_set_nocache() 384 return prot; ktext_set_nocache() 464 pgprot_t prot = init_pgprot(address); kernel_physical_mapping_init() local 465 *(pte_t *)pmd = pte_mkhuge(pfn_pte(pfn, prot)); kernel_physical_mapping_init() 468 pte[pte_ofs] = pfn_pte(pfn, prot); kernel_physical_mapping_init() 475 pgprot_t prot = init_pgprot(address); kernel_physical_mapping_init() local 476 pte[pte_ofs] = pfn_pte(pfn, prot); kernel_physical_mapping_init() 513 pgprot_t prot = construct_pgprot(PAGE_KERNEL_EXEC, kernel_physical_mapping_init() local 518 prot = hv_pte_set_mode(prot, kernel_physical_mapping_init() 521 prot = hv_pte_set_mode(prot, kernel_physical_mapping_init() 524 prot = hv_pte_set_mode(prot, kernel_physical_mapping_init() 528 prot = ktext_set_nocache(prot); kernel_physical_mapping_init() 542 prot = set_remote_cache_cpu(prot, cpu); kernel_physical_mapping_init() 547 pte[pte_ofs] = pfn_pte(pfn, prot); kernel_physical_mapping_init()
|
H A D | highmem.c | 198 void *kmap_atomic_prot(struct page *page, pgprot_t prot) kmap_atomic_prot() argument 208 BUG_ON(pte_exec(prot)); kmap_atomic_prot() 220 kmap_atomic_register(page, type, vaddr, pte, mk_pte(page, prot)); kmap_atomic_prot() 273 void *kmap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot) kmap_atomic_prot_pfn() argument 275 return kmap_atomic_prot(pfn_to_page(pfn), prot); kmap_atomic_prot_pfn()
|
H A D | pgtable.c | 371 pgprot_t set_remote_cache_cpu(pgprot_t prot, int cpu) set_remote_cache_cpu() argument 377 BUG_ON(hv_pte_get_mode(prot) != HV_PTE_MODE_CACHE_TILE_L3); set_remote_cache_cpu() 380 return hv_pte_set_lotar(prot, HV_XY_TO_LOTAR(x, y)); set_remote_cache_cpu() 383 int get_remote_cache_cpu(pgprot_t prot) get_remote_cache_cpu() argument 385 HV_LOTAR lotar = hv_pte_get_lotar(prot); get_remote_cache_cpu() 388 BUG_ON(hv_pte_get_mode(prot) != HV_PTE_MODE_CACHE_TILE_L3); get_remote_cache_cpu()
|
/linux-4.1.27/net/core/ |
H A D | sock.c | 1319 static struct sock *sk_prot_alloc(struct proto *prot, gfp_t priority, sk_prot_alloc() argument 1325 slab = prot->slab; sk_prot_alloc() 1331 if (prot->clear_sk) sk_prot_alloc() 1332 prot->clear_sk(sk, prot->obj_size); sk_prot_alloc() 1334 sk_prot_clear_nulls(sk, prot->obj_size); sk_prot_alloc() 1337 sk = kmalloc(prot->obj_size, priority); sk_prot_alloc() 1345 if (!try_module_get(prot->owner)) sk_prot_alloc() 1362 static void sk_prot_free(struct proto *prot, struct sock *sk) sk_prot_free() argument 1367 owner = prot->owner; sk_prot_free() 1368 slab = prot->slab; sk_prot_free() 1394 * @prot: struct proto associated with this new sock instance 1397 struct proto *prot) sk_alloc() 1401 sk = sk_prot_alloc(prot, priority | __GFP_ZERO, family); sk_alloc() 1408 sk->sk_prot = sk->sk_prot_creator = prot; sk_alloc() 2003 struct proto *prot = sk->sk_prot; __sk_mem_schedule() local 2031 if (atomic_read(&sk->sk_rmem_alloc) < prot->sysctl_rmem[0]) __sk_mem_schedule() 2036 if (sk->sk_wmem_queued < prot->sysctl_wmem[0]) __sk_mem_schedule() 2039 prot->sysctl_wmem[0]) __sk_mem_schedule() 2068 trace_sock_exceed_buf_limit(sk, prot, allocated); __sk_mem_schedule() 2644 void sock_prot_inuse_add(struct net *net, struct proto *prot, int val) sock_prot_inuse_add() argument 2646 __this_cpu_add(net->core.inuse->val[prot->inuse_idx], val); sock_prot_inuse_add() 2650 int sock_prot_inuse_get(struct net *net, struct proto *prot) sock_prot_inuse_get() argument 2652 int cpu, idx = prot->inuse_idx; sock_prot_inuse_get() 2690 void sock_prot_inuse_add(struct net *net, struct proto *prot, int val) sock_prot_inuse_add() argument 2692 __this_cpu_add(prot_inuse.val[prot->inuse_idx], val); sock_prot_inuse_add() 2696 int sock_prot_inuse_get(struct net *net, struct proto *prot) sock_prot_inuse_get() argument 2698 int cpu, idx = prot->inuse_idx; sock_prot_inuse_get() 2709 static void assign_proto_idx(struct proto *prot) assign_proto_idx() argument 2711 prot->inuse_idx = find_first_zero_bit(proto_inuse_idx, PROTO_INUSE_NR); assign_proto_idx() 2713 if (unlikely(prot->inuse_idx == PROTO_INUSE_NR - 1)) { assign_proto_idx() 2718 set_bit(prot->inuse_idx, proto_inuse_idx); assign_proto_idx() 2721 static void release_proto_idx(struct proto *prot) release_proto_idx() argument 2723 if (prot->inuse_idx != PROTO_INUSE_NR - 1) release_proto_idx() 2724 clear_bit(prot->inuse_idx, proto_inuse_idx); release_proto_idx() 2727 static inline void assign_proto_idx(struct proto *prot) assign_proto_idx() argument 2731 static inline void release_proto_idx(struct proto *prot) release_proto_idx() argument 2748 static int req_prot_init(const struct proto *prot) req_prot_init() argument 2750 struct request_sock_ops *rsk_prot = prot->rsk_prot; req_prot_init() 2756 prot->name); req_prot_init() 2766 prot->name); req_prot_init() 2772 int proto_register(struct proto *prot, int alloc_slab) proto_register() argument 2775 prot->slab = kmem_cache_create(prot->name, prot->obj_size, 0, proto_register() 2776 SLAB_HWCACHE_ALIGN | prot->slab_flags, proto_register() 2779 if (prot->slab == NULL) { proto_register() 2781 prot->name); proto_register() 2785 if (req_prot_init(prot)) proto_register() 2788 if (prot->twsk_prot != NULL) { proto_register() 2789 prot->twsk_prot->twsk_slab_name = kasprintf(GFP_KERNEL, "tw_sock_%s", prot->name); proto_register() 2791 if (prot->twsk_prot->twsk_slab_name == NULL) proto_register() 2794 prot->twsk_prot->twsk_slab = proto_register() 2795 kmem_cache_create(prot->twsk_prot->twsk_slab_name, proto_register() 2796 prot->twsk_prot->twsk_obj_size, proto_register() 2798 prot->slab_flags, proto_register() 2800 if (prot->twsk_prot->twsk_slab == NULL) proto_register() 2806 list_add(&prot->node, &proto_list); proto_register() 2807 assign_proto_idx(prot); proto_register() 2812 kfree(prot->twsk_prot->twsk_slab_name); proto_register() 2814 req_prot_cleanup(prot->rsk_prot); proto_register() 2816 kmem_cache_destroy(prot->slab); proto_register() 2817 prot->slab = NULL; proto_register() 2823 void proto_unregister(struct proto *prot) proto_unregister() argument 2826 release_proto_idx(prot); proto_unregister() 2827 list_del(&prot->node); proto_unregister() 2830 if (prot->slab != NULL) { proto_unregister() 2831 kmem_cache_destroy(prot->slab); proto_unregister() 2832 prot->slab = NULL; proto_unregister() 2835 req_prot_cleanup(prot->rsk_prot); proto_unregister() 2837 if (prot->twsk_prot != NULL && prot->twsk_prot->twsk_slab != NULL) { proto_unregister() 2838 kmem_cache_destroy(prot->twsk_prot->twsk_slab); proto_unregister() 2839 kfree(prot->twsk_prot->twsk_slab_name); proto_unregister() 2840 prot->twsk_prot->twsk_slab = NULL; proto_unregister() 1396 sk_alloc(struct net *net, int family, gfp_t priority, struct proto *prot) sk_alloc() argument
|
/linux-4.1.27/drivers/iommu/ |
H A D | io-pgtable-arm.c | 209 arm_lpae_iopte prot, int lvl, arm_lpae_init_pte() 212 arm_lpae_iopte pte = prot; arm_lpae_init_pte() 248 phys_addr_t paddr, size_t size, arm_lpae_iopte prot, __arm_lpae_map() 260 return arm_lpae_init_pte(data, iova, paddr, prot, lvl, ptep); __arm_lpae_map() 286 return __arm_lpae_map(data, iova, paddr, size, prot, lvl + 1, cptep); __arm_lpae_map() 290 int prot) arm_lpae_prot_to_pte() 298 if (!(prot & IOMMU_WRITE) && (prot & IOMMU_READ)) arm_lpae_prot_to_pte() 301 if (prot & IOMMU_CACHE) arm_lpae_prot_to_pte() 306 if (prot & IOMMU_READ) arm_lpae_prot_to_pte() 308 if (prot & IOMMU_WRITE) arm_lpae_prot_to_pte() 310 if (prot & IOMMU_CACHE) arm_lpae_prot_to_pte() 316 if (prot & IOMMU_NOEXEC) arm_lpae_prot_to_pte() 328 arm_lpae_iopte prot; arm_lpae_map() local 334 prot = arm_lpae_prot_to_pte(data, iommu_prot); arm_lpae_map() 335 return __arm_lpae_map(data, iova, paddr, size, prot, lvl, ptep); arm_lpae_map() 379 arm_lpae_iopte prot, int lvl, arm_lpae_split_blk_unmap() 401 if (__arm_lpae_map(data, blk_start, blk_paddr, size, prot, lvl, arm_lpae_split_blk_unmap() 207 arm_lpae_init_pte(struct arm_lpae_io_pgtable *data, unsigned long iova, phys_addr_t paddr, arm_lpae_iopte prot, int lvl, arm_lpae_iopte *ptep) arm_lpae_init_pte() argument 247 __arm_lpae_map(struct arm_lpae_io_pgtable *data, unsigned long iova, phys_addr_t paddr, size_t size, arm_lpae_iopte prot, int lvl, arm_lpae_iopte *ptep) __arm_lpae_map() argument 289 arm_lpae_prot_to_pte(struct arm_lpae_io_pgtable *data, int prot) arm_lpae_prot_to_pte() argument 377 arm_lpae_split_blk_unmap(struct arm_lpae_io_pgtable *data, unsigned long iova, size_t size, arm_lpae_iopte prot, int lvl, arm_lpae_iopte *ptep, size_t blk_size) arm_lpae_split_blk_unmap() argument
|
H A D | omap-iommu.c | 662 static int iopgd_alloc_section(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) iopgd_alloc_section() argument 672 *iopgd = (pa & IOSECTION_MASK) | prot | IOPGD_SECTION; iopgd_alloc_section() 677 static int iopgd_alloc_super(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) iopgd_alloc_super() argument 689 *(iopgd + i) = (pa & IOSUPER_MASK) | prot | IOPGD_SUPER; iopgd_alloc_super() 694 static int iopte_alloc_page(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) iopte_alloc_page() argument 702 *iopte = (pa & IOPAGE_MASK) | prot | IOPTE_SMALL; iopte_alloc_page() 711 static int iopte_alloc_large(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) iopte_alloc_large() argument 727 *(iopte + i) = (pa & IOLARGE_MASK) | prot | IOPTE_LARGE; iopte_alloc_large() 736 u32 prot; iopgtable_store_entry_core() local 761 prot = get_iopte_attr(e); iopgtable_store_entry_core() 764 err = fn(obj, e->da, e->pa, prot); iopgtable_store_entry_core() 1124 phys_addr_t pa, size_t bytes, int prot) omap_iommu_map() 1123 omap_iommu_map(struct iommu_domain *domain, unsigned long da, phys_addr_t pa, size_t bytes, int prot) omap_iommu_map() argument
|
H A D | rockchip-iommu.c | 227 /* TODO: set cache flags per prot IOMMU_CACHE */ rk_mk_pte() 228 static u32 rk_mk_pte(phys_addr_t page, int prot) rk_mk_pte() argument 231 flags |= (prot & IOMMU_READ) ? RK_PTE_PAGE_READABLE : 0; rk_mk_pte() 232 flags |= (prot & IOMMU_WRITE) ? RK_PTE_PAGE_WRITABLE : 0; rk_mk_pte() 612 int prot) rk_iommu_map_iova() 626 pte_addr[pte_count] = rk_mk_pte(paddr, prot); rk_iommu_map_iova() 640 pr_err("iova: %pad already mapped to %pa cannot remap to phys: %pa prot: %#x\n", rk_iommu_map_iova() 641 &iova, &page_phys, &paddr, prot); rk_iommu_map_iova() 647 phys_addr_t paddr, size_t size, int prot) rk_iommu_map() 671 ret = rk_iommu_map_iova(rk_domain, pte_addr, iova, paddr, size, prot); rk_iommu_map() 610 rk_iommu_map_iova(struct rk_iommu_domain *rk_domain, u32 *pte_addr, dma_addr_t iova, phys_addr_t paddr, size_t size, int prot) rk_iommu_map_iova() argument 646 rk_iommu_map(struct iommu_domain *domain, unsigned long _iova, phys_addr_t paddr, size_t size, int prot) rk_iommu_map() argument
|
H A D | fsl_pamu_domain.h | 28 int prot; member in struct:dma_window
|
H A D | fsl_pamu_domain.c | 114 sub_win_ptr[i].prot); map_subwins() 140 0, wnd->prot); map_win() 173 wnd->prot); update_liodn() 187 0, wnd->prot); update_liodn() 535 phys_addr_t paddr, u64 size, int prot) fsl_pamu_window_enable() 544 if (prot & IOMMU_READ) fsl_pamu_window_enable() 546 if (prot & IOMMU_WRITE) fsl_pamu_window_enable() 588 wnd->prot = pamu_prot; fsl_pamu_window_enable() 534 fsl_pamu_window_enable(struct iommu_domain *domain, u32 wnd_nr, phys_addr_t paddr, u64 size, int prot) fsl_pamu_window_enable() argument
|
/linux-4.1.27/arch/mips/include/asm/ |
H A D | pgtable-32.h | 110 pfn_pte(unsigned long pfn, pgprot_t prot) pfn_pte() argument 115 (pgprot_val(prot) & ~_PFNX_MASK); pfn_pte() 117 (pgprot_val(prot) & ~_PFN_MASK); pfn_pte() 127 #define pfn_pte(pfn, prot) __pte(((pfn) << (PAGE_SHIFT + 2)) | pgprot_val(prot)) 130 #define pfn_pte(pfn, prot) __pte(((unsigned long long)(pfn) << _PFN_SHIFT) | pgprot_val(prot))
|
H A D | pgtable.h | 408 unsigned long prot = pgprot_val(_prot); pgprot_noncached() local 410 prot = (prot & ~_CACHE_MASK) | _CACHE_UNCACHED; pgprot_noncached() 412 return __pgprot(prot); pgprot_noncached() 417 unsigned long prot = pgprot_val(_prot); pgprot_writecombine() local 420 prot = (prot & ~_CACHE_MASK) | cpu_data[0].writecombine; pgprot_writecombine() 422 return __pgprot(prot); pgprot_writecombine() 469 extern int remap_pfn_range(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot); 475 pgprot_t prot) io_remap_pfn_range() 478 return remap_pfn_range(vma, vaddr, phys_addr_high >> PAGE_SHIFT, size, prot); io_remap_pfn_range() 589 extern pmd_t mk_pmd(struct page *page, pgprot_t prot); 471 io_remap_pfn_range(struct vm_area_struct *vma, unsigned long vaddr, unsigned long pfn, unsigned long size, pgprot_t prot) io_remap_pfn_range() argument
|
H A D | pgtable-64.h | 229 #define pfn_pte(pfn, prot) __pte(((pfn) << (PAGE_SHIFT + 2)) | pgprot_val(prot)) 232 #define pfn_pte(pfn, prot) __pte(((pfn) << _PFN_SHIFT) | pgprot_val(prot)) 233 #define pfn_pmd(pfn, prot) __pmd(((pfn) << _PFN_SHIFT) | pgprot_val(prot))
|
/linux-4.1.27/arch/nios2/include/asm/ |
H A D | pgtable.h | 121 unsigned long prot = pgprot_val(_prot); pgprot_noncached() local 123 prot &= ~_PAGE_CACHED; pgprot_noncached() 125 return __pgprot(prot); pgprot_noncached() 198 #define pfn_pte(pfn, prot) (__pte(pfn | pgprot_val(prot))) 241 #define mk_pte(page, prot) (pfn_pte(page_to_pfn(page), prot))
|
/linux-4.1.27/mm/ |
H A D | mprotect.c | 336 unsigned long, prot) SYSCALL_DEFINE3() 341 const int grows = prot & (PROT_GROWSDOWN|PROT_GROWSUP); SYSCALL_DEFINE3() 342 prot &= ~(PROT_GROWSDOWN|PROT_GROWSUP); SYSCALL_DEFINE3() 354 if (!arch_validate_prot(prot)) SYSCALL_DEFINE3() 357 reqprot = prot; SYSCALL_DEFINE3() 361 if ((prot & PROT_READ) && (current->personality & READ_IMPLIES_EXEC)) SYSCALL_DEFINE3() 362 prot |= PROT_EXEC; SYSCALL_DEFINE3() 364 vm_flags = calc_vm_prot_bits(prot); SYSCALL_DEFINE3() 407 error = security_file_mprotect(vma, reqprot, prot); SYSCALL_DEFINE3()
|
H A D | early_ioremap.c | 49 phys_addr_t phys, pgprot_t prot) __late_set_fixmap() 97 __early_ioremap(resource_size_t phys_addr, unsigned long size, pgprot_t prot) __early_ioremap() argument 145 __late_set_fixmap(idx, phys_addr, prot); __early_ioremap() 147 __early_set_fixmap(idx, phys_addr, prot); __early_ioremap() 48 __late_set_fixmap(enum fixed_addresses idx, phys_addr_t phys, pgprot_t prot) __late_set_fixmap() argument
|
H A D | nommu.c | 280 void *__vmalloc(unsigned long size, gfp_t gfp_mask, pgprot_t prot) __vmalloc() argument 467 void *vmap(struct page **pages, unsigned int count, unsigned long flags, pgprot_t prot) vmap() argument 480 void *vm_map_ram(struct page **pages, unsigned int count, int node, pgprot_t prot) vm_map_ram() argument 951 unsigned long prot, validate_mmap_request() 1028 if ((prot & PROT_WRITE) && validate_mmap_request() 1052 if (prot & PROT_WRITE) validate_mmap_request() 1057 if (((prot & PROT_READ) && !(capabilities & NOMMU_MAP_READ)) || validate_mmap_request() 1058 ((prot & PROT_WRITE) && !(capabilities & NOMMU_MAP_WRITE)) || validate_mmap_request() 1059 ((prot & PROT_EXEC) && !(capabilities & NOMMU_MAP_EXEC)) validate_mmap_request() 1073 if (prot & PROT_EXEC) validate_mmap_request() 1075 } else if ((prot & PROT_READ) && !(prot & PROT_EXEC)) { validate_mmap_request() 1079 prot |= PROT_EXEC; validate_mmap_request() 1081 } else if ((prot & PROT_READ) && validate_mmap_request() 1082 (prot & PROT_EXEC) && validate_mmap_request() 1095 if ((prot & PROT_READ) && validate_mmap_request() 1097 prot |= PROT_EXEC; validate_mmap_request() 1115 unsigned long prot, determine_vm_flags() 1121 vm_flags = calc_vm_prot_bits(prot) | calc_vm_flag_bits(flags); determine_vm_flags() 1127 if (file && !(prot & PROT_WRITE)) determine_vm_flags() 1277 unsigned long prot, do_mmap_pgoff() 1288 kenter(",%lx,%lx,%lx,%lx,%lx", addr, len, prot, flags, pgoff); do_mmap_pgoff() 1294 ret = validate_mmap_request(file, addr, len, prot, flags, pgoff, do_mmap_pgoff() 1307 vm_flags = determine_vm_flags(file, prot, flags, capabilities); do_mmap_pgoff() 1508 unsigned long, prot, unsigned long, flags, SYSCALL_DEFINE6() 1523 retval = vm_mmap_pgoff(file, addr, len, prot, flags, pgoff); SYSCALL_DEFINE6() 1535 unsigned long prot; member in struct:mmap_arg_struct 1550 return sys_mmap_pgoff(a.addr, a.len, a.prot, a.flags, a.fd, SYSCALL_DEFINE1() 1855 unsigned long pfn, unsigned long size, pgprot_t prot) remap_pfn_range() 948 validate_mmap_request(struct file *file, unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long pgoff, unsigned long *_capabilities) validate_mmap_request() argument 1114 determine_vm_flags(struct file *file, unsigned long prot, unsigned long flags, unsigned long capabilities) determine_vm_flags() argument 1274 do_mmap_pgoff(struct file *file, unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long pgoff, unsigned long *populate) do_mmap_pgoff() argument 1854 remap_pfn_range(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn, unsigned long size, pgprot_t prot) remap_pfn_range() argument
|
H A D | vmalloc.c | 118 unsigned long end, pgprot_t prot, struct page **pages, int *nr) vmap_pte_range() 137 set_pte_at(&init_mm, addr, pte, mk_pte(page, prot)); vmap_pte_range() 144 unsigned long end, pgprot_t prot, struct page **pages, int *nr) vmap_pmd_range() 154 if (vmap_pte_range(pmd, addr, next, prot, pages, nr)) vmap_pmd_range() 161 unsigned long end, pgprot_t prot, struct page **pages, int *nr) vmap_pud_range() 171 if (vmap_pmd_range(pud, addr, next, prot, pages, nr)) vmap_pud_range() 178 * Set up page tables in kva (addr, end). The ptes shall have prot "prot", and 184 pgprot_t prot, struct page **pages) vmap_page_range_noflush() 196 err = vmap_pud_range(pgd, addr, next, prot, pages, &nr); vmap_page_range_noflush() 205 pgprot_t prot, struct page **pages) vmap_page_range() 209 ret = vmap_page_range_noflush(start, end, prot, pages); vmap_page_range() 1105 * @prot: memory protection to use. PAGE_KERNEL for regular RAM 1115 void *vm_map_ram(struct page **pages, unsigned int count, int node, pgprot_t prot) vm_map_ram() argument 1136 if (vmap_page_range(addr, addr + size, prot, pages) < 0) { vm_map_ram() 1233 * @prot: page protection flags to use 1249 pgprot_t prot, struct page **pages) map_kernel_range_noflush() 1251 return vmap_page_range_noflush(addr, addr + size, prot, pages); map_kernel_range_noflush() 1292 int map_vm_area(struct vm_struct *area, pgprot_t prot, struct page **pages) map_vm_area() argument 1298 err = vmap_page_range(addr, end, prot, pages); map_vm_area() 1546 * @prot: page protection for the mapping 1552 unsigned long flags, pgprot_t prot) vmap() 1566 if (map_vm_area(area, prot, pages)) { vmap() 1576 gfp_t gfp_mask, pgprot_t prot, 1579 pgprot_t prot, int node) __vmalloc_area_node() 1624 if (map_vm_area(area, prot, pages)) __vmalloc_area_node() 1643 * @prot: protection mask for the allocated pages 1650 * kernel virtual space, using a pagetable protection of @prot. 1654 pgprot_t prot, unsigned long vm_flags, int node, __vmalloc_node_range() 1670 addr = __vmalloc_area_node(area, gfp_mask, prot, node); __vmalloc_node_range() 1702 * @prot: protection mask for the allocated pages 1708 * kernel virtual space, using a pagetable protection of @prot. 1711 gfp_t gfp_mask, pgprot_t prot, __vmalloc_node() 1715 gfp_mask, prot, 0, node, caller); __vmalloc_node() 1718 void *__vmalloc(unsigned long size, gfp_t gfp_mask, pgprot_t prot) __vmalloc() argument 1720 return __vmalloc_node(size, 1, gfp_mask, prot, NUMA_NO_NODE, __vmalloc() 117 vmap_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end, pgprot_t prot, struct page **pages, int *nr) vmap_pte_range() argument 143 vmap_pmd_range(pud_t *pud, unsigned long addr, unsigned long end, pgprot_t prot, struct page **pages, int *nr) vmap_pmd_range() argument 160 vmap_pud_range(pgd_t *pgd, unsigned long addr, unsigned long end, pgprot_t prot, struct page **pages, int *nr) vmap_pud_range() argument 183 vmap_page_range_noflush(unsigned long start, unsigned long end, pgprot_t prot, struct page **pages) vmap_page_range_noflush() argument 204 vmap_page_range(unsigned long start, unsigned long end, pgprot_t prot, struct page **pages) vmap_page_range() argument 1248 map_kernel_range_noflush(unsigned long addr, unsigned long size, pgprot_t prot, struct page **pages) map_kernel_range_noflush() argument 1551 vmap(struct page **pages, unsigned int count, unsigned long flags, pgprot_t prot) vmap() argument 1578 __vmalloc_area_node(struct vm_struct *area, gfp_t gfp_mask, pgprot_t prot, int node) __vmalloc_area_node() argument 1652 __vmalloc_node_range(unsigned long size, unsigned long align, unsigned long start, unsigned long end, gfp_t gfp_mask, pgprot_t prot, unsigned long vm_flags, int node, const void *caller) __vmalloc_node_range() argument 1710 __vmalloc_node(unsigned long size, unsigned long align, gfp_t gfp_mask, pgprot_t prot, int node, const void *caller) __vmalloc_node() argument
|
H A D | util.c | 287 unsigned long len, unsigned long prot, vm_mmap_pgoff() 294 ret = security_mmap_file(file, prot, flag); vm_mmap_pgoff() 297 ret = do_mmap_pgoff(file, addr, len, prot, flag, pgoff, vm_mmap_pgoff() 307 unsigned long len, unsigned long prot, vm_mmap() 315 return vm_mmap_pgoff(file, addr, len, prot, flag, offset >> PAGE_SHIFT); vm_mmap() 286 vm_mmap_pgoff(struct file *file, unsigned long addr, unsigned long len, unsigned long prot, unsigned long flag, unsigned long pgoff) vm_mmap_pgoff() argument 306 vm_mmap(struct file *file, unsigned long addr, unsigned long len, unsigned long prot, unsigned long flag, unsigned long offset) vm_mmap() argument
|
/linux-4.1.27/include/uapi/linux/ |
H A D | agpgart.h | 77 * The "prot" down below needs still a "sleep" flag somehow ... 82 int prot; /* prot flags for mmap */ member in struct:_agp_segment
|
/linux-4.1.27/arch/ia64/mm/ |
H A D | ioremap.c | 40 pgprot_t prot; ioremap() local 74 prot = PAGE_KERNEL; ioremap() 92 (unsigned long) addr + size, phys_addr, prot)) { ioremap()
|
/linux-4.1.27/arch/metag/include/asm/ |
H A D | elf.h | 111 struct elf32_phdr *eppnt, int prot, int type, 116 struct elf32_phdr *eppnt, int prot, metag_elf_map() 119 return __metag_elf_map(filep, addr, eppnt, prot, type, total_size); metag_elf_map() 114 metag_elf_map(struct file *filep, unsigned long addr, struct elf32_phdr *eppnt, int prot, int type, unsigned long total_size) metag_elf_map() argument
|
H A D | pgtable.h | 106 #define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) 142 #define pgprot_writecombine(prot) \ pte_mkhuge() 143 __pgprot(pgprot_val(prot) & ~(_PAGE_CACHE_CTRL1 | _PAGE_CACHE_CTRL0)) pte_mkhuge() 145 #define pgprot_noncached(prot) \ pte_mkhuge() 146 __pgprot(pgprot_val(prot) & ~_PAGE_CACHEABLE) pte_mkhuge()
|
/linux-4.1.27/drivers/gpu/drm/ttm/ |
H A D | ttm_bo_util.c | 252 pgprot_t prot) ttm_copy_io_ttm_page() 263 dst = kmap_atomic_prot(d, prot); ttm_copy_io_ttm_page() 265 if (pgprot_val(prot) != pgprot_val(PAGE_KERNEL)) ttm_copy_io_ttm_page() 266 dst = vmap(&d, 1, 0, prot); ttm_copy_io_ttm_page() 278 if (pgprot_val(prot) != pgprot_val(PAGE_KERNEL)) ttm_copy_io_ttm_page() 289 pgprot_t prot) ttm_copy_ttm_io_page() 299 src = kmap_atomic_prot(s, prot); ttm_copy_ttm_io_page() 301 if (pgprot_val(prot) != pgprot_val(PAGE_KERNEL)) ttm_copy_ttm_io_page() 302 src = vmap(&s, 1, 0, prot); ttm_copy_ttm_io_page() 314 if (pgprot_val(prot) != pgprot_val(PAGE_KERNEL)) ttm_copy_ttm_io_page() 384 pgprot_t prot = ttm_io_prot(old_mem->placement, ttm_bo_move_memcpy() local 387 prot); ttm_bo_move_memcpy() 389 pgprot_t prot = ttm_io_prot(new_mem->placement, ttm_bo_move_memcpy() local 392 prot); ttm_bo_move_memcpy() 533 struct ttm_mem_reg *mem = &bo->mem; pgprot_t prot; ttm_bo_kmap_ttm() local 559 prot = ttm_io_prot(mem->placement, PAGE_KERNEL); ttm_bo_kmap_ttm() 562 0, prot); ttm_bo_kmap_ttm() 250 ttm_copy_io_ttm_page(struct ttm_tt *ttm, void *src, unsigned long page, pgprot_t prot) ttm_copy_io_ttm_page() argument 287 ttm_copy_ttm_io_page(struct ttm_tt *ttm, void *dst, unsigned long page, pgprot_t prot) ttm_copy_ttm_io_page() argument
|
/linux-4.1.27/arch/tile/kernel/ |
H A D | sys.c | 78 unsigned long, prot, unsigned long, flags, SYSCALL_DEFINE6() 84 return sys_mmap_pgoff(addr, len, prot, flags, fd, SYSCALL_DEFINE6() 90 unsigned long, prot, unsigned long, flags, SYSCALL_DEFINE6() 95 return sys_mmap_pgoff(addr, len, prot, flags, fd, SYSCALL_DEFINE6()
|
/linux-4.1.27/arch/um/os-Linux/skas/ |
H A D | mem.c | 170 int map(struct mm_id * mm_idp, unsigned long virt, unsigned long len, int prot, map() argument 174 unsigned long args[] = { virt, len, prot, map() 198 unsigned int prot, int done, void **data) protect() 201 unsigned long args[] = { addr, len, prot, 0, 0, 0 }; protect() 197 protect(struct mm_id * mm_idp, unsigned long addr, unsigned long len, unsigned int prot, int done, void **data) protect() argument
|
/linux-4.1.27/arch/sh/include/cpu-sh4/cpu/ |
H A D | sq.h | 32 const char *name, pgprot_t prot);
|
/linux-4.1.27/arch/mips/kernel/ |
H A D | syscall.c | 60 unsigned long, prot, unsigned long, flags, unsigned long, SYSCALL_DEFINE6() 69 result = sys_mmap_pgoff(addr, len, prot, flags, fd, offset >> PAGE_SHIFT); SYSCALL_DEFINE6() 76 unsigned long, prot, unsigned long, flags, unsigned long, fd, SYSCALL_DEFINE6() 82 return sys_mmap_pgoff(addr, len, prot, flags, fd, pgoff >> (PAGE_SHIFT-12)); SYSCALL_DEFINE6()
|
H A D | linux32.c | 65 unsigned long, prot, unsigned long, flags, unsigned long, fd, 73 error = sys_mmap_pgoff(addr, len, prot, flags, fd,
|
/linux-4.1.27/arch/openrisc/mm/ |
H A D | init.c | 78 pgprot_t prot; map_ram() local 117 prot = PAGE_KERNEL; for_each_memblock() 119 prot = PAGE_KERNEL_RO; for_each_memblock() 121 set_pte(pte, mk_pte_phys(p, prot)); for_each_memblock()
|
H A D | ioremap.c | 42 __ioremap(phys_addr_t addr, unsigned long size, pgprot_t prot) __ioremap() argument 73 if (ioremap_page_range(v, v + size, p, prot)) { __ioremap()
|
/linux-4.1.27/arch/cris/include/asm/ |
H A D | cacheflush.h | 30 int change_page_attr(struct page *page, int numpages, pgprot_t prot);
|
H A D | pgtable.h | 199 #define pgprot_noncached(prot) __pgprot((pgprot_val(prot) | _PAGE_NO_CACHE)) 253 #define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
|
/linux-4.1.27/arch/ia64/kernel/ |
H A D | sys_ia64.c | 137 sys_mmap2 (unsigned long addr, unsigned long len, int prot, int flags, int fd, long pgoff) sys_mmap2() argument 139 addr = sys_mmap_pgoff(addr, len, prot, flags, fd, pgoff); sys_mmap2() 146 sys_mmap (unsigned long addr, unsigned long len, int prot, int flags, int fd, long off) sys_mmap() argument 151 addr = sys_mmap_pgoff(addr, len, prot, flags, fd, off >> PAGE_SHIFT); sys_mmap()
|
/linux-4.1.27/arch/powerpc/boot/ |
H A D | mv64x60.c | 283 u32 i, base, size, enables, prot = 0, snoop_bits = 0; mv64x60_config_ctlr_windows() local 304 prot |= (0x3 << (i << 1)); /* RW access */ mv64x60_config_ctlr_windows() 314 out_le32((u32 *)(bridge_base + MV64x60_ENET2MEM_ACC_PROT_0), prot); mv64x60_config_ctlr_windows() 315 out_le32((u32 *)(bridge_base + MV64x60_ENET2MEM_ACC_PROT_1), prot); mv64x60_config_ctlr_windows() 316 out_le32((u32 *)(bridge_base + MV64x60_ENET2MEM_ACC_PROT_2), prot); mv64x60_config_ctlr_windows() 317 out_le32((u32 *)(bridge_base + MV64x60_MPSC2MEM_ACC_PROT_0), prot); mv64x60_config_ctlr_windows() 318 out_le32((u32 *)(bridge_base + MV64x60_MPSC2MEM_ACC_PROT_1), prot); mv64x60_config_ctlr_windows() 319 out_le32((u32 *)(bridge_base + MV64x60_IDMA2MEM_ACC_PROT_0), prot); mv64x60_config_ctlr_windows() 320 out_le32((u32 *)(bridge_base + MV64x60_IDMA2MEM_ACC_PROT_1), prot); mv64x60_config_ctlr_windows() 321 out_le32((u32 *)(bridge_base + MV64x60_IDMA2MEM_ACC_PROT_2), prot); mv64x60_config_ctlr_windows() 322 out_le32((u32 *)(bridge_base + MV64x60_IDMA2MEM_ACC_PROT_3), prot); mv64x60_config_ctlr_windows()
|
/linux-4.1.27/drivers/isdn/hisax/ |
H A D | l3dss1.c | 55 retval = p->prot.dss1.last_invoke_id + 1; /* try new id */ new_invoke_id() 56 while ((i) && (p->prot.dss1.invoke_used[retval >> 3] == 0xFF)) { new_invoke_id() 57 p->prot.dss1.last_invoke_id = (retval & 0xF8) + 8; new_invoke_id() 61 while (p->prot.dss1.invoke_used[retval >> 3] & (1 << (retval & 7))) new_invoke_id() 65 p->prot.dss1.last_invoke_id = retval; new_invoke_id() 66 p->prot.dss1.invoke_used[retval >> 3] |= (1 << (retval & 7)); new_invoke_id() 78 p->prot.dss1.invoke_used[id >> 3] &= ~(1 << (id & 7)); free_invoke_id() 92 proc->prot.dss1.invoke_id = 0; dss1_new_l3_process() 93 proc->prot.dss1.remote_operation = 0; dss1_new_l3_process() 94 proc->prot.dss1.uus1_data[0] = '\0'; dss1_new_l3_process() 105 free_invoke_id(p->st, p->prot.dss1.invoke_id); dss1_release_l3_process() 119 { if ((pc->callref == -1) && (pc->prot.dss1.invoke_id == id)) l3dss1_search_dummy_proc() 143 ic.parm.dss1_io.hl_id = pc->prot.dss1.invoke_id; l3dss1_dummy_return_result() 144 ic.parm.dss1_io.ll_id = pc->prot.dss1.ll_id; l3dss1_dummy_return_result() 145 ic.parm.dss1_io.proc = pc->prot.dss1.proc; l3dss1_dummy_return_result() 149 free_invoke_id(pc->st, pc->prot.dss1.invoke_id); l3dss1_dummy_return_result() 150 pc->prot.dss1.invoke_id = 0; /* reset id */ l3dss1_dummy_return_result() 176 ic.parm.dss1_io.hl_id = pc->prot.dss1.invoke_id; l3dss1_dummy_error_return() 177 ic.parm.dss1_io.ll_id = pc->prot.dss1.ll_id; l3dss1_dummy_error_return() 178 ic.parm.dss1_io.proc = pc->prot.dss1.proc; l3dss1_dummy_error_return() 182 free_invoke_id(pc->st, pc->prot.dss1.invoke_id); l3dss1_dummy_error_return() 183 pc->prot.dss1.invoke_id = 0; /* reset id */ l3dss1_dummy_error_return() 436 if ((pc->prot.dss1.invoke_id) && (pc->prot.dss1.invoke_id == id)) l3dss1_parse_facility() 438 free_invoke_id(st, pc->prot.dss1.invoke_id); l3dss1_parse_facility() 439 pc->prot.dss1.remote_result = 0; /* success */ l3dss1_parse_facility() 440 pc->prot.dss1.invoke_id = 0; l3dss1_parse_facility() 441 pc->redir_result = pc->prot.dss1.remote_result; l3dss1_parse_facility() 481 if ((pc->prot.dss1.invoke_id) && (pc->prot.dss1.invoke_id == id)) l3dss1_parse_facility() 483 free_invoke_id(st, pc->prot.dss1.invoke_id); l3dss1_parse_facility() 484 pc->prot.dss1.remote_result = err_ret; /* result */ l3dss1_parse_facility() 485 pc->prot.dss1.invoke_id = 0; l3dss1_parse_facility() 486 pc->redir_result = pc->prot.dss1.remote_result; l3dss1_parse_facility() 935 if (pc->prot.dss1.uus1_data[0]) l3dss1_msg_with_uus() 937 *p++ = strlen(pc->prot.dss1.uus1_data) + 1; l3dss1_msg_with_uus() 939 strcpy(p, pc->prot.dss1.uus1_data); l3dss1_msg_with_uus() 940 p += strlen(pc->prot.dss1.uus1_data); l3dss1_msg_with_uus() 941 pc->prot.dss1.uus1_data[0] = '\0'; l3dss1_msg_with_uus() 956 if (!pc->prot.dss1.uus1_data[0]) l3dss1_release_req() 1777 if (pc->prot.dss1.uus1_data[0]) l3dss1_disconnect_req() 1779 *p++ = strlen(pc->prot.dss1.uus1_data) + 1; l3dss1_disconnect_req() 1781 strcpy(p, pc->prot.dss1.uus1_data); l3dss1_disconnect_req() 1782 p += strlen(pc->prot.dss1.uus1_data); l3dss1_disconnect_req() 1783 pc->prot.dss1.uus1_data[0] = '\0'; l3dss1_disconnect_req() 1897 if (!pc->prot.dss1.uus1_data[0]) l3dss1_alert_req() 2095 strcpy(pc->prot.dss1.uus1_data, pc->chan->setup.eazmsn); /* copy uus element if available */ l3dss1_redir_req() 2102 if (pc->prot.dss1.invoke_id) l3dss1_redir_req() 2103 free_invoke_id(pc->st, pc->prot.dss1.invoke_id); l3dss1_redir_req() 2105 if (!(pc->prot.dss1.invoke_id = new_invoke_id(pc->st))) l3dss1_redir_req() 2121 *p++ = pc->prot.dss1.invoke_id; /* invoke id */ l3dss1_redir_req() 2211 pc->prot.dss1.ll_id = ic->parm.dss1_io.ll_id; /* remember id */ l3dss1_cmd_global() 2212 pc->prot.dss1.proc = ic->parm.dss1_io.proc; /* and procedure */ l3dss1_cmd_global() 2222 { pc->prot.dss1.invoke_id = id; /* remember id */ l3dss1_cmd_global() 2259 ic.parm.dss1_io.hl_id = pc->prot.dss1.invoke_id; l3dss1_io_timer() 2260 ic.parm.dss1_io.ll_id = pc->prot.dss1.ll_id; l3dss1_io_timer() 2261 ic.parm.dss1_io.proc = pc->prot.dss1.proc; l3dss1_io_timer() 2265 free_invoke_id(pc->st, pc->prot.dss1.invoke_id); l3dss1_io_timer() 2266 pc->prot.dss1.invoke_id = 0; /* reset id */ l3dss1_io_timer() 3205 st->prot.dss1.last_invoke_id = 0; setstack_dss1() 3206 st->prot.dss1.invoke_used[0] = 1; /* Bit 0 must always be set to 1 */ setstack_dss1() 3209 st->prot.dss1.invoke_used[i++] = 0; setstack_dss1() 3220 st->l3.global->prot.dss1.invoke_id = 0; setstack_dss1()
|
H A D | l3ni1.c | 54 retval = p->prot.ni1.last_invoke_id + 1; /* try new id */ new_invoke_id() 55 while ((i) && (p->prot.ni1.invoke_used[retval >> 3] == 0xFF)) { new_invoke_id() 56 p->prot.ni1.last_invoke_id = (retval & 0xF8) + 8; new_invoke_id() 60 while (p->prot.ni1.invoke_used[retval >> 3] & (1 << (retval & 7))) new_invoke_id() 64 p->prot.ni1.last_invoke_id = retval; new_invoke_id() 65 p->prot.ni1.invoke_used[retval >> 3] |= (1 << (retval & 7)); new_invoke_id() 77 p->prot.ni1.invoke_used[id >> 3] &= ~(1 << (id & 7)); free_invoke_id() 91 proc->prot.ni1.invoke_id = 0; ni1_new_l3_process() 92 proc->prot.ni1.remote_operation = 0; ni1_new_l3_process() 93 proc->prot.ni1.uus1_data[0] = '\0'; ni1_new_l3_process() 104 free_invoke_id(p->st, p->prot.ni1.invoke_id); ni1_release_l3_process() 118 { if ((pc->callref == -1) && (pc->prot.ni1.invoke_id == id)) l3ni1_search_dummy_proc() 142 ic.parm.ni1_io.hl_id = pc->prot.ni1.invoke_id; l3ni1_dummy_return_result() 143 ic.parm.ni1_io.ll_id = pc->prot.ni1.ll_id; l3ni1_dummy_return_result() 144 ic.parm.ni1_io.proc = pc->prot.ni1.proc; l3ni1_dummy_return_result() 148 free_invoke_id(pc->st, pc->prot.ni1.invoke_id); l3ni1_dummy_return_result() 149 pc->prot.ni1.invoke_id = 0; /* reset id */ l3ni1_dummy_return_result() 175 ic.parm.ni1_io.hl_id = pc->prot.ni1.invoke_id; l3ni1_dummy_error_return() 176 ic.parm.ni1_io.ll_id = pc->prot.ni1.ll_id; l3ni1_dummy_error_return() 177 ic.parm.ni1_io.proc = pc->prot.ni1.proc; l3ni1_dummy_error_return() 181 free_invoke_id(pc->st, pc->prot.ni1.invoke_id); l3ni1_dummy_error_return() 182 pc->prot.ni1.invoke_id = 0; /* reset id */ l3ni1_dummy_error_return() 365 if ((pc->prot.ni1.invoke_id) && (pc->prot.ni1.invoke_id == id)) l3ni1_parse_facility() 367 free_invoke_id(st, pc->prot.ni1.invoke_id); l3ni1_parse_facility() 368 pc->prot.ni1.remote_result = 0; /* success */ l3ni1_parse_facility() 369 pc->prot.ni1.invoke_id = 0; l3ni1_parse_facility() 370 pc->redir_result = pc->prot.ni1.remote_result; l3ni1_parse_facility() 410 if ((pc->prot.ni1.invoke_id) && (pc->prot.ni1.invoke_id == id)) l3ni1_parse_facility() 412 free_invoke_id(st, pc->prot.ni1.invoke_id); l3ni1_parse_facility() 413 pc->prot.ni1.remote_result = err_ret; /* result */ l3ni1_parse_facility() 414 pc->prot.ni1.invoke_id = 0; l3ni1_parse_facility() 415 pc->redir_result = pc->prot.ni1.remote_result; l3ni1_parse_facility() 885 if (pc->prot.ni1.uus1_data[0]) l3ni1_msg_with_uus() 887 *p++ = strlen(pc->prot.ni1.uus1_data) + 1; l3ni1_msg_with_uus() 889 strcpy(p, pc->prot.ni1.uus1_data); l3ni1_msg_with_uus() 890 p += strlen(pc->prot.ni1.uus1_data); l3ni1_msg_with_uus() 891 pc->prot.ni1.uus1_data[0] = '\0'; l3ni1_msg_with_uus() 906 if (!pc->prot.ni1.uus1_data[0]) l3ni1_release_req() 1631 if (pc->prot.ni1.uus1_data[0]) l3ni1_disconnect_req() 1633 *p++ = strlen(pc->prot.ni1.uus1_data) + 1; l3ni1_disconnect_req() 1635 strcpy(p, pc->prot.ni1.uus1_data); l3ni1_disconnect_req() 1636 p += strlen(pc->prot.ni1.uus1_data); l3ni1_disconnect_req() 1637 pc->prot.ni1.uus1_data[0] = '\0'; l3ni1_disconnect_req() 1753 if (!pc->prot.ni1.uus1_data[0]) l3ni1_alert_req() 1951 strcpy(pc->prot.ni1.uus1_data, pc->chan->setup.eazmsn); /* copy uus element if available */ l3ni1_redir_req() 1958 if (pc->prot.ni1.invoke_id) l3ni1_redir_req() 1959 free_invoke_id(pc->st, pc->prot.ni1.invoke_id); l3ni1_redir_req() 1961 if (!(pc->prot.ni1.invoke_id = new_invoke_id(pc->st))) l3ni1_redir_req() 1977 *p++ = pc->prot.ni1.invoke_id; /* invoke id */ l3ni1_redir_req() 2069 pc->prot.ni1.ll_id = ic->parm.ni1_io.ll_id; l3ni1_cmd_global() 2071 pc->prot.ni1.proc = ic->parm.ni1_io.proc; l3ni1_cmd_global() 2082 { pc->prot.ni1.invoke_id = id; /* remember id */ l3ni1_cmd_global() 2119 ic.parm.ni1_io.hl_id = pc->prot.ni1.invoke_id; l3ni1_io_timer() 2120 ic.parm.ni1_io.ll_id = pc->prot.ni1.ll_id; l3ni1_io_timer() 2121 ic.parm.ni1_io.proc = pc->prot.ni1.proc; l3ni1_io_timer() 2125 free_invoke_id(pc->st, pc->prot.ni1.invoke_id); l3ni1_io_timer() 2126 pc->prot.ni1.invoke_id = 0; /* reset id */ l3ni1_io_timer() 3161 st->prot.ni1.last_invoke_id = 0; setstack_ni1() 3162 st->prot.ni1.invoke_used[0] = 1; /* Bit 0 must always be set to 1 */ setstack_ni1() 3165 st->prot.ni1.invoke_used[i++] = 0; setstack_ni1() 3176 st->l3.global->prot.ni1.invoke_id = 0; setstack_ni1()
|
/linux-4.1.27/arch/s390/kernel/ |
H A D | sys_s390.c | 42 unsigned long prot; member in struct:s390_mmap_arg_struct 55 error = sys_mmap_pgoff(a.addr, a.len, a.prot, a.flags, a.fd, a.offset); SYSCALL_DEFINE1()
|
H A D | crash_dump.c | 188 unsigned long size, pgprot_t prot) remap_oldmem_pfn_range_kdump() 197 size_old, prot); remap_oldmem_pfn_range_kdump() 204 return remap_pfn_range(vma, from, pfn, size, prot); remap_oldmem_pfn_range_kdump() 216 unsigned long size, pgprot_t prot) remap_oldmem_pfn_range_zfcpdump() 229 return remap_pfn_range(vma, from, pfn, size, prot); remap_oldmem_pfn_range_zfcpdump() 236 unsigned long pfn, unsigned long size, pgprot_t prot) remap_oldmem_pfn_range() 239 return remap_oldmem_pfn_range_kdump(vma, from, pfn, size, prot); remap_oldmem_pfn_range() 242 prot); remap_oldmem_pfn_range() 186 remap_oldmem_pfn_range_kdump(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) remap_oldmem_pfn_range_kdump() argument 213 remap_oldmem_pfn_range_zfcpdump(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) remap_oldmem_pfn_range_zfcpdump() argument 235 remap_oldmem_pfn_range(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) remap_oldmem_pfn_range() argument
|
/linux-4.1.27/arch/blackfin/include/asm/ |
H A D | pgtable.h | 38 #define pgprot_noncached(prot) (prot)
|
/linux-4.1.27/arch/c6x/include/asm/ |
H A D | pgtable.h | 43 #define pgprot_noncached(prot) (prot)
|
/linux-4.1.27/arch/cris/mm/ |
H A D | ioremap.c | 30 void __iomem * __ioremap_prot(unsigned long phys_addr, unsigned long size, pgprot_t prot) __ioremap_prot() argument 56 phys_addr, prot)) { __ioremap_prot()
|
/linux-4.1.27/tools/perf/util/ |
H A D | event.c | 240 char prot[5]; perf_event__synthesize_mmap_events() local 255 &event->mmap2.start, &event->mmap2.len, prot, perf_event__synthesize_mmap_events() 277 event->mmap2.prot = 0; perf_event__synthesize_mmap_events() 279 if (prot[0] == 'r') perf_event__synthesize_mmap_events() 280 event->mmap2.prot |= PROT_READ; perf_event__synthesize_mmap_events() 281 if (prot[1] == 'w') perf_event__synthesize_mmap_events() 282 event->mmap2.prot |= PROT_WRITE; perf_event__synthesize_mmap_events() 283 if (prot[2] == 'x') perf_event__synthesize_mmap_events() 284 event->mmap2.prot |= PROT_EXEC; perf_event__synthesize_mmap_events() 286 if (prot[3] == 's') perf_event__synthesize_mmap_events() 291 if (prot[2] != 'x') { perf_event__synthesize_mmap_events() 292 if (!mmap_data || prot[0] != 'r') perf_event__synthesize_mmap_events() 712 (event->mmap2.prot & PROT_READ) ? 'r' : '-', perf_event__fprintf_mmap2() 713 (event->mmap2.prot & PROT_WRITE) ? 'w' : '-', perf_event__fprintf_mmap2() 714 (event->mmap2.prot & PROT_EXEC) ? 'x' : '-', perf_event__fprintf_mmap2()
|
/linux-4.1.27/drivers/infiniband/ulp/iser/ |
H A D | iser_initiator.c | 46 * os stored in task->prot[ISER_DIR_IN].data_len 66 struct iser_data_buf *pbuf_in = &iser_task->prot[ISER_DIR_IN]; iser_prepare_read_cmd() 97 * is stored at task->prot[ISER_DIR_OUT].data_len 121 struct iser_data_buf *pbuf_out = &iser_task->prot[ISER_DIR_OUT]; iser_prepare_write_cmd() 397 prot_buf = &iser_task->prot[ISER_DIR_IN]; iser_send_command() 400 prot_buf = &iser_task->prot[ISER_DIR_OUT]; iser_send_command() 660 iser_task->prot[ISER_DIR_IN].data_len = 0; iser_task_rdma_init() 661 iser_task->prot[ISER_DIR_OUT].data_len = 0; iser_task_rdma_init() 693 if (iser_task->prot[ISER_DIR_IN].orig_sg) { iser_task_rdma_finalize() 696 &iser_task->prot[ISER_DIR_IN], iser_task_rdma_finalize() 700 if (iser_task->prot[ISER_DIR_OUT].orig_sg) { iser_task_rdma_finalize() 703 &iser_task->prot[ISER_DIR_OUT], iser_task_rdma_finalize() 715 &iser_task->prot[ISER_DIR_IN], iser_task_rdma_finalize() 727 &iser_task->prot[ISER_DIR_OUT], iser_task_rdma_finalize()
|
/linux-4.1.27/arch/sparc/include/asm/ |
H A D | pgtable_32.h | 262 #define pfn_pte(pfn, prot) mk_pte(pfn_to_page(pfn), prot) 299 static inline pgprot_t pgprot_noncached(pgprot_t prot) pgprot_noncached() argument 301 prot &= ~__pgprot(SRMMU_CACHE); pgprot_noncached() 302 return prot; pgprot_noncached() 414 unsigned long size, pgprot_t prot) io_remap_pfn_range() 422 return remap_pfn_range(vma, from, phys_base >> PAGE_SHIFT, size, prot); io_remap_pfn_range() 412 io_remap_pfn_range(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) io_remap_pfn_range() argument
|
H A D | pgtable_64.h | 237 static inline pte_t pfn_pte(unsigned long pfn, pgprot_t prot) pfn_pte() argument 242 return __pte(paddr | pgprot_val(prot)); pfn_pte() 278 static inline pte_t pte_modify(pte_t pte, pgprot_t prot) pte_modify() argument 333 return __pte((pte_val(pte) & mask) | (pgprot_val(prot) & ~mask)); pte_modify() 347 static inline pgprot_t pgprot_noncached(pgprot_t prot) pgprot_noncached() argument 349 unsigned long val = pgprot_val(prot); pgprot_noncached() 907 #define move_pte(pte, prot, old_addr, new_addr) \ 975 unsigned long size, pgprot_t prot) io_remap_pfn_range() 983 return remap_pfn_range(vma, from, phys_base >> PAGE_SHIFT, size, prot); io_remap_pfn_range() 973 io_remap_pfn_range(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) io_remap_pfn_range() argument
|
/linux-4.1.27/arch/mips/pci/ |
H A D | pci.c | 328 unsigned long prot; pci_mmap_page_range() local 341 prot = pgprot_val(vma->vm_page_prot); pci_mmap_page_range() 342 prot = (prot & ~_CACHE_MASK) | _CACHE_UNCACHED; pci_mmap_page_range() 343 vma->vm_page_prot = __pgprot(prot); pci_mmap_page_range()
|
/linux-4.1.27/drivers/char/agp/ |
H A D | compat_ioctl.h | 56 * The "prot" down below needs still a "sleep" flag somehow ... 61 compat_int_t prot; /* prot flags for mmap */ member in struct:agp_segment32
|
/linux-4.1.27/drivers/gpu/drm/msm/ |
H A D | msm_iommu.c | 47 struct sg_table *sgt, unsigned len, int prot) msm_iommu_map() 65 ret = iommu_map(domain, da, pa, bytes, prot); msm_iommu_map() 46 msm_iommu_map(struct msm_mmu *mmu, uint32_t iova, struct sg_table *sgt, unsigned len, int prot) msm_iommu_map() argument
|
H A D | msm_mmu.h | 27 unsigned len, int prot);
|
/linux-4.1.27/arch/mn10300/include/asm/ |
H A D | pgtable.h | 318 #define pfn_pte(pfn, prot) __pte(__pfn_addr(pfn) | pgprot_val(prot)) 319 #define pfn_pmd(pfn, prot) __pmd(__pfn_addr(pfn) | pgprot_val(prot)) 378 #define pgprot_noncached(prot) __pgprot(pgprot_val(prot) & ~_PAGE_CACHE) 384 #define pgprot_through(prot) __pgprot(pgprot_val(prot) | _PAGE_CACHE_WT)
|
/linux-4.1.27/drivers/base/ |
H A D | dma-mapping.c | 279 unsigned long vm_flags, pgprot_t prot, dma_common_pages_remap() 290 if (map_vm_area(area, prot, pages)) { dma_common_pages_remap() 305 pgprot_t prot, const void *caller) dma_common_contiguous_remap() 319 ptr = dma_common_pages_remap(pages, size, vm_flags, prot, caller); dma_common_contiguous_remap() 278 dma_common_pages_remap(struct page **pages, size_t size, unsigned long vm_flags, pgprot_t prot, const void *caller) dma_common_pages_remap() argument 303 dma_common_contiguous_remap(struct page *page, size_t size, unsigned long vm_flags, pgprot_t prot, const void *caller) dma_common_contiguous_remap() argument
|
/linux-4.1.27/include/linux/sunrpc/ |
H A D | svcsock.h | 64 struct svc_xprt *svc_sock_create(struct svc_serv *serv, int prot);
|
/linux-4.1.27/arch/mn10300/unit-asb2305/ |
H A D | pci-asb2305.c | 219 unsigned long prot; pci_mmap_page_range() local 226 prot = pgprot_val(vma->vm_page_prot); pci_mmap_page_range() 227 prot &= ~_PAGE_CACHE; pci_mmap_page_range() 228 vma->vm_page_prot = __pgprot(prot); pci_mmap_page_range()
|
/linux-4.1.27/arch/openrisc/include/asm/ |
H A D | io.h | 36 pgprot_t prot);
|
/linux-4.1.27/arch/x86/pci/ |
H A D | i386.c | 417 unsigned long prot; pci_mmap_page_range() local 425 prot = pgprot_val(vma->vm_page_prot); pci_mmap_page_range() 436 prot |= cachemode2protval(_PAGE_CACHE_MODE_WC); pci_mmap_page_range() 443 prot |= cachemode2protval(_PAGE_CACHE_MODE_UC_MINUS); pci_mmap_page_range() 445 vma->vm_page_prot = __pgprot(prot); pci_mmap_page_range()
|
/linux-4.1.27/arch/xtensa/kernel/ |
H A D | pci.c | 344 int prot = pgprot_val(vma->vm_page_prot); __pci_mmap_set_pgprot() local 347 prot = (prot & _PAGE_CA_MASK) | _PAGE_CA_WT; __pci_mmap_set_pgprot() 350 prot |= _PAGE_WRITETHRU; __pci_mmap_set_pgprot() 352 vma->vm_page_prot = __pgprot(prot); __pci_mmap_set_pgprot()
|
/linux-4.1.27/arch/sh/kernel/cpu/sh4/ |
H A D | sq.c | 104 static int __sq_remap(struct sq_mapping *map, pgprot_t prot) __sq_remap() argument 117 vma->phys_addr, prot)) { __sq_remap() 139 * @prot: Protection bits. 146 const char *name, pgprot_t prot) sq_remap() 181 ret = __sq_remap(map, prot); sq_remap() 145 sq_remap(unsigned long phys, unsigned int size, const char *name, pgprot_t prot) sq_remap() argument
|
/linux-4.1.27/security/ |
H A D | security.c | 735 static inline unsigned long mmap_prot(struct file *file, unsigned long prot) mmap_prot() argument 741 if ((prot & (PROT_READ | PROT_EXEC)) != PROT_READ) mmap_prot() 742 return prot; mmap_prot() 744 return prot; mmap_prot() 749 return prot | PROT_EXEC; mmap_prot() 759 return prot; mmap_prot() 762 return prot | PROT_EXEC; mmap_prot() 765 return prot; mmap_prot() 768 int security_mmap_file(struct file *file, unsigned long prot, security_mmap_file() argument 772 ret = security_ops->mmap_file(file, prot, security_mmap_file() 773 mmap_prot(file, prot), flags); security_mmap_file() 776 return ima_file_mmap(file, prot); security_mmap_file() 785 unsigned long prot) security_file_mprotect() 787 return security_ops->file_mprotect(vma, reqprot, prot); security_file_mprotect() 784 security_file_mprotect(struct vm_area_struct *vma, unsigned long reqprot, unsigned long prot) security_file_mprotect() argument
|
/linux-4.1.27/arch/xtensa/include/asm/ |
H A D | pgtable.h | 276 #define pgprot_noncached(prot) (__pgprot(pgprot_val(prot) & ~_PAGE_CA_MASK)) 286 #define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) 287 #define mk_pte(page, prot) pfn_pte(page_to_pfn(page), prot)
|
/linux-4.1.27/arch/metag/kernel/ |
H A D | sys_metag.c | 41 unsigned long prot, unsigned long flags, sys_mmap2() 50 return sys_mmap_pgoff(addr, len, prot, flags, fd, pgoff); sys_mmap2() 40 sys_mmap2(unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long fd, unsigned long pgoff) sys_mmap2() argument
|
H A D | process.c | 379 struct elf_phdr *eppnt, int prot, int type, __metag_elf_map() 410 map_addr = vm_mmap(filep, addr, total_size, prot, type, off); __metag_elf_map() 414 map_addr = vm_mmap(filep, addr, size, prot, type, off); __metag_elf_map() 378 __metag_elf_map(struct file *filep, unsigned long addr, struct elf_phdr *eppnt, int prot, int type, unsigned long total_size) __metag_elf_map() argument
|
/linux-4.1.27/arch/microblaze/include/asm/ |
H A D | pgtable.h | 51 #define pgprot_noncached_wc(prot) prot 104 #define pgprot_noncached(prot) \ pte_mkspecial() 105 (__pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | \ pte_mkspecial() 108 #define pgprot_noncached_wc(prot) \ pte_mkspecial() 109 (__pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | \ pte_mkspecial() 312 #define pfn_pte(pfn, prot) \ 313 __pte(((pte_basic_t)(pfn) << PFN_SHIFT_OFFSET) | pgprot_val(prot))
|
H A D | highmem.h | 55 extern void *kmap_atomic_prot(struct page *page, pgprot_t prot);
|
/linux-4.1.27/arch/frv/mm/ |
H A D | dma-alloc.c | 50 static int map_page(unsigned long va, unsigned long pa, pgprot_t prot) map_page() argument 67 set_pte(pte, mk_pte_phys(pa & PAGE_MASK, prot)); map_page()
|
/linux-4.1.27/arch/ia64/pci/ |
H A D | pci.c | 594 pgprot_t prot; pci_mmap_page_range() local 612 prot = phys_mem_access_prot(NULL, vma->vm_pgoff, size, pci_mmap_page_range() 621 ((pgprot_val(prot) & _PAGE_MA_MASK) == _PAGE_MA_UC || pci_mmap_page_range() 622 (pgprot_val(prot) & _PAGE_MA_MASK) == _PAGE_MA_WC) && pci_mmap_page_range() 626 vma->vm_page_prot = prot; pci_mmap_page_range() 665 pgprot_t prot; pci_mmap_legacy_page_range() local 678 prot = phys_mem_access_prot(NULL, vma->vm_pgoff, size, pci_mmap_legacy_page_range() 686 vma->vm_page_prot = prot; pci_mmap_legacy_page_range()
|
/linux-4.1.27/security/apparmor/ |
H A D | lsm.c | 469 static int common_mmap(int op, struct file *file, unsigned long prot, common_mmap() argument 477 if (prot & PROT_READ) common_mmap() 483 if ((prot & PROT_WRITE) && !(flags & MAP_PRIVATE)) common_mmap() 485 if (prot & PROT_EXEC) common_mmap() 492 unsigned long prot, unsigned long flags) apparmor_mmap_file() 494 return common_mmap(OP_FMMAP, file, prot, flags); apparmor_mmap_file() 498 unsigned long reqprot, unsigned long prot) apparmor_file_mprotect() 500 return common_mmap(OP_FMPROT, vma->vm_file, prot, apparmor_file_mprotect() 491 apparmor_mmap_file(struct file *file, unsigned long reqprot, unsigned long prot, unsigned long flags) apparmor_mmap_file() argument 497 apparmor_file_mprotect(struct vm_area_struct *vma, unsigned long reqprot, unsigned long prot) apparmor_file_mprotect() argument
|
/linux-4.1.27/arch/arm/xen/ |
H A D | enlighten.c | 59 int *err_ptr, pgprot_t prot, xen_remap_domain_mfn_array() 64 prot, domid, pages); xen_remap_domain_mfn_array() 72 pgprot_t prot, unsigned domid, xen_remap_domain_mfn_range() 56 xen_remap_domain_mfn_array(struct vm_area_struct *vma, unsigned long addr, xen_pfn_t *mfn, int nr, int *err_ptr, pgprot_t prot, unsigned domid, struct page **pages) xen_remap_domain_mfn_array() argument 69 xen_remap_domain_mfn_range(struct vm_area_struct *vma, unsigned long addr, xen_pfn_t mfn, int nr, pgprot_t prot, unsigned domid, struct page **pages) xen_remap_domain_mfn_range() argument
|
/linux-4.1.27/arch/alpha/kernel/ |
H A D | proto.h | 195 pgprot_t prot; __alpha_remap_area_pages() local 197 prot = __pgprot(_PAGE_VALID | _PAGE_ASM | _PAGE_KRE __alpha_remap_area_pages() 199 return ioremap_page_range(address, address + size, phys_addr, prot); __alpha_remap_area_pages()
|
/linux-4.1.27/arch/arc/include/asm/ |
H A D | pgtable.h | 271 #define pfn_pte(pfn, prot) (__pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot))) 312 #define pgprot_noncached(prot) (__pgprot(pgprot_val(prot) & ~_PAGE_CACHEABLE)) 373 * remap a physical page `pfn' of size `size' with page protection `prot'
|
/linux-4.1.27/net/phonet/ |
H A D | af_phonet.c | 49 if (pp && !try_module_get(pp->prot->owner)) phonet_proto_get() 58 module_put(pp->prot->owner); phonet_proto_put() 100 sk = sk_alloc(net, PF_PHONET, GFP_KERNEL, pnp->prot); pn_socket_create() 480 err = proto_register(pp->prot, 1); phonet_proto_register() 502 proto_unregister(pp->prot); phonet_proto_unregister()
|
/linux-4.1.27/drivers/staging/android/ |
H A D | ashmem.c | 354 static inline vm_flags_t calc_vm_may_flags(unsigned long prot) calc_vm_may_flags() argument 356 return _calc_vm_trans(prot, PROT_READ, VM_MAYREAD) | calc_vm_may_flags() 357 _calc_vm_trans(prot, PROT_WRITE, VM_MAYWRITE) | calc_vm_may_flags() 358 _calc_vm_trans(prot, PROT_EXEC, VM_MAYEXEC); calc_vm_may_flags() 484 static int set_prot_mask(struct ashmem_area *asma, unsigned long prot) set_prot_mask() argument 491 if (unlikely((asma->prot_mask & prot) != prot)) { set_prot_mask() 497 if ((prot & PROT_READ) && (current->personality & READ_IMPLIES_EXEC)) set_prot_mask() 498 prot |= PROT_EXEC; set_prot_mask() 500 asma->prot_mask = prot; set_prot_mask()
|
/linux-4.1.27/fs/proc/ |
H A D | vmcore.c | 162 unsigned long size, pgprot_t prot) remap_oldmem_pfn_range() 164 return remap_pfn_range(vma, from, pfn, size, prot); remap_oldmem_pfn_range() 339 * @prot: protection bits 345 unsigned long size, pgprot_t prot) remap_oldmem_pfn_checked() 367 prot)) remap_oldmem_pfn_checked() 374 PAGE_SIZE, prot)) remap_oldmem_pfn_checked() 384 map_size, prot)) remap_oldmem_pfn_checked() 395 unsigned long size, pgprot_t prot) vmcore_remap_oldmem_pfn() 402 return remap_oldmem_pfn_checked(vma, from, pfn, size, prot); vmcore_remap_oldmem_pfn() 404 return remap_oldmem_pfn_range(vma, from, pfn, size, prot); vmcore_remap_oldmem_pfn() 160 remap_oldmem_pfn_range(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) remap_oldmem_pfn_range() argument 343 remap_oldmem_pfn_checked(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) remap_oldmem_pfn_checked() argument 393 vmcore_remap_oldmem_pfn(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) vmcore_remap_oldmem_pfn() argument
|
/linux-4.1.27/arch/frv/include/asm/ |
H A D | pgtable.h | 271 #define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) 272 #define pfn_pmd(pfn, prot) __pmd(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) 411 #define pgprot_noncached(prot) (__pgprot(pgprot_val(prot) | _PAGE_NOCACHE))
|
/linux-4.1.27/security/integrity/ima/ |
H A D | ima_main.c | 261 * @prot: contains the protection that will be applied by the kernel. 269 int ima_file_mmap(struct file *file, unsigned long prot) ima_file_mmap() argument 271 if (file && (prot & PROT_EXEC)) ima_file_mmap()
|
/linux-4.1.27/drivers/target/ |
H A D | target_core_file.c | 491 sector_t prot_length, prot; fd_do_prot_fill() local 501 for (prot = 0; prot < prot_length;) { fd_do_prot_fill() 502 sector_t len = min_t(sector_t, bufsize, prot_length - prot); fd_do_prot_fill() 503 ssize_t ret = kernel_write(prot_fd, buf, len, pos + prot); fd_do_prot_fill() 506 pr_err("vfs_write to prot file failed: %zd\n", ret); fd_do_prot_fill() 509 prot += ret; fd_do_prot_fill() 523 pr_err("Unable to allocate FILEIO prot buf\n"); fd_do_prot_unmap() 881 pr_err("Unable to allocate FILEIO prot buf\n"); fd_format_prot()
|
/linux-4.1.27/drivers/ata/ |
H A D | sata_sil24.c | 41 __le16 prot; member in struct:sil24_prb 802 u8 prot = qc->tf.protocol; sil24_qc_defer() local 823 int is_excl = (ata_is_atapi(prot) || sil24_qc_defer() 858 u16 prot = 0; sil24_qc_prep() local 861 prot |= PRB_PROT_NCQ; sil24_qc_prep() 863 prot |= PRB_PROT_WRITE; sil24_qc_prep() 865 prot |= PRB_PROT_READ; sil24_qc_prep() 866 prb->prot = cpu_to_le16(prot); sil24_qc_prep()
|
/linux-4.1.27/drivers/gpu/drm/tegra/ |
H A D | gem.c | 97 int prot = IOMMU_READ | IOMMU_WRITE; tegra_bo_iommu_map() local 118 bo->sgt->nents, prot); tegra_bo_iommu_map() 504 pgprot_t prot = vm_get_page_prot(vma->vm_flags); tegra_drm_mmap() local 509 vma->vm_page_prot = pgprot_writecombine(prot); tegra_drm_mmap()
|
/linux-4.1.27/fs/pstore/ |
H A D | ram_core.c | 389 pgprot_t prot; persistent_ram_vmap() local 397 prot = pgprot_noncached(PAGE_KERNEL); persistent_ram_vmap() 399 prot = pgprot_writecombine(PAGE_KERNEL); persistent_ram_vmap() 412 vaddr = vmap(pages, page_count, VM_MAP, prot); persistent_ram_vmap()
|
/linux-4.1.27/include/net/phonet/ |
H A D | phonet.h | 107 struct proto *prot; member in struct:phonet_protocol
|
/linux-4.1.27/arch/microblaze/pci/ |
H A D | pci-common.c | 239 pgprot_t prot = protection; __pci_mmap_set_pgprot() local 254 return pgprot_noncached(prot); __pci_mmap_set_pgprot() 265 pgprot_t prot) pci_phys_mem_access_prot() 273 return prot; pci_phys_mem_access_prot() 275 prot = pgprot_noncached(prot); for_each_pci_dev() 296 prot = pgprot_noncached_wc(prot); 300 pr_debug("PCI: Non-PCI map for %llx, prot: %lx\n", 301 (unsigned long long)offset, pgprot_val(prot)); 303 return prot; 262 pci_phys_mem_access_prot(struct file *file, unsigned long pfn, unsigned long size, pgprot_t prot) pci_phys_mem_access_prot() argument
|
/linux-4.1.27/arch/x86/kernel/ |
H A D | sys_x86_64.c | 87 unsigned long, prot, unsigned long, flags, SYSCALL_DEFINE6() 95 error = sys_mmap_pgoff(addr, len, prot, flags, fd, off >> PAGE_SHIFT); SYSCALL_DEFINE6()
|