Searched refs:prot (Results 1 - 200 of 386) sorted by relevance

12

/linux-4.1.27/arch/powerpc/include/asm/
H A Dmman.h21 static inline unsigned long arch_calc_vm_prot_bits(unsigned long prot) arch_calc_vm_prot_bits() argument
23 return (prot & PROT_SAO) ? VM_SAO : 0; arch_calc_vm_prot_bits()
25 #define arch_calc_vm_prot_bits(prot) arch_calc_vm_prot_bits(prot)
33 static inline int arch_validate_prot(unsigned long prot) arch_validate_prot() argument
35 if (prot & ~(PROT_READ | PROT_WRITE | PROT_EXEC | PROT_SEM | PROT_SAO)) arch_validate_prot()
37 if ((prot & PROT_SAO) && !cpu_has_feature(CPU_FTR_SAO)) arch_validate_prot()
41 #define arch_validate_prot(prot) arch_validate_prot(prot)
H A Dsyscalls.h12 unsigned long prot, unsigned long flags,
15 unsigned long prot, unsigned long flags,
H A Dpgtable.h187 #define pgprot_noncached(prot) (__pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | \
190 #define pgprot_noncached_wc(prot) (__pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | \
193 #define pgprot_cached(prot) (__pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | \
196 #define pgprot_cached_wthru(prot) (__pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | \
199 #define pgprot_cached_noncoherent(prot) \
200 (__pgprot(pgprot_val(prot) & ~_PAGE_CACHE_CTL))
H A Dpgtable-ppc64-4k.h85 #define remap_4k_pfn(vma, addr, pfn, prot) \
86 remap_pfn_range((vma), (addr), (pfn), PAGE_SIZE, (prot))
H A Dpte-hash64-64k.h97 #define remap_4k_pfn(vma, addr, pfn, prot) \
100 __pgprot(pgprot_val((prot)) | _PAGE_4K_PFN)))
H A Dhighmem.h63 extern void *kmap_atomic_prot(struct page *page, pgprot_t prot);
/linux-4.1.27/arch/x86/include/asm/
H A Dfb.h11 unsigned long prot; fb_pgprotect() local
13 prot = pgprot_val(vma->vm_page_prot) & ~_PAGE_CACHE_MASK; fb_pgprotect()
16 prot | cachemode2protval(_PAGE_CACHE_MODE_UC_MINUS); fb_pgprotect()
H A Diomap.h30 iomap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot);
36 iomap_create_wc(resource_size_t base, unsigned long size, pgprot_t *prot);
H A Dhighmem.h66 void *kmap_atomic_prot(struct page *page, pgprot_t prot);
70 void *kmap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot);
/linux-4.1.27/net/ipv6/
H A Dprotocol.c32 int inet6_add_protocol(const struct inet6_protocol *prot, unsigned char protocol) inet6_add_protocol() argument
35 NULL, prot) ? 0 : -1; inet6_add_protocol()
39 int inet6_del_protocol(const struct inet6_protocol *prot, unsigned char protocol) inet6_del_protocol() argument
44 prot, NULL) == prot) ? 0 : -1; inet6_del_protocol()
56 int inet6_add_offload(const struct net_offload *prot, unsigned char protocol) inet6_add_offload() argument
59 NULL, prot) ? 0 : -1; inet6_add_offload()
63 int inet6_del_offload(const struct net_offload *prot, unsigned char protocol) inet6_del_offload() argument
68 prot, NULL) == prot) ? 0 : -1; inet6_del_offload()
H A Dudplite.c65 .prot = &udplitev6_prot,
/linux-4.1.27/include/linux/
H A Dmman.h38 #define arch_calc_vm_prot_bits(prot) 0
50 * Returns true if the prot flags are valid
52 static inline int arch_validate_prot(unsigned long prot) arch_validate_prot() argument
54 return (prot & ~(PROT_READ | PROT_WRITE | PROT_EXEC | PROT_SEM)) == 0; arch_validate_prot()
70 * Combine the mmap "prot" argument into "vm_flags" used internally.
73 calc_vm_prot_bits(unsigned long prot) calc_vm_prot_bits() argument
75 return _calc_vm_trans(prot, PROT_READ, VM_READ ) | calc_vm_prot_bits()
76 _calc_vm_trans(prot, PROT_WRITE, VM_WRITE) | calc_vm_prot_bits()
77 _calc_vm_trans(prot, PROT_EXEC, VM_EXEC) | calc_vm_prot_bits()
78 arch_calc_vm_prot_bits(prot); calc_vm_prot_bits()
H A Dio-mapping.h41 pgprot_t prot; member in struct:io_mapping
55 pgprot_t prot; io_mapping_create_wc() local
61 if (iomap_create_wc(base, size, &prot)) io_mapping_create_wc()
66 iomap->prot = prot; io_mapping_create_wc()
93 return iomap_atomic_prot_pfn(pfn, mapping->prot); io_mapping_map_atomic_wc()
H A Dagpgart.h51 * The "prot" down below needs still a "sleep" flag somehow ...
56 int prot; /* prot flags for mmap */ member in struct:agp_segment
62 pgprot_t prot; member in struct:agp_segment_priv
H A Dima.h20 extern int ima_file_mmap(struct file *file, unsigned long prot);
40 static inline int ima_file_mmap(struct file *file, unsigned long prot) ima_file_mmap() argument
H A Dvmalloc.h58 int node, pgprot_t prot);
77 extern void *__vmalloc(unsigned long size, gfp_t gfp_mask, pgprot_t prot);
80 pgprot_t prot, unsigned long vm_flags, int node,
86 unsigned long flags, pgprot_t prot);
123 extern int map_vm_area(struct vm_struct *area, pgprot_t prot,
127 pgprot_t prot, struct page **pages);
133 pgprot_t prot, struct page **pages) map_kernel_range_noflush()
132 map_kernel_range_noflush(unsigned long start, unsigned long size, pgprot_t prot, struct page **pages) map_kernel_range_noflush() argument
H A Diommu.h148 phys_addr_t paddr, size_t size, int prot);
152 struct scatterlist *sg, unsigned int nents, int prot);
164 phys_addr_t paddr, u64 size, int prot);
197 phys_addr_t paddr, size_t size, int prot);
202 int prot);
245 int prot);
290 unsigned int nents, int prot) iommu_map_sg()
292 return domain->ops->map_sg(domain, iova, sg, nents, prot); iommu_map_sg()
336 phys_addr_t paddr, int gfp_order, int prot) iommu_map()
349 unsigned int nents, int prot) iommu_map_sg()
356 u64 size, int prot) iommu_domain_window_enable()
288 iommu_map_sg(struct iommu_domain *domain, unsigned long iova, struct scatterlist *sg, unsigned int nents, int prot) iommu_map_sg() argument
335 iommu_map(struct iommu_domain *domain, unsigned long iova, phys_addr_t paddr, int gfp_order, int prot) iommu_map() argument
347 iommu_map_sg(struct iommu_domain *domain, unsigned long iova, struct scatterlist *sg, unsigned int nents, int prot) iommu_map_sg() argument
354 iommu_domain_window_enable(struct iommu_domain *domain, u32 wnd_nr, phys_addr_t paddr, u64 size, int prot) iommu_domain_window_enable() argument
H A Dio.h32 phys_addr_t phys_addr, pgprot_t prot);
35 phys_addr_t phys_addr, pgprot_t prot) ioremap_page_range()
34 ioremap_page_range(unsigned long addr, unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_page_range() argument
H A Dcrash_dump.h23 unsigned long size, pgprot_t prot);
/linux-4.1.27/net/ipv4/
H A Dprotocol.c35 int inet_add_protocol(const struct net_protocol *prot, unsigned char protocol) inet_add_protocol() argument
37 if (!prot->netns_ok) { inet_add_protocol()
44 NULL, prot) ? 0 : -1; inet_add_protocol()
48 int inet_add_offload(const struct net_offload *prot, unsigned char protocol) inet_add_offload() argument
51 NULL, prot) ? 0 : -1; inet_add_offload()
55 int inet_del_protocol(const struct net_protocol *prot, unsigned char protocol) inet_del_protocol() argument
60 prot, NULL) == prot) ? 0 : -1; inet_del_protocol()
68 int inet_del_offload(const struct net_offload *prot, unsigned char protocol) inet_del_offload() argument
73 prot, NULL) == prot) ? 0 : -1; inet_del_offload()
H A Dudplite.c71 .prot = &udplite_prot,
/linux-4.1.27/arch/sh/include/asm/
H A Dsyscalls.h7 unsigned long prot, unsigned long flags,
10 unsigned long prot, unsigned long flags,
H A Dmmu.h62 unsigned long size, pgprot_t prot);
64 pgprot_t prot, void *caller);
71 unsigned long size, pgprot_t prot) pmb_bolt_mapping()
78 pgprot_t prot, void *caller) pmb_remap_caller()
99 pmb_remap(phys_addr_t phys, unsigned long size, pgprot_t prot) pmb_remap() argument
101 return pmb_remap_caller(phys, size, prot, __builtin_return_address(0)); pmb_remap()
70 pmb_bolt_mapping(unsigned long virt, phys_addr_t phys, unsigned long size, pgprot_t prot) pmb_bolt_mapping() argument
77 pmb_remap_caller(phys_addr_t phys, unsigned long size, pgprot_t prot, void *caller) pmb_remap_caller() argument
H A Dio.h270 pgprot_t prot, void *caller);
274 __ioremap(phys_addr_t offset, unsigned long size, pgprot_t prot) __ioremap() argument
276 return __ioremap_caller(offset, size, prot, __builtin_return_address(0)); __ioremap()
280 __ioremap_29bit(phys_addr_t offset, unsigned long size, pgprot_t prot) __ioremap_29bit() argument
292 u64 flags = pgprot_val(prot); __ioremap_29bit()
315 __ioremap_mode(phys_addr_t offset, unsigned long size, pgprot_t prot) __ioremap_mode() argument
323 ret = __ioremap_29bit(offset, size, prot); __ioremap_mode()
327 return __ioremap(offset, size, prot); __ioremap_mode()
330 #define __ioremap(offset, size, prot) ((void __iomem *)(offset))
331 #define __ioremap_mode(offset, size, prot) ((void __iomem *)(offset))
360 ioremap_fixed(phys_addr_t phys_addr, unsigned long size, pgprot_t prot) ioremap_fixed() argument
H A Dpgtable_32.h317 #define pfn_pte(pfn, prot) \
318 __pte(((unsigned long long)(pfn) << PAGE_SHIFT) | pgprot_val(prot))
319 #define pfn_pmd(pfn, prot) \
320 __pmd(((unsigned long long)(pfn) << PAGE_SHIFT) | pgprot_val(prot))
378 #define pgprot_writecombine(prot) \
379 __pgprot(pgprot_val(prot) & ~_PAGE_CACHABLE)
H A Dpgtable_64.h207 #define pgprot_writecombine(prot) __pgprot(pgprot_val(prot) & ~_PAGE_CACHABLE)
307 #define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
308 #define pfn_pmd(pfn, prot) __pmd(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
/linux-4.1.27/include/trace/events/
H A Dsock.h34 TP_PROTO(struct sock *sk, struct proto *prot, long allocated),
36 TP_ARGS(sk, prot, allocated),
47 strncpy(__entry->name, prot->name, 32);
48 __entry->sysctl_mem = prot->sysctl_mem;
50 __entry->sysctl_rmem = prot->sysctl_rmem[0];
/linux-4.1.27/arch/um/include/asm/
H A Dpgtable-2level.h41 #define pfn_pte(pfn, prot) __pte(pfn_to_phys(pfn) | pgprot_val(prot))
42 #define pfn_pmd(pfn, prot) __pmd(pfn_to_phys(pfn) | pgprot_val(prot))
H A Dpage.h48 #define pte_set_val(pte, phys, prot) \
50 (pte).pte_low = (phys) | pgprot_val(prot); })
77 #define pte_set_val(p, phys, prot) (p).pte = (phys | pgprot_val(prot))
/linux-4.1.27/arch/avr32/include/asm/
H A Dpgtable-2level.h42 #define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
43 #define pfn_pmd(pfn, prot) __pmd(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
H A Dpgtable.h263 * Mark the prot value as uncacheable and unbufferable
265 #define pgprot_noncached(prot) \
266 __pgprot(pgprot_val(prot) & ~(_PAGE_BUFFER | _PAGE_CACHABLE))
269 * Mark the prot value as uncacheable but bufferable
271 #define pgprot_writecombine(prot) \
272 __pgprot((pgprot_val(prot) & ~_PAGE_CACHABLE) | _PAGE_BUFFER)
/linux-4.1.27/arch/um/kernel/
H A Dsyscall.c18 unsigned long prot, unsigned long flags, old_mmap()
25 err = sys_mmap_pgoff(addr, len, prot, flags, fd, offset >> PAGE_SHIFT); old_mmap()
17 old_mmap(unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long fd, unsigned long offset) old_mmap() argument
H A Dtlb.c24 unsigned int prot; member in struct:host_vm_change::host_vm_op::__anon2946::__anon2947
35 unsigned int prot; member in struct:host_vm_change::host_vm_op::__anon2946::__anon2949
64 op->u.mmap.prot, op->u.mmap.fd, do_ops()
73 op->u.mprotect.len, op->u.mprotect.prot, do_ops()
88 unsigned int prot, struct host_vm_change *hvc) add_mmap()
99 (last->u.mmap.prot == prot) && (last->u.mmap.fd == fd) && add_mmap()
115 .prot = prot, add_mmap()
153 unsigned int prot, struct host_vm_change *hvc) add_mprotect()
162 (last->u.mprotect.prot == prot)) { add_mprotect()
177 .prot = prot } } }); add_mprotect()
188 int r, w, x, prot, ret = 0; update_pte_range() local
204 prot = ((r ? UM_PROT_READ : 0) | (w ? UM_PROT_WRITE : 0) | update_pte_range()
209 PAGE_SIZE, prot, hvc); update_pte_range()
213 ret = add_mprotect(addr, PAGE_SIZE, prot, hvc); update_pte_range()
391 int r, w, x, prot, err = 0; flush_tlb_page() local
420 prot = ((r ? UM_PROT_READ : 0) | (w ? UM_PROT_WRITE : 0) | flush_tlb_page()
428 err = map(mm_id, address, PAGE_SIZE, prot, fd, offset, flush_tlb_page()
434 err = protect(mm_id, address, PAGE_SIZE, prot, 1, &flush); flush_tlb_page()
87 add_mmap(unsigned long virt, unsigned long phys, unsigned long len, unsigned int prot, struct host_vm_change *hvc) add_mmap() argument
152 add_mprotect(unsigned long addr, unsigned long len, unsigned int prot, struct host_vm_change *hvc) add_mprotect() argument
/linux-4.1.27/include/asm-generic/
H A Dsyscalls.h14 unsigned long prot, unsigned long flags,
20 unsigned long prot, unsigned long flags,
H A Dpgtable.h246 #define move_pte(pte, prot, old_addr, new_addr) (pte)
258 #define pgprot_noncached(prot) (prot)
492 static inline int track_pfn_remap(struct vm_area_struct *vma, pgprot_t *prot, track_pfn_remap() argument
503 static inline int track_pfn_insert(struct vm_area_struct *vma, pgprot_t *prot, track_pfn_insert() argument
528 extern int track_pfn_remap(struct vm_area_struct *vma, pgprot_t *prot,
531 extern int track_pfn_insert(struct vm_area_struct *vma, pgprot_t *prot,
701 int pud_set_huge(pud_t *pud, phys_addr_t addr, pgprot_t prot);
702 int pmd_set_huge(pmd_t *pmd, phys_addr_t addr, pgprot_t prot);
706 static inline int pud_set_huge(pud_t *pud, phys_addr_t addr, pgprot_t prot) pud_set_huge() argument
710 static inline int pmd_set_huge(pmd_t *pmd, phys_addr_t addr, pgprot_t prot) pmd_set_huge() argument
/linux-4.1.27/arch/arm/mm/
H A Didmap.c22 unsigned long prot) idmap_add_pmd()
47 *pmd = __pmd((addr & PMD_MASK) | prot); idmap_add_pmd()
53 unsigned long prot) idmap_add_pmd()
57 addr = (addr & PMD_MASK) | prot; idmap_add_pmd()
66 unsigned long prot) idmap_add_pud()
73 idmap_add_pmd(pud, addr, next, prot); idmap_add_pud()
78 const char *text_end, unsigned long prot) identity_mapping_add()
87 prot |= PMD_TYPE_SECT | PMD_SECT_AP_WRITE | PMD_SECT_AF; identity_mapping_add()
90 prot |= PMD_BIT4; identity_mapping_add()
95 idmap_add_pud(pgd, addr, next, prot); identity_mapping_add()
21 idmap_add_pmd(pud_t *pud, unsigned long addr, unsigned long end, unsigned long prot) idmap_add_pmd() argument
52 idmap_add_pmd(pud_t *pud, unsigned long addr, unsigned long end, unsigned long prot) idmap_add_pmd() argument
65 idmap_add_pud(pgd_t *pgd, unsigned long addr, unsigned long end, unsigned long prot) idmap_add_pud() argument
77 identity_mapping_add(pgd_t *pgd, const char *text_start, const char *text_end, unsigned long prot) identity_mapping_add() argument
H A Ddma-mapping.c291 pgprot_t prot, struct page **ret_page,
295 pgprot_t prot, struct page **ret_page,
299 __dma_alloc_remap(struct page *page, size_t size, gfp_t gfp, pgprot_t prot, __dma_alloc_remap() argument
308 prot, caller); __dma_alloc_remap()
349 pgprot_t prot = pgprot_dmacoherent(PAGE_KERNEL); atomic_pool_init() local
359 ptr = __alloc_from_contiguous(NULL, atomic_pool_size, prot, atomic_pool_init()
362 ptr = __alloc_remap_buffer(NULL, atomic_pool_size, gfp, prot, atomic_pool_init()
453 pgprot_t prot = *(pgprot_t *)data; __dma_update_pte() local
455 set_pte_ext(pte, mk_pte(page, prot), 0); __dma_update_pte()
459 static void __dma_remap(struct page *page, size_t size, pgprot_t prot) __dma_remap() argument
464 apply_to_page_range(&init_mm, start, size, __dma_update_pte, &prot); __dma_remap()
469 pgprot_t prot, struct page **ret_page, __alloc_remap_buffer()
480 ptr = __dma_alloc_remap(page, size, gfp, prot, caller); __alloc_remap_buffer()
528 pgprot_t prot, struct page **ret_page, __alloc_from_contiguous()
546 ptr = __dma_alloc_remap(page, size, GFP_KERNEL, prot, caller); __alloc_from_contiguous()
552 __dma_remap(page, size, prot); __alloc_from_contiguous()
573 static inline pgprot_t __get_dma_pgprot(struct dma_attrs *attrs, pgprot_t prot) __get_dma_pgprot() argument
575 prot = dma_get_attr(DMA_ATTR_WRITE_COMBINE, attrs) ? __get_dma_pgprot()
576 pgprot_writecombine(prot) : __get_dma_pgprot()
577 pgprot_dmacoherent(prot); __get_dma_pgprot()
578 return prot; __get_dma_pgprot()
587 #define __get_dma_pgprot(attrs, prot) __pgprot(0)
588 #define __alloc_remap_buffer(dev, size, gfp, prot, ret, c, wv) NULL
590 #define __alloc_from_contiguous(dev, size, prot, ret, c, wv) NULL
612 gfp_t gfp, pgprot_t prot, bool is_coherent, __dma_alloc()
653 addr = __alloc_remap_buffer(dev, size, gfp, prot, &page, caller, want_vaddr); __dma_alloc()
655 addr = __alloc_from_contiguous(dev, size, prot, &page, caller, want_vaddr); __dma_alloc()
670 pgprot_t prot = __get_dma_pgprot(attrs, PAGE_KERNEL); arm_dma_alloc() local
676 return __dma_alloc(dev, size, handle, gfp, prot, false, arm_dma_alloc()
683 pgprot_t prot = __get_dma_pgprot(attrs, PAGE_KERNEL); arm_coherent_dma_alloc() local
689 return __dma_alloc(dev, size, handle, gfp, prot, true, arm_coherent_dma_alloc()
1226 __iommu_alloc_remap(struct page **pages, size_t size, gfp_t gfp, pgprot_t prot, __iommu_alloc_remap() argument
1230 VM_ARM_DMA_CONSISTENT | VM_USERMAP, prot, caller); __iommu_alloc_remap()
1347 pgprot_t prot = __get_dma_pgprot(attrs, PAGE_KERNEL); arm_iommu_alloc_attrs() local
1377 addr = __iommu_alloc_remap(pages, size, gfp, prot, arm_iommu_alloc_attrs()
1470 int prot; __dma_direction_to_prot() local
1474 prot = IOMMU_READ | IOMMU_WRITE; __dma_direction_to_prot()
1477 prot = IOMMU_READ; __dma_direction_to_prot()
1480 prot = IOMMU_WRITE; __dma_direction_to_prot()
1483 prot = 0; __dma_direction_to_prot()
1486 return prot; __dma_direction_to_prot()
1502 int prot; __map_sg_chunk() local
1519 prot = __dma_direction_to_prot(dir); __map_sg_chunk()
1521 ret = iommu_map(mapping->domain, iova, phys, len, prot); __map_sg_chunk()
1720 int ret, prot, len = PAGE_ALIGN(size + offset); arm_coherent_iommu_map_page() local
1726 prot = __dma_direction_to_prot(dir); arm_coherent_iommu_map_page()
1728 ret = iommu_map(mapping->domain, dma_addr, page_to_phys(page), len, prot); arm_coherent_iommu_map_page()
468 __alloc_remap_buffer(struct device *dev, size_t size, gfp_t gfp, pgprot_t prot, struct page **ret_page, const void *caller, bool want_vaddr) __alloc_remap_buffer() argument
527 __alloc_from_contiguous(struct device *dev, size_t size, pgprot_t prot, struct page **ret_page, const void *caller, bool want_vaddr) __alloc_from_contiguous() argument
611 __dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, gfp_t gfp, pgprot_t prot, bool is_coherent, struct dma_attrs *attrs, const void *caller) __dma_alloc() argument
H A Dinit.c575 pmdval_t prot; member in struct:section_perm
585 .prot = PMD_SECT_XN,
592 .prot = PMD_SECT_XN,
600 .prot = PMD_SECT_XN,
613 .prot = L_PMD_SECT_RDONLY,
616 .prot = PMD_SECT_APX | PMD_SECT_AP_WRITE,
629 pmdval_t prot) section_update()
638 pmd[0] = __pmd((pmd_val(pmd[0]) & mask) | prot); section_update()
641 pmd[1] = __pmd((pmd_val(pmd[1]) & mask) | prot); section_update()
643 pmd[0] = __pmd((pmd_val(pmd[0]) & mask) | prot); section_update()
684 set_section_perms(nx_perms, prot); fix_kernmem_perms()
690 set_section_perms(ro_perms, prot); mark_rodata_ro()
700 set_section_perms(ro_perms, prot); set_kernel_text_ro()
628 section_update(unsigned long addr, pmdval_t mask, pmdval_t prot) section_update() argument
H A Ddump.c221 u64 prot = val & pg_level[level].mask; note_page() local
225 st->current_prot = prot; note_page()
227 } else if (prot != st->current_prot || level != st->level || note_page()
252 st->current_prot = prot; note_page()
H A Dfault-armv.c243 pgprot_t prot = __pgprot_modify(PAGE_KERNEL, check_writebuffer_bugs() local
246 p1 = vmap(&page, 1, VM_IOREMAP, prot); check_writebuffer_bugs()
247 p2 = vmap(&page, 1, VM_IOREMAP, prot); check_writebuffer_bugs()
/linux-4.1.27/lib/
H A Dioremap.c54 unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_pte_range()
65 set_pte_at(&init_mm, addr, pte, pfn_pte(pfn, prot)); ioremap_pte_range()
72 unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_pmd_range()
87 if (pmd_set_huge(pmd, phys_addr + addr, prot)) ioremap_pmd_range()
91 if (ioremap_pte_range(pmd, addr, next, phys_addr + addr, prot)) ioremap_pmd_range()
98 unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_pud_range()
113 if (pud_set_huge(pud, phys_addr + addr, prot)) ioremap_pud_range()
117 if (ioremap_pmd_range(pud, addr, next, phys_addr + addr, prot)) ioremap_pud_range()
124 unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_page_range()
138 err = ioremap_pud_range(pgd, addr, next, phys_addr+addr, prot); ioremap_page_range()
53 ioremap_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_pte_range() argument
71 ioremap_pmd_range(pud_t *pud, unsigned long addr, unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_pmd_range() argument
97 ioremap_pud_range(pgd_t *pgd, unsigned long addr, unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_pud_range() argument
123 ioremap_page_range(unsigned long addr, unsigned long end, phys_addr_t phys_addr, pgprot_t prot) ioremap_page_range() argument
/linux-4.1.27/arch/x86/mm/
H A Diomap_32.c34 int iomap_create_wc(resource_size_t base, unsigned long size, pgprot_t *prot) iomap_create_wc() argument
46 *prot = __pgprot(__PAGE_KERNEL | cachemode2protval(pcm)); iomap_create_wc()
57 void *kmap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot) kmap_atomic_prot_pfn() argument
67 set_pte(kmap_pte - idx, pfn_pte(pfn, prot)); kmap_atomic_prot_pfn()
74 * Map 'pfn' using protections 'prot'
77 iomap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot) iomap_atomic_prot_pfn() argument
85 if (!pat_enabled && pgprot_val(prot) == iomap_atomic_prot_pfn()
87 prot = __pgprot(__PAGE_KERNEL | iomap_atomic_prot_pfn()
90 return (void __force __iomem *) kmap_atomic_prot_pfn(pfn, prot); iomap_atomic_prot_pfn()
H A Ddump_pagetables.c124 static void printk_prot(struct seq_file *m, pgprot_t prot, int level, bool dmsg) printk_prot() argument
126 pgprotval_t pr = pgprot_val(prot); printk_prot()
130 if (!pgprot_val(prot)) { printk_prot()
193 pgprotval_t prot, cur; note_page() local
201 prot = pgprot_val(new_prot) & PTE_FLAGS_MASK; note_page()
212 } else if (prot != cur || level != st->level || note_page()
275 pgprot_t prot = pte_pgprot(*start); walk_pte_level() local
278 note_page(m, st, prot, 4); walk_pte_level()
295 pgprotval_t prot = pmd_val(*start) & PTE_FLAGS_MASK; walk_pmd_level() local
298 note_page(m, st, __pgprot(prot), 3); walk_pmd_level()
327 pgprotval_t prot = pud_val(*start) & PTE_FLAGS_MASK; walk_pud_level() local
330 note_page(m, st, __pgprot(prot), 2); walk_pud_level()
365 pgprotval_t prot = pgd_val(*start) & PTE_FLAGS_MASK; ptdump_walk_pgd_level() local
368 note_page(m, &st, __pgprot(prot), 1); ptdump_walk_pgd_level()
H A Dioremap.c91 pgprot_t prot; __ioremap_caller() local
160 prot = PAGE_KERNEL_IO; __ioremap_caller()
164 prot = __pgprot(pgprot_val(prot) | __ioremap_caller()
168 prot = __pgprot(pgprot_val(prot) | __ioremap_caller()
172 prot = __pgprot(pgprot_val(prot) | __ioremap_caller()
191 if (ioremap_page_range(vaddr, vaddr + size, phys_addr, prot)) __ioremap_caller()
H A Dpat.c712 * Internal interface to reserve a range of physical memory with prot.
714 * this func also keeps identity mapping (if any) in sync with this new prot.
800 * If the vma has a linear pfn mapping for the entire range, we get the prot
806 unsigned long prot; track_pfn_copy() local
815 if (follow_phys(vma, vma->vm_start, 0, &prot, &paddr)) { track_pfn_copy()
819 pgprot = __pgprot(prot); track_pfn_copy()
827 * prot is passed in as a parameter for the new mapping. If the vma has a
831 int track_pfn_remap(struct vm_area_struct *vma, pgprot_t *prot, track_pfn_remap() argument
841 ret = reserve_pfn_range(paddr, size, prot, 0); track_pfn_remap()
851 * For anything smaller than the vma size we set prot based on the track_pfn_remap()
864 *prot = __pgprot((pgprot_val(vma->vm_page_prot) & (~_PAGE_CACHE_MASK)) | track_pfn_remap()
870 int track_pfn_insert(struct vm_area_struct *vma, pgprot_t *prot, track_pfn_insert() argument
878 /* Set prot based on lookup */ track_pfn_insert()
880 *prot = __pgprot((pgprot_val(vma->vm_page_prot) & (~_PAGE_CACHE_MASK)) | track_pfn_insert()
895 unsigned long prot; untrack_pfn() local
903 if (follow_phys(vma, vma->vm_start, 0, &prot, &paddr)) { untrack_pfn()
914 pgprot_t pgprot_writecombine(pgprot_t prot) pgprot_writecombine() argument
917 return __pgprot(pgprot_val(prot) | pgprot_writecombine()
920 return pgprot_noncached(prot); pgprot_writecombine()
H A Dhighmem_32.c33 void *kmap_atomic_prot(struct page *page, pgprot_t prot) kmap_atomic_prot() argument
48 set_pte(kmap_pte-idx, mk_pte(page, prot)); kmap_atomic_prot()
H A Dpgtable.c566 int pud_set_huge(pud_t *pud, phys_addr_t addr, pgprot_t prot) pud_set_huge() argument
578 prot = pgprot_4k_2_large(prot); pud_set_huge()
582 __pgprot(pgprot_val(prot) | _PAGE_PSE))); pud_set_huge()
587 int pmd_set_huge(pmd_t *pmd, phys_addr_t addr, pgprot_t prot) pmd_set_huge() argument
599 prot = pgprot_4k_2_large(prot); pmd_set_huge()
603 __pgprot(pgprot_val(prot) | _PAGE_PSE))); pmd_set_huge()
H A Dinit_64.c331 pgprot_t prot; __init_extra_mapping() local
333 pgprot_val(prot) = pgprot_val(PAGE_KERNEL_LARGE) | __init_extra_mapping()
351 set_pmd(pmd, __pmd(phys | pgprot_val(prot))); __init_extra_mapping()
403 pgprot_t prot) phys_pte_init()
437 set_pte(pte, pfn_pte(addr >> PAGE_SHIFT, prot)); phys_pte_init()
448 unsigned long page_size_mask, pgprot_t prot) phys_pmd_init()
458 pgprot_t new_prot = prot; phys_pmd_init()
474 end, prot); phys_pmd_init()
504 __pgprot(pgprot_val(prot) | _PAGE_PSE))); phys_pmd_init()
532 pgprot_t prot = PAGE_KERNEL; phys_pud_init() local
547 page_size_mask, prot); phys_pud_init()
569 prot = pte_pgprot(pte_clrhuge(*(pte_t *)pud)); phys_pud_init()
585 prot); phys_pud_init()
402 phys_pte_init(pte_t *pte_page, unsigned long addr, unsigned long end, pgprot_t prot) phys_pte_init() argument
447 phys_pmd_init(pmd_t *pmd_page, unsigned long address, unsigned long end, unsigned long page_size_mask, pgprot_t prot) phys_pmd_init() argument
/linux-4.1.27/arch/s390/mm/
H A Ddump_pagetables.c62 unsigned int prot, cur; note_page() local
70 prot = new_prot; note_page()
79 } else if (prot != cur || level != st->level || note_page()
113 unsigned int prot; walk_pte_level() local
120 prot = pte_val(*pte) & (_PAGE_PROTECT | _PAGE_INVALID); walk_pte_level()
121 note_page(m, st, prot, 4); walk_pte_level()
129 unsigned int prot; walk_pmd_level() local
138 prot = pmd_val(*pmd) & _SEGMENT_ENTRY_PROTECT; walk_pmd_level()
139 note_page(m, st, prot, 3); walk_pmd_level()
151 unsigned int prot; walk_pud_level() local
160 prot = pud_val(*pud) & _REGION3_ENTRY_RO; walk_pud_level()
161 note_page(m, st, prot, 2); walk_pud_level()
H A Dhugetlbpage.c19 * prot-none, clean, old 111.000000.1 -> 00..1...1...00 __pte_to_pmd()
20 * prot-none, clean, young 111.000001.1 -> 01..1...1...00 __pte_to_pmd()
21 * prot-none, dirty, old 111.000010.1 -> 10..1...1...00 __pte_to_pmd()
22 * prot-none, dirty, young 111.000011.1 -> 11..1...1...00 __pte_to_pmd()
56 * prot-none, clean, old 00..1...1...00 -> 111.000000.1 __pmd_to_pte()
57 * prot-none, clean, young 01..1...1...00 -> 111.000001.1 __pmd_to_pte()
58 * prot-none, dirty, old 10..1...1...00 -> 111.000010.1 __pmd_to_pte()
59 * prot-none, dirty, young 11..1...1...00 -> 111.000011.1 __pmd_to_pte()
/linux-4.1.27/include/net/
H A Dprotocol.h86 struct proto *prot; member in struct:inet_protosw
103 int inet_add_protocol(const struct net_protocol *prot, unsigned char num);
104 int inet_del_protocol(const struct net_protocol *prot, unsigned char num);
105 int inet_add_offload(const struct net_offload *prot, unsigned char num);
106 int inet_del_offload(const struct net_offload *prot, unsigned char num);
110 int udp_add_offload(struct udp_offload *prot);
111 void udp_del_offload(struct udp_offload *prot);
114 int inet6_add_protocol(const struct inet6_protocol *prot, unsigned char num);
115 int inet6_del_protocol(const struct inet6_protocol *prot, unsigned char num);
119 int inet6_add_offload(const struct net_offload *prot, unsigned char num);
120 int inet6_del_offload(const struct net_offload *prot, unsigned char num);
H A Dsock.h1083 int proto_register(struct proto *prot, int alloc_slab);
1084 void proto_unregister(struct proto *prot);
1177 struct proto *prot = sk->sk_prot; sk_leave_memory_pressure() local
1179 for (; cg_proto; cg_proto = parent_cg_proto(prot, cg_proto)) sk_leave_memory_pressure()
1192 struct proto *prot = sk->sk_prot; sk_enter_memory_pressure() local
1194 for (; cg_proto; cg_proto = parent_cg_proto(prot, cg_proto)) sk_enter_memory_pressure()
1203 long *prot = sk->sk_prot->sysctl_mem; sk_prot_mem_limits() local
1205 prot = sk->sk_cgrp->sysctl_mem; sk_prot_mem_limits()
1206 return prot[index]; sk_prot_mem_limits()
1209 static inline void memcg_memory_allocated_add(struct cg_proto *prot, memcg_memory_allocated_add() argument
1213 page_counter_charge(&prot->memory_allocated, amt); memcg_memory_allocated_add()
1215 if (page_counter_read(&prot->memory_allocated) > memcg_memory_allocated_add()
1216 prot->memory_allocated.limit) memcg_memory_allocated_add()
1220 static inline void memcg_memory_allocated_sub(struct cg_proto *prot, memcg_memory_allocated_sub() argument
1223 page_counter_uncharge(&prot->memory_allocated, amt); memcg_memory_allocated_sub()
1229 struct proto *prot = sk->sk_prot; sk_memory_allocated() local
1234 return atomic_long_read(prot->memory_allocated); sk_memory_allocated()
1240 struct proto *prot = sk->sk_prot; sk_memory_allocated_add() local
1245 atomic_long_add_return(amt, prot->memory_allocated); sk_memory_allocated_add()
1249 return atomic_long_add_return(amt, prot->memory_allocated); sk_memory_allocated_add()
1255 struct proto *prot = sk->sk_prot; sk_memory_allocated_sub() local
1260 atomic_long_sub(amt, prot->memory_allocated); sk_memory_allocated_sub()
1265 struct proto *prot = sk->sk_prot; sk_sockets_allocated_dec() local
1270 for (; cg_proto; cg_proto = parent_cg_proto(prot, cg_proto)) sk_sockets_allocated_dec()
1274 percpu_counter_dec(prot->sockets_allocated); sk_sockets_allocated_dec()
1279 struct proto *prot = sk->sk_prot; sk_sockets_allocated_inc() local
1284 for (; cg_proto; cg_proto = parent_cg_proto(prot, cg_proto)) sk_sockets_allocated_inc()
1288 percpu_counter_inc(prot->sockets_allocated); sk_sockets_allocated_inc()
1294 struct proto *prot = sk->sk_prot; sk_sockets_allocated_read_positive() local
1299 return percpu_counter_read_positive(prot->sockets_allocated); sk_sockets_allocated_read_positive()
1303 proto_sockets_allocated_sum_positive(struct proto *prot) proto_sockets_allocated_sum_positive() argument
1305 return percpu_counter_sum_positive(prot->sockets_allocated); proto_sockets_allocated_sum_positive()
1309 proto_memory_allocated(struct proto *prot) proto_memory_allocated() argument
1311 return atomic_long_read(prot->memory_allocated); proto_memory_allocated()
1315 proto_memory_pressure(struct proto *prot) proto_memory_pressure() argument
1317 if (!prot->memory_pressure) proto_memory_pressure()
1319 return !!*prot->memory_pressure; proto_memory_pressure()
1325 void sock_prot_inuse_add(struct net *net, struct proto *prot, int inc);
1328 static inline void sock_prot_inuse_add(struct net *net, struct proto *prot, sock_prot_inuse_add() argument
1528 struct proto *prot);
/linux-4.1.27/arch/tile/gxio/
H A Dkiorpc.c33 pgprot_t mmio_base, prot = { 0 }; iorpc_ioremap() local
55 prot = hv_pte_set_lotar(prot, hv_pte_get_lotar(mmio_base)); iorpc_ioremap()
58 return ioremap_prot(PFN_PHYS(pfn), size, prot); iorpc_ioremap()
/linux-4.1.27/arch/microblaze/kernel/
H A Dsys_microblaze.c37 unsigned long, prot, unsigned long, flags, unsigned long, fd, SYSCALL_DEFINE6()
43 return sys_mmap_pgoff(addr, len, prot, flags, fd, pgoff >> PAGE_SHIFT); SYSCALL_DEFINE6()
47 unsigned long, prot, unsigned long, flags, unsigned long, fd, SYSCALL_DEFINE6()
53 return sys_mmap_pgoff(addr, len, prot, flags, fd, SYSCALL_DEFINE6()
/linux-4.1.27/arch/score/kernel/
H A Dsys_score.c37 sys_mmap2(unsigned long addr, unsigned long len, unsigned long prot, sys_mmap2() argument
40 return sys_mmap_pgoff(addr, len, prot, flags, fd, pgoff); sys_mmap2()
44 sys_mmap(unsigned long addr, unsigned long len, unsigned long prot, sys_mmap() argument
49 return sys_mmap_pgoff(addr, len, prot, flags, fd, offset >> PAGE_SHIFT); sys_mmap()
/linux-4.1.27/arch/arm/include/asm/
H A Dfixmap.h21 void __set_fixmap(enum fixed_addresses idx, phys_addr_t phys, pgprot_t prot);
H A Dpgtable.h113 #define __pgprot_modify(prot,mask,bits) \
114 __pgprot((pgprot_val(prot) & ~(mask)) | (bits))
116 #define pgprot_noncached(prot) \
117 __pgprot_modify(prot, L_PTE_MT_MASK, L_PTE_MT_UNCACHED)
119 #define pgprot_writecombine(prot) \
120 __pgprot_modify(prot, L_PTE_MT_MASK, L_PTE_MT_BUFFERABLE)
122 #define pgprot_stronglyordered(prot) \
123 __pgprot_modify(prot, L_PTE_MT_MASK, L_PTE_MT_UNCACHED)
126 #define pgprot_dmacoherent(prot) \
127 __pgprot_modify(prot, L_PTE_MT_MASK, L_PTE_MT_BUFFERABLE | L_PTE_XN)
133 #define pgprot_dmacoherent(prot) \
134 __pgprot_modify(prot, L_PTE_MT_MASK, L_PTE_MT_UNCACHED | L_PTE_XN)
210 #define pfn_pte(pfn,prot) __pte(__pfn_to_phys(pfn) | pgprot_val(prot))
213 #define mk_pte(page,prot) pfn_pte(page_to_pfn(page), prot)
255 static inline pte_t clear_pte_bit(pte_t pte, pgprot_t prot) clear_pte_bit() argument
257 pte_val(pte) &= ~pgprot_val(prot); clear_pte_bit()
261 static inline pte_t set_pte_bit(pte_t pte, pgprot_t prot) set_pte_bit() argument
263 pte_val(pte) |= pgprot_val(prot); set_pte_bit()
H A Dpgalloc.h132 pmdval_t prot) __pmd_populate()
134 pmdval_t pmdval = (pte + PTE_HWTABLE_OFF) | prot; __pmd_populate()
161 pmdval_t prot; pmd_populate() local
164 prot = user_pmd_table; pmd_populate()
166 prot = _PAGE_USER_TABLE; pmd_populate()
168 __pmd_populate(pmdp, page_to_phys(ptep), prot); pmd_populate()
131 __pmd_populate(pmd_t *pmdp, phys_addr_t pte, pmdval_t prot) __pmd_populate() argument
H A Dpgtable-nommu.h64 * Mark the prot value as uncacheable and unbufferable.
66 #define pgprot_noncached(prot) __pgprot(0)
67 #define pgprot_writecombine(prot) __pgprot(0)
68 #define pgprot_dmacoherent(prot) __pgprot(0)
H A Dpgtable-3level.h257 #define pfn_pmd(pfn,prot) (__pmd(((phys_addr_t)(pfn) << PAGE_SHIFT) | pgprot_val(prot)))
258 #define mk_pmd(page,prot) pfn_pmd(page_to_pfn(page),prot)
/linux-4.1.27/arch/arm64/include/asm/
H A Dpgtable.h122 #define pfn_pte(pfn,prot) (__pte(((phys_addr_t)(pfn) << PAGE_SHIFT) | pgprot_val(prot)))
153 static inline pte_t clear_pte_bit(pte_t pte, pgprot_t prot) clear_pte_bit() argument
155 pte_val(pte) &= ~pgprot_val(prot); clear_pte_bit()
159 static inline pte_t set_pte_bit(pte_t pte, pgprot_t prot) set_pte_bit() argument
161 pte_val(pte) |= pgprot_val(prot); set_pte_bit()
268 static inline pgprot_t mk_sect_prot(pgprot_t prot) mk_sect_prot() argument
270 return __pgprot(pgprot_val(prot) & ~PTE_TABLE_BIT); mk_sect_prot()
305 #define pfn_pmd(pfn,prot) (__pmd(((phys_addr_t)(pfn) << PAGE_SHIFT) | pgprot_val(prot)))
306 #define mk_pmd(page,prot) pfn_pmd(page_to_pfn(page),prot)
318 #define __pgprot_modify(prot,mask,bits) \
319 __pgprot((pgprot_val(prot) & ~(mask)) | (bits))
322 * Mark the prot value as uncacheable and unbufferable.
324 #define pgprot_noncached(prot) \
325 __pgprot_modify(prot, PTE_ATTRINDX_MASK, PTE_ATTRINDX(MT_DEVICE_nGnRnE) | PTE_PXN | PTE_UXN)
326 #define pgprot_writecombine(prot) \
327 __pgprot_modify(prot, PTE_ATTRINDX_MASK, PTE_ATTRINDX(MT_NORMAL_NC) | PTE_PXN | PTE_UXN)
328 #define pgprot_device(prot) \
329 __pgprot_modify(prot, PTE_ATTRINDX_MASK, PTE_ATTRINDX(MT_DEVICE_nGnRE) | PTE_PXN | PTE_UXN)
377 #define mk_pte(page,prot) pfn_pte(page_to_pfn(page),prot)
H A Dmmu.h36 pgprot_t prot);
H A Dpgalloc.h111 pmdval_t prot) __pmd_populate()
113 set_pmd(pmdp, __pmd(pte | prot)); __pmd_populate()
110 __pmd_populate(pmd_t *pmdp, phys_addr_t pte, pmdval_t prot) __pmd_populate() argument
H A Dfixmap.h68 extern void __set_fixmap(enum fixed_addresses idx, phys_addr_t phys, pgprot_t prot);
/linux-4.1.27/arch/mn10300/kernel/
H A Dsys_mn10300.c27 unsigned long prot, unsigned long flags, old_mmap()
32 return sys_mmap_pgoff(addr, len, prot, flags, fd, offset >> PAGE_SHIFT); old_mmap()
26 old_mmap(unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long fd, unsigned long offset) old_mmap() argument
/linux-4.1.27/arch/cris/kernel/
H A Dsys_cris.c30 sys_mmap2(unsigned long addr, unsigned long len, unsigned long prot, sys_mmap2() argument
34 return sys_mmap_pgoff(addr, len, prot, flags, fd, pgoff); sys_mmap2()
/linux-4.1.27/arch/ia64/include/asm/
H A Dunistd.h40 int prot, int flags,
44 int prot, int flags,
H A Dpgtable.h351 #define pgprot_cacheable(prot) __pgprot((pgprot_val(prot) & ~_PAGE_MA_MASK) | _PAGE_MA_WB)
352 #define pgprot_noncached(prot) __pgprot((pgprot_val(prot) & ~_PAGE_MA_MASK) | _PAGE_MA_UC)
353 #define pgprot_writecombine(prot) __pgprot((pgprot_val(prot) & ~_PAGE_MA_MASK) | _PAGE_MA_WC)
/linux-4.1.27/arch/x86/um/shared/sysdep/
H A Dsyscalls_64.h27 unsigned long prot, unsigned long flags,
/linux-4.1.27/arch/sh/kernel/
H A Dsys_sh.c32 unsigned long prot, unsigned long flags, old_mmap()
37 return sys_mmap_pgoff(addr, len, prot, flags, fd, off>>PAGE_SHIFT); old_mmap()
41 unsigned long prot, unsigned long flags, sys_mmap2()
53 return sys_mmap_pgoff(addr, len, prot, flags, fd, pgoff); sys_mmap2()
31 old_mmap(unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, int fd, unsigned long off) old_mmap() argument
40 sys_mmap2(unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long fd, unsigned long pgoff) sys_mmap2() argument
/linux-4.1.27/arch/m32r/include/asm/
H A Dpgtable-2level.h71 #define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
72 #define pfn_pmd(pfn, prot) __pmd(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
H A Dpgtable.h267 unsigned long prot = pgprot_val(_prot); pgprot_noncached() local
269 prot |= _PAGE_NONCACHABLE; pgprot_noncached()
270 return __pgprot(prot); pgprot_noncached()
273 #define pgprot_writecombine(prot) pgprot_noncached(prot)
/linux-4.1.27/arch/arc/mm/
H A Dioremap.c47 pgprot_t prot = __pgprot(flags); ioremap_prot() local
59 prot = pgprot_noncached(prot); ioremap_prot()
75 (unsigned long)vaddr + size, paddr, prot)) { ioremap_prot()
/linux-4.1.27/arch/sh/boards/mach-landisk/
H A Dsetup.c57 pgprot_t prot; landisk_devices_setup() local
63 prot = PAGE_KERNEL_PCC(1, _PAGE_PCC_IO16); landisk_devices_setup()
64 cf_ide_base = ioremap_prot(paddrbase, PAGE_SIZE, pgprot_val(prot)); landisk_devices_setup()
/linux-4.1.27/arch/sh/boards/mach-lboxre2/
H A Dsetup.c53 pgprot_t prot; lboxre2_devices_setup() local
59 prot = PAGE_KERNEL_PCC(1, _PAGE_PCC_IO16); lboxre2_devices_setup()
60 cf0_io_base = (u32)ioremap_prot(paddrbase, psize, pgprot_val(prot)); lboxre2_devices_setup()
/linux-4.1.27/arch/sh/boards/mach-sh03/
H A Dsetup.c78 pgprot_t prot; sh03_devices_setup() local
84 prot = PAGE_KERNEL_PCC(1, _PAGE_PCC_IO16); sh03_devices_setup()
85 cf_ide_base = ioremap_prot(paddrbase, PAGE_SIZE, pgprot_val(prot)); sh03_devices_setup()
/linux-4.1.27/fs/affs/
H A Damigaffs.c388 prot_to_mode(u32 prot) prot_to_mode() argument
392 if (!(prot & FIBF_NOWRITE)) prot_to_mode()
394 if (!(prot & FIBF_NOREAD)) prot_to_mode()
396 if (!(prot & FIBF_NOEXECUTE)) prot_to_mode()
398 if (prot & FIBF_GRP_WRITE) prot_to_mode()
400 if (prot & FIBF_GRP_READ) prot_to_mode()
402 if (prot & FIBF_GRP_EXECUTE) prot_to_mode()
404 if (prot & FIBF_OTR_WRITE) prot_to_mode()
406 if (prot & FIBF_OTR_READ) prot_to_mode()
408 if (prot & FIBF_OTR_EXECUTE) prot_to_mode()
417 u32 prot = AFFS_I(inode)->i_protect; mode_to_prot() local
421 prot |= FIBF_NOEXECUTE; mode_to_prot()
423 prot |= FIBF_NOREAD; mode_to_prot()
425 prot |= FIBF_NOWRITE; mode_to_prot()
427 prot |= FIBF_GRP_EXECUTE; mode_to_prot()
429 prot |= FIBF_GRP_READ; mode_to_prot()
431 prot |= FIBF_GRP_WRITE; mode_to_prot()
433 prot |= FIBF_OTR_EXECUTE; mode_to_prot()
435 prot |= FIBF_OTR_READ; mode_to_prot()
437 prot |= FIBF_OTR_WRITE; mode_to_prot()
439 AFFS_I(inode)->i_protect = prot; mode_to_prot()
H A Dinode.c24 u32 prot; affs_iget() local
49 prot = be32_to_cpu(tail->protect); affs_iget()
56 AFFS_I(inode)->i_protect = prot; affs_iget()
72 inode->i_mode = prot_to_mode(prot); affs_iget()
/linux-4.1.27/include/xen/
H A Dxen-ops.h39 * @prot: page protection mask
52 int *err_ptr, pgprot_t prot,
61 * @prot: page protection mask
71 pgprot_t prot, unsigned domid,
78 int *err_ptr, pgprot_t prot,
/linux-4.1.27/arch/parisc/kernel/
H A Dentry.S50 /* space_to_prot macro creates a prot id from a space id */
53 .macro space_to_prot spc prot
54 depd,z \spc,62,31,\prot
57 .macro space_to_prot spc prot
58 extrd,u \spc,(64 - (SPACEID_SHIFT)),32,\prot
509 /* Drop prot bits and convert to page addr for iitlbt and idtlbt */
517 /* Convert the pte and prot to tlb insertion values. How
519 .macro make_insert_tlb spc,pte,prot
520 space_to_prot \spc \prot /* create prot id from space */
533 depd \pte,8,7,\prot
539 depdi 7,11,3,\prot
544 depd %r0,11,2,\prot /* If Gateway, Set PL2 to 0 */
553 depdi 1,12,1,\prot
555 /* Drop prot bits and convert to page addr for iitlbt and idtlbt */
562 .macro make_insert_tlb_11 spc,pte,prot
563 zdep \spc,30,15,\prot
564 dep \pte,8,7,\prot
566 depi 1,12,1,\prot
568 depi 7,11,3,\prot /* Set for user space (1 rsvd for read) */
570 depi 0,11,2,\prot /* If Gateway, Set PL2 to 0 */
572 /* Get rid of prot bits and convert to page addr for iitlba */
596 .macro do_alias spc,tmp,tmp1,va,pte,prot,fault,patype
617 ldi (_PAGE_REFTRAP|_PAGE_READ|_PAGE_WRITE),\prot
620 * instruction) nullify the next load so \prot is only T.
624 ldi (_PAGE_DIRTY|_PAGE_READ|_PAGE_WRITE),\prot
626 depd,z \prot,8,7,\prot
629 depw,z \prot,8,7,\prot
1134 prot = r17 /* prot bits */ define
1150 make_insert_tlb spc,pte,prot
1152 idtlbt pte,prot
1159 do_alias spc,t0,t1,va,pte,prot,dtlb_fault,20
1161 idtlbt pte,prot
1176 make_insert_tlb spc,pte,prot
1178 idtlbt pte,prot
1185 do_alias spc,t0,t1,va,pte,prot,nadtlb_emulate,20
1187 idtlbt pte,prot
1204 make_insert_tlb_11 spc,pte,prot
1210 idtlbp prot,(%sr1,va)
1219 do_alias spc,t0,t1,va,pte,prot,dtlb_fault,11
1222 idtlbp prot,(va)
1237 make_insert_tlb_11 spc,pte,prot
1243 idtlbp prot,(%sr1,va)
1252 do_alias spc,t0,t1,va,pte,prot,nadtlb_emulate,11
1255 idtlbp prot,(va)
1270 make_insert_tlb spc,pte,prot
1274 idtlbt pte,prot
1281 do_alias spc,t0,t1,va,pte,prot,dtlb_fault,20
1283 idtlbt pte,prot
1298 make_insert_tlb spc,pte,prot
1302 idtlbt pte,prot
1309 do_alias spc,t0,t1,va,pte,prot,nadtlb_emulate,20
1311 idtlbt pte,prot
1407 make_insert_tlb spc,pte,prot
1409 iitlbt pte,prot
1431 make_insert_tlb spc,pte,prot
1433 iitlbt pte,prot
1440 do_alias spc,t0,t1,va,pte,prot,naitlb_fault,20
1442 iitlbt pte,prot
1459 make_insert_tlb_11 spc,pte,prot
1465 iitlbp prot,(%sr1,va)
1483 make_insert_tlb_11 spc,pte,prot
1489 iitlbp prot,(%sr1,va)
1498 do_alias spc,t0,t1,va,pte,prot,itlb_fault,11
1501 iitlbp prot,(%sr0, va)
1517 make_insert_tlb spc,pte,prot
1521 iitlbt pte,prot
1537 make_insert_tlb spc,pte,prot
1541 iitlbt pte,prot
1548 do_alias spc,t0,t1,va,pte,prot,naitlb_fault,20
1550 iitlbt pte,prot
1569 make_insert_tlb spc,pte,prot
1571 idtlbt pte,prot
1589 make_insert_tlb_11 spc,pte,prot
1595 idtlbp prot,(%sr1,va)
1613 make_insert_tlb spc,pte,prot
1617 idtlbt pte,prot
H A Dpci.c226 unsigned long prot; pci_mmap_page_range() local
242 prot = pgprot_val(vma->vm_page_prot); pci_mmap_page_range()
243 prot |= _PAGE_NO_CACHE; pci_mmap_page_range()
244 vma->vm_page_prot = __pgprot(prot); pci_mmap_page_range()
H A Dsys_parisc.c265 unsigned long prot, unsigned long flags, unsigned long fd, sys_mmap2()
270 return sys_mmap_pgoff(addr, len, prot, flags, fd, sys_mmap2()
275 unsigned long prot, unsigned long flags, unsigned long fd, sys_mmap()
279 return sys_mmap_pgoff(addr, len, prot, flags, fd, sys_mmap()
371 asmlinkage unsigned long sys_alloc_hugepages(int key, unsigned long addr, unsigned long len, int prot, int flag) sys_alloc_hugepages() argument
264 sys_mmap2(unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long fd, unsigned long pgoff) sys_mmap2() argument
274 sys_mmap(unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long fd, unsigned long offset) sys_mmap() argument
/linux-4.1.27/arch/powerpc/kernel/
H A Dsyscalls.c45 unsigned long prot, unsigned long flags, do_mmap2()
50 if (!arch_validate_prot(prot)) do_mmap2()
59 ret = sys_mmap_pgoff(addr, len, prot, flags, fd, off); do_mmap2()
65 unsigned long prot, unsigned long flags, sys_mmap2()
68 return do_mmap2(addr, len, prot, flags, fd, pgoff, PAGE_SHIFT-12); sys_mmap2()
72 unsigned long prot, unsigned long flags, sys_mmap()
75 return do_mmap2(addr, len, prot, flags, fd, offset, PAGE_SHIFT); sys_mmap()
44 do_mmap2(unsigned long addr, size_t len, unsigned long prot, unsigned long flags, unsigned long fd, unsigned long off, int shift) do_mmap2() argument
64 sys_mmap2(unsigned long addr, size_t len, unsigned long prot, unsigned long flags, unsigned long fd, unsigned long pgoff) sys_mmap2() argument
71 sys_mmap(unsigned long addr, size_t len, unsigned long prot, unsigned long flags, unsigned long fd, off_t offset) sys_mmap() argument
H A Dsys_ppc32.c65 unsigned long prot, unsigned long flags, compat_sys_mmap2()
69 return sys_mmap(addr, len, prot, flags, fd, pgoff << 12); compat_sys_mmap2()
64 compat_sys_mmap2(unsigned long addr, size_t len, unsigned long prot, unsigned long flags, unsigned long fd, unsigned long pgoff) compat_sys_mmap2() argument
/linux-4.1.27/arch/m68k/include/asm/
H A Dpgtable_mm.h155 # define pgprot_noncached(prot) (__pgprot(pgprot_val(prot) | CF_PAGE_NOCACHE))
162 #define pgprot_noncached(prot) \
164 ? (__pgprot(pgprot_val(prot) | __SUN3_PAGE_NOCACHE)) \
166 ? (__pgprot(pgprot_val(prot) | _PAGE_NOCACHE030)) \
168 ? (__pgprot((pgprot_val(prot) & _CACHEMASK040) | _PAGE_NOCACHE_S)) \
169 : (prot)))
/linux-4.1.27/arch/arm64/mm/
H A Dmmu.c91 pgprot_t prot, alloc_init_pte()
107 set_pte(pte, pfn_pte(pfn, prot)); alloc_init_pte()
115 pgprot_t prot = __pgprot(pud_val(*old_pud) ^ addr); split_pud() local
119 set_pmd(pmd, __pmd(addr | prot)); split_pud()
126 phys_addr_t phys, pgprot_t prot, alloc_init_pmd()
156 pgprot_val(mk_sect_prot(prot)))); alloc_init_pmd()
171 prot, alloc); alloc_init_pmd()
191 phys_addr_t phys, pgprot_t prot, alloc_init_pud()
213 pgprot_val(mk_sect_prot(prot)))); alloc_init_pud()
231 alloc_init_pmd(mm, pud, addr, next, phys, prot, alloc); alloc_init_pud()
243 phys_addr_t size, pgprot_t prot, __create_mapping()
254 alloc_init_pud(mm, pgd, addr, next, phys, prot, alloc); __create_mapping()
270 phys_addr_t size, pgprot_t prot) create_mapping()
278 size, prot, early_alloc); create_mapping()
283 pgprot_t prot) create_pgd_mapping()
285 __create_mapping(mm, pgd_offset(mm, virt), phys, virt, size, prot, create_pgd_mapping()
290 phys_addr_t size, pgprot_t prot) create_mapping_late()
299 phys, virt, size, prot, late_alloc); create_mapping_late()
89 alloc_init_pte(pmd_t *pmd, unsigned long addr, unsigned long end, unsigned long pfn, pgprot_t prot, void *(*alloc)(unsigned long size)) alloc_init_pte() argument
124 alloc_init_pmd(struct mm_struct *mm, pud_t *pud, unsigned long addr, unsigned long end, phys_addr_t phys, pgprot_t prot, void *(*alloc)(unsigned long size)) alloc_init_pmd() argument
189 alloc_init_pud(struct mm_struct *mm, pgd_t *pgd, unsigned long addr, unsigned long end, phys_addr_t phys, pgprot_t prot, void *(*alloc)(unsigned long size)) alloc_init_pud() argument
241 __create_mapping(struct mm_struct *mm, pgd_t *pgd, phys_addr_t phys, unsigned long virt, phys_addr_t size, pgprot_t prot, void *(*alloc)(unsigned long size)) __create_mapping() argument
269 create_mapping(phys_addr_t phys, unsigned long virt, phys_addr_t size, pgprot_t prot) create_mapping() argument
281 create_pgd_mapping(struct mm_struct *mm, phys_addr_t phys, unsigned long virt, phys_addr_t size, pgprot_t prot) create_pgd_mapping() argument
289 create_mapping_late(phys_addr_t phys, unsigned long virt, phys_addr_t size, pgprot_t prot) create_mapping_late() argument
H A Dioremap.c33 pgprot_t prot, void *caller) __ioremap_caller()
67 err = ioremap_page_range(addr, addr + size, phys_addr, prot); __ioremap_caller()
76 void __iomem *__ioremap(phys_addr_t phys_addr, size_t size, pgprot_t prot) __ioremap() argument
78 return __ioremap_caller(phys_addr, size, prot, __ioremap()
32 __ioremap_caller(phys_addr_t phys_addr, size_t size, pgprot_t prot, void *caller) __ioremap_caller() argument
H A Ddump.c188 u64 prot = val & pg_level[level].mask; note_page() local
192 st->current_prot = prot; note_page()
195 } else if (prot != st->current_prot || level != st->level || note_page()
222 st->current_prot = prot; note_page()
H A Ddma-mapping.c34 static pgprot_t __get_dma_pgprot(struct dma_attrs *attrs, pgprot_t prot, __get_dma_pgprot() argument
38 return pgprot_writecombine(prot); __get_dma_pgprot()
39 return prot; __get_dma_pgprot()
358 pgprot_t prot = __pgprot(PROT_NORMAL_NC); atomic_pool_init() local
382 VM_USERMAP, prot, atomic_pool_init); atomic_pool_init()
/linux-4.1.27/drivers/vfio/
H A Dvfio_iommu_type1.c68 int prot; /* IOMMU_CACHE */ member in struct:vfio_domain
77 int prot; /* IOMMU_READ/WRITE */ member in struct:vfio_dma
219 static int put_pfn(unsigned long pfn, int prot) put_pfn() argument
223 if (prot & IOMMU_WRITE) put_pfn()
231 static int vaddr_get_pfn(unsigned long vaddr, int prot, unsigned long *pfn) vaddr_get_pfn() argument
237 if (get_user_pages_fast(vaddr, 1, !!(prot & IOMMU_WRITE), page) == 1) { vaddr_get_pfn()
263 int prot, unsigned long *pfn_base) vfio_pin_pages()
273 ret = vaddr_get_pfn(vaddr, prot, pfn_base); vfio_pin_pages()
280 put_pfn(*pfn_base, prot); vfio_pin_pages()
296 ret = vaddr_get_pfn(vaddr, prot, &pfn); vfio_pin_pages()
302 put_pfn(pfn, prot); vfio_pin_pages()
308 put_pfn(pfn, prot); vfio_pin_pages()
322 int prot, bool do_accounting) vfio_unpin_pages()
328 unlocked += put_pfn(pfn++, prot); vfio_unpin_pages()
387 dma->prot, false); vfio_unmap_unpin()
502 unsigned long pfn, long npage, int prot) map_try_harder()
510 PAGE_SIZE, prot | domain->prot); map_try_harder()
522 unsigned long pfn, long npage, int prot) vfio_iommu_map()
529 npage << PAGE_SHIFT, prot | d->prot); vfio_iommu_map()
532 map_try_harder(d, iova, pfn, npage, prot)) vfio_iommu_map()
555 int ret = 0, prot = 0; vfio_dma_do_map() local
570 prot |= IOMMU_WRITE; vfio_dma_do_map()
572 prot |= IOMMU_READ; vfio_dma_do_map()
574 if (!prot || !size || (size | iova | vaddr) & mask) vfio_dma_do_map()
596 dma->prot = prot; vfio_dma_do_map()
604 size >> PAGE_SHIFT, prot, &pfn); vfio_dma_do_map()
612 ret = vfio_iommu_map(iommu, iova + dma->size, pfn, npage, prot); vfio_dma_do_map()
614 vfio_unpin_pages(pfn, npage, prot, true); vfio_dma_do_map()
680 size, dma->prot | domain->prot); vfio_iommu_replay()
711 IOMMU_READ | IOMMU_WRITE | domain->prot); vfio_test_domain_fgsp()
790 domain->prot |= IOMMU_CACHE; vfio_iommu_type1_attach_group()
801 d->prot == domain->prot) { vfio_iommu_type1_attach_group()
945 if (!(domain->prot & IOMMU_CACHE)) { vfio_domains_have_iommu_cache()
262 vfio_pin_pages(unsigned long vaddr, long npage, int prot, unsigned long *pfn_base) vfio_pin_pages() argument
321 vfio_unpin_pages(unsigned long pfn, long npage, int prot, bool do_accounting) vfio_unpin_pages() argument
501 map_try_harder(struct vfio_domain *domain, dma_addr_t iova, unsigned long pfn, long npage, int prot) map_try_harder() argument
521 vfio_iommu_map(struct vfio_iommu *iommu, dma_addr_t iova, unsigned long pfn, long npage, int prot) vfio_iommu_map() argument
/linux-4.1.27/arch/sparc/kernel/
H A Dsys_sparc_32.c99 unsigned long prot, unsigned long flags, unsigned long fd, sys_mmap2()
104 return sys_mmap_pgoff(addr, len, prot, flags, fd, sys_mmap2()
109 unsigned long prot, unsigned long flags, unsigned long fd, sys_mmap()
113 return sys_mmap_pgoff(addr, len, prot, flags, fd, off >> PAGE_SHIFT); sys_mmap()
117 unsigned long prot, unsigned long pgoff, sparc_remap_file_pages()
123 return sys_remap_file_pages(start, size, prot, sparc_remap_file_pages()
98 sys_mmap2(unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long fd, unsigned long pgoff) sys_mmap2() argument
108 sys_mmap(unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long fd, unsigned long off) sys_mmap() argument
116 sparc_remap_file_pages(unsigned long start, unsigned long size, unsigned long prot, unsigned long pgoff, unsigned long flags) sparc_remap_file_pages() argument
H A Dsystbls.h17 unsigned long prot, unsigned long flags,
23 unsigned long prot, unsigned long flags,
26 unsigned long prot, unsigned long pgoff,
H A Dpci_sun4v.c40 unsigned long prot; /* IOMMU page protections */ member in struct:iommu_batch
50 static inline void iommu_batch_start(struct device *dev, unsigned long prot, unsigned long entry) iommu_batch_start() argument
55 p->prot = prot; iommu_batch_start()
65 unsigned long prot = p->prot; iommu_batch_flush() local
74 npages, prot, __pa(pglist)); iommu_batch_flush()
81 npages, prot, __pa(pglist), num); iommu_batch_flush()
244 unsigned long prot; dma_4v_map_page() local
265 prot = HV_PCI_MAP_ATTR_READ; dma_4v_map_page()
267 prot |= HV_PCI_MAP_ATTR_WRITE; dma_4v_map_page()
271 iommu_batch_start(dev, prot, entry); dma_4v_map_page()
328 unsigned long flags, handle, prot; dma_4v_map_sg() local
343 prot = HV_PCI_MAP_ATTR_READ; dma_4v_map_sg()
345 prot |= HV_PCI_MAP_ATTR_WRITE; dma_4v_map_sg()
357 iommu_batch_start(dev, prot, ~0UL); dma_4v_map_sg()
/linux-4.1.27/arch/sh/mm/
H A Dioremap_fixed.c47 ioremap_fixed(phys_addr_t phys_addr, unsigned long size, pgprot_t prot) ioremap_fixed() argument
88 pgprot_val(prot) |= _PAGE_WIRED; ioremap_fixed()
89 __set_fixmap(idx, phys_addr, prot); ioremap_fixed()
H A Dinit.c72 static void set_pte_phys(unsigned long addr, unsigned long phys, pgprot_t prot) set_pte_phys() argument
82 set_pte(pte, pfn_pte(phys >> PAGE_SHIFT, prot)); set_pte_phys()
85 if (pgprot_val(prot) & _PAGE_WIRED) set_pte_phys()
89 static void clear_pte_phys(unsigned long addr, pgprot_t prot) clear_pte_phys() argument
95 if (pgprot_val(prot) & _PAGE_WIRED) clear_pte_phys()
102 void __set_fixmap(enum fixed_addresses idx, unsigned long phys, pgprot_t prot) __set_fixmap() argument
111 set_pte_phys(address, phys, prot); __set_fixmap()
114 void __clear_fixmap(enum fixed_addresses idx, pgprot_t prot) __clear_fixmap() argument
123 clear_pte_phys(address, prot); __clear_fixmap()
H A Dnommu.c98 void __set_fixmap(enum fixed_addresses idx, unsigned long phys, pgprot_t prot) __set_fixmap() argument
H A Dpmb.c117 static inline unsigned long pgprot_to_pmb_flags(pgprot_t prot) pgprot_to_pmb_flags() argument
120 u64 flags = pgprot_val(prot); pgprot_to_pmb_flags()
207 static inline bool pmb_prot_valid(pgprot_t prot) pmb_prot_valid() argument
209 return (pgprot_val(prot) & _PAGE_USER) == 0; pmb_prot_valid()
336 unsigned long size, pgprot_t prot) pmb_bolt_mapping()
355 pmb_flags = pgprot_to_pmb_flags(prot); pmb_bolt_mapping()
411 pgprot_t prot, void *caller) pmb_remap_caller()
428 if (!pmb_prot_valid(prot)) pmb_remap_caller()
454 ret = pmb_bolt_mapping(vaddr, phys, size, prot); pmb_remap_caller()
335 pmb_bolt_mapping(unsigned long vaddr, phys_addr_t phys, unsigned long size, pgprot_t prot) pmb_bolt_mapping() argument
410 pmb_remap_caller(phys_addr_t phys, unsigned long size, pgprot_t prot, void *caller) pmb_remap_caller() argument
/linux-4.1.27/arch/metag/mm/
H A Dioremap.c34 pgprot_t prot; __ioremap() local
52 prot = __pgprot(_PAGE_PRESENT | _PAGE_WRITE | _PAGE_DIRTY | __ioremap()
64 if (ioremap_page_range(addr, addr + size, phys_addr, prot)) { __ioremap()
/linux-4.1.27/arch/avr32/mm/
H A Dioremap.c28 pgprot_t prot; __ioremap() local
56 prot = __pgprot(_PAGE_PRESENT | _PAGE_GLOBAL | _PAGE_RW | _PAGE_DIRTY __ioremap()
67 if (ioremap_page_range(addr, addr + size, phys_addr, prot)) { __ioremap()
/linux-4.1.27/arch/cris/arch-v32/drivers/pci/
H A Dbios.c20 unsigned long prot; pci_mmap_page_range() local
25 prot = pgprot_val(vma->vm_page_prot); pci_mmap_page_range()
26 vma->vm_page_prot = __pgprot(prot); pci_mmap_page_range()
/linux-4.1.27/arch/powerpc/mm/
H A DMakefile35 obj-$(CONFIG_PPC_SUBPAGE_PROT) += subpage-prot.o
H A Dhighmem.c32 void *kmap_atomic_prot(struct page *page, pgprot_t prot) kmap_atomic_prot() argument
48 __set_pte_at(&init_mm, vaddr, kmap_pte-idx, mk_pte(page, prot), 1); kmap_atomic_prot()
H A Dpgtable_32.c397 static int __change_page_attr(struct page *page, pgprot_t prot) __change_page_attr() argument
410 __set_pte_at(&init_mm, address, kpte, mk_pte(page, prot), 0); __change_page_attr()
423 static int change_page_attr(struct page *page, int numpages, pgprot_t prot) change_page_attr() argument
430 err = __change_page_attr(page, prot); change_page_attr()
H A Dhash_utils_64.c182 unsigned long pstart, unsigned long prot, htab_bolt_mapping()
192 prot = htab_convert_pte_flags(prot); htab_bolt_mapping()
195 vstart, vend, pstart, prot, psize, ssize); htab_bolt_mapping()
202 unsigned long tprot = prot; htab_bolt_mapping()
675 unsigned long prot; htab_initialize() local
743 prot = pgprot_val(PAGE_KERNEL); htab_initialize()
762 DBG("creating mapping for region: %lx..%lx (prot: %lx)\n", for_each_memblock()
763 base, size, prot); for_each_memblock()
780 __pa(base), prot, for_each_memblock()
787 prot, for_each_memblock()
794 prot, mmu_linear_psize, mmu_kernel_ssize)); for_each_memblock()
813 __pa(tce_alloc_start), prot,
181 htab_bolt_mapping(unsigned long vstart, unsigned long vend, unsigned long pstart, unsigned long prot, int psize, int ssize) htab_bolt_mapping() argument
/linux-4.1.27/arch/unicore32/include/asm/
H A Dpgtable.h153 #define pfn_pte(pfn, prot) (__pte(((pfn) << PAGE_SHIFT) \
154 | pgprot_val(prot)))
197 * Mark the prot value as uncacheable. pte_mkspecial()
199 #define pgprot_noncached(prot) \ pte_mkspecial()
200 __pgprot(pgprot_val(prot) & ~PTE_CACHEABLE) pte_mkspecial()
201 #define pgprot_writecombine(prot) \ pte_mkspecial()
202 __pgprot(pgprot_val(prot) & ~PTE_CACHEABLE) pte_mkspecial()
203 #define pgprot_dmacoherent(prot) \ pte_mkspecial()
204 __pgprot(pgprot_val(prot) & ~PTE_CACHEABLE) pte_mkspecial()
230 #define mk_pte(page, prot) pfn_pte(page_to_pfn(page), prot)
/linux-4.1.27/drivers/net/ethernet/mellanox/mlx4/
H A Dmcg.c159 u32 prot; new_steering_entry() local
207 prot = be32_to_cpu(mgm->members_count) >> 30; new_steering_entry()
222 mgm->members_count = cpu_to_be32(members_count | (prot << 30)); new_steering_entry()
429 u32 prot; add_promisc_qp() local
475 prot = be32_to_cpu(mgm->members_count) >> 30; add_promisc_qp()
506 (prot << 30)); add_promisc_qp()
696 u8 *gid, enum mlx4_protocol prot, find_entry()
705 u8 op_mod = (prot == MLX4_PROT_ETH) ? find_entry()
740 be32_to_cpu(mgm->members_count) >> 30 == prot) find_entry()
1098 int block_mcast_loopback, enum mlx4_protocol prot, mlx4_qp_attach_common()
1118 err = find_entry(dev, port, gid, prot, mlx4_qp_attach_common()
1164 mgm->members_count = cpu_to_be32(members_count | (u32) prot << 30); mlx4_qp_attach_common()
1184 if (prot == MLX4_PROT_ETH) { mlx4_qp_attach_common()
1207 enum mlx4_protocol prot, enum mlx4_steer_type steer) mlx4_qp_detach_common()
1226 err = find_entry(dev, port, gid, prot, mlx4_qp_detach_common()
1240 if (prot == MLX4_PROT_ETH && mlx4_qp_detach_common()
1261 mgm->members_count = cpu_to_be32(--members_count | (u32) prot << 30); mlx4_qp_detach_common()
1263 if (prot == MLX4_PROT_ETH) mlx4_qp_detach_common()
1266 if (members_count && (prot != MLX4_PROT_ETH || !removed_entry)) { mlx4_qp_detach_common()
1272 mgm->members_count = cpu_to_be32((u32) prot << 30); mlx4_qp_detach_common()
1329 enum mlx4_protocol prot) mlx4_QP_ATTACH()
1344 qpn |= (prot << 28); mlx4_QP_ATTACH()
1362 enum mlx4_protocol prot, u64 *reg_id) mlx4_trans_to_dmfs_attach()
1379 switch (prot) { mlx4_trans_to_dmfs_attach()
1401 enum mlx4_protocol prot, u64 *reg_id) mlx4_multicast_attach()
1405 if (prot == MLX4_PROT_ETH) mlx4_multicast_attach()
1409 if (prot == MLX4_PROT_ETH) mlx4_multicast_attach()
1414 block_mcast_loopback, prot); mlx4_multicast_attach()
1416 block_mcast_loopback, prot, mlx4_multicast_attach()
1422 prot, reg_id); mlx4_multicast_attach()
1430 enum mlx4_protocol prot, u64 reg_id) mlx4_multicast_detach()
1434 if (prot == MLX4_PROT_ETH) mlx4_multicast_detach()
1438 if (prot == MLX4_PROT_ETH) mlx4_multicast_detach()
1442 return mlx4_QP_ATTACH(dev, qp, gid, 0, 0, prot); mlx4_multicast_detach()
1444 return mlx4_qp_detach_common(dev, qp, gid, prot, mlx4_multicast_detach()
1516 int block_mcast_loopback, enum mlx4_protocol prot) mlx4_unicast_attach()
1518 if (prot == MLX4_PROT_ETH) mlx4_unicast_attach()
1523 block_mcast_loopback, prot); mlx4_unicast_attach()
1526 prot, MLX4_UC_STEER); mlx4_unicast_attach()
1531 u8 gid[16], enum mlx4_protocol prot) mlx4_unicast_detach()
1533 if (prot == MLX4_PROT_ETH) mlx4_unicast_detach()
1537 return mlx4_QP_ATTACH(dev, qp, gid, 0, 0, prot); mlx4_unicast_detach()
1539 return mlx4_qp_detach_common(dev, qp, gid, prot, MLX4_UC_STEER); mlx4_unicast_detach()
695 find_entry(struct mlx4_dev *dev, u8 port, u8 *gid, enum mlx4_protocol prot, struct mlx4_cmd_mailbox *mgm_mailbox, int *prev, int *index) find_entry() argument
1097 mlx4_qp_attach_common(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], int block_mcast_loopback, enum mlx4_protocol prot, enum mlx4_steer_type steer) mlx4_qp_attach_common() argument
1206 mlx4_qp_detach_common(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], enum mlx4_protocol prot, enum mlx4_steer_type steer) mlx4_qp_detach_common() argument
1327 mlx4_QP_ATTACH(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], u8 attach, u8 block_loopback, enum mlx4_protocol prot) mlx4_QP_ATTACH() argument
1359 mlx4_trans_to_dmfs_attach(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], u8 port, int block_mcast_loopback, enum mlx4_protocol prot, u64 *reg_id) mlx4_trans_to_dmfs_attach() argument
1399 mlx4_multicast_attach(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], u8 port, int block_mcast_loopback, enum mlx4_protocol prot, u64 *reg_id) mlx4_multicast_attach() argument
1429 mlx4_multicast_detach(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], enum mlx4_protocol prot, u64 reg_id) mlx4_multicast_detach() argument
1514 mlx4_unicast_attach(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], int block_mcast_loopback, enum mlx4_protocol prot) mlx4_unicast_attach() argument
1530 mlx4_unicast_detach(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16], enum mlx4_protocol prot) mlx4_unicast_detach() argument
/linux-4.1.27/arch/tile/include/asm/
H A Dhighmem.h65 void *kmap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot);
67 void *kmap_atomic_prot(struct page *page, pgprot_t prot);
H A Dmmu_context.h37 static inline void __install_page_table(pgd_t *pgdir, int asid, pgprot_t prot) __install_page_table() argument
40 int rc = hv_install_context(__pa(pgdir), prot, asid, __install_page_table()
H A Dpgtable.h275 extern pgprot_t set_remote_cache_cpu(pgprot_t prot, int cpu);
276 extern int get_remote_cache_cpu(pgprot_t prot);
278 static inline pte_t pfn_pte(unsigned long pfn, pgprot_t prot) pfn_pte() argument
280 return hv_pte_set_pa(prot, PFN_PHYS(pfn)); pfn_pte()
433 static inline pmd_t ptfn_pmd(unsigned long ptfn, pgprot_t prot) ptfn_pmd() argument
435 return pte_pmd(hv_pte_set_ptfn(prot, ptfn)); ptfn_pmd()
/linux-4.1.27/arch/microblaze/mm/
H A Dhighmem.c34 void *kmap_atomic_prot(struct page *page, pgprot_t prot) kmap_atomic_prot() argument
52 set_pte_at(&init_mm, vaddr, kmap_pte-idx, mk_pte(page, prot)); kmap_atomic_prot()
/linux-4.1.27/arch/score/include/asm/
H A Dpgtable.h67 #define pfn_pte(pfn, prot) \
68 __pte(((unsigned long long)(pfn) << PAGE_SHIFT) | pgprot_val(prot))
99 #define mk_pte(page, prot) pfn_pte(page_to_pfn(page), prot) pte_mkspecial()
155 unsigned long prot = pgprot_val(_prot); pgprot_noncached() local
157 prot = (prot & ~_CACHE_MASK); pgprot_noncached()
159 return __pgprot(prot); pgprot_noncached()
/linux-4.1.27/arch/mips/mm/
H A Dpgtable-64.c79 pmd_t mk_pmd(struct page *page, pgprot_t prot) mk_pmd() argument
83 pmd_val(pmd) = (page_to_pfn(page) << _PFN_SHIFT) | pgprot_val(prot); mk_pmd()
/linux-4.1.27/arch/arm64/kernel/
H A Dsys.c30 unsigned long prot, unsigned long flags, sys_mmap()
36 return sys_mmap_pgoff(addr, len, prot, flags, fd, off >> PAGE_SHIFT); sys_mmap()
29 sys_mmap(unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long fd, off_t off) sys_mmap() argument
H A Defi.c238 pgprot_t prot; efi_virtmap_init() local
259 prot = __pgprot(PROT_DEVICE_nGnRE); efi_virtmap_init()
262 prot = PAGE_KERNEL_EXEC; efi_virtmap_init()
264 prot = PAGE_KERNEL; efi_virtmap_init()
266 create_pgd_mapping(&efi_mm, paddr, md->virt_addr, size, prot); efi_virtmap_init()
/linux-4.1.27/arch/frv/kernel/
H A Dsys_frv.c31 unsigned long prot, unsigned long flags, sys_mmap2()
42 return sys_mmap_pgoff(addr, len, prot, flags, fd, sys_mmap2()
30 sys_mmap2(unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long fd, unsigned long pgoff) sys_mmap2() argument
/linux-4.1.27/arch/hexagon/mm/
H A Dioremap.c31 pgprot_t prot = __pgprot(_PAGE_PRESENT|_PAGE_READ|_PAGE_WRITE ioremap_nocache() local
46 if (ioremap_page_range(addr, addr+size, phys_addr, prot)) { ioremap_nocache()
/linux-4.1.27/include/linux/can/
H A Dcore.h34 * @prot: pointer to struct proto structure.
40 struct proto *prot; member in struct:can_proto
/linux-4.1.27/arch/um/os-Linux/
H A Dprocess.c136 int prot; os_map_memory() local
138 prot = (r ? PROT_READ : 0) | (w ? PROT_WRITE : 0) | os_map_memory()
141 loc = mmap64((void *) virt, len, prot, MAP_SHARED | MAP_FIXED, os_map_memory()
150 int prot = ((r ? PROT_READ : 0) | (w ? PROT_WRITE : 0) | os_protect_memory() local
153 if (mprotect(addr, len, prot) < 0) os_protect_memory()
/linux-4.1.27/drivers/xen/
H A Dxlate_mmu.c66 pgprot_t prot; member in struct:remap_data
82 pte_t pte = pte_mkspecial(pfn_pte(pfn, info->prot)); remap_pte_fn()
99 int *err_ptr, pgprot_t prot, xen_xlate_remap_gfn_array()
112 data.prot = prot; xen_xlate_remap_gfn_array()
96 xen_xlate_remap_gfn_array(struct vm_area_struct *vma, unsigned long addr, xen_pfn_t *mfn, int nr, int *err_ptr, pgprot_t prot, unsigned domid, struct page **pages) xen_xlate_remap_gfn_array() argument
/linux-4.1.27/arch/tile/mm/
H A Dinit.c210 static pgprot_t __init construct_pgprot(pgprot_t prot, int home) construct_pgprot() argument
212 prot = pte_set_home(prot, home); construct_pgprot()
215 prot = hv_pte_set_mode(prot, HV_PTE_MODE_CACHE_HASH_L3); construct_pgprot()
217 prot = hv_pte_set_mode(prot, HV_PTE_MODE_CACHE_NO_L3); construct_pgprot()
219 return prot; construct_pgprot()
378 static inline pgprot_t ktext_set_nocache(pgprot_t prot) ktext_set_nocache() argument
381 prot = hv_pte_set_nc(prot); ktext_set_nocache()
383 prot = hv_pte_set_no_alloc_l2(prot); ktext_set_nocache()
384 return prot; ktext_set_nocache()
464 pgprot_t prot = init_pgprot(address); kernel_physical_mapping_init() local
465 *(pte_t *)pmd = pte_mkhuge(pfn_pte(pfn, prot)); kernel_physical_mapping_init()
468 pte[pte_ofs] = pfn_pte(pfn, prot); kernel_physical_mapping_init()
475 pgprot_t prot = init_pgprot(address); kernel_physical_mapping_init() local
476 pte[pte_ofs] = pfn_pte(pfn, prot); kernel_physical_mapping_init()
513 pgprot_t prot = construct_pgprot(PAGE_KERNEL_EXEC, kernel_physical_mapping_init() local
518 prot = hv_pte_set_mode(prot, kernel_physical_mapping_init()
521 prot = hv_pte_set_mode(prot, kernel_physical_mapping_init()
524 prot = hv_pte_set_mode(prot, kernel_physical_mapping_init()
528 prot = ktext_set_nocache(prot); kernel_physical_mapping_init()
542 prot = set_remote_cache_cpu(prot, cpu); kernel_physical_mapping_init()
547 pte[pte_ofs] = pfn_pte(pfn, prot); kernel_physical_mapping_init()
H A Dhighmem.c198 void *kmap_atomic_prot(struct page *page, pgprot_t prot) kmap_atomic_prot() argument
208 BUG_ON(pte_exec(prot)); kmap_atomic_prot()
220 kmap_atomic_register(page, type, vaddr, pte, mk_pte(page, prot)); kmap_atomic_prot()
273 void *kmap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot) kmap_atomic_prot_pfn() argument
275 return kmap_atomic_prot(pfn_to_page(pfn), prot); kmap_atomic_prot_pfn()
H A Dpgtable.c371 pgprot_t set_remote_cache_cpu(pgprot_t prot, int cpu) set_remote_cache_cpu() argument
377 BUG_ON(hv_pte_get_mode(prot) != HV_PTE_MODE_CACHE_TILE_L3); set_remote_cache_cpu()
380 return hv_pte_set_lotar(prot, HV_XY_TO_LOTAR(x, y)); set_remote_cache_cpu()
383 int get_remote_cache_cpu(pgprot_t prot) get_remote_cache_cpu() argument
385 HV_LOTAR lotar = hv_pte_get_lotar(prot); get_remote_cache_cpu()
388 BUG_ON(hv_pte_get_mode(prot) != HV_PTE_MODE_CACHE_TILE_L3); get_remote_cache_cpu()
/linux-4.1.27/net/core/
H A Dsock.c1319 static struct sock *sk_prot_alloc(struct proto *prot, gfp_t priority, sk_prot_alloc() argument
1325 slab = prot->slab; sk_prot_alloc()
1331 if (prot->clear_sk) sk_prot_alloc()
1332 prot->clear_sk(sk, prot->obj_size); sk_prot_alloc()
1334 sk_prot_clear_nulls(sk, prot->obj_size); sk_prot_alloc()
1337 sk = kmalloc(prot->obj_size, priority); sk_prot_alloc()
1345 if (!try_module_get(prot->owner)) sk_prot_alloc()
1362 static void sk_prot_free(struct proto *prot, struct sock *sk) sk_prot_free() argument
1367 owner = prot->owner; sk_prot_free()
1368 slab = prot->slab; sk_prot_free()
1394 * @prot: struct proto associated with this new sock instance
1397 struct proto *prot) sk_alloc()
1401 sk = sk_prot_alloc(prot, priority | __GFP_ZERO, family); sk_alloc()
1408 sk->sk_prot = sk->sk_prot_creator = prot; sk_alloc()
2003 struct proto *prot = sk->sk_prot; __sk_mem_schedule() local
2031 if (atomic_read(&sk->sk_rmem_alloc) < prot->sysctl_rmem[0]) __sk_mem_schedule()
2036 if (sk->sk_wmem_queued < prot->sysctl_wmem[0]) __sk_mem_schedule()
2039 prot->sysctl_wmem[0]) __sk_mem_schedule()
2068 trace_sock_exceed_buf_limit(sk, prot, allocated); __sk_mem_schedule()
2644 void sock_prot_inuse_add(struct net *net, struct proto *prot, int val) sock_prot_inuse_add() argument
2646 __this_cpu_add(net->core.inuse->val[prot->inuse_idx], val); sock_prot_inuse_add()
2650 int sock_prot_inuse_get(struct net *net, struct proto *prot) sock_prot_inuse_get() argument
2652 int cpu, idx = prot->inuse_idx; sock_prot_inuse_get()
2690 void sock_prot_inuse_add(struct net *net, struct proto *prot, int val) sock_prot_inuse_add() argument
2692 __this_cpu_add(prot_inuse.val[prot->inuse_idx], val); sock_prot_inuse_add()
2696 int sock_prot_inuse_get(struct net *net, struct proto *prot) sock_prot_inuse_get() argument
2698 int cpu, idx = prot->inuse_idx; sock_prot_inuse_get()
2709 static void assign_proto_idx(struct proto *prot) assign_proto_idx() argument
2711 prot->inuse_idx = find_first_zero_bit(proto_inuse_idx, PROTO_INUSE_NR); assign_proto_idx()
2713 if (unlikely(prot->inuse_idx == PROTO_INUSE_NR - 1)) { assign_proto_idx()
2718 set_bit(prot->inuse_idx, proto_inuse_idx); assign_proto_idx()
2721 static void release_proto_idx(struct proto *prot) release_proto_idx() argument
2723 if (prot->inuse_idx != PROTO_INUSE_NR - 1) release_proto_idx()
2724 clear_bit(prot->inuse_idx, proto_inuse_idx); release_proto_idx()
2727 static inline void assign_proto_idx(struct proto *prot) assign_proto_idx() argument
2731 static inline void release_proto_idx(struct proto *prot) release_proto_idx() argument
2748 static int req_prot_init(const struct proto *prot) req_prot_init() argument
2750 struct request_sock_ops *rsk_prot = prot->rsk_prot; req_prot_init()
2756 prot->name); req_prot_init()
2766 prot->name); req_prot_init()
2772 int proto_register(struct proto *prot, int alloc_slab) proto_register() argument
2775 prot->slab = kmem_cache_create(prot->name, prot->obj_size, 0, proto_register()
2776 SLAB_HWCACHE_ALIGN | prot->slab_flags, proto_register()
2779 if (prot->slab == NULL) { proto_register()
2781 prot->name); proto_register()
2785 if (req_prot_init(prot)) proto_register()
2788 if (prot->twsk_prot != NULL) { proto_register()
2789 prot->twsk_prot->twsk_slab_name = kasprintf(GFP_KERNEL, "tw_sock_%s", prot->name); proto_register()
2791 if (prot->twsk_prot->twsk_slab_name == NULL) proto_register()
2794 prot->twsk_prot->twsk_slab = proto_register()
2795 kmem_cache_create(prot->twsk_prot->twsk_slab_name, proto_register()
2796 prot->twsk_prot->twsk_obj_size, proto_register()
2798 prot->slab_flags, proto_register()
2800 if (prot->twsk_prot->twsk_slab == NULL) proto_register()
2806 list_add(&prot->node, &proto_list); proto_register()
2807 assign_proto_idx(prot); proto_register()
2812 kfree(prot->twsk_prot->twsk_slab_name); proto_register()
2814 req_prot_cleanup(prot->rsk_prot); proto_register()
2816 kmem_cache_destroy(prot->slab); proto_register()
2817 prot->slab = NULL; proto_register()
2823 void proto_unregister(struct proto *prot) proto_unregister() argument
2826 release_proto_idx(prot); proto_unregister()
2827 list_del(&prot->node); proto_unregister()
2830 if (prot->slab != NULL) { proto_unregister()
2831 kmem_cache_destroy(prot->slab); proto_unregister()
2832 prot->slab = NULL; proto_unregister()
2835 req_prot_cleanup(prot->rsk_prot); proto_unregister()
2837 if (prot->twsk_prot != NULL && prot->twsk_prot->twsk_slab != NULL) { proto_unregister()
2838 kmem_cache_destroy(prot->twsk_prot->twsk_slab); proto_unregister()
2839 kfree(prot->twsk_prot->twsk_slab_name); proto_unregister()
2840 prot->twsk_prot->twsk_slab = NULL; proto_unregister()
1396 sk_alloc(struct net *net, int family, gfp_t priority, struct proto *prot) sk_alloc() argument
/linux-4.1.27/drivers/iommu/
H A Dio-pgtable-arm.c209 arm_lpae_iopte prot, int lvl, arm_lpae_init_pte()
212 arm_lpae_iopte pte = prot; arm_lpae_init_pte()
248 phys_addr_t paddr, size_t size, arm_lpae_iopte prot, __arm_lpae_map()
260 return arm_lpae_init_pte(data, iova, paddr, prot, lvl, ptep); __arm_lpae_map()
286 return __arm_lpae_map(data, iova, paddr, size, prot, lvl + 1, cptep); __arm_lpae_map()
290 int prot) arm_lpae_prot_to_pte()
298 if (!(prot & IOMMU_WRITE) && (prot & IOMMU_READ)) arm_lpae_prot_to_pte()
301 if (prot & IOMMU_CACHE) arm_lpae_prot_to_pte()
306 if (prot & IOMMU_READ) arm_lpae_prot_to_pte()
308 if (prot & IOMMU_WRITE) arm_lpae_prot_to_pte()
310 if (prot & IOMMU_CACHE) arm_lpae_prot_to_pte()
316 if (prot & IOMMU_NOEXEC) arm_lpae_prot_to_pte()
328 arm_lpae_iopte prot; arm_lpae_map() local
334 prot = arm_lpae_prot_to_pte(data, iommu_prot); arm_lpae_map()
335 return __arm_lpae_map(data, iova, paddr, size, prot, lvl, ptep); arm_lpae_map()
379 arm_lpae_iopte prot, int lvl, arm_lpae_split_blk_unmap()
401 if (__arm_lpae_map(data, blk_start, blk_paddr, size, prot, lvl, arm_lpae_split_blk_unmap()
207 arm_lpae_init_pte(struct arm_lpae_io_pgtable *data, unsigned long iova, phys_addr_t paddr, arm_lpae_iopte prot, int lvl, arm_lpae_iopte *ptep) arm_lpae_init_pte() argument
247 __arm_lpae_map(struct arm_lpae_io_pgtable *data, unsigned long iova, phys_addr_t paddr, size_t size, arm_lpae_iopte prot, int lvl, arm_lpae_iopte *ptep) __arm_lpae_map() argument
289 arm_lpae_prot_to_pte(struct arm_lpae_io_pgtable *data, int prot) arm_lpae_prot_to_pte() argument
377 arm_lpae_split_blk_unmap(struct arm_lpae_io_pgtable *data, unsigned long iova, size_t size, arm_lpae_iopte prot, int lvl, arm_lpae_iopte *ptep, size_t blk_size) arm_lpae_split_blk_unmap() argument
H A Domap-iommu.c662 static int iopgd_alloc_section(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) iopgd_alloc_section() argument
672 *iopgd = (pa & IOSECTION_MASK) | prot | IOPGD_SECTION; iopgd_alloc_section()
677 static int iopgd_alloc_super(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) iopgd_alloc_super() argument
689 *(iopgd + i) = (pa & IOSUPER_MASK) | prot | IOPGD_SUPER; iopgd_alloc_super()
694 static int iopte_alloc_page(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) iopte_alloc_page() argument
702 *iopte = (pa & IOPAGE_MASK) | prot | IOPTE_SMALL; iopte_alloc_page()
711 static int iopte_alloc_large(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) iopte_alloc_large() argument
727 *(iopte + i) = (pa & IOLARGE_MASK) | prot | IOPTE_LARGE; iopte_alloc_large()
736 u32 prot; iopgtable_store_entry_core() local
761 prot = get_iopte_attr(e); iopgtable_store_entry_core()
764 err = fn(obj, e->da, e->pa, prot); iopgtable_store_entry_core()
1124 phys_addr_t pa, size_t bytes, int prot) omap_iommu_map()
1123 omap_iommu_map(struct iommu_domain *domain, unsigned long da, phys_addr_t pa, size_t bytes, int prot) omap_iommu_map() argument
H A Drockchip-iommu.c227 /* TODO: set cache flags per prot IOMMU_CACHE */ rk_mk_pte()
228 static u32 rk_mk_pte(phys_addr_t page, int prot) rk_mk_pte() argument
231 flags |= (prot & IOMMU_READ) ? RK_PTE_PAGE_READABLE : 0; rk_mk_pte()
232 flags |= (prot & IOMMU_WRITE) ? RK_PTE_PAGE_WRITABLE : 0; rk_mk_pte()
612 int prot) rk_iommu_map_iova()
626 pte_addr[pte_count] = rk_mk_pte(paddr, prot); rk_iommu_map_iova()
640 pr_err("iova: %pad already mapped to %pa cannot remap to phys: %pa prot: %#x\n", rk_iommu_map_iova()
641 &iova, &page_phys, &paddr, prot); rk_iommu_map_iova()
647 phys_addr_t paddr, size_t size, int prot) rk_iommu_map()
671 ret = rk_iommu_map_iova(rk_domain, pte_addr, iova, paddr, size, prot); rk_iommu_map()
610 rk_iommu_map_iova(struct rk_iommu_domain *rk_domain, u32 *pte_addr, dma_addr_t iova, phys_addr_t paddr, size_t size, int prot) rk_iommu_map_iova() argument
646 rk_iommu_map(struct iommu_domain *domain, unsigned long _iova, phys_addr_t paddr, size_t size, int prot) rk_iommu_map() argument
H A Dfsl_pamu_domain.h28 int prot; member in struct:dma_window
H A Dfsl_pamu_domain.c114 sub_win_ptr[i].prot); map_subwins()
140 0, wnd->prot); map_win()
173 wnd->prot); update_liodn()
187 0, wnd->prot); update_liodn()
535 phys_addr_t paddr, u64 size, int prot) fsl_pamu_window_enable()
544 if (prot & IOMMU_READ) fsl_pamu_window_enable()
546 if (prot & IOMMU_WRITE) fsl_pamu_window_enable()
588 wnd->prot = pamu_prot; fsl_pamu_window_enable()
534 fsl_pamu_window_enable(struct iommu_domain *domain, u32 wnd_nr, phys_addr_t paddr, u64 size, int prot) fsl_pamu_window_enable() argument
/linux-4.1.27/arch/mips/include/asm/
H A Dpgtable-32.h110 pfn_pte(unsigned long pfn, pgprot_t prot) pfn_pte() argument
115 (pgprot_val(prot) & ~_PFNX_MASK); pfn_pte()
117 (pgprot_val(prot) & ~_PFN_MASK); pfn_pte()
127 #define pfn_pte(pfn, prot) __pte(((pfn) << (PAGE_SHIFT + 2)) | pgprot_val(prot))
130 #define pfn_pte(pfn, prot) __pte(((unsigned long long)(pfn) << _PFN_SHIFT) | pgprot_val(prot))
H A Dpgtable.h408 unsigned long prot = pgprot_val(_prot); pgprot_noncached() local
410 prot = (prot & ~_CACHE_MASK) | _CACHE_UNCACHED; pgprot_noncached()
412 return __pgprot(prot); pgprot_noncached()
417 unsigned long prot = pgprot_val(_prot); pgprot_writecombine() local
420 prot = (prot & ~_CACHE_MASK) | cpu_data[0].writecombine; pgprot_writecombine()
422 return __pgprot(prot); pgprot_writecombine()
469 extern int remap_pfn_range(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot);
475 pgprot_t prot) io_remap_pfn_range()
478 return remap_pfn_range(vma, vaddr, phys_addr_high >> PAGE_SHIFT, size, prot); io_remap_pfn_range()
589 extern pmd_t mk_pmd(struct page *page, pgprot_t prot);
471 io_remap_pfn_range(struct vm_area_struct *vma, unsigned long vaddr, unsigned long pfn, unsigned long size, pgprot_t prot) io_remap_pfn_range() argument
H A Dpgtable-64.h229 #define pfn_pte(pfn, prot) __pte(((pfn) << (PAGE_SHIFT + 2)) | pgprot_val(prot))
232 #define pfn_pte(pfn, prot) __pte(((pfn) << _PFN_SHIFT) | pgprot_val(prot))
233 #define pfn_pmd(pfn, prot) __pmd(((pfn) << _PFN_SHIFT) | pgprot_val(prot))
/linux-4.1.27/arch/nios2/include/asm/
H A Dpgtable.h121 unsigned long prot = pgprot_val(_prot); pgprot_noncached() local
123 prot &= ~_PAGE_CACHED; pgprot_noncached()
125 return __pgprot(prot); pgprot_noncached()
198 #define pfn_pte(pfn, prot) (__pte(pfn | pgprot_val(prot)))
241 #define mk_pte(page, prot) (pfn_pte(page_to_pfn(page), prot))
/linux-4.1.27/mm/
H A Dmprotect.c336 unsigned long, prot) SYSCALL_DEFINE3()
341 const int grows = prot & (PROT_GROWSDOWN|PROT_GROWSUP); SYSCALL_DEFINE3()
342 prot &= ~(PROT_GROWSDOWN|PROT_GROWSUP); SYSCALL_DEFINE3()
354 if (!arch_validate_prot(prot)) SYSCALL_DEFINE3()
357 reqprot = prot; SYSCALL_DEFINE3()
361 if ((prot & PROT_READ) && (current->personality & READ_IMPLIES_EXEC)) SYSCALL_DEFINE3()
362 prot |= PROT_EXEC; SYSCALL_DEFINE3()
364 vm_flags = calc_vm_prot_bits(prot); SYSCALL_DEFINE3()
407 error = security_file_mprotect(vma, reqprot, prot); SYSCALL_DEFINE3()
H A Dearly_ioremap.c49 phys_addr_t phys, pgprot_t prot) __late_set_fixmap()
97 __early_ioremap(resource_size_t phys_addr, unsigned long size, pgprot_t prot) __early_ioremap() argument
145 __late_set_fixmap(idx, phys_addr, prot); __early_ioremap()
147 __early_set_fixmap(idx, phys_addr, prot); __early_ioremap()
48 __late_set_fixmap(enum fixed_addresses idx, phys_addr_t phys, pgprot_t prot) __late_set_fixmap() argument
H A Dnommu.c280 void *__vmalloc(unsigned long size, gfp_t gfp_mask, pgprot_t prot) __vmalloc() argument
467 void *vmap(struct page **pages, unsigned int count, unsigned long flags, pgprot_t prot) vmap() argument
480 void *vm_map_ram(struct page **pages, unsigned int count, int node, pgprot_t prot) vm_map_ram() argument
951 unsigned long prot, validate_mmap_request()
1028 if ((prot & PROT_WRITE) && validate_mmap_request()
1052 if (prot & PROT_WRITE) validate_mmap_request()
1057 if (((prot & PROT_READ) && !(capabilities & NOMMU_MAP_READ)) || validate_mmap_request()
1058 ((prot & PROT_WRITE) && !(capabilities & NOMMU_MAP_WRITE)) || validate_mmap_request()
1059 ((prot & PROT_EXEC) && !(capabilities & NOMMU_MAP_EXEC)) validate_mmap_request()
1073 if (prot & PROT_EXEC) validate_mmap_request()
1075 } else if ((prot & PROT_READ) && !(prot & PROT_EXEC)) { validate_mmap_request()
1079 prot |= PROT_EXEC; validate_mmap_request()
1081 } else if ((prot & PROT_READ) && validate_mmap_request()
1082 (prot & PROT_EXEC) && validate_mmap_request()
1095 if ((prot & PROT_READ) && validate_mmap_request()
1097 prot |= PROT_EXEC; validate_mmap_request()
1115 unsigned long prot, determine_vm_flags()
1121 vm_flags = calc_vm_prot_bits(prot) | calc_vm_flag_bits(flags); determine_vm_flags()
1127 if (file && !(prot & PROT_WRITE)) determine_vm_flags()
1277 unsigned long prot, do_mmap_pgoff()
1288 kenter(",%lx,%lx,%lx,%lx,%lx", addr, len, prot, flags, pgoff); do_mmap_pgoff()
1294 ret = validate_mmap_request(file, addr, len, prot, flags, pgoff, do_mmap_pgoff()
1307 vm_flags = determine_vm_flags(file, prot, flags, capabilities); do_mmap_pgoff()
1508 unsigned long, prot, unsigned long, flags, SYSCALL_DEFINE6()
1523 retval = vm_mmap_pgoff(file, addr, len, prot, flags, pgoff); SYSCALL_DEFINE6()
1535 unsigned long prot; member in struct:mmap_arg_struct
1550 return sys_mmap_pgoff(a.addr, a.len, a.prot, a.flags, a.fd, SYSCALL_DEFINE1()
1855 unsigned long pfn, unsigned long size, pgprot_t prot) remap_pfn_range()
948 validate_mmap_request(struct file *file, unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long pgoff, unsigned long *_capabilities) validate_mmap_request() argument
1114 determine_vm_flags(struct file *file, unsigned long prot, unsigned long flags, unsigned long capabilities) determine_vm_flags() argument
1274 do_mmap_pgoff(struct file *file, unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long pgoff, unsigned long *populate) do_mmap_pgoff() argument
1854 remap_pfn_range(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn, unsigned long size, pgprot_t prot) remap_pfn_range() argument
H A Dvmalloc.c118 unsigned long end, pgprot_t prot, struct page **pages, int *nr) vmap_pte_range()
137 set_pte_at(&init_mm, addr, pte, mk_pte(page, prot)); vmap_pte_range()
144 unsigned long end, pgprot_t prot, struct page **pages, int *nr) vmap_pmd_range()
154 if (vmap_pte_range(pmd, addr, next, prot, pages, nr)) vmap_pmd_range()
161 unsigned long end, pgprot_t prot, struct page **pages, int *nr) vmap_pud_range()
171 if (vmap_pmd_range(pud, addr, next, prot, pages, nr)) vmap_pud_range()
178 * Set up page tables in kva (addr, end). The ptes shall have prot "prot", and
184 pgprot_t prot, struct page **pages) vmap_page_range_noflush()
196 err = vmap_pud_range(pgd, addr, next, prot, pages, &nr); vmap_page_range_noflush()
205 pgprot_t prot, struct page **pages) vmap_page_range()
209 ret = vmap_page_range_noflush(start, end, prot, pages); vmap_page_range()
1105 * @prot: memory protection to use. PAGE_KERNEL for regular RAM
1115 void *vm_map_ram(struct page **pages, unsigned int count, int node, pgprot_t prot) vm_map_ram() argument
1136 if (vmap_page_range(addr, addr + size, prot, pages) < 0) { vm_map_ram()
1233 * @prot: page protection flags to use
1249 pgprot_t prot, struct page **pages) map_kernel_range_noflush()
1251 return vmap_page_range_noflush(addr, addr + size, prot, pages); map_kernel_range_noflush()
1292 int map_vm_area(struct vm_struct *area, pgprot_t prot, struct page **pages) map_vm_area() argument
1298 err = vmap_page_range(addr, end, prot, pages); map_vm_area()
1546 * @prot: page protection for the mapping
1552 unsigned long flags, pgprot_t prot) vmap()
1566 if (map_vm_area(area, prot, pages)) { vmap()
1576 gfp_t gfp_mask, pgprot_t prot,
1579 pgprot_t prot, int node) __vmalloc_area_node()
1624 if (map_vm_area(area, prot, pages)) __vmalloc_area_node()
1643 * @prot: protection mask for the allocated pages
1650 * kernel virtual space, using a pagetable protection of @prot.
1654 pgprot_t prot, unsigned long vm_flags, int node, __vmalloc_node_range()
1670 addr = __vmalloc_area_node(area, gfp_mask, prot, node); __vmalloc_node_range()
1702 * @prot: protection mask for the allocated pages
1708 * kernel virtual space, using a pagetable protection of @prot.
1711 gfp_t gfp_mask, pgprot_t prot, __vmalloc_node()
1715 gfp_mask, prot, 0, node, caller); __vmalloc_node()
1718 void *__vmalloc(unsigned long size, gfp_t gfp_mask, pgprot_t prot) __vmalloc() argument
1720 return __vmalloc_node(size, 1, gfp_mask, prot, NUMA_NO_NODE, __vmalloc()
117 vmap_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end, pgprot_t prot, struct page **pages, int *nr) vmap_pte_range() argument
143 vmap_pmd_range(pud_t *pud, unsigned long addr, unsigned long end, pgprot_t prot, struct page **pages, int *nr) vmap_pmd_range() argument
160 vmap_pud_range(pgd_t *pgd, unsigned long addr, unsigned long end, pgprot_t prot, struct page **pages, int *nr) vmap_pud_range() argument
183 vmap_page_range_noflush(unsigned long start, unsigned long end, pgprot_t prot, struct page **pages) vmap_page_range_noflush() argument
204 vmap_page_range(unsigned long start, unsigned long end, pgprot_t prot, struct page **pages) vmap_page_range() argument
1248 map_kernel_range_noflush(unsigned long addr, unsigned long size, pgprot_t prot, struct page **pages) map_kernel_range_noflush() argument
1551 vmap(struct page **pages, unsigned int count, unsigned long flags, pgprot_t prot) vmap() argument
1578 __vmalloc_area_node(struct vm_struct *area, gfp_t gfp_mask, pgprot_t prot, int node) __vmalloc_area_node() argument
1652 __vmalloc_node_range(unsigned long size, unsigned long align, unsigned long start, unsigned long end, gfp_t gfp_mask, pgprot_t prot, unsigned long vm_flags, int node, const void *caller) __vmalloc_node_range() argument
1710 __vmalloc_node(unsigned long size, unsigned long align, gfp_t gfp_mask, pgprot_t prot, int node, const void *caller) __vmalloc_node() argument
H A Dutil.c287 unsigned long len, unsigned long prot, vm_mmap_pgoff()
294 ret = security_mmap_file(file, prot, flag); vm_mmap_pgoff()
297 ret = do_mmap_pgoff(file, addr, len, prot, flag, pgoff, vm_mmap_pgoff()
307 unsigned long len, unsigned long prot, vm_mmap()
315 return vm_mmap_pgoff(file, addr, len, prot, flag, offset >> PAGE_SHIFT); vm_mmap()
286 vm_mmap_pgoff(struct file *file, unsigned long addr, unsigned long len, unsigned long prot, unsigned long flag, unsigned long pgoff) vm_mmap_pgoff() argument
306 vm_mmap(struct file *file, unsigned long addr, unsigned long len, unsigned long prot, unsigned long flag, unsigned long offset) vm_mmap() argument
/linux-4.1.27/include/uapi/linux/
H A Dagpgart.h77 * The "prot" down below needs still a "sleep" flag somehow ...
82 int prot; /* prot flags for mmap */ member in struct:_agp_segment
/linux-4.1.27/arch/ia64/mm/
H A Dioremap.c40 pgprot_t prot; ioremap() local
74 prot = PAGE_KERNEL; ioremap()
92 (unsigned long) addr + size, phys_addr, prot)) { ioremap()
/linux-4.1.27/arch/metag/include/asm/
H A Delf.h111 struct elf32_phdr *eppnt, int prot, int type,
116 struct elf32_phdr *eppnt, int prot, metag_elf_map()
119 return __metag_elf_map(filep, addr, eppnt, prot, type, total_size); metag_elf_map()
114 metag_elf_map(struct file *filep, unsigned long addr, struct elf32_phdr *eppnt, int prot, int type, unsigned long total_size) metag_elf_map() argument
H A Dpgtable.h106 #define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
142 #define pgprot_writecombine(prot) \ pte_mkhuge()
143 __pgprot(pgprot_val(prot) & ~(_PAGE_CACHE_CTRL1 | _PAGE_CACHE_CTRL0)) pte_mkhuge()
145 #define pgprot_noncached(prot) \ pte_mkhuge()
146 __pgprot(pgprot_val(prot) & ~_PAGE_CACHEABLE) pte_mkhuge()
/linux-4.1.27/drivers/gpu/drm/ttm/
H A Dttm_bo_util.c252 pgprot_t prot) ttm_copy_io_ttm_page()
263 dst = kmap_atomic_prot(d, prot); ttm_copy_io_ttm_page()
265 if (pgprot_val(prot) != pgprot_val(PAGE_KERNEL)) ttm_copy_io_ttm_page()
266 dst = vmap(&d, 1, 0, prot); ttm_copy_io_ttm_page()
278 if (pgprot_val(prot) != pgprot_val(PAGE_KERNEL)) ttm_copy_io_ttm_page()
289 pgprot_t prot) ttm_copy_ttm_io_page()
299 src = kmap_atomic_prot(s, prot); ttm_copy_ttm_io_page()
301 if (pgprot_val(prot) != pgprot_val(PAGE_KERNEL)) ttm_copy_ttm_io_page()
302 src = vmap(&s, 1, 0, prot); ttm_copy_ttm_io_page()
314 if (pgprot_val(prot) != pgprot_val(PAGE_KERNEL)) ttm_copy_ttm_io_page()
384 pgprot_t prot = ttm_io_prot(old_mem->placement, ttm_bo_move_memcpy() local
387 prot); ttm_bo_move_memcpy()
389 pgprot_t prot = ttm_io_prot(new_mem->placement, ttm_bo_move_memcpy() local
392 prot); ttm_bo_move_memcpy()
533 struct ttm_mem_reg *mem = &bo->mem; pgprot_t prot; ttm_bo_kmap_ttm() local
559 prot = ttm_io_prot(mem->placement, PAGE_KERNEL); ttm_bo_kmap_ttm()
562 0, prot); ttm_bo_kmap_ttm()
250 ttm_copy_io_ttm_page(struct ttm_tt *ttm, void *src, unsigned long page, pgprot_t prot) ttm_copy_io_ttm_page() argument
287 ttm_copy_ttm_io_page(struct ttm_tt *ttm, void *dst, unsigned long page, pgprot_t prot) ttm_copy_ttm_io_page() argument
/linux-4.1.27/arch/tile/kernel/
H A Dsys.c78 unsigned long, prot, unsigned long, flags, SYSCALL_DEFINE6()
84 return sys_mmap_pgoff(addr, len, prot, flags, fd, SYSCALL_DEFINE6()
90 unsigned long, prot, unsigned long, flags, SYSCALL_DEFINE6()
95 return sys_mmap_pgoff(addr, len, prot, flags, fd, SYSCALL_DEFINE6()
/linux-4.1.27/arch/um/os-Linux/skas/
H A Dmem.c170 int map(struct mm_id * mm_idp, unsigned long virt, unsigned long len, int prot, map() argument
174 unsigned long args[] = { virt, len, prot, map()
198 unsigned int prot, int done, void **data) protect()
201 unsigned long args[] = { addr, len, prot, 0, 0, 0 }; protect()
197 protect(struct mm_id * mm_idp, unsigned long addr, unsigned long len, unsigned int prot, int done, void **data) protect() argument
/linux-4.1.27/arch/sh/include/cpu-sh4/cpu/
H A Dsq.h32 const char *name, pgprot_t prot);
/linux-4.1.27/arch/mips/kernel/
H A Dsyscall.c60 unsigned long, prot, unsigned long, flags, unsigned long, SYSCALL_DEFINE6()
69 result = sys_mmap_pgoff(addr, len, prot, flags, fd, offset >> PAGE_SHIFT); SYSCALL_DEFINE6()
76 unsigned long, prot, unsigned long, flags, unsigned long, fd, SYSCALL_DEFINE6()
82 return sys_mmap_pgoff(addr, len, prot, flags, fd, pgoff >> (PAGE_SHIFT-12)); SYSCALL_DEFINE6()
H A Dlinux32.c65 unsigned long, prot, unsigned long, flags, unsigned long, fd,
73 error = sys_mmap_pgoff(addr, len, prot, flags, fd,
/linux-4.1.27/arch/openrisc/mm/
H A Dinit.c78 pgprot_t prot; map_ram() local
117 prot = PAGE_KERNEL; for_each_memblock()
119 prot = PAGE_KERNEL_RO; for_each_memblock()
121 set_pte(pte, mk_pte_phys(p, prot)); for_each_memblock()
H A Dioremap.c42 __ioremap(phys_addr_t addr, unsigned long size, pgprot_t prot) __ioremap() argument
73 if (ioremap_page_range(v, v + size, p, prot)) { __ioremap()
/linux-4.1.27/arch/cris/include/asm/
H A Dcacheflush.h30 int change_page_attr(struct page *page, int numpages, pgprot_t prot);
H A Dpgtable.h199 #define pgprot_noncached(prot) __pgprot((pgprot_val(prot) | _PAGE_NO_CACHE))
253 #define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
/linux-4.1.27/arch/ia64/kernel/
H A Dsys_ia64.c137 sys_mmap2 (unsigned long addr, unsigned long len, int prot, int flags, int fd, long pgoff) sys_mmap2() argument
139 addr = sys_mmap_pgoff(addr, len, prot, flags, fd, pgoff); sys_mmap2()
146 sys_mmap (unsigned long addr, unsigned long len, int prot, int flags, int fd, long off) sys_mmap() argument
151 addr = sys_mmap_pgoff(addr, len, prot, flags, fd, off >> PAGE_SHIFT); sys_mmap()
/linux-4.1.27/arch/powerpc/boot/
H A Dmv64x60.c283 u32 i, base, size, enables, prot = 0, snoop_bits = 0; mv64x60_config_ctlr_windows() local
304 prot |= (0x3 << (i << 1)); /* RW access */ mv64x60_config_ctlr_windows()
314 out_le32((u32 *)(bridge_base + MV64x60_ENET2MEM_ACC_PROT_0), prot); mv64x60_config_ctlr_windows()
315 out_le32((u32 *)(bridge_base + MV64x60_ENET2MEM_ACC_PROT_1), prot); mv64x60_config_ctlr_windows()
316 out_le32((u32 *)(bridge_base + MV64x60_ENET2MEM_ACC_PROT_2), prot); mv64x60_config_ctlr_windows()
317 out_le32((u32 *)(bridge_base + MV64x60_MPSC2MEM_ACC_PROT_0), prot); mv64x60_config_ctlr_windows()
318 out_le32((u32 *)(bridge_base + MV64x60_MPSC2MEM_ACC_PROT_1), prot); mv64x60_config_ctlr_windows()
319 out_le32((u32 *)(bridge_base + MV64x60_IDMA2MEM_ACC_PROT_0), prot); mv64x60_config_ctlr_windows()
320 out_le32((u32 *)(bridge_base + MV64x60_IDMA2MEM_ACC_PROT_1), prot); mv64x60_config_ctlr_windows()
321 out_le32((u32 *)(bridge_base + MV64x60_IDMA2MEM_ACC_PROT_2), prot); mv64x60_config_ctlr_windows()
322 out_le32((u32 *)(bridge_base + MV64x60_IDMA2MEM_ACC_PROT_3), prot); mv64x60_config_ctlr_windows()
/linux-4.1.27/drivers/isdn/hisax/
H A Dl3dss1.c55 retval = p->prot.dss1.last_invoke_id + 1; /* try new id */ new_invoke_id()
56 while ((i) && (p->prot.dss1.invoke_used[retval >> 3] == 0xFF)) { new_invoke_id()
57 p->prot.dss1.last_invoke_id = (retval & 0xF8) + 8; new_invoke_id()
61 while (p->prot.dss1.invoke_used[retval >> 3] & (1 << (retval & 7))) new_invoke_id()
65 p->prot.dss1.last_invoke_id = retval; new_invoke_id()
66 p->prot.dss1.invoke_used[retval >> 3] |= (1 << (retval & 7)); new_invoke_id()
78 p->prot.dss1.invoke_used[id >> 3] &= ~(1 << (id & 7)); free_invoke_id()
92 proc->prot.dss1.invoke_id = 0; dss1_new_l3_process()
93 proc->prot.dss1.remote_operation = 0; dss1_new_l3_process()
94 proc->prot.dss1.uus1_data[0] = '\0'; dss1_new_l3_process()
105 free_invoke_id(p->st, p->prot.dss1.invoke_id); dss1_release_l3_process()
119 { if ((pc->callref == -1) && (pc->prot.dss1.invoke_id == id)) l3dss1_search_dummy_proc()
143 ic.parm.dss1_io.hl_id = pc->prot.dss1.invoke_id; l3dss1_dummy_return_result()
144 ic.parm.dss1_io.ll_id = pc->prot.dss1.ll_id; l3dss1_dummy_return_result()
145 ic.parm.dss1_io.proc = pc->prot.dss1.proc; l3dss1_dummy_return_result()
149 free_invoke_id(pc->st, pc->prot.dss1.invoke_id); l3dss1_dummy_return_result()
150 pc->prot.dss1.invoke_id = 0; /* reset id */ l3dss1_dummy_return_result()
176 ic.parm.dss1_io.hl_id = pc->prot.dss1.invoke_id; l3dss1_dummy_error_return()
177 ic.parm.dss1_io.ll_id = pc->prot.dss1.ll_id; l3dss1_dummy_error_return()
178 ic.parm.dss1_io.proc = pc->prot.dss1.proc; l3dss1_dummy_error_return()
182 free_invoke_id(pc->st, pc->prot.dss1.invoke_id); l3dss1_dummy_error_return()
183 pc->prot.dss1.invoke_id = 0; /* reset id */ l3dss1_dummy_error_return()
436 if ((pc->prot.dss1.invoke_id) && (pc->prot.dss1.invoke_id == id)) l3dss1_parse_facility()
438 free_invoke_id(st, pc->prot.dss1.invoke_id); l3dss1_parse_facility()
439 pc->prot.dss1.remote_result = 0; /* success */ l3dss1_parse_facility()
440 pc->prot.dss1.invoke_id = 0; l3dss1_parse_facility()
441 pc->redir_result = pc->prot.dss1.remote_result; l3dss1_parse_facility()
481 if ((pc->prot.dss1.invoke_id) && (pc->prot.dss1.invoke_id == id)) l3dss1_parse_facility()
483 free_invoke_id(st, pc->prot.dss1.invoke_id); l3dss1_parse_facility()
484 pc->prot.dss1.remote_result = err_ret; /* result */ l3dss1_parse_facility()
485 pc->prot.dss1.invoke_id = 0; l3dss1_parse_facility()
486 pc->redir_result = pc->prot.dss1.remote_result; l3dss1_parse_facility()
935 if (pc->prot.dss1.uus1_data[0]) l3dss1_msg_with_uus()
937 *p++ = strlen(pc->prot.dss1.uus1_data) + 1; l3dss1_msg_with_uus()
939 strcpy(p, pc->prot.dss1.uus1_data); l3dss1_msg_with_uus()
940 p += strlen(pc->prot.dss1.uus1_data); l3dss1_msg_with_uus()
941 pc->prot.dss1.uus1_data[0] = '\0'; l3dss1_msg_with_uus()
956 if (!pc->prot.dss1.uus1_data[0]) l3dss1_release_req()
1777 if (pc->prot.dss1.uus1_data[0]) l3dss1_disconnect_req()
1779 *p++ = strlen(pc->prot.dss1.uus1_data) + 1; l3dss1_disconnect_req()
1781 strcpy(p, pc->prot.dss1.uus1_data); l3dss1_disconnect_req()
1782 p += strlen(pc->prot.dss1.uus1_data); l3dss1_disconnect_req()
1783 pc->prot.dss1.uus1_data[0] = '\0'; l3dss1_disconnect_req()
1897 if (!pc->prot.dss1.uus1_data[0]) l3dss1_alert_req()
2095 strcpy(pc->prot.dss1.uus1_data, pc->chan->setup.eazmsn); /* copy uus element if available */ l3dss1_redir_req()
2102 if (pc->prot.dss1.invoke_id) l3dss1_redir_req()
2103 free_invoke_id(pc->st, pc->prot.dss1.invoke_id); l3dss1_redir_req()
2105 if (!(pc->prot.dss1.invoke_id = new_invoke_id(pc->st))) l3dss1_redir_req()
2121 *p++ = pc->prot.dss1.invoke_id; /* invoke id */ l3dss1_redir_req()
2211 pc->prot.dss1.ll_id = ic->parm.dss1_io.ll_id; /* remember id */ l3dss1_cmd_global()
2212 pc->prot.dss1.proc = ic->parm.dss1_io.proc; /* and procedure */ l3dss1_cmd_global()
2222 { pc->prot.dss1.invoke_id = id; /* remember id */ l3dss1_cmd_global()
2259 ic.parm.dss1_io.hl_id = pc->prot.dss1.invoke_id; l3dss1_io_timer()
2260 ic.parm.dss1_io.ll_id = pc->prot.dss1.ll_id; l3dss1_io_timer()
2261 ic.parm.dss1_io.proc = pc->prot.dss1.proc; l3dss1_io_timer()
2265 free_invoke_id(pc->st, pc->prot.dss1.invoke_id); l3dss1_io_timer()
2266 pc->prot.dss1.invoke_id = 0; /* reset id */ l3dss1_io_timer()
3205 st->prot.dss1.last_invoke_id = 0; setstack_dss1()
3206 st->prot.dss1.invoke_used[0] = 1; /* Bit 0 must always be set to 1 */ setstack_dss1()
3209 st->prot.dss1.invoke_used[i++] = 0; setstack_dss1()
3220 st->l3.global->prot.dss1.invoke_id = 0; setstack_dss1()
H A Dl3ni1.c54 retval = p->prot.ni1.last_invoke_id + 1; /* try new id */ new_invoke_id()
55 while ((i) && (p->prot.ni1.invoke_used[retval >> 3] == 0xFF)) { new_invoke_id()
56 p->prot.ni1.last_invoke_id = (retval & 0xF8) + 8; new_invoke_id()
60 while (p->prot.ni1.invoke_used[retval >> 3] & (1 << (retval & 7))) new_invoke_id()
64 p->prot.ni1.last_invoke_id = retval; new_invoke_id()
65 p->prot.ni1.invoke_used[retval >> 3] |= (1 << (retval & 7)); new_invoke_id()
77 p->prot.ni1.invoke_used[id >> 3] &= ~(1 << (id & 7)); free_invoke_id()
91 proc->prot.ni1.invoke_id = 0; ni1_new_l3_process()
92 proc->prot.ni1.remote_operation = 0; ni1_new_l3_process()
93 proc->prot.ni1.uus1_data[0] = '\0'; ni1_new_l3_process()
104 free_invoke_id(p->st, p->prot.ni1.invoke_id); ni1_release_l3_process()
118 { if ((pc->callref == -1) && (pc->prot.ni1.invoke_id == id)) l3ni1_search_dummy_proc()
142 ic.parm.ni1_io.hl_id = pc->prot.ni1.invoke_id; l3ni1_dummy_return_result()
143 ic.parm.ni1_io.ll_id = pc->prot.ni1.ll_id; l3ni1_dummy_return_result()
144 ic.parm.ni1_io.proc = pc->prot.ni1.proc; l3ni1_dummy_return_result()
148 free_invoke_id(pc->st, pc->prot.ni1.invoke_id); l3ni1_dummy_return_result()
149 pc->prot.ni1.invoke_id = 0; /* reset id */ l3ni1_dummy_return_result()
175 ic.parm.ni1_io.hl_id = pc->prot.ni1.invoke_id; l3ni1_dummy_error_return()
176 ic.parm.ni1_io.ll_id = pc->prot.ni1.ll_id; l3ni1_dummy_error_return()
177 ic.parm.ni1_io.proc = pc->prot.ni1.proc; l3ni1_dummy_error_return()
181 free_invoke_id(pc->st, pc->prot.ni1.invoke_id); l3ni1_dummy_error_return()
182 pc->prot.ni1.invoke_id = 0; /* reset id */ l3ni1_dummy_error_return()
365 if ((pc->prot.ni1.invoke_id) && (pc->prot.ni1.invoke_id == id)) l3ni1_parse_facility()
367 free_invoke_id(st, pc->prot.ni1.invoke_id); l3ni1_parse_facility()
368 pc->prot.ni1.remote_result = 0; /* success */ l3ni1_parse_facility()
369 pc->prot.ni1.invoke_id = 0; l3ni1_parse_facility()
370 pc->redir_result = pc->prot.ni1.remote_result; l3ni1_parse_facility()
410 if ((pc->prot.ni1.invoke_id) && (pc->prot.ni1.invoke_id == id)) l3ni1_parse_facility()
412 free_invoke_id(st, pc->prot.ni1.invoke_id); l3ni1_parse_facility()
413 pc->prot.ni1.remote_result = err_ret; /* result */ l3ni1_parse_facility()
414 pc->prot.ni1.invoke_id = 0; l3ni1_parse_facility()
415 pc->redir_result = pc->prot.ni1.remote_result; l3ni1_parse_facility()
885 if (pc->prot.ni1.uus1_data[0]) l3ni1_msg_with_uus()
887 *p++ = strlen(pc->prot.ni1.uus1_data) + 1; l3ni1_msg_with_uus()
889 strcpy(p, pc->prot.ni1.uus1_data); l3ni1_msg_with_uus()
890 p += strlen(pc->prot.ni1.uus1_data); l3ni1_msg_with_uus()
891 pc->prot.ni1.uus1_data[0] = '\0'; l3ni1_msg_with_uus()
906 if (!pc->prot.ni1.uus1_data[0]) l3ni1_release_req()
1631 if (pc->prot.ni1.uus1_data[0]) l3ni1_disconnect_req()
1633 *p++ = strlen(pc->prot.ni1.uus1_data) + 1; l3ni1_disconnect_req()
1635 strcpy(p, pc->prot.ni1.uus1_data); l3ni1_disconnect_req()
1636 p += strlen(pc->prot.ni1.uus1_data); l3ni1_disconnect_req()
1637 pc->prot.ni1.uus1_data[0] = '\0'; l3ni1_disconnect_req()
1753 if (!pc->prot.ni1.uus1_data[0]) l3ni1_alert_req()
1951 strcpy(pc->prot.ni1.uus1_data, pc->chan->setup.eazmsn); /* copy uus element if available */ l3ni1_redir_req()
1958 if (pc->prot.ni1.invoke_id) l3ni1_redir_req()
1959 free_invoke_id(pc->st, pc->prot.ni1.invoke_id); l3ni1_redir_req()
1961 if (!(pc->prot.ni1.invoke_id = new_invoke_id(pc->st))) l3ni1_redir_req()
1977 *p++ = pc->prot.ni1.invoke_id; /* invoke id */ l3ni1_redir_req()
2069 pc->prot.ni1.ll_id = ic->parm.ni1_io.ll_id; l3ni1_cmd_global()
2071 pc->prot.ni1.proc = ic->parm.ni1_io.proc; l3ni1_cmd_global()
2082 { pc->prot.ni1.invoke_id = id; /* remember id */ l3ni1_cmd_global()
2119 ic.parm.ni1_io.hl_id = pc->prot.ni1.invoke_id; l3ni1_io_timer()
2120 ic.parm.ni1_io.ll_id = pc->prot.ni1.ll_id; l3ni1_io_timer()
2121 ic.parm.ni1_io.proc = pc->prot.ni1.proc; l3ni1_io_timer()
2125 free_invoke_id(pc->st, pc->prot.ni1.invoke_id); l3ni1_io_timer()
2126 pc->prot.ni1.invoke_id = 0; /* reset id */ l3ni1_io_timer()
3161 st->prot.ni1.last_invoke_id = 0; setstack_ni1()
3162 st->prot.ni1.invoke_used[0] = 1; /* Bit 0 must always be set to 1 */ setstack_ni1()
3165 st->prot.ni1.invoke_used[i++] = 0; setstack_ni1()
3176 st->l3.global->prot.ni1.invoke_id = 0; setstack_ni1()
/linux-4.1.27/arch/s390/kernel/
H A Dsys_s390.c42 unsigned long prot; member in struct:s390_mmap_arg_struct
55 error = sys_mmap_pgoff(a.addr, a.len, a.prot, a.flags, a.fd, a.offset); SYSCALL_DEFINE1()
H A Dcrash_dump.c188 unsigned long size, pgprot_t prot) remap_oldmem_pfn_range_kdump()
197 size_old, prot); remap_oldmem_pfn_range_kdump()
204 return remap_pfn_range(vma, from, pfn, size, prot); remap_oldmem_pfn_range_kdump()
216 unsigned long size, pgprot_t prot) remap_oldmem_pfn_range_zfcpdump()
229 return remap_pfn_range(vma, from, pfn, size, prot); remap_oldmem_pfn_range_zfcpdump()
236 unsigned long pfn, unsigned long size, pgprot_t prot) remap_oldmem_pfn_range()
239 return remap_oldmem_pfn_range_kdump(vma, from, pfn, size, prot); remap_oldmem_pfn_range()
242 prot); remap_oldmem_pfn_range()
186 remap_oldmem_pfn_range_kdump(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) remap_oldmem_pfn_range_kdump() argument
213 remap_oldmem_pfn_range_zfcpdump(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) remap_oldmem_pfn_range_zfcpdump() argument
235 remap_oldmem_pfn_range(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) remap_oldmem_pfn_range() argument
/linux-4.1.27/arch/blackfin/include/asm/
H A Dpgtable.h38 #define pgprot_noncached(prot) (prot)
/linux-4.1.27/arch/c6x/include/asm/
H A Dpgtable.h43 #define pgprot_noncached(prot) (prot)
/linux-4.1.27/arch/cris/mm/
H A Dioremap.c30 void __iomem * __ioremap_prot(unsigned long phys_addr, unsigned long size, pgprot_t prot) __ioremap_prot() argument
56 phys_addr, prot)) { __ioremap_prot()
/linux-4.1.27/tools/perf/util/
H A Devent.c240 char prot[5]; perf_event__synthesize_mmap_events() local
255 &event->mmap2.start, &event->mmap2.len, prot, perf_event__synthesize_mmap_events()
277 event->mmap2.prot = 0; perf_event__synthesize_mmap_events()
279 if (prot[0] == 'r') perf_event__synthesize_mmap_events()
280 event->mmap2.prot |= PROT_READ; perf_event__synthesize_mmap_events()
281 if (prot[1] == 'w') perf_event__synthesize_mmap_events()
282 event->mmap2.prot |= PROT_WRITE; perf_event__synthesize_mmap_events()
283 if (prot[2] == 'x') perf_event__synthesize_mmap_events()
284 event->mmap2.prot |= PROT_EXEC; perf_event__synthesize_mmap_events()
286 if (prot[3] == 's') perf_event__synthesize_mmap_events()
291 if (prot[2] != 'x') { perf_event__synthesize_mmap_events()
292 if (!mmap_data || prot[0] != 'r') perf_event__synthesize_mmap_events()
712 (event->mmap2.prot & PROT_READ) ? 'r' : '-', perf_event__fprintf_mmap2()
713 (event->mmap2.prot & PROT_WRITE) ? 'w' : '-', perf_event__fprintf_mmap2()
714 (event->mmap2.prot & PROT_EXEC) ? 'x' : '-', perf_event__fprintf_mmap2()
/linux-4.1.27/drivers/infiniband/ulp/iser/
H A Diser_initiator.c46 * os stored in task->prot[ISER_DIR_IN].data_len
66 struct iser_data_buf *pbuf_in = &iser_task->prot[ISER_DIR_IN]; iser_prepare_read_cmd()
97 * is stored at task->prot[ISER_DIR_OUT].data_len
121 struct iser_data_buf *pbuf_out = &iser_task->prot[ISER_DIR_OUT]; iser_prepare_write_cmd()
397 prot_buf = &iser_task->prot[ISER_DIR_IN]; iser_send_command()
400 prot_buf = &iser_task->prot[ISER_DIR_OUT]; iser_send_command()
660 iser_task->prot[ISER_DIR_IN].data_len = 0; iser_task_rdma_init()
661 iser_task->prot[ISER_DIR_OUT].data_len = 0; iser_task_rdma_init()
693 if (iser_task->prot[ISER_DIR_IN].orig_sg) { iser_task_rdma_finalize()
696 &iser_task->prot[ISER_DIR_IN], iser_task_rdma_finalize()
700 if (iser_task->prot[ISER_DIR_OUT].orig_sg) { iser_task_rdma_finalize()
703 &iser_task->prot[ISER_DIR_OUT], iser_task_rdma_finalize()
715 &iser_task->prot[ISER_DIR_IN], iser_task_rdma_finalize()
727 &iser_task->prot[ISER_DIR_OUT], iser_task_rdma_finalize()
/linux-4.1.27/arch/sparc/include/asm/
H A Dpgtable_32.h262 #define pfn_pte(pfn, prot) mk_pte(pfn_to_page(pfn), prot)
299 static inline pgprot_t pgprot_noncached(pgprot_t prot) pgprot_noncached() argument
301 prot &= ~__pgprot(SRMMU_CACHE); pgprot_noncached()
302 return prot; pgprot_noncached()
414 unsigned long size, pgprot_t prot) io_remap_pfn_range()
422 return remap_pfn_range(vma, from, phys_base >> PAGE_SHIFT, size, prot); io_remap_pfn_range()
412 io_remap_pfn_range(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) io_remap_pfn_range() argument
H A Dpgtable_64.h237 static inline pte_t pfn_pte(unsigned long pfn, pgprot_t prot) pfn_pte() argument
242 return __pte(paddr | pgprot_val(prot)); pfn_pte()
278 static inline pte_t pte_modify(pte_t pte, pgprot_t prot) pte_modify() argument
333 return __pte((pte_val(pte) & mask) | (pgprot_val(prot) & ~mask)); pte_modify()
347 static inline pgprot_t pgprot_noncached(pgprot_t prot) pgprot_noncached() argument
349 unsigned long val = pgprot_val(prot); pgprot_noncached()
907 #define move_pte(pte, prot, old_addr, new_addr) \
975 unsigned long size, pgprot_t prot) io_remap_pfn_range()
983 return remap_pfn_range(vma, from, phys_base >> PAGE_SHIFT, size, prot); io_remap_pfn_range()
973 io_remap_pfn_range(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) io_remap_pfn_range() argument
/linux-4.1.27/arch/mips/pci/
H A Dpci.c328 unsigned long prot; pci_mmap_page_range() local
341 prot = pgprot_val(vma->vm_page_prot); pci_mmap_page_range()
342 prot = (prot & ~_CACHE_MASK) | _CACHE_UNCACHED; pci_mmap_page_range()
343 vma->vm_page_prot = __pgprot(prot); pci_mmap_page_range()
/linux-4.1.27/drivers/char/agp/
H A Dcompat_ioctl.h56 * The "prot" down below needs still a "sleep" flag somehow ...
61 compat_int_t prot; /* prot flags for mmap */ member in struct:agp_segment32
/linux-4.1.27/drivers/gpu/drm/msm/
H A Dmsm_iommu.c47 struct sg_table *sgt, unsigned len, int prot) msm_iommu_map()
65 ret = iommu_map(domain, da, pa, bytes, prot); msm_iommu_map()
46 msm_iommu_map(struct msm_mmu *mmu, uint32_t iova, struct sg_table *sgt, unsigned len, int prot) msm_iommu_map() argument
H A Dmsm_mmu.h27 unsigned len, int prot);
/linux-4.1.27/arch/mn10300/include/asm/
H A Dpgtable.h318 #define pfn_pte(pfn, prot) __pte(__pfn_addr(pfn) | pgprot_val(prot))
319 #define pfn_pmd(pfn, prot) __pmd(__pfn_addr(pfn) | pgprot_val(prot))
378 #define pgprot_noncached(prot) __pgprot(pgprot_val(prot) & ~_PAGE_CACHE)
384 #define pgprot_through(prot) __pgprot(pgprot_val(prot) | _PAGE_CACHE_WT)
/linux-4.1.27/drivers/base/
H A Ddma-mapping.c279 unsigned long vm_flags, pgprot_t prot, dma_common_pages_remap()
290 if (map_vm_area(area, prot, pages)) { dma_common_pages_remap()
305 pgprot_t prot, const void *caller) dma_common_contiguous_remap()
319 ptr = dma_common_pages_remap(pages, size, vm_flags, prot, caller); dma_common_contiguous_remap()
278 dma_common_pages_remap(struct page **pages, size_t size, unsigned long vm_flags, pgprot_t prot, const void *caller) dma_common_pages_remap() argument
303 dma_common_contiguous_remap(struct page *page, size_t size, unsigned long vm_flags, pgprot_t prot, const void *caller) dma_common_contiguous_remap() argument
/linux-4.1.27/include/linux/sunrpc/
H A Dsvcsock.h64 struct svc_xprt *svc_sock_create(struct svc_serv *serv, int prot);
/linux-4.1.27/arch/mn10300/unit-asb2305/
H A Dpci-asb2305.c219 unsigned long prot; pci_mmap_page_range() local
226 prot = pgprot_val(vma->vm_page_prot); pci_mmap_page_range()
227 prot &= ~_PAGE_CACHE; pci_mmap_page_range()
228 vma->vm_page_prot = __pgprot(prot); pci_mmap_page_range()
/linux-4.1.27/arch/openrisc/include/asm/
H A Dio.h36 pgprot_t prot);
/linux-4.1.27/arch/x86/pci/
H A Di386.c417 unsigned long prot; pci_mmap_page_range() local
425 prot = pgprot_val(vma->vm_page_prot); pci_mmap_page_range()
436 prot |= cachemode2protval(_PAGE_CACHE_MODE_WC); pci_mmap_page_range()
443 prot |= cachemode2protval(_PAGE_CACHE_MODE_UC_MINUS); pci_mmap_page_range()
445 vma->vm_page_prot = __pgprot(prot); pci_mmap_page_range()
/linux-4.1.27/arch/xtensa/kernel/
H A Dpci.c344 int prot = pgprot_val(vma->vm_page_prot); __pci_mmap_set_pgprot() local
347 prot = (prot & _PAGE_CA_MASK) | _PAGE_CA_WT; __pci_mmap_set_pgprot()
350 prot |= _PAGE_WRITETHRU; __pci_mmap_set_pgprot()
352 vma->vm_page_prot = __pgprot(prot); __pci_mmap_set_pgprot()
/linux-4.1.27/arch/sh/kernel/cpu/sh4/
H A Dsq.c104 static int __sq_remap(struct sq_mapping *map, pgprot_t prot) __sq_remap() argument
117 vma->phys_addr, prot)) { __sq_remap()
139 * @prot: Protection bits.
146 const char *name, pgprot_t prot) sq_remap()
181 ret = __sq_remap(map, prot); sq_remap()
145 sq_remap(unsigned long phys, unsigned int size, const char *name, pgprot_t prot) sq_remap() argument
/linux-4.1.27/security/
H A Dsecurity.c735 static inline unsigned long mmap_prot(struct file *file, unsigned long prot) mmap_prot() argument
741 if ((prot & (PROT_READ | PROT_EXEC)) != PROT_READ) mmap_prot()
742 return prot; mmap_prot()
744 return prot; mmap_prot()
749 return prot | PROT_EXEC; mmap_prot()
759 return prot; mmap_prot()
762 return prot | PROT_EXEC; mmap_prot()
765 return prot; mmap_prot()
768 int security_mmap_file(struct file *file, unsigned long prot, security_mmap_file() argument
772 ret = security_ops->mmap_file(file, prot, security_mmap_file()
773 mmap_prot(file, prot), flags); security_mmap_file()
776 return ima_file_mmap(file, prot); security_mmap_file()
785 unsigned long prot) security_file_mprotect()
787 return security_ops->file_mprotect(vma, reqprot, prot); security_file_mprotect()
784 security_file_mprotect(struct vm_area_struct *vma, unsigned long reqprot, unsigned long prot) security_file_mprotect() argument
/linux-4.1.27/arch/xtensa/include/asm/
H A Dpgtable.h276 #define pgprot_noncached(prot) (__pgprot(pgprot_val(prot) & ~_PAGE_CA_MASK))
286 #define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
287 #define mk_pte(page, prot) pfn_pte(page_to_pfn(page), prot)
/linux-4.1.27/arch/metag/kernel/
H A Dsys_metag.c41 unsigned long prot, unsigned long flags, sys_mmap2()
50 return sys_mmap_pgoff(addr, len, prot, flags, fd, pgoff); sys_mmap2()
40 sys_mmap2(unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, unsigned long fd, unsigned long pgoff) sys_mmap2() argument
H A Dprocess.c379 struct elf_phdr *eppnt, int prot, int type, __metag_elf_map()
410 map_addr = vm_mmap(filep, addr, total_size, prot, type, off); __metag_elf_map()
414 map_addr = vm_mmap(filep, addr, size, prot, type, off); __metag_elf_map()
378 __metag_elf_map(struct file *filep, unsigned long addr, struct elf_phdr *eppnt, int prot, int type, unsigned long total_size) __metag_elf_map() argument
/linux-4.1.27/arch/microblaze/include/asm/
H A Dpgtable.h51 #define pgprot_noncached_wc(prot) prot
104 #define pgprot_noncached(prot) \ pte_mkspecial()
105 (__pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | \ pte_mkspecial()
108 #define pgprot_noncached_wc(prot) \ pte_mkspecial()
109 (__pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | \ pte_mkspecial()
312 #define pfn_pte(pfn, prot) \
313 __pte(((pte_basic_t)(pfn) << PFN_SHIFT_OFFSET) | pgprot_val(prot))
H A Dhighmem.h55 extern void *kmap_atomic_prot(struct page *page, pgprot_t prot);
/linux-4.1.27/arch/frv/mm/
H A Ddma-alloc.c50 static int map_page(unsigned long va, unsigned long pa, pgprot_t prot) map_page() argument
67 set_pte(pte, mk_pte_phys(pa & PAGE_MASK, prot)); map_page()
/linux-4.1.27/arch/ia64/pci/
H A Dpci.c594 pgprot_t prot; pci_mmap_page_range() local
612 prot = phys_mem_access_prot(NULL, vma->vm_pgoff, size, pci_mmap_page_range()
621 ((pgprot_val(prot) & _PAGE_MA_MASK) == _PAGE_MA_UC || pci_mmap_page_range()
622 (pgprot_val(prot) & _PAGE_MA_MASK) == _PAGE_MA_WC) && pci_mmap_page_range()
626 vma->vm_page_prot = prot; pci_mmap_page_range()
665 pgprot_t prot; pci_mmap_legacy_page_range() local
678 prot = phys_mem_access_prot(NULL, vma->vm_pgoff, size, pci_mmap_legacy_page_range()
686 vma->vm_page_prot = prot; pci_mmap_legacy_page_range()
/linux-4.1.27/security/apparmor/
H A Dlsm.c469 static int common_mmap(int op, struct file *file, unsigned long prot, common_mmap() argument
477 if (prot & PROT_READ) common_mmap()
483 if ((prot & PROT_WRITE) && !(flags & MAP_PRIVATE)) common_mmap()
485 if (prot & PROT_EXEC) common_mmap()
492 unsigned long prot, unsigned long flags) apparmor_mmap_file()
494 return common_mmap(OP_FMMAP, file, prot, flags); apparmor_mmap_file()
498 unsigned long reqprot, unsigned long prot) apparmor_file_mprotect()
500 return common_mmap(OP_FMPROT, vma->vm_file, prot, apparmor_file_mprotect()
491 apparmor_mmap_file(struct file *file, unsigned long reqprot, unsigned long prot, unsigned long flags) apparmor_mmap_file() argument
497 apparmor_file_mprotect(struct vm_area_struct *vma, unsigned long reqprot, unsigned long prot) apparmor_file_mprotect() argument
/linux-4.1.27/arch/arm/xen/
H A Denlighten.c59 int *err_ptr, pgprot_t prot, xen_remap_domain_mfn_array()
64 prot, domid, pages); xen_remap_domain_mfn_array()
72 pgprot_t prot, unsigned domid, xen_remap_domain_mfn_range()
56 xen_remap_domain_mfn_array(struct vm_area_struct *vma, unsigned long addr, xen_pfn_t *mfn, int nr, int *err_ptr, pgprot_t prot, unsigned domid, struct page **pages) xen_remap_domain_mfn_array() argument
69 xen_remap_domain_mfn_range(struct vm_area_struct *vma, unsigned long addr, xen_pfn_t mfn, int nr, pgprot_t prot, unsigned domid, struct page **pages) xen_remap_domain_mfn_range() argument
/linux-4.1.27/arch/alpha/kernel/
H A Dproto.h195 pgprot_t prot; __alpha_remap_area_pages() local
197 prot = __pgprot(_PAGE_VALID | _PAGE_ASM | _PAGE_KRE __alpha_remap_area_pages()
199 return ioremap_page_range(address, address + size, phys_addr, prot); __alpha_remap_area_pages()
/linux-4.1.27/arch/arc/include/asm/
H A Dpgtable.h271 #define pfn_pte(pfn, prot) (__pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot)))
312 #define pgprot_noncached(prot) (__pgprot(pgprot_val(prot) & ~_PAGE_CACHEABLE))
373 * remap a physical page `pfn' of size `size' with page protection `prot'
/linux-4.1.27/net/phonet/
H A Daf_phonet.c49 if (pp && !try_module_get(pp->prot->owner)) phonet_proto_get()
58 module_put(pp->prot->owner); phonet_proto_put()
100 sk = sk_alloc(net, PF_PHONET, GFP_KERNEL, pnp->prot); pn_socket_create()
480 err = proto_register(pp->prot, 1); phonet_proto_register()
502 proto_unregister(pp->prot); phonet_proto_unregister()
/linux-4.1.27/drivers/staging/android/
H A Dashmem.c354 static inline vm_flags_t calc_vm_may_flags(unsigned long prot) calc_vm_may_flags() argument
356 return _calc_vm_trans(prot, PROT_READ, VM_MAYREAD) | calc_vm_may_flags()
357 _calc_vm_trans(prot, PROT_WRITE, VM_MAYWRITE) | calc_vm_may_flags()
358 _calc_vm_trans(prot, PROT_EXEC, VM_MAYEXEC); calc_vm_may_flags()
484 static int set_prot_mask(struct ashmem_area *asma, unsigned long prot) set_prot_mask() argument
491 if (unlikely((asma->prot_mask & prot) != prot)) { set_prot_mask()
497 if ((prot & PROT_READ) && (current->personality & READ_IMPLIES_EXEC)) set_prot_mask()
498 prot |= PROT_EXEC; set_prot_mask()
500 asma->prot_mask = prot; set_prot_mask()
/linux-4.1.27/fs/proc/
H A Dvmcore.c162 unsigned long size, pgprot_t prot) remap_oldmem_pfn_range()
164 return remap_pfn_range(vma, from, pfn, size, prot); remap_oldmem_pfn_range()
339 * @prot: protection bits
345 unsigned long size, pgprot_t prot) remap_oldmem_pfn_checked()
367 prot)) remap_oldmem_pfn_checked()
374 PAGE_SIZE, prot)) remap_oldmem_pfn_checked()
384 map_size, prot)) remap_oldmem_pfn_checked()
395 unsigned long size, pgprot_t prot) vmcore_remap_oldmem_pfn()
402 return remap_oldmem_pfn_checked(vma, from, pfn, size, prot); vmcore_remap_oldmem_pfn()
404 return remap_oldmem_pfn_range(vma, from, pfn, size, prot); vmcore_remap_oldmem_pfn()
160 remap_oldmem_pfn_range(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) remap_oldmem_pfn_range() argument
343 remap_oldmem_pfn_checked(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) remap_oldmem_pfn_checked() argument
393 vmcore_remap_oldmem_pfn(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot) vmcore_remap_oldmem_pfn() argument
/linux-4.1.27/arch/frv/include/asm/
H A Dpgtable.h271 #define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
272 #define pfn_pmd(pfn, prot) __pmd(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
411 #define pgprot_noncached(prot) (__pgprot(pgprot_val(prot) | _PAGE_NOCACHE))
/linux-4.1.27/security/integrity/ima/
H A Dima_main.c261 * @prot: contains the protection that will be applied by the kernel.
269 int ima_file_mmap(struct file *file, unsigned long prot) ima_file_mmap() argument
271 if (file && (prot & PROT_EXEC)) ima_file_mmap()
/linux-4.1.27/drivers/target/
H A Dtarget_core_file.c491 sector_t prot_length, prot; fd_do_prot_fill() local
501 for (prot = 0; prot < prot_length;) { fd_do_prot_fill()
502 sector_t len = min_t(sector_t, bufsize, prot_length - prot); fd_do_prot_fill()
503 ssize_t ret = kernel_write(prot_fd, buf, len, pos + prot); fd_do_prot_fill()
506 pr_err("vfs_write to prot file failed: %zd\n", ret); fd_do_prot_fill()
509 prot += ret; fd_do_prot_fill()
523 pr_err("Unable to allocate FILEIO prot buf\n"); fd_do_prot_unmap()
881 pr_err("Unable to allocate FILEIO prot buf\n"); fd_format_prot()
/linux-4.1.27/drivers/ata/
H A Dsata_sil24.c41 __le16 prot; member in struct:sil24_prb
802 u8 prot = qc->tf.protocol; sil24_qc_defer() local
823 int is_excl = (ata_is_atapi(prot) || sil24_qc_defer()
858 u16 prot = 0; sil24_qc_prep() local
861 prot |= PRB_PROT_NCQ; sil24_qc_prep()
863 prot |= PRB_PROT_WRITE; sil24_qc_prep()
865 prot |= PRB_PROT_READ; sil24_qc_prep()
866 prb->prot = cpu_to_le16(prot); sil24_qc_prep()
/linux-4.1.27/drivers/gpu/drm/tegra/
H A Dgem.c97 int prot = IOMMU_READ | IOMMU_WRITE; tegra_bo_iommu_map() local
118 bo->sgt->nents, prot); tegra_bo_iommu_map()
504 pgprot_t prot = vm_get_page_prot(vma->vm_flags); tegra_drm_mmap() local
509 vma->vm_page_prot = pgprot_writecombine(prot); tegra_drm_mmap()
/linux-4.1.27/fs/pstore/
H A Dram_core.c389 pgprot_t prot; persistent_ram_vmap() local
397 prot = pgprot_noncached(PAGE_KERNEL); persistent_ram_vmap()
399 prot = pgprot_writecombine(PAGE_KERNEL); persistent_ram_vmap()
412 vaddr = vmap(pages, page_count, VM_MAP, prot); persistent_ram_vmap()
/linux-4.1.27/include/net/phonet/
H A Dphonet.h107 struct proto *prot; member in struct:phonet_protocol
/linux-4.1.27/arch/microblaze/pci/
H A Dpci-common.c239 pgprot_t prot = protection; __pci_mmap_set_pgprot() local
254 return pgprot_noncached(prot); __pci_mmap_set_pgprot()
265 pgprot_t prot) pci_phys_mem_access_prot()
273 return prot; pci_phys_mem_access_prot()
275 prot = pgprot_noncached(prot); for_each_pci_dev()
296 prot = pgprot_noncached_wc(prot);
300 pr_debug("PCI: Non-PCI map for %llx, prot: %lx\n",
301 (unsigned long long)offset, pgprot_val(prot));
303 return prot;
262 pci_phys_mem_access_prot(struct file *file, unsigned long pfn, unsigned long size, pgprot_t prot) pci_phys_mem_access_prot() argument
/linux-4.1.27/arch/x86/kernel/
H A Dsys_x86_64.c87 unsigned long, prot, unsigned long, flags, SYSCALL_DEFINE6()
95 error = sys_mmap_pgoff(addr, len, prot, flags, fd, off >> PAGE_SHIFT); SYSCALL_DEFINE6()

Completed in 8547 milliseconds

12