prot 142 arch/alpha/include/asm/pgtable.h #define pgprot_noncached(prot) (prot) prot 178 arch/alpha/kernel/osf_sys.c unsigned long, prot, unsigned long, flags, unsigned long, fd, prot 192 arch/alpha/kernel/osf_sys.c ret = ksys_mmap_pgoff(addr, len, prot, flags, fd, off >> PAGE_SHIFT); prot 196 arch/alpha/kernel/proto.h pgprot_t prot; prot 198 arch/alpha/kernel/proto.h prot = __pgprot(_PAGE_VALID | _PAGE_ASM | _PAGE_KRE prot 200 arch/alpha/kernel/proto.h return ioremap_page_range(address, address + size, phys_addr, prot); prot 38 arch/arc/include/asm/hugepage.h #define mk_pmd(page, prot) pte_pmd(mk_pte(page, prot)) prot 42 arch/arc/include/asm/hugepage.h #define pfn_pmd(pfn, prot) (__pmd(((pfn) << PAGE_SHIFT) | pgprot_val(prot))) prot 280 arch/arc/include/asm/pgtable.h #define mk_pte(page, prot) pfn_pte(page_to_pfn(page), prot) prot 281 arch/arc/include/asm/pgtable.h #define pfn_pte(pfn, prot) __pte(__pfn_to_phys(pfn) | pgprot_val(prot)) prot 326 arch/arc/include/asm/pgtable.h #define pgprot_noncached(prot) (__pgprot(pgprot_val(prot) & ~_PAGE_CACHEABLE)) prot 59 arch/arc/mm/ioremap.c pgprot_t prot = __pgprot(flags); prot 71 arch/arc/mm/ioremap.c prot = pgprot_noncached(prot); prot 86 arch/arc/mm/ioremap.c if (ioremap_page_range(vaddr, vaddr + size, paddr, prot)) { prot 56 arch/arm/include/asm/fixmap.h void __set_fixmap(enum fixed_addresses idx, phys_addr_t phys, pgprot_t prot); prot 46 arch/arm/include/asm/kvm_mmu.h int create_hyp_mappings(void *from, void *to, pgprot_t prot); prot 73 arch/arm/include/asm/kvm_mmu.h #define kvm_pfn_pte(pfn, prot) pfn_pte(pfn, prot) prot 74 arch/arm/include/asm/kvm_mmu.h #define kvm_pfn_pmd(pfn, prot) pfn_pmd(pfn, prot) prot 75 arch/arm/include/asm/kvm_mmu.h #define kvm_pfn_pud(pfn, prot) (__pud(0)) prot 112 arch/arm/include/asm/pgalloc.h pmdval_t prot) prot 114 arch/arm/include/asm/pgalloc.h pmdval_t pmdval = (pte + PTE_HWTABLE_OFF) | prot; prot 141 arch/arm/include/asm/pgalloc.h pmdval_t prot; prot 144 arch/arm/include/asm/pgalloc.h prot = user_pmd_table; prot 146 arch/arm/include/asm/pgalloc.h prot = _PAGE_USER_TABLE; prot 148 arch/arm/include/asm/pgalloc.h __pmd_populate(pmdp, page_to_phys(ptep), prot); prot 236 arch/arm/include/asm/pgtable-3level.h #define pfn_pmd(pfn,prot) (__pmd(((phys_addr_t)(pfn) << PAGE_SHIFT) | pgprot_val(prot))) prot 237 arch/arm/include/asm/pgtable-3level.h #define mk_pmd(page,prot) pfn_pmd(page_to_pfn(page),prot) prot 63 arch/arm/include/asm/pgtable-nommu.h #define pgprot_noncached(prot) (prot) prot 64 arch/arm/include/asm/pgtable-nommu.h #define pgprot_writecombine(prot) (prot) prot 65 arch/arm/include/asm/pgtable-nommu.h #define pgprot_device(prot) (prot) prot 113 arch/arm/include/asm/pgtable.h #define __pgprot_modify(prot,mask,bits) \ prot 114 arch/arm/include/asm/pgtable.h __pgprot((pgprot_val(prot) & ~(mask)) | (bits)) prot 116 arch/arm/include/asm/pgtable.h #define pgprot_noncached(prot) \ prot 117 arch/arm/include/asm/pgtable.h __pgprot_modify(prot, L_PTE_MT_MASK, L_PTE_MT_UNCACHED) prot 119 arch/arm/include/asm/pgtable.h #define pgprot_writecombine(prot) \ prot 120 arch/arm/include/asm/pgtable.h __pgprot_modify(prot, L_PTE_MT_MASK, L_PTE_MT_BUFFERABLE) prot 122 arch/arm/include/asm/pgtable.h #define pgprot_stronglyordered(prot) \ prot 123 arch/arm/include/asm/pgtable.h __pgprot_modify(prot, L_PTE_MT_MASK, L_PTE_MT_UNCACHED) prot 125 arch/arm/include/asm/pgtable.h #define pgprot_device(prot) \ prot 126 arch/arm/include/asm/pgtable.h __pgprot_modify(prot, L_PTE_MT_MASK, L_PTE_MT_DEV_SHARED | L_PTE_SHARED | L_PTE_DIRTY | L_PTE_XN) prot 129 arch/arm/include/asm/pgtable.h #define pgprot_dmacoherent(prot) \ prot 130 arch/arm/include/asm/pgtable.h __pgprot_modify(prot, L_PTE_MT_MASK, L_PTE_MT_BUFFERABLE | L_PTE_XN) prot 136 arch/arm/include/asm/pgtable.h #define pgprot_dmacoherent(prot) \ prot 137 arch/arm/include/asm/pgtable.h __pgprot_modify(prot, L_PTE_MT_MASK, L_PTE_MT_UNCACHED | L_PTE_XN) prot 212 arch/arm/include/asm/pgtable.h #define pfn_pte(pfn,prot) __pte(__pfn_to_phys(pfn) | pgprot_val(prot)) prot 215 arch/arm/include/asm/pgtable.h #define mk_pte(page,prot) pfn_pte(page_to_pfn(page), prot) prot 269 arch/arm/include/asm/pgtable.h static inline pte_t clear_pte_bit(pte_t pte, pgprot_t prot) prot 271 arch/arm/include/asm/pgtable.h pte_val(pte) &= ~pgprot_val(prot); prot 275 arch/arm/include/asm/pgtable.h static inline pte_t set_pte_bit(pte_t pte, pgprot_t prot) prot 277 arch/arm/include/asm/pgtable.h pte_val(pte) |= pgprot_val(prot); prot 633 arch/arm/mach-sa1100/assabet.c int prot = PMD_TYPE_SECT | PMD_SECT_AP_WRITE | PMD_DOMAIN(DOMAIN_IO); prot 637 arch/arm/mach-sa1100/assabet.c *pmd = __pmd(phys | prot); prot 48 arch/arm/mm/dma-mapping.c pgprot_t prot; prot 335 arch/arm/mm/dma-mapping.c pgprot_t prot, struct page **ret_page, prot 340 arch/arm/mm/dma-mapping.c pgprot_t prot, struct page **ret_page, prot 360 arch/arm/mm/dma-mapping.c pgprot_t prot = pgprot_dmacoherent(PAGE_KERNEL); prot 373 arch/arm/mm/dma-mapping.c ptr = __alloc_from_contiguous(NULL, atomic_pool_size, prot, prot 377 arch/arm/mm/dma-mapping.c ptr = __alloc_remap_buffer(NULL, atomic_pool_size, gfp, prot, prot 467 arch/arm/mm/dma-mapping.c pgprot_t prot = *(pgprot_t *)data; prot 469 arch/arm/mm/dma-mapping.c set_pte_ext(pte, mk_pte(page, prot), 0); prot 473 arch/arm/mm/dma-mapping.c static void __dma_remap(struct page *page, size_t size, pgprot_t prot) prot 478 arch/arm/mm/dma-mapping.c apply_to_page_range(&init_mm, start, size, __dma_update_pte, &prot); prot 483 arch/arm/mm/dma-mapping.c pgprot_t prot, struct page **ret_page, prot 498 arch/arm/mm/dma-mapping.c ptr = dma_common_contiguous_remap(page, size, prot, caller); prot 546 arch/arm/mm/dma-mapping.c pgprot_t prot, struct page **ret_page, prot 565 arch/arm/mm/dma-mapping.c ptr = dma_common_contiguous_remap(page, size, prot, caller); prot 571 arch/arm/mm/dma-mapping.c __dma_remap(page, size, prot); prot 592 arch/arm/mm/dma-mapping.c static inline pgprot_t __get_dma_pgprot(unsigned long attrs, pgprot_t prot) prot 594 arch/arm/mm/dma-mapping.c prot = (attrs & DMA_ATTR_WRITE_COMBINE) ? prot 595 arch/arm/mm/dma-mapping.c pgprot_writecombine(prot) : prot 596 arch/arm/mm/dma-mapping.c pgprot_dmacoherent(prot); prot 597 arch/arm/mm/dma-mapping.c return prot; prot 633 arch/arm/mm/dma-mapping.c return __alloc_from_contiguous(args->dev, args->size, args->prot, prot 670 arch/arm/mm/dma-mapping.c args->prot, ret_page, args->caller, prot 688 arch/arm/mm/dma-mapping.c gfp_t gfp, pgprot_t prot, bool is_coherent, prot 700 arch/arm/mm/dma-mapping.c .prot = prot, prot 774 arch/arm/mm/dma-mapping.c pgprot_t prot = __get_dma_pgprot(attrs, PAGE_KERNEL); prot 776 arch/arm/mm/dma-mapping.c return __dma_alloc(dev, size, handle, gfp, prot, false, prot 1119 arch/arm/mm/dma-mapping.c int prot = 0; prot 1122 arch/arm/mm/dma-mapping.c prot |= IOMMU_PRIV; prot 1126 arch/arm/mm/dma-mapping.c return prot | IOMMU_READ | IOMMU_WRITE; prot 1128 arch/arm/mm/dma-mapping.c return prot | IOMMU_READ; prot 1130 arch/arm/mm/dma-mapping.c return prot | IOMMU_WRITE; prot 1132 arch/arm/mm/dma-mapping.c return prot; prot 1465 arch/arm/mm/dma-mapping.c pgprot_t prot = __get_dma_pgprot(attrs, PAGE_KERNEL); prot 1496 arch/arm/mm/dma-mapping.c addr = dma_common_pages_remap(pages, size, prot, prot 1625 arch/arm/mm/dma-mapping.c int prot; prot 1641 arch/arm/mm/dma-mapping.c prot = __dma_info_to_prot(dir, attrs); prot 1643 arch/arm/mm/dma-mapping.c ret = iommu_map(mapping->domain, iova, phys, len, prot); prot 1843 arch/arm/mm/dma-mapping.c int ret, prot, len = PAGE_ALIGN(size + offset); prot 1849 arch/arm/mm/dma-mapping.c prot = __dma_info_to_prot(dir, attrs); prot 1851 arch/arm/mm/dma-mapping.c ret = iommu_map(mapping->domain, dma_addr, page_to_phys(page), len, prot); prot 1946 arch/arm/mm/dma-mapping.c int ret, prot; prot 1955 arch/arm/mm/dma-mapping.c prot = __dma_info_to_prot(dir, attrs) | IOMMU_MMIO; prot 1957 arch/arm/mm/dma-mapping.c ret = iommu_map(mapping->domain, dma_addr, addr, len, prot); prot 258 arch/arm/mm/dump.c u64 prot = val & pg_level[level].mask; prot 262 arch/arm/mm/dump.c st->current_prot = prot; prot 265 arch/arm/mm/dump.c } else if (prot != st->current_prot || level != st->level || prot 296 arch/arm/mm/dump.c st->current_prot = prot; prot 240 arch/arm/mm/fault-armv.c pgprot_t prot = __pgprot_modify(PAGE_KERNEL, prot 243 arch/arm/mm/fault-armv.c p1 = vmap(&page, 1, VM_IOREMAP, prot); prot 244 arch/arm/mm/fault-armv.c p2 = vmap(&page, 1, VM_IOREMAP, prot); prot 25 arch/arm/mm/idmap.c unsigned long prot) prot 50 arch/arm/mm/idmap.c *pmd = __pmd((addr & PMD_MASK) | prot); prot 56 arch/arm/mm/idmap.c unsigned long prot) prot 60 arch/arm/mm/idmap.c addr = (addr & PMD_MASK) | prot; prot 69 arch/arm/mm/idmap.c unsigned long prot) prot 76 arch/arm/mm/idmap.c idmap_add_pmd(pud, addr, next, prot); prot 81 arch/arm/mm/idmap.c const char *text_end, unsigned long prot) prot 90 arch/arm/mm/idmap.c prot |= PMD_TYPE_SECT | PMD_SECT_AP_WRITE | PMD_SECT_AF; prot 93 arch/arm/mm/idmap.c prot |= PMD_BIT4; prot 98 arch/arm/mm/idmap.c idmap_add_pud(pgd, addr, next, prot); prot 512 arch/arm/mm/init.c pmdval_t prot; prot 526 arch/arm/mm/init.c .prot = PMD_SECT_XN, prot 534 arch/arm/mm/init.c .prot = PMD_SECT_XN, prot 542 arch/arm/mm/init.c .prot = PMD_SECT_XN, prot 554 arch/arm/mm/init.c .prot = L_PMD_SECT_RDONLY | PMD_SECT_AP2, prot 557 arch/arm/mm/init.c .prot = PMD_SECT_APX | PMD_SECT_AP_WRITE, prot 569 arch/arm/mm/init.c pmdval_t prot, struct mm_struct *mm) prot 576 arch/arm/mm/init.c pmd[0] = __pmd((pmd_val(pmd[0]) & mask) | prot); prot 579 arch/arm/mm/init.c pmd[1] = __pmd((pmd_val(pmd[1]) & mask) | prot); prot 581 arch/arm/mm/init.c pmd[0] = __pmd((pmd_val(pmd[0]) & mask) | prot); prot 618 arch/arm/mm/init.c set ? perms[i].prot : perms[i].clear, mm); prot 406 arch/arm/mm/mmu.c void __set_fixmap(enum fixed_addresses idx, phys_addr_t phys, pgprot_t prot) prot 417 arch/arm/mm/mmu.c if (WARN_ON(pgprot_val(prot) != pgprot_val(FIXMAP_PAGE_IO) && prot 421 arch/arm/mm/mmu.c if (pgprot_val(prot)) prot 423 arch/arm/mm/mmu.c pfn_pte(phys >> PAGE_SHIFT, prot)); prot 740 arch/arm/mm/mmu.c unsigned long prot, prot 745 arch/arm/mm/mmu.c __pmd_populate(pmd, __pa(pte), prot); prot 752 arch/arm/mm/mmu.c unsigned long prot) prot 754 arch/arm/mm/mmu.c return arm_pte_alloc(pmd, addr, prot, early_alloc); prot 107 arch/arm64/include/asm/fixmap.h extern void __set_fixmap(enum fixed_addresses idx, phys_addr_t phys, pgprot_t prot); prot 165 arch/arm64/include/asm/io.h extern void __iomem *__ioremap(phys_addr_t phys_addr, size_t size, pgprot_t prot); prot 147 arch/arm64/include/asm/kvm_mmu.h int create_hyp_mappings(void *from, void *to, pgprot_t prot); prot 179 arch/arm64/include/asm/kvm_mmu.h #define kvm_pfn_pte(pfn, prot) pfn_pte(pfn, prot) prot 180 arch/arm64/include/asm/kvm_mmu.h #define kvm_pfn_pmd(pfn, prot) pfn_pmd(pfn, prot) prot 181 arch/arm64/include/asm/kvm_mmu.h #define kvm_pfn_pud(pfn, prot) pfn_pud(pfn, prot) prot 128 arch/arm64/include/asm/mmu.h pgprot_t prot, bool page_mappings_only); prot 129 arch/arm64/include/asm/mmu.h extern void *fixmap_remap_fdt(phys_addr_t dt_phys, int *size, pgprot_t prot); prot 47 arch/arm64/include/asm/pgalloc.h static inline void __pud_populate(pud_t *pudp, phys_addr_t pmdp, pudval_t prot) prot 49 arch/arm64/include/asm/pgalloc.h set_pud(pudp, __pud(__phys_to_pud_val(pmdp) | prot)); prot 57 arch/arm64/include/asm/pgalloc.h static inline void __pud_populate(pud_t *pudp, phys_addr_t pmdp, pudval_t prot) prot 76 arch/arm64/include/asm/pgalloc.h static inline void __pgd_populate(pgd_t *pgdp, phys_addr_t pudp, pgdval_t prot) prot 78 arch/arm64/include/asm/pgalloc.h set_pgd(pgdp, __pgd(__phys_to_pgd_val(pudp) | prot)); prot 86 arch/arm64/include/asm/pgalloc.h static inline void __pgd_populate(pgd_t *pgdp, phys_addr_t pudp, pgdval_t prot) prot 96 arch/arm64/include/asm/pgalloc.h pmdval_t prot) prot 98 arch/arm64/include/asm/pgalloc.h set_pmd(pmdp, __pmd(__phys_to_pmd_val(ptep) | prot)); prot 66 arch/arm64/include/asm/pgtable.h #define pfn_pte(pfn,prot) \ prot 67 arch/arm64/include/asm/pgtable.h __pte(__phys_to_pte_val((phys_addr_t)(pfn) << PAGE_SHIFT) | pgprot_val(prot)) prot 126 arch/arm64/include/asm/pgtable.h static inline pte_t clear_pte_bit(pte_t pte, pgprot_t prot) prot 128 arch/arm64/include/asm/pgtable.h pte_val(pte) &= ~pgprot_val(prot); prot 132 arch/arm64/include/asm/pgtable.h static inline pte_t set_pte_bit(pte_t pte, pgprot_t prot) prot 134 arch/arm64/include/asm/pgtable.h pte_val(pte) |= pgprot_val(prot); prot 326 arch/arm64/include/asm/pgtable.h static inline pgprot_t mk_pud_sect_prot(pgprot_t prot) prot 328 arch/arm64/include/asm/pgtable.h return __pgprot((pgprot_val(prot) & ~PUD_TABLE_BIT) | PUD_TYPE_SECT); prot 331 arch/arm64/include/asm/pgtable.h static inline pgprot_t mk_pmd_sect_prot(pgprot_t prot) prot 333 arch/arm64/include/asm/pgtable.h return __pgprot((pgprot_val(prot) & ~PMD_TABLE_BIT) | PMD_TYPE_SECT); prot 388 arch/arm64/include/asm/pgtable.h #define pfn_pmd(pfn,prot) __pmd(__phys_to_pmd_val((phys_addr_t)(pfn) << PAGE_SHIFT) | pgprot_val(prot)) prot 389 arch/arm64/include/asm/pgtable.h #define mk_pmd(page,prot) pfn_pmd(page_to_pfn(page),prot) prot 400 arch/arm64/include/asm/pgtable.h #define pfn_pud(pfn,prot) __pud(__phys_to_pud_val((phys_addr_t)(pfn) << PAGE_SHIFT) | pgprot_val(prot)) prot 407 arch/arm64/include/asm/pgtable.h #define __pgprot_modify(prot,mask,bits) \ prot 408 arch/arm64/include/asm/pgtable.h __pgprot((pgprot_val(prot) & ~(mask)) | (bits)) prot 413 arch/arm64/include/asm/pgtable.h #define pgprot_noncached(prot) \ prot 414 arch/arm64/include/asm/pgtable.h __pgprot_modify(prot, PTE_ATTRINDX_MASK, PTE_ATTRINDX(MT_DEVICE_nGnRnE) | PTE_PXN | PTE_UXN) prot 415 arch/arm64/include/asm/pgtable.h #define pgprot_writecombine(prot) \ prot 416 arch/arm64/include/asm/pgtable.h __pgprot_modify(prot, PTE_ATTRINDX_MASK, PTE_ATTRINDX(MT_NORMAL_NC) | PTE_PXN | PTE_UXN) prot 417 arch/arm64/include/asm/pgtable.h #define pgprot_device(prot) \ prot 418 arch/arm64/include/asm/pgtable.h __pgprot_modify(prot, PTE_ATTRINDX_MASK, PTE_ATTRINDX(MT_DEVICE_nGnRE) | PTE_PXN | PTE_UXN) prot 427 arch/arm64/include/asm/pgtable.h #define pgprot_dmacoherent(prot) \ prot 428 arch/arm64/include/asm/pgtable.h __pgprot_modify(prot, PTE_ATTRINDX_MASK, \ prot 519 arch/arm64/include/asm/pgtable.h #define mk_pte(page,prot) pfn_pte(page_to_pfn(page),prot) prot 22 arch/arm64/kernel/sys.c unsigned long, prot, unsigned long, flags, prot 28 arch/arm64/kernel/sys.c return ksys_mmap_pgoff(addr, len, prot, flags, fd, off >> PAGE_SHIFT); prot 56 arch/arm64/kernel/sys32.c unsigned long, prot, unsigned long, flags, prot 64 arch/arm64/kernel/sys32.c return ksys_mmap_pgoff(addr, len, prot, flags, fd, off_4k); prot 247 arch/arm64/mm/dump.c u64 prot = val & pg_level[level].mask; prot 251 arch/arm64/mm/dump.c st->current_prot = prot; prot 254 arch/arm64/mm/dump.c } else if (prot != st->current_prot || level != st->level || prot 284 arch/arm64/mm/dump.c st->current_prot = prot; prot 22 arch/arm64/mm/ioremap.c pgprot_t prot, void *caller) prot 56 arch/arm64/mm/ioremap.c err = ioremap_page_range(addr, addr + size, phys_addr, prot); prot 65 arch/arm64/mm/ioremap.c void __iomem *__ioremap(phys_addr_t phys_addr, size_t size, pgprot_t prot) prot 67 arch/arm64/mm/ioremap.c return __ioremap_caller(phys_addr, size, prot, prot 141 arch/arm64/mm/mmu.c phys_addr_t phys, pgprot_t prot) prot 149 arch/arm64/mm/mmu.c set_pte(ptep, pfn_pte(__phys_to_pfn(phys), prot)); prot 166 arch/arm64/mm/mmu.c pgprot_t prot, prot 184 arch/arm64/mm/mmu.c pgprot_t __prot = prot; prot 191 arch/arm64/mm/mmu.c __prot = __pgprot(pgprot_val(prot) | PTE_CONT); prot 200 arch/arm64/mm/mmu.c phys_addr_t phys, pgprot_t prot, prot 215 arch/arm64/mm/mmu.c pmd_set_huge(pmdp, phys, prot); prot 224 arch/arm64/mm/mmu.c alloc_init_cont_pte(pmdp, addr, next, phys, prot, prot 238 arch/arm64/mm/mmu.c pgprot_t prot, prot 258 arch/arm64/mm/mmu.c pgprot_t __prot = prot; prot 265 arch/arm64/mm/mmu.c __prot = __pgprot(pgprot_val(prot) | PTE_CONT); prot 286 arch/arm64/mm/mmu.c phys_addr_t phys, pgprot_t prot, prot 314 arch/arm64/mm/mmu.c pud_set_huge(pudp, phys, prot); prot 323 arch/arm64/mm/mmu.c alloc_init_cont_pmd(pudp, addr, next, phys, prot, prot 337 arch/arm64/mm/mmu.c pgprot_t prot, prot 358 arch/arm64/mm/mmu.c alloc_init_pud(pgdp, addr, next, phys, prot, pgtable_alloc, prot 400 arch/arm64/mm/mmu.c phys_addr_t size, pgprot_t prot) prot 407 arch/arm64/mm/mmu.c __create_pgd_mapping(init_mm.pgd, phys, virt, size, prot, NULL, prot 413 arch/arm64/mm/mmu.c pgprot_t prot, bool page_mappings_only) prot 422 arch/arm64/mm/mmu.c __create_pgd_mapping(mm->pgd, phys, virt, size, prot, prot 427 arch/arm64/mm/mmu.c phys_addr_t size, pgprot_t prot) prot 435 arch/arm64/mm/mmu.c __create_pgd_mapping(init_mm.pgd, phys, virt, size, prot, NULL, prot 443 arch/arm64/mm/mmu.c phys_addr_t end, pgprot_t prot, int flags) prot 446 arch/arm64/mm/mmu.c prot, early_pgtable_alloc, flags); prot 541 arch/arm64/mm/mmu.c pgprot_t prot, struct vm_struct *vma, prot 550 arch/arm64/mm/mmu.c __create_pgd_mapping(pgdp, pa_start, (unsigned long)va_start, size, prot, prot 586 arch/arm64/mm/mmu.c pgprot_t prot = rodata_enabled ? PAGE_KERNEL_ROX : PAGE_KERNEL_EXEC; prot 590 arch/arm64/mm/mmu.c pgprot_val(prot) &= ~PTE_NG; prot 595 arch/arm64/mm/mmu.c prot, __pgd_pgtable_alloc, 0); prot 598 arch/arm64/mm/mmu.c __set_fixmap(FIX_ENTRY_TRAMP_TEXT, pa_start, prot); prot 882 arch/arm64/mm/mmu.c void *__init fixmap_remap_fdt(phys_addr_t dt_phys, int *size, pgprot_t prot) prot 919 arch/arm64/mm/mmu.c dt_virt_base, SWAPPER_BLOCK_SIZE, prot); prot 930 arch/arm64/mm/mmu.c round_up(offset + *size, SWAPPER_BLOCK_SIZE), prot); prot 956 arch/arm64/mm/mmu.c int pud_set_huge(pud_t *pudp, phys_addr_t phys, pgprot_t prot) prot 958 arch/arm64/mm/mmu.c pud_t new_pud = pfn_pud(__phys_to_pfn(phys), mk_pud_sect_prot(prot)); prot 970 arch/arm64/mm/mmu.c int pmd_set_huge(pmd_t *pmdp, phys_addr_t phys, pgprot_t prot) prot 972 arch/arm64/mm/mmu.c pmd_t new_pmd = pfn_pmd(__phys_to_pfn(phys), mk_pmd_sect_prot(prot)); prot 40 arch/c6x/include/asm/pgtable.h #define pgprot_noncached(prot) (prot) prot 40 arch/csky/include/asm/io.h extern void __iomem *__ioremap(phys_addr_t addr, size_t size, pgprot_t prot); prot 52 arch/csky/include/asm/pgtable.h #define pfn_pte(pfn, prot) __pte(((unsigned long long)(pfn) << PAGE_SHIFT) \ prot 53 arch/csky/include/asm/pgtable.h | pgprot_val(prot)) prot 259 arch/csky/include/asm/pgtable.h unsigned long prot = pgprot_val(_prot); prot 261 arch/csky/include/asm/pgtable.h prot = (prot & ~_CACHE_MASK) | _CACHE_UNCACHED | _PAGE_SO; prot 263 arch/csky/include/asm/pgtable.h return __pgprot(prot); prot 269 arch/csky/include/asm/pgtable.h unsigned long prot = pgprot_val(_prot); prot 271 arch/csky/include/asm/pgtable.h prot = (prot & ~_CACHE_MASK) | _CACHE_UNCACHED; prot 273 arch/csky/include/asm/pgtable.h return __pgprot(prot); prot 309 arch/csky/include/asm/pgtable.h #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ prot 310 arch/csky/include/asm/pgtable.h remap_pfn_range(vma, vaddr, pfn, size, prot) prot 20 arch/csky/kernel/syscall.c unsigned long, prot, prot 28 arch/csky/kernel/syscall.c return ksys_mmap_pgoff(addr, len, prot, flags, fd, prot 12 arch/csky/mm/ioremap.c pgprot_t prot, void *caller) prot 32 arch/csky/mm/ioremap.c if (ioremap_page_range(vaddr, vaddr + size, addr, prot)) { prot 40 arch/csky/mm/ioremap.c void __iomem *__ioremap(phys_addr_t phys_addr, size_t size, pgprot_t prot) prot 42 arch/csky/mm/ioremap.c return __ioremap_caller(phys_addr, size, prot, prot 19 arch/h8300/include/asm/pgtable.h #define pgprot_writecombine(prot) (prot) prot 341 arch/hexagon/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t prot) prot 344 arch/hexagon/include/asm/pgtable.h pte_val(pte) |= pgprot_val(prot); prot 18 arch/hexagon/mm/ioremap.c pgprot_t prot = __pgprot(_PAGE_PRESENT|_PAGE_READ|_PAGE_WRITE prot 33 arch/hexagon/mm/ioremap.c if (ioremap_page_range(addr, addr+size, phys_addr, prot)) { prot 926 arch/ia64/include/asm/pal.h ia64_pal_cache_prot_info (u64 cache_level, u64 cache_type, pal_cache_protection_info_t *prot) prot 933 arch/ia64/include/asm/pal.h prot->pcpi_status = iprv.status; prot 934 arch/ia64/include/asm/pal.h prot->pcp_info[0].pcpi_data = iprv.v0 & 0xffffffff; prot 935 arch/ia64/include/asm/pal.h prot->pcp_info[1].pcpi_data = iprv.v0 >> 32; prot 936 arch/ia64/include/asm/pal.h prot->pcp_info[2].pcpi_data = iprv.v1 & 0xffffffff; prot 937 arch/ia64/include/asm/pal.h prot->pcp_info[3].pcpi_data = iprv.v1 >> 32; prot 938 arch/ia64/include/asm/pal.h prot->pcp_info[4].pcpi_data = iprv.v2 & 0xffffffff; prot 939 arch/ia64/include/asm/pal.h prot->pcp_info[5].pcpi_data = iprv.v2 >> 32; prot 352 arch/ia64/include/asm/pgtable.h #define pgprot_cacheable(prot) __pgprot((pgprot_val(prot) & ~_PAGE_MA_MASK) | _PAGE_MA_WB) prot 353 arch/ia64/include/asm/pgtable.h #define pgprot_noncached(prot) __pgprot((pgprot_val(prot) & ~_PAGE_MA_MASK) | _PAGE_MA_UC) prot 354 arch/ia64/include/asm/pgtable.h #define pgprot_writecombine(prot) __pgprot((pgprot_val(prot) & ~_PAGE_MA_MASK) | _PAGE_MA_WC) prot 28 arch/ia64/include/asm/unistd.h int prot, int flags, prot 32 arch/ia64/include/asm/unistd.h int prot, int flags, prot 140 arch/ia64/kernel/sys_ia64.c sys_mmap2 (unsigned long addr, unsigned long len, int prot, int flags, int fd, long pgoff) prot 142 arch/ia64/kernel/sys_ia64.c addr = ksys_mmap_pgoff(addr, len, prot, flags, fd, pgoff); prot 149 arch/ia64/kernel/sys_ia64.c sys_mmap (unsigned long addr, unsigned long len, int prot, int flags, int fd, long off) prot 154 arch/ia64/kernel/sys_ia64.c addr = ksys_mmap_pgoff(addr, len, prot, flags, fd, off >> PAGE_SHIFT); prot 37 arch/ia64/mm/ioremap.c pgprot_t prot; prot 71 arch/ia64/mm/ioremap.c prot = PAGE_KERNEL; prot 89 arch/ia64/mm/ioremap.c (unsigned long) addr + size, phys_addr, prot)) { prot 443 arch/ia64/pci/pci.c pgprot_t prot; prot 456 arch/ia64/pci/pci.c prot = phys_mem_access_prot(NULL, vma->vm_pgoff, size, prot 464 arch/ia64/pci/pci.c vma->vm_page_prot = prot; prot 404 arch/m68k/include/asm/mcf_pgtable.h #define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) prot 136 arch/m68k/include/asm/motorola_pgtable.h #define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) prot 156 arch/m68k/include/asm/pgtable_mm.h # define pgprot_noncached(prot) (__pgprot(pgprot_val(prot) | CF_PAGE_NOCACHE)) prot 163 arch/m68k/include/asm/pgtable_mm.h #define pgprot_noncached(prot) \ prot 165 arch/m68k/include/asm/pgtable_mm.h ? (__pgprot(pgprot_val(prot) | __SUN3_PAGE_NOCACHE)) \ prot 167 arch/m68k/include/asm/pgtable_mm.h ? (__pgprot(pgprot_val(prot) | _PAGE_NOCACHE030)) \ prot 169 arch/m68k/include/asm/pgtable_mm.h ? (__pgprot((pgprot_val(prot) & _CACHEMASK040) | _PAGE_NOCACHE_S)) \ prot 170 arch/m68k/include/asm/pgtable_mm.h : (prot))) prot 172 arch/m68k/include/asm/pgtable_mm.h pgprot_t pgprot_dmacoherent(pgprot_t prot); prot 173 arch/m68k/include/asm/pgtable_mm.h #define pgprot_dmacoherent(prot) pgprot_dmacoherent(prot) prot 26 arch/m68k/kernel/dma.c pgprot_t pgprot_dmacoherent(pgprot_t prot) prot 29 arch/m68k/kernel/dma.c pgprot_val(prot) &= ~_PAGE_CACHE040; prot 30 arch/m68k/kernel/dma.c pgprot_val(prot) |= _PAGE_GLOBAL040 | _PAGE_NOCACHE_S; prot 32 arch/m68k/kernel/dma.c pgprot_val(prot) |= _PAGE_NOCACHE030; prot 34 arch/m68k/kernel/dma.c return prot; prot 41 arch/m68k/kernel/sys_m68k.c unsigned long prot, unsigned long flags, prot 49 arch/m68k/kernel/sys_m68k.c return ksys_mmap_pgoff(addr, len, prot, flags, fd, pgoff); prot 56 arch/microblaze/include/asm/highmem.h extern void *kmap_atomic_prot(struct page *page, pgprot_t prot); prot 67 arch/microblaze/include/asm/pci.h pgprot_t prot); prot 51 arch/microblaze/include/asm/pgtable.h #define pgprot_noncached_wc(prot) prot prot 104 arch/microblaze/include/asm/pgtable.h #define pgprot_noncached(prot) \ prot 105 arch/microblaze/include/asm/pgtable.h (__pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | \ prot 108 arch/microblaze/include/asm/pgtable.h #define pgprot_noncached_wc(prot) \ prot 109 arch/microblaze/include/asm/pgtable.h (__pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | \ prot 312 arch/microblaze/include/asm/pgtable.h #define pfn_pte(pfn, prot) \ prot 313 arch/microblaze/include/asm/pgtable.h __pte(((pte_basic_t)(pfn) << PFN_SHIFT_OFFSET) | pgprot_val(prot)) prot 37 arch/microblaze/kernel/sys_microblaze.c unsigned long, prot, unsigned long, flags, unsigned long, fd, prot 43 arch/microblaze/kernel/sys_microblaze.c return ksys_mmap_pgoff(addr, len, prot, flags, fd, pgoff >> PAGE_SHIFT); prot 47 arch/microblaze/kernel/sys_microblaze.c unsigned long, prot, unsigned long, flags, unsigned long, fd, prot 53 arch/microblaze/kernel/sys_microblaze.c return ksys_mmap_pgoff(addr, len, prot, flags, fd, prot 35 arch/microblaze/mm/highmem.c void *kmap_atomic_prot(struct page *page, pgprot_t prot) prot 53 arch/microblaze/mm/highmem.c set_pte_at(&init_mm, vaddr, kmap_pte-idx, mk_pte(page, prot)); prot 176 arch/microblaze/pci/pci-common.c pgprot_t prot) prot 184 arch/microblaze/pci/pci-common.c return prot; prot 186 arch/microblaze/pci/pci-common.c prot = pgprot_noncached(prot); prot 207 arch/microblaze/pci/pci-common.c prot = pgprot_noncached_wc(prot); prot 212 arch/microblaze/pci/pci-common.c (unsigned long long)offset, pgprot_val(prot)); prot 214 arch/microblaze/pci/pci-common.c return prot; prot 160 arch/mips/include/asm/pgtable-32.h pfn_pte(unsigned long pfn, pgprot_t prot) prot 165 arch/mips/include/asm/pgtable-32.h (pgprot_val(prot) & ~_PFNX_MASK); prot 167 arch/mips/include/asm/pgtable-32.h (pgprot_val(prot) & ~_PFN_MASK); prot 175 arch/mips/include/asm/pgtable-32.h static inline pte_t pfn_pte(unsigned long pfn, pgprot_t prot) prot 179 arch/mips/include/asm/pgtable-32.h pte.pte_high = (pfn << 6) | (pgprot_val(prot) & 0x3f); prot 180 arch/mips/include/asm/pgtable-32.h pte.pte_low = pgprot_val(prot); prot 189 arch/mips/include/asm/pgtable-32.h #define pfn_pte(pfn, prot) __pte(((pfn) << (PAGE_SHIFT + 2)) | pgprot_val(prot)) prot 192 arch/mips/include/asm/pgtable-32.h #define pfn_pte(pfn, prot) __pte(((unsigned long long)(pfn) << _PFN_SHIFT) | pgprot_val(prot)) prot 193 arch/mips/include/asm/pgtable-32.h #define pfn_pmd(pfn, prot) __pmd(((unsigned long long)(pfn) << _PFN_SHIFT) | pgprot_val(prot)) prot 315 arch/mips/include/asm/pgtable-64.h #define pfn_pte(pfn, prot) __pte(((pfn) << (PAGE_SHIFT + 2)) | pgprot_val(prot)) prot 318 arch/mips/include/asm/pgtable-64.h #define pfn_pte(pfn, prot) __pte(((pfn) << _PFN_SHIFT) | pgprot_val(prot)) prot 319 arch/mips/include/asm/pgtable-64.h #define pfn_pmd(pfn, prot) __pmd(((pfn) << _PFN_SHIFT) | pgprot_val(prot)) prot 422 arch/mips/include/asm/pgtable.h unsigned long prot = pgprot_val(_prot); prot 424 arch/mips/include/asm/pgtable.h prot = (prot & ~_CACHE_MASK) | _CACHE_UNCACHED; prot 426 arch/mips/include/asm/pgtable.h return __pgprot(prot); prot 433 arch/mips/include/asm/pgtable.h unsigned long prot = pgprot_val(_prot); prot 436 arch/mips/include/asm/pgtable.h prot = (prot & ~_CACHE_MASK) | cpu_data[0].writecombine; prot 438 arch/mips/include/asm/pgtable.h return __pgprot(prot); prot 495 arch/mips/include/asm/pgtable.h extern int remap_pfn_range(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot); prot 501 arch/mips/include/asm/pgtable.h pgprot_t prot) prot 504 arch/mips/include/asm/pgtable.h return remap_pfn_range(vma, vaddr, phys_addr_high >> PAGE_SHIFT, size, prot); prot 596 arch/mips/include/asm/pgtable.h extern pmd_t mk_pmd(struct page *page, pgprot_t prot); prot 615 arch/mips/include/asm/r4kcache.h #define __BUILD_BLAST_CACHE_RANGE(pfx, desc, hitop, prot, extra) \ prot 616 arch/mips/include/asm/r4kcache.h static inline void prot##extra##blast_##pfx##cache##_range(unsigned long start, \ prot 624 arch/mips/include/asm/r4kcache.h prot##cache_op(hitop, addr); \ prot 61 arch/mips/kernel/syscall.c unsigned long, prot, unsigned long, flags, unsigned long, prot 66 arch/mips/kernel/syscall.c return ksys_mmap_pgoff(addr, len, prot, flags, fd, prot 71 arch/mips/kernel/syscall.c unsigned long, prot, unsigned long, flags, unsigned long, fd, prot 77 arch/mips/kernel/syscall.c return ksys_mmap_pgoff(addr, len, prot, flags, fd, prot 84 arch/mips/mm/init.c static void *__kmap_pgprot(struct page *page, unsigned long addr, pgprot_t prot) prot 100 arch/mips/mm/init.c pte = mk_pte(page, prot); prot 35 arch/mips/mm/pgtable-32.c pmd_t mk_pmd(struct page *page, pgprot_t prot) prot 39 arch/mips/mm/pgtable-32.c pmd_val(pmd) = (page_to_pfn(page) << _PFN_SHIFT) | pgprot_val(prot); prot 91 arch/mips/mm/pgtable-64.c pmd_t mk_pmd(struct page *page, pgprot_t prot) prot 95 arch/mips/mm/pgtable-64.c pmd_val(pmd) = (page_to_pfn(page) << _PFN_SHIFT) | pgprot_val(prot); prot 26 arch/nds32/include/asm/fixmap.h void __set_fixmap(enum fixed_addresses idx, phys_addr_t phys, pgprot_t prot); prot 188 arch/nds32/include/asm/pgtable.h #define pfn_pte(pfn,prot) (__pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot))) prot 307 arch/nds32/include/asm/pgtable.h #define pgprot_noncached(prot) __pgprot((pgprot_val(prot)&~_PAGE_C_MASK) | _PAGE_C_DEV) prot 308 arch/nds32/include/asm/pgtable.h #define pgprot_writecombine(prot) __pgprot((pgprot_val(prot)&~_PAGE_C_MASK) | _PAGE_C_DEV_WB) prot 317 arch/nds32/include/asm/pgtable.h static inline pmd_t __mk_pmd(pte_t * ptep, unsigned long prot) prot 327 arch/nds32/include/asm/pgtable.h pmd_val(pmd) = __virt_to_phys(ptr) | prot; prot 342 arch/nds32/include/asm/pgtable.h #define mk_pte(page,prot) pfn_pte(page_to_pfn(page),prot) prot 354 arch/nds32/include/asm/pgtable.h #define page_pte_prot(page,prot) mk_pte(page, prot) prot 13 arch/nds32/kernel/sys_nds32.c unsigned long, prot, unsigned long, flags, prot 19 arch/nds32/kernel/sys_nds32.c return sys_mmap_pgoff(addr, len, prot, flags, fd, prot 120 arch/nds32/kernel/vdso.c pgprot_t prot; prot 170 arch/nds32/kernel/vdso.c prot = __pgprot(_PAGE_V | _PAGE_M_UR_KR | _PAGE_D | _PAGE_C_DEV); prot 172 arch/nds32/kernel/vdso.c PAGE_SIZE, prot); prot 16 arch/nds32/mm/ioremap.c pgprot_t prot; prot 39 arch/nds32/mm/ioremap.c prot = __pgprot(_PAGE_V | _PAGE_M_KRW | _PAGE_D | prot 41 arch/nds32/mm/ioremap.c if (ioremap_page_range(addr, addr + size, phys_addr, prot)) { prot 122 arch/nios2/include/asm/pgtable.h unsigned long prot = pgprot_val(_prot); prot 124 arch/nios2/include/asm/pgtable.h prot &= ~_PAGE_CACHED; prot 126 arch/nios2/include/asm/pgtable.h return __pgprot(prot); prot 199 arch/nios2/include/asm/pgtable.h #define pfn_pte(pfn, prot) (__pte(pfn | pgprot_val(prot))) prot 241 arch/nios2/include/asm/pgtable.h #define mk_pte(page, prot) (pfn_pte(page_to_pfn(page), prot)) prot 402 arch/openrisc/include/asm/pgtable.h #define pfn_pte(pfn, prot) __pte((((pfn) << PAGE_SHIFT)) | pgprot_val(prot)) prot 72 arch/openrisc/mm/init.c pgprot_t prot; prot 114 arch/openrisc/mm/init.c prot = PAGE_KERNEL; prot 116 arch/openrisc/mm/init.c prot = PAGE_KERNEL_RO; prot 118 arch/openrisc/mm/init.c set_pte(pte, mk_pte_phys(p, prot)); prot 132 arch/parisc/include/asm/cacheflush.h #define kmap_atomic_prot(page, prot) kmap_atomic(page) prot 569 arch/parisc/include/asm/pgtable.h #define pgprot_noncached(prot) __pgprot(pgprot_val(prot) | _PAGE_NO_CACHE) prot 256 arch/parisc/kernel/sys_parisc.c unsigned long prot, unsigned long flags, unsigned long fd, prot 261 arch/parisc/kernel/sys_parisc.c return ksys_mmap_pgoff(addr, len, prot, flags, fd, prot 266 arch/parisc/kernel/sys_parisc.c unsigned long prot, unsigned long flags, unsigned long fd, prot 270 arch/parisc/kernel/sys_parisc.c return ksys_mmap_pgoff(addr, len, prot, flags, fd, prot 433 arch/parisc/mm/init.c pgprot_t prot; prot 437 arch/parisc/mm/init.c prot = pgprot; prot 440 arch/parisc/mm/init.c prot = PAGE_KERNEL; prot 443 arch/parisc/mm/init.c prot = PAGE_KERNEL_RWX; prot 447 arch/parisc/mm/init.c prot = (address < ro_end) ? prot 451 arch/parisc/mm/init.c prot = PAGE_KERNEL; prot 454 arch/parisc/mm/init.c pte = __mk_pte(address, prot); prot 140 arch/powerpc/include/asm/book3s/32/pgtable.h int map_kernel_page(unsigned long va, phys_addr_t pa, pgprot_t prot); prot 583 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pgprot_t pgprot_noncached(pgprot_t prot) prot 585 arch/powerpc/include/asm/book3s/32/pgtable.h return __pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | prot 590 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pgprot_t pgprot_noncached_wc(pgprot_t prot) prot 592 arch/powerpc/include/asm/book3s/32/pgtable.h return __pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | prot 597 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pgprot_t pgprot_cached(pgprot_t prot) prot 599 arch/powerpc/include/asm/book3s/32/pgtable.h return __pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | prot 604 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pgprot_t pgprot_cached_wthru(pgprot_t prot) prot 606 arch/powerpc/include/asm/book3s/32/pgtable.h return __pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | prot 611 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pgprot_t pgprot_cached_noncoherent(pgprot_t prot) prot 613 arch/powerpc/include/asm/book3s/32/pgtable.h return __pgprot(pgprot_val(prot) & ~_PAGE_CACHE_CTL); prot 617 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pgprot_t pgprot_writecombine(pgprot_t prot) prot 619 arch/powerpc/include/asm/book3s/32/pgtable.h return pgprot_noncached_wc(prot); prot 69 arch/powerpc/include/asm/book3s/64/hash-4k.h #define remap_4k_pfn(vma, addr, pfn, prot) \ prot 70 arch/powerpc/include/asm/book3s/64/hash-4k.h remap_pfn_range((vma), (addr), (pfn), PAGE_SIZE, (prot)) prot 163 arch/powerpc/include/asm/book3s/64/hash-64k.h unsigned long pfn, pgprot_t prot) prot 170 arch/powerpc/include/asm/book3s/64/hash-64k.h __pgprot(pgprot_val(prot) | H_PAGE_4K_PFN)); prot 247 arch/powerpc/include/asm/book3s/64/hash.h int hash__map_kernel_page(unsigned long ea, unsigned long pa, pgprot_t prot); prot 491 arch/powerpc/include/asm/book3s/64/mmu-hash.h unsigned long pstart, unsigned long prot, prot 68 arch/powerpc/include/asm/book3s/64/pgtable-64k.h unsigned long pfn, pgprot_t prot) prot 72 arch/powerpc/include/asm/book3s/64/pgtable-64k.h return hash__remap_4k_pfn(vma, addr, pfn, prot); prot 826 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pgprot_t pgprot_noncached(pgprot_t prot) prot 828 arch/powerpc/include/asm/book3s/64/pgtable.h return __pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | prot 833 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pgprot_t pgprot_noncached_wc(pgprot_t prot) prot 835 arch/powerpc/include/asm/book3s/64/pgtable.h return __pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | prot 840 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pgprot_t pgprot_cached(pgprot_t prot) prot 842 arch/powerpc/include/asm/book3s/64/pgtable.h return __pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL)); prot 846 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pgprot_t pgprot_writecombine(pgprot_t prot) prot 848 arch/powerpc/include/asm/book3s/64/pgtable.h return pgprot_noncached_wc(prot); prot 1037 arch/powerpc/include/asm/book3s/64/pgtable.h static inline int map_kernel_page(unsigned long ea, unsigned long pa, pgprot_t prot) prot 1044 arch/powerpc/include/asm/book3s/64/pgtable.h return radix__map_kernel_page(ea, pa, prot, PAGE_SIZE); prot 1046 arch/powerpc/include/asm/book3s/64/pgtable.h return hash__map_kernel_page(ea, pa, prot); prot 64 arch/powerpc/include/asm/highmem.h extern void *kmap_atomic_prot(struct page *page, pgprot_t prot); prot 53 arch/powerpc/include/asm/io-workarounds.h pgprot_t prot, void *caller); prot 726 arch/powerpc/include/asm/io.h unsigned long size, pgprot_t prot); prot 728 arch/powerpc/include/asm/io.h pgprot_t prot, void *caller); prot 731 arch/powerpc/include/asm/io.h pgprot_t prot, void *caller); prot 734 arch/powerpc/include/asm/io.h unsigned long size, pgprot_t prot); prot 20 arch/powerpc/include/asm/mman.h static inline unsigned long arch_calc_vm_prot_bits(unsigned long prot, prot 24 arch/powerpc/include/asm/mman.h return (((prot & PROT_SAO) ? VM_SAO : 0) | pkey_to_vmflag_bits(pkey)); prot 26 arch/powerpc/include/asm/mman.h return ((prot & PROT_SAO) ? VM_SAO : 0); prot 29 arch/powerpc/include/asm/mman.h #define arch_calc_vm_prot_bits(prot, pkey) arch_calc_vm_prot_bits(prot, pkey) prot 43 arch/powerpc/include/asm/mman.h static inline bool arch_validate_prot(unsigned long prot, unsigned long addr) prot 45 arch/powerpc/include/asm/mman.h if (prot & ~(PROT_READ | PROT_WRITE | PROT_EXEC | PROT_SEM | PROT_SAO)) prot 47 arch/powerpc/include/asm/mman.h if ((prot & PROT_SAO) && !cpu_has_feature(CPU_FTR_SAO)) prot 67 arch/powerpc/include/asm/nohash/32/pgtable.h int map_kernel_page(unsigned long va, phys_addr_t pa, pgprot_t prot); prot 90 arch/powerpc/include/asm/nohash/64/pgtable-4k.h #define remap_4k_pfn(vma, addr, pfn, prot) \ prot 91 arch/powerpc/include/asm/nohash/64/pgtable-4k.h remap_pfn_range((vma), (addr), (pfn), PAGE_SIZE, (prot)) prot 358 arch/powerpc/include/asm/nohash/64/pgtable.h int map_kernel_page(unsigned long ea, unsigned long pa, pgprot_t prot); prot 240 arch/powerpc/include/asm/nohash/pgtable.h #define pgprot_noncached(prot) (__pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | \ prot 243 arch/powerpc/include/asm/nohash/pgtable.h #define pgprot_noncached_wc(prot) (__pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | \ prot 246 arch/powerpc/include/asm/nohash/pgtable.h #define pgprot_cached(prot) (__pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | \ prot 250 arch/powerpc/include/asm/nohash/pgtable.h #define pgprot_cached_wthru(prot) (__pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | \ prot 253 arch/powerpc/include/asm/nohash/pgtable.h #define pgprot_cached_wthru(prot) pgprot_noncached(prot) prot 256 arch/powerpc/include/asm/nohash/pgtable.h #define pgprot_cached_noncoherent(prot) \ prot 257 arch/powerpc/include/asm/nohash/pgtable.h (__pgprot(pgprot_val(prot) & ~_PAGE_CACHE_CTL)) prot 113 arch/powerpc/include/asm/pci.h pgprot_t prot); prot 160 arch/powerpc/include/asm/pkeys.h int prot, int pkey); prot 162 arch/powerpc/include/asm/pkeys.h int prot, int pkey) prot 174 arch/powerpc/include/asm/pkeys.h return __arch_override_mprotect_pkey(vma, prot, pkey); prot 13 arch/powerpc/include/asm/syscalls.h unsigned long prot, unsigned long flags, prot 16 arch/powerpc/include/asm/syscalls.h unsigned long prot, unsigned long flags, prot 153 arch/powerpc/kernel/io-workarounds.c pgprot_t prot, void *caller) prot 156 arch/powerpc/kernel/io-workarounds.c void __iomem *res = __ioremap_caller(addr, size, prot, caller); prot 440 arch/powerpc/kernel/pci-common.c pgprot_t prot) prot 448 arch/powerpc/kernel/pci-common.c return prot; prot 450 arch/powerpc/kernel/pci-common.c prot = pgprot_noncached(prot); prot 471 arch/powerpc/kernel/pci-common.c prot = pgprot_noncached_wc(prot); prot 476 arch/powerpc/kernel/pci-common.c (unsigned long long)offset, pgprot_val(prot)); prot 478 arch/powerpc/kernel/pci-common.c return prot; prot 52 arch/powerpc/kernel/sys_ppc32.c unsigned long prot, unsigned long flags, prot 56 arch/powerpc/kernel/sys_ppc32.c return sys_mmap(addr, len, prot, flags, fd, pgoff << 12); prot 41 arch/powerpc/kernel/syscalls.c unsigned long prot, unsigned long flags, prot 46 arch/powerpc/kernel/syscalls.c if (!arch_validate_prot(prot, addr)) prot 55 arch/powerpc/kernel/syscalls.c ret = ksys_mmap_pgoff(addr, len, prot, flags, fd, off); prot 61 arch/powerpc/kernel/syscalls.c unsigned long, prot, unsigned long, flags, prot 64 arch/powerpc/kernel/syscalls.c return do_mmap2(addr, len, prot, flags, fd, pgoff, PAGE_SHIFT-12); prot 68 arch/powerpc/kernel/syscalls.c unsigned long, prot, unsigned long, flags, prot 71 arch/powerpc/kernel/syscalls.c return do_mmap2(addr, len, prot, flags, fd, offset, PAGE_SHIFT); prot 123 arch/powerpc/mm/book3s32/mmu.c unsigned int size, pgprot_t prot) prot 128 arch/powerpc/mm/book3s32/mmu.c unsigned long flags = pgprot_val(prot); prot 250 arch/powerpc/mm/book3s32/mmu.c unsigned int size, pgprot_t prot) prot 255 arch/powerpc/mm/book3s32/mmu.c unsigned long flags = pgprot_val(prot); prot 148 arch/powerpc/mm/book3s64/hash_pgtable.c int hash__map_kernel_page(unsigned long ea, unsigned long pa, pgprot_t prot) prot 167 arch/powerpc/mm/book3s64/hash_pgtable.c set_pte_at(&init_mm, ea, ptep, pfn_pte(pa >> PAGE_SHIFT, prot)); prot 175 arch/powerpc/mm/book3s64/hash_pgtable.c if (htab_bolt_mapping(ea, ea + PAGE_SIZE, pa, pgprot_val(prot), prot 245 arch/powerpc/mm/book3s64/hash_utils.c unsigned long pstart, unsigned long prot, prot 255 arch/powerpc/mm/book3s64/hash_utils.c prot = htab_convert_pte_flags(prot); prot 258 arch/powerpc/mm/book3s64/hash_utils.c vstart, vend, pstart, prot, psize, ssize); prot 265 arch/powerpc/mm/book3s64/hash_utils.c unsigned long tprot = prot; prot 840 arch/powerpc/mm/book3s64/hash_utils.c unsigned long prot; prot 916 arch/powerpc/mm/book3s64/hash_utils.c prot = pgprot_val(PAGE_KERNEL); prot 936 arch/powerpc/mm/book3s64/hash_utils.c base, size, prot); prot 944 arch/powerpc/mm/book3s64/hash_utils.c prot, mmu_linear_psize, mmu_kernel_ssize)); prot 963 arch/powerpc/mm/book3s64/hash_utils.c __pa(tce_alloc_start), prot, prot 337 arch/powerpc/mm/book3s64/pkeys.c int __arch_override_mprotect_pkey(struct vm_area_struct *vma, int prot, prot 344 arch/powerpc/mm/book3s64/pkeys.c if (vma_is_pkey_exec_only(vma) && (prot != PROT_EXEC)) prot 351 arch/powerpc/mm/book3s64/pkeys.c if (prot == PROT_EXEC) { prot 259 arch/powerpc/mm/book3s64/radix_pgtable.c pgprot_t prot; prot 288 arch/powerpc/mm/book3s64/radix_pgtable.c prot = PAGE_KERNEL_X; prot 291 arch/powerpc/mm/book3s64/radix_pgtable.c prot = PAGE_KERNEL; prot 300 arch/powerpc/mm/book3s64/radix_pgtable.c rc = __map_kernel_page(vaddr, addr, prot, mapping_size, nid, start, end); prot 1100 arch/powerpc/mm/book3s64/radix_pgtable.c int pud_set_huge(pud_t *pud, phys_addr_t addr, pgprot_t prot) prot 1103 arch/powerpc/mm/book3s64/radix_pgtable.c pte_t new_pud = pfn_pte(__phys_to_pfn(addr), prot); prot 1147 arch/powerpc/mm/book3s64/radix_pgtable.c int pmd_set_huge(pmd_t *pmd, phys_addr_t addr, pgprot_t prot) prot 1150 arch/powerpc/mm/book3s64/radix_pgtable.c pte_t new_pmd = pfn_pte(__phys_to_pfn(addr), prot); prot 33 arch/powerpc/mm/highmem.c void *kmap_atomic_prot(struct page *page, pgprot_t prot) prot 47 arch/powerpc/mm/highmem.c __set_pte_at(&init_mm, vaddr, kmap_pte-idx, mk_pte(page, prot), 1); prot 13 arch/powerpc/mm/ioremap.c pgprot_t prot = pgprot_noncached(PAGE_KERNEL); prot 17 arch/powerpc/mm/ioremap.c return iowa_ioremap(addr, size, prot, caller); prot 18 arch/powerpc/mm/ioremap.c return __ioremap_caller(addr, size, prot, caller); prot 24 arch/powerpc/mm/ioremap.c pgprot_t prot = pgprot_noncached_wc(PAGE_KERNEL); prot 28 arch/powerpc/mm/ioremap.c return iowa_ioremap(addr, size, prot, caller); prot 29 arch/powerpc/mm/ioremap.c return __ioremap_caller(addr, size, prot, caller); prot 35 arch/powerpc/mm/ioremap.c pgprot_t prot = pgprot_cached(PAGE_KERNEL); prot 39 arch/powerpc/mm/ioremap.c return iowa_ioremap(addr, size, prot, caller); prot 40 arch/powerpc/mm/ioremap.c return __ioremap_caller(addr, size, prot, caller); prot 63 arch/powerpc/mm/ioremap.c unsigned long size, pgprot_t prot) prot 68 arch/powerpc/mm/ioremap.c int err = map_kernel_page(ea + i, pa + i, prot); prot 78 arch/powerpc/mm/ioremap.c pgprot_t prot, void *caller) prot 91 arch/powerpc/mm/ioremap.c ret = ioremap_page_range(va, va + size, pa, prot); prot 11 arch/powerpc/mm/ioremap_32.c pgprot_t prot = pgprot_cached_wthru(PAGE_KERNEL); prot 13 arch/powerpc/mm/ioremap_32.c return __ioremap_caller(addr, size, prot, __builtin_return_address(0)); prot 18 arch/powerpc/mm/ioremap_32.c __ioremap_caller(phys_addr_t addr, unsigned long size, pgprot_t prot, void *caller) prot 66 arch/powerpc/mm/ioremap_32.c return do_ioremap(p, offset, size, prot, caller); prot 72 arch/powerpc/mm/ioremap_32.c err = early_ioremap_range(ioremap_bot - size, p, size, prot); prot 10 arch/powerpc/mm/ioremap_64.c void __iomem *__ioremap_at(phys_addr_t pa, void *ea, unsigned long size, pgprot_t prot) prot 16 arch/powerpc/mm/ioremap_64.c if (pgprot_val(prot) & H_PAGE_4K_PFN) prot 29 arch/powerpc/mm/ioremap_64.c ret = ioremap_page_range(va, va + size, pa, prot); prot 33 arch/powerpc/mm/ioremap_64.c ret = early_ioremap_range(va, pa, size, prot); prot 58 arch/powerpc/mm/ioremap_64.c pgprot_t prot, void *caller) prot 65 arch/powerpc/mm/ioremap_64.c if (pgprot_val(prot) & H_PAGE_4K_PFN) prot 82 arch/powerpc/mm/ioremap_64.c return do_ioremap(paligned, offset, size, prot, caller); prot 84 arch/powerpc/mm/ioremap_64.c err = early_ioremap_range(ioremap_bot, paligned, size, prot); prot 23 arch/powerpc/mm/kasan/kasan_init_32.c static void kasan_populate_pte(pte_t *ptep, pgprot_t prot) prot 30 arch/powerpc/mm/kasan/kasan_init_32.c __set_pte_at(&init_mm, va, ptep, pfn_pte(PHYS_PFN(pa), prot), 0); prot 37 arch/powerpc/mm/kasan/kasan_init_32.c pgprot_t prot = slab_is_available() ? kasan_prot_ro() : PAGE_KERNEL; prot 55 arch/powerpc/mm/kasan/kasan_init_32.c kasan_populate_pte(new, prot); prot 112 arch/powerpc/mm/kasan/kasan_init_32.c pgprot_t prot = kasan_prot_ro(); prot 118 arch/powerpc/mm/kasan/kasan_init_32.c kasan_populate_pte(kasan_early_shadow_pte, prot); prot 127 arch/powerpc/mm/kasan/kasan_init_32.c __set_pte_at(&init_mm, k_cur, ptep, pfn_pte(PHYS_PFN(pa), prot), 0); prot 98 arch/powerpc/mm/mmu_decl.h unsigned int size, pgprot_t prot); prot 73 arch/powerpc/mm/nohash/book3e_pgtable.c int __ref map_kernel_page(unsigned long ea, unsigned long pa, pgprot_t prot) prot 112 arch/powerpc/mm/nohash/book3e_pgtable.c set_pte_at(&init_mm, ea, ptep, pfn_pte(pa >> PAGE_SHIFT, prot)); prot 59 arch/powerpc/mm/pgtable_32.c int __ref map_kernel_page(unsigned long va, phys_addr_t pa, pgprot_t prot) prot 77 arch/powerpc/mm/pgtable_32.c BUG_ON((pte_present(*pg) | pte_hashpte(*pg)) && pgprot_val(prot)); prot 78 arch/powerpc/mm/pgtable_32.c set_pte_at(&init_mm, va, pg, pfn_pte(pa >> PAGE_SHIFT, prot)); prot 161 arch/powerpc/mm/pgtable_32.c static int __change_page_attr_noflush(struct page *page, pgprot_t prot) prot 174 arch/powerpc/mm/pgtable_32.c __set_pte_at(&init_mm, address, kpte, mk_pte(page, prot), 0); prot 185 arch/powerpc/mm/pgtable_32.c static int change_page_attr(struct page *page, int numpages, pgprot_t prot) prot 193 arch/powerpc/mm/pgtable_32.c err = __change_page_attr_noflush(page, prot); prot 173 arch/powerpc/platforms/cell/iommu.c const unsigned long prot = 0xc48; prot 175 arch/powerpc/platforms/cell/iommu.c ((prot << (52 + 4 * direction)) & prot 41 arch/riscv/include/asm/fixmap.h phys_addr_t phys, pgprot_t prot); prot 68 arch/riscv/include/asm/pgtable-64.h static inline pmd_t pfn_pmd(unsigned long pfn, pgprot_t prot) prot 70 arch/riscv/include/asm/pgtable-64.h return __pmd((pfn << _PAGE_PFN_SHIFT) | pgprot_val(prot)); prot 148 arch/riscv/include/asm/pgtable.h static inline pgd_t pfn_pgd(unsigned long pfn, pgprot_t prot) prot 150 arch/riscv/include/asm/pgtable.h return __pgd((pfn << _PAGE_PFN_SHIFT) | pgprot_val(prot)); prot 187 arch/riscv/include/asm/pgtable.h static inline pte_t pfn_pte(unsigned long pfn, pgprot_t prot) prot 189 arch/riscv/include/asm/pgtable.h return __pte((pfn << _PAGE_PFN_SHIFT) | pgprot_val(prot)); prot 192 arch/riscv/include/asm/pgtable.h #define mk_pte(page, prot) pfn_pte(page_to_pfn(page), prot) prot 13 arch/riscv/kernel/sys_riscv.c unsigned long prot, unsigned long flags, prot 19 arch/riscv/kernel/sys_riscv.c return ksys_mmap_pgoff(addr, len, prot, flags, fd, prot 25 arch/riscv/kernel/sys_riscv.c unsigned long, prot, unsigned long, flags, prot 28 arch/riscv/kernel/sys_riscv.c return riscv_sys_mmap(addr, len, prot, flags, fd, offset, 0); prot 32 arch/riscv/kernel/sys_riscv.c unsigned long, prot, unsigned long, flags, prot 39 arch/riscv/kernel/sys_riscv.c return riscv_sys_mmap(addr, len, prot, flags, fd, offset, 12); prot 161 arch/riscv/mm/init.c void __set_fixmap(enum fixed_addresses idx, phys_addr_t phys, pgprot_t prot) prot 170 arch/riscv/mm/init.c if (pgprot_val(prot)) { prot 171 arch/riscv/mm/init.c set_pte(ptep, pfn_pte(phys >> PAGE_SHIFT, prot)); prot 201 arch/riscv/mm/init.c phys_addr_t sz, pgprot_t prot) prot 208 arch/riscv/mm/init.c ptep[pte_index] = pfn_pte(PFN_DOWN(pa), prot); prot 247 arch/riscv/mm/init.c phys_addr_t sz, pgprot_t prot) prot 255 arch/riscv/mm/init.c pmdp[pmd_index] = pfn_pmd(PFN_DOWN(pa), prot); prot 269 arch/riscv/mm/init.c create_pte_mapping(ptep, va, pa, sz, prot); prot 291 arch/riscv/mm/init.c phys_addr_t sz, pgprot_t prot) prot 299 arch/riscv/mm/init.c pgdp[pgd_index] = pfn_pgd(PFN_DOWN(pa), prot); prot 313 arch/riscv/mm/init.c create_pgd_next_mapping(nextp, va, pa, sz, prot); prot 24 arch/riscv/mm/ioremap.c pgprot_t prot, void *caller) prot 45 arch/riscv/mm/ioremap.c if (ioremap_page_range(vaddr, vaddr + size, addr, prot)) { prot 143 arch/s390/include/asm/gmap.h unsigned long len, int prot); prot 1145 arch/s390/include/asm/pgtable.h pte_t *ptep, int prot, unsigned long bit); prot 427 arch/s390/include/uapi/asm/ptrace.h ptprot_flags prot; prot 198 arch/s390/kernel/compat_linux.c compat_ulong_t prot; prot 212 arch/s390/kernel/compat_linux.c return ksys_mmap_pgoff(a.addr, a.len, a.prot, a.flags, a.fd, prot 222 arch/s390/kernel/compat_linux.c return ksys_mmap_pgoff(a.addr, a.len, a.prot, a.flags, a.fd, a.offset); prot 241 arch/s390/kernel/crash_dump.c unsigned long size, pgprot_t prot) prot 250 arch/s390/kernel/crash_dump.c size_old, prot); prot 257 arch/s390/kernel/crash_dump.c return remap_pfn_range(vma, from, pfn, size, prot); prot 269 arch/s390/kernel/crash_dump.c unsigned long size, pgprot_t prot) prot 282 arch/s390/kernel/crash_dump.c return remap_pfn_range(vma, from, pfn, size, prot); prot 289 arch/s390/kernel/crash_dump.c unsigned long pfn, unsigned long size, pgprot_t prot) prot 292 arch/s390/kernel/crash_dump.c return remap_oldmem_pfn_range_kdump(vma, from, pfn, size, prot); prot 295 arch/s390/kernel/crash_dump.c prot); prot 43 arch/s390/kernel/sys_s390.c unsigned long prot; prot 56 arch/s390/kernel/sys_s390.c error = ksys_mmap_pgoff(a.addr, a.len, a.prot, a.flags, a.fd, a.offset); prot 494 arch/s390/kvm/gaccess.c u8 ar, enum gacc_mode mode, enum prot_type prot) prot 505 arch/s390/kvm/gaccess.c switch (prot) { prot 616 arch/s390/kvm/gaccess.c enum gacc_mode mode, enum prot_type *prot) prot 765 arch/s390/kvm/gaccess.c *prot = PROT_TYPE_DAT; prot 769 arch/s390/kvm/gaccess.c *prot = PROT_TYPE_IEP; prot 803 arch/s390/kvm/gaccess.c enum prot_type prot; prot 813 arch/s390/kvm/gaccess.c rc = guest_translate(vcpu, ga, pages, asce, mode, &prot); prot 822 arch/s390/kvm/gaccess.c return trans_exc(vcpu, rc, ga, ar, mode, prot); prot 908 arch/s390/kvm/gaccess.c enum prot_type prot; prot 923 arch/s390/kvm/gaccess.c rc = guest_translate(vcpu, gva, gpa, asce, mode, &prot); prot 925 arch/s390/kvm/gaccess.c return trans_exc(vcpu, rc, gva, 0, mode, prot); prot 73 arch/s390/mm/dump_pagetables.c unsigned int prot, cur; prot 81 arch/s390/mm/dump_pagetables.c prot = new_prot; prot 90 arch/s390/mm/dump_pagetables.c } else if (prot != cur || level != st->level || prot 117 arch/s390/mm/dump_pagetables.c unsigned int prot; prot 119 arch/s390/mm/dump_pagetables.c prot = pte_val(*kasan_early_shadow_pte) & prot 121 arch/s390/mm/dump_pagetables.c note_page(m, st, prot, 4); prot 136 arch/s390/mm/dump_pagetables.c unsigned int prot; prot 143 arch/s390/mm/dump_pagetables.c prot = pte_val(*pte) & prot 145 arch/s390/mm/dump_pagetables.c note_page(m, st, prot, 4); prot 153 arch/s390/mm/dump_pagetables.c unsigned int prot; prot 169 arch/s390/mm/dump_pagetables.c prot = pmd_val(*pmd) & prot 172 arch/s390/mm/dump_pagetables.c note_page(m, st, prot, 3); prot 184 arch/s390/mm/dump_pagetables.c unsigned int prot; prot 200 arch/s390/mm/dump_pagetables.c prot = pud_val(*pud) & prot 203 arch/s390/mm/dump_pagetables.c note_page(m, st, prot, 2); prot 874 arch/s390/mm/gmap.c unsigned long vmaddr, int prot) prot 881 arch/s390/mm/gmap.c fault_flags = (prot == PROT_WRITE) ? FAULT_FLAG_WRITE : 0; prot 960 arch/s390/mm/gmap.c pmd_t *pmdp, int prot, unsigned long bits) prot 967 arch/s390/mm/gmap.c if ((pmd_i && (prot != PROT_NONE)) || (pmd_p && (prot == PROT_WRITE))) prot 970 arch/s390/mm/gmap.c if (prot == PROT_NONE && !pmd_i) { prot 975 arch/s390/mm/gmap.c if (prot == PROT_READ && !pmd_p) { prot 1005 arch/s390/mm/gmap.c pmd_t *pmdp, int prot, unsigned long bits) prot 1022 arch/s390/mm/gmap.c rc = ptep_force_prot(gmap->mm, gaddr, ptep, prot, pbits); prot 1041 arch/s390/mm/gmap.c unsigned long len, int prot, unsigned long bits) prot 1053 arch/s390/mm/gmap.c rc = gmap_protect_pte(gmap, gaddr, pmdp, prot, prot 1060 arch/s390/mm/gmap.c rc = gmap_protect_pmd(gmap, gaddr, pmdp, prot, prot 1078 arch/s390/mm/gmap.c rc = gmap_pte_op_fixup(gmap, gaddr, vmaddr, prot); prot 1101 arch/s390/mm/gmap.c unsigned long len, int prot) prot 1107 arch/s390/mm/gmap.c if (!MACHINE_HAS_ESOP && prot == PROT_READ) prot 1110 arch/s390/mm/gmap.c rc = gmap_protect_range(gmap, gaddr, len, prot, GMAP_NOTIFY_MPROT); prot 2112 arch/s390/mm/gmap.c int prot; prot 2117 arch/s390/mm/gmap.c prot = (pte_val(pte) & _PAGE_PROTECT) ? PROT_READ : PROT_WRITE; prot 2159 arch/s390/mm/gmap.c rc = gmap_pte_op_fixup(parent, paddr, vmaddr, prot); prot 112 arch/s390/mm/pageattr.c unsigned long pte_addr, prot; prot 123 arch/s390/mm/pageattr.c prot = pgprot_val(ro ? PAGE_KERNEL_RO : PAGE_KERNEL); prot 125 arch/s390/mm/pageattr.c prot &= ~_PAGE_NOEXEC; prot 128 arch/s390/mm/pageattr.c pte_val(*ptep) = pte_addr | prot; prot 189 arch/s390/mm/pageattr.c unsigned long pmd_addr, prot; prot 200 arch/s390/mm/pageattr.c prot = pgprot_val(ro ? SEGMENT_KERNEL_RO : SEGMENT_KERNEL); prot 202 arch/s390/mm/pageattr.c prot &= ~_SEGMENT_ENTRY_NOEXEC; prot 205 arch/s390/mm/pageattr.c pmd_val(*pmdp) = pmd_addr | prot; prot 596 arch/s390/mm/pgtable.c pte_t *ptep, int prot, unsigned long bit) prot 607 arch/s390/mm/pgtable.c if ((pte_i && (prot != PROT_NONE)) || prot 608 arch/s390/mm/pgtable.c (pte_p && (prot & PROT_WRITE))) { prot 614 arch/s390/mm/pgtable.c if (prot == PROT_NONE && !pte_i) { prot 619 arch/s390/mm/pgtable.c if (prot == PROT_READ && !pte_p) { prot 54 arch/sh/boards/mach-landisk/setup.c pgprot_t prot; prot 60 arch/sh/boards/mach-landisk/setup.c prot = PAGE_KERNEL_PCC(1, _PAGE_PCC_IO16); prot 61 arch/sh/boards/mach-landisk/setup.c cf_ide_base = ioremap_prot(paddrbase, PAGE_SIZE, pgprot_val(prot)); prot 49 arch/sh/boards/mach-lboxre2/setup.c pgprot_t prot; prot 55 arch/sh/boards/mach-lboxre2/setup.c prot = PAGE_KERNEL_PCC(1, _PAGE_PCC_IO16); prot 56 arch/sh/boards/mach-lboxre2/setup.c cf0_io_base = (u32)ioremap_prot(paddrbase, psize, pgprot_val(prot)); prot 71 arch/sh/boards/mach-sh03/setup.c pgprot_t prot; prot 77 arch/sh/boards/mach-sh03/setup.c prot = PAGE_KERNEL_PCC(1, _PAGE_PCC_IO16); prot 78 arch/sh/boards/mach-sh03/setup.c cf_ide_base = ioremap_prot(paddrbase, PAGE_SIZE, pgprot_val(prot)); prot 269 arch/sh/include/asm/io.h pgprot_t prot, void *caller); prot 273 arch/sh/include/asm/io.h __ioremap(phys_addr_t offset, unsigned long size, pgprot_t prot) prot 275 arch/sh/include/asm/io.h return __ioremap_caller(offset, size, prot, __builtin_return_address(0)); prot 279 arch/sh/include/asm/io.h __ioremap_29bit(phys_addr_t offset, unsigned long size, pgprot_t prot) prot 291 arch/sh/include/asm/io.h u64 flags = pgprot_val(prot); prot 314 arch/sh/include/asm/io.h __ioremap_mode(phys_addr_t offset, unsigned long size, pgprot_t prot) prot 322 arch/sh/include/asm/io.h ret = __ioremap_29bit(offset, size, prot); prot 326 arch/sh/include/asm/io.h return __ioremap(offset, size, prot); prot 329 arch/sh/include/asm/io.h #define __ioremap(offset, size, prot) ((void __iomem *)(offset)) prot 330 arch/sh/include/asm/io.h #define __ioremap_mode(offset, size, prot) ((void __iomem *)(offset)) prot 360 arch/sh/include/asm/io.h ioremap_fixed(phys_addr_t phys_addr, unsigned long size, pgprot_t prot) prot 63 arch/sh/include/asm/mmu.h unsigned long size, pgprot_t prot); prot 65 arch/sh/include/asm/mmu.h pgprot_t prot, void *caller); prot 72 arch/sh/include/asm/mmu.h unsigned long size, pgprot_t prot) prot 79 arch/sh/include/asm/mmu.h pgprot_t prot, void *caller) prot 100 arch/sh/include/asm/mmu.h pmb_remap(phys_addr_t phys, unsigned long size, pgprot_t prot) prot 102 arch/sh/include/asm/mmu.h return pmb_remap_caller(phys, size, prot, __builtin_return_address(0)); prot 147 arch/sh/include/asm/pgtable.h static inline bool __pte_access_permitted(pte_t pte, u64 prot) prot 149 arch/sh/include/asm/pgtable.h return (pte_val(pte) & (prot | _PAGE_SPECIAL)) == prot; prot 155 arch/sh/include/asm/pgtable.h u64 prot = _PAGE_PRESENT; prot 157 arch/sh/include/asm/pgtable.h prot |= _PAGE_EXT(_PAGE_EXT_KERN_READ | _PAGE_EXT_USER_READ); prot 159 arch/sh/include/asm/pgtable.h prot |= _PAGE_EXT(_PAGE_EXT_KERN_WRITE | _PAGE_EXT_USER_WRITE); prot 160 arch/sh/include/asm/pgtable.h return __pte_access_permitted(pte, prot); prot 165 arch/sh/include/asm/pgtable.h u64 prot = _PAGE_PRESENT | _PAGE_USER | _PAGE_READ; prot 168 arch/sh/include/asm/pgtable.h prot |= _PAGE_WRITE; prot 169 arch/sh/include/asm/pgtable.h return __pte_access_permitted(pte, prot); prot 174 arch/sh/include/asm/pgtable.h u64 prot = _PAGE_PRESENT | _PAGE_USER; prot 177 arch/sh/include/asm/pgtable.h prot |= _PAGE_RW; prot 178 arch/sh/include/asm/pgtable.h return __pte_access_permitted(pte, prot); prot 318 arch/sh/include/asm/pgtable_32.h #define pfn_pte(pfn, prot) \ prot 319 arch/sh/include/asm/pgtable_32.h __pte(((unsigned long long)(pfn) << PAGE_SHIFT) | pgprot_val(prot)) prot 320 arch/sh/include/asm/pgtable_32.h #define pfn_pmd(pfn, prot) \ prot 321 arch/sh/include/asm/pgtable_32.h __pmd(((unsigned long long)(pfn) << PAGE_SHIFT) | pgprot_val(prot)) prot 379 arch/sh/include/asm/pgtable_32.h #define pgprot_writecombine(prot) \ prot 380 arch/sh/include/asm/pgtable_32.h __pgprot(pgprot_val(prot) & ~_PAGE_CACHABLE) prot 204 arch/sh/include/asm/pgtable_64.h #define pgprot_writecombine(prot) __pgprot(pgprot_val(prot) & ~_PAGE_CACHABLE) prot 304 arch/sh/include/asm/pgtable_64.h #define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) prot 305 arch/sh/include/asm/pgtable_64.h #define pfn_pmd(pfn, prot) __pmd(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) prot 8 arch/sh/include/asm/syscalls.h unsigned long prot, unsigned long flags, prot 11 arch/sh/include/asm/syscalls.h unsigned long prot, unsigned long flags, prot 29 arch/sh/include/cpu-sh4/cpu/sq.h const char *name, pgprot_t prot); prot 101 arch/sh/kernel/cpu/sh4/sq.c static int __sq_remap(struct sq_mapping *map, pgprot_t prot) prot 114 arch/sh/kernel/cpu/sh4/sq.c vma->phys_addr, prot)) { prot 143 arch/sh/kernel/cpu/sh4/sq.c const char *name, pgprot_t prot) prot 178 arch/sh/kernel/cpu/sh4/sq.c ret = __sq_remap(map, prot); prot 33 arch/sh/kernel/sys_sh.c unsigned long prot, unsigned long flags, prot 38 arch/sh/kernel/sys_sh.c return ksys_mmap_pgoff(addr, len, prot, flags, fd, off>>PAGE_SHIFT); prot 42 arch/sh/kernel/sys_sh.c unsigned long prot, unsigned long flags, prot 54 arch/sh/kernel/sys_sh.c return ksys_mmap_pgoff(addr, len, prot, flags, fd, pgoff); prot 72 arch/sh/mm/init.c static void set_pte_phys(unsigned long addr, unsigned long phys, pgprot_t prot) prot 82 arch/sh/mm/init.c set_pte(pte, pfn_pte(phys >> PAGE_SHIFT, prot)); prot 85 arch/sh/mm/init.c if (pgprot_val(prot) & _PAGE_WIRED) prot 89 arch/sh/mm/init.c static void clear_pte_phys(unsigned long addr, pgprot_t prot) prot 95 arch/sh/mm/init.c if (pgprot_val(prot) & _PAGE_WIRED) prot 102 arch/sh/mm/init.c void __set_fixmap(enum fixed_addresses idx, unsigned long phys, pgprot_t prot) prot 111 arch/sh/mm/init.c set_pte_phys(address, phys, prot); prot 114 arch/sh/mm/init.c void __clear_fixmap(enum fixed_addresses idx, pgprot_t prot) prot 123 arch/sh/mm/init.c clear_pte_phys(address, prot); prot 48 arch/sh/mm/ioremap_fixed.c ioremap_fixed(phys_addr_t phys_addr, unsigned long size, pgprot_t prot) prot 89 arch/sh/mm/ioremap_fixed.c pgprot_val(prot) |= _PAGE_WIRED; prot 90 arch/sh/mm/ioremap_fixed.c __set_fixmap(idx, phys_addr, prot); prot 97 arch/sh/mm/nommu.c void __set_fixmap(enum fixed_addresses idx, unsigned long phys, pgprot_t prot) prot 117 arch/sh/mm/pmb.c static inline unsigned long pgprot_to_pmb_flags(pgprot_t prot) prot 120 arch/sh/mm/pmb.c u64 flags = pgprot_val(prot); prot 207 arch/sh/mm/pmb.c static inline bool pmb_prot_valid(pgprot_t prot) prot 209 arch/sh/mm/pmb.c return (pgprot_val(prot) & _PAGE_USER) == 0; prot 336 arch/sh/mm/pmb.c unsigned long size, pgprot_t prot) prot 355 arch/sh/mm/pmb.c pmb_flags = pgprot_to_pmb_flags(prot); prot 411 arch/sh/mm/pmb.c pgprot_t prot, void *caller) prot 428 arch/sh/mm/pmb.c if (!pmb_prot_valid(prot)) prot 454 arch/sh/mm/pmb.c ret = pmb_bolt_mapping(vaddr, phys, size, prot); prot 30 arch/sparc/include/asm/mman.h #define arch_calc_vm_prot_bits(prot, pkey) sparc_calc_vm_prot_bits(prot) prot 31 arch/sparc/include/asm/mman.h static inline unsigned long sparc_calc_vm_prot_bits(unsigned long prot) prot 33 arch/sparc/include/asm/mman.h if (adi_capable() && (prot & PROT_ADI)) { prot 55 arch/sparc/include/asm/mman.h #define arch_validate_prot(prot, addr) sparc_validate_prot(prot, addr) prot 56 arch/sparc/include/asm/mman.h static inline int sparc_validate_prot(unsigned long prot, unsigned long addr) prot 58 arch/sparc/include/asm/mman.h if (prot & ~(PROT_READ | PROT_WRITE | PROT_EXEC | PROT_SEM | PROT_ADI)) prot 60 arch/sparc/include/asm/mman.h if (prot & PROT_ADI) { prot 263 arch/sparc/include/asm/pgtable_32.h #define pfn_pte(pfn, prot) mk_pte(pfn_to_page(pfn), prot) prot 300 arch/sparc/include/asm/pgtable_32.h static inline pgprot_t pgprot_noncached(pgprot_t prot) prot 302 arch/sparc/include/asm/pgtable_32.h pgprot_val(prot) &= ~pgprot_val(__pgprot(SRMMU_CACHE)); prot 303 arch/sparc/include/asm/pgtable_32.h return prot; prot 415 arch/sparc/include/asm/pgtable_32.h unsigned long size, pgprot_t prot) prot 423 arch/sparc/include/asm/pgtable_32.h return remap_pfn_range(vma, from, phys_base >> PAGE_SHIFT, size, prot); prot 239 arch/sparc/include/asm/pgtable_64.h static inline pte_t pfn_pte(unsigned long pfn, pgprot_t prot) prot 244 arch/sparc/include/asm/pgtable_64.h return __pte(paddr | pgprot_val(prot)); prot 280 arch/sparc/include/asm/pgtable_64.h static inline pte_t pte_modify(pte_t pte, pgprot_t prot) prot 335 arch/sparc/include/asm/pgtable_64.h return __pte((pte_val(pte) & mask) | (pgprot_val(prot) & ~mask)); prot 349 arch/sparc/include/asm/pgtable_64.h static inline pgprot_t pgprot_noncached(pgprot_t prot) prot 351 arch/sparc/include/asm/pgtable_64.h unsigned long val = pgprot_val(prot); prot 968 arch/sparc/include/asm/pgtable_64.h #define move_pte(pte, prot, old_addr, new_addr) \ prot 1069 arch/sparc/include/asm/pgtable_64.h unsigned long size, pgprot_t prot) prot 1077 arch/sparc/include/asm/pgtable_64.h return remap_pfn_range(vma, from, phys_base >> PAGE_SHIFT, size, prot); prot 1106 arch/sparc/include/asm/pgtable_64.h u64 prot; prot 1109 arch/sparc/include/asm/pgtable_64.h prot = _PAGE_PRESENT_4V | _PAGE_P_4V; prot 1111 arch/sparc/include/asm/pgtable_64.h prot |= _PAGE_WRITE_4V; prot 1113 arch/sparc/include/asm/pgtable_64.h prot = _PAGE_PRESENT_4U | _PAGE_P_4U; prot 1115 arch/sparc/include/asm/pgtable_64.h prot |= _PAGE_WRITE_4U; prot 1118 arch/sparc/include/asm/pgtable_64.h return (pte_val(pte) & (prot | _PAGE_SPECIAL)) == prot; prot 437 arch/sparc/kernel/iommu.c unsigned long flags, handle, prot, ctx; prot 460 arch/sparc/kernel/iommu.c prot = IOPTE_STREAMING(ctx); prot 462 arch/sparc/kernel/iommu.c prot = IOPTE_CONSISTENT(ctx); prot 464 arch/sparc/kernel/iommu.c prot |= IOPTE_WRITE; prot 512 arch/sparc/kernel/iommu.c iopte_val(*base) = prot | paddr; prot 56 arch/sparc/kernel/pci_sun4v.c unsigned long prot; /* IOMMU page protections */ prot 66 arch/sparc/kernel/pci_sun4v.c static inline void iommu_batch_start(struct device *dev, unsigned long prot, unsigned long entry) prot 71 arch/sparc/kernel/pci_sun4v.c p->prot = prot; prot 88 arch/sparc/kernel/pci_sun4v.c unsigned long prot = p->prot; prot 97 arch/sparc/kernel/pci_sun4v.c prot &= (HV_PCI_MAP_ATTR_READ | HV_PCI_MAP_ATTR_WRITE); prot 104 arch/sparc/kernel/pci_sun4v.c prot, prot 111 arch/sparc/kernel/pci_sun4v.c npages, prot, __pa(pglist), prot 121 arch/sparc/kernel/pci_sun4v.c prot, prot 128 arch/sparc/kernel/pci_sun4v.c index_count, prot, prot 185 arch/sparc/kernel/pci_sun4v.c unsigned long prot = 0; prot 201 arch/sparc/kernel/pci_sun4v.c prot = HV_PCI_MAP_ATTR_RELAXED_ORDER; prot 231 arch/sparc/kernel/pci_sun4v.c (HV_PCI_MAP_ATTR_READ | prot | prot 364 arch/sparc/kernel/pci_sun4v.c unsigned long prot; prot 393 arch/sparc/kernel/pci_sun4v.c prot = HV_PCI_MAP_ATTR_READ; prot 395 arch/sparc/kernel/pci_sun4v.c prot |= HV_PCI_MAP_ATTR_WRITE; prot 398 arch/sparc/kernel/pci_sun4v.c prot |= HV_PCI_MAP_ATTR_RELAXED_ORDER; prot 402 arch/sparc/kernel/pci_sun4v.c iommu_batch_start(dev, prot, entry); prot 472 arch/sparc/kernel/pci_sun4v.c unsigned long flags, handle, prot; prot 491 arch/sparc/kernel/pci_sun4v.c prot = HV_PCI_MAP_ATTR_READ; prot 493 arch/sparc/kernel/pci_sun4v.c prot |= HV_PCI_MAP_ATTR_WRITE; prot 496 arch/sparc/kernel/pci_sun4v.c prot |= HV_PCI_MAP_ATTR_RELAXED_ORDER; prot 508 arch/sparc/kernel/pci_sun4v.c iommu_batch_start(dev, prot, ~0UL); prot 102 arch/sparc/kernel/sys_sparc_32.c unsigned long, prot, unsigned long, flags, unsigned long, fd, prot 107 arch/sparc/kernel/sys_sparc_32.c return ksys_mmap_pgoff(addr, len, prot, flags, fd, prot 112 arch/sparc/kernel/sys_sparc_32.c unsigned long, prot, unsigned long, flags, unsigned long, fd, prot 116 arch/sparc/kernel/sys_sparc_32.c return ksys_mmap_pgoff(addr, len, prot, flags, fd, off >> PAGE_SHIFT); prot 120 arch/sparc/kernel/sys_sparc_32.c unsigned long, prot, unsigned long, pgoff, prot 126 arch/sparc/kernel/sys_sparc_32.c return sys_remap_file_pages(start, size, prot, prot 456 arch/sparc/kernel/sys_sparc_64.c unsigned long, prot, unsigned long, flags, unsigned long, fd, prot 465 arch/sparc/kernel/sys_sparc_64.c retval = ksys_mmap_pgoff(addr, len, prot, flags, fd, off >> PAGE_SHIFT); prot 18 arch/sparc/kernel/systbls.h unsigned long prot, unsigned long flags, prot 24 arch/sparc/kernel/systbls.h unsigned long prot, unsigned long flags, prot 27 arch/sparc/kernel/systbls.h unsigned long prot, unsigned long pgoff, prot 1787 arch/sparc/mm/init_64.c unsigned long pend, pgprot_t prot, prot 1855 arch/sparc/mm/init_64.c pte_val(*pte) = (paddr | pgprot_val(prot)); prot 2805 arch/sparc/mm/init_64.c pte_t mk_pte_io(unsigned long page, pgprot_t prot, int space, unsigned long page_size) prot 2809 arch/sparc/mm/init_64.c pte_val(pte) = page | pgprot_val(pgprot_noncached(prot)); prot 159 arch/sparc/vdso/vdso2c.c static void map_input(const char *name, void **addr, size_t *len, int prot) prot 173 arch/sparc/vdso/vdso2c.c *addr = mmap(NULL, tmp_len, prot, MAP_PRIVATE, fd, 0); prot 47 arch/um/include/asm/page.h #define pte_set_val(p, phys, prot) \ prot 48 arch/um/include/asm/page.h ({ (p).pte = (phys) | pgprot_val(prot); }) prot 74 arch/um/include/asm/page.h #define pte_set_val(p, phys, prot) (p).pte = (phys | pgprot_val(prot)) prot 42 arch/um/include/asm/pgtable-2level.h #define pfn_pte(pfn, prot) __pte(pfn_to_phys(pfn) | pgprot_val(prot)) prot 43 arch/um/include/asm/pgtable-2level.h #define pfn_pmd(pfn, prot) __pmd(pfn_to_phys(pfn) | pgprot_val(prot)) prot 275 arch/um/include/shared/os.h unsigned long len, int prot, int phys_fd, prot 280 arch/um/include/shared/os.h unsigned long len, unsigned int prot, int done, void **data); prot 18 arch/um/kernel/syscall.c unsigned long prot, unsigned long flags, prot 25 arch/um/kernel/syscall.c err = ksys_mmap_pgoff(addr, len, prot, flags, fd, offset >> PAGE_SHIFT); prot 25 arch/um/kernel/tlb.c unsigned int prot; prot 36 arch/um/kernel/tlb.c unsigned int prot; prot 75 arch/um/kernel/tlb.c op->u.mmap.len, op->u.mmap.prot, prot 100 arch/um/kernel/tlb.c op->u.mprotect.prot, prot 123 arch/um/kernel/tlb.c unsigned int prot, struct host_vm_change *hvc) prot 137 arch/um/kernel/tlb.c (last->u.mmap.prot == prot) && (last->u.mmap.fd == fd) && prot 153 arch/um/kernel/tlb.c .prot = prot, prot 191 arch/um/kernel/tlb.c unsigned int prot, struct host_vm_change *hvc) prot 200 arch/um/kernel/tlb.c (last->u.mprotect.prot == prot)) { prot 215 arch/um/kernel/tlb.c .prot = prot } } }); prot 226 arch/um/kernel/tlb.c int r, w, x, prot, ret = 0; prot 242 arch/um/kernel/tlb.c prot = ((r ? UM_PROT_READ : 0) | (w ? UM_PROT_WRITE : 0) | prot 248 arch/um/kernel/tlb.c PAGE_SIZE, prot, hvc); prot 252 arch/um/kernel/tlb.c ret = add_mprotect(addr, PAGE_SIZE, prot, hvc); prot 432 arch/um/kernel/tlb.c int r, w, x, prot, err = 0; prot 461 arch/um/kernel/tlb.c prot = ((r ? UM_PROT_READ : 0) | (w ? UM_PROT_WRITE : 0) | prot 469 arch/um/kernel/tlb.c err = map(mm_id, address, PAGE_SIZE, prot, fd, offset, prot 475 arch/um/kernel/tlb.c err = protect(mm_id, address, PAGE_SIZE, prot, 1, &flush); prot 142 arch/um/os-Linux/process.c int prot; prot 144 arch/um/os-Linux/process.c prot = (r ? PROT_READ : 0) | (w ? PROT_WRITE : 0) | prot 147 arch/um/os-Linux/process.c loc = mmap64((void *) virt, len, prot, MAP_SHARED | MAP_FIXED, prot 156 arch/um/os-Linux/process.c int prot = ((r ? PROT_READ : 0) | (w ? PROT_WRITE : 0) | prot 159 arch/um/os-Linux/process.c if (mprotect(addr, len, prot) < 0) prot 170 arch/um/os-Linux/skas/mem.c int map(struct mm_id * mm_idp, unsigned long virt, unsigned long len, int prot, prot 174 arch/um/os-Linux/skas/mem.c unsigned long args[] = { virt, len, prot, prot 198 arch/um/os-Linux/skas/mem.c unsigned int prot, int done, void **data) prot 201 arch/um/os-Linux/skas/mem.c unsigned long args[] = { addr, len, prot, 0, 0, 0 }; prot 151 arch/unicore32/include/asm/pgtable.h #define pfn_pte(pfn, prot) (__pte(((pfn) << PAGE_SHIFT) \ prot 152 arch/unicore32/include/asm/pgtable.h | pgprot_val(prot))) prot 197 arch/unicore32/include/asm/pgtable.h #define pgprot_noncached(prot) \ prot 198 arch/unicore32/include/asm/pgtable.h __pgprot(pgprot_val(prot) & ~PTE_CACHEABLE) prot 199 arch/unicore32/include/asm/pgtable.h #define pgprot_writecombine(prot) \ prot 200 arch/unicore32/include/asm/pgtable.h __pgprot(pgprot_val(prot) & ~PTE_CACHEABLE) prot 226 arch/unicore32/include/asm/pgtable.h #define mk_pte(page, prot) pfn_pte(page_to_pfn(page), prot) prot 142 arch/unicore32/mm/mmu.c unsigned long prot) prot 152 arch/unicore32/mm/mmu.c __pmd_populate(pmd, __pa(pte) | prot); prot 181 arch/x86/entry/vdso/vdso2c.c static void map_input(const char *name, void **addr, size_t *len, int prot) prot 194 arch/x86/entry/vdso/vdso2c.c *addr = mmap(NULL, tmp_len, prot, MAP_PRIVATE, fd, 0); prot 285 arch/x86/events/intel/ds.c static void ds_update_cea(void *cea, void *addr, size_t size, pgprot_t prot) prot 295 arch/x86/events/intel/ds.c cea_set_pte(cea, pa, prot); prot 153 arch/x86/ia32/sys_ia32.c unsigned int prot; prot 169 arch/x86/ia32/sys_ia32.c return ksys_mmap_pgoff(a.addr, a.len, a.prot, a.flags, a.fd, prot 12 arch/x86/include/asm/fb.h unsigned long prot; prot 14 arch/x86/include/asm/fb.h prot = pgprot_val(vma->vm_page_prot) & ~_PAGE_CACHE_MASK; prot 17 arch/x86/include/asm/fb.h prot | cachemode2protval(_PAGE_CACHE_MODE_UC_MINUS); prot 67 arch/x86/include/asm/highmem.h void *kmap_atomic_prot(struct page *page, pgprot_t prot); prot 71 arch/x86/include/asm/highmem.h void *kmap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot); prot 17 arch/x86/include/asm/iomap.h iomap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot); prot 23 arch/x86/include/asm/iomap.h iomap_create_wc(resource_size_t base, unsigned long size, pgprot_t *prot); prot 12 arch/x86/include/asm/pgtable.h #define pgprot_noncached(prot) \ prot 14 arch/x86/include/asm/pgtable.h ? (__pgprot(pgprot_val(prot) | \ prot 16 arch/x86/include/asm/pgtable.h : (prot)) prot 21 arch/x86/include/asm/pgtable.h #define pgprot_encrypted(prot) __pgprot(__sme_set(pgprot_val(prot))) prot 22 arch/x86/include/asm/pgtable.h #define pgprot_decrypted(prot) __pgprot(__sme_clr(pgprot_val(prot))) prot 647 arch/x86/include/asm/pgtable.h static inline pgprot_t arch_filter_pgprot(pgprot_t prot) prot 649 arch/x86/include/asm/pgtable.h return canon_pgprot(prot); prot 1463 arch/x86/include/asm/pgtable.h extern bool pfn_modify_allowed(unsigned long pfn, pgprot_t prot); prot 515 arch/x86/include/asm/pgtable_types.h extern pgprot_t pgprot_writecombine(pgprot_t prot); prot 518 arch/x86/include/asm/pgtable_types.h extern pgprot_t pgprot_writethrough(pgprot_t prot); prot 31 arch/x86/include/asm/pkeys.h int prot, int pkey); prot 33 arch/x86/include/asm/pkeys.h int prot, int pkey) prot 38 arch/x86/include/asm/pkeys.h return __arch_override_mprotect_pkey(vma, prot, pkey); prot 22 arch/x86/include/uapi/asm/mman.h #define arch_calc_vm_prot_bits(prot, key) ( \ prot 126 arch/x86/kernel/machine_kexec_64.c pgprot_t prot = PAGE_KERNEL_EXEC_NOENC; prot 171 arch/x86/kernel/machine_kexec_64.c prot = PAGE_KERNEL_EXEC; prot 173 arch/x86/kernel/machine_kexec_64.c set_pte(pte, pfn_pte(paddr >> PAGE_SHIFT, prot)); prot 92 arch/x86/kernel/sys_x86_64.c unsigned long, prot, unsigned long, flags, prot 100 arch/x86/kernel/sys_x86_64.c error = ksys_mmap_pgoff(addr, len, prot, flags, fd, off >> PAGE_SHIFT); prot 105 arch/x86/kernel/tboot.c pgprot_t prot) prot 126 arch/x86/kernel/tboot.c set_pte_at(&tboot_mm, vaddr, pte, pfn_pte(pfn, prot)); prot 49 arch/x86/mm/cpu_entry_area.c cea_map_percpu_pages(void *cea_vaddr, void *ptr, int pages, pgprot_t prot) prot 52 arch/x86/mm/cpu_entry_area.c cea_set_pte(cea_vaddr, per_cpu_ptr_to_phys(ptr), prot); prot 177 arch/x86/mm/dump_pagetables.c static void printk_prot(struct seq_file *m, pgprot_t prot, int level, bool dmsg) prot 179 arch/x86/mm/dump_pagetables.c pgprotval_t pr = pgprot_val(prot); prot 271 arch/x86/mm/dump_pagetables.c pgprotval_t prot, cur, eff; prot 279 arch/x86/mm/dump_pagetables.c prot = pgprot_val(new_prot); prot 292 arch/x86/mm/dump_pagetables.c } else if (prot != cur || new_eff != eff || level != st->level || prot 362 arch/x86/mm/dump_pagetables.c pgprotval_t prot, eff; prot 367 arch/x86/mm/dump_pagetables.c prot = pte_flags(*pte); prot 368 arch/x86/mm/dump_pagetables.c eff = effective_prot(eff_in, prot); prot 369 arch/x86/mm/dump_pagetables.c note_page(m, st, __pgprot(prot), eff, 5); prot 389 arch/x86/mm/dump_pagetables.c pgprotval_t prot = pte_flags(kasan_early_shadow_pte[0]); prot 390 arch/x86/mm/dump_pagetables.c note_page(m, st, __pgprot(prot), 0, 5); prot 410 arch/x86/mm/dump_pagetables.c pgprotval_t prot, eff; prot 416 arch/x86/mm/dump_pagetables.c prot = pmd_flags(*start); prot 417 arch/x86/mm/dump_pagetables.c eff = effective_prot(eff_in, prot); prot 419 arch/x86/mm/dump_pagetables.c note_page(m, st, __pgprot(prot), eff, 4); prot 443 arch/x86/mm/dump_pagetables.c pgprotval_t prot, eff; prot 450 arch/x86/mm/dump_pagetables.c prot = pud_flags(*start); prot 451 arch/x86/mm/dump_pagetables.c eff = effective_prot(eff_in, prot); prot 453 arch/x86/mm/dump_pagetables.c note_page(m, st, __pgprot(prot), eff, 3); prot 476 arch/x86/mm/dump_pagetables.c pgprotval_t prot, eff; prot 486 arch/x86/mm/dump_pagetables.c prot = p4d_flags(*start); prot 487 arch/x86/mm/dump_pagetables.c eff = effective_prot(eff_in, prot); prot 489 arch/x86/mm/dump_pagetables.c note_page(m, st, __pgprot(prot), eff, 2); prot 522 arch/x86/mm/dump_pagetables.c pgprotval_t prot, eff; prot 538 arch/x86/mm/dump_pagetables.c prot = pgd_flags(*start); prot 542 arch/x86/mm/dump_pagetables.c eff = prot; prot 545 arch/x86/mm/dump_pagetables.c note_page(m, &st, __pgprot(prot), eff, 1); prot 34 arch/x86/mm/highmem_32.c void *kmap_atomic_prot(struct page *page, pgprot_t prot) prot 49 arch/x86/mm/highmem_32.c set_pte(kmap_pte-idx, mk_pte(page, prot)); prot 317 arch/x86/mm/init_32.c pgprot_t prot = PAGE_KERNEL_LARGE; prot 332 arch/x86/mm/init_32.c prot = PAGE_KERNEL_LARGE_EXEC; prot 338 arch/x86/mm/init_32.c set_pmd(pmd, pfn_pmd(pfn, prot)); prot 349 arch/x86/mm/init_32.c pgprot_t prot = PAGE_KERNEL; prot 357 arch/x86/mm/init_32.c prot = PAGE_KERNEL_EXEC; prot 364 arch/x86/mm/init_32.c set_pte(pte, pfn_pte(pfn, prot)); prot 367 arch/x86/mm/init_64.c pgprot_t prot; prot 369 arch/x86/mm/init_64.c pgprot_val(prot) = pgprot_val(PAGE_KERNEL_LARGE) | prot 393 arch/x86/mm/init_64.c set_pmd(pmd, __pmd(phys | pgprot_val(prot))); prot 449 arch/x86/mm/init_64.c pgprot_t prot, bool init) prot 487 arch/x86/mm/init_64.c set_pte_init(pte, pfn_pte(paddr >> PAGE_SHIFT, prot), init); prot 503 arch/x86/mm/init_64.c unsigned long page_size_mask, pgprot_t prot, bool init) prot 513 arch/x86/mm/init_64.c pgprot_t new_prot = prot; prot 531 arch/x86/mm/init_64.c paddr_end, prot, prot 562 arch/x86/mm/init_64.c __pgprot(pgprot_val(prot) | _PAGE_PSE)), prot 598 arch/x86/mm/init_64.c pgprot_t prot = PAGE_KERNEL; prot 620 arch/x86/mm/init_64.c prot, init); prot 641 arch/x86/mm/init_64.c prot = pte_pgprot(pte_clrhuge(*(pte_t *)pud)); prot 658 arch/x86/mm/init_64.c page_size_mask, prot, init); prot 21 arch/x86/mm/iomap_32.c int iomap_create_wc(resource_size_t base, unsigned long size, pgprot_t *prot) prot 33 arch/x86/mm/iomap_32.c *prot = __pgprot(__PAGE_KERNEL | cachemode2protval(pcm)); prot 35 arch/x86/mm/iomap_32.c pgprot_val(*prot) &= __default_kernel_pte_mask; prot 47 arch/x86/mm/iomap_32.c void *kmap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot) prot 58 arch/x86/mm/iomap_32.c set_pte(kmap_pte - idx, pfn_pte(pfn, prot)); prot 68 arch/x86/mm/iomap_32.c iomap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot) prot 77 arch/x86/mm/iomap_32.c if (!pat_enabled() && pgprot2cachemode(prot) != _PAGE_CACHE_MODE_WB) prot 78 arch/x86/mm/iomap_32.c prot = __pgprot(__PAGE_KERNEL | prot 82 arch/x86/mm/iomap_32.c pgprot_val(prot) &= __default_kernel_pte_mask; prot 84 arch/x86/mm/iomap_32.c return (void __force __iomem *) kmap_atomic_prot_pfn(pfn, prot); prot 186 arch/x86/mm/ioremap.c pgprot_t prot; prot 244 arch/x86/mm/ioremap.c prot = PAGE_KERNEL_IO; prot 246 arch/x86/mm/ioremap.c prot = pgprot_encrypted(prot); prot 251 arch/x86/mm/ioremap.c prot = __pgprot(pgprot_val(prot) | prot 255 arch/x86/mm/ioremap.c prot = __pgprot(pgprot_val(prot) | prot 259 arch/x86/mm/ioremap.c prot = __pgprot(pgprot_val(prot) | prot 263 arch/x86/mm/ioremap.c prot = __pgprot(pgprot_val(prot) | prot 282 arch/x86/mm/ioremap.c if (ioremap_page_range(vaddr, vaddr + size, phys_addr, prot)) prot 728 arch/x86/mm/ioremap.c pgprot_t prot) prot 733 arch/x86/mm/ioremap.c return prot; prot 746 arch/x86/mm/ioremap.c return encrypted_prot ? pgprot_encrypted(prot) prot 747 arch/x86/mm/ioremap.c : pgprot_decrypted(prot); prot 383 arch/x86/mm/kasan_init_64.c pgprot_t prot; prot 385 arch/x86/mm/kasan_init_64.c prot = __pgprot(__PAGE_KERNEL_RO | _PAGE_ENC); prot 386 arch/x86/mm/kasan_init_64.c pgprot_val(prot) &= __default_kernel_pte_mask; prot 388 arch/x86/mm/kasan_init_64.c pte = __pte(__pa(kasan_early_shadow_page) | pgprot_val(prot)); prot 238 arch/x86/mm/mmap.c bool pfn_modify_allowed(unsigned long pfn, pgprot_t prot) prot 242 arch/x86/mm/mmap.c if (!__pte_needs_invert(pgprot_val(prot))) prot 489 arch/x86/mm/pageattr.c static inline bool conflicts(pgprot_t prot, pgprotval_t val) prot 491 arch/x86/mm/pageattr.c return (pgprot_val(prot) & ~val) != pgprot_val(prot); prot 494 arch/x86/mm/pageattr.c static inline void check_conflict(int warnlvl, pgprot_t prot, pgprotval_t val, prot 504 arch/x86/mm/pageattr.c if (warnlvl > cpa_warn_level || !conflicts(prot, val)) prot 508 arch/x86/mm/pageattr.c lvltxt[warnlvl], txt, start, end, pfn, (unsigned long long)pgprot_val(prot), prot 518 arch/x86/mm/pageattr.c static inline pgprot_t static_protections(pgprot_t prot, unsigned long start, prot 529 arch/x86/mm/pageattr.c if (!(pgprot_val(prot) & _PAGE_PRESENT)) prot 530 arch/x86/mm/pageattr.c return prot; prot 536 arch/x86/mm/pageattr.c check_conflict(warnlvl, prot, res, start, end, pfn, "Text NX"); prot 547 arch/x86/mm/pageattr.c check_conflict(warnlvl, prot, res, start, end, pfn, "Text RO"); prot 553 arch/x86/mm/pageattr.c check_conflict(warnlvl, prot, res, start, end, pfn, "PCIBIOS NX"); prot 557 arch/x86/mm/pageattr.c check_conflict(warnlvl, prot, res, start, end, pfn, "Rodata RO"); prot 560 arch/x86/mm/pageattr.c return __pgprot(pgprot_val(prot) & ~forbidden); prot 729 arch/x86/mm/pageattr.c static pgprot_t pgprot_clear_protnone_bits(pgprot_t prot) prot 740 arch/x86/mm/pageattr.c if (!(pgprot_val(prot) & _PAGE_PRESENT)) prot 741 arch/x86/mm/pageattr.c pgprot_val(prot) &= ~_PAGE_GLOBAL; prot 743 arch/x86/mm/pageattr.c return prot; prot 909 arch/x86/mm/pageattr.c pgprot_t prot; prot 919 arch/x86/mm/pageattr.c prot = static_protections(ref_prot, address, pfn, npg, 0, CPA_PROTECT); prot 921 arch/x86/mm/pageattr.c if (pgprot_val(prot) == pgprot_val(ref_prot)) prot 933 arch/x86/mm/pageattr.c ref_prot = prot; prot 968 arch/x86/mm/pat.c unsigned long prot; prot 977 arch/x86/mm/pat.c if (follow_phys(vma, vma->vm_start, 0, &prot, &paddr)) { prot 981 arch/x86/mm/pat.c pgprot = __pgprot(prot); prot 994 arch/x86/mm/pat.c int track_pfn_remap(struct vm_area_struct *vma, pgprot_t *prot, prot 1005 arch/x86/mm/pat.c ret = reserve_pfn_range(paddr, size, prot, 0); prot 1028 arch/x86/mm/pat.c *prot = __pgprot((pgprot_val(*prot) & (~_PAGE_CACHE_MASK)) | prot 1034 arch/x86/mm/pat.c void track_pfn_insert(struct vm_area_struct *vma, pgprot_t *prot, pfn_t pfn) prot 1043 arch/x86/mm/pat.c *prot = __pgprot((pgprot_val(*prot) & (~_PAGE_CACHE_MASK)) | prot 1056 arch/x86/mm/pat.c unsigned long prot; prot 1064 arch/x86/mm/pat.c if (follow_phys(vma, vma->vm_start, 0, &prot, &paddr)) { prot 1086 arch/x86/mm/pat.c pgprot_t pgprot_writecombine(pgprot_t prot) prot 1088 arch/x86/mm/pat.c return __pgprot(pgprot_val(prot) | prot 1093 arch/x86/mm/pat.c pgprot_t pgprot_writethrough(pgprot_t prot) prot 1095 arch/x86/mm/pat.c return __pgprot(pgprot_val(prot) | prot 662 arch/x86/mm/pgtable.c int p4d_set_huge(p4d_t *p4d, phys_addr_t addr, pgprot_t prot) prot 696 arch/x86/mm/pgtable.c int pud_set_huge(pud_t *pud, phys_addr_t addr, pgprot_t prot) prot 709 arch/x86/mm/pgtable.c prot = pgprot_4k_2_large(prot); prot 713 arch/x86/mm/pgtable.c __pgprot(pgprot_val(prot) | _PAGE_PSE))); prot 725 arch/x86/mm/pgtable.c int pmd_set_huge(pmd_t *pmd, phys_addr_t addr, pgprot_t prot) prot 741 arch/x86/mm/pgtable.c prot = pgprot_4k_2_large(prot); prot 745 arch/x86/mm/pgtable.c __pgprot(pgprot_val(prot) | _PAGE_PSE))); prot 77 arch/x86/mm/pkeys.c int __arch_override_mprotect_pkey(struct vm_area_struct *vma, int prot, int pkey) prot 92 arch/x86/mm/pkeys.c if (prot == PROT_EXEC) { prot 27 arch/x86/um/shared/sysdep/syscalls_64.h unsigned long prot, unsigned long flags, prot 348 arch/x86/xen/enlighten_pv.c static void set_aliased_prot(void *v, pgprot_t prot) prot 363 arch/x86/xen/enlighten_pv.c pte = pfn_pte(pfn, prot); prot 1620 arch/x86/xen/mmu_pv.c static inline void __set_pfn_prot(unsigned long pfn, pgprot_t prot) prot 1627 arch/x86/xen/mmu_pv.c pfn_pte(pfn, prot), 0); prot 1761 arch/x86/xen/mmu_pv.c static void __init set_page_prot_flags(void *addr, pgprot_t prot, prot 1765 arch/x86/xen/mmu_pv.c pte_t pte = pfn_pte(pfn, prot); prot 1770 arch/x86/xen/mmu_pv.c static void __init set_page_prot(void *addr, pgprot_t prot) prot 1772 arch/x86/xen/mmu_pv.c return set_page_prot_flags(addr, prot, UVMF_NONE); prot 2295 arch/x86/xen/mmu_pv.c static void xen_set_fixmap(unsigned idx, phys_addr_t phys, pgprot_t prot) prot 2312 arch/x86/xen/mmu_pv.c pte = pfn_pte(phys, prot); prot 2334 arch/x86/xen/mmu_pv.c pte = mfn_pte(phys, prot); prot 2339 arch/x86/xen/mmu_pv.c pte = mfn_pte(phys, prot); prot 2687 arch/x86/xen/mmu_pv.c pgprot_t prot; prot 2694 arch/x86/xen/mmu_pv.c pte_t pte = pte_mkspecial(mfn_pte(*rmd->pfn, rmd->prot)); prot 2716 arch/x86/xen/mmu_pv.c xen_pfn_t *pfn, int nr, int *err_ptr, pgprot_t prot, prot 2728 arch/x86/xen/mmu_pv.c rmd.prot = prot; prot 287 arch/xtensa/include/asm/pgtable.h #define pgprot_noncached(prot) (__pgprot(pgprot_val(prot) & ~_PAGE_CA_MASK)) prot 297 arch/xtensa/include/asm/pgtable.h #define pfn_pte(pfn, prot) __pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot)) prot 298 arch/xtensa/include/asm/pgtable.h #define mk_pte(page, prot) pfn_pte(page_to_pfn(page), prot) prot 15 arch/xtensa/mm/ioremap.c pgprot_t prot) prot 36 arch/xtensa/mm/ioremap.c err = ioremap_page_range(vaddr, vaddr + size, paddr, prot); prot 136 drivers/acpi/apei/ghes.c pgprot_t prot; prot 139 drivers/acpi/apei/ghes.c prot = arch_apei_get_mem_attribute(paddr); prot 140 drivers/acpi/apei/ghes.c __set_fixmap(fixmap_idx, paddr, prot); prot 834 drivers/acpi/arm64/iort.c int prot = IOMMU_WRITE | IOMMU_NOEXEC | IOMMU_MMIO; prot 838 drivers/acpi/arm64/iort.c prot, IOMMU_RESV_MSI); prot 5426 drivers/ata/libata-core.c u8 prot = qc->tf.protocol; prot 5434 drivers/ata/libata-core.c if (ata_is_ncq(prot)) { prot 5454 drivers/ata/libata-core.c if (ata_is_data(prot) && (!qc->sg || !qc->n_elem || !qc->nbytes)) prot 5457 drivers/ata/libata-core.c if (ata_is_dma(prot) || (ata_is_pio(prot) && prot 31 drivers/ata/sata_sil24.c __le16 prot; prot 792 drivers/ata/sata_sil24.c u8 prot = qc->tf.protocol; prot 813 drivers/ata/sata_sil24.c int is_excl = (ata_is_atapi(prot) || prot 848 drivers/ata/sata_sil24.c u16 prot = 0; prot 851 drivers/ata/sata_sil24.c prot |= PRB_PROT_NCQ; prot 853 drivers/ata/sata_sil24.c prot |= PRB_PROT_WRITE; prot 855 drivers/ata/sata_sil24.c prot |= PRB_PROT_READ; prot 856 drivers/ata/sata_sil24.c prb->prot = cpu_to_le16(prot); prot 125 drivers/char/agp/compat_ioctl.c ksegment[seg].prot = usegment[seg].prot; prot 61 drivers/char/agp/compat_ioctl.h compat_int_t prot; /* prot flags for mmap */ prot 117 drivers/char/agp/frontend.c (pgprot_val(seg[i].prot) == pgprot_val(page_prot))) { prot 155 drivers/char/agp/frontend.c static pgprot_t agp_convert_mmap_flags(int prot) prot 159 drivers/char/agp/frontend.c prot_bits = calc_vm_prot_bits(prot, 0) | VM_SHARED; prot 181 drivers/char/agp/frontend.c seg[i].prot = agp_convert_mmap_flags(user_seg[i].prot); prot 163 drivers/crypto/ccp/ccp-dev-v5.c #define CCP5_CMD_PROT(p) (CCP5_CMD_DW0(p).prot) prot 586 drivers/crypto/ccp/ccp-dev.h unsigned int prot:1; prot 1330 drivers/crypto/ux500/cryp/cryp_core.c struct cryp_protection_config prot = { prot 1406 drivers/crypto/ux500/cryp/cryp_core.c if (cryp_configure_protection(device_data, &prot)) { prot 49 drivers/gpu/drm/etnaviv/etnaviv_iommu.c size_t size, int prot) prot 94 drivers/gpu/drm/etnaviv/etnaviv_iommu_v2.c size_t size, int prot) prot 106 drivers/gpu/drm/etnaviv/etnaviv_iommu_v2.c if (prot & ETNAVIV_PROT_WRITE) prot 41 drivers/gpu/drm/etnaviv/etnaviv_mmu.c size_t size, int prot) prot 56 drivers/gpu/drm/etnaviv/etnaviv_mmu.c prot); prot 73 drivers/gpu/drm/etnaviv/etnaviv_mmu.c struct sg_table *sgt, unsigned len, int prot) prot 88 drivers/gpu/drm/etnaviv/etnaviv_mmu.c ret = etnaviv_context_map(context, da, pa, bytes, prot); prot 26 drivers/gpu/drm/etnaviv/etnaviv_mmu.h phys_addr_t paddr, size_t size, int prot); prot 35 drivers/gpu/drm/i915/i915_mm.c pgprot_t prot; prot 43 drivers/gpu/drm/i915/i915_mm.c set_pte_at(r->mm, addr, pte, pte_mkspecial(pfn_pte(r->pfn, r->prot))); prot 72 drivers/gpu/drm/i915/i915_mm.c r.prot = __pgprot((pgprot_val(iomap->prot) & _PAGE_CACHE_MASK) | prot 114 drivers/gpu/drm/lima/lima_gem.c pgprot_t prot = vm_get_page_prot(vma->vm_flags); prot 118 drivers/gpu/drm/lima/lima_gem.c vma->vm_page_prot = pgprot_writecombine(prot); prot 242 drivers/gpu/drm/msm/msm_drv.h struct msm_gem_vma *vma, int prot, prot 424 drivers/gpu/drm/msm/msm_gem.c int prot = IOMMU_READ; prot 427 drivers/gpu/drm/msm/msm_gem.c prot |= IOMMU_WRITE; prot 442 drivers/gpu/drm/msm/msm_gem.c return msm_gem_map_vma(aspace, vma, prot, prot 60 drivers/gpu/drm/msm/msm_gem_vma.c struct msm_gem_vma *vma, int prot, prot 79 drivers/gpu/drm/msm/msm_gem_vma.c size, prot); prot 36 drivers/gpu/drm/msm/msm_gpummu.c struct sg_table *sgt, unsigned len, int prot) prot 44 drivers/gpu/drm/msm/msm_gpummu.c if (prot & IOMMU_WRITE) prot 46 drivers/gpu/drm/msm/msm_gpummu.c if (prot & IOMMU_READ) prot 43 drivers/gpu/drm/msm/msm_iommu.c struct sg_table *sgt, unsigned len, int prot) prot 48 drivers/gpu/drm/msm/msm_iommu.c ret = iommu_map_sg(iommu->domain, iova, sgt->sgl, sgt->nents, prot); prot 16 drivers/gpu/drm/msm/msm_mmu.h unsigned len, int prot); prot 249 drivers/gpu/drm/panfrost/panfrost_mmu.c u64 iova, int prot, struct sg_table *sgt) prot 265 drivers/gpu/drm/panfrost/panfrost_mmu.c ops->map(ops, iova, paddr, pgsize, prot); prot 283 drivers/gpu/drm/panfrost/panfrost_mmu.c int prot = IOMMU_READ | IOMMU_WRITE; prot 289 drivers/gpu/drm/panfrost/panfrost_mmu.c prot |= IOMMU_NOEXEC; prot 296 drivers/gpu/drm/panfrost/panfrost_mmu.c prot, sgt); prot 22 drivers/gpu/drm/rockchip/rockchip_drm_gem.c int prot = IOMMU_READ | IOMMU_WRITE; prot 39 drivers/gpu/drm/rockchip/rockchip_drm_gem.c rk_obj->sgt->nents, prot); prot 116 drivers/gpu/drm/tegra/gem.c int prot = IOMMU_READ | IOMMU_WRITE; prot 139 drivers/gpu/drm/tegra/gem.c bo->sgt->nents, prot); prot 473 drivers/gpu/drm/tegra/gem.c pgprot_t prot = vm_get_page_prot(vma->vm_flags); prot 478 drivers/gpu/drm/tegra/gem.c vma->vm_page_prot = pgprot_writecombine(prot); prot 287 drivers/gpu/drm/ttm/ttm_bo_util.c void *ttm_kmap_atomic_prot(struct page *page, pgprot_t prot) prot 289 drivers/gpu/drm/ttm/ttm_bo_util.c if (pgprot_val(prot) == pgprot_val(PAGE_KERNEL)) prot 292 drivers/gpu/drm/ttm/ttm_bo_util.c return __ttm_kmap_atomic_prot(page, prot); prot 303 drivers/gpu/drm/ttm/ttm_bo_util.c void ttm_kunmap_atomic_prot(void *addr, pgprot_t prot) prot 305 drivers/gpu/drm/ttm/ttm_bo_util.c if (pgprot_val(prot) == pgprot_val(PAGE_KERNEL)) prot 314 drivers/gpu/drm/ttm/ttm_bo_util.c pgprot_t prot) prot 323 drivers/gpu/drm/ttm/ttm_bo_util.c dst = ttm_kmap_atomic_prot(d, prot); prot 329 drivers/gpu/drm/ttm/ttm_bo_util.c ttm_kunmap_atomic_prot(dst, prot); prot 336 drivers/gpu/drm/ttm/ttm_bo_util.c pgprot_t prot) prot 345 drivers/gpu/drm/ttm/ttm_bo_util.c src = ttm_kmap_atomic_prot(s, prot); prot 351 drivers/gpu/drm/ttm/ttm_bo_util.c ttm_kunmap_atomic_prot(src, prot); prot 421 drivers/gpu/drm/ttm/ttm_bo_util.c pgprot_t prot = ttm_io_prot(old_mem->placement, prot 424 drivers/gpu/drm/ttm/ttm_bo_util.c prot); prot 426 drivers/gpu/drm/ttm/ttm_bo_util.c pgprot_t prot = ttm_io_prot(new_mem->placement, prot 429 drivers/gpu/drm/ttm/ttm_bo_util.c prot); prot 588 drivers/gpu/drm/ttm/ttm_bo_util.c pgprot_t prot; prot 611 drivers/gpu/drm/ttm/ttm_bo_util.c prot = ttm_io_prot(mem->placement, PAGE_KERNEL); prot 614 drivers/gpu/drm/ttm/ttm_bo_util.c 0, prot); prot 938 drivers/infiniband/core/uverbs_main.c unsigned long pfn, unsigned long size, pgprot_t prot) prot 959 drivers/infiniband/core/uverbs_main.c vma->vm_page_prot = prot; prot 960 drivers/infiniband/core/uverbs_main.c if (io_remap_pfn_range(vma, vma->vm_start, pfn, size, prot)) { prot 319 drivers/infiniband/hw/cxgb4/t4.h static inline pgprot_t t4_pgprot_wc(pgprot_t prot) prot 322 drivers/infiniband/hw/cxgb4/t4.h return pgprot_writecombine(prot); prot 324 drivers/infiniband/hw/cxgb4/t4.h return pgprot_noncached(prot); prot 1863 drivers/infiniband/hw/mlx4/main.c enum mlx4_protocol prot = MLX4_PROT_IB_IPV6; prot 1876 drivers/infiniband/hw/mlx4/main.c prot, ®_id.id); prot 1888 drivers/infiniband/hw/mlx4/main.c prot, ®_id.mirror); prot 1908 drivers/infiniband/hw/mlx4/main.c prot, reg_id.id); prot 1911 drivers/infiniband/hw/mlx4/main.c prot, reg_id.mirror); prot 1943 drivers/infiniband/hw/mlx4/main.c enum mlx4_protocol prot = MLX4_PROT_IB_IPV6; prot 1966 drivers/infiniband/hw/mlx4/main.c prot, reg_id.id); prot 1972 drivers/infiniband/hw/mlx4/main.c prot, reg_id.mirror); prot 2088 drivers/infiniband/hw/mlx5/main.c pgprot_t prot; prot 2122 drivers/infiniband/hw/mlx5/main.c prot = pgprot_writecombine(vma->vm_page_prot); prot 2125 drivers/infiniband/hw/mlx5/main.c prot = pgprot_noncached(vma->vm_page_prot); prot 2168 drivers/infiniband/hw/mlx5/main.c prot); prot 4602 drivers/infiniband/hw/mlx5/qp.c bool prot = false; prot 4615 drivers/infiniband/hw/mlx5/qp.c prot = true; prot 4618 drivers/infiniband/hw/mlx5/qp.c if (!prot || (data_key == prot_key && data_va == prot_va && prot 518 drivers/infiniband/ulp/iser/iscsi_iser.h struct iser_data_buf prot[ISER_DIRS_NUM]; prot 65 drivers/infiniband/ulp/iser/iser_initiator.c struct iser_data_buf *pbuf_in = &iser_task->prot[ISER_DIR_IN]; prot 119 drivers/infiniband/ulp/iser/iser_initiator.c struct iser_data_buf *pbuf_out = &iser_task->prot[ISER_DIR_OUT]; prot 384 drivers/infiniband/ulp/iser/iser_initiator.c prot_buf = &iser_task->prot[ISER_DIR_IN]; prot 387 drivers/infiniband/ulp/iser/iser_initiator.c prot_buf = &iser_task->prot[ISER_DIR_OUT]; prot 749 drivers/infiniband/ulp/iser/iser_initiator.c iser_task->prot[ISER_DIR_IN].data_len = 0; prot 750 drivers/infiniband/ulp/iser/iser_initiator.c iser_task->prot[ISER_DIR_OUT].data_len = 0; prot 752 drivers/infiniband/ulp/iser/iser_initiator.c iser_task->prot[ISER_DIR_IN].dma_nents = 0; prot 753 drivers/infiniband/ulp/iser/iser_initiator.c iser_task->prot[ISER_DIR_OUT].dma_nents = 0; prot 772 drivers/infiniband/ulp/iser/iser_initiator.c &iser_task->prot[ISER_DIR_IN], prot 783 drivers/infiniband/ulp/iser/iser_initiator.c &iser_task->prot[ISER_DIR_OUT], prot 525 drivers/infiniband/ulp/iser/iser_memory.c err = iser_reg_sig_mr(task, mem, &task->prot[dir], prot 1662 drivers/iommu/amd_iommu.c int prot, prot 1674 drivers/iommu/amd_iommu.c if (!(prot & IOMMU_PROT_MASK)) prot 1696 drivers/iommu/amd_iommu.c if (prot & IOMMU_PROT_IR) prot 1698 drivers/iommu/amd_iommu.c if (prot & IOMMU_PROT_IW) prot 2421 drivers/iommu/amd_iommu.c int prot = 0; prot 2431 drivers/iommu/amd_iommu.c prot = dir2prot(direction); prot 2436 drivers/iommu/amd_iommu.c PAGE_SIZE, prot, GFP_ATOMIC); prot 2579 drivers/iommu/amd_iommu.c int mapped_pages = 0, npages = 0, prot = 0, i; prot 2600 drivers/iommu/amd_iommu.c prot = dir2prot(direction); prot 2612 drivers/iommu/amd_iommu.c PAGE_SIZE, prot, prot 3102 drivers/iommu/amd_iommu.c int prot = 0; prot 3109 drivers/iommu/amd_iommu.c prot |= IOMMU_PROT_IR; prot 3111 drivers/iommu/amd_iommu.c prot |= IOMMU_PROT_IW; prot 3114 drivers/iommu/amd_iommu.c ret = iommu_map_page(domain, iova, paddr, page_size, prot, GFP_KERNEL); prot 3188 drivers/iommu/amd_iommu.c int type, prot = 0; prot 3196 drivers/iommu/amd_iommu.c if (entry->prot & IOMMU_PROT_IR) prot 3197 drivers/iommu/amd_iommu.c prot |= IOMMU_READ; prot 3198 drivers/iommu/amd_iommu.c if (entry->prot & IOMMU_PROT_IW) prot 3199 drivers/iommu/amd_iommu.c prot |= IOMMU_WRITE; prot 3200 drivers/iommu/amd_iommu.c if (entry->prot & IOMMU_UNITY_MAP_FLAG_EXCL_RANGE) prot 3205 drivers/iommu/amd_iommu.c length, prot, type); prot 2136 drivers/iommu/amd_iommu_init.c e->prot = m->flags >> 1; prot 700 drivers/iommu/amd_iommu_types.h int prot; prot 2453 drivers/iommu/arm-smmu-v3.c phys_addr_t paddr, size_t size, int prot) prot 2460 drivers/iommu/arm-smmu-v3.c return ops->map(ops, iova, paddr, size, prot); prot 2703 drivers/iommu/arm-smmu-v3.c int prot = IOMMU_WRITE | IOMMU_NOEXEC | IOMMU_MMIO; prot 2706 drivers/iommu/arm-smmu-v3.c prot, IOMMU_RESV_SW_MSI); prot 1163 drivers/iommu/arm-smmu.c phys_addr_t paddr, size_t size, int prot) prot 1173 drivers/iommu/arm-smmu.c ret = ops->map(ops, iova, paddr, size, prot); prot 1534 drivers/iommu/arm-smmu.c int prot = IOMMU_WRITE | IOMMU_NOEXEC | IOMMU_MMIO; prot 1537 drivers/iommu/arm-smmu.c prot, IOMMU_RESV_SW_MSI); prot 367 drivers/iommu/dma-iommu.c int prot = coherent ? IOMMU_CACHE : 0; prot 370 drivers/iommu/dma-iommu.c prot |= IOMMU_PRIV; prot 374 drivers/iommu/dma-iommu.c return prot | IOMMU_READ | IOMMU_WRITE; prot 376 drivers/iommu/dma-iommu.c return prot | IOMMU_READ; prot 378 drivers/iommu/dma-iommu.c return prot | IOMMU_WRITE; prot 464 drivers/iommu/dma-iommu.c size_t size, int prot) prot 478 drivers/iommu/dma-iommu.c if (iommu_map(domain, iova, phys - iova_off, size, prot)) { prot 572 drivers/iommu/dma-iommu.c pgprot_t prot = dma_pgprot(dev, PAGE_KERNEL, attrs); prot 617 drivers/iommu/dma-iommu.c vaddr = dma_common_pages_remap(pages, size, prot, prot 710 drivers/iommu/dma-iommu.c int prot = dma_info_to_prot(dir, coherent, attrs); prot 713 drivers/iommu/dma-iommu.c dma_handle =__iommu_dma_map(dev, phys, size, prot); prot 817 drivers/iommu/dma-iommu.c int prot = dma_info_to_prot(dir, dev_is_dma_coherent(dev), attrs); prot 873 drivers/iommu/dma-iommu.c if (iommu_map_sg(domain, iova, sg, nents, prot) < iova_len) prot 976 drivers/iommu/dma-iommu.c pgprot_t prot = dma_pgprot(dev, PAGE_KERNEL, attrs); prot 979 drivers/iommu/dma-iommu.c prot, __builtin_return_address(0)); prot 1144 drivers/iommu/dma-iommu.c int prot = IOMMU_WRITE | IOMMU_NOEXEC | IOMMU_MMIO; prot 1160 drivers/iommu/dma-iommu.c if (iommu_map(domain, iova, msi_addr, size, prot)) prot 124 drivers/iommu/exynos-iommu.c #define mk_lv1ent_sect(pa, prot) ((pa >> PG_ENT_SHIFT) | LV1_PROT[prot] | 2) prot 126 drivers/iommu/exynos-iommu.c #define mk_lv2ent_lpage(pa, prot) ((pa >> PG_ENT_SHIFT) | LV2_PROT[prot] | 1) prot 127 drivers/iommu/exynos-iommu.c #define mk_lv2ent_spage(pa, prot) ((pa >> PG_ENT_SHIFT) | LV2_PROT[prot] | 2) prot 978 drivers/iommu/exynos-iommu.c phys_addr_t paddr, int prot, short *pgcnt) prot 997 drivers/iommu/exynos-iommu.c update_pte(sent, mk_lv1ent_sect(paddr, prot)); prot 1015 drivers/iommu/exynos-iommu.c int prot, short *pgcnt) prot 1021 drivers/iommu/exynos-iommu.c update_pte(pent, mk_lv2ent_spage(paddr, prot)); prot 1037 drivers/iommu/exynos-iommu.c *pent = mk_lv2ent_lpage(paddr, prot); prot 1076 drivers/iommu/exynos-iommu.c int prot) prot 1085 drivers/iommu/exynos-iommu.c prot &= SYSMMU_SUPPORTED_PROT_BITS; prot 1092 drivers/iommu/exynos-iommu.c ret = lv1set_section(domain, entry, iova, paddr, prot, prot 1103 drivers/iommu/exynos-iommu.c ret = lv2set_page(pent, paddr, size, prot, prot 324 drivers/iommu/fsl_pamu.c u32 subwin_cnt, int prot) prot 388 drivers/iommu/fsl_pamu.c set_bf(ppaace->addr_bitfields, PAACE_AF_AP, prot); prot 416 drivers/iommu/fsl_pamu.c u32 snoopid, u32 stashid, int enable, int prot) prot 458 drivers/iommu/fsl_pamu.c set_bf(paace->addr_bitfields, PAACE_AF_AP, prot); prot 389 drivers/iommu/fsl_pamu.h u32 subwin_cnt, int prot); prot 392 drivers/iommu/fsl_pamu.h uint32_t snoopid, u32 stashid, int enable, int prot); prot 104 drivers/iommu/fsl_pamu_domain.c sub_win_ptr[i].prot); prot 130 drivers/iommu/fsl_pamu_domain.c 0, wnd->prot); prot 163 drivers/iommu/fsl_pamu_domain.c wnd->prot); prot 177 drivers/iommu/fsl_pamu_domain.c 0, wnd->prot); prot 525 drivers/iommu/fsl_pamu_domain.c phys_addr_t paddr, u64 size, int prot) prot 534 drivers/iommu/fsl_pamu_domain.c if (prot & IOMMU_READ) prot 536 drivers/iommu/fsl_pamu_domain.c if (prot & IOMMU_WRITE) prot 578 drivers/iommu/fsl_pamu_domain.c wnd->prot = pamu_prot; prot 16 drivers/iommu/fsl_pamu_domain.h int prot; prot 2225 drivers/iommu/intel-iommu.c unsigned long nr_pages, int prot) prot 2235 drivers/iommu/intel-iommu.c if ((prot & (DMA_PTE_READ|DMA_PTE_WRITE)) == 0) prot 2238 drivers/iommu/intel-iommu.c prot &= DMA_PTE_READ | DMA_PTE_WRITE | DMA_PTE_SNP; prot 2242 drivers/iommu/intel-iommu.c pteval = ((phys_addr_t)phys_pfn << VTD_PAGE_SHIFT) | prot; prot 2254 drivers/iommu/intel-iommu.c pteval = (sg_phys(sg) - pgoff) | prot; prot 2340 drivers/iommu/intel-iommu.c unsigned long nr_pages, int prot) prot 2346 drivers/iommu/intel-iommu.c ret = __domain_mapping(domain, iov_pfn, sg, phys_pfn, nr_pages, prot); prot 2360 drivers/iommu/intel-iommu.c int prot) prot 2362 drivers/iommu/intel-iommu.c return domain_mapping(domain, iov_pfn, sg, 0, nr_pages, prot); prot 2367 drivers/iommu/intel-iommu.c int prot) prot 2369 drivers/iommu/intel-iommu.c return domain_mapping(domain, iov_pfn, NULL, phys_pfn, nr_pages, prot); prot 3498 drivers/iommu/intel-iommu.c int prot = 0; prot 3522 drivers/iommu/intel-iommu.c prot |= DMA_PTE_READ; prot 3524 drivers/iommu/intel-iommu.c prot |= DMA_PTE_WRITE; prot 3532 drivers/iommu/intel-iommu.c mm_to_dma_pfn(paddr_pfn), size, prot); prot 3714 drivers/iommu/intel-iommu.c int prot = 0; prot 3747 drivers/iommu/intel-iommu.c prot |= DMA_PTE_READ; prot 3749 drivers/iommu/intel-iommu.c prot |= DMA_PTE_WRITE; prot 3753 drivers/iommu/intel-iommu.c ret = domain_sg_mapping(domain, start_vpfn, sglist, size, prot); prot 3817 drivers/iommu/intel-iommu.c int prot = 0; prot 3840 drivers/iommu/intel-iommu.c prot |= DMA_PTE_READ; prot 3842 drivers/iommu/intel-iommu.c prot |= DMA_PTE_WRITE; prot 3873 drivers/iommu/intel-iommu.c tlb_addr >> VTD_PAGE_SHIFT, nrpages, prot); prot 5445 drivers/iommu/intel-iommu.c int prot = 0; prot 5449 drivers/iommu/intel-iommu.c prot |= DMA_PTE_READ; prot 5451 drivers/iommu/intel-iommu.c prot |= DMA_PTE_WRITE; prot 5453 drivers/iommu/intel-iommu.c prot |= DMA_PTE_SNP; prot 5473 drivers/iommu/intel-iommu.c hpa >> VTD_PAGE_SHIFT, size, prot); prot 5669 drivers/iommu/intel-iommu.c int prot = DMA_PTE_READ | DMA_PTE_WRITE; prot 5693 drivers/iommu/intel-iommu.c length, prot, type); prot 5707 drivers/iommu/intel-iommu.c reg = iommu_alloc_resv_region(0, 1UL << 24, prot, prot 316 drivers/iommu/io-pgtable-arm-v7s.c static arm_v7s_iopte arm_v7s_prot_to_pte(int prot, int lvl, prot 322 drivers/iommu/io-pgtable-arm-v7s.c if (!(prot & IOMMU_MMIO)) prot 326 drivers/iommu/io-pgtable-arm-v7s.c if (!(prot & IOMMU_PRIV)) prot 328 drivers/iommu/io-pgtable-arm-v7s.c if (!(prot & IOMMU_WRITE)) prot 333 drivers/iommu/io-pgtable-arm-v7s.c if ((prot & IOMMU_NOEXEC) && ap) prot 335 drivers/iommu/io-pgtable-arm-v7s.c if (prot & IOMMU_MMIO) prot 337 drivers/iommu/io-pgtable-arm-v7s.c else if (prot & IOMMU_CACHE) prot 349 drivers/iommu/io-pgtable-arm-v7s.c int prot = IOMMU_READ; prot 353 drivers/iommu/io-pgtable-arm-v7s.c prot |= IOMMU_WRITE; prot 355 drivers/iommu/io-pgtable-arm-v7s.c prot |= IOMMU_PRIV; prot 357 drivers/iommu/io-pgtable-arm-v7s.c prot |= IOMMU_MMIO; prot 359 drivers/iommu/io-pgtable-arm-v7s.c prot |= IOMMU_CACHE; prot 361 drivers/iommu/io-pgtable-arm-v7s.c prot |= IOMMU_NOEXEC; prot 363 drivers/iommu/io-pgtable-arm-v7s.c return prot; prot 413 drivers/iommu/io-pgtable-arm-v7s.c unsigned long iova, phys_addr_t paddr, int prot, prot 439 drivers/iommu/io-pgtable-arm-v7s.c pte = arm_v7s_prot_to_pte(prot, lvl, cfg); prot 474 drivers/iommu/io-pgtable-arm-v7s.c phys_addr_t paddr, size_t size, int prot, prot 486 drivers/iommu/io-pgtable-arm-v7s.c return arm_v7s_init_pte(data, iova, paddr, prot, prot 517 drivers/iommu/io-pgtable-arm-v7s.c return __arm_v7s_map(data, iova, paddr, size, prot, lvl + 1, cptep); prot 521 drivers/iommu/io-pgtable-arm-v7s.c phys_addr_t paddr, size_t size, int prot) prot 528 drivers/iommu/io-pgtable-arm-v7s.c if (!(prot & (IOMMU_READ | IOMMU_WRITE))) prot 535 drivers/iommu/io-pgtable-arm-v7s.c ret = __arm_v7s_map(data, iova, paddr, size, prot, 1, data->pgd); prot 300 drivers/iommu/io-pgtable-arm.c phys_addr_t paddr, arm_lpae_iopte prot, prot 303 drivers/iommu/io-pgtable-arm.c arm_lpae_iopte pte = prot; prot 323 drivers/iommu/io-pgtable-arm.c arm_lpae_iopte prot, int lvl, prot 347 drivers/iommu/io-pgtable-arm.c __arm_lpae_init_pte(data, paddr, prot, lvl, ptep); prot 383 drivers/iommu/io-pgtable-arm.c phys_addr_t paddr, size_t size, arm_lpae_iopte prot, prot 396 drivers/iommu/io-pgtable-arm.c return arm_lpae_init_pte(data, iova, paddr, prot, lvl, ptep); prot 425 drivers/iommu/io-pgtable-arm.c return __arm_lpae_map(data, iova, paddr, size, prot, lvl + 1, cptep); prot 429 drivers/iommu/io-pgtable-arm.c int prot) prot 436 drivers/iommu/io-pgtable-arm.c if (!(prot & IOMMU_WRITE) && (prot & IOMMU_READ)) prot 438 drivers/iommu/io-pgtable-arm.c if (!(prot & IOMMU_PRIV)) prot 442 drivers/iommu/io-pgtable-arm.c if (prot & IOMMU_READ) prot 444 drivers/iommu/io-pgtable-arm.c if (prot & IOMMU_WRITE) prot 454 drivers/iommu/io-pgtable-arm.c if (prot & IOMMU_MMIO) prot 456 drivers/iommu/io-pgtable-arm.c else if (prot & IOMMU_CACHE) prot 461 drivers/iommu/io-pgtable-arm.c if (prot & IOMMU_MMIO) prot 464 drivers/iommu/io-pgtable-arm.c else if (prot & IOMMU_CACHE) prot 467 drivers/iommu/io-pgtable-arm.c else if (prot & IOMMU_QCOM_SYS_CACHE) prot 472 drivers/iommu/io-pgtable-arm.c if (prot & IOMMU_NOEXEC) prot 484 drivers/iommu/io-pgtable-arm.c arm_lpae_iopte prot; prot 494 drivers/iommu/io-pgtable-arm.c prot = arm_lpae_prot_to_pte(data, iommu_prot); prot 495 drivers/iommu/io-pgtable-arm.c ret = __arm_lpae_map(data, iova, paddr, size, prot, lvl, ptep); prot 299 drivers/iommu/iommu.c new->prot, new->type); prot 663 drivers/iommu/iommu.c ret = iommu_map(domain, addr, addr, pg_size, entry->prot); prot 1859 drivers/iommu/iommu.c phys_addr_t paddr, size_t size, int prot) prot 1897 drivers/iommu/iommu.c ret = ops->map(domain, iova, paddr, pgsize, prot); prot 1996 drivers/iommu/iommu.c struct scatterlist *sg, unsigned int nents, int prot) prot 2007 drivers/iommu/iommu.c ret = iommu_map(domain, iova + mapped, start, len, prot); prot 2038 drivers/iommu/iommu.c phys_addr_t paddr, u64 size, int prot) prot 2044 drivers/iommu/iommu.c prot); prot 2173 drivers/iommu/iommu.c size_t length, int prot, prot 2185 drivers/iommu/iommu.c region->prot = prot; prot 727 drivers/iommu/ipmmu-vmsa.c phys_addr_t paddr, size_t size, int prot) prot 734 drivers/iommu/ipmmu-vmsa.c return domain->iop->map(domain->iop, iova, paddr, size, prot); prot 507 drivers/iommu/msm_iommu.c phys_addr_t pa, size_t len, int prot) prot 514 drivers/iommu/msm_iommu.c ret = priv->iop->map(priv->iop, iova, pa, len, prot); prot 430 drivers/iommu/mtk_iommu.c phys_addr_t paddr, size_t size, int prot) prot 442 drivers/iommu/mtk_iommu.c ret = dom->iop->map(dom->iop, iova, paddr, size, prot); prot 60 drivers/iommu/mtk_iommu_v1.c #define F_MMU_TF_PROTECT_SEL(prot) (((prot) & 0x3) << 5) prot 298 drivers/iommu/mtk_iommu_v1.c phys_addr_t paddr, size_t size, int prot) prot 556 drivers/iommu/omap-iommu.c static int iopgd_alloc_section(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) prot 567 drivers/iommu/omap-iommu.c *iopgd = (pa & IOSECTION_MASK) | prot | IOPGD_SECTION; prot 572 drivers/iommu/omap-iommu.c static int iopgd_alloc_super(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) prot 585 drivers/iommu/omap-iommu.c *(iopgd + i) = (pa & IOSUPER_MASK) | prot | IOPGD_SUPER; prot 590 drivers/iommu/omap-iommu.c static int iopte_alloc_page(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) prot 600 drivers/iommu/omap-iommu.c *iopte = (pa & IOPAGE_MASK) | prot | IOPTE_SMALL; prot 609 drivers/iommu/omap-iommu.c static int iopte_alloc_large(struct omap_iommu *obj, u32 da, u32 pa, u32 prot) prot 627 drivers/iommu/omap-iommu.c *(iopte + i) = (pa & IOLARGE_MASK) | prot | IOPTE_LARGE; prot 636 drivers/iommu/omap-iommu.c u32 prot; prot 663 drivers/iommu/omap-iommu.c prot = get_iopte_attr(e); prot 666 drivers/iommu/omap-iommu.c err = fn(obj, e->da, e->pa, prot); prot 1342 drivers/iommu/omap-iommu.c phys_addr_t pa, size_t bytes, int prot) prot 422 drivers/iommu/qcom_iommu.c phys_addr_t paddr, size_t size, int prot) prot 433 drivers/iommu/qcom_iommu.c ret = ops->map(ops, iova, paddr, size, prot); prot 229 drivers/iommu/rockchip-iommu.c static u32 rk_mk_pte(phys_addr_t page, int prot) prot 232 drivers/iommu/rockchip-iommu.c flags |= (prot & IOMMU_READ) ? RK_PTE_PAGE_READABLE : 0; prot 233 drivers/iommu/rockchip-iommu.c flags |= (prot & IOMMU_WRITE) ? RK_PTE_PAGE_WRITABLE : 0; prot 717 drivers/iommu/rockchip-iommu.c phys_addr_t paddr, size_t size, int prot) prot 731 drivers/iommu/rockchip-iommu.c pte_addr[pte_count] = rk_mk_pte(paddr, prot); prot 755 drivers/iommu/rockchip-iommu.c &iova, &page_phys, &paddr, prot); prot 761 drivers/iommu/rockchip-iommu.c phys_addr_t paddr, size_t size, int prot) prot 790 drivers/iommu/rockchip-iommu.c paddr, size, prot); prot 268 drivers/iommu/s390-iommu.c phys_addr_t paddr, size_t size, int prot) prot 273 drivers/iommu/s390-iommu.c if (!(prot & IOMMU_READ)) prot 276 drivers/iommu/s390-iommu.c if (!(prot & IOMMU_WRITE)) prot 181 drivers/iommu/tegra-gart.c phys_addr_t pa, size_t bytes, int prot) prot 654 drivers/iommu/tegra-smmu.c phys_addr_t paddr, size_t size, int prot) prot 671 drivers/iommu/tegra-smmu.c if (prot & IOMMU_READ) prot 674 drivers/iommu/tegra-smmu.c if (prot & IOMMU_WRITE) prot 427 drivers/iommu/virtio-iommu.c unsigned long prot = IOMMU_WRITE | IOMMU_NOEXEC | IOMMU_MMIO; prot 450 drivers/iommu/virtio-iommu.c region = iommu_alloc_resv_region(start, size, prot, prot 718 drivers/iommu/virtio-iommu.c phys_addr_t paddr, size_t size, int prot) prot 725 drivers/iommu/virtio-iommu.c flags = (prot & IOMMU_READ ? VIRTIO_IOMMU_MAP_F_READ : 0) | prot 726 drivers/iommu/virtio-iommu.c (prot & IOMMU_WRITE ? VIRTIO_IOMMU_MAP_F_WRITE : 0) | prot 727 drivers/iommu/virtio-iommu.c (prot & IOMMU_MMIO ? VIRTIO_IOMMU_MAP_F_MMIO : 0); prot 815 drivers/iommu/virtio-iommu.c int prot = IOMMU_WRITE | IOMMU_NOEXEC | IOMMU_MMIO; prot 833 drivers/iommu/virtio-iommu.c prot, IOMMU_RESV_SW_MSI); prot 49 drivers/misc/mic/scif/scif_debugfs.c window->nr_pages, window->nr_contig_chunks, window->prot); prot 342 drivers/misc/mic/scif/scif_dma.c scif_register_temp(scif_epd_t epd, unsigned long addr, size_t len, int prot, prot 353 drivers/misc/mic/scif/scif_dma.c aligned_len, &prot, 0, &pinned_pages); prot 357 drivers/misc/mic/scif/scif_dma.c pinned_pages->prot = prot; prot 377 drivers/misc/mic/scif/scif_dma.c (*out_window)->prot = pinned_pages->prot; prot 1679 drivers/misc/mic/scif/scif_dma.c remote_req.prot = dir == SCIF_LOCAL_TO_REMOTE ? VM_WRITE : VM_READ; prot 1707 drivers/misc/mic/scif/scif_dma.c req.prot = (dir == SCIF_LOCAL_TO_REMOTE ? prot 1718 drivers/misc/mic/scif/scif_dma.c req.nr_bytes, req.prot, prot 1746 drivers/misc/mic/scif/scif_dma.c req.prot = dir == SCIF_LOCAL_TO_REMOTE ? VM_READ : VM_WRITE; prot 314 drivers/misc/mic/scif/scif_fd.c reg.offset, reg.prot, reg.flags); prot 306 drivers/misc/mic/scif/scif_fence.c req.prot = SCIF_PROT_WRITE; prot 33 drivers/misc/mic/scif/scif_mmap.c req.prot = recv_window->prot; prot 237 drivers/misc/mic/scif/scif_mmap.c req.prot = 0; prot 276 drivers/misc/mic/scif/scif_mmap.c (*pages)->prot_flags = window->prot; prot 565 drivers/misc/mic/scif/scif_mmap.c req.prot = vma->vm_flags & (VM_READ | VM_WRITE); prot 637 drivers/misc/mic/scif/scif_mmap.c req.prot = vma->vm_flags & (VM_READ | VM_WRITE); prot 81 drivers/misc/mic/scif/scif_rma.c scif_create_pinned_pages(int nr_pages, int prot) prot 94 drivers/misc/mic/scif/scif_rma.c pin->prot = prot; prot 113 drivers/misc/mic/scif/scif_rma.c int writeable = pin->prot & SCIF_PROT_WRITE; prot 917 drivers/misc/mic/scif/scif_rma.c remote_window->prot = window->prot; prot 1186 drivers/misc/mic/scif/scif_rma.c req.prot = 0; prot 1314 drivers/misc/mic/scif/scif_rma.c int prot = *out_prot; prot 1324 drivers/misc/mic/scif/scif_rma.c if (prot & ~(SCIF_PROT_READ | SCIF_PROT_WRITE)) prot 1338 drivers/misc/mic/scif/scif_rma.c pinned_pages = scif_create_pinned_pages(nr_pages, prot); prot 1365 drivers/misc/mic/scif/scif_rma.c if (prot == SCIF_PROT_READ) prot 1367 drivers/misc/mic/scif/scif_rma.c prot |= SCIF_PROT_WRITE; prot 1381 drivers/misc/mic/scif/scif_rma.c (prot & SCIF_PROT_WRITE) ? FOLL_WRITE : 0, prot 1393 drivers/misc/mic/scif/scif_rma.c prot &= ~SCIF_PROT_WRITE; prot 1407 drivers/misc/mic/scif/scif_rma.c *out_prot = prot; prot 1424 drivers/misc/mic/scif/scif_rma.c int scif_pin_pages(void *addr, size_t len, int prot, prot 1427 drivers/misc/mic/scif/scif_rma.c return __scif_pin_pages(addr, len, &prot, map_flags, pages); prot 1524 drivers/misc/mic/scif/scif_rma.c window->prot = pinned_pages->prot; prot 1569 drivers/misc/mic/scif/scif_rma.c int prot, int map_flags) prot 1581 drivers/misc/mic/scif/scif_rma.c epd, addr, len, offset, prot, map_flags); prot 1597 drivers/misc/mic/scif/scif_rma.c if (prot & ~(SCIF_PROT_READ | SCIF_PROT_WRITE)) prot 1646 drivers/misc/mic/scif/scif_rma.c err = __scif_pin_pages(addr, len, &prot, prot 1656 drivers/misc/mic/scif/scif_rma.c window->prot = pinned_pages->prot; prot 1724 drivers/misc/mic/scif/scif_rma.c req.prot = 0; prot 188 drivers/misc/mic/scif/scif_rma.h int prot; prot 258 drivers/misc/mic/scif/scif_rma.h int prot; prot 106 drivers/misc/mic/scif/scif_rma_list.c if ((window->prot & req->prot) == req->prot) { prot 155 drivers/misc/mic/scif/scif_rma_list.c if ((window->prot & req->prot) != req->prot) prot 30 drivers/misc/mic/scif/scif_rma_list.h int prot; prot 2223 drivers/mtd/chips/cfi_cmdset_0001.c u_char *buf, u_int size, u_long prot, u_int grpno, u_int grpsz) prot 2256 drivers/mtd/chips/cfi_cmdset_0001.c u_char *buf, u_int size, u_long prot, u_int grpno, u_int grpsz) prot 2281 drivers/mtd/chips/cfi_cmdset_0001.c u_char *buf, u_int size, u_long prot, u_int grpno, u_int grpsz) prot 2292 drivers/mtd/chips/cfi_cmdset_0001.c return do_write_oneword(map, chip, prot, datum, FL_OTP_WRITE); prot 3340 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c __u8 prot = 0; prot 3349 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c prot = ipv6_hdr(skb)->nexthdr; prot 3352 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c prot = ip_hdr(skb)->protocol; prot 3366 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c if (prot == IPPROTO_TCP) prot 2918 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c int dpaa2_eth_cls_fld_off(int prot, int field) prot 2923 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c if (dist_fields[i].cls_prot == prot && prot 2986 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c key->extract.from_hdr.prot = dist_fields[i].cls_prot; prot 523 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.h int dpaa2_eth_cls_fld_off(int prot, int field); prot 448 drivers/net/ethernet/freescale/dpaa2/dpkg.h enum net_prot prot; prot 407 drivers/net/ethernet/freescale/dpaa2/dpni-cmd.h u8 prot; prot 37 drivers/net/ethernet/freescale/dpaa2/dpni.c extr->prot = cfg->extracts[i].extract.from_hdr.prot; prot 309 drivers/net/ethernet/freescale/enetc/enetc_pf.c static int enetc_vlan_rx_add_vid(struct net_device *ndev, __be16 prot, u16 vid) prot 327 drivers/net/ethernet/freescale/enetc/enetc_pf.c static int enetc_vlan_rx_del_vid(struct net_device *ndev, __be16 prot, u16 vid) prot 2716 drivers/net/ethernet/mellanox/mlx4/fw.c u8 prot; prot 2757 drivers/net/ethernet/mellanox/mlx4/fw.c prot = ((u8 *)(&mgm->members_count))[0] >> 6; prot 2764 drivers/net/ethernet/mellanox/mlx4/fw.c prot, 0); prot 2769 drivers/net/ethernet/mellanox/mlx4/fw.c , 0, prot, prot 158 drivers/net/ethernet/mellanox/mlx4/mcg.c u32 prot; prot 206 drivers/net/ethernet/mellanox/mlx4/mcg.c prot = be32_to_cpu(mgm->members_count) >> 30; prot 221 drivers/net/ethernet/mellanox/mlx4/mcg.c mgm->members_count = cpu_to_be32(members_count | (prot << 30)); prot 428 drivers/net/ethernet/mellanox/mlx4/mcg.c u32 prot; prot 474 drivers/net/ethernet/mellanox/mlx4/mcg.c prot = be32_to_cpu(mgm->members_count) >> 30; prot 505 drivers/net/ethernet/mellanox/mlx4/mcg.c (prot << 30)); prot 695 drivers/net/ethernet/mellanox/mlx4/mcg.c u8 *gid, enum mlx4_protocol prot, prot 704 drivers/net/ethernet/mellanox/mlx4/mcg.c u8 op_mod = (prot == MLX4_PROT_ETH) ? prot 739 drivers/net/ethernet/mellanox/mlx4/mcg.c be32_to_cpu(mgm->members_count) >> 30 == prot) prot 1105 drivers/net/ethernet/mellanox/mlx4/mcg.c int block_mcast_loopback, enum mlx4_protocol prot, prot 1125 drivers/net/ethernet/mellanox/mlx4/mcg.c err = find_entry(dev, port, gid, prot, prot 1171 drivers/net/ethernet/mellanox/mlx4/mcg.c mgm->members_count = cpu_to_be32(members_count | (u32) prot << 30); prot 1191 drivers/net/ethernet/mellanox/mlx4/mcg.c if (prot == MLX4_PROT_ETH && index != -1) { prot 1215 drivers/net/ethernet/mellanox/mlx4/mcg.c enum mlx4_protocol prot, enum mlx4_steer_type steer) prot 1234 drivers/net/ethernet/mellanox/mlx4/mcg.c err = find_entry(dev, port, gid, prot, prot 1248 drivers/net/ethernet/mellanox/mlx4/mcg.c if (prot == MLX4_PROT_ETH && prot 1269 drivers/net/ethernet/mellanox/mlx4/mcg.c mgm->members_count = cpu_to_be32(--members_count | (u32) prot << 30); prot 1271 drivers/net/ethernet/mellanox/mlx4/mcg.c if (prot == MLX4_PROT_ETH) prot 1274 drivers/net/ethernet/mellanox/mlx4/mcg.c if (members_count && (prot != MLX4_PROT_ETH || !removed_entry)) { prot 1280 drivers/net/ethernet/mellanox/mlx4/mcg.c mgm->members_count = cpu_to_be32((u32) prot << 30); prot 1337 drivers/net/ethernet/mellanox/mlx4/mcg.c enum mlx4_protocol prot) prot 1352 drivers/net/ethernet/mellanox/mlx4/mcg.c qpn |= (prot << 28); prot 1370 drivers/net/ethernet/mellanox/mlx4/mcg.c enum mlx4_protocol prot, u64 *reg_id) prot 1387 drivers/net/ethernet/mellanox/mlx4/mcg.c switch (prot) { prot 1409 drivers/net/ethernet/mellanox/mlx4/mcg.c enum mlx4_protocol prot, u64 *reg_id) prot 1413 drivers/net/ethernet/mellanox/mlx4/mcg.c if (prot == MLX4_PROT_ETH) prot 1418 drivers/net/ethernet/mellanox/mlx4/mcg.c if (prot == MLX4_PROT_ETH) prot 1423 drivers/net/ethernet/mellanox/mlx4/mcg.c block_mcast_loopback, prot); prot 1425 drivers/net/ethernet/mellanox/mlx4/mcg.c block_mcast_loopback, prot, prot 1431 drivers/net/ethernet/mellanox/mlx4/mcg.c prot, reg_id); prot 1439 drivers/net/ethernet/mellanox/mlx4/mcg.c enum mlx4_protocol prot, u64 reg_id) prot 1443 drivers/net/ethernet/mellanox/mlx4/mcg.c if (prot == MLX4_PROT_ETH) prot 1448 drivers/net/ethernet/mellanox/mlx4/mcg.c if (prot == MLX4_PROT_ETH) prot 1452 drivers/net/ethernet/mellanox/mlx4/mcg.c return mlx4_QP_ATTACH(dev, qp, gid, 0, 0, prot); prot 1454 drivers/net/ethernet/mellanox/mlx4/mcg.c return mlx4_qp_detach_common(dev, qp, gid, prot, prot 1531 drivers/net/ethernet/mellanox/mlx4/mcg.c int block_mcast_loopback, enum mlx4_protocol prot) prot 1533 drivers/net/ethernet/mellanox/mlx4/mcg.c if (prot == MLX4_PROT_ETH) prot 1538 drivers/net/ethernet/mellanox/mlx4/mcg.c block_mcast_loopback, prot); prot 1541 drivers/net/ethernet/mellanox/mlx4/mcg.c prot, MLX4_UC_STEER); prot 1546 drivers/net/ethernet/mellanox/mlx4/mcg.c u8 gid[16], enum mlx4_protocol prot) prot 1548 drivers/net/ethernet/mellanox/mlx4/mcg.c if (prot == MLX4_PROT_ETH) prot 1552 drivers/net/ethernet/mellanox/mlx4/mcg.c return mlx4_QP_ATTACH(dev, qp, gid, 0, 0, prot); prot 1554 drivers/net/ethernet/mellanox/mlx4/mcg.c return mlx4_qp_detach_common(dev, qp, gid, prot, MLX4_UC_STEER); prot 1322 drivers/net/ethernet/mellanox/mlx4/mlx4.h enum mlx4_protocol prot, enum mlx4_steer_type steer); prot 1324 drivers/net/ethernet/mellanox/mlx4/mlx4.h int block_mcast_loopback, enum mlx4_protocol prot, prot 1329 drivers/net/ethernet/mellanox/mlx4/mlx4.h enum mlx4_protocol prot, u64 *reg_id); prot 90 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c enum mlx4_protocol prot; prot 4028 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c u8 *gid, enum mlx4_protocol prot, prot 4044 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c res->prot = prot; prot 4056 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c u8 *gid, enum mlx4_protocol prot, prot 4064 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c if (!res || res->prot != prot || res->steer != steer) prot 4078 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c u8 gid[16], int block_loopback, enum mlx4_protocol prot, prot 4087 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c block_loopback, prot, prot 4091 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c if (prot == MLX4_PROT_ETH) { prot 4098 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c block_loopback, prot, type); prot 4105 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c u8 gid[16], enum mlx4_protocol prot, prot 4112 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c return mlx4_qp_detach_common(dev, qp, gid, prot, type); prot 4119 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c u8 *gid, enum mlx4_protocol prot) prot 4123 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c if (prot != MLX4_PROT_ETH) prot 4145 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c enum mlx4_protocol prot = (vhcr->in_modifier >> 28) & 0x7; prot 4162 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c err = qp_attach(dev, slave, &qp, gid, block_loopback, prot, prot 4168 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c err = add_mcg_res(dev, slave, rqp, gid, prot, type, reg_id); prot 4172 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c err = mlx4_adjust_port(dev, slave, gid, prot); prot 4176 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c err = rem_mcg_res(dev, slave, rqp, gid, prot, type, ®_id); prot 4180 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c err = qp_detach(dev, &qp, gid, prot, type, reg_id); prot 4189 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c qp_detach(dev, &qp, gid, prot, type, reg_id); prot 4595 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c rgid->prot, rgid->steer); prot 869 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c u32 prot[6]; prot 874 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c for (i = 0; i < ARRAY_SIZE(prot); i++) { prot 875 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c prot[i] = mt76_rr(dev, MT_CCK_PROT_CFG + i * 4); prot 876 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c prot[i] &= ~MT_PROT_CFG_CTRL; prot 878 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c prot[i] &= ~MT_PROT_CFG_RATE; prot 889 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c prot[0] |= MT_PROT_CTRL_RTS_CTS; prot 892 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c prot[1] |= MT_PROT_CTRL_CTS2SELF; prot 894 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c prot[2] |= MT_PROT_RATE_CCK_11; prot 895 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c prot[3] |= MT_PROT_RATE_CCK_11; prot 896 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c prot[4] |= MT_PROT_RATE_CCK_11; prot 897 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c prot[5] |= MT_PROT_RATE_CCK_11; prot 904 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c prot[1] |= MT_PROT_CTRL_RTS_CTS; prot 906 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c prot[2] |= MT_PROT_RATE_OFDM_24; prot 907 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c prot[3] |= MT_PROT_RATE_DUP_OFDM_24; prot 908 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c prot[4] |= MT_PROT_RATE_OFDM_24; prot 909 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c prot[5] |= MT_PROT_RATE_DUP_OFDM_24; prot 919 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c prot[2] |= MT_PROT_CTRL_RTS_CTS; prot 920 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c prot[3] |= MT_PROT_CTRL_RTS_CTS; prot 921 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c prot[4] |= MT_PROT_CTRL_RTS_CTS; prot 922 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c prot[5] |= MT_PROT_CTRL_RTS_CTS; prot 928 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c prot[3] |= MT_PROT_CTRL_RTS_CTS; prot 929 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c prot[5] |= MT_PROT_CTRL_RTS_CTS; prot 936 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c prot[4] |= MT_PROT_CTRL_RTS_CTS; prot 937 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c prot[5] |= MT_PROT_CTRL_RTS_CTS; prot 940 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c for (i = 0; i < ARRAY_SIZE(prot); i++) prot 941 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c mt76_wr(dev, MT_CCK_PROT_CFG + i * 4, prot[i]); prot 205 drivers/net/wireless/mediatek/mt7601u/mac.c u32 prot[6]; prot 209 drivers/net/wireless/mediatek/mt7601u/mac.c prot[0] = MT_PROT_NAV_SHORT | prot 212 drivers/net/wireless/mediatek/mt7601u/mac.c prot[1] = prot[0]; prot 214 drivers/net/wireless/mediatek/mt7601u/mac.c prot[1] |= MT_PROT_CTRL_CTS2SELF; prot 216 drivers/net/wireless/mediatek/mt7601u/mac.c prot[2] = prot[4] = MT_PROT_NAV_SHORT | MT_PROT_TXOP_ALLOW_BW20; prot 217 drivers/net/wireless/mediatek/mt7601u/mac.c prot[3] = prot[5] = MT_PROT_NAV_SHORT | MT_PROT_TXOP_ALLOW_ALL; prot 220 drivers/net/wireless/mediatek/mt7601u/mac.c prot[2] |= MT_PROT_RATE_CCK_11; prot 221 drivers/net/wireless/mediatek/mt7601u/mac.c prot[3] |= MT_PROT_RATE_CCK_11; prot 222 drivers/net/wireless/mediatek/mt7601u/mac.c prot[4] |= MT_PROT_RATE_CCK_11; prot 223 drivers/net/wireless/mediatek/mt7601u/mac.c prot[5] |= MT_PROT_RATE_CCK_11; prot 225 drivers/net/wireless/mediatek/mt7601u/mac.c prot[2] |= MT_PROT_RATE_OFDM_24; prot 226 drivers/net/wireless/mediatek/mt7601u/mac.c prot[3] |= MT_PROT_RATE_DUP_OFDM_24; prot 227 drivers/net/wireless/mediatek/mt7601u/mac.c prot[4] |= MT_PROT_RATE_OFDM_24; prot 228 drivers/net/wireless/mediatek/mt7601u/mac.c prot[5] |= MT_PROT_RATE_DUP_OFDM_24; prot 253 drivers/net/wireless/mediatek/mt7601u/mac.c prot[i + 2] |= MT_PROT_CTRL_RTS_CTS; prot 256 drivers/net/wireless/mediatek/mt7601u/mac.c mt7601u_wr(dev, MT_CCK_PROT_CFG + i * 4, prot[i]); prot 757 drivers/power/supply/sbs-battery.c static void sbs_alert(struct i2c_client *client, enum i2c_alert_protocol prot, prot 237 drivers/power/supply/sbs-manager.c static void sbsm_alert(struct i2c_client *client, enum i2c_alert_protocol prot, prot 200 drivers/s390/net/qeth_core.h #define qeth_is_ipafunc_supported(c, prot, f) \ prot 201 drivers/s390/net/qeth_core.h ((prot == QETH_PROT_IPV6) ? \ prot 203 drivers/s390/net/qeth_core.h #define qeth_is_ipafunc_enabled(c, prot, f) \ prot 204 drivers/s390/net/qeth_core.h ((prot == QETH_PROT_IPV6) ? \ prot 910 drivers/s390/net/qeth_core.h __be16 prot = veth->h_vlan_proto; prot 912 drivers/s390/net/qeth_core.h if (prot == htons(ETH_P_8021Q)) prot 913 drivers/s390/net/qeth_core.h prot = veth->h_vlan_encapsulated_proto; prot 915 drivers/s390/net/qeth_core.h switch (prot) { prot 981 drivers/s390/net/qeth_core.h enum qeth_prot_versions prot); prot 1038 drivers/s390/net/qeth_core.h enum qeth_prot_versions prot, prot 1047 drivers/s390/net/qeth_core.h enum qeth_prot_versions prot); prot 2748 drivers/s390/net/qeth_core_main.c enum qeth_prot_versions prot, prot 2771 drivers/s390/net/qeth_core_main.c hdr->prot_version = prot; prot 2938 drivers/s390/net/qeth_core_main.c enum qeth_prot_versions prot) prot 2943 drivers/s390/net/qeth_core_main.c QETH_CARD_TEXT_(card, 2, "qipassi%i", prot); prot 2944 drivers/s390/net/qeth_core_main.c iob = qeth_ipa_alloc_cmd(card, IPA_CMD_QIPASSIST, prot, 0); prot 5362 drivers/s390/net/qeth_core_main.c enum qeth_prot_versions prot) prot 5369 drivers/s390/net/qeth_core_main.c iob = qeth_ipa_alloc_cmd(card, IPA_CMD_SETASSPARMS, prot, prot 5389 drivers/s390/net/qeth_core_main.c enum qeth_prot_versions prot) prot 5394 drivers/s390/net/qeth_core_main.c QETH_CARD_TEXT_(card, 4, "simassp%i", prot); prot 5395 drivers/s390/net/qeth_core_main.c iob = qeth_get_setassparms_cmd(card, ipa_func, cmd_code, length, prot); prot 5917 drivers/s390/net/qeth_core_main.c enum qeth_prot_versions prot) prot 5920 drivers/s390/net/qeth_core_main.c NULL, prot); prot 5924 drivers/s390/net/qeth_core_main.c enum qeth_prot_versions prot) prot 5933 drivers/s390/net/qeth_core_main.c if (IS_LAYER3(card) && prot == QETH_PROT_IPV4 && prot 5938 drivers/s390/net/qeth_core_main.c prot); prot 5947 drivers/s390/net/qeth_core_main.c qeth_set_csum_off(card, cstype, prot); prot 5953 drivers/s390/net/qeth_core_main.c prot); prot 5955 drivers/s390/net/qeth_core_main.c qeth_set_csum_off(card, cstype, prot); prot 5964 drivers/s390/net/qeth_core_main.c qeth_set_csum_off(card, cstype, prot); prot 5970 drivers/s390/net/qeth_core_main.c qeth_set_csum_off(card, cstype, prot); prot 5975 drivers/s390/net/qeth_core_main.c cstype == IPA_INBOUND_CHECKSUM ? "in" : "out", prot); prot 5985 drivers/s390/net/qeth_core_main.c enum qeth_prot_versions prot) prot 5987 drivers/s390/net/qeth_core_main.c return on ? qeth_set_csum_on(card, cstype, prot) : prot 5988 drivers/s390/net/qeth_core_main.c qeth_set_csum_off(card, cstype, prot); prot 6006 drivers/s390/net/qeth_core_main.c enum qeth_prot_versions prot) prot 6009 drivers/s390/net/qeth_core_main.c IPA_CMD_ASS_STOP, NULL, prot); prot 6013 drivers/s390/net/qeth_core_main.c enum qeth_prot_versions prot) prot 6021 drivers/s390/net/qeth_core_main.c IPA_CMD_ASS_START, 0, prot); prot 6030 drivers/s390/net/qeth_core_main.c qeth_set_tso_off(card, prot); prot 6036 drivers/s390/net/qeth_core_main.c SETASS_DATA_SIZEOF(caps), prot); prot 6038 drivers/s390/net/qeth_core_main.c qeth_set_tso_off(card, prot); prot 6047 drivers/s390/net/qeth_core_main.c qeth_set_tso_off(card, prot); prot 6053 drivers/s390/net/qeth_core_main.c qeth_set_tso_off(card, prot); prot 6057 drivers/s390/net/qeth_core_main.c dev_info(&card->gdev->dev, "TSOv%u enabled (MSS: %u)\n", prot, prot 6063 drivers/s390/net/qeth_core_main.c enum qeth_prot_versions prot) prot 6065 drivers/s390/net/qeth_core_main.c return on ? qeth_set_tso_on(card, prot) : qeth_set_tso_off(card, prot); prot 67 drivers/s390/net/qeth_l3_main.c static struct qeth_ipaddr *qeth_l3_get_addr_buffer(enum qeth_prot_versions prot) prot 72 drivers/s390/net/qeth_l3_main.c qeth_l3_init_ipaddr(addr, QETH_IP_TYPE_NORMAL, prot); prot 458 drivers/s390/net/qeth_l3_main.c enum qeth_routing_types type, enum qeth_prot_versions prot) prot 465 drivers/s390/net/qeth_l3_main.c iob = qeth_ipa_alloc_cmd(card, IPA_CMD_SETRTG, prot, prot 477 drivers/s390/net/qeth_l3_main.c enum qeth_routing_types *type, enum qeth_prot_versions prot) prot 496 drivers/s390/net/qeth_l3_main.c if (qeth_is_ipafunc_supported(card, prot, prot 1324 drivers/s390/net/qeth_l3_main.c u16 prot = (hdr->hdr.l3.flags & QETH_HDR_IPV6) ? ETH_P_IPV6 : prot 1331 drivers/s390/net/qeth_l3_main.c if (prot == ETH_P_IP) prot 1348 drivers/s390/net/qeth_l3_main.c card->dev->header_ops->create(skb, card->dev, prot, prot 1352 drivers/s390/net/qeth_l3_main.c card->dev->header_ops->create(skb, card->dev, prot, prot 1598 drivers/s390/net/qeth_l3_main.c static int arpentry_matches_prot(struct qeth_arp_entrytype *type, __u16 prot) prot 1600 drivers/s390/net/qeth_l3_main.c return (type->ip == QETHARP_IP_ADDR_V4 && prot == QETH_PROT_IPV4) || prot 1601 drivers/s390/net/qeth_l3_main.c (type->ip == QETHARP_IP_ADDR_V6 && prot == QETH_PROT_IPV6); prot 1684 drivers/s390/net/qeth_l3_main.c enum qeth_prot_versions prot, prot 1691 drivers/s390/net/qeth_l3_main.c QETH_CARD_TEXT_(card, 3, "qarpipv%i", prot); prot 1695 drivers/s390/net/qeth_l3_main.c SETASS_DATA_SIZEOF(query_arp), prot); prot 70 drivers/s390/net/qeth_l3_sys.c struct qeth_routing_info *route, enum qeth_prot_versions prot, prot 95 drivers/s390/net/qeth_l3_sys.c if (prot == QETH_PROT_IPV4) prot 97 drivers/s390/net/qeth_l3_sys.c else if (prot == QETH_PROT_IPV6) prot 1218 drivers/scsi/cxgbi/libcxgbi.c unsigned int prot) prot 1220 drivers/scsi/cxgbi/libcxgbi.c struct scsi_data_buffer *sdb = prot ? scsi_prot(sc) : &sc->sdb; prot 526 drivers/scsi/hisi_sas/hisi_sas.h u32 prot[7]; prot 1110 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c struct hisi_sas_protect_iu_v3_hw *prot) prot 1118 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c prot->dw0 |= T10_INSRT_EN_MSK; prot 1119 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c prot->lbrtgv = lbrt_chk_val; prot 1122 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c prot->dw0 |= (T10_RMV_EN_MSK | T10_CHK_EN_MSK); prot 1123 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c prot->lbrtcv = lbrt_chk_val; prot 1124 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c prot->dw4 |= get_prot_chk_msk_v3_hw(scsi_cmnd); prot 1127 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c prot->dw0 |= T10_CHK_EN_MSK; prot 1128 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c prot->lbrtcv = lbrt_chk_val; prot 1129 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c prot->dw4 |= get_prot_chk_msk_v3_hw(scsi_cmnd); prot 1132 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c prot->dw0 |= T10_INSRT_EN_MSK; prot 1133 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c prot->lbrtgv = lbrt_chk_val; prot 1136 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c prot->dw0 |= (T10_RMV_EN_MSK | T10_CHK_EN_MSK); prot 1137 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c prot->lbrtcv = lbrt_chk_val; prot 1140 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c prot->dw0 |= T10_CHK_EN_MSK; prot 1141 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c prot->lbrtcv = lbrt_chk_val; prot 1142 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c prot->dw4 |= get_prot_chk_msk_v3_hw(scsi_cmnd); prot 1153 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c prot->dw0 |= (0x1 << USR_DATA_BLOCK_SZ_OFF); prot 1156 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c prot->dw0 |= (0x2 << USR_DATA_BLOCK_SZ_OFF); prot 1164 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c prot->dw0 |= INCR_LBRT_MSK; prot 1256 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c struct hisi_sas_protect_iu_v3_hw prot; prot 1265 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c memset(&prot, 0, sizeof(struct hisi_sas_protect_iu_v3_hw)); prot 1266 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c fill_prot_v3_hw(scsi_cmnd, &prot); prot 1267 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c memcpy(buf_cmd_prot, &prot, prot 2721 drivers/scsi/lpfc/lpfc_scsi.c int prot, protsegcnt; prot 2732 drivers/scsi/lpfc/lpfc_scsi.c prot = scsi_get_prot_op(cmd); prot 2733 drivers/scsi/lpfc/lpfc_scsi.c if ((prot == SCSI_PROT_READ_STRIP) || prot 2734 drivers/scsi/lpfc/lpfc_scsi.c (prot == SCSI_PROT_WRITE_INSERT) || prot 2735 drivers/scsi/lpfc/lpfc_scsi.c (prot == SCSI_PROT_NORMAL)) prot 2826 drivers/scsi/qla2xxx/qla_attr.c int prot = 0, guard; prot 2832 drivers/scsi/qla2xxx/qla_attr.c prot = SHOST_DIX_TYPE0_PROTECTION; prot 2834 drivers/scsi/qla2xxx/qla_attr.c prot | SHOST_DIF_TYPE1_PROTECTION prot 3351 drivers/scsi/qla2xxx/qla_os.c int prot = 0, guard; prot 3357 drivers/scsi/qla2xxx/qla_os.c prot = SHOST_DIX_TYPE0_PROTECTION; prot 3362 drivers/scsi/qla2xxx/qla_os.c prot | SHOST_DIF_TYPE1_PROTECTION prot 1137 drivers/scsi/scsi_lib.c void *prot = cmd->prot_sdb; prot 1156 drivers/scsi/scsi_lib.c cmd->prot_sdb = prot; prot 949 drivers/slimbus/qcom-ngd-ctrl.c wbuf[txn.msg->num_bytes++] = exp << 4 | rt->prot; prot 951 drivers/slimbus/qcom-ngd-ctrl.c if (rt->prot == SLIM_PROTO_ISO) prot 337 drivers/slimbus/slimbus.h enum slim_transport_protocol prot; prot 230 drivers/slimbus/stream.c rt->prot = SLIM_PROTO_PUSH; prot 232 drivers/slimbus/stream.c rt->prot = SLIM_PROTO_PULL; prot 234 drivers/slimbus/stream.c rt->prot = SLIM_PROTO_ISO; prot 276 drivers/slimbus/stream.c if (stream->prot != SLIM_PROTO_ISO) prot 311 drivers/slimbus/stream.c wbuf[2] = (stream->prot << 4) | ((port->ch.seg_dist & 0xF00) >> 8); prot 312 drivers/slimbus/stream.c if (stream->prot == SLIM_PROTO_ISO) prot 47 drivers/soc/aspeed/aspeed-lpc-ctrl.c pgprot_t prot = vma->vm_page_prot; prot 53 drivers/soc/aspeed/aspeed-lpc-ctrl.c prot = pgprot_noncached(prot); prot 57 drivers/soc/aspeed/aspeed-lpc-ctrl.c vsize, prot)) prot 103 drivers/soc/aspeed/aspeed-p2a-ctrl.c pgprot_t prot; prot 111 drivers/soc/aspeed/aspeed-p2a-ctrl.c prot = vma->vm_page_prot; prot 117 drivers/soc/aspeed/aspeed-p2a-ctrl.c prot = pgprot_noncached(prot); prot 121 drivers/soc/aspeed/aspeed-p2a-ctrl.c vsize, prot)) prot 347 drivers/staging/android/ashmem.c static inline vm_flags_t calc_vm_may_flags(unsigned long prot) prot 349 drivers/staging/android/ashmem.c return _calc_vm_trans(prot, PROT_READ, VM_MAYREAD) | prot 350 drivers/staging/android/ashmem.c _calc_vm_trans(prot, PROT_WRITE, VM_MAYWRITE) | prot 351 drivers/staging/android/ashmem.c _calc_vm_trans(prot, PROT_EXEC, VM_MAYEXEC); prot 527 drivers/staging/android/ashmem.c static int set_prot_mask(struct ashmem_area *asma, unsigned long prot) prot 534 drivers/staging/android/ashmem.c if ((asma->prot_mask & prot) != prot) { prot 540 drivers/staging/android/ashmem.c if ((prot & PROT_READ) && (current->personality & READ_IMPLIES_EXEC)) prot 541 drivers/staging/android/ashmem.c prot |= PROT_EXEC; prot 543 drivers/staging/android/ashmem.c asma->prot_mask = prot; prot 499 drivers/target/target_core_file.c sector_t prot_length, prot; prot 509 drivers/target/target_core_file.c for (prot = 0; prot < prot_length;) { prot 510 drivers/target/target_core_file.c sector_t len = min_t(sector_t, bufsize, prot_length - prot); prot 517 drivers/target/target_core_file.c prot += ret; prot 1914 drivers/vfio/vfio.c int prot, unsigned long *phys_pfn) prot 1939 drivers/vfio/vfio.c npage, prot, phys_pfn); prot 79 drivers/vfio/vfio_iommu_type1.c int prot; /* IOMMU_CACHE */ prot 88 drivers/vfio/vfio_iommu_type1.c int prot; /* IOMMU_READ/WRITE */ prot 127 drivers/vfio/vfio_iommu_type1.c static int put_pfn(unsigned long pfn, int prot); prot 262 drivers/vfio/vfio_iommu_type1.c ret = put_pfn(vpfn->pfn, dma->prot); prot 326 drivers/vfio/vfio_iommu_type1.c static int put_pfn(unsigned long pfn, int prot) prot 330 drivers/vfio/vfio_iommu_type1.c if (prot & IOMMU_WRITE) prot 339 drivers/vfio/vfio_iommu_type1.c int prot, unsigned long *pfn) prot 347 drivers/vfio/vfio_iommu_type1.c if (prot & IOMMU_WRITE) prot 410 drivers/vfio/vfio_iommu_type1.c ret = vaddr_get_pfn(current->mm, vaddr, dma->prot, pfn_base); prot 423 drivers/vfio/vfio_iommu_type1.c put_pfn(*pfn_base, dma->prot); prot 437 drivers/vfio/vfio_iommu_type1.c ret = vaddr_get_pfn(current->mm, vaddr, dma->prot, &pfn); prot 443 drivers/vfio/vfio_iommu_type1.c put_pfn(pfn, dma->prot); prot 450 drivers/vfio/vfio_iommu_type1.c put_pfn(pfn, dma->prot); prot 467 drivers/vfio/vfio_iommu_type1.c put_pfn(pfn, dma->prot); prot 484 drivers/vfio/vfio_iommu_type1.c if (put_pfn(pfn++, dma->prot)) { prot 507 drivers/vfio/vfio_iommu_type1.c ret = vaddr_get_pfn(mm, vaddr, dma->prot, pfn_base); prot 511 drivers/vfio/vfio_iommu_type1.c put_pfn(*pfn_base, dma->prot); prot 543 drivers/vfio/vfio_iommu_type1.c int npage, int prot, prot 585 drivers/vfio/vfio_iommu_type1.c if ((dma->prot & prot) != prot) { prot 986 drivers/vfio/vfio_iommu_type1.c unsigned long pfn, long npage, int prot) prot 993 drivers/vfio/vfio_iommu_type1.c npage << PAGE_SHIFT, prot | d->prot); prot 1031 drivers/vfio/vfio_iommu_type1.c dma->prot); prot 1077 drivers/vfio/vfio_iommu_type1.c int ret = 0, prot = 0; prot 1091 drivers/vfio/vfio_iommu_type1.c prot |= IOMMU_WRITE; prot 1093 drivers/vfio/vfio_iommu_type1.c prot |= IOMMU_READ; prot 1095 drivers/vfio/vfio_iommu_type1.c if (!prot || !size || (size | iova | vaddr) & mask) prot 1128 drivers/vfio/vfio_iommu_type1.c dma->prot = prot; prot 1251 drivers/vfio/vfio_iommu_type1.c size, dma->prot | domain->prot); prot 1282 drivers/vfio/vfio_iommu_type1.c IOMMU_READ | IOMMU_WRITE | domain->prot); prot 1793 drivers/vfio/vfio_iommu_type1.c domain->prot |= IOMMU_CACHE; prot 1804 drivers/vfio/vfio_iommu_type1.c d->prot == domain->prot) { prot 2137 drivers/vfio/vfio_iommu_type1.c if (!(domain->prot & IOMMU_CACHE)) { prot 992 drivers/video/fbdev/vermilion/vermilion.c unsigned long prot; prot 998 drivers/video/fbdev/vermilion/vermilion.c prot = pgprot_val(vma->vm_page_prot) & ~_PAGE_CACHE_MASK; prot 1000 drivers/video/fbdev/vermilion/vermilion.c prot | cachemode2protval(_PAGE_CACHE_MODE_UC_MINUS); prot 67 drivers/xen/xlate_mmu.c pgprot_t prot; prot 100 drivers/xen/xlate_mmu.c pte_t pte = pte_mkspecial(pfn_pte(page_to_pfn(page), info->prot)); prot 146 drivers/xen/xlate_mmu.c int *err_ptr, pgprot_t prot, prot 160 drivers/xen/xlate_mmu.c data.prot = prot; prot 269 drivers/xen/xlate_mmu.c pgprot_t prot; prot 277 drivers/xen/xlate_mmu.c pte_t pte = pte_mkspecial(pfn_pte(page_to_pfn(page), r->prot)); prot 291 drivers/xen/xlate_mmu.c .prot = vma->vm_page_prot, prot 143 fs/affs/affs.h extern umode_t affs_prot_to_mode(u32 prot); prot 391 fs/affs/amigaffs.c affs_prot_to_mode(u32 prot) prot 395 fs/affs/amigaffs.c if (!(prot & FIBF_NOWRITE)) prot 397 fs/affs/amigaffs.c if (!(prot & FIBF_NOREAD)) prot 399 fs/affs/amigaffs.c if (!(prot & FIBF_NOEXECUTE)) prot 401 fs/affs/amigaffs.c if (prot & FIBF_GRP_WRITE) prot 403 fs/affs/amigaffs.c if (prot & FIBF_GRP_READ) prot 405 fs/affs/amigaffs.c if (prot & FIBF_GRP_EXECUTE) prot 407 fs/affs/amigaffs.c if (prot & FIBF_OTR_WRITE) prot 409 fs/affs/amigaffs.c if (prot & FIBF_OTR_READ) prot 411 fs/affs/amigaffs.c if (prot & FIBF_OTR_EXECUTE) prot 420 fs/affs/amigaffs.c u32 prot = AFFS_I(inode)->i_protect; prot 424 fs/affs/amigaffs.c prot |= FIBF_NOEXECUTE; prot 426 fs/affs/amigaffs.c prot |= FIBF_NOREAD; prot 428 fs/affs/amigaffs.c prot |= FIBF_NOWRITE; prot 430 fs/affs/amigaffs.c prot |= FIBF_GRP_EXECUTE; prot 432 fs/affs/amigaffs.c prot |= FIBF_GRP_READ; prot 434 fs/affs/amigaffs.c prot |= FIBF_GRP_WRITE; prot 436 fs/affs/amigaffs.c prot |= FIBF_OTR_EXECUTE; prot 438 fs/affs/amigaffs.c prot |= FIBF_OTR_READ; prot 440 fs/affs/amigaffs.c prot |= FIBF_OTR_WRITE; prot 442 fs/affs/amigaffs.c AFFS_I(inode)->i_protect = prot; prot 26 fs/affs/inode.c u32 prot; prot 51 fs/affs/inode.c prot = be32_to_cpu(tail->protect); prot 58 fs/affs/inode.c AFFS_I(inode)->i_protect = prot; prot 74 fs/affs/inode.c inode->i_mode = affs_prot_to_mode(prot); prot 102 fs/binfmt_elf.c static int set_brk(unsigned long start, unsigned long end, int prot) prot 113 fs/binfmt_elf.c prot & PROT_EXEC ? VM_EXEC : 0); prot 349 fs/binfmt_elf.c const struct elf_phdr *eppnt, int prot, int type, prot 373 fs/binfmt_elf.c map_addr = vm_mmap(filep, addr, total_size, prot, type, off); prot 377 fs/binfmt_elf.c map_addr = vm_mmap(filep, addr, size, prot, type, off); prot 530 fs/binfmt_elf.c int prot = 0; prot 533 fs/binfmt_elf.c prot |= PROT_READ; prot 535 fs/binfmt_elf.c prot |= PROT_WRITE; prot 537 fs/binfmt_elf.c prot |= PROT_EXEC; prot 538 fs/binfmt_elf.c return prot; prot 1023 fs/binfmt_elf_fdpic.c int prot = 0, flags; prot 1035 fs/binfmt_elf_fdpic.c if (phdr->p_flags & PF_R) prot |= PROT_READ; prot 1036 fs/binfmt_elf_fdpic.c if (phdr->p_flags & PF_W) prot |= PROT_WRITE; prot 1037 fs/binfmt_elf_fdpic.c if (phdr->p_flags & PF_X) prot |= PROT_EXEC; prot 1083 fs/binfmt_elf_fdpic.c maddr = vm_mmap(file, maddr, phdr->p_memsz + disp, prot, flags, prot 1087 fs/binfmt_elf_fdpic.c loop, phdr->p_memsz + disp, prot, flags, prot 1107 fs/binfmt_elf_fdpic.c if (prot & PROT_WRITE && disp > 0) { prot 1129 fs/binfmt_elf_fdpic.c prot, flags, 0); prot 1133 fs/binfmt_elf_fdpic.c loop, xaddr, excess - excess1, prot, flags, prot 1140 fs/binfmt_elf_fdpic.c if (prot & PROT_WRITE && excess1 > 0) { prot 189 fs/proc/vmcore.c unsigned long size, pgprot_t prot) prot 191 fs/proc/vmcore.c prot = pgprot_encrypted(prot); prot 192 fs/proc/vmcore.c return remap_pfn_range(vma, from, pfn, size, prot); prot 489 fs/proc/vmcore.c unsigned long size, pgprot_t prot) prot 511 fs/proc/vmcore.c prot)) prot 518 fs/proc/vmcore.c PAGE_SIZE, prot)) prot 528 fs/proc/vmcore.c map_size, prot)) prot 539 fs/proc/vmcore.c unsigned long size, pgprot_t prot) prot 546 fs/proc/vmcore.c return remap_oldmem_pfn_checked(vma, from, pfn, size, prot); prot 548 fs/proc/vmcore.c return remap_oldmem_pfn_range(vma, from, pfn, size, prot); prot 407 fs/pstore/ram_core.c pgprot_t prot; prot 415 fs/pstore/ram_core.c prot = pgprot_noncached(PAGE_KERNEL); prot 417 fs/pstore/ram_core.c prot = pgprot_writecombine(PAGE_KERNEL); prot 430 fs/pstore/ram_core.c vaddr = vmap(pages, page_count, VM_MAP, prot); prot 482 include/asm-generic/pgtable.h #define move_pte(pte, prot, old_addr, new_addr) (pte) prot 494 include/asm-generic/pgtable.h #define pgprot_noncached(prot) (prot) prot 672 include/asm-generic/pgtable.h #define pgprot_encrypted(prot) (prot) prot 676 include/asm-generic/pgtable.h #define pgprot_decrypted(prot) (prot) prot 805 include/asm-generic/pgtable.h static inline int track_pfn_remap(struct vm_area_struct *vma, pgprot_t *prot, prot 816 include/asm-generic/pgtable.h static inline void track_pfn_insert(struct vm_area_struct *vma, pgprot_t *prot, prot 847 include/asm-generic/pgtable.h extern int track_pfn_remap(struct vm_area_struct *vma, pgprot_t *prot, prot 850 include/asm-generic/pgtable.h extern void track_pfn_insert(struct vm_area_struct *vma, pgprot_t *prot, prot 1042 include/asm-generic/pgtable.h int p4d_set_huge(p4d_t *p4d, phys_addr_t addr, pgprot_t prot); prot 1045 include/asm-generic/pgtable.h static inline int p4d_set_huge(p4d_t *p4d, phys_addr_t addr, pgprot_t prot) prot 1055 include/asm-generic/pgtable.h int pud_set_huge(pud_t *pud, phys_addr_t addr, pgprot_t prot); prot 1056 include/asm-generic/pgtable.h int pmd_set_huge(pmd_t *pmd, phys_addr_t addr, pgprot_t prot); prot 1063 include/asm-generic/pgtable.h static inline int p4d_set_huge(p4d_t *p4d, phys_addr_t addr, pgprot_t prot) prot 1067 include/asm-generic/pgtable.h static inline int pud_set_huge(pud_t *pud, phys_addr_t addr, pgprot_t prot) prot 1071 include/asm-generic/pgtable.h static inline int pmd_set_huge(pmd_t *pmd, phys_addr_t addr, pgprot_t prot) prot 1131 include/asm-generic/pgtable.h static inline bool pfn_modify_allowed(unsigned long pfn, pgprot_t prot) prot 15 include/asm-generic/syscalls.h unsigned long prot, unsigned long flags, prot 21 include/asm-generic/syscalls.h unsigned long prot, unsigned long flags, prot 737 include/drm/ttm/ttm_bo_api.h void *ttm_kmap_atomic_prot(struct page *page, pgprot_t prot); prot 739 include/drm/ttm/ttm_bo_api.h void ttm_kunmap_atomic_prot(void *addr, pgprot_t prot); prot 56 include/linux/agpgart.h int prot; /* prot flags for mmap */ prot 62 include/linux/agpgart.h pgprot_t prot; prot 41 include/linux/can/core.h struct proto *prot; prot 25 include/linux/crash_dump.h unsigned long size, pgprot_t prot); prot 630 include/linux/dma-mapping.h pgprot_t prot, const void *caller); prot 633 include/linux/dma-mapping.h pgprot_t prot, const void *caller); prot 55 include/linux/dma-noncoherent.h #define pgprot_dmacoherent(prot) pgprot_noncached(prot) prot 58 include/linux/dma-noncoherent.h pgprot_t dma_pgprot(struct device *dev, pgprot_t prot, unsigned long attrs); prot 60 include/linux/dma-noncoherent.h static inline pgprot_t dma_pgprot(struct device *dev, pgprot_t prot, prot 63 include/linux/dma-noncoherent.h return prot; /* no protection bits supported without page tables */ prot 97 include/linux/highmem.h #define kmap_atomic_prot(page, prot) kmap_atomic(page) prot 268 include/linux/huge_mm.h #define mk_huge_pmd(page, prot) pmd_mkhuge(mk_pmd(page, prot)) prot 20 include/linux/ima.h extern int ima_file_mmap(struct file *file, unsigned long prot); prot 67 include/linux/ima.h static inline int ima_file_mmap(struct file *file, unsigned long prot) prot 25 include/linux/io-mapping.h pgprot_t prot; prot 44 include/linux/io-mapping.h pgprot_t prot; prot 46 include/linux/io-mapping.h if (iomap_create_wc(base, size, &prot)) prot 51 include/linux/io-mapping.h iomap->prot = prot; prot 72 include/linux/io-mapping.h return iomap_atomic_prot_pfn(pfn, mapping->prot); prot 115 include/linux/io-mapping.h iomap->prot = pgprot_noncached_wc(PAGE_KERNEL); prot 117 include/linux/io-mapping.h iomap->prot = pgprot_writecombine(PAGE_KERNEL); prot 119 include/linux/io-mapping.h iomap->prot = pgprot_noncached(PAGE_KERNEL); prot 139 include/linux/io-pgtable.h phys_addr_t paddr, size_t size, int prot); prot 25 include/linux/io.h phys_addr_t phys_addr, pgprot_t prot); prot 28 include/linux/io.h phys_addr_t phys_addr, pgprot_t prot) prot 161 include/linux/iommu.h int prot; prot 259 include/linux/iommu.h phys_addr_t paddr, size_t size, int prot); prot 284 include/linux/iommu.h phys_addr_t paddr, u64 size, int prot); prot 423 include/linux/iommu.h phys_addr_t paddr, size_t size, int prot); prot 430 include/linux/iommu.h struct scatterlist *sg,unsigned int nents, int prot); prot 443 include/linux/iommu.h iommu_alloc_resv_region(phys_addr_t start, size_t length, int prot, prot 493 include/linux/iommu.h int prot); prot 660 include/linux/iommu.h phys_addr_t paddr, size_t size, int prot) prot 680 include/linux/iommu.h unsigned int nents, int prot) prot 696 include/linux/iommu.h u64 size, int prot) prot 1029 include/linux/libata.h static inline bool ata_is_atapi(u8 prot) prot 1031 include/linux/libata.h return prot & ATA_PROT_FLAG_ATAPI; prot 1034 include/linux/libata.h static inline bool ata_is_pio(u8 prot) prot 1036 include/linux/libata.h return prot & ATA_PROT_FLAG_PIO; prot 1039 include/linux/libata.h static inline bool ata_is_dma(u8 prot) prot 1041 include/linux/libata.h return prot & ATA_PROT_FLAG_DMA; prot 1044 include/linux/libata.h static inline bool ata_is_ncq(u8 prot) prot 1046 include/linux/libata.h return prot & ATA_PROT_FLAG_NCQ; prot 1049 include/linux/libata.h static inline bool ata_is_data(u8 prot) prot 1051 include/linux/libata.h return prot & (ATA_PROT_FLAG_PIO | ATA_PROT_FLAG_DMA); prot 1607 include/linux/lsm_hooks.h unsigned long prot, unsigned long flags); prot 1609 include/linux/lsm_hooks.h unsigned long prot); prot 1160 include/linux/mlx4/device.h int block_mcast_loopback, enum mlx4_protocol prot); prot 1162 include/linux/mlx4/device.h enum mlx4_protocol prot); prot 1475 include/linux/mm.h unsigned int flags, unsigned long *prot, resource_size_t *phys); prot 2346 include/linux/mm.h unsigned long len, unsigned long prot, unsigned long flags, prot 2356 include/linux/mm.h unsigned long len, unsigned long prot, unsigned long flags, prot 2360 include/linux/mm.h return do_mmap(file, addr, len, prot, flags, 0, pgoff, populate, uf); prot 81 include/linux/mman.h #define arch_calc_vm_prot_bits(prot, pkey) 0 prot 95 include/linux/mman.h static inline bool arch_validate_prot(unsigned long prot, unsigned long addr) prot 97 include/linux/mman.h return (prot & ~(PROT_READ | PROT_WRITE | PROT_EXEC | PROT_SEM)) == 0; prot 117 include/linux/mman.h calc_vm_prot_bits(unsigned long prot, unsigned long pkey) prot 119 include/linux/mman.h return _calc_vm_trans(prot, PROT_READ, VM_READ ) | prot 120 include/linux/mman.h _calc_vm_trans(prot, PROT_WRITE, VM_WRITE) | prot 121 include/linux/mman.h _calc_vm_trans(prot, PROT_EXEC, VM_EXEC) | prot 122 include/linux/mman.h arch_calc_vm_prot_bits(prot, pkey); prot 12 include/linux/pkeys.h #define arch_override_mprotect_pkey(vma, prot, pkey) (0) prot 152 include/linux/security.h unsigned long prot, unsigned long flags); prot 363 include/linux/security.h int security_mmap_file(struct file *file, unsigned long prot, prot 367 include/linux/security.h unsigned long prot); prot 908 include/linux/security.h static inline int security_mmap_file(struct file *file, unsigned long prot, prot 921 include/linux/security.h unsigned long prot) prot 374 include/linux/sunrpc/svc.h u32 prot; /* protocol (UDP or TCP) */ prot 66 include/linux/sunrpc/svcsock.h struct svc_xprt *svc_sock_create(struct svc_serv *serv, int prot); prot 193 include/linux/sunrpc/xprt.h int prot; /* IP protocol */ prot 868 include/linux/syscalls.h unsigned long prot); prot 878 include/linux/syscalls.h unsigned long prot, unsigned long pgoff, prot 984 include/linux/syscalls.h unsigned long prot, int pkey); prot 1214 include/linux/syscalls.h unsigned long prot, unsigned long flags, prot 1268 include/linux/syscalls.h unsigned long prot, unsigned long flags, prot 76 include/linux/vfio.h int npage, int prot, prot 106 include/linux/vfio.h int npage, int prot, unsigned long *phys_pfn); prot 78 include/linux/vmalloc.h int node, pgprot_t prot); prot 99 include/linux/vmalloc.h extern void *__vmalloc(unsigned long size, gfp_t gfp_mask, pgprot_t prot); prot 102 include/linux/vmalloc.h pgprot_t prot, unsigned long vm_flags, int node, prot 120 include/linux/vmalloc.h unsigned long flags, pgprot_t prot); prot 158 include/linux/vmalloc.h extern int map_vm_area(struct vm_struct *area, pgprot_t prot, prot 162 include/linux/vmalloc.h pgprot_t prot, struct page **pages); prot 175 include/linux/vmalloc.h pgprot_t prot, struct page **pages) prot 99 include/net/llc_conn.h struct proto *prot, int kern); prot 94 include/net/phonet/phonet.h struct proto *prot; prot 88 include/net/protocol.h struct proto *prot; prot 105 include/net/protocol.h int inet_add_protocol(const struct net_protocol *prot, unsigned char num); prot 106 include/net/protocol.h int inet_del_protocol(const struct net_protocol *prot, unsigned char num); prot 107 include/net/protocol.h int inet_add_offload(const struct net_offload *prot, unsigned char num); prot 108 include/net/protocol.h int inet_del_offload(const struct net_offload *prot, unsigned char num); prot 113 include/net/protocol.h int inet6_add_protocol(const struct inet6_protocol *prot, unsigned char num); prot 114 include/net/protocol.h int inet6_del_protocol(const struct inet6_protocol *prot, unsigned char num); prot 118 include/net/protocol.h int inet6_add_offload(const struct net_offload *prot, unsigned char num); prot 119 include/net/protocol.h int inet6_del_offload(const struct net_offload *prot, unsigned char num); prot 1189 include/net/sock.h int proto_register(struct proto *prot, int alloc_slab); prot 1190 include/net/sock.h void proto_unregister(struct proto *prot); prot 1306 include/net/sock.h proto_sockets_allocated_sum_positive(struct proto *prot) prot 1308 include/net/sock.h return percpu_counter_sum_positive(prot->sockets_allocated); prot 1312 include/net/sock.h proto_memory_allocated(struct proto *prot) prot 1314 include/net/sock.h return atomic_long_read(prot->memory_allocated); prot 1318 include/net/sock.h proto_memory_pressure(struct proto *prot) prot 1320 include/net/sock.h if (!prot->memory_pressure) prot 1322 include/net/sock.h return !!*prot->memory_pressure; prot 1328 include/net/sock.h void sock_prot_inuse_add(struct net *net, struct proto *prot, int inc); prot 1332 include/net/sock.h static inline void sock_prot_inuse_add(struct net *net, struct proto *prot, prot 1603 include/net/sock.h struct proto *prot, int kern); prot 492 include/net/tls.h struct tls_prot_info *prot, prot 495 include/net/tls.h if (tls_bigint_increment(ctx->rec_seq, prot->rec_seq_size)) prot 498 include/net/tls.h if (prot->version != TLS_1_3_VERSION) prot 500 include/net/tls.h prot->iv_size); prot 509 include/net/tls.h struct tls_prot_info *prot = &ctx->prot_info; prot 510 include/net/tls.h size_t pkt_len, iv_size = prot->iv_size; prot 512 include/net/tls.h pkt_len = plaintext_len + prot->tag_size; prot 2797 include/rdma/ib_verbs.h unsigned long pfn, unsigned long size, pgprot_t prot); prot 2802 include/rdma/ib_verbs.h pgprot_t prot) prot 94 include/trace/events/sock.h TP_PROTO(struct sock *sk, struct proto *prot, long allocated, int kind), prot 96 include/trace/events/sock.h TP_ARGS(sk, prot, allocated, kind), prot 111 include/trace/events/sock.h strncpy(__entry->name, prot->name, 32); prot 112 include/trace/events/sock.h __entry->sysctl_mem = prot->sysctl_mem; prot 114 include/trace/events/sock.h __entry->sysctl_rmem = sk_get_rmem0(sk, prot); prot 116 include/trace/events/sock.h __entry->sysctl_wmem = sk_get_wmem0(sk, prot); prot 83 include/uapi/linux/agpgart.h int prot; /* prot flags for mmap */ prot 123 include/uapi/linux/scif_ioctl.h __s32 prot; prot 66 include/xen/xen-ops.h xen_pfn_t *pfn, int nr, int *err_ptr, pgprot_t prot, prot 71 include/xen/xen-ops.h pgprot_t prot, unsigned int domid, prot 85 include/xen/xen-ops.h int *err_ptr, pgprot_t prot, prot 98 include/xen/xen-ops.h int *err_ptr, pgprot_t prot, prot 135 include/xen/xen-ops.h int *err_ptr, pgprot_t prot, prot 141 include/xen/xen-ops.h prot, domid, pages); prot 148 include/xen/xen-ops.h return xen_remap_pfn(vma, addr, gfn, nr, err_ptr, prot, domid, prot 172 include/xen/xen-ops.h pgprot_t prot, unsigned int domid, prot 178 include/xen/xen-ops.h return xen_remap_pfn(vma, addr, mfn, nr, err_ptr, prot, domid, prot 197 include/xen/xen-ops.h pgprot_t prot, unsigned int domid, prot 203 include/xen/xen-ops.h return xen_remap_pfn(vma, addr, &gfn, nr, NULL, prot, domid, false, prot 1427 ipc/shm.c unsigned long prot; prot 1462 ipc/shm.c prot = PROT_READ; prot 1466 ipc/shm.c prot = PROT_READ | PROT_WRITE; prot 1471 ipc/shm.c prot |= PROT_EXEC; prot 1543 ipc/shm.c err = security_mmap_file(file, prot, flags); prot 1561 ipc/shm.c addr = do_mmap_pgoff(file, addr, size, prot, flags, 0, &populate, NULL); prot 170 kernel/dma/mapping.c pgprot_t dma_pgprot(struct device *dev, pgprot_t prot, unsigned long attrs) prot 173 kernel/dma/mapping.c prot = pgprot_decrypted(prot); prot 177 kernel/dma/mapping.c return prot; prot 180 kernel/dma/mapping.c return pgprot_writecombine(prot); prot 182 kernel/dma/mapping.c return pgprot_dmacoherent(prot); prot 24 kernel/dma/remap.c size_t size, pgprot_t prot, const void *caller) prot 32 kernel/dma/remap.c if (map_vm_area(area, prot, pages)) { prot 45 kernel/dma/remap.c pgprot_t prot, const void *caller) prot 49 kernel/dma/remap.c area = __dma_common_pages_remap(pages, size, prot, caller); prot 63 kernel/dma/remap.c pgprot_t prot, const void *caller) prot 76 kernel/dma/remap.c area = __dma_common_pages_remap(pages, size, prot, caller); prot 7363 kernel/events/core.c u32 prot, flags; prot 7406 kernel/events/core.c mmap_event->event_id.header.size += sizeof(mmap_event->prot); prot 7426 kernel/events/core.c perf_output_put(&handle, mmap_event->prot); prot 7447 kernel/events/core.c u32 prot = 0, flags = 0; prot 7454 kernel/events/core.c prot |= PROT_READ; prot 7456 kernel/events/core.c prot |= PROT_WRITE; prot 7458 kernel/events/core.c prot |= PROT_EXEC; prot 7546 kernel/events/core.c mmap_event->prot = prot; prot 64 lib/ioremap.c unsigned long end, phys_addr_t phys_addr, pgprot_t prot) prot 75 lib/ioremap.c set_pte_at(&init_mm, addr, pte, pfn_pte(pfn, prot)); prot 83 lib/ioremap.c pgprot_t prot) prot 100 lib/ioremap.c return pmd_set_huge(pmd, phys_addr, prot); prot 104 lib/ioremap.c unsigned long end, phys_addr_t phys_addr, pgprot_t prot) prot 115 lib/ioremap.c if (ioremap_try_huge_pmd(pmd, addr, next, phys_addr, prot)) prot 118 lib/ioremap.c if (ioremap_pte_range(pmd, addr, next, phys_addr, prot)) prot 126 lib/ioremap.c pgprot_t prot) prot 143 lib/ioremap.c return pud_set_huge(pud, phys_addr, prot); prot 147 lib/ioremap.c unsigned long end, phys_addr_t phys_addr, pgprot_t prot) prot 158 lib/ioremap.c if (ioremap_try_huge_pud(pud, addr, next, phys_addr, prot)) prot 161 lib/ioremap.c if (ioremap_pmd_range(pud, addr, next, phys_addr, prot)) prot 169 lib/ioremap.c pgprot_t prot) prot 186 lib/ioremap.c return p4d_set_huge(p4d, phys_addr, prot); prot 190 lib/ioremap.c unsigned long end, phys_addr_t phys_addr, pgprot_t prot) prot 201 lib/ioremap.c if (ioremap_try_huge_p4d(p4d, addr, next, phys_addr, prot)) prot 204 lib/ioremap.c if (ioremap_pud_range(p4d, addr, next, phys_addr, prot)) prot 211 lib/ioremap.c unsigned long end, phys_addr_t phys_addr, pgprot_t prot) prot 225 lib/ioremap.c err = ioremap_p4d_range(pgd, addr, next, phys_addr, prot); prot 36 mm/early_ioremap.c pgprot_t prot) prot 38 mm/early_ioremap.c return prot; prot 58 mm/early_ioremap.c phys_addr_t phys, pgprot_t prot) prot 106 mm/early_ioremap.c __early_ioremap(resource_size_t phys_addr, unsigned long size, pgprot_t prot) prot 154 mm/early_ioremap.c __late_set_fixmap(idx, phys_addr, prot); prot 156 mm/early_ioremap.c __early_set_fixmap(idx, phys_addr, prot); prot 226 mm/early_ioremap.c pgprot_t prot = early_memremap_pgprot_adjust(phys_addr, size, prot 229 mm/early_ioremap.c return (__force void *)__early_ioremap(phys_addr, size, prot); prot 235 mm/early_ioremap.c pgprot_t prot = early_memremap_pgprot_adjust(phys_addr, size, prot 238 mm/early_ioremap.c return (__force void *)__early_ioremap(phys_addr, size, prot); prot 770 mm/huge_memory.c pmd_t *pmd, pfn_t pfn, pgprot_t prot, bool write, prot 793 mm/huge_memory.c entry = pmd_mkhuge(pfn_t_pmd(pfn, prot)); prot 859 mm/huge_memory.c pud_t *pud, pfn_t pfn, pgprot_t prot, bool write) prot 880 mm/huge_memory.c entry = pud_mkhuge(pfn_t_pud(pfn, prot)); prot 1423 mm/memory.c struct page *page, pgprot_t prot) prot 1446 mm/memory.c set_pte_at(mm, addr, pte, mk_pte(page, prot)); prot 1582 mm/memory.c pfn_t pfn, pgprot_t prot, bool mkwrite) prot 1617 mm/memory.c entry = pte_mkdevmap(pfn_t_pte(pfn, prot)); prot 1619 mm/memory.c entry = pte_mkspecial(pfn_t_pte(pfn, prot)); prot 1793 mm/memory.c unsigned long pfn, pgprot_t prot) prot 1805 mm/memory.c if (!pfn_modify_allowed(pfn, prot)) { prot 1809 mm/memory.c set_pte_at(mm, addr, pte, pte_mkspecial(pfn_pte(pfn, prot))); prot 1819 mm/memory.c unsigned long pfn, pgprot_t prot) prot 1833 mm/memory.c pfn + (addr >> PAGE_SHIFT), prot); prot 1842 mm/memory.c unsigned long pfn, pgprot_t prot) prot 1855 mm/memory.c pfn + (addr >> PAGE_SHIFT), prot); prot 1864 mm/memory.c unsigned long pfn, pgprot_t prot) prot 1877 mm/memory.c pfn + (addr >> PAGE_SHIFT), prot); prot 1897 mm/memory.c unsigned long pfn, unsigned long size, pgprot_t prot) prot 1930 mm/memory.c err = track_pfn_remap(vma, &prot, remap_pfn, addr, PAGE_ALIGN(size)); prot 1943 mm/memory.c pfn + (addr >> PAGE_SHIFT), prot); prot 4257 mm/memory.c unsigned long *prot, resource_size_t *phys) prot 4273 mm/memory.c *prot = pgprot_val(pte_pgprot(pte)); prot 4287 mm/memory.c unsigned long prot = 0; prot 4291 mm/memory.c if (follow_phys(vma, addr, write, &prot, &phys_addr)) prot 4294 mm/memory.c maddr = ioremap_prot(phys_addr, PAGE_ALIGN(len + offset), prot); prot 100 mm/mmap.c static inline pgprot_t arch_filter_pgprot(pgprot_t prot) prot 102 mm/mmap.c return prot; prot 1390 mm/mmap.c unsigned long len, unsigned long prot, prot 1409 mm/mmap.c if ((prot & PROT_READ) && (current->personality & READ_IMPLIES_EXEC)) prot 1411 mm/mmap.c prot |= PROT_EXEC; prot 1447 mm/mmap.c if (prot == PROT_EXEC) { prot 1457 mm/mmap.c vm_flags |= calc_vm_prot_bits(prot, pkey) | calc_vm_flag_bits(flags) | prot 1490 mm/mmap.c if (prot & PROT_WRITE) { prot 1578 mm/mmap.c unsigned long prot, unsigned long flags, prot 1619 mm/mmap.c retval = vm_mmap_pgoff(file, addr, len, prot, flags, pgoff); prot 1627 mm/mmap.c unsigned long, prot, unsigned long, flags, prot 1630 mm/mmap.c return ksys_mmap_pgoff(addr, len, prot, flags, fd, pgoff); prot 1637 mm/mmap.c unsigned long prot; prot 1652 mm/mmap.c return ksys_mmap_pgoff(a.addr, a.len, a.prot, a.flags, a.fd, prot 2892 mm/mmap.c unsigned long, prot, unsigned long, pgoff, unsigned long, flags) prot 2904 mm/mmap.c if (prot) prot 2949 mm/mmap.c prot |= vma->vm_flags & VM_READ ? PROT_READ : 0; prot 2950 mm/mmap.c prot |= vma->vm_flags & VM_WRITE ? PROT_WRITE : 0; prot 2951 mm/mmap.c prot |= vma->vm_flags & VM_EXEC ? PROT_EXEC : 0; prot 2976 mm/mmap.c prot, flags, pgoff, &populate, NULL); prot 487 mm/mprotect.c unsigned long prot, int pkey) prot 492 mm/mprotect.c const int grows = prot & (PROT_GROWSDOWN|PROT_GROWSUP); prot 494 mm/mprotect.c (prot & PROT_READ); prot 498 mm/mprotect.c prot &= ~(PROT_GROWSDOWN|PROT_GROWSUP); prot 510 mm/mprotect.c if (!arch_validate_prot(prot, start)) prot 513 mm/mprotect.c reqprot = prot; prot 560 mm/mprotect.c prot |= PROT_EXEC; prot 570 mm/mprotect.c new_vma_pkey = arch_override_mprotect_pkey(vma, prot, pkey); prot 571 mm/mprotect.c newflags = calc_vm_prot_bits(prot, new_vma_pkey); prot 580 mm/mprotect.c error = security_file_mprotect(vma, reqprot, prot); prot 602 mm/mprotect.c prot = reqprot; prot 610 mm/mprotect.c unsigned long, prot) prot 612 mm/mprotect.c return do_mprotect_pkey(start, len, prot, -1); prot 618 mm/mprotect.c unsigned long, prot, int, pkey) prot 620 mm/mprotect.c return do_mprotect_pkey(start, len, prot, pkey); prot 143 mm/nommu.c void *__vmalloc(unsigned long size, gfp_t gfp_mask, pgprot_t prot) prot 330 mm/nommu.c void *vmap(struct page **pages, unsigned int count, unsigned long flags, pgprot_t prot) prot 343 mm/nommu.c void *vm_map_ram(struct page **pages, unsigned int count, int node, pgprot_t prot) prot 787 mm/nommu.c unsigned long prot, prot 860 mm/nommu.c if ((prot & PROT_WRITE) && prot 884 mm/nommu.c if (prot & PROT_WRITE) prot 889 mm/nommu.c if (((prot & PROT_READ) && !(capabilities & NOMMU_MAP_READ)) || prot 890 mm/nommu.c ((prot & PROT_WRITE) && !(capabilities & NOMMU_MAP_WRITE)) || prot 891 mm/nommu.c ((prot & PROT_EXEC) && !(capabilities & NOMMU_MAP_EXEC)) prot 904 mm/nommu.c if (prot & PROT_EXEC) prot 906 mm/nommu.c } else if ((prot & PROT_READ) && !(prot & PROT_EXEC)) { prot 910 mm/nommu.c prot |= PROT_EXEC; prot 912 mm/nommu.c } else if ((prot & PROT_READ) && prot 913 mm/nommu.c (prot & PROT_EXEC) && prot 926 mm/nommu.c if ((prot & PROT_READ) && prot 928 mm/nommu.c prot |= PROT_EXEC; prot 946 mm/nommu.c unsigned long prot, prot 952 mm/nommu.c vm_flags = calc_vm_prot_bits(prot, 0) | calc_vm_flag_bits(flags); prot 958 mm/nommu.c if (file && !(prot & PROT_WRITE)) prot 1101 mm/nommu.c unsigned long prot, prot 1118 mm/nommu.c ret = validate_mmap_request(file, addr, len, prot, flags, pgoff, prot 1129 mm/nommu.c vm_flags |= determine_vm_flags(file, prot, flags, capabilities); prot 1325 mm/nommu.c unsigned long prot, unsigned long flags, prot 1340 mm/nommu.c retval = vm_mmap_pgoff(file, addr, len, prot, flags, pgoff); prot 1349 mm/nommu.c unsigned long, prot, unsigned long, flags, prot 1352 mm/nommu.c return ksys_mmap_pgoff(addr, len, prot, flags, fd, pgoff); prot 1359 mm/nommu.c unsigned long prot; prot 1374 mm/nommu.c return ksys_mmap_pgoff(a.addr, a.len, a.prot, a.flags, a.fd, prot 1652 mm/nommu.c unsigned long pfn, unsigned long size, pgprot_t prot) prot 484 mm/util.c unsigned long len, unsigned long prot, prot 492 mm/util.c ret = security_mmap_file(file, prot, flag); prot 496 mm/util.c ret = do_mmap_pgoff(file, addr, len, prot, flag, pgoff, prot 507 mm/util.c unsigned long len, unsigned long prot, prot 515 mm/util.c return vm_mmap_pgoff(file, addr, len, prot, flag, offset >> PAGE_SHIFT); prot 139 mm/vmalloc.c unsigned long end, pgprot_t prot, struct page **pages, int *nr) prot 158 mm/vmalloc.c set_pte_at(&init_mm, addr, pte, mk_pte(page, prot)); prot 165 mm/vmalloc.c unsigned long end, pgprot_t prot, struct page **pages, int *nr) prot 175 mm/vmalloc.c if (vmap_pte_range(pmd, addr, next, prot, pages, nr)) prot 182 mm/vmalloc.c unsigned long end, pgprot_t prot, struct page **pages, int *nr) prot 192 mm/vmalloc.c if (vmap_pmd_range(pud, addr, next, prot, pages, nr)) prot 199 mm/vmalloc.c unsigned long end, pgprot_t prot, struct page **pages, int *nr) prot 209 mm/vmalloc.c if (vmap_pud_range(p4d, addr, next, prot, pages, nr)) prot 222 mm/vmalloc.c pgprot_t prot, struct page **pages) prot 234 mm/vmalloc.c err = vmap_p4d_range(pgd, addr, next, prot, pages, &nr); prot 243 mm/vmalloc.c pgprot_t prot, struct page **pages) prot 247 mm/vmalloc.c ret = vmap_page_range_noflush(start, end, prot, pages); prot 1778 mm/vmalloc.c void *vm_map_ram(struct page **pages, unsigned int count, int node, pgprot_t prot) prot 1799 mm/vmalloc.c if (vmap_page_range(addr, addr + size, prot, pages) < 0) { prot 1963 mm/vmalloc.c pgprot_t prot, struct page **pages) prot 1965 mm/vmalloc.c return vmap_page_range_noflush(addr, addr + size, prot, pages); prot 2006 mm/vmalloc.c int map_vm_area(struct vm_struct *area, pgprot_t prot, struct page **pages) prot 2012 mm/vmalloc.c err = vmap_page_range(addr, end, prot, pages); prot 2371 mm/vmalloc.c unsigned long flags, pgprot_t prot) prot 2386 mm/vmalloc.c if (map_vm_area(area, prot, pages)) { prot 2396 mm/vmalloc.c gfp_t gfp_mask, pgprot_t prot, prot 2399 mm/vmalloc.c pgprot_t prot, int node) prot 2449 mm/vmalloc.c if (map_vm_area(area, prot, pages)) prot 2481 mm/vmalloc.c pgprot_t prot, unsigned long vm_flags, int node, prot 2497 mm/vmalloc.c addr = __vmalloc_area_node(area, gfp_mask, prot, node); prot 2549 mm/vmalloc.c gfp_t gfp_mask, pgprot_t prot, prot 2553 mm/vmalloc.c gfp_mask, prot, 0, node, caller); prot 2556 mm/vmalloc.c void *__vmalloc(unsigned long size, gfp_t gfp_mask, pgprot_t prot) prot 2558 mm/vmalloc.c return __vmalloc_node(size, 1, gfp_mask, prot, NUMA_NO_NODE, prot 229 net/atm/br2684.c unsigned short prot = ntohs(skb->protocol); prot 232 net/atm/br2684.c switch (prot) { prot 1032 net/caif/caif_socket.c static struct proto prot = {.name = "PF_CAIF", prot 1061 net/caif/caif_socket.c sk = sk_alloc(net, PF_CAIF, GFP_KERNEL, &prot, kern); prot 102 net/can/af_can.c if (cp && !try_module_get(cp->prot->owner)) prot 111 net/can/af_can.c module_put(cp->prot->owner); prot 158 net/can/af_can.c sk = sk_alloc(net, PF_CAN, GFP_KERNEL, cp->prot, kern); prot 729 net/can/af_can.c err = proto_register(cp->prot, 0); prot 745 net/can/af_can.c proto_unregister(cp->prot); prot 766 net/can/af_can.c proto_unregister(cp->prot); prot 1670 net/can/bcm.c .prot = &bcm_proto, prot 1222 net/can/j1939/socket.c .prot = &j1939_proto, prot 877 net/can/raw.c .prot = &raw_proto, prot 1451 net/core/neighbour.c __be16 prot = n->tbl->protocol; prot 1460 net/core/neighbour.c dev->header_ops->cache(n, hh, prot); prot 1591 net/core/sock.c static struct sock *sk_prot_alloc(struct proto *prot, gfp_t priority, prot 1597 net/core/sock.c slab = prot->slab; prot 1603 net/core/sock.c sk_prot_clear_nulls(sk, prot->obj_size); prot 1605 net/core/sock.c sk = kmalloc(prot->obj_size, priority); prot 1611 net/core/sock.c if (!try_module_get(prot->owner)) prot 1628 net/core/sock.c static void sk_prot_free(struct proto *prot, struct sock *sk) prot 1633 net/core/sock.c owner = prot->owner; prot 1634 net/core/sock.c slab = prot->slab; prot 1655 net/core/sock.c struct proto *prot, int kern) prot 1659 net/core/sock.c sk = sk_prot_alloc(prot, priority | __GFP_ZERO, family); prot 1666 net/core/sock.c sk->sk_prot = sk->sk_prot_creator = prot; prot 2502 net/core/sock.c struct proto *prot = sk->sk_prot; prot 2526 net/core/sock.c if (atomic_read(&sk->sk_rmem_alloc) < sk_get_rmem0(sk, prot)) prot 2530 net/core/sock.c int wmem0 = sk_get_wmem0(sk, prot); prot 2566 net/core/sock.c trace_sock_exceed_buf_limit(sk, prot, allocated, kind); prot 3232 net/core/sock.c void sock_prot_inuse_add(struct net *net, struct proto *prot, int val) prot 3234 net/core/sock.c __this_cpu_add(net->core.prot_inuse->val[prot->inuse_idx], val); prot 3238 net/core/sock.c int sock_prot_inuse_get(struct net *net, struct proto *prot) prot 3240 net/core/sock.c int cpu, idx = prot->inuse_idx; prot 3305 net/core/sock.c static int assign_proto_idx(struct proto *prot) prot 3307 net/core/sock.c prot->inuse_idx = find_first_zero_bit(proto_inuse_idx, PROTO_INUSE_NR); prot 3309 net/core/sock.c if (unlikely(prot->inuse_idx == PROTO_INUSE_NR - 1)) { prot 3314 net/core/sock.c set_bit(prot->inuse_idx, proto_inuse_idx); prot 3318 net/core/sock.c static void release_proto_idx(struct proto *prot) prot 3320 net/core/sock.c if (prot->inuse_idx != PROTO_INUSE_NR - 1) prot 3321 net/core/sock.c clear_bit(prot->inuse_idx, proto_inuse_idx); prot 3324 net/core/sock.c static inline int assign_proto_idx(struct proto *prot) prot 3329 net/core/sock.c static inline void release_proto_idx(struct proto *prot) prot 3348 net/core/sock.c static int req_prot_init(const struct proto *prot) prot 3350 net/core/sock.c struct request_sock_ops *rsk_prot = prot->rsk_prot; prot 3356 net/core/sock.c prot->name); prot 3362 net/core/sock.c SLAB_ACCOUNT | prot->slab_flags, prot 3367 net/core/sock.c prot->name); prot 3373 net/core/sock.c int proto_register(struct proto *prot, int alloc_slab) prot 3378 net/core/sock.c prot->slab = kmem_cache_create_usercopy(prot->name, prot 3379 net/core/sock.c prot->obj_size, 0, prot 3381 net/core/sock.c prot->slab_flags, prot 3382 net/core/sock.c prot->useroffset, prot->usersize, prot 3385 net/core/sock.c if (prot->slab == NULL) { prot 3387 net/core/sock.c prot->name); prot 3391 net/core/sock.c if (req_prot_init(prot)) prot 3394 net/core/sock.c if (prot->twsk_prot != NULL) { prot 3395 net/core/sock.c prot->twsk_prot->twsk_slab_name = kasprintf(GFP_KERNEL, "tw_sock_%s", prot->name); prot 3397 net/core/sock.c if (prot->twsk_prot->twsk_slab_name == NULL) prot 3400 net/core/sock.c prot->twsk_prot->twsk_slab = prot 3401 net/core/sock.c kmem_cache_create(prot->twsk_prot->twsk_slab_name, prot 3402 net/core/sock.c prot->twsk_prot->twsk_obj_size, prot 3405 net/core/sock.c prot->slab_flags, prot 3407 net/core/sock.c if (prot->twsk_prot->twsk_slab == NULL) prot 3413 net/core/sock.c ret = assign_proto_idx(prot); prot 3418 net/core/sock.c list_add(&prot->node, &proto_list); prot 3423 net/core/sock.c if (alloc_slab && prot->twsk_prot) prot 3424 net/core/sock.c kfree(prot->twsk_prot->twsk_slab_name); prot 3427 net/core/sock.c req_prot_cleanup(prot->rsk_prot); prot 3429 net/core/sock.c kmem_cache_destroy(prot->slab); prot 3430 net/core/sock.c prot->slab = NULL; prot 3437 net/core/sock.c void proto_unregister(struct proto *prot) prot 3440 net/core/sock.c release_proto_idx(prot); prot 3441 net/core/sock.c list_del(&prot->node); prot 3444 net/core/sock.c kmem_cache_destroy(prot->slab); prot 3445 net/core/sock.c prot->slab = NULL; prot 3447 net/core/sock.c req_prot_cleanup(prot->rsk_prot); prot 3449 net/core/sock.c if (prot->twsk_prot != NULL && prot->twsk_prot->twsk_slab != NULL) { prot 3450 net/core/sock.c kmem_cache_destroy(prot->twsk_prot->twsk_slab); prot 3451 net/core/sock.c kfree(prot->twsk_prot->twsk_slab_name); prot 3452 net/core/sock.c prot->twsk_prot->twsk_slab = NULL; prot 1009 net/dccp/ipv4.c .prot = &dccp_v4_prot, prot 1093 net/dccp/ipv6.c .prot = &dccp_v6_prot, prot 314 net/ipv4/af_inet.c answer_prot = answer->prot; prot 1093 net/ipv4/af_inet.c .prot = &tcp_prot, prot 1102 net/ipv4/af_inet.c .prot = &udp_prot, prot 1110 net/ipv4/af_inet.c .prot = &ping_prot, prot 1118 net/ipv4/af_inet.c .prot = &raw_prot, prot 143 net/ipv4/fou.c int prot; prot 147 net/ipv4/fou.c prot = IPPROTO_IPIP; prot 150 net/ipv4/fou.c prot = IPPROTO_IPV6; prot 159 net/ipv4/fou.c return -prot; prot 32 net/ipv4/protocol.c int inet_add_protocol(const struct net_protocol *prot, unsigned char protocol) prot 34 net/ipv4/protocol.c if (!prot->netns_ok) { prot 41 net/ipv4/protocol.c NULL, prot) ? 0 : -1; prot 45 net/ipv4/protocol.c int inet_add_offload(const struct net_offload *prot, unsigned char protocol) prot 48 net/ipv4/protocol.c NULL, prot) ? 0 : -1; prot 52 net/ipv4/protocol.c int inet_del_protocol(const struct net_protocol *prot, unsigned char protocol) prot 57 net/ipv4/protocol.c prot, NULL) == prot) ? 0 : -1; prot 65 net/ipv4/protocol.c int inet_del_offload(const struct net_offload *prot, unsigned char protocol) prot 70 net/ipv4/protocol.c prot, NULL) == prot) ? 0 : -1; prot 528 net/ipv4/route.c u8 prot, u32 mark, int flow_flags) prot 536 net/ipv4/route.c prot = inet->hdrincl ? IPPROTO_RAW : sk->sk_protocol; prot 539 net/ipv4/route.c RT_SCOPE_UNIVERSE, prot, prot 552 net/ipv4/route.c u8 prot = iph->protocol; prot 555 net/ipv4/route.c __build_flow_key(net, fl4, sk, iph, oif, tos, prot, mark, 0); prot 826 net/ipv4/route.c u8 prot = iph->protocol; prot 831 net/ipv4/route.c __build_flow_key(net, &fl4, sk, iph, oif, tos, prot, mark, 0); prot 600 net/ipv4/tcp_bpf.c static void tcp_bpf_rebuild_protos(struct proto prot[TCP_BPF_NUM_CFGS], prot 603 net/ipv4/tcp_bpf.c prot[TCP_BPF_BASE] = *base; prot 604 net/ipv4/tcp_bpf.c prot[TCP_BPF_BASE].unhash = tcp_bpf_unhash; prot 605 net/ipv4/tcp_bpf.c prot[TCP_BPF_BASE].close = tcp_bpf_close; prot 606 net/ipv4/tcp_bpf.c prot[TCP_BPF_BASE].recvmsg = tcp_bpf_recvmsg; prot 607 net/ipv4/tcp_bpf.c prot[TCP_BPF_BASE].stream_memory_read = tcp_bpf_stream_read; prot 609 net/ipv4/tcp_bpf.c prot[TCP_BPF_TX] = prot[TCP_BPF_BASE]; prot 610 net/ipv4/tcp_bpf.c prot[TCP_BPF_TX].sendmsg = tcp_bpf_sendmsg; prot 611 net/ipv4/tcp_bpf.c prot[TCP_BPF_TX].sendpage = tcp_bpf_sendpage; prot 69 net/ipv4/udplite.c .prot = &udplite_prot, prot 173 net/ipv6/af_inet6.c answer_prot = answer->prot; prot 203 net/ipv6/af_inet6.c sk->sk_backlog_rcv = answer->prot->backlog_rcv; prot 228 net/ipv6/ipv6_sockglue.c struct proto *prot = &udp_prot; prot 231 net/ipv6/ipv6_sockglue.c prot = &udplite_prot; prot 234 net/ipv6/ipv6_sockglue.c sock_prot_inuse_add(net, prot, 1); prot 236 net/ipv6/ipv6_sockglue.c sk->sk_prot = prot; prot 186 net/ipv6/ping.c .prot = &pingv6_prot, prot 28 net/ipv6/protocol.c int inet6_add_protocol(const struct inet6_protocol *prot, unsigned char protocol) prot 31 net/ipv6/protocol.c NULL, prot) ? 0 : -1; prot 35 net/ipv6/protocol.c int inet6_del_protocol(const struct inet6_protocol *prot, unsigned char protocol) prot 40 net/ipv6/protocol.c prot, NULL) == prot) ? 0 : -1; prot 52 net/ipv6/protocol.c int inet6_add_offload(const struct net_offload *prot, unsigned char protocol) prot 55 net/ipv6/protocol.c NULL, prot) ? 0 : -1; prot 59 net/ipv6/protocol.c int inet6_del_offload(const struct net_offload *prot, unsigned char protocol) prot 64 net/ipv6/protocol.c prot, NULL) == prot) ? 0 : -1; prot 1388 net/ipv6/raw.c .prot = &rawv6_prot, prot 2067 net/ipv6/tcp_ipv6.c .prot = &tcpv6_prot, prot 1689 net/ipv6/udp.c .prot = &udpv6_prot, prot 64 net/ipv6/udplite.c .prot = &udplitev6_prot, prot 650 net/l2tp/l2tp_ip.c .prot = &l2tp_ip_prot, prot 783 net/l2tp/l2tp_ip6.c .prot = &l2tp_ip6_prot, prot 915 net/llc/llc_conn.c struct sock *llc_sk_alloc(struct net *net, int family, gfp_t priority, struct proto *prot, int kern) prot 917 net/llc/llc_conn.c struct sock *sk = sk_alloc(net, family, priority, prot, kern); prot 608 net/mac80211/tdls.c u16 prot = IEEE80211_HT_OP_MODE_PROTECTION_NONHT_MIXED | prot 614 net/mac80211/tdls.c &sdata->vif.bss_conf.chandef, prot, prot 36 net/phonet/af_phonet.c if (pp && !try_module_get(pp->prot->owner)) prot 45 net/phonet/af_phonet.c module_put(pp->prot->owner); prot 87 net/phonet/af_phonet.c sk = sk_alloc(net, PF_PHONET, GFP_KERNEL, pnp->prot, kern); prot 467 net/phonet/af_phonet.c err = proto_register(pp->prot, 1); prot 490 net/phonet/af_phonet.c proto_unregister(pp->prot); prot 187 net/phonet/datagram.c .prot = &pn_proto, prot 1343 net/phonet/pep.c .prot = &pep_proto, prot 1043 net/sctp/ipv6.c .prot = &sctpv6_prot, prot 1050 net/sctp/ipv6.c .prot = &sctpv6_prot, prot 1048 net/sctp/protocol.c .prot = &sctp_prot, prot 1055 net/sctp/protocol.c .prot = &sctp_prot, prot 218 net/smc/af_smc.c struct proto *prot; prot 221 net/smc/af_smc.c prot = (protocol == SMCPROTO_SMC6) ? &smc_proto6 : &smc_proto; prot 222 net/smc/af_smc.c sk = sk_alloc(net, PF_SMC, GFP_KERNEL, prot, 0); prot 193 net/smc/smc_diag.c static int smc_diag_dump_proto(struct proto *prot, struct sk_buff *skb, prot 202 net/smc/smc_diag.c read_lock(&prot->h.smc_hash->lock); prot 203 net/smc/smc_diag.c head = &prot->h.smc_hash->ht; prot 216 net/smc/smc_diag.c read_unlock(&prot->h.smc_hash->lock); prot 450 net/sunrpc/rpcb_clnt.c int rpcb_register(struct net *net, u32 prog, u32 vers, int prot, unsigned short port) prot 455 net/sunrpc/rpcb_clnt.c .r_prot = prot, prot 466 net/sunrpc/rpcb_clnt.c prog, vers, prot, port); prot 698 net/sunrpc/rpcb_clnt.c xprt->servername, clnt->cl_prog, clnt->cl_vers, xprt->prot); prot 753 net/sunrpc/rpcb_clnt.c xprt->prot, bind_version, prot 771 net/sunrpc/rpcb_clnt.c map->r_prot = xprt->prot; prot 1563 net/sunrpc/svc.c rqstp->rq_prot = req->rq_xprt->prot; prot 1192 net/sunrpc/svc_xprt.c dr->prot = rqstp->rq_prot; prot 1227 net/sunrpc/svc_xprt.c rqstp->rq_prot = dr->prot; prot 309 net/sunrpc/xprtrdma/svc_rdma_backchannel.c xprt->prot = XPRT_TRANSPORT_BC_RDMA; prot 343 net/sunrpc/xprtrdma/transport.c xprt->prot = IPPROTO_TCP; prot 2917 net/sunrpc/xprtsock.c xprt->prot = 0; prot 2986 net/sunrpc/xprtsock.c xprt->prot = IPPROTO_UDP; prot 3066 net/sunrpc/xprtsock.c xprt->prot = IPPROTO_TCP; prot 3139 net/sunrpc/xprtsock.c xprt->prot = IPPROTO_TCP; prot 261 net/tls/tls_device.c struct tls_prot_info *prot = &ctx->prot_info; prot 273 net/tls/tls_device.c tls_advance_record_sn(sk, prot, &ctx->tx); prot 295 net/tls/tls_device.c struct tls_prot_info *prot = &ctx->prot_info; prot 304 net/tls/tls_device.c if (likely(skb_page_frag_refill(prot->tag_size, pfrag, prot 307 net/tls/tls_device.c tls_append_frag(record, pfrag, prot->tag_size); prot 309 net/tls/tls_device.c ret = prot->tag_size; prot 310 net/tls/tls_device.c if (record->len <= prot->overhead_size) prot 316 net/tls/tls_device.c record->len - prot->overhead_size, prot 317 net/tls/tls_device.c record_type, prot->version); prot 406 net/tls/tls_device.c struct tls_prot_info *prot = &tls_ctx->prot_info; prot 441 net/tls/tls_device.c prot->prepend_size; prot 444 net/tls/tls_device.c prot->prepend_size); prot 462 net/tls/tls_device.c } else if (record->len > prot->prepend_size) { prot 682 net/tls/tls_device.c struct tls_prot_info *prot; prot 690 net/tls/tls_device.c prot = &tls_ctx->prot_info; prot 692 net/tls/tls_device.c memcpy(rcd_sn, tls_ctx->rx.rec_seq, prot->rec_seq_size); prot 717 net/tls/tls_device.c tls_bigint_increment(rcd_sn, prot->rec_seq_size); prot 759 net/tls/tls_device.c struct tls_prot_info *prot = &tls_ctx->prot_info; prot 762 net/tls/tls_device.c memcpy(rcd_sn, tls_ctx->rx.rec_seq, prot->rec_seq_size); prot 763 net/tls/tls_device.c tls_bigint_increment(rcd_sn, prot->rec_seq_size); prot 908 net/tls/tls_device.c struct tls_prot_info *prot = &tls_ctx->prot_info; prot 961 net/tls/tls_device.c prot->version = crypto_info->version; prot 962 net/tls/tls_device.c prot->cipher_type = crypto_info->cipher_type; prot 963 net/tls/tls_device.c prot->prepend_size = TLS_HEADER_SIZE + nonce_size; prot 964 net/tls/tls_device.c prot->tag_size = tag_size; prot 965 net/tls/tls_device.c prot->overhead_size = prot->prepend_size + prot->tag_size; prot 966 net/tls/tls_device.c prot->iv_size = iv_size; prot 976 net/tls/tls_device.c prot->rec_seq_size = rec_seq_size; prot 65 net/tls/tls_main.c static void build_protos(struct proto prot[TLS_NUM_CONFIG][TLS_NUM_CONFIG], prot 723 net/tls/tls_main.c static void build_protos(struct proto prot[TLS_NUM_CONFIG][TLS_NUM_CONFIG], prot 726 net/tls/tls_main.c prot[TLS_BASE][TLS_BASE] = *base; prot 727 net/tls/tls_main.c prot[TLS_BASE][TLS_BASE].setsockopt = tls_setsockopt; prot 728 net/tls/tls_main.c prot[TLS_BASE][TLS_BASE].getsockopt = tls_getsockopt; prot 729 net/tls/tls_main.c prot[TLS_BASE][TLS_BASE].close = tls_sk_proto_close; prot 731 net/tls/tls_main.c prot[TLS_SW][TLS_BASE] = prot[TLS_BASE][TLS_BASE]; prot 732 net/tls/tls_main.c prot[TLS_SW][TLS_BASE].sendmsg = tls_sw_sendmsg; prot 733 net/tls/tls_main.c prot[TLS_SW][TLS_BASE].sendpage = tls_sw_sendpage; prot 735 net/tls/tls_main.c prot[TLS_BASE][TLS_SW] = prot[TLS_BASE][TLS_BASE]; prot 736 net/tls/tls_main.c prot[TLS_BASE][TLS_SW].recvmsg = tls_sw_recvmsg; prot 737 net/tls/tls_main.c prot[TLS_BASE][TLS_SW].stream_memory_read = tls_sw_stream_read; prot 738 net/tls/tls_main.c prot[TLS_BASE][TLS_SW].close = tls_sk_proto_close; prot 740 net/tls/tls_main.c prot[TLS_SW][TLS_SW] = prot[TLS_SW][TLS_BASE]; prot 741 net/tls/tls_main.c prot[TLS_SW][TLS_SW].recvmsg = tls_sw_recvmsg; prot 742 net/tls/tls_main.c prot[TLS_SW][TLS_SW].stream_memory_read = tls_sw_stream_read; prot 743 net/tls/tls_main.c prot[TLS_SW][TLS_SW].close = tls_sk_proto_close; prot 746 net/tls/tls_main.c prot[TLS_HW][TLS_BASE] = prot[TLS_BASE][TLS_BASE]; prot 747 net/tls/tls_main.c prot[TLS_HW][TLS_BASE].sendmsg = tls_device_sendmsg; prot 748 net/tls/tls_main.c prot[TLS_HW][TLS_BASE].sendpage = tls_device_sendpage; prot 750 net/tls/tls_main.c prot[TLS_HW][TLS_SW] = prot[TLS_BASE][TLS_SW]; prot 751 net/tls/tls_main.c prot[TLS_HW][TLS_SW].sendmsg = tls_device_sendmsg; prot 752 net/tls/tls_main.c prot[TLS_HW][TLS_SW].sendpage = tls_device_sendpage; prot 754 net/tls/tls_main.c prot[TLS_BASE][TLS_HW] = prot[TLS_BASE][TLS_SW]; prot 756 net/tls/tls_main.c prot[TLS_SW][TLS_HW] = prot[TLS_SW][TLS_SW]; prot 758 net/tls/tls_main.c prot[TLS_HW][TLS_HW] = prot[TLS_HW][TLS_SW]; prot 761 net/tls/tls_main.c prot[TLS_HW_RECORD][TLS_HW_RECORD] = *base; prot 762 net/tls/tls_main.c prot[TLS_HW_RECORD][TLS_HW_RECORD].hash = tls_hw_hash; prot 763 net/tls/tls_main.c prot[TLS_HW_RECORD][TLS_HW_RECORD].unhash = tls_hw_unhash; prot 122 net/tls/tls_sw.c struct tls_prot_info *prot, struct sk_buff *skb) prot 128 net/tls/tls_sw.c if (prot->version == TLS_1_3_VERSION) { prot 134 net/tls/tls_sw.c if (back > rxm->full_len - prot->prepend_size) prot 158 net/tls/tls_sw.c struct tls_prot_info *prot; prot 167 net/tls/tls_sw.c prot = &tls_ctx->prot_info; prot 177 net/tls/tls_sw.c pad = padding_length(ctx, prot, skb); prot 183 net/tls/tls_sw.c rxm->offset += prot->prepend_size; prot 184 net/tls/tls_sw.c rxm->full_len -= prot->overhead_size; prot 224 net/tls/tls_sw.c struct tls_prot_info *prot = &tls_ctx->prot_info; prot 229 net/tls/tls_sw.c aead_request_set_ad(aead_req, prot->aad_size); prot 231 net/tls/tls_sw.c data_len + prot->tag_size, prot 269 net/tls/tls_sw.c struct tls_prot_info *prot = &tls_ctx->prot_info; prot 275 net/tls/tls_sw.c target_size += prot->overhead_size; prot 292 net/tls/tls_sw.c struct tls_prot_info *prot = &tls_ctx->prot_info; prot 308 net/tls/tls_sw.c skip = prot->prepend_size + msg_pl->sg.size; prot 316 net/tls/tls_sw.c struct tls_prot_info *prot = &tls_ctx->prot_info; prot 335 net/tls/tls_sw.c sg_set_buf(&rec->sg_aead_in[0], rec->aad_space, prot->aad_size); prot 339 net/tls/tls_sw.c sg_set_buf(&rec->sg_aead_out[0], rec->aad_space, prot->aad_size); prot 428 net/tls/tls_sw.c struct tls_prot_info *prot = &tls_ctx->prot_info; prot 440 net/tls/tls_sw.c sge->offset -= prot->prepend_size; prot 441 net/tls/tls_sw.c sge->length += prot->prepend_size; prot 490 net/tls/tls_sw.c struct tls_prot_info *prot = &tls_ctx->prot_info; prot 497 net/tls/tls_sw.c if (prot->cipher_type == TLS_CIPHER_AES_CCM_128) { prot 503 net/tls/tls_sw.c prot->iv_size + prot->salt_size); prot 505 net/tls/tls_sw.c xor_iv_with_seq(prot->version, rec->iv_data, tls_ctx->tx.rec_seq); prot 507 net/tls/tls_sw.c sge->offset += prot->prepend_size; prot 508 net/tls/tls_sw.c sge->length -= prot->prepend_size; prot 513 net/tls/tls_sw.c aead_request_set_ad(aead_req, prot->aad_size); prot 528 net/tls/tls_sw.c sge->offset -= prot->prepend_size; prot 529 net/tls/tls_sw.c sge->length += prot->prepend_size; prot 541 net/tls/tls_sw.c tls_advance_record_sn(sk, prot, &tls_ctx->tx); prot 667 net/tls/tls_sw.c struct tls_prot_info *prot = &tls_ctx->prot_info; prot 686 net/tls/tls_sw.c prot->overhead_size > msg_en->sg.size) || prot 689 net/tls/tls_sw.c prot->overhead_size > msg_en->sg.size))) { prot 695 net/tls/tls_sw.c split_point, prot->overhead_size, prot 711 net/tls/tls_sw.c prot->overhead_size); prot 721 net/tls/tls_sw.c if (prot->version == TLS_1_3_VERSION) { prot 747 net/tls/tls_sw.c tls_make_aad(rec->aad_space, msg_pl->sg.size + prot->tail_size, prot 748 net/tls/tls_sw.c tls_ctx->tx.rec_seq, prot->rec_seq_size, prot 749 net/tls/tls_sw.c record_type, prot->version); prot 754 net/tls/tls_sw.c msg_pl->sg.size + prot->tail_size, prot 755 net/tls/tls_sw.c record_type, prot->version); prot 760 net/tls/tls_sw.c msg_pl->sg.size + prot->tail_size, i); prot 774 net/tls/tls_sw.c sk_msg_trim(sk, msg_en, msg_pl->sg.size + prot->overhead_size); prot 916 net/tls/tls_sw.c struct tls_prot_info *prot = &tls_ctx->prot_info; prot 979 net/tls/tls_sw.c prot->overhead_size; prot 1047 net/tls/tls_sw.c msg_pl->sg.size + prot->overhead_size); prot 1141 net/tls/tls_sw.c struct tls_prot_info *prot = &tls_ctx->prot_info; prot 1183 net/tls/tls_sw.c required_size = msg_pl->sg.size + copy + prot->overhead_size; prot 1397 net/tls/tls_sw.c struct tls_prot_info *prot = &tls_ctx->prot_info; prot 1405 net/tls/tls_sw.c const int data_len = rxm->full_len - prot->overhead_size + prot 1406 net/tls/tls_sw.c prot->tail_size; prot 1414 net/tls/tls_sw.c n_sgin = skb_nsg(skb, rxm->offset + prot->prepend_size, prot 1415 net/tls/tls_sw.c rxm->full_len - prot->prepend_size); prot 1432 net/tls/tls_sw.c mem_size = mem_size + prot->aad_size; prot 1448 net/tls/tls_sw.c iv = aad + prot->aad_size; prot 1451 net/tls/tls_sw.c if (prot->cipher_type == TLS_CIPHER_AES_CCM_128) { prot 1458 net/tls/tls_sw.c iv + iv_offset + prot->salt_size, prot 1459 net/tls/tls_sw.c prot->iv_size); prot 1464 net/tls/tls_sw.c if (prot->version == TLS_1_3_VERSION) prot 1468 net/tls/tls_sw.c memcpy(iv + iv_offset, tls_ctx->rx.iv, prot->salt_size); prot 1470 net/tls/tls_sw.c xor_iv_with_seq(prot->version, iv, tls_ctx->rx.rec_seq); prot 1473 net/tls/tls_sw.c tls_make_aad(aad, rxm->full_len - prot->overhead_size + prot 1474 net/tls/tls_sw.c prot->tail_size, prot 1475 net/tls/tls_sw.c tls_ctx->rx.rec_seq, prot->rec_seq_size, prot 1476 net/tls/tls_sw.c ctx->control, prot->version); prot 1480 net/tls/tls_sw.c sg_set_buf(&sgin[0], aad, prot->aad_size); prot 1482 net/tls/tls_sw.c rxm->offset + prot->prepend_size, prot 1483 net/tls/tls_sw.c rxm->full_len - prot->prepend_size); prot 1492 net/tls/tls_sw.c sg_set_buf(&sgout[0], aad, prot->aad_size); prot 1533 net/tls/tls_sw.c struct tls_prot_info *prot = &tls_ctx->prot_info; prot 1550 net/tls/tls_sw.c tls_advance_record_sn(sk, prot, prot 1559 net/tls/tls_sw.c pad = padding_length(ctx, prot, skb); prot 1564 net/tls/tls_sw.c rxm->offset += prot->prepend_size; prot 1565 net/tls/tls_sw.c rxm->full_len -= prot->overhead_size; prot 1566 net/tls/tls_sw.c tls_advance_record_sn(sk, prot, &tls_ctx->rx); prot 1727 net/tls/tls_sw.c struct tls_prot_info *prot = &tls_ctx->prot_info; prot 1791 net/tls/tls_sw.c if (prot->version == TLS_1_3_VERSION) prot 1799 net/tls/tls_sw.c to_decrypt = rxm->full_len - prot->overhead_size; prot 1803 net/tls/tls_sw.c prot->version != TLS_1_3_VERSION) prot 1822 net/tls/tls_sw.c } else if (prot->version == TLS_1_3_VERSION) { prot 2023 net/tls/tls_sw.c struct tls_prot_info *prot = &tls_ctx->prot_info; prot 2031 net/tls/tls_sw.c if (rxm->offset + prot->prepend_size > skb->len) prot 2035 net/tls/tls_sw.c if (WARN_ON(prot->prepend_size > sizeof(header))) { prot 2041 net/tls/tls_sw.c ret = skb_copy_bits(skb, rxm->offset, header, prot->prepend_size); prot 2050 net/tls/tls_sw.c cipher_overhead = prot->tag_size; prot 2051 net/tls/tls_sw.c if (prot->version != TLS_1_3_VERSION) prot 2052 net/tls/tls_sw.c cipher_overhead += prot->iv_size; prot 2055 net/tls/tls_sw.c prot->tail_size) { prot 2261 net/tls/tls_sw.c struct tls_prot_info *prot = &tls_ctx->prot_info; prot 2392 net/tls/tls_sw.c prot->aad_size = TLS_HEADER_SIZE; prot 2393 net/tls/tls_sw.c prot->tail_size = 1; prot 2395 net/tls/tls_sw.c prot->aad_size = TLS_AAD_SPACE_SIZE; prot 2396 net/tls/tls_sw.c prot->tail_size = 0; prot 2399 net/tls/tls_sw.c prot->version = crypto_info->version; prot 2400 net/tls/tls_sw.c prot->cipher_type = crypto_info->cipher_type; prot 2401 net/tls/tls_sw.c prot->prepend_size = TLS_HEADER_SIZE + nonce_size; prot 2402 net/tls/tls_sw.c prot->tag_size = tag_size; prot 2403 net/tls/tls_sw.c prot->overhead_size = prot->prepend_size + prot 2404 net/tls/tls_sw.c prot->tag_size + prot->tail_size; prot 2405 net/tls/tls_sw.c prot->iv_size = iv_size; prot 2406 net/tls/tls_sw.c prot->salt_size = salt_size; prot 2413 net/tls/tls_sw.c prot->rec_seq_size = rec_seq_size; prot 2438 net/tls/tls_sw.c rc = crypto_aead_setauthsize(*aead, prot->tag_size); prot 481 security/apparmor/lsm.c static int common_mmap(const char *op, struct file *file, unsigned long prot, prot 489 security/apparmor/lsm.c if (prot & PROT_READ) prot 495 security/apparmor/lsm.c if ((prot & PROT_WRITE) && !(flags & MAP_PRIVATE)) prot 497 security/apparmor/lsm.c if (prot & PROT_EXEC) prot 504 security/apparmor/lsm.c unsigned long prot, unsigned long flags) prot 506 security/apparmor/lsm.c return common_mmap(OP_FMMAP, file, prot, flags); prot 510 security/apparmor/lsm.c unsigned long reqprot, unsigned long prot) prot 512 security/apparmor/lsm.c return common_mmap(OP_FMPROT, vma->vm_file, prot, prot 1336 security/commoncap.c unsigned long prot, unsigned long flags) prot 381 security/integrity/ima/ima_main.c int ima_file_mmap(struct file *file, unsigned long prot) prot 385 security/integrity/ima/ima_main.c if (file && (prot & PROT_EXEC)) { prot 1412 security/security.c static inline unsigned long mmap_prot(struct file *file, unsigned long prot) prot 1418 security/security.c if ((prot & (PROT_READ | PROT_EXEC)) != PROT_READ) prot 1419 security/security.c return prot; prot 1421 security/security.c return prot; prot 1426 security/security.c return prot | PROT_EXEC; prot 1436 security/security.c return prot; prot 1439 security/security.c return prot | PROT_EXEC; prot 1442 security/security.c return prot; prot 1445 security/security.c int security_mmap_file(struct file *file, unsigned long prot, prot 1449 security/security.c ret = call_int_hook(mmap_file, 0, file, prot, prot 1450 security/security.c mmap_prot(file, prot), flags); prot 1453 security/security.c return ima_file_mmap(file, prot); prot 1462 security/security.c unsigned long prot) prot 1464 security/security.c return call_int_hook(file_mprotect, 0, vma, reqprot, prot); prot 3652 security/selinux/hooks.c static int file_map_prot_check(struct file *file, unsigned long prot, int shared) prot 3659 security/selinux/hooks.c (prot & PROT_EXEC) && (!file || IS_PRIVATE(file_inode(file)) || prot 3660 security/selinux/hooks.c (!shared && (prot & PROT_WRITE)))) { prot 3678 security/selinux/hooks.c if (shared && (prot & PROT_WRITE)) prot 3681 security/selinux/hooks.c if (prot & PROT_EXEC) prot 3706 security/selinux/hooks.c unsigned long prot, unsigned long flags) prot 3721 security/selinux/hooks.c prot = reqprot; prot 3723 security/selinux/hooks.c return file_map_prot_check(file, prot, prot 3729 security/selinux/hooks.c unsigned long prot) prot 3735 security/selinux/hooks.c prot = reqprot; prot 3738 security/selinux/hooks.c (prot & PROT_EXEC) && !(vma->vm_flags & VM_EXEC)) { prot 3766 security/selinux/hooks.c return file_map_prot_check(vma->vm_file, prot, vma->vm_flags&VM_SHARED); prot 1657 security/smack/smack_lsm.c unsigned long reqprot, unsigned long prot, prot 65 sound/core/sgbuf.c pgprot_t prot = PAGE_KERNEL; prot 75 sound/core/sgbuf.c prot = pgprot_noncached(PAGE_KERNEL); prot 125 sound/core/sgbuf.c dmab->area = vmap(sgbuf->page_table, sgbuf->pages, VM_MAP, prot); prot 427 tools/arch/s390/include/uapi/asm/ptrace.h ptprot_flags prot; prot 736 tools/perf/builtin-report.c map->prot & PROT_READ ? 'r' : '-', prot 737 tools/perf/builtin-report.c map->prot & PROT_WRITE ? 'w' : '-', prot 738 tools/perf/builtin-report.c map->prot & PROT_EXEC ? 'x' : '-', prot 30 tools/perf/lib/include/perf/event.h __u32 prot; prot 9 tools/perf/trace/beauty/mmap.c int printed = 0, prot = arg->val; prot 12 tools/perf/trace/beauty/mmap.c if (prot == PROT_NONE) prot 15 tools/perf/trace/beauty/mmap.c if (prot & PROT_##n) { \ prot 17 tools/perf/trace/beauty/mmap.c prot &= ~PROT_##n; \ prot 28 tools/perf/trace/beauty/mmap.c if (prot) prot 29 tools/perf/trace/beauty/mmap.c printed += scnprintf(bf + printed, size - printed, "%s%#x", printed ? "|" : "", prot); prot 95 tools/perf/util/auxtrace.c mm->base = mmap(NULL, mp->len, mp->prot, MAP_SHARED, fd, mp->offset); prot 122 tools/perf/util/auxtrace.c mp->prot = PROT_READ | (auxtrace_overwrite ? 0 : PROT_WRITE); prot 297 tools/perf/util/auxtrace.h int prot; prot 272 tools/perf/util/event.c (event->mmap2.prot & PROT_READ) ? 'r' : '-', prot 273 tools/perf/util/event.c (event->mmap2.prot & PROT_WRITE) ? 'w' : '-', prot 274 tools/perf/util/event.c (event->mmap2.prot & PROT_EXEC) ? 'x' : '-', prot 647 tools/perf/util/evlist.c mp->prot = PROT_READ | PROT_WRITE; prot 660 tools/perf/util/evlist.c mp->prot &= ~PROT_WRITE; prot 465 tools/perf/util/jitdump.c event->mmap2.prot = st.st_mode; prot 557 tools/perf/util/jitdump.c event->mmap2.prot = st.st_mode; prot 1677 tools/perf/util/machine.c event->mmap2.prot, prot 1706 tools/perf/util/machine.c u32 prot = 0; prot 1726 tools/perf/util/machine.c prot = PROT_EXEC; prot 1730 tools/perf/util/machine.c 0, 0, 0, 0, prot, 0, prot 150 tools/perf/util/map.c u64 ino_gen, u32 prot, u32 flags, char *filename, prot 171 tools/perf/util/map.c map->prot = prot; prot 175 tools/perf/util/map.c if ((anon || no_dso) && nsi && (prot & PROT_EXEC)) { prot 215 tools/perf/util/map.c if (!(prot & PROT_EXEC)) prot 31 tools/perf/util/map.h u32 prot; prot 115 tools/perf/util/map.h u64 ino_gen, u32 prot, u32 flags, prot 374 tools/perf/util/mmap.c map->core.base = mmap(NULL, perf_mmap__mmap_len(map), mp->prot, prot 40 tools/perf/util/mmap.h int prot, mask, nr_cblocks, affinity, flush, comp_level; prot 1278 tools/perf/util/sort.c map && !(map->prot & PROT_EXEC) && prot 1302 tools/perf/util/symbol.c err = file__read_maps(fd, map->prot & PROT_EXEC, kcore_mapfn, &md, prot 1394 tools/perf/util/symbol.c if (map->prot & PROT_EXEC) prot 312 tools/perf/util/synthetic-events.c char prot[5]; prot 336 tools/perf/util/synthetic-events.c &event->mmap2.start, &event->mmap2.len, prot, prot 358 tools/perf/util/synthetic-events.c event->mmap2.prot = 0; prot 360 tools/perf/util/synthetic-events.c if (prot[0] == 'r') prot 361 tools/perf/util/synthetic-events.c event->mmap2.prot |= PROT_READ; prot 362 tools/perf/util/synthetic-events.c if (prot[1] == 'w') prot 363 tools/perf/util/synthetic-events.c event->mmap2.prot |= PROT_WRITE; prot 364 tools/perf/util/synthetic-events.c if (prot[2] == 'x') prot 365 tools/perf/util/synthetic-events.c event->mmap2.prot |= PROT_EXEC; prot 367 tools/perf/util/synthetic-events.c if (prot[3] == 's') prot 372 tools/perf/util/synthetic-events.c if (prot[2] != 'x') { prot 373 tools/perf/util/synthetic-events.c if (!mmap_data || prot[0] != 'r') prot 680 tools/testing/selftests/x86/protection_keys.c int prot; prot 685 tools/testing/selftests/x86/protection_keys.c void record_pkey_malloc(void *ptr, long size, int prot) prot 717 tools/testing/selftests/x86/protection_keys.c rec->prot = prot; prot 750 tools/testing/selftests/x86/protection_keys.c void *malloc_pkey_with_mprotect(long size, int prot, u16 pkey) prot 757 tools/testing/selftests/x86/protection_keys.c size, prot, pkey); prot 759 tools/testing/selftests/x86/protection_keys.c ptr = mmap(NULL, size, prot, MAP_ANONYMOUS|MAP_PRIVATE, -1, 0); prot 761 tools/testing/selftests/x86/protection_keys.c ret = mprotect_pkey((void *)ptr, PAGE_SIZE, prot, pkey); prot 763 tools/testing/selftests/x86/protection_keys.c record_pkey_malloc(ptr, size, prot); prot 770 tools/testing/selftests/x86/protection_keys.c void *malloc_pkey_anon_huge(long size, int prot, u16 pkey) prot 776 tools/testing/selftests/x86/protection_keys.c size, prot, pkey); prot 784 tools/testing/selftests/x86/protection_keys.c record_pkey_malloc(ptr, size, prot); prot 785 tools/testing/selftests/x86/protection_keys.c mprotect_pkey(ptr, size, prot, pkey); prot 842 tools/testing/selftests/x86/protection_keys.c void *malloc_pkey_hugetlb(long size, int prot, u16 pkey) prot 850 tools/testing/selftests/x86/protection_keys.c dprintf1("doing %s(%ld, %x, %x)\n", __func__, size, prot, pkey); prot 855 tools/testing/selftests/x86/protection_keys.c mprotect_pkey(ptr, size, prot, pkey); prot 857 tools/testing/selftests/x86/protection_keys.c record_pkey_malloc(ptr, size, prot); prot 863 tools/testing/selftests/x86/protection_keys.c void *malloc_pkey_mmap_dax(long size, int prot, u16 pkey) prot 869 tools/testing/selftests/x86/protection_keys.c size, prot, pkey); prot 874 tools/testing/selftests/x86/protection_keys.c ptr = mmap(0, size, prot, MAP_SHARED, fd, 0); prot 877 tools/testing/selftests/x86/protection_keys.c mprotect_pkey(ptr, size, prot, pkey); prot 879 tools/testing/selftests/x86/protection_keys.c record_pkey_malloc(ptr, size, prot); prot 886 tools/testing/selftests/x86/protection_keys.c void *(*pkey_malloc[])(long size, int prot, u16 pkey) = { prot 897 tools/testing/selftests/x86/protection_keys.c void *malloc_pkey(long size, int prot, u16 pkey) prot 908 tools/testing/selftests/x86/protection_keys.c ret = pkey_malloc[malloc_type](size, prot, pkey); prot 923 tools/testing/selftests/x86/protection_keys.c size, prot, pkey, ret); prot 1216 tools/testing/selftests/x86/protection_keys.c int prot; prot 1228 tools/testing/selftests/x86/protection_keys.c prot = pkey_last_malloc_record->prot; prot 1231 tools/testing/selftests/x86/protection_keys.c mprotect_pkey(ptr, size, prot, 0); prot 1234 tools/testing/selftests/x86/protection_keys.c mprotect_pkey(ptr, size, prot, pkey); prot 1441 tools/testing/selftests/x86/protection_keys.c int prot = PROT_READ|PROT_WRITE; prot 1453 tools/testing/selftests/x86/protection_keys.c ptr = malloc_pkey(PAGE_SIZE, prot, pkey); prot 612 virt/kvm/arm/mmu.c pgprot_t prot) prot 620 virt/kvm/arm/mmu.c kvm_set_pte(pte, kvm_pfn_pte(pfn, prot)); prot 628 virt/kvm/arm/mmu.c pgprot_t prot) prot 652 virt/kvm/arm/mmu.c create_hyp_pte_mappings(pmd, addr, next, pfn, prot); prot 661 virt/kvm/arm/mmu.c pgprot_t prot) prot 683 virt/kvm/arm/mmu.c ret = create_hyp_pmd_mappings(pud, addr, next, pfn, prot); prot 694 virt/kvm/arm/mmu.c unsigned long pfn, pgprot_t prot) prot 719 virt/kvm/arm/mmu.c err = create_hyp_pud_mappings(pgd, addr, next, pfn, prot); prot 750 virt/kvm/arm/mmu.c int create_hyp_mappings(void *from, void *to, pgprot_t prot) prot 770 virt/kvm/arm/mmu.c prot); prot 779 virt/kvm/arm/mmu.c unsigned long *haddr, pgprot_t prot) prot 818 virt/kvm/arm/mmu.c __phys_to_pfn(phys_addr), prot);