prot              142 arch/alpha/include/asm/pgtable.h #define pgprot_noncached(prot)	(prot)
prot              178 arch/alpha/kernel/osf_sys.c 		unsigned long, prot, unsigned long, flags, unsigned long, fd,
prot              192 arch/alpha/kernel/osf_sys.c 	ret = ksys_mmap_pgoff(addr, len, prot, flags, fd, off >> PAGE_SHIFT);
prot              196 arch/alpha/kernel/proto.h 	pgprot_t prot;
prot              198 arch/alpha/kernel/proto.h 	prot = __pgprot(_PAGE_VALID | _PAGE_ASM | _PAGE_KRE
prot              200 arch/alpha/kernel/proto.h 	return ioremap_page_range(address, address + size, phys_addr, prot);
prot               38 arch/arc/include/asm/hugepage.h #define mk_pmd(page, prot)	pte_pmd(mk_pte(page, prot))
prot               42 arch/arc/include/asm/hugepage.h #define pfn_pmd(pfn, prot)	(__pmd(((pfn) << PAGE_SHIFT) | pgprot_val(prot)))
prot              280 arch/arc/include/asm/pgtable.h #define mk_pte(page, prot)	pfn_pte(page_to_pfn(page), prot)
prot              281 arch/arc/include/asm/pgtable.h #define pfn_pte(pfn, prot)	__pte(__pfn_to_phys(pfn) | pgprot_val(prot))
prot              326 arch/arc/include/asm/pgtable.h #define pgprot_noncached(prot)	(__pgprot(pgprot_val(prot) & ~_PAGE_CACHEABLE))
prot               59 arch/arc/mm/ioremap.c 	pgprot_t prot = __pgprot(flags);
prot               71 arch/arc/mm/ioremap.c 	prot = pgprot_noncached(prot);
prot               86 arch/arc/mm/ioremap.c 	if (ioremap_page_range(vaddr, vaddr + size, paddr, prot)) {
prot               56 arch/arm/include/asm/fixmap.h void __set_fixmap(enum fixed_addresses idx, phys_addr_t phys, pgprot_t prot);
prot               46 arch/arm/include/asm/kvm_mmu.h int create_hyp_mappings(void *from, void *to, pgprot_t prot);
prot               73 arch/arm/include/asm/kvm_mmu.h #define kvm_pfn_pte(pfn, prot)	pfn_pte(pfn, prot)
prot               74 arch/arm/include/asm/kvm_mmu.h #define kvm_pfn_pmd(pfn, prot)	pfn_pmd(pfn, prot)
prot               75 arch/arm/include/asm/kvm_mmu.h #define kvm_pfn_pud(pfn, prot)	(__pud(0))
prot              112 arch/arm/include/asm/pgalloc.h 				  pmdval_t prot)
prot              114 arch/arm/include/asm/pgalloc.h 	pmdval_t pmdval = (pte + PTE_HWTABLE_OFF) | prot;
prot              141 arch/arm/include/asm/pgalloc.h 	pmdval_t prot;
prot              144 arch/arm/include/asm/pgalloc.h 		prot = user_pmd_table;
prot              146 arch/arm/include/asm/pgalloc.h 		prot = _PAGE_USER_TABLE;
prot              148 arch/arm/include/asm/pgalloc.h 	__pmd_populate(pmdp, page_to_phys(ptep), prot);
prot              236 arch/arm/include/asm/pgtable-3level.h #define pfn_pmd(pfn,prot)	(__pmd(((phys_addr_t)(pfn) << PAGE_SHIFT) | pgprot_val(prot)))
prot              237 arch/arm/include/asm/pgtable-3level.h #define mk_pmd(page,prot)	pfn_pmd(page_to_pfn(page),prot)
prot               63 arch/arm/include/asm/pgtable-nommu.h #define pgprot_noncached(prot)	(prot)
prot               64 arch/arm/include/asm/pgtable-nommu.h #define pgprot_writecombine(prot) (prot)
prot               65 arch/arm/include/asm/pgtable-nommu.h #define pgprot_device(prot)	(prot)
prot              113 arch/arm/include/asm/pgtable.h #define __pgprot_modify(prot,mask,bits)		\
prot              114 arch/arm/include/asm/pgtable.h 	__pgprot((pgprot_val(prot) & ~(mask)) | (bits))
prot              116 arch/arm/include/asm/pgtable.h #define pgprot_noncached(prot) \
prot              117 arch/arm/include/asm/pgtable.h 	__pgprot_modify(prot, L_PTE_MT_MASK, L_PTE_MT_UNCACHED)
prot              119 arch/arm/include/asm/pgtable.h #define pgprot_writecombine(prot) \
prot              120 arch/arm/include/asm/pgtable.h 	__pgprot_modify(prot, L_PTE_MT_MASK, L_PTE_MT_BUFFERABLE)
prot              122 arch/arm/include/asm/pgtable.h #define pgprot_stronglyordered(prot) \
prot              123 arch/arm/include/asm/pgtable.h 	__pgprot_modify(prot, L_PTE_MT_MASK, L_PTE_MT_UNCACHED)
prot              125 arch/arm/include/asm/pgtable.h #define pgprot_device(prot) \
prot              126 arch/arm/include/asm/pgtable.h 	__pgprot_modify(prot, L_PTE_MT_MASK, L_PTE_MT_DEV_SHARED | L_PTE_SHARED | L_PTE_DIRTY | L_PTE_XN)
prot              129 arch/arm/include/asm/pgtable.h #define pgprot_dmacoherent(prot) \
prot              130 arch/arm/include/asm/pgtable.h 	__pgprot_modify(prot, L_PTE_MT_MASK, L_PTE_MT_BUFFERABLE | L_PTE_XN)
prot              136 arch/arm/include/asm/pgtable.h #define pgprot_dmacoherent(prot) \
prot              137 arch/arm/include/asm/pgtable.h 	__pgprot_modify(prot, L_PTE_MT_MASK, L_PTE_MT_UNCACHED | L_PTE_XN)
prot              212 arch/arm/include/asm/pgtable.h #define pfn_pte(pfn,prot)	__pte(__pfn_to_phys(pfn) | pgprot_val(prot))
prot              215 arch/arm/include/asm/pgtable.h #define mk_pte(page,prot)	pfn_pte(page_to_pfn(page), prot)
prot              269 arch/arm/include/asm/pgtable.h static inline pte_t clear_pte_bit(pte_t pte, pgprot_t prot)
prot              271 arch/arm/include/asm/pgtable.h 	pte_val(pte) &= ~pgprot_val(prot);
prot              275 arch/arm/include/asm/pgtable.h static inline pte_t set_pte_bit(pte_t pte, pgprot_t prot)
prot              277 arch/arm/include/asm/pgtable.h 	pte_val(pte) |= pgprot_val(prot);
prot              633 arch/arm/mach-sa1100/assabet.c 	int prot = PMD_TYPE_SECT | PMD_SECT_AP_WRITE | PMD_DOMAIN(DOMAIN_IO);
prot              637 arch/arm/mach-sa1100/assabet.c 	*pmd = __pmd(phys | prot);
prot               48 arch/arm/mm/dma-mapping.c 	pgprot_t prot;
prot              335 arch/arm/mm/dma-mapping.c 				     pgprot_t prot, struct page **ret_page,
prot              340 arch/arm/mm/dma-mapping.c 				 pgprot_t prot, struct page **ret_page,
prot              360 arch/arm/mm/dma-mapping.c 	pgprot_t prot = pgprot_dmacoherent(PAGE_KERNEL);
prot              373 arch/arm/mm/dma-mapping.c 		ptr = __alloc_from_contiguous(NULL, atomic_pool_size, prot,
prot              377 arch/arm/mm/dma-mapping.c 		ptr = __alloc_remap_buffer(NULL, atomic_pool_size, gfp, prot,
prot              467 arch/arm/mm/dma-mapping.c 	pgprot_t prot = *(pgprot_t *)data;
prot              469 arch/arm/mm/dma-mapping.c 	set_pte_ext(pte, mk_pte(page, prot), 0);
prot              473 arch/arm/mm/dma-mapping.c static void __dma_remap(struct page *page, size_t size, pgprot_t prot)
prot              478 arch/arm/mm/dma-mapping.c 	apply_to_page_range(&init_mm, start, size, __dma_update_pte, &prot);
prot              483 arch/arm/mm/dma-mapping.c 				 pgprot_t prot, struct page **ret_page,
prot              498 arch/arm/mm/dma-mapping.c 	ptr = dma_common_contiguous_remap(page, size, prot, caller);
prot              546 arch/arm/mm/dma-mapping.c 				     pgprot_t prot, struct page **ret_page,
prot              565 arch/arm/mm/dma-mapping.c 		ptr = dma_common_contiguous_remap(page, size, prot, caller);
prot              571 arch/arm/mm/dma-mapping.c 		__dma_remap(page, size, prot);
prot              592 arch/arm/mm/dma-mapping.c static inline pgprot_t __get_dma_pgprot(unsigned long attrs, pgprot_t prot)
prot              594 arch/arm/mm/dma-mapping.c 	prot = (attrs & DMA_ATTR_WRITE_COMBINE) ?
prot              595 arch/arm/mm/dma-mapping.c 			pgprot_writecombine(prot) :
prot              596 arch/arm/mm/dma-mapping.c 			pgprot_dmacoherent(prot);
prot              597 arch/arm/mm/dma-mapping.c 	return prot;
prot              633 arch/arm/mm/dma-mapping.c 	return __alloc_from_contiguous(args->dev, args->size, args->prot,
prot              670 arch/arm/mm/dma-mapping.c 				    args->prot, ret_page, args->caller,
prot              688 arch/arm/mm/dma-mapping.c 			 gfp_t gfp, pgprot_t prot, bool is_coherent,
prot              700 arch/arm/mm/dma-mapping.c 		.prot = prot,
prot              774 arch/arm/mm/dma-mapping.c 	pgprot_t prot = __get_dma_pgprot(attrs, PAGE_KERNEL);
prot              776 arch/arm/mm/dma-mapping.c 	return __dma_alloc(dev, size, handle, gfp, prot, false,
prot             1119 arch/arm/mm/dma-mapping.c 	int prot = 0;
prot             1122 arch/arm/mm/dma-mapping.c 		prot |= IOMMU_PRIV;
prot             1126 arch/arm/mm/dma-mapping.c 		return prot | IOMMU_READ | IOMMU_WRITE;
prot             1128 arch/arm/mm/dma-mapping.c 		return prot | IOMMU_READ;
prot             1130 arch/arm/mm/dma-mapping.c 		return prot | IOMMU_WRITE;
prot             1132 arch/arm/mm/dma-mapping.c 		return prot;
prot             1465 arch/arm/mm/dma-mapping.c 	pgprot_t prot = __get_dma_pgprot(attrs, PAGE_KERNEL);
prot             1496 arch/arm/mm/dma-mapping.c 	addr = dma_common_pages_remap(pages, size, prot,
prot             1625 arch/arm/mm/dma-mapping.c 	int prot;
prot             1641 arch/arm/mm/dma-mapping.c 		prot = __dma_info_to_prot(dir, attrs);
prot             1643 arch/arm/mm/dma-mapping.c 		ret = iommu_map(mapping->domain, iova, phys, len, prot);
prot             1843 arch/arm/mm/dma-mapping.c 	int ret, prot, len = PAGE_ALIGN(size + offset);
prot             1849 arch/arm/mm/dma-mapping.c 	prot = __dma_info_to_prot(dir, attrs);
prot             1851 arch/arm/mm/dma-mapping.c 	ret = iommu_map(mapping->domain, dma_addr, page_to_phys(page), len, prot);
prot             1946 arch/arm/mm/dma-mapping.c 	int ret, prot;
prot             1955 arch/arm/mm/dma-mapping.c 	prot = __dma_info_to_prot(dir, attrs) | IOMMU_MMIO;
prot             1957 arch/arm/mm/dma-mapping.c 	ret = iommu_map(mapping->domain, dma_addr, addr, len, prot);
prot              258 arch/arm/mm/dump.c 	u64 prot = val & pg_level[level].mask;
prot              262 arch/arm/mm/dump.c 		st->current_prot = prot;
prot              265 arch/arm/mm/dump.c 	} else if (prot != st->current_prot || level != st->level ||
prot              296 arch/arm/mm/dump.c 		st->current_prot = prot;
prot              240 arch/arm/mm/fault-armv.c 		pgprot_t prot = __pgprot_modify(PAGE_KERNEL,
prot              243 arch/arm/mm/fault-armv.c 		p1 = vmap(&page, 1, VM_IOREMAP, prot);
prot              244 arch/arm/mm/fault-armv.c 		p2 = vmap(&page, 1, VM_IOREMAP, prot);
prot               25 arch/arm/mm/idmap.c 	unsigned long prot)
prot               50 arch/arm/mm/idmap.c 		*pmd = __pmd((addr & PMD_MASK) | prot);
prot               56 arch/arm/mm/idmap.c 	unsigned long prot)
prot               60 arch/arm/mm/idmap.c 	addr = (addr & PMD_MASK) | prot;
prot               69 arch/arm/mm/idmap.c 	unsigned long prot)
prot               76 arch/arm/mm/idmap.c 		idmap_add_pmd(pud, addr, next, prot);
prot               81 arch/arm/mm/idmap.c 				 const char *text_end, unsigned long prot)
prot               90 arch/arm/mm/idmap.c 	prot |= PMD_TYPE_SECT | PMD_SECT_AP_WRITE | PMD_SECT_AF;
prot               93 arch/arm/mm/idmap.c 		prot |= PMD_BIT4;
prot               98 arch/arm/mm/idmap.c 		idmap_add_pud(pgd, addr, next, prot);
prot              512 arch/arm/mm/init.c 	pmdval_t prot;
prot              526 arch/arm/mm/init.c 		.prot	= PMD_SECT_XN,
prot              534 arch/arm/mm/init.c 		.prot	= PMD_SECT_XN,
prot              542 arch/arm/mm/init.c 		.prot   = PMD_SECT_XN,
prot              554 arch/arm/mm/init.c 		.prot   = L_PMD_SECT_RDONLY | PMD_SECT_AP2,
prot              557 arch/arm/mm/init.c 		.prot   = PMD_SECT_APX | PMD_SECT_AP_WRITE,
prot              569 arch/arm/mm/init.c 				  pmdval_t prot, struct mm_struct *mm)
prot              576 arch/arm/mm/init.c 	pmd[0] = __pmd((pmd_val(pmd[0]) & mask) | prot);
prot              579 arch/arm/mm/init.c 		pmd[1] = __pmd((pmd_val(pmd[1]) & mask) | prot);
prot              581 arch/arm/mm/init.c 		pmd[0] = __pmd((pmd_val(pmd[0]) & mask) | prot);
prot              618 arch/arm/mm/init.c 				set ? perms[i].prot : perms[i].clear, mm);
prot              406 arch/arm/mm/mmu.c void __set_fixmap(enum fixed_addresses idx, phys_addr_t phys, pgprot_t prot)
prot              417 arch/arm/mm/mmu.c 	if (WARN_ON(pgprot_val(prot) != pgprot_val(FIXMAP_PAGE_IO) &&
prot              421 arch/arm/mm/mmu.c 	if (pgprot_val(prot))
prot              423 arch/arm/mm/mmu.c 			pfn_pte(phys >> PAGE_SHIFT, prot));
prot              740 arch/arm/mm/mmu.c 				unsigned long prot,
prot              745 arch/arm/mm/mmu.c 		__pmd_populate(pmd, __pa(pte), prot);
prot              752 arch/arm/mm/mmu.c 				      unsigned long prot)
prot              754 arch/arm/mm/mmu.c 	return arm_pte_alloc(pmd, addr, prot, early_alloc);
prot              107 arch/arm64/include/asm/fixmap.h extern void __set_fixmap(enum fixed_addresses idx, phys_addr_t phys, pgprot_t prot);
prot              165 arch/arm64/include/asm/io.h extern void __iomem *__ioremap(phys_addr_t phys_addr, size_t size, pgprot_t prot);
prot              147 arch/arm64/include/asm/kvm_mmu.h int create_hyp_mappings(void *from, void *to, pgprot_t prot);
prot              179 arch/arm64/include/asm/kvm_mmu.h #define kvm_pfn_pte(pfn, prot)		pfn_pte(pfn, prot)
prot              180 arch/arm64/include/asm/kvm_mmu.h #define kvm_pfn_pmd(pfn, prot)		pfn_pmd(pfn, prot)
prot              181 arch/arm64/include/asm/kvm_mmu.h #define kvm_pfn_pud(pfn, prot)		pfn_pud(pfn, prot)
prot              128 arch/arm64/include/asm/mmu.h 			       pgprot_t prot, bool page_mappings_only);
prot              129 arch/arm64/include/asm/mmu.h extern void *fixmap_remap_fdt(phys_addr_t dt_phys, int *size, pgprot_t prot);
prot               47 arch/arm64/include/asm/pgalloc.h static inline void __pud_populate(pud_t *pudp, phys_addr_t pmdp, pudval_t prot)
prot               49 arch/arm64/include/asm/pgalloc.h 	set_pud(pudp, __pud(__phys_to_pud_val(pmdp) | prot));
prot               57 arch/arm64/include/asm/pgalloc.h static inline void __pud_populate(pud_t *pudp, phys_addr_t pmdp, pudval_t prot)
prot               76 arch/arm64/include/asm/pgalloc.h static inline void __pgd_populate(pgd_t *pgdp, phys_addr_t pudp, pgdval_t prot)
prot               78 arch/arm64/include/asm/pgalloc.h 	set_pgd(pgdp, __pgd(__phys_to_pgd_val(pudp) | prot));
prot               86 arch/arm64/include/asm/pgalloc.h static inline void __pgd_populate(pgd_t *pgdp, phys_addr_t pudp, pgdval_t prot)
prot               96 arch/arm64/include/asm/pgalloc.h 				  pmdval_t prot)
prot               98 arch/arm64/include/asm/pgalloc.h 	set_pmd(pmdp, __pmd(__phys_to_pmd_val(ptep) | prot));
prot               66 arch/arm64/include/asm/pgtable.h #define pfn_pte(pfn,prot)	\
prot               67 arch/arm64/include/asm/pgtable.h 	__pte(__phys_to_pte_val((phys_addr_t)(pfn) << PAGE_SHIFT) | pgprot_val(prot))
prot              126 arch/arm64/include/asm/pgtable.h static inline pte_t clear_pte_bit(pte_t pte, pgprot_t prot)
prot              128 arch/arm64/include/asm/pgtable.h 	pte_val(pte) &= ~pgprot_val(prot);
prot              132 arch/arm64/include/asm/pgtable.h static inline pte_t set_pte_bit(pte_t pte, pgprot_t prot)
prot              134 arch/arm64/include/asm/pgtable.h 	pte_val(pte) |= pgprot_val(prot);
prot              326 arch/arm64/include/asm/pgtable.h static inline pgprot_t mk_pud_sect_prot(pgprot_t prot)
prot              328 arch/arm64/include/asm/pgtable.h 	return __pgprot((pgprot_val(prot) & ~PUD_TABLE_BIT) | PUD_TYPE_SECT);
prot              331 arch/arm64/include/asm/pgtable.h static inline pgprot_t mk_pmd_sect_prot(pgprot_t prot)
prot              333 arch/arm64/include/asm/pgtable.h 	return __pgprot((pgprot_val(prot) & ~PMD_TABLE_BIT) | PMD_TYPE_SECT);
prot              388 arch/arm64/include/asm/pgtable.h #define pfn_pmd(pfn,prot)	__pmd(__phys_to_pmd_val((phys_addr_t)(pfn) << PAGE_SHIFT) | pgprot_val(prot))
prot              389 arch/arm64/include/asm/pgtable.h #define mk_pmd(page,prot)	pfn_pmd(page_to_pfn(page),prot)
prot              400 arch/arm64/include/asm/pgtable.h #define pfn_pud(pfn,prot)	__pud(__phys_to_pud_val((phys_addr_t)(pfn) << PAGE_SHIFT) | pgprot_val(prot))
prot              407 arch/arm64/include/asm/pgtable.h #define __pgprot_modify(prot,mask,bits) \
prot              408 arch/arm64/include/asm/pgtable.h 	__pgprot((pgprot_val(prot) & ~(mask)) | (bits))
prot              413 arch/arm64/include/asm/pgtable.h #define pgprot_noncached(prot) \
prot              414 arch/arm64/include/asm/pgtable.h 	__pgprot_modify(prot, PTE_ATTRINDX_MASK, PTE_ATTRINDX(MT_DEVICE_nGnRnE) | PTE_PXN | PTE_UXN)
prot              415 arch/arm64/include/asm/pgtable.h #define pgprot_writecombine(prot) \
prot              416 arch/arm64/include/asm/pgtable.h 	__pgprot_modify(prot, PTE_ATTRINDX_MASK, PTE_ATTRINDX(MT_NORMAL_NC) | PTE_PXN | PTE_UXN)
prot              417 arch/arm64/include/asm/pgtable.h #define pgprot_device(prot) \
prot              418 arch/arm64/include/asm/pgtable.h 	__pgprot_modify(prot, PTE_ATTRINDX_MASK, PTE_ATTRINDX(MT_DEVICE_nGnRE) | PTE_PXN | PTE_UXN)
prot              427 arch/arm64/include/asm/pgtable.h #define pgprot_dmacoherent(prot) \
prot              428 arch/arm64/include/asm/pgtable.h 	__pgprot_modify(prot, PTE_ATTRINDX_MASK, \
prot              519 arch/arm64/include/asm/pgtable.h #define mk_pte(page,prot)	pfn_pte(page_to_pfn(page),prot)
prot               22 arch/arm64/kernel/sys.c 		unsigned long, prot, unsigned long, flags,
prot               28 arch/arm64/kernel/sys.c 	return ksys_mmap_pgoff(addr, len, prot, flags, fd, off >> PAGE_SHIFT);
prot               56 arch/arm64/kernel/sys32.c 		       unsigned long, prot, unsigned long, flags,
prot               64 arch/arm64/kernel/sys32.c 	return ksys_mmap_pgoff(addr, len, prot, flags, fd, off_4k);
prot              247 arch/arm64/mm/dump.c 	u64 prot = val & pg_level[level].mask;
prot              251 arch/arm64/mm/dump.c 		st->current_prot = prot;
prot              254 arch/arm64/mm/dump.c 	} else if (prot != st->current_prot || level != st->level ||
prot              284 arch/arm64/mm/dump.c 		st->current_prot = prot;
prot               22 arch/arm64/mm/ioremap.c 				      pgprot_t prot, void *caller)
prot               56 arch/arm64/mm/ioremap.c 	err = ioremap_page_range(addr, addr + size, phys_addr, prot);
prot               65 arch/arm64/mm/ioremap.c void __iomem *__ioremap(phys_addr_t phys_addr, size_t size, pgprot_t prot)
prot               67 arch/arm64/mm/ioremap.c 	return __ioremap_caller(phys_addr, size, prot,
prot              141 arch/arm64/mm/mmu.c 		     phys_addr_t phys, pgprot_t prot)
prot              149 arch/arm64/mm/mmu.c 		set_pte(ptep, pfn_pte(__phys_to_pfn(phys), prot));
prot              166 arch/arm64/mm/mmu.c 				pgprot_t prot,
prot              184 arch/arm64/mm/mmu.c 		pgprot_t __prot = prot;
prot              191 arch/arm64/mm/mmu.c 			__prot = __pgprot(pgprot_val(prot) | PTE_CONT);
prot              200 arch/arm64/mm/mmu.c 		     phys_addr_t phys, pgprot_t prot,
prot              215 arch/arm64/mm/mmu.c 			pmd_set_huge(pmdp, phys, prot);
prot              224 arch/arm64/mm/mmu.c 			alloc_init_cont_pte(pmdp, addr, next, phys, prot,
prot              238 arch/arm64/mm/mmu.c 				pgprot_t prot,
prot              258 arch/arm64/mm/mmu.c 		pgprot_t __prot = prot;
prot              265 arch/arm64/mm/mmu.c 			__prot = __pgprot(pgprot_val(prot) | PTE_CONT);
prot              286 arch/arm64/mm/mmu.c 			   phys_addr_t phys, pgprot_t prot,
prot              314 arch/arm64/mm/mmu.c 			pud_set_huge(pudp, phys, prot);
prot              323 arch/arm64/mm/mmu.c 			alloc_init_cont_pmd(pudp, addr, next, phys, prot,
prot              337 arch/arm64/mm/mmu.c 				 pgprot_t prot,
prot              358 arch/arm64/mm/mmu.c 		alloc_init_pud(pgdp, addr, next, phys, prot, pgtable_alloc,
prot              400 arch/arm64/mm/mmu.c 				  phys_addr_t size, pgprot_t prot)
prot              407 arch/arm64/mm/mmu.c 	__create_pgd_mapping(init_mm.pgd, phys, virt, size, prot, NULL,
prot              413 arch/arm64/mm/mmu.c 			       pgprot_t prot, bool page_mappings_only)
prot              422 arch/arm64/mm/mmu.c 	__create_pgd_mapping(mm->pgd, phys, virt, size, prot,
prot              427 arch/arm64/mm/mmu.c 				phys_addr_t size, pgprot_t prot)
prot              435 arch/arm64/mm/mmu.c 	__create_pgd_mapping(init_mm.pgd, phys, virt, size, prot, NULL,
prot              443 arch/arm64/mm/mmu.c 				  phys_addr_t end, pgprot_t prot, int flags)
prot              446 arch/arm64/mm/mmu.c 			     prot, early_pgtable_alloc, flags);
prot              541 arch/arm64/mm/mmu.c 				      pgprot_t prot, struct vm_struct *vma,
prot              550 arch/arm64/mm/mmu.c 	__create_pgd_mapping(pgdp, pa_start, (unsigned long)va_start, size, prot,
prot              586 arch/arm64/mm/mmu.c 	pgprot_t prot = rodata_enabled ? PAGE_KERNEL_ROX : PAGE_KERNEL_EXEC;
prot              590 arch/arm64/mm/mmu.c 	pgprot_val(prot) &= ~PTE_NG;
prot              595 arch/arm64/mm/mmu.c 			     prot, __pgd_pgtable_alloc, 0);
prot              598 arch/arm64/mm/mmu.c 	__set_fixmap(FIX_ENTRY_TRAMP_TEXT, pa_start, prot);
prot              882 arch/arm64/mm/mmu.c void *__init fixmap_remap_fdt(phys_addr_t dt_phys, int *size, pgprot_t prot)
prot              919 arch/arm64/mm/mmu.c 			dt_virt_base, SWAPPER_BLOCK_SIZE, prot);
prot              930 arch/arm64/mm/mmu.c 			       round_up(offset + *size, SWAPPER_BLOCK_SIZE), prot);
prot              956 arch/arm64/mm/mmu.c int pud_set_huge(pud_t *pudp, phys_addr_t phys, pgprot_t prot)
prot              958 arch/arm64/mm/mmu.c 	pud_t new_pud = pfn_pud(__phys_to_pfn(phys), mk_pud_sect_prot(prot));
prot              970 arch/arm64/mm/mmu.c int pmd_set_huge(pmd_t *pmdp, phys_addr_t phys, pgprot_t prot)
prot              972 arch/arm64/mm/mmu.c 	pmd_t new_pmd = pfn_pmd(__phys_to_pfn(phys), mk_pmd_sect_prot(prot));
prot               40 arch/c6x/include/asm/pgtable.h #define pgprot_noncached(prot)	(prot)
prot               40 arch/csky/include/asm/io.h extern void __iomem *__ioremap(phys_addr_t addr, size_t size, pgprot_t prot);
prot               52 arch/csky/include/asm/pgtable.h #define pfn_pte(pfn, prot) __pte(((unsigned long long)(pfn) << PAGE_SHIFT) \
prot               53 arch/csky/include/asm/pgtable.h 				| pgprot_val(prot))
prot              259 arch/csky/include/asm/pgtable.h 	unsigned long prot = pgprot_val(_prot);
prot              261 arch/csky/include/asm/pgtable.h 	prot = (prot & ~_CACHE_MASK) | _CACHE_UNCACHED | _PAGE_SO;
prot              263 arch/csky/include/asm/pgtable.h 	return __pgprot(prot);
prot              269 arch/csky/include/asm/pgtable.h 	unsigned long prot = pgprot_val(_prot);
prot              271 arch/csky/include/asm/pgtable.h 	prot = (prot & ~_CACHE_MASK) | _CACHE_UNCACHED;
prot              273 arch/csky/include/asm/pgtable.h 	return __pgprot(prot);
prot              309 arch/csky/include/asm/pgtable.h #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \
prot              310 arch/csky/include/asm/pgtable.h 	remap_pfn_range(vma, vaddr, pfn, size, prot)
prot               20 arch/csky/kernel/syscall.c 	unsigned long, prot,
prot               28 arch/csky/kernel/syscall.c 	return ksys_mmap_pgoff(addr, len, prot, flags, fd,
prot               12 arch/csky/mm/ioremap.c 				      pgprot_t prot, void *caller)
prot               32 arch/csky/mm/ioremap.c 	if (ioremap_page_range(vaddr, vaddr + size, addr, prot)) {
prot               40 arch/csky/mm/ioremap.c void __iomem *__ioremap(phys_addr_t phys_addr, size_t size, pgprot_t prot)
prot               42 arch/csky/mm/ioremap.c 	return __ioremap_caller(phys_addr, size, prot,
prot               19 arch/h8300/include/asm/pgtable.h #define pgprot_writecombine(prot)  (prot)
prot              341 arch/hexagon/include/asm/pgtable.h static inline pte_t pte_modify(pte_t pte, pgprot_t prot)
prot              344 arch/hexagon/include/asm/pgtable.h 	pte_val(pte) |= pgprot_val(prot);
prot               18 arch/hexagon/mm/ioremap.c 	pgprot_t prot = __pgprot(_PAGE_PRESENT|_PAGE_READ|_PAGE_WRITE
prot               33 arch/hexagon/mm/ioremap.c 	if (ioremap_page_range(addr, addr+size, phys_addr, prot)) {
prot              926 arch/ia64/include/asm/pal.h ia64_pal_cache_prot_info (u64 cache_level, u64 cache_type, pal_cache_protection_info_t *prot)
prot              933 arch/ia64/include/asm/pal.h 		prot->pcpi_status           = iprv.status;
prot              934 arch/ia64/include/asm/pal.h 		prot->pcp_info[0].pcpi_data = iprv.v0 & 0xffffffff;
prot              935 arch/ia64/include/asm/pal.h 		prot->pcp_info[1].pcpi_data = iprv.v0 >> 32;
prot              936 arch/ia64/include/asm/pal.h 		prot->pcp_info[2].pcpi_data = iprv.v1 & 0xffffffff;
prot              937 arch/ia64/include/asm/pal.h 		prot->pcp_info[3].pcpi_data = iprv.v1 >> 32;
prot              938 arch/ia64/include/asm/pal.h 		prot->pcp_info[4].pcpi_data = iprv.v2 & 0xffffffff;
prot              939 arch/ia64/include/asm/pal.h 		prot->pcp_info[5].pcpi_data = iprv.v2 >> 32;
prot              352 arch/ia64/include/asm/pgtable.h #define pgprot_cacheable(prot)		__pgprot((pgprot_val(prot) & ~_PAGE_MA_MASK) | _PAGE_MA_WB)
prot              353 arch/ia64/include/asm/pgtable.h #define pgprot_noncached(prot)		__pgprot((pgprot_val(prot) & ~_PAGE_MA_MASK) | _PAGE_MA_UC)
prot              354 arch/ia64/include/asm/pgtable.h #define pgprot_writecombine(prot)	__pgprot((pgprot_val(prot) & ~_PAGE_MA_MASK) | _PAGE_MA_WC)
prot               28 arch/ia64/include/asm/unistd.h 				int prot, int flags,
prot               32 arch/ia64/include/asm/unistd.h 				int prot, int flags,
prot              140 arch/ia64/kernel/sys_ia64.c sys_mmap2 (unsigned long addr, unsigned long len, int prot, int flags, int fd, long pgoff)
prot              142 arch/ia64/kernel/sys_ia64.c 	addr = ksys_mmap_pgoff(addr, len, prot, flags, fd, pgoff);
prot              149 arch/ia64/kernel/sys_ia64.c sys_mmap (unsigned long addr, unsigned long len, int prot, int flags, int fd, long off)
prot              154 arch/ia64/kernel/sys_ia64.c 	addr = ksys_mmap_pgoff(addr, len, prot, flags, fd, off >> PAGE_SHIFT);
prot               37 arch/ia64/mm/ioremap.c 	pgprot_t prot;
prot               71 arch/ia64/mm/ioremap.c 		prot = PAGE_KERNEL;
prot               89 arch/ia64/mm/ioremap.c 				(unsigned long) addr + size, phys_addr, prot)) {
prot              443 arch/ia64/pci/pci.c 	pgprot_t prot;
prot              456 arch/ia64/pci/pci.c 	prot = phys_mem_access_prot(NULL, vma->vm_pgoff, size,
prot              464 arch/ia64/pci/pci.c 	vma->vm_page_prot = prot;
prot              404 arch/m68k/include/asm/mcf_pgtable.h #define pfn_pte(pfn, prot)	__pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
prot              136 arch/m68k/include/asm/motorola_pgtable.h #define pfn_pte(pfn, prot)	__pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
prot              156 arch/m68k/include/asm/pgtable_mm.h # define pgprot_noncached(prot) (__pgprot(pgprot_val(prot) | CF_PAGE_NOCACHE))
prot              163 arch/m68k/include/asm/pgtable_mm.h #define pgprot_noncached(prot)							\
prot              165 arch/m68k/include/asm/pgtable_mm.h 	 ? (__pgprot(pgprot_val(prot) | __SUN3_PAGE_NOCACHE))			\
prot              167 arch/m68k/include/asm/pgtable_mm.h 	    ? (__pgprot(pgprot_val(prot) | _PAGE_NOCACHE030))			\
prot              169 arch/m68k/include/asm/pgtable_mm.h 	    ? (__pgprot((pgprot_val(prot) & _CACHEMASK040) | _PAGE_NOCACHE_S))	\
prot              170 arch/m68k/include/asm/pgtable_mm.h 	    : (prot)))
prot              172 arch/m68k/include/asm/pgtable_mm.h pgprot_t pgprot_dmacoherent(pgprot_t prot);
prot              173 arch/m68k/include/asm/pgtable_mm.h #define pgprot_dmacoherent(prot)	pgprot_dmacoherent(prot)
prot               26 arch/m68k/kernel/dma.c pgprot_t pgprot_dmacoherent(pgprot_t prot)
prot               29 arch/m68k/kernel/dma.c 		pgprot_val(prot) &= ~_PAGE_CACHE040;
prot               30 arch/m68k/kernel/dma.c 		pgprot_val(prot) |= _PAGE_GLOBAL040 | _PAGE_NOCACHE_S;
prot               32 arch/m68k/kernel/dma.c 		pgprot_val(prot) |= _PAGE_NOCACHE030;
prot               34 arch/m68k/kernel/dma.c 	return prot;
prot               41 arch/m68k/kernel/sys_m68k.c 	unsigned long prot, unsigned long flags,
prot               49 arch/m68k/kernel/sys_m68k.c 	return ksys_mmap_pgoff(addr, len, prot, flags, fd, pgoff);
prot               56 arch/microblaze/include/asm/highmem.h extern void *kmap_atomic_prot(struct page *page, pgprot_t prot);
prot               67 arch/microblaze/include/asm/pci.h 					 pgprot_t prot);
prot               51 arch/microblaze/include/asm/pgtable.h #define pgprot_noncached_wc(prot)	prot
prot              104 arch/microblaze/include/asm/pgtable.h #define pgprot_noncached(prot) \
prot              105 arch/microblaze/include/asm/pgtable.h 			(__pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | \
prot              108 arch/microblaze/include/asm/pgtable.h #define pgprot_noncached_wc(prot) \
prot              109 arch/microblaze/include/asm/pgtable.h 			 (__pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | \
prot              312 arch/microblaze/include/asm/pgtable.h #define pfn_pte(pfn, prot) \
prot              313 arch/microblaze/include/asm/pgtable.h 	__pte(((pte_basic_t)(pfn) << PFN_SHIFT_OFFSET) | pgprot_val(prot))
prot               37 arch/microblaze/kernel/sys_microblaze.c 		unsigned long, prot, unsigned long, flags, unsigned long, fd,
prot               43 arch/microblaze/kernel/sys_microblaze.c 	return ksys_mmap_pgoff(addr, len, prot, flags, fd, pgoff >> PAGE_SHIFT);
prot               47 arch/microblaze/kernel/sys_microblaze.c 		unsigned long, prot, unsigned long, flags, unsigned long, fd,
prot               53 arch/microblaze/kernel/sys_microblaze.c 	return ksys_mmap_pgoff(addr, len, prot, flags, fd,
prot               35 arch/microblaze/mm/highmem.c void *kmap_atomic_prot(struct page *page, pgprot_t prot)
prot               53 arch/microblaze/mm/highmem.c 	set_pte_at(&init_mm, vaddr, kmap_pte-idx, mk_pte(page, prot));
prot              176 arch/microblaze/pci/pci-common.c 				  pgprot_t prot)
prot              184 arch/microblaze/pci/pci-common.c 		return prot;
prot              186 arch/microblaze/pci/pci-common.c 	prot = pgprot_noncached(prot);
prot              207 arch/microblaze/pci/pci-common.c 			prot = pgprot_noncached_wc(prot);
prot              212 arch/microblaze/pci/pci-common.c 		 (unsigned long long)offset, pgprot_val(prot));
prot              214 arch/microblaze/pci/pci-common.c 	return prot;
prot              160 arch/mips/include/asm/pgtable-32.h pfn_pte(unsigned long pfn, pgprot_t prot)
prot              165 arch/mips/include/asm/pgtable-32.h 				(pgprot_val(prot) & ~_PFNX_MASK);
prot              167 arch/mips/include/asm/pgtable-32.h 				(pgprot_val(prot) & ~_PFN_MASK);
prot              175 arch/mips/include/asm/pgtable-32.h static inline pte_t pfn_pte(unsigned long pfn, pgprot_t prot)
prot              179 arch/mips/include/asm/pgtable-32.h 	pte.pte_high = (pfn << 6) | (pgprot_val(prot) & 0x3f);
prot              180 arch/mips/include/asm/pgtable-32.h 	pte.pte_low = pgprot_val(prot);
prot              189 arch/mips/include/asm/pgtable-32.h #define pfn_pte(pfn, prot)	__pte(((pfn) << (PAGE_SHIFT + 2)) | pgprot_val(prot))
prot              192 arch/mips/include/asm/pgtable-32.h #define pfn_pte(pfn, prot)	__pte(((unsigned long long)(pfn) << _PFN_SHIFT) | pgprot_val(prot))
prot              193 arch/mips/include/asm/pgtable-32.h #define pfn_pmd(pfn, prot)	__pmd(((unsigned long long)(pfn) << _PFN_SHIFT) | pgprot_val(prot))
prot              315 arch/mips/include/asm/pgtable-64.h #define pfn_pte(pfn, prot)	__pte(((pfn) << (PAGE_SHIFT + 2)) | pgprot_val(prot))
prot              318 arch/mips/include/asm/pgtable-64.h #define pfn_pte(pfn, prot)	__pte(((pfn) << _PFN_SHIFT) | pgprot_val(prot))
prot              319 arch/mips/include/asm/pgtable-64.h #define pfn_pmd(pfn, prot)	__pmd(((pfn) << _PFN_SHIFT) | pgprot_val(prot))
prot              422 arch/mips/include/asm/pgtable.h 	unsigned long prot = pgprot_val(_prot);
prot              424 arch/mips/include/asm/pgtable.h 	prot = (prot & ~_CACHE_MASK) | _CACHE_UNCACHED;
prot              426 arch/mips/include/asm/pgtable.h 	return __pgprot(prot);
prot              433 arch/mips/include/asm/pgtable.h 	unsigned long prot = pgprot_val(_prot);
prot              436 arch/mips/include/asm/pgtable.h 	prot = (prot & ~_CACHE_MASK) | cpu_data[0].writecombine;
prot              438 arch/mips/include/asm/pgtable.h 	return __pgprot(prot);
prot              495 arch/mips/include/asm/pgtable.h extern int remap_pfn_range(struct vm_area_struct *vma, unsigned long from, unsigned long pfn, unsigned long size, pgprot_t prot);
prot              501 arch/mips/include/asm/pgtable.h 		pgprot_t prot)
prot              504 arch/mips/include/asm/pgtable.h 	return remap_pfn_range(vma, vaddr, phys_addr_high >> PAGE_SHIFT, size, prot);
prot              596 arch/mips/include/asm/pgtable.h extern pmd_t mk_pmd(struct page *page, pgprot_t prot);
prot              615 arch/mips/include/asm/r4kcache.h #define __BUILD_BLAST_CACHE_RANGE(pfx, desc, hitop, prot, extra)	\
prot              616 arch/mips/include/asm/r4kcache.h static inline void prot##extra##blast_##pfx##cache##_range(unsigned long start, \
prot              624 arch/mips/include/asm/r4kcache.h 		prot##cache_op(hitop, addr);				\
prot               61 arch/mips/kernel/syscall.c 	unsigned long, prot, unsigned long, flags, unsigned long,
prot               66 arch/mips/kernel/syscall.c 	return ksys_mmap_pgoff(addr, len, prot, flags, fd,
prot               71 arch/mips/kernel/syscall.c 	unsigned long, prot, unsigned long, flags, unsigned long, fd,
prot               77 arch/mips/kernel/syscall.c 	return ksys_mmap_pgoff(addr, len, prot, flags, fd,
prot               84 arch/mips/mm/init.c static void *__kmap_pgprot(struct page *page, unsigned long addr, pgprot_t prot)
prot              100 arch/mips/mm/init.c 	pte = mk_pte(page, prot);
prot               35 arch/mips/mm/pgtable-32.c pmd_t mk_pmd(struct page *page, pgprot_t prot)
prot               39 arch/mips/mm/pgtable-32.c 	pmd_val(pmd) = (page_to_pfn(page) << _PFN_SHIFT) | pgprot_val(prot);
prot               91 arch/mips/mm/pgtable-64.c pmd_t mk_pmd(struct page *page, pgprot_t prot)
prot               95 arch/mips/mm/pgtable-64.c 	pmd_val(pmd) = (page_to_pfn(page) << _PFN_SHIFT) | pgprot_val(prot);
prot               26 arch/nds32/include/asm/fixmap.h void __set_fixmap(enum fixed_addresses idx, phys_addr_t phys, pgprot_t prot);
prot              188 arch/nds32/include/asm/pgtable.h #define pfn_pte(pfn,prot)	(__pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot)))
prot              307 arch/nds32/include/asm/pgtable.h #define pgprot_noncached(prot)	   __pgprot((pgprot_val(prot)&~_PAGE_C_MASK) | _PAGE_C_DEV)
prot              308 arch/nds32/include/asm/pgtable.h #define pgprot_writecombine(prot)  __pgprot((pgprot_val(prot)&~_PAGE_C_MASK) | _PAGE_C_DEV_WB)
prot              317 arch/nds32/include/asm/pgtable.h static inline pmd_t __mk_pmd(pte_t * ptep, unsigned long prot)
prot              327 arch/nds32/include/asm/pgtable.h 	pmd_val(pmd) = __virt_to_phys(ptr) | prot;
prot              342 arch/nds32/include/asm/pgtable.h #define mk_pte(page,prot)	pfn_pte(page_to_pfn(page),prot)
prot              354 arch/nds32/include/asm/pgtable.h #define page_pte_prot(page,prot)     	mk_pte(page, prot)
prot               13 arch/nds32/kernel/sys_nds32.c 	       unsigned long, prot, unsigned long, flags,
prot               19 arch/nds32/kernel/sys_nds32.c 	return sys_mmap_pgoff(addr, len, prot, flags, fd,
prot              120 arch/nds32/kernel/vdso.c 	pgprot_t prot;
prot              170 arch/nds32/kernel/vdso.c 	prot = __pgprot(_PAGE_V | _PAGE_M_UR_KR | _PAGE_D |  _PAGE_C_DEV);
prot              172 arch/nds32/kernel/vdso.c 				 PAGE_SIZE, prot);
prot               16 arch/nds32/mm/ioremap.c 	pgprot_t prot;
prot               39 arch/nds32/mm/ioremap.c 	prot = __pgprot(_PAGE_V | _PAGE_M_KRW | _PAGE_D |
prot               41 arch/nds32/mm/ioremap.c 	if (ioremap_page_range(addr, addr + size, phys_addr, prot)) {
prot              122 arch/nios2/include/asm/pgtable.h 	unsigned long prot = pgprot_val(_prot);
prot              124 arch/nios2/include/asm/pgtable.h 	prot &= ~_PAGE_CACHED;
prot              126 arch/nios2/include/asm/pgtable.h 	return __pgprot(prot);
prot              199 arch/nios2/include/asm/pgtable.h #define pfn_pte(pfn, prot)	(__pte(pfn | pgprot_val(prot)))
prot              241 arch/nios2/include/asm/pgtable.h #define mk_pte(page, prot)	(pfn_pte(page_to_pfn(page), prot))
prot              402 arch/openrisc/include/asm/pgtable.h #define pfn_pte(pfn, prot)  __pte((((pfn) << PAGE_SHIFT)) | pgprot_val(prot))
prot               72 arch/openrisc/mm/init.c 	pgprot_t prot;
prot              114 arch/openrisc/mm/init.c 					prot = PAGE_KERNEL;
prot              116 arch/openrisc/mm/init.c 					prot = PAGE_KERNEL_RO;
prot              118 arch/openrisc/mm/init.c 				set_pte(pte, mk_pte_phys(p, prot));
prot              132 arch/parisc/include/asm/cacheflush.h #define kmap_atomic_prot(page, prot)	kmap_atomic(page)
prot              569 arch/parisc/include/asm/pgtable.h #define pgprot_noncached(prot) __pgprot(pgprot_val(prot) | _PAGE_NO_CACHE)
prot              256 arch/parisc/kernel/sys_parisc.c 	unsigned long prot, unsigned long flags, unsigned long fd,
prot              261 arch/parisc/kernel/sys_parisc.c 	return ksys_mmap_pgoff(addr, len, prot, flags, fd,
prot              266 arch/parisc/kernel/sys_parisc.c 		unsigned long prot, unsigned long flags, unsigned long fd,
prot              270 arch/parisc/kernel/sys_parisc.c 		return ksys_mmap_pgoff(addr, len, prot, flags, fd,
prot              433 arch/parisc/mm/init.c 				pgprot_t prot;
prot              437 arch/parisc/mm/init.c 					prot = pgprot;
prot              440 arch/parisc/mm/init.c 					prot = PAGE_KERNEL;
prot              443 arch/parisc/mm/init.c 					prot = PAGE_KERNEL_RWX;
prot              447 arch/parisc/mm/init.c 					prot = (address < ro_end) ?
prot              451 arch/parisc/mm/init.c 					prot = PAGE_KERNEL;
prot              454 arch/parisc/mm/init.c 				pte = __mk_pte(address, prot);
prot              140 arch/powerpc/include/asm/book3s/32/pgtable.h int map_kernel_page(unsigned long va, phys_addr_t pa, pgprot_t prot);
prot              583 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pgprot_t pgprot_noncached(pgprot_t prot)
prot              585 arch/powerpc/include/asm/book3s/32/pgtable.h 	return __pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) |
prot              590 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pgprot_t pgprot_noncached_wc(pgprot_t prot)
prot              592 arch/powerpc/include/asm/book3s/32/pgtable.h 	return __pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) |
prot              597 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pgprot_t pgprot_cached(pgprot_t prot)
prot              599 arch/powerpc/include/asm/book3s/32/pgtable.h 	return __pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) |
prot              604 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pgprot_t pgprot_cached_wthru(pgprot_t prot)
prot              606 arch/powerpc/include/asm/book3s/32/pgtable.h 	return __pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) |
prot              611 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pgprot_t pgprot_cached_noncoherent(pgprot_t prot)
prot              613 arch/powerpc/include/asm/book3s/32/pgtable.h 	return __pgprot(pgprot_val(prot) & ~_PAGE_CACHE_CTL);
prot              617 arch/powerpc/include/asm/book3s/32/pgtable.h static inline pgprot_t pgprot_writecombine(pgprot_t prot)
prot              619 arch/powerpc/include/asm/book3s/32/pgtable.h 	return pgprot_noncached_wc(prot);
prot               69 arch/powerpc/include/asm/book3s/64/hash-4k.h #define remap_4k_pfn(vma, addr, pfn, prot)	\
prot               70 arch/powerpc/include/asm/book3s/64/hash-4k.h 	remap_pfn_range((vma), (addr), (pfn), PAGE_SIZE, (prot))
prot              163 arch/powerpc/include/asm/book3s/64/hash-64k.h 				 unsigned long pfn, pgprot_t prot)
prot              170 arch/powerpc/include/asm/book3s/64/hash-64k.h 			       __pgprot(pgprot_val(prot) | H_PAGE_4K_PFN));
prot              247 arch/powerpc/include/asm/book3s/64/hash.h int hash__map_kernel_page(unsigned long ea, unsigned long pa, pgprot_t prot);
prot              491 arch/powerpc/include/asm/book3s/64/mmu-hash.h 			     unsigned long pstart, unsigned long prot,
prot               68 arch/powerpc/include/asm/book3s/64/pgtable-64k.h 			       unsigned long pfn, pgprot_t prot)
prot               72 arch/powerpc/include/asm/book3s/64/pgtable-64k.h 	return hash__remap_4k_pfn(vma, addr, pfn, prot);
prot              826 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pgprot_t pgprot_noncached(pgprot_t prot)
prot              828 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) |
prot              833 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pgprot_t pgprot_noncached_wc(pgprot_t prot)
prot              835 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) |
prot              840 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pgprot_t pgprot_cached(pgprot_t prot)
prot              842 arch/powerpc/include/asm/book3s/64/pgtable.h 	return __pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL));
prot              846 arch/powerpc/include/asm/book3s/64/pgtable.h static inline pgprot_t pgprot_writecombine(pgprot_t prot)
prot              848 arch/powerpc/include/asm/book3s/64/pgtable.h 	return pgprot_noncached_wc(prot);
prot             1037 arch/powerpc/include/asm/book3s/64/pgtable.h static inline int map_kernel_page(unsigned long ea, unsigned long pa, pgprot_t prot)
prot             1044 arch/powerpc/include/asm/book3s/64/pgtable.h 		return radix__map_kernel_page(ea, pa, prot, PAGE_SIZE);
prot             1046 arch/powerpc/include/asm/book3s/64/pgtable.h 	return hash__map_kernel_page(ea, pa, prot);
prot               64 arch/powerpc/include/asm/highmem.h extern void *kmap_atomic_prot(struct page *page, pgprot_t prot);
prot               53 arch/powerpc/include/asm/io-workarounds.h 			   pgprot_t prot, void *caller);
prot              726 arch/powerpc/include/asm/io.h 			unsigned long size, pgprot_t prot);
prot              728 arch/powerpc/include/asm/io.h 			 pgprot_t prot, void *caller);
prot              731 arch/powerpc/include/asm/io.h 				      pgprot_t prot, void *caller);
prot              734 arch/powerpc/include/asm/io.h 				   unsigned long size, pgprot_t prot);
prot               20 arch/powerpc/include/asm/mman.h static inline unsigned long arch_calc_vm_prot_bits(unsigned long prot,
prot               24 arch/powerpc/include/asm/mman.h 	return (((prot & PROT_SAO) ? VM_SAO : 0) | pkey_to_vmflag_bits(pkey));
prot               26 arch/powerpc/include/asm/mman.h 	return ((prot & PROT_SAO) ? VM_SAO : 0);
prot               29 arch/powerpc/include/asm/mman.h #define arch_calc_vm_prot_bits(prot, pkey) arch_calc_vm_prot_bits(prot, pkey)
prot               43 arch/powerpc/include/asm/mman.h static inline bool arch_validate_prot(unsigned long prot, unsigned long addr)
prot               45 arch/powerpc/include/asm/mman.h 	if (prot & ~(PROT_READ | PROT_WRITE | PROT_EXEC | PROT_SEM | PROT_SAO))
prot               47 arch/powerpc/include/asm/mman.h 	if ((prot & PROT_SAO) && !cpu_has_feature(CPU_FTR_SAO))
prot               67 arch/powerpc/include/asm/nohash/32/pgtable.h int map_kernel_page(unsigned long va, phys_addr_t pa, pgprot_t prot);
prot               90 arch/powerpc/include/asm/nohash/64/pgtable-4k.h #define remap_4k_pfn(vma, addr, pfn, prot)	\
prot               91 arch/powerpc/include/asm/nohash/64/pgtable-4k.h 	remap_pfn_range((vma), (addr), (pfn), PAGE_SIZE, (prot))
prot              358 arch/powerpc/include/asm/nohash/64/pgtable.h int map_kernel_page(unsigned long ea, unsigned long pa, pgprot_t prot);
prot              240 arch/powerpc/include/asm/nohash/pgtable.h #define pgprot_noncached(prot)	  (__pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | \
prot              243 arch/powerpc/include/asm/nohash/pgtable.h #define pgprot_noncached_wc(prot) (__pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | \
prot              246 arch/powerpc/include/asm/nohash/pgtable.h #define pgprot_cached(prot)       (__pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | \
prot              250 arch/powerpc/include/asm/nohash/pgtable.h #define pgprot_cached_wthru(prot) (__pgprot((pgprot_val(prot) & ~_PAGE_CACHE_CTL) | \
prot              253 arch/powerpc/include/asm/nohash/pgtable.h #define pgprot_cached_wthru(prot)	pgprot_noncached(prot)
prot              256 arch/powerpc/include/asm/nohash/pgtable.h #define pgprot_cached_noncoherent(prot) \
prot              257 arch/powerpc/include/asm/nohash/pgtable.h 		(__pgprot(pgprot_val(prot) & ~_PAGE_CACHE_CTL))
prot              113 arch/powerpc/include/asm/pci.h 					 pgprot_t prot);
prot              160 arch/powerpc/include/asm/pkeys.h 					 int prot, int pkey);
prot              162 arch/powerpc/include/asm/pkeys.h 					      int prot, int pkey)
prot              174 arch/powerpc/include/asm/pkeys.h 	return __arch_override_mprotect_pkey(vma, prot, pkey);
prot               13 arch/powerpc/include/asm/syscalls.h 		unsigned long prot, unsigned long flags,
prot               16 arch/powerpc/include/asm/syscalls.h 		unsigned long prot, unsigned long flags,
prot              153 arch/powerpc/kernel/io-workarounds.c 			   pgprot_t prot, void *caller)
prot              156 arch/powerpc/kernel/io-workarounds.c 	void __iomem *res = __ioremap_caller(addr, size, prot, caller);
prot              440 arch/powerpc/kernel/pci-common.c 				  pgprot_t prot)
prot              448 arch/powerpc/kernel/pci-common.c 		return prot;
prot              450 arch/powerpc/kernel/pci-common.c 	prot = pgprot_noncached(prot);
prot              471 arch/powerpc/kernel/pci-common.c 			prot = pgprot_noncached_wc(prot);
prot              476 arch/powerpc/kernel/pci-common.c 		 (unsigned long long)offset, pgprot_val(prot));
prot              478 arch/powerpc/kernel/pci-common.c 	return prot;
prot               52 arch/powerpc/kernel/sys_ppc32.c 			  unsigned long prot, unsigned long flags,
prot               56 arch/powerpc/kernel/sys_ppc32.c 	return sys_mmap(addr, len, prot, flags, fd, pgoff << 12);
prot               41 arch/powerpc/kernel/syscalls.c 			unsigned long prot, unsigned long flags,
prot               46 arch/powerpc/kernel/syscalls.c 	if (!arch_validate_prot(prot, addr))
prot               55 arch/powerpc/kernel/syscalls.c 	ret = ksys_mmap_pgoff(addr, len, prot, flags, fd, off);
prot               61 arch/powerpc/kernel/syscalls.c 		unsigned long, prot, unsigned long, flags,
prot               64 arch/powerpc/kernel/syscalls.c 	return do_mmap2(addr, len, prot, flags, fd, pgoff, PAGE_SHIFT-12);
prot               68 arch/powerpc/kernel/syscalls.c 		unsigned long, prot, unsigned long, flags,
prot               71 arch/powerpc/kernel/syscalls.c 	return do_mmap2(addr, len, prot, flags, fd, offset, PAGE_SHIFT);
prot              123 arch/powerpc/mm/book3s32/mmu.c 		    unsigned int size, pgprot_t prot)
prot              128 arch/powerpc/mm/book3s32/mmu.c 	unsigned long flags = pgprot_val(prot);
prot              250 arch/powerpc/mm/book3s32/mmu.c 		   unsigned int size, pgprot_t prot)
prot              255 arch/powerpc/mm/book3s32/mmu.c 	unsigned long flags = pgprot_val(prot);
prot              148 arch/powerpc/mm/book3s64/hash_pgtable.c int hash__map_kernel_page(unsigned long ea, unsigned long pa, pgprot_t prot)
prot              167 arch/powerpc/mm/book3s64/hash_pgtable.c 		set_pte_at(&init_mm, ea, ptep, pfn_pte(pa >> PAGE_SHIFT, prot));
prot              175 arch/powerpc/mm/book3s64/hash_pgtable.c 		if (htab_bolt_mapping(ea, ea + PAGE_SIZE, pa, pgprot_val(prot),
prot              245 arch/powerpc/mm/book3s64/hash_utils.c 		      unsigned long pstart, unsigned long prot,
prot              255 arch/powerpc/mm/book3s64/hash_utils.c 	prot = htab_convert_pte_flags(prot);
prot              258 arch/powerpc/mm/book3s64/hash_utils.c 	    vstart, vend, pstart, prot, psize, ssize);
prot              265 arch/powerpc/mm/book3s64/hash_utils.c 		unsigned long tprot = prot;
prot              840 arch/powerpc/mm/book3s64/hash_utils.c 	unsigned long prot;
prot              916 arch/powerpc/mm/book3s64/hash_utils.c 	prot = pgprot_val(PAGE_KERNEL);
prot              936 arch/powerpc/mm/book3s64/hash_utils.c 		    base, size, prot);
prot              944 arch/powerpc/mm/book3s64/hash_utils.c 				prot, mmu_linear_psize, mmu_kernel_ssize));
prot              963 arch/powerpc/mm/book3s64/hash_utils.c 					 __pa(tce_alloc_start), prot,
prot              337 arch/powerpc/mm/book3s64/pkeys.c int __arch_override_mprotect_pkey(struct vm_area_struct *vma, int prot,
prot              344 arch/powerpc/mm/book3s64/pkeys.c 	if (vma_is_pkey_exec_only(vma) && (prot != PROT_EXEC))
prot              351 arch/powerpc/mm/book3s64/pkeys.c 	if (prot == PROT_EXEC) {
prot              259 arch/powerpc/mm/book3s64/radix_pgtable.c 	pgprot_t prot;
prot              288 arch/powerpc/mm/book3s64/radix_pgtable.c 			prot = PAGE_KERNEL_X;
prot              291 arch/powerpc/mm/book3s64/radix_pgtable.c 			prot = PAGE_KERNEL;
prot              300 arch/powerpc/mm/book3s64/radix_pgtable.c 		rc = __map_kernel_page(vaddr, addr, prot, mapping_size, nid, start, end);
prot             1100 arch/powerpc/mm/book3s64/radix_pgtable.c int pud_set_huge(pud_t *pud, phys_addr_t addr, pgprot_t prot)
prot             1103 arch/powerpc/mm/book3s64/radix_pgtable.c 	pte_t new_pud = pfn_pte(__phys_to_pfn(addr), prot);
prot             1147 arch/powerpc/mm/book3s64/radix_pgtable.c int pmd_set_huge(pmd_t *pmd, phys_addr_t addr, pgprot_t prot)
prot             1150 arch/powerpc/mm/book3s64/radix_pgtable.c 	pte_t new_pmd = pfn_pte(__phys_to_pfn(addr), prot);
prot               33 arch/powerpc/mm/highmem.c void *kmap_atomic_prot(struct page *page, pgprot_t prot)
prot               47 arch/powerpc/mm/highmem.c 	__set_pte_at(&init_mm, vaddr, kmap_pte-idx, mk_pte(page, prot), 1);
prot               13 arch/powerpc/mm/ioremap.c 	pgprot_t prot = pgprot_noncached(PAGE_KERNEL);
prot               17 arch/powerpc/mm/ioremap.c 		return iowa_ioremap(addr, size, prot, caller);
prot               18 arch/powerpc/mm/ioremap.c 	return __ioremap_caller(addr, size, prot, caller);
prot               24 arch/powerpc/mm/ioremap.c 	pgprot_t prot = pgprot_noncached_wc(PAGE_KERNEL);
prot               28 arch/powerpc/mm/ioremap.c 		return iowa_ioremap(addr, size, prot, caller);
prot               29 arch/powerpc/mm/ioremap.c 	return __ioremap_caller(addr, size, prot, caller);
prot               35 arch/powerpc/mm/ioremap.c 	pgprot_t prot = pgprot_cached(PAGE_KERNEL);
prot               39 arch/powerpc/mm/ioremap.c 		return iowa_ioremap(addr, size, prot, caller);
prot               40 arch/powerpc/mm/ioremap.c 	return __ioremap_caller(addr, size, prot, caller);
prot               63 arch/powerpc/mm/ioremap.c 			unsigned long size, pgprot_t prot)
prot               68 arch/powerpc/mm/ioremap.c 		int err = map_kernel_page(ea + i, pa + i, prot);
prot               78 arch/powerpc/mm/ioremap.c 			 pgprot_t prot, void *caller)
prot               91 arch/powerpc/mm/ioremap.c 	ret = ioremap_page_range(va, va + size, pa, prot);
prot               11 arch/powerpc/mm/ioremap_32.c 	pgprot_t prot = pgprot_cached_wthru(PAGE_KERNEL);
prot               13 arch/powerpc/mm/ioremap_32.c 	return __ioremap_caller(addr, size, prot, __builtin_return_address(0));
prot               18 arch/powerpc/mm/ioremap_32.c __ioremap_caller(phys_addr_t addr, unsigned long size, pgprot_t prot, void *caller)
prot               66 arch/powerpc/mm/ioremap_32.c 		return do_ioremap(p, offset, size, prot, caller);
prot               72 arch/powerpc/mm/ioremap_32.c 	err = early_ioremap_range(ioremap_bot - size, p, size, prot);
prot               10 arch/powerpc/mm/ioremap_64.c void __iomem *__ioremap_at(phys_addr_t pa, void *ea, unsigned long size, pgprot_t prot)
prot               16 arch/powerpc/mm/ioremap_64.c 	if (pgprot_val(prot) & H_PAGE_4K_PFN)
prot               29 arch/powerpc/mm/ioremap_64.c 		ret = ioremap_page_range(va, va + size, pa, prot);
prot               33 arch/powerpc/mm/ioremap_64.c 		ret = early_ioremap_range(va, pa, size, prot);
prot               58 arch/powerpc/mm/ioremap_64.c 			       pgprot_t prot, void *caller)
prot               65 arch/powerpc/mm/ioremap_64.c 	if (pgprot_val(prot) & H_PAGE_4K_PFN)
prot               82 arch/powerpc/mm/ioremap_64.c 		return do_ioremap(paligned, offset, size, prot, caller);
prot               84 arch/powerpc/mm/ioremap_64.c 	err = early_ioremap_range(ioremap_bot, paligned, size, prot);
prot               23 arch/powerpc/mm/kasan/kasan_init_32.c static void kasan_populate_pte(pte_t *ptep, pgprot_t prot)
prot               30 arch/powerpc/mm/kasan/kasan_init_32.c 		__set_pte_at(&init_mm, va, ptep, pfn_pte(PHYS_PFN(pa), prot), 0);
prot               37 arch/powerpc/mm/kasan/kasan_init_32.c 	pgprot_t prot = slab_is_available() ? kasan_prot_ro() : PAGE_KERNEL;
prot               55 arch/powerpc/mm/kasan/kasan_init_32.c 		kasan_populate_pte(new, prot);
prot              112 arch/powerpc/mm/kasan/kasan_init_32.c 	pgprot_t prot = kasan_prot_ro();
prot              118 arch/powerpc/mm/kasan/kasan_init_32.c 	kasan_populate_pte(kasan_early_shadow_pte, prot);
prot              127 arch/powerpc/mm/kasan/kasan_init_32.c 		__set_pte_at(&init_mm, k_cur, ptep, pfn_pte(PHYS_PFN(pa), prot), 0);
prot               98 arch/powerpc/mm/mmu_decl.h 		   unsigned int size, pgprot_t prot);
prot               73 arch/powerpc/mm/nohash/book3e_pgtable.c int __ref map_kernel_page(unsigned long ea, unsigned long pa, pgprot_t prot)
prot              112 arch/powerpc/mm/nohash/book3e_pgtable.c 	set_pte_at(&init_mm, ea, ptep, pfn_pte(pa >> PAGE_SHIFT, prot));
prot               59 arch/powerpc/mm/pgtable_32.c int __ref map_kernel_page(unsigned long va, phys_addr_t pa, pgprot_t prot)
prot               77 arch/powerpc/mm/pgtable_32.c 		BUG_ON((pte_present(*pg) | pte_hashpte(*pg)) && pgprot_val(prot));
prot               78 arch/powerpc/mm/pgtable_32.c 		set_pte_at(&init_mm, va, pg, pfn_pte(pa >> PAGE_SHIFT, prot));
prot              161 arch/powerpc/mm/pgtable_32.c static int __change_page_attr_noflush(struct page *page, pgprot_t prot)
prot              174 arch/powerpc/mm/pgtable_32.c 	__set_pte_at(&init_mm, address, kpte, mk_pte(page, prot), 0);
prot              185 arch/powerpc/mm/pgtable_32.c static int change_page_attr(struct page *page, int numpages, pgprot_t prot)
prot              193 arch/powerpc/mm/pgtable_32.c 		err = __change_page_attr_noflush(page, prot);
prot              173 arch/powerpc/platforms/cell/iommu.c 	const unsigned long prot = 0xc48;
prot              175 arch/powerpc/platforms/cell/iommu.c 		((prot << (52 + 4 * direction)) &
prot               41 arch/riscv/include/asm/fixmap.h 			 phys_addr_t phys, pgprot_t prot);
prot               68 arch/riscv/include/asm/pgtable-64.h static inline pmd_t pfn_pmd(unsigned long pfn, pgprot_t prot)
prot               70 arch/riscv/include/asm/pgtable-64.h 	return __pmd((pfn << _PAGE_PFN_SHIFT) | pgprot_val(prot));
prot              148 arch/riscv/include/asm/pgtable.h static inline pgd_t pfn_pgd(unsigned long pfn, pgprot_t prot)
prot              150 arch/riscv/include/asm/pgtable.h 	return __pgd((pfn << _PAGE_PFN_SHIFT) | pgprot_val(prot));
prot              187 arch/riscv/include/asm/pgtable.h static inline pte_t pfn_pte(unsigned long pfn, pgprot_t prot)
prot              189 arch/riscv/include/asm/pgtable.h 	return __pte((pfn << _PAGE_PFN_SHIFT) | pgprot_val(prot));
prot              192 arch/riscv/include/asm/pgtable.h #define mk_pte(page, prot)       pfn_pte(page_to_pfn(page), prot)
prot               13 arch/riscv/kernel/sys_riscv.c 			   unsigned long prot, unsigned long flags,
prot               19 arch/riscv/kernel/sys_riscv.c 	return ksys_mmap_pgoff(addr, len, prot, flags, fd,
prot               25 arch/riscv/kernel/sys_riscv.c 	unsigned long, prot, unsigned long, flags,
prot               28 arch/riscv/kernel/sys_riscv.c 	return riscv_sys_mmap(addr, len, prot, flags, fd, offset, 0);
prot               32 arch/riscv/kernel/sys_riscv.c 	unsigned long, prot, unsigned long, flags,
prot               39 arch/riscv/kernel/sys_riscv.c 	return riscv_sys_mmap(addr, len, prot, flags, fd, offset, 12);
prot              161 arch/riscv/mm/init.c void __set_fixmap(enum fixed_addresses idx, phys_addr_t phys, pgprot_t prot)
prot              170 arch/riscv/mm/init.c 	if (pgprot_val(prot)) {
prot              171 arch/riscv/mm/init.c 		set_pte(ptep, pfn_pte(phys >> PAGE_SHIFT, prot));
prot              201 arch/riscv/mm/init.c 				      phys_addr_t sz, pgprot_t prot)
prot              208 arch/riscv/mm/init.c 		ptep[pte_index] = pfn_pte(PFN_DOWN(pa), prot);
prot              247 arch/riscv/mm/init.c 				      phys_addr_t sz, pgprot_t prot)
prot              255 arch/riscv/mm/init.c 			pmdp[pmd_index] = pfn_pmd(PFN_DOWN(pa), prot);
prot              269 arch/riscv/mm/init.c 	create_pte_mapping(ptep, va, pa, sz, prot);
prot              291 arch/riscv/mm/init.c 				      phys_addr_t sz, pgprot_t prot)
prot              299 arch/riscv/mm/init.c 			pgdp[pgd_index] = pfn_pgd(PFN_DOWN(pa), prot);
prot              313 arch/riscv/mm/init.c 	create_pgd_next_mapping(nextp, va, pa, sz, prot);
prot               24 arch/riscv/mm/ioremap.c 	pgprot_t prot, void *caller)
prot               45 arch/riscv/mm/ioremap.c 	if (ioremap_page_range(vaddr, vaddr + size, addr, prot)) {
prot              143 arch/s390/include/asm/gmap.h 			 unsigned long len, int prot);
prot             1145 arch/s390/include/asm/pgtable.h 		    pte_t *ptep, int prot, unsigned long bit);
prot              427 arch/s390/include/uapi/asm/ptrace.h 	ptprot_flags prot;
prot              198 arch/s390/kernel/compat_linux.c 	compat_ulong_t prot;
prot              212 arch/s390/kernel/compat_linux.c 	return ksys_mmap_pgoff(a.addr, a.len, a.prot, a.flags, a.fd,
prot              222 arch/s390/kernel/compat_linux.c 	return ksys_mmap_pgoff(a.addr, a.len, a.prot, a.flags, a.fd, a.offset);
prot              241 arch/s390/kernel/crash_dump.c 					unsigned long size, pgprot_t prot)
prot              250 arch/s390/kernel/crash_dump.c 				     size_old, prot);
prot              257 arch/s390/kernel/crash_dump.c 	return remap_pfn_range(vma, from, pfn, size, prot);
prot              269 arch/s390/kernel/crash_dump.c 					   unsigned long size, pgprot_t prot)
prot              282 arch/s390/kernel/crash_dump.c 	return remap_pfn_range(vma, from, pfn, size, prot);
prot              289 arch/s390/kernel/crash_dump.c 			   unsigned long pfn, unsigned long size, pgprot_t prot)
prot              292 arch/s390/kernel/crash_dump.c 		return remap_oldmem_pfn_range_kdump(vma, from, pfn, size, prot);
prot              295 arch/s390/kernel/crash_dump.c 						       prot);
prot               43 arch/s390/kernel/sys_s390.c 	unsigned long prot;
prot               56 arch/s390/kernel/sys_s390.c 	error = ksys_mmap_pgoff(a.addr, a.len, a.prot, a.flags, a.fd, a.offset);
prot              494 arch/s390/kvm/gaccess.c 		     u8 ar, enum gacc_mode mode, enum prot_type prot)
prot              505 arch/s390/kvm/gaccess.c 		switch (prot) {
prot              616 arch/s390/kvm/gaccess.c 				     enum gacc_mode mode, enum prot_type *prot)
prot              765 arch/s390/kvm/gaccess.c 		*prot = PROT_TYPE_DAT;
prot              769 arch/s390/kvm/gaccess.c 		*prot = PROT_TYPE_IEP;
prot              803 arch/s390/kvm/gaccess.c 	enum prot_type prot;
prot              813 arch/s390/kvm/gaccess.c 			rc = guest_translate(vcpu, ga, pages, asce, mode, &prot);
prot              822 arch/s390/kvm/gaccess.c 			return trans_exc(vcpu, rc, ga, ar, mode, prot);
prot              908 arch/s390/kvm/gaccess.c 	enum prot_type prot;
prot              923 arch/s390/kvm/gaccess.c 		rc = guest_translate(vcpu, gva, gpa, asce, mode, &prot);
prot              925 arch/s390/kvm/gaccess.c 			return trans_exc(vcpu, rc, gva, 0, mode, prot);
prot               73 arch/s390/mm/dump_pagetables.c 	unsigned int prot, cur;
prot               81 arch/s390/mm/dump_pagetables.c 	prot = new_prot;
prot               90 arch/s390/mm/dump_pagetables.c 	} else if (prot != cur || level != st->level ||
prot              117 arch/s390/mm/dump_pagetables.c 	unsigned int prot;
prot              119 arch/s390/mm/dump_pagetables.c 	prot = pte_val(*kasan_early_shadow_pte) &
prot              121 arch/s390/mm/dump_pagetables.c 	note_page(m, st, prot, 4);
prot              136 arch/s390/mm/dump_pagetables.c 	unsigned int prot;
prot              143 arch/s390/mm/dump_pagetables.c 		prot = pte_val(*pte) &
prot              145 arch/s390/mm/dump_pagetables.c 		note_page(m, st, prot, 4);
prot              153 arch/s390/mm/dump_pagetables.c 	unsigned int prot;
prot              169 arch/s390/mm/dump_pagetables.c 				prot = pmd_val(*pmd) &
prot              172 arch/s390/mm/dump_pagetables.c 				note_page(m, st, prot, 3);
prot              184 arch/s390/mm/dump_pagetables.c 	unsigned int prot;
prot              200 arch/s390/mm/dump_pagetables.c 				prot = pud_val(*pud) &
prot              203 arch/s390/mm/dump_pagetables.c 				note_page(m, st, prot, 2);
prot              874 arch/s390/mm/gmap.c 			     unsigned long vmaddr, int prot)
prot              881 arch/s390/mm/gmap.c 	fault_flags = (prot == PROT_WRITE) ? FAULT_FLAG_WRITE : 0;
prot              960 arch/s390/mm/gmap.c 			    pmd_t *pmdp, int prot, unsigned long bits)
prot              967 arch/s390/mm/gmap.c 	if ((pmd_i && (prot != PROT_NONE)) || (pmd_p && (prot == PROT_WRITE)))
prot              970 arch/s390/mm/gmap.c 	if (prot == PROT_NONE && !pmd_i) {
prot              975 arch/s390/mm/gmap.c 	if (prot == PROT_READ && !pmd_p) {
prot             1005 arch/s390/mm/gmap.c 			    pmd_t *pmdp, int prot, unsigned long bits)
prot             1022 arch/s390/mm/gmap.c 	rc = ptep_force_prot(gmap->mm, gaddr, ptep, prot, pbits);
prot             1041 arch/s390/mm/gmap.c 			      unsigned long len, int prot, unsigned long bits)
prot             1053 arch/s390/mm/gmap.c 				rc = gmap_protect_pte(gmap, gaddr, pmdp, prot,
prot             1060 arch/s390/mm/gmap.c 				rc = gmap_protect_pmd(gmap, gaddr, pmdp, prot,
prot             1078 arch/s390/mm/gmap.c 			rc = gmap_pte_op_fixup(gmap, gaddr, vmaddr, prot);
prot             1101 arch/s390/mm/gmap.c 			 unsigned long len, int prot)
prot             1107 arch/s390/mm/gmap.c 	if (!MACHINE_HAS_ESOP && prot == PROT_READ)
prot             1110 arch/s390/mm/gmap.c 	rc = gmap_protect_range(gmap, gaddr, len, prot, GMAP_NOTIFY_MPROT);
prot             2112 arch/s390/mm/gmap.c 	int prot;
prot             2117 arch/s390/mm/gmap.c 	prot = (pte_val(pte) & _PAGE_PROTECT) ? PROT_READ : PROT_WRITE;
prot             2159 arch/s390/mm/gmap.c 		rc = gmap_pte_op_fixup(parent, paddr, vmaddr, prot);
prot              112 arch/s390/mm/pageattr.c 	unsigned long pte_addr, prot;
prot              123 arch/s390/mm/pageattr.c 	prot = pgprot_val(ro ? PAGE_KERNEL_RO : PAGE_KERNEL);
prot              125 arch/s390/mm/pageattr.c 		prot &= ~_PAGE_NOEXEC;
prot              128 arch/s390/mm/pageattr.c 		pte_val(*ptep) = pte_addr | prot;
prot              189 arch/s390/mm/pageattr.c 	unsigned long pmd_addr, prot;
prot              200 arch/s390/mm/pageattr.c 	prot = pgprot_val(ro ? SEGMENT_KERNEL_RO : SEGMENT_KERNEL);
prot              202 arch/s390/mm/pageattr.c 		prot &= ~_SEGMENT_ENTRY_NOEXEC;
prot              205 arch/s390/mm/pageattr.c 		pmd_val(*pmdp) = pmd_addr | prot;
prot              596 arch/s390/mm/pgtable.c 		    pte_t *ptep, int prot, unsigned long bit)
prot              607 arch/s390/mm/pgtable.c 	if ((pte_i && (prot != PROT_NONE)) ||
prot              608 arch/s390/mm/pgtable.c 	    (pte_p && (prot & PROT_WRITE))) {
prot              614 arch/s390/mm/pgtable.c 	if (prot == PROT_NONE && !pte_i) {
prot              619 arch/s390/mm/pgtable.c 	if (prot == PROT_READ && !pte_p) {
prot               54 arch/sh/boards/mach-landisk/setup.c 	pgprot_t prot;
prot               60 arch/sh/boards/mach-landisk/setup.c 	prot = PAGE_KERNEL_PCC(1, _PAGE_PCC_IO16);
prot               61 arch/sh/boards/mach-landisk/setup.c 	cf_ide_base = ioremap_prot(paddrbase, PAGE_SIZE, pgprot_val(prot));
prot               49 arch/sh/boards/mach-lboxre2/setup.c 	pgprot_t prot;
prot               55 arch/sh/boards/mach-lboxre2/setup.c 	prot = PAGE_KERNEL_PCC(1, _PAGE_PCC_IO16);
prot               56 arch/sh/boards/mach-lboxre2/setup.c 	cf0_io_base = (u32)ioremap_prot(paddrbase, psize, pgprot_val(prot));
prot               71 arch/sh/boards/mach-sh03/setup.c 	pgprot_t prot;
prot               77 arch/sh/boards/mach-sh03/setup.c 	prot = PAGE_KERNEL_PCC(1, _PAGE_PCC_IO16);
prot               78 arch/sh/boards/mach-sh03/setup.c 	cf_ide_base = ioremap_prot(paddrbase, PAGE_SIZE, pgprot_val(prot));
prot              269 arch/sh/include/asm/io.h 			       pgprot_t prot, void *caller);
prot              273 arch/sh/include/asm/io.h __ioremap(phys_addr_t offset, unsigned long size, pgprot_t prot)
prot              275 arch/sh/include/asm/io.h 	return __ioremap_caller(offset, size, prot, __builtin_return_address(0));
prot              279 arch/sh/include/asm/io.h __ioremap_29bit(phys_addr_t offset, unsigned long size, pgprot_t prot)
prot              291 arch/sh/include/asm/io.h 		u64 flags = pgprot_val(prot);
prot              314 arch/sh/include/asm/io.h __ioremap_mode(phys_addr_t offset, unsigned long size, pgprot_t prot)
prot              322 arch/sh/include/asm/io.h 	ret = __ioremap_29bit(offset, size, prot);
prot              326 arch/sh/include/asm/io.h 	return __ioremap(offset, size, prot);
prot              329 arch/sh/include/asm/io.h #define __ioremap(offset, size, prot)		((void __iomem *)(offset))
prot              330 arch/sh/include/asm/io.h #define __ioremap_mode(offset, size, prot)	((void __iomem *)(offset))
prot              360 arch/sh/include/asm/io.h ioremap_fixed(phys_addr_t phys_addr, unsigned long size, pgprot_t prot)
prot               63 arch/sh/include/asm/mmu.h 		     unsigned long size, pgprot_t prot);
prot               65 arch/sh/include/asm/mmu.h 			       pgprot_t prot, void *caller);
prot               72 arch/sh/include/asm/mmu.h 		 unsigned long size, pgprot_t prot)
prot               79 arch/sh/include/asm/mmu.h 		 pgprot_t prot, void *caller)
prot              100 arch/sh/include/asm/mmu.h pmb_remap(phys_addr_t phys, unsigned long size, pgprot_t prot)
prot              102 arch/sh/include/asm/mmu.h 	return pmb_remap_caller(phys, size, prot, __builtin_return_address(0));
prot              147 arch/sh/include/asm/pgtable.h static inline bool __pte_access_permitted(pte_t pte, u64 prot)
prot              149 arch/sh/include/asm/pgtable.h 	return (pte_val(pte) & (prot | _PAGE_SPECIAL)) == prot;
prot              155 arch/sh/include/asm/pgtable.h 	u64 prot = _PAGE_PRESENT;
prot              157 arch/sh/include/asm/pgtable.h 	prot |= _PAGE_EXT(_PAGE_EXT_KERN_READ | _PAGE_EXT_USER_READ);
prot              159 arch/sh/include/asm/pgtable.h 		prot |= _PAGE_EXT(_PAGE_EXT_KERN_WRITE | _PAGE_EXT_USER_WRITE);
prot              160 arch/sh/include/asm/pgtable.h 	return __pte_access_permitted(pte, prot);
prot              165 arch/sh/include/asm/pgtable.h 	u64 prot = _PAGE_PRESENT | _PAGE_USER | _PAGE_READ;
prot              168 arch/sh/include/asm/pgtable.h 		prot |= _PAGE_WRITE;
prot              169 arch/sh/include/asm/pgtable.h 	return __pte_access_permitted(pte, prot);
prot              174 arch/sh/include/asm/pgtable.h 	u64 prot = _PAGE_PRESENT | _PAGE_USER;
prot              177 arch/sh/include/asm/pgtable.h 		prot |= _PAGE_RW;
prot              178 arch/sh/include/asm/pgtable.h 	return __pte_access_permitted(pte, prot);
prot              318 arch/sh/include/asm/pgtable_32.h #define pfn_pte(pfn, prot) \
prot              319 arch/sh/include/asm/pgtable_32.h 	__pte(((unsigned long long)(pfn) << PAGE_SHIFT) | pgprot_val(prot))
prot              320 arch/sh/include/asm/pgtable_32.h #define pfn_pmd(pfn, prot) \
prot              321 arch/sh/include/asm/pgtable_32.h 	__pmd(((unsigned long long)(pfn) << PAGE_SHIFT) | pgprot_val(prot))
prot              379 arch/sh/include/asm/pgtable_32.h #define pgprot_writecombine(prot) \
prot              380 arch/sh/include/asm/pgtable_32.h 	__pgprot(pgprot_val(prot) & ~_PAGE_CACHABLE)
prot              204 arch/sh/include/asm/pgtable_64.h #define pgprot_writecombine(prot) __pgprot(pgprot_val(prot) & ~_PAGE_CACHABLE)
prot              304 arch/sh/include/asm/pgtable_64.h #define pfn_pte(pfn, prot)	__pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
prot              305 arch/sh/include/asm/pgtable_64.h #define pfn_pmd(pfn, prot)	__pmd(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
prot                8 arch/sh/include/asm/syscalls.h 			unsigned long prot, unsigned long flags,
prot               11 arch/sh/include/asm/syscalls.h 			  unsigned long prot, unsigned long flags,
prot               29 arch/sh/include/cpu-sh4/cpu/sq.h 		       const char *name, pgprot_t prot);
prot              101 arch/sh/kernel/cpu/sh4/sq.c static int __sq_remap(struct sq_mapping *map, pgprot_t prot)
prot              114 arch/sh/kernel/cpu/sh4/sq.c 			       vma->phys_addr, prot)) {
prot              143 arch/sh/kernel/cpu/sh4/sq.c 		       const char *name, pgprot_t prot)
prot              178 arch/sh/kernel/cpu/sh4/sq.c 	ret = __sq_remap(map, prot);
prot               33 arch/sh/kernel/sys_sh.c 	unsigned long prot, unsigned long flags,
prot               38 arch/sh/kernel/sys_sh.c 	return ksys_mmap_pgoff(addr, len, prot, flags, fd, off>>PAGE_SHIFT);
prot               42 arch/sh/kernel/sys_sh.c 	unsigned long prot, unsigned long flags,
prot               54 arch/sh/kernel/sys_sh.c 	return ksys_mmap_pgoff(addr, len, prot, flags, fd, pgoff);
prot               72 arch/sh/mm/init.c static void set_pte_phys(unsigned long addr, unsigned long phys, pgprot_t prot)
prot               82 arch/sh/mm/init.c 	set_pte(pte, pfn_pte(phys >> PAGE_SHIFT, prot));
prot               85 arch/sh/mm/init.c 	if (pgprot_val(prot) & _PAGE_WIRED)
prot               89 arch/sh/mm/init.c static void clear_pte_phys(unsigned long addr, pgprot_t prot)
prot               95 arch/sh/mm/init.c 	if (pgprot_val(prot) & _PAGE_WIRED)
prot              102 arch/sh/mm/init.c void __set_fixmap(enum fixed_addresses idx, unsigned long phys, pgprot_t prot)
prot              111 arch/sh/mm/init.c 	set_pte_phys(address, phys, prot);
prot              114 arch/sh/mm/init.c void __clear_fixmap(enum fixed_addresses idx, pgprot_t prot)
prot              123 arch/sh/mm/init.c 	clear_pte_phys(address, prot);
prot               48 arch/sh/mm/ioremap_fixed.c ioremap_fixed(phys_addr_t phys_addr, unsigned long size, pgprot_t prot)
prot               89 arch/sh/mm/ioremap_fixed.c 		pgprot_val(prot) |= _PAGE_WIRED;
prot               90 arch/sh/mm/ioremap_fixed.c 		__set_fixmap(idx, phys_addr, prot);
prot               97 arch/sh/mm/nommu.c void __set_fixmap(enum fixed_addresses idx, unsigned long phys, pgprot_t prot)
prot              117 arch/sh/mm/pmb.c static inline unsigned long pgprot_to_pmb_flags(pgprot_t prot)
prot              120 arch/sh/mm/pmb.c 	u64 flags = pgprot_val(prot);
prot              207 arch/sh/mm/pmb.c static inline bool pmb_prot_valid(pgprot_t prot)
prot              209 arch/sh/mm/pmb.c 	return (pgprot_val(prot) & _PAGE_USER) == 0;
prot              336 arch/sh/mm/pmb.c 		     unsigned long size, pgprot_t prot)
prot              355 arch/sh/mm/pmb.c 	pmb_flags = pgprot_to_pmb_flags(prot);
prot              411 arch/sh/mm/pmb.c 			       pgprot_t prot, void *caller)
prot              428 arch/sh/mm/pmb.c 	if (!pmb_prot_valid(prot))
prot              454 arch/sh/mm/pmb.c 	ret = pmb_bolt_mapping(vaddr, phys, size, prot);
prot               30 arch/sparc/include/asm/mman.h #define arch_calc_vm_prot_bits(prot, pkey) sparc_calc_vm_prot_bits(prot)
prot               31 arch/sparc/include/asm/mman.h static inline unsigned long sparc_calc_vm_prot_bits(unsigned long prot)
prot               33 arch/sparc/include/asm/mman.h 	if (adi_capable() && (prot & PROT_ADI)) {
prot               55 arch/sparc/include/asm/mman.h #define arch_validate_prot(prot, addr) sparc_validate_prot(prot, addr)
prot               56 arch/sparc/include/asm/mman.h static inline int sparc_validate_prot(unsigned long prot, unsigned long addr)
prot               58 arch/sparc/include/asm/mman.h 	if (prot & ~(PROT_READ | PROT_WRITE | PROT_EXEC | PROT_SEM | PROT_ADI))
prot               60 arch/sparc/include/asm/mman.h 	if (prot & PROT_ADI) {
prot              263 arch/sparc/include/asm/pgtable_32.h #define pfn_pte(pfn, prot)		mk_pte(pfn_to_page(pfn), prot)
prot              300 arch/sparc/include/asm/pgtable_32.h static inline pgprot_t pgprot_noncached(pgprot_t prot)
prot              302 arch/sparc/include/asm/pgtable_32.h 	pgprot_val(prot) &= ~pgprot_val(__pgprot(SRMMU_CACHE));
prot              303 arch/sparc/include/asm/pgtable_32.h 	return prot;
prot              415 arch/sparc/include/asm/pgtable_32.h 				     unsigned long size, pgprot_t prot)
prot              423 arch/sparc/include/asm/pgtable_32.h 	return remap_pfn_range(vma, from, phys_base >> PAGE_SHIFT, size, prot);
prot              239 arch/sparc/include/asm/pgtable_64.h static inline pte_t pfn_pte(unsigned long pfn, pgprot_t prot)
prot              244 arch/sparc/include/asm/pgtable_64.h 	return __pte(paddr | pgprot_val(prot));
prot              280 arch/sparc/include/asm/pgtable_64.h static inline pte_t pte_modify(pte_t pte, pgprot_t prot)
prot              335 arch/sparc/include/asm/pgtable_64.h 	return __pte((pte_val(pte) & mask) | (pgprot_val(prot) & ~mask));
prot              349 arch/sparc/include/asm/pgtable_64.h static inline pgprot_t pgprot_noncached(pgprot_t prot)
prot              351 arch/sparc/include/asm/pgtable_64.h 	unsigned long val = pgprot_val(prot);
prot              968 arch/sparc/include/asm/pgtable_64.h #define move_pte(pte, prot, old_addr, new_addr)				\
prot             1069 arch/sparc/include/asm/pgtable_64.h 				     unsigned long size, pgprot_t prot)
prot             1077 arch/sparc/include/asm/pgtable_64.h 	return remap_pfn_range(vma, from, phys_base >> PAGE_SHIFT, size, prot);
prot             1106 arch/sparc/include/asm/pgtable_64.h 	u64 prot;
prot             1109 arch/sparc/include/asm/pgtable_64.h 		prot = _PAGE_PRESENT_4V | _PAGE_P_4V;
prot             1111 arch/sparc/include/asm/pgtable_64.h 			prot |= _PAGE_WRITE_4V;
prot             1113 arch/sparc/include/asm/pgtable_64.h 		prot = _PAGE_PRESENT_4U | _PAGE_P_4U;
prot             1115 arch/sparc/include/asm/pgtable_64.h 			prot |= _PAGE_WRITE_4U;
prot             1118 arch/sparc/include/asm/pgtable_64.h 	return (pte_val(pte) & (prot | _PAGE_SPECIAL)) == prot;
prot              437 arch/sparc/kernel/iommu.c 	unsigned long flags, handle, prot, ctx;
prot              460 arch/sparc/kernel/iommu.c 		prot = IOPTE_STREAMING(ctx);
prot              462 arch/sparc/kernel/iommu.c 		prot = IOPTE_CONSISTENT(ctx);
prot              464 arch/sparc/kernel/iommu.c 		prot |= IOPTE_WRITE;
prot              512 arch/sparc/kernel/iommu.c 			iopte_val(*base) = prot | paddr;
prot               56 arch/sparc/kernel/pci_sun4v.c 	unsigned long	prot;		/* IOMMU page protections	*/
prot               66 arch/sparc/kernel/pci_sun4v.c static inline void iommu_batch_start(struct device *dev, unsigned long prot, unsigned long entry)
prot               71 arch/sparc/kernel/pci_sun4v.c 	p->prot		= prot;
prot               88 arch/sparc/kernel/pci_sun4v.c 	unsigned long prot = p->prot;
prot               97 arch/sparc/kernel/pci_sun4v.c 		prot &= (HV_PCI_MAP_ATTR_READ | HV_PCI_MAP_ATTR_WRITE);
prot              104 arch/sparc/kernel/pci_sun4v.c 						  prot,
prot              111 arch/sparc/kernel/pci_sun4v.c 						   npages, prot, __pa(pglist),
prot              121 arch/sparc/kernel/pci_sun4v.c 						  prot,
prot              128 arch/sparc/kernel/pci_sun4v.c 						   index_count, prot,
prot              185 arch/sparc/kernel/pci_sun4v.c 	unsigned long prot = 0;
prot              201 arch/sparc/kernel/pci_sun4v.c 		prot = HV_PCI_MAP_ATTR_RELAXED_ORDER;
prot              231 arch/sparc/kernel/pci_sun4v.c 			  (HV_PCI_MAP_ATTR_READ | prot |
prot              364 arch/sparc/kernel/pci_sun4v.c 	unsigned long prot;
prot              393 arch/sparc/kernel/pci_sun4v.c 	prot = HV_PCI_MAP_ATTR_READ;
prot              395 arch/sparc/kernel/pci_sun4v.c 		prot |= HV_PCI_MAP_ATTR_WRITE;
prot              398 arch/sparc/kernel/pci_sun4v.c 		prot |= HV_PCI_MAP_ATTR_RELAXED_ORDER;
prot              402 arch/sparc/kernel/pci_sun4v.c 	iommu_batch_start(dev, prot, entry);
prot              472 arch/sparc/kernel/pci_sun4v.c 	unsigned long flags, handle, prot;
prot              491 arch/sparc/kernel/pci_sun4v.c 	prot = HV_PCI_MAP_ATTR_READ;
prot              493 arch/sparc/kernel/pci_sun4v.c 		prot |= HV_PCI_MAP_ATTR_WRITE;
prot              496 arch/sparc/kernel/pci_sun4v.c 		prot |= HV_PCI_MAP_ATTR_RELAXED_ORDER;
prot              508 arch/sparc/kernel/pci_sun4v.c 	iommu_batch_start(dev, prot, ~0UL);
prot              102 arch/sparc/kernel/sys_sparc_32.c 	unsigned long, prot, unsigned long, flags, unsigned long, fd,
prot              107 arch/sparc/kernel/sys_sparc_32.c 	return ksys_mmap_pgoff(addr, len, prot, flags, fd,
prot              112 arch/sparc/kernel/sys_sparc_32.c 	unsigned long, prot, unsigned long, flags, unsigned long, fd,
prot              116 arch/sparc/kernel/sys_sparc_32.c 	return ksys_mmap_pgoff(addr, len, prot, flags, fd, off >> PAGE_SHIFT);
prot              120 arch/sparc/kernel/sys_sparc_32.c 			   unsigned long, prot, unsigned long, pgoff,
prot              126 arch/sparc/kernel/sys_sparc_32.c 	return sys_remap_file_pages(start, size, prot,
prot              456 arch/sparc/kernel/sys_sparc_64.c 		unsigned long, prot, unsigned long, flags, unsigned long, fd,
prot              465 arch/sparc/kernel/sys_sparc_64.c 	retval = ksys_mmap_pgoff(addr, len, prot, flags, fd, off >> PAGE_SHIFT);
prot               18 arch/sparc/kernel/systbls.h 			 unsigned long prot, unsigned long flags,
prot               24 arch/sparc/kernel/systbls.h 			  unsigned long prot, unsigned long flags,
prot               27 arch/sparc/kernel/systbls.h 			    unsigned long prot, unsigned long pgoff,
prot             1787 arch/sparc/mm/init_64.c 					    unsigned long pend, pgprot_t prot,
prot             1855 arch/sparc/mm/init_64.c 			pte_val(*pte) = (paddr | pgprot_val(prot));
prot             2805 arch/sparc/mm/init_64.c pte_t mk_pte_io(unsigned long page, pgprot_t prot, int space, unsigned long page_size)
prot             2809 arch/sparc/mm/init_64.c 	pte_val(pte)  = page | pgprot_val(pgprot_noncached(prot));
prot              159 arch/sparc/vdso/vdso2c.c static void map_input(const char *name, void **addr, size_t *len, int prot)
prot              173 arch/sparc/vdso/vdso2c.c 	*addr = mmap(NULL, tmp_len, prot, MAP_PRIVATE, fd, 0);
prot               47 arch/um/include/asm/page.h #define pte_set_val(p, phys, prot) \
prot               48 arch/um/include/asm/page.h 	({ (p).pte = (phys) | pgprot_val(prot); })
prot               74 arch/um/include/asm/page.h #define pte_set_val(p, phys, prot) (p).pte = (phys | pgprot_val(prot))
prot               42 arch/um/include/asm/pgtable-2level.h #define pfn_pte(pfn, prot) __pte(pfn_to_phys(pfn) | pgprot_val(prot))
prot               43 arch/um/include/asm/pgtable-2level.h #define pfn_pmd(pfn, prot) __pmd(pfn_to_phys(pfn) | pgprot_val(prot))
prot              275 arch/um/include/shared/os.h 	       unsigned long len, int prot, int phys_fd,
prot              280 arch/um/include/shared/os.h 		   unsigned long len, unsigned int prot, int done, void **data);
prot               18 arch/um/kernel/syscall.c 	      unsigned long prot, unsigned long flags,
prot               25 arch/um/kernel/syscall.c 	err = ksys_mmap_pgoff(addr, len, prot, flags, fd, offset >> PAGE_SHIFT);
prot               25 arch/um/kernel/tlb.c 				unsigned int prot;
prot               36 arch/um/kernel/tlb.c 				unsigned int prot;
prot               75 arch/um/kernel/tlb.c 					  op->u.mmap.len, op->u.mmap.prot,
prot              100 arch/um/kernel/tlb.c 					      op->u.mprotect.prot,
prot              123 arch/um/kernel/tlb.c 		    unsigned int prot, struct host_vm_change *hvc)
prot              137 arch/um/kernel/tlb.c 		   (last->u.mmap.prot == prot) && (last->u.mmap.fd == fd) &&
prot              153 arch/um/kernel/tlb.c 						     .prot	= prot,
prot              191 arch/um/kernel/tlb.c 			unsigned int prot, struct host_vm_change *hvc)
prot              200 arch/um/kernel/tlb.c 		   (last->u.mprotect.prot == prot)) {
prot              215 arch/um/kernel/tlb.c 							 .prot	= prot } } });
prot              226 arch/um/kernel/tlb.c 	int r, w, x, prot, ret = 0;
prot              242 arch/um/kernel/tlb.c 		prot = ((r ? UM_PROT_READ : 0) | (w ? UM_PROT_WRITE : 0) |
prot              248 arch/um/kernel/tlb.c 						       PAGE_SIZE, prot, hvc);
prot              252 arch/um/kernel/tlb.c 			ret = add_mprotect(addr, PAGE_SIZE, prot, hvc);
prot              432 arch/um/kernel/tlb.c 	int r, w, x, prot, err = 0;
prot              461 arch/um/kernel/tlb.c 	prot = ((r ? UM_PROT_READ : 0) | (w ? UM_PROT_WRITE : 0) |
prot              469 arch/um/kernel/tlb.c 			err = map(mm_id, address, PAGE_SIZE, prot, fd, offset,
prot              475 arch/um/kernel/tlb.c 		err = protect(mm_id, address, PAGE_SIZE, prot, 1, &flush);
prot              142 arch/um/os-Linux/process.c 	int prot;
prot              144 arch/um/os-Linux/process.c 	prot = (r ? PROT_READ : 0) | (w ? PROT_WRITE : 0) |
prot              147 arch/um/os-Linux/process.c 	loc = mmap64((void *) virt, len, prot, MAP_SHARED | MAP_FIXED,
prot              156 arch/um/os-Linux/process.c 	int prot = ((r ? PROT_READ : 0) | (w ? PROT_WRITE : 0) |
prot              159 arch/um/os-Linux/process.c 	if (mprotect(addr, len, prot) < 0)
prot              170 arch/um/os-Linux/skas/mem.c int map(struct mm_id * mm_idp, unsigned long virt, unsigned long len, int prot,
prot              174 arch/um/os-Linux/skas/mem.c 	unsigned long args[] = { virt, len, prot,
prot              198 arch/um/os-Linux/skas/mem.c 	    unsigned int prot, int done, void **data)
prot              201 arch/um/os-Linux/skas/mem.c 	unsigned long args[] = { addr, len, prot, 0, 0, 0 };
prot              151 arch/unicore32/include/asm/pgtable.h #define pfn_pte(pfn, prot)		(__pte(((pfn) << PAGE_SHIFT) \
prot              152 arch/unicore32/include/asm/pgtable.h 						| pgprot_val(prot)))
prot              197 arch/unicore32/include/asm/pgtable.h #define pgprot_noncached(prot)		\
prot              198 arch/unicore32/include/asm/pgtable.h 	__pgprot(pgprot_val(prot) & ~PTE_CACHEABLE)
prot              199 arch/unicore32/include/asm/pgtable.h #define pgprot_writecombine(prot)	\
prot              200 arch/unicore32/include/asm/pgtable.h 	__pgprot(pgprot_val(prot) & ~PTE_CACHEABLE)
prot              226 arch/unicore32/include/asm/pgtable.h #define mk_pte(page, prot)	pfn_pte(page_to_pfn(page), prot)
prot              142 arch/unicore32/mm/mmu.c 		unsigned long prot)
prot              152 arch/unicore32/mm/mmu.c 		__pmd_populate(pmd, __pa(pte) | prot);
prot              181 arch/x86/entry/vdso/vdso2c.c static void map_input(const char *name, void **addr, size_t *len, int prot)
prot              194 arch/x86/entry/vdso/vdso2c.c 	*addr = mmap(NULL, tmp_len, prot, MAP_PRIVATE, fd, 0);
prot              285 arch/x86/events/intel/ds.c static void ds_update_cea(void *cea, void *addr, size_t size, pgprot_t prot)
prot              295 arch/x86/events/intel/ds.c 		cea_set_pte(cea, pa, prot);
prot              153 arch/x86/ia32/sys_ia32.c 	unsigned int prot;
prot              169 arch/x86/ia32/sys_ia32.c 	return ksys_mmap_pgoff(a.addr, a.len, a.prot, a.flags, a.fd,
prot               12 arch/x86/include/asm/fb.h 	unsigned long prot;
prot               14 arch/x86/include/asm/fb.h 	prot = pgprot_val(vma->vm_page_prot) & ~_PAGE_CACHE_MASK;
prot               17 arch/x86/include/asm/fb.h 			prot | cachemode2protval(_PAGE_CACHE_MODE_UC_MINUS);
prot               67 arch/x86/include/asm/highmem.h void *kmap_atomic_prot(struct page *page, pgprot_t prot);
prot               71 arch/x86/include/asm/highmem.h void *kmap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot);
prot               17 arch/x86/include/asm/iomap.h iomap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot);
prot               23 arch/x86/include/asm/iomap.h iomap_create_wc(resource_size_t base, unsigned long size, pgprot_t *prot);
prot               12 arch/x86/include/asm/pgtable.h #define pgprot_noncached(prot)						\
prot               14 arch/x86/include/asm/pgtable.h 	 ? (__pgprot(pgprot_val(prot) |					\
prot               16 arch/x86/include/asm/pgtable.h 	 : (prot))
prot               21 arch/x86/include/asm/pgtable.h #define pgprot_encrypted(prot)	__pgprot(__sme_set(pgprot_val(prot)))
prot               22 arch/x86/include/asm/pgtable.h #define pgprot_decrypted(prot)	__pgprot(__sme_clr(pgprot_val(prot)))
prot              647 arch/x86/include/asm/pgtable.h static inline pgprot_t arch_filter_pgprot(pgprot_t prot)
prot              649 arch/x86/include/asm/pgtable.h 	return canon_pgprot(prot);
prot             1463 arch/x86/include/asm/pgtable.h extern bool pfn_modify_allowed(unsigned long pfn, pgprot_t prot);
prot              515 arch/x86/include/asm/pgtable_types.h extern pgprot_t pgprot_writecombine(pgprot_t prot);
prot              518 arch/x86/include/asm/pgtable_types.h extern pgprot_t pgprot_writethrough(pgprot_t prot);
prot               31 arch/x86/include/asm/pkeys.h 		int prot, int pkey);
prot               33 arch/x86/include/asm/pkeys.h 		int prot, int pkey)
prot               38 arch/x86/include/asm/pkeys.h 	return __arch_override_mprotect_pkey(vma, prot, pkey);
prot               22 arch/x86/include/uapi/asm/mman.h #define arch_calc_vm_prot_bits(prot, key) (		\
prot              126 arch/x86/kernel/machine_kexec_64.c 	pgprot_t prot = PAGE_KERNEL_EXEC_NOENC;
prot              171 arch/x86/kernel/machine_kexec_64.c 		prot = PAGE_KERNEL_EXEC;
prot              173 arch/x86/kernel/machine_kexec_64.c 	set_pte(pte, pfn_pte(paddr >> PAGE_SHIFT, prot));
prot               92 arch/x86/kernel/sys_x86_64.c 		unsigned long, prot, unsigned long, flags,
prot              100 arch/x86/kernel/sys_x86_64.c 	error = ksys_mmap_pgoff(addr, len, prot, flags, fd, off >> PAGE_SHIFT);
prot              105 arch/x86/kernel/tboot.c 			  pgprot_t prot)
prot              126 arch/x86/kernel/tboot.c 	set_pte_at(&tboot_mm, vaddr, pte, pfn_pte(pfn, prot));
prot               49 arch/x86/mm/cpu_entry_area.c cea_map_percpu_pages(void *cea_vaddr, void *ptr, int pages, pgprot_t prot)
prot               52 arch/x86/mm/cpu_entry_area.c 		cea_set_pte(cea_vaddr, per_cpu_ptr_to_phys(ptr), prot);
prot              177 arch/x86/mm/dump_pagetables.c static void printk_prot(struct seq_file *m, pgprot_t prot, int level, bool dmsg)
prot              179 arch/x86/mm/dump_pagetables.c 	pgprotval_t pr = pgprot_val(prot);
prot              271 arch/x86/mm/dump_pagetables.c 	pgprotval_t prot, cur, eff;
prot              279 arch/x86/mm/dump_pagetables.c 	prot = pgprot_val(new_prot);
prot              292 arch/x86/mm/dump_pagetables.c 	} else if (prot != cur || new_eff != eff || level != st->level ||
prot              362 arch/x86/mm/dump_pagetables.c 	pgprotval_t prot, eff;
prot              367 arch/x86/mm/dump_pagetables.c 		prot = pte_flags(*pte);
prot              368 arch/x86/mm/dump_pagetables.c 		eff = effective_prot(eff_in, prot);
prot              369 arch/x86/mm/dump_pagetables.c 		note_page(m, st, __pgprot(prot), eff, 5);
prot              389 arch/x86/mm/dump_pagetables.c 		pgprotval_t prot = pte_flags(kasan_early_shadow_pte[0]);
prot              390 arch/x86/mm/dump_pagetables.c 		note_page(m, st, __pgprot(prot), 0, 5);
prot              410 arch/x86/mm/dump_pagetables.c 	pgprotval_t prot, eff;
prot              416 arch/x86/mm/dump_pagetables.c 			prot = pmd_flags(*start);
prot              417 arch/x86/mm/dump_pagetables.c 			eff = effective_prot(eff_in, prot);
prot              419 arch/x86/mm/dump_pagetables.c 				note_page(m, st, __pgprot(prot), eff, 4);
prot              443 arch/x86/mm/dump_pagetables.c 	pgprotval_t prot, eff;
prot              450 arch/x86/mm/dump_pagetables.c 			prot = pud_flags(*start);
prot              451 arch/x86/mm/dump_pagetables.c 			eff = effective_prot(eff_in, prot);
prot              453 arch/x86/mm/dump_pagetables.c 				note_page(m, st, __pgprot(prot), eff, 3);
prot              476 arch/x86/mm/dump_pagetables.c 	pgprotval_t prot, eff;
prot              486 arch/x86/mm/dump_pagetables.c 			prot = p4d_flags(*start);
prot              487 arch/x86/mm/dump_pagetables.c 			eff = effective_prot(eff_in, prot);
prot              489 arch/x86/mm/dump_pagetables.c 				note_page(m, st, __pgprot(prot), eff, 2);
prot              522 arch/x86/mm/dump_pagetables.c 	pgprotval_t prot, eff;
prot              538 arch/x86/mm/dump_pagetables.c 			prot = pgd_flags(*start);
prot              542 arch/x86/mm/dump_pagetables.c 			eff = prot;
prot              545 arch/x86/mm/dump_pagetables.c 				note_page(m, &st, __pgprot(prot), eff, 1);
prot               34 arch/x86/mm/highmem_32.c void *kmap_atomic_prot(struct page *page, pgprot_t prot)
prot               49 arch/x86/mm/highmem_32.c 	set_pte(kmap_pte-idx, mk_pte(page, prot));
prot              317 arch/x86/mm/init_32.c 				pgprot_t prot = PAGE_KERNEL_LARGE;
prot              332 arch/x86/mm/init_32.c 					prot = PAGE_KERNEL_LARGE_EXEC;
prot              338 arch/x86/mm/init_32.c 					set_pmd(pmd, pfn_pmd(pfn, prot));
prot              349 arch/x86/mm/init_32.c 				pgprot_t prot = PAGE_KERNEL;
prot              357 arch/x86/mm/init_32.c 					prot = PAGE_KERNEL_EXEC;
prot              364 arch/x86/mm/init_32.c 					set_pte(pte, pfn_pte(pfn, prot));
prot              367 arch/x86/mm/init_64.c 	pgprot_t prot;
prot              369 arch/x86/mm/init_64.c 	pgprot_val(prot) = pgprot_val(PAGE_KERNEL_LARGE) |
prot              393 arch/x86/mm/init_64.c 		set_pmd(pmd, __pmd(phys | pgprot_val(prot)));
prot              449 arch/x86/mm/init_64.c 	      pgprot_t prot, bool init)
prot              487 arch/x86/mm/init_64.c 		set_pte_init(pte, pfn_pte(paddr >> PAGE_SHIFT, prot), init);
prot              503 arch/x86/mm/init_64.c 	      unsigned long page_size_mask, pgprot_t prot, bool init)
prot              513 arch/x86/mm/init_64.c 		pgprot_t new_prot = prot;
prot              531 arch/x86/mm/init_64.c 							   paddr_end, prot,
prot              562 arch/x86/mm/init_64.c 					     __pgprot(pgprot_val(prot) | _PAGE_PSE)),
prot              598 arch/x86/mm/init_64.c 		pgprot_t prot = PAGE_KERNEL;
prot              620 arch/x86/mm/init_64.c 							   prot, init);
prot              641 arch/x86/mm/init_64.c 			prot = pte_pgprot(pte_clrhuge(*(pte_t *)pud));
prot              658 arch/x86/mm/init_64.c 					   page_size_mask, prot, init);
prot               21 arch/x86/mm/iomap_32.c int iomap_create_wc(resource_size_t base, unsigned long size, pgprot_t *prot)
prot               33 arch/x86/mm/iomap_32.c 	*prot = __pgprot(__PAGE_KERNEL | cachemode2protval(pcm));
prot               35 arch/x86/mm/iomap_32.c 	pgprot_val(*prot) &= __default_kernel_pte_mask;
prot               47 arch/x86/mm/iomap_32.c void *kmap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot)
prot               58 arch/x86/mm/iomap_32.c 	set_pte(kmap_pte - idx, pfn_pte(pfn, prot));
prot               68 arch/x86/mm/iomap_32.c iomap_atomic_prot_pfn(unsigned long pfn, pgprot_t prot)
prot               77 arch/x86/mm/iomap_32.c 	if (!pat_enabled() && pgprot2cachemode(prot) != _PAGE_CACHE_MODE_WB)
prot               78 arch/x86/mm/iomap_32.c 		prot = __pgprot(__PAGE_KERNEL |
prot               82 arch/x86/mm/iomap_32.c 	pgprot_val(prot) &= __default_kernel_pte_mask;
prot               84 arch/x86/mm/iomap_32.c 	return (void __force __iomem *) kmap_atomic_prot_pfn(pfn, prot);
prot              186 arch/x86/mm/ioremap.c 	pgprot_t prot;
prot              244 arch/x86/mm/ioremap.c 	prot = PAGE_KERNEL_IO;
prot              246 arch/x86/mm/ioremap.c 		prot = pgprot_encrypted(prot);
prot              251 arch/x86/mm/ioremap.c 		prot = __pgprot(pgprot_val(prot) |
prot              255 arch/x86/mm/ioremap.c 		prot = __pgprot(pgprot_val(prot) |
prot              259 arch/x86/mm/ioremap.c 		prot = __pgprot(pgprot_val(prot) |
prot              263 arch/x86/mm/ioremap.c 		prot = __pgprot(pgprot_val(prot) |
prot              282 arch/x86/mm/ioremap.c 	if (ioremap_page_range(vaddr, vaddr + size, phys_addr, prot))
prot              728 arch/x86/mm/ioremap.c 					     pgprot_t prot)
prot              733 arch/x86/mm/ioremap.c 		return prot;
prot              746 arch/x86/mm/ioremap.c 	return encrypted_prot ? pgprot_encrypted(prot)
prot              747 arch/x86/mm/ioremap.c 			      : pgprot_decrypted(prot);
prot              383 arch/x86/mm/kasan_init_64.c 		pgprot_t prot;
prot              385 arch/x86/mm/kasan_init_64.c 		prot = __pgprot(__PAGE_KERNEL_RO | _PAGE_ENC);
prot              386 arch/x86/mm/kasan_init_64.c 		pgprot_val(prot) &= __default_kernel_pte_mask;
prot              388 arch/x86/mm/kasan_init_64.c 		pte = __pte(__pa(kasan_early_shadow_page) | pgprot_val(prot));
prot              238 arch/x86/mm/mmap.c bool pfn_modify_allowed(unsigned long pfn, pgprot_t prot)
prot              242 arch/x86/mm/mmap.c 	if (!__pte_needs_invert(pgprot_val(prot)))
prot              489 arch/x86/mm/pageattr.c static inline bool conflicts(pgprot_t prot, pgprotval_t val)
prot              491 arch/x86/mm/pageattr.c 	return (pgprot_val(prot) & ~val) != pgprot_val(prot);
prot              494 arch/x86/mm/pageattr.c static inline void check_conflict(int warnlvl, pgprot_t prot, pgprotval_t val,
prot              504 arch/x86/mm/pageattr.c 	if (warnlvl > cpa_warn_level || !conflicts(prot, val))
prot              508 arch/x86/mm/pageattr.c 		lvltxt[warnlvl], txt, start, end, pfn, (unsigned long long)pgprot_val(prot),
prot              518 arch/x86/mm/pageattr.c static inline pgprot_t static_protections(pgprot_t prot, unsigned long start,
prot              529 arch/x86/mm/pageattr.c 	if (!(pgprot_val(prot) & _PAGE_PRESENT))
prot              530 arch/x86/mm/pageattr.c 		return prot;
prot              536 arch/x86/mm/pageattr.c 	check_conflict(warnlvl, prot, res, start, end, pfn, "Text NX");
prot              547 arch/x86/mm/pageattr.c 		check_conflict(warnlvl, prot, res, start, end, pfn, "Text RO");
prot              553 arch/x86/mm/pageattr.c 	check_conflict(warnlvl, prot, res, start, end, pfn, "PCIBIOS NX");
prot              557 arch/x86/mm/pageattr.c 	check_conflict(warnlvl, prot, res, start, end, pfn, "Rodata RO");
prot              560 arch/x86/mm/pageattr.c 	return __pgprot(pgprot_val(prot) & ~forbidden);
prot              729 arch/x86/mm/pageattr.c static pgprot_t pgprot_clear_protnone_bits(pgprot_t prot)
prot              740 arch/x86/mm/pageattr.c 	if (!(pgprot_val(prot) & _PAGE_PRESENT))
prot              741 arch/x86/mm/pageattr.c 		pgprot_val(prot) &= ~_PAGE_GLOBAL;
prot              743 arch/x86/mm/pageattr.c 	return prot;
prot              909 arch/x86/mm/pageattr.c 	pgprot_t prot;
prot              919 arch/x86/mm/pageattr.c 	prot = static_protections(ref_prot, address, pfn, npg, 0, CPA_PROTECT);
prot              921 arch/x86/mm/pageattr.c 	if (pgprot_val(prot) == pgprot_val(ref_prot))
prot              933 arch/x86/mm/pageattr.c 		ref_prot = prot;
prot              968 arch/x86/mm/pat.c 	unsigned long prot;
prot              977 arch/x86/mm/pat.c 		if (follow_phys(vma, vma->vm_start, 0, &prot, &paddr)) {
prot              981 arch/x86/mm/pat.c 		pgprot = __pgprot(prot);
prot              994 arch/x86/mm/pat.c int track_pfn_remap(struct vm_area_struct *vma, pgprot_t *prot,
prot             1005 arch/x86/mm/pat.c 		ret = reserve_pfn_range(paddr, size, prot, 0);
prot             1028 arch/x86/mm/pat.c 	*prot = __pgprot((pgprot_val(*prot) & (~_PAGE_CACHE_MASK)) |
prot             1034 arch/x86/mm/pat.c void track_pfn_insert(struct vm_area_struct *vma, pgprot_t *prot, pfn_t pfn)
prot             1043 arch/x86/mm/pat.c 	*prot = __pgprot((pgprot_val(*prot) & (~_PAGE_CACHE_MASK)) |
prot             1056 arch/x86/mm/pat.c 	unsigned long prot;
prot             1064 arch/x86/mm/pat.c 		if (follow_phys(vma, vma->vm_start, 0, &prot, &paddr)) {
prot             1086 arch/x86/mm/pat.c pgprot_t pgprot_writecombine(pgprot_t prot)
prot             1088 arch/x86/mm/pat.c 	return __pgprot(pgprot_val(prot) |
prot             1093 arch/x86/mm/pat.c pgprot_t pgprot_writethrough(pgprot_t prot)
prot             1095 arch/x86/mm/pat.c 	return __pgprot(pgprot_val(prot) |
prot              662 arch/x86/mm/pgtable.c int p4d_set_huge(p4d_t *p4d, phys_addr_t addr, pgprot_t prot)
prot              696 arch/x86/mm/pgtable.c int pud_set_huge(pud_t *pud, phys_addr_t addr, pgprot_t prot)
prot              709 arch/x86/mm/pgtable.c 	prot = pgprot_4k_2_large(prot);
prot              713 arch/x86/mm/pgtable.c 		__pgprot(pgprot_val(prot) | _PAGE_PSE)));
prot              725 arch/x86/mm/pgtable.c int pmd_set_huge(pmd_t *pmd, phys_addr_t addr, pgprot_t prot)
prot              741 arch/x86/mm/pgtable.c 	prot = pgprot_4k_2_large(prot);
prot              745 arch/x86/mm/pgtable.c 		__pgprot(pgprot_val(prot) | _PAGE_PSE)));
prot               77 arch/x86/mm/pkeys.c int __arch_override_mprotect_pkey(struct vm_area_struct *vma, int prot, int pkey)
prot               92 arch/x86/mm/pkeys.c 	if (prot == PROT_EXEC) {
prot               27 arch/x86/um/shared/sysdep/syscalls_64.h 		     unsigned long prot, unsigned long flags,
prot              348 arch/x86/xen/enlighten_pv.c static void set_aliased_prot(void *v, pgprot_t prot)
prot              363 arch/x86/xen/enlighten_pv.c 	pte = pfn_pte(pfn, prot);
prot             1620 arch/x86/xen/mmu_pv.c static inline void __set_pfn_prot(unsigned long pfn, pgprot_t prot)
prot             1627 arch/x86/xen/mmu_pv.c 				pfn_pte(pfn, prot), 0);
prot             1761 arch/x86/xen/mmu_pv.c static void __init set_page_prot_flags(void *addr, pgprot_t prot,
prot             1765 arch/x86/xen/mmu_pv.c 	pte_t pte = pfn_pte(pfn, prot);
prot             1770 arch/x86/xen/mmu_pv.c static void __init set_page_prot(void *addr, pgprot_t prot)
prot             1772 arch/x86/xen/mmu_pv.c 	return set_page_prot_flags(addr, prot, UVMF_NONE);
prot             2295 arch/x86/xen/mmu_pv.c static void xen_set_fixmap(unsigned idx, phys_addr_t phys, pgprot_t prot)
prot             2312 arch/x86/xen/mmu_pv.c 		pte = pfn_pte(phys, prot);
prot             2334 arch/x86/xen/mmu_pv.c 		pte = mfn_pte(phys, prot);
prot             2339 arch/x86/xen/mmu_pv.c 		pte = mfn_pte(phys, prot);
prot             2687 arch/x86/xen/mmu_pv.c 	pgprot_t prot;
prot             2694 arch/x86/xen/mmu_pv.c 	pte_t pte = pte_mkspecial(mfn_pte(*rmd->pfn, rmd->prot));
prot             2716 arch/x86/xen/mmu_pv.c 		  xen_pfn_t *pfn, int nr, int *err_ptr, pgprot_t prot,
prot             2728 arch/x86/xen/mmu_pv.c 	rmd.prot = prot;
prot              287 arch/xtensa/include/asm/pgtable.h #define pgprot_noncached(prot) (__pgprot(pgprot_val(prot) & ~_PAGE_CA_MASK))
prot              297 arch/xtensa/include/asm/pgtable.h #define pfn_pte(pfn, prot)	__pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
prot              298 arch/xtensa/include/asm/pgtable.h #define mk_pte(page, prot)	pfn_pte(page_to_pfn(page), prot)
prot               15 arch/xtensa/mm/ioremap.c 				    pgprot_t prot)
prot               36 arch/xtensa/mm/ioremap.c 	err = ioremap_page_range(vaddr, vaddr + size, paddr, prot);
prot              136 drivers/acpi/apei/ghes.c 	pgprot_t prot;
prot              139 drivers/acpi/apei/ghes.c 	prot = arch_apei_get_mem_attribute(paddr);
prot              140 drivers/acpi/apei/ghes.c 	__set_fixmap(fixmap_idx, paddr, prot);
prot              834 drivers/acpi/arm64/iort.c 			int prot = IOMMU_WRITE | IOMMU_NOEXEC | IOMMU_MMIO;
prot              838 drivers/acpi/arm64/iort.c 							 prot, IOMMU_RESV_MSI);
prot             5426 drivers/ata/libata-core.c 	u8 prot = qc->tf.protocol;
prot             5434 drivers/ata/libata-core.c 	if (ata_is_ncq(prot)) {
prot             5454 drivers/ata/libata-core.c 	if (ata_is_data(prot) && (!qc->sg || !qc->n_elem || !qc->nbytes))
prot             5457 drivers/ata/libata-core.c 	if (ata_is_dma(prot) || (ata_is_pio(prot) &&
prot               31 drivers/ata/sata_sil24.c 	__le16	prot;
prot              792 drivers/ata/sata_sil24.c 	u8 prot = qc->tf.protocol;
prot              813 drivers/ata/sata_sil24.c 	int is_excl = (ata_is_atapi(prot) ||
prot              848 drivers/ata/sata_sil24.c 			u16 prot = 0;
prot              851 drivers/ata/sata_sil24.c 				prot |= PRB_PROT_NCQ;
prot              853 drivers/ata/sata_sil24.c 				prot |= PRB_PROT_WRITE;
prot              855 drivers/ata/sata_sil24.c 				prot |= PRB_PROT_READ;
prot              856 drivers/ata/sata_sil24.c 			prb->prot = cpu_to_le16(prot);
prot              125 drivers/char/agp/compat_ioctl.c 			ksegment[seg].prot = usegment[seg].prot;
prot               61 drivers/char/agp/compat_ioctl.h 	compat_int_t prot;		/* prot flags for mmap          */
prot              117 drivers/char/agp/frontend.c 		    (pgprot_val(seg[i].prot) == pgprot_val(page_prot))) {
prot              155 drivers/char/agp/frontend.c static pgprot_t agp_convert_mmap_flags(int prot)
prot              159 drivers/char/agp/frontend.c 	prot_bits = calc_vm_prot_bits(prot, 0) | VM_SHARED;
prot              181 drivers/char/agp/frontend.c 		seg[i].prot = agp_convert_mmap_flags(user_seg[i].prot);
prot              163 drivers/crypto/ccp/ccp-dev-v5.c #define CCP5_CMD_PROT(p)	(CCP5_CMD_DW0(p).prot)
prot              586 drivers/crypto/ccp/ccp-dev.h 	unsigned int prot:1;
prot             1330 drivers/crypto/ux500/cryp/cryp_core.c 	struct cryp_protection_config prot = {
prot             1406 drivers/crypto/ux500/cryp/cryp_core.c 	if (cryp_configure_protection(device_data, &prot)) {
prot               49 drivers/gpu/drm/etnaviv/etnaviv_iommu.c 			       size_t size, int prot)
prot               94 drivers/gpu/drm/etnaviv/etnaviv_iommu_v2.c 			       size_t size, int prot)
prot              106 drivers/gpu/drm/etnaviv/etnaviv_iommu_v2.c 	if (prot & ETNAVIV_PROT_WRITE)
prot               41 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 			      size_t size, int prot)
prot               56 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 						prot);
prot               73 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 			     struct sg_table *sgt, unsigned len, int prot)
prot               88 drivers/gpu/drm/etnaviv/etnaviv_mmu.c 		ret = etnaviv_context_map(context, da, pa, bytes, prot);
prot               26 drivers/gpu/drm/etnaviv/etnaviv_mmu.h 		   phys_addr_t paddr, size_t size, int prot);
prot               35 drivers/gpu/drm/i915/i915_mm.c 	pgprot_t prot;
prot               43 drivers/gpu/drm/i915/i915_mm.c 	set_pte_at(r->mm, addr, pte, pte_mkspecial(pfn_pte(r->pfn, r->prot)));
prot               72 drivers/gpu/drm/i915/i915_mm.c 	r.prot = __pgprot((pgprot_val(iomap->prot) & _PAGE_CACHE_MASK) |
prot              114 drivers/gpu/drm/lima/lima_gem.c 	pgprot_t prot = vm_get_page_prot(vma->vm_flags);
prot              118 drivers/gpu/drm/lima/lima_gem.c 	vma->vm_page_prot = pgprot_writecombine(prot);
prot              242 drivers/gpu/drm/msm/msm_drv.h 		struct msm_gem_vma *vma, int prot,
prot              424 drivers/gpu/drm/msm/msm_gem.c 	int prot = IOMMU_READ;
prot              427 drivers/gpu/drm/msm/msm_gem.c 		prot |= IOMMU_WRITE;
prot              442 drivers/gpu/drm/msm/msm_gem.c 	return msm_gem_map_vma(aspace, vma, prot,
prot               60 drivers/gpu/drm/msm/msm_gem_vma.c 		struct msm_gem_vma *vma, int prot,
prot               79 drivers/gpu/drm/msm/msm_gem_vma.c 				size, prot);
prot               36 drivers/gpu/drm/msm/msm_gpummu.c 		struct sg_table *sgt, unsigned len, int prot)
prot               44 drivers/gpu/drm/msm/msm_gpummu.c 	if (prot & IOMMU_WRITE)
prot               46 drivers/gpu/drm/msm/msm_gpummu.c 	if (prot & IOMMU_READ)
prot               43 drivers/gpu/drm/msm/msm_iommu.c 		struct sg_table *sgt, unsigned len, int prot)
prot               48 drivers/gpu/drm/msm/msm_iommu.c 	ret = iommu_map_sg(iommu->domain, iova, sgt->sgl, sgt->nents, prot);
prot               16 drivers/gpu/drm/msm/msm_mmu.h 			unsigned len, int prot);
prot              249 drivers/gpu/drm/panfrost/panfrost_mmu.c 		      u64 iova, int prot, struct sg_table *sgt)
prot              265 drivers/gpu/drm/panfrost/panfrost_mmu.c 			ops->map(ops, iova, paddr, pgsize, prot);
prot              283 drivers/gpu/drm/panfrost/panfrost_mmu.c 	int prot = IOMMU_READ | IOMMU_WRITE;
prot              289 drivers/gpu/drm/panfrost/panfrost_mmu.c 		prot |= IOMMU_NOEXEC;
prot              296 drivers/gpu/drm/panfrost/panfrost_mmu.c 		   prot, sgt);
prot               22 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 	int prot = IOMMU_READ | IOMMU_WRITE;
prot               39 drivers/gpu/drm/rockchip/rockchip_drm_gem.c 			   rk_obj->sgt->nents, prot);
prot              116 drivers/gpu/drm/tegra/gem.c 	int prot = IOMMU_READ | IOMMU_WRITE;
prot              139 drivers/gpu/drm/tegra/gem.c 				bo->sgt->nents, prot);
prot              473 drivers/gpu/drm/tegra/gem.c 		pgprot_t prot = vm_get_page_prot(vma->vm_flags);
prot              478 drivers/gpu/drm/tegra/gem.c 		vma->vm_page_prot = pgprot_writecombine(prot);
prot              287 drivers/gpu/drm/ttm/ttm_bo_util.c void *ttm_kmap_atomic_prot(struct page *page, pgprot_t prot)
prot              289 drivers/gpu/drm/ttm/ttm_bo_util.c 	if (pgprot_val(prot) == pgprot_val(PAGE_KERNEL))
prot              292 drivers/gpu/drm/ttm/ttm_bo_util.c 		return __ttm_kmap_atomic_prot(page, prot);
prot              303 drivers/gpu/drm/ttm/ttm_bo_util.c void ttm_kunmap_atomic_prot(void *addr, pgprot_t prot)
prot              305 drivers/gpu/drm/ttm/ttm_bo_util.c 	if (pgprot_val(prot) == pgprot_val(PAGE_KERNEL))
prot              314 drivers/gpu/drm/ttm/ttm_bo_util.c 				pgprot_t prot)
prot              323 drivers/gpu/drm/ttm/ttm_bo_util.c 	dst = ttm_kmap_atomic_prot(d, prot);
prot              329 drivers/gpu/drm/ttm/ttm_bo_util.c 	ttm_kunmap_atomic_prot(dst, prot);
prot              336 drivers/gpu/drm/ttm/ttm_bo_util.c 				pgprot_t prot)
prot              345 drivers/gpu/drm/ttm/ttm_bo_util.c 	src = ttm_kmap_atomic_prot(s, prot);
prot              351 drivers/gpu/drm/ttm/ttm_bo_util.c 	ttm_kunmap_atomic_prot(src, prot);
prot              421 drivers/gpu/drm/ttm/ttm_bo_util.c 			pgprot_t prot = ttm_io_prot(old_mem->placement,
prot              424 drivers/gpu/drm/ttm/ttm_bo_util.c 						   prot);
prot              426 drivers/gpu/drm/ttm/ttm_bo_util.c 			pgprot_t prot = ttm_io_prot(new_mem->placement,
prot              429 drivers/gpu/drm/ttm/ttm_bo_util.c 						   prot);
prot              588 drivers/gpu/drm/ttm/ttm_bo_util.c 	pgprot_t prot;
prot              611 drivers/gpu/drm/ttm/ttm_bo_util.c 		prot = ttm_io_prot(mem->placement, PAGE_KERNEL);
prot              614 drivers/gpu/drm/ttm/ttm_bo_util.c 				    0, prot);
prot              938 drivers/infiniband/core/uverbs_main.c 		      unsigned long pfn, unsigned long size, pgprot_t prot)
prot              959 drivers/infiniband/core/uverbs_main.c 	vma->vm_page_prot = prot;
prot              960 drivers/infiniband/core/uverbs_main.c 	if (io_remap_pfn_range(vma, vma->vm_start, pfn, size, prot)) {
prot              319 drivers/infiniband/hw/cxgb4/t4.h static inline pgprot_t t4_pgprot_wc(pgprot_t prot)
prot              322 drivers/infiniband/hw/cxgb4/t4.h 	return pgprot_writecombine(prot);
prot              324 drivers/infiniband/hw/cxgb4/t4.h 	return pgprot_noncached(prot);
prot             1863 drivers/infiniband/hw/mlx4/main.c 	enum mlx4_protocol prot = MLX4_PROT_IB_IPV6;
prot             1876 drivers/infiniband/hw/mlx4/main.c 				    prot, &reg_id.id);
prot             1888 drivers/infiniband/hw/mlx4/main.c 					    prot, &reg_id.mirror);
prot             1908 drivers/infiniband/hw/mlx4/main.c 			      prot, reg_id.id);
prot             1911 drivers/infiniband/hw/mlx4/main.c 				      prot, reg_id.mirror);
prot             1943 drivers/infiniband/hw/mlx4/main.c 	enum mlx4_protocol prot =  MLX4_PROT_IB_IPV6;
prot             1966 drivers/infiniband/hw/mlx4/main.c 				    prot, reg_id.id);
prot             1972 drivers/infiniband/hw/mlx4/main.c 					    prot, reg_id.mirror);
prot             2088 drivers/infiniband/hw/mlx5/main.c 	pgprot_t prot;
prot             2122 drivers/infiniband/hw/mlx5/main.c 		prot = pgprot_writecombine(vma->vm_page_prot);
prot             2125 drivers/infiniband/hw/mlx5/main.c 		prot = pgprot_noncached(vma->vm_page_prot);
prot             2168 drivers/infiniband/hw/mlx5/main.c 				prot);
prot             4602 drivers/infiniband/hw/mlx5/qp.c 	bool prot = false;
prot             4615 drivers/infiniband/hw/mlx5/qp.c 		prot = true;
prot             4618 drivers/infiniband/hw/mlx5/qp.c 	if (!prot || (data_key == prot_key && data_va == prot_va &&
prot              518 drivers/infiniband/ulp/iser/iscsi_iser.h 	struct iser_data_buf         prot[ISER_DIRS_NUM];
prot               65 drivers/infiniband/ulp/iser/iser_initiator.c 		struct iser_data_buf *pbuf_in = &iser_task->prot[ISER_DIR_IN];
prot              119 drivers/infiniband/ulp/iser/iser_initiator.c 		struct iser_data_buf *pbuf_out = &iser_task->prot[ISER_DIR_OUT];
prot              384 drivers/infiniband/ulp/iser/iser_initiator.c 		prot_buf = &iser_task->prot[ISER_DIR_IN];
prot              387 drivers/infiniband/ulp/iser/iser_initiator.c 		prot_buf = &iser_task->prot[ISER_DIR_OUT];
prot              749 drivers/infiniband/ulp/iser/iser_initiator.c 	iser_task->prot[ISER_DIR_IN].data_len  = 0;
prot              750 drivers/infiniband/ulp/iser/iser_initiator.c 	iser_task->prot[ISER_DIR_OUT].data_len = 0;
prot              752 drivers/infiniband/ulp/iser/iser_initiator.c 	iser_task->prot[ISER_DIR_IN].dma_nents = 0;
prot              753 drivers/infiniband/ulp/iser/iser_initiator.c 	iser_task->prot[ISER_DIR_OUT].dma_nents = 0;
prot              772 drivers/infiniband/ulp/iser/iser_initiator.c 						 &iser_task->prot[ISER_DIR_IN],
prot              783 drivers/infiniband/ulp/iser/iser_initiator.c 						 &iser_task->prot[ISER_DIR_OUT],
prot              525 drivers/infiniband/ulp/iser/iser_memory.c 		err = iser_reg_sig_mr(task, mem, &task->prot[dir],
prot             1662 drivers/iommu/amd_iommu.c 			  int prot,
prot             1674 drivers/iommu/amd_iommu.c 	if (!(prot & IOMMU_PROT_MASK))
prot             1696 drivers/iommu/amd_iommu.c 	if (prot & IOMMU_PROT_IR)
prot             1698 drivers/iommu/amd_iommu.c 	if (prot & IOMMU_PROT_IW)
prot             2421 drivers/iommu/amd_iommu.c 	int prot = 0;
prot             2431 drivers/iommu/amd_iommu.c 	prot = dir2prot(direction);
prot             2436 drivers/iommu/amd_iommu.c 				     PAGE_SIZE, prot, GFP_ATOMIC);
prot             2579 drivers/iommu/amd_iommu.c 	int mapped_pages = 0, npages = 0, prot = 0, i;
prot             2600 drivers/iommu/amd_iommu.c 	prot = dir2prot(direction);
prot             2612 drivers/iommu/amd_iommu.c 					     PAGE_SIZE, prot,
prot             3102 drivers/iommu/amd_iommu.c 	int prot = 0;
prot             3109 drivers/iommu/amd_iommu.c 		prot |= IOMMU_PROT_IR;
prot             3111 drivers/iommu/amd_iommu.c 		prot |= IOMMU_PROT_IW;
prot             3114 drivers/iommu/amd_iommu.c 	ret = iommu_map_page(domain, iova, paddr, page_size, prot, GFP_KERNEL);
prot             3188 drivers/iommu/amd_iommu.c 		int type, prot = 0;
prot             3196 drivers/iommu/amd_iommu.c 		if (entry->prot & IOMMU_PROT_IR)
prot             3197 drivers/iommu/amd_iommu.c 			prot |= IOMMU_READ;
prot             3198 drivers/iommu/amd_iommu.c 		if (entry->prot & IOMMU_PROT_IW)
prot             3199 drivers/iommu/amd_iommu.c 			prot |= IOMMU_WRITE;
prot             3200 drivers/iommu/amd_iommu.c 		if (entry->prot & IOMMU_UNITY_MAP_FLAG_EXCL_RANGE)
prot             3205 drivers/iommu/amd_iommu.c 						 length, prot, type);
prot             2136 drivers/iommu/amd_iommu_init.c 	e->prot = m->flags >> 1;
prot              700 drivers/iommu/amd_iommu_types.h 	int prot;
prot             2453 drivers/iommu/arm-smmu-v3.c 			phys_addr_t paddr, size_t size, int prot)
prot             2460 drivers/iommu/arm-smmu-v3.c 	return ops->map(ops, iova, paddr, size, prot);
prot             2703 drivers/iommu/arm-smmu-v3.c 	int prot = IOMMU_WRITE | IOMMU_NOEXEC | IOMMU_MMIO;
prot             2706 drivers/iommu/arm-smmu-v3.c 					 prot, IOMMU_RESV_SW_MSI);
prot             1163 drivers/iommu/arm-smmu.c 			phys_addr_t paddr, size_t size, int prot)
prot             1173 drivers/iommu/arm-smmu.c 	ret = ops->map(ops, iova, paddr, size, prot);
prot             1534 drivers/iommu/arm-smmu.c 	int prot = IOMMU_WRITE | IOMMU_NOEXEC | IOMMU_MMIO;
prot             1537 drivers/iommu/arm-smmu.c 					 prot, IOMMU_RESV_SW_MSI);
prot              367 drivers/iommu/dma-iommu.c 	int prot = coherent ? IOMMU_CACHE : 0;
prot              370 drivers/iommu/dma-iommu.c 		prot |= IOMMU_PRIV;
prot              374 drivers/iommu/dma-iommu.c 		return prot | IOMMU_READ | IOMMU_WRITE;
prot              376 drivers/iommu/dma-iommu.c 		return prot | IOMMU_READ;
prot              378 drivers/iommu/dma-iommu.c 		return prot | IOMMU_WRITE;
prot              464 drivers/iommu/dma-iommu.c 		size_t size, int prot)
prot              478 drivers/iommu/dma-iommu.c 	if (iommu_map(domain, iova, phys - iova_off, size, prot)) {
prot              572 drivers/iommu/dma-iommu.c 	pgprot_t prot = dma_pgprot(dev, PAGE_KERNEL, attrs);
prot              617 drivers/iommu/dma-iommu.c 	vaddr = dma_common_pages_remap(pages, size, prot,
prot              710 drivers/iommu/dma-iommu.c 	int prot = dma_info_to_prot(dir, coherent, attrs);
prot              713 drivers/iommu/dma-iommu.c 	dma_handle =__iommu_dma_map(dev, phys, size, prot);
prot              817 drivers/iommu/dma-iommu.c 	int prot = dma_info_to_prot(dir, dev_is_dma_coherent(dev), attrs);
prot              873 drivers/iommu/dma-iommu.c 	if (iommu_map_sg(domain, iova, sg, nents, prot) < iova_len)
prot              976 drivers/iommu/dma-iommu.c 		pgprot_t prot = dma_pgprot(dev, PAGE_KERNEL, attrs);
prot              979 drivers/iommu/dma-iommu.c 				prot, __builtin_return_address(0));
prot             1144 drivers/iommu/dma-iommu.c 	int prot = IOMMU_WRITE | IOMMU_NOEXEC | IOMMU_MMIO;
prot             1160 drivers/iommu/dma-iommu.c 	if (iommu_map(domain, iova, msi_addr, size, prot))
prot              124 drivers/iommu/exynos-iommu.c #define mk_lv1ent_sect(pa, prot) ((pa >> PG_ENT_SHIFT) | LV1_PROT[prot] | 2)
prot              126 drivers/iommu/exynos-iommu.c #define mk_lv2ent_lpage(pa, prot) ((pa >> PG_ENT_SHIFT) | LV2_PROT[prot] | 1)
prot              127 drivers/iommu/exynos-iommu.c #define mk_lv2ent_spage(pa, prot) ((pa >> PG_ENT_SHIFT) | LV2_PROT[prot] | 2)
prot              978 drivers/iommu/exynos-iommu.c 			  phys_addr_t paddr, int prot, short *pgcnt)
prot              997 drivers/iommu/exynos-iommu.c 	update_pte(sent, mk_lv1ent_sect(paddr, prot));
prot             1015 drivers/iommu/exynos-iommu.c 		       int prot, short *pgcnt)
prot             1021 drivers/iommu/exynos-iommu.c 		update_pte(pent, mk_lv2ent_spage(paddr, prot));
prot             1037 drivers/iommu/exynos-iommu.c 			*pent = mk_lv2ent_lpage(paddr, prot);
prot             1076 drivers/iommu/exynos-iommu.c 			    int prot)
prot             1085 drivers/iommu/exynos-iommu.c 	prot &= SYSMMU_SUPPORTED_PROT_BITS;
prot             1092 drivers/iommu/exynos-iommu.c 		ret = lv1set_section(domain, entry, iova, paddr, prot,
prot             1103 drivers/iommu/exynos-iommu.c 			ret = lv2set_page(pent, paddr, size, prot,
prot              324 drivers/iommu/fsl_pamu.c 		       u32 subwin_cnt, int prot)
prot              388 drivers/iommu/fsl_pamu.c 		set_bf(ppaace->addr_bitfields, PAACE_AF_AP, prot);
prot              416 drivers/iommu/fsl_pamu.c 		       u32 snoopid, u32 stashid, int enable, int prot)
prot              458 drivers/iommu/fsl_pamu.c 	set_bf(paace->addr_bitfields, PAACE_AF_AP, prot);
prot              389 drivers/iommu/fsl_pamu.h 		       u32 subwin_cnt, int prot);
prot              392 drivers/iommu/fsl_pamu.h 		       uint32_t snoopid, u32 stashid, int enable, int prot);
prot              104 drivers/iommu/fsl_pamu_domain.c 						 sub_win_ptr[i].prot);
prot              130 drivers/iommu/fsl_pamu_domain.c 				 0, wnd->prot);
prot              163 drivers/iommu/fsl_pamu_domain.c 					 wnd->prot);
prot              177 drivers/iommu/fsl_pamu_domain.c 					 0, wnd->prot);
prot              525 drivers/iommu/fsl_pamu_domain.c 				  phys_addr_t paddr, u64 size, int prot)
prot              534 drivers/iommu/fsl_pamu_domain.c 	if (prot & IOMMU_READ)
prot              536 drivers/iommu/fsl_pamu_domain.c 	if (prot & IOMMU_WRITE)
prot              578 drivers/iommu/fsl_pamu_domain.c 		wnd->prot = pamu_prot;
prot               16 drivers/iommu/fsl_pamu_domain.h 	int prot;
prot             2225 drivers/iommu/intel-iommu.c 			    unsigned long nr_pages, int prot)
prot             2235 drivers/iommu/intel-iommu.c 	if ((prot & (DMA_PTE_READ|DMA_PTE_WRITE)) == 0)
prot             2238 drivers/iommu/intel-iommu.c 	prot &= DMA_PTE_READ | DMA_PTE_WRITE | DMA_PTE_SNP;
prot             2242 drivers/iommu/intel-iommu.c 		pteval = ((phys_addr_t)phys_pfn << VTD_PAGE_SHIFT) | prot;
prot             2254 drivers/iommu/intel-iommu.c 			pteval = (sg_phys(sg) - pgoff) | prot;
prot             2340 drivers/iommu/intel-iommu.c 			  unsigned long nr_pages, int prot)
prot             2346 drivers/iommu/intel-iommu.c 	ret = __domain_mapping(domain, iov_pfn, sg, phys_pfn, nr_pages, prot);
prot             2360 drivers/iommu/intel-iommu.c 				    int prot)
prot             2362 drivers/iommu/intel-iommu.c 	return domain_mapping(domain, iov_pfn, sg, 0, nr_pages, prot);
prot             2367 drivers/iommu/intel-iommu.c 				     int prot)
prot             2369 drivers/iommu/intel-iommu.c 	return domain_mapping(domain, iov_pfn, NULL, phys_pfn, nr_pages, prot);
prot             3498 drivers/iommu/intel-iommu.c 	int prot = 0;
prot             3522 drivers/iommu/intel-iommu.c 		prot |= DMA_PTE_READ;
prot             3524 drivers/iommu/intel-iommu.c 		prot |= DMA_PTE_WRITE;
prot             3532 drivers/iommu/intel-iommu.c 				 mm_to_dma_pfn(paddr_pfn), size, prot);
prot             3714 drivers/iommu/intel-iommu.c 	int prot = 0;
prot             3747 drivers/iommu/intel-iommu.c 		prot |= DMA_PTE_READ;
prot             3749 drivers/iommu/intel-iommu.c 		prot |= DMA_PTE_WRITE;
prot             3753 drivers/iommu/intel-iommu.c 	ret = domain_sg_mapping(domain, start_vpfn, sglist, size, prot);
prot             3817 drivers/iommu/intel-iommu.c 	int prot = 0;
prot             3840 drivers/iommu/intel-iommu.c 		prot |= DMA_PTE_READ;
prot             3842 drivers/iommu/intel-iommu.c 		prot |= DMA_PTE_WRITE;
prot             3873 drivers/iommu/intel-iommu.c 				 tlb_addr >> VTD_PAGE_SHIFT, nrpages, prot);
prot             5445 drivers/iommu/intel-iommu.c 	int prot = 0;
prot             5449 drivers/iommu/intel-iommu.c 		prot |= DMA_PTE_READ;
prot             5451 drivers/iommu/intel-iommu.c 		prot |= DMA_PTE_WRITE;
prot             5453 drivers/iommu/intel-iommu.c 		prot |= DMA_PTE_SNP;
prot             5473 drivers/iommu/intel-iommu.c 				 hpa >> VTD_PAGE_SHIFT, size, prot);
prot             5669 drivers/iommu/intel-iommu.c 	int prot = DMA_PTE_READ | DMA_PTE_WRITE;
prot             5693 drivers/iommu/intel-iommu.c 						       length, prot, type);
prot             5707 drivers/iommu/intel-iommu.c 			reg = iommu_alloc_resv_region(0, 1UL << 24, prot,
prot              316 drivers/iommu/io-pgtable-arm-v7s.c static arm_v7s_iopte arm_v7s_prot_to_pte(int prot, int lvl,
prot              322 drivers/iommu/io-pgtable-arm-v7s.c 	if (!(prot & IOMMU_MMIO))
prot              326 drivers/iommu/io-pgtable-arm-v7s.c 		if (!(prot & IOMMU_PRIV))
prot              328 drivers/iommu/io-pgtable-arm-v7s.c 		if (!(prot & IOMMU_WRITE))
prot              333 drivers/iommu/io-pgtable-arm-v7s.c 	if ((prot & IOMMU_NOEXEC) && ap)
prot              335 drivers/iommu/io-pgtable-arm-v7s.c 	if (prot & IOMMU_MMIO)
prot              337 drivers/iommu/io-pgtable-arm-v7s.c 	else if (prot & IOMMU_CACHE)
prot              349 drivers/iommu/io-pgtable-arm-v7s.c 	int prot = IOMMU_READ;
prot              353 drivers/iommu/io-pgtable-arm-v7s.c 		prot |= IOMMU_WRITE;
prot              355 drivers/iommu/io-pgtable-arm-v7s.c 		prot |= IOMMU_PRIV;
prot              357 drivers/iommu/io-pgtable-arm-v7s.c 		prot |= IOMMU_MMIO;
prot              359 drivers/iommu/io-pgtable-arm-v7s.c 		prot |= IOMMU_CACHE;
prot              361 drivers/iommu/io-pgtable-arm-v7s.c 		prot |= IOMMU_NOEXEC;
prot              363 drivers/iommu/io-pgtable-arm-v7s.c 	return prot;
prot              413 drivers/iommu/io-pgtable-arm-v7s.c 			    unsigned long iova, phys_addr_t paddr, int prot,
prot              439 drivers/iommu/io-pgtable-arm-v7s.c 	pte = arm_v7s_prot_to_pte(prot, lvl, cfg);
prot              474 drivers/iommu/io-pgtable-arm-v7s.c 			 phys_addr_t paddr, size_t size, int prot,
prot              486 drivers/iommu/io-pgtable-arm-v7s.c 		return arm_v7s_init_pte(data, iova, paddr, prot,
prot              517 drivers/iommu/io-pgtable-arm-v7s.c 	return __arm_v7s_map(data, iova, paddr, size, prot, lvl + 1, cptep);
prot              521 drivers/iommu/io-pgtable-arm-v7s.c 			phys_addr_t paddr, size_t size, int prot)
prot              528 drivers/iommu/io-pgtable-arm-v7s.c 	if (!(prot & (IOMMU_READ | IOMMU_WRITE)))
prot              535 drivers/iommu/io-pgtable-arm-v7s.c 	ret = __arm_v7s_map(data, iova, paddr, size, prot, 1, data->pgd);
prot              300 drivers/iommu/io-pgtable-arm.c 				phys_addr_t paddr, arm_lpae_iopte prot,
prot              303 drivers/iommu/io-pgtable-arm.c 	arm_lpae_iopte pte = prot;
prot              323 drivers/iommu/io-pgtable-arm.c 			     arm_lpae_iopte prot, int lvl,
prot              347 drivers/iommu/io-pgtable-arm.c 	__arm_lpae_init_pte(data, paddr, prot, lvl, ptep);
prot              383 drivers/iommu/io-pgtable-arm.c 			  phys_addr_t paddr, size_t size, arm_lpae_iopte prot,
prot              396 drivers/iommu/io-pgtable-arm.c 		return arm_lpae_init_pte(data, iova, paddr, prot, lvl, ptep);
prot              425 drivers/iommu/io-pgtable-arm.c 	return __arm_lpae_map(data, iova, paddr, size, prot, lvl + 1, cptep);
prot              429 drivers/iommu/io-pgtable-arm.c 					   int prot)
prot              436 drivers/iommu/io-pgtable-arm.c 		if (!(prot & IOMMU_WRITE) && (prot & IOMMU_READ))
prot              438 drivers/iommu/io-pgtable-arm.c 		if (!(prot & IOMMU_PRIV))
prot              442 drivers/iommu/io-pgtable-arm.c 		if (prot & IOMMU_READ)
prot              444 drivers/iommu/io-pgtable-arm.c 		if (prot & IOMMU_WRITE)
prot              454 drivers/iommu/io-pgtable-arm.c 		if (prot & IOMMU_MMIO)
prot              456 drivers/iommu/io-pgtable-arm.c 		else if (prot & IOMMU_CACHE)
prot              461 drivers/iommu/io-pgtable-arm.c 		if (prot & IOMMU_MMIO)
prot              464 drivers/iommu/io-pgtable-arm.c 		else if (prot & IOMMU_CACHE)
prot              467 drivers/iommu/io-pgtable-arm.c 		else if (prot & IOMMU_QCOM_SYS_CACHE)
prot              472 drivers/iommu/io-pgtable-arm.c 	if (prot & IOMMU_NOEXEC)
prot              484 drivers/iommu/io-pgtable-arm.c 	arm_lpae_iopte prot;
prot              494 drivers/iommu/io-pgtable-arm.c 	prot = arm_lpae_prot_to_pte(data, iommu_prot);
prot              495 drivers/iommu/io-pgtable-arm.c 	ret = __arm_lpae_map(data, iova, paddr, size, prot, lvl, ptep);
prot              299 drivers/iommu/iommu.c 				     new->prot, new->type);
prot              663 drivers/iommu/iommu.c 			ret = iommu_map(domain, addr, addr, pg_size, entry->prot);
prot             1859 drivers/iommu/iommu.c 	      phys_addr_t paddr, size_t size, int prot)
prot             1897 drivers/iommu/iommu.c 		ret = ops->map(domain, iova, paddr, pgsize, prot);
prot             1996 drivers/iommu/iommu.c 		    struct scatterlist *sg, unsigned int nents, int prot)
prot             2007 drivers/iommu/iommu.c 			ret = iommu_map(domain, iova + mapped, start, len, prot);
prot             2038 drivers/iommu/iommu.c 			       phys_addr_t paddr, u64 size, int prot)
prot             2044 drivers/iommu/iommu.c 						 prot);
prot             2173 drivers/iommu/iommu.c 						  size_t length, int prot,
prot             2185 drivers/iommu/iommu.c 	region->prot = prot;
prot              727 drivers/iommu/ipmmu-vmsa.c 		     phys_addr_t paddr, size_t size, int prot)
prot              734 drivers/iommu/ipmmu-vmsa.c 	return domain->iop->map(domain->iop, iova, paddr, size, prot);
prot              507 drivers/iommu/msm_iommu.c 			 phys_addr_t pa, size_t len, int prot)
prot              514 drivers/iommu/msm_iommu.c 	ret = priv->iop->map(priv->iop, iova, pa, len, prot);
prot              430 drivers/iommu/mtk_iommu.c 			 phys_addr_t paddr, size_t size, int prot)
prot              442 drivers/iommu/mtk_iommu.c 	ret = dom->iop->map(dom->iop, iova, paddr, size, prot);
prot               60 drivers/iommu/mtk_iommu_v1.c #define F_MMU_TF_PROTECT_SEL(prot)		(((prot) & 0x3) << 5)
prot              298 drivers/iommu/mtk_iommu_v1.c 			 phys_addr_t paddr, size_t size, int prot)
prot              556 drivers/iommu/omap-iommu.c static int iopgd_alloc_section(struct omap_iommu *obj, u32 da, u32 pa, u32 prot)
prot              567 drivers/iommu/omap-iommu.c 	*iopgd = (pa & IOSECTION_MASK) | prot | IOPGD_SECTION;
prot              572 drivers/iommu/omap-iommu.c static int iopgd_alloc_super(struct omap_iommu *obj, u32 da, u32 pa, u32 prot)
prot              585 drivers/iommu/omap-iommu.c 		*(iopgd + i) = (pa & IOSUPER_MASK) | prot | IOPGD_SUPER;
prot              590 drivers/iommu/omap-iommu.c static int iopte_alloc_page(struct omap_iommu *obj, u32 da, u32 pa, u32 prot)
prot              600 drivers/iommu/omap-iommu.c 	*iopte = (pa & IOPAGE_MASK) | prot | IOPTE_SMALL;
prot              609 drivers/iommu/omap-iommu.c static int iopte_alloc_large(struct omap_iommu *obj, u32 da, u32 pa, u32 prot)
prot              627 drivers/iommu/omap-iommu.c 		*(iopte + i) = (pa & IOLARGE_MASK) | prot | IOPTE_LARGE;
prot              636 drivers/iommu/omap-iommu.c 	u32 prot;
prot              663 drivers/iommu/omap-iommu.c 	prot = get_iopte_attr(e);
prot              666 drivers/iommu/omap-iommu.c 	err = fn(obj, e->da, e->pa, prot);
prot             1342 drivers/iommu/omap-iommu.c 			  phys_addr_t pa, size_t bytes, int prot)
prot              422 drivers/iommu/qcom_iommu.c 			  phys_addr_t paddr, size_t size, int prot)
prot              433 drivers/iommu/qcom_iommu.c 	ret = ops->map(ops, iova, paddr, size, prot);
prot              229 drivers/iommu/rockchip-iommu.c static u32 rk_mk_pte(phys_addr_t page, int prot)
prot              232 drivers/iommu/rockchip-iommu.c 	flags |= (prot & IOMMU_READ) ? RK_PTE_PAGE_READABLE : 0;
prot              233 drivers/iommu/rockchip-iommu.c 	flags |= (prot & IOMMU_WRITE) ? RK_PTE_PAGE_WRITABLE : 0;
prot              717 drivers/iommu/rockchip-iommu.c 			     phys_addr_t paddr, size_t size, int prot)
prot              731 drivers/iommu/rockchip-iommu.c 		pte_addr[pte_count] = rk_mk_pte(paddr, prot);
prot              755 drivers/iommu/rockchip-iommu.c 	       &iova, &page_phys, &paddr, prot);
prot              761 drivers/iommu/rockchip-iommu.c 			phys_addr_t paddr, size_t size, int prot)
prot              790 drivers/iommu/rockchip-iommu.c 				paddr, size, prot);
prot              268 drivers/iommu/s390-iommu.c 			  phys_addr_t paddr, size_t size, int prot)
prot              273 drivers/iommu/s390-iommu.c 	if (!(prot & IOMMU_READ))
prot              276 drivers/iommu/s390-iommu.c 	if (!(prot & IOMMU_WRITE))
prot              181 drivers/iommu/tegra-gart.c 			  phys_addr_t pa, size_t bytes, int prot)
prot              654 drivers/iommu/tegra-smmu.c 			  phys_addr_t paddr, size_t size, int prot)
prot              671 drivers/iommu/tegra-smmu.c 	if (prot & IOMMU_READ)
prot              674 drivers/iommu/tegra-smmu.c 	if (prot & IOMMU_WRITE)
prot              427 drivers/iommu/virtio-iommu.c 	unsigned long prot = IOMMU_WRITE | IOMMU_NOEXEC | IOMMU_MMIO;
prot              450 drivers/iommu/virtio-iommu.c 		region = iommu_alloc_resv_region(start, size, prot,
prot              718 drivers/iommu/virtio-iommu.c 		      phys_addr_t paddr, size_t size, int prot)
prot              725 drivers/iommu/virtio-iommu.c 	flags = (prot & IOMMU_READ ? VIRTIO_IOMMU_MAP_F_READ : 0) |
prot              726 drivers/iommu/virtio-iommu.c 		(prot & IOMMU_WRITE ? VIRTIO_IOMMU_MAP_F_WRITE : 0) |
prot              727 drivers/iommu/virtio-iommu.c 		(prot & IOMMU_MMIO ? VIRTIO_IOMMU_MAP_F_MMIO : 0);
prot              815 drivers/iommu/virtio-iommu.c 	int prot = IOMMU_WRITE | IOMMU_NOEXEC | IOMMU_MMIO;
prot              833 drivers/iommu/virtio-iommu.c 					      prot, IOMMU_RESV_SW_MSI);
prot               49 drivers/misc/mic/scif/scif_debugfs.c 		   window->nr_pages, window->nr_contig_chunks, window->prot);
prot              342 drivers/misc/mic/scif/scif_dma.c scif_register_temp(scif_epd_t epd, unsigned long addr, size_t len, int prot,
prot              353 drivers/misc/mic/scif/scif_dma.c 			       aligned_len, &prot, 0, &pinned_pages);
prot              357 drivers/misc/mic/scif/scif_dma.c 	pinned_pages->prot = prot;
prot              377 drivers/misc/mic/scif/scif_dma.c 	(*out_window)->prot = pinned_pages->prot;
prot             1679 drivers/misc/mic/scif/scif_dma.c 	remote_req.prot = dir == SCIF_LOCAL_TO_REMOTE ? VM_WRITE : VM_READ;
prot             1707 drivers/misc/mic/scif/scif_dma.c 		req.prot = (dir == SCIF_LOCAL_TO_REMOTE ?
prot             1718 drivers/misc/mic/scif/scif_dma.c 						 req.nr_bytes, req.prot,
prot             1746 drivers/misc/mic/scif/scif_dma.c 		req.prot = dir == SCIF_LOCAL_TO_REMOTE ? VM_READ : VM_WRITE;
prot              314 drivers/misc/mic/scif/scif_fd.c 				    reg.offset, reg.prot, reg.flags);
prot              306 drivers/misc/mic/scif/scif_fence.c 	req.prot = SCIF_PROT_WRITE;
prot               33 drivers/misc/mic/scif/scif_mmap.c 	req.prot = recv_window->prot;
prot              237 drivers/misc/mic/scif/scif_mmap.c 	req.prot = 0;
prot              276 drivers/misc/mic/scif/scif_mmap.c 	(*pages)->prot_flags = window->prot;
prot              565 drivers/misc/mic/scif/scif_mmap.c 	req.prot = vma->vm_flags & (VM_READ | VM_WRITE);
prot              637 drivers/misc/mic/scif/scif_mmap.c 	req.prot = vma->vm_flags & (VM_READ | VM_WRITE);
prot               81 drivers/misc/mic/scif/scif_rma.c scif_create_pinned_pages(int nr_pages, int prot)
prot               94 drivers/misc/mic/scif/scif_rma.c 	pin->prot = prot;
prot              113 drivers/misc/mic/scif/scif_rma.c 	int writeable = pin->prot & SCIF_PROT_WRITE;
prot              917 drivers/misc/mic/scif/scif_rma.c 	remote_window->prot = window->prot;
prot             1186 drivers/misc/mic/scif/scif_rma.c 	req.prot = 0;
prot             1314 drivers/misc/mic/scif/scif_rma.c 	int prot = *out_prot;
prot             1324 drivers/misc/mic/scif/scif_rma.c 	if (prot & ~(SCIF_PROT_READ | SCIF_PROT_WRITE))
prot             1338 drivers/misc/mic/scif/scif_rma.c 	pinned_pages = scif_create_pinned_pages(nr_pages, prot);
prot             1365 drivers/misc/mic/scif/scif_rma.c 		if (prot == SCIF_PROT_READ)
prot             1367 drivers/misc/mic/scif/scif_rma.c 		prot |= SCIF_PROT_WRITE;
prot             1381 drivers/misc/mic/scif/scif_rma.c 				(prot & SCIF_PROT_WRITE) ? FOLL_WRITE : 0,
prot             1393 drivers/misc/mic/scif/scif_rma.c 				prot &= ~SCIF_PROT_WRITE;
prot             1407 drivers/misc/mic/scif/scif_rma.c 	*out_prot = prot;
prot             1424 drivers/misc/mic/scif/scif_rma.c int scif_pin_pages(void *addr, size_t len, int prot,
prot             1427 drivers/misc/mic/scif/scif_rma.c 	return __scif_pin_pages(addr, len, &prot, map_flags, pages);
prot             1524 drivers/misc/mic/scif/scif_rma.c 	window->prot = pinned_pages->prot;
prot             1569 drivers/misc/mic/scif/scif_rma.c 		    int prot, int map_flags)
prot             1581 drivers/misc/mic/scif/scif_rma.c 		epd, addr, len, offset, prot, map_flags);
prot             1597 drivers/misc/mic/scif/scif_rma.c 	if (prot & ~(SCIF_PROT_READ | SCIF_PROT_WRITE))
prot             1646 drivers/misc/mic/scif/scif_rma.c 	err = __scif_pin_pages(addr, len, &prot,
prot             1656 drivers/misc/mic/scif/scif_rma.c 	window->prot = pinned_pages->prot;
prot             1724 drivers/misc/mic/scif/scif_rma.c 	req.prot = 0;
prot              188 drivers/misc/mic/scif/scif_rma.h 	int prot;
prot              258 drivers/misc/mic/scif/scif_rma.h 	int prot;
prot              106 drivers/misc/mic/scif/scif_rma_list.c 		if ((window->prot & req->prot) == req->prot) {
prot              155 drivers/misc/mic/scif/scif_rma_list.c 		if ((window->prot & req->prot) != req->prot)
prot               30 drivers/misc/mic/scif/scif_rma_list.h 	int prot;
prot             2223 drivers/mtd/chips/cfi_cmdset_0001.c 	    u_char *buf, u_int size, u_long prot, u_int grpno, u_int grpsz)
prot             2256 drivers/mtd/chips/cfi_cmdset_0001.c 	     u_char *buf, u_int size, u_long prot, u_int grpno, u_int grpsz)
prot             2281 drivers/mtd/chips/cfi_cmdset_0001.c 	    u_char *buf, u_int size, u_long prot, u_int grpno, u_int grpsz)
prot             2292 drivers/mtd/chips/cfi_cmdset_0001.c 	return do_write_oneword(map, chip, prot, datum, FL_OTP_WRITE);
prot             3340 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 	__u8 prot = 0;
prot             3349 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 		prot = ipv6_hdr(skb)->nexthdr;
prot             3352 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 		prot = ip_hdr(skb)->protocol;
prot             3366 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 	if (prot == IPPROTO_TCP)
prot             2918 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c int dpaa2_eth_cls_fld_off(int prot, int field)
prot             2923 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 		if (dist_fields[i].cls_prot == prot &&
prot             2986 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c 		key->extract.from_hdr.prot = dist_fields[i].cls_prot;
prot              523 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.h int dpaa2_eth_cls_fld_off(int prot, int field);
prot              448 drivers/net/ethernet/freescale/dpaa2/dpkg.h 			enum net_prot			prot;
prot              407 drivers/net/ethernet/freescale/dpaa2/dpni-cmd.h 	u8 prot;
prot               37 drivers/net/ethernet/freescale/dpaa2/dpni.c 			extr->prot = cfg->extracts[i].extract.from_hdr.prot;
prot              309 drivers/net/ethernet/freescale/enetc/enetc_pf.c static int enetc_vlan_rx_add_vid(struct net_device *ndev, __be16 prot, u16 vid)
prot              327 drivers/net/ethernet/freescale/enetc/enetc_pf.c static int enetc_vlan_rx_del_vid(struct net_device *ndev, __be16 prot, u16 vid)
prot             2716 drivers/net/ethernet/mellanox/mlx4/fw.c 	u8 prot;
prot             2757 drivers/net/ethernet/mellanox/mlx4/fw.c 			prot = ((u8 *)(&mgm->members_count))[0] >> 6;
prot             2764 drivers/net/ethernet/mellanox/mlx4/fw.c 								    prot, 0);
prot             2769 drivers/net/ethernet/mellanox/mlx4/fw.c 								    , 0, prot,
prot              158 drivers/net/ethernet/mellanox/mlx4/mcg.c 	u32 prot;
prot              206 drivers/net/ethernet/mellanox/mlx4/mcg.c 	prot = be32_to_cpu(mgm->members_count) >> 30;
prot              221 drivers/net/ethernet/mellanox/mlx4/mcg.c 	mgm->members_count = cpu_to_be32(members_count | (prot << 30));
prot              428 drivers/net/ethernet/mellanox/mlx4/mcg.c 	u32 prot;
prot              474 drivers/net/ethernet/mellanox/mlx4/mcg.c 			prot = be32_to_cpu(mgm->members_count) >> 30;
prot              505 drivers/net/ethernet/mellanox/mlx4/mcg.c 						    (prot << 30));
prot              695 drivers/net/ethernet/mellanox/mlx4/mcg.c 		      u8 *gid, enum mlx4_protocol prot,
prot              704 drivers/net/ethernet/mellanox/mlx4/mcg.c 	u8 op_mod = (prot == MLX4_PROT_ETH) ?
prot              739 drivers/net/ethernet/mellanox/mlx4/mcg.c 		    be32_to_cpu(mgm->members_count) >> 30 == prot)
prot             1105 drivers/net/ethernet/mellanox/mlx4/mcg.c 			  int block_mcast_loopback, enum mlx4_protocol prot,
prot             1125 drivers/net/ethernet/mellanox/mlx4/mcg.c 	err = find_entry(dev, port, gid, prot,
prot             1171 drivers/net/ethernet/mellanox/mlx4/mcg.c 	mgm->members_count = cpu_to_be32(members_count | (u32) prot << 30);
prot             1191 drivers/net/ethernet/mellanox/mlx4/mcg.c 	if (prot == MLX4_PROT_ETH && index != -1) {
prot             1215 drivers/net/ethernet/mellanox/mlx4/mcg.c 			  enum mlx4_protocol prot, enum mlx4_steer_type steer)
prot             1234 drivers/net/ethernet/mellanox/mlx4/mcg.c 	err = find_entry(dev, port, gid, prot,
prot             1248 drivers/net/ethernet/mellanox/mlx4/mcg.c 	if (prot == MLX4_PROT_ETH &&
prot             1269 drivers/net/ethernet/mellanox/mlx4/mcg.c 	mgm->members_count = cpu_to_be32(--members_count | (u32) prot << 30);
prot             1271 drivers/net/ethernet/mellanox/mlx4/mcg.c 	if (prot == MLX4_PROT_ETH)
prot             1274 drivers/net/ethernet/mellanox/mlx4/mcg.c 	if (members_count && (prot != MLX4_PROT_ETH || !removed_entry)) {
prot             1280 drivers/net/ethernet/mellanox/mlx4/mcg.c 	mgm->members_count = cpu_to_be32((u32) prot << 30);
prot             1337 drivers/net/ethernet/mellanox/mlx4/mcg.c 			  enum mlx4_protocol prot)
prot             1352 drivers/net/ethernet/mellanox/mlx4/mcg.c 	qpn |= (prot << 28);
prot             1370 drivers/net/ethernet/mellanox/mlx4/mcg.c 			      enum mlx4_protocol prot, u64 *reg_id)
prot             1387 drivers/net/ethernet/mellanox/mlx4/mcg.c 		switch (prot) {
prot             1409 drivers/net/ethernet/mellanox/mlx4/mcg.c 			  enum mlx4_protocol prot, u64 *reg_id)
prot             1413 drivers/net/ethernet/mellanox/mlx4/mcg.c 		if (prot == MLX4_PROT_ETH)
prot             1418 drivers/net/ethernet/mellanox/mlx4/mcg.c 		if (prot == MLX4_PROT_ETH)
prot             1423 drivers/net/ethernet/mellanox/mlx4/mcg.c 					      block_mcast_loopback, prot);
prot             1425 drivers/net/ethernet/mellanox/mlx4/mcg.c 					     block_mcast_loopback, prot,
prot             1431 drivers/net/ethernet/mellanox/mlx4/mcg.c 						 prot, reg_id);
prot             1439 drivers/net/ethernet/mellanox/mlx4/mcg.c 			  enum mlx4_protocol prot, u64 reg_id)
prot             1443 drivers/net/ethernet/mellanox/mlx4/mcg.c 		if (prot == MLX4_PROT_ETH)
prot             1448 drivers/net/ethernet/mellanox/mlx4/mcg.c 		if (prot == MLX4_PROT_ETH)
prot             1452 drivers/net/ethernet/mellanox/mlx4/mcg.c 			return mlx4_QP_ATTACH(dev, qp, gid, 0, 0, prot);
prot             1454 drivers/net/ethernet/mellanox/mlx4/mcg.c 		return mlx4_qp_detach_common(dev, qp, gid, prot,
prot             1531 drivers/net/ethernet/mellanox/mlx4/mcg.c 			int block_mcast_loopback, enum mlx4_protocol prot)
prot             1533 drivers/net/ethernet/mellanox/mlx4/mcg.c 	if (prot == MLX4_PROT_ETH)
prot             1538 drivers/net/ethernet/mellanox/mlx4/mcg.c 					block_mcast_loopback, prot);
prot             1541 drivers/net/ethernet/mellanox/mlx4/mcg.c 					prot, MLX4_UC_STEER);
prot             1546 drivers/net/ethernet/mellanox/mlx4/mcg.c 			       u8 gid[16], enum mlx4_protocol prot)
prot             1548 drivers/net/ethernet/mellanox/mlx4/mcg.c 	if (prot == MLX4_PROT_ETH)
prot             1552 drivers/net/ethernet/mellanox/mlx4/mcg.c 		return mlx4_QP_ATTACH(dev, qp, gid, 0, 0, prot);
prot             1554 drivers/net/ethernet/mellanox/mlx4/mcg.c 	return mlx4_qp_detach_common(dev, qp, gid, prot, MLX4_UC_STEER);
prot             1322 drivers/net/ethernet/mellanox/mlx4/mlx4.h 			  enum mlx4_protocol prot, enum mlx4_steer_type steer);
prot             1324 drivers/net/ethernet/mellanox/mlx4/mlx4.h 			  int block_mcast_loopback, enum mlx4_protocol prot,
prot             1329 drivers/net/ethernet/mellanox/mlx4/mlx4.h 			      enum mlx4_protocol prot, u64 *reg_id);
prot               90 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	enum mlx4_protocol	prot;
prot             4028 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 		       u8 *gid, enum mlx4_protocol prot,
prot             4044 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 		res->prot = prot;
prot             4056 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 		       u8 *gid, enum mlx4_protocol prot,
prot             4064 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	if (!res || res->prot != prot || res->steer != steer)
prot             4078 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 		     u8 gid[16], int block_loopback, enum mlx4_protocol prot,
prot             4087 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 						block_loopback, prot,
prot             4091 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 		if (prot == MLX4_PROT_ETH) {
prot             4098 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 					    block_loopback, prot, type);
prot             4105 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 		     u8 gid[16], enum mlx4_protocol prot,
prot             4112 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 		return mlx4_qp_detach_common(dev, qp, gid, prot, type);
prot             4119 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 			    u8 *gid, enum mlx4_protocol prot)
prot             4123 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	if (prot != MLX4_PROT_ETH)
prot             4145 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	enum mlx4_protocol prot = (vhcr->in_modifier >> 28) & 0x7;
prot             4162 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 		err = qp_attach(dev, slave, &qp, gid, block_loopback, prot,
prot             4168 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 		err = add_mcg_res(dev, slave, rqp, gid, prot, type, reg_id);
prot             4172 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 		err = mlx4_adjust_port(dev, slave, gid, prot);
prot             4176 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 		err = rem_mcg_res(dev, slave, rqp, gid, prot, type, &reg_id);
prot             4180 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 		err = qp_detach(dev, &qp, gid, prot, type, reg_id);
prot             4189 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 	qp_detach(dev, &qp, gid, prot, type, reg_id);
prot             4595 drivers/net/ethernet/mellanox/mlx4/resource_tracker.c 						     rgid->prot, rgid->steer);
prot              869 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 	u32 prot[6];
prot              874 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 	for (i = 0; i < ARRAY_SIZE(prot); i++) {
prot              875 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 		prot[i] = mt76_rr(dev, MT_CCK_PROT_CFG + i * 4);
prot              876 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 		prot[i] &= ~MT_PROT_CFG_CTRL;
prot              878 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 			prot[i] &= ~MT_PROT_CFG_RATE;
prot              889 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 		prot[0] |= MT_PROT_CTRL_RTS_CTS;
prot              892 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 		prot[1] |= MT_PROT_CTRL_CTS2SELF;
prot              894 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 		prot[2] |= MT_PROT_RATE_CCK_11;
prot              895 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 		prot[3] |= MT_PROT_RATE_CCK_11;
prot              896 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 		prot[4] |= MT_PROT_RATE_CCK_11;
prot              897 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 		prot[5] |= MT_PROT_RATE_CCK_11;
prot              904 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 			prot[1] |= MT_PROT_CTRL_RTS_CTS;
prot              906 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 		prot[2] |= MT_PROT_RATE_OFDM_24;
prot              907 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 		prot[3] |= MT_PROT_RATE_DUP_OFDM_24;
prot              908 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 		prot[4] |= MT_PROT_RATE_OFDM_24;
prot              909 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 		prot[5] |= MT_PROT_RATE_DUP_OFDM_24;
prot              919 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 		prot[2] |= MT_PROT_CTRL_RTS_CTS;
prot              920 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 		prot[3] |= MT_PROT_CTRL_RTS_CTS;
prot              921 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 		prot[4] |= MT_PROT_CTRL_RTS_CTS;
prot              922 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 		prot[5] |= MT_PROT_CTRL_RTS_CTS;
prot              928 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 		prot[3] |= MT_PROT_CTRL_RTS_CTS;
prot              929 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 		prot[5] |= MT_PROT_CTRL_RTS_CTS;
prot              936 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 		prot[4] |= MT_PROT_CTRL_RTS_CTS;
prot              937 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 		prot[5] |= MT_PROT_CTRL_RTS_CTS;
prot              940 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 	for (i = 0; i < ARRAY_SIZE(prot); i++)
prot              941 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c 		mt76_wr(dev, MT_CCK_PROT_CFG + i * 4, prot[i]);
prot              205 drivers/net/wireless/mediatek/mt7601u/mac.c 	u32 prot[6];
prot              209 drivers/net/wireless/mediatek/mt7601u/mac.c 	prot[0] = MT_PROT_NAV_SHORT |
prot              212 drivers/net/wireless/mediatek/mt7601u/mac.c 	prot[1] = prot[0];
prot              214 drivers/net/wireless/mediatek/mt7601u/mac.c 		prot[1] |= MT_PROT_CTRL_CTS2SELF;
prot              216 drivers/net/wireless/mediatek/mt7601u/mac.c 	prot[2] = prot[4] = MT_PROT_NAV_SHORT | MT_PROT_TXOP_ALLOW_BW20;
prot              217 drivers/net/wireless/mediatek/mt7601u/mac.c 	prot[3] = prot[5] = MT_PROT_NAV_SHORT | MT_PROT_TXOP_ALLOW_ALL;
prot              220 drivers/net/wireless/mediatek/mt7601u/mac.c 		prot[2] |= MT_PROT_RATE_CCK_11;
prot              221 drivers/net/wireless/mediatek/mt7601u/mac.c 		prot[3] |= MT_PROT_RATE_CCK_11;
prot              222 drivers/net/wireless/mediatek/mt7601u/mac.c 		prot[4] |= MT_PROT_RATE_CCK_11;
prot              223 drivers/net/wireless/mediatek/mt7601u/mac.c 		prot[5] |= MT_PROT_RATE_CCK_11;
prot              225 drivers/net/wireless/mediatek/mt7601u/mac.c 		prot[2] |= MT_PROT_RATE_OFDM_24;
prot              226 drivers/net/wireless/mediatek/mt7601u/mac.c 		prot[3] |= MT_PROT_RATE_DUP_OFDM_24;
prot              227 drivers/net/wireless/mediatek/mt7601u/mac.c 		prot[4] |= MT_PROT_RATE_OFDM_24;
prot              228 drivers/net/wireless/mediatek/mt7601u/mac.c 		prot[5] |= MT_PROT_RATE_DUP_OFDM_24;
prot              253 drivers/net/wireless/mediatek/mt7601u/mac.c 			prot[i + 2] |= MT_PROT_CTRL_RTS_CTS;
prot              256 drivers/net/wireless/mediatek/mt7601u/mac.c 		mt7601u_wr(dev, MT_CCK_PROT_CFG + i * 4, prot[i]);
prot              757 drivers/power/supply/sbs-battery.c static void sbs_alert(struct i2c_client *client, enum i2c_alert_protocol prot,
prot              237 drivers/power/supply/sbs-manager.c static void sbsm_alert(struct i2c_client *client, enum i2c_alert_protocol prot,
prot              200 drivers/s390/net/qeth_core.h #define qeth_is_ipafunc_supported(c, prot, f) \
prot              201 drivers/s390/net/qeth_core.h 	 ((prot == QETH_PROT_IPV6) ? \
prot              203 drivers/s390/net/qeth_core.h #define qeth_is_ipafunc_enabled(c, prot, f) \
prot              204 drivers/s390/net/qeth_core.h 	 ((prot == QETH_PROT_IPV6) ? \
prot              910 drivers/s390/net/qeth_core.h 	__be16 prot = veth->h_vlan_proto;
prot              912 drivers/s390/net/qeth_core.h 	if (prot == htons(ETH_P_8021Q))
prot              913 drivers/s390/net/qeth_core.h 		prot = veth->h_vlan_encapsulated_proto;
prot              915 drivers/s390/net/qeth_core.h 	switch (prot) {
prot              981 drivers/s390/net/qeth_core.h 				      enum qeth_prot_versions prot);
prot             1038 drivers/s390/net/qeth_core.h 					   enum qeth_prot_versions prot,
prot             1047 drivers/s390/net/qeth_core.h 						 enum qeth_prot_versions prot);
prot             2748 drivers/s390/net/qeth_core_main.c 					   enum qeth_prot_versions prot,
prot             2771 drivers/s390/net/qeth_core_main.c 	hdr->prot_version = prot;
prot             2938 drivers/s390/net/qeth_core_main.c 				enum qeth_prot_versions prot)
prot             2943 drivers/s390/net/qeth_core_main.c 	QETH_CARD_TEXT_(card, 2, "qipassi%i", prot);
prot             2944 drivers/s390/net/qeth_core_main.c 	iob = qeth_ipa_alloc_cmd(card, IPA_CMD_QIPASSIST, prot, 0);
prot             5362 drivers/s390/net/qeth_core_main.c 						 enum qeth_prot_versions prot)
prot             5369 drivers/s390/net/qeth_core_main.c 	iob = qeth_ipa_alloc_cmd(card, IPA_CMD_SETASSPARMS, prot,
prot             5389 drivers/s390/net/qeth_core_main.c 				      enum qeth_prot_versions prot)
prot             5394 drivers/s390/net/qeth_core_main.c 	QETH_CARD_TEXT_(card, 4, "simassp%i", prot);
prot             5395 drivers/s390/net/qeth_core_main.c 	iob = qeth_get_setassparms_cmd(card, ipa_func, cmd_code, length, prot);
prot             5917 drivers/s390/net/qeth_core_main.c 			     enum qeth_prot_versions prot)
prot             5920 drivers/s390/net/qeth_core_main.c 						 NULL, prot);
prot             5924 drivers/s390/net/qeth_core_main.c 			    enum qeth_prot_versions prot)
prot             5933 drivers/s390/net/qeth_core_main.c 	if (IS_LAYER3(card) && prot == QETH_PROT_IPV4 &&
prot             5938 drivers/s390/net/qeth_core_main.c 				       prot);
prot             5947 drivers/s390/net/qeth_core_main.c 		qeth_set_csum_off(card, cstype, prot);
prot             5953 drivers/s390/net/qeth_core_main.c 				       prot);
prot             5955 drivers/s390/net/qeth_core_main.c 		qeth_set_csum_off(card, cstype, prot);
prot             5964 drivers/s390/net/qeth_core_main.c 		qeth_set_csum_off(card, cstype, prot);
prot             5970 drivers/s390/net/qeth_core_main.c 		qeth_set_csum_off(card, cstype, prot);
prot             5975 drivers/s390/net/qeth_core_main.c 		 cstype == IPA_INBOUND_CHECKSUM ? "in" : "out", prot);
prot             5985 drivers/s390/net/qeth_core_main.c 			     enum qeth_prot_versions prot)
prot             5987 drivers/s390/net/qeth_core_main.c 	return on ? qeth_set_csum_on(card, cstype, prot) :
prot             5988 drivers/s390/net/qeth_core_main.c 		    qeth_set_csum_off(card, cstype, prot);
prot             6006 drivers/s390/net/qeth_core_main.c 			    enum qeth_prot_versions prot)
prot             6009 drivers/s390/net/qeth_core_main.c 						 IPA_CMD_ASS_STOP, NULL, prot);
prot             6013 drivers/s390/net/qeth_core_main.c 			   enum qeth_prot_versions prot)
prot             6021 drivers/s390/net/qeth_core_main.c 				       IPA_CMD_ASS_START, 0, prot);
prot             6030 drivers/s390/net/qeth_core_main.c 		qeth_set_tso_off(card, prot);
prot             6036 drivers/s390/net/qeth_core_main.c 				       SETASS_DATA_SIZEOF(caps), prot);
prot             6038 drivers/s390/net/qeth_core_main.c 		qeth_set_tso_off(card, prot);
prot             6047 drivers/s390/net/qeth_core_main.c 		qeth_set_tso_off(card, prot);
prot             6053 drivers/s390/net/qeth_core_main.c 		qeth_set_tso_off(card, prot);
prot             6057 drivers/s390/net/qeth_core_main.c 	dev_info(&card->gdev->dev, "TSOv%u enabled (MSS: %u)\n", prot,
prot             6063 drivers/s390/net/qeth_core_main.c 			    enum qeth_prot_versions prot)
prot             6065 drivers/s390/net/qeth_core_main.c 	return on ? qeth_set_tso_on(card, prot) : qeth_set_tso_off(card, prot);
prot               67 drivers/s390/net/qeth_l3_main.c static struct qeth_ipaddr *qeth_l3_get_addr_buffer(enum qeth_prot_versions prot)
prot               72 drivers/s390/net/qeth_l3_main.c 		qeth_l3_init_ipaddr(addr, QETH_IP_TYPE_NORMAL, prot);
prot              458 drivers/s390/net/qeth_l3_main.c 	enum qeth_routing_types type, enum qeth_prot_versions prot)
prot              465 drivers/s390/net/qeth_l3_main.c 	iob = qeth_ipa_alloc_cmd(card, IPA_CMD_SETRTG, prot,
prot              477 drivers/s390/net/qeth_l3_main.c 		enum qeth_routing_types *type, enum qeth_prot_versions prot)
prot              496 drivers/s390/net/qeth_l3_main.c 			if (qeth_is_ipafunc_supported(card, prot,
prot             1324 drivers/s390/net/qeth_l3_main.c 		u16 prot = (hdr->hdr.l3.flags & QETH_HDR_IPV6) ? ETH_P_IPV6 :
prot             1331 drivers/s390/net/qeth_l3_main.c 			if (prot == ETH_P_IP)
prot             1348 drivers/s390/net/qeth_l3_main.c 			card->dev->header_ops->create(skb, card->dev, prot,
prot             1352 drivers/s390/net/qeth_l3_main.c 			card->dev->header_ops->create(skb, card->dev, prot,
prot             1598 drivers/s390/net/qeth_l3_main.c static int arpentry_matches_prot(struct qeth_arp_entrytype *type, __u16 prot)
prot             1600 drivers/s390/net/qeth_l3_main.c 	return (type->ip == QETHARP_IP_ADDR_V4 && prot == QETH_PROT_IPV4) ||
prot             1601 drivers/s390/net/qeth_l3_main.c 		(type->ip == QETHARP_IP_ADDR_V6 && prot == QETH_PROT_IPV6);
prot             1684 drivers/s390/net/qeth_l3_main.c 	enum qeth_prot_versions prot,
prot             1691 drivers/s390/net/qeth_l3_main.c 	QETH_CARD_TEXT_(card, 3, "qarpipv%i", prot);
prot             1695 drivers/s390/net/qeth_l3_main.c 				       SETASS_DATA_SIZEOF(query_arp), prot);
prot               70 drivers/s390/net/qeth_l3_sys.c 		struct qeth_routing_info *route, enum qeth_prot_versions prot,
prot               95 drivers/s390/net/qeth_l3_sys.c 		if (prot == QETH_PROT_IPV4)
prot               97 drivers/s390/net/qeth_l3_sys.c 		else if (prot == QETH_PROT_IPV6)
prot             1218 drivers/scsi/cxgbi/libcxgbi.c 		unsigned int prot)
prot             1220 drivers/scsi/cxgbi/libcxgbi.c 	struct scsi_data_buffer *sdb = prot ? scsi_prot(sc) : &sc->sdb;
prot              526 drivers/scsi/hisi_sas/hisi_sas.h 			u32 prot[7];
prot             1110 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 			    struct hisi_sas_protect_iu_v3_hw *prot)
prot             1118 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		prot->dw0 |= T10_INSRT_EN_MSK;
prot             1119 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		prot->lbrtgv = lbrt_chk_val;
prot             1122 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		prot->dw0 |= (T10_RMV_EN_MSK | T10_CHK_EN_MSK);
prot             1123 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		prot->lbrtcv = lbrt_chk_val;
prot             1124 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		prot->dw4 |= get_prot_chk_msk_v3_hw(scsi_cmnd);
prot             1127 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		prot->dw0 |= T10_CHK_EN_MSK;
prot             1128 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		prot->lbrtcv = lbrt_chk_val;
prot             1129 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		prot->dw4 |= get_prot_chk_msk_v3_hw(scsi_cmnd);
prot             1132 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		prot->dw0 |= T10_INSRT_EN_MSK;
prot             1133 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		prot->lbrtgv = lbrt_chk_val;
prot             1136 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		prot->dw0 |= (T10_RMV_EN_MSK | T10_CHK_EN_MSK);
prot             1137 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		prot->lbrtcv = lbrt_chk_val;
prot             1140 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		prot->dw0 |= T10_CHK_EN_MSK;
prot             1141 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		prot->lbrtcv = lbrt_chk_val;
prot             1142 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		prot->dw4 |= get_prot_chk_msk_v3_hw(scsi_cmnd);
prot             1153 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		prot->dw0 |= (0x1 << USR_DATA_BLOCK_SZ_OFF);
prot             1156 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		prot->dw0 |= (0x2 << USR_DATA_BLOCK_SZ_OFF);
prot             1164 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 	prot->dw0 |= INCR_LBRT_MSK;
prot             1256 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		struct hisi_sas_protect_iu_v3_hw prot;
prot             1265 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		memset(&prot, 0, sizeof(struct hisi_sas_protect_iu_v3_hw));
prot             1266 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		fill_prot_v3_hw(scsi_cmnd, &prot);
prot             1267 drivers/scsi/hisi_sas/hisi_sas_v3_hw.c 		memcpy(buf_cmd_prot, &prot,
prot             2721 drivers/scsi/lpfc/lpfc_scsi.c 	int prot, protsegcnt;
prot             2732 drivers/scsi/lpfc/lpfc_scsi.c 	prot = scsi_get_prot_op(cmd);
prot             2733 drivers/scsi/lpfc/lpfc_scsi.c 	if ((prot == SCSI_PROT_READ_STRIP) ||
prot             2734 drivers/scsi/lpfc/lpfc_scsi.c 	    (prot == SCSI_PROT_WRITE_INSERT) ||
prot             2735 drivers/scsi/lpfc/lpfc_scsi.c 	    (prot == SCSI_PROT_NORMAL))
prot             2826 drivers/scsi/qla2xxx/qla_attr.c 			int prot = 0, guard;
prot             2832 drivers/scsi/qla2xxx/qla_attr.c 				prot = SHOST_DIX_TYPE0_PROTECTION;
prot             2834 drivers/scsi/qla2xxx/qla_attr.c 			    prot | SHOST_DIF_TYPE1_PROTECTION
prot             3351 drivers/scsi/qla2xxx/qla_os.c 			int prot = 0, guard;
prot             3357 drivers/scsi/qla2xxx/qla_os.c 				prot = SHOST_DIX_TYPE0_PROTECTION;
prot             3362 drivers/scsi/qla2xxx/qla_os.c 				    prot | SHOST_DIF_TYPE1_PROTECTION
prot             1137 drivers/scsi/scsi_lib.c 	void *prot = cmd->prot_sdb;
prot             1156 drivers/scsi/scsi_lib.c 	cmd->prot_sdb = prot;
prot              949 drivers/slimbus/qcom-ngd-ctrl.c 			wbuf[txn.msg->num_bytes++] = exp << 4 | rt->prot;
prot              951 drivers/slimbus/qcom-ngd-ctrl.c 			if (rt->prot == SLIM_PROTO_ISO)
prot              337 drivers/slimbus/slimbus.h 	enum slim_transport_protocol prot;
prot              230 drivers/slimbus/stream.c 			rt->prot = SLIM_PROTO_PUSH;
prot              232 drivers/slimbus/stream.c 			rt->prot = SLIM_PROTO_PULL;
prot              234 drivers/slimbus/stream.c 		rt->prot = SLIM_PROTO_ISO;
prot              276 drivers/slimbus/stream.c 	if (stream->prot != SLIM_PROTO_ISO)
prot              311 drivers/slimbus/stream.c 	wbuf[2] = (stream->prot << 4) | ((port->ch.seg_dist & 0xF00) >> 8);
prot              312 drivers/slimbus/stream.c 	if (stream->prot == SLIM_PROTO_ISO)
prot               47 drivers/soc/aspeed/aspeed-lpc-ctrl.c 	pgprot_t prot = vma->vm_page_prot;
prot               53 drivers/soc/aspeed/aspeed-lpc-ctrl.c 	prot = pgprot_noncached(prot);
prot               57 drivers/soc/aspeed/aspeed-lpc-ctrl.c 		vsize, prot))
prot              103 drivers/soc/aspeed/aspeed-p2a-ctrl.c 	pgprot_t prot;
prot              111 drivers/soc/aspeed/aspeed-p2a-ctrl.c 	prot = vma->vm_page_prot;
prot              117 drivers/soc/aspeed/aspeed-p2a-ctrl.c 	prot = pgprot_noncached(prot);
prot              121 drivers/soc/aspeed/aspeed-p2a-ctrl.c 		vsize, prot))
prot              347 drivers/staging/android/ashmem.c static inline vm_flags_t calc_vm_may_flags(unsigned long prot)
prot              349 drivers/staging/android/ashmem.c 	return _calc_vm_trans(prot, PROT_READ,  VM_MAYREAD) |
prot              350 drivers/staging/android/ashmem.c 	       _calc_vm_trans(prot, PROT_WRITE, VM_MAYWRITE) |
prot              351 drivers/staging/android/ashmem.c 	       _calc_vm_trans(prot, PROT_EXEC,  VM_MAYEXEC);
prot              527 drivers/staging/android/ashmem.c static int set_prot_mask(struct ashmem_area *asma, unsigned long prot)
prot              534 drivers/staging/android/ashmem.c 	if ((asma->prot_mask & prot) != prot) {
prot              540 drivers/staging/android/ashmem.c 	if ((prot & PROT_READ) && (current->personality & READ_IMPLIES_EXEC))
prot              541 drivers/staging/android/ashmem.c 		prot |= PROT_EXEC;
prot              543 drivers/staging/android/ashmem.c 	asma->prot_mask = prot;
prot              499 drivers/target/target_core_file.c 	sector_t prot_length, prot;
prot              509 drivers/target/target_core_file.c 	for (prot = 0; prot < prot_length;) {
prot              510 drivers/target/target_core_file.c 		sector_t len = min_t(sector_t, bufsize, prot_length - prot);
prot              517 drivers/target/target_core_file.c 		prot += ret;
prot             1914 drivers/vfio/vfio.c 		   int prot, unsigned long *phys_pfn)
prot             1939 drivers/vfio/vfio.c 					     npage, prot, phys_pfn);
prot               79 drivers/vfio/vfio_iommu_type1.c 	int			prot;		/* IOMMU_CACHE */
prot               88 drivers/vfio/vfio_iommu_type1.c 	int			prot;		/* IOMMU_READ/WRITE */
prot              127 drivers/vfio/vfio_iommu_type1.c static int put_pfn(unsigned long pfn, int prot);
prot              262 drivers/vfio/vfio_iommu_type1.c 		ret = put_pfn(vpfn->pfn, dma->prot);
prot              326 drivers/vfio/vfio_iommu_type1.c static int put_pfn(unsigned long pfn, int prot)
prot              330 drivers/vfio/vfio_iommu_type1.c 		if (prot & IOMMU_WRITE)
prot              339 drivers/vfio/vfio_iommu_type1.c 			 int prot, unsigned long *pfn)
prot              347 drivers/vfio/vfio_iommu_type1.c 	if (prot & IOMMU_WRITE)
prot              410 drivers/vfio/vfio_iommu_type1.c 	ret = vaddr_get_pfn(current->mm, vaddr, dma->prot, pfn_base);
prot              423 drivers/vfio/vfio_iommu_type1.c 			put_pfn(*pfn_base, dma->prot);
prot              437 drivers/vfio/vfio_iommu_type1.c 		ret = vaddr_get_pfn(current->mm, vaddr, dma->prot, &pfn);
prot              443 drivers/vfio/vfio_iommu_type1.c 			put_pfn(pfn, dma->prot);
prot              450 drivers/vfio/vfio_iommu_type1.c 				put_pfn(pfn, dma->prot);
prot              467 drivers/vfio/vfio_iommu_type1.c 				put_pfn(pfn, dma->prot);
prot              484 drivers/vfio/vfio_iommu_type1.c 		if (put_pfn(pfn++, dma->prot)) {
prot              507 drivers/vfio/vfio_iommu_type1.c 	ret = vaddr_get_pfn(mm, vaddr, dma->prot, pfn_base);
prot              511 drivers/vfio/vfio_iommu_type1.c 			put_pfn(*pfn_base, dma->prot);
prot              543 drivers/vfio/vfio_iommu_type1.c 				      int npage, int prot,
prot              585 drivers/vfio/vfio_iommu_type1.c 		if ((dma->prot & prot) != prot) {
prot              986 drivers/vfio/vfio_iommu_type1.c 			  unsigned long pfn, long npage, int prot)
prot              993 drivers/vfio/vfio_iommu_type1.c 				npage << PAGE_SHIFT, prot | d->prot);
prot             1031 drivers/vfio/vfio_iommu_type1.c 				     dma->prot);
prot             1077 drivers/vfio/vfio_iommu_type1.c 	int ret = 0, prot = 0;
prot             1091 drivers/vfio/vfio_iommu_type1.c 		prot |= IOMMU_WRITE;
prot             1093 drivers/vfio/vfio_iommu_type1.c 		prot |= IOMMU_READ;
prot             1095 drivers/vfio/vfio_iommu_type1.c 	if (!prot || !size || (size | iova | vaddr) & mask)
prot             1128 drivers/vfio/vfio_iommu_type1.c 	dma->prot = prot;
prot             1251 drivers/vfio/vfio_iommu_type1.c 					size, dma->prot | domain->prot);
prot             1282 drivers/vfio/vfio_iommu_type1.c 			IOMMU_READ | IOMMU_WRITE | domain->prot);
prot             1793 drivers/vfio/vfio_iommu_type1.c 		domain->prot |= IOMMU_CACHE;
prot             1804 drivers/vfio/vfio_iommu_type1.c 		    d->prot == domain->prot) {
prot             2137 drivers/vfio/vfio_iommu_type1.c 		if (!(domain->prot & IOMMU_CACHE)) {
prot              992 drivers/video/fbdev/vermilion/vermilion.c 	unsigned long prot;
prot              998 drivers/video/fbdev/vermilion/vermilion.c 	prot = pgprot_val(vma->vm_page_prot) & ~_PAGE_CACHE_MASK;
prot             1000 drivers/video/fbdev/vermilion/vermilion.c 		prot | cachemode2protval(_PAGE_CACHE_MODE_UC_MINUS);
prot               67 drivers/xen/xlate_mmu.c 	pgprot_t prot;
prot              100 drivers/xen/xlate_mmu.c 	pte_t pte = pte_mkspecial(pfn_pte(page_to_pfn(page), info->prot));
prot              146 drivers/xen/xlate_mmu.c 			      int *err_ptr, pgprot_t prot,
prot              160 drivers/xen/xlate_mmu.c 	data.prot  = prot;
prot              269 drivers/xen/xlate_mmu.c 	pgprot_t prot;
prot              277 drivers/xen/xlate_mmu.c 	pte_t pte = pte_mkspecial(pfn_pte(page_to_pfn(page), r->prot));
prot              291 drivers/xen/xlate_mmu.c 		.prot = vma->vm_page_prot,
prot              143 fs/affs/affs.h extern umode_t	affs_prot_to_mode(u32 prot);
prot              391 fs/affs/amigaffs.c affs_prot_to_mode(u32 prot)
prot              395 fs/affs/amigaffs.c 	if (!(prot & FIBF_NOWRITE))
prot              397 fs/affs/amigaffs.c 	if (!(prot & FIBF_NOREAD))
prot              399 fs/affs/amigaffs.c 	if (!(prot & FIBF_NOEXECUTE))
prot              401 fs/affs/amigaffs.c 	if (prot & FIBF_GRP_WRITE)
prot              403 fs/affs/amigaffs.c 	if (prot & FIBF_GRP_READ)
prot              405 fs/affs/amigaffs.c 	if (prot & FIBF_GRP_EXECUTE)
prot              407 fs/affs/amigaffs.c 	if (prot & FIBF_OTR_WRITE)
prot              409 fs/affs/amigaffs.c 	if (prot & FIBF_OTR_READ)
prot              411 fs/affs/amigaffs.c 	if (prot & FIBF_OTR_EXECUTE)
prot              420 fs/affs/amigaffs.c 	u32 prot = AFFS_I(inode)->i_protect;
prot              424 fs/affs/amigaffs.c 		prot |= FIBF_NOEXECUTE;
prot              426 fs/affs/amigaffs.c 		prot |= FIBF_NOREAD;
prot              428 fs/affs/amigaffs.c 		prot |= FIBF_NOWRITE;
prot              430 fs/affs/amigaffs.c 		prot |= FIBF_GRP_EXECUTE;
prot              432 fs/affs/amigaffs.c 		prot |= FIBF_GRP_READ;
prot              434 fs/affs/amigaffs.c 		prot |= FIBF_GRP_WRITE;
prot              436 fs/affs/amigaffs.c 		prot |= FIBF_OTR_EXECUTE;
prot              438 fs/affs/amigaffs.c 		prot |= FIBF_OTR_READ;
prot              440 fs/affs/amigaffs.c 		prot |= FIBF_OTR_WRITE;
prot              442 fs/affs/amigaffs.c 	AFFS_I(inode)->i_protect = prot;
prot               26 fs/affs/inode.c 	u32			 prot;
prot               51 fs/affs/inode.c 	prot = be32_to_cpu(tail->protect);
prot               58 fs/affs/inode.c 	AFFS_I(inode)->i_protect = prot;
prot               74 fs/affs/inode.c 		inode->i_mode = affs_prot_to_mode(prot);
prot              102 fs/binfmt_elf.c static int set_brk(unsigned long start, unsigned long end, int prot)
prot              113 fs/binfmt_elf.c 				prot & PROT_EXEC ? VM_EXEC : 0);
prot              349 fs/binfmt_elf.c 		const struct elf_phdr *eppnt, int prot, int type,
prot              373 fs/binfmt_elf.c 		map_addr = vm_mmap(filep, addr, total_size, prot, type, off);
prot              377 fs/binfmt_elf.c 		map_addr = vm_mmap(filep, addr, size, prot, type, off);
prot              530 fs/binfmt_elf.c 	int prot = 0;
prot              533 fs/binfmt_elf.c 		prot |= PROT_READ;
prot              535 fs/binfmt_elf.c 		prot |= PROT_WRITE;
prot              537 fs/binfmt_elf.c 		prot |= PROT_EXEC;
prot              538 fs/binfmt_elf.c 	return prot;
prot             1023 fs/binfmt_elf_fdpic.c 		int prot = 0, flags;
prot             1035 fs/binfmt_elf_fdpic.c 		if (phdr->p_flags & PF_R) prot |= PROT_READ;
prot             1036 fs/binfmt_elf_fdpic.c 		if (phdr->p_flags & PF_W) prot |= PROT_WRITE;
prot             1037 fs/binfmt_elf_fdpic.c 		if (phdr->p_flags & PF_X) prot |= PROT_EXEC;
prot             1083 fs/binfmt_elf_fdpic.c 		maddr = vm_mmap(file, maddr, phdr->p_memsz + disp, prot, flags,
prot             1087 fs/binfmt_elf_fdpic.c 		       loop, phdr->p_memsz + disp, prot, flags,
prot             1107 fs/binfmt_elf_fdpic.c 		if (prot & PROT_WRITE && disp > 0) {
prot             1129 fs/binfmt_elf_fdpic.c 					 prot, flags, 0);
prot             1133 fs/binfmt_elf_fdpic.c 			       loop, xaddr, excess - excess1, prot, flags,
prot             1140 fs/binfmt_elf_fdpic.c 		if (prot & PROT_WRITE && excess1 > 0) {
prot              189 fs/proc/vmcore.c 				  unsigned long size, pgprot_t prot)
prot              191 fs/proc/vmcore.c 	prot = pgprot_encrypted(prot);
prot              192 fs/proc/vmcore.c 	return remap_pfn_range(vma, from, pfn, size, prot);
prot              489 fs/proc/vmcore.c 				    unsigned long size, pgprot_t prot)
prot              511 fs/proc/vmcore.c 							   prot))
prot              518 fs/proc/vmcore.c 						   PAGE_SIZE, prot))
prot              528 fs/proc/vmcore.c 					   map_size, prot))
prot              539 fs/proc/vmcore.c 			    unsigned long size, pgprot_t prot)
prot              546 fs/proc/vmcore.c 		return remap_oldmem_pfn_checked(vma, from, pfn, size, prot);
prot              548 fs/proc/vmcore.c 		return remap_oldmem_pfn_range(vma, from, pfn, size, prot);
prot              407 fs/pstore/ram_core.c 	pgprot_t prot;
prot              415 fs/pstore/ram_core.c 		prot = pgprot_noncached(PAGE_KERNEL);
prot              417 fs/pstore/ram_core.c 		prot = pgprot_writecombine(PAGE_KERNEL);
prot              430 fs/pstore/ram_core.c 	vaddr = vmap(pages, page_count, VM_MAP, prot);
prot              482 include/asm-generic/pgtable.h #define move_pte(pte, prot, old_addr, new_addr)	(pte)
prot              494 include/asm-generic/pgtable.h #define pgprot_noncached(prot)	(prot)
prot              672 include/asm-generic/pgtable.h #define pgprot_encrypted(prot)	(prot)
prot              676 include/asm-generic/pgtable.h #define pgprot_decrypted(prot)	(prot)
prot              805 include/asm-generic/pgtable.h static inline int track_pfn_remap(struct vm_area_struct *vma, pgprot_t *prot,
prot              816 include/asm-generic/pgtable.h static inline void track_pfn_insert(struct vm_area_struct *vma, pgprot_t *prot,
prot              847 include/asm-generic/pgtable.h extern int track_pfn_remap(struct vm_area_struct *vma, pgprot_t *prot,
prot              850 include/asm-generic/pgtable.h extern void track_pfn_insert(struct vm_area_struct *vma, pgprot_t *prot,
prot             1042 include/asm-generic/pgtable.h int p4d_set_huge(p4d_t *p4d, phys_addr_t addr, pgprot_t prot);
prot             1045 include/asm-generic/pgtable.h static inline int p4d_set_huge(p4d_t *p4d, phys_addr_t addr, pgprot_t prot)
prot             1055 include/asm-generic/pgtable.h int pud_set_huge(pud_t *pud, phys_addr_t addr, pgprot_t prot);
prot             1056 include/asm-generic/pgtable.h int pmd_set_huge(pmd_t *pmd, phys_addr_t addr, pgprot_t prot);
prot             1063 include/asm-generic/pgtable.h static inline int p4d_set_huge(p4d_t *p4d, phys_addr_t addr, pgprot_t prot)
prot             1067 include/asm-generic/pgtable.h static inline int pud_set_huge(pud_t *pud, phys_addr_t addr, pgprot_t prot)
prot             1071 include/asm-generic/pgtable.h static inline int pmd_set_huge(pmd_t *pmd, phys_addr_t addr, pgprot_t prot)
prot             1131 include/asm-generic/pgtable.h static inline bool pfn_modify_allowed(unsigned long pfn, pgprot_t prot)
prot               15 include/asm-generic/syscalls.h 			unsigned long prot, unsigned long flags,
prot               21 include/asm-generic/syscalls.h 			unsigned long prot, unsigned long flags,
prot              737 include/drm/ttm/ttm_bo_api.h void *ttm_kmap_atomic_prot(struct page *page, pgprot_t prot);
prot              739 include/drm/ttm/ttm_bo_api.h void ttm_kunmap_atomic_prot(void *addr, pgprot_t prot);
prot               56 include/linux/agpgart.h 	int prot;		/* prot flags for mmap          */
prot               62 include/linux/agpgart.h 	pgprot_t prot;
prot               41 include/linux/can/core.h 	struct proto *prot;
prot               25 include/linux/crash_dump.h 				  unsigned long size, pgprot_t prot);
prot              630 include/linux/dma-mapping.h 			pgprot_t prot, const void *caller);
prot              633 include/linux/dma-mapping.h 			pgprot_t prot, const void *caller);
prot               55 include/linux/dma-noncoherent.h #define pgprot_dmacoherent(prot)	pgprot_noncached(prot)
prot               58 include/linux/dma-noncoherent.h pgprot_t dma_pgprot(struct device *dev, pgprot_t prot, unsigned long attrs);
prot               60 include/linux/dma-noncoherent.h static inline pgprot_t dma_pgprot(struct device *dev, pgprot_t prot,
prot               63 include/linux/dma-noncoherent.h 	return prot;	/* no protection bits supported without page tables */
prot               97 include/linux/highmem.h #define kmap_atomic_prot(page, prot)	kmap_atomic(page)
prot              268 include/linux/huge_mm.h #define mk_huge_pmd(page, prot) pmd_mkhuge(mk_pmd(page, prot))
prot               20 include/linux/ima.h extern int ima_file_mmap(struct file *file, unsigned long prot);
prot               67 include/linux/ima.h static inline int ima_file_mmap(struct file *file, unsigned long prot)
prot               25 include/linux/io-mapping.h 	pgprot_t prot;
prot               44 include/linux/io-mapping.h 	pgprot_t prot;
prot               46 include/linux/io-mapping.h 	if (iomap_create_wc(base, size, &prot))
prot               51 include/linux/io-mapping.h 	iomap->prot = prot;
prot               72 include/linux/io-mapping.h 	return iomap_atomic_prot_pfn(pfn, mapping->prot);
prot              115 include/linux/io-mapping.h 	iomap->prot = pgprot_noncached_wc(PAGE_KERNEL);
prot              117 include/linux/io-mapping.h 	iomap->prot = pgprot_writecombine(PAGE_KERNEL);
prot              119 include/linux/io-mapping.h 	iomap->prot = pgprot_noncached(PAGE_KERNEL);
prot              139 include/linux/io-pgtable.h 		   phys_addr_t paddr, size_t size, int prot);
prot               25 include/linux/io.h 		       phys_addr_t phys_addr, pgprot_t prot);
prot               28 include/linux/io.h 				     phys_addr_t phys_addr, pgprot_t prot)
prot              161 include/linux/iommu.h 	int			prot;
prot              259 include/linux/iommu.h 		   phys_addr_t paddr, size_t size, int prot);
prot              284 include/linux/iommu.h 				    phys_addr_t paddr, u64 size, int prot);
prot              423 include/linux/iommu.h 		     phys_addr_t paddr, size_t size, int prot);
prot              430 include/linux/iommu.h 			   struct scatterlist *sg,unsigned int nents, int prot);
prot              443 include/linux/iommu.h iommu_alloc_resv_region(phys_addr_t start, size_t length, int prot,
prot              493 include/linux/iommu.h 				      int prot);
prot              660 include/linux/iommu.h 			    phys_addr_t paddr, size_t size, int prot)
prot              680 include/linux/iommu.h 				  unsigned int nents, int prot)
prot              696 include/linux/iommu.h 					     u64 size, int prot)
prot             1029 include/linux/libata.h static inline bool ata_is_atapi(u8 prot)
prot             1031 include/linux/libata.h 	return prot & ATA_PROT_FLAG_ATAPI;
prot             1034 include/linux/libata.h static inline bool ata_is_pio(u8 prot)
prot             1036 include/linux/libata.h 	return prot & ATA_PROT_FLAG_PIO;
prot             1039 include/linux/libata.h static inline bool ata_is_dma(u8 prot)
prot             1041 include/linux/libata.h 	return prot & ATA_PROT_FLAG_DMA;
prot             1044 include/linux/libata.h static inline bool ata_is_ncq(u8 prot)
prot             1046 include/linux/libata.h 	return prot & ATA_PROT_FLAG_NCQ;
prot             1049 include/linux/libata.h static inline bool ata_is_data(u8 prot)
prot             1051 include/linux/libata.h 	return prot & (ATA_PROT_FLAG_PIO | ATA_PROT_FLAG_DMA);
prot             1607 include/linux/lsm_hooks.h 				unsigned long prot, unsigned long flags);
prot             1609 include/linux/lsm_hooks.h 				unsigned long prot);
prot             1160 include/linux/mlx4/device.h 			int block_mcast_loopback, enum mlx4_protocol prot);
prot             1162 include/linux/mlx4/device.h 			enum mlx4_protocol prot);
prot             1475 include/linux/mm.h 		unsigned int flags, unsigned long *prot, resource_size_t *phys);
prot             2346 include/linux/mm.h 	unsigned long len, unsigned long prot, unsigned long flags,
prot             2356 include/linux/mm.h 	unsigned long len, unsigned long prot, unsigned long flags,
prot             2360 include/linux/mm.h 	return do_mmap(file, addr, len, prot, flags, 0, pgoff, populate, uf);
prot               81 include/linux/mman.h #define arch_calc_vm_prot_bits(prot, pkey) 0
prot               95 include/linux/mman.h static inline bool arch_validate_prot(unsigned long prot, unsigned long addr)
prot               97 include/linux/mman.h 	return (prot & ~(PROT_READ | PROT_WRITE | PROT_EXEC | PROT_SEM)) == 0;
prot              117 include/linux/mman.h calc_vm_prot_bits(unsigned long prot, unsigned long pkey)
prot              119 include/linux/mman.h 	return _calc_vm_trans(prot, PROT_READ,  VM_READ ) |
prot              120 include/linux/mman.h 	       _calc_vm_trans(prot, PROT_WRITE, VM_WRITE) |
prot              121 include/linux/mman.h 	       _calc_vm_trans(prot, PROT_EXEC,  VM_EXEC) |
prot              122 include/linux/mman.h 	       arch_calc_vm_prot_bits(prot, pkey);
prot               12 include/linux/pkeys.h #define arch_override_mprotect_pkey(vma, prot, pkey) (0)
prot              152 include/linux/security.h 			 unsigned long prot, unsigned long flags);
prot              363 include/linux/security.h int security_mmap_file(struct file *file, unsigned long prot,
prot              367 include/linux/security.h 			   unsigned long prot);
prot              908 include/linux/security.h static inline int security_mmap_file(struct file *file, unsigned long prot,
prot              921 include/linux/security.h 					 unsigned long prot)
prot              374 include/linux/sunrpc/svc.h 	u32			prot;	/* protocol (UDP or TCP) */
prot               66 include/linux/sunrpc/svcsock.h struct svc_xprt *svc_sock_create(struct svc_serv *serv, int prot);
prot              193 include/linux/sunrpc/xprt.h 	int			prot;		/* IP protocol */
prot              868 include/linux/syscalls.h 				unsigned long prot);
prot              878 include/linux/syscalls.h 			unsigned long prot, unsigned long pgoff,
prot              984 include/linux/syscalls.h 				  unsigned long prot, int pkey);
prot             1214 include/linux/syscalls.h 			unsigned long prot, unsigned long flags,
prot             1268 include/linux/syscalls.h 			      unsigned long prot, unsigned long flags,
prot               76 include/linux/vfio.h 				     int npage, int prot,
prot              106 include/linux/vfio.h 			  int npage, int prot, unsigned long *phys_pfn);
prot               78 include/linux/vmalloc.h 				int node, pgprot_t prot);
prot               99 include/linux/vmalloc.h extern void *__vmalloc(unsigned long size, gfp_t gfp_mask, pgprot_t prot);
prot              102 include/linux/vmalloc.h 			pgprot_t prot, unsigned long vm_flags, int node,
prot              120 include/linux/vmalloc.h 			unsigned long flags, pgprot_t prot);
prot              158 include/linux/vmalloc.h extern int map_vm_area(struct vm_struct *area, pgprot_t prot,
prot              162 include/linux/vmalloc.h 				    pgprot_t prot, struct page **pages);
prot              175 include/linux/vmalloc.h 			pgprot_t prot, struct page **pages)
prot               99 include/net/llc_conn.h 			  struct proto *prot, int kern);
prot               94 include/net/phonet/phonet.h 	struct proto		*prot;
prot               88 include/net/protocol.h 	struct proto	 *prot;
prot              105 include/net/protocol.h int inet_add_protocol(const struct net_protocol *prot, unsigned char num);
prot              106 include/net/protocol.h int inet_del_protocol(const struct net_protocol *prot, unsigned char num);
prot              107 include/net/protocol.h int inet_add_offload(const struct net_offload *prot, unsigned char num);
prot              108 include/net/protocol.h int inet_del_offload(const struct net_offload *prot, unsigned char num);
prot              113 include/net/protocol.h int inet6_add_protocol(const struct inet6_protocol *prot, unsigned char num);
prot              114 include/net/protocol.h int inet6_del_protocol(const struct inet6_protocol *prot, unsigned char num);
prot              118 include/net/protocol.h int inet6_add_offload(const struct net_offload *prot, unsigned char num);
prot              119 include/net/protocol.h int inet6_del_offload(const struct net_offload *prot, unsigned char num);
prot             1189 include/net/sock.h int proto_register(struct proto *prot, int alloc_slab);
prot             1190 include/net/sock.h void proto_unregister(struct proto *prot);
prot             1306 include/net/sock.h proto_sockets_allocated_sum_positive(struct proto *prot)
prot             1308 include/net/sock.h 	return percpu_counter_sum_positive(prot->sockets_allocated);
prot             1312 include/net/sock.h proto_memory_allocated(struct proto *prot)
prot             1314 include/net/sock.h 	return atomic_long_read(prot->memory_allocated);
prot             1318 include/net/sock.h proto_memory_pressure(struct proto *prot)
prot             1320 include/net/sock.h 	if (!prot->memory_pressure)
prot             1322 include/net/sock.h 	return !!*prot->memory_pressure;
prot             1328 include/net/sock.h void sock_prot_inuse_add(struct net *net, struct proto *prot, int inc);
prot             1332 include/net/sock.h static inline void sock_prot_inuse_add(struct net *net, struct proto *prot,
prot             1603 include/net/sock.h 		      struct proto *prot, int kern);
prot              492 include/net/tls.h 					 struct tls_prot_info *prot,
prot              495 include/net/tls.h 	if (tls_bigint_increment(ctx->rec_seq, prot->rec_seq_size))
prot              498 include/net/tls.h 	if (prot->version != TLS_1_3_VERSION)
prot              500 include/net/tls.h 				     prot->iv_size);
prot              509 include/net/tls.h 	struct tls_prot_info *prot = &ctx->prot_info;
prot              510 include/net/tls.h 	size_t pkt_len, iv_size = prot->iv_size;
prot              512 include/net/tls.h 	pkt_len = plaintext_len + prot->tag_size;
prot             2797 include/rdma/ib_verbs.h 		      unsigned long pfn, unsigned long size, pgprot_t prot);
prot             2802 include/rdma/ib_verbs.h 				    pgprot_t prot)
prot               94 include/trace/events/sock.h 	TP_PROTO(struct sock *sk, struct proto *prot, long allocated, int kind),
prot               96 include/trace/events/sock.h 	TP_ARGS(sk, prot, allocated, kind),
prot              111 include/trace/events/sock.h 		strncpy(__entry->name, prot->name, 32);
prot              112 include/trace/events/sock.h 		__entry->sysctl_mem = prot->sysctl_mem;
prot              114 include/trace/events/sock.h 		__entry->sysctl_rmem = sk_get_rmem0(sk, prot);
prot              116 include/trace/events/sock.h 		__entry->sysctl_wmem = sk_get_wmem0(sk, prot);
prot               83 include/uapi/linux/agpgart.h 	int prot;			/* prot flags for mmap          */
prot              123 include/uapi/linux/scif_ioctl.h 	__s32		prot;
prot               66 include/xen/xen-ops.h 		  xen_pfn_t *pfn, int nr, int *err_ptr, pgprot_t prot,
prot               71 include/xen/xen-ops.h 				pgprot_t prot,  unsigned int domid,
prot               85 include/xen/xen-ops.h 			      int *err_ptr, pgprot_t prot,
prot               98 include/xen/xen-ops.h 					    int *err_ptr, pgprot_t prot,
prot              135 include/xen/xen-ops.h 					     int *err_ptr, pgprot_t prot,
prot              141 include/xen/xen-ops.h 						 prot, domid, pages);
prot              148 include/xen/xen-ops.h 	return xen_remap_pfn(vma, addr, gfn, nr, err_ptr, prot, domid,
prot              172 include/xen/xen-ops.h 					     pgprot_t prot, unsigned int domid,
prot              178 include/xen/xen-ops.h 	return xen_remap_pfn(vma, addr, mfn, nr, err_ptr, prot, domid,
prot              197 include/xen/xen-ops.h 					     pgprot_t prot, unsigned int domid,
prot              203 include/xen/xen-ops.h 	return xen_remap_pfn(vma, addr, &gfn, nr, NULL, prot, domid, false,
prot             1427 ipc/shm.c      	unsigned long prot;
prot             1462 ipc/shm.c      		prot = PROT_READ;
prot             1466 ipc/shm.c      		prot = PROT_READ | PROT_WRITE;
prot             1471 ipc/shm.c      		prot |= PROT_EXEC;
prot             1543 ipc/shm.c      	err = security_mmap_file(file, prot, flags);
prot             1561 ipc/shm.c      	addr = do_mmap_pgoff(file, addr, size, prot, flags, 0, &populate, NULL);
prot              170 kernel/dma/mapping.c pgprot_t dma_pgprot(struct device *dev, pgprot_t prot, unsigned long attrs)
prot              173 kernel/dma/mapping.c 		prot = pgprot_decrypted(prot);
prot              177 kernel/dma/mapping.c 		return prot;
prot              180 kernel/dma/mapping.c 		return pgprot_writecombine(prot);
prot              182 kernel/dma/mapping.c 	return pgprot_dmacoherent(prot);
prot               24 kernel/dma/remap.c 			size_t size, pgprot_t prot, const void *caller)
prot               32 kernel/dma/remap.c 	if (map_vm_area(area, prot, pages)) {
prot               45 kernel/dma/remap.c 			 pgprot_t prot, const void *caller)
prot               49 kernel/dma/remap.c 	area = __dma_common_pages_remap(pages, size, prot, caller);
prot               63 kernel/dma/remap.c 			pgprot_t prot, const void *caller)
prot               76 kernel/dma/remap.c 	area = __dma_common_pages_remap(pages, size, prot, caller);
prot             7363 kernel/events/core.c 	u32			prot, flags;
prot             7406 kernel/events/core.c 		mmap_event->event_id.header.size += sizeof(mmap_event->prot);
prot             7426 kernel/events/core.c 		perf_output_put(&handle, mmap_event->prot);
prot             7447 kernel/events/core.c 	u32 prot = 0, flags = 0;
prot             7454 kernel/events/core.c 		prot |= PROT_READ;
prot             7456 kernel/events/core.c 		prot |= PROT_WRITE;
prot             7458 kernel/events/core.c 		prot |= PROT_EXEC;
prot             7546 kernel/events/core.c 	mmap_event->prot = prot;
prot               64 lib/ioremap.c  		unsigned long end, phys_addr_t phys_addr, pgprot_t prot)
prot               75 lib/ioremap.c  		set_pte_at(&init_mm, addr, pte, pfn_pte(pfn, prot));
prot               83 lib/ioremap.c  				pgprot_t prot)
prot              100 lib/ioremap.c  	return pmd_set_huge(pmd, phys_addr, prot);
prot              104 lib/ioremap.c  		unsigned long end, phys_addr_t phys_addr, pgprot_t prot)
prot              115 lib/ioremap.c  		if (ioremap_try_huge_pmd(pmd, addr, next, phys_addr, prot))
prot              118 lib/ioremap.c  		if (ioremap_pte_range(pmd, addr, next, phys_addr, prot))
prot              126 lib/ioremap.c  				pgprot_t prot)
prot              143 lib/ioremap.c  	return pud_set_huge(pud, phys_addr, prot);
prot              147 lib/ioremap.c  		unsigned long end, phys_addr_t phys_addr, pgprot_t prot)
prot              158 lib/ioremap.c  		if (ioremap_try_huge_pud(pud, addr, next, phys_addr, prot))
prot              161 lib/ioremap.c  		if (ioremap_pmd_range(pud, addr, next, phys_addr, prot))
prot              169 lib/ioremap.c  				pgprot_t prot)
prot              186 lib/ioremap.c  	return p4d_set_huge(p4d, phys_addr, prot);
prot              190 lib/ioremap.c  		unsigned long end, phys_addr_t phys_addr, pgprot_t prot)
prot              201 lib/ioremap.c  		if (ioremap_try_huge_p4d(p4d, addr, next, phys_addr, prot))
prot              204 lib/ioremap.c  		if (ioremap_pud_range(p4d, addr, next, phys_addr, prot))
prot              211 lib/ioremap.c  		       unsigned long end, phys_addr_t phys_addr, pgprot_t prot)
prot              225 lib/ioremap.c  		err = ioremap_p4d_range(pgd, addr, next, phys_addr, prot);
prot               36 mm/early_ioremap.c 						    pgprot_t prot)
prot               38 mm/early_ioremap.c 	return prot;
prot               58 mm/early_ioremap.c 					    phys_addr_t phys, pgprot_t prot)
prot              106 mm/early_ioremap.c __early_ioremap(resource_size_t phys_addr, unsigned long size, pgprot_t prot)
prot              154 mm/early_ioremap.c 			__late_set_fixmap(idx, phys_addr, prot);
prot              156 mm/early_ioremap.c 			__early_set_fixmap(idx, phys_addr, prot);
prot              226 mm/early_ioremap.c 	pgprot_t prot = early_memremap_pgprot_adjust(phys_addr, size,
prot              229 mm/early_ioremap.c 	return (__force void *)__early_ioremap(phys_addr, size, prot);
prot              235 mm/early_ioremap.c 	pgprot_t prot = early_memremap_pgprot_adjust(phys_addr, size,
prot              238 mm/early_ioremap.c 	return (__force void *)__early_ioremap(phys_addr, size, prot);
prot              770 mm/huge_memory.c 		pmd_t *pmd, pfn_t pfn, pgprot_t prot, bool write,
prot              793 mm/huge_memory.c 	entry = pmd_mkhuge(pfn_t_pmd(pfn, prot));
prot              859 mm/huge_memory.c 		pud_t *pud, pfn_t pfn, pgprot_t prot, bool write)
prot              880 mm/huge_memory.c 	entry = pud_mkhuge(pfn_t_pud(pfn, prot));
prot             1423 mm/memory.c    			struct page *page, pgprot_t prot)
prot             1446 mm/memory.c    	set_pte_at(mm, addr, pte, mk_pte(page, prot));
prot             1582 mm/memory.c    			pfn_t pfn, pgprot_t prot, bool mkwrite)
prot             1617 mm/memory.c    		entry = pte_mkdevmap(pfn_t_pte(pfn, prot));
prot             1619 mm/memory.c    		entry = pte_mkspecial(pfn_t_pte(pfn, prot));
prot             1793 mm/memory.c    			unsigned long pfn, pgprot_t prot)
prot             1805 mm/memory.c    		if (!pfn_modify_allowed(pfn, prot)) {
prot             1809 mm/memory.c    		set_pte_at(mm, addr, pte, pte_mkspecial(pfn_pte(pfn, prot)));
prot             1819 mm/memory.c    			unsigned long pfn, pgprot_t prot)
prot             1833 mm/memory.c    				pfn + (addr >> PAGE_SHIFT), prot);
prot             1842 mm/memory.c    			unsigned long pfn, pgprot_t prot)
prot             1855 mm/memory.c    				pfn + (addr >> PAGE_SHIFT), prot);
prot             1864 mm/memory.c    			unsigned long pfn, pgprot_t prot)
prot             1877 mm/memory.c    				pfn + (addr >> PAGE_SHIFT), prot);
prot             1897 mm/memory.c    		    unsigned long pfn, unsigned long size, pgprot_t prot)
prot             1930 mm/memory.c    	err = track_pfn_remap(vma, &prot, remap_pfn, addr, PAGE_ALIGN(size));
prot             1943 mm/memory.c    				pfn + (addr >> PAGE_SHIFT), prot);
prot             4257 mm/memory.c    		unsigned long *prot, resource_size_t *phys)
prot             4273 mm/memory.c    	*prot = pgprot_val(pte_pgprot(pte));
prot             4287 mm/memory.c    	unsigned long prot = 0;
prot             4291 mm/memory.c    	if (follow_phys(vma, addr, write, &prot, &phys_addr))
prot             4294 mm/memory.c    	maddr = ioremap_prot(phys_addr, PAGE_ALIGN(len + offset), prot);
prot              100 mm/mmap.c      static inline pgprot_t arch_filter_pgprot(pgprot_t prot)
prot              102 mm/mmap.c      	return prot;
prot             1390 mm/mmap.c      			unsigned long len, unsigned long prot,
prot             1409 mm/mmap.c      	if ((prot & PROT_READ) && (current->personality & READ_IMPLIES_EXEC))
prot             1411 mm/mmap.c      			prot |= PROT_EXEC;
prot             1447 mm/mmap.c      	if (prot == PROT_EXEC) {
prot             1457 mm/mmap.c      	vm_flags |= calc_vm_prot_bits(prot, pkey) | calc_vm_flag_bits(flags) |
prot             1490 mm/mmap.c      			if (prot & PROT_WRITE) {
prot             1578 mm/mmap.c      			      unsigned long prot, unsigned long flags,
prot             1619 mm/mmap.c      	retval = vm_mmap_pgoff(file, addr, len, prot, flags, pgoff);
prot             1627 mm/mmap.c      		unsigned long, prot, unsigned long, flags,
prot             1630 mm/mmap.c      	return ksys_mmap_pgoff(addr, len, prot, flags, fd, pgoff);
prot             1637 mm/mmap.c      	unsigned long prot;
prot             1652 mm/mmap.c      	return ksys_mmap_pgoff(a.addr, a.len, a.prot, a.flags, a.fd,
prot             2892 mm/mmap.c      		unsigned long, prot, unsigned long, pgoff, unsigned long, flags)
prot             2904 mm/mmap.c      	if (prot)
prot             2949 mm/mmap.c      	prot |= vma->vm_flags & VM_READ ? PROT_READ : 0;
prot             2950 mm/mmap.c      	prot |= vma->vm_flags & VM_WRITE ? PROT_WRITE : 0;
prot             2951 mm/mmap.c      	prot |= vma->vm_flags & VM_EXEC ? PROT_EXEC : 0;
prot             2976 mm/mmap.c      			prot, flags, pgoff, &populate, NULL);
prot              487 mm/mprotect.c  		unsigned long prot, int pkey)
prot              492 mm/mprotect.c  	const int grows = prot & (PROT_GROWSDOWN|PROT_GROWSUP);
prot              494 mm/mprotect.c  				(prot & PROT_READ);
prot              498 mm/mprotect.c  	prot &= ~(PROT_GROWSDOWN|PROT_GROWSUP);
prot              510 mm/mprotect.c  	if (!arch_validate_prot(prot, start))
prot              513 mm/mprotect.c  	reqprot = prot;
prot              560 mm/mprotect.c  			prot |= PROT_EXEC;
prot              570 mm/mprotect.c  		new_vma_pkey = arch_override_mprotect_pkey(vma, prot, pkey);
prot              571 mm/mprotect.c  		newflags = calc_vm_prot_bits(prot, new_vma_pkey);
prot              580 mm/mprotect.c  		error = security_file_mprotect(vma, reqprot, prot);
prot              602 mm/mprotect.c  		prot = reqprot;
prot              610 mm/mprotect.c  		unsigned long, prot)
prot              612 mm/mprotect.c  	return do_mprotect_pkey(start, len, prot, -1);
prot              618 mm/mprotect.c  		unsigned long, prot, int, pkey)
prot              620 mm/mprotect.c  	return do_mprotect_pkey(start, len, prot, pkey);
prot              143 mm/nommu.c     void *__vmalloc(unsigned long size, gfp_t gfp_mask, pgprot_t prot)
prot              330 mm/nommu.c     void *vmap(struct page **pages, unsigned int count, unsigned long flags, pgprot_t prot)
prot              343 mm/nommu.c     void *vm_map_ram(struct page **pages, unsigned int count, int node, pgprot_t prot)
prot              787 mm/nommu.c     				 unsigned long prot,
prot              860 mm/nommu.c     			if ((prot & PROT_WRITE) &&
prot              884 mm/nommu.c     			if (prot & PROT_WRITE)
prot              889 mm/nommu.c     			if (((prot & PROT_READ)  && !(capabilities & NOMMU_MAP_READ))  ||
prot              890 mm/nommu.c     			    ((prot & PROT_WRITE) && !(capabilities & NOMMU_MAP_WRITE)) ||
prot              891 mm/nommu.c     			    ((prot & PROT_EXEC)  && !(capabilities & NOMMU_MAP_EXEC))
prot              904 mm/nommu.c     			if (prot & PROT_EXEC)
prot              906 mm/nommu.c     		} else if ((prot & PROT_READ) && !(prot & PROT_EXEC)) {
prot              910 mm/nommu.c     					prot |= PROT_EXEC;
prot              912 mm/nommu.c     		} else if ((prot & PROT_READ) &&
prot              913 mm/nommu.c     			 (prot & PROT_EXEC) &&
prot              926 mm/nommu.c     		if ((prot & PROT_READ) &&
prot              928 mm/nommu.c     			prot |= PROT_EXEC;
prot              946 mm/nommu.c     					unsigned long prot,
prot              952 mm/nommu.c     	vm_flags = calc_vm_prot_bits(prot, 0) | calc_vm_flag_bits(flags);
prot              958 mm/nommu.c     		if (file && !(prot & PROT_WRITE))
prot             1101 mm/nommu.c     			unsigned long prot,
prot             1118 mm/nommu.c     	ret = validate_mmap_request(file, addr, len, prot, flags, pgoff,
prot             1129 mm/nommu.c     	vm_flags |= determine_vm_flags(file, prot, flags, capabilities);
prot             1325 mm/nommu.c     			      unsigned long prot, unsigned long flags,
prot             1340 mm/nommu.c     	retval = vm_mmap_pgoff(file, addr, len, prot, flags, pgoff);
prot             1349 mm/nommu.c     		unsigned long, prot, unsigned long, flags,
prot             1352 mm/nommu.c     	return ksys_mmap_pgoff(addr, len, prot, flags, fd, pgoff);
prot             1359 mm/nommu.c     	unsigned long prot;
prot             1374 mm/nommu.c     	return ksys_mmap_pgoff(a.addr, a.len, a.prot, a.flags, a.fd,
prot             1652 mm/nommu.c     		unsigned long pfn, unsigned long size, pgprot_t prot)
prot              484 mm/util.c      	unsigned long len, unsigned long prot,
prot              492 mm/util.c      	ret = security_mmap_file(file, prot, flag);
prot              496 mm/util.c      		ret = do_mmap_pgoff(file, addr, len, prot, flag, pgoff,
prot              507 mm/util.c      	unsigned long len, unsigned long prot,
prot              515 mm/util.c      	return vm_mmap_pgoff(file, addr, len, prot, flag, offset >> PAGE_SHIFT);
prot              139 mm/vmalloc.c   		unsigned long end, pgprot_t prot, struct page **pages, int *nr)
prot              158 mm/vmalloc.c   		set_pte_at(&init_mm, addr, pte, mk_pte(page, prot));
prot              165 mm/vmalloc.c   		unsigned long end, pgprot_t prot, struct page **pages, int *nr)
prot              175 mm/vmalloc.c   		if (vmap_pte_range(pmd, addr, next, prot, pages, nr))
prot              182 mm/vmalloc.c   		unsigned long end, pgprot_t prot, struct page **pages, int *nr)
prot              192 mm/vmalloc.c   		if (vmap_pmd_range(pud, addr, next, prot, pages, nr))
prot              199 mm/vmalloc.c   		unsigned long end, pgprot_t prot, struct page **pages, int *nr)
prot              209 mm/vmalloc.c   		if (vmap_pud_range(p4d, addr, next, prot, pages, nr))
prot              222 mm/vmalloc.c   				   pgprot_t prot, struct page **pages)
prot              234 mm/vmalloc.c   		err = vmap_p4d_range(pgd, addr, next, prot, pages, &nr);
prot              243 mm/vmalloc.c   			   pgprot_t prot, struct page **pages)
prot              247 mm/vmalloc.c   	ret = vmap_page_range_noflush(start, end, prot, pages);
prot             1778 mm/vmalloc.c   void *vm_map_ram(struct page **pages, unsigned int count, int node, pgprot_t prot)
prot             1799 mm/vmalloc.c   	if (vmap_page_range(addr, addr + size, prot, pages) < 0) {
prot             1963 mm/vmalloc.c   			     pgprot_t prot, struct page **pages)
prot             1965 mm/vmalloc.c   	return vmap_page_range_noflush(addr, addr + size, prot, pages);
prot             2006 mm/vmalloc.c   int map_vm_area(struct vm_struct *area, pgprot_t prot, struct page **pages)
prot             2012 mm/vmalloc.c   	err = vmap_page_range(addr, end, prot, pages);
prot             2371 mm/vmalloc.c   	   unsigned long flags, pgprot_t prot)
prot             2386 mm/vmalloc.c   	if (map_vm_area(area, prot, pages)) {
prot             2396 mm/vmalloc.c   			    gfp_t gfp_mask, pgprot_t prot,
prot             2399 mm/vmalloc.c   				 pgprot_t prot, int node)
prot             2449 mm/vmalloc.c   	if (map_vm_area(area, prot, pages))
prot             2481 mm/vmalloc.c   			pgprot_t prot, unsigned long vm_flags, int node,
prot             2497 mm/vmalloc.c   	addr = __vmalloc_area_node(area, gfp_mask, prot, node);
prot             2549 mm/vmalloc.c   			    gfp_t gfp_mask, pgprot_t prot,
prot             2553 mm/vmalloc.c   				gfp_mask, prot, 0, node, caller);
prot             2556 mm/vmalloc.c   void *__vmalloc(unsigned long size, gfp_t gfp_mask, pgprot_t prot)
prot             2558 mm/vmalloc.c   	return __vmalloc_node(size, 1, gfp_mask, prot, NUMA_NO_NODE,
prot              229 net/atm/br2684.c 			unsigned short prot = ntohs(skb->protocol);
prot              232 net/atm/br2684.c 			switch (prot) {
prot             1032 net/caif/caif_socket.c 	static struct proto prot = {.name = "PF_CAIF",
prot             1061 net/caif/caif_socket.c 	sk = sk_alloc(net, PF_CAIF, GFP_KERNEL, &prot, kern);
prot              102 net/can/af_can.c 	if (cp && !try_module_get(cp->prot->owner))
prot              111 net/can/af_can.c 	module_put(cp->prot->owner);
prot              158 net/can/af_can.c 	sk = sk_alloc(net, PF_CAN, GFP_KERNEL, cp->prot, kern);
prot              729 net/can/af_can.c 	err = proto_register(cp->prot, 0);
prot              745 net/can/af_can.c 		proto_unregister(cp->prot);
prot              766 net/can/af_can.c 	proto_unregister(cp->prot);
prot             1670 net/can/bcm.c  	.prot       = &bcm_proto,
prot             1222 net/can/j1939/socket.c 	.prot = &j1939_proto,
prot              877 net/can/raw.c  	.prot       = &raw_proto,
prot             1451 net/core/neighbour.c 	__be16 prot = n->tbl->protocol;
prot             1460 net/core/neighbour.c 		dev->header_ops->cache(n, hh, prot);
prot             1591 net/core/sock.c static struct sock *sk_prot_alloc(struct proto *prot, gfp_t priority,
prot             1597 net/core/sock.c 	slab = prot->slab;
prot             1603 net/core/sock.c 			sk_prot_clear_nulls(sk, prot->obj_size);
prot             1605 net/core/sock.c 		sk = kmalloc(prot->obj_size, priority);
prot             1611 net/core/sock.c 		if (!try_module_get(prot->owner))
prot             1628 net/core/sock.c static void sk_prot_free(struct proto *prot, struct sock *sk)
prot             1633 net/core/sock.c 	owner = prot->owner;
prot             1634 net/core/sock.c 	slab = prot->slab;
prot             1655 net/core/sock.c 		      struct proto *prot, int kern)
prot             1659 net/core/sock.c 	sk = sk_prot_alloc(prot, priority | __GFP_ZERO, family);
prot             1666 net/core/sock.c 		sk->sk_prot = sk->sk_prot_creator = prot;
prot             2502 net/core/sock.c 	struct proto *prot = sk->sk_prot;
prot             2526 net/core/sock.c 		if (atomic_read(&sk->sk_rmem_alloc) < sk_get_rmem0(sk, prot))
prot             2530 net/core/sock.c 		int wmem0 = sk_get_wmem0(sk, prot);
prot             2566 net/core/sock.c 		trace_sock_exceed_buf_limit(sk, prot, allocated, kind);
prot             3232 net/core/sock.c void sock_prot_inuse_add(struct net *net, struct proto *prot, int val)
prot             3234 net/core/sock.c 	__this_cpu_add(net->core.prot_inuse->val[prot->inuse_idx], val);
prot             3238 net/core/sock.c int sock_prot_inuse_get(struct net *net, struct proto *prot)
prot             3240 net/core/sock.c 	int cpu, idx = prot->inuse_idx;
prot             3305 net/core/sock.c static int assign_proto_idx(struct proto *prot)
prot             3307 net/core/sock.c 	prot->inuse_idx = find_first_zero_bit(proto_inuse_idx, PROTO_INUSE_NR);
prot             3309 net/core/sock.c 	if (unlikely(prot->inuse_idx == PROTO_INUSE_NR - 1)) {
prot             3314 net/core/sock.c 	set_bit(prot->inuse_idx, proto_inuse_idx);
prot             3318 net/core/sock.c static void release_proto_idx(struct proto *prot)
prot             3320 net/core/sock.c 	if (prot->inuse_idx != PROTO_INUSE_NR - 1)
prot             3321 net/core/sock.c 		clear_bit(prot->inuse_idx, proto_inuse_idx);
prot             3324 net/core/sock.c static inline int assign_proto_idx(struct proto *prot)
prot             3329 net/core/sock.c static inline void release_proto_idx(struct proto *prot)
prot             3348 net/core/sock.c static int req_prot_init(const struct proto *prot)
prot             3350 net/core/sock.c 	struct request_sock_ops *rsk_prot = prot->rsk_prot;
prot             3356 net/core/sock.c 					prot->name);
prot             3362 net/core/sock.c 					   SLAB_ACCOUNT | prot->slab_flags,
prot             3367 net/core/sock.c 			prot->name);
prot             3373 net/core/sock.c int proto_register(struct proto *prot, int alloc_slab)
prot             3378 net/core/sock.c 		prot->slab = kmem_cache_create_usercopy(prot->name,
prot             3379 net/core/sock.c 					prot->obj_size, 0,
prot             3381 net/core/sock.c 					prot->slab_flags,
prot             3382 net/core/sock.c 					prot->useroffset, prot->usersize,
prot             3385 net/core/sock.c 		if (prot->slab == NULL) {
prot             3387 net/core/sock.c 				prot->name);
prot             3391 net/core/sock.c 		if (req_prot_init(prot))
prot             3394 net/core/sock.c 		if (prot->twsk_prot != NULL) {
prot             3395 net/core/sock.c 			prot->twsk_prot->twsk_slab_name = kasprintf(GFP_KERNEL, "tw_sock_%s", prot->name);
prot             3397 net/core/sock.c 			if (prot->twsk_prot->twsk_slab_name == NULL)
prot             3400 net/core/sock.c 			prot->twsk_prot->twsk_slab =
prot             3401 net/core/sock.c 				kmem_cache_create(prot->twsk_prot->twsk_slab_name,
prot             3402 net/core/sock.c 						  prot->twsk_prot->twsk_obj_size,
prot             3405 net/core/sock.c 						  prot->slab_flags,
prot             3407 net/core/sock.c 			if (prot->twsk_prot->twsk_slab == NULL)
prot             3413 net/core/sock.c 	ret = assign_proto_idx(prot);
prot             3418 net/core/sock.c 	list_add(&prot->node, &proto_list);
prot             3423 net/core/sock.c 	if (alloc_slab && prot->twsk_prot)
prot             3424 net/core/sock.c 		kfree(prot->twsk_prot->twsk_slab_name);
prot             3427 net/core/sock.c 		req_prot_cleanup(prot->rsk_prot);
prot             3429 net/core/sock.c 		kmem_cache_destroy(prot->slab);
prot             3430 net/core/sock.c 		prot->slab = NULL;
prot             3437 net/core/sock.c void proto_unregister(struct proto *prot)
prot             3440 net/core/sock.c 	release_proto_idx(prot);
prot             3441 net/core/sock.c 	list_del(&prot->node);
prot             3444 net/core/sock.c 	kmem_cache_destroy(prot->slab);
prot             3445 net/core/sock.c 	prot->slab = NULL;
prot             3447 net/core/sock.c 	req_prot_cleanup(prot->rsk_prot);
prot             3449 net/core/sock.c 	if (prot->twsk_prot != NULL && prot->twsk_prot->twsk_slab != NULL) {
prot             3450 net/core/sock.c 		kmem_cache_destroy(prot->twsk_prot->twsk_slab);
prot             3451 net/core/sock.c 		kfree(prot->twsk_prot->twsk_slab_name);
prot             3452 net/core/sock.c 		prot->twsk_prot->twsk_slab = NULL;
prot             1009 net/dccp/ipv4.c 	.prot		= &dccp_v4_prot,
prot             1093 net/dccp/ipv6.c 	.prot		= &dccp_v6_prot,
prot              314 net/ipv4/af_inet.c 	answer_prot = answer->prot;
prot             1093 net/ipv4/af_inet.c 		.prot =       &tcp_prot,
prot             1102 net/ipv4/af_inet.c 		.prot =       &udp_prot,
prot             1110 net/ipv4/af_inet.c 		.prot =       &ping_prot,
prot             1118 net/ipv4/af_inet.c 	       .prot =       &raw_prot,
prot              143 net/ipv4/fou.c 		int prot;
prot              147 net/ipv4/fou.c 			prot = IPPROTO_IPIP;
prot              150 net/ipv4/fou.c 			prot = IPPROTO_IPV6;
prot              159 net/ipv4/fou.c 		return -prot;
prot               32 net/ipv4/protocol.c int inet_add_protocol(const struct net_protocol *prot, unsigned char protocol)
prot               34 net/ipv4/protocol.c 	if (!prot->netns_ok) {
prot               41 net/ipv4/protocol.c 			NULL, prot) ? 0 : -1;
prot               45 net/ipv4/protocol.c int inet_add_offload(const struct net_offload *prot, unsigned char protocol)
prot               48 net/ipv4/protocol.c 			NULL, prot) ? 0 : -1;
prot               52 net/ipv4/protocol.c int inet_del_protocol(const struct net_protocol *prot, unsigned char protocol)
prot               57 net/ipv4/protocol.c 		       prot, NULL) == prot) ? 0 : -1;
prot               65 net/ipv4/protocol.c int inet_del_offload(const struct net_offload *prot, unsigned char protocol)
prot               70 net/ipv4/protocol.c 		       prot, NULL) == prot) ? 0 : -1;
prot              528 net/ipv4/route.c 			     u8 prot, u32 mark, int flow_flags)
prot              536 net/ipv4/route.c 		prot = inet->hdrincl ? IPPROTO_RAW : sk->sk_protocol;
prot              539 net/ipv4/route.c 			   RT_SCOPE_UNIVERSE, prot,
prot              552 net/ipv4/route.c 	u8 prot = iph->protocol;
prot              555 net/ipv4/route.c 	__build_flow_key(net, fl4, sk, iph, oif, tos, prot, mark, 0);
prot              826 net/ipv4/route.c 	u8 prot = iph->protocol;
prot              831 net/ipv4/route.c 	__build_flow_key(net, &fl4, sk, iph, oif, tos, prot, mark, 0);
prot              600 net/ipv4/tcp_bpf.c static void tcp_bpf_rebuild_protos(struct proto prot[TCP_BPF_NUM_CFGS],
prot              603 net/ipv4/tcp_bpf.c 	prot[TCP_BPF_BASE]			= *base;
prot              604 net/ipv4/tcp_bpf.c 	prot[TCP_BPF_BASE].unhash		= tcp_bpf_unhash;
prot              605 net/ipv4/tcp_bpf.c 	prot[TCP_BPF_BASE].close		= tcp_bpf_close;
prot              606 net/ipv4/tcp_bpf.c 	prot[TCP_BPF_BASE].recvmsg		= tcp_bpf_recvmsg;
prot              607 net/ipv4/tcp_bpf.c 	prot[TCP_BPF_BASE].stream_memory_read	= tcp_bpf_stream_read;
prot              609 net/ipv4/tcp_bpf.c 	prot[TCP_BPF_TX]			= prot[TCP_BPF_BASE];
prot              610 net/ipv4/tcp_bpf.c 	prot[TCP_BPF_TX].sendmsg		= tcp_bpf_sendmsg;
prot              611 net/ipv4/tcp_bpf.c 	prot[TCP_BPF_TX].sendpage		= tcp_bpf_sendpage;
prot               69 net/ipv4/udplite.c 	.prot		=  &udplite_prot,
prot              173 net/ipv6/af_inet6.c 	answer_prot = answer->prot;
prot              203 net/ipv6/af_inet6.c 	sk->sk_backlog_rcv	= answer->prot->backlog_rcv;
prot              228 net/ipv6/ipv6_sockglue.c 				struct proto *prot = &udp_prot;
prot              231 net/ipv6/ipv6_sockglue.c 					prot = &udplite_prot;
prot              234 net/ipv6/ipv6_sockglue.c 				sock_prot_inuse_add(net, prot, 1);
prot              236 net/ipv6/ipv6_sockglue.c 				sk->sk_prot = prot;
prot              186 net/ipv6/ping.c 	.prot =      &pingv6_prot,
prot               28 net/ipv6/protocol.c int inet6_add_protocol(const struct inet6_protocol *prot, unsigned char protocol)
prot               31 net/ipv6/protocol.c 			NULL, prot) ? 0 : -1;
prot               35 net/ipv6/protocol.c int inet6_del_protocol(const struct inet6_protocol *prot, unsigned char protocol)
prot               40 net/ipv6/protocol.c 		       prot, NULL) == prot) ? 0 : -1;
prot               52 net/ipv6/protocol.c int inet6_add_offload(const struct net_offload *prot, unsigned char protocol)
prot               55 net/ipv6/protocol.c 			NULL, prot) ? 0 : -1;
prot               59 net/ipv6/protocol.c int inet6_del_offload(const struct net_offload *prot, unsigned char protocol)
prot               64 net/ipv6/protocol.c 		       prot, NULL) == prot) ? 0 : -1;
prot             1388 net/ipv6/raw.c 	.prot		= &rawv6_prot,
prot             2067 net/ipv6/tcp_ipv6.c 	.prot		=	&tcpv6_prot,
prot             1689 net/ipv6/udp.c 	.prot =      &udpv6_prot,
prot               64 net/ipv6/udplite.c 	.prot		= &udplitev6_prot,
prot              650 net/l2tp/l2tp_ip.c 	.prot		= &l2tp_ip_prot,
prot              783 net/l2tp/l2tp_ip6.c 	.prot		= &l2tp_ip6_prot,
prot              915 net/llc/llc_conn.c struct sock *llc_sk_alloc(struct net *net, int family, gfp_t priority, struct proto *prot, int kern)
prot              917 net/llc/llc_conn.c 	struct sock *sk = sk_alloc(net, family, priority, prot, kern);
prot              608 net/mac80211/tdls.c 		u16 prot = IEEE80211_HT_OP_MODE_PROTECTION_NONHT_MIXED |
prot              614 net/mac80211/tdls.c 					   &sdata->vif.bss_conf.chandef, prot,
prot               36 net/phonet/af_phonet.c 	if (pp && !try_module_get(pp->prot->owner))
prot               45 net/phonet/af_phonet.c 	module_put(pp->prot->owner);
prot               87 net/phonet/af_phonet.c 	sk = sk_alloc(net, PF_PHONET, GFP_KERNEL, pnp->prot, kern);
prot              467 net/phonet/af_phonet.c 	err = proto_register(pp->prot, 1);
prot              490 net/phonet/af_phonet.c 	proto_unregister(pp->prot);
prot              187 net/phonet/datagram.c 	.prot		= &pn_proto,
prot             1343 net/phonet/pep.c 	.prot		= &pep_proto,
prot             1043 net/sctp/ipv6.c 	.prot 	       = &sctpv6_prot,
prot             1050 net/sctp/ipv6.c 	.prot 	       = &sctpv6_prot,
prot             1048 net/sctp/protocol.c 	.prot       = &sctp_prot,
prot             1055 net/sctp/protocol.c 	.prot       = &sctp_prot,
prot              218 net/smc/af_smc.c 	struct proto *prot;
prot              221 net/smc/af_smc.c 	prot = (protocol == SMCPROTO_SMC6) ? &smc_proto6 : &smc_proto;
prot              222 net/smc/af_smc.c 	sk = sk_alloc(net, PF_SMC, GFP_KERNEL, prot, 0);
prot              193 net/smc/smc_diag.c static int smc_diag_dump_proto(struct proto *prot, struct sk_buff *skb,
prot              202 net/smc/smc_diag.c 	read_lock(&prot->h.smc_hash->lock);
prot              203 net/smc/smc_diag.c 	head = &prot->h.smc_hash->ht;
prot              216 net/smc/smc_diag.c 	read_unlock(&prot->h.smc_hash->lock);
prot              450 net/sunrpc/rpcb_clnt.c int rpcb_register(struct net *net, u32 prog, u32 vers, int prot, unsigned short port)
prot              455 net/sunrpc/rpcb_clnt.c 		.r_prot		= prot,
prot              466 net/sunrpc/rpcb_clnt.c 			prog, vers, prot, port);
prot              698 net/sunrpc/rpcb_clnt.c 		xprt->servername, clnt->cl_prog, clnt->cl_vers, xprt->prot);
prot              753 net/sunrpc/rpcb_clnt.c 				xprt->prot, bind_version,
prot              771 net/sunrpc/rpcb_clnt.c 	map->r_prot = xprt->prot;
prot             1563 net/sunrpc/svc.c 	rqstp->rq_prot = req->rq_xprt->prot;
prot             1192 net/sunrpc/svc_xprt.c 		dr->prot = rqstp->rq_prot;
prot             1227 net/sunrpc/svc_xprt.c 	rqstp->rq_prot        = dr->prot;
prot              309 net/sunrpc/xprtrdma/svc_rdma_backchannel.c 	xprt->prot = XPRT_TRANSPORT_BC_RDMA;
prot              343 net/sunrpc/xprtrdma/transport.c 	xprt->prot = IPPROTO_TCP;
prot             2917 net/sunrpc/xprtsock.c 	xprt->prot = 0;
prot             2986 net/sunrpc/xprtsock.c 	xprt->prot = IPPROTO_UDP;
prot             3066 net/sunrpc/xprtsock.c 	xprt->prot = IPPROTO_TCP;
prot             3139 net/sunrpc/xprtsock.c 	xprt->prot = IPPROTO_TCP;
prot              261 net/tls/tls_device.c 	struct tls_prot_info *prot = &ctx->prot_info;
prot              273 net/tls/tls_device.c 	tls_advance_record_sn(sk, prot, &ctx->tx);
prot              295 net/tls/tls_device.c 	struct tls_prot_info *prot = &ctx->prot_info;
prot              304 net/tls/tls_device.c 	if (likely(skb_page_frag_refill(prot->tag_size, pfrag,
prot              307 net/tls/tls_device.c 		tls_append_frag(record, pfrag, prot->tag_size);
prot              309 net/tls/tls_device.c 		ret = prot->tag_size;
prot              310 net/tls/tls_device.c 		if (record->len <= prot->overhead_size)
prot              316 net/tls/tls_device.c 			 record->len - prot->overhead_size,
prot              317 net/tls/tls_device.c 			 record_type, prot->version);
prot              406 net/tls/tls_device.c 	struct tls_prot_info *prot = &tls_ctx->prot_info;
prot              441 net/tls/tls_device.c 			      prot->prepend_size;
prot              444 net/tls/tls_device.c 				       prot->prepend_size);
prot              462 net/tls/tls_device.c 			} else if (record->len > prot->prepend_size) {
prot              682 net/tls/tls_device.c 	struct tls_prot_info *prot;
prot              690 net/tls/tls_device.c 	prot = &tls_ctx->prot_info;
prot              692 net/tls/tls_device.c 	memcpy(rcd_sn, tls_ctx->rx.rec_seq, prot->rec_seq_size);
prot              717 net/tls/tls_device.c 		tls_bigint_increment(rcd_sn, prot->rec_seq_size);
prot              759 net/tls/tls_device.c 		struct tls_prot_info *prot = &tls_ctx->prot_info;
prot              762 net/tls/tls_device.c 		memcpy(rcd_sn, tls_ctx->rx.rec_seq, prot->rec_seq_size);
prot              763 net/tls/tls_device.c 		tls_bigint_increment(rcd_sn, prot->rec_seq_size);
prot              908 net/tls/tls_device.c 	struct tls_prot_info *prot = &tls_ctx->prot_info;
prot              961 net/tls/tls_device.c 	prot->version = crypto_info->version;
prot              962 net/tls/tls_device.c 	prot->cipher_type = crypto_info->cipher_type;
prot              963 net/tls/tls_device.c 	prot->prepend_size = TLS_HEADER_SIZE + nonce_size;
prot              964 net/tls/tls_device.c 	prot->tag_size = tag_size;
prot              965 net/tls/tls_device.c 	prot->overhead_size = prot->prepend_size + prot->tag_size;
prot              966 net/tls/tls_device.c 	prot->iv_size = iv_size;
prot              976 net/tls/tls_device.c 	prot->rec_seq_size = rec_seq_size;
prot               65 net/tls/tls_main.c static void build_protos(struct proto prot[TLS_NUM_CONFIG][TLS_NUM_CONFIG],
prot              723 net/tls/tls_main.c static void build_protos(struct proto prot[TLS_NUM_CONFIG][TLS_NUM_CONFIG],
prot              726 net/tls/tls_main.c 	prot[TLS_BASE][TLS_BASE] = *base;
prot              727 net/tls/tls_main.c 	prot[TLS_BASE][TLS_BASE].setsockopt	= tls_setsockopt;
prot              728 net/tls/tls_main.c 	prot[TLS_BASE][TLS_BASE].getsockopt	= tls_getsockopt;
prot              729 net/tls/tls_main.c 	prot[TLS_BASE][TLS_BASE].close		= tls_sk_proto_close;
prot              731 net/tls/tls_main.c 	prot[TLS_SW][TLS_BASE] = prot[TLS_BASE][TLS_BASE];
prot              732 net/tls/tls_main.c 	prot[TLS_SW][TLS_BASE].sendmsg		= tls_sw_sendmsg;
prot              733 net/tls/tls_main.c 	prot[TLS_SW][TLS_BASE].sendpage		= tls_sw_sendpage;
prot              735 net/tls/tls_main.c 	prot[TLS_BASE][TLS_SW] = prot[TLS_BASE][TLS_BASE];
prot              736 net/tls/tls_main.c 	prot[TLS_BASE][TLS_SW].recvmsg		  = tls_sw_recvmsg;
prot              737 net/tls/tls_main.c 	prot[TLS_BASE][TLS_SW].stream_memory_read = tls_sw_stream_read;
prot              738 net/tls/tls_main.c 	prot[TLS_BASE][TLS_SW].close		  = tls_sk_proto_close;
prot              740 net/tls/tls_main.c 	prot[TLS_SW][TLS_SW] = prot[TLS_SW][TLS_BASE];
prot              741 net/tls/tls_main.c 	prot[TLS_SW][TLS_SW].recvmsg		= tls_sw_recvmsg;
prot              742 net/tls/tls_main.c 	prot[TLS_SW][TLS_SW].stream_memory_read	= tls_sw_stream_read;
prot              743 net/tls/tls_main.c 	prot[TLS_SW][TLS_SW].close		= tls_sk_proto_close;
prot              746 net/tls/tls_main.c 	prot[TLS_HW][TLS_BASE] = prot[TLS_BASE][TLS_BASE];
prot              747 net/tls/tls_main.c 	prot[TLS_HW][TLS_BASE].sendmsg		= tls_device_sendmsg;
prot              748 net/tls/tls_main.c 	prot[TLS_HW][TLS_BASE].sendpage		= tls_device_sendpage;
prot              750 net/tls/tls_main.c 	prot[TLS_HW][TLS_SW] = prot[TLS_BASE][TLS_SW];
prot              751 net/tls/tls_main.c 	prot[TLS_HW][TLS_SW].sendmsg		= tls_device_sendmsg;
prot              752 net/tls/tls_main.c 	prot[TLS_HW][TLS_SW].sendpage		= tls_device_sendpage;
prot              754 net/tls/tls_main.c 	prot[TLS_BASE][TLS_HW] = prot[TLS_BASE][TLS_SW];
prot              756 net/tls/tls_main.c 	prot[TLS_SW][TLS_HW] = prot[TLS_SW][TLS_SW];
prot              758 net/tls/tls_main.c 	prot[TLS_HW][TLS_HW] = prot[TLS_HW][TLS_SW];
prot              761 net/tls/tls_main.c 	prot[TLS_HW_RECORD][TLS_HW_RECORD] = *base;
prot              762 net/tls/tls_main.c 	prot[TLS_HW_RECORD][TLS_HW_RECORD].hash		= tls_hw_hash;
prot              763 net/tls/tls_main.c 	prot[TLS_HW_RECORD][TLS_HW_RECORD].unhash	= tls_hw_unhash;
prot              122 net/tls/tls_sw.c 			  struct tls_prot_info *prot, struct sk_buff *skb)
prot              128 net/tls/tls_sw.c 	if (prot->version == TLS_1_3_VERSION) {
prot              134 net/tls/tls_sw.c 			if (back > rxm->full_len - prot->prepend_size)
prot              158 net/tls/tls_sw.c 	struct tls_prot_info *prot;
prot              167 net/tls/tls_sw.c 	prot = &tls_ctx->prot_info;
prot              177 net/tls/tls_sw.c 		pad = padding_length(ctx, prot, skb);
prot              183 net/tls/tls_sw.c 			rxm->offset += prot->prepend_size;
prot              184 net/tls/tls_sw.c 			rxm->full_len -= prot->overhead_size;
prot              224 net/tls/tls_sw.c 	struct tls_prot_info *prot = &tls_ctx->prot_info;
prot              229 net/tls/tls_sw.c 	aead_request_set_ad(aead_req, prot->aad_size);
prot              231 net/tls/tls_sw.c 			       data_len + prot->tag_size,
prot              269 net/tls/tls_sw.c 	struct tls_prot_info *prot = &tls_ctx->prot_info;
prot              275 net/tls/tls_sw.c 		target_size += prot->overhead_size;
prot              292 net/tls/tls_sw.c 	struct tls_prot_info *prot = &tls_ctx->prot_info;
prot              308 net/tls/tls_sw.c 	skip = prot->prepend_size + msg_pl->sg.size;
prot              316 net/tls/tls_sw.c 	struct tls_prot_info *prot = &tls_ctx->prot_info;
prot              335 net/tls/tls_sw.c 	sg_set_buf(&rec->sg_aead_in[0], rec->aad_space, prot->aad_size);
prot              339 net/tls/tls_sw.c 	sg_set_buf(&rec->sg_aead_out[0], rec->aad_space, prot->aad_size);
prot              428 net/tls/tls_sw.c 	struct tls_prot_info *prot = &tls_ctx->prot_info;
prot              440 net/tls/tls_sw.c 	sge->offset -= prot->prepend_size;
prot              441 net/tls/tls_sw.c 	sge->length += prot->prepend_size;
prot              490 net/tls/tls_sw.c 	struct tls_prot_info *prot = &tls_ctx->prot_info;
prot              497 net/tls/tls_sw.c 	if (prot->cipher_type == TLS_CIPHER_AES_CCM_128) {
prot              503 net/tls/tls_sw.c 	       prot->iv_size + prot->salt_size);
prot              505 net/tls/tls_sw.c 	xor_iv_with_seq(prot->version, rec->iv_data, tls_ctx->tx.rec_seq);
prot              507 net/tls/tls_sw.c 	sge->offset += prot->prepend_size;
prot              508 net/tls/tls_sw.c 	sge->length -= prot->prepend_size;
prot              513 net/tls/tls_sw.c 	aead_request_set_ad(aead_req, prot->aad_size);
prot              528 net/tls/tls_sw.c 		sge->offset -= prot->prepend_size;
prot              529 net/tls/tls_sw.c 		sge->length += prot->prepend_size;
prot              541 net/tls/tls_sw.c 	tls_advance_record_sn(sk, prot, &tls_ctx->tx);
prot              667 net/tls/tls_sw.c 	struct tls_prot_info *prot = &tls_ctx->prot_info;
prot              686 net/tls/tls_sw.c 		      prot->overhead_size > msg_en->sg.size) ||
prot              689 net/tls/tls_sw.c 		      prot->overhead_size > msg_en->sg.size))) {
prot              695 net/tls/tls_sw.c 					   split_point, prot->overhead_size,
prot              711 net/tls/tls_sw.c 			    prot->overhead_size);
prot              721 net/tls/tls_sw.c 	if (prot->version == TLS_1_3_VERSION) {
prot              747 net/tls/tls_sw.c 	tls_make_aad(rec->aad_space, msg_pl->sg.size + prot->tail_size,
prot              748 net/tls/tls_sw.c 		     tls_ctx->tx.rec_seq, prot->rec_seq_size,
prot              749 net/tls/tls_sw.c 		     record_type, prot->version);
prot              754 net/tls/tls_sw.c 			 msg_pl->sg.size + prot->tail_size,
prot              755 net/tls/tls_sw.c 			 record_type, prot->version);
prot              760 net/tls/tls_sw.c 			       msg_pl->sg.size + prot->tail_size, i);
prot              774 net/tls/tls_sw.c 		sk_msg_trim(sk, msg_en, msg_pl->sg.size + prot->overhead_size);
prot              916 net/tls/tls_sw.c 	struct tls_prot_info *prot = &tls_ctx->prot_info;
prot              979 net/tls/tls_sw.c 				prot->overhead_size;
prot             1047 net/tls/tls_sw.c 				    msg_pl->sg.size + prot->overhead_size);
prot             1141 net/tls/tls_sw.c 	struct tls_prot_info *prot = &tls_ctx->prot_info;
prot             1183 net/tls/tls_sw.c 		required_size = msg_pl->sg.size + copy + prot->overhead_size;
prot             1397 net/tls/tls_sw.c 	struct tls_prot_info *prot = &tls_ctx->prot_info;
prot             1405 net/tls/tls_sw.c 	const int data_len = rxm->full_len - prot->overhead_size +
prot             1406 net/tls/tls_sw.c 			     prot->tail_size;
prot             1414 net/tls/tls_sw.c 		n_sgin = skb_nsg(skb, rxm->offset + prot->prepend_size,
prot             1415 net/tls/tls_sw.c 				 rxm->full_len - prot->prepend_size);
prot             1432 net/tls/tls_sw.c 	mem_size = mem_size + prot->aad_size;
prot             1448 net/tls/tls_sw.c 	iv = aad + prot->aad_size;
prot             1451 net/tls/tls_sw.c 	if (prot->cipher_type == TLS_CIPHER_AES_CCM_128) {
prot             1458 net/tls/tls_sw.c 			    iv + iv_offset + prot->salt_size,
prot             1459 net/tls/tls_sw.c 			    prot->iv_size);
prot             1464 net/tls/tls_sw.c 	if (prot->version == TLS_1_3_VERSION)
prot             1468 net/tls/tls_sw.c 		memcpy(iv + iv_offset, tls_ctx->rx.iv, prot->salt_size);
prot             1470 net/tls/tls_sw.c 	xor_iv_with_seq(prot->version, iv, tls_ctx->rx.rec_seq);
prot             1473 net/tls/tls_sw.c 	tls_make_aad(aad, rxm->full_len - prot->overhead_size +
prot             1474 net/tls/tls_sw.c 		     prot->tail_size,
prot             1475 net/tls/tls_sw.c 		     tls_ctx->rx.rec_seq, prot->rec_seq_size,
prot             1476 net/tls/tls_sw.c 		     ctx->control, prot->version);
prot             1480 net/tls/tls_sw.c 	sg_set_buf(&sgin[0], aad, prot->aad_size);
prot             1482 net/tls/tls_sw.c 			   rxm->offset + prot->prepend_size,
prot             1483 net/tls/tls_sw.c 			   rxm->full_len - prot->prepend_size);
prot             1492 net/tls/tls_sw.c 			sg_set_buf(&sgout[0], aad, prot->aad_size);
prot             1533 net/tls/tls_sw.c 	struct tls_prot_info *prot = &tls_ctx->prot_info;
prot             1550 net/tls/tls_sw.c 					tls_advance_record_sn(sk, prot,
prot             1559 net/tls/tls_sw.c 		pad = padding_length(ctx, prot, skb);
prot             1564 net/tls/tls_sw.c 		rxm->offset += prot->prepend_size;
prot             1565 net/tls/tls_sw.c 		rxm->full_len -= prot->overhead_size;
prot             1566 net/tls/tls_sw.c 		tls_advance_record_sn(sk, prot, &tls_ctx->rx);
prot             1727 net/tls/tls_sw.c 	struct tls_prot_info *prot = &tls_ctx->prot_info;
prot             1791 net/tls/tls_sw.c 			if (prot->version == TLS_1_3_VERSION)
prot             1799 net/tls/tls_sw.c 		to_decrypt = rxm->full_len - prot->overhead_size;
prot             1803 net/tls/tls_sw.c 		    prot->version != TLS_1_3_VERSION)
prot             1822 net/tls/tls_sw.c 		} else if (prot->version == TLS_1_3_VERSION) {
prot             2023 net/tls/tls_sw.c 	struct tls_prot_info *prot = &tls_ctx->prot_info;
prot             2031 net/tls/tls_sw.c 	if (rxm->offset + prot->prepend_size > skb->len)
prot             2035 net/tls/tls_sw.c 	if (WARN_ON(prot->prepend_size > sizeof(header))) {
prot             2041 net/tls/tls_sw.c 	ret = skb_copy_bits(skb, rxm->offset, header, prot->prepend_size);
prot             2050 net/tls/tls_sw.c 	cipher_overhead = prot->tag_size;
prot             2051 net/tls/tls_sw.c 	if (prot->version != TLS_1_3_VERSION)
prot             2052 net/tls/tls_sw.c 		cipher_overhead += prot->iv_size;
prot             2055 net/tls/tls_sw.c 	    prot->tail_size) {
prot             2261 net/tls/tls_sw.c 	struct tls_prot_info *prot = &tls_ctx->prot_info;
prot             2392 net/tls/tls_sw.c 		prot->aad_size = TLS_HEADER_SIZE;
prot             2393 net/tls/tls_sw.c 		prot->tail_size = 1;
prot             2395 net/tls/tls_sw.c 		prot->aad_size = TLS_AAD_SPACE_SIZE;
prot             2396 net/tls/tls_sw.c 		prot->tail_size = 0;
prot             2399 net/tls/tls_sw.c 	prot->version = crypto_info->version;
prot             2400 net/tls/tls_sw.c 	prot->cipher_type = crypto_info->cipher_type;
prot             2401 net/tls/tls_sw.c 	prot->prepend_size = TLS_HEADER_SIZE + nonce_size;
prot             2402 net/tls/tls_sw.c 	prot->tag_size = tag_size;
prot             2403 net/tls/tls_sw.c 	prot->overhead_size = prot->prepend_size +
prot             2404 net/tls/tls_sw.c 			      prot->tag_size + prot->tail_size;
prot             2405 net/tls/tls_sw.c 	prot->iv_size = iv_size;
prot             2406 net/tls/tls_sw.c 	prot->salt_size = salt_size;
prot             2413 net/tls/tls_sw.c 	prot->rec_seq_size = rec_seq_size;
prot             2438 net/tls/tls_sw.c 	rc = crypto_aead_setauthsize(*aead, prot->tag_size);
prot              481 security/apparmor/lsm.c static int common_mmap(const char *op, struct file *file, unsigned long prot,
prot              489 security/apparmor/lsm.c 	if (prot & PROT_READ)
prot              495 security/apparmor/lsm.c 	if ((prot & PROT_WRITE) && !(flags & MAP_PRIVATE))
prot              497 security/apparmor/lsm.c 	if (prot & PROT_EXEC)
prot              504 security/apparmor/lsm.c 			      unsigned long prot, unsigned long flags)
prot              506 security/apparmor/lsm.c 	return common_mmap(OP_FMMAP, file, prot, flags);
prot              510 security/apparmor/lsm.c 				  unsigned long reqprot, unsigned long prot)
prot              512 security/apparmor/lsm.c 	return common_mmap(OP_FMPROT, vma->vm_file, prot,
prot             1336 security/commoncap.c 		  unsigned long prot, unsigned long flags)
prot              381 security/integrity/ima/ima_main.c int ima_file_mmap(struct file *file, unsigned long prot)
prot              385 security/integrity/ima/ima_main.c 	if (file && (prot & PROT_EXEC)) {
prot             1412 security/security.c static inline unsigned long mmap_prot(struct file *file, unsigned long prot)
prot             1418 security/security.c 	if ((prot & (PROT_READ | PROT_EXEC)) != PROT_READ)
prot             1419 security/security.c 		return prot;
prot             1421 security/security.c 		return prot;
prot             1426 security/security.c 		return prot | PROT_EXEC;
prot             1436 security/security.c 				return prot;
prot             1439 security/security.c 		return prot | PROT_EXEC;
prot             1442 security/security.c 	return prot;
prot             1445 security/security.c int security_mmap_file(struct file *file, unsigned long prot,
prot             1449 security/security.c 	ret = call_int_hook(mmap_file, 0, file, prot,
prot             1450 security/security.c 					mmap_prot(file, prot), flags);
prot             1453 security/security.c 	return ima_file_mmap(file, prot);
prot             1462 security/security.c 			    unsigned long prot)
prot             1464 security/security.c 	return call_int_hook(file_mprotect, 0, vma, reqprot, prot);
prot             3652 security/selinux/hooks.c static int file_map_prot_check(struct file *file, unsigned long prot, int shared)
prot             3659 security/selinux/hooks.c 	    (prot & PROT_EXEC) && (!file || IS_PRIVATE(file_inode(file)) ||
prot             3660 security/selinux/hooks.c 				   (!shared && (prot & PROT_WRITE)))) {
prot             3678 security/selinux/hooks.c 		if (shared && (prot & PROT_WRITE))
prot             3681 security/selinux/hooks.c 		if (prot & PROT_EXEC)
prot             3706 security/selinux/hooks.c 			     unsigned long prot, unsigned long flags)
prot             3721 security/selinux/hooks.c 		prot = reqprot;
prot             3723 security/selinux/hooks.c 	return file_map_prot_check(file, prot,
prot             3729 security/selinux/hooks.c 				 unsigned long prot)
prot             3735 security/selinux/hooks.c 		prot = reqprot;
prot             3738 security/selinux/hooks.c 	    (prot & PROT_EXEC) && !(vma->vm_flags & VM_EXEC)) {
prot             3766 security/selinux/hooks.c 	return file_map_prot_check(vma->vm_file, prot, vma->vm_flags&VM_SHARED);
prot             1657 security/smack/smack_lsm.c 			   unsigned long reqprot, unsigned long prot,
prot               65 sound/core/sgbuf.c 	pgprot_t prot = PAGE_KERNEL;
prot               75 sound/core/sgbuf.c 		prot = pgprot_noncached(PAGE_KERNEL);
prot              125 sound/core/sgbuf.c 	dmab->area = vmap(sgbuf->page_table, sgbuf->pages, VM_MAP, prot);
prot              427 tools/arch/s390/include/uapi/asm/ptrace.h 	ptprot_flags prot;
prot              736 tools/perf/builtin-report.c 				   map->prot & PROT_READ ? 'r' : '-',
prot              737 tools/perf/builtin-report.c 				   map->prot & PROT_WRITE ? 'w' : '-',
prot              738 tools/perf/builtin-report.c 				   map->prot & PROT_EXEC ? 'x' : '-',
prot               30 tools/perf/lib/include/perf/event.h 	__u32			 prot;
prot                9 tools/perf/trace/beauty/mmap.c 	int printed = 0, prot = arg->val;
prot               12 tools/perf/trace/beauty/mmap.c 	if (prot == PROT_NONE)
prot               15 tools/perf/trace/beauty/mmap.c 	if (prot & PROT_##n) { \
prot               17 tools/perf/trace/beauty/mmap.c 		prot &= ~PROT_##n; \
prot               28 tools/perf/trace/beauty/mmap.c 	if (prot)
prot               29 tools/perf/trace/beauty/mmap.c 		printed += scnprintf(bf + printed, size - printed, "%s%#x", printed ? "|" : "", prot);
prot               95 tools/perf/util/auxtrace.c 	mm->base = mmap(NULL, mp->len, mp->prot, MAP_SHARED, fd, mp->offset);
prot              122 tools/perf/util/auxtrace.c 		mp->prot = PROT_READ | (auxtrace_overwrite ? 0 : PROT_WRITE);
prot              297 tools/perf/util/auxtrace.h 	int		prot;
prot              272 tools/perf/util/event.c 		       (event->mmap2.prot & PROT_READ) ? 'r' : '-',
prot              273 tools/perf/util/event.c 		       (event->mmap2.prot & PROT_WRITE) ? 'w' : '-',
prot              274 tools/perf/util/event.c 		       (event->mmap2.prot & PROT_EXEC) ? 'x' : '-',
prot              647 tools/perf/util/evlist.c 		mp->prot = PROT_READ | PROT_WRITE;
prot              660 tools/perf/util/evlist.c 			mp->prot &= ~PROT_WRITE;
prot              465 tools/perf/util/jitdump.c 	event->mmap2.prot  = st.st_mode;
prot              557 tools/perf/util/jitdump.c 	event->mmap2.prot  = st.st_mode;
prot             1677 tools/perf/util/machine.c 			event->mmap2.prot,
prot             1706 tools/perf/util/machine.c 	u32 prot = 0;
prot             1726 tools/perf/util/machine.c 		prot = PROT_EXEC;
prot             1730 tools/perf/util/machine.c 			0, 0, 0, 0, prot, 0,
prot              150 tools/perf/util/map.c 		     u64 ino_gen, u32 prot, u32 flags, char *filename,
prot              171 tools/perf/util/map.c 		map->prot = prot;
prot              175 tools/perf/util/map.c 		if ((anon || no_dso) && nsi && (prot & PROT_EXEC)) {
prot              215 tools/perf/util/map.c 			if (!(prot & PROT_EXEC))
prot               31 tools/perf/util/map.h 	u32			prot;
prot              115 tools/perf/util/map.h 		     u64 ino_gen, u32 prot, u32 flags,
prot              374 tools/perf/util/mmap.c 	map->core.base = mmap(NULL, perf_mmap__mmap_len(map), mp->prot,
prot               40 tools/perf/util/mmap.h 	int prot, mask, nr_cblocks, affinity, flush, comp_level;
prot             1278 tools/perf/util/sort.c 		     map && !(map->prot & PROT_EXEC) &&
prot             1302 tools/perf/util/symbol.c 	err = file__read_maps(fd, map->prot & PROT_EXEC, kcore_mapfn, &md,
prot             1394 tools/perf/util/symbol.c 	if (map->prot & PROT_EXEC)
prot              312 tools/perf/util/synthetic-events.c 		char prot[5];
prot              336 tools/perf/util/synthetic-events.c 		       &event->mmap2.start, &event->mmap2.len, prot,
prot              358 tools/perf/util/synthetic-events.c 		event->mmap2.prot = 0;
prot              360 tools/perf/util/synthetic-events.c 		if (prot[0] == 'r')
prot              361 tools/perf/util/synthetic-events.c 			event->mmap2.prot |= PROT_READ;
prot              362 tools/perf/util/synthetic-events.c 		if (prot[1] == 'w')
prot              363 tools/perf/util/synthetic-events.c 			event->mmap2.prot |= PROT_WRITE;
prot              364 tools/perf/util/synthetic-events.c 		if (prot[2] == 'x')
prot              365 tools/perf/util/synthetic-events.c 			event->mmap2.prot |= PROT_EXEC;
prot              367 tools/perf/util/synthetic-events.c 		if (prot[3] == 's')
prot              372 tools/perf/util/synthetic-events.c 		if (prot[2] != 'x') {
prot              373 tools/perf/util/synthetic-events.c 			if (!mmap_data || prot[0] != 'r')
prot              680 tools/testing/selftests/x86/protection_keys.c 	int prot;
prot              685 tools/testing/selftests/x86/protection_keys.c void record_pkey_malloc(void *ptr, long size, int prot)
prot              717 tools/testing/selftests/x86/protection_keys.c 	rec->prot = prot;
prot              750 tools/testing/selftests/x86/protection_keys.c void *malloc_pkey_with_mprotect(long size, int prot, u16 pkey)
prot              757 tools/testing/selftests/x86/protection_keys.c 			size, prot, pkey);
prot              759 tools/testing/selftests/x86/protection_keys.c 	ptr = mmap(NULL, size, prot, MAP_ANONYMOUS|MAP_PRIVATE, -1, 0);
prot              761 tools/testing/selftests/x86/protection_keys.c 	ret = mprotect_pkey((void *)ptr, PAGE_SIZE, prot, pkey);
prot              763 tools/testing/selftests/x86/protection_keys.c 	record_pkey_malloc(ptr, size, prot);
prot              770 tools/testing/selftests/x86/protection_keys.c void *malloc_pkey_anon_huge(long size, int prot, u16 pkey)
prot              776 tools/testing/selftests/x86/protection_keys.c 			size, prot, pkey);
prot              784 tools/testing/selftests/x86/protection_keys.c 	record_pkey_malloc(ptr, size, prot);
prot              785 tools/testing/selftests/x86/protection_keys.c 	mprotect_pkey(ptr, size, prot, pkey);
prot              842 tools/testing/selftests/x86/protection_keys.c void *malloc_pkey_hugetlb(long size, int prot, u16 pkey)
prot              850 tools/testing/selftests/x86/protection_keys.c 	dprintf1("doing %s(%ld, %x, %x)\n", __func__, size, prot, pkey);
prot              855 tools/testing/selftests/x86/protection_keys.c 	mprotect_pkey(ptr, size, prot, pkey);
prot              857 tools/testing/selftests/x86/protection_keys.c 	record_pkey_malloc(ptr, size, prot);
prot              863 tools/testing/selftests/x86/protection_keys.c void *malloc_pkey_mmap_dax(long size, int prot, u16 pkey)
prot              869 tools/testing/selftests/x86/protection_keys.c 			size, prot, pkey);
prot              874 tools/testing/selftests/x86/protection_keys.c 	ptr = mmap(0, size, prot, MAP_SHARED, fd, 0);
prot              877 tools/testing/selftests/x86/protection_keys.c 	mprotect_pkey(ptr, size, prot, pkey);
prot              879 tools/testing/selftests/x86/protection_keys.c 	record_pkey_malloc(ptr, size, prot);
prot              886 tools/testing/selftests/x86/protection_keys.c void *(*pkey_malloc[])(long size, int prot, u16 pkey) = {
prot              897 tools/testing/selftests/x86/protection_keys.c void *malloc_pkey(long size, int prot, u16 pkey)
prot              908 tools/testing/selftests/x86/protection_keys.c 		ret = pkey_malloc[malloc_type](size, prot, pkey);
prot              923 tools/testing/selftests/x86/protection_keys.c 			size, prot, pkey, ret);
prot             1216 tools/testing/selftests/x86/protection_keys.c 	int prot;
prot             1228 tools/testing/selftests/x86/protection_keys.c 	prot = pkey_last_malloc_record->prot;
prot             1231 tools/testing/selftests/x86/protection_keys.c 	mprotect_pkey(ptr, size, prot, 0);
prot             1234 tools/testing/selftests/x86/protection_keys.c 	mprotect_pkey(ptr, size, prot, pkey);
prot             1441 tools/testing/selftests/x86/protection_keys.c 	int prot = PROT_READ|PROT_WRITE;
prot             1453 tools/testing/selftests/x86/protection_keys.c 		ptr = malloc_pkey(PAGE_SIZE, prot, pkey);
prot              612 virt/kvm/arm/mmu.c 				    pgprot_t prot)
prot              620 virt/kvm/arm/mmu.c 		kvm_set_pte(pte, kvm_pfn_pte(pfn, prot));
prot              628 virt/kvm/arm/mmu.c 				   pgprot_t prot)
prot              652 virt/kvm/arm/mmu.c 		create_hyp_pte_mappings(pmd, addr, next, pfn, prot);
prot              661 virt/kvm/arm/mmu.c 				   pgprot_t prot)
prot              683 virt/kvm/arm/mmu.c 		ret = create_hyp_pmd_mappings(pud, addr, next, pfn, prot);
prot              694 virt/kvm/arm/mmu.c 				 unsigned long pfn, pgprot_t prot)
prot              719 virt/kvm/arm/mmu.c 		err = create_hyp_pud_mappings(pgd, addr, next, pfn, prot);
prot              750 virt/kvm/arm/mmu.c int create_hyp_mappings(void *from, void *to, pgprot_t prot)
prot              770 virt/kvm/arm/mmu.c 					    prot);
prot              779 virt/kvm/arm/mmu.c 					unsigned long *haddr, pgprot_t prot)
prot              818 virt/kvm/arm/mmu.c 				    __phys_to_pfn(phys_addr), prot);