prot_sect 206 arch/arm/mm/ioremap.c pmd[0] = __pmd(__pfn_to_phys(pfn) | type->prot_sect); prot_sect 208 arch/arm/mm/ioremap.c pmd[1] = __pmd(__pfn_to_phys(pfn) | type->prot_sect); prot_sect 240 arch/arm/mm/ioremap.c super_pmd_val = __pfn_to_phys(pfn) | type->prot_sect | prot_sect 48 arch/arm/mm/mm.h pmdval_t prot_sect; prot_sect 253 arch/arm/mm/mmu.c .prot_sect = PROT_SECT_DEVICE | PMD_SECT_S, prot_sect 259 arch/arm/mm/mmu.c .prot_sect = PROT_SECT_DEVICE, prot_sect 265 arch/arm/mm/mmu.c .prot_sect = PROT_SECT_DEVICE | PMD_SECT_WB, prot_sect 271 arch/arm/mm/mmu.c .prot_sect = PROT_SECT_DEVICE, prot_sect 277 arch/arm/mm/mmu.c .prot_sect = PMD_TYPE_SECT | PMD_SECT_XN, prot_sect 281 arch/arm/mm/mmu.c .prot_sect = PMD_TYPE_SECT | PMD_SECT_XN, prot_sect 286 arch/arm/mm/mmu.c .prot_sect = PMD_TYPE_SECT | PMD_SECT_XN | PMD_SECT_MINICACHE, prot_sect 305 arch/arm/mm/mmu.c .prot_sect = PMD_TYPE_SECT | PMD_SECT_AP_WRITE, prot_sect 312 arch/arm/mm/mmu.c .prot_sect = PMD_TYPE_SECT | PMD_SECT_AP_WRITE, prot_sect 316 arch/arm/mm/mmu.c .prot_sect = PMD_TYPE_SECT, prot_sect 323 arch/arm/mm/mmu.c .prot_sect = PMD_TYPE_SECT | PMD_SECT_AP_WRITE, prot_sect 330 arch/arm/mm/mmu.c .prot_sect = PMD_TYPE_SECT | PMD_SECT_XN, prot_sect 342 arch/arm/mm/mmu.c .prot_sect = PMD_TYPE_SECT | PMD_SECT_AP_WRITE | PMD_SECT_S | prot_sect 474 arch/arm/mm/mmu.c mem_types[i].prot_sect &= ~PMD_SECT_TEX(7); prot_sect 477 arch/arm/mm/mmu.c mem_types[i].prot_sect &= ~PMD_SECT_S; prot_sect 486 arch/arm/mm/mmu.c mem_types[i].prot_sect &= ~PMD_BIT4; prot_sect 493 arch/arm/mm/mmu.c if (mem_types[i].prot_sect) prot_sect 494 arch/arm/mm/mmu.c mem_types[i].prot_sect |= PMD_BIT4; prot_sect 507 arch/arm/mm/mmu.c mem_types[MT_DEVICE].prot_sect |= PMD_SECT_XN; prot_sect 508 arch/arm/mm/mmu.c mem_types[MT_DEVICE_NONSHARED].prot_sect |= PMD_SECT_XN; prot_sect 509 arch/arm/mm/mmu.c mem_types[MT_DEVICE_CACHED].prot_sect |= PMD_SECT_XN; prot_sect 510 arch/arm/mm/mmu.c mem_types[MT_DEVICE_WC].prot_sect |= PMD_SECT_XN; prot_sect 513 arch/arm/mm/mmu.c mem_types[MT_MEMORY_RW].prot_sect |= PMD_SECT_XN; prot_sect 523 arch/arm/mm/mmu.c mem_types[MT_DEVICE].prot_sect |= PMD_SECT_TEX(1); prot_sect 524 arch/arm/mm/mmu.c mem_types[MT_DEVICE_NONSHARED].prot_sect |= PMD_SECT_TEX(1); prot_sect 525 arch/arm/mm/mmu.c mem_types[MT_DEVICE_WC].prot_sect |= PMD_SECT_BUFFERABLE; prot_sect 534 arch/arm/mm/mmu.c mem_types[MT_DEVICE].prot_sect |= PMD_SECT_TEX(1) | PMD_SECT_BUFFERED; prot_sect 535 arch/arm/mm/mmu.c mem_types[MT_DEVICE_NONSHARED].prot_sect |= PMD_SECT_TEX(2); prot_sect 536 arch/arm/mm/mmu.c mem_types[MT_DEVICE_WC].prot_sect |= PMD_SECT_TEX(1); prot_sect 545 arch/arm/mm/mmu.c mem_types[MT_DEVICE].prot_sect |= PMD_SECT_BUFFERED; prot_sect 546 arch/arm/mm/mmu.c mem_types[MT_DEVICE_NONSHARED].prot_sect |= PMD_SECT_TEX(2); prot_sect 547 arch/arm/mm/mmu.c mem_types[MT_DEVICE_WC].prot_sect |= PMD_SECT_TEX(1); prot_sect 553 arch/arm/mm/mmu.c mem_types[MT_DEVICE_WC].prot_sect |= PMD_SECT_BUFFERABLE; prot_sect 593 arch/arm/mm/mmu.c mem_types[MT_ROM].prot_sect |= PMD_SECT_APX|PMD_SECT_AP_WRITE; prot_sect 594 arch/arm/mm/mmu.c mem_types[MT_MINICLEAN].prot_sect |= PMD_SECT_APX|PMD_SECT_AP_WRITE; prot_sect 595 arch/arm/mm/mmu.c mem_types[MT_CACHECLEAN].prot_sect |= PMD_SECT_APX|PMD_SECT_AP_WRITE; prot_sect 608 arch/arm/mm/mmu.c mem_types[MT_DEVICE_WC].prot_sect |= PMD_SECT_S; prot_sect 610 arch/arm/mm/mmu.c mem_types[MT_DEVICE_CACHED].prot_sect |= PMD_SECT_S; prot_sect 612 arch/arm/mm/mmu.c mem_types[MT_MEMORY_RWX].prot_sect |= PMD_SECT_S; prot_sect 614 arch/arm/mm/mmu.c mem_types[MT_MEMORY_RW].prot_sect |= PMD_SECT_S; prot_sect 617 arch/arm/mm/mmu.c mem_types[MT_MEMORY_RWX_NONCACHED].prot_sect |= PMD_SECT_S; prot_sect 629 arch/arm/mm/mmu.c mem_types[MT_MEMORY_RWX_NONCACHED].prot_sect |= prot_sect 633 arch/arm/mm/mmu.c mem_types[MT_MEMORY_RWX_NONCACHED].prot_sect |= prot_sect 637 arch/arm/mm/mmu.c mem_types[MT_MEMORY_RWX_NONCACHED].prot_sect |= PMD_SECT_BUFFERABLE; prot_sect 646 arch/arm/mm/mmu.c if (mem_types[i].prot_sect) prot_sect 647 arch/arm/mm/mmu.c mem_types[i].prot_sect |= PMD_SECT_AF; prot_sect 675 arch/arm/mm/mmu.c mem_types[MT_MEMORY_RWX].prot_sect |= ecc_mask | cp->pmd; prot_sect 677 arch/arm/mm/mmu.c mem_types[MT_MEMORY_RW].prot_sect |= ecc_mask | cp->pmd; prot_sect 680 arch/arm/mm/mmu.c mem_types[MT_MEMORY_RWX_NONCACHED].prot_sect |= ecc_mask; prot_sect 681 arch/arm/mm/mmu.c mem_types[MT_ROM].prot_sect |= cp->pmd; prot_sect 685 arch/arm/mm/mmu.c mem_types[MT_CACHECLEAN].prot_sect |= PMD_SECT_WT; prot_sect 689 arch/arm/mm/mmu.c mem_types[MT_CACHECLEAN].prot_sect |= PMD_SECT_WB; prot_sect 699 arch/arm/mm/mmu.c if (t->prot_sect) prot_sect 700 arch/arm/mm/mmu.c t->prot_sect |= PMD_DOMAIN(t->domain); prot_sect 791 arch/arm/mm/mmu.c *pmd = __pmd(phys | type->prot_sect | (ng ? PMD_SECT_nG : 0)); prot_sect 817 arch/arm/mm/mmu.c if (type->prot_sect && prot_sect 897 arch/arm/mm/mmu.c *pmd++ = __pmd(phys | type->prot_sect | PMD_SECT_SUPER | prot_sect 114 arch/unicore32/mm/ioremap.c set_pmd(pmd, __pmd(__pfn_to_phys(pfn) | type->prot_sect)); prot_sect 30 arch/unicore32/mm/mm.h unsigned int prot_sect; prot_sect 91 arch/unicore32/mm/mmu.c .prot_sect = PROT_SECT_DEVICE, prot_sect 101 arch/unicore32/mm/mmu.c .prot_sect = PROT_SECT_DEVICE, prot_sect 113 arch/unicore32/mm/mmu.c .prot_sect = PMD_TYPE_SECT | PMD_PRESENT | PMD_SECT_CACHEABLE | prot_sect 117 arch/unicore32/mm/mmu.c .prot_sect = PMD_TYPE_SECT | PMD_PRESENT | PMD_SECT_CACHEABLE | prot_sect 183 arch/unicore32/mm/mmu.c set_pmd(pmd, __pmd(phys | type->prot_sect));