Searched refs:mem_types (Results 1 - 3 of 3) sorted by relevance

/linux-4.4.14/arch/arm/mm/
H A Dmmu.c246 static struct mem_type mem_types[] = { variable in typeref:struct:mem_type
357 return type < ARRAY_SIZE(mem_types) ? &mem_types[type] : NULL; get_mem_type()
469 for (i = 0; i < ARRAY_SIZE(mem_types); i++) build_mem_type_table()
470 mem_types[i].prot_sect &= ~PMD_SECT_TEX(7); build_mem_type_table()
472 for (i = 0; i < ARRAY_SIZE(mem_types); i++) build_mem_type_table()
473 mem_types[i].prot_sect &= ~PMD_SECT_S; build_mem_type_table()
481 for (i = 0; i < ARRAY_SIZE(mem_types); i++) { build_mem_type_table()
482 mem_types[i].prot_sect &= ~PMD_BIT4; build_mem_type_table()
483 mem_types[i].prot_l1 &= ~PMD_BIT4; build_mem_type_table()
486 for (i = 0; i < ARRAY_SIZE(mem_types); i++) { build_mem_type_table()
487 if (mem_types[i].prot_l1) build_mem_type_table()
488 mem_types[i].prot_l1 |= PMD_BIT4; build_mem_type_table()
489 if (mem_types[i].prot_sect) build_mem_type_table()
490 mem_types[i].prot_sect |= PMD_BIT4; build_mem_type_table()
503 mem_types[MT_DEVICE].prot_sect |= PMD_SECT_XN; build_mem_type_table()
504 mem_types[MT_DEVICE_NONSHARED].prot_sect |= PMD_SECT_XN; build_mem_type_table()
505 mem_types[MT_DEVICE_CACHED].prot_sect |= PMD_SECT_XN; build_mem_type_table()
506 mem_types[MT_DEVICE_WC].prot_sect |= PMD_SECT_XN; build_mem_type_table()
509 mem_types[MT_MEMORY_RW].prot_sect |= PMD_SECT_XN; build_mem_type_table()
519 mem_types[MT_DEVICE].prot_sect |= PMD_SECT_TEX(1); build_mem_type_table()
520 mem_types[MT_DEVICE_NONSHARED].prot_sect |= PMD_SECT_TEX(1); build_mem_type_table()
521 mem_types[MT_DEVICE_WC].prot_sect |= PMD_SECT_BUFFERABLE; build_mem_type_table()
530 mem_types[MT_DEVICE].prot_sect |= PMD_SECT_TEX(1) | PMD_SECT_BUFFERED; build_mem_type_table()
531 mem_types[MT_DEVICE_NONSHARED].prot_sect |= PMD_SECT_TEX(2); build_mem_type_table()
532 mem_types[MT_DEVICE_WC].prot_sect |= PMD_SECT_TEX(1); build_mem_type_table()
541 mem_types[MT_DEVICE].prot_sect |= PMD_SECT_BUFFERED; build_mem_type_table()
542 mem_types[MT_DEVICE_NONSHARED].prot_sect |= PMD_SECT_TEX(2); build_mem_type_table()
543 mem_types[MT_DEVICE_WC].prot_sect |= PMD_SECT_TEX(1); build_mem_type_table()
549 mem_types[MT_DEVICE_WC].prot_sect |= PMD_SECT_BUFFERABLE; build_mem_type_table()
558 hyp_device_pgprot = mem_types[MT_DEVICE].prot_pte; build_mem_type_table()
559 s2_device_pgprot = mem_types[MT_DEVICE].prot_pte_s2; build_mem_type_table()
589 mem_types[MT_ROM].prot_sect |= PMD_SECT_APX|PMD_SECT_AP_WRITE; build_mem_type_table()
590 mem_types[MT_MINICLEAN].prot_sect |= PMD_SECT_APX|PMD_SECT_AP_WRITE; build_mem_type_table()
591 mem_types[MT_CACHECLEAN].prot_sect |= PMD_SECT_APX|PMD_SECT_AP_WRITE; build_mem_type_table()
604 mem_types[MT_DEVICE_WC].prot_sect |= PMD_SECT_S; build_mem_type_table()
605 mem_types[MT_DEVICE_WC].prot_pte |= L_PTE_SHARED; build_mem_type_table()
606 mem_types[MT_DEVICE_CACHED].prot_sect |= PMD_SECT_S; build_mem_type_table()
607 mem_types[MT_DEVICE_CACHED].prot_pte |= L_PTE_SHARED; build_mem_type_table()
608 mem_types[MT_MEMORY_RWX].prot_sect |= PMD_SECT_S; build_mem_type_table()
609 mem_types[MT_MEMORY_RWX].prot_pte |= L_PTE_SHARED; build_mem_type_table()
610 mem_types[MT_MEMORY_RW].prot_sect |= PMD_SECT_S; build_mem_type_table()
611 mem_types[MT_MEMORY_RW].prot_pte |= L_PTE_SHARED; build_mem_type_table()
612 mem_types[MT_MEMORY_DMA_READY].prot_pte |= L_PTE_SHARED; build_mem_type_table()
613 mem_types[MT_MEMORY_RWX_NONCACHED].prot_sect |= PMD_SECT_S; build_mem_type_table()
614 mem_types[MT_MEMORY_RWX_NONCACHED].prot_pte |= L_PTE_SHARED; build_mem_type_table()
625 mem_types[MT_MEMORY_RWX_NONCACHED].prot_sect |= build_mem_type_table()
629 mem_types[MT_MEMORY_RWX_NONCACHED].prot_sect |= build_mem_type_table()
633 mem_types[MT_MEMORY_RWX_NONCACHED].prot_sect |= PMD_SECT_BUFFERABLE; build_mem_type_table()
640 for (i = 0; i < ARRAY_SIZE(mem_types); i++) { build_mem_type_table()
641 mem_types[i].prot_pte |= PTE_EXT_AF; build_mem_type_table()
642 if (mem_types[i].prot_sect) build_mem_type_table()
643 mem_types[i].prot_sect |= PMD_SECT_AF; build_mem_type_table()
659 mem_types[MT_LOW_VECTORS].prot_pte |= vecs_pgprot; build_mem_type_table()
660 mem_types[MT_HIGH_VECTORS].prot_pte |= vecs_pgprot; build_mem_type_table()
669 mem_types[MT_LOW_VECTORS].prot_l1 |= ecc_mask; build_mem_type_table()
670 mem_types[MT_HIGH_VECTORS].prot_l1 |= ecc_mask; build_mem_type_table()
671 mem_types[MT_MEMORY_RWX].prot_sect |= ecc_mask | cp->pmd; build_mem_type_table()
672 mem_types[MT_MEMORY_RWX].prot_pte |= kern_pgprot; build_mem_type_table()
673 mem_types[MT_MEMORY_RW].prot_sect |= ecc_mask | cp->pmd; build_mem_type_table()
674 mem_types[MT_MEMORY_RW].prot_pte |= kern_pgprot; build_mem_type_table()
675 mem_types[MT_MEMORY_DMA_READY].prot_pte |= kern_pgprot; build_mem_type_table()
676 mem_types[MT_MEMORY_RWX_NONCACHED].prot_sect |= ecc_mask; build_mem_type_table()
677 mem_types[MT_ROM].prot_sect |= cp->pmd; build_mem_type_table()
681 mem_types[MT_CACHECLEAN].prot_sect |= PMD_SECT_WT; build_mem_type_table()
685 mem_types[MT_CACHECLEAN].prot_sect |= PMD_SECT_WB; build_mem_type_table()
691 for (i = 0; i < ARRAY_SIZE(mem_types); i++) { build_mem_type_table()
692 struct mem_type *t = &mem_types[i]; build_mem_type_table()
906 type = &mem_types[md->type]; create_mapping()
/linux-4.4.14/arch/unicore32/mm/
H A Dmmu.c91 static struct mem_type mem_types[] = { variable in typeref:struct:mem_type
128 return type < ARRAY_SIZE(mem_types) ? &mem_types[type] : NULL; get_mem_type()
228 type = &mem_types[md->type]; create_mapping()
/linux-4.4.14/drivers/edac/
H A Dedac_mc_sysfs.c94 static const char * const mem_types[] = { variable
199 return sprintf(data, "%s\n", mem_types[csrow->channels[0]->dimm->mtype]); csrow_mem_type_show()
534 return sprintf(data, "%s\n", mem_types[dimm->mtype]); dimmdev_mem_type_show()

Completed in 153 milliseconds