Searched refs:mem_types (Results 1 - 3 of 3) sorted by relevance

/linux-4.1.27/arch/arm/mm/
H A Dmmu.c245 static struct mem_type mem_types[] = { variable in typeref:struct:mem_type
356 return type < ARRAY_SIZE(mem_types) ? &mem_types[type] : NULL; get_mem_type()
427 for (i = 0; i < ARRAY_SIZE(mem_types); i++) build_mem_type_table()
428 mem_types[i].prot_sect &= ~PMD_SECT_TEX(7); build_mem_type_table()
430 for (i = 0; i < ARRAY_SIZE(mem_types); i++) build_mem_type_table()
431 mem_types[i].prot_sect &= ~PMD_SECT_S; build_mem_type_table()
439 for (i = 0; i < ARRAY_SIZE(mem_types); i++) { build_mem_type_table()
440 mem_types[i].prot_sect &= ~PMD_BIT4; build_mem_type_table()
441 mem_types[i].prot_l1 &= ~PMD_BIT4; build_mem_type_table()
444 for (i = 0; i < ARRAY_SIZE(mem_types); i++) { build_mem_type_table()
445 if (mem_types[i].prot_l1) build_mem_type_table()
446 mem_types[i].prot_l1 |= PMD_BIT4; build_mem_type_table()
447 if (mem_types[i].prot_sect) build_mem_type_table()
448 mem_types[i].prot_sect |= PMD_BIT4; build_mem_type_table()
461 mem_types[MT_DEVICE].prot_sect |= PMD_SECT_XN; build_mem_type_table()
462 mem_types[MT_DEVICE_NONSHARED].prot_sect |= PMD_SECT_XN; build_mem_type_table()
463 mem_types[MT_DEVICE_CACHED].prot_sect |= PMD_SECT_XN; build_mem_type_table()
464 mem_types[MT_DEVICE_WC].prot_sect |= PMD_SECT_XN; build_mem_type_table()
467 mem_types[MT_MEMORY_RW].prot_sect |= PMD_SECT_XN; build_mem_type_table()
477 mem_types[MT_DEVICE].prot_sect |= PMD_SECT_TEX(1); build_mem_type_table()
478 mem_types[MT_DEVICE_NONSHARED].prot_sect |= PMD_SECT_TEX(1); build_mem_type_table()
479 mem_types[MT_DEVICE_WC].prot_sect |= PMD_SECT_BUFFERABLE; build_mem_type_table()
488 mem_types[MT_DEVICE].prot_sect |= PMD_SECT_TEX(1) | PMD_SECT_BUFFERED; build_mem_type_table()
489 mem_types[MT_DEVICE_NONSHARED].prot_sect |= PMD_SECT_TEX(2); build_mem_type_table()
490 mem_types[MT_DEVICE_WC].prot_sect |= PMD_SECT_TEX(1); build_mem_type_table()
499 mem_types[MT_DEVICE].prot_sect |= PMD_SECT_BUFFERED; build_mem_type_table()
500 mem_types[MT_DEVICE_NONSHARED].prot_sect |= PMD_SECT_TEX(2); build_mem_type_table()
501 mem_types[MT_DEVICE_WC].prot_sect |= PMD_SECT_TEX(1); build_mem_type_table()
507 mem_types[MT_DEVICE_WC].prot_sect |= PMD_SECT_BUFFERABLE; build_mem_type_table()
516 hyp_device_pgprot = mem_types[MT_DEVICE].prot_pte; build_mem_type_table()
517 s2_device_pgprot = mem_types[MT_DEVICE].prot_pte_s2; build_mem_type_table()
547 mem_types[MT_ROM].prot_sect |= PMD_SECT_APX|PMD_SECT_AP_WRITE; build_mem_type_table()
548 mem_types[MT_MINICLEAN].prot_sect |= PMD_SECT_APX|PMD_SECT_AP_WRITE; build_mem_type_table()
549 mem_types[MT_CACHECLEAN].prot_sect |= PMD_SECT_APX|PMD_SECT_AP_WRITE; build_mem_type_table()
562 mem_types[MT_DEVICE_WC].prot_sect |= PMD_SECT_S; build_mem_type_table()
563 mem_types[MT_DEVICE_WC].prot_pte |= L_PTE_SHARED; build_mem_type_table()
564 mem_types[MT_DEVICE_CACHED].prot_sect |= PMD_SECT_S; build_mem_type_table()
565 mem_types[MT_DEVICE_CACHED].prot_pte |= L_PTE_SHARED; build_mem_type_table()
566 mem_types[MT_MEMORY_RWX].prot_sect |= PMD_SECT_S; build_mem_type_table()
567 mem_types[MT_MEMORY_RWX].prot_pte |= L_PTE_SHARED; build_mem_type_table()
568 mem_types[MT_MEMORY_RW].prot_sect |= PMD_SECT_S; build_mem_type_table()
569 mem_types[MT_MEMORY_RW].prot_pte |= L_PTE_SHARED; build_mem_type_table()
570 mem_types[MT_MEMORY_DMA_READY].prot_pte |= L_PTE_SHARED; build_mem_type_table()
571 mem_types[MT_MEMORY_RWX_NONCACHED].prot_sect |= PMD_SECT_S; build_mem_type_table()
572 mem_types[MT_MEMORY_RWX_NONCACHED].prot_pte |= L_PTE_SHARED; build_mem_type_table()
583 mem_types[MT_MEMORY_RWX_NONCACHED].prot_sect |= build_mem_type_table()
587 mem_types[MT_MEMORY_RWX_NONCACHED].prot_sect |= build_mem_type_table()
591 mem_types[MT_MEMORY_RWX_NONCACHED].prot_sect |= PMD_SECT_BUFFERABLE; build_mem_type_table()
598 for (i = 0; i < ARRAY_SIZE(mem_types); i++) { build_mem_type_table()
599 mem_types[i].prot_pte |= PTE_EXT_AF; build_mem_type_table()
600 if (mem_types[i].prot_sect) build_mem_type_table()
601 mem_types[i].prot_sect |= PMD_SECT_AF; build_mem_type_table()
617 mem_types[MT_LOW_VECTORS].prot_pte |= vecs_pgprot; build_mem_type_table()
618 mem_types[MT_HIGH_VECTORS].prot_pte |= vecs_pgprot; build_mem_type_table()
627 mem_types[MT_LOW_VECTORS].prot_l1 |= ecc_mask; build_mem_type_table()
628 mem_types[MT_HIGH_VECTORS].prot_l1 |= ecc_mask; build_mem_type_table()
629 mem_types[MT_MEMORY_RWX].prot_sect |= ecc_mask | cp->pmd; build_mem_type_table()
630 mem_types[MT_MEMORY_RWX].prot_pte |= kern_pgprot; build_mem_type_table()
631 mem_types[MT_MEMORY_RW].prot_sect |= ecc_mask | cp->pmd; build_mem_type_table()
632 mem_types[MT_MEMORY_RW].prot_pte |= kern_pgprot; build_mem_type_table()
633 mem_types[MT_MEMORY_DMA_READY].prot_pte |= kern_pgprot; build_mem_type_table()
634 mem_types[MT_MEMORY_RWX_NONCACHED].prot_sect |= ecc_mask; build_mem_type_table()
635 mem_types[MT_ROM].prot_sect |= cp->pmd; build_mem_type_table()
639 mem_types[MT_CACHECLEAN].prot_sect |= PMD_SECT_WT; build_mem_type_table()
643 mem_types[MT_CACHECLEAN].prot_sect |= PMD_SECT_WB; build_mem_type_table()
649 for (i = 0; i < ARRAY_SIZE(mem_types); i++) { build_mem_type_table()
650 struct mem_type *t = &mem_types[i]; build_mem_type_table()
864 type = &mem_types[md->type]; create_mapping()
/linux-4.1.27/arch/unicore32/mm/
H A Dmmu.c91 static struct mem_type mem_types[] = { variable in typeref:struct:mem_type
128 return type < ARRAY_SIZE(mem_types) ? &mem_types[type] : NULL; get_mem_type()
228 type = &mem_types[md->type]; create_mapping()
/linux-4.1.27/drivers/edac/
H A Dedac_mc_sysfs.c94 static const char * const mem_types[] = { variable
199 return sprintf(data, "%s\n", mem_types[csrow->channels[0]->dimm->mtype]); csrow_mem_type_show()
525 return sprintf(data, "%s\n", mem_types[dimm->mtype]); dimmdev_mem_type_show()

Completed in 181 milliseconds