m_end            1008 arch/sparc/mm/init_64.c 	u64 ret_end, pa_start, m_mask, m_match, m_end;
m_end            1073 arch/sparc/mm/init_64.c 	m_end = m_match + (1ul << __ffs(m_mask)) - mblock->offset;
m_end            1074 arch/sparc/mm/init_64.c 	m_end += pa_start & ~((1ul << fls64(m_mask)) - 1);
m_end            1075 arch/sparc/mm/init_64.c 	ret_end = m_end > end ? end : m_end;
m_end            1105 arch/x86/kernel/cpu/amd.c #define AMD_MODEL_RANGE(f, m_start, s_start, m_end, s_end) \
m_end            1106 arch/x86/kernel/cpu/amd.c 	((f << 24) | (m_start << 16) | (s_start << 12) | (m_end << 4) | (s_end))
m_end             219 drivers/firmware/efi/memmap.c 	u64 m_start, m_end;
m_end             228 drivers/firmware/efi/memmap.c 	m_end = range->end;
m_end             232 drivers/firmware/efi/memmap.c 		if (start < m_end && m_end < end)
m_end             238 drivers/firmware/efi/memmap.c 		if (m_end < end)
m_end             241 drivers/firmware/efi/memmap.c 		if (end <= m_end)
m_end             260 drivers/firmware/efi/memmap.c 	u64 m_start, m_end, m_attr;
m_end             267 drivers/firmware/efi/memmap.c 	m_end = mem->range.end;
m_end             276 drivers/firmware/efi/memmap.c 	    !IS_ALIGNED(m_end + 1, EFI_PAGE_SIZE)) {
m_end             291 drivers/firmware/efi/memmap.c 		if (m_start <= start && end <= m_end)
m_end             295 drivers/firmware/efi/memmap.c 		    (start < m_end && m_end < end)) {
m_end             298 drivers/firmware/efi/memmap.c 			md->num_pages = (m_end - md->phys_addr + 1) >>
m_end             304 drivers/firmware/efi/memmap.c 			md->phys_addr = m_end + 1;
m_end             309 drivers/firmware/efi/memmap.c 		if ((start < m_start && m_start < end) && m_end < end) {
m_end             319 drivers/firmware/efi/memmap.c 			md->num_pages = (m_end - m_start + 1) >>
m_end             325 drivers/firmware/efi/memmap.c 			md->phys_addr = m_end + 1;
m_end             326 drivers/firmware/efi/memmap.c 			md->num_pages = (end - m_end) >>
m_end             331 drivers/firmware/efi/memmap.c 		    (end <= m_end)) {
m_end             464 fs/ntfs/mft.c  	unsigned int block_start, block_end, m_start, m_end, page_ofs;
m_end             515 fs/ntfs/mft.c  	m_end = m_start + vol->mft_record_size;
m_end             521 fs/ntfs/mft.c  		if (unlikely(block_start >= m_end))
m_end             571 fs/ntfs/mft.c  		BUG_ON((nr_bhs >= max_bhs) && (m_end != block_end));
m_end             670 fs/ntfs/mft.c  	unsigned int block_start, block_end, m_start, m_end;
m_end             695 fs/ntfs/mft.c  	m_end = m_start + vol->mft_record_size;
m_end             701 fs/ntfs/mft.c  		if (unlikely(block_start >= m_end))
m_end             758 fs/ntfs/mft.c  		BUG_ON((nr_bhs >= max_bhs) && (m_end != block_end));
m_end            1032 mm/memblock.c  		phys_addr_t m_end = m->base + m->size;
m_end            1042 mm/memblock.c  				*out_end = m_end;
m_end            1065 mm/memblock.c  			if (r_start >= m_end)
m_end            1073 mm/memblock.c  					*out_end = min(m_end, r_end);
m_end            1080 mm/memblock.c  				if (m_end <= r_end)
m_end            1136 mm/memblock.c  		phys_addr_t m_end = m->base + m->size;
m_end            1146 mm/memblock.c  				*out_end = m_end;
m_end            1172 mm/memblock.c  			if (m_end > r_start) {
m_end            1176 mm/memblock.c  					*out_end = min(m_end, r_end);