m_end 1008 arch/sparc/mm/init_64.c u64 ret_end, pa_start, m_mask, m_match, m_end; m_end 1073 arch/sparc/mm/init_64.c m_end = m_match + (1ul << __ffs(m_mask)) - mblock->offset; m_end 1074 arch/sparc/mm/init_64.c m_end += pa_start & ~((1ul << fls64(m_mask)) - 1); m_end 1075 arch/sparc/mm/init_64.c ret_end = m_end > end ? end : m_end; m_end 1105 arch/x86/kernel/cpu/amd.c #define AMD_MODEL_RANGE(f, m_start, s_start, m_end, s_end) \ m_end 1106 arch/x86/kernel/cpu/amd.c ((f << 24) | (m_start << 16) | (s_start << 12) | (m_end << 4) | (s_end)) m_end 219 drivers/firmware/efi/memmap.c u64 m_start, m_end; m_end 228 drivers/firmware/efi/memmap.c m_end = range->end; m_end 232 drivers/firmware/efi/memmap.c if (start < m_end && m_end < end) m_end 238 drivers/firmware/efi/memmap.c if (m_end < end) m_end 241 drivers/firmware/efi/memmap.c if (end <= m_end) m_end 260 drivers/firmware/efi/memmap.c u64 m_start, m_end, m_attr; m_end 267 drivers/firmware/efi/memmap.c m_end = mem->range.end; m_end 276 drivers/firmware/efi/memmap.c !IS_ALIGNED(m_end + 1, EFI_PAGE_SIZE)) { m_end 291 drivers/firmware/efi/memmap.c if (m_start <= start && end <= m_end) m_end 295 drivers/firmware/efi/memmap.c (start < m_end && m_end < end)) { m_end 298 drivers/firmware/efi/memmap.c md->num_pages = (m_end - md->phys_addr + 1) >> m_end 304 drivers/firmware/efi/memmap.c md->phys_addr = m_end + 1; m_end 309 drivers/firmware/efi/memmap.c if ((start < m_start && m_start < end) && m_end < end) { m_end 319 drivers/firmware/efi/memmap.c md->num_pages = (m_end - m_start + 1) >> m_end 325 drivers/firmware/efi/memmap.c md->phys_addr = m_end + 1; m_end 326 drivers/firmware/efi/memmap.c md->num_pages = (end - m_end) >> m_end 331 drivers/firmware/efi/memmap.c (end <= m_end)) { m_end 464 fs/ntfs/mft.c unsigned int block_start, block_end, m_start, m_end, page_ofs; m_end 515 fs/ntfs/mft.c m_end = m_start + vol->mft_record_size; m_end 521 fs/ntfs/mft.c if (unlikely(block_start >= m_end)) m_end 571 fs/ntfs/mft.c BUG_ON((nr_bhs >= max_bhs) && (m_end != block_end)); m_end 670 fs/ntfs/mft.c unsigned int block_start, block_end, m_start, m_end; m_end 695 fs/ntfs/mft.c m_end = m_start + vol->mft_record_size; m_end 701 fs/ntfs/mft.c if (unlikely(block_start >= m_end)) m_end 758 fs/ntfs/mft.c BUG_ON((nr_bhs >= max_bhs) && (m_end != block_end)); m_end 1032 mm/memblock.c phys_addr_t m_end = m->base + m->size; m_end 1042 mm/memblock.c *out_end = m_end; m_end 1065 mm/memblock.c if (r_start >= m_end) m_end 1073 mm/memblock.c *out_end = min(m_end, r_end); m_end 1080 mm/memblock.c if (m_end <= r_end) m_end 1136 mm/memblock.c phys_addr_t m_end = m->base + m->size; m_end 1146 mm/memblock.c *out_end = m_end; m_end 1172 mm/memblock.c if (m_end > r_start) { m_end 1176 mm/memblock.c *out_end = min(m_end, r_end);