part 158 arch/arm/include/asm/hardware/cache-l2x0.h void l2x0_pmu_register(void __iomem *base, u32 part); part 162 arch/arm/include/asm/hardware/cache-l2x0.h static inline void l2x0_pmu_register(void __iomem *base, u32 part) {} part 57 arch/arm/mach-omap2/cm.h int (*wait_module_ready)(u8 part, s16 prcm_mod, u16 idlest_reg, part 59 arch/arm/mach-omap2/cm.h int (*wait_module_idle)(u8 part, s16 prcm_mod, u16 idlest_reg, part 61 arch/arm/mach-omap2/cm.h void (*module_enable)(u8 mode, u8 part, u16 inst, u16 clkctrl_offs); part 62 arch/arm/mach-omap2/cm.h void (*module_disable)(u8 part, u16 inst, u16 clkctrl_offs); part 63 arch/arm/mach-omap2/cm.h u32 (*xlate_clkctrl)(u8 part, u16 inst, u16 clkctrl_offs); part 68 arch/arm/mach-omap2/cm.h int omap_cm_wait_module_ready(u8 part, s16 prcm_mod, u16 idlest_reg, part 70 arch/arm/mach-omap2/cm.h int omap_cm_wait_module_idle(u8 part, s16 prcm_mod, u16 idlest_reg, part 72 arch/arm/mach-omap2/cm.h int omap_cm_module_enable(u8 mode, u8 part, u16 inst, u16 clkctrl_offs); part 73 arch/arm/mach-omap2/cm.h int omap_cm_module_disable(u8 part, u16 inst, u16 clkctrl_offs); part 74 arch/arm/mach-omap2/cm.h u32 omap_cm_xlate_clkctrl(u8 part, u16 inst, u16 clkctrl_offs); part 245 arch/arm/mach-omap2/cm2xxx.c int omap2xxx_cm_wait_module_ready(u8 part, s16 prcm_mod, u16 idlest_id, part 54 arch/arm/mach-omap2/cm2xxx.h int omap2xxx_cm_wait_module_ready(u8 part, s16 prcm_mod, u16 idlest_id, part 229 arch/arm/mach-omap2/cm33xx.c static int am33xx_cm_wait_module_ready(u8 part, s16 inst, u16 clkctrl_offs, part 252 arch/arm/mach-omap2/cm33xx.c static int am33xx_cm_wait_module_idle(u8 part, s16 inst, u16 clkctrl_offs, part 273 arch/arm/mach-omap2/cm33xx.c static void am33xx_cm_module_enable(u8 mode, u8 part, u16 inst, part 292 arch/arm/mach-omap2/cm33xx.c static void am33xx_cm_module_disable(u8 part, u16 inst, u16 clkctrl_offs) part 347 arch/arm/mach-omap2/cm33xx.c static u32 am33xx_cm_xlate_clkctrl(u8 part, u16 inst, u16 offset) part 88 arch/arm/mach-omap2/cm3xxx.c static int omap3xxx_cm_wait_module_ready(u8 part, s16 prcm_mod, u16 idlest_id, part 94 arch/arm/mach-omap2/cm_common.c int omap_cm_wait_module_ready(u8 part, s16 prcm_mod, u16 idlest_reg, part 103 arch/arm/mach-omap2/cm_common.c return cm_ll_data->wait_module_ready(part, prcm_mod, idlest_reg, part 120 arch/arm/mach-omap2/cm_common.c int omap_cm_wait_module_idle(u8 part, s16 prcm_mod, u16 idlest_reg, part 129 arch/arm/mach-omap2/cm_common.c return cm_ll_data->wait_module_idle(part, prcm_mod, idlest_reg, part 144 arch/arm/mach-omap2/cm_common.c int omap_cm_module_enable(u8 mode, u8 part, u16 inst, u16 clkctrl_offs) part 152 arch/arm/mach-omap2/cm_common.c cm_ll_data->module_enable(mode, part, inst, clkctrl_offs); part 166 arch/arm/mach-omap2/cm_common.c int omap_cm_module_disable(u8 part, u16 inst, u16 clkctrl_offs) part 174 arch/arm/mach-omap2/cm_common.c cm_ll_data->module_disable(part, inst, clkctrl_offs); part 178 arch/arm/mach-omap2/cm_common.c u32 omap_cm_xlate_clkctrl(u8 part, u16 inst, u16 clkctrl_offs) part 185 arch/arm/mach-omap2/cm_common.c return cm_ll_data->xlate_clkctrl(part, inst, clkctrl_offs); part 74 arch/arm/mach-omap2/cminst44xx.c static u32 omap4_cminst_read_inst_reg(u8 part, u16 inst, u16 idx); part 85 arch/arm/mach-omap2/cminst44xx.c static u32 _clkctrl_idlest(u8 part, u16 inst, u16 clkctrl_offs) part 87 arch/arm/mach-omap2/cminst44xx.c u32 v = omap4_cminst_read_inst_reg(part, inst, clkctrl_offs); part 102 arch/arm/mach-omap2/cminst44xx.c static bool _is_module_ready(u8 part, u16 inst, u16 clkctrl_offs) part 106 arch/arm/mach-omap2/cminst44xx.c v = _clkctrl_idlest(part, inst, clkctrl_offs); part 113 arch/arm/mach-omap2/cminst44xx.c static u32 omap4_cminst_read_inst_reg(u8 part, u16 inst, u16 idx) part 115 arch/arm/mach-omap2/cminst44xx.c BUG_ON(part >= OMAP4_MAX_PRCM_PARTITIONS || part 116 arch/arm/mach-omap2/cminst44xx.c part == OMAP4430_INVALID_PRCM_PARTITION || part 117 arch/arm/mach-omap2/cminst44xx.c !_cm_bases[part].va); part 118 arch/arm/mach-omap2/cminst44xx.c return readl_relaxed(_cm_bases[part].va + inst + idx); part 122 arch/arm/mach-omap2/cminst44xx.c static void omap4_cminst_write_inst_reg(u32 val, u8 part, u16 inst, u16 idx) part 124 arch/arm/mach-omap2/cminst44xx.c BUG_ON(part >= OMAP4_MAX_PRCM_PARTITIONS || part 125 arch/arm/mach-omap2/cminst44xx.c part == OMAP4430_INVALID_PRCM_PARTITION || part 126 arch/arm/mach-omap2/cminst44xx.c !_cm_bases[part].va); part 127 arch/arm/mach-omap2/cminst44xx.c writel_relaxed(val, _cm_bases[part].va + inst + idx); part 131 arch/arm/mach-omap2/cminst44xx.c static u32 omap4_cminst_rmw_inst_reg_bits(u32 mask, u32 bits, u8 part, u16 inst, part 136 arch/arm/mach-omap2/cminst44xx.c v = omap4_cminst_read_inst_reg(part, inst, idx); part 139 arch/arm/mach-omap2/cminst44xx.c omap4_cminst_write_inst_reg(v, part, inst, idx); part 144 arch/arm/mach-omap2/cminst44xx.c static u32 omap4_cminst_set_inst_reg_bits(u32 bits, u8 part, u16 inst, s16 idx) part 146 arch/arm/mach-omap2/cminst44xx.c return omap4_cminst_rmw_inst_reg_bits(bits, bits, part, inst, idx); part 149 arch/arm/mach-omap2/cminst44xx.c static u32 omap4_cminst_clear_inst_reg_bits(u32 bits, u8 part, u16 inst, part 152 arch/arm/mach-omap2/cminst44xx.c return omap4_cminst_rmw_inst_reg_bits(bits, 0x0, part, inst, idx); part 155 arch/arm/mach-omap2/cminst44xx.c static u32 omap4_cminst_read_inst_reg_bits(u8 part, u16 inst, s16 idx, u32 mask) part 159 arch/arm/mach-omap2/cminst44xx.c v = omap4_cminst_read_inst_reg(part, inst, idx); part 180 arch/arm/mach-omap2/cminst44xx.c static void _clktrctrl_write(u8 c, u8 part, u16 inst, u16 cdoffs) part 184 arch/arm/mach-omap2/cminst44xx.c v = omap4_cminst_read_inst_reg(part, inst, cdoffs + OMAP4_CM_CLKSTCTRL); part 187 arch/arm/mach-omap2/cminst44xx.c omap4_cminst_write_inst_reg(v, part, inst, cdoffs + OMAP4_CM_CLKSTCTRL); part 199 arch/arm/mach-omap2/cminst44xx.c static bool omap4_cminst_is_clkdm_in_hwsup(u8 part, u16 inst, u16 cdoffs) part 203 arch/arm/mach-omap2/cminst44xx.c v = omap4_cminst_read_inst_reg(part, inst, cdoffs + OMAP4_CM_CLKSTCTRL); part 219 arch/arm/mach-omap2/cminst44xx.c static void omap4_cminst_clkdm_enable_hwsup(u8 part, u16 inst, u16 cdoffs) part 221 arch/arm/mach-omap2/cminst44xx.c _clktrctrl_write(OMAP34XX_CLKSTCTRL_ENABLE_AUTO, part, inst, cdoffs); part 234 arch/arm/mach-omap2/cminst44xx.c static void omap4_cminst_clkdm_disable_hwsup(u8 part, u16 inst, u16 cdoffs) part 236 arch/arm/mach-omap2/cminst44xx.c _clktrctrl_write(OMAP34XX_CLKSTCTRL_DISABLE_AUTO, part, inst, cdoffs); part 248 arch/arm/mach-omap2/cminst44xx.c static void omap4_cminst_clkdm_force_wakeup(u8 part, u16 inst, u16 cdoffs) part 250 arch/arm/mach-omap2/cminst44xx.c _clktrctrl_write(OMAP34XX_CLKSTCTRL_FORCE_WAKEUP, part, inst, cdoffs); part 257 arch/arm/mach-omap2/cminst44xx.c static void omap4_cminst_clkdm_force_sleep(u8 part, u16 inst, u16 cdoffs) part 259 arch/arm/mach-omap2/cminst44xx.c _clktrctrl_write(OMAP34XX_CLKSTCTRL_FORCE_SLEEP, part, inst, cdoffs); part 274 arch/arm/mach-omap2/cminst44xx.c static int omap4_cminst_wait_module_ready(u8 part, s16 inst, u16 clkctrl_offs, part 279 arch/arm/mach-omap2/cminst44xx.c omap_test_timeout(_is_module_ready(part, inst, clkctrl_offs), part 297 arch/arm/mach-omap2/cminst44xx.c static int omap4_cminst_wait_module_idle(u8 part, s16 inst, u16 clkctrl_offs, part 302 arch/arm/mach-omap2/cminst44xx.c omap_test_timeout((_clkctrl_idlest(part, inst, clkctrl_offs) == part 318 arch/arm/mach-omap2/cminst44xx.c static void omap4_cminst_module_enable(u8 mode, u8 part, u16 inst, part 323 arch/arm/mach-omap2/cminst44xx.c v = omap4_cminst_read_inst_reg(part, inst, clkctrl_offs); part 326 arch/arm/mach-omap2/cminst44xx.c omap4_cminst_write_inst_reg(v, part, inst, clkctrl_offs); part 337 arch/arm/mach-omap2/cminst44xx.c static void omap4_cminst_module_disable(u8 part, u16 inst, u16 clkctrl_offs) part 341 arch/arm/mach-omap2/cminst44xx.c v = omap4_cminst_read_inst_reg(part, inst, clkctrl_offs); part 343 arch/arm/mach-omap2/cminst44xx.c omap4_cminst_write_inst_reg(v, part, inst, clkctrl_offs); part 476 arch/arm/mach-omap2/cminst44xx.c static u32 omap4_cminst_xlate_clkctrl(u8 part, u16 inst, u16 offset) part 478 arch/arm/mach-omap2/cminst44xx.c return _cm_bases[part].pa + inst + offset; part 138 arch/arm/mach-omap2/prm.h bool (*was_any_context_lost_old)(u8 part, s16 inst, u16 idx); part 139 arch/arm/mach-omap2/prm.h void (*clear_context_loss_flags_old)(u8 part, s16 inst, u16 idx); part 141 arch/arm/mach-omap2/prm.h int (*assert_hardreset)(u8 shift, u8 part, s16 prm_mod, u16 offset); part 142 arch/arm/mach-omap2/prm.h int (*deassert_hardreset)(u8 shift, u8 st_shift, u8 part, s16 prm_mod, part 144 arch/arm/mach-omap2/prm.h int (*is_hardreset_asserted)(u8 shift, u8 part, s16 prm_mod, part 155 arch/arm/mach-omap2/prm.h int omap_prm_assert_hardreset(u8 shift, u8 part, s16 prm_mod, u16 offset); part 156 arch/arm/mach-omap2/prm.h int omap_prm_deassert_hardreset(u8 shift, u8 st_shift, u8 part, s16 prm_mod, part 158 arch/arm/mach-omap2/prm.h int omap_prm_is_hardreset_asserted(u8 shift, u8 part, s16 prm_mod, u16 offset); part 160 arch/arm/mach-omap2/prm.h extern bool prm_was_any_context_lost_old(u8 part, s16 inst, u16 idx); part 161 arch/arm/mach-omap2/prm.h extern void prm_clear_context_loss_flags_old(u8 part, s16 inst, u16 idx); part 33 arch/arm/mach-omap2/prm2xxx_3xxx.c int omap2_prm_is_hardreset_asserted(u8 shift, u8 part, s16 prm_mod, u16 offset) part 53 arch/arm/mach-omap2/prm2xxx_3xxx.c int omap2_prm_assert_hardreset(u8 shift, u8 part, s16 prm_mod, u16 offset) part 82 arch/arm/mach-omap2/prm2xxx_3xxx.c int omap2_prm_deassert_hardreset(u8 rst_shift, u8 st_shift, u8 part, part 100 arch/arm/mach-omap2/prm2xxx_3xxx.h int omap2_prm_is_hardreset_asserted(u8 shift, u8 part, s16 prm_mod, u16 offset); part 101 arch/arm/mach-omap2/prm2xxx_3xxx.h int omap2_prm_assert_hardreset(u8 shift, u8 part, s16 prm_mod, part 103 arch/arm/mach-omap2/prm2xxx_3xxx.h int omap2_prm_deassert_hardreset(u8 rst_shift, u8 st_shift, u8 part, part 67 arch/arm/mach-omap2/prm33xx.c static int am33xx_prm_is_hardreset_asserted(u8 shift, u8 part, s16 inst, part 93 arch/arm/mach-omap2/prm33xx.c static int am33xx_prm_assert_hardreset(u8 shift, u8 part, s16 inst, part 122 arch/arm/mach-omap2/prm33xx.c static int am33xx_prm_deassert_hardreset(u8 shift, u8 st_shift, u8 part, part 404 arch/arm/mach-omap2/prm44xx.c static bool omap44xx_prm_was_any_context_lost_old(u8 part, s16 inst, u16 idx) part 406 arch/arm/mach-omap2/prm44xx.c return (omap4_prminst_read_inst_reg(part, inst, idx)) ? 1 : 0; part 419 arch/arm/mach-omap2/prm44xx.c static void omap44xx_prm_clear_context_loss_flags_old(u8 part, s16 inst, part 422 arch/arm/mach-omap2/prm44xx.c omap4_prminst_write_inst_reg(0xffffffff, part, inst, idx); part 394 arch/arm/mach-omap2/prm_common.c bool prm_was_any_context_lost_old(u8 part, s16 inst, u16 idx) part 399 arch/arm/mach-omap2/prm_common.c ret = prm_ll_data->was_any_context_lost_old(part, inst, idx); part 418 arch/arm/mach-omap2/prm_common.c void prm_clear_context_loss_flags_old(u8 part, s16 inst, u16 idx) part 421 arch/arm/mach-omap2/prm_common.c prm_ll_data->clear_context_loss_flags_old(part, inst, idx); part 436 arch/arm/mach-omap2/prm_common.c int omap_prm_assert_hardreset(u8 shift, u8 part, s16 prm_mod, u16 offset) part 444 arch/arm/mach-omap2/prm_common.c return prm_ll_data->assert_hardreset(shift, part, prm_mod, offset); part 458 arch/arm/mach-omap2/prm_common.c int omap_prm_deassert_hardreset(u8 shift, u8 st_shift, u8 part, s16 prm_mod, part 467 arch/arm/mach-omap2/prm_common.c return prm_ll_data->deassert_hardreset(shift, st_shift, part, prm_mod, part 480 arch/arm/mach-omap2/prm_common.c int omap_prm_is_hardreset_asserted(u8 shift, u8 part, s16 prm_mod, u16 offset) part 488 arch/arm/mach-omap2/prm_common.c return prm_ll_data->is_hardreset_asserted(shift, part, prm_mod, offset); part 58 arch/arm/mach-omap2/prminst44xx.c u32 omap4_prminst_read_inst_reg(u8 part, s16 inst, u16 idx) part 60 arch/arm/mach-omap2/prminst44xx.c BUG_ON(part >= OMAP4_MAX_PRCM_PARTITIONS || part 61 arch/arm/mach-omap2/prminst44xx.c part == OMAP4430_INVALID_PRCM_PARTITION || part 62 arch/arm/mach-omap2/prminst44xx.c !_prm_bases[part].va); part 63 arch/arm/mach-omap2/prminst44xx.c return readl_relaxed(_prm_bases[part].va + inst + idx); part 67 arch/arm/mach-omap2/prminst44xx.c void omap4_prminst_write_inst_reg(u32 val, u8 part, s16 inst, u16 idx) part 69 arch/arm/mach-omap2/prminst44xx.c BUG_ON(part >= OMAP4_MAX_PRCM_PARTITIONS || part 70 arch/arm/mach-omap2/prminst44xx.c part == OMAP4430_INVALID_PRCM_PARTITION || part 71 arch/arm/mach-omap2/prminst44xx.c !_prm_bases[part].va); part 72 arch/arm/mach-omap2/prminst44xx.c writel_relaxed(val, _prm_bases[part].va + inst + idx); part 76 arch/arm/mach-omap2/prminst44xx.c u32 omap4_prminst_rmw_inst_reg_bits(u32 mask, u32 bits, u8 part, s16 inst, part 81 arch/arm/mach-omap2/prminst44xx.c v = omap4_prminst_read_inst_reg(part, inst, idx); part 84 arch/arm/mach-omap2/prminst44xx.c omap4_prminst_write_inst_reg(v, part, inst, idx); part 99 arch/arm/mach-omap2/prminst44xx.c int omap4_prminst_is_hardreset_asserted(u8 shift, u8 part, s16 inst, part 104 arch/arm/mach-omap2/prminst44xx.c v = omap4_prminst_read_inst_reg(part, inst, rstctrl_offs); part 123 arch/arm/mach-omap2/prminst44xx.c int omap4_prminst_assert_hardreset(u8 shift, u8 part, s16 inst, part 128 arch/arm/mach-omap2/prminst44xx.c omap4_prminst_rmw_inst_reg_bits(mask, mask, part, inst, rstctrl_offs); part 152 arch/arm/mach-omap2/prminst44xx.c int omap4_prminst_deassert_hardreset(u8 shift, u8 st_shift, u8 part, s16 inst, part 160 arch/arm/mach-omap2/prminst44xx.c if (omap4_prminst_is_hardreset_asserted(shift, part, inst, part 165 arch/arm/mach-omap2/prminst44xx.c omap4_prminst_rmw_inst_reg_bits(0xffffffff, st_mask, part, inst, part 168 arch/arm/mach-omap2/prminst44xx.c omap4_prminst_rmw_inst_reg_bits(mask, 0, part, inst, rstctrl_offs); part 170 arch/arm/mach-omap2/prminst44xx.c omap_test_timeout(omap4_prminst_is_hardreset_asserted(st_shift, part, part 20 arch/arm/mach-omap2/prminst44xx.h extern u32 omap4_prminst_read_inst_reg(u8 part, s16 inst, u16 idx); part 21 arch/arm/mach-omap2/prminst44xx.h extern void omap4_prminst_write_inst_reg(u32 val, u8 part, s16 inst, u16 idx); part 22 arch/arm/mach-omap2/prminst44xx.h extern u32 omap4_prminst_rmw_inst_reg_bits(u32 mask, u32 bits, u8 part, part 27 arch/arm/mach-omap2/prminst44xx.h extern int omap4_prminst_is_hardreset_asserted(u8 shift, u8 part, s16 inst, part 29 arch/arm/mach-omap2/prminst44xx.h extern int omap4_prminst_assert_hardreset(u8 shift, u8 part, s16 inst, part 31 arch/arm/mach-omap2/prminst44xx.h int omap4_prminst_deassert_hardreset(u8 shift, u8 st_shift, u8 part, part 474 arch/arm/mm/cache-l2x0-pmu.c void __init l2x0_pmu_register(void __iomem *base, u32 part) part 489 arch/arm/mm/cache-l2x0-pmu.c switch (part & L2X0_CACHE_ID_PART_MASK) { part 292 arch/ia64/include/asm/pal.h part : 8, /* 31-24 part of the part 305 arch/ia64/include/asm/pal.h part : 8, /* 31-24 part of the part 325 arch/ia64/include/asm/pal.h #define pclid_read_part pclid_info_read.part part 334 arch/ia64/include/asm/pal.h #define pclid_write_part pclid_info_write.part part 118 arch/mips/include/asm/xtalk/xwidget.h u32 part; part 581 arch/mips/net/ebpf_jit.c u64 part = (value >> (16 * (3 - i))) & 0xffff; part 583 arch/mips/net/ebpf_jit.c if (seen_part && needed_shift > 0 && (part || i == 3)) { part 587 arch/mips/net/ebpf_jit.c if (part) { part 588 arch/mips/net/ebpf_jit.c if (i == 0 || (!seen_part && i < 3 && part < 0x8000)) { part 589 arch/mips/net/ebpf_jit.c emit_instr(ctx, lui, dst, (s32)(s16)part); part 594 arch/mips/net/ebpf_jit.c (unsigned int)part); part 78 arch/powerpc/include/asm/nvram.h extern int __init nvram_init_os_partition(struct nvram_os_partition *part); part 84 arch/powerpc/include/asm/nvram.h extern int nvram_read_partition(struct nvram_os_partition *part, char *buff, part 89 arch/powerpc/include/asm/nvram.h extern int nvram_write_os_partition(struct nvram_os_partition *part, part 180 arch/powerpc/kernel/nvram_64.c int nvram_write_os_partition(struct nvram_os_partition *part, part 189 arch/powerpc/kernel/nvram_64.c if (part->index == -1) part 192 arch/powerpc/kernel/nvram_64.c if (length > part->size) part 193 arch/powerpc/kernel/nvram_64.c length = part->size; part 198 arch/powerpc/kernel/nvram_64.c tmp_index = part->index; part 219 arch/powerpc/kernel/nvram_64.c int nvram_read_partition(struct nvram_os_partition *part, char *buff, part 227 arch/powerpc/kernel/nvram_64.c if (part->index == -1) part 230 arch/powerpc/kernel/nvram_64.c if (length > part->size) part 231 arch/powerpc/kernel/nvram_64.c length = part->size; part 233 arch/powerpc/kernel/nvram_64.c tmp_index = part->index; part 235 arch/powerpc/kernel/nvram_64.c if (part->os_partition) { part 249 arch/powerpc/kernel/nvram_64.c if (part->os_partition) { part 274 arch/powerpc/kernel/nvram_64.c int __init nvram_init_os_partition(struct nvram_os_partition *part) part 280 arch/powerpc/kernel/nvram_64.c p = nvram_find_partition(part->name, NVRAM_SIG_OS, &size); part 283 arch/powerpc/kernel/nvram_64.c if (p && size < part->min_size) { part 285 arch/powerpc/kernel/nvram_64.c " removing it...\n", part->name); part 286 arch/powerpc/kernel/nvram_64.c nvram_remove_partition(part->name, NVRAM_SIG_OS, NULL); part 292 arch/powerpc/kernel/nvram_64.c p = nvram_create_partition(part->name, NVRAM_SIG_OS, part 293 arch/powerpc/kernel/nvram_64.c part->req_size, part->min_size); part 297 arch/powerpc/kernel/nvram_64.c part->name); part 300 arch/powerpc/kernel/nvram_64.c p = nvram_create_partition(part->name, NVRAM_SIG_OS, part 301 arch/powerpc/kernel/nvram_64.c part->req_size, part->min_size); part 307 arch/powerpc/kernel/nvram_64.c " partition, err %d\n", part->name, (int)p); part 311 arch/powerpc/kernel/nvram_64.c part->index = p; part 312 arch/powerpc/kernel/nvram_64.c part->size = nvram_get_partition_size(p) - sizeof(struct err_log_info); part 392 arch/powerpc/kernel/nvram_64.c if (record->part > 1 || (record->type != PSTORE_TYPE_DMESG)) part 412 arch/powerpc/kernel/nvram_64.c record->id = record->part; part 425 arch/powerpc/kernel/nvram_64.c struct nvram_os_partition *part = NULL; part 434 arch/powerpc/kernel/nvram_64.c part = &oops_log_partition; part 439 arch/powerpc/kernel/nvram_64.c part = &common_partition; part 447 arch/powerpc/kernel/nvram_64.c part = &rtas_log_partition; part 454 arch/powerpc/kernel/nvram_64.c part = &of_config_partition; part 464 arch/powerpc/kernel/nvram_64.c part = &skiboot_partition; part 475 arch/powerpc/kernel/nvram_64.c if (!part->os_partition) { part 476 arch/powerpc/kernel/nvram_64.c p = nvram_find_partition(part->name, sig, &size); part 479 arch/powerpc/kernel/nvram_64.c "err %d\n", part->name, (int)p); part 482 arch/powerpc/kernel/nvram_64.c part->index = p; part 483 arch/powerpc/kernel/nvram_64.c part->size = size; part 486 arch/powerpc/kernel/nvram_64.c buff = kmalloc(part->size, GFP_KERNEL); part 491 arch/powerpc/kernel/nvram_64.c if (nvram_read_partition(part, buff, part->size, &err_type, &id_no)) { part 498 arch/powerpc/kernel/nvram_64.c if (part->os_partition) part 531 arch/powerpc/kernel/nvram_64.c return part->size; part 724 arch/powerpc/kernel/nvram_64.c static int __init nvram_write_header(struct nvram_partition * part) part 730 arch/powerpc/kernel/nvram_64.c memcpy(&phead, &part->header, NVRAM_HEADER_LEN); part 733 arch/powerpc/kernel/nvram_64.c tmp_index = part->index; part 758 arch/powerpc/kernel/nvram_64.c static int nvram_can_remove_partition(struct nvram_partition *part, part 761 arch/powerpc/kernel/nvram_64.c if (part->header.signature != sig) part 764 arch/powerpc/kernel/nvram_64.c if (strncmp(name, part->header.name, 12)) part 769 arch/powerpc/kernel/nvram_64.c if (!strncmp(*except, part->header.name, 12)) part 788 arch/powerpc/kernel/nvram_64.c struct nvram_partition *part, *prev, *tmp; part 791 arch/powerpc/kernel/nvram_64.c list_for_each_entry(part, &nvram_partitions, partition) { part 792 arch/powerpc/kernel/nvram_64.c if (!nvram_can_remove_partition(part, name, sig, exceptions)) part 796 arch/powerpc/kernel/nvram_64.c part->header.signature = NVRAM_SIG_FREE; part 797 arch/powerpc/kernel/nvram_64.c memset(part->header.name, 'w', 12); part 798 arch/powerpc/kernel/nvram_64.c part->header.checksum = nvram_checksum(&part->header); part 799 arch/powerpc/kernel/nvram_64.c rc = nvram_write_header(part); part 808 arch/powerpc/kernel/nvram_64.c list_for_each_entry_safe(part, tmp, &nvram_partitions, partition) { part 809 arch/powerpc/kernel/nvram_64.c if (part->header.signature != NVRAM_SIG_FREE) { part 814 arch/powerpc/kernel/nvram_64.c prev->header.length += part->header.length; part 821 arch/powerpc/kernel/nvram_64.c list_del(&part->partition); part 822 arch/powerpc/kernel/nvram_64.c kfree(part); part 824 arch/powerpc/kernel/nvram_64.c prev = part; part 846 arch/powerpc/kernel/nvram_64.c struct nvram_partition *part; part 874 arch/powerpc/kernel/nvram_64.c list_for_each_entry(part, &nvram_partitions, partition) { part 875 arch/powerpc/kernel/nvram_64.c if (part->header.signature != NVRAM_SIG_FREE) part 878 arch/powerpc/kernel/nvram_64.c if (part->header.length >= req_size) { part 880 arch/powerpc/kernel/nvram_64.c free_part = part; part 883 arch/powerpc/kernel/nvram_64.c if (part->header.length > size && part 884 arch/powerpc/kernel/nvram_64.c part->header.length >= min_size) { part 885 arch/powerpc/kernel/nvram_64.c size = part->header.length; part 886 arch/powerpc/kernel/nvram_64.c free_part = part; part 952 arch/powerpc/kernel/nvram_64.c struct nvram_partition *part; part 954 arch/powerpc/kernel/nvram_64.c list_for_each_entry(part, &nvram_partitions, partition) { part 955 arch/powerpc/kernel/nvram_64.c if (part->index + NVRAM_HEADER_LEN == data_index) part 956 arch/powerpc/kernel/nvram_64.c return (part->header.length - 1) * NVRAM_BLOCK_LEN; part 127 arch/powerpc/kernel/ptrace32.c u32 part; part 135 arch/powerpc/kernel/ptrace32.c part = 1; /* want the 2nd half of the register (right-most). */ part 137 arch/powerpc/kernel/ptrace32.c part = 0; /* want the 1st half of the register (left-most). */ part 157 arch/powerpc/kernel/ptrace32.c reg32bits = ((u32*)&tmp)[part]; part 350 arch/powerpc/platforms/powermac/setup.c void note_bootable_part(dev_t dev, int part, int goodness); part 357 arch/powerpc/platforms/powermac/setup.c void __ref note_bootable_part(dev_t dev, int part, int goodness) part 370 arch/powerpc/platforms/powermac/setup.c ROOT_DEV = dev + part; part 95 arch/x86/platform/uv/bios_uv.c union partition_info_u part; part 102 arch/x86/platform/uv/bios_uv.c part.val = v0; part 104 arch/x86/platform/uv/bios_uv.c *uvtype = part.hub_version; part 106 arch/x86/platform/uv/bios_uv.c *partid = part.partition_id; part 108 arch/x86/platform/uv/bios_uv.c *coher = part.coherence_id; part 110 arch/x86/platform/uv/bios_uv.c *region = part.region_size; part 1755 block/bio.c void update_io_ticks(struct hd_struct *part, unsigned long now) part 1759 block/bio.c stamp = READ_ONCE(part->stamp); part 1761 block/bio.c if (likely(cmpxchg(&part->stamp, stamp, now) == stamp)) { part 1762 block/bio.c __part_stat_add(part, io_ticks, 1); part 1765 block/bio.c if (part->partno) { part 1766 block/bio.c part = &part_to_disk(part)->part0; part 1772 block/bio.c unsigned long sectors, struct hd_struct *part) part 1778 block/bio.c update_io_ticks(part, jiffies); part 1779 block/bio.c part_stat_inc(part, ios[sgrp]); part 1780 block/bio.c part_stat_add(part, sectors[sgrp], sectors); part 1781 block/bio.c part_inc_in_flight(q, part, op_is_write(op)); part 1788 block/bio.c struct hd_struct *part, unsigned long start_time) part 1796 block/bio.c update_io_ticks(part, now); part 1797 block/bio.c part_stat_add(part, nsecs[sgrp], jiffies_to_nsecs(duration)); part 1798 block/bio.c part_stat_add(part, time_in_queue, duration); part 1799 block/bio.c part_dec_in_flight(q, part, op_is_write(req_op)); part 772 block/blk-cgroup.c int key_len, part; part 782 block/blk-cgroup.c disk = get_gendisk(MKDEV(major, minor), &part); part 785 block/blk-cgroup.c if (part) { part 121 block/blk-core.c rq->part = NULL; part 762 block/blk-core.c static bool should_fail_request(struct hd_struct *part, unsigned int bytes) part 764 block/blk-core.c return part->make_it_fail && should_fail(&fail_make_request, bytes); part 779 block/blk-core.c static inline bool should_fail_request(struct hd_struct *part, part 787 block/blk-core.c static inline bool bio_check_ro(struct bio *bio, struct hd_struct *part) part 791 block/blk-core.c if (part->policy && op_is_write(op)) { part 800 block/blk-core.c bio_devname(bio, b), part->partno); part 1313 block/blk-core.c struct hd_struct *part; part 1316 block/blk-core.c part = req->part; part 1317 block/blk-core.c part_stat_add(part, sectors[sgrp], bytes >> 9); part 1331 block/blk-core.c struct hd_struct *part; part 1334 block/blk-core.c part = req->part; part 1336 block/blk-core.c update_io_ticks(part, jiffies); part 1337 block/blk-core.c part_stat_inc(part, ios[sgrp]); part 1338 block/blk-core.c part_stat_add(part, nsecs[sgrp], now - req->start_time_ns); part 1339 block/blk-core.c part_stat_add(part, time_in_queue, nsecs_to_jiffies64(now - req->start_time_ns)); part 1340 block/blk-core.c part_dec_in_flight(req->q, part, rq_data_dir(req)); part 1342 block/blk-core.c hd_struct_put(part); part 1349 block/blk-core.c struct hd_struct *part; part 1358 block/blk-core.c part = rq->part; part 1359 block/blk-core.c part_stat_inc(part, merges[rw]); part 1361 block/blk-core.c part = disk_map_sector_rcu(rq->rq_disk, blk_rq_pos(rq)); part 1362 block/blk-core.c if (!hd_struct_try_get(part)) { part 1371 block/blk-core.c part = &rq->rq_disk->part0; part 1372 block/blk-core.c hd_struct_get(part); part 1374 block/blk-core.c part_inc_in_flight(rq->q, part, rw); part 1375 block/blk-core.c rq->part = part; part 1378 block/blk-core.c update_io_ticks(part, jiffies); part 682 block/blk-merge.c struct hd_struct *part; part 685 block/blk-merge.c part = req->part; part 687 block/blk-merge.c part_dec_in_flight(req->q, part, rq_data_dir(req)); part 689 block/blk-merge.c hd_struct_put(part); part 95 block/blk-mq.c struct hd_struct *part; part 108 block/blk-mq.c if (rq->part == mi->part) part 114 block/blk-mq.c unsigned int blk_mq_in_flight(struct request_queue *q, struct hd_struct *part) part 117 block/blk-mq.c struct mq_inflight mi = { .part = part, .inflight = inflight, }; part 131 block/blk-mq.c if (rq->part == mi->part) part 137 block/blk-mq.c void blk_mq_in_flight_rw(struct request_queue *q, struct hd_struct *part, part 140 block/blk-mq.c struct mq_inflight mi = { .part = part, .inflight = inflight, }; part 328 block/blk-mq.c rq->part = NULL; part 188 block/blk-mq.h unsigned int blk_mq_in_flight(struct request_queue *q, struct hd_struct *part); part 189 block/blk-mq.h void blk_mq_in_flight_rw(struct request_queue *q, struct hd_struct *part, part 49 block/genhd.c void part_inc_in_flight(struct request_queue *q, struct hd_struct *part, int rw) part 54 block/genhd.c part_stat_local_inc(part, in_flight[rw]); part 55 block/genhd.c if (part->partno) part 56 block/genhd.c part_stat_local_inc(&part_to_disk(part)->part0, in_flight[rw]); part 59 block/genhd.c void part_dec_in_flight(struct request_queue *q, struct hd_struct *part, int rw) part 64 block/genhd.c part_stat_local_dec(part, in_flight[rw]); part 65 block/genhd.c if (part->partno) part 66 block/genhd.c part_stat_local_dec(&part_to_disk(part)->part0, in_flight[rw]); part 69 block/genhd.c unsigned int part_in_flight(struct request_queue *q, struct hd_struct *part) part 75 block/genhd.c return blk_mq_in_flight(q, part); part 80 block/genhd.c inflight += part_stat_local_read_cpu(part, in_flight[0], cpu) + part 81 block/genhd.c part_stat_local_read_cpu(part, in_flight[1], cpu); part 89 block/genhd.c void part_in_flight_rw(struct request_queue *q, struct hd_struct *part, part 95 block/genhd.c blk_mq_in_flight_rw(q, part, inflight); part 102 block/genhd.c inflight[0] += part_stat_local_read_cpu(part, in_flight[0], cpu); part 103 block/genhd.c inflight[1] += part_stat_local_read_cpu(part, in_flight[1], cpu); part 117 block/genhd.c return rcu_dereference(ptbl->part[partno]); part 136 block/genhd.c struct hd_struct *part; part 139 block/genhd.c part = __disk_get_part(disk, partno); part 140 block/genhd.c if (part) part 141 block/genhd.c get_device(part_to_dev(part)); part 144 block/genhd.c return part; part 168 block/genhd.c piter->part = NULL; part 198 block/genhd.c disk_put_part(piter->part); part 199 block/genhd.c piter->part = NULL; part 220 block/genhd.c struct hd_struct *part; part 222 block/genhd.c part = rcu_dereference(ptbl->part[piter->idx]); part 223 block/genhd.c if (!part) part 225 block/genhd.c if (!part_nr_sects_read(part) && part 231 block/genhd.c get_device(part_to_dev(part)); part 232 block/genhd.c piter->part = part; part 239 block/genhd.c return piter->part; part 254 block/genhd.c disk_put_part(piter->part); part 255 block/genhd.c piter->part = NULL; part 259 block/genhd.c static inline int sector_in_part(struct hd_struct *part, sector_t sector) part 261 block/genhd.c return part->start_sect <= sector && part 262 block/genhd.c sector < part->start_sect + part_nr_sects_read(part); part 283 block/genhd.c struct hd_struct *part; part 288 block/genhd.c part = rcu_dereference(ptbl->last_lookup); part 289 block/genhd.c if (part && sector_in_part(part, sector)) part 290 block/genhd.c return part; part 293 block/genhd.c part = rcu_dereference(ptbl->part[i]); part 295 block/genhd.c if (part && sector_in_part(part, sector)) { part 296 block/genhd.c rcu_assign_pointer(ptbl->last_lookup, part); part 297 block/genhd.c return part; part 488 block/genhd.c int blk_alloc_devt(struct hd_struct *part, dev_t *devt) part 490 block/genhd.c struct gendisk *disk = part_to_disk(part); part 494 block/genhd.c if (part->partno < disk->minors) { part 495 block/genhd.c *devt = MKDEV(disk->major, disk->first_minor + part->partno); part 503 block/genhd.c idx = idr_alloc(&ext_devt_idr, part, 0, NR_EXT_DEVT, GFP_NOWAIT); part 602 block/genhd.c struct hd_struct *part; part 667 block/genhd.c while ((part = disk_part_iter_next(&piter))) part 668 block/genhd.c kobject_uevent(&part_to_dev(part)->kobj, KOBJ_ADD); part 776 block/genhd.c struct hd_struct *part; part 789 block/genhd.c while ((part = disk_part_iter_next(&piter))) { part 790 block/genhd.c invalidate_partition(disk, part->partno); part 791 block/genhd.c bdev_unhash_inode(part_devt(part)); part 792 block/genhd.c delete_partition(disk, part->partno); part 882 block/genhd.c struct hd_struct *part; part 885 block/genhd.c part = idr_find(&ext_devt_idr, blk_mangle_minor(MINOR(devt))); part 886 block/genhd.c if (part && get_disk_and_module(part_to_disk(part))) { part 887 block/genhd.c *partno = part->partno; part 888 block/genhd.c disk = part_to_disk(part); part 928 block/genhd.c struct hd_struct *part; part 931 block/genhd.c part = disk_get_part(disk, partno); part 932 block/genhd.c if (part) part 933 block/genhd.c bdev = bdget(part_devt(part)); part 934 block/genhd.c disk_put_part(part); part 954 block/genhd.c struct hd_struct *part; part 972 block/genhd.c while ((part = disk_part_iter_next(&piter))) { part 973 block/genhd.c bool is_part0 = part == &disk->part0; part 976 block/genhd.c bdevt_str(part_devt(part), devt_buf), part 977 block/genhd.c (unsigned long long)part_nr_sects_read(part) >> 1 part 978 block/genhd.c , disk_name(disk, part->partno, name_buf), part 979 block/genhd.c part->info ? part->info->uuid : ""); part 1055 block/genhd.c struct hd_struct *part; part 1067 block/genhd.c while ((part = disk_part_iter_next(&piter))) part 1069 block/genhd.c MAJOR(part_devt(part)), MINOR(part_devt(part)), part 1070 block/genhd.c (unsigned long long)part_nr_sects_read(part) >> 1, part 1071 block/genhd.c disk_name(sgp, part->partno, buf)); part 1309 block/genhd.c new_ptbl = kzalloc_node(struct_size(new_ptbl, part, target), GFP_KERNEL, part 1317 block/genhd.c rcu_assign_pointer(new_ptbl->part[i], old_ptbl->part[i]); part 1438 block/genhd.c struct hd_struct *part; part 1451 block/genhd.c part = disk_get_part(disk, partno); part 1452 block/genhd.c if (part) { part 1453 block/genhd.c devt = part_devt(part); part 1454 block/genhd.c disk_put_part(part); part 1457 block/genhd.c disk_put_part(part); part 1490 block/genhd.c rcu_assign_pointer(ptbl->part[0], &disk->part0); part 1580 block/genhd.c struct hd_struct *part; part 1588 block/genhd.c while ((part = disk_part_iter_next(&piter))) part 1589 block/genhd.c part->policy = flag; part 18 block/ioctl.c struct hd_struct *part, *lpart; part 58 block/ioctl.c while ((part = disk_part_iter_next(&piter))) { part 59 block/ioctl.c if (!(start + length <= part->start_sect || part 60 block/ioctl.c start >= part->start_sect + part->nr_sects)) { part 69 block/ioctl.c part = add_partition(disk, partno, start, length, part 72 block/ioctl.c return PTR_ERR_OR_ZERO(part); part 74 block/ioctl.c part = disk_get_part(disk, partno); part 75 block/ioctl.c if (!part) part 78 block/ioctl.c bdevp = bdget(part_devt(part)); part 79 block/ioctl.c disk_put_part(part); part 112 block/ioctl.c part = disk_get_part(disk, partno); part 113 block/ioctl.c if (!part) part 115 block/ioctl.c bdevp = bdget(part_devt(part)); part 117 block/ioctl.c disk_put_part(part); part 122 block/ioctl.c if (start != part->start_sect) { part 126 block/ioctl.c disk_put_part(part); part 141 block/ioctl.c disk_put_part(part); part 146 block/ioctl.c part_nr_sects_write(part, (sector_t)length); part 151 block/ioctl.c disk_put_part(part); part 233 block/partition-generic.c struct hd_struct *part = dev_to_part(dev); part 235 block/partition-generic.c add_uevent_var(env, "PARTN=%u", part->partno); part 236 block/partition-generic.c if (part->info && part->info->volname[0]) part 237 block/partition-generic.c add_uevent_var(env, "PARTNAME=%s", part->info->volname); part 250 block/partition-generic.c struct hd_struct *part = container_of(to_rcu_work(work), struct hd_struct, part 253 block/partition-generic.c part->start_sect = 0; part 254 block/partition-generic.c part->nr_sects = 0; part 255 block/partition-generic.c part_stat_set_all(part, 0); part 256 block/partition-generic.c put_device(part_to_dev(part)); part 261 block/partition-generic.c struct hd_struct *part = container_of(ref, struct hd_struct, ref); part 262 block/partition-generic.c INIT_RCU_WORK(&part->rcu_work, delete_partition_work_fn); part 263 block/partition-generic.c queue_rcu_work(system_wq, &part->rcu_work); part 274 block/partition-generic.c struct hd_struct *part; part 279 block/partition-generic.c part = rcu_dereference_protected(ptbl->part[partno], 1); part 280 block/partition-generic.c if (!part) part 283 block/partition-generic.c rcu_assign_pointer(ptbl->part[partno], NULL); part 285 block/partition-generic.c kobject_put(part->holder_dir); part 286 block/partition-generic.c device_del(part_to_dev(part)); part 294 block/partition-generic.c blk_invalidate_devt(part_devt(part)); part 295 block/partition-generic.c hd_struct_kill(part); part 326 block/partition-generic.c if (ptbl->part[partno]) part 402 block/partition-generic.c rcu_assign_pointer(ptbl->part[partno], p); part 445 block/partition-generic.c struct hd_struct *part; part 455 block/partition-generic.c while ((part = disk_part_iter_next(&piter))) part 456 block/partition-generic.c delete_partition(disk, part->partno); part 515 block/partition-generic.c struct hd_struct *part; part 625 block/partition-generic.c part = add_partition(disk, p, from, size, part 628 block/partition-generic.c if (IS_ERR(part)) { part 630 block/partition-generic.c disk->disk_name, p, -PTR_ERR(part)); part 635 block/partition-generic.c md_autodetect_dev(part_to_dev(part)->devt); part 66 block/partitions/acorn.c struct riscix_part part[8]; part 87 block/partitions/acorn.c int part; part 92 block/partitions/acorn.c for (part = 0; part < 8; part++) { part 93 block/partitions/acorn.c if (rr->part[part].one && part 94 block/partitions/acorn.c memcmp(rr->part[part].name, "All\0", 4)) { part 96 block/partitions/acorn.c le32_to_cpu(rr->part[part].start), part 97 block/partitions/acorn.c le32_to_cpu(rr->part[part].length)); part 99 block/partitions/acorn.c strlcat(state->pp_buf, rr->part[part].name, PAGE_SIZE); part 35 block/partitions/amiga.c int start_sect, nr_sects, blk, part, res = 0; part 84 block/partitions/amiga.c for (part = 1; blk>0 && part<=16; part++, put_dev_sector(sect)) { part 59 block/partitions/atari.c if (!VALID_PARTITION(&rs->part[0], hd_size) && part 60 block/partitions/atari.c !VALID_PARTITION(&rs->part[1], hd_size) && part 61 block/partitions/atari.c !VALID_PARTITION(&rs->part[2], hd_size) && part 62 block/partitions/atari.c !VALID_PARTITION(&rs->part[3], hd_size)) { part 72 block/partitions/atari.c pi = &rs->part[0]; part 74 block/partitions/atari.c for (slot = 1; pi < &rs->part[4] && slot < state->limit; slot++, pi++) { part 103 block/partitions/atari.c if (!(xrs->part[0].flg & 1)) { part 110 block/partitions/atari.c partsect + be32_to_cpu(xrs->part[0].st), part 111 block/partitions/atari.c be32_to_cpu(xrs->part[0].siz)); part 113 block/partitions/atari.c if (!(xrs->part[1].flg & 1)) { part 118 block/partitions/atari.c if (memcmp( xrs->part[1].id, "XGM", 3 ) != 0) { part 124 block/partitions/atari.c partsect = be32_to_cpu(xrs->part[1].st) + extensect; part 31 block/partitions/atari.h struct partition_info part[4]; part 142 block/partitions/efi.c static inline int pmbr_part_valid(gpt_mbr_record *part) part 144 block/partitions/efi.c if (part->os_type != EFI_PMBR_OSTYPE_EFI_GPT) part 148 block/partitions/efi.c if (le32_to_cpu(part->starting_lba) != GPT_PRIMARY_PARTITION_TABLE_LBA) part 177 block/partitions/efi.c int i, part = 0, ret = 0; /* invalid by default */ part 185 block/partitions/efi.c part = i; part 217 block/partitions/efi.c sz = le32_to_cpu(mbr->partition_record[part].size_in_lba); part 576 block/partitions/ldm.c struct vblk_part *part; part 592 block/partitions/ldm.c part = &vb->vblk.part; part 594 block/partitions/ldm.c if (part->disk_id != disk->obj_id) part 598 block/partitions/ldm.c part->start, part->size); part 931 block/partitions/ldm.c struct vblk_part *part; part 980 block/partitions/ldm.c part = &vb->vblk.part; part 981 block/partitions/ldm.c part->start = get_unaligned_be64(buffer + 0x24 + r_name); part 982 block/partitions/ldm.c part->volume_offset = get_unaligned_be64(buffer + 0x2C + r_name); part 983 block/partitions/ldm.c part->size = ldm_get_vnum(buffer + 0x34 + r_name); part 984 block/partitions/ldm.c part->parent_id = ldm_get_vnum(buffer + 0x34 + r_size); part 985 block/partitions/ldm.c part->disk_id = ldm_get_vnum(buffer + 0x34 + r_parent); part 987 block/partitions/ldm.c part->partnum = buffer[0x35 + r_diskid]; part 989 block/partitions/ldm.c part->partnum = 0; part 1203 block/partitions/ldm.c if ((v->vblk.part.disk_id == vb->vblk.part.disk_id) && part 1204 block/partitions/ldm.c (v->vblk.part.start > vb->vblk.part.start)) { part 179 block/partitions/ldm.h struct vblk_part part; part 16 block/partitions/mac.c extern void note_bootable_part(dev_t dev, int part, int goodness); part 41 block/partitions/mac.c struct mac_partition *part; part 59 block/partitions/mac.c if (partoffset + sizeof(*part) > datasize) part 61 block/partitions/mac.c part = (struct mac_partition *) (data + partoffset); part 62 block/partitions/mac.c if (be16_to_cpu(part->signature) != MAC_PARTITION_MAGIC) { part 66 block/partitions/mac.c blocks_in_map = be32_to_cpu(part->map_count); part 82 block/partitions/mac.c part = (struct mac_partition *) (data + pos%512); part 83 block/partitions/mac.c if (be16_to_cpu(part->signature) != MAC_PARTITION_MAGIC) part 86 block/partitions/mac.c be32_to_cpu(part->start_block) * (secsize/512), part 87 block/partitions/mac.c be32_to_cpu(part->block_count) * (secsize/512)); part 89 block/partitions/mac.c if (!strncasecmp(part->type, "Linux_RAID", 10)) part 99 block/partitions/mac.c mac_fix_string(part->processor, 16); part 100 block/partitions/mac.c mac_fix_string(part->name, 32); part 101 block/partitions/mac.c mac_fix_string(part->type, 32); part 103 block/partitions/mac.c if ((be32_to_cpu(part->status) & MAC_STATUS_BOOTABLE) part 104 block/partitions/mac.c && strcasecmp(part->processor, "powerpc") == 0) part 107 block/partitions/mac.c if (strcasecmp(part->type, "Apple_UNIX_SVR2") == 0 part 108 block/partitions/mac.c || (strncasecmp(part->type, "Linux", 5) == 0 part 109 block/partitions/mac.c && strcasecmp(part->type, "Linux_swap") != 0)) { part 113 block/partitions/mac.c l = strlen(part->name); part 114 block/partitions/mac.c if (strcmp(part->name, "/") == 0) part 117 block/partitions/mac.c if (strncasecmp(part->name + i, "root", part 123 block/partitions/mac.c if (strncasecmp(part->name, "swap", 4) == 0) part 23 drivers/acpi/acpica/utmath.c struct uint64_struct part; part 62 drivers/acpi/acpica/utmath.c ACPI_MUL_64_BY_32(0, multiplicand_ovl.part.hi, multiplier, part 63 drivers/acpi/acpica/utmath.c product.part.hi, carry32); part 65 drivers/acpi/acpica/utmath.c ACPI_MUL_64_BY_32(0, multiplicand_ovl.part.lo, multiplier, part 66 drivers/acpi/acpica/utmath.c product.part.lo, carry32); part 68 drivers/acpi/acpica/utmath.c product.part.hi += carry32; part 100 drivers/acpi/acpica/utmath.c operand_ovl.part.hi = operand_ovl.part.lo; part 101 drivers/acpi/acpica/utmath.c operand_ovl.part.lo = 0; part 104 drivers/acpi/acpica/utmath.c ACPI_SHIFT_LEFT_64_BY_32(operand_ovl.part.hi, part 105 drivers/acpi/acpica/utmath.c operand_ovl.part.lo, count); part 137 drivers/acpi/acpica/utmath.c operand_ovl.part.lo = operand_ovl.part.hi; part 138 drivers/acpi/acpica/utmath.c operand_ovl.part.hi = 0; part 141 drivers/acpi/acpica/utmath.c ACPI_SHIFT_RIGHT_64_BY_32(operand_ovl.part.hi, part 142 drivers/acpi/acpica/utmath.c operand_ovl.part.lo, count); part 278 drivers/acpi/acpica/utmath.c ACPI_DIV_64_BY_32(0, dividend_ovl.part.hi, divisor, part 279 drivers/acpi/acpica/utmath.c quotient.part.hi, remainder32); part 281 drivers/acpi/acpica/utmath.c ACPI_DIV_64_BY_32(remainder32, dividend_ovl.part.lo, divisor, part 282 drivers/acpi/acpica/utmath.c quotient.part.lo, remainder32); part 336 drivers/acpi/acpica/utmath.c if (divisor.part.hi == 0) { part 341 drivers/acpi/acpica/utmath.c remainder.part.hi = 0; part 347 drivers/acpi/acpica/utmath.c ACPI_DIV_64_BY_32(0, dividend.part.hi, divisor.part.lo, part 348 drivers/acpi/acpica/utmath.c quotient.part.hi, partial1); part 350 drivers/acpi/acpica/utmath.c ACPI_DIV_64_BY_32(partial1, dividend.part.lo, divisor.part.lo, part 351 drivers/acpi/acpica/utmath.c quotient.part.lo, remainder.part.lo); part 359 drivers/acpi/acpica/utmath.c quotient.part.hi = 0; part 366 drivers/acpi/acpica/utmath.c ACPI_SHIFT_RIGHT_64(normalized_divisor.part.hi, part 367 drivers/acpi/acpica/utmath.c normalized_divisor.part.lo); part 368 drivers/acpi/acpica/utmath.c ACPI_SHIFT_RIGHT_64(normalized_dividend.part.hi, part 369 drivers/acpi/acpica/utmath.c normalized_dividend.part.lo); part 371 drivers/acpi/acpica/utmath.c } while (normalized_divisor.part.hi != 0); part 375 drivers/acpi/acpica/utmath.c ACPI_DIV_64_BY_32(normalized_dividend.part.hi, part 376 drivers/acpi/acpica/utmath.c normalized_dividend.part.lo, part 377 drivers/acpi/acpica/utmath.c normalized_divisor.part.lo, quotient.part.lo, part 384 drivers/acpi/acpica/utmath.c partial1 = quotient.part.lo * divisor.part.hi; part 385 drivers/acpi/acpica/utmath.c partial2.full = (u64) quotient.part.lo * divisor.part.lo; part 386 drivers/acpi/acpica/utmath.c partial3.full = (u64) partial2.part.hi + partial1; part 388 drivers/acpi/acpica/utmath.c remainder.part.hi = partial3.part.lo; part 389 drivers/acpi/acpica/utmath.c remainder.part.lo = partial2.part.lo; part 391 drivers/acpi/acpica/utmath.c if (partial3.part.hi == 0) { part 392 drivers/acpi/acpica/utmath.c if (partial3.part.lo >= dividend.part.hi) { part 393 drivers/acpi/acpica/utmath.c if (partial3.part.lo == dividend.part.hi) { part 394 drivers/acpi/acpica/utmath.c if (partial2.part.lo > dividend.part.lo) { part 395 drivers/acpi/acpica/utmath.c quotient.part.lo--; part 399 drivers/acpi/acpica/utmath.c quotient.part.lo--; part 405 drivers/acpi/acpica/utmath.c remainder.part.hi = (u32)-((s32)remainder.part.hi); part 406 drivers/acpi/acpica/utmath.c remainder.part.lo = (u32)-((s32)remainder.part.lo); part 408 drivers/acpi/acpica/utmath.c if (remainder.part.lo) { part 409 drivers/acpi/acpica/utmath.c remainder.part.hi--; part 1849 drivers/block/amiflop.c static struct kobject *floppy_find(dev_t dev, int *part, void *data) part 1851 drivers/block/amiflop.c int drive = *part & 3; part 1854 drivers/block/amiflop.c *part = 0; part 1964 drivers/block/ataflop.c static struct kobject *floppy_find(dev_t dev, int *part, void *data) part 1966 drivers/block/ataflop.c int drive = *part & 3; part 1967 drivers/block/ataflop.c int type = *part >> 2; part 1970 drivers/block/ataflop.c *part = 0; part 456 drivers/block/brd.c static struct kobject *brd_probe(dev_t dev, int *part, void *data) part 468 drivers/block/brd.c *part = 0; part 4527 drivers/block/floppy.c static struct kobject *floppy_find(dev_t dev, int *part, void *data) part 4529 drivers/block/floppy.c int drive = (*part & 3) | ((*part & 0x80) >> 5); part 4532 drivers/block/floppy.c if (((*part >> 2) & 0x1f) >= ARRAY_SIZE(floppy_type)) part 4534 drivers/block/floppy.c *part = 0; part 2154 drivers/block/loop.c static struct kobject *loop_probe(dev_t dev, int *part, void *data) part 2170 drivers/block/loop.c *part = 0; part 766 drivers/block/swim.c static struct kobject *floppy_find(dev_t dev, int *part, void *data) part 769 drivers/block/swim.c int drive = (*part & 3); part 774 drivers/block/swim.c *part = 0; part 333 drivers/block/z2ram.c static struct kobject *z2_find(dev_t dev, int *part, void *data) part 335 drivers/block/z2ram.c *part = 0; part 298 drivers/char/nvram.c int part, offset; part 300 drivers/char/nvram.c if (copy_from_user(&part, (void __user *)arg, part 301 drivers/char/nvram.c sizeof(part)) != 0) part 303 drivers/char/nvram.c if (part < pmac_nvram_OF || part > pmac_nvram_NR) part 305 drivers/char/nvram.c offset = pmac_get_partition(part); part 33 drivers/firmware/efi/efi-pstore.c static inline u64 generic_id(u64 timestamp, unsigned int part, int count) part 35 drivers/firmware/efi/efi-pstore.c return (timestamp * 100 + part) * 1000 + count; part 45 drivers/firmware/efi/efi-pstore.c unsigned int part; part 56 drivers/firmware/efi/efi-pstore.c &record->type, &part, &cnt, &time, &data_type) == 5) { part 57 drivers/firmware/efi/efi-pstore.c record->id = generic_id(time, part, cnt); part 58 drivers/firmware/efi/efi-pstore.c record->part = part; part 68 drivers/firmware/efi/efi-pstore.c &record->type, &part, &cnt, &time) == 4) { part 69 drivers/firmware/efi/efi-pstore.c record->id = generic_id(time, part, cnt); part 70 drivers/firmware/efi/efi-pstore.c record->part = part; part 77 drivers/firmware/efi/efi-pstore.c &record->type, &part, &time) == 3) { part 83 drivers/firmware/efi/efi-pstore.c record->id = generic_id(time, part, 0); part 84 drivers/firmware/efi/efi-pstore.c record->part = part; part 249 drivers/firmware/efi/efi-pstore.c record->id = generic_id(record->time.tv_sec, record->part, part 256 drivers/firmware/efi/efi-pstore.c record->type, record->part, record->count, part 333 drivers/firmware/efi/efi-pstore.c record->type, record->part, record->count, part 340 drivers/firmware/efi/efi-pstore.c record->type, record->part, (long long)record->time.tv_sec); part 300 drivers/gpu/drm/i915/display/intel_display_types.h int i, u32 *part); part 5770 drivers/gpu/drm/i915/display/intel_dp.c int i, u32 *part) part 5778 drivers/gpu/drm/i915/display/intel_dp.c DP_AUX_HDCP_V_PRIME(i), part, part 1415 drivers/gpu/drm/i915/display/intel_hdmi.c int i, u32 *part) part 1423 drivers/gpu/drm/i915/display/intel_hdmi.c part, DRM_HDCP_V_PRIME_PART_LEN); part 74 drivers/gpu/drm/nouveau/dispnv50/wndw.c args.nv50.part = NV50_DMA_V0_PART_256; part 78 drivers/gpu/drm/nouveau/dispnv50/wndw.c args.nv50.part = NV50_DMA_V0_PART_256; part 33 drivers/gpu/drm/nouveau/include/nvif/cl0002.h __u8 part; part 17 drivers/gpu/drm/nouveau/include/nvkm/subdev/mmu.h bool part:1; /* Region was split from an allocated region by map(). */ part 78 drivers/gpu/drm/nouveau/nvkm/engine/dma/usernv50.c u32 user, part, comp, kind; part 97 drivers/gpu/drm/nouveau/nvkm/engine/dma/usernv50.c args->v0.priv, args->v0.part, args->v0.comp, part 100 drivers/gpu/drm/nouveau/nvkm/engine/dma/usernv50.c part = args->v0.part; part 107 drivers/gpu/drm/nouveau/nvkm/engine/dma/usernv50.c part = NV50_DMA_V0_PART_256; part 112 drivers/gpu/drm/nouveau/nvkm/engine/dma/usernv50.c part = NV50_DMA_V0_PART_VM; part 119 drivers/gpu/drm/nouveau/nvkm/engine/dma/usernv50.c if (user > 2 || part > 2 || comp > 3 || kind > 0x7f) part 123 drivers/gpu/drm/nouveau/nvkm/engine/dma/usernv50.c dmaobj->flags5 = (part << 16); part 125 drivers/gpu/drm/nouveau/nvkm/subdev/devinit/gt215.c u32 part = (addr / mmio[2]) & 7; part 128 drivers/gpu/drm/nouveau/nvkm/subdev/devinit/gt215.c if (part >= hweight8((init->r001540 >> 16) & 0xff)) part 112 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgf100.c u32 part = nvkm_rd32(device, 0x022438), i; part 119 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgf100.c for (i = 0; (magic & 0x80000000) && i < part; addr += 0x1000, i++) { part 249 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c if (ret = -ENOENT, !vma || vma->addr != addr || vma->part) { part 251 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c vma ? vma->addr : ~0ULL, vma ? vma->part : 0); part 651 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c u64 part = ALIGN(addr, next) - addr; part 652 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (size - part >= next) part 653 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c block = (part >> page[i].shift) << page[i].shift; part 776 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c new->part = vma->part; part 934 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c vma->part = true; part 944 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c tmp->part = true; part 961 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c vma->part ? 'P' : '-', part 1157 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (vma->addr == addr && vma->part && (prev = node(vma, prev))) { part 1163 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (!next->part || part 1343 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (vma->part && (prev = node(vma, prev)) && prev->mapped) part 1345 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if ((next = node(vma, next)) && (!next->part || next->mapped)) part 1546 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c BUG_ON(vma->part); part 1557 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c while ((next = node(next, next)) && next->part && part 1576 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c } while (next && next->part); part 1587 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c } while ((next = node(vma, next)) && next->part); part 1786 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (inst && vmm && vmm->func->part) { part 1788 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c vmm->func->part(vmm, inst); part 141 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h void (*part)(struct nvkm_vmm *, struct nvkm_memory *inst); part 373 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c .part = gf100_vmm_part, part 388 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c .part = gf100_vmm_part, part 70 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgk104.c .part = gf100_vmm_part, part 85 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgk104.c .part = gf100_vmm_part, part 39 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgk20a.c .part = gf100_vmm_part, part 54 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgk20a.c .part = gf100_vmm_part, part 112 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c .part = gf100_vmm_part, part 128 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c .part = gf100_vmm_part, part 27 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm20b.c .part = gf100_vmm_part, part 43 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm20b.c .part = gf100_vmm_part, part 490 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c .part = gf100_vmm_part, part 27 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp10b.c .part = gf100_vmm_part, part 65 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgv100.c .part = gf100_vmm_part, part 27 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmmcp77.c .part = nv50_vmm_part, part 367 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c .part = nv50_vmm_part, part 55 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmtu102.c .part = gf100_vmm_part, part 225 drivers/i3c/device.c u16 part = I3C_PID_PART_ID(devinfo.pid); part 234 drivers/i3c/device.c if (manuf != id->manuf_id || part != id->part_id) part 258 drivers/i3c/master.c u16 manuf, part, ext; part 262 drivers/i3c/master.c part = I3C_PID_PART_ID(devinfo.pid); part 271 drivers/i3c/master.c devinfo.dcr, manuf, part, ext); part 911 drivers/ide/ide-probe.c static struct kobject *ata_probe(dev_t dev, int *part, void *data) part 914 drivers/ide/ide-probe.c int unit = *part >> PARTN_BITS; part 932 drivers/ide/ide-probe.c static struct kobject *exact_match(dev_t dev, int *part, void *data) part 935 drivers/ide/ide-probe.c *part &= (1 << PARTN_BITS) - 1; part 153 drivers/iio/light/si1145.c u8 part; part 1305 drivers/iio/light/si1145.c if (part_id != data->part_info->part) { part 1307 drivers/iio/light/si1145.c part_id, data->part_info->part); part 84 drivers/infiniband/sw/siw/siw_qp_tx.c unsigned long part = bytes - (PAGE_SIZE - off); part 86 drivers/infiniband/sw/siw/siw_qp_tx.c memcpy(paddr, buffer + off, part); part 91 drivers/infiniband/sw/siw/siw_qp_tx.c sge->laddr + part); part 94 drivers/infiniband/sw/siw/siw_qp_tx.c sge->laddr + part, part 100 drivers/infiniband/sw/siw/siw_qp_tx.c memcpy(paddr + part, buffer, bytes - part); part 1631 drivers/irqchip/irq-gic-v3.c struct partition_affinity *part; part 1634 drivers/irqchip/irq-gic-v3.c part = &parts[part_idx]; part 1636 drivers/irqchip/irq-gic-v3.c part->partition_id = of_node_to_fwnode(child_part); part 1665 drivers/irqchip/irq-gic-v3.c cpumask_set_cpu(cpu, &part->mask); part 25 drivers/irqchip/irq-partition-percpu.c static bool partition_check_cpu(struct partition_desc *part, part 28 drivers/irqchip/irq-partition-percpu.c return cpumask_test_cpu(cpu, &part->parts[hwirq].mask); part 33 drivers/irqchip/irq-partition-percpu.c struct partition_desc *part = irq_data_get_irq_chip_data(d); part 34 drivers/irqchip/irq-partition-percpu.c struct irq_chip *chip = irq_desc_get_chip(part->chained_desc); part 35 drivers/irqchip/irq-partition-percpu.c struct irq_data *data = irq_desc_get_irq_data(part->chained_desc); part 37 drivers/irqchip/irq-partition-percpu.c if (partition_check_cpu(part, smp_processor_id(), d->hwirq) && part 44 drivers/irqchip/irq-partition-percpu.c struct partition_desc *part = irq_data_get_irq_chip_data(d); part 45 drivers/irqchip/irq-partition-percpu.c struct irq_chip *chip = irq_desc_get_chip(part->chained_desc); part 46 drivers/irqchip/irq-partition-percpu.c struct irq_data *data = irq_desc_get_irq_data(part->chained_desc); part 48 drivers/irqchip/irq-partition-percpu.c if (partition_check_cpu(part, smp_processor_id(), d->hwirq) && part 57 drivers/irqchip/irq-partition-percpu.c struct partition_desc *part = irq_data_get_irq_chip_data(d); part 58 drivers/irqchip/irq-partition-percpu.c struct irq_chip *chip = irq_desc_get_chip(part->chained_desc); part 59 drivers/irqchip/irq-partition-percpu.c struct irq_data *data = irq_desc_get_irq_data(part->chained_desc); part 61 drivers/irqchip/irq-partition-percpu.c if (partition_check_cpu(part, smp_processor_id(), d->hwirq) && part 72 drivers/irqchip/irq-partition-percpu.c struct partition_desc *part = irq_data_get_irq_chip_data(d); part 73 drivers/irqchip/irq-partition-percpu.c struct irq_chip *chip = irq_desc_get_chip(part->chained_desc); part 74 drivers/irqchip/irq-partition-percpu.c struct irq_data *data = irq_desc_get_irq_data(part->chained_desc); part 76 drivers/irqchip/irq-partition-percpu.c if (partition_check_cpu(part, smp_processor_id(), d->hwirq) && part 85 drivers/irqchip/irq-partition-percpu.c struct partition_desc *part = irq_data_get_irq_chip_data(d); part 86 drivers/irqchip/irq-partition-percpu.c struct irq_chip *chip = irq_desc_get_chip(part->chained_desc); part 87 drivers/irqchip/irq-partition-percpu.c struct irq_data *data = irq_desc_get_irq_data(part->chained_desc); part 97 drivers/irqchip/irq-partition-percpu.c struct partition_desc *part = irq_data_get_irq_chip_data(d); part 98 drivers/irqchip/irq-partition-percpu.c struct irq_chip *chip = irq_desc_get_chip(part->chained_desc); part 99 drivers/irqchip/irq-partition-percpu.c struct irq_data *data = irq_desc_get_irq_data(part->chained_desc); part 115 drivers/irqchip/irq-partition-percpu.c struct partition_desc *part = irq_desc_get_handler_data(desc); part 122 drivers/irqchip/irq-partition-percpu.c for_each_set_bit(hwirq, part->bitmap, part->nr_parts) { part 123 drivers/irqchip/irq-partition-percpu.c if (partition_check_cpu(part, cpu, hwirq)) part 127 drivers/irqchip/irq-partition-percpu.c if (unlikely(hwirq == part->nr_parts)) { part 131 drivers/irqchip/irq-partition-percpu.c irq = irq_find_mapping(part->domain, hwirq); part 145 drivers/irqchip/irq-partition-percpu.c struct partition_desc *part; part 152 drivers/irqchip/irq-partition-percpu.c part = domain->host_data; part 154 drivers/irqchip/irq-partition-percpu.c set_bit(hwirq, part->bitmap); part 155 drivers/irqchip/irq-partition-percpu.c irq_set_chained_handler_and_data(irq_desc_get_irq(part->chained_desc), part 156 drivers/irqchip/irq-partition-percpu.c partition_handle_irq, part); part 157 drivers/irqchip/irq-partition-percpu.c irq_set_percpu_devid_partition(virq, &part->parts[hwirq].mask); part 158 drivers/irqchip/irq-partition-percpu.c irq_domain_set_info(domain, virq, hwirq, &partition_irq_chip, part, part 179 drivers/irqchip/irq-partition-percpu.c struct partition_affinity *part = NULL; part 184 drivers/irqchip/irq-partition-percpu.c part = &desc->parts[i]; part 189 drivers/irqchip/irq-partition-percpu.c if (WARN_ON(!part)) { part 1022 drivers/macintosh/smu.c const struct smu_sdbp_header *part; part 1039 drivers/macintosh/smu.c part = of_get_property(smu->of_node, pname, size); part 1040 drivers/macintosh/smu.c if (part == NULL) { part 1042 drivers/macintosh/smu.c part = smu_create_sdb_partition(id); part 1043 drivers/macintosh/smu.c if (part != NULL && size) part 1044 drivers/macintosh/smu.c *size = part->len << 2; part 1047 drivers/macintosh/smu.c return part; part 1130 drivers/macintosh/smu.c const struct smu_sdbp_header *part; part 1131 drivers/macintosh/smu.c part = __smu_get_sdb_partition(hdr.cmd, NULL, 1); part 1132 drivers/macintosh/smu.c if (part == NULL) part 1134 drivers/macintosh/smu.c else if (IS_ERR(part)) part 1135 drivers/macintosh/smu.c return PTR_ERR(part); part 632 drivers/md/dm.c struct hd_struct *part = &dm_disk(md)->part0; part 636 drivers/md/dm.c sum += part_stat_local_read_cpu(part, in_flight[0], cpu); part 637 drivers/md/dm.c sum += part_stat_local_read_cpu(part, in_flight[1], cpu); part 1601 drivers/md/dm.c #define __dm_part_stat_sub(part, field, subnd) \ part 1602 drivers/md/dm.c (part_stat_get(part, field) -= (subnd)) part 70 drivers/md/md.c static void autostart_arrays(int part); part 5542 drivers/md/md.c static struct kobject *md_probe(dev_t dev, int *part, void *data) part 6269 drivers/md/md.c static void autorun_devices(int part) part 6296 drivers/md/md.c if (part) { part 9556 drivers/md/md.c static void autostart_arrays(int part) part 9593 drivers/md/md.c autorun_devices(part); part 509 drivers/misc/bh1770glc.c u8 manu, part; part 519 drivers/misc/bh1770glc.c part = (u8)ret; part 521 drivers/misc/bh1770glc.c chip->revision = (part & BH1770_REV_MASK) >> BH1770_REV_SHIFT; part 527 drivers/misc/bh1770glc.c ((part & BH1770_PART_MASK) == BH1770_PART)) { part 533 drivers/misc/bh1770glc.c ((part & BH1770_PART_MASK) == BH1770_PART)) { part 664 drivers/misc/sgi-xp/xpc.h xpc_wakeup_channel_mgr(struct xpc_partition *part) part 666 drivers/misc/sgi-xp/xpc.h if (atomic_inc_return(&part->channel_mgr_requests) == 1) part 667 drivers/misc/sgi-xp/xpc.h wake_up(&part->channel_mgr_wq); part 698 drivers/misc/sgi-xp/xpc.h xpc_part_deref(struct xpc_partition *part) part 700 drivers/misc/sgi-xp/xpc.h s32 refs = atomic_dec_return(&part->references); part 703 drivers/misc/sgi-xp/xpc.h if (refs == 0 && part->setup_state == XPC_P_SS_WTEARDOWN) part 704 drivers/misc/sgi-xp/xpc.h wake_up(&part->teardown_wq); part 708 drivers/misc/sgi-xp/xpc.h xpc_part_ref(struct xpc_partition *part) part 712 drivers/misc/sgi-xp/xpc.h atomic_inc(&part->references); part 713 drivers/misc/sgi-xp/xpc.h setup = (part->setup_state == XPC_P_SS_SETUP); part 715 drivers/misc/sgi-xp/xpc.h xpc_part_deref(part); part 82 drivers/misc/sgi-xp/xpc_channel.c struct xpc_partition *part = &xpc_partitions[ch->partid]; part 101 drivers/misc/sgi-xp/xpc_channel.c if (part->act_state == XPC_P_AS_DEACTIVATING) { part 157 drivers/misc/sgi-xp/xpc_channel.c atomic_dec(&part->nchannels_active); part 168 drivers/misc/sgi-xp/xpc_channel.c if (part->act_state != XPC_P_AS_DEACTIVATING) { part 170 drivers/misc/sgi-xp/xpc_channel.c spin_lock(&part->chctl_lock); part 171 drivers/misc/sgi-xp/xpc_channel.c part->chctl.flags[ch->number] |= part 173 drivers/misc/sgi-xp/xpc_channel.c spin_unlock(&part->chctl_lock); part 183 drivers/misc/sgi-xp/xpc_channel.c xpc_process_openclose_chctl_flags(struct xpc_partition *part, int ch_number, part 188 drivers/misc/sgi-xp/xpc_channel.c &part->remote_openclose_args[ch_number]; part 189 drivers/misc/sgi-xp/xpc_channel.c struct xpc_channel *ch = &part->channels[ch_number]; part 238 drivers/misc/sgi-xp/xpc_channel.c if (part->chctl.flags[ch_number] & part 242 drivers/misc/sgi-xp/xpc_channel.c spin_lock(&part->chctl_lock); part 243 drivers/misc/sgi-xp/xpc_channel.c part->chctl.flags[ch_number] |= part 245 drivers/misc/sgi-xp/xpc_channel.c spin_unlock(&part->chctl_lock); part 253 drivers/misc/sgi-xp/xpc_channel.c atomic_inc(&part->nchannels_active); part 289 drivers/misc/sgi-xp/xpc_channel.c DBUG_ON(part->act_state != XPC_P_AS_DEACTIVATING); part 296 drivers/misc/sgi-xp/xpc_channel.c if (part->chctl.flags[ch_number] & part 300 drivers/misc/sgi-xp/xpc_channel.c spin_lock(&part->chctl_lock); part 301 drivers/misc/sgi-xp/xpc_channel.c part->chctl.flags[ch_number] |= part 303 drivers/misc/sgi-xp/xpc_channel.c spin_unlock(&part->chctl_lock); part 323 drivers/misc/sgi-xp/xpc_channel.c if (part->act_state == XPC_P_AS_DEACTIVATING || part 362 drivers/misc/sgi-xp/xpc_channel.c atomic_inc(&part->nchannels_active); part 546 drivers/misc/sgi-xp/xpc_channel.c xpc_process_sent_chctl_flags(struct xpc_partition *part) part 554 drivers/misc/sgi-xp/xpc_channel.c chctl.all_flags = xpc_arch_ops.get_chctl_all_flags(part); part 563 drivers/misc/sgi-xp/xpc_channel.c for (ch_number = 0; ch_number < part->nchannels; ch_number++) { part 564 drivers/misc/sgi-xp/xpc_channel.c ch = &part->channels[ch_number]; part 572 drivers/misc/sgi-xp/xpc_channel.c xpc_process_openclose_chctl_flags(part, ch_number, part 585 drivers/misc/sgi-xp/xpc_channel.c if (part->act_state == XPC_P_AS_DEACTIVATING) part 603 drivers/misc/sgi-xp/xpc_channel.c xpc_arch_ops.process_msg_chctl_flags(part, ch_number); part 617 drivers/misc/sgi-xp/xpc_channel.c xpc_partition_going_down(struct xpc_partition *part, enum xp_retval reason) part 624 drivers/misc/sgi-xp/xpc_channel.c XPC_PARTID(part), reason); part 626 drivers/misc/sgi-xp/xpc_channel.c if (!xpc_part_ref(part)) { part 633 drivers/misc/sgi-xp/xpc_channel.c for (ch_number = 0; ch_number < part->nchannels; ch_number++) { part 634 drivers/misc/sgi-xp/xpc_channel.c ch = &part->channels[ch_number]; part 645 drivers/misc/sgi-xp/xpc_channel.c xpc_wakeup_channel_mgr(part); part 647 drivers/misc/sgi-xp/xpc_channel.c xpc_part_deref(part); part 658 drivers/misc/sgi-xp/xpc_channel.c struct xpc_partition *part; part 663 drivers/misc/sgi-xp/xpc_channel.c part = &xpc_partitions[partid]; part 665 drivers/misc/sgi-xp/xpc_channel.c if (xpc_part_ref(part)) { part 670 drivers/misc/sgi-xp/xpc_channel.c xpc_wakeup_channel_mgr(part); part 671 drivers/misc/sgi-xp/xpc_channel.c xpc_part_deref(part); part 711 drivers/misc/sgi-xp/xpc_channel.c struct xpc_partition *part; part 718 drivers/misc/sgi-xp/xpc_channel.c part = &xpc_partitions[partid]; part 720 drivers/misc/sgi-xp/xpc_channel.c if (xpc_part_ref(part)) { part 721 drivers/misc/sgi-xp/xpc_channel.c ch = &part->channels[ch_number]; part 736 drivers/misc/sgi-xp/xpc_channel.c xpc_part_deref(part); part 875 drivers/misc/sgi-xp/xpc_channel.c struct xpc_partition *part = &xpc_partitions[partid]; part 882 drivers/misc/sgi-xp/xpc_channel.c DBUG_ON(ch_number < 0 || ch_number >= part->nchannels); part 885 drivers/misc/sgi-xp/xpc_channel.c if (xpc_part_ref(part)) { part 886 drivers/misc/sgi-xp/xpc_channel.c ret = xpc_arch_ops.send_payload(&part->channels[ch_number], part 888 drivers/misc/sgi-xp/xpc_channel.c xpc_part_deref(part); part 926 drivers/misc/sgi-xp/xpc_channel.c struct xpc_partition *part = &xpc_partitions[partid]; part 933 drivers/misc/sgi-xp/xpc_channel.c DBUG_ON(ch_number < 0 || ch_number >= part->nchannels); part 937 drivers/misc/sgi-xp/xpc_channel.c if (xpc_part_ref(part)) { part 938 drivers/misc/sgi-xp/xpc_channel.c ret = xpc_arch_ops.send_payload(&part->channels[ch_number], part 940 drivers/misc/sgi-xp/xpc_channel.c xpc_part_deref(part); part 1000 drivers/misc/sgi-xp/xpc_channel.c struct xpc_partition *part = &xpc_partitions[partid]; part 1004 drivers/misc/sgi-xp/xpc_channel.c DBUG_ON(ch_number < 0 || ch_number >= part->nchannels); part 1006 drivers/misc/sgi-xp/xpc_channel.c ch = &part->channels[ch_number]; part 177 drivers/misc/sgi-xp/xpc_main.c struct xpc_partition *part = from_timer(part, t, disengage_timer); part 179 drivers/misc/sgi-xp/xpc_main.c DBUG_ON(time_is_after_jiffies(part->disengage_timeout)); part 181 drivers/misc/sgi-xp/xpc_main.c (void)xpc_partition_disengaged(part); part 183 drivers/misc/sgi-xp/xpc_main.c DBUG_ON(part->disengage_timeout != 0); part 184 drivers/misc/sgi-xp/xpc_main.c DBUG_ON(xpc_arch_ops.partition_engaged(XPC_PARTID(part))); part 226 drivers/misc/sgi-xp/xpc_main.c struct xpc_partition *part; part 238 drivers/misc/sgi-xp/xpc_main.c part = &xpc_partitions[partid]; part 240 drivers/misc/sgi-xp/xpc_main.c if (part->act_state == XPC_P_AS_INACTIVE || part 241 drivers/misc/sgi-xp/xpc_main.c part->act_state == XPC_P_AS_DEACTIVATING) { part 245 drivers/misc/sgi-xp/xpc_main.c ret = xpc_arch_ops.get_remote_heartbeat(part); part 247 drivers/misc/sgi-xp/xpc_main.c XPC_DEACTIVATE_PARTITION(part, ret); part 339 drivers/misc/sgi-xp/xpc_main.c xpc_channel_mgr(struct xpc_partition *part) part 341 drivers/misc/sgi-xp/xpc_main.c while (part->act_state != XPC_P_AS_DEACTIVATING || part 342 drivers/misc/sgi-xp/xpc_main.c atomic_read(&part->nchannels_active) > 0 || part 343 drivers/misc/sgi-xp/xpc_main.c !xpc_partition_disengaged(part)) { part 345 drivers/misc/sgi-xp/xpc_main.c xpc_process_sent_chctl_flags(part); part 360 drivers/misc/sgi-xp/xpc_main.c atomic_dec(&part->channel_mgr_requests); part 361 drivers/misc/sgi-xp/xpc_main.c (void)wait_event_interruptible(part->channel_mgr_wq, part 362 drivers/misc/sgi-xp/xpc_main.c (atomic_read(&part->channel_mgr_requests) > 0 || part 363 drivers/misc/sgi-xp/xpc_main.c part->chctl.all_flags != 0 || part 364 drivers/misc/sgi-xp/xpc_main.c (part->act_state == XPC_P_AS_DEACTIVATING && part 365 drivers/misc/sgi-xp/xpc_main.c atomic_read(&part->nchannels_active) == 0 && part 366 drivers/misc/sgi-xp/xpc_main.c xpc_partition_disengaged(part)))); part 367 drivers/misc/sgi-xp/xpc_main.c atomic_set(&part->channel_mgr_requests, 1); part 400 drivers/misc/sgi-xp/xpc_main.c xpc_setup_ch_structures(struct xpc_partition *part) part 405 drivers/misc/sgi-xp/xpc_main.c short partid = XPC_PARTID(part); part 411 drivers/misc/sgi-xp/xpc_main.c DBUG_ON(part->channels != NULL); part 412 drivers/misc/sgi-xp/xpc_main.c part->channels = kcalloc(XPC_MAX_NCHANNELS, part 415 drivers/misc/sgi-xp/xpc_main.c if (part->channels == NULL) { part 422 drivers/misc/sgi-xp/xpc_main.c part->remote_openclose_args = part 424 drivers/misc/sgi-xp/xpc_main.c GFP_KERNEL, &part-> part 426 drivers/misc/sgi-xp/xpc_main.c if (part->remote_openclose_args == NULL) { part 432 drivers/misc/sgi-xp/xpc_main.c part->chctl.all_flags = 0; part 433 drivers/misc/sgi-xp/xpc_main.c spin_lock_init(&part->chctl_lock); part 435 drivers/misc/sgi-xp/xpc_main.c atomic_set(&part->channel_mgr_requests, 1); part 436 drivers/misc/sgi-xp/xpc_main.c init_waitqueue_head(&part->channel_mgr_wq); part 438 drivers/misc/sgi-xp/xpc_main.c part->nchannels = XPC_MAX_NCHANNELS; part 440 drivers/misc/sgi-xp/xpc_main.c atomic_set(&part->nchannels_active, 0); part 441 drivers/misc/sgi-xp/xpc_main.c atomic_set(&part->nchannels_engaged, 0); part 443 drivers/misc/sgi-xp/xpc_main.c for (ch_number = 0; ch_number < part->nchannels; ch_number++) { part 444 drivers/misc/sgi-xp/xpc_main.c ch = &part->channels[ch_number]; part 465 drivers/misc/sgi-xp/xpc_main.c ret = xpc_arch_ops.setup_ch_structures(part); part 473 drivers/misc/sgi-xp/xpc_main.c part->setup_state = XPC_P_SS_SETUP; part 479 drivers/misc/sgi-xp/xpc_main.c kfree(part->remote_openclose_args_base); part 480 drivers/misc/sgi-xp/xpc_main.c part->remote_openclose_args = NULL; part 482 drivers/misc/sgi-xp/xpc_main.c kfree(part->channels); part 483 drivers/misc/sgi-xp/xpc_main.c part->channels = NULL; part 492 drivers/misc/sgi-xp/xpc_main.c xpc_teardown_ch_structures(struct xpc_partition *part) part 494 drivers/misc/sgi-xp/xpc_main.c DBUG_ON(atomic_read(&part->nchannels_engaged) != 0); part 495 drivers/misc/sgi-xp/xpc_main.c DBUG_ON(atomic_read(&part->nchannels_active) != 0); part 502 drivers/misc/sgi-xp/xpc_main.c DBUG_ON(part->setup_state != XPC_P_SS_SETUP); part 503 drivers/misc/sgi-xp/xpc_main.c part->setup_state = XPC_P_SS_WTEARDOWN; part 505 drivers/misc/sgi-xp/xpc_main.c wait_event(part->teardown_wq, (atomic_read(&part->references) == 0)); part 509 drivers/misc/sgi-xp/xpc_main.c xpc_arch_ops.teardown_ch_structures(part); part 511 drivers/misc/sgi-xp/xpc_main.c kfree(part->remote_openclose_args_base); part 512 drivers/misc/sgi-xp/xpc_main.c part->remote_openclose_args = NULL; part 513 drivers/misc/sgi-xp/xpc_main.c kfree(part->channels); part 514 drivers/misc/sgi-xp/xpc_main.c part->channels = NULL; part 516 drivers/misc/sgi-xp/xpc_main.c part->setup_state = XPC_P_SS_TORNDOWN; part 534 drivers/misc/sgi-xp/xpc_main.c struct xpc_partition *part = &xpc_partitions[partid]; part 539 drivers/misc/sgi-xp/xpc_main.c spin_lock_irqsave(&part->act_lock, irq_flags); part 541 drivers/misc/sgi-xp/xpc_main.c if (part->act_state == XPC_P_AS_DEACTIVATING) { part 542 drivers/misc/sgi-xp/xpc_main.c part->act_state = XPC_P_AS_INACTIVE; part 543 drivers/misc/sgi-xp/xpc_main.c spin_unlock_irqrestore(&part->act_lock, irq_flags); part 544 drivers/misc/sgi-xp/xpc_main.c part->remote_rp_pa = 0; part 549 drivers/misc/sgi-xp/xpc_main.c DBUG_ON(part->act_state != XPC_P_AS_ACTIVATION_REQ); part 550 drivers/misc/sgi-xp/xpc_main.c part->act_state = XPC_P_AS_ACTIVATING; part 552 drivers/misc/sgi-xp/xpc_main.c XPC_SET_REASON(part, 0, 0); part 553 drivers/misc/sgi-xp/xpc_main.c spin_unlock_irqrestore(&part->act_lock, irq_flags); part 559 drivers/misc/sgi-xp/xpc_main.c if (xpc_setup_ch_structures(part) == xpSuccess) { part 560 drivers/misc/sgi-xp/xpc_main.c (void)xpc_part_ref(part); /* this will always succeed */ part 562 drivers/misc/sgi-xp/xpc_main.c if (xpc_arch_ops.make_first_contact(part) == xpSuccess) { part 563 drivers/misc/sgi-xp/xpc_main.c xpc_mark_partition_active(part); part 564 drivers/misc/sgi-xp/xpc_main.c xpc_channel_mgr(part); part 568 drivers/misc/sgi-xp/xpc_main.c xpc_part_deref(part); part 569 drivers/misc/sgi-xp/xpc_main.c xpc_teardown_ch_structures(part); part 573 drivers/misc/sgi-xp/xpc_main.c xpc_mark_partition_inactive(part); part 575 drivers/misc/sgi-xp/xpc_main.c if (part->reason == xpReactivating) { part 577 drivers/misc/sgi-xp/xpc_main.c xpc_arch_ops.request_partition_reactivation(part); part 584 drivers/misc/sgi-xp/xpc_main.c xpc_activate_partition(struct xpc_partition *part) part 586 drivers/misc/sgi-xp/xpc_main.c short partid = XPC_PARTID(part); part 590 drivers/misc/sgi-xp/xpc_main.c spin_lock_irqsave(&part->act_lock, irq_flags); part 592 drivers/misc/sgi-xp/xpc_main.c DBUG_ON(part->act_state != XPC_P_AS_INACTIVE); part 594 drivers/misc/sgi-xp/xpc_main.c part->act_state = XPC_P_AS_ACTIVATION_REQ; part 595 drivers/misc/sgi-xp/xpc_main.c XPC_SET_REASON(part, xpCloneKThread, __LINE__); part 597 drivers/misc/sgi-xp/xpc_main.c spin_unlock_irqrestore(&part->act_lock, irq_flags); part 602 drivers/misc/sgi-xp/xpc_main.c spin_lock_irqsave(&part->act_lock, irq_flags); part 603 drivers/misc/sgi-xp/xpc_main.c part->act_state = XPC_P_AS_INACTIVE; part 604 drivers/misc/sgi-xp/xpc_main.c XPC_SET_REASON(part, xpCloneKThreadFailed, __LINE__); part 605 drivers/misc/sgi-xp/xpc_main.c spin_unlock_irqrestore(&part->act_lock, irq_flags); part 648 drivers/misc/sgi-xp/xpc_main.c xpc_kthread_waitmsgs(struct xpc_partition *part, struct xpc_channel *ch) part 685 drivers/misc/sgi-xp/xpc_main.c struct xpc_partition *part = &xpc_partitions[partid]; part 695 drivers/misc/sgi-xp/xpc_main.c ch = &part->channels[ch_number]; part 727 drivers/misc/sgi-xp/xpc_main.c xpc_kthread_waitmsgs(part, ch); part 746 drivers/misc/sgi-xp/xpc_main.c atomic_dec_return(&part->nchannels_engaged) == 0) { part 747 drivers/misc/sgi-xp/xpc_main.c xpc_arch_ops.indicate_partition_disengaged(part); part 755 drivers/misc/sgi-xp/xpc_main.c xpc_part_deref(part); part 777 drivers/misc/sgi-xp/xpc_main.c struct xpc_partition *part = &xpc_partitions[ch->partid]; part 801 drivers/misc/sgi-xp/xpc_main.c atomic_inc_return(&part->nchannels_engaged) == 1) { part 802 drivers/misc/sgi-xp/xpc_main.c xpc_arch_ops.indicate_partition_engaged(part); part 804 drivers/misc/sgi-xp/xpc_main.c (void)xpc_part_ref(part); part 823 drivers/misc/sgi-xp/xpc_main.c atomic_dec_return(&part->nchannels_engaged) == 0) { part 824 drivers/misc/sgi-xp/xpc_main.c indicate_partition_disengaged(part); part 827 drivers/misc/sgi-xp/xpc_main.c xpc_part_deref(part); part 851 drivers/misc/sgi-xp/xpc_main.c struct xpc_partition *part; part 857 drivers/misc/sgi-xp/xpc_main.c part = &xpc_partitions[partid]; part 859 drivers/misc/sgi-xp/xpc_main.c if (!xpc_part_ref(part)) part 862 drivers/misc/sgi-xp/xpc_main.c ch = &part->channels[ch_number]; part 865 drivers/misc/sgi-xp/xpc_main.c xpc_part_deref(part); part 876 drivers/misc/sgi-xp/xpc_main.c if (part->act_state != XPC_P_AS_DEACTIVATING) { part 877 drivers/misc/sgi-xp/xpc_main.c spin_lock(&part->chctl_lock); part 878 drivers/misc/sgi-xp/xpc_main.c part->chctl.flags[ch->number] |= part 880 drivers/misc/sgi-xp/xpc_main.c spin_unlock(&part->chctl_lock); part 890 drivers/misc/sgi-xp/xpc_main.c xpc_wakeup_channel_mgr(part); part 892 drivers/misc/sgi-xp/xpc_main.c xpc_part_deref(part); part 900 drivers/misc/sgi-xp/xpc_main.c struct xpc_partition *part; part 919 drivers/misc/sgi-xp/xpc_main.c part = &xpc_partitions[partid]; part 921 drivers/misc/sgi-xp/xpc_main.c DBUG_ON((u64)part != L1_CACHE_ALIGN((u64)part)); part 923 drivers/misc/sgi-xp/xpc_main.c part->activate_IRQ_rcvd = 0; part 924 drivers/misc/sgi-xp/xpc_main.c spin_lock_init(&part->act_lock); part 925 drivers/misc/sgi-xp/xpc_main.c part->act_state = XPC_P_AS_INACTIVE; part 926 drivers/misc/sgi-xp/xpc_main.c XPC_SET_REASON(part, 0, 0); part 928 drivers/misc/sgi-xp/xpc_main.c timer_setup(&part->disengage_timer, part 931 drivers/misc/sgi-xp/xpc_main.c part->setup_state = XPC_P_SS_UNSET; part 932 drivers/misc/sgi-xp/xpc_main.c init_waitqueue_head(&part->teardown_wq); part 933 drivers/misc/sgi-xp/xpc_main.c atomic_set(&part->references, 0); part 951 drivers/misc/sgi-xp/xpc_main.c struct xpc_partition *part; part 983 drivers/misc/sgi-xp/xpc_main.c part = &xpc_partitions[partid]; part 985 drivers/misc/sgi-xp/xpc_main.c if (xpc_partition_disengaged(part) && part 986 drivers/misc/sgi-xp/xpc_main.c part->act_state == XPC_P_AS_INACTIVE) { part 992 drivers/misc/sgi-xp/xpc_main.c XPC_DEACTIVATE_PARTITION(part, reason); part 994 drivers/misc/sgi-xp/xpc_main.c if (part->disengage_timeout > disengage_timeout) part 995 drivers/misc/sgi-xp/xpc_main.c disengage_timeout = part->disengage_timeout; part 1086 drivers/misc/sgi-xp/xpc_main.c struct xpc_partition *part; part 1101 drivers/misc/sgi-xp/xpc_main.c part = &xpc_partitions[partid]; part 1104 drivers/misc/sgi-xp/xpc_main.c part->act_state != XPC_P_AS_INACTIVE) { part 1105 drivers/misc/sgi-xp/xpc_main.c xpc_arch_ops.request_partition_deactivation(part); part 1106 drivers/misc/sgi-xp/xpc_main.c xpc_arch_ops.indicate_partition_disengaged(part); part 265 drivers/misc/sgi-xp/xpc_partition.c xpc_partition_disengaged(struct xpc_partition *part) part 267 drivers/misc/sgi-xp/xpc_partition.c short partid = XPC_PARTID(part); part 271 drivers/misc/sgi-xp/xpc_partition.c if (part->disengage_timeout) { part 273 drivers/misc/sgi-xp/xpc_partition.c if (time_is_after_jiffies(part->disengage_timeout)) { part 289 drivers/misc/sgi-xp/xpc_partition.c part->disengage_timeout = 0; part 293 drivers/misc/sgi-xp/xpc_partition.c del_singleshot_timer_sync(&part->disengage_timer); part 295 drivers/misc/sgi-xp/xpc_partition.c DBUG_ON(part->act_state != XPC_P_AS_DEACTIVATING && part 296 drivers/misc/sgi-xp/xpc_partition.c part->act_state != XPC_P_AS_INACTIVE); part 297 drivers/misc/sgi-xp/xpc_partition.c if (part->act_state != XPC_P_AS_INACTIVE) part 298 drivers/misc/sgi-xp/xpc_partition.c xpc_wakeup_channel_mgr(part); part 300 drivers/misc/sgi-xp/xpc_partition.c xpc_arch_ops.cancel_partition_deactivation_request(part); part 309 drivers/misc/sgi-xp/xpc_partition.c xpc_mark_partition_active(struct xpc_partition *part) part 314 drivers/misc/sgi-xp/xpc_partition.c dev_dbg(xpc_part, "setting partition %d to ACTIVE\n", XPC_PARTID(part)); part 316 drivers/misc/sgi-xp/xpc_partition.c spin_lock_irqsave(&part->act_lock, irq_flags); part 317 drivers/misc/sgi-xp/xpc_partition.c if (part->act_state == XPC_P_AS_ACTIVATING) { part 318 drivers/misc/sgi-xp/xpc_partition.c part->act_state = XPC_P_AS_ACTIVE; part 321 drivers/misc/sgi-xp/xpc_partition.c DBUG_ON(part->reason == xpSuccess); part 322 drivers/misc/sgi-xp/xpc_partition.c ret = part->reason; part 324 drivers/misc/sgi-xp/xpc_partition.c spin_unlock_irqrestore(&part->act_lock, irq_flags); part 333 drivers/misc/sgi-xp/xpc_partition.c xpc_deactivate_partition(const int line, struct xpc_partition *part, part 338 drivers/misc/sgi-xp/xpc_partition.c spin_lock_irqsave(&part->act_lock, irq_flags); part 340 drivers/misc/sgi-xp/xpc_partition.c if (part->act_state == XPC_P_AS_INACTIVE) { part 341 drivers/misc/sgi-xp/xpc_partition.c XPC_SET_REASON(part, reason, line); part 342 drivers/misc/sgi-xp/xpc_partition.c spin_unlock_irqrestore(&part->act_lock, irq_flags); part 345 drivers/misc/sgi-xp/xpc_partition.c xpc_arch_ops.request_partition_reactivation(part); part 349 drivers/misc/sgi-xp/xpc_partition.c if (part->act_state == XPC_P_AS_DEACTIVATING) { part 350 drivers/misc/sgi-xp/xpc_partition.c if ((part->reason == xpUnloading && reason != xpUnloading) || part 352 drivers/misc/sgi-xp/xpc_partition.c XPC_SET_REASON(part, reason, line); part 354 drivers/misc/sgi-xp/xpc_partition.c spin_unlock_irqrestore(&part->act_lock, irq_flags); part 358 drivers/misc/sgi-xp/xpc_partition.c part->act_state = XPC_P_AS_DEACTIVATING; part 359 drivers/misc/sgi-xp/xpc_partition.c XPC_SET_REASON(part, reason, line); part 361 drivers/misc/sgi-xp/xpc_partition.c spin_unlock_irqrestore(&part->act_lock, irq_flags); part 364 drivers/misc/sgi-xp/xpc_partition.c xpc_arch_ops.request_partition_deactivation(part); part 367 drivers/misc/sgi-xp/xpc_partition.c part->disengage_timeout = jiffies + (xpc_disengage_timelimit * HZ); part 368 drivers/misc/sgi-xp/xpc_partition.c part->disengage_timer.expires = part->disengage_timeout; part 369 drivers/misc/sgi-xp/xpc_partition.c add_timer(&part->disengage_timer); part 372 drivers/misc/sgi-xp/xpc_partition.c XPC_PARTID(part), reason); part 374 drivers/misc/sgi-xp/xpc_partition.c xpc_partition_going_down(part, reason); part 381 drivers/misc/sgi-xp/xpc_partition.c xpc_mark_partition_inactive(struct xpc_partition *part) part 386 drivers/misc/sgi-xp/xpc_partition.c XPC_PARTID(part)); part 388 drivers/misc/sgi-xp/xpc_partition.c spin_lock_irqsave(&part->act_lock, irq_flags); part 389 drivers/misc/sgi-xp/xpc_partition.c part->act_state = XPC_P_AS_INACTIVE; part 390 drivers/misc/sgi-xp/xpc_partition.c spin_unlock_irqrestore(&part->act_lock, irq_flags); part 391 drivers/misc/sgi-xp/xpc_partition.c part->remote_rp_pa = 0; part 521 drivers/misc/sgi-xp/xpc_partition.c struct xpc_partition *part; part 524 drivers/misc/sgi-xp/xpc_partition.c part = &xpc_partitions[partid]; part 525 drivers/misc/sgi-xp/xpc_partition.c if (part->remote_rp_pa == 0) part 530 drivers/misc/sgi-xp/xpc_partition.c part_nasid_pa = (unsigned long)XPC_RP_PART_NASIDS(part->remote_rp_pa); part 376 drivers/misc/sgi-xp/xpc_uv.c struct xpc_partition *part; part 383 drivers/misc/sgi-xp/xpc_uv.c part = &xpc_partitions[partid]; part 385 drivers/misc/sgi-xp/xpc_uv.c if (part->sn.uv.act_state_req == 0) part 391 drivers/misc/sgi-xp/xpc_uv.c act_state_req = part->sn.uv.act_state_req; part 392 drivers/misc/sgi-xp/xpc_uv.c part->sn.uv.act_state_req = 0; part 396 drivers/misc/sgi-xp/xpc_uv.c if (part->act_state == XPC_P_AS_INACTIVE) part 397 drivers/misc/sgi-xp/xpc_uv.c xpc_activate_partition(part); part 398 drivers/misc/sgi-xp/xpc_uv.c else if (part->act_state == XPC_P_AS_DEACTIVATING) part 399 drivers/misc/sgi-xp/xpc_uv.c XPC_DEACTIVATE_PARTITION(part, xpReactivating); part 402 drivers/misc/sgi-xp/xpc_uv.c if (part->act_state == XPC_P_AS_INACTIVE) part 403 drivers/misc/sgi-xp/xpc_uv.c xpc_activate_partition(part); part 405 drivers/misc/sgi-xp/xpc_uv.c XPC_DEACTIVATE_PARTITION(part, xpReactivating); part 408 drivers/misc/sgi-xp/xpc_uv.c XPC_DEACTIVATE_PARTITION(part, part->sn.uv.reason); part 423 drivers/misc/sgi-xp/xpc_uv.c xpc_handle_activate_mq_msg_uv(struct xpc_partition *part, part 429 drivers/misc/sgi-xp/xpc_uv.c struct xpc_partition_uv *part_uv = &part->sn.uv; part 454 drivers/misc/sgi-xp/xpc_uv.c part->remote_rp_pa = msg->rp_gpa; /* !!! _pa is _gpa */ part 455 drivers/misc/sgi-xp/xpc_uv.c part->remote_rp_ts_jiffies = msg_hdr->rp_ts_jiffies; part 496 drivers/misc/sgi-xp/xpc_uv.c args = &part->remote_openclose_args[msg->ch_number]; part 499 drivers/misc/sgi-xp/xpc_uv.c spin_lock_irqsave(&part->chctl_lock, irq_flags); part 500 drivers/misc/sgi-xp/xpc_uv.c part->chctl.flags[msg->ch_number] |= XPC_CHCTL_CLOSEREQUEST; part 501 drivers/misc/sgi-xp/xpc_uv.c spin_unlock_irqrestore(&part->chctl_lock, irq_flags); part 503 drivers/misc/sgi-xp/xpc_uv.c xpc_wakeup_channel_mgr(part); part 516 drivers/misc/sgi-xp/xpc_uv.c spin_lock_irqsave(&part->chctl_lock, irq_flags); part 517 drivers/misc/sgi-xp/xpc_uv.c part->chctl.flags[msg->ch_number] |= XPC_CHCTL_CLOSEREPLY; part 518 drivers/misc/sgi-xp/xpc_uv.c spin_unlock_irqrestore(&part->chctl_lock, irq_flags); part 520 drivers/misc/sgi-xp/xpc_uv.c xpc_wakeup_channel_mgr(part); part 532 drivers/misc/sgi-xp/xpc_uv.c args = &part->remote_openclose_args[msg->ch_number]; part 536 drivers/misc/sgi-xp/xpc_uv.c spin_lock_irqsave(&part->chctl_lock, irq_flags); part 537 drivers/misc/sgi-xp/xpc_uv.c part->chctl.flags[msg->ch_number] |= XPC_CHCTL_OPENREQUEST; part 538 drivers/misc/sgi-xp/xpc_uv.c spin_unlock_irqrestore(&part->chctl_lock, irq_flags); part 540 drivers/misc/sgi-xp/xpc_uv.c xpc_wakeup_channel_mgr(part); part 551 drivers/misc/sgi-xp/xpc_uv.c args = &part->remote_openclose_args[msg->ch_number]; part 556 drivers/misc/sgi-xp/xpc_uv.c spin_lock_irqsave(&part->chctl_lock, irq_flags); part 557 drivers/misc/sgi-xp/xpc_uv.c part->chctl.flags[msg->ch_number] |= XPC_CHCTL_OPENREPLY; part 558 drivers/misc/sgi-xp/xpc_uv.c spin_unlock_irqrestore(&part->chctl_lock, irq_flags); part 560 drivers/misc/sgi-xp/xpc_uv.c xpc_wakeup_channel_mgr(part); part 571 drivers/misc/sgi-xp/xpc_uv.c spin_lock_irqsave(&part->chctl_lock, irq_flags); part 572 drivers/misc/sgi-xp/xpc_uv.c part->chctl.flags[msg->ch_number] |= XPC_CHCTL_OPENCOMPLETE; part 573 drivers/misc/sgi-xp/xpc_uv.c spin_unlock_irqrestore(&part->chctl_lock, irq_flags); part 575 drivers/misc/sgi-xp/xpc_uv.c xpc_wakeup_channel_mgr(part); part 592 drivers/misc/sgi-xp/xpc_uv.c "from partition=%d\n", msg_hdr->type, XPC_PARTID(part)); part 606 drivers/misc/sgi-xp/xpc_uv.c if (msg_hdr->rp_ts_jiffies != part->remote_rp_ts_jiffies && part 607 drivers/misc/sgi-xp/xpc_uv.c part->remote_rp_ts_jiffies != 0) { part 627 drivers/misc/sgi-xp/xpc_uv.c struct xpc_partition *part; part 642 drivers/misc/sgi-xp/xpc_uv.c part = &xpc_partitions[partid]; part 644 drivers/misc/sgi-xp/xpc_uv.c part_referenced = xpc_part_ref(part); part 645 drivers/misc/sgi-xp/xpc_uv.c xpc_handle_activate_mq_msg_uv(part, msg_hdr, part 649 drivers/misc/sgi-xp/xpc_uv.c xpc_part_deref(part); part 676 drivers/misc/sgi-xp/xpc_uv.c xpc_send_activate_IRQ_uv(struct xpc_partition *part, void *msg, size_t msg_size, part 680 drivers/misc/sgi-xp/xpc_uv.c struct xpc_partition_uv *part_uv = &part->sn.uv; part 689 drivers/misc/sgi-xp/xpc_uv.c msg_hdr->act_state = part->act_state; part 732 drivers/misc/sgi-xp/xpc_uv.c xpc_send_activate_IRQ_part_uv(struct xpc_partition *part, void *msg, part 737 drivers/misc/sgi-xp/xpc_uv.c ret = xpc_send_activate_IRQ_uv(part, msg, msg_size, msg_type); part 739 drivers/misc/sgi-xp/xpc_uv.c XPC_DEACTIVATE_PARTITION(part, ret); part 746 drivers/misc/sgi-xp/xpc_uv.c struct xpc_partition *part = &xpc_partitions[ch->partid]; part 749 drivers/misc/sgi-xp/xpc_uv.c ret = xpc_send_activate_IRQ_uv(part, msg, msg_size, msg_type); part 754 drivers/misc/sgi-xp/xpc_uv.c XPC_DEACTIVATE_PARTITION(part, ret); part 762 drivers/misc/sgi-xp/xpc_uv.c xpc_send_local_activate_IRQ_uv(struct xpc_partition *part, int act_state_req) part 765 drivers/misc/sgi-xp/xpc_uv.c struct xpc_partition_uv *part_uv = &part->sn.uv; part 875 drivers/misc/sgi-xp/xpc_uv.c xpc_get_remote_heartbeat_uv(struct xpc_partition *part) part 877 drivers/misc/sgi-xp/xpc_uv.c struct xpc_partition_uv *part_uv = &part->sn.uv; part 886 drivers/misc/sgi-xp/xpc_uv.c if (part_uv->cached_heartbeat.value == part->last_heartbeat && part 891 drivers/misc/sgi-xp/xpc_uv.c part->last_heartbeat = part_uv->cached_heartbeat.value; part 901 drivers/misc/sgi-xp/xpc_uv.c struct xpc_partition *part = &xpc_partitions[partid]; part 904 drivers/misc/sgi-xp/xpc_uv.c part->remote_rp_pa = remote_rp_gpa; /* !!! _pa here is really _gpa */ part 905 drivers/misc/sgi-xp/xpc_uv.c part->remote_rp_ts_jiffies = remote_rp->ts_jiffies; part 906 drivers/misc/sgi-xp/xpc_uv.c part->sn.uv.heartbeat_gpa = remote_rp->sn.uv.heartbeat_gpa; part 907 drivers/misc/sgi-xp/xpc_uv.c part->sn.uv.activate_gru_mq_desc_gpa = part 914 drivers/misc/sgi-xp/xpc_uv.c if (part->sn.uv.remote_act_state == XPC_P_AS_INACTIVE) { part 919 drivers/misc/sgi-xp/xpc_uv.c xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg), part 923 drivers/misc/sgi-xp/xpc_uv.c if (part->act_state == XPC_P_AS_INACTIVE) part 924 drivers/misc/sgi-xp/xpc_uv.c xpc_send_local_activate_IRQ_uv(part, XPC_P_ASR_ACTIVATE_UV); part 928 drivers/misc/sgi-xp/xpc_uv.c xpc_request_partition_reactivation_uv(struct xpc_partition *part) part 930 drivers/misc/sgi-xp/xpc_uv.c xpc_send_local_activate_IRQ_uv(part, XPC_P_ASR_ACTIVATE_UV); part 934 drivers/misc/sgi-xp/xpc_uv.c xpc_request_partition_deactivation_uv(struct xpc_partition *part) part 942 drivers/misc/sgi-xp/xpc_uv.c if (part->sn.uv.remote_act_state != XPC_P_AS_DEACTIVATING && part 943 drivers/misc/sgi-xp/xpc_uv.c part->sn.uv.remote_act_state != XPC_P_AS_INACTIVE) { part 945 drivers/misc/sgi-xp/xpc_uv.c msg.reason = part->reason; part 946 drivers/misc/sgi-xp/xpc_uv.c xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg), part 952 drivers/misc/sgi-xp/xpc_uv.c xpc_cancel_partition_deactivation_request_uv(struct xpc_partition *part) part 1016 drivers/misc/sgi-xp/xpc_uv.c xpc_setup_ch_structures_uv(struct xpc_partition *part) part 1021 drivers/misc/sgi-xp/xpc_uv.c for (ch_number = 0; ch_number < part->nchannels; ch_number++) { part 1022 drivers/misc/sgi-xp/xpc_uv.c ch_uv = &part->channels[ch_number].sn.uv; part 1035 drivers/misc/sgi-xp/xpc_uv.c xpc_teardown_ch_structures_uv(struct xpc_partition *part) part 1042 drivers/misc/sgi-xp/xpc_uv.c xpc_make_first_contact_uv(struct xpc_partition *part) part 1051 drivers/misc/sgi-xp/xpc_uv.c xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg), part 1054 drivers/misc/sgi-xp/xpc_uv.c while (!((part->sn.uv.remote_act_state == XPC_P_AS_ACTIVATING) || part 1055 drivers/misc/sgi-xp/xpc_uv.c (part->sn.uv.remote_act_state == XPC_P_AS_ACTIVE))) { part 1058 drivers/misc/sgi-xp/xpc_uv.c "partition %d\n", XPC_PARTID(part)); part 1063 drivers/misc/sgi-xp/xpc_uv.c if (part->act_state == XPC_P_AS_DEACTIVATING) part 1064 drivers/misc/sgi-xp/xpc_uv.c return part->reason; part 1071 drivers/misc/sgi-xp/xpc_uv.c xpc_get_chctl_all_flags_uv(struct xpc_partition *part) part 1076 drivers/misc/sgi-xp/xpc_uv.c spin_lock_irqsave(&part->chctl_lock, irq_flags); part 1077 drivers/misc/sgi-xp/xpc_uv.c chctl = part->chctl; part 1079 drivers/misc/sgi-xp/xpc_uv.c part->chctl.all_flags = 0; part 1081 drivers/misc/sgi-xp/xpc_uv.c spin_unlock_irqrestore(&part->chctl_lock, irq_flags); part 1260 drivers/misc/sgi-xp/xpc_uv.c xpc_send_chctl_local_msgrequest_uv(struct xpc_partition *part, int ch_number) part 1264 drivers/misc/sgi-xp/xpc_uv.c spin_lock_irqsave(&part->chctl_lock, irq_flags); part 1265 drivers/misc/sgi-xp/xpc_uv.c part->chctl.flags[ch_number] |= XPC_CHCTL_MSGREQUEST; part 1266 drivers/misc/sgi-xp/xpc_uv.c spin_unlock_irqrestore(&part->chctl_lock, irq_flags); part 1268 drivers/misc/sgi-xp/xpc_uv.c xpc_wakeup_channel_mgr(part); part 1283 drivers/misc/sgi-xp/xpc_uv.c xpc_indicate_partition_engaged_uv(struct xpc_partition *part) part 1287 drivers/misc/sgi-xp/xpc_uv.c xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg), part 1292 drivers/misc/sgi-xp/xpc_uv.c xpc_indicate_partition_disengaged_uv(struct xpc_partition *part) part 1296 drivers/misc/sgi-xp/xpc_uv.c xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg), part 1410 drivers/misc/sgi-xp/xpc_uv.c xpc_handle_notify_mq_msg_uv(struct xpc_partition *part, part 1413 drivers/misc/sgi-xp/xpc_uv.c struct xpc_partition_uv *part_uv = &part->sn.uv; part 1420 drivers/misc/sgi-xp/xpc_uv.c if (unlikely(ch_number >= part->nchannels)) { part 1423 drivers/misc/sgi-xp/xpc_uv.c ch_number, XPC_PARTID(part)); part 1437 drivers/misc/sgi-xp/xpc_uv.c ch = &part->channels[ch_number]; part 1473 drivers/misc/sgi-xp/xpc_uv.c xpc_send_chctl_local_msgrequest_uv(part, ch->number); part 1483 drivers/misc/sgi-xp/xpc_uv.c struct xpc_partition *part; part 1493 drivers/misc/sgi-xp/xpc_uv.c part = &xpc_partitions[partid]; part 1495 drivers/misc/sgi-xp/xpc_uv.c if (xpc_part_ref(part)) { part 1496 drivers/misc/sgi-xp/xpc_uv.c xpc_handle_notify_mq_msg_uv(part, msg); part 1497 drivers/misc/sgi-xp/xpc_uv.c xpc_part_deref(part); part 1514 drivers/misc/sgi-xp/xpc_uv.c xpc_process_msg_chctl_flags_uv(struct xpc_partition *part, int ch_number) part 1516 drivers/misc/sgi-xp/xpc_uv.c struct xpc_channel *ch = &part->channels[ch_number]; part 31 drivers/misc/sram-exec.c struct sram_partition *part) part 33 drivers/misc/sram-exec.c unsigned long base = (unsigned long)part->base; part 45 drivers/misc/sram-exec.c int sram_add_protect_exec(struct sram_partition *part) part 48 drivers/misc/sram-exec.c list_add_tail(&part->list, &exec_pool_list); part 84 drivers/misc/sram-exec.c struct sram_partition *part = NULL, *p; part 92 drivers/misc/sram-exec.c part = p; part 96 drivers/misc/sram-exec.c if (!part) part 102 drivers/misc/sram-exec.c base = (unsigned long)part->base; part 105 drivers/misc/sram-exec.c mutex_lock(&part->lock); part 115 drivers/misc/sram-exec.c mutex_unlock(&part->lock); part 29 drivers/misc/sram.c struct sram_partition *part; part 31 drivers/misc/sram.c part = container_of(attr, struct sram_partition, battr); part 33 drivers/misc/sram.c mutex_lock(&part->lock); part 34 drivers/misc/sram.c memcpy_fromio(buf, part->base + pos, count); part 35 drivers/misc/sram.c mutex_unlock(&part->lock); part 44 drivers/misc/sram.c struct sram_partition *part; part 46 drivers/misc/sram.c part = container_of(attr, struct sram_partition, battr); part 48 drivers/misc/sram.c mutex_lock(&part->lock); part 49 drivers/misc/sram.c memcpy_toio(part->base + pos, buf, count); part 50 drivers/misc/sram.c mutex_unlock(&part->lock); part 56 drivers/misc/sram.c phys_addr_t start, struct sram_partition *part) part 60 drivers/misc/sram.c part->pool = devm_gen_pool_create(sram->dev, ilog2(SRAM_GRANULARITY), part 62 drivers/misc/sram.c if (IS_ERR(part->pool)) part 63 drivers/misc/sram.c return PTR_ERR(part->pool); part 65 drivers/misc/sram.c ret = gen_pool_add_virt(part->pool, (unsigned long)part->base, start, part 76 drivers/misc/sram.c phys_addr_t start, struct sram_partition *part) part 78 drivers/misc/sram.c sysfs_bin_attr_init(&part->battr); part 79 drivers/misc/sram.c part->battr.attr.name = devm_kasprintf(sram->dev, GFP_KERNEL, part 82 drivers/misc/sram.c if (!part->battr.attr.name) part 85 drivers/misc/sram.c part->battr.attr.mode = S_IRUSR | S_IWUSR; part 86 drivers/misc/sram.c part->battr.read = sram_read; part 87 drivers/misc/sram.c part->battr.write = sram_write; part 88 drivers/misc/sram.c part->battr.size = block->size; part 90 drivers/misc/sram.c return device_create_bin_file(sram->dev, &part->battr); part 97 drivers/misc/sram.c struct sram_partition *part = &sram->partition[sram->partitions]; part 99 drivers/misc/sram.c mutex_init(&part->lock); part 100 drivers/misc/sram.c part->base = sram->virt_base + block->start; part 103 drivers/misc/sram.c ret = sram_add_pool(sram, block, start, part); part 108 drivers/misc/sram.c ret = sram_add_export(sram, block, start, part); part 113 drivers/misc/sram.c ret = sram_check_protect_exec(sram, block, part); part 117 drivers/misc/sram.c ret = sram_add_pool(sram, block, start, part); part 121 drivers/misc/sram.c sram_add_protect_exec(part); part 131 drivers/misc/sram.c struct sram_partition *part; part 136 drivers/misc/sram.c part = &sram->partition[sram->partitions - 1]; part 137 drivers/misc/sram.c for (; sram->partitions; sram->partitions--, part--) { part 138 drivers/misc/sram.c if (part->battr.size) part 139 drivers/misc/sram.c device_remove_bin_file(sram->dev, &part->battr); part 141 drivers/misc/sram.c if (part->pool && part 142 drivers/misc/sram.c gen_pool_avail(part->pool) < gen_pool_size(part->pool)) part 40 drivers/misc/sram.h struct sram_partition *part); part 41 drivers/misc/sram.h int sram_add_protect_exec(struct sram_partition *part); part 45 drivers/misc/sram.h struct sram_partition *part) part 50 drivers/misc/sram.h static inline int sram_add_protect_exec(struct sram_partition *part) part 106 drivers/mmc/core/block.c struct list_head part; part 261 drivers/mmc/core/block.c list_for_each_entry(part_md, &md->part, part) part 2292 drivers/mmc/core/block.c INIT_LIST_HEAD(&md->part); part 2409 drivers/mmc/core/block.c list_add(&part_md->part, &md->part); part 2585 drivers/mmc/core/block.c if (card->part[idx].area_type & MMC_BLK_DATA_AREA_RPMB) { part 2593 drivers/mmc/core/block.c card->part[idx].part_cfg, part 2594 drivers/mmc/core/block.c card->part[idx].size >> 9, part 2595 drivers/mmc/core/block.c card->part[idx].name); part 2598 drivers/mmc/core/block.c } else if (card->part[idx].size) { part 2600 drivers/mmc/core/block.c card->part[idx].part_cfg, part 2601 drivers/mmc/core/block.c card->part[idx].size >> 9, part 2602 drivers/mmc/core/block.c card->part[idx].force_ro, part 2603 drivers/mmc/core/block.c card->part[idx].name, part 2604 drivers/mmc/core/block.c card->part[idx].area_type); part 2652 drivers/mmc/core/block.c list_for_each_safe(pos, q, &md->part) { part 2653 drivers/mmc/core/block.c part_md = list_entry(pos, struct mmc_blk_data, part); part 2905 drivers/mmc/core/block.c list_for_each_entry(part_md, &md->part, part) { part 2960 drivers/mmc/core/block.c list_for_each_entry(part_md, &md->part, part) { part 2992 drivers/mmc/core/block.c list_for_each_entry(part_md, &md->part, part) { part 304 drivers/mmc/core/mmc.c card->part[card->nr_parts].size = size; part 305 drivers/mmc/core/mmc.c card->part[card->nr_parts].part_cfg = part_cfg; part 306 drivers/mmc/core/mmc.c sprintf(card->part[card->nr_parts].name, name, idx); part 307 drivers/mmc/core/mmc.c card->part[card->nr_parts].force_ro = ro; part 308 drivers/mmc/core/mmc.c card->part[card->nr_parts].area_type = area_type; part 151 drivers/mtd/ftl.c static int scan_header(partition_t *part) part 157 drivers/mtd/ftl.c part->header.FormattedSize = 0; part 158 drivers/mtd/ftl.c max_offset = (0x100000<part->mbd.mtd->size)?0x100000:part->mbd.mtd->size; part 162 drivers/mtd/ftl.c offset += part->mbd.mtd->erasesize ? : 0x2000) { part 164 drivers/mtd/ftl.c err = mtd_read(part->mbd.mtd, offset, sizeof(header), &ret, part 183 drivers/mtd/ftl.c if ((1 << header.EraseUnitSize) != part->mbd.mtd->erasesize) { part 185 drivers/mtd/ftl.c 1 << header.EraseUnitSize,part->mbd.mtd->erasesize); part 188 drivers/mtd/ftl.c part->header = header; part 192 drivers/mtd/ftl.c static int build_maps(partition_t *part) part 202 drivers/mtd/ftl.c part->DataUnits = le16_to_cpu(part->header.NumEraseUnits) - part 203 drivers/mtd/ftl.c part->header.NumTransferUnits; part 204 drivers/mtd/ftl.c part->EUNInfo = kmalloc_array(part->DataUnits, sizeof(struct eun_info_t), part 206 drivers/mtd/ftl.c if (!part->EUNInfo) part 208 drivers/mtd/ftl.c for (i = 0; i < part->DataUnits; i++) part 209 drivers/mtd/ftl.c part->EUNInfo[i].Offset = 0xffffffff; part 210 drivers/mtd/ftl.c part->XferInfo = part 211 drivers/mtd/ftl.c kmalloc_array(part->header.NumTransferUnits, part 214 drivers/mtd/ftl.c if (!part->XferInfo) part 218 drivers/mtd/ftl.c for (i = 0; i < le16_to_cpu(part->header.NumEraseUnits); i++) { part 219 drivers/mtd/ftl.c offset = ((i + le16_to_cpu(part->header.FirstPhysicalEUN)) part 220 drivers/mtd/ftl.c << part->header.EraseUnitSize); part 221 drivers/mtd/ftl.c ret = mtd_read(part->mbd.mtd, offset, sizeof(header), &retval, part 230 drivers/mtd/ftl.c if (hdr_ok && (le16_to_cpu(header.LogicalEUN) < part->DataUnits) && part 231 drivers/mtd/ftl.c (part->EUNInfo[le16_to_cpu(header.LogicalEUN)].Offset == 0xffffffff)) { part 232 drivers/mtd/ftl.c part->EUNInfo[le16_to_cpu(header.LogicalEUN)].Offset = offset; part 233 drivers/mtd/ftl.c part->EUNInfo[le16_to_cpu(header.LogicalEUN)].EraseCount = part 237 drivers/mtd/ftl.c if (xtrans == part->header.NumTransferUnits) { part 243 drivers/mtd/ftl.c part->XferInfo[xtrans].state = XFER_PREPARED; part 244 drivers/mtd/ftl.c part->XferInfo[xtrans].EraseCount = le32_to_cpu(header.EraseCount); part 246 drivers/mtd/ftl.c part->XferInfo[xtrans].state = XFER_UNKNOWN; part 248 drivers/mtd/ftl.c part->XferInfo[xtrans].EraseCount = part 249 drivers/mtd/ftl.c le32_to_cpu(part->header.EraseCount); part 251 drivers/mtd/ftl.c part->XferInfo[xtrans].Offset = offset; part 256 drivers/mtd/ftl.c header = part->header; part 266 drivers/mtd/ftl.c part->VirtualBlockMap = vmalloc(array_size(blocks, sizeof(uint32_t))); part 267 drivers/mtd/ftl.c if (!part->VirtualBlockMap) part 270 drivers/mtd/ftl.c memset(part->VirtualBlockMap, 0xff, blocks * sizeof(uint32_t)); part 271 drivers/mtd/ftl.c part->BlocksPerUnit = (1 << header.EraseUnitSize) >> header.BlockSize; part 273 drivers/mtd/ftl.c part->bam_cache = kmalloc_array(part->BlocksPerUnit, sizeof(uint32_t), part 275 drivers/mtd/ftl.c if (!part->bam_cache) part 278 drivers/mtd/ftl.c part->bam_index = 0xffff; part 279 drivers/mtd/ftl.c part->FreeTotal = 0; part 281 drivers/mtd/ftl.c for (i = 0; i < part->DataUnits; i++) { part 282 drivers/mtd/ftl.c part->EUNInfo[i].Free = 0; part 283 drivers/mtd/ftl.c part->EUNInfo[i].Deleted = 0; part 284 drivers/mtd/ftl.c offset = part->EUNInfo[i].Offset + le32_to_cpu(header.BAMOffset); part 286 drivers/mtd/ftl.c ret = mtd_read(part->mbd.mtd, offset, part 287 drivers/mtd/ftl.c part->BlocksPerUnit * sizeof(uint32_t), &retval, part 288 drivers/mtd/ftl.c (unsigned char *)part->bam_cache); part 293 drivers/mtd/ftl.c for (j = 0; j < part->BlocksPerUnit; j++) { part 294 drivers/mtd/ftl.c if (BLOCK_FREE(le32_to_cpu(part->bam_cache[j]))) { part 295 drivers/mtd/ftl.c part->EUNInfo[i].Free++; part 296 drivers/mtd/ftl.c part->FreeTotal++; part 297 drivers/mtd/ftl.c } else if ((BLOCK_TYPE(le32_to_cpu(part->bam_cache[j])) == BLOCK_DATA) && part 298 drivers/mtd/ftl.c (BLOCK_NUMBER(le32_to_cpu(part->bam_cache[j])) < blocks)) part 299 drivers/mtd/ftl.c part->VirtualBlockMap[BLOCK_NUMBER(le32_to_cpu(part->bam_cache[j]))] = part 301 drivers/mtd/ftl.c else if (BLOCK_DELETED(le32_to_cpu(part->bam_cache[j]))) part 302 drivers/mtd/ftl.c part->EUNInfo[i].Deleted++; part 310 drivers/mtd/ftl.c kfree(part->bam_cache); part 312 drivers/mtd/ftl.c vfree(part->VirtualBlockMap); part 314 drivers/mtd/ftl.c kfree(part->XferInfo); part 316 drivers/mtd/ftl.c kfree(part->EUNInfo); part 328 drivers/mtd/ftl.c static int erase_xfer(partition_t *part, part 335 drivers/mtd/ftl.c xfer = &part->XferInfo[xfernum]; part 347 drivers/mtd/ftl.c erase->len = 1 << part->header.EraseUnitSize; part 349 drivers/mtd/ftl.c ret = mtd_erase(part->mbd.mtd, erase); part 370 drivers/mtd/ftl.c static int prepare_xfer(partition_t *part, int i) part 379 drivers/mtd/ftl.c xfer = &part->XferInfo[i]; part 385 drivers/mtd/ftl.c header = part->header; part 389 drivers/mtd/ftl.c ret = mtd_write(part->mbd.mtd, xfer->Offset, sizeof(header), &retlen, part 397 drivers/mtd/ftl.c nbam = DIV_ROUND_UP(part->BlocksPerUnit * sizeof(uint32_t) + part 398 drivers/mtd/ftl.c le32_to_cpu(part->header.BAMOffset), SECTOR_SIZE); part 400 drivers/mtd/ftl.c offset = xfer->Offset + le32_to_cpu(part->header.BAMOffset); part 405 drivers/mtd/ftl.c ret = mtd_write(part->mbd.mtd, offset, sizeof(uint32_t), &retlen, part 428 drivers/mtd/ftl.c static int copy_erase_unit(partition_t *part, uint16_t srcunit, part 441 drivers/mtd/ftl.c eun = &part->EUNInfo[srcunit]; part 442 drivers/mtd/ftl.c xfer = &part->XferInfo[xferunit]; part 448 drivers/mtd/ftl.c if (part->bam_index != srcunit) { part 450 drivers/mtd/ftl.c offset = eun->Offset + le32_to_cpu(part->header.BAMOffset); part 452 drivers/mtd/ftl.c ret = mtd_read(part->mbd.mtd, offset, part 453 drivers/mtd/ftl.c part->BlocksPerUnit * sizeof(uint32_t), &retlen, part 454 drivers/mtd/ftl.c (u_char *)(part->bam_cache)); part 457 drivers/mtd/ftl.c part->bam_index = 0xffff; part 470 drivers/mtd/ftl.c ret = mtd_write(part->mbd.mtd, offset, sizeof(uint16_t), &retlen, part 483 drivers/mtd/ftl.c for (i = 0; i < part->BlocksPerUnit; i++) { part 484 drivers/mtd/ftl.c switch (BLOCK_TYPE(le32_to_cpu(part->bam_cache[i]))) { part 490 drivers/mtd/ftl.c ret = mtd_read(part->mbd.mtd, src, SECTOR_SIZE, &retlen, part 498 drivers/mtd/ftl.c ret = mtd_write(part->mbd.mtd, dest, SECTOR_SIZE, &retlen, part 508 drivers/mtd/ftl.c part->bam_cache[i] = cpu_to_le32(0xffffffff); part 517 drivers/mtd/ftl.c ret = mtd_write(part->mbd.mtd, part 518 drivers/mtd/ftl.c xfer->Offset + le32_to_cpu(part->header.BAMOffset), part 519 drivers/mtd/ftl.c part->BlocksPerUnit * sizeof(int32_t), part 521 drivers/mtd/ftl.c (u_char *)part->bam_cache); part 529 drivers/mtd/ftl.c ret = mtd_write(part->mbd.mtd, xfer->Offset + 20, sizeof(uint16_t), part 541 drivers/mtd/ftl.c part->FreeTotal -= eun->Free; part 542 drivers/mtd/ftl.c part->FreeTotal += free; part 547 drivers/mtd/ftl.c part->bam_index = srcunit; part 568 drivers/mtd/ftl.c static int reclaim_block(partition_t *part) part 575 drivers/mtd/ftl.c pr_debug("NumTransferUnits == %x\n", part->header.NumTransferUnits); part 580 drivers/mtd/ftl.c for (i = 0; i < part->header.NumTransferUnits; i++) { part 582 drivers/mtd/ftl.c if (part->XferInfo[i].state == XFER_UNKNOWN) { part 585 drivers/mtd/ftl.c erase_xfer(part, i); part 587 drivers/mtd/ftl.c if (part->XferInfo[i].state == XFER_ERASING) { part 592 drivers/mtd/ftl.c else if (part->XferInfo[i].state == XFER_ERASED) { part 595 drivers/mtd/ftl.c prepare_xfer(part, i); part 597 drivers/mtd/ftl.c if (part->XferInfo[i].state == XFER_PREPARED) { part 600 drivers/mtd/ftl.c if (part->XferInfo[i].EraseCount <= best) { part 601 drivers/mtd/ftl.c best = part->XferInfo[i].EraseCount; part 606 drivers/mtd/ftl.c pr_debug("XferInfo[%d].state == %x\n",i, part->XferInfo[i].state); part 613 drivers/mtd/ftl.c mtd_sync(part->mbd.mtd); part 632 drivers/mtd/ftl.c for (i = 0; i < part->DataUnits; i++) part 633 drivers/mtd/ftl.c if (part->EUNInfo[i].EraseCount <= best) { part 634 drivers/mtd/ftl.c best = part->EUNInfo[i].EraseCount; part 639 drivers/mtd/ftl.c for (i = 0; i < part->DataUnits; i++) part 640 drivers/mtd/ftl.c if (part->EUNInfo[i].Deleted >= best) { part 641 drivers/mtd/ftl.c best = part->EUNInfo[i].Deleted; part 656 drivers/mtd/ftl.c ret = copy_erase_unit(part, eun, xfer); part 658 drivers/mtd/ftl.c erase_xfer(part, xfer); part 675 drivers/mtd/ftl.c static void dump_lists(partition_t *part) part 678 drivers/mtd/ftl.c printk(KERN_DEBUG "ftl_cs: Free total = %d\n", part->FreeTotal); part 679 drivers/mtd/ftl.c for (i = 0; i < part->DataUnits; i++) part 682 drivers/mtd/ftl.c part->EUNInfo[i].Offset >> part->header.EraseUnitSize, part 683 drivers/mtd/ftl.c part->EUNInfo[i].Free, part->EUNInfo[i].Deleted); part 687 drivers/mtd/ftl.c static uint32_t find_free(partition_t *part) part 695 drivers/mtd/ftl.c stop = (part->bam_index == 0xffff) ? 0 : part->bam_index; part 698 drivers/mtd/ftl.c if (part->EUNInfo[eun].Free != 0) break; part 700 drivers/mtd/ftl.c if (++eun == part->DataUnits) eun = 0; part 703 drivers/mtd/ftl.c if (part->EUNInfo[eun].Free == 0) part 707 drivers/mtd/ftl.c if (eun != part->bam_index) { part 709 drivers/mtd/ftl.c part->bam_index = 0xffff; part 711 drivers/mtd/ftl.c ret = mtd_read(part->mbd.mtd, part 712 drivers/mtd/ftl.c part->EUNInfo[eun].Offset + le32_to_cpu(part->header.BAMOffset), part 713 drivers/mtd/ftl.c part->BlocksPerUnit * sizeof(uint32_t), part 715 drivers/mtd/ftl.c (u_char *)(part->bam_cache)); part 721 drivers/mtd/ftl.c part->bam_index = eun; part 725 drivers/mtd/ftl.c for (blk = 0; blk < part->BlocksPerUnit; blk++) part 726 drivers/mtd/ftl.c if (BLOCK_FREE(le32_to_cpu(part->bam_cache[blk]))) break; part 727 drivers/mtd/ftl.c if (blk == part->BlocksPerUnit) { part 731 drivers/mtd/ftl.c dump_lists(part); part 748 drivers/mtd/ftl.c static int ftl_read(partition_t *part, caddr_t buffer, part 757 drivers/mtd/ftl.c part, sector, nblocks); part 758 drivers/mtd/ftl.c if (!(part->state & FTL_FORMATTED)) { part 762 drivers/mtd/ftl.c bsize = 1 << part->header.EraseUnitSize; part 765 drivers/mtd/ftl.c if (((sector+i) * SECTOR_SIZE) >= le32_to_cpu(part->header.FormattedSize)) { part 769 drivers/mtd/ftl.c log_addr = part->VirtualBlockMap[sector+i]; part 773 drivers/mtd/ftl.c offset = (part->EUNInfo[log_addr / bsize].Offset part 775 drivers/mtd/ftl.c ret = mtd_read(part->mbd.mtd, offset, SECTOR_SIZE, &retlen, part 794 drivers/mtd/ftl.c static int set_bam_entry(partition_t *part, uint32_t log_addr, part 806 drivers/mtd/ftl.c part, log_addr, virt_addr); part 807 drivers/mtd/ftl.c bsize = 1 << part->header.EraseUnitSize; part 810 drivers/mtd/ftl.c offset = (part->EUNInfo[eun].Offset + blk * sizeof(uint32_t) + part 811 drivers/mtd/ftl.c le32_to_cpu(part->header.BAMOffset)); part 814 drivers/mtd/ftl.c ret = mtd_read(part->mbd.mtd, offset, sizeof(uint32_t), &retlen, part 835 drivers/mtd/ftl.c if (part->bam_index == eun) { part 837 drivers/mtd/ftl.c if (le32_to_cpu(part->bam_cache[blk]) != old_addr) { part 844 drivers/mtd/ftl.c le32_to_cpu(part->bam_cache[blk]), old_addr); part 849 drivers/mtd/ftl.c part->bam_cache[blk] = le_virt_addr; part 851 drivers/mtd/ftl.c ret = mtd_write(part->mbd.mtd, offset, sizeof(uint32_t), &retlen, part 862 drivers/mtd/ftl.c static int ftl_write(partition_t *part, caddr_t buffer, part 871 drivers/mtd/ftl.c part, sector, nblocks); part 872 drivers/mtd/ftl.c if (!(part->state & FTL_FORMATTED)) { part 877 drivers/mtd/ftl.c while (part->FreeTotal < nblocks) { part 878 drivers/mtd/ftl.c ret = reclaim_block(part); part 883 drivers/mtd/ftl.c bsize = 1 << part->header.EraseUnitSize; part 887 drivers/mtd/ftl.c if (virt_addr >= le32_to_cpu(part->header.FormattedSize)) { part 893 drivers/mtd/ftl.c blk = find_free(part); part 903 drivers/mtd/ftl.c log_addr = part->bam_index * bsize + blk * SECTOR_SIZE; part 904 drivers/mtd/ftl.c part->EUNInfo[part->bam_index].Free--; part 905 drivers/mtd/ftl.c part->FreeTotal--; part 906 drivers/mtd/ftl.c if (set_bam_entry(part, log_addr, 0xfffffffe)) part 908 drivers/mtd/ftl.c part->EUNInfo[part->bam_index].Deleted++; part 909 drivers/mtd/ftl.c offset = (part->EUNInfo[part->bam_index].Offset + part 911 drivers/mtd/ftl.c ret = mtd_write(part->mbd.mtd, offset, SECTOR_SIZE, &retlen, buffer); part 922 drivers/mtd/ftl.c old_addr = part->VirtualBlockMap[sector+i]; part 924 drivers/mtd/ftl.c part->VirtualBlockMap[sector+i] = 0xffffffff; part 925 drivers/mtd/ftl.c part->EUNInfo[old_addr/bsize].Deleted++; part 926 drivers/mtd/ftl.c if (set_bam_entry(part, old_addr, 0)) part 931 drivers/mtd/ftl.c if (set_bam_entry(part, log_addr, virt_addr)) part 933 drivers/mtd/ftl.c part->VirtualBlockMap[sector+i] = log_addr; part 934 drivers/mtd/ftl.c part->EUNInfo[part->bam_index].Deleted--; part 944 drivers/mtd/ftl.c partition_t *part = (void *)dev; part 948 drivers/mtd/ftl.c sect = le32_to_cpu(part->header.FormattedSize)/SECTOR_SIZE; part 972 drivers/mtd/ftl.c partition_t *part = (void *)dev; part 973 drivers/mtd/ftl.c uint32_t bsize = 1 << part->header.EraseUnitSize; part 979 drivers/mtd/ftl.c uint32_t old_addr = part->VirtualBlockMap[sector]; part 981 drivers/mtd/ftl.c part->VirtualBlockMap[sector] = 0xffffffff; part 982 drivers/mtd/ftl.c part->EUNInfo[old_addr/bsize].Deleted++; part 983 drivers/mtd/ftl.c if (set_bam_entry(part, old_addr, 0)) part 994 drivers/mtd/ftl.c static void ftl_freepart(partition_t *part) part 996 drivers/mtd/ftl.c vfree(part->VirtualBlockMap); part 997 drivers/mtd/ftl.c part->VirtualBlockMap = NULL; part 998 drivers/mtd/ftl.c kfree(part->EUNInfo); part 999 drivers/mtd/ftl.c part->EUNInfo = NULL; part 1000 drivers/mtd/ftl.c kfree(part->XferInfo); part 1001 drivers/mtd/ftl.c part->XferInfo = NULL; part 1002 drivers/mtd/ftl.c kfree(part->bam_cache); part 1003 drivers/mtd/ftl.c part->bam_cache = NULL; part 52 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 57 drivers/mtd/mtdpart.c return part_absolute_offset(part->parent) + part->offset; part 68 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 72 drivers/mtd/mtdpart.c stats = part->parent->ecc_stats; part 73 drivers/mtd/mtdpart.c res = part->parent->_read(part->parent, from + part->offset, len, part 77 drivers/mtd/mtdpart.c part->parent->ecc_stats.failed - stats.failed; part 80 drivers/mtd/mtdpart.c part->parent->ecc_stats.corrected - stats.corrected; part 87 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 89 drivers/mtd/mtdpart.c return part->parent->_point(part->parent, from + part->offset, len, part 95 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 97 drivers/mtd/mtdpart.c return part->parent->_unpoint(part->parent, from + part->offset, len); part 103 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 107 drivers/mtd/mtdpart.c stats = part->parent->ecc_stats; part 108 drivers/mtd/mtdpart.c res = part->parent->_read_oob(part->parent, from + part->offset, ops); part 111 drivers/mtd/mtdpart.c part->parent->ecc_stats.failed - stats.failed; part 114 drivers/mtd/mtdpart.c part->parent->ecc_stats.corrected - stats.corrected; part 121 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 122 drivers/mtd/mtdpart.c return part->parent->_read_user_prot_reg(part->parent, from, len, part 129 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 130 drivers/mtd/mtdpart.c return part->parent->_get_user_prot_info(part->parent, len, retlen, part 137 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 138 drivers/mtd/mtdpart.c return part->parent->_read_fact_prot_reg(part->parent, from, len, part 145 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 146 drivers/mtd/mtdpart.c return part->parent->_get_fact_prot_info(part->parent, len, retlen, part 153 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 154 drivers/mtd/mtdpart.c return part->parent->_write(part->parent, to + part->offset, len, part 161 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 162 drivers/mtd/mtdpart.c return part->parent->_panic_write(part->parent, to + part->offset, len, part 169 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 171 drivers/mtd/mtdpart.c return part->parent->_write_oob(part->parent, to + part->offset, ops); part 177 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 178 drivers/mtd/mtdpart.c return part->parent->_write_user_prot_reg(part->parent, from, len, part 185 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 186 drivers/mtd/mtdpart.c return part->parent->_lock_user_prot_reg(part->parent, from, len); part 192 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 193 drivers/mtd/mtdpart.c return part->parent->_writev(part->parent, vecs, count, part 194 drivers/mtd/mtdpart.c to + part->offset, retlen); part 199 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 202 drivers/mtd/mtdpart.c instr->addr += part->offset; part 203 drivers/mtd/mtdpart.c ret = part->parent->_erase(part->parent, instr); part 205 drivers/mtd/mtdpart.c instr->fail_addr -= part->offset; part 206 drivers/mtd/mtdpart.c instr->addr -= part->offset; part 213 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 214 drivers/mtd/mtdpart.c return part->parent->_lock(part->parent, ofs + part->offset, len); part 219 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 220 drivers/mtd/mtdpart.c return part->parent->_unlock(part->parent, ofs + part->offset, len); part 225 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 226 drivers/mtd/mtdpart.c return part->parent->_is_locked(part->parent, ofs + part->offset, len); part 231 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 232 drivers/mtd/mtdpart.c part->parent->_sync(part->parent); part 237 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 238 drivers/mtd/mtdpart.c return part->parent->_suspend(part->parent); part 243 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 244 drivers/mtd/mtdpart.c part->parent->_resume(part->parent); part 249 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 250 drivers/mtd/mtdpart.c ofs += part->offset; part 251 drivers/mtd/mtdpart.c return part->parent->_block_isreserved(part->parent, ofs); part 256 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 257 drivers/mtd/mtdpart.c ofs += part->offset; part 258 drivers/mtd/mtdpart.c return part->parent->_block_isbad(part->parent, ofs); part 263 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 266 drivers/mtd/mtdpart.c ofs += part->offset; part 267 drivers/mtd/mtdpart.c res = part->parent->_block_markbad(part->parent, ofs); part 275 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 276 drivers/mtd/mtdpart.c return part->parent->_get_device(part->parent); part 281 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 282 drivers/mtd/mtdpart.c part->parent->_put_device(part->parent); part 288 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 290 drivers/mtd/mtdpart.c return mtd_ooblayout_ecc(part->parent, section, oobregion); part 296 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 298 drivers/mtd/mtdpart.c return mtd_ooblayout_free(part->parent, section, oobregion); part 308 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 310 drivers/mtd/mtdpart.c return part->parent->_max_bad_blocks(part->parent, part 311 drivers/mtd/mtdpart.c ofs + part->offset, len); part 321 drivers/mtd/mtdpart.c const struct mtd_partition *part, int partno, part 333 drivers/mtd/mtdpart.c name = kstrdup(part->name, GFP_KERNEL); part 344 drivers/mtd/mtdpart.c slave->mtd.flags = parent->orig_flags & ~part->mask_flags; part 346 drivers/mtd/mtdpart.c slave->mtd.size = part->size; part 368 drivers/mtd/mtdpart.c slave->mtd.dev.of_node = part->of_node; part 430 drivers/mtd/mtdpart.c slave->offset = part->offset; part 452 drivers/mtd/mtdpart.c part->name, parent->size - slave->offset, part 474 drivers/mtd/mtdpart.c part->name); part 480 drivers/mtd/mtdpart.c part->name, parent->name, (unsigned long long)slave->mtd.size); part 524 drivers/mtd/mtdpart.c part->name); part 532 drivers/mtd/mtdpart.c part->name); part 560 drivers/mtd/mtdpart.c struct mtd_part *part = mtd_to_part(mtd); part 561 drivers/mtd/mtdpart.c return snprintf(buf, PAGE_SIZE, "%llu\n", part->offset); part 583 drivers/mtd/mtdpart.c struct mtd_partition part; part 598 drivers/mtd/mtdpart.c memset(&part, 0, sizeof(part)); part 599 drivers/mtd/mtdpart.c part.name = name; part 600 drivers/mtd/mtdpart.c part.size = length; part 601 drivers/mtd/mtdpart.c part.offset = offset; part 603 drivers/mtd/mtdpart.c new = allocate_partition(parent, &part, -1, offset); part 1028 drivers/mtd/mtdpart.c struct mtd_part *part; part 1032 drivers/mtd/mtdpart.c list_for_each_entry(part, &mtd_partitions, list) part 1033 drivers/mtd/mtdpart.c if (&part->mtd == mtd) { part 1346 drivers/mtd/mtdswap.c unsigned long part; part 1356 drivers/mtd/mtdswap.c if (kstrtoul(this_opt, 0, &part) < 0) part 1359 drivers/mtd/mtdswap.c if (mtd->index == part) part 1363 drivers/mtd/mtdswap.c if (mtd->index != part) part 1422 drivers/mtd/mtdswap.c MTDSWAP_PREFIX, part, swap_size / 1024, spare_cnt, bad_blocks); part 120 drivers/mtd/parsers/afs.c u_int off, struct mtd_partition *part) part 201 drivers/mtd/parsers/afs.c part->name = kstrdup(iis.name, GFP_KERNEL); part 202 drivers/mtd/parsers/afs.c if (!part->name) part 205 drivers/mtd/parsers/afs.c part->size = (iis.length + mtd->erasesize - 1) & ~(mtd->erasesize - 1); part 206 drivers/mtd/parsers/afs.c part->offset = img_ptr; part 207 drivers/mtd/parsers/afs.c part->mask_flags = 0; part 210 drivers/mtd/parsers/afs.c img_ptr, part->size / 1024, part 211 drivers/mtd/parsers/afs.c iis.imageNumber, part->name); part 217 drivers/mtd/parsers/afs.c u_int off, struct mtd_partition *part) part 312 drivers/mtd/parsers/afs.c part->name = kstrdup(name, GFP_KERNEL); part 313 drivers/mtd/parsers/afs.c if (!part->name) part 315 drivers/mtd/parsers/afs.c part->offset = region_start; part 316 drivers/mtd/parsers/afs.c part->size = region_end - region_start; part 317 drivers/mtd/parsers/afs.c part->mask_flags = 0; part 54 drivers/mtd/parsers/bcm47xxpart.c static void bcm47xxpart_add_part(struct mtd_partition *part, const char *name, part 57 drivers/mtd/parsers/bcm47xxpart.c part->name = name; part 58 drivers/mtd/parsers/bcm47xxpart.c part->offset = offset; part 59 drivers/mtd/parsers/bcm47xxpart.c part->mask_flags = mask_flags; part 302 drivers/mtd/parsers/cmdlinepart.c struct cmdline_mtd_partition *part; part 316 drivers/mtd/parsers/cmdlinepart.c for (part = partitions; part; part = part->next) { part 317 drivers/mtd/parsers/cmdlinepart.c if ((!mtd_id) || (!strcmp(part->mtd_id, mtd_id))) part 321 drivers/mtd/parsers/cmdlinepart.c if (!part) part 324 drivers/mtd/parsers/cmdlinepart.c for (i = 0, offset = 0; i < part->num_parts; i++) { part 325 drivers/mtd/parsers/cmdlinepart.c if (part->parts[i].offset == OFFSET_CONTINUOUS) part 326 drivers/mtd/parsers/cmdlinepart.c part->parts[i].offset = offset; part 328 drivers/mtd/parsers/cmdlinepart.c offset = part->parts[i].offset; part 330 drivers/mtd/parsers/cmdlinepart.c if (part->parts[i].size == SIZE_REMAINING) part 331 drivers/mtd/parsers/cmdlinepart.c part->parts[i].size = master->size - offset; part 333 drivers/mtd/parsers/cmdlinepart.c if (offset + part->parts[i].size > master->size) { part 335 drivers/mtd/parsers/cmdlinepart.c part->mtd_id); part 336 drivers/mtd/parsers/cmdlinepart.c part->parts[i].size = master->size - offset; part 338 drivers/mtd/parsers/cmdlinepart.c offset += part->parts[i].size; part 340 drivers/mtd/parsers/cmdlinepart.c if (part->parts[i].size == 0) { part 342 drivers/mtd/parsers/cmdlinepart.c part->mtd_id); part 343 drivers/mtd/parsers/cmdlinepart.c part->num_parts--; part 344 drivers/mtd/parsers/cmdlinepart.c memmove(&part->parts[i], &part->parts[i + 1], part 345 drivers/mtd/parsers/cmdlinepart.c sizeof(*part->parts) * (part->num_parts - i)); part 350 drivers/mtd/parsers/cmdlinepart.c *pparts = kmemdup(part->parts, sizeof(*part->parts) * part->num_parts, part 355 drivers/mtd/parsers/cmdlinepart.c return part->num_parts; part 160 drivers/mtd/parsers/ofpart.c } *part; part 168 drivers/mtd/parsers/ofpart.c part = of_get_property(dp, "partitions", &plen); part 169 drivers/mtd/parsers/ofpart.c if (!part) part 174 drivers/mtd/parsers/ofpart.c nr_parts = plen / sizeof(part[0]); part 183 drivers/mtd/parsers/ofpart.c parts[i].offset = be32_to_cpu(part->offset); part 184 drivers/mtd/parsers/ofpart.c parts[i].size = be32_to_cpu(part->len) & ~1; part 186 drivers/mtd/parsers/ofpart.c if (be32_to_cpu(part->len) & 1) part 199 drivers/mtd/parsers/ofpart.c part++; part 55 drivers/mtd/parsers/parser_trx.c struct mtd_partition *part; part 80 drivers/mtd/parsers/parser_trx.c part = &parts[curr_part++]; part 81 drivers/mtd/parsers/parser_trx.c part->name = "loader"; part 82 drivers/mtd/parsers/parser_trx.c part->offset = trx.offset[i]; part 87 drivers/mtd/parsers/parser_trx.c part = &parts[curr_part++]; part 88 drivers/mtd/parsers/parser_trx.c part->name = "linux"; part 89 drivers/mtd/parsers/parser_trx.c part->offset = trx.offset[i]; part 94 drivers/mtd/parsers/parser_trx.c part = &parts[curr_part++]; part 95 drivers/mtd/parsers/parser_trx.c part->name = parser_trx_data_part_name(mtd, trx.offset[i]); part 96 drivers/mtd/parsers/parser_trx.c part->offset = trx.offset[i]; part 93 drivers/mtd/rfd_ftl.c static int build_block_map(struct partition *part, int block_no) part 95 drivers/mtd/rfd_ftl.c struct block *block = &part->blocks[block_no]; part 98 drivers/mtd/rfd_ftl.c block->offset = part->block_size * block_no; part 100 drivers/mtd/rfd_ftl.c if (le16_to_cpu(part->header_cache[0]) != RFD_MAGIC) { part 107 drivers/mtd/rfd_ftl.c for (i=0; i<part->data_sectors_per_block; i++) { part 110 drivers/mtd/rfd_ftl.c entry = le16_to_cpu(part->header_cache[HEADER_MAP_OFFSET + i]); part 123 drivers/mtd/rfd_ftl.c if (entry >= part->sector_count) { part 127 drivers/mtd/rfd_ftl.c part->mbd.mtd->name, block_no, i, entry); part 131 drivers/mtd/rfd_ftl.c if (part->sector_map[entry] != -1) { part 134 drivers/mtd/rfd_ftl.c part->mbd.mtd->name, entry); part 135 drivers/mtd/rfd_ftl.c part->errors = 1; part 139 drivers/mtd/rfd_ftl.c part->sector_map[entry] = block->offset + part 140 drivers/mtd/rfd_ftl.c (i + part->header_sectors_per_block) * SECTOR_SIZE; part 145 drivers/mtd/rfd_ftl.c if (block->free_sectors == part->data_sectors_per_block) part 146 drivers/mtd/rfd_ftl.c part->reserved_block = block_no; part 151 drivers/mtd/rfd_ftl.c static int scan_header(struct partition *part) part 158 drivers/mtd/rfd_ftl.c sectors_per_block = part->block_size / SECTOR_SIZE; part 159 drivers/mtd/rfd_ftl.c part->total_blocks = (u32)part->mbd.mtd->size / part->block_size; part 161 drivers/mtd/rfd_ftl.c if (part->total_blocks < 2) part 165 drivers/mtd/rfd_ftl.c part->header_sectors_per_block = part 169 drivers/mtd/rfd_ftl.c part->data_sectors_per_block = sectors_per_block - part 170 drivers/mtd/rfd_ftl.c part->header_sectors_per_block; part 172 drivers/mtd/rfd_ftl.c part->header_size = (HEADER_MAP_OFFSET + part 173 drivers/mtd/rfd_ftl.c part->data_sectors_per_block) * sizeof(u16); part 175 drivers/mtd/rfd_ftl.c part->cylinders = (part->data_sectors_per_block * part 176 drivers/mtd/rfd_ftl.c (part->total_blocks - 1) - 1) / SECTORS_PER_TRACK; part 178 drivers/mtd/rfd_ftl.c part->sector_count = part->cylinders * SECTORS_PER_TRACK; part 180 drivers/mtd/rfd_ftl.c part->current_block = -1; part 181 drivers/mtd/rfd_ftl.c part->reserved_block = -1; part 182 drivers/mtd/rfd_ftl.c part->is_reclaiming = 0; part 184 drivers/mtd/rfd_ftl.c part->header_cache = kmalloc(part->header_size, GFP_KERNEL); part 185 drivers/mtd/rfd_ftl.c if (!part->header_cache) part 188 drivers/mtd/rfd_ftl.c part->blocks = kcalloc(part->total_blocks, sizeof(struct block), part 190 drivers/mtd/rfd_ftl.c if (!part->blocks) part 193 drivers/mtd/rfd_ftl.c part->sector_map = vmalloc(array_size(sizeof(u_long), part 194 drivers/mtd/rfd_ftl.c part->sector_count)); part 195 drivers/mtd/rfd_ftl.c if (!part->sector_map) { part 197 drivers/mtd/rfd_ftl.c "sector map", part->mbd.mtd->name); part 201 drivers/mtd/rfd_ftl.c for (i=0; i<part->sector_count; i++) part 202 drivers/mtd/rfd_ftl.c part->sector_map[i] = -1; part 204 drivers/mtd/rfd_ftl.c for (i=0, blocks_found=0; i<part->total_blocks; i++) { part 205 drivers/mtd/rfd_ftl.c rc = mtd_read(part->mbd.mtd, i * part->block_size, part 206 drivers/mtd/rfd_ftl.c part->header_size, &retlen, part 207 drivers/mtd/rfd_ftl.c (u_char *)part->header_cache); part 209 drivers/mtd/rfd_ftl.c if (!rc && retlen != part->header_size) part 215 drivers/mtd/rfd_ftl.c if (!build_block_map(part, i)) part 221 drivers/mtd/rfd_ftl.c part->mbd.mtd->name); part 226 drivers/mtd/rfd_ftl.c if (part->reserved_block == -1) { part 228 drivers/mtd/rfd_ftl.c part->mbd.mtd->name); part 230 drivers/mtd/rfd_ftl.c part->errors = 1; part 236 drivers/mtd/rfd_ftl.c vfree(part->sector_map); part 237 drivers/mtd/rfd_ftl.c kfree(part->header_cache); part 238 drivers/mtd/rfd_ftl.c kfree(part->blocks); part 245 drivers/mtd/rfd_ftl.c struct partition *part = (struct partition*)dev; part 250 drivers/mtd/rfd_ftl.c if (sector >= part->sector_count) part 253 drivers/mtd/rfd_ftl.c addr = part->sector_map[sector]; part 255 drivers/mtd/rfd_ftl.c rc = mtd_read(part->mbd.mtd, addr, SECTOR_SIZE, &retlen, part 262 drivers/mtd/rfd_ftl.c "0x%lx\n", part->mbd.mtd->name, addr); part 271 drivers/mtd/rfd_ftl.c static int erase_block(struct partition *part, int block) part 280 drivers/mtd/rfd_ftl.c erase->addr = part->blocks[block].offset; part 281 drivers/mtd/rfd_ftl.c erase->len = part->block_size; part 283 drivers/mtd/rfd_ftl.c part->blocks[block].state = BLOCK_ERASING; part 284 drivers/mtd/rfd_ftl.c part->blocks[block].free_sectors = 0; part 286 drivers/mtd/rfd_ftl.c rc = mtd_erase(part->mbd.mtd, erase); part 290 drivers/mtd/rfd_ftl.c (unsigned long long)erase->len, part->mbd.mtd->name); part 291 drivers/mtd/rfd_ftl.c part->blocks[block].state = BLOCK_FAILED; part 292 drivers/mtd/rfd_ftl.c part->blocks[block].free_sectors = 0; part 293 drivers/mtd/rfd_ftl.c part->blocks[block].used_sectors = 0; part 298 drivers/mtd/rfd_ftl.c part->blocks[block].state = BLOCK_ERASED; part 299 drivers/mtd/rfd_ftl.c part->blocks[block].free_sectors = part->data_sectors_per_block; part 300 drivers/mtd/rfd_ftl.c part->blocks[block].used_sectors = 0; part 301 drivers/mtd/rfd_ftl.c part->blocks[block].erases++; part 303 drivers/mtd/rfd_ftl.c rc = mtd_write(part->mbd.mtd, part->blocks[block].offset, part 310 drivers/mtd/rfd_ftl.c part->mbd.mtd->name, part->blocks[block].offset); part 311 drivers/mtd/rfd_ftl.c part->blocks[block].state = BLOCK_FAILED; part 313 drivers/mtd/rfd_ftl.c part->blocks[block].state = BLOCK_OK; part 322 drivers/mtd/rfd_ftl.c static int move_block_contents(struct partition *part, int block_no, u_long *old_sector) part 329 drivers/mtd/rfd_ftl.c part->is_reclaiming = 1; part 335 drivers/mtd/rfd_ftl.c map = kmalloc(part->header_size, GFP_KERNEL); part 339 drivers/mtd/rfd_ftl.c rc = mtd_read(part->mbd.mtd, part->blocks[block_no].offset, part 340 drivers/mtd/rfd_ftl.c part->header_size, &retlen, (u_char *)map); part 342 drivers/mtd/rfd_ftl.c if (!rc && retlen != part->header_size) part 347 drivers/mtd/rfd_ftl.c "0x%lx\n", part->mbd.mtd->name, part 348 drivers/mtd/rfd_ftl.c part->blocks[block_no].offset); part 353 drivers/mtd/rfd_ftl.c for (i=0; i<part->data_sectors_per_block; i++) { part 365 drivers/mtd/rfd_ftl.c if (entry >= part->sector_count) part 368 drivers/mtd/rfd_ftl.c addr = part->blocks[block_no].offset + part 369 drivers/mtd/rfd_ftl.c (i + part->header_sectors_per_block) * SECTOR_SIZE; part 373 drivers/mtd/rfd_ftl.c if (!part->blocks[block_no].used_sectors--) { part 374 drivers/mtd/rfd_ftl.c rc = erase_block(part, block_no); part 379 drivers/mtd/rfd_ftl.c rc = mtd_read(part->mbd.mtd, addr, SECTOR_SIZE, &retlen, part 388 drivers/mtd/rfd_ftl.c part->mbd.mtd->name); part 393 drivers/mtd/rfd_ftl.c rc = rfd_ftl_writesect((struct mtd_blktrans_dev*)part, part 405 drivers/mtd/rfd_ftl.c part->is_reclaiming = 0; part 410 drivers/mtd/rfd_ftl.c static int reclaim_block(struct partition *part, u_long *old_sector) part 416 drivers/mtd/rfd_ftl.c mtd_sync(part->mbd.mtd); part 421 drivers/mtd/rfd_ftl.c old_sector_block = *old_sector / part->block_size; part 425 drivers/mtd/rfd_ftl.c for (block=0; block<part->total_blocks; block++) { part 428 drivers/mtd/rfd_ftl.c if (block == part->reserved_block) part 436 drivers/mtd/rfd_ftl.c if (part->blocks[block].free_sectors) part 439 drivers/mtd/rfd_ftl.c this_score = part->blocks[block].used_sectors; part 445 drivers/mtd/rfd_ftl.c if (part->blocks[block].used_sectors == part 446 drivers/mtd/rfd_ftl.c part->data_sectors_per_block) part 450 drivers/mtd/rfd_ftl.c this_score += part->blocks[block].erases; part 461 drivers/mtd/rfd_ftl.c part->current_block = -1; part 462 drivers/mtd/rfd_ftl.c part->reserved_block = best_block; part 466 drivers/mtd/rfd_ftl.c part->blocks[best_block].used_sectors, part 467 drivers/mtd/rfd_ftl.c part->blocks[best_block].free_sectors); part 469 drivers/mtd/rfd_ftl.c if (part->blocks[best_block].used_sectors) part 470 drivers/mtd/rfd_ftl.c rc = move_block_contents(part, best_block, old_sector); part 472 drivers/mtd/rfd_ftl.c rc = erase_block(part, best_block); part 482 drivers/mtd/rfd_ftl.c static int find_free_block(struct partition *part) part 486 drivers/mtd/rfd_ftl.c block = part->current_block == -1 ? part 487 drivers/mtd/rfd_ftl.c jiffies % part->total_blocks : part->current_block; part 491 drivers/mtd/rfd_ftl.c if (part->blocks[block].free_sectors && part 492 drivers/mtd/rfd_ftl.c block != part->reserved_block) part 495 drivers/mtd/rfd_ftl.c if (part->blocks[block].state == BLOCK_UNUSED) part 496 drivers/mtd/rfd_ftl.c erase_block(part, block); part 498 drivers/mtd/rfd_ftl.c if (++block >= part->total_blocks) part 506 drivers/mtd/rfd_ftl.c static int find_writable_block(struct partition *part, u_long *old_sector) part 511 drivers/mtd/rfd_ftl.c block = find_free_block(part); part 514 drivers/mtd/rfd_ftl.c if (!part->is_reclaiming) { part 515 drivers/mtd/rfd_ftl.c rc = reclaim_block(part, old_sector); part 519 drivers/mtd/rfd_ftl.c block = find_free_block(part); part 528 drivers/mtd/rfd_ftl.c rc = mtd_read(part->mbd.mtd, part->blocks[block].offset, part 529 drivers/mtd/rfd_ftl.c part->header_size, &retlen, part 530 drivers/mtd/rfd_ftl.c (u_char *)part->header_cache); part 532 drivers/mtd/rfd_ftl.c if (!rc && retlen != part->header_size) part 537 drivers/mtd/rfd_ftl.c "0x%lx\n", part->mbd.mtd->name, part 538 drivers/mtd/rfd_ftl.c part->blocks[block].offset); part 542 drivers/mtd/rfd_ftl.c part->current_block = block; part 548 drivers/mtd/rfd_ftl.c static int mark_sector_deleted(struct partition *part, u_long old_addr) part 555 drivers/mtd/rfd_ftl.c block = old_addr / part->block_size; part 556 drivers/mtd/rfd_ftl.c offset = (old_addr % part->block_size) / SECTOR_SIZE - part 557 drivers/mtd/rfd_ftl.c part->header_sectors_per_block; part 559 drivers/mtd/rfd_ftl.c addr = part->blocks[block].offset + part 561 drivers/mtd/rfd_ftl.c rc = mtd_write(part->mbd.mtd, addr, sizeof(del), &retlen, part 569 drivers/mtd/rfd_ftl.c "0x%lx\n", part->mbd.mtd->name, addr); part 572 drivers/mtd/rfd_ftl.c if (block == part->current_block) part 573 drivers/mtd/rfd_ftl.c part->header_cache[offset + HEADER_MAP_OFFSET] = del; part 575 drivers/mtd/rfd_ftl.c part->blocks[block].used_sectors--; part 577 drivers/mtd/rfd_ftl.c if (!part->blocks[block].used_sectors && part 578 drivers/mtd/rfd_ftl.c !part->blocks[block].free_sectors) part 579 drivers/mtd/rfd_ftl.c rc = erase_block(part, block); part 585 drivers/mtd/rfd_ftl.c static int find_free_sector(const struct partition *part, const struct block *block) part 589 drivers/mtd/rfd_ftl.c i = stop = part->data_sectors_per_block - block->free_sectors; part 592 drivers/mtd/rfd_ftl.c if (le16_to_cpu(part->header_cache[HEADER_MAP_OFFSET + i]) part 596 drivers/mtd/rfd_ftl.c if (++i == part->data_sectors_per_block) part 606 drivers/mtd/rfd_ftl.c struct partition *part = (struct partition*)dev; part 614 drivers/mtd/rfd_ftl.c if (part->current_block == -1 || part 615 drivers/mtd/rfd_ftl.c !part->blocks[part->current_block].free_sectors) { part 617 drivers/mtd/rfd_ftl.c rc = find_writable_block(part, old_addr); part 622 drivers/mtd/rfd_ftl.c block = &part->blocks[part->current_block]; part 624 drivers/mtd/rfd_ftl.c i = find_free_sector(part, block); part 631 drivers/mtd/rfd_ftl.c addr = (i + part->header_sectors_per_block) * SECTOR_SIZE + part 633 drivers/mtd/rfd_ftl.c rc = mtd_write(part->mbd.mtd, addr, SECTOR_SIZE, &retlen, part 641 drivers/mtd/rfd_ftl.c part->mbd.mtd->name, addr); part 645 drivers/mtd/rfd_ftl.c part->sector_map[sector] = addr; part 649 drivers/mtd/rfd_ftl.c part->header_cache[i + HEADER_MAP_OFFSET] = entry; part 652 drivers/mtd/rfd_ftl.c rc = mtd_write(part->mbd.mtd, addr, sizeof(entry), &retlen, part 660 drivers/mtd/rfd_ftl.c part->mbd.mtd->name, addr); part 672 drivers/mtd/rfd_ftl.c struct partition *part = (struct partition*)dev; part 679 drivers/mtd/rfd_ftl.c if (part->reserved_block == -1) { part 684 drivers/mtd/rfd_ftl.c if (sector >= part->sector_count) { part 689 drivers/mtd/rfd_ftl.c old_addr = part->sector_map[sector]; part 702 drivers/mtd/rfd_ftl.c part->sector_map[sector] = -1; part 705 drivers/mtd/rfd_ftl.c rc = mark_sector_deleted(part, old_addr); part 713 drivers/mtd/rfd_ftl.c struct partition *part = (struct partition*)dev; part 717 drivers/mtd/rfd_ftl.c geo->cylinders = part->cylinders; part 724 drivers/mtd/rfd_ftl.c struct partition *part; part 729 drivers/mtd/rfd_ftl.c part = kzalloc(sizeof(struct partition), GFP_KERNEL); part 730 drivers/mtd/rfd_ftl.c if (!part) part 733 drivers/mtd/rfd_ftl.c part->mbd.mtd = mtd; part 736 drivers/mtd/rfd_ftl.c part->block_size = block_size; part 742 drivers/mtd/rfd_ftl.c part->block_size = mtd->erasesize; part 745 drivers/mtd/rfd_ftl.c if (scan_header(part) == 0) { part 746 drivers/mtd/rfd_ftl.c part->mbd.size = part->sector_count; part 747 drivers/mtd/rfd_ftl.c part->mbd.tr = tr; part 748 drivers/mtd/rfd_ftl.c part->mbd.devnum = -1; part 750 drivers/mtd/rfd_ftl.c part->mbd.readonly = 1; part 751 drivers/mtd/rfd_ftl.c else if (part->errors) { part 754 drivers/mtd/rfd_ftl.c part->mbd.readonly = 1; part 760 drivers/mtd/rfd_ftl.c if (!add_mtd_blktrans_dev((void*)part)) part 764 drivers/mtd/rfd_ftl.c kfree(part); part 769 drivers/mtd/rfd_ftl.c struct partition *part = (struct partition*)dev; part 772 drivers/mtd/rfd_ftl.c for (i=0; i<part->total_blocks; i++) { part 774 drivers/mtd/rfd_ftl.c part->mbd.mtd->name, i, part->blocks[i].erases); part 778 drivers/mtd/rfd_ftl.c vfree(part->sector_map); part 779 drivers/mtd/rfd_ftl.c kfree(part->header_cache); part 780 drivers/mtd/rfd_ftl.c kfree(part->blocks); part 829 drivers/mtd/spi-nor/intel-spi.c struct mtd_partition *part) part 834 drivers/mtd/spi-nor/intel-spi.c memset(part, 0, sizeof(*part)); part 837 drivers/mtd/spi-nor/intel-spi.c part->size = 4096; part 838 drivers/mtd/spi-nor/intel-spi.c part->name = "BIOS"; part 862 drivers/mtd/spi-nor/intel-spi.c if (end > part->size) part 863 drivers/mtd/spi-nor/intel-spi.c part->size = end; part 875 drivers/mtd/spi-nor/intel-spi.c struct mtd_partition part; part 912 drivers/mtd/spi-nor/intel-spi.c intel_spi_fill_partition(ispi, &part); part 918 drivers/mtd/spi-nor/intel-spi.c ret = mtd_device_register(&ispi->nor.mtd, &part, 1); part 401 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c u32 part; part 415 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c part = (speed * 100) / line_speed; part 417 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c if (line_speed < speed || !part) { part 425 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c bp->pending_max = part; part 427 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c bnx2x_update_max_mf_config(bp, part); part 280 drivers/net/ethernet/brocade/bna/bfa_defs.h struct bfa_flash_part_attr part[BFA_FLASH_PART_MAX]; part 3146 drivers/net/ethernet/brocade/bna/bfa_ioc.c attr->part[i].part_type = part 3147 drivers/net/ethernet/brocade/bna/bfa_ioc.c be32_to_cpu(f->part[i].part_type); part 3148 drivers/net/ethernet/brocade/bna/bfa_ioc.c attr->part[i].part_instance = part 3149 drivers/net/ethernet/brocade/bna/bfa_ioc.c be32_to_cpu(f->part[i].part_instance); part 3150 drivers/net/ethernet/brocade/bna/bfa_ioc.c attr->part[i].part_off = part 3151 drivers/net/ethernet/brocade/bna/bfa_ioc.c be32_to_cpu(f->part[i].part_off); part 3152 drivers/net/ethernet/brocade/bna/bfa_ioc.c attr->part[i].part_size = part 3153 drivers/net/ethernet/brocade/bna/bfa_ioc.c be32_to_cpu(f->part[i].part_size); part 3154 drivers/net/ethernet/brocade/bna/bfa_ioc.c attr->part[i].part_len = part 3155 drivers/net/ethernet/brocade/bna/bfa_ioc.c be32_to_cpu(f->part[i].part_len); part 3156 drivers/net/ethernet/brocade/bna/bfa_ioc.c attr->part[i].part_status = part 3157 drivers/net/ethernet/brocade/bna/bfa_ioc.c be32_to_cpu(f->part[i].part_status); part 975 drivers/net/ethernet/brocade/bna/bnad_ethtool.c if (offset >= flash_attr->part[i].part_off && part 976 drivers/net/ethernet/brocade/bna/bnad_ethtool.c offset < (flash_attr->part[i].part_off + part 977 drivers/net/ethernet/brocade/bna/bnad_ethtool.c flash_attr->part[i].part_size)) { part 978 drivers/net/ethernet/brocade/bna/bnad_ethtool.c flash_part = flash_attr->part[i].part_type; part 979 drivers/net/ethernet/brocade/bna/bnad_ethtool.c *base_offset = flash_attr->part[i].part_off; part 8953 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c unsigned int part, manufacturer; part 8973 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c for (part = 0; part < ARRAY_SIZE(supported_flash); part++) part 8974 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c if (supported_flash[part].vendor_and_model_id == flashid) { part 8975 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c adap->params.sf_size = supported_flash[part].size_mb; part 74 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c struct mlxsw_sp1_kvdl_part *part, *min_part = NULL; part 78 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c part = kvdl->parts[i]; part 79 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c if (alloc_size <= part->info.alloc_size && part 81 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c part->info.alloc_size <= min_part->info.alloc_size)) part 82 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c min_part = part; part 91 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c struct mlxsw_sp1_kvdl_part *part; part 95 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c part = kvdl->parts[i]; part 96 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c if (kvdl_index >= part->info.start_index && part 97 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c kvdl_index <= part->info.end_index) part 98 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c return part; part 118 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c static int mlxsw_sp1_kvdl_part_alloc(struct mlxsw_sp1_kvdl_part *part, part 121 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c const struct mlxsw_sp1_kvdl_part_info *info = &part->info; part 126 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c entry_index = find_first_zero_bit(part->usage, nr_entries); part 129 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c __set_bit(entry_index, part->usage); part 136 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c static void mlxsw_sp1_kvdl_part_free(struct mlxsw_sp1_kvdl_part *part, part 139 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c const struct mlxsw_sp1_kvdl_part_info *info = &part->info; part 143 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c __clear_bit(entry_index, part->usage); part 152 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c struct mlxsw_sp1_kvdl_part *part; part 157 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c part = mlxsw_sp1_kvdl_alloc_size_part(kvdl, entry_count); part 158 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c if (IS_ERR(part)) part 159 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c return PTR_ERR(part); part 161 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c return mlxsw_sp1_kvdl_part_alloc(part, p_entry_index); part 169 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c struct mlxsw_sp1_kvdl_part *part; part 171 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c part = mlxsw_sp1_kvdl_index_part(kvdl, entry_index); part 172 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c if (IS_ERR(part)) part 174 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c mlxsw_sp1_kvdl_part_free(part, entry_index); part 184 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c struct mlxsw_sp1_kvdl_part *part; part 186 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c part = mlxsw_sp1_kvdl_alloc_size_part(kvdl, entry_count); part 187 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c if (IS_ERR(part)) part 188 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c return PTR_ERR(part); part 190 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c *p_alloc_size = part->info.alloc_size; part 195 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c static void mlxsw_sp1_kvdl_part_update(struct mlxsw_sp1_kvdl_part *part, part 200 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c part->info.end_index = size - 1; part 202 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c part->info.start_index = part_prev->info.end_index + 1; part 203 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c part->info.end_index = part->info.start_index + size - 1; part 213 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c struct mlxsw_sp1_kvdl_part *part; part 229 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c part = kzalloc(sizeof(*part) + usage_size, GFP_KERNEL); part 230 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c if (!part) part 233 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c memcpy(&part->info, info, sizeof(part->info)); part 236 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c mlxsw_sp1_kvdl_part_update(part, part_prev, resource_size); part 237 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c return part; part 240 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c static void mlxsw_sp1_kvdl_part_fini(struct mlxsw_sp1_kvdl_part *part) part 242 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c kfree(part); part 278 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c static u64 mlxsw_sp1_kvdl_part_occ(struct mlxsw_sp1_kvdl_part *part) part 280 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c const struct mlxsw_sp1_kvdl_part_info *info = &part->info; part 288 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c while ((bit = find_next_bit(part->usage, nr_entries, bit + 1)) part 309 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c struct mlxsw_sp1_kvdl_part *part; part 311 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c part = kvdl->parts[MLXSW_SP1_KVDL_PART_ID_SINGLE]; part 312 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c return mlxsw_sp1_kvdl_part_occ(part); part 318 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c struct mlxsw_sp1_kvdl_part *part; part 320 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c part = kvdl->parts[MLXSW_SP1_KVDL_PART_ID_CHUNKS]; part 321 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c return mlxsw_sp1_kvdl_part_occ(part); part 327 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c struct mlxsw_sp1_kvdl_part *part; part 329 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c part = kvdl->parts[MLXSW_SP1_KVDL_PART_ID_LARGE_CHUNKS]; part 330 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c return mlxsw_sp1_kvdl_part_occ(part); part 55 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c static int mlxsw_sp2_kvdl_part_find_zero_bits(struct mlxsw_sp2_kvdl_part *part, part 64 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c start_bit = part->last_allocated_bit + 1; part 65 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c if (start_bit == part->usage_bit_count) part 69 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c bit = find_next_zero_bit(part->usage, part->usage_bit_count, bit); part 70 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c if (!wrap && bit + bit_count >= part->usage_bit_count) { part 78 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c if (test_bit(bit + i, part->usage)) { part 87 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c static int mlxsw_sp2_kvdl_part_alloc(struct mlxsw_sp2_kvdl_part *part, part 96 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c bit_count = DIV_ROUND_UP(size, part->indexes_per_usage_bit); part 97 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c err = mlxsw_sp2_kvdl_part_find_zero_bits(part, bit_count, &bit); part 101 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c __set_bit(bit + i, part->usage); part 102 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c *p_kvdl_index = bit * part->indexes_per_usage_bit; part 124 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c struct mlxsw_sp2_kvdl_part *part, part 133 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c err = mlxsw_sp2_kvdl_rec_del(mlxsw_sp, part->info->res_type, part 138 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c bit_count = DIV_ROUND_UP(size, part->indexes_per_usage_bit); part 139 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c bit = kvdl_index / part->indexes_per_usage_bit; part 141 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c __clear_bit(bit + i, part->usage); part 151 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c struct mlxsw_sp2_kvdl_part *part = kvdl->parts[type]; part 153 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c return mlxsw_sp2_kvdl_part_alloc(part, size, p_entry_index); part 163 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c struct mlxsw_sp2_kvdl_part *part = kvdl->parts[type]; part 165 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c return mlxsw_sp2_kvdl_part_free(mlxsw_sp, part, size, entry_index); part 183 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c struct mlxsw_sp2_kvdl_part *part; part 205 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c part = kzalloc(sizeof(*part) + usage_size, GFP_KERNEL); part 206 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c if (!part) part 208 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c part->info = info; part 209 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c part->usage_bit_count = usage_bit_count; part 210 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c part->indexes_per_usage_bit = indexes_per_usage_bit; part 211 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c part->last_allocated_bit = usage_bit_count - 1; part 212 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c return part; part 215 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c static void mlxsw_sp2_kvdl_part_fini(struct mlxsw_sp2_kvdl_part *part) part 217 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c kfree(part); part 262 drivers/net/ethernet/netronome/nfp/nfp_devlink.c const char *sn, *vendor, *part; part 272 drivers/net/ethernet/netronome/nfp/nfp_devlink.c part = nfp_hwinfo_lookup(pf->hwinfo, "assembly.partno"); part 274 drivers/net/ethernet/netronome/nfp/nfp_devlink.c if (vendor && part && sn) { part 277 drivers/net/ethernet/netronome/nfp/nfp_devlink.c buf = kmalloc(strlen(vendor) + strlen(part) + strlen(sn) + 1, part 284 drivers/net/ethernet/netronome/nfp/nfp_devlink.c strcat(buf, part); part 6052 drivers/net/ethernet/sfc/ef10.c struct efx_mcdi_mtd_partition *part, part 6094 drivers/net/ethernet/sfc/ef10.c part->nvram_type = type; part 6105 drivers/net/ethernet/sfc/ef10.c part->fw_subtype = MCDI_DWORD(outbuf, part 6108 drivers/net/ethernet/sfc/ef10.c part->common.dev_type_name = "EF10 NVRAM manager"; part 6109 drivers/net/ethernet/sfc/ef10.c part->common.type_name = info->name; part 6111 drivers/net/ethernet/sfc/ef10.c part->common.mtd.type = MTD_NORFLASH; part 6112 drivers/net/ethernet/sfc/ef10.c part->common.mtd.flags = MTD_CAP_NORFLASH; part 6113 drivers/net/ethernet/sfc/ef10.c part->common.mtd.size = size; part 6114 drivers/net/ethernet/sfc/ef10.c part->common.mtd.erasesize = erase_size; part 6117 drivers/net/ethernet/sfc/ef10.c part->common.mtd.flags |= MTD_NO_ERASE; part 740 drivers/net/ethernet/sfc/falcon/falcon.c falcon_spi_slow_wait(struct falcon_mtd_partition *part, bool uninterruptible) part 742 drivers/net/ethernet/sfc/falcon/falcon.c const struct falcon_spi_device *spi = part->spi; part 743 drivers/net/ethernet/sfc/falcon/falcon.c struct ef4_nic *efx = part->common.mtd.priv; part 762 drivers/net/ethernet/sfc/falcon/falcon.c part->common.name, part->common.dev_type_name); part 804 drivers/net/ethernet/sfc/falcon/falcon.c falcon_spi_erase(struct falcon_mtd_partition *part, loff_t start, size_t len) part 806 drivers/net/ethernet/sfc/falcon/falcon.c const struct falcon_spi_device *spi = part->spi; part 807 drivers/net/ethernet/sfc/falcon/falcon.c struct ef4_nic *efx = part->common.mtd.priv; part 829 drivers/net/ethernet/sfc/falcon/falcon.c rc = falcon_spi_slow_wait(part, false); part 851 drivers/net/ethernet/sfc/falcon/falcon.c static void falcon_mtd_rename(struct ef4_mtd_partition *part) part 853 drivers/net/ethernet/sfc/falcon/falcon.c struct ef4_nic *efx = part->mtd.priv; part 855 drivers/net/ethernet/sfc/falcon/falcon.c snprintf(part->name, sizeof(part->name), "%s %s", part 856 drivers/net/ethernet/sfc/falcon/falcon.c efx->name, part->type_name); part 862 drivers/net/ethernet/sfc/falcon/falcon.c struct falcon_mtd_partition *part = to_falcon_mtd_partition(mtd); part 870 drivers/net/ethernet/sfc/falcon/falcon.c rc = falcon_spi_read(efx, part->spi, part->offset + start, part 878 drivers/net/ethernet/sfc/falcon/falcon.c struct falcon_mtd_partition *part = to_falcon_mtd_partition(mtd); part 886 drivers/net/ethernet/sfc/falcon/falcon.c rc = falcon_spi_erase(part, part->offset + start, len); part 894 drivers/net/ethernet/sfc/falcon/falcon.c struct falcon_mtd_partition *part = to_falcon_mtd_partition(mtd); part 902 drivers/net/ethernet/sfc/falcon/falcon.c rc = falcon_spi_write(efx, part->spi, part->offset + start, part 910 drivers/net/ethernet/sfc/falcon/falcon.c struct falcon_mtd_partition *part = to_falcon_mtd_partition(mtd); part 916 drivers/net/ethernet/sfc/falcon/falcon.c rc = falcon_spi_slow_wait(part, true); part 30 drivers/net/ethernet/sfc/falcon/mtd.c struct ef4_mtd_partition *part = to_ef4_mtd_partition(mtd); part 37 drivers/net/ethernet/sfc/falcon/mtd.c part->name, part->dev_type_name, rc); part 40 drivers/net/ethernet/sfc/falcon/mtd.c static void ef4_mtd_remove_partition(struct ef4_mtd_partition *part) part 45 drivers/net/ethernet/sfc/falcon/mtd.c rc = mtd_device_unregister(&part->mtd); part 51 drivers/net/ethernet/sfc/falcon/mtd.c list_del(&part->node); part 57 drivers/net/ethernet/sfc/falcon/mtd.c struct ef4_mtd_partition *part; part 61 drivers/net/ethernet/sfc/falcon/mtd.c part = (struct ef4_mtd_partition *)((char *)parts + part 64 drivers/net/ethernet/sfc/falcon/mtd.c part->mtd.writesize = 1; part 66 drivers/net/ethernet/sfc/falcon/mtd.c part->mtd.owner = THIS_MODULE; part 67 drivers/net/ethernet/sfc/falcon/mtd.c part->mtd.priv = efx; part 68 drivers/net/ethernet/sfc/falcon/mtd.c part->mtd.name = part->name; part 69 drivers/net/ethernet/sfc/falcon/mtd.c part->mtd._erase = ef4_mtd_erase; part 70 drivers/net/ethernet/sfc/falcon/mtd.c part->mtd._read = efx->type->mtd_read; part 71 drivers/net/ethernet/sfc/falcon/mtd.c part->mtd._write = efx->type->mtd_write; part 72 drivers/net/ethernet/sfc/falcon/mtd.c part->mtd._sync = ef4_mtd_sync; part 74 drivers/net/ethernet/sfc/falcon/mtd.c efx->type->mtd_rename(part); part 76 drivers/net/ethernet/sfc/falcon/mtd.c if (mtd_device_register(&part->mtd, NULL, 0)) part 80 drivers/net/ethernet/sfc/falcon/mtd.c list_add_tail(&part->node, &efx->mtd_list); part 87 drivers/net/ethernet/sfc/falcon/mtd.c part = (struct ef4_mtd_partition *)((char *)parts + part 89 drivers/net/ethernet/sfc/falcon/mtd.c ef4_mtd_remove_partition(part); part 97 drivers/net/ethernet/sfc/falcon/mtd.c struct ef4_mtd_partition *parts, *part, *next; part 107 drivers/net/ethernet/sfc/falcon/mtd.c list_for_each_entry_safe(part, next, &efx->mtd_list, node) part 108 drivers/net/ethernet/sfc/falcon/mtd.c ef4_mtd_remove_partition(part); part 115 drivers/net/ethernet/sfc/falcon/mtd.c struct ef4_mtd_partition *part; part 119 drivers/net/ethernet/sfc/falcon/mtd.c list_for_each_entry(part, &efx->mtd_list, node) part 120 drivers/net/ethernet/sfc/falcon/mtd.c efx->type->mtd_rename(part); part 1130 drivers/net/ethernet/sfc/falcon/net_driver.h void (*mtd_rename)(struct ef4_mtd_partition *part); part 2204 drivers/net/ethernet/sfc/mcdi.c struct efx_mcdi_mtd_partition *part = to_efx_mcdi_mtd_partition(mtd); part 2213 drivers/net/ethernet/sfc/mcdi.c rc = efx_mcdi_nvram_read(efx, part->nvram_type, offset, part 2227 drivers/net/ethernet/sfc/mcdi.c struct efx_mcdi_mtd_partition *part = to_efx_mcdi_mtd_partition(mtd); part 2231 drivers/net/ethernet/sfc/mcdi.c size_t chunk = part->common.mtd.erasesize; part 2234 drivers/net/ethernet/sfc/mcdi.c if (!part->updating) { part 2235 drivers/net/ethernet/sfc/mcdi.c rc = efx_mcdi_nvram_update_start(efx, part->nvram_type); part 2238 drivers/net/ethernet/sfc/mcdi.c part->updating = true; part 2245 drivers/net/ethernet/sfc/mcdi.c rc = efx_mcdi_nvram_erase(efx, part->nvram_type, offset, part 2258 drivers/net/ethernet/sfc/mcdi.c struct efx_mcdi_mtd_partition *part = to_efx_mcdi_mtd_partition(mtd); part 2265 drivers/net/ethernet/sfc/mcdi.c if (!part->updating) { part 2266 drivers/net/ethernet/sfc/mcdi.c rc = efx_mcdi_nvram_update_start(efx, part->nvram_type); part 2269 drivers/net/ethernet/sfc/mcdi.c part->updating = true; part 2274 drivers/net/ethernet/sfc/mcdi.c rc = efx_mcdi_nvram_write(efx, part->nvram_type, offset, part 2288 drivers/net/ethernet/sfc/mcdi.c struct efx_mcdi_mtd_partition *part = to_efx_mcdi_mtd_partition(mtd); part 2292 drivers/net/ethernet/sfc/mcdi.c if (part->updating) { part 2293 drivers/net/ethernet/sfc/mcdi.c part->updating = false; part 2294 drivers/net/ethernet/sfc/mcdi.c rc = efx_mcdi_nvram_update_finish(efx, part->nvram_type); part 2300 drivers/net/ethernet/sfc/mcdi.c void efx_mcdi_mtd_rename(struct efx_mtd_partition *part) part 2303 drivers/net/ethernet/sfc/mcdi.c container_of(part, struct efx_mcdi_mtd_partition, common); part 2304 drivers/net/ethernet/sfc/mcdi.c struct efx_nic *efx = part->mtd.priv; part 2306 drivers/net/ethernet/sfc/mcdi.c snprintf(part->name, sizeof(part->name), "%s %s:%02x", part 2307 drivers/net/ethernet/sfc/mcdi.c efx->name, part->type_name, mcdi_part->fw_subtype); part 380 drivers/net/ethernet/sfc/mcdi.h void efx_mcdi_mtd_rename(struct efx_mtd_partition *part); part 30 drivers/net/ethernet/sfc/mtd.c struct efx_mtd_partition *part = to_efx_mtd_partition(mtd); part 37 drivers/net/ethernet/sfc/mtd.c part->name, part->dev_type_name, rc); part 40 drivers/net/ethernet/sfc/mtd.c static void efx_mtd_remove_partition(struct efx_mtd_partition *part) part 45 drivers/net/ethernet/sfc/mtd.c rc = mtd_device_unregister(&part->mtd); part 51 drivers/net/ethernet/sfc/mtd.c list_del(&part->node); part 57 drivers/net/ethernet/sfc/mtd.c struct efx_mtd_partition *part; part 61 drivers/net/ethernet/sfc/mtd.c part = (struct efx_mtd_partition *)((char *)parts + part 64 drivers/net/ethernet/sfc/mtd.c part->mtd.writesize = 1; part 66 drivers/net/ethernet/sfc/mtd.c if (!(part->mtd.flags & MTD_NO_ERASE)) part 67 drivers/net/ethernet/sfc/mtd.c part->mtd.flags |= MTD_WRITEABLE; part 69 drivers/net/ethernet/sfc/mtd.c part->mtd.owner = THIS_MODULE; part 70 drivers/net/ethernet/sfc/mtd.c part->mtd.priv = efx; part 71 drivers/net/ethernet/sfc/mtd.c part->mtd.name = part->name; part 72 drivers/net/ethernet/sfc/mtd.c part->mtd._erase = efx_mtd_erase; part 73 drivers/net/ethernet/sfc/mtd.c part->mtd._read = efx->type->mtd_read; part 74 drivers/net/ethernet/sfc/mtd.c part->mtd._write = efx->type->mtd_write; part 75 drivers/net/ethernet/sfc/mtd.c part->mtd._sync = efx_mtd_sync; part 77 drivers/net/ethernet/sfc/mtd.c efx->type->mtd_rename(part); part 79 drivers/net/ethernet/sfc/mtd.c if (mtd_device_register(&part->mtd, NULL, 0)) part 83 drivers/net/ethernet/sfc/mtd.c list_add_tail(&part->node, &efx->mtd_list); part 90 drivers/net/ethernet/sfc/mtd.c part = (struct efx_mtd_partition *)((char *)parts + part 92 drivers/net/ethernet/sfc/mtd.c efx_mtd_remove_partition(part); part 100 drivers/net/ethernet/sfc/mtd.c struct efx_mtd_partition *parts, *part, *next; part 110 drivers/net/ethernet/sfc/mtd.c list_for_each_entry_safe(part, next, &efx->mtd_list, node) part 111 drivers/net/ethernet/sfc/mtd.c efx_mtd_remove_partition(part); part 118 drivers/net/ethernet/sfc/mtd.c struct efx_mtd_partition *part; part 122 drivers/net/ethernet/sfc/mtd.c list_for_each_entry(part, &efx->mtd_list, node) part 123 drivers/net/ethernet/sfc/mtd.c efx->type->mtd_rename(part); part 1375 drivers/net/ethernet/sfc/net_driver.h void (*mtd_rename)(struct efx_mtd_partition *part); part 848 drivers/net/ethernet/sfc/siena.c struct efx_mcdi_mtd_partition *part, part 871 drivers/net/ethernet/sfc/siena.c part->nvram_type = type; part 872 drivers/net/ethernet/sfc/siena.c part->common.dev_type_name = "Siena NVRAM manager"; part 873 drivers/net/ethernet/sfc/siena.c part->common.type_name = info->name; part 875 drivers/net/ethernet/sfc/siena.c part->common.mtd.type = MTD_NORFLASH; part 876 drivers/net/ethernet/sfc/siena.c part->common.mtd.flags = MTD_CAP_NORFLASH; part 877 drivers/net/ethernet/sfc/siena.c part->common.mtd.size = size; part 878 drivers/net/ethernet/sfc/siena.c part->common.mtd.erasesize = erase_size; part 1508 drivers/net/ieee802154/at86rf230.c unsigned int part, version, val; part 1523 drivers/net/ieee802154/at86rf230.c rc = __at86rf230_read(lp, RG_PART_NUM, &part); part 1553 drivers/net/ieee802154/at86rf230.c switch (part) { part 963 drivers/net/wireless/cisco/airo.c } part; // saves partial message word across update() calls part 1680 drivers/net/wireless/cisco/airo.c context->part.d8[byte_position++] = *pOctets++; part 1684 drivers/net/wireless/cisco/airo.c MIC_ACCUM(ntohl(context->part.d32)); part 1698 drivers/net/wireless/cisco/airo.c context->part.d8[byte_position++] = *pOctets++; part 1722 drivers/net/wireless/cisco/airo.c val = ntohl(context->part.d32); part 19 drivers/net/wireless/mediatek/mt76/eeprom.c const char *part; part 41 drivers/net/wireless/mediatek/mt76/eeprom.c part = of_get_property(np, "label", NULL); part 42 drivers/net/wireless/mediatek/mt76/eeprom.c if (!part) part 43 drivers/net/wireless/mediatek/mt76/eeprom.c part = np->name; part 45 drivers/net/wireless/mediatek/mt76/eeprom.c mtd = get_mtd_device_nm(part); part 1061 drivers/net/wireless/ti/wlcore/debugfs.c struct wlcore_partition_set part, old_part; part 1079 drivers/net/wireless/ti/wlcore/debugfs.c memset(&part, 0, sizeof(part)); part 1080 drivers/net/wireless/ti/wlcore/debugfs.c part.mem.start = *ppos; part 1081 drivers/net/wireless/ti/wlcore/debugfs.c part.mem.size = bytes; part 1102 drivers/net/wireless/ti/wlcore/debugfs.c ret = wlcore_set_partition(wl, &part); part 1143 drivers/net/wireless/ti/wlcore/debugfs.c struct wlcore_partition_set part, old_part; part 1161 drivers/net/wireless/ti/wlcore/debugfs.c memset(&part, 0, sizeof(part)); part 1162 drivers/net/wireless/ti/wlcore/debugfs.c part.mem.start = *ppos; part 1163 drivers/net/wireless/ti/wlcore/debugfs.c part.mem.size = bytes; part 1184 drivers/net/wireless/ti/wlcore/debugfs.c ret = wlcore_set_partition(wl, &part); part 57 drivers/net/wireless/ti/wlcore/io.c struct wlcore_partition_set *part = &wl->curr_part; part 69 drivers/net/wireless/ti/wlcore/io.c if ((addr >= part->mem.start) && part 70 drivers/net/wireless/ti/wlcore/io.c (addr < part->mem.start + part->mem.size)) part 71 drivers/net/wireless/ti/wlcore/io.c return addr - part->mem.start; part 72 drivers/net/wireless/ti/wlcore/io.c else if ((addr >= part->reg.start) && part 73 drivers/net/wireless/ti/wlcore/io.c (addr < part->reg.start + part->reg.size)) part 74 drivers/net/wireless/ti/wlcore/io.c return addr - part->reg.start + part->mem.size; part 75 drivers/net/wireless/ti/wlcore/io.c else if ((addr >= part->mem2.start) && part 76 drivers/net/wireless/ti/wlcore/io.c (addr < part->mem2.start + part->mem2.size)) part 77 drivers/net/wireless/ti/wlcore/io.c return addr - part->mem2.start + part->mem.size + part 78 drivers/net/wireless/ti/wlcore/io.c part->reg.size; part 79 drivers/net/wireless/ti/wlcore/io.c else if ((addr >= part->mem3.start) && part 80 drivers/net/wireless/ti/wlcore/io.c (addr < part->mem3.start + part->mem3.size)) part 81 drivers/net/wireless/ti/wlcore/io.c return addr - part->mem3.start + part->mem.size + part 82 drivers/net/wireless/ti/wlcore/io.c part->reg.size + part->mem2.size; part 482 drivers/ntb/hw/idt/ntb_hw_idt.c unsigned char pidx, port, part; part 491 drivers/ntb/hw/idt/ntb_hw_idt.c ndev->part = GET_FIELD(SWPORTxSTS_SWPART, portsts); part 510 drivers/ntb/hw/idt/ntb_hw_idt.c part = GET_FIELD(SWPORTxSTS_SWPART, portsts); part 513 drivers/ntb/hw/idt/ntb_hw_idt.c partsts = idt_sw_read(ndev, partdata_tbl[part].sts); part 522 drivers/ntb/hw/idt/ntb_hw_idt.c ndev->peers[ndev->peer_cnt].part = part; part 525 drivers/ntb/hw/idt/ntb_hw_idt.c ndev->part_idx_map[part] = ndev->peer_cnt; part 645 drivers/ntb/hw/idt/ntb_hw_idt.c part_mask = ~BIT(ndev->part); part 648 drivers/ntb/hw/idt/ntb_hw_idt.c part_mask &= ~BIT(ndev->peers[pidx].part); part 746 drivers/ntb/hw/idt/ntb_hw_idt.c SET_FIELD(NTMTBLDATA_PART, 0, ndev->part) | part 749 drivers/ntb/hw/idt/ntb_hw_idt.c idt_nt_write(ndev, IDT_NT_NTMTBLADDR, ndev->part); part 755 drivers/ntb/hw/idt/ntb_hw_idt.c idt_sw_write(ndev, IDT_SW_SEGSIGSTS, (u32)1 << ndev->part); part 776 drivers/ntb/hw/idt/ntb_hw_idt.c idt_nt_write(ndev, IDT_NT_NTMTBLADDR, ndev->part); part 782 drivers/ntb/hw/idt/ntb_hw_idt.c idt_sw_write(ndev, IDT_SW_SEGSIGSTS, (u32)1 << ndev->part); part 815 drivers/ntb/hw/idt/ntb_hw_idt.c idt_nt_write(ndev, IDT_NT_NTMTBLADDR, ndev->part); part 861 drivers/ntb/hw/idt/ntb_hw_idt.c idt_nt_write(ndev, IDT_NT_NTMTBLADDR, ndev->peers[pidx].part); part 1316 drivers/ntb/hw/idt/ntb_hw_idt.c data = SET_FIELD(BARSETUP_TPART, data, ndev->peers[pidx].part); part 1331 drivers/ntb/hw/idt/ntb_hw_idt.c data = SET_FIELD(LUTUDATA_PART, 0, ndev->peers[pidx].part) | part 1802 drivers/ntb/hw/idt/ntb_hw_idt.c SET_FIELD(SWPxMSGCTL_PART, 0, ndev->peers[pidx].part); part 1807 drivers/ntb/hw/idt/ntb_hw_idt.c idt_sw_write(ndev, partdata_tbl[ndev->part].msgctl[midx], swpmsgctl); part 2361 drivers/ntb/hw/idt/ntb_hw_idt.c "Local Port %hhu, Partition %hhu\n", ndev->port, ndev->part); part 2368 drivers/ntb/hw/idt/ntb_hw_idt.c idx, ndev->peers[idx].port, ndev->peers[idx].part); part 1086 drivers/ntb/hw/idt/ntb_hw_idt.h unsigned char part; part 1129 drivers/ntb/hw/idt/ntb_hw_idt.h unsigned char part; part 683 drivers/pci/switch/switchtec.c s->part[i] = reg; part 893 drivers/pci/switch/switchtec.c int i, part; part 902 drivers/pci/switch/switchtec.c for (part = 0; part < stdev->partition_count; part++) { part 903 drivers/pci/switch/switchtec.c pcfg = &stdev->mmio_part_cfg_all[part]; part 904 drivers/pci/switch/switchtec.c p.partition = part; part 682 drivers/rtc/rtc-abx80x.c unsigned int part = id->driver_data; part 724 drivers/rtc/rtc-abx80x.c if (part == RV1805) { part 780 drivers/rtc/rtc-abx80x.c if (part == ABX80X) { part 789 drivers/rtc/rtc-abx80x.c part = i; part 792 drivers/rtc/rtc-abx80x.c if (partnumber != abx80x_caps[part].pn) { part 794 drivers/rtc/rtc-abx80x.c partnumber, abx80x_caps[part].pn); part 798 drivers/rtc/rtc-abx80x.c if (np && abx80x_caps[part].has_tc) part 825 drivers/rtc/rtc-abx80x.c if (abx80x_caps[part].has_wdog) { part 435 drivers/s390/block/dasd.c struct hd_struct *part; part 447 drivers/s390/block/dasd.c while ((part = disk_part_iter_next(&piter))) part 448 drivers/s390/block/dasd.c kobject_uevent(&part_to_dev(part)->kobj, KOBJ_CHANGE); part 462 drivers/s390/block/dasd.c struct hd_struct *part; part 474 drivers/s390/block/dasd.c while ((part = disk_part_iter_next(&piter))) part 475 drivers/s390/block/dasd.c kobject_uevent(&part_to_dev(part)->kobj, KOBJ_CHANGE); part 1140 drivers/scsi/bfa/bfa_defs.h struct bfa_flash_part_attr_s part[BFA_FLASH_PART_MAX]; part 4381 drivers/scsi/bfa/bfa_ioc.c attr->part[i].part_type = part 4382 drivers/scsi/bfa/bfa_ioc.c be32_to_cpu(f->part[i].part_type); part 4383 drivers/scsi/bfa/bfa_ioc.c attr->part[i].part_instance = part 4384 drivers/scsi/bfa/bfa_ioc.c be32_to_cpu(f->part[i].part_instance); part 4385 drivers/scsi/bfa/bfa_ioc.c attr->part[i].part_off = part 4386 drivers/scsi/bfa/bfa_ioc.c be32_to_cpu(f->part[i].part_off); part 4387 drivers/scsi/bfa/bfa_ioc.c attr->part[i].part_size = part 4388 drivers/scsi/bfa/bfa_ioc.c be32_to_cpu(f->part[i].part_size); part 4389 drivers/scsi/bfa/bfa_ioc.c attr->part[i].part_len = part 4390 drivers/scsi/bfa/bfa_ioc.c be32_to_cpu(f->part[i].part_len); part 4391 drivers/scsi/bfa/bfa_ioc.c attr->part[i].part_status = part 4392 drivers/scsi/bfa/bfa_ioc.c be32_to_cpu(f->part[i].part_status); part 774 drivers/scsi/csiostor/csio_hw.c u32 part, manufacturer; part 788 drivers/scsi/csiostor/csio_hw.c for (part = 0; part < ARRAY_SIZE(supported_flash); part++) part 789 drivers/scsi/csiostor/csio_hw.c if (supported_flash[part].vendor_and_model_id == flashid) { part 790 drivers/scsi/csiostor/csio_hw.c hw->params.sf_size = supported_flash[part].size_mb; part 120 drivers/staging/gs_fpgaboot/gs_fpgaboot.c pr_info("part: %s\n", fimage->part); part 142 drivers/staging/gs_fpgaboot/gs_fpgaboot.c err = readinfo_bitstream(bitdata, fimage->part, MAX_STR, &offset); part 36 drivers/staging/gs_fpgaboot/gs_fpgaboot.h char part[MAX_STR]; part 232 drivers/vhost/vringh.c size_t part, len = sizeof(struct vring_desc); part 238 drivers/vhost/vringh.c part = len; part 241 drivers/vhost/vringh.c if (!rcheck(vrh, addr, &part, range, getrange)) part 244 drivers/vhost/vringh.c err = copy(dst, src, part); part 248 drivers/vhost/vringh.c dst += part; part 249 drivers/vhost/vringh.c src += part; part 250 drivers/vhost/vringh.c len -= part; part 422 drivers/vhost/vringh.c u16 part = vrh->vring.num - off; part 423 drivers/vhost/vringh.c err = putused(&used_ring->ring[off], used, part); part 425 drivers/vhost/vringh.c err = putused(&used_ring->ring[0], used + part, part 426 drivers/vhost/vringh.c num_used - part); part 653 drivers/video/fbdev/amba-clcd.c unsigned int part; part 679 drivers/video/fbdev/amba-clcd.c if (amba_part(fb->dev) != panels[i].part) part 3008 fs/buffer.c struct hd_struct *part; part 3011 fs/buffer.c part = __disk_get_part(bio->bi_disk, bio->bi_partno); part 3012 fs/buffer.c if (part) part 3013 fs/buffer.c maxsector = part_nr_sects_read(part); part 457 fs/char_dev.c static struct kobject *exact_match(dev_t dev, int *part, void *data) part 656 fs/char_dev.c static struct kobject *base_probe(dev_t dev, int *part, void *data) part 157 fs/hfs/hfs_fs.h int session, part; part 82 fs/hfs/part_tbl.c (HFS_SB(sb)->part < 0 || HFS_SB(sb)->part == i)) { part 98 fs/hfs/part_tbl.c (HFS_SB(sb)->part < 0 || HFS_SB(sb)->part == i)) { part 149 fs/hfs/super.c if (sbi->part >= 0) part 150 fs/hfs/super.c seq_printf(seq, ",part=%u", sbi->part); part 237 fs/hfs/super.c hsb->part = -1; part 294 fs/hfs/super.c if (match_int(&args[0], &hsb->part)) { part 187 fs/hfsplus/hfsplus_fs.h int part, session; part 59 fs/hfsplus/options.c opts->part = -1; part 157 fs/hfsplus/options.c if (match_int(&args[0], &sbi->part)) { part 228 fs/hfsplus/options.c if (sbi->part >= 0) part 229 fs/hfsplus/options.c seq_printf(seq, ",part=%u", sbi->part); part 81 fs/hfsplus/part_tbl.c (sbi->part < 0 || sbi->part == i)) { part 102 fs/hfsplus/part_tbl.c (sbi->part < 0 || sbi->part == i)) { part 17 fs/orangefs/dir.c struct orangefs_dir_part *part; part 134 fs/orangefs/dir.c struct orangefs_dir_part *part, *new; part 138 fs/orangefs/dir.c part = od->part; part 139 fs/orangefs/dir.c while (part) { part 141 fs/orangefs/dir.c if (part->next) part 142 fs/orangefs/dir.c part = part->next; part 151 fs/orangefs/dir.c if (!od->part) part 152 fs/orangefs/dir.c od->part = new; part 154 fs/orangefs/dir.c part->next = new; part 189 fs/orangefs/dir.c static int fill_from_part(struct orangefs_dir_part *part, part 200 fs/orangefs/dir.c if (i > part->len) part 210 fs/orangefs/dir.c while (i < part->len) { part 211 fs/orangefs/dir.c if (part->len < i + sizeof *len) part 213 fs/orangefs/dir.c len = (void *)part + offset + i; part 220 fs/orangefs/dir.c if (part->len < i + padlen + sizeof *khandle) part 222 fs/orangefs/dir.c s = (void *)part + offset + i + sizeof *len; part 225 fs/orangefs/dir.c khandle = (void *)part + offset + i + padlen; part 232 fs/orangefs/dir.c BUG_ON(i > part->len); part 245 fs/orangefs/dir.c struct orangefs_dir_part *part; part 250 fs/orangefs/dir.c part = od->part; part 251 fs/orangefs/dir.c while (part->next && count) { part 253 fs/orangefs/dir.c part = part->next; part 261 fs/orangefs/dir.c while (part && part->len) { part 263 fs/orangefs/dir.c r = fill_from_part(part, ctx); part 276 fs/orangefs/dir.c part = part->next; part 291 fs/orangefs/dir.c struct orangefs_dir_part *part = od->part; part 292 fs/orangefs/dir.c while (part) { part 293 fs/orangefs/dir.c struct orangefs_dir_part *next = part->next; part 294 fs/orangefs/dir.c vfree(part); part 295 fs/orangefs/dir.c part = next; part 298 fs/orangefs/dir.c od->part = NULL; part 379 fs/orangefs/dir.c od->part = NULL; part 388 fs/orangefs/dir.c struct orangefs_dir_part *part = od->part; part 389 fs/orangefs/dir.c while (part) { part 390 fs/orangefs/dir.c struct orangefs_dir_part *next = part->next; part 391 fs/orangefs/dir.c vfree(part); part 392 fs/orangefs/dir.c part = next; part 393 fs/pstore/platform.c unsigned int part = 1; part 424 fs/pstore/platform.c record.part = part; part 437 fs/pstore/platform.c oopscount, part); part 466 fs/pstore/platform.c part++; part 402 fs/pstore/ram.c if (record->part != 1) part 2496 fs/udf/super.c unsigned int part = sbi->s_partition; part 2497 fs/udf/super.c int ptype = sbi->s_partmaps[part].s_partition_type; part 2500 fs/udf/super.c part = sbi->s_partmaps[part].s_type_specific.s_metadata. part 2514 fs/udf/super.c if (le32_to_cpu(lvid->numOfPartitions) > part) { part 2516 fs/udf/super.c lvid->freeSpaceTable[part]); part 2525 fs/udf/super.c map = &sbi->s_partmaps[part]; part 96 include/linux/amba/bus.h #define AMBA_LINUX_ID(conf, rev, part) \ part 98 include/linux/amba/bus.h AMBA_VENDOR_LINUX << 12 | ((part) & 0xfff)) part 51 include/linux/bcm963xx_nvram.h #define BCM963XX_NVRAM_NAND_PART_OFFSET(nvram, part) \ part 52 include/linux/bcm963xx_nvram.h bcm963xx_nvram_nand_part_offset(nvram, BCM963XX_NVRAM_NAND_PART_ ##part) part 56 include/linux/bcm963xx_nvram.h enum bcm963xx_nvram_nand_part part) part 58 include/linux/bcm963xx_nvram.h return nvram->nand_part_offset[part] * SZ_1K; part 61 include/linux/bcm963xx_nvram.h #define BCM963XX_NVRAM_NAND_PART_SIZE(nvram, part) \ part 62 include/linux/bcm963xx_nvram.h bcm963xx_nvram_nand_part_size(nvram, BCM963XX_NVRAM_NAND_PART_ ##part) part 66 include/linux/bcm963xx_nvram.h enum bcm963xx_nvram_nand_part part) part 68 include/linux/bcm963xx_nvram.h return nvram->nand_part_size[part] * SZ_1K; part 456 include/linux/bio.h unsigned long sectors, struct hd_struct *part); part 458 include/linux/bio.h struct hd_struct *part, part 196 include/linux/blkdev.h struct hd_struct *part; part 238 include/linux/clk/ti.h int (*cm_wait_module_ready)(u8 part, s16 prcm_mod, u16 idlest_reg, part 27 include/linux/genhd.h #define part_to_dev(part) (&((part)->__dev)) part 164 include/linux/genhd.h struct hd_struct __rcu *part[]; part 224 include/linux/genhd.h static inline struct gendisk *part_to_disk(struct hd_struct *part) part 226 include/linux/genhd.h if (likely(part)) { part 227 include/linux/genhd.h if (part->partno) part 228 include/linux/genhd.h return dev_to_disk(part_to_dev(part)->parent); part 230 include/linux/genhd.h return dev_to_disk(part_to_dev(part)); part 253 include/linux/genhd.h static inline dev_t part_devt(struct hd_struct *part) part 255 include/linux/genhd.h return part_to_dev(part)->devt; part 261 include/linux/genhd.h static inline void disk_put_part(struct hd_struct *part) part 263 include/linux/genhd.h if (likely(part)) part 264 include/linux/genhd.h put_device(part_to_dev(part)); part 277 include/linux/genhd.h struct hd_struct *part; part 306 include/linux/genhd.h #define part_stat_get_cpu(part, field, cpu) \ part 307 include/linux/genhd.h (per_cpu_ptr((part)->dkstats, (cpu))->field) part 309 include/linux/genhd.h #define part_stat_get(part, field) \ part 310 include/linux/genhd.h part_stat_get_cpu(part, field, smp_processor_id()) part 312 include/linux/genhd.h #define part_stat_read(part, field) \ part 314 include/linux/genhd.h typeof((part)->dkstats->field) res = 0; \ part 317 include/linux/genhd.h res += per_cpu_ptr((part)->dkstats, _cpu)->field; \ part 321 include/linux/genhd.h static inline void part_stat_set_all(struct hd_struct *part, int value) part 326 include/linux/genhd.h memset(per_cpu_ptr(part->dkstats, i), value, part 330 include/linux/genhd.h static inline int init_part_stats(struct hd_struct *part) part 332 include/linux/genhd.h part->dkstats = alloc_percpu(struct disk_stats); part 333 include/linux/genhd.h if (!part->dkstats) part 338 include/linux/genhd.h static inline void free_part_stats(struct hd_struct *part) part 340 include/linux/genhd.h free_percpu(part->dkstats); part 347 include/linux/genhd.h #define part_stat_get(part, field) ((part)->dkstats.field) part 348 include/linux/genhd.h #define part_stat_get_cpu(part, field, cpu) part_stat_get(part, field) part 349 include/linux/genhd.h #define part_stat_read(part, field) part_stat_get(part, field) part 351 include/linux/genhd.h static inline void part_stat_set_all(struct hd_struct *part, int value) part 353 include/linux/genhd.h memset(&part->dkstats, value, sizeof(struct disk_stats)); part 356 include/linux/genhd.h static inline int init_part_stats(struct hd_struct *part) part 361 include/linux/genhd.h static inline void free_part_stats(struct hd_struct *part) part 367 include/linux/genhd.h #define part_stat_read_msecs(part, which) \ part 368 include/linux/genhd.h div_u64(part_stat_read(part, nsecs[which]), NSEC_PER_MSEC) part 370 include/linux/genhd.h #define part_stat_read_accum(part, field) \ part 371 include/linux/genhd.h (part_stat_read(part, field[STAT_READ]) + \ part 372 include/linux/genhd.h part_stat_read(part, field[STAT_WRITE]) + \ part 373 include/linux/genhd.h part_stat_read(part, field[STAT_DISCARD])) part 375 include/linux/genhd.h #define __part_stat_add(part, field, addnd) \ part 376 include/linux/genhd.h (part_stat_get(part, field) += (addnd)) part 378 include/linux/genhd.h #define part_stat_add(part, field, addnd) do { \ part 379 include/linux/genhd.h __part_stat_add((part), field, addnd); \ part 380 include/linux/genhd.h if ((part)->partno) \ part 381 include/linux/genhd.h __part_stat_add(&part_to_disk((part))->part0, \ part 401 include/linux/genhd.h unsigned int part_in_flight(struct request_queue *q, struct hd_struct *part); part 402 include/linux/genhd.h void part_in_flight_rw(struct request_queue *q, struct hd_struct *part, part 404 include/linux/genhd.h void part_dec_in_flight(struct request_queue *q, struct hd_struct *part, part 406 include/linux/genhd.h void part_inc_in_flight(struct request_queue *q, struct hd_struct *part, part 417 include/linux/genhd.h static inline void free_part_info(struct hd_struct *part) part 419 include/linux/genhd.h kfree(part->info); part 422 include/linux/genhd.h void update_io_ticks(struct hd_struct *part, unsigned long now); part 618 include/linux/genhd.h extern int blk_alloc_devt(struct hd_struct *part, dev_t *devt); part 679 include/linux/genhd.h static inline int hd_ref_init(struct hd_struct *part) part 681 include/linux/genhd.h if (percpu_ref_init(&part->ref, __delete_partition, 0, part 687 include/linux/genhd.h static inline void hd_struct_get(struct hd_struct *part) part 689 include/linux/genhd.h percpu_ref_get(&part->ref); part 692 include/linux/genhd.h static inline int hd_struct_try_get(struct hd_struct *part) part 694 include/linux/genhd.h return percpu_ref_tryget_live(&part->ref); part 697 include/linux/genhd.h static inline void hd_struct_put(struct hd_struct *part) part 699 include/linux/genhd.h percpu_ref_put(&part->ref); part 702 include/linux/genhd.h static inline void hd_struct_kill(struct hd_struct *part) part 704 include/linux/genhd.h percpu_ref_kill(&part->ref); part 707 include/linux/genhd.h static inline void hd_free_part(struct hd_struct *part) part 709 include/linux/genhd.h free_part_stats(part); part 710 include/linux/genhd.h free_part_info(part); part 711 include/linux/genhd.h percpu_ref_exit(&part->ref); part 723 include/linux/genhd.h static inline sector_t part_nr_sects_read(struct hd_struct *part) part 729 include/linux/genhd.h seq = read_seqcount_begin(&part->nr_sects_seq); part 730 include/linux/genhd.h nr_sects = part->nr_sects; part 731 include/linux/genhd.h } while (read_seqcount_retry(&part->nr_sects_seq, seq)); part 737 include/linux/genhd.h nr_sects = part->nr_sects; part 741 include/linux/genhd.h return part->nr_sects; part 750 include/linux/genhd.h static inline void part_nr_sects_write(struct hd_struct *part, sector_t size) part 753 include/linux/genhd.h write_seqcount_begin(&part->nr_sects_seq); part 754 include/linux/genhd.h part->nr_sects = size; part 755 include/linux/genhd.h write_seqcount_end(&part->nr_sects_seq); part 758 include/linux/genhd.h part->nr_sects = size; part 761 include/linux/genhd.h part->nr_sects = size; part 308 include/linux/mmc/card.h struct mmc_part part[MMC_NUM_PHY_PARTITION]; /* physical partitions */ part 80 include/linux/pstore.h unsigned int part; part 58 include/uapi/linux/switchtec_ioctl.h __u32 part[48]; part 67 include/uapi/linux/switchtec_ioctl.h __u32 part[48]; part 81 init/do_mounts.c struct hd_struct *part = dev_to_part(dev); part 83 init/do_mounts.c if (!part->info) part 86 init/do_mounts.c if (strncasecmp(cmp->uuid, part->info->uuid, cmp->len)) part 114 init/do_mounts.c struct hd_struct *part; part 154 init/do_mounts.c part = disk_get_part(disk, dev_to_part(dev)->partno + offset); part 155 init/do_mounts.c if (part) { part 156 init/do_mounts.c res = part_devt(part); part 157 init/do_mounts.c put_device(part_to_dev(part)); part 183 init/do_mounts.c struct hd_struct *part = dev_to_part(dev); part 185 init/do_mounts.c if (part->info && !strcmp(label, part->info->volname)) part 227 init/do_mounts.c int part; part 295 init/do_mounts.c part = simple_strtoul(p, NULL, 10); part 297 init/do_mounts.c res = blk_lookup_devt(s, part); part 305 init/do_mounts.c res = blk_lookup_devt(s, part); part 458 kernel/trace/blktrace.c struct hd_struct *part = NULL; part 461 kernel/trace/blktrace.c part = bdev->bd_part; part 463 kernel/trace/blktrace.c if (part) { part 464 kernel/trace/blktrace.c bt->start_lba = part->start_sect; part 465 kernel/trace/blktrace.c bt->end_lba = part->start_sect + part->nr_sects; part 142 net/rds/transport.c unsigned int part; part 153 net/rds/transport.c part = trans->stats_info_copy(iter, avail); part 154 net/rds/transport.c avail -= min(avail, part); part 155 net/rds/transport.c total += part; part 1392 sound/soc/codecs/cs47l15.c cs47l15->core.adsp[0].part = "cs47l15"; part 1236 sound/soc/codecs/cs47l24.c cs47l24->core.adsp[i].part = "cs47l24"; part 1685 sound/soc/codecs/cs47l35.c cs47l35->core.adsp[i].part = "cs47l35"; part 2633 sound/soc/codecs/cs47l85.c cs47l85->core.adsp[i].part = "cs47l85"; part 2536 sound/soc/codecs/cs47l90.c cs47l90->core.adsp[i].part = "cs47l90"; part 1944 sound/soc/codecs/cs47l92.c cs47l92->core.adsp[0].part = "cs47l92"; part 2214 sound/soc/codecs/wm2200.c wm2200->dsp[i].part = "wm2200"; part 2050 sound/soc/codecs/wm5102.c wm5102->core.adsp[0].part = "wm5102"; part 2410 sound/soc/codecs/wm5110.c wm5110->core.adsp[i].part = "wm5110"; part 1819 sound/soc/codecs/wm_adsp.c snprintf(file, PAGE_SIZE, "%s-%s-%s.wmfw", dsp->part, dsp->fwf_name, part 2473 sound/soc/codecs/wm_adsp.c snprintf(file, PAGE_SIZE, "%s-%s-%s.bin", dsp->part, dsp->fwf_name, part 3816 sound/soc/codecs/wm_adsp.c buf->name = kasprintf(GFP_KERNEL, "%s-dsp-%s", ctl->dsp->part, part 57 sound/soc/codecs/wm_adsp.h const char *part; part 1037 sound/soc/intel/atom/sst-atom-controls.c cmd.param.part.cfg.rate = 2; /* 48khz */ part 1039 sound/soc/intel/atom/sst-atom-controls.c cmd.param.part.cfg.format = ids->format; /* stereo/Mono */ part 1040 sound/soc/intel/atom/sst-atom-controls.c cmd.param.part.cfg.s_length = 1; /* 24bit left justified */ part 425 sound/soc/intel/atom/sst-atom-controls.h } part; part 200 sound/soc/intel/atom/sst-mfld-dsp.h } part; part 223 sound/soc/intel/atom/sst-mfld-dsp.h } part; part 58 sound/soc/intel/atom/sst/sst.c if (isr.part.done_interrupt) { part 63 sound/soc/intel/atom/sst/sst.c header.p.header_high.part.done = 0; part 67 sound/soc/intel/atom/sst/sst.c isr.part.done_interrupt = 1; part 76 sound/soc/intel/atom/sst/sst.c if (isr.part.busy_interrupt) { part 80 sound/soc/intel/atom/sst/sst.c imr.part.busy_interrupt = 1; part 85 sound/soc/intel/atom/sst/sst.c if (sst_create_ipc_msg(&msg, header.p.header_high.part.large)) { part 90 sound/soc/intel/atom/sst/sst.c if (header.p.header_high.part.large) { part 104 sound/soc/intel/atom/sst/sst.c sst_is_process_reply(header.p.header_high.part.msg_id); part 76 sound/soc/intel/atom/sst/sst.h } part; part 87 sound/soc/intel/atom/sst/sst.h } part; part 103 sound/soc/intel/atom/sst/sst.h } part; part 123 sound/soc/intel/atom/sst/sst.h } part; part 132 sound/soc/intel/atom/sst/sst.h } part; part 141 sound/soc/intel/atom/sst/sst.h } part; part 129 sound/soc/intel/atom/sst/sst_ipc.c while (header.p.header_high.part.busy) { part 149 sound/soc/intel/atom/sst/sst_ipc.c if (header.p.header_high.part.busy) { part 165 sound/soc/intel/atom/sst/sst_ipc.c if (msg->mrfld_header.p.header_high.part.large) part 191 sound/soc/intel/atom/sst/sst_ipc.c isr.part.busy_interrupt = 1; part 197 sound/soc/intel/atom/sst/sst_ipc.c clear_ipc.p.header_high.part.busy = 0; part 198 sound/soc/intel/atom/sst/sst_ipc.c clear_ipc.p.header_high.part.done = 1; part 202 sound/soc/intel/atom/sst/sst_ipc.c imr.part.busy_interrupt = 0; part 335 sound/soc/intel/atom/sst/sst_ipc.c drv_id = msg_high.part.drv_id; part 345 sound/soc/intel/atom/sst/sst_ipc.c if (msg_high.part.result && drv_id && !msg_high.part.large) { part 348 sound/soc/intel/atom/sst/sst_ipc.c sst_wake_up_block(sst_drv_ctx, msg_high.part.result, part 349 sound/soc/intel/atom/sst/sst_ipc.c msg_high.part.drv_id, part 350 sound/soc/intel/atom/sst/sst_ipc.c msg_high.part.msg_id, NULL, 0); part 359 sound/soc/intel/atom/sst/sst_ipc.c if (msg_high.part.large) { part 366 sound/soc/intel/atom/sst/sst_ipc.c if (sst_wake_up_block(sst_drv_ctx, msg_high.part.result, part 367 sound/soc/intel/atom/sst/sst_ipc.c msg_high.part.drv_id, part 368 sound/soc/intel/atom/sst/sst_ipc.c msg_high.part.msg_id, data, msg_low)) part 371 sound/soc/intel/atom/sst/sst_ipc.c sst_wake_up_block(sst_drv_ctx, msg_high.part.result, part 372 sound/soc/intel/atom/sst/sst_ipc.c msg_high.part.drv_id, part 373 sound/soc/intel/atom/sst/sst_ipc.c msg_high.part.msg_id, NULL, 0); part 97 sound/soc/intel/atom/sst/sst_loader.c csr.part.xt_snoop = 1; part 251 sound/soc/intel/atom/sst/sst_pvt.c msg->mrfld_header.p.header_high.part.res_rqd = !sync; part 255 sound/soc/intel/atom/sst/sst_pvt.c msg->mrfld_header.p.header_high.part.res_rqd); part 307 sound/soc/intel/atom/sst/sst_pvt.c header->p.header_high.part.msg_id = msg; part 308 sound/soc/intel/atom/sst/sst_pvt.c header->p.header_high.part.task_id = task_id; part 309 sound/soc/intel/atom/sst/sst_pvt.c header->p.header_high.part.large = large; part 310 sound/soc/intel/atom/sst/sst_pvt.c header->p.header_high.part.drv_id = drv_id; part 311 sound/soc/intel/atom/sst/sst_pvt.c header->p.header_high.part.done = 0; part 312 sound/soc/intel/atom/sst/sst_pvt.c header->p.header_high.part.busy = 1; part 313 sound/soc/intel/atom/sst/sst_pvt.c header->p.header_high.part.res_rqd = 1; part 189 sound/soc/intel/atom/sst/sst_stream.c msg->mrfld_header.p.header_high.part.res_rqd = bytes->block; part 190 tools/usb/usbip/libsrc/vhci_driver.c int part; part 221 tools/usb/usbip/libsrc/vhci_driver.c for (part = 0; part < 3; part++) { part 222 tools/usb/usbip/libsrc/vhci_driver.c end = strchr(start, delim[part]); part 223 tools/usb/usbip/libsrc/vhci_driver.c if (end == NULL || (end - start) > max_len[part]) {