part              158 arch/arm/include/asm/hardware/cache-l2x0.h void l2x0_pmu_register(void __iomem *base, u32 part);
part              162 arch/arm/include/asm/hardware/cache-l2x0.h static inline void l2x0_pmu_register(void __iomem *base, u32 part) {}
part               57 arch/arm/mach-omap2/cm.h 	int (*wait_module_ready)(u8 part, s16 prcm_mod, u16 idlest_reg,
part               59 arch/arm/mach-omap2/cm.h 	int (*wait_module_idle)(u8 part, s16 prcm_mod, u16 idlest_reg,
part               61 arch/arm/mach-omap2/cm.h 	void (*module_enable)(u8 mode, u8 part, u16 inst, u16 clkctrl_offs);
part               62 arch/arm/mach-omap2/cm.h 	void (*module_disable)(u8 part, u16 inst, u16 clkctrl_offs);
part               63 arch/arm/mach-omap2/cm.h 	u32 (*xlate_clkctrl)(u8 part, u16 inst, u16 clkctrl_offs);
part               68 arch/arm/mach-omap2/cm.h int omap_cm_wait_module_ready(u8 part, s16 prcm_mod, u16 idlest_reg,
part               70 arch/arm/mach-omap2/cm.h int omap_cm_wait_module_idle(u8 part, s16 prcm_mod, u16 idlest_reg,
part               72 arch/arm/mach-omap2/cm.h int omap_cm_module_enable(u8 mode, u8 part, u16 inst, u16 clkctrl_offs);
part               73 arch/arm/mach-omap2/cm.h int omap_cm_module_disable(u8 part, u16 inst, u16 clkctrl_offs);
part               74 arch/arm/mach-omap2/cm.h u32 omap_cm_xlate_clkctrl(u8 part, u16 inst, u16 clkctrl_offs);
part              245 arch/arm/mach-omap2/cm2xxx.c int omap2xxx_cm_wait_module_ready(u8 part, s16 prcm_mod, u16 idlest_id,
part               54 arch/arm/mach-omap2/cm2xxx.h int omap2xxx_cm_wait_module_ready(u8 part, s16 prcm_mod, u16 idlest_id,
part              229 arch/arm/mach-omap2/cm33xx.c static int am33xx_cm_wait_module_ready(u8 part, s16 inst, u16 clkctrl_offs,
part              252 arch/arm/mach-omap2/cm33xx.c static int am33xx_cm_wait_module_idle(u8 part, s16 inst, u16 clkctrl_offs,
part              273 arch/arm/mach-omap2/cm33xx.c static void am33xx_cm_module_enable(u8 mode, u8 part, u16 inst,
part              292 arch/arm/mach-omap2/cm33xx.c static void am33xx_cm_module_disable(u8 part, u16 inst, u16 clkctrl_offs)
part              347 arch/arm/mach-omap2/cm33xx.c static u32 am33xx_cm_xlate_clkctrl(u8 part, u16 inst, u16 offset)
part               88 arch/arm/mach-omap2/cm3xxx.c static int omap3xxx_cm_wait_module_ready(u8 part, s16 prcm_mod, u16 idlest_id,
part               94 arch/arm/mach-omap2/cm_common.c int omap_cm_wait_module_ready(u8 part, s16 prcm_mod, u16 idlest_reg,
part              103 arch/arm/mach-omap2/cm_common.c 	return cm_ll_data->wait_module_ready(part, prcm_mod, idlest_reg,
part              120 arch/arm/mach-omap2/cm_common.c int omap_cm_wait_module_idle(u8 part, s16 prcm_mod, u16 idlest_reg,
part              129 arch/arm/mach-omap2/cm_common.c 	return cm_ll_data->wait_module_idle(part, prcm_mod, idlest_reg,
part              144 arch/arm/mach-omap2/cm_common.c int omap_cm_module_enable(u8 mode, u8 part, u16 inst, u16 clkctrl_offs)
part              152 arch/arm/mach-omap2/cm_common.c 	cm_ll_data->module_enable(mode, part, inst, clkctrl_offs);
part              166 arch/arm/mach-omap2/cm_common.c int omap_cm_module_disable(u8 part, u16 inst, u16 clkctrl_offs)
part              174 arch/arm/mach-omap2/cm_common.c 	cm_ll_data->module_disable(part, inst, clkctrl_offs);
part              178 arch/arm/mach-omap2/cm_common.c u32 omap_cm_xlate_clkctrl(u8 part, u16 inst, u16 clkctrl_offs)
part              185 arch/arm/mach-omap2/cm_common.c 	return cm_ll_data->xlate_clkctrl(part, inst, clkctrl_offs);
part               74 arch/arm/mach-omap2/cminst44xx.c static u32 omap4_cminst_read_inst_reg(u8 part, u16 inst, u16 idx);
part               85 arch/arm/mach-omap2/cminst44xx.c static u32 _clkctrl_idlest(u8 part, u16 inst, u16 clkctrl_offs)
part               87 arch/arm/mach-omap2/cminst44xx.c 	u32 v = omap4_cminst_read_inst_reg(part, inst, clkctrl_offs);
part              102 arch/arm/mach-omap2/cminst44xx.c static bool _is_module_ready(u8 part, u16 inst, u16 clkctrl_offs)
part              106 arch/arm/mach-omap2/cminst44xx.c 	v = _clkctrl_idlest(part, inst, clkctrl_offs);
part              113 arch/arm/mach-omap2/cminst44xx.c static u32 omap4_cminst_read_inst_reg(u8 part, u16 inst, u16 idx)
part              115 arch/arm/mach-omap2/cminst44xx.c 	BUG_ON(part >= OMAP4_MAX_PRCM_PARTITIONS ||
part              116 arch/arm/mach-omap2/cminst44xx.c 	       part == OMAP4430_INVALID_PRCM_PARTITION ||
part              117 arch/arm/mach-omap2/cminst44xx.c 	       !_cm_bases[part].va);
part              118 arch/arm/mach-omap2/cminst44xx.c 	return readl_relaxed(_cm_bases[part].va + inst + idx);
part              122 arch/arm/mach-omap2/cminst44xx.c static void omap4_cminst_write_inst_reg(u32 val, u8 part, u16 inst, u16 idx)
part              124 arch/arm/mach-omap2/cminst44xx.c 	BUG_ON(part >= OMAP4_MAX_PRCM_PARTITIONS ||
part              125 arch/arm/mach-omap2/cminst44xx.c 	       part == OMAP4430_INVALID_PRCM_PARTITION ||
part              126 arch/arm/mach-omap2/cminst44xx.c 	       !_cm_bases[part].va);
part              127 arch/arm/mach-omap2/cminst44xx.c 	writel_relaxed(val, _cm_bases[part].va + inst + idx);
part              131 arch/arm/mach-omap2/cminst44xx.c static u32 omap4_cminst_rmw_inst_reg_bits(u32 mask, u32 bits, u8 part, u16 inst,
part              136 arch/arm/mach-omap2/cminst44xx.c 	v = omap4_cminst_read_inst_reg(part, inst, idx);
part              139 arch/arm/mach-omap2/cminst44xx.c 	omap4_cminst_write_inst_reg(v, part, inst, idx);
part              144 arch/arm/mach-omap2/cminst44xx.c static u32 omap4_cminst_set_inst_reg_bits(u32 bits, u8 part, u16 inst, s16 idx)
part              146 arch/arm/mach-omap2/cminst44xx.c 	return omap4_cminst_rmw_inst_reg_bits(bits, bits, part, inst, idx);
part              149 arch/arm/mach-omap2/cminst44xx.c static u32 omap4_cminst_clear_inst_reg_bits(u32 bits, u8 part, u16 inst,
part              152 arch/arm/mach-omap2/cminst44xx.c 	return omap4_cminst_rmw_inst_reg_bits(bits, 0x0, part, inst, idx);
part              155 arch/arm/mach-omap2/cminst44xx.c static u32 omap4_cminst_read_inst_reg_bits(u8 part, u16 inst, s16 idx, u32 mask)
part              159 arch/arm/mach-omap2/cminst44xx.c 	v = omap4_cminst_read_inst_reg(part, inst, idx);
part              180 arch/arm/mach-omap2/cminst44xx.c static void _clktrctrl_write(u8 c, u8 part, u16 inst, u16 cdoffs)
part              184 arch/arm/mach-omap2/cminst44xx.c 	v = omap4_cminst_read_inst_reg(part, inst, cdoffs + OMAP4_CM_CLKSTCTRL);
part              187 arch/arm/mach-omap2/cminst44xx.c 	omap4_cminst_write_inst_reg(v, part, inst, cdoffs + OMAP4_CM_CLKSTCTRL);
part              199 arch/arm/mach-omap2/cminst44xx.c static bool omap4_cminst_is_clkdm_in_hwsup(u8 part, u16 inst, u16 cdoffs)
part              203 arch/arm/mach-omap2/cminst44xx.c 	v = omap4_cminst_read_inst_reg(part, inst, cdoffs + OMAP4_CM_CLKSTCTRL);
part              219 arch/arm/mach-omap2/cminst44xx.c static void omap4_cminst_clkdm_enable_hwsup(u8 part, u16 inst, u16 cdoffs)
part              221 arch/arm/mach-omap2/cminst44xx.c 	_clktrctrl_write(OMAP34XX_CLKSTCTRL_ENABLE_AUTO, part, inst, cdoffs);
part              234 arch/arm/mach-omap2/cminst44xx.c static void omap4_cminst_clkdm_disable_hwsup(u8 part, u16 inst, u16 cdoffs)
part              236 arch/arm/mach-omap2/cminst44xx.c 	_clktrctrl_write(OMAP34XX_CLKSTCTRL_DISABLE_AUTO, part, inst, cdoffs);
part              248 arch/arm/mach-omap2/cminst44xx.c static void omap4_cminst_clkdm_force_wakeup(u8 part, u16 inst, u16 cdoffs)
part              250 arch/arm/mach-omap2/cminst44xx.c 	_clktrctrl_write(OMAP34XX_CLKSTCTRL_FORCE_WAKEUP, part, inst, cdoffs);
part              257 arch/arm/mach-omap2/cminst44xx.c static void omap4_cminst_clkdm_force_sleep(u8 part, u16 inst, u16 cdoffs)
part              259 arch/arm/mach-omap2/cminst44xx.c 	_clktrctrl_write(OMAP34XX_CLKSTCTRL_FORCE_SLEEP, part, inst, cdoffs);
part              274 arch/arm/mach-omap2/cminst44xx.c static int omap4_cminst_wait_module_ready(u8 part, s16 inst, u16 clkctrl_offs,
part              279 arch/arm/mach-omap2/cminst44xx.c 	omap_test_timeout(_is_module_ready(part, inst, clkctrl_offs),
part              297 arch/arm/mach-omap2/cminst44xx.c static int omap4_cminst_wait_module_idle(u8 part, s16 inst, u16 clkctrl_offs,
part              302 arch/arm/mach-omap2/cminst44xx.c 	omap_test_timeout((_clkctrl_idlest(part, inst, clkctrl_offs) ==
part              318 arch/arm/mach-omap2/cminst44xx.c static void omap4_cminst_module_enable(u8 mode, u8 part, u16 inst,
part              323 arch/arm/mach-omap2/cminst44xx.c 	v = omap4_cminst_read_inst_reg(part, inst, clkctrl_offs);
part              326 arch/arm/mach-omap2/cminst44xx.c 	omap4_cminst_write_inst_reg(v, part, inst, clkctrl_offs);
part              337 arch/arm/mach-omap2/cminst44xx.c static void omap4_cminst_module_disable(u8 part, u16 inst, u16 clkctrl_offs)
part              341 arch/arm/mach-omap2/cminst44xx.c 	v = omap4_cminst_read_inst_reg(part, inst, clkctrl_offs);
part              343 arch/arm/mach-omap2/cminst44xx.c 	omap4_cminst_write_inst_reg(v, part, inst, clkctrl_offs);
part              476 arch/arm/mach-omap2/cminst44xx.c static u32 omap4_cminst_xlate_clkctrl(u8 part, u16 inst, u16 offset)
part              478 arch/arm/mach-omap2/cminst44xx.c 	return _cm_bases[part].pa + inst + offset;
part              138 arch/arm/mach-omap2/prm.h 	bool (*was_any_context_lost_old)(u8 part, s16 inst, u16 idx);
part              139 arch/arm/mach-omap2/prm.h 	void (*clear_context_loss_flags_old)(u8 part, s16 inst, u16 idx);
part              141 arch/arm/mach-omap2/prm.h 	int (*assert_hardreset)(u8 shift, u8 part, s16 prm_mod, u16 offset);
part              142 arch/arm/mach-omap2/prm.h 	int (*deassert_hardreset)(u8 shift, u8 st_shift, u8 part, s16 prm_mod,
part              144 arch/arm/mach-omap2/prm.h 	int (*is_hardreset_asserted)(u8 shift, u8 part, s16 prm_mod,
part              155 arch/arm/mach-omap2/prm.h int omap_prm_assert_hardreset(u8 shift, u8 part, s16 prm_mod, u16 offset);
part              156 arch/arm/mach-omap2/prm.h int omap_prm_deassert_hardreset(u8 shift, u8 st_shift, u8 part, s16 prm_mod,
part              158 arch/arm/mach-omap2/prm.h int omap_prm_is_hardreset_asserted(u8 shift, u8 part, s16 prm_mod, u16 offset);
part              160 arch/arm/mach-omap2/prm.h extern bool prm_was_any_context_lost_old(u8 part, s16 inst, u16 idx);
part              161 arch/arm/mach-omap2/prm.h extern void prm_clear_context_loss_flags_old(u8 part, s16 inst, u16 idx);
part               33 arch/arm/mach-omap2/prm2xxx_3xxx.c int omap2_prm_is_hardreset_asserted(u8 shift, u8 part, s16 prm_mod, u16 offset)
part               53 arch/arm/mach-omap2/prm2xxx_3xxx.c int omap2_prm_assert_hardreset(u8 shift, u8 part, s16 prm_mod, u16 offset)
part               82 arch/arm/mach-omap2/prm2xxx_3xxx.c int omap2_prm_deassert_hardreset(u8 rst_shift, u8 st_shift, u8 part,
part              100 arch/arm/mach-omap2/prm2xxx_3xxx.h int omap2_prm_is_hardreset_asserted(u8 shift, u8 part, s16 prm_mod, u16 offset);
part              101 arch/arm/mach-omap2/prm2xxx_3xxx.h int omap2_prm_assert_hardreset(u8 shift, u8 part, s16 prm_mod,
part              103 arch/arm/mach-omap2/prm2xxx_3xxx.h int omap2_prm_deassert_hardreset(u8 rst_shift, u8 st_shift, u8 part,
part               67 arch/arm/mach-omap2/prm33xx.c static int am33xx_prm_is_hardreset_asserted(u8 shift, u8 part, s16 inst,
part               93 arch/arm/mach-omap2/prm33xx.c static int am33xx_prm_assert_hardreset(u8 shift, u8 part, s16 inst,
part              122 arch/arm/mach-omap2/prm33xx.c static int am33xx_prm_deassert_hardreset(u8 shift, u8 st_shift, u8 part,
part              404 arch/arm/mach-omap2/prm44xx.c static bool omap44xx_prm_was_any_context_lost_old(u8 part, s16 inst, u16 idx)
part              406 arch/arm/mach-omap2/prm44xx.c 	return (omap4_prminst_read_inst_reg(part, inst, idx)) ? 1 : 0;
part              419 arch/arm/mach-omap2/prm44xx.c static void omap44xx_prm_clear_context_loss_flags_old(u8 part, s16 inst,
part              422 arch/arm/mach-omap2/prm44xx.c 	omap4_prminst_write_inst_reg(0xffffffff, part, inst, idx);
part              394 arch/arm/mach-omap2/prm_common.c bool prm_was_any_context_lost_old(u8 part, s16 inst, u16 idx)
part              399 arch/arm/mach-omap2/prm_common.c 		ret = prm_ll_data->was_any_context_lost_old(part, inst, idx);
part              418 arch/arm/mach-omap2/prm_common.c void prm_clear_context_loss_flags_old(u8 part, s16 inst, u16 idx)
part              421 arch/arm/mach-omap2/prm_common.c 		prm_ll_data->clear_context_loss_flags_old(part, inst, idx);
part              436 arch/arm/mach-omap2/prm_common.c int omap_prm_assert_hardreset(u8 shift, u8 part, s16 prm_mod, u16 offset)
part              444 arch/arm/mach-omap2/prm_common.c 	return prm_ll_data->assert_hardreset(shift, part, prm_mod, offset);
part              458 arch/arm/mach-omap2/prm_common.c int omap_prm_deassert_hardreset(u8 shift, u8 st_shift, u8 part, s16 prm_mod,
part              467 arch/arm/mach-omap2/prm_common.c 	return prm_ll_data->deassert_hardreset(shift, st_shift, part, prm_mod,
part              480 arch/arm/mach-omap2/prm_common.c int omap_prm_is_hardreset_asserted(u8 shift, u8 part, s16 prm_mod, u16 offset)
part              488 arch/arm/mach-omap2/prm_common.c 	return prm_ll_data->is_hardreset_asserted(shift, part, prm_mod, offset);
part               58 arch/arm/mach-omap2/prminst44xx.c u32 omap4_prminst_read_inst_reg(u8 part, s16 inst, u16 idx)
part               60 arch/arm/mach-omap2/prminst44xx.c 	BUG_ON(part >= OMAP4_MAX_PRCM_PARTITIONS ||
part               61 arch/arm/mach-omap2/prminst44xx.c 	       part == OMAP4430_INVALID_PRCM_PARTITION ||
part               62 arch/arm/mach-omap2/prminst44xx.c 	       !_prm_bases[part].va);
part               63 arch/arm/mach-omap2/prminst44xx.c 	return readl_relaxed(_prm_bases[part].va + inst + idx);
part               67 arch/arm/mach-omap2/prminst44xx.c void omap4_prminst_write_inst_reg(u32 val, u8 part, s16 inst, u16 idx)
part               69 arch/arm/mach-omap2/prminst44xx.c 	BUG_ON(part >= OMAP4_MAX_PRCM_PARTITIONS ||
part               70 arch/arm/mach-omap2/prminst44xx.c 	       part == OMAP4430_INVALID_PRCM_PARTITION ||
part               71 arch/arm/mach-omap2/prminst44xx.c 	       !_prm_bases[part].va);
part               72 arch/arm/mach-omap2/prminst44xx.c 	writel_relaxed(val, _prm_bases[part].va + inst + idx);
part               76 arch/arm/mach-omap2/prminst44xx.c u32 omap4_prminst_rmw_inst_reg_bits(u32 mask, u32 bits, u8 part, s16 inst,
part               81 arch/arm/mach-omap2/prminst44xx.c 	v = omap4_prminst_read_inst_reg(part, inst, idx);
part               84 arch/arm/mach-omap2/prminst44xx.c 	omap4_prminst_write_inst_reg(v, part, inst, idx);
part               99 arch/arm/mach-omap2/prminst44xx.c int omap4_prminst_is_hardreset_asserted(u8 shift, u8 part, s16 inst,
part              104 arch/arm/mach-omap2/prminst44xx.c 	v = omap4_prminst_read_inst_reg(part, inst, rstctrl_offs);
part              123 arch/arm/mach-omap2/prminst44xx.c int omap4_prminst_assert_hardreset(u8 shift, u8 part, s16 inst,
part              128 arch/arm/mach-omap2/prminst44xx.c 	omap4_prminst_rmw_inst_reg_bits(mask, mask, part, inst, rstctrl_offs);
part              152 arch/arm/mach-omap2/prminst44xx.c int omap4_prminst_deassert_hardreset(u8 shift, u8 st_shift, u8 part, s16 inst,
part              160 arch/arm/mach-omap2/prminst44xx.c 	if (omap4_prminst_is_hardreset_asserted(shift, part, inst,
part              165 arch/arm/mach-omap2/prminst44xx.c 	omap4_prminst_rmw_inst_reg_bits(0xffffffff, st_mask, part, inst,
part              168 arch/arm/mach-omap2/prminst44xx.c 	omap4_prminst_rmw_inst_reg_bits(mask, 0, part, inst, rstctrl_offs);
part              170 arch/arm/mach-omap2/prminst44xx.c 	omap_test_timeout(omap4_prminst_is_hardreset_asserted(st_shift, part,
part               20 arch/arm/mach-omap2/prminst44xx.h extern u32 omap4_prminst_read_inst_reg(u8 part, s16 inst, u16 idx);
part               21 arch/arm/mach-omap2/prminst44xx.h extern void omap4_prminst_write_inst_reg(u32 val, u8 part, s16 inst, u16 idx);
part               22 arch/arm/mach-omap2/prminst44xx.h extern u32 omap4_prminst_rmw_inst_reg_bits(u32 mask, u32 bits, u8 part,
part               27 arch/arm/mach-omap2/prminst44xx.h extern int omap4_prminst_is_hardreset_asserted(u8 shift, u8 part, s16 inst,
part               29 arch/arm/mach-omap2/prminst44xx.h extern int omap4_prminst_assert_hardreset(u8 shift, u8 part, s16 inst,
part               31 arch/arm/mach-omap2/prminst44xx.h int omap4_prminst_deassert_hardreset(u8 shift, u8 st_shift, u8 part,
part              474 arch/arm/mm/cache-l2x0-pmu.c void __init l2x0_pmu_register(void __iomem *base, u32 part)
part              489 arch/arm/mm/cache-l2x0-pmu.c 	switch (part & L2X0_CACHE_ID_PART_MASK) {
part              292 arch/ia64/include/asm/pal.h 				part		: 8,	/* 31-24 part of the
part              305 arch/ia64/include/asm/pal.h 				part		: 8,	/* 31-24 part of the
part              325 arch/ia64/include/asm/pal.h #define pclid_read_part		pclid_info_read.part
part              334 arch/ia64/include/asm/pal.h #define pclid_write_part	pclid_info_write.part
part              118 arch/mips/include/asm/xtalk/xwidget.h 	u32 part;
part              581 arch/mips/net/ebpf_jit.c 			u64 part = (value >> (16 * (3 - i))) & 0xffff;
part              583 arch/mips/net/ebpf_jit.c 			if (seen_part && needed_shift > 0 && (part || i == 3)) {
part              587 arch/mips/net/ebpf_jit.c 			if (part) {
part              588 arch/mips/net/ebpf_jit.c 				if (i == 0 || (!seen_part && i < 3 && part < 0x8000)) {
part              589 arch/mips/net/ebpf_jit.c 					emit_instr(ctx, lui, dst, (s32)(s16)part);
part              594 arch/mips/net/ebpf_jit.c 						   (unsigned int)part);
part               78 arch/powerpc/include/asm/nvram.h extern int __init nvram_init_os_partition(struct nvram_os_partition *part);
part               84 arch/powerpc/include/asm/nvram.h extern int nvram_read_partition(struct nvram_os_partition *part, char *buff,
part               89 arch/powerpc/include/asm/nvram.h extern int nvram_write_os_partition(struct nvram_os_partition *part,
part              180 arch/powerpc/kernel/nvram_64.c int nvram_write_os_partition(struct nvram_os_partition *part,
part              189 arch/powerpc/kernel/nvram_64.c 	if (part->index == -1)
part              192 arch/powerpc/kernel/nvram_64.c 	if (length > part->size)
part              193 arch/powerpc/kernel/nvram_64.c 		length = part->size;
part              198 arch/powerpc/kernel/nvram_64.c 	tmp_index = part->index;
part              219 arch/powerpc/kernel/nvram_64.c int nvram_read_partition(struct nvram_os_partition *part, char *buff,
part              227 arch/powerpc/kernel/nvram_64.c 	if (part->index == -1)
part              230 arch/powerpc/kernel/nvram_64.c 	if (length > part->size)
part              231 arch/powerpc/kernel/nvram_64.c 		length = part->size;
part              233 arch/powerpc/kernel/nvram_64.c 	tmp_index = part->index;
part              235 arch/powerpc/kernel/nvram_64.c 	if (part->os_partition) {
part              249 arch/powerpc/kernel/nvram_64.c 	if (part->os_partition) {
part              274 arch/powerpc/kernel/nvram_64.c int __init nvram_init_os_partition(struct nvram_os_partition *part)
part              280 arch/powerpc/kernel/nvram_64.c 	p = nvram_find_partition(part->name, NVRAM_SIG_OS, &size);
part              283 arch/powerpc/kernel/nvram_64.c 	if (p && size < part->min_size) {
part              285 arch/powerpc/kernel/nvram_64.c 					" removing it...\n", part->name);
part              286 arch/powerpc/kernel/nvram_64.c 		nvram_remove_partition(part->name, NVRAM_SIG_OS, NULL);
part              292 arch/powerpc/kernel/nvram_64.c 		p = nvram_create_partition(part->name, NVRAM_SIG_OS,
part              293 arch/powerpc/kernel/nvram_64.c 					part->req_size, part->min_size);
part              297 arch/powerpc/kernel/nvram_64.c 				part->name);
part              300 arch/powerpc/kernel/nvram_64.c 			p = nvram_create_partition(part->name, NVRAM_SIG_OS,
part              301 arch/powerpc/kernel/nvram_64.c 					part->req_size, part->min_size);
part              307 arch/powerpc/kernel/nvram_64.c 		       " partition, err %d\n", part->name, (int)p);
part              311 arch/powerpc/kernel/nvram_64.c 	part->index = p;
part              312 arch/powerpc/kernel/nvram_64.c 	part->size = nvram_get_partition_size(p) - sizeof(struct err_log_info);
part              392 arch/powerpc/kernel/nvram_64.c 	if (record->part > 1 || (record->type != PSTORE_TYPE_DMESG))
part              412 arch/powerpc/kernel/nvram_64.c 	record->id = record->part;
part              425 arch/powerpc/kernel/nvram_64.c 	struct nvram_os_partition *part = NULL;
part              434 arch/powerpc/kernel/nvram_64.c 		part = &oops_log_partition;
part              439 arch/powerpc/kernel/nvram_64.c 		part = &common_partition;
part              447 arch/powerpc/kernel/nvram_64.c 		part = &rtas_log_partition;
part              454 arch/powerpc/kernel/nvram_64.c 		part = &of_config_partition;
part              464 arch/powerpc/kernel/nvram_64.c 		part = &skiboot_partition;
part              475 arch/powerpc/kernel/nvram_64.c 	if (!part->os_partition) {
part              476 arch/powerpc/kernel/nvram_64.c 		p = nvram_find_partition(part->name, sig, &size);
part              479 arch/powerpc/kernel/nvram_64.c 				"err %d\n", part->name, (int)p);
part              482 arch/powerpc/kernel/nvram_64.c 		part->index = p;
part              483 arch/powerpc/kernel/nvram_64.c 		part->size = size;
part              486 arch/powerpc/kernel/nvram_64.c 	buff = kmalloc(part->size, GFP_KERNEL);
part              491 arch/powerpc/kernel/nvram_64.c 	if (nvram_read_partition(part, buff, part->size, &err_type, &id_no)) {
part              498 arch/powerpc/kernel/nvram_64.c 	if (part->os_partition)
part              531 arch/powerpc/kernel/nvram_64.c 	return part->size;
part              724 arch/powerpc/kernel/nvram_64.c static int __init nvram_write_header(struct nvram_partition * part)
part              730 arch/powerpc/kernel/nvram_64.c 	memcpy(&phead, &part->header, NVRAM_HEADER_LEN);
part              733 arch/powerpc/kernel/nvram_64.c 	tmp_index = part->index;
part              758 arch/powerpc/kernel/nvram_64.c static int nvram_can_remove_partition(struct nvram_partition *part,
part              761 arch/powerpc/kernel/nvram_64.c 	if (part->header.signature != sig)
part              764 arch/powerpc/kernel/nvram_64.c 		if (strncmp(name, part->header.name, 12))
part              769 arch/powerpc/kernel/nvram_64.c 			if (!strncmp(*except, part->header.name, 12))
part              788 arch/powerpc/kernel/nvram_64.c 	struct nvram_partition *part, *prev, *tmp;
part              791 arch/powerpc/kernel/nvram_64.c 	list_for_each_entry(part, &nvram_partitions, partition) {
part              792 arch/powerpc/kernel/nvram_64.c 		if (!nvram_can_remove_partition(part, name, sig, exceptions))
part              796 arch/powerpc/kernel/nvram_64.c 		part->header.signature = NVRAM_SIG_FREE;
part              797 arch/powerpc/kernel/nvram_64.c 		memset(part->header.name, 'w', 12);
part              798 arch/powerpc/kernel/nvram_64.c 		part->header.checksum = nvram_checksum(&part->header);
part              799 arch/powerpc/kernel/nvram_64.c 		rc = nvram_write_header(part);
part              808 arch/powerpc/kernel/nvram_64.c 	list_for_each_entry_safe(part, tmp, &nvram_partitions, partition) {
part              809 arch/powerpc/kernel/nvram_64.c 		if (part->header.signature != NVRAM_SIG_FREE) {
part              814 arch/powerpc/kernel/nvram_64.c 			prev->header.length += part->header.length;
part              821 arch/powerpc/kernel/nvram_64.c 			list_del(&part->partition);
part              822 arch/powerpc/kernel/nvram_64.c 			kfree(part);
part              824 arch/powerpc/kernel/nvram_64.c 			prev = part;
part              846 arch/powerpc/kernel/nvram_64.c 	struct nvram_partition *part;
part              874 arch/powerpc/kernel/nvram_64.c 	list_for_each_entry(part, &nvram_partitions, partition) {
part              875 arch/powerpc/kernel/nvram_64.c 		if (part->header.signature != NVRAM_SIG_FREE)
part              878 arch/powerpc/kernel/nvram_64.c 		if (part->header.length >= req_size) {
part              880 arch/powerpc/kernel/nvram_64.c 			free_part = part;
part              883 arch/powerpc/kernel/nvram_64.c 		if (part->header.length > size &&
part              884 arch/powerpc/kernel/nvram_64.c 		    part->header.length >= min_size) {
part              885 arch/powerpc/kernel/nvram_64.c 			size = part->header.length;
part              886 arch/powerpc/kernel/nvram_64.c 			free_part = part;
part              952 arch/powerpc/kernel/nvram_64.c 	struct nvram_partition *part;
part              954 arch/powerpc/kernel/nvram_64.c 	list_for_each_entry(part, &nvram_partitions, partition) {
part              955 arch/powerpc/kernel/nvram_64.c 		if (part->index + NVRAM_HEADER_LEN == data_index)
part              956 arch/powerpc/kernel/nvram_64.c 			return (part->header.length - 1) * NVRAM_BLOCK_LEN;
part              127 arch/powerpc/kernel/ptrace32.c 		u32 part;
part              135 arch/powerpc/kernel/ptrace32.c 			part = 1;  /* want the 2nd half of the register (right-most). */
part              137 arch/powerpc/kernel/ptrace32.c 			part = 0;  /* want the 1st half of the register (left-most). */
part              157 arch/powerpc/kernel/ptrace32.c 		reg32bits = ((u32*)&tmp)[part];
part              350 arch/powerpc/platforms/powermac/setup.c void note_bootable_part(dev_t dev, int part, int goodness);
part              357 arch/powerpc/platforms/powermac/setup.c void __ref note_bootable_part(dev_t dev, int part, int goodness)
part              370 arch/powerpc/platforms/powermac/setup.c 	ROOT_DEV = dev + part;
part               95 arch/x86/platform/uv/bios_uv.c 	union partition_info_u part;
part              102 arch/x86/platform/uv/bios_uv.c 	part.val = v0;
part              104 arch/x86/platform/uv/bios_uv.c 		*uvtype = part.hub_version;
part              106 arch/x86/platform/uv/bios_uv.c 		*partid = part.partition_id;
part              108 arch/x86/platform/uv/bios_uv.c 		*coher = part.coherence_id;
part              110 arch/x86/platform/uv/bios_uv.c 		*region = part.region_size;
part             1755 block/bio.c    void update_io_ticks(struct hd_struct *part, unsigned long now)
part             1759 block/bio.c    	stamp = READ_ONCE(part->stamp);
part             1761 block/bio.c    		if (likely(cmpxchg(&part->stamp, stamp, now) == stamp)) {
part             1762 block/bio.c    			__part_stat_add(part, io_ticks, 1);
part             1765 block/bio.c    	if (part->partno) {
part             1766 block/bio.c    		part = &part_to_disk(part)->part0;
part             1772 block/bio.c    			   unsigned long sectors, struct hd_struct *part)
part             1778 block/bio.c    	update_io_ticks(part, jiffies);
part             1779 block/bio.c    	part_stat_inc(part, ios[sgrp]);
part             1780 block/bio.c    	part_stat_add(part, sectors[sgrp], sectors);
part             1781 block/bio.c    	part_inc_in_flight(q, part, op_is_write(op));
part             1788 block/bio.c    			 struct hd_struct *part, unsigned long start_time)
part             1796 block/bio.c    	update_io_ticks(part, now);
part             1797 block/bio.c    	part_stat_add(part, nsecs[sgrp], jiffies_to_nsecs(duration));
part             1798 block/bio.c    	part_stat_add(part, time_in_queue, duration);
part             1799 block/bio.c    	part_dec_in_flight(q, part, op_is_write(req_op));
part              772 block/blk-cgroup.c 	int key_len, part;
part              782 block/blk-cgroup.c 	disk = get_gendisk(MKDEV(major, minor), &part);
part              785 block/blk-cgroup.c 	if (part) {
part              121 block/blk-core.c 	rq->part = NULL;
part              762 block/blk-core.c static bool should_fail_request(struct hd_struct *part, unsigned int bytes)
part              764 block/blk-core.c 	return part->make_it_fail && should_fail(&fail_make_request, bytes);
part              779 block/blk-core.c static inline bool should_fail_request(struct hd_struct *part,
part              787 block/blk-core.c static inline bool bio_check_ro(struct bio *bio, struct hd_struct *part)
part              791 block/blk-core.c 	if (part->policy && op_is_write(op)) {
part              800 block/blk-core.c 			bio_devname(bio, b), part->partno);
part             1313 block/blk-core.c 		struct hd_struct *part;
part             1316 block/blk-core.c 		part = req->part;
part             1317 block/blk-core.c 		part_stat_add(part, sectors[sgrp], bytes >> 9);
part             1331 block/blk-core.c 		struct hd_struct *part;
part             1334 block/blk-core.c 		part = req->part;
part             1336 block/blk-core.c 		update_io_ticks(part, jiffies);
part             1337 block/blk-core.c 		part_stat_inc(part, ios[sgrp]);
part             1338 block/blk-core.c 		part_stat_add(part, nsecs[sgrp], now - req->start_time_ns);
part             1339 block/blk-core.c 		part_stat_add(part, time_in_queue, nsecs_to_jiffies64(now - req->start_time_ns));
part             1340 block/blk-core.c 		part_dec_in_flight(req->q, part, rq_data_dir(req));
part             1342 block/blk-core.c 		hd_struct_put(part);
part             1349 block/blk-core.c 	struct hd_struct *part;
part             1358 block/blk-core.c 		part = rq->part;
part             1359 block/blk-core.c 		part_stat_inc(part, merges[rw]);
part             1361 block/blk-core.c 		part = disk_map_sector_rcu(rq->rq_disk, blk_rq_pos(rq));
part             1362 block/blk-core.c 		if (!hd_struct_try_get(part)) {
part             1371 block/blk-core.c 			part = &rq->rq_disk->part0;
part             1372 block/blk-core.c 			hd_struct_get(part);
part             1374 block/blk-core.c 		part_inc_in_flight(rq->q, part, rw);
part             1375 block/blk-core.c 		rq->part = part;
part             1378 block/blk-core.c 	update_io_ticks(part, jiffies);
part              682 block/blk-merge.c 		struct hd_struct *part;
part              685 block/blk-merge.c 		part = req->part;
part              687 block/blk-merge.c 		part_dec_in_flight(req->q, part, rq_data_dir(req));
part              689 block/blk-merge.c 		hd_struct_put(part);
part               95 block/blk-mq.c 	struct hd_struct *part;
part              108 block/blk-mq.c 	if (rq->part == mi->part)
part              114 block/blk-mq.c unsigned int blk_mq_in_flight(struct request_queue *q, struct hd_struct *part)
part              117 block/blk-mq.c 	struct mq_inflight mi = { .part = part, .inflight = inflight, };
part              131 block/blk-mq.c 	if (rq->part == mi->part)
part              137 block/blk-mq.c void blk_mq_in_flight_rw(struct request_queue *q, struct hd_struct *part,
part              140 block/blk-mq.c 	struct mq_inflight mi = { .part = part, .inflight = inflight, };
part              328 block/blk-mq.c 	rq->part = NULL;
part              188 block/blk-mq.h unsigned int blk_mq_in_flight(struct request_queue *q, struct hd_struct *part);
part              189 block/blk-mq.h void blk_mq_in_flight_rw(struct request_queue *q, struct hd_struct *part,
part               49 block/genhd.c  void part_inc_in_flight(struct request_queue *q, struct hd_struct *part, int rw)
part               54 block/genhd.c  	part_stat_local_inc(part, in_flight[rw]);
part               55 block/genhd.c  	if (part->partno)
part               56 block/genhd.c  		part_stat_local_inc(&part_to_disk(part)->part0, in_flight[rw]);
part               59 block/genhd.c  void part_dec_in_flight(struct request_queue *q, struct hd_struct *part, int rw)
part               64 block/genhd.c  	part_stat_local_dec(part, in_flight[rw]);
part               65 block/genhd.c  	if (part->partno)
part               66 block/genhd.c  		part_stat_local_dec(&part_to_disk(part)->part0, in_flight[rw]);
part               69 block/genhd.c  unsigned int part_in_flight(struct request_queue *q, struct hd_struct *part)
part               75 block/genhd.c  		return blk_mq_in_flight(q, part);
part               80 block/genhd.c  		inflight += part_stat_local_read_cpu(part, in_flight[0], cpu) +
part               81 block/genhd.c  			    part_stat_local_read_cpu(part, in_flight[1], cpu);
part               89 block/genhd.c  void part_in_flight_rw(struct request_queue *q, struct hd_struct *part,
part               95 block/genhd.c  		blk_mq_in_flight_rw(q, part, inflight);
part              102 block/genhd.c  		inflight[0] += part_stat_local_read_cpu(part, in_flight[0], cpu);
part              103 block/genhd.c  		inflight[1] += part_stat_local_read_cpu(part, in_flight[1], cpu);
part              117 block/genhd.c  	return rcu_dereference(ptbl->part[partno]);
part              136 block/genhd.c  	struct hd_struct *part;
part              139 block/genhd.c  	part = __disk_get_part(disk, partno);
part              140 block/genhd.c  	if (part)
part              141 block/genhd.c  		get_device(part_to_dev(part));
part              144 block/genhd.c  	return part;
part              168 block/genhd.c  	piter->part = NULL;
part              198 block/genhd.c  	disk_put_part(piter->part);
part              199 block/genhd.c  	piter->part = NULL;
part              220 block/genhd.c  		struct hd_struct *part;
part              222 block/genhd.c  		part = rcu_dereference(ptbl->part[piter->idx]);
part              223 block/genhd.c  		if (!part)
part              225 block/genhd.c  		if (!part_nr_sects_read(part) &&
part              231 block/genhd.c  		get_device(part_to_dev(part));
part              232 block/genhd.c  		piter->part = part;
part              239 block/genhd.c  	return piter->part;
part              254 block/genhd.c  	disk_put_part(piter->part);
part              255 block/genhd.c  	piter->part = NULL;
part              259 block/genhd.c  static inline int sector_in_part(struct hd_struct *part, sector_t sector)
part              261 block/genhd.c  	return part->start_sect <= sector &&
part              262 block/genhd.c  		sector < part->start_sect + part_nr_sects_read(part);
part              283 block/genhd.c  	struct hd_struct *part;
part              288 block/genhd.c  	part = rcu_dereference(ptbl->last_lookup);
part              289 block/genhd.c  	if (part && sector_in_part(part, sector))
part              290 block/genhd.c  		return part;
part              293 block/genhd.c  		part = rcu_dereference(ptbl->part[i]);
part              295 block/genhd.c  		if (part && sector_in_part(part, sector)) {
part              296 block/genhd.c  			rcu_assign_pointer(ptbl->last_lookup, part);
part              297 block/genhd.c  			return part;
part              488 block/genhd.c  int blk_alloc_devt(struct hd_struct *part, dev_t *devt)
part              490 block/genhd.c  	struct gendisk *disk = part_to_disk(part);
part              494 block/genhd.c  	if (part->partno < disk->minors) {
part              495 block/genhd.c  		*devt = MKDEV(disk->major, disk->first_minor + part->partno);
part              503 block/genhd.c  	idx = idr_alloc(&ext_devt_idr, part, 0, NR_EXT_DEVT, GFP_NOWAIT);
part              602 block/genhd.c  	struct hd_struct *part;
part              667 block/genhd.c  	while ((part = disk_part_iter_next(&piter)))
part              668 block/genhd.c  		kobject_uevent(&part_to_dev(part)->kobj, KOBJ_ADD);
part              776 block/genhd.c  	struct hd_struct *part;
part              789 block/genhd.c  	while ((part = disk_part_iter_next(&piter))) {
part              790 block/genhd.c  		invalidate_partition(disk, part->partno);
part              791 block/genhd.c  		bdev_unhash_inode(part_devt(part));
part              792 block/genhd.c  		delete_partition(disk, part->partno);
part              882 block/genhd.c  		struct hd_struct *part;
part              885 block/genhd.c  		part = idr_find(&ext_devt_idr, blk_mangle_minor(MINOR(devt)));
part              886 block/genhd.c  		if (part && get_disk_and_module(part_to_disk(part))) {
part              887 block/genhd.c  			*partno = part->partno;
part              888 block/genhd.c  			disk = part_to_disk(part);
part              928 block/genhd.c  	struct hd_struct *part;
part              931 block/genhd.c  	part = disk_get_part(disk, partno);
part              932 block/genhd.c  	if (part)
part              933 block/genhd.c  		bdev = bdget(part_devt(part));
part              934 block/genhd.c  	disk_put_part(part);
part              954 block/genhd.c  		struct hd_struct *part;
part              972 block/genhd.c  		while ((part = disk_part_iter_next(&piter))) {
part              973 block/genhd.c  			bool is_part0 = part == &disk->part0;
part              976 block/genhd.c  			       bdevt_str(part_devt(part), devt_buf),
part              977 block/genhd.c  			       (unsigned long long)part_nr_sects_read(part) >> 1
part              978 block/genhd.c  			       , disk_name(disk, part->partno, name_buf),
part              979 block/genhd.c  			       part->info ? part->info->uuid : "");
part             1055 block/genhd.c  	struct hd_struct *part;
part             1067 block/genhd.c  	while ((part = disk_part_iter_next(&piter)))
part             1069 block/genhd.c  			   MAJOR(part_devt(part)), MINOR(part_devt(part)),
part             1070 block/genhd.c  			   (unsigned long long)part_nr_sects_read(part) >> 1,
part             1071 block/genhd.c  			   disk_name(sgp, part->partno, buf));
part             1309 block/genhd.c  	new_ptbl = kzalloc_node(struct_size(new_ptbl, part, target), GFP_KERNEL,
part             1317 block/genhd.c  		rcu_assign_pointer(new_ptbl->part[i], old_ptbl->part[i]);
part             1438 block/genhd.c  		struct hd_struct *part;
part             1451 block/genhd.c  		part = disk_get_part(disk, partno);
part             1452 block/genhd.c  		if (part) {
part             1453 block/genhd.c  			devt = part_devt(part);
part             1454 block/genhd.c  			disk_put_part(part);
part             1457 block/genhd.c  		disk_put_part(part);
part             1490 block/genhd.c  		rcu_assign_pointer(ptbl->part[0], &disk->part0);
part             1580 block/genhd.c  	struct hd_struct *part;
part             1588 block/genhd.c  	while ((part = disk_part_iter_next(&piter)))
part             1589 block/genhd.c  		part->policy = flag;
part               18 block/ioctl.c  	struct hd_struct *part, *lpart;
part               58 block/ioctl.c  			while ((part = disk_part_iter_next(&piter))) {
part               59 block/ioctl.c  				if (!(start + length <= part->start_sect ||
part               60 block/ioctl.c  				      start >= part->start_sect + part->nr_sects)) {
part               69 block/ioctl.c  			part = add_partition(disk, partno, start, length,
part               72 block/ioctl.c  			return PTR_ERR_OR_ZERO(part);
part               74 block/ioctl.c  			part = disk_get_part(disk, partno);
part               75 block/ioctl.c  			if (!part)
part               78 block/ioctl.c  			bdevp = bdget(part_devt(part));
part               79 block/ioctl.c  			disk_put_part(part);
part              112 block/ioctl.c  			part = disk_get_part(disk, partno);
part              113 block/ioctl.c  			if (!part)
part              115 block/ioctl.c  			bdevp = bdget(part_devt(part));
part              117 block/ioctl.c  				disk_put_part(part);
part              122 block/ioctl.c  			if (start != part->start_sect) {
part              126 block/ioctl.c  				disk_put_part(part);
part              141 block/ioctl.c  					disk_put_part(part);
part              146 block/ioctl.c  			part_nr_sects_write(part, (sector_t)length);
part              151 block/ioctl.c  			disk_put_part(part);
part              233 block/partition-generic.c 	struct hd_struct *part = dev_to_part(dev);
part              235 block/partition-generic.c 	add_uevent_var(env, "PARTN=%u", part->partno);
part              236 block/partition-generic.c 	if (part->info && part->info->volname[0])
part              237 block/partition-generic.c 		add_uevent_var(env, "PARTNAME=%s", part->info->volname);
part              250 block/partition-generic.c 	struct hd_struct *part = container_of(to_rcu_work(work), struct hd_struct,
part              253 block/partition-generic.c 	part->start_sect = 0;
part              254 block/partition-generic.c 	part->nr_sects = 0;
part              255 block/partition-generic.c 	part_stat_set_all(part, 0);
part              256 block/partition-generic.c 	put_device(part_to_dev(part));
part              261 block/partition-generic.c 	struct hd_struct *part = container_of(ref, struct hd_struct, ref);
part              262 block/partition-generic.c 	INIT_RCU_WORK(&part->rcu_work, delete_partition_work_fn);
part              263 block/partition-generic.c 	queue_rcu_work(system_wq, &part->rcu_work);
part              274 block/partition-generic.c 	struct hd_struct *part;
part              279 block/partition-generic.c 	part = rcu_dereference_protected(ptbl->part[partno], 1);
part              280 block/partition-generic.c 	if (!part)
part              283 block/partition-generic.c 	rcu_assign_pointer(ptbl->part[partno], NULL);
part              285 block/partition-generic.c 	kobject_put(part->holder_dir);
part              286 block/partition-generic.c 	device_del(part_to_dev(part));
part              294 block/partition-generic.c 	blk_invalidate_devt(part_devt(part));
part              295 block/partition-generic.c 	hd_struct_kill(part);
part              326 block/partition-generic.c 	if (ptbl->part[partno])
part              402 block/partition-generic.c 	rcu_assign_pointer(ptbl->part[partno], p);
part              445 block/partition-generic.c 	struct hd_struct *part;
part              455 block/partition-generic.c 	while ((part = disk_part_iter_next(&piter)))
part              456 block/partition-generic.c 		delete_partition(disk, part->partno);
part              515 block/partition-generic.c 	struct hd_struct *part;
part              625 block/partition-generic.c 		part = add_partition(disk, p, from, size,
part              628 block/partition-generic.c 		if (IS_ERR(part)) {
part              630 block/partition-generic.c 			       disk->disk_name, p, -PTR_ERR(part));
part              635 block/partition-generic.c 			md_autodetect_dev(part_to_dev(part)->devt);
part               66 block/partitions/acorn.c 	struct riscix_part part[8];
part               87 block/partitions/acorn.c 		int part;
part               92 block/partitions/acorn.c 		for (part = 0; part < 8; part++) {
part               93 block/partitions/acorn.c 			if (rr->part[part].one &&
part               94 block/partitions/acorn.c 			    memcmp(rr->part[part].name, "All\0", 4)) {
part               96 block/partitions/acorn.c 					le32_to_cpu(rr->part[part].start),
part               97 block/partitions/acorn.c 					le32_to_cpu(rr->part[part].length));
part               99 block/partitions/acorn.c 				strlcat(state->pp_buf, rr->part[part].name, PAGE_SIZE);
part               35 block/partitions/amiga.c 	int start_sect, nr_sects, blk, part, res = 0;
part               84 block/partitions/amiga.c 	for (part = 1; blk>0 && part<=16; part++, put_dev_sector(sect)) {
part               59 block/partitions/atari.c 	if (!VALID_PARTITION(&rs->part[0], hd_size) &&
part               60 block/partitions/atari.c 	    !VALID_PARTITION(&rs->part[1], hd_size) &&
part               61 block/partitions/atari.c 	    !VALID_PARTITION(&rs->part[2], hd_size) &&
part               62 block/partitions/atari.c 	    !VALID_PARTITION(&rs->part[3], hd_size)) {
part               72 block/partitions/atari.c 	pi = &rs->part[0];
part               74 block/partitions/atari.c 	for (slot = 1; pi < &rs->part[4] && slot < state->limit; slot++, pi++) {
part              103 block/partitions/atari.c 			if (!(xrs->part[0].flg & 1)) {
part              110 block/partitions/atari.c 				   partsect + be32_to_cpu(xrs->part[0].st),
part              111 block/partitions/atari.c 				   be32_to_cpu(xrs->part[0].siz));
part              113 block/partitions/atari.c 			if (!(xrs->part[1].flg & 1)) {
part              118 block/partitions/atari.c 			if (memcmp( xrs->part[1].id, "XGM", 3 ) != 0) {
part              124 block/partitions/atari.c 			partsect = be32_to_cpu(xrs->part[1].st) + extensect;
part               31 block/partitions/atari.h   struct partition_info part[4];
part              142 block/partitions/efi.c static inline int pmbr_part_valid(gpt_mbr_record *part)
part              144 block/partitions/efi.c 	if (part->os_type != EFI_PMBR_OSTYPE_EFI_GPT)
part              148 block/partitions/efi.c 	if (le32_to_cpu(part->starting_lba) != GPT_PRIMARY_PARTITION_TABLE_LBA)
part              177 block/partitions/efi.c 	int i, part = 0, ret = 0; /* invalid by default */
part              185 block/partitions/efi.c 			part = i;
part              217 block/partitions/efi.c 		sz = le32_to_cpu(mbr->partition_record[part].size_in_lba);
part              576 block/partitions/ldm.c 	struct vblk_part *part;
part              592 block/partitions/ldm.c 		part = &vb->vblk.part;
part              594 block/partitions/ldm.c 		if (part->disk_id != disk->obj_id)
part              598 block/partitions/ldm.c 				part->start, part->size);
part              931 block/partitions/ldm.c 	struct vblk_part *part;
part              980 block/partitions/ldm.c 	part = &vb->vblk.part;
part              981 block/partitions/ldm.c 	part->start = get_unaligned_be64(buffer + 0x24 + r_name);
part              982 block/partitions/ldm.c 	part->volume_offset = get_unaligned_be64(buffer + 0x2C + r_name);
part              983 block/partitions/ldm.c 	part->size = ldm_get_vnum(buffer + 0x34 + r_name);
part              984 block/partitions/ldm.c 	part->parent_id = ldm_get_vnum(buffer + 0x34 + r_size);
part              985 block/partitions/ldm.c 	part->disk_id = ldm_get_vnum(buffer + 0x34 + r_parent);
part              987 block/partitions/ldm.c 		part->partnum = buffer[0x35 + r_diskid];
part              989 block/partitions/ldm.c 		part->partnum = 0;
part             1203 block/partitions/ldm.c 			if ((v->vblk.part.disk_id == vb->vblk.part.disk_id) &&
part             1204 block/partitions/ldm.c 			    (v->vblk.part.start > vb->vblk.part.start)) {
part              179 block/partitions/ldm.h 		struct vblk_part part;
part               16 block/partitions/mac.c extern void note_bootable_part(dev_t dev, int part, int goodness);
part               41 block/partitions/mac.c 	struct mac_partition *part;
part               59 block/partitions/mac.c 	if (partoffset + sizeof(*part) > datasize)
part               61 block/partitions/mac.c 	part = (struct mac_partition *) (data + partoffset);
part               62 block/partitions/mac.c 	if (be16_to_cpu(part->signature) != MAC_PARTITION_MAGIC) {
part               66 block/partitions/mac.c 	blocks_in_map = be32_to_cpu(part->map_count);
part               82 block/partitions/mac.c 		part = (struct mac_partition *) (data + pos%512);
part               83 block/partitions/mac.c 		if (be16_to_cpu(part->signature) != MAC_PARTITION_MAGIC)
part               86 block/partitions/mac.c 			be32_to_cpu(part->start_block) * (secsize/512),
part               87 block/partitions/mac.c 			be32_to_cpu(part->block_count) * (secsize/512));
part               89 block/partitions/mac.c 		if (!strncasecmp(part->type, "Linux_RAID", 10))
part               99 block/partitions/mac.c 			mac_fix_string(part->processor, 16);
part              100 block/partitions/mac.c 			mac_fix_string(part->name, 32);
part              101 block/partitions/mac.c 			mac_fix_string(part->type, 32);					
part              103 block/partitions/mac.c 			if ((be32_to_cpu(part->status) & MAC_STATUS_BOOTABLE)
part              104 block/partitions/mac.c 			    && strcasecmp(part->processor, "powerpc") == 0)
part              107 block/partitions/mac.c 			if (strcasecmp(part->type, "Apple_UNIX_SVR2") == 0
part              108 block/partitions/mac.c 			    || (strncasecmp(part->type, "Linux", 5) == 0
part              109 block/partitions/mac.c 			        && strcasecmp(part->type, "Linux_swap") != 0)) {
part              113 block/partitions/mac.c 				l = strlen(part->name);
part              114 block/partitions/mac.c 				if (strcmp(part->name, "/") == 0)
part              117 block/partitions/mac.c 					if (strncasecmp(part->name + i, "root",
part              123 block/partitions/mac.c 				if (strncasecmp(part->name, "swap", 4) == 0)
part               23 drivers/acpi/acpica/utmath.c 	struct uint64_struct part;
part               62 drivers/acpi/acpica/utmath.c 	ACPI_MUL_64_BY_32(0, multiplicand_ovl.part.hi, multiplier,
part               63 drivers/acpi/acpica/utmath.c 			  product.part.hi, carry32);
part               65 drivers/acpi/acpica/utmath.c 	ACPI_MUL_64_BY_32(0, multiplicand_ovl.part.lo, multiplier,
part               66 drivers/acpi/acpica/utmath.c 			  product.part.lo, carry32);
part               68 drivers/acpi/acpica/utmath.c 	product.part.hi += carry32;
part              100 drivers/acpi/acpica/utmath.c 		operand_ovl.part.hi = operand_ovl.part.lo;
part              101 drivers/acpi/acpica/utmath.c 		operand_ovl.part.lo = 0;
part              104 drivers/acpi/acpica/utmath.c 	ACPI_SHIFT_LEFT_64_BY_32(operand_ovl.part.hi,
part              105 drivers/acpi/acpica/utmath.c 				 operand_ovl.part.lo, count);
part              137 drivers/acpi/acpica/utmath.c 		operand_ovl.part.lo = operand_ovl.part.hi;
part              138 drivers/acpi/acpica/utmath.c 		operand_ovl.part.hi = 0;
part              141 drivers/acpi/acpica/utmath.c 	ACPI_SHIFT_RIGHT_64_BY_32(operand_ovl.part.hi,
part              142 drivers/acpi/acpica/utmath.c 				  operand_ovl.part.lo, count);
part              278 drivers/acpi/acpica/utmath.c 	ACPI_DIV_64_BY_32(0, dividend_ovl.part.hi, divisor,
part              279 drivers/acpi/acpica/utmath.c 			  quotient.part.hi, remainder32);
part              281 drivers/acpi/acpica/utmath.c 	ACPI_DIV_64_BY_32(remainder32, dividend_ovl.part.lo, divisor,
part              282 drivers/acpi/acpica/utmath.c 			  quotient.part.lo, remainder32);
part              336 drivers/acpi/acpica/utmath.c 	if (divisor.part.hi == 0) {
part              341 drivers/acpi/acpica/utmath.c 		remainder.part.hi = 0;
part              347 drivers/acpi/acpica/utmath.c 		ACPI_DIV_64_BY_32(0, dividend.part.hi, divisor.part.lo,
part              348 drivers/acpi/acpica/utmath.c 				  quotient.part.hi, partial1);
part              350 drivers/acpi/acpica/utmath.c 		ACPI_DIV_64_BY_32(partial1, dividend.part.lo, divisor.part.lo,
part              351 drivers/acpi/acpica/utmath.c 				  quotient.part.lo, remainder.part.lo);
part              359 drivers/acpi/acpica/utmath.c 		quotient.part.hi = 0;
part              366 drivers/acpi/acpica/utmath.c 			ACPI_SHIFT_RIGHT_64(normalized_divisor.part.hi,
part              367 drivers/acpi/acpica/utmath.c 					    normalized_divisor.part.lo);
part              368 drivers/acpi/acpica/utmath.c 			ACPI_SHIFT_RIGHT_64(normalized_dividend.part.hi,
part              369 drivers/acpi/acpica/utmath.c 					    normalized_dividend.part.lo);
part              371 drivers/acpi/acpica/utmath.c 		} while (normalized_divisor.part.hi != 0);
part              375 drivers/acpi/acpica/utmath.c 		ACPI_DIV_64_BY_32(normalized_dividend.part.hi,
part              376 drivers/acpi/acpica/utmath.c 				  normalized_dividend.part.lo,
part              377 drivers/acpi/acpica/utmath.c 				  normalized_divisor.part.lo, quotient.part.lo,
part              384 drivers/acpi/acpica/utmath.c 		partial1 = quotient.part.lo * divisor.part.hi;
part              385 drivers/acpi/acpica/utmath.c 		partial2.full = (u64) quotient.part.lo * divisor.part.lo;
part              386 drivers/acpi/acpica/utmath.c 		partial3.full = (u64) partial2.part.hi + partial1;
part              388 drivers/acpi/acpica/utmath.c 		remainder.part.hi = partial3.part.lo;
part              389 drivers/acpi/acpica/utmath.c 		remainder.part.lo = partial2.part.lo;
part              391 drivers/acpi/acpica/utmath.c 		if (partial3.part.hi == 0) {
part              392 drivers/acpi/acpica/utmath.c 			if (partial3.part.lo >= dividend.part.hi) {
part              393 drivers/acpi/acpica/utmath.c 				if (partial3.part.lo == dividend.part.hi) {
part              394 drivers/acpi/acpica/utmath.c 					if (partial2.part.lo > dividend.part.lo) {
part              395 drivers/acpi/acpica/utmath.c 						quotient.part.lo--;
part              399 drivers/acpi/acpica/utmath.c 					quotient.part.lo--;
part              405 drivers/acpi/acpica/utmath.c 			remainder.part.hi = (u32)-((s32)remainder.part.hi);
part              406 drivers/acpi/acpica/utmath.c 			remainder.part.lo = (u32)-((s32)remainder.part.lo);
part              408 drivers/acpi/acpica/utmath.c 			if (remainder.part.lo) {
part              409 drivers/acpi/acpica/utmath.c 				remainder.part.hi--;
part             1849 drivers/block/amiflop.c static struct kobject *floppy_find(dev_t dev, int *part, void *data)
part             1851 drivers/block/amiflop.c 	int drive = *part & 3;
part             1854 drivers/block/amiflop.c 	*part = 0;
part             1964 drivers/block/ataflop.c static struct kobject *floppy_find(dev_t dev, int *part, void *data)
part             1966 drivers/block/ataflop.c 	int drive = *part & 3;
part             1967 drivers/block/ataflop.c 	int type  = *part >> 2;
part             1970 drivers/block/ataflop.c 	*part = 0;
part              456 drivers/block/brd.c static struct kobject *brd_probe(dev_t dev, int *part, void *data)
part              468 drivers/block/brd.c 		*part = 0;
part             4527 drivers/block/floppy.c static struct kobject *floppy_find(dev_t dev, int *part, void *data)
part             4529 drivers/block/floppy.c 	int drive = (*part & 3) | ((*part & 0x80) >> 5);
part             4532 drivers/block/floppy.c 	if (((*part >> 2) & 0x1f) >= ARRAY_SIZE(floppy_type))
part             4534 drivers/block/floppy.c 	*part = 0;
part             2154 drivers/block/loop.c static struct kobject *loop_probe(dev_t dev, int *part, void *data)
part             2170 drivers/block/loop.c 	*part = 0;
part              766 drivers/block/swim.c static struct kobject *floppy_find(dev_t dev, int *part, void *data)
part              769 drivers/block/swim.c 	int drive = (*part & 3);
part              774 drivers/block/swim.c 	*part = 0;
part              333 drivers/block/z2ram.c static struct kobject *z2_find(dev_t dev, int *part, void *data)
part              335 drivers/block/z2ram.c 	*part = 0;
part              298 drivers/char/nvram.c 			int part, offset;
part              300 drivers/char/nvram.c 			if (copy_from_user(&part, (void __user *)arg,
part              301 drivers/char/nvram.c 					   sizeof(part)) != 0)
part              303 drivers/char/nvram.c 			if (part < pmac_nvram_OF || part > pmac_nvram_NR)
part              305 drivers/char/nvram.c 			offset = pmac_get_partition(part);
part               33 drivers/firmware/efi/efi-pstore.c static inline u64 generic_id(u64 timestamp, unsigned int part, int count)
part               35 drivers/firmware/efi/efi-pstore.c 	return (timestamp * 100 + part) * 1000 + count;
part               45 drivers/firmware/efi/efi-pstore.c 	unsigned int part;
part               56 drivers/firmware/efi/efi-pstore.c 		   &record->type, &part, &cnt, &time, &data_type) == 5) {
part               57 drivers/firmware/efi/efi-pstore.c 		record->id = generic_id(time, part, cnt);
part               58 drivers/firmware/efi/efi-pstore.c 		record->part = part;
part               68 drivers/firmware/efi/efi-pstore.c 		   &record->type, &part, &cnt, &time) == 4) {
part               69 drivers/firmware/efi/efi-pstore.c 		record->id = generic_id(time, part, cnt);
part               70 drivers/firmware/efi/efi-pstore.c 		record->part = part;
part               77 drivers/firmware/efi/efi-pstore.c 			  &record->type, &part, &time) == 3) {
part               83 drivers/firmware/efi/efi-pstore.c 		record->id = generic_id(time, part, 0);
part               84 drivers/firmware/efi/efi-pstore.c 		record->part = part;
part              249 drivers/firmware/efi/efi-pstore.c 	record->id = generic_id(record->time.tv_sec, record->part,
part              256 drivers/firmware/efi/efi-pstore.c 		 record->type, record->part, record->count,
part              333 drivers/firmware/efi/efi-pstore.c 		 record->type, record->part, record->count,
part              340 drivers/firmware/efi/efi-pstore.c 		record->type, record->part, (long long)record->time.tv_sec);
part              300 drivers/gpu/drm/i915/display/intel_display_types.h 				 int i, u32 *part);
part             5770 drivers/gpu/drm/i915/display/intel_dp.c 				    int i, u32 *part)
part             5778 drivers/gpu/drm/i915/display/intel_dp.c 			       DP_AUX_HDCP_V_PRIME(i), part,
part             1415 drivers/gpu/drm/i915/display/intel_hdmi.c 				      int i, u32 *part)
part             1423 drivers/gpu/drm/i915/display/intel_hdmi.c 				   part, DRM_HDCP_V_PRIME_PART_LEN);
part               74 drivers/gpu/drm/nouveau/dispnv50/wndw.c 		args.nv50.part = NV50_DMA_V0_PART_256;
part               78 drivers/gpu/drm/nouveau/dispnv50/wndw.c 		args.nv50.part = NV50_DMA_V0_PART_256;
part               33 drivers/gpu/drm/nouveau/include/nvif/cl0002.h 	__u8  part;
part               17 drivers/gpu/drm/nouveau/include/nvkm/subdev/mmu.h 	bool part:1; /* Region was split from an allocated region by map(). */
part               78 drivers/gpu/drm/nouveau/nvkm/engine/dma/usernv50.c 	u32 user, part, comp, kind;
part               97 drivers/gpu/drm/nouveau/nvkm/engine/dma/usernv50.c 			   args->v0.priv, args->v0.part, args->v0.comp,
part              100 drivers/gpu/drm/nouveau/nvkm/engine/dma/usernv50.c 		part = args->v0.part;
part              107 drivers/gpu/drm/nouveau/nvkm/engine/dma/usernv50.c 			part = NV50_DMA_V0_PART_256;
part              112 drivers/gpu/drm/nouveau/nvkm/engine/dma/usernv50.c 			part = NV50_DMA_V0_PART_VM;
part              119 drivers/gpu/drm/nouveau/nvkm/engine/dma/usernv50.c 	if (user > 2 || part > 2 || comp > 3 || kind > 0x7f)
part              123 drivers/gpu/drm/nouveau/nvkm/engine/dma/usernv50.c 	dmaobj->flags5 = (part << 16);
part              125 drivers/gpu/drm/nouveau/nvkm/subdev/devinit/gt215.c 			u32 part = (addr / mmio[2]) & 7;
part              128 drivers/gpu/drm/nouveau/nvkm/subdev/devinit/gt215.c 			if (part >= hweight8((init->r001540 >> 16) & 0xff))
part              112 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgf100.c 	u32 part = nvkm_rd32(device, 0x022438), i;
part              119 drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgf100.c 	for (i = 0; (magic & 0x80000000) && i < part; addr += 0x1000, i++) {
part              249 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c 	if (ret = -ENOENT, !vma || vma->addr != addr || vma->part) {
part              251 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c 			  vma ? vma->addr : ~0ULL, vma ? vma->part : 0);
part              651 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			u64 part = ALIGN(addr, next) - addr;
part              652 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			if (size - part >= next)
part              653 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 				block = (part >> page[i].shift) << page[i].shift;
part              776 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	new->part = vma->part;
part              934 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		vma->part = true;
part              944 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		tmp->part = true;
part              961 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	       vma->part ? 'P' : '-',
part             1157 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	if (vma->addr == addr && vma->part && (prev = node(vma, prev))) {
part             1163 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		if (!next->part ||
part             1343 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	if (vma->part && (prev = node(vma, prev)) && prev->mapped)
part             1345 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	if ((next = node(vma, next)) && (!next->part || next->mapped))
part             1546 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	BUG_ON(vma->part);
part             1557 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 			while ((next = node(next, next)) && next->part &&
part             1576 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		} while (next && next->part);
part             1587 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	} while ((next = node(vma, next)) && next->part);
part             1786 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 	if (inst && vmm && vmm->func->part) {
part             1788 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c 		vmm->func->part(vmm, inst);
part              141 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h 	void (*part)(struct nvkm_vmm *, struct nvkm_memory *inst);
part              373 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c 	.part = gf100_vmm_part,
part              388 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c 	.part = gf100_vmm_part,
part               70 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgk104.c 	.part = gf100_vmm_part,
part               85 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgk104.c 	.part = gf100_vmm_part,
part               39 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgk20a.c 	.part = gf100_vmm_part,
part               54 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgk20a.c 	.part = gf100_vmm_part,
part              112 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c 	.part = gf100_vmm_part,
part              128 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c 	.part = gf100_vmm_part,
part               27 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm20b.c 	.part = gf100_vmm_part,
part               43 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm20b.c 	.part = gf100_vmm_part,
part              490 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c 	.part = gf100_vmm_part,
part               27 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp10b.c 	.part = gf100_vmm_part,
part               65 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgv100.c 	.part = gf100_vmm_part,
part               27 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmmcp77.c 	.part = nv50_vmm_part,
part              367 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c 	.part = nv50_vmm_part,
part               55 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmtu102.c 	.part = gf100_vmm_part,
part              225 drivers/i3c/device.c 		u16 part = I3C_PID_PART_ID(devinfo.pid);
part              234 drivers/i3c/device.c 			if (manuf != id->manuf_id || part != id->part_id)
part              258 drivers/i3c/master.c 	u16 manuf, part, ext;
part              262 drivers/i3c/master.c 	part = I3C_PID_PART_ID(devinfo.pid);
part              271 drivers/i3c/master.c 			      devinfo.dcr, manuf, part, ext);
part              911 drivers/ide/ide-probe.c static struct kobject *ata_probe(dev_t dev, int *part, void *data)
part              914 drivers/ide/ide-probe.c 	int unit = *part >> PARTN_BITS;
part              932 drivers/ide/ide-probe.c static struct kobject *exact_match(dev_t dev, int *part, void *data)
part              935 drivers/ide/ide-probe.c 	*part &= (1 << PARTN_BITS) - 1;
part              153 drivers/iio/light/si1145.c 	u8 part;
part             1305 drivers/iio/light/si1145.c 	if (part_id != data->part_info->part) {
part             1307 drivers/iio/light/si1145.c 				part_id, data->part_info->part);
part               84 drivers/infiniband/sw/siw/siw_qp_tx.c 				unsigned long part = bytes - (PAGE_SIZE - off);
part               86 drivers/infiniband/sw/siw/siw_qp_tx.c 				memcpy(paddr, buffer + off, part);
part               91 drivers/infiniband/sw/siw/siw_qp_tx.c 							  sge->laddr + part);
part               94 drivers/infiniband/sw/siw/siw_qp_tx.c 							    sge->laddr + part,
part              100 drivers/infiniband/sw/siw/siw_qp_tx.c 				memcpy(paddr + part, buffer, bytes - part);
part             1631 drivers/irqchip/irq-gic-v3.c 		struct partition_affinity *part;
part             1634 drivers/irqchip/irq-gic-v3.c 		part = &parts[part_idx];
part             1636 drivers/irqchip/irq-gic-v3.c 		part->partition_id = of_node_to_fwnode(child_part);
part             1665 drivers/irqchip/irq-gic-v3.c 			cpumask_set_cpu(cpu, &part->mask);
part               25 drivers/irqchip/irq-partition-percpu.c static bool partition_check_cpu(struct partition_desc *part,
part               28 drivers/irqchip/irq-partition-percpu.c 	return cpumask_test_cpu(cpu, &part->parts[hwirq].mask);
part               33 drivers/irqchip/irq-partition-percpu.c 	struct partition_desc *part = irq_data_get_irq_chip_data(d);
part               34 drivers/irqchip/irq-partition-percpu.c 	struct irq_chip *chip = irq_desc_get_chip(part->chained_desc);
part               35 drivers/irqchip/irq-partition-percpu.c 	struct irq_data *data = irq_desc_get_irq_data(part->chained_desc);
part               37 drivers/irqchip/irq-partition-percpu.c 	if (partition_check_cpu(part, smp_processor_id(), d->hwirq) &&
part               44 drivers/irqchip/irq-partition-percpu.c 	struct partition_desc *part = irq_data_get_irq_chip_data(d);
part               45 drivers/irqchip/irq-partition-percpu.c 	struct irq_chip *chip = irq_desc_get_chip(part->chained_desc);
part               46 drivers/irqchip/irq-partition-percpu.c 	struct irq_data *data = irq_desc_get_irq_data(part->chained_desc);
part               48 drivers/irqchip/irq-partition-percpu.c 	if (partition_check_cpu(part, smp_processor_id(), d->hwirq) &&
part               57 drivers/irqchip/irq-partition-percpu.c 	struct partition_desc *part = irq_data_get_irq_chip_data(d);
part               58 drivers/irqchip/irq-partition-percpu.c 	struct irq_chip *chip = irq_desc_get_chip(part->chained_desc);
part               59 drivers/irqchip/irq-partition-percpu.c 	struct irq_data *data = irq_desc_get_irq_data(part->chained_desc);
part               61 drivers/irqchip/irq-partition-percpu.c 	if (partition_check_cpu(part, smp_processor_id(), d->hwirq) &&
part               72 drivers/irqchip/irq-partition-percpu.c 	struct partition_desc *part = irq_data_get_irq_chip_data(d);
part               73 drivers/irqchip/irq-partition-percpu.c 	struct irq_chip *chip = irq_desc_get_chip(part->chained_desc);
part               74 drivers/irqchip/irq-partition-percpu.c 	struct irq_data *data = irq_desc_get_irq_data(part->chained_desc);
part               76 drivers/irqchip/irq-partition-percpu.c 	if (partition_check_cpu(part, smp_processor_id(), d->hwirq) &&
part               85 drivers/irqchip/irq-partition-percpu.c 	struct partition_desc *part = irq_data_get_irq_chip_data(d);
part               86 drivers/irqchip/irq-partition-percpu.c 	struct irq_chip *chip = irq_desc_get_chip(part->chained_desc);
part               87 drivers/irqchip/irq-partition-percpu.c 	struct irq_data *data = irq_desc_get_irq_data(part->chained_desc);
part               97 drivers/irqchip/irq-partition-percpu.c 	struct partition_desc *part = irq_data_get_irq_chip_data(d);
part               98 drivers/irqchip/irq-partition-percpu.c 	struct irq_chip *chip = irq_desc_get_chip(part->chained_desc);
part               99 drivers/irqchip/irq-partition-percpu.c 	struct irq_data *data = irq_desc_get_irq_data(part->chained_desc);
part              115 drivers/irqchip/irq-partition-percpu.c 	struct partition_desc *part = irq_desc_get_handler_data(desc);
part              122 drivers/irqchip/irq-partition-percpu.c 	for_each_set_bit(hwirq, part->bitmap, part->nr_parts) {
part              123 drivers/irqchip/irq-partition-percpu.c 		if (partition_check_cpu(part, cpu, hwirq))
part              127 drivers/irqchip/irq-partition-percpu.c 	if (unlikely(hwirq == part->nr_parts)) {
part              131 drivers/irqchip/irq-partition-percpu.c 		irq = irq_find_mapping(part->domain, hwirq);
part              145 drivers/irqchip/irq-partition-percpu.c 	struct partition_desc *part;
part              152 drivers/irqchip/irq-partition-percpu.c 	part = domain->host_data;
part              154 drivers/irqchip/irq-partition-percpu.c 	set_bit(hwirq, part->bitmap);
part              155 drivers/irqchip/irq-partition-percpu.c 	irq_set_chained_handler_and_data(irq_desc_get_irq(part->chained_desc),
part              156 drivers/irqchip/irq-partition-percpu.c 					 partition_handle_irq, part);
part              157 drivers/irqchip/irq-partition-percpu.c 	irq_set_percpu_devid_partition(virq, &part->parts[hwirq].mask);
part              158 drivers/irqchip/irq-partition-percpu.c 	irq_domain_set_info(domain, virq, hwirq, &partition_irq_chip, part,
part              179 drivers/irqchip/irq-partition-percpu.c 	struct partition_affinity *part = NULL;
part              184 drivers/irqchip/irq-partition-percpu.c 			part = &desc->parts[i];
part              189 drivers/irqchip/irq-partition-percpu.c 	if (WARN_ON(!part)) {
part             1022 drivers/macintosh/smu.c 	const struct smu_sdbp_header *part;
part             1039 drivers/macintosh/smu.c 	part = of_get_property(smu->of_node, pname, size);
part             1040 drivers/macintosh/smu.c 	if (part == NULL) {
part             1042 drivers/macintosh/smu.c 		part = smu_create_sdb_partition(id);
part             1043 drivers/macintosh/smu.c 		if (part != NULL && size)
part             1044 drivers/macintosh/smu.c 			*size = part->len << 2;
part             1047 drivers/macintosh/smu.c 	return part;
part             1130 drivers/macintosh/smu.c 		const struct smu_sdbp_header *part;
part             1131 drivers/macintosh/smu.c 		part = __smu_get_sdb_partition(hdr.cmd, NULL, 1);
part             1132 drivers/macintosh/smu.c 		if (part == NULL)
part             1134 drivers/macintosh/smu.c 		else if (IS_ERR(part))
part             1135 drivers/macintosh/smu.c 			return PTR_ERR(part);
part              632 drivers/md/dm.c 	struct hd_struct *part = &dm_disk(md)->part0;
part              636 drivers/md/dm.c 		sum += part_stat_local_read_cpu(part, in_flight[0], cpu);
part              637 drivers/md/dm.c 		sum += part_stat_local_read_cpu(part, in_flight[1], cpu);
part             1601 drivers/md/dm.c #define __dm_part_stat_sub(part, field, subnd)	\
part             1602 drivers/md/dm.c 	(part_stat_get(part, field) -= (subnd))
part               70 drivers/md/md.c static void autostart_arrays(int part);
part             5542 drivers/md/md.c static struct kobject *md_probe(dev_t dev, int *part, void *data)
part             6269 drivers/md/md.c static void autorun_devices(int part)
part             6296 drivers/md/md.c 		if (part) {
part             9556 drivers/md/md.c static void autostart_arrays(int part)
part             9593 drivers/md/md.c 	autorun_devices(part);
part              509 drivers/misc/bh1770glc.c 	u8 manu, part;
part              519 drivers/misc/bh1770glc.c 	part = (u8)ret;
part              521 drivers/misc/bh1770glc.c 	chip->revision = (part & BH1770_REV_MASK) >> BH1770_REV_SHIFT;
part              527 drivers/misc/bh1770glc.c 	    ((part & BH1770_PART_MASK) == BH1770_PART)) {
part              533 drivers/misc/bh1770glc.c 	    ((part & BH1770_PART_MASK) == BH1770_PART)) {
part              664 drivers/misc/sgi-xp/xpc.h xpc_wakeup_channel_mgr(struct xpc_partition *part)
part              666 drivers/misc/sgi-xp/xpc.h 	if (atomic_inc_return(&part->channel_mgr_requests) == 1)
part              667 drivers/misc/sgi-xp/xpc.h 		wake_up(&part->channel_mgr_wq);
part              698 drivers/misc/sgi-xp/xpc.h xpc_part_deref(struct xpc_partition *part)
part              700 drivers/misc/sgi-xp/xpc.h 	s32 refs = atomic_dec_return(&part->references);
part              703 drivers/misc/sgi-xp/xpc.h 	if (refs == 0 && part->setup_state == XPC_P_SS_WTEARDOWN)
part              704 drivers/misc/sgi-xp/xpc.h 		wake_up(&part->teardown_wq);
part              708 drivers/misc/sgi-xp/xpc.h xpc_part_ref(struct xpc_partition *part)
part              712 drivers/misc/sgi-xp/xpc.h 	atomic_inc(&part->references);
part              713 drivers/misc/sgi-xp/xpc.h 	setup = (part->setup_state == XPC_P_SS_SETUP);
part              715 drivers/misc/sgi-xp/xpc.h 		xpc_part_deref(part);
part               82 drivers/misc/sgi-xp/xpc_channel.c 	struct xpc_partition *part = &xpc_partitions[ch->partid];
part              101 drivers/misc/sgi-xp/xpc_channel.c 	if (part->act_state == XPC_P_AS_DEACTIVATING) {
part              157 drivers/misc/sgi-xp/xpc_channel.c 	atomic_dec(&part->nchannels_active);
part              168 drivers/misc/sgi-xp/xpc_channel.c 		if (part->act_state != XPC_P_AS_DEACTIVATING) {
part              170 drivers/misc/sgi-xp/xpc_channel.c 			spin_lock(&part->chctl_lock);
part              171 drivers/misc/sgi-xp/xpc_channel.c 			part->chctl.flags[ch->number] |=
part              173 drivers/misc/sgi-xp/xpc_channel.c 			spin_unlock(&part->chctl_lock);
part              183 drivers/misc/sgi-xp/xpc_channel.c xpc_process_openclose_chctl_flags(struct xpc_partition *part, int ch_number,
part              188 drivers/misc/sgi-xp/xpc_channel.c 	    &part->remote_openclose_args[ch_number];
part              189 drivers/misc/sgi-xp/xpc_channel.c 	struct xpc_channel *ch = &part->channels[ch_number];
part              238 drivers/misc/sgi-xp/xpc_channel.c 				if (part->chctl.flags[ch_number] &
part              242 drivers/misc/sgi-xp/xpc_channel.c 					spin_lock(&part->chctl_lock);
part              243 drivers/misc/sgi-xp/xpc_channel.c 					part->chctl.flags[ch_number] |=
part              245 drivers/misc/sgi-xp/xpc_channel.c 					spin_unlock(&part->chctl_lock);
part              253 drivers/misc/sgi-xp/xpc_channel.c 			atomic_inc(&part->nchannels_active);
part              289 drivers/misc/sgi-xp/xpc_channel.c 			DBUG_ON(part->act_state != XPC_P_AS_DEACTIVATING);
part              296 drivers/misc/sgi-xp/xpc_channel.c 			if (part->chctl.flags[ch_number] &
part              300 drivers/misc/sgi-xp/xpc_channel.c 				spin_lock(&part->chctl_lock);
part              301 drivers/misc/sgi-xp/xpc_channel.c 				part->chctl.flags[ch_number] |=
part              303 drivers/misc/sgi-xp/xpc_channel.c 				spin_unlock(&part->chctl_lock);
part              323 drivers/misc/sgi-xp/xpc_channel.c 		if (part->act_state == XPC_P_AS_DEACTIVATING ||
part              362 drivers/misc/sgi-xp/xpc_channel.c 			atomic_inc(&part->nchannels_active);
part              546 drivers/misc/sgi-xp/xpc_channel.c xpc_process_sent_chctl_flags(struct xpc_partition *part)
part              554 drivers/misc/sgi-xp/xpc_channel.c 	chctl.all_flags = xpc_arch_ops.get_chctl_all_flags(part);
part              563 drivers/misc/sgi-xp/xpc_channel.c 	for (ch_number = 0; ch_number < part->nchannels; ch_number++) {
part              564 drivers/misc/sgi-xp/xpc_channel.c 		ch = &part->channels[ch_number];
part              572 drivers/misc/sgi-xp/xpc_channel.c 			xpc_process_openclose_chctl_flags(part, ch_number,
part              585 drivers/misc/sgi-xp/xpc_channel.c 		if (part->act_state == XPC_P_AS_DEACTIVATING)
part              603 drivers/misc/sgi-xp/xpc_channel.c 			xpc_arch_ops.process_msg_chctl_flags(part, ch_number);
part              617 drivers/misc/sgi-xp/xpc_channel.c xpc_partition_going_down(struct xpc_partition *part, enum xp_retval reason)
part              624 drivers/misc/sgi-xp/xpc_channel.c 		XPC_PARTID(part), reason);
part              626 drivers/misc/sgi-xp/xpc_channel.c 	if (!xpc_part_ref(part)) {
part              633 drivers/misc/sgi-xp/xpc_channel.c 	for (ch_number = 0; ch_number < part->nchannels; ch_number++) {
part              634 drivers/misc/sgi-xp/xpc_channel.c 		ch = &part->channels[ch_number];
part              645 drivers/misc/sgi-xp/xpc_channel.c 	xpc_wakeup_channel_mgr(part);
part              647 drivers/misc/sgi-xp/xpc_channel.c 	xpc_part_deref(part);
part              658 drivers/misc/sgi-xp/xpc_channel.c 	struct xpc_partition *part;
part              663 drivers/misc/sgi-xp/xpc_channel.c 		part = &xpc_partitions[partid];
part              665 drivers/misc/sgi-xp/xpc_channel.c 		if (xpc_part_ref(part)) {
part              670 drivers/misc/sgi-xp/xpc_channel.c 			xpc_wakeup_channel_mgr(part);
part              671 drivers/misc/sgi-xp/xpc_channel.c 			xpc_part_deref(part);
part              711 drivers/misc/sgi-xp/xpc_channel.c 	struct xpc_partition *part;
part              718 drivers/misc/sgi-xp/xpc_channel.c 		part = &xpc_partitions[partid];
part              720 drivers/misc/sgi-xp/xpc_channel.c 		if (xpc_part_ref(part)) {
part              721 drivers/misc/sgi-xp/xpc_channel.c 			ch = &part->channels[ch_number];
part              736 drivers/misc/sgi-xp/xpc_channel.c 			xpc_part_deref(part);
part              875 drivers/misc/sgi-xp/xpc_channel.c 	struct xpc_partition *part = &xpc_partitions[partid];
part              882 drivers/misc/sgi-xp/xpc_channel.c 	DBUG_ON(ch_number < 0 || ch_number >= part->nchannels);
part              885 drivers/misc/sgi-xp/xpc_channel.c 	if (xpc_part_ref(part)) {
part              886 drivers/misc/sgi-xp/xpc_channel.c 		ret = xpc_arch_ops.send_payload(&part->channels[ch_number],
part              888 drivers/misc/sgi-xp/xpc_channel.c 		xpc_part_deref(part);
part              926 drivers/misc/sgi-xp/xpc_channel.c 	struct xpc_partition *part = &xpc_partitions[partid];
part              933 drivers/misc/sgi-xp/xpc_channel.c 	DBUG_ON(ch_number < 0 || ch_number >= part->nchannels);
part              937 drivers/misc/sgi-xp/xpc_channel.c 	if (xpc_part_ref(part)) {
part              938 drivers/misc/sgi-xp/xpc_channel.c 		ret = xpc_arch_ops.send_payload(&part->channels[ch_number],
part              940 drivers/misc/sgi-xp/xpc_channel.c 		xpc_part_deref(part);
part             1000 drivers/misc/sgi-xp/xpc_channel.c 	struct xpc_partition *part = &xpc_partitions[partid];
part             1004 drivers/misc/sgi-xp/xpc_channel.c 	DBUG_ON(ch_number < 0 || ch_number >= part->nchannels);
part             1006 drivers/misc/sgi-xp/xpc_channel.c 	ch = &part->channels[ch_number];
part              177 drivers/misc/sgi-xp/xpc_main.c 	struct xpc_partition *part = from_timer(part, t, disengage_timer);
part              179 drivers/misc/sgi-xp/xpc_main.c 	DBUG_ON(time_is_after_jiffies(part->disengage_timeout));
part              181 drivers/misc/sgi-xp/xpc_main.c 	(void)xpc_partition_disengaged(part);
part              183 drivers/misc/sgi-xp/xpc_main.c 	DBUG_ON(part->disengage_timeout != 0);
part              184 drivers/misc/sgi-xp/xpc_main.c 	DBUG_ON(xpc_arch_ops.partition_engaged(XPC_PARTID(part)));
part              226 drivers/misc/sgi-xp/xpc_main.c 	struct xpc_partition *part;
part              238 drivers/misc/sgi-xp/xpc_main.c 		part = &xpc_partitions[partid];
part              240 drivers/misc/sgi-xp/xpc_main.c 		if (part->act_state == XPC_P_AS_INACTIVE ||
part              241 drivers/misc/sgi-xp/xpc_main.c 		    part->act_state == XPC_P_AS_DEACTIVATING) {
part              245 drivers/misc/sgi-xp/xpc_main.c 		ret = xpc_arch_ops.get_remote_heartbeat(part);
part              247 drivers/misc/sgi-xp/xpc_main.c 			XPC_DEACTIVATE_PARTITION(part, ret);
part              339 drivers/misc/sgi-xp/xpc_main.c xpc_channel_mgr(struct xpc_partition *part)
part              341 drivers/misc/sgi-xp/xpc_main.c 	while (part->act_state != XPC_P_AS_DEACTIVATING ||
part              342 drivers/misc/sgi-xp/xpc_main.c 	       atomic_read(&part->nchannels_active) > 0 ||
part              343 drivers/misc/sgi-xp/xpc_main.c 	       !xpc_partition_disengaged(part)) {
part              345 drivers/misc/sgi-xp/xpc_main.c 		xpc_process_sent_chctl_flags(part);
part              360 drivers/misc/sgi-xp/xpc_main.c 		atomic_dec(&part->channel_mgr_requests);
part              361 drivers/misc/sgi-xp/xpc_main.c 		(void)wait_event_interruptible(part->channel_mgr_wq,
part              362 drivers/misc/sgi-xp/xpc_main.c 				(atomic_read(&part->channel_mgr_requests) > 0 ||
part              363 drivers/misc/sgi-xp/xpc_main.c 				 part->chctl.all_flags != 0 ||
part              364 drivers/misc/sgi-xp/xpc_main.c 				 (part->act_state == XPC_P_AS_DEACTIVATING &&
part              365 drivers/misc/sgi-xp/xpc_main.c 				 atomic_read(&part->nchannels_active) == 0 &&
part              366 drivers/misc/sgi-xp/xpc_main.c 				 xpc_partition_disengaged(part))));
part              367 drivers/misc/sgi-xp/xpc_main.c 		atomic_set(&part->channel_mgr_requests, 1);
part              400 drivers/misc/sgi-xp/xpc_main.c xpc_setup_ch_structures(struct xpc_partition *part)
part              405 drivers/misc/sgi-xp/xpc_main.c 	short partid = XPC_PARTID(part);
part              411 drivers/misc/sgi-xp/xpc_main.c 	DBUG_ON(part->channels != NULL);
part              412 drivers/misc/sgi-xp/xpc_main.c 	part->channels = kcalloc(XPC_MAX_NCHANNELS,
part              415 drivers/misc/sgi-xp/xpc_main.c 	if (part->channels == NULL) {
part              422 drivers/misc/sgi-xp/xpc_main.c 	part->remote_openclose_args =
part              424 drivers/misc/sgi-xp/xpc_main.c 					  GFP_KERNEL, &part->
part              426 drivers/misc/sgi-xp/xpc_main.c 	if (part->remote_openclose_args == NULL) {
part              432 drivers/misc/sgi-xp/xpc_main.c 	part->chctl.all_flags = 0;
part              433 drivers/misc/sgi-xp/xpc_main.c 	spin_lock_init(&part->chctl_lock);
part              435 drivers/misc/sgi-xp/xpc_main.c 	atomic_set(&part->channel_mgr_requests, 1);
part              436 drivers/misc/sgi-xp/xpc_main.c 	init_waitqueue_head(&part->channel_mgr_wq);
part              438 drivers/misc/sgi-xp/xpc_main.c 	part->nchannels = XPC_MAX_NCHANNELS;
part              440 drivers/misc/sgi-xp/xpc_main.c 	atomic_set(&part->nchannels_active, 0);
part              441 drivers/misc/sgi-xp/xpc_main.c 	atomic_set(&part->nchannels_engaged, 0);
part              443 drivers/misc/sgi-xp/xpc_main.c 	for (ch_number = 0; ch_number < part->nchannels; ch_number++) {
part              444 drivers/misc/sgi-xp/xpc_main.c 		ch = &part->channels[ch_number];
part              465 drivers/misc/sgi-xp/xpc_main.c 	ret = xpc_arch_ops.setup_ch_structures(part);
part              473 drivers/misc/sgi-xp/xpc_main.c 	part->setup_state = XPC_P_SS_SETUP;
part              479 drivers/misc/sgi-xp/xpc_main.c 	kfree(part->remote_openclose_args_base);
part              480 drivers/misc/sgi-xp/xpc_main.c 	part->remote_openclose_args = NULL;
part              482 drivers/misc/sgi-xp/xpc_main.c 	kfree(part->channels);
part              483 drivers/misc/sgi-xp/xpc_main.c 	part->channels = NULL;
part              492 drivers/misc/sgi-xp/xpc_main.c xpc_teardown_ch_structures(struct xpc_partition *part)
part              494 drivers/misc/sgi-xp/xpc_main.c 	DBUG_ON(atomic_read(&part->nchannels_engaged) != 0);
part              495 drivers/misc/sgi-xp/xpc_main.c 	DBUG_ON(atomic_read(&part->nchannels_active) != 0);
part              502 drivers/misc/sgi-xp/xpc_main.c 	DBUG_ON(part->setup_state != XPC_P_SS_SETUP);
part              503 drivers/misc/sgi-xp/xpc_main.c 	part->setup_state = XPC_P_SS_WTEARDOWN;
part              505 drivers/misc/sgi-xp/xpc_main.c 	wait_event(part->teardown_wq, (atomic_read(&part->references) == 0));
part              509 drivers/misc/sgi-xp/xpc_main.c 	xpc_arch_ops.teardown_ch_structures(part);
part              511 drivers/misc/sgi-xp/xpc_main.c 	kfree(part->remote_openclose_args_base);
part              512 drivers/misc/sgi-xp/xpc_main.c 	part->remote_openclose_args = NULL;
part              513 drivers/misc/sgi-xp/xpc_main.c 	kfree(part->channels);
part              514 drivers/misc/sgi-xp/xpc_main.c 	part->channels = NULL;
part              516 drivers/misc/sgi-xp/xpc_main.c 	part->setup_state = XPC_P_SS_TORNDOWN;
part              534 drivers/misc/sgi-xp/xpc_main.c 	struct xpc_partition *part = &xpc_partitions[partid];
part              539 drivers/misc/sgi-xp/xpc_main.c 	spin_lock_irqsave(&part->act_lock, irq_flags);
part              541 drivers/misc/sgi-xp/xpc_main.c 	if (part->act_state == XPC_P_AS_DEACTIVATING) {
part              542 drivers/misc/sgi-xp/xpc_main.c 		part->act_state = XPC_P_AS_INACTIVE;
part              543 drivers/misc/sgi-xp/xpc_main.c 		spin_unlock_irqrestore(&part->act_lock, irq_flags);
part              544 drivers/misc/sgi-xp/xpc_main.c 		part->remote_rp_pa = 0;
part              549 drivers/misc/sgi-xp/xpc_main.c 	DBUG_ON(part->act_state != XPC_P_AS_ACTIVATION_REQ);
part              550 drivers/misc/sgi-xp/xpc_main.c 	part->act_state = XPC_P_AS_ACTIVATING;
part              552 drivers/misc/sgi-xp/xpc_main.c 	XPC_SET_REASON(part, 0, 0);
part              553 drivers/misc/sgi-xp/xpc_main.c 	spin_unlock_irqrestore(&part->act_lock, irq_flags);
part              559 drivers/misc/sgi-xp/xpc_main.c 	if (xpc_setup_ch_structures(part) == xpSuccess) {
part              560 drivers/misc/sgi-xp/xpc_main.c 		(void)xpc_part_ref(part);	/* this will always succeed */
part              562 drivers/misc/sgi-xp/xpc_main.c 		if (xpc_arch_ops.make_first_contact(part) == xpSuccess) {
part              563 drivers/misc/sgi-xp/xpc_main.c 			xpc_mark_partition_active(part);
part              564 drivers/misc/sgi-xp/xpc_main.c 			xpc_channel_mgr(part);
part              568 drivers/misc/sgi-xp/xpc_main.c 		xpc_part_deref(part);
part              569 drivers/misc/sgi-xp/xpc_main.c 		xpc_teardown_ch_structures(part);
part              573 drivers/misc/sgi-xp/xpc_main.c 	xpc_mark_partition_inactive(part);
part              575 drivers/misc/sgi-xp/xpc_main.c 	if (part->reason == xpReactivating) {
part              577 drivers/misc/sgi-xp/xpc_main.c 		xpc_arch_ops.request_partition_reactivation(part);
part              584 drivers/misc/sgi-xp/xpc_main.c xpc_activate_partition(struct xpc_partition *part)
part              586 drivers/misc/sgi-xp/xpc_main.c 	short partid = XPC_PARTID(part);
part              590 drivers/misc/sgi-xp/xpc_main.c 	spin_lock_irqsave(&part->act_lock, irq_flags);
part              592 drivers/misc/sgi-xp/xpc_main.c 	DBUG_ON(part->act_state != XPC_P_AS_INACTIVE);
part              594 drivers/misc/sgi-xp/xpc_main.c 	part->act_state = XPC_P_AS_ACTIVATION_REQ;
part              595 drivers/misc/sgi-xp/xpc_main.c 	XPC_SET_REASON(part, xpCloneKThread, __LINE__);
part              597 drivers/misc/sgi-xp/xpc_main.c 	spin_unlock_irqrestore(&part->act_lock, irq_flags);
part              602 drivers/misc/sgi-xp/xpc_main.c 		spin_lock_irqsave(&part->act_lock, irq_flags);
part              603 drivers/misc/sgi-xp/xpc_main.c 		part->act_state = XPC_P_AS_INACTIVE;
part              604 drivers/misc/sgi-xp/xpc_main.c 		XPC_SET_REASON(part, xpCloneKThreadFailed, __LINE__);
part              605 drivers/misc/sgi-xp/xpc_main.c 		spin_unlock_irqrestore(&part->act_lock, irq_flags);
part              648 drivers/misc/sgi-xp/xpc_main.c xpc_kthread_waitmsgs(struct xpc_partition *part, struct xpc_channel *ch)
part              685 drivers/misc/sgi-xp/xpc_main.c 	struct xpc_partition *part = &xpc_partitions[partid];
part              695 drivers/misc/sgi-xp/xpc_main.c 	ch = &part->channels[ch_number];
part              727 drivers/misc/sgi-xp/xpc_main.c 		xpc_kthread_waitmsgs(part, ch);
part              746 drivers/misc/sgi-xp/xpc_main.c 	    atomic_dec_return(&part->nchannels_engaged) == 0) {
part              747 drivers/misc/sgi-xp/xpc_main.c 		xpc_arch_ops.indicate_partition_disengaged(part);
part              755 drivers/misc/sgi-xp/xpc_main.c 	xpc_part_deref(part);
part              777 drivers/misc/sgi-xp/xpc_main.c 	struct xpc_partition *part = &xpc_partitions[ch->partid];
part              801 drivers/misc/sgi-xp/xpc_main.c 			   atomic_inc_return(&part->nchannels_engaged) == 1) {
part              802 drivers/misc/sgi-xp/xpc_main.c 			xpc_arch_ops.indicate_partition_engaged(part);
part              804 drivers/misc/sgi-xp/xpc_main.c 		(void)xpc_part_ref(part);
part              823 drivers/misc/sgi-xp/xpc_main.c 			    atomic_dec_return(&part->nchannels_engaged) == 0) {
part              824 drivers/misc/sgi-xp/xpc_main.c 				indicate_partition_disengaged(part);
part              827 drivers/misc/sgi-xp/xpc_main.c 			xpc_part_deref(part);
part              851 drivers/misc/sgi-xp/xpc_main.c 	struct xpc_partition *part;
part              857 drivers/misc/sgi-xp/xpc_main.c 		part = &xpc_partitions[partid];
part              859 drivers/misc/sgi-xp/xpc_main.c 		if (!xpc_part_ref(part))
part              862 drivers/misc/sgi-xp/xpc_main.c 		ch = &part->channels[ch_number];
part              865 drivers/misc/sgi-xp/xpc_main.c 			xpc_part_deref(part);
part              876 drivers/misc/sgi-xp/xpc_main.c 			if (part->act_state != XPC_P_AS_DEACTIVATING) {
part              877 drivers/misc/sgi-xp/xpc_main.c 				spin_lock(&part->chctl_lock);
part              878 drivers/misc/sgi-xp/xpc_main.c 				part->chctl.flags[ch->number] |=
part              880 drivers/misc/sgi-xp/xpc_main.c 				spin_unlock(&part->chctl_lock);
part              890 drivers/misc/sgi-xp/xpc_main.c 			xpc_wakeup_channel_mgr(part);
part              892 drivers/misc/sgi-xp/xpc_main.c 		xpc_part_deref(part);
part              900 drivers/misc/sgi-xp/xpc_main.c 	struct xpc_partition *part;
part              919 drivers/misc/sgi-xp/xpc_main.c 		part = &xpc_partitions[partid];
part              921 drivers/misc/sgi-xp/xpc_main.c 		DBUG_ON((u64)part != L1_CACHE_ALIGN((u64)part));
part              923 drivers/misc/sgi-xp/xpc_main.c 		part->activate_IRQ_rcvd = 0;
part              924 drivers/misc/sgi-xp/xpc_main.c 		spin_lock_init(&part->act_lock);
part              925 drivers/misc/sgi-xp/xpc_main.c 		part->act_state = XPC_P_AS_INACTIVE;
part              926 drivers/misc/sgi-xp/xpc_main.c 		XPC_SET_REASON(part, 0, 0);
part              928 drivers/misc/sgi-xp/xpc_main.c 		timer_setup(&part->disengage_timer,
part              931 drivers/misc/sgi-xp/xpc_main.c 		part->setup_state = XPC_P_SS_UNSET;
part              932 drivers/misc/sgi-xp/xpc_main.c 		init_waitqueue_head(&part->teardown_wq);
part              933 drivers/misc/sgi-xp/xpc_main.c 		atomic_set(&part->references, 0);
part              951 drivers/misc/sgi-xp/xpc_main.c 	struct xpc_partition *part;
part              983 drivers/misc/sgi-xp/xpc_main.c 			part = &xpc_partitions[partid];
part              985 drivers/misc/sgi-xp/xpc_main.c 			if (xpc_partition_disengaged(part) &&
part              986 drivers/misc/sgi-xp/xpc_main.c 			    part->act_state == XPC_P_AS_INACTIVE) {
part              992 drivers/misc/sgi-xp/xpc_main.c 			XPC_DEACTIVATE_PARTITION(part, reason);
part              994 drivers/misc/sgi-xp/xpc_main.c 			if (part->disengage_timeout > disengage_timeout)
part              995 drivers/misc/sgi-xp/xpc_main.c 				disengage_timeout = part->disengage_timeout;
part             1086 drivers/misc/sgi-xp/xpc_main.c 	struct xpc_partition *part;
part             1101 drivers/misc/sgi-xp/xpc_main.c 		part = &xpc_partitions[partid];
part             1104 drivers/misc/sgi-xp/xpc_main.c 		    part->act_state != XPC_P_AS_INACTIVE) {
part             1105 drivers/misc/sgi-xp/xpc_main.c 			xpc_arch_ops.request_partition_deactivation(part);
part             1106 drivers/misc/sgi-xp/xpc_main.c 			xpc_arch_ops.indicate_partition_disengaged(part);
part              265 drivers/misc/sgi-xp/xpc_partition.c xpc_partition_disengaged(struct xpc_partition *part)
part              267 drivers/misc/sgi-xp/xpc_partition.c 	short partid = XPC_PARTID(part);
part              271 drivers/misc/sgi-xp/xpc_partition.c 	if (part->disengage_timeout) {
part              273 drivers/misc/sgi-xp/xpc_partition.c 			if (time_is_after_jiffies(part->disengage_timeout)) {
part              289 drivers/misc/sgi-xp/xpc_partition.c 		part->disengage_timeout = 0;
part              293 drivers/misc/sgi-xp/xpc_partition.c 			del_singleshot_timer_sync(&part->disengage_timer);
part              295 drivers/misc/sgi-xp/xpc_partition.c 		DBUG_ON(part->act_state != XPC_P_AS_DEACTIVATING &&
part              296 drivers/misc/sgi-xp/xpc_partition.c 			part->act_state != XPC_P_AS_INACTIVE);
part              297 drivers/misc/sgi-xp/xpc_partition.c 		if (part->act_state != XPC_P_AS_INACTIVE)
part              298 drivers/misc/sgi-xp/xpc_partition.c 			xpc_wakeup_channel_mgr(part);
part              300 drivers/misc/sgi-xp/xpc_partition.c 		xpc_arch_ops.cancel_partition_deactivation_request(part);
part              309 drivers/misc/sgi-xp/xpc_partition.c xpc_mark_partition_active(struct xpc_partition *part)
part              314 drivers/misc/sgi-xp/xpc_partition.c 	dev_dbg(xpc_part, "setting partition %d to ACTIVE\n", XPC_PARTID(part));
part              316 drivers/misc/sgi-xp/xpc_partition.c 	spin_lock_irqsave(&part->act_lock, irq_flags);
part              317 drivers/misc/sgi-xp/xpc_partition.c 	if (part->act_state == XPC_P_AS_ACTIVATING) {
part              318 drivers/misc/sgi-xp/xpc_partition.c 		part->act_state = XPC_P_AS_ACTIVE;
part              321 drivers/misc/sgi-xp/xpc_partition.c 		DBUG_ON(part->reason == xpSuccess);
part              322 drivers/misc/sgi-xp/xpc_partition.c 		ret = part->reason;
part              324 drivers/misc/sgi-xp/xpc_partition.c 	spin_unlock_irqrestore(&part->act_lock, irq_flags);
part              333 drivers/misc/sgi-xp/xpc_partition.c xpc_deactivate_partition(const int line, struct xpc_partition *part,
part              338 drivers/misc/sgi-xp/xpc_partition.c 	spin_lock_irqsave(&part->act_lock, irq_flags);
part              340 drivers/misc/sgi-xp/xpc_partition.c 	if (part->act_state == XPC_P_AS_INACTIVE) {
part              341 drivers/misc/sgi-xp/xpc_partition.c 		XPC_SET_REASON(part, reason, line);
part              342 drivers/misc/sgi-xp/xpc_partition.c 		spin_unlock_irqrestore(&part->act_lock, irq_flags);
part              345 drivers/misc/sgi-xp/xpc_partition.c 			xpc_arch_ops.request_partition_reactivation(part);
part              349 drivers/misc/sgi-xp/xpc_partition.c 	if (part->act_state == XPC_P_AS_DEACTIVATING) {
part              350 drivers/misc/sgi-xp/xpc_partition.c 		if ((part->reason == xpUnloading && reason != xpUnloading) ||
part              352 drivers/misc/sgi-xp/xpc_partition.c 			XPC_SET_REASON(part, reason, line);
part              354 drivers/misc/sgi-xp/xpc_partition.c 		spin_unlock_irqrestore(&part->act_lock, irq_flags);
part              358 drivers/misc/sgi-xp/xpc_partition.c 	part->act_state = XPC_P_AS_DEACTIVATING;
part              359 drivers/misc/sgi-xp/xpc_partition.c 	XPC_SET_REASON(part, reason, line);
part              361 drivers/misc/sgi-xp/xpc_partition.c 	spin_unlock_irqrestore(&part->act_lock, irq_flags);
part              364 drivers/misc/sgi-xp/xpc_partition.c 	xpc_arch_ops.request_partition_deactivation(part);
part              367 drivers/misc/sgi-xp/xpc_partition.c 	part->disengage_timeout = jiffies + (xpc_disengage_timelimit * HZ);
part              368 drivers/misc/sgi-xp/xpc_partition.c 	part->disengage_timer.expires = part->disengage_timeout;
part              369 drivers/misc/sgi-xp/xpc_partition.c 	add_timer(&part->disengage_timer);
part              372 drivers/misc/sgi-xp/xpc_partition.c 		XPC_PARTID(part), reason);
part              374 drivers/misc/sgi-xp/xpc_partition.c 	xpc_partition_going_down(part, reason);
part              381 drivers/misc/sgi-xp/xpc_partition.c xpc_mark_partition_inactive(struct xpc_partition *part)
part              386 drivers/misc/sgi-xp/xpc_partition.c 		XPC_PARTID(part));
part              388 drivers/misc/sgi-xp/xpc_partition.c 	spin_lock_irqsave(&part->act_lock, irq_flags);
part              389 drivers/misc/sgi-xp/xpc_partition.c 	part->act_state = XPC_P_AS_INACTIVE;
part              390 drivers/misc/sgi-xp/xpc_partition.c 	spin_unlock_irqrestore(&part->act_lock, irq_flags);
part              391 drivers/misc/sgi-xp/xpc_partition.c 	part->remote_rp_pa = 0;
part              521 drivers/misc/sgi-xp/xpc_partition.c 	struct xpc_partition *part;
part              524 drivers/misc/sgi-xp/xpc_partition.c 	part = &xpc_partitions[partid];
part              525 drivers/misc/sgi-xp/xpc_partition.c 	if (part->remote_rp_pa == 0)
part              530 drivers/misc/sgi-xp/xpc_partition.c 	part_nasid_pa = (unsigned long)XPC_RP_PART_NASIDS(part->remote_rp_pa);
part              376 drivers/misc/sgi-xp/xpc_uv.c 	struct xpc_partition *part;
part              383 drivers/misc/sgi-xp/xpc_uv.c 		part = &xpc_partitions[partid];
part              385 drivers/misc/sgi-xp/xpc_uv.c 		if (part->sn.uv.act_state_req == 0)
part              391 drivers/misc/sgi-xp/xpc_uv.c 		act_state_req = part->sn.uv.act_state_req;
part              392 drivers/misc/sgi-xp/xpc_uv.c 		part->sn.uv.act_state_req = 0;
part              396 drivers/misc/sgi-xp/xpc_uv.c 			if (part->act_state == XPC_P_AS_INACTIVE)
part              397 drivers/misc/sgi-xp/xpc_uv.c 				xpc_activate_partition(part);
part              398 drivers/misc/sgi-xp/xpc_uv.c 			else if (part->act_state == XPC_P_AS_DEACTIVATING)
part              399 drivers/misc/sgi-xp/xpc_uv.c 				XPC_DEACTIVATE_PARTITION(part, xpReactivating);
part              402 drivers/misc/sgi-xp/xpc_uv.c 			if (part->act_state == XPC_P_AS_INACTIVE)
part              403 drivers/misc/sgi-xp/xpc_uv.c 				xpc_activate_partition(part);
part              405 drivers/misc/sgi-xp/xpc_uv.c 				XPC_DEACTIVATE_PARTITION(part, xpReactivating);
part              408 drivers/misc/sgi-xp/xpc_uv.c 			XPC_DEACTIVATE_PARTITION(part, part->sn.uv.reason);
part              423 drivers/misc/sgi-xp/xpc_uv.c xpc_handle_activate_mq_msg_uv(struct xpc_partition *part,
part              429 drivers/misc/sgi-xp/xpc_uv.c 	struct xpc_partition_uv *part_uv = &part->sn.uv;
part              454 drivers/misc/sgi-xp/xpc_uv.c 		part->remote_rp_pa = msg->rp_gpa; /* !!! _pa is _gpa */
part              455 drivers/misc/sgi-xp/xpc_uv.c 		part->remote_rp_ts_jiffies = msg_hdr->rp_ts_jiffies;
part              496 drivers/misc/sgi-xp/xpc_uv.c 		args = &part->remote_openclose_args[msg->ch_number];
part              499 drivers/misc/sgi-xp/xpc_uv.c 		spin_lock_irqsave(&part->chctl_lock, irq_flags);
part              500 drivers/misc/sgi-xp/xpc_uv.c 		part->chctl.flags[msg->ch_number] |= XPC_CHCTL_CLOSEREQUEST;
part              501 drivers/misc/sgi-xp/xpc_uv.c 		spin_unlock_irqrestore(&part->chctl_lock, irq_flags);
part              503 drivers/misc/sgi-xp/xpc_uv.c 		xpc_wakeup_channel_mgr(part);
part              516 drivers/misc/sgi-xp/xpc_uv.c 		spin_lock_irqsave(&part->chctl_lock, irq_flags);
part              517 drivers/misc/sgi-xp/xpc_uv.c 		part->chctl.flags[msg->ch_number] |= XPC_CHCTL_CLOSEREPLY;
part              518 drivers/misc/sgi-xp/xpc_uv.c 		spin_unlock_irqrestore(&part->chctl_lock, irq_flags);
part              520 drivers/misc/sgi-xp/xpc_uv.c 		xpc_wakeup_channel_mgr(part);
part              532 drivers/misc/sgi-xp/xpc_uv.c 		args = &part->remote_openclose_args[msg->ch_number];
part              536 drivers/misc/sgi-xp/xpc_uv.c 		spin_lock_irqsave(&part->chctl_lock, irq_flags);
part              537 drivers/misc/sgi-xp/xpc_uv.c 		part->chctl.flags[msg->ch_number] |= XPC_CHCTL_OPENREQUEST;
part              538 drivers/misc/sgi-xp/xpc_uv.c 		spin_unlock_irqrestore(&part->chctl_lock, irq_flags);
part              540 drivers/misc/sgi-xp/xpc_uv.c 		xpc_wakeup_channel_mgr(part);
part              551 drivers/misc/sgi-xp/xpc_uv.c 		args = &part->remote_openclose_args[msg->ch_number];
part              556 drivers/misc/sgi-xp/xpc_uv.c 		spin_lock_irqsave(&part->chctl_lock, irq_flags);
part              557 drivers/misc/sgi-xp/xpc_uv.c 		part->chctl.flags[msg->ch_number] |= XPC_CHCTL_OPENREPLY;
part              558 drivers/misc/sgi-xp/xpc_uv.c 		spin_unlock_irqrestore(&part->chctl_lock, irq_flags);
part              560 drivers/misc/sgi-xp/xpc_uv.c 		xpc_wakeup_channel_mgr(part);
part              571 drivers/misc/sgi-xp/xpc_uv.c 		spin_lock_irqsave(&part->chctl_lock, irq_flags);
part              572 drivers/misc/sgi-xp/xpc_uv.c 		part->chctl.flags[msg->ch_number] |= XPC_CHCTL_OPENCOMPLETE;
part              573 drivers/misc/sgi-xp/xpc_uv.c 		spin_unlock_irqrestore(&part->chctl_lock, irq_flags);
part              575 drivers/misc/sgi-xp/xpc_uv.c 		xpc_wakeup_channel_mgr(part);
part              592 drivers/misc/sgi-xp/xpc_uv.c 			"from partition=%d\n", msg_hdr->type, XPC_PARTID(part));
part              606 drivers/misc/sgi-xp/xpc_uv.c 	if (msg_hdr->rp_ts_jiffies != part->remote_rp_ts_jiffies &&
part              607 drivers/misc/sgi-xp/xpc_uv.c 	    part->remote_rp_ts_jiffies != 0) {
part              627 drivers/misc/sgi-xp/xpc_uv.c 	struct xpc_partition *part;
part              642 drivers/misc/sgi-xp/xpc_uv.c 			part = &xpc_partitions[partid];
part              644 drivers/misc/sgi-xp/xpc_uv.c 			part_referenced = xpc_part_ref(part);
part              645 drivers/misc/sgi-xp/xpc_uv.c 			xpc_handle_activate_mq_msg_uv(part, msg_hdr,
part              649 drivers/misc/sgi-xp/xpc_uv.c 				xpc_part_deref(part);
part              676 drivers/misc/sgi-xp/xpc_uv.c xpc_send_activate_IRQ_uv(struct xpc_partition *part, void *msg, size_t msg_size,
part              680 drivers/misc/sgi-xp/xpc_uv.c 	struct xpc_partition_uv *part_uv = &part->sn.uv;
part              689 drivers/misc/sgi-xp/xpc_uv.c 	msg_hdr->act_state = part->act_state;
part              732 drivers/misc/sgi-xp/xpc_uv.c xpc_send_activate_IRQ_part_uv(struct xpc_partition *part, void *msg,
part              737 drivers/misc/sgi-xp/xpc_uv.c 	ret = xpc_send_activate_IRQ_uv(part, msg, msg_size, msg_type);
part              739 drivers/misc/sgi-xp/xpc_uv.c 		XPC_DEACTIVATE_PARTITION(part, ret);
part              746 drivers/misc/sgi-xp/xpc_uv.c 	struct xpc_partition *part = &xpc_partitions[ch->partid];
part              749 drivers/misc/sgi-xp/xpc_uv.c 	ret = xpc_send_activate_IRQ_uv(part, msg, msg_size, msg_type);
part              754 drivers/misc/sgi-xp/xpc_uv.c 		XPC_DEACTIVATE_PARTITION(part, ret);
part              762 drivers/misc/sgi-xp/xpc_uv.c xpc_send_local_activate_IRQ_uv(struct xpc_partition *part, int act_state_req)
part              765 drivers/misc/sgi-xp/xpc_uv.c 	struct xpc_partition_uv *part_uv = &part->sn.uv;
part              875 drivers/misc/sgi-xp/xpc_uv.c xpc_get_remote_heartbeat_uv(struct xpc_partition *part)
part              877 drivers/misc/sgi-xp/xpc_uv.c 	struct xpc_partition_uv *part_uv = &part->sn.uv;
part              886 drivers/misc/sgi-xp/xpc_uv.c 	if (part_uv->cached_heartbeat.value == part->last_heartbeat &&
part              891 drivers/misc/sgi-xp/xpc_uv.c 		part->last_heartbeat = part_uv->cached_heartbeat.value;
part              901 drivers/misc/sgi-xp/xpc_uv.c 	struct xpc_partition *part = &xpc_partitions[partid];
part              904 drivers/misc/sgi-xp/xpc_uv.c 	part->remote_rp_pa = remote_rp_gpa; /* !!! _pa here is really _gpa */
part              905 drivers/misc/sgi-xp/xpc_uv.c 	part->remote_rp_ts_jiffies = remote_rp->ts_jiffies;
part              906 drivers/misc/sgi-xp/xpc_uv.c 	part->sn.uv.heartbeat_gpa = remote_rp->sn.uv.heartbeat_gpa;
part              907 drivers/misc/sgi-xp/xpc_uv.c 	part->sn.uv.activate_gru_mq_desc_gpa =
part              914 drivers/misc/sgi-xp/xpc_uv.c 	if (part->sn.uv.remote_act_state == XPC_P_AS_INACTIVE) {
part              919 drivers/misc/sgi-xp/xpc_uv.c 		xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg),
part              923 drivers/misc/sgi-xp/xpc_uv.c 	if (part->act_state == XPC_P_AS_INACTIVE)
part              924 drivers/misc/sgi-xp/xpc_uv.c 		xpc_send_local_activate_IRQ_uv(part, XPC_P_ASR_ACTIVATE_UV);
part              928 drivers/misc/sgi-xp/xpc_uv.c xpc_request_partition_reactivation_uv(struct xpc_partition *part)
part              930 drivers/misc/sgi-xp/xpc_uv.c 	xpc_send_local_activate_IRQ_uv(part, XPC_P_ASR_ACTIVATE_UV);
part              934 drivers/misc/sgi-xp/xpc_uv.c xpc_request_partition_deactivation_uv(struct xpc_partition *part)
part              942 drivers/misc/sgi-xp/xpc_uv.c 	if (part->sn.uv.remote_act_state != XPC_P_AS_DEACTIVATING &&
part              943 drivers/misc/sgi-xp/xpc_uv.c 	    part->sn.uv.remote_act_state != XPC_P_AS_INACTIVE) {
part              945 drivers/misc/sgi-xp/xpc_uv.c 		msg.reason = part->reason;
part              946 drivers/misc/sgi-xp/xpc_uv.c 		xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg),
part              952 drivers/misc/sgi-xp/xpc_uv.c xpc_cancel_partition_deactivation_request_uv(struct xpc_partition *part)
part             1016 drivers/misc/sgi-xp/xpc_uv.c xpc_setup_ch_structures_uv(struct xpc_partition *part)
part             1021 drivers/misc/sgi-xp/xpc_uv.c 	for (ch_number = 0; ch_number < part->nchannels; ch_number++) {
part             1022 drivers/misc/sgi-xp/xpc_uv.c 		ch_uv = &part->channels[ch_number].sn.uv;
part             1035 drivers/misc/sgi-xp/xpc_uv.c xpc_teardown_ch_structures_uv(struct xpc_partition *part)
part             1042 drivers/misc/sgi-xp/xpc_uv.c xpc_make_first_contact_uv(struct xpc_partition *part)
part             1051 drivers/misc/sgi-xp/xpc_uv.c 	xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg),
part             1054 drivers/misc/sgi-xp/xpc_uv.c 	while (!((part->sn.uv.remote_act_state == XPC_P_AS_ACTIVATING) ||
part             1055 drivers/misc/sgi-xp/xpc_uv.c 		 (part->sn.uv.remote_act_state == XPC_P_AS_ACTIVE))) {
part             1058 drivers/misc/sgi-xp/xpc_uv.c 			"partition %d\n", XPC_PARTID(part));
part             1063 drivers/misc/sgi-xp/xpc_uv.c 		if (part->act_state == XPC_P_AS_DEACTIVATING)
part             1064 drivers/misc/sgi-xp/xpc_uv.c 			return part->reason;
part             1071 drivers/misc/sgi-xp/xpc_uv.c xpc_get_chctl_all_flags_uv(struct xpc_partition *part)
part             1076 drivers/misc/sgi-xp/xpc_uv.c 	spin_lock_irqsave(&part->chctl_lock, irq_flags);
part             1077 drivers/misc/sgi-xp/xpc_uv.c 	chctl = part->chctl;
part             1079 drivers/misc/sgi-xp/xpc_uv.c 		part->chctl.all_flags = 0;
part             1081 drivers/misc/sgi-xp/xpc_uv.c 	spin_unlock_irqrestore(&part->chctl_lock, irq_flags);
part             1260 drivers/misc/sgi-xp/xpc_uv.c xpc_send_chctl_local_msgrequest_uv(struct xpc_partition *part, int ch_number)
part             1264 drivers/misc/sgi-xp/xpc_uv.c 	spin_lock_irqsave(&part->chctl_lock, irq_flags);
part             1265 drivers/misc/sgi-xp/xpc_uv.c 	part->chctl.flags[ch_number] |= XPC_CHCTL_MSGREQUEST;
part             1266 drivers/misc/sgi-xp/xpc_uv.c 	spin_unlock_irqrestore(&part->chctl_lock, irq_flags);
part             1268 drivers/misc/sgi-xp/xpc_uv.c 	xpc_wakeup_channel_mgr(part);
part             1283 drivers/misc/sgi-xp/xpc_uv.c xpc_indicate_partition_engaged_uv(struct xpc_partition *part)
part             1287 drivers/misc/sgi-xp/xpc_uv.c 	xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg),
part             1292 drivers/misc/sgi-xp/xpc_uv.c xpc_indicate_partition_disengaged_uv(struct xpc_partition *part)
part             1296 drivers/misc/sgi-xp/xpc_uv.c 	xpc_send_activate_IRQ_part_uv(part, &msg, sizeof(msg),
part             1410 drivers/misc/sgi-xp/xpc_uv.c xpc_handle_notify_mq_msg_uv(struct xpc_partition *part,
part             1413 drivers/misc/sgi-xp/xpc_uv.c 	struct xpc_partition_uv *part_uv = &part->sn.uv;
part             1420 drivers/misc/sgi-xp/xpc_uv.c 	if (unlikely(ch_number >= part->nchannels)) {
part             1423 drivers/misc/sgi-xp/xpc_uv.c 			ch_number, XPC_PARTID(part));
part             1437 drivers/misc/sgi-xp/xpc_uv.c 	ch = &part->channels[ch_number];
part             1473 drivers/misc/sgi-xp/xpc_uv.c 			xpc_send_chctl_local_msgrequest_uv(part, ch->number);
part             1483 drivers/misc/sgi-xp/xpc_uv.c 	struct xpc_partition *part;
part             1493 drivers/misc/sgi-xp/xpc_uv.c 			part = &xpc_partitions[partid];
part             1495 drivers/misc/sgi-xp/xpc_uv.c 			if (xpc_part_ref(part)) {
part             1496 drivers/misc/sgi-xp/xpc_uv.c 				xpc_handle_notify_mq_msg_uv(part, msg);
part             1497 drivers/misc/sgi-xp/xpc_uv.c 				xpc_part_deref(part);
part             1514 drivers/misc/sgi-xp/xpc_uv.c xpc_process_msg_chctl_flags_uv(struct xpc_partition *part, int ch_number)
part             1516 drivers/misc/sgi-xp/xpc_uv.c 	struct xpc_channel *ch = &part->channels[ch_number];
part               31 drivers/misc/sram-exec.c 			    struct sram_partition *part)
part               33 drivers/misc/sram-exec.c 	unsigned long base = (unsigned long)part->base;
part               45 drivers/misc/sram-exec.c int sram_add_protect_exec(struct sram_partition *part)
part               48 drivers/misc/sram-exec.c 	list_add_tail(&part->list, &exec_pool_list);
part               84 drivers/misc/sram-exec.c 	struct sram_partition *part = NULL, *p;
part               92 drivers/misc/sram-exec.c 			part = p;
part               96 drivers/misc/sram-exec.c 	if (!part)
part              102 drivers/misc/sram-exec.c 	base = (unsigned long)part->base;
part              105 drivers/misc/sram-exec.c 	mutex_lock(&part->lock);
part              115 drivers/misc/sram-exec.c 	mutex_unlock(&part->lock);
part               29 drivers/misc/sram.c 	struct sram_partition *part;
part               31 drivers/misc/sram.c 	part = container_of(attr, struct sram_partition, battr);
part               33 drivers/misc/sram.c 	mutex_lock(&part->lock);
part               34 drivers/misc/sram.c 	memcpy_fromio(buf, part->base + pos, count);
part               35 drivers/misc/sram.c 	mutex_unlock(&part->lock);
part               44 drivers/misc/sram.c 	struct sram_partition *part;
part               46 drivers/misc/sram.c 	part = container_of(attr, struct sram_partition, battr);
part               48 drivers/misc/sram.c 	mutex_lock(&part->lock);
part               49 drivers/misc/sram.c 	memcpy_toio(part->base + pos, buf, count);
part               50 drivers/misc/sram.c 	mutex_unlock(&part->lock);
part               56 drivers/misc/sram.c 			 phys_addr_t start, struct sram_partition *part)
part               60 drivers/misc/sram.c 	part->pool = devm_gen_pool_create(sram->dev, ilog2(SRAM_GRANULARITY),
part               62 drivers/misc/sram.c 	if (IS_ERR(part->pool))
part               63 drivers/misc/sram.c 		return PTR_ERR(part->pool);
part               65 drivers/misc/sram.c 	ret = gen_pool_add_virt(part->pool, (unsigned long)part->base, start,
part               76 drivers/misc/sram.c 			   phys_addr_t start, struct sram_partition *part)
part               78 drivers/misc/sram.c 	sysfs_bin_attr_init(&part->battr);
part               79 drivers/misc/sram.c 	part->battr.attr.name = devm_kasprintf(sram->dev, GFP_KERNEL,
part               82 drivers/misc/sram.c 	if (!part->battr.attr.name)
part               85 drivers/misc/sram.c 	part->battr.attr.mode = S_IRUSR | S_IWUSR;
part               86 drivers/misc/sram.c 	part->battr.read = sram_read;
part               87 drivers/misc/sram.c 	part->battr.write = sram_write;
part               88 drivers/misc/sram.c 	part->battr.size = block->size;
part               90 drivers/misc/sram.c 	return device_create_bin_file(sram->dev, &part->battr);
part               97 drivers/misc/sram.c 	struct sram_partition *part = &sram->partition[sram->partitions];
part               99 drivers/misc/sram.c 	mutex_init(&part->lock);
part              100 drivers/misc/sram.c 	part->base = sram->virt_base + block->start;
part              103 drivers/misc/sram.c 		ret = sram_add_pool(sram, block, start, part);
part              108 drivers/misc/sram.c 		ret = sram_add_export(sram, block, start, part);
part              113 drivers/misc/sram.c 		ret = sram_check_protect_exec(sram, block, part);
part              117 drivers/misc/sram.c 		ret = sram_add_pool(sram, block, start, part);
part              121 drivers/misc/sram.c 		sram_add_protect_exec(part);
part              131 drivers/misc/sram.c 	struct sram_partition *part;
part              136 drivers/misc/sram.c 	part = &sram->partition[sram->partitions - 1];
part              137 drivers/misc/sram.c 	for (; sram->partitions; sram->partitions--, part--) {
part              138 drivers/misc/sram.c 		if (part->battr.size)
part              139 drivers/misc/sram.c 			device_remove_bin_file(sram->dev, &part->battr);
part              141 drivers/misc/sram.c 		if (part->pool &&
part              142 drivers/misc/sram.c 		    gen_pool_avail(part->pool) < gen_pool_size(part->pool))
part               40 drivers/misc/sram.h 			    struct sram_partition *part);
part               41 drivers/misc/sram.h int sram_add_protect_exec(struct sram_partition *part);
part               45 drivers/misc/sram.h 					  struct sram_partition *part)
part               50 drivers/misc/sram.h static inline int sram_add_protect_exec(struct sram_partition *part)
part              106 drivers/mmc/core/block.c 	struct list_head part;
part              261 drivers/mmc/core/block.c 		list_for_each_entry(part_md, &md->part, part)
part             2292 drivers/mmc/core/block.c 	INIT_LIST_HEAD(&md->part);
part             2409 drivers/mmc/core/block.c 	list_add(&part_md->part, &md->part);
part             2585 drivers/mmc/core/block.c 		if (card->part[idx].area_type & MMC_BLK_DATA_AREA_RPMB) {
part             2593 drivers/mmc/core/block.c 				card->part[idx].part_cfg,
part             2594 drivers/mmc/core/block.c 				card->part[idx].size >> 9,
part             2595 drivers/mmc/core/block.c 				card->part[idx].name);
part             2598 drivers/mmc/core/block.c 		} else if (card->part[idx].size) {
part             2600 drivers/mmc/core/block.c 				card->part[idx].part_cfg,
part             2601 drivers/mmc/core/block.c 				card->part[idx].size >> 9,
part             2602 drivers/mmc/core/block.c 				card->part[idx].force_ro,
part             2603 drivers/mmc/core/block.c 				card->part[idx].name,
part             2604 drivers/mmc/core/block.c 				card->part[idx].area_type);
part             2652 drivers/mmc/core/block.c 	list_for_each_safe(pos, q, &md->part) {
part             2653 drivers/mmc/core/block.c 		part_md = list_entry(pos, struct mmc_blk_data, part);
part             2905 drivers/mmc/core/block.c 	list_for_each_entry(part_md, &md->part, part) {
part             2960 drivers/mmc/core/block.c 		list_for_each_entry(part_md, &md->part, part) {
part             2992 drivers/mmc/core/block.c 		list_for_each_entry(part_md, &md->part, part) {
part              304 drivers/mmc/core/mmc.c 	card->part[card->nr_parts].size = size;
part              305 drivers/mmc/core/mmc.c 	card->part[card->nr_parts].part_cfg = part_cfg;
part              306 drivers/mmc/core/mmc.c 	sprintf(card->part[card->nr_parts].name, name, idx);
part              307 drivers/mmc/core/mmc.c 	card->part[card->nr_parts].force_ro = ro;
part              308 drivers/mmc/core/mmc.c 	card->part[card->nr_parts].area_type = area_type;
part              151 drivers/mtd/ftl.c static int scan_header(partition_t *part)
part              157 drivers/mtd/ftl.c     part->header.FormattedSize = 0;
part              158 drivers/mtd/ftl.c     max_offset = (0x100000<part->mbd.mtd->size)?0x100000:part->mbd.mtd->size;
part              162 drivers/mtd/ftl.c 	 offset += part->mbd.mtd->erasesize ? : 0x2000) {
part              164 drivers/mtd/ftl.c 	err = mtd_read(part->mbd.mtd, offset, sizeof(header), &ret,
part              183 drivers/mtd/ftl.c     if ((1 << header.EraseUnitSize) != part->mbd.mtd->erasesize) {
part              185 drivers/mtd/ftl.c 	       1 << header.EraseUnitSize,part->mbd.mtd->erasesize);
part              188 drivers/mtd/ftl.c     part->header = header;
part              192 drivers/mtd/ftl.c static int build_maps(partition_t *part)
part              202 drivers/mtd/ftl.c     part->DataUnits = le16_to_cpu(part->header.NumEraseUnits) -
part              203 drivers/mtd/ftl.c 	part->header.NumTransferUnits;
part              204 drivers/mtd/ftl.c     part->EUNInfo = kmalloc_array(part->DataUnits, sizeof(struct eun_info_t),
part              206 drivers/mtd/ftl.c     if (!part->EUNInfo)
part              208 drivers/mtd/ftl.c     for (i = 0; i < part->DataUnits; i++)
part              209 drivers/mtd/ftl.c 	part->EUNInfo[i].Offset = 0xffffffff;
part              210 drivers/mtd/ftl.c     part->XferInfo =
part              211 drivers/mtd/ftl.c 	kmalloc_array(part->header.NumTransferUnits,
part              214 drivers/mtd/ftl.c     if (!part->XferInfo)
part              218 drivers/mtd/ftl.c     for (i = 0; i < le16_to_cpu(part->header.NumEraseUnits); i++) {
part              219 drivers/mtd/ftl.c 	offset = ((i + le16_to_cpu(part->header.FirstPhysicalEUN))
part              220 drivers/mtd/ftl.c 		      << part->header.EraseUnitSize);
part              221 drivers/mtd/ftl.c 	ret = mtd_read(part->mbd.mtd, offset, sizeof(header), &retval,
part              230 drivers/mtd/ftl.c 	if (hdr_ok && (le16_to_cpu(header.LogicalEUN) < part->DataUnits) &&
part              231 drivers/mtd/ftl.c 	    (part->EUNInfo[le16_to_cpu(header.LogicalEUN)].Offset == 0xffffffff)) {
part              232 drivers/mtd/ftl.c 	    part->EUNInfo[le16_to_cpu(header.LogicalEUN)].Offset = offset;
part              233 drivers/mtd/ftl.c 	    part->EUNInfo[le16_to_cpu(header.LogicalEUN)].EraseCount =
part              237 drivers/mtd/ftl.c 	    if (xtrans == part->header.NumTransferUnits) {
part              243 drivers/mtd/ftl.c 		part->XferInfo[xtrans].state = XFER_PREPARED;
part              244 drivers/mtd/ftl.c 		part->XferInfo[xtrans].EraseCount = le32_to_cpu(header.EraseCount);
part              246 drivers/mtd/ftl.c 		part->XferInfo[xtrans].state = XFER_UNKNOWN;
part              248 drivers/mtd/ftl.c 		part->XferInfo[xtrans].EraseCount =
part              249 drivers/mtd/ftl.c 		    le32_to_cpu(part->header.EraseCount);
part              251 drivers/mtd/ftl.c 	    part->XferInfo[xtrans].Offset = offset;
part              256 drivers/mtd/ftl.c     header = part->header;
part              266 drivers/mtd/ftl.c     part->VirtualBlockMap = vmalloc(array_size(blocks, sizeof(uint32_t)));
part              267 drivers/mtd/ftl.c     if (!part->VirtualBlockMap)
part              270 drivers/mtd/ftl.c     memset(part->VirtualBlockMap, 0xff, blocks * sizeof(uint32_t));
part              271 drivers/mtd/ftl.c     part->BlocksPerUnit = (1 << header.EraseUnitSize) >> header.BlockSize;
part              273 drivers/mtd/ftl.c     part->bam_cache = kmalloc_array(part->BlocksPerUnit, sizeof(uint32_t),
part              275 drivers/mtd/ftl.c     if (!part->bam_cache)
part              278 drivers/mtd/ftl.c     part->bam_index = 0xffff;
part              279 drivers/mtd/ftl.c     part->FreeTotal = 0;
part              281 drivers/mtd/ftl.c     for (i = 0; i < part->DataUnits; i++) {
part              282 drivers/mtd/ftl.c 	part->EUNInfo[i].Free = 0;
part              283 drivers/mtd/ftl.c 	part->EUNInfo[i].Deleted = 0;
part              284 drivers/mtd/ftl.c 	offset = part->EUNInfo[i].Offset + le32_to_cpu(header.BAMOffset);
part              286 drivers/mtd/ftl.c 	ret = mtd_read(part->mbd.mtd, offset,
part              287 drivers/mtd/ftl.c                        part->BlocksPerUnit * sizeof(uint32_t), &retval,
part              288 drivers/mtd/ftl.c                        (unsigned char *)part->bam_cache);
part              293 drivers/mtd/ftl.c 	for (j = 0; j < part->BlocksPerUnit; j++) {
part              294 drivers/mtd/ftl.c 	    if (BLOCK_FREE(le32_to_cpu(part->bam_cache[j]))) {
part              295 drivers/mtd/ftl.c 		part->EUNInfo[i].Free++;
part              296 drivers/mtd/ftl.c 		part->FreeTotal++;
part              297 drivers/mtd/ftl.c 	    } else if ((BLOCK_TYPE(le32_to_cpu(part->bam_cache[j])) == BLOCK_DATA) &&
part              298 drivers/mtd/ftl.c 		     (BLOCK_NUMBER(le32_to_cpu(part->bam_cache[j])) < blocks))
part              299 drivers/mtd/ftl.c 		part->VirtualBlockMap[BLOCK_NUMBER(le32_to_cpu(part->bam_cache[j]))] =
part              301 drivers/mtd/ftl.c 	    else if (BLOCK_DELETED(le32_to_cpu(part->bam_cache[j])))
part              302 drivers/mtd/ftl.c 		part->EUNInfo[i].Deleted++;
part              310 drivers/mtd/ftl.c     kfree(part->bam_cache);
part              312 drivers/mtd/ftl.c     vfree(part->VirtualBlockMap);
part              314 drivers/mtd/ftl.c     kfree(part->XferInfo);
part              316 drivers/mtd/ftl.c     kfree(part->EUNInfo);
part              328 drivers/mtd/ftl.c static int erase_xfer(partition_t *part,
part              335 drivers/mtd/ftl.c     xfer = &part->XferInfo[xfernum];
part              347 drivers/mtd/ftl.c     erase->len = 1 << part->header.EraseUnitSize;
part              349 drivers/mtd/ftl.c     ret = mtd_erase(part->mbd.mtd, erase);
part              370 drivers/mtd/ftl.c static int prepare_xfer(partition_t *part, int i)
part              379 drivers/mtd/ftl.c     xfer = &part->XferInfo[i];
part              385 drivers/mtd/ftl.c     header = part->header;
part              389 drivers/mtd/ftl.c     ret = mtd_write(part->mbd.mtd, xfer->Offset, sizeof(header), &retlen,
part              397 drivers/mtd/ftl.c     nbam = DIV_ROUND_UP(part->BlocksPerUnit * sizeof(uint32_t) +
part              398 drivers/mtd/ftl.c 			le32_to_cpu(part->header.BAMOffset), SECTOR_SIZE);
part              400 drivers/mtd/ftl.c     offset = xfer->Offset + le32_to_cpu(part->header.BAMOffset);
part              405 drivers/mtd/ftl.c 	ret = mtd_write(part->mbd.mtd, offset, sizeof(uint32_t), &retlen,
part              428 drivers/mtd/ftl.c static int copy_erase_unit(partition_t *part, uint16_t srcunit,
part              441 drivers/mtd/ftl.c     eun = &part->EUNInfo[srcunit];
part              442 drivers/mtd/ftl.c     xfer = &part->XferInfo[xferunit];
part              448 drivers/mtd/ftl.c     if (part->bam_index != srcunit) {
part              450 drivers/mtd/ftl.c 	offset = eun->Offset + le32_to_cpu(part->header.BAMOffset);
part              452 drivers/mtd/ftl.c 	ret = mtd_read(part->mbd.mtd, offset,
part              453 drivers/mtd/ftl.c                        part->BlocksPerUnit * sizeof(uint32_t), &retlen,
part              454 drivers/mtd/ftl.c                        (u_char *)(part->bam_cache));
part              457 drivers/mtd/ftl.c 	part->bam_index = 0xffff;
part              470 drivers/mtd/ftl.c     ret = mtd_write(part->mbd.mtd, offset, sizeof(uint16_t), &retlen,
part              483 drivers/mtd/ftl.c     for (i = 0; i < part->BlocksPerUnit; i++) {
part              484 drivers/mtd/ftl.c 	switch (BLOCK_TYPE(le32_to_cpu(part->bam_cache[i]))) {
part              490 drivers/mtd/ftl.c 	    ret = mtd_read(part->mbd.mtd, src, SECTOR_SIZE, &retlen,
part              498 drivers/mtd/ftl.c 	    ret = mtd_write(part->mbd.mtd, dest, SECTOR_SIZE, &retlen,
part              508 drivers/mtd/ftl.c 	    part->bam_cache[i] = cpu_to_le32(0xffffffff);
part              517 drivers/mtd/ftl.c     ret = mtd_write(part->mbd.mtd,
part              518 drivers/mtd/ftl.c                     xfer->Offset + le32_to_cpu(part->header.BAMOffset),
part              519 drivers/mtd/ftl.c                     part->BlocksPerUnit * sizeof(int32_t),
part              521 drivers/mtd/ftl.c                     (u_char *)part->bam_cache);
part              529 drivers/mtd/ftl.c     ret = mtd_write(part->mbd.mtd, xfer->Offset + 20, sizeof(uint16_t),
part              541 drivers/mtd/ftl.c     part->FreeTotal -= eun->Free;
part              542 drivers/mtd/ftl.c     part->FreeTotal += free;
part              547 drivers/mtd/ftl.c     part->bam_index = srcunit;
part              568 drivers/mtd/ftl.c static int reclaim_block(partition_t *part)
part              575 drivers/mtd/ftl.c     pr_debug("NumTransferUnits == %x\n", part->header.NumTransferUnits);
part              580 drivers/mtd/ftl.c 	for (i = 0; i < part->header.NumTransferUnits; i++) {
part              582 drivers/mtd/ftl.c 	    if (part->XferInfo[i].state == XFER_UNKNOWN) {
part              585 drivers/mtd/ftl.c 		erase_xfer(part, i);
part              587 drivers/mtd/ftl.c 	    if (part->XferInfo[i].state == XFER_ERASING) {
part              592 drivers/mtd/ftl.c 	    else if (part->XferInfo[i].state == XFER_ERASED) {
part              595 drivers/mtd/ftl.c 		prepare_xfer(part, i);
part              597 drivers/mtd/ftl.c 	    if (part->XferInfo[i].state == XFER_PREPARED) {
part              600 drivers/mtd/ftl.c 		if (part->XferInfo[i].EraseCount <= best) {
part              601 drivers/mtd/ftl.c 		    best = part->XferInfo[i].EraseCount;
part              606 drivers/mtd/ftl.c 		    pr_debug("XferInfo[%d].state == %x\n",i, part->XferInfo[i].state);
part              613 drivers/mtd/ftl.c 		mtd_sync(part->mbd.mtd);
part              632 drivers/mtd/ftl.c 	for (i = 0; i < part->DataUnits; i++)
part              633 drivers/mtd/ftl.c 	    if (part->EUNInfo[i].EraseCount <= best) {
part              634 drivers/mtd/ftl.c 		best = part->EUNInfo[i].EraseCount;
part              639 drivers/mtd/ftl.c 	for (i = 0; i < part->DataUnits; i++)
part              640 drivers/mtd/ftl.c 	    if (part->EUNInfo[i].Deleted >= best) {
part              641 drivers/mtd/ftl.c 		best = part->EUNInfo[i].Deleted;
part              656 drivers/mtd/ftl.c     ret = copy_erase_unit(part, eun, xfer);
part              658 drivers/mtd/ftl.c 	erase_xfer(part, xfer);
part              675 drivers/mtd/ftl.c static void dump_lists(partition_t *part)
part              678 drivers/mtd/ftl.c     printk(KERN_DEBUG "ftl_cs: Free total = %d\n", part->FreeTotal);
part              679 drivers/mtd/ftl.c     for (i = 0; i < part->DataUnits; i++)
part              682 drivers/mtd/ftl.c 	       part->EUNInfo[i].Offset >> part->header.EraseUnitSize,
part              683 drivers/mtd/ftl.c 	       part->EUNInfo[i].Free, part->EUNInfo[i].Deleted);
part              687 drivers/mtd/ftl.c static uint32_t find_free(partition_t *part)
part              695 drivers/mtd/ftl.c     stop = (part->bam_index == 0xffff) ? 0 : part->bam_index;
part              698 drivers/mtd/ftl.c 	if (part->EUNInfo[eun].Free != 0) break;
part              700 drivers/mtd/ftl.c 	if (++eun == part->DataUnits) eun = 0;
part              703 drivers/mtd/ftl.c     if (part->EUNInfo[eun].Free == 0)
part              707 drivers/mtd/ftl.c     if (eun != part->bam_index) {
part              709 drivers/mtd/ftl.c 	part->bam_index = 0xffff;
part              711 drivers/mtd/ftl.c 	ret = mtd_read(part->mbd.mtd,
part              712 drivers/mtd/ftl.c                        part->EUNInfo[eun].Offset + le32_to_cpu(part->header.BAMOffset),
part              713 drivers/mtd/ftl.c                        part->BlocksPerUnit * sizeof(uint32_t),
part              715 drivers/mtd/ftl.c                        (u_char *)(part->bam_cache));
part              721 drivers/mtd/ftl.c 	part->bam_index = eun;
part              725 drivers/mtd/ftl.c     for (blk = 0; blk < part->BlocksPerUnit; blk++)
part              726 drivers/mtd/ftl.c 	if (BLOCK_FREE(le32_to_cpu(part->bam_cache[blk]))) break;
part              727 drivers/mtd/ftl.c     if (blk == part->BlocksPerUnit) {
part              731 drivers/mtd/ftl.c 	    dump_lists(part);
part              748 drivers/mtd/ftl.c static int ftl_read(partition_t *part, caddr_t buffer,
part              757 drivers/mtd/ftl.c 	  part, sector, nblocks);
part              758 drivers/mtd/ftl.c     if (!(part->state & FTL_FORMATTED)) {
part              762 drivers/mtd/ftl.c     bsize = 1 << part->header.EraseUnitSize;
part              765 drivers/mtd/ftl.c 	if (((sector+i) * SECTOR_SIZE) >= le32_to_cpu(part->header.FormattedSize)) {
part              769 drivers/mtd/ftl.c 	log_addr = part->VirtualBlockMap[sector+i];
part              773 drivers/mtd/ftl.c 	    offset = (part->EUNInfo[log_addr / bsize].Offset
part              775 drivers/mtd/ftl.c 	    ret = mtd_read(part->mbd.mtd, offset, SECTOR_SIZE, &retlen,
part              794 drivers/mtd/ftl.c static int set_bam_entry(partition_t *part, uint32_t log_addr,
part              806 drivers/mtd/ftl.c 	  part, log_addr, virt_addr);
part              807 drivers/mtd/ftl.c     bsize = 1 << part->header.EraseUnitSize;
part              810 drivers/mtd/ftl.c     offset = (part->EUNInfo[eun].Offset + blk * sizeof(uint32_t) +
part              811 drivers/mtd/ftl.c 		  le32_to_cpu(part->header.BAMOffset));
part              814 drivers/mtd/ftl.c     ret = mtd_read(part->mbd.mtd, offset, sizeof(uint32_t), &retlen,
part              835 drivers/mtd/ftl.c     if (part->bam_index == eun) {
part              837 drivers/mtd/ftl.c 	if (le32_to_cpu(part->bam_cache[blk]) != old_addr) {
part              844 drivers/mtd/ftl.c 		       le32_to_cpu(part->bam_cache[blk]), old_addr);
part              849 drivers/mtd/ftl.c 	part->bam_cache[blk] = le_virt_addr;
part              851 drivers/mtd/ftl.c     ret = mtd_write(part->mbd.mtd, offset, sizeof(uint32_t), &retlen,
part              862 drivers/mtd/ftl.c static int ftl_write(partition_t *part, caddr_t buffer,
part              871 drivers/mtd/ftl.c 	  part, sector, nblocks);
part              872 drivers/mtd/ftl.c     if (!(part->state & FTL_FORMATTED)) {
part              877 drivers/mtd/ftl.c     while (part->FreeTotal < nblocks) {
part              878 drivers/mtd/ftl.c 	ret = reclaim_block(part);
part              883 drivers/mtd/ftl.c     bsize = 1 << part->header.EraseUnitSize;
part              887 drivers/mtd/ftl.c 	if (virt_addr >= le32_to_cpu(part->header.FormattedSize)) {
part              893 drivers/mtd/ftl.c 	blk = find_free(part);
part              903 drivers/mtd/ftl.c 	log_addr = part->bam_index * bsize + blk * SECTOR_SIZE;
part              904 drivers/mtd/ftl.c 	part->EUNInfo[part->bam_index].Free--;
part              905 drivers/mtd/ftl.c 	part->FreeTotal--;
part              906 drivers/mtd/ftl.c 	if (set_bam_entry(part, log_addr, 0xfffffffe))
part              908 drivers/mtd/ftl.c 	part->EUNInfo[part->bam_index].Deleted++;
part              909 drivers/mtd/ftl.c 	offset = (part->EUNInfo[part->bam_index].Offset +
part              911 drivers/mtd/ftl.c 	ret = mtd_write(part->mbd.mtd, offset, SECTOR_SIZE, &retlen, buffer);
part              922 drivers/mtd/ftl.c 	old_addr = part->VirtualBlockMap[sector+i];
part              924 drivers/mtd/ftl.c 	    part->VirtualBlockMap[sector+i] = 0xffffffff;
part              925 drivers/mtd/ftl.c 	    part->EUNInfo[old_addr/bsize].Deleted++;
part              926 drivers/mtd/ftl.c 	    if (set_bam_entry(part, old_addr, 0))
part              931 drivers/mtd/ftl.c 	if (set_bam_entry(part, log_addr, virt_addr))
part              933 drivers/mtd/ftl.c 	part->VirtualBlockMap[sector+i] = log_addr;
part              934 drivers/mtd/ftl.c 	part->EUNInfo[part->bam_index].Deleted--;
part              944 drivers/mtd/ftl.c 	partition_t *part = (void *)dev;
part              948 drivers/mtd/ftl.c 	sect = le32_to_cpu(part->header.FormattedSize)/SECTOR_SIZE;
part              972 drivers/mtd/ftl.c 	partition_t *part = (void *)dev;
part              973 drivers/mtd/ftl.c 	uint32_t bsize = 1 << part->header.EraseUnitSize;
part              979 drivers/mtd/ftl.c 		uint32_t old_addr = part->VirtualBlockMap[sector];
part              981 drivers/mtd/ftl.c 			part->VirtualBlockMap[sector] = 0xffffffff;
part              982 drivers/mtd/ftl.c 			part->EUNInfo[old_addr/bsize].Deleted++;
part              983 drivers/mtd/ftl.c 			if (set_bam_entry(part, old_addr, 0))
part              994 drivers/mtd/ftl.c static void ftl_freepart(partition_t *part)
part              996 drivers/mtd/ftl.c 	vfree(part->VirtualBlockMap);
part              997 drivers/mtd/ftl.c 	part->VirtualBlockMap = NULL;
part              998 drivers/mtd/ftl.c 	kfree(part->EUNInfo);
part              999 drivers/mtd/ftl.c 	part->EUNInfo = NULL;
part             1000 drivers/mtd/ftl.c 	kfree(part->XferInfo);
part             1001 drivers/mtd/ftl.c 	part->XferInfo = NULL;
part             1002 drivers/mtd/ftl.c 	kfree(part->bam_cache);
part             1003 drivers/mtd/ftl.c 	part->bam_cache = NULL;
part               52 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part               57 drivers/mtd/mtdpart.c 	return part_absolute_offset(part->parent) + part->offset;
part               68 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part               72 drivers/mtd/mtdpart.c 	stats = part->parent->ecc_stats;
part               73 drivers/mtd/mtdpart.c 	res = part->parent->_read(part->parent, from + part->offset, len,
part               77 drivers/mtd/mtdpart.c 			part->parent->ecc_stats.failed - stats.failed;
part               80 drivers/mtd/mtdpart.c 			part->parent->ecc_stats.corrected - stats.corrected;
part               87 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part               89 drivers/mtd/mtdpart.c 	return part->parent->_point(part->parent, from + part->offset, len,
part               95 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part               97 drivers/mtd/mtdpart.c 	return part->parent->_unpoint(part->parent, from + part->offset, len);
part              103 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              107 drivers/mtd/mtdpart.c 	stats = part->parent->ecc_stats;
part              108 drivers/mtd/mtdpart.c 	res = part->parent->_read_oob(part->parent, from + part->offset, ops);
part              111 drivers/mtd/mtdpart.c 			part->parent->ecc_stats.failed - stats.failed;
part              114 drivers/mtd/mtdpart.c 			part->parent->ecc_stats.corrected - stats.corrected;
part              121 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              122 drivers/mtd/mtdpart.c 	return part->parent->_read_user_prot_reg(part->parent, from, len,
part              129 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              130 drivers/mtd/mtdpart.c 	return part->parent->_get_user_prot_info(part->parent, len, retlen,
part              137 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              138 drivers/mtd/mtdpart.c 	return part->parent->_read_fact_prot_reg(part->parent, from, len,
part              145 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              146 drivers/mtd/mtdpart.c 	return part->parent->_get_fact_prot_info(part->parent, len, retlen,
part              153 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              154 drivers/mtd/mtdpart.c 	return part->parent->_write(part->parent, to + part->offset, len,
part              161 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              162 drivers/mtd/mtdpart.c 	return part->parent->_panic_write(part->parent, to + part->offset, len,
part              169 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              171 drivers/mtd/mtdpart.c 	return part->parent->_write_oob(part->parent, to + part->offset, ops);
part              177 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              178 drivers/mtd/mtdpart.c 	return part->parent->_write_user_prot_reg(part->parent, from, len,
part              185 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              186 drivers/mtd/mtdpart.c 	return part->parent->_lock_user_prot_reg(part->parent, from, len);
part              192 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              193 drivers/mtd/mtdpart.c 	return part->parent->_writev(part->parent, vecs, count,
part              194 drivers/mtd/mtdpart.c 				     to + part->offset, retlen);
part              199 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              202 drivers/mtd/mtdpart.c 	instr->addr += part->offset;
part              203 drivers/mtd/mtdpart.c 	ret = part->parent->_erase(part->parent, instr);
part              205 drivers/mtd/mtdpart.c 		instr->fail_addr -= part->offset;
part              206 drivers/mtd/mtdpart.c 	instr->addr -= part->offset;
part              213 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              214 drivers/mtd/mtdpart.c 	return part->parent->_lock(part->parent, ofs + part->offset, len);
part              219 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              220 drivers/mtd/mtdpart.c 	return part->parent->_unlock(part->parent, ofs + part->offset, len);
part              225 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              226 drivers/mtd/mtdpart.c 	return part->parent->_is_locked(part->parent, ofs + part->offset, len);
part              231 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              232 drivers/mtd/mtdpart.c 	part->parent->_sync(part->parent);
part              237 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              238 drivers/mtd/mtdpart.c 	return part->parent->_suspend(part->parent);
part              243 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              244 drivers/mtd/mtdpart.c 	part->parent->_resume(part->parent);
part              249 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              250 drivers/mtd/mtdpart.c 	ofs += part->offset;
part              251 drivers/mtd/mtdpart.c 	return part->parent->_block_isreserved(part->parent, ofs);
part              256 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              257 drivers/mtd/mtdpart.c 	ofs += part->offset;
part              258 drivers/mtd/mtdpart.c 	return part->parent->_block_isbad(part->parent, ofs);
part              263 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              266 drivers/mtd/mtdpart.c 	ofs += part->offset;
part              267 drivers/mtd/mtdpart.c 	res = part->parent->_block_markbad(part->parent, ofs);
part              275 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              276 drivers/mtd/mtdpart.c 	return part->parent->_get_device(part->parent);
part              281 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              282 drivers/mtd/mtdpart.c 	part->parent->_put_device(part->parent);
part              288 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              290 drivers/mtd/mtdpart.c 	return mtd_ooblayout_ecc(part->parent, section, oobregion);
part              296 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              298 drivers/mtd/mtdpart.c 	return mtd_ooblayout_free(part->parent, section, oobregion);
part              308 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              310 drivers/mtd/mtdpart.c 	return part->parent->_max_bad_blocks(part->parent,
part              311 drivers/mtd/mtdpart.c 					     ofs + part->offset, len);
part              321 drivers/mtd/mtdpart.c 			const struct mtd_partition *part, int partno,
part              333 drivers/mtd/mtdpart.c 	name = kstrdup(part->name, GFP_KERNEL);
part              344 drivers/mtd/mtdpart.c 	slave->mtd.flags = parent->orig_flags & ~part->mask_flags;
part              346 drivers/mtd/mtdpart.c 	slave->mtd.size = part->size;
part              368 drivers/mtd/mtdpart.c 	slave->mtd.dev.of_node = part->of_node;
part              430 drivers/mtd/mtdpart.c 	slave->offset = part->offset;
part              452 drivers/mtd/mtdpart.c 				part->name, parent->size - slave->offset,
part              474 drivers/mtd/mtdpart.c 			part->name);
part              480 drivers/mtd/mtdpart.c 			part->name, parent->name, (unsigned long long)slave->mtd.size);
part              524 drivers/mtd/mtdpart.c 			part->name);
part              532 drivers/mtd/mtdpart.c 			part->name);
part              560 drivers/mtd/mtdpart.c 	struct mtd_part *part = mtd_to_part(mtd);
part              561 drivers/mtd/mtdpart.c 	return snprintf(buf, PAGE_SIZE, "%llu\n", part->offset);
part              583 drivers/mtd/mtdpart.c 	struct mtd_partition part;
part              598 drivers/mtd/mtdpart.c 	memset(&part, 0, sizeof(part));
part              599 drivers/mtd/mtdpart.c 	part.name = name;
part              600 drivers/mtd/mtdpart.c 	part.size = length;
part              601 drivers/mtd/mtdpart.c 	part.offset = offset;
part              603 drivers/mtd/mtdpart.c 	new = allocate_partition(parent, &part, -1, offset);
part             1028 drivers/mtd/mtdpart.c 	struct mtd_part *part;
part             1032 drivers/mtd/mtdpart.c 	list_for_each_entry(part, &mtd_partitions, list)
part             1033 drivers/mtd/mtdpart.c 		if (&part->mtd == mtd) {
part             1346 drivers/mtd/mtdswap.c 	unsigned long part;
part             1356 drivers/mtd/mtdswap.c 		if (kstrtoul(this_opt, 0, &part) < 0)
part             1359 drivers/mtd/mtdswap.c 		if (mtd->index == part)
part             1363 drivers/mtd/mtdswap.c 	if (mtd->index != part)
part             1422 drivers/mtd/mtdswap.c 		MTDSWAP_PREFIX, part, swap_size / 1024, spare_cnt, bad_blocks);
part              120 drivers/mtd/parsers/afs.c 				  u_int off, struct mtd_partition *part)
part              201 drivers/mtd/parsers/afs.c 	part->name = kstrdup(iis.name, GFP_KERNEL);
part              202 drivers/mtd/parsers/afs.c 	if (!part->name)
part              205 drivers/mtd/parsers/afs.c 	part->size = (iis.length + mtd->erasesize - 1) & ~(mtd->erasesize - 1);
part              206 drivers/mtd/parsers/afs.c 	part->offset = img_ptr;
part              207 drivers/mtd/parsers/afs.c 	part->mask_flags = 0;
part              210 drivers/mtd/parsers/afs.c 	       img_ptr, part->size / 1024,
part              211 drivers/mtd/parsers/afs.c 	       iis.imageNumber, part->name);
part              217 drivers/mtd/parsers/afs.c 				  u_int off, struct mtd_partition *part)
part              312 drivers/mtd/parsers/afs.c 		part->name = kstrdup(name, GFP_KERNEL);
part              313 drivers/mtd/parsers/afs.c 		if (!part->name)
part              315 drivers/mtd/parsers/afs.c 		part->offset = region_start;
part              316 drivers/mtd/parsers/afs.c 		part->size = region_end - region_start;
part              317 drivers/mtd/parsers/afs.c 		part->mask_flags = 0;
part               54 drivers/mtd/parsers/bcm47xxpart.c static void bcm47xxpart_add_part(struct mtd_partition *part, const char *name,
part               57 drivers/mtd/parsers/bcm47xxpart.c 	part->name = name;
part               58 drivers/mtd/parsers/bcm47xxpart.c 	part->offset = offset;
part               59 drivers/mtd/parsers/bcm47xxpart.c 	part->mask_flags = mask_flags;
part              302 drivers/mtd/parsers/cmdlinepart.c 	struct cmdline_mtd_partition *part;
part              316 drivers/mtd/parsers/cmdlinepart.c 	for (part = partitions; part; part = part->next) {
part              317 drivers/mtd/parsers/cmdlinepart.c 		if ((!mtd_id) || (!strcmp(part->mtd_id, mtd_id)))
part              321 drivers/mtd/parsers/cmdlinepart.c 	if (!part)
part              324 drivers/mtd/parsers/cmdlinepart.c 	for (i = 0, offset = 0; i < part->num_parts; i++) {
part              325 drivers/mtd/parsers/cmdlinepart.c 		if (part->parts[i].offset == OFFSET_CONTINUOUS)
part              326 drivers/mtd/parsers/cmdlinepart.c 			part->parts[i].offset = offset;
part              328 drivers/mtd/parsers/cmdlinepart.c 			offset = part->parts[i].offset;
part              330 drivers/mtd/parsers/cmdlinepart.c 		if (part->parts[i].size == SIZE_REMAINING)
part              331 drivers/mtd/parsers/cmdlinepart.c 			part->parts[i].size = master->size - offset;
part              333 drivers/mtd/parsers/cmdlinepart.c 		if (offset + part->parts[i].size > master->size) {
part              335 drivers/mtd/parsers/cmdlinepart.c 				part->mtd_id);
part              336 drivers/mtd/parsers/cmdlinepart.c 			part->parts[i].size = master->size - offset;
part              338 drivers/mtd/parsers/cmdlinepart.c 		offset += part->parts[i].size;
part              340 drivers/mtd/parsers/cmdlinepart.c 		if (part->parts[i].size == 0) {
part              342 drivers/mtd/parsers/cmdlinepart.c 				part->mtd_id);
part              343 drivers/mtd/parsers/cmdlinepart.c 			part->num_parts--;
part              344 drivers/mtd/parsers/cmdlinepart.c 			memmove(&part->parts[i], &part->parts[i + 1],
part              345 drivers/mtd/parsers/cmdlinepart.c 				sizeof(*part->parts) * (part->num_parts - i));
part              350 drivers/mtd/parsers/cmdlinepart.c 	*pparts = kmemdup(part->parts, sizeof(*part->parts) * part->num_parts,
part              355 drivers/mtd/parsers/cmdlinepart.c 	return part->num_parts;
part              160 drivers/mtd/parsers/ofpart.c 	} *part;
part              168 drivers/mtd/parsers/ofpart.c 	part = of_get_property(dp, "partitions", &plen);
part              169 drivers/mtd/parsers/ofpart.c 	if (!part)
part              174 drivers/mtd/parsers/ofpart.c 	nr_parts = plen / sizeof(part[0]);
part              183 drivers/mtd/parsers/ofpart.c 		parts[i].offset = be32_to_cpu(part->offset);
part              184 drivers/mtd/parsers/ofpart.c 		parts[i].size   = be32_to_cpu(part->len) & ~1;
part              186 drivers/mtd/parsers/ofpart.c 		if (be32_to_cpu(part->len) & 1)
part              199 drivers/mtd/parsers/ofpart.c 		part++;
part               55 drivers/mtd/parsers/parser_trx.c 	struct mtd_partition *part;
part               80 drivers/mtd/parsers/parser_trx.c 		part = &parts[curr_part++];
part               81 drivers/mtd/parsers/parser_trx.c 		part->name = "loader";
part               82 drivers/mtd/parsers/parser_trx.c 		part->offset = trx.offset[i];
part               87 drivers/mtd/parsers/parser_trx.c 		part = &parts[curr_part++];
part               88 drivers/mtd/parsers/parser_trx.c 		part->name = "linux";
part               89 drivers/mtd/parsers/parser_trx.c 		part->offset = trx.offset[i];
part               94 drivers/mtd/parsers/parser_trx.c 		part = &parts[curr_part++];
part               95 drivers/mtd/parsers/parser_trx.c 		part->name = parser_trx_data_part_name(mtd, trx.offset[i]);
part               96 drivers/mtd/parsers/parser_trx.c 		part->offset = trx.offset[i];
part               93 drivers/mtd/rfd_ftl.c static int build_block_map(struct partition *part, int block_no)
part               95 drivers/mtd/rfd_ftl.c 	struct block *block = &part->blocks[block_no];
part               98 drivers/mtd/rfd_ftl.c 	block->offset = part->block_size * block_no;
part              100 drivers/mtd/rfd_ftl.c 	if (le16_to_cpu(part->header_cache[0]) != RFD_MAGIC) {
part              107 drivers/mtd/rfd_ftl.c 	for (i=0; i<part->data_sectors_per_block; i++) {
part              110 drivers/mtd/rfd_ftl.c 		entry = le16_to_cpu(part->header_cache[HEADER_MAP_OFFSET + i]);
part              123 drivers/mtd/rfd_ftl.c 		if (entry >= part->sector_count) {
part              127 drivers/mtd/rfd_ftl.c 				part->mbd.mtd->name, block_no, i, entry);
part              131 drivers/mtd/rfd_ftl.c 		if (part->sector_map[entry] != -1) {
part              134 drivers/mtd/rfd_ftl.c 				part->mbd.mtd->name, entry);
part              135 drivers/mtd/rfd_ftl.c 			part->errors = 1;
part              139 drivers/mtd/rfd_ftl.c 		part->sector_map[entry] = block->offset +
part              140 drivers/mtd/rfd_ftl.c 			(i + part->header_sectors_per_block) * SECTOR_SIZE;
part              145 drivers/mtd/rfd_ftl.c 	if (block->free_sectors == part->data_sectors_per_block)
part              146 drivers/mtd/rfd_ftl.c 		part->reserved_block = block_no;
part              151 drivers/mtd/rfd_ftl.c static int scan_header(struct partition *part)
part              158 drivers/mtd/rfd_ftl.c 	sectors_per_block = part->block_size / SECTOR_SIZE;
part              159 drivers/mtd/rfd_ftl.c 	part->total_blocks = (u32)part->mbd.mtd->size / part->block_size;
part              161 drivers/mtd/rfd_ftl.c 	if (part->total_blocks < 2)
part              165 drivers/mtd/rfd_ftl.c 	part->header_sectors_per_block =
part              169 drivers/mtd/rfd_ftl.c 	part->data_sectors_per_block = sectors_per_block -
part              170 drivers/mtd/rfd_ftl.c 			part->header_sectors_per_block;
part              172 drivers/mtd/rfd_ftl.c 	part->header_size = (HEADER_MAP_OFFSET +
part              173 drivers/mtd/rfd_ftl.c 			part->data_sectors_per_block) * sizeof(u16);
part              175 drivers/mtd/rfd_ftl.c 	part->cylinders = (part->data_sectors_per_block *
part              176 drivers/mtd/rfd_ftl.c 			(part->total_blocks - 1) - 1) / SECTORS_PER_TRACK;
part              178 drivers/mtd/rfd_ftl.c 	part->sector_count = part->cylinders * SECTORS_PER_TRACK;
part              180 drivers/mtd/rfd_ftl.c 	part->current_block = -1;
part              181 drivers/mtd/rfd_ftl.c 	part->reserved_block = -1;
part              182 drivers/mtd/rfd_ftl.c 	part->is_reclaiming = 0;
part              184 drivers/mtd/rfd_ftl.c 	part->header_cache = kmalloc(part->header_size, GFP_KERNEL);
part              185 drivers/mtd/rfd_ftl.c 	if (!part->header_cache)
part              188 drivers/mtd/rfd_ftl.c 	part->blocks = kcalloc(part->total_blocks, sizeof(struct block),
part              190 drivers/mtd/rfd_ftl.c 	if (!part->blocks)
part              193 drivers/mtd/rfd_ftl.c 	part->sector_map = vmalloc(array_size(sizeof(u_long),
part              194 drivers/mtd/rfd_ftl.c 					      part->sector_count));
part              195 drivers/mtd/rfd_ftl.c 	if (!part->sector_map) {
part              197 drivers/mtd/rfd_ftl.c 			"sector map", part->mbd.mtd->name);
part              201 drivers/mtd/rfd_ftl.c 	for (i=0; i<part->sector_count; i++)
part              202 drivers/mtd/rfd_ftl.c 		part->sector_map[i] = -1;
part              204 drivers/mtd/rfd_ftl.c 	for (i=0, blocks_found=0; i<part->total_blocks; i++) {
part              205 drivers/mtd/rfd_ftl.c 		rc = mtd_read(part->mbd.mtd, i * part->block_size,
part              206 drivers/mtd/rfd_ftl.c 			      part->header_size, &retlen,
part              207 drivers/mtd/rfd_ftl.c 			      (u_char *)part->header_cache);
part              209 drivers/mtd/rfd_ftl.c 		if (!rc && retlen != part->header_size)
part              215 drivers/mtd/rfd_ftl.c 		if (!build_block_map(part, i))
part              221 drivers/mtd/rfd_ftl.c 				part->mbd.mtd->name);
part              226 drivers/mtd/rfd_ftl.c 	if (part->reserved_block == -1) {
part              228 drivers/mtd/rfd_ftl.c 				part->mbd.mtd->name);
part              230 drivers/mtd/rfd_ftl.c 		part->errors = 1;
part              236 drivers/mtd/rfd_ftl.c 	vfree(part->sector_map);
part              237 drivers/mtd/rfd_ftl.c 	kfree(part->header_cache);
part              238 drivers/mtd/rfd_ftl.c 	kfree(part->blocks);
part              245 drivers/mtd/rfd_ftl.c 	struct partition *part = (struct partition*)dev;
part              250 drivers/mtd/rfd_ftl.c 	if (sector >= part->sector_count)
part              253 drivers/mtd/rfd_ftl.c 	addr = part->sector_map[sector];
part              255 drivers/mtd/rfd_ftl.c 		rc = mtd_read(part->mbd.mtd, addr, SECTOR_SIZE, &retlen,
part              262 drivers/mtd/rfd_ftl.c 				"0x%lx\n", part->mbd.mtd->name, addr);
part              271 drivers/mtd/rfd_ftl.c static int erase_block(struct partition *part, int block)
part              280 drivers/mtd/rfd_ftl.c 	erase->addr = part->blocks[block].offset;
part              281 drivers/mtd/rfd_ftl.c 	erase->len = part->block_size;
part              283 drivers/mtd/rfd_ftl.c 	part->blocks[block].state = BLOCK_ERASING;
part              284 drivers/mtd/rfd_ftl.c 	part->blocks[block].free_sectors = 0;
part              286 drivers/mtd/rfd_ftl.c 	rc = mtd_erase(part->mbd.mtd, erase);
part              290 drivers/mtd/rfd_ftl.c 				(unsigned long long)erase->len, part->mbd.mtd->name);
part              291 drivers/mtd/rfd_ftl.c 		part->blocks[block].state = BLOCK_FAILED;
part              292 drivers/mtd/rfd_ftl.c 		part->blocks[block].free_sectors = 0;
part              293 drivers/mtd/rfd_ftl.c 		part->blocks[block].used_sectors = 0;
part              298 drivers/mtd/rfd_ftl.c 		part->blocks[block].state = BLOCK_ERASED;
part              299 drivers/mtd/rfd_ftl.c 		part->blocks[block].free_sectors = part->data_sectors_per_block;
part              300 drivers/mtd/rfd_ftl.c 		part->blocks[block].used_sectors = 0;
part              301 drivers/mtd/rfd_ftl.c 		part->blocks[block].erases++;
part              303 drivers/mtd/rfd_ftl.c 		rc = mtd_write(part->mbd.mtd, part->blocks[block].offset,
part              310 drivers/mtd/rfd_ftl.c 			       part->mbd.mtd->name, part->blocks[block].offset);
part              311 drivers/mtd/rfd_ftl.c 			part->blocks[block].state = BLOCK_FAILED;
part              313 drivers/mtd/rfd_ftl.c 			part->blocks[block].state = BLOCK_OK;
part              322 drivers/mtd/rfd_ftl.c static int move_block_contents(struct partition *part, int block_no, u_long *old_sector)
part              329 drivers/mtd/rfd_ftl.c 	part->is_reclaiming = 1;
part              335 drivers/mtd/rfd_ftl.c 	map = kmalloc(part->header_size, GFP_KERNEL);
part              339 drivers/mtd/rfd_ftl.c 	rc = mtd_read(part->mbd.mtd, part->blocks[block_no].offset,
part              340 drivers/mtd/rfd_ftl.c 		      part->header_size, &retlen, (u_char *)map);
part              342 drivers/mtd/rfd_ftl.c 	if (!rc && retlen != part->header_size)
part              347 drivers/mtd/rfd_ftl.c 			"0x%lx\n", part->mbd.mtd->name,
part              348 drivers/mtd/rfd_ftl.c 			part->blocks[block_no].offset);
part              353 drivers/mtd/rfd_ftl.c 	for (i=0; i<part->data_sectors_per_block; i++) {
part              365 drivers/mtd/rfd_ftl.c 		if (entry >= part->sector_count)
part              368 drivers/mtd/rfd_ftl.c 		addr = part->blocks[block_no].offset +
part              369 drivers/mtd/rfd_ftl.c 			(i + part->header_sectors_per_block) * SECTOR_SIZE;
part              373 drivers/mtd/rfd_ftl.c 			if (!part->blocks[block_no].used_sectors--) {
part              374 drivers/mtd/rfd_ftl.c 				rc = erase_block(part, block_no);
part              379 drivers/mtd/rfd_ftl.c 		rc = mtd_read(part->mbd.mtd, addr, SECTOR_SIZE, &retlen,
part              388 drivers/mtd/rfd_ftl.c 				part->mbd.mtd->name);
part              393 drivers/mtd/rfd_ftl.c 		rc = rfd_ftl_writesect((struct mtd_blktrans_dev*)part,
part              405 drivers/mtd/rfd_ftl.c 	part->is_reclaiming = 0;
part              410 drivers/mtd/rfd_ftl.c static int reclaim_block(struct partition *part, u_long *old_sector)
part              416 drivers/mtd/rfd_ftl.c 	mtd_sync(part->mbd.mtd);
part              421 drivers/mtd/rfd_ftl.c 		old_sector_block = *old_sector / part->block_size;
part              425 drivers/mtd/rfd_ftl.c 	for (block=0; block<part->total_blocks; block++) {
part              428 drivers/mtd/rfd_ftl.c 		if (block == part->reserved_block)
part              436 drivers/mtd/rfd_ftl.c 		if (part->blocks[block].free_sectors)
part              439 drivers/mtd/rfd_ftl.c 		this_score = part->blocks[block].used_sectors;
part              445 drivers/mtd/rfd_ftl.c 			if (part->blocks[block].used_sectors ==
part              446 drivers/mtd/rfd_ftl.c 					part->data_sectors_per_block)
part              450 drivers/mtd/rfd_ftl.c 		this_score += part->blocks[block].erases;
part              461 drivers/mtd/rfd_ftl.c 	part->current_block = -1;
part              462 drivers/mtd/rfd_ftl.c 	part->reserved_block = best_block;
part              466 drivers/mtd/rfd_ftl.c 		 part->blocks[best_block].used_sectors,
part              467 drivers/mtd/rfd_ftl.c 		 part->blocks[best_block].free_sectors);
part              469 drivers/mtd/rfd_ftl.c 	if (part->blocks[best_block].used_sectors)
part              470 drivers/mtd/rfd_ftl.c 		rc = move_block_contents(part, best_block, old_sector);
part              472 drivers/mtd/rfd_ftl.c 		rc = erase_block(part, best_block);
part              482 drivers/mtd/rfd_ftl.c static int find_free_block(struct partition *part)
part              486 drivers/mtd/rfd_ftl.c 	block = part->current_block == -1 ?
part              487 drivers/mtd/rfd_ftl.c 			jiffies % part->total_blocks : part->current_block;
part              491 drivers/mtd/rfd_ftl.c 		if (part->blocks[block].free_sectors &&
part              492 drivers/mtd/rfd_ftl.c 				block != part->reserved_block)
part              495 drivers/mtd/rfd_ftl.c 		if (part->blocks[block].state == BLOCK_UNUSED)
part              496 drivers/mtd/rfd_ftl.c 			erase_block(part, block);
part              498 drivers/mtd/rfd_ftl.c 		if (++block >= part->total_blocks)
part              506 drivers/mtd/rfd_ftl.c static int find_writable_block(struct partition *part, u_long *old_sector)
part              511 drivers/mtd/rfd_ftl.c 	block = find_free_block(part);
part              514 drivers/mtd/rfd_ftl.c 		if (!part->is_reclaiming) {
part              515 drivers/mtd/rfd_ftl.c 			rc = reclaim_block(part, old_sector);
part              519 drivers/mtd/rfd_ftl.c 			block = find_free_block(part);
part              528 drivers/mtd/rfd_ftl.c 	rc = mtd_read(part->mbd.mtd, part->blocks[block].offset,
part              529 drivers/mtd/rfd_ftl.c 		      part->header_size, &retlen,
part              530 drivers/mtd/rfd_ftl.c 		      (u_char *)part->header_cache);
part              532 drivers/mtd/rfd_ftl.c 	if (!rc && retlen != part->header_size)
part              537 drivers/mtd/rfd_ftl.c 				"0x%lx\n", part->mbd.mtd->name,
part              538 drivers/mtd/rfd_ftl.c 				part->blocks[block].offset);
part              542 drivers/mtd/rfd_ftl.c 	part->current_block = block;
part              548 drivers/mtd/rfd_ftl.c static int mark_sector_deleted(struct partition *part, u_long old_addr)
part              555 drivers/mtd/rfd_ftl.c 	block = old_addr / part->block_size;
part              556 drivers/mtd/rfd_ftl.c 	offset = (old_addr % part->block_size) / SECTOR_SIZE -
part              557 drivers/mtd/rfd_ftl.c 		part->header_sectors_per_block;
part              559 drivers/mtd/rfd_ftl.c 	addr = part->blocks[block].offset +
part              561 drivers/mtd/rfd_ftl.c 	rc = mtd_write(part->mbd.mtd, addr, sizeof(del), &retlen,
part              569 drivers/mtd/rfd_ftl.c 			"0x%lx\n", part->mbd.mtd->name, addr);
part              572 drivers/mtd/rfd_ftl.c 	if (block == part->current_block)
part              573 drivers/mtd/rfd_ftl.c 		part->header_cache[offset + HEADER_MAP_OFFSET] = del;
part              575 drivers/mtd/rfd_ftl.c 	part->blocks[block].used_sectors--;
part              577 drivers/mtd/rfd_ftl.c 	if (!part->blocks[block].used_sectors &&
part              578 drivers/mtd/rfd_ftl.c 	    !part->blocks[block].free_sectors)
part              579 drivers/mtd/rfd_ftl.c 		rc = erase_block(part, block);
part              585 drivers/mtd/rfd_ftl.c static int find_free_sector(const struct partition *part, const struct block *block)
part              589 drivers/mtd/rfd_ftl.c 	i = stop = part->data_sectors_per_block - block->free_sectors;
part              592 drivers/mtd/rfd_ftl.c 		if (le16_to_cpu(part->header_cache[HEADER_MAP_OFFSET + i])
part              596 drivers/mtd/rfd_ftl.c 		if (++i == part->data_sectors_per_block)
part              606 drivers/mtd/rfd_ftl.c 	struct partition *part = (struct partition*)dev;
part              614 drivers/mtd/rfd_ftl.c 	if (part->current_block == -1 ||
part              615 drivers/mtd/rfd_ftl.c 		!part->blocks[part->current_block].free_sectors) {
part              617 drivers/mtd/rfd_ftl.c 		rc = find_writable_block(part, old_addr);
part              622 drivers/mtd/rfd_ftl.c 	block = &part->blocks[part->current_block];
part              624 drivers/mtd/rfd_ftl.c 	i = find_free_sector(part, block);
part              631 drivers/mtd/rfd_ftl.c 	addr = (i + part->header_sectors_per_block) * SECTOR_SIZE +
part              633 drivers/mtd/rfd_ftl.c 	rc = mtd_write(part->mbd.mtd, addr, SECTOR_SIZE, &retlen,
part              641 drivers/mtd/rfd_ftl.c 				part->mbd.mtd->name, addr);
part              645 drivers/mtd/rfd_ftl.c 	part->sector_map[sector] = addr;
part              649 drivers/mtd/rfd_ftl.c 	part->header_cache[i + HEADER_MAP_OFFSET] = entry;
part              652 drivers/mtd/rfd_ftl.c 	rc = mtd_write(part->mbd.mtd, addr, sizeof(entry), &retlen,
part              660 drivers/mtd/rfd_ftl.c 				part->mbd.mtd->name, addr);
part              672 drivers/mtd/rfd_ftl.c 	struct partition *part = (struct partition*)dev;
part              679 drivers/mtd/rfd_ftl.c 	if (part->reserved_block == -1) {
part              684 drivers/mtd/rfd_ftl.c 	if (sector >= part->sector_count) {
part              689 drivers/mtd/rfd_ftl.c 	old_addr = part->sector_map[sector];
part              702 drivers/mtd/rfd_ftl.c 		part->sector_map[sector] = -1;
part              705 drivers/mtd/rfd_ftl.c 		rc = mark_sector_deleted(part, old_addr);
part              713 drivers/mtd/rfd_ftl.c 	struct partition *part = (struct partition*)dev;
part              717 drivers/mtd/rfd_ftl.c 	geo->cylinders = part->cylinders;
part              724 drivers/mtd/rfd_ftl.c 	struct partition *part;
part              729 drivers/mtd/rfd_ftl.c 	part = kzalloc(sizeof(struct partition), GFP_KERNEL);
part              730 drivers/mtd/rfd_ftl.c 	if (!part)
part              733 drivers/mtd/rfd_ftl.c 	part->mbd.mtd = mtd;
part              736 drivers/mtd/rfd_ftl.c 		part->block_size = block_size;
part              742 drivers/mtd/rfd_ftl.c 			part->block_size = mtd->erasesize;
part              745 drivers/mtd/rfd_ftl.c 	if (scan_header(part) == 0) {
part              746 drivers/mtd/rfd_ftl.c 		part->mbd.size = part->sector_count;
part              747 drivers/mtd/rfd_ftl.c 		part->mbd.tr = tr;
part              748 drivers/mtd/rfd_ftl.c 		part->mbd.devnum = -1;
part              750 drivers/mtd/rfd_ftl.c 			part->mbd.readonly = 1;
part              751 drivers/mtd/rfd_ftl.c 		else if (part->errors) {
part              754 drivers/mtd/rfd_ftl.c 			part->mbd.readonly = 1;
part              760 drivers/mtd/rfd_ftl.c 		if (!add_mtd_blktrans_dev((void*)part))
part              764 drivers/mtd/rfd_ftl.c 	kfree(part);
part              769 drivers/mtd/rfd_ftl.c 	struct partition *part = (struct partition*)dev;
part              772 drivers/mtd/rfd_ftl.c 	for (i=0; i<part->total_blocks; i++) {
part              774 drivers/mtd/rfd_ftl.c 			part->mbd.mtd->name, i, part->blocks[i].erases);
part              778 drivers/mtd/rfd_ftl.c 	vfree(part->sector_map);
part              779 drivers/mtd/rfd_ftl.c 	kfree(part->header_cache);
part              780 drivers/mtd/rfd_ftl.c 	kfree(part->blocks);
part              829 drivers/mtd/spi-nor/intel-spi.c 				     struct mtd_partition *part)
part              834 drivers/mtd/spi-nor/intel-spi.c 	memset(part, 0, sizeof(*part));
part              837 drivers/mtd/spi-nor/intel-spi.c 	part->size = 4096;
part              838 drivers/mtd/spi-nor/intel-spi.c 	part->name = "BIOS";
part              862 drivers/mtd/spi-nor/intel-spi.c 		if (end > part->size)
part              863 drivers/mtd/spi-nor/intel-spi.c 			part->size = end;
part              875 drivers/mtd/spi-nor/intel-spi.c 	struct mtd_partition part;
part              912 drivers/mtd/spi-nor/intel-spi.c 	intel_spi_fill_partition(ispi, &part);
part              918 drivers/mtd/spi-nor/intel-spi.c 	ret = mtd_device_register(&ispi->nor.mtd, &part, 1);
part              401 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c 		u32 part;
part              415 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c 		part = (speed * 100) / line_speed;
part              417 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c 		if (line_speed < speed || !part) {
part              425 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c 			bp->pending_max = part;
part              427 drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c 			bnx2x_update_max_mf_config(bp, part);
part              280 drivers/net/ethernet/brocade/bna/bfa_defs.h 	struct bfa_flash_part_attr part[BFA_FLASH_PART_MAX];
part             3146 drivers/net/ethernet/brocade/bna/bfa_ioc.c 				attr->part[i].part_type =
part             3147 drivers/net/ethernet/brocade/bna/bfa_ioc.c 					be32_to_cpu(f->part[i].part_type);
part             3148 drivers/net/ethernet/brocade/bna/bfa_ioc.c 				attr->part[i].part_instance =
part             3149 drivers/net/ethernet/brocade/bna/bfa_ioc.c 					be32_to_cpu(f->part[i].part_instance);
part             3150 drivers/net/ethernet/brocade/bna/bfa_ioc.c 				attr->part[i].part_off =
part             3151 drivers/net/ethernet/brocade/bna/bfa_ioc.c 					be32_to_cpu(f->part[i].part_off);
part             3152 drivers/net/ethernet/brocade/bna/bfa_ioc.c 				attr->part[i].part_size =
part             3153 drivers/net/ethernet/brocade/bna/bfa_ioc.c 					be32_to_cpu(f->part[i].part_size);
part             3154 drivers/net/ethernet/brocade/bna/bfa_ioc.c 				attr->part[i].part_len =
part             3155 drivers/net/ethernet/brocade/bna/bfa_ioc.c 					be32_to_cpu(f->part[i].part_len);
part             3156 drivers/net/ethernet/brocade/bna/bfa_ioc.c 				attr->part[i].part_status =
part             3157 drivers/net/ethernet/brocade/bna/bfa_ioc.c 					be32_to_cpu(f->part[i].part_status);
part              975 drivers/net/ethernet/brocade/bna/bnad_ethtool.c 			if (offset >= flash_attr->part[i].part_off &&
part              976 drivers/net/ethernet/brocade/bna/bnad_ethtool.c 			    offset < (flash_attr->part[i].part_off +
part              977 drivers/net/ethernet/brocade/bna/bnad_ethtool.c 				      flash_attr->part[i].part_size)) {
part              978 drivers/net/ethernet/brocade/bna/bnad_ethtool.c 				flash_part = flash_attr->part[i].part_type;
part              979 drivers/net/ethernet/brocade/bna/bnad_ethtool.c 				*base_offset = flash_attr->part[i].part_off;
part             8953 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	unsigned int part, manufacturer;
part             8973 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 	for (part = 0; part < ARRAY_SIZE(supported_flash); part++)
part             8974 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 		if (supported_flash[part].vendor_and_model_id == flashid) {
part             8975 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 			adap->params.sf_size = supported_flash[part].size_mb;
part               74 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	struct mlxsw_sp1_kvdl_part *part, *min_part = NULL;
part               78 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 		part = kvdl->parts[i];
part               79 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 		if (alloc_size <= part->info.alloc_size &&
part               81 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 		     part->info.alloc_size <= min_part->info.alloc_size))
part               82 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 			min_part = part;
part               91 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	struct mlxsw_sp1_kvdl_part *part;
part               95 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 		part = kvdl->parts[i];
part               96 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 		if (kvdl_index >= part->info.start_index &&
part               97 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 		    kvdl_index <= part->info.end_index)
part               98 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 			return part;
part              118 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c static int mlxsw_sp1_kvdl_part_alloc(struct mlxsw_sp1_kvdl_part *part,
part              121 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	const struct mlxsw_sp1_kvdl_part_info *info = &part->info;
part              126 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	entry_index = find_first_zero_bit(part->usage, nr_entries);
part              129 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	__set_bit(entry_index, part->usage);
part              136 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c static void mlxsw_sp1_kvdl_part_free(struct mlxsw_sp1_kvdl_part *part,
part              139 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	const struct mlxsw_sp1_kvdl_part_info *info = &part->info;
part              143 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	__clear_bit(entry_index, part->usage);
part              152 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	struct mlxsw_sp1_kvdl_part *part;
part              157 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	part = mlxsw_sp1_kvdl_alloc_size_part(kvdl, entry_count);
part              158 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	if (IS_ERR(part))
part              159 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 		return PTR_ERR(part);
part              161 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	return mlxsw_sp1_kvdl_part_alloc(part, p_entry_index);
part              169 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	struct mlxsw_sp1_kvdl_part *part;
part              171 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	part = mlxsw_sp1_kvdl_index_part(kvdl, entry_index);
part              172 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	if (IS_ERR(part))
part              174 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	mlxsw_sp1_kvdl_part_free(part, entry_index);
part              184 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	struct mlxsw_sp1_kvdl_part *part;
part              186 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	part = mlxsw_sp1_kvdl_alloc_size_part(kvdl, entry_count);
part              187 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	if (IS_ERR(part))
part              188 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 		return PTR_ERR(part);
part              190 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	*p_alloc_size = part->info.alloc_size;
part              195 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c static void mlxsw_sp1_kvdl_part_update(struct mlxsw_sp1_kvdl_part *part,
part              200 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 		part->info.end_index = size - 1;
part              202 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 		part->info.start_index = part_prev->info.end_index + 1;
part              203 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 		part->info.end_index = part->info.start_index + size - 1;
part              213 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	struct mlxsw_sp1_kvdl_part *part;
part              229 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	part = kzalloc(sizeof(*part) + usage_size, GFP_KERNEL);
part              230 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	if (!part)
part              233 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	memcpy(&part->info, info, sizeof(part->info));
part              236 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 		mlxsw_sp1_kvdl_part_update(part, part_prev, resource_size);
part              237 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	return part;
part              240 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c static void mlxsw_sp1_kvdl_part_fini(struct mlxsw_sp1_kvdl_part *part)
part              242 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	kfree(part);
part              278 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c static u64 mlxsw_sp1_kvdl_part_occ(struct mlxsw_sp1_kvdl_part *part)
part              280 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	const struct mlxsw_sp1_kvdl_part_info *info = &part->info;
part              288 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	while ((bit = find_next_bit(part->usage, nr_entries, bit + 1))
part              309 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	struct mlxsw_sp1_kvdl_part *part;
part              311 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	part = kvdl->parts[MLXSW_SP1_KVDL_PART_ID_SINGLE];
part              312 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	return mlxsw_sp1_kvdl_part_occ(part);
part              318 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	struct mlxsw_sp1_kvdl_part *part;
part              320 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	part = kvdl->parts[MLXSW_SP1_KVDL_PART_ID_CHUNKS];
part              321 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	return mlxsw_sp1_kvdl_part_occ(part);
part              327 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	struct mlxsw_sp1_kvdl_part *part;
part              329 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	part = kvdl->parts[MLXSW_SP1_KVDL_PART_ID_LARGE_CHUNKS];
part              330 drivers/net/ethernet/mellanox/mlxsw/spectrum1_kvdl.c 	return mlxsw_sp1_kvdl_part_occ(part);
part               55 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c static int mlxsw_sp2_kvdl_part_find_zero_bits(struct mlxsw_sp2_kvdl_part *part,
part               64 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 	start_bit = part->last_allocated_bit + 1;
part               65 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 	if (start_bit == part->usage_bit_count)
part               69 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 	bit = find_next_zero_bit(part->usage, part->usage_bit_count, bit);
part               70 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 	if (!wrap && bit + bit_count >= part->usage_bit_count) {
part               78 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 		if (test_bit(bit + i, part->usage)) {
part               87 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c static int mlxsw_sp2_kvdl_part_alloc(struct mlxsw_sp2_kvdl_part *part,
part               96 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 	bit_count = DIV_ROUND_UP(size, part->indexes_per_usage_bit);
part               97 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 	err = mlxsw_sp2_kvdl_part_find_zero_bits(part, bit_count, &bit);
part              101 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 		__set_bit(bit + i, part->usage);
part              102 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 	*p_kvdl_index = bit * part->indexes_per_usage_bit;
part              124 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 				     struct mlxsw_sp2_kvdl_part *part,
part              133 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 	err = mlxsw_sp2_kvdl_rec_del(mlxsw_sp, part->info->res_type,
part              138 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 	bit_count = DIV_ROUND_UP(size, part->indexes_per_usage_bit);
part              139 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 	bit = kvdl_index / part->indexes_per_usage_bit;
part              141 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 		__clear_bit(bit + i, part->usage);
part              151 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 	struct mlxsw_sp2_kvdl_part *part = kvdl->parts[type];
part              153 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 	return mlxsw_sp2_kvdl_part_alloc(part, size, p_entry_index);
part              163 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 	struct mlxsw_sp2_kvdl_part *part = kvdl->parts[type];
part              165 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 	return mlxsw_sp2_kvdl_part_free(mlxsw_sp, part, size, entry_index);
part              183 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 	struct mlxsw_sp2_kvdl_part *part;
part              205 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 	part = kzalloc(sizeof(*part) + usage_size, GFP_KERNEL);
part              206 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 	if (!part)
part              208 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 	part->info = info;
part              209 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 	part->usage_bit_count = usage_bit_count;
part              210 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 	part->indexes_per_usage_bit = indexes_per_usage_bit;
part              211 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 	part->last_allocated_bit = usage_bit_count - 1;
part              212 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 	return part;
part              215 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c static void mlxsw_sp2_kvdl_part_fini(struct mlxsw_sp2_kvdl_part *part)
part              217 drivers/net/ethernet/mellanox/mlxsw/spectrum2_kvdl.c 	kfree(part);
part              262 drivers/net/ethernet/netronome/nfp/nfp_devlink.c 	const char *sn, *vendor, *part;
part              272 drivers/net/ethernet/netronome/nfp/nfp_devlink.c 	part = nfp_hwinfo_lookup(pf->hwinfo, "assembly.partno");
part              274 drivers/net/ethernet/netronome/nfp/nfp_devlink.c 	if (vendor && part && sn) {
part              277 drivers/net/ethernet/netronome/nfp/nfp_devlink.c 		buf = kmalloc(strlen(vendor) + strlen(part) + strlen(sn) + 1,
part              284 drivers/net/ethernet/netronome/nfp/nfp_devlink.c 		strcat(buf, part);
part             6052 drivers/net/ethernet/sfc/ef10.c 					struct efx_mcdi_mtd_partition *part,
part             6094 drivers/net/ethernet/sfc/ef10.c 	part->nvram_type = type;
part             6105 drivers/net/ethernet/sfc/ef10.c 		part->fw_subtype = MCDI_DWORD(outbuf,
part             6108 drivers/net/ethernet/sfc/ef10.c 	part->common.dev_type_name = "EF10 NVRAM manager";
part             6109 drivers/net/ethernet/sfc/ef10.c 	part->common.type_name = info->name;
part             6111 drivers/net/ethernet/sfc/ef10.c 	part->common.mtd.type = MTD_NORFLASH;
part             6112 drivers/net/ethernet/sfc/ef10.c 	part->common.mtd.flags = MTD_CAP_NORFLASH;
part             6113 drivers/net/ethernet/sfc/ef10.c 	part->common.mtd.size = size;
part             6114 drivers/net/ethernet/sfc/ef10.c 	part->common.mtd.erasesize = erase_size;
part             6117 drivers/net/ethernet/sfc/ef10.c 		part->common.mtd.flags |= MTD_NO_ERASE;
part              740 drivers/net/ethernet/sfc/falcon/falcon.c falcon_spi_slow_wait(struct falcon_mtd_partition *part, bool uninterruptible)
part              742 drivers/net/ethernet/sfc/falcon/falcon.c 	const struct falcon_spi_device *spi = part->spi;
part              743 drivers/net/ethernet/sfc/falcon/falcon.c 	struct ef4_nic *efx = part->common.mtd.priv;
part              762 drivers/net/ethernet/sfc/falcon/falcon.c 	       part->common.name, part->common.dev_type_name);
part              804 drivers/net/ethernet/sfc/falcon/falcon.c falcon_spi_erase(struct falcon_mtd_partition *part, loff_t start, size_t len)
part              806 drivers/net/ethernet/sfc/falcon/falcon.c 	const struct falcon_spi_device *spi = part->spi;
part              807 drivers/net/ethernet/sfc/falcon/falcon.c 	struct ef4_nic *efx = part->common.mtd.priv;
part              829 drivers/net/ethernet/sfc/falcon/falcon.c 	rc = falcon_spi_slow_wait(part, false);
part              851 drivers/net/ethernet/sfc/falcon/falcon.c static void falcon_mtd_rename(struct ef4_mtd_partition *part)
part              853 drivers/net/ethernet/sfc/falcon/falcon.c 	struct ef4_nic *efx = part->mtd.priv;
part              855 drivers/net/ethernet/sfc/falcon/falcon.c 	snprintf(part->name, sizeof(part->name), "%s %s",
part              856 drivers/net/ethernet/sfc/falcon/falcon.c 		 efx->name, part->type_name);
part              862 drivers/net/ethernet/sfc/falcon/falcon.c 	struct falcon_mtd_partition *part = to_falcon_mtd_partition(mtd);
part              870 drivers/net/ethernet/sfc/falcon/falcon.c 	rc = falcon_spi_read(efx, part->spi, part->offset + start,
part              878 drivers/net/ethernet/sfc/falcon/falcon.c 	struct falcon_mtd_partition *part = to_falcon_mtd_partition(mtd);
part              886 drivers/net/ethernet/sfc/falcon/falcon.c 	rc = falcon_spi_erase(part, part->offset + start, len);
part              894 drivers/net/ethernet/sfc/falcon/falcon.c 	struct falcon_mtd_partition *part = to_falcon_mtd_partition(mtd);
part              902 drivers/net/ethernet/sfc/falcon/falcon.c 	rc = falcon_spi_write(efx, part->spi, part->offset + start,
part              910 drivers/net/ethernet/sfc/falcon/falcon.c 	struct falcon_mtd_partition *part = to_falcon_mtd_partition(mtd);
part              916 drivers/net/ethernet/sfc/falcon/falcon.c 	rc = falcon_spi_slow_wait(part, true);
part               30 drivers/net/ethernet/sfc/falcon/mtd.c 	struct ef4_mtd_partition *part = to_ef4_mtd_partition(mtd);
part               37 drivers/net/ethernet/sfc/falcon/mtd.c 		       part->name, part->dev_type_name, rc);
part               40 drivers/net/ethernet/sfc/falcon/mtd.c static void ef4_mtd_remove_partition(struct ef4_mtd_partition *part)
part               45 drivers/net/ethernet/sfc/falcon/mtd.c 		rc = mtd_device_unregister(&part->mtd);
part               51 drivers/net/ethernet/sfc/falcon/mtd.c 	list_del(&part->node);
part               57 drivers/net/ethernet/sfc/falcon/mtd.c 	struct ef4_mtd_partition *part;
part               61 drivers/net/ethernet/sfc/falcon/mtd.c 		part = (struct ef4_mtd_partition *)((char *)parts +
part               64 drivers/net/ethernet/sfc/falcon/mtd.c 		part->mtd.writesize = 1;
part               66 drivers/net/ethernet/sfc/falcon/mtd.c 		part->mtd.owner = THIS_MODULE;
part               67 drivers/net/ethernet/sfc/falcon/mtd.c 		part->mtd.priv = efx;
part               68 drivers/net/ethernet/sfc/falcon/mtd.c 		part->mtd.name = part->name;
part               69 drivers/net/ethernet/sfc/falcon/mtd.c 		part->mtd._erase = ef4_mtd_erase;
part               70 drivers/net/ethernet/sfc/falcon/mtd.c 		part->mtd._read = efx->type->mtd_read;
part               71 drivers/net/ethernet/sfc/falcon/mtd.c 		part->mtd._write = efx->type->mtd_write;
part               72 drivers/net/ethernet/sfc/falcon/mtd.c 		part->mtd._sync = ef4_mtd_sync;
part               74 drivers/net/ethernet/sfc/falcon/mtd.c 		efx->type->mtd_rename(part);
part               76 drivers/net/ethernet/sfc/falcon/mtd.c 		if (mtd_device_register(&part->mtd, NULL, 0))
part               80 drivers/net/ethernet/sfc/falcon/mtd.c 		list_add_tail(&part->node, &efx->mtd_list);
part               87 drivers/net/ethernet/sfc/falcon/mtd.c 		part = (struct ef4_mtd_partition *)((char *)parts +
part               89 drivers/net/ethernet/sfc/falcon/mtd.c 		ef4_mtd_remove_partition(part);
part               97 drivers/net/ethernet/sfc/falcon/mtd.c 	struct ef4_mtd_partition *parts, *part, *next;
part              107 drivers/net/ethernet/sfc/falcon/mtd.c 	list_for_each_entry_safe(part, next, &efx->mtd_list, node)
part              108 drivers/net/ethernet/sfc/falcon/mtd.c 		ef4_mtd_remove_partition(part);
part              115 drivers/net/ethernet/sfc/falcon/mtd.c 	struct ef4_mtd_partition *part;
part              119 drivers/net/ethernet/sfc/falcon/mtd.c 	list_for_each_entry(part, &efx->mtd_list, node)
part              120 drivers/net/ethernet/sfc/falcon/mtd.c 		efx->type->mtd_rename(part);
part             1130 drivers/net/ethernet/sfc/falcon/net_driver.h 	void (*mtd_rename)(struct ef4_mtd_partition *part);
part             2204 drivers/net/ethernet/sfc/mcdi.c 	struct efx_mcdi_mtd_partition *part = to_efx_mcdi_mtd_partition(mtd);
part             2213 drivers/net/ethernet/sfc/mcdi.c 		rc = efx_mcdi_nvram_read(efx, part->nvram_type, offset,
part             2227 drivers/net/ethernet/sfc/mcdi.c 	struct efx_mcdi_mtd_partition *part = to_efx_mcdi_mtd_partition(mtd);
part             2231 drivers/net/ethernet/sfc/mcdi.c 	size_t chunk = part->common.mtd.erasesize;
part             2234 drivers/net/ethernet/sfc/mcdi.c 	if (!part->updating) {
part             2235 drivers/net/ethernet/sfc/mcdi.c 		rc = efx_mcdi_nvram_update_start(efx, part->nvram_type);
part             2238 drivers/net/ethernet/sfc/mcdi.c 		part->updating = true;
part             2245 drivers/net/ethernet/sfc/mcdi.c 		rc = efx_mcdi_nvram_erase(efx, part->nvram_type, offset,
part             2258 drivers/net/ethernet/sfc/mcdi.c 	struct efx_mcdi_mtd_partition *part = to_efx_mcdi_mtd_partition(mtd);
part             2265 drivers/net/ethernet/sfc/mcdi.c 	if (!part->updating) {
part             2266 drivers/net/ethernet/sfc/mcdi.c 		rc = efx_mcdi_nvram_update_start(efx, part->nvram_type);
part             2269 drivers/net/ethernet/sfc/mcdi.c 		part->updating = true;
part             2274 drivers/net/ethernet/sfc/mcdi.c 		rc = efx_mcdi_nvram_write(efx, part->nvram_type, offset,
part             2288 drivers/net/ethernet/sfc/mcdi.c 	struct efx_mcdi_mtd_partition *part = to_efx_mcdi_mtd_partition(mtd);
part             2292 drivers/net/ethernet/sfc/mcdi.c 	if (part->updating) {
part             2293 drivers/net/ethernet/sfc/mcdi.c 		part->updating = false;
part             2294 drivers/net/ethernet/sfc/mcdi.c 		rc = efx_mcdi_nvram_update_finish(efx, part->nvram_type);
part             2300 drivers/net/ethernet/sfc/mcdi.c void efx_mcdi_mtd_rename(struct efx_mtd_partition *part)
part             2303 drivers/net/ethernet/sfc/mcdi.c 		container_of(part, struct efx_mcdi_mtd_partition, common);
part             2304 drivers/net/ethernet/sfc/mcdi.c 	struct efx_nic *efx = part->mtd.priv;
part             2306 drivers/net/ethernet/sfc/mcdi.c 	snprintf(part->name, sizeof(part->name), "%s %s:%02x",
part             2307 drivers/net/ethernet/sfc/mcdi.c 		 efx->name, part->type_name, mcdi_part->fw_subtype);
part              380 drivers/net/ethernet/sfc/mcdi.h void efx_mcdi_mtd_rename(struct efx_mtd_partition *part);
part               30 drivers/net/ethernet/sfc/mtd.c 	struct efx_mtd_partition *part = to_efx_mtd_partition(mtd);
part               37 drivers/net/ethernet/sfc/mtd.c 		       part->name, part->dev_type_name, rc);
part               40 drivers/net/ethernet/sfc/mtd.c static void efx_mtd_remove_partition(struct efx_mtd_partition *part)
part               45 drivers/net/ethernet/sfc/mtd.c 		rc = mtd_device_unregister(&part->mtd);
part               51 drivers/net/ethernet/sfc/mtd.c 	list_del(&part->node);
part               57 drivers/net/ethernet/sfc/mtd.c 	struct efx_mtd_partition *part;
part               61 drivers/net/ethernet/sfc/mtd.c 		part = (struct efx_mtd_partition *)((char *)parts +
part               64 drivers/net/ethernet/sfc/mtd.c 		part->mtd.writesize = 1;
part               66 drivers/net/ethernet/sfc/mtd.c 		if (!(part->mtd.flags & MTD_NO_ERASE))
part               67 drivers/net/ethernet/sfc/mtd.c 			part->mtd.flags |= MTD_WRITEABLE;
part               69 drivers/net/ethernet/sfc/mtd.c 		part->mtd.owner = THIS_MODULE;
part               70 drivers/net/ethernet/sfc/mtd.c 		part->mtd.priv = efx;
part               71 drivers/net/ethernet/sfc/mtd.c 		part->mtd.name = part->name;
part               72 drivers/net/ethernet/sfc/mtd.c 		part->mtd._erase = efx_mtd_erase;
part               73 drivers/net/ethernet/sfc/mtd.c 		part->mtd._read = efx->type->mtd_read;
part               74 drivers/net/ethernet/sfc/mtd.c 		part->mtd._write = efx->type->mtd_write;
part               75 drivers/net/ethernet/sfc/mtd.c 		part->mtd._sync = efx_mtd_sync;
part               77 drivers/net/ethernet/sfc/mtd.c 		efx->type->mtd_rename(part);
part               79 drivers/net/ethernet/sfc/mtd.c 		if (mtd_device_register(&part->mtd, NULL, 0))
part               83 drivers/net/ethernet/sfc/mtd.c 		list_add_tail(&part->node, &efx->mtd_list);
part               90 drivers/net/ethernet/sfc/mtd.c 		part = (struct efx_mtd_partition *)((char *)parts +
part               92 drivers/net/ethernet/sfc/mtd.c 		efx_mtd_remove_partition(part);
part              100 drivers/net/ethernet/sfc/mtd.c 	struct efx_mtd_partition *parts, *part, *next;
part              110 drivers/net/ethernet/sfc/mtd.c 	list_for_each_entry_safe(part, next, &efx->mtd_list, node)
part              111 drivers/net/ethernet/sfc/mtd.c 		efx_mtd_remove_partition(part);
part              118 drivers/net/ethernet/sfc/mtd.c 	struct efx_mtd_partition *part;
part              122 drivers/net/ethernet/sfc/mtd.c 	list_for_each_entry(part, &efx->mtd_list, node)
part              123 drivers/net/ethernet/sfc/mtd.c 		efx->type->mtd_rename(part);
part             1375 drivers/net/ethernet/sfc/net_driver.h 	void (*mtd_rename)(struct efx_mtd_partition *part);
part              848 drivers/net/ethernet/sfc/siena.c 				     struct efx_mcdi_mtd_partition *part,
part              871 drivers/net/ethernet/sfc/siena.c 	part->nvram_type = type;
part              872 drivers/net/ethernet/sfc/siena.c 	part->common.dev_type_name = "Siena NVRAM manager";
part              873 drivers/net/ethernet/sfc/siena.c 	part->common.type_name = info->name;
part              875 drivers/net/ethernet/sfc/siena.c 	part->common.mtd.type = MTD_NORFLASH;
part              876 drivers/net/ethernet/sfc/siena.c 	part->common.mtd.flags = MTD_CAP_NORFLASH;
part              877 drivers/net/ethernet/sfc/siena.c 	part->common.mtd.size = size;
part              878 drivers/net/ethernet/sfc/siena.c 	part->common.mtd.erasesize = erase_size;
part             1508 drivers/net/ieee802154/at86rf230.c 	unsigned int part, version, val;
part             1523 drivers/net/ieee802154/at86rf230.c 	rc = __at86rf230_read(lp, RG_PART_NUM, &part);
part             1553 drivers/net/ieee802154/at86rf230.c 	switch (part) {
part              963 drivers/net/wireless/cisco/airo.c 	} part;	// saves partial message word across update() calls
part             1680 drivers/net/wireless/cisco/airo.c 			context->part.d8[byte_position++] = *pOctets++;
part             1684 drivers/net/wireless/cisco/airo.c 		MIC_ACCUM(ntohl(context->part.d32));
part             1698 drivers/net/wireless/cisco/airo.c 		context->part.d8[byte_position++] = *pOctets++;
part             1722 drivers/net/wireless/cisco/airo.c 		val = ntohl(context->part.d32);
part               19 drivers/net/wireless/mediatek/mt76/eeprom.c 	const char *part;
part               41 drivers/net/wireless/mediatek/mt76/eeprom.c 	part = of_get_property(np, "label", NULL);
part               42 drivers/net/wireless/mediatek/mt76/eeprom.c 	if (!part)
part               43 drivers/net/wireless/mediatek/mt76/eeprom.c 		part = np->name;
part               45 drivers/net/wireless/mediatek/mt76/eeprom.c 	mtd = get_mtd_device_nm(part);
part             1061 drivers/net/wireless/ti/wlcore/debugfs.c 	struct wlcore_partition_set part, old_part;
part             1079 drivers/net/wireless/ti/wlcore/debugfs.c 	memset(&part, 0, sizeof(part));
part             1080 drivers/net/wireless/ti/wlcore/debugfs.c 	part.mem.start = *ppos;
part             1081 drivers/net/wireless/ti/wlcore/debugfs.c 	part.mem.size = bytes;
part             1102 drivers/net/wireless/ti/wlcore/debugfs.c 	ret = wlcore_set_partition(wl, &part);
part             1143 drivers/net/wireless/ti/wlcore/debugfs.c 	struct wlcore_partition_set part, old_part;
part             1161 drivers/net/wireless/ti/wlcore/debugfs.c 	memset(&part, 0, sizeof(part));
part             1162 drivers/net/wireless/ti/wlcore/debugfs.c 	part.mem.start = *ppos;
part             1163 drivers/net/wireless/ti/wlcore/debugfs.c 	part.mem.size = bytes;
part             1184 drivers/net/wireless/ti/wlcore/debugfs.c 	ret = wlcore_set_partition(wl, &part);
part               57 drivers/net/wireless/ti/wlcore/io.c 	struct wlcore_partition_set *part = &wl->curr_part;
part               69 drivers/net/wireless/ti/wlcore/io.c 	if ((addr >= part->mem.start) &&
part               70 drivers/net/wireless/ti/wlcore/io.c 	    (addr < part->mem.start + part->mem.size))
part               71 drivers/net/wireless/ti/wlcore/io.c 		return addr - part->mem.start;
part               72 drivers/net/wireless/ti/wlcore/io.c 	else if ((addr >= part->reg.start) &&
part               73 drivers/net/wireless/ti/wlcore/io.c 		 (addr < part->reg.start + part->reg.size))
part               74 drivers/net/wireless/ti/wlcore/io.c 		return addr - part->reg.start + part->mem.size;
part               75 drivers/net/wireless/ti/wlcore/io.c 	else if ((addr >= part->mem2.start) &&
part               76 drivers/net/wireless/ti/wlcore/io.c 		 (addr < part->mem2.start + part->mem2.size))
part               77 drivers/net/wireless/ti/wlcore/io.c 		return addr - part->mem2.start + part->mem.size +
part               78 drivers/net/wireless/ti/wlcore/io.c 			part->reg.size;
part               79 drivers/net/wireless/ti/wlcore/io.c 	else if ((addr >= part->mem3.start) &&
part               80 drivers/net/wireless/ti/wlcore/io.c 		 (addr < part->mem3.start + part->mem3.size))
part               81 drivers/net/wireless/ti/wlcore/io.c 		return addr - part->mem3.start + part->mem.size +
part               82 drivers/net/wireless/ti/wlcore/io.c 			part->reg.size + part->mem2.size;
part              482 drivers/ntb/hw/idt/ntb_hw_idt.c 	unsigned char pidx, port, part;
part              491 drivers/ntb/hw/idt/ntb_hw_idt.c 	ndev->part = GET_FIELD(SWPORTxSTS_SWPART, portsts);
part              510 drivers/ntb/hw/idt/ntb_hw_idt.c 		part = GET_FIELD(SWPORTxSTS_SWPART, portsts);
part              513 drivers/ntb/hw/idt/ntb_hw_idt.c 		partsts = idt_sw_read(ndev, partdata_tbl[part].sts);
part              522 drivers/ntb/hw/idt/ntb_hw_idt.c 			ndev->peers[ndev->peer_cnt].part = part;
part              525 drivers/ntb/hw/idt/ntb_hw_idt.c 			ndev->part_idx_map[part] = ndev->peer_cnt;
part              645 drivers/ntb/hw/idt/ntb_hw_idt.c 	part_mask = ~BIT(ndev->part);
part              648 drivers/ntb/hw/idt/ntb_hw_idt.c 		part_mask &= ~BIT(ndev->peers[pidx].part);
part              746 drivers/ntb/hw/idt/ntb_hw_idt.c 		   SET_FIELD(NTMTBLDATA_PART, 0, ndev->part) |
part              749 drivers/ntb/hw/idt/ntb_hw_idt.c 	idt_nt_write(ndev, IDT_NT_NTMTBLADDR, ndev->part);
part              755 drivers/ntb/hw/idt/ntb_hw_idt.c 	idt_sw_write(ndev, IDT_SW_SEGSIGSTS, (u32)1 << ndev->part);
part              776 drivers/ntb/hw/idt/ntb_hw_idt.c 	idt_nt_write(ndev, IDT_NT_NTMTBLADDR, ndev->part);
part              782 drivers/ntb/hw/idt/ntb_hw_idt.c 	idt_sw_write(ndev, IDT_SW_SEGSIGSTS, (u32)1 << ndev->part);
part              815 drivers/ntb/hw/idt/ntb_hw_idt.c 	idt_nt_write(ndev, IDT_NT_NTMTBLADDR, ndev->part);
part              861 drivers/ntb/hw/idt/ntb_hw_idt.c 	idt_nt_write(ndev, IDT_NT_NTMTBLADDR, ndev->peers[pidx].part);
part             1316 drivers/ntb/hw/idt/ntb_hw_idt.c 		data = SET_FIELD(BARSETUP_TPART, data, ndev->peers[pidx].part);
part             1331 drivers/ntb/hw/idt/ntb_hw_idt.c 		data = SET_FIELD(LUTUDATA_PART, 0, ndev->peers[pidx].part) |
part             1802 drivers/ntb/hw/idt/ntb_hw_idt.c 		    SET_FIELD(SWPxMSGCTL_PART, 0, ndev->peers[pidx].part);
part             1807 drivers/ntb/hw/idt/ntb_hw_idt.c 	idt_sw_write(ndev, partdata_tbl[ndev->part].msgctl[midx], swpmsgctl);
part             2361 drivers/ntb/hw/idt/ntb_hw_idt.c 		"Local Port %hhu, Partition %hhu\n", ndev->port, ndev->part);
part             2368 drivers/ntb/hw/idt/ntb_hw_idt.c 			idx, ndev->peers[idx].port, ndev->peers[idx].part);
part             1086 drivers/ntb/hw/idt/ntb_hw_idt.h 	unsigned char part;
part             1129 drivers/ntb/hw/idt/ntb_hw_idt.h 	unsigned char part;
part              683 drivers/pci/switch/switchtec.c 		s->part[i] = reg;
part              893 drivers/pci/switch/switchtec.c 	int i, part;
part              902 drivers/pci/switch/switchtec.c 	for (part = 0; part < stdev->partition_count; part++) {
part              903 drivers/pci/switch/switchtec.c 		pcfg = &stdev->mmio_part_cfg_all[part];
part              904 drivers/pci/switch/switchtec.c 		p.partition = part;
part              682 drivers/rtc/rtc-abx80x.c 	unsigned int part = id->driver_data;
part              724 drivers/rtc/rtc-abx80x.c 	if (part == RV1805) {
part              780 drivers/rtc/rtc-abx80x.c 	if (part == ABX80X) {
part              789 drivers/rtc/rtc-abx80x.c 		part = i;
part              792 drivers/rtc/rtc-abx80x.c 	if (partnumber != abx80x_caps[part].pn) {
part              794 drivers/rtc/rtc-abx80x.c 			partnumber, abx80x_caps[part].pn);
part              798 drivers/rtc/rtc-abx80x.c 	if (np && abx80x_caps[part].has_tc)
part              825 drivers/rtc/rtc-abx80x.c 	if (abx80x_caps[part].has_wdog) {
part              435 drivers/s390/block/dasd.c 	struct hd_struct *part;
part              447 drivers/s390/block/dasd.c 		while ((part = disk_part_iter_next(&piter)))
part              448 drivers/s390/block/dasd.c 			kobject_uevent(&part_to_dev(part)->kobj, KOBJ_CHANGE);
part              462 drivers/s390/block/dasd.c 	struct hd_struct *part;
part              474 drivers/s390/block/dasd.c 		while ((part = disk_part_iter_next(&piter)))
part              475 drivers/s390/block/dasd.c 			kobject_uevent(&part_to_dev(part)->kobj, KOBJ_CHANGE);
part             1140 drivers/scsi/bfa/bfa_defs.h 	struct bfa_flash_part_attr_s part[BFA_FLASH_PART_MAX];
part             4381 drivers/scsi/bfa/bfa_ioc.c 				attr->part[i].part_type =
part             4382 drivers/scsi/bfa/bfa_ioc.c 					be32_to_cpu(f->part[i].part_type);
part             4383 drivers/scsi/bfa/bfa_ioc.c 				attr->part[i].part_instance =
part             4384 drivers/scsi/bfa/bfa_ioc.c 					be32_to_cpu(f->part[i].part_instance);
part             4385 drivers/scsi/bfa/bfa_ioc.c 				attr->part[i].part_off =
part             4386 drivers/scsi/bfa/bfa_ioc.c 					be32_to_cpu(f->part[i].part_off);
part             4387 drivers/scsi/bfa/bfa_ioc.c 				attr->part[i].part_size =
part             4388 drivers/scsi/bfa/bfa_ioc.c 					be32_to_cpu(f->part[i].part_size);
part             4389 drivers/scsi/bfa/bfa_ioc.c 				attr->part[i].part_len =
part             4390 drivers/scsi/bfa/bfa_ioc.c 					be32_to_cpu(f->part[i].part_len);
part             4391 drivers/scsi/bfa/bfa_ioc.c 				attr->part[i].part_status =
part             4392 drivers/scsi/bfa/bfa_ioc.c 					be32_to_cpu(f->part[i].part_status);
part              774 drivers/scsi/csiostor/csio_hw.c 	u32 part, manufacturer;
part              788 drivers/scsi/csiostor/csio_hw.c 	for (part = 0; part < ARRAY_SIZE(supported_flash); part++)
part              789 drivers/scsi/csiostor/csio_hw.c 		if (supported_flash[part].vendor_and_model_id == flashid) {
part              790 drivers/scsi/csiostor/csio_hw.c 			hw->params.sf_size = supported_flash[part].size_mb;
part              120 drivers/staging/gs_fpgaboot/gs_fpgaboot.c 	pr_info("part: %s\n", fimage->part);
part              142 drivers/staging/gs_fpgaboot/gs_fpgaboot.c 	err = readinfo_bitstream(bitdata, fimage->part, MAX_STR, &offset);
part               36 drivers/staging/gs_fpgaboot/gs_fpgaboot.h 	char	part[MAX_STR];
part              232 drivers/vhost/vringh.c 	size_t part, len = sizeof(struct vring_desc);
part              238 drivers/vhost/vringh.c 		part = len;
part              241 drivers/vhost/vringh.c 		if (!rcheck(vrh, addr, &part, range, getrange))
part              244 drivers/vhost/vringh.c 		err = copy(dst, src, part);
part              248 drivers/vhost/vringh.c 		dst += part;
part              249 drivers/vhost/vringh.c 		src += part;
part              250 drivers/vhost/vringh.c 		len -= part;
part              422 drivers/vhost/vringh.c 		u16 part = vrh->vring.num - off;
part              423 drivers/vhost/vringh.c 		err = putused(&used_ring->ring[off], used, part);
part              425 drivers/vhost/vringh.c 			err = putused(&used_ring->ring[0], used + part,
part              426 drivers/vhost/vringh.c 				      num_used - part);
part              653 drivers/video/fbdev/amba-clcd.c 		unsigned int part;
part              679 drivers/video/fbdev/amba-clcd.c 		if (amba_part(fb->dev) != panels[i].part)
part             3008 fs/buffer.c    	struct hd_struct *part;
part             3011 fs/buffer.c    	part = __disk_get_part(bio->bi_disk, bio->bi_partno);
part             3012 fs/buffer.c    	if (part)
part             3013 fs/buffer.c    		maxsector = part_nr_sects_read(part);
part              457 fs/char_dev.c  static struct kobject *exact_match(dev_t dev, int *part, void *data)
part              656 fs/char_dev.c  static struct kobject *base_probe(dev_t dev, int *part, void *data)
part              157 fs/hfs/hfs_fs.h 	int session, part;
part               82 fs/hfs/part_tbl.c 			    (HFS_SB(sb)->part < 0 || HFS_SB(sb)->part == i)) {
part               98 fs/hfs/part_tbl.c 			    (HFS_SB(sb)->part < 0 || HFS_SB(sb)->part == i)) {
part              149 fs/hfs/super.c 	if (sbi->part >= 0)
part              150 fs/hfs/super.c 		seq_printf(seq, ",part=%u", sbi->part);
part              237 fs/hfs/super.c 	hsb->part = -1;
part              294 fs/hfs/super.c 			if (match_int(&args[0], &hsb->part)) {
part              187 fs/hfsplus/hfsplus_fs.h 	int part, session;
part               59 fs/hfsplus/options.c 	opts->part = -1;
part              157 fs/hfsplus/options.c 			if (match_int(&args[0], &sbi->part)) {
part              228 fs/hfsplus/options.c 	if (sbi->part >= 0)
part              229 fs/hfsplus/options.c 		seq_printf(seq, ",part=%u", sbi->part);
part               81 fs/hfsplus/part_tbl.c 		    (sbi->part < 0 || sbi->part == i)) {
part              102 fs/hfsplus/part_tbl.c 		    (sbi->part < 0 || sbi->part == i)) {
part               17 fs/orangefs/dir.c 	struct orangefs_dir_part *part;
part              134 fs/orangefs/dir.c 	struct orangefs_dir_part *part, *new;
part              138 fs/orangefs/dir.c 	part = od->part;
part              139 fs/orangefs/dir.c 	while (part) {
part              141 fs/orangefs/dir.c 		if (part->next)
part              142 fs/orangefs/dir.c 			part = part->next;
part              151 fs/orangefs/dir.c 	if (!od->part)
part              152 fs/orangefs/dir.c 		od->part = new;
part              154 fs/orangefs/dir.c 		part->next = new;
part              189 fs/orangefs/dir.c static int fill_from_part(struct orangefs_dir_part *part,
part              200 fs/orangefs/dir.c 	if (i > part->len)
part              210 fs/orangefs/dir.c 	while (i < part->len) {
part              211 fs/orangefs/dir.c 		if (part->len < i + sizeof *len)
part              213 fs/orangefs/dir.c 		len = (void *)part + offset + i;
part              220 fs/orangefs/dir.c 		if (part->len < i + padlen + sizeof *khandle)
part              222 fs/orangefs/dir.c 		s = (void *)part + offset + i + sizeof *len;
part              225 fs/orangefs/dir.c 		khandle = (void *)part + offset + i + padlen;
part              232 fs/orangefs/dir.c 		BUG_ON(i > part->len);
part              245 fs/orangefs/dir.c 	struct orangefs_dir_part *part;
part              250 fs/orangefs/dir.c 	part = od->part;
part              251 fs/orangefs/dir.c 	while (part->next && count) {
part              253 fs/orangefs/dir.c 		part = part->next;
part              261 fs/orangefs/dir.c 	while (part && part->len) {
part              263 fs/orangefs/dir.c 		r = fill_from_part(part, ctx);
part              276 fs/orangefs/dir.c 			part = part->next;
part              291 fs/orangefs/dir.c 		struct orangefs_dir_part *part = od->part;
part              292 fs/orangefs/dir.c 		while (part) {
part              293 fs/orangefs/dir.c 			struct orangefs_dir_part *next = part->next;
part              294 fs/orangefs/dir.c 			vfree(part);
part              295 fs/orangefs/dir.c 			part = next;
part              298 fs/orangefs/dir.c 		od->part = NULL;
part              379 fs/orangefs/dir.c 	od->part = NULL;
part              388 fs/orangefs/dir.c 	struct orangefs_dir_part *part = od->part;
part              389 fs/orangefs/dir.c 	while (part) {
part              390 fs/orangefs/dir.c 		struct orangefs_dir_part *next = part->next;
part              391 fs/orangefs/dir.c 		vfree(part);
part              392 fs/orangefs/dir.c 		part = next;
part              393 fs/pstore/platform.c 	unsigned int	part = 1;
part              424 fs/pstore/platform.c 		record.part = part;
part              437 fs/pstore/platform.c 				 oopscount, part);
part              466 fs/pstore/platform.c 		part++;
part              402 fs/pstore/ram.c 	if (record->part != 1)
part             2496 fs/udf/super.c 	unsigned int part = sbi->s_partition;
part             2497 fs/udf/super.c 	int ptype = sbi->s_partmaps[part].s_partition_type;
part             2500 fs/udf/super.c 		part = sbi->s_partmaps[part].s_type_specific.s_metadata.
part             2514 fs/udf/super.c 		if (le32_to_cpu(lvid->numOfPartitions) > part) {
part             2516 fs/udf/super.c 					lvid->freeSpaceTable[part]);
part             2525 fs/udf/super.c 	map = &sbi->s_partmaps[part];
part               96 include/linux/amba/bus.h #define AMBA_LINUX_ID(conf, rev, part) \
part               98 include/linux/amba/bus.h 	AMBA_VENDOR_LINUX << 12 | ((part) & 0xfff))
part               51 include/linux/bcm963xx_nvram.h #define BCM963XX_NVRAM_NAND_PART_OFFSET(nvram, part) \
part               52 include/linux/bcm963xx_nvram.h 	bcm963xx_nvram_nand_part_offset(nvram, BCM963XX_NVRAM_NAND_PART_ ##part)
part               56 include/linux/bcm963xx_nvram.h 	enum bcm963xx_nvram_nand_part part)
part               58 include/linux/bcm963xx_nvram.h 	return nvram->nand_part_offset[part] * SZ_1K;
part               61 include/linux/bcm963xx_nvram.h #define BCM963XX_NVRAM_NAND_PART_SIZE(nvram, part) \
part               62 include/linux/bcm963xx_nvram.h 	bcm963xx_nvram_nand_part_size(nvram, BCM963XX_NVRAM_NAND_PART_ ##part)
part               66 include/linux/bcm963xx_nvram.h 	enum bcm963xx_nvram_nand_part part)
part               68 include/linux/bcm963xx_nvram.h 	return nvram->nand_part_size[part] * SZ_1K;
part              456 include/linux/bio.h 				unsigned long sectors, struct hd_struct *part);
part              458 include/linux/bio.h 				struct hd_struct *part,
part              196 include/linux/blkdev.h 	struct hd_struct *part;
part              238 include/linux/clk/ti.h 	int	(*cm_wait_module_ready)(u8 part, s16 prcm_mod, u16 idlest_reg,
part               27 include/linux/genhd.h #define part_to_dev(part)	(&((part)->__dev))
part              164 include/linux/genhd.h 	struct hd_struct __rcu *part[];
part              224 include/linux/genhd.h static inline struct gendisk *part_to_disk(struct hd_struct *part)
part              226 include/linux/genhd.h 	if (likely(part)) {
part              227 include/linux/genhd.h 		if (part->partno)
part              228 include/linux/genhd.h 			return dev_to_disk(part_to_dev(part)->parent);
part              230 include/linux/genhd.h 			return dev_to_disk(part_to_dev(part));
part              253 include/linux/genhd.h static inline dev_t part_devt(struct hd_struct *part)
part              255 include/linux/genhd.h 	return part_to_dev(part)->devt;
part              261 include/linux/genhd.h static inline void disk_put_part(struct hd_struct *part)
part              263 include/linux/genhd.h 	if (likely(part))
part              264 include/linux/genhd.h 		put_device(part_to_dev(part));
part              277 include/linux/genhd.h 	struct hd_struct	*part;
part              306 include/linux/genhd.h #define part_stat_get_cpu(part, field, cpu)					\
part              307 include/linux/genhd.h 	(per_cpu_ptr((part)->dkstats, (cpu))->field)
part              309 include/linux/genhd.h #define part_stat_get(part, field)					\
part              310 include/linux/genhd.h 	part_stat_get_cpu(part, field, smp_processor_id())
part              312 include/linux/genhd.h #define part_stat_read(part, field)					\
part              314 include/linux/genhd.h 	typeof((part)->dkstats->field) res = 0;				\
part              317 include/linux/genhd.h 		res += per_cpu_ptr((part)->dkstats, _cpu)->field;	\
part              321 include/linux/genhd.h static inline void part_stat_set_all(struct hd_struct *part, int value)
part              326 include/linux/genhd.h 		memset(per_cpu_ptr(part->dkstats, i), value,
part              330 include/linux/genhd.h static inline int init_part_stats(struct hd_struct *part)
part              332 include/linux/genhd.h 	part->dkstats = alloc_percpu(struct disk_stats);
part              333 include/linux/genhd.h 	if (!part->dkstats)
part              338 include/linux/genhd.h static inline void free_part_stats(struct hd_struct *part)
part              340 include/linux/genhd.h 	free_percpu(part->dkstats);
part              347 include/linux/genhd.h #define part_stat_get(part, field)		((part)->dkstats.field)
part              348 include/linux/genhd.h #define part_stat_get_cpu(part, field, cpu)	part_stat_get(part, field)
part              349 include/linux/genhd.h #define part_stat_read(part, field)		part_stat_get(part, field)
part              351 include/linux/genhd.h static inline void part_stat_set_all(struct hd_struct *part, int value)
part              353 include/linux/genhd.h 	memset(&part->dkstats, value, sizeof(struct disk_stats));
part              356 include/linux/genhd.h static inline int init_part_stats(struct hd_struct *part)
part              361 include/linux/genhd.h static inline void free_part_stats(struct hd_struct *part)
part              367 include/linux/genhd.h #define part_stat_read_msecs(part, which)				\
part              368 include/linux/genhd.h 	div_u64(part_stat_read(part, nsecs[which]), NSEC_PER_MSEC)
part              370 include/linux/genhd.h #define part_stat_read_accum(part, field)				\
part              371 include/linux/genhd.h 	(part_stat_read(part, field[STAT_READ]) +			\
part              372 include/linux/genhd.h 	 part_stat_read(part, field[STAT_WRITE]) +			\
part              373 include/linux/genhd.h 	 part_stat_read(part, field[STAT_DISCARD]))
part              375 include/linux/genhd.h #define __part_stat_add(part, field, addnd)				\
part              376 include/linux/genhd.h 	(part_stat_get(part, field) += (addnd))
part              378 include/linux/genhd.h #define part_stat_add(part, field, addnd)	do {			\
part              379 include/linux/genhd.h 	__part_stat_add((part), field, addnd);				\
part              380 include/linux/genhd.h 	if ((part)->partno)						\
part              381 include/linux/genhd.h 		__part_stat_add(&part_to_disk((part))->part0,		\
part              401 include/linux/genhd.h unsigned int part_in_flight(struct request_queue *q, struct hd_struct *part);
part              402 include/linux/genhd.h void part_in_flight_rw(struct request_queue *q, struct hd_struct *part,
part              404 include/linux/genhd.h void part_dec_in_flight(struct request_queue *q, struct hd_struct *part,
part              406 include/linux/genhd.h void part_inc_in_flight(struct request_queue *q, struct hd_struct *part,
part              417 include/linux/genhd.h static inline void free_part_info(struct hd_struct *part)
part              419 include/linux/genhd.h 	kfree(part->info);
part              422 include/linux/genhd.h void update_io_ticks(struct hd_struct *part, unsigned long now);
part              618 include/linux/genhd.h extern int blk_alloc_devt(struct hd_struct *part, dev_t *devt);
part              679 include/linux/genhd.h static inline int hd_ref_init(struct hd_struct *part)
part              681 include/linux/genhd.h 	if (percpu_ref_init(&part->ref, __delete_partition, 0,
part              687 include/linux/genhd.h static inline void hd_struct_get(struct hd_struct *part)
part              689 include/linux/genhd.h 	percpu_ref_get(&part->ref);
part              692 include/linux/genhd.h static inline int hd_struct_try_get(struct hd_struct *part)
part              694 include/linux/genhd.h 	return percpu_ref_tryget_live(&part->ref);
part              697 include/linux/genhd.h static inline void hd_struct_put(struct hd_struct *part)
part              699 include/linux/genhd.h 	percpu_ref_put(&part->ref);
part              702 include/linux/genhd.h static inline void hd_struct_kill(struct hd_struct *part)
part              704 include/linux/genhd.h 	percpu_ref_kill(&part->ref);
part              707 include/linux/genhd.h static inline void hd_free_part(struct hd_struct *part)
part              709 include/linux/genhd.h 	free_part_stats(part);
part              710 include/linux/genhd.h 	free_part_info(part);
part              711 include/linux/genhd.h 	percpu_ref_exit(&part->ref);
part              723 include/linux/genhd.h static inline sector_t part_nr_sects_read(struct hd_struct *part)
part              729 include/linux/genhd.h 		seq = read_seqcount_begin(&part->nr_sects_seq);
part              730 include/linux/genhd.h 		nr_sects = part->nr_sects;
part              731 include/linux/genhd.h 	} while (read_seqcount_retry(&part->nr_sects_seq, seq));
part              737 include/linux/genhd.h 	nr_sects = part->nr_sects;
part              741 include/linux/genhd.h 	return part->nr_sects;
part              750 include/linux/genhd.h static inline void part_nr_sects_write(struct hd_struct *part, sector_t size)
part              753 include/linux/genhd.h 	write_seqcount_begin(&part->nr_sects_seq);
part              754 include/linux/genhd.h 	part->nr_sects = size;
part              755 include/linux/genhd.h 	write_seqcount_end(&part->nr_sects_seq);
part              758 include/linux/genhd.h 	part->nr_sects = size;
part              761 include/linux/genhd.h 	part->nr_sects = size;
part              308 include/linux/mmc/card.h 	struct mmc_part	part[MMC_NUM_PHY_PARTITION]; /* physical partitions */
part               80 include/linux/pstore.h 	unsigned int		part;
part               58 include/uapi/linux/switchtec_ioctl.h 	__u32 part[48];
part               67 include/uapi/linux/switchtec_ioctl.h 	__u32 part[48];
part               81 init/do_mounts.c 	struct hd_struct *part = dev_to_part(dev);
part               83 init/do_mounts.c 	if (!part->info)
part               86 init/do_mounts.c 	if (strncasecmp(cmp->uuid, part->info->uuid, cmp->len))
part              114 init/do_mounts.c 	struct hd_struct *part;
part              154 init/do_mounts.c 	part = disk_get_part(disk, dev_to_part(dev)->partno + offset);
part              155 init/do_mounts.c 	if (part) {
part              156 init/do_mounts.c 		res = part_devt(part);
part              157 init/do_mounts.c 		put_device(part_to_dev(part));
part              183 init/do_mounts.c 	struct hd_struct *part = dev_to_part(dev);
part              185 init/do_mounts.c 	if (part->info && !strcmp(label, part->info->volname))
part              227 init/do_mounts.c 	int part;
part              295 init/do_mounts.c 	part = simple_strtoul(p, NULL, 10);
part              297 init/do_mounts.c 	res = blk_lookup_devt(s, part);
part              305 init/do_mounts.c 	res = blk_lookup_devt(s, part);
part              458 kernel/trace/blktrace.c 	struct hd_struct *part = NULL;
part              461 kernel/trace/blktrace.c 		part = bdev->bd_part;
part              463 kernel/trace/blktrace.c 	if (part) {
part              464 kernel/trace/blktrace.c 		bt->start_lba = part->start_sect;
part              465 kernel/trace/blktrace.c 		bt->end_lba = part->start_sect + part->nr_sects;
part              142 net/rds/transport.c 	unsigned int part;
part              153 net/rds/transport.c 		part = trans->stats_info_copy(iter, avail);
part              154 net/rds/transport.c 		avail -= min(avail, part);
part              155 net/rds/transport.c 		total += part;
part             1392 sound/soc/codecs/cs47l15.c 	cs47l15->core.adsp[0].part = "cs47l15";
part             1236 sound/soc/codecs/cs47l24.c 		cs47l24->core.adsp[i].part = "cs47l24";
part             1685 sound/soc/codecs/cs47l35.c 		cs47l35->core.adsp[i].part = "cs47l35";
part             2633 sound/soc/codecs/cs47l85.c 		cs47l85->core.adsp[i].part = "cs47l85";
part             2536 sound/soc/codecs/cs47l90.c 		cs47l90->core.adsp[i].part = "cs47l90";
part             1944 sound/soc/codecs/cs47l92.c 	cs47l92->core.adsp[0].part = "cs47l92";
part             2214 sound/soc/codecs/wm2200.c 		wm2200->dsp[i].part = "wm2200";
part             2050 sound/soc/codecs/wm5102.c 	wm5102->core.adsp[0].part = "wm5102";
part             2410 sound/soc/codecs/wm5110.c 		wm5110->core.adsp[i].part = "wm5110";
part             1819 sound/soc/codecs/wm_adsp.c 	snprintf(file, PAGE_SIZE, "%s-%s-%s.wmfw", dsp->part, dsp->fwf_name,
part             2473 sound/soc/codecs/wm_adsp.c 	snprintf(file, PAGE_SIZE, "%s-%s-%s.bin", dsp->part, dsp->fwf_name,
part             3816 sound/soc/codecs/wm_adsp.c 	buf->name = kasprintf(GFP_KERNEL, "%s-dsp-%s", ctl->dsp->part,
part               57 sound/soc/codecs/wm_adsp.h 	const char *part;
part             1037 sound/soc/intel/atom/sst-atom-controls.c 	cmd.param.part.cfg.rate = 2; /* 48khz */
part             1039 sound/soc/intel/atom/sst-atom-controls.c 	cmd.param.part.cfg.format = ids->format; /* stereo/Mono */
part             1040 sound/soc/intel/atom/sst-atom-controls.c 	cmd.param.part.cfg.s_length = 1; /* 24bit left justified */
part              425 sound/soc/intel/atom/sst-atom-controls.h 	} part;
part              200 sound/soc/intel/atom/sst-mfld-dsp.h 	} part;
part              223 sound/soc/intel/atom/sst-mfld-dsp.h 	} part;
part               58 sound/soc/intel/atom/sst/sst.c 	if (isr.part.done_interrupt) {
part               63 sound/soc/intel/atom/sst/sst.c 		header.p.header_high.part.done = 0;
part               67 sound/soc/intel/atom/sst/sst.c 		isr.part.done_interrupt = 1;
part               76 sound/soc/intel/atom/sst/sst.c 	if (isr.part.busy_interrupt) {
part               80 sound/soc/intel/atom/sst/sst.c 		imr.part.busy_interrupt = 1;
part               85 sound/soc/intel/atom/sst/sst.c 		if (sst_create_ipc_msg(&msg, header.p.header_high.part.large)) {
part               90 sound/soc/intel/atom/sst/sst.c 		if (header.p.header_high.part.large) {
part              104 sound/soc/intel/atom/sst/sst.c 			sst_is_process_reply(header.p.header_high.part.msg_id);
part               76 sound/soc/intel/atom/sst/sst.h 	} part;
part               87 sound/soc/intel/atom/sst/sst.h 	} part;
part              103 sound/soc/intel/atom/sst/sst.h 	} part;
part              123 sound/soc/intel/atom/sst/sst.h 	} part;
part              132 sound/soc/intel/atom/sst/sst.h 	} part;
part              141 sound/soc/intel/atom/sst/sst.h 	} part;
part              129 sound/soc/intel/atom/sst/sst_ipc.c 		while (header.p.header_high.part.busy) {
part              149 sound/soc/intel/atom/sst/sst_ipc.c 		if (header.p.header_high.part.busy) {
part              165 sound/soc/intel/atom/sst/sst_ipc.c 	if (msg->mrfld_header.p.header_high.part.large)
part              191 sound/soc/intel/atom/sst/sst_ipc.c 	isr.part.busy_interrupt = 1;
part              197 sound/soc/intel/atom/sst/sst_ipc.c 	clear_ipc.p.header_high.part.busy = 0;
part              198 sound/soc/intel/atom/sst/sst_ipc.c 	clear_ipc.p.header_high.part.done = 1;
part              202 sound/soc/intel/atom/sst/sst_ipc.c 	imr.part.busy_interrupt = 0;
part              335 sound/soc/intel/atom/sst/sst_ipc.c 	drv_id = msg_high.part.drv_id;
part              345 sound/soc/intel/atom/sst/sst_ipc.c 	if (msg_high.part.result && drv_id && !msg_high.part.large) {
part              348 sound/soc/intel/atom/sst/sst_ipc.c 		sst_wake_up_block(sst_drv_ctx, msg_high.part.result,
part              349 sound/soc/intel/atom/sst/sst_ipc.c 			msg_high.part.drv_id,
part              350 sound/soc/intel/atom/sst/sst_ipc.c 			msg_high.part.msg_id, NULL, 0);
part              359 sound/soc/intel/atom/sst/sst_ipc.c 	if (msg_high.part.large) {
part              366 sound/soc/intel/atom/sst/sst_ipc.c 		if (sst_wake_up_block(sst_drv_ctx, msg_high.part.result,
part              367 sound/soc/intel/atom/sst/sst_ipc.c 				msg_high.part.drv_id,
part              368 sound/soc/intel/atom/sst/sst_ipc.c 				msg_high.part.msg_id, data, msg_low))
part              371 sound/soc/intel/atom/sst/sst_ipc.c 		sst_wake_up_block(sst_drv_ctx, msg_high.part.result,
part              372 sound/soc/intel/atom/sst/sst_ipc.c 				msg_high.part.drv_id,
part              373 sound/soc/intel/atom/sst/sst_ipc.c 				msg_high.part.msg_id, NULL, 0);
part               97 sound/soc/intel/atom/sst/sst_loader.c 	csr.part.xt_snoop = 1;
part              251 sound/soc/intel/atom/sst/sst_pvt.c 	msg->mrfld_header.p.header_high.part.res_rqd = !sync;
part              255 sound/soc/intel/atom/sst/sst_pvt.c 			msg->mrfld_header.p.header_high.part.res_rqd);
part              307 sound/soc/intel/atom/sst/sst_pvt.c 	header->p.header_high.part.msg_id = msg;
part              308 sound/soc/intel/atom/sst/sst_pvt.c 	header->p.header_high.part.task_id = task_id;
part              309 sound/soc/intel/atom/sst/sst_pvt.c 	header->p.header_high.part.large = large;
part              310 sound/soc/intel/atom/sst/sst_pvt.c 	header->p.header_high.part.drv_id = drv_id;
part              311 sound/soc/intel/atom/sst/sst_pvt.c 	header->p.header_high.part.done = 0;
part              312 sound/soc/intel/atom/sst/sst_pvt.c 	header->p.header_high.part.busy = 1;
part              313 sound/soc/intel/atom/sst/sst_pvt.c 	header->p.header_high.part.res_rqd = 1;
part              189 sound/soc/intel/atom/sst/sst_stream.c 	msg->mrfld_header.p.header_high.part.res_rqd = bytes->block;
part              190 tools/usb/usbip/libsrc/vhci_driver.c 	int part;
part              221 tools/usb/usbip/libsrc/vhci_driver.c 	for (part = 0; part < 3; part++) {
part              222 tools/usb/usbip/libsrc/vhci_driver.c 		end = strchr(start, delim[part]);
part              223 tools/usb/usbip/libsrc/vhci_driver.c 		if (end == NULL || (end - start) > max_len[part]) {