rh                 73 arch/arm/vfp/vfp.h 	u64 rh, rma, rmb, rl;
rh                 86 arch/arm/vfp/vfp.h 	rh = (u64)nh * mh;
rh                 87 arch/arm/vfp/vfp.h 	rh += ((u64)(rma < rmb) << 32) + (rma >> 32);
rh                 91 arch/arm/vfp/vfp.h 	rh += (rl < rma);
rh                 94 arch/arm/vfp/vfp.h 	*resh = rh;
rh                105 arch/arm/vfp/vfp.h 	u64 rh, rl;
rh                106 arch/arm/vfp/vfp.h 	mul64to128(&rh, &rl, n, m);
rh                107 arch/arm/vfp/vfp.h 	return rh | (rl != 0);
rh                464 arch/ia64/kernel/mca.c 	sal_log_record_header_t *rh;
rh                483 arch/ia64/kernel/mca.c 	rh = (sal_log_record_header_t *)buffer;
rh                484 arch/ia64/kernel/mca.c 	if (rh->severity == sal_log_severity_corrected)
rh               1335 arch/ia64/kernel/mca.c 		sal_log_record_header_t *rh = IA64_LOG_CURR_BUFFER(SAL_INFO_TYPE_MCA);
rh               1336 arch/ia64/kernel/mca.c 		rh->severity = sal_log_severity_corrected;
rh                386 arch/ia64/kernel/salinfo.c 	sal_log_record_header_t *rh;
rh                388 arch/ia64/kernel/salinfo.c 	rh = (sal_log_record_header_t *)(data->log_buffer);
rh                390 arch/ia64/kernel/salinfo.c 	if (rh->severity == sal_log_severity_corrected)
rh                408 arch/ia64/kernel/salinfo.c 			sal_log_record_header_t *rh = (sal_log_record_header_t *)(data_saved->buffer);
rh                410 arch/ia64/kernel/salinfo.c 			memcpy(data->log_buffer, rh, data->log_size);
rh                412 arch/ia64/kernel/salinfo.c 			if (rh->id == data_saved->id) {
rh                465 arch/ia64/kernel/salinfo.c 	sal_log_record_header_t *rh;
rh                479 arch/ia64/kernel/salinfo.c 	rh = (sal_log_record_header_t *)(data->log_buffer);
rh                481 arch/ia64/kernel/salinfo.c 	if (rh->severity != sal_log_severity_corrected)
rh                 27 arch/powerpc/include/asm/fsl_85xx_cache_sram.h 	rh_info_t *rh;
rh                 47 arch/powerpc/sysdev/fsl_85xx_cache_sram.c 	offset = rh_alloc_align(cache_sram->rh, size, align, NULL);
rh                 65 arch/powerpc/sysdev/fsl_85xx_cache_sram.c 	rh_free(cache_sram->rh, ptr - cache_sram->base_virt);
rh                106 arch/powerpc/sysdev/fsl_85xx_cache_sram.c 	cache_sram->rh = rh_create(sizeof(unsigned int));
rh                107 arch/powerpc/sysdev/fsl_85xx_cache_sram.c 	if (IS_ERR(cache_sram->rh)) {
rh                110 arch/powerpc/sysdev/fsl_85xx_cache_sram.c 		ret = PTR_ERR(cache_sram->rh);
rh                114 arch/powerpc/sysdev/fsl_85xx_cache_sram.c 	rh_attach_region(cache_sram->rh, 0, cache_sram->size);
rh                137 arch/powerpc/sysdev/fsl_85xx_cache_sram.c 	rh_detach_region(cache_sram->rh, 0, cache_sram->size);
rh                138 arch/powerpc/sysdev/fsl_85xx_cache_sram.c 	rh_destroy(cache_sram->rh);
rh                103 crypto/vmac.c  #define ADD128(rh, rl, ih, il)						\
rh                108 crypto/vmac.c  			(rh)++;						\
rh                109 crypto/vmac.c  		(rh) += (ih);						\
rh                114 crypto/vmac.c  #define PMUL64(rh, rl, i1, i2)	/* Assumes m doesn't overflow */	\
rh                118 crypto/vmac.c  		rh = MUL32(_i1>>32, _i2>>32);				\
rh                120 crypto/vmac.c  		ADD128(rh, rl, (m >> 32), (m << 32));			\
rh                123 crypto/vmac.c  #define MUL64(rh, rl, i1, i2)						\
rh                128 crypto/vmac.c  		rh = MUL32(_i1>>32, _i2>>32);				\
rh                130 crypto/vmac.c  		ADD128(rh, rl, (m1 >> 32), (m1 << 32));			\
rh                131 crypto/vmac.c  		ADD128(rh, rl, (m2 >> 32), (m2 << 32));			\
rh                148 crypto/vmac.c  #define nh_16(mp, kp, nw, rh, rl)					\
rh                151 crypto/vmac.c  		rh = rl = 0;						\
rh                155 crypto/vmac.c  			ADD128(rh, rl, th, tl);				\
rh                159 crypto/vmac.c  #define nh_16_2(mp, kp, nw, rh, rl, rh1, rl1)				\
rh                162 crypto/vmac.c  		rh1 = rl1 = rh = rl = 0;				\
rh                166 crypto/vmac.c  			ADD128(rh, rl, th, tl);				\
rh                174 crypto/vmac.c  #define nh_vmac_nhbytes(mp, kp, nw, rh, rl)				\
rh                177 crypto/vmac.c  		rh = rl = 0;						\
rh                181 crypto/vmac.c  			ADD128(rh, rl, th, tl);				\
rh                184 crypto/vmac.c  			ADD128(rh, rl, th, tl);				\
rh                187 crypto/vmac.c  			ADD128(rh, rl, th, tl);				\
rh                190 crypto/vmac.c  			ADD128(rh, rl, th, tl);				\
rh                194 crypto/vmac.c  #define nh_vmac_nhbytes_2(mp, kp, nw, rh, rl, rh1, rl1)			\
rh                197 crypto/vmac.c  		rh1 = rl1 = rh = rl = 0;				\
rh                201 crypto/vmac.c  			ADD128(rh, rl, th, tl);				\
rh                207 crypto/vmac.c  			ADD128(rh, rl, th, tl);				\
rh                213 crypto/vmac.c  			ADD128(rh, rl, th, tl);				\
rh                219 crypto/vmac.c  			ADD128(rh, rl, th, tl);				\
rh                253 crypto/vmac.c  #define nh_16(mp, kp, nw, rh, rl)					\
rh                257 crypto/vmac.c  		rh = rl = t = 0;					\
rh                263 crypto/vmac.c  			ADD128(rh, rl, MUL32(t1 >> 32, t2 >> 32),	\
rh                265 crypto/vmac.c  			rh += (u64)(u32)(m1 >> 32)			\
rh                269 crypto/vmac.c  		ADD128(rh, rl, (t >> 32), (t << 32));			\
rh                340 crypto/vmac.c  #define nh_16_2(mp, kp, nw, rh, rl, rh2, rl2)				\
rh                342 crypto/vmac.c  		nh_16(mp, kp, nw, rh, rl);				\
rh                347 crypto/vmac.c  #define nh_vmac_nhbytes(mp, kp, nw, rh, rl)				\
rh                348 crypto/vmac.c  	nh_16(mp, kp, nw, rh, rl)
rh                351 crypto/vmac.c  #define nh_vmac_nhbytes_2(mp, kp, nw, rh, rl, rh2, rl2)			\
rh                353 crypto/vmac.c  		nh_vmac_nhbytes(mp, kp, nw, rh, rl);			\
rh                360 crypto/vmac.c  	u64 rh, rl, t, z = 0;
rh                385 crypto/vmac.c  	MUL64(rh, rl, p1, p2);
rh                386 crypto/vmac.c  	t = rh >> 56;
rh                387 crypto/vmac.c  	ADD128(t, rl, z, rh);
rh                388 crypto/vmac.c  	rh <<= 8;
rh                389 crypto/vmac.c  	ADD128(t, rl, z, rh);
rh                407 crypto/vmac.c  	u64 rh, rl;
rh                411 crypto/vmac.c  		nh_vmac_nhbytes(mptr, kptr, VMAC_NHBYTES/8, rh, rl);
rh                412 crypto/vmac.c  		rh &= m62;
rh                413 crypto/vmac.c  		ADD128(ch, cl, rh, rl);
rh                419 crypto/vmac.c  		nh_vmac_nhbytes(mptr, kptr, VMAC_NHBYTES/8, rh, rl);
rh                420 crypto/vmac.c  		rh &= m62;
rh                421 crypto/vmac.c  		poly_step(ch, cl, pkh, pkl, rh, rl);
rh                547 crypto/vmac.c  		u64 rh, rl;
rh                550 crypto/vmac.c  		nh_16(dctx->partial_words, tctx->nhkey, n / 8, rh, rl);
rh                551 crypto/vmac.c  		rh &= m62;
rh                554 crypto/vmac.c  				  rh, rl);
rh                556 crypto/vmac.c  			ADD128(ch, cl, rh, rl);
rh                489 drivers/crypto/cavium/nitrox/nitrox_csr.h 		u64 rh : 1;
rh                495 drivers/crypto/cavium/nitrox/nitrox_csr.h 		u64 rh : 1;
rh                205 drivers/crypto/cavium/nitrox/nitrox_hal.c 	pkt_slc_ctl.s.rh = 1;
rh                 97 drivers/dma/bestcomm/sram.c 	bcom_sram->rh = rh_create(4);
rh                110 drivers/dma/bestcomm/sram.c 		rh_attach_region(bcom_sram->rh, 0, bcom_sram->size);
rh                115 drivers/dma/bestcomm/sram.c 			rh_attach_region(bcom_sram->rh, zbase - bcom_sram->base_phys, regaddr_p[1]);
rh                140 drivers/dma/bestcomm/sram.c 		rh_destroy(bcom_sram->rh);
rh                154 drivers/dma/bestcomm/sram.c 	offset = rh_alloc_align(bcom_sram->rh, size, align, NULL);
rh                175 drivers/dma/bestcomm/sram.c 	rh_free(bcom_sram->rh, offset);
rh                 65 drivers/md/dm-raid1.c 	struct dm_region_hash *rh;
rh                336 drivers/md/dm-raid1.c 	sector_t region_size = dm_rh_get_region_size(ms->rh);
rh                341 drivers/md/dm-raid1.c 	from.sector = m->offset + dm_rh_region_to_sector(ms->rh, key);
rh                360 drivers/md/dm-raid1.c 		dest->sector = m->offset + dm_rh_region_to_sector(ms->rh, key);
rh                387 drivers/md/dm-raid1.c 	struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
rh                392 drivers/md/dm-raid1.c 	dm_rh_recovery_prepare(ms->rh);
rh                397 drivers/md/dm-raid1.c 	while ((reg = dm_rh_recovery_start(ms->rh)))
rh                439 drivers/md/dm-raid1.c 	struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
rh                440 drivers/md/dm-raid1.c 	region_t region = dm_rh_bio_to_region(ms->rh, bio);
rh                555 drivers/md/dm-raid1.c 	int state = dm_rh_get_state(ms->rh, region, may_block);
rh                566 drivers/md/dm-raid1.c 		region = dm_rh_bio_to_region(ms->rh, bio);
rh                685 drivers/md/dm-raid1.c 	struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
rh                706 drivers/md/dm-raid1.c 		region = dm_rh_bio_to_region(ms->rh, bio);
rh                714 drivers/md/dm-raid1.c 		state = dm_rh_get_state(ms->rh, region, 1);
rh                749 drivers/md/dm-raid1.c 	dm_rh_inc_pending(ms->rh, &sync);
rh                750 drivers/md/dm-raid1.c 	dm_rh_inc_pending(ms->rh, &nosync);
rh                757 drivers/md/dm-raid1.c 	ms->log_failure = dm_rh_flush(ms->rh) ? 1 : ms->log_failure;
rh                772 drivers/md/dm-raid1.c 		dm_rh_delay(ms->rh, bio);
rh                814 drivers/md/dm-raid1.c 			dm_rh_mark_nosync(ms->rh, bio);
rh                866 drivers/md/dm-raid1.c 	dm_rh_update_states(ms->rh, errors_handled(ms));
rh                911 drivers/md/dm-raid1.c 	ms->rh = dm_region_hash_create(ms, dispatch_bios, wakeup_mirrord,
rh                915 drivers/md/dm-raid1.c 	if (IS_ERR(ms->rh)) {
rh                932 drivers/md/dm-raid1.c 	dm_region_hash_destroy(ms->rh);
rh               1112 drivers/md/dm-raid1.c 	r = dm_set_target_max_io_len(ti, dm_rh_get_region_size(ms->rh));
rh               1189 drivers/md/dm-raid1.c 	struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
rh               1197 drivers/md/dm-raid1.c 		bio_record->write_region = dm_rh_bio_to_region(ms->rh, bio);
rh               1202 drivers/md/dm-raid1.c 	r = log->type->in_sync(log, dm_rh_bio_to_region(ms->rh, bio), 0);
rh               1249 drivers/md/dm-raid1.c 			dm_rh_dec(ms->rh, bio_record->write_region);
rh               1303 drivers/md/dm-raid1.c 	struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
rh               1328 drivers/md/dm-raid1.c 	dm_rh_stop_recovery(ms->rh);
rh               1331 drivers/md/dm-raid1.c 		   !dm_rh_recovery_in_flight(ms->rh));
rh               1349 drivers/md/dm-raid1.c 	struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
rh               1359 drivers/md/dm-raid1.c 	struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
rh               1365 drivers/md/dm-raid1.c 	dm_rh_start_recovery(ms->rh);
rh               1399 drivers/md/dm-raid1.c 	struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
rh                103 drivers/md/dm-region-hash.c 	struct dm_region_hash *rh;	/* FIXME: can we get rid of this ? */
rh                117 drivers/md/dm-region-hash.c static region_t dm_rh_sector_to_region(struct dm_region_hash *rh, sector_t sector)
rh                119 drivers/md/dm-region-hash.c 	return sector >> rh->region_shift;
rh                122 drivers/md/dm-region-hash.c sector_t dm_rh_region_to_sector(struct dm_region_hash *rh, region_t region)
rh                124 drivers/md/dm-region-hash.c 	return region << rh->region_shift;
rh                128 drivers/md/dm-region-hash.c region_t dm_rh_bio_to_region(struct dm_region_hash *rh, struct bio *bio)
rh                130 drivers/md/dm-region-hash.c 	return dm_rh_sector_to_region(rh, bio->bi_iter.bi_sector -
rh                131 drivers/md/dm-region-hash.c 				      rh->target_begin);
rh                137 drivers/md/dm-region-hash.c 	return reg->rh->context;
rh                147 drivers/md/dm-region-hash.c sector_t dm_rh_get_region_size(struct dm_region_hash *rh)
rh                149 drivers/md/dm-region-hash.c 	return rh->region_size;
rh                170 drivers/md/dm-region-hash.c 	struct dm_region_hash *rh;
rh                184 drivers/md/dm-region-hash.c 	rh = kzalloc(sizeof(*rh), GFP_KERNEL);
rh                185 drivers/md/dm-region-hash.c 	if (!rh) {
rh                190 drivers/md/dm-region-hash.c 	rh->context = context;
rh                191 drivers/md/dm-region-hash.c 	rh->dispatch_bios = dispatch_bios;
rh                192 drivers/md/dm-region-hash.c 	rh->wakeup_workers = wakeup_workers;
rh                193 drivers/md/dm-region-hash.c 	rh->wakeup_all_recovery_waiters = wakeup_all_recovery_waiters;
rh                194 drivers/md/dm-region-hash.c 	rh->target_begin = target_begin;
rh                195 drivers/md/dm-region-hash.c 	rh->max_recovery = max_recovery;
rh                196 drivers/md/dm-region-hash.c 	rh->log = log;
rh                197 drivers/md/dm-region-hash.c 	rh->region_size = region_size;
rh                198 drivers/md/dm-region-hash.c 	rh->region_shift = __ffs(region_size);
rh                199 drivers/md/dm-region-hash.c 	rwlock_init(&rh->hash_lock);
rh                200 drivers/md/dm-region-hash.c 	rh->mask = nr_buckets - 1;
rh                201 drivers/md/dm-region-hash.c 	rh->nr_buckets = nr_buckets;
rh                203 drivers/md/dm-region-hash.c 	rh->shift = RH_HASH_SHIFT;
rh                204 drivers/md/dm-region-hash.c 	rh->prime = RH_HASH_MULT;
rh                206 drivers/md/dm-region-hash.c 	rh->buckets = vmalloc(array_size(nr_buckets, sizeof(*rh->buckets)));
rh                207 drivers/md/dm-region-hash.c 	if (!rh->buckets) {
rh                209 drivers/md/dm-region-hash.c 		kfree(rh);
rh                214 drivers/md/dm-region-hash.c 		INIT_LIST_HEAD(rh->buckets + i);
rh                216 drivers/md/dm-region-hash.c 	spin_lock_init(&rh->region_lock);
rh                217 drivers/md/dm-region-hash.c 	sema_init(&rh->recovery_count, 0);
rh                218 drivers/md/dm-region-hash.c 	atomic_set(&rh->recovery_in_flight, 0);
rh                219 drivers/md/dm-region-hash.c 	INIT_LIST_HEAD(&rh->clean_regions);
rh                220 drivers/md/dm-region-hash.c 	INIT_LIST_HEAD(&rh->quiesced_regions);
rh                221 drivers/md/dm-region-hash.c 	INIT_LIST_HEAD(&rh->recovered_regions);
rh                222 drivers/md/dm-region-hash.c 	INIT_LIST_HEAD(&rh->failed_recovered_regions);
rh                223 drivers/md/dm-region-hash.c 	rh->flush_failure = 0;
rh                225 drivers/md/dm-region-hash.c 	ret = mempool_init_kmalloc_pool(&rh->region_pool, MIN_REGIONS,
rh                228 drivers/md/dm-region-hash.c 		vfree(rh->buckets);
rh                229 drivers/md/dm-region-hash.c 		kfree(rh);
rh                230 drivers/md/dm-region-hash.c 		rh = ERR_PTR(-ENOMEM);
rh                233 drivers/md/dm-region-hash.c 	return rh;
rh                237 drivers/md/dm-region-hash.c void dm_region_hash_destroy(struct dm_region_hash *rh)
rh                242 drivers/md/dm-region-hash.c 	BUG_ON(!list_empty(&rh->quiesced_regions));
rh                243 drivers/md/dm-region-hash.c 	for (h = 0; h < rh->nr_buckets; h++) {
rh                244 drivers/md/dm-region-hash.c 		list_for_each_entry_safe(reg, nreg, rh->buckets + h,
rh                247 drivers/md/dm-region-hash.c 			mempool_free(reg, &rh->region_pool);
rh                251 drivers/md/dm-region-hash.c 	if (rh->log)
rh                252 drivers/md/dm-region-hash.c 		dm_dirty_log_destroy(rh->log);
rh                254 drivers/md/dm-region-hash.c 	mempool_exit(&rh->region_pool);
rh                255 drivers/md/dm-region-hash.c 	vfree(rh->buckets);
rh                256 drivers/md/dm-region-hash.c 	kfree(rh);
rh                260 drivers/md/dm-region-hash.c struct dm_dirty_log *dm_rh_dirty_log(struct dm_region_hash *rh)
rh                262 drivers/md/dm-region-hash.c 	return rh->log;
rh                266 drivers/md/dm-region-hash.c static unsigned rh_hash(struct dm_region_hash *rh, region_t region)
rh                268 drivers/md/dm-region-hash.c 	return (unsigned) ((region * rh->prime) >> rh->shift) & rh->mask;
rh                271 drivers/md/dm-region-hash.c static struct dm_region *__rh_lookup(struct dm_region_hash *rh, region_t region)
rh                274 drivers/md/dm-region-hash.c 	struct list_head *bucket = rh->buckets + rh_hash(rh, region);
rh                283 drivers/md/dm-region-hash.c static void __rh_insert(struct dm_region_hash *rh, struct dm_region *reg)
rh                285 drivers/md/dm-region-hash.c 	list_add(&reg->hash_list, rh->buckets + rh_hash(rh, reg->key));
rh                288 drivers/md/dm-region-hash.c static struct dm_region *__rh_alloc(struct dm_region_hash *rh, region_t region)
rh                292 drivers/md/dm-region-hash.c 	nreg = mempool_alloc(&rh->region_pool, GFP_ATOMIC);
rh                296 drivers/md/dm-region-hash.c 	nreg->state = rh->log->type->in_sync(rh->log, region, 1) ?
rh                298 drivers/md/dm-region-hash.c 	nreg->rh = rh;
rh                304 drivers/md/dm-region-hash.c 	write_lock_irq(&rh->hash_lock);
rh                305 drivers/md/dm-region-hash.c 	reg = __rh_lookup(rh, region);
rh                308 drivers/md/dm-region-hash.c 		mempool_free(nreg, &rh->region_pool);
rh                310 drivers/md/dm-region-hash.c 		__rh_insert(rh, nreg);
rh                312 drivers/md/dm-region-hash.c 			spin_lock(&rh->region_lock);
rh                313 drivers/md/dm-region-hash.c 			list_add(&nreg->list, &rh->clean_regions);
rh                314 drivers/md/dm-region-hash.c 			spin_unlock(&rh->region_lock);
rh                319 drivers/md/dm-region-hash.c 	write_unlock_irq(&rh->hash_lock);
rh                324 drivers/md/dm-region-hash.c static struct dm_region *__rh_find(struct dm_region_hash *rh, region_t region)
rh                328 drivers/md/dm-region-hash.c 	reg = __rh_lookup(rh, region);
rh                330 drivers/md/dm-region-hash.c 		read_unlock(&rh->hash_lock);
rh                331 drivers/md/dm-region-hash.c 		reg = __rh_alloc(rh, region);
rh                332 drivers/md/dm-region-hash.c 		read_lock(&rh->hash_lock);
rh                338 drivers/md/dm-region-hash.c int dm_rh_get_state(struct dm_region_hash *rh, region_t region, int may_block)
rh                343 drivers/md/dm-region-hash.c 	read_lock(&rh->hash_lock);
rh                344 drivers/md/dm-region-hash.c 	reg = __rh_lookup(rh, region);
rh                345 drivers/md/dm-region-hash.c 	read_unlock(&rh->hash_lock);
rh                354 drivers/md/dm-region-hash.c 	r = rh->log->type->in_sync(rh->log, region, may_block);
rh                366 drivers/md/dm-region-hash.c 	struct dm_region_hash *rh = reg->rh;
rh                368 drivers/md/dm-region-hash.c 	rh->log->type->set_region_sync(rh->log, reg->key, success);
rh                379 drivers/md/dm-region-hash.c 	rh->dispatch_bios(rh->context, &reg->delayed_bios);
rh                380 drivers/md/dm-region-hash.c 	if (atomic_dec_and_test(&rh->recovery_in_flight))
rh                381 drivers/md/dm-region-hash.c 		rh->wakeup_all_recovery_waiters(rh->context);
rh                382 drivers/md/dm-region-hash.c 	up(&rh->recovery_count);
rh                395 drivers/md/dm-region-hash.c void dm_rh_mark_nosync(struct dm_region_hash *rh, struct bio *bio)
rh                398 drivers/md/dm-region-hash.c 	struct dm_dirty_log *log = rh->log;
rh                400 drivers/md/dm-region-hash.c 	region_t region = dm_rh_bio_to_region(rh, bio);
rh                404 drivers/md/dm-region-hash.c 		rh->flush_failure = 1;
rh                414 drivers/md/dm-region-hash.c 	read_lock(&rh->hash_lock);
rh                415 drivers/md/dm-region-hash.c 	reg = __rh_find(rh, region);
rh                416 drivers/md/dm-region-hash.c 	read_unlock(&rh->hash_lock);
rh                422 drivers/md/dm-region-hash.c 	spin_lock_irqsave(&rh->region_lock, flags);
rh                433 drivers/md/dm-region-hash.c 	spin_unlock_irqrestore(&rh->region_lock, flags);
rh                440 drivers/md/dm-region-hash.c void dm_rh_update_states(struct dm_region_hash *rh, int errors_handled)
rh                451 drivers/md/dm-region-hash.c 	write_lock_irq(&rh->hash_lock);
rh                452 drivers/md/dm-region-hash.c 	spin_lock(&rh->region_lock);
rh                453 drivers/md/dm-region-hash.c 	if (!list_empty(&rh->clean_regions)) {
rh                454 drivers/md/dm-region-hash.c 		list_splice_init(&rh->clean_regions, &clean);
rh                460 drivers/md/dm-region-hash.c 	if (!list_empty(&rh->recovered_regions)) {
rh                461 drivers/md/dm-region-hash.c 		list_splice_init(&rh->recovered_regions, &recovered);
rh                467 drivers/md/dm-region-hash.c 	if (!list_empty(&rh->failed_recovered_regions)) {
rh                468 drivers/md/dm-region-hash.c 		list_splice_init(&rh->failed_recovered_regions,
rh                475 drivers/md/dm-region-hash.c 	spin_unlock(&rh->region_lock);
rh                476 drivers/md/dm-region-hash.c 	write_unlock_irq(&rh->hash_lock);
rh                484 drivers/md/dm-region-hash.c 		rh->log->type->clear_region(rh->log, reg->key);
rh                486 drivers/md/dm-region-hash.c 		mempool_free(reg, &rh->region_pool);
rh                491 drivers/md/dm-region-hash.c 		mempool_free(reg, &rh->region_pool);
rh                495 drivers/md/dm-region-hash.c 		rh->log->type->clear_region(rh->log, reg->key);
rh                496 drivers/md/dm-region-hash.c 		mempool_free(reg, &rh->region_pool);
rh                499 drivers/md/dm-region-hash.c 	rh->log->type->flush(rh->log);
rh                503 drivers/md/dm-region-hash.c static void rh_inc(struct dm_region_hash *rh, region_t region)
rh                507 drivers/md/dm-region-hash.c 	read_lock(&rh->hash_lock);
rh                508 drivers/md/dm-region-hash.c 	reg = __rh_find(rh, region);
rh                510 drivers/md/dm-region-hash.c 	spin_lock_irq(&rh->region_lock);
rh                516 drivers/md/dm-region-hash.c 		spin_unlock_irq(&rh->region_lock);
rh                518 drivers/md/dm-region-hash.c 		rh->log->type->mark_region(rh->log, reg->key);
rh                520 drivers/md/dm-region-hash.c 		spin_unlock_irq(&rh->region_lock);
rh                523 drivers/md/dm-region-hash.c 	read_unlock(&rh->hash_lock);
rh                526 drivers/md/dm-region-hash.c void dm_rh_inc_pending(struct dm_region_hash *rh, struct bio_list *bios)
rh                533 drivers/md/dm-region-hash.c 		rh_inc(rh, dm_rh_bio_to_region(rh, bio));
rh                538 drivers/md/dm-region-hash.c void dm_rh_dec(struct dm_region_hash *rh, region_t region)
rh                544 drivers/md/dm-region-hash.c 	read_lock(&rh->hash_lock);
rh                545 drivers/md/dm-region-hash.c 	reg = __rh_lookup(rh, region);
rh                546 drivers/md/dm-region-hash.c 	read_unlock(&rh->hash_lock);
rh                548 drivers/md/dm-region-hash.c 	spin_lock_irqsave(&rh->region_lock, flags);
rh                562 drivers/md/dm-region-hash.c 		if (unlikely(rh->flush_failure)) {
rh                570 drivers/md/dm-region-hash.c 			list_add_tail(&reg->list, &rh->quiesced_regions);
rh                573 drivers/md/dm-region-hash.c 			list_add(&reg->list, &rh->clean_regions);
rh                577 drivers/md/dm-region-hash.c 	spin_unlock_irqrestore(&rh->region_lock, flags);
rh                580 drivers/md/dm-region-hash.c 		rh->wakeup_workers(rh->context);
rh                587 drivers/md/dm-region-hash.c static int __rh_recovery_prepare(struct dm_region_hash *rh)
rh                596 drivers/md/dm-region-hash.c 	r = rh->log->type->get_resync_work(rh->log, &region);
rh                604 drivers/md/dm-region-hash.c 	read_lock(&rh->hash_lock);
rh                605 drivers/md/dm-region-hash.c 	reg = __rh_find(rh, region);
rh                606 drivers/md/dm-region-hash.c 	read_unlock(&rh->hash_lock);
rh                608 drivers/md/dm-region-hash.c 	spin_lock_irq(&rh->region_lock);
rh                615 drivers/md/dm-region-hash.c 		list_move(&reg->list, &rh->quiesced_regions);
rh                617 drivers/md/dm-region-hash.c 	spin_unlock_irq(&rh->region_lock);
rh                622 drivers/md/dm-region-hash.c void dm_rh_recovery_prepare(struct dm_region_hash *rh)
rh                625 drivers/md/dm-region-hash.c 	atomic_inc(&rh->recovery_in_flight);
rh                627 drivers/md/dm-region-hash.c 	while (!down_trylock(&rh->recovery_count)) {
rh                628 drivers/md/dm-region-hash.c 		atomic_inc(&rh->recovery_in_flight);
rh                629 drivers/md/dm-region-hash.c 		if (__rh_recovery_prepare(rh) <= 0) {
rh                630 drivers/md/dm-region-hash.c 			atomic_dec(&rh->recovery_in_flight);
rh                631 drivers/md/dm-region-hash.c 			up(&rh->recovery_count);
rh                637 drivers/md/dm-region-hash.c 	if (atomic_dec_and_test(&rh->recovery_in_flight))
rh                638 drivers/md/dm-region-hash.c 		rh->wakeup_all_recovery_waiters(rh->context);
rh                645 drivers/md/dm-region-hash.c struct dm_region *dm_rh_recovery_start(struct dm_region_hash *rh)
rh                649 drivers/md/dm-region-hash.c 	spin_lock_irq(&rh->region_lock);
rh                650 drivers/md/dm-region-hash.c 	if (!list_empty(&rh->quiesced_regions)) {
rh                651 drivers/md/dm-region-hash.c 		reg = list_entry(rh->quiesced_regions.next,
rh                655 drivers/md/dm-region-hash.c 	spin_unlock_irq(&rh->region_lock);
rh                663 drivers/md/dm-region-hash.c 	struct dm_region_hash *rh = reg->rh;
rh                665 drivers/md/dm-region-hash.c 	spin_lock_irq(&rh->region_lock);
rh                667 drivers/md/dm-region-hash.c 		list_add(&reg->list, &reg->rh->recovered_regions);
rh                669 drivers/md/dm-region-hash.c 		list_add(&reg->list, &reg->rh->failed_recovered_regions);
rh                671 drivers/md/dm-region-hash.c 	spin_unlock_irq(&rh->region_lock);
rh                673 drivers/md/dm-region-hash.c 	rh->wakeup_workers(rh->context);
rh                678 drivers/md/dm-region-hash.c int dm_rh_recovery_in_flight(struct dm_region_hash *rh)
rh                680 drivers/md/dm-region-hash.c 	return atomic_read(&rh->recovery_in_flight);
rh                684 drivers/md/dm-region-hash.c int dm_rh_flush(struct dm_region_hash *rh)
rh                686 drivers/md/dm-region-hash.c 	return rh->log->type->flush(rh->log);
rh                690 drivers/md/dm-region-hash.c void dm_rh_delay(struct dm_region_hash *rh, struct bio *bio)
rh                694 drivers/md/dm-region-hash.c 	read_lock(&rh->hash_lock);
rh                695 drivers/md/dm-region-hash.c 	reg = __rh_find(rh, dm_rh_bio_to_region(rh, bio));
rh                697 drivers/md/dm-region-hash.c 	read_unlock(&rh->hash_lock);
rh                701 drivers/md/dm-region-hash.c void dm_rh_stop_recovery(struct dm_region_hash *rh)
rh                706 drivers/md/dm-region-hash.c 	for (i = 0; i < rh->max_recovery; i++)
rh                707 drivers/md/dm-region-hash.c 		down(&rh->recovery_count);
rh                711 drivers/md/dm-region-hash.c void dm_rh_start_recovery(struct dm_region_hash *rh)
rh                715 drivers/md/dm-region-hash.c 	for (i = 0; i < rh->max_recovery; i++)
rh                716 drivers/md/dm-region-hash.c 		up(&rh->recovery_count);
rh                718 drivers/md/dm-region-hash.c 	rh->wakeup_workers(rh->context);
rh                156 drivers/media/usb/uvc/uvc_v4l2.c 	u16 rw, rh;
rh                191 drivers/media/usb/uvc/uvc_v4l2.c 	rh = fmt->fmt.pix.height;
rh                198 drivers/media/usb/uvc/uvc_v4l2.c 		d = min(w, rw) * min(h, rh);
rh                199 drivers/media/usb/uvc/uvc_v4l2.c 		d = w*h + rw*rh - 2*d;
rh                662 drivers/net/ethernet/broadcom/b44.c 	struct rx_header *rh;
rh                706 drivers/net/ethernet/broadcom/b44.c 	rh = (struct rx_header *) skb->data;
rh                708 drivers/net/ethernet/broadcom/b44.c 	rh->len = 0;
rh                709 drivers/net/ethernet/broadcom/b44.c 	rh->flags = 0;
rh                737 drivers/net/ethernet/broadcom/b44.c 	struct rx_header *rh;
rh                748 drivers/net/ethernet/broadcom/b44.c 	rh = (struct rx_header *) src_map->skb->data;
rh                749 drivers/net/ethernet/broadcom/b44.c 	rh->len = 0;
rh                750 drivers/net/ethernet/broadcom/b44.c 	rh->flags = 0;
rh                793 drivers/net/ethernet/broadcom/b44.c 		struct rx_header *rh;
rh                799 drivers/net/ethernet/broadcom/b44.c 		rh = (struct rx_header *) skb->data;
rh                800 drivers/net/ethernet/broadcom/b44.c 		len = le16_to_cpu(rh->len);
rh                802 drivers/net/ethernet/broadcom/b44.c 		    (rh->flags & cpu_to_le16(RX_FLAG_ERRORS))) {
rh                816 drivers/net/ethernet/broadcom/b44.c 				len = le16_to_cpu(rh->len);
rh                570 drivers/net/ethernet/cavium/liquidio/lio_core.c 		     union octeon_rh *rh,
rh                625 drivers/net/ethernet/cavium/liquidio/lio_core.c 		r_dh_off = (rh->r_dh.len - 1) * BYTES_PER_DHLEN_UNIT;
rh                628 drivers/net/ethernet/cavium/liquidio/lio_core.c 			if (rh->r_dh.has_hwtstamp) {
rh                649 drivers/net/ethernet/cavium/liquidio/lio_core.c 		if (rh->r_dh.has_hash) {
rh                657 drivers/net/ethernet/cavium/liquidio/lio_core.c 		skb_pull(skb, rh->r_dh.len * BYTES_PER_DHLEN_UNIT);
rh                661 drivers/net/ethernet/cavium/liquidio/lio_core.c 		    (((rh->r_dh.encap_on) &&
rh                662 drivers/net/ethernet/cavium/liquidio/lio_core.c 		      (rh->r_dh.csum_verified & CNNIC_TUN_CSUM_VERIFIED)) ||
rh                663 drivers/net/ethernet/cavium/liquidio/lio_core.c 		     (!(rh->r_dh.encap_on) &&
rh                664 drivers/net/ethernet/cavium/liquidio/lio_core.c 		      ((rh->r_dh.csum_verified & CNNIC_CSUM_VERIFIED) ==
rh                674 drivers/net/ethernet/cavium/liquidio/lio_core.c 		if (rh->r_dh.encap_on) {
rh                682 drivers/net/ethernet/cavium/liquidio/lio_core.c 		    rh->r_dh.vlan) {
rh                683 drivers/net/ethernet/cavium/liquidio/lio_core.c 			u16 priority = rh->r_dh.priority;
rh                684 drivers/net/ethernet/cavium/liquidio/lio_core.c 			u16 vid = rh->r_dh.vlan;
rh                693 drivers/net/ethernet/cavium/liquidio/lio_core.c 			rh->r_dh.len * BYTES_PER_DHLEN_UNIT;
rh                 37 drivers/net/ethernet/cavium/liquidio/lio_ethtool.c 	u64     rh;
rh                 43 drivers/net/ethernet/cavium/liquidio/lio_ethtool.c 	u64 rh;
rh                102 drivers/net/ethernet/cavium/liquidio/lio_main.c 	u64 rh;
rh                108 drivers/net/ethernet/cavium/liquidio/lio_main.c 	u64 rh;
rh               3316 drivers/net/ethernet/cavium/liquidio/lio_main.c 			recv_pkt->rh.r_nic_info.gmxport);
rh               3320 drivers/net/ethernet/cavium/liquidio/lio_main.c 	gmxport = recv_pkt->rh.r_nic_info.gmxport;
rh               4011 drivers/net/ethernet/cavium/liquidio/lio_main.c 	notice = recv_pkt->rh.r.ossp;
rh                 44 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c 	u64 rh;
rh               1892 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c 			recv_pkt->rh.r_nic_info.gmxport);
rh               1896 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c 	gmxport = recv_pkt->rh.r_nic_info.gmxport;
rh                300 drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c 	union octeon_rh *rh;
rh                309 drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c 	rh = &recv_pkt->rh;
rh                310 drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c 	ifidx = rh->r.ossp;
rh                328 drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c 	skb_pull(skb, rh->r_dh.len * BYTES_PER_DHLEN_UNIT);
rh                910 drivers/net/ethernet/cavium/liquidio/liquidio_common.h static inline int opcode_slow_path(union octeon_rh *rh)
rh                914 drivers/net/ethernet/cavium/liquidio/liquidio_common.h 	subcode1 = OPCODE_SUBCODE((rh)->r.opcode, (rh)->r.subcode);
rh               1041 drivers/net/ethernet/cavium/liquidio/liquidio_common.h 	u64 rh;
rh               1212 drivers/net/ethernet/cavium/liquidio/octeon_device.c 		(u32)recv_pkt->rh.r_core_drv_init.app_mode),
rh               1214 drivers/net/ethernet/cavium/liquidio/octeon_device.c 	oct->app_mode = (u32)recv_pkt->rh.r_core_drv_init.app_mode;
rh               1215 drivers/net/ethernet/cavium/liquidio/octeon_device.c 	if (recv_pkt->rh.r_core_drv_init.app_mode == CVM_DRV_NIC_APP) {
rh               1217 drivers/net/ethernet/cavium/liquidio/octeon_device.c 			(u32)recv_pkt->rh.r_core_drv_init.max_nic_ports;
rh               1219 drivers/net/ethernet/cavium/liquidio/octeon_device.c 			(u32)recv_pkt->rh.r_core_drv_init.num_gmx_ports;
rh               1228 drivers/net/ethernet/cavium/liquidio/octeon_device.c 	oct->fw_info.app_cap_flags = recv_pkt->rh.r_core_drv_init.app_cap_flags;
rh               1229 drivers/net/ethernet/cavium/liquidio/octeon_device.c 	oct->fw_info.app_mode = (u32)recv_pkt->rh.r_core_drv_init.app_mode;
rh               1230 drivers/net/ethernet/cavium/liquidio/octeon_device.c 	oct->pfvf_hsword.app_mode = (u32)recv_pkt->rh.r_core_drv_init.app_mode;
rh               1232 drivers/net/ethernet/cavium/liquidio/octeon_device.c 	oct->pfvf_hsword.pkind = recv_pkt->rh.r_core_drv_init.pkind;
rh                354 drivers/net/ethernet/cavium/liquidio/octeon_droq.c 	recv_pkt->rh = info->rh;
rh                533 drivers/net/ethernet/cavium/liquidio/octeon_droq.c 			 union octeon_rh *rh,
rh                542 drivers/net/ethernet/cavium/liquidio/octeon_droq.c 	disp_fn = octeon_get_dispatch(oct, (u16)rh->r.opcode,
rh                543 drivers/net/ethernet/cavium/liquidio/octeon_droq.c 				      (u16)rh->r.subcode);
rh                551 drivers/net/ethernet/cavium/liquidio/octeon_droq.c 			rinfo->recv_pkt->rh = *rh;
rh                559 drivers/net/ethernet/cavium/liquidio/octeon_droq.c 			(unsigned int)rh->r.opcode,
rh                560 drivers/net/ethernet/cavium/liquidio/octeon_droq.c 			(unsigned int)rh->r.subcode);
rh                602 drivers/net/ethernet/cavium/liquidio/octeon_droq.c 	union octeon_rh *rh;
rh                627 drivers/net/ethernet/cavium/liquidio/octeon_droq.c 		rh = &info->rh;
rh                630 drivers/net/ethernet/cavium/liquidio/octeon_droq.c 		rh->r_dh.len += (ROUNDUP8(OCT_DROQ_INFO_SIZE) / sizeof(u64));
rh                632 drivers/net/ethernet/cavium/liquidio/octeon_droq.c 		if (opcode_slow_path(rh)) {
rh                635 drivers/net/ethernet/cavium/liquidio/octeon_droq.c 			buf_cnt = octeon_droq_dispatch_pkt(oct, droq, rh, info);
rh                695 drivers/net/ethernet/cavium/liquidio/octeon_droq.c 						       rh, &droq->napi,
rh                769 drivers/net/ethernet/cavium/liquidio/octeon_droq.c 				(u16)rdisp->rinfo->recv_pkt->rh.r.opcode,
rh                770 drivers/net/ethernet/cavium/liquidio/octeon_droq.c 				(u16)rdisp->rinfo->recv_pkt->rh.r.subcode));
rh                819 drivers/net/ethernet/cavium/liquidio/octeon_droq.c 				(u16)rdisp->rinfo->recv_pkt->rh.r.opcode,
rh                820 drivers/net/ethernet/cavium/liquidio/octeon_droq.c 				(u16)rdisp->rinfo->recv_pkt->rh.r.subcode));
rh                 58 drivers/net/ethernet/cavium/liquidio/octeon_droq.h 	union octeon_rh rh;
rh                148 drivers/net/ethernet/cavium/liquidio/octeon_droq.h 	union octeon_rh rh;
rh                 39 drivers/net/ethernet/cavium/liquidio/octeon_network.h 	u64 rh;
rh                 69 drivers/net/ethernet/cavium/liquidio/octeon_network.h 	u64     rh;
rh                 75 drivers/net/ethernet/cavium/liquidio/octeon_network.h 	u64     rh;
rh                 86 drivers/net/ethernet/cavium/liquidio/octeon_network.h 	u64 rh;
rh                 47 drivers/net/ethernet/cisco/enic/vnic_dev.c 	struct vnic_resource_header __iomem *rh;
rh                 60 drivers/net/ethernet/cisco/enic/vnic_dev.c 	rh  = bar->vaddr;
rh                 62 drivers/net/ethernet/cisco/enic/vnic_dev.c 	if (!rh) {
rh                 68 drivers/net/ethernet/cisco/enic/vnic_dev.c 	if ((ioread32(&rh->magic) != VNIC_RES_MAGIC) ||
rh                 69 drivers/net/ethernet/cisco/enic/vnic_dev.c 		(ioread32(&rh->version) != VNIC_RES_VERSION)) {
rh                 75 drivers/net/ethernet/cisco/enic/vnic_dev.c 				 ioread32(&rh->magic), ioread32(&rh->version));
rh                 83 drivers/net/ethernet/cisco/enic/vnic_dev.c 		r = (struct vnic_resource __iomem *)(rh + 1);
rh                293 drivers/net/ethernet/microchip/enc28j60.c 	int rl, rh;
rh                297 drivers/net/ethernet/microchip/enc28j60.c 	rh = spi_read_op(priv, ENC28J60_READ_CTRL_REG, address + 1);
rh                299 drivers/net/ethernet/microchip/enc28j60.c 	return (rh << 8) | rl;
rh               3408 drivers/net/ethernet/sun/niu.c 	struct rx_pkt_hdr1 *rh;
rh               3470 drivers/net/ethernet/sun/niu.c 	len += sizeof(*rh);
rh               3471 drivers/net/ethernet/sun/niu.c 	len = min_t(int, len, sizeof(*rh) + VLAN_ETH_HLEN);
rh               3474 drivers/net/ethernet/sun/niu.c 	rh = (struct rx_pkt_hdr1 *) skb->data;
rh               3477 drivers/net/ethernet/sun/niu.c 			     ((u32)rh->hashval2_0 << 24 |
rh               3478 drivers/net/ethernet/sun/niu.c 			      (u32)rh->hashval2_1 << 16 |
rh               3479 drivers/net/ethernet/sun/niu.c 			      (u32)rh->hashval1_1 << 8 |
rh               3480 drivers/net/ethernet/sun/niu.c 			      (u32)rh->hashval1_2 << 0),
rh               3482 drivers/net/ethernet/sun/niu.c 	skb_pull(skb, sizeof(*rh));
rh                 90 drivers/scsi/fnic/vnic_dev.c 	struct vnic_resource_header __iomem *rh;
rh                 99 drivers/scsi/fnic/vnic_dev.c 	rh = bar->vaddr;
rh                100 drivers/scsi/fnic/vnic_dev.c 	if (!rh) {
rh                105 drivers/scsi/fnic/vnic_dev.c 	if (ioread32(&rh->magic) != VNIC_RES_MAGIC ||
rh                106 drivers/scsi/fnic/vnic_dev.c 	    ioread32(&rh->version) != VNIC_RES_VERSION) {
rh                110 drivers/scsi/fnic/vnic_dev.c 			ioread32(&rh->magic), ioread32(&rh->version));
rh                114 drivers/scsi/fnic/vnic_dev.c 	r = (struct vnic_resource __iomem *)(rh + 1);
rh               3026 drivers/scsi/lpfc/lpfc_ct.c 	struct lpfc_fdmi_reg_hba *rh;
rh               3083 drivers/scsi/lpfc/lpfc_ct.c 		rh = (struct lpfc_fdmi_reg_hba *)&CtReq->un.PortID;
rh               3085 drivers/scsi/lpfc/lpfc_ct.c 		memcpy(&rh->hi.PortName, &phba->pport->fc_sparam.portName,
rh               3091 drivers/scsi/lpfc/lpfc_ct.c 			rh->rpl.EntryCnt = cpu_to_be32(1);
rh               3092 drivers/scsi/lpfc/lpfc_ct.c 			memcpy(&rh->rpl.pe, &phba->pport->fc_sparam.portName,
rh               3101 drivers/scsi/lpfc/lpfc_ct.c 		ab = (struct lpfc_fdmi_attr_block *)((uint8_t *)rh + size);
rh               3116 drivers/scsi/lpfc/lpfc_ct.c 					     ((uint8_t *)rh + size));
rh               3135 drivers/scsi/lpfc/lpfc_ct.c 			rh = (struct lpfc_fdmi_reg_hba *)pab;
rh               3137 drivers/scsi/lpfc/lpfc_ct.c 			memcpy(&rh->hi.PortName,
rh                 85 drivers/scsi/snic/vnic_dev.c 	struct vnic_resource_header __iomem *rh;
rh                 98 drivers/scsi/snic/vnic_dev.c 	rh = bar->vaddr;
rh                 99 drivers/scsi/snic/vnic_dev.c 	if (!rh) {
rh                105 drivers/scsi/snic/vnic_dev.c 	if (ioread32(&rh->magic) != VNIC_RES_MAGIC ||
rh                106 drivers/scsi/snic/vnic_dev.c 	    ioread32(&rh->version) != VNIC_RES_VERSION) {
rh                109 drivers/scsi/snic/vnic_dev.c 			ioread32(&rh->magic), ioread32(&rh->version));
rh                114 drivers/scsi/snic/vnic_dev.c 	r = (struct vnic_resource __iomem *)(rh + 1);
rh                137 drivers/soc/qcom/cmd-db.c 			     const struct rsc_hdr **rh)
rh                161 drivers/soc/qcom/cmd-db.c 				if (rh)
rh                162 drivers/soc/qcom/cmd-db.c 					*rh = rsc_hdr;
rh                539 drivers/usb/host/ohci-hub.c 	u32		rh = roothub_a (ohci);
rh                543 drivers/usb/host/ohci-hub.c 	desc->bPwrOn2PwrGood = (rh & RH_A_POTPGT) >> 24;
rh                551 drivers/usb/host/ohci-hub.c 	if (rh & RH_A_NPS)		/* no power switching? */
rh                553 drivers/usb/host/ohci-hub.c 	if (rh & RH_A_PSM)		/* per-port power switching? */
rh                555 drivers/usb/host/ohci-hub.c 	if (rh & RH_A_NOCP)		/* no overcurrent reporting? */
rh                557 drivers/usb/host/ohci-hub.c 	else if (rh & RH_A_OCPM)	/* per-port overcurrent reporting? */
rh                562 drivers/usb/host/ohci-hub.c 	rh = roothub_b (ohci);
rh                565 drivers/usb/host/ohci-hub.c 	desc->u.hs.DeviceRemovable[0] = rh & RH_B_DR;
rh                567 drivers/usb/host/ohci-hub.c 		desc->u.hs.DeviceRemovable[1] = (rh & RH_B_DR) >> 8;
rh                256 drivers/usb/host/ohci-omap.c 		u32	rh = roothub_a (ohci);
rh                259 drivers/usb/host/ohci-omap.c 		rh &= ~RH_A_NPS;
rh                265 drivers/usb/host/ohci-omap.c 			rh &= ~RH_A_NOCP;
rh                275 drivers/usb/host/ohci-omap.c 		ohci_writel(ohci, rh, &ohci->regs->roothub.a);
rh                998 drivers/usb/host/r8a66597-hcd.c 	struct r8a66597_root_hub *rh = &r8a66597->root_hub[port];
rh               1000 drivers/usb/host/r8a66597-hcd.c 	rh->old_syssts = r8a66597_read(r8a66597, get_syssts_reg(port)) & LNST;
rh               1001 drivers/usb/host/r8a66597-hcd.c 	rh->scount = R8A66597_MAX_SAMPLING;
rh               1003 drivers/usb/host/r8a66597-hcd.c 		rh->port |= USB_PORT_STAT_CONNECTION;
rh               1005 drivers/usb/host/r8a66597-hcd.c 		rh->port &= ~USB_PORT_STAT_CONNECTION;
rh               1006 drivers/usb/host/r8a66597-hcd.c 	rh->port |= USB_PORT_STAT_C_CONNECTION << 16;
rh               1042 drivers/usb/host/r8a66597-hcd.c 	struct r8a66597_root_hub *rh = &r8a66597->root_hub[port];
rh               1044 drivers/usb/host/r8a66597-hcd.c 	rh->port &= ~(USB_PORT_STAT_HIGH_SPEED | USB_PORT_STAT_LOW_SPEED);
rh               1046 drivers/usb/host/r8a66597-hcd.c 		rh->port |= USB_PORT_STAT_HIGH_SPEED;
rh               1048 drivers/usb/host/r8a66597-hcd.c 		rh->port |= USB_PORT_STAT_LOW_SPEED;
rh               1050 drivers/usb/host/r8a66597-hcd.c 	rh->port &= ~USB_PORT_STAT_RESET;
rh               1051 drivers/usb/host/r8a66597-hcd.c 	rh->port |= USB_PORT_STAT_ENABLE;
rh               1688 drivers/usb/host/r8a66597-hcd.c 	struct r8a66597_root_hub *rh = &r8a66597->root_hub[port];
rh               1690 drivers/usb/host/r8a66597-hcd.c 	if (rh->port & USB_PORT_STAT_RESET) {
rh               1702 drivers/usb/host/r8a66597-hcd.c 	if (!(rh->port & USB_PORT_STAT_CONNECTION)) {
rh               1707 drivers/usb/host/r8a66597-hcd.c 	if (rh->scount > 0) {
rh               1709 drivers/usb/host/r8a66597-hcd.c 		if (tmp == rh->old_syssts) {
rh               1710 drivers/usb/host/r8a66597-hcd.c 			rh->scount--;
rh               1711 drivers/usb/host/r8a66597-hcd.c 			if (rh->scount == 0)
rh               1716 drivers/usb/host/r8a66597-hcd.c 			rh->scount = R8A66597_MAX_SAMPLING;
rh               1717 drivers/usb/host/r8a66597-hcd.c 			rh->old_syssts = tmp;
rh               2148 drivers/usb/host/r8a66597-hcd.c 	struct r8a66597_root_hub *rh = &r8a66597->root_hub[port];
rh               2173 drivers/usb/host/r8a66597-hcd.c 			rh->port &= ~USB_PORT_STAT_POWER;
rh               2189 drivers/usb/host/r8a66597-hcd.c 		rh->port &= ~(1 << wValue);
rh               2201 drivers/usb/host/r8a66597-hcd.c 		*(__le32 *)buf = cpu_to_le32(rh->port);
rh               2214 drivers/usb/host/r8a66597-hcd.c 			rh->port |= USB_PORT_STAT_POWER;
rh               2217 drivers/usb/host/r8a66597-hcd.c 			struct r8a66597_device *dev = rh->dev;
rh               2219 drivers/usb/host/r8a66597-hcd.c 			rh->port |= USB_PORT_STAT_RESET;
rh               2233 drivers/usb/host/r8a66597-hcd.c 		rh->port |= 1 << wValue;
rh               2254 drivers/usb/host/r8a66597-hcd.c 		struct r8a66597_root_hub *rh = &r8a66597->root_hub[port];
rh               2257 drivers/usb/host/r8a66597-hcd.c 		if (!(rh->port & USB_PORT_STAT_ENABLE))
rh               2260 drivers/usb/host/r8a66597-hcd.c 		dev_dbg(&rh->dev->udev->dev, "suspend port = %d\n", port);
rh               2262 drivers/usb/host/r8a66597-hcd.c 		rh->port |= USB_PORT_STAT_SUSPEND;
rh               2264 drivers/usb/host/r8a66597-hcd.c 		if (rh->dev->udev->do_remote_wakeup) {
rh               2285 drivers/usb/host/r8a66597-hcd.c 		struct r8a66597_root_hub *rh = &r8a66597->root_hub[port];
rh               2288 drivers/usb/host/r8a66597-hcd.c 		if (!(rh->port & USB_PORT_STAT_SUSPEND))
rh               2291 drivers/usb/host/r8a66597-hcd.c 		dev_dbg(&rh->dev->udev->dev, "resume port = %d\n", port);
rh               2292 drivers/usb/host/r8a66597-hcd.c 		rh->port &= ~USB_PORT_STAT_SUSPEND;
rh               2293 drivers/usb/host/r8a66597-hcd.c 		rh->port |= USB_PORT_STAT_C_SUSPEND << 16;
rh               2352 drivers/usb/host/r8a66597-hcd.c 		struct r8a66597_root_hub *rh = &r8a66597->root_hub[port];
rh               2354 drivers/usb/host/r8a66597-hcd.c 		rh->port = 0x00000000;
rh                 43 include/linux/dm-region-hash.h void dm_region_hash_destroy(struct dm_region_hash *rh);
rh                 45 include/linux/dm-region-hash.h struct dm_dirty_log *dm_rh_dirty_log(struct dm_region_hash *rh);
rh                 50 include/linux/dm-region-hash.h region_t dm_rh_bio_to_region(struct dm_region_hash *rh, struct bio *bio);
rh                 51 include/linux/dm-region-hash.h sector_t dm_rh_region_to_sector(struct dm_region_hash *rh, region_t region);
rh                 57 include/linux/dm-region-hash.h sector_t dm_rh_get_region_size(struct dm_region_hash *rh);
rh                 64 include/linux/dm-region-hash.h int dm_rh_get_state(struct dm_region_hash *rh, region_t region, int may_block);
rh                 65 include/linux/dm-region-hash.h void dm_rh_set_state(struct dm_region_hash *rh, region_t region,
rh                 69 include/linux/dm-region-hash.h void dm_rh_update_states(struct dm_region_hash *rh, int errors_handled);
rh                 72 include/linux/dm-region-hash.h int dm_rh_flush(struct dm_region_hash *rh);
rh                 75 include/linux/dm-region-hash.h void dm_rh_inc_pending(struct dm_region_hash *rh, struct bio_list *bios);
rh                 76 include/linux/dm-region-hash.h void dm_rh_dec(struct dm_region_hash *rh, region_t region);
rh                 79 include/linux/dm-region-hash.h void dm_rh_delay(struct dm_region_hash *rh, struct bio *bio);
rh                 81 include/linux/dm-region-hash.h void dm_rh_mark_nosync(struct dm_region_hash *rh, struct bio *bio);
rh                 88 include/linux/dm-region-hash.h void dm_rh_recovery_prepare(struct dm_region_hash *rh);
rh                 91 include/linux/dm-region-hash.h struct dm_region *dm_rh_recovery_start(struct dm_region_hash *rh);
rh                 97 include/linux/dm-region-hash.h int dm_rh_recovery_in_flight(struct dm_region_hash *rh);
rh                100 include/linux/dm-region-hash.h void dm_rh_start_recovery(struct dm_region_hash *rh);
rh                101 include/linux/dm-region-hash.h void dm_rh_stop_recovery(struct dm_region_hash *rh);
rh                 28 include/linux/fsl/bestcomm/sram.h 	rh_info_t		*rh;
rh                221 include/linux/math64.h 	} rl, rm, rn, rh, a0, b0;
rh                230 include/linux/math64.h 	rh.ll = mul_u32_u32(a0.l.high, b0.l.high);
rh                238 include/linux/math64.h 	rh.l.low = c = (c >> 32) + rm.l.high + rn.l.high + rh.l.low;
rh                239 include/linux/math64.h 	rh.l.high = (c >> 32) + rh.l.high;
rh                248 include/linux/math64.h 		return (rl.ll >> shift) | (rh.ll << (64 - shift));
rh                249 include/linux/math64.h 	return rh.ll >> (shift & 63);
rh                267 include/linux/math64.h 	} u, rl, rh;
rh                271 include/linux/math64.h 	rh.ll = mul_u32_u32(u.l.high, mul) + rl.l.high;
rh                274 include/linux/math64.h 	rl.l.high = do_div(rh.ll, divisor);
rh                279 include/linux/math64.h 	rl.l.high = rh.l.low;
rh                147 include/math-emu/op-2.h #define __FP_FRAC_ADD_2(rh, rl, xh, xl, yh, yl)	\
rh                148 include/math-emu/op-2.h   (rh = xh + yh + ((rl = xl + yl) < xl))
rh                151 include/math-emu/op-2.h #define __FP_FRAC_SUB_2(rh, rl, xh, xl, yh, yl)	\
rh                152 include/math-emu/op-2.h   (rh = xh - yh - ((rl = xl - yl) > xl))
rh               1635 kernel/rcu/rcutorture.c 	struct rcu_head rh;
rh               1646 kernel/rcu/rcutorture.c 	struct fwd_cb_state *fcsp = container_of(rhp, struct fwd_cb_state, rh);
rh               1652 kernel/rcu/rcutorture.c 	cur_ops->call(&fcsp->rh, rcu_torture_fwd_prog_cb);
rh               1657 kernel/rcu/rcutorture.c 	struct rcu_head rh;
rh               1707 kernel/rcu/rcutorture.c 	struct rcu_fwd_cb *rfcp = container_of(rhp, struct rcu_fwd_cb, rh);
rh               1781 kernel/rcu/rcutorture.c 		init_rcu_head_on_stack(&fcs.rh);
rh               1790 kernel/rcu/rcutorture.c 		cur_ops->call(&fcs.rh, rcu_torture_fwd_prog_cb);
rh               1826 kernel/rcu/rcutorture.c 		destroy_rcu_head_on_stack(&fcs.rh);
rh               1892 kernel/rcu/rcutorture.c 		cur_ops->call(&rfcp->rh, rcu_torture_fwd_cb_cr);
rh                135 net/dccp/ccids/lib/loss_interval.c int tfrc_lh_interval_add(struct tfrc_loss_hist *lh, struct tfrc_rx_hist *rh,
rh                140 net/dccp/ccids/lib/loss_interval.c 	if (cur != NULL && !tfrc_lh_is_new_loss(cur, tfrc_rx_hist_loss_prev(rh)))
rh                149 net/dccp/ccids/lib/loss_interval.c 	new->li_seqno	  = tfrc_rx_hist_loss_prev(rh)->tfrchrx_seqno;
rh                150 net/dccp/ccids/lib/loss_interval.c 	new->li_ccval	  = tfrc_rx_hist_loss_prev(rh)->tfrchrx_ccval;
rh                158 net/dccp/ccids/lib/loss_interval.c 				  tfrc_rx_hist_last_rcv(rh)->tfrchrx_seqno) + 1;
rh                222 net/ipv6/exthdrs_core.c 			struct ipv6_rt_hdr _rh, *rh;
rh                224 net/ipv6/exthdrs_core.c 			rh = skb_header_pointer(skb, start, sizeof(_rh),
rh                226 net/ipv6/exthdrs_core.c 			if (!rh)
rh                230 net/ipv6/exthdrs_core.c 			    rh->segments_left == 0)
rh                 39 net/ipv6/netfilter/ip6t_rt.c 	const struct ipv6_rt_hdr *rh;
rh                 56 net/ipv6/netfilter/ip6t_rt.c 	rh = skb_header_pointer(skb, ptr, sizeof(_route), &_route);
rh                 57 net/ipv6/netfilter/ip6t_rt.c 	if (rh == NULL) {
rh                 62 net/ipv6/netfilter/ip6t_rt.c 	hdrlen = ipv6_optlen(rh);
rh                 68 net/ipv6/netfilter/ip6t_rt.c 	pr_debug("IPv6 RT LEN %u %u ", hdrlen, rh->hdrlen);
rh                 69 net/ipv6/netfilter/ip6t_rt.c 	pr_debug("TYPE %04X ", rh->type);
rh                 70 net/ipv6/netfilter/ip6t_rt.c 	pr_debug("SGS_LEFT %u %02X\n", rh->segments_left, rh->segments_left);
rh                 74 net/ipv6/netfilter/ip6t_rt.c 				rh->segments_left,
rh                 77 net/ipv6/netfilter/ip6t_rt.c 		 rtinfo->rt_type, rh->type,
rh                 79 net/ipv6/netfilter/ip6t_rt.c 		  ((rtinfo->rt_type == rh->type) ^
rh                 88 net/ipv6/netfilter/ip6t_rt.c 		 ((const struct rt0_hdr *)rh)->reserved,
rh                 90 net/ipv6/netfilter/ip6t_rt.c 		   (((const struct rt0_hdr *)rh)->reserved)));
rh                 92 net/ipv6/netfilter/ip6t_rt.c 	ret = (rh != NULL) &&
rh                 94 net/ipv6/netfilter/ip6t_rt.c 			      rh->segments_left,
rh                100 net/ipv6/netfilter/ip6t_rt.c 	       ((rtinfo->rt_type == rh->type) ^
rh                996 sound/soc/codecs/rt1305.c 	unsigned int rh, rl, rhl, r0ohm;
rh               1060 sound/soc/codecs/rt1305.c 	regmap_read(rt1305->regmap, RT1305_PR_BASE + 0x55, &rh);
rh               1062 sound/soc/codecs/rt1305.c 	rhl = (rh << 16) | rl;
rh               1065 sound/soc/codecs/rt1305.c 	pr_debug("Left_rhl = 0x%x rh=0x%x rl=0x%x\n", rhl, rh, rl);
rh               1077 sound/soc/codecs/rt1305.c 	regmap_read(rt1305->regmap, RT1305_PR_BASE + 0x55, &rh);
rh               1079 sound/soc/codecs/rt1305.c 	rhl = (rh << 16) | rl;
rh               1082 sound/soc/codecs/rt1305.c 	pr_debug("Right_rhl = 0x%x rh=0x%x rl=0x%x\n", rhl, rh, rl);