rh 73 arch/arm/vfp/vfp.h u64 rh, rma, rmb, rl; rh 86 arch/arm/vfp/vfp.h rh = (u64)nh * mh; rh 87 arch/arm/vfp/vfp.h rh += ((u64)(rma < rmb) << 32) + (rma >> 32); rh 91 arch/arm/vfp/vfp.h rh += (rl < rma); rh 94 arch/arm/vfp/vfp.h *resh = rh; rh 105 arch/arm/vfp/vfp.h u64 rh, rl; rh 106 arch/arm/vfp/vfp.h mul64to128(&rh, &rl, n, m); rh 107 arch/arm/vfp/vfp.h return rh | (rl != 0); rh 464 arch/ia64/kernel/mca.c sal_log_record_header_t *rh; rh 483 arch/ia64/kernel/mca.c rh = (sal_log_record_header_t *)buffer; rh 484 arch/ia64/kernel/mca.c if (rh->severity == sal_log_severity_corrected) rh 1335 arch/ia64/kernel/mca.c sal_log_record_header_t *rh = IA64_LOG_CURR_BUFFER(SAL_INFO_TYPE_MCA); rh 1336 arch/ia64/kernel/mca.c rh->severity = sal_log_severity_corrected; rh 386 arch/ia64/kernel/salinfo.c sal_log_record_header_t *rh; rh 388 arch/ia64/kernel/salinfo.c rh = (sal_log_record_header_t *)(data->log_buffer); rh 390 arch/ia64/kernel/salinfo.c if (rh->severity == sal_log_severity_corrected) rh 408 arch/ia64/kernel/salinfo.c sal_log_record_header_t *rh = (sal_log_record_header_t *)(data_saved->buffer); rh 410 arch/ia64/kernel/salinfo.c memcpy(data->log_buffer, rh, data->log_size); rh 412 arch/ia64/kernel/salinfo.c if (rh->id == data_saved->id) { rh 465 arch/ia64/kernel/salinfo.c sal_log_record_header_t *rh; rh 479 arch/ia64/kernel/salinfo.c rh = (sal_log_record_header_t *)(data->log_buffer); rh 481 arch/ia64/kernel/salinfo.c if (rh->severity != sal_log_severity_corrected) rh 27 arch/powerpc/include/asm/fsl_85xx_cache_sram.h rh_info_t *rh; rh 47 arch/powerpc/sysdev/fsl_85xx_cache_sram.c offset = rh_alloc_align(cache_sram->rh, size, align, NULL); rh 65 arch/powerpc/sysdev/fsl_85xx_cache_sram.c rh_free(cache_sram->rh, ptr - cache_sram->base_virt); rh 106 arch/powerpc/sysdev/fsl_85xx_cache_sram.c cache_sram->rh = rh_create(sizeof(unsigned int)); rh 107 arch/powerpc/sysdev/fsl_85xx_cache_sram.c if (IS_ERR(cache_sram->rh)) { rh 110 arch/powerpc/sysdev/fsl_85xx_cache_sram.c ret = PTR_ERR(cache_sram->rh); rh 114 arch/powerpc/sysdev/fsl_85xx_cache_sram.c rh_attach_region(cache_sram->rh, 0, cache_sram->size); rh 137 arch/powerpc/sysdev/fsl_85xx_cache_sram.c rh_detach_region(cache_sram->rh, 0, cache_sram->size); rh 138 arch/powerpc/sysdev/fsl_85xx_cache_sram.c rh_destroy(cache_sram->rh); rh 103 crypto/vmac.c #define ADD128(rh, rl, ih, il) \ rh 108 crypto/vmac.c (rh)++; \ rh 109 crypto/vmac.c (rh) += (ih); \ rh 114 crypto/vmac.c #define PMUL64(rh, rl, i1, i2) /* Assumes m doesn't overflow */ \ rh 118 crypto/vmac.c rh = MUL32(_i1>>32, _i2>>32); \ rh 120 crypto/vmac.c ADD128(rh, rl, (m >> 32), (m << 32)); \ rh 123 crypto/vmac.c #define MUL64(rh, rl, i1, i2) \ rh 128 crypto/vmac.c rh = MUL32(_i1>>32, _i2>>32); \ rh 130 crypto/vmac.c ADD128(rh, rl, (m1 >> 32), (m1 << 32)); \ rh 131 crypto/vmac.c ADD128(rh, rl, (m2 >> 32), (m2 << 32)); \ rh 148 crypto/vmac.c #define nh_16(mp, kp, nw, rh, rl) \ rh 151 crypto/vmac.c rh = rl = 0; \ rh 155 crypto/vmac.c ADD128(rh, rl, th, tl); \ rh 159 crypto/vmac.c #define nh_16_2(mp, kp, nw, rh, rl, rh1, rl1) \ rh 162 crypto/vmac.c rh1 = rl1 = rh = rl = 0; \ rh 166 crypto/vmac.c ADD128(rh, rl, th, tl); \ rh 174 crypto/vmac.c #define nh_vmac_nhbytes(mp, kp, nw, rh, rl) \ rh 177 crypto/vmac.c rh = rl = 0; \ rh 181 crypto/vmac.c ADD128(rh, rl, th, tl); \ rh 184 crypto/vmac.c ADD128(rh, rl, th, tl); \ rh 187 crypto/vmac.c ADD128(rh, rl, th, tl); \ rh 190 crypto/vmac.c ADD128(rh, rl, th, tl); \ rh 194 crypto/vmac.c #define nh_vmac_nhbytes_2(mp, kp, nw, rh, rl, rh1, rl1) \ rh 197 crypto/vmac.c rh1 = rl1 = rh = rl = 0; \ rh 201 crypto/vmac.c ADD128(rh, rl, th, tl); \ rh 207 crypto/vmac.c ADD128(rh, rl, th, tl); \ rh 213 crypto/vmac.c ADD128(rh, rl, th, tl); \ rh 219 crypto/vmac.c ADD128(rh, rl, th, tl); \ rh 253 crypto/vmac.c #define nh_16(mp, kp, nw, rh, rl) \ rh 257 crypto/vmac.c rh = rl = t = 0; \ rh 263 crypto/vmac.c ADD128(rh, rl, MUL32(t1 >> 32, t2 >> 32), \ rh 265 crypto/vmac.c rh += (u64)(u32)(m1 >> 32) \ rh 269 crypto/vmac.c ADD128(rh, rl, (t >> 32), (t << 32)); \ rh 340 crypto/vmac.c #define nh_16_2(mp, kp, nw, rh, rl, rh2, rl2) \ rh 342 crypto/vmac.c nh_16(mp, kp, nw, rh, rl); \ rh 347 crypto/vmac.c #define nh_vmac_nhbytes(mp, kp, nw, rh, rl) \ rh 348 crypto/vmac.c nh_16(mp, kp, nw, rh, rl) rh 351 crypto/vmac.c #define nh_vmac_nhbytes_2(mp, kp, nw, rh, rl, rh2, rl2) \ rh 353 crypto/vmac.c nh_vmac_nhbytes(mp, kp, nw, rh, rl); \ rh 360 crypto/vmac.c u64 rh, rl, t, z = 0; rh 385 crypto/vmac.c MUL64(rh, rl, p1, p2); rh 386 crypto/vmac.c t = rh >> 56; rh 387 crypto/vmac.c ADD128(t, rl, z, rh); rh 388 crypto/vmac.c rh <<= 8; rh 389 crypto/vmac.c ADD128(t, rl, z, rh); rh 407 crypto/vmac.c u64 rh, rl; rh 411 crypto/vmac.c nh_vmac_nhbytes(mptr, kptr, VMAC_NHBYTES/8, rh, rl); rh 412 crypto/vmac.c rh &= m62; rh 413 crypto/vmac.c ADD128(ch, cl, rh, rl); rh 419 crypto/vmac.c nh_vmac_nhbytes(mptr, kptr, VMAC_NHBYTES/8, rh, rl); rh 420 crypto/vmac.c rh &= m62; rh 421 crypto/vmac.c poly_step(ch, cl, pkh, pkl, rh, rl); rh 547 crypto/vmac.c u64 rh, rl; rh 550 crypto/vmac.c nh_16(dctx->partial_words, tctx->nhkey, n / 8, rh, rl); rh 551 crypto/vmac.c rh &= m62; rh 554 crypto/vmac.c rh, rl); rh 556 crypto/vmac.c ADD128(ch, cl, rh, rl); rh 489 drivers/crypto/cavium/nitrox/nitrox_csr.h u64 rh : 1; rh 495 drivers/crypto/cavium/nitrox/nitrox_csr.h u64 rh : 1; rh 205 drivers/crypto/cavium/nitrox/nitrox_hal.c pkt_slc_ctl.s.rh = 1; rh 97 drivers/dma/bestcomm/sram.c bcom_sram->rh = rh_create(4); rh 110 drivers/dma/bestcomm/sram.c rh_attach_region(bcom_sram->rh, 0, bcom_sram->size); rh 115 drivers/dma/bestcomm/sram.c rh_attach_region(bcom_sram->rh, zbase - bcom_sram->base_phys, regaddr_p[1]); rh 140 drivers/dma/bestcomm/sram.c rh_destroy(bcom_sram->rh); rh 154 drivers/dma/bestcomm/sram.c offset = rh_alloc_align(bcom_sram->rh, size, align, NULL); rh 175 drivers/dma/bestcomm/sram.c rh_free(bcom_sram->rh, offset); rh 65 drivers/md/dm-raid1.c struct dm_region_hash *rh; rh 336 drivers/md/dm-raid1.c sector_t region_size = dm_rh_get_region_size(ms->rh); rh 341 drivers/md/dm-raid1.c from.sector = m->offset + dm_rh_region_to_sector(ms->rh, key); rh 360 drivers/md/dm-raid1.c dest->sector = m->offset + dm_rh_region_to_sector(ms->rh, key); rh 387 drivers/md/dm-raid1.c struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); rh 392 drivers/md/dm-raid1.c dm_rh_recovery_prepare(ms->rh); rh 397 drivers/md/dm-raid1.c while ((reg = dm_rh_recovery_start(ms->rh))) rh 439 drivers/md/dm-raid1.c struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); rh 440 drivers/md/dm-raid1.c region_t region = dm_rh_bio_to_region(ms->rh, bio); rh 555 drivers/md/dm-raid1.c int state = dm_rh_get_state(ms->rh, region, may_block); rh 566 drivers/md/dm-raid1.c region = dm_rh_bio_to_region(ms->rh, bio); rh 685 drivers/md/dm-raid1.c struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); rh 706 drivers/md/dm-raid1.c region = dm_rh_bio_to_region(ms->rh, bio); rh 714 drivers/md/dm-raid1.c state = dm_rh_get_state(ms->rh, region, 1); rh 749 drivers/md/dm-raid1.c dm_rh_inc_pending(ms->rh, &sync); rh 750 drivers/md/dm-raid1.c dm_rh_inc_pending(ms->rh, &nosync); rh 757 drivers/md/dm-raid1.c ms->log_failure = dm_rh_flush(ms->rh) ? 1 : ms->log_failure; rh 772 drivers/md/dm-raid1.c dm_rh_delay(ms->rh, bio); rh 814 drivers/md/dm-raid1.c dm_rh_mark_nosync(ms->rh, bio); rh 866 drivers/md/dm-raid1.c dm_rh_update_states(ms->rh, errors_handled(ms)); rh 911 drivers/md/dm-raid1.c ms->rh = dm_region_hash_create(ms, dispatch_bios, wakeup_mirrord, rh 915 drivers/md/dm-raid1.c if (IS_ERR(ms->rh)) { rh 932 drivers/md/dm-raid1.c dm_region_hash_destroy(ms->rh); rh 1112 drivers/md/dm-raid1.c r = dm_set_target_max_io_len(ti, dm_rh_get_region_size(ms->rh)); rh 1189 drivers/md/dm-raid1.c struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); rh 1197 drivers/md/dm-raid1.c bio_record->write_region = dm_rh_bio_to_region(ms->rh, bio); rh 1202 drivers/md/dm-raid1.c r = log->type->in_sync(log, dm_rh_bio_to_region(ms->rh, bio), 0); rh 1249 drivers/md/dm-raid1.c dm_rh_dec(ms->rh, bio_record->write_region); rh 1303 drivers/md/dm-raid1.c struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); rh 1328 drivers/md/dm-raid1.c dm_rh_stop_recovery(ms->rh); rh 1331 drivers/md/dm-raid1.c !dm_rh_recovery_in_flight(ms->rh)); rh 1349 drivers/md/dm-raid1.c struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); rh 1359 drivers/md/dm-raid1.c struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); rh 1365 drivers/md/dm-raid1.c dm_rh_start_recovery(ms->rh); rh 1399 drivers/md/dm-raid1.c struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); rh 103 drivers/md/dm-region-hash.c struct dm_region_hash *rh; /* FIXME: can we get rid of this ? */ rh 117 drivers/md/dm-region-hash.c static region_t dm_rh_sector_to_region(struct dm_region_hash *rh, sector_t sector) rh 119 drivers/md/dm-region-hash.c return sector >> rh->region_shift; rh 122 drivers/md/dm-region-hash.c sector_t dm_rh_region_to_sector(struct dm_region_hash *rh, region_t region) rh 124 drivers/md/dm-region-hash.c return region << rh->region_shift; rh 128 drivers/md/dm-region-hash.c region_t dm_rh_bio_to_region(struct dm_region_hash *rh, struct bio *bio) rh 130 drivers/md/dm-region-hash.c return dm_rh_sector_to_region(rh, bio->bi_iter.bi_sector - rh 131 drivers/md/dm-region-hash.c rh->target_begin); rh 137 drivers/md/dm-region-hash.c return reg->rh->context; rh 147 drivers/md/dm-region-hash.c sector_t dm_rh_get_region_size(struct dm_region_hash *rh) rh 149 drivers/md/dm-region-hash.c return rh->region_size; rh 170 drivers/md/dm-region-hash.c struct dm_region_hash *rh; rh 184 drivers/md/dm-region-hash.c rh = kzalloc(sizeof(*rh), GFP_KERNEL); rh 185 drivers/md/dm-region-hash.c if (!rh) { rh 190 drivers/md/dm-region-hash.c rh->context = context; rh 191 drivers/md/dm-region-hash.c rh->dispatch_bios = dispatch_bios; rh 192 drivers/md/dm-region-hash.c rh->wakeup_workers = wakeup_workers; rh 193 drivers/md/dm-region-hash.c rh->wakeup_all_recovery_waiters = wakeup_all_recovery_waiters; rh 194 drivers/md/dm-region-hash.c rh->target_begin = target_begin; rh 195 drivers/md/dm-region-hash.c rh->max_recovery = max_recovery; rh 196 drivers/md/dm-region-hash.c rh->log = log; rh 197 drivers/md/dm-region-hash.c rh->region_size = region_size; rh 198 drivers/md/dm-region-hash.c rh->region_shift = __ffs(region_size); rh 199 drivers/md/dm-region-hash.c rwlock_init(&rh->hash_lock); rh 200 drivers/md/dm-region-hash.c rh->mask = nr_buckets - 1; rh 201 drivers/md/dm-region-hash.c rh->nr_buckets = nr_buckets; rh 203 drivers/md/dm-region-hash.c rh->shift = RH_HASH_SHIFT; rh 204 drivers/md/dm-region-hash.c rh->prime = RH_HASH_MULT; rh 206 drivers/md/dm-region-hash.c rh->buckets = vmalloc(array_size(nr_buckets, sizeof(*rh->buckets))); rh 207 drivers/md/dm-region-hash.c if (!rh->buckets) { rh 209 drivers/md/dm-region-hash.c kfree(rh); rh 214 drivers/md/dm-region-hash.c INIT_LIST_HEAD(rh->buckets + i); rh 216 drivers/md/dm-region-hash.c spin_lock_init(&rh->region_lock); rh 217 drivers/md/dm-region-hash.c sema_init(&rh->recovery_count, 0); rh 218 drivers/md/dm-region-hash.c atomic_set(&rh->recovery_in_flight, 0); rh 219 drivers/md/dm-region-hash.c INIT_LIST_HEAD(&rh->clean_regions); rh 220 drivers/md/dm-region-hash.c INIT_LIST_HEAD(&rh->quiesced_regions); rh 221 drivers/md/dm-region-hash.c INIT_LIST_HEAD(&rh->recovered_regions); rh 222 drivers/md/dm-region-hash.c INIT_LIST_HEAD(&rh->failed_recovered_regions); rh 223 drivers/md/dm-region-hash.c rh->flush_failure = 0; rh 225 drivers/md/dm-region-hash.c ret = mempool_init_kmalloc_pool(&rh->region_pool, MIN_REGIONS, rh 228 drivers/md/dm-region-hash.c vfree(rh->buckets); rh 229 drivers/md/dm-region-hash.c kfree(rh); rh 230 drivers/md/dm-region-hash.c rh = ERR_PTR(-ENOMEM); rh 233 drivers/md/dm-region-hash.c return rh; rh 237 drivers/md/dm-region-hash.c void dm_region_hash_destroy(struct dm_region_hash *rh) rh 242 drivers/md/dm-region-hash.c BUG_ON(!list_empty(&rh->quiesced_regions)); rh 243 drivers/md/dm-region-hash.c for (h = 0; h < rh->nr_buckets; h++) { rh 244 drivers/md/dm-region-hash.c list_for_each_entry_safe(reg, nreg, rh->buckets + h, rh 247 drivers/md/dm-region-hash.c mempool_free(reg, &rh->region_pool); rh 251 drivers/md/dm-region-hash.c if (rh->log) rh 252 drivers/md/dm-region-hash.c dm_dirty_log_destroy(rh->log); rh 254 drivers/md/dm-region-hash.c mempool_exit(&rh->region_pool); rh 255 drivers/md/dm-region-hash.c vfree(rh->buckets); rh 256 drivers/md/dm-region-hash.c kfree(rh); rh 260 drivers/md/dm-region-hash.c struct dm_dirty_log *dm_rh_dirty_log(struct dm_region_hash *rh) rh 262 drivers/md/dm-region-hash.c return rh->log; rh 266 drivers/md/dm-region-hash.c static unsigned rh_hash(struct dm_region_hash *rh, region_t region) rh 268 drivers/md/dm-region-hash.c return (unsigned) ((region * rh->prime) >> rh->shift) & rh->mask; rh 271 drivers/md/dm-region-hash.c static struct dm_region *__rh_lookup(struct dm_region_hash *rh, region_t region) rh 274 drivers/md/dm-region-hash.c struct list_head *bucket = rh->buckets + rh_hash(rh, region); rh 283 drivers/md/dm-region-hash.c static void __rh_insert(struct dm_region_hash *rh, struct dm_region *reg) rh 285 drivers/md/dm-region-hash.c list_add(®->hash_list, rh->buckets + rh_hash(rh, reg->key)); rh 288 drivers/md/dm-region-hash.c static struct dm_region *__rh_alloc(struct dm_region_hash *rh, region_t region) rh 292 drivers/md/dm-region-hash.c nreg = mempool_alloc(&rh->region_pool, GFP_ATOMIC); rh 296 drivers/md/dm-region-hash.c nreg->state = rh->log->type->in_sync(rh->log, region, 1) ? rh 298 drivers/md/dm-region-hash.c nreg->rh = rh; rh 304 drivers/md/dm-region-hash.c write_lock_irq(&rh->hash_lock); rh 305 drivers/md/dm-region-hash.c reg = __rh_lookup(rh, region); rh 308 drivers/md/dm-region-hash.c mempool_free(nreg, &rh->region_pool); rh 310 drivers/md/dm-region-hash.c __rh_insert(rh, nreg); rh 312 drivers/md/dm-region-hash.c spin_lock(&rh->region_lock); rh 313 drivers/md/dm-region-hash.c list_add(&nreg->list, &rh->clean_regions); rh 314 drivers/md/dm-region-hash.c spin_unlock(&rh->region_lock); rh 319 drivers/md/dm-region-hash.c write_unlock_irq(&rh->hash_lock); rh 324 drivers/md/dm-region-hash.c static struct dm_region *__rh_find(struct dm_region_hash *rh, region_t region) rh 328 drivers/md/dm-region-hash.c reg = __rh_lookup(rh, region); rh 330 drivers/md/dm-region-hash.c read_unlock(&rh->hash_lock); rh 331 drivers/md/dm-region-hash.c reg = __rh_alloc(rh, region); rh 332 drivers/md/dm-region-hash.c read_lock(&rh->hash_lock); rh 338 drivers/md/dm-region-hash.c int dm_rh_get_state(struct dm_region_hash *rh, region_t region, int may_block) rh 343 drivers/md/dm-region-hash.c read_lock(&rh->hash_lock); rh 344 drivers/md/dm-region-hash.c reg = __rh_lookup(rh, region); rh 345 drivers/md/dm-region-hash.c read_unlock(&rh->hash_lock); rh 354 drivers/md/dm-region-hash.c r = rh->log->type->in_sync(rh->log, region, may_block); rh 366 drivers/md/dm-region-hash.c struct dm_region_hash *rh = reg->rh; rh 368 drivers/md/dm-region-hash.c rh->log->type->set_region_sync(rh->log, reg->key, success); rh 379 drivers/md/dm-region-hash.c rh->dispatch_bios(rh->context, ®->delayed_bios); rh 380 drivers/md/dm-region-hash.c if (atomic_dec_and_test(&rh->recovery_in_flight)) rh 381 drivers/md/dm-region-hash.c rh->wakeup_all_recovery_waiters(rh->context); rh 382 drivers/md/dm-region-hash.c up(&rh->recovery_count); rh 395 drivers/md/dm-region-hash.c void dm_rh_mark_nosync(struct dm_region_hash *rh, struct bio *bio) rh 398 drivers/md/dm-region-hash.c struct dm_dirty_log *log = rh->log; rh 400 drivers/md/dm-region-hash.c region_t region = dm_rh_bio_to_region(rh, bio); rh 404 drivers/md/dm-region-hash.c rh->flush_failure = 1; rh 414 drivers/md/dm-region-hash.c read_lock(&rh->hash_lock); rh 415 drivers/md/dm-region-hash.c reg = __rh_find(rh, region); rh 416 drivers/md/dm-region-hash.c read_unlock(&rh->hash_lock); rh 422 drivers/md/dm-region-hash.c spin_lock_irqsave(&rh->region_lock, flags); rh 433 drivers/md/dm-region-hash.c spin_unlock_irqrestore(&rh->region_lock, flags); rh 440 drivers/md/dm-region-hash.c void dm_rh_update_states(struct dm_region_hash *rh, int errors_handled) rh 451 drivers/md/dm-region-hash.c write_lock_irq(&rh->hash_lock); rh 452 drivers/md/dm-region-hash.c spin_lock(&rh->region_lock); rh 453 drivers/md/dm-region-hash.c if (!list_empty(&rh->clean_regions)) { rh 454 drivers/md/dm-region-hash.c list_splice_init(&rh->clean_regions, &clean); rh 460 drivers/md/dm-region-hash.c if (!list_empty(&rh->recovered_regions)) { rh 461 drivers/md/dm-region-hash.c list_splice_init(&rh->recovered_regions, &recovered); rh 467 drivers/md/dm-region-hash.c if (!list_empty(&rh->failed_recovered_regions)) { rh 468 drivers/md/dm-region-hash.c list_splice_init(&rh->failed_recovered_regions, rh 475 drivers/md/dm-region-hash.c spin_unlock(&rh->region_lock); rh 476 drivers/md/dm-region-hash.c write_unlock_irq(&rh->hash_lock); rh 484 drivers/md/dm-region-hash.c rh->log->type->clear_region(rh->log, reg->key); rh 486 drivers/md/dm-region-hash.c mempool_free(reg, &rh->region_pool); rh 491 drivers/md/dm-region-hash.c mempool_free(reg, &rh->region_pool); rh 495 drivers/md/dm-region-hash.c rh->log->type->clear_region(rh->log, reg->key); rh 496 drivers/md/dm-region-hash.c mempool_free(reg, &rh->region_pool); rh 499 drivers/md/dm-region-hash.c rh->log->type->flush(rh->log); rh 503 drivers/md/dm-region-hash.c static void rh_inc(struct dm_region_hash *rh, region_t region) rh 507 drivers/md/dm-region-hash.c read_lock(&rh->hash_lock); rh 508 drivers/md/dm-region-hash.c reg = __rh_find(rh, region); rh 510 drivers/md/dm-region-hash.c spin_lock_irq(&rh->region_lock); rh 516 drivers/md/dm-region-hash.c spin_unlock_irq(&rh->region_lock); rh 518 drivers/md/dm-region-hash.c rh->log->type->mark_region(rh->log, reg->key); rh 520 drivers/md/dm-region-hash.c spin_unlock_irq(&rh->region_lock); rh 523 drivers/md/dm-region-hash.c read_unlock(&rh->hash_lock); rh 526 drivers/md/dm-region-hash.c void dm_rh_inc_pending(struct dm_region_hash *rh, struct bio_list *bios) rh 533 drivers/md/dm-region-hash.c rh_inc(rh, dm_rh_bio_to_region(rh, bio)); rh 538 drivers/md/dm-region-hash.c void dm_rh_dec(struct dm_region_hash *rh, region_t region) rh 544 drivers/md/dm-region-hash.c read_lock(&rh->hash_lock); rh 545 drivers/md/dm-region-hash.c reg = __rh_lookup(rh, region); rh 546 drivers/md/dm-region-hash.c read_unlock(&rh->hash_lock); rh 548 drivers/md/dm-region-hash.c spin_lock_irqsave(&rh->region_lock, flags); rh 562 drivers/md/dm-region-hash.c if (unlikely(rh->flush_failure)) { rh 570 drivers/md/dm-region-hash.c list_add_tail(®->list, &rh->quiesced_regions); rh 573 drivers/md/dm-region-hash.c list_add(®->list, &rh->clean_regions); rh 577 drivers/md/dm-region-hash.c spin_unlock_irqrestore(&rh->region_lock, flags); rh 580 drivers/md/dm-region-hash.c rh->wakeup_workers(rh->context); rh 587 drivers/md/dm-region-hash.c static int __rh_recovery_prepare(struct dm_region_hash *rh) rh 596 drivers/md/dm-region-hash.c r = rh->log->type->get_resync_work(rh->log, ®ion); rh 604 drivers/md/dm-region-hash.c read_lock(&rh->hash_lock); rh 605 drivers/md/dm-region-hash.c reg = __rh_find(rh, region); rh 606 drivers/md/dm-region-hash.c read_unlock(&rh->hash_lock); rh 608 drivers/md/dm-region-hash.c spin_lock_irq(&rh->region_lock); rh 615 drivers/md/dm-region-hash.c list_move(®->list, &rh->quiesced_regions); rh 617 drivers/md/dm-region-hash.c spin_unlock_irq(&rh->region_lock); rh 622 drivers/md/dm-region-hash.c void dm_rh_recovery_prepare(struct dm_region_hash *rh) rh 625 drivers/md/dm-region-hash.c atomic_inc(&rh->recovery_in_flight); rh 627 drivers/md/dm-region-hash.c while (!down_trylock(&rh->recovery_count)) { rh 628 drivers/md/dm-region-hash.c atomic_inc(&rh->recovery_in_flight); rh 629 drivers/md/dm-region-hash.c if (__rh_recovery_prepare(rh) <= 0) { rh 630 drivers/md/dm-region-hash.c atomic_dec(&rh->recovery_in_flight); rh 631 drivers/md/dm-region-hash.c up(&rh->recovery_count); rh 637 drivers/md/dm-region-hash.c if (atomic_dec_and_test(&rh->recovery_in_flight)) rh 638 drivers/md/dm-region-hash.c rh->wakeup_all_recovery_waiters(rh->context); rh 645 drivers/md/dm-region-hash.c struct dm_region *dm_rh_recovery_start(struct dm_region_hash *rh) rh 649 drivers/md/dm-region-hash.c spin_lock_irq(&rh->region_lock); rh 650 drivers/md/dm-region-hash.c if (!list_empty(&rh->quiesced_regions)) { rh 651 drivers/md/dm-region-hash.c reg = list_entry(rh->quiesced_regions.next, rh 655 drivers/md/dm-region-hash.c spin_unlock_irq(&rh->region_lock); rh 663 drivers/md/dm-region-hash.c struct dm_region_hash *rh = reg->rh; rh 665 drivers/md/dm-region-hash.c spin_lock_irq(&rh->region_lock); rh 667 drivers/md/dm-region-hash.c list_add(®->list, ®->rh->recovered_regions); rh 669 drivers/md/dm-region-hash.c list_add(®->list, ®->rh->failed_recovered_regions); rh 671 drivers/md/dm-region-hash.c spin_unlock_irq(&rh->region_lock); rh 673 drivers/md/dm-region-hash.c rh->wakeup_workers(rh->context); rh 678 drivers/md/dm-region-hash.c int dm_rh_recovery_in_flight(struct dm_region_hash *rh) rh 680 drivers/md/dm-region-hash.c return atomic_read(&rh->recovery_in_flight); rh 684 drivers/md/dm-region-hash.c int dm_rh_flush(struct dm_region_hash *rh) rh 686 drivers/md/dm-region-hash.c return rh->log->type->flush(rh->log); rh 690 drivers/md/dm-region-hash.c void dm_rh_delay(struct dm_region_hash *rh, struct bio *bio) rh 694 drivers/md/dm-region-hash.c read_lock(&rh->hash_lock); rh 695 drivers/md/dm-region-hash.c reg = __rh_find(rh, dm_rh_bio_to_region(rh, bio)); rh 697 drivers/md/dm-region-hash.c read_unlock(&rh->hash_lock); rh 701 drivers/md/dm-region-hash.c void dm_rh_stop_recovery(struct dm_region_hash *rh) rh 706 drivers/md/dm-region-hash.c for (i = 0; i < rh->max_recovery; i++) rh 707 drivers/md/dm-region-hash.c down(&rh->recovery_count); rh 711 drivers/md/dm-region-hash.c void dm_rh_start_recovery(struct dm_region_hash *rh) rh 715 drivers/md/dm-region-hash.c for (i = 0; i < rh->max_recovery; i++) rh 716 drivers/md/dm-region-hash.c up(&rh->recovery_count); rh 718 drivers/md/dm-region-hash.c rh->wakeup_workers(rh->context); rh 156 drivers/media/usb/uvc/uvc_v4l2.c u16 rw, rh; rh 191 drivers/media/usb/uvc/uvc_v4l2.c rh = fmt->fmt.pix.height; rh 198 drivers/media/usb/uvc/uvc_v4l2.c d = min(w, rw) * min(h, rh); rh 199 drivers/media/usb/uvc/uvc_v4l2.c d = w*h + rw*rh - 2*d; rh 662 drivers/net/ethernet/broadcom/b44.c struct rx_header *rh; rh 706 drivers/net/ethernet/broadcom/b44.c rh = (struct rx_header *) skb->data; rh 708 drivers/net/ethernet/broadcom/b44.c rh->len = 0; rh 709 drivers/net/ethernet/broadcom/b44.c rh->flags = 0; rh 737 drivers/net/ethernet/broadcom/b44.c struct rx_header *rh; rh 748 drivers/net/ethernet/broadcom/b44.c rh = (struct rx_header *) src_map->skb->data; rh 749 drivers/net/ethernet/broadcom/b44.c rh->len = 0; rh 750 drivers/net/ethernet/broadcom/b44.c rh->flags = 0; rh 793 drivers/net/ethernet/broadcom/b44.c struct rx_header *rh; rh 799 drivers/net/ethernet/broadcom/b44.c rh = (struct rx_header *) skb->data; rh 800 drivers/net/ethernet/broadcom/b44.c len = le16_to_cpu(rh->len); rh 802 drivers/net/ethernet/broadcom/b44.c (rh->flags & cpu_to_le16(RX_FLAG_ERRORS))) { rh 816 drivers/net/ethernet/broadcom/b44.c len = le16_to_cpu(rh->len); rh 570 drivers/net/ethernet/cavium/liquidio/lio_core.c union octeon_rh *rh, rh 625 drivers/net/ethernet/cavium/liquidio/lio_core.c r_dh_off = (rh->r_dh.len - 1) * BYTES_PER_DHLEN_UNIT; rh 628 drivers/net/ethernet/cavium/liquidio/lio_core.c if (rh->r_dh.has_hwtstamp) { rh 649 drivers/net/ethernet/cavium/liquidio/lio_core.c if (rh->r_dh.has_hash) { rh 657 drivers/net/ethernet/cavium/liquidio/lio_core.c skb_pull(skb, rh->r_dh.len * BYTES_PER_DHLEN_UNIT); rh 661 drivers/net/ethernet/cavium/liquidio/lio_core.c (((rh->r_dh.encap_on) && rh 662 drivers/net/ethernet/cavium/liquidio/lio_core.c (rh->r_dh.csum_verified & CNNIC_TUN_CSUM_VERIFIED)) || rh 663 drivers/net/ethernet/cavium/liquidio/lio_core.c (!(rh->r_dh.encap_on) && rh 664 drivers/net/ethernet/cavium/liquidio/lio_core.c ((rh->r_dh.csum_verified & CNNIC_CSUM_VERIFIED) == rh 674 drivers/net/ethernet/cavium/liquidio/lio_core.c if (rh->r_dh.encap_on) { rh 682 drivers/net/ethernet/cavium/liquidio/lio_core.c rh->r_dh.vlan) { rh 683 drivers/net/ethernet/cavium/liquidio/lio_core.c u16 priority = rh->r_dh.priority; rh 684 drivers/net/ethernet/cavium/liquidio/lio_core.c u16 vid = rh->r_dh.vlan; rh 693 drivers/net/ethernet/cavium/liquidio/lio_core.c rh->r_dh.len * BYTES_PER_DHLEN_UNIT; rh 37 drivers/net/ethernet/cavium/liquidio/lio_ethtool.c u64 rh; rh 43 drivers/net/ethernet/cavium/liquidio/lio_ethtool.c u64 rh; rh 102 drivers/net/ethernet/cavium/liquidio/lio_main.c u64 rh; rh 108 drivers/net/ethernet/cavium/liquidio/lio_main.c u64 rh; rh 3316 drivers/net/ethernet/cavium/liquidio/lio_main.c recv_pkt->rh.r_nic_info.gmxport); rh 3320 drivers/net/ethernet/cavium/liquidio/lio_main.c gmxport = recv_pkt->rh.r_nic_info.gmxport; rh 4011 drivers/net/ethernet/cavium/liquidio/lio_main.c notice = recv_pkt->rh.r.ossp; rh 44 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c u64 rh; rh 1892 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c recv_pkt->rh.r_nic_info.gmxport); rh 1896 drivers/net/ethernet/cavium/liquidio/lio_vf_main.c gmxport = recv_pkt->rh.r_nic_info.gmxport; rh 300 drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c union octeon_rh *rh; rh 309 drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c rh = &recv_pkt->rh; rh 310 drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c ifidx = rh->r.ossp; rh 328 drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c skb_pull(skb, rh->r_dh.len * BYTES_PER_DHLEN_UNIT); rh 910 drivers/net/ethernet/cavium/liquidio/liquidio_common.h static inline int opcode_slow_path(union octeon_rh *rh) rh 914 drivers/net/ethernet/cavium/liquidio/liquidio_common.h subcode1 = OPCODE_SUBCODE((rh)->r.opcode, (rh)->r.subcode); rh 1041 drivers/net/ethernet/cavium/liquidio/liquidio_common.h u64 rh; rh 1212 drivers/net/ethernet/cavium/liquidio/octeon_device.c (u32)recv_pkt->rh.r_core_drv_init.app_mode), rh 1214 drivers/net/ethernet/cavium/liquidio/octeon_device.c oct->app_mode = (u32)recv_pkt->rh.r_core_drv_init.app_mode; rh 1215 drivers/net/ethernet/cavium/liquidio/octeon_device.c if (recv_pkt->rh.r_core_drv_init.app_mode == CVM_DRV_NIC_APP) { rh 1217 drivers/net/ethernet/cavium/liquidio/octeon_device.c (u32)recv_pkt->rh.r_core_drv_init.max_nic_ports; rh 1219 drivers/net/ethernet/cavium/liquidio/octeon_device.c (u32)recv_pkt->rh.r_core_drv_init.num_gmx_ports; rh 1228 drivers/net/ethernet/cavium/liquidio/octeon_device.c oct->fw_info.app_cap_flags = recv_pkt->rh.r_core_drv_init.app_cap_flags; rh 1229 drivers/net/ethernet/cavium/liquidio/octeon_device.c oct->fw_info.app_mode = (u32)recv_pkt->rh.r_core_drv_init.app_mode; rh 1230 drivers/net/ethernet/cavium/liquidio/octeon_device.c oct->pfvf_hsword.app_mode = (u32)recv_pkt->rh.r_core_drv_init.app_mode; rh 1232 drivers/net/ethernet/cavium/liquidio/octeon_device.c oct->pfvf_hsword.pkind = recv_pkt->rh.r_core_drv_init.pkind; rh 354 drivers/net/ethernet/cavium/liquidio/octeon_droq.c recv_pkt->rh = info->rh; rh 533 drivers/net/ethernet/cavium/liquidio/octeon_droq.c union octeon_rh *rh, rh 542 drivers/net/ethernet/cavium/liquidio/octeon_droq.c disp_fn = octeon_get_dispatch(oct, (u16)rh->r.opcode, rh 543 drivers/net/ethernet/cavium/liquidio/octeon_droq.c (u16)rh->r.subcode); rh 551 drivers/net/ethernet/cavium/liquidio/octeon_droq.c rinfo->recv_pkt->rh = *rh; rh 559 drivers/net/ethernet/cavium/liquidio/octeon_droq.c (unsigned int)rh->r.opcode, rh 560 drivers/net/ethernet/cavium/liquidio/octeon_droq.c (unsigned int)rh->r.subcode); rh 602 drivers/net/ethernet/cavium/liquidio/octeon_droq.c union octeon_rh *rh; rh 627 drivers/net/ethernet/cavium/liquidio/octeon_droq.c rh = &info->rh; rh 630 drivers/net/ethernet/cavium/liquidio/octeon_droq.c rh->r_dh.len += (ROUNDUP8(OCT_DROQ_INFO_SIZE) / sizeof(u64)); rh 632 drivers/net/ethernet/cavium/liquidio/octeon_droq.c if (opcode_slow_path(rh)) { rh 635 drivers/net/ethernet/cavium/liquidio/octeon_droq.c buf_cnt = octeon_droq_dispatch_pkt(oct, droq, rh, info); rh 695 drivers/net/ethernet/cavium/liquidio/octeon_droq.c rh, &droq->napi, rh 769 drivers/net/ethernet/cavium/liquidio/octeon_droq.c (u16)rdisp->rinfo->recv_pkt->rh.r.opcode, rh 770 drivers/net/ethernet/cavium/liquidio/octeon_droq.c (u16)rdisp->rinfo->recv_pkt->rh.r.subcode)); rh 819 drivers/net/ethernet/cavium/liquidio/octeon_droq.c (u16)rdisp->rinfo->recv_pkt->rh.r.opcode, rh 820 drivers/net/ethernet/cavium/liquidio/octeon_droq.c (u16)rdisp->rinfo->recv_pkt->rh.r.subcode)); rh 58 drivers/net/ethernet/cavium/liquidio/octeon_droq.h union octeon_rh rh; rh 148 drivers/net/ethernet/cavium/liquidio/octeon_droq.h union octeon_rh rh; rh 39 drivers/net/ethernet/cavium/liquidio/octeon_network.h u64 rh; rh 69 drivers/net/ethernet/cavium/liquidio/octeon_network.h u64 rh; rh 75 drivers/net/ethernet/cavium/liquidio/octeon_network.h u64 rh; rh 86 drivers/net/ethernet/cavium/liquidio/octeon_network.h u64 rh; rh 47 drivers/net/ethernet/cisco/enic/vnic_dev.c struct vnic_resource_header __iomem *rh; rh 60 drivers/net/ethernet/cisco/enic/vnic_dev.c rh = bar->vaddr; rh 62 drivers/net/ethernet/cisco/enic/vnic_dev.c if (!rh) { rh 68 drivers/net/ethernet/cisco/enic/vnic_dev.c if ((ioread32(&rh->magic) != VNIC_RES_MAGIC) || rh 69 drivers/net/ethernet/cisco/enic/vnic_dev.c (ioread32(&rh->version) != VNIC_RES_VERSION)) { rh 75 drivers/net/ethernet/cisco/enic/vnic_dev.c ioread32(&rh->magic), ioread32(&rh->version)); rh 83 drivers/net/ethernet/cisco/enic/vnic_dev.c r = (struct vnic_resource __iomem *)(rh + 1); rh 293 drivers/net/ethernet/microchip/enc28j60.c int rl, rh; rh 297 drivers/net/ethernet/microchip/enc28j60.c rh = spi_read_op(priv, ENC28J60_READ_CTRL_REG, address + 1); rh 299 drivers/net/ethernet/microchip/enc28j60.c return (rh << 8) | rl; rh 3408 drivers/net/ethernet/sun/niu.c struct rx_pkt_hdr1 *rh; rh 3470 drivers/net/ethernet/sun/niu.c len += sizeof(*rh); rh 3471 drivers/net/ethernet/sun/niu.c len = min_t(int, len, sizeof(*rh) + VLAN_ETH_HLEN); rh 3474 drivers/net/ethernet/sun/niu.c rh = (struct rx_pkt_hdr1 *) skb->data; rh 3477 drivers/net/ethernet/sun/niu.c ((u32)rh->hashval2_0 << 24 | rh 3478 drivers/net/ethernet/sun/niu.c (u32)rh->hashval2_1 << 16 | rh 3479 drivers/net/ethernet/sun/niu.c (u32)rh->hashval1_1 << 8 | rh 3480 drivers/net/ethernet/sun/niu.c (u32)rh->hashval1_2 << 0), rh 3482 drivers/net/ethernet/sun/niu.c skb_pull(skb, sizeof(*rh)); rh 90 drivers/scsi/fnic/vnic_dev.c struct vnic_resource_header __iomem *rh; rh 99 drivers/scsi/fnic/vnic_dev.c rh = bar->vaddr; rh 100 drivers/scsi/fnic/vnic_dev.c if (!rh) { rh 105 drivers/scsi/fnic/vnic_dev.c if (ioread32(&rh->magic) != VNIC_RES_MAGIC || rh 106 drivers/scsi/fnic/vnic_dev.c ioread32(&rh->version) != VNIC_RES_VERSION) { rh 110 drivers/scsi/fnic/vnic_dev.c ioread32(&rh->magic), ioread32(&rh->version)); rh 114 drivers/scsi/fnic/vnic_dev.c r = (struct vnic_resource __iomem *)(rh + 1); rh 3026 drivers/scsi/lpfc/lpfc_ct.c struct lpfc_fdmi_reg_hba *rh; rh 3083 drivers/scsi/lpfc/lpfc_ct.c rh = (struct lpfc_fdmi_reg_hba *)&CtReq->un.PortID; rh 3085 drivers/scsi/lpfc/lpfc_ct.c memcpy(&rh->hi.PortName, &phba->pport->fc_sparam.portName, rh 3091 drivers/scsi/lpfc/lpfc_ct.c rh->rpl.EntryCnt = cpu_to_be32(1); rh 3092 drivers/scsi/lpfc/lpfc_ct.c memcpy(&rh->rpl.pe, &phba->pport->fc_sparam.portName, rh 3101 drivers/scsi/lpfc/lpfc_ct.c ab = (struct lpfc_fdmi_attr_block *)((uint8_t *)rh + size); rh 3116 drivers/scsi/lpfc/lpfc_ct.c ((uint8_t *)rh + size)); rh 3135 drivers/scsi/lpfc/lpfc_ct.c rh = (struct lpfc_fdmi_reg_hba *)pab; rh 3137 drivers/scsi/lpfc/lpfc_ct.c memcpy(&rh->hi.PortName, rh 85 drivers/scsi/snic/vnic_dev.c struct vnic_resource_header __iomem *rh; rh 98 drivers/scsi/snic/vnic_dev.c rh = bar->vaddr; rh 99 drivers/scsi/snic/vnic_dev.c if (!rh) { rh 105 drivers/scsi/snic/vnic_dev.c if (ioread32(&rh->magic) != VNIC_RES_MAGIC || rh 106 drivers/scsi/snic/vnic_dev.c ioread32(&rh->version) != VNIC_RES_VERSION) { rh 109 drivers/scsi/snic/vnic_dev.c ioread32(&rh->magic), ioread32(&rh->version)); rh 114 drivers/scsi/snic/vnic_dev.c r = (struct vnic_resource __iomem *)(rh + 1); rh 137 drivers/soc/qcom/cmd-db.c const struct rsc_hdr **rh) rh 161 drivers/soc/qcom/cmd-db.c if (rh) rh 162 drivers/soc/qcom/cmd-db.c *rh = rsc_hdr; rh 539 drivers/usb/host/ohci-hub.c u32 rh = roothub_a (ohci); rh 543 drivers/usb/host/ohci-hub.c desc->bPwrOn2PwrGood = (rh & RH_A_POTPGT) >> 24; rh 551 drivers/usb/host/ohci-hub.c if (rh & RH_A_NPS) /* no power switching? */ rh 553 drivers/usb/host/ohci-hub.c if (rh & RH_A_PSM) /* per-port power switching? */ rh 555 drivers/usb/host/ohci-hub.c if (rh & RH_A_NOCP) /* no overcurrent reporting? */ rh 557 drivers/usb/host/ohci-hub.c else if (rh & RH_A_OCPM) /* per-port overcurrent reporting? */ rh 562 drivers/usb/host/ohci-hub.c rh = roothub_b (ohci); rh 565 drivers/usb/host/ohci-hub.c desc->u.hs.DeviceRemovable[0] = rh & RH_B_DR; rh 567 drivers/usb/host/ohci-hub.c desc->u.hs.DeviceRemovable[1] = (rh & RH_B_DR) >> 8; rh 256 drivers/usb/host/ohci-omap.c u32 rh = roothub_a (ohci); rh 259 drivers/usb/host/ohci-omap.c rh &= ~RH_A_NPS; rh 265 drivers/usb/host/ohci-omap.c rh &= ~RH_A_NOCP; rh 275 drivers/usb/host/ohci-omap.c ohci_writel(ohci, rh, &ohci->regs->roothub.a); rh 998 drivers/usb/host/r8a66597-hcd.c struct r8a66597_root_hub *rh = &r8a66597->root_hub[port]; rh 1000 drivers/usb/host/r8a66597-hcd.c rh->old_syssts = r8a66597_read(r8a66597, get_syssts_reg(port)) & LNST; rh 1001 drivers/usb/host/r8a66597-hcd.c rh->scount = R8A66597_MAX_SAMPLING; rh 1003 drivers/usb/host/r8a66597-hcd.c rh->port |= USB_PORT_STAT_CONNECTION; rh 1005 drivers/usb/host/r8a66597-hcd.c rh->port &= ~USB_PORT_STAT_CONNECTION; rh 1006 drivers/usb/host/r8a66597-hcd.c rh->port |= USB_PORT_STAT_C_CONNECTION << 16; rh 1042 drivers/usb/host/r8a66597-hcd.c struct r8a66597_root_hub *rh = &r8a66597->root_hub[port]; rh 1044 drivers/usb/host/r8a66597-hcd.c rh->port &= ~(USB_PORT_STAT_HIGH_SPEED | USB_PORT_STAT_LOW_SPEED); rh 1046 drivers/usb/host/r8a66597-hcd.c rh->port |= USB_PORT_STAT_HIGH_SPEED; rh 1048 drivers/usb/host/r8a66597-hcd.c rh->port |= USB_PORT_STAT_LOW_SPEED; rh 1050 drivers/usb/host/r8a66597-hcd.c rh->port &= ~USB_PORT_STAT_RESET; rh 1051 drivers/usb/host/r8a66597-hcd.c rh->port |= USB_PORT_STAT_ENABLE; rh 1688 drivers/usb/host/r8a66597-hcd.c struct r8a66597_root_hub *rh = &r8a66597->root_hub[port]; rh 1690 drivers/usb/host/r8a66597-hcd.c if (rh->port & USB_PORT_STAT_RESET) { rh 1702 drivers/usb/host/r8a66597-hcd.c if (!(rh->port & USB_PORT_STAT_CONNECTION)) { rh 1707 drivers/usb/host/r8a66597-hcd.c if (rh->scount > 0) { rh 1709 drivers/usb/host/r8a66597-hcd.c if (tmp == rh->old_syssts) { rh 1710 drivers/usb/host/r8a66597-hcd.c rh->scount--; rh 1711 drivers/usb/host/r8a66597-hcd.c if (rh->scount == 0) rh 1716 drivers/usb/host/r8a66597-hcd.c rh->scount = R8A66597_MAX_SAMPLING; rh 1717 drivers/usb/host/r8a66597-hcd.c rh->old_syssts = tmp; rh 2148 drivers/usb/host/r8a66597-hcd.c struct r8a66597_root_hub *rh = &r8a66597->root_hub[port]; rh 2173 drivers/usb/host/r8a66597-hcd.c rh->port &= ~USB_PORT_STAT_POWER; rh 2189 drivers/usb/host/r8a66597-hcd.c rh->port &= ~(1 << wValue); rh 2201 drivers/usb/host/r8a66597-hcd.c *(__le32 *)buf = cpu_to_le32(rh->port); rh 2214 drivers/usb/host/r8a66597-hcd.c rh->port |= USB_PORT_STAT_POWER; rh 2217 drivers/usb/host/r8a66597-hcd.c struct r8a66597_device *dev = rh->dev; rh 2219 drivers/usb/host/r8a66597-hcd.c rh->port |= USB_PORT_STAT_RESET; rh 2233 drivers/usb/host/r8a66597-hcd.c rh->port |= 1 << wValue; rh 2254 drivers/usb/host/r8a66597-hcd.c struct r8a66597_root_hub *rh = &r8a66597->root_hub[port]; rh 2257 drivers/usb/host/r8a66597-hcd.c if (!(rh->port & USB_PORT_STAT_ENABLE)) rh 2260 drivers/usb/host/r8a66597-hcd.c dev_dbg(&rh->dev->udev->dev, "suspend port = %d\n", port); rh 2262 drivers/usb/host/r8a66597-hcd.c rh->port |= USB_PORT_STAT_SUSPEND; rh 2264 drivers/usb/host/r8a66597-hcd.c if (rh->dev->udev->do_remote_wakeup) { rh 2285 drivers/usb/host/r8a66597-hcd.c struct r8a66597_root_hub *rh = &r8a66597->root_hub[port]; rh 2288 drivers/usb/host/r8a66597-hcd.c if (!(rh->port & USB_PORT_STAT_SUSPEND)) rh 2291 drivers/usb/host/r8a66597-hcd.c dev_dbg(&rh->dev->udev->dev, "resume port = %d\n", port); rh 2292 drivers/usb/host/r8a66597-hcd.c rh->port &= ~USB_PORT_STAT_SUSPEND; rh 2293 drivers/usb/host/r8a66597-hcd.c rh->port |= USB_PORT_STAT_C_SUSPEND << 16; rh 2352 drivers/usb/host/r8a66597-hcd.c struct r8a66597_root_hub *rh = &r8a66597->root_hub[port]; rh 2354 drivers/usb/host/r8a66597-hcd.c rh->port = 0x00000000; rh 43 include/linux/dm-region-hash.h void dm_region_hash_destroy(struct dm_region_hash *rh); rh 45 include/linux/dm-region-hash.h struct dm_dirty_log *dm_rh_dirty_log(struct dm_region_hash *rh); rh 50 include/linux/dm-region-hash.h region_t dm_rh_bio_to_region(struct dm_region_hash *rh, struct bio *bio); rh 51 include/linux/dm-region-hash.h sector_t dm_rh_region_to_sector(struct dm_region_hash *rh, region_t region); rh 57 include/linux/dm-region-hash.h sector_t dm_rh_get_region_size(struct dm_region_hash *rh); rh 64 include/linux/dm-region-hash.h int dm_rh_get_state(struct dm_region_hash *rh, region_t region, int may_block); rh 65 include/linux/dm-region-hash.h void dm_rh_set_state(struct dm_region_hash *rh, region_t region, rh 69 include/linux/dm-region-hash.h void dm_rh_update_states(struct dm_region_hash *rh, int errors_handled); rh 72 include/linux/dm-region-hash.h int dm_rh_flush(struct dm_region_hash *rh); rh 75 include/linux/dm-region-hash.h void dm_rh_inc_pending(struct dm_region_hash *rh, struct bio_list *bios); rh 76 include/linux/dm-region-hash.h void dm_rh_dec(struct dm_region_hash *rh, region_t region); rh 79 include/linux/dm-region-hash.h void dm_rh_delay(struct dm_region_hash *rh, struct bio *bio); rh 81 include/linux/dm-region-hash.h void dm_rh_mark_nosync(struct dm_region_hash *rh, struct bio *bio); rh 88 include/linux/dm-region-hash.h void dm_rh_recovery_prepare(struct dm_region_hash *rh); rh 91 include/linux/dm-region-hash.h struct dm_region *dm_rh_recovery_start(struct dm_region_hash *rh); rh 97 include/linux/dm-region-hash.h int dm_rh_recovery_in_flight(struct dm_region_hash *rh); rh 100 include/linux/dm-region-hash.h void dm_rh_start_recovery(struct dm_region_hash *rh); rh 101 include/linux/dm-region-hash.h void dm_rh_stop_recovery(struct dm_region_hash *rh); rh 28 include/linux/fsl/bestcomm/sram.h rh_info_t *rh; rh 221 include/linux/math64.h } rl, rm, rn, rh, a0, b0; rh 230 include/linux/math64.h rh.ll = mul_u32_u32(a0.l.high, b0.l.high); rh 238 include/linux/math64.h rh.l.low = c = (c >> 32) + rm.l.high + rn.l.high + rh.l.low; rh 239 include/linux/math64.h rh.l.high = (c >> 32) + rh.l.high; rh 248 include/linux/math64.h return (rl.ll >> shift) | (rh.ll << (64 - shift)); rh 249 include/linux/math64.h return rh.ll >> (shift & 63); rh 267 include/linux/math64.h } u, rl, rh; rh 271 include/linux/math64.h rh.ll = mul_u32_u32(u.l.high, mul) + rl.l.high; rh 274 include/linux/math64.h rl.l.high = do_div(rh.ll, divisor); rh 279 include/linux/math64.h rl.l.high = rh.l.low; rh 147 include/math-emu/op-2.h #define __FP_FRAC_ADD_2(rh, rl, xh, xl, yh, yl) \ rh 148 include/math-emu/op-2.h (rh = xh + yh + ((rl = xl + yl) < xl)) rh 151 include/math-emu/op-2.h #define __FP_FRAC_SUB_2(rh, rl, xh, xl, yh, yl) \ rh 152 include/math-emu/op-2.h (rh = xh - yh - ((rl = xl - yl) > xl)) rh 1635 kernel/rcu/rcutorture.c struct rcu_head rh; rh 1646 kernel/rcu/rcutorture.c struct fwd_cb_state *fcsp = container_of(rhp, struct fwd_cb_state, rh); rh 1652 kernel/rcu/rcutorture.c cur_ops->call(&fcsp->rh, rcu_torture_fwd_prog_cb); rh 1657 kernel/rcu/rcutorture.c struct rcu_head rh; rh 1707 kernel/rcu/rcutorture.c struct rcu_fwd_cb *rfcp = container_of(rhp, struct rcu_fwd_cb, rh); rh 1781 kernel/rcu/rcutorture.c init_rcu_head_on_stack(&fcs.rh); rh 1790 kernel/rcu/rcutorture.c cur_ops->call(&fcs.rh, rcu_torture_fwd_prog_cb); rh 1826 kernel/rcu/rcutorture.c destroy_rcu_head_on_stack(&fcs.rh); rh 1892 kernel/rcu/rcutorture.c cur_ops->call(&rfcp->rh, rcu_torture_fwd_cb_cr); rh 135 net/dccp/ccids/lib/loss_interval.c int tfrc_lh_interval_add(struct tfrc_loss_hist *lh, struct tfrc_rx_hist *rh, rh 140 net/dccp/ccids/lib/loss_interval.c if (cur != NULL && !tfrc_lh_is_new_loss(cur, tfrc_rx_hist_loss_prev(rh))) rh 149 net/dccp/ccids/lib/loss_interval.c new->li_seqno = tfrc_rx_hist_loss_prev(rh)->tfrchrx_seqno; rh 150 net/dccp/ccids/lib/loss_interval.c new->li_ccval = tfrc_rx_hist_loss_prev(rh)->tfrchrx_ccval; rh 158 net/dccp/ccids/lib/loss_interval.c tfrc_rx_hist_last_rcv(rh)->tfrchrx_seqno) + 1; rh 222 net/ipv6/exthdrs_core.c struct ipv6_rt_hdr _rh, *rh; rh 224 net/ipv6/exthdrs_core.c rh = skb_header_pointer(skb, start, sizeof(_rh), rh 226 net/ipv6/exthdrs_core.c if (!rh) rh 230 net/ipv6/exthdrs_core.c rh->segments_left == 0) rh 39 net/ipv6/netfilter/ip6t_rt.c const struct ipv6_rt_hdr *rh; rh 56 net/ipv6/netfilter/ip6t_rt.c rh = skb_header_pointer(skb, ptr, sizeof(_route), &_route); rh 57 net/ipv6/netfilter/ip6t_rt.c if (rh == NULL) { rh 62 net/ipv6/netfilter/ip6t_rt.c hdrlen = ipv6_optlen(rh); rh 68 net/ipv6/netfilter/ip6t_rt.c pr_debug("IPv6 RT LEN %u %u ", hdrlen, rh->hdrlen); rh 69 net/ipv6/netfilter/ip6t_rt.c pr_debug("TYPE %04X ", rh->type); rh 70 net/ipv6/netfilter/ip6t_rt.c pr_debug("SGS_LEFT %u %02X\n", rh->segments_left, rh->segments_left); rh 74 net/ipv6/netfilter/ip6t_rt.c rh->segments_left, rh 77 net/ipv6/netfilter/ip6t_rt.c rtinfo->rt_type, rh->type, rh 79 net/ipv6/netfilter/ip6t_rt.c ((rtinfo->rt_type == rh->type) ^ rh 88 net/ipv6/netfilter/ip6t_rt.c ((const struct rt0_hdr *)rh)->reserved, rh 90 net/ipv6/netfilter/ip6t_rt.c (((const struct rt0_hdr *)rh)->reserved))); rh 92 net/ipv6/netfilter/ip6t_rt.c ret = (rh != NULL) && rh 94 net/ipv6/netfilter/ip6t_rt.c rh->segments_left, rh 100 net/ipv6/netfilter/ip6t_rt.c ((rtinfo->rt_type == rh->type) ^ rh 996 sound/soc/codecs/rt1305.c unsigned int rh, rl, rhl, r0ohm; rh 1060 sound/soc/codecs/rt1305.c regmap_read(rt1305->regmap, RT1305_PR_BASE + 0x55, &rh); rh 1062 sound/soc/codecs/rt1305.c rhl = (rh << 16) | rl; rh 1065 sound/soc/codecs/rt1305.c pr_debug("Left_rhl = 0x%x rh=0x%x rl=0x%x\n", rhl, rh, rl); rh 1077 sound/soc/codecs/rt1305.c regmap_read(rt1305->regmap, RT1305_PR_BASE + 0x55, &rh); rh 1079 sound/soc/codecs/rt1305.c rhl = (rh << 16) | rl; rh 1082 sound/soc/codecs/rt1305.c pr_debug("Right_rhl = 0x%x rh=0x%x rl=0x%x\n", rhl, rh, rl);