Searched refs:rh (Results 1 - 57 of 57) sorted by relevance

/linux-4.4.14/drivers/md/
H A Ddm-region-hash.c102 struct dm_region_hash *rh; /* FIXME: can we get rid of this ? */ member in struct:dm_region
116 static region_t dm_rh_sector_to_region(struct dm_region_hash *rh, sector_t sector) dm_rh_sector_to_region() argument
118 return sector >> rh->region_shift; dm_rh_sector_to_region()
121 sector_t dm_rh_region_to_sector(struct dm_region_hash *rh, region_t region) dm_rh_region_to_sector() argument
123 return region << rh->region_shift; dm_rh_region_to_sector()
127 region_t dm_rh_bio_to_region(struct dm_region_hash *rh, struct bio *bio) dm_rh_bio_to_region() argument
129 return dm_rh_sector_to_region(rh, bio->bi_iter.bi_sector - dm_rh_bio_to_region()
130 rh->target_begin); dm_rh_bio_to_region()
136 return reg->rh->context; dm_rh_region_context()
146 sector_t dm_rh_get_region_size(struct dm_region_hash *rh) dm_rh_get_region_size() argument
148 return rh->region_size; dm_rh_get_region_size()
169 struct dm_region_hash *rh; dm_region_hash_create() local
182 rh = kmalloc(sizeof(*rh), GFP_KERNEL); dm_region_hash_create()
183 if (!rh) { dm_region_hash_create()
188 rh->context = context; dm_region_hash_create()
189 rh->dispatch_bios = dispatch_bios; dm_region_hash_create()
190 rh->wakeup_workers = wakeup_workers; dm_region_hash_create()
191 rh->wakeup_all_recovery_waiters = wakeup_all_recovery_waiters; dm_region_hash_create()
192 rh->target_begin = target_begin; dm_region_hash_create()
193 rh->max_recovery = max_recovery; dm_region_hash_create()
194 rh->log = log; dm_region_hash_create()
195 rh->region_size = region_size; dm_region_hash_create()
196 rh->region_shift = __ffs(region_size); dm_region_hash_create()
197 rwlock_init(&rh->hash_lock); dm_region_hash_create()
198 rh->mask = nr_buckets - 1; dm_region_hash_create()
199 rh->nr_buckets = nr_buckets; dm_region_hash_create()
201 rh->shift = RH_HASH_SHIFT; dm_region_hash_create()
202 rh->prime = RH_HASH_MULT; dm_region_hash_create()
204 rh->buckets = vmalloc(nr_buckets * sizeof(*rh->buckets)); dm_region_hash_create()
205 if (!rh->buckets) { dm_region_hash_create()
207 kfree(rh); dm_region_hash_create()
212 INIT_LIST_HEAD(rh->buckets + i); dm_region_hash_create()
214 spin_lock_init(&rh->region_lock); dm_region_hash_create()
215 sema_init(&rh->recovery_count, 0); dm_region_hash_create()
216 atomic_set(&rh->recovery_in_flight, 0); dm_region_hash_create()
217 INIT_LIST_HEAD(&rh->clean_regions); dm_region_hash_create()
218 INIT_LIST_HEAD(&rh->quiesced_regions); dm_region_hash_create()
219 INIT_LIST_HEAD(&rh->recovered_regions); dm_region_hash_create()
220 INIT_LIST_HEAD(&rh->failed_recovered_regions); dm_region_hash_create()
221 rh->flush_failure = 0; dm_region_hash_create()
223 rh->region_pool = mempool_create_kmalloc_pool(MIN_REGIONS, dm_region_hash_create()
225 if (!rh->region_pool) { dm_region_hash_create()
226 vfree(rh->buckets); dm_region_hash_create()
227 kfree(rh); dm_region_hash_create()
228 rh = ERR_PTR(-ENOMEM); dm_region_hash_create()
231 return rh; dm_region_hash_create()
235 void dm_region_hash_destroy(struct dm_region_hash *rh) dm_region_hash_destroy() argument
240 BUG_ON(!list_empty(&rh->quiesced_regions)); dm_region_hash_destroy()
241 for (h = 0; h < rh->nr_buckets; h++) { dm_region_hash_destroy()
242 list_for_each_entry_safe(reg, nreg, rh->buckets + h, dm_region_hash_destroy()
245 mempool_free(reg, rh->region_pool); dm_region_hash_destroy()
249 if (rh->log) dm_region_hash_destroy()
250 dm_dirty_log_destroy(rh->log); dm_region_hash_destroy()
252 mempool_destroy(rh->region_pool); dm_region_hash_destroy()
253 vfree(rh->buckets); dm_region_hash_destroy()
254 kfree(rh); dm_region_hash_destroy()
258 struct dm_dirty_log *dm_rh_dirty_log(struct dm_region_hash *rh) dm_rh_dirty_log() argument
260 return rh->log; dm_rh_dirty_log()
264 static unsigned rh_hash(struct dm_region_hash *rh, region_t region) rh_hash() argument
266 return (unsigned) ((region * rh->prime) >> rh->shift) & rh->mask; rh_hash()
269 static struct dm_region *__rh_lookup(struct dm_region_hash *rh, region_t region) __rh_lookup() argument
272 struct list_head *bucket = rh->buckets + rh_hash(rh, region); __rh_lookup()
281 static void __rh_insert(struct dm_region_hash *rh, struct dm_region *reg) __rh_insert() argument
283 list_add(&reg->hash_list, rh->buckets + rh_hash(rh, reg->key)); __rh_insert()
286 static struct dm_region *__rh_alloc(struct dm_region_hash *rh, region_t region) __rh_alloc() argument
290 nreg = mempool_alloc(rh->region_pool, GFP_ATOMIC); __rh_alloc()
294 nreg->state = rh->log->type->in_sync(rh->log, region, 1) ? __rh_alloc()
296 nreg->rh = rh; __rh_alloc()
302 write_lock_irq(&rh->hash_lock); __rh_alloc()
303 reg = __rh_lookup(rh, region); __rh_alloc()
306 mempool_free(nreg, rh->region_pool); __rh_alloc()
308 __rh_insert(rh, nreg); __rh_alloc()
310 spin_lock(&rh->region_lock); __rh_alloc()
311 list_add(&nreg->list, &rh->clean_regions); __rh_alloc()
312 spin_unlock(&rh->region_lock); __rh_alloc()
317 write_unlock_irq(&rh->hash_lock); __rh_alloc()
322 static struct dm_region *__rh_find(struct dm_region_hash *rh, region_t region) __rh_find() argument
326 reg = __rh_lookup(rh, region); __rh_find()
328 read_unlock(&rh->hash_lock); __rh_find()
329 reg = __rh_alloc(rh, region); __rh_find()
330 read_lock(&rh->hash_lock); __rh_find()
336 int dm_rh_get_state(struct dm_region_hash *rh, region_t region, int may_block) dm_rh_get_state() argument
341 read_lock(&rh->hash_lock); dm_rh_get_state()
342 reg = __rh_lookup(rh, region); dm_rh_get_state()
343 read_unlock(&rh->hash_lock); dm_rh_get_state()
352 r = rh->log->type->in_sync(rh->log, region, may_block); dm_rh_get_state()
364 struct dm_region_hash *rh = reg->rh; complete_resync_work() local
366 rh->log->type->set_region_sync(rh->log, reg->key, success); complete_resync_work()
377 rh->dispatch_bios(rh->context, &reg->delayed_bios); complete_resync_work()
378 if (atomic_dec_and_test(&rh->recovery_in_flight)) complete_resync_work()
379 rh->wakeup_all_recovery_waiters(rh->context); complete_resync_work()
380 up(&rh->recovery_count); complete_resync_work()
393 void dm_rh_mark_nosync(struct dm_region_hash *rh, struct bio *bio) dm_rh_mark_nosync() argument
396 struct dm_dirty_log *log = rh->log; dm_rh_mark_nosync()
398 region_t region = dm_rh_bio_to_region(rh, bio); dm_rh_mark_nosync()
402 rh->flush_failure = 1; dm_rh_mark_nosync()
412 read_lock(&rh->hash_lock); dm_rh_mark_nosync()
413 reg = __rh_find(rh, region); dm_rh_mark_nosync()
414 read_unlock(&rh->hash_lock); dm_rh_mark_nosync()
420 spin_lock_irqsave(&rh->region_lock, flags); dm_rh_mark_nosync()
431 spin_unlock_irqrestore(&rh->region_lock, flags); dm_rh_mark_nosync()
438 void dm_rh_update_states(struct dm_region_hash *rh, int errors_handled) dm_rh_update_states() argument
449 write_lock_irq(&rh->hash_lock); dm_rh_update_states()
450 spin_lock(&rh->region_lock); dm_rh_update_states()
451 if (!list_empty(&rh->clean_regions)) { dm_rh_update_states()
452 list_splice_init(&rh->clean_regions, &clean); dm_rh_update_states()
458 if (!list_empty(&rh->recovered_regions)) { dm_rh_update_states()
459 list_splice_init(&rh->recovered_regions, &recovered); dm_rh_update_states()
465 if (!list_empty(&rh->failed_recovered_regions)) { dm_rh_update_states()
466 list_splice_init(&rh->failed_recovered_regions, dm_rh_update_states()
473 spin_unlock(&rh->region_lock); dm_rh_update_states()
474 write_unlock_irq(&rh->hash_lock); dm_rh_update_states()
482 rh->log->type->clear_region(rh->log, reg->key); dm_rh_update_states()
484 mempool_free(reg, rh->region_pool); dm_rh_update_states()
489 mempool_free(reg, rh->region_pool); dm_rh_update_states()
493 rh->log->type->clear_region(rh->log, reg->key); dm_rh_update_states()
494 mempool_free(reg, rh->region_pool); dm_rh_update_states()
497 rh->log->type->flush(rh->log); dm_rh_update_states()
501 static void rh_inc(struct dm_region_hash *rh, region_t region) rh_inc() argument
505 read_lock(&rh->hash_lock); rh_inc()
506 reg = __rh_find(rh, region); rh_inc()
508 spin_lock_irq(&rh->region_lock); rh_inc()
514 spin_unlock_irq(&rh->region_lock); rh_inc()
516 rh->log->type->mark_region(rh->log, reg->key); rh_inc()
518 spin_unlock_irq(&rh->region_lock); rh_inc()
521 read_unlock(&rh->hash_lock); rh_inc()
524 void dm_rh_inc_pending(struct dm_region_hash *rh, struct bio_list *bios) dm_rh_inc_pending() argument
531 rh_inc(rh, dm_rh_bio_to_region(rh, bio)); dm_rh_inc_pending()
536 void dm_rh_dec(struct dm_region_hash *rh, region_t region) dm_rh_dec() argument
542 read_lock(&rh->hash_lock); dm_rh_dec()
543 reg = __rh_lookup(rh, region); dm_rh_dec()
544 read_unlock(&rh->hash_lock); dm_rh_dec()
546 spin_lock_irqsave(&rh->region_lock, flags); dm_rh_dec()
560 if (unlikely(rh->flush_failure)) { dm_rh_dec()
568 list_add_tail(&reg->list, &rh->quiesced_regions); dm_rh_dec()
571 list_add(&reg->list, &rh->clean_regions); dm_rh_dec()
575 spin_unlock_irqrestore(&rh->region_lock, flags); dm_rh_dec()
578 rh->wakeup_workers(rh->context); dm_rh_dec()
585 static int __rh_recovery_prepare(struct dm_region_hash *rh) __rh_recovery_prepare() argument
594 r = rh->log->type->get_resync_work(rh->log, &region); __rh_recovery_prepare()
602 read_lock(&rh->hash_lock); __rh_recovery_prepare()
603 reg = __rh_find(rh, region); __rh_recovery_prepare()
604 read_unlock(&rh->hash_lock); __rh_recovery_prepare()
606 spin_lock_irq(&rh->region_lock); __rh_recovery_prepare()
613 list_move(&reg->list, &rh->quiesced_regions); __rh_recovery_prepare()
615 spin_unlock_irq(&rh->region_lock); __rh_recovery_prepare()
620 void dm_rh_recovery_prepare(struct dm_region_hash *rh) dm_rh_recovery_prepare() argument
623 atomic_inc(&rh->recovery_in_flight); dm_rh_recovery_prepare()
625 while (!down_trylock(&rh->recovery_count)) { dm_rh_recovery_prepare()
626 atomic_inc(&rh->recovery_in_flight); dm_rh_recovery_prepare()
627 if (__rh_recovery_prepare(rh) <= 0) { dm_rh_recovery_prepare()
628 atomic_dec(&rh->recovery_in_flight); dm_rh_recovery_prepare()
629 up(&rh->recovery_count); dm_rh_recovery_prepare()
635 if (atomic_dec_and_test(&rh->recovery_in_flight)) dm_rh_recovery_prepare()
636 rh->wakeup_all_recovery_waiters(rh->context); dm_rh_recovery_prepare()
643 struct dm_region *dm_rh_recovery_start(struct dm_region_hash *rh) dm_rh_recovery_start() argument
647 spin_lock_irq(&rh->region_lock); dm_rh_recovery_start()
648 if (!list_empty(&rh->quiesced_regions)) { dm_rh_recovery_start()
649 reg = list_entry(rh->quiesced_regions.next, dm_rh_recovery_start()
653 spin_unlock_irq(&rh->region_lock); dm_rh_recovery_start()
661 struct dm_region_hash *rh = reg->rh; dm_rh_recovery_end() local
663 spin_lock_irq(&rh->region_lock); dm_rh_recovery_end()
665 list_add(&reg->list, &reg->rh->recovered_regions); dm_rh_recovery_end()
667 list_add(&reg->list, &reg->rh->failed_recovered_regions); dm_rh_recovery_end()
669 spin_unlock_irq(&rh->region_lock); dm_rh_recovery_end()
671 rh->wakeup_workers(rh->context); dm_rh_recovery_end()
676 int dm_rh_recovery_in_flight(struct dm_region_hash *rh) dm_rh_recovery_in_flight() argument
678 return atomic_read(&rh->recovery_in_flight); dm_rh_recovery_in_flight()
682 int dm_rh_flush(struct dm_region_hash *rh) dm_rh_flush() argument
684 return rh->log->type->flush(rh->log); dm_rh_flush()
688 void dm_rh_delay(struct dm_region_hash *rh, struct bio *bio) dm_rh_delay() argument
692 read_lock(&rh->hash_lock); dm_rh_delay()
693 reg = __rh_find(rh, dm_rh_bio_to_region(rh, bio)); dm_rh_delay()
695 read_unlock(&rh->hash_lock); dm_rh_delay()
699 void dm_rh_stop_recovery(struct dm_region_hash *rh) dm_rh_stop_recovery() argument
704 for (i = 0; i < rh->max_recovery; i++) dm_rh_stop_recovery()
705 down(&rh->recovery_count); dm_rh_stop_recovery()
709 void dm_rh_start_recovery(struct dm_region_hash *rh) dm_rh_start_recovery() argument
713 for (i = 0; i < rh->max_recovery; i++) dm_rh_start_recovery()
714 up(&rh->recovery_count); dm_rh_start_recovery()
716 rh->wakeup_workers(rh->context); dm_rh_start_recovery()
H A Ddm-raid1.c63 struct dm_region_hash *rh; member in struct:mirror_set
336 sector_t region_size = dm_rh_get_region_size(ms->rh); recover()
341 from.sector = m->offset + dm_rh_region_to_sector(ms->rh, key); recover()
360 dest->sector = m->offset + dm_rh_region_to_sector(ms->rh, key); recover()
389 struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); do_recovery()
395 dm_rh_recovery_prepare(ms->rh); do_recovery()
400 while ((reg = dm_rh_recovery_start(ms->rh))) { do_recovery()
445 struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); mirror_available()
446 region_t region = dm_rh_bio_to_region(ms->rh, bio); mirror_available()
560 int state = dm_rh_get_state(ms->rh, region, may_block); region_in_sync()
571 region = dm_rh_bio_to_region(ms->rh, bio); do_reads()
689 struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); do_writes()
710 region = dm_rh_bio_to_region(ms->rh, bio); do_writes()
718 state = dm_rh_get_state(ms->rh, region, 1); do_writes()
753 dm_rh_inc_pending(ms->rh, &sync); do_writes()
754 dm_rh_inc_pending(ms->rh, &nosync); do_writes()
761 ms->log_failure = dm_rh_flush(ms->rh) ? 1 : ms->log_failure; do_writes()
776 dm_rh_delay(ms->rh, bio); do_writes()
818 dm_rh_mark_nosync(ms->rh, bio); do_failures()
870 dm_rh_update_states(ms->rh, errors_handled(ms)); do_mirror()
918 ms->rh = dm_region_hash_create(ms, dispatch_bios, wakeup_mirrord, alloc_context()
922 if (IS_ERR(ms->rh)) { alloc_context()
939 dm_region_hash_destroy(ms->rh); free_context()
1118 r = dm_set_target_max_io_len(ti, dm_rh_get_region_size(ms->rh)); mirror_ctr()
1196 struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); mirror_map()
1204 bio_record->write_region = dm_rh_bio_to_region(ms->rh, bio); mirror_map()
1209 r = log->type->in_sync(log, dm_rh_bio_to_region(ms->rh, bio), 0); mirror_map()
1254 dm_rh_dec(ms->rh, bio_record->write_region); mirror_end_io()
1307 struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); mirror_presuspend()
1332 dm_rh_stop_recovery(ms->rh); mirror_presuspend()
1335 !dm_rh_recovery_in_flight(ms->rh)); mirror_presuspend()
1353 struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); mirror_postsuspend()
1363 struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); mirror_resume()
1369 dm_rh_start_recovery(ms->rh); mirror_resume()
1403 struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); mirror_status()
/linux-4.4.14/include/linux/
H A Ddm-region-hash.h43 void dm_region_hash_destroy(struct dm_region_hash *rh);
45 struct dm_dirty_log *dm_rh_dirty_log(struct dm_region_hash *rh);
50 region_t dm_rh_bio_to_region(struct dm_region_hash *rh, struct bio *bio);
51 sector_t dm_rh_region_to_sector(struct dm_region_hash *rh, region_t region);
57 sector_t dm_rh_get_region_size(struct dm_region_hash *rh);
64 int dm_rh_get_state(struct dm_region_hash *rh, region_t region, int may_block);
65 void dm_rh_set_state(struct dm_region_hash *rh, region_t region,
69 void dm_rh_update_states(struct dm_region_hash *rh, int errors_handled);
72 int dm_rh_flush(struct dm_region_hash *rh);
75 void dm_rh_inc_pending(struct dm_region_hash *rh, struct bio_list *bios);
76 void dm_rh_dec(struct dm_region_hash *rh, region_t region);
79 void dm_rh_delay(struct dm_region_hash *rh, struct bio *bio);
81 void dm_rh_mark_nosync(struct dm_region_hash *rh, struct bio *bio);
88 void dm_rh_recovery_prepare(struct dm_region_hash *rh);
91 struct dm_region *dm_rh_recovery_start(struct dm_region_hash *rh);
97 int dm_rh_recovery_in_flight(struct dm_region_hash *rh);
100 void dm_rh_start_recovery(struct dm_region_hash *rh);
101 void dm_rh_stop_recovery(struct dm_region_hash *rh);
H A Dmath64.h183 } rl, rm, rn, rh, a0, b0; mul_u64_u64_shr() local
192 rh.ll = (u64)a0.l.high * b0.l.high; mul_u64_u64_shr()
200 rh.l.low = c = (c >> 32) + rm.l.high + rn.l.high + rh.l.low; mul_u64_u64_shr()
201 rh.l.high = (c >> 32) + rh.l.high; mul_u64_u64_shr()
204 * The 128-bit result of the multiplication is in rl.ll and rh.ll, mul_u64_u64_shr()
210 return (rl.ll >> shift) | (rh.ll << (64 - shift)); mul_u64_u64_shr()
211 return rh.ll >> (shift & 63); mul_u64_u64_shr()
229 } u, rl, rh; mul_u64_u32_div() local
233 rh.ll = (u64)u.l.high * mul + rl.l.high; mul_u64_u32_div()
235 /* Bits 32-63 of the result will be in rh.l.low. */ mul_u64_u32_div()
236 rl.l.high = do_div(rh.ll, divisor); mul_u64_u32_div()
241 rl.l.high = rh.l.low; mul_u64_u32_div()
/linux-4.4.14/drivers/usb/wusbcore/
H A DMakefile14 rh.o \
H A Dwusbhc.h39 * rh Root Hub emulation (part of the HCD glue)
144 * See rh.c for more information.
H A Dwa-hc.h68 * and call into the rh.c:hwahc_rc_port_reset() code to authenticate
H A Ddevconnect.c69 * __wusb_dev_disable() Called by rh.c:wusbhc_rh_clear_port_feat() when
/linux-4.4.14/net/ipv6/netfilter/
H A Dip6t_rt.c42 const struct ipv6_rt_hdr *rh; rt_mt6() local
59 rh = skb_header_pointer(skb, ptr, sizeof(_route), &_route); rt_mt6()
60 if (rh == NULL) { rt_mt6()
65 hdrlen = ipv6_optlen(rh); rt_mt6()
71 pr_debug("IPv6 RT LEN %u %u ", hdrlen, rh->hdrlen); rt_mt6()
72 pr_debug("TYPE %04X ", rh->type); rt_mt6()
73 pr_debug("SGS_LEFT %u %02X\n", rh->segments_left, rh->segments_left); rt_mt6()
77 rh->segments_left, rt_mt6()
80 rtinfo->rt_type, rh->type, rt_mt6()
82 ((rtinfo->rt_type == rh->type) ^ rt_mt6()
91 ((const struct rt0_hdr *)rh)->reserved, rt_mt6()
93 (((const struct rt0_hdr *)rh)->reserved))); rt_mt6()
95 ret = (rh != NULL) && rt_mt6()
97 rh->segments_left, rt_mt6()
103 ((rtinfo->rt_type == rh->type) ^ rt_mt6()
/linux-4.4.14/crypto/
H A Dvmac.c66 #define ADD128(rh, rl, ih, il) \
71 (rh)++; \
72 (rh) += (ih); \
77 #define PMUL64(rh, rl, i1, i2) /* Assumes m doesn't overflow */ \
81 rh = MUL32(_i1>>32, _i2>>32); \
83 ADD128(rh, rl, (m >> 32), (m << 32)); \
86 #define MUL64(rh, rl, i1, i2) \
91 rh = MUL32(_i1>>32, _i2>>32); \
93 ADD128(rh, rl, (m1 >> 32), (m1 << 32)); \
94 ADD128(rh, rl, (m2 >> 32), (m2 << 32)); \
111 #define nh_16(mp, kp, nw, rh, rl) \
114 rh = rl = 0; \
118 ADD128(rh, rl, th, tl); \
122 #define nh_16_2(mp, kp, nw, rh, rl, rh1, rl1) \
125 rh1 = rl1 = rh = rl = 0; \
129 ADD128(rh, rl, th, tl); \
137 #define nh_vmac_nhbytes(mp, kp, nw, rh, rl) \
140 rh = rl = 0; \
144 ADD128(rh, rl, th, tl); \
147 ADD128(rh, rl, th, tl); \
150 ADD128(rh, rl, th, tl); \
153 ADD128(rh, rl, th, tl); \
157 #define nh_vmac_nhbytes_2(mp, kp, nw, rh, rl, rh1, rl1) \
160 rh1 = rl1 = rh = rl = 0; \
164 ADD128(rh, rl, th, tl); \
170 ADD128(rh, rl, th, tl); \
176 ADD128(rh, rl, th, tl); \
182 ADD128(rh, rl, th, tl); \
216 #define nh_16(mp, kp, nw, rh, rl) \
220 rh = rl = t = 0; \
226 ADD128(rh, rl, MUL32(t1 >> 32, t2 >> 32), \
228 rh += (u64)(u32)(m1 >> 32) \
232 ADD128(rh, rl, (t >> 32), (t << 32)); \
303 #define nh_16_2(mp, kp, nw, rh, rl, rh2, rl2) \
305 nh_16(mp, kp, nw, rh, rl); \
310 #define nh_vmac_nhbytes(mp, kp, nw, rh, rl) \
311 nh_16(mp, kp, nw, rh, rl)
314 #define nh_vmac_nhbytes_2(mp, kp, nw, rh, rl, rh2, rl2) \
316 nh_vmac_nhbytes(mp, kp, nw, rh, rl); \
330 u64 rh, rl, t, z = 0; l3hash() local
355 MUL64(rh, rl, p1, p2); l3hash()
356 t = rh >> 56; l3hash()
357 ADD128(t, rl, z, rh); l3hash()
358 rh <<= 8; l3hash()
359 ADD128(t, rl, z, rh); l3hash()
371 u64 rh, rl, *mptr; vhash_update() local
391 nh_vmac_nhbytes(mptr, kptr, VMAC_NHBYTES/8, rh, rl); vhash_update()
392 rh &= m62; vhash_update()
393 ADD128(ch, cl, rh, rl); vhash_update()
399 nh_vmac_nhbytes(mptr, kptr, VMAC_NHBYTES/8, rh, rl); vhash_update()
400 rh &= m62; vhash_update()
401 poly_step(ch, cl, pkh, pkl, rh, rl); vhash_update()
412 u64 rh, rl, *mptr; vhash() local
444 nh_vmac_nhbytes(mptr, kptr, VMAC_NHBYTES/8, rh, rl); vhash()
445 rh &= m62; vhash()
446 poly_step(ch, cl, pkh, pkl, rh, rl); vhash()
450 nh_16(mptr, kptr, 2*((remaining+15)/16), rh, rl); vhash()
451 rh &= m62; vhash()
452 poly_step(ch, cl, pkh, pkl, rh, rl); vhash()
/linux-4.4.14/drivers/dma/bestcomm/
H A Dsram.c97 bcom_sram->rh = rh_create(4); bcom_sram_init()
110 rh_attach_region(bcom_sram->rh, 0, bcom_sram->size); bcom_sram_init()
115 rh_attach_region(bcom_sram->rh, zbase - bcom_sram->base_phys, regaddr_p[1]); bcom_sram_init()
140 rh_destroy(bcom_sram->rh); bcom_sram_cleanup()
154 offset = rh_alloc_align(bcom_sram->rh, size, align, NULL); bcom_sram_alloc()
175 rh_free(bcom_sram->rh, offset); bcom_sram_free()
/linux-4.4.14/arch/powerpc/sysdev/
H A Dfsl_85xx_cache_sram.c60 offset = rh_alloc_align(cache_sram->rh, size, align, NULL); mpc85xx_cache_sram_alloc()
78 rh_free(cache_sram->rh, ptr - cache_sram->base_virt); mpc85xx_cache_sram_free()
119 cache_sram->rh = rh_create(sizeof(unsigned int)); instantiate_cache_sram()
120 if (IS_ERR(cache_sram->rh)) { instantiate_cache_sram()
123 ret = PTR_ERR(cache_sram->rh); instantiate_cache_sram()
127 rh_attach_region(cache_sram->rh, 0, cache_sram->size); instantiate_cache_sram()
150 rh_detach_region(cache_sram->rh, 0, cache_sram->size); remove_cache_sram()
151 rh_destroy(cache_sram->rh); remove_cache_sram()
H A Dppc4xx_ocm.c53 rh_info_t *rh; member in struct:ocm_region
97 rh_free(ocm_reg->rh, offset); ocm_free_region()
216 ocm->nc.rh = rh_create(ocm->alignment); ocm_init_node()
217 rh_attach_region(ocm->nc.rh, 0, ocm->nc.memtotal); ocm_init_node()
221 ocm->c.rh = rh_create(ocm->alignment); ocm_init_node()
222 rh_attach_region(ocm->c.rh, 0, ocm->c.memtotal); ocm_init_node()
337 offset = rh_alloc_align(ocm_reg->rh, size, align, NULL); ppc4xx_ocm_alloc()
345 rh_free(ocm_reg->rh, offset); ppc4xx_ocm_alloc()
/linux-4.4.14/net/dccp/ccids/lib/
H A Dloss_interval.c133 * @rh: Receive history containing a fresh loss event
139 int tfrc_lh_interval_add(struct tfrc_loss_hist *lh, struct tfrc_rx_hist *rh, tfrc_lh_interval_add() argument
144 if (cur != NULL && !tfrc_lh_is_new_loss(cur, tfrc_rx_hist_loss_prev(rh))) tfrc_lh_interval_add()
153 new->li_seqno = tfrc_rx_hist_loss_prev(rh)->tfrchrx_seqno; tfrc_lh_interval_add()
154 new->li_ccval = tfrc_rx_hist_loss_prev(rh)->tfrchrx_ccval; tfrc_lh_interval_add()
162 tfrc_rx_hist_last_rcv(rh)->tfrchrx_seqno) + 1; tfrc_lh_interval_add()
/linux-4.4.14/arch/arm/vfp/
H A Dvfp.h76 u64 rh, rma, rmb, rl; mul64to128() local
89 rh = (u64)nh * mh; mul64to128()
90 rh += ((u64)(rma < rmb) << 32) + (rma >> 32); mul64to128()
94 rh += (rl < rma); mul64to128()
97 *resh = rh; mul64to128()
108 u64 rh, rl; vfp_hi64multiply64() local
109 mul64to128(&rh, &rl, n, m); vfp_hi64multiply64()
110 return rh | (rl != 0); vfp_hi64multiply64()
/linux-4.4.14/drivers/net/ethernet/cavium/liquidio/
H A Docteon_droq.c369 recv_pkt->rh = info->rh; octeon_create_recv_info()
517 union octeon_rh *rh, octeon_droq_dispatch_pkt()
526 disp_fn = octeon_get_dispatch(oct, (u16)rh->r.opcode, octeon_droq_dispatch_pkt()
527 (u16)rh->r.subcode); octeon_droq_dispatch_pkt()
535 rinfo->recv_pkt->rh = *rh; octeon_droq_dispatch_pkt()
581 union octeon_rh *rh; octeon_droq_fast_process_packets() local
605 rh = &info->rh; octeon_droq_fast_process_packets()
609 if (OPCODE_SLOW_PATH(rh)) { octeon_droq_fast_process_packets()
612 buf_cnt = octeon_droq_dispatch_pkt(oct, droq, rh, info); octeon_droq_fast_process_packets()
674 rh, &droq->napi); octeon_droq_fast_process_packets()
741 (u16)rdisp->rinfo->recv_pkt->rh.r.opcode, octeon_droq_process_packets()
742 (u16)rdisp->rinfo->recv_pkt->rh.r.subcode)); octeon_droq_process_packets()
798 (u16)rdisp->rinfo->recv_pkt->rh.r.opcode, octeon_droq_process_poll_pkts()
799 (u16)rdisp->rinfo->recv_pkt->rh.r.subcode)); octeon_droq_process_poll_pkts()
515 octeon_droq_dispatch_pkt(struct octeon_device *oct, struct octeon_droq *droq, union octeon_rh *rh, struct octeon_droq_info *info) octeon_droq_dispatch_pkt() argument
H A Docteon_device.c1104 (u32)recv_pkt->rh.r_core_drv_init.app_mode), octeon_core_drv_init()
1106 oct->app_mode = (u32)recv_pkt->rh.r_core_drv_init.app_mode; octeon_core_drv_init()
1107 if (recv_pkt->rh.r_core_drv_init.app_mode == CVM_DRV_NIC_APP) { octeon_core_drv_init()
1109 (u32)recv_pkt->rh.r_core_drv_init.max_nic_ports; octeon_core_drv_init()
1111 (u32)recv_pkt->rh.r_core_drv_init.num_gmx_ports; octeon_core_drv_init()
1120 oct->fw_info.app_cap_flags = recv_pkt->rh.r_core_drv_init.app_cap_flags; octeon_core_drv_init()
1121 oct->fw_info.app_mode = (u32)recv_pkt->rh.r_core_drv_init.app_mode; octeon_core_drv_init()
H A Dliquidio_common.h81 #define OPCODE_SLOW_PATH(rh) \
82 (OPCODE_SUBCODE(rh->r.opcode, rh->r.subcode) != \
H A Docteon_droq.h60 union octeon_rh rh; member in struct:octeon_droq_info
138 union octeon_rh rh; member in struct:octeon_recv_pkt
H A Dlio_main.c108 u64 rh; member in struct:liquidio_if_cfg_resp
114 u64 rh; member in struct:oct_link_status_resp
120 u64 rh; member in struct:oct_timestamp_resp
1762 * @param rh - Control header associated with the packet
1769 union octeon_rh *rh, liquidio_push_packet()
1778 (struct net_device *)oct->props[rh->r_dh.link].netdev; liquidio_push_packet()
1794 if (rh->r_dh.has_hwtstamp) { liquidio_push_packet()
1814 (rh->r_dh.csum_verified == CNNIC_CSUM_VERIFIED)) liquidio_push_packet()
3092 (recv_pkt->rh.r_nic_info.ifidx > oct->ifcount)) { lio_nic_info()
3095 recv_pkt->rh.r_nic_info.ifidx); lio_nic_info()
3099 ifidx = recv_pkt->rh.r_nic_info.ifidx; lio_nic_info()
1766 liquidio_push_packet(u32 octeon_id, void *skbuff, u32 len, union octeon_rh *rh, void *param) liquidio_push_packet() argument
H A Dlio_ethtool.c50 u64 rh; member in struct:oct_mdio_cmd_resp
/linux-4.4.14/include/linux/fsl/bestcomm/
H A Dsram.h28 rh_info_t *rh; member in struct:bcom_sram
/linux-4.4.14/arch/arm/mm/
H A Dproc-v7-3level.S70 #define rh r2 define
73 #define rh r3 define
87 tst rh, #1 << (57 - 32) @ L_PTE_NONE
91 eor ip, rh, #1 << (55 - 32) @ toggle L_PTE_DIRTY in temp reg to
/linux-4.4.14/drivers/usb/host/
H A Dr8a66597-hcd.c1011 struct r8a66597_root_hub *rh = &r8a66597->root_hub[port]; start_root_hub_sampling() local
1013 rh->old_syssts = r8a66597_read(r8a66597, get_syssts_reg(port)) & LNST; start_root_hub_sampling()
1014 rh->scount = R8A66597_MAX_SAMPLING; start_root_hub_sampling()
1016 rh->port |= USB_PORT_STAT_CONNECTION; start_root_hub_sampling()
1018 rh->port &= ~USB_PORT_STAT_CONNECTION; start_root_hub_sampling()
1019 rh->port |= USB_PORT_STAT_C_CONNECTION << 16; start_root_hub_sampling()
1055 struct r8a66597_root_hub *rh = &r8a66597->root_hub[port]; r8a66597_usb_connect() local
1057 rh->port &= ~(USB_PORT_STAT_HIGH_SPEED | USB_PORT_STAT_LOW_SPEED); r8a66597_usb_connect()
1059 rh->port |= USB_PORT_STAT_HIGH_SPEED; r8a66597_usb_connect()
1061 rh->port |= USB_PORT_STAT_LOW_SPEED; r8a66597_usb_connect()
1063 rh->port &= ~USB_PORT_STAT_RESET; r8a66597_usb_connect()
1064 rh->port |= USB_PORT_STAT_ENABLE; r8a66597_usb_connect()
1701 struct r8a66597_root_hub *rh = &r8a66597->root_hub[port]; r8a66597_root_hub_control() local
1703 if (rh->port & USB_PORT_STAT_RESET) { r8a66597_root_hub_control()
1715 if (!(rh->port & USB_PORT_STAT_CONNECTION)) { r8a66597_root_hub_control()
1720 if (rh->scount > 0) { r8a66597_root_hub_control()
1722 if (tmp == rh->old_syssts) { r8a66597_root_hub_control()
1723 rh->scount--; r8a66597_root_hub_control()
1724 if (rh->scount == 0) r8a66597_root_hub_control()
1729 rh->scount = R8A66597_MAX_SAMPLING; r8a66597_root_hub_control()
1730 rh->old_syssts = tmp; r8a66597_root_hub_control()
2157 struct r8a66597_root_hub *rh = &r8a66597->root_hub[port]; r8a66597_hub_control() local
2182 rh->port &= ~USB_PORT_STAT_POWER; r8a66597_hub_control()
2198 rh->port &= ~(1 << wValue); r8a66597_hub_control()
2210 *(__le32 *)buf = cpu_to_le32(rh->port); r8a66597_hub_control()
2223 rh->port |= USB_PORT_STAT_POWER; r8a66597_hub_control()
2226 struct r8a66597_device *dev = rh->dev; r8a66597_hub_control()
2228 rh->port |= USB_PORT_STAT_RESET; r8a66597_hub_control()
2242 rh->port |= 1 << wValue; r8a66597_hub_control()
2263 struct r8a66597_root_hub *rh = &r8a66597->root_hub[port]; r8a66597_bus_suspend() local
2266 if (!(rh->port & USB_PORT_STAT_ENABLE)) r8a66597_bus_suspend()
2269 dev_dbg(&rh->dev->udev->dev, "suspend port = %d\n", port); r8a66597_bus_suspend()
2271 rh->port |= USB_PORT_STAT_SUSPEND; r8a66597_bus_suspend()
2273 if (rh->dev->udev->do_remote_wakeup) { r8a66597_bus_suspend()
2294 struct r8a66597_root_hub *rh = &r8a66597->root_hub[port]; r8a66597_bus_resume() local
2297 if (!(rh->port & USB_PORT_STAT_SUSPEND)) r8a66597_bus_resume()
2300 dev_dbg(&rh->dev->udev->dev, "resume port = %d\n", port); r8a66597_bus_resume()
2301 rh->port &= ~USB_PORT_STAT_SUSPEND; r8a66597_bus_resume()
2302 rh->port |= USB_PORT_STAT_C_SUSPEND << 16; r8a66597_bus_resume()
2361 struct r8a66597_root_hub *rh = &r8a66597->root_hub[port]; r8a66597_suspend() local
2363 rh->port = 0x00000000; r8a66597_suspend()
H A Dohci-hub.c536 u32 rh = roothub_a (ohci); ohci_hub_descriptor() local
540 desc->bPwrOn2PwrGood = (rh & RH_A_POTPGT) >> 24; ohci_hub_descriptor()
548 if (rh & RH_A_NPS) /* no power switching? */ ohci_hub_descriptor()
550 if (rh & RH_A_PSM) /* per-port power switching? */ ohci_hub_descriptor()
552 if (rh & RH_A_NOCP) /* no overcurrent reporting? */ ohci_hub_descriptor()
554 else if (rh & RH_A_OCPM) /* per-port overcurrent reporting? */ ohci_hub_descriptor()
559 rh = roothub_b (ohci); ohci_hub_descriptor()
562 desc->u.hs.DeviceRemovable[0] = rh & RH_B_DR; ohci_hub_descriptor()
564 desc->u.hs.DeviceRemovable[1] = (rh & RH_B_DR) >> 8; ohci_hub_descriptor()
H A Dohci-omap.c259 u32 rh = roothub_a (ohci); ohci_omap_reset() local
262 rh &= ~RH_A_NPS; ohci_omap_reset()
268 rh &= ~RH_A_NOCP; ohci_omap_reset()
278 ohci_writel(ohci, rh, &ohci->regs->roothub.a); ohci_omap_reset()
H A Dohci-dbg.c107 "OHCI %d.%d, %s legacy support registers, rh state %s\n", ohci_dump_status()
H A Dohci.h404 unsigned autostop:1; /* rh auto stopping/stopped */
H A Dehci-dbg.c812 "EHCI %x.%02x, rh state %s\n", fill_registers_buffer()
H A Dfotg210-hcd.c684 "EHCI %x.%02x, rh state %s\n", fill_registers_buffer()
/linux-4.4.14/net/ipv4/
H A Dinetpeer.c247 int lh, rh; peer_avl_rebalance() local
255 rh = node_height(r); peer_avl_rebalance()
256 if (lh > rh + 1) { /* l: RH+2 */ peer_avl_rebalance()
275 node->avl_height = rh + 1; /* node: RH+1 */ peer_avl_rebalance()
278 l->avl_height = rh + 1; /* l: RH+1 */ peer_avl_rebalance()
281 lr->avl_height = rh + 2; peer_avl_rebalance()
284 } else if (rh > lh + 1) { /* r: LH+2 */ peer_avl_rebalance()
313 node->avl_height = (lh > rh ? lh : rh) + 1; peer_avl_rebalance()
/linux-4.4.14/drivers/scsi/snic/
H A Dvnic_dev.c85 struct vnic_resource_header __iomem *rh; vnic_dev_discover_res() local
98 rh = bar->vaddr; vnic_dev_discover_res()
99 if (!rh) { vnic_dev_discover_res()
105 if (ioread32(&rh->magic) != VNIC_RES_MAGIC || vnic_dev_discover_res()
106 ioread32(&rh->version) != VNIC_RES_VERSION) { vnic_dev_discover_res()
109 ioread32(&rh->magic), ioread32(&rh->version)); vnic_dev_discover_res()
114 r = (struct vnic_resource __iomem *)(rh + 1); vnic_dev_discover_res()
/linux-4.4.14/drivers/scsi/fnic/
H A Dvnic_dev.c66 struct vnic_resource_header __iomem *rh; vnic_dev_discover_res() local
75 rh = bar->vaddr; vnic_dev_discover_res()
76 if (!rh) { vnic_dev_discover_res()
81 if (ioread32(&rh->magic) != VNIC_RES_MAGIC || vnic_dev_discover_res()
82 ioread32(&rh->version) != VNIC_RES_VERSION) { vnic_dev_discover_res()
86 ioread32(&rh->magic), ioread32(&rh->version)); vnic_dev_discover_res()
90 r = (struct vnic_resource __iomem *)(rh + 1); vnic_dev_discover_res()
/linux-4.4.14/arch/powerpc/include/asm/
H A Dfsl_85xx_cache_sram.h40 rh_info_t *rh; member in struct:mpc85xx_cache_sram
/linux-4.4.14/net/ipv6/
H A Dexthdrs_core.c225 struct ipv6_rt_hdr _rh, *rh; ipv6_find_hdr() local
227 rh = skb_header_pointer(skb, start, sizeof(_rh), ipv6_find_hdr()
229 if (!rh) ipv6_find_hdr()
233 rh->segments_left == 0) ipv6_find_hdr()
/linux-4.4.14/arch/ia64/kernel/
H A Dsalinfo.c414 sal_log_record_header_t *rh; salinfo_log_read_cpu() local
416 rh = (sal_log_record_header_t *)(data->log_buffer); salinfo_log_read_cpu()
418 if (rh->severity == sal_log_severity_corrected) salinfo_log_read_cpu()
435 sal_log_record_header_t *rh = (sal_log_record_header_t *)(data_saved->buffer); salinfo_log_new_read() local
437 memcpy(data->log_buffer, rh, data->log_size); salinfo_log_new_read()
439 if (rh->id == data_saved->id) { salinfo_log_new_read()
490 sal_log_record_header_t *rh; salinfo_log_clear() local
504 rh = (sal_log_record_header_t *)(data->log_buffer); salinfo_log_clear()
506 if (rh->severity != sal_log_severity_corrected) salinfo_log_clear()
H A Dmca.c458 sal_log_record_header_t *rh; ia64_mca_log_sal_error_record() local
477 rh = (sal_log_record_header_t *)buffer; ia64_mca_log_sal_error_record()
478 if (rh->severity == sal_log_severity_corrected) ia64_mca_log_sal_error_record()
1335 sal_log_record_header_t *rh = IA64_LOG_CURR_BUFFER(SAL_INFO_TYPE_MCA); ia64_mca_handler() local
1336 rh->severity = sal_log_severity_corrected; ia64_mca_handler()
/linux-4.4.14/drivers/scsi/lpfc/
H A Dlpfc_ct.c1411 struct lpfc_fdmi_reg_hba *rh; lpfc_fdmi_cmd() local
1478 rh = (struct lpfc_fdmi_reg_hba *)&CtReq->un.PortID; lpfc_fdmi_cmd()
1480 memcpy(&rh->hi.PortName, &vport->fc_sparam.portName, lpfc_fdmi_cmd()
1486 rh->rpl.EntryCnt = cpu_to_be32(1); lpfc_fdmi_cmd()
1487 memcpy(&rh->rpl.pe, &vport->fc_sparam.portName, lpfc_fdmi_cmd()
1497 ((uint8_t *)rh + size); lpfc_fdmi_cmd()
1506 ((uint8_t *)rh + size); lpfc_fdmi_cmd()
1522 ((uint8_t *)rh + size); lpfc_fdmi_cmd()
1540 ((uint8_t *)rh + size); lpfc_fdmi_cmd()
1558 ((uint8_t *)rh + size); lpfc_fdmi_cmd()
1575 ((uint8_t *)rh + size); lpfc_fdmi_cmd()
1592 ((uint8_t *)rh + size); lpfc_fdmi_cmd()
1619 ((uint8_t *)rh + size); lpfc_fdmi_cmd()
1637 ((uint8_t *)rh + size); lpfc_fdmi_cmd()
1655 ((uint8_t *)rh + size); lpfc_fdmi_cmd()
1673 ((uint8_t *)rh + size); lpfc_fdmi_cmd()
1694 ((uint8_t *)rh + size); lpfc_fdmi_cmd()
1715 ((uint8_t *)rh + size); lpfc_fdmi_cmd()
1739 rh = (struct lpfc_fdmi_reg_hba *) lpfc_fdmi_cmd()
1742 memcpy(&rh->hi.PortName, lpfc_fdmi_cmd()
1746 &rh->rpl.EntryCnt; lpfc_fdmi_cmd()
/linux-4.4.14/net/netfilter/
H A Dxt_connlimit.c107 union nf_inet_addr lh, rh; same_source_net() local
112 rh.ip6[i] = u3->ip6[i] & mask->ip6[i]; same_source_net()
115 return memcmp(&lh.ip6, &rh.ip6, sizeof(lh.ip6)); same_source_net()
/linux-4.4.14/include/math-emu/
H A Dop-2.h152 #define __FP_FRAC_ADD_2(rh, rl, xh, xl, yh, yl) \
153 (rh = xh + yh + ((rl = xl + yl) < xl))
156 #define __FP_FRAC_SUB_2(rh, rl, xh, xl, yh, yl) \
157 (rh = xh - yh - ((rl = xl - yl) > xl))
/linux-4.4.14/drivers/net/ethernet/cisco/enic/
H A Dvnic_dev.c47 struct vnic_resource_header __iomem *rh; vnic_dev_discover_res() local
60 rh = bar->vaddr; vnic_dev_discover_res()
62 if (!rh) { vnic_dev_discover_res()
68 if ((ioread32(&rh->magic) != VNIC_RES_MAGIC) || vnic_dev_discover_res()
69 (ioread32(&rh->version) != VNIC_RES_VERSION)) { vnic_dev_discover_res()
75 ioread32(&rh->magic), ioread32(&rh->version)); vnic_dev_discover_res()
83 r = (struct vnic_resource __iomem *)(rh + 1); vnic_dev_discover_res()
/linux-4.4.14/arch/x86/kvm/
H A Di8254.c63 u64 rl, rh; muldiv64() local
67 rh = (u64)u.l.high * (u64)b; muldiv64()
68 rh += (rl >> 32); muldiv64()
69 res.l.high = div64_u64(rh, c); muldiv64()
70 res.l.low = div64_u64(((mod_64(rh, c) << 32) + (rl & 0xffffffff)), c); muldiv64()
/linux-4.4.14/drivers/net/ethernet/broadcom/
H A Db44.c662 struct rx_header *rh; b44_alloc_rx_skb() local
706 rh = (struct rx_header *) skb->data; b44_alloc_rx_skb()
708 rh->len = 0; b44_alloc_rx_skb()
709 rh->flags = 0; b44_alloc_rx_skb()
737 struct rx_header *rh; b44_recycle_rx() local
748 rh = (struct rx_header *) src_map->skb->data; b44_recycle_rx()
749 rh->len = 0; b44_recycle_rx()
750 rh->flags = 0; b44_recycle_rx()
793 struct rx_header *rh; b44_rx() local
799 rh = (struct rx_header *) skb->data; b44_rx()
800 len = le16_to_cpu(rh->len); b44_rx()
802 (rh->flags & cpu_to_le16(RX_FLAG_ERRORS))) { b44_rx()
816 len = le16_to_cpu(rh->len); b44_rx()
/linux-4.4.14/fs/nfsd/
H A Dnfscache.c348 struct list_head *rh = &b->lru_head; nfsd_cache_search() local
351 list_for_each_entry(rp, rh, c_lru) { list_for_each_entry()
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/subdev/fb/
H A Dnv50.c221 * cause IOMMU "read from address 0" errors (rh#561267) nv50_fb_init()
/linux-4.4.14/drivers/media/usb/uvc/
H A Duvc_v4l2.c151 __u16 rw, rh; uvc_v4l2_try_format() local
186 rh = fmt->fmt.pix.height; uvc_v4l2_try_format()
193 d = min(w, rw) * min(h, rh); uvc_v4l2_try_format()
194 d = w*h + rw*rh - 2*d; uvc_v4l2_try_format()
/linux-4.4.14/include/trace/events/
H A Dkvm.h156 (__entry->address & (1<<3)) ? "|rh" : "")
/linux-4.4.14/drivers/usb/musb/
H A Dmusb_virthub.c126 dev_dbg(musb->controller, "bogus rh suspend? %s\n", musb_port_suspend()
/linux-4.4.14/drivers/net/ethernet/microchip/
H A Denc28j60.c301 int rl, rh; nolock_regw_read() local
305 rh = spi_read_op(priv, ENC28J60_READ_CTRL_REG, address + 1); nolock_regw_read()
307 return (rh << 8) | rl; nolock_regw_read()
/linux-4.4.14/arch/powerpc/lib/
H A Drheap.c128 printk(KERN_ERR "rh: out of slots; crash is imminent.\n"); get_slot()
/linux-4.4.14/drivers/net/ethernet/sun/
H A Dniu.c3425 struct rx_pkt_hdr1 *rh; niu_process_rx_pkt() local
3488 len += sizeof(*rh); niu_process_rx_pkt()
3489 len = min_t(int, len, sizeof(*rh) + VLAN_ETH_HLEN); niu_process_rx_pkt()
3492 rh = (struct rx_pkt_hdr1 *) skb->data; niu_process_rx_pkt()
3495 ((u32)rh->hashval2_0 << 24 | niu_process_rx_pkt()
3496 (u32)rh->hashval2_1 << 16 | niu_process_rx_pkt()
3497 (u32)rh->hashval1_1 << 8 | niu_process_rx_pkt()
3498 (u32)rh->hashval1_2 << 0), niu_process_rx_pkt()
3500 skb_pull(skb, sizeof(*rh)); niu_process_rx_pkt()
/linux-4.4.14/kernel/rcu/
H A Dtree_plugin.h2877 struct rcu_head rh; member in struct:rcu_sysidle_head
2892 rshp = container_of(rhp, struct rcu_sysidle_head, rh); rcu_sysidle_cb()
2959 call_rcu(&rsh.rh, rcu_sysidle_cb);
/linux-4.4.14/drivers/gpu/drm/nouveau/
H A Dnouveau_connector.c366 * valid - it's not (rh#613284) nouveau_connector_detect_lvds()
/linux-4.4.14/include/linux/usb/
H A Dhcd.h112 #define HCD_FLAG_POLL_RH 2 /* poll for rh status? */
/linux-4.4.14/drivers/usb/core/
H A Dhcd.c786 dev_dbg (hcd->self.controller, "not queuing rh status urb\n"); rh_queue_status()
/linux-4.4.14/drivers/block/
H A Dfloppy.c2311 pr_info("rh=%d h=%d\n", R_HEAD, HEAD); rw_interrupt()
/linux-4.4.14/fs/xfs/libxfs/
H A Dxfs_btree.c4031 /* now read rh sibling block for next iteration */ xfs_btree_block_change_owner()

Completed in 2720 milliseconds