/linux-4.4.14/drivers/md/ |
H A D | dm-region-hash.c | 102 struct dm_region_hash *rh; /* FIXME: can we get rid of this ? */ member in struct:dm_region 116 static region_t dm_rh_sector_to_region(struct dm_region_hash *rh, sector_t sector) dm_rh_sector_to_region() argument 118 return sector >> rh->region_shift; dm_rh_sector_to_region() 121 sector_t dm_rh_region_to_sector(struct dm_region_hash *rh, region_t region) dm_rh_region_to_sector() argument 123 return region << rh->region_shift; dm_rh_region_to_sector() 127 region_t dm_rh_bio_to_region(struct dm_region_hash *rh, struct bio *bio) dm_rh_bio_to_region() argument 129 return dm_rh_sector_to_region(rh, bio->bi_iter.bi_sector - dm_rh_bio_to_region() 130 rh->target_begin); dm_rh_bio_to_region() 136 return reg->rh->context; dm_rh_region_context() 146 sector_t dm_rh_get_region_size(struct dm_region_hash *rh) dm_rh_get_region_size() argument 148 return rh->region_size; dm_rh_get_region_size() 169 struct dm_region_hash *rh; dm_region_hash_create() local 182 rh = kmalloc(sizeof(*rh), GFP_KERNEL); dm_region_hash_create() 183 if (!rh) { dm_region_hash_create() 188 rh->context = context; dm_region_hash_create() 189 rh->dispatch_bios = dispatch_bios; dm_region_hash_create() 190 rh->wakeup_workers = wakeup_workers; dm_region_hash_create() 191 rh->wakeup_all_recovery_waiters = wakeup_all_recovery_waiters; dm_region_hash_create() 192 rh->target_begin = target_begin; dm_region_hash_create() 193 rh->max_recovery = max_recovery; dm_region_hash_create() 194 rh->log = log; dm_region_hash_create() 195 rh->region_size = region_size; dm_region_hash_create() 196 rh->region_shift = __ffs(region_size); dm_region_hash_create() 197 rwlock_init(&rh->hash_lock); dm_region_hash_create() 198 rh->mask = nr_buckets - 1; dm_region_hash_create() 199 rh->nr_buckets = nr_buckets; dm_region_hash_create() 201 rh->shift = RH_HASH_SHIFT; dm_region_hash_create() 202 rh->prime = RH_HASH_MULT; dm_region_hash_create() 204 rh->buckets = vmalloc(nr_buckets * sizeof(*rh->buckets)); dm_region_hash_create() 205 if (!rh->buckets) { dm_region_hash_create() 207 kfree(rh); dm_region_hash_create() 212 INIT_LIST_HEAD(rh->buckets + i); dm_region_hash_create() 214 spin_lock_init(&rh->region_lock); dm_region_hash_create() 215 sema_init(&rh->recovery_count, 0); dm_region_hash_create() 216 atomic_set(&rh->recovery_in_flight, 0); dm_region_hash_create() 217 INIT_LIST_HEAD(&rh->clean_regions); dm_region_hash_create() 218 INIT_LIST_HEAD(&rh->quiesced_regions); dm_region_hash_create() 219 INIT_LIST_HEAD(&rh->recovered_regions); dm_region_hash_create() 220 INIT_LIST_HEAD(&rh->failed_recovered_regions); dm_region_hash_create() 221 rh->flush_failure = 0; dm_region_hash_create() 223 rh->region_pool = mempool_create_kmalloc_pool(MIN_REGIONS, dm_region_hash_create() 225 if (!rh->region_pool) { dm_region_hash_create() 226 vfree(rh->buckets); dm_region_hash_create() 227 kfree(rh); dm_region_hash_create() 228 rh = ERR_PTR(-ENOMEM); dm_region_hash_create() 231 return rh; dm_region_hash_create() 235 void dm_region_hash_destroy(struct dm_region_hash *rh) dm_region_hash_destroy() argument 240 BUG_ON(!list_empty(&rh->quiesced_regions)); dm_region_hash_destroy() 241 for (h = 0; h < rh->nr_buckets; h++) { dm_region_hash_destroy() 242 list_for_each_entry_safe(reg, nreg, rh->buckets + h, dm_region_hash_destroy() 245 mempool_free(reg, rh->region_pool); dm_region_hash_destroy() 249 if (rh->log) dm_region_hash_destroy() 250 dm_dirty_log_destroy(rh->log); dm_region_hash_destroy() 252 mempool_destroy(rh->region_pool); dm_region_hash_destroy() 253 vfree(rh->buckets); dm_region_hash_destroy() 254 kfree(rh); dm_region_hash_destroy() 258 struct dm_dirty_log *dm_rh_dirty_log(struct dm_region_hash *rh) dm_rh_dirty_log() argument 260 return rh->log; dm_rh_dirty_log() 264 static unsigned rh_hash(struct dm_region_hash *rh, region_t region) rh_hash() argument 266 return (unsigned) ((region * rh->prime) >> rh->shift) & rh->mask; rh_hash() 269 static struct dm_region *__rh_lookup(struct dm_region_hash *rh, region_t region) __rh_lookup() argument 272 struct list_head *bucket = rh->buckets + rh_hash(rh, region); __rh_lookup() 281 static void __rh_insert(struct dm_region_hash *rh, struct dm_region *reg) __rh_insert() argument 283 list_add(®->hash_list, rh->buckets + rh_hash(rh, reg->key)); __rh_insert() 286 static struct dm_region *__rh_alloc(struct dm_region_hash *rh, region_t region) __rh_alloc() argument 290 nreg = mempool_alloc(rh->region_pool, GFP_ATOMIC); __rh_alloc() 294 nreg->state = rh->log->type->in_sync(rh->log, region, 1) ? __rh_alloc() 296 nreg->rh = rh; __rh_alloc() 302 write_lock_irq(&rh->hash_lock); __rh_alloc() 303 reg = __rh_lookup(rh, region); __rh_alloc() 306 mempool_free(nreg, rh->region_pool); __rh_alloc() 308 __rh_insert(rh, nreg); __rh_alloc() 310 spin_lock(&rh->region_lock); __rh_alloc() 311 list_add(&nreg->list, &rh->clean_regions); __rh_alloc() 312 spin_unlock(&rh->region_lock); __rh_alloc() 317 write_unlock_irq(&rh->hash_lock); __rh_alloc() 322 static struct dm_region *__rh_find(struct dm_region_hash *rh, region_t region) __rh_find() argument 326 reg = __rh_lookup(rh, region); __rh_find() 328 read_unlock(&rh->hash_lock); __rh_find() 329 reg = __rh_alloc(rh, region); __rh_find() 330 read_lock(&rh->hash_lock); __rh_find() 336 int dm_rh_get_state(struct dm_region_hash *rh, region_t region, int may_block) dm_rh_get_state() argument 341 read_lock(&rh->hash_lock); dm_rh_get_state() 342 reg = __rh_lookup(rh, region); dm_rh_get_state() 343 read_unlock(&rh->hash_lock); dm_rh_get_state() 352 r = rh->log->type->in_sync(rh->log, region, may_block); dm_rh_get_state() 364 struct dm_region_hash *rh = reg->rh; complete_resync_work() local 366 rh->log->type->set_region_sync(rh->log, reg->key, success); complete_resync_work() 377 rh->dispatch_bios(rh->context, ®->delayed_bios); complete_resync_work() 378 if (atomic_dec_and_test(&rh->recovery_in_flight)) complete_resync_work() 379 rh->wakeup_all_recovery_waiters(rh->context); complete_resync_work() 380 up(&rh->recovery_count); complete_resync_work() 393 void dm_rh_mark_nosync(struct dm_region_hash *rh, struct bio *bio) dm_rh_mark_nosync() argument 396 struct dm_dirty_log *log = rh->log; dm_rh_mark_nosync() 398 region_t region = dm_rh_bio_to_region(rh, bio); dm_rh_mark_nosync() 402 rh->flush_failure = 1; dm_rh_mark_nosync() 412 read_lock(&rh->hash_lock); dm_rh_mark_nosync() 413 reg = __rh_find(rh, region); dm_rh_mark_nosync() 414 read_unlock(&rh->hash_lock); dm_rh_mark_nosync() 420 spin_lock_irqsave(&rh->region_lock, flags); dm_rh_mark_nosync() 431 spin_unlock_irqrestore(&rh->region_lock, flags); dm_rh_mark_nosync() 438 void dm_rh_update_states(struct dm_region_hash *rh, int errors_handled) dm_rh_update_states() argument 449 write_lock_irq(&rh->hash_lock); dm_rh_update_states() 450 spin_lock(&rh->region_lock); dm_rh_update_states() 451 if (!list_empty(&rh->clean_regions)) { dm_rh_update_states() 452 list_splice_init(&rh->clean_regions, &clean); dm_rh_update_states() 458 if (!list_empty(&rh->recovered_regions)) { dm_rh_update_states() 459 list_splice_init(&rh->recovered_regions, &recovered); dm_rh_update_states() 465 if (!list_empty(&rh->failed_recovered_regions)) { dm_rh_update_states() 466 list_splice_init(&rh->failed_recovered_regions, dm_rh_update_states() 473 spin_unlock(&rh->region_lock); dm_rh_update_states() 474 write_unlock_irq(&rh->hash_lock); dm_rh_update_states() 482 rh->log->type->clear_region(rh->log, reg->key); dm_rh_update_states() 484 mempool_free(reg, rh->region_pool); dm_rh_update_states() 489 mempool_free(reg, rh->region_pool); dm_rh_update_states() 493 rh->log->type->clear_region(rh->log, reg->key); dm_rh_update_states() 494 mempool_free(reg, rh->region_pool); dm_rh_update_states() 497 rh->log->type->flush(rh->log); dm_rh_update_states() 501 static void rh_inc(struct dm_region_hash *rh, region_t region) rh_inc() argument 505 read_lock(&rh->hash_lock); rh_inc() 506 reg = __rh_find(rh, region); rh_inc() 508 spin_lock_irq(&rh->region_lock); rh_inc() 514 spin_unlock_irq(&rh->region_lock); rh_inc() 516 rh->log->type->mark_region(rh->log, reg->key); rh_inc() 518 spin_unlock_irq(&rh->region_lock); rh_inc() 521 read_unlock(&rh->hash_lock); rh_inc() 524 void dm_rh_inc_pending(struct dm_region_hash *rh, struct bio_list *bios) dm_rh_inc_pending() argument 531 rh_inc(rh, dm_rh_bio_to_region(rh, bio)); dm_rh_inc_pending() 536 void dm_rh_dec(struct dm_region_hash *rh, region_t region) dm_rh_dec() argument 542 read_lock(&rh->hash_lock); dm_rh_dec() 543 reg = __rh_lookup(rh, region); dm_rh_dec() 544 read_unlock(&rh->hash_lock); dm_rh_dec() 546 spin_lock_irqsave(&rh->region_lock, flags); dm_rh_dec() 560 if (unlikely(rh->flush_failure)) { dm_rh_dec() 568 list_add_tail(®->list, &rh->quiesced_regions); dm_rh_dec() 571 list_add(®->list, &rh->clean_regions); dm_rh_dec() 575 spin_unlock_irqrestore(&rh->region_lock, flags); dm_rh_dec() 578 rh->wakeup_workers(rh->context); dm_rh_dec() 585 static int __rh_recovery_prepare(struct dm_region_hash *rh) __rh_recovery_prepare() argument 594 r = rh->log->type->get_resync_work(rh->log, ®ion); __rh_recovery_prepare() 602 read_lock(&rh->hash_lock); __rh_recovery_prepare() 603 reg = __rh_find(rh, region); __rh_recovery_prepare() 604 read_unlock(&rh->hash_lock); __rh_recovery_prepare() 606 spin_lock_irq(&rh->region_lock); __rh_recovery_prepare() 613 list_move(®->list, &rh->quiesced_regions); __rh_recovery_prepare() 615 spin_unlock_irq(&rh->region_lock); __rh_recovery_prepare() 620 void dm_rh_recovery_prepare(struct dm_region_hash *rh) dm_rh_recovery_prepare() argument 623 atomic_inc(&rh->recovery_in_flight); dm_rh_recovery_prepare() 625 while (!down_trylock(&rh->recovery_count)) { dm_rh_recovery_prepare() 626 atomic_inc(&rh->recovery_in_flight); dm_rh_recovery_prepare() 627 if (__rh_recovery_prepare(rh) <= 0) { dm_rh_recovery_prepare() 628 atomic_dec(&rh->recovery_in_flight); dm_rh_recovery_prepare() 629 up(&rh->recovery_count); dm_rh_recovery_prepare() 635 if (atomic_dec_and_test(&rh->recovery_in_flight)) dm_rh_recovery_prepare() 636 rh->wakeup_all_recovery_waiters(rh->context); dm_rh_recovery_prepare() 643 struct dm_region *dm_rh_recovery_start(struct dm_region_hash *rh) dm_rh_recovery_start() argument 647 spin_lock_irq(&rh->region_lock); dm_rh_recovery_start() 648 if (!list_empty(&rh->quiesced_regions)) { dm_rh_recovery_start() 649 reg = list_entry(rh->quiesced_regions.next, dm_rh_recovery_start() 653 spin_unlock_irq(&rh->region_lock); dm_rh_recovery_start() 661 struct dm_region_hash *rh = reg->rh; dm_rh_recovery_end() local 663 spin_lock_irq(&rh->region_lock); dm_rh_recovery_end() 665 list_add(®->list, ®->rh->recovered_regions); dm_rh_recovery_end() 667 list_add(®->list, ®->rh->failed_recovered_regions); dm_rh_recovery_end() 669 spin_unlock_irq(&rh->region_lock); dm_rh_recovery_end() 671 rh->wakeup_workers(rh->context); dm_rh_recovery_end() 676 int dm_rh_recovery_in_flight(struct dm_region_hash *rh) dm_rh_recovery_in_flight() argument 678 return atomic_read(&rh->recovery_in_flight); dm_rh_recovery_in_flight() 682 int dm_rh_flush(struct dm_region_hash *rh) dm_rh_flush() argument 684 return rh->log->type->flush(rh->log); dm_rh_flush() 688 void dm_rh_delay(struct dm_region_hash *rh, struct bio *bio) dm_rh_delay() argument 692 read_lock(&rh->hash_lock); dm_rh_delay() 693 reg = __rh_find(rh, dm_rh_bio_to_region(rh, bio)); dm_rh_delay() 695 read_unlock(&rh->hash_lock); dm_rh_delay() 699 void dm_rh_stop_recovery(struct dm_region_hash *rh) dm_rh_stop_recovery() argument 704 for (i = 0; i < rh->max_recovery; i++) dm_rh_stop_recovery() 705 down(&rh->recovery_count); dm_rh_stop_recovery() 709 void dm_rh_start_recovery(struct dm_region_hash *rh) dm_rh_start_recovery() argument 713 for (i = 0; i < rh->max_recovery; i++) dm_rh_start_recovery() 714 up(&rh->recovery_count); dm_rh_start_recovery() 716 rh->wakeup_workers(rh->context); dm_rh_start_recovery()
|
H A D | dm-raid1.c | 63 struct dm_region_hash *rh; member in struct:mirror_set 336 sector_t region_size = dm_rh_get_region_size(ms->rh); recover() 341 from.sector = m->offset + dm_rh_region_to_sector(ms->rh, key); recover() 360 dest->sector = m->offset + dm_rh_region_to_sector(ms->rh, key); recover() 389 struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); do_recovery() 395 dm_rh_recovery_prepare(ms->rh); do_recovery() 400 while ((reg = dm_rh_recovery_start(ms->rh))) { do_recovery() 445 struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); mirror_available() 446 region_t region = dm_rh_bio_to_region(ms->rh, bio); mirror_available() 560 int state = dm_rh_get_state(ms->rh, region, may_block); region_in_sync() 571 region = dm_rh_bio_to_region(ms->rh, bio); do_reads() 689 struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); do_writes() 710 region = dm_rh_bio_to_region(ms->rh, bio); do_writes() 718 state = dm_rh_get_state(ms->rh, region, 1); do_writes() 753 dm_rh_inc_pending(ms->rh, &sync); do_writes() 754 dm_rh_inc_pending(ms->rh, &nosync); do_writes() 761 ms->log_failure = dm_rh_flush(ms->rh) ? 1 : ms->log_failure; do_writes() 776 dm_rh_delay(ms->rh, bio); do_writes() 818 dm_rh_mark_nosync(ms->rh, bio); do_failures() 870 dm_rh_update_states(ms->rh, errors_handled(ms)); do_mirror() 918 ms->rh = dm_region_hash_create(ms, dispatch_bios, wakeup_mirrord, alloc_context() 922 if (IS_ERR(ms->rh)) { alloc_context() 939 dm_region_hash_destroy(ms->rh); free_context() 1118 r = dm_set_target_max_io_len(ti, dm_rh_get_region_size(ms->rh)); mirror_ctr() 1196 struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); mirror_map() 1204 bio_record->write_region = dm_rh_bio_to_region(ms->rh, bio); mirror_map() 1209 r = log->type->in_sync(log, dm_rh_bio_to_region(ms->rh, bio), 0); mirror_map() 1254 dm_rh_dec(ms->rh, bio_record->write_region); mirror_end_io() 1307 struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); mirror_presuspend() 1332 dm_rh_stop_recovery(ms->rh); mirror_presuspend() 1335 !dm_rh_recovery_in_flight(ms->rh)); mirror_presuspend() 1353 struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); mirror_postsuspend() 1363 struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); mirror_resume() 1369 dm_rh_start_recovery(ms->rh); mirror_resume() 1403 struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); mirror_status()
|
/linux-4.4.14/include/linux/ |
H A D | dm-region-hash.h | 43 void dm_region_hash_destroy(struct dm_region_hash *rh); 45 struct dm_dirty_log *dm_rh_dirty_log(struct dm_region_hash *rh); 50 region_t dm_rh_bio_to_region(struct dm_region_hash *rh, struct bio *bio); 51 sector_t dm_rh_region_to_sector(struct dm_region_hash *rh, region_t region); 57 sector_t dm_rh_get_region_size(struct dm_region_hash *rh); 64 int dm_rh_get_state(struct dm_region_hash *rh, region_t region, int may_block); 65 void dm_rh_set_state(struct dm_region_hash *rh, region_t region, 69 void dm_rh_update_states(struct dm_region_hash *rh, int errors_handled); 72 int dm_rh_flush(struct dm_region_hash *rh); 75 void dm_rh_inc_pending(struct dm_region_hash *rh, struct bio_list *bios); 76 void dm_rh_dec(struct dm_region_hash *rh, region_t region); 79 void dm_rh_delay(struct dm_region_hash *rh, struct bio *bio); 81 void dm_rh_mark_nosync(struct dm_region_hash *rh, struct bio *bio); 88 void dm_rh_recovery_prepare(struct dm_region_hash *rh); 91 struct dm_region *dm_rh_recovery_start(struct dm_region_hash *rh); 97 int dm_rh_recovery_in_flight(struct dm_region_hash *rh); 100 void dm_rh_start_recovery(struct dm_region_hash *rh); 101 void dm_rh_stop_recovery(struct dm_region_hash *rh);
|
H A D | math64.h | 183 } rl, rm, rn, rh, a0, b0; mul_u64_u64_shr() local 192 rh.ll = (u64)a0.l.high * b0.l.high; mul_u64_u64_shr() 200 rh.l.low = c = (c >> 32) + rm.l.high + rn.l.high + rh.l.low; mul_u64_u64_shr() 201 rh.l.high = (c >> 32) + rh.l.high; mul_u64_u64_shr() 204 * The 128-bit result of the multiplication is in rl.ll and rh.ll, mul_u64_u64_shr() 210 return (rl.ll >> shift) | (rh.ll << (64 - shift)); mul_u64_u64_shr() 211 return rh.ll >> (shift & 63); mul_u64_u64_shr() 229 } u, rl, rh; mul_u64_u32_div() local 233 rh.ll = (u64)u.l.high * mul + rl.l.high; mul_u64_u32_div() 235 /* Bits 32-63 of the result will be in rh.l.low. */ mul_u64_u32_div() 236 rl.l.high = do_div(rh.ll, divisor); mul_u64_u32_div() 241 rl.l.high = rh.l.low; mul_u64_u32_div()
|
/linux-4.4.14/drivers/usb/wusbcore/ |
H A D | Makefile | 14 rh.o \
|
H A D | wusbhc.h | 39 * rh Root Hub emulation (part of the HCD glue) 144 * See rh.c for more information.
|
H A D | wa-hc.h | 68 * and call into the rh.c:hwahc_rc_port_reset() code to authenticate
|
H A D | devconnect.c | 69 * __wusb_dev_disable() Called by rh.c:wusbhc_rh_clear_port_feat() when
|
/linux-4.4.14/net/ipv6/netfilter/ |
H A D | ip6t_rt.c | 42 const struct ipv6_rt_hdr *rh; rt_mt6() local 59 rh = skb_header_pointer(skb, ptr, sizeof(_route), &_route); rt_mt6() 60 if (rh == NULL) { rt_mt6() 65 hdrlen = ipv6_optlen(rh); rt_mt6() 71 pr_debug("IPv6 RT LEN %u %u ", hdrlen, rh->hdrlen); rt_mt6() 72 pr_debug("TYPE %04X ", rh->type); rt_mt6() 73 pr_debug("SGS_LEFT %u %02X\n", rh->segments_left, rh->segments_left); rt_mt6() 77 rh->segments_left, rt_mt6() 80 rtinfo->rt_type, rh->type, rt_mt6() 82 ((rtinfo->rt_type == rh->type) ^ rt_mt6() 91 ((const struct rt0_hdr *)rh)->reserved, rt_mt6() 93 (((const struct rt0_hdr *)rh)->reserved))); rt_mt6() 95 ret = (rh != NULL) && rt_mt6() 97 rh->segments_left, rt_mt6() 103 ((rtinfo->rt_type == rh->type) ^ rt_mt6()
|
/linux-4.4.14/crypto/ |
H A D | vmac.c | 66 #define ADD128(rh, rl, ih, il) \ 71 (rh)++; \ 72 (rh) += (ih); \ 77 #define PMUL64(rh, rl, i1, i2) /* Assumes m doesn't overflow */ \ 81 rh = MUL32(_i1>>32, _i2>>32); \ 83 ADD128(rh, rl, (m >> 32), (m << 32)); \ 86 #define MUL64(rh, rl, i1, i2) \ 91 rh = MUL32(_i1>>32, _i2>>32); \ 93 ADD128(rh, rl, (m1 >> 32), (m1 << 32)); \ 94 ADD128(rh, rl, (m2 >> 32), (m2 << 32)); \ 111 #define nh_16(mp, kp, nw, rh, rl) \ 114 rh = rl = 0; \ 118 ADD128(rh, rl, th, tl); \ 122 #define nh_16_2(mp, kp, nw, rh, rl, rh1, rl1) \ 125 rh1 = rl1 = rh = rl = 0; \ 129 ADD128(rh, rl, th, tl); \ 137 #define nh_vmac_nhbytes(mp, kp, nw, rh, rl) \ 140 rh = rl = 0; \ 144 ADD128(rh, rl, th, tl); \ 147 ADD128(rh, rl, th, tl); \ 150 ADD128(rh, rl, th, tl); \ 153 ADD128(rh, rl, th, tl); \ 157 #define nh_vmac_nhbytes_2(mp, kp, nw, rh, rl, rh1, rl1) \ 160 rh1 = rl1 = rh = rl = 0; \ 164 ADD128(rh, rl, th, tl); \ 170 ADD128(rh, rl, th, tl); \ 176 ADD128(rh, rl, th, tl); \ 182 ADD128(rh, rl, th, tl); \ 216 #define nh_16(mp, kp, nw, rh, rl) \ 220 rh = rl = t = 0; \ 226 ADD128(rh, rl, MUL32(t1 >> 32, t2 >> 32), \ 228 rh += (u64)(u32)(m1 >> 32) \ 232 ADD128(rh, rl, (t >> 32), (t << 32)); \ 303 #define nh_16_2(mp, kp, nw, rh, rl, rh2, rl2) \ 305 nh_16(mp, kp, nw, rh, rl); \ 310 #define nh_vmac_nhbytes(mp, kp, nw, rh, rl) \ 311 nh_16(mp, kp, nw, rh, rl) 314 #define nh_vmac_nhbytes_2(mp, kp, nw, rh, rl, rh2, rl2) \ 316 nh_vmac_nhbytes(mp, kp, nw, rh, rl); \ 330 u64 rh, rl, t, z = 0; l3hash() local 355 MUL64(rh, rl, p1, p2); l3hash() 356 t = rh >> 56; l3hash() 357 ADD128(t, rl, z, rh); l3hash() 358 rh <<= 8; l3hash() 359 ADD128(t, rl, z, rh); l3hash() 371 u64 rh, rl, *mptr; vhash_update() local 391 nh_vmac_nhbytes(mptr, kptr, VMAC_NHBYTES/8, rh, rl); vhash_update() 392 rh &= m62; vhash_update() 393 ADD128(ch, cl, rh, rl); vhash_update() 399 nh_vmac_nhbytes(mptr, kptr, VMAC_NHBYTES/8, rh, rl); vhash_update() 400 rh &= m62; vhash_update() 401 poly_step(ch, cl, pkh, pkl, rh, rl); vhash_update() 412 u64 rh, rl, *mptr; vhash() local 444 nh_vmac_nhbytes(mptr, kptr, VMAC_NHBYTES/8, rh, rl); vhash() 445 rh &= m62; vhash() 446 poly_step(ch, cl, pkh, pkl, rh, rl); vhash() 450 nh_16(mptr, kptr, 2*((remaining+15)/16), rh, rl); vhash() 451 rh &= m62; vhash() 452 poly_step(ch, cl, pkh, pkl, rh, rl); vhash()
|
/linux-4.4.14/drivers/dma/bestcomm/ |
H A D | sram.c | 97 bcom_sram->rh = rh_create(4); bcom_sram_init() 110 rh_attach_region(bcom_sram->rh, 0, bcom_sram->size); bcom_sram_init() 115 rh_attach_region(bcom_sram->rh, zbase - bcom_sram->base_phys, regaddr_p[1]); bcom_sram_init() 140 rh_destroy(bcom_sram->rh); bcom_sram_cleanup() 154 offset = rh_alloc_align(bcom_sram->rh, size, align, NULL); bcom_sram_alloc() 175 rh_free(bcom_sram->rh, offset); bcom_sram_free()
|
/linux-4.4.14/arch/powerpc/sysdev/ |
H A D | fsl_85xx_cache_sram.c | 60 offset = rh_alloc_align(cache_sram->rh, size, align, NULL); mpc85xx_cache_sram_alloc() 78 rh_free(cache_sram->rh, ptr - cache_sram->base_virt); mpc85xx_cache_sram_free() 119 cache_sram->rh = rh_create(sizeof(unsigned int)); instantiate_cache_sram() 120 if (IS_ERR(cache_sram->rh)) { instantiate_cache_sram() 123 ret = PTR_ERR(cache_sram->rh); instantiate_cache_sram() 127 rh_attach_region(cache_sram->rh, 0, cache_sram->size); instantiate_cache_sram() 150 rh_detach_region(cache_sram->rh, 0, cache_sram->size); remove_cache_sram() 151 rh_destroy(cache_sram->rh); remove_cache_sram()
|
H A D | ppc4xx_ocm.c | 53 rh_info_t *rh; member in struct:ocm_region 97 rh_free(ocm_reg->rh, offset); ocm_free_region() 216 ocm->nc.rh = rh_create(ocm->alignment); ocm_init_node() 217 rh_attach_region(ocm->nc.rh, 0, ocm->nc.memtotal); ocm_init_node() 221 ocm->c.rh = rh_create(ocm->alignment); ocm_init_node() 222 rh_attach_region(ocm->c.rh, 0, ocm->c.memtotal); ocm_init_node() 337 offset = rh_alloc_align(ocm_reg->rh, size, align, NULL); ppc4xx_ocm_alloc() 345 rh_free(ocm_reg->rh, offset); ppc4xx_ocm_alloc()
|
/linux-4.4.14/net/dccp/ccids/lib/ |
H A D | loss_interval.c | 133 * @rh: Receive history containing a fresh loss event 139 int tfrc_lh_interval_add(struct tfrc_loss_hist *lh, struct tfrc_rx_hist *rh, tfrc_lh_interval_add() argument 144 if (cur != NULL && !tfrc_lh_is_new_loss(cur, tfrc_rx_hist_loss_prev(rh))) tfrc_lh_interval_add() 153 new->li_seqno = tfrc_rx_hist_loss_prev(rh)->tfrchrx_seqno; tfrc_lh_interval_add() 154 new->li_ccval = tfrc_rx_hist_loss_prev(rh)->tfrchrx_ccval; tfrc_lh_interval_add() 162 tfrc_rx_hist_last_rcv(rh)->tfrchrx_seqno) + 1; tfrc_lh_interval_add()
|
/linux-4.4.14/arch/arm/vfp/ |
H A D | vfp.h | 76 u64 rh, rma, rmb, rl; mul64to128() local 89 rh = (u64)nh * mh; mul64to128() 90 rh += ((u64)(rma < rmb) << 32) + (rma >> 32); mul64to128() 94 rh += (rl < rma); mul64to128() 97 *resh = rh; mul64to128() 108 u64 rh, rl; vfp_hi64multiply64() local 109 mul64to128(&rh, &rl, n, m); vfp_hi64multiply64() 110 return rh | (rl != 0); vfp_hi64multiply64()
|
/linux-4.4.14/drivers/net/ethernet/cavium/liquidio/ |
H A D | octeon_droq.c | 369 recv_pkt->rh = info->rh; octeon_create_recv_info() 517 union octeon_rh *rh, octeon_droq_dispatch_pkt() 526 disp_fn = octeon_get_dispatch(oct, (u16)rh->r.opcode, octeon_droq_dispatch_pkt() 527 (u16)rh->r.subcode); octeon_droq_dispatch_pkt() 535 rinfo->recv_pkt->rh = *rh; octeon_droq_dispatch_pkt() 581 union octeon_rh *rh; octeon_droq_fast_process_packets() local 605 rh = &info->rh; octeon_droq_fast_process_packets() 609 if (OPCODE_SLOW_PATH(rh)) { octeon_droq_fast_process_packets() 612 buf_cnt = octeon_droq_dispatch_pkt(oct, droq, rh, info); octeon_droq_fast_process_packets() 674 rh, &droq->napi); octeon_droq_fast_process_packets() 741 (u16)rdisp->rinfo->recv_pkt->rh.r.opcode, octeon_droq_process_packets() 742 (u16)rdisp->rinfo->recv_pkt->rh.r.subcode)); octeon_droq_process_packets() 798 (u16)rdisp->rinfo->recv_pkt->rh.r.opcode, octeon_droq_process_poll_pkts() 799 (u16)rdisp->rinfo->recv_pkt->rh.r.subcode)); octeon_droq_process_poll_pkts() 515 octeon_droq_dispatch_pkt(struct octeon_device *oct, struct octeon_droq *droq, union octeon_rh *rh, struct octeon_droq_info *info) octeon_droq_dispatch_pkt() argument
|
H A D | octeon_device.c | 1104 (u32)recv_pkt->rh.r_core_drv_init.app_mode), octeon_core_drv_init() 1106 oct->app_mode = (u32)recv_pkt->rh.r_core_drv_init.app_mode; octeon_core_drv_init() 1107 if (recv_pkt->rh.r_core_drv_init.app_mode == CVM_DRV_NIC_APP) { octeon_core_drv_init() 1109 (u32)recv_pkt->rh.r_core_drv_init.max_nic_ports; octeon_core_drv_init() 1111 (u32)recv_pkt->rh.r_core_drv_init.num_gmx_ports; octeon_core_drv_init() 1120 oct->fw_info.app_cap_flags = recv_pkt->rh.r_core_drv_init.app_cap_flags; octeon_core_drv_init() 1121 oct->fw_info.app_mode = (u32)recv_pkt->rh.r_core_drv_init.app_mode; octeon_core_drv_init()
|
H A D | liquidio_common.h | 81 #define OPCODE_SLOW_PATH(rh) \ 82 (OPCODE_SUBCODE(rh->r.opcode, rh->r.subcode) != \
|
H A D | octeon_droq.h | 60 union octeon_rh rh; member in struct:octeon_droq_info 138 union octeon_rh rh; member in struct:octeon_recv_pkt
|
H A D | lio_main.c | 108 u64 rh; member in struct:liquidio_if_cfg_resp 114 u64 rh; member in struct:oct_link_status_resp 120 u64 rh; member in struct:oct_timestamp_resp 1762 * @param rh - Control header associated with the packet 1769 union octeon_rh *rh, liquidio_push_packet() 1778 (struct net_device *)oct->props[rh->r_dh.link].netdev; liquidio_push_packet() 1794 if (rh->r_dh.has_hwtstamp) { liquidio_push_packet() 1814 (rh->r_dh.csum_verified == CNNIC_CSUM_VERIFIED)) liquidio_push_packet() 3092 (recv_pkt->rh.r_nic_info.ifidx > oct->ifcount)) { lio_nic_info() 3095 recv_pkt->rh.r_nic_info.ifidx); lio_nic_info() 3099 ifidx = recv_pkt->rh.r_nic_info.ifidx; lio_nic_info() 1766 liquidio_push_packet(u32 octeon_id, void *skbuff, u32 len, union octeon_rh *rh, void *param) liquidio_push_packet() argument
|
H A D | lio_ethtool.c | 50 u64 rh; member in struct:oct_mdio_cmd_resp
|
/linux-4.4.14/include/linux/fsl/bestcomm/ |
H A D | sram.h | 28 rh_info_t *rh; member in struct:bcom_sram
|
/linux-4.4.14/arch/arm/mm/ |
H A D | proc-v7-3level.S | 70 #define rh r2 define 73 #define rh r3 define 87 tst rh, #1 << (57 - 32) @ L_PTE_NONE 91 eor ip, rh, #1 << (55 - 32) @ toggle L_PTE_DIRTY in temp reg to
|
/linux-4.4.14/drivers/usb/host/ |
H A D | r8a66597-hcd.c | 1011 struct r8a66597_root_hub *rh = &r8a66597->root_hub[port]; start_root_hub_sampling() local 1013 rh->old_syssts = r8a66597_read(r8a66597, get_syssts_reg(port)) & LNST; start_root_hub_sampling() 1014 rh->scount = R8A66597_MAX_SAMPLING; start_root_hub_sampling() 1016 rh->port |= USB_PORT_STAT_CONNECTION; start_root_hub_sampling() 1018 rh->port &= ~USB_PORT_STAT_CONNECTION; start_root_hub_sampling() 1019 rh->port |= USB_PORT_STAT_C_CONNECTION << 16; start_root_hub_sampling() 1055 struct r8a66597_root_hub *rh = &r8a66597->root_hub[port]; r8a66597_usb_connect() local 1057 rh->port &= ~(USB_PORT_STAT_HIGH_SPEED | USB_PORT_STAT_LOW_SPEED); r8a66597_usb_connect() 1059 rh->port |= USB_PORT_STAT_HIGH_SPEED; r8a66597_usb_connect() 1061 rh->port |= USB_PORT_STAT_LOW_SPEED; r8a66597_usb_connect() 1063 rh->port &= ~USB_PORT_STAT_RESET; r8a66597_usb_connect() 1064 rh->port |= USB_PORT_STAT_ENABLE; r8a66597_usb_connect() 1701 struct r8a66597_root_hub *rh = &r8a66597->root_hub[port]; r8a66597_root_hub_control() local 1703 if (rh->port & USB_PORT_STAT_RESET) { r8a66597_root_hub_control() 1715 if (!(rh->port & USB_PORT_STAT_CONNECTION)) { r8a66597_root_hub_control() 1720 if (rh->scount > 0) { r8a66597_root_hub_control() 1722 if (tmp == rh->old_syssts) { r8a66597_root_hub_control() 1723 rh->scount--; r8a66597_root_hub_control() 1724 if (rh->scount == 0) r8a66597_root_hub_control() 1729 rh->scount = R8A66597_MAX_SAMPLING; r8a66597_root_hub_control() 1730 rh->old_syssts = tmp; r8a66597_root_hub_control() 2157 struct r8a66597_root_hub *rh = &r8a66597->root_hub[port]; r8a66597_hub_control() local 2182 rh->port &= ~USB_PORT_STAT_POWER; r8a66597_hub_control() 2198 rh->port &= ~(1 << wValue); r8a66597_hub_control() 2210 *(__le32 *)buf = cpu_to_le32(rh->port); r8a66597_hub_control() 2223 rh->port |= USB_PORT_STAT_POWER; r8a66597_hub_control() 2226 struct r8a66597_device *dev = rh->dev; r8a66597_hub_control() 2228 rh->port |= USB_PORT_STAT_RESET; r8a66597_hub_control() 2242 rh->port |= 1 << wValue; r8a66597_hub_control() 2263 struct r8a66597_root_hub *rh = &r8a66597->root_hub[port]; r8a66597_bus_suspend() local 2266 if (!(rh->port & USB_PORT_STAT_ENABLE)) r8a66597_bus_suspend() 2269 dev_dbg(&rh->dev->udev->dev, "suspend port = %d\n", port); r8a66597_bus_suspend() 2271 rh->port |= USB_PORT_STAT_SUSPEND; r8a66597_bus_suspend() 2273 if (rh->dev->udev->do_remote_wakeup) { r8a66597_bus_suspend() 2294 struct r8a66597_root_hub *rh = &r8a66597->root_hub[port]; r8a66597_bus_resume() local 2297 if (!(rh->port & USB_PORT_STAT_SUSPEND)) r8a66597_bus_resume() 2300 dev_dbg(&rh->dev->udev->dev, "resume port = %d\n", port); r8a66597_bus_resume() 2301 rh->port &= ~USB_PORT_STAT_SUSPEND; r8a66597_bus_resume() 2302 rh->port |= USB_PORT_STAT_C_SUSPEND << 16; r8a66597_bus_resume() 2361 struct r8a66597_root_hub *rh = &r8a66597->root_hub[port]; r8a66597_suspend() local 2363 rh->port = 0x00000000; r8a66597_suspend()
|
H A D | ohci-hub.c | 536 u32 rh = roothub_a (ohci); ohci_hub_descriptor() local 540 desc->bPwrOn2PwrGood = (rh & RH_A_POTPGT) >> 24; ohci_hub_descriptor() 548 if (rh & RH_A_NPS) /* no power switching? */ ohci_hub_descriptor() 550 if (rh & RH_A_PSM) /* per-port power switching? */ ohci_hub_descriptor() 552 if (rh & RH_A_NOCP) /* no overcurrent reporting? */ ohci_hub_descriptor() 554 else if (rh & RH_A_OCPM) /* per-port overcurrent reporting? */ ohci_hub_descriptor() 559 rh = roothub_b (ohci); ohci_hub_descriptor() 562 desc->u.hs.DeviceRemovable[0] = rh & RH_B_DR; ohci_hub_descriptor() 564 desc->u.hs.DeviceRemovable[1] = (rh & RH_B_DR) >> 8; ohci_hub_descriptor()
|
H A D | ohci-omap.c | 259 u32 rh = roothub_a (ohci); ohci_omap_reset() local 262 rh &= ~RH_A_NPS; ohci_omap_reset() 268 rh &= ~RH_A_NOCP; ohci_omap_reset() 278 ohci_writel(ohci, rh, &ohci->regs->roothub.a); ohci_omap_reset()
|
H A D | ohci-dbg.c | 107 "OHCI %d.%d, %s legacy support registers, rh state %s\n", ohci_dump_status()
|
H A D | ohci.h | 404 unsigned autostop:1; /* rh auto stopping/stopped */
|
H A D | ehci-dbg.c | 812 "EHCI %x.%02x, rh state %s\n", fill_registers_buffer()
|
H A D | fotg210-hcd.c | 684 "EHCI %x.%02x, rh state %s\n", fill_registers_buffer()
|
/linux-4.4.14/net/ipv4/ |
H A D | inetpeer.c | 247 int lh, rh; peer_avl_rebalance() local 255 rh = node_height(r); peer_avl_rebalance() 256 if (lh > rh + 1) { /* l: RH+2 */ peer_avl_rebalance() 275 node->avl_height = rh + 1; /* node: RH+1 */ peer_avl_rebalance() 278 l->avl_height = rh + 1; /* l: RH+1 */ peer_avl_rebalance() 281 lr->avl_height = rh + 2; peer_avl_rebalance() 284 } else if (rh > lh + 1) { /* r: LH+2 */ peer_avl_rebalance() 313 node->avl_height = (lh > rh ? lh : rh) + 1; peer_avl_rebalance()
|
/linux-4.4.14/drivers/scsi/snic/ |
H A D | vnic_dev.c | 85 struct vnic_resource_header __iomem *rh; vnic_dev_discover_res() local 98 rh = bar->vaddr; vnic_dev_discover_res() 99 if (!rh) { vnic_dev_discover_res() 105 if (ioread32(&rh->magic) != VNIC_RES_MAGIC || vnic_dev_discover_res() 106 ioread32(&rh->version) != VNIC_RES_VERSION) { vnic_dev_discover_res() 109 ioread32(&rh->magic), ioread32(&rh->version)); vnic_dev_discover_res() 114 r = (struct vnic_resource __iomem *)(rh + 1); vnic_dev_discover_res()
|
/linux-4.4.14/drivers/scsi/fnic/ |
H A D | vnic_dev.c | 66 struct vnic_resource_header __iomem *rh; vnic_dev_discover_res() local 75 rh = bar->vaddr; vnic_dev_discover_res() 76 if (!rh) { vnic_dev_discover_res() 81 if (ioread32(&rh->magic) != VNIC_RES_MAGIC || vnic_dev_discover_res() 82 ioread32(&rh->version) != VNIC_RES_VERSION) { vnic_dev_discover_res() 86 ioread32(&rh->magic), ioread32(&rh->version)); vnic_dev_discover_res() 90 r = (struct vnic_resource __iomem *)(rh + 1); vnic_dev_discover_res()
|
/linux-4.4.14/arch/powerpc/include/asm/ |
H A D | fsl_85xx_cache_sram.h | 40 rh_info_t *rh; member in struct:mpc85xx_cache_sram
|
/linux-4.4.14/net/ipv6/ |
H A D | exthdrs_core.c | 225 struct ipv6_rt_hdr _rh, *rh; ipv6_find_hdr() local 227 rh = skb_header_pointer(skb, start, sizeof(_rh), ipv6_find_hdr() 229 if (!rh) ipv6_find_hdr() 233 rh->segments_left == 0) ipv6_find_hdr()
|
/linux-4.4.14/arch/ia64/kernel/ |
H A D | salinfo.c | 414 sal_log_record_header_t *rh; salinfo_log_read_cpu() local 416 rh = (sal_log_record_header_t *)(data->log_buffer); salinfo_log_read_cpu() 418 if (rh->severity == sal_log_severity_corrected) salinfo_log_read_cpu() 435 sal_log_record_header_t *rh = (sal_log_record_header_t *)(data_saved->buffer); salinfo_log_new_read() local 437 memcpy(data->log_buffer, rh, data->log_size); salinfo_log_new_read() 439 if (rh->id == data_saved->id) { salinfo_log_new_read() 490 sal_log_record_header_t *rh; salinfo_log_clear() local 504 rh = (sal_log_record_header_t *)(data->log_buffer); salinfo_log_clear() 506 if (rh->severity != sal_log_severity_corrected) salinfo_log_clear()
|
H A D | mca.c | 458 sal_log_record_header_t *rh; ia64_mca_log_sal_error_record() local 477 rh = (sal_log_record_header_t *)buffer; ia64_mca_log_sal_error_record() 478 if (rh->severity == sal_log_severity_corrected) ia64_mca_log_sal_error_record() 1335 sal_log_record_header_t *rh = IA64_LOG_CURR_BUFFER(SAL_INFO_TYPE_MCA); ia64_mca_handler() local 1336 rh->severity = sal_log_severity_corrected; ia64_mca_handler()
|
/linux-4.4.14/drivers/scsi/lpfc/ |
H A D | lpfc_ct.c | 1411 struct lpfc_fdmi_reg_hba *rh; lpfc_fdmi_cmd() local 1478 rh = (struct lpfc_fdmi_reg_hba *)&CtReq->un.PortID; lpfc_fdmi_cmd() 1480 memcpy(&rh->hi.PortName, &vport->fc_sparam.portName, lpfc_fdmi_cmd() 1486 rh->rpl.EntryCnt = cpu_to_be32(1); lpfc_fdmi_cmd() 1487 memcpy(&rh->rpl.pe, &vport->fc_sparam.portName, lpfc_fdmi_cmd() 1497 ((uint8_t *)rh + size); lpfc_fdmi_cmd() 1506 ((uint8_t *)rh + size); lpfc_fdmi_cmd() 1522 ((uint8_t *)rh + size); lpfc_fdmi_cmd() 1540 ((uint8_t *)rh + size); lpfc_fdmi_cmd() 1558 ((uint8_t *)rh + size); lpfc_fdmi_cmd() 1575 ((uint8_t *)rh + size); lpfc_fdmi_cmd() 1592 ((uint8_t *)rh + size); lpfc_fdmi_cmd() 1619 ((uint8_t *)rh + size); lpfc_fdmi_cmd() 1637 ((uint8_t *)rh + size); lpfc_fdmi_cmd() 1655 ((uint8_t *)rh + size); lpfc_fdmi_cmd() 1673 ((uint8_t *)rh + size); lpfc_fdmi_cmd() 1694 ((uint8_t *)rh + size); lpfc_fdmi_cmd() 1715 ((uint8_t *)rh + size); lpfc_fdmi_cmd() 1739 rh = (struct lpfc_fdmi_reg_hba *) lpfc_fdmi_cmd() 1742 memcpy(&rh->hi.PortName, lpfc_fdmi_cmd() 1746 &rh->rpl.EntryCnt; lpfc_fdmi_cmd()
|
/linux-4.4.14/net/netfilter/ |
H A D | xt_connlimit.c | 107 union nf_inet_addr lh, rh; same_source_net() local 112 rh.ip6[i] = u3->ip6[i] & mask->ip6[i]; same_source_net() 115 return memcmp(&lh.ip6, &rh.ip6, sizeof(lh.ip6)); same_source_net()
|
/linux-4.4.14/include/math-emu/ |
H A D | op-2.h | 152 #define __FP_FRAC_ADD_2(rh, rl, xh, xl, yh, yl) \ 153 (rh = xh + yh + ((rl = xl + yl) < xl)) 156 #define __FP_FRAC_SUB_2(rh, rl, xh, xl, yh, yl) \ 157 (rh = xh - yh - ((rl = xl - yl) > xl))
|
/linux-4.4.14/drivers/net/ethernet/cisco/enic/ |
H A D | vnic_dev.c | 47 struct vnic_resource_header __iomem *rh; vnic_dev_discover_res() local 60 rh = bar->vaddr; vnic_dev_discover_res() 62 if (!rh) { vnic_dev_discover_res() 68 if ((ioread32(&rh->magic) != VNIC_RES_MAGIC) || vnic_dev_discover_res() 69 (ioread32(&rh->version) != VNIC_RES_VERSION)) { vnic_dev_discover_res() 75 ioread32(&rh->magic), ioread32(&rh->version)); vnic_dev_discover_res() 83 r = (struct vnic_resource __iomem *)(rh + 1); vnic_dev_discover_res()
|
/linux-4.4.14/arch/x86/kvm/ |
H A D | i8254.c | 63 u64 rl, rh; muldiv64() local 67 rh = (u64)u.l.high * (u64)b; muldiv64() 68 rh += (rl >> 32); muldiv64() 69 res.l.high = div64_u64(rh, c); muldiv64() 70 res.l.low = div64_u64(((mod_64(rh, c) << 32) + (rl & 0xffffffff)), c); muldiv64()
|
/linux-4.4.14/drivers/net/ethernet/broadcom/ |
H A D | b44.c | 662 struct rx_header *rh; b44_alloc_rx_skb() local 706 rh = (struct rx_header *) skb->data; b44_alloc_rx_skb() 708 rh->len = 0; b44_alloc_rx_skb() 709 rh->flags = 0; b44_alloc_rx_skb() 737 struct rx_header *rh; b44_recycle_rx() local 748 rh = (struct rx_header *) src_map->skb->data; b44_recycle_rx() 749 rh->len = 0; b44_recycle_rx() 750 rh->flags = 0; b44_recycle_rx() 793 struct rx_header *rh; b44_rx() local 799 rh = (struct rx_header *) skb->data; b44_rx() 800 len = le16_to_cpu(rh->len); b44_rx() 802 (rh->flags & cpu_to_le16(RX_FLAG_ERRORS))) { b44_rx() 816 len = le16_to_cpu(rh->len); b44_rx()
|
/linux-4.4.14/fs/nfsd/ |
H A D | nfscache.c | 348 struct list_head *rh = &b->lru_head; nfsd_cache_search() local 351 list_for_each_entry(rp, rh, c_lru) { list_for_each_entry()
|
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/subdev/fb/ |
H A D | nv50.c | 221 * cause IOMMU "read from address 0" errors (rh#561267) nv50_fb_init()
|
/linux-4.4.14/drivers/media/usb/uvc/ |
H A D | uvc_v4l2.c | 151 __u16 rw, rh; uvc_v4l2_try_format() local 186 rh = fmt->fmt.pix.height; uvc_v4l2_try_format() 193 d = min(w, rw) * min(h, rh); uvc_v4l2_try_format() 194 d = w*h + rw*rh - 2*d; uvc_v4l2_try_format()
|
/linux-4.4.14/include/trace/events/ |
H A D | kvm.h | 156 (__entry->address & (1<<3)) ? "|rh" : "")
|
/linux-4.4.14/drivers/usb/musb/ |
H A D | musb_virthub.c | 126 dev_dbg(musb->controller, "bogus rh suspend? %s\n", musb_port_suspend()
|
/linux-4.4.14/drivers/net/ethernet/microchip/ |
H A D | enc28j60.c | 301 int rl, rh; nolock_regw_read() local 305 rh = spi_read_op(priv, ENC28J60_READ_CTRL_REG, address + 1); nolock_regw_read() 307 return (rh << 8) | rl; nolock_regw_read()
|
/linux-4.4.14/arch/powerpc/lib/ |
H A D | rheap.c | 128 printk(KERN_ERR "rh: out of slots; crash is imminent.\n"); get_slot()
|
/linux-4.4.14/drivers/net/ethernet/sun/ |
H A D | niu.c | 3425 struct rx_pkt_hdr1 *rh; niu_process_rx_pkt() local 3488 len += sizeof(*rh); niu_process_rx_pkt() 3489 len = min_t(int, len, sizeof(*rh) + VLAN_ETH_HLEN); niu_process_rx_pkt() 3492 rh = (struct rx_pkt_hdr1 *) skb->data; niu_process_rx_pkt() 3495 ((u32)rh->hashval2_0 << 24 | niu_process_rx_pkt() 3496 (u32)rh->hashval2_1 << 16 | niu_process_rx_pkt() 3497 (u32)rh->hashval1_1 << 8 | niu_process_rx_pkt() 3498 (u32)rh->hashval1_2 << 0), niu_process_rx_pkt() 3500 skb_pull(skb, sizeof(*rh)); niu_process_rx_pkt()
|
/linux-4.4.14/kernel/rcu/ |
H A D | tree_plugin.h | 2877 struct rcu_head rh; member in struct:rcu_sysidle_head 2892 rshp = container_of(rhp, struct rcu_sysidle_head, rh); rcu_sysidle_cb() 2959 call_rcu(&rsh.rh, rcu_sysidle_cb);
|
/linux-4.4.14/drivers/gpu/drm/nouveau/ |
H A D | nouveau_connector.c | 366 * valid - it's not (rh#613284) nouveau_connector_detect_lvds()
|
/linux-4.4.14/include/linux/usb/ |
H A D | hcd.h | 112 #define HCD_FLAG_POLL_RH 2 /* poll for rh status? */
|
/linux-4.4.14/drivers/usb/core/ |
H A D | hcd.c | 786 dev_dbg (hcd->self.controller, "not queuing rh status urb\n"); rh_queue_status()
|
/linux-4.4.14/drivers/block/ |
H A D | floppy.c | 2311 pr_info("rh=%d h=%d\n", R_HEAD, HEAD); rw_interrupt()
|
/linux-4.4.14/fs/xfs/libxfs/ |
H A D | xfs_btree.c | 4031 /* now read rh sibling block for next iteration */ xfs_btree_block_change_owner()
|