h1 30 arch/alpha/kernel/console.c default_vga_hose_select(struct pci_controller *h1, struct pci_controller *h2) h1 32 arch/alpha/kernel/console.c if (h2->index < h1->index) h1 35 arch/alpha/kernel/console.c return h1; h1 122 arch/mips/boot/elf2ecoff.c const Elf32_Phdr *h1 = v1; h1 125 arch/mips/boot/elf2ecoff.c if (h1->p_vaddr > h2->p_vaddr) h1 127 arch/mips/boot/elf2ecoff.c else if (h1->p_vaddr < h2->p_vaddr) h1 97 crypto/poly1305_generic.c u32 h0, h1, h2, h3, h4; h1 115 crypto/poly1305_generic.c h1 = state->h[1]; h1 123 crypto/poly1305_generic.c h1 += (get_unaligned_le32(src + 3) >> 2) & 0x3ffffff; h1 129 crypto/poly1305_generic.c d0 = mlt(h0, r0) + mlt(h1, s4) + mlt(h2, s3) + h1 131 crypto/poly1305_generic.c d1 = mlt(h0, r1) + mlt(h1, r0) + mlt(h2, s4) + h1 133 crypto/poly1305_generic.c d2 = mlt(h0, r2) + mlt(h1, r1) + mlt(h2, r0) + h1 135 crypto/poly1305_generic.c d3 = mlt(h0, r3) + mlt(h1, r2) + mlt(h2, r1) + h1 137 crypto/poly1305_generic.c d4 = mlt(h0, r4) + mlt(h1, r3) + mlt(h2, r2) + h1 142 crypto/poly1305_generic.c d2 += sr(d1, 26); h1 = and(d1, 0x3ffffff); h1 146 crypto/poly1305_generic.c h1 += h0 >> 26; h0 = h0 & 0x3ffffff; h1 152 crypto/poly1305_generic.c state->h[1] = h1; h1 218 crypto/poly1305_generic.c u32 h0, h1, h2, h3, h4; h1 224 crypto/poly1305_generic.c h1 = state->h[1]; h1 229 crypto/poly1305_generic.c h2 += (h1 >> 26); h1 = h1 & 0x3ffffff; h1 233 crypto/poly1305_generic.c h1 += (h0 >> 26); h0 = h0 & 0x3ffffff; h1 237 crypto/poly1305_generic.c g1 = h1 + (g0 >> 26); g0 &= 0x3ffffff; h1 251 crypto/poly1305_generic.c h1 = (h1 & mask) | g1; h1 257 crypto/poly1305_generic.c put_unaligned_le32((h0 >> 0) | (h1 << 26), dst + 0); h1 258 crypto/poly1305_generic.c put_unaligned_le32((h1 >> 6) | (h2 << 20), dst + 4); h1 156 drivers/gpu/drm/amd/display/dc/dce110/dce110_timing_generator_v.c uint32_t h1 = 0; h1 163 drivers/gpu/drm/amd/display/dc/dce110/dce110_timing_generator_v.c h1 = get_reg_field_value( h1 185 drivers/gpu/drm/amd/display/dc/dce110/dce110_timing_generator_v.c if (h1 == h2 && v1 == v2) h1 473 drivers/gpu/drm/exynos/exynos_drm_fimc.c u32 cfg, h1, h2, v1, v2; h1 476 drivers/gpu/drm/exynos/exynos_drm_fimc.c h1 = buf->rect.x; h1 484 drivers/gpu/drm/exynos/exynos_drm_fimc.c DRM_DEV_DEBUG_KMS(ctx->dev, "h1[%d]h2[%d]v1[%d]v2[%d]\n", h1, h2, v1, h1 494 drivers/gpu/drm/exynos/exynos_drm_fimc.c cfg |= (EXYNOS_CIWDOFST_WINHOROFST(h1) | h1 6052 drivers/infiniband/hw/qib/qib_iba7322.c u32 pidx, unit, port, deflt, h1; h1 6093 drivers/infiniband/hw/qib/qib_iba7322.c h1 = 0; /* gcc thinks it might be used uninitted */ h1 6096 drivers/infiniband/hw/qib/qib_iba7322.c h1 = (u32)simple_strtoul(str, &nxt, 0); h1 6111 drivers/infiniband/hw/qib/qib_iba7322.c ppd->cpspec->h1_val = h1; h1 30 drivers/media/common/saa7146/saa7146_i2c.c int h1, h2; h1 58 drivers/media/common/saa7146/saa7146_i2c.c h1 = op_count/3; h2 = op_count%3; h1 59 drivers/media/common/saa7146/saa7146_i2c.c op[h1] |= cpu_to_le32( (u8)addr << ((3-h2)*8)); h1 60 drivers/media/common/saa7146/saa7146_i2c.c op[h1] |= cpu_to_le32(SAA7146_I2C_START << ((3-h2)*2)); h1 66 drivers/media/common/saa7146/saa7146_i2c.c h1 = op_count/3; h2 = op_count%3; h1 67 drivers/media/common/saa7146/saa7146_i2c.c op[h1] |= cpu_to_le32( (u32)((u8)m[i].buf[j]) << ((3-h2)*8)); h1 68 drivers/media/common/saa7146/saa7146_i2c.c op[h1] |= cpu_to_le32( SAA7146_I2C_CONT << ((3-h2)*2)); h1 76 drivers/media/common/saa7146/saa7146_i2c.c h1 = (op_count-1)/3; h2 = (op_count-1)%3; h1 77 drivers/media/common/saa7146/saa7146_i2c.c if ( SAA7146_I2C_CONT == (0x3 & (le32_to_cpu(op[h1]) >> ((3-h2)*2))) ) { h1 78 drivers/media/common/saa7146/saa7146_i2c.c op[h1] &= ~cpu_to_le32(0x2 << ((3-h2)*2)); h1 79 drivers/media/common/saa7146/saa7146_i2c.c op[h1] |= cpu_to_le32(SAA7146_I2C_STOP << ((3-h2)*2)); h1 33 drivers/mtd/inftlmount.c struct inftl_unittail h1; h1 90 drivers/mtd/inftlmount.c 8, &retlen,(char *)&h1); h1 540 drivers/mtd/inftlmount.c struct inftl_unittail h1; h1 597 drivers/mtd/inftlmount.c (char *)&h1) < 0) { h1 605 drivers/mtd/inftlmount.c erase_mark = le16_to_cpu((h1.EraseMark | h1.EraseMark1)); h1 27 drivers/mtd/nftlmount.c struct nftl_uci1 h1; h1 81 drivers/mtd/nftlmount.c (char *)&h1); h1 92 drivers/mtd/nftlmount.c if (le16_to_cpu(h1.EraseMark | h1.EraseMark1) != ERASE_MARK) { h1 95 drivers/mtd/nftlmount.c le16_to_cpu(h1.EraseMark), le16_to_cpu(h1.EraseMark1)); h1 494 drivers/mtd/nftlmount.c struct nftl_uci1 h1; h1 500 drivers/mtd/nftlmount.c &retlen, (char *)&h1) < 0) h1 503 drivers/mtd/nftlmount.c erase_mark = le16_to_cpu ((h1.EraseMark | h1.EraseMark1)); h1 511 drivers/mtd/nftlmount.c h1.EraseMark = cpu_to_le16(ERASE_MARK); h1 512 drivers/mtd/nftlmount.c h1.EraseMark1 = cpu_to_le16(ERASE_MARK); h1 513 drivers/mtd/nftlmount.c h1.WearInfo = cpu_to_le32(0); h1 516 drivers/mtd/nftlmount.c &retlen, (char *)&h1) < 0) h1 572 drivers/mtd/nftlmount.c struct nftl_uci1 h1; h1 604 drivers/mtd/nftlmount.c &retlen, (char *)&h1) < 0) { h1 612 drivers/mtd/nftlmount.c erase_mark = le16_to_cpu ((h1.EraseMark | h1.EraseMark1)); h1 381 drivers/s390/net/qeth_core.h static inline bool qeth_l2_same_vlan(struct qeth_hdr_layer2 *h1, h1 384 drivers/s390/net/qeth_core.h return !((h1->flags[2] ^ h2->flags[2]) & QETH_LAYER2_FLAG_VLAN) && h1 385 drivers/s390/net/qeth_core.h h1->vlan_id == h2->vlan_id; h1 388 drivers/s390/net/qeth_core.h static inline bool qeth_l3_iqd_same_vlan(struct qeth_hdr_layer3 *h1, h1 391 drivers/s390/net/qeth_core.h return !((h1->ext_flags ^ h2->ext_flags) & QETH_HDR_EXT_VLAN_FRAME) && h1 392 drivers/s390/net/qeth_core.h h1->vlan_id == h2->vlan_id; h1 395 drivers/s390/net/qeth_core.h static inline bool qeth_l3_same_next_hop(struct qeth_hdr_layer3 *h1, h1 398 drivers/s390/net/qeth_core.h return !((h1->flags ^ h2->flags) & QETH_HDR_IPV6) && h1 399 drivers/s390/net/qeth_core.h ipv6_addr_equal(&h1->next_hop.ipv6_addr, h1 376 drivers/tty/serial/ifx6x60.c u16 h1; h1 380 drivers/tty/serial/ifx6x60.c h1 = *in_buffer; h1 383 drivers/tty/serial/ifx6x60.c if (h1 == 0 && h2 == 0) { h1 387 drivers/tty/serial/ifx6x60.c } else if (h1 == 0xffff && h2 == 0xffff) { h1 393 drivers/tty/serial/ifx6x60.c *length = h1 & 0xfff; /* upper bits of byte are flags */ h1 31 fs/reiserfs/hashes.c b1 = h1; \ h1 41 fs/reiserfs/hashes.c h1 += b1; \ h1 48 fs/reiserfs/hashes.c u32 h0 = k[0], h1 = k[1]; h1 121 fs/reiserfs/hashes.c return h0 ^ h1; h1 40 include/keys/trusted.h unsigned int keylen, unsigned char *h1, h1 741 include/linux/if_vlan.h static inline unsigned long compare_vlan_header(const struct vlan_hdr *h1, h1 745 include/linux/if_vlan.h return *(u32 *)h1 ^ *(u32 *)h2; h1 747 include/linux/if_vlan.h return ((__force u32)h1->h_vlan_TCI ^ (__force u32)h2->h_vlan_TCI) | h1 748 include/linux/if_vlan.h ((__force u32)h1->h_vlan_encapsulated_proto ^ h1 139 include/linux/vmw_vmci_defs.h static inline bool vmci_handle_is_equal(struct vmci_handle h1, h1 142 include/linux/vmw_vmci_defs.h return h1.context == h2.context && h1.resource == h2.resource; h1 72 lib/test_hash.c u32 h0 = (u32)h64, h1, h2; h1 75 lib/test_hash.c hash_or[0][0] |= h1 = __hash_32(h0); h1 79 lib/test_hash.c if (h1 != h2) { h1 81 lib/test_hash.c h0, h1, h2); h1 92 lib/test_hash.c hash_or[0][k] |= h1 = hash_32(h0, k); h1 93 lib/test_hash.c if (h1 > m) { h1 94 lib/test_hash.c pr_err("hash_32(%#x, %d) = %#x > %#x", h0, k, h1, m); h1 100 lib/test_hash.c if (h1 != h2) { h1 102 lib/test_hash.c " = %#x", h0, k, h1, h2); h1 108 lib/test_hash.c h0, k, h1, m); h1 114 lib/test_hash.c hash_or[1][k] |= h1 = hash_64(h64, k); h1 115 lib/test_hash.c if (h1 > m) { h1 116 lib/test_hash.c pr_err("hash_64(%#llx, %d) = %#x > %#x", h64, k, h1, m); h1 122 lib/test_hash.c if (h1 != h2) { h1 124 lib/test_hash.c "= %#x", h64, k, h1, h2); h1 130 lib/test_hash.c h64, k, h1, m); h1 122 net/ipv6/ip6_gre.c unsigned int h1 = HASH_KEY(key); h1 131 net/ipv6/ip6_gre.c for_each_ip_tunnel_rcu(t, ign->tunnels_r_l[h0 ^ h1]) { h1 156 net/ipv6/ip6_gre.c for_each_ip_tunnel_rcu(t, ign->tunnels_r[h0 ^ h1]) { h1 180 net/ipv6/ip6_gre.c for_each_ip_tunnel_rcu(t, ign->tunnels_l[h1]) { h1 206 net/ipv6/ip6_gre.c for_each_ip_tunnel_rcu(t, ign->tunnels_wc[h1]) { h1 95 net/ipv6/sit.c unsigned int h1 = HASH(local); h1 100 net/ipv6/sit.c for_each_ip_tunnel_rcu(t, sitn->tunnels_r_l[h0 ^ h1]) { h1 115 net/ipv6/sit.c for_each_ip_tunnel_rcu(t, sitn->tunnels_l[h1]) { h1 113 net/netfilter/nf_conntrack_core.c static void nf_conntrack_double_unlock(unsigned int h1, unsigned int h2) h1 115 net/netfilter/nf_conntrack_core.c h1 %= CONNTRACK_LOCKS; h1 117 net/netfilter/nf_conntrack_core.c spin_unlock(&nf_conntrack_locks[h1]); h1 118 net/netfilter/nf_conntrack_core.c if (h1 != h2) h1 123 net/netfilter/nf_conntrack_core.c static bool nf_conntrack_double_lock(struct net *net, unsigned int h1, h1 126 net/netfilter/nf_conntrack_core.c h1 %= CONNTRACK_LOCKS; h1 128 net/netfilter/nf_conntrack_core.c if (h1 <= h2) { h1 129 net/netfilter/nf_conntrack_core.c nf_conntrack_lock(&nf_conntrack_locks[h1]); h1 130 net/netfilter/nf_conntrack_core.c if (h1 != h2) h1 135 net/netfilter/nf_conntrack_core.c spin_lock_nested(&nf_conntrack_locks[h1], h1 139 net/netfilter/nf_conntrack_core.c nf_conntrack_double_unlock(h1, h2); h1 158 net/packet/af_packet.c struct tpacket_hdr *h1; h1 368 net/packet/af_packet.c h.h1->tp_status = status; h1 369 net/packet/af_packet.c flush_dcache_page(pgv_to_page(&h.h1->tp_status)); h1 396 net/packet/af_packet.c flush_dcache_page(pgv_to_page(&h.h1->tp_status)); h1 397 net/packet/af_packet.c return h.h1->tp_status; h1 440 net/packet/af_packet.c h.h1->tp_sec = ts.tv_sec; h1 441 net/packet/af_packet.c h.h1->tp_usec = ts.tv_nsec / NSEC_PER_USEC; h1 457 net/packet/af_packet.c flush_dcache_page(pgv_to_page(&h.h1->tp_sec)); h1 759 net/packet/af_packet.c struct tpacket_hdr_v1 *h1 = &pbd1->hdr.bh1; h1 770 net/packet/af_packet.c h1->ts_last_pkt.ts_sec = last_pkt->tp_sec; h1 771 net/packet/af_packet.c h1->ts_last_pkt.ts_nsec = last_pkt->tp_nsec; h1 780 net/packet/af_packet.c h1->ts_last_pkt.ts_sec = ts.tv_sec; h1 781 net/packet/af_packet.c h1->ts_last_pkt.ts_nsec = ts.tv_nsec; h1 810 net/packet/af_packet.c struct tpacket_hdr_v1 *h1 = &pbd1->hdr.bh1; h1 824 net/packet/af_packet.c h1->ts_first_pkt.ts_sec = ts.tv_sec; h1 825 net/packet/af_packet.c h1->ts_first_pkt.ts_nsec = ts.tv_nsec; h1 2319 net/packet/af_packet.c h.h1->tp_len = skb->len; h1 2320 net/packet/af_packet.c h.h1->tp_snaplen = snaplen; h1 2321 net/packet/af_packet.c h.h1->tp_mac = macoff; h1 2322 net/packet/af_packet.c h.h1->tp_net = netoff; h1 2323 net/packet/af_packet.c h.h1->tp_sec = ts.tv_sec; h1 2324 net/packet/af_packet.c h.h1->tp_usec = ts.tv_nsec / NSEC_PER_USEC; h1 2325 net/packet/af_packet.c hdrlen = sizeof(*h.h1); h1 2571 net/packet/af_packet.c tp_len = ph.h1->tp_len; h1 2593 net/packet/af_packet.c off = ph.h1->tp_net; h1 2605 net/packet/af_packet.c off = ph.h1->tp_mac; h1 220 net/sched/cls_route.c unsigned int h1, h2; h1 222 net/sched/cls_route.c h1 = to_hash(handle); h1 223 net/sched/cls_route.c if (h1 > 256) h1 230 net/sched/cls_route.c b = rtnl_dereference(head->table[h1]); h1 279 net/sched/cls_route.c int h1, h2; h1 284 net/sched/cls_route.c for (h1 = 0; h1 <= 256; h1++) { h1 287 net/sched/cls_route.c b = rtnl_dereference(head->table[h1]); h1 304 net/sched/cls_route.c RCU_INIT_POINTER(head->table[h1], NULL); h1 320 net/sched/cls_route.c int i, h1; h1 364 net/sched/cls_route.c for (h1 = 0; h1 <= 256; h1++) { h1 365 net/sched/cls_route.c if (rcu_access_pointer(head->table[h1])) { h1 389 net/sched/cls_route.c unsigned int h1; h1 427 net/sched/cls_route.c h1 = to_hash(nhandle); h1 428 net/sched/cls_route.c b = rtnl_dereference(head->table[h1]); h1 434 net/sched/cls_route.c rcu_assign_pointer(head->table[h1], b); h1 565 net/sched/cls_route.c unsigned int h, h1; h1 574 net/sched/cls_route.c for (h1 = 0; h1 <= 32; h1++) { h1 577 net/sched/cls_route.c for (f = rtnl_dereference(b->ht[h1]); h1 133 net/sched/cls_rsvp.h unsigned int h1, h2; h1 167 net/sched/cls_rsvp.h h1 = hash_dst(dst, protocol, tunnelid); h1 170 net/sched/cls_rsvp.h for (s = rcu_dereference_bh(head->ht[h1]); s; h1 226 net/sched/cls_rsvp.h unsigned int h1 = h & 0xFF; h1 229 net/sched/cls_rsvp.h for (s = rtnl_dereference(head->ht[h1]); s; h1 252 net/sched/cls_rsvp.h unsigned int h1 = handle & 0xFF; h1 258 net/sched/cls_rsvp.h for (s = rtnl_dereference(head->ht[h1]); s; h1 315 net/sched/cls_rsvp.h int h1, h2; h1 320 net/sched/cls_rsvp.h for (h1 = 0; h1 < 256; h1++) { h1 323 net/sched/cls_rsvp.h while ((s = rtnl_dereference(data->ht[h1])) != NULL) { h1 324 net/sched/cls_rsvp.h RCU_INIT_POINTER(data->ht[h1], s->next); h1 349 net/sched/cls_rsvp.h int i, h1; h1 381 net/sched/cls_rsvp.h for (h1 = 0; h1 < 256; h1++) { h1 382 net/sched/cls_rsvp.h if (rcu_access_pointer(head->ht[h1])) { h1 423 net/sched/cls_rsvp.h int h1, h2; h1 427 net/sched/cls_rsvp.h for (h1 = 0; h1 < 256; h1++) { h1 429 net/sched/cls_rsvp.h for (s = rtnl_dereference(sht[h1]); s; h1 487 net/sched/cls_rsvp.h unsigned int h1, h2; h1 567 net/sched/cls_rsvp.h h1 = hash_dst(dst, pinfo ? pinfo->protocol : 0, pinfo ? pinfo->tunnelid : 0); h1 570 net/sched/cls_rsvp.h if ((f->handle = gen_handle(tp, h1 | (h2<<8))) == 0) h1 584 net/sched/cls_rsvp.h for (sp = &data->ht[h1]; h1 637 net/sched/cls_rsvp.h sp = &data->ht[h1]; h1 660 net/sched/cls_rsvp.h unsigned int h, h1; h1 670 net/sched/cls_rsvp.h for (h1 = 0; h1 <= 16; h1++) { h1 673 net/sched/cls_rsvp.h for (f = rtnl_dereference(s->ht[h1]); f; h1 123 security/keys/trusted.c unsigned int keylen, unsigned char *h1, h1 166 security/keys/trusted.c paramdigest, TPM_NONCE_SIZE, h1, h1 1171 sound/pci/asihpi/hpi_internal.h } h1; h1 230 sound/usb/card.c struct uac1_ac_header_descriptor *h1; h1 233 sound/usb/card.c h1 = snd_usb_find_csint_desc(host_iface->extra, h1 236 sound/usb/card.c if (!h1 || h1->bLength < sizeof(*h1)) { h1 242 sound/usb/card.c host_iface->extralen) - (void *)h1; h1 250 sound/usb/card.c if (rest_bytes < sizeof(*h1)) { h1 255 sound/usb/card.c if (!h1->bInCollection) { h1 260 sound/usb/card.c if (rest_bytes < h1->bLength) { h1 265 sound/usb/card.c if (h1->bLength < sizeof(*h1) + h1->bInCollection) { h1 270 sound/usb/card.c for (i = 0; i < h1->bInCollection; i++) h1 271 sound/usb/card.c snd_usb_create_stream(chip, ctrlif, h1->baInterfaceNr[i]); h1 82 tools/testing/selftests/net/psock_tpacket.c struct tpacket_hdr_v1 h1; h1 512 tools/testing/selftests/net/psock_tpacket.c if (__v3_prev_block_seq_num + 1 != pbd->h1.seq_num) { h1 516 tools/testing/selftests/net/psock_tpacket.c (uint64_t) pbd->h1.seq_num); h1 520 tools/testing/selftests/net/psock_tpacket.c __v3_prev_block_seq_num = pbd->h1.seq_num; h1 525 tools/testing/selftests/net/psock_tpacket.c if (pbd->h1.num_pkts && bytes != pbd->h1.blk_len) { h1 528 tools/testing/selftests/net/psock_tpacket.c pbd->h1.num_pkts, bytes, pbd->h1.blk_len); h1 535 tools/testing/selftests/net/psock_tpacket.c if ((pbd->h1.block_status & TP_STATUS_USER) == 0) { h1 545 tools/testing/selftests/net/psock_tpacket.c int num_pkts = pbd->h1.num_pkts, i; h1 552 tools/testing/selftests/net/psock_tpacket.c pbd->h1.offset_to_first_pkt); h1 577 tools/testing/selftests/net/psock_tpacket.c pbd->h1.block_status = TP_STATUS_KERNEL; h1 602 tools/testing/selftests/net/psock_tpacket.c while ((pbd->h1.block_status & TP_STATUS_USER) == 0)