gw 835 arch/s390/crypto/aes_s390.c static void gcm_walk_start(struct gcm_sg_walk *gw, struct scatterlist *sg, gw 838 arch/s390/crypto/aes_s390.c memset(gw, 0, sizeof(*gw)); gw 839 arch/s390/crypto/aes_s390.c gw->walk_bytes_remain = len; gw 840 arch/s390/crypto/aes_s390.c scatterwalk_start(&gw->walk, sg); gw 843 arch/s390/crypto/aes_s390.c static inline unsigned int _gcm_sg_clamp_and_map(struct gcm_sg_walk *gw) gw 847 arch/s390/crypto/aes_s390.c gw->walk_bytes = scatterwalk_clamp(&gw->walk, gw->walk_bytes_remain); gw 848 arch/s390/crypto/aes_s390.c while (!gw->walk_bytes) { gw 849 arch/s390/crypto/aes_s390.c nextsg = sg_next(gw->walk.sg); gw 852 arch/s390/crypto/aes_s390.c scatterwalk_start(&gw->walk, nextsg); gw 853 arch/s390/crypto/aes_s390.c gw->walk_bytes = scatterwalk_clamp(&gw->walk, gw 854 arch/s390/crypto/aes_s390.c gw->walk_bytes_remain); gw 856 arch/s390/crypto/aes_s390.c gw->walk_ptr = scatterwalk_map(&gw->walk); gw 857 arch/s390/crypto/aes_s390.c return gw->walk_bytes; gw 860 arch/s390/crypto/aes_s390.c static inline void _gcm_sg_unmap_and_advance(struct gcm_sg_walk *gw, gw 863 arch/s390/crypto/aes_s390.c gw->walk_bytes_remain -= nbytes; gw 864 arch/s390/crypto/aes_s390.c scatterwalk_unmap(&gw->walk); gw 865 arch/s390/crypto/aes_s390.c scatterwalk_advance(&gw->walk, nbytes); gw 866 arch/s390/crypto/aes_s390.c scatterwalk_done(&gw->walk, 0, gw->walk_bytes_remain); gw 867 arch/s390/crypto/aes_s390.c gw->walk_ptr = NULL; gw 870 arch/s390/crypto/aes_s390.c static int gcm_in_walk_go(struct gcm_sg_walk *gw, unsigned int minbytesneeded) gw 874 arch/s390/crypto/aes_s390.c if (gw->buf_bytes && gw->buf_bytes >= minbytesneeded) { gw 875 arch/s390/crypto/aes_s390.c gw->ptr = gw->buf; gw 876 arch/s390/crypto/aes_s390.c gw->nbytes = gw->buf_bytes; gw 880 arch/s390/crypto/aes_s390.c if (gw->walk_bytes_remain == 0) { gw 881 arch/s390/crypto/aes_s390.c gw->ptr = NULL; gw 882 arch/s390/crypto/aes_s390.c gw->nbytes = 0; gw 886 arch/s390/crypto/aes_s390.c if (!_gcm_sg_clamp_and_map(gw)) { gw 887 arch/s390/crypto/aes_s390.c gw->ptr = NULL; gw 888 arch/s390/crypto/aes_s390.c gw->nbytes = 0; gw 892 arch/s390/crypto/aes_s390.c if (!gw->buf_bytes && gw->walk_bytes >= minbytesneeded) { gw 893 arch/s390/crypto/aes_s390.c gw->ptr = gw->walk_ptr; gw 894 arch/s390/crypto/aes_s390.c gw->nbytes = gw->walk_bytes; gw 899 arch/s390/crypto/aes_s390.c n = min(gw->walk_bytes, AES_BLOCK_SIZE - gw->buf_bytes); gw 900 arch/s390/crypto/aes_s390.c memcpy(gw->buf + gw->buf_bytes, gw->walk_ptr, n); gw 901 arch/s390/crypto/aes_s390.c gw->buf_bytes += n; gw 902 arch/s390/crypto/aes_s390.c _gcm_sg_unmap_and_advance(gw, n); gw 903 arch/s390/crypto/aes_s390.c if (gw->buf_bytes >= minbytesneeded) { gw 904 arch/s390/crypto/aes_s390.c gw->ptr = gw->buf; gw 905 arch/s390/crypto/aes_s390.c gw->nbytes = gw->buf_bytes; gw 908 arch/s390/crypto/aes_s390.c if (!_gcm_sg_clamp_and_map(gw)) { gw 909 arch/s390/crypto/aes_s390.c gw->ptr = NULL; gw 910 arch/s390/crypto/aes_s390.c gw->nbytes = 0; gw 916 arch/s390/crypto/aes_s390.c return gw->nbytes; gw 919 arch/s390/crypto/aes_s390.c static int gcm_out_walk_go(struct gcm_sg_walk *gw, unsigned int minbytesneeded) gw 921 arch/s390/crypto/aes_s390.c if (gw->walk_bytes_remain == 0) { gw 922 arch/s390/crypto/aes_s390.c gw->ptr = NULL; gw 923 arch/s390/crypto/aes_s390.c gw->nbytes = 0; gw 927 arch/s390/crypto/aes_s390.c if (!_gcm_sg_clamp_and_map(gw)) { gw 928 arch/s390/crypto/aes_s390.c gw->ptr = NULL; gw 929 arch/s390/crypto/aes_s390.c gw->nbytes = 0; gw 933 arch/s390/crypto/aes_s390.c if (gw->walk_bytes >= minbytesneeded) { gw 934 arch/s390/crypto/aes_s390.c gw->ptr = gw->walk_ptr; gw 935 arch/s390/crypto/aes_s390.c gw->nbytes = gw->walk_bytes; gw 939 arch/s390/crypto/aes_s390.c scatterwalk_unmap(&gw->walk); gw 940 arch/s390/crypto/aes_s390.c gw->walk_ptr = NULL; gw 942 arch/s390/crypto/aes_s390.c gw->ptr = gw->buf; gw 943 arch/s390/crypto/aes_s390.c gw->nbytes = sizeof(gw->buf); gw 946 arch/s390/crypto/aes_s390.c return gw->nbytes; gw 949 arch/s390/crypto/aes_s390.c static int gcm_in_walk_done(struct gcm_sg_walk *gw, unsigned int bytesdone) gw 951 arch/s390/crypto/aes_s390.c if (gw->ptr == NULL) gw 954 arch/s390/crypto/aes_s390.c if (gw->ptr == gw->buf) { gw 955 arch/s390/crypto/aes_s390.c int n = gw->buf_bytes - bytesdone; gw 957 arch/s390/crypto/aes_s390.c memmove(gw->buf, gw->buf + bytesdone, n); gw 958 arch/s390/crypto/aes_s390.c gw->buf_bytes = n; gw 960 arch/s390/crypto/aes_s390.c gw->buf_bytes = 0; gw 962 arch/s390/crypto/aes_s390.c _gcm_sg_unmap_and_advance(gw, bytesdone); gw 967 arch/s390/crypto/aes_s390.c static int gcm_out_walk_done(struct gcm_sg_walk *gw, unsigned int bytesdone) gw 971 arch/s390/crypto/aes_s390.c if (gw->ptr == NULL) gw 974 arch/s390/crypto/aes_s390.c if (gw->ptr == gw->buf) { gw 976 arch/s390/crypto/aes_s390.c if (!_gcm_sg_clamp_and_map(gw)) gw 978 arch/s390/crypto/aes_s390.c n = min(gw->walk_bytes, bytesdone - i); gw 979 arch/s390/crypto/aes_s390.c memcpy(gw->walk_ptr, gw->buf + i, n); gw 980 arch/s390/crypto/aes_s390.c _gcm_sg_unmap_and_advance(gw, n); gw 983 arch/s390/crypto/aes_s390.c _gcm_sg_unmap_and_advance(gw, bytesdone); gw 676 arch/sparc/kernel/setup_64.c u32 cl, sv, gw; gw 680 arch/sparc/kernel/setup_64.c gw = prom_getintdefault (chosen, "gateway-ip", 0); gw 684 arch/sparc/kernel/setup_64.c if (gw) gw 685 arch/sparc/kernel/setup_64.c ic_gateway = gw; gw 556 arch/x86/kvm/paging_tmpl.h struct guest_walker *gw, int level) gw 559 arch/x86/kvm/paging_tmpl.h gpa_t base_gpa, pte_gpa = gw->pte_gpa[level - 1]; gw 569 arch/x86/kvm/paging_tmpl.h gw->prefetch_ptes, sizeof(gw->prefetch_ptes)); gw 570 arch/x86/kvm/paging_tmpl.h curr_pte = gw->prefetch_ptes[index]; gw 575 arch/x86/kvm/paging_tmpl.h return r || curr_pte != gw->ptes[level - 1]; gw 578 arch/x86/kvm/paging_tmpl.h static void FNAME(pte_prefetch)(struct kvm_vcpu *vcpu, struct guest_walker *gw, gw 582 arch/x86/kvm/paging_tmpl.h pt_element_t *gptep = gw->prefetch_ptes; gw 615 arch/x86/kvm/paging_tmpl.h struct guest_walker *gw, gw 622 arch/x86/kvm/paging_tmpl.h unsigned direct_access, access = gw->pt_access; gw 626 arch/x86/kvm/paging_tmpl.h direct_access = gw->pte_access; gw 637 arch/x86/kvm/paging_tmpl.h if (FNAME(gpte_changed)(vcpu, gw, top_level)) gw 644 arch/x86/kvm/paging_tmpl.h shadow_walk_okay(&it) && it.level > gw->level; gw 653 arch/x86/kvm/paging_tmpl.h table_gfn = gw->table_gfn[it.level - 2]; gw 662 arch/x86/kvm/paging_tmpl.h if (FNAME(gpte_changed)(vcpu, gw, it.level - 1)) gw 673 arch/x86/kvm/paging_tmpl.h gfn = gw->gfn | ((addr & PT_LVL_OFFSET_MASK(gw->level)) >> PAGE_SHIFT); gw 676 arch/x86/kvm/paging_tmpl.h trace_kvm_mmu_spte_requested(addr, gw->level, pfn); gw 704 arch/x86/kvm/paging_tmpl.h ret = mmu_set_spte(vcpu, it.sptep, gw->pte_access, write_fault, gw 706 arch/x86/kvm/paging_tmpl.h FNAME(pte_prefetch)(vcpu, gw, it.sptep); gw 37 drivers/gpio/gpio-gw-pld.c struct gw_pld *gw = gpiochip_get_data(gc); gw 39 drivers/gpio/gpio-gw-pld.c gw->out |= BIT(offset); gw 40 drivers/gpio/gpio-gw-pld.c return i2c_smbus_write_byte(gw->client, gw->out); gw 45 drivers/gpio/gpio-gw-pld.c struct gw_pld *gw = gpiochip_get_data(gc); gw 48 drivers/gpio/gpio-gw-pld.c val = i2c_smbus_read_byte(gw->client); gw 55 drivers/gpio/gpio-gw-pld.c struct gw_pld *gw = gpiochip_get_data(gc); gw 58 drivers/gpio/gpio-gw-pld.c gw->out |= BIT(offset); gw 60 drivers/gpio/gpio-gw-pld.c gw->out &= ~BIT(offset); gw 62 drivers/gpio/gpio-gw-pld.c return i2c_smbus_write_byte(gw->client, gw->out); gw 75 drivers/gpio/gpio-gw-pld.c struct gw_pld *gw; gw 78 drivers/gpio/gpio-gw-pld.c gw = devm_kzalloc(dev, sizeof(*gw), GFP_KERNEL); gw 79 drivers/gpio/gpio-gw-pld.c if (!gw) gw 82 drivers/gpio/gpio-gw-pld.c gw->chip.base = -1; gw 83 drivers/gpio/gpio-gw-pld.c gw->chip.can_sleep = true; gw 84 drivers/gpio/gpio-gw-pld.c gw->chip.parent = dev; gw 85 drivers/gpio/gpio-gw-pld.c gw->chip.of_node = np; gw 86 drivers/gpio/gpio-gw-pld.c gw->chip.owner = THIS_MODULE; gw 87 drivers/gpio/gpio-gw-pld.c gw->chip.label = dev_name(dev); gw 88 drivers/gpio/gpio-gw-pld.c gw->chip.ngpio = 8; gw 89 drivers/gpio/gpio-gw-pld.c gw->chip.direction_input = gw_pld_input8; gw 90 drivers/gpio/gpio-gw-pld.c gw->chip.get = gw_pld_get8; gw 91 drivers/gpio/gpio-gw-pld.c gw->chip.direction_output = gw_pld_output8; gw 92 drivers/gpio/gpio-gw-pld.c gw->chip.set = gw_pld_set8; gw 93 drivers/gpio/gpio-gw-pld.c gw->client = client; gw 101 drivers/gpio/gpio-gw-pld.c gw->out = 0xFF; gw 103 drivers/gpio/gpio-gw-pld.c i2c_set_clientdata(client, gw); gw 105 drivers/gpio/gpio-gw-pld.c ret = devm_gpiochip_add_data(dev, &gw->chip, gw); gw 2847 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c const struct in6_addr *gw, int ifindex, gw 2857 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c ipv6_addr_equal(gw, (struct in6_addr *) nh->gw_addr)) gw 2875 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c struct in6_addr *gw; gw 2880 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c gw = &fib6_nh->fib_nh_gw6; gw 2881 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c if (!mlxsw_sp_nexthop6_group_has_nexthop(nh_grp, gw, ifindex, gw 239 drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c union mlxsw_sp_l3addr gw, gw 247 drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c if (mlxsw_sp_l3addr_is_zero(gw)) gw 248 drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c gw = daddr; gw 250 drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c if (!edev || mlxsw_sp_span_dmac(tbl, &gw, edev, dmac)) gw 339 drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c union mlxsw_sp_l3addr gw = daddr; gw 351 drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c l3edev = mlxsw_sp_span_gretap4_route(to_dev, &saddr.addr4, &gw.addr4); gw 352 drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c return mlxsw_sp_span_entry_tunnel_parms_common(l3edev, saddr, daddr, gw, gw 440 drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c union mlxsw_sp_l3addr gw = daddr; gw 452 drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c l3edev = mlxsw_sp_span_gretap6_route(to_dev, &saddr.addr6, &gw.addr6); gw 453 drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c return mlxsw_sp_span_entry_tunnel_parms_common(l3edev, saddr, daddr, gw, gw 335 drivers/scsi/be2iscsi/be_iscsi.c u8 *ip = NULL, *subnet = NULL, *gw; gw 350 drivers/scsi/be2iscsi/be_iscsi.c gw = info->value; gw 351 drivers/scsi/be2iscsi/be_iscsi.c ret = beiscsi_if_set_gw(phba, BEISCSI_IP_TYPE_V4, gw); gw 434 drivers/scsi/be2iscsi/be_mgmt.c u32 action, u32 ip_type, u8 *gw) gw 449 drivers/scsi/be2iscsi/be_mgmt.c memcpy(req->ip_addr.addr, gw, gw 454 drivers/scsi/be2iscsi/be_mgmt.c int beiscsi_if_set_gw(struct beiscsi_hba *phba, u32 ip_type, u8 *gw) gw 477 drivers/scsi/be2iscsi/be_mgmt.c rt_val = beiscsi_if_mod_gw(phba, IP_ACTION_ADD, ip_type, gw); gw 639 drivers/scsi/be2iscsi/be_mgmt.c u8 *gw; gw 668 drivers/scsi/be2iscsi/be_mgmt.c gw = (u8 *)&gw_resp.ip_addr.addr; gw 669 drivers/scsi/be2iscsi/be_mgmt.c if (!beiscsi_if_zero_ip(gw, if_info->ip_addr.ip_type)) { gw 671 drivers/scsi/be2iscsi/be_mgmt.c if_info->ip_addr.ip_type, gw); gw 177 drivers/scsi/be2iscsi/be_mgmt.h int beiscsi_if_set_gw(struct beiscsi_hba *phba, u32 ip_type, u8 *gw); gw 1983 drivers/scsi/qedi/qedi_main.c char *fmt, *ip, *sub, *gw; gw 2000 drivers/scsi/qedi/qedi_main.c gw = ipv6_en ? initiator->ipv6.gateway.byte : gw 2007 drivers/scsi/qedi/qedi_main.c gw = ip; gw 2019 drivers/scsi/qedi/qedi_main.c rc = snprintf(buf, ip_len, fmt, gw); gw 335 include/net/ip6_route.h struct neighbour *ip6_neigh_lookup(const struct in6_addr *gw, gw 436 include/net/ip_fib.h int ip_fib_check_default(__be32 gw, struct net_device *dev); gw 9 include/net/netfilter/ipv4/nf_dup_ipv4.h const struct in_addr *gw, int oif); gw 8 include/net/netfilter/ipv6/nf_dup_ipv6.h const struct in6_addr *gw, int oif); gw 37 include/net/nexthop.h } gw; gw 35 include/trace/events/fib6.h __array( __u8, gw, 16 ) gw 73 include/trace/events/fib6.h in6 = (struct in6_addr *)__entry->gw; gw 77 include/trace/events/fib6.h in6 = (struct in6_addr *)__entry->gw; gw 86 include/trace/events/fib6.h __get_str(name), __entry->gw, __entry->err) gw 8 include/uapi/linux/netfilter/xt_TEE.h union nf_inet_addr gw; gw 953 include/video/pm3fb.h #define PM3DownloadGlyphwidth_GlyphWidth(gw) ((gw) & 0xffff) gw 2406 net/batman-adv/bat_iv_ogm.c atomic_set(&bat_priv->gw.sel_class, 20); gw 2422 net/batman-adv/bat_iv_ogm.c hlist_for_each_entry_rcu(gw_node, &bat_priv->gw.gateway_list, list) { gw 2438 net/batman-adv/bat_iv_ogm.c switch (atomic_read(&bat_priv->gw.sel_class)) { gw 2501 net/batman-adv/bat_iv_ogm.c if (atomic_read(&bat_priv->gw.sel_class) <= 2) gw 2536 net/batman-adv/bat_iv_ogm.c if ((atomic_read(&bat_priv->gw.sel_class) > 3) && gw 2537 net/batman-adv/bat_iv_ogm.c (orig_tq_avg - gw_tq_avg < atomic_read(&bat_priv->gw.sel_class))) gw 2610 net/batman-adv/bat_iv_ogm.c hlist_for_each_entry_rcu(gw_node, &bat_priv->gw.gateway_list, list) { gw 2715 net/batman-adv/bat_iv_ogm.c spin_lock_bh(&bat_priv->gw.list_lock); gw 2716 net/batman-adv/bat_iv_ogm.c cb->seq = bat_priv->gw.generation << 1 | 1; gw 2718 net/batman-adv/bat_iv_ogm.c hlist_for_each_entry(gw_node, &bat_priv->gw.gateway_list, list) { gw 2731 net/batman-adv/bat_iv_ogm.c spin_unlock_bh(&bat_priv->gw.list_lock); gw 2759 net/batman-adv/bat_iv_ogm.c .gw = { gw 666 net/batman-adv/bat_v.c atomic_set(&bat_priv->gw.sel_class, 50); gw 679 net/batman-adv/bat_v.c old_class = atomic_read(&bat_priv->gw.sel_class); gw 680 net/batman-adv/bat_v.c atomic_set(&bat_priv->gw.sel_class, class); gw 690 net/batman-adv/bat_v.c u32 class = atomic_read(&bat_priv->gw.sel_class); gw 751 net/batman-adv/bat_v.c hlist_for_each_entry_rcu(gw_node, &bat_priv->gw.gateway_list, list) { gw 792 net/batman-adv/bat_v.c threshold = atomic_read(&bat_priv->gw.sel_class); gw 890 net/batman-adv/bat_v.c hlist_for_each_entry_rcu(gw_node, &bat_priv->gw.gateway_list, list) { gw 1015 net/batman-adv/bat_v.c spin_lock_bh(&bat_priv->gw.list_lock); gw 1016 net/batman-adv/bat_v.c cb->seq = bat_priv->gw.generation << 1 | 1; gw 1018 net/batman-adv/bat_v.c hlist_for_each_entry(gw_node, &bat_priv->gw.gateway_list, list) { gw 1031 net/batman-adv/bat_v.c spin_unlock_bh(&bat_priv->gw.list_lock); gw 1060 net/batman-adv/bat_v.c .gw = { gw 95 net/batman-adv/gateway_client.c gw_node = rcu_dereference(bat_priv->gw.curr_gw); gw 144 net/batman-adv/gateway_client.c spin_lock_bh(&bat_priv->gw.list_lock); gw 149 net/batman-adv/gateway_client.c curr_gw_node = rcu_dereference_protected(bat_priv->gw.curr_gw, 1); gw 150 net/batman-adv/gateway_client.c rcu_assign_pointer(bat_priv->gw.curr_gw, new_gw_node); gw 155 net/batman-adv/gateway_client.c spin_unlock_bh(&bat_priv->gw.list_lock); gw 171 net/batman-adv/gateway_client.c atomic_set(&bat_priv->gw.reselect, 1); gw 186 net/batman-adv/gateway_client.c if (atomic_read(&bat_priv->gw.mode) != BATADV_GW_MODE_CLIENT) gw 218 net/batman-adv/gateway_client.c if (atomic_read(&bat_priv->gw.mode) != BATADV_GW_MODE_CLIENT) gw 221 net/batman-adv/gateway_client.c if (!bat_priv->algo_ops->gw.get_best_gw_node) gw 226 net/batman-adv/gateway_client.c if (!batadv_atomic_dec_not_zero(&bat_priv->gw.reselect) && curr_gw) gw 233 net/batman-adv/gateway_client.c next_gw = bat_priv->algo_ops->gw.get_best_gw_node(bat_priv); gw 311 net/batman-adv/gateway_client.c if (!bat_priv->algo_ops->gw.is_eligible) gw 322 net/batman-adv/gateway_client.c if (!bat_priv->algo_ops->gw.is_eligible(bat_priv, curr_gw_orig, gw 348 net/batman-adv/gateway_client.c lockdep_assert_held(&bat_priv->gw.list_lock); gw 365 net/batman-adv/gateway_client.c hlist_add_head_rcu(&gw_node->list, &bat_priv->gw.gateway_list); gw 366 net/batman-adv/gateway_client.c bat_priv->gw.generation++; gw 393 net/batman-adv/gateway_client.c hlist_for_each_entry_rcu(gw_node_tmp, &bat_priv->gw.gateway_list, gw 422 net/batman-adv/gateway_client.c spin_lock_bh(&bat_priv->gw.list_lock); gw 426 net/batman-adv/gateway_client.c spin_unlock_bh(&bat_priv->gw.list_lock); gw 429 net/batman-adv/gateway_client.c spin_unlock_bh(&bat_priv->gw.list_lock); gw 458 net/batman-adv/gateway_client.c spin_lock_bh(&bat_priv->gw.list_lock); gw 462 net/batman-adv/gateway_client.c bat_priv->gw.generation++; gw 464 net/batman-adv/gateway_client.c spin_unlock_bh(&bat_priv->gw.list_lock); gw 504 net/batman-adv/gateway_client.c spin_lock_bh(&bat_priv->gw.list_lock); gw 506 net/batman-adv/gateway_client.c &bat_priv->gw.gateway_list, list) { gw 509 net/batman-adv/gateway_client.c bat_priv->gw.generation++; gw 511 net/batman-adv/gateway_client.c spin_unlock_bh(&bat_priv->gw.list_lock); gw 540 net/batman-adv/gateway_client.c if (!bat_priv->algo_ops->gw.print) { gw 546 net/batman-adv/gateway_client.c bat_priv->algo_ops->gw.print(bat_priv, seq); gw 587 net/batman-adv/gateway_client.c if (!bat_priv->algo_ops->gw.dump) { gw 592 net/batman-adv/gateway_client.c bat_priv->algo_ops->gw.dump(msg, cb, bat_priv); gw 768 net/batman-adv/gateway_client.c switch (atomic_read(&bat_priv->gw.mode)) { gw 136 net/batman-adv/gateway_common.c struct batadv_tvlv_gateway_data gw; gw 140 net/batman-adv/gateway_common.c gw_mode = atomic_read(&bat_priv->gw.mode); gw 148 net/batman-adv/gateway_common.c down = atomic_read(&bat_priv->gw.bandwidth_down); gw 149 net/batman-adv/gateway_common.c up = atomic_read(&bat_priv->gw.bandwidth_up); gw 150 net/batman-adv/gateway_common.c gw.bandwidth_down = htonl(down); gw 151 net/batman-adv/gateway_common.c gw.bandwidth_up = htonl(up); gw 153 net/batman-adv/gateway_common.c &gw, sizeof(gw)); gw 177 net/batman-adv/gateway_common.c down_curr = (unsigned int)atomic_read(&bat_priv->gw.bandwidth_down); gw 178 net/batman-adv/gateway_common.c up_curr = (unsigned int)atomic_read(&bat_priv->gw.bandwidth_up); gw 202 net/batman-adv/gateway_common.c atomic_set(&bat_priv->gw.bandwidth_down, down_new); gw 203 net/batman-adv/gateway_common.c atomic_set(&bat_priv->gw.bandwidth_up, up_new); gw 246 net/batman-adv/gateway_common.c atomic_read(&bat_priv->gw.mode) == BATADV_GW_MODE_CLIENT) gw 256 net/batman-adv/gateway_common.c if (bat_priv->algo_ops->gw.init_sel_class) gw 257 net/batman-adv/gateway_common.c bat_priv->algo_ops->gw.init_sel_class(bat_priv); gw 259 net/batman-adv/gateway_common.c atomic_set(&bat_priv->gw.sel_class, 1); gw 169 net/batman-adv/main.c spin_lock_init(&bat_priv->gw.list_lock); gw 181 net/batman-adv/main.c INIT_HLIST_HEAD(&bat_priv->gw.gateway_list); gw 198 net/batman-adv/main.c bat_priv->gw.generation = 0; gw 227 net/batman-adv/main.c atomic_set(&bat_priv->gw.reselect, 0); gw 310 net/batman-adv/netlink.c atomic_read(&bat_priv->gw.bandwidth_down))) gw 314 net/batman-adv/netlink.c atomic_read(&bat_priv->gw.bandwidth_up))) gw 318 net/batman-adv/netlink.c atomic_read(&bat_priv->gw.mode))) gw 321 net/batman-adv/netlink.c if (bat_priv->algo_ops->gw.get_best_gw_node && gw 322 net/batman-adv/netlink.c bat_priv->algo_ops->gw.is_eligible) { gw 327 net/batman-adv/netlink.c atomic_read(&bat_priv->gw.sel_class))) gw 505 net/batman-adv/netlink.c atomic_set(&bat_priv->gw.bandwidth_down, nla_get_u32(attr)); gw 512 net/batman-adv/netlink.c atomic_set(&bat_priv->gw.bandwidth_up, nla_get_u32(attr)); gw 540 net/batman-adv/netlink.c atomic_set(&bat_priv->gw.mode, gw_mode); gw 546 net/batman-adv/netlink.c bat_priv->algo_ops->gw.get_best_gw_node && gw 547 net/batman-adv/netlink.c bat_priv->algo_ops->gw.is_eligible) { gw 558 net/batman-adv/netlink.c if (!bat_priv->algo_ops->gw.store_sel_class) gw 562 net/batman-adv/netlink.c atomic_set(&bat_priv->gw.sel_class, sel_class); gw 271 net/batman-adv/soft-interface.c gw_mode = atomic_read(&bat_priv->gw.mode); gw 781 net/batman-adv/soft-interface.c atomic_set(&bat_priv->gw.mode, BATADV_GW_MODE_OFF); gw 782 net/batman-adv/soft-interface.c atomic_set(&bat_priv->gw.bandwidth_down, 100); gw 783 net/batman-adv/soft-interface.c atomic_set(&bat_priv->gw.bandwidth_up, 20); gw 464 net/batman-adv/sysfs.c if (!bat_priv->algo_ops->gw.get_best_gw_node || gw 465 net/batman-adv/sysfs.c !bat_priv->algo_ops->gw.is_eligible) gw 468 net/batman-adv/sysfs.c switch (atomic_read(&bat_priv->gw.mode)) { gw 500 net/batman-adv/sysfs.c if (!bat_priv->algo_ops->gw.get_best_gw_node || gw 501 net/batman-adv/sysfs.c !bat_priv->algo_ops->gw.is_eligible) gw 526 net/batman-adv/sysfs.c if (atomic_read(&bat_priv->gw.mode) == gw_mode_tmp) gw 529 net/batman-adv/sysfs.c switch (atomic_read(&bat_priv->gw.mode)) { gw 558 net/batman-adv/sysfs.c atomic_set(&bat_priv->gw.mode, (unsigned int)gw_mode_tmp); gw 576 net/batman-adv/sysfs.c if (!bat_priv->algo_ops->gw.get_best_gw_node || gw 577 net/batman-adv/sysfs.c !bat_priv->algo_ops->gw.is_eligible) gw 580 net/batman-adv/sysfs.c if (bat_priv->algo_ops->gw.show_sel_class) gw 581 net/batman-adv/sysfs.c return bat_priv->algo_ops->gw.show_sel_class(bat_priv, buff); gw 583 net/batman-adv/sysfs.c return sprintf(buff, "%i\n", atomic_read(&bat_priv->gw.sel_class)); gw 598 net/batman-adv/sysfs.c if (!bat_priv->algo_ops->gw.get_best_gw_node || gw 599 net/batman-adv/sysfs.c !bat_priv->algo_ops->gw.is_eligible) gw 605 net/batman-adv/sysfs.c if (bat_priv->algo_ops->gw.store_sel_class) gw 606 net/batman-adv/sysfs.c return bat_priv->algo_ops->gw.store_sel_class(bat_priv, buff, gw 611 net/batman-adv/sysfs.c &bat_priv->gw.sel_class, gw 627 net/batman-adv/sysfs.c down = atomic_read(&bat_priv->gw.bandwidth_down); gw 628 net/batman-adv/sysfs.c up = atomic_read(&bat_priv->gw.bandwidth_up); gw 1734 net/batman-adv/types.h struct batadv_priv_gw gw; gw 2323 net/batman-adv/types.h struct batadv_algo_gw_ops gw; gw 33 net/bluetooth/6lowpan.c struct in6_addr gw; gw 172 net/bluetooth/6lowpan.c if (ipv6_addr_any(&lowpan_cb(skb)->gw)) { gw 180 net/bluetooth/6lowpan.c nexthop = &lowpan_cb(skb)->gw; gw 189 net/bluetooth/6lowpan.c memcpy(&lowpan_cb(skb)->gw, nexthop, sizeof(struct in6_addr)); gw 253 net/decnet/dn_table.c __le16 gw; gw 263 net/decnet/dn_table.c gw = gw_attr ? nla_get_le16(gw_attr) : 0; gw 265 net/decnet/dn_table.c if (gw && gw != nh->nh_gw) gw 431 net/ipv4/fib_semantics.c int ip_fib_check_default(__be32 gw, struct net_device *dev) gw 443 net/ipv4/fib_semantics.c nh->fib_nh_gw4 == gw && gw 2811 net/ipv4/fib_trie.c __be32 gw = 0; gw 2814 net/ipv4/fib_trie.c gw = nhc->nhc_gw.ipv4; gw 2820 net/ipv4/fib_trie.c prefix, gw, flags, 0, 0, gw 24 net/ipv4/netfilter/nf_dup_ipv4.c const struct in_addr *gw, int oif) gw 34 net/ipv4/netfilter/nf_dup_ipv4.c fl4.daddr = gw->s_addr; gw 51 net/ipv4/netfilter/nf_dup_ipv4.c const struct in_addr *gw, int oif) gw 86 net/ipv4/netfilter/nf_dup_ipv4.c if (nf_dup_ipv4_route(net, skb, gw, oif)) { gw 25 net/ipv4/netfilter/nft_dup_ipv4.c struct in_addr gw = { gw 30 net/ipv4/netfilter/nft_dup_ipv4.c nf_dup_ipv4(nft_net(pkt), pkt->skb, nft_hook(pkt), &gw, oif); gw 1174 net/ipv4/nexthop.c .fc_gw4 = cfg->gw.ipv4, gw 1175 net/ipv4/nexthop.c .fc_gw_family = cfg->gw.ipv4 ? AF_INET : 0, gw 1210 net/ipv4/nexthop.c .fc_gateway = cfg->gw.ipv6, gw 1217 net/ipv4/nexthop.c if (!ipv6_addr_any(&cfg->gw.ipv6)) gw 1445 net/ipv4/nexthop.c cfg->gw.ipv4 = nla_get_be32(gwa); gw 1452 net/ipv4/nexthop.c cfg->gw.ipv6 = nla_get_in6_addr(gwa); gw 644 net/ipv4/route.c __be32 gw, u32 pmtu, bool lock, gw 680 net/ipv4/route.c if (gw) gw 681 net/ipv4/route.c fnhe->fnhe_gw = gw; gw 707 net/ipv4/route.c fnhe->fnhe_gw = gw; gw 919 net/ipv4/route.c __be32 gw = rt_nexthop(rt, ip_hdr(skb)->daddr); gw 921 net/ipv4/route.c icmp_send(skb, ICMP_REDIRECT, ICMP_REDIR_HOST, gw); gw 929 net/ipv4/route.c &ip_hdr(skb)->daddr, &gw); gw 1814 net/ipv4/route.c __be32 gw; gw 1816 net/ipv4/route.c gw = nhc->nhc_gw_family == AF_INET ? nhc->nhc_gw.ipv4 : 0; gw 1818 net/ipv4/route.c inet_addr_onlink(out_dev, saddr, gw)) gw 20 net/ipv6/netfilter/nf_dup_ipv6.c const struct in6_addr *gw, int oif) gw 30 net/ipv6/netfilter/nf_dup_ipv6.c fl6.daddr = *gw; gw 48 net/ipv6/netfilter/nf_dup_ipv6.c const struct in6_addr *gw, int oif) gw 65 net/ipv6/netfilter/nf_dup_ipv6.c if (nf_dup_ipv6_route(net, skb, gw, oif)) { gw 25 net/ipv6/netfilter/nft_dup_ipv6.c struct in6_addr *gw = (struct in6_addr *)®s->data[priv->sreg_addr]; gw 28 net/ipv6/netfilter/nft_dup_ipv6.c nf_dup_ipv6(nft_net(pkt), pkt->skb, nft_hook(pkt), gw, oif); gw 200 net/ipv6/route.c struct neighbour *ip6_neigh_lookup(const struct in6_addr *gw, gw 207 net/ipv6/route.c daddr = choose_neigh_daddr(gw, skb, daddr); gw 1941 net/ipv6/route.c const struct in6_addr *gw; gw 1951 net/ipv6/route.c (arg->gw && !nh->fib_nh_gw_family) || gw 1952 net/ipv6/route.c (!arg->gw && nh->fib_nh_gw_family) || gw 1953 net/ipv6/route.c (arg->gw && !ipv6_addr_equal(arg->gw, &nh->fib_nh_gw6))) gw 1976 net/ipv6/route.c .gw = &rt->rt6i_gateway, gw 2774 net/ipv6/route.c .gw = &rt6->rt6i_gateway, gw 2872 net/ipv6/route.c const struct in6_addr *gw, gw 2886 net/ipv6/route.c if (!ipv6_addr_equal(gw, &nh->fib_nh_gw6)) { gw 2891 net/ipv6/route.c ipv6_addr_equal(gw, &rt_cache->rt6i_gateway)) { gw 2903 net/ipv6/route.c const struct in6_addr *gw; gw 2912 net/ipv6/route.c return ip6_redirect_nh_match(arg->res, arg->fl6, arg->gw, arg->ret); gw 2933 net/ipv6/route.c .gw = &rdfl->gateway, gw 4073 net/ipv6/route.c .gw = &rt->rt6i_gateway, gw 42 net/netfilter/xt_TEE.c nf_dup_ipv4(xt_net(par), skb, xt_hooknum(par), &info->gw.in, oif); gw 54 net/netfilter/xt_TEE.c nf_dup_ipv6(xt_net(par), skb, xt_hooknum(par), &info->gw.in6, oif); gw 99 net/netfilter/xt_TEE.c if (memcmp(&info->gw, &tee_zero_address, gw 24 samples/bpf/xdp_router_ipv4_kern.c __be32 gw; gw 166 samples/bpf/xdp_router_ipv4_kern.c if (!prefix_value->gw) gw 168 samples/bpf/xdp_router_ipv4_kern.c dest_ip = prefix_value->gw; gw 134 samples/bpf/xdp_router_ipv4_user.c __be32 dst, gw; gw 192 samples/bpf/xdp_router_ipv4_user.c route.gw = atoi(gws); gw 208 samples/bpf/xdp_router_ipv4_user.c __be32 gw; gw 242 samples/bpf/xdp_router_ipv4_user.c route.gw, route.dst_len, gw 251 samples/bpf/xdp_router_ipv4_user.c prefix_value->gw = route.gw; gw 293 samples/bpf/xdp_router_ipv4_user.c prefix_value->gw = route.gw;