gw                835 arch/s390/crypto/aes_s390.c static void gcm_walk_start(struct gcm_sg_walk *gw, struct scatterlist *sg,
gw                838 arch/s390/crypto/aes_s390.c 	memset(gw, 0, sizeof(*gw));
gw                839 arch/s390/crypto/aes_s390.c 	gw->walk_bytes_remain = len;
gw                840 arch/s390/crypto/aes_s390.c 	scatterwalk_start(&gw->walk, sg);
gw                843 arch/s390/crypto/aes_s390.c static inline unsigned int _gcm_sg_clamp_and_map(struct gcm_sg_walk *gw)
gw                847 arch/s390/crypto/aes_s390.c 	gw->walk_bytes = scatterwalk_clamp(&gw->walk, gw->walk_bytes_remain);
gw                848 arch/s390/crypto/aes_s390.c 	while (!gw->walk_bytes) {
gw                849 arch/s390/crypto/aes_s390.c 		nextsg = sg_next(gw->walk.sg);
gw                852 arch/s390/crypto/aes_s390.c 		scatterwalk_start(&gw->walk, nextsg);
gw                853 arch/s390/crypto/aes_s390.c 		gw->walk_bytes = scatterwalk_clamp(&gw->walk,
gw                854 arch/s390/crypto/aes_s390.c 						   gw->walk_bytes_remain);
gw                856 arch/s390/crypto/aes_s390.c 	gw->walk_ptr = scatterwalk_map(&gw->walk);
gw                857 arch/s390/crypto/aes_s390.c 	return gw->walk_bytes;
gw                860 arch/s390/crypto/aes_s390.c static inline void _gcm_sg_unmap_and_advance(struct gcm_sg_walk *gw,
gw                863 arch/s390/crypto/aes_s390.c 	gw->walk_bytes_remain -= nbytes;
gw                864 arch/s390/crypto/aes_s390.c 	scatterwalk_unmap(&gw->walk);
gw                865 arch/s390/crypto/aes_s390.c 	scatterwalk_advance(&gw->walk, nbytes);
gw                866 arch/s390/crypto/aes_s390.c 	scatterwalk_done(&gw->walk, 0, gw->walk_bytes_remain);
gw                867 arch/s390/crypto/aes_s390.c 	gw->walk_ptr = NULL;
gw                870 arch/s390/crypto/aes_s390.c static int gcm_in_walk_go(struct gcm_sg_walk *gw, unsigned int minbytesneeded)
gw                874 arch/s390/crypto/aes_s390.c 	if (gw->buf_bytes && gw->buf_bytes >= minbytesneeded) {
gw                875 arch/s390/crypto/aes_s390.c 		gw->ptr = gw->buf;
gw                876 arch/s390/crypto/aes_s390.c 		gw->nbytes = gw->buf_bytes;
gw                880 arch/s390/crypto/aes_s390.c 	if (gw->walk_bytes_remain == 0) {
gw                881 arch/s390/crypto/aes_s390.c 		gw->ptr = NULL;
gw                882 arch/s390/crypto/aes_s390.c 		gw->nbytes = 0;
gw                886 arch/s390/crypto/aes_s390.c 	if (!_gcm_sg_clamp_and_map(gw)) {
gw                887 arch/s390/crypto/aes_s390.c 		gw->ptr = NULL;
gw                888 arch/s390/crypto/aes_s390.c 		gw->nbytes = 0;
gw                892 arch/s390/crypto/aes_s390.c 	if (!gw->buf_bytes && gw->walk_bytes >= minbytesneeded) {
gw                893 arch/s390/crypto/aes_s390.c 		gw->ptr = gw->walk_ptr;
gw                894 arch/s390/crypto/aes_s390.c 		gw->nbytes = gw->walk_bytes;
gw                899 arch/s390/crypto/aes_s390.c 		n = min(gw->walk_bytes, AES_BLOCK_SIZE - gw->buf_bytes);
gw                900 arch/s390/crypto/aes_s390.c 		memcpy(gw->buf + gw->buf_bytes, gw->walk_ptr, n);
gw                901 arch/s390/crypto/aes_s390.c 		gw->buf_bytes += n;
gw                902 arch/s390/crypto/aes_s390.c 		_gcm_sg_unmap_and_advance(gw, n);
gw                903 arch/s390/crypto/aes_s390.c 		if (gw->buf_bytes >= minbytesneeded) {
gw                904 arch/s390/crypto/aes_s390.c 			gw->ptr = gw->buf;
gw                905 arch/s390/crypto/aes_s390.c 			gw->nbytes = gw->buf_bytes;
gw                908 arch/s390/crypto/aes_s390.c 		if (!_gcm_sg_clamp_and_map(gw)) {
gw                909 arch/s390/crypto/aes_s390.c 			gw->ptr = NULL;
gw                910 arch/s390/crypto/aes_s390.c 			gw->nbytes = 0;
gw                916 arch/s390/crypto/aes_s390.c 	return gw->nbytes;
gw                919 arch/s390/crypto/aes_s390.c static int gcm_out_walk_go(struct gcm_sg_walk *gw, unsigned int minbytesneeded)
gw                921 arch/s390/crypto/aes_s390.c 	if (gw->walk_bytes_remain == 0) {
gw                922 arch/s390/crypto/aes_s390.c 		gw->ptr = NULL;
gw                923 arch/s390/crypto/aes_s390.c 		gw->nbytes = 0;
gw                927 arch/s390/crypto/aes_s390.c 	if (!_gcm_sg_clamp_and_map(gw)) {
gw                928 arch/s390/crypto/aes_s390.c 		gw->ptr = NULL;
gw                929 arch/s390/crypto/aes_s390.c 		gw->nbytes = 0;
gw                933 arch/s390/crypto/aes_s390.c 	if (gw->walk_bytes >= minbytesneeded) {
gw                934 arch/s390/crypto/aes_s390.c 		gw->ptr = gw->walk_ptr;
gw                935 arch/s390/crypto/aes_s390.c 		gw->nbytes = gw->walk_bytes;
gw                939 arch/s390/crypto/aes_s390.c 	scatterwalk_unmap(&gw->walk);
gw                940 arch/s390/crypto/aes_s390.c 	gw->walk_ptr = NULL;
gw                942 arch/s390/crypto/aes_s390.c 	gw->ptr = gw->buf;
gw                943 arch/s390/crypto/aes_s390.c 	gw->nbytes = sizeof(gw->buf);
gw                946 arch/s390/crypto/aes_s390.c 	return gw->nbytes;
gw                949 arch/s390/crypto/aes_s390.c static int gcm_in_walk_done(struct gcm_sg_walk *gw, unsigned int bytesdone)
gw                951 arch/s390/crypto/aes_s390.c 	if (gw->ptr == NULL)
gw                954 arch/s390/crypto/aes_s390.c 	if (gw->ptr == gw->buf) {
gw                955 arch/s390/crypto/aes_s390.c 		int n = gw->buf_bytes - bytesdone;
gw                957 arch/s390/crypto/aes_s390.c 			memmove(gw->buf, gw->buf + bytesdone, n);
gw                958 arch/s390/crypto/aes_s390.c 			gw->buf_bytes = n;
gw                960 arch/s390/crypto/aes_s390.c 			gw->buf_bytes = 0;
gw                962 arch/s390/crypto/aes_s390.c 		_gcm_sg_unmap_and_advance(gw, bytesdone);
gw                967 arch/s390/crypto/aes_s390.c static int gcm_out_walk_done(struct gcm_sg_walk *gw, unsigned int bytesdone)
gw                971 arch/s390/crypto/aes_s390.c 	if (gw->ptr == NULL)
gw                974 arch/s390/crypto/aes_s390.c 	if (gw->ptr == gw->buf) {
gw                976 arch/s390/crypto/aes_s390.c 			if (!_gcm_sg_clamp_and_map(gw))
gw                978 arch/s390/crypto/aes_s390.c 			n = min(gw->walk_bytes, bytesdone - i);
gw                979 arch/s390/crypto/aes_s390.c 			memcpy(gw->walk_ptr, gw->buf + i, n);
gw                980 arch/s390/crypto/aes_s390.c 			_gcm_sg_unmap_and_advance(gw, n);
gw                983 arch/s390/crypto/aes_s390.c 		_gcm_sg_unmap_and_advance(gw, bytesdone);
gw                676 arch/sparc/kernel/setup_64.c 		u32 cl, sv, gw;
gw                680 arch/sparc/kernel/setup_64.c 		gw = prom_getintdefault (chosen, "gateway-ip", 0);
gw                684 arch/sparc/kernel/setup_64.c 			if (gw)
gw                685 arch/sparc/kernel/setup_64.c 				ic_gateway = gw;
gw                556 arch/x86/kvm/paging_tmpl.h 				struct guest_walker *gw, int level)
gw                559 arch/x86/kvm/paging_tmpl.h 	gpa_t base_gpa, pte_gpa = gw->pte_gpa[level - 1];
gw                569 arch/x86/kvm/paging_tmpl.h 				gw->prefetch_ptes, sizeof(gw->prefetch_ptes));
gw                570 arch/x86/kvm/paging_tmpl.h 		curr_pte = gw->prefetch_ptes[index];
gw                575 arch/x86/kvm/paging_tmpl.h 	return r || curr_pte != gw->ptes[level - 1];
gw                578 arch/x86/kvm/paging_tmpl.h static void FNAME(pte_prefetch)(struct kvm_vcpu *vcpu, struct guest_walker *gw,
gw                582 arch/x86/kvm/paging_tmpl.h 	pt_element_t *gptep = gw->prefetch_ptes;
gw                615 arch/x86/kvm/paging_tmpl.h 			 struct guest_walker *gw,
gw                622 arch/x86/kvm/paging_tmpl.h 	unsigned direct_access, access = gw->pt_access;
gw                626 arch/x86/kvm/paging_tmpl.h 	direct_access = gw->pte_access;
gw                637 arch/x86/kvm/paging_tmpl.h 	if (FNAME(gpte_changed)(vcpu, gw, top_level))
gw                644 arch/x86/kvm/paging_tmpl.h 	     shadow_walk_okay(&it) && it.level > gw->level;
gw                653 arch/x86/kvm/paging_tmpl.h 			table_gfn = gw->table_gfn[it.level - 2];
gw                662 arch/x86/kvm/paging_tmpl.h 		if (FNAME(gpte_changed)(vcpu, gw, it.level - 1))
gw                673 arch/x86/kvm/paging_tmpl.h 	gfn = gw->gfn | ((addr & PT_LVL_OFFSET_MASK(gw->level)) >> PAGE_SHIFT);
gw                676 arch/x86/kvm/paging_tmpl.h 	trace_kvm_mmu_spte_requested(addr, gw->level, pfn);
gw                704 arch/x86/kvm/paging_tmpl.h 	ret = mmu_set_spte(vcpu, it.sptep, gw->pte_access, write_fault,
gw                706 arch/x86/kvm/paging_tmpl.h 	FNAME(pte_prefetch)(vcpu, gw, it.sptep);
gw                 37 drivers/gpio/gpio-gw-pld.c 	struct gw_pld *gw = gpiochip_get_data(gc);
gw                 39 drivers/gpio/gpio-gw-pld.c 	gw->out |= BIT(offset);
gw                 40 drivers/gpio/gpio-gw-pld.c 	return i2c_smbus_write_byte(gw->client, gw->out);
gw                 45 drivers/gpio/gpio-gw-pld.c 	struct gw_pld *gw = gpiochip_get_data(gc);
gw                 48 drivers/gpio/gpio-gw-pld.c 	val = i2c_smbus_read_byte(gw->client);
gw                 55 drivers/gpio/gpio-gw-pld.c 	struct gw_pld *gw = gpiochip_get_data(gc);
gw                 58 drivers/gpio/gpio-gw-pld.c 		gw->out |= BIT(offset);
gw                 60 drivers/gpio/gpio-gw-pld.c 		gw->out &= ~BIT(offset);
gw                 62 drivers/gpio/gpio-gw-pld.c 	return i2c_smbus_write_byte(gw->client, gw->out);
gw                 75 drivers/gpio/gpio-gw-pld.c 	struct gw_pld *gw;
gw                 78 drivers/gpio/gpio-gw-pld.c 	gw = devm_kzalloc(dev, sizeof(*gw), GFP_KERNEL);
gw                 79 drivers/gpio/gpio-gw-pld.c 	if (!gw)
gw                 82 drivers/gpio/gpio-gw-pld.c 	gw->chip.base = -1;
gw                 83 drivers/gpio/gpio-gw-pld.c 	gw->chip.can_sleep = true;
gw                 84 drivers/gpio/gpio-gw-pld.c 	gw->chip.parent = dev;
gw                 85 drivers/gpio/gpio-gw-pld.c 	gw->chip.of_node = np;
gw                 86 drivers/gpio/gpio-gw-pld.c 	gw->chip.owner = THIS_MODULE;
gw                 87 drivers/gpio/gpio-gw-pld.c 	gw->chip.label = dev_name(dev);
gw                 88 drivers/gpio/gpio-gw-pld.c 	gw->chip.ngpio = 8;
gw                 89 drivers/gpio/gpio-gw-pld.c 	gw->chip.direction_input = gw_pld_input8;
gw                 90 drivers/gpio/gpio-gw-pld.c 	gw->chip.get = gw_pld_get8;
gw                 91 drivers/gpio/gpio-gw-pld.c 	gw->chip.direction_output = gw_pld_output8;
gw                 92 drivers/gpio/gpio-gw-pld.c 	gw->chip.set = gw_pld_set8;
gw                 93 drivers/gpio/gpio-gw-pld.c 	gw->client = client;
gw                101 drivers/gpio/gpio-gw-pld.c 	gw->out = 0xFF;
gw                103 drivers/gpio/gpio-gw-pld.c 	i2c_set_clientdata(client, gw);
gw                105 drivers/gpio/gpio-gw-pld.c 	ret = devm_gpiochip_add_data(dev, &gw->chip, gw);
gw               2847 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c 				    const struct in6_addr *gw, int ifindex,
gw               2857 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c 		    ipv6_addr_equal(gw, (struct in6_addr *) nh->gw_addr))
gw               2875 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c 		struct in6_addr *gw;
gw               2880 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c 		gw = &fib6_nh->fib_nh_gw6;
gw               2881 drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c 		if (!mlxsw_sp_nexthop6_group_has_nexthop(nh_grp, gw, ifindex,
gw                239 drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c 					union mlxsw_sp_l3addr gw,
gw                247 drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c 	if (mlxsw_sp_l3addr_is_zero(gw))
gw                248 drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c 		gw = daddr;
gw                250 drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c 	if (!edev || mlxsw_sp_span_dmac(tbl, &gw, edev, dmac))
gw                339 drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c 	union mlxsw_sp_l3addr gw = daddr;
gw                351 drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c 	l3edev = mlxsw_sp_span_gretap4_route(to_dev, &saddr.addr4, &gw.addr4);
gw                352 drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c 	return mlxsw_sp_span_entry_tunnel_parms_common(l3edev, saddr, daddr, gw,
gw                440 drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c 	union mlxsw_sp_l3addr gw = daddr;
gw                452 drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c 	l3edev = mlxsw_sp_span_gretap6_route(to_dev, &saddr.addr6, &gw.addr6);
gw                453 drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c 	return mlxsw_sp_span_entry_tunnel_parms_common(l3edev, saddr, daddr, gw,
gw                335 drivers/scsi/be2iscsi/be_iscsi.c 	u8 *ip = NULL, *subnet = NULL, *gw;
gw                350 drivers/scsi/be2iscsi/be_iscsi.c 		gw = info->value;
gw                351 drivers/scsi/be2iscsi/be_iscsi.c 		ret = beiscsi_if_set_gw(phba, BEISCSI_IP_TYPE_V4, gw);
gw                434 drivers/scsi/be2iscsi/be_mgmt.c 			     u32 action, u32 ip_type, u8 *gw)
gw                449 drivers/scsi/be2iscsi/be_mgmt.c 	memcpy(req->ip_addr.addr, gw,
gw                454 drivers/scsi/be2iscsi/be_mgmt.c int beiscsi_if_set_gw(struct beiscsi_hba *phba, u32 ip_type, u8 *gw)
gw                477 drivers/scsi/be2iscsi/be_mgmt.c 	rt_val = beiscsi_if_mod_gw(phba, IP_ACTION_ADD, ip_type, gw);
gw                639 drivers/scsi/be2iscsi/be_mgmt.c 	u8 *gw;
gw                668 drivers/scsi/be2iscsi/be_mgmt.c 	gw = (u8 *)&gw_resp.ip_addr.addr;
gw                669 drivers/scsi/be2iscsi/be_mgmt.c 	if (!beiscsi_if_zero_ip(gw, if_info->ip_addr.ip_type)) {
gw                671 drivers/scsi/be2iscsi/be_mgmt.c 				       if_info->ip_addr.ip_type, gw);
gw                177 drivers/scsi/be2iscsi/be_mgmt.h int beiscsi_if_set_gw(struct beiscsi_hba *phba, u32 ip_type, u8 *gw);
gw               1983 drivers/scsi/qedi/qedi_main.c 	char *fmt, *ip, *sub, *gw;
gw               2000 drivers/scsi/qedi/qedi_main.c 	gw = ipv6_en ? initiator->ipv6.gateway.byte :
gw               2007 drivers/scsi/qedi/qedi_main.c 		gw = ip;
gw               2019 drivers/scsi/qedi/qedi_main.c 		rc = snprintf(buf, ip_len, fmt, gw);
gw                335 include/net/ip6_route.h struct neighbour *ip6_neigh_lookup(const struct in6_addr *gw,
gw                436 include/net/ip_fib.h int ip_fib_check_default(__be32 gw, struct net_device *dev);
gw                  9 include/net/netfilter/ipv4/nf_dup_ipv4.h 		 const struct in_addr *gw, int oif);
gw                  8 include/net/netfilter/ipv6/nf_dup_ipv6.h 		 const struct in6_addr *gw, int oif);
gw                 37 include/net/nexthop.h 	} gw;
gw                 35 include/trace/events/fib6.h 		__array(		__u8,	gw,	16	 )
gw                 73 include/trace/events/fib6.h 			in6 = (struct in6_addr *)__entry->gw;
gw                 77 include/trace/events/fib6.h 			in6 = (struct in6_addr *)__entry->gw;
gw                 86 include/trace/events/fib6.h 		  __get_str(name), __entry->gw, __entry->err)
gw                  8 include/uapi/linux/netfilter/xt_TEE.h 	union nf_inet_addr gw;
gw                953 include/video/pm3fb.h 	#define PM3DownloadGlyphwidth_GlyphWidth(gw)	((gw) & 0xffff)
gw               2406 net/batman-adv/bat_iv_ogm.c 	atomic_set(&bat_priv->gw.sel_class, 20);
gw               2422 net/batman-adv/bat_iv_ogm.c 	hlist_for_each_entry_rcu(gw_node, &bat_priv->gw.gateway_list, list) {
gw               2438 net/batman-adv/bat_iv_ogm.c 		switch (atomic_read(&bat_priv->gw.sel_class)) {
gw               2501 net/batman-adv/bat_iv_ogm.c 	if (atomic_read(&bat_priv->gw.sel_class) <= 2)
gw               2536 net/batman-adv/bat_iv_ogm.c 	if ((atomic_read(&bat_priv->gw.sel_class) > 3) &&
gw               2537 net/batman-adv/bat_iv_ogm.c 	    (orig_tq_avg - gw_tq_avg < atomic_read(&bat_priv->gw.sel_class)))
gw               2610 net/batman-adv/bat_iv_ogm.c 	hlist_for_each_entry_rcu(gw_node, &bat_priv->gw.gateway_list, list) {
gw               2715 net/batman-adv/bat_iv_ogm.c 	spin_lock_bh(&bat_priv->gw.list_lock);
gw               2716 net/batman-adv/bat_iv_ogm.c 	cb->seq = bat_priv->gw.generation << 1 | 1;
gw               2718 net/batman-adv/bat_iv_ogm.c 	hlist_for_each_entry(gw_node, &bat_priv->gw.gateway_list, list) {
gw               2731 net/batman-adv/bat_iv_ogm.c 	spin_unlock_bh(&bat_priv->gw.list_lock);
gw               2759 net/batman-adv/bat_iv_ogm.c 	.gw = {
gw                666 net/batman-adv/bat_v.c 	atomic_set(&bat_priv->gw.sel_class, 50);
gw                679 net/batman-adv/bat_v.c 	old_class = atomic_read(&bat_priv->gw.sel_class);
gw                680 net/batman-adv/bat_v.c 	atomic_set(&bat_priv->gw.sel_class, class);
gw                690 net/batman-adv/bat_v.c 	u32 class = atomic_read(&bat_priv->gw.sel_class);
gw                751 net/batman-adv/bat_v.c 	hlist_for_each_entry_rcu(gw_node, &bat_priv->gw.gateway_list, list) {
gw                792 net/batman-adv/bat_v.c 	threshold = atomic_read(&bat_priv->gw.sel_class);
gw                890 net/batman-adv/bat_v.c 	hlist_for_each_entry_rcu(gw_node, &bat_priv->gw.gateway_list, list) {
gw               1015 net/batman-adv/bat_v.c 	spin_lock_bh(&bat_priv->gw.list_lock);
gw               1016 net/batman-adv/bat_v.c 	cb->seq = bat_priv->gw.generation << 1 | 1;
gw               1018 net/batman-adv/bat_v.c 	hlist_for_each_entry(gw_node, &bat_priv->gw.gateway_list, list) {
gw               1031 net/batman-adv/bat_v.c 	spin_unlock_bh(&bat_priv->gw.list_lock);
gw               1060 net/batman-adv/bat_v.c 	.gw = {
gw                 95 net/batman-adv/gateway_client.c 	gw_node = rcu_dereference(bat_priv->gw.curr_gw);
gw                144 net/batman-adv/gateway_client.c 	spin_lock_bh(&bat_priv->gw.list_lock);
gw                149 net/batman-adv/gateway_client.c 	curr_gw_node = rcu_dereference_protected(bat_priv->gw.curr_gw, 1);
gw                150 net/batman-adv/gateway_client.c 	rcu_assign_pointer(bat_priv->gw.curr_gw, new_gw_node);
gw                155 net/batman-adv/gateway_client.c 	spin_unlock_bh(&bat_priv->gw.list_lock);
gw                171 net/batman-adv/gateway_client.c 	atomic_set(&bat_priv->gw.reselect, 1);
gw                186 net/batman-adv/gateway_client.c 	if (atomic_read(&bat_priv->gw.mode) != BATADV_GW_MODE_CLIENT)
gw                218 net/batman-adv/gateway_client.c 	if (atomic_read(&bat_priv->gw.mode) != BATADV_GW_MODE_CLIENT)
gw                221 net/batman-adv/gateway_client.c 	if (!bat_priv->algo_ops->gw.get_best_gw_node)
gw                226 net/batman-adv/gateway_client.c 	if (!batadv_atomic_dec_not_zero(&bat_priv->gw.reselect) && curr_gw)
gw                233 net/batman-adv/gateway_client.c 	next_gw = bat_priv->algo_ops->gw.get_best_gw_node(bat_priv);
gw                311 net/batman-adv/gateway_client.c 	if (!bat_priv->algo_ops->gw.is_eligible)
gw                322 net/batman-adv/gateway_client.c 	if (!bat_priv->algo_ops->gw.is_eligible(bat_priv, curr_gw_orig,
gw                348 net/batman-adv/gateway_client.c 	lockdep_assert_held(&bat_priv->gw.list_lock);
gw                365 net/batman-adv/gateway_client.c 	hlist_add_head_rcu(&gw_node->list, &bat_priv->gw.gateway_list);
gw                366 net/batman-adv/gateway_client.c 	bat_priv->gw.generation++;
gw                393 net/batman-adv/gateway_client.c 	hlist_for_each_entry_rcu(gw_node_tmp, &bat_priv->gw.gateway_list,
gw                422 net/batman-adv/gateway_client.c 	spin_lock_bh(&bat_priv->gw.list_lock);
gw                426 net/batman-adv/gateway_client.c 		spin_unlock_bh(&bat_priv->gw.list_lock);
gw                429 net/batman-adv/gateway_client.c 	spin_unlock_bh(&bat_priv->gw.list_lock);
gw                458 net/batman-adv/gateway_client.c 		spin_lock_bh(&bat_priv->gw.list_lock);
gw                462 net/batman-adv/gateway_client.c 			bat_priv->gw.generation++;
gw                464 net/batman-adv/gateway_client.c 		spin_unlock_bh(&bat_priv->gw.list_lock);
gw                504 net/batman-adv/gateway_client.c 	spin_lock_bh(&bat_priv->gw.list_lock);
gw                506 net/batman-adv/gateway_client.c 				  &bat_priv->gw.gateway_list, list) {
gw                509 net/batman-adv/gateway_client.c 		bat_priv->gw.generation++;
gw                511 net/batman-adv/gateway_client.c 	spin_unlock_bh(&bat_priv->gw.list_lock);
gw                540 net/batman-adv/gateway_client.c 	if (!bat_priv->algo_ops->gw.print) {
gw                546 net/batman-adv/gateway_client.c 	bat_priv->algo_ops->gw.print(bat_priv, seq);
gw                587 net/batman-adv/gateway_client.c 	if (!bat_priv->algo_ops->gw.dump) {
gw                592 net/batman-adv/gateway_client.c 	bat_priv->algo_ops->gw.dump(msg, cb, bat_priv);
gw                768 net/batman-adv/gateway_client.c 	switch (atomic_read(&bat_priv->gw.mode)) {
gw                136 net/batman-adv/gateway_common.c 	struct batadv_tvlv_gateway_data gw;
gw                140 net/batman-adv/gateway_common.c 	gw_mode = atomic_read(&bat_priv->gw.mode);
gw                148 net/batman-adv/gateway_common.c 		down = atomic_read(&bat_priv->gw.bandwidth_down);
gw                149 net/batman-adv/gateway_common.c 		up = atomic_read(&bat_priv->gw.bandwidth_up);
gw                150 net/batman-adv/gateway_common.c 		gw.bandwidth_down = htonl(down);
gw                151 net/batman-adv/gateway_common.c 		gw.bandwidth_up = htonl(up);
gw                153 net/batman-adv/gateway_common.c 					       &gw, sizeof(gw));
gw                177 net/batman-adv/gateway_common.c 	down_curr = (unsigned int)atomic_read(&bat_priv->gw.bandwidth_down);
gw                178 net/batman-adv/gateway_common.c 	up_curr = (unsigned int)atomic_read(&bat_priv->gw.bandwidth_up);
gw                202 net/batman-adv/gateway_common.c 	atomic_set(&bat_priv->gw.bandwidth_down, down_new);
gw                203 net/batman-adv/gateway_common.c 	atomic_set(&bat_priv->gw.bandwidth_up, up_new);
gw                246 net/batman-adv/gateway_common.c 	    atomic_read(&bat_priv->gw.mode) == BATADV_GW_MODE_CLIENT)
gw                256 net/batman-adv/gateway_common.c 	if (bat_priv->algo_ops->gw.init_sel_class)
gw                257 net/batman-adv/gateway_common.c 		bat_priv->algo_ops->gw.init_sel_class(bat_priv);
gw                259 net/batman-adv/gateway_common.c 		atomic_set(&bat_priv->gw.sel_class, 1);
gw                169 net/batman-adv/main.c 	spin_lock_init(&bat_priv->gw.list_lock);
gw                181 net/batman-adv/main.c 	INIT_HLIST_HEAD(&bat_priv->gw.gateway_list);
gw                198 net/batman-adv/main.c 	bat_priv->gw.generation = 0;
gw                227 net/batman-adv/main.c 	atomic_set(&bat_priv->gw.reselect, 0);
gw                310 net/batman-adv/netlink.c 			atomic_read(&bat_priv->gw.bandwidth_down)))
gw                314 net/batman-adv/netlink.c 			atomic_read(&bat_priv->gw.bandwidth_up)))
gw                318 net/batman-adv/netlink.c 		       atomic_read(&bat_priv->gw.mode)))
gw                321 net/batman-adv/netlink.c 	if (bat_priv->algo_ops->gw.get_best_gw_node &&
gw                322 net/batman-adv/netlink.c 	    bat_priv->algo_ops->gw.is_eligible) {
gw                327 net/batman-adv/netlink.c 				atomic_read(&bat_priv->gw.sel_class)))
gw                505 net/batman-adv/netlink.c 		atomic_set(&bat_priv->gw.bandwidth_down, nla_get_u32(attr));
gw                512 net/batman-adv/netlink.c 		atomic_set(&bat_priv->gw.bandwidth_up, nla_get_u32(attr));
gw                540 net/batman-adv/netlink.c 			atomic_set(&bat_priv->gw.mode, gw_mode);
gw                546 net/batman-adv/netlink.c 	    bat_priv->algo_ops->gw.get_best_gw_node &&
gw                547 net/batman-adv/netlink.c 	    bat_priv->algo_ops->gw.is_eligible) {
gw                558 net/batman-adv/netlink.c 		if (!bat_priv->algo_ops->gw.store_sel_class)
gw                562 net/batman-adv/netlink.c 			atomic_set(&bat_priv->gw.sel_class, sel_class);
gw                271 net/batman-adv/soft-interface.c 	gw_mode = atomic_read(&bat_priv->gw.mode);
gw                781 net/batman-adv/soft-interface.c 	atomic_set(&bat_priv->gw.mode, BATADV_GW_MODE_OFF);
gw                782 net/batman-adv/soft-interface.c 	atomic_set(&bat_priv->gw.bandwidth_down, 100);
gw                783 net/batman-adv/soft-interface.c 	atomic_set(&bat_priv->gw.bandwidth_up, 20);
gw                464 net/batman-adv/sysfs.c 	if (!bat_priv->algo_ops->gw.get_best_gw_node ||
gw                465 net/batman-adv/sysfs.c 	    !bat_priv->algo_ops->gw.is_eligible)
gw                468 net/batman-adv/sysfs.c 	switch (atomic_read(&bat_priv->gw.mode)) {
gw                500 net/batman-adv/sysfs.c 	if (!bat_priv->algo_ops->gw.get_best_gw_node ||
gw                501 net/batman-adv/sysfs.c 	    !bat_priv->algo_ops->gw.is_eligible)
gw                526 net/batman-adv/sysfs.c 	if (atomic_read(&bat_priv->gw.mode) == gw_mode_tmp)
gw                529 net/batman-adv/sysfs.c 	switch (atomic_read(&bat_priv->gw.mode)) {
gw                558 net/batman-adv/sysfs.c 	atomic_set(&bat_priv->gw.mode, (unsigned int)gw_mode_tmp);
gw                576 net/batman-adv/sysfs.c 	if (!bat_priv->algo_ops->gw.get_best_gw_node ||
gw                577 net/batman-adv/sysfs.c 	    !bat_priv->algo_ops->gw.is_eligible)
gw                580 net/batman-adv/sysfs.c 	if (bat_priv->algo_ops->gw.show_sel_class)
gw                581 net/batman-adv/sysfs.c 		return bat_priv->algo_ops->gw.show_sel_class(bat_priv, buff);
gw                583 net/batman-adv/sysfs.c 	return sprintf(buff, "%i\n", atomic_read(&bat_priv->gw.sel_class));
gw                598 net/batman-adv/sysfs.c 	if (!bat_priv->algo_ops->gw.get_best_gw_node ||
gw                599 net/batman-adv/sysfs.c 	    !bat_priv->algo_ops->gw.is_eligible)
gw                605 net/batman-adv/sysfs.c 	if (bat_priv->algo_ops->gw.store_sel_class)
gw                606 net/batman-adv/sysfs.c 		return bat_priv->algo_ops->gw.store_sel_class(bat_priv, buff,
gw                611 net/batman-adv/sysfs.c 					  &bat_priv->gw.sel_class,
gw                627 net/batman-adv/sysfs.c 	down = atomic_read(&bat_priv->gw.bandwidth_down);
gw                628 net/batman-adv/sysfs.c 	up = atomic_read(&bat_priv->gw.bandwidth_up);
gw               1734 net/batman-adv/types.h 	struct batadv_priv_gw gw;
gw               2323 net/batman-adv/types.h 	struct batadv_algo_gw_ops gw;
gw                 33 net/bluetooth/6lowpan.c 	struct in6_addr gw;
gw                172 net/bluetooth/6lowpan.c 		if (ipv6_addr_any(&lowpan_cb(skb)->gw)) {
gw                180 net/bluetooth/6lowpan.c 			nexthop = &lowpan_cb(skb)->gw;
gw                189 net/bluetooth/6lowpan.c 		memcpy(&lowpan_cb(skb)->gw, nexthop, sizeof(struct in6_addr));
gw                253 net/decnet/dn_table.c 		__le16 gw;
gw                263 net/decnet/dn_table.c 			gw = gw_attr ? nla_get_le16(gw_attr) : 0;
gw                265 net/decnet/dn_table.c 			if (gw && gw != nh->nh_gw)
gw                431 net/ipv4/fib_semantics.c int ip_fib_check_default(__be32 gw, struct net_device *dev)
gw                443 net/ipv4/fib_semantics.c 		    nh->fib_nh_gw4 == gw &&
gw               2811 net/ipv4/fib_trie.c 			__be32 gw = 0;
gw               2814 net/ipv4/fib_trie.c 				gw = nhc->nhc_gw.ipv4;
gw               2820 net/ipv4/fib_trie.c 				   prefix, gw, flags, 0, 0,
gw                 24 net/ipv4/netfilter/nf_dup_ipv4.c 			      const struct in_addr *gw, int oif)
gw                 34 net/ipv4/netfilter/nf_dup_ipv4.c 	fl4.daddr = gw->s_addr;
gw                 51 net/ipv4/netfilter/nf_dup_ipv4.c 		 const struct in_addr *gw, int oif)
gw                 86 net/ipv4/netfilter/nf_dup_ipv4.c 	if (nf_dup_ipv4_route(net, skb, gw, oif)) {
gw                 25 net/ipv4/netfilter/nft_dup_ipv4.c 	struct in_addr gw = {
gw                 30 net/ipv4/netfilter/nft_dup_ipv4.c 	nf_dup_ipv4(nft_net(pkt), pkt->skb, nft_hook(pkt), &gw, oif);
gw               1174 net/ipv4/nexthop.c 		.fc_gw4   = cfg->gw.ipv4,
gw               1175 net/ipv4/nexthop.c 		.fc_gw_family = cfg->gw.ipv4 ? AF_INET : 0,
gw               1210 net/ipv4/nexthop.c 		.fc_gateway = cfg->gw.ipv6,
gw               1217 net/ipv4/nexthop.c 	if (!ipv6_addr_any(&cfg->gw.ipv6))
gw               1445 net/ipv4/nexthop.c 			cfg->gw.ipv4 = nla_get_be32(gwa);
gw               1452 net/ipv4/nexthop.c 			cfg->gw.ipv6 = nla_get_in6_addr(gwa);
gw                644 net/ipv4/route.c 				  __be32 gw, u32 pmtu, bool lock,
gw                680 net/ipv4/route.c 		if (gw)
gw                681 net/ipv4/route.c 			fnhe->fnhe_gw = gw;
gw                707 net/ipv4/route.c 		fnhe->fnhe_gw = gw;
gw                919 net/ipv4/route.c 		__be32 gw = rt_nexthop(rt, ip_hdr(skb)->daddr);
gw                921 net/ipv4/route.c 		icmp_send(skb, ICMP_REDIRECT, ICMP_REDIR_HOST, gw);
gw                929 net/ipv4/route.c 					     &ip_hdr(skb)->daddr, &gw);
gw               1814 net/ipv4/route.c 		__be32 gw;
gw               1816 net/ipv4/route.c 		gw = nhc->nhc_gw_family == AF_INET ? nhc->nhc_gw.ipv4 : 0;
gw               1818 net/ipv4/route.c 		    inet_addr_onlink(out_dev, saddr, gw))
gw                 20 net/ipv6/netfilter/nf_dup_ipv6.c 			      const struct in6_addr *gw, int oif)
gw                 30 net/ipv6/netfilter/nf_dup_ipv6.c 	fl6.daddr = *gw;
gw                 48 net/ipv6/netfilter/nf_dup_ipv6.c 		 const struct in6_addr *gw, int oif)
gw                 65 net/ipv6/netfilter/nf_dup_ipv6.c 	if (nf_dup_ipv6_route(net, skb, gw, oif)) {
gw                 25 net/ipv6/netfilter/nft_dup_ipv6.c 	struct in6_addr *gw = (struct in6_addr *)&regs->data[priv->sreg_addr];
gw                 28 net/ipv6/netfilter/nft_dup_ipv6.c 	nf_dup_ipv6(nft_net(pkt), pkt->skb, nft_hook(pkt), gw, oif);
gw                200 net/ipv6/route.c struct neighbour *ip6_neigh_lookup(const struct in6_addr *gw,
gw                207 net/ipv6/route.c 	daddr = choose_neigh_daddr(gw, skb, daddr);
gw               1941 net/ipv6/route.c 	const struct in6_addr	*gw;
gw               1951 net/ipv6/route.c 	    (arg->gw && !nh->fib_nh_gw_family) ||
gw               1952 net/ipv6/route.c 	    (!arg->gw && nh->fib_nh_gw_family) ||
gw               1953 net/ipv6/route.c 	    (arg->gw && !ipv6_addr_equal(arg->gw, &nh->fib_nh_gw6)))
gw               1976 net/ipv6/route.c 			.gw = &rt->rt6i_gateway,
gw               2774 net/ipv6/route.c 				.gw = &rt6->rt6i_gateway,
gw               2872 net/ipv6/route.c 				  const struct in6_addr *gw,
gw               2886 net/ipv6/route.c 	if (!ipv6_addr_equal(gw, &nh->fib_nh_gw6)) {
gw               2891 net/ipv6/route.c 		    ipv6_addr_equal(gw, &rt_cache->rt6i_gateway)) {
gw               2903 net/ipv6/route.c 	const struct in6_addr	*gw;
gw               2912 net/ipv6/route.c 	return ip6_redirect_nh_match(arg->res, arg->fl6, arg->gw, arg->ret);
gw               2933 net/ipv6/route.c 		.gw  = &rdfl->gateway,
gw               4073 net/ipv6/route.c 			.gw = &rt->rt6i_gateway,
gw                 42 net/netfilter/xt_TEE.c 	nf_dup_ipv4(xt_net(par), skb, xt_hooknum(par), &info->gw.in, oif);
gw                 54 net/netfilter/xt_TEE.c 	nf_dup_ipv6(xt_net(par), skb, xt_hooknum(par), &info->gw.in6, oif);
gw                 99 net/netfilter/xt_TEE.c 	if (memcmp(&info->gw, &tee_zero_address,
gw                 24 samples/bpf/xdp_router_ipv4_kern.c 	__be32 gw;
gw                166 samples/bpf/xdp_router_ipv4_kern.c 				if (!prefix_value->gw)
gw                168 samples/bpf/xdp_router_ipv4_kern.c 				dest_ip = prefix_value->gw;
gw                134 samples/bpf/xdp_router_ipv4_user.c 		__be32 dst, gw;
gw                192 samples/bpf/xdp_router_ipv4_user.c 		route.gw = atoi(gws);
gw                208 samples/bpf/xdp_router_ipv4_user.c 				__be32 gw;
gw                242 samples/bpf/xdp_router_ipv4_user.c 			       route.gw, route.dst_len,
gw                251 samples/bpf/xdp_router_ipv4_user.c 				prefix_value->gw = route.gw;
gw                293 samples/bpf/xdp_router_ipv4_user.c 					prefix_value->gw = route.gw;