ipt              1881 fs/io_uring.c  	struct io_poll_table ipt;
ipt              1902 fs/io_uring.c  	ipt.pt._qproc = io_poll_queue_proc;
ipt              1903 fs/io_uring.c  	ipt.pt._key = poll->events;
ipt              1904 fs/io_uring.c  	ipt.req = req;
ipt              1905 fs/io_uring.c  	ipt.error = -EINVAL; /* same as no support for IOCB_CMD_POLL */
ipt              1913 fs/io_uring.c  	mask = vfs_poll(poll->file, &ipt.pt) & poll->events;
ipt              1919 fs/io_uring.c  			if (ipt.error)
ipt              1921 fs/io_uring.c  			ipt.error = 0;
ipt              1924 fs/io_uring.c  		if (mask || ipt.error)
ipt              1933 fs/io_uring.c  		ipt.error = 0;
ipt              1942 fs/io_uring.c  	return ipt.error;
ipt                38 net/ipv4/netfilter/ip_tables.c 	return xt_alloc_initial_table(ipt, IPT);
ipt                79 net/sched/act_ipt.c 	struct tcf_ipt *ipt = to_ipt(a);
ipt                81 net/sched/act_ipt.c 	if (ipt->tcfi_t) {
ipt                82 net/sched/act_ipt.c 		ipt_destroy_target(ipt->tcfi_t, a->idrinfo->net);
ipt                83 net/sched/act_ipt.c 		kfree(ipt->tcfi_t);
ipt                85 net/sched/act_ipt.c 	kfree(ipt->tcfi_tname);
ipt               102 net/sched/act_ipt.c 	struct tcf_ipt *ipt;
ipt               180 net/sched/act_ipt.c 	ipt = to_ipt(*a);
ipt               182 net/sched/act_ipt.c 	spin_lock_bh(&ipt->tcf_lock);
ipt               184 net/sched/act_ipt.c 		ipt_destroy_target(ipt->tcfi_t, net);
ipt               185 net/sched/act_ipt.c 		kfree(ipt->tcfi_tname);
ipt               186 net/sched/act_ipt.c 		kfree(ipt->tcfi_t);
ipt               188 net/sched/act_ipt.c 	ipt->tcfi_tname = tname;
ipt               189 net/sched/act_ipt.c 	ipt->tcfi_t     = t;
ipt               190 net/sched/act_ipt.c 	ipt->tcfi_hook  = hook;
ipt               191 net/sched/act_ipt.c 	spin_unlock_bh(&ipt->tcf_lock);
ipt               227 net/sched/act_ipt.c 	struct tcf_ipt *ipt = to_ipt(a);
ipt               232 net/sched/act_ipt.c 		.hook	= ipt->tcfi_hook,
ipt               239 net/sched/act_ipt.c 	spin_lock(&ipt->tcf_lock);
ipt               241 net/sched/act_ipt.c 	tcf_lastuse_update(&ipt->tcf_tm);
ipt               242 net/sched/act_ipt.c 	bstats_update(&ipt->tcf_bstats, skb);
ipt               249 net/sched/act_ipt.c 	par.target   = ipt->tcfi_t->u.kernel.target;
ipt               250 net/sched/act_ipt.c 	par.targinfo = ipt->tcfi_t->data;
ipt               259 net/sched/act_ipt.c 		ipt->tcf_qstats.drops++;
ipt               270 net/sched/act_ipt.c 	spin_unlock(&ipt->tcf_lock);
ipt               279 net/sched/act_ipt.c 	struct tcf_ipt *ipt = to_ipt(a);
ipt               289 net/sched/act_ipt.c 	spin_lock_bh(&ipt->tcf_lock);
ipt               290 net/sched/act_ipt.c 	t = kmemdup(ipt->tcfi_t, ipt->tcfi_t->u.user.target_size, GFP_ATOMIC);
ipt               294 net/sched/act_ipt.c 	c.bindcnt = atomic_read(&ipt->tcf_bindcnt) - bind;
ipt               295 net/sched/act_ipt.c 	c.refcnt = refcount_read(&ipt->tcf_refcnt) - ref;
ipt               296 net/sched/act_ipt.c 	strcpy(t->u.user.name, ipt->tcfi_t->u.kernel.target->name);
ipt               298 net/sched/act_ipt.c 	if (nla_put(skb, TCA_IPT_TARG, ipt->tcfi_t->u.user.target_size, t) ||
ipt               299 net/sched/act_ipt.c 	    nla_put_u32(skb, TCA_IPT_INDEX, ipt->tcf_index) ||
ipt               300 net/sched/act_ipt.c 	    nla_put_u32(skb, TCA_IPT_HOOK, ipt->tcfi_hook) ||
ipt               302 net/sched/act_ipt.c 	    nla_put_string(skb, TCA_IPT_TABLE, ipt->tcfi_tname))
ipt               305 net/sched/act_ipt.c 	tcf_tm_dump(&tm, &ipt->tcf_tm);
ipt               309 net/sched/act_ipt.c 	spin_unlock_bh(&ipt->tcf_lock);
ipt               314 net/sched/act_ipt.c 	spin_unlock_bh(&ipt->tcf_lock);