ipt 1881 fs/io_uring.c struct io_poll_table ipt; ipt 1902 fs/io_uring.c ipt.pt._qproc = io_poll_queue_proc; ipt 1903 fs/io_uring.c ipt.pt._key = poll->events; ipt 1904 fs/io_uring.c ipt.req = req; ipt 1905 fs/io_uring.c ipt.error = -EINVAL; /* same as no support for IOCB_CMD_POLL */ ipt 1913 fs/io_uring.c mask = vfs_poll(poll->file, &ipt.pt) & poll->events; ipt 1919 fs/io_uring.c if (ipt.error) ipt 1921 fs/io_uring.c ipt.error = 0; ipt 1924 fs/io_uring.c if (mask || ipt.error) ipt 1933 fs/io_uring.c ipt.error = 0; ipt 1942 fs/io_uring.c return ipt.error; ipt 38 net/ipv4/netfilter/ip_tables.c return xt_alloc_initial_table(ipt, IPT); ipt 79 net/sched/act_ipt.c struct tcf_ipt *ipt = to_ipt(a); ipt 81 net/sched/act_ipt.c if (ipt->tcfi_t) { ipt 82 net/sched/act_ipt.c ipt_destroy_target(ipt->tcfi_t, a->idrinfo->net); ipt 83 net/sched/act_ipt.c kfree(ipt->tcfi_t); ipt 85 net/sched/act_ipt.c kfree(ipt->tcfi_tname); ipt 102 net/sched/act_ipt.c struct tcf_ipt *ipt; ipt 180 net/sched/act_ipt.c ipt = to_ipt(*a); ipt 182 net/sched/act_ipt.c spin_lock_bh(&ipt->tcf_lock); ipt 184 net/sched/act_ipt.c ipt_destroy_target(ipt->tcfi_t, net); ipt 185 net/sched/act_ipt.c kfree(ipt->tcfi_tname); ipt 186 net/sched/act_ipt.c kfree(ipt->tcfi_t); ipt 188 net/sched/act_ipt.c ipt->tcfi_tname = tname; ipt 189 net/sched/act_ipt.c ipt->tcfi_t = t; ipt 190 net/sched/act_ipt.c ipt->tcfi_hook = hook; ipt 191 net/sched/act_ipt.c spin_unlock_bh(&ipt->tcf_lock); ipt 227 net/sched/act_ipt.c struct tcf_ipt *ipt = to_ipt(a); ipt 232 net/sched/act_ipt.c .hook = ipt->tcfi_hook, ipt 239 net/sched/act_ipt.c spin_lock(&ipt->tcf_lock); ipt 241 net/sched/act_ipt.c tcf_lastuse_update(&ipt->tcf_tm); ipt 242 net/sched/act_ipt.c bstats_update(&ipt->tcf_bstats, skb); ipt 249 net/sched/act_ipt.c par.target = ipt->tcfi_t->u.kernel.target; ipt 250 net/sched/act_ipt.c par.targinfo = ipt->tcfi_t->data; ipt 259 net/sched/act_ipt.c ipt->tcf_qstats.drops++; ipt 270 net/sched/act_ipt.c spin_unlock(&ipt->tcf_lock); ipt 279 net/sched/act_ipt.c struct tcf_ipt *ipt = to_ipt(a); ipt 289 net/sched/act_ipt.c spin_lock_bh(&ipt->tcf_lock); ipt 290 net/sched/act_ipt.c t = kmemdup(ipt->tcfi_t, ipt->tcfi_t->u.user.target_size, GFP_ATOMIC); ipt 294 net/sched/act_ipt.c c.bindcnt = atomic_read(&ipt->tcf_bindcnt) - bind; ipt 295 net/sched/act_ipt.c c.refcnt = refcount_read(&ipt->tcf_refcnt) - ref; ipt 296 net/sched/act_ipt.c strcpy(t->u.user.name, ipt->tcfi_t->u.kernel.target->name); ipt 298 net/sched/act_ipt.c if (nla_put(skb, TCA_IPT_TARG, ipt->tcfi_t->u.user.target_size, t) || ipt 299 net/sched/act_ipt.c nla_put_u32(skb, TCA_IPT_INDEX, ipt->tcf_index) || ipt 300 net/sched/act_ipt.c nla_put_u32(skb, TCA_IPT_HOOK, ipt->tcfi_hook) || ipt 302 net/sched/act_ipt.c nla_put_string(skb, TCA_IPT_TABLE, ipt->tcfi_tname)) ipt 305 net/sched/act_ipt.c tcf_tm_dump(&tm, &ipt->tcf_tm); ipt 309 net/sched/act_ipt.c spin_unlock_bh(&ipt->tcf_lock); ipt 314 net/sched/act_ipt.c spin_unlock_bh(&ipt->tcf_lock);