po 298 arch/arm/mach-omap2/powerdomain.c int pwrdm_register_platform_funcs(struct pwrdm_ops *po) po 300 arch/arm/mach-omap2/powerdomain.c if (!po) po 306 arch/arm/mach-omap2/powerdomain.c arch_pwrdm = po; po 29 drivers/gpio/gpio-ml-ioh.c u32 po; po 99 drivers/gpio/gpio-ml-ioh.c reg_val = ioread32(&chip->reg->regs[chip->ch].po); po 105 drivers/gpio/gpio-ml-ioh.c iowrite32(reg_val, &chip->reg->regs[chip->ch].po); po 130 drivers/gpio/gpio-ml-ioh.c reg_val = ioread32(&chip->reg->regs[chip->ch].po); po 135 drivers/gpio/gpio-ml-ioh.c iowrite32(reg_val, &chip->reg->regs[chip->ch].po); po 168 drivers/gpio/gpio-ml-ioh.c ioread32(&chip->reg->regs[chip->ch].po); po 194 drivers/gpio/gpio-ml-ioh.c &chip->reg->regs[chip->ch].po); po 29 drivers/gpio/gpio-pch.c u32 po; po 104 drivers/gpio/gpio-pch.c reg_val = ioread32(&chip->reg->po); po 110 drivers/gpio/gpio-pch.c iowrite32(reg_val, &chip->reg->po); po 131 drivers/gpio/gpio-pch.c reg_val = ioread32(&chip->reg->po); po 136 drivers/gpio/gpio-pch.c iowrite32(reg_val, &chip->reg->po); po 169 drivers/gpio/gpio-pch.c chip->pch_gpio_reg.po_reg = ioread32(&chip->reg->po); po 186 drivers/gpio/gpio-pch.c iowrite32(chip->pch_gpio_reg.po_reg, &chip->reg->po); po 397 drivers/media/platform/sti/hva/hva-h264.c struct hva_h264_po po; po 780 drivers/media/platform/sti/hva/hva-h264.c offsetof(struct hva_h264_task, po); po 875 drivers/media/platform/sti/hva/hva-h264.c struct hva_h264_po *po = &task->po; po 877 drivers/media/platform/sti/hva/hva-h264.c return po->bitstream_size; po 882 drivers/media/platform/sti/hva/hva-h264.c struct hva_h264_po *po = &task->po; po 884 drivers/media/platform/sti/hva/hva-h264.c return po->stuffing_bits >> 3; po 179 drivers/net/ppp/pppoe.c static int __set_item(struct pppoe_net *pn, struct pppox_sock *po) po 181 drivers/net/ppp/pppoe.c int hash = hash_item(po->pppoe_pa.sid, po->pppoe_pa.remote); po 186 drivers/net/ppp/pppoe.c if (cmp_2_addr(&ret->pppoe_pa, &po->pppoe_pa) && po 187 drivers/net/ppp/pppoe.c ret->pppoe_ifindex == po->pppoe_ifindex) po 193 drivers/net/ppp/pppoe.c po->next = pn->hash_table[hash]; po 194 drivers/net/ppp/pppoe.c pn->hash_table[hash] = po; po 228 drivers/net/ppp/pppoe.c struct pppox_sock *po; po 231 drivers/net/ppp/pppoe.c po = __get_item(pn, sid, addr, ifindex); po 232 drivers/net/ppp/pppoe.c if (po) po 233 drivers/net/ppp/pppoe.c sock_hold(sk_pppox(po)); po 236 drivers/net/ppp/pppoe.c return po; po 283 drivers/net/ppp/pppoe.c struct pppox_sock *po = pn->hash_table[i]; po 286 drivers/net/ppp/pppoe.c while (po) { po 287 drivers/net/ppp/pppoe.c while (po && po->pppoe_dev != dev) { po 288 drivers/net/ppp/pppoe.c po = po->next; po 291 drivers/net/ppp/pppoe.c if (!po) po 294 drivers/net/ppp/pppoe.c sk = sk_pppox(po); po 308 drivers/net/ppp/pppoe.c if (po->pppoe_dev == dev && po 312 drivers/net/ppp/pppoe.c po->pppoe_dev = NULL; po 326 drivers/net/ppp/pppoe.c po = pn->hash_table[i]; po 369 drivers/net/ppp/pppoe.c struct pppox_sock *po = pppox_sk(sk); po 381 drivers/net/ppp/pppoe.c ppp_input(&po->chan, skb); po 384 drivers/net/ppp/pppoe.c &po->pppoe_relay); po 419 drivers/net/ppp/pppoe.c struct pppox_sock *po; po 449 drivers/net/ppp/pppoe.c po = get_item(pn, ph->sid, eth_hdr(skb)->h_source, dev->ifindex); po 450 drivers/net/ppp/pppoe.c if (!po) po 453 drivers/net/ppp/pppoe.c return sk_receive_skb(sk_pppox(po), skb, 0); po 463 drivers/net/ppp/pppoe.c struct pppox_sock *po = container_of(work, struct pppox_sock, po 465 drivers/net/ppp/pppoe.c struct sock *sk = sk_pppox(po); po 468 drivers/net/ppp/pppoe.c if (po->pppoe_dev) { po 469 drivers/net/ppp/pppoe.c dev_put(po->pppoe_dev); po 470 drivers/net/ppp/pppoe.c po->pppoe_dev = NULL; po 488 drivers/net/ppp/pppoe.c struct pppox_sock *po; po 506 drivers/net/ppp/pppoe.c po = get_item(pn, ph->sid, eth_hdr(skb)->h_source, dev->ifindex); po 507 drivers/net/ppp/pppoe.c if (po) po 508 drivers/net/ppp/pppoe.c if (!schedule_work(&po->proto.pppoe.padt_work)) po 509 drivers/net/ppp/pppoe.c sock_put(sk_pppox(po)); po 566 drivers/net/ppp/pppoe.c struct pppox_sock *po; po 579 drivers/net/ppp/pppoe.c po = pppox_sk(sk); po 581 drivers/net/ppp/pppoe.c if (po->pppoe_dev) { po 582 drivers/net/ppp/pppoe.c dev_put(po->pppoe_dev); po 583 drivers/net/ppp/pppoe.c po->pppoe_dev = NULL; po 598 drivers/net/ppp/pppoe.c delete_item(pn, po->pppoe_pa.sid, po->pppoe_pa.remote, po 599 drivers/net/ppp/pppoe.c po->pppoe_ifindex); po 616 drivers/net/ppp/pppoe.c struct pppox_sock *po = pppox_sk(sk); po 647 drivers/net/ppp/pppoe.c if (stage_session(po->pppoe_pa.sid)) { po 650 drivers/net/ppp/pppoe.c delete_item(pn, po->pppoe_pa.sid, po 651 drivers/net/ppp/pppoe.c po->pppoe_pa.remote, po->pppoe_ifindex); po 652 drivers/net/ppp/pppoe.c if (po->pppoe_dev) { po 653 drivers/net/ppp/pppoe.c dev_put(po->pppoe_dev); po 654 drivers/net/ppp/pppoe.c po->pppoe_dev = NULL; po 657 drivers/net/ppp/pppoe.c po->pppoe_ifindex = 0; po 658 drivers/net/ppp/pppoe.c memset(&po->pppoe_pa, 0, sizeof(po->pppoe_pa)); po 659 drivers/net/ppp/pppoe.c memset(&po->pppoe_relay, 0, sizeof(po->pppoe_relay)); po 660 drivers/net/ppp/pppoe.c memset(&po->chan, 0, sizeof(po->chan)); po 661 drivers/net/ppp/pppoe.c po->next = NULL; po 662 drivers/net/ppp/pppoe.c po->num = 0; po 675 drivers/net/ppp/pppoe.c po->pppoe_dev = dev; po 676 drivers/net/ppp/pppoe.c po->pppoe_ifindex = dev->ifindex; po 682 drivers/net/ppp/pppoe.c memcpy(&po->pppoe_pa, po 687 drivers/net/ppp/pppoe.c error = __set_item(pn, po); po 692 drivers/net/ppp/pppoe.c po->chan.hdrlen = (sizeof(struct pppoe_hdr) + po 695 drivers/net/ppp/pppoe.c po->chan.mtu = dev->mtu - sizeof(struct pppoe_hdr) - 2; po 696 drivers/net/ppp/pppoe.c po->chan.private = sk; po 697 drivers/net/ppp/pppoe.c po->chan.ops = &pppoe_chan_ops; po 699 drivers/net/ppp/pppoe.c error = ppp_register_net_channel(dev_net(dev), &po->chan); po 701 drivers/net/ppp/pppoe.c delete_item(pn, po->pppoe_pa.sid, po 702 drivers/net/ppp/pppoe.c po->pppoe_pa.remote, po->pppoe_ifindex); po 709 drivers/net/ppp/pppoe.c po->num = sp->sa_addr.pppoe.sid; po 715 drivers/net/ppp/pppoe.c if (po->pppoe_dev) { po 716 drivers/net/ppp/pppoe.c dev_put(po->pppoe_dev); po 717 drivers/net/ppp/pppoe.c po->pppoe_dev = NULL; po 742 drivers/net/ppp/pppoe.c struct pppox_sock *po = pppox_sk(sk); po 753 drivers/net/ppp/pppoe.c if (put_user(po->pppoe_dev->mtu - po 770 drivers/net/ppp/pppoe.c if (val < (po->pppoe_dev->mtu po 800 drivers/net/ppp/pppoe.c if (copy_from_user(&po->pppoe_relay, po 806 drivers/net/ppp/pppoe.c if (po->pppoe_relay.sa_family != AF_PPPOX || po 807 drivers/net/ppp/pppoe.c po->pppoe_relay.sa_protocol != PX_PROTO_OE) po 812 drivers/net/ppp/pppoe.c relay_po = get_item_by_addr(sock_net(sk), &po->pppoe_relay); po 843 drivers/net/ppp/pppoe.c struct pppox_sock *po = pppox_sk(sk); po 860 drivers/net/ppp/pppoe.c hdr.sid = po->num; po 862 drivers/net/ppp/pppoe.c dev = po->pppoe_dev; po 896 drivers/net/ppp/pppoe.c po->pppoe_pa.remote, NULL, total_len); po 916 drivers/net/ppp/pppoe.c struct pppox_sock *po = pppox_sk(sk); po 917 drivers/net/ppp/pppoe.c struct net_device *dev = po->pppoe_dev; po 948 drivers/net/ppp/pppoe.c ph->sid = po->num; po 955 drivers/net/ppp/pppoe.c po->pppoe_pa.remote, NULL, data_len); po 1015 drivers/net/ppp/pppoe.c struct pppox_sock *po; po 1023 drivers/net/ppp/pppoe.c po = v; po 1024 drivers/net/ppp/pppoe.c dev_name = po->pppoe_pa.dev; po 1027 drivers/net/ppp/pppoe.c po->pppoe_pa.sid, po->pppoe_pa.remote, dev_name); po 1034 drivers/net/ppp/pppoe.c struct pppox_sock *po; po 1038 drivers/net/ppp/pppoe.c po = pn->hash_table[i]; po 1039 drivers/net/ppp/pppoe.c while (po) { po 1042 drivers/net/ppp/pppoe.c po = po->next; po 1047 drivers/net/ppp/pppoe.c return po; po 1063 drivers/net/ppp/pppoe.c struct pppox_sock *po; po 1067 drivers/net/ppp/pppoe.c po = pppoe_get_idx(pn, 0); po 1070 drivers/net/ppp/pppoe.c po = v; po 1071 drivers/net/ppp/pppoe.c if (po->next) po 1072 drivers/net/ppp/pppoe.c po = po->next; po 1074 drivers/net/ppp/pppoe.c int hash = hash_item(po->pppoe_pa.sid, po->pppoe_pa.remote); po 1076 drivers/net/ppp/pppoe.c po = NULL; po 1078 drivers/net/ppp/pppoe.c po = pn->hash_table[hash]; po 1079 drivers/net/ppp/pppoe.c if (po) po 1085 drivers/net/ppp/pppoe.c return po; po 70 drivers/net/ppp/pppox.c struct pppox_sock *po = pppox_sk(sk); po 83 drivers/net/ppp/pppox.c index = ppp_channel_index(&po->chan); po 134 drivers/net/ppp/pptp.c struct pppox_sock *po = pppox_sk(sk); po 136 drivers/net/ppp/pptp.c struct pptp_opt *opt = &po->proto.pptp; po 151 drivers/net/ppp/pptp.c if (sk_pppox(po)->sk_state & PPPOX_DEAD) po 257 drivers/net/ppp/pptp.c struct pppox_sock *po = pppox_sk(sk); po 258 drivers/net/ppp/pptp.c struct pptp_opt *opt = &po->proto.pptp; po 325 drivers/net/ppp/pptp.c ppp_input(&po->chan, skb); po 336 drivers/net/ppp/pptp.c struct pppox_sock *po; po 358 drivers/net/ppp/pptp.c po = lookup_chan(htons(header->call_id), iph->saddr); po 359 drivers/net/ppp/pptp.c if (po) { po 362 drivers/net/ppp/pptp.c return sk_receive_skb(sk_pppox(po), skb, 0); po 374 drivers/net/ppp/pptp.c struct pppox_sock *po = pppox_sk(sk); po 392 drivers/net/ppp/pptp.c if (add_chan(po, &sp->sa_addr.pptp)) po 407 drivers/net/ppp/pptp.c struct pppox_sock *po = pppox_sk(sk); po 408 drivers/net/ppp/pptp.c struct pptp_opt *opt = &po->proto.pptp; po 440 drivers/net/ppp/pptp.c po->chan.private = sk; po 441 drivers/net/ppp/pptp.c po->chan.ops = &pptp_chan_ops; po 454 drivers/net/ppp/pptp.c po->chan.mtu = dst_mtu(&rt->dst); po 455 drivers/net/ppp/pptp.c if (!po->chan.mtu) po 456 drivers/net/ppp/pptp.c po->chan.mtu = PPP_MRU; po 457 drivers/net/ppp/pptp.c po->chan.mtu -= PPTP_HEADER_OVERHEAD; po 459 drivers/net/ppp/pptp.c po->chan.hdrlen = 2 + sizeof(struct pptp_gre_header); po 460 drivers/net/ppp/pptp.c error = ppp_register_channel(&po->chan); po 494 drivers/net/ppp/pptp.c struct pppox_sock *po; po 507 drivers/net/ppp/pptp.c po = pppox_sk(sk); po 508 drivers/net/ppp/pptp.c del_chan(po); po 537 drivers/net/ppp/pptp.c struct pppox_sock *po; po 556 drivers/net/ppp/pptp.c po = pppox_sk(sk); po 557 drivers/net/ppp/pptp.c opt = &po->proto.pptp; po 571 drivers/net/ppp/pptp.c struct pppox_sock *po = pppox_sk(sk); po 572 drivers/net/ppp/pptp.c struct pptp_opt *opt = &po->proto.pptp; po 248 drivers/scsi/sym53c8xx_2/sym_fw.c u_short *po; po 255 drivers/scsi/sym53c8xx_2/sym_fw.c po = (u_short *) fw->a_ofs; po 258 drivers/scsi/sym53c8xx_2/sym_fw.c pa[i] = np->scripta_ba + po[i]; po 263 drivers/scsi/sym53c8xx_2/sym_fw.c po = (u_short *) fw->b_ofs; po 266 drivers/scsi/sym53c8xx_2/sym_fw.c pa[i] = np->scriptb_ba + po[i]; po 271 drivers/scsi/sym53c8xx_2/sym_fw.c po = (u_short *) fw->z_ofs; po 274 drivers/scsi/sym53c8xx_2/sym_fw.c pa[i] = np->scriptz_ba + po[i]; po 957 drivers/tty/serial/ifx6x60.c gpio_set_value(ifx_dev->gpio.po, 0); po 962 drivers/tty/serial/ifx6x60.c gpio_set_value(ifx_dev->gpio.po, 1); po 964 drivers/tty/serial/ifx6x60.c gpio_set_value(ifx_dev->gpio.po, 0); po 1083 drivers/tty/serial/ifx6x60.c ifx_dev->gpio.po = pl_data->pwr_on; po 1089 drivers/tty/serial/ifx6x60.c ifx_dev->gpio.reset, ifx_dev->gpio.po, ifx_dev->gpio.mrdy, po 1108 drivers/tty/serial/ifx6x60.c ret = gpio_request(ifx_dev->gpio.po, "ifxModem"); po 1109 drivers/tty/serial/ifx6x60.c ret += gpio_direction_output(ifx_dev->gpio.po, 0); po 1110 drivers/tty/serial/ifx6x60.c ret += gpio_export(ifx_dev->gpio.po, 1); po 1113 drivers/tty/serial/ifx6x60.c ifx_dev->gpio.po); po 1211 drivers/tty/serial/ifx6x60.c gpio_free(ifx_dev->gpio.po); po 1243 drivers/tty/serial/ifx6x60.c gpio_free(ifx_dev->gpio.po); po 100 drivers/tty/serial/ifx6x60.h unsigned short po; /* modem-on gpio */ po 65 include/linux/if_pppox.h static inline struct sock *sk_pppox(struct pppox_sock *po) po 67 include/linux/if_pppox.h return (struct sock *)po; po 2616 include/linux/netdevice.h void dev_add_offload(struct packet_offload *po); po 2617 include/linux/netdevice.h void dev_remove_offload(struct packet_offload *po); po 66 net/ceph/pagevec.c int po = off & ~PAGE_MASK; po 71 net/ceph/pagevec.c l = min_t(int, PAGE_SIZE-po, left); po 72 net/ceph/pagevec.c bad = copy_from_user(page_address(pages[i]) + po, data, l); po 77 net/ceph/pagevec.c po += l - bad; po 78 net/ceph/pagevec.c if (po == PAGE_SIZE) { po 79 net/ceph/pagevec.c po = 0; po 92 net/ceph/pagevec.c size_t po = off & ~PAGE_MASK; po 96 net/ceph/pagevec.c size_t l = min_t(size_t, PAGE_SIZE-po, left); po 98 net/ceph/pagevec.c memcpy(page_address(pages[i]) + po, data, l); po 101 net/ceph/pagevec.c po += l; po 102 net/ceph/pagevec.c if (po == PAGE_SIZE) { po 103 net/ceph/pagevec.c po = 0; po 115 net/ceph/pagevec.c size_t po = off & ~PAGE_MASK; po 119 net/ceph/pagevec.c size_t l = min_t(size_t, PAGE_SIZE-po, left); po 121 net/ceph/pagevec.c memcpy(data, page_address(pages[i]) + po, l); po 124 net/ceph/pagevec.c po += l; po 125 net/ceph/pagevec.c if (po == PAGE_SIZE) { po 126 net/ceph/pagevec.c po = 0; po 400 net/core/dev.c void dev_add_offload(struct packet_offload *po) po 406 net/core/dev.c if (po->priority < elem->priority) po 409 net/core/dev.c list_add_rcu(&po->list, elem->list.prev); po 427 net/core/dev.c static void __dev_remove_offload(struct packet_offload *po) po 435 net/core/dev.c if (po == po1) { po 436 net/core/dev.c list_del_rcu(&po->list); po 441 net/core/dev.c pr_warn("dev_remove_offload: %p not found\n", po); po 458 net/core/dev.c void dev_remove_offload(struct packet_offload *po) po 460 net/core/dev.c __dev_remove_offload(po); po 236 net/l2tp/l2tp_ppp.c struct pppox_sock *po; po 242 net/l2tp/l2tp_ppp.c po = pppox_sk(sk); po 243 net/l2tp/l2tp_ppp.c ppp_input(&po->chan, skb); po 539 net/l2tp/l2tp_ppp.c struct pppox_sock *po = pppox_sk(sk); po 541 net/l2tp/l2tp_ppp.c seq_printf(m, " interface %s\n", ppp_dev_name(&po->chan)); po 665 net/l2tp/l2tp_ppp.c struct pppox_sock *po = pppox_sk(sk); po 816 net/l2tp/l2tp_ppp.c po->chan.hdrlen = PPPOL2TP_L2TP_HDR_SIZE_NOSEQ; po 818 net/l2tp/l2tp_ppp.c po->chan.private = sk; po 819 net/l2tp/l2tp_ppp.c po->chan.ops = &pppol2tp_chan_ops; po 820 net/l2tp/l2tp_ppp.c po->chan.mtu = pppol2tp_tunnel_mtu(tunnel); po 822 net/l2tp/l2tp_ppp.c error = ppp_register_net_channel(sock_net(sk), &po->chan); po 1198 net/l2tp/l2tp_ppp.c struct pppox_sock *po = pppox_sk(sk); po 1200 net/l2tp/l2tp_ppp.c po->chan.hdrlen = val ? PPPOL2TP_L2TP_HDR_SIZE_SEQ : po 1591 net/l2tp/l2tp_ppp.c struct pppox_sock *po = pppox_sk(sk); po 1593 net/l2tp/l2tp_ppp.c seq_printf(m, " interface %s\n", ppp_dev_name(&po->chan)); po 38 net/netfilter/xt_set.c #define ADT_OPT(n, f, d, fs, cfs, t, p, b, po, bo) \ po 47 net/netfilter/xt_set.c .ext.packets_op = po, \ po 186 net/packet/af_packet.c static void *packet_previous_frame(struct packet_sock *po, po 235 net/packet/af_packet.c static void __fanout_unlink(struct sock *sk, struct packet_sock *po); po 236 net/packet/af_packet.c static void __fanout_link(struct sock *sk, struct packet_sock *po); po 243 net/packet/af_packet.c static struct net_device *packet_cached_dev_get(struct packet_sock *po) po 248 net/packet/af_packet.c dev = rcu_dereference(po->cached_dev); po 256 net/packet/af_packet.c static void packet_cached_dev_assign(struct packet_sock *po, po 259 net/packet/af_packet.c rcu_assign_pointer(po->cached_dev, dev); po 262 net/packet/af_packet.c static void packet_cached_dev_reset(struct packet_sock *po) po 264 net/packet/af_packet.c RCU_INIT_POINTER(po->cached_dev, NULL); po 267 net/packet/af_packet.c static bool packet_use_direct_xmit(const struct packet_sock *po) po 269 net/packet/af_packet.c return po->xmit == packet_direct_xmit; po 299 net/packet/af_packet.c struct packet_sock *po = pkt_sk(sk); po 301 net/packet/af_packet.c if (!po->running) { po 302 net/packet/af_packet.c if (po->fanout) po 303 net/packet/af_packet.c __fanout_link(sk, po); po 305 net/packet/af_packet.c dev_add_pack(&po->prot_hook); po 308 net/packet/af_packet.c po->running = 1; po 326 net/packet/af_packet.c struct packet_sock *po = pkt_sk(sk); po 328 net/packet/af_packet.c lockdep_assert_held_once(&po->bind_lock); po 330 net/packet/af_packet.c po->running = 0; po 332 net/packet/af_packet.c if (po->fanout) po 333 net/packet/af_packet.c __fanout_unlink(sk, po); po 335 net/packet/af_packet.c __dev_remove_pack(&po->prot_hook); po 340 net/packet/af_packet.c spin_unlock(&po->bind_lock); po 342 net/packet/af_packet.c spin_lock(&po->bind_lock); po 348 net/packet/af_packet.c struct packet_sock *po = pkt_sk(sk); po 350 net/packet/af_packet.c if (po->running) po 361 net/packet/af_packet.c static void __packet_set_status(struct packet_sock *po, void *frame, int status) po 366 net/packet/af_packet.c switch (po->tp_version) { po 387 net/packet/af_packet.c static int __packet_get_status(const struct packet_sock *po, void *frame) po 394 net/packet/af_packet.c switch (po->tp_version) { po 427 net/packet/af_packet.c static __u32 __packet_set_timestamp(struct packet_sock *po, void *frame, po 434 net/packet/af_packet.c if (!(ts_status = tpacket_get_timestamp(skb, &ts, po->tp_tstamp))) po 438 net/packet/af_packet.c switch (po->tp_version) { po 463 net/packet/af_packet.c static void *packet_lookup_frame(const struct packet_sock *po, po 477 net/packet/af_packet.c if (status != __packet_get_status(po, h.raw)) po 483 net/packet/af_packet.c static void *packet_current_frame(struct packet_sock *po, po 487 net/packet/af_packet.c return packet_lookup_frame(po, rb, rb->head, status); po 495 net/packet/af_packet.c static void prb_shutdown_retire_blk_timer(struct packet_sock *po, po 500 net/packet/af_packet.c pkc = GET_PBDQC_FROM_RB(&po->rx_ring); po 509 net/packet/af_packet.c static void prb_setup_retire_blk_timer(struct packet_sock *po) po 513 net/packet/af_packet.c pkc = GET_PBDQC_FROM_RB(&po->rx_ring); po 519 net/packet/af_packet.c static int prb_calc_retire_blk_tmo(struct packet_sock *po, po 528 net/packet/af_packet.c dev = __dev_get_by_index(sock_net(&po->sk), po->ifindex); po 568 net/packet/af_packet.c static void init_prb_bdqc(struct packet_sock *po, po 584 net/packet/af_packet.c p1->hdrlen = po->tp_hdrlen; po 585 net/packet/af_packet.c p1->version = po->tp_version; po 587 net/packet/af_packet.c po->stats.stats3.tp_freeze_q_cnt = 0; po 591 net/packet/af_packet.c p1->retire_blk_tov = prb_calc_retire_blk_tmo(po, po 598 net/packet/af_packet.c prb_setup_retire_blk_timer(po); po 637 net/packet/af_packet.c struct packet_sock *po = po 638 net/packet/af_packet.c from_timer(po, t, rx_ring.prb_bdqc.retire_blk_timer); po 639 net/packet/af_packet.c struct tpacket_kbdq_core *pkc = GET_PBDQC_FROM_RB(&po->rx_ring); po 643 net/packet/af_packet.c spin_lock(&po->sk.sk_receive_queue.lock); po 673 net/packet/af_packet.c prb_retire_current_block(pkc, po, TP_STATUS_BLK_TMO); po 674 net/packet/af_packet.c if (!prb_dispatch_next_block(pkc, po)) po 706 net/packet/af_packet.c spin_unlock(&po->sk.sk_receive_queue.lock); po 754 net/packet/af_packet.c struct packet_sock *po, unsigned int stat) po 760 net/packet/af_packet.c struct sock *sk = &po->sk; po 762 net/packet/af_packet.c if (atomic_read(&po->tp_drops)) po 867 net/packet/af_packet.c struct packet_sock *po) po 870 net/packet/af_packet.c po->stats.stats3.tp_freeze_q_cnt++; po 882 net/packet/af_packet.c struct packet_sock *po) po 893 net/packet/af_packet.c prb_freeze_queue(pkc, po); po 907 net/packet/af_packet.c struct packet_sock *po, unsigned int status) po 928 net/packet/af_packet.c prb_close_block(pkc, pbd, po, status); po 1005 net/packet/af_packet.c static void *__packet_lookup_frame_in_block(struct packet_sock *po, po 1014 net/packet/af_packet.c pkc = GET_PBDQC_FROM_RB(&po->rx_ring); po 1049 net/packet/af_packet.c prb_retire_current_block(pkc, po, 0); po 1052 net/packet/af_packet.c curr = (char *)prb_dispatch_next_block(pkc, po); po 1066 net/packet/af_packet.c static void *packet_current_rx_frame(struct packet_sock *po, po 1071 net/packet/af_packet.c switch (po->tp_version) { po 1074 net/packet/af_packet.c curr = packet_lookup_frame(po, &po->rx_ring, po 1075 net/packet/af_packet.c po->rx_ring.head, status); po 1078 net/packet/af_packet.c return __packet_lookup_frame_in_block(po, skb, len); po 1086 net/packet/af_packet.c static void *prb_lookup_block(const struct packet_sock *po, po 1110 net/packet/af_packet.c static void *__prb_previous_block(struct packet_sock *po, po 1115 net/packet/af_packet.c return prb_lookup_block(po, rb, previous, status); po 1118 net/packet/af_packet.c static void *packet_previous_rx_frame(struct packet_sock *po, po 1122 net/packet/af_packet.c if (po->tp_version <= TPACKET_V2) po 1123 net/packet/af_packet.c return packet_previous_frame(po, rb, status); po 1125 net/packet/af_packet.c return __prb_previous_block(po, rb, status); po 1128 net/packet/af_packet.c static void packet_increment_rx_head(struct packet_sock *po, po 1131 net/packet/af_packet.c switch (po->tp_version) { po 1143 net/packet/af_packet.c static void *packet_previous_frame(struct packet_sock *po, po 1148 net/packet/af_packet.c return packet_lookup_frame(po, rb, previous, status); po 1181 net/packet/af_packet.c static int packet_alloc_pending(struct packet_sock *po) po 1183 net/packet/af_packet.c po->rx_ring.pending_refcnt = NULL; po 1185 net/packet/af_packet.c po->tx_ring.pending_refcnt = alloc_percpu(unsigned int); po 1186 net/packet/af_packet.c if (unlikely(po->tx_ring.pending_refcnt == NULL)) po 1192 net/packet/af_packet.c static void packet_free_pending(struct packet_sock *po) po 1194 net/packet/af_packet.c free_percpu(po->tx_ring.pending_refcnt); po 1202 net/packet/af_packet.c static bool __tpacket_has_room(const struct packet_sock *po, int pow_off) po 1206 net/packet/af_packet.c len = READ_ONCE(po->rx_ring.frame_max) + 1; po 1207 net/packet/af_packet.c idx = READ_ONCE(po->rx_ring.head); po 1212 net/packet/af_packet.c return packet_lookup_frame(po, &po->rx_ring, idx, TP_STATUS_KERNEL); po 1215 net/packet/af_packet.c static bool __tpacket_v3_has_room(const struct packet_sock *po, int pow_off) po 1219 net/packet/af_packet.c len = READ_ONCE(po->rx_ring.prb_bdqc.knum_blocks); po 1220 net/packet/af_packet.c idx = READ_ONCE(po->rx_ring.prb_bdqc.kactive_blk_num); po 1225 net/packet/af_packet.c return prb_lookup_block(po, &po->rx_ring, idx, TP_STATUS_KERNEL); po 1228 net/packet/af_packet.c static int __packet_rcv_has_room(const struct packet_sock *po, po 1231 net/packet/af_packet.c const struct sock *sk = &po->sk; po 1234 net/packet/af_packet.c if (po->prot_hook.func != tpacket_rcv) { po 1247 net/packet/af_packet.c if (po->tp_version == TPACKET_V3) { po 1248 net/packet/af_packet.c if (__tpacket_v3_has_room(po, ROOM_POW_OFF)) po 1250 net/packet/af_packet.c else if (__tpacket_v3_has_room(po, 0)) po 1253 net/packet/af_packet.c if (__tpacket_has_room(po, ROOM_POW_OFF)) po 1255 net/packet/af_packet.c else if (__tpacket_has_room(po, 0)) po 1262 net/packet/af_packet.c static int packet_rcv_has_room(struct packet_sock *po, struct sk_buff *skb) po 1266 net/packet/af_packet.c ret = __packet_rcv_has_room(po, skb); po 1269 net/packet/af_packet.c if (READ_ONCE(po->pressure) != pressure) po 1270 net/packet/af_packet.c WRITE_ONCE(po->pressure, pressure); po 1275 net/packet/af_packet.c static void packet_rcv_try_clear_pressure(struct packet_sock *po) po 1277 net/packet/af_packet.c if (READ_ONCE(po->pressure) && po 1278 net/packet/af_packet.c __packet_rcv_has_room(po, NULL) == ROOM_NORMAL) po 1279 net/packet/af_packet.c WRITE_ONCE(po->pressure, 0); po 1297 net/packet/af_packet.c static bool fanout_flow_is_huge(struct packet_sock *po, struct sk_buff *skb) po 1299 net/packet/af_packet.c u32 *history = po->rollover->history; po 1352 net/packet/af_packet.c struct packet_sock *po, *po_next, *po_skip = NULL; po 1355 net/packet/af_packet.c po = pkt_sk(f->arr[idx]); po 1358 net/packet/af_packet.c room = packet_rcv_has_room(po, skb); po 1360 net/packet/af_packet.c (room == ROOM_LOW && !fanout_flow_is_huge(po, skb))) po 1362 net/packet/af_packet.c po_skip = po; po 1365 net/packet/af_packet.c i = j = min_t(int, po->rollover->sock, num - 1); po 1371 net/packet/af_packet.c po->rollover->sock = i; po 1372 net/packet/af_packet.c atomic_long_inc(&po->rollover->num); po 1374 net/packet/af_packet.c atomic_long_inc(&po->rollover->num_huge); po 1382 net/packet/af_packet.c atomic_long_inc(&po->rollover->num_failed); po 1420 net/packet/af_packet.c struct packet_sock *po; po 1462 net/packet/af_packet.c po = pkt_sk(f->arr[idx]); po 1463 net/packet/af_packet.c return po->prot_hook.func(skb, dev, &po->prot_hook, orig_dev); po 1471 net/packet/af_packet.c static void __fanout_link(struct sock *sk, struct packet_sock *po) po 1473 net/packet/af_packet.c struct packet_fanout *f = po->fanout; po 1484 net/packet/af_packet.c static void __fanout_unlink(struct sock *sk, struct packet_sock *po) po 1486 net/packet/af_packet.c struct packet_fanout *f = po->fanout; po 1538 net/packet/af_packet.c static int fanout_set_data_cbpf(struct packet_sock *po, char __user *data, po 1545 net/packet/af_packet.c if (sock_flag(&po->sk, SOCK_FILTER_LOCKED)) po 1556 net/packet/af_packet.c __fanout_set_data_bpf(po->fanout, new); po 1560 net/packet/af_packet.c static int fanout_set_data_ebpf(struct packet_sock *po, char __user *data, po 1566 net/packet/af_packet.c if (sock_flag(&po->sk, SOCK_FILTER_LOCKED)) po 1577 net/packet/af_packet.c __fanout_set_data_bpf(po->fanout, new); po 1581 net/packet/af_packet.c static int fanout_set_data(struct packet_sock *po, char __user *data, po 1584 net/packet/af_packet.c switch (po->fanout->type) { po 1586 net/packet/af_packet.c return fanout_set_data_cbpf(po, data, len); po 1588 net/packet/af_packet.c return fanout_set_data_ebpf(po, data, len); po 1636 net/packet/af_packet.c struct packet_sock *po = pkt_sk(sk); po 1661 net/packet/af_packet.c if (po->fanout) po 1712 net/packet/af_packet.c match->prot_hook.type = po->prot_hook.type; po 1713 net/packet/af_packet.c match->prot_hook.dev = po->prot_hook.dev; po 1721 net/packet/af_packet.c spin_lock(&po->bind_lock); po 1722 net/packet/af_packet.c if (po->running && po 1724 net/packet/af_packet.c match->prot_hook.type == po->prot_hook.type && po 1725 net/packet/af_packet.c match->prot_hook.dev == po->prot_hook.dev) { po 1728 net/packet/af_packet.c __dev_remove_pack(&po->prot_hook); po 1729 net/packet/af_packet.c po->fanout = match; po 1730 net/packet/af_packet.c po->rollover = rollover; po 1733 net/packet/af_packet.c __fanout_link(sk, po); po 1737 net/packet/af_packet.c spin_unlock(&po->bind_lock); po 1757 net/packet/af_packet.c struct packet_sock *po = pkt_sk(sk); po 1761 net/packet/af_packet.c f = po->fanout; po 1763 net/packet/af_packet.c po->fanout = NULL; po 2049 net/packet/af_packet.c struct packet_sock *po; po 2059 net/packet/af_packet.c po = pkt_sk(sk); po 2111 net/packet/af_packet.c if (unlikely(po->origdev)) po 2134 net/packet/af_packet.c po->stats.stats1.tp_packets++; po 2143 net/packet/af_packet.c atomic_inc(&po->tp_drops); po 2163 net/packet/af_packet.c struct packet_sock *po; po 2189 net/packet/af_packet.c po = pkt_sk(sk); po 2210 net/packet/af_packet.c if (__packet_rcv_has_room(po, skb) == ROOM_NONE) { po 2211 net/packet/af_packet.c atomic_inc(&po->tp_drops); po 2226 net/packet/af_packet.c macoff = netoff = TPACKET_ALIGN(po->tp_hdrlen) + 16 + po 2227 net/packet/af_packet.c po->tp_reserve; po 2230 net/packet/af_packet.c netoff = TPACKET_ALIGN(po->tp_hdrlen + po 2232 net/packet/af_packet.c po->tp_reserve; po 2233 net/packet/af_packet.c if (po->has_vnet_hdr) { po 2239 net/packet/af_packet.c if (po->tp_version <= TPACKET_V2) { po 2240 net/packet/af_packet.c if (macoff + snaplen > po->rx_ring.frame_size) { po 2241 net/packet/af_packet.c if (po->copy_thresh && po 2252 net/packet/af_packet.c snaplen = po->rx_ring.frame_size - macoff; po 2259 net/packet/af_packet.c GET_PBDQC_FROM_RB(&po->rx_ring)->max_frame_len)) { po 2262 net/packet/af_packet.c nval = GET_PBDQC_FROM_RB(&po->rx_ring)->max_frame_len - macoff; po 2268 net/packet/af_packet.c macoff = GET_PBDQC_FROM_RB(&po->rx_ring)->max_frame_len; po 2273 net/packet/af_packet.c h.raw = packet_current_rx_frame(po, skb, po 2278 net/packet/af_packet.c if (po->tp_version <= TPACKET_V2) { po 2279 net/packet/af_packet.c slot_id = po->rx_ring.head; po 2280 net/packet/af_packet.c if (test_bit(slot_id, po->rx_ring.rx_owner_map)) po 2282 net/packet/af_packet.c __set_bit(slot_id, po->rx_ring.rx_owner_map); po 2291 net/packet/af_packet.c if (po->tp_version <= TPACKET_V2) { po 2292 net/packet/af_packet.c packet_increment_rx_head(po, &po->rx_ring); po 2299 net/packet/af_packet.c if (atomic_read(&po->tp_drops)) po 2303 net/packet/af_packet.c po->stats.stats1.tp_packets++; po 2312 net/packet/af_packet.c if (!(ts_status = tpacket_get_timestamp(skb, &ts, po->tp_tstamp))) po 2317 net/packet/af_packet.c switch (po->tp_version) { po 2369 net/packet/af_packet.c if (unlikely(po->origdev)) po 2377 net/packet/af_packet.c if (po->tp_version <= TPACKET_V2) { po 2389 net/packet/af_packet.c if (po->tp_version <= TPACKET_V2) { po 2391 net/packet/af_packet.c __packet_set_status(po, h.raw, status); po 2392 net/packet/af_packet.c __clear_bit(slot_id, po->rx_ring.rx_owner_map); po 2396 net/packet/af_packet.c prb_clear_blk_fill_status(&po->rx_ring); po 2413 net/packet/af_packet.c atomic_inc(&po->tp_drops); po 2423 net/packet/af_packet.c struct packet_sock *po = pkt_sk(skb->sk); po 2425 net/packet/af_packet.c if (likely(po->tx_ring.pg_vec)) { po 2430 net/packet/af_packet.c packet_dec_pending(&po->tx_ring); po 2432 net/packet/af_packet.c ts = __packet_set_timestamp(po, ph, skb); po 2433 net/packet/af_packet.c __packet_set_status(po, ph, TP_STATUS_AVAILABLE | ts); po 2435 net/packet/af_packet.c if (!packet_read_pending(&po->tx_ring)) po 2436 net/packet/af_packet.c complete(&po->skb_completion); po 2471 net/packet/af_packet.c static int tpacket_fill_skb(struct packet_sock *po, struct sk_buff *skb, po 2478 net/packet/af_packet.c struct socket *sock = po->sk.sk_socket; po 2486 net/packet/af_packet.c skb->priority = po->sk.sk_priority; po 2487 net/packet/af_packet.c skb->mark = po->sk.sk_mark; po 2524 net/packet/af_packet.c refcount_add(to_write, &po->sk.sk_wmem_alloc); po 2551 net/packet/af_packet.c static int tpacket_parse_header(struct packet_sock *po, void *frame, po 2559 net/packet/af_packet.c switch (po->tp_version) { po 2579 net/packet/af_packet.c if (unlikely(po->tp_tx_has_off)) { po 2582 net/packet/af_packet.c off_min = po->tp_hdrlen - sizeof(struct sockaddr_ll); po 2583 net/packet/af_packet.c off_max = po->tx_ring.frame_size - tp_len; po 2584 net/packet/af_packet.c if (po->sk.sk_type == SOCK_DGRAM) { po 2585 net/packet/af_packet.c switch (po->tp_version) { po 2597 net/packet/af_packet.c switch (po->tp_version) { po 2612 net/packet/af_packet.c off = po->tp_hdrlen - sizeof(struct sockaddr_ll); po 2619 net/packet/af_packet.c static int tpacket_snd(struct packet_sock *po, struct msghdr *msg) po 2638 net/packet/af_packet.c mutex_lock(&po->pg_vec_lock); po 2643 net/packet/af_packet.c if (unlikely(!po->tx_ring.pg_vec)) { po 2648 net/packet/af_packet.c dev = packet_cached_dev_get(po); po 2649 net/packet/af_packet.c proto = po->num; po 2659 net/packet/af_packet.c dev = dev_get_by_index(sock_net(&po->sk), saddr->sll_ifindex); po 2660 net/packet/af_packet.c if (po->sk.sk_socket->type == SOCK_DGRAM) { po 2675 net/packet/af_packet.c sockcm_init(&sockc, &po->sk); po 2677 net/packet/af_packet.c err = sock_cmsg_send(&po->sk, msg, &sockc); po 2682 net/packet/af_packet.c if (po->sk.sk_socket->type == SOCK_RAW) po 2684 net/packet/af_packet.c size_max = po->tx_ring.frame_size po 2685 net/packet/af_packet.c - (po->tp_hdrlen - sizeof(struct sockaddr_ll)); po 2687 net/packet/af_packet.c if ((size_max > dev->mtu + reserve + VLAN_HLEN) && !po->has_vnet_hdr) po 2690 net/packet/af_packet.c reinit_completion(&po->skb_completion); po 2693 net/packet/af_packet.c ph = packet_current_frame(po, &po->tx_ring, po 2697 net/packet/af_packet.c timeo = sock_sndtimeo(&po->sk, msg->msg_flags & MSG_DONTWAIT); po 2698 net/packet/af_packet.c timeo = wait_for_completion_interruptible_timeout(&po->skb_completion, timeo); po 2709 net/packet/af_packet.c tp_len = tpacket_parse_header(po, ph, size_max, &data); po 2716 net/packet/af_packet.c if (po->has_vnet_hdr) { po 2729 net/packet/af_packet.c skb = sock_alloc_send_skb(&po->sk, po 2740 net/packet/af_packet.c tp_len = tpacket_fill_skb(po, skb, ph, dev, data, tp_len, proto, po 2744 net/packet/af_packet.c !po->has_vnet_hdr && po 2750 net/packet/af_packet.c if (po->tp_loss) { po 2751 net/packet/af_packet.c __packet_set_status(po, ph, po 2753 net/packet/af_packet.c packet_increment_head(&po->tx_ring); po 2763 net/packet/af_packet.c if (po->has_vnet_hdr) { po 2772 net/packet/af_packet.c __packet_set_status(po, ph, TP_STATUS_SENDING); po 2773 net/packet/af_packet.c packet_inc_pending(&po->tx_ring); po 2776 net/packet/af_packet.c err = po->xmit(skb); po 2779 net/packet/af_packet.c if (err && __packet_get_status(po, ph) == po 2791 net/packet/af_packet.c packet_increment_head(&po->tx_ring); po 2800 net/packet/af_packet.c (need_wait && packet_read_pending(&po->tx_ring)))); po 2806 net/packet/af_packet.c __packet_set_status(po, ph, status); po 2811 net/packet/af_packet.c mutex_unlock(&po->pg_vec_lock); po 2851 net/packet/af_packet.c struct packet_sock *po = pkt_sk(sk); po 2861 net/packet/af_packet.c dev = packet_cached_dev_get(po); po 2862 net/packet/af_packet.c proto = po->num; po 2896 net/packet/af_packet.c if (po->has_vnet_hdr) { po 2978 net/packet/af_packet.c err = po->xmit(skb); po 2998 net/packet/af_packet.c struct packet_sock *po = pkt_sk(sk); po 3000 net/packet/af_packet.c if (po->tx_ring.pg_vec) po 3001 net/packet/af_packet.c return tpacket_snd(po, msg); po 3014 net/packet/af_packet.c struct packet_sock *po; po 3023 net/packet/af_packet.c po = pkt_sk(sk); po 3033 net/packet/af_packet.c spin_lock(&po->bind_lock); po 3035 net/packet/af_packet.c packet_cached_dev_reset(po); po 3037 net/packet/af_packet.c if (po->prot_hook.dev) { po 3038 net/packet/af_packet.c dev_put(po->prot_hook.dev); po 3039 net/packet/af_packet.c po->prot_hook.dev = NULL; po 3041 net/packet/af_packet.c spin_unlock(&po->bind_lock); po 3046 net/packet/af_packet.c if (po->rx_ring.pg_vec) { po 3051 net/packet/af_packet.c if (po->tx_ring.pg_vec) { po 3061 net/packet/af_packet.c kfree(po->rollover); po 3075 net/packet/af_packet.c packet_free_pending(po); po 3089 net/packet/af_packet.c struct packet_sock *po = pkt_sk(sk); po 3098 net/packet/af_packet.c spin_lock(&po->bind_lock); po 3101 net/packet/af_packet.c if (po->fanout) { po 3123 net/packet/af_packet.c proto_curr = po->prot_hook.type; po 3124 net/packet/af_packet.c dev_curr = po->prot_hook.dev; po 3129 net/packet/af_packet.c if (po->running) { po 3134 net/packet/af_packet.c po->num = 0; po 3137 net/packet/af_packet.c dev_curr = po->prot_hook.dev; po 3143 net/packet/af_packet.c BUG_ON(po->running); po 3144 net/packet/af_packet.c po->num = proto; po 3145 net/packet/af_packet.c po->prot_hook.type = proto; po 3149 net/packet/af_packet.c po->prot_hook.dev = NULL; po 3150 net/packet/af_packet.c po->ifindex = -1; po 3151 net/packet/af_packet.c packet_cached_dev_reset(po); po 3153 net/packet/af_packet.c po->prot_hook.dev = dev; po 3154 net/packet/af_packet.c po->ifindex = dev ? dev->ifindex : 0; po 3155 net/packet/af_packet.c packet_cached_dev_assign(po, dev); po 3174 net/packet/af_packet.c spin_unlock(&po->bind_lock); po 3236 net/packet/af_packet.c struct packet_sock *po; po 3259 net/packet/af_packet.c po = pkt_sk(sk); po 3260 net/packet/af_packet.c init_completion(&po->skb_completion); po 3262 net/packet/af_packet.c po->num = proto; po 3263 net/packet/af_packet.c po->xmit = dev_queue_xmit; po 3265 net/packet/af_packet.c err = packet_alloc_pending(po); po 3269 net/packet/af_packet.c packet_cached_dev_reset(po); po 3278 net/packet/af_packet.c spin_lock_init(&po->bind_lock); po 3279 net/packet/af_packet.c mutex_init(&po->pg_vec_lock); po 3280 net/packet/af_packet.c po->rollover = NULL; po 3281 net/packet/af_packet.c po->prot_hook.func = packet_rcv; po 3284 net/packet/af_packet.c po->prot_hook.func = packet_rcv_spkt; po 3286 net/packet/af_packet.c po->prot_hook.af_packet_priv = sk; po 3289 net/packet/af_packet.c po->prot_hook.type = proto; po 3481 net/packet/af_packet.c struct packet_sock *po = pkt_sk(sk); po 3488 net/packet/af_packet.c sll->sll_ifindex = po->ifindex; po 3489 net/packet/af_packet.c sll->sll_protocol = po->num; po 3492 net/packet/af_packet.c dev = dev_get_by_index_rcu(sock_net(sk), po->ifindex); po 3553 net/packet/af_packet.c struct packet_sock *po = pkt_sk(sk); po 3575 net/packet/af_packet.c for (ml = po->mclist; ml; ml = ml->next) { po 3593 net/packet/af_packet.c i->next = po->mclist; po 3594 net/packet/af_packet.c po->mclist = i; po 3597 net/packet/af_packet.c po->mclist = i->next; po 3634 net/packet/af_packet.c struct packet_sock *po = pkt_sk(sk); po 3637 net/packet/af_packet.c if (!po->mclist) po 3641 net/packet/af_packet.c while ((ml = po->mclist) != NULL) { po 3644 net/packet/af_packet.c po->mclist = ml->next; po 3657 net/packet/af_packet.c struct packet_sock *po = pkt_sk(sk); po 3692 net/packet/af_packet.c switch (po->tp_version) { po 3743 net/packet/af_packet.c if (po->rx_ring.pg_vec || po->tx_ring.pg_vec) { po 3746 net/packet/af_packet.c po->tp_version = val; po 3763 net/packet/af_packet.c if (po->rx_ring.pg_vec || po->tx_ring.pg_vec) { po 3766 net/packet/af_packet.c po->tp_reserve = val; po 3782 net/packet/af_packet.c if (po->rx_ring.pg_vec || po->tx_ring.pg_vec) { po 3785 net/packet/af_packet.c po->tp_loss = !!val; po 3801 net/packet/af_packet.c po->auxdata = !!val; po 3815 net/packet/af_packet.c po->origdev = !!val; po 3831 net/packet/af_packet.c if (po->rx_ring.pg_vec || po->tx_ring.pg_vec) { po 3834 net/packet/af_packet.c po->has_vnet_hdr = !!val; po 3849 net/packet/af_packet.c po->tp_tstamp = val; po 3865 net/packet/af_packet.c if (!po->fanout) po 3868 net/packet/af_packet.c return fanout_set_data(po, optval, optlen); po 3881 net/packet/af_packet.c po->prot_hook.ignore_outgoing = !!val; po 3894 net/packet/af_packet.c if (po->rx_ring.pg_vec || po->tx_ring.pg_vec) { po 3897 net/packet/af_packet.c po->tp_tx_has_off = !!val; po 3912 net/packet/af_packet.c po->xmit = val ? packet_direct_xmit : dev_queue_xmit; po 3926 net/packet/af_packet.c struct packet_sock *po = pkt_sk(sk); po 3944 net/packet/af_packet.c memcpy(&st, &po->stats, sizeof(st)); po 3945 net/packet/af_packet.c memset(&po->stats, 0, sizeof(po->stats)); po 3947 net/packet/af_packet.c drops = atomic_xchg(&po->tp_drops, 0); po 3949 net/packet/af_packet.c if (po->tp_version == TPACKET_V3) { po 3963 net/packet/af_packet.c val = po->auxdata; po 3966 net/packet/af_packet.c val = po->origdev; po 3969 net/packet/af_packet.c val = po->has_vnet_hdr; po 3972 net/packet/af_packet.c val = po->tp_version; po 3996 net/packet/af_packet.c val = po->tp_reserve; po 3999 net/packet/af_packet.c val = po->tp_loss; po 4002 net/packet/af_packet.c val = po->tp_tstamp; po 4005 net/packet/af_packet.c val = (po->fanout ? po 4006 net/packet/af_packet.c ((u32)po->fanout->id | po 4007 net/packet/af_packet.c ((u32)po->fanout->type << 16) | po 4008 net/packet/af_packet.c ((u32)po->fanout->flags << 24)) : po 4012 net/packet/af_packet.c val = po->prot_hook.ignore_outgoing; po 4015 net/packet/af_packet.c if (!po->rollover) po 4017 net/packet/af_packet.c rstats.tp_all = atomic_long_read(&po->rollover->num); po 4018 net/packet/af_packet.c rstats.tp_huge = atomic_long_read(&po->rollover->num_huge); po 4019 net/packet/af_packet.c rstats.tp_failed = atomic_long_read(&po->rollover->num_failed); po 4024 net/packet/af_packet.c val = po->tp_tx_has_off; po 4027 net/packet/af_packet.c val = packet_use_direct_xmit(po); po 4047 net/packet/af_packet.c struct packet_sock *po = pkt_sk(sock->sk); po 4053 net/packet/af_packet.c po->fanout && po->fanout->type == PACKET_FANOUT_CBPF) { po 4073 net/packet/af_packet.c struct packet_sock *po = pkt_sk(sk); po 4077 net/packet/af_packet.c if (po->mclist) po 4078 net/packet/af_packet.c packet_dev_mclist_delete(dev, &po->mclist); po 4082 net/packet/af_packet.c if (dev->ifindex == po->ifindex) { po 4083 net/packet/af_packet.c spin_lock(&po->bind_lock); po 4084 net/packet/af_packet.c if (po->running) { po 4091 net/packet/af_packet.c packet_cached_dev_reset(po); po 4092 net/packet/af_packet.c po->ifindex = -1; po 4093 net/packet/af_packet.c if (po->prot_hook.dev) po 4094 net/packet/af_packet.c dev_put(po->prot_hook.dev); po 4095 net/packet/af_packet.c po->prot_hook.dev = NULL; po 4097 net/packet/af_packet.c spin_unlock(&po->bind_lock); po 4101 net/packet/af_packet.c if (dev->ifindex == po->ifindex) { po 4102 net/packet/af_packet.c spin_lock(&po->bind_lock); po 4103 net/packet/af_packet.c if (po->num) po 4105 net/packet/af_packet.c spin_unlock(&po->bind_lock); po 4167 net/packet/af_packet.c struct packet_sock *po = pkt_sk(sk); po 4171 net/packet/af_packet.c if (po->rx_ring.pg_vec) { po 4172 net/packet/af_packet.c if (!packet_previous_rx_frame(po, &po->rx_ring, po 4176 net/packet/af_packet.c packet_rcv_try_clear_pressure(po); po 4179 net/packet/af_packet.c if (po->tx_ring.pg_vec) { po 4180 net/packet/af_packet.c if (packet_current_frame(po, &po->tx_ring, TP_STATUS_AVAILABLE)) po 4289 net/packet/af_packet.c struct packet_sock *po = pkt_sk(sk); po 4299 net/packet/af_packet.c rb = tx_ring ? &po->tx_ring : &po->rx_ring; po 4304 net/packet/af_packet.c if (atomic_read(&po->mapped)) po 4318 net/packet/af_packet.c switch (po->tp_version) { po 4320 net/packet/af_packet.c po->tp_hdrlen = TPACKET_HDRLEN; po 4323 net/packet/af_packet.c po->tp_hdrlen = TPACKET2_HDRLEN; po 4326 net/packet/af_packet.c po->tp_hdrlen = TPACKET3_HDRLEN; po 4335 net/packet/af_packet.c min_frame_size = po->tp_hdrlen + po->tp_reserve; po 4336 net/packet/af_packet.c if (po->tp_version >= TPACKET_V3 && po 4359 net/packet/af_packet.c switch (po->tp_version) { po 4363 net/packet/af_packet.c init_prb_bdqc(po, rb, pg_vec, req_u); po 4394 net/packet/af_packet.c spin_lock(&po->bind_lock); po 4395 net/packet/af_packet.c was_running = po->running; po 4396 net/packet/af_packet.c num = po->num; po 4398 net/packet/af_packet.c po->num = 0; po 4401 net/packet/af_packet.c spin_unlock(&po->bind_lock); po 4406 net/packet/af_packet.c mutex_lock(&po->pg_vec_lock); po 4407 net/packet/af_packet.c if (closing || atomic_read(&po->mapped) == 0) { po 4411 net/packet/af_packet.c if (po->tp_version <= TPACKET_V2) po 4422 net/packet/af_packet.c po->prot_hook.func = (po->rx_ring.pg_vec) ? po 4425 net/packet/af_packet.c if (atomic_read(&po->mapped)) po 4427 net/packet/af_packet.c atomic_read(&po->mapped)); po 4429 net/packet/af_packet.c mutex_unlock(&po->pg_vec_lock); po 4431 net/packet/af_packet.c spin_lock(&po->bind_lock); po 4433 net/packet/af_packet.c po->num = num; po 4436 net/packet/af_packet.c spin_unlock(&po->bind_lock); po 4437 net/packet/af_packet.c if (pg_vec && (po->tp_version > TPACKET_V2)) { po 4440 net/packet/af_packet.c prb_shutdown_retire_blk_timer(po, rb_queue); po 4455 net/packet/af_packet.c struct packet_sock *po = pkt_sk(sk); po 4465 net/packet/af_packet.c mutex_lock(&po->pg_vec_lock); po 4468 net/packet/af_packet.c for (rb = &po->rx_ring; rb <= &po->tx_ring; rb++) { po 4484 net/packet/af_packet.c for (rb = &po->rx_ring; rb <= &po->tx_ring; rb++) { po 4504 net/packet/af_packet.c atomic_inc(&po->mapped); po 4509 net/packet/af_packet.c mutex_unlock(&po->pg_vec_lock); po 4599 net/packet/af_packet.c const struct packet_sock *po = pkt_sk(s); po 4606 net/packet/af_packet.c ntohs(po->num), po 4607 net/packet/af_packet.c po->ifindex, po 4608 net/packet/af_packet.c po->running, po 13 net/packet/diag.c static int pdiag_put_info(const struct packet_sock *po, struct sk_buff *nlskb) po 17 net/packet/diag.c pinfo.pdi_index = po->ifindex; po 18 net/packet/diag.c pinfo.pdi_version = po->tp_version; po 19 net/packet/diag.c pinfo.pdi_reserve = po->tp_reserve; po 20 net/packet/diag.c pinfo.pdi_copy_thresh = po->copy_thresh; po 21 net/packet/diag.c pinfo.pdi_tstamp = po->tp_tstamp; po 24 net/packet/diag.c if (po->running) po 26 net/packet/diag.c if (po->auxdata) po 28 net/packet/diag.c if (po->origdev) po 30 net/packet/diag.c if (po->has_vnet_hdr) po 32 net/packet/diag.c if (po->tp_loss) po 38 net/packet/diag.c static int pdiag_put_mclist(const struct packet_sock *po, struct sk_buff *nlskb) po 48 net/packet/diag.c for (ml = po->mclist; ml; ml = ml->next) { po 98 net/packet/diag.c static int pdiag_put_rings_cfg(struct packet_sock *po, struct sk_buff *skb) po 102 net/packet/diag.c mutex_lock(&po->pg_vec_lock); po 103 net/packet/diag.c ret = pdiag_put_ring(&po->rx_ring, po->tp_version, po 106 net/packet/diag.c ret = pdiag_put_ring(&po->tx_ring, po->tp_version, po 108 net/packet/diag.c mutex_unlock(&po->pg_vec_lock); po 113 net/packet/diag.c static int pdiag_put_fanout(struct packet_sock *po, struct sk_buff *nlskb) po 118 net/packet/diag.c if (po->fanout) { po 121 net/packet/diag.c val = (u32)po->fanout->id | ((u32)po->fanout->type << 16); po 137 net/packet/diag.c struct packet_sock *po = pkt_sk(sk); po 146 net/packet/diag.c rp->pdiag_num = ntohs(po->num); po 151 net/packet/diag.c pdiag_put_info(po, skb)) po 160 net/packet/diag.c pdiag_put_mclist(po, skb)) po 164 net/packet/diag.c pdiag_put_rings_cfg(po, skb)) po 168 net/packet/diag.c pdiag_put_fanout(po, skb)) po 73 sound/usb/line6/playback.c short *po = (short *)urb_out->transfer_buffer; po 76 sound/usb/line6/playback.c po[0] = pi[0]; po 77 sound/usb/line6/playback.c po[1] = 0; po 79 sound/usb/line6/playback.c po += 2; po 84 sound/usb/line6/playback.c unsigned char *po = urb_out->transfer_buffer; po 88 sound/usb/line6/playback.c po[j] = pi[j]; po 91 sound/usb/line6/playback.c po[j] = 0; po 94 sound/usb/line6/playback.c po += bytes_per_frame; po 115 sound/usb/line6/playback.c __le16 *pi, *po, *buf_end; po 118 sound/usb/line6/playback.c po = (__le16 *)urb_out->transfer_buffer; po 119 sound/usb/line6/playback.c buf_end = po + urb_out->transfer_buffer_length / sizeof(*po); po 121 sound/usb/line6/playback.c for (; po < buf_end; ++pi, ++po) { po 122 sound/usb/line6/playback.c short pov = le16_to_cpu(*po); po 126 sound/usb/line6/playback.c *po = cpu_to_le16(pov);