iucv 1314 drivers/s390/net/qeth_l3_main.c struct af_iucv_trans_hdr *iucv = (struct af_iucv_trans_hdr *) skb->data; iucv 1317 drivers/s390/net/qeth_l3_main.c if (IS_IQD(card) && iucv->magic == ETH_P_AF_IUCV) { iucv 153 net/iucv/af_iucv.c struct iucv_sock *iucv; iucv 161 net/iucv/af_iucv.c iucv = iucv_sk(sk); iucv 175 net/iucv/af_iucv.c skb_queue_purge(&iucv->send_skb_q); iucv 176 net/iucv/af_iucv.c skb_queue_purge(&iucv->backlog_skb_q); iucv 288 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv 292 net/iucv/af_iucv.c if (iucv->transport == AF_IUCV_TRANS_IUCV) iucv 293 net/iucv/af_iucv.c return (skb_queue_len(&iucv->send_skb_q) < iucv->path->msglim); iucv 295 net/iucv/af_iucv.c return ((atomic_read(&iucv->msg_sent) < iucv->msglimit_peer) && iucv 296 net/iucv/af_iucv.c (atomic_read(&iucv->pendings) <= 0)); iucv 320 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sock); iucv 333 net/iucv/af_iucv.c phs_hdr->window = iucv->msglimit; iucv 335 net/iucv/af_iucv.c confirm_recv = atomic_read(&iucv->msg_recv); iucv 340 net/iucv/af_iucv.c memcpy(phs_hdr->destUserID, iucv->dst_user_id, 8); iucv 341 net/iucv/af_iucv.c memcpy(phs_hdr->destAppName, iucv->dst_name, 8); iucv 342 net/iucv/af_iucv.c memcpy(phs_hdr->srcUserID, iucv->src_user_id, 8); iucv 343 net/iucv/af_iucv.c memcpy(phs_hdr->srcAppName, iucv->src_name, 8); iucv 351 net/iucv/af_iucv.c skb->dev = iucv->hs_dev; iucv 379 net/iucv/af_iucv.c skb_queue_tail(&iucv->send_skb_q, nskb); iucv 382 net/iucv/af_iucv.c skb_unlink(nskb, &iucv->send_skb_q); iucv 385 net/iucv/af_iucv.c atomic_sub(confirm_recv, &iucv->msg_recv); iucv 386 net/iucv/af_iucv.c WARN_ON(atomic_read(&iucv->msg_recv) < 0); iucv 453 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv 454 net/iucv/af_iucv.c struct iucv_path *path = iucv->path; iucv 456 net/iucv/af_iucv.c if (iucv->path) { iucv 457 net/iucv/af_iucv.c iucv->path = NULL; iucv 459 net/iucv/af_iucv.c low_nmcpy(user_data, iucv->src_name); iucv 460 net/iucv/af_iucv.c high_nmcpy(user_data, iucv->dst_name); iucv 472 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv 479 net/iucv/af_iucv.c LL_RESERVED_SPACE(iucv->hs_dev); iucv 498 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv 510 net/iucv/af_iucv.c if (iucv->transport == AF_IUCV_TRANS_HIPER) { iucv 521 net/iucv/af_iucv.c if (!err && !skb_queue_empty(&iucv->send_skb_q)) { iucv 539 net/iucv/af_iucv.c skb_queue_purge(&iucv->send_skb_q); iucv 540 net/iucv/af_iucv.c skb_queue_purge(&iucv->backlog_skb_q); iucv 547 net/iucv/af_iucv.c if (iucv->hs_dev) { iucv 548 net/iucv/af_iucv.c dev_put(iucv->hs_dev); iucv 549 net/iucv/af_iucv.c iucv->hs_dev = NULL; iucv 570 net/iucv/af_iucv.c struct iucv_sock *iucv; iucv 575 net/iucv/af_iucv.c iucv = iucv_sk(sk); iucv 578 net/iucv/af_iucv.c INIT_LIST_HEAD(&iucv->accept_q); iucv 579 net/iucv/af_iucv.c spin_lock_init(&iucv->accept_q_lock); iucv 580 net/iucv/af_iucv.c skb_queue_head_init(&iucv->send_skb_q); iucv 581 net/iucv/af_iucv.c INIT_LIST_HEAD(&iucv->message_q.list); iucv 582 net/iucv/af_iucv.c spin_lock_init(&iucv->message_q.lock); iucv 583 net/iucv/af_iucv.c skb_queue_head_init(&iucv->backlog_skb_q); iucv 584 net/iucv/af_iucv.c iucv->send_tag = 0; iucv 585 net/iucv/af_iucv.c atomic_set(&iucv->pendings, 0); iucv 586 net/iucv/af_iucv.c iucv->flags = 0; iucv 587 net/iucv/af_iucv.c iucv->msglimit = 0; iucv 588 net/iucv/af_iucv.c atomic_set(&iucv->msg_sent, 0); iucv 589 net/iucv/af_iucv.c atomic_set(&iucv->msg_recv, 0); iucv 590 net/iucv/af_iucv.c iucv->path = NULL; iucv 591 net/iucv/af_iucv.c iucv->sk_txnotify = afiucv_hs_callback_txnotify; iucv 592 net/iucv/af_iucv.c memset(&iucv->src_user_id , 0, 32); iucv 594 net/iucv/af_iucv.c iucv->transport = AF_IUCV_TRANS_IUCV; iucv 596 net/iucv/af_iucv.c iucv->transport = AF_IUCV_TRANS_HIPER; iucv 713 net/iucv/af_iucv.c static void __iucv_auto_name(struct iucv_sock *iucv) iucv 722 net/iucv/af_iucv.c memcpy(iucv->src_name, name, 8); iucv 731 net/iucv/af_iucv.c struct iucv_sock *iucv; iucv 749 net/iucv/af_iucv.c iucv = iucv_sk(sk); iucv 754 net/iucv/af_iucv.c if (iucv->path) iucv 768 net/iucv/af_iucv.c memcpy(iucv->src_user_id, sa->siucv_user_id, 8); iucv 771 net/iucv/af_iucv.c __iucv_auto_name(iucv); iucv 773 net/iucv/af_iucv.c memcpy(iucv->src_name, sa->siucv_name, 8); iucv 775 net/iucv/af_iucv.c iucv->hs_dev = dev; iucv 778 net/iucv/af_iucv.c iucv->transport = AF_IUCV_TRANS_HIPER; iucv 779 net/iucv/af_iucv.c if (!iucv->msglimit) iucv 780 net/iucv/af_iucv.c iucv->msglimit = IUCV_HIPER_MSGLIM_DEFAULT; iucv 789 net/iucv/af_iucv.c memcpy(iucv->src_name, sa->siucv_name, 8); iucv 790 net/iucv/af_iucv.c memcpy(iucv->src_user_id, iucv_userid, 8); iucv 792 net/iucv/af_iucv.c iucv->transport = AF_IUCV_TRANS_IUCV; iucv 794 net/iucv/af_iucv.c if (!iucv->msglimit) iucv 795 net/iucv/af_iucv.c iucv->msglimit = IUCV_QUEUELEN_DEFAULT; iucv 811 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv 817 net/iucv/af_iucv.c memcpy(iucv->src_user_id, iucv_userid, 8); iucv 818 net/iucv/af_iucv.c iucv->transport = AF_IUCV_TRANS_IUCV; iucv 822 net/iucv/af_iucv.c __iucv_auto_name(iucv); iucv 825 net/iucv/af_iucv.c if (!iucv->msglimit) iucv 826 net/iucv/af_iucv.c iucv->msglimit = IUCV_QUEUELEN_DEFAULT; iucv 835 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv 840 net/iucv/af_iucv.c low_nmcpy(user_data, iucv->src_name); iucv 844 net/iucv/af_iucv.c iucv->path = iucv_path_alloc(iucv->msglimit, iucv 846 net/iucv/af_iucv.c if (!iucv->path) { iucv 850 net/iucv/af_iucv.c err = pr_iucv->path_connect(iucv->path, &af_iucv_handler, iucv 854 net/iucv/af_iucv.c iucv_path_free(iucv->path); iucv 855 net/iucv/af_iucv.c iucv->path = NULL; iucv 882 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv 892 net/iucv/af_iucv.c iucv->transport == AF_IUCV_TRANS_HIPER) iucv 907 net/iucv/af_iucv.c memcpy(iucv->dst_user_id, sa->siucv_user_id, 8); iucv 908 net/iucv/af_iucv.c memcpy(iucv->dst_name, sa->siucv_name, 8); iucv 910 net/iucv/af_iucv.c if (iucv->transport == AF_IUCV_TRANS_HIPER) iucv 925 net/iucv/af_iucv.c if (err && iucv->transport == AF_IUCV_TRANS_IUCV) iucv 1018 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv 1023 net/iucv/af_iucv.c memcpy(siucv->siucv_user_id, iucv->dst_user_id, 8); iucv 1024 net/iucv/af_iucv.c memcpy(siucv->siucv_name, iucv->dst_name, 8); iucv 1026 net/iucv/af_iucv.c memcpy(siucv->siucv_user_id, iucv->src_user_id, 8); iucv 1027 net/iucv/af_iucv.c memcpy(siucv->siucv_name, iucv->src_name, 8); iucv 1064 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv 1144 net/iucv/af_iucv.c if (iucv->transport == AF_IUCV_TRANS_HIPER) { iucv 1146 net/iucv/af_iucv.c LL_RESERVED_SPACE(iucv->hs_dev); iucv 1186 net/iucv/af_iucv.c txmsg.tag = iucv->send_tag++; iucv 1189 net/iucv/af_iucv.c if (iucv->transport == AF_IUCV_TRANS_HIPER) { iucv 1190 net/iucv/af_iucv.c atomic_inc(&iucv->msg_sent); iucv 1193 net/iucv/af_iucv.c atomic_dec(&iucv->msg_sent); iucv 1197 net/iucv/af_iucv.c skb_queue_tail(&iucv->send_skb_q, skb); iucv 1199 net/iucv/af_iucv.c if (((iucv->path->flags & IUCV_IPRMDATA) & iucv->flags) && iucv 1201 net/iucv/af_iucv.c err = iucv_send_iprm(iucv->path, &txmsg, skb); iucv 1206 net/iucv/af_iucv.c skb_unlink(skb, &iucv->send_skb_q); iucv 1213 net/iucv/af_iucv.c pr_iucv->path_sever(iucv->path, NULL); iucv 1214 net/iucv/af_iucv.c skb_unlink(skb, &iucv->send_skb_q); iucv 1232 net/iucv/af_iucv.c err = pr_iucv->message_send(iucv->path, &txmsg, iucv 1236 net/iucv/af_iucv.c err = pr_iucv->message_send(iucv->path, &txmsg, iucv 1242 net/iucv/af_iucv.c memcpy(user_id, iucv->dst_user_id, 8); iucv 1244 net/iucv/af_iucv.c memcpy(appl_id, iucv->dst_name, 8); iucv 1252 net/iucv/af_iucv.c skb_unlink(skb, &iucv->send_skb_q); iucv 1363 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv 1367 net/iucv/af_iucv.c list_for_each_entry_safe(p, n, &iucv->message_q.list, list) { iucv 1374 net/iucv/af_iucv.c if (!skb_queue_empty(&iucv->backlog_skb_q)) iucv 1384 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv 1391 net/iucv/af_iucv.c skb_queue_empty(&iucv->backlog_skb_q) && iucv 1393 net/iucv/af_iucv.c list_empty(&iucv->message_q.list)) iucv 1454 net/iucv/af_iucv.c if (iucv->transport == AF_IUCV_TRANS_HIPER) { iucv 1455 net/iucv/af_iucv.c atomic_inc(&iucv->msg_recv); iucv 1456 net/iucv/af_iucv.c if (atomic_read(&iucv->msg_recv) > iucv->msglimit) { iucv 1464 net/iucv/af_iucv.c spin_lock_bh(&iucv->message_q.lock); iucv 1465 net/iucv/af_iucv.c rskb = skb_dequeue(&iucv->backlog_skb_q); iucv 1470 net/iucv/af_iucv.c skb_queue_head(&iucv->backlog_skb_q, iucv 1474 net/iucv/af_iucv.c rskb = skb_dequeue(&iucv->backlog_skb_q); iucv 1476 net/iucv/af_iucv.c if (skb_queue_empty(&iucv->backlog_skb_q)) { iucv 1477 net/iucv/af_iucv.c if (!list_empty(&iucv->message_q.list)) iucv 1479 net/iucv/af_iucv.c if (atomic_read(&iucv->msg_recv) >= iucv 1480 net/iucv/af_iucv.c iucv->msglimit / 2) { iucv 1488 net/iucv/af_iucv.c spin_unlock_bh(&iucv->message_q.lock); iucv 1556 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv 1578 net/iucv/af_iucv.c if (iucv->transport == AF_IUCV_TRANS_IUCV) { iucv 1581 net/iucv/af_iucv.c err = pr_iucv->message_send(iucv->path, &txmsg, iucv 1602 net/iucv/af_iucv.c if ((iucv->transport == AF_IUCV_TRANS_IUCV) && iucv 1603 net/iucv/af_iucv.c iucv->path) { iucv 1604 net/iucv/af_iucv.c err = pr_iucv->path_quiesce(iucv->path, NULL); iucv 1640 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv 1659 net/iucv/af_iucv.c iucv->flags |= IUCV_IPRMDATA; iucv 1661 net/iucv/af_iucv.c iucv->flags &= ~IUCV_IPRMDATA; iucv 1670 net/iucv/af_iucv.c iucv->msglimit = val; iucv 1690 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv 1707 net/iucv/af_iucv.c val = (iucv->flags & IUCV_IPRMDATA) ? 1 : 0; iucv 1711 net/iucv/af_iucv.c val = (iucv->path != NULL) ? iucv->path->msglim /* connected */ iucv 1712 net/iucv/af_iucv.c : iucv->msglimit; /* default */ iucv 1718 net/iucv/af_iucv.c val = (iucv->hs_dev) ? iucv->hs_dev->mtu - iucv 1743 net/iucv/af_iucv.c struct iucv_sock *iucv, *niucv; iucv 1750 net/iucv/af_iucv.c iucv = NULL; iucv 1759 net/iucv/af_iucv.c iucv = iucv_sk(sk); iucv 1763 net/iucv/af_iucv.c if (!iucv) iucv 1770 net/iucv/af_iucv.c low_nmcpy(user_data, iucv->src_name); iucv 1771 net/iucv/af_iucv.c high_nmcpy(user_data, iucv->dst_name); iucv 1803 net/iucv/af_iucv.c memcpy(niucv->src_name, iucv->src_name, 8); iucv 1804 net/iucv/af_iucv.c memcpy(niucv->src_user_id, iucv->src_user_id, 8); iucv 1813 net/iucv/af_iucv.c niucv->msglimit = iucv->msglimit; iucv 1814 net/iucv/af_iucv.c path->msglim = iucv->msglimit; iucv 1844 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv 1854 net/iucv/af_iucv.c spin_lock(&iucv->message_q.lock); iucv 1856 net/iucv/af_iucv.c if (!list_empty(&iucv->message_q.list) || iucv 1857 net/iucv/af_iucv.c !skb_queue_empty(&iucv->backlog_skb_q)) iucv 1879 net/iucv/af_iucv.c list_add_tail(&save_msg->list, &iucv->message_q.list); iucv 1882 net/iucv/af_iucv.c spin_unlock(&iucv->message_q.lock); iucv 1981 net/iucv/af_iucv.c struct iucv_sock *iucv, *niucv; iucv 1984 net/iucv/af_iucv.c iucv = iucv_sk(sk); iucv 1985 net/iucv/af_iucv.c if (!iucv) { iucv 2010 net/iucv/af_iucv.c niucv->msglimit = iucv->msglimit; iucv 2017 net/iucv/af_iucv.c memcpy(niucv->src_name, iucv->src_name, 8); iucv 2018 net/iucv/af_iucv.c memcpy(niucv->src_user_id, iucv->src_user_id, 8); iucv 2020 net/iucv/af_iucv.c niucv->hs_dev = iucv->hs_dev; iucv 2044 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv 2046 net/iucv/af_iucv.c if (!iucv) iucv 2051 net/iucv/af_iucv.c iucv->msglimit_peer = iucv_trans_hdr(skb)->window; iucv 2065 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv 2067 net/iucv/af_iucv.c if (!iucv) iucv 2085 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv 2088 net/iucv/af_iucv.c if (!iucv) iucv 2106 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv 2108 net/iucv/af_iucv.c if (!iucv) iucv 2114 net/iucv/af_iucv.c atomic_sub(iucv_trans_hdr(skb)->window, &iucv->msg_sent); iucv 2124 net/iucv/af_iucv.c struct iucv_sock *iucv = iucv_sk(sk); iucv 2126 net/iucv/af_iucv.c if (!iucv) { iucv 2152 net/iucv/af_iucv.c spin_lock(&iucv->message_q.lock); iucv 2153 net/iucv/af_iucv.c if (skb_queue_empty(&iucv->backlog_skb_q)) { iucv 2156 net/iucv/af_iucv.c skb_queue_tail(&iucv->backlog_skb_q, skb); iucv 2159 net/iucv/af_iucv.c spin_unlock(&iucv->message_q.lock); iucv 2172 net/iucv/af_iucv.c struct iucv_sock *iucv; iucv 2189 net/iucv/af_iucv.c iucv = NULL; iucv 2201 net/iucv/af_iucv.c iucv = iucv_sk(sk); iucv 2213 net/iucv/af_iucv.c iucv = iucv_sk(sk); iucv 2219 net/iucv/af_iucv.c if (!iucv) iucv 2279 net/iucv/af_iucv.c struct iucv_sock *iucv = NULL; iucv 2288 net/iucv/af_iucv.c iucv = iucv_sk(sk); iucv 2293 net/iucv/af_iucv.c if (!iucv || sock_flag(sk, SOCK_ZAPPED)) iucv 2296 net/iucv/af_iucv.c list = &iucv->send_skb_q; iucv 2307 net/iucv/af_iucv.c atomic_inc(&iucv->pendings); iucv 2311 net/iucv/af_iucv.c atomic_dec(&iucv->pendings); iucv 2312 net/iucv/af_iucv.c if (atomic_read(&iucv->pendings) <= 0) iucv 2351 net/iucv/af_iucv.c struct iucv_sock *iucv; iucv 2357 net/iucv/af_iucv.c iucv = iucv_sk(sk); iucv 2358 net/iucv/af_iucv.c if ((iucv->hs_dev == event_dev) &&