iucv             1314 drivers/s390/net/qeth_l3_main.c 	struct af_iucv_trans_hdr *iucv = (struct af_iucv_trans_hdr *) skb->data;
iucv             1317 drivers/s390/net/qeth_l3_main.c 	if (IS_IQD(card) && iucv->magic == ETH_P_AF_IUCV) {
iucv              153 net/iucv/af_iucv.c 	struct iucv_sock *iucv;
iucv              161 net/iucv/af_iucv.c 		iucv = iucv_sk(sk);
iucv              175 net/iucv/af_iucv.c 		skb_queue_purge(&iucv->send_skb_q);
iucv              176 net/iucv/af_iucv.c 		skb_queue_purge(&iucv->backlog_skb_q);
iucv              288 net/iucv/af_iucv.c 	struct iucv_sock *iucv = iucv_sk(sk);
iucv              292 net/iucv/af_iucv.c 	if (iucv->transport == AF_IUCV_TRANS_IUCV)
iucv              293 net/iucv/af_iucv.c 		return (skb_queue_len(&iucv->send_skb_q) < iucv->path->msglim);
iucv              295 net/iucv/af_iucv.c 		return ((atomic_read(&iucv->msg_sent) < iucv->msglimit_peer) &&
iucv              296 net/iucv/af_iucv.c 			(atomic_read(&iucv->pendings) <= 0));
iucv              320 net/iucv/af_iucv.c 	struct iucv_sock *iucv = iucv_sk(sock);
iucv              333 net/iucv/af_iucv.c 		phs_hdr->window = iucv->msglimit;
iucv              335 net/iucv/af_iucv.c 		confirm_recv = atomic_read(&iucv->msg_recv);
iucv              340 net/iucv/af_iucv.c 	memcpy(phs_hdr->destUserID, iucv->dst_user_id, 8);
iucv              341 net/iucv/af_iucv.c 	memcpy(phs_hdr->destAppName, iucv->dst_name, 8);
iucv              342 net/iucv/af_iucv.c 	memcpy(phs_hdr->srcUserID, iucv->src_user_id, 8);
iucv              343 net/iucv/af_iucv.c 	memcpy(phs_hdr->srcAppName, iucv->src_name, 8);
iucv              351 net/iucv/af_iucv.c 	skb->dev = iucv->hs_dev;
iucv              379 net/iucv/af_iucv.c 	skb_queue_tail(&iucv->send_skb_q, nskb);
iucv              382 net/iucv/af_iucv.c 		skb_unlink(nskb, &iucv->send_skb_q);
iucv              385 net/iucv/af_iucv.c 		atomic_sub(confirm_recv, &iucv->msg_recv);
iucv              386 net/iucv/af_iucv.c 		WARN_ON(atomic_read(&iucv->msg_recv) < 0);
iucv              453 net/iucv/af_iucv.c 	struct iucv_sock *iucv = iucv_sk(sk);
iucv              454 net/iucv/af_iucv.c 	struct iucv_path *path = iucv->path;
iucv              456 net/iucv/af_iucv.c 	if (iucv->path) {
iucv              457 net/iucv/af_iucv.c 		iucv->path = NULL;
iucv              459 net/iucv/af_iucv.c 			low_nmcpy(user_data, iucv->src_name);
iucv              460 net/iucv/af_iucv.c 			high_nmcpy(user_data, iucv->dst_name);
iucv              472 net/iucv/af_iucv.c 	struct iucv_sock *iucv = iucv_sk(sk);
iucv              479 net/iucv/af_iucv.c 	       LL_RESERVED_SPACE(iucv->hs_dev);
iucv              498 net/iucv/af_iucv.c 	struct iucv_sock *iucv = iucv_sk(sk);
iucv              510 net/iucv/af_iucv.c 		if (iucv->transport == AF_IUCV_TRANS_HIPER) {
iucv              521 net/iucv/af_iucv.c 		if (!err && !skb_queue_empty(&iucv->send_skb_q)) {
iucv              539 net/iucv/af_iucv.c 		skb_queue_purge(&iucv->send_skb_q);
iucv              540 net/iucv/af_iucv.c 		skb_queue_purge(&iucv->backlog_skb_q);
iucv              547 net/iucv/af_iucv.c 	if (iucv->hs_dev) {
iucv              548 net/iucv/af_iucv.c 		dev_put(iucv->hs_dev);
iucv              549 net/iucv/af_iucv.c 		iucv->hs_dev = NULL;
iucv              570 net/iucv/af_iucv.c 	struct iucv_sock *iucv;
iucv              575 net/iucv/af_iucv.c 	iucv = iucv_sk(sk);
iucv              578 net/iucv/af_iucv.c 	INIT_LIST_HEAD(&iucv->accept_q);
iucv              579 net/iucv/af_iucv.c 	spin_lock_init(&iucv->accept_q_lock);
iucv              580 net/iucv/af_iucv.c 	skb_queue_head_init(&iucv->send_skb_q);
iucv              581 net/iucv/af_iucv.c 	INIT_LIST_HEAD(&iucv->message_q.list);
iucv              582 net/iucv/af_iucv.c 	spin_lock_init(&iucv->message_q.lock);
iucv              583 net/iucv/af_iucv.c 	skb_queue_head_init(&iucv->backlog_skb_q);
iucv              584 net/iucv/af_iucv.c 	iucv->send_tag = 0;
iucv              585 net/iucv/af_iucv.c 	atomic_set(&iucv->pendings, 0);
iucv              586 net/iucv/af_iucv.c 	iucv->flags = 0;
iucv              587 net/iucv/af_iucv.c 	iucv->msglimit = 0;
iucv              588 net/iucv/af_iucv.c 	atomic_set(&iucv->msg_sent, 0);
iucv              589 net/iucv/af_iucv.c 	atomic_set(&iucv->msg_recv, 0);
iucv              590 net/iucv/af_iucv.c 	iucv->path = NULL;
iucv              591 net/iucv/af_iucv.c 	iucv->sk_txnotify = afiucv_hs_callback_txnotify;
iucv              592 net/iucv/af_iucv.c 	memset(&iucv->src_user_id , 0, 32);
iucv              594 net/iucv/af_iucv.c 		iucv->transport = AF_IUCV_TRANS_IUCV;
iucv              596 net/iucv/af_iucv.c 		iucv->transport = AF_IUCV_TRANS_HIPER;
iucv              713 net/iucv/af_iucv.c static void __iucv_auto_name(struct iucv_sock *iucv)
iucv              722 net/iucv/af_iucv.c 	memcpy(iucv->src_name, name, 8);
iucv              731 net/iucv/af_iucv.c 	struct iucv_sock *iucv;
iucv              749 net/iucv/af_iucv.c 	iucv = iucv_sk(sk);
iucv              754 net/iucv/af_iucv.c 	if (iucv->path)
iucv              768 net/iucv/af_iucv.c 			memcpy(iucv->src_user_id, sa->siucv_user_id, 8);
iucv              771 net/iucv/af_iucv.c 				__iucv_auto_name(iucv);
iucv              773 net/iucv/af_iucv.c 				memcpy(iucv->src_name, sa->siucv_name, 8);
iucv              775 net/iucv/af_iucv.c 			iucv->hs_dev = dev;
iucv              778 net/iucv/af_iucv.c 			iucv->transport = AF_IUCV_TRANS_HIPER;
iucv              779 net/iucv/af_iucv.c 			if (!iucv->msglimit)
iucv              780 net/iucv/af_iucv.c 				iucv->msglimit = IUCV_HIPER_MSGLIM_DEFAULT;
iucv              789 net/iucv/af_iucv.c 		memcpy(iucv->src_name, sa->siucv_name, 8);
iucv              790 net/iucv/af_iucv.c 		memcpy(iucv->src_user_id, iucv_userid, 8);
iucv              792 net/iucv/af_iucv.c 		iucv->transport = AF_IUCV_TRANS_IUCV;
iucv              794 net/iucv/af_iucv.c 		if (!iucv->msglimit)
iucv              795 net/iucv/af_iucv.c 			iucv->msglimit = IUCV_QUEUELEN_DEFAULT;
iucv              811 net/iucv/af_iucv.c 	struct iucv_sock *iucv = iucv_sk(sk);
iucv              817 net/iucv/af_iucv.c 	memcpy(iucv->src_user_id, iucv_userid, 8);
iucv              818 net/iucv/af_iucv.c 	iucv->transport = AF_IUCV_TRANS_IUCV;
iucv              822 net/iucv/af_iucv.c 	__iucv_auto_name(iucv);
iucv              825 net/iucv/af_iucv.c 	if (!iucv->msglimit)
iucv              826 net/iucv/af_iucv.c 		iucv->msglimit = IUCV_QUEUELEN_DEFAULT;
iucv              835 net/iucv/af_iucv.c 	struct iucv_sock *iucv = iucv_sk(sk);
iucv              840 net/iucv/af_iucv.c 	low_nmcpy(user_data, iucv->src_name);
iucv              844 net/iucv/af_iucv.c 	iucv->path = iucv_path_alloc(iucv->msglimit,
iucv              846 net/iucv/af_iucv.c 	if (!iucv->path) {
iucv              850 net/iucv/af_iucv.c 	err = pr_iucv->path_connect(iucv->path, &af_iucv_handler,
iucv              854 net/iucv/af_iucv.c 		iucv_path_free(iucv->path);
iucv              855 net/iucv/af_iucv.c 		iucv->path = NULL;
iucv              882 net/iucv/af_iucv.c 	struct iucv_sock *iucv = iucv_sk(sk);
iucv              892 net/iucv/af_iucv.c 	    iucv->transport == AF_IUCV_TRANS_HIPER)
iucv              907 net/iucv/af_iucv.c 	memcpy(iucv->dst_user_id, sa->siucv_user_id, 8);
iucv              908 net/iucv/af_iucv.c 	memcpy(iucv->dst_name, sa->siucv_name, 8);
iucv              910 net/iucv/af_iucv.c 	if (iucv->transport == AF_IUCV_TRANS_HIPER)
iucv              925 net/iucv/af_iucv.c 	if (err && iucv->transport == AF_IUCV_TRANS_IUCV)
iucv             1018 net/iucv/af_iucv.c 	struct iucv_sock *iucv = iucv_sk(sk);
iucv             1023 net/iucv/af_iucv.c 		memcpy(siucv->siucv_user_id, iucv->dst_user_id, 8);
iucv             1024 net/iucv/af_iucv.c 		memcpy(siucv->siucv_name, iucv->dst_name, 8);
iucv             1026 net/iucv/af_iucv.c 		memcpy(siucv->siucv_user_id, iucv->src_user_id, 8);
iucv             1027 net/iucv/af_iucv.c 		memcpy(siucv->siucv_name, iucv->src_name, 8);
iucv             1064 net/iucv/af_iucv.c 	struct iucv_sock *iucv = iucv_sk(sk);
iucv             1144 net/iucv/af_iucv.c 	if (iucv->transport == AF_IUCV_TRANS_HIPER) {
iucv             1146 net/iucv/af_iucv.c 			   LL_RESERVED_SPACE(iucv->hs_dev);
iucv             1186 net/iucv/af_iucv.c 	txmsg.tag = iucv->send_tag++;
iucv             1189 net/iucv/af_iucv.c 	if (iucv->transport == AF_IUCV_TRANS_HIPER) {
iucv             1190 net/iucv/af_iucv.c 		atomic_inc(&iucv->msg_sent);
iucv             1193 net/iucv/af_iucv.c 			atomic_dec(&iucv->msg_sent);
iucv             1197 net/iucv/af_iucv.c 		skb_queue_tail(&iucv->send_skb_q, skb);
iucv             1199 net/iucv/af_iucv.c 		if (((iucv->path->flags & IUCV_IPRMDATA) & iucv->flags) &&
iucv             1201 net/iucv/af_iucv.c 			err = iucv_send_iprm(iucv->path, &txmsg, skb);
iucv             1206 net/iucv/af_iucv.c 				skb_unlink(skb, &iucv->send_skb_q);
iucv             1213 net/iucv/af_iucv.c 				pr_iucv->path_sever(iucv->path, NULL);
iucv             1214 net/iucv/af_iucv.c 				skb_unlink(skb, &iucv->send_skb_q);
iucv             1232 net/iucv/af_iucv.c 			err = pr_iucv->message_send(iucv->path, &txmsg,
iucv             1236 net/iucv/af_iucv.c 			err = pr_iucv->message_send(iucv->path, &txmsg,
iucv             1242 net/iucv/af_iucv.c 				memcpy(user_id, iucv->dst_user_id, 8);
iucv             1244 net/iucv/af_iucv.c 				memcpy(appl_id, iucv->dst_name, 8);
iucv             1252 net/iucv/af_iucv.c 			skb_unlink(skb, &iucv->send_skb_q);
iucv             1363 net/iucv/af_iucv.c 	struct iucv_sock *iucv = iucv_sk(sk);
iucv             1367 net/iucv/af_iucv.c 	list_for_each_entry_safe(p, n, &iucv->message_q.list, list) {
iucv             1374 net/iucv/af_iucv.c 		if (!skb_queue_empty(&iucv->backlog_skb_q))
iucv             1384 net/iucv/af_iucv.c 	struct iucv_sock *iucv = iucv_sk(sk);
iucv             1391 net/iucv/af_iucv.c 	    skb_queue_empty(&iucv->backlog_skb_q) &&
iucv             1393 net/iucv/af_iucv.c 	    list_empty(&iucv->message_q.list))
iucv             1454 net/iucv/af_iucv.c 		if (iucv->transport == AF_IUCV_TRANS_HIPER) {
iucv             1455 net/iucv/af_iucv.c 			atomic_inc(&iucv->msg_recv);
iucv             1456 net/iucv/af_iucv.c 			if (atomic_read(&iucv->msg_recv) > iucv->msglimit) {
iucv             1464 net/iucv/af_iucv.c 		spin_lock_bh(&iucv->message_q.lock);
iucv             1465 net/iucv/af_iucv.c 		rskb = skb_dequeue(&iucv->backlog_skb_q);
iucv             1470 net/iucv/af_iucv.c 				skb_queue_head(&iucv->backlog_skb_q,
iucv             1474 net/iucv/af_iucv.c 			rskb = skb_dequeue(&iucv->backlog_skb_q);
iucv             1476 net/iucv/af_iucv.c 		if (skb_queue_empty(&iucv->backlog_skb_q)) {
iucv             1477 net/iucv/af_iucv.c 			if (!list_empty(&iucv->message_q.list))
iucv             1479 net/iucv/af_iucv.c 			if (atomic_read(&iucv->msg_recv) >=
iucv             1480 net/iucv/af_iucv.c 							iucv->msglimit / 2) {
iucv             1488 net/iucv/af_iucv.c 		spin_unlock_bh(&iucv->message_q.lock);
iucv             1556 net/iucv/af_iucv.c 	struct iucv_sock *iucv = iucv_sk(sk);
iucv             1578 net/iucv/af_iucv.c 		if (iucv->transport == AF_IUCV_TRANS_IUCV) {
iucv             1581 net/iucv/af_iucv.c 			err = pr_iucv->message_send(iucv->path, &txmsg,
iucv             1602 net/iucv/af_iucv.c 		if ((iucv->transport == AF_IUCV_TRANS_IUCV) &&
iucv             1603 net/iucv/af_iucv.c 		    iucv->path) {
iucv             1604 net/iucv/af_iucv.c 			err = pr_iucv->path_quiesce(iucv->path, NULL);
iucv             1640 net/iucv/af_iucv.c 	struct iucv_sock *iucv = iucv_sk(sk);
iucv             1659 net/iucv/af_iucv.c 			iucv->flags |= IUCV_IPRMDATA;
iucv             1661 net/iucv/af_iucv.c 			iucv->flags &= ~IUCV_IPRMDATA;
iucv             1670 net/iucv/af_iucv.c 				iucv->msglimit = val;
iucv             1690 net/iucv/af_iucv.c 	struct iucv_sock *iucv = iucv_sk(sk);
iucv             1707 net/iucv/af_iucv.c 		val = (iucv->flags & IUCV_IPRMDATA) ? 1 : 0;
iucv             1711 net/iucv/af_iucv.c 		val = (iucv->path != NULL) ? iucv->path->msglim	/* connected */
iucv             1712 net/iucv/af_iucv.c 					   : iucv->msglimit;	/* default */
iucv             1718 net/iucv/af_iucv.c 		val = (iucv->hs_dev) ? iucv->hs_dev->mtu -
iucv             1743 net/iucv/af_iucv.c 	struct iucv_sock *iucv, *niucv;
iucv             1750 net/iucv/af_iucv.c 	iucv = NULL;
iucv             1759 net/iucv/af_iucv.c 			iucv = iucv_sk(sk);
iucv             1763 net/iucv/af_iucv.c 	if (!iucv)
iucv             1770 net/iucv/af_iucv.c 	low_nmcpy(user_data, iucv->src_name);
iucv             1771 net/iucv/af_iucv.c 	high_nmcpy(user_data, iucv->dst_name);
iucv             1803 net/iucv/af_iucv.c 	memcpy(niucv->src_name, iucv->src_name, 8);
iucv             1804 net/iucv/af_iucv.c 	memcpy(niucv->src_user_id, iucv->src_user_id, 8);
iucv             1813 net/iucv/af_iucv.c 	niucv->msglimit = iucv->msglimit;
iucv             1814 net/iucv/af_iucv.c 	path->msglim = iucv->msglimit;
iucv             1844 net/iucv/af_iucv.c 	struct iucv_sock *iucv = iucv_sk(sk);
iucv             1854 net/iucv/af_iucv.c 	spin_lock(&iucv->message_q.lock);
iucv             1856 net/iucv/af_iucv.c 	if (!list_empty(&iucv->message_q.list) ||
iucv             1857 net/iucv/af_iucv.c 	    !skb_queue_empty(&iucv->backlog_skb_q))
iucv             1879 net/iucv/af_iucv.c 	list_add_tail(&save_msg->list, &iucv->message_q.list);
iucv             1882 net/iucv/af_iucv.c 	spin_unlock(&iucv->message_q.lock);
iucv             1981 net/iucv/af_iucv.c 	struct iucv_sock *iucv, *niucv;
iucv             1984 net/iucv/af_iucv.c 	iucv = iucv_sk(sk);
iucv             1985 net/iucv/af_iucv.c 	if (!iucv) {
iucv             2010 net/iucv/af_iucv.c 	niucv->msglimit = iucv->msglimit;
iucv             2017 net/iucv/af_iucv.c 	memcpy(niucv->src_name, iucv->src_name, 8);
iucv             2018 net/iucv/af_iucv.c 	memcpy(niucv->src_user_id, iucv->src_user_id, 8);
iucv             2020 net/iucv/af_iucv.c 	niucv->hs_dev = iucv->hs_dev;
iucv             2044 net/iucv/af_iucv.c 	struct iucv_sock *iucv = iucv_sk(sk);
iucv             2046 net/iucv/af_iucv.c 	if (!iucv)
iucv             2051 net/iucv/af_iucv.c 	iucv->msglimit_peer = iucv_trans_hdr(skb)->window;
iucv             2065 net/iucv/af_iucv.c 	struct iucv_sock *iucv = iucv_sk(sk);
iucv             2067 net/iucv/af_iucv.c 	if (!iucv)
iucv             2085 net/iucv/af_iucv.c 	struct iucv_sock *iucv = iucv_sk(sk);
iucv             2088 net/iucv/af_iucv.c 	if (!iucv)
iucv             2106 net/iucv/af_iucv.c 	struct iucv_sock *iucv = iucv_sk(sk);
iucv             2108 net/iucv/af_iucv.c 	if (!iucv)
iucv             2114 net/iucv/af_iucv.c 	atomic_sub(iucv_trans_hdr(skb)->window, &iucv->msg_sent);
iucv             2124 net/iucv/af_iucv.c 	struct iucv_sock *iucv = iucv_sk(sk);
iucv             2126 net/iucv/af_iucv.c 	if (!iucv) {
iucv             2152 net/iucv/af_iucv.c 	spin_lock(&iucv->message_q.lock);
iucv             2153 net/iucv/af_iucv.c 	if (skb_queue_empty(&iucv->backlog_skb_q)) {
iucv             2156 net/iucv/af_iucv.c 			skb_queue_tail(&iucv->backlog_skb_q, skb);
iucv             2159 net/iucv/af_iucv.c 	spin_unlock(&iucv->message_q.lock);
iucv             2172 net/iucv/af_iucv.c 	struct iucv_sock *iucv;
iucv             2189 net/iucv/af_iucv.c 	iucv = NULL;
iucv             2201 net/iucv/af_iucv.c 				iucv = iucv_sk(sk);
iucv             2213 net/iucv/af_iucv.c 				iucv = iucv_sk(sk);
iucv             2219 net/iucv/af_iucv.c 	if (!iucv)
iucv             2279 net/iucv/af_iucv.c 	struct iucv_sock *iucv = NULL;
iucv             2288 net/iucv/af_iucv.c 			iucv = iucv_sk(sk);
iucv             2293 net/iucv/af_iucv.c 	if (!iucv || sock_flag(sk, SOCK_ZAPPED))
iucv             2296 net/iucv/af_iucv.c 	list = &iucv->send_skb_q;
iucv             2307 net/iucv/af_iucv.c 				atomic_inc(&iucv->pendings);
iucv             2311 net/iucv/af_iucv.c 				atomic_dec(&iucv->pendings);
iucv             2312 net/iucv/af_iucv.c 				if (atomic_read(&iucv->pendings) <= 0)
iucv             2351 net/iucv/af_iucv.c 	struct iucv_sock *iucv;
iucv             2357 net/iucv/af_iucv.c 			iucv = iucv_sk(sk);
iucv             2358 net/iucv/af_iucv.c 			if ((iucv->hs_dev == event_dev) &&