vsk               269 drivers/vhost/vsock.c vhost_transport_cancel_pkt(struct vsock_sock *vsk)
vsk               280 drivers/vhost/vsock.c 	vsock = vhost_vsock_get(vsk->remote_addr.svm_cid);
vsk               286 drivers/vhost/vsock.c 		if (pkt->vsk != vsk)
vsk               611 drivers/vhost/vsock.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk               619 drivers/vhost/vsock.c 	if (vhost_vsock_get(vsk->remote_addr.svm_cid))
vsk               625 drivers/vhost/vsock.c 	if (vsk->close_work_scheduled)
vsk               629 drivers/vhost/vsock.c 	vsk->peer_shutdown = SHUTDOWN_MASK;
vsk                26 include/linux/virtio_vsock.h 	struct vsock_sock *vsk;
vsk                53 include/linux/virtio_vsock.h 	struct vsock_sock *vsk;
vsk                63 include/linux/virtio_vsock.h 	struct vsock_sock *vsk;
vsk                81 include/linux/virtio_vsock.h virtio_transport_stream_dequeue(struct vsock_sock *vsk,
vsk                86 include/linux/virtio_vsock.h virtio_transport_dgram_dequeue(struct vsock_sock *vsk,
vsk                90 include/linux/virtio_vsock.h s64 virtio_transport_stream_has_data(struct vsock_sock *vsk);
vsk                91 include/linux/virtio_vsock.h s64 virtio_transport_stream_has_space(struct vsock_sock *vsk);
vsk                93 include/linux/virtio_vsock.h int virtio_transport_do_socket_init(struct vsock_sock *vsk,
vsk                95 include/linux/virtio_vsock.h u64 virtio_transport_get_buffer_size(struct vsock_sock *vsk);
vsk                96 include/linux/virtio_vsock.h u64 virtio_transport_get_min_buffer_size(struct vsock_sock *vsk);
vsk                97 include/linux/virtio_vsock.h u64 virtio_transport_get_max_buffer_size(struct vsock_sock *vsk);
vsk                98 include/linux/virtio_vsock.h void virtio_transport_set_buffer_size(struct vsock_sock *vsk, u64 val);
vsk                99 include/linux/virtio_vsock.h void virtio_transport_set_min_buffer_size(struct vsock_sock *vsk, u64 val);
vsk               102 include/linux/virtio_vsock.h virtio_transport_notify_poll_in(struct vsock_sock *vsk,
vsk               106 include/linux/virtio_vsock.h virtio_transport_notify_poll_out(struct vsock_sock *vsk,
vsk               110 include/linux/virtio_vsock.h int virtio_transport_notify_recv_init(struct vsock_sock *vsk,
vsk               112 include/linux/virtio_vsock.h int virtio_transport_notify_recv_pre_block(struct vsock_sock *vsk,
vsk               114 include/linux/virtio_vsock.h int virtio_transport_notify_recv_pre_dequeue(struct vsock_sock *vsk,
vsk               116 include/linux/virtio_vsock.h int virtio_transport_notify_recv_post_dequeue(struct vsock_sock *vsk,
vsk               119 include/linux/virtio_vsock.h int virtio_transport_notify_send_init(struct vsock_sock *vsk,
vsk               121 include/linux/virtio_vsock.h int virtio_transport_notify_send_pre_block(struct vsock_sock *vsk,
vsk               123 include/linux/virtio_vsock.h int virtio_transport_notify_send_pre_enqueue(struct vsock_sock *vsk,
vsk               125 include/linux/virtio_vsock.h int virtio_transport_notify_send_post_enqueue(struct vsock_sock *vsk,
vsk               128 include/linux/virtio_vsock.h u64 virtio_transport_stream_rcvhiwat(struct vsock_sock *vsk);
vsk               129 include/linux/virtio_vsock.h bool virtio_transport_stream_is_active(struct vsock_sock *vsk);
vsk               131 include/linux/virtio_vsock.h int virtio_transport_dgram_bind(struct vsock_sock *vsk,
vsk               135 include/linux/virtio_vsock.h int virtio_transport_connect(struct vsock_sock *vsk);
vsk               137 include/linux/virtio_vsock.h int virtio_transport_shutdown(struct vsock_sock *vsk, int mode);
vsk               139 include/linux/virtio_vsock.h void virtio_transport_release(struct vsock_sock *vsk);
vsk               142 include/linux/virtio_vsock.h virtio_transport_stream_enqueue(struct vsock_sock *vsk,
vsk               146 include/linux/virtio_vsock.h virtio_transport_dgram_enqueue(struct vsock_sock *vsk,
vsk               151 include/linux/virtio_vsock.h void virtio_transport_destruct(struct vsock_sock *vsk);
vsk                71 include/net/af_vsock.h s64 vsock_stream_has_data(struct vsock_sock *vsk);
vsk                72 include/net/af_vsock.h s64 vsock_stream_has_space(struct vsock_sock *vsk);
vsk                98 include/net/af_vsock.h 	int (*cancel_pkt)(struct vsock_sock *vsk);
vsk               105 include/net/af_vsock.h 	int (*dgram_dequeue)(struct vsock_sock *vsk, struct msghdr *msg,
vsk               173 include/net/af_vsock.h static inline bool __vsock_in_bound_table(struct vsock_sock *vsk)
vsk               175 include/net/af_vsock.h 	return !list_empty(&vsk->bound_table);
vsk               179 include/net/af_vsock.h static inline bool __vsock_in_connected_table(struct vsock_sock *vsk)
vsk               181 include/net/af_vsock.h 	return !list_empty(&vsk->connected_table);
vsk               188 include/net/af_vsock.h void vsock_insert_connected(struct vsock_sock *vsk);
vsk               189 include/net/af_vsock.h void vsock_remove_bound(struct vsock_sock *vsk);
vsk               190 include/net/af_vsock.h void vsock_remove_connected(struct vsock_sock *vsk);
vsk               194 include/net/af_vsock.h void vsock_remove_sock(struct vsock_sock *vsk);
vsk               169 net/vmw_vsock/af_vsock.c #define vsock_connected_sockets_vsk(vsk)				\
vsk               170 net/vmw_vsock/af_vsock.c 	vsock_connected_sockets(&(vsk)->remote_addr, &(vsk)->local_addr)
vsk               180 net/vmw_vsock/af_vsock.c static int vsock_auto_bind(struct vsock_sock *vsk)
vsk               182 net/vmw_vsock/af_vsock.c 	struct sock *sk = sk_vsock(vsk);
vsk               185 net/vmw_vsock/af_vsock.c 	if (vsock_addr_bound(&vsk->local_addr))
vsk               204 net/vmw_vsock/af_vsock.c 				 struct vsock_sock *vsk)
vsk               206 net/vmw_vsock/af_vsock.c 	sock_hold(&vsk->sk);
vsk               207 net/vmw_vsock/af_vsock.c 	list_add(&vsk->bound_table, list);
vsk               211 net/vmw_vsock/af_vsock.c 				     struct vsock_sock *vsk)
vsk               213 net/vmw_vsock/af_vsock.c 	sock_hold(&vsk->sk);
vsk               214 net/vmw_vsock/af_vsock.c 	list_add(&vsk->connected_table, list);
vsk               217 net/vmw_vsock/af_vsock.c static void __vsock_remove_bound(struct vsock_sock *vsk)
vsk               219 net/vmw_vsock/af_vsock.c 	list_del_init(&vsk->bound_table);
vsk               220 net/vmw_vsock/af_vsock.c 	sock_put(&vsk->sk);
vsk               223 net/vmw_vsock/af_vsock.c static void __vsock_remove_connected(struct vsock_sock *vsk)
vsk               225 net/vmw_vsock/af_vsock.c 	list_del_init(&vsk->connected_table);
vsk               226 net/vmw_vsock/af_vsock.c 	sock_put(&vsk->sk);
vsk               231 net/vmw_vsock/af_vsock.c 	struct vsock_sock *vsk;
vsk               233 net/vmw_vsock/af_vsock.c 	list_for_each_entry(vsk, vsock_bound_sockets(addr), bound_table)
vsk               234 net/vmw_vsock/af_vsock.c 		if (addr->svm_port == vsk->local_addr.svm_port)
vsk               235 net/vmw_vsock/af_vsock.c 			return sk_vsock(vsk);
vsk               243 net/vmw_vsock/af_vsock.c 	struct vsock_sock *vsk;
vsk               245 net/vmw_vsock/af_vsock.c 	list_for_each_entry(vsk, vsock_connected_sockets(src, dst),
vsk               247 net/vmw_vsock/af_vsock.c 		if (vsock_addr_equals_addr(src, &vsk->remote_addr) &&
vsk               248 net/vmw_vsock/af_vsock.c 		    dst->svm_port == vsk->local_addr.svm_port) {
vsk               249 net/vmw_vsock/af_vsock.c 			return sk_vsock(vsk);
vsk               256 net/vmw_vsock/af_vsock.c static void vsock_insert_unbound(struct vsock_sock *vsk)
vsk               259 net/vmw_vsock/af_vsock.c 	__vsock_insert_bound(vsock_unbound_sockets, vsk);
vsk               263 net/vmw_vsock/af_vsock.c void vsock_insert_connected(struct vsock_sock *vsk)
vsk               266 net/vmw_vsock/af_vsock.c 		&vsk->remote_addr, &vsk->local_addr);
vsk               269 net/vmw_vsock/af_vsock.c 	__vsock_insert_connected(list, vsk);
vsk               274 net/vmw_vsock/af_vsock.c void vsock_remove_bound(struct vsock_sock *vsk)
vsk               277 net/vmw_vsock/af_vsock.c 	if (__vsock_in_bound_table(vsk))
vsk               278 net/vmw_vsock/af_vsock.c 		__vsock_remove_bound(vsk);
vsk               283 net/vmw_vsock/af_vsock.c void vsock_remove_connected(struct vsock_sock *vsk)
vsk               286 net/vmw_vsock/af_vsock.c 	if (__vsock_in_connected_table(vsk))
vsk               287 net/vmw_vsock/af_vsock.c 		__vsock_remove_connected(vsk);
vsk               323 net/vmw_vsock/af_vsock.c void vsock_remove_sock(struct vsock_sock *vsk)
vsk               325 net/vmw_vsock/af_vsock.c 	vsock_remove_bound(vsk);
vsk               326 net/vmw_vsock/af_vsock.c 	vsock_remove_connected(vsk);
vsk               337 net/vmw_vsock/af_vsock.c 		struct vsock_sock *vsk;
vsk               338 net/vmw_vsock/af_vsock.c 		list_for_each_entry(vsk, &vsock_connected_table[i],
vsk               340 net/vmw_vsock/af_vsock.c 			fn(sk_vsock(vsk));
vsk               409 net/vmw_vsock/af_vsock.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk               410 net/vmw_vsock/af_vsock.c 	return list_empty(&vsk->accept_queue);
vsk               415 net/vmw_vsock/af_vsock.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk               416 net/vmw_vsock/af_vsock.c 	return !list_empty(&vsk->pending_links);
vsk               428 net/vmw_vsock/af_vsock.c 	struct vsock_sock *vsk;
vsk               431 net/vmw_vsock/af_vsock.c 	vsk = container_of(work, struct vsock_sock, pending_work.work);
vsk               432 net/vmw_vsock/af_vsock.c 	sk = sk_vsock(vsk);
vsk               433 net/vmw_vsock/af_vsock.c 	listener = vsk->listener;
vsk               443 net/vmw_vsock/af_vsock.c 	} else if (!vsk->rejected) {
vsk               457 net/vmw_vsock/af_vsock.c 	vsock_remove_connected(vsk);
vsk               473 net/vmw_vsock/af_vsock.c static int __vsock_bind_stream(struct vsock_sock *vsk,
vsk               516 net/vmw_vsock/af_vsock.c 	vsock_addr_init(&vsk->local_addr, new_addr.svm_cid, new_addr.svm_port);
vsk               522 net/vmw_vsock/af_vsock.c 	__vsock_remove_bound(vsk);
vsk               523 net/vmw_vsock/af_vsock.c 	__vsock_insert_bound(vsock_bound_sockets(&vsk->local_addr), vsk);
vsk               528 net/vmw_vsock/af_vsock.c static int __vsock_bind_dgram(struct vsock_sock *vsk,
vsk               531 net/vmw_vsock/af_vsock.c 	return transport->dgram_bind(vsk, addr);
vsk               536 net/vmw_vsock/af_vsock.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk               541 net/vmw_vsock/af_vsock.c 	if (vsock_addr_bound(&vsk->local_addr))
vsk               556 net/vmw_vsock/af_vsock.c 		retval = __vsock_bind_stream(vsk, addr);
vsk               561 net/vmw_vsock/af_vsock.c 		retval = __vsock_bind_dgram(vsk, addr);
vsk               583 net/vmw_vsock/af_vsock.c 	struct vsock_sock *vsk;
vsk               598 net/vmw_vsock/af_vsock.c 	vsk = vsock_sk(sk);
vsk               599 net/vmw_vsock/af_vsock.c 	vsock_addr_init(&vsk->local_addr, VMADDR_CID_ANY, VMADDR_PORT_ANY);
vsk               600 net/vmw_vsock/af_vsock.c 	vsock_addr_init(&vsk->remote_addr, VMADDR_CID_ANY, VMADDR_PORT_ANY);
vsk               606 net/vmw_vsock/af_vsock.c 	INIT_LIST_HEAD(&vsk->bound_table);
vsk               607 net/vmw_vsock/af_vsock.c 	INIT_LIST_HEAD(&vsk->connected_table);
vsk               608 net/vmw_vsock/af_vsock.c 	vsk->listener = NULL;
vsk               609 net/vmw_vsock/af_vsock.c 	INIT_LIST_HEAD(&vsk->pending_links);
vsk               610 net/vmw_vsock/af_vsock.c 	INIT_LIST_HEAD(&vsk->accept_queue);
vsk               611 net/vmw_vsock/af_vsock.c 	vsk->rejected = false;
vsk               612 net/vmw_vsock/af_vsock.c 	vsk->sent_request = false;
vsk               613 net/vmw_vsock/af_vsock.c 	vsk->ignore_connecting_rst = false;
vsk               614 net/vmw_vsock/af_vsock.c 	vsk->peer_shutdown = 0;
vsk               615 net/vmw_vsock/af_vsock.c 	INIT_DELAYED_WORK(&vsk->connect_work, vsock_connect_timeout);
vsk               616 net/vmw_vsock/af_vsock.c 	INIT_DELAYED_WORK(&vsk->pending_work, vsock_pending_work);
vsk               620 net/vmw_vsock/af_vsock.c 		vsk->trusted = psk->trusted;
vsk               621 net/vmw_vsock/af_vsock.c 		vsk->owner = get_cred(psk->owner);
vsk               622 net/vmw_vsock/af_vsock.c 		vsk->connect_timeout = psk->connect_timeout;
vsk               624 net/vmw_vsock/af_vsock.c 		vsk->trusted = capable(CAP_NET_ADMIN);
vsk               625 net/vmw_vsock/af_vsock.c 		vsk->owner = get_current_cred();
vsk               626 net/vmw_vsock/af_vsock.c 		vsk->connect_timeout = VSOCK_DEFAULT_CONNECT_TIMEOUT;
vsk               629 net/vmw_vsock/af_vsock.c 	if (transport->init(vsk, psk) < 0) {
vsk               635 net/vmw_vsock/af_vsock.c 		vsock_insert_unbound(vsk);
vsk               646 net/vmw_vsock/af_vsock.c 		struct vsock_sock *vsk;
vsk               648 net/vmw_vsock/af_vsock.c 		vsk = vsock_sk(sk);
vsk               654 net/vmw_vsock/af_vsock.c 		transport->release(vsk);
vsk               681 net/vmw_vsock/af_vsock.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk               683 net/vmw_vsock/af_vsock.c 	transport->destruct(vsk);
vsk               688 net/vmw_vsock/af_vsock.c 	vsock_addr_init(&vsk->local_addr, VMADDR_CID_ANY, VMADDR_PORT_ANY);
vsk               689 net/vmw_vsock/af_vsock.c 	vsock_addr_init(&vsk->remote_addr, VMADDR_CID_ANY, VMADDR_PORT_ANY);
vsk               691 net/vmw_vsock/af_vsock.c 	put_cred(vsk->owner);
vsk               705 net/vmw_vsock/af_vsock.c s64 vsock_stream_has_data(struct vsock_sock *vsk)
vsk               707 net/vmw_vsock/af_vsock.c 	return transport->stream_has_data(vsk);
vsk               711 net/vmw_vsock/af_vsock.c s64 vsock_stream_has_space(struct vsock_sock *vsk)
vsk               713 net/vmw_vsock/af_vsock.c 	return transport->stream_has_space(vsk);
vsk               750 net/vmw_vsock/af_vsock.c 	struct vsock_sock *vsk;
vsk               754 net/vmw_vsock/af_vsock.c 	vsk = vsock_sk(sk);
vsk               764 net/vmw_vsock/af_vsock.c 		vm_addr = &vsk->remote_addr;
vsk               766 net/vmw_vsock/af_vsock.c 		vm_addr = &vsk->local_addr;
vsk               842 net/vmw_vsock/af_vsock.c 	struct vsock_sock *vsk;
vsk               845 net/vmw_vsock/af_vsock.c 	vsk = vsock_sk(sk);
vsk               859 net/vmw_vsock/af_vsock.c 	     (vsk->peer_shutdown & SEND_SHUTDOWN))) {
vsk               864 net/vmw_vsock/af_vsock.c 	    vsk->peer_shutdown & SEND_SHUTDOWN) {
vsk               892 net/vmw_vsock/af_vsock.c 		if (transport->stream_is_active(vsk) &&
vsk               896 net/vmw_vsock/af_vsock.c 					vsk, 1, &data_ready_now);
vsk               911 net/vmw_vsock/af_vsock.c 		    vsk->peer_shutdown & SEND_SHUTDOWN) {
vsk               920 net/vmw_vsock/af_vsock.c 						vsk, 1, &space_avail_now);
vsk               955 net/vmw_vsock/af_vsock.c 	struct vsock_sock *vsk;
vsk               964 net/vmw_vsock/af_vsock.c 	vsk = vsock_sk(sk);
vsk               968 net/vmw_vsock/af_vsock.c 	err = vsock_auto_bind(vsk);
vsk               991 net/vmw_vsock/af_vsock.c 		remote_addr = &vsk->remote_addr;
vsk               999 net/vmw_vsock/af_vsock.c 		if (!vsock_addr_bound(&vsk->remote_addr)) {
vsk              1014 net/vmw_vsock/af_vsock.c 	err = transport->dgram_enqueue(vsk, remote_addr, msg, len);
vsk              1026 net/vmw_vsock/af_vsock.c 	struct vsock_sock *vsk;
vsk              1030 net/vmw_vsock/af_vsock.c 	vsk = vsock_sk(sk);
vsk              1035 net/vmw_vsock/af_vsock.c 		vsock_addr_init(&vsk->remote_addr, VMADDR_CID_ANY,
vsk              1045 net/vmw_vsock/af_vsock.c 	err = vsock_auto_bind(vsk);
vsk              1055 net/vmw_vsock/af_vsock.c 	memcpy(&vsk->remote_addr, remote_addr, sizeof(vsk->remote_addr));
vsk              1090 net/vmw_vsock/af_vsock.c static int vsock_transport_cancel_pkt(struct vsock_sock *vsk)
vsk              1095 net/vmw_vsock/af_vsock.c 	return transport->cancel_pkt(vsk);
vsk              1101 net/vmw_vsock/af_vsock.c 	struct vsock_sock *vsk;
vsk              1104 net/vmw_vsock/af_vsock.c 	vsk = container_of(work, struct vsock_sock, connect_work.work);
vsk              1105 net/vmw_vsock/af_vsock.c 	sk = sk_vsock(vsk);
vsk              1117 net/vmw_vsock/af_vsock.c 		vsock_transport_cancel_pkt(vsk);
vsk              1127 net/vmw_vsock/af_vsock.c 	struct vsock_sock *vsk;
vsk              1134 net/vmw_vsock/af_vsock.c 	vsk = vsock_sk(sk);
vsk              1172 net/vmw_vsock/af_vsock.c 		memcpy(&vsk->remote_addr, remote_addr,
vsk              1173 net/vmw_vsock/af_vsock.c 		       sizeof(vsk->remote_addr));
vsk              1175 net/vmw_vsock/af_vsock.c 		err = vsock_auto_bind(vsk);
vsk              1181 net/vmw_vsock/af_vsock.c 		err = transport->connect(vsk);
vsk              1196 net/vmw_vsock/af_vsock.c 	timeout = vsk->connect_timeout;
vsk              1208 net/vmw_vsock/af_vsock.c 			schedule_delayed_work(&vsk->connect_work, timeout);
vsk              1222 net/vmw_vsock/af_vsock.c 			vsock_transport_cancel_pkt(vsk);
vsk              1228 net/vmw_vsock/af_vsock.c 			vsock_transport_cancel_pkt(vsk);
vsk              1336 net/vmw_vsock/af_vsock.c 	struct vsock_sock *vsk;
vsk              1352 net/vmw_vsock/af_vsock.c 	vsk = vsock_sk(sk);
vsk              1354 net/vmw_vsock/af_vsock.c 	if (!vsock_addr_bound(&vsk->local_addr)) {
vsk              1377 net/vmw_vsock/af_vsock.c 	struct vsock_sock *vsk;
vsk              1397 net/vmw_vsock/af_vsock.c 	vsk = vsock_sk(sk);
vsk              1404 net/vmw_vsock/af_vsock.c 		transport->set_buffer_size(vsk, val);
vsk              1409 net/vmw_vsock/af_vsock.c 		transport->set_max_buffer_size(vsk, val);
vsk              1414 net/vmw_vsock/af_vsock.c 		transport->set_min_buffer_size(vsk, val);
vsk              1422 net/vmw_vsock/af_vsock.c 			vsk->connect_timeout = tv.tv_sec * HZ +
vsk              1424 net/vmw_vsock/af_vsock.c 			if (vsk->connect_timeout == 0)
vsk              1425 net/vmw_vsock/af_vsock.c 				vsk->connect_timeout =
vsk              1454 net/vmw_vsock/af_vsock.c 	struct vsock_sock *vsk;
vsk              1477 net/vmw_vsock/af_vsock.c 	vsk = vsock_sk(sk);
vsk              1481 net/vmw_vsock/af_vsock.c 		val = transport->get_buffer_size(vsk);
vsk              1486 net/vmw_vsock/af_vsock.c 		val = transport->get_max_buffer_size(vsk);
vsk              1491 net/vmw_vsock/af_vsock.c 		val = transport->get_min_buffer_size(vsk);
vsk              1497 net/vmw_vsock/af_vsock.c 		tv.tv_sec = vsk->connect_timeout / HZ;
vsk              1499 net/vmw_vsock/af_vsock.c 		    (vsk->connect_timeout -
vsk              1521 net/vmw_vsock/af_vsock.c 	struct vsock_sock *vsk;
vsk              1529 net/vmw_vsock/af_vsock.c 	vsk = vsock_sk(sk);
vsk              1546 net/vmw_vsock/af_vsock.c 	    vsk->peer_shutdown & RCV_SHUTDOWN) {
vsk              1552 net/vmw_vsock/af_vsock.c 	    !vsock_addr_bound(&vsk->local_addr)) {
vsk              1557 net/vmw_vsock/af_vsock.c 	if (!vsock_addr_bound(&vsk->remote_addr)) {
vsk              1565 net/vmw_vsock/af_vsock.c 	err = transport->notify_send_init(vsk, &send_data);
vsk              1573 net/vmw_vsock/af_vsock.c 		while (vsock_stream_has_space(vsk) == 0 &&
vsk              1576 net/vmw_vsock/af_vsock.c 		       !(vsk->peer_shutdown & RCV_SHUTDOWN)) {
vsk              1585 net/vmw_vsock/af_vsock.c 			err = transport->notify_send_pre_block(vsk, &send_data);
vsk              1614 net/vmw_vsock/af_vsock.c 			   (vsk->peer_shutdown & RCV_SHUTDOWN)) {
vsk              1619 net/vmw_vsock/af_vsock.c 		err = transport->notify_send_pre_enqueue(vsk, &send_data);
vsk              1630 net/vmw_vsock/af_vsock.c 				vsk, msg,
vsk              1640 net/vmw_vsock/af_vsock.c 				vsk, written, &send_data);
vsk              1660 net/vmw_vsock/af_vsock.c 	struct vsock_sock *vsk;
vsk              1670 net/vmw_vsock/af_vsock.c 	vsk = vsock_sk(sk);
vsk              1718 net/vmw_vsock/af_vsock.c 	if (target >= transport->stream_rcvhiwat(vsk)) {
vsk              1725 net/vmw_vsock/af_vsock.c 	err = transport->notify_recv_init(vsk, target, &recv_data);
vsk              1734 net/vmw_vsock/af_vsock.c 		ready = vsock_stream_has_data(vsk);
vsk              1739 net/vmw_vsock/af_vsock.c 			    (vsk->peer_shutdown & SEND_SHUTDOWN)) {
vsk              1751 net/vmw_vsock/af_vsock.c 					vsk, target, &recv_data);
vsk              1785 net/vmw_vsock/af_vsock.c 					vsk, target, &recv_data);
vsk              1790 net/vmw_vsock/af_vsock.c 					vsk, msg,
vsk              1800 net/vmw_vsock/af_vsock.c 					vsk, target, read,
vsk                17 net/vmw_vsock/diag.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk                37 net/vmw_vsock/diag.c 	rep->vdiag_src_cid = vsk->local_addr.svm_cid;
vsk                38 net/vmw_vsock/diag.c 	rep->vdiag_src_port = vsk->local_addr.svm_port;
vsk                39 net/vmw_vsock/diag.c 	rep->vdiag_dst_cid = vsk->remote_addr.svm_cid;
vsk                40 net/vmw_vsock/diag.c 	rep->vdiag_dst_port = vsk->remote_addr.svm_port;
vsk                51 net/vmw_vsock/diag.c 	struct vsock_sock *vsk;
vsk                76 net/vmw_vsock/diag.c 			list_for_each_entry(vsk, head, bound_table) {
vsk                77 net/vmw_vsock/diag.c 				struct sock *sk = sk_vsock(vsk);
vsk               106 net/vmw_vsock/diag.c 		list_for_each_entry(vsk, head, connected_table) {
vsk               107 net/vmw_vsock/diag.c 			struct sock *sk = sk_vsock(vsk);
vsk               110 net/vmw_vsock/diag.c 			if (__vsock_in_bound_table(vsk))
vsk                90 net/vmw_vsock/hyperv_transport.c 	struct vsock_sock *vsk;
vsk               237 net/vmw_vsock/hyperv_transport.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk               238 net/vmw_vsock/hyperv_transport.c 	struct hvsock *hvs = vsk->trans;
vsk               248 net/vmw_vsock/hyperv_transport.c static void hvs_do_close_lock_held(struct vsock_sock *vsk,
vsk               251 net/vmw_vsock/hyperv_transport.c 	struct sock *sk = sk_vsock(vsk);
vsk               254 net/vmw_vsock/hyperv_transport.c 	vsk->peer_shutdown = SHUTDOWN_MASK;
vsk               255 net/vmw_vsock/hyperv_transport.c 	if (vsock_stream_has_data(vsk) <= 0)
vsk               258 net/vmw_vsock/hyperv_transport.c 	if (vsk->close_work_scheduled &&
vsk               259 net/vmw_vsock/hyperv_transport.c 	    (!cancel_timeout || cancel_delayed_work(&vsk->close_work))) {
vsk               260 net/vmw_vsock/hyperv_transport.c 		vsk->close_work_scheduled = false;
vsk               261 net/vmw_vsock/hyperv_transport.c 		vsock_remove_sock(vsk);
vsk               421 net/vmw_vsock/hyperv_transport.c static int hvs_sock_init(struct vsock_sock *vsk, struct vsock_sock *psk)
vsk               424 net/vmw_vsock/hyperv_transport.c 	struct sock *sk = sk_vsock(vsk);
vsk               430 net/vmw_vsock/hyperv_transport.c 	vsk->trans = hvs;
vsk               431 net/vmw_vsock/hyperv_transport.c 	hvs->vsk = vsk;
vsk               437 net/vmw_vsock/hyperv_transport.c static int hvs_connect(struct vsock_sock *vsk)
vsk               440 net/vmw_vsock/hyperv_transport.c 	struct hvsock *h = vsk->trans;
vsk               443 net/vmw_vsock/hyperv_transport.c 	vm.svm_port = vsk->local_addr.svm_port;
vsk               447 net/vmw_vsock/hyperv_transport.c 	host.svm_port = vsk->remote_addr.svm_port;
vsk               465 net/vmw_vsock/hyperv_transport.c static int hvs_shutdown(struct vsock_sock *vsk, int mode)
vsk               467 net/vmw_vsock/hyperv_transport.c 	struct sock *sk = sk_vsock(vsk);
vsk               473 net/vmw_vsock/hyperv_transport.c 	hvs_shutdown_lock_held(vsk->trans, mode);
vsk               480 net/vmw_vsock/hyperv_transport.c 	struct vsock_sock *vsk =
vsk               482 net/vmw_vsock/hyperv_transport.c 	struct sock *sk = sk_vsock(vsk);
vsk               487 net/vmw_vsock/hyperv_transport.c 		hvs_do_close_lock_held(vsk, false);
vsk               489 net/vmw_vsock/hyperv_transport.c 	vsk->close_work_scheduled = false;
vsk               495 net/vmw_vsock/hyperv_transport.c static bool hvs_close_lock_held(struct vsock_sock *vsk)
vsk               497 net/vmw_vsock/hyperv_transport.c 	struct sock *sk = sk_vsock(vsk);
vsk               504 net/vmw_vsock/hyperv_transport.c 		hvs_shutdown_lock_held(vsk->trans, SHUTDOWN_MASK);
vsk               511 net/vmw_vsock/hyperv_transport.c 	INIT_DELAYED_WORK(&vsk->close_work, hvs_close_timeout);
vsk               512 net/vmw_vsock/hyperv_transport.c 	vsk->close_work_scheduled = true;
vsk               513 net/vmw_vsock/hyperv_transport.c 	schedule_delayed_work(&vsk->close_work, HVS_CLOSE_TIMEOUT);
vsk               517 net/vmw_vsock/hyperv_transport.c static void hvs_release(struct vsock_sock *vsk)
vsk               519 net/vmw_vsock/hyperv_transport.c 	struct sock *sk = sk_vsock(vsk);
vsk               523 net/vmw_vsock/hyperv_transport.c 	remove_sock = hvs_close_lock_held(vsk);
vsk               526 net/vmw_vsock/hyperv_transport.c 		vsock_remove_sock(vsk);
vsk               529 net/vmw_vsock/hyperv_transport.c static void hvs_destruct(struct vsock_sock *vsk)
vsk               531 net/vmw_vsock/hyperv_transport.c 	struct hvsock *hvs = vsk->trans;
vsk               540 net/vmw_vsock/hyperv_transport.c static int hvs_dgram_bind(struct vsock_sock *vsk, struct sockaddr_vm *addr)
vsk               545 net/vmw_vsock/hyperv_transport.c static int hvs_dgram_dequeue(struct vsock_sock *vsk, struct msghdr *msg,
vsk               551 net/vmw_vsock/hyperv_transport.c static int hvs_dgram_enqueue(struct vsock_sock *vsk,
vsk               575 net/vmw_vsock/hyperv_transport.c 		hvs->vsk->peer_shutdown |= SEND_SHUTDOWN;
vsk               583 net/vmw_vsock/hyperv_transport.c static ssize_t hvs_stream_dequeue(struct vsock_sock *vsk, struct msghdr *msg,
vsk               586 net/vmw_vsock/hyperv_transport.c 	struct hvsock *hvs = vsk->trans;
vsk               623 net/vmw_vsock/hyperv_transport.c static ssize_t hvs_stream_enqueue(struct vsock_sock *vsk, struct msghdr *msg,
vsk               626 net/vmw_vsock/hyperv_transport.c 	struct hvsock *hvs = vsk->trans;
vsk               671 net/vmw_vsock/hyperv_transport.c static s64 hvs_stream_has_data(struct vsock_sock *vsk)
vsk               673 net/vmw_vsock/hyperv_transport.c 	struct hvsock *hvs = vsk->trans;
vsk               684 net/vmw_vsock/hyperv_transport.c 		vsk->peer_shutdown |= SEND_SHUTDOWN;
vsk               695 net/vmw_vsock/hyperv_transport.c static s64 hvs_stream_has_space(struct vsock_sock *vsk)
vsk               697 net/vmw_vsock/hyperv_transport.c 	struct hvsock *hvs = vsk->trans;
vsk               702 net/vmw_vsock/hyperv_transport.c static u64 hvs_stream_rcvhiwat(struct vsock_sock *vsk)
vsk               707 net/vmw_vsock/hyperv_transport.c static bool hvs_stream_is_active(struct vsock_sock *vsk)
vsk               709 net/vmw_vsock/hyperv_transport.c 	struct hvsock *hvs = vsk->trans;
vsk               723 net/vmw_vsock/hyperv_transport.c int hvs_notify_poll_in(struct vsock_sock *vsk, size_t target, bool *readable)
vsk               725 net/vmw_vsock/hyperv_transport.c 	struct hvsock *hvs = vsk->trans;
vsk               732 net/vmw_vsock/hyperv_transport.c int hvs_notify_poll_out(struct vsock_sock *vsk, size_t target, bool *writable)
vsk               734 net/vmw_vsock/hyperv_transport.c 	*writable = hvs_stream_has_space(vsk) > 0;
vsk               740 net/vmw_vsock/hyperv_transport.c int hvs_notify_recv_init(struct vsock_sock *vsk, size_t target,
vsk               747 net/vmw_vsock/hyperv_transport.c int hvs_notify_recv_pre_block(struct vsock_sock *vsk, size_t target,
vsk               754 net/vmw_vsock/hyperv_transport.c int hvs_notify_recv_pre_dequeue(struct vsock_sock *vsk, size_t target,
vsk               761 net/vmw_vsock/hyperv_transport.c int hvs_notify_recv_post_dequeue(struct vsock_sock *vsk, size_t target,
vsk               769 net/vmw_vsock/hyperv_transport.c int hvs_notify_send_init(struct vsock_sock *vsk,
vsk               776 net/vmw_vsock/hyperv_transport.c int hvs_notify_send_pre_block(struct vsock_sock *vsk,
vsk               783 net/vmw_vsock/hyperv_transport.c int hvs_notify_send_pre_enqueue(struct vsock_sock *vsk,
vsk               790 net/vmw_vsock/hyperv_transport.c int hvs_notify_send_post_enqueue(struct vsock_sock *vsk, ssize_t written,
vsk               796 net/vmw_vsock/hyperv_transport.c static void hvs_set_buffer_size(struct vsock_sock *vsk, u64 val)
vsk               801 net/vmw_vsock/hyperv_transport.c static void hvs_set_min_buffer_size(struct vsock_sock *vsk, u64 val)
vsk               806 net/vmw_vsock/hyperv_transport.c static void hvs_set_max_buffer_size(struct vsock_sock *vsk, u64 val)
vsk               811 net/vmw_vsock/hyperv_transport.c static u64 hvs_get_buffer_size(struct vsock_sock *vsk)
vsk               816 net/vmw_vsock/hyperv_transport.c static u64 hvs_get_min_buffer_size(struct vsock_sock *vsk)
vsk               821 net/vmw_vsock/hyperv_transport.c static u64 hvs_get_max_buffer_size(struct vsock_sock *vsk)
vsk               243 net/vmw_vsock/virtio_transport.c virtio_transport_cancel_pkt(struct vsock_sock *vsk)
vsk               259 net/vmw_vsock/virtio_transport.c 		if (pkt->vsk != vsk)
vsk                64 net/vmw_vsock/virtio_transport_common.c 	pkt->vsk		= info->vsk;
vsk               163 net/vmw_vsock/virtio_transport_common.c static int virtio_transport_send_pkt_info(struct vsock_sock *vsk,
vsk               172 net/vmw_vsock/virtio_transport_common.c 	src_port = vsk->local_addr.svm_port;
vsk               174 net/vmw_vsock/virtio_transport_common.c 		dst_cid	= vsk->remote_addr.svm_cid;
vsk               175 net/vmw_vsock/virtio_transport_common.c 		dst_port = vsk->remote_addr.svm_port;
vsk               181 net/vmw_vsock/virtio_transport_common.c 	vvs = vsk->trans;
vsk               257 net/vmw_vsock/virtio_transport_common.c static int virtio_transport_send_credit_update(struct vsock_sock *vsk,
vsk               264 net/vmw_vsock/virtio_transport_common.c 		.vsk = vsk,
vsk               267 net/vmw_vsock/virtio_transport_common.c 	return virtio_transport_send_pkt_info(vsk, &info);
vsk               271 net/vmw_vsock/virtio_transport_common.c virtio_transport_stream_do_dequeue(struct vsock_sock *vsk,
vsk               275 net/vmw_vsock/virtio_transport_common.c 	struct virtio_vsock_sock *vvs = vsk->trans;
vsk               323 net/vmw_vsock/virtio_transport_common.c 		virtio_transport_send_credit_update(vsk,
vsk               337 net/vmw_vsock/virtio_transport_common.c virtio_transport_stream_dequeue(struct vsock_sock *vsk,
vsk               344 net/vmw_vsock/virtio_transport_common.c 	return virtio_transport_stream_do_dequeue(vsk, msg, len);
vsk               349 net/vmw_vsock/virtio_transport_common.c virtio_transport_dgram_dequeue(struct vsock_sock *vsk,
vsk               357 net/vmw_vsock/virtio_transport_common.c s64 virtio_transport_stream_has_data(struct vsock_sock *vsk)
vsk               359 net/vmw_vsock/virtio_transport_common.c 	struct virtio_vsock_sock *vvs = vsk->trans;
vsk               370 net/vmw_vsock/virtio_transport_common.c static s64 virtio_transport_has_space(struct vsock_sock *vsk)
vsk               372 net/vmw_vsock/virtio_transport_common.c 	struct virtio_vsock_sock *vvs = vsk->trans;
vsk               382 net/vmw_vsock/virtio_transport_common.c s64 virtio_transport_stream_has_space(struct vsock_sock *vsk)
vsk               384 net/vmw_vsock/virtio_transport_common.c 	struct virtio_vsock_sock *vvs = vsk->trans;
vsk               388 net/vmw_vsock/virtio_transport_common.c 	bytes = virtio_transport_has_space(vsk);
vsk               395 net/vmw_vsock/virtio_transport_common.c int virtio_transport_do_socket_init(struct vsock_sock *vsk,
vsk               404 net/vmw_vsock/virtio_transport_common.c 	vsk->trans = vvs;
vsk               405 net/vmw_vsock/virtio_transport_common.c 	vvs->vsk = vsk;
vsk               429 net/vmw_vsock/virtio_transport_common.c u64 virtio_transport_get_buffer_size(struct vsock_sock *vsk)
vsk               431 net/vmw_vsock/virtio_transport_common.c 	struct virtio_vsock_sock *vvs = vsk->trans;
vsk               437 net/vmw_vsock/virtio_transport_common.c u64 virtio_transport_get_min_buffer_size(struct vsock_sock *vsk)
vsk               439 net/vmw_vsock/virtio_transport_common.c 	struct virtio_vsock_sock *vvs = vsk->trans;
vsk               445 net/vmw_vsock/virtio_transport_common.c u64 virtio_transport_get_max_buffer_size(struct vsock_sock *vsk)
vsk               447 net/vmw_vsock/virtio_transport_common.c 	struct virtio_vsock_sock *vvs = vsk->trans;
vsk               453 net/vmw_vsock/virtio_transport_common.c void virtio_transport_set_buffer_size(struct vsock_sock *vsk, u64 val)
vsk               455 net/vmw_vsock/virtio_transport_common.c 	struct virtio_vsock_sock *vvs = vsk->trans;
vsk               466 net/vmw_vsock/virtio_transport_common.c 	virtio_transport_send_credit_update(vsk, VIRTIO_VSOCK_TYPE_STREAM,
vsk               471 net/vmw_vsock/virtio_transport_common.c void virtio_transport_set_min_buffer_size(struct vsock_sock *vsk, u64 val)
vsk               473 net/vmw_vsock/virtio_transport_common.c 	struct virtio_vsock_sock *vvs = vsk->trans;
vsk               483 net/vmw_vsock/virtio_transport_common.c void virtio_transport_set_max_buffer_size(struct vsock_sock *vsk, u64 val)
vsk               485 net/vmw_vsock/virtio_transport_common.c 	struct virtio_vsock_sock *vvs = vsk->trans;
vsk               496 net/vmw_vsock/virtio_transport_common.c virtio_transport_notify_poll_in(struct vsock_sock *vsk,
vsk               500 net/vmw_vsock/virtio_transport_common.c 	if (vsock_stream_has_data(vsk))
vsk               510 net/vmw_vsock/virtio_transport_common.c virtio_transport_notify_poll_out(struct vsock_sock *vsk,
vsk               516 net/vmw_vsock/virtio_transport_common.c 	free_space = vsock_stream_has_space(vsk);
vsk               526 net/vmw_vsock/virtio_transport_common.c int virtio_transport_notify_recv_init(struct vsock_sock *vsk,
vsk               533 net/vmw_vsock/virtio_transport_common.c int virtio_transport_notify_recv_pre_block(struct vsock_sock *vsk,
vsk               540 net/vmw_vsock/virtio_transport_common.c int virtio_transport_notify_recv_pre_dequeue(struct vsock_sock *vsk,
vsk               547 net/vmw_vsock/virtio_transport_common.c int virtio_transport_notify_recv_post_dequeue(struct vsock_sock *vsk,
vsk               555 net/vmw_vsock/virtio_transport_common.c int virtio_transport_notify_send_init(struct vsock_sock *vsk,
vsk               562 net/vmw_vsock/virtio_transport_common.c int virtio_transport_notify_send_pre_block(struct vsock_sock *vsk,
vsk               569 net/vmw_vsock/virtio_transport_common.c int virtio_transport_notify_send_pre_enqueue(struct vsock_sock *vsk,
vsk               576 net/vmw_vsock/virtio_transport_common.c int virtio_transport_notify_send_post_enqueue(struct vsock_sock *vsk,
vsk               583 net/vmw_vsock/virtio_transport_common.c u64 virtio_transport_stream_rcvhiwat(struct vsock_sock *vsk)
vsk               585 net/vmw_vsock/virtio_transport_common.c 	struct virtio_vsock_sock *vvs = vsk->trans;
vsk               591 net/vmw_vsock/virtio_transport_common.c bool virtio_transport_stream_is_active(struct vsock_sock *vsk)
vsk               603 net/vmw_vsock/virtio_transport_common.c int virtio_transport_dgram_bind(struct vsock_sock *vsk,
vsk               616 net/vmw_vsock/virtio_transport_common.c int virtio_transport_connect(struct vsock_sock *vsk)
vsk               621 net/vmw_vsock/virtio_transport_common.c 		.vsk = vsk,
vsk               624 net/vmw_vsock/virtio_transport_common.c 	return virtio_transport_send_pkt_info(vsk, &info);
vsk               628 net/vmw_vsock/virtio_transport_common.c int virtio_transport_shutdown(struct vsock_sock *vsk, int mode)
vsk               637 net/vmw_vsock/virtio_transport_common.c 		.vsk = vsk,
vsk               640 net/vmw_vsock/virtio_transport_common.c 	return virtio_transport_send_pkt_info(vsk, &info);
vsk               645 net/vmw_vsock/virtio_transport_common.c virtio_transport_dgram_enqueue(struct vsock_sock *vsk,
vsk               655 net/vmw_vsock/virtio_transport_common.c virtio_transport_stream_enqueue(struct vsock_sock *vsk,
vsk               664 net/vmw_vsock/virtio_transport_common.c 		.vsk = vsk,
vsk               667 net/vmw_vsock/virtio_transport_common.c 	return virtio_transport_send_pkt_info(vsk, &info);
vsk               671 net/vmw_vsock/virtio_transport_common.c void virtio_transport_destruct(struct vsock_sock *vsk)
vsk               673 net/vmw_vsock/virtio_transport_common.c 	struct virtio_vsock_sock *vvs = vsk->trans;
vsk               679 net/vmw_vsock/virtio_transport_common.c static int virtio_transport_reset(struct vsock_sock *vsk,
vsk               686 net/vmw_vsock/virtio_transport_common.c 		.vsk = vsk,
vsk               693 net/vmw_vsock/virtio_transport_common.c 	return virtio_transport_send_pkt_info(vsk, &info);
vsk               747 net/vmw_vsock/virtio_transport_common.c static void virtio_transport_do_close(struct vsock_sock *vsk,
vsk               750 net/vmw_vsock/virtio_transport_common.c 	struct sock *sk = sk_vsock(vsk);
vsk               753 net/vmw_vsock/virtio_transport_common.c 	vsk->peer_shutdown = SHUTDOWN_MASK;
vsk               754 net/vmw_vsock/virtio_transport_common.c 	if (vsock_stream_has_data(vsk) <= 0)
vsk               758 net/vmw_vsock/virtio_transport_common.c 	if (vsk->close_work_scheduled &&
vsk               759 net/vmw_vsock/virtio_transport_common.c 	    (!cancel_timeout || cancel_delayed_work(&vsk->close_work))) {
vsk               760 net/vmw_vsock/virtio_transport_common.c 		vsk->close_work_scheduled = false;
vsk               762 net/vmw_vsock/virtio_transport_common.c 		vsock_remove_sock(vsk);
vsk               771 net/vmw_vsock/virtio_transport_common.c 	struct vsock_sock *vsk =
vsk               773 net/vmw_vsock/virtio_transport_common.c 	struct sock *sk = sk_vsock(vsk);
vsk               779 net/vmw_vsock/virtio_transport_common.c 		(void)virtio_transport_reset(vsk, NULL);
vsk               781 net/vmw_vsock/virtio_transport_common.c 		virtio_transport_do_close(vsk, false);
vsk               784 net/vmw_vsock/virtio_transport_common.c 	vsk->close_work_scheduled = false;
vsk               791 net/vmw_vsock/virtio_transport_common.c static bool virtio_transport_close(struct vsock_sock *vsk)
vsk               793 net/vmw_vsock/virtio_transport_common.c 	struct sock *sk = &vsk->sk;
vsk               800 net/vmw_vsock/virtio_transport_common.c 	if ((vsk->peer_shutdown & SHUTDOWN_MASK) == SHUTDOWN_MASK) {
vsk               801 net/vmw_vsock/virtio_transport_common.c 		(void)virtio_transport_reset(vsk, NULL);
vsk               806 net/vmw_vsock/virtio_transport_common.c 		(void)virtio_transport_shutdown(vsk, SHUTDOWN_MASK);
vsk               816 net/vmw_vsock/virtio_transport_common.c 	INIT_DELAYED_WORK(&vsk->close_work,
vsk               818 net/vmw_vsock/virtio_transport_common.c 	vsk->close_work_scheduled = true;
vsk               819 net/vmw_vsock/virtio_transport_common.c 	schedule_delayed_work(&vsk->close_work, VSOCK_CLOSE_TIMEOUT);
vsk               823 net/vmw_vsock/virtio_transport_common.c void virtio_transport_release(struct vsock_sock *vsk)
vsk               825 net/vmw_vsock/virtio_transport_common.c 	struct virtio_vsock_sock *vvs = vsk->trans;
vsk               827 net/vmw_vsock/virtio_transport_common.c 	struct sock *sk = &vsk->sk;
vsk               832 net/vmw_vsock/virtio_transport_common.c 		remove_sock = virtio_transport_close(vsk);
vsk               841 net/vmw_vsock/virtio_transport_common.c 		vsock_remove_sock(vsk);
vsk               849 net/vmw_vsock/virtio_transport_common.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk               857 net/vmw_vsock/virtio_transport_common.c 		vsock_insert_connected(vsk);
vsk               874 net/vmw_vsock/virtio_transport_common.c 	virtio_transport_reset(vsk, pkt);
vsk               882 net/vmw_vsock/virtio_transport_common.c virtio_transport_recv_enqueue(struct vsock_sock *vsk,
vsk               885 net/vmw_vsock/virtio_transport_common.c 	struct virtio_vsock_sock *vvs = vsk->trans;
vsk               933 net/vmw_vsock/virtio_transport_common.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk               938 net/vmw_vsock/virtio_transport_common.c 		virtio_transport_recv_enqueue(vsk, pkt);
vsk               946 net/vmw_vsock/virtio_transport_common.c 			vsk->peer_shutdown |= RCV_SHUTDOWN;
vsk               948 net/vmw_vsock/virtio_transport_common.c 			vsk->peer_shutdown |= SEND_SHUTDOWN;
vsk               949 net/vmw_vsock/virtio_transport_common.c 		if (vsk->peer_shutdown == SHUTDOWN_MASK &&
vsk               950 net/vmw_vsock/virtio_transport_common.c 		    vsock_stream_has_data(vsk) <= 0 &&
vsk               952 net/vmw_vsock/virtio_transport_common.c 			(void)virtio_transport_reset(vsk, NULL);
vsk               954 net/vmw_vsock/virtio_transport_common.c 			virtio_transport_do_close(vsk, true);
vsk               960 net/vmw_vsock/virtio_transport_common.c 		virtio_transport_do_close(vsk, true);
vsk               975 net/vmw_vsock/virtio_transport_common.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk               978 net/vmw_vsock/virtio_transport_common.c 		virtio_transport_do_close(vsk, true);
vsk               982 net/vmw_vsock/virtio_transport_common.c virtio_transport_send_response(struct vsock_sock *vsk,
vsk               991 net/vmw_vsock/virtio_transport_common.c 		.vsk = vsk,
vsk               994 net/vmw_vsock/virtio_transport_common.c 	return virtio_transport_send_pkt_info(vsk, &info);
vsk              1001 net/vmw_vsock/virtio_transport_common.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk              1006 net/vmw_vsock/virtio_transport_common.c 		virtio_transport_reset(vsk, pkt);
vsk              1011 net/vmw_vsock/virtio_transport_common.c 		virtio_transport_reset(vsk, pkt);
vsk              1018 net/vmw_vsock/virtio_transport_common.c 		virtio_transport_reset(vsk, pkt);
vsk              1047 net/vmw_vsock/virtio_transport_common.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk              1048 net/vmw_vsock/virtio_transport_common.c 	struct virtio_vsock_sock *vvs = vsk->trans;
vsk              1055 net/vmw_vsock/virtio_transport_common.c 	space_available = virtio_transport_has_space(vsk);
vsk              1066 net/vmw_vsock/virtio_transport_common.c 	struct vsock_sock *vsk;
vsk              1101 net/vmw_vsock/virtio_transport_common.c 	vsk = vsock_sk(sk);
vsk              1108 net/vmw_vsock/virtio_transport_common.c 	vsk->local_addr.svm_cid = dst.svm_cid;
vsk               292 net/vmw_vsock/vmci_transport.c 	struct vsock_sock *vsk;
vsk               294 net/vmw_vsock/vmci_transport.c 	vsk = vsock_sk(sk);
vsk               296 net/vmw_vsock/vmci_transport.c 	if (!vsock_addr_bound(&vsk->local_addr))
vsk               299 net/vmw_vsock/vmci_transport.c 	if (!vsock_addr_bound(&vsk->remote_addr))
vsk               302 net/vmw_vsock/vmci_transport.c 	return vmci_transport_alloc_send_control_pkt(&vsk->local_addr,
vsk               303 net/vmw_vsock/vmci_transport.c 						     &vsk->remote_addr,
vsk               325 net/vmw_vsock/vmci_transport.c 	struct vsock_sock *vsk;
vsk               330 net/vmw_vsock/vmci_transport.c 	vsk = vsock_sk(sk);
vsk               332 net/vmw_vsock/vmci_transport.c 	if (!vsock_addr_bound(&vsk->local_addr))
vsk               335 net/vmw_vsock/vmci_transport.c 	if (vsock_addr_bound(&vsk->remote_addr)) {
vsk               336 net/vmw_vsock/vmci_transport.c 		dst_ptr = &vsk->remote_addr;
vsk               342 net/vmw_vsock/vmci_transport.c 	return vmci_transport_alloc_send_control_pkt(&vsk->local_addr, dst_ptr,
vsk               456 net/vmw_vsock/vmci_transport.c static int vmci_transport_shutdown(struct vsock_sock *vsk, int mode)
vsk               459 net/vmw_vsock/vmci_transport.c 					&vsk->sk,
vsk               623 net/vmw_vsock/vmci_transport.c 	struct vsock_sock *vsk;
vsk               636 net/vmw_vsock/vmci_transport.c 	vsk = vsock_sk(sk);
vsk               637 net/vmw_vsock/vmci_transport.c 	if (!vmci_transport_allow_dgram(vsk, dg->src.context))
vsk               684 net/vmw_vsock/vmci_transport.c 	struct vsock_sock *vsk;
vsk               753 net/vmw_vsock/vmci_transport.c 	vsk = vsock_sk(sk);
vsk               754 net/vmw_vsock/vmci_transport.c 	if (!vmci_transport_allow_dgram(vsk, pkt->dg.src.context)) {
vsk               768 net/vmw_vsock/vmci_transport.c 		vsk->local_addr.svm_cid = dst.svm_cid;
vsk               771 net/vmw_vsock/vmci_transport.c 			vmci_trans(vsk)->notify_ops->handle_notify_pkt(
vsk               812 net/vmw_vsock/vmci_transport.c 	struct vsock_sock *vsk;
vsk               814 net/vmw_vsock/vmci_transport.c 	vsk = vsock_sk(sk);
vsk               815 net/vmw_vsock/vmci_transport.c 	if (!vmci_handle_is_invalid(vmci_trans(vsk)->qp_handle)) {
vsk               821 net/vmw_vsock/vmci_transport.c 		vsk->peer_shutdown = SHUTDOWN_MASK;
vsk               830 net/vmw_vsock/vmci_transport.c 		if (vsk->local_addr.svm_cid == VMADDR_CID_HOST ||
vsk               831 net/vmw_vsock/vmci_transport.c 		    vsock_stream_has_data(vsk) <= 0) {
vsk              1279 net/vmw_vsock/vmci_transport.c 	struct vsock_sock *vsk;
vsk              1283 net/vmw_vsock/vmci_transport.c 	vsk = vsock_sk(sk);
vsk              1289 net/vmw_vsock/vmci_transport.c 					  vmci_trans(vsk)->qp_handle)) {
vsk              1302 net/vmw_vsock/vmci_transport.c 		vsock_insert_connected(vsk);
vsk              1309 net/vmw_vsock/vmci_transport.c 		    || pkt->dg.src.context != vsk->remote_addr.svm_cid
vsk              1310 net/vmw_vsock/vmci_transport.c 		    || pkt->src_port != vsk->remote_addr.svm_port
vsk              1311 net/vmw_vsock/vmci_transport.c 		    || !vmci_handle_is_invalid(vmci_trans(vsk)->qp_handle)
vsk              1312 net/vmw_vsock/vmci_transport.c 		    || vmci_trans(vsk)->qpair
vsk              1313 net/vmw_vsock/vmci_transport.c 		    || vmci_trans(vsk)->produce_size != 0
vsk              1314 net/vmw_vsock/vmci_transport.c 		    || vmci_trans(vsk)->consume_size != 0
vsk              1315 net/vmw_vsock/vmci_transport.c 		    || vmci_trans(vsk)->detach_sub_id != VMCI_INVALID_ID) {
vsk              1347 net/vmw_vsock/vmci_transport.c 		if (vsk->ignore_connecting_rst) {
vsk              1348 net/vmw_vsock/vmci_transport.c 			vsk->ignore_connecting_rst = false;
vsk              1379 net/vmw_vsock/vmci_transport.c 	struct vsock_sock *vsk;
vsk              1389 net/vmw_vsock/vmci_transport.c 	vsk = vsock_sk(sk);
vsk              1396 net/vmw_vsock/vmci_transport.c 	vsk->sent_request = false;
vsk              1397 net/vmw_vsock/vmci_transport.c 	vsk->ignore_connecting_rst = false;
vsk              1400 net/vmw_vsock/vmci_transport.c 	if (pkt->u.size < vmci_trans(vsk)->queue_pair_min_size ||
vsk              1401 net/vmw_vsock/vmci_transport.c 	    pkt->u.size > vmci_trans(vsk)->queue_pair_max_size) {
vsk              1408 net/vmw_vsock/vmci_transport.c 	if (vsk->local_addr.svm_cid == VMADDR_CID_ANY)
vsk              1409 net/vmw_vsock/vmci_transport.c 		vsk->local_addr.svm_cid = pkt->dg.dst.context;
vsk              1443 net/vmw_vsock/vmci_transport.c 				   vmci_trans(vsk), &detach_sub_id);
vsk              1451 net/vmw_vsock/vmci_transport.c 	is_local = vsk->remote_addr.svm_cid == vsk->local_addr.svm_cid;
vsk              1458 net/vmw_vsock/vmci_transport.c 					      vsk->remote_addr.svm_cid,
vsk              1461 net/vmw_vsock/vmci_transport.c 						  vsk,
vsk              1462 net/vmw_vsock/vmci_transport.c 						  vsk->
vsk              1473 net/vmw_vsock/vmci_transport.c 	vmci_trans(vsk)->qp_handle = handle;
vsk              1474 net/vmw_vsock/vmci_transport.c 	vmci_trans(vsk)->qpair = qpair;
vsk              1476 net/vmw_vsock/vmci_transport.c 	vmci_trans(vsk)->produce_size = vmci_trans(vsk)->consume_size =
vsk              1479 net/vmw_vsock/vmci_transport.c 	vmci_trans(vsk)->detach_sub_id = detach_sub_id;
vsk              1481 net/vmw_vsock/vmci_transport.c 	vmci_trans(vsk)->notify_ops->process_negotiate(sk);
vsk              1500 net/vmw_vsock/vmci_transport.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk              1502 net/vmw_vsock/vmci_transport.c 	if (vsk->sent_request) {
vsk              1503 net/vmw_vsock/vmci_transport.c 		vsk->sent_request = false;
vsk              1504 net/vmw_vsock/vmci_transport.c 		vsk->ignore_connecting_rst = true;
vsk              1507 net/vmw_vsock/vmci_transport.c 			sk, vmci_trans(vsk)->queue_pair_size);
vsk              1521 net/vmw_vsock/vmci_transport.c 	struct vsock_sock *vsk;
vsk              1535 net/vmw_vsock/vmci_transport.c 			vsk = vsock_sk(sk);
vsk              1537 net/vmw_vsock/vmci_transport.c 			vsk->peer_shutdown |= pkt->u.mode;
vsk              1543 net/vmw_vsock/vmci_transport.c 		vsk = vsock_sk(sk);
vsk              1554 net/vmw_vsock/vmci_transport.c 		vsk->peer_shutdown = SHUTDOWN_MASK;
vsk              1555 net/vmw_vsock/vmci_transport.c 		if (vsock_stream_has_data(vsk) <= 0)
vsk              1562 net/vmw_vsock/vmci_transport.c 		vsk = vsock_sk(sk);
vsk              1563 net/vmw_vsock/vmci_transport.c 		vmci_trans(vsk)->notify_ops->handle_notify_pkt(
vsk              1575 net/vmw_vsock/vmci_transport.c static int vmci_transport_socket_init(struct vsock_sock *vsk,
vsk              1578 net/vmw_vsock/vmci_transport.c 	vsk->trans = kmalloc(sizeof(struct vmci_transport), GFP_KERNEL);
vsk              1579 net/vmw_vsock/vmci_transport.c 	if (!vsk->trans)
vsk              1582 net/vmw_vsock/vmci_transport.c 	vmci_trans(vsk)->dg_handle = VMCI_INVALID_HANDLE;
vsk              1583 net/vmw_vsock/vmci_transport.c 	vmci_trans(vsk)->qp_handle = VMCI_INVALID_HANDLE;
vsk              1584 net/vmw_vsock/vmci_transport.c 	vmci_trans(vsk)->qpair = NULL;
vsk              1585 net/vmw_vsock/vmci_transport.c 	vmci_trans(vsk)->produce_size = vmci_trans(vsk)->consume_size = 0;
vsk              1586 net/vmw_vsock/vmci_transport.c 	vmci_trans(vsk)->detach_sub_id = VMCI_INVALID_ID;
vsk              1587 net/vmw_vsock/vmci_transport.c 	vmci_trans(vsk)->notify_ops = NULL;
vsk              1588 net/vmw_vsock/vmci_transport.c 	INIT_LIST_HEAD(&vmci_trans(vsk)->elem);
vsk              1589 net/vmw_vsock/vmci_transport.c 	vmci_trans(vsk)->sk = &vsk->sk;
vsk              1590 net/vmw_vsock/vmci_transport.c 	spin_lock_init(&vmci_trans(vsk)->lock);
vsk              1592 net/vmw_vsock/vmci_transport.c 		vmci_trans(vsk)->queue_pair_size =
vsk              1594 net/vmw_vsock/vmci_transport.c 		vmci_trans(vsk)->queue_pair_min_size =
vsk              1596 net/vmw_vsock/vmci_transport.c 		vmci_trans(vsk)->queue_pair_max_size =
vsk              1599 net/vmw_vsock/vmci_transport.c 		vmci_trans(vsk)->queue_pair_size =
vsk              1601 net/vmw_vsock/vmci_transport.c 		vmci_trans(vsk)->queue_pair_min_size =
vsk              1603 net/vmw_vsock/vmci_transport.c 		vmci_trans(vsk)->queue_pair_max_size =
vsk              1644 net/vmw_vsock/vmci_transport.c static void vmci_transport_destruct(struct vsock_sock *vsk)
vsk              1647 net/vmw_vsock/vmci_transport.c 	if (!vmci_trans(vsk))
vsk              1653 net/vmw_vsock/vmci_transport.c 	spin_lock_bh(&vmci_trans(vsk)->lock);
vsk              1654 net/vmw_vsock/vmci_transport.c 	vmci_trans(vsk)->sk = NULL;
vsk              1655 net/vmw_vsock/vmci_transport.c 	spin_unlock_bh(&vmci_trans(vsk)->lock);
vsk              1657 net/vmw_vsock/vmci_transport.c 	if (vmci_trans(vsk)->notify_ops)
vsk              1658 net/vmw_vsock/vmci_transport.c 		vmci_trans(vsk)->notify_ops->socket_destruct(vsk);
vsk              1661 net/vmw_vsock/vmci_transport.c 	list_add(&vmci_trans(vsk)->elem, &vmci_transport_cleanup_list);
vsk              1665 net/vmw_vsock/vmci_transport.c 	vsk->trans = NULL;
vsk              1668 net/vmw_vsock/vmci_transport.c static void vmci_transport_release(struct vsock_sock *vsk)
vsk              1670 net/vmw_vsock/vmci_transport.c 	vsock_remove_sock(vsk);
vsk              1672 net/vmw_vsock/vmci_transport.c 	if (!vmci_handle_is_invalid(vmci_trans(vsk)->dg_handle)) {
vsk              1673 net/vmw_vsock/vmci_transport.c 		vmci_datagram_destroy_handle(vmci_trans(vsk)->dg_handle);
vsk              1674 net/vmw_vsock/vmci_transport.c 		vmci_trans(vsk)->dg_handle = VMCI_INVALID_HANDLE;
vsk              1678 net/vmw_vsock/vmci_transport.c static int vmci_transport_dgram_bind(struct vsock_sock *vsk,
vsk              1699 net/vmw_vsock/vmci_transport.c 						 &vsk->sk,
vsk              1700 net/vmw_vsock/vmci_transport.c 						 &vmci_trans(vsk)->dg_handle);
vsk              1703 net/vmw_vsock/vmci_transport.c 	vsock_addr_init(&vsk->local_addr, addr->svm_cid,
vsk              1704 net/vmw_vsock/vmci_transport.c 			vmci_trans(vsk)->dg_handle.resource);
vsk              1710 net/vmw_vsock/vmci_transport.c 	struct vsock_sock *vsk,
vsk              1721 net/vmw_vsock/vmci_transport.c 	if (!vmci_transport_allow_dgram(vsk, remote_addr->svm_cid))
vsk              1733 net/vmw_vsock/vmci_transport.c 	dg->src = vmci_make_handle(vsk->local_addr.svm_cid,
vsk              1734 net/vmw_vsock/vmci_transport.c 				   vsk->local_addr.svm_port);
vsk              1745 net/vmw_vsock/vmci_transport.c static int vmci_transport_dgram_dequeue(struct vsock_sock *vsk,
vsk              1762 net/vmw_vsock/vmci_transport.c 	skb = skb_recv_datagram(&vsk->sk, flags, noblock, &err);
vsk              1797 net/vmw_vsock/vmci_transport.c 	skb_free_datagram(&vsk->sk, skb);
vsk              1813 net/vmw_vsock/vmci_transport.c static int vmci_transport_connect(struct vsock_sock *vsk)
vsk              1817 net/vmw_vsock/vmci_transport.c 	struct sock *sk = &vsk->sk;
vsk              1822 net/vmw_vsock/vmci_transport.c 			sk, vmci_trans(vsk)->queue_pair_size);
vsk              1831 net/vmw_vsock/vmci_transport.c 				sk, vmci_trans(vsk)->queue_pair_size,
vsk              1838 net/vmw_vsock/vmci_transport.c 		vsk->sent_request = true;
vsk              1845 net/vmw_vsock/vmci_transport.c 	struct vsock_sock *vsk,
vsk              1851 net/vmw_vsock/vmci_transport.c 		return vmci_qpair_peekv(vmci_trans(vsk)->qpair, msg, len, 0);
vsk              1853 net/vmw_vsock/vmci_transport.c 		return vmci_qpair_dequev(vmci_trans(vsk)->qpair, msg, len, 0);
vsk              1857 net/vmw_vsock/vmci_transport.c 	struct vsock_sock *vsk,
vsk              1861 net/vmw_vsock/vmci_transport.c 	return vmci_qpair_enquev(vmci_trans(vsk)->qpair, msg, len, 0);
vsk              1864 net/vmw_vsock/vmci_transport.c static s64 vmci_transport_stream_has_data(struct vsock_sock *vsk)
vsk              1866 net/vmw_vsock/vmci_transport.c 	return vmci_qpair_consume_buf_ready(vmci_trans(vsk)->qpair);
vsk              1869 net/vmw_vsock/vmci_transport.c static s64 vmci_transport_stream_has_space(struct vsock_sock *vsk)
vsk              1871 net/vmw_vsock/vmci_transport.c 	return vmci_qpair_produce_free_space(vmci_trans(vsk)->qpair);
vsk              1874 net/vmw_vsock/vmci_transport.c static u64 vmci_transport_stream_rcvhiwat(struct vsock_sock *vsk)
vsk              1876 net/vmw_vsock/vmci_transport.c 	return vmci_trans(vsk)->consume_size;
vsk              1879 net/vmw_vsock/vmci_transport.c static bool vmci_transport_stream_is_active(struct vsock_sock *vsk)
vsk              1881 net/vmw_vsock/vmci_transport.c 	return !vmci_handle_is_invalid(vmci_trans(vsk)->qp_handle);
vsk              1884 net/vmw_vsock/vmci_transport.c static u64 vmci_transport_get_buffer_size(struct vsock_sock *vsk)
vsk              1886 net/vmw_vsock/vmci_transport.c 	return vmci_trans(vsk)->queue_pair_size;
vsk              1889 net/vmw_vsock/vmci_transport.c static u64 vmci_transport_get_min_buffer_size(struct vsock_sock *vsk)
vsk              1891 net/vmw_vsock/vmci_transport.c 	return vmci_trans(vsk)->queue_pair_min_size;
vsk              1894 net/vmw_vsock/vmci_transport.c static u64 vmci_transport_get_max_buffer_size(struct vsock_sock *vsk)
vsk              1896 net/vmw_vsock/vmci_transport.c 	return vmci_trans(vsk)->queue_pair_max_size;
vsk              1899 net/vmw_vsock/vmci_transport.c static void vmci_transport_set_buffer_size(struct vsock_sock *vsk, u64 val)
vsk              1901 net/vmw_vsock/vmci_transport.c 	if (val < vmci_trans(vsk)->queue_pair_min_size)
vsk              1902 net/vmw_vsock/vmci_transport.c 		vmci_trans(vsk)->queue_pair_min_size = val;
vsk              1903 net/vmw_vsock/vmci_transport.c 	if (val > vmci_trans(vsk)->queue_pair_max_size)
vsk              1904 net/vmw_vsock/vmci_transport.c 		vmci_trans(vsk)->queue_pair_max_size = val;
vsk              1905 net/vmw_vsock/vmci_transport.c 	vmci_trans(vsk)->queue_pair_size = val;
vsk              1908 net/vmw_vsock/vmci_transport.c static void vmci_transport_set_min_buffer_size(struct vsock_sock *vsk,
vsk              1911 net/vmw_vsock/vmci_transport.c 	if (val > vmci_trans(vsk)->queue_pair_size)
vsk              1912 net/vmw_vsock/vmci_transport.c 		vmci_trans(vsk)->queue_pair_size = val;
vsk              1913 net/vmw_vsock/vmci_transport.c 	vmci_trans(vsk)->queue_pair_min_size = val;
vsk              1916 net/vmw_vsock/vmci_transport.c static void vmci_transport_set_max_buffer_size(struct vsock_sock *vsk,
vsk              1919 net/vmw_vsock/vmci_transport.c 	if (val < vmci_trans(vsk)->queue_pair_size)
vsk              1920 net/vmw_vsock/vmci_transport.c 		vmci_trans(vsk)->queue_pair_size = val;
vsk              1921 net/vmw_vsock/vmci_transport.c 	vmci_trans(vsk)->queue_pair_max_size = val;
vsk              1925 net/vmw_vsock/vmci_transport.c 	struct vsock_sock *vsk,
vsk              1929 net/vmw_vsock/vmci_transport.c 	return vmci_trans(vsk)->notify_ops->poll_in(
vsk              1930 net/vmw_vsock/vmci_transport.c 			&vsk->sk, target, data_ready_now);
vsk              1934 net/vmw_vsock/vmci_transport.c 	struct vsock_sock *vsk,
vsk              1938 net/vmw_vsock/vmci_transport.c 	return vmci_trans(vsk)->notify_ops->poll_out(
vsk              1939 net/vmw_vsock/vmci_transport.c 			&vsk->sk, target, space_available_now);
vsk              1943 net/vmw_vsock/vmci_transport.c 	struct vsock_sock *vsk,
vsk              1947 net/vmw_vsock/vmci_transport.c 	return vmci_trans(vsk)->notify_ops->recv_init(
vsk              1948 net/vmw_vsock/vmci_transport.c 			&vsk->sk, target,
vsk              1953 net/vmw_vsock/vmci_transport.c 	struct vsock_sock *vsk,
vsk              1957 net/vmw_vsock/vmci_transport.c 	return vmci_trans(vsk)->notify_ops->recv_pre_block(
vsk              1958 net/vmw_vsock/vmci_transport.c 			&vsk->sk, target,
vsk              1963 net/vmw_vsock/vmci_transport.c 	struct vsock_sock *vsk,
vsk              1967 net/vmw_vsock/vmci_transport.c 	return vmci_trans(vsk)->notify_ops->recv_pre_dequeue(
vsk              1968 net/vmw_vsock/vmci_transport.c 			&vsk->sk, target,
vsk              1973 net/vmw_vsock/vmci_transport.c 	struct vsock_sock *vsk,
vsk              1979 net/vmw_vsock/vmci_transport.c 	return vmci_trans(vsk)->notify_ops->recv_post_dequeue(
vsk              1980 net/vmw_vsock/vmci_transport.c 			&vsk->sk, target, copied, data_read,
vsk              1985 net/vmw_vsock/vmci_transport.c 	struct vsock_sock *vsk,
vsk              1988 net/vmw_vsock/vmci_transport.c 	return vmci_trans(vsk)->notify_ops->send_init(
vsk              1989 net/vmw_vsock/vmci_transport.c 			&vsk->sk,
vsk              1994 net/vmw_vsock/vmci_transport.c 	struct vsock_sock *vsk,
vsk              1997 net/vmw_vsock/vmci_transport.c 	return vmci_trans(vsk)->notify_ops->send_pre_block(
vsk              1998 net/vmw_vsock/vmci_transport.c 			&vsk->sk,
vsk              2003 net/vmw_vsock/vmci_transport.c 	struct vsock_sock *vsk,
vsk              2006 net/vmw_vsock/vmci_transport.c 	return vmci_trans(vsk)->notify_ops->send_pre_enqueue(
vsk              2007 net/vmw_vsock/vmci_transport.c 			&vsk->sk,
vsk              2012 net/vmw_vsock/vmci_transport.c 	struct vsock_sock *vsk,
vsk              2016 net/vmw_vsock/vmci_transport.c 	return vmci_trans(vsk)->notify_ops->send_post_enqueue(
vsk              2017 net/vmw_vsock/vmci_transport.c 			&vsk->sk, written,
vsk              2040 net/vmw_vsock/vmci_transport.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk              2047 net/vmw_vsock/vmci_transport.c 		vmci_trans(vsk)->notify_ops = &vmci_transport_notify_pkt_ops;
vsk              2053 net/vmw_vsock/vmci_transport.c 		vmci_trans(vsk)->notify_ops =
vsk              2062 net/vmw_vsock/vmci_transport.c 	vmci_trans(vsk)->notify_ops->socket_init(sk);
vsk                15 net/vmw_vsock/vmci_transport_notify.c #define PKT_FIELD(vsk, field_name) (vmci_trans(vsk)->notify.pkt.field_name)
vsk                17 net/vmw_vsock/vmci_transport_notify.c static bool vmci_transport_notify_waiting_write(struct vsock_sock *vsk)
vsk                23 net/vmw_vsock/vmci_transport_notify.c 	if (!PKT_FIELD(vsk, peer_waiting_write))
vsk                34 net/vmw_vsock/vmci_transport_notify.c 	if (!PKT_FIELD(vsk, peer_waiting_write_detected)) {
vsk                35 net/vmw_vsock/vmci_transport_notify.c 		PKT_FIELD(vsk, peer_waiting_write_detected) = true;
vsk                36 net/vmw_vsock/vmci_transport_notify.c 		if (PKT_FIELD(vsk, write_notify_window) < PAGE_SIZE) {
vsk                37 net/vmw_vsock/vmci_transport_notify.c 			PKT_FIELD(vsk, write_notify_window) =
vsk                38 net/vmw_vsock/vmci_transport_notify.c 			    PKT_FIELD(vsk, write_notify_min_window);
vsk                40 net/vmw_vsock/vmci_transport_notify.c 			PKT_FIELD(vsk, write_notify_window) -= PAGE_SIZE;
vsk                41 net/vmw_vsock/vmci_transport_notify.c 			if (PKT_FIELD(vsk, write_notify_window) <
vsk                42 net/vmw_vsock/vmci_transport_notify.c 			    PKT_FIELD(vsk, write_notify_min_window))
vsk                43 net/vmw_vsock/vmci_transport_notify.c 				PKT_FIELD(vsk, write_notify_window) =
vsk                44 net/vmw_vsock/vmci_transport_notify.c 				    PKT_FIELD(vsk, write_notify_min_window);
vsk                48 net/vmw_vsock/vmci_transport_notify.c 	notify_limit = vmci_trans(vsk)->consume_size -
vsk                49 net/vmw_vsock/vmci_transport_notify.c 		PKT_FIELD(vsk, write_notify_window);
vsk                68 net/vmw_vsock/vmci_transport_notify.c 	retval = vmci_qpair_consume_free_space(vmci_trans(vsk)->qpair) >
vsk                77 net/vmw_vsock/vmci_transport_notify.c 		PKT_FIELD(vsk, peer_waiting_write_detected) = false;
vsk                86 net/vmw_vsock/vmci_transport_notify.c static bool vmci_transport_notify_waiting_read(struct vsock_sock *vsk)
vsk                89 net/vmw_vsock/vmci_transport_notify.c 	if (!PKT_FIELD(vsk, peer_waiting_read))
vsk                98 net/vmw_vsock/vmci_transport_notify.c 	return vmci_qpair_produce_buf_ready(vmci_trans(vsk)->qpair) > 0;
vsk               112 net/vmw_vsock/vmci_transport_notify.c 	struct vsock_sock *vsk;
vsk               114 net/vmw_vsock/vmci_transport_notify.c 	vsk = vsock_sk(sk);
vsk               116 net/vmw_vsock/vmci_transport_notify.c 	PKT_FIELD(vsk, peer_waiting_read) = true;
vsk               117 net/vmw_vsock/vmci_transport_notify.c 	memcpy(&PKT_FIELD(vsk, peer_waiting_read_info), &pkt->u.wait,
vsk               118 net/vmw_vsock/vmci_transport_notify.c 	       sizeof(PKT_FIELD(vsk, peer_waiting_read_info)));
vsk               120 net/vmw_vsock/vmci_transport_notify.c 	if (vmci_transport_notify_waiting_read(vsk)) {
vsk               129 net/vmw_vsock/vmci_transport_notify.c 			PKT_FIELD(vsk, peer_waiting_read) = false;
vsk               142 net/vmw_vsock/vmci_transport_notify.c 	struct vsock_sock *vsk;
vsk               144 net/vmw_vsock/vmci_transport_notify.c 	vsk = vsock_sk(sk);
vsk               146 net/vmw_vsock/vmci_transport_notify.c 	PKT_FIELD(vsk, peer_waiting_write) = true;
vsk               147 net/vmw_vsock/vmci_transport_notify.c 	memcpy(&PKT_FIELD(vsk, peer_waiting_write_info), &pkt->u.wait,
vsk               148 net/vmw_vsock/vmci_transport_notify.c 	       sizeof(PKT_FIELD(vsk, peer_waiting_write_info)));
vsk               150 net/vmw_vsock/vmci_transport_notify.c 	if (vmci_transport_notify_waiting_write(vsk)) {
vsk               159 net/vmw_vsock/vmci_transport_notify.c 			PKT_FIELD(vsk, peer_waiting_write) = false;
vsk               171 net/vmw_vsock/vmci_transport_notify.c 	struct vsock_sock *vsk;
vsk               173 net/vmw_vsock/vmci_transport_notify.c 	vsk = vsock_sk(sk);
vsk               174 net/vmw_vsock/vmci_transport_notify.c 	PKT_FIELD(vsk, sent_waiting_write) = false;
vsk               183 net/vmw_vsock/vmci_transport_notify.c 	struct vsock_sock *vsk;
vsk               190 net/vmw_vsock/vmci_transport_notify.c 	vsk = vsock_sk(sk);
vsk               192 net/vmw_vsock/vmci_transport_notify.c 	if (PKT_FIELD(vsk, sent_waiting_read))
vsk               195 net/vmw_vsock/vmci_transport_notify.c 	if (PKT_FIELD(vsk, write_notify_window) <
vsk               196 net/vmw_vsock/vmci_transport_notify.c 			vmci_trans(vsk)->consume_size)
vsk               197 net/vmw_vsock/vmci_transport_notify.c 		PKT_FIELD(vsk, write_notify_window) =
vsk               198 net/vmw_vsock/vmci_transport_notify.c 		    min(PKT_FIELD(vsk, write_notify_window) + PAGE_SIZE,
vsk               199 net/vmw_vsock/vmci_transport_notify.c 			vmci_trans(vsk)->consume_size);
vsk               201 net/vmw_vsock/vmci_transport_notify.c 	vmci_qpair_get_consume_indexes(vmci_trans(vsk)->qpair, &tail, &head);
vsk               202 net/vmw_vsock/vmci_transport_notify.c 	room_left = vmci_trans(vsk)->consume_size - head;
vsk               206 net/vmw_vsock/vmci_transport_notify.c 		    PKT_FIELD(vsk, consume_q_generation) + 1;
vsk               209 net/vmw_vsock/vmci_transport_notify.c 		waiting_info.generation = PKT_FIELD(vsk, consume_q_generation);
vsk               214 net/vmw_vsock/vmci_transport_notify.c 		PKT_FIELD(vsk, sent_waiting_read) = true;
vsk               225 net/vmw_vsock/vmci_transport_notify.c 	struct vsock_sock *vsk;
vsk               232 net/vmw_vsock/vmci_transport_notify.c 	vsk = vsock_sk(sk);
vsk               234 net/vmw_vsock/vmci_transport_notify.c 	if (PKT_FIELD(vsk, sent_waiting_write))
vsk               237 net/vmw_vsock/vmci_transport_notify.c 	vmci_qpair_get_produce_indexes(vmci_trans(vsk)->qpair, &tail, &head);
vsk               238 net/vmw_vsock/vmci_transport_notify.c 	room_left = vmci_trans(vsk)->produce_size - tail;
vsk               242 net/vmw_vsock/vmci_transport_notify.c 		waiting_info.generation = PKT_FIELD(vsk, produce_q_generation);
vsk               246 net/vmw_vsock/vmci_transport_notify.c 		    PKT_FIELD(vsk, produce_q_generation) - 1;
vsk               251 net/vmw_vsock/vmci_transport_notify.c 		PKT_FIELD(vsk, sent_waiting_write) = true;
vsk               261 net/vmw_vsock/vmci_transport_notify.c 	struct vsock_sock *vsk;
vsk               266 net/vmw_vsock/vmci_transport_notify.c 	vsk = vsock_sk(sk);
vsk               271 net/vmw_vsock/vmci_transport_notify.c 	if (vmci_transport_notify_waiting_write(vsk)) {
vsk               279 net/vmw_vsock/vmci_transport_notify.c 		while (!(vsk->peer_shutdown & RCV_SHUTDOWN) &&
vsk               293 net/vmw_vsock/vmci_transport_notify.c 			PKT_FIELD(vsk, peer_waiting_write) = false;
vsk               307 net/vmw_vsock/vmci_transport_notify.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk               308 net/vmw_vsock/vmci_transport_notify.c 	PKT_FIELD(vsk, sent_waiting_read) = false;
vsk               315 net/vmw_vsock/vmci_transport_notify.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk               317 net/vmw_vsock/vmci_transport_notify.c 	PKT_FIELD(vsk, write_notify_window) = PAGE_SIZE;
vsk               318 net/vmw_vsock/vmci_transport_notify.c 	PKT_FIELD(vsk, write_notify_min_window) = PAGE_SIZE;
vsk               319 net/vmw_vsock/vmci_transport_notify.c 	PKT_FIELD(vsk, peer_waiting_read) = false;
vsk               320 net/vmw_vsock/vmci_transport_notify.c 	PKT_FIELD(vsk, peer_waiting_write) = false;
vsk               321 net/vmw_vsock/vmci_transport_notify.c 	PKT_FIELD(vsk, peer_waiting_write_detected) = false;
vsk               322 net/vmw_vsock/vmci_transport_notify.c 	PKT_FIELD(vsk, sent_waiting_read) = false;
vsk               323 net/vmw_vsock/vmci_transport_notify.c 	PKT_FIELD(vsk, sent_waiting_write) = false;
vsk               324 net/vmw_vsock/vmci_transport_notify.c 	PKT_FIELD(vsk, produce_q_generation) = 0;
vsk               325 net/vmw_vsock/vmci_transport_notify.c 	PKT_FIELD(vsk, consume_q_generation) = 0;
vsk               327 net/vmw_vsock/vmci_transport_notify.c 	memset(&PKT_FIELD(vsk, peer_waiting_read_info), 0,
vsk               328 net/vmw_vsock/vmci_transport_notify.c 	       sizeof(PKT_FIELD(vsk, peer_waiting_read_info)));
vsk               329 net/vmw_vsock/vmci_transport_notify.c 	memset(&PKT_FIELD(vsk, peer_waiting_write_info), 0,
vsk               330 net/vmw_vsock/vmci_transport_notify.c 	       sizeof(PKT_FIELD(vsk, peer_waiting_write_info)));
vsk               333 net/vmw_vsock/vmci_transport_notify.c static void vmci_transport_notify_pkt_socket_destruct(struct vsock_sock *vsk)
vsk               341 net/vmw_vsock/vmci_transport_notify.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk               343 net/vmw_vsock/vmci_transport_notify.c 	if (vsock_stream_has_data(vsk)) {
vsk               366 net/vmw_vsock/vmci_transport_notify.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk               368 net/vmw_vsock/vmci_transport_notify.c 	produce_q_free_space = vsock_stream_has_space(vsk);
vsk               396 net/vmw_vsock/vmci_transport_notify.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk               404 net/vmw_vsock/vmci_transport_notify.c 	if (PKT_FIELD(vsk, write_notify_min_window) < target + 1) {
vsk               405 net/vmw_vsock/vmci_transport_notify.c 		PKT_FIELD(vsk, write_notify_min_window) = target + 1;
vsk               406 net/vmw_vsock/vmci_transport_notify.c 		if (PKT_FIELD(vsk, write_notify_window) <
vsk               407 net/vmw_vsock/vmci_transport_notify.c 		    PKT_FIELD(vsk, write_notify_min_window)) {
vsk               415 net/vmw_vsock/vmci_transport_notify.c 			PKT_FIELD(vsk, write_notify_window) =
vsk               416 net/vmw_vsock/vmci_transport_notify.c 			    PKT_FIELD(vsk, write_notify_min_window);
vsk               458 net/vmw_vsock/vmci_transport_notify.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk               464 net/vmw_vsock/vmci_transport_notify.c 	vmci_qpair_get_consume_indexes(vmci_trans(vsk)->qpair,
vsk               480 net/vmw_vsock/vmci_transport_notify.c 	struct vsock_sock *vsk;
vsk               483 net/vmw_vsock/vmci_transport_notify.c 	vsk = vsock_sk(sk);
vsk               493 net/vmw_vsock/vmci_transport_notify.c 			vmci_trans(vsk)->consume_size - data->consume_head)
vsk               494 net/vmw_vsock/vmci_transport_notify.c 			PKT_FIELD(vsk, consume_q_generation)++;
vsk               535 net/vmw_vsock/vmci_transport_notify.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk               538 net/vmw_vsock/vmci_transport_notify.c 	vmci_qpair_get_produce_indexes(vmci_trans(vsk)->qpair,
vsk               553 net/vmw_vsock/vmci_transport_notify.c 	struct vsock_sock *vsk;
vsk               557 net/vmw_vsock/vmci_transport_notify.c 	vsk = vsock_sk(sk);
vsk               564 net/vmw_vsock/vmci_transport_notify.c 	if (written >= vmci_trans(vsk)->produce_size - data->produce_tail)
vsk               565 net/vmw_vsock/vmci_transport_notify.c 		PKT_FIELD(vsk, produce_q_generation)++;
vsk               569 net/vmw_vsock/vmci_transport_notify.c 	if (vmci_transport_notify_waiting_read(vsk)) {
vsk               575 net/vmw_vsock/vmci_transport_notify.c 		while (!(vsk->peer_shutdown & RCV_SHUTDOWN) &&
vsk               590 net/vmw_vsock/vmci_transport_notify.c 			PKT_FIELD(vsk, peer_waiting_read) = false;
vsk               635 net/vmw_vsock/vmci_transport_notify.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk               637 net/vmw_vsock/vmci_transport_notify.c 	PKT_FIELD(vsk, write_notify_window) = vmci_trans(vsk)->consume_size;
vsk               638 net/vmw_vsock/vmci_transport_notify.c 	if (vmci_trans(vsk)->consume_size <
vsk               639 net/vmw_vsock/vmci_transport_notify.c 		PKT_FIELD(vsk, write_notify_min_window))
vsk               640 net/vmw_vsock/vmci_transport_notify.c 		PKT_FIELD(vsk, write_notify_min_window) =
vsk               641 net/vmw_vsock/vmci_transport_notify.c 			vmci_trans(vsk)->consume_size;
vsk               646 net/vmw_vsock/vmci_transport_notify.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk               648 net/vmw_vsock/vmci_transport_notify.c 	PKT_FIELD(vsk, write_notify_window) = vmci_trans(vsk)->consume_size;
vsk               649 net/vmw_vsock/vmci_transport_notify.c 	if (vmci_trans(vsk)->consume_size <
vsk               650 net/vmw_vsock/vmci_transport_notify.c 		PKT_FIELD(vsk, write_notify_min_window))
vsk               651 net/vmw_vsock/vmci_transport_notify.c 		PKT_FIELD(vsk, write_notify_min_window) =
vsk               652 net/vmw_vsock/vmci_transport_notify.c 			vmci_trans(vsk)->consume_size;
vsk                41 net/vmw_vsock/vmci_transport_notify.h 	void (*socket_destruct) (struct vsock_sock *vsk);
vsk                15 net/vmw_vsock/vmci_transport_notify_qstate.c #define PKT_FIELD(vsk, field_name) \
vsk                16 net/vmw_vsock/vmci_transport_notify_qstate.c 	(vmci_trans(vsk)->notify.pkt_q_state.field_name)
vsk                18 net/vmw_vsock/vmci_transport_notify_qstate.c static bool vmci_transport_notify_waiting_write(struct vsock_sock *vsk)
vsk                23 net/vmw_vsock/vmci_transport_notify_qstate.c 	if (!PKT_FIELD(vsk, peer_waiting_write))
vsk                33 net/vmw_vsock/vmci_transport_notify_qstate.c 	if (!PKT_FIELD(vsk, peer_waiting_write_detected)) {
vsk                34 net/vmw_vsock/vmci_transport_notify_qstate.c 		PKT_FIELD(vsk, peer_waiting_write_detected) = true;
vsk                35 net/vmw_vsock/vmci_transport_notify_qstate.c 		if (PKT_FIELD(vsk, write_notify_window) < PAGE_SIZE) {
vsk                36 net/vmw_vsock/vmci_transport_notify_qstate.c 			PKT_FIELD(vsk, write_notify_window) =
vsk                37 net/vmw_vsock/vmci_transport_notify_qstate.c 			    PKT_FIELD(vsk, write_notify_min_window);
vsk                39 net/vmw_vsock/vmci_transport_notify_qstate.c 			PKT_FIELD(vsk, write_notify_window) -= PAGE_SIZE;
vsk                40 net/vmw_vsock/vmci_transport_notify_qstate.c 			if (PKT_FIELD(vsk, write_notify_window) <
vsk                41 net/vmw_vsock/vmci_transport_notify_qstate.c 			    PKT_FIELD(vsk, write_notify_min_window))
vsk                42 net/vmw_vsock/vmci_transport_notify_qstate.c 				PKT_FIELD(vsk, write_notify_window) =
vsk                43 net/vmw_vsock/vmci_transport_notify_qstate.c 				    PKT_FIELD(vsk, write_notify_min_window);
vsk                47 net/vmw_vsock/vmci_transport_notify_qstate.c 	notify_limit = vmci_trans(vsk)->consume_size -
vsk                48 net/vmw_vsock/vmci_transport_notify_qstate.c 		PKT_FIELD(vsk, write_notify_window);
vsk                59 net/vmw_vsock/vmci_transport_notify_qstate.c 	retval = vmci_qpair_consume_free_space(vmci_trans(vsk)->qpair) >
vsk                67 net/vmw_vsock/vmci_transport_notify_qstate.c 		PKT_FIELD(vsk, peer_waiting_write_detected) = false;
vsk                92 net/vmw_vsock/vmci_transport_notify_qstate.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk                94 net/vmw_vsock/vmci_transport_notify_qstate.c 	if (PKT_FIELD(vsk, write_notify_window) < vmci_trans(vsk)->consume_size)
vsk                95 net/vmw_vsock/vmci_transport_notify_qstate.c 		PKT_FIELD(vsk, write_notify_window) =
vsk                96 net/vmw_vsock/vmci_transport_notify_qstate.c 		    min(PKT_FIELD(vsk, write_notify_window) + PAGE_SIZE,
vsk                97 net/vmw_vsock/vmci_transport_notify_qstate.c 			vmci_trans(vsk)->consume_size);
vsk               102 net/vmw_vsock/vmci_transport_notify_qstate.c 	struct vsock_sock *vsk;
vsk               107 net/vmw_vsock/vmci_transport_notify_qstate.c 	vsk = vsock_sk(sk);
vsk               112 net/vmw_vsock/vmci_transport_notify_qstate.c 	if (vmci_transport_notify_waiting_write(vsk)) {
vsk               120 net/vmw_vsock/vmci_transport_notify_qstate.c 		while (!(vsk->peer_shutdown & RCV_SHUTDOWN) &&
vsk               134 net/vmw_vsock/vmci_transport_notify_qstate.c 			PKT_FIELD(vsk, peer_waiting_write) = false;
vsk               142 net/vmw_vsock/vmci_transport_notify_qstate.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk               144 net/vmw_vsock/vmci_transport_notify_qstate.c 	PKT_FIELD(vsk, write_notify_window) = PAGE_SIZE;
vsk               145 net/vmw_vsock/vmci_transport_notify_qstate.c 	PKT_FIELD(vsk, write_notify_min_window) = PAGE_SIZE;
vsk               146 net/vmw_vsock/vmci_transport_notify_qstate.c 	PKT_FIELD(vsk, peer_waiting_write) = false;
vsk               147 net/vmw_vsock/vmci_transport_notify_qstate.c 	PKT_FIELD(vsk, peer_waiting_write_detected) = false;
vsk               150 net/vmw_vsock/vmci_transport_notify_qstate.c static void vmci_transport_notify_pkt_socket_destruct(struct vsock_sock *vsk)
vsk               152 net/vmw_vsock/vmci_transport_notify_qstate.c 	PKT_FIELD(vsk, write_notify_window) = PAGE_SIZE;
vsk               153 net/vmw_vsock/vmci_transport_notify_qstate.c 	PKT_FIELD(vsk, write_notify_min_window) = PAGE_SIZE;
vsk               154 net/vmw_vsock/vmci_transport_notify_qstate.c 	PKT_FIELD(vsk, peer_waiting_write) = false;
vsk               155 net/vmw_vsock/vmci_transport_notify_qstate.c 	PKT_FIELD(vsk, peer_waiting_write_detected) = false;
vsk               162 net/vmw_vsock/vmci_transport_notify_qstate.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk               164 net/vmw_vsock/vmci_transport_notify_qstate.c 	if (vsock_stream_has_data(vsk)) {
vsk               184 net/vmw_vsock/vmci_transport_notify_qstate.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk               186 net/vmw_vsock/vmci_transport_notify_qstate.c 	produce_q_free_space = vsock_stream_has_space(vsk);
vsk               206 net/vmw_vsock/vmci_transport_notify_qstate.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk               212 net/vmw_vsock/vmci_transport_notify_qstate.c 	if (PKT_FIELD(vsk, write_notify_min_window) < target + 1) {
vsk               213 net/vmw_vsock/vmci_transport_notify_qstate.c 		PKT_FIELD(vsk, write_notify_min_window) = target + 1;
vsk               214 net/vmw_vsock/vmci_transport_notify_qstate.c 		if (PKT_FIELD(vsk, write_notify_window) <
vsk               215 net/vmw_vsock/vmci_transport_notify_qstate.c 		    PKT_FIELD(vsk, write_notify_min_window)) {
vsk               223 net/vmw_vsock/vmci_transport_notify_qstate.c 			PKT_FIELD(vsk, write_notify_window) =
vsk               224 net/vmw_vsock/vmci_transport_notify_qstate.c 			    PKT_FIELD(vsk, write_notify_min_window);
vsk               260 net/vmw_vsock/vmci_transport_notify_qstate.c 	struct vsock_sock *vsk;
vsk               265 net/vmw_vsock/vmci_transport_notify_qstate.c 	vsk = vsock_sk(sk);
vsk               272 net/vmw_vsock/vmci_transport_notify_qstate.c 			vmci_qpair_consume_free_space(vmci_trans(vsk)->qpair);
vsk               276 net/vmw_vsock/vmci_transport_notify_qstate.c 			PKT_FIELD(vsk, peer_waiting_write) = true;
vsk               309 net/vmw_vsock/vmci_transport_notify_qstate.c 	struct vsock_sock *vsk;
vsk               314 net/vmw_vsock/vmci_transport_notify_qstate.c 	vsk = vsock_sk(sk);
vsk               319 net/vmw_vsock/vmci_transport_notify_qstate.c 		vmci_qpair_produce_buf_ready(vmci_trans(vsk)->qpair) == written;
vsk               321 net/vmw_vsock/vmci_transport_notify_qstate.c 		while (!(vsk->peer_shutdown & RCV_SHUTDOWN) &&
vsk               368 net/vmw_vsock/vmci_transport_notify_qstate.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk               370 net/vmw_vsock/vmci_transport_notify_qstate.c 	PKT_FIELD(vsk, write_notify_window) = vmci_trans(vsk)->consume_size;
vsk               371 net/vmw_vsock/vmci_transport_notify_qstate.c 	if (vmci_trans(vsk)->consume_size <
vsk               372 net/vmw_vsock/vmci_transport_notify_qstate.c 		PKT_FIELD(vsk, write_notify_min_window))
vsk               373 net/vmw_vsock/vmci_transport_notify_qstate.c 		PKT_FIELD(vsk, write_notify_min_window) =
vsk               374 net/vmw_vsock/vmci_transport_notify_qstate.c 			vmci_trans(vsk)->consume_size;
vsk               379 net/vmw_vsock/vmci_transport_notify_qstate.c 	struct vsock_sock *vsk = vsock_sk(sk);
vsk               381 net/vmw_vsock/vmci_transport_notify_qstate.c 	PKT_FIELD(vsk, write_notify_window) = vmci_trans(vsk)->consume_size;
vsk               382 net/vmw_vsock/vmci_transport_notify_qstate.c 	if (vmci_trans(vsk)->consume_size <
vsk               383 net/vmw_vsock/vmci_transport_notify_qstate.c 		PKT_FIELD(vsk, write_notify_min_window))
vsk               384 net/vmw_vsock/vmci_transport_notify_qstate.c 		PKT_FIELD(vsk, write_notify_min_window) =
vsk               385 net/vmw_vsock/vmci_transport_notify_qstate.c 			vmci_trans(vsk)->consume_size;