vsk 269 drivers/vhost/vsock.c vhost_transport_cancel_pkt(struct vsock_sock *vsk) vsk 280 drivers/vhost/vsock.c vsock = vhost_vsock_get(vsk->remote_addr.svm_cid); vsk 286 drivers/vhost/vsock.c if (pkt->vsk != vsk) vsk 611 drivers/vhost/vsock.c struct vsock_sock *vsk = vsock_sk(sk); vsk 619 drivers/vhost/vsock.c if (vhost_vsock_get(vsk->remote_addr.svm_cid)) vsk 625 drivers/vhost/vsock.c if (vsk->close_work_scheduled) vsk 629 drivers/vhost/vsock.c vsk->peer_shutdown = SHUTDOWN_MASK; vsk 26 include/linux/virtio_vsock.h struct vsock_sock *vsk; vsk 53 include/linux/virtio_vsock.h struct vsock_sock *vsk; vsk 63 include/linux/virtio_vsock.h struct vsock_sock *vsk; vsk 81 include/linux/virtio_vsock.h virtio_transport_stream_dequeue(struct vsock_sock *vsk, vsk 86 include/linux/virtio_vsock.h virtio_transport_dgram_dequeue(struct vsock_sock *vsk, vsk 90 include/linux/virtio_vsock.h s64 virtio_transport_stream_has_data(struct vsock_sock *vsk); vsk 91 include/linux/virtio_vsock.h s64 virtio_transport_stream_has_space(struct vsock_sock *vsk); vsk 93 include/linux/virtio_vsock.h int virtio_transport_do_socket_init(struct vsock_sock *vsk, vsk 95 include/linux/virtio_vsock.h u64 virtio_transport_get_buffer_size(struct vsock_sock *vsk); vsk 96 include/linux/virtio_vsock.h u64 virtio_transport_get_min_buffer_size(struct vsock_sock *vsk); vsk 97 include/linux/virtio_vsock.h u64 virtio_transport_get_max_buffer_size(struct vsock_sock *vsk); vsk 98 include/linux/virtio_vsock.h void virtio_transport_set_buffer_size(struct vsock_sock *vsk, u64 val); vsk 99 include/linux/virtio_vsock.h void virtio_transport_set_min_buffer_size(struct vsock_sock *vsk, u64 val); vsk 102 include/linux/virtio_vsock.h virtio_transport_notify_poll_in(struct vsock_sock *vsk, vsk 106 include/linux/virtio_vsock.h virtio_transport_notify_poll_out(struct vsock_sock *vsk, vsk 110 include/linux/virtio_vsock.h int virtio_transport_notify_recv_init(struct vsock_sock *vsk, vsk 112 include/linux/virtio_vsock.h int virtio_transport_notify_recv_pre_block(struct vsock_sock *vsk, vsk 114 include/linux/virtio_vsock.h int virtio_transport_notify_recv_pre_dequeue(struct vsock_sock *vsk, vsk 116 include/linux/virtio_vsock.h int virtio_transport_notify_recv_post_dequeue(struct vsock_sock *vsk, vsk 119 include/linux/virtio_vsock.h int virtio_transport_notify_send_init(struct vsock_sock *vsk, vsk 121 include/linux/virtio_vsock.h int virtio_transport_notify_send_pre_block(struct vsock_sock *vsk, vsk 123 include/linux/virtio_vsock.h int virtio_transport_notify_send_pre_enqueue(struct vsock_sock *vsk, vsk 125 include/linux/virtio_vsock.h int virtio_transport_notify_send_post_enqueue(struct vsock_sock *vsk, vsk 128 include/linux/virtio_vsock.h u64 virtio_transport_stream_rcvhiwat(struct vsock_sock *vsk); vsk 129 include/linux/virtio_vsock.h bool virtio_transport_stream_is_active(struct vsock_sock *vsk); vsk 131 include/linux/virtio_vsock.h int virtio_transport_dgram_bind(struct vsock_sock *vsk, vsk 135 include/linux/virtio_vsock.h int virtio_transport_connect(struct vsock_sock *vsk); vsk 137 include/linux/virtio_vsock.h int virtio_transport_shutdown(struct vsock_sock *vsk, int mode); vsk 139 include/linux/virtio_vsock.h void virtio_transport_release(struct vsock_sock *vsk); vsk 142 include/linux/virtio_vsock.h virtio_transport_stream_enqueue(struct vsock_sock *vsk, vsk 146 include/linux/virtio_vsock.h virtio_transport_dgram_enqueue(struct vsock_sock *vsk, vsk 151 include/linux/virtio_vsock.h void virtio_transport_destruct(struct vsock_sock *vsk); vsk 71 include/net/af_vsock.h s64 vsock_stream_has_data(struct vsock_sock *vsk); vsk 72 include/net/af_vsock.h s64 vsock_stream_has_space(struct vsock_sock *vsk); vsk 98 include/net/af_vsock.h int (*cancel_pkt)(struct vsock_sock *vsk); vsk 105 include/net/af_vsock.h int (*dgram_dequeue)(struct vsock_sock *vsk, struct msghdr *msg, vsk 173 include/net/af_vsock.h static inline bool __vsock_in_bound_table(struct vsock_sock *vsk) vsk 175 include/net/af_vsock.h return !list_empty(&vsk->bound_table); vsk 179 include/net/af_vsock.h static inline bool __vsock_in_connected_table(struct vsock_sock *vsk) vsk 181 include/net/af_vsock.h return !list_empty(&vsk->connected_table); vsk 188 include/net/af_vsock.h void vsock_insert_connected(struct vsock_sock *vsk); vsk 189 include/net/af_vsock.h void vsock_remove_bound(struct vsock_sock *vsk); vsk 190 include/net/af_vsock.h void vsock_remove_connected(struct vsock_sock *vsk); vsk 194 include/net/af_vsock.h void vsock_remove_sock(struct vsock_sock *vsk); vsk 169 net/vmw_vsock/af_vsock.c #define vsock_connected_sockets_vsk(vsk) \ vsk 170 net/vmw_vsock/af_vsock.c vsock_connected_sockets(&(vsk)->remote_addr, &(vsk)->local_addr) vsk 180 net/vmw_vsock/af_vsock.c static int vsock_auto_bind(struct vsock_sock *vsk) vsk 182 net/vmw_vsock/af_vsock.c struct sock *sk = sk_vsock(vsk); vsk 185 net/vmw_vsock/af_vsock.c if (vsock_addr_bound(&vsk->local_addr)) vsk 204 net/vmw_vsock/af_vsock.c struct vsock_sock *vsk) vsk 206 net/vmw_vsock/af_vsock.c sock_hold(&vsk->sk); vsk 207 net/vmw_vsock/af_vsock.c list_add(&vsk->bound_table, list); vsk 211 net/vmw_vsock/af_vsock.c struct vsock_sock *vsk) vsk 213 net/vmw_vsock/af_vsock.c sock_hold(&vsk->sk); vsk 214 net/vmw_vsock/af_vsock.c list_add(&vsk->connected_table, list); vsk 217 net/vmw_vsock/af_vsock.c static void __vsock_remove_bound(struct vsock_sock *vsk) vsk 219 net/vmw_vsock/af_vsock.c list_del_init(&vsk->bound_table); vsk 220 net/vmw_vsock/af_vsock.c sock_put(&vsk->sk); vsk 223 net/vmw_vsock/af_vsock.c static void __vsock_remove_connected(struct vsock_sock *vsk) vsk 225 net/vmw_vsock/af_vsock.c list_del_init(&vsk->connected_table); vsk 226 net/vmw_vsock/af_vsock.c sock_put(&vsk->sk); vsk 231 net/vmw_vsock/af_vsock.c struct vsock_sock *vsk; vsk 233 net/vmw_vsock/af_vsock.c list_for_each_entry(vsk, vsock_bound_sockets(addr), bound_table) vsk 234 net/vmw_vsock/af_vsock.c if (addr->svm_port == vsk->local_addr.svm_port) vsk 235 net/vmw_vsock/af_vsock.c return sk_vsock(vsk); vsk 243 net/vmw_vsock/af_vsock.c struct vsock_sock *vsk; vsk 245 net/vmw_vsock/af_vsock.c list_for_each_entry(vsk, vsock_connected_sockets(src, dst), vsk 247 net/vmw_vsock/af_vsock.c if (vsock_addr_equals_addr(src, &vsk->remote_addr) && vsk 248 net/vmw_vsock/af_vsock.c dst->svm_port == vsk->local_addr.svm_port) { vsk 249 net/vmw_vsock/af_vsock.c return sk_vsock(vsk); vsk 256 net/vmw_vsock/af_vsock.c static void vsock_insert_unbound(struct vsock_sock *vsk) vsk 259 net/vmw_vsock/af_vsock.c __vsock_insert_bound(vsock_unbound_sockets, vsk); vsk 263 net/vmw_vsock/af_vsock.c void vsock_insert_connected(struct vsock_sock *vsk) vsk 266 net/vmw_vsock/af_vsock.c &vsk->remote_addr, &vsk->local_addr); vsk 269 net/vmw_vsock/af_vsock.c __vsock_insert_connected(list, vsk); vsk 274 net/vmw_vsock/af_vsock.c void vsock_remove_bound(struct vsock_sock *vsk) vsk 277 net/vmw_vsock/af_vsock.c if (__vsock_in_bound_table(vsk)) vsk 278 net/vmw_vsock/af_vsock.c __vsock_remove_bound(vsk); vsk 283 net/vmw_vsock/af_vsock.c void vsock_remove_connected(struct vsock_sock *vsk) vsk 286 net/vmw_vsock/af_vsock.c if (__vsock_in_connected_table(vsk)) vsk 287 net/vmw_vsock/af_vsock.c __vsock_remove_connected(vsk); vsk 323 net/vmw_vsock/af_vsock.c void vsock_remove_sock(struct vsock_sock *vsk) vsk 325 net/vmw_vsock/af_vsock.c vsock_remove_bound(vsk); vsk 326 net/vmw_vsock/af_vsock.c vsock_remove_connected(vsk); vsk 337 net/vmw_vsock/af_vsock.c struct vsock_sock *vsk; vsk 338 net/vmw_vsock/af_vsock.c list_for_each_entry(vsk, &vsock_connected_table[i], vsk 340 net/vmw_vsock/af_vsock.c fn(sk_vsock(vsk)); vsk 409 net/vmw_vsock/af_vsock.c struct vsock_sock *vsk = vsock_sk(sk); vsk 410 net/vmw_vsock/af_vsock.c return list_empty(&vsk->accept_queue); vsk 415 net/vmw_vsock/af_vsock.c struct vsock_sock *vsk = vsock_sk(sk); vsk 416 net/vmw_vsock/af_vsock.c return !list_empty(&vsk->pending_links); vsk 428 net/vmw_vsock/af_vsock.c struct vsock_sock *vsk; vsk 431 net/vmw_vsock/af_vsock.c vsk = container_of(work, struct vsock_sock, pending_work.work); vsk 432 net/vmw_vsock/af_vsock.c sk = sk_vsock(vsk); vsk 433 net/vmw_vsock/af_vsock.c listener = vsk->listener; vsk 443 net/vmw_vsock/af_vsock.c } else if (!vsk->rejected) { vsk 457 net/vmw_vsock/af_vsock.c vsock_remove_connected(vsk); vsk 473 net/vmw_vsock/af_vsock.c static int __vsock_bind_stream(struct vsock_sock *vsk, vsk 516 net/vmw_vsock/af_vsock.c vsock_addr_init(&vsk->local_addr, new_addr.svm_cid, new_addr.svm_port); vsk 522 net/vmw_vsock/af_vsock.c __vsock_remove_bound(vsk); vsk 523 net/vmw_vsock/af_vsock.c __vsock_insert_bound(vsock_bound_sockets(&vsk->local_addr), vsk); vsk 528 net/vmw_vsock/af_vsock.c static int __vsock_bind_dgram(struct vsock_sock *vsk, vsk 531 net/vmw_vsock/af_vsock.c return transport->dgram_bind(vsk, addr); vsk 536 net/vmw_vsock/af_vsock.c struct vsock_sock *vsk = vsock_sk(sk); vsk 541 net/vmw_vsock/af_vsock.c if (vsock_addr_bound(&vsk->local_addr)) vsk 556 net/vmw_vsock/af_vsock.c retval = __vsock_bind_stream(vsk, addr); vsk 561 net/vmw_vsock/af_vsock.c retval = __vsock_bind_dgram(vsk, addr); vsk 583 net/vmw_vsock/af_vsock.c struct vsock_sock *vsk; vsk 598 net/vmw_vsock/af_vsock.c vsk = vsock_sk(sk); vsk 599 net/vmw_vsock/af_vsock.c vsock_addr_init(&vsk->local_addr, VMADDR_CID_ANY, VMADDR_PORT_ANY); vsk 600 net/vmw_vsock/af_vsock.c vsock_addr_init(&vsk->remote_addr, VMADDR_CID_ANY, VMADDR_PORT_ANY); vsk 606 net/vmw_vsock/af_vsock.c INIT_LIST_HEAD(&vsk->bound_table); vsk 607 net/vmw_vsock/af_vsock.c INIT_LIST_HEAD(&vsk->connected_table); vsk 608 net/vmw_vsock/af_vsock.c vsk->listener = NULL; vsk 609 net/vmw_vsock/af_vsock.c INIT_LIST_HEAD(&vsk->pending_links); vsk 610 net/vmw_vsock/af_vsock.c INIT_LIST_HEAD(&vsk->accept_queue); vsk 611 net/vmw_vsock/af_vsock.c vsk->rejected = false; vsk 612 net/vmw_vsock/af_vsock.c vsk->sent_request = false; vsk 613 net/vmw_vsock/af_vsock.c vsk->ignore_connecting_rst = false; vsk 614 net/vmw_vsock/af_vsock.c vsk->peer_shutdown = 0; vsk 615 net/vmw_vsock/af_vsock.c INIT_DELAYED_WORK(&vsk->connect_work, vsock_connect_timeout); vsk 616 net/vmw_vsock/af_vsock.c INIT_DELAYED_WORK(&vsk->pending_work, vsock_pending_work); vsk 620 net/vmw_vsock/af_vsock.c vsk->trusted = psk->trusted; vsk 621 net/vmw_vsock/af_vsock.c vsk->owner = get_cred(psk->owner); vsk 622 net/vmw_vsock/af_vsock.c vsk->connect_timeout = psk->connect_timeout; vsk 624 net/vmw_vsock/af_vsock.c vsk->trusted = capable(CAP_NET_ADMIN); vsk 625 net/vmw_vsock/af_vsock.c vsk->owner = get_current_cred(); vsk 626 net/vmw_vsock/af_vsock.c vsk->connect_timeout = VSOCK_DEFAULT_CONNECT_TIMEOUT; vsk 629 net/vmw_vsock/af_vsock.c if (transport->init(vsk, psk) < 0) { vsk 635 net/vmw_vsock/af_vsock.c vsock_insert_unbound(vsk); vsk 646 net/vmw_vsock/af_vsock.c struct vsock_sock *vsk; vsk 648 net/vmw_vsock/af_vsock.c vsk = vsock_sk(sk); vsk 654 net/vmw_vsock/af_vsock.c transport->release(vsk); vsk 681 net/vmw_vsock/af_vsock.c struct vsock_sock *vsk = vsock_sk(sk); vsk 683 net/vmw_vsock/af_vsock.c transport->destruct(vsk); vsk 688 net/vmw_vsock/af_vsock.c vsock_addr_init(&vsk->local_addr, VMADDR_CID_ANY, VMADDR_PORT_ANY); vsk 689 net/vmw_vsock/af_vsock.c vsock_addr_init(&vsk->remote_addr, VMADDR_CID_ANY, VMADDR_PORT_ANY); vsk 691 net/vmw_vsock/af_vsock.c put_cred(vsk->owner); vsk 705 net/vmw_vsock/af_vsock.c s64 vsock_stream_has_data(struct vsock_sock *vsk) vsk 707 net/vmw_vsock/af_vsock.c return transport->stream_has_data(vsk); vsk 711 net/vmw_vsock/af_vsock.c s64 vsock_stream_has_space(struct vsock_sock *vsk) vsk 713 net/vmw_vsock/af_vsock.c return transport->stream_has_space(vsk); vsk 750 net/vmw_vsock/af_vsock.c struct vsock_sock *vsk; vsk 754 net/vmw_vsock/af_vsock.c vsk = vsock_sk(sk); vsk 764 net/vmw_vsock/af_vsock.c vm_addr = &vsk->remote_addr; vsk 766 net/vmw_vsock/af_vsock.c vm_addr = &vsk->local_addr; vsk 842 net/vmw_vsock/af_vsock.c struct vsock_sock *vsk; vsk 845 net/vmw_vsock/af_vsock.c vsk = vsock_sk(sk); vsk 859 net/vmw_vsock/af_vsock.c (vsk->peer_shutdown & SEND_SHUTDOWN))) { vsk 864 net/vmw_vsock/af_vsock.c vsk->peer_shutdown & SEND_SHUTDOWN) { vsk 892 net/vmw_vsock/af_vsock.c if (transport->stream_is_active(vsk) && vsk 896 net/vmw_vsock/af_vsock.c vsk, 1, &data_ready_now); vsk 911 net/vmw_vsock/af_vsock.c vsk->peer_shutdown & SEND_SHUTDOWN) { vsk 920 net/vmw_vsock/af_vsock.c vsk, 1, &space_avail_now); vsk 955 net/vmw_vsock/af_vsock.c struct vsock_sock *vsk; vsk 964 net/vmw_vsock/af_vsock.c vsk = vsock_sk(sk); vsk 968 net/vmw_vsock/af_vsock.c err = vsock_auto_bind(vsk); vsk 991 net/vmw_vsock/af_vsock.c remote_addr = &vsk->remote_addr; vsk 999 net/vmw_vsock/af_vsock.c if (!vsock_addr_bound(&vsk->remote_addr)) { vsk 1014 net/vmw_vsock/af_vsock.c err = transport->dgram_enqueue(vsk, remote_addr, msg, len); vsk 1026 net/vmw_vsock/af_vsock.c struct vsock_sock *vsk; vsk 1030 net/vmw_vsock/af_vsock.c vsk = vsock_sk(sk); vsk 1035 net/vmw_vsock/af_vsock.c vsock_addr_init(&vsk->remote_addr, VMADDR_CID_ANY, vsk 1045 net/vmw_vsock/af_vsock.c err = vsock_auto_bind(vsk); vsk 1055 net/vmw_vsock/af_vsock.c memcpy(&vsk->remote_addr, remote_addr, sizeof(vsk->remote_addr)); vsk 1090 net/vmw_vsock/af_vsock.c static int vsock_transport_cancel_pkt(struct vsock_sock *vsk) vsk 1095 net/vmw_vsock/af_vsock.c return transport->cancel_pkt(vsk); vsk 1101 net/vmw_vsock/af_vsock.c struct vsock_sock *vsk; vsk 1104 net/vmw_vsock/af_vsock.c vsk = container_of(work, struct vsock_sock, connect_work.work); vsk 1105 net/vmw_vsock/af_vsock.c sk = sk_vsock(vsk); vsk 1117 net/vmw_vsock/af_vsock.c vsock_transport_cancel_pkt(vsk); vsk 1127 net/vmw_vsock/af_vsock.c struct vsock_sock *vsk; vsk 1134 net/vmw_vsock/af_vsock.c vsk = vsock_sk(sk); vsk 1172 net/vmw_vsock/af_vsock.c memcpy(&vsk->remote_addr, remote_addr, vsk 1173 net/vmw_vsock/af_vsock.c sizeof(vsk->remote_addr)); vsk 1175 net/vmw_vsock/af_vsock.c err = vsock_auto_bind(vsk); vsk 1181 net/vmw_vsock/af_vsock.c err = transport->connect(vsk); vsk 1196 net/vmw_vsock/af_vsock.c timeout = vsk->connect_timeout; vsk 1208 net/vmw_vsock/af_vsock.c schedule_delayed_work(&vsk->connect_work, timeout); vsk 1222 net/vmw_vsock/af_vsock.c vsock_transport_cancel_pkt(vsk); vsk 1228 net/vmw_vsock/af_vsock.c vsock_transport_cancel_pkt(vsk); vsk 1336 net/vmw_vsock/af_vsock.c struct vsock_sock *vsk; vsk 1352 net/vmw_vsock/af_vsock.c vsk = vsock_sk(sk); vsk 1354 net/vmw_vsock/af_vsock.c if (!vsock_addr_bound(&vsk->local_addr)) { vsk 1377 net/vmw_vsock/af_vsock.c struct vsock_sock *vsk; vsk 1397 net/vmw_vsock/af_vsock.c vsk = vsock_sk(sk); vsk 1404 net/vmw_vsock/af_vsock.c transport->set_buffer_size(vsk, val); vsk 1409 net/vmw_vsock/af_vsock.c transport->set_max_buffer_size(vsk, val); vsk 1414 net/vmw_vsock/af_vsock.c transport->set_min_buffer_size(vsk, val); vsk 1422 net/vmw_vsock/af_vsock.c vsk->connect_timeout = tv.tv_sec * HZ + vsk 1424 net/vmw_vsock/af_vsock.c if (vsk->connect_timeout == 0) vsk 1425 net/vmw_vsock/af_vsock.c vsk->connect_timeout = vsk 1454 net/vmw_vsock/af_vsock.c struct vsock_sock *vsk; vsk 1477 net/vmw_vsock/af_vsock.c vsk = vsock_sk(sk); vsk 1481 net/vmw_vsock/af_vsock.c val = transport->get_buffer_size(vsk); vsk 1486 net/vmw_vsock/af_vsock.c val = transport->get_max_buffer_size(vsk); vsk 1491 net/vmw_vsock/af_vsock.c val = transport->get_min_buffer_size(vsk); vsk 1497 net/vmw_vsock/af_vsock.c tv.tv_sec = vsk->connect_timeout / HZ; vsk 1499 net/vmw_vsock/af_vsock.c (vsk->connect_timeout - vsk 1521 net/vmw_vsock/af_vsock.c struct vsock_sock *vsk; vsk 1529 net/vmw_vsock/af_vsock.c vsk = vsock_sk(sk); vsk 1546 net/vmw_vsock/af_vsock.c vsk->peer_shutdown & RCV_SHUTDOWN) { vsk 1552 net/vmw_vsock/af_vsock.c !vsock_addr_bound(&vsk->local_addr)) { vsk 1557 net/vmw_vsock/af_vsock.c if (!vsock_addr_bound(&vsk->remote_addr)) { vsk 1565 net/vmw_vsock/af_vsock.c err = transport->notify_send_init(vsk, &send_data); vsk 1573 net/vmw_vsock/af_vsock.c while (vsock_stream_has_space(vsk) == 0 && vsk 1576 net/vmw_vsock/af_vsock.c !(vsk->peer_shutdown & RCV_SHUTDOWN)) { vsk 1585 net/vmw_vsock/af_vsock.c err = transport->notify_send_pre_block(vsk, &send_data); vsk 1614 net/vmw_vsock/af_vsock.c (vsk->peer_shutdown & RCV_SHUTDOWN)) { vsk 1619 net/vmw_vsock/af_vsock.c err = transport->notify_send_pre_enqueue(vsk, &send_data); vsk 1630 net/vmw_vsock/af_vsock.c vsk, msg, vsk 1640 net/vmw_vsock/af_vsock.c vsk, written, &send_data); vsk 1660 net/vmw_vsock/af_vsock.c struct vsock_sock *vsk; vsk 1670 net/vmw_vsock/af_vsock.c vsk = vsock_sk(sk); vsk 1718 net/vmw_vsock/af_vsock.c if (target >= transport->stream_rcvhiwat(vsk)) { vsk 1725 net/vmw_vsock/af_vsock.c err = transport->notify_recv_init(vsk, target, &recv_data); vsk 1734 net/vmw_vsock/af_vsock.c ready = vsock_stream_has_data(vsk); vsk 1739 net/vmw_vsock/af_vsock.c (vsk->peer_shutdown & SEND_SHUTDOWN)) { vsk 1751 net/vmw_vsock/af_vsock.c vsk, target, &recv_data); vsk 1785 net/vmw_vsock/af_vsock.c vsk, target, &recv_data); vsk 1790 net/vmw_vsock/af_vsock.c vsk, msg, vsk 1800 net/vmw_vsock/af_vsock.c vsk, target, read, vsk 17 net/vmw_vsock/diag.c struct vsock_sock *vsk = vsock_sk(sk); vsk 37 net/vmw_vsock/diag.c rep->vdiag_src_cid = vsk->local_addr.svm_cid; vsk 38 net/vmw_vsock/diag.c rep->vdiag_src_port = vsk->local_addr.svm_port; vsk 39 net/vmw_vsock/diag.c rep->vdiag_dst_cid = vsk->remote_addr.svm_cid; vsk 40 net/vmw_vsock/diag.c rep->vdiag_dst_port = vsk->remote_addr.svm_port; vsk 51 net/vmw_vsock/diag.c struct vsock_sock *vsk; vsk 76 net/vmw_vsock/diag.c list_for_each_entry(vsk, head, bound_table) { vsk 77 net/vmw_vsock/diag.c struct sock *sk = sk_vsock(vsk); vsk 106 net/vmw_vsock/diag.c list_for_each_entry(vsk, head, connected_table) { vsk 107 net/vmw_vsock/diag.c struct sock *sk = sk_vsock(vsk); vsk 110 net/vmw_vsock/diag.c if (__vsock_in_bound_table(vsk)) vsk 90 net/vmw_vsock/hyperv_transport.c struct vsock_sock *vsk; vsk 237 net/vmw_vsock/hyperv_transport.c struct vsock_sock *vsk = vsock_sk(sk); vsk 238 net/vmw_vsock/hyperv_transport.c struct hvsock *hvs = vsk->trans; vsk 248 net/vmw_vsock/hyperv_transport.c static void hvs_do_close_lock_held(struct vsock_sock *vsk, vsk 251 net/vmw_vsock/hyperv_transport.c struct sock *sk = sk_vsock(vsk); vsk 254 net/vmw_vsock/hyperv_transport.c vsk->peer_shutdown = SHUTDOWN_MASK; vsk 255 net/vmw_vsock/hyperv_transport.c if (vsock_stream_has_data(vsk) <= 0) vsk 258 net/vmw_vsock/hyperv_transport.c if (vsk->close_work_scheduled && vsk 259 net/vmw_vsock/hyperv_transport.c (!cancel_timeout || cancel_delayed_work(&vsk->close_work))) { vsk 260 net/vmw_vsock/hyperv_transport.c vsk->close_work_scheduled = false; vsk 261 net/vmw_vsock/hyperv_transport.c vsock_remove_sock(vsk); vsk 421 net/vmw_vsock/hyperv_transport.c static int hvs_sock_init(struct vsock_sock *vsk, struct vsock_sock *psk) vsk 424 net/vmw_vsock/hyperv_transport.c struct sock *sk = sk_vsock(vsk); vsk 430 net/vmw_vsock/hyperv_transport.c vsk->trans = hvs; vsk 431 net/vmw_vsock/hyperv_transport.c hvs->vsk = vsk; vsk 437 net/vmw_vsock/hyperv_transport.c static int hvs_connect(struct vsock_sock *vsk) vsk 440 net/vmw_vsock/hyperv_transport.c struct hvsock *h = vsk->trans; vsk 443 net/vmw_vsock/hyperv_transport.c vm.svm_port = vsk->local_addr.svm_port; vsk 447 net/vmw_vsock/hyperv_transport.c host.svm_port = vsk->remote_addr.svm_port; vsk 465 net/vmw_vsock/hyperv_transport.c static int hvs_shutdown(struct vsock_sock *vsk, int mode) vsk 467 net/vmw_vsock/hyperv_transport.c struct sock *sk = sk_vsock(vsk); vsk 473 net/vmw_vsock/hyperv_transport.c hvs_shutdown_lock_held(vsk->trans, mode); vsk 480 net/vmw_vsock/hyperv_transport.c struct vsock_sock *vsk = vsk 482 net/vmw_vsock/hyperv_transport.c struct sock *sk = sk_vsock(vsk); vsk 487 net/vmw_vsock/hyperv_transport.c hvs_do_close_lock_held(vsk, false); vsk 489 net/vmw_vsock/hyperv_transport.c vsk->close_work_scheduled = false; vsk 495 net/vmw_vsock/hyperv_transport.c static bool hvs_close_lock_held(struct vsock_sock *vsk) vsk 497 net/vmw_vsock/hyperv_transport.c struct sock *sk = sk_vsock(vsk); vsk 504 net/vmw_vsock/hyperv_transport.c hvs_shutdown_lock_held(vsk->trans, SHUTDOWN_MASK); vsk 511 net/vmw_vsock/hyperv_transport.c INIT_DELAYED_WORK(&vsk->close_work, hvs_close_timeout); vsk 512 net/vmw_vsock/hyperv_transport.c vsk->close_work_scheduled = true; vsk 513 net/vmw_vsock/hyperv_transport.c schedule_delayed_work(&vsk->close_work, HVS_CLOSE_TIMEOUT); vsk 517 net/vmw_vsock/hyperv_transport.c static void hvs_release(struct vsock_sock *vsk) vsk 519 net/vmw_vsock/hyperv_transport.c struct sock *sk = sk_vsock(vsk); vsk 523 net/vmw_vsock/hyperv_transport.c remove_sock = hvs_close_lock_held(vsk); vsk 526 net/vmw_vsock/hyperv_transport.c vsock_remove_sock(vsk); vsk 529 net/vmw_vsock/hyperv_transport.c static void hvs_destruct(struct vsock_sock *vsk) vsk 531 net/vmw_vsock/hyperv_transport.c struct hvsock *hvs = vsk->trans; vsk 540 net/vmw_vsock/hyperv_transport.c static int hvs_dgram_bind(struct vsock_sock *vsk, struct sockaddr_vm *addr) vsk 545 net/vmw_vsock/hyperv_transport.c static int hvs_dgram_dequeue(struct vsock_sock *vsk, struct msghdr *msg, vsk 551 net/vmw_vsock/hyperv_transport.c static int hvs_dgram_enqueue(struct vsock_sock *vsk, vsk 575 net/vmw_vsock/hyperv_transport.c hvs->vsk->peer_shutdown |= SEND_SHUTDOWN; vsk 583 net/vmw_vsock/hyperv_transport.c static ssize_t hvs_stream_dequeue(struct vsock_sock *vsk, struct msghdr *msg, vsk 586 net/vmw_vsock/hyperv_transport.c struct hvsock *hvs = vsk->trans; vsk 623 net/vmw_vsock/hyperv_transport.c static ssize_t hvs_stream_enqueue(struct vsock_sock *vsk, struct msghdr *msg, vsk 626 net/vmw_vsock/hyperv_transport.c struct hvsock *hvs = vsk->trans; vsk 671 net/vmw_vsock/hyperv_transport.c static s64 hvs_stream_has_data(struct vsock_sock *vsk) vsk 673 net/vmw_vsock/hyperv_transport.c struct hvsock *hvs = vsk->trans; vsk 684 net/vmw_vsock/hyperv_transport.c vsk->peer_shutdown |= SEND_SHUTDOWN; vsk 695 net/vmw_vsock/hyperv_transport.c static s64 hvs_stream_has_space(struct vsock_sock *vsk) vsk 697 net/vmw_vsock/hyperv_transport.c struct hvsock *hvs = vsk->trans; vsk 702 net/vmw_vsock/hyperv_transport.c static u64 hvs_stream_rcvhiwat(struct vsock_sock *vsk) vsk 707 net/vmw_vsock/hyperv_transport.c static bool hvs_stream_is_active(struct vsock_sock *vsk) vsk 709 net/vmw_vsock/hyperv_transport.c struct hvsock *hvs = vsk->trans; vsk 723 net/vmw_vsock/hyperv_transport.c int hvs_notify_poll_in(struct vsock_sock *vsk, size_t target, bool *readable) vsk 725 net/vmw_vsock/hyperv_transport.c struct hvsock *hvs = vsk->trans; vsk 732 net/vmw_vsock/hyperv_transport.c int hvs_notify_poll_out(struct vsock_sock *vsk, size_t target, bool *writable) vsk 734 net/vmw_vsock/hyperv_transport.c *writable = hvs_stream_has_space(vsk) > 0; vsk 740 net/vmw_vsock/hyperv_transport.c int hvs_notify_recv_init(struct vsock_sock *vsk, size_t target, vsk 747 net/vmw_vsock/hyperv_transport.c int hvs_notify_recv_pre_block(struct vsock_sock *vsk, size_t target, vsk 754 net/vmw_vsock/hyperv_transport.c int hvs_notify_recv_pre_dequeue(struct vsock_sock *vsk, size_t target, vsk 761 net/vmw_vsock/hyperv_transport.c int hvs_notify_recv_post_dequeue(struct vsock_sock *vsk, size_t target, vsk 769 net/vmw_vsock/hyperv_transport.c int hvs_notify_send_init(struct vsock_sock *vsk, vsk 776 net/vmw_vsock/hyperv_transport.c int hvs_notify_send_pre_block(struct vsock_sock *vsk, vsk 783 net/vmw_vsock/hyperv_transport.c int hvs_notify_send_pre_enqueue(struct vsock_sock *vsk, vsk 790 net/vmw_vsock/hyperv_transport.c int hvs_notify_send_post_enqueue(struct vsock_sock *vsk, ssize_t written, vsk 796 net/vmw_vsock/hyperv_transport.c static void hvs_set_buffer_size(struct vsock_sock *vsk, u64 val) vsk 801 net/vmw_vsock/hyperv_transport.c static void hvs_set_min_buffer_size(struct vsock_sock *vsk, u64 val) vsk 806 net/vmw_vsock/hyperv_transport.c static void hvs_set_max_buffer_size(struct vsock_sock *vsk, u64 val) vsk 811 net/vmw_vsock/hyperv_transport.c static u64 hvs_get_buffer_size(struct vsock_sock *vsk) vsk 816 net/vmw_vsock/hyperv_transport.c static u64 hvs_get_min_buffer_size(struct vsock_sock *vsk) vsk 821 net/vmw_vsock/hyperv_transport.c static u64 hvs_get_max_buffer_size(struct vsock_sock *vsk) vsk 243 net/vmw_vsock/virtio_transport.c virtio_transport_cancel_pkt(struct vsock_sock *vsk) vsk 259 net/vmw_vsock/virtio_transport.c if (pkt->vsk != vsk) vsk 64 net/vmw_vsock/virtio_transport_common.c pkt->vsk = info->vsk; vsk 163 net/vmw_vsock/virtio_transport_common.c static int virtio_transport_send_pkt_info(struct vsock_sock *vsk, vsk 172 net/vmw_vsock/virtio_transport_common.c src_port = vsk->local_addr.svm_port; vsk 174 net/vmw_vsock/virtio_transport_common.c dst_cid = vsk->remote_addr.svm_cid; vsk 175 net/vmw_vsock/virtio_transport_common.c dst_port = vsk->remote_addr.svm_port; vsk 181 net/vmw_vsock/virtio_transport_common.c vvs = vsk->trans; vsk 257 net/vmw_vsock/virtio_transport_common.c static int virtio_transport_send_credit_update(struct vsock_sock *vsk, vsk 264 net/vmw_vsock/virtio_transport_common.c .vsk = vsk, vsk 267 net/vmw_vsock/virtio_transport_common.c return virtio_transport_send_pkt_info(vsk, &info); vsk 271 net/vmw_vsock/virtio_transport_common.c virtio_transport_stream_do_dequeue(struct vsock_sock *vsk, vsk 275 net/vmw_vsock/virtio_transport_common.c struct virtio_vsock_sock *vvs = vsk->trans; vsk 323 net/vmw_vsock/virtio_transport_common.c virtio_transport_send_credit_update(vsk, vsk 337 net/vmw_vsock/virtio_transport_common.c virtio_transport_stream_dequeue(struct vsock_sock *vsk, vsk 344 net/vmw_vsock/virtio_transport_common.c return virtio_transport_stream_do_dequeue(vsk, msg, len); vsk 349 net/vmw_vsock/virtio_transport_common.c virtio_transport_dgram_dequeue(struct vsock_sock *vsk, vsk 357 net/vmw_vsock/virtio_transport_common.c s64 virtio_transport_stream_has_data(struct vsock_sock *vsk) vsk 359 net/vmw_vsock/virtio_transport_common.c struct virtio_vsock_sock *vvs = vsk->trans; vsk 370 net/vmw_vsock/virtio_transport_common.c static s64 virtio_transport_has_space(struct vsock_sock *vsk) vsk 372 net/vmw_vsock/virtio_transport_common.c struct virtio_vsock_sock *vvs = vsk->trans; vsk 382 net/vmw_vsock/virtio_transport_common.c s64 virtio_transport_stream_has_space(struct vsock_sock *vsk) vsk 384 net/vmw_vsock/virtio_transport_common.c struct virtio_vsock_sock *vvs = vsk->trans; vsk 388 net/vmw_vsock/virtio_transport_common.c bytes = virtio_transport_has_space(vsk); vsk 395 net/vmw_vsock/virtio_transport_common.c int virtio_transport_do_socket_init(struct vsock_sock *vsk, vsk 404 net/vmw_vsock/virtio_transport_common.c vsk->trans = vvs; vsk 405 net/vmw_vsock/virtio_transport_common.c vvs->vsk = vsk; vsk 429 net/vmw_vsock/virtio_transport_common.c u64 virtio_transport_get_buffer_size(struct vsock_sock *vsk) vsk 431 net/vmw_vsock/virtio_transport_common.c struct virtio_vsock_sock *vvs = vsk->trans; vsk 437 net/vmw_vsock/virtio_transport_common.c u64 virtio_transport_get_min_buffer_size(struct vsock_sock *vsk) vsk 439 net/vmw_vsock/virtio_transport_common.c struct virtio_vsock_sock *vvs = vsk->trans; vsk 445 net/vmw_vsock/virtio_transport_common.c u64 virtio_transport_get_max_buffer_size(struct vsock_sock *vsk) vsk 447 net/vmw_vsock/virtio_transport_common.c struct virtio_vsock_sock *vvs = vsk->trans; vsk 453 net/vmw_vsock/virtio_transport_common.c void virtio_transport_set_buffer_size(struct vsock_sock *vsk, u64 val) vsk 455 net/vmw_vsock/virtio_transport_common.c struct virtio_vsock_sock *vvs = vsk->trans; vsk 466 net/vmw_vsock/virtio_transport_common.c virtio_transport_send_credit_update(vsk, VIRTIO_VSOCK_TYPE_STREAM, vsk 471 net/vmw_vsock/virtio_transport_common.c void virtio_transport_set_min_buffer_size(struct vsock_sock *vsk, u64 val) vsk 473 net/vmw_vsock/virtio_transport_common.c struct virtio_vsock_sock *vvs = vsk->trans; vsk 483 net/vmw_vsock/virtio_transport_common.c void virtio_transport_set_max_buffer_size(struct vsock_sock *vsk, u64 val) vsk 485 net/vmw_vsock/virtio_transport_common.c struct virtio_vsock_sock *vvs = vsk->trans; vsk 496 net/vmw_vsock/virtio_transport_common.c virtio_transport_notify_poll_in(struct vsock_sock *vsk, vsk 500 net/vmw_vsock/virtio_transport_common.c if (vsock_stream_has_data(vsk)) vsk 510 net/vmw_vsock/virtio_transport_common.c virtio_transport_notify_poll_out(struct vsock_sock *vsk, vsk 516 net/vmw_vsock/virtio_transport_common.c free_space = vsock_stream_has_space(vsk); vsk 526 net/vmw_vsock/virtio_transport_common.c int virtio_transport_notify_recv_init(struct vsock_sock *vsk, vsk 533 net/vmw_vsock/virtio_transport_common.c int virtio_transport_notify_recv_pre_block(struct vsock_sock *vsk, vsk 540 net/vmw_vsock/virtio_transport_common.c int virtio_transport_notify_recv_pre_dequeue(struct vsock_sock *vsk, vsk 547 net/vmw_vsock/virtio_transport_common.c int virtio_transport_notify_recv_post_dequeue(struct vsock_sock *vsk, vsk 555 net/vmw_vsock/virtio_transport_common.c int virtio_transport_notify_send_init(struct vsock_sock *vsk, vsk 562 net/vmw_vsock/virtio_transport_common.c int virtio_transport_notify_send_pre_block(struct vsock_sock *vsk, vsk 569 net/vmw_vsock/virtio_transport_common.c int virtio_transport_notify_send_pre_enqueue(struct vsock_sock *vsk, vsk 576 net/vmw_vsock/virtio_transport_common.c int virtio_transport_notify_send_post_enqueue(struct vsock_sock *vsk, vsk 583 net/vmw_vsock/virtio_transport_common.c u64 virtio_transport_stream_rcvhiwat(struct vsock_sock *vsk) vsk 585 net/vmw_vsock/virtio_transport_common.c struct virtio_vsock_sock *vvs = vsk->trans; vsk 591 net/vmw_vsock/virtio_transport_common.c bool virtio_transport_stream_is_active(struct vsock_sock *vsk) vsk 603 net/vmw_vsock/virtio_transport_common.c int virtio_transport_dgram_bind(struct vsock_sock *vsk, vsk 616 net/vmw_vsock/virtio_transport_common.c int virtio_transport_connect(struct vsock_sock *vsk) vsk 621 net/vmw_vsock/virtio_transport_common.c .vsk = vsk, vsk 624 net/vmw_vsock/virtio_transport_common.c return virtio_transport_send_pkt_info(vsk, &info); vsk 628 net/vmw_vsock/virtio_transport_common.c int virtio_transport_shutdown(struct vsock_sock *vsk, int mode) vsk 637 net/vmw_vsock/virtio_transport_common.c .vsk = vsk, vsk 640 net/vmw_vsock/virtio_transport_common.c return virtio_transport_send_pkt_info(vsk, &info); vsk 645 net/vmw_vsock/virtio_transport_common.c virtio_transport_dgram_enqueue(struct vsock_sock *vsk, vsk 655 net/vmw_vsock/virtio_transport_common.c virtio_transport_stream_enqueue(struct vsock_sock *vsk, vsk 664 net/vmw_vsock/virtio_transport_common.c .vsk = vsk, vsk 667 net/vmw_vsock/virtio_transport_common.c return virtio_transport_send_pkt_info(vsk, &info); vsk 671 net/vmw_vsock/virtio_transport_common.c void virtio_transport_destruct(struct vsock_sock *vsk) vsk 673 net/vmw_vsock/virtio_transport_common.c struct virtio_vsock_sock *vvs = vsk->trans; vsk 679 net/vmw_vsock/virtio_transport_common.c static int virtio_transport_reset(struct vsock_sock *vsk, vsk 686 net/vmw_vsock/virtio_transport_common.c .vsk = vsk, vsk 693 net/vmw_vsock/virtio_transport_common.c return virtio_transport_send_pkt_info(vsk, &info); vsk 747 net/vmw_vsock/virtio_transport_common.c static void virtio_transport_do_close(struct vsock_sock *vsk, vsk 750 net/vmw_vsock/virtio_transport_common.c struct sock *sk = sk_vsock(vsk); vsk 753 net/vmw_vsock/virtio_transport_common.c vsk->peer_shutdown = SHUTDOWN_MASK; vsk 754 net/vmw_vsock/virtio_transport_common.c if (vsock_stream_has_data(vsk) <= 0) vsk 758 net/vmw_vsock/virtio_transport_common.c if (vsk->close_work_scheduled && vsk 759 net/vmw_vsock/virtio_transport_common.c (!cancel_timeout || cancel_delayed_work(&vsk->close_work))) { vsk 760 net/vmw_vsock/virtio_transport_common.c vsk->close_work_scheduled = false; vsk 762 net/vmw_vsock/virtio_transport_common.c vsock_remove_sock(vsk); vsk 771 net/vmw_vsock/virtio_transport_common.c struct vsock_sock *vsk = vsk 773 net/vmw_vsock/virtio_transport_common.c struct sock *sk = sk_vsock(vsk); vsk 779 net/vmw_vsock/virtio_transport_common.c (void)virtio_transport_reset(vsk, NULL); vsk 781 net/vmw_vsock/virtio_transport_common.c virtio_transport_do_close(vsk, false); vsk 784 net/vmw_vsock/virtio_transport_common.c vsk->close_work_scheduled = false; vsk 791 net/vmw_vsock/virtio_transport_common.c static bool virtio_transport_close(struct vsock_sock *vsk) vsk 793 net/vmw_vsock/virtio_transport_common.c struct sock *sk = &vsk->sk; vsk 800 net/vmw_vsock/virtio_transport_common.c if ((vsk->peer_shutdown & SHUTDOWN_MASK) == SHUTDOWN_MASK) { vsk 801 net/vmw_vsock/virtio_transport_common.c (void)virtio_transport_reset(vsk, NULL); vsk 806 net/vmw_vsock/virtio_transport_common.c (void)virtio_transport_shutdown(vsk, SHUTDOWN_MASK); vsk 816 net/vmw_vsock/virtio_transport_common.c INIT_DELAYED_WORK(&vsk->close_work, vsk 818 net/vmw_vsock/virtio_transport_common.c vsk->close_work_scheduled = true; vsk 819 net/vmw_vsock/virtio_transport_common.c schedule_delayed_work(&vsk->close_work, VSOCK_CLOSE_TIMEOUT); vsk 823 net/vmw_vsock/virtio_transport_common.c void virtio_transport_release(struct vsock_sock *vsk) vsk 825 net/vmw_vsock/virtio_transport_common.c struct virtio_vsock_sock *vvs = vsk->trans; vsk 827 net/vmw_vsock/virtio_transport_common.c struct sock *sk = &vsk->sk; vsk 832 net/vmw_vsock/virtio_transport_common.c remove_sock = virtio_transport_close(vsk); vsk 841 net/vmw_vsock/virtio_transport_common.c vsock_remove_sock(vsk); vsk 849 net/vmw_vsock/virtio_transport_common.c struct vsock_sock *vsk = vsock_sk(sk); vsk 857 net/vmw_vsock/virtio_transport_common.c vsock_insert_connected(vsk); vsk 874 net/vmw_vsock/virtio_transport_common.c virtio_transport_reset(vsk, pkt); vsk 882 net/vmw_vsock/virtio_transport_common.c virtio_transport_recv_enqueue(struct vsock_sock *vsk, vsk 885 net/vmw_vsock/virtio_transport_common.c struct virtio_vsock_sock *vvs = vsk->trans; vsk 933 net/vmw_vsock/virtio_transport_common.c struct vsock_sock *vsk = vsock_sk(sk); vsk 938 net/vmw_vsock/virtio_transport_common.c virtio_transport_recv_enqueue(vsk, pkt); vsk 946 net/vmw_vsock/virtio_transport_common.c vsk->peer_shutdown |= RCV_SHUTDOWN; vsk 948 net/vmw_vsock/virtio_transport_common.c vsk->peer_shutdown |= SEND_SHUTDOWN; vsk 949 net/vmw_vsock/virtio_transport_common.c if (vsk->peer_shutdown == SHUTDOWN_MASK && vsk 950 net/vmw_vsock/virtio_transport_common.c vsock_stream_has_data(vsk) <= 0 && vsk 952 net/vmw_vsock/virtio_transport_common.c (void)virtio_transport_reset(vsk, NULL); vsk 954 net/vmw_vsock/virtio_transport_common.c virtio_transport_do_close(vsk, true); vsk 960 net/vmw_vsock/virtio_transport_common.c virtio_transport_do_close(vsk, true); vsk 975 net/vmw_vsock/virtio_transport_common.c struct vsock_sock *vsk = vsock_sk(sk); vsk 978 net/vmw_vsock/virtio_transport_common.c virtio_transport_do_close(vsk, true); vsk 982 net/vmw_vsock/virtio_transport_common.c virtio_transport_send_response(struct vsock_sock *vsk, vsk 991 net/vmw_vsock/virtio_transport_common.c .vsk = vsk, vsk 994 net/vmw_vsock/virtio_transport_common.c return virtio_transport_send_pkt_info(vsk, &info); vsk 1001 net/vmw_vsock/virtio_transport_common.c struct vsock_sock *vsk = vsock_sk(sk); vsk 1006 net/vmw_vsock/virtio_transport_common.c virtio_transport_reset(vsk, pkt); vsk 1011 net/vmw_vsock/virtio_transport_common.c virtio_transport_reset(vsk, pkt); vsk 1018 net/vmw_vsock/virtio_transport_common.c virtio_transport_reset(vsk, pkt); vsk 1047 net/vmw_vsock/virtio_transport_common.c struct vsock_sock *vsk = vsock_sk(sk); vsk 1048 net/vmw_vsock/virtio_transport_common.c struct virtio_vsock_sock *vvs = vsk->trans; vsk 1055 net/vmw_vsock/virtio_transport_common.c space_available = virtio_transport_has_space(vsk); vsk 1066 net/vmw_vsock/virtio_transport_common.c struct vsock_sock *vsk; vsk 1101 net/vmw_vsock/virtio_transport_common.c vsk = vsock_sk(sk); vsk 1108 net/vmw_vsock/virtio_transport_common.c vsk->local_addr.svm_cid = dst.svm_cid; vsk 292 net/vmw_vsock/vmci_transport.c struct vsock_sock *vsk; vsk 294 net/vmw_vsock/vmci_transport.c vsk = vsock_sk(sk); vsk 296 net/vmw_vsock/vmci_transport.c if (!vsock_addr_bound(&vsk->local_addr)) vsk 299 net/vmw_vsock/vmci_transport.c if (!vsock_addr_bound(&vsk->remote_addr)) vsk 302 net/vmw_vsock/vmci_transport.c return vmci_transport_alloc_send_control_pkt(&vsk->local_addr, vsk 303 net/vmw_vsock/vmci_transport.c &vsk->remote_addr, vsk 325 net/vmw_vsock/vmci_transport.c struct vsock_sock *vsk; vsk 330 net/vmw_vsock/vmci_transport.c vsk = vsock_sk(sk); vsk 332 net/vmw_vsock/vmci_transport.c if (!vsock_addr_bound(&vsk->local_addr)) vsk 335 net/vmw_vsock/vmci_transport.c if (vsock_addr_bound(&vsk->remote_addr)) { vsk 336 net/vmw_vsock/vmci_transport.c dst_ptr = &vsk->remote_addr; vsk 342 net/vmw_vsock/vmci_transport.c return vmci_transport_alloc_send_control_pkt(&vsk->local_addr, dst_ptr, vsk 456 net/vmw_vsock/vmci_transport.c static int vmci_transport_shutdown(struct vsock_sock *vsk, int mode) vsk 459 net/vmw_vsock/vmci_transport.c &vsk->sk, vsk 623 net/vmw_vsock/vmci_transport.c struct vsock_sock *vsk; vsk 636 net/vmw_vsock/vmci_transport.c vsk = vsock_sk(sk); vsk 637 net/vmw_vsock/vmci_transport.c if (!vmci_transport_allow_dgram(vsk, dg->src.context)) vsk 684 net/vmw_vsock/vmci_transport.c struct vsock_sock *vsk; vsk 753 net/vmw_vsock/vmci_transport.c vsk = vsock_sk(sk); vsk 754 net/vmw_vsock/vmci_transport.c if (!vmci_transport_allow_dgram(vsk, pkt->dg.src.context)) { vsk 768 net/vmw_vsock/vmci_transport.c vsk->local_addr.svm_cid = dst.svm_cid; vsk 771 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->notify_ops->handle_notify_pkt( vsk 812 net/vmw_vsock/vmci_transport.c struct vsock_sock *vsk; vsk 814 net/vmw_vsock/vmci_transport.c vsk = vsock_sk(sk); vsk 815 net/vmw_vsock/vmci_transport.c if (!vmci_handle_is_invalid(vmci_trans(vsk)->qp_handle)) { vsk 821 net/vmw_vsock/vmci_transport.c vsk->peer_shutdown = SHUTDOWN_MASK; vsk 830 net/vmw_vsock/vmci_transport.c if (vsk->local_addr.svm_cid == VMADDR_CID_HOST || vsk 831 net/vmw_vsock/vmci_transport.c vsock_stream_has_data(vsk) <= 0) { vsk 1279 net/vmw_vsock/vmci_transport.c struct vsock_sock *vsk; vsk 1283 net/vmw_vsock/vmci_transport.c vsk = vsock_sk(sk); vsk 1289 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->qp_handle)) { vsk 1302 net/vmw_vsock/vmci_transport.c vsock_insert_connected(vsk); vsk 1309 net/vmw_vsock/vmci_transport.c || pkt->dg.src.context != vsk->remote_addr.svm_cid vsk 1310 net/vmw_vsock/vmci_transport.c || pkt->src_port != vsk->remote_addr.svm_port vsk 1311 net/vmw_vsock/vmci_transport.c || !vmci_handle_is_invalid(vmci_trans(vsk)->qp_handle) vsk 1312 net/vmw_vsock/vmci_transport.c || vmci_trans(vsk)->qpair vsk 1313 net/vmw_vsock/vmci_transport.c || vmci_trans(vsk)->produce_size != 0 vsk 1314 net/vmw_vsock/vmci_transport.c || vmci_trans(vsk)->consume_size != 0 vsk 1315 net/vmw_vsock/vmci_transport.c || vmci_trans(vsk)->detach_sub_id != VMCI_INVALID_ID) { vsk 1347 net/vmw_vsock/vmci_transport.c if (vsk->ignore_connecting_rst) { vsk 1348 net/vmw_vsock/vmci_transport.c vsk->ignore_connecting_rst = false; vsk 1379 net/vmw_vsock/vmci_transport.c struct vsock_sock *vsk; vsk 1389 net/vmw_vsock/vmci_transport.c vsk = vsock_sk(sk); vsk 1396 net/vmw_vsock/vmci_transport.c vsk->sent_request = false; vsk 1397 net/vmw_vsock/vmci_transport.c vsk->ignore_connecting_rst = false; vsk 1400 net/vmw_vsock/vmci_transport.c if (pkt->u.size < vmci_trans(vsk)->queue_pair_min_size || vsk 1401 net/vmw_vsock/vmci_transport.c pkt->u.size > vmci_trans(vsk)->queue_pair_max_size) { vsk 1408 net/vmw_vsock/vmci_transport.c if (vsk->local_addr.svm_cid == VMADDR_CID_ANY) vsk 1409 net/vmw_vsock/vmci_transport.c vsk->local_addr.svm_cid = pkt->dg.dst.context; vsk 1443 net/vmw_vsock/vmci_transport.c vmci_trans(vsk), &detach_sub_id); vsk 1451 net/vmw_vsock/vmci_transport.c is_local = vsk->remote_addr.svm_cid == vsk->local_addr.svm_cid; vsk 1458 net/vmw_vsock/vmci_transport.c vsk->remote_addr.svm_cid, vsk 1461 net/vmw_vsock/vmci_transport.c vsk, vsk 1462 net/vmw_vsock/vmci_transport.c vsk-> vsk 1473 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->qp_handle = handle; vsk 1474 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->qpair = qpair; vsk 1476 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->produce_size = vmci_trans(vsk)->consume_size = vsk 1479 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->detach_sub_id = detach_sub_id; vsk 1481 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->notify_ops->process_negotiate(sk); vsk 1500 net/vmw_vsock/vmci_transport.c struct vsock_sock *vsk = vsock_sk(sk); vsk 1502 net/vmw_vsock/vmci_transport.c if (vsk->sent_request) { vsk 1503 net/vmw_vsock/vmci_transport.c vsk->sent_request = false; vsk 1504 net/vmw_vsock/vmci_transport.c vsk->ignore_connecting_rst = true; vsk 1507 net/vmw_vsock/vmci_transport.c sk, vmci_trans(vsk)->queue_pair_size); vsk 1521 net/vmw_vsock/vmci_transport.c struct vsock_sock *vsk; vsk 1535 net/vmw_vsock/vmci_transport.c vsk = vsock_sk(sk); vsk 1537 net/vmw_vsock/vmci_transport.c vsk->peer_shutdown |= pkt->u.mode; vsk 1543 net/vmw_vsock/vmci_transport.c vsk = vsock_sk(sk); vsk 1554 net/vmw_vsock/vmci_transport.c vsk->peer_shutdown = SHUTDOWN_MASK; vsk 1555 net/vmw_vsock/vmci_transport.c if (vsock_stream_has_data(vsk) <= 0) vsk 1562 net/vmw_vsock/vmci_transport.c vsk = vsock_sk(sk); vsk 1563 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->notify_ops->handle_notify_pkt( vsk 1575 net/vmw_vsock/vmci_transport.c static int vmci_transport_socket_init(struct vsock_sock *vsk, vsk 1578 net/vmw_vsock/vmci_transport.c vsk->trans = kmalloc(sizeof(struct vmci_transport), GFP_KERNEL); vsk 1579 net/vmw_vsock/vmci_transport.c if (!vsk->trans) vsk 1582 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->dg_handle = VMCI_INVALID_HANDLE; vsk 1583 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->qp_handle = VMCI_INVALID_HANDLE; vsk 1584 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->qpair = NULL; vsk 1585 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->produce_size = vmci_trans(vsk)->consume_size = 0; vsk 1586 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->detach_sub_id = VMCI_INVALID_ID; vsk 1587 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->notify_ops = NULL; vsk 1588 net/vmw_vsock/vmci_transport.c INIT_LIST_HEAD(&vmci_trans(vsk)->elem); vsk 1589 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->sk = &vsk->sk; vsk 1590 net/vmw_vsock/vmci_transport.c spin_lock_init(&vmci_trans(vsk)->lock); vsk 1592 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->queue_pair_size = vsk 1594 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->queue_pair_min_size = vsk 1596 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->queue_pair_max_size = vsk 1599 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->queue_pair_size = vsk 1601 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->queue_pair_min_size = vsk 1603 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->queue_pair_max_size = vsk 1644 net/vmw_vsock/vmci_transport.c static void vmci_transport_destruct(struct vsock_sock *vsk) vsk 1647 net/vmw_vsock/vmci_transport.c if (!vmci_trans(vsk)) vsk 1653 net/vmw_vsock/vmci_transport.c spin_lock_bh(&vmci_trans(vsk)->lock); vsk 1654 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->sk = NULL; vsk 1655 net/vmw_vsock/vmci_transport.c spin_unlock_bh(&vmci_trans(vsk)->lock); vsk 1657 net/vmw_vsock/vmci_transport.c if (vmci_trans(vsk)->notify_ops) vsk 1658 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->notify_ops->socket_destruct(vsk); vsk 1661 net/vmw_vsock/vmci_transport.c list_add(&vmci_trans(vsk)->elem, &vmci_transport_cleanup_list); vsk 1665 net/vmw_vsock/vmci_transport.c vsk->trans = NULL; vsk 1668 net/vmw_vsock/vmci_transport.c static void vmci_transport_release(struct vsock_sock *vsk) vsk 1670 net/vmw_vsock/vmci_transport.c vsock_remove_sock(vsk); vsk 1672 net/vmw_vsock/vmci_transport.c if (!vmci_handle_is_invalid(vmci_trans(vsk)->dg_handle)) { vsk 1673 net/vmw_vsock/vmci_transport.c vmci_datagram_destroy_handle(vmci_trans(vsk)->dg_handle); vsk 1674 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->dg_handle = VMCI_INVALID_HANDLE; vsk 1678 net/vmw_vsock/vmci_transport.c static int vmci_transport_dgram_bind(struct vsock_sock *vsk, vsk 1699 net/vmw_vsock/vmci_transport.c &vsk->sk, vsk 1700 net/vmw_vsock/vmci_transport.c &vmci_trans(vsk)->dg_handle); vsk 1703 net/vmw_vsock/vmci_transport.c vsock_addr_init(&vsk->local_addr, addr->svm_cid, vsk 1704 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->dg_handle.resource); vsk 1710 net/vmw_vsock/vmci_transport.c struct vsock_sock *vsk, vsk 1721 net/vmw_vsock/vmci_transport.c if (!vmci_transport_allow_dgram(vsk, remote_addr->svm_cid)) vsk 1733 net/vmw_vsock/vmci_transport.c dg->src = vmci_make_handle(vsk->local_addr.svm_cid, vsk 1734 net/vmw_vsock/vmci_transport.c vsk->local_addr.svm_port); vsk 1745 net/vmw_vsock/vmci_transport.c static int vmci_transport_dgram_dequeue(struct vsock_sock *vsk, vsk 1762 net/vmw_vsock/vmci_transport.c skb = skb_recv_datagram(&vsk->sk, flags, noblock, &err); vsk 1797 net/vmw_vsock/vmci_transport.c skb_free_datagram(&vsk->sk, skb); vsk 1813 net/vmw_vsock/vmci_transport.c static int vmci_transport_connect(struct vsock_sock *vsk) vsk 1817 net/vmw_vsock/vmci_transport.c struct sock *sk = &vsk->sk; vsk 1822 net/vmw_vsock/vmci_transport.c sk, vmci_trans(vsk)->queue_pair_size); vsk 1831 net/vmw_vsock/vmci_transport.c sk, vmci_trans(vsk)->queue_pair_size, vsk 1838 net/vmw_vsock/vmci_transport.c vsk->sent_request = true; vsk 1845 net/vmw_vsock/vmci_transport.c struct vsock_sock *vsk, vsk 1851 net/vmw_vsock/vmci_transport.c return vmci_qpair_peekv(vmci_trans(vsk)->qpair, msg, len, 0); vsk 1853 net/vmw_vsock/vmci_transport.c return vmci_qpair_dequev(vmci_trans(vsk)->qpair, msg, len, 0); vsk 1857 net/vmw_vsock/vmci_transport.c struct vsock_sock *vsk, vsk 1861 net/vmw_vsock/vmci_transport.c return vmci_qpair_enquev(vmci_trans(vsk)->qpair, msg, len, 0); vsk 1864 net/vmw_vsock/vmci_transport.c static s64 vmci_transport_stream_has_data(struct vsock_sock *vsk) vsk 1866 net/vmw_vsock/vmci_transport.c return vmci_qpair_consume_buf_ready(vmci_trans(vsk)->qpair); vsk 1869 net/vmw_vsock/vmci_transport.c static s64 vmci_transport_stream_has_space(struct vsock_sock *vsk) vsk 1871 net/vmw_vsock/vmci_transport.c return vmci_qpair_produce_free_space(vmci_trans(vsk)->qpair); vsk 1874 net/vmw_vsock/vmci_transport.c static u64 vmci_transport_stream_rcvhiwat(struct vsock_sock *vsk) vsk 1876 net/vmw_vsock/vmci_transport.c return vmci_trans(vsk)->consume_size; vsk 1879 net/vmw_vsock/vmci_transport.c static bool vmci_transport_stream_is_active(struct vsock_sock *vsk) vsk 1881 net/vmw_vsock/vmci_transport.c return !vmci_handle_is_invalid(vmci_trans(vsk)->qp_handle); vsk 1884 net/vmw_vsock/vmci_transport.c static u64 vmci_transport_get_buffer_size(struct vsock_sock *vsk) vsk 1886 net/vmw_vsock/vmci_transport.c return vmci_trans(vsk)->queue_pair_size; vsk 1889 net/vmw_vsock/vmci_transport.c static u64 vmci_transport_get_min_buffer_size(struct vsock_sock *vsk) vsk 1891 net/vmw_vsock/vmci_transport.c return vmci_trans(vsk)->queue_pair_min_size; vsk 1894 net/vmw_vsock/vmci_transport.c static u64 vmci_transport_get_max_buffer_size(struct vsock_sock *vsk) vsk 1896 net/vmw_vsock/vmci_transport.c return vmci_trans(vsk)->queue_pair_max_size; vsk 1899 net/vmw_vsock/vmci_transport.c static void vmci_transport_set_buffer_size(struct vsock_sock *vsk, u64 val) vsk 1901 net/vmw_vsock/vmci_transport.c if (val < vmci_trans(vsk)->queue_pair_min_size) vsk 1902 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->queue_pair_min_size = val; vsk 1903 net/vmw_vsock/vmci_transport.c if (val > vmci_trans(vsk)->queue_pair_max_size) vsk 1904 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->queue_pair_max_size = val; vsk 1905 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->queue_pair_size = val; vsk 1908 net/vmw_vsock/vmci_transport.c static void vmci_transport_set_min_buffer_size(struct vsock_sock *vsk, vsk 1911 net/vmw_vsock/vmci_transport.c if (val > vmci_trans(vsk)->queue_pair_size) vsk 1912 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->queue_pair_size = val; vsk 1913 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->queue_pair_min_size = val; vsk 1916 net/vmw_vsock/vmci_transport.c static void vmci_transport_set_max_buffer_size(struct vsock_sock *vsk, vsk 1919 net/vmw_vsock/vmci_transport.c if (val < vmci_trans(vsk)->queue_pair_size) vsk 1920 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->queue_pair_size = val; vsk 1921 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->queue_pair_max_size = val; vsk 1925 net/vmw_vsock/vmci_transport.c struct vsock_sock *vsk, vsk 1929 net/vmw_vsock/vmci_transport.c return vmci_trans(vsk)->notify_ops->poll_in( vsk 1930 net/vmw_vsock/vmci_transport.c &vsk->sk, target, data_ready_now); vsk 1934 net/vmw_vsock/vmci_transport.c struct vsock_sock *vsk, vsk 1938 net/vmw_vsock/vmci_transport.c return vmci_trans(vsk)->notify_ops->poll_out( vsk 1939 net/vmw_vsock/vmci_transport.c &vsk->sk, target, space_available_now); vsk 1943 net/vmw_vsock/vmci_transport.c struct vsock_sock *vsk, vsk 1947 net/vmw_vsock/vmci_transport.c return vmci_trans(vsk)->notify_ops->recv_init( vsk 1948 net/vmw_vsock/vmci_transport.c &vsk->sk, target, vsk 1953 net/vmw_vsock/vmci_transport.c struct vsock_sock *vsk, vsk 1957 net/vmw_vsock/vmci_transport.c return vmci_trans(vsk)->notify_ops->recv_pre_block( vsk 1958 net/vmw_vsock/vmci_transport.c &vsk->sk, target, vsk 1963 net/vmw_vsock/vmci_transport.c struct vsock_sock *vsk, vsk 1967 net/vmw_vsock/vmci_transport.c return vmci_trans(vsk)->notify_ops->recv_pre_dequeue( vsk 1968 net/vmw_vsock/vmci_transport.c &vsk->sk, target, vsk 1973 net/vmw_vsock/vmci_transport.c struct vsock_sock *vsk, vsk 1979 net/vmw_vsock/vmci_transport.c return vmci_trans(vsk)->notify_ops->recv_post_dequeue( vsk 1980 net/vmw_vsock/vmci_transport.c &vsk->sk, target, copied, data_read, vsk 1985 net/vmw_vsock/vmci_transport.c struct vsock_sock *vsk, vsk 1988 net/vmw_vsock/vmci_transport.c return vmci_trans(vsk)->notify_ops->send_init( vsk 1989 net/vmw_vsock/vmci_transport.c &vsk->sk, vsk 1994 net/vmw_vsock/vmci_transport.c struct vsock_sock *vsk, vsk 1997 net/vmw_vsock/vmci_transport.c return vmci_trans(vsk)->notify_ops->send_pre_block( vsk 1998 net/vmw_vsock/vmci_transport.c &vsk->sk, vsk 2003 net/vmw_vsock/vmci_transport.c struct vsock_sock *vsk, vsk 2006 net/vmw_vsock/vmci_transport.c return vmci_trans(vsk)->notify_ops->send_pre_enqueue( vsk 2007 net/vmw_vsock/vmci_transport.c &vsk->sk, vsk 2012 net/vmw_vsock/vmci_transport.c struct vsock_sock *vsk, vsk 2016 net/vmw_vsock/vmci_transport.c return vmci_trans(vsk)->notify_ops->send_post_enqueue( vsk 2017 net/vmw_vsock/vmci_transport.c &vsk->sk, written, vsk 2040 net/vmw_vsock/vmci_transport.c struct vsock_sock *vsk = vsock_sk(sk); vsk 2047 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->notify_ops = &vmci_transport_notify_pkt_ops; vsk 2053 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->notify_ops = vsk 2062 net/vmw_vsock/vmci_transport.c vmci_trans(vsk)->notify_ops->socket_init(sk); vsk 15 net/vmw_vsock/vmci_transport_notify.c #define PKT_FIELD(vsk, field_name) (vmci_trans(vsk)->notify.pkt.field_name) vsk 17 net/vmw_vsock/vmci_transport_notify.c static bool vmci_transport_notify_waiting_write(struct vsock_sock *vsk) vsk 23 net/vmw_vsock/vmci_transport_notify.c if (!PKT_FIELD(vsk, peer_waiting_write)) vsk 34 net/vmw_vsock/vmci_transport_notify.c if (!PKT_FIELD(vsk, peer_waiting_write_detected)) { vsk 35 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, peer_waiting_write_detected) = true; vsk 36 net/vmw_vsock/vmci_transport_notify.c if (PKT_FIELD(vsk, write_notify_window) < PAGE_SIZE) { vsk 37 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, write_notify_window) = vsk 38 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, write_notify_min_window); vsk 40 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, write_notify_window) -= PAGE_SIZE; vsk 41 net/vmw_vsock/vmci_transport_notify.c if (PKT_FIELD(vsk, write_notify_window) < vsk 42 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, write_notify_min_window)) vsk 43 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, write_notify_window) = vsk 44 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, write_notify_min_window); vsk 48 net/vmw_vsock/vmci_transport_notify.c notify_limit = vmci_trans(vsk)->consume_size - vsk 49 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, write_notify_window); vsk 68 net/vmw_vsock/vmci_transport_notify.c retval = vmci_qpair_consume_free_space(vmci_trans(vsk)->qpair) > vsk 77 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, peer_waiting_write_detected) = false; vsk 86 net/vmw_vsock/vmci_transport_notify.c static bool vmci_transport_notify_waiting_read(struct vsock_sock *vsk) vsk 89 net/vmw_vsock/vmci_transport_notify.c if (!PKT_FIELD(vsk, peer_waiting_read)) vsk 98 net/vmw_vsock/vmci_transport_notify.c return vmci_qpair_produce_buf_ready(vmci_trans(vsk)->qpair) > 0; vsk 112 net/vmw_vsock/vmci_transport_notify.c struct vsock_sock *vsk; vsk 114 net/vmw_vsock/vmci_transport_notify.c vsk = vsock_sk(sk); vsk 116 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, peer_waiting_read) = true; vsk 117 net/vmw_vsock/vmci_transport_notify.c memcpy(&PKT_FIELD(vsk, peer_waiting_read_info), &pkt->u.wait, vsk 118 net/vmw_vsock/vmci_transport_notify.c sizeof(PKT_FIELD(vsk, peer_waiting_read_info))); vsk 120 net/vmw_vsock/vmci_transport_notify.c if (vmci_transport_notify_waiting_read(vsk)) { vsk 129 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, peer_waiting_read) = false; vsk 142 net/vmw_vsock/vmci_transport_notify.c struct vsock_sock *vsk; vsk 144 net/vmw_vsock/vmci_transport_notify.c vsk = vsock_sk(sk); vsk 146 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, peer_waiting_write) = true; vsk 147 net/vmw_vsock/vmci_transport_notify.c memcpy(&PKT_FIELD(vsk, peer_waiting_write_info), &pkt->u.wait, vsk 148 net/vmw_vsock/vmci_transport_notify.c sizeof(PKT_FIELD(vsk, peer_waiting_write_info))); vsk 150 net/vmw_vsock/vmci_transport_notify.c if (vmci_transport_notify_waiting_write(vsk)) { vsk 159 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, peer_waiting_write) = false; vsk 171 net/vmw_vsock/vmci_transport_notify.c struct vsock_sock *vsk; vsk 173 net/vmw_vsock/vmci_transport_notify.c vsk = vsock_sk(sk); vsk 174 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, sent_waiting_write) = false; vsk 183 net/vmw_vsock/vmci_transport_notify.c struct vsock_sock *vsk; vsk 190 net/vmw_vsock/vmci_transport_notify.c vsk = vsock_sk(sk); vsk 192 net/vmw_vsock/vmci_transport_notify.c if (PKT_FIELD(vsk, sent_waiting_read)) vsk 195 net/vmw_vsock/vmci_transport_notify.c if (PKT_FIELD(vsk, write_notify_window) < vsk 196 net/vmw_vsock/vmci_transport_notify.c vmci_trans(vsk)->consume_size) vsk 197 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, write_notify_window) = vsk 198 net/vmw_vsock/vmci_transport_notify.c min(PKT_FIELD(vsk, write_notify_window) + PAGE_SIZE, vsk 199 net/vmw_vsock/vmci_transport_notify.c vmci_trans(vsk)->consume_size); vsk 201 net/vmw_vsock/vmci_transport_notify.c vmci_qpair_get_consume_indexes(vmci_trans(vsk)->qpair, &tail, &head); vsk 202 net/vmw_vsock/vmci_transport_notify.c room_left = vmci_trans(vsk)->consume_size - head; vsk 206 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, consume_q_generation) + 1; vsk 209 net/vmw_vsock/vmci_transport_notify.c waiting_info.generation = PKT_FIELD(vsk, consume_q_generation); vsk 214 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, sent_waiting_read) = true; vsk 225 net/vmw_vsock/vmci_transport_notify.c struct vsock_sock *vsk; vsk 232 net/vmw_vsock/vmci_transport_notify.c vsk = vsock_sk(sk); vsk 234 net/vmw_vsock/vmci_transport_notify.c if (PKT_FIELD(vsk, sent_waiting_write)) vsk 237 net/vmw_vsock/vmci_transport_notify.c vmci_qpair_get_produce_indexes(vmci_trans(vsk)->qpair, &tail, &head); vsk 238 net/vmw_vsock/vmci_transport_notify.c room_left = vmci_trans(vsk)->produce_size - tail; vsk 242 net/vmw_vsock/vmci_transport_notify.c waiting_info.generation = PKT_FIELD(vsk, produce_q_generation); vsk 246 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, produce_q_generation) - 1; vsk 251 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, sent_waiting_write) = true; vsk 261 net/vmw_vsock/vmci_transport_notify.c struct vsock_sock *vsk; vsk 266 net/vmw_vsock/vmci_transport_notify.c vsk = vsock_sk(sk); vsk 271 net/vmw_vsock/vmci_transport_notify.c if (vmci_transport_notify_waiting_write(vsk)) { vsk 279 net/vmw_vsock/vmci_transport_notify.c while (!(vsk->peer_shutdown & RCV_SHUTDOWN) && vsk 293 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, peer_waiting_write) = false; vsk 307 net/vmw_vsock/vmci_transport_notify.c struct vsock_sock *vsk = vsock_sk(sk); vsk 308 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, sent_waiting_read) = false; vsk 315 net/vmw_vsock/vmci_transport_notify.c struct vsock_sock *vsk = vsock_sk(sk); vsk 317 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, write_notify_window) = PAGE_SIZE; vsk 318 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, write_notify_min_window) = PAGE_SIZE; vsk 319 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, peer_waiting_read) = false; vsk 320 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, peer_waiting_write) = false; vsk 321 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, peer_waiting_write_detected) = false; vsk 322 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, sent_waiting_read) = false; vsk 323 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, sent_waiting_write) = false; vsk 324 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, produce_q_generation) = 0; vsk 325 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, consume_q_generation) = 0; vsk 327 net/vmw_vsock/vmci_transport_notify.c memset(&PKT_FIELD(vsk, peer_waiting_read_info), 0, vsk 328 net/vmw_vsock/vmci_transport_notify.c sizeof(PKT_FIELD(vsk, peer_waiting_read_info))); vsk 329 net/vmw_vsock/vmci_transport_notify.c memset(&PKT_FIELD(vsk, peer_waiting_write_info), 0, vsk 330 net/vmw_vsock/vmci_transport_notify.c sizeof(PKT_FIELD(vsk, peer_waiting_write_info))); vsk 333 net/vmw_vsock/vmci_transport_notify.c static void vmci_transport_notify_pkt_socket_destruct(struct vsock_sock *vsk) vsk 341 net/vmw_vsock/vmci_transport_notify.c struct vsock_sock *vsk = vsock_sk(sk); vsk 343 net/vmw_vsock/vmci_transport_notify.c if (vsock_stream_has_data(vsk)) { vsk 366 net/vmw_vsock/vmci_transport_notify.c struct vsock_sock *vsk = vsock_sk(sk); vsk 368 net/vmw_vsock/vmci_transport_notify.c produce_q_free_space = vsock_stream_has_space(vsk); vsk 396 net/vmw_vsock/vmci_transport_notify.c struct vsock_sock *vsk = vsock_sk(sk); vsk 404 net/vmw_vsock/vmci_transport_notify.c if (PKT_FIELD(vsk, write_notify_min_window) < target + 1) { vsk 405 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, write_notify_min_window) = target + 1; vsk 406 net/vmw_vsock/vmci_transport_notify.c if (PKT_FIELD(vsk, write_notify_window) < vsk 407 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, write_notify_min_window)) { vsk 415 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, write_notify_window) = vsk 416 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, write_notify_min_window); vsk 458 net/vmw_vsock/vmci_transport_notify.c struct vsock_sock *vsk = vsock_sk(sk); vsk 464 net/vmw_vsock/vmci_transport_notify.c vmci_qpair_get_consume_indexes(vmci_trans(vsk)->qpair, vsk 480 net/vmw_vsock/vmci_transport_notify.c struct vsock_sock *vsk; vsk 483 net/vmw_vsock/vmci_transport_notify.c vsk = vsock_sk(sk); vsk 493 net/vmw_vsock/vmci_transport_notify.c vmci_trans(vsk)->consume_size - data->consume_head) vsk 494 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, consume_q_generation)++; vsk 535 net/vmw_vsock/vmci_transport_notify.c struct vsock_sock *vsk = vsock_sk(sk); vsk 538 net/vmw_vsock/vmci_transport_notify.c vmci_qpair_get_produce_indexes(vmci_trans(vsk)->qpair, vsk 553 net/vmw_vsock/vmci_transport_notify.c struct vsock_sock *vsk; vsk 557 net/vmw_vsock/vmci_transport_notify.c vsk = vsock_sk(sk); vsk 564 net/vmw_vsock/vmci_transport_notify.c if (written >= vmci_trans(vsk)->produce_size - data->produce_tail) vsk 565 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, produce_q_generation)++; vsk 569 net/vmw_vsock/vmci_transport_notify.c if (vmci_transport_notify_waiting_read(vsk)) { vsk 575 net/vmw_vsock/vmci_transport_notify.c while (!(vsk->peer_shutdown & RCV_SHUTDOWN) && vsk 590 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, peer_waiting_read) = false; vsk 635 net/vmw_vsock/vmci_transport_notify.c struct vsock_sock *vsk = vsock_sk(sk); vsk 637 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, write_notify_window) = vmci_trans(vsk)->consume_size; vsk 638 net/vmw_vsock/vmci_transport_notify.c if (vmci_trans(vsk)->consume_size < vsk 639 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, write_notify_min_window)) vsk 640 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, write_notify_min_window) = vsk 641 net/vmw_vsock/vmci_transport_notify.c vmci_trans(vsk)->consume_size; vsk 646 net/vmw_vsock/vmci_transport_notify.c struct vsock_sock *vsk = vsock_sk(sk); vsk 648 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, write_notify_window) = vmci_trans(vsk)->consume_size; vsk 649 net/vmw_vsock/vmci_transport_notify.c if (vmci_trans(vsk)->consume_size < vsk 650 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, write_notify_min_window)) vsk 651 net/vmw_vsock/vmci_transport_notify.c PKT_FIELD(vsk, write_notify_min_window) = vsk 652 net/vmw_vsock/vmci_transport_notify.c vmci_trans(vsk)->consume_size; vsk 41 net/vmw_vsock/vmci_transport_notify.h void (*socket_destruct) (struct vsock_sock *vsk); vsk 15 net/vmw_vsock/vmci_transport_notify_qstate.c #define PKT_FIELD(vsk, field_name) \ vsk 16 net/vmw_vsock/vmci_transport_notify_qstate.c (vmci_trans(vsk)->notify.pkt_q_state.field_name) vsk 18 net/vmw_vsock/vmci_transport_notify_qstate.c static bool vmci_transport_notify_waiting_write(struct vsock_sock *vsk) vsk 23 net/vmw_vsock/vmci_transport_notify_qstate.c if (!PKT_FIELD(vsk, peer_waiting_write)) vsk 33 net/vmw_vsock/vmci_transport_notify_qstate.c if (!PKT_FIELD(vsk, peer_waiting_write_detected)) { vsk 34 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, peer_waiting_write_detected) = true; vsk 35 net/vmw_vsock/vmci_transport_notify_qstate.c if (PKT_FIELD(vsk, write_notify_window) < PAGE_SIZE) { vsk 36 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, write_notify_window) = vsk 37 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, write_notify_min_window); vsk 39 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, write_notify_window) -= PAGE_SIZE; vsk 40 net/vmw_vsock/vmci_transport_notify_qstate.c if (PKT_FIELD(vsk, write_notify_window) < vsk 41 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, write_notify_min_window)) vsk 42 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, write_notify_window) = vsk 43 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, write_notify_min_window); vsk 47 net/vmw_vsock/vmci_transport_notify_qstate.c notify_limit = vmci_trans(vsk)->consume_size - vsk 48 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, write_notify_window); vsk 59 net/vmw_vsock/vmci_transport_notify_qstate.c retval = vmci_qpair_consume_free_space(vmci_trans(vsk)->qpair) > vsk 67 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, peer_waiting_write_detected) = false; vsk 92 net/vmw_vsock/vmci_transport_notify_qstate.c struct vsock_sock *vsk = vsock_sk(sk); vsk 94 net/vmw_vsock/vmci_transport_notify_qstate.c if (PKT_FIELD(vsk, write_notify_window) < vmci_trans(vsk)->consume_size) vsk 95 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, write_notify_window) = vsk 96 net/vmw_vsock/vmci_transport_notify_qstate.c min(PKT_FIELD(vsk, write_notify_window) + PAGE_SIZE, vsk 97 net/vmw_vsock/vmci_transport_notify_qstate.c vmci_trans(vsk)->consume_size); vsk 102 net/vmw_vsock/vmci_transport_notify_qstate.c struct vsock_sock *vsk; vsk 107 net/vmw_vsock/vmci_transport_notify_qstate.c vsk = vsock_sk(sk); vsk 112 net/vmw_vsock/vmci_transport_notify_qstate.c if (vmci_transport_notify_waiting_write(vsk)) { vsk 120 net/vmw_vsock/vmci_transport_notify_qstate.c while (!(vsk->peer_shutdown & RCV_SHUTDOWN) && vsk 134 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, peer_waiting_write) = false; vsk 142 net/vmw_vsock/vmci_transport_notify_qstate.c struct vsock_sock *vsk = vsock_sk(sk); vsk 144 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, write_notify_window) = PAGE_SIZE; vsk 145 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, write_notify_min_window) = PAGE_SIZE; vsk 146 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, peer_waiting_write) = false; vsk 147 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, peer_waiting_write_detected) = false; vsk 150 net/vmw_vsock/vmci_transport_notify_qstate.c static void vmci_transport_notify_pkt_socket_destruct(struct vsock_sock *vsk) vsk 152 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, write_notify_window) = PAGE_SIZE; vsk 153 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, write_notify_min_window) = PAGE_SIZE; vsk 154 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, peer_waiting_write) = false; vsk 155 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, peer_waiting_write_detected) = false; vsk 162 net/vmw_vsock/vmci_transport_notify_qstate.c struct vsock_sock *vsk = vsock_sk(sk); vsk 164 net/vmw_vsock/vmci_transport_notify_qstate.c if (vsock_stream_has_data(vsk)) { vsk 184 net/vmw_vsock/vmci_transport_notify_qstate.c struct vsock_sock *vsk = vsock_sk(sk); vsk 186 net/vmw_vsock/vmci_transport_notify_qstate.c produce_q_free_space = vsock_stream_has_space(vsk); vsk 206 net/vmw_vsock/vmci_transport_notify_qstate.c struct vsock_sock *vsk = vsock_sk(sk); vsk 212 net/vmw_vsock/vmci_transport_notify_qstate.c if (PKT_FIELD(vsk, write_notify_min_window) < target + 1) { vsk 213 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, write_notify_min_window) = target + 1; vsk 214 net/vmw_vsock/vmci_transport_notify_qstate.c if (PKT_FIELD(vsk, write_notify_window) < vsk 215 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, write_notify_min_window)) { vsk 223 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, write_notify_window) = vsk 224 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, write_notify_min_window); vsk 260 net/vmw_vsock/vmci_transport_notify_qstate.c struct vsock_sock *vsk; vsk 265 net/vmw_vsock/vmci_transport_notify_qstate.c vsk = vsock_sk(sk); vsk 272 net/vmw_vsock/vmci_transport_notify_qstate.c vmci_qpair_consume_free_space(vmci_trans(vsk)->qpair); vsk 276 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, peer_waiting_write) = true; vsk 309 net/vmw_vsock/vmci_transport_notify_qstate.c struct vsock_sock *vsk; vsk 314 net/vmw_vsock/vmci_transport_notify_qstate.c vsk = vsock_sk(sk); vsk 319 net/vmw_vsock/vmci_transport_notify_qstate.c vmci_qpair_produce_buf_ready(vmci_trans(vsk)->qpair) == written; vsk 321 net/vmw_vsock/vmci_transport_notify_qstate.c while (!(vsk->peer_shutdown & RCV_SHUTDOWN) && vsk 368 net/vmw_vsock/vmci_transport_notify_qstate.c struct vsock_sock *vsk = vsock_sk(sk); vsk 370 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, write_notify_window) = vmci_trans(vsk)->consume_size; vsk 371 net/vmw_vsock/vmci_transport_notify_qstate.c if (vmci_trans(vsk)->consume_size < vsk 372 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, write_notify_min_window)) vsk 373 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, write_notify_min_window) = vsk 374 net/vmw_vsock/vmci_transport_notify_qstate.c vmci_trans(vsk)->consume_size; vsk 379 net/vmw_vsock/vmci_transport_notify_qstate.c struct vsock_sock *vsk = vsock_sk(sk); vsk 381 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, write_notify_window) = vmci_trans(vsk)->consume_size; vsk 382 net/vmw_vsock/vmci_transport_notify_qstate.c if (vmci_trans(vsk)->consume_size < vsk 383 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, write_notify_min_window)) vsk 384 net/vmw_vsock/vmci_transport_notify_qstate.c PKT_FIELD(vsk, write_notify_min_window) = vsk 385 net/vmw_vsock/vmci_transport_notify_qstate.c vmci_trans(vsk)->consume_size;