123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680 |
- /*
- * VMware vSockets Driver
- *
- * Copyright (C) 2009-2013 VMware, Inc. All rights reserved.
- *
- * This program is free software; you can redistribute it and/or modify it
- * under the terms of the GNU General Public License as published by the Free
- * Software Foundation version 2 and no later version.
- *
- * This program is distributed in the hope that it will be useful, but WITHOUT
- * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
- * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
- * more details.
- */
- #include <linux/types.h>
- #include <linux/socket.h>
- #include <linux/stddef.h>
- #include <net/sock.h>
- #include "vmci_transport_notify.h"
- #define PKT_FIELD(vsk, field_name) (vmci_trans(vsk)->notify.pkt.field_name)
- static bool vmci_transport_notify_waiting_write(struct vsock_sock *vsk)
- {
- #if defined(VSOCK_OPTIMIZATION_WAITING_NOTIFY)
- bool retval;
- u64 notify_limit;
- if (!PKT_FIELD(vsk, peer_waiting_write))
- return false;
- #ifdef VSOCK_OPTIMIZATION_FLOW_CONTROL
- /* When the sender blocks, we take that as a sign that the sender is
- * faster than the receiver. To reduce the transmit rate of the sender,
- * we delay the sending of the read notification by decreasing the
- * write_notify_window. The notification is delayed until the number of
- * bytes used in the queue drops below the write_notify_window.
- */
- if (!PKT_FIELD(vsk, peer_waiting_write_detected)) {
- PKT_FIELD(vsk, peer_waiting_write_detected) = true;
- if (PKT_FIELD(vsk, write_notify_window) < PAGE_SIZE) {
- PKT_FIELD(vsk, write_notify_window) =
- PKT_FIELD(vsk, write_notify_min_window);
- } else {
- PKT_FIELD(vsk, write_notify_window) -= PAGE_SIZE;
- if (PKT_FIELD(vsk, write_notify_window) <
- PKT_FIELD(vsk, write_notify_min_window))
- PKT_FIELD(vsk, write_notify_window) =
- PKT_FIELD(vsk, write_notify_min_window);
- }
- }
- notify_limit = vmci_trans(vsk)->consume_size -
- PKT_FIELD(vsk, write_notify_window);
- #else
- notify_limit = 0;
- #endif
- /* For now we ignore the wait information and just see if the free
- * space exceeds the notify limit. Note that improving this function
- * to be more intelligent will not require a protocol change and will
- * retain compatibility between endpoints with mixed versions of this
- * function.
- *
- * The notify_limit is used to delay notifications in the case where
- * flow control is enabled. Below the test is expressed in terms of
- * free space in the queue: if free_space > ConsumeSize -
- * write_notify_window then notify An alternate way of expressing this
- * is to rewrite the expression to use the data ready in the receive
- * queue: if write_notify_window > bufferReady then notify as
- * free_space == ConsumeSize - bufferReady.
- */
- retval = vmci_qpair_consume_free_space(vmci_trans(vsk)->qpair) >
- notify_limit;
- #ifdef VSOCK_OPTIMIZATION_FLOW_CONTROL
- if (retval) {
- /*
- * Once we notify the peer, we reset the detected flag so the
- * next wait will again cause a decrease in the window size.
- */
- PKT_FIELD(vsk, peer_waiting_write_detected) = false;
- }
- #endif
- return retval;
- #else
- return true;
- #endif
- }
- static bool vmci_transport_notify_waiting_read(struct vsock_sock *vsk)
- {
- #if defined(VSOCK_OPTIMIZATION_WAITING_NOTIFY)
- if (!PKT_FIELD(vsk, peer_waiting_read))
- return false;
- /* For now we ignore the wait information and just see if there is any
- * data for our peer to read. Note that improving this function to be
- * more intelligent will not require a protocol change and will retain
- * compatibility between endpoints with mixed versions of this
- * function.
- */
- return vmci_qpair_produce_buf_ready(vmci_trans(vsk)->qpair) > 0;
- #else
- return true;
- #endif
- }
- static void
- vmci_transport_handle_waiting_read(struct sock *sk,
- struct vmci_transport_packet *pkt,
- bool bottom_half,
- struct sockaddr_vm *dst,
- struct sockaddr_vm *src)
- {
- #if defined(VSOCK_OPTIMIZATION_WAITING_NOTIFY)
- struct vsock_sock *vsk;
- vsk = vsock_sk(sk);
- PKT_FIELD(vsk, peer_waiting_read) = true;
- memcpy(&PKT_FIELD(vsk, peer_waiting_read_info), &pkt->u.wait,
- sizeof(PKT_FIELD(vsk, peer_waiting_read_info)));
- if (vmci_transport_notify_waiting_read(vsk)) {
- bool sent;
- if (bottom_half)
- sent = vmci_transport_send_wrote_bh(dst, src) > 0;
- else
- sent = vmci_transport_send_wrote(sk) > 0;
- if (sent)
- PKT_FIELD(vsk, peer_waiting_read) = false;
- }
- #endif
- }
- static void
- vmci_transport_handle_waiting_write(struct sock *sk,
- struct vmci_transport_packet *pkt,
- bool bottom_half,
- struct sockaddr_vm *dst,
- struct sockaddr_vm *src)
- {
- #if defined(VSOCK_OPTIMIZATION_WAITING_NOTIFY)
- struct vsock_sock *vsk;
- vsk = vsock_sk(sk);
- PKT_FIELD(vsk, peer_waiting_write) = true;
- memcpy(&PKT_FIELD(vsk, peer_waiting_write_info), &pkt->u.wait,
- sizeof(PKT_FIELD(vsk, peer_waiting_write_info)));
- if (vmci_transport_notify_waiting_write(vsk)) {
- bool sent;
- if (bottom_half)
- sent = vmci_transport_send_read_bh(dst, src) > 0;
- else
- sent = vmci_transport_send_read(sk) > 0;
- if (sent)
- PKT_FIELD(vsk, peer_waiting_write) = false;
- }
- #endif
- }
- static void
- vmci_transport_handle_read(struct sock *sk,
- struct vmci_transport_packet *pkt,
- bool bottom_half,
- struct sockaddr_vm *dst, struct sockaddr_vm *src)
- {
- #if defined(VSOCK_OPTIMIZATION_WAITING_NOTIFY)
- struct vsock_sock *vsk;
- vsk = vsock_sk(sk);
- PKT_FIELD(vsk, sent_waiting_write) = false;
- #endif
- sk->sk_write_space(sk);
- }
- static bool send_waiting_read(struct sock *sk, u64 room_needed)
- {
- #if defined(VSOCK_OPTIMIZATION_WAITING_NOTIFY)
- struct vsock_sock *vsk;
- struct vmci_transport_waiting_info waiting_info;
- u64 tail;
- u64 head;
- u64 room_left;
- bool ret;
- vsk = vsock_sk(sk);
- if (PKT_FIELD(vsk, sent_waiting_read))
- return true;
- if (PKT_FIELD(vsk, write_notify_window) <
- vmci_trans(vsk)->consume_size)
- PKT_FIELD(vsk, write_notify_window) =
- min(PKT_FIELD(vsk, write_notify_window) + PAGE_SIZE,
- vmci_trans(vsk)->consume_size);
- vmci_qpair_get_consume_indexes(vmci_trans(vsk)->qpair, &tail, &head);
- room_left = vmci_trans(vsk)->consume_size - head;
- if (room_needed >= room_left) {
- waiting_info.offset = room_needed - room_left;
- waiting_info.generation =
- PKT_FIELD(vsk, consume_q_generation) + 1;
- } else {
- waiting_info.offset = head + room_needed;
- waiting_info.generation = PKT_FIELD(vsk, consume_q_generation);
- }
- ret = vmci_transport_send_waiting_read(sk, &waiting_info) > 0;
- if (ret)
- PKT_FIELD(vsk, sent_waiting_read) = true;
- return ret;
- #else
- return true;
- #endif
- }
- static bool send_waiting_write(struct sock *sk, u64 room_needed)
- {
- #if defined(VSOCK_OPTIMIZATION_WAITING_NOTIFY)
- struct vsock_sock *vsk;
- struct vmci_transport_waiting_info waiting_info;
- u64 tail;
- u64 head;
- u64 room_left;
- bool ret;
- vsk = vsock_sk(sk);
- if (PKT_FIELD(vsk, sent_waiting_write))
- return true;
- vmci_qpair_get_produce_indexes(vmci_trans(vsk)->qpair, &tail, &head);
- room_left = vmci_trans(vsk)->produce_size - tail;
- if (room_needed + 1 >= room_left) {
- /* Wraps around to current generation. */
- waiting_info.offset = room_needed + 1 - room_left;
- waiting_info.generation = PKT_FIELD(vsk, produce_q_generation);
- } else {
- waiting_info.offset = tail + room_needed + 1;
- waiting_info.generation =
- PKT_FIELD(vsk, produce_q_generation) - 1;
- }
- ret = vmci_transport_send_waiting_write(sk, &waiting_info) > 0;
- if (ret)
- PKT_FIELD(vsk, sent_waiting_write) = true;
- return ret;
- #else
- return true;
- #endif
- }
- static int vmci_transport_send_read_notification(struct sock *sk)
- {
- struct vsock_sock *vsk;
- bool sent_read;
- unsigned int retries;
- int err;
- vsk = vsock_sk(sk);
- sent_read = false;
- retries = 0;
- err = 0;
- if (vmci_transport_notify_waiting_write(vsk)) {
- /* Notify the peer that we have read, retrying the send on
- * failure up to our maximum value. XXX For now we just log
- * the failure, but later we should schedule a work item to
- * handle the resend until it succeeds. That would require
- * keeping track of work items in the vsk and cleaning them up
- * upon socket close.
- */
- while (!(vsk->peer_shutdown & RCV_SHUTDOWN) &&
- !sent_read &&
- retries < VMCI_TRANSPORT_MAX_DGRAM_RESENDS) {
- err = vmci_transport_send_read(sk);
- if (err >= 0)
- sent_read = true;
- retries++;
- }
- if (retries >= VMCI_TRANSPORT_MAX_DGRAM_RESENDS)
- pr_err("%p unable to send read notify to peer\n", sk);
- else
- #if defined(VSOCK_OPTIMIZATION_WAITING_NOTIFY)
- PKT_FIELD(vsk, peer_waiting_write) = false;
- #endif
- }
- return err;
- }
- static void
- vmci_transport_handle_wrote(struct sock *sk,
- struct vmci_transport_packet *pkt,
- bool bottom_half,
- struct sockaddr_vm *dst, struct sockaddr_vm *src)
- {
- #if defined(VSOCK_OPTIMIZATION_WAITING_NOTIFY)
- struct vsock_sock *vsk = vsock_sk(sk);
- PKT_FIELD(vsk, sent_waiting_read) = false;
- #endif
- sk->sk_data_ready(sk);
- }
- static void vmci_transport_notify_pkt_socket_init(struct sock *sk)
- {
- struct vsock_sock *vsk = vsock_sk(sk);
- PKT_FIELD(vsk, write_notify_window) = PAGE_SIZE;
- PKT_FIELD(vsk, write_notify_min_window) = PAGE_SIZE;
- PKT_FIELD(vsk, peer_waiting_read) = false;
- PKT_FIELD(vsk, peer_waiting_write) = false;
- PKT_FIELD(vsk, peer_waiting_write_detected) = false;
- PKT_FIELD(vsk, sent_waiting_read) = false;
- PKT_FIELD(vsk, sent_waiting_write) = false;
- PKT_FIELD(vsk, produce_q_generation) = 0;
- PKT_FIELD(vsk, consume_q_generation) = 0;
- memset(&PKT_FIELD(vsk, peer_waiting_read_info), 0,
- sizeof(PKT_FIELD(vsk, peer_waiting_read_info)));
- memset(&PKT_FIELD(vsk, peer_waiting_write_info), 0,
- sizeof(PKT_FIELD(vsk, peer_waiting_write_info)));
- }
- static void vmci_transport_notify_pkt_socket_destruct(struct vsock_sock *vsk)
- {
- }
- static int
- vmci_transport_notify_pkt_poll_in(struct sock *sk,
- size_t target, bool *data_ready_now)
- {
- struct vsock_sock *vsk = vsock_sk(sk);
- if (vsock_stream_has_data(vsk)) {
- *data_ready_now = true;
- } else {
- /* We can't read right now because there is nothing in the
- * queue. Ask for notifications when there is something to
- * read.
- */
- if (sk->sk_state == SS_CONNECTED) {
- if (!send_waiting_read(sk, 1))
- return -1;
- }
- *data_ready_now = false;
- }
- return 0;
- }
- static int
- vmci_transport_notify_pkt_poll_out(struct sock *sk,
- size_t target, bool *space_avail_now)
- {
- s64 produce_q_free_space;
- struct vsock_sock *vsk = vsock_sk(sk);
- produce_q_free_space = vsock_stream_has_space(vsk);
- if (produce_q_free_space > 0) {
- *space_avail_now = true;
- return 0;
- } else if (produce_q_free_space == 0) {
- /* This is a connected socket but we can't currently send data.
- * Notify the peer that we are waiting if the queue is full. We
- * only send a waiting write if the queue is full because
- * otherwise we end up in an infinite WAITING_WRITE, READ,
- * WAITING_WRITE, READ, etc. loop. Treat failing to send the
- * notification as a socket error, passing that back through
- * the mask.
- */
- if (!send_waiting_write(sk, 1))
- return -1;
- *space_avail_now = false;
- }
- return 0;
- }
- static int
- vmci_transport_notify_pkt_recv_init(
- struct sock *sk,
- size_t target,
- struct vmci_transport_recv_notify_data *data)
- {
- struct vsock_sock *vsk = vsock_sk(sk);
- #ifdef VSOCK_OPTIMIZATION_WAITING_NOTIFY
- data->consume_head = 0;
- data->produce_tail = 0;
- #ifdef VSOCK_OPTIMIZATION_FLOW_CONTROL
- data->notify_on_block = false;
- if (PKT_FIELD(vsk, write_notify_min_window) < target + 1) {
- PKT_FIELD(vsk, write_notify_min_window) = target + 1;
- if (PKT_FIELD(vsk, write_notify_window) <
- PKT_FIELD(vsk, write_notify_min_window)) {
- /* If the current window is smaller than the new
- * minimal window size, we need to reevaluate whether
- * we need to notify the sender. If the number of ready
- * bytes are smaller than the new window, we need to
- * send a notification to the sender before we block.
- */
- PKT_FIELD(vsk, write_notify_window) =
- PKT_FIELD(vsk, write_notify_min_window);
- data->notify_on_block = true;
- }
- }
- #endif
- #endif
- return 0;
- }
- static int
- vmci_transport_notify_pkt_recv_pre_block(
- struct sock *sk,
- size_t target,
- struct vmci_transport_recv_notify_data *data)
- {
- int err = 0;
- /* Notify our peer that we are waiting for data to read. */
- if (!send_waiting_read(sk, target)) {
- err = -EHOSTUNREACH;
- return err;
- }
- #ifdef VSOCK_OPTIMIZATION_FLOW_CONTROL
- if (data->notify_on_block) {
- err = vmci_transport_send_read_notification(sk);
- if (err < 0)
- return err;
- data->notify_on_block = false;
- }
- #endif
- return err;
- }
- static int
- vmci_transport_notify_pkt_recv_pre_dequeue(
- struct sock *sk,
- size_t target,
- struct vmci_transport_recv_notify_data *data)
- {
- struct vsock_sock *vsk = vsock_sk(sk);
- /* Now consume up to len bytes from the queue. Note that since we have
- * the socket locked we should copy at least ready bytes.
- */
- #if defined(VSOCK_OPTIMIZATION_WAITING_NOTIFY)
- vmci_qpair_get_consume_indexes(vmci_trans(vsk)->qpair,
- &data->produce_tail,
- &data->consume_head);
- #endif
- return 0;
- }
- static int
- vmci_transport_notify_pkt_recv_post_dequeue(
- struct sock *sk,
- size_t target,
- ssize_t copied,
- bool data_read,
- struct vmci_transport_recv_notify_data *data)
- {
- struct vsock_sock *vsk;
- int err;
- vsk = vsock_sk(sk);
- err = 0;
- if (data_read) {
- #if defined(VSOCK_OPTIMIZATION_WAITING_NOTIFY)
- /* Detect a wrap-around to maintain queue generation. Note
- * that this is safe since we hold the socket lock across the
- * two queue pair operations.
- */
- if (copied >=
- vmci_trans(vsk)->consume_size - data->consume_head)
- PKT_FIELD(vsk, consume_q_generation)++;
- #endif
- err = vmci_transport_send_read_notification(sk);
- if (err < 0)
- return err;
- }
- return err;
- }
- static int
- vmci_transport_notify_pkt_send_init(
- struct sock *sk,
- struct vmci_transport_send_notify_data *data)
- {
- #ifdef VSOCK_OPTIMIZATION_WAITING_NOTIFY
- data->consume_head = 0;
- data->produce_tail = 0;
- #endif
- return 0;
- }
- static int
- vmci_transport_notify_pkt_send_pre_block(
- struct sock *sk,
- struct vmci_transport_send_notify_data *data)
- {
- /* Notify our peer that we are waiting for room to write. */
- if (!send_waiting_write(sk, 1))
- return -EHOSTUNREACH;
- return 0;
- }
- static int
- vmci_transport_notify_pkt_send_pre_enqueue(
- struct sock *sk,
- struct vmci_transport_send_notify_data *data)
- {
- struct vsock_sock *vsk = vsock_sk(sk);
- #if defined(VSOCK_OPTIMIZATION_WAITING_NOTIFY)
- vmci_qpair_get_produce_indexes(vmci_trans(vsk)->qpair,
- &data->produce_tail,
- &data->consume_head);
- #endif
- return 0;
- }
- static int
- vmci_transport_notify_pkt_send_post_enqueue(
- struct sock *sk,
- ssize_t written,
- struct vmci_transport_send_notify_data *data)
- {
- int err = 0;
- struct vsock_sock *vsk;
- bool sent_wrote = false;
- int retries = 0;
- vsk = vsock_sk(sk);
- #if defined(VSOCK_OPTIMIZATION_WAITING_NOTIFY)
- /* Detect a wrap-around to maintain queue generation. Note that this
- * is safe since we hold the socket lock across the two queue pair
- * operations.
- */
- if (written >= vmci_trans(vsk)->produce_size - data->produce_tail)
- PKT_FIELD(vsk, produce_q_generation)++;
- #endif
- if (vmci_transport_notify_waiting_read(vsk)) {
- /* Notify the peer that we have written, retrying the send on
- * failure up to our maximum value. See the XXX comment for the
- * corresponding piece of code in StreamRecvmsg() for potential
- * improvements.
- */
- while (!(vsk->peer_shutdown & RCV_SHUTDOWN) &&
- !sent_wrote &&
- retries < VMCI_TRANSPORT_MAX_DGRAM_RESENDS) {
- err = vmci_transport_send_wrote(sk);
- if (err >= 0)
- sent_wrote = true;
- retries++;
- }
- if (retries >= VMCI_TRANSPORT_MAX_DGRAM_RESENDS) {
- pr_err("%p unable to send wrote notify to peer\n", sk);
- return err;
- } else {
- #if defined(VSOCK_OPTIMIZATION_WAITING_NOTIFY)
- PKT_FIELD(vsk, peer_waiting_read) = false;
- #endif
- }
- }
- return err;
- }
- static void
- vmci_transport_notify_pkt_handle_pkt(
- struct sock *sk,
- struct vmci_transport_packet *pkt,
- bool bottom_half,
- struct sockaddr_vm *dst,
- struct sockaddr_vm *src, bool *pkt_processed)
- {
- bool processed = false;
- switch (pkt->type) {
- case VMCI_TRANSPORT_PACKET_TYPE_WROTE:
- vmci_transport_handle_wrote(sk, pkt, bottom_half, dst, src);
- processed = true;
- break;
- case VMCI_TRANSPORT_PACKET_TYPE_READ:
- vmci_transport_handle_read(sk, pkt, bottom_half, dst, src);
- processed = true;
- break;
- case VMCI_TRANSPORT_PACKET_TYPE_WAITING_WRITE:
- vmci_transport_handle_waiting_write(sk, pkt, bottom_half,
- dst, src);
- processed = true;
- break;
- case VMCI_TRANSPORT_PACKET_TYPE_WAITING_READ:
- vmci_transport_handle_waiting_read(sk, pkt, bottom_half,
- dst, src);
- processed = true;
- break;
- }
- if (pkt_processed)
- *pkt_processed = processed;
- }
- static void vmci_transport_notify_pkt_process_request(struct sock *sk)
- {
- struct vsock_sock *vsk = vsock_sk(sk);
- PKT_FIELD(vsk, write_notify_window) = vmci_trans(vsk)->consume_size;
- if (vmci_trans(vsk)->consume_size <
- PKT_FIELD(vsk, write_notify_min_window))
- PKT_FIELD(vsk, write_notify_min_window) =
- vmci_trans(vsk)->consume_size;
- }
- static void vmci_transport_notify_pkt_process_negotiate(struct sock *sk)
- {
- struct vsock_sock *vsk = vsock_sk(sk);
- PKT_FIELD(vsk, write_notify_window) = vmci_trans(vsk)->consume_size;
- if (vmci_trans(vsk)->consume_size <
- PKT_FIELD(vsk, write_notify_min_window))
- PKT_FIELD(vsk, write_notify_min_window) =
- vmci_trans(vsk)->consume_size;
- }
- /* Socket control packet based operations. */
- struct vmci_transport_notify_ops vmci_transport_notify_pkt_ops = {
- vmci_transport_notify_pkt_socket_init,
- vmci_transport_notify_pkt_socket_destruct,
- vmci_transport_notify_pkt_poll_in,
- vmci_transport_notify_pkt_poll_out,
- vmci_transport_notify_pkt_handle_pkt,
- vmci_transport_notify_pkt_recv_init,
- vmci_transport_notify_pkt_recv_pre_block,
- vmci_transport_notify_pkt_recv_pre_dequeue,
- vmci_transport_notify_pkt_recv_post_dequeue,
- vmci_transport_notify_pkt_send_init,
- vmci_transport_notify_pkt_send_pre_block,
- vmci_transport_notify_pkt_send_pre_enqueue,
- vmci_transport_notify_pkt_send_post_enqueue,
- vmci_transport_notify_pkt_process_request,
- vmci_transport_notify_pkt_process_negotiate,
- };
|