linux/include/xen/interface/io/netif.h

940 lines
34 KiB
C

/******************************************************************************
* xen_netif.h
*
* Unified network-device I/O interface for Xen guest OSes.
*
* Permission is hereby granted, free of charge, to any person obtaining a copy
* of this software and associated documentation files (the "Software"), to
* deal in the Software without restriction, including without limitation the
* rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
* sell copies of the Software, and to permit persons to whom the Software is
* furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be included in
* all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
* AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
* DEALINGS IN THE SOFTWARE.
*
* Copyright (c) 2003-2004, Keir Fraser
*/
#ifndef __XEN_PUBLIC_IO_XEN_NETIF_H__
#define __XEN_PUBLIC_IO_XEN_NETIF_H__
#include "ring.h"
#include "../grant_table.h"
/*
* Older implementation of Xen network frontend / backend has an
* implicit dependency on the MAX_SKB_FRAGS as the maximum number of
* ring slots a skb can use. Netfront / netback may not work as
* expected when frontend and backend have different MAX_SKB_FRAGS.
*
* A better approach is to add mechanism for netfront / netback to
* negotiate this value. However we cannot fix all possible
* frontends, so we need to define a value which states the minimum
* slots backend must support.
*
* The minimum value derives from older Linux kernel's MAX_SKB_FRAGS
* (18), which is proved to work with most frontends. Any new backend
* which doesn't negotiate with frontend should expect frontend to
* send a valid packet using slots up to this value.
*/
#define XEN_NETIF_NR_SLOTS_MIN 18
/*
* Notifications after enqueuing any type of message should be conditional on
* the appropriate req_event or rsp_event field in the shared ring.
* If the client sends notification for rx requests then it should specify
* feature 'feature-rx-notify' via xenbus. Otherwise the backend will assume
* that it cannot safely queue packets (as it may not be kicked to send them).
*/
/*
* "feature-split-event-channels" is introduced to separate guest TX
* and RX notification. Backend either doesn't support this feature or
* advertises it via xenstore as 0 (disabled) or 1 (enabled).
*
* To make use of this feature, frontend should allocate two event
* channels for TX and RX, advertise them to backend as
* "event-channel-tx" and "event-channel-rx" respectively. If frontend
* doesn't want to use this feature, it just writes "event-channel"
* node as before.
*/
/*
* Multiple transmit and receive queues:
* If supported, the backend will write the key "multi-queue-max-queues" to
* the directory for that vif, and set its value to the maximum supported
* number of queues.
* Frontends that are aware of this feature and wish to use it can write the
* key "multi-queue-num-queues", set to the number they wish to use, which
* must be greater than zero, and no more than the value reported by the backend
* in "multi-queue-max-queues".
*
* Queues replicate the shared rings and event channels.
* "feature-split-event-channels" may optionally be used when using
* multiple queues, but is not mandatory.
*
* Each queue consists of one shared ring pair, i.e. there must be the same
* number of tx and rx rings.
*
* For frontends requesting just one queue, the usual event-channel and
* ring-ref keys are written as before, simplifying the backend processing
* to avoid distinguishing between a frontend that doesn't understand the
* multi-queue feature, and one that does, but requested only one queue.
*
* Frontends requesting two or more queues must not write the toplevel
* event-channel (or event-channel-{tx,rx}) and {tx,rx}-ring-ref keys,
* instead writing those keys under sub-keys having the name "queue-N" where
* N is the integer ID of the queue for which those keys belong. Queues
* are indexed from zero. For example, a frontend with two queues and split
* event channels must write the following set of queue-related keys:
*
* /local/domain/1/device/vif/0/multi-queue-num-queues = "2"
* /local/domain/1/device/vif/0/queue-0 = ""
* /local/domain/1/device/vif/0/queue-0/tx-ring-ref = "<ring-ref-tx0>"
* /local/domain/1/device/vif/0/queue-0/rx-ring-ref = "<ring-ref-rx0>"
* /local/domain/1/device/vif/0/queue-0/event-channel-tx = "<evtchn-tx0>"
* /local/domain/1/device/vif/0/queue-0/event-channel-rx = "<evtchn-rx0>"
* /local/domain/1/device/vif/0/queue-1 = ""
* /local/domain/1/device/vif/0/queue-1/tx-ring-ref = "<ring-ref-tx1>"
* /local/domain/1/device/vif/0/queue-1/rx-ring-ref = "<ring-ref-rx1"
* /local/domain/1/device/vif/0/queue-1/event-channel-tx = "<evtchn-tx1>"
* /local/domain/1/device/vif/0/queue-1/event-channel-rx = "<evtchn-rx1>"
*
* If there is any inconsistency in the XenStore data, the backend may
* choose not to connect any queues, instead treating the request as an
* error. This includes scenarios where more (or fewer) queues were
* requested than the frontend provided details for.
*
* Mapping of packets to queues is considered to be a function of the
* transmitting system (backend or frontend) and is not negotiated
* between the two. Guests are free to transmit packets on any queue
* they choose, provided it has been set up correctly. Guests must be
* prepared to receive packets on any queue they have requested be set up.
*/
/*
* "feature-no-csum-offload" should be used to turn IPv4 TCP/UDP checksum
* offload off or on. If it is missing then the feature is assumed to be on.
* "feature-ipv6-csum-offload" should be used to turn IPv6 TCP/UDP checksum
* offload on or off. If it is missing then the feature is assumed to be off.
*/
/*
* "feature-gso-tcpv4" and "feature-gso-tcpv6" advertise the capability to
* handle large TCP packets (in IPv4 or IPv6 form respectively). Neither
* frontends nor backends are assumed to be capable unless the flags are
* present.
*/
/*
* "feature-multicast-control" and "feature-dynamic-multicast-control"
* advertise the capability to filter ethernet multicast packets in the
* backend. If the frontend wishes to take advantage of this feature then
* it may set "request-multicast-control". If the backend only advertises
* "feature-multicast-control" then "request-multicast-control" must be set
* before the frontend moves into the connected state. The backend will
* sample the value on this state transition and any subsequent change in
* value will have no effect. However, if the backend also advertises
* "feature-dynamic-multicast-control" then "request-multicast-control"
* may be set by the frontend at any time. In this case, the backend will
* watch the value and re-sample on watch events.
*
* If the sampled value of "request-multicast-control" is set then the
* backend transmit side should no longer flood multicast packets to the
* frontend, it should instead drop any multicast packet that does not
* match in a filter list.
* The list is amended by the frontend by sending dummy transmit requests
* containing XEN_NETIF_EXTRA_TYPE_MCAST_{ADD,DEL} extra-info fragments as
* specified below.
* Note that the filter list may be amended even if the sampled value of
* "request-multicast-control" is not set, however the filter should only
* be applied if it is set.
*/
/*
* Control ring
* ============
*
* Some features, such as hashing (detailed below), require a
* significant amount of out-of-band data to be passed from frontend to
* backend. Use of xenstore is not suitable for large quantities of data
* because of quota limitations and so a dedicated 'control ring' is used.
* The ability of the backend to use a control ring is advertised by
* setting:
*
* /local/domain/X/backend/<domid>/<vif>/feature-ctrl-ring = "1"
*
* The frontend provides a control ring to the backend by setting:
*
* /local/domain/<domid>/device/vif/<vif>/ctrl-ring-ref = <gref>
* /local/domain/<domid>/device/vif/<vif>/event-channel-ctrl = <port>
*
* where <gref> is the grant reference of the shared page used to
* implement the control ring and <port> is an event channel to be used
* as a mailbox interrupt. These keys must be set before the frontend
* moves into the connected state.
*
* The control ring uses a fixed request/response message size and is
* balanced (i.e. one request to one response), so operationally it is much
* the same as a transmit or receive ring.
* Note that there is no requirement that responses are issued in the same
* order as requests.
*/
/*
* Hash types
* ==========
*
* For the purposes of the definitions below, 'Packet[]' is an array of
* octets containing an IP packet without options, 'Array[X..Y]' means a
* sub-array of 'Array' containing bytes X thru Y inclusive, and '+' is
* used to indicate concatenation of arrays.
*/
/*
* A hash calculated over an IP version 4 header as follows:
*
* Buffer[0..8] = Packet[12..15] (source address) +
* Packet[16..19] (destination address)
*
* Result = Hash(Buffer, 8)
*/
#define _XEN_NETIF_CTRL_HASH_TYPE_IPV4 0
#define XEN_NETIF_CTRL_HASH_TYPE_IPV4 \
(1 << _XEN_NETIF_CTRL_HASH_TYPE_IPV4)
/*
* A hash calculated over an IP version 4 header and TCP header as
* follows:
*
* Buffer[0..12] = Packet[12..15] (source address) +
* Packet[16..19] (destination address) +
* Packet[20..21] (source port) +
* Packet[22..23] (destination port)
*
* Result = Hash(Buffer, 12)
*/
#define _XEN_NETIF_CTRL_HASH_TYPE_IPV4_TCP 1
#define XEN_NETIF_CTRL_HASH_TYPE_IPV4_TCP \
(1 << _XEN_NETIF_CTRL_HASH_TYPE_IPV4_TCP)
/*
* A hash calculated over an IP version 6 header as follows:
*
* Buffer[0..32] = Packet[8..23] (source address ) +
* Packet[24..39] (destination address)
*
* Result = Hash(Buffer, 32)
*/
#define _XEN_NETIF_CTRL_HASH_TYPE_IPV6 2
#define XEN_NETIF_CTRL_HASH_TYPE_IPV6 \
(1 << _XEN_NETIF_CTRL_HASH_TYPE_IPV6)
/*
* A hash calculated over an IP version 6 header and TCP header as
* follows:
*
* Buffer[0..36] = Packet[8..23] (source address) +
* Packet[24..39] (destination address) +
* Packet[40..41] (source port) +
* Packet[42..43] (destination port)
*
* Result = Hash(Buffer, 36)
*/
#define _XEN_NETIF_CTRL_HASH_TYPE_IPV6_TCP 3
#define XEN_NETIF_CTRL_HASH_TYPE_IPV6_TCP \
(1 << _XEN_NETIF_CTRL_HASH_TYPE_IPV6_TCP)
/*
* Hash algorithms
* ===============
*/
#define XEN_NETIF_CTRL_HASH_ALGORITHM_NONE 0
/*
* Toeplitz hash:
*/
#define XEN_NETIF_CTRL_HASH_ALGORITHM_TOEPLITZ 1
/*
* This algorithm uses a 'key' as well as the data buffer itself.
* (Buffer[] and Key[] are treated as shift-registers where the MSB of
* Buffer/Key[0] is considered 'left-most' and the LSB of Buffer/Key[N-1]
* is the 'right-most').
*
* Value = 0
* For number of bits in Buffer[]
* If (left-most bit of Buffer[] is 1)
* Value ^= left-most 32 bits of Key[]
* Key[] << 1
* Buffer[] << 1
*
* The code below is provided for convenience where an operating system
* does not already provide an implementation.
*/
#ifdef XEN_NETIF_DEFINE_TOEPLITZ
static uint32_t xen_netif_toeplitz_hash(const uint8_t *key,
unsigned int keylen,
const uint8_t *buf, unsigned int buflen)
{
unsigned int keyi, bufi;
uint64_t prefix = 0;
uint64_t hash = 0;
/* Pre-load prefix with the first 8 bytes of the key */
for (keyi = 0; keyi < 8; keyi++) {
prefix <<= 8;
prefix |= (keyi < keylen) ? key[keyi] : 0;
}
for (bufi = 0; bufi < buflen; bufi++) {
uint8_t byte = buf[bufi];
unsigned int bit;
for (bit = 0; bit < 8; bit++) {
if (byte & 0x80)
hash ^= prefix;
prefix <<= 1;
byte <<= 1;
}
/*
* 'prefix' has now been left-shifted by 8, so
* OR in the next byte.
*/
prefix |= (keyi < keylen) ? key[keyi] : 0;
keyi++;
}
/* The valid part of the hash is in the upper 32 bits. */
return hash >> 32;
}
#endif /* XEN_NETIF_DEFINE_TOEPLITZ */
/*
* Control requests (struct xen_netif_ctrl_request)
* ================================================
*
* All requests have the following format:
*
* 0 1 2 3 4 5 6 7 octet
* +-----+-----+-----+-----+-----+-----+-----+-----+
* | id | type | data[0] |
* +-----+-----+-----+-----+-----+-----+-----+-----+
* | data[1] | data[2] |
* +-----+-----+-----+-----+-----------------------+
*
* id: the request identifier, echoed in response.
* type: the type of request (see below)
* data[]: any data associated with the request (determined by type)
*/
struct xen_netif_ctrl_request {
uint16_t id;
uint16_t type;
#define XEN_NETIF_CTRL_TYPE_INVALID 0
#define XEN_NETIF_CTRL_TYPE_GET_HASH_FLAGS 1
#define XEN_NETIF_CTRL_TYPE_SET_HASH_FLAGS 2
#define XEN_NETIF_CTRL_TYPE_SET_HASH_KEY 3
#define XEN_NETIF_CTRL_TYPE_GET_HASH_MAPPING_SIZE 4
#define XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING_SIZE 5
#define XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING 6
#define XEN_NETIF_CTRL_TYPE_SET_HASH_ALGORITHM 7
uint32_t data[3];
};
/*
* Control responses (struct xen_netif_ctrl_response)
* ==================================================
*
* All responses have the following format:
*
* 0 1 2 3 4 5 6 7 octet
* +-----+-----+-----+-----+-----+-----+-----+-----+
* | id | type | status |
* +-----+-----+-----+-----+-----+-----+-----+-----+
* | data |
* +-----+-----+-----+-----+
*
* id: the corresponding request identifier
* type: the type of the corresponding request
* status: the status of request processing
* data: any data associated with the response (determined by type and
* status)
*/
struct xen_netif_ctrl_response {
uint16_t id;
uint16_t type;
uint32_t status;
#define XEN_NETIF_CTRL_STATUS_SUCCESS 0
#define XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED 1
#define XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER 2
#define XEN_NETIF_CTRL_STATUS_BUFFER_OVERFLOW 3
uint32_t data;
};
/*
* Control messages
* ================
*
* XEN_NETIF_CTRL_TYPE_SET_HASH_ALGORITHM
* --------------------------------------
*
* This is sent by the frontend to set the desired hash algorithm.
*
* Request:
*
* type = XEN_NETIF_CTRL_TYPE_SET_HASH_ALGORITHM
* data[0] = a XEN_NETIF_CTRL_HASH_ALGORITHM_* value
* data[1] = 0
* data[2] = 0
*
* Response:
*
* status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED - Operation not
* supported
* XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER - The algorithm is not
* supported
* XEN_NETIF_CTRL_STATUS_SUCCESS - Operation successful
*
* NOTE: Setting data[0] to XEN_NETIF_CTRL_HASH_ALGORITHM_NONE disables
* hashing and the backend is free to choose how it steers packets
* to queues (which is the default behaviour).
*
* XEN_NETIF_CTRL_TYPE_GET_HASH_FLAGS
* ----------------------------------
*
* This is sent by the frontend to query the types of hash supported by
* the backend.
*
* Request:
*
* type = XEN_NETIF_CTRL_TYPE_GET_HASH_FLAGS
* data[0] = 0
* data[1] = 0
* data[2] = 0
*
* Response:
*
* status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED - Operation not supported
* XEN_NETIF_CTRL_STATUS_SUCCESS - Operation successful
* data = supported hash types (if operation was successful)
*
* NOTE: A valid hash algorithm must be selected before this operation can
* succeed.
*
* XEN_NETIF_CTRL_TYPE_SET_HASH_FLAGS
* ----------------------------------
*
* This is sent by the frontend to set the types of hash that the backend
* should calculate. (See above for hash type definitions).
* Note that the 'maximal' type of hash should always be chosen. For
* example, if the frontend sets both IPV4 and IPV4_TCP hash types then
* the latter hash type should be calculated for any TCP packet and the
* former only calculated for non-TCP packets.
*
* Request:
*
* type = XEN_NETIF_CTRL_TYPE_SET_HASH_FLAGS
* data[0] = bitwise OR of XEN_NETIF_CTRL_HASH_TYPE_* values
* data[1] = 0
* data[2] = 0
*
* Response:
*
* status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED - Operation not
* supported
* XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER - One or more flag
* value is invalid or
* unsupported
* XEN_NETIF_CTRL_STATUS_SUCCESS - Operation successful
* data = 0
*
* NOTE: A valid hash algorithm must be selected before this operation can
* succeed.
* Also, setting data[0] to zero disables hashing and the backend
* is free to choose how it steers packets to queues.
*
* XEN_NETIF_CTRL_TYPE_SET_HASH_KEY
* --------------------------------
*
* This is sent by the frontend to set the key of the hash if the algorithm
* requires it. (See hash algorithms above).
*
* Request:
*
* type = XEN_NETIF_CTRL_TYPE_SET_HASH_KEY
* data[0] = grant reference of page containing the key (assumed to
* start at beginning of grant)
* data[1] = size of key in octets
* data[2] = 0
*
* Response:
*
* status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED - Operation not
* supported
* XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER - Key size is invalid
* XEN_NETIF_CTRL_STATUS_BUFFER_OVERFLOW - Key size is larger
* than the backend
* supports
* XEN_NETIF_CTRL_STATUS_SUCCESS - Operation successful
* data = 0
*
* NOTE: Any key octets not specified are assumed to be zero (the key
* is assumed to be empty by default) and specifying a new key
* invalidates any previous key, hence specifying a key size of
* zero will clear the key (which ensures that the calculated hash
* will always be zero).
* The maximum size of key is algorithm and backend specific, but
* is also limited by the single grant reference.
* The grant reference may be read-only and must remain valid until
* the response has been processed.
*
* XEN_NETIF_CTRL_TYPE_GET_HASH_MAPPING_SIZE
* -----------------------------------------
*
* This is sent by the frontend to query the maximum size of mapping
* table supported by the backend. The size is specified in terms of
* table entries.
*
* Request:
*
* type = XEN_NETIF_CTRL_TYPE_GET_HASH_MAPPING_SIZE
* data[0] = 0
* data[1] = 0
* data[2] = 0
*
* Response:
*
* status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED - Operation not supported
* XEN_NETIF_CTRL_STATUS_SUCCESS - Operation successful
* data = maximum number of entries allowed in the mapping table
* (if operation was successful) or zero if a mapping table is
* not supported (i.e. hash mapping is done only by modular
* arithmetic).
*
* XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING_SIZE
* -------------------------------------
*
* This is sent by the frontend to set the actual size of the mapping
* table to be used by the backend. The size is specified in terms of
* table entries.
* Any previous table is invalidated by this message and any new table
* is assumed to be zero filled.
*
* Request:
*
* type = XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING_SIZE
* data[0] = number of entries in mapping table
* data[1] = 0
* data[2] = 0
*
* Response:
*
* status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED - Operation not
* supported
* XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER - Table size is invalid
* XEN_NETIF_CTRL_STATUS_SUCCESS - Operation successful
* data = 0
*
* NOTE: Setting data[0] to 0 means that hash mapping should be done
* using modular arithmetic.
*
* XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING
* ------------------------------------
*
* This is sent by the frontend to set the content of the table mapping
* hash value to queue number. The backend should calculate the hash from
* the packet header, use it as an index into the table (modulo the size
* of the table) and then steer the packet to the queue number found at
* that index.
*
* Request:
*
* type = XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING
* data[0] = grant reference of page containing the mapping (sub-)table
* (assumed to start at beginning of grant)
* data[1] = size of (sub-)table in entries
* data[2] = offset, in entries, of sub-table within overall table
*
* Response:
*
* status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED - Operation not
* supported
* XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER - Table size or content
* is invalid
* XEN_NETIF_CTRL_STATUS_BUFFER_OVERFLOW - Table size is larger
* than the backend
* supports
* XEN_NETIF_CTRL_STATUS_SUCCESS - Operation successful
* data = 0
*
* NOTE: The overall table has the following format:
*
* 0 1 2 3 4 5 6 7 octet
* +-----+-----+-----+-----+-----+-----+-----+-----+
* | mapping[0] | mapping[1] |
* +-----+-----+-----+-----+-----+-----+-----+-----+
* | . |
* | . |
* | . |
* +-----+-----+-----+-----+-----+-----+-----+-----+
* | mapping[N-2] | mapping[N-1] |
* +-----+-----+-----+-----+-----+-----+-----+-----+
*
* where N is specified by a XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING_SIZE
* message and each mapping must specifies a queue between 0 and
* "multi-queue-num-queues" (see above).
* The backend may support a mapping table larger than can be
* mapped by a single grant reference. Thus sub-tables within a
* larger table can be individually set by sending multiple messages
* with differing offset values. Specifying a new sub-table does not
* invalidate any table data outside that range.
* The grant reference may be read-only and must remain valid until
* the response has been processed.
*/
DEFINE_RING_TYPES(xen_netif_ctrl,
struct xen_netif_ctrl_request,
struct xen_netif_ctrl_response);
/*
* Guest transmit
* ==============
*
* This is the 'wire' format for transmit (frontend -> backend) packets:
*
* Fragment 1: xen_netif_tx_request_t - flags = XEN_NETTXF_*
* size = total packet size
* [Extra 1: xen_netif_extra_info_t] - (only if fragment 1 flags include
* XEN_NETTXF_extra_info)
* ...
* [Extra N: xen_netif_extra_info_t] - (only if extra N-1 flags include
* XEN_NETIF_EXTRA_MORE)
* ...
* Fragment N: xen_netif_tx_request_t - (only if fragment N-1 flags include
* XEN_NETTXF_more_data - flags on preceding
* extras are not relevant here)
* flags = 0
* size = fragment size
*
* NOTE:
*
* This format slightly is different from that used for receive
* (backend -> frontend) packets. Specifically, in a multi-fragment
* packet the actual size of fragment 1 can only be determined by
* subtracting the sizes of fragments 2..N from the total packet size.
*
* Ring slot size is 12 octets, however not all request/response
* structs use the full size.
*
* tx request data (xen_netif_tx_request_t)
* ------------------------------------
*
* 0 1 2 3 4 5 6 7 octet
* +-----+-----+-----+-----+-----+-----+-----+-----+
* | grant ref | offset | flags |
* +-----+-----+-----+-----+-----+-----+-----+-----+
* | id | size |
* +-----+-----+-----+-----+
*
* grant ref: Reference to buffer page.
* offset: Offset within buffer page.
* flags: XEN_NETTXF_*.
* id: request identifier, echoed in response.
* size: packet size in bytes.
*
* tx response (xen_netif_tx_response_t)
* ---------------------------------
*
* 0 1 2 3 4 5 6 7 octet
* +-----+-----+-----+-----+-----+-----+-----+-----+
* | id | status | unused |
* +-----+-----+-----+-----+-----+-----+-----+-----+
* | unused |
* +-----+-----+-----+-----+
*
* id: reflects id in transmit request
* status: XEN_NETIF_RSP_*
*
* Guest receive
* =============
*
* This is the 'wire' format for receive (backend -> frontend) packets:
*
* Fragment 1: xen_netif_rx_request_t - flags = XEN_NETRXF_*
* size = fragment size
* [Extra 1: xen_netif_extra_info_t] - (only if fragment 1 flags include
* XEN_NETRXF_extra_info)
* ...
* [Extra N: xen_netif_extra_info_t] - (only if extra N-1 flags include
* XEN_NETIF_EXTRA_MORE)
* ...
* Fragment N: xen_netif_rx_request_t - (only if fragment N-1 flags include
* XEN_NETRXF_more_data - flags on preceding
* extras are not relevant here)
* flags = 0
* size = fragment size
*
* NOTE:
*
* This format slightly is different from that used for transmit
* (frontend -> backend) packets. Specifically, in a multi-fragment
* packet the size of the packet can only be determined by summing the
* sizes of fragments 1..N.
*
* Ring slot size is 8 octets.
*
* rx request (xen_netif_rx_request_t)
* -------------------------------
*
* 0 1 2 3 4 5 6 7 octet
* +-----+-----+-----+-----+-----+-----+-----+-----+
* | id | pad | gref |
* +-----+-----+-----+-----+-----+-----+-----+-----+
*
* id: request identifier, echoed in response.
* gref: reference to incoming granted frame.
*
* rx response (xen_netif_rx_response_t)
* ---------------------------------
*
* 0 1 2 3 4 5 6 7 octet
* +-----+-----+-----+-----+-----+-----+-----+-----+
* | id | offset | flags | status |
* +-----+-----+-----+-----+-----+-----+-----+-----+
*
* id: reflects id in receive request
* offset: offset in page of start of received packet
* flags: XEN_NETRXF_*
* status: -ve: XEN_NETIF_RSP_*; +ve: Rx'ed pkt size.
*
* NOTE: Historically, to support GSO on the frontend receive side, Linux
* netfront does not make use of the rx response id (because, as
* described below, extra info structures overlay the id field).
* Instead it assumes that responses always appear in the same ring
* slot as their corresponding request. Thus, to maintain
* compatibility, backends must make sure this is the case.
*
* Extra Info
* ==========
*
* Can be present if initial request or response has NET{T,R}XF_extra_info,
* or previous extra request has XEN_NETIF_EXTRA_MORE.
*
* The struct therefore needs to fit into either a tx or rx slot and
* is therefore limited to 8 octets.
*
* NOTE: Because extra info data overlays the usual request/response
* structures, there is no id information in the opposite direction.
* So, if an extra info overlays an rx response the frontend can
* assume that it is in the same ring slot as the request that was
* consumed to make the slot available, and the backend must ensure
* this assumption is true.
*
* extra info (xen_netif_extra_info_t)
* -------------------------------
*
* General format:
*
* 0 1 2 3 4 5 6 7 octet
* +-----+-----+-----+-----+-----+-----+-----+-----+
* |type |flags| type specific data |
* +-----+-----+-----+-----+-----+-----+-----+-----+
* | padding for tx |
* +-----+-----+-----+-----+
*
* type: XEN_NETIF_EXTRA_TYPE_*
* flags: XEN_NETIF_EXTRA_FLAG_*
* padding for tx: present only in the tx case due to 8 octet limit
* from rx case. Not shown in type specific entries
* below.
*
* XEN_NETIF_EXTRA_TYPE_GSO:
*
* 0 1 2 3 4 5 6 7 octet
* +-----+-----+-----+-----+-----+-----+-----+-----+
* |type |flags| size |type | pad | features |
* +-----+-----+-----+-----+-----+-----+-----+-----+
*
* type: Must be XEN_NETIF_EXTRA_TYPE_GSO
* flags: XEN_NETIF_EXTRA_FLAG_*
* size: Maximum payload size of each segment. For example,
* for TCP this is just the path MSS.
* type: XEN_NETIF_GSO_TYPE_*: This determines the protocol of
* the packet and any extra features required to segment the
* packet properly.
* features: EN_XEN_NETIF_GSO_FEAT_*: This specifies any extra GSO
* features required to process this packet, such as ECN
* support for TCPv4.
*
* XEN_NETIF_EXTRA_TYPE_MCAST_{ADD,DEL}:
*
* 0 1 2 3 4 5 6 7 octet
* +-----+-----+-----+-----+-----+-----+-----+-----+
* |type |flags| addr |
* +-----+-----+-----+-----+-----+-----+-----+-----+
*
* type: Must be XEN_NETIF_EXTRA_TYPE_MCAST_{ADD,DEL}
* flags: XEN_NETIF_EXTRA_FLAG_*
* addr: address to add/remove
*
* XEN_NETIF_EXTRA_TYPE_HASH:
*
* A backend that supports teoplitz hashing is assumed to accept
* this type of extra info in transmit packets.
* A frontend that enables hashing is assumed to accept
* this type of extra info in receive packets.
*
* 0 1 2 3 4 5 6 7 octet
* +-----+-----+-----+-----+-----+-----+-----+-----+
* |type |flags|htype| alg |LSB ---- value ---- MSB|
* +-----+-----+-----+-----+-----+-----+-----+-----+
*
* type: Must be XEN_NETIF_EXTRA_TYPE_HASH
* flags: XEN_NETIF_EXTRA_FLAG_*
* htype: Hash type (one of _XEN_NETIF_CTRL_HASH_TYPE_* - see above)
* alg: The algorithm used to calculate the hash (one of
* XEN_NETIF_CTRL_HASH_TYPE_ALGORITHM_* - see above)
* value: Hash value
*/
/* Protocol checksum field is blank in the packet (hardware offload)? */
#define _XEN_NETTXF_csum_blank (0)
#define XEN_NETTXF_csum_blank (1U<<_XEN_NETTXF_csum_blank)
/* Packet data has been validated against protocol checksum. */
#define _XEN_NETTXF_data_validated (1)
#define XEN_NETTXF_data_validated (1U<<_XEN_NETTXF_data_validated)
/* Packet continues in the next request descriptor. */
#define _XEN_NETTXF_more_data (2)
#define XEN_NETTXF_more_data (1U<<_XEN_NETTXF_more_data)
/* Packet to be followed by extra descriptor(s). */
#define _XEN_NETTXF_extra_info (3)
#define XEN_NETTXF_extra_info (1U<<_XEN_NETTXF_extra_info)
#define XEN_NETIF_MAX_TX_SIZE 0xFFFF
struct xen_netif_tx_request {
grant_ref_t gref;
uint16_t offset;
uint16_t flags;
uint16_t id;
uint16_t size;
};
/* Types of xen_netif_extra_info descriptors. */
#define XEN_NETIF_EXTRA_TYPE_NONE (0) /* Never used - invalid */
#define XEN_NETIF_EXTRA_TYPE_GSO (1) /* u.gso */
#define XEN_NETIF_EXTRA_TYPE_MCAST_ADD (2) /* u.mcast */
#define XEN_NETIF_EXTRA_TYPE_MCAST_DEL (3) /* u.mcast */
#define XEN_NETIF_EXTRA_TYPE_HASH (4) /* u.hash */
#define XEN_NETIF_EXTRA_TYPE_MAX (5)
/* xen_netif_extra_info_t flags. */
#define _XEN_NETIF_EXTRA_FLAG_MORE (0)
#define XEN_NETIF_EXTRA_FLAG_MORE (1U<<_XEN_NETIF_EXTRA_FLAG_MORE)
/* GSO types */
#define XEN_NETIF_GSO_TYPE_NONE (0)
#define XEN_NETIF_GSO_TYPE_TCPV4 (1)
#define XEN_NETIF_GSO_TYPE_TCPV6 (2)
/*
* This structure needs to fit within both xen_netif_tx_request_t and
* xen_netif_rx_response_t for compatibility.
*/
struct xen_netif_extra_info {
uint8_t type;
uint8_t flags;
union {
struct {
uint16_t size;
uint8_t type;
uint8_t pad;
uint16_t features;
} gso;
struct {
uint8_t addr[6];
} mcast;
struct {
uint8_t type;
uint8_t algorithm;
uint8_t value[4];
} hash;
uint16_t pad[3];
} u;
};
struct xen_netif_tx_response {
uint16_t id;
int16_t status;
};
struct xen_netif_rx_request {
uint16_t id; /* Echoed in response message. */
uint16_t pad;
grant_ref_t gref;
};
/* Packet data has been validated against protocol checksum. */
#define _XEN_NETRXF_data_validated (0)
#define XEN_NETRXF_data_validated (1U<<_XEN_NETRXF_data_validated)
/* Protocol checksum field is blank in the packet (hardware offload)? */
#define _XEN_NETRXF_csum_blank (1)
#define XEN_NETRXF_csum_blank (1U<<_XEN_NETRXF_csum_blank)
/* Packet continues in the next request descriptor. */
#define _XEN_NETRXF_more_data (2)
#define XEN_NETRXF_more_data (1U<<_XEN_NETRXF_more_data)
/* Packet to be followed by extra descriptor(s). */
#define _XEN_NETRXF_extra_info (3)
#define XEN_NETRXF_extra_info (1U<<_XEN_NETRXF_extra_info)
/* Packet has GSO prefix. Deprecated but included for compatibility */
#define _XEN_NETRXF_gso_prefix (4)
#define XEN_NETRXF_gso_prefix (1U<<_XEN_NETRXF_gso_prefix)
struct xen_netif_rx_response {
uint16_t id;
uint16_t offset;
uint16_t flags;
int16_t status;
};
/*
* Generate xen_netif ring structures and types.
*/
DEFINE_RING_TYPES(xen_netif_tx, struct xen_netif_tx_request,
struct xen_netif_tx_response);
DEFINE_RING_TYPES(xen_netif_rx, struct xen_netif_rx_request,
struct xen_netif_rx_response);
#define XEN_NETIF_RSP_DROPPED -2
#define XEN_NETIF_RSP_ERROR -1
#define XEN_NETIF_RSP_OKAY 0
/* No response: used for auxiliary requests (e.g., xen_netif_extra_info_t). */
#define XEN_NETIF_RSP_NULL 1
#endif