1 /* SPDX-License-Identifier: MIT */
2 /******************************************************************************
3 * xen_netif.h
4 *
5 * Unified network-device I/O interface for Xen guest OSes.
6 *
7 * Copyright (c) 2003-2004, Keir Fraser
8 */
9
10 #ifndef __XEN_PUBLIC_IO_XEN_NETIF_H__
11 #define __XEN_PUBLIC_IO_XEN_NETIF_H__
12
13 #include "ring.h"
14 #include "../grant_table.h"
15
16 /*
17 * Older implementation of Xen network frontend / backend has an
18 * implicit dependency on the MAX_SKB_FRAGS as the maximum number of
19 * ring slots a skb can use. Netfront / netback may not work as
20 * expected when frontend and backend have different MAX_SKB_FRAGS.
21 *
22 * A better approach is to add mechanism for netfront / netback to
23 * negotiate this value. However we cannot fix all possible
24 * frontends, so we need to define a value which states the minimum
25 * slots backend must support.
26 *
27 * The minimum value derives from older Linux kernel's MAX_SKB_FRAGS
28 * (18), which is proved to work with most frontends. Any new backend
29 * which doesn't negotiate with frontend should expect frontend to
30 * send a valid packet using slots up to this value.
31 */
32 #define XEN_NETIF_NR_SLOTS_MIN 18
33
34 /*
35 * Notifications after enqueuing any type of message should be conditional on
36 * the appropriate req_event or rsp_event field in the shared ring.
37 * If the client sends notification for rx requests then it should specify
38 * feature 'feature-rx-notify' via xenbus. Otherwise the backend will assume
39 * that it cannot safely queue packets (as it may not be kicked to send them).
40 */
41
42 /*
43 * "feature-split-event-channels" is introduced to separate guest TX
44 * and RX notification. Backend either doesn't support this feature or
45 * advertises it via xenstore as 0 (disabled) or 1 (enabled).
46 *
47 * To make use of this feature, frontend should allocate two event
48 * channels for TX and RX, advertise them to backend as
49 * "event-channel-tx" and "event-channel-rx" respectively. If frontend
50 * doesn't want to use this feature, it just writes "event-channel"
51 * node as before.
52 */
53
54 /*
55 * Multiple transmit and receive queues:
56 * If supported, the backend will write the key "multi-queue-max-queues" to
57 * the directory for that vif, and set its value to the maximum supported
58 * number of queues.
59 * Frontends that are aware of this feature and wish to use it can write the
60 * key "multi-queue-num-queues", set to the number they wish to use, which
61 * must be greater than zero, and no more than the value reported by the backend
62 * in "multi-queue-max-queues".
63 *
64 * Queues replicate the shared rings and event channels.
65 * "feature-split-event-channels" may optionally be used when using
66 * multiple queues, but is not mandatory.
67 *
68 * Each queue consists of one shared ring pair, i.e. there must be the same
69 * number of tx and rx rings.
70 *
71 * For frontends requesting just one queue, the usual event-channel and
72 * ring-ref keys are written as before, simplifying the backend processing
73 * to avoid distinguishing between a frontend that doesn't understand the
74 * multi-queue feature, and one that does, but requested only one queue.
75 *
76 * Frontends requesting two or more queues must not write the toplevel
77 * event-channel (or event-channel-{tx,rx}) and {tx,rx}-ring-ref keys,
78 * instead writing those keys under sub-keys having the name "queue-N" where
79 * N is the integer ID of the queue for which those keys belong. Queues
80 * are indexed from zero. For example, a frontend with two queues and split
81 * event channels must write the following set of queue-related keys:
82 *
83 * /local/domain/1/device/vif/0/multi-queue-num-queues = "2"
84 * /local/domain/1/device/vif/0/queue-0 = ""
85 * /local/domain/1/device/vif/0/queue-0/tx-ring-ref = "<ring-ref-tx0>"
86 * /local/domain/1/device/vif/0/queue-0/rx-ring-ref = "<ring-ref-rx0>"
87 * /local/domain/1/device/vif/0/queue-0/event-channel-tx = "<evtchn-tx0>"
88 * /local/domain/1/device/vif/0/queue-0/event-channel-rx = "<evtchn-rx0>"
89 * /local/domain/1/device/vif/0/queue-1 = ""
90 * /local/domain/1/device/vif/0/queue-1/tx-ring-ref = "<ring-ref-tx1>"
91 * /local/domain/1/device/vif/0/queue-1/rx-ring-ref = "<ring-ref-rx1"
92 * /local/domain/1/device/vif/0/queue-1/event-channel-tx = "<evtchn-tx1>"
93 * /local/domain/1/device/vif/0/queue-1/event-channel-rx = "<evtchn-rx1>"
94 *
95 * If there is any inconsistency in the XenStore data, the backend may
96 * choose not to connect any queues, instead treating the request as an
97 * error. This includes scenarios where more (or fewer) queues were
98 * requested than the frontend provided details for.
99 *
100 * Mapping of packets to queues is considered to be a function of the
101 * transmitting system (backend or frontend) and is not negotiated
102 * between the two. Guests are free to transmit packets on any queue
103 * they choose, provided it has been set up correctly. Guests must be
104 * prepared to receive packets on any queue they have requested be set up.
105 */
106
107 /*
108 * "feature-no-csum-offload" should be used to turn IPv4 TCP/UDP checksum
109 * offload off or on. If it is missing then the feature is assumed to be on.
110 * "feature-ipv6-csum-offload" should be used to turn IPv6 TCP/UDP checksum
111 * offload on or off. If it is missing then the feature is assumed to be off.
112 */
113
114 /*
115 * "feature-gso-tcpv4" and "feature-gso-tcpv6" advertise the capability to
116 * handle large TCP packets (in IPv4 or IPv6 form respectively). Neither
117 * frontends nor backends are assumed to be capable unless the flags are
118 * present.
119 */
120
121 /*
122 * "feature-multicast-control" and "feature-dynamic-multicast-control"
123 * advertise the capability to filter ethernet multicast packets in the
124 * backend. If the frontend wishes to take advantage of this feature then
125 * it may set "request-multicast-control". If the backend only advertises
126 * "feature-multicast-control" then "request-multicast-control" must be set
127 * before the frontend moves into the connected state. The backend will
128 * sample the value on this state transition and any subsequent change in
129 * value will have no effect. However, if the backend also advertises
130 * "feature-dynamic-multicast-control" then "request-multicast-control"
131 * may be set by the frontend at any time. In this case, the backend will
132 * watch the value and re-sample on watch events.
133 *
134 * If the sampled value of "request-multicast-control" is set then the
135 * backend transmit side should no longer flood multicast packets to the
136 * frontend, it should instead drop any multicast packet that does not
137 * match in a filter list.
138 * The list is amended by the frontend by sending dummy transmit requests
139 * containing XEN_NETIF_EXTRA_TYPE_MCAST_{ADD,DEL} extra-info fragments as
140 * specified below.
141 * Note that the filter list may be amended even if the sampled value of
142 * "request-multicast-control" is not set, however the filter should only
143 * be applied if it is set.
144 */
145
146 /*
147 * "xdp-headroom" is used to request that extra space is added
148 * for XDP processing. The value is measured in bytes and passed by
149 * the frontend to be consistent between both ends.
150 * If the value is greater than zero that means that
151 * an RX response is going to be passed to an XDP program for processing.
152 * XEN_NETIF_MAX_XDP_HEADROOM defines the maximum headroom offset in bytes
153 *
154 * "feature-xdp-headroom" is set to "1" by the netback side like other features
155 * so a guest can check if an XDP program can be processed.
156 */
157 #define XEN_NETIF_MAX_XDP_HEADROOM 0x7FFF
158
159 /*
160 * Control ring
161 * ============
162 *
163 * Some features, such as hashing (detailed below), require a
164 * significant amount of out-of-band data to be passed from frontend to
165 * backend. Use of xenstore is not suitable for large quantities of data
166 * because of quota limitations and so a dedicated 'control ring' is used.
167 * The ability of the backend to use a control ring is advertised by
168 * setting:
169 *
170 * /local/domain/X/backend/<domid>/<vif>/feature-ctrl-ring = "1"
171 *
172 * The frontend provides a control ring to the backend by setting:
173 *
174 * /local/domain/<domid>/device/vif/<vif>/ctrl-ring-ref = <gref>
175 * /local/domain/<domid>/device/vif/<vif>/event-channel-ctrl = <port>
176 *
177 * where <gref> is the grant reference of the shared page used to
178 * implement the control ring and <port> is an event channel to be used
179 * as a mailbox interrupt. These keys must be set before the frontend
180 * moves into the connected state.
181 *
182 * The control ring uses a fixed request/response message size and is
183 * balanced (i.e. one request to one response), so operationally it is much
184 * the same as a transmit or receive ring.
185 * Note that there is no requirement that responses are issued in the same
186 * order as requests.
187 */
188
189 /*
190 * Hash types
191 * ==========
192 *
193 * For the purposes of the definitions below, 'Packet[]' is an array of
194 * octets containing an IP packet without options, 'Array[X..Y]' means a
195 * sub-array of 'Array' containing bytes X thru Y inclusive, and '+' is
196 * used to indicate concatenation of arrays.
197 */
198
199 /*
200 * A hash calculated over an IP version 4 header as follows:
201 *
202 * Buffer[0..8] = Packet[12..15] (source address) +
203 * Packet[16..19] (destination address)
204 *
205 * Result = Hash(Buffer, 8)
206 */
207 #define _XEN_NETIF_CTRL_HASH_TYPE_IPV4 0
208 #define XEN_NETIF_CTRL_HASH_TYPE_IPV4 \
209 (1 << _XEN_NETIF_CTRL_HASH_TYPE_IPV4)
210
211 /*
212 * A hash calculated over an IP version 4 header and TCP header as
213 * follows:
214 *
215 * Buffer[0..12] = Packet[12..15] (source address) +
216 * Packet[16..19] (destination address) +
217 * Packet[20..21] (source port) +
218 * Packet[22..23] (destination port)
219 *
220 * Result = Hash(Buffer, 12)
221 */
222 #define _XEN_NETIF_CTRL_HASH_TYPE_IPV4_TCP 1
223 #define XEN_NETIF_CTRL_HASH_TYPE_IPV4_TCP \
224 (1 << _XEN_NETIF_CTRL_HASH_TYPE_IPV4_TCP)
225
226 /*
227 * A hash calculated over an IP version 6 header as follows:
228 *
229 * Buffer[0..32] = Packet[8..23] (source address ) +
230 * Packet[24..39] (destination address)
231 *
232 * Result = Hash(Buffer, 32)
233 */
234 #define _XEN_NETIF_CTRL_HASH_TYPE_IPV6 2
235 #define XEN_NETIF_CTRL_HASH_TYPE_IPV6 \
236 (1 << _XEN_NETIF_CTRL_HASH_TYPE_IPV6)
237
238 /*
239 * A hash calculated over an IP version 6 header and TCP header as
240 * follows:
241 *
242 * Buffer[0..36] = Packet[8..23] (source address) +
243 * Packet[24..39] (destination address) +
244 * Packet[40..41] (source port) +
245 * Packet[42..43] (destination port)
246 *
247 * Result = Hash(Buffer, 36)
248 */
249 #define _XEN_NETIF_CTRL_HASH_TYPE_IPV6_TCP 3
250 #define XEN_NETIF_CTRL_HASH_TYPE_IPV6_TCP \
251 (1 << _XEN_NETIF_CTRL_HASH_TYPE_IPV6_TCP)
252
253 /*
254 * Hash algorithms
255 * ===============
256 */
257
258 #define XEN_NETIF_CTRL_HASH_ALGORITHM_NONE 0
259
260 /*
261 * Toeplitz hash:
262 */
263
264 #define XEN_NETIF_CTRL_HASH_ALGORITHM_TOEPLITZ 1
265
266 /*
267 * This algorithm uses a 'key' as well as the data buffer itself.
268 * (Buffer[] and Key[] are treated as shift-registers where the MSB of
269 * Buffer/Key[0] is considered 'left-most' and the LSB of Buffer/Key[N-1]
270 * is the 'right-most').
271 *
272 * Value = 0
273 * For number of bits in Buffer[]
274 * If (left-most bit of Buffer[] is 1)
275 * Value ^= left-most 32 bits of Key[]
276 * Key[] << 1
277 * Buffer[] << 1
278 *
279 * The code below is provided for convenience where an operating system
280 * does not already provide an implementation.
281 */
282 #ifdef XEN_NETIF_DEFINE_TOEPLITZ
xen_netif_toeplitz_hash(const uint8_t * key,unsigned int keylen,const uint8_t * buf,unsigned int buflen)283 static uint32_t xen_netif_toeplitz_hash(const uint8_t *key,
284 unsigned int keylen,
285 const uint8_t *buf, unsigned int buflen)
286 {
287 unsigned int keyi, bufi;
288 uint64_t prefix = 0;
289 uint64_t hash = 0;
290
291 /* Pre-load prefix with the first 8 bytes of the key */
292 for (keyi = 0; keyi < 8; keyi++) {
293 prefix <<= 8;
294 prefix |= (keyi < keylen) ? key[keyi] : 0;
295 }
296
297 for (bufi = 0; bufi < buflen; bufi++) {
298 uint8_t byte = buf[bufi];
299 unsigned int bit;
300
301 for (bit = 0; bit < 8; bit++) {
302 if (byte & 0x80)
303 hash ^= prefix;
304 prefix <<= 1;
305 byte <<= 1;
306 }
307
308 /*
309 * 'prefix' has now been left-shifted by 8, so
310 * OR in the next byte.
311 */
312 prefix |= (keyi < keylen) ? key[keyi] : 0;
313 keyi++;
314 }
315
316 /* The valid part of the hash is in the upper 32 bits. */
317 return hash >> 32;
318 }
319 #endif /* XEN_NETIF_DEFINE_TOEPLITZ */
320
321 /*
322 * Control requests (struct xen_netif_ctrl_request)
323 * ================================================
324 *
325 * All requests have the following format:
326 *
327 * 0 1 2 3 4 5 6 7 octet
328 * +-----+-----+-----+-----+-----+-----+-----+-----+
329 * | id | type | data[0] |
330 * +-----+-----+-----+-----+-----+-----+-----+-----+
331 * | data[1] | data[2] |
332 * +-----+-----+-----+-----+-----------------------+
333 *
334 * id: the request identifier, echoed in response.
335 * type: the type of request (see below)
336 * data[]: any data associated with the request (determined by type)
337 */
338
339 struct xen_netif_ctrl_request {
340 uint16_t id;
341 uint16_t type;
342
343 #define XEN_NETIF_CTRL_TYPE_INVALID 0
344 #define XEN_NETIF_CTRL_TYPE_GET_HASH_FLAGS 1
345 #define XEN_NETIF_CTRL_TYPE_SET_HASH_FLAGS 2
346 #define XEN_NETIF_CTRL_TYPE_SET_HASH_KEY 3
347 #define XEN_NETIF_CTRL_TYPE_GET_HASH_MAPPING_SIZE 4
348 #define XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING_SIZE 5
349 #define XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING 6
350 #define XEN_NETIF_CTRL_TYPE_SET_HASH_ALGORITHM 7
351
352 uint32_t data[3];
353 };
354
355 /*
356 * Control responses (struct xen_netif_ctrl_response)
357 * ==================================================
358 *
359 * All responses have the following format:
360 *
361 * 0 1 2 3 4 5 6 7 octet
362 * +-----+-----+-----+-----+-----+-----+-----+-----+
363 * | id | type | status |
364 * +-----+-----+-----+-----+-----+-----+-----+-----+
365 * | data |
366 * +-----+-----+-----+-----+
367 *
368 * id: the corresponding request identifier
369 * type: the type of the corresponding request
370 * status: the status of request processing
371 * data: any data associated with the response (determined by type and
372 * status)
373 */
374
375 struct xen_netif_ctrl_response {
376 uint16_t id;
377 uint16_t type;
378 uint32_t status;
379
380 #define XEN_NETIF_CTRL_STATUS_SUCCESS 0
381 #define XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED 1
382 #define XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER 2
383 #define XEN_NETIF_CTRL_STATUS_BUFFER_OVERFLOW 3
384
385 uint32_t data;
386 };
387
388 /*
389 * Control messages
390 * ================
391 *
392 * XEN_NETIF_CTRL_TYPE_SET_HASH_ALGORITHM
393 * --------------------------------------
394 *
395 * This is sent by the frontend to set the desired hash algorithm.
396 *
397 * Request:
398 *
399 * type = XEN_NETIF_CTRL_TYPE_SET_HASH_ALGORITHM
400 * data[0] = a XEN_NETIF_CTRL_HASH_ALGORITHM_* value
401 * data[1] = 0
402 * data[2] = 0
403 *
404 * Response:
405 *
406 * status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED - Operation not
407 * supported
408 * XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER - The algorithm is not
409 * supported
410 * XEN_NETIF_CTRL_STATUS_SUCCESS - Operation successful
411 *
412 * NOTE: Setting data[0] to XEN_NETIF_CTRL_HASH_ALGORITHM_NONE disables
413 * hashing and the backend is free to choose how it steers packets
414 * to queues (which is the default behaviour).
415 *
416 * XEN_NETIF_CTRL_TYPE_GET_HASH_FLAGS
417 * ----------------------------------
418 *
419 * This is sent by the frontend to query the types of hash supported by
420 * the backend.
421 *
422 * Request:
423 *
424 * type = XEN_NETIF_CTRL_TYPE_GET_HASH_FLAGS
425 * data[0] = 0
426 * data[1] = 0
427 * data[2] = 0
428 *
429 * Response:
430 *
431 * status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED - Operation not supported
432 * XEN_NETIF_CTRL_STATUS_SUCCESS - Operation successful
433 * data = supported hash types (if operation was successful)
434 *
435 * NOTE: A valid hash algorithm must be selected before this operation can
436 * succeed.
437 *
438 * XEN_NETIF_CTRL_TYPE_SET_HASH_FLAGS
439 * ----------------------------------
440 *
441 * This is sent by the frontend to set the types of hash that the backend
442 * should calculate. (See above for hash type definitions).
443 * Note that the 'maximal' type of hash should always be chosen. For
444 * example, if the frontend sets both IPV4 and IPV4_TCP hash types then
445 * the latter hash type should be calculated for any TCP packet and the
446 * former only calculated for non-TCP packets.
447 *
448 * Request:
449 *
450 * type = XEN_NETIF_CTRL_TYPE_SET_HASH_FLAGS
451 * data[0] = bitwise OR of XEN_NETIF_CTRL_HASH_TYPE_* values
452 * data[1] = 0
453 * data[2] = 0
454 *
455 * Response:
456 *
457 * status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED - Operation not
458 * supported
459 * XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER - One or more flag
460 * value is invalid or
461 * unsupported
462 * XEN_NETIF_CTRL_STATUS_SUCCESS - Operation successful
463 * data = 0
464 *
465 * NOTE: A valid hash algorithm must be selected before this operation can
466 * succeed.
467 * Also, setting data[0] to zero disables hashing and the backend
468 * is free to choose how it steers packets to queues.
469 *
470 * XEN_NETIF_CTRL_TYPE_SET_HASH_KEY
471 * --------------------------------
472 *
473 * This is sent by the frontend to set the key of the hash if the algorithm
474 * requires it. (See hash algorithms above).
475 *
476 * Request:
477 *
478 * type = XEN_NETIF_CTRL_TYPE_SET_HASH_KEY
479 * data[0] = grant reference of page containing the key (assumed to
480 * start at beginning of grant)
481 * data[1] = size of key in octets
482 * data[2] = 0
483 *
484 * Response:
485 *
486 * status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED - Operation not
487 * supported
488 * XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER - Key size is invalid
489 * XEN_NETIF_CTRL_STATUS_BUFFER_OVERFLOW - Key size is larger
490 * than the backend
491 * supports
492 * XEN_NETIF_CTRL_STATUS_SUCCESS - Operation successful
493 * data = 0
494 *
495 * NOTE: Any key octets not specified are assumed to be zero (the key
496 * is assumed to be empty by default) and specifying a new key
497 * invalidates any previous key, hence specifying a key size of
498 * zero will clear the key (which ensures that the calculated hash
499 * will always be zero).
500 * The maximum size of key is algorithm and backend specific, but
501 * is also limited by the single grant reference.
502 * The grant reference may be read-only and must remain valid until
503 * the response has been processed.
504 *
505 * XEN_NETIF_CTRL_TYPE_GET_HASH_MAPPING_SIZE
506 * -----------------------------------------
507 *
508 * This is sent by the frontend to query the maximum size of mapping
509 * table supported by the backend. The size is specified in terms of
510 * table entries.
511 *
512 * Request:
513 *
514 * type = XEN_NETIF_CTRL_TYPE_GET_HASH_MAPPING_SIZE
515 * data[0] = 0
516 * data[1] = 0
517 * data[2] = 0
518 *
519 * Response:
520 *
521 * status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED - Operation not supported
522 * XEN_NETIF_CTRL_STATUS_SUCCESS - Operation successful
523 * data = maximum number of entries allowed in the mapping table
524 * (if operation was successful) or zero if a mapping table is
525 * not supported (i.e. hash mapping is done only by modular
526 * arithmetic).
527 *
528 * XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING_SIZE
529 * -------------------------------------
530 *
531 * This is sent by the frontend to set the actual size of the mapping
532 * table to be used by the backend. The size is specified in terms of
533 * table entries.
534 * Any previous table is invalidated by this message and any new table
535 * is assumed to be zero filled.
536 *
537 * Request:
538 *
539 * type = XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING_SIZE
540 * data[0] = number of entries in mapping table
541 * data[1] = 0
542 * data[2] = 0
543 *
544 * Response:
545 *
546 * status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED - Operation not
547 * supported
548 * XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER - Table size is invalid
549 * XEN_NETIF_CTRL_STATUS_SUCCESS - Operation successful
550 * data = 0
551 *
552 * NOTE: Setting data[0] to 0 means that hash mapping should be done
553 * using modular arithmetic.
554 *
555 * XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING
556 * ------------------------------------
557 *
558 * This is sent by the frontend to set the content of the table mapping
559 * hash value to queue number. The backend should calculate the hash from
560 * the packet header, use it as an index into the table (modulo the size
561 * of the table) and then steer the packet to the queue number found at
562 * that index.
563 *
564 * Request:
565 *
566 * type = XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING
567 * data[0] = grant reference of page containing the mapping (sub-)table
568 * (assumed to start at beginning of grant)
569 * data[1] = size of (sub-)table in entries
570 * data[2] = offset, in entries, of sub-table within overall table
571 *
572 * Response:
573 *
574 * status = XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED - Operation not
575 * supported
576 * XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER - Table size or content
577 * is invalid
578 * XEN_NETIF_CTRL_STATUS_BUFFER_OVERFLOW - Table size is larger
579 * than the backend
580 * supports
581 * XEN_NETIF_CTRL_STATUS_SUCCESS - Operation successful
582 * data = 0
583 *
584 * NOTE: The overall table has the following format:
585 *
586 * 0 1 2 3 4 5 6 7 octet
587 * +-----+-----+-----+-----+-----+-----+-----+-----+
588 * | mapping[0] | mapping[1] |
589 * +-----+-----+-----+-----+-----+-----+-----+-----+
590 * | . |
591 * | . |
592 * | . |
593 * +-----+-----+-----+-----+-----+-----+-----+-----+
594 * | mapping[N-2] | mapping[N-1] |
595 * +-----+-----+-----+-----+-----+-----+-----+-----+
596 *
597 * where N is specified by a XEN_NETIF_CTRL_TYPE_SET_HASH_MAPPING_SIZE
598 * message and each mapping must specifies a queue between 0 and
599 * "multi-queue-num-queues" (see above).
600 * The backend may support a mapping table larger than can be
601 * mapped by a single grant reference. Thus sub-tables within a
602 * larger table can be individually set by sending multiple messages
603 * with differing offset values. Specifying a new sub-table does not
604 * invalidate any table data outside that range.
605 * The grant reference may be read-only and must remain valid until
606 * the response has been processed.
607 */
608
609 DEFINE_RING_TYPES(xen_netif_ctrl,
610 struct xen_netif_ctrl_request,
611 struct xen_netif_ctrl_response);
612
613 /*
614 * Guest transmit
615 * ==============
616 *
617 * This is the 'wire' format for transmit (frontend -> backend) packets:
618 *
619 * Fragment 1: xen_netif_tx_request_t - flags = XEN_NETTXF_*
620 * size = total packet size
621 * [Extra 1: xen_netif_extra_info_t] - (only if fragment 1 flags include
622 * XEN_NETTXF_extra_info)
623 * ...
624 * [Extra N: xen_netif_extra_info_t] - (only if extra N-1 flags include
625 * XEN_NETIF_EXTRA_MORE)
626 * ...
627 * Fragment N: xen_netif_tx_request_t - (only if fragment N-1 flags include
628 * XEN_NETTXF_more_data - flags on preceding
629 * extras are not relevant here)
630 * flags = 0
631 * size = fragment size
632 *
633 * NOTE:
634 *
635 * This format slightly is different from that used for receive
636 * (backend -> frontend) packets. Specifically, in a multi-fragment
637 * packet the actual size of fragment 1 can only be determined by
638 * subtracting the sizes of fragments 2..N from the total packet size.
639 *
640 * Ring slot size is 12 octets, however not all request/response
641 * structs use the full size.
642 *
643 * tx request data (xen_netif_tx_request_t)
644 * ------------------------------------
645 *
646 * 0 1 2 3 4 5 6 7 octet
647 * +-----+-----+-----+-----+-----+-----+-----+-----+
648 * | grant ref | offset | flags |
649 * +-----+-----+-----+-----+-----+-----+-----+-----+
650 * | id | size |
651 * +-----+-----+-----+-----+
652 *
653 * grant ref: Reference to buffer page.
654 * offset: Offset within buffer page.
655 * flags: XEN_NETTXF_*.
656 * id: request identifier, echoed in response.
657 * size: packet size in bytes.
658 *
659 * tx response (xen_netif_tx_response_t)
660 * ---------------------------------
661 *
662 * 0 1 2 3 4 5 6 7 octet
663 * +-----+-----+-----+-----+-----+-----+-----+-----+
664 * | id | status | unused |
665 * +-----+-----+-----+-----+-----+-----+-----+-----+
666 * | unused |
667 * +-----+-----+-----+-----+
668 *
669 * id: reflects id in transmit request
670 * status: XEN_NETIF_RSP_*
671 *
672 * Guest receive
673 * =============
674 *
675 * This is the 'wire' format for receive (backend -> frontend) packets:
676 *
677 * Fragment 1: xen_netif_rx_request_t - flags = XEN_NETRXF_*
678 * size = fragment size
679 * [Extra 1: xen_netif_extra_info_t] - (only if fragment 1 flags include
680 * XEN_NETRXF_extra_info)
681 * ...
682 * [Extra N: xen_netif_extra_info_t] - (only if extra N-1 flags include
683 * XEN_NETIF_EXTRA_MORE)
684 * ...
685 * Fragment N: xen_netif_rx_request_t - (only if fragment N-1 flags include
686 * XEN_NETRXF_more_data - flags on preceding
687 * extras are not relevant here)
688 * flags = 0
689 * size = fragment size
690 *
691 * NOTE:
692 *
693 * This format slightly is different from that used for transmit
694 * (frontend -> backend) packets. Specifically, in a multi-fragment
695 * packet the size of the packet can only be determined by summing the
696 * sizes of fragments 1..N.
697 *
698 * Ring slot size is 8 octets.
699 *
700 * rx request (xen_netif_rx_request_t)
701 * -------------------------------
702 *
703 * 0 1 2 3 4 5 6 7 octet
704 * +-----+-----+-----+-----+-----+-----+-----+-----+
705 * | id | pad | gref |
706 * +-----+-----+-----+-----+-----+-----+-----+-----+
707 *
708 * id: request identifier, echoed in response.
709 * gref: reference to incoming granted frame.
710 *
711 * rx response (xen_netif_rx_response_t)
712 * ---------------------------------
713 *
714 * 0 1 2 3 4 5 6 7 octet
715 * +-----+-----+-----+-----+-----+-----+-----+-----+
716 * | id | offset | flags | status |
717 * +-----+-----+-----+-----+-----+-----+-----+-----+
718 *
719 * id: reflects id in receive request
720 * offset: offset in page of start of received packet
721 * flags: XEN_NETRXF_*
722 * status: -ve: XEN_NETIF_RSP_*; +ve: Rx'ed pkt size.
723 *
724 * NOTE: Historically, to support GSO on the frontend receive side, Linux
725 * netfront does not make use of the rx response id (because, as
726 * described below, extra info structures overlay the id field).
727 * Instead it assumes that responses always appear in the same ring
728 * slot as their corresponding request. Thus, to maintain
729 * compatibility, backends must make sure this is the case.
730 *
731 * Extra Info
732 * ==========
733 *
734 * Can be present if initial request or response has NET{T,R}XF_extra_info,
735 * or previous extra request has XEN_NETIF_EXTRA_MORE.
736 *
737 * The struct therefore needs to fit into either a tx or rx slot and
738 * is therefore limited to 8 octets.
739 *
740 * NOTE: Because extra info data overlays the usual request/response
741 * structures, there is no id information in the opposite direction.
742 * So, if an extra info overlays an rx response the frontend can
743 * assume that it is in the same ring slot as the request that was
744 * consumed to make the slot available, and the backend must ensure
745 * this assumption is true.
746 *
747 * extra info (xen_netif_extra_info_t)
748 * -------------------------------
749 *
750 * General format:
751 *
752 * 0 1 2 3 4 5 6 7 octet
753 * +-----+-----+-----+-----+-----+-----+-----+-----+
754 * |type |flags| type specific data |
755 * +-----+-----+-----+-----+-----+-----+-----+-----+
756 * | padding for tx |
757 * +-----+-----+-----+-----+
758 *
759 * type: XEN_NETIF_EXTRA_TYPE_*
760 * flags: XEN_NETIF_EXTRA_FLAG_*
761 * padding for tx: present only in the tx case due to 8 octet limit
762 * from rx case. Not shown in type specific entries
763 * below.
764 *
765 * XEN_NETIF_EXTRA_TYPE_GSO:
766 *
767 * 0 1 2 3 4 5 6 7 octet
768 * +-----+-----+-----+-----+-----+-----+-----+-----+
769 * |type |flags| size |type | pad | features |
770 * +-----+-----+-----+-----+-----+-----+-----+-----+
771 *
772 * type: Must be XEN_NETIF_EXTRA_TYPE_GSO
773 * flags: XEN_NETIF_EXTRA_FLAG_*
774 * size: Maximum payload size of each segment. For example,
775 * for TCP this is just the path MSS.
776 * type: XEN_NETIF_GSO_TYPE_*: This determines the protocol of
777 * the packet and any extra features required to segment the
778 * packet properly.
779 * features: EN_XEN_NETIF_GSO_FEAT_*: This specifies any extra GSO
780 * features required to process this packet, such as ECN
781 * support for TCPv4.
782 *
783 * XEN_NETIF_EXTRA_TYPE_MCAST_{ADD,DEL}:
784 *
785 * 0 1 2 3 4 5 6 7 octet
786 * +-----+-----+-----+-----+-----+-----+-----+-----+
787 * |type |flags| addr |
788 * +-----+-----+-----+-----+-----+-----+-----+-----+
789 *
790 * type: Must be XEN_NETIF_EXTRA_TYPE_MCAST_{ADD,DEL}
791 * flags: XEN_NETIF_EXTRA_FLAG_*
792 * addr: address to add/remove
793 *
794 * XEN_NETIF_EXTRA_TYPE_HASH:
795 *
796 * A backend that supports teoplitz hashing is assumed to accept
797 * this type of extra info in transmit packets.
798 * A frontend that enables hashing is assumed to accept
799 * this type of extra info in receive packets.
800 *
801 * 0 1 2 3 4 5 6 7 octet
802 * +-----+-----+-----+-----+-----+-----+-----+-----+
803 * |type |flags|htype| alg |LSB ---- value ---- MSB|
804 * +-----+-----+-----+-----+-----+-----+-----+-----+
805 *
806 * type: Must be XEN_NETIF_EXTRA_TYPE_HASH
807 * flags: XEN_NETIF_EXTRA_FLAG_*
808 * htype: Hash type (one of _XEN_NETIF_CTRL_HASH_TYPE_* - see above)
809 * alg: The algorithm used to calculate the hash (one of
810 * XEN_NETIF_CTRL_HASH_TYPE_ALGORITHM_* - see above)
811 * value: Hash value
812 */
813
814 /* Protocol checksum field is blank in the packet (hardware offload)? */
815 #define _XEN_NETTXF_csum_blank (0)
816 #define XEN_NETTXF_csum_blank (1U<<_XEN_NETTXF_csum_blank)
817
818 /* Packet data has been validated against protocol checksum. */
819 #define _XEN_NETTXF_data_validated (1)
820 #define XEN_NETTXF_data_validated (1U<<_XEN_NETTXF_data_validated)
821
822 /* Packet continues in the next request descriptor. */
823 #define _XEN_NETTXF_more_data (2)
824 #define XEN_NETTXF_more_data (1U<<_XEN_NETTXF_more_data)
825
826 /* Packet to be followed by extra descriptor(s). */
827 #define _XEN_NETTXF_extra_info (3)
828 #define XEN_NETTXF_extra_info (1U<<_XEN_NETTXF_extra_info)
829
830 #define XEN_NETIF_MAX_TX_SIZE 0xFFFF
831 struct xen_netif_tx_request {
832 grant_ref_t gref;
833 uint16_t offset;
834 uint16_t flags;
835 uint16_t id;
836 uint16_t size;
837 };
838
839 /* Types of xen_netif_extra_info descriptors. */
840 #define XEN_NETIF_EXTRA_TYPE_NONE (0) /* Never used - invalid */
841 #define XEN_NETIF_EXTRA_TYPE_GSO (1) /* u.gso */
842 #define XEN_NETIF_EXTRA_TYPE_MCAST_ADD (2) /* u.mcast */
843 #define XEN_NETIF_EXTRA_TYPE_MCAST_DEL (3) /* u.mcast */
844 #define XEN_NETIF_EXTRA_TYPE_HASH (4) /* u.hash */
845 #define XEN_NETIF_EXTRA_TYPE_XDP (5) /* u.xdp */
846 #define XEN_NETIF_EXTRA_TYPE_MAX (6)
847
848 /* xen_netif_extra_info_t flags. */
849 #define _XEN_NETIF_EXTRA_FLAG_MORE (0)
850 #define XEN_NETIF_EXTRA_FLAG_MORE (1U<<_XEN_NETIF_EXTRA_FLAG_MORE)
851
852 /* GSO types */
853 #define XEN_NETIF_GSO_TYPE_NONE (0)
854 #define XEN_NETIF_GSO_TYPE_TCPV4 (1)
855 #define XEN_NETIF_GSO_TYPE_TCPV6 (2)
856
857 /*
858 * This structure needs to fit within both xen_netif_tx_request_t and
859 * xen_netif_rx_response_t for compatibility.
860 */
861 struct xen_netif_extra_info {
862 uint8_t type;
863 uint8_t flags;
864 union {
865 struct {
866 uint16_t size;
867 uint8_t type;
868 uint8_t pad;
869 uint16_t features;
870 } gso;
871 struct {
872 uint8_t addr[6];
873 } mcast;
874 struct {
875 uint8_t type;
876 uint8_t algorithm;
877 uint8_t value[4];
878 } hash;
879 struct {
880 uint16_t headroom;
881 uint16_t pad[2];
882 } xdp;
883 uint16_t pad[3];
884 } u;
885 };
886
887 struct xen_netif_tx_response {
888 uint16_t id;
889 int16_t status;
890 };
891
892 struct xen_netif_rx_request {
893 uint16_t id; /* Echoed in response message. */
894 uint16_t pad;
895 grant_ref_t gref;
896 };
897
898 /* Packet data has been validated against protocol checksum. */
899 #define _XEN_NETRXF_data_validated (0)
900 #define XEN_NETRXF_data_validated (1U<<_XEN_NETRXF_data_validated)
901
902 /* Protocol checksum field is blank in the packet (hardware offload)? */
903 #define _XEN_NETRXF_csum_blank (1)
904 #define XEN_NETRXF_csum_blank (1U<<_XEN_NETRXF_csum_blank)
905
906 /* Packet continues in the next request descriptor. */
907 #define _XEN_NETRXF_more_data (2)
908 #define XEN_NETRXF_more_data (1U<<_XEN_NETRXF_more_data)
909
910 /* Packet to be followed by extra descriptor(s). */
911 #define _XEN_NETRXF_extra_info (3)
912 #define XEN_NETRXF_extra_info (1U<<_XEN_NETRXF_extra_info)
913
914 /* Packet has GSO prefix. Deprecated but included for compatibility */
915 #define _XEN_NETRXF_gso_prefix (4)
916 #define XEN_NETRXF_gso_prefix (1U<<_XEN_NETRXF_gso_prefix)
917
918 struct xen_netif_rx_response {
919 uint16_t id;
920 uint16_t offset;
921 uint16_t flags;
922 int16_t status;
923 };
924
925 /*
926 * Generate xen_netif ring structures and types.
927 */
928
929 DEFINE_RING_TYPES(xen_netif_tx, struct xen_netif_tx_request,
930 struct xen_netif_tx_response);
931 DEFINE_RING_TYPES(xen_netif_rx, struct xen_netif_rx_request,
932 struct xen_netif_rx_response);
933
934 #define XEN_NETIF_RSP_DROPPED -2
935 #define XEN_NETIF_RSP_ERROR -1
936 #define XEN_NETIF_RSP_OKAY 0
937 /* No response: used for auxiliary requests (e.g., xen_netif_extra_info_t). */
938 #define XEN_NETIF_RSP_NULL 1
939
940 #endif
941