1 /* SPDX-License-Identifier: GPL-2.0-or-later */ 2 /**************************************************************************/ 3 /* */ 4 /* IBM System i and System p Virtual NIC Device Driver */ 5 /* Copyright (C) 2014 IBM Corp. */ 6 /* Santiago Leon (santi_leon@yahoo.com) */ 7 /* Thomas Falcon (tlfalcon@linux.vnet.ibm.com) */ 8 /* John Allen (jallen@linux.vnet.ibm.com) */ 9 /* */ 10 /* */ 11 /* This module contains the implementation of a virtual ethernet device */ 12 /* for use with IBM i/pSeries LPAR Linux. It utilizes the logical LAN */ 13 /* option of the RS/6000 Platform Architecture to interface with virtual */ 14 /* ethernet NICs that are presented to the partition by the hypervisor. */ 15 /* */ 16 /**************************************************************************/ 17 18 #define IBMVNIC_NAME "ibmvnic" 19 #define IBMVNIC_DRIVER_VERSION "1.0.1" 20 #define IBMVNIC_INVALID_MAP -1 21 #define IBMVNIC_OPEN_FAILED 3 22 23 /* basic structures plus 100 2k buffers */ 24 #define IBMVNIC_IO_ENTITLEMENT_DEFAULT 610305 25 26 /* Initial module_parameters */ 27 #define IBMVNIC_RX_WEIGHT 16 28 /* when changing this, update IBMVNIC_IO_ENTITLEMENT_DEFAULT */ 29 #define IBMVNIC_BUFFS_PER_POOL 100 30 #define IBMVNIC_MAX_QUEUES 16 31 #define IBMVNIC_MAX_QUEUE_SZ 4096 32 #define IBMVNIC_MAX_IND_DESCS 16 33 #define IBMVNIC_IND_ARR_SZ (IBMVNIC_MAX_IND_DESCS * 32) 34 35 #define IBMVNIC_TSO_BUF_SZ 65536 36 #define IBMVNIC_TSO_BUFS 64 37 #define IBMVNIC_TSO_POOL_MASK 0x80000000 38 39 /* A VNIC adapter has set of Rx and Tx pools (aka queues). Each Rx/Tx pool 40 * has a set of buffers. The size of each buffer is determined by the MTU. 41 * 42 * Each Rx/Tx pool is also associated with a DMA region that is shared 43 * with the "hardware" (VIOS) and used to send/receive packets. The DMA 44 * region is also referred to as a Long Term Buffer or LTB. 45 * 46 * The size of the DMA region required for an Rx/Tx pool depends on the 47 * number and size (MTU) of the buffers in the pool. At the max levels 48 * of 4096 jumbo frames (MTU=9000) we will need about 9K*4K = 36MB plus 49 * some padding. 50 * 51 * But the size of a single DMA region is limited by MAX_ORDER in the 52 * kernel (about 16MB currently). To support say 4K Jumbo frames, we 53 * use a set of LTBs (struct ltb_set) per pool. 54 * 55 * IBMVNIC_ONE_LTB_MAX - max size of each LTB supported by kernel 56 * IBMVNIC_ONE_LTB_SIZE - current max size of each LTB in an ltb_set 57 * (must be <= IBMVNIC_ONE_LTB_MAX) 58 * IBMVNIC_LTB_SET_SIZE - current size of all LTBs in an ltb_set 59 * 60 * Each VNIC can have upto 16 Rx, 16 Tx and 16 TSO pools. The TSO pools 61 * are of fixed length (IBMVNIC_TSO_BUF_SZ * IBMVNIC_TSO_BUFS) of 4MB. 62 * 63 * The Rx and Tx pools can have upto 4096 buffers. The max size of these 64 * buffers is about 9588 (for jumbo frames, including IBMVNIC_BUFFER_HLEN). 65 * So, setting the IBMVNIC_LTB_SET_SIZE for a pool to 4096 * 9588 ~= 38MB. 66 * 67 * There is a trade-off in setting IBMVNIC_ONE_LTB_SIZE. If it is large, 68 * the allocation of the LTB can fail when system is low in memory. If 69 * its too small, we would need several mappings for each of the Rx/ 70 * Tx/TSO pools but there is a limit of 255 mappings per vnic in the 71 * VNIC protocol. 72 * 73 * So setting IBMVNIC_ONE_LTB_SIZE to 8MB. With IBMVNIC_LTB_SET_SIZE set 74 * to 38MB, we will need 5 LTBs per Rx and Tx pool and 1 LTB per TSO 75 * pool for the 4MB. Thus the 16 Rx and Tx queues require 32 * 5 = 160 76 * plus 16 for the TSO pools for a total of 176 LTB mappings per VNIC. 77 */ 78 #define IBMVNIC_ONE_LTB_MAX ((u32)((1 << (MAX_ORDER - 1)) * PAGE_SIZE)) 79 #define IBMVNIC_ONE_LTB_SIZE min((u32)(8 << 20), IBMVNIC_ONE_LTB_MAX) 80 #define IBMVNIC_LTB_SET_SIZE (38 << 20) 81 82 #define IBMVNIC_BUFFER_HLEN 500 83 #define IBMVNIC_RESET_DELAY 100 84 85 struct ibmvnic_login_buffer { 86 __be32 len; 87 __be32 version; 88 #define INITIAL_VERSION_LB 1 89 __be32 num_txcomp_subcrqs; 90 __be32 off_txcomp_subcrqs; 91 __be32 num_rxcomp_subcrqs; 92 __be32 off_rxcomp_subcrqs; 93 __be32 login_rsp_ioba; 94 __be32 login_rsp_len; 95 __be32 client_data_offset; 96 __be32 client_data_len; 97 } __packed __aligned(8); 98 99 struct ibmvnic_login_rsp_buffer { 100 __be32 len; 101 __be32 version; 102 #define INITIAL_VERSION_LRB 1 103 __be32 num_txsubm_subcrqs; 104 __be32 off_txsubm_subcrqs; 105 __be32 num_rxadd_subcrqs; 106 __be32 off_rxadd_subcrqs; 107 __be32 off_rxadd_buff_size; 108 __be32 num_supp_tx_desc; 109 __be32 off_supp_tx_desc; 110 } __packed __aligned(8); 111 112 struct ibmvnic_query_ip_offload_buffer { 113 __be32 len; 114 __be32 version; 115 #define INITIAL_VERSION_IOB 1 116 u8 ipv4_chksum; 117 u8 ipv6_chksum; 118 u8 tcp_ipv4_chksum; 119 u8 tcp_ipv6_chksum; 120 u8 udp_ipv4_chksum; 121 u8 udp_ipv6_chksum; 122 u8 large_tx_ipv4; 123 u8 large_tx_ipv6; 124 u8 large_rx_ipv4; 125 u8 large_rx_ipv6; 126 u8 reserved1[14]; 127 __be16 max_ipv4_header_size; 128 __be16 max_ipv6_header_size; 129 __be16 max_tcp_header_size; 130 __be16 max_udp_header_size; 131 __be32 max_large_tx_size; 132 __be32 max_large_rx_size; 133 u8 reserved2[16]; 134 u8 ipv6_extension_header; 135 #define IPV6_EH_NOT_SUPPORTED 0x00 136 #define IPV6_EH_SUPPORTED_LIM 0x01 137 #define IPV6_EH_SUPPORTED 0xFF 138 u8 tcp_pseudosum_req; 139 #define TCP_PS_NOT_REQUIRED 0x00 140 #define TCP_PS_REQUIRED 0x01 141 u8 reserved3[30]; 142 __be16 num_ipv6_ext_headers; 143 __be32 off_ipv6_ext_headers; 144 u8 reserved4[154]; 145 } __packed __aligned(8); 146 147 struct ibmvnic_control_ip_offload_buffer { 148 __be32 len; 149 __be32 version; 150 #define INITIAL_VERSION_IOB 1 151 u8 ipv4_chksum; 152 u8 ipv6_chksum; 153 u8 tcp_ipv4_chksum; 154 u8 tcp_ipv6_chksum; 155 u8 udp_ipv4_chksum; 156 u8 udp_ipv6_chksum; 157 u8 large_tx_ipv4; 158 u8 large_tx_ipv6; 159 u8 bad_packet_rx; 160 u8 large_rx_ipv4; 161 u8 large_rx_ipv6; 162 u8 reserved4[111]; 163 } __packed __aligned(8); 164 165 struct ibmvnic_fw_component { 166 u8 name[48]; 167 __be32 trace_buff_size; 168 u8 correlator; 169 u8 trace_level; 170 u8 parent_correlator; 171 u8 error_check_level; 172 u8 trace_on; 173 u8 reserved[7]; 174 u8 description[192]; 175 } __packed __aligned(8); 176 177 struct ibmvnic_fw_trace_entry { 178 __be32 trace_id; 179 u8 num_valid_data; 180 u8 reserved[3]; 181 __be64 pmc_registers; 182 __be64 timebase; 183 __be64 trace_data[5]; 184 } __packed __aligned(8); 185 186 struct ibmvnic_statistics { 187 __be32 version; 188 __be32 promiscuous; 189 __be64 rx_packets; 190 __be64 rx_bytes; 191 __be64 tx_packets; 192 __be64 tx_bytes; 193 __be64 ucast_tx_packets; 194 __be64 ucast_rx_packets; 195 __be64 mcast_tx_packets; 196 __be64 mcast_rx_packets; 197 __be64 bcast_tx_packets; 198 __be64 bcast_rx_packets; 199 __be64 align_errors; 200 __be64 fcs_errors; 201 __be64 single_collision_frames; 202 __be64 multi_collision_frames; 203 __be64 sqe_test_errors; 204 __be64 deferred_tx; 205 __be64 late_collisions; 206 __be64 excess_collisions; 207 __be64 internal_mac_tx_errors; 208 __be64 carrier_sense; 209 __be64 too_long_frames; 210 __be64 internal_mac_rx_errors; 211 u8 reserved[72]; 212 } __packed __aligned(8); 213 214 #define NUM_TX_STATS 3 215 struct ibmvnic_tx_queue_stats { 216 u64 packets; 217 u64 bytes; 218 u64 dropped_packets; 219 }; 220 221 #define NUM_RX_STATS 3 222 struct ibmvnic_rx_queue_stats { 223 u64 packets; 224 u64 bytes; 225 u64 interrupts; 226 }; 227 228 struct ibmvnic_acl_buffer { 229 __be32 len; 230 __be32 version; 231 #define INITIAL_VERSION_IOB 1 232 u8 mac_acls_restrict; 233 u8 vlan_acls_restrict; 234 u8 reserved1[22]; 235 __be32 num_mac_addrs; 236 __be32 offset_mac_addrs; 237 __be32 num_vlan_ids; 238 __be32 offset_vlan_ids; 239 u8 reserved2[80]; 240 } __packed __aligned(8); 241 242 /* descriptors have been changed, how should this be defined? 1? 4? */ 243 244 #define IBMVNIC_TX_DESC_VERSIONS 3 245 246 /* is this still needed? */ 247 struct ibmvnic_tx_comp_desc { 248 u8 first; 249 u8 num_comps; 250 __be16 rcs[5]; 251 __be32 correlators[5]; 252 } __packed __aligned(8); 253 254 /* some flags that included in v0 descriptor, which is gone 255 * only used for IBMVNIC_TCP_CHKSUM and IBMVNIC_UDP_CHKSUM 256 * and only in some offload_flags variable that doesn't seem 257 * to be used anywhere, can probably be removed? 258 */ 259 260 #define IBMVNIC_TCP_CHKSUM 0x20 261 #define IBMVNIC_UDP_CHKSUM 0x08 262 263 struct ibmvnic_tx_desc { 264 u8 first; 265 u8 type; 266 267 #define IBMVNIC_TX_DESC 0x10 268 u8 n_crq_elem; 269 u8 n_sge; 270 u8 flags1; 271 #define IBMVNIC_TX_COMP_NEEDED 0x80 272 #define IBMVNIC_TX_CHKSUM_OFFLOAD 0x40 273 #define IBMVNIC_TX_LSO 0x20 274 #define IBMVNIC_TX_PROT_TCP 0x10 275 #define IBMVNIC_TX_PROT_UDP 0x08 276 #define IBMVNIC_TX_PROT_IPV4 0x04 277 #define IBMVNIC_TX_PROT_IPV6 0x02 278 #define IBMVNIC_TX_VLAN_PRESENT 0x01 279 u8 flags2; 280 #define IBMVNIC_TX_VLAN_INSERT 0x80 281 __be16 mss; 282 u8 reserved[4]; 283 __be32 correlator; 284 __be16 vlan_id; 285 __be16 dma_reg; 286 __be32 sge_len; 287 __be64 ioba; 288 } __packed __aligned(8); 289 290 struct ibmvnic_hdr_desc { 291 u8 first; 292 u8 type; 293 #define IBMVNIC_HDR_DESC 0x11 294 u8 len; 295 u8 l2_len; 296 __be16 l3_len; 297 u8 l4_len; 298 u8 flag; 299 u8 data[24]; 300 } __packed __aligned(8); 301 302 struct ibmvnic_hdr_ext_desc { 303 u8 first; 304 u8 type; 305 #define IBMVNIC_HDR_EXT_DESC 0x12 306 u8 len; 307 u8 data[29]; 308 } __packed __aligned(8); 309 310 struct ibmvnic_sge_desc { 311 u8 first; 312 u8 type; 313 #define IBMVNIC_SGE_DESC 0x30 314 __be16 sge1_dma_reg; 315 __be32 sge1_len; 316 __be64 sge1_ioba; 317 __be16 reserved; 318 __be16 sge2_dma_reg; 319 __be32 sge2_len; 320 __be64 sge2_ioba; 321 } __packed __aligned(8); 322 323 struct ibmvnic_rx_comp_desc { 324 u8 first; 325 u8 flags; 326 #define IBMVNIC_IP_CHKSUM_GOOD 0x80 327 #define IBMVNIC_TCP_UDP_CHKSUM_GOOD 0x40 328 #define IBMVNIC_END_FRAME 0x20 329 #define IBMVNIC_EXACT_MC 0x10 330 #define IBMVNIC_VLAN_STRIPPED 0x08 331 __be16 off_frame_data; 332 __be32 len; 333 __be64 correlator; 334 __be16 vlan_tci; 335 __be16 rc; 336 u8 reserved[12]; 337 } __packed __aligned(8); 338 339 struct ibmvnic_generic_scrq { 340 u8 first; 341 u8 reserved[31]; 342 } __packed __aligned(8); 343 344 struct ibmvnic_rx_buff_add_desc { 345 u8 first; 346 u8 reserved[7]; 347 __be64 correlator; 348 __be32 ioba; 349 u8 map_id; 350 __be32 len:24; 351 u8 reserved2[8]; 352 } __packed __aligned(8); 353 354 struct ibmvnic_rc { 355 u8 code; /* one of enum ibmvnic_rc_codes */ 356 u8 detailed_data[3]; 357 } __packed __aligned(4); 358 359 struct ibmvnic_generic_crq { 360 u8 first; 361 u8 cmd; 362 u8 params[10]; 363 struct ibmvnic_rc rc; 364 } __packed __aligned(8); 365 366 struct ibmvnic_version_exchange { 367 u8 first; 368 u8 cmd; 369 __be16 version; 370 #define IBMVNIC_INITIAL_VERSION 1 371 u8 reserved[8]; 372 struct ibmvnic_rc rc; 373 } __packed __aligned(8); 374 375 struct ibmvnic_capability { 376 u8 first; 377 u8 cmd; 378 __be16 capability; /* one of ibmvnic_capabilities */ 379 __be64 number; 380 struct ibmvnic_rc rc; 381 } __packed __aligned(8); 382 383 struct ibmvnic_login { 384 u8 first; 385 u8 cmd; 386 u8 reserved[6]; 387 __be32 ioba; 388 __be32 len; 389 } __packed __aligned(8); 390 391 struct ibmvnic_phys_parms { 392 u8 first; 393 u8 cmd; 394 u8 flags1; 395 #define IBMVNIC_EXTERNAL_LOOPBACK 0x80 396 #define IBMVNIC_INTERNAL_LOOPBACK 0x40 397 #define IBMVNIC_PROMISC 0x20 398 #define IBMVNIC_PHYS_LINK_ACTIVE 0x10 399 #define IBMVNIC_AUTONEG_DUPLEX 0x08 400 #define IBMVNIC_FULL_DUPLEX 0x04 401 #define IBMVNIC_HALF_DUPLEX 0x02 402 #define IBMVNIC_CAN_CHG_PHYS_PARMS 0x01 403 u8 flags2; 404 #define IBMVNIC_LOGICAL_LNK_ACTIVE 0x80 405 __be32 speed; 406 #define IBMVNIC_AUTONEG 0x80000000 407 #define IBMVNIC_10MBPS 0x40000000 408 #define IBMVNIC_100MBPS 0x20000000 409 #define IBMVNIC_1GBPS 0x10000000 410 #define IBMVNIC_10GBPS 0x08000000 411 #define IBMVNIC_40GBPS 0x04000000 412 #define IBMVNIC_100GBPS 0x02000000 413 #define IBMVNIC_25GBPS 0x01000000 414 #define IBMVNIC_50GBPS 0x00800000 415 #define IBMVNIC_200GBPS 0x00400000 416 __be32 mtu; 417 struct ibmvnic_rc rc; 418 } __packed __aligned(8); 419 420 struct ibmvnic_logical_link_state { 421 u8 first; 422 u8 cmd; 423 u8 link_state; 424 #define IBMVNIC_LOGICAL_LNK_DN 0x00 425 #define IBMVNIC_LOGICAL_LNK_UP 0x01 426 #define IBMVNIC_LOGICAL_LNK_QUERY 0xff 427 u8 reserved[9]; 428 struct ibmvnic_rc rc; 429 } __packed __aligned(8); 430 431 struct ibmvnic_query_ip_offload { 432 u8 first; 433 u8 cmd; 434 u8 reserved[2]; 435 __be32 len; 436 __be32 ioba; 437 struct ibmvnic_rc rc; 438 } __packed __aligned(8); 439 440 struct ibmvnic_control_ip_offload { 441 u8 first; 442 u8 cmd; 443 u8 reserved[2]; 444 __be32 ioba; 445 __be32 len; 446 struct ibmvnic_rc rc; 447 } __packed __aligned(8); 448 449 struct ibmvnic_request_statistics { 450 u8 first; 451 u8 cmd; 452 u8 flags; 453 #define IBMVNIC_PHYSICAL_PORT 0x80 454 u8 reserved1; 455 __be32 ioba; 456 __be32 len; 457 u8 reserved[4]; 458 } __packed __aligned(8); 459 460 struct ibmvnic_error_indication { 461 u8 first; 462 u8 cmd; 463 u8 flags; 464 #define IBMVNIC_FATAL_ERROR 0x80 465 u8 reserved1; 466 __be32 error_id; 467 __be32 detail_error_sz; 468 __be16 error_cause; 469 u8 reserved2[2]; 470 } __packed __aligned(8); 471 472 struct ibmvnic_link_state_indication { 473 u8 first; 474 u8 cmd; 475 u8 reserved1[2]; 476 u8 phys_link_state; 477 u8 logical_link_state; 478 u8 reserved2[10]; 479 } __packed __aligned(8); 480 481 struct ibmvnic_change_mac_addr { 482 u8 first; 483 u8 cmd; 484 u8 mac_addr[6]; 485 u8 reserved[4]; 486 struct ibmvnic_rc rc; 487 } __packed __aligned(8); 488 489 struct ibmvnic_multicast_ctrl { 490 u8 first; 491 u8 cmd; 492 u8 mac_addr[6]; 493 u8 flags; 494 #define IBMVNIC_ENABLE_MC 0x80 495 #define IBMVNIC_DISABLE_MC 0x40 496 #define IBMVNIC_ENABLE_ALL 0x20 497 #define IBMVNIC_DISABLE_ALL 0x10 498 u8 reserved1; 499 __be16 reserved2; /* was num_enabled_mc_addr; */ 500 struct ibmvnic_rc rc; 501 } __packed __aligned(8); 502 503 struct ibmvnic_get_vpd_size { 504 u8 first; 505 u8 cmd; 506 u8 reserved[14]; 507 } __packed __aligned(8); 508 509 struct ibmvnic_get_vpd_size_rsp { 510 u8 first; 511 u8 cmd; 512 u8 reserved[2]; 513 __be64 len; 514 struct ibmvnic_rc rc; 515 } __packed __aligned(8); 516 517 struct ibmvnic_get_vpd { 518 u8 first; 519 u8 cmd; 520 u8 reserved1[2]; 521 __be32 ioba; 522 __be32 len; 523 u8 reserved[4]; 524 } __packed __aligned(8); 525 526 struct ibmvnic_get_vpd_rsp { 527 u8 first; 528 u8 cmd; 529 u8 reserved[10]; 530 struct ibmvnic_rc rc; 531 } __packed __aligned(8); 532 533 struct ibmvnic_acl_change_indication { 534 u8 first; 535 u8 cmd; 536 __be16 change_type; 537 #define IBMVNIC_MAC_ACL 0 538 #define IBMVNIC_VLAN_ACL 1 539 u8 reserved[12]; 540 } __packed __aligned(8); 541 542 struct ibmvnic_acl_query { 543 u8 first; 544 u8 cmd; 545 u8 reserved1[2]; 546 __be32 ioba; 547 __be32 len; 548 u8 reserved2[4]; 549 } __packed __aligned(8); 550 551 struct ibmvnic_tune { 552 u8 first; 553 u8 cmd; 554 u8 reserved1[2]; 555 __be32 ioba; 556 __be32 len; 557 u8 reserved2[4]; 558 } __packed __aligned(8); 559 560 struct ibmvnic_request_map { 561 u8 first; 562 u8 cmd; 563 u8 reserved1; 564 u8 map_id; 565 __be32 ioba; 566 __be32 len; 567 u8 reserved2[4]; 568 } __packed __aligned(8); 569 570 struct ibmvnic_request_map_rsp { 571 u8 first; 572 u8 cmd; 573 u8 reserved1; 574 u8 map_id; 575 u8 reserved2[8]; 576 struct ibmvnic_rc rc; 577 } __packed __aligned(8); 578 579 struct ibmvnic_request_unmap { 580 u8 first; 581 u8 cmd; 582 u8 reserved1; 583 u8 map_id; 584 u8 reserved2[12]; 585 } __packed __aligned(8); 586 587 struct ibmvnic_request_unmap_rsp { 588 u8 first; 589 u8 cmd; 590 u8 reserved1; 591 u8 map_id; 592 u8 reserved2[8]; 593 struct ibmvnic_rc rc; 594 } __packed __aligned(8); 595 596 struct ibmvnic_query_map { 597 u8 first; 598 u8 cmd; 599 u8 reserved[14]; 600 } __packed __aligned(8); 601 602 struct ibmvnic_query_map_rsp { 603 u8 first; 604 u8 cmd; 605 u8 reserved; 606 u8 page_size; 607 __be32 tot_pages; 608 __be32 free_pages; 609 struct ibmvnic_rc rc; 610 } __packed __aligned(8); 611 612 union ibmvnic_crq { 613 struct ibmvnic_generic_crq generic; 614 struct ibmvnic_version_exchange version_exchange; 615 struct ibmvnic_version_exchange version_exchange_rsp; 616 struct ibmvnic_capability query_capability; 617 struct ibmvnic_capability query_capability_rsp; 618 struct ibmvnic_capability request_capability; 619 struct ibmvnic_capability request_capability_rsp; 620 struct ibmvnic_login login; 621 struct ibmvnic_generic_crq login_rsp; 622 struct ibmvnic_phys_parms query_phys_parms; 623 struct ibmvnic_phys_parms query_phys_parms_rsp; 624 struct ibmvnic_phys_parms query_phys_capabilities; 625 struct ibmvnic_phys_parms query_phys_capabilities_rsp; 626 struct ibmvnic_phys_parms set_phys_parms; 627 struct ibmvnic_phys_parms set_phys_parms_rsp; 628 struct ibmvnic_logical_link_state logical_link_state; 629 struct ibmvnic_logical_link_state logical_link_state_rsp; 630 struct ibmvnic_query_ip_offload query_ip_offload; 631 struct ibmvnic_query_ip_offload query_ip_offload_rsp; 632 struct ibmvnic_control_ip_offload control_ip_offload; 633 struct ibmvnic_control_ip_offload control_ip_offload_rsp; 634 struct ibmvnic_request_statistics request_statistics; 635 struct ibmvnic_generic_crq request_statistics_rsp; 636 struct ibmvnic_error_indication error_indication; 637 struct ibmvnic_link_state_indication link_state_indication; 638 struct ibmvnic_change_mac_addr change_mac_addr; 639 struct ibmvnic_change_mac_addr change_mac_addr_rsp; 640 struct ibmvnic_multicast_ctrl multicast_ctrl; 641 struct ibmvnic_multicast_ctrl multicast_ctrl_rsp; 642 struct ibmvnic_get_vpd_size get_vpd_size; 643 struct ibmvnic_get_vpd_size_rsp get_vpd_size_rsp; 644 struct ibmvnic_get_vpd get_vpd; 645 struct ibmvnic_get_vpd_rsp get_vpd_rsp; 646 struct ibmvnic_acl_change_indication acl_change_indication; 647 struct ibmvnic_acl_query acl_query; 648 struct ibmvnic_generic_crq acl_query_rsp; 649 struct ibmvnic_tune tune; 650 struct ibmvnic_generic_crq tune_rsp; 651 struct ibmvnic_request_map request_map; 652 struct ibmvnic_request_map_rsp request_map_rsp; 653 struct ibmvnic_request_unmap request_unmap; 654 struct ibmvnic_request_unmap_rsp request_unmap_rsp; 655 struct ibmvnic_query_map query_map; 656 struct ibmvnic_query_map_rsp query_map_rsp; 657 }; 658 659 enum ibmvnic_rc_codes { 660 SUCCESS = 0, 661 PARTIALSUCCESS = 1, 662 PERMISSION = 2, 663 NOMEMORY = 3, 664 PARAMETER = 4, 665 UNKNOWNCOMMAND = 5, 666 ABORTED = 6, 667 INVALIDSTATE = 7, 668 INVALIDIOBA = 8, 669 INVALIDLENGTH = 9, 670 UNSUPPORTEDOPTION = 10, 671 }; 672 673 enum ibmvnic_capabilities { 674 MIN_TX_QUEUES = 1, 675 MIN_RX_QUEUES = 2, 676 MIN_RX_ADD_QUEUES = 3, 677 MAX_TX_QUEUES = 4, 678 MAX_RX_QUEUES = 5, 679 MAX_RX_ADD_QUEUES = 6, 680 REQ_TX_QUEUES = 7, 681 REQ_RX_QUEUES = 8, 682 REQ_RX_ADD_QUEUES = 9, 683 MIN_TX_ENTRIES_PER_SUBCRQ = 10, 684 MIN_RX_ADD_ENTRIES_PER_SUBCRQ = 11, 685 MAX_TX_ENTRIES_PER_SUBCRQ = 12, 686 MAX_RX_ADD_ENTRIES_PER_SUBCRQ = 13, 687 REQ_TX_ENTRIES_PER_SUBCRQ = 14, 688 REQ_RX_ADD_ENTRIES_PER_SUBCRQ = 15, 689 TCP_IP_OFFLOAD = 16, 690 PROMISC_REQUESTED = 17, 691 PROMISC_SUPPORTED = 18, 692 MIN_MTU = 19, 693 MAX_MTU = 20, 694 REQ_MTU = 21, 695 MAX_MULTICAST_FILTERS = 22, 696 VLAN_HEADER_INSERTION = 23, 697 RX_VLAN_HEADER_INSERTION = 24, 698 MAX_TX_SG_ENTRIES = 25, 699 RX_SG_SUPPORTED = 26, 700 RX_SG_REQUESTED = 27, 701 OPT_TX_COMP_SUB_QUEUES = 28, 702 OPT_RX_COMP_QUEUES = 29, 703 OPT_RX_BUFADD_Q_PER_RX_COMP_Q = 30, 704 OPT_TX_ENTRIES_PER_SUBCRQ = 31, 705 OPT_RXBA_ENTRIES_PER_SUBCRQ = 32, 706 TX_RX_DESC_REQ = 33, 707 }; 708 709 enum ibmvnic_error_cause { 710 ADAPTER_PROBLEM = 0, 711 BUS_PROBLEM = 1, 712 FW_PROBLEM = 2, 713 DD_PROBLEM = 3, 714 EEH_RECOVERY = 4, 715 FW_UPDATED = 5, 716 LOW_MEMORY = 6, 717 }; 718 719 enum ibmvnic_commands { 720 VERSION_EXCHANGE = 0x01, 721 VERSION_EXCHANGE_RSP = 0x81, 722 QUERY_CAPABILITY = 0x02, 723 QUERY_CAPABILITY_RSP = 0x82, 724 REQUEST_CAPABILITY = 0x03, 725 REQUEST_CAPABILITY_RSP = 0x83, 726 LOGIN = 0x04, 727 LOGIN_RSP = 0x84, 728 QUERY_PHYS_PARMS = 0x05, 729 QUERY_PHYS_PARMS_RSP = 0x85, 730 QUERY_PHYS_CAPABILITIES = 0x06, 731 QUERY_PHYS_CAPABILITIES_RSP = 0x86, 732 SET_PHYS_PARMS = 0x07, 733 SET_PHYS_PARMS_RSP = 0x87, 734 ERROR_INDICATION = 0x08, 735 LOGICAL_LINK_STATE = 0x0C, 736 LOGICAL_LINK_STATE_RSP = 0x8C, 737 REQUEST_STATISTICS = 0x0D, 738 REQUEST_STATISTICS_RSP = 0x8D, 739 COLLECT_FW_TRACE = 0x11, 740 COLLECT_FW_TRACE_RSP = 0x91, 741 LINK_STATE_INDICATION = 0x12, 742 CHANGE_MAC_ADDR = 0x13, 743 CHANGE_MAC_ADDR_RSP = 0x93, 744 MULTICAST_CTRL = 0x14, 745 MULTICAST_CTRL_RSP = 0x94, 746 GET_VPD_SIZE = 0x15, 747 GET_VPD_SIZE_RSP = 0x95, 748 GET_VPD = 0x16, 749 GET_VPD_RSP = 0x96, 750 TUNE = 0x17, 751 TUNE_RSP = 0x97, 752 QUERY_IP_OFFLOAD = 0x18, 753 QUERY_IP_OFFLOAD_RSP = 0x98, 754 CONTROL_IP_OFFLOAD = 0x19, 755 CONTROL_IP_OFFLOAD_RSP = 0x99, 756 ACL_CHANGE_INDICATION = 0x1A, 757 ACL_QUERY = 0x1B, 758 ACL_QUERY_RSP = 0x9B, 759 QUERY_MAP = 0x1D, 760 QUERY_MAP_RSP = 0x9D, 761 REQUEST_MAP = 0x1E, 762 REQUEST_MAP_RSP = 0x9E, 763 REQUEST_UNMAP = 0x1F, 764 REQUEST_UNMAP_RSP = 0x9F, 765 VLAN_CTRL = 0x20, 766 VLAN_CTRL_RSP = 0xA0, 767 }; 768 769 enum ibmvnic_crq_type { 770 IBMVNIC_CRQ_CMD = 0x80, 771 IBMVNIC_CRQ_CMD_RSP = 0x80, 772 IBMVNIC_CRQ_INIT_CMD = 0xC0, 773 IBMVNIC_CRQ_INIT_RSP = 0xC0, 774 IBMVNIC_CRQ_XPORT_EVENT = 0xFF, 775 }; 776 777 enum ibmvfc_crq_format { 778 IBMVNIC_CRQ_INIT = 0x01, 779 IBMVNIC_CRQ_INIT_COMPLETE = 0x02, 780 IBMVNIC_PARTITION_MIGRATED = 0x06, 781 IBMVNIC_DEVICE_FAILOVER = 0x08, 782 }; 783 784 struct ibmvnic_crq_queue { 785 union ibmvnic_crq *msgs; 786 int size, cur; 787 dma_addr_t msg_token; 788 /* Used for serialization of msgs, cur */ 789 spinlock_t lock; 790 bool active; 791 char name[32]; 792 }; 793 794 union sub_crq { 795 struct ibmvnic_generic_scrq generic; 796 struct ibmvnic_tx_comp_desc tx_comp; 797 struct ibmvnic_tx_desc v1; 798 struct ibmvnic_hdr_desc hdr; 799 struct ibmvnic_hdr_ext_desc hdr_ext; 800 struct ibmvnic_sge_desc sge; 801 struct ibmvnic_rx_comp_desc rx_comp; 802 struct ibmvnic_rx_buff_add_desc rx_add; 803 }; 804 805 struct ibmvnic_ind_xmit_queue { 806 union sub_crq *indir_arr; 807 dma_addr_t indir_dma; 808 int index; 809 }; 810 811 struct ibmvnic_sub_crq_queue { 812 union sub_crq *msgs; 813 int size, cur; 814 dma_addr_t msg_token; 815 unsigned long crq_num; 816 unsigned long hw_irq; 817 unsigned int irq; 818 unsigned int pool_index; 819 int scrq_num; 820 /* Used for serialization of msgs, cur */ 821 spinlock_t lock; 822 struct sk_buff *rx_skb_top; 823 struct ibmvnic_adapter *adapter; 824 struct ibmvnic_ind_xmit_queue ind_buf; 825 atomic_t used; 826 char name[32]; 827 u64 handle; 828 } ____cacheline_aligned; 829 830 struct ibmvnic_long_term_buff { 831 unsigned char *buff; 832 dma_addr_t addr; 833 u64 size; 834 u8 map_id; 835 }; 836 837 struct ibmvnic_ltb_set { 838 int num_ltbs; 839 struct ibmvnic_long_term_buff *ltbs; 840 }; 841 842 struct ibmvnic_tx_buff { 843 struct sk_buff *skb; 844 int index; 845 int pool_index; 846 int num_entries; 847 }; 848 849 struct ibmvnic_tx_pool { 850 struct ibmvnic_tx_buff *tx_buff; 851 int *free_map; 852 int consumer_index; 853 int producer_index; 854 struct ibmvnic_ltb_set ltb_set; 855 int num_buffers; 856 int buf_size; 857 } ____cacheline_aligned; 858 859 struct ibmvnic_rx_buff { 860 struct sk_buff *skb; 861 dma_addr_t dma; 862 unsigned char *data; 863 int size; 864 int pool_index; 865 }; 866 867 struct ibmvnic_rx_pool { 868 struct ibmvnic_rx_buff *rx_buff; 869 int size; /* # of buffers in the pool */ 870 int index; 871 int buff_size; 872 atomic_t available; 873 int *free_map; 874 int next_free; 875 int next_alloc; 876 int active; 877 struct ibmvnic_ltb_set ltb_set; 878 } ____cacheline_aligned; 879 880 struct ibmvnic_vpd { 881 unsigned char *buff; 882 dma_addr_t dma_addr; 883 u64 len; 884 }; 885 886 enum vnic_state {VNIC_PROBING = 1, 887 VNIC_PROBED, 888 VNIC_OPENING, 889 VNIC_OPEN, 890 VNIC_CLOSING, 891 VNIC_CLOSED, 892 VNIC_REMOVING, 893 VNIC_REMOVED, 894 VNIC_DOWN}; 895 896 enum ibmvnic_reset_reason {VNIC_RESET_FAILOVER = 1, 897 VNIC_RESET_MOBILITY, 898 VNIC_RESET_FATAL, 899 VNIC_RESET_NON_FATAL, 900 VNIC_RESET_TIMEOUT, 901 VNIC_RESET_CHANGE_PARAM, 902 VNIC_RESET_PASSIVE_INIT}; 903 904 struct ibmvnic_rwi { 905 enum ibmvnic_reset_reason reset_reason; 906 struct list_head list; 907 }; 908 909 struct ibmvnic_tunables { 910 u64 rx_queues; 911 u64 tx_queues; 912 u64 rx_entries; 913 u64 tx_entries; 914 u64 mtu; 915 }; 916 917 struct ibmvnic_adapter { 918 struct vio_dev *vdev; 919 struct net_device *netdev; 920 struct ibmvnic_crq_queue crq; 921 u8 mac_addr[ETH_ALEN]; 922 struct ibmvnic_query_ip_offload_buffer ip_offload_buf; 923 dma_addr_t ip_offload_tok; 924 struct ibmvnic_control_ip_offload_buffer ip_offload_ctrl; 925 dma_addr_t ip_offload_ctrl_tok; 926 u32 msg_enable; 927 928 /* Vital Product Data (VPD) */ 929 struct ibmvnic_vpd *vpd; 930 char fw_version[32]; 931 932 /* Statistics */ 933 struct ibmvnic_statistics stats; 934 dma_addr_t stats_token; 935 struct completion stats_done; 936 int replenish_no_mem; 937 int replenish_add_buff_success; 938 int replenish_add_buff_failure; 939 int replenish_task_cycles; 940 int tx_send_failed; 941 int tx_map_failed; 942 943 struct ibmvnic_tx_queue_stats *tx_stats_buffers; 944 struct ibmvnic_rx_queue_stats *rx_stats_buffers; 945 946 int phys_link_state; 947 int logical_link_state; 948 949 u32 speed; 950 u8 duplex; 951 952 /* login data */ 953 struct ibmvnic_login_buffer *login_buf; 954 dma_addr_t login_buf_token; 955 int login_buf_sz; 956 957 struct ibmvnic_login_rsp_buffer *login_rsp_buf; 958 dma_addr_t login_rsp_buf_token; 959 int login_rsp_buf_sz; 960 961 atomic_t running_cap_crqs; 962 963 struct ibmvnic_sub_crq_queue **tx_scrq ____cacheline_aligned; 964 struct ibmvnic_sub_crq_queue **rx_scrq ____cacheline_aligned; 965 966 /* rx structs */ 967 struct napi_struct *napi; 968 struct ibmvnic_rx_pool *rx_pool; 969 u64 promisc; 970 971 struct ibmvnic_tx_pool *tx_pool; 972 struct ibmvnic_tx_pool *tso_pool; 973 struct completion probe_done; 974 struct completion init_done; 975 int init_done_rc; 976 977 struct completion fw_done; 978 /* Used for serialization of device commands */ 979 struct mutex fw_lock; 980 int fw_done_rc; 981 982 struct completion reset_done; 983 int reset_done_rc; 984 bool wait_for_reset; 985 986 /* partner capabilities */ 987 u64 min_tx_queues; 988 u64 min_rx_queues; 989 u64 min_rx_add_queues; 990 u64 max_tx_queues; 991 u64 max_rx_queues; 992 u64 max_rx_add_queues; 993 u64 req_tx_queues; 994 u64 req_rx_queues; 995 u64 req_rx_add_queues; 996 u64 min_tx_entries_per_subcrq; 997 u64 min_rx_add_entries_per_subcrq; 998 u64 max_tx_entries_per_subcrq; 999 u64 max_rx_add_entries_per_subcrq; 1000 u64 req_tx_entries_per_subcrq; 1001 u64 req_rx_add_entries_per_subcrq; 1002 u64 tcp_ip_offload; 1003 u64 promisc_requested; 1004 u64 promisc_supported; 1005 u64 min_mtu; 1006 u64 max_mtu; 1007 u64 req_mtu; 1008 u64 prev_mtu; 1009 u64 max_multicast_filters; 1010 u64 vlan_header_insertion; 1011 u64 rx_vlan_header_insertion; 1012 u64 max_tx_sg_entries; 1013 u64 rx_sg_supported; 1014 u64 rx_sg_requested; 1015 u64 opt_tx_comp_sub_queues; 1016 u64 opt_rx_comp_queues; 1017 u64 opt_rx_bufadd_q_per_rx_comp_q; 1018 u64 opt_tx_entries_per_subcrq; 1019 u64 opt_rxba_entries_per_subcrq; 1020 __be64 tx_rx_desc_req; 1021 #define MAX_MAP_ID 255 1022 DECLARE_BITMAP(map_ids, MAX_MAP_ID); 1023 u32 num_active_rx_scrqs; 1024 u32 num_active_rx_pools; 1025 u32 num_active_rx_napi; 1026 u32 num_active_tx_scrqs; 1027 u32 num_active_tx_pools; 1028 1029 u32 prev_rx_pool_size; 1030 u32 prev_tx_pool_size; 1031 u32 cur_rx_buf_sz; 1032 u32 prev_rx_buf_sz; 1033 1034 struct tasklet_struct tasklet; 1035 enum vnic_state state; 1036 /* Used for serialization of state field. When taking both state 1037 * and rwi locks, take state lock first. 1038 */ 1039 spinlock_t state_lock; 1040 enum ibmvnic_reset_reason reset_reason; 1041 struct list_head rwi_list; 1042 /* Used for serialization of rwi_list. When taking both state 1043 * and rwi locks, take state lock first 1044 */ 1045 spinlock_t rwi_lock; 1046 struct work_struct ibmvnic_reset; 1047 struct delayed_work ibmvnic_delayed_reset; 1048 unsigned long resetting; 1049 /* last device reset time */ 1050 unsigned long last_reset_time; 1051 1052 bool napi_enabled; 1053 bool from_passive_init; 1054 bool login_pending; 1055 /* protected by rcu */ 1056 bool tx_queues_active; 1057 bool failover_pending; 1058 bool force_reset_recovery; 1059 1060 struct ibmvnic_tunables desired; 1061 struct ibmvnic_tunables fallback; 1062 }; 1063