11. BNXT Poll Mode Driver
The Broadcom BNXT PMD (librte_net_bnxt) implements support for adapters based on Ethernet controllers belonging to the Broadcom BCM5741X/BCM575XX NetXtreme-E® Family of Ethernet Network Controllers.
A complete list is in the Supported Network Adapters section.
11.1. CPU Support
BNXT PMD supports multiple CPU architectures, including x86-32, x86-64, and ARMv8.
11.2. Kernel Dependency
BNXT PMD requires a kernel module (VFIO or UIO) for setting up a device, mapping device memory to userspace, registering interrupts, etc. VFIO is more secure than UIO, relying on IOMMU protection. UIO requires the IOMMU disabled or configured to pass-through mode.
The BNXT PMD supports operating with:
Linux vfio-pci
Linux uio_pci_generic
Linux igb_uio
BSD nic_uio
11.3. Running BNXT PMD
Bind the device to one of the kernel modules listed above
./dpdk-devbind.py -b vfio-pci|igb_uio|uio_pci_generic bus_id:device_id.function_id
The BNXT PMD can run on PF or VF.
PCI-SIG Single Root I/O Virtualization (SR-IOV) involves the direct assignment of part of the network port resources to guest operating systems using the SR-IOV standard. NIC is logically distributed among multiple virtual machines (VMs), while still having global data in common to share with the PF and other VFs.
Sysadmin can create and configure VFs:
echo num_vfs > /sys/bus/pci/devices/domain_id:bus_id:device_id:function_id/sriov_numvfs
(ex) echo 4 > /sys/bus/pci/devices/0000:82:00:0/sriov_numvfs
Sysadmin also can change the VF property such as MAC address, transparent VLAN, TX rate limit, and trusted VF:
ip link set pf_id vf vf_id mac (mac_address) vlan (vlan_id) txrate (rate_value) trust (enable|disable)
(ex) ip link set 0 vf 0 mac 00:11:22:33:44:55 vlan 0x100 txrate 100 trust disable
11.3.1. Running on VF
11.3.1.1. Flow Bifurcation
The Flow Bifurcation splits the incoming data traffic to user space applications (such as DPDK applications) and/or kernel space programs (such as the Linux kernel stack). It can direct some traffic, for example data plane traffic, to DPDK. Rest of the traffic, for example control plane traffic, would be redirected to the traditional Linux networking stack.
Refer to Flow Bifurcation How-to Guide
Benefits of the flow bifurcation include:
Better performance with less CPU overhead, as user application can directly access the NIC for data path
NIC is still being controlled by the kernel, as control traffic is forwarded only to the kernel driver
Control commands, e.g. ethtool, will work as usual
Running on a VF, the BXNT PMD supports the flow bifurcation with a combination of SR-IOV and packet classification and/or forwarding capability. In the simplest case of flow bifurcation, a PF driver configures a NIC to forward all user traffic directly to VFs with matching destination MAC address, while the rest of the traffic is forwarded to a PF. Note that the broadcast packets will be forwarded to both PF and VF.
(ex) ethtool --config-ntuple ens2f0 flow-type ether dst 00:01:02:03:00:01 vlan 10 vlan-mask 0xf000 action 0x100000000
11.3.1.2. Trusted VF
By default, VFs are not allowed to perform privileged operations, such as modifying the VF’s MAC address in the guest. These security measures are designed to prevent possible attacks. However, when a DPDK application can be trusted (e.g., OVS-DPDK, here), these operations performed by a VF would be legitimate and better be allowed.
To enable VF to request “trusted mode,” a new trusted VF concept was introduced in Linux kernel 4.4 and allowed VFs to become “trusted” and perform some privileged operations.
The BNXT PMD supports the trusted VF mode of operation. Only a PF can enable the trusted attribute on the VF. It is preferable to enable the Trusted setting on a VF before starting applications. However, the BNXT PMD handles dynamic changes in trusted settings as well.
Note that control commands, e.g., ethtool, will work via the kernel PF driver, not via the trusted VF driver.
Operations supported by trusted VF:
MAC address configuration
Promiscuous mode setting
Flow rule creation
Operations not supported by trusted VF:
Firmware upgrade
11.3.2. Running on PF
Unlike the VF when BNXT PMD runs on a PF there are no restrictions placed on the features which the PF can enable or request. In a multiport NIC, each port will have a corresponding PF. Also depending on the configuration of the NIC there can be more than one PF associated per port. A sysadmin can load the kernel driver on one PF, and run BNXT PMD on the other PF or run the PMD on both the PFs. In such cases, the firmware picks one of the PFs as a master PF.
Much like in the trusted VF, the DPDK application must be trusted and expected to be well-behaved.
11.4. Features
The BNXT PMD supports the following features:
- Port Control
Port MTU
LED
Flow Control and Autoneg
- Packet Filtering
Unicast MAC Filter
Multicast MAC Filter
VLAN Filtering
Allmulticast Mode
Promiscuous Mode
- Stateless Offloads
CRC Offload
Checksum Offload (IPv4, TCP, and UDP)
Multi-Queue (TSS and RSS)
Segmentation and Reassembly (TSO and LRO)
VLAN insert strip
Stats Collection
Generic Flow Offload
11.4.1. Port Control
Port MTU: BNXT PMD supports the MTU (Maximum Transmission Unit) up to 9,574 bytes:
testpmd> port config mtu (port_id) mtu_value
testpmd> show port info (port_id)
LED: Application tunes on (or off) a port LED, typically for a port identification:
int rte_eth_led_on (uint16_t port_id)
int rte_eth_led_off (uint16_t port_id)
Flow Control and Autoneg: Application tunes on (or off) flow control and/or auto-negotiation on a port:
testpmd> set flow_ctrl rx (on|off) (port_id)
testpmd> set flow_ctrl tx (on|off) (port_id)
testpmd> set flow_ctrl autoneg (on|off) (port_id)
Note that the BNXT PMD does not support some options and ignores them when requested:
high_water
low_water
pause_time
mac_ctrl_frame_fwd
send_xon
11.4.2. Packet Filtering
Applications control the packet-forwarding behaviors with packet filters.
The BNXT PMD supports hardware-based packet filtering:
- UC (Unicast) MAC Filters
No unicast packets are forwarded to an application except the one with DMAC address added to the port
At initialization, the station MAC address is added to the port
- MC (Multicast) MAC Filters
No multicast packets are forwarded to an application except the one with MC address added to the port
When the application listens to a multicast group, it adds the MC address to the port
- VLAN Filtering Mode
When enabled, no packets are forwarded to an application except the ones with the VLAN tag assigned to the port
- Allmulticast Mode
When enabled, every multicast packet received on the port is forwarded to the application
Typical usage is routing applications
- Promiscuous Mode
When enabled, every packet received on the port is forwarded to the application
11.4.2.1. Unicast MAC Filter
The application can add (or remove) MAC addresses to enable (or disable) filtering on MAC address used to accept packets.
testpmd> show port (port_id) macs
testpmd> mac_addr add port_id XX:XX:XX:XX:XX:XX
testpmd> mac_addr remove port_id XX:XX:XX:XX:XX:XX
11.4.2.2. Multicast MAC Filter
The application can add (or remove) Multicast addresses that enable (or disable) filtering on multicast MAC address used to accept packets.
testpmd> show port (port_id) mcast_macs
testpmd> mcast_addr add port_id XX:XX:XX:XX:XX:XX
testpmd> mcast_addr remove port_id XX:XX:XX:XX:XX:XX
Application adds (or removes) Multicast addresses to enable (or disable) allowlist filtering to accept packets.
Note that the BNXT PMD supports up to 16 MC MAC filters. if the user adds more than 16 MC MACs, the BNXT PMD puts the port into the Allmulticast mode.
11.4.2.3. VLAN Filtering
The application enables (or disables) VLAN filtering mode. When the mode is enabled, no packets are forwarded to an application except ones with VLAN tag assigned for the application.
testpmd> vlan set filter (on|off) (port_id)
testpmd> rx_vlan (add|rm) (vlan_id) (port_id)
11.4.2.4. Allmulticast Mode
The application enables (or disables) the allmulticast mode. When the mode is enabled, every multicast packet received is forwarded to the application.
testpmd> show port info (port_id)
testpmd> set allmulti (port_id) (on|off)
11.4.2.5. Promiscuous Mode
The application enables (or disables) the promiscuous mode. When the mode is enabled on a port, every packet received on the port is forwarded to the application.
testpmd> show port info (port_id)
testpmd> set promisc port_id (on|off)
11.4.3. Stateless Offloads
Like Linux, DPDK provides enabling hardware offload of some stateless processing (such as checksum calculation) of the stack, alleviating the CPU from having to burn cycles on every packet.
Listed below are the stateless offloads supported by the BNXT PMD:
CRC offload (for both TX and RX packets)
- Checksum Offload (for both TX and RX packets)
IPv4 Checksum Offload
TCP Checksum Offload
UDP Checksum Offload
- Segmentation/Reassembly Offloads
TCP Segmentation Offload (TSO)
Large Receive Offload (LRO)
- Multi-Queue
Transmit Side Scaling (TSS)
Receive Side Scaling (RSS)
Also, the BNXT PMD supports stateless offloads on inner frames for tunneled packets. Listed below are the tunneling protocols supported by the BNXT PMD:
VXLAN
GRE
NVGRE
Note that enabling (or disabling) stateless offloads requires applications to stop DPDK before changing configuration.
11.4.3.1. CRC Offload
The FCS (Frame Check Sequence) in the Ethernet frame is a four-octet CRC (Cyclic Redundancy Check) that allows detection of corrupted data within the entire frame as received on the receiver side.
The BNXT PMD supports hardware-based CRC offload:
TX: calculate and insert CRC
RX: check and remove CRC, notify the application on CRC error
Note that the CRC offload is always turned on.
11.4.3.2. Checksum Offload
The application enables hardware checksum calculation for IPv4, TCP, and UDP.
testpmd> port stop (port_id)
testpmd> csum set (ip|tcp|udp|outer-ip|outer-udp) (sw|hw) (port_id)
testpmd> set fwd csum
11.4.3.3. Multi-Queue
Multi-Queue, also known as TSS (Transmit Side Scaling) or RSS (Receive Side Scaling), is a common networking technique that allows for more efficient load balancing across multiple CPU cores.
The application enables multiple TX and RX queues when it is started.
dpdk-testpmd -l 1,3,5 --main-lcore 1 --txq=2 –rxq=2 --nb-cores=2
TSS
TSS distributes network transmit processing across several hardware-based transmit queues, allowing outbound network traffic to be processed by multiple CPU cores.
RSS
RSS distributes network receive processing across several hardware-based receive queues, allowing inbound network traffic to be processed by multiple CPU cores.
The application can select the RSS mode, i.e. select the header fields that are
included for hash calculation. The BNXT PMD supports the RSS mode of
default|ip|tcp|udp|none
, where default mode is L3 and L4.
For tunneled packets, RSS hash is calculated over inner frame header fields. Applications may want to select the tunnel header fields for hash calculation, and it will be supported in 20.08 using RSS level.
testpmd> port config (port_id) rss (all|default|ip|tcp|udp|none)
// note that the testpmd defaults the RSS mode to ip
// ensure to issue the command below to enable L4 header (TCP or UDP) along with IPv4 header
testpmd> port config (port_id) rss default
// to check the current RSS configuration, such as RSS function and RSS key
testpmd> show port (port_id) rss-hash key
// RSS is enabled by default. However, application can disable RSS as follows
testpmd> port config (port_id) rss none
Application can change the flow distribution, i.e. remap the received traffic to CPU cores, using RSS RETA (Redirection Table).
// application queries the current RSS RETA configuration
testpmd> show port (port_id) rss reta size (mask0, mask1)
// application changes the RSS RETA configuration
testpmd> port config (port_id) rss reta (hash, queue) [, (hash, queue)]
11.4.3.4. TSO
TSO (TCP Segmentation Offload), also known as LSO (Large Send Offload), enables the TCP/IP stack to pass to the NIC a larger datagram than the MTU (Maximum Transmit Unit). NIC breaks it into multiple segments before sending it to the network.
The BNXT PMD supports hardware-based TSO.
// display the status of TSO
testpmd> tso show (port_id)
// enable/disable TSO
testpmd> port config (port_id) tx_offload tcp_tso (on|off)
// set TSO segment size
testpmd> tso set segment_size (port_id)
The BNXT PMD also supports hardware-based tunneled TSO.
// display the status of tunneled TSO
testpmd> tunnel_tso show (port_id)
// enable/disable tunneled TSO
testpmd> port config (port_id) tx_offload vxlan_tnl_tso|gre_tnl_tso (on|off)
// set tunneled TSO segment size
testpmd> tunnel_tso set segment_size (port_id)
Note that the checksum offload is always assumed to be enabled for TSO.
11.4.3.5. LRO
LRO (Large Receive Offload) enables NIC to aggregate multiple incoming TCP/IP packets from a single stream into a larger buffer, before passing to the networking stack.
The BNXT PMD supports hardware-based LRO.
// display the status of LRO
testpmd> show port (port_id) rx_offload capabilities
testpmd> show port (port_id) rx_offload configuration
// enable/disable LRO
testpmd> port config (port_id) rx_offload tcp_lro (on|off)
// set max LRO packet (datagram) size
testpmd> port config (port_id) max-lro-pkt-size (max_size)
The BNXT PMD also supports tunneled LRO.
Some applications, such as routing, should not change the packet headers as they pass through (i.e. received from and sent back to the network). In such a case, GRO (Generic Receive Offload) should be used instead of LRO.
11.4.4. VLAN Insert/Strip
DPDK application offloads VLAN insert/strip to improve performance. The BNXT PMD supports hardware-based VLAN insert/strip offload for both single and double VLAN packets.
11.4.4.1. VLAN Insert
Application configures the VLAN TPID (Tag Protocol ID). By default, the TPID is 0x8100.
// configure outer TPID value for a port
testpmd> vlan set outer tpid (tpid_value) (port_id)
The inner TPID set will be rejected as the BNXT PMD supports inserting only an outer VLAN. Note that when a packet has a single VLAN, the tag is considered as outer, i.e. the inner VLAN is relevant only when a packet is double-tagged.
The BNXT PMD supports various TPID values shown below. Any other values will be rejected.
0x8100
0x88a8
0x9100
0x9200
0x9300
The BNXT PMD supports the VLAN insert offload per-packet basis. The application provides the TCI (Tag Control Info) for a packet via mbuf. In turn, the BNXT PMD inserts the VLAN tag (via hardware) using the provided TCI along with the configured TPID.
// enable VLAN insert offload
testpmd> port config (port_id) rx_offload vlan_insert|qinq_insert (on|off)
if (mbuf->ol_flags && RTE_MBUF_F_TX_QINQ) // case-1: insert VLAN to single-tagged packet
tci_value = mbuf->vlan_tci_outer
else if (mbuf->ol_flags && RTE_MBUF_F_TX_VLAN) // case-2: insert VLAN to untagged packet
tci_value = mbuf->vlan_tci
11.4.4.2. VLAN Strip
The application configures the per-port VLAN strip offload.
// enable VLAN strip on a port
testpmd> port config (port_id) tx_offload vlan_strip (on|off)
// notify application VLAN strip via mbuf
mbuf->ol_flags |= RTE_MBUF_F_RX_VLAN | RTE_MBUF_F_RX_STRIPPED // outer VLAN is found and stripped
mbuf->vlan_tci = tci_value // TCI of the stripped VLAN
11.4.5. Time Synchronization
System operators may run a PTP (Precision Time Protocol) client application to synchronize the time on the NIC (and optionally, on the system) to a PTP master.
The BNXT PMD supports a PTP client application to communicate with a PTP master clock using DPDK IEEE1588 APIs. Note that the PTP client application needs to run on PF and vector mode needs to be disabled.
testpmd> set fwd ieee1588 // enable IEEE 1588 mode
When enabled, the BNXT PMD configures hardware to insert IEEE 1588 timestamps to the outgoing PTP packets and reports IEEE 1588 timestamps from the incoming PTP packets to application via mbuf.
// RX packet completion will indicate whether the packet is PTP
mbuf->ol_flags |= RTE_MBUF_F_RX_IEEE1588_PTP
11.4.6. Statistics Collection
In Linux, the ethtool -S enables us to query the NIC stats. DPDK provides the similar functionalities via rte_eth_stats and rte_eth_xstats.
The BNXT PMD supports both basic and extended stats collection:
Basic stats
Extended stats
11.4.6.1. Basic Stats
The application collects per-port and per-queue stats using rte_eth_stats APIs.
testpmd> show port stats (port_id)
Basic stats include:
ipackets
ibytes
opackets
obytes
imissed
ierrors
oerrors
By default, per-queue stats for 16 queues are supported. For more than 16
queues, BNXT PMD should be compiled with RTE_ETHDEV_QUEUE_STAT_CNTRS
set to the desired number of queues.
11.4.6.2. Extended Stats
Unlike basic stats, the extended stats are vendor-specific, i.e. each vendor provides its own set of counters.
The BNXT PMD provides a rich set of counters, including per-flow counters, per-cos counters, per-priority counters, etc.
testpmd> show port xstats (port_id)
Shown below is the elaborated sequence to retrieve extended stats:
// application queries the number of xstats
len = rte_eth_xstats_get(port_id, NULL, 0);
// BNXT PMD returns the size of xstats array (i.e. the number of entries)
// BNXT PMD returns 0, if the feature is compiled out or disabled
// application allocates memory for xstats
struct rte_eth_xstats_name *names; // name is 64 character or less
struct rte_eth_xstats *xstats;
names = calloc(len, sizeof(*names));
xstats = calloc(len, sizeof(*xstats));
// application retrieves xstats // names and values
ret = rte_eth_xstats_get_names(port_id, *names, len);
ret = rte_eth_xstats_get(port_id, *xstats, len);
// application checks the xstats
// application may repeat the below:
len = rte_eth_xstats_reset(port_id); // reset the xstats
// reset can be skipped, if application wants to see accumulated stats
// run traffic
// probably stop the traffic
// retrieve xstats // no need to retrieve xstats names again
// check xstats
11.4.7. Generic Flow Offload
Applications can get benefit by offloading all or part of flow processing to hardware. For example, applications can offload packet classification only (partial offload) or whole match-action (full offload).
DPDK offers the Generic Flow API (rte_flow API) to configure hardware to perform flow processing.
11.4.7.1. TruFlow®
To fully support the generic flow offload, TruFlow was introduced in BNXT PMD. Before TruFlow, hardware flow processing resources were mapped to and handled by firmware. With TruFlow, hardware flow processing resources are mapped to and handled by driver.
Alleviating the limitations of firmware-based feature development, TruFlow not only increases the flow offload feature velocity but also improves the control plane performance (i.e., higher flow update rate).
11.5. Flow APIs, Items, and Actions
BNXT PMD supports thread-safe rte_flow operations for rte_flow APIs and rich set of flow items (i.e., matching patterns) and actions. Refer to the “Supported APIs” section for the list of rte_flow APIs as well as flow items and actions.
11.6. Flow Persistency
On stopping a device port, all the flows created on a port by the application will be flushed from the hardware and any tables maintained by the PMD. After stopping the device port, all flows on the port become invalid and are not represented in the system anymore. Instead of destroying or flushing such flows an application should discard all references to these flows and re-create the flows as required after the port is restarted.
Note: A VNIC represents a virtual interface in the hardware. It is a resource in the RX path of the chip and is used to setup various target actions such as RSS, MAC filtering etc. for the physical function in use.
11.7. Virtual Function Port Representors
The BNXT PMD supports the creation of VF port representors for the control
and monitoring of BNXT virtual function devices. Each port representor
corresponds to a single virtual function of that device that is connected to a
VF. When there is no hardware flow offload, each packet transmitted by the VF
will be received by the corresponding representor. Similarly each packet that is
sent to a representor will be received by the VF. Applications can take
advantage of this feature when SRIOV is enabled. The representor will allow the
first packet that is transmitted by the VF to be received by the DPDK
application which can then decide if the flow should be offloaded to the
hardware. Once the flow is offloaded in the hardware, any packet matching the
flow will be received by the VF while the DPDK application will not receive it
any more. The BNXT PMD supports creation and handling of the port representors
when the PMD is initialized on a PF or trusted-VF. The user can specify the list
of VF IDs of the VFs for which the representors are needed by using the
devargs
option representor
.:
-a DBDF,representor=[0,1,4]
Note that currently hot-plugging of representor ports is not supported so all the required representors must be specified on the creation of the PF or the trusted VF.
11.8. Application Support
11.8.1. Firmware
The BNXT PMD supports the application to retrieve the firmware version.
testpmd> show port info (port_id)
Note that the applications cannot update the firmware using BNXT PMD.
11.8.2. Multiple Processes
When two or more DPDK applications (e.g., testpmd and dpdk-pdump) share a single instance of DPDK, the BNXT PMD supports a single primary application and one or more secondary applications. Note that the DPDK-layer (not the PMD) ensures there is only one primary application.
There are two modes:
Manual mode
Application notifies whether it is primary or secondary using proc-type flag
1st process should be spawned with
--proc-type=primary
All subsequent processes should be spawned with
--proc-type=secondary
Auto detection mode
Application is using
proc-type=auto
flagA process is spawned as a secondary if a primary is already running
The BNXT PMD uses the info to skip a device initialization, i.e. performs a device initialization only when being brought up by a primary application.
11.8.3. Runtime Queue Setup
Typically, a DPDK application allocates TX and RX queues statically: i.e. queues are allocated at start. However, an application may want to increase (or decrease) the number of queues dynamically for various reasons, e.g. power savings.
The BNXT PMD supports applications to increase or decrease queues at runtime.
testpmd> port config all (rxq|txq) (num_queues)
Note that a DPDK application must allocate default queues (one for TX and one for RX at minimum) at initialization.
11.8.4. Descriptor Status
Applications may use the descriptor status for various reasons, e.g. for power savings. For example, an application may stop polling and change to interrupt mode when the descriptor status shows no packets to service for a while.
The BNXT PMD supports the application to retrieve both TX and RX descriptor status.
testpmd> show port (port_id) (rxq|txq) (queue_id) desc (desc_id) status
11.8.5. Bonding
DPDK implements a light-weight library to allow PMDs to be bonded together and provide a single logical PMD to the application.
dpdk-testpmd -l 0-3 -n4 --vdev 'net_bonding0,mode=0,slave=<PCI B:D.F device 1>,slave=<PCI B:D.F device 2>,mac=XX:XX:XX:XX:XX:XX’ – --socket_num=1 – -i --port-topology=chained
(ex) dpdk-testpmd -l 1,3,5,7,9 -n4 --vdev 'net_bonding0,mode=0,slave=0000:82:00.0,slave=0000:82:00.1,mac=00:1e:67:1d:fd:1d' – --socket-num=1 – -i --port-topology=chained
11.9. Vector Processing
Vector mode provides significantly improved performance over scalar mode, using SIMD (Single Instruction Multiple Data) instructions to operate on multiple packets in parallel.
The BNXT PMD provides vectorized burst transmit/receive function implementations on x86-based platforms and ARM-based platforms. The BNXT PMD supports SSE (Streaming SIMD Extensions) and AVX2 (Advanced Vector Extensions 2) instructions for x86-based platforms, and NEON instructions for ARM-based platforms.
The BNXT Vector PMD is enabled in DPDK builds by default. However, the vector mode is disabled when applying SIMD instructions does not improve the performance due to non-uniform packet handling. TX and RX vector mode can be enabled independently from each other, and the decision to disable vector mode is made at run-time when the port is started.
The vector mode is disabled with TX and RX offloads. However, a limited set of offloads can be enabled in a vector mode. Listed below are the TX and RX offloads with which the vector mode can be enabled:
TX offloads supported in vector mode
RTE_ETH_TX_OFFLOAD_MBUF_FAST_FREE
RX offloads supported in vector mode
RTE_ETH_RX_OFFLOAD_VLAN_STRIP RTE_ETH_RX_OFFLOAD_KEEP_CRC RTE_ETH_RX_OFFLOAD_IPV4_CKSUM RTE_ETH_RX_OFFLOAD_UDP_CKSUM RTE_ETH_RX_OFFLOAD_TCP_CKSUM RTE_ETH_RX_OFFLOAD_OUTER_IPV4_CKSUM RTE_ETH_RX_OFFLOAD_OUTER_UDP_CKSUM RTE_ETH_RX_OFFLOAD_RSS_HASH RTE_ETH_RX_OFFLOAD_VLAN_FILTER
Note that the offload configuration changes impacting the vector mode enablement are allowed only when the port is stopped.
11.10. Performance Report
Broadcom DPDK performance has been reported since 19.08 release. The reports provide not only the performance results but also test scenarios including test topology and detailed configurations. See the reports at DPDK performance link <https://core.dpdk.org/perf-reports/>.
11.11. Supported Network Adapters
Listed below are BCM57400 and BCM57500 NetXtreme-E® family of Ethernet network adapters. More information can be found in the NetXtreme|reg| Brand section of the Broadcom website.
11.11.1. BCM57400 NetXtreme-E® Family of Ethernet Network Controllers
11.11.1.1. PCIe NICs
P210P .... Dual-port 10 Gigabit Ethernet Adapter
P210TP ... Dual-port 10 Gigabit Ethernet Adapter
P225P .... Dual-port 25 Gigabit Ethernet Adapter
P150P .... Single-port 50 Gigabit Ethernet Adapter
11.11.1.2. OCP 2.0 NICs
M210P .... Dual-port 10 Gigabit Ethernet Adapter
M210TP ... Dual-port 10 Gigabit Ethernet Adapter
M125P .... Single-port 25 Gigabit Ethernet Adapter
M225P .... Dual-port 25 Gigabit Ethernet Adapter
M150P .... Single-port 50 Gigabit Ethernet Adapter
M150PM ... Single-port Multi-Host 50 Gigabit Ethernet Adapter
11.11.1.3. OCP 3.0 NICs
N210P .... Dual-port 10 Gigabit Ethernet Adapter
N210TP ... Dual-port 10 Gigabit Ethernet Adapter
N225P ... Dual-port 10 Gigabit Ethernet Adapter
11.11.2. BCM57500 NetXtreme-E® Family of Ethernet Network Controllers
11.11.2.1. PCIe NICs
P410SG ... Quad-port 10 Gigabit Ethernet Adapter
P410SGBT . Quad-port 10 Gigabit Ethernet Adapter
P425G .... Quad-port 25 Gigabit Ethernet Adapter
P1100G ... Single-port 100 Gigabit Ethernet Adapter
P2100G ... Dual-port 100 Gigabit Ethernet Adapter
P2200G ... Dual-port 200 Gigabit Ethernet Adapter
11.11.2.2. OCP 2.0 NICs
M1100G ... Single-port OCP 2.0 10/25/50/100 Gigabit Ethernet Adapter
11.11.2.3. OCP 3.0 NICs
N410SG ... Quad-port 10 Gigabit Ethernet Adapter
N410SGBT . Quad-port 10 Gigabit Ethernet Adapter
N425G .... Quad-port 25 Gigabit Ethernet Adapter
N150G .... Single-port 50 Gigabit Ethernet Adapter
N250G .... Dual-port 50 Gigabit Ethernet Adapter
N1100G ... Single-port 100 Gigabit Ethernet Adapter
N2100G ... Dual-port 100 Gigabit Ethernet Adapter
N2200G ... Dual-port 200 Gigabit Ethernet Adapter
11.12. Supported Firmware Versions
Shown below are Ethernet Network Adapters and their supported firmware versions (refer to section “Supported Network Adapters” for the list of adapters):
BCM57400 NetXtreme-E\ |reg| Family
… Firmware 219.0.0 or laterBCM57500 NetXtreme-E\ |reg| Family
… Firmware 219.0.0 or later
Shown below are DPDK LTS releases and their supported firmware versions:
DPDK Release 19.11
… Firmware 219.0.103 or laterDPDK Release 20.11
… Firmware 219.0.103 or laterDPDK Release 21.11
… Firmware 221.0.101 or later
11.13. Supported APIs
11.13.1. rte_eth APIs
Listed below are the rte_eth functions supported:
* rte_eth_allmulticast_disable
* rte_eth_allmulticast_enable
* rte_eth_allmulticast_get
* rte_eth_dev_close
* rte_eth_dev_conf_get
* rte_eth_dev_configure
* rte_eth_dev_default_mac_addr_set
* rte_eth_dev_flow_ctrl_get
* rte_eth_dev_flow_ctrl_set
* rte_eth_dev_fw_version_get
* rte_eth_dev_get_eeprom
* rte_eth_dev_get_eeprom
* rte_eth_dev_get_eeprom_length
* rte_eth_dev_get_eeprom_length
* rte_eth_dev_get_module_eeprom
* rte_eth_dev_get_module_info
* rte_eth_dev_get_mtu
* rte_eth_dev_get_name_by_port rte_eth_dev_get_port_by_name
* rte_eth_dev_get_supported_ptypes
* rte_eth_dev_get_vlan_offload
* rte_eth_dev_info_get
* rte_eth_dev_infos_get
* rte_eth_dev_mac_addr_add
* rte_eth_dev_mac_addr_remove
* rte_eth_dev_macaddr_get
* rte_eth_dev_macaddrs_get
* rte_eth_dev_owner_get
* rte_eth_dev_rss_hash_conf_get
* rte_eth_dev_rss_hash_update
* rte_eth_dev_rss_reta_query
* rte_eth_dev_rss_reta_update
* rte_eth_dev_rx_intr_disable
* rte_eth_dev_rx_intr_enable
* rte_eth_dev_rx_queue_start
* rte_eth_dev_rx_queue_stop
* rte_eth_dev_set_eeprom
* rte_eth_dev_set_link_down
* rte_eth_dev_set_link_up
* rte_eth_dev_set_mc_addr_list
* rte_eth_dev_set_mtu
* rte_eth_dev_set_vlan_ether_type
* rte_eth_dev_set_vlan_offload
* rte_eth_dev_set_vlan_pvid
* rte_eth_dev_start
* rte_eth_dev_stop
* rte_eth_dev_supported_ptypes_get
* rte_eth_dev_tx_queue_start
* rte_eth_dev_tx_queue_stop
* rte_eth_dev_udp_tunnel_port_add
* rte_eth_dev_udp_tunnel_port_delete
* rte_eth_dev_vlan_filter
* rte_eth_led_off
* rte_eth_led_on
* rte_eth_link_get
* rte_eth_link_get_nowait
* rte_eth_macaddr_get
* rte_eth_macaddrs_get
* rte_eth_promiscuous_disable
* rte_eth_promiscuous_enable
* rte_eth_promiscuous_get
* rte_eth_rx_burst_mode_get
* rte_eth_rx_queue_info_get
* rte_eth_rx_queue_setup
* rte_eth_stats_get
* rte_eth_stats_reset
* rte_eth_timesync_adjust_time
* rte_eth_timesync_disable
* rte_eth_timesync_enable
* rte_eth_timesync_read_rx_timestamp
* rte_eth_timesync_read_time
* rte_eth_timesync_read_tx_timestamp
* rte_eth_timesync_write_time
* rte_eth_tx_burst_mode_get
* rte_eth_tx_queue_info_get
* rte_eth_tx_queue_setup
* rte_eth_xstats_get
* rte_eth_xstats_get_names
11.13.2. rte_flow APIs
Listed below are the rte_flow functions supported:
* rte_flow_ops_get
* rte_flow_validate
* rte_flow_create
* rte_flow_destroy
* rte_flow_flush
* rte_flow_tunnel_action_decap_release
* rte_flow_tunnel_decap_set
* rte_flow_tunnel_item_release
* rte_flow_tunnel_match
11.13.3. rte_flow Items
Refer to rte_flow items availability in networking drivers.
Listed below are the rte_flow items supported:
any
eth
gre
icmp
icmp6
ipv4
ipv6
pf
phy_port
port_id
port_representor
represented_port
tcp
udp
vf
vlan
vxlan
11.13.4. rte_flow Actions
Refer to rte_flow actions availability in networking drivers.
Listed below are the rte_flow actions supported:
count
dec_ttl
drop
of_pop_vlan
of_push_vlan
of_set_vlan_pcp
of_set_vlan_vid
pf
phy_port
port_id
port_representor
represented_port
rss
set_ipv4_dst
set_ipv4_src
set_tp_dst
set_tp_src
vf
vxlan_decap
vxlan_encap