日韩性视频-久久久蜜桃-www中文字幕-在线中文字幕av-亚洲欧美一区二区三区四区-撸久久-香蕉视频一区-久久无码精品丰满人妻-国产高潮av-激情福利社-日韩av网址大全-国产精品久久999-日本五十路在线-性欧美在线-久久99精品波多结衣一区-男女午夜免费视频-黑人极品ⅴideos精品欧美棵-人人妻人人澡人人爽精品欧美一区-日韩一区在线看-欧美a级在线免费观看

歡迎訪問 生活随笔!

生活随笔

當前位置: 首頁 > 编程资源 > 综合教程 >内容正文

综合教程

【RDMA】mellanox RDMA网卡驱动安装和查看状态

發布時間:2023/12/2 综合教程 39 生活家
生活随笔 收集整理的這篇文章主要介紹了 【RDMA】mellanox RDMA网卡驱动安装和查看状态 小編覺得挺不錯的,現在分享給大家,幫大家做個參考.

目錄

1. 基礎知識

一、診斷工具--翻譯中

2. 驅動安裝

3.配置IP

4. 性能測試

5. 其他問題

RDMA 、InfiniBand、IB卡、IB驅動的關系

問題記錄?


2. 驅動安裝

  1. 下載驅動,進入網站選擇相應系統和軟件版本,archive versions這里可以下載舊版本驅動

    NVIDIA Mellanox InfiniBand and VPI Software | NVIDIA

    點擊,然后跳轉頁面,拉到底,勾選復選框,自動開始下載

    下載完成后傳到需要安裝驅動的環境上。

  2. 掛載或解壓,如果下載的iso則掛載,若是tgz就解壓,下面是掛載命令:

    sudo? mount -o ro,loop? MLNX_OFED_LINUX-5.0-2.1.8.0-ubuntu16.04-x86_64.iso? /mnt

  3. 安裝

    cd /mnt sudo ./mlnxofedinstall

    可能會提示你安裝一堆東西,復制,安裝就可以了。

定制系統的安裝

如果是非社區的定制系統(可能同不過腳本的判斷,報錯系統為不支持的系統,則需要加一些參數),例如公司的定制系統是基于centos9/redahat9,則可以加--distro rhel9.0?

./mlnxofedinstall --distro rhel9.0?

非匹配內核的安裝

非標準內核的,則可以加--add-kernel-support參數讓其根據當前的內核重新編譯安裝

執行下面命令查看包支持的內核版本和查看本機的內核版本:

cat .supported_kernels? ?列出包支持的內核版本

uname -r? ? 顯示本機內核版本
?

如果本機內核版本不在支持的內核列表里面,則加--add-kernel-support 參數讓它重新根據當前的內核編譯出包并安裝

./mlnxofedinstall --add-kernel-support ??

還可以根據系統增加加其他參數:

./mlnxofedinstall --add-kernel-support --distro rhel9.0 --with-nvmf --with-nfsrdma? ?

不檢查OS的發布版本進行編譯/不更新固件

--skip-distro-check 讓執行腳本不去檢查當前OS的發布版本號(如redhat7.5)

--without-fw-update 參數是不更新FW

更多的參數可以./mlnxofedinstall -h 查看。

mlnx_add_kernel_support.sh腳本也有同樣的功能:

./mlnx_add_kernel_support.sh -m?/opt/mellanox/MLNX_OFED_LINUX-5.0-2.1.8.0-rhel7.7-x86_64?--make-tgz?-t /tmp/

生成對應此OS內核的包,生成目錄默認在/tmp下,安裝過程中可能會有提示需要安裝相關OS的依賴包,配置yum源進行安裝即可:

mlnx_add_kernel_support.sh腳本文件用于在當前系統內核下進行重新編譯打包

其中-m后參數為原始驅動包文件夾路徑,-t后參數為存放重新編譯打包后的文件夾,--make-tgz表示已tgz壓縮包方式打包,--skip-repo 不檢查OS的發布版本進行編譯)

  1. 安裝成功截圖:

  2. 執行以下命令:

    sudo /etc/init.d/openibd restart sudo /etc/init.d/opensmd restart

  3. 查看網卡狀態:

    sudo hca_self_test.ofed

    沒有 failed 就對了。

    一些其它查看網卡信息的命令:

    ibstat ibstatus ibv_devinfo ibv_devices #查看本主機的infiniband設備 ibnodes #查看網絡中的infiniband設備

3.配置IP

見:《linux配置網口IP|RDMA配置IP|ens、eno、enp網口的區別》https://blog.csdn.net/bandaoyu/article/details/116308950

4. 性能測試

  1. 服務端運行:

    ib_send_bw -a -c UD -d mlx4_0 -i 1

    注意,參數 -i 指定端口,在一個網卡有多個網口的時候,需要指定測試的端口,具體哪個端口,通過 ibstatus 可以看到。

  2. 客戶端運行:

    ib_send_bw -a -c UD -d mlx4_0 -i 1 172.16.0.102

    最后面的ip地址是服務端infiniband網卡的ip地址。

    1. 其他測試項

https://blog.csdn.net/bandaoyu/article/details/115798045

其他測試工具說明:Mellanox Interconnect Community

測試工具源碼:https://github.com/linux-rdma/perftest

5. 其他問題

更換網卡工作模式:

有些網卡,當你安裝好驅動后,通過 ibstatus 命令,會出現下面的情況:

可以看到,該網卡現在處于 Ethernet 的工作模式,如果想要切換成infiniband模式,參考如下鏈接:

Mellanox Interconnect Community

查看當前工作模式:

sudo /sbin/connectx_port_config -s

輸入以下命令切換工作模式:

sudo /sbin/connectx_port_config

如果提示如圖,說明不支持infiniband模式,否則,就切換成功了,再次使用一下命令可以驗證:

sudo /sbin/connectx_port_config -s

不能切換到infiniband工作模式,并不代表不支持RDMA,處于Ethernet模式的網卡使用 RoCE 協議工作。

RDMA 協議:底層可以是以太網( RoCE 或者 iWARP )或者 Infiniband

有些網卡只支持Ethernet(RoCE),不支持Infiniband模式,也就是想從Ethernet切換到Infiniband模式時不能成功,這個要提前了解好。我目前了解到的,Connectx-3只支持Ethernet模式。

Mellanox Interconnect Community

RDMA 、InfiniBand、IB卡、IB驅動的關系
?

1、RDMA 是一種技術(遠程直接內存訪問技術),如何實現這種技術呢?

2、實現這種技術你可以選擇用 Infiniband 協議。(當然你也可以使用其他協議,eg:roce、iwarp)

3、如果你使用 Infiniband 協議,這個協議作為一個新一代網絡協議。它必須依靠專門的硬件才能實現。eg:專用INC(網卡)—— IB卡,或者專用交換機。

4、當你在服務器上安裝了IB卡之后,你還得安裝 IB驅動,才能使用 infiniband 。

原文鏈接:https://blog.csdn.net/ljlfather/article/details/102930714

其他文章《Infiniband的驅動安裝與配置》:Infiniband的驅動安裝與配置:https://blog.csdn.net/oPrinceme/article/details/51001849

?驗證RDMA內核模塊是否已加載

在centos7中,可以用以下命令查看RDMA內核模塊的狀態:

[root@power27 docs]# /etc/init.d/openibd status

? HCA driver loaded

Configured IPoIB devices:
ib0

Currently active IPoIB devices:
ib0
Configured Mellanox EN devices:

Currently active Mellanox devices:
ib0

The following OFED modules are loaded:

? rdma_ucm
? rdma_cm
? ib_ipoib
? mlx4_core
? mlx4_ib
? mlx4_en
? mlx5_core
? mlx5_ib
? ib_uverbs
? ib_umad
? ib_ucm
? ib_cm
? ib_core
? mlxfw
? mlx5_fpga_tools
?

HCA即支持RDMA的網卡,driver loaded表示驅動已加載。

LINUX 查看安裝的網卡

lspci | grep -i eth

檢查主機上是否有可用的RDMA網卡

libibverbs-utils工具:ibv_devices、ibv_devinfo?

ibv_devinfo?和?ibstat?命令輸出信息稍有不同(比如端口 MTU 信息是在?ibv_devinfo?而不是?ibstat?輸出中顯示,而端口 PUID 信息是在?ibstat?而不是?ibv_devinfo?輸出中顯示。同時有些信息的命名方式也不同,例如:ibstat?輸出中的基礎本地標識符(LID)與?ibv_devinfo?輸出中的?port_lid?是相同的信息。

ibv_devices

ibv_devices是一個包含在libibverbs-utils.rpm包里的工具,用于顯示本機上的RDMA設備:

[root@power27 docs]# ibv_devices
? ? device ? ? ? ? ? ? ? ? node GUID
? ? ------ ? ? ? ? ? ? ?----------------
? ? mlx4_0 ? ? ? ? ? ? ?e41d2d030050e830

?ibv_devinfo

ibv_devinfo? ? ? 輸出主要信息

ibv_devinfo? -v 輸出詳細信息

ibv_devinfo -d? mlx4_0? ? ?輸出指定設備的主要信息

ibv_devinfo -d? mlx4_0 -v?輸出指定設備的詳細信息
Options:
? -d, --ib-dev=<dev> ? ? use IB device <dev> (default first device found)
? -i, --ib-port=<port> ? use port <port> of IB device (default all ports)
? -l, --list ? ? ? ? ? ? print only the IB devices names
? -v, --verbose ? ? ? ? ?print all the attributes of the IB device(s)

?ibv_devinfo也是libibverbs-utils.rpm包中的一個工具,它會打開一個設備查詢設備的屬性,通過它可以驗證用戶空間和內核空間的RMDA棧是否能夠一起正常運作:

[root@power27 docs]# ibv_devinfo -d mlx4_0
hca_id: mlx4_0
? ? ? ? transport: ? ? ? ? ? ? ? ? ? ? ?InfiniBand (0)
? ? ? ? fw_ver: ? ? ? ? ? ? ? ? ? ? ? ? 2.42.5000
? ? ? ? node_guid: ? ? ? ? ? ? ? ? ? ? ?e41d:2d03:0050:e830
? ? ? ? sys_image_guid: ? ? ? ? ? ? ? ? e41d:2d03:0050:e833
? ? ? ? vendor_id: ? ? ? ? ? ? ? ? ? ? ?0x02c9
? ? ? ? vendor_part_id: ? ? ? ? ? ? ? ? 4099
? ? ? ? hw_ver: ? ? ? ? ? ? ? ? ? ? ? ? 0x1
? ? ? ? board_id: ? ? ? ? ? ? ? ? ? ? ? MT_1100120019
? ? ? ? phys_port_cnt: ? ? ? ? ? ? ? ? ?1
? ? ? ? Device ports:
? ? ? ? ? ? ? ? port: ? 1
? ? ? ? ? ? ? ? ? ? ? ? state: ? ? ? ? ? ? ? ? ?PORT_ACTIVE (4)
? ? ? ? ? ? ? ? ? ? ? ? max_mtu: ? ? ? ? ? ? ? ?4096 (5)
? ? ? ? ? ? ? ? ? ? ? ? active_mtu: ? ? ? ? ? ? 4096 (5)
? ? ? ? ? ? ? ? ? ? ? ? sm_lid: ? ? ? ? ? ? ? ? 1
? ? ? ? ? ? ? ? ? ? ? ? port_lid: ? ? ? ? ? ? ? 12
? ? ? ? ? ? ? ? ? ? ? ? port_lmc: ? ? ? ? ? ? ? 0x00
? ? ? ? ? ? ? ? ? ? ? ? link_layer: ? ? ? ? ? ? InfiniBand
?

至少要有一個端口的狀態是PORT_ACTIVE,才能說明RDMA相關組件已經正常運行起來。

?查看加載了哪些模塊?lsmod

ibstat

使用ibstat命令顯示InfiniBand的具體信息:

[root@node24 ~]# ibstatCA 'mlx4_0'	CA type: MT4099Number of ports: 1Firmware version: 2.30.3000Hardware version: 0Node GUID: 0x46d2c92000004820System image GUID: 0x46d2c92000004823Port 1:State: ActivePhysical state: LinkUpRate: 56Base lid: 72LMC: 0SM lid: 49Capability mask: 0x02514868Port GUID: 0x46d2c92000004821Link layer: InfiniBand[root@node25 ~]# ibstatCA 'mlx4_0'	CA type: MT4099Number of ports: 1Firmware version: 2.30.3000Hardware version: 0Node GUID: 0x46d2c92000003500System image GUID: 0x46d2c92000003503Port 1:State: ActivePhysical state: LinkUpRate: 56Base lid: 53LMC: 0SM lid: 49Capability mask: 0x0251486aPort GUID: 0x46d2c92000003501Link layer: InfiniBand

?查看加載了哪些模塊

lsmod

問題記錄?

ceph 的RDMA 連接 RDMAConnectedSocketImpl::activate()? 失敗。 dmesg 查看到:

402303.425662] QAT: Device not yet ready.
[402318.101625] QAT: Device not yet ready.
[402336.462612] QAT: Device not yet ready.
[402367.537465] QAT: Device not yet ready.
[402393.717441] QAT: Device not yet ready.
[402426.970661] QAT: Device not yet ready.
[402457.179681] QAT: Device not yet ready. 

猜測:osd open dvice之后,雖然TCP 建立連接正常讀寫了,但是rdma的device還沒有ready好。

所以TCP的infiniband->recv_msg(cct, tcp_fd, peer_msg);可以正常讀,走過,然后到activate()

RDMA 沒有準備好。

void RDMAConnectedSocketImpl::handle_connection() {
……
? int r = infiniband->recv_msg(cct, tcp_fd, peer_msg);
……
? ? if (!connected) {
? ? ? r = activate();
? ? ? assert(!r);
? ? }

……

}

關于Mellanox mlx5 ethtool計數器

(摘自:關于Mellanox mlx5 ethtool計數器 – https://www.houzhibo.com/archives/1373)

在處理一個業務網卡丟包的問題,監控查看丟包源是在機器進方向丟包的,通過ethtool查看計數器的值和ifconfig的dropped值哪個是一樣的(也可以相近的畢竟也有個別其他情況丟的包,比如crc等。),判斷丟包類型,通過rx_out_of_buffer計數器確認丟包類型,表面意思看起來和buffer有關系,但不確定具體的解釋,通過網上搜索,確認這個計數器的解釋,最終通過修改ring buffer降低了業務的丟包率。修改方法比較簡單,如下:

查看當前設置:

ethtool -g eth0
Ring parameters for eth0:
Pre-set maximums:
RX: 8192
RX Mini: 0
RX Jumbo: 0
TX: 8192
Current hardware settings:
RX: 8192
RX Mini: 0
RX Jumbo: 0
TX: 8192

修改設置:

ethtool -G eth0 rx 8192

ethtool -G eth0 tx 8192

ethtool -g eth0
Ring parameters for eth0:
Pre-set maximums:
RX: 8192
RX Mini: 0
RX Jumbo: 0
TX: 8192
Current hardware settings:
RX: 8192
RX Mini: 0
RX Jumbo: 0
TX: 8192

中午11:30左右修改前后丟包對比

?

在修改的過程中遇到了一個莫名其妙的bug,導致我嘗試在第四次修改的時候服務器自己重啟了,在這沒有去排查具體重啟的原因,重啟之后就可以正常順利的修改了。有重要業務的同學,建議停服再操作,我當時報錯的記錄如下:

?

由于所搜的歷程比較麻煩,所以做個記錄分享下,Mellanox mlx5 ethtool計數器的解釋,摘自Mellanox 官網。原鏈接:MyMellanoxGen2

Understanding mlx5 ethtool Counters

Description:

This post shows the list of ethtool counters applicable for ConnectX-4 and above?(mlx5 driver). All counters listed?here are available via ethtool starting with MLNX_OFED 4.0.

Note: The post also provides a reference to ConnectX-3/ConnectX-3 Pro counters that co-exist for the mlx4 driver (see notes below).
?

  • References
    • Release Notes
  • Counters Overview
    • Counter Groups
    • Counter Types
    • Acceleration Mechanism
  • Counters Description
    • Ring / Software Port Counters
      • Ring / Software Port Counter Table
    • vPort Counters
      • vPort Counter Table
    • Physical Port Counters
      • Physical Port Counter Table
    • Priority Port Counters
      • Priority Port Counter Table
    • Device Counters
    • Full List of Counters

References

  • MLNX_OFED User Manual
  • Netstat tool

Release Notes

This document was updated to match Linux kernel 4.20

Counters Overview

There are several counter groups, depends where the counter is being counted. In addition, each group of counters may have different counter types.

Counter Groups

  • Ring?– software ring counters
  • Software Port?– An aggregation of software ring counters.
  • vPort counters?– traffic counters and drops due to steering or no buffers. May indicate on NIC issues. These counters include Ethernet traffic counters (including Raw Ethernet) and RDMA/RoCE traffic counters.
  • Physical port counters?– the physical port connecting NIC to the network. May indicate on NIC issues or link or network issue. This measuring point holds information on standardized counters like IEEE 802.3, RFC2863, RFC 2819, RFC 3635 and additional counters like flow control, FEC and more. Physical port counters are not exposed to virtual machines.
  • Priority Port Counters?– A set of the?physical port counters, per priory per port.

Counter Types

Counters are divided to three Types

  • Traffic Informative Counters?– counters which counts traffic. These counters can be used for load estimation of for general debug.
  • Traffic Acceleration?Counters?– counters which counts traffic which was accelerated by Mellanox driver or by hardware. The counters are an additional layer to the informative counter set and the same traffic is counted in both informative and acceleration counters. Acceleration counters are marked with [A].
  • Error Counters?– Increment of these counters might indicate a problem. Each of these counter has an explanation and correction action.

Statistic can be fetched via the?ip?link?or?ethtool?commands.?ethtool?provides more detailed information.

ip –s link show <if-name>

ethtool -S <if-name>

Acceleration Mechanism

The following?acceleration mechanisms?have dedicated counters:

  • TSO?(TCP Segmentation Offload) – increasing outbound throughput and reducing CPU utilization by allowing the kernel to buffer multiple packets in a single large buffer. The NIC split the buffer into packet and transmits it
  • LRO?(Large Receive Offload) – increasing inbound throughput and reducing CPU utilization by aggregation of o multiple incoming packet of a single stream to a single buffer
  • CHECKSUM?(Checksum) – calculation of TCP checksum (by the NIC). The following CSUM offload are available (refer to?skbuff.h?for detailed explanation)
    • CHECKSUM_UNNECESSARY
    • CHECKSUM_NONE – no CSUM acceleration was used
    • CHECKSUM_COMPLETE – Device provided CSUM on the entire packet
    • CHECKSUM_PARTIAL – Device provided CSUM
  • CQE Compress?– compression of Completion Queue Events (CQE) used for sparing bandwidth on PCIe and hence achieve better performance.

Counters Description

Ring / Software Port Counters

The following counters are available per?ring?or?Software port.

These counters provide information on the amount of traffic that was accelerated by the NIC. The counters are counting the accelerated traffic in addition to the standard counters which counts it (i.e. accelerated traffic is counted twice).

The counter names in the table below refers to both ring and port counters. the notation for ring counters includes the [i] index without the braces. the notation for port counters doesn’t include the [i]. a counter name rx[i]_packets will be printed as rx0_packets for ring 0 and rx_packets for the software port

Ring / Software Port Counter Table

Counter Description Type
rx[i]_packets The number of packets received on ring i.ConnectX-3 naming :?rx[i]_packets Informative
rx[i]_bytes The number of bytes received on ring i.ConnectX-3 naming :?rx[i]_bytes Informative
tx[i]_packets The number of packets transmitted on ring i.ConnectX-3 naming :?tx[i]_packets Informative
tx[i]_bytes The number of bytes transmitted on ring i.ConnectX-3 naming :?tx[i]_bytes Informative
tx[i]_tso_packets The number of TSO packets transmitted on ring i [A]. Acceleration
tx[i]_tso_bytes The number of TSO bytes transmitted on ring i [A]. Acceleration
tx[i]_tso_inner_packets The number of TSO packets which are indicated to be carry internal encapsulation transmitted on ring i [A] Acceleration
tx[i]_tso_inner_bytes The number of TSO bytes which are indicated to be carry internal encapsulation transmitted on ring i [A]. Acceleration
rx[i]_lro_packets The number of LRO packets received on ring i [A]. Acceleration
rx[i]_lro_bytes The number of LRO bytes received on ring i [A]. Acceleration
rx[i]_csum_unnecessary Packets received with a CHECKSUM_UNNECESSARY on ring i [A]. Acceleration
rx[i]_csum_none Packets received with CHECKSUM_NONE on ring i [A]. Acceleration
rx[i]_csum_complete Packets received with a CHECKSUM_COMPLETE on ring i [A]. Acceleration
rx[i]_csum_unnecessary_inner Packets received with inner encapsulation with a CHECK_SUM UNNECESSARY on ring i [A]. Acceleration
tx[i]_csum_partial Packets transmitted with a CHECKSUM_PARTIAL on ring i [A]. Acceleration
tx[i]_csum_partial_inner Packets transmitted with inner encapsulation with a CHECKSUM_PARTIAL on ring i [A]. Acceleration
tx[i]_csum_none Packets transmitted with no hardware checksum acceleration on ring i. Informative
tx[i]_stoppedtx_queue_stopped?(1) Events where SQ was full on ring i. If this counter is increased, check the amount of buffers allocated for transmission. Error
tx[i]_waketx_queue_wake?(1) Events where SQ was full and has become not full on ring i. Error
tx[i]_droppedtx_queue_dropped?(1) Packets transmitted that were dropped due to DMA mapping failure on ring i. If this counter is increased, check the amount of buffers allocated for transmission. Error
rx[i]_wqe_err The number of wrong opcodes received on ring i. Error
tx[i]_nop The number of no WQEs (empty WQEs) inserted to the SQ (related to ring i) due to the reach of the end of the cyclic buffer. When reaching near to the end of cyclic buffer the driver may add those empty WQEs to avoid handling a state the a WQE start in the end of the queue and ends in the beginning of the queue. This is a normal condition. Informative
rx[i]_mpwqe_frag The number of WQEs that failed to allocate compound page and hence fragmented MPWQE’s (Multi Packet WQEs) were used on ring i. If this counter raise, it may suggest that there is no enough memory for large pages, the driver allocated fragmented pages. This is not abnormal condition. Informative
rx[i]_mpwqe_filler_cqes The number of filler CQEs events that where issued on ring i.berfore kernel 4.19 name was rx[i]_mpwqe_filler Informative
rx[i]_cqe_compress_blks The number of receive blocks with CQE compression on ring i [A]. Acceleration
rx[i]_cqe_compress_pkts The number of receive packets with CQE compression on ring i [A]. Acceleration
rx[i]_cache_reuse The number of events of successful reuse of a page from a driver’s internal page cache – supported from Kernel 4.9 Acceleration
rx[i]_cache_full The number of events of full internal page cache where driver can’t put a page back to the cache for recycling (page will be freed) – supported from Kernel 4.9 Acceleration
rx[i]_cache_empty The number of events where cache was empty – no page to give. driver shall allocate new page – supported from Kernel 4.9 Acceleration
rx[i]_cache_busy The number of events where cache head was busy and cannot be recycled. driver allocated new page – supported from Kernel 4.9 Acceleration
rx[i]_xmit_more The number of packets sent with xmit_more indication set on the skbuff (no doorbell) – Supported from kernel 4.8 Acceleration
tx[i]_cqes The number of completions received on the CQ of TX ring. Supported from kernel 4.19 Informative
ch[i]_poll The number of invocations of?NAPI?poll of channel. Supported from kernel 4.19 Informative
ch[i]_arm The number of times the NAPI poll function completed and armed the completion queues on channelSupported from kernel 4.19 Informative
ch[i]_aff_change The number of times the NAPI poll function explicitly stopped execution on a CPU due to a change in affinity, on channel. Supported from kernel 4.19 Informative
rx[i]_congst_umr The number of times an outstanding UMR request is delayed due to congestion, on ringSupported from kernel 4.19 Error
ch[i]_events The number of hard interrupt events on the completion queues of channel. Supported from kernel 4.19 Informative
rx[i]_mpwqe_filler_strides The number of strides consumed by filler CQEs on ring. Supported from kernel 4.19 Informative
rx[i]_xdp_tx_xmit The number of packets forwarded back to the port due to XDP program XDP_TX action (bouncing). these packets are not counted by other software counters. These packets are counted by physical port and vPort counters – supported from kernel 4.9Before kernel 4.19 name was rx[i]_xdp_tx Informative
rx[i]_xdp_tx_full The number of packets that should have been forwarded back to the port due to XDP_TX action but were dropped due to full tx queue. these packets are not counted by other software counters. These packets are counted by physical port and vPort countersyou may open more rx queues and spread traffic rx over all queues and/or increase rx ring sizesupported from kernel 4.9 Error
rx[i]_xdp_tx_err The number of times an XDP_TX error such as frame too long and frame too short occurred on XDP_TX ring of RX ring. Supported from kernel 4.19 Error
rx[i]_xdp_tx_cqesrx_xdp_tx_cqe?(1) The number of completions received on the CQ of the XDP-TX ring. Supported from kernel 4.19 Informative
rx[i]_xdp_drop The number of packets dropped due to XDP program XDP_DROP action. these packets are not counted by other software counters. These packets are counted by physical port and vPort counters – supported from kernel 4.9 Informative
rx[i]_xdp_redirect The number of times an XDP redirect action was triggered on ring. .Supported from kernel 4.19 Acceleration
tx[i]_xdp_xmit The number of packets redirected to the interface(due to XDP redirect). These packets are not counted by other software counters. These packets are counted by physical port and vPort counters – Supported from kernel 4.19 Informative
tx[i]_xdp_full The number of packets redirected to the interface(due to XDP redirect), but were dropped due to full tx queue. these packets are not counted by other software counters. you may enlarge tx queues. Supported from kernel 4.19 Informative
tx[i]_xdp_err The number of packets redirected to the interface(due to XDP redirect) but were dropped due to error such as frame too long and frame too short . Supported from kernel 4.19 Error
tx[i]_xdp_cqes The number of completions received for packets redirected to the interface(due to XDP redirect) on the CQ . Supported from kernel 4.19 Informative
rx[i]_cache_waive The number of cache evacuation. This can occur due to page move to another NUMA node or page was pfmemalloc-ed and should be freed as soon as possible. Supported from kernel 4.14 Acceleration

Notes:

(1) The corresponding ring and global counters do not share the same name (i.e. do not follow the common naming scheme).

vPort Counters

Counters on the eswitch port that is connected to the VNIC.

vPort Counter Table

Counter Description Type
rx_vport_unicast_packets Unicast packets received, steered to a port including Raw Ethernet QP/DPDK traffic, excluding RDMA traffic Informative
rx_vport_unicast_bytes Unicast bytes received, steered to a port including Raw Ethernet QP/DPDK traffic, excluding RDMA traffic Informative
tx_vport_unicast_packets Unicast packets transmitted, steered from a port including Raw Ethernet QP/DPDK traffic, excluding RDMA traffic Informative
tx_vport_unicast_bytes Unicast bytes transmitted, steered from a port including Raw Ethernet QP/DPDK traffic, excluding RDMA traffic Informative
rx_vport_multicast_packets Multicast packets received, steered to a port including Raw Ethernet QP/DPDK traffic, excluding RDMA traffic Informative
rx_vport_multicast_bytes Multicast bytes received, steered to a port including Raw Ethernet QP/DPDK traffic, excluding RDMA traffic Informative
tx_vport_multicast_packets Multicast packets transmitted, steered from a port including Raw Ethernet QP/DPDK traffic, excluding RDMA traffic Informative
tx_vport_multicast_bytes Multicast bytes transmitted, steered from a port including Raw Ethernet QP/DPDK traffic, excluding RDMA traffic Informative
rx_vport_broadcast_packets Broadcast packets received, steered to a port including Raw Ethernet QP/DPDK traffic, excluding RDMA traffic Informative
rx_vport_broadcast_bytes Broadcast bytes received, steered to a port including Raw Ethernet QP/DPDK traffic, excluding RDMA traffic Informative
tx_vport_broadcast_packets Broadcast packets transmitted, steered from a port including Raw Ethernet QP/DPDK traffic, excluding RDMA traffic Informative
tx_vport_broadcast_bytes Broadcast packets transmitted, steered from a port including Raw Ethernet QP/DPDK traffic, excluding RDMA traffic Informative
rx_vport_rdma_unicast_packets RDMA unicast packets received, steered to a port (counters counts RoCE/UD/RC traffic) [A] Acceleration
rx_vport_rdma_unicast_bytes RDMA unicast bytes received, steered to a port (counters counts RoCE/UD/RC traffic) [A] Acceleration
tx_vport_rdma_unicast_packets RDMA unicast packets transmitted, steered from a port (counters counts RoCE/UD/RC traffic) [A] Acceleration
tx_vport_rdma_unicast_bytes RDMA unicast bytes transmitted, steered from a port (counters counts RoCE/UD/RC traffic) [A] Acceleration
rx_vport_ rdma _multicast_packets RDMA multicast packets received, steered to a port (counters counts RoCE/UD/RC traffic) [A] Acceleration
rx_vport_ rdma _multicast_bytes RDMA multicast bytes received, steered to a port (counters counts RoCE/UD/RC traffic) [A] Acceleration
tx_vport_ rdma _multicast_packets RDMA multicast packets transmitted, steered from a port (counters counts RoCE/UD/RC traffic) [A] Acceleration
tx_vport_ rdma _multicast_bytes RDMA multicast bytes transmitted, steered from a port (counters counts RoCE/UD/RC traffic) [A] Acceleration
rx_steer_missed_packets Number of packets that was received by the NIC, however was discarded because it did not match any flow in the NIC flow table. supported from kernel 4.16 Error
rx_packets Representor?only: packets received, that were handled by the hypervisor. supported from kernel 4.18 Informative
rx_bytes Representor?only: bytes received, that were handled by the hypervisor. supported from kernel 4.18 Informative
tx_packets Representor??only: packets transmitted, that were handled by the hypervisor. supported from kernel 4.18 Informative
tx_bytes Representor??only: bytes transmitted, that were handled by the hypervisor. supported from kernel 4.18 Informative

Physical Port Counters

The physical port counters are the counters on the external port connecting adapter to the network. This measuring point holds information on standardized counters like IEEE 802.3, RFC2863, RFC 2819, RFC 3635 and additional counters like flow control, FEC and more.

Physical Port Counter Table

Counter Description Type
rx_packets_phy The number of packets received on the physical port. This counter doesn’t include packets that were discarded due to FCS, frame size and similar errors.ConnectX-3 naming : rx_packets Informative
tx_packets_phy The number of packets transmitted on the physical port.ConnectX-3 naming : tx_packets Informative
rx_bytes_phy The number of bytes received on the physical port, including Ethernet header and FCS.ConnectX-3 naming : rx_bytes Informative
tx_bytes_phy The number of bytes transmitted on the physical port.ConnectX-3 naming : tx_bytes Informative
rx_multicast_phy The number of multicast packets received on the physical port.ConnectX-3 naming : rx_multicast_packets Informative
tx_multicast_phy The number of multicast packets transmitted on the physical port.ConnectX-3 naming : tx_multicast_packets Informative
rx_broadcast_phy The number of broadcast packets received on the physical port.ConnectX-3 naming : rx_broadcast_packets Informative
tx_broadcast_phy The number of broadcast packets transmitted on the physical port.ConnectX-3 naming : tx_broadcast_packets Informative
rx_crc_errors_phy The number of dropped received packets due to FCS (Frame Check Sequence) error on the physical port. If this counter is increased in high rate, check the link quality using?rx_symbol_error_phy?and?rx_corrected_bits_phy counters below.ConnectX-3 naming : rx_crc_errors Error
rx_in_range_len_errors_phy The number of received packets dropped due to length/type errors on a physical port.ConnectX-3 naming : rx_in_range_length_error Error
rx_out_of_range_len_phy The number of received packets dropped due to length greater than allowed on a physical port.If this counter is increasing, it implies that the peer connected to the adapter has a?larger MTU?configured. Using same MTU configuration shall resolve this issue.ConnectX-3 naming : rx_out_range_length_error Error
rx_oversize_pkts_phy The number of dropped received packets due to length which exceed MTU size on a physical portIf this counter is increasing, it implies that the peer connected to the adapter has a?larger MTU?configured. Using same MTU configuration shall resolve this issue.ConnectX-3 naming : rx_frame_errors Error
rx_symbol_err_phy The number of received packets dropped due to physical coding errors (symbol errors) on a physical port. Error
rx_mac_control_phy The number of MAC control packets received on the physical port. Informative
tx_mac_control_phy The number of MAC control packets transmitted on the physical port. Informative
rx_pause_ctrl_phy The number of link layer pause packets received on a physical port. If this counter is increasing, it implies that the network is congested and cannot absorb the traffic coming from to the adapter. Informative
tx_pause_ctrl_phy The number of link layer pause packets transmitted on a physical port. If this counter is increasing, it implies that the NIC is congested and cannot absorb the traffic coming from the network. Informative
rx_unsupported_op_phy The number of MAC control packets received with unsupported opcode on a physical port. Error
rx_discards_phy The number of received packets dropped due to lack of buffers on a physical port. If this counter is increasing, it implies that the adapter is congested and cannot absorb the traffic coming from the network.ConnectX-3 naming : rx_fifo_errors Error
tx_discards_phy? The number of packets which were discarded on transmission, even no errors were detected. the drop might occur due to link in down state, head of line drop, pause from the network, etc Error?
tx_errors_phy The number of transmitted packets dropped due to a length which exceed MTU size on a physical port. Error
rx_undersize_pkts_phy The number of received packets dropped due to length which is shorter than 64 bytes on a physical port. If this counter is increasing, it implies that the peer connected to the adapter has a non-standard MTU configured or malformed packet had arrived. Error
rx_fragments_phy The number of received packets dropped due to a length which is shorter than 64 bytes and has FCS error on a physical port. If this counter is increasing, it implies that the peer connected to the adapter has a non-standard MTU configured. Error
rx_jabbers_phy The number of received packets d due to a length which is longer than 64 bytes and had FCS error on a physical port. Error
rx_64_bytes_phy The number of packets received on the physical port with size of 64 bytes. Informative
rx_65_to_127_bytes_phy The number of packets received on the physical port with size of 65 to 127 bytes. Informative
rx_128_to_255_bytes_phy The number of packets received on the physical port with size of 128 to 255 bytes. Informative
rx_256_to_511_bytes_phy The number of packets received on the physical port with size of 256 to 512 bytes. Informative
rx_512_to_1023_bytes_phy The number of packets received on the physical port with size of 512 to 1023 bytes. Informative
rx_1024_to_1518_bytes_phy The number of packets received on the physical port with size of 1024 to 1518 bytes. Informative
rx_1519_to_2047_bytes_phy The number of packets received on the physical port with size of 1519 to 2047 bytes. Informative
rx_2048_to_4095_bytes_phy The number of packets received on the physical port with size of 2048 to 4095 bytes. Informative
rx_4096_to_8191_bytes_phy The number of packets received on the physical port with size of 4096 to 8191 bytes. Informative
rx_8192_to_10239_bytes_phy The number of packets received on the physical port with size of 8192 to 10239 bytes. Informative
link_down_events_phy The number of times where the link operative state changed to down. In case this counter is increasing it may imply on port flapping. You may need to replace the cable/transceiver. Error
rx_out_of_buffer Number of times receive queue had no software buffers allocated for the adapter’s incoming traffic. Error
module_bus_stuck The number of times that module’s I2C bus (data or clock) short-wire was detected. You may need to replace the cable/transceiver – supported from kernel 4.10 Error
module_high_temp The number of times that the module temperature was too high. If this issue persist, you may need to check the ambient temperature or replace the cable/transceiver module – supported from kernel 4.10 Error
module_bad_shorted The number of times that the module cables were shorted. You may need to replace the cable/transceiver module – supported from kernel 4.10 Error
module_unplug The number of times that module was ejected – supported from kernel 4.10 Informative
rx_buffer_passed_thres_phy The number of events where the port receive buffer was over 85% full. Supported from kernel 4.14 Informative
tx_pause_storm_warning_events The number of times the device was sending pauses for a long period of time – supported from kernel 4.15 Informative
tx_pause_storm_error_events The number of times the device was sending pauses for a long period of time, reaching time out and disabling transmission of pause frames. on the period where pause frames were disabled, drop could have been occurred – supported from kernel 4.15 Error
rx[i]_buff_alloc_err / rx_buff_alloc_err Failed to allocate a buffer to received packet (or SKB) on port (or per ring) Error
rx_bits_phy This counter provides information on the total amount of traffic that could have been received and can be used as a guideline to measure the ratio of errored traffic in?rx_pcs_symbol_err_phy&?rx_corrected_bits_phy. Informative
rx_pcs_symbol_err_phy This counter counts the number of symbol errors that wasn’t corrected by FEC correction algorithm or that FEC algorithm was not active on this interface. If this counter is increasing, it implies that the link between the NIC and the network is suffering from high BER, and that traffic is lost. You may need to replace the cable/transceiver. The error rate is the number of?rx_pcs_symbol_err_phy?divided by the number of rx_phy_bits on a specific time frame. Error
rx_corrected_bits_phy The number of corrected bits on this port according to active FEC (RS/FC). If this counter is increasing, it implies that the link between the NIC and the network is suffering from high BER. The corrected bit rate is the number of?rx_corrected_bits_phy?divided by the number of?rx_phy_bits?on a specific time frame Error
phy_raw_errors_lane[l] This counter counts the number of physical raw errors per lane [l] index. The counter counts errors before FEC corrections. If this counter is increasing, it implies that the link between the NIC and the network is suffering from high BER, and that traffic might be lost. You may need to replace the cable/transceiver. Please check in accordance with?rx_corrected_bits_phy.
Supported from kernel 4.20
Error

Priority Port Counters

The following counters are physical port counters that being counted per L2 priority (0-7).

Note: ‘p’?in the counter name represents the priority.

Priority Port Counter Table

Counter Description Type
rx_prio[p]_bytes The number of bytes received with priority p on the physical port.ConnectX-3 naming :rx_prio_[p]_bytes. this counter also counts packets with no vlan Informative
rx_prio[p]_packets The number of packets received with priority p on the physical port.ConnectX-3 naming :?rx_prio_[p]_packets. this counter also counts packets with no vlan Informative
tx_prio[p]_bytes The number of bytes transmitted on priority p on the physical port.ConnectX-3 naming :tx_prio_[p]_bytes. Informative
tx_prio[p]_packets The number of packets transmitted on priority p on the physical port.ConnectX-3 naming :?tx_prio_[p]_packets. Informative
rx_prio[p]_pause The number of pause packets received with priority p on a physical port. If this counter is increasing, it implies that the network is congested and cannot absorb the traffic coming from the adapter.Note:?This counter is available only if PFC was enabled on priority p. Refer to?HowTo Configure PFC on ConnectX-4?.ConnectX-3 naming : rx_pause_prio_p Informative
rx_prio[p]_pause_duration The duration of pause received (in microSec) on priority p on the physical port. The counter represents the time the port did not send any traffic on this priority. If this counter is increasing, it implies that the network is congested and cannot absorb the traffic coming from the adapter.Note:?This counter is available only if PFC was enabled on priority p. Refer to?HowTo Configure PFC on ConnectX-4?.ConnectX-3 naming : rx_pause_duration_prio_p Informative
rx_prio[p]_pause_transition The number of times a transition from Xoff to Xon on priority p on the physical port has occurred.Note:?This counter is available only if PFC was enabled on priority p. Refer to?HowTo Configure PFC on ConnectX-4?.ConnectX-3 naming : rx_pause_transition_prio_p Informative
tx_prio[p]_pause The number of pause packets transmitted on priority p on a physical port. If this counter is increasing, it implies that the adapter is congested and cannot absorb the traffic coming from the network.Note:?This counter is available only if PFC was enabled on priority p. Refer to?HowTo Configure PFC on ConnectX-4?.ConnectX-3 naming : tx_pause_prio_p Informative
tx_prio[p]_pause_duration The duration of pause transmitter (in microSec) on priority p on the physical port.Note:?This counter is available only if PFC was enabled on priority p. Refer to?HowTo Configure PFC on ConnectX-4?.ConnectX-3 naming : tx_pause_duration_prio_p Informative
rx_prio[p]_buf_discard The number of packets discarded by device due to lack of per host receive buffers.?Supported from kernel 5.3 Informative
rx_prio[p]_cong_discard The number of packets discarded by device due to per host congestion.?Supported from kernel 5.3 Informative
rx_prio[p]_marked The number of packets ecn marked by device due to per host congestion.?Supported from kernel 5.3 Informative
rx_prio[p]_discard The number of packets discarded by device due to lack of receive buffers.?Supported from kernel 5.6 Infornative

Device Counters

Counter Description Type
rx_pci_signal_integrity Counts physical layer PCIe signal integrity errors, the number of transitions to recovery due to Framing errors and CRC (dlp and tlp).If this counter is raising, try moving the adapter card to a different slot to rule out a bad PCI slot. Validate that you are running with the latest firmware available and latest server BIOS version. Error
tx_pci_signal_integrity Counts physical layer PCIe signal integrity errors, the number of transition to recovery initiated by the other side (moving to recovery due to getting TS/EIEOS).If this counter is raising, try moving the adapter card to a different slot to rule out a bad PCI slot. Validate that you are running with the latest firmware available and latest server BIOS version. Error
outbound_pci_buffer_overflow The number of packets dropped due to pci buffer overflow. If this counter is raising in high rate, it might indicate that the receive traffic rate for a host is larger than the PCIe bus and therefore a congestion occurs. Supported from kernel 4.14 Informative
outbound_pci_stalled_rd The percentage (in the range 0…100) of time within the last second that the NIC had outbound non-posted reads requests but could not perform the operation due to insufficient posted credits. Supported from kernel 4.14 Informative
outbound_pci_stalled_wr The percentage (in the range 0…100) of time within the last second that the NIC had outbound posted writes requests but could not perform the operation due to insufficient posted credits. Supported from kernel 4.14 Informative
outbound_pci_stalled_rd_events The number of seconds where outbound_pci_stalled_rd was above 30%. Supported from kernel 4.14 Informative
outbound_pci_stalled_wr_events The number of seconds where outbound_pci_stalled_wr was above 30%. Supported from kernel 4.14 Informative
dev_out_of_buffer The number of times the device owned queue had not enough buffers allocated Error

Full List of Counters

# ethtool -S eth5

NIC statistics:

rx_packets: 10

rx_bytes: 3420

tx_packets: 18

tx_bytes: 1296

tx_tso_packets: 0

tx_tso_bytes: 0

tx_tso_inner_packets: 0

tx_tso_inner_bytes: 0

tx_added_vlan_packets: 0

tx_nop: 0

rx_lro_packets: 0

rx_lro_bytes: 0

rx_ecn_mark: 0

rx_removed_vlan_packets: 0

rx_csum_unnecessary: 0

rx_csum_none: 0

rx_csum_complete: 10

rx_csum_unnecessary_inner: 0

rx_xdp_drop: 0

rx_xdp_redirect: 0

rx_xdp_tx_xmit: 0

rx_xdp_tx_full: 0

rx_xdp_tx_err: 0

rx_xdp_tx_cqe: 0

tx_csum_none: 18

tx_csum_partial: 0

tx_csum_partial_inner: 0

tx_queue_stopped: 0

tx_queue_dropped: 0

tx_xmit_more: 0

tx_recover: 0

tx_cqes: 18

tx_queue_wake: 0

tx_udp_seg_rem: 0

tx_cqe_err: 0

tx_xdp_xmit: 0

tx_xdp_full: 0

tx_xdp_err: 0

tx_xdp_cqes: 0

rx_wqe_err: 0

rx_mpwqe_filler_cqes: 0

rx_mpwqe_filler_strides: 0

rx_buff_alloc_err: 0

rx_cqe_compress_blks: 0

rx_cqe_compress_pkts: 0

rx_page_reuse: 0

rx_cache_reuse: 0

rx_cache_full: 0

rx_cache_empty: 2688

rx_cache_busy: 0

rx_cache_waive: 0

rx_congst_umr: 0

rx_arfs_err: 0

ch_events: 75

ch_poll: 75

ch_arm: 75

ch_aff_change: 0

ch_eq_rearm: 0

rx_out_of_buffer: 0

rx_if_down_packets: 15

rx_steer_missed_packets: 0

rx_vport_unicast_packets: 0

rx_vport_unicast_bytes: 0

tx_vport_unicast_packets: 0

tx_vport_unicast_bytes: 0

rx_vport_multicast_packets: 2

rx_vport_multicast_bytes: 172

tx_vport_multicast_packets: 12

tx_vport_multicast_bytes: 936

rx_vport_broadcast_packets: 37

rx_vport_broadcast_bytes: 9270

tx_vport_broadcast_packets: 6

tx_vport_broadcast_bytes: 360

rx_vport_rdma_unicast_packets: 0

rx_vport_rdma_unicast_bytes: 0

tx_vport_rdma_unicast_packets: 0

tx_vport_rdma_unicast_bytes: 0

rx_vport_rdma_multicast_packets: 0

rx_vport_rdma_multicast_bytes: 0

tx_vport_rdma_multicast_packets: 0

tx_vport_rdma_multicast_bytes: 0

tx_packets_phy: 0

rx_packets_phy: 0

rx_crc_errors_phy: 0

tx_bytes_phy: 0

rx_bytes_phy: 0

tx_multicast_phy: 0

tx_broadcast_phy: 0

rx_multicast_phy: 0

rx_broadcast_phy: 0

rx_in_range_len_errors_phy: 0

rx_out_of_range_len_phy: 0

rx_oversize_pkts_phy: 0

rx_symbol_err_phy: 0

tx_mac_control_phy: 0

rx_mac_control_phy: 0

rx_unsupported_op_phy: 0

rx_pause_ctrl_phy: 0

tx_pause_ctrl_phy: 0

rx_discards_phy: 0

tx_discards_phy: 0

tx_errors_phy: 0

rx_undersize_pkts_phy: 0

rx_fragments_phy: 0

rx_jabbers_phy: 0

rx_64_bytes_phy: 0

rx_65_to_127_bytes_phy: 0

rx_128_to_255_bytes_phy: 0

rx_256_to_511_bytes_phy: 0

rx_512_to_1023_bytes_phy: 0

rx_1024_to_1518_bytes_phy: 0

rx_1519_to_2047_bytes_phy: 0

rx_2048_to_4095_bytes_phy: 0

rx_4096_to_8191_bytes_phy: 0

rx_8192_to_10239_bytes_phy: 0

link_down_events_phy: 0

rx_prio0_bytes: 0

rx_prio0_packets: 0

tx_prio0_bytes: 0

tx_prio0_packets: 0

rx_prio1_bytes: 0

rx_prio1_packets: 0

tx_prio1_bytes: 0

tx_prio1_packets: 0

rx_prio2_bytes: 0

rx_prio2_packets: 0

tx_prio2_bytes: 0

tx_prio2_packets: 0

rx_prio3_bytes: 0

rx_prio3_packets: 0

tx_prio3_bytes: 0

tx_prio3_packets: 0

rx_prio4_bytes: 0

rx_prio4_packets: 0

tx_prio4_bytes: 0

tx_prio4_packets: 0

rx_prio5_bytes: 0

rx_prio5_packets: 0

tx_prio5_bytes: 0

tx_prio5_packets: 0

rx_prio6_bytes: 0

rx_prio6_packets: 0

tx_prio6_bytes: 0

tx_prio6_packets: 0

rx_prio7_bytes: 0

rx_prio7_packets: 0

tx_prio7_bytes: 0

tx_prio7_packets: 0

module_unplug: 0

module_bus_stuck: 0

module_high_temp: 0

module_bad_shorted: 0

ch0_events: 9

ch0_poll: 9

ch0_arm: 9

ch0_aff_change: 0

ch0_eq_rearm: 0

ch1_events: 23

ch1_poll: 23

ch1_arm: 23

ch1_aff_change: 0

ch1_eq_rearm: 0

ch2_events: 8

ch2_poll: 8

ch2_arm: 8

ch2_aff_change: 0

ch2_eq_rearm: 0

ch3_events: 19

ch3_poll: 19

ch3_arm: 19

ch3_aff_change: 0

ch3_eq_rearm: 0

ch4_events: 8

ch4_poll: 8

ch4_arm: 8

ch4_aff_change: 0

ch4_eq_rearm: 0

ch5_events: 8

ch5_poll: 8

ch5_arm: 8

ch5_aff_change: 0

ch5_eq_rearm: 0

rx0_packets: 0

rx0_bytes: 0

rx0_csum_complete: 0

rx0_csum_unnecessary: 0

rx0_csum_unnecessary_inner: 0

rx0_csum_none: 0

rx0_xdp_drop: 0

rx0_xdp_redirect: 0

rx0_lro_packets: 0

rx0_lro_bytes: 0

rx0_ecn_mark: 0

rx0_removed_vlan_packets: 0

rx0_wqe_err: 0

rx0_mpwqe_filler_cqes: 0

rx0_mpwqe_filler_strides: 0

rx0_buff_alloc_err: 0

rx0_cqe_compress_blks: 0

rx0_cqe_compress_pkts: 0

rx0_page_reuse: 0

rx0_cache_reuse: 0

rx0_cache_full: 0

rx0_cache_empty: 448

rx0_cache_busy: 0

rx0_cache_waive: 0

rx0_congst_umr: 0

rx0_arfs_err: 0

rx0_xdp_tx_xmit: 0

rx0_xdp_tx_full: 0

rx0_xdp_tx_err: 0

rx0_xdp_tx_cqes: 0

rx1_packets: 10

rx1_bytes: 3420

rx1_csum_complete: 10

rx1_csum_unnecessary: 0

rx1_csum_unnecessary_inner: 0

rx1_csum_none: 0

rx1_xdp_drop: 0

rx1_xdp_redirect: 0

rx1_lro_packets: 0

rx1_lro_bytes: 0

rx1_ecn_mark: 0

rx1_removed_vlan_packets: 0

rx1_wqe_err: 0

rx1_mpwqe_filler_cqes: 0

rx1_mpwqe_filler_strides: 0

rx1_buff_alloc_err: 0

rx1_cqe_compress_blks: 0

rx1_cqe_compress_pkts: 0

rx1_page_reuse: 0

rx1_cache_reuse: 0

rx1_cache_full: 0

rx1_cache_empty: 448

rx1_cache_busy: 0

rx1_cache_waive: 0

rx1_congst_umr: 0

rx1_arfs_err: 0

rx1_xdp_tx_xmit: 0

rx1_xdp_tx_full: 0

rx1_xdp_tx_err: 0

rx1_xdp_tx_cqes: 0

rx2_packets: 0

rx2_bytes: 0

rx2_csum_complete: 0

rx2_csum_unnecessary: 0

rx2_csum_unnecessary_inner: 0

rx2_csum_none: 0

rx2_xdp_drop: 0

rx2_xdp_redirect: 0

rx2_lro_packets: 0

rx2_lro_bytes: 0

rx2_ecn_mark: 0

rx2_removed_vlan_packets: 0

rx2_wqe_err: 0

rx2_mpwqe_filler_cqes: 0

rx2_mpwqe_filler_strides: 0

rx2_buff_alloc_err: 0

rx2_cqe_compress_blks: 0

rx2_cqe_compress_pkts: 0

rx2_page_reuse: 0

rx2_cache_reuse: 0

rx2_cache_full: 0

rx2_cache_empty: 448

rx2_cache_busy: 0

rx2_cache_waive: 0

rx2_congst_umr: 0

rx2_arfs_err: 0

rx2_xdp_tx_xmit: 0

rx2_xdp_tx_full: 0

rx2_xdp_tx_err: 0

rx2_xdp_tx_cqes: 0

tx0_packets: 1

tx0_bytes: 60

tx0_tso_packets: 0

tx0_tso_bytes: 0

tx0_tso_inner_packets: 0

tx0_tso_inner_bytes: 0

tx0_csum_partial: 0

tx0_csum_partial_inner: 0

tx0_added_vlan_packets: 0

tx0_nop: 0

tx0_csum_none: 1

tx0_stopped: 0

tx0_dropped: 0

tx0_xmit_more: 0

tx0_recover: 0

tx0_cqes: 1

tx0_wake: 0

tx0_cqe_err: 0

tx1_packets: 5

tx1_bytes: 300

tx1_tso_packets: 0

tx1_tso_bytes: 0

tx1_tso_inner_packets: 0

tx1_tso_inner_bytes: 0

tx1_csum_partial: 0

tx1_csum_partial_inner: 0

tx1_added_vlan_packets: 0

tx1_nop: 0

tx1_csum_none: 5

tx1_stopped: 0

tx1_dropped: 0

tx1_xmit_more: 0

tx1_recover: 0

tx1_cqes: 5

tx1_wake: 0

tx1_cqe_err: 0

tx2_packets: 0

tx2_bytes: 0

tx2_tso_packets: 0

tx2_tso_bytes: 0

tx2_tso_inner_packets: 0

tx2_tso_inner_bytes: 0

tx2_csum_partial: 0

tx2_csum_partial_inner: 0

tx2_added_vlan_packets: 0

tx2_nop: 0

tx2_csum_none: 0

tx2_stopped: 0

tx2_dropped: 0

tx2_xmit_more: 0

tx2_recover: 0

tx2_cqes: 0

tx2_wake: 0

tx2_cqe_err: 0

The number of packets dropped due to XDP program XDP_DROP action. these packets are not counted by other software counters. These packets are counted by physical port and vPort counters – supported from kernel 4.9

?

總結

以上是生活随笔為你收集整理的【RDMA】mellanox RDMA网卡驱动安装和查看状态的全部內容,希望文章能夠幫你解決所遇到的問題。

如果覺得生活随笔網站內容還不錯,歡迎將生活随笔推薦給好友。