forked from luck/tmp_suning_uos_patched
mlx4: give precise rx/tx bytes/packets counters
mlx4 stats are chaotic because a deferred work queue is responsible to update them every 250 ms. Even sampling stats every one second with "sar -n DEV 1" gives variations like the following : lpaa23:~# sar -n DEV 1 10 | grep eth0 | cut -c1-65 07:39:22 eth0 146877.00 3265554.00 9467.15 4828168.50 07:39:23 eth0 146587.00 3260329.00 9448.15 4820445.98 07:39:24 eth0 146894.00 3259989.00 9468.55 4819943.26 07:39:25 eth0 110368.00 2454497.00 7113.95 3629012.17 <<>> 07:39:26 eth0 146563.00 3257502.00 9447.25 4816266.23 07:39:27 eth0 145678.00 3258292.00 9389.79 4817414.39 07:39:28 eth0 145268.00 3253171.00 9363.85 4809852.46 07:39:29 eth0 146439.00 3262185.00 9438.97 4823172.48 07:39:30 eth0 146758.00 3264175.00 9459.94 4826124.13 07:39:31 eth0 146843.00 3256903.00 9465.44 4815381.97 Average: eth0 142827.50 3179259.70 9206.30 4700578.16 This patch allows rx/tx bytes/packets counters being folded at the time we need stats. We now can fetch stats every 1 ms if we want to check NIC behavior on a small time window. It is also easier to detect anomalies. lpaa23:~# sar -n DEV 1 10 | grep eth0 | cut -c1-65 07:42:50 eth0 142915.00 3177696.00 9212.06 4698270.42 07:42:51 eth0 143741.00 3200232.00 9265.15 4731593.02 07:42:52 eth0 142781.00 3171600.00 9202.92 4689260.16 07:42:53 eth0 143835.00 3192932.00 9271.80 4720761.39 07:42:54 eth0 141922.00 3165174.00 9147.64 4679759.21 07:42:55 eth0 142993.00 3207038.00 9216.78 4741653.05 07:42:56 eth0 141394.06 3154335.64 9113.85 4663731.73 07:42:57 eth0 141850.00 3161202.00 9144.48 4673866.07 07:42:58 eth0 143439.00 3180736.00 9246.05 4702755.35 07:42:59 eth0 143501.00 3210992.00 9249.99 4747501.84 Average: eth0 142835.66 3182165.93 9206.98 4704874.08 Signed-off-by: Eric Dumazet <edumazet@google.com> Cc: Tariq Toukan <tariqt@mellanox.com> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
parent
31ac1c1945
commit
40931b8511
|
@ -367,6 +367,8 @@ static void mlx4_en_get_ethtool_stats(struct net_device *dev,
|
||||||
|
|
||||||
spin_lock_bh(&priv->stats_lock);
|
spin_lock_bh(&priv->stats_lock);
|
||||||
|
|
||||||
|
mlx4_en_fold_software_stats(dev);
|
||||||
|
|
||||||
for (i = 0; i < NUM_MAIN_STATS; i++, bitmap_iterator_inc(&it))
|
for (i = 0; i < NUM_MAIN_STATS; i++, bitmap_iterator_inc(&it))
|
||||||
if (bitmap_iterator_test(&it))
|
if (bitmap_iterator_test(&it))
|
||||||
data[index++] = ((unsigned long *)&dev->stats)[i];
|
data[index++] = ((unsigned long *)&dev->stats)[i];
|
||||||
|
|
|
@ -1324,6 +1324,7 @@ mlx4_en_get_stats64(struct net_device *dev, struct rtnl_link_stats64 *stats)
|
||||||
struct mlx4_en_priv *priv = netdev_priv(dev);
|
struct mlx4_en_priv *priv = netdev_priv(dev);
|
||||||
|
|
||||||
spin_lock_bh(&priv->stats_lock);
|
spin_lock_bh(&priv->stats_lock);
|
||||||
|
mlx4_en_fold_software_stats(dev);
|
||||||
netdev_stats_to_stats64(stats, &dev->stats);
|
netdev_stats_to_stats64(stats, &dev->stats);
|
||||||
spin_unlock_bh(&priv->stats_lock);
|
spin_unlock_bh(&priv->stats_lock);
|
||||||
|
|
||||||
|
|
|
@ -147,6 +147,39 @@ static unsigned long en_stats_adder(__be64 *start, __be64 *next, int num)
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void mlx4_en_fold_software_stats(struct net_device *dev)
|
||||||
|
{
|
||||||
|
struct mlx4_en_priv *priv = netdev_priv(dev);
|
||||||
|
struct mlx4_en_dev *mdev = priv->mdev;
|
||||||
|
unsigned long packets, bytes;
|
||||||
|
int i;
|
||||||
|
|
||||||
|
if (mlx4_is_master(mdev->dev))
|
||||||
|
return;
|
||||||
|
|
||||||
|
packets = 0;
|
||||||
|
bytes = 0;
|
||||||
|
for (i = 0; i < priv->rx_ring_num; i++) {
|
||||||
|
const struct mlx4_en_rx_ring *ring = priv->rx_ring[i];
|
||||||
|
|
||||||
|
packets += READ_ONCE(ring->packets);
|
||||||
|
bytes += READ_ONCE(ring->bytes);
|
||||||
|
}
|
||||||
|
dev->stats.rx_packets = packets;
|
||||||
|
dev->stats.rx_bytes = bytes;
|
||||||
|
|
||||||
|
packets = 0;
|
||||||
|
bytes = 0;
|
||||||
|
for (i = 0; i < priv->tx_ring_num[TX]; i++) {
|
||||||
|
const struct mlx4_en_tx_ring *ring = priv->tx_ring[TX][i];
|
||||||
|
|
||||||
|
packets += READ_ONCE(ring->packets);
|
||||||
|
bytes += READ_ONCE(ring->bytes);
|
||||||
|
}
|
||||||
|
dev->stats.tx_packets = packets;
|
||||||
|
dev->stats.tx_bytes = bytes;
|
||||||
|
}
|
||||||
|
|
||||||
int mlx4_en_DUMP_ETH_STATS(struct mlx4_en_dev *mdev, u8 port, u8 reset)
|
int mlx4_en_DUMP_ETH_STATS(struct mlx4_en_dev *mdev, u8 port, u8 reset)
|
||||||
{
|
{
|
||||||
struct mlx4_counter tmp_counter_stats;
|
struct mlx4_counter tmp_counter_stats;
|
||||||
|
@ -159,6 +192,7 @@ int mlx4_en_DUMP_ETH_STATS(struct mlx4_en_dev *mdev, u8 port, u8 reset)
|
||||||
u64 in_mod = reset << 8 | port;
|
u64 in_mod = reset << 8 | port;
|
||||||
int err;
|
int err;
|
||||||
int i, counter_index;
|
int i, counter_index;
|
||||||
|
unsigned long sw_tx_dropped = 0;
|
||||||
unsigned long sw_rx_dropped = 0;
|
unsigned long sw_rx_dropped = 0;
|
||||||
|
|
||||||
mailbox = mlx4_alloc_cmd_mailbox(mdev->dev);
|
mailbox = mlx4_alloc_cmd_mailbox(mdev->dev);
|
||||||
|
@ -174,8 +208,8 @@ int mlx4_en_DUMP_ETH_STATS(struct mlx4_en_dev *mdev, u8 port, u8 reset)
|
||||||
|
|
||||||
spin_lock_bh(&priv->stats_lock);
|
spin_lock_bh(&priv->stats_lock);
|
||||||
|
|
||||||
stats->rx_packets = 0;
|
mlx4_en_fold_software_stats(dev);
|
||||||
stats->rx_bytes = 0;
|
|
||||||
priv->port_stats.rx_chksum_good = 0;
|
priv->port_stats.rx_chksum_good = 0;
|
||||||
priv->port_stats.rx_chksum_none = 0;
|
priv->port_stats.rx_chksum_none = 0;
|
||||||
priv->port_stats.rx_chksum_complete = 0;
|
priv->port_stats.rx_chksum_complete = 0;
|
||||||
|
@ -183,19 +217,16 @@ int mlx4_en_DUMP_ETH_STATS(struct mlx4_en_dev *mdev, u8 port, u8 reset)
|
||||||
priv->xdp_stats.rx_xdp_tx = 0;
|
priv->xdp_stats.rx_xdp_tx = 0;
|
||||||
priv->xdp_stats.rx_xdp_tx_full = 0;
|
priv->xdp_stats.rx_xdp_tx_full = 0;
|
||||||
for (i = 0; i < priv->rx_ring_num; i++) {
|
for (i = 0; i < priv->rx_ring_num; i++) {
|
||||||
stats->rx_packets += priv->rx_ring[i]->packets;
|
const struct mlx4_en_rx_ring *ring = priv->rx_ring[i];
|
||||||
stats->rx_bytes += priv->rx_ring[i]->bytes;
|
|
||||||
sw_rx_dropped += priv->rx_ring[i]->dropped;
|
sw_rx_dropped += READ_ONCE(ring->dropped);
|
||||||
priv->port_stats.rx_chksum_good += priv->rx_ring[i]->csum_ok;
|
priv->port_stats.rx_chksum_good += READ_ONCE(ring->csum_ok);
|
||||||
priv->port_stats.rx_chksum_none += priv->rx_ring[i]->csum_none;
|
priv->port_stats.rx_chksum_none += READ_ONCE(ring->csum_none);
|
||||||
priv->port_stats.rx_chksum_complete += priv->rx_ring[i]->csum_complete;
|
priv->port_stats.rx_chksum_complete += READ_ONCE(ring->csum_complete);
|
||||||
priv->xdp_stats.rx_xdp_drop += priv->rx_ring[i]->xdp_drop;
|
priv->xdp_stats.rx_xdp_drop += READ_ONCE(ring->xdp_drop);
|
||||||
priv->xdp_stats.rx_xdp_tx += priv->rx_ring[i]->xdp_tx;
|
priv->xdp_stats.rx_xdp_tx += READ_ONCE(ring->xdp_tx);
|
||||||
priv->xdp_stats.rx_xdp_tx_full += priv->rx_ring[i]->xdp_tx_full;
|
priv->xdp_stats.rx_xdp_tx_full += READ_ONCE(ring->xdp_tx_full);
|
||||||
}
|
}
|
||||||
stats->tx_packets = 0;
|
|
||||||
stats->tx_bytes = 0;
|
|
||||||
stats->tx_dropped = 0;
|
|
||||||
priv->port_stats.tx_chksum_offload = 0;
|
priv->port_stats.tx_chksum_offload = 0;
|
||||||
priv->port_stats.queue_stopped = 0;
|
priv->port_stats.queue_stopped = 0;
|
||||||
priv->port_stats.wake_queue = 0;
|
priv->port_stats.wake_queue = 0;
|
||||||
|
@ -205,15 +236,14 @@ int mlx4_en_DUMP_ETH_STATS(struct mlx4_en_dev *mdev, u8 port, u8 reset)
|
||||||
for (i = 0; i < priv->tx_ring_num[TX]; i++) {
|
for (i = 0; i < priv->tx_ring_num[TX]; i++) {
|
||||||
const struct mlx4_en_tx_ring *ring = priv->tx_ring[TX][i];
|
const struct mlx4_en_tx_ring *ring = priv->tx_ring[TX][i];
|
||||||
|
|
||||||
stats->tx_packets += ring->packets;
|
sw_tx_dropped += READ_ONCE(ring->tx_dropped);
|
||||||
stats->tx_bytes += ring->bytes;
|
priv->port_stats.tx_chksum_offload += READ_ONCE(ring->tx_csum);
|
||||||
stats->tx_dropped += ring->tx_dropped;
|
priv->port_stats.queue_stopped += READ_ONCE(ring->queue_stopped);
|
||||||
priv->port_stats.tx_chksum_offload += ring->tx_csum;
|
priv->port_stats.wake_queue += READ_ONCE(ring->wake_queue);
|
||||||
priv->port_stats.queue_stopped += ring->queue_stopped;
|
priv->port_stats.tso_packets += READ_ONCE(ring->tso_packets);
|
||||||
priv->port_stats.wake_queue += ring->wake_queue;
|
priv->port_stats.xmit_more += READ_ONCE(ring->xmit_more);
|
||||||
priv->port_stats.tso_packets += ring->tso_packets;
|
|
||||||
priv->port_stats.xmit_more += ring->xmit_more;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if (mlx4_is_master(mdev->dev)) {
|
if (mlx4_is_master(mdev->dev)) {
|
||||||
stats->rx_packets = en_stats_adder(&mlx4_en_stats->RTOT_prio_0,
|
stats->rx_packets = en_stats_adder(&mlx4_en_stats->RTOT_prio_0,
|
||||||
&mlx4_en_stats->RTOT_prio_1,
|
&mlx4_en_stats->RTOT_prio_1,
|
||||||
|
@ -251,7 +281,8 @@ int mlx4_en_DUMP_ETH_STATS(struct mlx4_en_dev *mdev, u8 port, u8 reset)
|
||||||
stats->rx_length_errors = be32_to_cpu(mlx4_en_stats->RdropLength);
|
stats->rx_length_errors = be32_to_cpu(mlx4_en_stats->RdropLength);
|
||||||
stats->rx_crc_errors = be32_to_cpu(mlx4_en_stats->RCRC);
|
stats->rx_crc_errors = be32_to_cpu(mlx4_en_stats->RCRC);
|
||||||
stats->rx_fifo_errors = be32_to_cpu(mlx4_en_stats->RdropOvflw);
|
stats->rx_fifo_errors = be32_to_cpu(mlx4_en_stats->RdropOvflw);
|
||||||
stats->tx_dropped += be32_to_cpu(mlx4_en_stats->TDROP);
|
stats->tx_dropped = be32_to_cpu(mlx4_en_stats->TDROP) +
|
||||||
|
sw_tx_dropped;
|
||||||
|
|
||||||
/* RX stats */
|
/* RX stats */
|
||||||
priv->pkstats.rx_multicast_packets = stats->multicast;
|
priv->pkstats.rx_multicast_packets = stats->multicast;
|
||||||
|
|
|
@ -755,6 +755,7 @@ void mlx4_en_rx_irq(struct mlx4_cq *mcq);
|
||||||
int mlx4_SET_MCAST_FLTR(struct mlx4_dev *dev, u8 port, u64 mac, u64 clear, u8 mode);
|
int mlx4_SET_MCAST_FLTR(struct mlx4_dev *dev, u8 port, u64 mac, u64 clear, u8 mode);
|
||||||
int mlx4_SET_VLAN_FLTR(struct mlx4_dev *dev, struct mlx4_en_priv *priv);
|
int mlx4_SET_VLAN_FLTR(struct mlx4_dev *dev, struct mlx4_en_priv *priv);
|
||||||
|
|
||||||
|
void mlx4_en_fold_software_stats(struct net_device *dev);
|
||||||
int mlx4_en_DUMP_ETH_STATS(struct mlx4_en_dev *mdev, u8 port, u8 reset);
|
int mlx4_en_DUMP_ETH_STATS(struct mlx4_en_dev *mdev, u8 port, u8 reset);
|
||||||
int mlx4_en_QUERY_PORT(struct mlx4_en_dev *mdev, u8 port);
|
int mlx4_en_QUERY_PORT(struct mlx4_en_dev *mdev, u8 port);
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue
Block a user