forked from luck/tmp_suning_uos_patched
f646968f8f
Rename the hardware VLAN acceleration features to include "CTAG" to indicate that they only support CTAGs. Follow up patches will introduce 802.1ad server provider tagging (STAGs) and require the distinction for hardware not supporting acclerating both. Signed-off-by: Patrick McHardy <kaber@trash.net> Signed-off-by: David S. Miller <davem@davemloft.net>
640 lines
19 KiB
C
640 lines
19 KiB
C
/*
|
|
* Linux driver for VMware's vmxnet3 ethernet NIC.
|
|
*
|
|
* Copyright (C) 2008-2009, VMware, Inc. All Rights Reserved.
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
* under the terms of the GNU General Public License as published by the
|
|
* Free Software Foundation; version 2 of the License and no later version.
|
|
*
|
|
* This program is distributed in the hope that it will be useful, but
|
|
* WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or
|
|
* NON INFRINGEMENT. See the GNU General Public License for more
|
|
* details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
|
|
*
|
|
* The full GNU General Public License is included in this distribution in
|
|
* the file called "COPYING".
|
|
*
|
|
* Maintained by: Shreyas Bhatewara <pv-drivers@vmware.com>
|
|
*
|
|
*/
|
|
|
|
|
|
#include "vmxnet3_int.h"
|
|
|
|
struct vmxnet3_stat_desc {
|
|
char desc[ETH_GSTRING_LEN];
|
|
int offset;
|
|
};
|
|
|
|
|
|
/* per tq stats maintained by the device */
|
|
static const struct vmxnet3_stat_desc
|
|
vmxnet3_tq_dev_stats[] = {
|
|
/* description, offset */
|
|
{ "Tx Queue#", 0 },
|
|
{ " TSO pkts tx", offsetof(struct UPT1_TxStats, TSOPktsTxOK) },
|
|
{ " TSO bytes tx", offsetof(struct UPT1_TxStats, TSOBytesTxOK) },
|
|
{ " ucast pkts tx", offsetof(struct UPT1_TxStats, ucastPktsTxOK) },
|
|
{ " ucast bytes tx", offsetof(struct UPT1_TxStats, ucastBytesTxOK) },
|
|
{ " mcast pkts tx", offsetof(struct UPT1_TxStats, mcastPktsTxOK) },
|
|
{ " mcast bytes tx", offsetof(struct UPT1_TxStats, mcastBytesTxOK) },
|
|
{ " bcast pkts tx", offsetof(struct UPT1_TxStats, bcastPktsTxOK) },
|
|
{ " bcast bytes tx", offsetof(struct UPT1_TxStats, bcastBytesTxOK) },
|
|
{ " pkts tx err", offsetof(struct UPT1_TxStats, pktsTxError) },
|
|
{ " pkts tx discard", offsetof(struct UPT1_TxStats, pktsTxDiscard) },
|
|
};
|
|
|
|
/* per tq stats maintained by the driver */
|
|
static const struct vmxnet3_stat_desc
|
|
vmxnet3_tq_driver_stats[] = {
|
|
/* description, offset */
|
|
{" drv dropped tx total", offsetof(struct vmxnet3_tq_driver_stats,
|
|
drop_total) },
|
|
{ " too many frags", offsetof(struct vmxnet3_tq_driver_stats,
|
|
drop_too_many_frags) },
|
|
{ " giant hdr", offsetof(struct vmxnet3_tq_driver_stats,
|
|
drop_oversized_hdr) },
|
|
{ " hdr err", offsetof(struct vmxnet3_tq_driver_stats,
|
|
drop_hdr_inspect_err) },
|
|
{ " tso", offsetof(struct vmxnet3_tq_driver_stats,
|
|
drop_tso) },
|
|
{ " ring full", offsetof(struct vmxnet3_tq_driver_stats,
|
|
tx_ring_full) },
|
|
{ " pkts linearized", offsetof(struct vmxnet3_tq_driver_stats,
|
|
linearized) },
|
|
{ " hdr cloned", offsetof(struct vmxnet3_tq_driver_stats,
|
|
copy_skb_header) },
|
|
{ " giant hdr", offsetof(struct vmxnet3_tq_driver_stats,
|
|
oversized_hdr) },
|
|
};
|
|
|
|
/* per rq stats maintained by the device */
|
|
static const struct vmxnet3_stat_desc
|
|
vmxnet3_rq_dev_stats[] = {
|
|
{ "Rx Queue#", 0 },
|
|
{ " LRO pkts rx", offsetof(struct UPT1_RxStats, LROPktsRxOK) },
|
|
{ " LRO byte rx", offsetof(struct UPT1_RxStats, LROBytesRxOK) },
|
|
{ " ucast pkts rx", offsetof(struct UPT1_RxStats, ucastPktsRxOK) },
|
|
{ " ucast bytes rx", offsetof(struct UPT1_RxStats, ucastBytesRxOK) },
|
|
{ " mcast pkts rx", offsetof(struct UPT1_RxStats, mcastPktsRxOK) },
|
|
{ " mcast bytes rx", offsetof(struct UPT1_RxStats, mcastBytesRxOK) },
|
|
{ " bcast pkts rx", offsetof(struct UPT1_RxStats, bcastPktsRxOK) },
|
|
{ " bcast bytes rx", offsetof(struct UPT1_RxStats, bcastBytesRxOK) },
|
|
{ " pkts rx OOB", offsetof(struct UPT1_RxStats, pktsRxOutOfBuf) },
|
|
{ " pkts rx err", offsetof(struct UPT1_RxStats, pktsRxError) },
|
|
};
|
|
|
|
/* per rq stats maintained by the driver */
|
|
static const struct vmxnet3_stat_desc
|
|
vmxnet3_rq_driver_stats[] = {
|
|
/* description, offset */
|
|
{ " drv dropped rx total", offsetof(struct vmxnet3_rq_driver_stats,
|
|
drop_total) },
|
|
{ " err", offsetof(struct vmxnet3_rq_driver_stats,
|
|
drop_err) },
|
|
{ " fcs", offsetof(struct vmxnet3_rq_driver_stats,
|
|
drop_fcs) },
|
|
{ " rx buf alloc fail", offsetof(struct vmxnet3_rq_driver_stats,
|
|
rx_buf_alloc_failure) },
|
|
};
|
|
|
|
/* gloabl stats maintained by the driver */
|
|
static const struct vmxnet3_stat_desc
|
|
vmxnet3_global_stats[] = {
|
|
/* description, offset */
|
|
{ "tx timeout count", offsetof(struct vmxnet3_adapter,
|
|
tx_timeout_count) }
|
|
};
|
|
|
|
|
|
struct rtnl_link_stats64 *
|
|
vmxnet3_get_stats64(struct net_device *netdev,
|
|
struct rtnl_link_stats64 *stats)
|
|
{
|
|
struct vmxnet3_adapter *adapter;
|
|
struct vmxnet3_tq_driver_stats *drvTxStats;
|
|
struct vmxnet3_rq_driver_stats *drvRxStats;
|
|
struct UPT1_TxStats *devTxStats;
|
|
struct UPT1_RxStats *devRxStats;
|
|
unsigned long flags;
|
|
int i;
|
|
|
|
adapter = netdev_priv(netdev);
|
|
|
|
/* Collect the dev stats into the shared area */
|
|
spin_lock_irqsave(&adapter->cmd_lock, flags);
|
|
VMXNET3_WRITE_BAR1_REG(adapter, VMXNET3_REG_CMD, VMXNET3_CMD_GET_STATS);
|
|
spin_unlock_irqrestore(&adapter->cmd_lock, flags);
|
|
|
|
for (i = 0; i < adapter->num_tx_queues; i++) {
|
|
devTxStats = &adapter->tqd_start[i].stats;
|
|
drvTxStats = &adapter->tx_queue[i].stats;
|
|
stats->tx_packets += devTxStats->ucastPktsTxOK +
|
|
devTxStats->mcastPktsTxOK +
|
|
devTxStats->bcastPktsTxOK;
|
|
stats->tx_bytes += devTxStats->ucastBytesTxOK +
|
|
devTxStats->mcastBytesTxOK +
|
|
devTxStats->bcastBytesTxOK;
|
|
stats->tx_errors += devTxStats->pktsTxError;
|
|
stats->tx_dropped += drvTxStats->drop_total;
|
|
}
|
|
|
|
for (i = 0; i < adapter->num_rx_queues; i++) {
|
|
devRxStats = &adapter->rqd_start[i].stats;
|
|
drvRxStats = &adapter->rx_queue[i].stats;
|
|
stats->rx_packets += devRxStats->ucastPktsRxOK +
|
|
devRxStats->mcastPktsRxOK +
|
|
devRxStats->bcastPktsRxOK;
|
|
|
|
stats->rx_bytes += devRxStats->ucastBytesRxOK +
|
|
devRxStats->mcastBytesRxOK +
|
|
devRxStats->bcastBytesRxOK;
|
|
|
|
stats->rx_errors += devRxStats->pktsRxError;
|
|
stats->rx_dropped += drvRxStats->drop_total;
|
|
stats->multicast += devRxStats->mcastPktsRxOK;
|
|
}
|
|
|
|
return stats;
|
|
}
|
|
|
|
static int
|
|
vmxnet3_get_sset_count(struct net_device *netdev, int sset)
|
|
{
|
|
struct vmxnet3_adapter *adapter = netdev_priv(netdev);
|
|
switch (sset) {
|
|
case ETH_SS_STATS:
|
|
return (ARRAY_SIZE(vmxnet3_tq_dev_stats) +
|
|
ARRAY_SIZE(vmxnet3_tq_driver_stats)) *
|
|
adapter->num_tx_queues +
|
|
(ARRAY_SIZE(vmxnet3_rq_dev_stats) +
|
|
ARRAY_SIZE(vmxnet3_rq_driver_stats)) *
|
|
adapter->num_rx_queues +
|
|
ARRAY_SIZE(vmxnet3_global_stats);
|
|
default:
|
|
return -EOPNOTSUPP;
|
|
}
|
|
}
|
|
|
|
|
|
/* Should be multiple of 4 */
|
|
#define NUM_TX_REGS 8
|
|
#define NUM_RX_REGS 12
|
|
|
|
static int
|
|
vmxnet3_get_regs_len(struct net_device *netdev)
|
|
{
|
|
struct vmxnet3_adapter *adapter = netdev_priv(netdev);
|
|
return (adapter->num_tx_queues * NUM_TX_REGS * sizeof(u32) +
|
|
adapter->num_rx_queues * NUM_RX_REGS * sizeof(u32));
|
|
}
|
|
|
|
|
|
static void
|
|
vmxnet3_get_drvinfo(struct net_device *netdev, struct ethtool_drvinfo *drvinfo)
|
|
{
|
|
struct vmxnet3_adapter *adapter = netdev_priv(netdev);
|
|
|
|
strlcpy(drvinfo->driver, vmxnet3_driver_name, sizeof(drvinfo->driver));
|
|
|
|
strlcpy(drvinfo->version, VMXNET3_DRIVER_VERSION_REPORT,
|
|
sizeof(drvinfo->version));
|
|
|
|
strlcpy(drvinfo->bus_info, pci_name(adapter->pdev),
|
|
sizeof(drvinfo->bus_info));
|
|
drvinfo->n_stats = vmxnet3_get_sset_count(netdev, ETH_SS_STATS);
|
|
drvinfo->testinfo_len = 0;
|
|
drvinfo->eedump_len = 0;
|
|
drvinfo->regdump_len = vmxnet3_get_regs_len(netdev);
|
|
}
|
|
|
|
|
|
static void
|
|
vmxnet3_get_strings(struct net_device *netdev, u32 stringset, u8 *buf)
|
|
{
|
|
struct vmxnet3_adapter *adapter = netdev_priv(netdev);
|
|
if (stringset == ETH_SS_STATS) {
|
|
int i, j;
|
|
for (j = 0; j < adapter->num_tx_queues; j++) {
|
|
for (i = 0; i < ARRAY_SIZE(vmxnet3_tq_dev_stats); i++) {
|
|
memcpy(buf, vmxnet3_tq_dev_stats[i].desc,
|
|
ETH_GSTRING_LEN);
|
|
buf += ETH_GSTRING_LEN;
|
|
}
|
|
for (i = 0; i < ARRAY_SIZE(vmxnet3_tq_driver_stats);
|
|
i++) {
|
|
memcpy(buf, vmxnet3_tq_driver_stats[i].desc,
|
|
ETH_GSTRING_LEN);
|
|
buf += ETH_GSTRING_LEN;
|
|
}
|
|
}
|
|
|
|
for (j = 0; j < adapter->num_rx_queues; j++) {
|
|
for (i = 0; i < ARRAY_SIZE(vmxnet3_rq_dev_stats); i++) {
|
|
memcpy(buf, vmxnet3_rq_dev_stats[i].desc,
|
|
ETH_GSTRING_LEN);
|
|
buf += ETH_GSTRING_LEN;
|
|
}
|
|
for (i = 0; i < ARRAY_SIZE(vmxnet3_rq_driver_stats);
|
|
i++) {
|
|
memcpy(buf, vmxnet3_rq_driver_stats[i].desc,
|
|
ETH_GSTRING_LEN);
|
|
buf += ETH_GSTRING_LEN;
|
|
}
|
|
}
|
|
|
|
for (i = 0; i < ARRAY_SIZE(vmxnet3_global_stats); i++) {
|
|
memcpy(buf, vmxnet3_global_stats[i].desc,
|
|
ETH_GSTRING_LEN);
|
|
buf += ETH_GSTRING_LEN;
|
|
}
|
|
}
|
|
}
|
|
|
|
int vmxnet3_set_features(struct net_device *netdev, netdev_features_t features)
|
|
{
|
|
struct vmxnet3_adapter *adapter = netdev_priv(netdev);
|
|
unsigned long flags;
|
|
netdev_features_t changed = features ^ netdev->features;
|
|
|
|
if (changed & (NETIF_F_RXCSUM | NETIF_F_LRO |
|
|
NETIF_F_HW_VLAN_CTAG_RX)) {
|
|
if (features & NETIF_F_RXCSUM)
|
|
adapter->shared->devRead.misc.uptFeatures |=
|
|
UPT1_F_RXCSUM;
|
|
else
|
|
adapter->shared->devRead.misc.uptFeatures &=
|
|
~UPT1_F_RXCSUM;
|
|
|
|
/* update harware LRO capability accordingly */
|
|
if (features & NETIF_F_LRO)
|
|
adapter->shared->devRead.misc.uptFeatures |=
|
|
UPT1_F_LRO;
|
|
else
|
|
adapter->shared->devRead.misc.uptFeatures &=
|
|
~UPT1_F_LRO;
|
|
|
|
if (features & NETIF_F_HW_VLAN_CTAG_RX)
|
|
adapter->shared->devRead.misc.uptFeatures |=
|
|
UPT1_F_RXVLAN;
|
|
else
|
|
adapter->shared->devRead.misc.uptFeatures &=
|
|
~UPT1_F_RXVLAN;
|
|
|
|
spin_lock_irqsave(&adapter->cmd_lock, flags);
|
|
VMXNET3_WRITE_BAR1_REG(adapter, VMXNET3_REG_CMD,
|
|
VMXNET3_CMD_UPDATE_FEATURE);
|
|
spin_unlock_irqrestore(&adapter->cmd_lock, flags);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static void
|
|
vmxnet3_get_ethtool_stats(struct net_device *netdev,
|
|
struct ethtool_stats *stats, u64 *buf)
|
|
{
|
|
struct vmxnet3_adapter *adapter = netdev_priv(netdev);
|
|
unsigned long flags;
|
|
u8 *base;
|
|
int i;
|
|
int j = 0;
|
|
|
|
spin_lock_irqsave(&adapter->cmd_lock, flags);
|
|
VMXNET3_WRITE_BAR1_REG(adapter, VMXNET3_REG_CMD, VMXNET3_CMD_GET_STATS);
|
|
spin_unlock_irqrestore(&adapter->cmd_lock, flags);
|
|
|
|
/* this does assume each counter is 64-bit wide */
|
|
for (j = 0; j < adapter->num_tx_queues; j++) {
|
|
base = (u8 *)&adapter->tqd_start[j].stats;
|
|
*buf++ = (u64)j;
|
|
for (i = 1; i < ARRAY_SIZE(vmxnet3_tq_dev_stats); i++)
|
|
*buf++ = *(u64 *)(base +
|
|
vmxnet3_tq_dev_stats[i].offset);
|
|
|
|
base = (u8 *)&adapter->tx_queue[j].stats;
|
|
for (i = 0; i < ARRAY_SIZE(vmxnet3_tq_driver_stats); i++)
|
|
*buf++ = *(u64 *)(base +
|
|
vmxnet3_tq_driver_stats[i].offset);
|
|
}
|
|
|
|
for (j = 0; j < adapter->num_tx_queues; j++) {
|
|
base = (u8 *)&adapter->rqd_start[j].stats;
|
|
*buf++ = (u64) j;
|
|
for (i = 1; i < ARRAY_SIZE(vmxnet3_rq_dev_stats); i++)
|
|
*buf++ = *(u64 *)(base +
|
|
vmxnet3_rq_dev_stats[i].offset);
|
|
|
|
base = (u8 *)&adapter->rx_queue[j].stats;
|
|
for (i = 0; i < ARRAY_SIZE(vmxnet3_rq_driver_stats); i++)
|
|
*buf++ = *(u64 *)(base +
|
|
vmxnet3_rq_driver_stats[i].offset);
|
|
}
|
|
|
|
base = (u8 *)adapter;
|
|
for (i = 0; i < ARRAY_SIZE(vmxnet3_global_stats); i++)
|
|
*buf++ = *(u64 *)(base + vmxnet3_global_stats[i].offset);
|
|
}
|
|
|
|
|
|
static void
|
|
vmxnet3_get_regs(struct net_device *netdev, struct ethtool_regs *regs, void *p)
|
|
{
|
|
struct vmxnet3_adapter *adapter = netdev_priv(netdev);
|
|
u32 *buf = p;
|
|
int i = 0, j = 0;
|
|
|
|
memset(p, 0, vmxnet3_get_regs_len(netdev));
|
|
|
|
regs->version = 1;
|
|
|
|
/* Update vmxnet3_get_regs_len if we want to dump more registers */
|
|
|
|
/* make each ring use multiple of 16 bytes */
|
|
for (i = 0; i < adapter->num_tx_queues; i++) {
|
|
buf[j++] = adapter->tx_queue[i].tx_ring.next2fill;
|
|
buf[j++] = adapter->tx_queue[i].tx_ring.next2comp;
|
|
buf[j++] = adapter->tx_queue[i].tx_ring.gen;
|
|
buf[j++] = 0;
|
|
|
|
buf[j++] = adapter->tx_queue[i].comp_ring.next2proc;
|
|
buf[j++] = adapter->tx_queue[i].comp_ring.gen;
|
|
buf[j++] = adapter->tx_queue[i].stopped;
|
|
buf[j++] = 0;
|
|
}
|
|
|
|
for (i = 0; i < adapter->num_rx_queues; i++) {
|
|
buf[j++] = adapter->rx_queue[i].rx_ring[0].next2fill;
|
|
buf[j++] = adapter->rx_queue[i].rx_ring[0].next2comp;
|
|
buf[j++] = adapter->rx_queue[i].rx_ring[0].gen;
|
|
buf[j++] = 0;
|
|
|
|
buf[j++] = adapter->rx_queue[i].rx_ring[1].next2fill;
|
|
buf[j++] = adapter->rx_queue[i].rx_ring[1].next2comp;
|
|
buf[j++] = adapter->rx_queue[i].rx_ring[1].gen;
|
|
buf[j++] = 0;
|
|
|
|
buf[j++] = adapter->rx_queue[i].comp_ring.next2proc;
|
|
buf[j++] = adapter->rx_queue[i].comp_ring.gen;
|
|
buf[j++] = 0;
|
|
buf[j++] = 0;
|
|
}
|
|
|
|
}
|
|
|
|
|
|
static void
|
|
vmxnet3_get_wol(struct net_device *netdev, struct ethtool_wolinfo *wol)
|
|
{
|
|
struct vmxnet3_adapter *adapter = netdev_priv(netdev);
|
|
|
|
wol->supported = WAKE_UCAST | WAKE_ARP | WAKE_MAGIC;
|
|
wol->wolopts = adapter->wol;
|
|
}
|
|
|
|
|
|
static int
|
|
vmxnet3_set_wol(struct net_device *netdev, struct ethtool_wolinfo *wol)
|
|
{
|
|
struct vmxnet3_adapter *adapter = netdev_priv(netdev);
|
|
|
|
if (wol->wolopts & (WAKE_PHY | WAKE_MCAST | WAKE_BCAST |
|
|
WAKE_MAGICSECURE)) {
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
adapter->wol = wol->wolopts;
|
|
|
|
device_set_wakeup_enable(&adapter->pdev->dev, adapter->wol);
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
static int
|
|
vmxnet3_get_settings(struct net_device *netdev, struct ethtool_cmd *ecmd)
|
|
{
|
|
struct vmxnet3_adapter *adapter = netdev_priv(netdev);
|
|
|
|
ecmd->supported = SUPPORTED_10000baseT_Full | SUPPORTED_1000baseT_Full |
|
|
SUPPORTED_TP;
|
|
ecmd->advertising = ADVERTISED_TP;
|
|
ecmd->port = PORT_TP;
|
|
ecmd->transceiver = XCVR_INTERNAL;
|
|
|
|
if (adapter->link_speed) {
|
|
ethtool_cmd_speed_set(ecmd, adapter->link_speed);
|
|
ecmd->duplex = DUPLEX_FULL;
|
|
} else {
|
|
ethtool_cmd_speed_set(ecmd, -1);
|
|
ecmd->duplex = -1;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
|
|
static void
|
|
vmxnet3_get_ringparam(struct net_device *netdev,
|
|
struct ethtool_ringparam *param)
|
|
{
|
|
struct vmxnet3_adapter *adapter = netdev_priv(netdev);
|
|
|
|
param->rx_max_pending = VMXNET3_RX_RING_MAX_SIZE;
|
|
param->tx_max_pending = VMXNET3_TX_RING_MAX_SIZE;
|
|
param->rx_mini_max_pending = 0;
|
|
param->rx_jumbo_max_pending = 0;
|
|
|
|
param->rx_pending = adapter->rx_queue[0].rx_ring[0].size;
|
|
param->tx_pending = adapter->tx_queue[0].tx_ring.size;
|
|
param->rx_mini_pending = 0;
|
|
param->rx_jumbo_pending = 0;
|
|
}
|
|
|
|
|
|
static int
|
|
vmxnet3_set_ringparam(struct net_device *netdev,
|
|
struct ethtool_ringparam *param)
|
|
{
|
|
struct vmxnet3_adapter *adapter = netdev_priv(netdev);
|
|
u32 new_tx_ring_size, new_rx_ring_size;
|
|
u32 sz;
|
|
int err = 0;
|
|
|
|
if (param->tx_pending == 0 || param->tx_pending >
|
|
VMXNET3_TX_RING_MAX_SIZE)
|
|
return -EINVAL;
|
|
|
|
if (param->rx_pending == 0 || param->rx_pending >
|
|
VMXNET3_RX_RING_MAX_SIZE)
|
|
return -EINVAL;
|
|
|
|
/* if adapter not yet initialized, do nothing */
|
|
if (adapter->rx_buf_per_pkt == 0) {
|
|
netdev_err(netdev, "adapter not completely initialized, "
|
|
"ring size cannot be changed yet\n");
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
/* round it up to a multiple of VMXNET3_RING_SIZE_ALIGN */
|
|
new_tx_ring_size = (param->tx_pending + VMXNET3_RING_SIZE_MASK) &
|
|
~VMXNET3_RING_SIZE_MASK;
|
|
new_tx_ring_size = min_t(u32, new_tx_ring_size,
|
|
VMXNET3_TX_RING_MAX_SIZE);
|
|
if (new_tx_ring_size > VMXNET3_TX_RING_MAX_SIZE || (new_tx_ring_size %
|
|
VMXNET3_RING_SIZE_ALIGN) != 0)
|
|
return -EINVAL;
|
|
|
|
/* ring0 has to be a multiple of
|
|
* rx_buf_per_pkt * VMXNET3_RING_SIZE_ALIGN
|
|
*/
|
|
sz = adapter->rx_buf_per_pkt * VMXNET3_RING_SIZE_ALIGN;
|
|
new_rx_ring_size = (param->rx_pending + sz - 1) / sz * sz;
|
|
new_rx_ring_size = min_t(u32, new_rx_ring_size,
|
|
VMXNET3_RX_RING_MAX_SIZE / sz * sz);
|
|
if (new_rx_ring_size > VMXNET3_RX_RING_MAX_SIZE || (new_rx_ring_size %
|
|
sz) != 0)
|
|
return -EINVAL;
|
|
|
|
if (new_tx_ring_size == adapter->tx_queue[0].tx_ring.size &&
|
|
new_rx_ring_size == adapter->rx_queue[0].rx_ring[0].size) {
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Reset_work may be in the middle of resetting the device, wait for its
|
|
* completion.
|
|
*/
|
|
while (test_and_set_bit(VMXNET3_STATE_BIT_RESETTING, &adapter->state))
|
|
msleep(1);
|
|
|
|
if (netif_running(netdev)) {
|
|
vmxnet3_quiesce_dev(adapter);
|
|
vmxnet3_reset_dev(adapter);
|
|
|
|
/* recreate the rx queue and the tx queue based on the
|
|
* new sizes */
|
|
vmxnet3_tq_destroy_all(adapter);
|
|
vmxnet3_rq_destroy_all(adapter);
|
|
|
|
err = vmxnet3_create_queues(adapter, new_tx_ring_size,
|
|
new_rx_ring_size, VMXNET3_DEF_RX_RING_SIZE);
|
|
|
|
if (err) {
|
|
/* failed, most likely because of OOM, try default
|
|
* size */
|
|
netdev_err(netdev, "failed to apply new sizes, "
|
|
"try the default ones\n");
|
|
err = vmxnet3_create_queues(adapter,
|
|
VMXNET3_DEF_TX_RING_SIZE,
|
|
VMXNET3_DEF_RX_RING_SIZE,
|
|
VMXNET3_DEF_RX_RING_SIZE);
|
|
if (err) {
|
|
netdev_err(netdev, "failed to create queues "
|
|
"with default sizes. Closing it\n");
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
err = vmxnet3_activate_dev(adapter);
|
|
if (err)
|
|
netdev_err(netdev, "failed to re-activate, error %d."
|
|
" Closing it\n", err);
|
|
}
|
|
|
|
out:
|
|
clear_bit(VMXNET3_STATE_BIT_RESETTING, &adapter->state);
|
|
if (err)
|
|
vmxnet3_force_close(adapter);
|
|
|
|
return err;
|
|
}
|
|
|
|
|
|
static int
|
|
vmxnet3_get_rxnfc(struct net_device *netdev, struct ethtool_rxnfc *info,
|
|
u32 *rules)
|
|
{
|
|
struct vmxnet3_adapter *adapter = netdev_priv(netdev);
|
|
switch (info->cmd) {
|
|
case ETHTOOL_GRXRINGS:
|
|
info->data = adapter->num_rx_queues;
|
|
return 0;
|
|
}
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
#ifdef VMXNET3_RSS
|
|
static u32
|
|
vmxnet3_get_rss_indir_size(struct net_device *netdev)
|
|
{
|
|
struct vmxnet3_adapter *adapter = netdev_priv(netdev);
|
|
struct UPT1_RSSConf *rssConf = adapter->rss_conf;
|
|
|
|
return rssConf->indTableSize;
|
|
}
|
|
|
|
static int
|
|
vmxnet3_get_rss_indir(struct net_device *netdev, u32 *p)
|
|
{
|
|
struct vmxnet3_adapter *adapter = netdev_priv(netdev);
|
|
struct UPT1_RSSConf *rssConf = adapter->rss_conf;
|
|
unsigned int n = rssConf->indTableSize;
|
|
|
|
while (n--)
|
|
p[n] = rssConf->indTable[n];
|
|
return 0;
|
|
|
|
}
|
|
|
|
static int
|
|
vmxnet3_set_rss_indir(struct net_device *netdev, const u32 *p)
|
|
{
|
|
unsigned int i;
|
|
unsigned long flags;
|
|
struct vmxnet3_adapter *adapter = netdev_priv(netdev);
|
|
struct UPT1_RSSConf *rssConf = adapter->rss_conf;
|
|
|
|
for (i = 0; i < rssConf->indTableSize; i++)
|
|
rssConf->indTable[i] = p[i];
|
|
|
|
spin_lock_irqsave(&adapter->cmd_lock, flags);
|
|
VMXNET3_WRITE_BAR1_REG(adapter, VMXNET3_REG_CMD,
|
|
VMXNET3_CMD_UPDATE_RSSIDT);
|
|
spin_unlock_irqrestore(&adapter->cmd_lock, flags);
|
|
|
|
return 0;
|
|
|
|
}
|
|
#endif
|
|
|
|
static const struct ethtool_ops vmxnet3_ethtool_ops = {
|
|
.get_settings = vmxnet3_get_settings,
|
|
.get_drvinfo = vmxnet3_get_drvinfo,
|
|
.get_regs_len = vmxnet3_get_regs_len,
|
|
.get_regs = vmxnet3_get_regs,
|
|
.get_wol = vmxnet3_get_wol,
|
|
.set_wol = vmxnet3_set_wol,
|
|
.get_link = ethtool_op_get_link,
|
|
.get_strings = vmxnet3_get_strings,
|
|
.get_sset_count = vmxnet3_get_sset_count,
|
|
.get_ethtool_stats = vmxnet3_get_ethtool_stats,
|
|
.get_ringparam = vmxnet3_get_ringparam,
|
|
.set_ringparam = vmxnet3_set_ringparam,
|
|
.get_rxnfc = vmxnet3_get_rxnfc,
|
|
#ifdef VMXNET3_RSS
|
|
.get_rxfh_indir_size = vmxnet3_get_rss_indir_size,
|
|
.get_rxfh_indir = vmxnet3_get_rss_indir,
|
|
.set_rxfh_indir = vmxnet3_set_rss_indir,
|
|
#endif
|
|
};
|
|
|
|
void vmxnet3_set_ethtool_ops(struct net_device *netdev)
|
|
{
|
|
SET_ETHTOOL_OPS(netdev, &vmxnet3_ethtool_ops);
|
|
}
|