Merge branch 'bnxt_en-Bug-fixes'
Michael Chan says: ==================== bnxt_en: Bug fixes. 3 bug fixes in this series. Fix a crash in bnxt_get_stats64() that can happen if the device is closing and freeing the statistics block at the same time. The 2nd one fixes ethtool -L failing when changing from combined to non-combined mode or vice versa. The last one fixes SRIOV failure on big-endian systems because we were setting a bitmap wrong in a firmware message. ==================== Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
commit
4c17e1febd
|
@ -3458,13 +3458,18 @@ static int bnxt_hwrm_func_drv_rgtr(struct bnxt *bp)
|
|||
req.ver_upd = DRV_VER_UPD;
|
||||
|
||||
if (BNXT_PF(bp)) {
|
||||
DECLARE_BITMAP(vf_req_snif_bmap, 256);
|
||||
u32 *data = (u32 *)vf_req_snif_bmap;
|
||||
u32 data[8];
|
||||
int i;
|
||||
|
||||
memset(vf_req_snif_bmap, 0, sizeof(vf_req_snif_bmap));
|
||||
for (i = 0; i < ARRAY_SIZE(bnxt_vf_req_snif); i++)
|
||||
__set_bit(bnxt_vf_req_snif[i], vf_req_snif_bmap);
|
||||
memset(data, 0, sizeof(data));
|
||||
for (i = 0; i < ARRAY_SIZE(bnxt_vf_req_snif); i++) {
|
||||
u16 cmd = bnxt_vf_req_snif[i];
|
||||
unsigned int bit, idx;
|
||||
|
||||
idx = cmd / 32;
|
||||
bit = cmd % 32;
|
||||
data[idx] |= 1 << bit;
|
||||
}
|
||||
|
||||
for (i = 0; i < 8; i++)
|
||||
req.vf_req_fwd[i] = cpu_to_le32(data[i]);
|
||||
|
@ -6279,6 +6284,12 @@ static int bnxt_open(struct net_device *dev)
|
|||
return __bnxt_open_nic(bp, true, true);
|
||||
}
|
||||
|
||||
static bool bnxt_drv_busy(struct bnxt *bp)
|
||||
{
|
||||
return (test_bit(BNXT_STATE_IN_SP_TASK, &bp->state) ||
|
||||
test_bit(BNXT_STATE_READ_STATS, &bp->state));
|
||||
}
|
||||
|
||||
int bnxt_close_nic(struct bnxt *bp, bool irq_re_init, bool link_re_init)
|
||||
{
|
||||
int rc = 0;
|
||||
|
@ -6297,7 +6308,7 @@ int bnxt_close_nic(struct bnxt *bp, bool irq_re_init, bool link_re_init)
|
|||
|
||||
clear_bit(BNXT_STATE_OPEN, &bp->state);
|
||||
smp_mb__after_atomic();
|
||||
while (test_bit(BNXT_STATE_IN_SP_TASK, &bp->state))
|
||||
while (bnxt_drv_busy(bp))
|
||||
msleep(20);
|
||||
|
||||
/* Flush rings and and disable interrupts */
|
||||
|
@ -6358,8 +6369,15 @@ bnxt_get_stats64(struct net_device *dev, struct rtnl_link_stats64 *stats)
|
|||
u32 i;
|
||||
struct bnxt *bp = netdev_priv(dev);
|
||||
|
||||
if (!bp->bnapi)
|
||||
set_bit(BNXT_STATE_READ_STATS, &bp->state);
|
||||
/* Make sure bnxt_close_nic() sees that we are reading stats before
|
||||
* we check the BNXT_STATE_OPEN flag.
|
||||
*/
|
||||
smp_mb__after_atomic();
|
||||
if (!test_bit(BNXT_STATE_OPEN, &bp->state)) {
|
||||
clear_bit(BNXT_STATE_READ_STATS, &bp->state);
|
||||
return;
|
||||
}
|
||||
|
||||
/* TODO check if we need to synchronize with bnxt_close path */
|
||||
for (i = 0; i < bp->cp_nr_rings; i++) {
|
||||
|
@ -6406,6 +6424,7 @@ bnxt_get_stats64(struct net_device *dev, struct rtnl_link_stats64 *stats)
|
|||
stats->tx_fifo_errors = le64_to_cpu(tx->tx_fifo_underruns);
|
||||
stats->tx_errors = le64_to_cpu(tx->tx_err);
|
||||
}
|
||||
clear_bit(BNXT_STATE_READ_STATS, &bp->state);
|
||||
}
|
||||
|
||||
static bool bnxt_mc_list_updated(struct bnxt *bp, u32 *rx_mask)
|
||||
|
@ -6904,16 +6923,13 @@ static void bnxt_sp_task(struct work_struct *work)
|
|||
}
|
||||
|
||||
/* Under rtnl_lock */
|
||||
int bnxt_reserve_rings(struct bnxt *bp, int tx, int rx, int tcs, int tx_xdp)
|
||||
int bnxt_reserve_rings(struct bnxt *bp, int tx, int rx, bool sh, int tcs,
|
||||
int tx_xdp)
|
||||
{
|
||||
int max_rx, max_tx, tx_sets = 1;
|
||||
int tx_rings_needed;
|
||||
bool sh = true;
|
||||
int rc;
|
||||
|
||||
if (!(bp->flags & BNXT_FLAG_SHARED_RINGS))
|
||||
sh = false;
|
||||
|
||||
if (tcs)
|
||||
tx_sets = tcs;
|
||||
|
||||
|
@ -7121,7 +7137,7 @@ int bnxt_setup_mq_tc(struct net_device *dev, u8 tc)
|
|||
sh = true;
|
||||
|
||||
rc = bnxt_reserve_rings(bp, bp->tx_nr_rings_per_tc, bp->rx_nr_rings,
|
||||
tc, bp->tx_nr_rings_xdp);
|
||||
sh, tc, bp->tx_nr_rings_xdp);
|
||||
if (rc)
|
||||
return rc;
|
||||
|
||||
|
|
|
@ -1117,6 +1117,7 @@ struct bnxt {
|
|||
unsigned long state;
|
||||
#define BNXT_STATE_OPEN 0
|
||||
#define BNXT_STATE_IN_SP_TASK 1
|
||||
#define BNXT_STATE_READ_STATS 2
|
||||
|
||||
struct bnxt_irq *irq_tbl;
|
||||
int total_irqs;
|
||||
|
@ -1300,7 +1301,8 @@ int bnxt_open_nic(struct bnxt *, bool, bool);
|
|||
int bnxt_half_open_nic(struct bnxt *bp);
|
||||
void bnxt_half_close_nic(struct bnxt *bp);
|
||||
int bnxt_close_nic(struct bnxt *, bool, bool);
|
||||
int bnxt_reserve_rings(struct bnxt *bp, int tx, int rx, int tcs, int tx_xdp);
|
||||
int bnxt_reserve_rings(struct bnxt *bp, int tx, int rx, bool sh, int tcs,
|
||||
int tx_xdp);
|
||||
int bnxt_setup_mq_tc(struct net_device *dev, u8 tc);
|
||||
int bnxt_get_max_rings(struct bnxt *, int *, int *, bool);
|
||||
void bnxt_restore_pf_fw_resources(struct bnxt *bp);
|
||||
|
|
|
@ -432,7 +432,8 @@ static int bnxt_set_channels(struct net_device *dev,
|
|||
}
|
||||
tx_xdp = req_rx_rings;
|
||||
}
|
||||
rc = bnxt_reserve_rings(bp, req_tx_rings, req_rx_rings, tcs, tx_xdp);
|
||||
rc = bnxt_reserve_rings(bp, req_tx_rings, req_rx_rings, sh, tcs,
|
||||
tx_xdp);
|
||||
if (rc) {
|
||||
netdev_warn(dev, "Unable to allocate the requested rings\n");
|
||||
return rc;
|
||||
|
|
|
@ -170,7 +170,7 @@ static int bnxt_xdp_set(struct bnxt *bp, struct bpf_prog *prog)
|
|||
if (!tc)
|
||||
tc = 1;
|
||||
rc = bnxt_reserve_rings(bp, bp->tx_nr_rings_per_tc, bp->rx_nr_rings,
|
||||
tc, tx_xdp);
|
||||
true, tc, tx_xdp);
|
||||
if (rc) {
|
||||
netdev_warn(dev, "Unable to reserve enough TX rings to support XDP.\n");
|
||||
return rc;
|
||||
|
|
Loading…
Reference in New Issue