Merge branch 'cpsw-common-res-usage'

Ivan Khoronzhuk says:

====================
net: ethernet: ti: cpsw: correct common res usage

This series is intended to remove unneeded redundancies connected with
common resource usage function.

Since v1:
- changed name to cpsw_get_usage_count()
- added comments to open/closw for cpsw_get_usage_count()
- added patch:
  net: ethernet: ti: cpsw: clarify ethtool ops changing num of descs

Based on net-next/master
====================

Reviewed-by: Grygorii Strashko <grygorii.strashko@ti.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
David S. Miller 2017-01-20 12:35:11 -05:00
commit cc154783e7
1 changed files with 88 additions and 112 deletions

View File

@ -357,7 +357,6 @@ struct cpsw_slave {
struct phy_device *phy; struct phy_device *phy;
struct net_device *ndev; struct net_device *ndev;
u32 port_vlan; u32 port_vlan;
u32 open_stat;
}; };
static inline u32 slave_read(struct cpsw_slave *slave, u32 offset) static inline u32 slave_read(struct cpsw_slave *slave, u32 offset)
@ -672,6 +671,18 @@ static void cpsw_intr_disable(struct cpsw_common *cpsw)
return; return;
} }
static int cpsw_get_usage_count(struct cpsw_common *cpsw)
{
u32 i;
u32 usage_count = 0;
for (i = 0; i < cpsw->data.slaves; i++)
if (netif_running(cpsw->slaves[i].ndev))
usage_count++;
return usage_count;
}
static void cpsw_tx_handler(void *token, int len, int status) static void cpsw_tx_handler(void *token, int len, int status)
{ {
struct netdev_queue *txq; struct netdev_queue *txq;
@ -704,18 +715,10 @@ static void cpsw_rx_handler(void *token, int len, int status)
cpsw_dual_emac_src_port_detect(cpsw, status, ndev, skb); cpsw_dual_emac_src_port_detect(cpsw, status, ndev, skb);
if (unlikely(status < 0) || unlikely(!netif_running(ndev))) { if (unlikely(status < 0) || unlikely(!netif_running(ndev))) {
bool ndev_status = false; /* In dual emac mode check for all interfaces */
struct cpsw_slave *slave = cpsw->slaves; if (cpsw->data.dual_emac &&
int n; cpsw_get_usage_count(cpsw) &&
(status >= 0)) {
if (cpsw->data.dual_emac) {
/* In dual emac mode check for all interfaces */
for (n = cpsw->data.slaves; n; n--, slave++)
if (netif_running(slave->ndev))
ndev_status = true;
}
if (ndev_status && (status >= 0)) {
/* The packet received is for the interface which /* The packet received is for the interface which
* is already down and the other interface is up * is already down and the other interface is up
* and running, instead of freeing which results * and running, instead of freeing which results
@ -1235,21 +1238,6 @@ static void cpsw_get_ethtool_stats(struct net_device *ndev,
} }
} }
static int cpsw_common_res_usage_state(struct cpsw_common *cpsw)
{
u32 i;
u32 usage_count = 0;
if (!cpsw->data.dual_emac)
return 0;
for (i = 0; i < cpsw->data.slaves; i++)
if (cpsw->slaves[i].open_stat)
usage_count++;
return usage_count;
}
static inline int cpsw_tx_packet_submit(struct cpsw_priv *priv, static inline int cpsw_tx_packet_submit(struct cpsw_priv *priv,
struct sk_buff *skb, struct sk_buff *skb,
struct cpdma_chan *txch) struct cpdma_chan *txch)
@ -1483,8 +1471,6 @@ static int cpsw_ndo_open(struct net_device *ndev)
return ret; return ret;
} }
if (!cpsw_common_res_usage_state(cpsw))
cpsw_intr_disable(cpsw);
netif_carrier_off(ndev); netif_carrier_off(ndev);
/* Notify the stack of the actual queue counts. */ /* Notify the stack of the actual queue counts. */
@ -1506,8 +1492,11 @@ static int cpsw_ndo_open(struct net_device *ndev)
CPSW_MAJOR_VERSION(reg), CPSW_MINOR_VERSION(reg), CPSW_MAJOR_VERSION(reg), CPSW_MINOR_VERSION(reg),
CPSW_RTL_VERSION(reg)); CPSW_RTL_VERSION(reg));
/* initialize host and slave ports */ /* Initialize host and slave ports.
if (!cpsw_common_res_usage_state(cpsw)) * Given ndev is marked as opened already, so init port only if 1 ndev
* is opened
*/
if (cpsw_get_usage_count(cpsw) < 2)
cpsw_init_host_port(priv); cpsw_init_host_port(priv);
for_each_slave(priv, cpsw_slave_open, priv); for_each_slave(priv, cpsw_slave_open, priv);
@ -1518,7 +1507,10 @@ static int cpsw_ndo_open(struct net_device *ndev)
cpsw_ale_add_vlan(cpsw->ale, cpsw->data.default_vlan, cpsw_ale_add_vlan(cpsw->ale, cpsw->data.default_vlan,
ALE_ALL_PORTS, ALE_ALL_PORTS, 0, 0); ALE_ALL_PORTS, ALE_ALL_PORTS, 0, 0);
if (!cpsw_common_res_usage_state(cpsw)) { /* Given ndev is marked as opened already, so if more ndev
* are opened - no need to init shared resources.
*/
if (cpsw_get_usage_count(cpsw) < 2) {
/* disable priority elevation */ /* disable priority elevation */
__raw_writel(0, &cpsw->regs->ptype); __raw_writel(0, &cpsw->regs->ptype);
@ -1561,9 +1553,6 @@ static int cpsw_ndo_open(struct net_device *ndev)
cpdma_ctlr_start(cpsw->dma); cpdma_ctlr_start(cpsw->dma);
cpsw_intr_enable(cpsw); cpsw_intr_enable(cpsw);
if (cpsw->data.dual_emac)
cpsw->slaves[priv->emac_port].open_stat = true;
return 0; return 0;
err_cleanup: err_cleanup:
@ -1583,7 +1572,10 @@ static int cpsw_ndo_stop(struct net_device *ndev)
netif_tx_stop_all_queues(priv->ndev); netif_tx_stop_all_queues(priv->ndev);
netif_carrier_off(priv->ndev); netif_carrier_off(priv->ndev);
if (cpsw_common_res_usage_state(cpsw) <= 1) { /* Given ndev is marked as close already,
* so disable shared resources if no open devices
*/
if (!cpsw_get_usage_count(cpsw)) {
napi_disable(&cpsw->napi_rx); napi_disable(&cpsw->napi_rx);
napi_disable(&cpsw->napi_tx); napi_disable(&cpsw->napi_tx);
cpts_unregister(cpsw->cpts); cpts_unregister(cpsw->cpts);
@ -1597,8 +1589,6 @@ static int cpsw_ndo_stop(struct net_device *ndev)
cpsw_split_res(ndev); cpsw_split_res(ndev);
pm_runtime_put_sync(cpsw->dev); pm_runtime_put_sync(cpsw->dev);
if (cpsw->data.dual_emac)
cpsw->slaves[priv->emac_port].open_stat = false;
return 0; return 0;
} }
@ -2368,17 +2358,11 @@ static int cpsw_update_channels(struct cpsw_priv *priv,
return 0; return 0;
} }
static int cpsw_set_channels(struct net_device *ndev, static void cpsw_suspend_data_pass(struct net_device *ndev)
struct ethtool_channels *chs)
{ {
struct cpsw_priv *priv = netdev_priv(ndev); struct cpsw_common *cpsw = ndev_to_cpsw(ndev);
struct cpsw_common *cpsw = priv->cpsw;
struct cpsw_slave *slave; struct cpsw_slave *slave;
int i, ret; int i;
ret = cpsw_check_ch_settings(cpsw, chs);
if (ret < 0)
return ret;
/* Disable NAPI scheduling */ /* Disable NAPI scheduling */
cpsw_intr_disable(cpsw); cpsw_intr_disable(cpsw);
@ -2396,6 +2380,51 @@ static int cpsw_set_channels(struct net_device *ndev,
/* Handle rest of tx packets and stop cpdma channels */ /* Handle rest of tx packets and stop cpdma channels */
cpdma_ctlr_stop(cpsw->dma); cpdma_ctlr_stop(cpsw->dma);
}
static int cpsw_resume_data_pass(struct net_device *ndev)
{
struct cpsw_priv *priv = netdev_priv(ndev);
struct cpsw_common *cpsw = priv->cpsw;
struct cpsw_slave *slave;
int i, ret;
/* Allow rx packets handling */
for (i = cpsw->data.slaves, slave = cpsw->slaves; i; i--, slave++)
if (slave->ndev && netif_running(slave->ndev))
netif_dormant_off(slave->ndev);
/* After this receive is started */
if (cpsw_get_usage_count(cpsw)) {
ret = cpsw_fill_rx_channels(priv);
if (ret)
return ret;
cpdma_ctlr_start(cpsw->dma);
cpsw_intr_enable(cpsw);
}
/* Resume transmit for every affected interface */
for (i = cpsw->data.slaves, slave = cpsw->slaves; i; i--, slave++)
if (slave->ndev && netif_running(slave->ndev))
netif_tx_start_all_queues(slave->ndev);
return 0;
}
static int cpsw_set_channels(struct net_device *ndev,
struct ethtool_channels *chs)
{
struct cpsw_priv *priv = netdev_priv(ndev);
struct cpsw_common *cpsw = priv->cpsw;
struct cpsw_slave *slave;
int i, ret;
ret = cpsw_check_ch_settings(cpsw, chs);
if (ret < 0)
return ret;
cpsw_suspend_data_pass(ndev);
ret = cpsw_update_channels(priv, chs); ret = cpsw_update_channels(priv, chs);
if (ret) if (ret)
goto err; goto err;
@ -2418,30 +2447,14 @@ static int cpsw_set_channels(struct net_device *ndev,
dev_err(priv->dev, "cannot set real number of rx queues\n"); dev_err(priv->dev, "cannot set real number of rx queues\n");
goto err; goto err;
} }
/* Enable rx packets handling */
netif_dormant_off(slave->ndev);
} }
if (cpsw_common_res_usage_state(cpsw)) { if (cpsw_get_usage_count(cpsw))
ret = cpsw_fill_rx_channels(priv);
if (ret)
goto err;
cpsw_split_res(ndev); cpsw_split_res(ndev);
/* After this receive is started */ ret = cpsw_resume_data_pass(ndev);
cpdma_ctlr_start(cpsw->dma); if (!ret)
cpsw_intr_enable(cpsw); return 0;
}
/* Resume transmit for every affected interface */
for (i = cpsw->data.slaves, slave = cpsw->slaves; i; i--, slave++) {
if (!(slave->ndev && netif_running(slave->ndev)))
continue;
netif_tx_start_all_queues(slave->ndev);
}
return 0;
err: err:
dev_err(priv->dev, "cannot update channels number, closing device\n"); dev_err(priv->dev, "cannot update channels number, closing device\n");
dev_close(ndev); dev_close(ndev);
@ -2502,8 +2515,7 @@ static int cpsw_set_ringparam(struct net_device *ndev,
{ {
struct cpsw_priv *priv = netdev_priv(ndev); struct cpsw_priv *priv = netdev_priv(ndev);
struct cpsw_common *cpsw = priv->cpsw; struct cpsw_common *cpsw = priv->cpsw;
struct cpsw_slave *slave; int ret;
int i, ret;
/* ignore ering->tx_pending - only rx_pending adjustment is supported */ /* ignore ering->tx_pending - only rx_pending adjustment is supported */
@ -2515,54 +2527,18 @@ static int cpsw_set_ringparam(struct net_device *ndev,
if (ering->rx_pending == cpdma_get_num_rx_descs(cpsw->dma)) if (ering->rx_pending == cpdma_get_num_rx_descs(cpsw->dma))
return 0; return 0;
/* Disable NAPI scheduling */ cpsw_suspend_data_pass(ndev);
cpsw_intr_disable(cpsw);
/* Stop all transmit queues for every network device.
* Disable re-using rx descriptors with dormant_on.
*/
for (i = cpsw->data.slaves, slave = cpsw->slaves; i; i--, slave++) {
if (!(slave->ndev && netif_running(slave->ndev)))
continue;
netif_tx_stop_all_queues(slave->ndev);
netif_dormant_on(slave->ndev);
}
/* Handle rest of tx packets and stop cpdma channels */
cpdma_ctlr_stop(cpsw->dma);
cpdma_set_num_rx_descs(cpsw->dma, ering->rx_pending); cpdma_set_num_rx_descs(cpsw->dma, ering->rx_pending);
for (i = cpsw->data.slaves, slave = cpsw->slaves; i; i--, slave++) { if (cpsw_get_usage_count(cpsw))
if (!(slave->ndev && netif_running(slave->ndev)))
continue;
/* Enable rx packets handling */
netif_dormant_off(slave->ndev);
}
if (cpsw_common_res_usage_state(cpsw)) {
cpdma_chan_split_pool(cpsw->dma); cpdma_chan_split_pool(cpsw->dma);
ret = cpsw_fill_rx_channels(priv); ret = cpsw_resume_data_pass(ndev);
if (ret) if (!ret)
goto err; return 0;
/* After this receive is started */ dev_err(&ndev->dev, "cannot set ring params, closing device\n");
cpdma_ctlr_start(cpsw->dma);
cpsw_intr_enable(cpsw);
}
/* Resume transmit for every affected interface */
for (i = cpsw->data.slaves, slave = cpsw->slaves; i; i--, slave++) {
if (!(slave->ndev && netif_running(slave->ndev)))
continue;
netif_tx_start_all_queues(slave->ndev);
}
return 0;
err:
dev_err(priv->dev, "cannot set ring params, closing device\n");
dev_close(ndev); dev_close(ndev);
return ret; return ret;
} }