mirror of https://gitee.com/openkylin/linux.git
net: mvneta: adjust multiline comments to net/ style
As reported by checkpatch, the multiline comments for net/ and drivers/net/ have a slightly different format than the one used in the rest of the kernel, so we adjust our multiline comments accordingly. Signed-off-by: Thomas Petazzoni <thomas.petazzoni@free-electrons.com>
This commit is contained in:
parent
b07812f15e
commit
6a20c1758d
|
@ -178,8 +178,7 @@
|
||||||
/* Napi polling weight */
|
/* Napi polling weight */
|
||||||
#define MVNETA_RX_POLL_WEIGHT 64
|
#define MVNETA_RX_POLL_WEIGHT 64
|
||||||
|
|
||||||
/*
|
/* The two bytes Marvell header. Either contains a special value used
|
||||||
* The two bytes Marvell header. Either contains a special value used
|
|
||||||
* by Marvell switches when a specific hardware mode is enabled (not
|
* by Marvell switches when a specific hardware mode is enabled (not
|
||||||
* supported by this driver) or is filled automatically by zeroes on
|
* supported by this driver) or is filled automatically by zeroes on
|
||||||
* the RX side. Those two bytes being at the front of the Ethernet
|
* the RX side. Those two bytes being at the front of the Ethernet
|
||||||
|
@ -259,8 +258,7 @@ struct mvneta_port {
|
||||||
unsigned int speed;
|
unsigned int speed;
|
||||||
};
|
};
|
||||||
|
|
||||||
/*
|
/* The mvneta_tx_desc and mvneta_rx_desc structures describe the
|
||||||
* The mvneta_tx_desc and mvneta_rx_desc structures describe the
|
|
||||||
* layout of the transmit and reception DMA descriptors, and their
|
* layout of the transmit and reception DMA descriptors, and their
|
||||||
* layout is therefore defined by the hardware design
|
* layout is therefore defined by the hardware design
|
||||||
*/
|
*/
|
||||||
|
@ -318,7 +316,8 @@ struct mvneta_tx_queue {
|
||||||
int size;
|
int size;
|
||||||
|
|
||||||
/* Number of currently used TX DMA descriptor in the
|
/* Number of currently used TX DMA descriptor in the
|
||||||
* descriptor ring */
|
* descriptor ring
|
||||||
|
*/
|
||||||
int count;
|
int count;
|
||||||
|
|
||||||
/* Array of transmitted skb */
|
/* Array of transmitted skb */
|
||||||
|
@ -454,8 +453,7 @@ struct rtnl_link_stats64 *mvneta_get_stats64(struct net_device *dev,
|
||||||
|
|
||||||
/* Rx descriptors helper methods */
|
/* Rx descriptors helper methods */
|
||||||
|
|
||||||
/*
|
/* Checks whether the given RX descriptor is both the first and the
|
||||||
* Checks whether the given RX descriptor is both the first and the
|
|
||||||
* last descriptor for the RX packet. Each RX packet is currently
|
* last descriptor for the RX packet. Each RX packet is currently
|
||||||
* received through a single RX descriptor, so not having each RX
|
* received through a single RX descriptor, so not having each RX
|
||||||
* descriptor with its first and last bits set is an error
|
* descriptor with its first and last bits set is an error
|
||||||
|
@ -472,7 +470,8 @@ static void mvneta_rxq_non_occup_desc_add(struct mvneta_port *pp,
|
||||||
int ndescs)
|
int ndescs)
|
||||||
{
|
{
|
||||||
/* Only MVNETA_RXQ_ADD_NON_OCCUPIED_MAX (255) descriptors can
|
/* Only MVNETA_RXQ_ADD_NON_OCCUPIED_MAX (255) descriptors can
|
||||||
* be added at once */
|
* be added at once
|
||||||
|
*/
|
||||||
while (ndescs > MVNETA_RXQ_ADD_NON_OCCUPIED_MAX) {
|
while (ndescs > MVNETA_RXQ_ADD_NON_OCCUPIED_MAX) {
|
||||||
mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id),
|
mvreg_write(pp, MVNETA_RXQ_STATUS_UPDATE_REG(rxq->id),
|
||||||
(MVNETA_RXQ_ADD_NON_OCCUPIED_MAX <<
|
(MVNETA_RXQ_ADD_NON_OCCUPIED_MAX <<
|
||||||
|
@ -494,8 +493,7 @@ static int mvneta_rxq_busy_desc_num_get(struct mvneta_port *pp,
|
||||||
return val & MVNETA_RXQ_OCCUPIED_ALL_MASK;
|
return val & MVNETA_RXQ_OCCUPIED_ALL_MASK;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/* Update num of rx desc called upon return from rx path or
|
||||||
* Update num of rx desc called upon return from rx path or
|
|
||||||
* from mvneta_rxq_drop_pkts().
|
* from mvneta_rxq_drop_pkts().
|
||||||
*/
|
*/
|
||||||
static void mvneta_rxq_desc_num_update(struct mvneta_port *pp,
|
static void mvneta_rxq_desc_num_update(struct mvneta_port *pp,
|
||||||
|
@ -580,7 +578,8 @@ static void mvneta_txq_pend_desc_add(struct mvneta_port *pp,
|
||||||
u32 val;
|
u32 val;
|
||||||
|
|
||||||
/* Only 255 descriptors can be added at once ; Assume caller
|
/* Only 255 descriptors can be added at once ; Assume caller
|
||||||
process TX desriptors in quanta less than 256 */
|
* process TX desriptors in quanta less than 256
|
||||||
|
*/
|
||||||
val = pend_desc;
|
val = pend_desc;
|
||||||
mvreg_write(pp, MVNETA_TXQ_UPDATE_REG(txq->id), val);
|
mvreg_write(pp, MVNETA_TXQ_UPDATE_REG(txq->id), val);
|
||||||
}
|
}
|
||||||
|
@ -596,7 +595,8 @@ mvneta_txq_next_desc_get(struct mvneta_tx_queue *txq)
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Release the last allocated TX descriptor. Useful to handle DMA
|
/* Release the last allocated TX descriptor. Useful to handle DMA
|
||||||
* mapping failures in the TX path. */
|
* mapping failures in the TX path.
|
||||||
|
*/
|
||||||
static void mvneta_txq_desc_put(struct mvneta_tx_queue *txq)
|
static void mvneta_txq_desc_put(struct mvneta_tx_queue *txq)
|
||||||
{
|
{
|
||||||
if (txq->next_desc_to_proc == 0)
|
if (txq->next_desc_to_proc == 0)
|
||||||
|
@ -714,7 +714,8 @@ static void mvneta_port_down(struct mvneta_port *pp)
|
||||||
} while (val & 0xff);
|
} while (val & 0xff);
|
||||||
|
|
||||||
/* Stop Tx port activity. Check port Tx activity. Issue stop
|
/* Stop Tx port activity. Check port Tx activity. Issue stop
|
||||||
command for active channels only */
|
* command for active channels only
|
||||||
|
*/
|
||||||
val = (mvreg_read(pp, MVNETA_TXQ_CMD)) & MVNETA_TXQ_ENABLE_MASK;
|
val = (mvreg_read(pp, MVNETA_TXQ_CMD)) & MVNETA_TXQ_ENABLE_MASK;
|
||||||
|
|
||||||
if (val != 0)
|
if (val != 0)
|
||||||
|
@ -865,7 +866,8 @@ static void mvneta_defaults_set(struct mvneta_port *pp)
|
||||||
mvreg_write(pp, MVNETA_MBUS_RETRY, 0x20);
|
mvreg_write(pp, MVNETA_MBUS_RETRY, 0x20);
|
||||||
|
|
||||||
/* Set CPU queue access map - all CPUs have access to all RX
|
/* Set CPU queue access map - all CPUs have access to all RX
|
||||||
queues and to all TX queues */
|
* queues and to all TX queues
|
||||||
|
*/
|
||||||
for (cpu = 0; cpu < CONFIG_NR_CPUS; cpu++)
|
for (cpu = 0; cpu < CONFIG_NR_CPUS; cpu++)
|
||||||
mvreg_write(pp, MVNETA_CPU_MAP(cpu),
|
mvreg_write(pp, MVNETA_CPU_MAP(cpu),
|
||||||
(MVNETA_CPU_RXQ_ACCESS_ALL_MASK |
|
(MVNETA_CPU_RXQ_ACCESS_ALL_MASK |
|
||||||
|
@ -1010,9 +1012,8 @@ static void mvneta_mac_addr_set(struct mvneta_port *pp, unsigned char *addr,
|
||||||
mvneta_set_ucast_addr(pp, addr[5], queue);
|
mvneta_set_ucast_addr(pp, addr[5], queue);
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/* Set the number of packets that will be received before RX interrupt
|
||||||
* Set the number of packets that will be received before
|
* will be generated by HW.
|
||||||
* RX interrupt will be generated by HW.
|
|
||||||
*/
|
*/
|
||||||
static void mvneta_rx_pkts_coal_set(struct mvneta_port *pp,
|
static void mvneta_rx_pkts_coal_set(struct mvneta_port *pp,
|
||||||
struct mvneta_rx_queue *rxq, u32 value)
|
struct mvneta_rx_queue *rxq, u32 value)
|
||||||
|
@ -1022,9 +1023,8 @@ static void mvneta_rx_pkts_coal_set(struct mvneta_port *pp,
|
||||||
rxq->pkts_coal = value;
|
rxq->pkts_coal = value;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/* Set the time delay in usec before RX interrupt will be generated by
|
||||||
* Set the time delay in usec before
|
* HW.
|
||||||
* RX interrupt will be generated by HW.
|
|
||||||
*/
|
*/
|
||||||
static void mvneta_rx_time_coal_set(struct mvneta_port *pp,
|
static void mvneta_rx_time_coal_set(struct mvneta_port *pp,
|
||||||
struct mvneta_rx_queue *rxq, u32 value)
|
struct mvneta_rx_queue *rxq, u32 value)
|
||||||
|
@ -1102,8 +1102,7 @@ static int mvneta_txq_sent_desc_num_get(struct mvneta_port *pp,
|
||||||
return sent_desc;
|
return sent_desc;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/* Get number of sent descriptors and decrement counter.
|
||||||
* Get number of sent descriptors and decrement counter.
|
|
||||||
* The number of sent descriptors is returned.
|
* The number of sent descriptors is returned.
|
||||||
*/
|
*/
|
||||||
static int mvneta_txq_sent_desc_proc(struct mvneta_port *pp,
|
static int mvneta_txq_sent_desc_proc(struct mvneta_port *pp,
|
||||||
|
@ -1128,8 +1127,9 @@ static u32 mvneta_txq_desc_csum(int l3_offs, int l3_proto,
|
||||||
u32 command;
|
u32 command;
|
||||||
|
|
||||||
/* Fields: L3_offset, IP_hdrlen, L3_type, G_IPv4_chk,
|
/* Fields: L3_offset, IP_hdrlen, L3_type, G_IPv4_chk,
|
||||||
G_L4_chk, L4_type; required only for checksum
|
* G_L4_chk, L4_type; required only for checksum
|
||||||
calculation */
|
* calculation
|
||||||
|
*/
|
||||||
command = l3_offs << MVNETA_TX_L3_OFF_SHIFT;
|
command = l3_offs << MVNETA_TX_L3_OFF_SHIFT;
|
||||||
command |= ip_hdr_len << MVNETA_TX_IP_HLEN_SHIFT;
|
command |= ip_hdr_len << MVNETA_TX_IP_HLEN_SHIFT;
|
||||||
|
|
||||||
|
@ -1305,8 +1305,7 @@ static u32 mvneta_skb_tx_csum(struct mvneta_port *pp, struct sk_buff *skb)
|
||||||
return MVNETA_TX_L4_CSUM_NOT;
|
return MVNETA_TX_L4_CSUM_NOT;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/* Returns rx queue pointer (find last set bit) according to causeRxTx
|
||||||
* Returns rx queue pointer (find last set bit) according to causeRxTx
|
|
||||||
* value
|
* value
|
||||||
*/
|
*/
|
||||||
static struct mvneta_rx_queue *mvneta_rx_policy(struct mvneta_port *pp,
|
static struct mvneta_rx_queue *mvneta_rx_policy(struct mvneta_port *pp,
|
||||||
|
@ -1454,7 +1453,8 @@ static int mvneta_tx_frag_process(struct mvneta_port *pp, struct sk_buff *skb,
|
||||||
|
|
||||||
error:
|
error:
|
||||||
/* Release all descriptors that were used to map fragments of
|
/* Release all descriptors that were used to map fragments of
|
||||||
* this packet, as well as the corresponding DMA mappings */
|
* this packet, as well as the corresponding DMA mappings
|
||||||
|
*/
|
||||||
for (i = i - 1; i >= 0; i--) {
|
for (i = i - 1; i >= 0; i--) {
|
||||||
tx_desc = txq->descs + i;
|
tx_desc = txq->descs + i;
|
||||||
dma_unmap_single(pp->dev->dev.parent,
|
dma_unmap_single(pp->dev->dev.parent,
|
||||||
|
@ -1546,7 +1546,8 @@ static int mvneta_tx(struct sk_buff *skb, struct net_device *dev)
|
||||||
mvneta_txq_done(pp, txq);
|
mvneta_txq_done(pp, txq);
|
||||||
|
|
||||||
/* If after calling mvneta_txq_done, count equals
|
/* If after calling mvneta_txq_done, count equals
|
||||||
frags, we need to set the timer */
|
* frags, we need to set the timer
|
||||||
|
*/
|
||||||
if (txq->count == frags && frags > 0)
|
if (txq->count == frags && frags > 0)
|
||||||
mvneta_add_tx_done_timer(pp);
|
mvneta_add_tx_done_timer(pp);
|
||||||
|
|
||||||
|
@ -1598,8 +1599,7 @@ static u32 mvneta_tx_done_gbe(struct mvneta_port *pp, u32 cause_tx_done,
|
||||||
return tx_done;
|
return tx_done;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/* Compute crc8 of the specified address, using a unique algorithm ,
|
||||||
* Compute crc8 of the specified address, using a unique algorithm ,
|
|
||||||
* according to hw spec, different than generic crc8 algorithm
|
* according to hw spec, different than generic crc8 algorithm
|
||||||
*/
|
*/
|
||||||
static int mvneta_addr_crc(unsigned char *addr)
|
static int mvneta_addr_crc(unsigned char *addr)
|
||||||
|
@ -1828,8 +1828,7 @@ static int mvneta_poll(struct napi_struct *napi, int budget)
|
||||||
cause_rx_tx = mvreg_read(pp, MVNETA_INTR_NEW_CAUSE) &
|
cause_rx_tx = mvreg_read(pp, MVNETA_INTR_NEW_CAUSE) &
|
||||||
MVNETA_RX_INTR_MASK(rxq_number);
|
MVNETA_RX_INTR_MASK(rxq_number);
|
||||||
|
|
||||||
/*
|
/* For the case where the last mvneta_poll did not process all
|
||||||
* For the case where the last mvneta_poll did not process all
|
|
||||||
* RX packets
|
* RX packets
|
||||||
*/
|
*/
|
||||||
cause_rx_tx |= pp->cause_rx_tx;
|
cause_rx_tx |= pp->cause_rx_tx;
|
||||||
|
@ -1847,10 +1846,12 @@ static int mvneta_poll(struct napi_struct *napi, int budget)
|
||||||
rx_done += count;
|
rx_done += count;
|
||||||
budget -= count;
|
budget -= count;
|
||||||
if (budget > 0) {
|
if (budget > 0) {
|
||||||
/* set off the rx bit of the corresponding bit
|
/* set off the rx bit of the
|
||||||
in the cause rx tx register, so that next
|
* corresponding bit in the cause rx
|
||||||
iteration will find the next rx queue where
|
* tx register, so that next iteration
|
||||||
packets are received on */
|
* will find the next rx queue where
|
||||||
|
* packets are received on
|
||||||
|
*/
|
||||||
cause_rx_tx &= ~((1 << rxq->id) << 8);
|
cause_rx_tx &= ~((1 << rxq->id) << 8);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -1925,7 +1926,8 @@ static int mvneta_rxq_fill(struct mvneta_port *pp, struct mvneta_rx_queue *rxq,
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Add this number of RX descriptors as non occupied (ready to
|
/* Add this number of RX descriptors as non occupied (ready to
|
||||||
get packets) */
|
* get packets)
|
||||||
|
*/
|
||||||
mvneta_rxq_non_occup_desc_add(pp, rxq, i);
|
mvneta_rxq_non_occup_desc_add(pp, rxq, i);
|
||||||
|
|
||||||
return i;
|
return i;
|
||||||
|
@ -2201,7 +2203,7 @@ static int mvneta_check_mtu_valid(struct net_device *dev, int mtu)
|
||||||
return -EINVAL;
|
return -EINVAL;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* 9676 == 9700 - 20 and rounding to 8 */
|
/* 9676 == 9700 - 20 and rounding to 8 */
|
||||||
if (mtu > 9676) {
|
if (mtu > 9676) {
|
||||||
netdev_info(dev, "Illegal MTU value %d, round to 9676\n", mtu);
|
netdev_info(dev, "Illegal MTU value %d, round to 9676\n", mtu);
|
||||||
mtu = 9676;
|
mtu = 9676;
|
||||||
|
@ -2231,8 +2233,7 @@ static int mvneta_change_mtu(struct net_device *dev, int mtu)
|
||||||
if (!netif_running(dev))
|
if (!netif_running(dev))
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
/*
|
/* The interface is running, so we have to force a
|
||||||
* The interface is running, so we have to force a
|
|
||||||
* reallocation of the RXQs
|
* reallocation of the RXQs
|
||||||
*/
|
*/
|
||||||
mvneta_stop_dev(pp);
|
mvneta_stop_dev(pp);
|
||||||
|
@ -2677,8 +2678,7 @@ static int __devinit mvneta_probe(struct platform_device *pdev)
|
||||||
int phy_mode;
|
int phy_mode;
|
||||||
int err;
|
int err;
|
||||||
|
|
||||||
/*
|
/* Our multiqueue support is not complete, so for now, only
|
||||||
* Our multiqueue support is not complete, so for now, only
|
|
||||||
* allow the usage of the first RX queue
|
* allow the usage of the first RX queue
|
||||||
*/
|
*/
|
||||||
if (rxq_def != 0) {
|
if (rxq_def != 0) {
|
||||||
|
|
Loading…
Reference in New Issue