|
@@ -189,7 +189,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
if (tx_q->dirty_tx != tx_q->cur_tx)
|
|
if (tx_q->dirty_tx != tx_q->cur_tx)
|
|
return -EBUSY; /* still unfinished work */
|
|
return -EBUSY; /* still unfinished work */
|
|
-@@ -1309,7 +1309,7 @@ static void stmmac_display_rx_rings(stru
|
|
|
|
|
|
+@@ -1310,7 +1310,7 @@ static void stmmac_display_rx_rings(stru
|
|
|
|
|
|
/* Display RX rings */
|
|
/* Display RX rings */
|
|
for (queue = 0; queue < rx_cnt; queue++) {
|
|
for (queue = 0; queue < rx_cnt; queue++) {
|
|
@@ -198,7 +198,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
pr_info("\tRX Queue %u rings\n", queue);
|
|
pr_info("\tRX Queue %u rings\n", queue);
|
|
|
|
|
|
-@@ -1322,7 +1322,7 @@ static void stmmac_display_rx_rings(stru
|
|
|
|
|
|
+@@ -1323,7 +1323,7 @@ static void stmmac_display_rx_rings(stru
|
|
}
|
|
}
|
|
|
|
|
|
/* Display RX ring */
|
|
/* Display RX ring */
|
|
@@ -207,7 +207,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
rx_q->dma_rx_phy, desc_size);
|
|
rx_q->dma_rx_phy, desc_size);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
-@@ -1336,7 +1336,7 @@ static void stmmac_display_tx_rings(stru
|
|
|
|
|
|
+@@ -1337,7 +1337,7 @@ static void stmmac_display_tx_rings(stru
|
|
|
|
|
|
/* Display TX rings */
|
|
/* Display TX rings */
|
|
for (queue = 0; queue < tx_cnt; queue++) {
|
|
for (queue = 0; queue < tx_cnt; queue++) {
|
|
@@ -216,7 +216,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
pr_info("\tTX Queue %d rings\n", queue);
|
|
pr_info("\tTX Queue %d rings\n", queue);
|
|
|
|
|
|
-@@ -1351,7 +1351,7 @@ static void stmmac_display_tx_rings(stru
|
|
|
|
|
|
+@@ -1352,7 +1352,7 @@ static void stmmac_display_tx_rings(stru
|
|
desc_size = sizeof(struct dma_desc);
|
|
desc_size = sizeof(struct dma_desc);
|
|
}
|
|
}
|
|
|
|
|
|
@@ -225,7 +225,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
tx_q->dma_tx_phy, desc_size);
|
|
tx_q->dma_tx_phy, desc_size);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
-@@ -1392,21 +1392,21 @@ static int stmmac_set_bfsize(int mtu, in
|
|
|
|
|
|
+@@ -1393,21 +1393,21 @@ static int stmmac_set_bfsize(int mtu, in
|
|
*/
|
|
*/
|
|
static void stmmac_clear_rx_descriptors(struct stmmac_priv *priv, u32 queue)
|
|
static void stmmac_clear_rx_descriptors(struct stmmac_priv *priv, u32 queue)
|
|
{
|
|
{
|
|
@@ -253,7 +253,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
}
|
|
}
|
|
|
|
|
|
/**
|
|
/**
|
|
-@@ -1418,12 +1418,12 @@ static void stmmac_clear_rx_descriptors(
|
|
|
|
|
|
+@@ -1419,12 +1419,12 @@ static void stmmac_clear_rx_descriptors(
|
|
*/
|
|
*/
|
|
static void stmmac_clear_tx_descriptors(struct stmmac_priv *priv, u32 queue)
|
|
static void stmmac_clear_tx_descriptors(struct stmmac_priv *priv, u32 queue)
|
|
{
|
|
{
|
|
@@ -269,7 +269,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
struct dma_desc *p;
|
|
struct dma_desc *p;
|
|
|
|
|
|
if (priv->extend_desc)
|
|
if (priv->extend_desc)
|
|
-@@ -1471,7 +1471,7 @@ static void stmmac_clear_descriptors(str
|
|
|
|
|
|
+@@ -1472,7 +1472,7 @@ static void stmmac_clear_descriptors(str
|
|
static int stmmac_init_rx_buffers(struct stmmac_priv *priv, struct dma_desc *p,
|
|
static int stmmac_init_rx_buffers(struct stmmac_priv *priv, struct dma_desc *p,
|
|
int i, gfp_t flags, u32 queue)
|
|
int i, gfp_t flags, u32 queue)
|
|
{
|
|
{
|
|
@@ -278,7 +278,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
struct stmmac_rx_buffer *buf = &rx_q->buf_pool[i];
|
|
struct stmmac_rx_buffer *buf = &rx_q->buf_pool[i];
|
|
|
|
|
|
if (!buf->page) {
|
|
if (!buf->page) {
|
|
-@@ -1496,7 +1496,7 @@ static int stmmac_init_rx_buffers(struct
|
|
|
|
|
|
+@@ -1497,7 +1497,7 @@ static int stmmac_init_rx_buffers(struct
|
|
buf->addr = page_pool_get_dma_addr(buf->page) + buf->page_offset;
|
|
buf->addr = page_pool_get_dma_addr(buf->page) + buf->page_offset;
|
|
|
|
|
|
stmmac_set_desc_addr(priv, p, buf->addr);
|
|
stmmac_set_desc_addr(priv, p, buf->addr);
|
|
@@ -287,7 +287,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
stmmac_init_desc3(priv, p);
|
|
stmmac_init_desc3(priv, p);
|
|
|
|
|
|
return 0;
|
|
return 0;
|
|
-@@ -1510,7 +1510,7 @@ static int stmmac_init_rx_buffers(struct
|
|
|
|
|
|
+@@ -1511,7 +1511,7 @@ static int stmmac_init_rx_buffers(struct
|
|
*/
|
|
*/
|
|
static void stmmac_free_rx_buffer(struct stmmac_priv *priv, u32 queue, int i)
|
|
static void stmmac_free_rx_buffer(struct stmmac_priv *priv, u32 queue, int i)
|
|
{
|
|
{
|
|
@@ -296,7 +296,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
struct stmmac_rx_buffer *buf = &rx_q->buf_pool[i];
|
|
struct stmmac_rx_buffer *buf = &rx_q->buf_pool[i];
|
|
|
|
|
|
if (buf->page)
|
|
if (buf->page)
|
|
-@@ -1530,7 +1530,7 @@ static void stmmac_free_rx_buffer(struct
|
|
|
|
|
|
+@@ -1531,7 +1531,7 @@ static void stmmac_free_rx_buffer(struct
|
|
*/
|
|
*/
|
|
static void stmmac_free_tx_buffer(struct stmmac_priv *priv, u32 queue, int i)
|
|
static void stmmac_free_tx_buffer(struct stmmac_priv *priv, u32 queue, int i)
|
|
{
|
|
{
|
|
@@ -305,7 +305,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
if (tx_q->tx_skbuff_dma[i].buf &&
|
|
if (tx_q->tx_skbuff_dma[i].buf &&
|
|
tx_q->tx_skbuff_dma[i].buf_type != STMMAC_TXBUF_T_XDP_TX) {
|
|
tx_q->tx_skbuff_dma[i].buf_type != STMMAC_TXBUF_T_XDP_TX) {
|
|
-@@ -1575,17 +1575,17 @@ static void dma_free_rx_skbufs(struct st
|
|
|
|
|
|
+@@ -1576,17 +1576,17 @@ static void dma_free_rx_skbufs(struct st
|
|
{
|
|
{
|
|
int i;
|
|
int i;
|
|
|
|
|
|
@@ -326,7 +326,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
struct dma_desc *p;
|
|
struct dma_desc *p;
|
|
int ret;
|
|
int ret;
|
|
|
|
|
|
-@@ -1612,10 +1612,10 @@ static int stmmac_alloc_rx_buffers(struc
|
|
|
|
|
|
+@@ -1613,10 +1613,10 @@ static int stmmac_alloc_rx_buffers(struc
|
|
*/
|
|
*/
|
|
static void dma_free_rx_xskbufs(struct stmmac_priv *priv, u32 queue)
|
|
static void dma_free_rx_xskbufs(struct stmmac_priv *priv, u32 queue)
|
|
{
|
|
{
|
|
@@ -339,7 +339,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
struct stmmac_rx_buffer *buf = &rx_q->buf_pool[i];
|
|
struct stmmac_rx_buffer *buf = &rx_q->buf_pool[i];
|
|
|
|
|
|
if (!buf->xdp)
|
|
if (!buf->xdp)
|
|
-@@ -1628,10 +1628,10 @@ static void dma_free_rx_xskbufs(struct s
|
|
|
|
|
|
+@@ -1629,10 +1629,10 @@ static void dma_free_rx_xskbufs(struct s
|
|
|
|
|
|
static int stmmac_alloc_rx_buffers_zc(struct stmmac_priv *priv, u32 queue)
|
|
static int stmmac_alloc_rx_buffers_zc(struct stmmac_priv *priv, u32 queue)
|
|
{
|
|
{
|
|
@@ -352,7 +352,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
struct stmmac_rx_buffer *buf;
|
|
struct stmmac_rx_buffer *buf;
|
|
dma_addr_t dma_addr;
|
|
dma_addr_t dma_addr;
|
|
struct dma_desc *p;
|
|
struct dma_desc *p;
|
|
-@@ -1674,7 +1674,7 @@ static struct xsk_buff_pool *stmmac_get_
|
|
|
|
|
|
+@@ -1675,7 +1675,7 @@ static struct xsk_buff_pool *stmmac_get_
|
|
*/
|
|
*/
|
|
static int __init_dma_rx_desc_rings(struct stmmac_priv *priv, u32 queue, gfp_t flags)
|
|
static int __init_dma_rx_desc_rings(struct stmmac_priv *priv, u32 queue, gfp_t flags)
|
|
{
|
|
{
|
|
@@ -361,7 +361,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
int ret;
|
|
int ret;
|
|
|
|
|
|
netif_dbg(priv, probe, priv->dev,
|
|
netif_dbg(priv, probe, priv->dev,
|
|
-@@ -1720,11 +1720,11 @@ static int __init_dma_rx_desc_rings(stru
|
|
|
|
|
|
+@@ -1721,11 +1721,11 @@ static int __init_dma_rx_desc_rings(stru
|
|
if (priv->extend_desc)
|
|
if (priv->extend_desc)
|
|
stmmac_mode_init(priv, rx_q->dma_erx,
|
|
stmmac_mode_init(priv, rx_q->dma_erx,
|
|
rx_q->dma_rx_phy,
|
|
rx_q->dma_rx_phy,
|
|
@@ -375,7 +375,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
}
|
|
}
|
|
|
|
|
|
return 0;
|
|
return 0;
|
|
-@@ -1751,7 +1751,7 @@ static int init_dma_rx_desc_rings(struct
|
|
|
|
|
|
+@@ -1752,7 +1752,7 @@ static int init_dma_rx_desc_rings(struct
|
|
|
|
|
|
err_init_rx_buffers:
|
|
err_init_rx_buffers:
|
|
while (queue >= 0) {
|
|
while (queue >= 0) {
|
|
@@ -384,7 +384,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
if (rx_q->xsk_pool)
|
|
if (rx_q->xsk_pool)
|
|
dma_free_rx_xskbufs(priv, queue);
|
|
dma_free_rx_xskbufs(priv, queue);
|
|
-@@ -1780,7 +1780,7 @@ err_init_rx_buffers:
|
|
|
|
|
|
+@@ -1781,7 +1781,7 @@ err_init_rx_buffers:
|
|
*/
|
|
*/
|
|
static int __init_dma_tx_desc_rings(struct stmmac_priv *priv, u32 queue)
|
|
static int __init_dma_tx_desc_rings(struct stmmac_priv *priv, u32 queue)
|
|
{
|
|
{
|
|
@@ -393,7 +393,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
int i;
|
|
int i;
|
|
|
|
|
|
netif_dbg(priv, probe, priv->dev,
|
|
netif_dbg(priv, probe, priv->dev,
|
|
-@@ -1792,16 +1792,16 @@ static int __init_dma_tx_desc_rings(stru
|
|
|
|
|
|
+@@ -1793,16 +1793,16 @@ static int __init_dma_tx_desc_rings(stru
|
|
if (priv->extend_desc)
|
|
if (priv->extend_desc)
|
|
stmmac_mode_init(priv, tx_q->dma_etx,
|
|
stmmac_mode_init(priv, tx_q->dma_etx,
|
|
tx_q->dma_tx_phy,
|
|
tx_q->dma_tx_phy,
|
|
@@ -413,7 +413,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
struct dma_desc *p;
|
|
struct dma_desc *p;
|
|
|
|
|
|
if (priv->extend_desc)
|
|
if (priv->extend_desc)
|
|
-@@ -1871,12 +1871,12 @@ static int init_dma_desc_rings(struct ne
|
|
|
|
|
|
+@@ -1872,12 +1872,12 @@ static int init_dma_desc_rings(struct ne
|
|
*/
|
|
*/
|
|
static void dma_free_tx_skbufs(struct stmmac_priv *priv, u32 queue)
|
|
static void dma_free_tx_skbufs(struct stmmac_priv *priv, u32 queue)
|
|
{
|
|
{
|
|
@@ -428,7 +428,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
stmmac_free_tx_buffer(priv, queue, i);
|
|
stmmac_free_tx_buffer(priv, queue, i);
|
|
|
|
|
|
if (tx_q->xsk_pool && tx_q->xsk_frames_done) {
|
|
if (tx_q->xsk_pool && tx_q->xsk_frames_done) {
|
|
-@@ -1906,7 +1906,7 @@ static void stmmac_free_tx_skbufs(struct
|
|
|
|
|
|
+@@ -1907,7 +1907,7 @@ static void stmmac_free_tx_skbufs(struct
|
|
*/
|
|
*/
|
|
static void __free_dma_rx_desc_resources(struct stmmac_priv *priv, u32 queue)
|
|
static void __free_dma_rx_desc_resources(struct stmmac_priv *priv, u32 queue)
|
|
{
|
|
{
|
|
@@ -437,7 +437,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
/* Release the DMA RX socket buffers */
|
|
/* Release the DMA RX socket buffers */
|
|
if (rx_q->xsk_pool)
|
|
if (rx_q->xsk_pool)
|
|
-@@ -1919,11 +1919,11 @@ static void __free_dma_rx_desc_resources
|
|
|
|
|
|
+@@ -1920,11 +1920,11 @@ static void __free_dma_rx_desc_resources
|
|
|
|
|
|
/* Free DMA regions of consistent memory previously allocated */
|
|
/* Free DMA regions of consistent memory previously allocated */
|
|
if (!priv->extend_desc)
|
|
if (!priv->extend_desc)
|
|
@@ -451,7 +451,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
sizeof(struct dma_extended_desc),
|
|
sizeof(struct dma_extended_desc),
|
|
rx_q->dma_erx, rx_q->dma_rx_phy);
|
|
rx_q->dma_erx, rx_q->dma_rx_phy);
|
|
|
|
|
|
-@@ -1952,7 +1952,7 @@ static void free_dma_rx_desc_resources(s
|
|
|
|
|
|
+@@ -1953,7 +1953,7 @@ static void free_dma_rx_desc_resources(s
|
|
*/
|
|
*/
|
|
static void __free_dma_tx_desc_resources(struct stmmac_priv *priv, u32 queue)
|
|
static void __free_dma_tx_desc_resources(struct stmmac_priv *priv, u32 queue)
|
|
{
|
|
{
|
|
@@ -460,7 +460,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
size_t size;
|
|
size_t size;
|
|
void *addr;
|
|
void *addr;
|
|
|
|
|
|
-@@ -1970,7 +1970,7 @@ static void __free_dma_tx_desc_resources
|
|
|
|
|
|
+@@ -1971,7 +1971,7 @@ static void __free_dma_tx_desc_resources
|
|
addr = tx_q->dma_tx;
|
|
addr = tx_q->dma_tx;
|
|
}
|
|
}
|
|
|
|
|
|
@@ -469,7 +469,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
dma_free_coherent(priv->device, size, addr, tx_q->dma_tx_phy);
|
|
dma_free_coherent(priv->device, size, addr, tx_q->dma_tx_phy);
|
|
|
|
|
|
-@@ -1999,7 +1999,7 @@ static void free_dma_tx_desc_resources(s
|
|
|
|
|
|
+@@ -2000,7 +2000,7 @@ static void free_dma_tx_desc_resources(s
|
|
*/
|
|
*/
|
|
static int __alloc_dma_rx_desc_resources(struct stmmac_priv *priv, u32 queue)
|
|
static int __alloc_dma_rx_desc_resources(struct stmmac_priv *priv, u32 queue)
|
|
{
|
|
{
|
|
@@ -478,7 +478,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
struct stmmac_channel *ch = &priv->channel[queue];
|
|
struct stmmac_channel *ch = &priv->channel[queue];
|
|
bool xdp_prog = stmmac_xdp_is_enabled(priv);
|
|
bool xdp_prog = stmmac_xdp_is_enabled(priv);
|
|
struct page_pool_params pp_params = { 0 };
|
|
struct page_pool_params pp_params = { 0 };
|
|
-@@ -2011,8 +2011,8 @@ static int __alloc_dma_rx_desc_resources
|
|
|
|
|
|
+@@ -2012,8 +2012,8 @@ static int __alloc_dma_rx_desc_resources
|
|
rx_q->priv_data = priv;
|
|
rx_q->priv_data = priv;
|
|
|
|
|
|
pp_params.flags = PP_FLAG_DMA_MAP | PP_FLAG_DMA_SYNC_DEV;
|
|
pp_params.flags = PP_FLAG_DMA_MAP | PP_FLAG_DMA_SYNC_DEV;
|
|
@@ -489,7 +489,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
pp_params.order = ilog2(num_pages);
|
|
pp_params.order = ilog2(num_pages);
|
|
pp_params.nid = dev_to_node(priv->device);
|
|
pp_params.nid = dev_to_node(priv->device);
|
|
pp_params.dev = priv->device;
|
|
pp_params.dev = priv->device;
|
|
-@@ -2027,7 +2027,7 @@ static int __alloc_dma_rx_desc_resources
|
|
|
|
|
|
+@@ -2028,7 +2028,7 @@ static int __alloc_dma_rx_desc_resources
|
|
return ret;
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
|
|
@@ -498,7 +498,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
sizeof(*rx_q->buf_pool),
|
|
sizeof(*rx_q->buf_pool),
|
|
GFP_KERNEL);
|
|
GFP_KERNEL);
|
|
if (!rx_q->buf_pool)
|
|
if (!rx_q->buf_pool)
|
|
-@@ -2035,7 +2035,7 @@ static int __alloc_dma_rx_desc_resources
|
|
|
|
|
|
+@@ -2036,7 +2036,7 @@ static int __alloc_dma_rx_desc_resources
|
|
|
|
|
|
if (priv->extend_desc) {
|
|
if (priv->extend_desc) {
|
|
rx_q->dma_erx = dma_alloc_coherent(priv->device,
|
|
rx_q->dma_erx = dma_alloc_coherent(priv->device,
|
|
@@ -507,7 +507,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
sizeof(struct dma_extended_desc),
|
|
sizeof(struct dma_extended_desc),
|
|
&rx_q->dma_rx_phy,
|
|
&rx_q->dma_rx_phy,
|
|
GFP_KERNEL);
|
|
GFP_KERNEL);
|
|
-@@ -2044,7 +2044,7 @@ static int __alloc_dma_rx_desc_resources
|
|
|
|
|
|
+@@ -2045,7 +2045,7 @@ static int __alloc_dma_rx_desc_resources
|
|
|
|
|
|
} else {
|
|
} else {
|
|
rx_q->dma_rx = dma_alloc_coherent(priv->device,
|
|
rx_q->dma_rx = dma_alloc_coherent(priv->device,
|
|
@@ -516,7 +516,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
sizeof(struct dma_desc),
|
|
sizeof(struct dma_desc),
|
|
&rx_q->dma_rx_phy,
|
|
&rx_q->dma_rx_phy,
|
|
GFP_KERNEL);
|
|
GFP_KERNEL);
|
|
-@@ -2101,20 +2101,20 @@ err_dma:
|
|
|
|
|
|
+@@ -2102,20 +2102,20 @@ err_dma:
|
|
*/
|
|
*/
|
|
static int __alloc_dma_tx_desc_resources(struct stmmac_priv *priv, u32 queue)
|
|
static int __alloc_dma_tx_desc_resources(struct stmmac_priv *priv, u32 queue)
|
|
{
|
|
{
|
|
@@ -540,7 +540,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
sizeof(struct sk_buff *),
|
|
sizeof(struct sk_buff *),
|
|
GFP_KERNEL);
|
|
GFP_KERNEL);
|
|
if (!tx_q->tx_skbuff)
|
|
if (!tx_q->tx_skbuff)
|
|
-@@ -2127,7 +2127,7 @@ static int __alloc_dma_tx_desc_resources
|
|
|
|
|
|
+@@ -2128,7 +2128,7 @@ static int __alloc_dma_tx_desc_resources
|
|
else
|
|
else
|
|
size = sizeof(struct dma_desc);
|
|
size = sizeof(struct dma_desc);
|
|
|
|
|
|
@@ -549,7 +549,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
addr = dma_alloc_coherent(priv->device, size,
|
|
addr = dma_alloc_coherent(priv->device, size,
|
|
&tx_q->dma_tx_phy, GFP_KERNEL);
|
|
&tx_q->dma_tx_phy, GFP_KERNEL);
|
|
-@@ -2371,7 +2371,7 @@ static void stmmac_dma_operation_mode(st
|
|
|
|
|
|
+@@ -2372,7 +2372,7 @@ static void stmmac_dma_operation_mode(st
|
|
|
|
|
|
/* configure all channels */
|
|
/* configure all channels */
|
|
for (chan = 0; chan < rx_channels_count; chan++) {
|
|
for (chan = 0; chan < rx_channels_count; chan++) {
|
|
@@ -558,7 +558,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
u32 buf_size;
|
|
u32 buf_size;
|
|
|
|
|
|
qmode = priv->plat->rx_queues_cfg[chan].mode_to_use;
|
|
qmode = priv->plat->rx_queues_cfg[chan].mode_to_use;
|
|
-@@ -2386,7 +2386,7 @@ static void stmmac_dma_operation_mode(st
|
|
|
|
|
|
+@@ -2387,7 +2387,7 @@ static void stmmac_dma_operation_mode(st
|
|
chan);
|
|
chan);
|
|
} else {
|
|
} else {
|
|
stmmac_set_dma_bfsize(priv, priv->ioaddr,
|
|
stmmac_set_dma_bfsize(priv, priv->ioaddr,
|
|
@@ -567,7 +567,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
chan);
|
|
chan);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
-@@ -2402,7 +2402,7 @@ static void stmmac_dma_operation_mode(st
|
|
|
|
|
|
+@@ -2403,7 +2403,7 @@ static void stmmac_dma_operation_mode(st
|
|
static bool stmmac_xdp_xmit_zc(struct stmmac_priv *priv, u32 queue, u32 budget)
|
|
static bool stmmac_xdp_xmit_zc(struct stmmac_priv *priv, u32 queue, u32 budget)
|
|
{
|
|
{
|
|
struct netdev_queue *nq = netdev_get_tx_queue(priv->dev, queue);
|
|
struct netdev_queue *nq = netdev_get_tx_queue(priv->dev, queue);
|
|
@@ -576,7 +576,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
struct xsk_buff_pool *pool = tx_q->xsk_pool;
|
|
struct xsk_buff_pool *pool = tx_q->xsk_pool;
|
|
unsigned int entry = tx_q->cur_tx;
|
|
unsigned int entry = tx_q->cur_tx;
|
|
struct dma_desc *tx_desc = NULL;
|
|
struct dma_desc *tx_desc = NULL;
|
|
-@@ -2477,7 +2477,7 @@ static bool stmmac_xdp_xmit_zc(struct st
|
|
|
|
|
|
+@@ -2478,7 +2478,7 @@ static bool stmmac_xdp_xmit_zc(struct st
|
|
|
|
|
|
stmmac_enable_dma_transmission(priv, priv->ioaddr);
|
|
stmmac_enable_dma_transmission(priv, priv->ioaddr);
|
|
|
|
|
|
@@ -585,7 +585,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
entry = tx_q->cur_tx;
|
|
entry = tx_q->cur_tx;
|
|
}
|
|
}
|
|
|
|
|
|
-@@ -2503,7 +2503,7 @@ static bool stmmac_xdp_xmit_zc(struct st
|
|
|
|
|
|
+@@ -2504,7 +2504,7 @@ static bool stmmac_xdp_xmit_zc(struct st
|
|
*/
|
|
*/
|
|
static int stmmac_tx_clean(struct stmmac_priv *priv, int budget, u32 queue)
|
|
static int stmmac_tx_clean(struct stmmac_priv *priv, int budget, u32 queue)
|
|
{
|
|
{
|
|
@@ -594,7 +594,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
unsigned int bytes_compl = 0, pkts_compl = 0;
|
|
unsigned int bytes_compl = 0, pkts_compl = 0;
|
|
unsigned int entry, xmits = 0, count = 0;
|
|
unsigned int entry, xmits = 0, count = 0;
|
|
|
|
|
|
-@@ -2516,7 +2516,7 @@ static int stmmac_tx_clean(struct stmmac
|
|
|
|
|
|
+@@ -2517,7 +2517,7 @@ static int stmmac_tx_clean(struct stmmac
|
|
entry = tx_q->dirty_tx;
|
|
entry = tx_q->dirty_tx;
|
|
|
|
|
|
/* Try to clean all TX complete frame in 1 shot */
|
|
/* Try to clean all TX complete frame in 1 shot */
|
|
@@ -603,7 +603,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
struct xdp_frame *xdpf;
|
|
struct xdp_frame *xdpf;
|
|
struct sk_buff *skb;
|
|
struct sk_buff *skb;
|
|
struct dma_desc *p;
|
|
struct dma_desc *p;
|
|
-@@ -2616,7 +2616,7 @@ static int stmmac_tx_clean(struct stmmac
|
|
|
|
|
|
+@@ -2617,7 +2617,7 @@ static int stmmac_tx_clean(struct stmmac
|
|
|
|
|
|
stmmac_release_tx_desc(priv, p, priv->mode);
|
|
stmmac_release_tx_desc(priv, p, priv->mode);
|
|
|
|
|
|
@@ -612,7 +612,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
}
|
|
}
|
|
tx_q->dirty_tx = entry;
|
|
tx_q->dirty_tx = entry;
|
|
|
|
|
|
-@@ -2681,7 +2681,7 @@ static int stmmac_tx_clean(struct stmmac
|
|
|
|
|
|
+@@ -2682,7 +2682,7 @@ static int stmmac_tx_clean(struct stmmac
|
|
*/
|
|
*/
|
|
static void stmmac_tx_err(struct stmmac_priv *priv, u32 chan)
|
|
static void stmmac_tx_err(struct stmmac_priv *priv, u32 chan)
|
|
{
|
|
{
|
|
@@ -621,7 +621,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
netif_tx_stop_queue(netdev_get_tx_queue(priv->dev, chan));
|
|
netif_tx_stop_queue(netdev_get_tx_queue(priv->dev, chan));
|
|
|
|
|
|
-@@ -2748,8 +2748,8 @@ static int stmmac_napi_check(struct stmm
|
|
|
|
|
|
+@@ -2749,8 +2749,8 @@ static int stmmac_napi_check(struct stmm
|
|
{
|
|
{
|
|
int status = stmmac_dma_interrupt_status(priv, priv->ioaddr,
|
|
int status = stmmac_dma_interrupt_status(priv, priv->ioaddr,
|
|
&priv->xstats, chan, dir);
|
|
&priv->xstats, chan, dir);
|
|
@@ -632,7 +632,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
struct stmmac_channel *ch = &priv->channel[chan];
|
|
struct stmmac_channel *ch = &priv->channel[chan];
|
|
struct napi_struct *rx_napi;
|
|
struct napi_struct *rx_napi;
|
|
struct napi_struct *tx_napi;
|
|
struct napi_struct *tx_napi;
|
|
-@@ -2925,7 +2925,7 @@ static int stmmac_init_dma_engine(struct
|
|
|
|
|
|
+@@ -2926,7 +2926,7 @@ static int stmmac_init_dma_engine(struct
|
|
|
|
|
|
/* DMA RX Channel Configuration */
|
|
/* DMA RX Channel Configuration */
|
|
for (chan = 0; chan < rx_channels_count; chan++) {
|
|
for (chan = 0; chan < rx_channels_count; chan++) {
|
|
@@ -641,7 +641,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
stmmac_init_rx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
|
|
stmmac_init_rx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
|
|
rx_q->dma_rx_phy, chan);
|
|
rx_q->dma_rx_phy, chan);
|
|
-@@ -2939,7 +2939,7 @@ static int stmmac_init_dma_engine(struct
|
|
|
|
|
|
+@@ -2940,7 +2940,7 @@ static int stmmac_init_dma_engine(struct
|
|
|
|
|
|
/* DMA TX Channel Configuration */
|
|
/* DMA TX Channel Configuration */
|
|
for (chan = 0; chan < tx_channels_count; chan++) {
|
|
for (chan = 0; chan < tx_channels_count; chan++) {
|
|
@@ -650,7 +650,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
|
|
stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
|
|
tx_q->dma_tx_phy, chan);
|
|
tx_q->dma_tx_phy, chan);
|
|
-@@ -2954,7 +2954,7 @@ static int stmmac_init_dma_engine(struct
|
|
|
|
|
|
+@@ -2955,7 +2955,7 @@ static int stmmac_init_dma_engine(struct
|
|
|
|
|
|
static void stmmac_tx_timer_arm(struct stmmac_priv *priv, u32 queue)
|
|
static void stmmac_tx_timer_arm(struct stmmac_priv *priv, u32 queue)
|
|
{
|
|
{
|
|
@@ -659,7 +659,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
hrtimer_start(&tx_q->txtimer,
|
|
hrtimer_start(&tx_q->txtimer,
|
|
STMMAC_COAL_TIMER(priv->tx_coal_timer[queue]),
|
|
STMMAC_COAL_TIMER(priv->tx_coal_timer[queue]),
|
|
-@@ -3004,7 +3004,7 @@ static void stmmac_init_coalesce(struct
|
|
|
|
|
|
+@@ -3005,7 +3005,7 @@ static void stmmac_init_coalesce(struct
|
|
u32 chan;
|
|
u32 chan;
|
|
|
|
|
|
for (chan = 0; chan < tx_channel_count; chan++) {
|
|
for (chan = 0; chan < tx_channel_count; chan++) {
|
|
@@ -668,7 +668,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
priv->tx_coal_frames[chan] = STMMAC_TX_FRAMES;
|
|
priv->tx_coal_frames[chan] = STMMAC_TX_FRAMES;
|
|
priv->tx_coal_timer[chan] = STMMAC_COAL_TX_TIMER;
|
|
priv->tx_coal_timer[chan] = STMMAC_COAL_TX_TIMER;
|
|
-@@ -3026,12 +3026,12 @@ static void stmmac_set_rings_length(stru
|
|
|
|
|
|
+@@ -3027,12 +3027,12 @@ static void stmmac_set_rings_length(stru
|
|
/* set TX ring length */
|
|
/* set TX ring length */
|
|
for (chan = 0; chan < tx_channels_count; chan++)
|
|
for (chan = 0; chan < tx_channels_count; chan++)
|
|
stmmac_set_tx_ring_len(priv, priv->ioaddr,
|
|
stmmac_set_tx_ring_len(priv, priv->ioaddr,
|
|
@@ -683,7 +683,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
}
|
|
}
|
|
|
|
|
|
/**
|
|
/**
|
|
-@@ -3366,7 +3366,7 @@ static int stmmac_hw_setup(struct net_de
|
|
|
|
|
|
+@@ -3367,7 +3367,7 @@ static int stmmac_hw_setup(struct net_de
|
|
/* Enable TSO */
|
|
/* Enable TSO */
|
|
if (priv->tso) {
|
|
if (priv->tso) {
|
|
for (chan = 0; chan < tx_cnt; chan++) {
|
|
for (chan = 0; chan < tx_cnt; chan++) {
|
|
@@ -692,7 +692,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
/* TSO and TBS cannot co-exist */
|
|
/* TSO and TBS cannot co-exist */
|
|
if (tx_q->tbs & STMMAC_TBS_AVAIL)
|
|
if (tx_q->tbs & STMMAC_TBS_AVAIL)
|
|
-@@ -3388,7 +3388,7 @@ static int stmmac_hw_setup(struct net_de
|
|
|
|
|
|
+@@ -3389,7 +3389,7 @@ static int stmmac_hw_setup(struct net_de
|
|
|
|
|
|
/* TBS */
|
|
/* TBS */
|
|
for (chan = 0; chan < tx_cnt; chan++) {
|
|
for (chan = 0; chan < tx_cnt; chan++) {
|
|
@@ -701,7 +701,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
int enable = tx_q->tbs & STMMAC_TBS_AVAIL;
|
|
int enable = tx_q->tbs & STMMAC_TBS_AVAIL;
|
|
|
|
|
|
stmmac_enable_tbs(priv, priv->ioaddr, enable, chan);
|
|
stmmac_enable_tbs(priv, priv->ioaddr, enable, chan);
|
|
-@@ -3432,7 +3432,7 @@ static void stmmac_free_irq(struct net_d
|
|
|
|
|
|
+@@ -3433,7 +3433,7 @@ static void stmmac_free_irq(struct net_d
|
|
for (j = irq_idx - 1; j >= 0; j--) {
|
|
for (j = irq_idx - 1; j >= 0; j--) {
|
|
if (priv->tx_irq[j] > 0) {
|
|
if (priv->tx_irq[j] > 0) {
|
|
irq_set_affinity_hint(priv->tx_irq[j], NULL);
|
|
irq_set_affinity_hint(priv->tx_irq[j], NULL);
|
|
@@ -710,7 +710,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
}
|
|
}
|
|
}
|
|
}
|
|
irq_idx = priv->plat->rx_queues_to_use;
|
|
irq_idx = priv->plat->rx_queues_to_use;
|
|
-@@ -3441,7 +3441,7 @@ static void stmmac_free_irq(struct net_d
|
|
|
|
|
|
+@@ -3442,7 +3442,7 @@ static void stmmac_free_irq(struct net_d
|
|
for (j = irq_idx - 1; j >= 0; j--) {
|
|
for (j = irq_idx - 1; j >= 0; j--) {
|
|
if (priv->rx_irq[j] > 0) {
|
|
if (priv->rx_irq[j] > 0) {
|
|
irq_set_affinity_hint(priv->rx_irq[j], NULL);
|
|
irq_set_affinity_hint(priv->rx_irq[j], NULL);
|
|
@@ -719,7 +719,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
-@@ -3574,7 +3574,7 @@ static int stmmac_request_irq_multi_msi(
|
|
|
|
|
|
+@@ -3575,7 +3575,7 @@ static int stmmac_request_irq_multi_msi(
|
|
sprintf(int_name, "%s:%s-%d", dev->name, "rx", i);
|
|
sprintf(int_name, "%s:%s-%d", dev->name, "rx", i);
|
|
ret = request_irq(priv->rx_irq[i],
|
|
ret = request_irq(priv->rx_irq[i],
|
|
stmmac_msi_intr_rx,
|
|
stmmac_msi_intr_rx,
|
|
@@ -728,7 +728,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
if (unlikely(ret < 0)) {
|
|
if (unlikely(ret < 0)) {
|
|
netdev_err(priv->dev,
|
|
netdev_err(priv->dev,
|
|
"%s: alloc rx-%d MSI %d (error: %d)\n",
|
|
"%s: alloc rx-%d MSI %d (error: %d)\n",
|
|
-@@ -3597,7 +3597,7 @@ static int stmmac_request_irq_multi_msi(
|
|
|
|
|
|
+@@ -3598,7 +3598,7 @@ static int stmmac_request_irq_multi_msi(
|
|
sprintf(int_name, "%s:%s-%d", dev->name, "tx", i);
|
|
sprintf(int_name, "%s:%s-%d", dev->name, "tx", i);
|
|
ret = request_irq(priv->tx_irq[i],
|
|
ret = request_irq(priv->tx_irq[i],
|
|
stmmac_msi_intr_tx,
|
|
stmmac_msi_intr_tx,
|
|
@@ -737,7 +737,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
if (unlikely(ret < 0)) {
|
|
if (unlikely(ret < 0)) {
|
|
netdev_err(priv->dev,
|
|
netdev_err(priv->dev,
|
|
"%s: alloc tx-%d MSI %d (error: %d)\n",
|
|
"%s: alloc tx-%d MSI %d (error: %d)\n",
|
|
-@@ -3728,21 +3728,21 @@ static int stmmac_open(struct net_device
|
|
|
|
|
|
+@@ -3729,21 +3729,21 @@ static int stmmac_open(struct net_device
|
|
bfsize = 0;
|
|
bfsize = 0;
|
|
|
|
|
|
if (bfsize < BUF_SIZE_16KiB)
|
|
if (bfsize < BUF_SIZE_16KiB)
|
|
@@ -766,7 +766,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
int tbs_en = priv->plat->tx_queues_cfg[chan].tbs_en;
|
|
int tbs_en = priv->plat->tx_queues_cfg[chan].tbs_en;
|
|
|
|
|
|
/* Setup per-TXQ tbs flag before TX descriptor alloc */
|
|
/* Setup per-TXQ tbs flag before TX descriptor alloc */
|
|
-@@ -3800,7 +3800,7 @@ irq_error:
|
|
|
|
|
|
+@@ -3801,7 +3801,7 @@ irq_error:
|
|
phylink_stop(priv->phylink);
|
|
phylink_stop(priv->phylink);
|
|
|
|
|
|
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
|
|
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
|
|
@@ -775,7 +775,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
stmmac_hw_teardown(dev);
|
|
stmmac_hw_teardown(dev);
|
|
init_error:
|
|
init_error:
|
|
-@@ -3842,7 +3842,7 @@ static int stmmac_release(struct net_dev
|
|
|
|
|
|
+@@ -3843,7 +3843,7 @@ static int stmmac_release(struct net_dev
|
|
stmmac_disable_all_queues(priv);
|
|
stmmac_disable_all_queues(priv);
|
|
|
|
|
|
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
|
|
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
|
|
@@ -784,7 +784,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
netif_tx_disable(dev);
|
|
netif_tx_disable(dev);
|
|
|
|
|
|
-@@ -3906,7 +3906,7 @@ static bool stmmac_vlan_insert(struct st
|
|
|
|
|
|
+@@ -3907,7 +3907,7 @@ static bool stmmac_vlan_insert(struct st
|
|
return false;
|
|
return false;
|
|
|
|
|
|
stmmac_set_tx_owner(priv, p);
|
|
stmmac_set_tx_owner(priv, p);
|
|
@@ -793,7 +793,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
return true;
|
|
return true;
|
|
}
|
|
}
|
|
|
|
|
|
-@@ -3924,7 +3924,7 @@ static bool stmmac_vlan_insert(struct st
|
|
|
|
|
|
+@@ -3925,7 +3925,7 @@ static bool stmmac_vlan_insert(struct st
|
|
static void stmmac_tso_allocator(struct stmmac_priv *priv, dma_addr_t des,
|
|
static void stmmac_tso_allocator(struct stmmac_priv *priv, dma_addr_t des,
|
|
int total_len, bool last_segment, u32 queue)
|
|
int total_len, bool last_segment, u32 queue)
|
|
{
|
|
{
|
|
@@ -802,7 +802,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
struct dma_desc *desc;
|
|
struct dma_desc *desc;
|
|
u32 buff_size;
|
|
u32 buff_size;
|
|
int tmp_len;
|
|
int tmp_len;
|
|
-@@ -3935,7 +3935,7 @@ static void stmmac_tso_allocator(struct
|
|
|
|
|
|
+@@ -3936,7 +3936,7 @@ static void stmmac_tso_allocator(struct
|
|
dma_addr_t curr_addr;
|
|
dma_addr_t curr_addr;
|
|
|
|
|
|
tx_q->cur_tx = STMMAC_GET_ENTRY(tx_q->cur_tx,
|
|
tx_q->cur_tx = STMMAC_GET_ENTRY(tx_q->cur_tx,
|
|
@@ -811,7 +811,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
WARN_ON(tx_q->tx_skbuff[tx_q->cur_tx]);
|
|
WARN_ON(tx_q->tx_skbuff[tx_q->cur_tx]);
|
|
|
|
|
|
if (tx_q->tbs & STMMAC_TBS_AVAIL)
|
|
if (tx_q->tbs & STMMAC_TBS_AVAIL)
|
|
-@@ -3963,7 +3963,7 @@ static void stmmac_tso_allocator(struct
|
|
|
|
|
|
+@@ -3964,7 +3964,7 @@ static void stmmac_tso_allocator(struct
|
|
|
|
|
|
static void stmmac_flush_tx_descriptors(struct stmmac_priv *priv, int queue)
|
|
static void stmmac_flush_tx_descriptors(struct stmmac_priv *priv, int queue)
|
|
{
|
|
{
|
|
@@ -820,7 +820,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
int desc_size;
|
|
int desc_size;
|
|
|
|
|
|
if (likely(priv->extend_desc))
|
|
if (likely(priv->extend_desc))
|
|
-@@ -4025,7 +4025,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
|
|
|
|
|
|
+@@ -4026,7 +4026,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
|
|
dma_addr_t des;
|
|
dma_addr_t des;
|
|
int i;
|
|
int i;
|
|
|
|
|
|
@@ -829,7 +829,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
first_tx = tx_q->cur_tx;
|
|
first_tx = tx_q->cur_tx;
|
|
|
|
|
|
/* Compute header lengths */
|
|
/* Compute header lengths */
|
|
-@@ -4065,7 +4065,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
|
|
|
|
|
|
+@@ -4066,7 +4066,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
|
|
stmmac_set_mss(priv, mss_desc, mss);
|
|
stmmac_set_mss(priv, mss_desc, mss);
|
|
tx_q->mss = mss;
|
|
tx_q->mss = mss;
|
|
tx_q->cur_tx = STMMAC_GET_ENTRY(tx_q->cur_tx,
|
|
tx_q->cur_tx = STMMAC_GET_ENTRY(tx_q->cur_tx,
|
|
@@ -838,7 +838,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
WARN_ON(tx_q->tx_skbuff[tx_q->cur_tx]);
|
|
WARN_ON(tx_q->tx_skbuff[tx_q->cur_tx]);
|
|
}
|
|
}
|
|
|
|
|
|
-@@ -4177,7 +4177,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
|
|
|
|
|
|
+@@ -4178,7 +4178,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
|
|
* ndo_start_xmit will fill this descriptor the next time it's
|
|
* ndo_start_xmit will fill this descriptor the next time it's
|
|
* called and stmmac_tx_clean may clean up to this descriptor.
|
|
* called and stmmac_tx_clean may clean up to this descriptor.
|
|
*/
|
|
*/
|
|
@@ -847,7 +847,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
if (unlikely(stmmac_tx_avail(priv, queue) <= (MAX_SKB_FRAGS + 1))) {
|
|
if (unlikely(stmmac_tx_avail(priv, queue) <= (MAX_SKB_FRAGS + 1))) {
|
|
netif_dbg(priv, hw, priv->dev, "%s: stop transmitted packets\n",
|
|
netif_dbg(priv, hw, priv->dev, "%s: stop transmitted packets\n",
|
|
-@@ -4265,7 +4265,7 @@ static netdev_tx_t stmmac_xmit(struct sk
|
|
|
|
|
|
+@@ -4266,7 +4266,7 @@ static netdev_tx_t stmmac_xmit(struct sk
|
|
int entry, first_tx;
|
|
int entry, first_tx;
|
|
dma_addr_t des;
|
|
dma_addr_t des;
|
|
|
|
|
|
@@ -856,7 +856,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
first_tx = tx_q->cur_tx;
|
|
first_tx = tx_q->cur_tx;
|
|
|
|
|
|
if (priv->tx_path_in_lpi_mode && priv->eee_sw_timer_en)
|
|
if (priv->tx_path_in_lpi_mode && priv->eee_sw_timer_en)
|
|
-@@ -4328,7 +4328,7 @@ static netdev_tx_t stmmac_xmit(struct sk
|
|
|
|
|
|
+@@ -4329,7 +4329,7 @@ static netdev_tx_t stmmac_xmit(struct sk
|
|
int len = skb_frag_size(frag);
|
|
int len = skb_frag_size(frag);
|
|
bool last_segment = (i == (nfrags - 1));
|
|
bool last_segment = (i == (nfrags - 1));
|
|
|
|
|
|
@@ -865,7 +865,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
WARN_ON(tx_q->tx_skbuff[entry]);
|
|
WARN_ON(tx_q->tx_skbuff[entry]);
|
|
|
|
|
|
if (likely(priv->extend_desc))
|
|
if (likely(priv->extend_desc))
|
|
-@@ -4399,7 +4399,7 @@ static netdev_tx_t stmmac_xmit(struct sk
|
|
|
|
|
|
+@@ -4400,7 +4400,7 @@ static netdev_tx_t stmmac_xmit(struct sk
|
|
* ndo_start_xmit will fill this descriptor the next time it's
|
|
* ndo_start_xmit will fill this descriptor the next time it's
|
|
* called and stmmac_tx_clean may clean up to this descriptor.
|
|
* called and stmmac_tx_clean may clean up to this descriptor.
|
|
*/
|
|
*/
|
|
@@ -874,7 +874,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
tx_q->cur_tx = entry;
|
|
tx_q->cur_tx = entry;
|
|
|
|
|
|
if (netif_msg_pktdata(priv)) {
|
|
if (netif_msg_pktdata(priv)) {
|
|
-@@ -4514,7 +4514,7 @@ static void stmmac_rx_vlan(struct net_de
|
|
|
|
|
|
+@@ -4515,7 +4515,7 @@ static void stmmac_rx_vlan(struct net_de
|
|
*/
|
|
*/
|
|
static inline void stmmac_rx_refill(struct stmmac_priv *priv, u32 queue)
|
|
static inline void stmmac_rx_refill(struct stmmac_priv *priv, u32 queue)
|
|
{
|
|
{
|
|
@@ -883,7 +883,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
int dirty = stmmac_rx_dirty(priv, queue);
|
|
int dirty = stmmac_rx_dirty(priv, queue);
|
|
unsigned int entry = rx_q->dirty_rx;
|
|
unsigned int entry = rx_q->dirty_rx;
|
|
|
|
|
|
-@@ -4564,7 +4564,7 @@ static inline void stmmac_rx_refill(stru
|
|
|
|
|
|
+@@ -4565,7 +4565,7 @@ static inline void stmmac_rx_refill(stru
|
|
dma_wmb();
|
|
dma_wmb();
|
|
stmmac_set_rx_owner(priv, p, use_rx_wd);
|
|
stmmac_set_rx_owner(priv, p, use_rx_wd);
|
|
|
|
|
|
@@ -892,7 +892,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
}
|
|
}
|
|
rx_q->dirty_rx = entry;
|
|
rx_q->dirty_rx = entry;
|
|
rx_q->rx_tail_addr = rx_q->dma_rx_phy +
|
|
rx_q->rx_tail_addr = rx_q->dma_rx_phy +
|
|
-@@ -4592,12 +4592,12 @@ static unsigned int stmmac_rx_buf1_len(s
|
|
|
|
|
|
+@@ -4593,12 +4593,12 @@ static unsigned int stmmac_rx_buf1_len(s
|
|
|
|
|
|
/* First descriptor, not last descriptor and not split header */
|
|
/* First descriptor, not last descriptor and not split header */
|
|
if (status & rx_not_ls)
|
|
if (status & rx_not_ls)
|
|
@@ -907,7 +907,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
}
|
|
}
|
|
|
|
|
|
static unsigned int stmmac_rx_buf2_len(struct stmmac_priv *priv,
|
|
static unsigned int stmmac_rx_buf2_len(struct stmmac_priv *priv,
|
|
-@@ -4613,7 +4613,7 @@ static unsigned int stmmac_rx_buf2_len(s
|
|
|
|
|
|
+@@ -4614,7 +4614,7 @@ static unsigned int stmmac_rx_buf2_len(s
|
|
|
|
|
|
/* Not last descriptor */
|
|
/* Not last descriptor */
|
|
if (status & rx_not_ls)
|
|
if (status & rx_not_ls)
|
|
@@ -916,7 +916,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
plen = stmmac_get_rx_frame_len(priv, p, coe);
|
|
plen = stmmac_get_rx_frame_len(priv, p, coe);
|
|
|
|
|
|
-@@ -4624,7 +4624,7 @@ static unsigned int stmmac_rx_buf2_len(s
|
|
|
|
|
|
+@@ -4625,7 +4625,7 @@ static unsigned int stmmac_rx_buf2_len(s
|
|
static int stmmac_xdp_xmit_xdpf(struct stmmac_priv *priv, int queue,
|
|
static int stmmac_xdp_xmit_xdpf(struct stmmac_priv *priv, int queue,
|
|
struct xdp_frame *xdpf, bool dma_map)
|
|
struct xdp_frame *xdpf, bool dma_map)
|
|
{
|
|
{
|
|
@@ -925,7 +925,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
unsigned int entry = tx_q->cur_tx;
|
|
unsigned int entry = tx_q->cur_tx;
|
|
struct dma_desc *tx_desc;
|
|
struct dma_desc *tx_desc;
|
|
dma_addr_t dma_addr;
|
|
dma_addr_t dma_addr;
|
|
-@@ -4687,7 +4687,7 @@ static int stmmac_xdp_xmit_xdpf(struct s
|
|
|
|
|
|
+@@ -4688,7 +4688,7 @@ static int stmmac_xdp_xmit_xdpf(struct s
|
|
|
|
|
|
stmmac_enable_dma_transmission(priv, priv->ioaddr);
|
|
stmmac_enable_dma_transmission(priv, priv->ioaddr);
|
|
|
|
|
|
@@ -934,7 +934,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
tx_q->cur_tx = entry;
|
|
tx_q->cur_tx = entry;
|
|
|
|
|
|
return STMMAC_XDP_TX;
|
|
return STMMAC_XDP_TX;
|
|
-@@ -4861,7 +4861,7 @@ static void stmmac_dispatch_skb_zc(struc
|
|
|
|
|
|
+@@ -4862,7 +4862,7 @@ static void stmmac_dispatch_skb_zc(struc
|
|
|
|
|
|
static bool stmmac_rx_refill_zc(struct stmmac_priv *priv, u32 queue, u32 budget)
|
|
static bool stmmac_rx_refill_zc(struct stmmac_priv *priv, u32 queue, u32 budget)
|
|
{
|
|
{
|
|
@@ -943,7 +943,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
unsigned int entry = rx_q->dirty_rx;
|
|
unsigned int entry = rx_q->dirty_rx;
|
|
struct dma_desc *rx_desc = NULL;
|
|
struct dma_desc *rx_desc = NULL;
|
|
bool ret = true;
|
|
bool ret = true;
|
|
-@@ -4904,7 +4904,7 @@ static bool stmmac_rx_refill_zc(struct s
|
|
|
|
|
|
+@@ -4905,7 +4905,7 @@ static bool stmmac_rx_refill_zc(struct s
|
|
dma_wmb();
|
|
dma_wmb();
|
|
stmmac_set_rx_owner(priv, rx_desc, use_rx_wd);
|
|
stmmac_set_rx_owner(priv, rx_desc, use_rx_wd);
|
|
|
|
|
|
@@ -952,7 +952,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
}
|
|
}
|
|
|
|
|
|
if (rx_desc) {
|
|
if (rx_desc) {
|
|
-@@ -4919,7 +4919,7 @@ static bool stmmac_rx_refill_zc(struct s
|
|
|
|
|
|
+@@ -4920,7 +4920,7 @@ static bool stmmac_rx_refill_zc(struct s
|
|
|
|
|
|
static int stmmac_rx_zc(struct stmmac_priv *priv, int limit, u32 queue)
|
|
static int stmmac_rx_zc(struct stmmac_priv *priv, int limit, u32 queue)
|
|
{
|
|
{
|
|
@@ -961,7 +961,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
unsigned int count = 0, error = 0, len = 0;
|
|
unsigned int count = 0, error = 0, len = 0;
|
|
int dirty = stmmac_rx_dirty(priv, queue);
|
|
int dirty = stmmac_rx_dirty(priv, queue);
|
|
unsigned int next_entry = rx_q->cur_rx;
|
|
unsigned int next_entry = rx_q->cur_rx;
|
|
-@@ -4941,7 +4941,7 @@ static int stmmac_rx_zc(struct stmmac_pr
|
|
|
|
|
|
+@@ -4942,7 +4942,7 @@ static int stmmac_rx_zc(struct stmmac_pr
|
|
desc_size = sizeof(struct dma_desc);
|
|
desc_size = sizeof(struct dma_desc);
|
|
}
|
|
}
|
|
|
|
|
|
@@ -970,7 +970,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
rx_q->dma_rx_phy, desc_size);
|
|
rx_q->dma_rx_phy, desc_size);
|
|
}
|
|
}
|
|
while (count < limit) {
|
|
while (count < limit) {
|
|
-@@ -4988,7 +4988,7 @@ read_again:
|
|
|
|
|
|
+@@ -4989,7 +4989,7 @@ read_again:
|
|
|
|
|
|
/* Prefetch the next RX descriptor */
|
|
/* Prefetch the next RX descriptor */
|
|
rx_q->cur_rx = STMMAC_GET_ENTRY(rx_q->cur_rx,
|
|
rx_q->cur_rx = STMMAC_GET_ENTRY(rx_q->cur_rx,
|
|
@@ -979,7 +979,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
next_entry = rx_q->cur_rx;
|
|
next_entry = rx_q->cur_rx;
|
|
|
|
|
|
if (priv->extend_desc)
|
|
if (priv->extend_desc)
|
|
-@@ -5109,7 +5109,7 @@ read_again:
|
|
|
|
|
|
+@@ -5110,7 +5110,7 @@ read_again:
|
|
*/
|
|
*/
|
|
static int stmmac_rx(struct stmmac_priv *priv, int limit, u32 queue)
|
|
static int stmmac_rx(struct stmmac_priv *priv, int limit, u32 queue)
|
|
{
|
|
{
|
|
@@ -988,7 +988,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
struct stmmac_channel *ch = &priv->channel[queue];
|
|
struct stmmac_channel *ch = &priv->channel[queue];
|
|
unsigned int count = 0, error = 0, len = 0;
|
|
unsigned int count = 0, error = 0, len = 0;
|
|
int status = 0, coe = priv->hw->rx_csum;
|
|
int status = 0, coe = priv->hw->rx_csum;
|
|
-@@ -5122,7 +5122,7 @@ static int stmmac_rx(struct stmmac_priv
|
|
|
|
|
|
+@@ -5123,7 +5123,7 @@ static int stmmac_rx(struct stmmac_priv
|
|
int buf_sz;
|
|
int buf_sz;
|
|
|
|
|
|
dma_dir = page_pool_get_dma_dir(rx_q->page_pool);
|
|
dma_dir = page_pool_get_dma_dir(rx_q->page_pool);
|
|
@@ -997,7 +997,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
if (netif_msg_rx_status(priv)) {
|
|
if (netif_msg_rx_status(priv)) {
|
|
void *rx_head;
|
|
void *rx_head;
|
|
-@@ -5136,7 +5136,7 @@ static int stmmac_rx(struct stmmac_priv
|
|
|
|
|
|
+@@ -5137,7 +5137,7 @@ static int stmmac_rx(struct stmmac_priv
|
|
desc_size = sizeof(struct dma_desc);
|
|
desc_size = sizeof(struct dma_desc);
|
|
}
|
|
}
|
|
|
|
|
|
@@ -1006,7 +1006,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
rx_q->dma_rx_phy, desc_size);
|
|
rx_q->dma_rx_phy, desc_size);
|
|
}
|
|
}
|
|
while (count < limit) {
|
|
while (count < limit) {
|
|
-@@ -5180,7 +5180,7 @@ read_again:
|
|
|
|
|
|
+@@ -5181,7 +5181,7 @@ read_again:
|
|
break;
|
|
break;
|
|
|
|
|
|
rx_q->cur_rx = STMMAC_GET_ENTRY(rx_q->cur_rx,
|
|
rx_q->cur_rx = STMMAC_GET_ENTRY(rx_q->cur_rx,
|
|
@@ -1015,7 +1015,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
next_entry = rx_q->cur_rx;
|
|
next_entry = rx_q->cur_rx;
|
|
|
|
|
|
if (priv->extend_desc)
|
|
if (priv->extend_desc)
|
|
-@@ -5314,7 +5314,7 @@ read_again:
|
|
|
|
|
|
+@@ -5315,7 +5315,7 @@ read_again:
|
|
buf1_len, dma_dir);
|
|
buf1_len, dma_dir);
|
|
skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags,
|
|
skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags,
|
|
buf->page, buf->page_offset, buf1_len,
|
|
buf->page, buf->page_offset, buf1_len,
|
|
@@ -1024,7 +1024,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
/* Data payload appended into SKB */
|
|
/* Data payload appended into SKB */
|
|
page_pool_release_page(rx_q->page_pool, buf->page);
|
|
page_pool_release_page(rx_q->page_pool, buf->page);
|
|
-@@ -5326,7 +5326,7 @@ read_again:
|
|
|
|
|
|
+@@ -5327,7 +5327,7 @@ read_again:
|
|
buf2_len, dma_dir);
|
|
buf2_len, dma_dir);
|
|
skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags,
|
|
skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags,
|
|
buf->sec_page, 0, buf2_len,
|
|
buf->sec_page, 0, buf2_len,
|
|
@@ -1033,7 +1033,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
/* Data payload appended into SKB */
|
|
/* Data payload appended into SKB */
|
|
page_pool_release_page(rx_q->page_pool, buf->sec_page);
|
|
page_pool_release_page(rx_q->page_pool, buf->sec_page);
|
|
-@@ -5768,11 +5768,13 @@ static irqreturn_t stmmac_safety_interru
|
|
|
|
|
|
+@@ -5770,11 +5770,13 @@ static irqreturn_t stmmac_safety_interru
|
|
static irqreturn_t stmmac_msi_intr_tx(int irq, void *data)
|
|
static irqreturn_t stmmac_msi_intr_tx(int irq, void *data)
|
|
{
|
|
{
|
|
struct stmmac_tx_queue *tx_q = (struct stmmac_tx_queue *)data;
|
|
struct stmmac_tx_queue *tx_q = (struct stmmac_tx_queue *)data;
|
|
@@ -1048,7 +1048,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
if (unlikely(!data)) {
|
|
if (unlikely(!data)) {
|
|
netdev_err(priv->dev, "%s: invalid dev pointer\n", __func__);
|
|
netdev_err(priv->dev, "%s: invalid dev pointer\n", __func__);
|
|
-@@ -5812,10 +5814,12 @@ static irqreturn_t stmmac_msi_intr_tx(in
|
|
|
|
|
|
+@@ -5814,10 +5816,12 @@ static irqreturn_t stmmac_msi_intr_tx(in
|
|
static irqreturn_t stmmac_msi_intr_rx(int irq, void *data)
|
|
static irqreturn_t stmmac_msi_intr_rx(int irq, void *data)
|
|
{
|
|
{
|
|
struct stmmac_rx_queue *rx_q = (struct stmmac_rx_queue *)data;
|
|
struct stmmac_rx_queue *rx_q = (struct stmmac_rx_queue *)data;
|
|
@@ -1062,7 +1062,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
if (unlikely(!data)) {
|
|
if (unlikely(!data)) {
|
|
netdev_err(priv->dev, "%s: invalid dev pointer\n", __func__);
|
|
netdev_err(priv->dev, "%s: invalid dev pointer\n", __func__);
|
|
-@@ -5846,10 +5850,10 @@ static void stmmac_poll_controller(struc
|
|
|
|
|
|
+@@ -5848,10 +5852,10 @@ static void stmmac_poll_controller(struc
|
|
|
|
|
|
if (priv->plat->multi_msi_en) {
|
|
if (priv->plat->multi_msi_en) {
|
|
for (i = 0; i < priv->plat->rx_queues_to_use; i++)
|
|
for (i = 0; i < priv->plat->rx_queues_to_use; i++)
|
|
@@ -1075,7 +1075,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
} else {
|
|
} else {
|
|
disable_irq(dev->irq);
|
|
disable_irq(dev->irq);
|
|
stmmac_interrupt(dev->irq, dev);
|
|
stmmac_interrupt(dev->irq, dev);
|
|
-@@ -6030,34 +6034,34 @@ static int stmmac_rings_status_show(stru
|
|
|
|
|
|
+@@ -6032,34 +6036,34 @@ static int stmmac_rings_status_show(stru
|
|
return 0;
|
|
return 0;
|
|
|
|
|
|
for (queue = 0; queue < rx_count; queue++) {
|
|
for (queue = 0; queue < rx_count; queue++) {
|
|
@@ -1116,7 +1116,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
-@@ -6404,7 +6408,7 @@ void stmmac_disable_rx_queue(struct stmm
|
|
|
|
|
|
+@@ -6406,7 +6410,7 @@ void stmmac_disable_rx_queue(struct stmm
|
|
|
|
|
|
void stmmac_enable_rx_queue(struct stmmac_priv *priv, u32 queue)
|
|
void stmmac_enable_rx_queue(struct stmmac_priv *priv, u32 queue)
|
|
{
|
|
{
|
|
@@ -1125,7 +1125,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
struct stmmac_channel *ch = &priv->channel[queue];
|
|
struct stmmac_channel *ch = &priv->channel[queue];
|
|
unsigned long flags;
|
|
unsigned long flags;
|
|
u32 buf_size;
|
|
u32 buf_size;
|
|
-@@ -6441,7 +6445,7 @@ void stmmac_enable_rx_queue(struct stmma
|
|
|
|
|
|
+@@ -6443,7 +6447,7 @@ void stmmac_enable_rx_queue(struct stmma
|
|
rx_q->queue_index);
|
|
rx_q->queue_index);
|
|
} else {
|
|
} else {
|
|
stmmac_set_dma_bfsize(priv, priv->ioaddr,
|
|
stmmac_set_dma_bfsize(priv, priv->ioaddr,
|
|
@@ -1134,7 +1134,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
rx_q->queue_index);
|
|
rx_q->queue_index);
|
|
}
|
|
}
|
|
|
|
|
|
-@@ -6467,7 +6471,7 @@ void stmmac_disable_tx_queue(struct stmm
|
|
|
|
|
|
+@@ -6469,7 +6473,7 @@ void stmmac_disable_tx_queue(struct stmm
|
|
|
|
|
|
void stmmac_enable_tx_queue(struct stmmac_priv *priv, u32 queue)
|
|
void stmmac_enable_tx_queue(struct stmmac_priv *priv, u32 queue)
|
|
{
|
|
{
|
|
@@ -1143,7 +1143,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
struct stmmac_channel *ch = &priv->channel[queue];
|
|
struct stmmac_channel *ch = &priv->channel[queue];
|
|
unsigned long flags;
|
|
unsigned long flags;
|
|
int ret;
|
|
int ret;
|
|
-@@ -6517,7 +6521,7 @@ void stmmac_xdp_release(struct net_devic
|
|
|
|
|
|
+@@ -6519,7 +6523,7 @@ void stmmac_xdp_release(struct net_devic
|
|
stmmac_disable_all_queues(priv);
|
|
stmmac_disable_all_queues(priv);
|
|
|
|
|
|
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
|
|
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
|
|
@@ -1152,7 +1152,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
/* Free the IRQ lines */
|
|
/* Free the IRQ lines */
|
|
stmmac_free_irq(dev, REQ_IRQ_ERR_ALL, 0);
|
|
stmmac_free_irq(dev, REQ_IRQ_ERR_ALL, 0);
|
|
-@@ -6576,7 +6580,7 @@ int stmmac_xdp_open(struct net_device *d
|
|
|
|
|
|
+@@ -6578,7 +6582,7 @@ int stmmac_xdp_open(struct net_device *d
|
|
|
|
|
|
/* DMA RX Channel Configuration */
|
|
/* DMA RX Channel Configuration */
|
|
for (chan = 0; chan < rx_cnt; chan++) {
|
|
for (chan = 0; chan < rx_cnt; chan++) {
|
|
@@ -1161,7 +1161,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
stmmac_init_rx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
|
|
stmmac_init_rx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
|
|
rx_q->dma_rx_phy, chan);
|
|
rx_q->dma_rx_phy, chan);
|
|
-@@ -6594,7 +6598,7 @@ int stmmac_xdp_open(struct net_device *d
|
|
|
|
|
|
+@@ -6596,7 +6600,7 @@ int stmmac_xdp_open(struct net_device *d
|
|
rx_q->queue_index);
|
|
rx_q->queue_index);
|
|
} else {
|
|
} else {
|
|
stmmac_set_dma_bfsize(priv, priv->ioaddr,
|
|
stmmac_set_dma_bfsize(priv, priv->ioaddr,
|
|
@@ -1170,7 +1170,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
rx_q->queue_index);
|
|
rx_q->queue_index);
|
|
}
|
|
}
|
|
|
|
|
|
-@@ -6603,7 +6607,7 @@ int stmmac_xdp_open(struct net_device *d
|
|
|
|
|
|
+@@ -6605,7 +6609,7 @@ int stmmac_xdp_open(struct net_device *d
|
|
|
|
|
|
/* DMA TX Channel Configuration */
|
|
/* DMA TX Channel Configuration */
|
|
for (chan = 0; chan < tx_cnt; chan++) {
|
|
for (chan = 0; chan < tx_cnt; chan++) {
|
|
@@ -1179,7 +1179,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
|
|
stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
|
|
tx_q->dma_tx_phy, chan);
|
|
tx_q->dma_tx_phy, chan);
|
|
-@@ -6636,7 +6640,7 @@ int stmmac_xdp_open(struct net_device *d
|
|
|
|
|
|
+@@ -6638,7 +6642,7 @@ int stmmac_xdp_open(struct net_device *d
|
|
|
|
|
|
irq_error:
|
|
irq_error:
|
|
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
|
|
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
|
|
@@ -1188,7 +1188,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
stmmac_hw_teardown(dev);
|
|
stmmac_hw_teardown(dev);
|
|
init_error:
|
|
init_error:
|
|
-@@ -6663,8 +6667,8 @@ int stmmac_xsk_wakeup(struct net_device
|
|
|
|
|
|
+@@ -6665,8 +6669,8 @@ int stmmac_xsk_wakeup(struct net_device
|
|
queue >= priv->plat->tx_queues_to_use)
|
|
queue >= priv->plat->tx_queues_to_use)
|
|
return -EINVAL;
|
|
return -EINVAL;
|
|
|
|
|
|
@@ -1199,7 +1199,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
ch = &priv->channel[queue];
|
|
ch = &priv->channel[queue];
|
|
|
|
|
|
if (!rx_q->xsk_pool && !tx_q->xsk_pool)
|
|
if (!rx_q->xsk_pool && !tx_q->xsk_pool)
|
|
-@@ -6924,8 +6928,8 @@ int stmmac_reinit_ringparam(struct net_d
|
|
|
|
|
|
+@@ -6926,8 +6930,8 @@ int stmmac_reinit_ringparam(struct net_d
|
|
if (netif_running(dev))
|
|
if (netif_running(dev))
|
|
stmmac_release(dev);
|
|
stmmac_release(dev);
|
|
|
|
|
|
@@ -1210,7 +1210,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
if (netif_running(dev))
|
|
if (netif_running(dev))
|
|
ret = stmmac_open(dev);
|
|
ret = stmmac_open(dev);
|
|
-@@ -7357,7 +7361,7 @@ int stmmac_suspend(struct device *dev)
|
|
|
|
|
|
+@@ -7362,7 +7366,7 @@ int stmmac_suspend(struct device *dev)
|
|
stmmac_disable_all_queues(priv);
|
|
stmmac_disable_all_queues(priv);
|
|
|
|
|
|
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
|
|
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
|
|
@@ -1219,7 +1219,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
if (priv->eee_enabled) {
|
|
if (priv->eee_enabled) {
|
|
priv->tx_path_in_lpi_mode = false;
|
|
priv->tx_path_in_lpi_mode = false;
|
|
-@@ -7408,7 +7412,7 @@ EXPORT_SYMBOL_GPL(stmmac_suspend);
|
|
|
|
|
|
+@@ -7414,7 +7418,7 @@ EXPORT_SYMBOL_GPL(stmmac_suspend);
|
|
|
|
|
|
static void stmmac_reset_rx_queue(struct stmmac_priv *priv, u32 queue)
|
|
static void stmmac_reset_rx_queue(struct stmmac_priv *priv, u32 queue)
|
|
{
|
|
{
|
|
@@ -1228,7 +1228,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
rx_q->cur_rx = 0;
|
|
rx_q->cur_rx = 0;
|
|
rx_q->dirty_rx = 0;
|
|
rx_q->dirty_rx = 0;
|
|
-@@ -7416,7 +7420,7 @@ static void stmmac_reset_rx_queue(struct
|
|
|
|
|
|
+@@ -7422,7 +7426,7 @@ static void stmmac_reset_rx_queue(struct
|
|
|
|
|
|
static void stmmac_reset_tx_queue(struct stmmac_priv *priv, u32 queue)
|
|
static void stmmac_reset_tx_queue(struct stmmac_priv *priv, u32 queue)
|
|
{
|
|
{
|
|
@@ -1270,7 +1270,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
if (i >= priv->plat->tx_queues_to_use)
|
|
if (i >= priv->plat->tx_queues_to_use)
|
|
--- a/drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
|
|
--- a/drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
|
|
+++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
|
|
+++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
|
|
-@@ -970,13 +970,13 @@ static int tc_setup_etf(struct stmmac_pr
|
|
|
|
|
|
+@@ -971,13 +971,13 @@ static int tc_setup_etf(struct stmmac_pr
|
|
return -EOPNOTSUPP;
|
|
return -EOPNOTSUPP;
|
|
if (qopt->queue >= priv->plat->tx_queues_to_use)
|
|
if (qopt->queue >= priv->plat->tx_queues_to_use)
|
|
return -EINVAL;
|
|
return -EINVAL;
|