|
@@ -775,7 +775,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
stmmac_hw_teardown(dev);
|
|
|
init_error:
|
|
|
-@@ -3844,7 +3844,7 @@ static int stmmac_release(struct net_dev
|
|
|
+@@ -3846,7 +3846,7 @@ static int stmmac_release(struct net_dev
|
|
|
stmmac_disable_all_queues(priv);
|
|
|
|
|
|
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
|
|
@@ -784,7 +784,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
netif_tx_disable(dev);
|
|
|
|
|
|
-@@ -3908,7 +3908,7 @@ static bool stmmac_vlan_insert(struct st
|
|
|
+@@ -3910,7 +3910,7 @@ static bool stmmac_vlan_insert(struct st
|
|
|
return false;
|
|
|
|
|
|
stmmac_set_tx_owner(priv, p);
|
|
@@ -793,7 +793,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
return true;
|
|
|
}
|
|
|
|
|
|
-@@ -3926,7 +3926,7 @@ static bool stmmac_vlan_insert(struct st
|
|
|
+@@ -3928,7 +3928,7 @@ static bool stmmac_vlan_insert(struct st
|
|
|
static void stmmac_tso_allocator(struct stmmac_priv *priv, dma_addr_t des,
|
|
|
int total_len, bool last_segment, u32 queue)
|
|
|
{
|
|
@@ -802,7 +802,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
struct dma_desc *desc;
|
|
|
u32 buff_size;
|
|
|
int tmp_len;
|
|
|
-@@ -3937,7 +3937,7 @@ static void stmmac_tso_allocator(struct
|
|
|
+@@ -3939,7 +3939,7 @@ static void stmmac_tso_allocator(struct
|
|
|
dma_addr_t curr_addr;
|
|
|
|
|
|
tx_q->cur_tx = STMMAC_GET_ENTRY(tx_q->cur_tx,
|
|
@@ -811,7 +811,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
WARN_ON(tx_q->tx_skbuff[tx_q->cur_tx]);
|
|
|
|
|
|
if (tx_q->tbs & STMMAC_TBS_AVAIL)
|
|
|
-@@ -3965,7 +3965,7 @@ static void stmmac_tso_allocator(struct
|
|
|
+@@ -3967,7 +3967,7 @@ static void stmmac_tso_allocator(struct
|
|
|
|
|
|
static void stmmac_flush_tx_descriptors(struct stmmac_priv *priv, int queue)
|
|
|
{
|
|
@@ -820,7 +820,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
int desc_size;
|
|
|
|
|
|
if (likely(priv->extend_desc))
|
|
|
-@@ -4027,7 +4027,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
|
|
|
+@@ -4029,7 +4029,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
|
|
|
dma_addr_t des;
|
|
|
int i;
|
|
|
|
|
@@ -829,7 +829,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
first_tx = tx_q->cur_tx;
|
|
|
|
|
|
/* Compute header lengths */
|
|
|
-@@ -4067,7 +4067,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
|
|
|
+@@ -4069,7 +4069,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
|
|
|
stmmac_set_mss(priv, mss_desc, mss);
|
|
|
tx_q->mss = mss;
|
|
|
tx_q->cur_tx = STMMAC_GET_ENTRY(tx_q->cur_tx,
|
|
@@ -838,7 +838,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
WARN_ON(tx_q->tx_skbuff[tx_q->cur_tx]);
|
|
|
}
|
|
|
|
|
|
-@@ -4179,7 +4179,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
|
|
|
+@@ -4181,7 +4181,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
|
|
|
* ndo_start_xmit will fill this descriptor the next time it's
|
|
|
* called and stmmac_tx_clean may clean up to this descriptor.
|
|
|
*/
|
|
@@ -847,7 +847,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
if (unlikely(stmmac_tx_avail(priv, queue) <= (MAX_SKB_FRAGS + 1))) {
|
|
|
netif_dbg(priv, hw, priv->dev, "%s: stop transmitted packets\n",
|
|
|
-@@ -4267,7 +4267,7 @@ static netdev_tx_t stmmac_xmit(struct sk
|
|
|
+@@ -4269,7 +4269,7 @@ static netdev_tx_t stmmac_xmit(struct sk
|
|
|
int entry, first_tx;
|
|
|
dma_addr_t des;
|
|
|
|
|
@@ -856,7 +856,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
first_tx = tx_q->cur_tx;
|
|
|
|
|
|
if (priv->tx_path_in_lpi_mode && priv->eee_sw_timer_en)
|
|
|
-@@ -4330,7 +4330,7 @@ static netdev_tx_t stmmac_xmit(struct sk
|
|
|
+@@ -4332,7 +4332,7 @@ static netdev_tx_t stmmac_xmit(struct sk
|
|
|
int len = skb_frag_size(frag);
|
|
|
bool last_segment = (i == (nfrags - 1));
|
|
|
|
|
@@ -865,7 +865,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
WARN_ON(tx_q->tx_skbuff[entry]);
|
|
|
|
|
|
if (likely(priv->extend_desc))
|
|
|
-@@ -4401,7 +4401,7 @@ static netdev_tx_t stmmac_xmit(struct sk
|
|
|
+@@ -4403,7 +4403,7 @@ static netdev_tx_t stmmac_xmit(struct sk
|
|
|
* ndo_start_xmit will fill this descriptor the next time it's
|
|
|
* called and stmmac_tx_clean may clean up to this descriptor.
|
|
|
*/
|
|
@@ -874,7 +874,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
tx_q->cur_tx = entry;
|
|
|
|
|
|
if (netif_msg_pktdata(priv)) {
|
|
|
-@@ -4513,7 +4513,7 @@ static void stmmac_rx_vlan(struct net_de
|
|
|
+@@ -4515,7 +4515,7 @@ static void stmmac_rx_vlan(struct net_de
|
|
|
*/
|
|
|
static inline void stmmac_rx_refill(struct stmmac_priv *priv, u32 queue)
|
|
|
{
|
|
@@ -883,7 +883,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
int dirty = stmmac_rx_dirty(priv, queue);
|
|
|
unsigned int entry = rx_q->dirty_rx;
|
|
|
|
|
|
-@@ -4563,7 +4563,7 @@ static inline void stmmac_rx_refill(stru
|
|
|
+@@ -4565,7 +4565,7 @@ static inline void stmmac_rx_refill(stru
|
|
|
dma_wmb();
|
|
|
stmmac_set_rx_owner(priv, p, use_rx_wd);
|
|
|
|
|
@@ -892,7 +892,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
}
|
|
|
rx_q->dirty_rx = entry;
|
|
|
rx_q->rx_tail_addr = rx_q->dma_rx_phy +
|
|
|
-@@ -4591,12 +4591,12 @@ static unsigned int stmmac_rx_buf1_len(s
|
|
|
+@@ -4593,12 +4593,12 @@ static unsigned int stmmac_rx_buf1_len(s
|
|
|
|
|
|
/* First descriptor, not last descriptor and not split header */
|
|
|
if (status & rx_not_ls)
|
|
@@ -907,7 +907,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
}
|
|
|
|
|
|
static unsigned int stmmac_rx_buf2_len(struct stmmac_priv *priv,
|
|
|
-@@ -4612,7 +4612,7 @@ static unsigned int stmmac_rx_buf2_len(s
|
|
|
+@@ -4614,7 +4614,7 @@ static unsigned int stmmac_rx_buf2_len(s
|
|
|
|
|
|
/* Not last descriptor */
|
|
|
if (status & rx_not_ls)
|
|
@@ -916,7 +916,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
plen = stmmac_get_rx_frame_len(priv, p, coe);
|
|
|
|
|
|
-@@ -4623,7 +4623,7 @@ static unsigned int stmmac_rx_buf2_len(s
|
|
|
+@@ -4625,7 +4625,7 @@ static unsigned int stmmac_rx_buf2_len(s
|
|
|
static int stmmac_xdp_xmit_xdpf(struct stmmac_priv *priv, int queue,
|
|
|
struct xdp_frame *xdpf, bool dma_map)
|
|
|
{
|
|
@@ -925,7 +925,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
unsigned int entry = tx_q->cur_tx;
|
|
|
struct dma_desc *tx_desc;
|
|
|
dma_addr_t dma_addr;
|
|
|
-@@ -4686,7 +4686,7 @@ static int stmmac_xdp_xmit_xdpf(struct s
|
|
|
+@@ -4688,7 +4688,7 @@ static int stmmac_xdp_xmit_xdpf(struct s
|
|
|
|
|
|
stmmac_enable_dma_transmission(priv, priv->ioaddr);
|
|
|
|
|
@@ -934,7 +934,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
tx_q->cur_tx = entry;
|
|
|
|
|
|
return STMMAC_XDP_TX;
|
|
|
-@@ -4860,7 +4860,7 @@ static void stmmac_dispatch_skb_zc(struc
|
|
|
+@@ -4862,7 +4862,7 @@ static void stmmac_dispatch_skb_zc(struc
|
|
|
|
|
|
static bool stmmac_rx_refill_zc(struct stmmac_priv *priv, u32 queue, u32 budget)
|
|
|
{
|
|
@@ -943,7 +943,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
unsigned int entry = rx_q->dirty_rx;
|
|
|
struct dma_desc *rx_desc = NULL;
|
|
|
bool ret = true;
|
|
|
-@@ -4903,7 +4903,7 @@ static bool stmmac_rx_refill_zc(struct s
|
|
|
+@@ -4905,7 +4905,7 @@ static bool stmmac_rx_refill_zc(struct s
|
|
|
dma_wmb();
|
|
|
stmmac_set_rx_owner(priv, rx_desc, use_rx_wd);
|
|
|
|
|
@@ -952,7 +952,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
}
|
|
|
|
|
|
if (rx_desc) {
|
|
|
-@@ -4918,7 +4918,7 @@ static bool stmmac_rx_refill_zc(struct s
|
|
|
+@@ -4920,7 +4920,7 @@ static bool stmmac_rx_refill_zc(struct s
|
|
|
|
|
|
static int stmmac_rx_zc(struct stmmac_priv *priv, int limit, u32 queue)
|
|
|
{
|
|
@@ -961,7 +961,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
unsigned int count = 0, error = 0, len = 0;
|
|
|
int dirty = stmmac_rx_dirty(priv, queue);
|
|
|
unsigned int next_entry = rx_q->cur_rx;
|
|
|
-@@ -4940,7 +4940,7 @@ static int stmmac_rx_zc(struct stmmac_pr
|
|
|
+@@ -4942,7 +4942,7 @@ static int stmmac_rx_zc(struct stmmac_pr
|
|
|
desc_size = sizeof(struct dma_desc);
|
|
|
}
|
|
|
|
|
@@ -970,7 +970,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
rx_q->dma_rx_phy, desc_size);
|
|
|
}
|
|
|
while (count < limit) {
|
|
|
-@@ -4987,7 +4987,7 @@ read_again:
|
|
|
+@@ -4989,7 +4989,7 @@ read_again:
|
|
|
|
|
|
/* Prefetch the next RX descriptor */
|
|
|
rx_q->cur_rx = STMMAC_GET_ENTRY(rx_q->cur_rx,
|
|
@@ -979,7 +979,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
next_entry = rx_q->cur_rx;
|
|
|
|
|
|
if (priv->extend_desc)
|
|
|
-@@ -5108,7 +5108,7 @@ read_again:
|
|
|
+@@ -5110,7 +5110,7 @@ read_again:
|
|
|
*/
|
|
|
static int stmmac_rx(struct stmmac_priv *priv, int limit, u32 queue)
|
|
|
{
|
|
@@ -988,7 +988,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
struct stmmac_channel *ch = &priv->channel[queue];
|
|
|
unsigned int count = 0, error = 0, len = 0;
|
|
|
int status = 0, coe = priv->hw->rx_csum;
|
|
|
-@@ -5121,7 +5121,7 @@ static int stmmac_rx(struct stmmac_priv
|
|
|
+@@ -5123,7 +5123,7 @@ static int stmmac_rx(struct stmmac_priv
|
|
|
int buf_sz;
|
|
|
|
|
|
dma_dir = page_pool_get_dma_dir(rx_q->page_pool);
|
|
@@ -997,7 +997,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
if (netif_msg_rx_status(priv)) {
|
|
|
void *rx_head;
|
|
|
-@@ -5135,7 +5135,7 @@ static int stmmac_rx(struct stmmac_priv
|
|
|
+@@ -5137,7 +5137,7 @@ static int stmmac_rx(struct stmmac_priv
|
|
|
desc_size = sizeof(struct dma_desc);
|
|
|
}
|
|
|
|
|
@@ -1006,7 +1006,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
rx_q->dma_rx_phy, desc_size);
|
|
|
}
|
|
|
while (count < limit) {
|
|
|
-@@ -5179,7 +5179,7 @@ read_again:
|
|
|
+@@ -5181,7 +5181,7 @@ read_again:
|
|
|
break;
|
|
|
|
|
|
rx_q->cur_rx = STMMAC_GET_ENTRY(rx_q->cur_rx,
|
|
@@ -1015,7 +1015,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
next_entry = rx_q->cur_rx;
|
|
|
|
|
|
if (priv->extend_desc)
|
|
|
-@@ -5313,7 +5313,7 @@ read_again:
|
|
|
+@@ -5315,7 +5315,7 @@ read_again:
|
|
|
buf1_len, dma_dir);
|
|
|
skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags,
|
|
|
buf->page, buf->page_offset, buf1_len,
|
|
@@ -1024,7 +1024,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
/* Data payload appended into SKB */
|
|
|
page_pool_release_page(rx_q->page_pool, buf->page);
|
|
|
-@@ -5325,7 +5325,7 @@ read_again:
|
|
|
+@@ -5327,7 +5327,7 @@ read_again:
|
|
|
buf2_len, dma_dir);
|
|
|
skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags,
|
|
|
buf->sec_page, 0, buf2_len,
|
|
@@ -1033,7 +1033,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
/* Data payload appended into SKB */
|
|
|
page_pool_release_page(rx_q->page_pool, buf->sec_page);
|
|
|
-@@ -5758,11 +5758,13 @@ static irqreturn_t stmmac_safety_interru
|
|
|
+@@ -5760,11 +5760,13 @@ static irqreturn_t stmmac_safety_interru
|
|
|
static irqreturn_t stmmac_msi_intr_tx(int irq, void *data)
|
|
|
{
|
|
|
struct stmmac_tx_queue *tx_q = (struct stmmac_tx_queue *)data;
|
|
@@ -1048,7 +1048,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
/* Check if adapter is up */
|
|
|
if (test_bit(STMMAC_DOWN, &priv->state))
|
|
|
-@@ -5797,10 +5799,12 @@ static irqreturn_t stmmac_msi_intr_tx(in
|
|
|
+@@ -5799,10 +5801,12 @@ static irqreturn_t stmmac_msi_intr_tx(in
|
|
|
static irqreturn_t stmmac_msi_intr_rx(int irq, void *data)
|
|
|
{
|
|
|
struct stmmac_rx_queue *rx_q = (struct stmmac_rx_queue *)data;
|
|
@@ -1062,7 +1062,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
/* Check if adapter is up */
|
|
|
if (test_bit(STMMAC_DOWN, &priv->state))
|
|
|
-@@ -5826,10 +5830,10 @@ static void stmmac_poll_controller(struc
|
|
|
+@@ -5828,10 +5832,10 @@ static void stmmac_poll_controller(struc
|
|
|
|
|
|
if (priv->plat->multi_msi_en) {
|
|
|
for (i = 0; i < priv->plat->rx_queues_to_use; i++)
|
|
@@ -1075,7 +1075,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
} else {
|
|
|
disable_irq(dev->irq);
|
|
|
stmmac_interrupt(dev->irq, dev);
|
|
|
-@@ -6010,34 +6014,34 @@ static int stmmac_rings_status_show(stru
|
|
|
+@@ -6012,34 +6016,34 @@ static int stmmac_rings_status_show(stru
|
|
|
return 0;
|
|
|
|
|
|
for (queue = 0; queue < rx_count; queue++) {
|
|
@@ -1116,7 +1116,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
}
|
|
|
}
|
|
|
|
|
|
-@@ -6384,7 +6388,7 @@ void stmmac_disable_rx_queue(struct stmm
|
|
|
+@@ -6386,7 +6390,7 @@ void stmmac_disable_rx_queue(struct stmm
|
|
|
|
|
|
void stmmac_enable_rx_queue(struct stmmac_priv *priv, u32 queue)
|
|
|
{
|
|
@@ -1125,7 +1125,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
struct stmmac_channel *ch = &priv->channel[queue];
|
|
|
unsigned long flags;
|
|
|
u32 buf_size;
|
|
|
-@@ -6421,7 +6425,7 @@ void stmmac_enable_rx_queue(struct stmma
|
|
|
+@@ -6423,7 +6427,7 @@ void stmmac_enable_rx_queue(struct stmma
|
|
|
rx_q->queue_index);
|
|
|
} else {
|
|
|
stmmac_set_dma_bfsize(priv, priv->ioaddr,
|
|
@@ -1134,7 +1134,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
rx_q->queue_index);
|
|
|
}
|
|
|
|
|
|
-@@ -6447,7 +6451,7 @@ void stmmac_disable_tx_queue(struct stmm
|
|
|
+@@ -6449,7 +6453,7 @@ void stmmac_disable_tx_queue(struct stmm
|
|
|
|
|
|
void stmmac_enable_tx_queue(struct stmmac_priv *priv, u32 queue)
|
|
|
{
|
|
@@ -1143,7 +1143,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
struct stmmac_channel *ch = &priv->channel[queue];
|
|
|
unsigned long flags;
|
|
|
int ret;
|
|
|
-@@ -6497,7 +6501,7 @@ void stmmac_xdp_release(struct net_devic
|
|
|
+@@ -6499,7 +6503,7 @@ void stmmac_xdp_release(struct net_devic
|
|
|
stmmac_disable_all_queues(priv);
|
|
|
|
|
|
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
|
|
@@ -1152,7 +1152,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
/* Free the IRQ lines */
|
|
|
stmmac_free_irq(dev, REQ_IRQ_ERR_ALL, 0);
|
|
|
-@@ -6556,7 +6560,7 @@ int stmmac_xdp_open(struct net_device *d
|
|
|
+@@ -6558,7 +6562,7 @@ int stmmac_xdp_open(struct net_device *d
|
|
|
|
|
|
/* DMA RX Channel Configuration */
|
|
|
for (chan = 0; chan < rx_cnt; chan++) {
|
|
@@ -1161,7 +1161,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
stmmac_init_rx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
|
|
|
rx_q->dma_rx_phy, chan);
|
|
|
-@@ -6574,7 +6578,7 @@ int stmmac_xdp_open(struct net_device *d
|
|
|
+@@ -6576,7 +6580,7 @@ int stmmac_xdp_open(struct net_device *d
|
|
|
rx_q->queue_index);
|
|
|
} else {
|
|
|
stmmac_set_dma_bfsize(priv, priv->ioaddr,
|
|
@@ -1170,7 +1170,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
rx_q->queue_index);
|
|
|
}
|
|
|
|
|
|
-@@ -6583,7 +6587,7 @@ int stmmac_xdp_open(struct net_device *d
|
|
|
+@@ -6585,7 +6589,7 @@ int stmmac_xdp_open(struct net_device *d
|
|
|
|
|
|
/* DMA TX Channel Configuration */
|
|
|
for (chan = 0; chan < tx_cnt; chan++) {
|
|
@@ -1179,7 +1179,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
|
|
|
tx_q->dma_tx_phy, chan);
|
|
|
-@@ -6616,7 +6620,7 @@ int stmmac_xdp_open(struct net_device *d
|
|
|
+@@ -6618,7 +6622,7 @@ int stmmac_xdp_open(struct net_device *d
|
|
|
|
|
|
irq_error:
|
|
|
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
|
|
@@ -1188,7 +1188,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
stmmac_hw_teardown(dev);
|
|
|
init_error:
|
|
|
-@@ -6643,8 +6647,8 @@ int stmmac_xsk_wakeup(struct net_device
|
|
|
+@@ -6645,8 +6649,8 @@ int stmmac_xsk_wakeup(struct net_device
|
|
|
queue >= priv->plat->tx_queues_to_use)
|
|
|
return -EINVAL;
|
|
|
|
|
@@ -1199,7 +1199,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
ch = &priv->channel[queue];
|
|
|
|
|
|
if (!rx_q->xsk_pool && !tx_q->xsk_pool)
|
|
|
-@@ -6904,8 +6908,8 @@ int stmmac_reinit_ringparam(struct net_d
|
|
|
+@@ -6906,8 +6910,8 @@ int stmmac_reinit_ringparam(struct net_d
|
|
|
if (netif_running(dev))
|
|
|
stmmac_release(dev);
|
|
|
|
|
@@ -1210,7 +1210,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
if (netif_running(dev))
|
|
|
ret = stmmac_open(dev);
|
|
|
-@@ -7343,7 +7347,7 @@ int stmmac_suspend(struct device *dev)
|
|
|
+@@ -7345,7 +7349,7 @@ int stmmac_suspend(struct device *dev)
|
|
|
stmmac_disable_all_queues(priv);
|
|
|
|
|
|
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
|
|
@@ -1219,7 +1219,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
if (priv->eee_enabled) {
|
|
|
priv->tx_path_in_lpi_mode = false;
|
|
|
-@@ -7395,7 +7399,7 @@ EXPORT_SYMBOL_GPL(stmmac_suspend);
|
|
|
+@@ -7397,7 +7401,7 @@ EXPORT_SYMBOL_GPL(stmmac_suspend);
|
|
|
|
|
|
static void stmmac_reset_rx_queue(struct stmmac_priv *priv, u32 queue)
|
|
|
{
|
|
@@ -1228,7 +1228,7 @@ Signed-off-by: Jakub Kicinski <[email protected]>
|
|
|
|
|
|
rx_q->cur_rx = 0;
|
|
|
rx_q->dirty_rx = 0;
|
|
|
-@@ -7403,7 +7407,7 @@ static void stmmac_reset_rx_queue(struct
|
|
|
+@@ -7405,7 +7409,7 @@ static void stmmac_reset_rx_queue(struct
|
|
|
|
|
|
static void stmmac_reset_tx_queue(struct stmmac_priv *priv, u32 queue)
|
|
|
{
|