|
1143 | 1143 | struct stmmac_channel *ch = &priv->channel[queue];
|
1144 | 1144 | unsigned long flags;
|
1145 | 1145 | int ret;
|
1146 |
| -@@ -6493,7 +6497,7 @@ void stmmac_xdp_release(struct net_devic |
| 1146 | +@@ -6496,7 +6500,7 @@ void stmmac_xdp_release(struct net_devic |
1147 | 1147 | stmmac_disable_all_queues(priv);
|
1148 | 1148 |
|
1149 | 1149 | for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
|
|
1152 | 1152 |
|
1153 | 1153 | /* Free the IRQ lines */
|
1154 | 1154 | stmmac_free_irq(dev, REQ_IRQ_ERR_ALL, 0);
|
1155 |
| -@@ -6552,7 +6556,7 @@ int stmmac_xdp_open(struct net_device *d |
| 1155 | +@@ -6555,7 +6559,7 @@ int stmmac_xdp_open(struct net_device *d |
1156 | 1156 |
|
1157 | 1157 | /* DMA RX Channel Configuration */
|
1158 | 1158 | for (chan = 0; chan < rx_cnt; chan++) {
|
|
1161 | 1161 |
|
1162 | 1162 | stmmac_init_rx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
|
1163 | 1163 | rx_q->dma_rx_phy, chan);
|
1164 |
| -@@ -6570,7 +6574,7 @@ int stmmac_xdp_open(struct net_device *d |
| 1164 | +@@ -6573,7 +6577,7 @@ int stmmac_xdp_open(struct net_device *d |
1165 | 1165 | rx_q->queue_index);
|
1166 | 1166 | } else {
|
1167 | 1167 | stmmac_set_dma_bfsize(priv, priv->ioaddr,
|
|
1170 | 1170 | rx_q->queue_index);
|
1171 | 1171 | }
|
1172 | 1172 |
|
1173 |
| -@@ -6579,7 +6583,7 @@ int stmmac_xdp_open(struct net_device *d |
| 1173 | +@@ -6582,7 +6586,7 @@ int stmmac_xdp_open(struct net_device *d |
1174 | 1174 |
|
1175 | 1175 | /* DMA TX Channel Configuration */
|
1176 | 1176 | for (chan = 0; chan < tx_cnt; chan++) {
|
|
1179 | 1179 |
|
1180 | 1180 | stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
|
1181 | 1181 | tx_q->dma_tx_phy, chan);
|
1182 |
| -@@ -6612,7 +6616,7 @@ int stmmac_xdp_open(struct net_device *d |
| 1182 | +@@ -6615,7 +6619,7 @@ int stmmac_xdp_open(struct net_device *d |
1183 | 1183 |
|
1184 | 1184 | irq_error:
|
1185 | 1185 | for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
|
|
1188 | 1188 |
|
1189 | 1189 | stmmac_hw_teardown(dev);
|
1190 | 1190 | init_error:
|
1191 |
| -@@ -6639,8 +6643,8 @@ int stmmac_xsk_wakeup(struct net_device |
| 1191 | +@@ -6642,8 +6646,8 @@ int stmmac_xsk_wakeup(struct net_device |
1192 | 1192 | queue >= priv->plat->tx_queues_to_use)
|
1193 | 1193 | return -EINVAL;
|
1194 | 1194 |
|
|
1199 | 1199 | ch = &priv->channel[queue];
|
1200 | 1200 |
|
1201 | 1201 | if (!rx_q->xsk_pool && !tx_q->xsk_pool)
|
1202 |
| -@@ -6896,8 +6900,8 @@ int stmmac_reinit_ringparam(struct net_d |
| 1202 | +@@ -6899,8 +6903,8 @@ int stmmac_reinit_ringparam(struct net_d |
1203 | 1203 | if (netif_running(dev))
|
1204 | 1204 | stmmac_release(dev);
|
1205 | 1205 |
|
|
1210 | 1210 |
|
1211 | 1211 | if (netif_running(dev))
|
1212 | 1212 | ret = stmmac_open(dev);
|
1213 |
| -@@ -7333,7 +7337,7 @@ int stmmac_suspend(struct device *dev) |
| 1213 | +@@ -7336,7 +7340,7 @@ int stmmac_suspend(struct device *dev) |
1214 | 1214 | stmmac_disable_all_queues(priv);
|
1215 | 1215 |
|
1216 | 1216 | for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
|
|
1219 | 1219 |
|
1220 | 1220 | if (priv->eee_enabled) {
|
1221 | 1221 | priv->tx_path_in_lpi_mode = false;
|
1222 |
| -@@ -7384,7 +7388,7 @@ EXPORT_SYMBOL_GPL(stmmac_suspend); |
| 1222 | +@@ -7387,7 +7391,7 @@ EXPORT_SYMBOL_GPL(stmmac_suspend); |
1223 | 1223 |
|
1224 | 1224 | static void stmmac_reset_rx_queue(struct stmmac_priv *priv, u32 queue)
|
1225 | 1225 | {
|
|
1228 | 1228 |
|
1229 | 1229 | rx_q->cur_rx = 0;
|
1230 | 1230 | rx_q->dirty_rx = 0;
|
1231 |
| -@@ -7392,7 +7396,7 @@ static void stmmac_reset_rx_queue(struct |
| 1231 | +@@ -7395,7 +7399,7 @@ static void stmmac_reset_rx_queue(struct |
1232 | 1232 |
|
1233 | 1233 | static void stmmac_reset_tx_queue(struct stmmac_priv *priv, u32 queue)
|
1234 | 1234 | {
|
|
0 commit comments