From nobody Wed Apr 15 07:02:46 2026 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 458ACC43334 for ; Sat, 23 Jul 2022 14:30:01 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S238134AbiGWO36 (ORCPT ); Sat, 23 Jul 2022 10:29:58 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:33414 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S235728AbiGWO3u (ORCPT ); Sat, 23 Jul 2022 10:29:50 -0400 Received: from mail-wr1-x42e.google.com (mail-wr1-x42e.google.com [IPv6:2a00:1450:4864:20::42e]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 213F819C19; Sat, 23 Jul 2022 07:29:49 -0700 (PDT) Received: by mail-wr1-x42e.google.com with SMTP id h8so10032153wrw.1; Sat, 23 Jul 2022 07:29:49 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20210112; h=from:to:cc:subject:date:message-id:in-reply-to:references :mime-version:content-transfer-encoding; bh=dtK+f+e69ktRtm1DfmKkH7OgqSKD4ay2ViO2yXXI+mw=; b=BRKTixbhnurH9pcfogJZP+9+9XbOYRMbxDufycEZFZc8tpLD+IfE7sDdgURmuvojSb mx6bjoYVKP5r8Y/ktV+gn4iSMhObNTo6afdTGZNOQgaKQi4DaXEonBi/YQf6qTyeHyCS c4gm8WFbaUMcHH5XF5X1rjnhpMynADfN1xdfxtCuBqLbyXEmojT7QQt6xz9i1dENaeut wqtCHIW/32q30tWK2DyAYSr+1VDV25kwnGHSmIdEwcpXHqMo1l5BEtpOJjrU/Qv9XTgY T42zHG0pqL6dT/GJa012PbeuB/yx+X685jDREvV2rslvY3b4bjKX9MnKiWcBwmzzHveU BUvA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references:mime-version:content-transfer-encoding; bh=dtK+f+e69ktRtm1DfmKkH7OgqSKD4ay2ViO2yXXI+mw=; b=ki4miqpKuOapnrdpjqUnAEq6+KU0jkVBGHdw/jw4iS+vDvLS5D6iNM7Ol0AjCe6Ueq WH6EsqlU20PbW/88lMLLuH/z/cUbhLizr8MnD7RUmSm2j1PdXtuCzYStq8TqsRsLPDQm y8g8h05sYMwBsXXAp1lqNH2hcoP1/W0+pr/bWukkNEEspu2avJe8scK6i7jP12Qnlcro 69INZvmcDLaR9FFvHDZqA+Mo1zyIsEY+CE66PwKfPiVYs127Z0YI4YRmeUofctMQxFWF n6aW7TZSMHxQATRp7nNy3szvvGGt/l/HoWSCzG0y4aTUtD4Jel9hunvbRVp6U2XacNBk /Klg== X-Gm-Message-State: AJIora8zT08CeWFEGSRCpxdSXdPBly7vCKDTtJeOvfUXR30mVG9PF/DG TUcpVUygEyxZ/Ly1ondvXRE= X-Google-Smtp-Source: AGRyM1vLn8trzceLunwqhaXT5AIv6dL8/i2Cw1MFuKkaH+23TDjqBDi7C+PEM5t9iwwpDbzI1v8hQQ== X-Received: by 2002:a05:6000:186:b0:21e:5874:2bfb with SMTP id p6-20020a056000018600b0021e58742bfbmr3016866wrx.510.1658586587614; Sat, 23 Jul 2022 07:29:47 -0700 (PDT) Received: from localhost.localdomain (host-87-7-207-127.retail.telecomitalia.it. [87.7.207.127]) by smtp.googlemail.com with ESMTPSA id q6-20020a1cf306000000b0039c5ab7167dsm11689717wmq.48.2022.07.23.07.29.44 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Sat, 23 Jul 2022 07:29:47 -0700 (PDT) From: Christian Marangi To: Giuseppe Cavallaro , Alexandre Torgue , Jose Abreu , "David S. Miller" , Eric Dumazet , Jakub Kicinski , Paolo Abeni , Maxime Coquelin , Russell King , netdev@vger.kernel.org, linux-stm32@st-md-mailman.stormreply.com, linux-arm-kernel@lists.infradead.org, linux-kernel@vger.kernel.org Cc: Christian Marangi Subject: [net-next PATCH v5 1/5] net: ethernet: stmicro: stmmac: move queue reset to dedicated functions Date: Sat, 23 Jul 2022 16:29:29 +0200 Message-Id: <20220723142933.16030-2-ansuelsmth@gmail.com> X-Mailer: git-send-email 2.36.1 In-Reply-To: <20220723142933.16030-1-ansuelsmth@gmail.com> References: <20220723142933.16030-1-ansuelsmth@gmail.com> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Type: text/plain; charset="utf-8" Move queue reset to dedicated functions. This aside from a simple cleanup is also required to allocate a dma conf without resetting the tx queue while the device is temporarily detached as now the reset is not part of the dma init function and can be done later in the code flow. Signed-off-by: Christian Marangi --- .../net/ethernet/stmicro/stmmac/stmmac_main.c | 59 ++++++++++--------- 1 file changed, 31 insertions(+), 28 deletions(-) diff --git a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c b/drivers/ne= t/ethernet/stmicro/stmmac/stmmac_main.c index 6f14b00c0b14..5578abb14949 100644 --- a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c +++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c @@ -130,6 +130,9 @@ static irqreturn_t stmmac_mac_interrupt(int irq, void *= dev_id); static irqreturn_t stmmac_safety_interrupt(int irq, void *dev_id); static irqreturn_t stmmac_msi_intr_tx(int irq, void *data); static irqreturn_t stmmac_msi_intr_rx(int irq, void *data); +static void stmmac_reset_rx_queue(struct stmmac_priv *priv, u32 queue); +static void stmmac_reset_tx_queue(struct stmmac_priv *priv, u32 queue); +static void stmmac_reset_queues_param(struct stmmac_priv *priv); static void stmmac_tx_timer_arm(struct stmmac_priv *priv, u32 queue); static void stmmac_flush_tx_descriptors(struct stmmac_priv *priv, int queu= e); static void stmmac_set_dma_operation_mode(struct stmmac_priv *priv, u32 tx= mode, @@ -1648,9 +1651,6 @@ static int __init_dma_rx_desc_rings(struct stmmac_pri= v *priv, u32 queue, gfp_t f return -ENOMEM; } =20 - rx_q->cur_rx =3D 0; - rx_q->dirty_rx =3D 0; - /* Setup the chained descriptor addresses */ if (priv->mode =3D=3D STMMAC_CHAIN_MODE) { if (priv->extend_desc) @@ -1753,12 +1753,6 @@ static int __init_dma_tx_desc_rings(struct stmmac_pr= iv *priv, u32 queue) tx_q->tx_skbuff[i] =3D NULL; } =20 - tx_q->dirty_tx =3D 0; - tx_q->cur_tx =3D 0; - tx_q->mss =3D 0; - - netdev_tx_reset_queue(netdev_get_tx_queue(priv->dev, queue)); - return 0; } =20 @@ -2644,10 +2638,7 @@ static void stmmac_tx_err(struct stmmac_priv *priv, = u32 chan) stmmac_stop_tx_dma(priv, chan); dma_free_tx_skbufs(priv, chan); stmmac_clear_tx_descriptors(priv, chan); - tx_q->dirty_tx =3D 0; - tx_q->cur_tx =3D 0; - tx_q->mss =3D 0; - netdev_tx_reset_queue(netdev_get_tx_queue(priv->dev, chan)); + stmmac_reset_tx_queue(priv, chan); stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg, tx_q->dma_tx_phy, chan); stmmac_start_tx_dma(priv, chan); @@ -3706,6 +3697,8 @@ static int stmmac_open(struct net_device *dev) goto init_error; } =20 + stmmac_reset_queues_param(priv); + ret =3D stmmac_hw_setup(dev, true); if (ret < 0) { netdev_err(priv->dev, "%s: Hw setup failed\n", __func__); @@ -6332,6 +6325,7 @@ void stmmac_enable_rx_queue(struct stmmac_priv *priv,= u32 queue) return; } =20 + stmmac_reset_rx_queue(priv, queue); stmmac_clear_rx_descriptors(priv, queue); =20 stmmac_init_rx_chan(priv, priv->ioaddr, priv->plat->dma_cfg, @@ -6393,6 +6387,7 @@ void stmmac_enable_tx_queue(struct stmmac_priv *priv,= u32 queue) return; } =20 + stmmac_reset_tx_queue(priv, queue); stmmac_clear_tx_descriptors(priv, queue); =20 stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg, @@ -7319,6 +7314,25 @@ int stmmac_suspend(struct device *dev) } EXPORT_SYMBOL_GPL(stmmac_suspend); =20 +static void stmmac_reset_rx_queue(struct stmmac_priv *priv, u32 queue) +{ + struct stmmac_rx_queue *rx_q =3D &priv->rx_queue[queue]; + + rx_q->cur_rx =3D 0; + rx_q->dirty_rx =3D 0; +} + +static void stmmac_reset_tx_queue(struct stmmac_priv *priv, u32 queue) +{ + struct stmmac_tx_queue *tx_q =3D &priv->tx_queue[queue]; + + tx_q->cur_tx =3D 0; + tx_q->dirty_tx =3D 0; + tx_q->mss =3D 0; + + netdev_tx_reset_queue(netdev_get_tx_queue(priv->dev, queue)); +} + /** * stmmac_reset_queues_param - reset queue parameters * @priv: device pointer @@ -7329,22 +7343,11 @@ static void stmmac_reset_queues_param(struct stmmac= _priv *priv) u32 tx_cnt =3D priv->plat->tx_queues_to_use; u32 queue; =20 - for (queue =3D 0; queue < rx_cnt; queue++) { - struct stmmac_rx_queue *rx_q =3D &priv->rx_queue[queue]; - - rx_q->cur_rx =3D 0; - rx_q->dirty_rx =3D 0; - } - - for (queue =3D 0; queue < tx_cnt; queue++) { - struct stmmac_tx_queue *tx_q =3D &priv->tx_queue[queue]; + for (queue =3D 0; queue < rx_cnt; queue++) + stmmac_reset_rx_queue(priv, queue); =20 - tx_q->cur_tx =3D 0; - tx_q->dirty_tx =3D 0; - tx_q->mss =3D 0; - - netdev_tx_reset_queue(netdev_get_tx_queue(priv->dev, queue)); - } + for (queue =3D 0; queue < tx_cnt; queue++) + stmmac_reset_tx_queue(priv, queue); } =20 /** --=20 2.36.1 From nobody Wed Apr 15 07:02:46 2026 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 2E350C43334 for ; Sat, 23 Jul 2022 14:30:04 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S238078AbiGWOaC (ORCPT ); Sat, 23 Jul 2022 10:30:02 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:33580 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S238014AbiGWO3z (ORCPT ); Sat, 23 Jul 2022 10:29:55 -0400 Received: from mail-wm1-x332.google.com (mail-wm1-x332.google.com [IPv6:2a00:1450:4864:20::332]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 5CF9B1BEBF; Sat, 23 Jul 2022 07:29:52 -0700 (PDT) Received: by mail-wm1-x332.google.com with SMTP id v67-20020a1cac46000000b003a1888b9d36so6766757wme.0; Sat, 23 Jul 2022 07:29:52 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20210112; h=from:to:cc:subject:date:message-id:in-reply-to:references :mime-version:content-transfer-encoding; bh=rjiJgm+ePUWk4K7ivlZ0MC/qqioPkcS003xuwMGC70A=; b=ixXDJxh8BydzdrlUp+5pUGg9ENcZKGe2KUprYBMZl2POERKwenf+RWBmrRBU+7gIcN WLWnVORyug8+zftCXTTwjU3z1OijT1bJouuZZdo5g7Fb0hODAOPGAGpTxMaDBaWsxVJ9 16SW4wUw0BASYukDBptpq07R+d/cHE1fTmUQAvANxHcpukXeTR70lz6lszlDXXtO9Iin woCuS7iso2EZKipkiL+zFeYXghQKpbDUGN7/bazdWzTs/f3xUL2eVmfC+MhZQf3B2bC5 XEU6Ofz4Rublo81gGXPxVRKSeWRMrdPO29YVc7RKI3JucyaKK3NWdj7RwR6J03iX3+ZV EKTw== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references:mime-version:content-transfer-encoding; bh=rjiJgm+ePUWk4K7ivlZ0MC/qqioPkcS003xuwMGC70A=; b=hd4duwQfvM3l3h0c0XPvcdxFHbw1ome3rCD2qAxJ2QfQOmT8tQKzXI8dIEoQMtsGb5 qy7gNtEto6v1TehkWlQc0GTEN3flvy3N51OIt9V6WMgkS3VX378+GX7ItIaZDL204VeC Ph9nPbnGasvTUVT0bPzX9wT+Y5RTnO58v1eUuq81j9Kru4BKWDrZ11rFnCXT0QmAJCuh hY2zSSXhrXWidymlldkeKpIbwzl9eX695ZhxMFF+p8hn4gBG3llL/T+m/lUdzokIiSCP O6vUXKPO+SdiDottzYj+G9v8eVV05Sj+6k7VyYQkzJJAjPGK2xQPmlyh4XWT6usixYeO 9MiQ== X-Gm-Message-State: AJIora/MkW7z2dhsBH9CEHF6Fr8S3NgbXGe/5aPpt8HrNKGy+oHFfxhK QWNj/PHNGzQo29RGxDSPtXA= X-Google-Smtp-Source: AGRyM1vVaFgb3GqjCNsvMbzlaBRd98PPg1L/7OqXzOm1rUViGRihb4msdECZSlhE1Jheg38qChktQg== X-Received: by 2002:a05:600c:4f49:b0:3a3:2717:27fc with SMTP id m9-20020a05600c4f4900b003a3271727fcmr15822873wmq.36.1658586590737; Sat, 23 Jul 2022 07:29:50 -0700 (PDT) Received: from localhost.localdomain (host-87-7-207-127.retail.telecomitalia.it. [87.7.207.127]) by smtp.googlemail.com with ESMTPSA id q6-20020a1cf306000000b0039c5ab7167dsm11689717wmq.48.2022.07.23.07.29.47 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Sat, 23 Jul 2022 07:29:50 -0700 (PDT) From: Christian Marangi To: Giuseppe Cavallaro , Alexandre Torgue , Jose Abreu , "David S. Miller" , Eric Dumazet , Jakub Kicinski , Paolo Abeni , Maxime Coquelin , Russell King , netdev@vger.kernel.org, linux-stm32@st-md-mailman.stormreply.com, linux-arm-kernel@lists.infradead.org, linux-kernel@vger.kernel.org Cc: Christian Marangi Subject: [net-next PATCH v5 2/5] net: ethernet: stmicro: stmmac: first disable all queues and disconnect in release Date: Sat, 23 Jul 2022 16:29:30 +0200 Message-Id: <20220723142933.16030-3-ansuelsmth@gmail.com> X-Mailer: git-send-email 2.36.1 In-Reply-To: <20220723142933.16030-1-ansuelsmth@gmail.com> References: <20220723142933.16030-1-ansuelsmth@gmail.com> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Type: text/plain; charset="utf-8" Disable all queues and disconnect before tx_disable in stmmac_release to prevent a corner case where packet may be still queued at the same time tx_disable is called resulting in kernel panic if some packet still has to be processed. Signed-off-by: Christian Marangi --- drivers/net/ethernet/stmicro/stmmac/stmmac_main.c | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c b/drivers/ne= t/ethernet/stmicro/stmmac/stmmac_main.c index 5578abb14949..96e55dadcfba 100644 --- a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c +++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c @@ -3758,8 +3758,6 @@ static int stmmac_release(struct net_device *dev) struct stmmac_priv *priv =3D netdev_priv(dev); u32 chan; =20 - netif_tx_disable(dev); - if (device_may_wakeup(priv->device)) phylink_speed_down(priv->phylink, false); /* Stop and disconnect the PHY */ @@ -3771,6 +3769,8 @@ static int stmmac_release(struct net_device *dev) for (chan =3D 0; chan < priv->plat->tx_queues_to_use; chan++) hrtimer_cancel(&priv->tx_queue[chan].txtimer); =20 + netif_tx_disable(dev); + /* Free the IRQ lines */ stmmac_free_irq(dev, REQ_IRQ_ERR_ALL, 0); =20 --=20 2.36.1 From nobody Wed Apr 15 07:02:46 2026 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 91AA9CCA483 for ; Sat, 23 Jul 2022 14:30:20 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S238356AbiGWOaS (ORCPT ); Sat, 23 Jul 2022 10:30:18 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:33634 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S238210AbiGWOaC (ORCPT ); Sat, 23 Jul 2022 10:30:02 -0400 Received: from mail-wr1-x42b.google.com (mail-wr1-x42b.google.com [IPv6:2a00:1450:4864:20::42b]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 5A9D81A063; Sat, 23 Jul 2022 07:29:56 -0700 (PDT) Received: by mail-wr1-x42b.google.com with SMTP id n12so10011880wrc.8; Sat, 23 Jul 2022 07:29:56 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20210112; h=from:to:cc:subject:date:message-id:in-reply-to:references :mime-version:content-transfer-encoding; bh=uUKbuZfmlvDhe/kIfDNrq8YP/dD8cuc75407ZRuXpG4=; b=R9Nt5gr7d2cJjDh9wgbPDPnBx3dve3w/2ccz8HfQtopEhFf4S6tvEIs4vtjcYkE+pn ZLywqh3C/AmuuYhV78bCttlN2oYtjtvdTdt6XDs5Ix/363rDMKytCGmuImLta1o6hNG4 /ERhe2Af3B8GMLq0VGbDQvh0bt+4l8PgB54N2vCSOWbaFM0XdfLgTjgtCTw54q94ibFQ B/A3g3u8qTOrKNrhoajLo3T50GM/dvmVmbcTgHRHIsuh41Kxlh3RWcbqKr6tq5XdE2mA np30zk8987EEnHXIN+q8GwDsQ0nWxjAQjFGcc5aXqpYGkWkehnq9ZLvncA9kxdfgBbrX 1mdA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references:mime-version:content-transfer-encoding; bh=uUKbuZfmlvDhe/kIfDNrq8YP/dD8cuc75407ZRuXpG4=; b=ugizA5Jf31LLLB23JB/ttlbNBDfqXoXIJ0i8Pt2tYI+l70y8q+8irObOVpWff7uUAw S5BJncKaIF8f6Z7NqovtTf3qnM2TrPmTENh18g94wBWN79XpzYlByu8syk5AKk7K02UO H4SV/Wo1+RRF82W8hFEhvdHlFTrSvJ5It1nLEifNnH0ayL7NZV4wusrJGpAiB4LeEoQS iihGUoGILHFxBjQauYMisr5FEmcChkzVSx36E+8+YfL1W2zTJ9CQV4KQTIoHr83ucIRB G16PGbqkqB3hzhhAVsPEQONqMnq3aA4X2jSjHWpYsVjxegifO9xPXeXPcCzcgmNpoHk1 a+Wg== X-Gm-Message-State: AJIora+xtqr37kBnYeRrGeSjSZAxpmETP1ipJLb82q1X5yjhNO5P6vsb 86u9TTzWbhFuAJvdBviVMrQ= X-Google-Smtp-Source: AGRyM1uguSquzYI1DOiq5G/Htu3ZUmgYldlD+imD8Won6YjMMH4UJI8mPn5lNxRhgJkRGpERu0i3Ig== X-Received: by 2002:a5d:498f:0:b0:21e:4074:8c49 with SMTP id r15-20020a5d498f000000b0021e40748c49mr3013474wrq.70.1658586594524; Sat, 23 Jul 2022 07:29:54 -0700 (PDT) Received: from localhost.localdomain (host-87-7-207-127.retail.telecomitalia.it. [87.7.207.127]) by smtp.googlemail.com with ESMTPSA id q6-20020a1cf306000000b0039c5ab7167dsm11689717wmq.48.2022.07.23.07.29.50 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Sat, 23 Jul 2022 07:29:54 -0700 (PDT) From: Christian Marangi To: Giuseppe Cavallaro , Alexandre Torgue , Jose Abreu , "David S. Miller" , Eric Dumazet , Jakub Kicinski , Paolo Abeni , Maxime Coquelin , Russell King , netdev@vger.kernel.org, linux-stm32@st-md-mailman.stormreply.com, linux-arm-kernel@lists.infradead.org, linux-kernel@vger.kernel.org Cc: Christian Marangi Subject: [net-next PATCH v5 3/5] net: ethernet: stmicro: stmmac: move dma conf to dedicated struct Date: Sat, 23 Jul 2022 16:29:31 +0200 Message-Id: <20220723142933.16030-4-ansuelsmth@gmail.com> X-Mailer: git-send-email 2.36.1 In-Reply-To: <20220723142933.16030-1-ansuelsmth@gmail.com> References: <20220723142933.16030-1-ansuelsmth@gmail.com> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Type: text/plain; charset="utf-8" Move dma buf conf to dedicated struct. This in preparation for code rework that will permit to allocate separate dma_conf without affecting the priv struct. Signed-off-by: Christian Marangi --- .../net/ethernet/stmicro/stmmac/chain_mode.c | 6 +- .../net/ethernet/stmicro/stmmac/ring_mode.c | 4 +- drivers/net/ethernet/stmicro/stmmac/stmmac.h | 21 +- .../ethernet/stmicro/stmmac/stmmac_ethtool.c | 4 +- .../net/ethernet/stmicro/stmmac/stmmac_main.c | 286 +++++++++--------- .../stmicro/stmmac/stmmac_selftests.c | 8 +- .../net/ethernet/stmicro/stmmac/stmmac_tc.c | 6 +- 7 files changed, 172 insertions(+), 163 deletions(-) diff --git a/drivers/net/ethernet/stmicro/stmmac/chain_mode.c b/drivers/net= /ethernet/stmicro/stmmac/chain_mode.c index d2cdc02d9f94..2e8744ac6b91 100644 --- a/drivers/net/ethernet/stmicro/stmmac/chain_mode.c +++ b/drivers/net/ethernet/stmicro/stmmac/chain_mode.c @@ -46,7 +46,7 @@ static int jumbo_frm(void *p, struct sk_buff *skb, int cs= um) =20 while (len !=3D 0) { tx_q->tx_skbuff[entry] =3D NULL; - entry =3D STMMAC_GET_ENTRY(entry, priv->dma_tx_size); + entry =3D STMMAC_GET_ENTRY(entry, priv->dma_conf.dma_tx_size); desc =3D tx_q->dma_tx + entry; =20 if (len > bmax) { @@ -137,7 +137,7 @@ static void refill_desc3(void *priv_ptr, struct dma_des= c *p) */ p->des3 =3D cpu_to_le32((unsigned int)(rx_q->dma_rx_phy + (((rx_q->dirty_rx) + 1) % - priv->dma_rx_size) * + priv->dma_conf.dma_rx_size) * sizeof(struct dma_desc))); } =20 @@ -155,7 +155,7 @@ static void clean_desc3(void *priv_ptr, struct dma_desc= *p) */ p->des3 =3D cpu_to_le32((unsigned int)((tx_q->dma_tx_phy + ((tx_q->dirty_tx + 1) % - priv->dma_tx_size)) + priv->dma_conf.dma_tx_size)) * sizeof(struct dma_desc))); } =20 diff --git a/drivers/net/ethernet/stmicro/stmmac/ring_mode.c b/drivers/net/= ethernet/stmicro/stmmac/ring_mode.c index 8ad900949dc8..2b5b17d8b8a0 100644 --- a/drivers/net/ethernet/stmicro/stmmac/ring_mode.c +++ b/drivers/net/ethernet/stmicro/stmmac/ring_mode.c @@ -51,7 +51,7 @@ static int jumbo_frm(void *p, struct sk_buff *skb, int cs= um) stmmac_prepare_tx_desc(priv, desc, 1, bmax, csum, STMMAC_RING_MODE, 0, false, skb->len); tx_q->tx_skbuff[entry] =3D NULL; - entry =3D STMMAC_GET_ENTRY(entry, priv->dma_tx_size); + entry =3D STMMAC_GET_ENTRY(entry, priv->dma_conf.dma_tx_size); =20 if (priv->extend_desc) desc =3D (struct dma_desc *)(tx_q->dma_etx + entry); @@ -107,7 +107,7 @@ static void refill_desc3(void *priv_ptr, struct dma_des= c *p) struct stmmac_priv *priv =3D rx_q->priv_data; =20 /* Fill DES3 in case of RING mode */ - if (priv->dma_buf_sz =3D=3D BUF_SIZE_16KiB) + if (priv->dma_conf.dma_buf_sz =3D=3D BUF_SIZE_16KiB) p->des3 =3D cpu_to_le32(le32_to_cpu(p->des2) + BUF_SIZE_8KiB); } =20 diff --git a/drivers/net/ethernet/stmicro/stmmac/stmmac.h b/drivers/net/eth= ernet/stmicro/stmmac/stmmac.h index 57970ae2178d..8ef44c9d84f4 100644 --- a/drivers/net/ethernet/stmicro/stmmac/stmmac.h +++ b/drivers/net/ethernet/stmicro/stmmac/stmmac.h @@ -188,6 +188,18 @@ struct stmmac_rfs_entry { int tc; }; =20 +struct stmmac_dma_conf { + unsigned int dma_buf_sz; + + /* RX Queue */ + struct stmmac_rx_queue rx_queue[MTL_MAX_RX_QUEUES]; + unsigned int dma_rx_size; + + /* TX Queue */ + struct stmmac_tx_queue tx_queue[MTL_MAX_TX_QUEUES]; + unsigned int dma_tx_size; +}; + struct stmmac_priv { /* Frequently used values are kept adjacent for cache effect */ u32 tx_coal_frames[MTL_MAX_TX_QUEUES]; @@ -201,7 +213,6 @@ struct stmmac_priv { int sph_cap; u32 sarc_type; =20 - unsigned int dma_buf_sz; unsigned int rx_copybreak; u32 rx_riwt[MTL_MAX_TX_QUEUES]; int hwts_rx_en; @@ -213,13 +224,7 @@ struct stmmac_priv { int (*hwif_quirks)(struct stmmac_priv *priv); struct mutex lock; =20 - /* RX Queue */ - struct stmmac_rx_queue rx_queue[MTL_MAX_RX_QUEUES]; - unsigned int dma_rx_size; - - /* TX Queue */ - struct stmmac_tx_queue tx_queue[MTL_MAX_TX_QUEUES]; - unsigned int dma_tx_size; + struct stmmac_dma_conf dma_conf; =20 /* Generic channel for NAPI */ struct stmmac_channel channel[STMMAC_CH_MAX]; diff --git a/drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c b/drivers= /net/ethernet/stmicro/stmmac/stmmac_ethtool.c index abfb3cd5958d..fdf5575aedb8 100644 --- a/drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c +++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c @@ -485,8 +485,8 @@ static void stmmac_get_ringparam(struct net_device *net= dev, =20 ring->rx_max_pending =3D DMA_MAX_RX_SIZE; ring->tx_max_pending =3D DMA_MAX_TX_SIZE; - ring->rx_pending =3D priv->dma_rx_size; - ring->tx_pending =3D priv->dma_tx_size; + ring->rx_pending =3D priv->dma_conf.dma_rx_size; + ring->tx_pending =3D priv->dma_conf.dma_tx_size; } =20 static int stmmac_set_ringparam(struct net_device *netdev, diff --git a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c b/drivers/ne= t/ethernet/stmicro/stmmac/stmmac_main.c index 96e55dadcfba..4d1a445baa07 100644 --- a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c +++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c @@ -74,8 +74,8 @@ static int phyaddr =3D -1; module_param(phyaddr, int, 0444); MODULE_PARM_DESC(phyaddr, "Physical device address"); =20 -#define STMMAC_TX_THRESH(x) ((x)->dma_tx_size / 4) -#define STMMAC_RX_THRESH(x) ((x)->dma_rx_size / 4) +#define STMMAC_TX_THRESH(x) ((x)->dma_conf.dma_tx_size / 4) +#define STMMAC_RX_THRESH(x) ((x)->dma_conf.dma_rx_size / 4) =20 /* Limit to make sure XDP TX and slow path can coexist */ #define STMMAC_XSK_TX_BUDGET_MAX 256 @@ -234,7 +234,7 @@ static void stmmac_disable_all_queues(struct stmmac_pri= v *priv) =20 /* synchronize_rcu() needed for pending XDP buffers to drain */ for (queue =3D 0; queue < rx_queues_cnt; queue++) { - rx_q =3D &priv->rx_queue[queue]; + rx_q =3D &priv->dma_conf.rx_queue[queue]; if (rx_q->xsk_pool) { synchronize_rcu(); break; @@ -360,13 +360,13 @@ static void print_pkt(unsigned char *buf, int len) =20 static inline u32 stmmac_tx_avail(struct stmmac_priv *priv, u32 queue) { - struct stmmac_tx_queue *tx_q =3D &priv->tx_queue[queue]; + struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[queue]; u32 avail; =20 if (tx_q->dirty_tx > tx_q->cur_tx) avail =3D tx_q->dirty_tx - tx_q->cur_tx - 1; else - avail =3D priv->dma_tx_size - tx_q->cur_tx + tx_q->dirty_tx - 1; + avail =3D priv->dma_conf.dma_tx_size - tx_q->cur_tx + tx_q->dirty_tx - 1; =20 return avail; } @@ -378,13 +378,13 @@ static inline u32 stmmac_tx_avail(struct stmmac_priv = *priv, u32 queue) */ static inline u32 stmmac_rx_dirty(struct stmmac_priv *priv, u32 queue) { - struct stmmac_rx_queue *rx_q =3D &priv->rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; u32 dirty; =20 if (rx_q->dirty_rx <=3D rx_q->cur_rx) dirty =3D rx_q->cur_rx - rx_q->dirty_rx; else - dirty =3D priv->dma_rx_size - rx_q->dirty_rx + rx_q->cur_rx; + dirty =3D priv->dma_conf.dma_rx_size - rx_q->dirty_rx + rx_q->cur_rx; =20 return dirty; } @@ -412,7 +412,7 @@ static int stmmac_enable_eee_mode(struct stmmac_priv *p= riv) =20 /* check if all TX queues have the work finished */ for (queue =3D 0; queue < tx_cnt; queue++) { - struct stmmac_tx_queue *tx_q =3D &priv->tx_queue[queue]; + struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[queue]; =20 if (tx_q->dirty_tx !=3D tx_q->cur_tx) return -EBUSY; /* still unfinished work */ @@ -1241,7 +1241,7 @@ static void stmmac_display_rx_rings(struct stmmac_pri= v *priv) =20 /* Display RX rings */ for (queue =3D 0; queue < rx_cnt; queue++) { - struct stmmac_rx_queue *rx_q =3D &priv->rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; =20 pr_info("\tRX Queue %u rings\n", queue); =20 @@ -1254,7 +1254,7 @@ static void stmmac_display_rx_rings(struct stmmac_pri= v *priv) } =20 /* Display RX ring */ - stmmac_display_ring(priv, head_rx, priv->dma_rx_size, true, + stmmac_display_ring(priv, head_rx, priv->dma_conf.dma_rx_size, true, rx_q->dma_rx_phy, desc_size); } } @@ -1268,7 +1268,7 @@ static void stmmac_display_tx_rings(struct stmmac_pri= v *priv) =20 /* Display TX rings */ for (queue =3D 0; queue < tx_cnt; queue++) { - struct stmmac_tx_queue *tx_q =3D &priv->tx_queue[queue]; + struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[queue]; =20 pr_info("\tTX Queue %d rings\n", queue); =20 @@ -1283,7 +1283,7 @@ static void stmmac_display_tx_rings(struct stmmac_pri= v *priv) desc_size =3D sizeof(struct dma_desc); } =20 - stmmac_display_ring(priv, head_tx, priv->dma_tx_size, false, + stmmac_display_ring(priv, head_tx, priv->dma_conf.dma_tx_size, false, tx_q->dma_tx_phy, desc_size); } } @@ -1324,21 +1324,21 @@ static int stmmac_set_bfsize(int mtu, int bufsize) */ static void stmmac_clear_rx_descriptors(struct stmmac_priv *priv, u32 queu= e) { - struct stmmac_rx_queue *rx_q =3D &priv->rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; int i; =20 /* Clear the RX descriptors */ - for (i =3D 0; i < priv->dma_rx_size; i++) + for (i =3D 0; i < priv->dma_conf.dma_rx_size; i++) if (priv->extend_desc) stmmac_init_rx_desc(priv, &rx_q->dma_erx[i].basic, priv->use_riwt, priv->mode, - (i =3D=3D priv->dma_rx_size - 1), - priv->dma_buf_sz); + (i =3D=3D priv->dma_conf.dma_rx_size - 1), + priv->dma_conf.dma_buf_sz); else stmmac_init_rx_desc(priv, &rx_q->dma_rx[i], priv->use_riwt, priv->mode, - (i =3D=3D priv->dma_rx_size - 1), - priv->dma_buf_sz); + (i =3D=3D priv->dma_conf.dma_rx_size - 1), + priv->dma_conf.dma_buf_sz); } =20 /** @@ -1350,12 +1350,12 @@ static void stmmac_clear_rx_descriptors(struct stmm= ac_priv *priv, u32 queue) */ static void stmmac_clear_tx_descriptors(struct stmmac_priv *priv, u32 queu= e) { - struct stmmac_tx_queue *tx_q =3D &priv->tx_queue[queue]; + struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[queue]; int i; =20 /* Clear the TX descriptors */ - for (i =3D 0; i < priv->dma_tx_size; i++) { - int last =3D (i =3D=3D (priv->dma_tx_size - 1)); + for (i =3D 0; i < priv->dma_conf.dma_tx_size; i++) { + int last =3D (i =3D=3D (priv->dma_conf.dma_tx_size - 1)); struct dma_desc *p; =20 if (priv->extend_desc) @@ -1403,7 +1403,7 @@ static void stmmac_clear_descriptors(struct stmmac_pr= iv *priv) static int stmmac_init_rx_buffers(struct stmmac_priv *priv, struct dma_des= c *p, int i, gfp_t flags, u32 queue) { - struct stmmac_rx_queue *rx_q =3D &priv->rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; struct stmmac_rx_buffer *buf =3D &rx_q->buf_pool[i]; gfp_t gfp =3D (GFP_ATOMIC | __GFP_NOWARN); =20 @@ -1432,7 +1432,7 @@ static int stmmac_init_rx_buffers(struct stmmac_priv = *priv, struct dma_desc *p, buf->addr =3D page_pool_get_dma_addr(buf->page) + buf->page_offset; =20 stmmac_set_desc_addr(priv, p, buf->addr); - if (priv->dma_buf_sz =3D=3D BUF_SIZE_16KiB) + if (priv->dma_conf.dma_buf_sz =3D=3D BUF_SIZE_16KiB) stmmac_init_desc3(priv, p); =20 return 0; @@ -1446,7 +1446,7 @@ static int stmmac_init_rx_buffers(struct stmmac_priv = *priv, struct dma_desc *p, */ static void stmmac_free_rx_buffer(struct stmmac_priv *priv, u32 queue, int= i) { - struct stmmac_rx_queue *rx_q =3D &priv->rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; struct stmmac_rx_buffer *buf =3D &rx_q->buf_pool[i]; =20 if (buf->page) @@ -1466,7 +1466,7 @@ static void stmmac_free_rx_buffer(struct stmmac_priv = *priv, u32 queue, int i) */ static void stmmac_free_tx_buffer(struct stmmac_priv *priv, u32 queue, int= i) { - struct stmmac_tx_queue *tx_q =3D &priv->tx_queue[queue]; + struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[queue]; =20 if (tx_q->tx_skbuff_dma[i].buf && tx_q->tx_skbuff_dma[i].buf_type !=3D STMMAC_TXBUF_T_XDP_TX) { @@ -1511,17 +1511,17 @@ static void dma_free_rx_skbufs(struct stmmac_priv *= priv, u32 queue) { int i; =20 - for (i =3D 0; i < priv->dma_rx_size; i++) + for (i =3D 0; i < priv->dma_conf.dma_rx_size; i++) stmmac_free_rx_buffer(priv, queue, i); } =20 static int stmmac_alloc_rx_buffers(struct stmmac_priv *priv, u32 queue, gfp_t flags) { - struct stmmac_rx_queue *rx_q =3D &priv->rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; int i; =20 - for (i =3D 0; i < priv->dma_rx_size; i++) { + for (i =3D 0; i < priv->dma_conf.dma_rx_size; i++) { struct dma_desc *p; int ret; =20 @@ -1548,10 +1548,10 @@ static int stmmac_alloc_rx_buffers(struct stmmac_pr= iv *priv, u32 queue, */ static void dma_free_rx_xskbufs(struct stmmac_priv *priv, u32 queue) { - struct stmmac_rx_queue *rx_q =3D &priv->rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; int i; =20 - for (i =3D 0; i < priv->dma_rx_size; i++) { + for (i =3D 0; i < priv->dma_conf.dma_rx_size; i++) { struct stmmac_rx_buffer *buf =3D &rx_q->buf_pool[i]; =20 if (!buf->xdp) @@ -1564,10 +1564,10 @@ static void dma_free_rx_xskbufs(struct stmmac_priv = *priv, u32 queue) =20 static int stmmac_alloc_rx_buffers_zc(struct stmmac_priv *priv, u32 queue) { - struct stmmac_rx_queue *rx_q =3D &priv->rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; int i; =20 - for (i =3D 0; i < priv->dma_rx_size; i++) { + for (i =3D 0; i < priv->dma_conf.dma_rx_size; i++) { struct stmmac_rx_buffer *buf; dma_addr_t dma_addr; struct dma_desc *p; @@ -1610,7 +1610,7 @@ static struct xsk_buff_pool *stmmac_get_xsk_pool(stru= ct stmmac_priv *priv, u32 q */ static int __init_dma_rx_desc_rings(struct stmmac_priv *priv, u32 queue, g= fp_t flags) { - struct stmmac_rx_queue *rx_q =3D &priv->rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; int ret; =20 netif_dbg(priv, probe, priv->dev, @@ -1656,11 +1656,11 @@ static int __init_dma_rx_desc_rings(struct stmmac_p= riv *priv, u32 queue, gfp_t f if (priv->extend_desc) stmmac_mode_init(priv, rx_q->dma_erx, rx_q->dma_rx_phy, - priv->dma_rx_size, 1); + priv->dma_conf.dma_rx_size, 1); else stmmac_mode_init(priv, rx_q->dma_rx, rx_q->dma_rx_phy, - priv->dma_rx_size, 0); + priv->dma_conf.dma_rx_size, 0); } =20 return 0; @@ -1687,7 +1687,7 @@ static int init_dma_rx_desc_rings(struct net_device *= dev, gfp_t flags) =20 err_init_rx_buffers: while (queue >=3D 0) { - struct stmmac_rx_queue *rx_q =3D &priv->rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; =20 if (rx_q->xsk_pool) dma_free_rx_xskbufs(priv, queue); @@ -1713,7 +1713,7 @@ static int init_dma_rx_desc_rings(struct net_device *= dev, gfp_t flags) */ static int __init_dma_tx_desc_rings(struct stmmac_priv *priv, u32 queue) { - struct stmmac_tx_queue *tx_q =3D &priv->tx_queue[queue]; + struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[queue]; int i; =20 netif_dbg(priv, probe, priv->dev, @@ -1725,16 +1725,16 @@ static int __init_dma_tx_desc_rings(struct stmmac_p= riv *priv, u32 queue) if (priv->extend_desc) stmmac_mode_init(priv, tx_q->dma_etx, tx_q->dma_tx_phy, - priv->dma_tx_size, 1); + priv->dma_conf.dma_tx_size, 1); else if (!(tx_q->tbs & STMMAC_TBS_AVAIL)) stmmac_mode_init(priv, tx_q->dma_tx, tx_q->dma_tx_phy, - priv->dma_tx_size, 0); + priv->dma_conf.dma_tx_size, 0); } =20 tx_q->xsk_pool =3D stmmac_get_xsk_pool(priv, queue); =20 - for (i =3D 0; i < priv->dma_tx_size; i++) { + for (i =3D 0; i < priv->dma_conf.dma_tx_size; i++) { struct dma_desc *p; =20 if (priv->extend_desc) @@ -1804,12 +1804,12 @@ static int init_dma_desc_rings(struct net_device *d= ev, gfp_t flags) */ static void dma_free_tx_skbufs(struct stmmac_priv *priv, u32 queue) { - struct stmmac_tx_queue *tx_q =3D &priv->tx_queue[queue]; + struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[queue]; int i; =20 tx_q->xsk_frames_done =3D 0; =20 - for (i =3D 0; i < priv->dma_tx_size; i++) + for (i =3D 0; i < priv->dma_conf.dma_tx_size; i++) stmmac_free_tx_buffer(priv, queue, i); =20 if (tx_q->xsk_pool && tx_q->xsk_frames_done) { @@ -1839,7 +1839,7 @@ static void stmmac_free_tx_skbufs(struct stmmac_priv = *priv) */ static void __free_dma_rx_desc_resources(struct stmmac_priv *priv, u32 que= ue) { - struct stmmac_rx_queue *rx_q =3D &priv->rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; =20 /* Release the DMA RX socket buffers */ if (rx_q->xsk_pool) @@ -1852,11 +1852,11 @@ static void __free_dma_rx_desc_resources(struct stm= mac_priv *priv, u32 queue) =20 /* Free DMA regions of consistent memory previously allocated */ if (!priv->extend_desc) - dma_free_coherent(priv->device, priv->dma_rx_size * + dma_free_coherent(priv->device, priv->dma_conf.dma_rx_size * sizeof(struct dma_desc), rx_q->dma_rx, rx_q->dma_rx_phy); else - dma_free_coherent(priv->device, priv->dma_rx_size * + dma_free_coherent(priv->device, priv->dma_conf.dma_rx_size * sizeof(struct dma_extended_desc), rx_q->dma_erx, rx_q->dma_rx_phy); =20 @@ -1885,7 +1885,7 @@ static void free_dma_rx_desc_resources(struct stmmac_= priv *priv) */ static void __free_dma_tx_desc_resources(struct stmmac_priv *priv, u32 que= ue) { - struct stmmac_tx_queue *tx_q =3D &priv->tx_queue[queue]; + struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[queue]; size_t size; void *addr; =20 @@ -1903,7 +1903,7 @@ static void __free_dma_tx_desc_resources(struct stmma= c_priv *priv, u32 queue) addr =3D tx_q->dma_tx; } =20 - size *=3D priv->dma_tx_size; + size *=3D priv->dma_conf.dma_tx_size; =20 dma_free_coherent(priv->device, size, addr, tx_q->dma_tx_phy); =20 @@ -1932,7 +1932,7 @@ static void free_dma_tx_desc_resources(struct stmmac_= priv *priv) */ static int __alloc_dma_rx_desc_resources(struct stmmac_priv *priv, u32 que= ue) { - struct stmmac_rx_queue *rx_q =3D &priv->rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; struct stmmac_channel *ch =3D &priv->channel[queue]; bool xdp_prog =3D stmmac_xdp_is_enabled(priv); struct page_pool_params pp_params =3D { 0 }; @@ -1944,8 +1944,8 @@ static int __alloc_dma_rx_desc_resources(struct stmma= c_priv *priv, u32 queue) rx_q->priv_data =3D priv; =20 pp_params.flags =3D PP_FLAG_DMA_MAP | PP_FLAG_DMA_SYNC_DEV; - pp_params.pool_size =3D priv->dma_rx_size; - num_pages =3D DIV_ROUND_UP(priv->dma_buf_sz, PAGE_SIZE); + pp_params.pool_size =3D priv->dma_conf.dma_rx_size; + num_pages =3D DIV_ROUND_UP(priv->dma_conf.dma_buf_sz, PAGE_SIZE); pp_params.order =3D ilog2(num_pages); pp_params.nid =3D dev_to_node(priv->device); pp_params.dev =3D priv->device; @@ -1960,7 +1960,7 @@ static int __alloc_dma_rx_desc_resources(struct stmma= c_priv *priv, u32 queue) return ret; } =20 - rx_q->buf_pool =3D kcalloc(priv->dma_rx_size, + rx_q->buf_pool =3D kcalloc(priv->dma_conf.dma_rx_size, sizeof(*rx_q->buf_pool), GFP_KERNEL); if (!rx_q->buf_pool) @@ -1968,7 +1968,7 @@ static int __alloc_dma_rx_desc_resources(struct stmma= c_priv *priv, u32 queue) =20 if (priv->extend_desc) { rx_q->dma_erx =3D dma_alloc_coherent(priv->device, - priv->dma_rx_size * + priv->dma_conf.dma_rx_size * sizeof(struct dma_extended_desc), &rx_q->dma_rx_phy, GFP_KERNEL); @@ -1977,7 +1977,7 @@ static int __alloc_dma_rx_desc_resources(struct stmma= c_priv *priv, u32 queue) =20 } else { rx_q->dma_rx =3D dma_alloc_coherent(priv->device, - priv->dma_rx_size * + priv->dma_conf.dma_rx_size * sizeof(struct dma_desc), &rx_q->dma_rx_phy, GFP_KERNEL); @@ -2034,20 +2034,20 @@ static int alloc_dma_rx_desc_resources(struct stmma= c_priv *priv) */ static int __alloc_dma_tx_desc_resources(struct stmmac_priv *priv, u32 que= ue) { - struct stmmac_tx_queue *tx_q =3D &priv->tx_queue[queue]; + struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[queue]; size_t size; void *addr; =20 tx_q->queue_index =3D queue; tx_q->priv_data =3D priv; =20 - tx_q->tx_skbuff_dma =3D kcalloc(priv->dma_tx_size, + tx_q->tx_skbuff_dma =3D kcalloc(priv->dma_conf.dma_tx_size, sizeof(*tx_q->tx_skbuff_dma), GFP_KERNEL); if (!tx_q->tx_skbuff_dma) return -ENOMEM; =20 - tx_q->tx_skbuff =3D kcalloc(priv->dma_tx_size, + tx_q->tx_skbuff =3D kcalloc(priv->dma_conf.dma_tx_size, sizeof(struct sk_buff *), GFP_KERNEL); if (!tx_q->tx_skbuff) @@ -2060,7 +2060,7 @@ static int __alloc_dma_tx_desc_resources(struct stmma= c_priv *priv, u32 queue) else size =3D sizeof(struct dma_desc); =20 - size *=3D priv->dma_tx_size; + size *=3D priv->dma_conf.dma_tx_size; =20 addr =3D dma_alloc_coherent(priv->device, size, &tx_q->dma_tx_phy, GFP_KERNEL); @@ -2304,7 +2304,7 @@ static void stmmac_dma_operation_mode(struct stmmac_p= riv *priv) =20 /* configure all channels */ for (chan =3D 0; chan < rx_channels_count; chan++) { - struct stmmac_rx_queue *rx_q =3D &priv->rx_queue[chan]; + struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[chan]; u32 buf_size; =20 qmode =3D priv->plat->rx_queues_cfg[chan].mode_to_use; @@ -2319,7 +2319,7 @@ static void stmmac_dma_operation_mode(struct stmmac_p= riv *priv) chan); } else { stmmac_set_dma_bfsize(priv, priv->ioaddr, - priv->dma_buf_sz, + priv->dma_conf.dma_buf_sz, chan); } } @@ -2335,7 +2335,7 @@ static void stmmac_dma_operation_mode(struct stmmac_p= riv *priv) static bool stmmac_xdp_xmit_zc(struct stmmac_priv *priv, u32 queue, u32 bu= dget) { struct netdev_queue *nq =3D netdev_get_tx_queue(priv->dev, queue); - struct stmmac_tx_queue *tx_q =3D &priv->tx_queue[queue]; + struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[queue]; struct xsk_buff_pool *pool =3D tx_q->xsk_pool; unsigned int entry =3D tx_q->cur_tx; struct dma_desc *tx_desc =3D NULL; @@ -2410,7 +2410,7 @@ static bool stmmac_xdp_xmit_zc(struct stmmac_priv *pr= iv, u32 queue, u32 budget) =20 stmmac_enable_dma_transmission(priv, priv->ioaddr); =20 - tx_q->cur_tx =3D STMMAC_GET_ENTRY(tx_q->cur_tx, priv->dma_tx_size); + tx_q->cur_tx =3D STMMAC_GET_ENTRY(tx_q->cur_tx, priv->dma_conf.dma_tx_si= ze); entry =3D tx_q->cur_tx; } =20 @@ -2451,7 +2451,7 @@ static void stmmac_bump_dma_threshold(struct stmmac_p= riv *priv, u32 chan) */ static int stmmac_tx_clean(struct stmmac_priv *priv, int budget, u32 queue) { - struct stmmac_tx_queue *tx_q =3D &priv->tx_queue[queue]; + struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[queue]; unsigned int bytes_compl =3D 0, pkts_compl =3D 0; unsigned int entry, xmits =3D 0, count =3D 0; =20 @@ -2464,7 +2464,7 @@ static int stmmac_tx_clean(struct stmmac_priv *priv, = int budget, u32 queue) entry =3D tx_q->dirty_tx; =20 /* Try to clean all TX complete frame in 1 shot */ - while ((entry !=3D tx_q->cur_tx) && count < priv->dma_tx_size) { + while ((entry !=3D tx_q->cur_tx) && count < priv->dma_conf.dma_tx_size) { struct xdp_frame *xdpf; struct sk_buff *skb; struct dma_desc *p; @@ -2566,7 +2566,7 @@ static int stmmac_tx_clean(struct stmmac_priv *priv, = int budget, u32 queue) =20 stmmac_release_tx_desc(priv, p, priv->mode); =20 - entry =3D STMMAC_GET_ENTRY(entry, priv->dma_tx_size); + entry =3D STMMAC_GET_ENTRY(entry, priv->dma_conf.dma_tx_size); } tx_q->dirty_tx =3D entry; =20 @@ -2631,7 +2631,7 @@ static int stmmac_tx_clean(struct stmmac_priv *priv, = int budget, u32 queue) */ static void stmmac_tx_err(struct stmmac_priv *priv, u32 chan) { - struct stmmac_tx_queue *tx_q =3D &priv->tx_queue[chan]; + struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[chan]; =20 netif_tx_stop_queue(netdev_get_tx_queue(priv->dev, chan)); =20 @@ -2698,8 +2698,8 @@ static int stmmac_napi_check(struct stmmac_priv *priv= , u32 chan, u32 dir) { int status =3D stmmac_dma_interrupt_status(priv, priv->ioaddr, &priv->xstats, chan, dir); - struct stmmac_rx_queue *rx_q =3D &priv->rx_queue[chan]; - struct stmmac_tx_queue *tx_q =3D &priv->tx_queue[chan]; + struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[chan]; + struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[chan]; struct stmmac_channel *ch =3D &priv->channel[chan]; struct napi_struct *rx_napi; struct napi_struct *tx_napi; @@ -2865,7 +2865,7 @@ static int stmmac_init_dma_engine(struct stmmac_priv = *priv) =20 /* DMA RX Channel Configuration */ for (chan =3D 0; chan < rx_channels_count; chan++) { - rx_q =3D &priv->rx_queue[chan]; + rx_q =3D &priv->dma_conf.rx_queue[chan]; =20 stmmac_init_rx_chan(priv, priv->ioaddr, priv->plat->dma_cfg, rx_q->dma_rx_phy, chan); @@ -2879,7 +2879,7 @@ static int stmmac_init_dma_engine(struct stmmac_priv = *priv) =20 /* DMA TX Channel Configuration */ for (chan =3D 0; chan < tx_channels_count; chan++) { - tx_q =3D &priv->tx_queue[chan]; + tx_q =3D &priv->dma_conf.tx_queue[chan]; =20 stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg, tx_q->dma_tx_phy, chan); @@ -2894,7 +2894,7 @@ static int stmmac_init_dma_engine(struct stmmac_priv = *priv) =20 static void stmmac_tx_timer_arm(struct stmmac_priv *priv, u32 queue) { - struct stmmac_tx_queue *tx_q =3D &priv->tx_queue[queue]; + struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[queue]; =20 hrtimer_start(&tx_q->txtimer, STMMAC_COAL_TIMER(priv->tx_coal_timer[queue]), @@ -2944,7 +2944,7 @@ static void stmmac_init_coalesce(struct stmmac_priv *= priv) u32 chan; =20 for (chan =3D 0; chan < tx_channel_count; chan++) { - struct stmmac_tx_queue *tx_q =3D &priv->tx_queue[chan]; + struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[chan]; =20 priv->tx_coal_frames[chan] =3D STMMAC_TX_FRAMES; priv->tx_coal_timer[chan] =3D STMMAC_COAL_TX_TIMER; @@ -2966,12 +2966,12 @@ static void stmmac_set_rings_length(struct stmmac_p= riv *priv) /* set TX ring length */ for (chan =3D 0; chan < tx_channels_count; chan++) stmmac_set_tx_ring_len(priv, priv->ioaddr, - (priv->dma_tx_size - 1), chan); + (priv->dma_conf.dma_tx_size - 1), chan); =20 /* set RX ring length */ for (chan =3D 0; chan < rx_channels_count; chan++) stmmac_set_rx_ring_len(priv, priv->ioaddr, - (priv->dma_rx_size - 1), chan); + (priv->dma_conf.dma_rx_size - 1), chan); } =20 /** @@ -3298,7 +3298,7 @@ static int stmmac_hw_setup(struct net_device *dev, bo= ol ptp_register) /* Enable TSO */ if (priv->tso) { for (chan =3D 0; chan < tx_cnt; chan++) { - struct stmmac_tx_queue *tx_q =3D &priv->tx_queue[chan]; + struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[chan]; =20 /* TSO and TBS cannot co-exist */ if (tx_q->tbs & STMMAC_TBS_AVAIL) @@ -3320,7 +3320,7 @@ static int stmmac_hw_setup(struct net_device *dev, bo= ol ptp_register) =20 /* TBS */ for (chan =3D 0; chan < tx_cnt; chan++) { - struct stmmac_tx_queue *tx_q =3D &priv->tx_queue[chan]; + struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[chan]; int enable =3D tx_q->tbs & STMMAC_TBS_AVAIL; =20 stmmac_enable_tbs(priv, priv->ioaddr, enable, chan); @@ -3364,7 +3364,7 @@ static void stmmac_free_irq(struct net_device *dev, for (j =3D irq_idx - 1; j >=3D 0; j--) { if (priv->tx_irq[j] > 0) { irq_set_affinity_hint(priv->tx_irq[j], NULL); - free_irq(priv->tx_irq[j], &priv->tx_queue[j]); + free_irq(priv->tx_irq[j], &priv->dma_conf.tx_queue[j]); } } irq_idx =3D priv->plat->rx_queues_to_use; @@ -3373,7 +3373,7 @@ static void stmmac_free_irq(struct net_device *dev, for (j =3D irq_idx - 1; j >=3D 0; j--) { if (priv->rx_irq[j] > 0) { irq_set_affinity_hint(priv->rx_irq[j], NULL); - free_irq(priv->rx_irq[j], &priv->rx_queue[j]); + free_irq(priv->rx_irq[j], &priv->dma_conf.rx_queue[j]); } } =20 @@ -3508,7 +3508,7 @@ static int stmmac_request_irq_multi_msi(struct net_de= vice *dev) sprintf(int_name, "%s:%s-%d", dev->name, "rx", i); ret =3D request_irq(priv->rx_irq[i], stmmac_msi_intr_rx, - 0, int_name, &priv->rx_queue[i]); + 0, int_name, &priv->dma_conf.rx_queue[i]); if (unlikely(ret < 0)) { netdev_err(priv->dev, "%s: alloc rx-%d MSI %d (error: %d)\n", @@ -3533,7 +3533,7 @@ static int stmmac_request_irq_multi_msi(struct net_de= vice *dev) sprintf(int_name, "%s:%s-%d", dev->name, "tx", i); ret =3D request_irq(priv->tx_irq[i], stmmac_msi_intr_tx, - 0, int_name, &priv->tx_queue[i]); + 0, int_name, &priv->dma_conf.tx_queue[i]); if (unlikely(ret < 0)) { netdev_err(priv->dev, "%s: alloc tx-%d MSI %d (error: %d)\n", @@ -3662,21 +3662,21 @@ static int stmmac_open(struct net_device *dev) bfsize =3D 0; =20 if (bfsize < BUF_SIZE_16KiB) - bfsize =3D stmmac_set_bfsize(dev->mtu, priv->dma_buf_sz); + bfsize =3D stmmac_set_bfsize(dev->mtu, priv->dma_conf.dma_buf_sz); =20 - priv->dma_buf_sz =3D bfsize; + priv->dma_conf.dma_buf_sz =3D bfsize; buf_sz =3D bfsize; =20 priv->rx_copybreak =3D STMMAC_RX_COPYBREAK; =20 - if (!priv->dma_tx_size) - priv->dma_tx_size =3D DMA_DEFAULT_TX_SIZE; - if (!priv->dma_rx_size) - priv->dma_rx_size =3D DMA_DEFAULT_RX_SIZE; + if (!priv->dma_conf.dma_tx_size) + priv->dma_conf.dma_tx_size =3D DMA_DEFAULT_TX_SIZE; + if (!priv->dma_conf.dma_rx_size) + priv->dma_conf.dma_rx_size =3D DMA_DEFAULT_RX_SIZE; =20 /* Earlier check for TBS */ for (chan =3D 0; chan < priv->plat->tx_queues_to_use; chan++) { - struct stmmac_tx_queue *tx_q =3D &priv->tx_queue[chan]; + struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[chan]; int tbs_en =3D priv->plat->tx_queues_cfg[chan].tbs_en; =20 /* Setup per-TXQ tbs flag before TX descriptor alloc */ @@ -3725,7 +3725,7 @@ static int stmmac_open(struct net_device *dev) phylink_stop(priv->phylink); =20 for (chan =3D 0; chan < priv->plat->tx_queues_to_use; chan++) - hrtimer_cancel(&priv->tx_queue[chan].txtimer); + hrtimer_cancel(&priv->dma_conf.tx_queue[chan].txtimer); =20 stmmac_hw_teardown(dev); init_error: @@ -3767,7 +3767,7 @@ static int stmmac_release(struct net_device *dev) stmmac_disable_all_queues(priv); =20 for (chan =3D 0; chan < priv->plat->tx_queues_to_use; chan++) - hrtimer_cancel(&priv->tx_queue[chan].txtimer); + hrtimer_cancel(&priv->dma_conf.tx_queue[chan].txtimer); =20 netif_tx_disable(dev); =20 @@ -3827,7 +3827,7 @@ static bool stmmac_vlan_insert(struct stmmac_priv *pr= iv, struct sk_buff *skb, return false; =20 stmmac_set_tx_owner(priv, p); - tx_q->cur_tx =3D STMMAC_GET_ENTRY(tx_q->cur_tx, priv->dma_tx_size); + tx_q->cur_tx =3D STMMAC_GET_ENTRY(tx_q->cur_tx, priv->dma_conf.dma_tx_siz= e); return true; } =20 @@ -3845,7 +3845,7 @@ static bool stmmac_vlan_insert(struct stmmac_priv *pr= iv, struct sk_buff *skb, static void stmmac_tso_allocator(struct stmmac_priv *priv, dma_addr_t des, int total_len, bool last_segment, u32 queue) { - struct stmmac_tx_queue *tx_q =3D &priv->tx_queue[queue]; + struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[queue]; struct dma_desc *desc; u32 buff_size; int tmp_len; @@ -3856,7 +3856,7 @@ static void stmmac_tso_allocator(struct stmmac_priv *= priv, dma_addr_t des, dma_addr_t curr_addr; =20 tx_q->cur_tx =3D STMMAC_GET_ENTRY(tx_q->cur_tx, - priv->dma_tx_size); + priv->dma_conf.dma_tx_size); WARN_ON(tx_q->tx_skbuff[tx_q->cur_tx]); =20 if (tx_q->tbs & STMMAC_TBS_AVAIL) @@ -3884,7 +3884,7 @@ static void stmmac_tso_allocator(struct stmmac_priv *= priv, dma_addr_t des, =20 static void stmmac_flush_tx_descriptors(struct stmmac_priv *priv, int queu= e) { - struct stmmac_tx_queue *tx_q =3D &priv->tx_queue[queue]; + struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[queue]; int desc_size; =20 if (likely(priv->extend_desc)) @@ -3946,7 +3946,7 @@ static netdev_tx_t stmmac_tso_xmit(struct sk_buff *sk= b, struct net_device *dev) dma_addr_t des; int i; =20 - tx_q =3D &priv->tx_queue[queue]; + tx_q =3D &priv->dma_conf.tx_queue[queue]; first_tx =3D tx_q->cur_tx; =20 /* Compute header lengths */ @@ -3986,7 +3986,7 @@ static netdev_tx_t stmmac_tso_xmit(struct sk_buff *sk= b, struct net_device *dev) stmmac_set_mss(priv, mss_desc, mss); tx_q->mss =3D mss; tx_q->cur_tx =3D STMMAC_GET_ENTRY(tx_q->cur_tx, - priv->dma_tx_size); + priv->dma_conf.dma_tx_size); WARN_ON(tx_q->tx_skbuff[tx_q->cur_tx]); } =20 @@ -4098,7 +4098,7 @@ static netdev_tx_t stmmac_tso_xmit(struct sk_buff *sk= b, struct net_device *dev) * ndo_start_xmit will fill this descriptor the next time it's * called and stmmac_tx_clean may clean up to this descriptor. */ - tx_q->cur_tx =3D STMMAC_GET_ENTRY(tx_q->cur_tx, priv->dma_tx_size); + tx_q->cur_tx =3D STMMAC_GET_ENTRY(tx_q->cur_tx, priv->dma_conf.dma_tx_siz= e); =20 if (unlikely(stmmac_tx_avail(priv, queue) <=3D (MAX_SKB_FRAGS + 1))) { netif_dbg(priv, hw, priv->dev, "%s: stop transmitted packets\n", @@ -4186,7 +4186,7 @@ static netdev_tx_t stmmac_xmit(struct sk_buff *skb, s= truct net_device *dev) int entry, first_tx; dma_addr_t des; =20 - tx_q =3D &priv->tx_queue[queue]; + tx_q =3D &priv->dma_conf.tx_queue[queue]; first_tx =3D tx_q->cur_tx; =20 if (priv->tx_path_in_lpi_mode && priv->eee_sw_timer_en) @@ -4249,7 +4249,7 @@ static netdev_tx_t stmmac_xmit(struct sk_buff *skb, s= truct net_device *dev) int len =3D skb_frag_size(frag); bool last_segment =3D (i =3D=3D (nfrags - 1)); =20 - entry =3D STMMAC_GET_ENTRY(entry, priv->dma_tx_size); + entry =3D STMMAC_GET_ENTRY(entry, priv->dma_conf.dma_tx_size); WARN_ON(tx_q->tx_skbuff[entry]); =20 if (likely(priv->extend_desc)) @@ -4320,7 +4320,7 @@ static netdev_tx_t stmmac_xmit(struct sk_buff *skb, s= truct net_device *dev) * ndo_start_xmit will fill this descriptor the next time it's * called and stmmac_tx_clean may clean up to this descriptor. */ - entry =3D STMMAC_GET_ENTRY(entry, priv->dma_tx_size); + entry =3D STMMAC_GET_ENTRY(entry, priv->dma_conf.dma_tx_size); tx_q->cur_tx =3D entry; =20 if (netif_msg_pktdata(priv)) { @@ -4435,7 +4435,7 @@ static void stmmac_rx_vlan(struct net_device *dev, st= ruct sk_buff *skb) */ static inline void stmmac_rx_refill(struct stmmac_priv *priv, u32 queue) { - struct stmmac_rx_queue *rx_q =3D &priv->rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; int dirty =3D stmmac_rx_dirty(priv, queue); unsigned int entry =3D rx_q->dirty_rx; gfp_t gfp =3D (GFP_ATOMIC | __GFP_NOWARN); @@ -4489,7 +4489,7 @@ static inline void stmmac_rx_refill(struct stmmac_pri= v *priv, u32 queue) dma_wmb(); stmmac_set_rx_owner(priv, p, use_rx_wd); =20 - entry =3D STMMAC_GET_ENTRY(entry, priv->dma_rx_size); + entry =3D STMMAC_GET_ENTRY(entry, priv->dma_conf.dma_rx_size); } rx_q->dirty_rx =3D entry; rx_q->rx_tail_addr =3D rx_q->dma_rx_phy + @@ -4517,12 +4517,12 @@ static unsigned int stmmac_rx_buf1_len(struct stmma= c_priv *priv, =20 /* First descriptor, not last descriptor and not split header */ if (status & rx_not_ls) - return priv->dma_buf_sz; + return priv->dma_conf.dma_buf_sz; =20 plen =3D stmmac_get_rx_frame_len(priv, p, coe); =20 /* First descriptor and last descriptor and not split header */ - return min_t(unsigned int, priv->dma_buf_sz, plen); + return min_t(unsigned int, priv->dma_conf.dma_buf_sz, plen); } =20 static unsigned int stmmac_rx_buf2_len(struct stmmac_priv *priv, @@ -4538,7 +4538,7 @@ static unsigned int stmmac_rx_buf2_len(struct stmmac_= priv *priv, =20 /* Not last descriptor */ if (status & rx_not_ls) - return priv->dma_buf_sz; + return priv->dma_conf.dma_buf_sz; =20 plen =3D stmmac_get_rx_frame_len(priv, p, coe); =20 @@ -4549,7 +4549,7 @@ static unsigned int stmmac_rx_buf2_len(struct stmmac_= priv *priv, static int stmmac_xdp_xmit_xdpf(struct stmmac_priv *priv, int queue, struct xdp_frame *xdpf, bool dma_map) { - struct stmmac_tx_queue *tx_q =3D &priv->tx_queue[queue]; + struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[queue]; unsigned int entry =3D tx_q->cur_tx; struct dma_desc *tx_desc; dma_addr_t dma_addr; @@ -4612,7 +4612,7 @@ static int stmmac_xdp_xmit_xdpf(struct stmmac_priv *p= riv, int queue, =20 stmmac_enable_dma_transmission(priv, priv->ioaddr); =20 - entry =3D STMMAC_GET_ENTRY(entry, priv->dma_tx_size); + entry =3D STMMAC_GET_ENTRY(entry, priv->dma_conf.dma_tx_size); tx_q->cur_tx =3D entry; =20 return STMMAC_XDP_TX; @@ -4786,7 +4786,7 @@ static void stmmac_dispatch_skb_zc(struct stmmac_priv= *priv, u32 queue, =20 static bool stmmac_rx_refill_zc(struct stmmac_priv *priv, u32 queue, u32 b= udget) { - struct stmmac_rx_queue *rx_q =3D &priv->rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; unsigned int entry =3D rx_q->dirty_rx; struct dma_desc *rx_desc =3D NULL; bool ret =3D true; @@ -4829,7 +4829,7 @@ static bool stmmac_rx_refill_zc(struct stmmac_priv *p= riv, u32 queue, u32 budget) dma_wmb(); stmmac_set_rx_owner(priv, rx_desc, use_rx_wd); =20 - entry =3D STMMAC_GET_ENTRY(entry, priv->dma_rx_size); + entry =3D STMMAC_GET_ENTRY(entry, priv->dma_conf.dma_rx_size); } =20 if (rx_desc) { @@ -4844,7 +4844,7 @@ static bool stmmac_rx_refill_zc(struct stmmac_priv *p= riv, u32 queue, u32 budget) =20 static int stmmac_rx_zc(struct stmmac_priv *priv, int limit, u32 queue) { - struct stmmac_rx_queue *rx_q =3D &priv->rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; unsigned int count =3D 0, error =3D 0, len =3D 0; int dirty =3D stmmac_rx_dirty(priv, queue); unsigned int next_entry =3D rx_q->cur_rx; @@ -4866,7 +4866,7 @@ static int stmmac_rx_zc(struct stmmac_priv *priv, int= limit, u32 queue) desc_size =3D sizeof(struct dma_desc); } =20 - stmmac_display_ring(priv, rx_head, priv->dma_rx_size, true, + stmmac_display_ring(priv, rx_head, priv->dma_conf.dma_rx_size, true, rx_q->dma_rx_phy, desc_size); } while (count < limit) { @@ -4913,7 +4913,7 @@ static int stmmac_rx_zc(struct stmmac_priv *priv, int= limit, u32 queue) =20 /* Prefetch the next RX descriptor */ rx_q->cur_rx =3D STMMAC_GET_ENTRY(rx_q->cur_rx, - priv->dma_rx_size); + priv->dma_conf.dma_rx_size); next_entry =3D rx_q->cur_rx; =20 if (priv->extend_desc) @@ -5034,7 +5034,7 @@ static int stmmac_rx_zc(struct stmmac_priv *priv, int= limit, u32 queue) */ static int stmmac_rx(struct stmmac_priv *priv, int limit, u32 queue) { - struct stmmac_rx_queue *rx_q =3D &priv->rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; struct stmmac_channel *ch =3D &priv->channel[queue]; unsigned int count =3D 0, error =3D 0, len =3D 0; int status =3D 0, coe =3D priv->hw->rx_csum; @@ -5047,7 +5047,7 @@ static int stmmac_rx(struct stmmac_priv *priv, int li= mit, u32 queue) int buf_sz; =20 dma_dir =3D page_pool_get_dma_dir(rx_q->page_pool); - buf_sz =3D DIV_ROUND_UP(priv->dma_buf_sz, PAGE_SIZE) * PAGE_SIZE; + buf_sz =3D DIV_ROUND_UP(priv->dma_conf.dma_buf_sz, PAGE_SIZE) * PAGE_SIZE; =20 if (netif_msg_rx_status(priv)) { void *rx_head; @@ -5061,7 +5061,7 @@ static int stmmac_rx(struct stmmac_priv *priv, int li= mit, u32 queue) desc_size =3D sizeof(struct dma_desc); } =20 - stmmac_display_ring(priv, rx_head, priv->dma_rx_size, true, + stmmac_display_ring(priv, rx_head, priv->dma_conf.dma_rx_size, true, rx_q->dma_rx_phy, desc_size); } while (count < limit) { @@ -5105,7 +5105,7 @@ static int stmmac_rx(struct stmmac_priv *priv, int li= mit, u32 queue) break; =20 rx_q->cur_rx =3D STMMAC_GET_ENTRY(rx_q->cur_rx, - priv->dma_rx_size); + priv->dma_conf.dma_rx_size); next_entry =3D rx_q->cur_rx; =20 if (priv->extend_desc) @@ -5240,7 +5240,7 @@ static int stmmac_rx(struct stmmac_priv *priv, int li= mit, u32 queue) buf1_len, dma_dir); skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, buf->page, buf->page_offset, buf1_len, - priv->dma_buf_sz); + priv->dma_conf.dma_buf_sz); =20 /* Data payload appended into SKB */ page_pool_release_page(rx_q->page_pool, buf->page); @@ -5252,7 +5252,7 @@ static int stmmac_rx(struct stmmac_priv *priv, int li= mit, u32 queue) buf2_len, dma_dir); skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, buf->sec_page, 0, buf2_len, - priv->dma_buf_sz); + priv->dma_conf.dma_buf_sz); =20 /* Data payload appended into SKB */ page_pool_release_page(rx_q->page_pool, buf->sec_page); @@ -5694,11 +5694,13 @@ static irqreturn_t stmmac_safety_interrupt(int irq,= void *dev_id) static irqreturn_t stmmac_msi_intr_tx(int irq, void *data) { struct stmmac_tx_queue *tx_q =3D (struct stmmac_tx_queue *)data; + struct stmmac_dma_conf *dma_conf; int chan =3D tx_q->queue_index; struct stmmac_priv *priv; int status; =20 - priv =3D container_of(tx_q, struct stmmac_priv, tx_queue[chan]); + dma_conf =3D container_of(tx_q, struct stmmac_dma_conf, tx_queue[chan]); + priv =3D container_of(dma_conf, struct stmmac_priv, dma_conf); =20 if (unlikely(!data)) { netdev_err(priv->dev, "%s: invalid dev pointer\n", __func__); @@ -5724,10 +5726,12 @@ static irqreturn_t stmmac_msi_intr_tx(int irq, void= *data) static irqreturn_t stmmac_msi_intr_rx(int irq, void *data) { struct stmmac_rx_queue *rx_q =3D (struct stmmac_rx_queue *)data; + struct stmmac_dma_conf *dma_conf; int chan =3D rx_q->queue_index; struct stmmac_priv *priv; =20 - priv =3D container_of(rx_q, struct stmmac_priv, rx_queue[chan]); + dma_conf =3D container_of(rx_q, struct stmmac_dma_conf, rx_queue[chan]); + priv =3D container_of(dma_conf, struct stmmac_priv, dma_conf); =20 if (unlikely(!data)) { netdev_err(priv->dev, "%s: invalid dev pointer\n", __func__); @@ -5758,10 +5762,10 @@ static void stmmac_poll_controller(struct net_devic= e *dev) =20 if (priv->plat->multi_msi_en) { for (i =3D 0; i < priv->plat->rx_queues_to_use; i++) - stmmac_msi_intr_rx(0, &priv->rx_queue[i]); + stmmac_msi_intr_rx(0, &priv->dma_conf.rx_queue[i]); =20 for (i =3D 0; i < priv->plat->tx_queues_to_use; i++) - stmmac_msi_intr_tx(0, &priv->tx_queue[i]); + stmmac_msi_intr_tx(0, &priv->dma_conf.tx_queue[i]); } else { disable_irq(dev->irq); stmmac_interrupt(dev->irq, dev); @@ -5940,34 +5944,34 @@ static int stmmac_rings_status_show(struct seq_file= *seq, void *v) return 0; =20 for (queue =3D 0; queue < rx_count; queue++) { - struct stmmac_rx_queue *rx_q =3D &priv->rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; =20 seq_printf(seq, "RX Queue %d:\n", queue); =20 if (priv->extend_desc) { seq_printf(seq, "Extended descriptor ring:\n"); sysfs_display_ring((void *)rx_q->dma_erx, - priv->dma_rx_size, 1, seq, rx_q->dma_rx_phy); + priv->dma_conf.dma_rx_size, 1, seq, rx_q->dma_rx_phy); } else { seq_printf(seq, "Descriptor ring:\n"); sysfs_display_ring((void *)rx_q->dma_rx, - priv->dma_rx_size, 0, seq, rx_q->dma_rx_phy); + priv->dma_conf.dma_rx_size, 0, seq, rx_q->dma_rx_phy); } } =20 for (queue =3D 0; queue < tx_count; queue++) { - struct stmmac_tx_queue *tx_q =3D &priv->tx_queue[queue]; + struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[queue]; =20 seq_printf(seq, "TX Queue %d:\n", queue); =20 if (priv->extend_desc) { seq_printf(seq, "Extended descriptor ring:\n"); sysfs_display_ring((void *)tx_q->dma_etx, - priv->dma_tx_size, 1, seq, tx_q->dma_tx_phy); + priv->dma_conf.dma_tx_size, 1, seq, tx_q->dma_tx_phy); } else if (!(tx_q->tbs & STMMAC_TBS_AVAIL)) { seq_printf(seq, "Descriptor ring:\n"); sysfs_display_ring((void *)tx_q->dma_tx, - priv->dma_tx_size, 0, seq, tx_q->dma_tx_phy); + priv->dma_conf.dma_tx_size, 0, seq, tx_q->dma_tx_phy); } } =20 @@ -6306,7 +6310,7 @@ void stmmac_disable_rx_queue(struct stmmac_priv *priv= , u32 queue) =20 void stmmac_enable_rx_queue(struct stmmac_priv *priv, u32 queue) { - struct stmmac_rx_queue *rx_q =3D &priv->rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; struct stmmac_channel *ch =3D &priv->channel[queue]; unsigned long flags; u32 buf_size; @@ -6343,7 +6347,7 @@ void stmmac_enable_rx_queue(struct stmmac_priv *priv,= u32 queue) rx_q->queue_index); } else { stmmac_set_dma_bfsize(priv, priv->ioaddr, - priv->dma_buf_sz, + priv->dma_conf.dma_buf_sz, rx_q->queue_index); } =20 @@ -6369,7 +6373,7 @@ void stmmac_disable_tx_queue(struct stmmac_priv *priv= , u32 queue) =20 void stmmac_enable_tx_queue(struct stmmac_priv *priv, u32 queue) { - struct stmmac_tx_queue *tx_q =3D &priv->tx_queue[queue]; + struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[queue]; struct stmmac_channel *ch =3D &priv->channel[queue]; unsigned long flags; int ret; @@ -6416,7 +6420,7 @@ void stmmac_xdp_release(struct net_device *dev) stmmac_disable_all_queues(priv); =20 for (chan =3D 0; chan < priv->plat->tx_queues_to_use; chan++) - hrtimer_cancel(&priv->tx_queue[chan].txtimer); + hrtimer_cancel(&priv->dma_conf.tx_queue[chan].txtimer); =20 /* Free the IRQ lines */ stmmac_free_irq(dev, REQ_IRQ_ERR_ALL, 0); @@ -6475,7 +6479,7 @@ int stmmac_xdp_open(struct net_device *dev) =20 /* DMA RX Channel Configuration */ for (chan =3D 0; chan < rx_cnt; chan++) { - rx_q =3D &priv->rx_queue[chan]; + rx_q =3D &priv->dma_conf.rx_queue[chan]; =20 stmmac_init_rx_chan(priv, priv->ioaddr, priv->plat->dma_cfg, rx_q->dma_rx_phy, chan); @@ -6493,7 +6497,7 @@ int stmmac_xdp_open(struct net_device *dev) rx_q->queue_index); } else { stmmac_set_dma_bfsize(priv, priv->ioaddr, - priv->dma_buf_sz, + priv->dma_conf.dma_buf_sz, rx_q->queue_index); } =20 @@ -6502,7 +6506,7 @@ int stmmac_xdp_open(struct net_device *dev) =20 /* DMA TX Channel Configuration */ for (chan =3D 0; chan < tx_cnt; chan++) { - tx_q =3D &priv->tx_queue[chan]; + tx_q =3D &priv->dma_conf.tx_queue[chan]; =20 stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg, tx_q->dma_tx_phy, chan); @@ -6535,7 +6539,7 @@ int stmmac_xdp_open(struct net_device *dev) =20 irq_error: for (chan =3D 0; chan < priv->plat->tx_queues_to_use; chan++) - hrtimer_cancel(&priv->tx_queue[chan].txtimer); + hrtimer_cancel(&priv->dma_conf.tx_queue[chan].txtimer); =20 stmmac_hw_teardown(dev); init_error: @@ -6562,8 +6566,8 @@ int stmmac_xsk_wakeup(struct net_device *dev, u32 que= ue, u32 flags) queue >=3D priv->plat->tx_queues_to_use) return -EINVAL; =20 - rx_q =3D &priv->rx_queue[queue]; - tx_q =3D &priv->tx_queue[queue]; + rx_q =3D &priv->dma_conf.rx_queue[queue]; + tx_q =3D &priv->dma_conf.tx_queue[queue]; ch =3D &priv->channel[queue]; =20 if (!rx_q->xsk_pool && !tx_q->xsk_pool) @@ -6818,8 +6822,8 @@ int stmmac_reinit_ringparam(struct net_device *dev, u= 32 rx_size, u32 tx_size) if (netif_running(dev)) stmmac_release(dev); =20 - priv->dma_rx_size =3D rx_size; - priv->dma_tx_size =3D tx_size; + priv->dma_conf.dma_rx_size =3D rx_size; + priv->dma_conf.dma_tx_size =3D tx_size; =20 if (netif_running(dev)) ret =3D stmmac_open(dev); @@ -7265,7 +7269,7 @@ int stmmac_suspend(struct device *dev) stmmac_disable_all_queues(priv); =20 for (chan =3D 0; chan < priv->plat->tx_queues_to_use; chan++) - hrtimer_cancel(&priv->tx_queue[chan].txtimer); + hrtimer_cancel(&priv->dma_conf.tx_queue[chan].txtimer); =20 if (priv->eee_enabled) { priv->tx_path_in_lpi_mode =3D false; @@ -7316,7 +7320,7 @@ EXPORT_SYMBOL_GPL(stmmac_suspend); =20 static void stmmac_reset_rx_queue(struct stmmac_priv *priv, u32 queue) { - struct stmmac_rx_queue *rx_q =3D &priv->rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; =20 rx_q->cur_rx =3D 0; rx_q->dirty_rx =3D 0; @@ -7324,7 +7328,7 @@ static void stmmac_reset_rx_queue(struct stmmac_priv = *priv, u32 queue) =20 static void stmmac_reset_tx_queue(struct stmmac_priv *priv, u32 queue) { - struct stmmac_tx_queue *tx_q =3D &priv->tx_queue[queue]; + struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[queue]; =20 tx_q->cur_tx =3D 0; tx_q->dirty_tx =3D 0; diff --git a/drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c b/drive= rs/net/ethernet/stmicro/stmmac/stmmac_selftests.c index 2fc51dc5eb0b..49af7e78b7f5 100644 --- a/drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c +++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c @@ -795,8 +795,8 @@ static int stmmac_test_flowctrl(struct stmmac_priv *pri= v) struct stmmac_channel *ch =3D &priv->channel[i]; u32 tail; =20 - tail =3D priv->rx_queue[i].dma_rx_phy + - (priv->dma_rx_size * sizeof(struct dma_desc)); + tail =3D priv->dma_conf.rx_queue[i].dma_rx_phy + + (priv->dma_conf.dma_rx_size * sizeof(struct dma_desc)); =20 stmmac_set_rx_tail_ptr(priv, priv->ioaddr, tail, i); stmmac_start_rx(priv, priv->ioaddr, i); @@ -1680,7 +1680,7 @@ static int stmmac_test_arpoffload(struct stmmac_priv = *priv) static int __stmmac_test_jumbo(struct stmmac_priv *priv, u16 queue) { struct stmmac_packet_attrs attr =3D { }; - int size =3D priv->dma_buf_sz; + int size =3D priv->dma_conf.dma_buf_sz; =20 attr.dst =3D priv->dev->dev_addr; attr.max_size =3D size - ETH_FCS_LEN; @@ -1763,7 +1763,7 @@ static int stmmac_test_tbs(struct stmmac_priv *priv) =20 /* Find first TBS enabled Queue, if any */ for (i =3D 0; i < priv->plat->tx_queues_to_use; i++) - if (priv->tx_queue[i].tbs & STMMAC_TBS_AVAIL) + if (priv->dma_conf.tx_queue[i].tbs & STMMAC_TBS_AVAIL) break; =20 if (i >=3D priv->plat->tx_queues_to_use) diff --git a/drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c b/drivers/net/= ethernet/stmicro/stmmac/stmmac_tc.c index d61766eeac6d..773e415cc2de 100644 --- a/drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c +++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c @@ -1091,13 +1091,13 @@ static int tc_setup_etf(struct stmmac_priv *priv, return -EOPNOTSUPP; if (qopt->queue >=3D priv->plat->tx_queues_to_use) return -EINVAL; - if (!(priv->tx_queue[qopt->queue].tbs & STMMAC_TBS_AVAIL)) + if (!(priv->dma_conf.tx_queue[qopt->queue].tbs & STMMAC_TBS_AVAIL)) return -EINVAL; =20 if (qopt->enable) - priv->tx_queue[qopt->queue].tbs |=3D STMMAC_TBS_EN; + priv->dma_conf.tx_queue[qopt->queue].tbs |=3D STMMAC_TBS_EN; else - priv->tx_queue[qopt->queue].tbs &=3D ~STMMAC_TBS_EN; + priv->dma_conf.tx_queue[qopt->queue].tbs &=3D ~STMMAC_TBS_EN; =20 netdev_info(priv->dev, "%s ETF for Queue %d\n", qopt->enable ? "enabled" : "disabled", qopt->queue); --=20 2.36.1 From nobody Wed Apr 15 07:02:46 2026 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 2C998C433EF for ; Sat, 23 Jul 2022 14:30:27 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S238244AbiGWOaY (ORCPT ); Sat, 23 Jul 2022 10:30:24 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:33560 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S237990AbiGWOaG (ORCPT ); Sat, 23 Jul 2022 10:30:06 -0400 Received: from mail-wm1-x32e.google.com (mail-wm1-x32e.google.com [IPv6:2a00:1450:4864:20::32e]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id A5F201CB0B; Sat, 23 Jul 2022 07:29:59 -0700 (PDT) Received: by mail-wm1-x32e.google.com with SMTP id b6so4280013wmq.5; Sat, 23 Jul 2022 07:29:59 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20210112; h=from:to:cc:subject:date:message-id:in-reply-to:references :mime-version:content-transfer-encoding; bh=ADzuSRgNV+lYAfSPNQC46tH2sv/EHvScw/QSutAFZWM=; b=HCue1zpXBVKy/GffG3VhgV6QuAL9k5kh1npU8rjfQ55n0Iape56gbWbdTyPaOOO6e8 V1+oap8TNfEDau8Wq322Gj0az3jokgBQkl6Ow85kjBI+yDuTFzFy35RL4C0gE3XLTUbu bAHhTzpGxUH77GiWdOpgtX5dpH7Ih+YqwY36hi9wxpVudhtmseij6KFOtyG7q40Vnl9z h5/BNjTZtoN1WaiMfVHASjNC5lO7z5BPYQ7rnU6dM/EupNzuQ6Rfsn/1Spji++AazUQH IHaChMTHLyPa2NOvAWG1zqADbGuV7INcelAzKAuDZZ+wY8ZRlnKfvsy2V31vYuTVm5tE BJjg== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references:mime-version:content-transfer-encoding; bh=ADzuSRgNV+lYAfSPNQC46tH2sv/EHvScw/QSutAFZWM=; b=RN/XDST0WralmtqxvXHG6t5McEP9SXapv8QTelM/6BjnwiN8K/AfsLfcjoPwxFxmLT DhoaLGCCKRv+s3iZcxJsExC1ywyc+qfNiKKiOKvzM21PlGk5fN486s6Ifu4mgVVKALUj VNWBSnH33je0nLQVR8AtN/Tf1IZyi8IHRrTVxSFc1KauY1VXp+Q+I9s/mfEydBkQBPxv s3QNLGJbVlGvZ0CFbsDtu8Nab1BSxjzULQgtywdkcXk2+nui2p17eDJnU4QX01r5Gn4M mWoSR2Y2Yk/Bji8e8HhqBu+b4ltp5vQi6foSuTiNvv9T9EPsD35GnGejEkKw4e0e/BJK 6H9g== X-Gm-Message-State: AJIora9eyLK8oIDuUg4gvdpdfx0JJyE4vmCkZUBhPu4VXYnOrpA/WdsC D1f0cSu7jan2bk49z8VElt0= X-Google-Smtp-Source: AGRyM1tCs+800F8fopOIcyJvlUkjLEUjVO5dypIdiBrB8uRSkkGX62rw7PoGl9pvyDTr3iA33PWFqw== X-Received: by 2002:a05:600c:1554:b0:3a3:2e32:8a7d with SMTP id f20-20020a05600c155400b003a32e328a7dmr11595593wmg.70.1658586597935; Sat, 23 Jul 2022 07:29:57 -0700 (PDT) Received: from localhost.localdomain (host-87-7-207-127.retail.telecomitalia.it. [87.7.207.127]) by smtp.googlemail.com with ESMTPSA id q6-20020a1cf306000000b0039c5ab7167dsm11689717wmq.48.2022.07.23.07.29.54 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Sat, 23 Jul 2022 07:29:57 -0700 (PDT) From: Christian Marangi To: Giuseppe Cavallaro , Alexandre Torgue , Jose Abreu , "David S. Miller" , Eric Dumazet , Jakub Kicinski , Paolo Abeni , Maxime Coquelin , Russell King , netdev@vger.kernel.org, linux-stm32@st-md-mailman.stormreply.com, linux-arm-kernel@lists.infradead.org, linux-kernel@vger.kernel.org Cc: Christian Marangi Subject: [net-next PATCH v5 4/5] net: ethernet: stmicro: stmmac: generate stmmac dma conf before open Date: Sat, 23 Jul 2022 16:29:32 +0200 Message-Id: <20220723142933.16030-5-ansuelsmth@gmail.com> X-Mailer: git-send-email 2.36.1 In-Reply-To: <20220723142933.16030-1-ansuelsmth@gmail.com> References: <20220723142933.16030-1-ansuelsmth@gmail.com> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Type: text/plain; charset="utf-8" Rework the driver to generate the stmmac dma_conf before stmmac_open. This permits a function to first check if it's possible to allocate a new dma_config and then pass it directly to __stmmac_open and "open" the interface with the new configuration. Signed-off-by: Christian Marangi --- .../net/ethernet/stmicro/stmmac/stmmac_main.c | 462 +++++++++++------- 1 file changed, 289 insertions(+), 173 deletions(-) diff --git a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c b/drivers/ne= t/ethernet/stmicro/stmmac/stmmac_main.c index 4d1a445baa07..9a927ce17941 100644 --- a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c +++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c @@ -1232,7 +1232,8 @@ static int stmmac_phy_setup(struct stmmac_priv *priv) return 0; } =20 -static void stmmac_display_rx_rings(struct stmmac_priv *priv) +static void stmmac_display_rx_rings(struct stmmac_priv *priv, + struct stmmac_dma_conf *dma_conf) { u32 rx_cnt =3D priv->plat->rx_queues_to_use; unsigned int desc_size; @@ -1241,7 +1242,7 @@ static void stmmac_display_rx_rings(struct stmmac_pri= v *priv) =20 /* Display RX rings */ for (queue =3D 0; queue < rx_cnt; queue++) { - struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &dma_conf->rx_queue[queue]; =20 pr_info("\tRX Queue %u rings\n", queue); =20 @@ -1254,12 +1255,13 @@ static void stmmac_display_rx_rings(struct stmmac_p= riv *priv) } =20 /* Display RX ring */ - stmmac_display_ring(priv, head_rx, priv->dma_conf.dma_rx_size, true, + stmmac_display_ring(priv, head_rx, dma_conf->dma_rx_size, true, rx_q->dma_rx_phy, desc_size); } } =20 -static void stmmac_display_tx_rings(struct stmmac_priv *priv) +static void stmmac_display_tx_rings(struct stmmac_priv *priv, + struct stmmac_dma_conf *dma_conf) { u32 tx_cnt =3D priv->plat->tx_queues_to_use; unsigned int desc_size; @@ -1268,7 +1270,7 @@ static void stmmac_display_tx_rings(struct stmmac_pri= v *priv) =20 /* Display TX rings */ for (queue =3D 0; queue < tx_cnt; queue++) { - struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[queue]; + struct stmmac_tx_queue *tx_q =3D &dma_conf->tx_queue[queue]; =20 pr_info("\tTX Queue %d rings\n", queue); =20 @@ -1283,18 +1285,19 @@ static void stmmac_display_tx_rings(struct stmmac_p= riv *priv) desc_size =3D sizeof(struct dma_desc); } =20 - stmmac_display_ring(priv, head_tx, priv->dma_conf.dma_tx_size, false, + stmmac_display_ring(priv, head_tx, dma_conf->dma_tx_size, false, tx_q->dma_tx_phy, desc_size); } } =20 -static void stmmac_display_rings(struct stmmac_priv *priv) +static void stmmac_display_rings(struct stmmac_priv *priv, + struct stmmac_dma_conf *dma_conf) { /* Display RX ring */ - stmmac_display_rx_rings(priv); + stmmac_display_rx_rings(priv, dma_conf); =20 /* Display TX ring */ - stmmac_display_tx_rings(priv); + stmmac_display_tx_rings(priv, dma_conf); } =20 static int stmmac_set_bfsize(int mtu, int bufsize) @@ -1318,44 +1321,50 @@ static int stmmac_set_bfsize(int mtu, int bufsize) /** * stmmac_clear_rx_descriptors - clear RX descriptors * @priv: driver private structure + * @dma_conf: structure to take the dma data * @queue: RX queue index * Description: this function is called to clear the RX descriptors * in case of both basic and extended descriptors are used. */ -static void stmmac_clear_rx_descriptors(struct stmmac_priv *priv, u32 queu= e) +static void stmmac_clear_rx_descriptors(struct stmmac_priv *priv, + struct stmmac_dma_conf *dma_conf, + u32 queue) { - struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &dma_conf->rx_queue[queue]; int i; =20 /* Clear the RX descriptors */ - for (i =3D 0; i < priv->dma_conf.dma_rx_size; i++) + for (i =3D 0; i < dma_conf->dma_rx_size; i++) if (priv->extend_desc) stmmac_init_rx_desc(priv, &rx_q->dma_erx[i].basic, priv->use_riwt, priv->mode, - (i =3D=3D priv->dma_conf.dma_rx_size - 1), - priv->dma_conf.dma_buf_sz); + (i =3D=3D dma_conf->dma_rx_size - 1), + dma_conf->dma_buf_sz); else stmmac_init_rx_desc(priv, &rx_q->dma_rx[i], priv->use_riwt, priv->mode, - (i =3D=3D priv->dma_conf.dma_rx_size - 1), - priv->dma_conf.dma_buf_sz); + (i =3D=3D dma_conf->dma_rx_size - 1), + dma_conf->dma_buf_sz); } =20 /** * stmmac_clear_tx_descriptors - clear tx descriptors * @priv: driver private structure + * @dma_conf: structure to take the dma data * @queue: TX queue index. * Description: this function is called to clear the TX descriptors * in case of both basic and extended descriptors are used. */ -static void stmmac_clear_tx_descriptors(struct stmmac_priv *priv, u32 queu= e) +static void stmmac_clear_tx_descriptors(struct stmmac_priv *priv, + struct stmmac_dma_conf *dma_conf, + u32 queue) { - struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[queue]; + struct stmmac_tx_queue *tx_q =3D &dma_conf->tx_queue[queue]; int i; =20 /* Clear the TX descriptors */ - for (i =3D 0; i < priv->dma_conf.dma_tx_size; i++) { - int last =3D (i =3D=3D (priv->dma_conf.dma_tx_size - 1)); + for (i =3D 0; i < dma_conf->dma_tx_size; i++) { + int last =3D (i =3D=3D (dma_conf->dma_tx_size - 1)); struct dma_desc *p; =20 if (priv->extend_desc) @@ -1372,10 +1381,12 @@ static void stmmac_clear_tx_descriptors(struct stmm= ac_priv *priv, u32 queue) /** * stmmac_clear_descriptors - clear descriptors * @priv: driver private structure + * @dma_conf: structure to take the dma data * Description: this function is called to clear the TX and RX descriptors * in case of both basic and extended descriptors are used. */ -static void stmmac_clear_descriptors(struct stmmac_priv *priv) +static void stmmac_clear_descriptors(struct stmmac_priv *priv, + struct stmmac_dma_conf *dma_conf) { u32 rx_queue_cnt =3D priv->plat->rx_queues_to_use; u32 tx_queue_cnt =3D priv->plat->tx_queues_to_use; @@ -1383,16 +1394,17 @@ static void stmmac_clear_descriptors(struct stmmac_= priv *priv) =20 /* Clear the RX descriptors */ for (queue =3D 0; queue < rx_queue_cnt; queue++) - stmmac_clear_rx_descriptors(priv, queue); + stmmac_clear_rx_descriptors(priv, dma_conf, queue); =20 /* Clear the TX descriptors */ for (queue =3D 0; queue < tx_queue_cnt; queue++) - stmmac_clear_tx_descriptors(priv, queue); + stmmac_clear_tx_descriptors(priv, dma_conf, queue); } =20 /** * stmmac_init_rx_buffers - init the RX descriptor buffer. * @priv: driver private structure + * @dma_conf: structure to take the dma data * @p: descriptor pointer * @i: descriptor index * @flags: gfp flag @@ -1400,10 +1412,12 @@ static void stmmac_clear_descriptors(struct stmmac_= priv *priv) * Description: this function is called to allocate a receive buffer, perf= orm * the DMA mapping and init the descriptor. */ -static int stmmac_init_rx_buffers(struct stmmac_priv *priv, struct dma_des= c *p, +static int stmmac_init_rx_buffers(struct stmmac_priv *priv, + struct stmmac_dma_conf *dma_conf, + struct dma_desc *p, int i, gfp_t flags, u32 queue) { - struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &dma_conf->rx_queue[queue]; struct stmmac_rx_buffer *buf =3D &rx_q->buf_pool[i]; gfp_t gfp =3D (GFP_ATOMIC | __GFP_NOWARN); =20 @@ -1432,7 +1446,7 @@ static int stmmac_init_rx_buffers(struct stmmac_priv = *priv, struct dma_desc *p, buf->addr =3D page_pool_get_dma_addr(buf->page) + buf->page_offset; =20 stmmac_set_desc_addr(priv, p, buf->addr); - if (priv->dma_conf.dma_buf_sz =3D=3D BUF_SIZE_16KiB) + if (dma_conf->dma_buf_sz =3D=3D BUF_SIZE_16KiB) stmmac_init_desc3(priv, p); =20 return 0; @@ -1441,12 +1455,13 @@ static int stmmac_init_rx_buffers(struct stmmac_pri= v *priv, struct dma_desc *p, /** * stmmac_free_rx_buffer - free RX dma buffers * @priv: private structure - * @queue: RX queue index + * @rx_q: RX queue * @i: buffer index. */ -static void stmmac_free_rx_buffer(struct stmmac_priv *priv, u32 queue, int= i) +static void stmmac_free_rx_buffer(struct stmmac_priv *priv, + struct stmmac_rx_queue *rx_q, + int i) { - struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; struct stmmac_rx_buffer *buf =3D &rx_q->buf_pool[i]; =20 if (buf->page) @@ -1461,12 +1476,15 @@ static void stmmac_free_rx_buffer(struct stmmac_pri= v *priv, u32 queue, int i) /** * stmmac_free_tx_buffer - free RX dma buffers * @priv: private structure + * @dma_conf: structure to take the dma data * @queue: RX queue index * @i: buffer index. */ -static void stmmac_free_tx_buffer(struct stmmac_priv *priv, u32 queue, int= i) +static void stmmac_free_tx_buffer(struct stmmac_priv *priv, + struct stmmac_dma_conf *dma_conf, + u32 queue, int i) { - struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[queue]; + struct stmmac_tx_queue *tx_q =3D &dma_conf->tx_queue[queue]; =20 if (tx_q->tx_skbuff_dma[i].buf && tx_q->tx_skbuff_dma[i].buf_type !=3D STMMAC_TXBUF_T_XDP_TX) { @@ -1505,23 +1523,28 @@ static void stmmac_free_tx_buffer(struct stmmac_pri= v *priv, u32 queue, int i) /** * dma_free_rx_skbufs - free RX dma buffers * @priv: private structure + * @dma_conf: structure to take the dma data * @queue: RX queue index */ -static void dma_free_rx_skbufs(struct stmmac_priv *priv, u32 queue) +static void dma_free_rx_skbufs(struct stmmac_priv *priv, + struct stmmac_dma_conf *dma_conf, + u32 queue) { + struct stmmac_rx_queue *rx_q =3D &dma_conf->rx_queue[queue]; int i; =20 - for (i =3D 0; i < priv->dma_conf.dma_rx_size; i++) - stmmac_free_rx_buffer(priv, queue, i); + for (i =3D 0; i < dma_conf->dma_rx_size; i++) + stmmac_free_rx_buffer(priv, rx_q, i); } =20 -static int stmmac_alloc_rx_buffers(struct stmmac_priv *priv, u32 queue, - gfp_t flags) +static int stmmac_alloc_rx_buffers(struct stmmac_priv *priv, + struct stmmac_dma_conf *dma_conf, + u32 queue, gfp_t flags) { - struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &dma_conf->rx_queue[queue]; int i; =20 - for (i =3D 0; i < priv->dma_conf.dma_rx_size; i++) { + for (i =3D 0; i < dma_conf->dma_rx_size; i++) { struct dma_desc *p; int ret; =20 @@ -1530,7 +1553,7 @@ static int stmmac_alloc_rx_buffers(struct stmmac_priv= *priv, u32 queue, else p =3D rx_q->dma_rx + i; =20 - ret =3D stmmac_init_rx_buffers(priv, p, i, flags, + ret =3D stmmac_init_rx_buffers(priv, dma_conf, p, i, flags, queue); if (ret) return ret; @@ -1544,14 +1567,17 @@ static int stmmac_alloc_rx_buffers(struct stmmac_pr= iv *priv, u32 queue, /** * dma_free_rx_xskbufs - free RX dma buffers from XSK pool * @priv: private structure + * @dma_conf: structure to take the dma data * @queue: RX queue index */ -static void dma_free_rx_xskbufs(struct stmmac_priv *priv, u32 queue) +static void dma_free_rx_xskbufs(struct stmmac_priv *priv, + struct stmmac_dma_conf *dma_conf, + u32 queue) { - struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &dma_conf->rx_queue[queue]; int i; =20 - for (i =3D 0; i < priv->dma_conf.dma_rx_size; i++) { + for (i =3D 0; i < dma_conf->dma_rx_size; i++) { struct stmmac_rx_buffer *buf =3D &rx_q->buf_pool[i]; =20 if (!buf->xdp) @@ -1562,12 +1588,14 @@ static void dma_free_rx_xskbufs(struct stmmac_priv = *priv, u32 queue) } } =20 -static int stmmac_alloc_rx_buffers_zc(struct stmmac_priv *priv, u32 queue) +static int stmmac_alloc_rx_buffers_zc(struct stmmac_priv *priv, + struct stmmac_dma_conf *dma_conf, + u32 queue) { - struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &dma_conf->rx_queue[queue]; int i; =20 - for (i =3D 0; i < priv->dma_conf.dma_rx_size; i++) { + for (i =3D 0; i < dma_conf->dma_rx_size; i++) { struct stmmac_rx_buffer *buf; dma_addr_t dma_addr; struct dma_desc *p; @@ -1602,22 +1630,25 @@ static struct xsk_buff_pool *stmmac_get_xsk_pool(st= ruct stmmac_priv *priv, u32 q /** * __init_dma_rx_desc_rings - init the RX descriptor ring (per queue) * @priv: driver private structure + * @dma_conf: structure to take the dma data * @queue: RX queue index * @flags: gfp flag. * Description: this function initializes the DMA RX descriptors * and allocates the socket buffers. It supports the chained and ring * modes. */ -static int __init_dma_rx_desc_rings(struct stmmac_priv *priv, u32 queue, g= fp_t flags) +static int __init_dma_rx_desc_rings(struct stmmac_priv *priv, + struct stmmac_dma_conf *dma_conf, + u32 queue, gfp_t flags) { - struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &dma_conf->rx_queue[queue]; int ret; =20 netif_dbg(priv, probe, priv->dev, "(%s) dma_rx_phy=3D0x%08x\n", __func__, (u32)rx_q->dma_rx_phy); =20 - stmmac_clear_rx_descriptors(priv, queue); + stmmac_clear_rx_descriptors(priv, dma_conf, queue); =20 xdp_rxq_info_unreg_mem_model(&rx_q->xdp_rxq); =20 @@ -1644,9 +1675,9 @@ static int __init_dma_rx_desc_rings(struct stmmac_pri= v *priv, u32 queue, gfp_t f /* RX XDP ZC buffer pool may not be populated, e.g. * xdpsock TX-only. */ - stmmac_alloc_rx_buffers_zc(priv, queue); + stmmac_alloc_rx_buffers_zc(priv, dma_conf, queue); } else { - ret =3D stmmac_alloc_rx_buffers(priv, queue, flags); + ret =3D stmmac_alloc_rx_buffers(priv, dma_conf, queue, flags); if (ret < 0) return -ENOMEM; } @@ -1656,17 +1687,19 @@ static int __init_dma_rx_desc_rings(struct stmmac_p= riv *priv, u32 queue, gfp_t f if (priv->extend_desc) stmmac_mode_init(priv, rx_q->dma_erx, rx_q->dma_rx_phy, - priv->dma_conf.dma_rx_size, 1); + dma_conf->dma_rx_size, 1); else stmmac_mode_init(priv, rx_q->dma_rx, rx_q->dma_rx_phy, - priv->dma_conf.dma_rx_size, 0); + dma_conf->dma_rx_size, 0); } =20 return 0; } =20 -static int init_dma_rx_desc_rings(struct net_device *dev, gfp_t flags) +static int init_dma_rx_desc_rings(struct net_device *dev, + struct stmmac_dma_conf *dma_conf, + gfp_t flags) { struct stmmac_priv *priv =3D netdev_priv(dev); u32 rx_count =3D priv->plat->rx_queues_to_use; @@ -1678,7 +1711,7 @@ static int init_dma_rx_desc_rings(struct net_device *= dev, gfp_t flags) "SKB addresses:\nskb\t\tskb data\tdma data\n"); =20 for (queue =3D 0; queue < rx_count; queue++) { - ret =3D __init_dma_rx_desc_rings(priv, queue, flags); + ret =3D __init_dma_rx_desc_rings(priv, dma_conf, queue, flags); if (ret) goto err_init_rx_buffers; } @@ -1687,12 +1720,12 @@ static int init_dma_rx_desc_rings(struct net_device= *dev, gfp_t flags) =20 err_init_rx_buffers: while (queue >=3D 0) { - struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &dma_conf->rx_queue[queue]; =20 if (rx_q->xsk_pool) - dma_free_rx_xskbufs(priv, queue); + dma_free_rx_xskbufs(priv, dma_conf, queue); else - dma_free_rx_skbufs(priv, queue); + dma_free_rx_skbufs(priv, dma_conf, queue); =20 rx_q->buf_alloc_num =3D 0; rx_q->xsk_pool =3D NULL; @@ -1706,14 +1739,17 @@ static int init_dma_rx_desc_rings(struct net_device= *dev, gfp_t flags) /** * __init_dma_tx_desc_rings - init the TX descriptor ring (per queue) * @priv: driver private structure - * @queue : TX queue index + * @dma_conf: structure to take the dma data + * @queue: TX queue index * Description: this function initializes the DMA TX descriptors * and allocates the socket buffers. It supports the chained and ring * modes. */ -static int __init_dma_tx_desc_rings(struct stmmac_priv *priv, u32 queue) +static int __init_dma_tx_desc_rings(struct stmmac_priv *priv, + struct stmmac_dma_conf *dma_conf, + u32 queue) { - struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[queue]; + struct stmmac_tx_queue *tx_q =3D &dma_conf->tx_queue[queue]; int i; =20 netif_dbg(priv, probe, priv->dev, @@ -1725,16 +1761,16 @@ static int __init_dma_tx_desc_rings(struct stmmac_p= riv *priv, u32 queue) if (priv->extend_desc) stmmac_mode_init(priv, tx_q->dma_etx, tx_q->dma_tx_phy, - priv->dma_conf.dma_tx_size, 1); + dma_conf->dma_tx_size, 1); else if (!(tx_q->tbs & STMMAC_TBS_AVAIL)) stmmac_mode_init(priv, tx_q->dma_tx, tx_q->dma_tx_phy, - priv->dma_conf.dma_tx_size, 0); + dma_conf->dma_tx_size, 0); } =20 tx_q->xsk_pool =3D stmmac_get_xsk_pool(priv, queue); =20 - for (i =3D 0; i < priv->dma_conf.dma_tx_size; i++) { + for (i =3D 0; i < dma_conf->dma_tx_size; i++) { struct dma_desc *p; =20 if (priv->extend_desc) @@ -1756,7 +1792,8 @@ static int __init_dma_tx_desc_rings(struct stmmac_pri= v *priv, u32 queue) return 0; } =20 -static int init_dma_tx_desc_rings(struct net_device *dev) +static int init_dma_tx_desc_rings(struct net_device *dev, + struct stmmac_dma_conf *dma_conf) { struct stmmac_priv *priv =3D netdev_priv(dev); u32 tx_queue_cnt; @@ -1765,7 +1802,7 @@ static int init_dma_tx_desc_rings(struct net_device *= dev) tx_queue_cnt =3D priv->plat->tx_queues_to_use; =20 for (queue =3D 0; queue < tx_queue_cnt; queue++) - __init_dma_tx_desc_rings(priv, queue); + __init_dma_tx_desc_rings(priv, dma_conf, queue); =20 return 0; } @@ -1773,26 +1810,29 @@ static int init_dma_tx_desc_rings(struct net_device= *dev) /** * init_dma_desc_rings - init the RX/TX descriptor rings * @dev: net device structure + * @dma_conf: structure to take the dma data * @flags: gfp flag. * Description: this function initializes the DMA RX/TX descriptors * and allocates the socket buffers. It supports the chained and ring * modes. */ -static int init_dma_desc_rings(struct net_device *dev, gfp_t flags) +static int init_dma_desc_rings(struct net_device *dev, + struct stmmac_dma_conf *dma_conf, + gfp_t flags) { struct stmmac_priv *priv =3D netdev_priv(dev); int ret; =20 - ret =3D init_dma_rx_desc_rings(dev, flags); + ret =3D init_dma_rx_desc_rings(dev, dma_conf, flags); if (ret) return ret; =20 - ret =3D init_dma_tx_desc_rings(dev); + ret =3D init_dma_tx_desc_rings(dev, dma_conf); =20 - stmmac_clear_descriptors(priv); + stmmac_clear_descriptors(priv, dma_conf); =20 if (netif_msg_hw(priv)) - stmmac_display_rings(priv); + stmmac_display_rings(priv, dma_conf); =20 return ret; } @@ -1800,17 +1840,20 @@ static int init_dma_desc_rings(struct net_device *d= ev, gfp_t flags) /** * dma_free_tx_skbufs - free TX dma buffers * @priv: private structure + * @dma_conf: structure to take the dma data * @queue: TX queue index */ -static void dma_free_tx_skbufs(struct stmmac_priv *priv, u32 queue) +static void dma_free_tx_skbufs(struct stmmac_priv *priv, + struct stmmac_dma_conf *dma_conf, + u32 queue) { - struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[queue]; + struct stmmac_tx_queue *tx_q =3D &dma_conf->tx_queue[queue]; int i; =20 tx_q->xsk_frames_done =3D 0; =20 - for (i =3D 0; i < priv->dma_conf.dma_tx_size; i++) - stmmac_free_tx_buffer(priv, queue, i); + for (i =3D 0; i < dma_conf->dma_tx_size; i++) + stmmac_free_tx_buffer(priv, dma_conf, queue, i); =20 if (tx_q->xsk_pool && tx_q->xsk_frames_done) { xsk_tx_completed(tx_q->xsk_pool, tx_q->xsk_frames_done); @@ -1829,34 +1872,37 @@ static void stmmac_free_tx_skbufs(struct stmmac_pri= v *priv) u32 queue; =20 for (queue =3D 0; queue < tx_queue_cnt; queue++) - dma_free_tx_skbufs(priv, queue); + dma_free_tx_skbufs(priv, &priv->dma_conf, queue); } =20 /** * __free_dma_rx_desc_resources - free RX dma desc resources (per queue) * @priv: private structure + * @dma_conf: structure to take the dma data * @queue: RX queue index */ -static void __free_dma_rx_desc_resources(struct stmmac_priv *priv, u32 que= ue) +static void __free_dma_rx_desc_resources(struct stmmac_priv *priv, + struct stmmac_dma_conf *dma_conf, + u32 queue) { - struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &dma_conf->rx_queue[queue]; =20 /* Release the DMA RX socket buffers */ if (rx_q->xsk_pool) - dma_free_rx_xskbufs(priv, queue); + dma_free_rx_xskbufs(priv, dma_conf, queue); else - dma_free_rx_skbufs(priv, queue); + dma_free_rx_skbufs(priv, dma_conf, queue); =20 rx_q->buf_alloc_num =3D 0; rx_q->xsk_pool =3D NULL; =20 /* Free DMA regions of consistent memory previously allocated */ if (!priv->extend_desc) - dma_free_coherent(priv->device, priv->dma_conf.dma_rx_size * + dma_free_coherent(priv->device, dma_conf->dma_rx_size * sizeof(struct dma_desc), rx_q->dma_rx, rx_q->dma_rx_phy); else - dma_free_coherent(priv->device, priv->dma_conf.dma_rx_size * + dma_free_coherent(priv->device, dma_conf->dma_rx_size * sizeof(struct dma_extended_desc), rx_q->dma_erx, rx_q->dma_rx_phy); =20 @@ -1868,29 +1914,33 @@ static void __free_dma_rx_desc_resources(struct stm= mac_priv *priv, u32 queue) page_pool_destroy(rx_q->page_pool); } =20 -static void free_dma_rx_desc_resources(struct stmmac_priv *priv) +static void free_dma_rx_desc_resources(struct stmmac_priv *priv, + struct stmmac_dma_conf *dma_conf) { u32 rx_count =3D priv->plat->rx_queues_to_use; u32 queue; =20 /* Free RX queue resources */ for (queue =3D 0; queue < rx_count; queue++) - __free_dma_rx_desc_resources(priv, queue); + __free_dma_rx_desc_resources(priv, dma_conf, queue); } =20 /** * __free_dma_tx_desc_resources - free TX dma desc resources (per queue) * @priv: private structure + * @dma_conf: structure to take the dma data * @queue: TX queue index */ -static void __free_dma_tx_desc_resources(struct stmmac_priv *priv, u32 que= ue) +static void __free_dma_tx_desc_resources(struct stmmac_priv *priv, + struct stmmac_dma_conf *dma_conf, + u32 queue) { - struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[queue]; + struct stmmac_tx_queue *tx_q =3D &dma_conf->tx_queue[queue]; size_t size; void *addr; =20 /* Release the DMA TX socket buffers */ - dma_free_tx_skbufs(priv, queue); + dma_free_tx_skbufs(priv, dma_conf, queue); =20 if (priv->extend_desc) { size =3D sizeof(struct dma_extended_desc); @@ -1903,7 +1953,7 @@ static void __free_dma_tx_desc_resources(struct stmma= c_priv *priv, u32 queue) addr =3D tx_q->dma_tx; } =20 - size *=3D priv->dma_conf.dma_tx_size; + size *=3D dma_conf->dma_tx_size; =20 dma_free_coherent(priv->device, size, addr, tx_q->dma_tx_phy); =20 @@ -1911,28 +1961,32 @@ static void __free_dma_tx_desc_resources(struct stm= mac_priv *priv, u32 queue) kfree(tx_q->tx_skbuff); } =20 -static void free_dma_tx_desc_resources(struct stmmac_priv *priv) +static void free_dma_tx_desc_resources(struct stmmac_priv *priv, + struct stmmac_dma_conf *dma_conf) { u32 tx_count =3D priv->plat->tx_queues_to_use; u32 queue; =20 /* Free TX queue resources */ for (queue =3D 0; queue < tx_count; queue++) - __free_dma_tx_desc_resources(priv, queue); + __free_dma_tx_desc_resources(priv, dma_conf, queue); } =20 /** * __alloc_dma_rx_desc_resources - alloc RX resources (per queue). * @priv: private structure + * @dma_conf: structure to take the dma data * @queue: RX queue index * Description: according to which descriptor can be used (extend or basic) * this function allocates the resources for TX and RX paths. In case of * reception, for example, it pre-allocated the RX socket buffer in order = to * allow zero-copy mechanism. */ -static int __alloc_dma_rx_desc_resources(struct stmmac_priv *priv, u32 que= ue) +static int __alloc_dma_rx_desc_resources(struct stmmac_priv *priv, + struct stmmac_dma_conf *dma_conf, + u32 queue) { - struct stmmac_rx_queue *rx_q =3D &priv->dma_conf.rx_queue[queue]; + struct stmmac_rx_queue *rx_q =3D &dma_conf->rx_queue[queue]; struct stmmac_channel *ch =3D &priv->channel[queue]; bool xdp_prog =3D stmmac_xdp_is_enabled(priv); struct page_pool_params pp_params =3D { 0 }; @@ -1944,8 +1998,8 @@ static int __alloc_dma_rx_desc_resources(struct stmma= c_priv *priv, u32 queue) rx_q->priv_data =3D priv; =20 pp_params.flags =3D PP_FLAG_DMA_MAP | PP_FLAG_DMA_SYNC_DEV; - pp_params.pool_size =3D priv->dma_conf.dma_rx_size; - num_pages =3D DIV_ROUND_UP(priv->dma_conf.dma_buf_sz, PAGE_SIZE); + pp_params.pool_size =3D dma_conf->dma_rx_size; + num_pages =3D DIV_ROUND_UP(dma_conf->dma_buf_sz, PAGE_SIZE); pp_params.order =3D ilog2(num_pages); pp_params.nid =3D dev_to_node(priv->device); pp_params.dev =3D priv->device; @@ -1960,7 +2014,7 @@ static int __alloc_dma_rx_desc_resources(struct stmma= c_priv *priv, u32 queue) return ret; } =20 - rx_q->buf_pool =3D kcalloc(priv->dma_conf.dma_rx_size, + rx_q->buf_pool =3D kcalloc(dma_conf->dma_rx_size, sizeof(*rx_q->buf_pool), GFP_KERNEL); if (!rx_q->buf_pool) @@ -1968,7 +2022,7 @@ static int __alloc_dma_rx_desc_resources(struct stmma= c_priv *priv, u32 queue) =20 if (priv->extend_desc) { rx_q->dma_erx =3D dma_alloc_coherent(priv->device, - priv->dma_conf.dma_rx_size * + dma_conf->dma_rx_size * sizeof(struct dma_extended_desc), &rx_q->dma_rx_phy, GFP_KERNEL); @@ -1977,7 +2031,7 @@ static int __alloc_dma_rx_desc_resources(struct stmma= c_priv *priv, u32 queue) =20 } else { rx_q->dma_rx =3D dma_alloc_coherent(priv->device, - priv->dma_conf.dma_rx_size * + dma_conf->dma_rx_size * sizeof(struct dma_desc), &rx_q->dma_rx_phy, GFP_KERNEL); @@ -2002,7 +2056,8 @@ static int __alloc_dma_rx_desc_resources(struct stmma= c_priv *priv, u32 queue) return 0; } =20 -static int alloc_dma_rx_desc_resources(struct stmmac_priv *priv) +static int alloc_dma_rx_desc_resources(struct stmmac_priv *priv, + struct stmmac_dma_conf *dma_conf) { u32 rx_count =3D priv->plat->rx_queues_to_use; u32 queue; @@ -2010,7 +2065,7 @@ static int alloc_dma_rx_desc_resources(struct stmmac_= priv *priv) =20 /* RX queues buffers and DMA */ for (queue =3D 0; queue < rx_count; queue++) { - ret =3D __alloc_dma_rx_desc_resources(priv, queue); + ret =3D __alloc_dma_rx_desc_resources(priv, dma_conf, queue); if (ret) goto err_dma; } @@ -2018,7 +2073,7 @@ static int alloc_dma_rx_desc_resources(struct stmmac_= priv *priv) return 0; =20 err_dma: - free_dma_rx_desc_resources(priv); + free_dma_rx_desc_resources(priv, dma_conf); =20 return ret; } @@ -2026,28 +2081,31 @@ static int alloc_dma_rx_desc_resources(struct stmma= c_priv *priv) /** * __alloc_dma_tx_desc_resources - alloc TX resources (per queue). * @priv: private structure + * @dma_conf: structure to take the dma data * @queue: TX queue index * Description: according to which descriptor can be used (extend or basic) * this function allocates the resources for TX and RX paths. In case of * reception, for example, it pre-allocated the RX socket buffer in order = to * allow zero-copy mechanism. */ -static int __alloc_dma_tx_desc_resources(struct stmmac_priv *priv, u32 que= ue) +static int __alloc_dma_tx_desc_resources(struct stmmac_priv *priv, + struct stmmac_dma_conf *dma_conf, + u32 queue) { - struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[queue]; + struct stmmac_tx_queue *tx_q =3D &dma_conf->tx_queue[queue]; size_t size; void *addr; =20 tx_q->queue_index =3D queue; tx_q->priv_data =3D priv; =20 - tx_q->tx_skbuff_dma =3D kcalloc(priv->dma_conf.dma_tx_size, + tx_q->tx_skbuff_dma =3D kcalloc(dma_conf->dma_tx_size, sizeof(*tx_q->tx_skbuff_dma), GFP_KERNEL); if (!tx_q->tx_skbuff_dma) return -ENOMEM; =20 - tx_q->tx_skbuff =3D kcalloc(priv->dma_conf.dma_tx_size, + tx_q->tx_skbuff =3D kcalloc(dma_conf->dma_tx_size, sizeof(struct sk_buff *), GFP_KERNEL); if (!tx_q->tx_skbuff) @@ -2060,7 +2118,7 @@ static int __alloc_dma_tx_desc_resources(struct stmma= c_priv *priv, u32 queue) else size =3D sizeof(struct dma_desc); =20 - size *=3D priv->dma_conf.dma_tx_size; + size *=3D dma_conf->dma_tx_size; =20 addr =3D dma_alloc_coherent(priv->device, size, &tx_q->dma_tx_phy, GFP_KERNEL); @@ -2077,7 +2135,8 @@ static int __alloc_dma_tx_desc_resources(struct stmma= c_priv *priv, u32 queue) return 0; } =20 -static int alloc_dma_tx_desc_resources(struct stmmac_priv *priv) +static int alloc_dma_tx_desc_resources(struct stmmac_priv *priv, + struct stmmac_dma_conf *dma_conf) { u32 tx_count =3D priv->plat->tx_queues_to_use; u32 queue; @@ -2085,7 +2144,7 @@ static int alloc_dma_tx_desc_resources(struct stmmac_= priv *priv) =20 /* TX queues buffers and DMA */ for (queue =3D 0; queue < tx_count; queue++) { - ret =3D __alloc_dma_tx_desc_resources(priv, queue); + ret =3D __alloc_dma_tx_desc_resources(priv, dma_conf, queue); if (ret) goto err_dma; } @@ -2093,27 +2152,29 @@ static int alloc_dma_tx_desc_resources(struct stmma= c_priv *priv) return 0; =20 err_dma: - free_dma_tx_desc_resources(priv); + free_dma_tx_desc_resources(priv, dma_conf); return ret; } =20 /** * alloc_dma_desc_resources - alloc TX/RX resources. * @priv: private structure + * @dma_conf: structure to take the dma data * Description: according to which descriptor can be used (extend or basic) * this function allocates the resources for TX and RX paths. In case of * reception, for example, it pre-allocated the RX socket buffer in order = to * allow zero-copy mechanism. */ -static int alloc_dma_desc_resources(struct stmmac_priv *priv) +static int alloc_dma_desc_resources(struct stmmac_priv *priv, + struct stmmac_dma_conf *dma_conf) { /* RX Allocation */ - int ret =3D alloc_dma_rx_desc_resources(priv); + int ret =3D alloc_dma_rx_desc_resources(priv, dma_conf); =20 if (ret) return ret; =20 - ret =3D alloc_dma_tx_desc_resources(priv); + ret =3D alloc_dma_tx_desc_resources(priv, dma_conf); =20 return ret; } @@ -2121,16 +2182,18 @@ static int alloc_dma_desc_resources(struct stmmac_p= riv *priv) /** * free_dma_desc_resources - free dma desc resources * @priv: private structure + * @dma_conf: structure to take the dma data */ -static void free_dma_desc_resources(struct stmmac_priv *priv) +static void free_dma_desc_resources(struct stmmac_priv *priv, + struct stmmac_dma_conf *dma_conf) { /* Release the DMA TX socket buffers */ - free_dma_tx_desc_resources(priv); + free_dma_tx_desc_resources(priv, dma_conf); =20 /* Release the DMA RX socket buffers later * to ensure all pending XDP_TX buffers are returned. */ - free_dma_rx_desc_resources(priv); + free_dma_rx_desc_resources(priv, dma_conf); } =20 /** @@ -2636,8 +2699,8 @@ static void stmmac_tx_err(struct stmmac_priv *priv, u= 32 chan) netif_tx_stop_queue(netdev_get_tx_queue(priv->dev, chan)); =20 stmmac_stop_tx_dma(priv, chan); - dma_free_tx_skbufs(priv, chan); - stmmac_clear_tx_descriptors(priv, chan); + dma_free_tx_skbufs(priv, &priv->dma_conf, chan); + stmmac_clear_tx_descriptors(priv, &priv->dma_conf, chan); stmmac_reset_tx_queue(priv, chan); stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg, tx_q->dma_tx_phy, chan); @@ -3620,19 +3683,93 @@ static int stmmac_request_irq(struct net_device *de= v) } =20 /** - * stmmac_open - open entry point of the driver + * stmmac_setup_dma_desc - Generate a dma_conf and allocate DMA queue + * @priv: driver private structure + * @mtu: MTU to setup the dma queue and buf with + * Description: Allocate and generate a dma_conf based on the provided MT= U. + * Allocate the Tx/Rx DMA queue and init them. + * Return value: + * the dma_conf allocated struct on success and an appropriate ERR_PTR on= failure. + */ +static struct stmmac_dma_conf * +stmmac_setup_dma_desc(struct stmmac_priv *priv, unsigned int mtu) +{ + struct stmmac_dma_conf *dma_conf; + int chan, bfsize, ret; + + dma_conf =3D kzalloc(sizeof(*dma_conf), GFP_KERNEL); + if (!dma_conf) { + netdev_err(priv->dev, "%s: DMA conf allocation failed\n", + __func__); + return ERR_PTR(-ENOMEM); + } + + bfsize =3D stmmac_set_16kib_bfsize(priv, mtu); + if (bfsize < 0) + bfsize =3D 0; + + if (bfsize < BUF_SIZE_16KiB) + bfsize =3D stmmac_set_bfsize(mtu, 0); + + dma_conf->dma_buf_sz =3D bfsize; + /* Chose the tx/rx size from the already defined one in the + * priv struct. (if defined) + */ + dma_conf->dma_tx_size =3D priv->dma_conf.dma_tx_size; + dma_conf->dma_rx_size =3D priv->dma_conf.dma_rx_size; + + if (!dma_conf->dma_tx_size) + dma_conf->dma_tx_size =3D DMA_DEFAULT_TX_SIZE; + if (!dma_conf->dma_rx_size) + dma_conf->dma_rx_size =3D DMA_DEFAULT_RX_SIZE; + + /* Earlier check for TBS */ + for (chan =3D 0; chan < priv->plat->tx_queues_to_use; chan++) { + struct stmmac_tx_queue *tx_q =3D &dma_conf->tx_queue[chan]; + int tbs_en =3D priv->plat->tx_queues_cfg[chan].tbs_en; + + /* Setup per-TXQ tbs flag before TX descriptor alloc */ + tx_q->tbs |=3D tbs_en ? STMMAC_TBS_AVAIL : 0; + } + + ret =3D alloc_dma_desc_resources(priv, dma_conf); + if (ret < 0) { + netdev_err(priv->dev, "%s: DMA descriptors allocation failed\n", + __func__); + goto alloc_error; + } + + ret =3D init_dma_desc_rings(priv->dev, dma_conf, GFP_KERNEL); + if (ret < 0) { + netdev_err(priv->dev, "%s: DMA descriptors initialization failed\n", + __func__); + goto init_error; + } + + return dma_conf; + +init_error: + free_dma_desc_resources(priv, dma_conf); +alloc_error: + kfree(dma_conf); + return ERR_PTR(ret); +} + +/** + * __stmmac_open - open entry point of the driver * @dev : pointer to the device structure. + * @dma_conf : structure to take the dma data * Description: * This function is the open entry point of the driver. * Return value: * 0 on success and an appropriate (-)ve integer as defined in errno.h * file on failure. */ -static int stmmac_open(struct net_device *dev) +static int __stmmac_open(struct net_device *dev, + struct stmmac_dma_conf *dma_conf) { struct stmmac_priv *priv =3D netdev_priv(dev); int mode =3D priv->plat->phy_interface; - int bfsize =3D 0; u32 chan; int ret; =20 @@ -3657,45 +3794,10 @@ static int stmmac_open(struct net_device *dev) memset(&priv->xstats, 0, sizeof(struct stmmac_extra_stats)); priv->xstats.threshold =3D tc; =20 - bfsize =3D stmmac_set_16kib_bfsize(priv, dev->mtu); - if (bfsize < 0) - bfsize =3D 0; - - if (bfsize < BUF_SIZE_16KiB) - bfsize =3D stmmac_set_bfsize(dev->mtu, priv->dma_conf.dma_buf_sz); - - priv->dma_conf.dma_buf_sz =3D bfsize; - buf_sz =3D bfsize; - priv->rx_copybreak =3D STMMAC_RX_COPYBREAK; =20 - if (!priv->dma_conf.dma_tx_size) - priv->dma_conf.dma_tx_size =3D DMA_DEFAULT_TX_SIZE; - if (!priv->dma_conf.dma_rx_size) - priv->dma_conf.dma_rx_size =3D DMA_DEFAULT_RX_SIZE; - - /* Earlier check for TBS */ - for (chan =3D 0; chan < priv->plat->tx_queues_to_use; chan++) { - struct stmmac_tx_queue *tx_q =3D &priv->dma_conf.tx_queue[chan]; - int tbs_en =3D priv->plat->tx_queues_cfg[chan].tbs_en; - - /* Setup per-TXQ tbs flag before TX descriptor alloc */ - tx_q->tbs |=3D tbs_en ? STMMAC_TBS_AVAIL : 0; - } - - ret =3D alloc_dma_desc_resources(priv); - if (ret < 0) { - netdev_err(priv->dev, "%s: DMA descriptors allocation failed\n", - __func__); - goto dma_desc_error; - } - - ret =3D init_dma_desc_rings(dev, GFP_KERNEL); - if (ret < 0) { - netdev_err(priv->dev, "%s: DMA descriptors initialization failed\n", - __func__); - goto init_error; - } + buf_sz =3D dma_conf->dma_buf_sz; + memcpy(&priv->dma_conf, dma_conf, sizeof(*dma_conf)); =20 stmmac_reset_queues_param(priv); =20 @@ -3729,14 +3831,28 @@ static int stmmac_open(struct net_device *dev) =20 stmmac_hw_teardown(dev); init_error: - free_dma_desc_resources(priv); -dma_desc_error: + free_dma_desc_resources(priv, &priv->dma_conf); phylink_disconnect_phy(priv->phylink); init_phy_error: pm_runtime_put(priv->device); return ret; } =20 +static int stmmac_open(struct net_device *dev) +{ + struct stmmac_priv *priv =3D netdev_priv(dev); + struct stmmac_dma_conf *dma_conf; + int ret; + + dma_conf =3D stmmac_setup_dma_desc(priv, dev->mtu); + if (IS_ERR(dma_conf)) + return PTR_ERR(dma_conf); + + ret =3D __stmmac_open(dev, dma_conf); + kfree(dma_conf); + return ret; +} + static void stmmac_fpe_stop_wq(struct stmmac_priv *priv) { set_bit(__FPE_REMOVING, &priv->fpe_task_state); @@ -3783,7 +3899,7 @@ static int stmmac_release(struct net_device *dev) stmmac_stop_all_dma(priv); =20 /* Release and free the Rx/Tx resources */ - free_dma_desc_resources(priv); + free_dma_desc_resources(priv, &priv->dma_conf); =20 /* Disable the MAC Rx/Tx */ stmmac_mac_set(priv, priv->ioaddr, false); @@ -6305,7 +6421,7 @@ void stmmac_disable_rx_queue(struct stmmac_priv *priv= , u32 queue) spin_unlock_irqrestore(&ch->lock, flags); =20 stmmac_stop_rx_dma(priv, queue); - __free_dma_rx_desc_resources(priv, queue); + __free_dma_rx_desc_resources(priv, &priv->dma_conf, queue); } =20 void stmmac_enable_rx_queue(struct stmmac_priv *priv, u32 queue) @@ -6316,21 +6432,21 @@ void stmmac_enable_rx_queue(struct stmmac_priv *pri= v, u32 queue) u32 buf_size; int ret; =20 - ret =3D __alloc_dma_rx_desc_resources(priv, queue); + ret =3D __alloc_dma_rx_desc_resources(priv, &priv->dma_conf, queue); if (ret) { netdev_err(priv->dev, "Failed to alloc RX desc.\n"); return; } =20 - ret =3D __init_dma_rx_desc_rings(priv, queue, GFP_KERNEL); + ret =3D __init_dma_rx_desc_rings(priv, &priv->dma_conf, queue, GFP_KERNEL= ); if (ret) { - __free_dma_rx_desc_resources(priv, queue); + __free_dma_rx_desc_resources(priv, &priv->dma_conf, queue); netdev_err(priv->dev, "Failed to init RX desc.\n"); return; } =20 stmmac_reset_rx_queue(priv, queue); - stmmac_clear_rx_descriptors(priv, queue); + stmmac_clear_rx_descriptors(priv, &priv->dma_conf, queue); =20 stmmac_init_rx_chan(priv, priv->ioaddr, priv->plat->dma_cfg, rx_q->dma_rx_phy, rx_q->queue_index); @@ -6368,7 +6484,7 @@ void stmmac_disable_tx_queue(struct stmmac_priv *priv= , u32 queue) spin_unlock_irqrestore(&ch->lock, flags); =20 stmmac_stop_tx_dma(priv, queue); - __free_dma_tx_desc_resources(priv, queue); + __free_dma_tx_desc_resources(priv, &priv->dma_conf, queue); } =20 void stmmac_enable_tx_queue(struct stmmac_priv *priv, u32 queue) @@ -6378,21 +6494,21 @@ void stmmac_enable_tx_queue(struct stmmac_priv *pri= v, u32 queue) unsigned long flags; int ret; =20 - ret =3D __alloc_dma_tx_desc_resources(priv, queue); + ret =3D __alloc_dma_tx_desc_resources(priv, &priv->dma_conf, queue); if (ret) { netdev_err(priv->dev, "Failed to alloc TX desc.\n"); return; } =20 - ret =3D __init_dma_tx_desc_rings(priv, queue); + ret =3D __init_dma_tx_desc_rings(priv, &priv->dma_conf, queue); if (ret) { - __free_dma_tx_desc_resources(priv, queue); + __free_dma_tx_desc_resources(priv, &priv->dma_conf, queue); netdev_err(priv->dev, "Failed to init TX desc.\n"); return; } =20 stmmac_reset_tx_queue(priv, queue); - stmmac_clear_tx_descriptors(priv, queue); + stmmac_clear_tx_descriptors(priv, &priv->dma_conf, queue); =20 stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg, tx_q->dma_tx_phy, tx_q->queue_index); @@ -6429,7 +6545,7 @@ void stmmac_xdp_release(struct net_device *dev) stmmac_stop_all_dma(priv); =20 /* Release and free the Rx/Tx resources */ - free_dma_desc_resources(priv); + free_dma_desc_resources(priv, &priv->dma_conf); =20 /* Disable the MAC Rx/Tx */ stmmac_mac_set(priv, priv->ioaddr, false); @@ -6454,14 +6570,14 @@ int stmmac_xdp_open(struct net_device *dev) u32 chan; int ret; =20 - ret =3D alloc_dma_desc_resources(priv); + ret =3D alloc_dma_desc_resources(priv, &priv->dma_conf); if (ret < 0) { netdev_err(dev, "%s: DMA descriptors allocation failed\n", __func__); goto dma_desc_error; } =20 - ret =3D init_dma_desc_rings(dev, GFP_KERNEL); + ret =3D init_dma_desc_rings(dev, &priv->dma_conf, GFP_KERNEL); if (ret < 0) { netdev_err(dev, "%s: DMA descriptors initialization failed\n", __func__); @@ -6543,7 +6659,7 @@ int stmmac_xdp_open(struct net_device *dev) =20 stmmac_hw_teardown(dev); init_error: - free_dma_desc_resources(priv); + free_dma_desc_resources(priv, &priv->dma_conf); dma_desc_error: return ret; } @@ -7411,7 +7527,7 @@ int stmmac_resume(struct device *dev) stmmac_reset_queues_param(priv); =20 stmmac_free_tx_skbufs(priv); - stmmac_clear_descriptors(priv); + stmmac_clear_descriptors(priv, &priv->dma_conf); =20 stmmac_hw_setup(ndev, false); stmmac_init_coalesce(priv); --=20 2.36.1 From nobody Wed Apr 15 07:02:46 2026 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id D0719C43334 for ; Sat, 23 Jul 2022 14:30:33 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S238424AbiGWOaa (ORCPT ); Sat, 23 Jul 2022 10:30:30 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:33942 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S238104AbiGWOaG (ORCPT ); Sat, 23 Jul 2022 10:30:06 -0400 Received: from mail-wm1-x334.google.com (mail-wm1-x334.google.com [IPv6:2a00:1450:4864:20::334]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 41FEB1CB3A; Sat, 23 Jul 2022 07:30:02 -0700 (PDT) Received: by mail-wm1-x334.google.com with SMTP id v5so4302916wmj.0; Sat, 23 Jul 2022 07:30:02 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20210112; h=from:to:cc:subject:date:message-id:in-reply-to:references :mime-version:content-transfer-encoding; bh=GX0e7m5ZrkPjpxkJdIBL+iV+EyM6eAVNIs5WQVoDC5I=; b=GuUs35D+Od53ehyBq7jSWrPuxnPg2k3rYm9X0U1icEB1hrTVCqlqQBL8TjQJp8E9Fu UyD0jy55G4vAFxREf4ZMZVOKBCT5oCkzuLx9FMnVNRjKfN3h6OQ2eYIzl0RItI+9XMJr vNZ1uImREQXKLyZO9AvoWamfOWSQhYgUSQh1ZxrlAOxZLgYdU3dpew+Cbc+4tvsBPULQ 8MtbxGJe+v2wSYdspSqQPa8PFQYd685D1/Vwj3nmTOCHo7JlCqSmJ9qL9s2NoxmC+5G0 5Ho7csQZ/+WVE5zDHwe1NoeSCbh2Wkn43qweWXZNJgS01xQuE7cpYZ74F6srqK4GThyi aWFQ== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references:mime-version:content-transfer-encoding; bh=GX0e7m5ZrkPjpxkJdIBL+iV+EyM6eAVNIs5WQVoDC5I=; b=IAe/TPuHUs+JGLieZHTl+1Nu2u7W1G+1JZ96e2DH5EsG1NpEiw8NH8QBGYRA7NMNvS LHCJS05QNGd4vATbQvcPPieTM2+82sk2+K5qqL/+ed1sYhV7aLklsEjjtMRAsVvM++Vj uCd93lNPuoIBCtWepSzNI5RT5PU8DANkect6aCRTYq27wGKtaQ2RhkVPnaGeDapVMj/i qP9TtPPAi3mZhL5Jxfmqx32JzOr1cBc881TsTFHWgHe/5qrCfvXekli6JZw7KexQnzLx lh4W8Cbv8oB+JKnVUaHueiL/4Ttvv1sYQ12iC15XBkgQoeNaMKdOAzWO70Ou4Z4NZRNo v1gQ== X-Gm-Message-State: AJIora8gn5BsautO1vJG3+Yb05dXAZm/YEb/YDVIsBNC6rFp8FjJMvKD YkogtgJ98pFcNvksIldeupWZzcQYX1U= X-Google-Smtp-Source: AGRyM1vyEESD+gOYudb18RYonLwuCA9ecgk+Boyi9uzgAedUAyeJVXmG9G/HRzwQjgJRpYQqd/XNAQ== X-Received: by 2002:a05:600c:3ba3:b0:3a3:5dd:f10f with SMTP id n35-20020a05600c3ba300b003a305ddf10fmr3105539wms.185.1658586601145; Sat, 23 Jul 2022 07:30:01 -0700 (PDT) Received: from localhost.localdomain (host-87-7-207-127.retail.telecomitalia.it. [87.7.207.127]) by smtp.googlemail.com with ESMTPSA id q6-20020a1cf306000000b0039c5ab7167dsm11689717wmq.48.2022.07.23.07.29.58 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Sat, 23 Jul 2022 07:30:00 -0700 (PDT) From: Christian Marangi To: Giuseppe Cavallaro , Alexandre Torgue , Jose Abreu , "David S. Miller" , Eric Dumazet , Jakub Kicinski , Paolo Abeni , Maxime Coquelin , Russell King , netdev@vger.kernel.org, linux-stm32@st-md-mailman.stormreply.com, linux-arm-kernel@lists.infradead.org, linux-kernel@vger.kernel.org Cc: Christian Marangi Subject: [net-next PATCH v5 5/5] net: ethernet: stmicro: stmmac: permit MTU change with interface up Date: Sat, 23 Jul 2022 16:29:33 +0200 Message-Id: <20220723142933.16030-6-ansuelsmth@gmail.com> X-Mailer: git-send-email 2.36.1 In-Reply-To: <20220723142933.16030-1-ansuelsmth@gmail.com> References: <20220723142933.16030-1-ansuelsmth@gmail.com> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Type: text/plain; charset="utf-8" Remove the limitation where the interface needs to be down to change MTU by releasing and opening the stmmac driver to set the new MTU. Also call the set_filter function to correctly init the port. This permits to remove the EBUSY error while the ethernet port is running permitting a correct MTU change if for example a DSA request a MTU change for a switch CPU port. Signed-off-by: Christian Marangi --- .../net/ethernet/stmicro/stmmac/stmmac_main.c | 30 +++++++++++++++---- 1 file changed, 24 insertions(+), 6 deletions(-) diff --git a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c b/drivers/ne= t/ethernet/stmicro/stmmac/stmmac_main.c index 9a927ce17941..083d08b5ab0e 100644 --- a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c +++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c @@ -5552,18 +5552,15 @@ static int stmmac_change_mtu(struct net_device *dev= , int new_mtu) { struct stmmac_priv *priv =3D netdev_priv(dev); int txfifosz =3D priv->plat->tx_fifo_size; + struct stmmac_dma_conf *dma_conf; const int mtu =3D new_mtu; + int ret; =20 if (txfifosz =3D=3D 0) txfifosz =3D priv->dma_cap.tx_fifo_size; =20 txfifosz /=3D priv->plat->tx_queues_to_use; =20 - if (netif_running(dev)) { - netdev_err(priv->dev, "must be stopped to change its MTU\n"); - return -EBUSY; - } - if (stmmac_xdp_is_enabled(priv) && new_mtu > ETH_DATA_LEN) { netdev_dbg(priv->dev, "Jumbo frames not supported for XDP\n"); return -EINVAL; @@ -5575,8 +5572,29 @@ static int stmmac_change_mtu(struct net_device *dev,= int new_mtu) if ((txfifosz < new_mtu) || (new_mtu > BUF_SIZE_16KiB)) return -EINVAL; =20 - dev->mtu =3D mtu; + if (netif_running(dev)) { + netdev_dbg(priv->dev, "restarting interface to change its MTU\n"); + /* Try to allocate the new DMA conf with the new mtu */ + dma_conf =3D stmmac_setup_dma_desc(priv, mtu); + if (IS_ERR(dma_conf)) { + netdev_err(priv->dev, "failed allocating new dma conf for new MTU %d\n", + mtu); + return PTR_ERR(dma_conf); + } =20 + stmmac_release(dev); + + ret =3D __stmmac_open(dev, dma_conf); + kfree(dma_conf); + if (ret) { + netdev_err(priv->dev, "failed reopening the interface after MTU change\= n"); + return ret; + } + + stmmac_set_rx_mode(dev); + } + + dev->mtu =3D mtu; netdev_update_features(dev); =20 return 0; --=20 2.36.1