From nobody Mon Nov 25 16:51:50 2024 Received: from szxga02-in.huawei.com (szxga02-in.huawei.com [45.249.212.188]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 294341DFD8; Sat, 26 Oct 2024 03:57:48 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=45.249.212.188 ARC-Seal: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1729915071; cv=none; b=QRE/Qt1oMTKgBGh7l/Vwt2Ec2g/VuOB/jqevnxg6+rgZ4VhnkCildoM5e5DjBQ3IY1XU0vOFbDrx5oy4DrzGKirBX5H8vNkgw28o7Ujx1hPkebj3ar/Y6ooQbODoI/iB/rD3Ob8CSnAOiLHXe94ARaf4LA676p88jvOyIYVkgpM= ARC-Message-Signature: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1729915071; c=relaxed/simple; bh=hUSRMfyu9RY2XLrAWKOMXbLeFv+ti9pDxOGiJ/kd2CI=; h=From:To:CC:Subject:Date:Message-ID:In-Reply-To:References: MIME-Version:Content-Type; b=hk0czgEnpnQfldhaMEMHwAC+HpEGkFsSbGb77StaYyUQ7KISmFq2+0NhZEa1N+enMazttBTNwNXupu1pa83g1NzBSQziW9ZFZPkJytur4sD4D5FkyZCCro6AHvbh5trzlqGpALlfJKCwtblSxf4TAsZ73PPLXw0P7ZnzMmJaFyM= ARC-Authentication-Results: i=1; smtp.subspace.kernel.org; dmarc=pass (p=quarantine dis=none) header.from=huawei.com; spf=pass smtp.mailfrom=huawei.com; arc=none smtp.client-ip=45.249.212.188 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=quarantine dis=none) header.from=huawei.com Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=huawei.com Received: from mail.maildlp.com (unknown [172.19.163.174]) by szxga02-in.huawei.com (SkyGuard) with ESMTP id 4Xb5R92507zlXH0; Sat, 26 Oct 2024 11:55:49 +0800 (CST) Received: from dggpemf500002.china.huawei.com (unknown [7.185.36.57]) by mail.maildlp.com (Postfix) with ESMTPS id B3FD01401F4; Sat, 26 Oct 2024 11:57:45 +0800 (CST) Received: from huawei.com (10.175.101.6) by dggpemf500002.china.huawei.com (7.185.36.57) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.2.1544.11; Sat, 26 Oct 2024 11:57:44 +0800 From: Yue Haibing To: , , , , , , , , , , , , , , , , , CC: , , , , Subject: [PATCH v4 net-next 4/4] ixgbevf: Fix passing 0 to ERR_PTR in ixgbevf_run_xdp() Date: Sat, 26 Oct 2024 12:12:49 +0800 Message-ID: <20241026041249.1267664-5-yuehaibing@huawei.com> X-Mailer: git-send-email 2.34.1 In-Reply-To: <20241026041249.1267664-1-yuehaibing@huawei.com> References: <20241026041249.1267664-1-yuehaibing@huawei.com> Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable X-ClientProxiedBy: dggems702-chm.china.huawei.com (10.3.19.179) To dggpemf500002.china.huawei.com (7.185.36.57) Content-Type: text/plain; charset="utf-8" ixgbevf_run_xdp() converts customed xdp action to a negative error code with the sk_buff pointer type which be checked with IS_ERR in ixgbevf_clean_rx_irq(). Remove this error pointer handing instead use plain int return value. Fixes: c7aec59657b6 ("ixgbevf: Add XDP support for pass and drop actions") Reviewed-by: Jacob Keller Reviewed-by: Maciej Fijalkowski Signed-off-by: Yue Haibing Reviewed-by: Simon Horman Tested-by: Chandan Kumar Rout (A Contingent Worke= r at Intel) --- .../net/ethernet/intel/ixgbevf/ixgbevf_main.c | 23 ++++++++----------- 1 file changed, 10 insertions(+), 13 deletions(-) diff --git a/drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c b/drivers/ne= t/ethernet/intel/ixgbevf/ixgbevf_main.c index 149911e3002a..183d2305d058 100644 --- a/drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c +++ b/drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c @@ -732,10 +732,6 @@ static bool ixgbevf_cleanup_headers(struct ixgbevf_rin= g *rx_ring, union ixgbe_adv_rx_desc *rx_desc, struct sk_buff *skb) { - /* XDP packets use error pointer so abort at this point */ - if (IS_ERR(skb)) - return true; - /* verify that the packet does not have any known errors */ if (unlikely(ixgbevf_test_staterr(rx_desc, IXGBE_RXDADV_ERR_FRAME_ERR_MASK))) { @@ -1044,9 +1040,9 @@ static int ixgbevf_xmit_xdp_ring(struct ixgbevf_ring = *ring, return IXGBEVF_XDP_TX; } =20 -static struct sk_buff *ixgbevf_run_xdp(struct ixgbevf_adapter *adapter, - struct ixgbevf_ring *rx_ring, - struct xdp_buff *xdp) +static int ixgbevf_run_xdp(struct ixgbevf_adapter *adapter, + struct ixgbevf_ring *rx_ring, + struct xdp_buff *xdp) { int result =3D IXGBEVF_XDP_PASS; struct ixgbevf_ring *xdp_ring; @@ -1080,7 +1076,7 @@ static struct sk_buff *ixgbevf_run_xdp(struct ixgbevf= _adapter *adapter, break; } xdp_out: - return ERR_PTR(-result); + return result; } =20 static unsigned int ixgbevf_rx_frame_truesize(struct ixgbevf_ring *rx_ring, @@ -1122,6 +1118,7 @@ static int ixgbevf_clean_rx_irq(struct ixgbevf_q_vect= or *q_vector, struct sk_buff *skb =3D rx_ring->skb; bool xdp_xmit =3D false; struct xdp_buff xdp; + int xdp_res =3D 0; =20 /* Frame size depend on rx_ring setup when PAGE_SIZE=3D4K */ #if (PAGE_SIZE < 8192) @@ -1165,11 +1162,11 @@ static int ixgbevf_clean_rx_irq(struct ixgbevf_q_ve= ctor *q_vector, /* At larger PAGE_SIZE, frame_sz depend on len size */ xdp.frame_sz =3D ixgbevf_rx_frame_truesize(rx_ring, size); #endif - skb =3D ixgbevf_run_xdp(adapter, rx_ring, &xdp); + xdp_res =3D ixgbevf_run_xdp(adapter, rx_ring, &xdp); } =20 - if (IS_ERR(skb)) { - if (PTR_ERR(skb) =3D=3D -IXGBEVF_XDP_TX) { + if (xdp_res) { + if (xdp_res =3D=3D IXGBEVF_XDP_TX) { xdp_xmit =3D true; ixgbevf_rx_buffer_flip(rx_ring, rx_buffer, size); @@ -1189,7 +1186,7 @@ static int ixgbevf_clean_rx_irq(struct ixgbevf_q_vect= or *q_vector, } =20 /* exit if we failed to retrieve a buffer */ - if (!skb) { + if (!xdp_res && !skb) { rx_ring->rx_stats.alloc_rx_buff_failed++; rx_buffer->pagecnt_bias++; break; @@ -1203,7 +1200,7 @@ static int ixgbevf_clean_rx_irq(struct ixgbevf_q_vect= or *q_vector, continue; =20 /* verify the packet layout is correct */ - if (ixgbevf_cleanup_headers(rx_ring, rx_desc, skb)) { + if (xdp_res || ixgbevf_cleanup_headers(rx_ring, rx_desc, skb)) { skb =3D NULL; continue; } --=20 2.34.1