From nobody Fri Oct 10 23:13:06 2025 Received: from mail-wr1-f49.google.com (mail-wr1-f49.google.com [209.85.221.49]) (using TLSv1.2 with cipher ECDHE-RSA-AES128-GCM-SHA256 (128/128 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 733D62701AB for ; Thu, 12 Jun 2025 15:00:13 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=209.85.221.49 ARC-Seal: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1749740416; cv=none; b=lQOp1I3c2l/sV8OBNXjw7XIU1D28tAFpuveMP/nX7mLpIy/nkLQC6Ak2/Z+u3kCMYZTrOOuPc9xAo+0cdJR+UOCr/TXE+4fGw5E6rxULQPaH/WeNVz25eIKd2apSTCQ0LOmsq+QnHXW/TWAvFVAW336jlqOGoWMksmPpa1YHwjs= ARC-Message-Signature: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1749740416; c=relaxed/simple; bh=gjOtxRDgxx8Tt0j75yME+dF5SDkmuxsE4rL0G/YMYlo=; h=From:To:Cc:Subject:Date:Message-ID:In-Reply-To:References: MIME-Version:Content-Type; b=tdeBPBTUGPkQVyYLHyc/Vui2gljUoEoNLoITCU1GRGGBqMd2tyUuFUJmxTLFMPySDMU662VRIeYOrXl1jaXlJ5vnzBIWzUaZDU/cxDTqjIKX/+/LJXafg60MaemfwpUvSotXCyZktIo+WExR00n8wyjencUhFdqpSM1K0qCSH/8= ARC-Authentication-Results: i=1; smtp.subspace.kernel.org; dmarc=none (p=none dis=none) header.from=ventanamicro.com; spf=pass smtp.mailfrom=ventanamicro.com; dkim=pass (2048-bit key) header.d=ventanamicro.com header.i=@ventanamicro.com header.b=G6EG9hC/; arc=none smtp.client-ip=209.85.221.49 Authentication-Results: smtp.subspace.kernel.org; dmarc=none (p=none dis=none) header.from=ventanamicro.com Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=ventanamicro.com Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=ventanamicro.com header.i=@ventanamicro.com header.b="G6EG9hC/" Received: by mail-wr1-f49.google.com with SMTP id ffacd0b85a97d-3a4ebbfb18fso50287f8f.3 for ; Thu, 12 Jun 2025 08:00:13 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=ventanamicro.com; s=google; t=1749740411; x=1750345211; darn=vger.kernel.org; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:from:to:cc:subject:date :message-id:reply-to; bh=YcQYAB/ZlkNpFLEJF5CWOp3cdYZoBXO0yKKyITW0/lA=; b=G6EG9hC/Sk9nR42jSs7IxZwhhmhZxGeQqwxoetw6ElbE6Iac89X/CY5q+IIwAkM3ya 6wxr8xqptckY2BaWYlVlowhXxEMMkRWdpAz72fNsu7zsaFRMst3FcxBFzb/Z1BXljHM8 pC44MbctFCpt8zw3T0GzzfaFsrn5ALbA0zofkEi3cOekvdzkD33gBOGShFQazJCUIXKa bSWD5rpvAx4lL/CBi1reaqvOd8maJ4UdBY9+NNdny6YwKtLPjzKpZryo1zjNzAH+UAOP nliJDnt8yyXGA/76mTmSV6T5iP5SHegF60sf0YisgFrXVulfLWq1ImuQIGGq2RWuBw9E +xew== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1749740412; x=1750345212; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:x-gm-message-state:from:to:cc :subject:date:message-id:reply-to; bh=YcQYAB/ZlkNpFLEJF5CWOp3cdYZoBXO0yKKyITW0/lA=; b=OUMGI85qcXdtU8Was0gRtiokXLaoEhDD1gPXrT/4iQ6EuQcVtGRla8qxElcZ5Q6Qf5 rw4oyiP2yQx9nrxqNcfsBGW2lET66a2As2zGTc4ZfHpEUkNPY7sWiJfODePKDJlPw/+H SetvjXlHsNvBHLnhmMl0V+Js/qsjComUYk597SZmkskp6NGmXCCaXpX7/ojbZCry3uew sToCPnKbgSartD2hP6DUsGfk7mMaXYNUki83RBcb0v5e1M0o3ZTfzWVOx08p8oFBHULb SWGL1gqTlh6ud+AktON1O5QLg63PBgSD2PhYrXGfNwpJ/ThnX4wGs1kIoON1MbdgORVh Smaw== X-Gm-Message-State: AOJu0YwFB18jCGjcorRxeDyPEqXcFXGnjRqg/yOhnmqAL+8ERT4ZhYkf tuM7OjXQvhJ7zTEg4E79T0nJOTCGisAyHpTIV578ndm3b/bd97Vu1qL5jTB78wL9I/c= X-Gm-Gg: ASbGnctJUUHFdXyRC+kCs5r8LpU7MkCIOmeK/YovaKFG0dmBDXM6BrWQw+Q32WGLsEe RPxGtUgOyuotTgeM4JuZwiTFFVEuMRiIuUCEd4uUrBwu3lK9XNFTP3podSZH+TsBYUNH6pmcwpd oJ1scAJP/ReOjvZAF09GtODKcLUk3NjIWqAKKHaYBDaTvhx01SDuIetzxTsgHD2pLWX/RWwUjP8 xvB9xOEEXC7IGBnhgX7E1RT4KSxEe3tsukJjIsG+B+D8qkyMxuc+wsda/8tgLc8peIhC7Ol0Gw8 LAojbzlWABLsA6eaLxFkyLMAMZ4tOMDdFw5aI9HGMZNALoMU5gYx9DFxyKzEUaXhadjQ2A== X-Google-Smtp-Source: AGHT+IEdv88UTi+FoedSsSS0XwV7MjztONQlfWNB/7ITxC5u6gxePJnjvgxuWV/3jUIF07rsxOlOCA== X-Received: by 2002:adf:b60c:0:b0:3a4:f912:86b2 with SMTP id ffacd0b85a97d-3a55880311dmr2172489f8f.11.1749740411478; Thu, 12 Jun 2025 08:00:11 -0700 (PDT) Received: from localhost ([2a02:8308:a00c:e200:e05f:7af3:1bf9:f101]) by smtp.gmail.com with ESMTPSA id 5b1f17b1804b1-4532e2354fbsm22860915e9.15.2025.06.12.08.00.10 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Thu, 12 Jun 2025 08:00:10 -0700 (PDT) From: =?UTF-8?q?Radim=20Kr=C4=8Dm=C3=A1=C5=99?= To: linux-riscv@lists.infradead.org Cc: linux-kernel@vger.kernel.org, Paul Walmsley , Palmer Dabbelt , Albert Ou , Alexandre Ghiti , Atish Patra , Andrew Jones , =?UTF-8?q?Cl=C3=A9ment=20L=C3=A9ger?= , Anup Patel Subject: [PATCH 2/2] RISC-V: make use of variadic sbi_ecall Date: Thu, 12 Jun 2025 16:57:55 +0200 Message-ID: <20250612145754.2126147-4-rkrcmar@ventanamicro.com> X-Mailer: git-send-email 2.49.0 In-Reply-To: <20250612145754.2126147-2-rkrcmar@ventanamicro.com> References: <20250612145754.2126147-2-rkrcmar@ventanamicro.com> Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Type: text/plain; charset="utf-8" Content-Transfer-Encoding: quoted-printable The new sbi_ecall doesn't have to list all 8 arguments anymore, so only pass the actual numbers of arguments for each SBI function. Trailing 0 is sometimes intentional. Signed-off-by: Radim Kr=C4=8Dm=C3=A1=C5=99 --- arch/riscv/include/asm/kvm_nacl.h | 4 +-- arch/riscv/kernel/cpu_ops_sbi.c | 6 ++-- arch/riscv/kernel/paravirt.c | 2 +- arch/riscv/kernel/sbi.c | 57 ++++++++++++++----------------- arch/riscv/kernel/sbi_ecall.c | 2 +- arch/riscv/kernel/suspend.c | 4 +-- arch/riscv/kvm/nacl.c | 7 ++-- drivers/acpi/riscv/cppc.c | 4 +-- drivers/perf/riscv_pmu_sbi.c | 48 +++++++++++++------------- 9 files changed, 65 insertions(+), 69 deletions(-) diff --git a/arch/riscv/include/asm/kvm_nacl.h b/arch/riscv/include/asm/kvm= _nacl.h index 4124d5e06a0f..5849af11c945 100644 --- a/arch/riscv/include/asm/kvm_nacl.h +++ b/arch/riscv/include/asm/kvm_nacl.h @@ -96,7 +96,7 @@ do { \ =20 #define nacl_sync_hfence(__e) \ sbi_ecall(SBI_EXT_NACL, SBI_EXT_NACL_SYNC_HFENCE, \ - (__e), 0, 0, 0, 0, 0) + (__e)) =20 #define nacl_hfence_mkconfig(__type, __order, __vmid, __asid) \ ({ \ @@ -196,7 +196,7 @@ do { \ =20 #define nacl_sync_csr(__csr) \ sbi_ecall(SBI_EXT_NACL, SBI_EXT_NACL_SYNC_CSR, \ - (__csr), 0, 0, 0, 0, 0) + (__csr)) =20 /* * Each ncsr_xyz() macro defined below has it's own static-branch so every diff --git a/arch/riscv/kernel/cpu_ops_sbi.c b/arch/riscv/kernel/cpu_ops_sb= i.c index e6fbaaf54956..d5de532ca082 100644 --- a/arch/riscv/kernel/cpu_ops_sbi.c +++ b/arch/riscv/kernel/cpu_ops_sbi.c @@ -29,7 +29,7 @@ static int sbi_hsm_hart_start(unsigned long hartid, unsig= ned long saddr, struct sbiret ret; =20 ret =3D sbi_ecall(SBI_EXT_HSM, SBI_EXT_HSM_HART_START, - hartid, saddr, priv, 0, 0, 0); + hartid, saddr, priv); if (ret.error) return sbi_err_map_linux_errno(ret.error); else @@ -41,7 +41,7 @@ static int sbi_hsm_hart_stop(void) { struct sbiret ret; =20 - ret =3D sbi_ecall(SBI_EXT_HSM, SBI_EXT_HSM_HART_STOP, 0, 0, 0, 0, 0, 0); + ret =3D sbi_ecall(SBI_EXT_HSM, SBI_EXT_HSM_HART_STOP); =20 if (ret.error) return sbi_err_map_linux_errno(ret.error); @@ -54,7 +54,7 @@ static int sbi_hsm_hart_get_status(unsigned long hartid) struct sbiret ret; =20 ret =3D sbi_ecall(SBI_EXT_HSM, SBI_EXT_HSM_HART_STATUS, - hartid, 0, 0, 0, 0, 0); + hartid); if (ret.error) return sbi_err_map_linux_errno(ret.error); else diff --git a/arch/riscv/kernel/paravirt.c b/arch/riscv/kernel/paravirt.c index fa6b0339a65d..9d00743a96c1 100644 --- a/arch/riscv/kernel/paravirt.c +++ b/arch/riscv/kernel/paravirt.c @@ -60,7 +60,7 @@ static int sbi_sta_steal_time_set_shmem(unsigned long lo,= unsigned long hi, struct sbiret ret; =20 ret =3D sbi_ecall(SBI_EXT_STA, SBI_EXT_STA_STEAL_TIME_SET_SHMEM, - lo, hi, flags, 0, 0, 0); + lo, hi, flags); if (ret.error) { if (lo =3D=3D SBI_SHMEM_DISABLE && hi =3D=3D SBI_SHMEM_DISABLE) pr_warn("Failed to disable steal-time shmem"); diff --git a/arch/riscv/kernel/sbi.c b/arch/riscv/kernel/sbi.c index 53836a9235e3..c2401efd84e5 100644 --- a/arch/riscv/kernel/sbi.c +++ b/arch/riscv/kernel/sbi.c @@ -57,7 +57,7 @@ static unsigned long __sbi_v01_cpumask_to_hartmask(const = struct cpumask *cpu_mas */ void sbi_console_putchar(int ch) { - sbi_ecall(SBI_EXT_0_1_CONSOLE_PUTCHAR, 0, ch, 0, 0, 0, 0, 0); + sbi_ecall(SBI_EXT_0_1_CONSOLE_PUTCHAR, 0, ch); } EXPORT_SYMBOL(sbi_console_putchar); =20 @@ -70,7 +70,7 @@ int sbi_console_getchar(void) { struct sbiret ret; =20 - ret =3D sbi_ecall(SBI_EXT_0_1_CONSOLE_GETCHAR, 0, 0, 0, 0, 0, 0, 0); + ret =3D sbi_ecall(SBI_EXT_0_1_CONSOLE_GETCHAR); =20 return ret.error; } @@ -83,7 +83,7 @@ EXPORT_SYMBOL(sbi_console_getchar); */ void sbi_shutdown(void) { - sbi_ecall(SBI_EXT_0_1_SHUTDOWN, 0, 0, 0, 0, 0, 0, 0); + sbi_ecall(SBI_EXT_0_1_SHUTDOWN); } EXPORT_SYMBOL(sbi_shutdown); =20 @@ -97,9 +97,9 @@ static void __sbi_set_timer_v01(uint64_t stime_value) { #if __riscv_xlen =3D=3D 32 sbi_ecall(SBI_EXT_0_1_SET_TIMER, 0, stime_value, - stime_value >> 32, 0, 0, 0, 0); + stime_value >> 32); #else - sbi_ecall(SBI_EXT_0_1_SET_TIMER, 0, stime_value, 0, 0, 0, 0, 0); + sbi_ecall(SBI_EXT_0_1_SET_TIMER, 0, stime_value); #endif } =20 @@ -107,8 +107,7 @@ static void __sbi_send_ipi_v01(unsigned int cpu) { unsigned long hart_mask =3D __sbi_v01_cpumask_to_hartmask(cpumask_of(cpu)); - sbi_ecall(SBI_EXT_0_1_SEND_IPI, 0, (unsigned long)(&hart_mask), - 0, 0, 0, 0, 0); + sbi_ecall(SBI_EXT_0_1_SEND_IPI, 0, (unsigned long)(&hart_mask)); } =20 static int __sbi_rfence_v01(int fid, const struct cpumask *cpu_mask, @@ -126,17 +125,16 @@ static int __sbi_rfence_v01(int fid, const struct cpu= mask *cpu_mask, switch (fid) { case SBI_EXT_RFENCE_REMOTE_FENCE_I: sbi_ecall(SBI_EXT_0_1_REMOTE_FENCE_I, 0, - (unsigned long)&hart_mask, 0, 0, 0, 0, 0); + (unsigned long)&hart_mask); break; case SBI_EXT_RFENCE_REMOTE_SFENCE_VMA: sbi_ecall(SBI_EXT_0_1_REMOTE_SFENCE_VMA, 0, - (unsigned long)&hart_mask, start, size, - 0, 0, 0); + (unsigned long)&hart_mask, start, size); break; case SBI_EXT_RFENCE_REMOTE_SFENCE_VMA_ASID: sbi_ecall(SBI_EXT_0_1_REMOTE_SFENCE_VMA_ASID, 0, (unsigned long)&hart_mask, start, size, - arg4, 0, 0); + arg4); break; default: pr_err("SBI call [%d]not supported in SBI v0.1\n", fid); @@ -180,10 +178,9 @@ static void __sbi_set_timer_v02(uint64_t stime_value) { #if __riscv_xlen =3D=3D 32 sbi_ecall(SBI_EXT_TIME, SBI_EXT_TIME_SET_TIMER, stime_value, - stime_value >> 32, 0, 0, 0, 0); + stime_value >> 32); #else - sbi_ecall(SBI_EXT_TIME, SBI_EXT_TIME_SET_TIMER, stime_value, 0, - 0, 0, 0, 0); + sbi_ecall(SBI_EXT_TIME, SBI_EXT_TIME_SET_TIMER, stime_value); #endif } =20 @@ -193,7 +190,7 @@ static void __sbi_send_ipi_v02(unsigned int cpu) struct sbiret ret =3D {0}; =20 ret =3D sbi_ecall(SBI_EXT_IPI, SBI_EXT_IPI_SEND_IPI, - 1UL, cpuid_to_hartid_map(cpu), 0, 0, 0, 0); + 1UL, cpuid_to_hartid_map(cpu)); if (ret.error) { result =3D sbi_err_map_linux_errno(ret.error); pr_err("%s: hbase =3D [%lu] failed (error [%d])\n", @@ -212,32 +209,32 @@ static int __sbi_rfence_v02_call(unsigned long fid, u= nsigned long hmask, =20 switch (fid) { case SBI_EXT_RFENCE_REMOTE_FENCE_I: - ret =3D sbi_ecall(ext, fid, hmask, hbase, 0, 0, 0, 0); + ret =3D sbi_ecall(ext, fid, hmask, hbase); break; case SBI_EXT_RFENCE_REMOTE_SFENCE_VMA: ret =3D sbi_ecall(ext, fid, hmask, hbase, start, - size, 0, 0); + size); break; case SBI_EXT_RFENCE_REMOTE_SFENCE_VMA_ASID: ret =3D sbi_ecall(ext, fid, hmask, hbase, start, - size, arg4, 0); + size, arg4); break; =20 case SBI_EXT_RFENCE_REMOTE_HFENCE_GVMA: ret =3D sbi_ecall(ext, fid, hmask, hbase, start, - size, 0, 0); + size); break; case SBI_EXT_RFENCE_REMOTE_HFENCE_GVMA_VMID: ret =3D sbi_ecall(ext, fid, hmask, hbase, start, - size, arg4, 0); + size, arg4); break; case SBI_EXT_RFENCE_REMOTE_HFENCE_VVMA: ret =3D sbi_ecall(ext, fid, hmask, hbase, start, - size, 0, 0); + size); break; case SBI_EXT_RFENCE_REMOTE_HFENCE_VVMA_ASID: ret =3D sbi_ecall(ext, fid, hmask, hbase, start, - size, arg4, 0); + size, arg4); break; default: pr_err("unknown function ID [%lu] for SBI extension [%d]\n", @@ -334,7 +331,7 @@ int sbi_fwft_set(u32 feature, unsigned long value, unsi= gned long flags) return -EOPNOTSUPP; =20 ret =3D sbi_ecall(SBI_EXT_FWFT, SBI_EXT_FWFT_SET, - feature, value, flags, 0, 0, 0); + feature, value, flags); =20 return sbi_err_map_linux_errno(ret.error); } @@ -510,8 +507,7 @@ EXPORT_SYMBOL(sbi_remote_hfence_vvma_asid); =20 static void sbi_srst_reset(unsigned long type, unsigned long reason) { - sbi_ecall(SBI_EXT_SRST, SBI_EXT_SRST_RESET, type, reason, - 0, 0, 0, 0); + sbi_ecall(SBI_EXT_SRST, SBI_EXT_SRST_RESET, type, reason); pr_warn("%s: type=3D0x%lx reason=3D0x%lx failed\n", __func__, type, reason); } @@ -544,8 +540,7 @@ long sbi_probe_extension(int extid) { struct sbiret ret; =20 - ret =3D sbi_ecall(SBI_EXT_BASE, SBI_EXT_BASE_PROBE_EXT, extid, - 0, 0, 0, 0, 0); + ret =3D sbi_ecall(SBI_EXT_BASE, SBI_EXT_BASE_PROBE_EXT, extid); if (!ret.error) return ret.value; =20 @@ -607,10 +602,10 @@ int sbi_debug_console_write(const char *bytes, unsign= ed int num_bytes) if (IS_ENABLED(CONFIG_32BIT)) ret =3D sbi_ecall(SBI_EXT_DBCN, SBI_EXT_DBCN_CONSOLE_WRITE, num_bytes, lower_32_bits(base_addr), - upper_32_bits(base_addr), 0, 0, 0); + upper_32_bits(base_addr)); else ret =3D sbi_ecall(SBI_EXT_DBCN, SBI_EXT_DBCN_CONSOLE_WRITE, - num_bytes, base_addr, 0, 0, 0, 0); + num_bytes, base_addr); =20 if (ret.error =3D=3D SBI_ERR_FAILURE) return -EIO; @@ -636,10 +631,10 @@ int sbi_debug_console_read(char *bytes, unsigned int = num_bytes) if (IS_ENABLED(CONFIG_32BIT)) ret =3D sbi_ecall(SBI_EXT_DBCN, SBI_EXT_DBCN_CONSOLE_READ, num_bytes, lower_32_bits(base_addr), - upper_32_bits(base_addr), 0, 0, 0); + upper_32_bits(base_addr)); else ret =3D sbi_ecall(SBI_EXT_DBCN, SBI_EXT_DBCN_CONSOLE_READ, - num_bytes, base_addr, 0, 0, 0, 0); + num_bytes, base_addr); =20 if (ret.error =3D=3D SBI_ERR_FAILURE) return -EIO; diff --git a/arch/riscv/kernel/sbi_ecall.c b/arch/riscv/kernel/sbi_ecall.c index 24aabb4fbde3..714a16103cd7 100644 --- a/arch/riscv/kernel/sbi_ecall.c +++ b/arch/riscv/kernel/sbi_ecall.c @@ -9,7 +9,7 @@ long __sbi_base_ecall(int fid) { struct sbiret ret; =20 - ret =3D sbi_ecall(SBI_EXT_BASE, fid, 0, 0, 0, 0, 0, 0); + ret =3D sbi_ecall(SBI_EXT_BASE, fid); if (!ret.error) return ret.value; else diff --git a/arch/riscv/kernel/suspend.c b/arch/riscv/kernel/suspend.c index 24b3f57d467f..bef83043e8c7 100644 --- a/arch/riscv/kernel/suspend.c +++ b/arch/riscv/kernel/suspend.c @@ -115,7 +115,7 @@ static int sbi_system_suspend(unsigned long sleep_type, struct sbiret ret; =20 ret =3D sbi_ecall(SBI_EXT_SUSP, SBI_EXT_SUSP_SYSTEM_SUSPEND, - sleep_type, resume_addr, opaque, 0, 0, 0); + sleep_type, resume_addr, opaque); if (ret.error) return sbi_err_map_linux_errno(ret.error); =20 @@ -153,7 +153,7 @@ static int sbi_suspend_finisher(unsigned long suspend_t= ype, struct sbiret ret; =20 ret =3D sbi_ecall(SBI_EXT_HSM, SBI_EXT_HSM_HART_SUSPEND, - suspend_type, resume_addr, opaque, 0, 0, 0); + suspend_type, resume_addr, opaque); =20 return (ret.error) ? sbi_err_map_linux_errno(ret.error) : 0; } diff --git a/arch/riscv/kvm/nacl.c b/arch/riscv/kvm/nacl.c index 08a95ad9ada2..bc0ea3157645 100644 --- a/arch/riscv/kvm/nacl.c +++ b/arch/riscv/kvm/nacl.c @@ -61,7 +61,7 @@ int kvm_riscv_nacl_enable(void) nacl =3D this_cpu_ptr(&kvm_riscv_nacl); =20 ret =3D sbi_ecall(SBI_EXT_NACL, SBI_EXT_NACL_SET_SHMEM, - nacl->shmem_phys, 0, 0, 0, 0, 0); + nacl->shmem_phys, 0, 0); rc =3D sbi_err_map_linux_errno(ret.error); if (rc) return rc; @@ -75,7 +75,7 @@ void kvm_riscv_nacl_disable(void) return; =20 sbi_ecall(SBI_EXT_NACL, SBI_EXT_NACL_SET_SHMEM, - SBI_SHMEM_DISABLE, SBI_SHMEM_DISABLE, 0, 0, 0, 0); + SBI_SHMEM_DISABLE, SBI_SHMEM_DISABLE, 0); } =20 void kvm_riscv_nacl_exit(void) @@ -106,8 +106,7 @@ static long nacl_probe_feature(long feature_id) if (!kvm_riscv_nacl_available()) return 0; =20 - ret =3D sbi_ecall(SBI_EXT_NACL, SBI_EXT_NACL_PROBE_FEATURE, - feature_id, 0, 0, 0, 0, 0); + ret =3D sbi_ecall(SBI_EXT_NACL, SBI_EXT_NACL_PROBE_FEATURE, feature_id); return ret.value; } =20 diff --git a/drivers/acpi/riscv/cppc.c b/drivers/acpi/riscv/cppc.c index 4cdff387deff..21e2051b3e35 100644 --- a/drivers/acpi/riscv/cppc.c +++ b/drivers/acpi/riscv/cppc.c @@ -53,7 +53,7 @@ static void sbi_cppc_read(void *read_data) struct sbi_cppc_data *data =3D (struct sbi_cppc_data *)read_data; =20 data->ret =3D sbi_ecall(SBI_EXT_CPPC, SBI_CPPC_READ, - data->reg, 0, 0, 0, 0, 0); + data->reg); } =20 static void sbi_cppc_write(void *write_data) @@ -61,7 +61,7 @@ static void sbi_cppc_write(void *write_data) struct sbi_cppc_data *data =3D (struct sbi_cppc_data *)write_data; =20 data->ret =3D sbi_ecall(SBI_EXT_CPPC, SBI_CPPC_WRITE, - data->reg, data->val, 0, 0, 0, 0); + data->reg, data->val); } =20 static void cppc_ffh_csr_read(void *read_data) diff --git a/drivers/perf/riscv_pmu_sbi.c b/drivers/perf/riscv_pmu_sbi.c index 698de8ddf895..752096b1ef9a 100644 --- a/drivers/perf/riscv_pmu_sbi.c +++ b/drivers/perf/riscv_pmu_sbi.c @@ -303,10 +303,10 @@ static void pmu_sbi_check_event(struct sbi_pmu_event_= data *edata) struct sbiret ret; =20 ret =3D sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_CFG_MATCH, - 0, cmask, 0, edata->event_idx, 0, 0); + 0, cmask, 0, edata->event_idx, 0); if (!ret.error) { sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_STOP, - ret.value, 0x1, SBI_PMU_STOP_FLAG_RESET, 0, 0, 0); + ret.value, 0x1, SBI_PMU_STOP_FLAG_RESET); } else if (ret.error =3D=3D SBI_ERR_NOT_SUPPORTED) { /* This event cannot be monitored by any counter */ edata->event_idx =3D -ENOENT; @@ -433,7 +433,7 @@ static int pmu_sbi_ctr_get_idx(struct perf_event *event) hwc->config >> 32); #else ret =3D sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_CFG_MATCH, cbase, - cmask, cflags, hwc->event_base, hwc->config, 0); + cmask, cflags, hwc->event_base, hwc->config); #endif if (ret.error) { pr_debug("Not able to find a counter for event %lx config %llx\n", @@ -606,7 +606,7 @@ static int pmu_sbi_snapshot_disable(void) struct sbiret ret; =20 ret =3D sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_SNAPSHOT_SET_SHMEM, SBI_SHMEM_= DISABLE, - SBI_SHMEM_DISABLE, 0, 0, 0, 0); + SBI_SHMEM_DISABLE); if (ret.error) { pr_warn("failed to disable snapshot shared memory\n"); return sbi_err_map_linux_errno(ret.error); @@ -630,10 +630,10 @@ static int pmu_sbi_snapshot_setup(struct riscv_pmu *p= mu, int cpu) if (IS_ENABLED(CONFIG_32BIT)) ret =3D sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_SNAPSHOT_SET_SHMEM, cpu_hw_evt->snapshot_addr_phys, - (u64)(cpu_hw_evt->snapshot_addr_phys) >> 32, 0, 0, 0, 0); + (u64)(cpu_hw_evt->snapshot_addr_phys) >> 32); else ret =3D sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_SNAPSHOT_SET_SHMEM, - cpu_hw_evt->snapshot_addr_phys, 0, 0, 0, 0, 0); + cpu_hw_evt->snapshot_addr_phys); =20 /* Free up the snapshot area memory and fall back to SBI PMU calls withou= t snapshot */ if (ret.error) { @@ -667,14 +667,14 @@ static u64 pmu_sbi_ctr_read(struct perf_event *event) =20 if (pmu_sbi_is_fw_event(event)) { ret =3D sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_FW_READ, - hwc->idx, 0, 0, 0, 0, 0); + hwc->idx); if (ret.error) return 0; =20 val =3D ret.value; if (IS_ENABLED(CONFIG_32BIT) && sbi_v2_available && info.width >=3D 32) { ret =3D sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_FW_READ_HI, - hwc->idx, 0, 0, 0, 0, 0); + hwc->idx); if (!ret.error) val |=3D ((u64)ret.value << 32); else @@ -717,10 +717,10 @@ static void pmu_sbi_ctr_start(struct perf_event *even= t, u64 ival) /* There is no benefit setting SNAPSHOT FLAG for a single counter */ #if defined(CONFIG_32BIT) ret =3D sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_START, hwc->idx, - 1, flag, ival, ival >> 32, 0); + 1, flag, ival, ival >> 32); #else ret =3D sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_START, hwc->idx, - 1, flag, ival, 0, 0); + 1, flag, ival); #endif if (ret.error && (ret.error !=3D SBI_ERR_ALREADY_STARTED)) pr_err("Starting counter idx %d failed with error %d\n", @@ -746,7 +746,8 @@ static void pmu_sbi_ctr_stop(struct perf_event *event, = unsigned long flag) if (sbi_pmu_snapshot_available()) flag |=3D SBI_PMU_STOP_FLAG_TAKE_SNAPSHOT; =20 - ret =3D sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_STOP, hwc->idx, 1, fla= g, 0, 0, 0); + ret =3D sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_STOP, + hwc->idx, 1, flag); if (!ret.error && sbi_pmu_snapshot_available()) { /* * The counter snapshot is based on the index base specified by hwc->idx. @@ -771,7 +772,7 @@ static int pmu_sbi_find_num_ctrs(void) { struct sbiret ret; =20 - ret =3D sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_NUM_COUNTERS, 0, 0, 0, 0, 0, 0= ); + ret =3D sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_NUM_COUNTERS); if (!ret.error) return ret.value; else @@ -789,7 +790,7 @@ static int pmu_sbi_get_ctrinfo(int nctr, unsigned long = *mask) return -ENOMEM; =20 for (i =3D 0; i < nctr; i++) { - ret =3D sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_GET_INFO, i, 0, 0, 0,= 0, 0); + ret =3D sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_GET_INFO, i); if (ret.error) /* The logical counter ids are not expected to be contiguous */ continue; @@ -816,7 +817,7 @@ static inline void pmu_sbi_stop_all(struct riscv_pmu *p= mu) * which may include counters that are not enabled yet. */ sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_STOP, - 0, pmu->cmask, SBI_PMU_STOP_FLAG_RESET, 0, 0, 0); + 0, pmu->cmask, SBI_PMU_STOP_FLAG_RESET); } =20 static inline void pmu_sbi_stop_hw_ctrs(struct riscv_pmu *pmu) @@ -837,7 +838,7 @@ static inline void pmu_sbi_stop_hw_ctrs(struct riscv_pm= u *pmu) for (i =3D 0; i < BITS_TO_LONGS(RISCV_MAX_COUNTERS); i++) { /* No need to check the error here as we can't do anything about the err= or */ ret =3D sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_STOP, i * BITS_PER_LO= NG, - cpu_hw_evt->used_hw_ctrs[i], flag, 0, 0, 0); + cpu_hw_evt->used_hw_ctrs[i], flag); if (!ret.error && sbi_pmu_snapshot_available()) { /* Save the counter values to avoid clobbering */ for_each_set_bit(idx, &cpu_hw_evt->used_hw_ctrs[i], BITS_PER_LONG) @@ -877,8 +878,8 @@ static inline void pmu_sbi_start_ovf_ctrs_sbi(struct cp= u_hw_events *cpu_hw_evt, for (i =3D 0; i < BITS_TO_LONGS(RISCV_MAX_COUNTERS); i++) { ctr_start_mask =3D cpu_hw_evt->used_hw_ctrs[i] & ~ctr_ovf_mask; /* Start all the counters that did not overflow in a single shot */ - sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_START, i * BITS_PER_LONG, ctr= _start_mask, - 0, 0, 0, 0); + sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_START, + i * BITS_PER_LONG, ctr_start_mask, 0); } =20 /* Reinitialize and start all the counter that overflowed */ @@ -889,11 +890,11 @@ static inline void pmu_sbi_start_ovf_ctrs_sbi(struct = cpu_hw_events *cpu_hw_evt, max_period =3D riscv_pmu_ctr_get_width_mask(event); init_val =3D local64_read(&hwc->prev_count) & max_period; #if defined(CONFIG_32BIT) - sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_START, idx, 1, - flag, init_val, init_val >> 32, 0); + sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_START, idx, + 1, flag, init_val, init_val >> 32); #else - sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_START, idx, 1, - flag, init_val, 0, 0); + sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_START, idx, + 1, flag, init_val); #endif perf_event_update_userpage(event); } @@ -932,8 +933,9 @@ static inline void pmu_sbi_start_ovf_ctrs_snapshot(stru= ct cpu_hw_events *cpu_hw_ sdata->ctr_values[idx] =3D cpu_hw_evt->snapshot_cval_shcopy[idx + i * BITS_PER_LONG]; /* Start all the counters in a single shot */ - sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_START, idx * BITS_PER_LONG, - cpu_hw_evt->used_hw_ctrs[i], flag, 0, 0, 0); + sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_START, + idx * BITS_PER_LONG, + cpu_hw_evt->used_hw_ctrs[i], flag); } } =20 --=20 2.49.0