Add FFA_RUN support and remove mention that YIELD and INTERRUPT are not
supported in the comment as we will now return them.
Make FFA_FEATURE answer supported for RUN, YIELD and INTERRUPT.
Share the SMCCC plumbing used by SEND_DIRECT and RUN via a common
ffa_finish_direct_req_run() helper so canonical success and error cases
are handled in one place.
The dispatcher now routes FFA_RUN through ffa_handle_run(), and direct
requests bail out early if a guest targets itself or a non-secure
endpoint. This simplifies the direct path while adding support for
INTERRUPT or YIELD answers from secure world.
Signed-off-by: Bertrand Marquis <bertrand.marquis@arm.com>
---
Changes in v1:
- Rework commit message to clearly mention that we add FFA_RUN support
- Remove mentions of FFA_MSG_YIELD/INTERRUPT/RUN not being supported
- Report YIELD, INTERRUPT an RUN as supported in ffa_features handling
---
xen/arch/arm/tee/ffa.c | 9 ++-
xen/arch/arm/tee/ffa_msg.c | 111 ++++++++++++++++++++++++---------
xen/arch/arm/tee/ffa_private.h | 1 +
3 files changed, 87 insertions(+), 34 deletions(-)
diff --git a/xen/arch/arm/tee/ffa.c b/xen/arch/arm/tee/ffa.c
index 2c09d10ae6a1..0f07efe5a7b3 100644
--- a/xen/arch/arm/tee/ffa.c
+++ b/xen/arch/arm/tee/ffa.c
@@ -24,9 +24,6 @@
* o FFA_MEM_DONATE_* and FFA_MEM_LEND_* - Used when tranferring ownership
* or access of a memory region
* o FFA_MSG_SEND2 and FFA_MSG_WAIT - Used for indirect messaging
- * o FFA_MSG_YIELD
- * o FFA_INTERRUPT - Used to report preemption
- * o FFA_RUN
*
* Limitations in the implemented FF-A interfaces:
* o FFA_RXTX_MAP_*:
@@ -266,6 +263,9 @@ static void handle_features(struct cpu_user_regs *regs)
case FFA_MSG_SEND_DIRECT_REQ_32:
case FFA_MSG_SEND_DIRECT_REQ_64:
case FFA_MSG_SEND2:
+ case FFA_RUN:
+ case FFA_INTERRUPT:
+ case FFA_MSG_YIELD:
ffa_set_regs_success(regs, 0, 0);
break;
case FFA_MEM_SHARE_64:
@@ -353,6 +353,9 @@ static bool ffa_handle_call(struct cpu_user_regs *regs)
case FFA_MSG_SEND_DIRECT_REQ_64:
ffa_handle_msg_send_direct_req(regs, fid);
return true;
+ case FFA_RUN:
+ ffa_handle_run(regs, fid);
+ return true;
case FFA_MSG_SEND2:
e = ffa_handle_msg_send2(regs);
break;
diff --git a/xen/arch/arm/tee/ffa_msg.c b/xen/arch/arm/tee/ffa_msg.c
index d60eed6d8811..5a4cb1bb8295 100644
--- a/xen/arch/arm/tee/ffa_msg.c
+++ b/xen/arch/arm/tee/ffa_msg.c
@@ -21,42 +21,74 @@ struct ffa_part_msg_rxtx {
uint32_t msg_size;
};
-void ffa_handle_msg_send_direct_req(struct cpu_user_regs *regs, uint32_t fid)
+static void ffa_finish_direct_req_run(struct cpu_user_regs *regs,
+ struct arm_smccc_1_2_regs *req)
{
- struct arm_smccc_1_2_regs arg = { .a0 = fid, };
struct arm_smccc_1_2_regs resp = { };
- struct domain *d = current->domain;
- uint32_t src_dst;
uint64_t mask;
- if ( smccc_is_conv_64(fid) )
+ arm_smccc_1_2_smc(req, &resp);
+
+ switch ( resp.a0 )
+ {
+ case FFA_ERROR:
+ case FFA_SUCCESS_32:
+ case FFA_SUCCESS_64:
+ case FFA_MSG_SEND_DIRECT_RESP_32:
+ case FFA_MSG_SEND_DIRECT_RESP_64:
+ case FFA_MSG_YIELD:
+ case FFA_INTERRUPT:
+ break;
+ default:
+ /* Bad fid, report back to the caller. */
+ ffa_set_regs_error(regs, FFA_RET_ABORTED);
+ return;
+ }
+
+ if ( smccc_is_conv_64(resp.a0) )
mask = GENMASK_ULL(63, 0);
else
mask = GENMASK_ULL(31, 0);
+ ffa_set_regs(regs, resp.a0, resp.a1 & mask, resp.a2 & mask, resp.a3 & mask,
+ resp.a4 & mask, resp.a5 & mask, resp.a6 & mask,
+ resp.a7 & mask);
+}
+
+void ffa_handle_msg_send_direct_req(struct cpu_user_regs *regs, uint32_t fid)
+{
+ struct arm_smccc_1_2_regs arg = { .a0 = fid, };
+ struct domain *d = current->domain;
+ uint32_t src_dst;
+ uint64_t mask;
+ int32_t ret;
+
if ( !ffa_fw_supports_fid(fid) )
{
- resp.a0 = FFA_ERROR;
- resp.a2 = FFA_RET_NOT_SUPPORTED;
+ ret = FFA_RET_NOT_SUPPORTED;
goto out;
}
src_dst = get_user_reg(regs, 1);
- if ( (src_dst >> 16) != ffa_get_vm_id(d) )
+ if ( (src_dst >> 16) != ffa_get_vm_id(d) ||
+ (src_dst & GENMASK(15,0)) == ffa_get_vm_id(d) )
{
- resp.a0 = FFA_ERROR;
- resp.a2 = FFA_RET_INVALID_PARAMETERS;
+ ret = FFA_RET_INVALID_PARAMETERS;
goto out;
}
/* we do not support direct messages to VMs */
if ( !FFA_ID_IS_SECURE(src_dst & GENMASK(15,0)) )
{
- resp.a0 = FFA_ERROR;
- resp.a2 = FFA_RET_NOT_SUPPORTED;
+ ret = FFA_RET_NOT_SUPPORTED;
goto out;
}
+ if ( smccc_is_conv_64(fid) )
+ mask = GENMASK_ULL(63, 0);
+ else
+ mask = GENMASK_ULL(31, 0);
+
arg.a1 = src_dst;
arg.a2 = get_user_reg(regs, 2) & mask;
arg.a3 = get_user_reg(regs, 3) & mask;
@@ -65,27 +97,11 @@ void ffa_handle_msg_send_direct_req(struct cpu_user_regs *regs, uint32_t fid)
arg.a6 = get_user_reg(regs, 6) & mask;
arg.a7 = get_user_reg(regs, 7) & mask;
- arm_smccc_1_2_smc(&arg, &resp);
- switch ( resp.a0 )
- {
- case FFA_ERROR:
- case FFA_SUCCESS_32:
- case FFA_SUCCESS_64:
- case FFA_MSG_SEND_DIRECT_RESP_32:
- case FFA_MSG_SEND_DIRECT_RESP_64:
- break;
- default:
- /* Bad fid, report back to the caller. */
- memset(&resp, 0, sizeof(resp));
- resp.a0 = FFA_ERROR;
- resp.a1 = src_dst;
- resp.a2 = FFA_RET_ABORTED;
- }
+ ffa_finish_direct_req_run(regs, &arg);
+ return;
out:
- ffa_set_regs(regs, resp.a0, resp.a1 & mask, resp.a2 & mask, resp.a3 & mask,
- resp.a4 & mask, resp.a5 & mask, resp.a6 & mask,
- resp.a7 & mask);
+ ffa_set_regs_error(regs, ret);
}
static int32_t ffa_msg_send2_vm(uint16_t dst_id, const void *src_buf,
@@ -217,3 +233,36 @@ out:
ffa_tx_release(src_ctx);
return ret;
}
+
+void ffa_handle_run(struct cpu_user_regs *regs, uint32_t fid)
+{
+ struct arm_smccc_1_2_regs arg = { .a0 = fid, };
+ uint32_t dst = get_user_reg(regs, 1);
+ int32_t ret;
+
+ if ( !ffa_fw_supports_fid(fid) )
+ {
+ ret = FFA_RET_NOT_SUPPORTED;
+ goto out;
+ }
+
+ /*
+ * We do not support FFA_RUN to VMs.
+ * Destination endpoint ID is in bits [31:16], bits[15:0] contain the
+ * vCPU ID.
+ */
+ if ( !FFA_ID_IS_SECURE(dst >> 16) )
+ {
+ ret = FFA_RET_NOT_SUPPORTED;
+ goto out;
+ }
+
+ arg.a1 = dst;
+
+ ffa_finish_direct_req_run(regs, &arg);
+
+ return;
+
+out:
+ ffa_set_regs_error(regs, ret);
+}
diff --git a/xen/arch/arm/tee/ffa_private.h b/xen/arch/arm/tee/ffa_private.h
index d883114948b1..030e6724743c 100644
--- a/xen/arch/arm/tee/ffa_private.h
+++ b/xen/arch/arm/tee/ffa_private.h
@@ -474,6 +474,7 @@ static inline void ffa_raise_rx_buffer_full(struct domain *d)
void ffa_handle_msg_send_direct_req(struct cpu_user_regs *regs, uint32_t fid);
int32_t ffa_handle_msg_send2(struct cpu_user_regs *regs);
+void ffa_handle_run(struct cpu_user_regs *regs, uint32_t fid);
#ifdef CONFIG_FFA_VM_TO_VM
static inline uint16_t get_ffa_vm_count(void)
--
2.51.2
Hi Bertrand,
On Fri, Dec 5, 2025 at 11:37 AM Bertrand Marquis
<bertrand.marquis@arm.com> wrote:
>
> Add FFA_RUN support and remove mention that YIELD and INTERRUPT are not
> supported in the comment as we will now return them.
>
> Make FFA_FEATURE answer supported for RUN, YIELD and INTERRUPT.
>
> Share the SMCCC plumbing used by SEND_DIRECT and RUN via a common
> ffa_finish_direct_req_run() helper so canonical success and error cases
> are handled in one place.
>
> The dispatcher now routes FFA_RUN through ffa_handle_run(), and direct
> requests bail out early if a guest targets itself or a non-secure
> endpoint. This simplifies the direct path while adding support for
> INTERRUPT or YIELD answers from secure world.
>
> Signed-off-by: Bertrand Marquis <bertrand.marquis@arm.com>
> ---
> Changes in v1:
> - Rework commit message to clearly mention that we add FFA_RUN support
> - Remove mentions of FFA_MSG_YIELD/INTERRUPT/RUN not being supported
> - Report YIELD, INTERRUPT an RUN as supported in ffa_features handling
> ---
> xen/arch/arm/tee/ffa.c | 9 ++-
> xen/arch/arm/tee/ffa_msg.c | 111 ++++++++++++++++++++++++---------
> xen/arch/arm/tee/ffa_private.h | 1 +
> 3 files changed, 87 insertions(+), 34 deletions(-)
Looks good
Reviewed-by: Jens Wiklander <jens.wiklander@linaro.org>
Cheers,
Jens
>
> diff --git a/xen/arch/arm/tee/ffa.c b/xen/arch/arm/tee/ffa.c
> index 2c09d10ae6a1..0f07efe5a7b3 100644
> --- a/xen/arch/arm/tee/ffa.c
> +++ b/xen/arch/arm/tee/ffa.c
> @@ -24,9 +24,6 @@
> * o FFA_MEM_DONATE_* and FFA_MEM_LEND_* - Used when tranferring ownership
> * or access of a memory region
> * o FFA_MSG_SEND2 and FFA_MSG_WAIT - Used for indirect messaging
> - * o FFA_MSG_YIELD
> - * o FFA_INTERRUPT - Used to report preemption
> - * o FFA_RUN
> *
> * Limitations in the implemented FF-A interfaces:
> * o FFA_RXTX_MAP_*:
> @@ -266,6 +263,9 @@ static void handle_features(struct cpu_user_regs *regs)
> case FFA_MSG_SEND_DIRECT_REQ_32:
> case FFA_MSG_SEND_DIRECT_REQ_64:
> case FFA_MSG_SEND2:
> + case FFA_RUN:
> + case FFA_INTERRUPT:
> + case FFA_MSG_YIELD:
> ffa_set_regs_success(regs, 0, 0);
> break;
> case FFA_MEM_SHARE_64:
> @@ -353,6 +353,9 @@ static bool ffa_handle_call(struct cpu_user_regs *regs)
> case FFA_MSG_SEND_DIRECT_REQ_64:
> ffa_handle_msg_send_direct_req(regs, fid);
> return true;
> + case FFA_RUN:
> + ffa_handle_run(regs, fid);
> + return true;
> case FFA_MSG_SEND2:
> e = ffa_handle_msg_send2(regs);
> break;
> diff --git a/xen/arch/arm/tee/ffa_msg.c b/xen/arch/arm/tee/ffa_msg.c
> index d60eed6d8811..5a4cb1bb8295 100644
> --- a/xen/arch/arm/tee/ffa_msg.c
> +++ b/xen/arch/arm/tee/ffa_msg.c
> @@ -21,42 +21,74 @@ struct ffa_part_msg_rxtx {
> uint32_t msg_size;
> };
>
> -void ffa_handle_msg_send_direct_req(struct cpu_user_regs *regs, uint32_t fid)
> +static void ffa_finish_direct_req_run(struct cpu_user_regs *regs,
> + struct arm_smccc_1_2_regs *req)
> {
> - struct arm_smccc_1_2_regs arg = { .a0 = fid, };
> struct arm_smccc_1_2_regs resp = { };
> - struct domain *d = current->domain;
> - uint32_t src_dst;
> uint64_t mask;
>
> - if ( smccc_is_conv_64(fid) )
> + arm_smccc_1_2_smc(req, &resp);
> +
> + switch ( resp.a0 )
> + {
> + case FFA_ERROR:
> + case FFA_SUCCESS_32:
> + case FFA_SUCCESS_64:
> + case FFA_MSG_SEND_DIRECT_RESP_32:
> + case FFA_MSG_SEND_DIRECT_RESP_64:
> + case FFA_MSG_YIELD:
> + case FFA_INTERRUPT:
> + break;
> + default:
> + /* Bad fid, report back to the caller. */
> + ffa_set_regs_error(regs, FFA_RET_ABORTED);
> + return;
> + }
> +
> + if ( smccc_is_conv_64(resp.a0) )
> mask = GENMASK_ULL(63, 0);
> else
> mask = GENMASK_ULL(31, 0);
>
> + ffa_set_regs(regs, resp.a0, resp.a1 & mask, resp.a2 & mask, resp.a3 & mask,
> + resp.a4 & mask, resp.a5 & mask, resp.a6 & mask,
> + resp.a7 & mask);
> +}
> +
> +void ffa_handle_msg_send_direct_req(struct cpu_user_regs *regs, uint32_t fid)
> +{
> + struct arm_smccc_1_2_regs arg = { .a0 = fid, };
> + struct domain *d = current->domain;
> + uint32_t src_dst;
> + uint64_t mask;
> + int32_t ret;
> +
> if ( !ffa_fw_supports_fid(fid) )
> {
> - resp.a0 = FFA_ERROR;
> - resp.a2 = FFA_RET_NOT_SUPPORTED;
> + ret = FFA_RET_NOT_SUPPORTED;
> goto out;
> }
>
> src_dst = get_user_reg(regs, 1);
> - if ( (src_dst >> 16) != ffa_get_vm_id(d) )
> + if ( (src_dst >> 16) != ffa_get_vm_id(d) ||
> + (src_dst & GENMASK(15,0)) == ffa_get_vm_id(d) )
> {
> - resp.a0 = FFA_ERROR;
> - resp.a2 = FFA_RET_INVALID_PARAMETERS;
> + ret = FFA_RET_INVALID_PARAMETERS;
> goto out;
> }
>
> /* we do not support direct messages to VMs */
> if ( !FFA_ID_IS_SECURE(src_dst & GENMASK(15,0)) )
> {
> - resp.a0 = FFA_ERROR;
> - resp.a2 = FFA_RET_NOT_SUPPORTED;
> + ret = FFA_RET_NOT_SUPPORTED;
> goto out;
> }
>
> + if ( smccc_is_conv_64(fid) )
> + mask = GENMASK_ULL(63, 0);
> + else
> + mask = GENMASK_ULL(31, 0);
> +
> arg.a1 = src_dst;
> arg.a2 = get_user_reg(regs, 2) & mask;
> arg.a3 = get_user_reg(regs, 3) & mask;
> @@ -65,27 +97,11 @@ void ffa_handle_msg_send_direct_req(struct cpu_user_regs *regs, uint32_t fid)
> arg.a6 = get_user_reg(regs, 6) & mask;
> arg.a7 = get_user_reg(regs, 7) & mask;
>
> - arm_smccc_1_2_smc(&arg, &resp);
> - switch ( resp.a0 )
> - {
> - case FFA_ERROR:
> - case FFA_SUCCESS_32:
> - case FFA_SUCCESS_64:
> - case FFA_MSG_SEND_DIRECT_RESP_32:
> - case FFA_MSG_SEND_DIRECT_RESP_64:
> - break;
> - default:
> - /* Bad fid, report back to the caller. */
> - memset(&resp, 0, sizeof(resp));
> - resp.a0 = FFA_ERROR;
> - resp.a1 = src_dst;
> - resp.a2 = FFA_RET_ABORTED;
> - }
> + ffa_finish_direct_req_run(regs, &arg);
> + return;
>
> out:
> - ffa_set_regs(regs, resp.a0, resp.a1 & mask, resp.a2 & mask, resp.a3 & mask,
> - resp.a4 & mask, resp.a5 & mask, resp.a6 & mask,
> - resp.a7 & mask);
> + ffa_set_regs_error(regs, ret);
> }
>
> static int32_t ffa_msg_send2_vm(uint16_t dst_id, const void *src_buf,
> @@ -217,3 +233,36 @@ out:
> ffa_tx_release(src_ctx);
> return ret;
> }
> +
> +void ffa_handle_run(struct cpu_user_regs *regs, uint32_t fid)
> +{
> + struct arm_smccc_1_2_regs arg = { .a0 = fid, };
> + uint32_t dst = get_user_reg(regs, 1);
> + int32_t ret;
> +
> + if ( !ffa_fw_supports_fid(fid) )
> + {
> + ret = FFA_RET_NOT_SUPPORTED;
> + goto out;
> + }
> +
> + /*
> + * We do not support FFA_RUN to VMs.
> + * Destination endpoint ID is in bits [31:16], bits[15:0] contain the
> + * vCPU ID.
> + */
> + if ( !FFA_ID_IS_SECURE(dst >> 16) )
> + {
> + ret = FFA_RET_NOT_SUPPORTED;
> + goto out;
> + }
> +
> + arg.a1 = dst;
> +
> + ffa_finish_direct_req_run(regs, &arg);
> +
> + return;
> +
> +out:
> + ffa_set_regs_error(regs, ret);
> +}
> diff --git a/xen/arch/arm/tee/ffa_private.h b/xen/arch/arm/tee/ffa_private.h
> index d883114948b1..030e6724743c 100644
> --- a/xen/arch/arm/tee/ffa_private.h
> +++ b/xen/arch/arm/tee/ffa_private.h
> @@ -474,6 +474,7 @@ static inline void ffa_raise_rx_buffer_full(struct domain *d)
>
> void ffa_handle_msg_send_direct_req(struct cpu_user_regs *regs, uint32_t fid);
> int32_t ffa_handle_msg_send2(struct cpu_user_regs *regs);
> +void ffa_handle_run(struct cpu_user_regs *regs, uint32_t fid);
>
> #ifdef CONFIG_FFA_VM_TO_VM
> static inline uint16_t get_ffa_vm_count(void)
> --
> 2.51.2
>
© 2016 - 2025 Red Hat, Inc.