Refactor CMDQ invalidation paths to carry security state and apply
cache invalidation per sec_sid instead of globally. Extend common
IOTLB/config invalidation helpers with sec_sid filtering, while keeping
SMMU_SEC_SID_NUM as the full-invalidate mode.
In smmuv3, propagate sec_sid/ssec through CFGI and TLBI handling, and
gate VMID usage on queue stage-2 capability (including SEL2 for secure
CMDQ). Update tracepoints to include ssec for better observability.
Signed-off-by: Tao Tang <tangtao1634@phytium.com.cn>
---
hw/arm/smmu-common.c | 96 ++++++++++++++++++++++++++++++++----
hw/arm/smmuv3.c | 67 ++++++++++++++++++-------
hw/arm/trace-events | 6 +--
include/hw/arm/smmu-common.h | 8 +--
4 files changed, 144 insertions(+), 33 deletions(-)
diff --git a/hw/arm/smmu-common.c b/hw/arm/smmu-common.c
index bb43430cc3b..5dece2024a4 100644
--- a/hw/arm/smmu-common.c
+++ b/hw/arm/smmu-common.c
@@ -201,7 +201,7 @@ void smmu_iotlb_insert(SMMUState *bs, SMMUTransCfg *cfg, SMMUTLBEntry *new)
uint8_t tg = (new->granule - 10) / 2;
if (g_hash_table_size(bs->iotlb) >= SMMU_IOTLB_MAX_SIZE) {
- smmu_iotlb_inv_all(bs);
+ smmu_iotlb_inv_all(bs, SMMU_SEC_SID_NUM);
}
*key = smmu_get_iotlb_key(cfg->asid, cfg->s2cfg.vmid, new->entry.iova,
@@ -211,10 +211,23 @@ void smmu_iotlb_insert(SMMUState *bs, SMMUTransCfg *cfg, SMMUTLBEntry *new)
g_hash_table_insert(bs->iotlb, key, new);
}
-void smmu_iotlb_inv_all(SMMUState *s)
+static gboolean smmu_hash_remove_by_sec_sid(gpointer key, gpointer value,
+ gpointer user_data)
+{
+ SMMUIOTLBKey *iotlb_key = (SMMUIOTLBKey *)key;
+ SMMUSecSID *sec_sid = (SMMUSecSID *)user_data;
+
+ return SMMU_IOTLB_SEC_SID(*iotlb_key) == *sec_sid;
+}
+
+void smmu_iotlb_inv_all(SMMUState *s, SMMUSecSID sec_sid)
{
trace_smmu_iotlb_inv_all();
- g_hash_table_remove_all(s->iotlb);
+ if (sec_sid == SMMU_SEC_SID_NUM) {
+ g_hash_table_remove_all(s->iotlb);
+ return;
+ }
+ g_hash_table_foreach_remove(s->iotlb, smmu_hash_remove_by_sec_sid, &sec_sid);
}
static gboolean smmu_hash_remove_by_asid_vmid(gpointer key, gpointer value,
@@ -292,6 +305,16 @@ static gboolean smmu_hash_remove_by_vmid_ipa(gpointer key, gpointer value,
((entry->iova & ~info->mask) == info->iova);
}
+typedef struct SMMUConfigInvRangeInfo {
+ SMMUSIDRange sid_range;
+ SMMUSecSID sec_sid;
+} SMMUConfigInvRangeInfo;
+
+typedef struct SMMUConfigInvSdevInfo {
+ SMMUDevice *sdev;
+ SMMUSecSID sec_sid;
+} SMMUConfigInvSdevInfo;
+
static gboolean
smmu_hash_remove_by_sid_range(gpointer key, gpointer value, gpointer user_data)
{
@@ -307,11 +330,41 @@ smmu_hash_remove_by_sid_range(gpointer key, gpointer value, gpointer user_data)
return true;
}
-void smmu_configs_inv_sid_range(SMMUState *s, SMMUSIDRange sid_range)
+static gboolean
+smmu_hash_remove_by_sid_range_sec(gpointer key, gpointer value,
+ gpointer user_data)
+{
+ SMMUConfigKey *config_key = (SMMUConfigKey *)key;
+ SMMUConfigInvRangeInfo *info = (SMMUConfigInvRangeInfo *)user_data;
+ SMMUDevice *sdev = config_key->sdev;
+ uint32_t sid = smmu_get_sid(sdev);
+
+ if (config_key->sec_sid != info->sec_sid) {
+ return false;
+ }
+ if (sid < info->sid_range.start || sid > info->sid_range.end) {
+ return false;
+ }
+ trace_smmu_config_cache_inv(sid);
+ return true;
+}
+
+void smmu_configs_inv_sid_range(SMMUState *s, SMMUSIDRange sid_range,
+ SMMUSecSID sec_sid)
{
+ SMMUConfigInvRangeInfo info = {
+ .sid_range = sid_range,
+ .sec_sid = sec_sid,
+ };
+
trace_smmu_configs_inv_sid_range(sid_range.start, sid_range.end);
- g_hash_table_foreach_remove(s->configs, smmu_hash_remove_by_sid_range,
- &sid_range);
+ if (sec_sid == SMMU_SEC_SID_NUM) {
+ g_hash_table_foreach_remove(s->configs, smmu_hash_remove_by_sid_range,
+ &sid_range);
+ return;
+ }
+ g_hash_table_foreach_remove(s->configs, smmu_hash_remove_by_sid_range_sec,
+ &info);
}
static gboolean smmu_hash_remove_by_sdev(gpointer key, gpointer value,
@@ -327,9 +380,35 @@ static gboolean smmu_hash_remove_by_sdev(gpointer key, gpointer value,
return true;
}
-void smmu_configs_inv_sdev(SMMUState *s, SMMUDevice *sdev)
+static gboolean smmu_hash_remove_by_sdev_sec(gpointer key, gpointer value,
+ gpointer user_data)
{
- g_hash_table_foreach_remove(s->configs, smmu_hash_remove_by_sdev, sdev);
+ SMMUConfigKey *config_key = (SMMUConfigKey *)key;
+ SMMUConfigInvSdevInfo *info = (SMMUConfigInvSdevInfo *)user_data;
+
+ if (config_key->sdev != info->sdev) {
+ return false;
+ }
+ if (config_key->sec_sid != info->sec_sid) {
+ return false;
+ }
+ trace_smmu_config_cache_inv(smmu_get_sid(info->sdev));
+ return true;
+}
+
+void smmu_configs_inv_sdev(SMMUState *s, SMMUDevice *sdev,
+ SMMUSecSID sec_sid)
+{
+ SMMUConfigInvSdevInfo info = {
+ .sdev = sdev,
+ .sec_sid = sec_sid,
+ };
+
+ if (sec_sid == SMMU_SEC_SID_NUM) {
+ g_hash_table_foreach_remove(s->configs, smmu_hash_remove_by_sdev, sdev);
+ return;
+ }
+ g_hash_table_foreach_remove(s->configs, smmu_hash_remove_by_sdev_sec, &info);
}
void smmu_iotlb_inv_iova(SMMUState *s, int asid, int vmid, dma_addr_t iova,
@@ -1193,4 +1272,3 @@ static void smmu_base_register_types(void)
}
type_init(smmu_base_register_types)
-
diff --git a/hw/arm/smmuv3.c b/hw/arm/smmuv3.c
index d4c58c0c724..29e862b8ae3 100644
--- a/hw/arm/smmuv3.c
+++ b/hw/arm/smmuv3.c
@@ -959,12 +959,13 @@ static SMMUTransCfg *smmuv3_get_config(SMMUDevice *sdev, SMMUEventInfo *event,
return cfg;
}
-static void smmuv3_flush_config(SMMUDevice *sdev)
+/* Flush all config caches when sec_sid == SMMU_SEC_SID_NUM */
+static void smmuv3_flush_config(SMMUDevice *sdev, SMMUSecSID sec_sid)
{
SMMUv3State *s = sdev->smmu;
SMMUState *bc = &s->smmu_state;
- smmu_configs_inv_sdev(bc, sdev);
+ smmu_configs_inv_sdev(bc, sdev, sec_sid);
}
/* Do translation with TLB lookup. */
@@ -1289,7 +1290,7 @@ static void smmuv3_inv_notifiers_iova(SMMUState *s, int asid, int vmid,
}
static void smmuv3_range_inval(SMMUState *s, Cmd *cmd, SMMUStage stage,
- SMMUSecSID sec_sid)
+ SMMUSecSID sec_sid, bool use_vmid)
{
dma_addr_t end, addr = CMD_ADDR(cmd);
uint8_t type = CMD_TYPE(cmd);
@@ -1302,10 +1303,8 @@ static void smmuv3_range_inval(SMMUState *s, Cmd *cmd, SMMUStage stage,
uint64_t num_pages;
uint8_t granule;
int asid = -1;
- SMMUv3State *smmuv3 = ARM_SMMUV3(s);
- /* Only consider VMID if stage-2 is supported. */
- if (STAGE2_SUPPORTED(smmuv3)) {
+ if (use_vmid) {
vmid = CMD_VMID(cmd);
}
@@ -1351,6 +1350,25 @@ static void smmuv3_range_inval(SMMUState *s, Cmd *cmd, SMMUStage stage,
}
}
+static inline bool smmu_cmdq_stage2_supported(SMMUv3State *s, SMMUSecSID sec_sid)
+{
+ /* IDR0.S2P: Stage 2 translation supported */
+ bool s2p = STAGE2_SUPPORTED(s);
+ if (!s2p) {
+ return false;
+ }
+
+ /*
+ * For Secure Command queue, Secure stage 2 is additionally gated by SEL2
+ * (SEL2 is 0 if S2P is 0).
+ */
+ if (sec_sid == SMMU_SEC_SID_S) {
+ return FIELD_EX32(s->bank[SMMU_SEC_SID_S].idr[1], S_IDR1, SEL2);
+ }
+
+ return true;
+}
+
static int smmuv3_cmdq_consume(SMMUv3State *s, Error **errp, SMMUSecSID sec_sid)
{
SMMUState *bs = ARM_SMMU(s);
@@ -1362,6 +1380,7 @@ static int smmuv3_cmdq_consume(SMMUv3State *s, Error **errp, SMMUSecSID sec_sid)
AddressSpace *as = smmu_get_address_space(bs, sec_sid);
/* Secure AddressSpace must be available, assert if not. */
g_assert(as);
+ bool queue_stage2_supported = smmu_cmdq_stage2_supported(s, sec_sid);
if (!smmuv3_cmdq_enabled(s, sec_sid)) {
return 0;
@@ -1376,6 +1395,7 @@ static int smmuv3_cmdq_consume(SMMUv3State *s, Error **errp, SMMUSecSID sec_sid)
while (!smmuv3_q_empty(q)) {
uint32_t pending = bank->gerror ^ bank->gerrorn;
Cmd cmd;
+ SMMUSecSID ssec = SMMU_SEC_SID_NS;
trace_smmuv3_cmdq_consume(sec_sid, Q_PROD(q), Q_CONS(q),
Q_PROD_WRAP(q), Q_CONS_WRAP(q));
@@ -1399,6 +1419,7 @@ static int smmuv3_cmdq_consume(SMMUv3State *s, Error **errp, SMMUSecSID sec_sid)
cmd_error = SMMU_CERROR_ILL;
break;
}
+ ssec = SMMU_SEC_SID_S;
}
type = CMD_TYPE(&cmd);
@@ -1424,12 +1445,12 @@ static int smmuv3_cmdq_consume(SMMUv3State *s, Error **errp, SMMUSecSID sec_sid)
break;
}
- trace_smmuv3_cmdq_cfgi_ste(sid);
+ trace_smmuv3_cmdq_cfgi_ste(ssec, sid);
if (!smmuv3_accel_install_ste(s, sdev, sid, errp)) {
cmd_error = SMMU_CERROR_ILL;
break;
}
- smmuv3_flush_config(sdev);
+ smmuv3_flush_config(sdev, ssec);
break;
}
@@ -1443,12 +1464,12 @@ static int smmuv3_cmdq_consume(SMMUv3State *s, Error **errp, SMMUSecSID sec_sid)
sid_range.start = sid & ~mask;
sid_range.end = sid_range.start + mask;
- trace_smmuv3_cmdq_cfgi_ste_range(sid_range.start, sid_range.end);
+ trace_smmuv3_cmdq_cfgi_ste_range(ssec, sid_range.start, sid_range.end);
if (!smmuv3_accel_install_ste_range(s, &sid_range, errp)) {
cmd_error = SMMU_CERROR_ILL;
break;
}
- smmu_configs_inv_sid_range(bs, sid_range);
+ smmu_configs_inv_sid_range(bs, sid_range, ssec);
break;
}
case SMMU_CMD_CFGI_CD:
@@ -1470,8 +1491,8 @@ static int smmuv3_cmdq_consume(SMMUv3State *s, Error **errp, SMMUSecSID sec_sid)
break;
}
- trace_smmuv3_cmdq_cfgi_cd(sid);
- smmuv3_flush_config(sdev);
+ trace_smmuv3_cmdq_cfgi_cd(ssec, sid);
+ smmuv3_flush_config(sdev, ssec);
if (!smmuv3_accel_issue_inv_cmd(s, &cmd, sdev, errp)) {
cmd_error = SMMU_CERROR_ILL;
break;
@@ -1492,7 +1513,7 @@ static int smmuv3_cmdq_consume(SMMUv3State *s, Error **errp, SMMUSecSID sec_sid)
* VMID is only matched when stage 2 is supported, otherwise set it
* to -1 as the value used for stage-1 only VMIDs.
*/
- if (STAGE2_SUPPORTED(s)) {
+ if (queue_stage2_supported) {
vmid = CMD_VMID(&cmd);
}
@@ -1518,18 +1539,27 @@ static int smmuv3_cmdq_consume(SMMUv3State *s, Error **errp, SMMUSecSID sec_sid)
* If stage-2 is supported, invalidate for this VMID only, otherwise
* invalidate the whole thing.
*/
- if (STAGE2_SUPPORTED(s)) {
+ if (queue_stage2_supported) {
vmid = CMD_VMID(&cmd);
trace_smmuv3_cmdq_tlbi_nh(sec_sid, vmid);
smmu_iotlb_inv_vmid_s1(bs, vmid, sec_sid);
break;
}
- QEMU_FALLTHROUGH;
+ trace_smmuv3_cmdq_tlbi_nh(sec_sid, vmid);
+ smmu_inv_notifiers_all(&s->smmu_state);
+ smmu_iotlb_inv_all(bs, sec_sid);
+ break;
}
case SMMU_CMD_TLBI_NSNH_ALL:
trace_smmuv3_cmdq_tlbi_nsnh();
smmu_inv_notifiers_all(&s->smmu_state);
- smmu_iotlb_inv_all(bs);
+ /*
+ * According to (IHI 0070G.b) 4.4.4.1 CMD_TLBI_NSNH_ALL, Page 194:
+ * "When issuing to the Realm programming interface, even though
+ * this command has NS in its name, it only applies to Realm entries."
+ */
+ smmu_iotlb_inv_all(bs, sec_sid > SMMU_SEC_SID_S ?
+ sec_sid : SMMU_SEC_SID_NS);
if (!smmuv3_accel_issue_inv_cmd(s, &cmd, NULL, errp)) {
cmd_error = SMMU_CERROR_ILL;
break;
@@ -1541,7 +1571,8 @@ static int smmuv3_cmdq_consume(SMMUv3State *s, Error **errp, SMMUSecSID sec_sid)
cmd_error = SMMU_CERROR_ILL;
break;
}
- smmuv3_range_inval(bs, &cmd, SMMU_STAGE_1, SMMU_SEC_SID_NS);
+ smmuv3_range_inval(bs, &cmd, SMMU_STAGE_1, sec_sid,
+ queue_stage2_supported);
if (!smmuv3_accel_issue_inv_cmd(s, &cmd, NULL, errp)) {
cmd_error = SMMU_CERROR_ILL;
break;
@@ -1570,7 +1601,7 @@ static int smmuv3_cmdq_consume(SMMUv3State *s, Error **errp, SMMUSecSID sec_sid)
* As currently only either s1 or s2 are supported
* we can reuse same function for s2.
*/
- smmuv3_range_inval(bs, &cmd, SMMU_STAGE_2, SMMU_SEC_SID_NS);
+ smmuv3_range_inval(bs, &cmd, SMMU_STAGE_2, SMMU_SEC_SID_NS, true);
break;
case SMMU_CMD_ATC_INV:
{
diff --git a/hw/arm/trace-events b/hw/arm/trace-events
index ca8485c96af..64f308a8d35 100644
--- a/hw/arm/trace-events
+++ b/hw/arm/trace-events
@@ -51,9 +51,9 @@ smmuv3_translate_success(const char *n, uint16_t sid, uint64_t iova, uint64_t tr
smmuv3_get_cd(uint64_t addr) "CD addr: 0x%"PRIx64
smmuv3_decode_cd(uint32_t oas) "oas=%d"
smmuv3_decode_cd_tt(int i, uint32_t tsz, uint64_t ttb, uint32_t granule_sz, bool had) "TT[%d]:tsz:%d ttb:0x%"PRIx64" granule_sz:%d had:%d"
-smmuv3_cmdq_cfgi_ste(int streamid) "streamid= 0x%x"
-smmuv3_cmdq_cfgi_ste_range(int start, int end) "start=0x%x - end=0x%x"
-smmuv3_cmdq_cfgi_cd(uint32_t sid) "sid=0x%x"
+smmuv3_cmdq_cfgi_ste(int ssec, int streamid) "ssec=%d streamid= 0x%x"
+smmuv3_cmdq_cfgi_ste_range(int ssec, int start, int end) "ssec=%d start=0x%x - end=0x%x"
+smmuv3_cmdq_cfgi_cd(int ssec, uint32_t sid) "ssec=%d sid=0x%x"
smmuv3_config_cache_hit(uint32_t sid, uint32_t hits, uint32_t misses, uint32_t perc) "Config cache HIT for sid=0x%x (hits=%d, misses=%d, hit rate=%d)"
smmuv3_config_cache_miss(uint32_t sid, uint32_t hits, uint32_t misses, uint32_t perc) "Config cache MISS for sid=0x%x (hits=%d, misses=%d, hit rate=%d)"
smmuv3_range_inval(int sec_sid, int vmid, int asid, uint64_t addr, uint8_t tg, uint64_t num_pages, uint8_t ttl, bool leaf, int stage) "sec_sid=%d vmid=%d asid=%d addr=0x%"PRIx64" tg=%d num_pages=0x%"PRIx64" ttl=%d leaf=%d stage=%d"
diff --git a/include/hw/arm/smmu-common.h b/include/hw/arm/smmu-common.h
index 7d1d0936921..d05cf6ae53b 100644
--- a/include/hw/arm/smmu-common.h
+++ b/include/hw/arm/smmu-common.h
@@ -259,7 +259,7 @@ void smmu_iotlb_insert(SMMUState *bs, SMMUTransCfg *cfg, SMMUTLBEntry *entry);
SMMUIOTLBKey smmu_get_iotlb_key(int asid, int vmid, uint64_t iova,
uint8_t tg, uint8_t level, SMMUSecSID sec_sid);
SMMUConfigKey smmu_get_config_key(SMMUDevice *sdev, SMMUSecSID sec_sid);
-void smmu_iotlb_inv_all(SMMUState *s);
+void smmu_iotlb_inv_all(SMMUState *s, SMMUSecSID sec_sid);
void smmu_iotlb_inv_asid_vmid(SMMUState *s, int asid, int vmid,
SMMUSecSID sec_sid);
void smmu_iotlb_inv_vmid(SMMUState *s, int vmid, SMMUSecSID sec_sid);
@@ -270,8 +270,10 @@ void smmu_iotlb_inv_iova(SMMUState *s, int asid, int vmid, dma_addr_t iova,
void smmu_iotlb_inv_ipa(SMMUState *s, int vmid, dma_addr_t ipa, uint8_t tg,
uint64_t num_pages, uint8_t ttl,
SMMUSecSID sec_sid);
-void smmu_configs_inv_sid_range(SMMUState *s, SMMUSIDRange sid_range);
-void smmu_configs_inv_sdev(SMMUState *s, SMMUDevice *sdev);
+void smmu_configs_inv_sid_range(SMMUState *s, SMMUSIDRange sid_range,
+ SMMUSecSID sec_sid);
+void smmu_configs_inv_sdev(SMMUState *s, SMMUDevice *sdev,
+ SMMUSecSID sec_sid);
/* Unmap the range of all the notifiers registered to any IOMMU mr */
void smmu_inv_notifiers_all(SMMUState *s);
--
2.34.1
On 2/21/26 2:17 AM, Tao Tang wrote:
> Refactor CMDQ invalidation paths to carry security state and apply
> cache invalidation per sec_sid instead of globally. Extend common
> IOTLB/config invalidation helpers with sec_sid filtering, while keeping
> SMMU_SEC_SID_NUM as the full-invalidate mode.
>
> In smmuv3, propagate sec_sid/ssec through CFGI and TLBI handling, and
> gate VMID usage on queue stage-2 capability (including SEL2 for secure
> CMDQ). Update tracepoints to include ssec for better observability.
>
> Signed-off-by: Tao Tang <tangtao1634@phytium.com.cn>
> ---
> hw/arm/smmu-common.c | 96 ++++++++++++++++++++++++++++++++----
> hw/arm/smmuv3.c | 67 ++++++++++++++++++-------
> hw/arm/trace-events | 6 +--
> include/hw/arm/smmu-common.h | 8 +--
> 4 files changed, 144 insertions(+), 33 deletions(-)
>
> diff --git a/hw/arm/smmu-common.c b/hw/arm/smmu-common.c
> index bb43430cc3b..5dece2024a4 100644
> --- a/hw/arm/smmu-common.c
> +++ b/hw/arm/smmu-common.c
> @@ -201,7 +201,7 @@ void smmu_iotlb_insert(SMMUState *bs, SMMUTransCfg *cfg, SMMUTLBEntry *new)
> uint8_t tg = (new->granule - 10) / 2;
>
> if (g_hash_table_size(bs->iotlb) >= SMMU_IOTLB_MAX_SIZE) {
> - smmu_iotlb_inv_all(bs);
> + smmu_iotlb_inv_all(bs, SMMU_SEC_SID_NUM);
> }
>
> *key = smmu_get_iotlb_key(cfg->asid, cfg->s2cfg.vmid, new->entry.iova,
> @@ -211,10 +211,23 @@ void smmu_iotlb_insert(SMMUState *bs, SMMUTransCfg *cfg, SMMUTLBEntry *new)
> g_hash_table_insert(bs->iotlb, key, new);
> }
>
> -void smmu_iotlb_inv_all(SMMUState *s)
> +static gboolean smmu_hash_remove_by_sec_sid(gpointer key, gpointer value,
> + gpointer user_data)
> +{
> + SMMUIOTLBKey *iotlb_key = (SMMUIOTLBKey *)key;
> + SMMUSecSID *sec_sid = (SMMUSecSID *)user_data;
> +
> + return SMMU_IOTLB_SEC_SID(*iotlb_key) == *sec_sid;
> +}
> +
> +void smmu_iotlb_inv_all(SMMUState *s, SMMUSecSID sec_sid)
> {
> trace_smmu_iotlb_inv_all();
> - g_hash_table_remove_all(s->iotlb);
> + if (sec_sid == SMMU_SEC_SID_NUM) {
> + g_hash_table_remove_all(s->iotlb);
> + return;
> + }
> + g_hash_table_foreach_remove(s->iotlb, smmu_hash_remove_by_sec_sid, &sec_sid);
> }
>
> static gboolean smmu_hash_remove_by_asid_vmid(gpointer key, gpointer value,
> @@ -292,6 +305,16 @@ static gboolean smmu_hash_remove_by_vmid_ipa(gpointer key, gpointer value,
> ((entry->iova & ~info->mask) == info->iova);
> }
>
> +typedef struct SMMUConfigInvRangeInfo {
> + SMMUSIDRange sid_range;
> + SMMUSecSID sec_sid;
> +} SMMUConfigInvRangeInfo;
> +
> +typedef struct SMMUConfigInvSdevInfo {
> + SMMUDevice *sdev;
> + SMMUSecSID sec_sid;
> +} SMMUConfigInvSdevInfo;
> +
> static gboolean
> smmu_hash_remove_by_sid_range(gpointer key, gpointer value, gpointer user_data)
> {
> @@ -307,11 +330,41 @@ smmu_hash_remove_by_sid_range(gpointer key, gpointer value, gpointer user_data)
> return true;
> }
>
> -void smmu_configs_inv_sid_range(SMMUState *s, SMMUSIDRange sid_range)
> +static gboolean
> +smmu_hash_remove_by_sid_range_sec(gpointer key, gpointer value,
> + gpointer user_data)
> +{
> + SMMUConfigKey *config_key = (SMMUConfigKey *)key;
> + SMMUConfigInvRangeInfo *info = (SMMUConfigInvRangeInfo *)user_data;
> + SMMUDevice *sdev = config_key->sdev;
> + uint32_t sid = smmu_get_sid(sdev);
> +
> + if (config_key->sec_sid != info->sec_sid) {
> + return false;
> + }
> + if (sid < info->sid_range.start || sid > info->sid_range.end) {
> + return false;
> + }
> + trace_smmu_config_cache_inv(sid);
> + return true;
> +}
> +
> +void smmu_configs_inv_sid_range(SMMUState *s, SMMUSIDRange sid_range,
> + SMMUSecSID sec_sid)
> {
> + SMMUConfigInvRangeInfo info = {
> + .sid_range = sid_range,
> + .sec_sid = sec_sid,
> + };
> +
> trace_smmu_configs_inv_sid_range(sid_range.start, sid_range.end);
> - g_hash_table_foreach_remove(s->configs, smmu_hash_remove_by_sid_range,
> - &sid_range);
> + if (sec_sid == SMMU_SEC_SID_NUM) {
> + g_hash_table_foreach_remove(s->configs, smmu_hash_remove_by_sid_range,
> + &sid_range);
> + return;
> + }
> + g_hash_table_foreach_remove(s->configs, smmu_hash_remove_by_sid_range_sec,
> + &info);
> }
>
> static gboolean smmu_hash_remove_by_sdev(gpointer key, gpointer value,
> @@ -327,9 +380,35 @@ static gboolean smmu_hash_remove_by_sdev(gpointer key, gpointer value,
> return true;
> }
>
> -void smmu_configs_inv_sdev(SMMUState *s, SMMUDevice *sdev)
> +static gboolean smmu_hash_remove_by_sdev_sec(gpointer key, gpointer value,
> + gpointer user_data)
> {
> - g_hash_table_foreach_remove(s->configs, smmu_hash_remove_by_sdev, sdev);
> + SMMUConfigKey *config_key = (SMMUConfigKey *)key;
> + SMMUConfigInvSdevInfo *info = (SMMUConfigInvSdevInfo *)user_data;
> +
> + if (config_key->sdev != info->sdev) {
> + return false;
> + }
> + if (config_key->sec_sid != info->sec_sid) {
> + return false;
> + }
> + trace_smmu_config_cache_inv(smmu_get_sid(info->sdev));
> + return true;
> +}
> +
> +void smmu_configs_inv_sdev(SMMUState *s, SMMUDevice *sdev,
> + SMMUSecSID sec_sid)
> +{
> + SMMUConfigInvSdevInfo info = {
> + .sdev = sdev,
> + .sec_sid = sec_sid,
> + };
> +
> + if (sec_sid == SMMU_SEC_SID_NUM) {
> + g_hash_table_foreach_remove(s->configs, smmu_hash_remove_by_sdev, sdev);
> + return;
> + }
> + g_hash_table_foreach_remove(s->configs, smmu_hash_remove_by_sdev_sec, &info);
> }
>
> void smmu_iotlb_inv_iova(SMMUState *s, int asid, int vmid, dma_addr_t iova,
> @@ -1193,4 +1272,3 @@ static void smmu_base_register_types(void)
> }
>
> type_init(smmu_base_register_types)
> -
> diff --git a/hw/arm/smmuv3.c b/hw/arm/smmuv3.c
> index d4c58c0c724..29e862b8ae3 100644
> --- a/hw/arm/smmuv3.c
> +++ b/hw/arm/smmuv3.c
> @@ -959,12 +959,13 @@ static SMMUTransCfg *smmuv3_get_config(SMMUDevice *sdev, SMMUEventInfo *event,
> return cfg;
> }
>
> -static void smmuv3_flush_config(SMMUDevice *sdev)
> +/* Flush all config caches when sec_sid == SMMU_SEC_SID_NUM */
> +static void smmuv3_flush_config(SMMUDevice *sdev, SMMUSecSID sec_sid)
> {
> SMMUv3State *s = sdev->smmu;
> SMMUState *bc = &s->smmu_state;
>
> - smmu_configs_inv_sdev(bc, sdev);
> + smmu_configs_inv_sdev(bc, sdev, sec_sid);
> }
>
> /* Do translation with TLB lookup. */
> @@ -1289,7 +1290,7 @@ static void smmuv3_inv_notifiers_iova(SMMUState *s, int asid, int vmid,
> }
>
> static void smmuv3_range_inval(SMMUState *s, Cmd *cmd, SMMUStage stage,
> - SMMUSecSID sec_sid)
> + SMMUSecSID sec_sid, bool use_vmid)
> {
> dma_addr_t end, addr = CMD_ADDR(cmd);
> uint8_t type = CMD_TYPE(cmd);
> @@ -1302,10 +1303,8 @@ static void smmuv3_range_inval(SMMUState *s, Cmd *cmd, SMMUStage stage,
> uint64_t num_pages;
> uint8_t granule;
> int asid = -1;
> - SMMUv3State *smmuv3 = ARM_SMMUV3(s);
>
> - /* Only consider VMID if stage-2 is supported. */
> - if (STAGE2_SUPPORTED(smmuv3)) {
> + if (use_vmid) {
> vmid = CMD_VMID(cmd);
> }
>
> @@ -1351,6 +1350,25 @@ static void smmuv3_range_inval(SMMUState *s, Cmd *cmd, SMMUStage stage,
> }
> }
>
> +static inline bool smmu_cmdq_stage2_supported(SMMUv3State *s, SMMUSecSID sec_sid)
> +{
> + /* IDR0.S2P: Stage 2 translation supported */
> + bool s2p = STAGE2_SUPPORTED(s);
> + if (!s2p) {
> + return false;
> + }
> +
> + /*
> + * For Secure Command queue, Secure stage 2 is additionally gated by SEL2
> + * (SEL2 is 0 if S2P is 0).
> + */
> + if (sec_sid == SMMU_SEC_SID_S) {
> + return FIELD_EX32(s->bank[SMMU_SEC_SID_S].idr[1], S_IDR1, SEL2);
> + }
> +
> + return true;
> +}
> +
> static int smmuv3_cmdq_consume(SMMUv3State *s, Error **errp, SMMUSecSID sec_sid)
> {
> SMMUState *bs = ARM_SMMU(s);
> @@ -1362,6 +1380,7 @@ static int smmuv3_cmdq_consume(SMMUv3State *s, Error **errp, SMMUSecSID sec_sid)
> AddressSpace *as = smmu_get_address_space(bs, sec_sid);
> /* Secure AddressSpace must be available, assert if not. */
> g_assert(as);
> + bool queue_stage2_supported = smmu_cmdq_stage2_supported(s, sec_sid);
>
> if (!smmuv3_cmdq_enabled(s, sec_sid)) {
> return 0;
> @@ -1376,6 +1395,7 @@ static int smmuv3_cmdq_consume(SMMUv3State *s, Error **errp, SMMUSecSID sec_sid)
> while (!smmuv3_q_empty(q)) {
> uint32_t pending = bank->gerror ^ bank->gerrorn;
> Cmd cmd;
> + SMMUSecSID ssec = SMMU_SEC_SID_NS;
>
> trace_smmuv3_cmdq_consume(sec_sid, Q_PROD(q), Q_CONS(q),
> Q_PROD_WRAP(q), Q_CONS_WRAP(q));
> @@ -1399,6 +1419,7 @@ static int smmuv3_cmdq_consume(SMMUv3State *s, Error **errp, SMMUSecSID sec_sid)
> cmd_error = SMMU_CERROR_ILL;
> break;
> }
> + ssec = SMMU_SEC_SID_S;
> }
>
> type = CMD_TYPE(&cmd);
> @@ -1424,12 +1445,12 @@ static int smmuv3_cmdq_consume(SMMUv3State *s, Error **errp, SMMUSecSID sec_sid)
> break;
> }
>
> - trace_smmuv3_cmdq_cfgi_ste(sid);
> + trace_smmuv3_cmdq_cfgi_ste(ssec, sid);
> if (!smmuv3_accel_install_ste(s, sdev, sid, errp)) {
> cmd_error = SMMU_CERROR_ILL;
> break;
> }
> - smmuv3_flush_config(sdev);
> + smmuv3_flush_config(sdev, ssec);
>
> break;
> }
> @@ -1443,12 +1464,12 @@ static int smmuv3_cmdq_consume(SMMUv3State *s, Error **errp, SMMUSecSID sec_sid)
> sid_range.start = sid & ~mask;
> sid_range.end = sid_range.start + mask;
>
> - trace_smmuv3_cmdq_cfgi_ste_range(sid_range.start, sid_range.end);
> + trace_smmuv3_cmdq_cfgi_ste_range(ssec, sid_range.start, sid_range.end);
> if (!smmuv3_accel_install_ste_range(s, &sid_range, errp)) {
> cmd_error = SMMU_CERROR_ILL;
> break;
> }
> - smmu_configs_inv_sid_range(bs, sid_range);
> + smmu_configs_inv_sid_range(bs, sid_range, ssec);
> break;
> }
> case SMMU_CMD_CFGI_CD:
> @@ -1470,8 +1491,8 @@ static int smmuv3_cmdq_consume(SMMUv3State *s, Error **errp, SMMUSecSID sec_sid)
> break;
> }
>
> - trace_smmuv3_cmdq_cfgi_cd(sid);
> - smmuv3_flush_config(sdev);
> + trace_smmuv3_cmdq_cfgi_cd(ssec, sid);
> + smmuv3_flush_config(sdev, ssec);
> if (!smmuv3_accel_issue_inv_cmd(s, &cmd, sdev, errp)) {
> cmd_error = SMMU_CERROR_ILL;
> break;
> @@ -1492,7 +1513,7 @@ static int smmuv3_cmdq_consume(SMMUv3State *s, Error **errp, SMMUSecSID sec_sid)
> * VMID is only matched when stage 2 is supported, otherwise set it
> * to -1 as the value used for stage-1 only VMIDs.
> */
> - if (STAGE2_SUPPORTED(s)) {
> + if (queue_stage2_supported) {
> vmid = CMD_VMID(&cmd);
> }
>
> @@ -1518,18 +1539,27 @@ static int smmuv3_cmdq_consume(SMMUv3State *s, Error **errp, SMMUSecSID sec_sid)
> * If stage-2 is supported, invalidate for this VMID only, otherwise
> * invalidate the whole thing.
> */
> - if (STAGE2_SUPPORTED(s)) {
> + if (queue_stage2_supported) {
> vmid = CMD_VMID(&cmd);
> trace_smmuv3_cmdq_tlbi_nh(sec_sid, vmid);
> smmu_iotlb_inv_vmid_s1(bs, vmid, sec_sid);
> break;
> }
> - QEMU_FALLTHROUGH;
> + trace_smmuv3_cmdq_tlbi_nh(sec_sid, vmid);
> + smmu_inv_notifiers_all(&s->smmu_state);
> + smmu_iotlb_inv_all(bs, sec_sid);
> + break;
> }
> case SMMU_CMD_TLBI_NSNH_ALL:
> trace_smmuv3_cmdq_tlbi_nsnh();
> smmu_inv_notifiers_all(&s->smmu_state);
> - smmu_iotlb_inv_all(bs);
> + /*
> + * According to (IHI 0070G.b) 4.4.4.1 CMD_TLBI_NSNH_ALL, Page 194:
> + * "When issuing to the Realm programming interface, even though
> + * this command has NS in its name, it only applies to Realm entries."
> + */
> + smmu_iotlb_inv_all(bs, sec_sid > SMMU_SEC_SID_S ?
> + sec_sid : SMMU_SEC_SID_NS);
> if (!smmuv3_accel_issue_inv_cmd(s, &cmd, NULL, errp)) {
> cmd_error = SMMU_CERROR_ILL;
> break;
> @@ -1541,7 +1571,8 @@ static int smmuv3_cmdq_consume(SMMUv3State *s, Error **errp, SMMUSecSID sec_sid)
> cmd_error = SMMU_CERROR_ILL;
> break;
> }
> - smmuv3_range_inval(bs, &cmd, SMMU_STAGE_1, SMMU_SEC_SID_NS);
> + smmuv3_range_inval(bs, &cmd, SMMU_STAGE_1, sec_sid,
> + queue_stage2_supported);
> if (!smmuv3_accel_issue_inv_cmd(s, &cmd, NULL, errp)) {
> cmd_error = SMMU_CERROR_ILL;
> break;
> @@ -1570,7 +1601,7 @@ static int smmuv3_cmdq_consume(SMMUv3State *s, Error **errp, SMMUSecSID sec_sid)
> * As currently only either s1 or s2 are supported
> * we can reuse same function for s2.
> */
> - smmuv3_range_inval(bs, &cmd, SMMU_STAGE_2, SMMU_SEC_SID_NS);
> + smmuv3_range_inval(bs, &cmd, SMMU_STAGE_2, SMMU_SEC_SID_NS, true);
> break;
> case SMMU_CMD_ATC_INV:
> {
> diff --git a/hw/arm/trace-events b/hw/arm/trace-events
> index ca8485c96af..64f308a8d35 100644
> --- a/hw/arm/trace-events
> +++ b/hw/arm/trace-events
> @@ -51,9 +51,9 @@ smmuv3_translate_success(const char *n, uint16_t sid, uint64_t iova, uint64_t tr
> smmuv3_get_cd(uint64_t addr) "CD addr: 0x%"PRIx64
> smmuv3_decode_cd(uint32_t oas) "oas=%d"
> smmuv3_decode_cd_tt(int i, uint32_t tsz, uint64_t ttb, uint32_t granule_sz, bool had) "TT[%d]:tsz:%d ttb:0x%"PRIx64" granule_sz:%d had:%d"
> -smmuv3_cmdq_cfgi_ste(int streamid) "streamid= 0x%x"
> -smmuv3_cmdq_cfgi_ste_range(int start, int end) "start=0x%x - end=0x%x"
> -smmuv3_cmdq_cfgi_cd(uint32_t sid) "sid=0x%x"
> +smmuv3_cmdq_cfgi_ste(int ssec, int streamid) "ssec=%d streamid= 0x%x"
> +smmuv3_cmdq_cfgi_ste_range(int ssec, int start, int end) "ssec=%d start=0x%x - end=0x%x"
> +smmuv3_cmdq_cfgi_cd(int ssec, uint32_t sid) "ssec=%d sid=0x%x"
> smmuv3_config_cache_hit(uint32_t sid, uint32_t hits, uint32_t misses, uint32_t perc) "Config cache HIT for sid=0x%x (hits=%d, misses=%d, hit rate=%d)"
> smmuv3_config_cache_miss(uint32_t sid, uint32_t hits, uint32_t misses, uint32_t perc) "Config cache MISS for sid=0x%x (hits=%d, misses=%d, hit rate=%d)"
> smmuv3_range_inval(int sec_sid, int vmid, int asid, uint64_t addr, uint8_t tg, uint64_t num_pages, uint8_t ttl, bool leaf, int stage) "sec_sid=%d vmid=%d asid=%d addr=0x%"PRIx64" tg=%d num_pages=0x%"PRIx64" ttl=%d leaf=%d stage=%d"
> diff --git a/include/hw/arm/smmu-common.h b/include/hw/arm/smmu-common.h
> index 7d1d0936921..d05cf6ae53b 100644
> --- a/include/hw/arm/smmu-common.h
> +++ b/include/hw/arm/smmu-common.h
> @@ -259,7 +259,7 @@ void smmu_iotlb_insert(SMMUState *bs, SMMUTransCfg *cfg, SMMUTLBEntry *entry);
> SMMUIOTLBKey smmu_get_iotlb_key(int asid, int vmid, uint64_t iova,
> uint8_t tg, uint8_t level, SMMUSecSID sec_sid);
> SMMUConfigKey smmu_get_config_key(SMMUDevice *sdev, SMMUSecSID sec_sid);
> -void smmu_iotlb_inv_all(SMMUState *s);
> +void smmu_iotlb_inv_all(SMMUState *s, SMMUSecSID sec_sid);
> void smmu_iotlb_inv_asid_vmid(SMMUState *s, int asid, int vmid,
> SMMUSecSID sec_sid);
> void smmu_iotlb_inv_vmid(SMMUState *s, int vmid, SMMUSecSID sec_sid);
> @@ -270,8 +270,10 @@ void smmu_iotlb_inv_iova(SMMUState *s, int asid, int vmid, dma_addr_t iova,
> void smmu_iotlb_inv_ipa(SMMUState *s, int vmid, dma_addr_t ipa, uint8_t tg,
> uint64_t num_pages, uint8_t ttl,
> SMMUSecSID sec_sid);
> -void smmu_configs_inv_sid_range(SMMUState *s, SMMUSIDRange sid_range);
> -void smmu_configs_inv_sdev(SMMUState *s, SMMUDevice *sdev);
> +void smmu_configs_inv_sid_range(SMMUState *s, SMMUSIDRange sid_range,
> + SMMUSecSID sec_sid);
> +void smmu_configs_inv_sdev(SMMUState *s, SMMUDevice *sdev,
> + SMMUSecSID sec_sid);
> /* Unmap the range of all the notifiers registered to any IOMMU mr */
> void smmu_inv_notifiers_all(SMMUState *s);
>
Looks ok overall.
Using SMMU_SEC_SID_NUM as a special value to trigger all
flush/invalidate works, but it's a bit surprising to read to be honest.
It could be more explicit to introduce the two next functions:
smmu_iotlb_inv_by_sec_sid(SMMUState *s, SMMUSecSID sec_sid)
smmuv3_flush_config_by_sec_sid(SMMUDevice *sdev, SMMUSecSID sec_sid)
And in addition:
smmu_iotlb_inv_all with the current semantic (invalidate all entries).
Same for smmuv3_flush_config (flush all configs).
What do you think about it?
Hi Pierrick,
On 2026/2/26 05:47, Pierrick Bouvier wrote:
> On 2/21/26 2:17 AM, Tao Tang wrote:
>> Refactor CMDQ invalidation paths to carry security state and apply
>> cache invalidation per sec_sid instead of globally. Extend common
>> IOTLB/config invalidation helpers with sec_sid filtering, while keeping
>> SMMU_SEC_SID_NUM as the full-invalidate mode.
>>
>> In smmuv3, propagate sec_sid/ssec through CFGI and TLBI handling, and
>> gate VMID usage on queue stage-2 capability (including SEL2 for secure
>> CMDQ). Update tracepoints to include ssec for better observability.
>>
>> Signed-off-by: Tao Tang <tangtao1634@phytium.com.cn>
>> ---
>> hw/arm/smmu-common.c | 96 ++++++++++++++++++++++++++++++++----
>> hw/arm/smmuv3.c | 67 ++++++++++++++++++-------
>> hw/arm/trace-events | 6 +--
>> include/hw/arm/smmu-common.h | 8 +--
>> 4 files changed, 144 insertions(+), 33 deletions(-)
>>
>> diff --git a/hw/arm/smmu-common.c b/hw/arm/smmu-common.c
>> index bb43430cc3b..5dece2024a4 100644
>> --- a/hw/arm/smmu-common.c
>> +++ b/hw/arm/smmu-common.c
>> @@ -201,7 +201,7 @@ void smmu_iotlb_insert(SMMUState *bs,
>> SMMUTransCfg *cfg, SMMUTLBEntry *new)
>> uint8_t tg = (new->granule - 10) / 2;
>> if (g_hash_table_size(bs->iotlb) >= SMMU_IOTLB_MAX_SIZE) {
>> - smmu_iotlb_inv_all(bs);
>> + smmu_iotlb_inv_all(bs, SMMU_SEC_SID_NUM);
>> }
>> *key = smmu_get_iotlb_key(cfg->asid, cfg->s2cfg.vmid,
>> new->entry.iova,
>> @@ -211,10 +211,23 @@ void smmu_iotlb_insert(SMMUState *bs,
>> SMMUTransCfg *cfg, SMMUTLBEntry *new)
>> g_hash_table_insert(bs->iotlb, key, new);
>> }
>> -void smmu_iotlb_inv_all(SMMUState *s)
>> +static gboolean smmu_hash_remove_by_sec_sid(gpointer key, gpointer
>> value,
>> + gpointer user_data)
>> +{
>> + SMMUIOTLBKey *iotlb_key = (SMMUIOTLBKey *)key;
>> + SMMUSecSID *sec_sid = (SMMUSecSID *)user_data;
>> +
>> + return SMMU_IOTLB_SEC_SID(*iotlb_key) == *sec_sid;
>> +}
>> +
>> +void smmu_iotlb_inv_all(SMMUState *s, SMMUSecSID sec_sid)
>> {
>> trace_smmu_iotlb_inv_all();
>> - g_hash_table_remove_all(s->iotlb);
>> + if (sec_sid == SMMU_SEC_SID_NUM) {
>> + g_hash_table_remove_all(s->iotlb);
>> + return;
>> + }
>> + g_hash_table_foreach_remove(s->iotlb,
>> smmu_hash_remove_by_sec_sid, &sec_sid);
>> }
>> static gboolean smmu_hash_remove_by_asid_vmid(gpointer key,
>> gpointer value,
>> @@ -292,6 +305,16 @@ static gboolean
>> smmu_hash_remove_by_vmid_ipa(gpointer key, gpointer value,
>> ((entry->iova & ~info->mask) == info->iova);
>> }
>> +typedef struct SMMUConfigInvRangeInfo {
>> + SMMUSIDRange sid_range;
>> + SMMUSecSID sec_sid;
>> +} SMMUConfigInvRangeInfo;
>> +
>> +typedef struct SMMUConfigInvSdevInfo {
>> + SMMUDevice *sdev;
>> + SMMUSecSID sec_sid;
>> +} SMMUConfigInvSdevInfo;
>> +
>> static gboolean
>> smmu_hash_remove_by_sid_range(gpointer key, gpointer value,
>> gpointer user_data)
>> {
>> @@ -307,11 +330,41 @@ smmu_hash_remove_by_sid_range(gpointer key,
>> gpointer value, gpointer user_data)
>> return true;
>> }
>> -void smmu_configs_inv_sid_range(SMMUState *s, SMMUSIDRange sid_range)
>> +static gboolean
>> +smmu_hash_remove_by_sid_range_sec(gpointer key, gpointer value,
>> + gpointer user_data)
>> +{
>> + SMMUConfigKey *config_key = (SMMUConfigKey *)key;
>> + SMMUConfigInvRangeInfo *info = (SMMUConfigInvRangeInfo *)user_data;
>> + SMMUDevice *sdev = config_key->sdev;
>> + uint32_t sid = smmu_get_sid(sdev);
>> +
>> + if (config_key->sec_sid != info->sec_sid) {
>> + return false;
>> + }
>> + if (sid < info->sid_range.start || sid > info->sid_range.end) {
>> + return false;
>> + }
>> + trace_smmu_config_cache_inv(sid);
>> + return true;
>> +}
>> +
>> +void smmu_configs_inv_sid_range(SMMUState *s, SMMUSIDRange sid_range,
>> + SMMUSecSID sec_sid)
>> {
>> + SMMUConfigInvRangeInfo info = {
>> + .sid_range = sid_range,
>> + .sec_sid = sec_sid,
>> + };
>> +
>> trace_smmu_configs_inv_sid_range(sid_range.start, sid_range.end);
>> - g_hash_table_foreach_remove(s->configs,
>> smmu_hash_remove_by_sid_range,
>> - &sid_range);
>> + if (sec_sid == SMMU_SEC_SID_NUM) {
>> + g_hash_table_foreach_remove(s->configs,
>> smmu_hash_remove_by_sid_range,
>> + &sid_range);
>> + return;
>> + }
>> + g_hash_table_foreach_remove(s->configs,
>> smmu_hash_remove_by_sid_range_sec,
>> + &info);
>> }
>> static gboolean smmu_hash_remove_by_sdev(gpointer key, gpointer
>> value,
>> @@ -327,9 +380,35 @@ static gboolean
>> smmu_hash_remove_by_sdev(gpointer key, gpointer value,
>> return true;
>> }
>> -void smmu_configs_inv_sdev(SMMUState *s, SMMUDevice *sdev)
>> +static gboolean smmu_hash_remove_by_sdev_sec(gpointer key, gpointer
>> value,
>> + gpointer user_data)
>> {
>> - g_hash_table_foreach_remove(s->configs,
>> smmu_hash_remove_by_sdev, sdev);
>> + SMMUConfigKey *config_key = (SMMUConfigKey *)key;
>> + SMMUConfigInvSdevInfo *info = (SMMUConfigInvSdevInfo *)user_data;
>> +
>> + if (config_key->sdev != info->sdev) {
>> + return false;
>> + }
>> + if (config_key->sec_sid != info->sec_sid) {
>> + return false;
>> + }
>> + trace_smmu_config_cache_inv(smmu_get_sid(info->sdev));
>> + return true;
>> +}
>> +
>> +void smmu_configs_inv_sdev(SMMUState *s, SMMUDevice *sdev,
>> + SMMUSecSID sec_sid)
>> +{
>> + SMMUConfigInvSdevInfo info = {
>> + .sdev = sdev,
>> + .sec_sid = sec_sid,
>> + };
>> +
>> + if (sec_sid == SMMU_SEC_SID_NUM) {
>> + g_hash_table_foreach_remove(s->configs,
>> smmu_hash_remove_by_sdev, sdev);
>> + return;
>> + }
>> + g_hash_table_foreach_remove(s->configs,
>> smmu_hash_remove_by_sdev_sec, &info);
>> }
>> void smmu_iotlb_inv_iova(SMMUState *s, int asid, int vmid,
>> dma_addr_t iova,
>> @@ -1193,4 +1272,3 @@ static void smmu_base_register_types(void)
>> }
>> type_init(smmu_base_register_types)
>> -
>> diff --git a/hw/arm/smmuv3.c b/hw/arm/smmuv3.c
>> index d4c58c0c724..29e862b8ae3 100644
>> --- a/hw/arm/smmuv3.c
>> +++ b/hw/arm/smmuv3.c
>> @@ -959,12 +959,13 @@ static SMMUTransCfg
>> *smmuv3_get_config(SMMUDevice *sdev, SMMUEventInfo *event,
>> return cfg;
>> }
>> -static void smmuv3_flush_config(SMMUDevice *sdev)
>> +/* Flush all config caches when sec_sid == SMMU_SEC_SID_NUM */
>> +static void smmuv3_flush_config(SMMUDevice *sdev, SMMUSecSID sec_sid)
>> {
>> SMMUv3State *s = sdev->smmu;
>> SMMUState *bc = &s->smmu_state;
>> - smmu_configs_inv_sdev(bc, sdev);
>> + smmu_configs_inv_sdev(bc, sdev, sec_sid);
>> }
>> /* Do translation with TLB lookup. */
>> @@ -1289,7 +1290,7 @@ static void smmuv3_inv_notifiers_iova(SMMUState
>> *s, int asid, int vmid,
>> }
>> static void smmuv3_range_inval(SMMUState *s, Cmd *cmd, SMMUStage
>> stage,
>> - SMMUSecSID sec_sid)
>> + SMMUSecSID sec_sid, bool use_vmid)
>> {
>> dma_addr_t end, addr = CMD_ADDR(cmd);
>> uint8_t type = CMD_TYPE(cmd);
>> @@ -1302,10 +1303,8 @@ static void smmuv3_range_inval(SMMUState *s,
>> Cmd *cmd, SMMUStage stage,
>> uint64_t num_pages;
>> uint8_t granule;
>> int asid = -1;
>> - SMMUv3State *smmuv3 = ARM_SMMUV3(s);
>> - /* Only consider VMID if stage-2 is supported. */
>> - if (STAGE2_SUPPORTED(smmuv3)) {
>> + if (use_vmid) {
>> vmid = CMD_VMID(cmd);
>> }
>> @@ -1351,6 +1350,25 @@ static void smmuv3_range_inval(SMMUState *s,
>> Cmd *cmd, SMMUStage stage,
>> }
>> }
>> +static inline bool smmu_cmdq_stage2_supported(SMMUv3State *s,
>> SMMUSecSID sec_sid)
>> +{
>> + /* IDR0.S2P: Stage 2 translation supported */
>> + bool s2p = STAGE2_SUPPORTED(s);
>> + if (!s2p) {
>> + return false;
>> + }
>> +
>> + /*
>> + * For Secure Command queue, Secure stage 2 is additionally
>> gated by SEL2
>> + * (SEL2 is 0 if S2P is 0).
>> + */
>> + if (sec_sid == SMMU_SEC_SID_S) {
>> + return FIELD_EX32(s->bank[SMMU_SEC_SID_S].idr[1], S_IDR1,
>> SEL2);
>> + }
>> +
>> + return true;
>> +}
>> +
>> static int smmuv3_cmdq_consume(SMMUv3State *s, Error **errp,
>> SMMUSecSID sec_sid)
>> {
>> SMMUState *bs = ARM_SMMU(s);
>> @@ -1362,6 +1380,7 @@ static int smmuv3_cmdq_consume(SMMUv3State *s,
>> Error **errp, SMMUSecSID sec_sid)
>> AddressSpace *as = smmu_get_address_space(bs, sec_sid);
>> /* Secure AddressSpace must be available, assert if not. */
>> g_assert(as);
>> + bool queue_stage2_supported = smmu_cmdq_stage2_supported(s,
>> sec_sid);
>> if (!smmuv3_cmdq_enabled(s, sec_sid)) {
>> return 0;
>> @@ -1376,6 +1395,7 @@ static int smmuv3_cmdq_consume(SMMUv3State *s,
>> Error **errp, SMMUSecSID sec_sid)
>> while (!smmuv3_q_empty(q)) {
>> uint32_t pending = bank->gerror ^ bank->gerrorn;
>> Cmd cmd;
>> + SMMUSecSID ssec = SMMU_SEC_SID_NS;
>> trace_smmuv3_cmdq_consume(sec_sid, Q_PROD(q), Q_CONS(q),
>> Q_PROD_WRAP(q), Q_CONS_WRAP(q));
>> @@ -1399,6 +1419,7 @@ static int smmuv3_cmdq_consume(SMMUv3State *s,
>> Error **errp, SMMUSecSID sec_sid)
>> cmd_error = SMMU_CERROR_ILL;
>> break;
>> }
>> + ssec = SMMU_SEC_SID_S;
>> }
>> type = CMD_TYPE(&cmd);
>> @@ -1424,12 +1445,12 @@ static int smmuv3_cmdq_consume(SMMUv3State
>> *s, Error **errp, SMMUSecSID sec_sid)
>> break;
>> }
>> - trace_smmuv3_cmdq_cfgi_ste(sid);
>> + trace_smmuv3_cmdq_cfgi_ste(ssec, sid);
>> if (!smmuv3_accel_install_ste(s, sdev, sid, errp)) {
>> cmd_error = SMMU_CERROR_ILL;
>> break;
>> }
>> - smmuv3_flush_config(sdev);
>> + smmuv3_flush_config(sdev, ssec);
>> break;
>> }
>> @@ -1443,12 +1464,12 @@ static int smmuv3_cmdq_consume(SMMUv3State
>> *s, Error **errp, SMMUSecSID sec_sid)
>> sid_range.start = sid & ~mask;
>> sid_range.end = sid_range.start + mask;
>> - trace_smmuv3_cmdq_cfgi_ste_range(sid_range.start,
>> sid_range.end);
>> + trace_smmuv3_cmdq_cfgi_ste_range(ssec, sid_range.start,
>> sid_range.end);
>> if (!smmuv3_accel_install_ste_range(s, &sid_range,
>> errp)) {
>> cmd_error = SMMU_CERROR_ILL;
>> break;
>> }
>> - smmu_configs_inv_sid_range(bs, sid_range);
>> + smmu_configs_inv_sid_range(bs, sid_range, ssec);
>> break;
>> }
>> case SMMU_CMD_CFGI_CD:
>> @@ -1470,8 +1491,8 @@ static int smmuv3_cmdq_consume(SMMUv3State *s,
>> Error **errp, SMMUSecSID sec_sid)
>> break;
>> }
>> - trace_smmuv3_cmdq_cfgi_cd(sid);
>> - smmuv3_flush_config(sdev);
>> + trace_smmuv3_cmdq_cfgi_cd(ssec, sid);
>> + smmuv3_flush_config(sdev, ssec);
>> if (!smmuv3_accel_issue_inv_cmd(s, &cmd, sdev, errp)) {
>> cmd_error = SMMU_CERROR_ILL;
>> break;
>> @@ -1492,7 +1513,7 @@ static int smmuv3_cmdq_consume(SMMUv3State *s,
>> Error **errp, SMMUSecSID sec_sid)
>> * VMID is only matched when stage 2 is supported,
>> otherwise set it
>> * to -1 as the value used for stage-1 only VMIDs.
>> */
>> - if (STAGE2_SUPPORTED(s)) {
>> + if (queue_stage2_supported) {
>> vmid = CMD_VMID(&cmd);
>> }
>> @@ -1518,18 +1539,27 @@ static int smmuv3_cmdq_consume(SMMUv3State
>> *s, Error **errp, SMMUSecSID sec_sid)
>> * If stage-2 is supported, invalidate for this VMID
>> only, otherwise
>> * invalidate the whole thing.
>> */
>> - if (STAGE2_SUPPORTED(s)) {
>> + if (queue_stage2_supported) {
>> vmid = CMD_VMID(&cmd);
>> trace_smmuv3_cmdq_tlbi_nh(sec_sid, vmid);
>> smmu_iotlb_inv_vmid_s1(bs, vmid, sec_sid);
>> break;
>> }
>> - QEMU_FALLTHROUGH;
>> + trace_smmuv3_cmdq_tlbi_nh(sec_sid, vmid);
>> + smmu_inv_notifiers_all(&s->smmu_state);
>> + smmu_iotlb_inv_all(bs, sec_sid);
>> + break;
>> }
>> case SMMU_CMD_TLBI_NSNH_ALL:
>> trace_smmuv3_cmdq_tlbi_nsnh();
>> smmu_inv_notifiers_all(&s->smmu_state);
>> - smmu_iotlb_inv_all(bs);
>> + /*
>> + * According to (IHI 0070G.b) 4.4.4.1 CMD_TLBI_NSNH_ALL,
>> Page 194:
>> + * "When issuing to the Realm programming interface,
>> even though
>> + * this command has NS in its name, it only applies to
>> Realm entries."
>> + */
>> + smmu_iotlb_inv_all(bs, sec_sid > SMMU_SEC_SID_S ?
>> + sec_sid : SMMU_SEC_SID_NS);
>> if (!smmuv3_accel_issue_inv_cmd(s, &cmd, NULL, errp)) {
>> cmd_error = SMMU_CERROR_ILL;
>> break;
>> @@ -1541,7 +1571,8 @@ static int smmuv3_cmdq_consume(SMMUv3State *s,
>> Error **errp, SMMUSecSID sec_sid)
>> cmd_error = SMMU_CERROR_ILL;
>> break;
>> }
>> - smmuv3_range_inval(bs, &cmd, SMMU_STAGE_1,
>> SMMU_SEC_SID_NS);
>> + smmuv3_range_inval(bs, &cmd, SMMU_STAGE_1, sec_sid,
>> + queue_stage2_supported);
>> if (!smmuv3_accel_issue_inv_cmd(s, &cmd, NULL, errp)) {
>> cmd_error = SMMU_CERROR_ILL;
>> break;
>> @@ -1570,7 +1601,7 @@ static int smmuv3_cmdq_consume(SMMUv3State *s,
>> Error **errp, SMMUSecSID sec_sid)
>> * As currently only either s1 or s2 are supported
>> * we can reuse same function for s2.
>> */
>> - smmuv3_range_inval(bs, &cmd, SMMU_STAGE_2,
>> SMMU_SEC_SID_NS);
>> + smmuv3_range_inval(bs, &cmd, SMMU_STAGE_2,
>> SMMU_SEC_SID_NS, true);
>> break;
>> case SMMU_CMD_ATC_INV:
>> {
>> diff --git a/hw/arm/trace-events b/hw/arm/trace-events
>> index ca8485c96af..64f308a8d35 100644
>> --- a/hw/arm/trace-events
>> +++ b/hw/arm/trace-events
>> @@ -51,9 +51,9 @@ smmuv3_translate_success(const char *n, uint16_t
>> sid, uint64_t iova, uint64_t tr
>> smmuv3_get_cd(uint64_t addr) "CD addr: 0x%"PRIx64
>> smmuv3_decode_cd(uint32_t oas) "oas=%d"
>> smmuv3_decode_cd_tt(int i, uint32_t tsz, uint64_t ttb, uint32_t
>> granule_sz, bool had) "TT[%d]:tsz:%d ttb:0x%"PRIx64" granule_sz:%d
>> had:%d"
>> -smmuv3_cmdq_cfgi_ste(int streamid) "streamid= 0x%x"
>> -smmuv3_cmdq_cfgi_ste_range(int start, int end) "start=0x%x - end=0x%x"
>> -smmuv3_cmdq_cfgi_cd(uint32_t sid) "sid=0x%x"
>> +smmuv3_cmdq_cfgi_ste(int ssec, int streamid) "ssec=%d streamid= 0x%x"
>> +smmuv3_cmdq_cfgi_ste_range(int ssec, int start, int end) "ssec=%d
>> start=0x%x - end=0x%x"
>> +smmuv3_cmdq_cfgi_cd(int ssec, uint32_t sid) "ssec=%d sid=0x%x"
>> smmuv3_config_cache_hit(uint32_t sid, uint32_t hits, uint32_t
>> misses, uint32_t perc) "Config cache HIT for sid=0x%x (hits=%d,
>> misses=%d, hit rate=%d)"
>> smmuv3_config_cache_miss(uint32_t sid, uint32_t hits, uint32_t
>> misses, uint32_t perc) "Config cache MISS for sid=0x%x (hits=%d,
>> misses=%d, hit rate=%d)"
>> smmuv3_range_inval(int sec_sid, int vmid, int asid, uint64_t addr,
>> uint8_t tg, uint64_t num_pages, uint8_t ttl, bool leaf, int stage)
>> "sec_sid=%d vmid=%d asid=%d addr=0x%"PRIx64" tg=%d
>> num_pages=0x%"PRIx64" ttl=%d leaf=%d stage=%d"
>> diff --git a/include/hw/arm/smmu-common.h b/include/hw/arm/smmu-common.h
>> index 7d1d0936921..d05cf6ae53b 100644
>> --- a/include/hw/arm/smmu-common.h
>> +++ b/include/hw/arm/smmu-common.h
>> @@ -259,7 +259,7 @@ void smmu_iotlb_insert(SMMUState *bs,
>> SMMUTransCfg *cfg, SMMUTLBEntry *entry);
>> SMMUIOTLBKey smmu_get_iotlb_key(int asid, int vmid, uint64_t iova,
>> uint8_t tg, uint8_t level,
>> SMMUSecSID sec_sid);
>> SMMUConfigKey smmu_get_config_key(SMMUDevice *sdev, SMMUSecSID
>> sec_sid);
>> -void smmu_iotlb_inv_all(SMMUState *s);
>> +void smmu_iotlb_inv_all(SMMUState *s, SMMUSecSID sec_sid);
>> void smmu_iotlb_inv_asid_vmid(SMMUState *s, int asid, int vmid,
>> SMMUSecSID sec_sid);
>> void smmu_iotlb_inv_vmid(SMMUState *s, int vmid, SMMUSecSID sec_sid);
>> @@ -270,8 +270,10 @@ void smmu_iotlb_inv_iova(SMMUState *s, int asid,
>> int vmid, dma_addr_t iova,
>> void smmu_iotlb_inv_ipa(SMMUState *s, int vmid, dma_addr_t ipa,
>> uint8_t tg,
>> uint64_t num_pages, uint8_t ttl,
>> SMMUSecSID sec_sid);
>> -void smmu_configs_inv_sid_range(SMMUState *s, SMMUSIDRange sid_range);
>> -void smmu_configs_inv_sdev(SMMUState *s, SMMUDevice *sdev);
>> +void smmu_configs_inv_sid_range(SMMUState *s, SMMUSIDRange sid_range,
>> + SMMUSecSID sec_sid);
>> +void smmu_configs_inv_sdev(SMMUState *s, SMMUDevice *sdev,
>> + SMMUSecSID sec_sid);
>> /* Unmap the range of all the notifiers registered to any IOMMU mr */
>> void smmu_inv_notifiers_all(SMMUState *s);
>
> Looks ok overall.
>
> Using SMMU_SEC_SID_NUM as a special value to trigger all
> flush/invalidate works, but it's a bit surprising to read to be honest.
>
> It could be more explicit to introduce the two next functions:
> smmu_iotlb_inv_by_sec_sid(SMMUState *s, SMMUSecSID sec_sid)
> smmuv3_flush_config_by_sec_sid(SMMUDevice *sdev, SMMUSecSID sec_sid)
>
> And in addition:
> smmu_iotlb_inv_all with the current semantic (invalidate all entries).
> Same for smmuv3_flush_config (flush all configs).
>
> What do you think about it?
Thank you for the review and the suggestion. I agree that using
SMMU_SEC_SID_NUM as a special value to trigger a full flush/invalidate
is not ideal in terms of readability.
I will adopt your approach and split the functions in the next version.
Best regards,
Tao
© 2016 - 2026 Red Hat, Inc.