Replace uses of QINVAL_ENTRY_ORDER and QINVAL_INDEX_SHIFT, such that
the constants can be dropped. Move the remaining QINVAL_* ones to the
single source file using them.
Signed-off-by: Jan Beulich <jbeulich@suse.com>
--- a/xen/drivers/passthrough/vtd/iommu.h
+++ b/xen/drivers/passthrough/vtd/iommu.h
@@ -451,17 +451,6 @@ struct qinval_entry {
}q;
};
-/* Each entry is 16 bytes, so 2^8 entries per page */
-#define QINVAL_ENTRY_ORDER ( PAGE_SHIFT - 4 )
-#define QINVAL_MAX_ENTRY_NR (1u << (7 + QINVAL_ENTRY_ORDER))
-
-/* Status data flag */
-#define QINVAL_STAT_INIT 0
-#define QINVAL_STAT_DONE 1
-
-/* Queue invalidation head/tail shift */
-#define QINVAL_INDEX_SHIFT 4
-
#define TYPE_INVAL_CONTEXT 0x1
#define TYPE_INVAL_IOTLB 0x2
#define TYPE_INVAL_DEVICE_IOTLB 0x3
--- a/xen/drivers/passthrough/vtd/qinval.c
+++ b/xen/drivers/passthrough/vtd/qinval.c
@@ -29,6 +29,13 @@
#include "extern.h"
#include "../ats.h"
+/* Each entry is 16 bytes, and there can be up to 2^7 pages. */
+#define QINVAL_MAX_ENTRY_NR (1u << (7 + PAGE_SHIFT_4K - 4))
+
+/* Status data flag */
+#define QINVAL_STAT_INIT 0
+#define QINVAL_STAT_DONE 1
+
static unsigned int __read_mostly qi_pg_order;
static unsigned int __read_mostly qi_entry_nr;
@@ -45,11 +52,11 @@ static unsigned int qinval_next_index(st
{
unsigned int tail = dmar_readl(iommu->reg, DMAR_IQT_REG);
- tail >>= QINVAL_INDEX_SHIFT;
+ tail /= sizeof(struct qinval_entry);
/* (tail+1 == head) indicates a full queue, wait for HW */
while ( ((tail + 1) & (qi_entry_nr - 1)) ==
- (dmar_readl(iommu->reg, DMAR_IQH_REG) >> QINVAL_INDEX_SHIFT) )
+ (dmar_readl(iommu->reg, DMAR_IQH_REG) / sizeof(struct qinval_entry)) )
{
printk_once(XENLOG_ERR VTDPREFIX " IOMMU#%u: no QI slot available\n",
iommu->index);
@@ -66,7 +73,7 @@ static void qinval_update_qtail(struct v
/* Need hold register lock when update tail */
ASSERT( spin_is_locked(&iommu->register_lock) );
val = (index + 1) & (qi_entry_nr - 1);
- dmar_writel(iommu->reg, DMAR_IQT_REG, val << QINVAL_INDEX_SHIFT);
+ dmar_writel(iommu->reg, DMAR_IQT_REG, val * sizeof(struct qinval_entry));
}
static struct qinval_entry *qi_map_entry(const struct vtd_iommu *iommu,
@@ -413,17 +420,18 @@ int enable_qinval(struct vtd_iommu *iomm
* only one entry left.
*/
BUILD_BUG_ON(CONFIG_NR_CPUS * 2 >= QINVAL_MAX_ENTRY_NR);
- qi_pg_order = get_order_from_bytes((num_present_cpus() * 2 + 1) <<
- (PAGE_SHIFT -
- QINVAL_ENTRY_ORDER));
- qi_entry_nr = 1u << (qi_pg_order + QINVAL_ENTRY_ORDER);
+ qi_pg_order = get_order_from_bytes((num_present_cpus() * 2 + 1) *
+ sizeof(struct qinval_entry));
+ qi_entry_nr = (PAGE_SIZE << qi_pg_order) /
+ sizeof(struct qinval_entry);
dprintk(XENLOG_INFO VTDPREFIX,
"QI: using %u-entry ring(s)\n", qi_entry_nr);
}
iommu->qinval_maddr =
- alloc_pgtable_maddr(qi_entry_nr >> QINVAL_ENTRY_ORDER,
+ alloc_pgtable_maddr(PFN_DOWN(qi_entry_nr *
+ sizeof(struct qinval_entry)),
iommu->node);
if ( iommu->qinval_maddr == 0 )
{
> From: Jan Beulich <jbeulich@suse.com> > Sent: Wednesday, June 9, 2021 5:30 PM > > Replace uses of QINVAL_ENTRY_ORDER and QINVAL_INDEX_SHIFT, such that > the constants can be dropped. Move the remaining QINVAL_* ones to the > single source file using them. > > Signed-off-by: Jan Beulich <jbeulich@suse.com> Reviewed-by: Kevin Tian <kevin.tian@intel.com> > > --- a/xen/drivers/passthrough/vtd/iommu.h > +++ b/xen/drivers/passthrough/vtd/iommu.h > @@ -451,17 +451,6 @@ struct qinval_entry { > }q; > }; > > -/* Each entry is 16 bytes, so 2^8 entries per page */ > -#define QINVAL_ENTRY_ORDER ( PAGE_SHIFT - 4 ) > -#define QINVAL_MAX_ENTRY_NR (1u << (7 + QINVAL_ENTRY_ORDER)) > - > -/* Status data flag */ > -#define QINVAL_STAT_INIT 0 > -#define QINVAL_STAT_DONE 1 > - > -/* Queue invalidation head/tail shift */ > -#define QINVAL_INDEX_SHIFT 4 > - > #define TYPE_INVAL_CONTEXT 0x1 > #define TYPE_INVAL_IOTLB 0x2 > #define TYPE_INVAL_DEVICE_IOTLB 0x3 > --- a/xen/drivers/passthrough/vtd/qinval.c > +++ b/xen/drivers/passthrough/vtd/qinval.c > @@ -29,6 +29,13 @@ > #include "extern.h" > #include "../ats.h" > > +/* Each entry is 16 bytes, and there can be up to 2^7 pages. */ > +#define QINVAL_MAX_ENTRY_NR (1u << (7 + PAGE_SHIFT_4K - 4)) > + > +/* Status data flag */ > +#define QINVAL_STAT_INIT 0 > +#define QINVAL_STAT_DONE 1 > + > static unsigned int __read_mostly qi_pg_order; > static unsigned int __read_mostly qi_entry_nr; > > @@ -45,11 +52,11 @@ static unsigned int qinval_next_index(st > { > unsigned int tail = dmar_readl(iommu->reg, DMAR_IQT_REG); > > - tail >>= QINVAL_INDEX_SHIFT; > + tail /= sizeof(struct qinval_entry); > > /* (tail+1 == head) indicates a full queue, wait for HW */ > while ( ((tail + 1) & (qi_entry_nr - 1)) == > - (dmar_readl(iommu->reg, DMAR_IQH_REG) >> > QINVAL_INDEX_SHIFT) ) > + (dmar_readl(iommu->reg, DMAR_IQH_REG) / sizeof(struct > qinval_entry)) ) > { > printk_once(XENLOG_ERR VTDPREFIX " IOMMU#%u: no QI slot > available\n", > iommu->index); > @@ -66,7 +73,7 @@ static void qinval_update_qtail(struct v > /* Need hold register lock when update tail */ > ASSERT( spin_is_locked(&iommu->register_lock) ); > val = (index + 1) & (qi_entry_nr - 1); > - dmar_writel(iommu->reg, DMAR_IQT_REG, val << QINVAL_INDEX_SHIFT); > + dmar_writel(iommu->reg, DMAR_IQT_REG, val * sizeof(struct > qinval_entry)); > } > > static struct qinval_entry *qi_map_entry(const struct vtd_iommu *iommu, > @@ -413,17 +420,18 @@ int enable_qinval(struct vtd_iommu *iomm > * only one entry left. > */ > BUILD_BUG_ON(CONFIG_NR_CPUS * 2 >= QINVAL_MAX_ENTRY_NR); > - qi_pg_order = get_order_from_bytes((num_present_cpus() * 2 + 1) > << > - (PAGE_SHIFT - > - QINVAL_ENTRY_ORDER)); > - qi_entry_nr = 1u << (qi_pg_order + QINVAL_ENTRY_ORDER); > + qi_pg_order = get_order_from_bytes((num_present_cpus() * 2 + 1) * > + sizeof(struct qinval_entry)); > + qi_entry_nr = (PAGE_SIZE << qi_pg_order) / > + sizeof(struct qinval_entry); > > dprintk(XENLOG_INFO VTDPREFIX, > "QI: using %u-entry ring(s)\n", qi_entry_nr); > } > > iommu->qinval_maddr = > - alloc_pgtable_maddr(qi_entry_nr >> QINVAL_ENTRY_ORDER, > + alloc_pgtable_maddr(PFN_DOWN(qi_entry_nr * > + sizeof(struct qinval_entry)), > iommu->node); > if ( iommu->qinval_maddr == 0 ) > {
© 2016 - 2025 Red Hat, Inc.