[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <Z5oN3sAk2RY4vyOr@google.com>
Date: Wed, 29 Jan 2025 11:15:42 +0000
From: Mostafa Saleh <smostafa@...gle.com>
To: Robin Murphy <robin.murphy@....com>
Cc: iommu@...ts.linux.dev, kvmarm@...ts.linux.dev,
linux-kernel@...r.kernel.org, linux-arm-kernel@...ts.infradead.org,
catalin.marinas@....com, will@...nel.org, maz@...nel.org,
oliver.upton@...ux.dev, joey.gouly@....com, suzuki.poulose@....com,
yuzenghui@...wei.com, robdclark@...il.com, joro@...tes.org,
jean-philippe@...aro.org, jgg@...pe.ca, nicolinc@...dia.com,
vdonnefort@...gle.com, qperret@...gle.com, tabba@...gle.com,
danielmentz@...gle.com, tzukui@...gle.com
Subject: Re: [RFC PATCH v2 27/58] KVM: arm64: smmu-v3: Setup command queue
On Thu, Jan 23, 2025 at 01:01:55PM +0000, Robin Murphy wrote:
> On 2024-12-12 6:03 pm, Mostafa Saleh wrote:
> > From: Jean-Philippe Brucker <jean-philippe@...aro.org>
> >
> > Map the command queue allocated by the host into the hypervisor address
> > space. When the host mappings are finalized, the queue is unmapped from
> > the host.
>
> Don't forget the fun of reimplementing the errata workarounds to avoid
> generating certain problematic command sequences - beware it's mostly
> implicit in the current kernel driver :)
Thanks, I see I missed “ARM_SMMU_OPT_CMDQ_FORCE_SYNC”, I will try to re-use
as much of the command queue code as possible in v3, although it’s unlikely
the hypervisor will have the some insertion algorithm as the host, but at
least for the command population.
Thanks,
Mostafa
>
> Thanks,
> Robin.
>
> > Signed-off-by: Jean-Philippe Brucker <jean-philippe@...aro.org>
> > Signed-off-by: Mostafa Saleh <smostafa@...gle.com>
> > ---
> > arch/arm64/kvm/hyp/nvhe/iommu/arm-smmu-v3.c | 165 ++++++++++++++++++++
> > include/kvm/arm_smmu_v3.h | 4 +
> > 2 files changed, 169 insertions(+)
> >
> > diff --git a/arch/arm64/kvm/hyp/nvhe/iommu/arm-smmu-v3.c b/arch/arm64/kvm/hyp/nvhe/iommu/arm-smmu-v3.c
> > index f7e60c188cb0..e15356509424 100644
> > --- a/arch/arm64/kvm/hyp/nvhe/iommu/arm-smmu-v3.c
> > +++ b/arch/arm64/kvm/hyp/nvhe/iommu/arm-smmu-v3.c
> > @@ -41,6 +41,15 @@ struct hyp_arm_smmu_v3_device *kvm_hyp_arm_smmu_v3_smmus;
> > __ret; \
> > })
> > +#define smmu_wait_event(_smmu, _cond) \
> > +({ \
> > + if ((_smmu)->features & ARM_SMMU_FEAT_SEV) { \
> > + while (!(_cond)) \
> > + wfe(); \
> > + } \
> > + smmu_wait(_cond); \
> > +})
> > +
> > static int smmu_write_cr0(struct hyp_arm_smmu_v3_device *smmu, u32 val)
> > {
> > writel_relaxed(val, smmu->base + ARM_SMMU_CR0);
> > @@ -60,6 +69,123 @@ static void smmu_reclaim_pages(u64 phys, size_t size)
> > WARN_ON(__pkvm_hyp_donate_host(phys >> PAGE_SHIFT, size >> PAGE_SHIFT));
> > }
> > +#define Q_WRAP(smmu, reg) ((reg) & (1 << (smmu)->cmdq_log2size))
> > +#define Q_IDX(smmu, reg) ((reg) & ((1 << (smmu)->cmdq_log2size) - 1))
> > +
> > +static bool smmu_cmdq_full(struct hyp_arm_smmu_v3_device *smmu)
> > +{
> > + u64 cons = readl_relaxed(smmu->base + ARM_SMMU_CMDQ_CONS);
> > +
> > + return Q_IDX(smmu, smmu->cmdq_prod) == Q_IDX(smmu, cons) &&
> > + Q_WRAP(smmu, smmu->cmdq_prod) != Q_WRAP(smmu, cons);
> > +}
> > +
> > +static bool smmu_cmdq_empty(struct hyp_arm_smmu_v3_device *smmu)
> > +{
> > + u64 cons = readl_relaxed(smmu->base + ARM_SMMU_CMDQ_CONS);
> > +
> > + return Q_IDX(smmu, smmu->cmdq_prod) == Q_IDX(smmu, cons) &&
> > + Q_WRAP(smmu, smmu->cmdq_prod) == Q_WRAP(smmu, cons);
> > +}
> > +
> > +static int smmu_add_cmd(struct hyp_arm_smmu_v3_device *smmu,
> > + struct arm_smmu_cmdq_ent *ent)
> > +{
> > + int i;
> > + int ret;
> > + u64 cmd[CMDQ_ENT_DWORDS] = {};
> > + int idx = Q_IDX(smmu, smmu->cmdq_prod);
> > + u64 *slot = smmu->cmdq_base + idx * CMDQ_ENT_DWORDS;
> > +
> > + if (smmu->iommu.power_is_off)
> > + return -EPIPE;
> > +
> > + ret = smmu_wait_event(smmu, !smmu_cmdq_full(smmu));
> > + if (ret)
> > + return ret;
> > +
> > + cmd[0] |= FIELD_PREP(CMDQ_0_OP, ent->opcode);
> > +
> > + switch (ent->opcode) {
> > + case CMDQ_OP_CFGI_ALL:
> > + cmd[1] |= FIELD_PREP(CMDQ_CFGI_1_RANGE, 31);
> > + break;
> > + case CMDQ_OP_CFGI_CD:
> > + cmd[0] |= FIELD_PREP(CMDQ_CFGI_0_SSID, ent->cfgi.ssid);
> > + fallthrough;
> > + case CMDQ_OP_CFGI_STE:
> > + cmd[0] |= FIELD_PREP(CMDQ_CFGI_0_SID, ent->cfgi.sid);
> > + cmd[1] |= FIELD_PREP(CMDQ_CFGI_1_LEAF, ent->cfgi.leaf);
> > + break;
> > + case CMDQ_OP_TLBI_NH_VA:
> > + cmd[0] |= FIELD_PREP(CMDQ_TLBI_0_VMID, ent->tlbi.vmid);
> > + cmd[0] |= FIELD_PREP(CMDQ_TLBI_0_NUM, ent->tlbi.num);
> > + cmd[0] |= FIELD_PREP(CMDQ_TLBI_0_SCALE, ent->tlbi.scale);
> > + cmd[0] |= FIELD_PREP(CMDQ_TLBI_0_ASID, ent->tlbi.asid);
> > + cmd[1] |= FIELD_PREP(CMDQ_TLBI_1_LEAF, ent->tlbi.leaf);
> > + cmd[1] |= FIELD_PREP(CMDQ_TLBI_1_TTL, ent->tlbi.ttl);
> > + cmd[1] |= FIELD_PREP(CMDQ_TLBI_1_TG, ent->tlbi.tg);
> > + cmd[1] |= ent->tlbi.addr & CMDQ_TLBI_1_VA_MASK;
> > + break;
> > + case CMDQ_OP_TLBI_NSNH_ALL:
> > + break;
> > + case CMDQ_OP_TLBI_NH_ASID:
> > + cmd[0] |= FIELD_PREP(CMDQ_TLBI_0_ASID, ent->tlbi.asid);
> > + fallthrough;
> > + case CMDQ_OP_TLBI_S12_VMALL:
> > + cmd[0] |= FIELD_PREP(CMDQ_TLBI_0_VMID, ent->tlbi.vmid);
> > + break;
> > + case CMDQ_OP_TLBI_S2_IPA:
> > + cmd[0] |= FIELD_PREP(CMDQ_TLBI_0_NUM, ent->tlbi.num);
> > + cmd[0] |= FIELD_PREP(CMDQ_TLBI_0_SCALE, ent->tlbi.scale);
> > + cmd[0] |= FIELD_PREP(CMDQ_TLBI_0_VMID, ent->tlbi.vmid);
> > + cmd[1] |= FIELD_PREP(CMDQ_TLBI_1_LEAF, ent->tlbi.leaf);
> > + cmd[1] |= FIELD_PREP(CMDQ_TLBI_1_TTL, ent->tlbi.ttl);
> > + cmd[1] |= FIELD_PREP(CMDQ_TLBI_1_TG, ent->tlbi.tg);
> > + cmd[1] |= ent->tlbi.addr & CMDQ_TLBI_1_IPA_MASK;
> > + break;
> > + case CMDQ_OP_CMD_SYNC:
> > + cmd[0] |= FIELD_PREP(CMDQ_SYNC_0_CS, CMDQ_SYNC_0_CS_SEV);
> > + break;
> > + default:
> > + return -EINVAL;
> > + }
> > +
> > + for (i = 0; i < CMDQ_ENT_DWORDS; i++)
> > + slot[i] = cpu_to_le64(cmd[i]);
> > +
> > + smmu->cmdq_prod++;
> > + writel(Q_IDX(smmu, smmu->cmdq_prod) | Q_WRAP(smmu, smmu->cmdq_prod),
> > + smmu->base + ARM_SMMU_CMDQ_PROD);
> > + return 0;
> > +}
> > +
> > +static int smmu_sync_cmd(struct hyp_arm_smmu_v3_device *smmu)
> > +{
> > + int ret;
> > + struct arm_smmu_cmdq_ent cmd = {
> > + .opcode = CMDQ_OP_CMD_SYNC,
> > + };
> > +
> > + ret = smmu_add_cmd(smmu, &cmd);
> > + if (ret)
> > + return ret;
> > +
> > + return smmu_wait_event(smmu, smmu_cmdq_empty(smmu));
> > +}
> > +
> > +__maybe_unused
> > +static int smmu_send_cmd(struct hyp_arm_smmu_v3_device *smmu,
> > + struct arm_smmu_cmdq_ent *cmd)
> > +{
> > + int ret = smmu_add_cmd(smmu, cmd);
> > +
> > + if (ret)
> > + return ret;
> > +
> > + return smmu_sync_cmd(smmu);
> > +}
> > +
> > static int smmu_init_registers(struct hyp_arm_smmu_v3_device *smmu)
> > {
> > u64 val, old;
> > @@ -94,6 +220,41 @@ static int smmu_init_registers(struct hyp_arm_smmu_v3_device *smmu)
> > return 0;
> > }
> > +static int smmu_init_cmdq(struct hyp_arm_smmu_v3_device *smmu)
> > +{
> > + u64 cmdq_base;
> > + size_t cmdq_nr_entries, cmdq_size;
> > + int ret;
> > + enum kvm_pgtable_prot prot = PAGE_HYP;
> > +
> > + cmdq_base = readq_relaxed(smmu->base + ARM_SMMU_CMDQ_BASE);
> > + if (cmdq_base & ~(Q_BASE_RWA | Q_BASE_ADDR_MASK | Q_BASE_LOG2SIZE))
> > + return -EINVAL;
> > +
> > + smmu->cmdq_log2size = cmdq_base & Q_BASE_LOG2SIZE;
> > + cmdq_nr_entries = 1 << smmu->cmdq_log2size;
> > + cmdq_size = cmdq_nr_entries * CMDQ_ENT_DWORDS * 8;
> > +
> > + cmdq_base &= Q_BASE_ADDR_MASK;
> > +
> > + if (!(smmu->features & ARM_SMMU_FEAT_COHERENCY))
> > + prot |= KVM_PGTABLE_PROT_NORMAL_NC;
> > +
> > + ret = ___pkvm_host_donate_hyp_prot(cmdq_base >> PAGE_SHIFT,
> > + PAGE_ALIGN(cmdq_size) >> PAGE_SHIFT,
> > + false, prot);
> > + if (ret)
> > + return ret;
> > +
> > + smmu->cmdq_base = hyp_phys_to_virt(cmdq_base);
> > +
> > + memset(smmu->cmdq_base, 0, cmdq_size);
> > + writel_relaxed(0, smmu->base + ARM_SMMU_CMDQ_PROD);
> > + writel_relaxed(0, smmu->base + ARM_SMMU_CMDQ_CONS);
> > +
> > + return 0;
> > +}
> > +
> > static int smmu_init_device(struct hyp_arm_smmu_v3_device *smmu)
> > {
> > int ret;
> > @@ -113,6 +274,10 @@ static int smmu_init_device(struct hyp_arm_smmu_v3_device *smmu)
> > if (ret)
> > return ret;
> > + ret = smmu_init_cmdq(smmu);
> > + if (ret)
> > + return ret;
> > +
> > return kvm_iommu_init_device(&smmu->iommu);
> > }
> > diff --git a/include/kvm/arm_smmu_v3.h b/include/kvm/arm_smmu_v3.h
> > index fb24bcef1624..393a1a04edba 100644
> > --- a/include/kvm/arm_smmu_v3.h
> > +++ b/include/kvm/arm_smmu_v3.h
> > @@ -16,8 +16,12 @@ struct hyp_arm_smmu_v3_device {
> > struct kvm_hyp_iommu iommu;
> > phys_addr_t mmio_addr;
> > size_t mmio_size;
> > + unsigned long features;
> > void __iomem *base;
> > + u32 cmdq_prod;
> > + u64 *cmdq_base;
> > + size_t cmdq_log2size;
> > };
> > extern size_t kvm_nvhe_sym(kvm_hyp_arm_smmu_v3_count);
>
Powered by blists - more mailing lists