lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <Z5oN3sAk2RY4vyOr@google.com>
Date: Wed, 29 Jan 2025 11:15:42 +0000
From: Mostafa Saleh <smostafa@...gle.com>
To: Robin Murphy <robin.murphy@....com>
Cc: iommu@...ts.linux.dev, kvmarm@...ts.linux.dev,
	linux-kernel@...r.kernel.org, linux-arm-kernel@...ts.infradead.org,
	catalin.marinas@....com, will@...nel.org, maz@...nel.org,
	oliver.upton@...ux.dev, joey.gouly@....com, suzuki.poulose@....com,
	yuzenghui@...wei.com, robdclark@...il.com, joro@...tes.org,
	jean-philippe@...aro.org, jgg@...pe.ca, nicolinc@...dia.com,
	vdonnefort@...gle.com, qperret@...gle.com, tabba@...gle.com,
	danielmentz@...gle.com, tzukui@...gle.com
Subject: Re: [RFC PATCH v2 27/58] KVM: arm64: smmu-v3: Setup command queue

On Thu, Jan 23, 2025 at 01:01:55PM +0000, Robin Murphy wrote:
> On 2024-12-12 6:03 pm, Mostafa Saleh wrote:
> > From: Jean-Philippe Brucker <jean-philippe@...aro.org>
> > 
> > Map the command queue allocated by the host into the hypervisor address
> > space. When the host mappings are finalized, the queue is unmapped from
> > the host.
> 
> Don't forget the fun of reimplementing the errata workarounds to avoid
> generating certain problematic command sequences - beware it's mostly
> implicit in the current kernel driver :)

Thanks, I see I missed “ARM_SMMU_OPT_CMDQ_FORCE_SYNC”, I will try to re-use
as much of the command queue code as possible in v3, although it’s unlikely
the hypervisor will have the some insertion algorithm as the host, but at
least for the command population.

Thanks,
Mostafa

> 
> Thanks,
> Robin.
> 
> > Signed-off-by: Jean-Philippe Brucker <jean-philippe@...aro.org>
> > Signed-off-by: Mostafa Saleh <smostafa@...gle.com>
> > ---
> >   arch/arm64/kvm/hyp/nvhe/iommu/arm-smmu-v3.c | 165 ++++++++++++++++++++
> >   include/kvm/arm_smmu_v3.h                   |   4 +
> >   2 files changed, 169 insertions(+)
> > 
> > diff --git a/arch/arm64/kvm/hyp/nvhe/iommu/arm-smmu-v3.c b/arch/arm64/kvm/hyp/nvhe/iommu/arm-smmu-v3.c
> > index f7e60c188cb0..e15356509424 100644
> > --- a/arch/arm64/kvm/hyp/nvhe/iommu/arm-smmu-v3.c
> > +++ b/arch/arm64/kvm/hyp/nvhe/iommu/arm-smmu-v3.c
> > @@ -41,6 +41,15 @@ struct hyp_arm_smmu_v3_device *kvm_hyp_arm_smmu_v3_smmus;
> >   	__ret;							\
> >   })
> > +#define smmu_wait_event(_smmu, _cond)				\
> > +({								\
> > +	if ((_smmu)->features & ARM_SMMU_FEAT_SEV) {		\
> > +		while (!(_cond))				\
> > +			wfe();					\
> > +	}							\
> > +	smmu_wait(_cond);					\
> > +})
> > +
> >   static int smmu_write_cr0(struct hyp_arm_smmu_v3_device *smmu, u32 val)
> >   {
> >   	writel_relaxed(val, smmu->base + ARM_SMMU_CR0);
> > @@ -60,6 +69,123 @@ static void smmu_reclaim_pages(u64 phys, size_t size)
> >   	WARN_ON(__pkvm_hyp_donate_host(phys >> PAGE_SHIFT, size >> PAGE_SHIFT));
> >   }
> > +#define Q_WRAP(smmu, reg)	((reg) & (1 << (smmu)->cmdq_log2size))
> > +#define Q_IDX(smmu, reg)	((reg) & ((1 << (smmu)->cmdq_log2size) - 1))
> > +
> > +static bool smmu_cmdq_full(struct hyp_arm_smmu_v3_device *smmu)
> > +{
> > +	u64 cons = readl_relaxed(smmu->base + ARM_SMMU_CMDQ_CONS);
> > +
> > +	return Q_IDX(smmu, smmu->cmdq_prod) == Q_IDX(smmu, cons) &&
> > +	       Q_WRAP(smmu, smmu->cmdq_prod) != Q_WRAP(smmu, cons);
> > +}
> > +
> > +static bool smmu_cmdq_empty(struct hyp_arm_smmu_v3_device *smmu)
> > +{
> > +	u64 cons = readl_relaxed(smmu->base + ARM_SMMU_CMDQ_CONS);
> > +
> > +	return Q_IDX(smmu, smmu->cmdq_prod) == Q_IDX(smmu, cons) &&
> > +	       Q_WRAP(smmu, smmu->cmdq_prod) == Q_WRAP(smmu, cons);
> > +}
> > +
> > +static int smmu_add_cmd(struct hyp_arm_smmu_v3_device *smmu,
> > +			struct arm_smmu_cmdq_ent *ent)
> > +{
> > +	int i;
> > +	int ret;
> > +	u64 cmd[CMDQ_ENT_DWORDS] = {};
> > +	int idx = Q_IDX(smmu, smmu->cmdq_prod);
> > +	u64 *slot = smmu->cmdq_base + idx * CMDQ_ENT_DWORDS;
> > +
> > +	if (smmu->iommu.power_is_off)
> > +		return -EPIPE;
> > +
> > +	ret = smmu_wait_event(smmu, !smmu_cmdq_full(smmu));
> > +	if (ret)
> > +		return ret;
> > +
> > +	cmd[0] |= FIELD_PREP(CMDQ_0_OP, ent->opcode);
> > +
> > +	switch (ent->opcode) {
> > +	case CMDQ_OP_CFGI_ALL:
> > +		cmd[1] |= FIELD_PREP(CMDQ_CFGI_1_RANGE, 31);
> > +		break;
> > +	case CMDQ_OP_CFGI_CD:
> > +		cmd[0] |= FIELD_PREP(CMDQ_CFGI_0_SSID, ent->cfgi.ssid);
> > +		fallthrough;
> > +	case CMDQ_OP_CFGI_STE:
> > +		cmd[0] |= FIELD_PREP(CMDQ_CFGI_0_SID, ent->cfgi.sid);
> > +		cmd[1] |= FIELD_PREP(CMDQ_CFGI_1_LEAF, ent->cfgi.leaf);
> > +		break;
> > +	case CMDQ_OP_TLBI_NH_VA:
> > +		cmd[0] |= FIELD_PREP(CMDQ_TLBI_0_VMID, ent->tlbi.vmid);
> > +		cmd[0] |= FIELD_PREP(CMDQ_TLBI_0_NUM, ent->tlbi.num);
> > +		cmd[0] |= FIELD_PREP(CMDQ_TLBI_0_SCALE, ent->tlbi.scale);
> > +		cmd[0] |= FIELD_PREP(CMDQ_TLBI_0_ASID, ent->tlbi.asid);
> > +		cmd[1] |= FIELD_PREP(CMDQ_TLBI_1_LEAF, ent->tlbi.leaf);
> > +		cmd[1] |= FIELD_PREP(CMDQ_TLBI_1_TTL, ent->tlbi.ttl);
> > +		cmd[1] |= FIELD_PREP(CMDQ_TLBI_1_TG, ent->tlbi.tg);
> > +		cmd[1] |= ent->tlbi.addr & CMDQ_TLBI_1_VA_MASK;
> > +		break;
> > +	case CMDQ_OP_TLBI_NSNH_ALL:
> > +		break;
> > +	case CMDQ_OP_TLBI_NH_ASID:
> > +		cmd[0] |= FIELD_PREP(CMDQ_TLBI_0_ASID, ent->tlbi.asid);
> > +		fallthrough;
> > +	case CMDQ_OP_TLBI_S12_VMALL:
> > +		cmd[0] |= FIELD_PREP(CMDQ_TLBI_0_VMID, ent->tlbi.vmid);
> > +		break;
> > +	case CMDQ_OP_TLBI_S2_IPA:
> > +		cmd[0] |= FIELD_PREP(CMDQ_TLBI_0_NUM, ent->tlbi.num);
> > +		cmd[0] |= FIELD_PREP(CMDQ_TLBI_0_SCALE, ent->tlbi.scale);
> > +		cmd[0] |= FIELD_PREP(CMDQ_TLBI_0_VMID, ent->tlbi.vmid);
> > +		cmd[1] |= FIELD_PREP(CMDQ_TLBI_1_LEAF, ent->tlbi.leaf);
> > +		cmd[1] |= FIELD_PREP(CMDQ_TLBI_1_TTL, ent->tlbi.ttl);
> > +		cmd[1] |= FIELD_PREP(CMDQ_TLBI_1_TG, ent->tlbi.tg);
> > +		cmd[1] |= ent->tlbi.addr & CMDQ_TLBI_1_IPA_MASK;
> > +		break;
> > +	case CMDQ_OP_CMD_SYNC:
> > +		cmd[0] |= FIELD_PREP(CMDQ_SYNC_0_CS, CMDQ_SYNC_0_CS_SEV);
> > +		break;
> > +	default:
> > +		return -EINVAL;
> > +	}
> > +
> > +	for (i = 0; i < CMDQ_ENT_DWORDS; i++)
> > +		slot[i] = cpu_to_le64(cmd[i]);
> > +
> > +	smmu->cmdq_prod++;
> > +	writel(Q_IDX(smmu, smmu->cmdq_prod) | Q_WRAP(smmu, smmu->cmdq_prod),
> > +	       smmu->base + ARM_SMMU_CMDQ_PROD);
> > +	return 0;
> > +}
> > +
> > +static int smmu_sync_cmd(struct hyp_arm_smmu_v3_device *smmu)
> > +{
> > +	int ret;
> > +	struct arm_smmu_cmdq_ent cmd = {
> > +		.opcode = CMDQ_OP_CMD_SYNC,
> > +	};
> > +
> > +	ret = smmu_add_cmd(smmu, &cmd);
> > +	if (ret)
> > +		return ret;
> > +
> > +	return smmu_wait_event(smmu, smmu_cmdq_empty(smmu));
> > +}
> > +
> > +__maybe_unused
> > +static int smmu_send_cmd(struct hyp_arm_smmu_v3_device *smmu,
> > +			 struct arm_smmu_cmdq_ent *cmd)
> > +{
> > +	int ret = smmu_add_cmd(smmu, cmd);
> > +
> > +	if (ret)
> > +		return ret;
> > +
> > +	return smmu_sync_cmd(smmu);
> > +}
> > +
> >   static int smmu_init_registers(struct hyp_arm_smmu_v3_device *smmu)
> >   {
> >   	u64 val, old;
> > @@ -94,6 +220,41 @@ static int smmu_init_registers(struct hyp_arm_smmu_v3_device *smmu)
> >   	return 0;
> >   }
> > +static int smmu_init_cmdq(struct hyp_arm_smmu_v3_device *smmu)
> > +{
> > +	u64 cmdq_base;
> > +	size_t cmdq_nr_entries, cmdq_size;
> > +	int ret;
> > +	enum kvm_pgtable_prot prot = PAGE_HYP;
> > +
> > +	cmdq_base = readq_relaxed(smmu->base + ARM_SMMU_CMDQ_BASE);
> > +	if (cmdq_base & ~(Q_BASE_RWA | Q_BASE_ADDR_MASK | Q_BASE_LOG2SIZE))
> > +		return -EINVAL;
> > +
> > +	smmu->cmdq_log2size = cmdq_base & Q_BASE_LOG2SIZE;
> > +	cmdq_nr_entries = 1 << smmu->cmdq_log2size;
> > +	cmdq_size = cmdq_nr_entries * CMDQ_ENT_DWORDS * 8;
> > +
> > +	cmdq_base &= Q_BASE_ADDR_MASK;
> > +
> > +	if (!(smmu->features & ARM_SMMU_FEAT_COHERENCY))
> > +		prot |= KVM_PGTABLE_PROT_NORMAL_NC;
> > +
> > +	ret = ___pkvm_host_donate_hyp_prot(cmdq_base >> PAGE_SHIFT,
> > +					   PAGE_ALIGN(cmdq_size) >> PAGE_SHIFT,
> > +					   false, prot);
> > +	if (ret)
> > +		return ret;
> > +
> > +	smmu->cmdq_base = hyp_phys_to_virt(cmdq_base);
> > +
> > +	memset(smmu->cmdq_base, 0, cmdq_size);
> > +	writel_relaxed(0, smmu->base + ARM_SMMU_CMDQ_PROD);
> > +	writel_relaxed(0, smmu->base + ARM_SMMU_CMDQ_CONS);
> > +
> > +	return 0;
> > +}
> > +
> >   static int smmu_init_device(struct hyp_arm_smmu_v3_device *smmu)
> >   {
> >   	int ret;
> > @@ -113,6 +274,10 @@ static int smmu_init_device(struct hyp_arm_smmu_v3_device *smmu)
> >   	if (ret)
> >   		return ret;
> > +	ret = smmu_init_cmdq(smmu);
> > +	if (ret)
> > +		return ret;
> > +
> >   	return kvm_iommu_init_device(&smmu->iommu);
> >   }
> > diff --git a/include/kvm/arm_smmu_v3.h b/include/kvm/arm_smmu_v3.h
> > index fb24bcef1624..393a1a04edba 100644
> > --- a/include/kvm/arm_smmu_v3.h
> > +++ b/include/kvm/arm_smmu_v3.h
> > @@ -16,8 +16,12 @@ struct hyp_arm_smmu_v3_device {
> >   	struct kvm_hyp_iommu	iommu;
> >   	phys_addr_t		mmio_addr;
> >   	size_t			mmio_size;
> > +	unsigned long		features;
> >   	void __iomem		*base;
> > +	u32			cmdq_prod;
> > +	u64			*cmdq_base;
> > +	size_t			cmdq_log2size;
> >   };
> >   extern size_t kvm_nvhe_sym(kvm_hyp_arm_smmu_v3_count);
> 

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ