lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20150430043750.GH15238@voom.redhat.com>
Date:	Thu, 30 Apr 2015 14:37:50 +1000
From:	David Gibson <david@...son.dropbear.id.au>
To:	Alexey Kardashevskiy <aik@...abs.ru>
Cc:	linuxppc-dev@...ts.ozlabs.org,
	Benjamin Herrenschmidt <benh@...nel.crashing.org>,
	Paul Mackerras <paulus@...ba.org>,
	Alex Williamson <alex.williamson@...hat.com>,
	Gavin Shan <gwshan@...ux.vnet.ibm.com>,
	linux-kernel@...r.kernel.org
Subject: Re: [PATCH kernel v9 23/32] powerpc/powernv/ioda: Define and
 implement DMA table/window management callbacks

On Wed, Apr 29, 2015 at 07:44:20PM +1000, Alexey Kardashevskiy wrote:
> On 04/29/2015 03:30 PM, David Gibson wrote:
> >On Sat, Apr 25, 2015 at 10:14:47PM +1000, Alexey Kardashevskiy wrote:
> >>This extends iommu_table_group_ops by a set of callbacks to support
> >>dynamic DMA windows management.
> >>
> >>create_table() creates a TCE table with specific parameters.
> >>it receives iommu_table_group to know nodeid in order to allocate
> >>TCE table memory closer to the PHB. The exact format of allocated
> >>multi-level table might be also specific to the PHB model (not
> >>the case now though).
> >>This callback calculated the DMA window offset on a PCI bus from @num
> >>and stores it in a just created table.
> >>
> >>set_window() sets the window at specified TVT index + @num on PHB.
> >>
> >>unset_window() unsets the window from specified TVT.
> >>
> >>This adds a free() callback to iommu_table_ops to free the memory
> >>(potentially a tree of tables) allocated for the TCE table.
> >
> >Doesn't the free callback belong with the previous patch introducing
> >multi-level tables?
> 
> 
> 
> If I did that, you would say "why is it here if nothing calls it" on
> "multilevel" patch and "I see the allocation but I do not see memory
> release" ;)

Yeah, fair enough ;)

> I need some rule of thumb here. I think it is a bit cleaner if the same
> patch adds a callback for memory allocation and its counterpart, no?

On further consideration, yes, I think you're right.

> >>create_table() and free() are supposed to be called once per
> >>VFIO container and set_window()/unset_window() are supposed to be
> >>called for every group in a container.
> >>
> >>This adds IOMMU capabilities to iommu_table_group such as default
> >>32bit window parameters and others.
> >>
> >>Signed-off-by: Alexey Kardashevskiy <aik@...abs.ru>
> >>---
> >>  arch/powerpc/include/asm/iommu.h            | 19 ++++++++
> >>  arch/powerpc/platforms/powernv/pci-ioda.c   | 75 ++++++++++++++++++++++++++---
> >>  arch/powerpc/platforms/powernv/pci-p5ioc2.c | 12 +++--
> >>  3 files changed, 96 insertions(+), 10 deletions(-)
> >>
> >>diff --git a/arch/powerpc/include/asm/iommu.h b/arch/powerpc/include/asm/iommu.h
> >>index 0f50ee2..7694546 100644
> >>--- a/arch/powerpc/include/asm/iommu.h
> >>+++ b/arch/powerpc/include/asm/iommu.h
> >>@@ -70,6 +70,7 @@ struct iommu_table_ops {
> >>  	/* get() returns a physical address */
> >>  	unsigned long (*get)(struct iommu_table *tbl, long index);
> >>  	void (*flush)(struct iommu_table *tbl);
> >>+	void (*free)(struct iommu_table *tbl);
> >>  };
> >>
> >>  /* These are used by VIO */
> >>@@ -148,6 +149,17 @@ extern struct iommu_table *iommu_init_table(struct iommu_table * tbl,
> >>  struct iommu_table_group;
> >>
> >>  struct iommu_table_group_ops {
> >>+	long (*create_table)(struct iommu_table_group *table_group,
> >>+			int num,
> >>+			__u32 page_shift,
> >>+			__u64 window_size,
> >>+			__u32 levels,
> >>+			struct iommu_table *tbl);
> >>+	long (*set_window)(struct iommu_table_group *table_group,
> >>+			int num,
> >>+			struct iommu_table *tblnew);
> >>+	long (*unset_window)(struct iommu_table_group *table_group,
> >>+			int num);
> >>  	/*
> >>  	 * Switches ownership from the kernel itself to an external
> >>  	 * user. While onwership is taken, the kernel cannot use IOMMU itself.
> >>@@ -160,6 +172,13 @@ struct iommu_table_group {
> >>  #ifdef CONFIG_IOMMU_API
> >>  	struct iommu_group *group;
> >>  #endif
> >>+	/* Some key properties of IOMMU */
> >>+	__u32 tce32_start;
> >>+	__u32 tce32_size;
> >>+	__u64 pgsizes; /* Bitmap of supported page sizes */
> >>+	__u32 max_dynamic_windows_supported;
> >>+	__u32 max_levels;
> >
> >With this information, table_group seems even more like a bad name.
> >"iommu_state" maybe?
> 
> 
> Please, no. We will never come to agreement then :( And "iommu_state" is too
> general anyway, it won't pass.
> 
> 
> >>  	struct iommu_table tables[IOMMU_TABLE_GROUP_MAX_TABLES];
> >>  	struct iommu_table_group_ops *ops;
> >>  };
> >>diff --git a/arch/powerpc/platforms/powernv/pci-ioda.c b/arch/powerpc/platforms/powernv/pci-ioda.c
> >>index cc1d09c..4828837 100644
> >>--- a/arch/powerpc/platforms/powernv/pci-ioda.c
> >>+++ b/arch/powerpc/platforms/powernv/pci-ioda.c
> >>@@ -24,6 +24,7 @@
> >>  #include <linux/msi.h>
> >>  #include <linux/memblock.h>
> >>  #include <linux/iommu.h>
> >>+#include <linux/sizes.h>
> >>
> >>  #include <asm/sections.h>
> >>  #include <asm/io.h>
> >>@@ -1846,6 +1847,7 @@ static struct iommu_table_ops pnv_ioda2_iommu_ops = {
> >>  #endif
> >>  	.clear = pnv_ioda2_tce_free,
> >>  	.get = pnv_tce_get,
> >>+	.free = pnv_pci_free_table,
> >>  };
> >>
> >>  static void pnv_pci_ioda_setup_opal_tce_kill(struct pnv_phb *phb,
> >>@@ -1936,6 +1938,8 @@ static void pnv_pci_ioda_setup_dma_pe(struct pnv_phb *phb,
> >>  				 TCE_PCI_SWINV_PAIR);
> >>
> >>  	tbl->it_ops = &pnv_ioda1_iommu_ops;
> >>+	pe->table_group.tce32_start = tbl->it_offset << tbl->it_page_shift;
> >>+	pe->table_group.tce32_size = tbl->it_size << tbl->it_page_shift;
> >>  	iommu_init_table(tbl, phb->hose->node);
> >>
> >>  	if (pe->flags & PNV_IODA_PE_DEV) {
> >>@@ -1961,7 +1965,7 @@ static void pnv_pci_ioda_setup_dma_pe(struct pnv_phb *phb,
> >>  }
> >>
> >>  static long pnv_pci_ioda2_set_window(struct iommu_table_group *table_group,
> >>-		struct iommu_table *tbl)
> >>+		int num, struct iommu_table *tbl)
> >>  {
> >>  	struct pnv_ioda_pe *pe = container_of(table_group, struct pnv_ioda_pe,
> >>  			table_group);
> >>@@ -1972,9 +1976,10 @@ static long pnv_pci_ioda2_set_window(struct iommu_table_group *table_group,
> >>  	const __u64 start_addr = tbl->it_offset << tbl->it_page_shift;
> >>  	const __u64 win_size = tbl->it_size << tbl->it_page_shift;
> >>
> >>-	pe_info(pe, "Setting up window at %llx..%llx "
> >>+	pe_info(pe, "Setting up window#%d at %llx..%llx "
> >>  			"pgsize=0x%x tablesize=0x%lx "
> >>  			"levels=%d levelsize=%x\n",
> >>+			num,
> >>  			start_addr, start_addr + win_size - 1,
> >>  			1UL << tbl->it_page_shift, tbl->it_size << 3,
> >>  			tbl->it_indirect_levels + 1, tbl->it_level_size << 3);
> >>@@ -1987,7 +1992,7 @@ static long pnv_pci_ioda2_set_window(struct iommu_table_group *table_group,
> >>  	 */
> >>  	rc = opal_pci_map_pe_dma_window(phb->opal_id,
> >>  			pe->pe_number,
> >>-			pe->pe_number << 1,
> >>+			(pe->pe_number << 1) + num,
> >
> >Heh, yes, well, that makes it rather clear that only 2 tables are possible.
> >
> >>  			tbl->it_indirect_levels + 1,
> >>  			__pa(tbl->it_base),
> >>  			size << 3,
> >>@@ -2000,7 +2005,7 @@ static long pnv_pci_ioda2_set_window(struct iommu_table_group *table_group,
> >>  	pnv_pci_ioda2_tvt_invalidate(pe);
> >>
> >>  	/* Store fully initialized *tbl (may be external) in PE */
> >>-	pe->table_group.tables[0] = *tbl;
> >>+	pe->table_group.tables[num] = *tbl;
> >
> >I'm a bit confused by this whole set_window thing.  Is the idea that
> >with multiple groups in a container you have multiple table_group s
> >each with different copies of the iommu_table structures, but pointing
> >to the same actual TCE entries (it_base)?
> 
> Yes.
> 
> >It seems to me not terribly
> >obvious when you "create" a table and when you "set" a window.
> 
> 
> A table is not attached anywhere until its address is programmed (in
> set_window()) to the hardware, it is just a table in memory. For
> POWER8/IODA2, I create a table before I attach any group to a container,
> then I program this table to every attached container, right now it is done
> in container's attach_group(). So later we can hotplug any host PCI device
> to a container - it will program same TCE table to every new group in the
> container.

So you "create" once, then "set" it to one or more table_groups?  It
seems odd that "create" is a table_group callback in that case.

> >It's also kind of hard to assess whether the relative lifetimes are
> >correct of the table_group, struct iommu_table and the actual TCE tables.
> 
> That is true. Do not know how to improve this though.

So I think the scheme I suggested in reply to an earlier patch helps
this.  With that the lifetime of the struct iommu_table represents the
lifetime of the TCE table in the hardware sense as well, which I think
makes things clearer.

> 
> 
> >Would it make more sense for table_group to become the
> >non-vfio-specific counterpart to the vfio container?
> >i.e. representing one set of DMA mappings, which one or more PEs could
> >be bound to.
> 
> 
> table_group is embedded into PE and table/table_group callbacks access PE
> when invalidating TCE table. So I will need something to access PE. Or just
> have an array of 2 iommu_table.
> 
> 
> 
> >
> >>  	return 0;
> >>  fail:
> >>@@ -2061,6 +2066,53 @@ static void pnv_pci_ioda2_setup_bypass_pe(struct pnv_phb *phb,
> >>  }
> >>
> >>  #ifdef CONFIG_IOMMU_API
> >>+static long pnv_pci_ioda2_create_table(struct iommu_table_group *table_group,
> >>+		int num, __u32 page_shift, __u64 window_size, __u32 levels,
> >>+		struct iommu_table *tbl)
> >>+{
> >>+	struct pnv_ioda_pe *pe = container_of(table_group, struct pnv_ioda_pe,
> >>+			table_group);
> >>+	int nid = pe->phb->hose->node;
> >>+	__u64 bus_offset = num ? pe->tce_bypass_base : 0;
> >>+	long ret;
> >>+
> >>+	ret = pnv_pci_create_table(table_group, nid, bus_offset, page_shift,
> >>+			window_size, levels, tbl);
> >>+	if (ret)
> >>+		return ret;
> >>+
> >>+	tbl->it_ops = &pnv_ioda2_iommu_ops;
> >>+	if (pe->tce_inval_reg)
> >>+		tbl->it_type |= (TCE_PCI_SWINV_CREATE | TCE_PCI_SWINV_FREE);
> >>+
> >>+	return 0;
> >>+}
> >>+
> >>+static long pnv_pci_ioda2_unset_window(struct iommu_table_group *table_group,
> >>+		int num)
> >>+{
> >>+	struct pnv_ioda_pe *pe = container_of(table_group, struct pnv_ioda_pe,
> >>+			table_group);
> >>+	struct pnv_phb *phb = pe->phb;
> >>+	struct iommu_table *tbl = &pe->table_group.tables[num];
> >>+	long ret;
> >>+
> >>+	pe_info(pe, "Removing DMA window #%d\n", num);
> >>+
> >>+	ret = opal_pci_map_pe_dma_window(phb->opal_id, pe->pe_number,
> >>+			(pe->pe_number << 1) + num,
> >>+			0/* levels */, 0/* table address */,
> >>+			0/* table size */, 0/* page size */);
> >>+	if (ret)
> >>+		pe_warn(pe, "Unmapping failed, ret = %ld\n", ret);
> >>+	else
> >>+		pnv_pci_ioda2_tvt_invalidate(pe);
> >>+
> >>+	memset(tbl, 0, sizeof(*tbl));
> >>+
> >>+	return ret;
> >>+}
> >>+
> >>  static void pnv_ioda2_take_ownership(struct iommu_table_group *table_group)
> >>  {
> >>  	struct pnv_ioda_pe *pe = container_of(table_group, struct pnv_ioda_pe,
> >>@@ -2080,6 +2132,9 @@ static void pnv_ioda2_release_ownership(struct iommu_table_group *table_group)
> >>  }
> >>
> >>  static struct iommu_table_group_ops pnv_pci_ioda2_ops = {
> >>+	.create_table = pnv_pci_ioda2_create_table,
> >>+	.set_window = pnv_pci_ioda2_set_window,
> >>+	.unset_window = pnv_pci_ioda2_unset_window,
> >>  	.take_ownership = pnv_ioda2_take_ownership,
> >>  	.release_ownership = pnv_ioda2_release_ownership,
> >>  };
> >>@@ -2102,8 +2157,16 @@ static void pnv_pci_ioda2_setup_dma_pe(struct pnv_phb *phb,
> >>  	pe_info(pe, "Setting up 32-bit TCE table at 0..%08x\n",
> >>  		phb->ioda.m32_pci_base);
> >>
> >>+	pe->table_group.tce32_start = 0;
> >>+	pe->table_group.tce32_size = phb->ioda.m32_pci_base;
> >>+	pe->table_group.max_dynamic_windows_supported =
> >>+			IOMMU_TABLE_GROUP_MAX_TABLES;
> >>+	pe->table_group.max_levels = POWERNV_IOMMU_MAX_LEVELS;
> >>+	pe->table_group.pgsizes = SZ_4K | SZ_64K | SZ_16M;
> >>+
> >>  	rc = pnv_pci_create_table(&pe->table_group, pe->phb->hose->node,
> >>-			0, IOMMU_PAGE_SHIFT_4K, phb->ioda.m32_pci_base,
> >>+			pe->table_group.tce32_start, IOMMU_PAGE_SHIFT_4K,
> >>+			pe->table_group.tce32_size,
> >>  			POWERNV_IOMMU_DEFAULT_LEVELS, tbl);
> >>  	if (rc) {
> >>  		pe_err(pe, "Failed to create 32-bit TCE table, err %ld", rc);
> >>@@ -2119,7 +2182,7 @@ static void pnv_pci_ioda2_setup_dma_pe(struct pnv_phb *phb,
> >>  	pe->table_group.ops = &pnv_pci_ioda2_ops;
> >>  #endif
> >>
> >>-	rc = pnv_pci_ioda2_set_window(&pe->table_group, tbl);
> >>+	rc = pnv_pci_ioda2_set_window(&pe->table_group, 0, tbl);
> >>  	if (rc) {
> >>  		pe_err(pe, "Failed to configure 32-bit TCE table,"
> >>  		       " err %ld\n", rc);
> >>diff --git a/arch/powerpc/platforms/powernv/pci-p5ioc2.c b/arch/powerpc/platforms/powernv/pci-p5ioc2.c
> >>index 7a6fd92..d9de4c7 100644
> >>--- a/arch/powerpc/platforms/powernv/pci-p5ioc2.c
> >>+++ b/arch/powerpc/platforms/powernv/pci-p5ioc2.c
> >>@@ -116,6 +116,8 @@ static void __init pnv_pci_init_p5ioc2_phb(struct device_node *np, u64 hub_id,
> >>  	u64 phb_id;
> >>  	int64_t rc;
> >>  	static int primary = 1;
> >>+	struct iommu_table_group *table_group;
> >>+	struct iommu_table *tbl;
> >>
> >>  	pr_info(" Initializing p5ioc2 PHB %s\n", np->full_name);
> >>
> >>@@ -181,14 +183,16 @@ static void __init pnv_pci_init_p5ioc2_phb(struct device_node *np, u64 hub_id,
> >>  	pnv_pci_init_p5ioc2_msis(phb);
> >>
> >>  	/* Setup iommu */
> >>-	phb->p5ioc2.table_group.tables[0].it_table_group =
> >>-			&phb->p5ioc2.table_group;
> >>+	table_group = &phb->p5ioc2.table_group;
> >>+	tbl = &phb->p5ioc2.table_group.tables[0];
> >>+	tbl->it_table_group = table_group;
> >>
> >>  	/* Setup TCEs */
> >>  	phb->dma_dev_setup = pnv_pci_p5ioc2_dma_dev_setup;
> >>-	pnv_pci_setup_iommu_table(&phb->p5ioc2.table_group.tables[0],
> >>-				  tce_mem, tce_size, 0,
> >>+	pnv_pci_setup_iommu_table(tbl, tce_mem, tce_size, 0,
> >>  				  IOMMU_PAGE_SHIFT_4K);
> >>+	table_group->tce32_start = tbl->it_offset << tbl->it_page_shift;
> >>+	table_group->tce32_size = tbl->it_size << tbl->it_page_shift;
> >
> >Doesn't pgsizes need to be set here (although it will only include 4K,
> >I'm assuming).
> 
> 
> No, pgsizes are not returned to the userspace for p5ioc2/ioda1 as they do
> not support DDW. No pgsize => no DDW.

Ah, ok.

-- 
David Gibson			| I'll have my music baroque, and my code
david AT gibson.dropbear.id.au	| minimalist, thank you.  NOT _the_ _other_
				| _way_ _around_!
http://www.ozlabs.org/~dgibson

Content of type "application/pgp-signature" skipped

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ