lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20091013053726.GE17547@sequoia.sous-sol.org>
Date:	Mon, 12 Oct 2009 22:37:26 -0700
From:	Chris Wright <chrisw@...s-sol.org>
To:	Alok Kataria <akataria@...are.com>
Cc:	Chris Wright <chrisw@...s-sol.org>,
	James Bottomley <James.Bottomley@...e.de>,
	Randy Dunlap <randy.dunlap@...cle.com>,
	Mike Christie <michaelc@...wisc.edu>,
	Bart Van Assche <bvanassche@....org>,
	"linux-scsi@...r.kernel.org" <linux-scsi@...r.kernel.org>,
	Matthew Wilcox <matthew@....cx>,
	"pv-drivers@...are.com" <pv-drivers@...are.com>,
	Roland Dreier <rdreier@...co.com>,
	LKML <linux-kernel@...r.kernel.org>,
	"Chetan.Loke@...lex.Com" <Chetan.Loke@...lex.Com>,
	Brian King <brking@...ux.vnet.ibm.com>,
	Rolf Eike Beer <eike-kernel@...tec.de>,
	Robert Love <robert.w.love@...el.com>,
	Andrew Morton <akpm@...ux-foundation.org>,
	Daniel Walker <dwalker@...o99.com>, Greg KH <gregkh@...e.de>,
	"virtualization@...ts.linux-foundataion.org" 
	<virtualization@...ts.linux-foundataion.org>
Subject: Re: SCSI driver for VMware's virtual HBA - V5.

mostly just nits

* Alok Kataria (akataria@...are.com) wrote:
> +#include <linux/kernel.h>
> +#include <linux/module.h>
> +#include <linux/moduleparam.h>

shouldn't be needed

> +#include <linux/types.h>

shouldn't be needed

> +#include <linux/interrupt.h>
> +#include <linux/workqueue.h>
> +#include <linux/pci.h>
> +
> +#include <scsi/scsi.h>
> +#include <scsi/scsi_host.h>
> +#include <scsi/scsi_cmnd.h>
> +#include <scsi/scsi_device.h>
> +
> +#include "vmw_pvscsi.h"
> +
> +#define PVSCSI_LINUX_DRIVER_DESC "VMware PVSCSI driver"
> +
> +MODULE_DESCRIPTION(PVSCSI_LINUX_DRIVER_DESC);
> +MODULE_AUTHOR("VMware, Inc.");
> +MODULE_LICENSE("GPL");
> +MODULE_VERSION(PVSCSI_DRIVER_VERSION_STRING);
> +
> +#define PVSCSI_DEFAULT_NUM_PAGES_PER_RING	8
> +#define PVSCSI_DEFAULT_NUM_PAGES_MSG_RING	1
> +#define PVSCSI_DEFAULT_QUEUE_DEPTH		64
> +#define SGL_SIZE				PAGE_SIZE
> +
> +#define pvscsi_dev(adapter) (&(adapter->dev->dev))

easy to make it static inline and get some type checking for free

> +struct pvscsi_sg_list {
> +	struct PVSCSISGElement sge[PVSCSI_MAX_NUM_SG_ENTRIES_PER_SEGMENT];
> +};
> +
> +struct pvscsi_ctx {
> +	/*
> +	 * The index of the context in cmd_map serves as the context ID for a
> +	 * 1-to-1 mapping completions back to requests.
> +	 */
> +	struct scsi_cmnd	*cmd;
> +	struct pvscsi_sg_list	*sgl;
> +	struct list_head	list;
> +	dma_addr_t		dataPA;
> +	dma_addr_t		sensePA;
> +	dma_addr_t		sglPA;
> +};
> +
> +struct pvscsi_adapter {
> +	char				*mmioBase;
> +	unsigned int			irq;
> +	u8				rev;
> +	bool				use_msi;
> +	bool				use_msix;
> +	bool				use_msg;
> +
> +	spinlock_t			hw_lock;
> +
> +	struct workqueue_struct		*workqueue;
> +	struct work_struct		work;
> +
> +	struct PVSCSIRingReqDesc	*req_ring;
> +	unsigned			req_pages;
> +	unsigned			req_depth;
> +	dma_addr_t			reqRingPA;
> +
> +	struct PVSCSIRingCmpDesc	*cmp_ring;
> +	unsigned			cmp_pages;
> +	dma_addr_t			cmpRingPA;
> +
> +	struct PVSCSIRingMsgDesc	*msg_ring;
> +	unsigned			msg_pages;
> +	dma_addr_t			msgRingPA;
> +
> +	struct PVSCSIRingsState		*rings_state;
> +	dma_addr_t			ringStatePA;
> +
> +	struct pci_dev			*dev;
> +	struct Scsi_Host		*host;
> +
> +	struct list_head		cmd_pool;
> +	struct pvscsi_ctx		*cmd_map;
> +};
> +
> +
> +/* Command line parameters */
> +static int pvscsi_ring_pages     = PVSCSI_DEFAULT_NUM_PAGES_PER_RING;
> +static int pvscsi_msg_ring_pages = PVSCSI_DEFAULT_NUM_PAGES_MSG_RING;
> +static int pvscsi_cmd_per_lun    = PVSCSI_DEFAULT_QUEUE_DEPTH;
> +static bool pvscsi_disable_msi;
> +static bool pvscsi_disable_msix;
> +static bool pvscsi_use_msg       = true;
> +
> +#define PVSCSI_RW (S_IRUSR | S_IWUSR)
> +
> +module_param_named(ring_pages, pvscsi_ring_pages, int, PVSCSI_RW);
> +MODULE_PARM_DESC(ring_pages, "Number of pages per req/cmp ring - (default="
> +		 __stringify(PVSCSI_DEFAULT_NUM_PAGES_PER_RING) ")");
> +
> +module_param_named(msg_ring_pages, pvscsi_msg_ring_pages, int, PVSCSI_RW);
> +MODULE_PARM_DESC(msg_ring_pages, "Number of pages for the msg ring - (default="
> +		 __stringify(PVSCSI_DEFAULT_NUM_PAGES_MSG_RING) ")");
> +
> +module_param_named(cmd_per_lun, pvscsi_cmd_per_lun, int, PVSCSI_RW);
> +MODULE_PARM_DESC(cmd_per_lun, "Maximum commands per lun - (default="
> +		 __stringify(PVSCSI_MAX_REQ_QUEUE_DEPTH) ")");
> +
> +module_param_named(disable_msi, pvscsi_disable_msi, bool, PVSCSI_RW);
> +MODULE_PARM_DESC(disable_msi, "Disable MSI use in driver - (default=0)");
> +
> +module_param_named(disable_msix, pvscsi_disable_msix, bool, PVSCSI_RW);
> +MODULE_PARM_DESC(disable_msix, "Disable MSI-X use in driver - (default=0)");
> +
> +module_param_named(use_msg, pvscsi_use_msg, bool, PVSCSI_RW);
> +MODULE_PARM_DESC(use_msg, "Use msg ring when available - (default=1)");
> +
> +static const struct pci_device_id pvscsi_pci_tbl[] = {
> +	{ PCI_VDEVICE(VMWARE, PCI_DEVICE_ID_VMWARE_PVSCSI) },
> +	{ 0 }
> +};
> +
> +MODULE_DEVICE_TABLE(pci, pvscsi_pci_tbl);
> +
> +static struct pvscsi_ctx *
> +pvscsi_find_context(const struct pvscsi_adapter *adapter, struct scsi_cmnd *cmd)
> +{
> +	struct pvscsi_ctx *ctx, *end;
> +
> +	end = &adapter->cmd_map[adapter->req_depth];
> +	for (ctx = adapter->cmd_map; ctx < end; ctx++)
> +		if (ctx->cmd == cmd)
> +			return ctx;
> +
> +	return NULL;
> +}
> +
> +static struct pvscsi_ctx *
> +pvscsi_acquire_context(struct pvscsi_adapter *adapter, struct scsi_cmnd *cmd)
> +{
> +	struct pvscsi_ctx *ctx;
> +
> +	if (list_empty(&adapter->cmd_pool))
> +		return NULL;
> +
> +	ctx = list_first_entry(&adapter->cmd_pool, struct pvscsi_ctx, list);
> +	ctx->cmd = cmd;
> +	list_del(&ctx->list);
> +
> +	return ctx;
> +}
> +
> +static void pvscsi_release_context(struct pvscsi_adapter *adapter,
> +				   struct pvscsi_ctx *ctx)
> +{
> +	ctx->cmd = NULL;
> +	list_add(&ctx->list, &adapter->cmd_pool);
> +}

These list manipulations are protected by hw_lock?  Looks like all cases
are covered.

<snip>
> +/*
> + * Allocate scatter gather lists.
> + *
> + * These are statically allocated.  Trying to be clever was not worth it.
> + *
> + * Dynamic allocation can fail, and we can't go deeep into the memory
> + * allocator, since we're a SCSI driver, and trying too hard to allocate
> + * memory might generate disk I/O.  We also don't want to fail disk I/O
> + * in that case because we can't get an allocation - the I/O could be
> + * trying to swap out data to free memory.  Since that is pathological,
> + * just use a statically allocated scatter list.
> + *
> + */
> +static int __devinit pvscsi_allocate_sg(struct pvscsi_adapter *adapter)
> +{
> +	struct pvscsi_ctx *ctx;
> +	int i;
> +
> +	ctx = adapter->cmd_map;
> +	BUILD_BUG_ON(sizeof(struct pvscsi_sg_list) > SGL_SIZE);
> +
> +	for (i = 0; i < adapter->req_depth; ++i, ++ctx) {
> +		ctx->sgl = kmalloc(SGL_SIZE, GFP_KERNEL);
> +		ctx->sglPA = 0;
> +		BUG_ON(!IS_ALIGNED(((unsigned long)ctx->sgl), PAGE_SIZE));

Why not simply allocate a page?  Seems different allocator or debugging
options could trigger this.

> +		if (!ctx->sgl) {
> +			for (; i >= 0; --i, --ctx) {
> +				kfree(ctx->sgl);
> +				ctx->sgl = NULL;
> +			}
> +			return -ENOMEM;
> +		}
> +	}
> +
> +	return 0;
> +}
> +
> +static int __devinit pvscsi_probe(struct pci_dev *pdev,
> +				  const struct pci_device_id *id)
> +{
> +	struct pvscsi_adapter *adapter;
> +	struct Scsi_Host *host;
> +	unsigned int i;
> +	int error;
> +
> +	error = -ENODEV;
> +
> +	if (pci_enable_device(pdev))
> +		return error;

looks mmio only, pci_enable_device_mem()

> +
> +	if (pci_set_dma_mask(pdev, DMA_BIT_MASK(64)) == 0 &&
> +	    pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(64)) == 0) {
> +		printk(KERN_INFO "vmw_pvscsi: using 64bit dma\n");
> +	} else if (pci_set_dma_mask(pdev, DMA_BIT_MASK(32)) == 0 &&
> +		   pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(32)) == 0) {
> +		printk(KERN_INFO "vmw_pvscsi: using 32bit dma\n");
> +	} else {
> +		printk(KERN_ERR "vmw_pvscsi: failed to set DMA mask\n");
> +		goto out_disable_device;
> +	}
> +
> +	pvscsi_template.can_queue =
> +		min(PVSCSI_MAX_NUM_PAGES_REQ_RING, pvscsi_ring_pages) *
> +		PVSCSI_MAX_NUM_REQ_ENTRIES_PER_PAGE;
> +	pvscsi_template.cmd_per_lun =
> +		min(pvscsi_template.can_queue, pvscsi_cmd_per_lun);

When/how are these tunables used?  Are they still useful?

> +	host = scsi_host_alloc(&pvscsi_template, sizeof(struct pvscsi_adapter));
> +	if (!host) {
> +		printk(KERN_ERR "vmw_pvscsi: failed to allocate host\n");
> +		goto out_disable_device;
> +	}
> +
> +	adapter = shost_priv(host);
> +	memset(adapter, 0, sizeof(*adapter));
> +	adapter->dev  = pdev;
> +	adapter->host = host;
> +
> +	spin_lock_init(&adapter->hw_lock);
> +
> +	host->max_channel = 0;
> +	host->max_id      = 16;
> +	host->max_lun     = 1;
> +	host->max_cmd_len = 16;
> +
> +	adapter->rev = pdev->revision;
> +
> +	if (pci_request_regions(pdev, "vmw_pvscsi")) {
> +		printk(KERN_ERR "vmw_pvscsi: pci memory selection failed\n");
> +		goto out_free_host;
> +	}
> +
> +	for (i = 0; i < DEVICE_COUNT_RESOURCE; i++) {
> +		if ((pci_resource_flags(pdev, i) & PCI_BASE_ADDRESS_SPACE_IO))
> +			continue;
> +
> +		if (pci_resource_len(pdev, i) < PVSCSI_MEM_SPACE_SIZE)
> +			continue;
> +
> +		break;
> +	}
> +
> +	if (i == DEVICE_COUNT_RESOURCE) {
> +		printk(KERN_ERR
> +		       "vmw_pvscsi: adapter has no suitable MMIO region\n");
> +		goto out_release_resources;
> +	}

Could simplify and just do pci_request_selected_regions.

> +	adapter->mmioBase = pci_iomap(pdev, i, PVSCSI_MEM_SPACE_SIZE);
> +
> +	if (!adapter->mmioBase) {
> +		printk(KERN_ERR
> +		       "vmw_pvscsi: can't iomap for BAR %d memsize %lu\n",
> +		       i, PVSCSI_MEM_SPACE_SIZE);
> +		goto out_release_resources;
> +	}
> +
> +	pci_set_master(pdev);
> +	pci_set_drvdata(pdev, host);
> +
> +	ll_adapter_reset(adapter);
> +
> +	adapter->use_msg = pvscsi_setup_msg_workqueue(adapter);
> +
> +	error = pvscsi_allocate_rings(adapter);
> +	if (error) {
> +		printk(KERN_ERR "vmw_pvscsi: unable to allocate ring memory\n");
> +		goto out_release_resources;
> +	}
> +
> +	/*
> +	 * From this point on we should reset the adapter if anything goes
> +	 * wrong.
> +	 */
> +	pvscsi_setup_all_rings(adapter);
> +
> +	adapter->cmd_map = kcalloc(adapter->req_depth,
> +				   sizeof(struct pvscsi_ctx), GFP_KERNEL);
> +	if (!adapter->cmd_map) {
> +		printk(KERN_ERR "vmw_pvscsi: failed to allocate memory.\n");
> +		error = -ENOMEM;
> +		goto out_reset_adapter;
> +	}
> +
> +	INIT_LIST_HEAD(&adapter->cmd_pool);
> +	for (i = 0; i < adapter->req_depth; i++) {
> +		struct pvscsi_ctx *ctx = adapter->cmd_map + i;
> +		list_add(&ctx->list, &adapter->cmd_pool);
> +	}
> +
> +	error = pvscsi_allocate_sg(adapter);
> +	if (error) {
> +		printk(KERN_ERR "vmw_pvscsi: unable to allocate s/g table\n");
> +		goto out_reset_adapter;
> +	}
> +
> +	if (!pvscsi_disable_msix &&
> +	    pvscsi_setup_msix(adapter, &adapter->irq) == 0) {
> +		printk(KERN_INFO "vmw_pvscsi: using MSI-X\n");
> +		adapter->use_msix = 1;
> +	} else if (!pvscsi_disable_msi && pci_enable_msi(pdev) == 0) {
> +		printk(KERN_INFO "vmw_pvscsi: using MSI\n");
> +		adapter->use_msi = 1;
> +		adapter->irq = pdev->irq;
> +	} else {
> +		printk(KERN_INFO "vmw_pvscsi: using INTx\n");
> +		adapter->irq = pdev->irq;
> +	}
> +
> +	error = request_irq(adapter->irq, pvscsi_isr, IRQF_SHARED,
> +			    "vmw_pvscsi", adapter);

Typically IRQF_SHARED w/ INTx, not MSI and MSI-X.

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ