[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <2721c3ee-88d1-a8e9-1f1e-ffc3eef1d1ca@redhat.com>
Date: Thu, 20 Sep 2018 13:21:18 +0200
From: David Hildenbrand <david@...hat.com>
To: Pankaj Gupta <pagupta@...hat.com>, linux-kernel@...r.kernel.org,
kvm@...r.kernel.org, qemu-devel@...gnu.org,
linux-nvdimm@...1.01.org
Cc: jack@...e.cz, stefanha@...hat.com, dan.j.williams@...el.com,
riel@...riel.com, nilal@...hat.com, kwolf@...hat.com,
pbonzini@...hat.com, ross.zwisler@...el.com,
xiaoguangrong.eric@...il.com, hch@...radead.org, mst@...hat.com,
niteshnarayanlal@...mail.com, lcapitulino@...hat.com,
imammedo@...hat.com, eblake@...hat.com
Subject: Re: [PATCH] qemu: Add virtio pmem device
> @@ -0,0 +1,241 @@
> +/*
> + * Virtio pmem device
> + *
> + * Copyright (C) 2018 Red Hat, Inc.
> + * Copyright (C) 2018 Pankaj Gupta <pagupta@...hat.com>
> + *
> + * This work is licensed under the terms of the GNU GPL, version 2.
> + * See the COPYING file in the top-level directory.
> + *
> + */
> +
> +#include "qemu/osdep.h"
> +#include "qapi/error.h"
> +#include "qemu-common.h"
> +#include "qemu/error-report.h"
> +#include "hw/virtio/virtio-access.h"
> +#include "hw/virtio/virtio-pmem.h"
> +#include "hw/mem/memory-device.h"
> +#include "block/aio.h"
> +#include "block/thread-pool.h"
> +
> +typedef struct VirtIOPMEMresp {
> + int ret;
> +} VirtIOPMEMResp;
> +
> +typedef struct VirtIODeviceRequest {
> + VirtQueueElement elem;
> + int fd;
> + VirtIOPMEM *pmem;
> + VirtIOPMEMResp resp;
> +} VirtIODeviceRequest;
Both, response and request have to go to a linux header (and a header
sync patch).
Also, you are using the same request for host<->guest handling and
internal purposes. The fd or pmem pointer definitely don't belong here.
Use a separate struct for internal handling purposes. (passing to worker_cb)
> +
> +static int worker_cb(void *opaque)
> +{
> + VirtIODeviceRequest *req = opaque;
> + int err = 0;
> +
> + /* flush raw backing image */
> + err = fsync(req->fd);
> + if (err != 0) {
> + err = EIO;
> + }
> + req->resp.ret = err;
> +
> + return 0;
> +}
> +
> +static void done_cb(void *opaque, int ret)
> +{
> + VirtIODeviceRequest *req = opaque;
> + int len = iov_from_buf(req->elem.in_sg, req->elem.in_num, 0,
> + &req->resp, sizeof(VirtIOPMEMResp));
> +
> + /* Callbacks are serialized, so no need to use atomic ops. */
> + virtqueue_push(req->pmem->rq_vq, &req->elem, len);
> + virtio_notify((VirtIODevice *)req->pmem, req->pmem->rq_vq);
> + g_free(req);
> +}
> +
> +static void virtio_pmem_flush(VirtIODevice *vdev, VirtQueue *vq)
> +{
> + VirtIODeviceRequest *req;
> + VirtIOPMEM *pmem = VIRTIO_PMEM(vdev);
> + HostMemoryBackend *backend = MEMORY_BACKEND(pmem->memdev);
> + ThreadPool *pool = aio_get_thread_pool(qemu_get_aio_context());
> +
> + req = virtqueue_pop(vq, sizeof(VirtIODeviceRequest));
> + if (!req) {
> + virtio_error(vdev, "virtio-pmem missing request data");
> + return;
> + }
> +
> + if (req->elem.out_num < 1 || req->elem.in_num < 1) {
> + virtio_error(vdev, "virtio-pmem request not proper");
> + g_free(req);
> + return;
> + }
> + req->fd = memory_region_get_fd(&backend->mr);
> + req->pmem = pmem;
> + thread_pool_submit_aio(pool, worker_cb, req, done_cb, req);
> +}
> +
> +static void virtio_pmem_get_config(VirtIODevice *vdev, uint8_t *config)
> +{
> + VirtIOPMEM *pmem = VIRTIO_PMEM(vdev);
> + struct virtio_pmem_config *pmemcfg = (struct virtio_pmem_config *) config;
> +
> + virtio_stq_p(vdev, &pmemcfg->start, pmem->start);
> + virtio_stq_p(vdev, &pmemcfg->size, pmem->size);
> +}
> +
> +static uint64_t virtio_pmem_get_features(VirtIODevice *vdev, uint64_t features,
> + Error **errp)
> +{
> + return features;
> +}
> +
> +static void virtio_pmem_realize(DeviceState *dev, Error **errp)
> +{
> + VirtIODevice *vdev = VIRTIO_DEVICE(dev);
> + VirtIOPMEM *pmem = VIRTIO_PMEM(dev);
> + MachineState *ms = MACHINE(qdev_get_machine());
> + uint64_t align;
> + Error *local_err = NULL;
> + MemoryRegion *mr;
> +
> + if (!pmem->memdev) {
> + error_setg(errp, "virtio-pmem memdev not set");
> + return;
> + }
> +
> + mr = host_memory_backend_get_memory(pmem->memdev);
> + align = memory_region_get_alignment(mr);
> + pmem->size = QEMU_ALIGN_DOWN(memory_region_size(mr), align);
> + pmem->start = memory_device_get_free_addr(ms, NULL, align, pmem->size,
> + &local_err);
> + if (local_err) {
> + error_setg(errp, "Can't get free address in mem device");
> + return;
> + }
> + memory_region_init_alias(&pmem->mr, OBJECT(pmem),
> + "virtio_pmem-memory", mr, 0, pmem->size);
> + memory_device_plug_region(ms, &pmem->mr, pmem->start);
> +
> + host_memory_backend_set_mapped(pmem->memdev, true);
> + virtio_init(vdev, TYPE_VIRTIO_PMEM, VIRTIO_ID_PMEM,
> + sizeof(struct virtio_pmem_config));
> + pmem->rq_vq = virtio_add_queue(vdev, 128, virtio_pmem_flush);
> +}
> +
> +static void virtio_mem_check_memdev(Object *obj, const char *name, Object *val,
> + Error **errp)
> +{
> + if (host_memory_backend_is_mapped(MEMORY_BACKEND(val))) {
> + char *path = object_get_canonical_path_component(val);
> + error_setg(errp, "Can't use already busy memdev: %s", path);
> + g_free(path);
> + return;
> + }
> +
> + qdev_prop_allow_set_link_before_realize(obj, name, val, errp);
> +}
> +
> +static const char *virtio_pmem_get_device_id(VirtIOPMEM *vm)
> +{
> + Object *obj = OBJECT(vm);
> + DeviceState *parent_dev;
> +
> + /* always use the ID of the proxy device */
> + if (obj->parent && object_dynamic_cast(obj->parent, TYPE_DEVICE)) {
> + parent_dev = DEVICE(obj->parent);
> + return parent_dev->id;
> + }
> + return NULL;
> +}
> +
> +static void virtio_pmem_md_fill_device_info(const MemoryDeviceState *md,
> + MemoryDeviceInfo *info)
> +{
> + VirtioPMemDeviceInfo *vi = g_new0(VirtioPMemDeviceInfo, 1);
> + VirtIOPMEM *vm = VIRTIO_PMEM(md);
> + const char *id = virtio_pmem_get_device_id(vm);
> +
> + if (id) {
> + vi->has_id = true;
> + vi->id = g_strdup(id);
> + }
> +
> + vi->start = vm->start;
> + vi->size = vm->size;
> + vi->memdev = object_get_canonical_path(OBJECT(vm->memdev));
> +
> + info->u.virtio_pmem.data = vi;
> + info->type = MEMORY_DEVICE_INFO_KIND_VIRTIO_PMEM;
> +}
> +
> +static uint64_t virtio_pmem_md_get_addr(const MemoryDeviceState *md)
> +{
> + VirtIOPMEM *vm = VIRTIO_PMEM(md);
> +
> + return vm->start;
> +}
> +
> +static uint64_t virtio_pmem_md_get_plugged_size(const MemoryDeviceState *md)
> +{
> + VirtIOPMEM *vm = VIRTIO_PMEM(md);
> +
> + return vm->size;
> +}
> +
> +static uint64_t virtio_pmem_md_get_region_size(const MemoryDeviceState *md)
> +{
> + VirtIOPMEM *vm = VIRTIO_PMEM(md);
> +
> + return vm->size;
> +}
> +
> +static void virtio_pmem_instance_init(Object *obj)
> +{
> + VirtIOPMEM *vm = VIRTIO_PMEM(obj);
> + object_property_add_link(obj, "memdev", TYPE_MEMORY_BACKEND,
> + (Object **)&vm->memdev,
> + (void *) virtio_mem_check_memdev,
> + OBJ_PROP_LINK_STRONG,
> + &error_abort);
> +}
> +
> +
> +static void virtio_pmem_class_init(ObjectClass *klass, void *data)
> +{
> + VirtioDeviceClass *vdc = VIRTIO_DEVICE_CLASS(klass);
> + MemoryDeviceClass *mdc = MEMORY_DEVICE_CLASS(klass);
> +
> + vdc->realize = virtio_pmem_realize;
> + vdc->get_config = virtio_pmem_get_config;
> + vdc->get_features = virtio_pmem_get_features;
> +
> + mdc->get_addr = virtio_pmem_md_get_addr;
> + mdc->get_plugged_size = virtio_pmem_md_get_plugged_size;
> + mdc->get_region_size = virtio_pmem_md_get_region_size;
> + mdc->fill_device_info = virtio_pmem_md_fill_device_info;
> +}
> +
> +static TypeInfo virtio_pmem_info = {
> + .name = TYPE_VIRTIO_PMEM,
> + .parent = TYPE_VIRTIO_DEVICE,
> + .class_init = virtio_pmem_class_init,
> + .instance_size = sizeof(VirtIOPMEM),
> + .instance_init = virtio_pmem_instance_init,
> + .interfaces = (InterfaceInfo[]) {
> + { TYPE_MEMORY_DEVICE },
> + { }
> + },
> +};
> +
> +static void virtio_register_types(void)
> +{
> + type_register_static(&virtio_pmem_info);
> +}
> +
> +type_init(virtio_register_types)
> diff --git a/include/hw/pci/pci.h b/include/hw/pci/pci.h
> index 990d6fcbde..28829b6437 100644
> --- a/include/hw/pci/pci.h
> +++ b/include/hw/pci/pci.h
> @@ -85,6 +85,7 @@ extern bool pci_available;
> #define PCI_DEVICE_ID_VIRTIO_RNG 0x1005
> #define PCI_DEVICE_ID_VIRTIO_9P 0x1009
> #define PCI_DEVICE_ID_VIRTIO_VSOCK 0x1012
> +#define PCI_DEVICE_ID_VIRTIO_PMEM 0x1013
>
> #define PCI_VENDOR_ID_REDHAT 0x1b36
> #define PCI_DEVICE_ID_REDHAT_BRIDGE 0x0001
> diff --git a/include/hw/virtio/virtio-pmem.h b/include/hw/virtio/virtio-pmem.h
> new file mode 100644
> index 0000000000..fda3ee691c
> --- /dev/null
> +++ b/include/hw/virtio/virtio-pmem.h
> @@ -0,0 +1,42 @@
> +/*
> + * Virtio pmem Device
> + *
> + * Copyright Red Hat, Inc. 2018
> + * Copyright Pankaj Gupta <pagupta@...hat.com>
> + *
> + * This work is licensed under the terms of the GNU GPL, version 2 or
> + * (at your option) any later version. See the COPYING file in the
> + * top-level directory.
> + */
> +
> +#ifndef QEMU_VIRTIO_PMEM_H
> +#define QEMU_VIRTIO_PMEM_H
> +
> +#include "hw/virtio/virtio.h"
> +#include "exec/memory.h"
> +#include "sysemu/hostmem.h"
> +#include "standard-headers/linux/virtio_ids.h"
> +#include "hw/boards.h"
> +#include "hw/i386/pc.h"
> +
> +#define TYPE_VIRTIO_PMEM "virtio-pmem"
> +
> +#define VIRTIO_PMEM(obj) \
> + OBJECT_CHECK(VirtIOPMEM, (obj), TYPE_VIRTIO_PMEM)
> +
> +/* VirtIOPMEM device structure */
> +typedef struct VirtIOPMEM {
> + VirtIODevice parent_obj;
> +
> + VirtQueue *rq_vq;
> + uint64_t start;
> + uint64_t size;
> + MemoryRegion mr;
> + HostMemoryBackend *memdev;
> +} VirtIOPMEM;
> +
> +struct virtio_pmem_config {
> + uint64_t start;
> + uint64_t size;
> +};
> +#endif
> diff --git a/include/standard-headers/linux/virtio_ids.h b/include/standard-headers/linux/virtio_ids.h
> index 6d5c3b2d4f..346389565a 100644
> --- a/include/standard-headers/linux/virtio_ids.h
> +++ b/include/standard-headers/linux/virtio_ids.h
> @@ -43,5 +43,6 @@
> #define VIRTIO_ID_INPUT 18 /* virtio input */
> #define VIRTIO_ID_VSOCK 19 /* virtio vsock transport */
> #define VIRTIO_ID_CRYPTO 20 /* virtio crypto */
> +#define VIRTIO_ID_PMEM 25 /* virtio pmem */
This should be moved to a linux header sync patch.
--
Thanks,
David / dhildenb
Powered by blists - more mailing lists