lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CANr-nt2EAkd7uqCe2W14wxzYC3fSTCCHRJs65Zh_S4UrPfg6Hw@mail.gmail.com>
Date:   Thu, 1 Aug 2019 09:15:59 +0200
From:   Hans Holmberg <hans@...tronix.com>
To:     Javier González <javier@...igon.com>
Cc:     Matias Bjørling <mb@...htnvm.io>,
        Christoph Hellwig <hch@....de>, linux-block@...r.kernel.org,
        Linux Kernel Mailing List <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH 2/4] lightnvm: move metadata mapping to lower level driver

On Wed, Jul 31, 2019 at 3:59 PM Javier González <javier@...igon.com> wrote:
>
> > On 31 Jul 2019, at 11.41, Hans Holmberg <hans@...tronix.com> wrote:
> >
> > Now that blk_rq_map_kern can map both kmem and vmem, move
> > internal metadata mapping down to the lower level driver.
> >
> > Signed-off-by: Hans Holmberg <hans@...tronix.com>
> > ---
> > drivers/lightnvm/core.c          |  16 +++---
> > drivers/lightnvm/pblk-core.c     | 113 +++++----------------------------------
> > drivers/lightnvm/pblk-read.c     |  22 ++------
> > drivers/lightnvm/pblk-recovery.c |  39 ++------------
> > drivers/lightnvm/pblk-write.c    |  20 ++-----
> > drivers/lightnvm/pblk.h          |   8 +--
> > drivers/nvme/host/lightnvm.c     |  20 +++++--
> > include/linux/lightnvm.h         |   6 +--
> > 8 files changed, 54 insertions(+), 190 deletions(-)
> >
> > diff --git a/drivers/lightnvm/core.c b/drivers/lightnvm/core.c
> > index 01d098fb96ac..3cd03582a2ed 100644
> > --- a/drivers/lightnvm/core.c
> > +++ b/drivers/lightnvm/core.c
> > @@ -731,7 +731,7 @@ static int nvm_set_flags(struct nvm_geo *geo, struct nvm_rq *rqd)
> >       return flags;
> > }
> >
> > -int nvm_submit_io(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd)
> > +int nvm_submit_io(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd, void *buf)
> > {
> >       struct nvm_dev *dev = tgt_dev->parent;
> >       int ret;
> > @@ -745,7 +745,7 @@ int nvm_submit_io(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd)
> >       rqd->flags = nvm_set_flags(&tgt_dev->geo, rqd);
> >
> >       /* In case of error, fail with right address format */
> > -     ret = dev->ops->submit_io(dev, rqd);
> > +     ret = dev->ops->submit_io(dev, rqd, buf);
> >       if (ret)
> >               nvm_rq_dev_to_tgt(tgt_dev, rqd);
> >       return ret;
> > @@ -759,7 +759,8 @@ static void nvm_sync_end_io(struct nvm_rq *rqd)
> >       complete(waiting);
> > }
> >
> > -static int nvm_submit_io_wait(struct nvm_dev *dev, struct nvm_rq *rqd)
> > +static int nvm_submit_io_wait(struct nvm_dev *dev, struct nvm_rq *rqd,
> > +                           void *buf)
> > {
> >       DECLARE_COMPLETION_ONSTACK(wait);
> >       int ret = 0;
> > @@ -767,7 +768,7 @@ static int nvm_submit_io_wait(struct nvm_dev *dev, struct nvm_rq *rqd)
> >       rqd->end_io = nvm_sync_end_io;
> >       rqd->private = &wait;
> >
> > -     ret = dev->ops->submit_io(dev, rqd);
> > +     ret = dev->ops->submit_io(dev, rqd, buf);
> >       if (ret)
> >               return ret;
> >
> > @@ -776,7 +777,8 @@ static int nvm_submit_io_wait(struct nvm_dev *dev, struct nvm_rq *rqd)
> >       return 0;
> > }
> >
> > -int nvm_submit_io_sync(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd)
> > +int nvm_submit_io_sync(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd,
> > +                    void *buf)
> > {
> >       struct nvm_dev *dev = tgt_dev->parent;
> >       int ret;
> > @@ -789,7 +791,7 @@ int nvm_submit_io_sync(struct nvm_tgt_dev *tgt_dev, struct nvm_rq *rqd)
> >       rqd->dev = tgt_dev;
> >       rqd->flags = nvm_set_flags(&tgt_dev->geo, rqd);
> >
> > -     ret = nvm_submit_io_wait(dev, rqd);
> > +     ret = nvm_submit_io_wait(dev, rqd, buf);
> >
> >       return ret;
> > }
> > @@ -816,7 +818,7 @@ static int nvm_submit_io_sync_raw(struct nvm_dev *dev, struct nvm_rq *rqd)
> >       rqd->dev = NULL;
> >       rqd->flags = nvm_set_flags(&dev->geo, rqd);
> >
> > -     return nvm_submit_io_wait(dev, rqd);
> > +     return nvm_submit_io_wait(dev, rqd, NULL);
> > }
> >
> > static int nvm_bb_chunk_sense(struct nvm_dev *dev, struct ppa_addr ppa)
> > diff --git a/drivers/lightnvm/pblk-core.c b/drivers/lightnvm/pblk-core.c
> > index f546e6f28b8a..a58d3c84a3f2 100644
> > --- a/drivers/lightnvm/pblk-core.c
> > +++ b/drivers/lightnvm/pblk-core.c
> > @@ -507,7 +507,7 @@ void pblk_set_sec_per_write(struct pblk *pblk, int sec_per_write)
> >       pblk->sec_per_write = sec_per_write;
> > }
> >
> > -int pblk_submit_io(struct pblk *pblk, struct nvm_rq *rqd)
> > +int pblk_submit_io(struct pblk *pblk, struct nvm_rq *rqd, void *buf)
> > {
> >       struct nvm_tgt_dev *dev = pblk->dev;
> >
> > @@ -518,7 +518,7 @@ int pblk_submit_io(struct pblk *pblk, struct nvm_rq *rqd)
> >               return NVM_IO_ERR;
> > #endif
> >
> > -     return nvm_submit_io(dev, rqd);
> > +     return nvm_submit_io(dev, rqd, buf);
> > }
> >
> > void pblk_check_chunk_state_update(struct pblk *pblk, struct nvm_rq *rqd)
> > @@ -541,7 +541,7 @@ void pblk_check_chunk_state_update(struct pblk *pblk, struct nvm_rq *rqd)
> >       }
> > }
> >
> > -int pblk_submit_io_sync(struct pblk *pblk, struct nvm_rq *rqd)
> > +int pblk_submit_io_sync(struct pblk *pblk, struct nvm_rq *rqd, void *buf)
> > {
> >       struct nvm_tgt_dev *dev = pblk->dev;
> >       int ret;
> > @@ -553,7 +553,7 @@ int pblk_submit_io_sync(struct pblk *pblk, struct nvm_rq *rqd)
> >               return NVM_IO_ERR;
> > #endif
> >
> > -     ret = nvm_submit_io_sync(dev, rqd);
> > +     ret = nvm_submit_io_sync(dev, rqd, buf);
> >
> >       if (trace_pblk_chunk_state_enabled() && !ret &&
> >           rqd->opcode == NVM_OP_PWRITE)
> > @@ -562,65 +562,19 @@ int pblk_submit_io_sync(struct pblk *pblk, struct nvm_rq *rqd)
> >       return ret;
> > }
> >
> > -int pblk_submit_io_sync_sem(struct pblk *pblk, struct nvm_rq *rqd)
> > +static int pblk_submit_io_sync_sem(struct pblk *pblk, struct nvm_rq *rqd,
> > +                                void *buf)
> > {
> >       struct ppa_addr *ppa_list = nvm_rq_to_ppa_list(rqd);
> >       int ret;
> >
> >       pblk_down_chunk(pblk, ppa_list[0]);
> > -     ret = pblk_submit_io_sync(pblk, rqd);
> > +     ret = pblk_submit_io_sync(pblk, rqd, buf);
> >       pblk_up_chunk(pblk, ppa_list[0]);
> >
> >       return ret;
> > }
> >
> > -static void pblk_bio_map_addr_endio(struct bio *bio)
> > -{
> > -     bio_put(bio);
> > -}
> > -
> > -struct bio *pblk_bio_map_addr(struct pblk *pblk, void *data,
> > -                           unsigned int nr_secs, unsigned int len,
> > -                           int alloc_type, gfp_t gfp_mask)
> > -{
> > -     struct nvm_tgt_dev *dev = pblk->dev;
> > -     void *kaddr = data;
> > -     struct page *page;
> > -     struct bio *bio;
> > -     int i, ret;
> > -
> > -     if (alloc_type == PBLK_KMALLOC_META)
> > -             return bio_map_kern(dev->q, kaddr, len, gfp_mask);
> > -
> > -     bio = bio_kmalloc(gfp_mask, nr_secs);
> > -     if (!bio)
> > -             return ERR_PTR(-ENOMEM);
> > -
> > -     for (i = 0; i < nr_secs; i++) {
> > -             page = vmalloc_to_page(kaddr);
> > -             if (!page) {
> > -                     pblk_err(pblk, "could not map vmalloc bio\n");
> > -                     bio_put(bio);
> > -                     bio = ERR_PTR(-ENOMEM);
> > -                     goto out;
> > -             }
> > -
> > -             ret = bio_add_pc_page(dev->q, bio, page, PAGE_SIZE, 0);
> > -             if (ret != PAGE_SIZE) {
> > -                     pblk_err(pblk, "could not add page to bio\n");
> > -                     bio_put(bio);
> > -                     bio = ERR_PTR(-ENOMEM);
> > -                     goto out;
> > -             }
> > -
> > -             kaddr += PAGE_SIZE;
> > -     }
> > -
> > -     bio->bi_end_io = pblk_bio_map_addr_endio;
> > -out:
> > -     return bio;
> > -}
> > -
> > int pblk_calc_secs(struct pblk *pblk, unsigned long secs_avail,
> >                  unsigned long secs_to_flush, bool skip_meta)
> > {
> > @@ -722,9 +676,7 @@ u64 pblk_line_smeta_start(struct pblk *pblk, struct pblk_line *line)
> >
> > int pblk_line_smeta_read(struct pblk *pblk, struct pblk_line *line)
> > {
> > -     struct nvm_tgt_dev *dev = pblk->dev;
> >       struct pblk_line_meta *lm = &pblk->lm;
> > -     struct bio *bio;
> >       struct ppa_addr *ppa_list;
> >       struct nvm_rq rqd;
> >       u64 paddr = pblk_line_smeta_start(pblk, line);
> > @@ -736,16 +688,6 @@ int pblk_line_smeta_read(struct pblk *pblk, struct pblk_line *line)
> >       if (ret)
> >               return ret;
> >
> > -     bio = bio_map_kern(dev->q, line->smeta, lm->smeta_len, GFP_KERNEL);
> > -     if (IS_ERR(bio)) {
> > -             ret = PTR_ERR(bio);
> > -             goto clear_rqd;
> > -     }
> > -
> > -     bio->bi_iter.bi_sector = 0; /* internal bio */
> > -     bio_set_op_attrs(bio, REQ_OP_READ, 0);
> > -
> > -     rqd.bio = bio;
> >       rqd.opcode = NVM_OP_PREAD;
> >       rqd.nr_ppas = lm->smeta_sec;
> >       rqd.is_seq = 1;
> > @@ -754,10 +696,9 @@ int pblk_line_smeta_read(struct pblk *pblk, struct pblk_line *line)
> >       for (i = 0; i < lm->smeta_sec; i++, paddr++)
> >               ppa_list[i] = addr_to_gen_ppa(pblk, paddr, line->id);
> >
> > -     ret = pblk_submit_io_sync(pblk, &rqd);
> > +     ret = pblk_submit_io_sync(pblk, &rqd, line->smeta);
> >       if (ret) {
> >               pblk_err(pblk, "smeta I/O submission failed: %d\n", ret);
> > -             bio_put(bio);
> >               goto clear_rqd;
> >       }
> >
> > @@ -776,9 +717,7 @@ int pblk_line_smeta_read(struct pblk *pblk, struct pblk_line *line)
> > static int pblk_line_smeta_write(struct pblk *pblk, struct pblk_line *line,
> >                                u64 paddr)
> > {
> > -     struct nvm_tgt_dev *dev = pblk->dev;
> >       struct pblk_line_meta *lm = &pblk->lm;
> > -     struct bio *bio;
> >       struct ppa_addr *ppa_list;
> >       struct nvm_rq rqd;
> >       __le64 *lba_list = emeta_to_lbas(pblk, line->emeta->buf);
> > @@ -791,16 +730,6 @@ static int pblk_line_smeta_write(struct pblk *pblk, struct pblk_line *line,
> >       if (ret)
> >               return ret;
> >
> > -     bio = bio_map_kern(dev->q, line->smeta, lm->smeta_len, GFP_KERNEL);
> > -     if (IS_ERR(bio)) {
> > -             ret = PTR_ERR(bio);
> > -             goto clear_rqd;
> > -     }
> > -
> > -     bio->bi_iter.bi_sector = 0; /* internal bio */
> > -     bio_set_op_attrs(bio, REQ_OP_WRITE, 0);
> > -
> > -     rqd.bio = bio;
> >       rqd.opcode = NVM_OP_PWRITE;
> >       rqd.nr_ppas = lm->smeta_sec;
> >       rqd.is_seq = 1;
> > @@ -814,10 +743,9 @@ static int pblk_line_smeta_write(struct pblk *pblk, struct pblk_line *line,
> >               meta->lba = lba_list[paddr] = addr_empty;
> >       }
> >
> > -     ret = pblk_submit_io_sync_sem(pblk, &rqd);
> > +     ret = pblk_submit_io_sync_sem(pblk, &rqd, line->smeta);
> >       if (ret) {
> >               pblk_err(pblk, "smeta I/O submission failed: %d\n", ret);
> > -             bio_put(bio);
> >               goto clear_rqd;
> >       }
> >
> > @@ -838,10 +766,8 @@ int pblk_line_emeta_read(struct pblk *pblk, struct pblk_line *line,
> > {
> >       struct nvm_tgt_dev *dev = pblk->dev;
> >       struct nvm_geo *geo = &dev->geo;
> > -     struct pblk_line_mgmt *l_mg = &pblk->l_mg;
> >       struct pblk_line_meta *lm = &pblk->lm;
> >       void *ppa_list_buf, *meta_list;
> > -     struct bio *bio;
> >       struct ppa_addr *ppa_list;
> >       struct nvm_rq rqd;
> >       u64 paddr = line->emeta_ssec;
> > @@ -867,17 +793,6 @@ int pblk_line_emeta_read(struct pblk *pblk, struct pblk_line *line,
> >       rq_ppas = pblk_calc_secs(pblk, left_ppas, 0, false);
> >       rq_len = rq_ppas * geo->csecs;
> >
> > -     bio = pblk_bio_map_addr(pblk, emeta_buf, rq_ppas, rq_len,
> > -                                     l_mg->emeta_alloc_type, GFP_KERNEL);
> > -     if (IS_ERR(bio)) {
> > -             ret = PTR_ERR(bio);
> > -             goto free_rqd_dma;
> > -     }
> > -
> > -     bio->bi_iter.bi_sector = 0; /* internal bio */
> > -     bio_set_op_attrs(bio, REQ_OP_READ, 0);
> > -
> > -     rqd.bio = bio;
> >       rqd.meta_list = meta_list;
> >       rqd.ppa_list = ppa_list_buf;
> >       rqd.dma_meta_list = dma_meta_list;
> > @@ -896,7 +811,6 @@ int pblk_line_emeta_read(struct pblk *pblk, struct pblk_line *line,
> >               while (test_bit(pos, line->blk_bitmap)) {
> >                       paddr += min;
> >                       if (pblk_boundary_paddr_checks(pblk, paddr)) {
> > -                             bio_put(bio);
> >                               ret = -EINTR;
> >                               goto free_rqd_dma;
> >                       }
> > @@ -906,7 +820,6 @@ int pblk_line_emeta_read(struct pblk *pblk, struct pblk_line *line,
> >               }
> >
> >               if (pblk_boundary_paddr_checks(pblk, paddr + min)) {
> > -                     bio_put(bio);
> >                       ret = -EINTR;
> >                       goto free_rqd_dma;
> >               }
> > @@ -915,10 +828,9 @@ int pblk_line_emeta_read(struct pblk *pblk, struct pblk_line *line,
> >                       ppa_list[i] = addr_to_gen_ppa(pblk, paddr, line_id);
> >       }
> >
> > -     ret = pblk_submit_io_sync(pblk, &rqd);
> > +     ret = pblk_submit_io_sync(pblk, &rqd, emeta_buf);
> >       if (ret) {
> >               pblk_err(pblk, "emeta I/O submission failed: %d\n", ret);
> > -             bio_put(bio);
> >               goto free_rqd_dma;
> >       }
> >
> > @@ -963,7 +875,7 @@ static int pblk_blk_erase_sync(struct pblk *pblk, struct ppa_addr ppa)
> >       /* The write thread schedules erases so that it minimizes disturbances
> >        * with writes. Thus, there is no need to take the LUN semaphore.
> >        */
> > -     ret = pblk_submit_io_sync(pblk, &rqd);
> > +     ret = pblk_submit_io_sync(pblk, &rqd, NULL);
> >       rqd.private = pblk;
> >       __pblk_end_io_erase(pblk, &rqd);
> >
> > @@ -1792,7 +1704,7 @@ int pblk_blk_erase_async(struct pblk *pblk, struct ppa_addr ppa)
> >       /* The write thread schedules erases so that it minimizes disturbances
> >        * with writes. Thus, there is no need to take the LUN semaphore.
> >        */
> > -     err = pblk_submit_io(pblk, rqd);
> > +     err = pblk_submit_io(pblk, rqd, NULL);
> >       if (err) {
> >               struct nvm_tgt_dev *dev = pblk->dev;
> >               struct nvm_geo *geo = &dev->geo;
> > @@ -1923,7 +1835,6 @@ void pblk_line_close_meta(struct pblk *pblk, struct pblk_line *line)
> > static void pblk_save_lba_list(struct pblk *pblk, struct pblk_line *line)
> > {
> >       struct pblk_line_meta *lm = &pblk->lm;
> > -     struct pblk_line_mgmt *l_mg = &pblk->l_mg;
> >       unsigned int lba_list_size = lm->emeta_len[2];
> >       struct pblk_w_err_gc *w_err_gc = line->w_err_gc;
> >       struct pblk_emeta *emeta = line->emeta;
> > diff --git a/drivers/lightnvm/pblk-read.c b/drivers/lightnvm/pblk-read.c
> > index d98ea392fe33..d572d4559e4e 100644
> > --- a/drivers/lightnvm/pblk-read.c
> > +++ b/drivers/lightnvm/pblk-read.c
> > @@ -342,7 +342,7 @@ void pblk_submit_read(struct pblk *pblk, struct bio *bio)
> >               bio_put(int_bio);
> >               int_bio = bio_clone_fast(bio, GFP_KERNEL, &pblk_bio_set);
> >               goto split_retry;
> > -     } else if (pblk_submit_io(pblk, rqd)) {
> > +     } else if (pblk_submit_io(pblk, rqd, NULL)) {
> >               /* Submitting IO to drive failed, let's report an error */
> >               rqd->error = -ENODEV;
> >               pblk_end_io_read(rqd);
> > @@ -419,7 +419,6 @@ int pblk_submit_read_gc(struct pblk *pblk, struct pblk_gc_rq *gc_rq)
> > {
> >       struct nvm_tgt_dev *dev = pblk->dev;
> >       struct nvm_geo *geo = &dev->geo;
> > -     struct bio *bio;
> >       struct nvm_rq rqd;
> >       int data_len;
> >       int ret = NVM_IO_OK;
> > @@ -447,25 +446,12 @@ int pblk_submit_read_gc(struct pblk *pblk, struct pblk_gc_rq *gc_rq)
> >               goto out;
> >
> >       data_len = (gc_rq->secs_to_gc) * geo->csecs;
> > -     bio = pblk_bio_map_addr(pblk, gc_rq->data, gc_rq->secs_to_gc, data_len,
> > -                                             PBLK_VMALLOC_META, GFP_KERNEL);
> > -     if (IS_ERR(bio)) {
> > -             pblk_err(pblk, "could not allocate GC bio (%lu)\n",
> > -                                                             PTR_ERR(bio));
> > -             ret = PTR_ERR(bio);
> > -             goto err_free_dma;
> > -     }
> > -
> > -     bio->bi_iter.bi_sector = 0; /* internal bio */
> > -     bio_set_op_attrs(bio, REQ_OP_READ, 0);
> > -
> >       rqd.opcode = NVM_OP_PREAD;
> >       rqd.nr_ppas = gc_rq->secs_to_gc;
> > -     rqd.bio = bio;
> >
> > -     if (pblk_submit_io_sync(pblk, &rqd)) {
> > +     if (pblk_submit_io_sync(pblk, &rqd, gc_rq->data)) {
> >               ret = -EIO;
> > -             goto err_free_bio;
> > +             goto err_free_dma;
> >       }
> >
> >       pblk_read_check_rand(pblk, &rqd, gc_rq->lba_list, gc_rq->nr_secs);
> > @@ -489,8 +475,6 @@ int pblk_submit_read_gc(struct pblk *pblk, struct pblk_gc_rq *gc_rq)
> >       pblk_free_rqd_meta(pblk, &rqd);
> >       return ret;
> >
> > -err_free_bio:
> > -     bio_put(bio);
> > err_free_dma:
> >       pblk_free_rqd_meta(pblk, &rqd);
> >       return ret;
> > diff --git a/drivers/lightnvm/pblk-recovery.c b/drivers/lightnvm/pblk-recovery.c
> > index e6dda04de144..d5e210c3c5b7 100644
> > --- a/drivers/lightnvm/pblk-recovery.c
> > +++ b/drivers/lightnvm/pblk-recovery.c
> > @@ -178,12 +178,11 @@ static int pblk_recov_pad_line(struct pblk *pblk, struct pblk_line *line,
> >       void *meta_list;
> >       struct pblk_pad_rq *pad_rq;
> >       struct nvm_rq *rqd;
> > -     struct bio *bio;
> >       struct ppa_addr *ppa_list;
> >       void *data;
> >       __le64 *lba_list = emeta_to_lbas(pblk, line->emeta->buf);
> >       u64 w_ptr = line->cur_sec;
> > -     int left_line_ppas, rq_ppas, rq_len;
> > +     int left_line_ppas, rq_ppas;
> >       int i, j;
> >       int ret = 0;
> >
> > @@ -212,28 +211,15 @@ static int pblk_recov_pad_line(struct pblk *pblk, struct pblk_line *line,
> >               goto fail_complete;
> >       }
> >
> > -     rq_len = rq_ppas * geo->csecs;
> > -
> > -     bio = pblk_bio_map_addr(pblk, data, rq_ppas, rq_len,
> > -                                             PBLK_VMALLOC_META, GFP_KERNEL);
> > -     if (IS_ERR(bio)) {
> > -             ret = PTR_ERR(bio);
> > -             goto fail_complete;
> > -     }
> > -
> > -     bio->bi_iter.bi_sector = 0; /* internal bio */
> > -     bio_set_op_attrs(bio, REQ_OP_WRITE, 0);
> > -
> >       rqd = pblk_alloc_rqd(pblk, PBLK_WRITE_INT);
> >
> >       ret = pblk_alloc_rqd_meta(pblk, rqd);
> >       if (ret) {
> >               pblk_free_rqd(pblk, rqd, PBLK_WRITE_INT);
> > -             bio_put(bio);
> >               goto fail_complete;
> >       }
> >
> > -     rqd->bio = bio;
> > +     rqd->bio = NULL;
> >       rqd->opcode = NVM_OP_PWRITE;
> >       rqd->is_seq = 1;
> >       rqd->nr_ppas = rq_ppas;
> > @@ -275,13 +261,12 @@ static int pblk_recov_pad_line(struct pblk *pblk, struct pblk_line *line,
> >       kref_get(&pad_rq->ref);
> >       pblk_down_chunk(pblk, ppa_list[0]);
> >
> > -     ret = pblk_submit_io(pblk, rqd);
> > +     ret = pblk_submit_io(pblk, rqd, data);
> >       if (ret) {
> >               pblk_err(pblk, "I/O submission failed: %d\n", ret);
> >               pblk_up_chunk(pblk, ppa_list[0]);
> >               kref_put(&pad_rq->ref, pblk_recov_complete);
> >               pblk_free_rqd(pblk, rqd, PBLK_WRITE_INT);
> > -             bio_put(bio);
> >               goto fail_complete;
> >       }
> >
> > @@ -375,7 +360,6 @@ static int pblk_recov_scan_oob(struct pblk *pblk, struct pblk_line *line,
> >       struct ppa_addr *ppa_list;
> >       void *meta_list;
> >       struct nvm_rq *rqd;
> > -     struct bio *bio;
> >       void *data;
> >       dma_addr_t dma_ppa_list, dma_meta_list;
> >       __le64 *lba_list;
> > @@ -407,15 +391,7 @@ static int pblk_recov_scan_oob(struct pblk *pblk, struct pblk_line *line,
> >       rq_len = rq_ppas * geo->csecs;
> >
> > retry_rq:
> > -     bio = bio_map_kern(dev->q, data, rq_len, GFP_KERNEL);
> > -     if (IS_ERR(bio))
> > -             return PTR_ERR(bio);
> > -
> > -     bio->bi_iter.bi_sector = 0; /* internal bio */
> > -     bio_set_op_attrs(bio, REQ_OP_READ, 0);
> > -     bio_get(bio);
> > -
> > -     rqd->bio = bio;
> > +     rqd->bio = NULL;
> >       rqd->opcode = NVM_OP_PREAD;
> >       rqd->meta_list = meta_list;
> >       rqd->nr_ppas = rq_ppas;
> > @@ -445,10 +421,9 @@ static int pblk_recov_scan_oob(struct pblk *pblk, struct pblk_line *line,
> >                               addr_to_gen_ppa(pblk, paddr + j, line->id);
> >       }
> >
> > -     ret = pblk_submit_io_sync(pblk, rqd);
> > +     ret = pblk_submit_io_sync(pblk, rqd, data);
> >       if (ret) {
> >               pblk_err(pblk, "I/O submission failed: %d\n", ret);
> > -             bio_put(bio);
> >               return ret;
> >       }
> >
> > @@ -460,24 +435,20 @@ static int pblk_recov_scan_oob(struct pblk *pblk, struct pblk_line *line,
> >
> >               if (padded) {
> >                       pblk_log_read_err(pblk, rqd);
> > -                     bio_put(bio);
> >                       return -EINTR;
> >               }
> >
> >               pad_distance = pblk_pad_distance(pblk, line);
> >               ret = pblk_recov_pad_line(pblk, line, pad_distance);
> >               if (ret) {
> > -                     bio_put(bio);
> >                       return ret;
> >               }
> >
> >               padded = true;
> > -             bio_put(bio);
> >               goto retry_rq;
> >       }
> >
> >       pblk_get_packed_meta(pblk, rqd);
> > -     bio_put(bio);
> >
> >       for (i = 0; i < rqd->nr_ppas; i++) {
> >               struct pblk_sec_meta *meta = pblk_get_meta(pblk, meta_list, i);
> > diff --git a/drivers/lightnvm/pblk-write.c b/drivers/lightnvm/pblk-write.c
> > index 4e63f9b5954c..b9a2aeba95ab 100644
> > --- a/drivers/lightnvm/pblk-write.c
> > +++ b/drivers/lightnvm/pblk-write.c
> > @@ -373,7 +373,6 @@ int pblk_submit_meta_io(struct pblk *pblk, struct pblk_line *meta_line)
> >       struct pblk_emeta *emeta = meta_line->emeta;
> >       struct ppa_addr *ppa_list;
> >       struct pblk_g_ctx *m_ctx;
> > -     struct bio *bio;
> >       struct nvm_rq *rqd;
> >       void *data;
> >       u64 paddr;
> > @@ -391,20 +390,9 @@ int pblk_submit_meta_io(struct pblk *pblk, struct pblk_line *meta_line)
> >       rq_len = rq_ppas * geo->csecs;
> >       data = ((void *)emeta->buf) + emeta->mem;
> >
> > -     bio = pblk_bio_map_addr(pblk, data, rq_ppas, rq_len,
> > -                                     l_mg->emeta_alloc_type, GFP_KERNEL);
> > -     if (IS_ERR(bio)) {
> > -             pblk_err(pblk, "failed to map emeta io");
> > -             ret = PTR_ERR(bio);
> > -             goto fail_free_rqd;
> > -     }
> > -     bio->bi_iter.bi_sector = 0; /* internal bio */
> > -     bio_set_op_attrs(bio, REQ_OP_WRITE, 0);
> > -     rqd->bio = bio;
> > -
> >       ret = pblk_alloc_w_rq(pblk, rqd, rq_ppas, pblk_end_io_write_meta);
> >       if (ret)
> > -             goto fail_free_bio;
> > +             goto fail_free_rqd;
> >
> >       ppa_list = nvm_rq_to_ppa_list(rqd);
> >       for (i = 0; i < rqd->nr_ppas; ) {
> > @@ -423,7 +411,7 @@ int pblk_submit_meta_io(struct pblk *pblk, struct pblk_line *meta_line)
> >
> >       pblk_down_chunk(pblk, ppa_list[0]);
> >
> > -     ret = pblk_submit_io(pblk, rqd);
> > +     ret = pblk_submit_io(pblk, rqd, data);
> >       if (ret) {
> >               pblk_err(pblk, "emeta I/O submission failed: %d\n", ret);
> >               goto fail_rollback;
> > @@ -437,8 +425,6 @@ int pblk_submit_meta_io(struct pblk *pblk, struct pblk_line *meta_line)
> >       pblk_dealloc_page(pblk, meta_line, rq_ppas);
> >       list_add(&meta_line->list, &meta_line->list);
> >       spin_unlock(&l_mg->close_lock);
> > -fail_free_bio:
> > -     bio_put(bio);
> > fail_free_rqd:
> >       pblk_free_rqd(pblk, rqd, PBLK_WRITE_INT);
> >       return ret;
> > @@ -523,7 +509,7 @@ static int pblk_submit_io_set(struct pblk *pblk, struct nvm_rq *rqd)
> >       meta_line = pblk_should_submit_meta_io(pblk, rqd);
> >
> >       /* Submit data write for current data line */
> > -     err = pblk_submit_io(pblk, rqd);
> > +     err = pblk_submit_io(pblk, rqd, NULL);
> >       if (err) {
> >               pblk_err(pblk, "data I/O submission failed: %d\n", err);
> >               return NVM_IO_ERR;
> > diff --git a/drivers/lightnvm/pblk.h b/drivers/lightnvm/pblk.h
> > index a67855387f53..d515d3409a74 100644
> > --- a/drivers/lightnvm/pblk.h
> > +++ b/drivers/lightnvm/pblk.h
> > @@ -783,14 +783,10 @@ struct nvm_chk_meta *pblk_chunk_get_off(struct pblk *pblk,
> >                                             struct ppa_addr ppa);
> > void pblk_log_write_err(struct pblk *pblk, struct nvm_rq *rqd);
> > void pblk_log_read_err(struct pblk *pblk, struct nvm_rq *rqd);
> > -int pblk_submit_io(struct pblk *pblk, struct nvm_rq *rqd);
> > -int pblk_submit_io_sync(struct pblk *pblk, struct nvm_rq *rqd);
> > -int pblk_submit_io_sync_sem(struct pblk *pblk, struct nvm_rq *rqd);
> > +int pblk_submit_io(struct pblk *pblk, struct nvm_rq *rqd, void *buf);
> > +int pblk_submit_io_sync(struct pblk *pblk, struct nvm_rq *rqd, void *buf);
> > int pblk_submit_meta_io(struct pblk *pblk, struct pblk_line *meta_line);
> > void pblk_check_chunk_state_update(struct pblk *pblk, struct nvm_rq *rqd);
> > -struct bio *pblk_bio_map_addr(struct pblk *pblk, void *data,
> > -                           unsigned int nr_secs, unsigned int len,
> > -                           int alloc_type, gfp_t gfp_mask);
> > struct pblk_line *pblk_line_get(struct pblk *pblk);
> > struct pblk_line *pblk_line_get_first_data(struct pblk *pblk);
> > struct pblk_line *pblk_line_replace_data(struct pblk *pblk);
> > diff --git a/drivers/nvme/host/lightnvm.c b/drivers/nvme/host/lightnvm.c
> > index d6f121452d5d..ec46693f6b64 100644
> > --- a/drivers/nvme/host/lightnvm.c
> > +++ b/drivers/nvme/host/lightnvm.c
> > @@ -667,11 +667,14 @@ static struct request *nvme_nvm_alloc_request(struct request_queue *q,
> >       return rq;
> > }
> >
> > -static int nvme_nvm_submit_io(struct nvm_dev *dev, struct nvm_rq *rqd)
> > +static int nvme_nvm_submit_io(struct nvm_dev *dev, struct nvm_rq *rqd,
> > +                           void *buf)
> > {
> > +     struct nvm_geo *geo = &dev->geo;
> >       struct request_queue *q = dev->q;
> >       struct nvme_nvm_command *cmd;
> >       struct request *rq;
> > +     int ret;
> >
> >       cmd = kzalloc(sizeof(struct nvme_nvm_command), GFP_KERNEL);
> >       if (!cmd)
> > @@ -679,8 +682,15 @@ static int nvme_nvm_submit_io(struct nvm_dev *dev, struct nvm_rq *rqd)
> >
> >       rq = nvme_nvm_alloc_request(q, rqd, cmd);
> >       if (IS_ERR(rq)) {
> > -             kfree(cmd);
> > -             return PTR_ERR(rq);
> > +             ret = PTR_ERR(rq);
> > +             goto err_free_cmd;
> > +     }
> > +
> > +     if (buf) {
> > +             ret = blk_rq_map_kern(q, rq, buf, geo->csecs * rqd->nr_ppas,
> > +                             GFP_KERNEL);
> > +             if (ret)
> > +                     goto err_free_cmd;
> >       }
> >
> >       rq->end_io_data = rqd;
> > @@ -688,6 +698,10 @@ static int nvme_nvm_submit_io(struct nvm_dev *dev, struct nvm_rq *rqd)
> >       blk_execute_rq_nowait(q, NULL, rq, 0, nvme_nvm_end_io);
> >
> >       return 0;
> > +
> > +err_free_cmd:
> > +     kfree(cmd);
> > +     return ret;
> > }
> >
> > static void *nvme_nvm_create_dma_pool(struct nvm_dev *nvmdev, char *name,
> > diff --git a/include/linux/lightnvm.h b/include/linux/lightnvm.h
> > index 8891647b24b1..ee8ec2e68055 100644
> > --- a/include/linux/lightnvm.h
> > +++ b/include/linux/lightnvm.h
> > @@ -88,7 +88,7 @@ typedef int (nvm_op_bb_tbl_fn)(struct nvm_dev *, struct ppa_addr, u8 *);
> > typedef int (nvm_op_set_bb_fn)(struct nvm_dev *, struct ppa_addr *, int, int);
> > typedef int (nvm_get_chk_meta_fn)(struct nvm_dev *, sector_t, int,
> >                                                       struct nvm_chk_meta *);
> > -typedef int (nvm_submit_io_fn)(struct nvm_dev *, struct nvm_rq *);
> > +typedef int (nvm_submit_io_fn)(struct nvm_dev *, struct nvm_rq *, void *);
> > typedef void *(nvm_create_dma_pool_fn)(struct nvm_dev *, char *, int);
> > typedef void (nvm_destroy_dma_pool_fn)(void *);
> > typedef void *(nvm_dev_dma_alloc_fn)(struct nvm_dev *, void *, gfp_t,
> > @@ -680,8 +680,8 @@ extern int nvm_get_chunk_meta(struct nvm_tgt_dev *, struct ppa_addr,
> >                             int, struct nvm_chk_meta *);
> > extern int nvm_set_chunk_meta(struct nvm_tgt_dev *, struct ppa_addr *,
> >                             int, int);
> > -extern int nvm_submit_io(struct nvm_tgt_dev *, struct nvm_rq *);
> > -extern int nvm_submit_io_sync(struct nvm_tgt_dev *, struct nvm_rq *);
> > +extern int nvm_submit_io(struct nvm_tgt_dev *, struct nvm_rq *, void *);
> > +extern int nvm_submit_io_sync(struct nvm_tgt_dev *, struct nvm_rq *, void *);
> > extern void nvm_end_io(struct nvm_rq *);
> >
> > #else /* CONFIG_NVM */
> > --
> > 2.7.4
>
> It’s very good to get rid of pblk_bio_map_addr(). Need to look at this
> closer because I remember a number of corner cases due to the metadata
> allocation. Have you tested the vmalloc allocation path? Note that you
> need big lines for this to happen as it will try to do kmalloc if
> possible

Yeah, i checked that vmalloc worked for the emeta buffers(by switching
over manually), so we should be good.

>
> Javier

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ