lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <59E71AE2.6080202@huawei.com>
Date:   Wed, 18 Oct 2017 17:12:02 +0800
From:   "Wei Hu (Xavier)" <xavier.huwei@...wei.com>
To:     Leon Romanovsky <leon@...nel.org>
CC:     <lijun_nudt@....com>, <linux-rdma@...r.kernel.org>,
        <shaobohsu@....com>, <linuxarm@...wei.com>,
        <linux-kernel@...r.kernel.org>, <dledford@...hat.com>,
        <zhangxiping3@...wei.com>, <shaoboxu@....com>,
        <shaobo.xu@...el.com>, "Doug Ledford" <dledford@...hat.com>,
        "Liuyixian (Eason)" <liuyixian@...wei.com>,
        "Chenxin (Charles)" <charles.chenxin@...wei.com>
Subject: Re: [PATCH for-next 2/4] RDMA/hns: Add IOMMU enable support in hip08



On 2017/10/18 16:42, Wei Hu (Xavier) wrote:
>
>
> On 2017/10/1 0:10, Leon Romanovsky wrote:
>> On Sat, Sep 30, 2017 at 05:28:59PM +0800, Wei Hu (Xavier) wrote:
>>> If the IOMMU is enabled, the length of sg obtained from
>>> __iommu_map_sg_attrs is not 4kB. When the IOVA is set with the sg
>>> dma address, the IOVA will not be page continuous. and the VA
>>> returned from dma_alloc_coherent is a vmalloc address. However,
>>> the VA obtained by the page_address is a discontinuous VA. Under
>>> these circumstances, the IOVA should be calculated based on the
>>> sg length, and record the VA returned from dma_alloc_coherent
>>> in the struct of hem.
>>>
>>> Signed-off-by: Wei Hu (Xavier) <xavier.huwei@...wei.com>
>>> Signed-off-by: Shaobo Xu <xushaobo2@...wei.com>
>>> Signed-off-by: Lijun Ou <oulijun@...wei.com>
>>> ---
>> Doug,
>>
>> I didn't invest time in reviewing it, but having "is_vmalloc_addr" in
>> driver code to deal with dma_alloc_coherent is most probably wrong.
>>
>> Thanks
>>
> Hi, Doug
>     When running in ARM64 platform, there probably be calltrace 
> currently.
>     Now our colleague will report it to iommu maillist and try to 
> solve it.
>     I also think RoCE driver shouldn't sense the difference.
>     I will pull it out of this series and send v2.
>     Thanks.
>
Hi, Doug & Leon
     I have sent patch v2.
     Thanks

     Regards
Wei Hu
> Regards
> Wei Hu
>
>>> drivers/infiniband/hw/hns/hns_roce_alloc.c |  5 ++++-
>>>   drivers/infiniband/hw/hns/hns_roce_hem.c   | 30 
>>> +++++++++++++++++++++++++++---
>>>   drivers/infiniband/hw/hns/hns_roce_hem.h   |  6 ++++++
>>>   drivers/infiniband/hw/hns/hns_roce_hw_v2.c | 22 
>>> +++++++++++++++-------
>>>   4 files changed, 52 insertions(+), 11 deletions(-)
>>>
>>> diff --git a/drivers/infiniband/hw/hns/hns_roce_alloc.c 
>>> b/drivers/infiniband/hw/hns/hns_roce_alloc.c
>>> index 3e4c525..a69cd4b 100644
>>> --- a/drivers/infiniband/hw/hns/hns_roce_alloc.c
>>> +++ b/drivers/infiniband/hw/hns/hns_roce_alloc.c
>>> @@ -243,7 +243,10 @@ int hns_roce_buf_alloc(struct hns_roce_dev 
>>> *hr_dev, u32 size, u32 max_direct,
>>>                   goto err_free;
>>>
>>>               for (i = 0; i < buf->nbufs; ++i)
>>> -                pages[i] = virt_to_page(buf->page_list[i].buf);
>>> +                pages[i] =
>>> +                    is_vmalloc_addr(buf->page_list[i].buf) ?
>>> +                    vmalloc_to_page(buf->page_list[i].buf) :
>>> +                    virt_to_page(buf->page_list[i].buf);
>>>
>>>               buf->direct.buf = vmap(pages, buf->nbufs, VM_MAP,
>>>                              PAGE_KERNEL);
>>> diff --git a/drivers/infiniband/hw/hns/hns_roce_hem.c 
>>> b/drivers/infiniband/hw/hns/hns_roce_hem.c
>>> index 8388ae2..4a3d1d4 100644
>>> --- a/drivers/infiniband/hw/hns/hns_roce_hem.c
>>> +++ b/drivers/infiniband/hw/hns/hns_roce_hem.c
>>> @@ -200,6 +200,7 @@ static struct hns_roce_hem 
>>> *hns_roce_alloc_hem(struct hns_roce_dev *hr_dev,
>>>                              gfp_t gfp_mask)
>>>   {
>>>       struct hns_roce_hem_chunk *chunk = NULL;
>>> +    struct hns_roce_vmalloc *vmalloc;
>>>       struct hns_roce_hem *hem;
>>>       struct scatterlist *mem;
>>>       int order;
>>> @@ -227,6 +228,7 @@ static struct hns_roce_hem 
>>> *hns_roce_alloc_hem(struct hns_roce_dev *hr_dev,
>>>               sg_init_table(chunk->mem, HNS_ROCE_HEM_CHUNK_LEN);
>>>               chunk->npages = 0;
>>>               chunk->nsg = 0;
>>> +            memset(chunk->vmalloc, 0, sizeof(chunk->vmalloc));
>>>               list_add_tail(&chunk->list, &hem->chunk_list);
>>>           }
>>>
>>> @@ -243,7 +245,15 @@ static struct hns_roce_hem 
>>> *hns_roce_alloc_hem(struct hns_roce_dev *hr_dev,
>>>           if (!buf)
>>>               goto fail;
>>>
>>> -        sg_set_buf(mem, buf, PAGE_SIZE << order);
>>> +        if (is_vmalloc_addr(buf)) {
>>> +            vmalloc = &chunk->vmalloc[chunk->npages];
>>> +            vmalloc->is_vmalloc_addr = true;
>>> +            vmalloc->vmalloc_addr = buf;
>>> +            sg_set_page(mem, vmalloc_to_page(buf),
>>> +                    PAGE_SIZE << order, offset_in_page(buf));
>>> +        } else {
>>> +            sg_set_buf(mem, buf, PAGE_SIZE << order);
>>> +        }
>>>           WARN_ON(mem->offset);
>>>           sg_dma_len(mem) = PAGE_SIZE << order;
>>>
>>> @@ -262,17 +272,25 @@ static struct hns_roce_hem 
>>> *hns_roce_alloc_hem(struct hns_roce_dev *hr_dev,
>>>   void hns_roce_free_hem(struct hns_roce_dev *hr_dev, struct 
>>> hns_roce_hem *hem)
>>>   {
>>>       struct hns_roce_hem_chunk *chunk, *tmp;
>>> +    void *cpu_addr;
>>>       int i;
>>>
>>>       if (!hem)
>>>           return;
>>>
>>>       list_for_each_entry_safe(chunk, tmp, &hem->chunk_list, list) {
>>> -        for (i = 0; i < chunk->npages; ++i)
>>> +        for (i = 0; i < chunk->npages; ++i) {
>>> +            if (chunk->vmalloc[i].is_vmalloc_addr)
>>> +                cpu_addr = chunk->vmalloc[i].vmalloc_addr;
>>> +            else
>>> +                cpu_addr =
>>> + lowmem_page_address(sg_page(&chunk->mem[i]));
>>> +
>>>               dma_free_coherent(hr_dev->dev,
>>>                      chunk->mem[i].length,
>>> - lowmem_page_address(sg_page(&chunk->mem[i])),
>>> +                   cpu_addr,
>>>                      sg_dma_address(&chunk->mem[i]));
>>> +        }
>>>           kfree(chunk);
>>>       }
>>>
>>> @@ -774,6 +792,12 @@ void *hns_roce_table_find(struct hns_roce_dev 
>>> *hr_dev,
>>>
>>>               if (chunk->mem[i].length > (u32)offset) {
>>>                   page = sg_page(&chunk->mem[i]);
>>> +                if (chunk->vmalloc[i].is_vmalloc_addr) {
>>> +                    mutex_unlock(&table->mutex);
>>> +                    return page ?
>>> +                        chunk->vmalloc[i].vmalloc_addr
>>> +                        + offset : NULL;
>>> +                }
>>>                   goto out;
>>>               }
>>>               offset -= chunk->mem[i].length;
>>> diff --git a/drivers/infiniband/hw/hns/hns_roce_hem.h 
>>> b/drivers/infiniband/hw/hns/hns_roce_hem.h
>>> index af28bbf..62d712a 100644
>>> --- a/drivers/infiniband/hw/hns/hns_roce_hem.h
>>> +++ b/drivers/infiniband/hw/hns/hns_roce_hem.h
>>> @@ -72,11 +72,17 @@ enum {
>>>        HNS_ROCE_HEM_PAGE_SIZE  = 1 << HNS_ROCE_HEM_PAGE_SHIFT,
>>>   };
>>>
>>> +struct hns_roce_vmalloc {
>>> +    bool    is_vmalloc_addr;
>>> +    void    *vmalloc_addr;
>>> +};
>>> +
>>>   struct hns_roce_hem_chunk {
>>>       struct list_head     list;
>>>       int             npages;
>>>       int             nsg;
>>>       struct scatterlist     mem[HNS_ROCE_HEM_CHUNK_LEN];
>>> +    struct hns_roce_vmalloc vmalloc[HNS_ROCE_HEM_CHUNK_LEN];
>>>   };
>>>
>>>   struct hns_roce_hem {
>>> diff --git a/drivers/infiniband/hw/hns/hns_roce_hw_v2.c 
>>> b/drivers/infiniband/hw/hns/hns_roce_hw_v2.c
>>> index b99d70a..9e19bf1 100644
>>> --- a/drivers/infiniband/hw/hns/hns_roce_hw_v2.c
>>> +++ b/drivers/infiniband/hw/hns/hns_roce_hw_v2.c
>>> @@ -1093,9 +1093,11 @@ static int hns_roce_v2_write_mtpt(void 
>>> *mb_buf, struct hns_roce_mr *mr,
>>>   {
>>>       struct hns_roce_v2_mpt_entry *mpt_entry;
>>>       struct scatterlist *sg;
>>> +    u64 page_addr = 0;
>>>       u64 *pages;
>>> +    int i = 0, j = 0;
>>> +    int len = 0;
>>>       int entry;
>>> -    int i;
>>>
>>>       mpt_entry = mb_buf;
>>>       memset(mpt_entry, 0, sizeof(*mpt_entry));
>>> @@ -1153,14 +1155,20 @@ static int hns_roce_v2_write_mtpt(void 
>>> *mb_buf, struct hns_roce_mr *mr,
>>>
>>>       i = 0;
>>>       for_each_sg(mr->umem->sg_head.sgl, sg, mr->umem->nmap, entry) {
>>> -        pages[i] = ((u64)sg_dma_address(sg)) >> 6;
>>> -
>>> -        /* Record the first 2 entry directly to MTPT table */
>>> -        if (i >= HNS_ROCE_V2_MAX_INNER_MTPT_NUM - 1)
>>> -            break;
>>> -        i++;
>>> +        len = sg_dma_len(sg) >> PAGE_SHIFT;
>>> +        for (j = 0; j < len; ++j) {
>>> +            page_addr = sg_dma_address(sg) +
>>> +                    (j << mr->umem->page_shift);
>>> +            pages[i] = page_addr >> 6;
>>> +
>>> +            /* Record the first 2 entry directly to MTPT table */
>>> +            if (i >= HNS_ROCE_V2_MAX_INNER_MTPT_NUM - 1)
>>> +                goto found;
>>> +            i++;
>>> +        }
>>>       }
>>>
>>> +found:
>>>       mpt_entry->pa0_l = cpu_to_le32(lower_32_bits(pages[0]));
>>>       roce_set_field(mpt_entry->byte_56_pa0_h, V2_MPT_BYTE_56_PA0_H_M,
>>>                  V2_MPT_BYTE_56_PA0_H_S,
>>> -- 
>>> 1.9.1
>>>
>
>
> _______________________________________________
> linuxarm mailing list
> linuxarm@...wei.com
> http://rnd-openeuler.huawei.com/mailman/listinfo/linuxarm
>
> .
>


Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ