[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <473d10c5-b2cb-e976-a923-b5add22bcde6@nvidia.com>
Date: Wed, 16 Nov 2016 09:13:37 +0530
From: Kirti Wankhede <kwankhede@...dia.com>
To: Alex Williamson <alex.williamson@...hat.com>
CC: <pbonzini@...hat.com>, <kraxel@...hat.com>, <cjia@...dia.com>,
<qemu-devel@...gnu.org>, <kvm@...r.kernel.org>,
<kevin.tian@...el.com>, <jike.song@...el.com>,
<bjsdjshi@...ux.vnet.ibm.com>, <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH v13 11/22] vfio iommu: Add blocking notifier to notify
DMA_UNMAP
On 11/16/2016 8:55 AM, Alex Williamson wrote:
> On Tue, 15 Nov 2016 20:16:12 -0700
> Alex Williamson <alex.williamson@...hat.com> wrote:
>
>> On Wed, 16 Nov 2016 08:16:15 +0530
>> Kirti Wankhede <kwankhede@...dia.com> wrote:
>>
>>> On 11/16/2016 3:49 AM, Alex Williamson wrote:
>>>> On Tue, 15 Nov 2016 20:59:54 +0530
>>>> Kirti Wankhede <kwankhede@...dia.com> wrote:
>>>>
>>> ...
>>>
>>>>> @@ -854,7 +857,28 @@ static int vfio_dma_do_unmap(struct vfio_iommu *iommu,
>>>>> */
>>>>> if (dma->task->mm != current->mm)
>>>>> break;
>>>>> +
>>>>> unmapped += dma->size;
>>>>> +
>>>>> + if (iommu->external_domain && !RB_EMPTY_ROOT(&dma->pfn_list)) {
>>>>> + struct vfio_iommu_type1_dma_unmap nb_unmap;
>>>>> +
>>>>> + nb_unmap.iova = dma->iova;
>>>>> + nb_unmap.size = dma->size;
>>>>> +
>>>>> + /*
>>>>> + * Notifier callback would call vfio_unpin_pages() which
>>>>> + * would acquire iommu->lock. Release lock here and
>>>>> + * reacquire it again.
>>>>> + */
>>>>> + mutex_unlock(&iommu->lock);
>>>>> + blocking_notifier_call_chain(&iommu->notifier,
>>>>> + VFIO_IOMMU_NOTIFY_DMA_UNMAP,
>>>>> + &nb_unmap);
>>>>> + mutex_lock(&iommu->lock);
>>>>> + if (WARN_ON(!RB_EMPTY_ROOT(&dma->pfn_list)))
>>>>> + break;
>>>>> + }
>>>>
>>>>
>>>> Why exactly do we need to notify per vfio_dma rather than per unmap
>>>> request? If we do the latter we can send the notify first, limiting us
>>>> to races where a page is pinned between the notify and the locking,
>>>> whereas here, even our dma pointer is suspect once we re-acquire the
>>>> lock, we don't technically know if another unmap could have removed
>>>> that already. Perhaps something like this (untested):
>>>>
>>>
>>> There are checks to validate unmap request, like v2 check and who is
>>> calling unmap and is it allowed for that task to unmap. Before these
>>> checks its not sure that unmap region range which asked for would be
>>> unmapped all. Notify call should be at the place where its sure that the
>>> range provided to notify call is definitely going to be removed. My
>>> change do that.
>>
>> Ok, but that does solve the problem. What about this (untested):
>
> s/does/does not/
>
> BTW, I like how the retries here fill the gap in my previous proposal
> where we could still race re-pinning. We've given it an honest shot or
> someone is not participating if we've retried 10 times. I don't
> understand why the test for iommu->external_domain was there, clearly
> if the list is not empty, we need to notify. Thanks,
>
Ok. Retry is good to give a chance to unpin all. But is it really
required to use BUG_ON() that would panic the host. I think WARN_ON
should be fine and then when container is closed or when the last group
is removed from the container, vfio_iommu_type1_release() is called and
we have a chance to unpin it all.
Thanks,
Kirti
> Alex
>
>> diff --git a/drivers/vfio/vfio_iommu_type1.c b/drivers/vfio/vfio_iommu_type1.c
>> index ee9a680..50cafdf 100644
>> --- a/drivers/vfio/vfio_iommu_type1.c
>> +++ b/drivers/vfio/vfio_iommu_type1.c
>> @@ -782,9 +782,9 @@ static int vfio_dma_do_unmap(struct vfio_iommu *iommu,
>> struct vfio_iommu_type1_dma_unmap *unmap)
>> {
>> uint64_t mask;
>> - struct vfio_dma *dma;
>> + struct vfio_dma *dma, *dma_last = NULL;
>> size_t unmapped = 0;
>> - int ret = 0;
>> + int ret = 0, retries;
>>
>> mask = ((uint64_t)1 << __ffs(vfio_pgsize_bitmap(iommu))) - 1;
>>
>> @@ -794,7 +794,7 @@ static int vfio_dma_do_unmap(struct vfio_iommu *iommu,
>> return -EINVAL;
>>
>> WARN_ON(mask & PAGE_MASK);
>> -
>> +again:
>> mutex_lock(&iommu->lock);
>>
>> /*
>> @@ -851,11 +851,16 @@ static int vfio_dma_do_unmap(struct vfio_iommu *iommu,
>> if (dma->task->mm != current->mm)
>> break;
>>
>> - unmapped += dma->size;
>> -
>> - if (iommu->external_domain && !RB_EMPTY_ROOT(&dma->pfn_list)) {
>> + if (!RB_EMPTY_ROOT(&dma->pfn_list)) {
>> struct vfio_iommu_type1_dma_unmap nb_unmap;
>>
>> + if (dma_last == dma) {
>> + BUG_ON(++retries > 10);
>> + } else {
>> + dma_last = dma;
>> + retries = 0;
>> + }
>> +
>> nb_unmap.iova = dma->iova;
>> nb_unmap.size = dma->size;
>>
>> @@ -868,11 +873,11 @@ static int vfio_dma_do_unmap(struct vfio_iommu *iommu,
>> blocking_notifier_call_chain(&iommu->notifier,
>> VFIO_IOMMU_NOTIFY_DMA_UNMAP,
>> &nb_unmap);
>> - mutex_lock(&iommu->lock);
>> - if (WARN_ON(!RB_EMPTY_ROOT(&dma->pfn_list)))
>> - break;
>> + goto again:
>> }
>> + unmapped += dma->size;
>> vfio_remove_dma(iommu, dma);
>> +
>> }
>>
>> unlock:
>
Powered by blists - more mailing lists