qemu-stable
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [PATCH for 7.2? V2] vhost: fix vq dirty bitmap syncing when vIOMMU i


From: Eric Auger
Subject: Re: [PATCH for 7.2? V2] vhost: fix vq dirty bitmap syncing when vIOMMU is enabled
Date: Tue, 29 Nov 2022 17:08:49 +0100
User-agent: Mozilla/5.0 (X11; Linux x86_64; rv:102.0) Gecko/20100101 Thunderbird/102.3.1

Hi Michael,

On 11/29/22 16:44, Michael S. Tsirkin wrote:
> On Tue, Nov 29, 2022 at 10:52:29AM +0100, Eric Auger wrote:
>> Hi Jason,
>>
>> On 11/29/22 05:02, Jason Wang wrote:
>>> When vIOMMU is enabled, the vq->used_phys is actually the IOVA not
>>> GPA. So we need to translate it to GPA before the syncing otherwise we
>>> may hit the following crash since IOVA could be out of the scope of
>>> the GPA log size. This could be noted when using virtio-IOMMU with
>>> vhost using 1G memory.
>>>
>>> Fixes: c471ad0e9bd46 ("vhost_net: device IOTLB support")
>>> Cc: qemu-stable@nongnu.org
>>> Tested-by: Lei Yang <leiyang@redhat.com>
>>> Reported-by: Yalan Zhang <yalzhang@redhat.com>
>>> Signed-off-by: Jason Wang <jasowang@redhat.com>
>>> ---
>>> Changes since V1:
>>> - Fix the address calculation when used ring is not page aligned
>>> - Fix the length for each round of dirty bitmap syncing
>>> - Use LOG_GUEST_ERROR to log wrong used adddress
>>> - Various other tweaks
>>> ---
>>>  hw/virtio/vhost.c | 76 ++++++++++++++++++++++++++++++++++-------------
>>>  1 file changed, 56 insertions(+), 20 deletions(-)
>>>
>>> diff --git a/hw/virtio/vhost.c b/hw/virtio/vhost.c
>>> index d1c4c20b8c..0cd5f25fcb 100644
>>> --- a/hw/virtio/vhost.c
>>> +++ b/hw/virtio/vhost.c
>>> @@ -20,6 +20,7 @@
>>>  #include "qemu/range.h"
>>>  #include "qemu/error-report.h"
>>>  #include "qemu/memfd.h"
>>> +#include "qemu/log.h"
>>>  #include "standard-headers/linux/vhost_types.h"
>>>  #include "hw/virtio/virtio-bus.h"
>>>  #include "hw/virtio/virtio-access.h"
>>> @@ -106,6 +107,24 @@ static void vhost_dev_sync_region(struct vhost_dev 
>>> *dev,
>>>      }
>>>  }
>>>  
>>> +static bool vhost_dev_has_iommu(struct vhost_dev *dev)
>>> +{
>>> +    VirtIODevice *vdev = dev->vdev;
>>> +
>>> +    /*
>>> +     * For vhost, VIRTIO_F_IOMMU_PLATFORM means the backend support
>>> +     * incremental memory mapping API via IOTLB API. For platform that
>>> +     * does not have IOMMU, there's no need to enable this feature
>>> +     * which may cause unnecessary IOTLB miss/update transactions.
>>> +     */
>>> +    if (vdev) {
>>> +        return virtio_bus_device_iommu_enabled(vdev) &&
>>> +            virtio_host_has_feature(vdev, VIRTIO_F_IOMMU_PLATFORM);
>>> +    } else {
>>> +        return false;
>>> +    }
>>> +}
>>> +
>>>  static int vhost_sync_dirty_bitmap(struct vhost_dev *dev,
>>>                                     MemoryRegionSection *section,
>>>                                     hwaddr first,
>>> @@ -137,8 +156,43 @@ static int vhost_sync_dirty_bitmap(struct vhost_dev 
>>> *dev,
>>>              continue;
>>>          }
>>>  
>>> -        vhost_dev_sync_region(dev, section, start_addr, end_addr, 
>>> vq->used_phys,
>>> -                              range_get_last(vq->used_phys, 
>>> vq->used_size));
>>> +        if (vhost_dev_has_iommu(dev)) {
>>> +            IOMMUTLBEntry iotlb;
>>> +            hwaddr used_phys = vq->used_phys, used_size = vq->used_size;
>>> +            hwaddr phys, s;
>>> +
>>> +            while (used_size) {
>>> +                rcu_read_lock();
>>> +                iotlb = address_space_get_iotlb_entry(dev->vdev->dma_as,
>>> +                                                      used_phys,
>>> +                                                      true, 
>>> MEMTXATTRS_UNSPECIFIED);
>>> +                rcu_read_unlock();
>>> +
>>> +                if (!iotlb.target_as) {
>>> +                    qemu_log_mask(LOG_GUEST_ERROR, "translation "
>>> +                                  "failure for used_phys %"PRIx64"\n", 
>>> used_phys);
>> looks weird to see translation of "used_phys" whereas it is an iova. At
>> least I would reword the msg
>>> +                    return -EINVAL;
>>> +                }
>>> +
>>> +                phys = iotlb.translated_addr + (used_phys & 
>>> iotlb.addr_mask);
>> you may use a local variable storing this offset =
>>
>> used_phys & iotlb.addr_mask
>>
>>> +
>>> +                /* Distance from start of used ring until last byte of
>>> +                   IOMMU page */
>> you can avoid checkpatch warnings here
>>> +                s = iotlb.addr_mask - (used_phys & iotlb.addr_mask);
>>> +                /* Size of used ring, or of the part of it until end
>>> +                   of IOMMU page */
>> and here
>>
>> I would suggest to rewrite this into
>> s =iotlb.addr_mask - (used_phys & iotlb.addr_mask) + 1
>> s = MIN(s, used_size);
> This does not work - if iotlb.addr_mask - (used_phys & iotlb.addr_mask)
> is all-ones then + 1 gives you 0 and MIN gives you 0.
> Theoretical but worth being safe here IMHO.
Ah OK, I should have read your previous discussion more thoroughly ...
Maybe just add a short comment then to justify the gym below and avoid
tempting sbdy else to rewrite it in a more common but wrong way.

Thanks

Eric
>
>
>>> +                s = MIN(s, used_size - 1) + 1;
>>> +
>>> +                vhost_dev_sync_region(dev, section, start_addr, end_addr, 
>>> phys,
>>> +                                      range_get_last(phys, s));
>>> +                used_size -= s;
>>> +                used_phys += s;
>>> +            }
>>> +        } else {
>>> +            vhost_dev_sync_region(dev, section, start_addr,
>>> +                                  end_addr, vq->used_phys,
>>> +                                  range_get_last(vq->used_phys, 
>>> vq->used_size));
>>> +        }
>>>      }
>>>      return 0;
>>>  }
>>> @@ -306,24 +360,6 @@ static inline void vhost_dev_log_resize(struct 
>>> vhost_dev *dev, uint64_t size)
>>>      dev->log_size = size;
>>>  }
>>>  
>>> -static bool vhost_dev_has_iommu(struct vhost_dev *dev)
>>> -{
>>> -    VirtIODevice *vdev = dev->vdev;
>>> -
>>> -    /*
>>> -     * For vhost, VIRTIO_F_IOMMU_PLATFORM means the backend support
>>> -     * incremental memory mapping API via IOTLB API. For platform that
>>> -     * does not have IOMMU, there's no need to enable this feature
>>> -     * which may cause unnecessary IOTLB miss/update transactions.
>>> -     */
>>> -    if (vdev) {
>>> -        return virtio_bus_device_iommu_enabled(vdev) &&
>>> -            virtio_host_has_feature(vdev, VIRTIO_F_IOMMU_PLATFORM);
>>> -    } else {
>>> -        return false;
>>> -    }
>>> -}
>>> -
>>>  static void *vhost_memory_map(struct vhost_dev *dev, hwaddr addr,
>>>                                hwaddr *plen, bool is_write)
>>>  {
>> Besides,
>>
>> Tested-by: Eric Auger <eric.auger@redhat.com>
>>
>> Eric




reply via email to

[Prev in Thread] Current Thread [Next in Thread]