qemu-devel
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [PATCH v6 1/2] vhost-user: remove VirtQ notifier restore


From: Michael S. Tsirkin
Subject: Re: [PATCH v6 1/2] vhost-user: remove VirtQ notifier restore
Date: Fri, 4 Feb 2022 07:25:57 -0500

I dropped this for now as I'm a bit lost with this patchset.
Cc Raphael maybe he'll understand it better.

On Wed, Jan 12, 2022 at 03:05:15PM +0000, Xueming(Steven) Li wrote:
> On Wed, 2021-11-03 at 16:30 -0400, Michael S. Tsirkin wrote:
> > On Wed, Nov 03, 2021 at 02:48:41PM +0000, Xueming(Steven) Li wrote:
> > > On Tue, 2021-11-02 at 02:49 -0400, Michael S. Tsirkin wrote:
> > > > On Tue, Nov 02, 2021 at 06:08:39AM +0000, Xueming(Steven) Li wrote:
> > > > > On Mon, 2021-11-01 at 17:06 -0400, Michael S. Tsirkin wrote:
> > > > > > On Mon, Nov 01, 2021 at 04:38:12PM +0800, Xueming Li wrote:
> > > > > > > When vhost-user vdpa client suspend, backend may close all 
> > > > > > > resources,
> > > > > > > VQ notifier mmap address become invalid, restore MR which contains
> > > > > > > the invalid address is wrong. vdpa client will set VQ notifier 
> > > > > > > after
> > > > > > > reconnect.
> > > > > > > 
> > > > > > > This patch removes VQ notifier restore and related flags to avoid 
> > > > > > > reusing
> > > > > > > invalid address.
> > > > > > > 
> > > > > > > Fixes: 44866521bd6e ("vhost-user: support registering external 
> > > > > > > host notifiers")
> > > > > > > Cc: qemu-stable@nongnu.org
> > > > > > > Cc: Yuwei Zhang <zhangyuwei.9149@bytedance.com>
> > > > > > > Signed-off-by: Xueming Li <xuemingl@nvidia.com>
> > > > > > > ---
> > > > > > >  hw/virtio/vhost-user.c         | 19 +------------------
> > > > > > >  include/hw/virtio/vhost-user.h |  1 -
> > > > > > >  2 files changed, 1 insertion(+), 19 deletions(-)
> > > > > > > 
> > > > > > > diff --git a/hw/virtio/vhost-user.c b/hw/virtio/vhost-user.c
> > > > > > > index bf6e50223c..c671719e9b 100644
> > > > > > > --- a/hw/virtio/vhost-user.c
> > > > > > > +++ b/hw/virtio/vhost-user.c
> > > > > > > @@ -1143,19 +1143,6 @@ static int vhost_user_set_vring_num(struct 
> > > > > > > vhost_dev *dev,
> > > > > > >      return vhost_set_vring(dev, VHOST_USER_SET_VRING_NUM, ring);
> > > > > > >  }
> > > > > > >  
> > > > > > > -static void vhost_user_host_notifier_restore(struct vhost_dev 
> > > > > > > *dev,
> > > > > > > -                                             int queue_idx)
> > > > > > > -{
> > > > > > > -    struct vhost_user *u = dev->opaque;
> > > > > > > -    VhostUserHostNotifier *n = &u->user->notifier[queue_idx];
> > > > > > > -    VirtIODevice *vdev = dev->vdev;
> > > > > > > -
> > > > > > > -    if (n->addr && !n->set) {
> > > > > > > -        virtio_queue_set_host_notifier_mr(vdev, queue_idx, 
> > > > > > > &n->mr, true);
> > > > > > > -        n->set = true;
> > > > > > > -    }
> > > > > > > -}
> > > > > > > -
> > > > > > >  static void vhost_user_host_notifier_remove(struct vhost_dev 
> > > > > > > *dev,
> > > > > > >                                              int queue_idx)
> > > > > > >  {
> > > > > > > @@ -1163,17 +1150,14 @@ static void 
> > > > > > > vhost_user_host_notifier_remove(struct vhost_dev *dev,
> > > > > > >      VhostUserHostNotifier *n = &u->user->notifier[queue_idx];
> > > > > > >      VirtIODevice *vdev = dev->vdev;
> > > > > > >  
> > > > > > > -    if (n->addr && n->set) {
> > > > > > > +    if (n->addr) {
> > > > > > >          virtio_queue_set_host_notifier_mr(vdev, queue_idx, 
> > > > > > > &n->mr, false);
> > > > > > > -        n->set = false;
> > > > > > >      }
> > > > > > >  }
> > > > > > > 
> > > > > > 
> > > > > > So on vq stop we still remove the notifier...
> > > > > >   
> > > > > > >  static int vhost_user_set_vring_base(struct vhost_dev *dev,
> > > > > > >                                       struct vhost_vring_state 
> > > > > > > *ring)
> > > > > > >  {
> > > > > > > -    vhost_user_host_notifier_restore(dev, ring->index);
> > > > > > > -
> > > > > > >      return vhost_set_vring(dev, VHOST_USER_SET_VRING_BASE, ring);
> > > > > > >  }
> > > > > > >  
> > > > > > 
> > > > > > but on vq start we do not reinstate it? Does this not mean that
> > > > > > notifiers won't work after stop then start?
> > > > > 
> > > > > 
> > > > > Yes, backend initially work w/o host notifier, request VM driver to
> > > > > install notifier if needed - call this function through slave socket.
> > > > 
> > > > I think it's cleaner if qemu handles this itself like it did before, it
> > > > knows vm is stopped without getting called.
> > > 
> > > If vhost play as server, there are 2 scenario that remove the notifier:
> > > 1. VM suspend: backend still there, it's okay to keep mmap address.
> > > 2. vhost backend stopped or process killed: resources from backend
> > > should be released. That's why patch 2/2 munmap in notifier remove
> > > function. Then the restore function get nothing to restore, maybe I
> > > shouldn't reverse patch order.
> > 
> > I can't say I understand what you mean here. Do you plan to change
> > the patchset in some way?
> > When you do, pls include a cover letter with a changelog and
> > Cc all people you include on patches on the cover letter too. 
> 
> Here is the detail of the problem I encountered, my vhost backend is
> DPDK vdpa user space application. Notifier address is set when vdpa ask
> qemu to mmap a FD and an offset from vdpa, see function
> vhost_user_slave_handle_vring_host_notifier(). If the vdpa application
> restart of get killed for some reason,
> vhost_user_host_notifier_remove() is called and notifier MR is
> uninstalled, the notifier address that retrieved from mmap is
> referencing to an invalid FD, not released. This will cause HW
> resources on kernel side still referenced, most important, when vdpa
> connection restored, this invalid notifier will be be restored as
> notifier MR.
> 
> To resolve it, have to remove the notifer restore mechanism, vDPA
> application will issue client socket request again to install notifier
> to VM, so no concern that the notifier will be lost after resume.
> 
> Since vdpa might be killed, no chance to notify qemu to remove
> notifier. An alternative solution is to detect sock disconnection and
> unmmap notifier, but it looks more complex to me. How do you think?
> 
> 
> > 
> > > > 
> > > > > > 
> > > > > > 
> > > > > > > @@ -1538,7 +1522,6 @@ static int 
> > > > > > > vhost_user_slave_handle_vring_host_notifier(struct vhost_dev *dev,
> > > > > > >      }
> > > > > > >  
> > > > > > >      n->addr = addr;
> > > > > > > -    n->set = true;
> > > > > > >  
> > > > > > >      return 0;
> > > > > > >  }
> > > > > > > diff --git a/include/hw/virtio/vhost-user.h 
> > > > > > > b/include/hw/virtio/vhost-user.h
> > > > > > > index a9abca3288..f6012b2078 100644
> > > > > > > --- a/include/hw/virtio/vhost-user.h
> > > > > > > +++ b/include/hw/virtio/vhost-user.h
> > > > > > > @@ -14,7 +14,6 @@
> > > > > > >  typedef struct VhostUserHostNotifier {
> > > > > > >      MemoryRegion mr;
> > > > > > >      void *addr;
> > > > > > > -    bool set;
> > > > > > >  } VhostUserHostNotifier;
> > > > > > >  
> > > > > > >  typedef struct VhostUserState {
> > > > > > > -- 
> > > > > > > 2.33.0
> > > > > > 
> > > > > 
> > > > 
> > > 
> > 
> 




reply via email to

[Prev in Thread] Current Thread [Next in Thread]