Jason Wang
2020-May-28 10:06 UTC
[PATCH] vdpa: bypass waking up vhost_woker for vdpa vq kick
On 2020/5/26 ??1:32, Zhu Lingshan wrote:> Standard vhost devices rely on waking up a vhost_worker to kick > a virtquque. However vdpa devices have hardware backends, so it > does not need this waking up routin. In this commit, vdpa device > will kick a virtqueue directly, reduce the performance overhead > caused by waking up a vhost_woker.Thanks for the patch. It would be helpful if you can share some performance numbers. And the title should be "vhost-vdpa:" instead of "vdpa:" This patch is important since we want to get rid of ktrhead and use_mm()/unuse_mm() stuffs which allows us to implement doorbell mapping.> > Signed-off-by: Zhu Lingshan <lingshan.zhu at intel.com> > Suggested-by: Jason Wang <jasowang at redhat.com> > --- > drivers/vhost/vdpa.c | 100 +++++++++++++++++++++++++++++++++++++++++++++++++++ > 1 file changed, 100 insertions(+) > > diff --git a/drivers/vhost/vdpa.c b/drivers/vhost/vdpa.c > index 0968361..d3a2aca 100644 > --- a/drivers/vhost/vdpa.c > +++ b/drivers/vhost/vdpa.c > @@ -287,6 +287,66 @@ static long vhost_vdpa_get_vring_num(struct vhost_vdpa *v, u16 __user *argp) > > return 0; > } > +void vhost_vdpa_poll_stop(struct vhost_virtqueue *vq) > +{ > + vhost_poll_stop(&vq->poll); > +} > + > +int vhost_vdpa_poll_start(struct vhost_virtqueue *vq) > +{ > + struct vhost_poll *poll = &vq->poll; > + struct file *file = vq->kick; > + __poll_t mask; > + > + > + if (poll->wqh) > + return 0; > + > + mask = vfs_poll(file, &poll->table); > + if (mask) > + vq->handle_kick(&vq->poll.work); > + if (mask & EPOLLERR) { > + vhost_poll_stop(poll); > + return -EINVAL; > + } > + > + return 0; > +}So this basically a duplication of vhost_poll_start()?> + > +static long vhost_vdpa_set_vring_kick(struct vhost_virtqueue *vq, > + void __user *argp) > +{ > + bool pollstart = false, pollstop = false; > + struct file *eventfp, *filep = NULL; > + struct vhost_vring_file f; > + long r; > + > + if (copy_from_user(&f, argp, sizeof(f))) > + return -EFAULT; > + > + eventfp = f.fd == -1 ? NULL : eventfd_fget(f.fd); > + if (IS_ERR(eventfp)) { > + r = PTR_ERR(eventfp); > + return r; > + } > + > + if (eventfp != vq->kick) { > + pollstop = (filep = vq->kick) != NULL; > + pollstart = (vq->kick = eventfp) != NULL; > + } else > + filep = eventfp; > + > + if (pollstop && vq->handle_kick) > + vhost_vdpa_poll_stop(vq); > + > + if (filep) > + fput(filep); > + > + if (pollstart && vq->handle_kick) > + r = vhost_vdpa_poll_start(vq); > + > + return r; > +} > > static long vhost_vdpa_vring_ioctl(struct vhost_vdpa *v, unsigned int cmd, > void __user *argp) > @@ -316,6 +376,11 @@ static long vhost_vdpa_vring_ioctl(struct vhost_vdpa *v, unsigned int cmd, > return 0; > } > > + if (cmd == VHOST_SET_VRING_KICK) { > + r = vhost_vdpa_set_vring_kick(vq, argp); > + return r; > + } > + > if (cmd == VHOST_GET_VRING_BASE) > vq->last_avail_idx = ops->get_vq_state(v->vdpa, idx); > > @@ -667,6 +732,39 @@ static void vhost_vdpa_free_domain(struct vhost_vdpa *v) > v->domain = NULL; > } > > +static int vhost_vdpa_poll_worker(wait_queue_entry_t *wait, unsigned int mode, > + int sync, void *key) > +{ > + struct vhost_poll *poll = container_of(wait, struct vhost_poll, wait); > + struct vhost_virtqueue *vq = container_of(poll, struct vhost_virtqueue, > + poll); > + > + if (!(key_to_poll(key) & poll->mask)) > + return 0; > + > + vq->handle_kick(&vq->poll.work); > + > + return 0; > +} > + > +void vhost_vdpa_poll_init(struct vhost_dev *dev) > +{ > + struct vhost_virtqueue *vq; > + struct vhost_poll *poll; > + int i; > + > + for (i = 0; i < dev->nvqs; i++) { > + vq = dev->vqs[i]; > + poll = &vq->poll; > + if (vq->handle_kick) { > + init_waitqueue_func_entry(&poll->wait, > + vhost_vdpa_poll_worker); > + poll->work.fn = vq->handle_kick;Why this is needed?> + } > + > + } > +} > + > static int vhost_vdpa_open(struct inode *inode, struct file *filep) > { > struct vhost_vdpa *v; > @@ -697,6 +795,8 @@ static int vhost_vdpa_open(struct inode *inode, struct file *filep) > vhost_dev_init(dev, vqs, nvqs, 0, 0, 0, > vhost_vdpa_process_iotlb_msg); > > + vhost_vdpa_poll_init(dev); > + > dev->iotlb = vhost_iotlb_alloc(0, 0); > if (!dev->iotlb) { > r = -ENOMEM;So my feeling here is that you want to reuse the infrastructure in vhost.c as much as possible If this is true, let's just avoid duplicating the codes. How about adding something like in vhost_poll_wakeup(): ??? struct vhost_poll *poll = container_of(wait, struct vhost_poll, wait); ??? struct vhost_work *work = &poll->work; ??? if (!(key_to_poll(key) & poll->mask)) ??? ??? return 0; ??? if (!poll->dev->use_worker) ??? ??? work->fn(work); ??? else ??? ??? vhost_poll_queue(poll); Then modify vhost_dev_init() to set use_worker (all true except for vdpa)? Thanks
Apparently Analagous Threads
- [PATCH] vdpa: bypass waking up vhost_woker for vdpa vq kick
- [PATCH] vdpa: bypass waking up vhost_woker for vdpa vq kick
- [PATCH] vdpa: bypass waking up vhost_woker for vdpa vq kick
- [PATCH] vdpa: bypass waking up vhost_woker for vdpa vq kick
- [RFC PATCH] vdpa: vhost_vdpa_poll_stop() can be static