Jakub Staroń
2019-May-07 20:25 UTC
[Qemu-devel] [PATCH v7 2/6] virtio-pmem: Add virtio pmem driver
On 4/25/19 10:00 PM, Pankaj Gupta wrote:> +void host_ack(struct virtqueue *vq) > +{ > + unsigned int len; > + unsigned long flags; > + struct virtio_pmem_request *req, *req_buf; > + struct virtio_pmem *vpmem = vq->vdev->priv; > + > + spin_lock_irqsave(&vpmem->pmem_lock, flags); > + while ((req = virtqueue_get_buf(vq, &len)) != NULL) { > + req->done = true; > + wake_up(&req->host_acked); > + > + if (!list_empty(&vpmem->req_list)) { > + req_buf = list_first_entry(&vpmem->req_list, > + struct virtio_pmem_request, list); > + list_del(&vpmem->req_list);Shouldn't it be rather `list_del(vpmem->req_list.next)`? We are trying to unlink first element of the list and `vpmem->req_list` is just the list head.> +int virtio_pmem_flush(struct nd_region *nd_region) > +{ > + int err; > + unsigned long flags; > + struct scatterlist *sgs[2], sg, ret; > + struct virtio_device *vdev = nd_region->provider_data; > + struct virtio_pmem *vpmem = vdev->priv; > + struct virtio_pmem_request *req; > + > + might_sleep(); > + req = kmalloc(sizeof(*req), GFP_KERNEL); > + if (!req) > + return -ENOMEM; > + > + req->done = req->wq_buf_avail = false; > + strcpy(req->name, "FLUSH"); > + init_waitqueue_head(&req->host_acked); > + init_waitqueue_head(&req->wq_buf); > + sg_init_one(&sg, req->name, strlen(req->name)); > + sgs[0] = &sg; > + sg_init_one(&ret, &req->ret, sizeof(req->ret)); > + sgs[1] = &ret; > + > + spin_lock_irqsave(&vpmem->pmem_lock, flags); > + err = virtqueue_add_sgs(vpmem->req_vq, sgs, 1, 1, req, GFP_ATOMIC); > + if (err) { > + dev_err(&vdev->dev, "failed to send command to virtio pmem device\n"); > + > + list_add_tail(&vpmem->req_list, &req->list); > + spin_unlock_irqrestore(&vpmem->pmem_lock, flags); > + > + /* When host has read buffer, this completes via host_ack */ > + wait_event(req->wq_buf, req->wq_buf_avail); > + spin_lock_irqsave(&vpmem->pmem_lock, flags); > + }Aren't the arguments in `list_add_tail` swapped? The element we are adding should be first, the list should be second. Also, shouldn't we resubmit the request after waking up from `wait_event(req->wq_buf, req->wq_buf_avail)`? I propose rewriting it like that: diff --git a/drivers/nvdimm/virtio_pmem.c b/drivers/nvdimm/virtio_pmem.c index 66b582f751a3..ff0556b04e86 100644 --- a/drivers/nvdimm/virtio_pmem.c +++ b/drivers/nvdimm/virtio_pmem.c @@ -25,7 +25,7 @@ void host_ack(struct virtqueue *vq) if (!list_empty(&vpmem->req_list)) { req_buf = list_first_entry(&vpmem->req_list, struct virtio_pmem_request, list); - list_del(&vpmem->req_list); + list_del(vpmem->req_list.next); req_buf->wq_buf_avail = true; wake_up(&req_buf->wq_buf); } @@ -59,17 +59,33 @@ int virtio_pmem_flush(struct nd_region *nd_region) sgs[1] = &ret; spin_lock_irqsave(&vpmem->pmem_lock, flags); - err = virtqueue_add_sgs(vpmem->req_vq, sgs, 1, 1, req, GFP_ATOMIC); - if (err) { - dev_err(&vdev->dev, "failed to send command to virtio pmem device\n"); + /* + * If virtqueue_add_sgs returns -ENOSPC then req_vq virtual queue does not + * have free descriptor slots. We add the request to req_list and wait + * for host_ack to wake us up when free slots are available. + */ + while ((err = virtqueue_add_sgs(vpmem->req_vq, sgs, 1, 1, req, GFP_ATOMIC)) == -ENOSPC) { + dev_err(&vdev->dev, "failed to send command to virtio pmem device, no free slots in the virtqueue, postponing request\n"); + req->wq_buf_avail = false; - list_add_tail(&vpmem->req_list, &req->list); + list_add_tail(&req->list, &vpmem->req_list); spin_unlock_irqrestore(&vpmem->pmem_lock, flags); /* When host has read buffer, this completes via host_ack */ wait_event(req->wq_buf, req->wq_buf_avail); spin_lock_irqsave(&vpmem->pmem_lock, flags); } + + /* + * virtqueue_add_sgs failed with error different than -ENOSPC, we can't + * do anything about that. + */ + if (err) { + dev_info(&vdev->dev, "failed to send command to virtio pmem device, error code %d\n", err); + spin_unlock_irqrestore(&vpmem->pmem_lock, flags); + err = -EIO; + goto ret; + } err = virtqueue_kick(vpmem->req_vq); spin_unlock_irqrestore(&vpmem->pmem_lock, flags); Let me know if it looks reasonable to you. Thank you, Jakub Staron
Pankaj Gupta
2019-May-08 11:12 UTC
[Qemu-devel] [PATCH v7 2/6] virtio-pmem: Add virtio pmem driver
> > On 4/25/19 10:00 PM, Pankaj Gupta wrote: > > > +void host_ack(struct virtqueue *vq) > > +{ > > + unsigned int len; > > + unsigned long flags; > > + struct virtio_pmem_request *req, *req_buf; > > + struct virtio_pmem *vpmem = vq->vdev->priv; > > + > > + spin_lock_irqsave(&vpmem->pmem_lock, flags); > > + while ((req = virtqueue_get_buf(vq, &len)) != NULL) { > > + req->done = true; > > + wake_up(&req->host_acked); > > + > > + if (!list_empty(&vpmem->req_list)) { > > + req_buf = list_first_entry(&vpmem->req_list, > > + struct virtio_pmem_request, list); > > + list_del(&vpmem->req_list); > > Shouldn't it be rather `list_del(vpmem->req_list.next)`? We are trying to > unlink > first element of the list and `vpmem->req_list` is just the list head.This looks correct. We are not deleting head but first entry in 'req_list' which is device corresponding list of pending requests. Please see below: /** * Retrieve the first list entry for the given list pointer. * * Example: * struct foo *first; * first = list_first_entry(&bar->list_of_foos, struct foo, list_of_foos); * * @param ptr The list head * @param type Data type of the list element to retrieve * @param member Member name of the struct list_head field in the list element. * @return A pointer to the first list element. */ #define list_first_entry(ptr, type, member) \ list_entry((ptr)->next, type, member)> > > +int virtio_pmem_flush(struct nd_region *nd_region) > > +{ > > + int err; > > + unsigned long flags; > > + struct scatterlist *sgs[2], sg, ret; > > + struct virtio_device *vdev = nd_region->provider_data; > > + struct virtio_pmem *vpmem = vdev->priv; > > + struct virtio_pmem_request *req; > > + > > + might_sleep(); > > + req = kmalloc(sizeof(*req), GFP_KERNEL); > > + if (!req) > > + return -ENOMEM; > > + > > + req->done = req->wq_buf_avail = false; > > + strcpy(req->name, "FLUSH"); > > + init_waitqueue_head(&req->host_acked); > > + init_waitqueue_head(&req->wq_buf); > > + sg_init_one(&sg, req->name, strlen(req->name)); > > + sgs[0] = &sg; > > + sg_init_one(&ret, &req->ret, sizeof(req->ret)); > > + sgs[1] = &ret; > > + > > + spin_lock_irqsave(&vpmem->pmem_lock, flags); > > + err = virtqueue_add_sgs(vpmem->req_vq, sgs, 1, 1, req, GFP_ATOMIC); > > + if (err) { > > + dev_err(&vdev->dev, "failed to send command to virtio pmem device\n"); > > + > > + list_add_tail(&vpmem->req_list, &req->list); > > + spin_unlock_irqrestore(&vpmem->pmem_lock, flags); > > + > > + /* When host has read buffer, this completes via host_ack */ > > + wait_event(req->wq_buf, req->wq_buf_avail); > > + spin_lock_irqsave(&vpmem->pmem_lock, flags); > > + } > > Aren't the arguments in `list_add_tail` swapped? The element we are addingNo, this is intentional. 'vpmem->req_list' maintains a list of pending requests for entire pmem device. 'req->list'is per request list and maintains pending request on virtio queue add failure. I think we don't need this list.> should > be first, the list should be second. Also, shouldn't we resubmit the request > after > waking up from `wait_event(req->wq_buf, req->wq_buf_avail)`?Yes. we should. Good point.> > I propose rewriting it like that: > > diff --git a/drivers/nvdimm/virtio_pmem.c b/drivers/nvdimm/virtio_pmem.c > index 66b582f751a3..ff0556b04e86 100644 > --- a/drivers/nvdimm/virtio_pmem.c > +++ b/drivers/nvdimm/virtio_pmem.c > @@ -25,7 +25,7 @@ void host_ack(struct virtqueue *vq) > if (!list_empty(&vpmem->req_list)) { > req_buf = list_first_entry(&vpmem->req_list, > struct virtio_pmem_request, list); > - list_del(&vpmem->req_list); > + list_del(vpmem->req_list.next);Don't think its correct.> req_buf->wq_buf_avail = true; > wake_up(&req_buf->wq_buf); > } > @@ -59,17 +59,33 @@ int virtio_pmem_flush(struct nd_region *nd_region) > sgs[1] = &ret; > > spin_lock_irqsave(&vpmem->pmem_lock, flags); > - err = virtqueue_add_sgs(vpmem->req_vq, sgs, 1, 1, req, GFP_ATOMIC); > - if (err) { > - dev_err(&vdev->dev, "failed to send command to virtio pmem device\n"); > + /* > + * If virtqueue_add_sgs returns -ENOSPC then req_vq virtual queue does not > + * have free descriptor slots. We add the request to req_list and wait > + * for host_ack to wake us up when free slots are available. > + */ > + while ((err = virtqueue_add_sgs(vpmem->req_vq, sgs, 1, 1, req, GFP_ATOMIC)) > == -ENOSPC) { > + dev_err(&vdev->dev, "failed to send command to virtio pmem device, no free > slots in the virtqueue, postponing request\n"); > + req->wq_buf_avail = false; > > - list_add_tail(&vpmem->req_list, &req->list); > + list_add_tail(&req->list, &vpmem->req_list); > spin_unlock_irqrestore(&vpmem->pmem_lock, flags); > > /* When host has read buffer, this completes via host_ack */ > wait_event(req->wq_buf, req->wq_buf_avail); > spin_lock_irqsave(&vpmem->pmem_lock, flags); > } > + > + /* > + * virtqueue_add_sgs failed with error different than -ENOSPC, we can't > + * do anything about that. > + */ > + if (err) { > + dev_info(&vdev->dev, "failed to send command to virtio pmem device, error > code %d\n", err); > + spin_unlock_irqrestore(&vpmem->pmem_lock, flags); > + err = -EIO; > + goto ret; > + } > err = virtqueue_kick(vpmem->req_vq); > spin_unlock_irqrestore(&vpmem->pmem_lock, flags); > > > Let me know if it looks reasonable to you.Don't think this is fulfilling entire logic correctly. But thanks, I spotted a bug in my code :) Will fix it.> > Thank you, > Jakub Staron > >
Pankaj Gupta
2019-May-08 15:23 UTC
[Qemu-devel] [PATCH v7 2/6] virtio-pmem: Add virtio pmem driver
> > > > > +int virtio_pmem_flush(struct nd_region *nd_region) > > > +{ > > > + int err; > > > + unsigned long flags; > > > + struct scatterlist *sgs[2], sg, ret; > > > + struct virtio_device *vdev = nd_region->provider_data; > > > + struct virtio_pmem *vpmem = vdev->priv; > > > + struct virtio_pmem_request *req; > > > + > > > + might_sleep(); > > > + req = kmalloc(sizeof(*req), GFP_KERNEL); > > > + if (!req) > > > + return -ENOMEM; > > > + > > > + req->done = req->wq_buf_avail = false; > > > + strcpy(req->name, "FLUSH"); > > > + init_waitqueue_head(&req->host_acked); > > > + init_waitqueue_head(&req->wq_buf); > > > + sg_init_one(&sg, req->name, strlen(req->name)); > > > + sgs[0] = &sg; > > > + sg_init_one(&ret, &req->ret, sizeof(req->ret)); > > > + sgs[1] = &ret; > > > + > > > + spin_lock_irqsave(&vpmem->pmem_lock, flags); > > > + err = virtqueue_add_sgs(vpmem->req_vq, sgs, 1, 1, req, GFP_ATOMIC); > > > + if (err) { > > > + dev_err(&vdev->dev, "failed to send command to virtio pmem device\n"); > > > + > > > + list_add_tail(&vpmem->req_list, &req->list); > > > + spin_unlock_irqrestore(&vpmem->pmem_lock, flags); > > > + > > > + /* When host has read buffer, this completes via host_ack */ > > > + wait_event(req->wq_buf, req->wq_buf_avail); > > > + spin_lock_irqsave(&vpmem->pmem_lock, flags); > > > + } > > > > Aren't the arguments in `list_add_tail` swapped? The element we are adding >Yes, arguments for 'list_add_tail' should be swapped. list_add_tail(&req->list, &vpmem->req_list); Thank you, Pankaj
Jakub Staroń
2019-May-08 19:05 UTC
[Qemu-devel] [PATCH v7 2/6] virtio-pmem: Add virtio pmem driver
On 5/8/19 4:12 AM, Pankaj Gupta wrote:> >> >> On 4/25/19 10:00 PM, Pankaj Gupta wrote: >> >>> +void host_ack(struct virtqueue *vq) >>> +{ >>> + unsigned int len; >>> + unsigned long flags; >>> + struct virtio_pmem_request *req, *req_buf; >>> + struct virtio_pmem *vpmem = vq->vdev->priv; >>> + >>> + spin_lock_irqsave(&vpmem->pmem_lock, flags); >>> + while ((req = virtqueue_get_buf(vq, &len)) != NULL) { >>> + req->done = true; >>> + wake_up(&req->host_acked); >>> + >>> + if (!list_empty(&vpmem->req_list)) { >>> + req_buf = list_first_entry(&vpmem->req_list, >>> + struct virtio_pmem_request, list); >>> + list_del(&vpmem->req_list); >> >> Shouldn't it be rather `list_del(vpmem->req_list.next)`? We are trying to >> unlink >> first element of the list and `vpmem->req_list` is just the list head. > > This looks correct. We are not deleting head but first entry in 'req_list' > which is device corresponding list of pending requests. > > Please see below: > > /** > * Retrieve the first list entry for the given list pointer. > * > * Example: > * struct foo *first; > * first = list_first_entry(&bar->list_of_foos, struct foo, list_of_foos); > * > * @param ptr The list head > * @param type Data type of the list element to retrieve > * @param member Member name of the struct list_head field in the list element. > * @return A pointer to the first list element. > */ > #define list_first_entry(ptr, type, member) \ > list_entry((ptr)->next, type, member)Please look at this StackOverflow question: https://stackoverflow.com/questions/19675419/deleting-first-element-of-a-list-h-list Author asks about deleting first element of the queue. In our case (and also in the question's author case), `vpmem->req_list` is not element of any request struct and not an element of the list. It's just a list head storing `next` and `prev` pointers which are then pointing to respectively first and last element of the list. We want to unlink the first element of the list, so we need to pass pointer to the first element of the list to the `list_del` function - that is, the `vpmem->req_list.next`. Thank you, Jakub Staron
Seemingly Similar Threads
- [Qemu-devel] [PATCH v7 2/6] virtio-pmem: Add virtio pmem driver
- [Qemu-devel] [PATCH v7 2/6] virtio-pmem: Add virtio pmem driver
- [PATCH v12 2/7] virtio-pmem: Add virtio pmem driver
- [PATCH v8 2/6] virtio-pmem: Add virtio pmem driver
- [PATCH v4 2/5] virtio-pmem: Add virtio pmem driver