On Sun, Mar 15, 2015 at 02:50:27PM +0800, Linhaifeng wrote:> Hi,Michael > > I had tested the start_xmit function by the follow code found that the tx queue's state is stopped and can't send any packets anymore.Why don't you Cc all maintainers on this email? Pls check the file MAINTAINERS for the full list. I added Cc for now.> > static netdev_tx_t start_xmit(struct sk_buff *skb, struct net_device *dev) > { > ... ... > > > capacity = 10; //########## test code : force to call netif_stop_queue > > if (capacity < 2+MAX_SKB_FRAGS) { > netif_stop_queue(dev);So you changed code to make it think we are out of capacity, now it stops the queue.> > if (unlikely(!virtqueue_enable_cb_delayed(vi->svq))) { > /* More just got used, free them then recheck. */ > capacity += free_old_xmit_skbs(vi); > dev_warn(&dev->dev, "free_old_xmit_skbs capacity =%d MAX_SKB_FRAGS=%d", capacity, MAX_SKB_FRAGS); > > capacity = 10; //########## test code : force not to call netif_start_queue > > if (capacity >= 2+MAX_SKB_FRAGS) { > netif_start_queue(dev); > virtqueue_disable_cb(vi->svq); > } else { > //########## OTOH if often enter this branch tx queue maybe stopped. > }and changed it here so it won't restart queue if host consumed all buffers. unsurprisingly this makes driver not work.> > } > > //########## Should we start queue here? I found that sometimes skb_xmit_done run before netif_stop_queue if this occurred the queue's state is > //########## stopped and have to reload virtio-net module to restore network.With or without your changes? Is this the condition you describe? if (sq->vq->num_free < 2+MAX_SKB_FRAGS) { ---> at this point, skb_xmit_done runs. this does: /* Suppress further interrupts. */ virtqueue_disable_cb(vq); /* We were probably waiting for more output buffers. */ netif_wake_subqueue(vi->dev, vq2txq(vq)); ---> netif_stop_subqueue(dev, qnum); ---> queue is now stopped if (unlikely(!virtqueue_enable_cb_delayed(sq->vq))) { ----> this re-enables interrupts, after an interrupt skb_xmit_done will run again. /* More just got used, free them then recheck. * */ free_old_xmit_skbs(sq); if (sq->vq->num_free >= 2+MAX_SKB_FRAGS) { netif_start_subqueue(dev, qnum); virtqueue_disable_cb(sq->vq); } } } I can't see a race condition from your description above.> } > > } > > ping 9.62.1.2 -i 0.1 > 64 bytes from 9.62.1.2: icmp_seq=19 ttl=64 time=0.115 ms > 64 bytes from 9.62.1.2: icmp_seq=20 ttl=64 time=0.101 ms > 64 bytes from 9.62.1.2: icmp_seq=21 ttl=64 time=0.094 ms > 64 bytes from 9.62.1.2: icmp_seq=22 ttl=64 time=0.098 ms > 64 bytes from 9.62.1.2: icmp_seq=23 ttl=64 time=0.097 ms > 64 bytes from 9.62.1.2: icmp_seq=24 ttl=64 time=0.095 ms > 64 bytes from 9.62.1.2: icmp_seq=25 ttl=64 time=0.095 ms > .... > ping: sendmsg: No buffer space available > ping: sendmsg: No buffer space available > ping: sendmsg: No buffer space available > ping: sendmsg: No buffer space available > ping: sendmsg: No buffer space available > ping: sendmsg: No buffer space available > .... > > -- > Regards, > HaifengI can't say what does your code-changing experiment show. It might be better to introduce delay by calling something like cpu_relax at specific points (maybe multiple times in a loop). -- MST
On 2015/3/15 16:40, Michael S. Tsirkin wrote:> On Sun, Mar 15, 2015 at 02:50:27PM +0800, Linhaifeng wrote: >> Hi,Michael >> >> I had tested the start_xmit function by the follow code found that the tx queue's state is stopped and can't send any packets anymore. > > Why don't you Cc all maintainers on this email? > Pls check the file MAINTAINERS for the full list. > I added Cc for now. >Thank you.>> >> static netdev_tx_t start_xmit(struct sk_buff *skb, struct net_device *dev) >> { >> ... ... >> >> >> capacity = 10; //########## test code : force to call netif_stop_queue >> >> if (capacity < 2+MAX_SKB_FRAGS) { >> netif_stop_queue(dev); > > So you changed code to make it think we are out of capacity, now it > stops the queue. > >> >> if (unlikely(!virtqueue_enable_cb_delayed(vi->svq))) { >> /* More just got used, free them then recheck. */ >> capacity += free_old_xmit_skbs(vi); >> dev_warn(&dev->dev, "free_old_xmit_skbs capacity =%d MAX_SKB_FRAGS=%d", capacity, MAX_SKB_FRAGS); >> >> capacity = 10; //########## test code : force not to call netif_start_queue >> >> if (capacity >= 2+MAX_SKB_FRAGS) { >> netif_start_queue(dev); >> virtqueue_disable_cb(vi->svq); >> } else { >> //########## OTOH if often enter this branch tx queue maybe stopped. >> } > > and changed it here so it won't restart queue if host consumed > all buffers. > unsurprisingly this makes driver not work. > > >> >> } >> >> //########## Should we start queue here? I found that sometimes skb_xmit_done run before netif_stop_queue if this occurred the queue's state is >> //########## stopped and have to reload virtio-net module to restore network. > > With or without your changes?without> Is this the condition you describe? > > > if (sq->vq->num_free < 2+MAX_SKB_FRAGS) { > > ---> at this point, skb_xmit_done runs. this does: > /* Suppress further interrupts. */ > virtqueue_disable_cb(vq); > > /* We were probably waiting for more output buffers. */ > netif_wake_subqueue(vi->dev, vq2txq(vq)); > ---> > > >Because i use vhost-user(poll mode) with virtio_net so at this time vhost had received all packets.> netif_stop_subqueue(dev, qnum); > > ---> queue is now stopped > > if (unlikely(!virtqueue_enable_cb_delayed(sq->vq))) { > > ----> this re-enables interrupts, after an interrupt skb_xmit_done > will run again. >Before netif_stop_subqueue called vhost had received all packets so virtio_net will never receive any skb_xmit_done. If vhost is in poll mode should we need or not to stop tx queue? Can i add a flag VHOST_F_POLL_MODE to support poll mode vhost(vhost-user)?> /* More just got used, free them then recheck. > * */ > free_old_xmit_skbs(sq); > if (sq->vq->num_free >= 2+MAX_SKB_FRAGS) { > netif_start_subqueue(dev, qnum); > virtqueue_disable_cb(sq->vq); > } > } > } > > > I can't see a race condition from your description above. > >> } >> >> } >> >> ping 9.62.1.2 -i 0.1 >> 64 bytes from 9.62.1.2: icmp_seq=19 ttl=64 time=0.115 ms >> 64 bytes from 9.62.1.2: icmp_seq=20 ttl=64 time=0.101 ms >> 64 bytes from 9.62.1.2: icmp_seq=21 ttl=64 time=0.094 ms >> 64 bytes from 9.62.1.2: icmp_seq=22 ttl=64 time=0.098 ms >> 64 bytes from 9.62.1.2: icmp_seq=23 ttl=64 time=0.097 ms >> 64 bytes from 9.62.1.2: icmp_seq=24 ttl=64 time=0.095 ms >> 64 bytes from 9.62.1.2: icmp_seq=25 ttl=64 time=0.095 ms >> .... >> ping: sendmsg: No buffer space available >> ping: sendmsg: No buffer space available >> ping: sendmsg: No buffer space available >> ping: sendmsg: No buffer space available >> ping: sendmsg: No buffer space available >> ping: sendmsg: No buffer space available >> .... >> >> -- >> Regards, >> Haifeng > > I can't say what does your code-changing experiment show. > It might be better to introduce delay by calling something like > cpu_relax at specific points (maybe multiple times in a loop). >-- Regards, Haifeng
On Mon, Mar 16, 2015 at 05:24:07PM +0800, Linhaifeng wrote:> > > On 2015/3/15 16:40, Michael S. Tsirkin wrote: > > On Sun, Mar 15, 2015 at 02:50:27PM +0800, Linhaifeng wrote: > >> Hi,Michael > >> > >> I had tested the start_xmit function by the follow code found that the tx queue's state is stopped and can't send any packets anymore. > > > > Why don't you Cc all maintainers on this email? > > Pls check the file MAINTAINERS for the full list. > > I added Cc for now. > > > > Thank you. > > >> > >> static netdev_tx_t start_xmit(struct sk_buff *skb, struct net_device *dev) > >> { > >> ... ... > >> > >> > >> capacity = 10; //########## test code : force to call netif_stop_queue > >> > >> if (capacity < 2+MAX_SKB_FRAGS) { > >> netif_stop_queue(dev); > > > > So you changed code to make it think we are out of capacity, now it > > stops the queue. > > > >> > >> if (unlikely(!virtqueue_enable_cb_delayed(vi->svq))) { > >> /* More just got used, free them then recheck. */ > >> capacity += free_old_xmit_skbs(vi); > >> dev_warn(&dev->dev, "free_old_xmit_skbs capacity =%d MAX_SKB_FRAGS=%d", capacity, MAX_SKB_FRAGS); > >> > >> capacity = 10; //########## test code : force not to call netif_start_queue > >> > >> if (capacity >= 2+MAX_SKB_FRAGS) { > >> netif_start_queue(dev); > >> virtqueue_disable_cb(vi->svq); > >> } else { > >> //########## OTOH if often enter this branch tx queue maybe stopped. > >> } > > > > and changed it here so it won't restart queue if host consumed > > all buffers. > > unsurprisingly this makes driver not work. > > > > > >> > >> } > >> > >> //########## Should we start queue here? I found that sometimes skb_xmit_done run before netif_stop_queue if this occurred the queue's state is > >> //########## stopped and have to reload virtio-net module to restore network. > > > > With or without your changes? > > without > > > Is this the condition you describe? > > > > > > if (sq->vq->num_free < 2+MAX_SKB_FRAGS) { > > > > ---> at this point, skb_xmit_done runs. this does: > > /* Suppress further interrupts. */ > > virtqueue_disable_cb(vq); > > > > /* We were probably waiting for more output buffers. */ > > netif_wake_subqueue(vi->dev, vq2txq(vq)); > > ---> > > > > > > > > Because i use vhost-user(poll mode) with virtio_net so at this time vhost > had received all packets.Must likely a vhost-user bug then.> > netif_stop_subqueue(dev, qnum); > > > > ---> queue is now stopped > > > > if (unlikely(!virtqueue_enable_cb_delayed(sq->vq))) { > > > > ----> this re-enables interrupts, after an interrupt skb_xmit_done > > will run again. > > > > Before netif_stop_subqueue called vhost had received all packets so virtio_net > will never receive any skb_xmit_done.And completed them in the used ring? In that case virtqueue_enable_cb_delayed will return false, so we'll call free_old_xmit_skbs below, and restart ring.> If vhost is in poll mode should we need or not to stop tx queue? > Can i add a flag VHOST_F_POLL_MODE to support poll mode vhost(vhost-user)?Host just needs to be spec-compliant. It must send interrupts unless they are disabled. So this sounds like a VHOST_F_FIX_A_BUG to me. Just fix races in vhost-user code, and no need for extra flags.> > /* More just got used, free them then recheck. > > * */ > > free_old_xmit_skbs(sq); > > if (sq->vq->num_free >= 2+MAX_SKB_FRAGS) { > > netif_start_subqueue(dev, qnum); > > virtqueue_disable_cb(sq->vq); > > } > > } > > } > > > > > > I can't see a race condition from your description above. > > > >> } > >> > >> } > >> > >> ping 9.62.1.2 -i 0.1 > >> 64 bytes from 9.62.1.2: icmp_seq=19 ttl=64 time=0.115 ms > >> 64 bytes from 9.62.1.2: icmp_seq=20 ttl=64 time=0.101 ms > >> 64 bytes from 9.62.1.2: icmp_seq=21 ttl=64 time=0.094 ms > >> 64 bytes from 9.62.1.2: icmp_seq=22 ttl=64 time=0.098 ms > >> 64 bytes from 9.62.1.2: icmp_seq=23 ttl=64 time=0.097 ms > >> 64 bytes from 9.62.1.2: icmp_seq=24 ttl=64 time=0.095 ms > >> 64 bytes from 9.62.1.2: icmp_seq=25 ttl=64 time=0.095 ms > >> .... > >> ping: sendmsg: No buffer space available > >> ping: sendmsg: No buffer space available > >> ping: sendmsg: No buffer space available > >> ping: sendmsg: No buffer space available > >> ping: sendmsg: No buffer space available > >> ping: sendmsg: No buffer space available > >> .... > >> > >> -- > >> Regards, > >> Haifeng > > > > I can't say what does your code-changing experiment show. > > It might be better to introduce delay by calling something like > > cpu_relax at specific points (maybe multiple times in a loop). > > > > > > -- > Regards, > Haifeng