search for: locklessly

Displaying 20 results from an estimated 213 matches for "locklessly".

Did you mean: lockless
2018 Mar 04
3
sqlinux weirdness
Every now and then I get an alert like this one. I have no clue what this "rear" subsystem is, or why madam would be trying to write to its log file. Can anyone enlighten me? thanks in advance! ------------------------- SELinux is preventing /usr/sbin/mdadm from write access on the file /var/log/rear/rear-fcshome.log.lockless. ***** Plugin restorecon (93.9 confidence) suggests
2017 Mar 30
1
[PATCH net-next 8/8] vhost_net: use lockless peeking for skb array during busy polling
On Thu, Mar 30, 2017 at 10:16:15AM +0800, Jason Wang wrote: > > > On 2017?03?29? 20:07, Michael S. Tsirkin wrote: > > On Tue, Mar 21, 2017 at 12:04:47PM +0800, Jason Wang wrote: > > > For the socket that exports its skb array, we can use lockless polling > > > to avoid touching spinlock during busy polling. > > > > > > Signed-off-by: Jason Wang
2017 Mar 30
1
[PATCH net-next 8/8] vhost_net: use lockless peeking for skb array during busy polling
On Thu, Mar 30, 2017 at 10:16:15AM +0800, Jason Wang wrote: > > > On 2017?03?29? 20:07, Michael S. Tsirkin wrote: > > On Tue, Mar 21, 2017 at 12:04:47PM +0800, Jason Wang wrote: > > > For the socket that exports its skb array, we can use lockless polling > > > to avoid touching spinlock during busy polling. > > > > > > Signed-off-by: Jason Wang
2017 Mar 29
2
[PATCH net-next 8/8] vhost_net: use lockless peeking for skb array during busy polling
On Tue, Mar 21, 2017 at 12:04:47PM +0800, Jason Wang wrote: > For the socket that exports its skb array, we can use lockless polling > to avoid touching spinlock during busy polling. > > Signed-off-by: Jason Wang <jasowang at redhat.com> > --- > drivers/vhost/net.c | 7 +++++-- > 1 file changed, 5 insertions(+), 2 deletions(-) > > diff --git
2017 Mar 29
2
[PATCH net-next 8/8] vhost_net: use lockless peeking for skb array during busy polling
On Tue, Mar 21, 2017 at 12:04:47PM +0800, Jason Wang wrote: > For the socket that exports its skb array, we can use lockless polling > to avoid touching spinlock during busy polling. > > Signed-off-by: Jason Wang <jasowang at redhat.com> > --- > drivers/vhost/net.c | 7 +++++-- > 1 file changed, 5 insertions(+), 2 deletions(-) > > diff --git
2017 Mar 21
0
[PATCH net-next 8/8] vhost_net: use lockless peeking for skb array during busy polling
For the socket that exports its skb array, we can use lockless polling to avoid touching spinlock during busy polling. Signed-off-by: Jason Wang <jasowang at redhat.com> --- drivers/vhost/net.c | 7 +++++-- 1 file changed, 5 insertions(+), 2 deletions(-) diff --git a/drivers/vhost/net.c b/drivers/vhost/net.c index 53f09f2..41153a3 100644 --- a/drivers/vhost/net.c +++ b/drivers/vhost/net.c
2017 Mar 30
0
[PATCH net-next 8/8] vhost_net: use lockless peeking for skb array during busy polling
On 2017?03?29? 20:07, Michael S. Tsirkin wrote: > On Tue, Mar 21, 2017 at 12:04:47PM +0800, Jason Wang wrote: >> For the socket that exports its skb array, we can use lockless polling >> to avoid touching spinlock during busy polling. >> >> Signed-off-by: Jason Wang <jasowang at redhat.com> >> --- >> drivers/vhost/net.c | 7 +++++-- >> 1 file
2007 Apr 07
8
How to get rid of locks
Although Dovecot is already read-lockless and it uses only short- lived write locks, it's be really nice to just get rid of the locking completely. :) I just figured out that O_APPEND is pretty great. If the operating system updates seek position after writing to a file opened with O_APPEND, writes to Dovecot's transaction log file can be made lockless. I see that this works with
2016 Apr 26
0
[PATCH 2/2] vhost: lockless enqueuing
We use spinlock to synchronize the work list now which may cause unnecessary contentions. So this patch switch to use llist to remove this contention. Pktgen tests shows about 5% improvement: Before: ~1300000 pps After: ~1370000 pps Signed-off-by: Jason Wang <jasowang at redhat.com> --- drivers/vhost/vhost.c | 52 +++++++++++++++++++++++++-------------------------- drivers/vhost/vhost.h |
2016 Apr 26
0
[PATCH 2/2] vhost: lockless enqueuing
On 04/26/2016 02:24 PM, Pankaj Gupta wrote: > Hi Jason, > > Overall patches look good. Just one doubt I have is below: >> We use spinlock to synchronize the work list now which may cause >> unnecessary contentions. So this patch switch to use llist to remove >> this contention. Pktgen tests shows about 5% improvement: >> >> Before: >> ~1300000 pps
2016 Apr 26
2
[PATCH 2/2] vhost: lockless enqueuing
Hi Jason, Overall patches look good. Just one doubt I have is below: > > We use spinlock to synchronize the work list now which may cause > unnecessary contentions. So this patch switch to use llist to remove > this contention. Pktgen tests shows about 5% improvement: > > Before: > ~1300000 pps > After: > ~1370000 pps > > Signed-off-by: Jason Wang <jasowang
2016 Apr 26
2
[PATCH 2/2] vhost: lockless enqueuing
Hi Jason, Overall patches look good. Just one doubt I have is below: > > We use spinlock to synchronize the work list now which may cause > unnecessary contentions. So this patch switch to use llist to remove > this contention. Pktgen tests shows about 5% improvement: > > Before: > ~1300000 pps > After: > ~1370000 pps > > Signed-off-by: Jason Wang <jasowang
2020 Jun 27
0
C7, mdadm generating selinux warnings
Hi! I'm getting these frequently. several times I've done the steps listed there to suppress the messages, but I keep getting them. Anyone got a good idea how to deal with this? Thanks in advance! ---------------------------- SELinux is preventing mdadm from 'read, open' accesses on the file /var/log/rear/rear-fcshome.log.lockless. ***** Plugin restorecon (99.5 confidence)
2003 Aug 24
3
0.99.11-test6
http://dovecot.procontrol.fi/test/ I think I've fixed the problems that were corrupting indexes with the new code. It happened every time you had expunged some messages and new messages arrived. I think this release might actually work :) disable_plaintext_auth = yes will be default from now on. It allows plaintext authentication from localhost (127.*, ::1) however. I just figured out
2013 Apr 23
1
[PATCH virtio-next] caif_virtio: Remove bouncing email addresses
Remove our (soon to be) bouncing email addresses, and update Dmitri's address. Dmitry will take over as maintainer for CAIF from now on. Cc: Vikram Arv <vikram.arv at stericsson.com> Cc: Dmitry Tarnyagin <dmitry.tarnyagin at stericsson.com> Cc: Dmitry Tarnyagin <dmitry.tarnyagin at lockless.no> Signed-off-by: Sjur Br?ndeland <sjur.brandeland at stericsson.com> ---
2013 Apr 23
1
[PATCH virtio-next] caif_virtio: Remove bouncing email addresses
Remove our (soon to be) bouncing email addresses, and update Dmitri's address. Dmitry will take over as maintainer for CAIF from now on. Cc: Vikram Arv <vikram.arv at stericsson.com> Cc: Dmitry Tarnyagin <dmitry.tarnyagin at stericsson.com> Cc: Dmitry Tarnyagin <dmitry.tarnyagin at lockless.no> Signed-off-by: Sjur Br?ndeland <sjur.brandeland at stericsson.com> ---
2019 Feb 25
4
Making LLD PDB generation faster
How do you compile LLD? There's a big difference between when using MSVC vs Clang. The parallel ghash patch I was mentioning is almost 2x as fast when using Clang 7.0+ vs. MSVC 15.9+, I don't know exactly why. I also suggest you use the Release target. You should also grab this patch: https://reviews.llvm.org/D55056 - I had to revert it because it was causing issues with LLDB. But it
2019 Jun 20
2
[PATCH v4 02/12] drm/virtio: switch virtio_gpu_wait_ioctl() to gem helper.
Use drm_gem_reservation_object_wait() in virtio_gpu_wait_ioctl(). This also makes the ioctl run lockless. v2: use reservation_object_test_signaled_rcu for VIRTGPU_WAIT_NOWAIT. Signed-off-by: Gerd Hoffmann <kraxel at redhat.com> Reviewed-by: Daniel Vetter <daniel.vetter at ffwll.ch> --- drivers/gpu/drm/virtio/virtgpu_ioctl.c | 24 ++++++++++-------------- 1 file changed, 10
2019 Jun 20
2
[PATCH v4 02/12] drm/virtio: switch virtio_gpu_wait_ioctl() to gem helper.
Use drm_gem_reservation_object_wait() in virtio_gpu_wait_ioctl(). This also makes the ioctl run lockless. v2: use reservation_object_test_signaled_rcu for VIRTGPU_WAIT_NOWAIT. Signed-off-by: Gerd Hoffmann <kraxel at redhat.com> Reviewed-by: Daniel Vetter <daniel.vetter at ffwll.ch> --- drivers/gpu/drm/virtio/virtgpu_ioctl.c | 24 ++++++++++-------------- 1 file changed, 10
2019 Feb 24
2
Making LLD PDB generation faster
Leonardo, to answer to your questions, yes to all of them  You can take a look at this prototype/proposal: https://reviews.llvm.org/D55585 Overall, computing ghashes in parallel at link-time and merging Types with them is less costly that the current approach to merging. The 35sec you’re seeing for merging should go down to about 15sec. The patch doesn’t parallelize (yet) the Type merging