search for: futexes

Displaying 20 results from an estimated 269 matches for "futexes".

Did you mean: mutexes
2007 Nov 05
1
xend crash at startup
xend crashes at startup with following stack trace: ... File "/usr/lib64/python/xen/xend/XendPIF.py", line 68, in linux_get_phy_ifaces phy_ifaces.append((name, int(mtu), bridged_mac)) UnboundLocalError: local variable ''bridged_mac'' referenced before assignment The simple patch attached below prevent the crash, and allow xend to start normally. It seems the crash
2012 Jun 25
6
puppet freezes on FUTEX_WAKE_PRIVATE
Hi, i just want to ask whether somebody else has this problem of mine and if it can be solved. Many of my ( 30+ ) puppet installation freeze up after some time. The process is waiting for some private futex, but it stays like that forever. This is what strace looks like when the problem occures: [pid 29173] futex(0x3d35ce7a84, FUTEX_WAIT_BITSET_PRIVATE|FUTEX_CLOCK_REALTIME, 43406739,
2019 Jan 02
3
[lit] check-all hanging
Hi, >From time to time, I see check-all hang during running of lit tests. The hang always happens at the > 90% completion stage and I'm pretty sure all tests have been run and check-all is just waiting for lit/python to exit. I see a single python processing running, taking very little CPU time. An strace of that process shows this: select(0, NULL, NULL, NULL, {0, 50000}) = 0
2019 May 07
4
[PATCH RFC] vhost: don't use kmap() to log dirty pages
Vhost log dirty pages directly to a userspace bitmap through GUP and kmap_atomic() since kernel doesn't have a set_bit_to_user() helper. This will cause issues for the arch that has virtually tagged caches. The way to fix is to keep using userspace virtual address. Fortunately, futex has a cmpxchg to userspace memory helper futex_atomic_cmpxchg_inatomic(). So switch to use it to exchange the
2019 May 07
4
[PATCH RFC] vhost: don't use kmap() to log dirty pages
Vhost log dirty pages directly to a userspace bitmap through GUP and kmap_atomic() since kernel doesn't have a set_bit_to_user() helper. This will cause issues for the arch that has virtually tagged caches. The way to fix is to keep using userspace virtual address. Fortunately, futex has a cmpxchg to userspace memory helper futex_atomic_cmpxchg_inatomic(). So switch to use it to exchange the
2019 Jan 02
4
[lit] check-all hanging
Hi David, Chandler, I see lldb tests hang often, and then I kill the dotest process. I'd like to stop running check-all too, but I feel it's important when I modify FileCheck. The flakiness that Chandler mentioned makes it time-consuming to verify test results. Joel On Wed, Jan 2, 2019 at 4:41 PM Chandler Carruth via llvm-dev < llvm-dev at lists.llvm.org> wrote: > What
2023 Jun 06
2
virsh not connecting to libvertd ?
I have identical two hypervisors same operating system: Ubuntu 22.04.2 LTS Recently both virsh stopped talking to the libvirtd. Both stopped within a few days of each other. Currently if I run: virsh uri virsh version virsh list # virsh list ..nothing just hangs When I ran strace on these broken machines it get stuck at same spot: strace virsh list ...
2011 Sep 27
2
Slow DNS lookups on CentOS6
Most of my desktops are still running CentOS5, but I have installed CentOS6 on a few of them. The users on those desktops are reporting that DNS lookups are slow, and from my brief tests, that does appear to be the case. After some googling, I found a suggestion to disable IPv6, but that didn't help. So I tried to figure out where the delay occurs using strace and ltrace, but that
2009 May 21
0
1.4.24.1 -> 1.6.0.9: segfault
I'm testing an upgrade of an i686 production machine running 1.4.24.1 to 1.6.0.9. I've installed dahdi-linux-2.1.0.4. But: asterisk -cvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvv Asterisk 1.6.0.9, Copyright (C) 1999 - 2008 Digium, Inc. and others. Created by Mark Spencer <markster at digium.com> Asterisk comes with ABSOLUTELY NO WARRANTY; type 'core show warranty' for details.
2006 Sep 05
0
futex hang workaround by rebooting first (?)
The previously noted upgrade-to-4.4 recipe: yum update yum sqlite python-sqlite ; yum update worked more consistently than running "yum upgrade," but it too fell victim to the futex hang. In my limited testing, a reboot of the machine before starting the upgrade gave me a 100% success rate. Given two identical machines, one that had been used for a few days and one that was freshly
2013 Nov 06
3
Calling 'virsh' from within a hook script - avoid deadlock?
I'm trying to create a hook that attaches several virtio-scsi disks to a starting vm, and it seems like the hook enters a deadlock while attempting to do so. Is there any workaround around this? Any better way, forking virsh execution somehow? Could anyone explain why this is creating a deadlock in the first place? Isn't libvirt multithreaded? Thanks for the help! Shlomi ### hook
2019 May 07
0
[PATCH RFC] vhost: don't use kmap() to log dirty pages
On Mon, May 06, 2019 at 10:23:29PM -0400, Jason Wang wrote: > Note: there're archs (few non popular ones) that don't implement > futex helper, we can't log dirty pages. We can fix them on top or > simply disable LOG_ALL features of vhost. That means vhost now has to depend on HAVE_FUTEX_CMPXCHG to make sure we have a working implementation. > #include
2006 Apr 28
0
Processes stuck in futex()
I'm occasionally seeing processes on my CentOS 4.3 system get hung up, which strace reveals to be waiting on a call to futex(). Previously I'd only ever seen this in mysql on RH9, and google searches for explanations haven't turned up anything useful. Any pointers to where I should look? Is this another one of those things where using the default SMP kernel with a hyperthreaded
2012 Aug 10
1
problems with bind update of 9.8.2.0.10
Everyone, I have two x86_64 servers with CentOS 6.3 that seem to have some difficulty with the last version of bind : 32:bind-9.8.2-0.10.rc1.el6_3.2.x86_64 When I run htop I am getting a CPU usage of 70 to 80% all of the time. I am not seeing problems identified in the log files on either machine. This may have been present before the update of 9.8.2.0, but I certainly did not notice it. Has
2019 May 13
5
[PATCH net] vhost: don't use kmap() to log dirty pages
Vhost log dirty pages directly to a userspace bitmap through GUP and kmap_atomic() since kernel doesn't have a set_bit_to_user() helper. This will cause issues for the arch that has virtually tagged caches. The way to fix is to keep using userspace virtual address. Fortunately, futex has arch_futex_atomic_op_inuser() which could be used for setting a bit to user. Note there're several
2019 May 13
5
[PATCH net] vhost: don't use kmap() to log dirty pages
Vhost log dirty pages directly to a userspace bitmap through GUP and kmap_atomic() since kernel doesn't have a set_bit_to_user() helper. This will cause issues for the arch that has virtually tagged caches. The way to fix is to keep using userspace virtual address. Fortunately, futex has arch_futex_atomic_op_inuser() which could be used for setting a bit to user. Note there're several
2006 Jun 21
0
[LLVMdev] size of generated machine code ?
On Tue, 20 Jun 2006 22:38:24 -0500 (CDT) Chris Lattner <sabre at nondot.org> wrote: > > >> Is this really the number of bytes of machine code ? > > Yes. > > To be specific, this is the number of bytes of machine code JIT'd. Thus, > this only includes the stuff reachable from main, for example. To see > specifically what functions it is compiling and
2006 Jun 21
3
[LLVMdev] size of generated machine code ?
On Tue, 20 Jun 2006, Chris Lattner wrote: > On Wed, 21 Jun 2006, Simon Burton wrote: >> This from the output: >> >> 24620 x86-emitter - Number of machine instructions emitted >> >> (i had to write a dummy main function to get this to work) >> >> Is this really the number of bytes of machine code ? > Yes. To be specific, this is the
2010 Dec 13
3
Slow I/O on ocfs2 file system
Hello, I have found, that ocfs2 is very slow when doing I/O operation without cache. See a simple test: ng-vvv1:~# dd if=/data/verejna/dd-1G bs=1k | dd of=/dev/null 1048576+0 records in 1048576+0 records out 1073741824 bytes (1.1 GB) copied, 395.183 s, 2.7 MB/s 2097152+0 records in 2097152+0 records out 1073741824 bytes (1.1 GB) copied, 395.184 s, 2.7 MB/s The underlying block device is quite
2008 Jan 10
4
yum hangs when Reading Local RPMDB
All of a sudden yum hangs on a Centos 4.5 (updated to the latest patch before 4.6) when I try to use it. "clean metadata" didn't help. The output of -d5 shows it gets to the point of "Reading Local RPMDB" and then sits there. The process doesn't die, but doesn't seem to do anything from that point on. strace shows this: futex(0xb76dcae8, FUTEX_WAIT, 2, NULL I