search for: wakening

Displaying 20 results from an estimated 39 matches for "wakening".

Did you mean: weakening
2008 Mar 31
4
[04/17] [PATCH] Add kvm arch-specific core code for kvm/ia64.-V8
Zhang, Xiantao wrote: >>From 62895ff991d48398a77afdbf7f2bef127e802230 Mon Sep 17 00:00:00 2001 > From: Xiantao Zhang <xiantao.zhang at intel.com> > Date: Fri, 28 Mar 2008 09:49:57 +0800 > Subject: [PATCH] KVM: IA64: Add kvm arch-specific core code for > kvm/ia64. > > kvm_ia64.c is created to handle kvm ia64-specific core logic. > Signed-off-by: Xiantao Zhang
2008 Mar 31
4
[04/17] [PATCH] Add kvm arch-specific core code for kvm/ia64.-V8
Zhang, Xiantao wrote: >>From 62895ff991d48398a77afdbf7f2bef127e802230 Mon Sep 17 00:00:00 2001 > From: Xiantao Zhang <xiantao.zhang at intel.com> > Date: Fri, 28 Mar 2008 09:49:57 +0800 > Subject: [PATCH] KVM: IA64: Add kvm arch-specific core code for > kvm/ia64. > > kvm_ia64.c is created to handle kvm ia64-specific core logic. > Signed-off-by: Xiantao Zhang
2015 Apr 09
2
[PATCH v15 13/15] pvqspinlock: Only kick CPU at unlock time
On Mon, Apr 06, 2015 at 10:55:48PM -0400, Waiman Long wrote: > @@ -219,24 +236,30 @@ static void pv_wait_node(struct mcs_spinlock *node) > } > > /* > + * Called after setting next->locked = 1 & lock acquired. > + * Check if the the CPU has been halted. If so, set the _Q_SLOW_VAL flag > + * and put an entry into the lock hash table to be waken up at unlock time.
2015 Apr 09
2
[PATCH v15 13/15] pvqspinlock: Only kick CPU at unlock time
On Mon, Apr 06, 2015 at 10:55:48PM -0400, Waiman Long wrote: > @@ -219,24 +236,30 @@ static void pv_wait_node(struct mcs_spinlock *node) > } > > /* > + * Called after setting next->locked = 1 & lock acquired. > + * Check if the the CPU has been halted. If so, set the _Q_SLOW_VAL flag > + * and put an entry into the lock hash table to be waken up at unlock time.
2014 Mar 06
0
backup-dir over nfs
Hi All. Since it's a lengthy one to explain, I'll jump right into my issue: I run rsync on 3 debian systems for the purpose of backing up approximately 500GBs of various types of data. The general strategy is two systems hold full backups and the third holds incremental's using date-based --backup --backup-dir=. The directory in question is an NFS mount. Here's the problem:
2016 Jan 01
2
[PATCH RFC] virtio_balloon: refill by config handler
Peter Mladek reported that balloon might never refill completely after restore. This is because fill_balloon is only called once there. Calling fill_balloon repeatedly seems too aggressive, especially in light of the fact that it sleeps on failure: let's wake the config change handler and fill it asynchronously. Reported-by: Petr Mladek <pmladek at suse.com> Signed-off-by: Michael S.
2016 Jan 01
2
[PATCH RFC] virtio_balloon: refill by config handler
Peter Mladek reported that balloon might never refill completely after restore. This is because fill_balloon is only called once there. Calling fill_balloon repeatedly seems too aggressive, especially in light of the fact that it sleeps on failure: let's wake the config change handler and fill it asynchronously. Reported-by: Petr Mladek <pmladek at suse.com> Signed-off-by: Michael S.
2016 Jan 05
0
[PATCH RFC] virtio_balloon: refill by config handler
On Fri 2016-01-01 12:13:48, Michael S. Tsirkin wrote: > Peter Mladek reported that balloon might never refill completely after > restore. This is because fill_balloon is only called once there. > Calling fill_balloon repeatedly seems too aggressive, > especially in light of the fact that it sleeps on failure: let's > wake the config change handler and fill it asynchronously.
2006 Apr 21
1
RE: [PATCH]Check the values of MAX_VIRT_CPUS and NR_CPUSfor SMP
>From: Tristan Gingold [mailto:Tristan.Gingold@bull.net] >Sent: 2006年4月21日 15:26 >> > >> >See how we solve this on x86 near the top of do_boot_cpu() in >> >arch/x86/smpboot.c. Essentially we allocate extra idle domains on >> >demand to give us enough idle vcpus for one per physical cpu. >> > >> > -- Keir >> >> Yes,
2001 Oct 24
1
Softtware raid for root?
I had been running a system with / as a raid0 array under Redhat 7.1 and the original smp kernel. Is it advisable to use ext-3 and RedHat 7.2 for a software raid0 root filesystem? I understand 7.2 uses a 2.4.7 kernel. I would like to use the smp kernel as this is a 2 proc. box. Stuart __________________________________________________________ I fear that all we have done is to waken a sleeping
2005 Aug 29
0
bcmxcp in development tag.
Hi. FYI. Have tested the bcmxcp driver from cvs development tag. And it work as expected on the 5115 with the serial interface. So to Wolfgang, no bug introduced in the serial module when you split'ed the driver. If i get some time to spare, i have a look on the serial module to make it wake up the ups. The 9125 is type pnp so it need to be waken up. Then there are some cleanup to be done
2017 Dec 12
0
[PATCH v19 2/7] xbitmap: potential improvement
This patch made some changes to the original xbitmap implementation from the linux-dax tree: - remove xb_fill() and xb_zero() from xbitmap.h since they are not implemented; - xb_test_bit: changed "ebit > BITS_PER_LONG" to "ebit >= BITS_PER_LONG", because bit 64 beyonds the "unsigned long" exceptional entry (0 to 63); - xb_set_bit: delete the new inserted
2015 Mar 26
2
[PATCH 0/9] qspinlock stuff -v15
On Wed, Mar 25, 2015 at 03:47:39PM -0400, Konrad Rzeszutek Wilk wrote: > Ah nice. That could be spun out as a seperate patch to optimize the existing > ticket locks I presume. Yes I suppose we can do something similar for the ticket and patch in the right increment. We'd need to restructure the code a bit, but its not fundamentally impossible. We could equally apply the head hashing to
2015 Mar 26
2
[PATCH 0/9] qspinlock stuff -v15
On Wed, Mar 25, 2015 at 03:47:39PM -0400, Konrad Rzeszutek Wilk wrote: > Ah nice. That could be spun out as a seperate patch to optimize the existing > ticket locks I presume. Yes I suppose we can do something similar for the ticket and patch in the right increment. We'd need to restructure the code a bit, but its not fundamentally impossible. We could equally apply the head hashing to
2010 Oct 26
8
[Bug 31139] New: Wake from sleep doesn't turn on display(s)
https://bugs.freedesktop.org/show_bug.cgi?id=31139 Summary: Wake from sleep doesn't turn on display(s) Product: xorg Version: git Platform: x86-64 (AMD64) OS/Version: Linux (All) Status: NEW Severity: normal Priority: medium Component: Driver/nouveau AssignedTo: nouveau at lists.freedesktop.org
2004 Aug 06
3
Is Speex realy patent free?
Hi all, <p>i have readed all the realy sad stuff about software patents in europe and what allread exists. There is one side related to speex and i dont know if this is allready discussed here. Take a look at: http://swpat.ffii.org/patents/effects/voip/index.en.html Is speex realy petant free or does patent free only means that is is not patent by the speex authors? Who have checked that
2015 Apr 07
0
[PATCH v15 13/15] pvqspinlock: Only kick CPU at unlock time
Before this patch, a CPU may have been kicked twice before getting the lock - one before it becomes queue head and once before it gets the lock. All these CPU kicking and halting (VMEXIT) can be expensive and slow down system performance, especially in an overcommitted guest. This patch add a new vCPU state (vcpu_hashed) which enables the code to delay CPU kicking until at unlock time. Once this
2016 Jan 25
2
[PATCH v5 0/2] virtio_balloon: Conversion to workqueue + parallel stats
The previous version did not prevent updating the balloon when it was being removed. There was an idea to update the stats in a separate work and cancel it after the config reset. But this does not seem to work. The code for updating the stats communicates with the host. IMHO, it should not run after the reset (flushing buffers). Therefore I decided to use a boolean value and a spin lock (inspired
2016 Jan 25
2
[PATCH v5 0/2] virtio_balloon: Conversion to workqueue + parallel stats
The previous version did not prevent updating the balloon when it was being removed. There was an idea to update the stats in a separate work and cancel it after the config reset. But this does not seem to work. The code for updating the stats communicates with the host. IMHO, it should not run after the reset (flushing buffers). Therefore I decided to use a boolean value and a spin lock (inspired
2014 May 30
0
[PATCH v11 14/16] pvqspinlock: Add qspinlock para-virtualization support
This patch adds base para-virtualization support to the queue spinlock in the same way as was done in the PV ticket lock code. In essence, the lock waiters will spin for a specified number of times (QSPIN_THRESHOLD = 2^14) and then halted itself. The queue head waiter, unlike the other waiter, will spins 2*QSPIN_THRESHOLD times before halting itself. Before being halted, the queue head waiter