similar to: [PATCH 0/2] vhost: improve livepatch switching for heavily loaded vhost worker kthreads

Displaying 20 results from an estimated 200 matches similar to: "[PATCH 0/2] vhost: improve livepatch switching for heavily loaded vhost worker kthreads"

2023 Jan 27
0
[PATCH 0/2] vhost: improve livepatch switching for heavily loaded vhost worker kthreads
On Thu 2023-01-26 15:12:35, Seth Forshee (DigitalOcean) wrote: > On Thu, Jan 26, 2023 at 06:03:16PM +0100, Petr Mladek wrote: > > On Fri 2023-01-20 16:12:20, Seth Forshee (DigitalOcean) wrote: > > > We've fairly regularaly seen liveptches which cannot transition within kpatch's > > > timeout period due to busy vhost worker kthreads. > > > > I have
2023 Jan 22
0
[PATCH 0/2] vhost: improve livepatch switching for heavily loaded vhost worker kthreads
On Fri, Jan 20, 2023 at 04:12:20PM -0600, Seth Forshee (DigitalOcean) wrote: > We've fairly regularaly seen liveptches which cannot transition within kpatch's > timeout period due to busy vhost worker kthreads. In looking for a solution the > only answer I found was to call klp_update_patch_state() from a safe location. > I tried adding this call to vhost_worker(), and it
2023 Jan 26
0
[PATCH 0/2] vhost: improve livepatch switching for heavily loaded vhost worker kthreads
On Fri 2023-01-20 16:12:20, Seth Forshee (DigitalOcean) wrote: > We've fairly regularaly seen liveptches which cannot transition within kpatch's > timeout period due to busy vhost worker kthreads. I have missed this detail. Miroslav told me that we have solved something similar some time ago, see https://lore.kernel.org/all/20220507174628.2086373-1-song at kernel.org/ Honestly,
2023 Jan 26
1
[PATCH 2/2] vhost: check for pending livepatches from vhost worker kthreads
On Wed 2023-01-25 10:57:30, Seth Forshee wrote: > On Wed, Jan 25, 2023 at 12:34:26PM +0100, Petr Mladek wrote: > > On Tue 2023-01-24 11:21:39, Seth Forshee wrote: > > > On Tue, Jan 24, 2023 at 03:17:43PM +0100, Petr Mladek wrote: > > > > On Fri 2023-01-20 16:12:22, Seth Forshee (DigitalOcean) wrote: > > > > > Livepatch relies on stack checking of
2019 Oct 09
2
Livepatch: Linux kernel updates without rebooting
Hi, I am running CentOS Linux release 7.7.1908 (Core). Does CentOS Linux kernel 3.10.0-1062.1.1.el7.x86_64 support kernel updates without rebooting (Live Patching)? I look forward to hearing from you and thanks in advance. Best Regards, Kaushal
2019 Oct 09
0
Livepatch: Linux kernel updates without rebooting
On Wed, Oct 9, 2019 at 11:27 AM Kaushal Shriyan <kaushalshriyan at gmail.com> wrote: > Hi, > > I am running CentOS Linux release 7.7.1908 (Core). Does CentOS Linux kernel > 3.10.0-1062.1.1.el7.x86_64 support kernel updates without rebooting (Live > Patching)? > > I look forward to hearing from you and thanks in advance. > > > this was just discussed in length
2020 Nov 03
0
[patch V3 24/37] sched: highmem: Store local kmaps in task struct
Instead of storing the map per CPU provide and use per task storage. That prepares for local kmaps which are preemptible. The context switch code is preparatory and not yet in use because kmap_atomic() runs with preemption disabled. Will be made usable in the next step. The context switch logic is safe even when an interrupt happens after clearing or before restoring the kmaps. The kmap index in
2005 May 17
2
Junk at the beginning, Warning, flexibel rate not heavily tested!
Hi,all I am newer to Asterisk.My Asterisk version is the newest CVS-HEAD.now something appears in the console CLI like below these,I don't know what's happen to my Asterisk Server.Could anybody help me? Thanks Junk at the beginning Warning, flexibel rate not heavily tested! Junk at the beginning Warning, flexibel rate not heavily tested! Warning, flexibel rate not heavily tested!
2004 Apr 04
0
Upgrading Heavily Modified RH9 Machines
Hello, I have several heavily modified RH9 machines (custom kernels, reiser FS, etc.) Just wondering if CentOS could do an upgrade install on machines such as this? I know it's always better to do a clean install, but these machines have been tweaked a lot over the last year and are running perfectly (a lot of time and effort went into this)... an upgrade install would be great. My main
2018 Jul 11
0
LMTP crashing heavily for my 2.2.36 installation
On 11 Jul 2018, at 8.41, Wolfgang Rosenauer <wrosenauer at gmail.com> wrote: > > Hi, > > I'm running 2.2.36 (as provided by openSUSE in their server:mail repository) and at least at one of my systems LMTP is crashing regularly on certain messages (apparently a lot of them). > > Sometimes (but not always a backtrace is posted to the logs: > >
2018 Jul 11
0
LMTP crashing heavily for my 2.2.36 installation
follow up question. Is there a commit which is reasonable to backport for me into the packages or is it too intrusive or based on heavily changed code? Thanks, Wolfgang On Wed, Jul 11, 2018 at 3:32 PM, Wolfgang Rosenauer <wrosenauer at gmail.com> wrote: > > On Wed, Jul 11, 2018 at 10:46 AM, Timo Sirainen <tss at iki.fi> wrote: > >> On 11 Jul 2018, at 8.41, Wolfgang
2018 Jul 12
0
LMTP crashing heavily for my 2.2.36 installation
On Wed, Jul 11, 2018 at 6:03 PM, Aki Tuomi <aki.tuomi at dovecot.fi> wrote: > One alternative is to migrate into sdbox format, in which this is supported > > > > --- > Aki Tuomi > Dovecot oy > > -------- Original message -------- > From: Wolfgang Rosenauer <wrosenauer at gmail.com> > Date: 11/07/2018 18:14 (GMT+02:00) > To: Timo Sirainen <tss at
2018 Jul 12
0
LMTP crashing heavily for my 2.2.36 installation (and now with 2.3.2.1)
Hi, I will try to create a coredump later but now I see version 2.3.2.1 also crashing in LMTP :-( 2018-07-12T10:09:57.336062+02:00 saruman dovecot: lmtp(an007498)<11814><zrPDEdUMR1smLgAAQ/KzDw>: Fatal: master: service(lmtp): child 11814 killed with signal 6 (core dumps disabled - https://dovecot.org/bugreport.html#coredumps) 2018-07-12T10:09:57.382925+02:00 saruman dovecot:
2019 May 05
3
CentOS 7 Xen 4.12 libvirt/virt-manager wrong path for qemu-system-i386
Hello, While testing Virt-SIG Xen 4.12 rpms on CentOS7 I noticed the following problem with libvirt/virt-manager when manually installing a new HVM guest from virt-manager GUI.. basicly the VM installation won't start, because libvirt/virt-manager is not able to start the VM, due to "missing" qemu-system-i386 binary: Unable to complete install: 'unsupported configuration:
2017 Jul 31
0
[PATCH net] virtio_net: fix truesize for mergeable buffers
Seth Forshee noticed a performance degradation with some workloads. This turns out to be due to packet drops. Euan Kemp noticed that this is because we drop all packets where length exceeds the truesize, but for some packets we add in extra memory without updating the truesize. This in turn was kept around unchanged from ab7db91705e95 ("virtio-net: auto-tune mergeable rx buffer size for
2017 Jul 31
0
[PATCH net] virtio_net: fix truesize for mergeable buffers
Seth Forshee noticed a performance degradation with some workloads. This turns out to be due to packet drops. Euan Kemp noticed that this is because we drop all packets where length exceeds the truesize, but for some packets we add in extra memory without updating the truesize. This in turn was kept around unchanged from ab7db91705e95 ("virtio-net: auto-tune mergeable rx buffer size for
1998 Oct 15
3
Performance tips for heavily loaded servers
Does nayone have a collection of performance tips for heavily loaded samba servers? I have a server that routinely has 45-90 connections and just runs really poorly. We've thrown more memory at it w/ some benefit, but not significant. This is on a linux 2.0.35 based machine. Does anyone have any suggestions? -- Nathan ------------------------------------------------------------ Nathan
2006 Mar 12
0
ERB and Builder template engines shouldn''t be so heavily wired in ActionView::Base
Problem: 1. ERB and Builder template engines are heavily wired in ActionView::Base 2. Should be optional. If we use different engine - why load all their stuff if not needed? 3. There are template engines which expect files in their own locations and cache them in their own way. For instance a template engine may store templates in a database. These engines don''t need any central source
2004 Mar 03
4
heavily fragmented file system.. How to defrag it on-line??
Hi, all, I got machines running continuously for long time, but then the underlying ext3 file systems becomes quite heavily fragmented (94% non-contiguous). We just don't have a chance to shutdown the machines since they are always busy.. I tried the defrag 0.70 version comes with e2fsprog package and standalone 0.73 packages, but neither help me since the defrag tool can not handle ext3.
2006 Jan 24
3
Merging two heavily referenced records into one
Here''s the situation: I have a "people" table whose "id" column serves as a foreign key to several other tables (authorships, editorships, members, users, etc.). Every now and then I will find two distinct records in the "people" table that represent the same human being. They generally differ by small differences in spelling but should really be the