similar to: [Bug 93458] New: page allocation failure: order:5, mode:0x240c0c0

Displaying 20 results from an estimated 1000 matches similar to: "[Bug 93458] New: page allocation failure: order:5, mode:0x240c0c0"

2017 Sep 02
2
Xen CentOS 7.3 server + CentOS 7.3 VM fails to boot after CR updates (applied to VM)!
On 09/01/2017 02:41 PM, Kevin Stange wrote: > On 08/31/2017 07:50 AM, PJ Welsh wrote: >> A recently created and fully functional CentOS 7.3 VM fails to boot >> after applying CR updates: > <snip> >> Server OS is CentOS 7.3 using Xen (no CR updates): >> rpm -qa xen\* >> xen-hypervisor-4.6.3-15.el7.x86_64 >> xen-4.6.3-15.el7.x86_64 >>
2017 Sep 04
2
Xen CentOS 7.3 server + CentOS 7.3 VM fails to boot after CR updates (applied to VM)!
On 09/04/2017 03:59 PM, Kevin Stange wrote: > On 09/02/2017 08:11 AM, Johnny Hughes wrote: >> On 09/01/2017 02:41 PM, Kevin Stange wrote: >>> On 08/31/2017 07:50 AM, PJ Welsh wrote: >>>> A recently created and fully functional CentOS 7.3 VM fails to boot >>>> after applying CR updates: >>> <snip> >>>> Server OS is CentOS 7.3
2017 Aug 31
5
Xen CentOS 7.3 server + CentOS 7.3 VM fails to boot after CR updates (applied to VM)!
A recently created and fully functional CentOS 7.3 VM fails to boot after applying CR updates: ... ;-1f1;-1f CentOS Linux 7 (Core)1;-1f1;-1f CentOS Linux 7 (Core)1;-1f1;-1f CentOS Linux 7 (Core)1;-1f1;-1f CentOS Linux 7 (Core)1;-1f1;-1f CentOS Linux 7 (Core)1;-1f1;-1f CentOS Linux 7 (Core)1;-1f1;-1f CentOS Linux 7 (Core)1;-1f1;-1f CentOS Linux 7 (Core)1;-1f1;-1f CentOS Linux 7 (Core)1;-1f1;-1f
2017 Oct 23
1
problems running a vol over IPoIB, and qemu off it?
hi people I wonder if anybody experience any problems with vols in replica mode that run across IPoIB links and libvirt stores qcow image on such a volume? I wonder if maybe devel could confirm it should just work, and then hardware/Infiniband I should blame. I have a direct IPoIB link between two hosts, gluster replica volume, libvirt store disk images there. I start a guest on hostA and
2019 Jul 12
2
Out of memory: kill process
Hello, On my bridge head DC, i can see in kern.log lots of < Out of memory : kill process > I'm using Samba 4.9.6 (11147 objects) on Debian Stretch 64 This DC synchronize from/to 20 others DC in bridge head mode (no mesh) Here is my VM (HyperV 2016) config : - 4 x Vcpu (intel Xeon Silver 4110) - 2 Go Ram - 1 Go swap I'm really not an expert on this
2017 Oct 18
2
[PATCH] virtio: avoid possible OOM lockup at virtballoon_oom_notify()
Tetsuo Handa wrote: > 20171016-deflate.log.xz continued printing "puff" messages without any OOM > killer messages, for fill_balloon() always inflates faster than leak_balloon() > deflates. > > Since the OOM killer cannot be invoked unless leak_balloon() completely > deflates faster than fill_balloon() inflates, the guest remained unusable > (e.g. unable to login
2017 Oct 18
2
[PATCH] virtio: avoid possible OOM lockup at virtballoon_oom_notify()
Tetsuo Handa wrote: > 20171016-deflate.log.xz continued printing "puff" messages without any OOM > killer messages, for fill_balloon() always inflates faster than leak_balloon() > deflates. > > Since the OOM killer cannot be invoked unless leak_balloon() completely > deflates faster than fill_balloon() inflates, the guest remained unusable > (e.g. unable to login
2019 Dec 04
5
[PATCH] virtio-balloon: fix managed page counts when migrating pages between zones
In case we have to migrate a ballon page to a newpage of another zone, the managed page count of both zones is wrong. Paired with memory offlining (which will adjust the managed page count), we can trigger kernel crashes and all kinds of different symptoms. One way to reproduce: 1. Start a QEMU guest with 4GB, no NUMA 2. Hotplug a 1GB DIMM and only the memory to ZONE_NORMAL 3. Inflate the balloon
2019 Dec 04
5
[PATCH] virtio-balloon: fix managed page counts when migrating pages between zones
In case we have to migrate a ballon page to a newpage of another zone, the managed page count of both zones is wrong. Paired with memory offlining (which will adjust the managed page count), we can trigger kernel crashes and all kinds of different symptoms. One way to reproduce: 1. Start a QEMU guest with 4GB, no NUMA 2. Hotplug a 1GB DIMM and only the memory to ZONE_NORMAL 3. Inflate the balloon
2013 Sep 23
6
btrfs: qgroup scan failed with -12
Not sure if it''s anything interesting - I had the following entry in dmesg a few days ago, on a server with 32 GB RAM. The system is still working fine. [1878432.675210] btrfs-qgroup-re: page allocation failure: order:5, mode:0x104050 [1878432.675319] CPU: 5 PID: 22251 Comm: btrfs-qgroup-re Not tainted 3.11.0-rc7 #2 [1878432.675417] Hardware name: System manufacturer System Product
2019 Mar 06
2
director in rings
Hi Maby stupid question :) It possible to have 3 directors (frontend) but without rings ? All directors connect to this same dovecot (backend) - all backad have this same login_trusted_networks -- Maciej Mi?aszewski IQ PL Sp. z o.o. Starszy Administrator Systemowy
2019 Dec 05
2
[PATCH v2] virtio-balloon: fix managed page counts when migrating pages between zones
In case we have to migrate a ballon page to a newpage of another zone, the managed page count of both zones is wrong. Paired with memory offlining (which will adjust the managed page count), we can trigger kernel crashes and all kinds of different symptoms. One way to reproduce: 1. Start a QEMU guest with 4GB, no NUMA 2. Hotplug a 1GB DIMM and only the memory to ZONE_NORMAL 3. Inflate the balloon
2019 Dec 05
2
[PATCH v2] virtio-balloon: fix managed page counts when migrating pages between zones
In case we have to migrate a ballon page to a newpage of another zone, the managed page count of both zones is wrong. Paired with memory offlining (which will adjust the managed page count), we can trigger kernel crashes and all kinds of different symptoms. One way to reproduce: 1. Start a QEMU guest with 4GB, no NUMA 2. Hotplug a 1GB DIMM and only the memory to ZONE_NORMAL 3. Inflate the balloon
2011 May 13
3
XEN Dom 0 boot and Out of Memory issue
Hi all, Recently I build XEN 4.1 (from the xen.org) on centIOS 5.5 (kernel 2.6.38.4). After changing the grub file and rebooting the machine, I am continuously getting out of memory error. The content of my grub file: default=0 timeout=5 serial --unit=0 --speed=115200 terminal --timeout=10 console serial splashimage=(hd0,0)/grub/splash.xpm.gz hiddenmenu title CentOS (2.6.38.4) root
2013 Mar 05
4
[RFC PATCH] drm/nouveau: use vmalloc for pgt allocation
Page tables on nv50 take 48kB, which can be hard to allocate in one piece. Let's use vmalloc. Signed-off-by: Marcin Slusarz <marcin.slusarz at gmail.com> Cc: stable at vger.kernel.org [3.7+] --- drivers/gpu/drm/nouveau/core/subdev/vm/base.c | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/drivers/gpu/drm/nouveau/core/subdev/vm/base.c
2012 Nov 15
3
Likely mem leak in 3.7
Starting with 3.7 rc1, my workstation seems to loose ram. Up until (and including) 3.6, used-(buffers+cached) was roughly the same as sum(rss) (taking shared into account). Now there is an approx 6G gap. When the box first starts, it is clearly less swappy than with <= 3.6; I can''t tell whether that is related. The reduced swappiness persists. It seems to get worse when I update
2013 Mar 27
2
system death under oom - 3.7.9
Hello, My system died last night apparently due to OOM conditions. Note that I don't have any swap set up, but my understanding is that this is not required. The full log is at: http://pastebin.com/YCYUXWvV. It was in my messages, so I guess the system took a bit to die completely. nouveau is somewhat implicated, as it is the first thing that hits an allocation failure in nouveau_vm_create,
2013 Dec 30
2
oom situation
I have continous oom&panic situation unresolved. I am not sure system fills up all the ram (36GB). Why this system triggered this oom situation? Is it about some other memory? highmem? lowmem? stack size? Best Regards, Kernel 3.10.24 Dec 27 09:19:05 2013 kernel: : [277622.359064] squid invoked oom-killer: gfp_mask=0x42d0, order=3, oom_score_adj=0 Dec 27 09:19:05 2013 kernel: :
2017 Oct 16
2
[PATCH] virtio: avoid possible OOM lockup at virtballoon_oom_notify()
Tetsuo Handa wrote: > Michael S. Tsirkin wrote: > > > > > > > > The proper fix isn't that hard - just avoid allocations under lock. > > > > > > > > Patch posted, pls take a look. > > > > > > Your patch allocates pages in order to inflate the balloon, but > > > your patch will allow leak_balloon() to deflate the
2017 Oct 16
2
[PATCH] virtio: avoid possible OOM lockup at virtballoon_oom_notify()
Tetsuo Handa wrote: > Michael S. Tsirkin wrote: > > > > > > > > The proper fix isn't that hard - just avoid allocations under lock. > > > > > > > > Patch posted, pls take a look. > > > > > > Your patch allocates pages in order to inflate the balloon, but > > > your patch will allow leak_balloon() to deflate the