similar to: Linux to Windows

Displaying 20 results from an estimated 500 matches similar to: "Linux to Windows"

2012 Mar 06
1
Correctly using libvirt to mount /dev/pts and /dev in an lxc application container
We would like to mount /dev and /dev/pts correctly using the libvirt xml config file instead of doing a bind mount in a container init script we use. Currently the container config is: <domain type='lxc'> <name>CentOS_57</name> <uuid>ff5d3c04-49e6-a3cc-0a14-ff13625eca3c</uuid> <memory>262144</memory>
2004 Nov 24
1
what are "false_alarms"?
Hi all. I'm sorry if this is a simple, obvious question, but I've looked around, and can not find an answer to this. What exactly are the false_alarms in the rsync summary? I am getting a number of false alarms in several of my rsync processes. Here's my output from one of my logfiles: total: matches=9724 tag_hits=741242 false_alarms=36 data=20850152 wrote 11731892 bytes read
2011 May 13
3
XEN Dom 0 boot and Out of Memory issue
Hi all, Recently I build XEN 4.1 (from the xen.org) on centIOS 5.5 (kernel 2.6.38.4). After changing the grub file and rebooting the machine, I am continuously getting out of memory error. The content of my grub file: default=0 timeout=5 serial --unit=0 --speed=115200 terminal --timeout=10 console serial splashimage=(hd0,0)/grub/splash.xpm.gz hiddenmenu title CentOS (2.6.38.4) root
2017 Sep 02
2
Xen CentOS 7.3 server + CentOS 7.3 VM fails to boot after CR updates (applied to VM)!
On 09/01/2017 02:41 PM, Kevin Stange wrote: > On 08/31/2017 07:50 AM, PJ Welsh wrote: >> A recently created and fully functional CentOS 7.3 VM fails to boot >> after applying CR updates: > <snip> >> Server OS is CentOS 7.3 using Xen (no CR updates): >> rpm -qa xen\* >> xen-hypervisor-4.6.3-15.el7.x86_64 >> xen-4.6.3-15.el7.x86_64 >>
2005 Aug 23
1
2.6 Kernel Out of memory
Hi guys, Another strange behavior with some test servers. They are booted using 2.6.11.12, xen-2.0.6, Debian. The only daemons running on dom0 are : xend, smtp, quagga (ospf), portmap (nfs client), sshd. Memory usage at startup is aroung 30 MB. Dom0 reserved memory is 256 MB. The only things that happen on this server are domU''s start/stop (xm create, xm shutdown) and a
2015 Dec 20
8
[Bug 93458] New: page allocation failure: order:5, mode:0x240c0c0
https://bugs.freedesktop.org/show_bug.cgi?id=93458 Bug ID: 93458 Summary: page allocation failure: order:5, mode:0x240c0c0 Product: xorg Version: 7.7 (2012.06) Hardware: x86-64 (AMD64) OS: Linux (All) Status: NEW Severity: normal Priority: medium Component: Driver/nouveau
2017 Sep 04
2
Xen CentOS 7.3 server + CentOS 7.3 VM fails to boot after CR updates (applied to VM)!
On 09/04/2017 03:59 PM, Kevin Stange wrote: > On 09/02/2017 08:11 AM, Johnny Hughes wrote: >> On 09/01/2017 02:41 PM, Kevin Stange wrote: >>> On 08/31/2017 07:50 AM, PJ Welsh wrote: >>>> A recently created and fully functional CentOS 7.3 VM fails to boot >>>> after applying CR updates: >>> <snip> >>>> Server OS is CentOS 7.3
2011 Jul 01
1
[79030.229547] motion: page allocation failure: order:6, mode:0xd4
Hi Konrad, In one of my domU''s to which I pci passthrough USB cards with a videograbber connected to it, the videocapture bails out after running for quite some time with the stacktrace below. I guess i''m interpreting the mem-info below wrong, but i fail to see why it can''t allocate those 185088 bytes. Could you shed some light ? it''s probably the coherent
2017 Oct 23
1
problems running a vol over IPoIB, and qemu off it?
hi people I wonder if anybody experience any problems with vols in replica mode that run across IPoIB links and libvirt stores qcow image on such a volume? I wonder if maybe devel could confirm it should just work, and then hardware/Infiniband I should blame. I have a direct IPoIB link between two hosts, gluster replica volume, libvirt store disk images there. I start a guest on hostA and
2013 Sep 23
6
btrfs: qgroup scan failed with -12
Not sure if it''s anything interesting - I had the following entry in dmesg a few days ago, on a server with 32 GB RAM. The system is still working fine. [1878432.675210] btrfs-qgroup-re: page allocation failure: order:5, mode:0x104050 [1878432.675319] CPU: 5 PID: 22251 Comm: btrfs-qgroup-re Not tainted 3.11.0-rc7 #2 [1878432.675417] Hardware name: System manufacturer System Product
2011 Oct 25
1
Page allocation failure
Dear list, I am seeing an error across multiple machines during heavy I/O, either disk or network. The VMs are on different Intel CPUs (Core 2 Quad, Core i5, Xeon) with varying boards (Abit, Asus, Supermicro). Machines that get this error are running either BackupPC, Zabbix (MySQL) or SABNZBd. I can also reproduce the error on the Supermicros with a looped wget of an ubuntu ISO as they are
2017 Oct 18
2
[PATCH] virtio: avoid possible OOM lockup at virtballoon_oom_notify()
Tetsuo Handa wrote: > 20171016-deflate.log.xz continued printing "puff" messages without any OOM > killer messages, for fill_balloon() always inflates faster than leak_balloon() > deflates. > > Since the OOM killer cannot be invoked unless leak_balloon() completely > deflates faster than fill_balloon() inflates, the guest remained unusable > (e.g. unable to login
2017 Oct 18
2
[PATCH] virtio: avoid possible OOM lockup at virtballoon_oom_notify()
Tetsuo Handa wrote: > 20171016-deflate.log.xz continued printing "puff" messages without any OOM > killer messages, for fill_balloon() always inflates faster than leak_balloon() > deflates. > > Since the OOM killer cannot be invoked unless leak_balloon() completely > deflates faster than fill_balloon() inflates, the guest remained unusable > (e.g. unable to login
2019 Jul 12
2
Out of memory: kill process
Hello, On my bridge head DC, i can see in kern.log lots of < Out of memory : kill process > I'm using Samba 4.9.6 (11147 objects) on Debian Stretch 64 This DC synchronize from/to 20 others DC in bridge head mode (no mesh) Here is my VM (HyperV 2016) config : - 4 x Vcpu (intel Xeon Silver 4110) - 2 Go Ram - 1 Go swap I'm really not an expert on this
2013 Dec 30
2
oom situation
I have continous oom&panic situation unresolved. I am not sure system fills up all the ram (36GB). Why this system triggered this oom situation? Is it about some other memory? highmem? lowmem? stack size? Best Regards, Kernel 3.10.24 Dec 27 09:19:05 2013 kernel: : [277622.359064] squid invoked oom-killer: gfp_mask=0x42d0, order=3, oom_score_adj=0 Dec 27 09:19:05 2013 kernel: :
2005 Oct 18
4
dom0 oom-killer: gfp_mask=0x1d
I had the dom0 which unfortunately didn''t have a console on it hit a race condition and saw oom errors on it also. This happened after it was running for over 36 hours with a domU whose load average was avg was around 3 most of the time. changeset: 7396:9b51e7637676 Dom0 - UP i686, Centos 4.1, 768 megs domU-1 92 megs snmpd domU-2 92 megs snmpdd domU-3 410 megs postgres, tomcat 5.5,
2014 Feb 03
3
Memory leak - how to investigate
My web & name server runs out of memory from time to time, to the point where it's completely unresponsive to anything. At that point reset is the only alternative. (Or, as this is a virtual guest, I just say "virsh destroy"). But why this happens - I would like to know. The host in question is a KVM guest, and runs CentOS 6.4. From "top" (situation now): Mem:
2013 Nov 19
5
xenwatch: page allocation failure: order:4, mode:0x10c0d0 xen_netback:xenvif_alloc: Could not allocate netdev for vif16.0
Hi Wei, I ran into the following problem when trying to boot another guest after less than a day of uptime. (the system started 15 guests at boot already which went fine). dom0 is allocated a fixed 1536M. Both host as pv guests run the same kernel, some hvm''s run a slightly older kernel (3.9 f.e.) The are quite some granttable messages in xl dmesg, i also included these and a
2009 Jul 29
2
out of memory
I am getting this message quite often lately. Centos 5.3, AMD dual core 5050e system x64 1 GIG ram, 4 GIG swap I dont have that much running that the kernel should be cutting out my processes. Any thoughts? Jerry ------------ emTotal: 766264 kB MemFree: 583984 kB Buffers: 9572 kB Cached: 31004 kB SwapCached: 36280 kB Active: 20240 kB Inactive:
2019 Dec 05
2
[PATCH v2] virtio-balloon: fix managed page counts when migrating pages between zones
In case we have to migrate a ballon page to a newpage of another zone, the managed page count of both zones is wrong. Paired with memory offlining (which will adjust the managed page count), we can trigger kernel crashes and all kinds of different symptoms. One way to reproduce: 1. Start a QEMU guest with 4GB, no NUMA 2. Hotplug a 1GB DIMM and only the memory to ZONE_NORMAL 3. Inflate the balloon