similar to: Guests can't connect to each other

Displaying 20 results from an estimated 2000 matches similar to: "Guests can't connect to each other"

2017 Apr 26
0
Re: Tunnelled migrate Windows7 VMs halted
On Wed, Apr 26, 2017 at 08:51:39AM -0500, Eric Blake wrote: > > > > I migrated a Windows 7 VM with libvirtd tunnelled, the VM halted > > on the target although the status is running. What do you mean by halted ? The guest OS has shutdown, or QEMU has crashed, or something else ? > > > > > > [root@test15 ~]# virsh migrate --live --p2p --tunnelled
2014 Feb 12
2
Re: Help? Running into problems with migrateToURI2() and virDomainDefCheckABIStability()
On 02/11/2014 04:45 PM, Cole Robinson wrote: > On 02/10/2014 06:46 PM, Chris Friesen wrote: >> Hi, >> >> We've run into a problem with libvirt 1.1.2 and are looking for some comments >> on whether this is a bug or design intent. >> >> We're trying to use migrateToURI() but we're using a few things (numatune, >> vcpu mask, etc.) that may need
2023 May 12
1
Question regaring correct usage of CPU shares
Hi there, I have a question regarding the shares option of the cputune section. I want to illustrate my question with the following example. Let's assume I have two virtual machines like the following on four dedicated core with two threads each: VM1: <cputune> <shares>512</shares> <vcpupin vcpu="0" cpuset="0"/> <vcpupin
2018 Sep 17
0
Re: NUMA issues on virtualized hosts
Hello, so the current domain configuration: <cpu mode='host-passthrough'><topology sockets='8' cores='4' threads='1'/><numa><cell cpus='0-3' memory='62000000' /><cell cpus='4-7' memory='62000000' /><cell cpus='8-11' memory='62000000' /><cell cpus='12-15'
2018 Sep 14
0
Re: NUMA issues on virtualized hosts
Hello, ok, I found that cpu pinning was wrong, so I corrected it to be 1:1. The issue with iozone remains the same. The spec is running, however, it runs slower than 1-NUMA case. The corrected XML looks like follows: <cpu mode='host-passthrough'><topology sockets='8' cores='4' threads='1'/><numa><cell cpus='0-3'
2017 Apr 19
2
virsh error: driver is not whitelisted
Hi, I'm using virsh to instance a VM in my environment, but I'm running on some issues. I created the following domain file: <domain type='kvm'> <name>demovm</name> <uuid>4a9b3f53-fa2a-47f3-a757-dd87720d9d1d</uuid> <memory unit='KiB'>4194304</memory> <currentMemory
2018 Sep 14
1
Re: NUMA issues on virtualized hosts
Hello again, when the iozone writes slow. This is how slabtop looks like: 62476752 62476728 0% 0.10K 1601968 39 6407872K buffer_head 1000678 999168 0% 0.56K 142954 7 571816K radix_tree_node 132184 125911 0% 0.03K 1066 124 4264K kmalloc-32 118496 118224 0% 0.12K 3703 32 14812K kmalloc-node 73206 56467 0% 0.19K 3486 21
2017 Apr 26
3
Tunnelled migrate Windows7 VMs halted
[moderator note: I'm forwarding a stripped down version of the original mail which was rejected in the moderator queue. I stripped the 3.3 megabyte .tar.bz2 of the log file attachment, which is inappropriate for a technical list. Either trim the log to the relevant portion, or host the log externally and have your list email merely give a URL of the externally-hosted file] >
2018 Sep 17
2
Re: NUMA issues on virtualized hosts
On 09/14/2018 03:36 PM, Lukas Hejtmanek wrote: > Hello, > > ok, I found that cpu pinning was wrong, so I corrected it to be 1:1. The issue > with iozone remains the same. > > The spec is running, however, it runs slower than 1-NUMA case. > > The corrected XML looks like follows: [Reformated XML for better reading] <cpu mode="host-passthrough">
2018 Sep 18
1
Re: NUMA issues on virtualized hosts
On 09/17/2018 04:59 PM, Lukas Hejtmanek wrote: > Hello, > > so the current domain configuration: > <cpu mode='host-passthrough'><topology sockets='8' cores='4' threads='1'/><numa><cell cpus='0-3' memory='62000000' /><cell cpus='4-7' memory='62000000' /><cell cpus='8-11'
2018 Sep 14
3
NUMA issues on virtualized hosts
Hello, I have cluster with AMD EPYC 7351 cpu. Two CPUs per node. I have performance 8-NUMA configuration: This is from hypervizor: [root@hde10 ~]# lscpu Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Byte Order: Little Endian CPU(s): 64 On-line CPU(s) list: 0-63 Thread(s) per core: 2 Core(s) per socket: 16 Socket(s): 2 NUMA
2013 Jun 14
0
can virsh set the cpuset attribute of <vcpu ..> (CPU Allocation) ?
Is it possible to use virsh to set the cpuset attribute of the CPU Allocation element in a domain ? <domain> ... <vcpu placement='static' cpuset="1-4,^3,6" current="1">2</vcpu> ... </domain> I have seen that virsh vcpupin and virsh emulatorpin can be used to query and set the cpusets of the <vcpupin> and <emulatorpin>
2012 Feb 25
2
Problems running virtual machines after an upgrade
Hi all, I upgraded qemu-kvm and libvirt to the new versions qemu-kvm at : 1.0 libvirt at : git version 0.9.10 I lost the original machines but the xml files were are all present (it is not showing up in virsh list --all) But when I import an xml I get an error: Attached is some information which might be helpful Can someone help please ? If this is a compatibility issue please can
2017 Mar 21
0
Re: CPU Pinning Help
2018 Sep 05
2
Domain vCPU threads affinity
Hello, According to the docs, vcpupin will use either cgroups or sched_setaffinity to pin vcpu threads to cpus. How is this decision made? I observe differences even on different hosts featuring the same version of libvirtd (1.3.1): on one host vcpupin affects cpuset.cpus (cgroup), and on the other it affects vcpu threads affinity (observed through taskset). Thanks, Nikos -------------- next
2014 Jul 24
2
Failed to bind to uuid (GUID)._msdcs.DOMAIN NT_STATUS_NO_LOGON_SERVERS & IRPC callback failed for DsReplicaSync - NT_STATUS_IO_TIMEOUT
Hello everyone. After all the problems I had demoting my DC and managed it to work again, it was working for only a week. Now i am getting this messages on the log: [2014/07/22 16:13:11.745783, 0] .../source4/librpc/rpc/dcerpc_util.c:681(dcerpc_pipe_auth_recv) Failed to bind to uuid e3514235-4b06-11d1-ab04-00c04fc2dcd2 for e3514235-4b06-11d1-ab04-00c04fc2dcd2 at
2014 Jan 15
2
Does libvirt lxc driver support "cpuset" attribute?
Dear all I allocate only one vcpu for the container by the following statement, that is, I want to pin the vcpu to physical core "2". <vcpu placement='static' cpuset="2" >1</vcpu> My host has 4 physical cores. Before test, all the 4 cores are idle. After I run 4 processes in the container, I found all the 4 cores in the host are 100% used. That is, the
2019 Aug 29
0
[libvirtd] qemu_process: reset CPU affinity to all enabled CPUs, when runs in custom cpuset
Hello All, Since 4.5.0-23.el7 version (Red Hat 7.7), when I launch pinned VM, libvirtd reset CPU affinity to all enabled in host CPUs, if it runs in custom cpuset. I can't reproduce this behavior with 4.5.0-10.el7_6.12 with the same kernel version (Red Hat 7.7). Libvirt runs in a custom cpuset 'libvirt', where the number of available cpus is restricted to 0,2,4,6,8. And this
2018 Sep 05
0
Re: Domain vCPU threads affinity
On Wed, Sep 05, 2018 at 03:48:45PM +0300, Nikos Anastopoulos wrote: >Hello, > >According to the docs, vcpupin will use either cgroups or sched_setaffinity >to pin vcpu threads to cpus. How is this decision made? >I observe differences even on different hosts featuring the same version of >libvirtd (1.3.1): on one host vcpupin affects cpuset.cpus (cgroup), and on >the other it
2016 May 19
1
libvirt on Raspberry Pi, block by parameter "-S"
Hi I have used libvirt on Raspberry Pi Debian. I can start the vm, but it just starts and doesn't continue. I find the qemu-system-arm command generated by virsh has a parameter "-S", when I copy the command, remove the parameter, run it in the console, the vm can run normally. But I don't know how to remove the parameter in the xml file, or in virsh. Regards Thank you Xu