Displaying 20 results from an estimated 20000 matches similar to: "Configure QCow2 L2 Cache through virt domain XML"
2016 Sep 20
0
Re: How to set QEMU qcow2 l2-cache-size using libvirt xml?
On Mon, Sep 19, 2016 at 11:32:26AM -0400, Frank Myhr wrote:
>QEMU's default qcow2 L2 cache size is too small for large images (and small cluster sizes), resulting in very bad performance.
>
>https://blogs.igalia.com/berto/2015/12/17/improving-disk-io-performance-in-qemu-2-5-with-the-qcow2-l2-cache/
>shows huge performance hit for a 20GB qcow2 with default 64kB cluster size:
>
2016 Sep 19
2
How to set QEMU qcow2 l2-cache-size using libvirt xml?
QEMU's default qcow2 L2 cache size is too small for large images (and small cluster sizes), resulting in very bad performance.
https://blogs.igalia.com/berto/2015/12/17/improving-disk-io-performance-in-qemu-2-5-with-the-qcow2-l2-cache/
shows huge performance hit for a 20GB qcow2 with default 64kB cluster size:
L2 Cache, MiB Average IOPS
1 (default) 5100
1.5
2014 Jan 16
0
Re: Double fault panic in L2 upon v2v conversion
There is a bug which is still not fixed:Bug 956580 -[ RFE ]Support for VMware & Citrix OVF image import into RHEV
https://bugzilla.redhat.com/show_bug.cgi?id=956580
Best regards,
Tingting Zheng(郑婷婷)
----- Original Message -----
From: "Rom Freiman" <rom@stratoscale.com>
To: libguestfs@redhat.com
Sent: Wednesday, January 15, 2014 10:35:29 PM
Subject: [Libguestfs] Double fault
2014 Jan 15
0
Double fault panic in L2 upon v2v conversion
Hi everybody,
Wanted to hear your opinion and to receive a smart advice.
I'm trying to use virt-v2v in order to convert ova image (exported from
vmware) to run on libvirt/kvm - all this inside a VM of fedora.
The converted image is also a fedora.
During the conversion process, in some point of libguestfs activity, I get
double fault panic from L2 (printed as part of libguest output) and the
2014 Feb 12
3
Debugging nested KVM guest (L2) booting with libguestfs/gdb
Heya,
With latest Fedora Rawhide Kernel, I see a nested KVM guest hanging at
boot (not unusual). Rich once suggested this[1] to try to attach gdb to
the nested L2 guest to find out where_ it's stuck, tonight I set out to
try it out (with KVM & TCG).
Below is all what I tried.
In guest hypervisor (L1):
$ git clone git://github.com/libguestfs/libguestfs.git
$ git log | head -1
2020 Jul 16
1
Cannot pass secret id for backing file after taking external snapshot on encrypted qcow2 file
Hi,
I used 'virsh snapshot-create' create an encrypted external snapshot, when I try to use 'qemu-img check' top file, found no entrance to pass backing-file's secret-id
1、Version
centos-release-8.2-2.2004.0.1.el8.x86_64
libvirt.x86_64 6.0.0-17.el8
qemu-kvm.x86_64
2023 Apr 05
1
backup-begin
Den 2023-04-05 kl. 09:47, skrev Peter Krempa:
> The backup operation is quite complex so it is possible.
> Please have a look into /var/log/libvirt/qemu/$VMNAME.log to see whether
> qemu logged something like an assertion failure before crashing.
>
> Additionally you can have a look into 'coredumpctl' whether there are
> any recorded crashes of
2024 Dec 02
1
Disk size and virtual size drive me crazy!
qemu-img info 100/vm-100-disk-0.qcow2
image: 100/vm-100-disk-0.qcow2
file format: qcow2
virtual size: 120 GiB (128849018880 bytes)
disk size: 916 GiB
cluster_size: 65536
Format specific information:
compat: 1.1
compression type: zlib
lazy refcounts: false
refcount bits: 16
corrupt: false
extended l2: false
Child node '/file':
filename: 100/vm-100-disk-0.qcow2
2007 Apr 07
1
OT: general question re processor, l2 and l3 cache etc
Greetings
Please forgive the OT question yet I highly value the experience and wisdom
on this list
I am wondering if anyone here can address the performance difference between
having a processor board with say 256KB L2 *and* 2048KB L3 cache *VERSUS*
just having the same processor board with just the L2 cache in a centos
server environment...
Please figure that all other necessary and related
2024 Nov 11
1
Disk size and virtual size drive me crazy!
Hi there.
I can't understand why I am having this different values:
proxmox01:/vms/images# df
Sist. Arq. Tam. Usado Disp. Uso% Montado em
udev 252G 0 252G 0% /dev
tmpfs 51G 9,4M 51G 1% /run
/dev/sda4 433G 20G 413G 5% /
tmpfs 252G 63M 252G 1% /dev/shm
tmpfs 5,0M 0 5,0M 0% /run/lock
efivarfs 496K 335K
2024 Nov 20
1
Disk size and virtual size drive me crazy!
What's your gluster version ?
Best Regards,Strahil Nikolov
? ??????????, 11 ??????? 2024 ?. ? 20:57:50 ?. ???????+2, Gilberto Ferreira <gilberto.nunes32 at gmail.com> ??????:
Hi there.
I can't understand why I am having? this different values:
proxmox01:/vms/images# df
Sist. Arq. ? ? ?Tam. Usado Disp. Uso% Montado em
udev ? ? ? ? ? ?252G ? ? 0 ?252G ? 0% /dev
tmpfs ? ? ?
2018 Apr 30
0
Re: Create qcow2 v3 volumes via libvirt
Il 30-01-2018 13:17 Gionatan Danti ha scritto:
> Hi all,
> on a fully patched CentOS 7.4 x86-64, I see the following behavior:
>
> - when creating a new volumes using vol-create-as, the resulting file
> is a qcow2 version 2 (compat=0.10) file. Example:
>
> [root@gdanti-lenovo vmimages]# virsh vol-create-as default zzz.qcow2
> 8589934592 --format=qcow2 --backing-vol
2024 Nov 20
1
Disk size and virtual size drive me crazy!
11.1
---
Gilberto Nunes Ferreira
+55 (47) 99676-7530
Proxmox VE
VinChin Backup & Restore
Em qua., 20 de nov. de 2024, 19:28, Strahil Nikolov <hunter86_bg at yahoo.com>
escreveu:
> What's your gluster version ?
>
> Best Regards,
> Strahil Nikolov
>
> ? ??????????, 11 ??????? 2024 ?. ? 20:57:50 ?. ???????+2, Gilberto
> Ferreira <gilberto.nunes32 at
2024 Nov 29
1
Disk size and virtual size drive me crazy!
Have you figured it out ?
Have you tried setting storage.reserve to 0 ?
Best Regards,
Strahil Nikolov
On Thu, Nov 21, 2024 at 0:39, Gilberto Ferreira<gilberto.nunes32 at gmail.com> wrote:
11.1
---
Gilberto Nunes Ferreira
+55 (47) 99676-7530
Proxmox VE
VinChin Backup & Restore
Em qua., 20 de nov. de 2024, 19:28, Strahil Nikolov <hunter86_bg at yahoo.com> escreveu:
2024 Nov 29
1
Disk size and virtual size drive me crazy!
No! I didn't! I wasn't aware of this option.
I will try.
Thanks
Em sex., 29 de nov. de 2024 ?s 16:43, Strahil Nikolov <hunter86_bg at yahoo.com>
escreveu:
> Have you figured it out ?
>
> Have you tried setting storage.reserve to 0 ?
>
> Best Regards,
> Strahil Nikolov
>
> On Thu, Nov 21, 2024 at 0:39, Gilberto Ferreira
> <gilberto.nunes32 at
2014 Jan 04
0
IRC question: virt-builder & size of qcow2 output files
This question was asked on IRC:
$ virt-builder centos-6 --output box.img --format qcow2 --size 40G # ...
$ qemu-img convert -O qcow2 box.img box-qcow2.img
$ ls -lAhs
1.6G -rw-r--r--. 1 james 41G Jan 2 21:53 box.img
1.1G -rw-r--r--. 1 james 1.1G Jan 3 18:54 box-qcow2.img
$ file box*
box.img: QEMU QCOW Image (v2), 42949672960 bytes
box-qcow2.img: QEMU QCOW Image (v2),
2009 May 03
1
[LLVMdev] L1, L2 Cache line sizes in TargetData?
Hello,
Is there any way for a pass to determine the L1 or L2 cacheline size
of the target before the IR is lowered to machine instructions?
Thanks,
--
Nick Johnson
2016 Nov 07
0
Disable L2 cache on nvidia gpu
Dear all,
Is there a command or low level instruction to disable the L2 cache in
GPUs. Can a certain section of memory be defined as uncachable.
Can you please point me the approach to do it.
Thank you
Vishwas
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.freedesktop.org/archives/nouveau/attachments/20161107/a1b678b7/attachment.html>
2019 Nov 28
2
virt-df -a xxx.qcow2 failure, after about ten minutes , failed
use the virt-df to query vm disk usage like this:
virt-df -a /home/centos7_mini.qcow2
then check:
virsh list --all
Id Name State
1 guestfs-p1km7lhfowv8o0q1running //virt-df will create a vm in temp filesystem
but after a few minutes , still fails to get the vm disk usage , why ??
[root@bogon libguestfs-1.38.2]# uname -ar
Linux bogon 3.10.0 #2 SMP Thu Nov 28 06:06:42 UTC 2019 x86_64 x86_64
2016 May 17
0
Re: Can I specify qcow2 preallocation method with virt-builder?
On Tue, May 17, 2016 at 6:01 PM, Richard W.M. Jones <rjones@redhat.com>
wrote:
> On Tue, May 17, 2016 at 05:53:31PM +0300, Yaniv Kaul wrote:
> > On Tue, May 17, 2016 at 5:48 PM, Richard W.M. Jones <rjones@redhat.com>
> > wrote:
> >
> > > On Tue, May 17, 2016 at 03:41:53PM +0100, Richard W.M. Jones wrote:
> > > > On Tue, May 17, 2016 at