Gilberto Ferreira
2024-Nov-11 18:57 UTC
[Gluster-users] Disk size and virtual size drive me crazy!
Hi there. I can't understand why I am having this different values: proxmox01:/vms/images# df Sist. Arq. Tam. Usado Disp. Uso% Montado em udev 252G 0 252G 0% /dev tmpfs 51G 9,4M 51G 1% /run /dev/sda4 433G 20G 413G 5% / tmpfs 252G 63M 252G 1% /dev/shm tmpfs 5,0M 0 5,0M 0% /run/lock efivarfs 496K 335K 157K 69% /sys/firmware/efi/efivars /dev/sda2 1,8G 204M 1,5G 12% /boot /dev/sda1 1,9G 12M 1,9G 1% /boot/efi /dev/sdb 932G 728G 204G 79% /disco1TB-0 /dev/sdc 932G 718G 214G 78% /disco1TB-1 /dev/sde 932G 720G 212G 78% /disco1TB-2 /dev/sdd 1,9T 1,5T 387G 80% /disco2TB-0 tmpfs 51G 4,0K 51G 1% /run/user/0 *gluster1:VMS 4,6T 3,6T 970G 80% /vms* /dev/fuse 128M 36K 128M 1% /etc/pve proxmox01:/vms/images# cd 103 proxmox01:/vms/images/103# ls vm-103-disk-0.qcow2 vm-103-disk-1.qcow2 proxmox01:/vms/images/103# ls -lh total 21T *-rw-r----- 1 root root 101G nov 11 15:53 vm-103-disk-0.qcow2-rw-r----- 1 root root 210G nov 11 15:45 vm-103-disk-1.qcow2* proxmox01:/vms/images/103# qemu-img info vm-103-disk-0.qcow2 image: vm-103-disk-0.qcow2 file format: qcow2 *virtual size: 100 GiB (107374182400 bytes)disk size: 3.78 TiB* cluster_size: 65536 Format specific information: compat: 1.1 compression type: zlib lazy refcounts: false refcount bits: 16 corrupt: false extended l2: false Child node '/file': filename: vm-103-disk-0.qcow2 protocol type: file * file length: 100 GiB (107390828544 bytes) disk size: 3.78 TiB* proxmox01:/vms/images/103# qemu-img info vm-103-disk-1.qcow2 image: vm-103-disk-1.qcow2 file format: qcow2 *virtual size: 2 TiB (2199023255552 bytes)disk size: 16.3 TiB* cluster_size: 65536 Format specific information: compat: 1.1 compression type: zlib lazy refcounts: false refcount bits: 16 corrupt: false extended l2: false Child node '/file': filename: vm-103-disk-1.qcow2 protocol type: file * file length: 210 GiB (225117732864 bytes) disk size: 16.3 TiB* proxmox01:/vms/images/103# Here is the vol info. proxmox01:/vms/images/103# gluster vol info Volume Name: VMS Type: Distributed-Replicate Volume ID: a98f7944-4308-499f-994e-9029f3be56c0 Status: Started Snapshot Count: 0 Number of Bricks: 4 x 2 = 8 Transport-type: tcp Bricks: Brick1: gluster1:/disco2TB-0/vms Brick2: gluster2:/disco2TB-0/vms Brick3: gluster1:/disco1TB-0/vms Brick4: gluster2:/disco1TB-0/vms Brick5: gluster1:/disco1TB-1/vms Brick6: gluster2:/disco1TB-1/vms Brick7: gluster1:/disco1TB-2/vms Brick8: gluster2:/disco1TB-2/vms Options Reconfigured: cluster.lookup-optimize: off server.keepalive-count: 5 server.keepalive-interval: 2 server.keepalive-time: 10 server.tcp-user-timeout: 20 server.event-threads: 4 client.event-threads: 4 cluster.choose-local: off cluster.shd-wait-qlength: 10000 cluster.shd-max-threads: 8 cluster.locking-scheme: granular cluster.server-quorum-type: none cluster.quorum-type: fixed network.remote-dio: disable performance.client-io-threads: on performance.strict-o-direct: on performance.low-prio-threads: 32 performance.io-cache: off performance.read-ahead: off performance.quick-read: off performance.flush-behind: off performance.write-behind: off cluster.data-self-heal-algorithm: full cluster.favorite-child-policy: mtime network.ping-timeout: 20 cluster.quorum-count: 1 cluster.quorum-reads: false cluster.self-heal-daemon: enable cluster.heal-timeout: 5 user.cifs: off features.shard: on cluster.granular-entry-heal: enable storage.fips-mode-rchecksum: on transport.address-family: inet nfs.disable: on --- Gilberto Nunes Ferreira -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20241111/8d5aba17/attachment.html>
Strahil Nikolov
2024-Nov-20 22:28 UTC
[Gluster-users] Disk size and virtual size drive me crazy!
What's your gluster version ? Best Regards,Strahil Nikolov ? ??????????, 11 ??????? 2024 ?. ? 20:57:50 ?. ???????+2, Gilberto Ferreira <gilberto.nunes32 at gmail.com> ??????: Hi there. I can't understand why I am having? this different values: proxmox01:/vms/images# df Sist. Arq. ? ? ?Tam. Usado Disp. Uso% Montado em udev ? ? ? ? ? ?252G ? ? 0 ?252G ? 0% /dev tmpfs ? ? ? ? ? ?51G ?9,4M ? 51G ? 1% /run /dev/sda4 ? ? ? 433G ? 20G ?413G ? 5% / tmpfs ? ? ? ? ? 252G ? 63M ?252G ? 1% /dev/shm tmpfs ? ? ? ? ? 5,0M ? ? 0 ?5,0M ? 0% /run/lock efivarfs ? ? ? ?496K ?335K ?157K ?69% /sys/firmware/efi/efivars /dev/sda2 ? ? ? 1,8G ?204M ?1,5G ?12% /boot /dev/sda1 ? ? ? 1,9G ? 12M ?1,9G ? 1% /boot/efi /dev/sdb ? ? ? ?932G ?728G ?204G ?79% /disco1TB-0 /dev/sdc ? ? ? ?932G ?718G ?214G ?78% /disco1TB-1 /dev/sde ? ? ? ?932G ?720G ?212G ?78% /disco1TB-2 /dev/sdd ? ? ? ?1,9T ?1,5T ?387G ?80% /disco2TB-0 tmpfs ? ? ? ? ? ?51G ?4,0K ? 51G ? 1% /run/user/0 gluster1:VMS ? ?4,6T ?3,6T ?970G ?80% /vms /dev/fuse ? ? ? 128M ? 36K ?128M ? 1% /etc/pve proxmox01:/vms/images# cd 103 proxmox01:/vms/images/103# ls vm-103-disk-0.qcow2 ?vm-103-disk-1.qcow2 proxmox01:/vms/images/103# ls -lh total 21T -rw-r----- 1 root root 101G nov 11 15:53 vm-103-disk-0.qcow2 -rw-r----- 1 root root 210G nov 11 15:45 vm-103-disk-1.qcow2 proxmox01:/vms/images/103# qemu-img info vm-103-disk-0.qcow2 image: vm-103-disk-0.qcow2 file format: qcow2 virtual size: 100 GiB (107374182400 bytes) disk size: 3.78 TiB cluster_size: 65536 Format specific information: ? ? compat: 1.1 ? ? compression type: zlib ? ? lazy refcounts: false ? ? refcount bits: 16 ? ? corrupt: false ? ? extended l2: false Child node '/file': ? ? filename: vm-103-disk-0.qcow2 ? ? protocol type: file ? ? file length: 100 GiB (107390828544 bytes) ? ? disk size: 3.78 TiB proxmox01:/vms/images/103# qemu-img info vm-103-disk-1.qcow2 image: vm-103-disk-1.qcow2 file format: qcow2 virtual size: 2 TiB (2199023255552 bytes) disk size: 16.3 TiB cluster_size: 65536 Format specific information: ? ? compat: 1.1 ? ? compression type: zlib ? ? lazy refcounts: false ? ? refcount bits: 16 ? ? corrupt: false ? ? extended l2: false Child node '/file': ? ? filename: vm-103-disk-1.qcow2 ? ? protocol type: file ? ? file length: 210 GiB (225117732864 bytes) ? ? disk size: 16.3 TiB proxmox01:/vms/images/103# ?Here is the vol info. proxmox01:/vms/images/103# gluster vol info ? Volume Name: VMS Type: Distributed-Replicate Volume ID: a98f7944-4308-499f-994e-9029f3be56c0 Status: Started Snapshot Count: 0 Number of Bricks: 4 x 2 = 8 Transport-type: tcp Bricks: Brick1: gluster1:/disco2TB-0/vms Brick2: gluster2:/disco2TB-0/vms Brick3: gluster1:/disco1TB-0/vms Brick4: gluster2:/disco1TB-0/vms Brick5: gluster1:/disco1TB-1/vms Brick6: gluster2:/disco1TB-1/vms Brick7: gluster1:/disco1TB-2/vms Brick8: gluster2:/disco1TB-2/vms Options Reconfigured: cluster.lookup-optimize: off server.keepalive-count: 5 server.keepalive-interval: 2 server.keepalive-time: 10 server.tcp-user-timeout: 20 server.event-threads: 4 client.event-threads: 4 cluster.choose-local: off cluster.shd-wait-qlength: 10000 cluster.shd-max-threads: 8 cluster.locking-scheme: granular cluster.server-quorum-type: none cluster.quorum-type: fixed network.remote-dio: disable performance.client-io-threads: on performance.strict-o-direct: on performance.low-prio-threads: 32 performance.io-cache: off performance.read-ahead: off performance.quick-read: off performance.flush-behind: off performance.write-behind: off cluster.data-self-heal-algorithm: full cluster.favorite-child-policy: mtime network.ping-timeout: 20 cluster.quorum-count: 1 cluster.quorum-reads: false cluster.self-heal-daemon: enable cluster.heal-timeout: 5 user.cifs: off features.shard: on cluster.granular-entry-heal: enable storage.fips-mode-rchecksum: on transport.address-family: inet nfs.disable: on--- Gilberto Nunes Ferreira ________ Community Meeting Calendar: Schedule - Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC Bridge: https://meet.google.com/cpu-eiue-hvk Gluster-users mailing list Gluster-users at gluster.org https://lists.gluster.org/mailman/listinfo/gluster-users -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20241120/d90854ad/attachment.html>
Apparently Analagous Threads
- Disk size and virtual size drive me crazy!
- Add an arbiter when have multiple bricks at same server.
- Add an arbiter when have multiple bricks at same server.
- Add an arbiter when have multiple bricks at same server.
- Add an arbiter when have multiple bricks at same server.