Dario Lesca
2021-Sep-06 22:13 UTC
[Gluster-users] gluster volume status show second node is offline
Hello everybody!
I'm a novice with gluster. I have setup my first cluster with two
nodes?
This is the current volume info:
[root at s-virt1 ~]# gluster volume info gfsvol1
Volume Name: gfsvol1
Type: Replicate
Volume ID: 5bad4a23-58cc-44d7-8195-88409720b941
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: virt1.local:/gfsvol1/brick1
Brick2: virt2.local:/gfsvol1/brick1
Options Reconfigured:
performance.client-io-threads: off
nfs.disable: on
transport.address-family: inet
storage.fips-mode-rchecksum: on
cluster.granular-entry-heal: on
storage.owner-uid: 107
storage.owner-gid: 107
server.allow-insecure: on
For now all seem work fine.
I have mount the gfs volume on all two nodes and use the VM into it
But today I noticed that the second node (virt2) is offline:
[root at s-virt1 ~]# gluster volume status
Status of volume: gfsvol1
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick virt1.local:/gfsvol1/brick1 49152 0 Y 3090
Brick virt2.local:/gfsvol1/brick1 N/A N/A N N/A
Self-heal Daemon on localhost N/A N/A Y 3105
Self-heal Daemon on virt2.local N/A N/A Y 3140
Task Status of Volume gfsvol1
------------------------------------------------------------------------------
There are no active volume tasks
[root at s-virt1 ~]# gluster volume status gfsvol1 detail
Status of volume: gfsvol1
------------------------------------------------------------------------------
Brick : Brick virt1.local:/gfsvol1/brick1
TCP Port : 49152
RDMA Port : 0
Online : Y
Pid : 3090
File System : xfs
Device : /dev/mapper/rl-gfsvol1
Mount Options :
rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,sunit=128,swidth=128,noquota
Inode Size : 512
Disk Space Free : 146.4GB
Total Disk Space : 999.9GB
Inode Count : 307030856
Free Inodes : 307026149
------------------------------------------------------------------------------
Brick : Brick virt2.local:/gfsvol1/brick1
TCP Port : N/A
RDMA Port : N/A
Online : N
Pid : N/A
File System : xfs
Device : /dev/mapper/rl-gfsvol1
Mount Options :
rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,sunit=128,swidth=128,noquota
Inode Size : 512
Disk Space Free : 146.4GB
Total Disk Space : 999.9GB
Inode Count : 307052016
Free Inodes : 307047307
What does it mean?
What's wrong?
Is this normal or I missing some setting?
If you need more information let me know
Many thanks for your help
--
Dario Lesca
(inviato dal mio Linux Fedora 34 Workstation)
Strahil Nikolov
2021-Sep-07 05:28 UTC
[Gluster-users] gluster volume status show second node is offline
No, it's not normal.Go to the virt2 and in /var/log/gluster directory you will find 'bricks' . Check the logs in bricks for more information. Best Regards,Strahil Nikolov On Tue, Sep 7, 2021 at 1:13, Dario Lesca<d.lesca at solinos.it> wrote: Hello everybody! I'm a novice with gluster. I have setup my first cluster with two nodes? This is the current volume info: ? [root at s-virt1 ~]# gluster volume info gfsvol1 ? Volume Name: gfsvol1 ? Type: Replicate ? Volume ID: 5bad4a23-58cc-44d7-8195-88409720b941 ? Status: Started ? Snapshot Count: 0 ? Number of Bricks: 1 x 2 = 2 ? Transport-type: tcp ? Bricks: ? Brick1: virt1.local:/gfsvol1/brick1 ? Brick2: virt2.local:/gfsvol1/brick1 ? Options Reconfigured: ? performance.client-io-threads: off ? nfs.disable: on ? transport.address-family: inet ? storage.fips-mode-rchecksum: on ? cluster.granular-entry-heal: on ? storage.owner-uid: 107 ? storage.owner-gid: 107 ? server.allow-insecure: on For now all seem work fine. I have mount the gfs volume on all two nodes and use the VM into it But today I noticed that the second node (virt2) is offline: ? [root at s-virt1 ~]# gluster volume status ? Status of volume: gfsvol1 ? Gluster process? ? ? ? ? ? ? ? ? ? ? ? ? ? TCP Port? RDMA Port? Online? Pid ? ------------------------------------------------------------------------------ ? Brick virt1.local:/gfsvol1/brick1? ? ? ? ? 49152? ? 0? ? ? ? ? Y? ? ? 3090 ? Brick virt2.local:/gfsvol1/brick1? ? ? ? ? N/A? ? ? N/A? ? ? ? N? ? ? N/A? ? Self-heal Daemon on localhost? ? ? ? ? ? ? N/A? ? ? N/A? ? ? ? Y? ? ? 3105 ? Self-heal Daemon on virt2.local? ? ? ? ? ? N/A? ? ? N/A? ? ? ? Y? ? ? 3140 ? ? ? Task Status of Volume gfsvol1 ? ------------------------------------------------------------------------------ ? There are no active volume tasks ? ? [root at s-virt1 ~]# gluster volume status gfsvol1 detail ? Status of volume: gfsvol1 ? ------------------------------------------------------------------------------ ? Brick? ? ? ? ? ? ? ? : Brick virt1.local:/gfsvol1/brick1 ? TCP Port? ? ? ? ? ? : 49152? ? ? ? ? ? ? ? RDMA Port? ? ? ? ? ? : 0? ? ? ? ? ? ? ? ? ? Online? ? ? ? ? ? ? : Y? ? ? ? ? ? ? ? ? ? Pid? ? ? ? ? ? ? ? ? : 3090? ? ? ? ? ? ? ? ? File System? ? ? ? ? : xfs? ? ? ? ? ? ? ? ? Device? ? ? ? ? ? ? : /dev/mapper/rl-gfsvol1 ? Mount Options? ? ? ? : rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,sunit=128,swidth=128,noquota ? Inode Size? ? ? ? ? : 512? ? ? ? ? ? ? ? ? Disk Space Free? ? ? : 146.4GB? ? ? ? ? ? ? Total Disk Space? ? : 999.9GB? ? ? ? ? ? ? Inode Count? ? ? ? ? : 307030856? ? ? ? ? ? Free Inodes? ? ? ? ? : 307026149? ? ? ? ? ? ------------------------------------------------------------------------------ ? Brick? ? ? ? ? ? ? ? : Brick virt2.local:/gfsvol1/brick1 ? TCP Port? ? ? ? ? ? : N/A? ? ? ? ? ? ? ? ? RDMA Port? ? ? ? ? ? : N/A? ? ? ? ? ? ? ? ? Online? ? ? ? ? ? ? : N? ? ? ? ? ? ? ? ? ? Pid? ? ? ? ? ? ? ? ? : N/A? ? ? ? ? ? ? ? ? File System? ? ? ? ? : xfs? ? ? ? ? ? ? ? ? Device? ? ? ? ? ? ? : /dev/mapper/rl-gfsvol1 ? Mount Options? ? ? ? : rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,sunit=128,swidth=128,noquota ? Inode Size? ? ? ? ? : 512? ? ? ? ? ? ? ? ? Disk Space Free? ? ? : 146.4GB? ? ? ? ? ? ? Total Disk Space? ? : 999.9GB? ? ? ? ? ? ? Inode Count? ? ? ? ? : 307052016? ? ? ? ? ? Free Inodes? ? ? ? ? : 307047307 ? What does it mean? What's wrong? Is this normal or I missing some setting? If you need more information let me know Many thanks for your help -- Dario Lesca (inviato dal mio Linux Fedora 34 Workstation) ________ Community Meeting Calendar: Schedule - Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC Bridge: https://meet.google.com/cpu-eiue-hvk Gluster-users mailing list Gluster-users at gluster.org https://lists.gluster.org/mailman/listinfo/gluster-users -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20210907/240b44e6/attachment.html>