Martin Decker
2023-Sep-29 15:36 UTC
[Gluster-users] gluster volume status shows -> Online "N" after node reboot.
Hi list, I am using a replica volume (3 nodes) gluster in an ovirt environment and after setting one node in maintenance mode and rebooting it, the "Online" flag in gluster volume status does not go to "Y" again. [root at node1 glusterfs]# gluster volume status Status of volume: my_volume Gluster process TCP Port RDMA Port Online Pid ------------------------------------------------------------------------------ Brick 10.22.1.95:/gluster_bricks/my_glust er/my_gluster N/A N/A N N/A Brick 10.22.1.97:/gluster_bricks/my_glust er/my_gluster 49152 0 Y 4954 Brick 10.22.1.94:/gluster_bricks/my_glust er/my_gluster 49152 0 Y 3574 Self-heal Daemon on localhost N/A N/A Y 3585 Self-heal Daemon on node2 N/A N/A Y 3557 Self-heal Daemon on node3 N/A N/A Y 4973 Task Status of Volume my_volume ------------------------------------------------------------------------------ There are no active volume tasks Shouldn?t it go back to Online Y automatically? This is the output from gluster volume info from the same node: [root at node1 glusterfs]# gluster volume info Volume Name: my_volume Type: Replicate Volume ID: 78b9299c-1df5-4780-b108-4d3a6dee225d Status: Started Snapshot Count: 0 Number of Bricks: 1 x 3 = 3 Transport-type: tcp Bricks: Brick1: 10.22.1.95:/gluster_bricks/my_gluster/my_gluster Brick2: 10.22.1.97:/gluster_bricks/my_gluster/my_gluster Brick3: 10.22.1.94:/gluster_bricks/my_gluster/my_gluster Options Reconfigured: cluster.granular-entry-heal: enable storage.owner-gid: 36 storage.owner-uid: 36 cluster.lookup-optimize: off server.keepalive-count: 5 server.keepalive-interval: 2 server.keepalive-time: 10 server.tcp-user-timeout: 20 network.ping-timeout: 30 server.event-threads: 4 client.event-threads: 4 cluster.choose-local: off features.shard: on cluster.shd-wait-qlength: 10000 cluster.shd-max-threads: 8 cluster.locking-scheme: granular cluster.data-self-heal-algorithm: full cluster.server-quorum-type: server cluster.quorum-type: auto cluster.eager-lock: enable performance.strict-o-direct: on network.remote-dio: off performance.low-prio-threads: 32 performance.io-cache: off performance.read-ahead: off performance.quick-read: off auth.allow: * user.cifs: off storage.fips-mode-rchecksum: on transport.address-family: inet nfs.disable: on performance.client-io-threads: on Regards, Martin -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20230929/78ff49c5/attachment.html>
Possibly Parallel Threads
- Gluster volume not mounted
- Gluster volume not mounted
- After restoring the failed host and synchronizing the data, it prompts that there are unsynchronized items
- Volume options appear twice
- Gluster very poor performance when copying small files (1x (2+1) = 3, SSD)