Strahil Nikolov
2019-Nov-09 17:37 UTC
[Gluster-users] Strange gluster behaviour after snapshot restore
Hello Community, today was the first time I had to rollback from a gluster snapshot. Here is what I did: 1. Killed the HostedEngine VM2. Stopped gluster volume3. Run 'gluster snapshot restore <snap-from-severam-minutes-ago>'4. Started my volume5. Created a new snapshot as the previous one was removed (according to docs - this is expected) Now I don't see my gluster bricks as before: '/gluster_bricks/engine/engine' but like this: # gluster volume info engine ? Volume Name: engine Type: Replicate Volume ID: 30ca1cc2-f2f7-4749-9e2e-cee9d7099ded Status: Started Snapshot Count: 3 Number of Bricks: 1 x (2 + 1) = 3 Transport-type: tcp Bricks: Brick1: gluster1:/run/gluster/snaps/abe6122484624d9f85dd89652fb8d207/brick1/engine Brick2: gluster2:/run/gluster/snaps/abe6122484624d9f85dd89652fb8d207/brick2/engine Brick3: ovirt3:/run/gluster/snaps/abe6122484624d9f85dd89652fb8d207/brick3/engine (arbiter) Options Reconfigured: features.barrier: disable cluster.choose-local: on performance.client-io-threads: off nfs.disable: on transport.address-family: inet performance.quick-read: off performance.read-ahead: off performance.io-cache: off performance.low-prio-threads: 32 network.remote-dio: on cluster.eager-lock: enable cluster.quorum-type: auto cluster.server-quorum-type: server cluster.data-self-heal-algorithm: full cluster.locking-scheme: granular cluster.shd-max-threads: 8 cluster.shd-wait-qlength: 10000 features.shard: on user.cifs: off storage.owner-uid: 36 storage.owner-gid: 36 network.ping-timeout: 30 performance.strict-o-direct: on cluster.granular-entry-heal: enable server.event-threads: 4 client.event-threads: 4 features.quota: off features.inode-quota: off features.quota-deem-statfs: off cluster.enable-shared-storage: enable Also, I have noticed 3 new features were added (quota,inode-quota,quota-deem-statfs). So why do I see my bricks as '/run/gluster/snaps/.../brickX/engine' ?I have created snapshots before and I never seen such behaviour. Best Regards,Strahil Nikolov -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20191109/745aa461/attachment.html>