Kingsley Tart
2023-Mar-14 19:31 UTC
[Gluster-users] can't set up geo-replication: can't fetch slave details
Hi, using Gluster 9.2 on debian 11 I'm trying to set up geo replication. I am following this guide: https://docs.gluster.org/en/main/Administrator-Guide/Geo-Replication/#password-less-ssh I have a volume called "ansible" which is only a small volume and seemed like an ideal test case. Firstly, for a bit of feedback (this isn't my issue as I worked around it) I had this problem with instructions from the above guide: root at glusterA:/data/brick# gluster volume geo-replication ansible geoaccount at glusterX::ansible create push-pem gluster command not found on glusterX for user geoaccount. geo-replication command failed Once I'd made a symlink to the gluster executable into the path for the geoaccount user, I ran the command again, and then got this error: root at glusterA:/data/brick# gluster volume geo-replication ansible geoaccount at glusterX::ansible create push-pem Unable to mount and fetch slave volume details. Please check the log: /var/log/glusterfs/geo-replication/gverify-slavemnt.log geo-replication command failed That log file contained this: [2023-03-14 19:13:48.904461 +0000] I [MSGID: 100030] [glusterfsd.c:2685:main] 0-glusterfs: Started running version [{arg=glusterfs}, {version=9.2}, {cmdlinestr=glusterfs --xlator- option=*dht.lookup-unhashed=off --volfile-server glusterX --volfile-id ansible -l /var/log/glusterfs/geo-replication/gverify-slavemnt.log /tmp/gverify.sh.txIgka}] [2023-03-14 19:13:48.905883 +0000] I [glusterfsd.c:2421:daemonize] 0- glusterfs: Pid of current running process is 3466942 [2023-03-14 19:13:48.912723 +0000] I [MSGID: 101190] [event- epoll.c:669:event_dispatch_epoll_worker] 0-epoll: Started thread with index [{index=1}] [2023-03-14 19:13:48.912759 +0000] I [MSGID: 101190] [event- epoll.c:669:event_dispatch_epoll_worker] 0-epoll: Started thread with index [{index=0}] [2023-03-14 19:13:48.914529 +0000] E [glusterfsd- mgmt.c:2137:mgmt_getspec_cbk] 0-glusterfs: failed to get the 'volume file' from server [2023-03-14 19:13:48.914549 +0000] E [glusterfsd- mgmt.c:2338:mgmt_getspec_cbk] 0-mgmt: failed to fetch volume file (key:ansible) [2023-03-14 19:13:48.914739 +0000] W [glusterfsd.c:1432:cleanup_and_exit] (-->/lib/x86_64-linux- gnu/libgfrpc.so.0(+0xfe3b) [0x7f1ca0a6ce3b] -->glusterfs(mgmt_getspec_cbk+0x762) [0x558ace8e8412] -- >glusterfs(cleanup_and_exit+0x57) [0x558ace8df4f7] ) 0-: receivedsignum (0), shutting down [2023-03-14 19:13:48.914767 +0000] I [fuse-bridge.c:7063:fini] 0-fuse: Unmounting '/tmp/gverify.sh.txIgka'. [2023-03-14 19:13:48.917289 +0000] I [fuse-bridge.c:7067:fini] 0-fuse: Closing fuse connection to '/tmp/gverify.sh.txIgka'. [2023-03-14 19:13:48.917358 +0000] W [glusterfsd.c:1432:cleanup_and_exit] (-->/lib/x86_64-linux- gnu/libpthread.so.0(+0x7ea7) [0x7f1ca0a28ea7] -->glusterfs(glusterfs_sigwaiter+0xc5) [0x558ace8e7175] -- >glusterfs(cleanup_and_exit+0x57) [0x558ace8df4f7] ) 0-: receivedsignum (15), shutting down In case it was a permissions issue, I copied the ssh pubkey to root's authorized_keys file on the replication slave and tried again with root@ but the error was the same. Where should I look next? Cheers, Kingsley. -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20230314/42f40228/attachment.html>
Kingsley Tart
2023-Mar-21 13:35 UTC
[Gluster-users] can't set up geo-replication: can't fetch slave details
Hi, is this a rare problem? Cheers, Kingsley. On Tue, 2023-03-14 at 19:31 +0000, Kingsley Tart wrote:> Hi, > > using Gluster 9.2 on debian 11 I'm trying to set up geo replication. > I am following this guide: > >https://docs.gluster.org/en/main/Administrator-Guide/Geo-Replication/#password-less-ssh> > I have a volume called "ansible" which is only a small volume and > seemed like an ideal test case. > > > Firstly, for a bit of feedback (this isn't my issue as I worked > around it) I had this problem with instructions from the above guide: > > root at glusterA:/data/brick# gluster volume geo-replication ansible > geoaccount at glusterX::ansible create push-pem > gluster command not found on glusterX for user geoaccount. > geo-replication command failed > > > Once I'd made a symlink to the gluster executable into the path for > the geoaccount user, I ran the command again, and then got this > error: > > root at glusterA:/data/brick# gluster volume geo-replication ansible > geoaccount at glusterX::ansible create push-pem > Unable to mount and fetch slave volume details. Please check the log: > /var/log/glusterfs/geo-replication/gverify-slavemnt.log > geo-replication command failed > > That log file contained this: > > [2023-03-14 19:13:48.904461 +0000] I [MSGID: 100030] > [glusterfsd.c:2685:main] 0-glusterfs: Started running version > [{arg=glusterfs}, {version=9.2}, {cmdlinestr=glusterfs --xlator- > option=*dht.lookup-unhashed=off --volfile-server glusterX --volfile- > id ansible -l /var/log/glusterfs/geo-replication/gverify-slavemnt.log > /tmp/gverify.sh.txIgka}] > [2023-03-14 19:13:48.905883 +0000] I [glusterfsd.c:2421:daemonize] 0- > glusterfs: Pid of current running process is 3466942 > [2023-03-14 19:13:48.912723 +0000] I [MSGID: 101190] [event- > epoll.c:669:event_dispatch_epoll_worker] 0-epoll: Started thread with > index [{index=1}] > [2023-03-14 19:13:48.912759 +0000] I [MSGID: 101190] [event- > epoll.c:669:event_dispatch_epoll_worker] 0-epoll: Started thread with > index [{index=0}] > [2023-03-14 19:13:48.914529 +0000] E [glusterfsd- > mgmt.c:2137:mgmt_getspec_cbk] 0-glusterfs: failed to get the 'volume > file' from server > [2023-03-14 19:13:48.914549 +0000] E [glusterfsd- > mgmt.c:2338:mgmt_getspec_cbk] 0-mgmt: failed to fetch volume file > (key:ansible) > [2023-03-14 19:13:48.914739 +0000] W > [glusterfsd.c:1432:cleanup_and_exit] (-->/lib/x86_64-linux- > gnu/libgfrpc.so.0(+0xfe3b) [0x7f1ca0a6ce3b] -- > >glusterfs(mgmt_getspec_cbk+0x762) [0x558ace8e8412] -- > >glusterfs(cleanup_and_exit+0x57) [0x558ace8df4f7] ) 0-: received > signum (0), shutting down > [2023-03-14 19:13:48.914767 +0000] I [fuse-bridge.c:7063:fini] 0- > fuse: Unmounting '/tmp/gverify.sh.txIgka'. > [2023-03-14 19:13:48.917289 +0000] I [fuse-bridge.c:7067:fini] 0- > fuse: Closing fuse connection to '/tmp/gverify.sh.txIgka'. > [2023-03-14 19:13:48.917358 +0000] W > [glusterfsd.c:1432:cleanup_and_exit] (-->/lib/x86_64-linux- > gnu/libpthread.so.0(+0x7ea7) [0x7f1ca0a28ea7] -- > >glusterfs(glusterfs_sigwaiter+0xc5) [0x558ace8e7175] -- > >glusterfs(cleanup_and_exit+0x57) [0x558ace8df4f7] ) 0-: received > signum (15), shutting down > > > In case it was a permissions issue, I copied the ssh pubkey to root's > authorized_keys file on the replication slave and tried again with > root@ but the error was the same. > > Where should I look next? > > Cheers, > Kingsley. > ________ > > > > Community Meeting Calendar: > > Schedule - > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC > Bridge: https://meet.google.com/cpu-eiue-hvk > Gluster-users mailing list > Gluster-users at gluster.org > https://lists.gluster.org/mailman/listinfo/gluster-users-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20230321/ae46f9d6/attachment.html>