Alexandr Porunov
2016-Nov-25 21:03 UTC
[Gluster-users] Geo-replication slaves are faulty after startup
Hello, I want to create geo replication between two volumes. Volumes works just fine. But geo-replication doesn't work at all. My master volume nodes are: 192.168.0.120 192.168.0.121 192.168.0.122 My slave volume nodes are: 192.168.0.123 192.168.0.124 192.168.0.125 My OS is: CentOS 7 I am running GlusterFS 3.8.5 Here is the status of geo-replication session: # gluster volume geo-replication gv0 geoaccount at 192.168.0.123::gv0 status MASTER NODE MASTER VOL MASTER BRICK SLAVE USER SLAVE SLAVE NODE STATUS CRAWL STATUS LAST_SYNCED ------------------------------------------------------------------------------------------------------------------------------------------------------------------ 192.168.0.120 gv0 /data/brick1/gv0 geoaccount geoaccount at 192.168.0.123::gv0 192.168.0.123 Active Changelog Crawl 2016-11-25 22:25:12 192.168.0.121 gv0 /data/brick1/gv0 geoaccount geoaccount at 192.168.0.123::gv0 N/A Faulty N/A N/A 192.168.0.122 gv0 /data/brick1/gv0 geoaccount geoaccount at 192.168.0.123::gv0 N/A Faulty N/A N/A I don't understand why it doesn't work. Here are interesting log files from the master node (192.168.0.120): /var/log/glusterfs/etc-glusterfs-glusterd.vol.log - http://paste.openstack.org/show/590503/ /var/log/glusterfs/mnt.log - http://paste.openstack.org/show/590504/ /var/log/glusterfs/run-gluster-shared_storage.log - http://paste.openstack.org/show/590505/ /var/log/glusterfs/geo-replication/gv0/ssh%3A%2F%2Fgeoaccount%40192.168.0.123%3Agluster%3A%2F%2F127.0.0.1%3Agv0.log - http://paste.openstack.org/show/590506/ Here is a log file from the slave node (192.168.0.123): /var/log/glusterfs/geo-replication-slaves/5afe64e3-d4e9-452b-a9cf-10674e052616\:gluster%3A%2F%2F127.0.0.1%3Agv0.gluster.log - http://paste.openstack.org/show/590507/ Here is how I have created a session: On slave nodes: useradd geoaccount groupadd geogroup usermod -a -G geogroup geoaccount usermod -a -G geogroup root passwd geoaccount mkdir -p /var/mountbroker-root chown root:root -R /var/mountbroker-root chmod 0711 -R /var/mountbroker-root chown root:geogroup -R /var/lib/glusterd/geo-replication/* chmod g=rwx,u=rwx,o-rwx -R /var/lib/glusterd/geo-replication/* On the slave (192.168.0.123): gluster system:: execute mountbroker opt mountbroker-root /var/mountbroker-root gluster system:: execute mountbroker opt geo-replication-log-group geogroup gluster system:: execute mountbroker opt rpc-auth-allow-insecure on gluster system:: execute mountbroker user geoaccount gv0 /usr/libexec/glusterfs/set_geo_rep_pem_keys.sh geoaccount gv0 gv0 gluster volume set all cluster.enable-shared-storage enable Then I have restarted all the slaves: systemctl restart glusterd On the master node (192.168.0.120): ssh-keygen ssh-copy-id geoaccount at 192.168.0.123 gluster system:: execute gsec_create container gluster volume set all cluster.enable-shared-storage enable gluster volume geo-replication gv0 geoaccount at 192.168.0.123::gv0 create ssh-port 22 push-pem gluster volume geo-replication gv0 geoaccount at 192.168.0.123::gv0 config remote-gsyncd /usr/libexec/glusterfs/gsyncd gluster volume geo-replication gv0 geoaccount at 192.168.0.123::gv0 config use-meta-volume true gluster volume geo-replication gv0 geoaccount at 192.168.0.123::gv0 config sync-jobs 3 gluster volume geo-replication gv0 geoaccount at 192.168.0.123::gv0 start Does somebody know what is wrong with this installation? I tried to install geo-replication for several times but without success.. Please help me Sincerely, Alexandr -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20161125/4c99eedc/attachment.html>
Aravinda
2016-Nov-28 08:14 UTC
[Gluster-users] Geo-replication slaves are faulty after startup
Following command should be run after the Geo-rep create push-pem step /usr/libexec/glusterfs/set_geo_rep_pem_keys.sh geoaccount gv0 gv0 Please run that command in any one of the slave node and observe the Geo-replication status. Admin guide: http://gluster.readthedocs.io/en/latest/Administrator%20Guide/Geo%20Replication/ regards Aravinda On 11/26/2016 02:33 AM, Alexandr Porunov wrote:> Hello, > > I want to create geo replication between two volumes. Volumes works > just fine. But geo-replication doesn't work at all. > > My master volume nodes are: > 192.168.0.120 > 192.168.0.121 > 192.168.0.122 > > My slave volume nodes are: > 192.168.0.123 > 192.168.0.124 > 192.168.0.125 > > My OS is: CentOS 7 > I am running GlusterFS 3.8.5 > > Here is the status of geo-replication session: > # gluster volume geo-replication gv0 geoaccount at 192.168.0.123::gv0 status > MASTER NODE MASTER VOL MASTER BRICK SLAVE USER SLAVE > SLAVE NODE STATUS CRAWL STATUS > LAST_SYNCED > ------------------------------------------------------------------------------------------------------------------------------------------------------------------ > 192.168.0.120 gv0 /data/brick1/gv0 geoaccount > geoaccount at 192.168.0.123::gv0 192.168.0.123 Active Changelog > Crawl 2016-11-25 22:25:12 > 192.168.0.121 gv0 /data/brick1/gv0 geoaccount > geoaccount at 192.168.0.123::gv0 N/A Faulty N/A N/A > 192.168.0.122 gv0 /data/brick1/gv0 geoaccount > geoaccount at 192.168.0.123::gv0 N/A Faulty N/A N/A > > > I don't understand why it doesn't work. Here are interesting log files > from the master node (192.168.0.120): > /var/log/glusterfs/etc-glusterfs-glusterd.vol.log - > http://paste.openstack.org/show/590503/ > > /var/log/glusterfs/mnt.log - http://paste.openstack.org/show/590504/ > > /var/log/glusterfs/run-gluster-shared_storage.log - > http://paste.openstack.org/show/590505/ > > /var/log/glusterfs/geo-replication/gv0/ssh%3A%2F%2Fgeoaccount%40192.168.0.123%3Agluster%3A%2F%2F127.0.0.1%3Agv0.log > - http://paste.openstack.org/show/590506/ > > Here is a log file from the slave node (192.168.0.123): > /var/log/glusterfs/geo-replication-slaves/5afe64e3-d4e9-452b-a9cf-10674e052616\:gluster%3A%2F%2F127.0.0.1%3Agv0.gluster.log > - http://paste.openstack.org/show/590507/ > > Here is how I have created a session: > On slave nodes: > useradd geoaccount > groupadd geogroup > usermod -a -G geogroup geoaccount > usermod -a -G geogroup root > passwd geoaccount > mkdir -p /var/mountbroker-root > chown root:root -R /var/mountbroker-root > chmod 0711 -R /var/mountbroker-root > chown root:geogroup -R /var/lib/glusterd/geo-replication/* > chmod g=rwx,u=rwx,o-rwx -R /var/lib/glusterd/geo-replication/* > > On the slave (192.168.0.123): > gluster system:: execute mountbroker opt mountbroker-root > /var/mountbroker-root > gluster system:: execute mountbroker opt geo-replication-log-group > geogroup > gluster system:: execute mountbroker opt rpc-auth-allow-insecure on > gluster system:: execute mountbroker user geoaccount gv0 > /usr/libexec/glusterfs/set_geo_rep_pem_keys.sh geoaccount gv0 gv0 > gluster volume set all cluster.enable-shared-storage enable > > Then I have restarted all the slaves: > systemctl restart glusterd > > On the master node (192.168.0.120): > ssh-keygen > ssh-copy-id geoaccount at 192.168.0.123 <mailto:geoaccount at 192.168.0.123> > gluster system:: execute gsec_create container > gluster volume set all cluster.enable-shared-storage enable > gluster volume geo-replication gv0 geoaccount at 192.168.0.123::gv0 > create ssh-port 22 push-pem > gluster volume geo-replication gv0 geoaccount at 192.168.0.123::gv0 > config remote-gsyncd /usr/libexec/glusterfs/gsyncd > gluster volume geo-replication gv0 geoaccount at 192.168.0.123::gv0 > config use-meta-volume true > gluster volume geo-replication gv0 geoaccount at 192.168.0.123::gv0 > config sync-jobs 3 > gluster volume geo-replication gv0 geoaccount at 192.168.0.123::gv0 start > > Does somebody know what is wrong with this installation? I tried to > install geo-replication for several times but without success.. Please > help me > > Sincerely, > Alexandr > > > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://www.gluster.org/mailman/listinfo/gluster-users-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20161128/78b573da/attachment.html>