Hello, I have tried to config a geo-replication volume , all the master nodes configuration are the same, When I start this volume, the status shows partial faulty as following: gluster volume geo-replication filews glusterfs01.sh3.ctripcorp.com::filews_slave status MASTER NODE MASTER VOL MASTER BRICK SLAVE STATUS CHECKPOINT STATUS CRAWL STATUS ------------------------------------------------------------------------------------------------------------------------------------------------- SVR8048HW2285 filews /export/sdb/filews glusterfs01.sh3.ctripcorp.com::filews_slave faulty N/A N/A SVR8050HW2285 filews /export/sdb/filews glusterfs03.sh3.ctripcorp.com::filews_slave Passive N/A N/A SVR8047HW2285 filews /export/sdb/filews glusterfs01.sh3.ctripcorp.com::filews_slave Active N/A Hybrid Crawl SVR8049HW2285 filews /export/sdb/filews glusterfs05.sh3.ctripcorp.com::filews_slave Active N/A Hybrid Crawl SH02SVR5951 filews /export/sdb/brick1 glusterfs06.sh3.ctripcorp.com::filews_slave Passive N/A N/A SH02SVR5953 filews /export/sdb/brick1 glusterfs01.sh3.ctripcorp.com::filews_slave faulty N/A N/A SVR6995HW2285 filews /export/sdb/filews glusterfs01.sh3.ctripcorp.com::filews_slave faulty N/A N/A SH02SVR5954 filews /export/sdb/brick1 glusterfs01.sh3.ctripcorp.com::filews_slave faulty N/A N/A SVR6994HW2285 filews /export/sdb/filews glusterfs02.sh3.ctripcorp.com::filews_slave Passive N/A N/A SVR6993HW2285 filews /export/sdb/filews glusterfs01.sh3.ctripcorp.com::filews_slave faulty N/A N/A SH02SVR5952 filews /export/sdb/brick1 glusterfs01.sh3.ctripcorp.com::filews_slave faulty N/A N/A SVR6996HW2285 filews /export/sdb/filews glusterfs04.sh3.ctripcorp.com::filews_slave Passive N/A N/A On the faulty node, log file /var/log/glusterfs/geo-replication/filews shows worker(/export/sdb/filews) died before establishing connection [2016-05-18 16:55:46.402622] I [monitor(monitor):215:monitor] Monitor: ------------------------------------------------------------ [2016-05-18 16:55:46.402930] I [monitor(monitor):216:monitor] Monitor: starting gsyncd worker [2016-05-18 16:55:46.517460] I [changelogagent(agent):72:__init__] ChangelogAgent: Agent listining... [2016-05-18 16:55:46.518066] I [repce(agent):92:service_loop] RepceServer: terminating on reaching EOF. [2016-05-18 16:55:46.518279] I [syncdutils(agent):214:finalize] <top>: exiting. [2016-05-18 16:55:46.518194] I [monitor(monitor):267:monitor] Monitor: worker(/export/sdb/filews) died before establishing connection [2016-05-18 16:55:56.697036] I [monitor(monitor):215:monitor] Monitor: ------------------------------------------------------------ Any advice and suggestions will be greatly appreciated. Best Regards ??? Yuyang Yang -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160519/0685b624/attachment.html>
Saravanakumar Arumugam
2016-May-19 05:58 UTC
[Gluster-users] 答复: geo-replication status partial faulty
Hi, There seems to be some issue in glusterfs01.sh3.ctripcorp.com slave node. Can you share the complete logs ? You can increase verbosity of debug messages like this: gluster volume geo-replication <master volume> <slave host>::<slave volume>config log-level DEBUG Also, check /root/.ssh/authorized_keys in glusterfs01.sh3.ctripcorp.com It should have entries in /var/lib/glusterd/geo-replication/common_secret.pem.pub (present in master node). Have a look at this one for example: https://www.gluster.org/pipermail/gluster-users/2015-August/023174.html Thanks, Saravana On 05/19/2016 07:53 AM, vyyy??? wrote:> > Hello, > > I have tried to config a geo-replication volume , all the master nodes > configuration are the same, When I start this volume, the status shows > partial faulty as following: > > gluster volume geo-replication filews > glusterfs01.sh3.ctripcorp.com::filews_slave status > > MASTER NODE MASTER VOL MASTER BRICK SLAVE STATUS > CHECKPOINT STATUS CRAWL STATUS > > ------------------------------------------------------------------------------------------------------------------------------------------------- > > SVR8048HW2285 filews /export/sdb/filews > glusterfs01.sh3.ctripcorp.com::filews_slave faulty > N/A N/A > > SVR8050HW2285 filews /export/sdb/filews > glusterfs03.sh3.ctripcorp.com::filews_slave Passive > N/A N/A > > SVR8047HW2285 filews /export/sdb/filews > glusterfs01.sh3.ctripcorp.com::filews_slave Active > N/A Hybrid Crawl > > SVR8049HW2285 filews /export/sdb/filews > glusterfs05.sh3.ctripcorp.com::filews_slave Active > N/A Hybrid Crawl > > SH02SVR5951 filews /export/sdb/brick1 > glusterfs06.sh3.ctripcorp.com::filews_slave Passive > N/A N/A > > SH02SVR5953 filews /export/sdb/brick1 > glusterfs01.sh3.ctripcorp.com::filews_slave faulty > N/A N/A > > SVR6995HW2285 filews /export/sdb/filews > glusterfs01.sh3.ctripcorp.com::filews_slave faulty > N/A N/A > > SH02SVR5954 filews /export/sdb/brick1 > glusterfs01.sh3.ctripcorp.com::filews_slave faulty > N/A N/A > > SVR6994HW2285 filews /export/sdb/filews > glusterfs02.sh3.ctripcorp.com::filews_slave Passive > N/A N/A > > SVR6993HW2285 filews /export/sdb/filews > glusterfs01.sh3.ctripcorp.com::filews_slave faulty > N/A N/A > > SH02SVR5952 filews /export/sdb/brick1 > glusterfs01.sh3.ctripcorp.com::filews_slave faulty > N/A N/A > > SVR6996HW2285 filews /export/sdb/filews > glusterfs04.sh3.ctripcorp.com::filews_slave Passive > N/A N/A > > On the faulty node, log file /var/log/glusterfs/geo-replication/filews > shows worker(/export/sdb/filews) died before establishing connection > > [2016-05-18 16:55:46.402622] I [monitor(monitor):215:monitor] Monitor: > ------------------------------------------------------------ > > [2016-05-18 16:55:46.402930] I [monitor(monitor):216:monitor] Monitor: > starting gsyncd worker > > [2016-05-18 16:55:46.517460] I [changelogagent(agent):72:__init__] > ChangelogAgent: Agent listining... > > [2016-05-18 16:55:46.518066] I [repce(agent):92:service_loop] > RepceServer: terminating on reaching EOF. > > [2016-05-18 16:55:46.518279] I [syncdutils(agent):214:finalize] <top>: > exiting. > > [2016-05-18 16:55:46.518194] I [monitor(monitor):267:monitor] Monitor: > worker(/export/sdb/filews) died before establishing connection > > [2016-05-18 16:55:56.697036] I [monitor(monitor):215:monitor] Monitor: > ------------------------------------------------------------ > > Any advice and suggestions will be greatly appreciated. > > Best Regards ** > > *??????**Yuyang Yang*** > > > > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://www.gluster.org/mailman/listinfo/gluster-users-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160519/5d7bd597/attachment.html>