Hi, I'm currently trying to use g?o-rep on the local data-node into a directory but it fails with status "faulty" Volume : Volume Name: athena Type: Distributed-Replicate Status: Started Number of Bricks: 2 x 2 = 4 Transport-type: tcp Bricks: Brick1: ylal3020:/users/exp1 Brick2: yval1010:/users/exp3 Brick3: ylal3030:/users/exp2 Brick4: yval1000:/users/exp4 Options Reconfigured: geo-replication.indexing: on diagnostics.count-fop-hits: on diagnostics.latency-measurement: on performance.cache-max-file-size: 256MB network.ping-timeout: 5 performance.cache-size: 512MB performance.cache-refresh-timeout: 60 nfs.port: 2049 I've done this cmd : # gluster volume geo-replication athena /soft/venus config # gluster volume geo-replication athena /soft/venus start # gluster volume geo-replication athena /soft/venus status MASTER SLAVE STATUS -------------------------------------------------------------------------------- athena /soft/venus faulty Here is the log file in Debug mod : [2011-05-16 13:28:55.268006] I [monitor(monitor):42:monitor] Monitor: ------------------------------------------------------------ [2011-05-16 13:28:55.268281] I [monitor(monitor):43:monitor] Monitor: starting gsyncd worker [2011-05-16 13:28:55.326309] I [gsyncd:287:main_i] <top>: syncing: gluster://localhost:athena -> file:///soft/venus [2011-05-16 13:28:55.327905] D [repce:131:push] RepceClient: call 10888:47702589471600:1305545335.33 __repce_version__() ... [2011-05-16 13:28:55.462613] D [repce:141:__call__] RepceClient: call 10888:47702589471600:1305545335.33 __repce_version__ -> 1.0 [2011-05-16 13:28:55.462886] D [repce:131:push] RepceClient: call 10888:47702589471600:1305545335.46 version() ... [2011-05-16 13:28:55.463330] D [repce:141:__call__] RepceClient: call 10888:47702589471600:1305545335.46 version -> 1.0 [2011-05-16 13:28:55.480202] D [resource:381:connect] GLUSTER: auxiliary glusterfs mount in place [2011-05-16 13:28:55.682863] D [resource:393:connect] GLUSTER: auxiliary glusterfs mount prepared [2011-05-16 13:28:55.684926] D [monitor(monitor):57:monitor] Monitor: worker got connected in 0 sec, waiting 59 more to make sure it's fine [2011-05-16 13:28:55.685096] D [repce:131:push] RepceClient: call 10888:1115703616:1305545335.68 keep_alive(None,) ... [2011-05-16 13:28:55.685859] D [repce:141:__call__] RepceClient: call 10888:1115703616:1305545335.68 keep_alive -> 1 [2011-05-16 13:28:59.546574] D [master:167:volinfo_state_machine] <top>: (None, None) << (None, 28521f8f) -> (None, 28521f8f) [2011-05-16 13:28:59.546863] I [master:184:crawl] GMaster: new master is 28521f8f-49d3-4e2a-b984-f664f44f5289 [2011-05-16 13:28:59.547034] I [master:191:crawl] GMaster: primary master with volume id 28521f8f-49d3-4e2a-b984-f664f44f5289 ... [2011-05-16 13:28:59.547180] D [master:199:crawl] GMaster: entering . [2011-05-16 13:28:59.548289] D [repce:131:push] RepceClient: call 10888:47702589471600:1305545339.55 xtime('.', '28521f8f-49d3-4e2a-b984-f664f44f5289') ... [2011-05-16 13:28:59.596978] E [syncdutils:131:log_raise_exception] <top>: FAIL: Traceback (most recent call last): File "/usr/local/libexec/glusterfs/python/syncdaemon/syncdutils.py", line 152, in twrap tf(*aa) File "/usr/local/libexec/glusterfs/python/syncdaemon/repce.py", line 118, in listen rid, exc, res = recv(self.inf) File "/usr/local/libexec/glusterfs/python/syncdaemon/repce.py", line 42, in recv return pickle.load(inf) EOFError Does anyone already got those errors ? -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20110516/62dc5154/attachment.html>
Hi - Do you have passwordless ssh login to slave machine? After setting passwordless login ,please try this - #gluster volume geo-replication athena root@$(hostname):/soft/venus start or #gluster volume geo-replication athena $(hostname):/soft/venus start wait for few seconds then verify the status. For minimum requirement ,checkout this http://www.gluster.com/community/documentation/index.php/Gluster_3.2:_Checking_Geo-replication_Minimum_Requirements HTH -- ---- Cheers, Lakshmipathi.G FOSS Programmer. ----- Original Message ----- From: "anthony garnier" <sokar6012 at hotmail.com> To: gluster-users at gluster.org Sent: Monday, May 16, 2011 5:06:22 PM Subject: [Gluster-users] G?o-rep fail Hi, I'm currently trying to use g?o-rep on the local data-node into a directory but it fails with status "faulty" Volume : Volume Name: athena Type: Distributed-Replicate Status: Started Number of Bricks: 2 x 2 = 4 Transport-type: tcp Bricks: Brick1: ylal3020:/users/exp1 Brick2: yval1010:/users/exp3 Brick3: ylal3030:/users/exp2 Brick4: yval1000:/users/exp4 Options Reconfigured: geo-replication.indexing: on diagnostics.count-fop-hits: on diagnostics.latency-measurement: on performance.cache-max-file-size: 256MB network.ping-timeout: 5 performance.cache-size: 512MB performance.cache-refresh-timeout: 60 nfs.port: 2049 I've done this cmd : # gluster volume geo-replication athena /soft/venus config # gluster volume geo-replication athena /soft/venus start # gluster volume geo-replication athena /soft/venus status MASTER SLAVE STATUS -------------------------------------------------------------------------------- athena /soft/venus faulty Here is the log file in Debug mod : [2011-05-16 13:28:55.268006] I [monitor(monitor):42:monitor] Monitor: ------------------------------------------------------------ [2011-05-16 13:28:55.268281] I [monitor(monitor):43:monitor] Monitor: starting gsyncd worker [2011-05-16 13:28:55.326309] I [gsyncd:287:main_i] <top>: syncing: gluster://localhost:athena -> file:///soft/venus [2011-05-16 13:28:55.327905] D [repce:131:push] RepceClient: call 10888:47702589471600:1305545335.33 __repce_version__() ... [2011-05-16 13:28:55.462613] D [repce:141:__call__] RepceClient: call 10888:47702589471600:1305545335.33 __repce_version__ -> 1.0 [2011-05-16 13:28:55.462886] D [repce:131:push] RepceClient: call 10888:47702589471600:1305545335.46 version() ... [2011-05-16 13:28:55.463330] D [repce:141:__call__] RepceClient: call 10888:47702589471600:1305545335.46 version -> 1.0 [2011-05-16 13:28:55.480202] D [resource:381:connect] GLUSTER: auxiliary glusterfs mount in place [2011-05-16 13:28:55.682863] D [resource:393:connect] GLUSTER: auxiliary glusterfs mount prepared [2011-05-16 13:28:55.684926] D [monitor(monitor):57:monitor] Monitor: worker got connected in 0 sec, waiting 59 more to make sure it's fine [2011-05-16 13:28:55.685096] D [repce:131:push] RepceClient: call 10888:1115703616:1305545335.68 keep_alive(None,) ... [2011-05-16 13:28:55.685859] D [repce:141:__call__] RepceClient: call 10888:1115703616:1305545335.68 keep_alive -> 1 [2011-05-16 13:28:59.546574] D [master:167:volinfo_state_machine] <top>: (None, None) << (None, 28521f8f) -> (None, 28521f8f) [2011-05-16 13:28:59.546863] I [master:184:crawl] GMaster: new master is 28521f8f-49d3-4e2a-b984-f664f44f5289 [2011-05-16 13:28:59.547034] I [master:191:crawl] GMaster: primary master with volume id 28521f8f-49d3-4e2a-b984-f664f44f5289 ... [2011-05-16 13:28:59.547180] D [master:199:crawl] GMaster: entering . [2011-05-16 13:28:59.548289] D [repce:131:push] RepceClient: call 10888:47702589471600:1305545339.55 xtime('.', '28521f8f-49d3-4e2a-b984-f664f44f5289') ... [2011-05-16 13:28:59.596978] E [syncdutils:131:log_raise_exception] <top>: FAIL: Traceback (most recent call last): File "/usr/local/libexec/glusterfs/python/syncdaemon/syncdutils.py", line 152, in twrap tf(*aa) File "/usr/local/libexec/glusterfs/python/syncdaemon/repce.py", line 118, in listen rid, exc, res = recv(self.inf) File "/usr/local/libexec/glusterfs/python/syncdaemon/repce.py", line 42, in recv return pickle.load(inf) EOFError Does anyone already got those errors ? _______________________________________________ Gluster-users mailing list Gluster-users at gluster.org http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
On 05/16/11 17:06, anthony garnier wrote:> Hi, > I'm currently trying to use g?o-rep on the local data-node into a > directory but it fails with status "faulty"[...]> I've done this cmd : > # gluster volume geo-replication athena /soft/venus config > > # gluster volume geo-replication athena /soft/venus start > > # gluster volume geo-replication athena /soft/venus status > MASTER SLAVE STATUS > -------------------------------------------------------------------------------- > athena /soft/venus faulty > > > Here is the log file in Debug mod : > > [2011-05-16 13:28:55.268006] I [monitor(monitor):42:monitor] Monitor: > ------------------------------------------------------------ > [2011-05-16 13:28:55.268281] I [monitor(monitor):43:monitor] Monitor: > starting gsyncd worker[...]> [2011-05-16 13:28:59.547034] I [master:191:crawl] GMaster: primary > master with volume id 28521f8f-49d3-4e2a-b984-f664f44f5289 ... > [2011-05-16 13:28:59.547180] D [master:199:crawl] GMaster: entering . > [2011-05-16 13:28:59.548289] D [repce:131:push] RepceClient: call > 10888:47702589471600:1305545339.55 xtime('.', > '28521f8f-49d3-4e2a-b984-f664f44f5289') ... > [2011-05-16 13:28:59.596978] E [syncdutils:131:log_raise_exception] > <top>: FAIL: > Traceback (most recent call last): > File "/usr/local/libexec/glusterfs/python/syncdaemon/syncdutils.py", > line 152, in twrap > tf(*aa) > File "/usr/local/libexec/glusterfs/python/syncdaemon/repce.py", line > 118, in listen > rid, exc, res = recv(self.inf) > File "/usr/local/libexec/glusterfs/python/syncdaemon/repce.py", line 42, > in recv > return pickle.load(inf) > EOFError > > > Does anyone already got those errors ?This means slave gsyncd instance could not properly start up. To debug this further, we need to see the slave side logs. In your case, the following commands will set a debug log level for the slave (takes effect if done before starting the geo-replication session) and locate its log file: # gluster volume geo-replication /soft/venus config log-level DEBUG # gluster volume geo-replication /soft/venus config log-file The output of the latter will contain an unresolved parameter ${session-owner}. To get its actual value, run # gluster volume geo-replication athena /soft/venus config session-owner -- please post the content of the actual log file, path to which you get after the substitution. (Also, cf. http://www.gluster.com/community/documentation/index.php/Gluster_3.2:_Configuring_Geo-replication , slave-side logs are illustrated there.) Csaba