Hi,
I'm currently trying to use g?o-rep on the local data-node into a directory
but it fails with status "faulty"
Volume :
Volume Name: athena
Type: Distributed-Replicate
Status: Started
Number of Bricks: 2 x 2 = 4
Transport-type: tcp
Bricks:
Brick1: ylal3020:/users/exp1
Brick2: yval1010:/users/exp3
Brick3: ylal3030:/users/exp2
Brick4: yval1000:/users/exp4
Options Reconfigured:
geo-replication.indexing: on
diagnostics.count-fop-hits: on
diagnostics.latency-measurement: on
performance.cache-max-file-size: 256MB
network.ping-timeout: 5
performance.cache-size: 512MB
performance.cache-refresh-timeout: 60
nfs.port: 2049
I've done this cmd :
# gluster volume geo-replication athena /soft/venus config
# gluster volume geo-replication athena /soft/venus start
# gluster volume geo-replication athena /soft/venus status
MASTER SLAVE STATUS
--------------------------------------------------------------------------------
athena /soft/venus faulty
Here is the log file in Debug mod :
[2011-05-16 13:28:55.268006] I [monitor(monitor):42:monitor] Monitor:
------------------------------------------------------------
[2011-05-16 13:28:55.268281] I [monitor(monitor):43:monitor] Monitor: starting
gsyncd worker
[2011-05-16 13:28:55.326309] I [gsyncd:287:main_i] <top>: syncing:
gluster://localhost:athena -> file:///soft/venus
[2011-05-16 13:28:55.327905] D [repce:131:push] RepceClient: call
10888:47702589471600:1305545335.33 __repce_version__() ...
[2011-05-16 13:28:55.462613] D [repce:141:__call__] RepceClient: call
10888:47702589471600:1305545335.33 __repce_version__ -> 1.0
[2011-05-16 13:28:55.462886] D [repce:131:push] RepceClient: call
10888:47702589471600:1305545335.46 version() ...
[2011-05-16 13:28:55.463330] D [repce:141:__call__] RepceClient: call
10888:47702589471600:1305545335.46 version -> 1.0
[2011-05-16 13:28:55.480202] D [resource:381:connect] GLUSTER: auxiliary
glusterfs mount in place
[2011-05-16 13:28:55.682863] D [resource:393:connect] GLUSTER: auxiliary
glusterfs mount prepared
[2011-05-16 13:28:55.684926] D [monitor(monitor):57:monitor] Monitor: worker got
connected in 0 sec, waiting 59 more to make sure it's fine
[2011-05-16 13:28:55.685096] D [repce:131:push] RepceClient: call
10888:1115703616:1305545335.68 keep_alive(None,) ...
[2011-05-16 13:28:55.685859] D [repce:141:__call__] RepceClient: call
10888:1115703616:1305545335.68 keep_alive -> 1
[2011-05-16 13:28:59.546574] D [master:167:volinfo_state_machine] <top>:
(None, None) << (None, 28521f8f) -> (None, 28521f8f)
[2011-05-16 13:28:59.546863] I [master:184:crawl] GMaster: new master is
28521f8f-49d3-4e2a-b984-f664f44f5289
[2011-05-16 13:28:59.547034] I [master:191:crawl] GMaster: primary master with
volume id 28521f8f-49d3-4e2a-b984-f664f44f5289 ...
[2011-05-16 13:28:59.547180] D [master:199:crawl] GMaster: entering .
[2011-05-16 13:28:59.548289] D [repce:131:push] RepceClient: call
10888:47702589471600:1305545339.55 xtime('.',
'28521f8f-49d3-4e2a-b984-f664f44f5289') ...
[2011-05-16 13:28:59.596978] E [syncdutils:131:log_raise_exception] <top>:
FAIL:
Traceback (most recent call last):
File "/usr/local/libexec/glusterfs/python/syncdaemon/syncdutils.py",
line 152, in twrap
tf(*aa)
File "/usr/local/libexec/glusterfs/python/syncdaemon/repce.py", line
118, in listen
rid, exc, res = recv(self.inf)
File "/usr/local/libexec/glusterfs/python/syncdaemon/repce.py", line
42, in recv
return pickle.load(inf)
EOFError
Does anyone already got those errors ?
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
<http://supercolony.gluster.org/pipermail/gluster-users/attachments/20110516/62dc5154/attachment.html>
Hi -
Do you have passwordless ssh login to slave machine? After setting passwordless
login ,please try this -
#gluster volume geo-replication athena root@$(hostname):/soft/venus start
or
#gluster volume geo-replication athena $(hostname):/soft/venus start
wait for few seconds then verify the status.
For minimum requirement ,checkout this
http://www.gluster.com/community/documentation/index.php/Gluster_3.2:_Checking_Geo-replication_Minimum_Requirements
HTH
--
----
Cheers,
Lakshmipathi.G
FOSS Programmer.
----- Original Message -----
From: "anthony garnier" <sokar6012 at hotmail.com>
To: gluster-users at gluster.org
Sent: Monday, May 16, 2011 5:06:22 PM
Subject: [Gluster-users] G?o-rep fail
Hi,
I'm currently trying to use g?o-rep on the local data-node into a directory
but it fails with status "faulty"
Volume :
Volume Name: athena
Type: Distributed-Replicate
Status: Started
Number of Bricks: 2 x 2 = 4
Transport-type: tcp
Bricks:
Brick1: ylal3020:/users/exp1
Brick2: yval1010:/users/exp3
Brick3: ylal3030:/users/exp2
Brick4: yval1000:/users/exp4
Options Reconfigured:
geo-replication.indexing: on
diagnostics.count-fop-hits: on
diagnostics.latency-measurement: on
performance.cache-max-file-size: 256MB
network.ping-timeout: 5
performance.cache-size: 512MB
performance.cache-refresh-timeout: 60
nfs.port: 2049
I've done this cmd :
# gluster volume geo-replication athena /soft/venus config
# gluster volume geo-replication athena /soft/venus start
# gluster volume geo-replication athena /soft/venus status
MASTER SLAVE STATUS
--------------------------------------------------------------------------------
athena /soft/venus faulty
Here is the log file in Debug mod :
[2011-05-16 13:28:55.268006] I [monitor(monitor):42:monitor] Monitor:
------------------------------------------------------------
[2011-05-16 13:28:55.268281] I [monitor(monitor):43:monitor] Monitor: starting
gsyncd worker
[2011-05-16 13:28:55.326309] I [gsyncd:287:main_i] <top>: syncing:
gluster://localhost:athena -> file:///soft/venus
[2011-05-16 13:28:55.327905] D [repce:131:push] RepceClient: call
10888:47702589471600:1305545335.33 __repce_version__() ...
[2011-05-16 13:28:55.462613] D [repce:141:__call__] RepceClient: call
10888:47702589471600:1305545335.33 __repce_version__ -> 1.0
[2011-05-16 13:28:55.462886] D [repce:131:push] RepceClient: call
10888:47702589471600:1305545335.46 version() ...
[2011-05-16 13:28:55.463330] D [repce:141:__call__] RepceClient: call
10888:47702589471600:1305545335.46 version -> 1.0
[2011-05-16 13:28:55.480202] D [resource:381:connect] GLUSTER: auxiliary
glusterfs mount in place
[2011-05-16 13:28:55.682863] D [resource:393:connect] GLUSTER: auxiliary
glusterfs mount prepared
[2011-05-16 13:28:55.684926] D [monitor(monitor):57:monitor] Monitor: worker got
connected in 0 sec, waiting 59 more to make sure it's fine
[2011-05-16 13:28:55.685096] D [repce:131:push] RepceClient: call
10888:1115703616:1305545335.68 keep_alive(None,) ...
[2011-05-16 13:28:55.685859] D [repce:141:__call__] RepceClient: call
10888:1115703616:1305545335.68 keep_alive -> 1
[2011-05-16 13:28:59.546574] D [master:167:volinfo_state_machine] <top>:
(None, None) << (None, 28521f8f) -> (None, 28521f8f)
[2011-05-16 13:28:59.546863] I [master:184:crawl] GMaster: new master is
28521f8f-49d3-4e2a-b984-f664f44f5289
[2011-05-16 13:28:59.547034] I [master:191:crawl] GMaster: primary master with
volume id 28521f8f-49d3-4e2a-b984-f664f44f5289 ...
[2011-05-16 13:28:59.547180] D [master:199:crawl] GMaster: entering .
[2011-05-16 13:28:59.548289] D [repce:131:push] RepceClient: call
10888:47702589471600:1305545339.55 xtime('.',
'28521f8f-49d3-4e2a-b984-f664f44f5289') ...
[2011-05-16 13:28:59.596978] E [syncdutils:131:log_raise_exception] <top>:
FAIL:
Traceback (most recent call last):
File "/usr/local/libexec/glusterfs/python/syncdaemon/syncdutils.py",
line 152, in twrap
tf(*aa)
File "/usr/local/libexec/glusterfs/python/syncdaemon/repce.py", line
118, in listen
rid, exc, res = recv(self.inf)
File "/usr/local/libexec/glusterfs/python/syncdaemon/repce.py", line
42, in recv
return pickle.load(inf)
EOFError
Does anyone already got those errors ?
_______________________________________________
Gluster-users mailing list
Gluster-users at gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
On 05/16/11 17:06, anthony garnier wrote:> Hi, > I'm currently trying to use g?o-rep on the local data-node into a > directory but it fails with status "faulty"[...]> I've done this cmd : > # gluster volume geo-replication athena /soft/venus config > > # gluster volume geo-replication athena /soft/venus start > > # gluster volume geo-replication athena /soft/venus status > MASTER SLAVE STATUS > -------------------------------------------------------------------------------- > athena /soft/venus faulty > > > Here is the log file in Debug mod : > > [2011-05-16 13:28:55.268006] I [monitor(monitor):42:monitor] Monitor: > ------------------------------------------------------------ > [2011-05-16 13:28:55.268281] I [monitor(monitor):43:monitor] Monitor: > starting gsyncd worker[...]> [2011-05-16 13:28:59.547034] I [master:191:crawl] GMaster: primary > master with volume id 28521f8f-49d3-4e2a-b984-f664f44f5289 ... > [2011-05-16 13:28:59.547180] D [master:199:crawl] GMaster: entering . > [2011-05-16 13:28:59.548289] D [repce:131:push] RepceClient: call > 10888:47702589471600:1305545339.55 xtime('.', > '28521f8f-49d3-4e2a-b984-f664f44f5289') ... > [2011-05-16 13:28:59.596978] E [syncdutils:131:log_raise_exception] > <top>: FAIL: > Traceback (most recent call last): > File "/usr/local/libexec/glusterfs/python/syncdaemon/syncdutils.py", > line 152, in twrap > tf(*aa) > File "/usr/local/libexec/glusterfs/python/syncdaemon/repce.py", line > 118, in listen > rid, exc, res = recv(self.inf) > File "/usr/local/libexec/glusterfs/python/syncdaemon/repce.py", line 42, > in recv > return pickle.load(inf) > EOFError > > > Does anyone already got those errors ?This means slave gsyncd instance could not properly start up. To debug this further, we need to see the slave side logs. In your case, the following commands will set a debug log level for the slave (takes effect if done before starting the geo-replication session) and locate its log file: # gluster volume geo-replication /soft/venus config log-level DEBUG # gluster volume geo-replication /soft/venus config log-file The output of the latter will contain an unresolved parameter ${session-owner}. To get its actual value, run # gluster volume geo-replication athena /soft/venus config session-owner -- please post the content of the actual log file, path to which you get after the substitution. (Also, cf. http://www.gluster.com/community/documentation/index.php/Gluster_3.2:_Configuring_Geo-replication , slave-side logs are illustrated there.) Csaba