I?ve rebuilt the cluster again, making a fresh installation. And now the error is different. MASTER NODE MASTER VOL MASTER BRICK SLAVE USER SLAVE SLAVE NODE STATUS CRAWL STATUS LAST_SYNCED ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- master-host01.me.com geotest /gpool/brick03/geotest guser guser at slave-host01::geotestdr N/A Faulty N/A N/A master-host02.me.com geotest /gpool/brick03/geotest guser guser at slave-host01::geotestdr slave-host01 Passive N/A N/A master-host03.me.com geotest /gpool/brick03/geotest guser guser at slave-host01::geotestdr slave-host03 Passive N/A N/A [2016-03-30 22:09:31.326898] I [monitor(monitor):221:monitor] Monitor: ------------------------------------------------------------ [2016-03-30 22:09:31.327461] I [monitor(monitor):222:monitor] Monitor: starting gsyncd worker [2016-03-30 22:09:31.544631] I [gsyncd(/gpool/brick03/geotest):649:main_i] <top>: syncing: gluster://localhost:geotest -> ssh://guser at slave-host02:gluster://localhost:geotestdr [2016-03-30 22:09:31.547542] I [changelogagent(agent):75:__init__] ChangelogAgent: Agent listining... [2016-03-30 22:09:31.830554] E [syncdutils(/gpool/brick03/geotest):252:log_raise_exception] <top>: connection to peer is broken [2016-03-30 22:09:31.831017] W [syncdutils(/gpool/brick03/geotest):256:log_raise_exception] <top>: !!!!!!!!!!!!! [2016-03-30 22:09:31.831258] W [syncdutils(/gpool/brick03/geotest):257:log_raise_exception] <top>: !!! getting "No such file or directory" errors is most likely due to MISCONFIGURATION, please consult https://access.redhat.com/site/documentation/en-US/Red_Hat_Storage/2.1/html/Administration_Guide/chap-User_Guide-Geo_Rep-Preparation-Settingup_Environment.html [2016-03-30 22:09:31.831502] W [syncdutils(/gpool/brick03/geotest):265:log_raise_exception] <top>: !!!!!!!!!!!!! [2016-03-30 22:09:31.836395] E [resource(/gpool/brick03/geotest):222:errlog] Popen: command "ssh -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i /var/lib/glusterd/geo-replication/secret.pem -oControlMaster=auto -S /tmp/gsyncd-aux-ssh-SfXvbB/de372ce5774b5d259c58c5c9522ffc8f.sock guser at slave-host02 /nonexistent/gsyncd --session-owner ec473e17-b933-4bf7-9eed-4c393f7aaf5d -N --listen --timeout 120 gluster://localhost:geotestdr" returned with 127, saying: [2016-03-30 22:09:31.836694] E [resource(/gpool/brick03/geotest):226:logerr] Popen: ssh> bash: /nonexistent/gsyncd: No such file or directory [2016-03-30 22:09:31.837193] I [syncdutils(/gpool/brick03/geotest):220:finalize] <top>: exiting. [2016-03-30 22:09:31.840569] I [repce(agent):92:service_loop] RepceServer: terminating on reaching EOF. [2016-03-30 22:09:31.840993] I [syncdutils(agent):220:finalize] <top>: exiting. [2016-03-30 22:09:31.840742] I [monitor(monitor):274:monitor] Monitor: worker(/gpool/brick03/geotest) died before establishing connection [2016-03-30 22:09:42.130866] I [monitor(monitor):221:monitor] Monitor: ------------------------------------------------------------ [2016-03-30 22:09:42.131448] I [monitor(monitor):222:monitor] Monitor: starting gsyncd worker [2016-03-30 22:09:42.348165] I [gsyncd(/gpool/brick03/geotest):649:main_i] <top>: syncing: gluster://localhost:geotest -> ssh://guser at slave-host02:gluster://localhost:geotestdr [2016-03-30 22:09:42.349118] I [changelogagent(agent):75:__init__] ChangelogAgent: Agent listining... [2016-03-30 22:09:42.653141] E [syncdutils(/gpool/brick03/geotest):252:log_raise_exception] <top>: connection to peer is broken [2016-03-30 22:09:42.653656] W [syncdutils(/gpool/brick03/geotest):256:log_raise_exception] <top>: !!!!!!!!!!!!! [2016-03-30 22:09:42.653898] W [syncdutils(/gpool/brick03/geotest):257:log_raise_exception] <top>: !!! getting "No such file or directory" errors is most likely due to MISCONFIGURATION, please consult https://access.redhat.com/site/documentation/en-US/Red_Hat_Storage/2.1/html/Administration_Guide/chap-User_Guide-Geo_Rep-Preparation-Settingup_Environment.html [2016-03-30 22:09:42.654129] W [syncdutils(/gpool/brick03/geotest):265:log_raise_exception] <top>: !!!!!!!!!!!!! [2016-03-30 22:09:42.659329] E [resource(/gpool/brick03/geotest):222:errlog] Popen: command "ssh -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i /var/lib/glusterd/geo-replication/secret.pem -oControlMaster=auto -S /tmp/gsyncd-aux-ssh-6r8rxx/de372ce5774b5d259c58c5c9522ffc8f.sock guser at slave-host02 /nonexistent/gsyncd --session-owner ec473e17-b933-4bf7-9eed-4c393f7aaf5d -N --listen --timeout 120 gluster://localhost:geotestdr" returned with 127, saying: [2016-03-30 22:09:42.659626] E [resource(/gpool/brick03/geotest):226:logerr] Popen: ssh> bash: /nonexistent/gsyncd: No such file or directory [2016-03-30 22:09:42.660140] I [syncdutils(/gpool/brick03/geotest):220:finalize] <top>: exiting. [2016-03-30 22:09:42.662802] I [repce(agent):92:service_loop] RepceServer: terminating on reaching EOF. [2016-03-30 22:09:42.663197] I [syncdutils(agent):220:finalize] <top>: exiting. [2016-03-30 22:09:42.663024] I [monitor(monitor):274:monitor] Monitor: worker(/gpool/brick03/geotest) died before establishing connection ?Bishoy> On Mar 30, 2016, at 10:50 AM, Gmail <b.s.mikhael at gmail.com> wrote: > > I?ve tried changing the permissions to 777 on /var/log/glusterfs on all the slave nodes, but still no luck :( > > here is the log from the master node where I created and started the geo-replication session. > > [2016-03-30 17:14:53.463150] I [monitor(monitor):221:monitor] Monitor: ------------------------------------------------------------ > [2016-03-30 17:14:53.463669] I [monitor(monitor):222:monitor] Monitor: starting gsyncd worker > [2016-03-30 17:14:53.603774] I [changelogagent(agent):75:__init__] ChangelogAgent: Agent listining... > [2016-03-30 17:14:53.604080] I [gsyncd(/mnt/brick10/xfsvol2):649:main_i] <top>: syncing: gluster://localhost:xfsvol2 <gluster://localhost:xfsvol2> -> ssh://guser at slave-host01:gluster://localhost:xfsvol2dr <ssh://guser at slave-host01:gluster://localhost:xfsvol2dr> > [2016-03-30 17:14:54.210602] E [syncdutils(/mnt/brick10/xfsvol2):252:log_raise_exception] <top>: connection to peer is broken > [2016-03-30 17:14:54.211117] E [resource(/mnt/brick10/xfsvol2):222:errlog] Popen: command "ssh -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i /var/lib/glusterd/geo-replication/secret.pem -oControlMaster=auto -S /tmp/gsyncd-aux-ssh-evONxc/3bda60dc6e900c0833fed4e4fdfbd480.sock guser at slave-host01 /nonexistent/gsyncd --session-owner ef9ccae5-0def-4a47-9a96-881a1896755c -N --listen --timeout 120 gluster://localhost:xfsvol2dr <gluster://localhost:xfsvol2dr>" returned with 1, saying: > [2016-03-30 17:14:54.211376] E [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> [2016-03-30 17:14:53.933174] I [cli.c:720:main] 0-cli: Started running /usr/sbin/gluster with version 3.7.3 > [2016-03-30 17:14:54.211631] E [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> [2016-03-30 17:14:53.933225] I [cli.c:608:cli_rpc_init] 0-cli: Connecting to remote glusterd at localhost > [2016-03-30 17:14:54.211828] E [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> [2016-03-30 17:14:54.074207] I [MSGID: 101190] [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread with index 1 > [2016-03-30 17:14:54.212017] E [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> [2016-03-30 17:14:54.074302] I [socket.c:2409:socket_event_handler] 0-transport: disconnecting now > [2016-03-30 17:14:54.212199] E [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> [2016-03-30 17:14:54.077207] I [cli-rpc-ops.c:6230:gf_cli_getwd_cbk] 0-cli: Received resp to getwd > [2016-03-30 17:14:54.212380] E [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> [2016-03-30 17:14:54.077269] I [input.c:36:cli_batch] 0-: Exiting with: 0 > [2016-03-30 17:14:54.212584] E [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> ERROR:root:FAIL: > [2016-03-30 17:14:54.212774] E [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> Traceback (most recent call last): > [2016-03-30 17:14:54.212954] E [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 165, in main > [2016-03-30 17:14:54.213131] E [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> main_i() > [2016-03-30 17:14:54.213308] E [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 638, in main_i > [2016-03-30 17:14:54.213500] E [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> startup(go_daemon=go_daemon, log_file=log_file, label=label) > [2016-03-30 17:14:54.213690] E [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 144, in startup > [2016-03-30 17:14:54.213890] E [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> GLogger._gsyncd_loginit(**kw) > [2016-03-30 17:14:54.214068] E [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 107, in _gsyncd_loginit > [2016-03-30 17:14:54.214246] E [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> cls.setup(label=kw.get('label'), **lkw) > [2016-03-30 17:14:54.214422] E [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 79, in setup > [2016-03-30 17:14:54.214622] E [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> logging_handler = handlers.WatchedFileHandler(lprm['filename']) > [2016-03-30 17:14:54.214802] E [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> File "/usr/lib64/python2.6/logging/handlers.py", line 377, in __init__ > [2016-03-30 17:14:54.214977] E [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> logging.FileHandler.__init__(self, filename, mode, encoding, delay) > [2016-03-30 17:14:54.215152] E [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> File "/usr/lib64/python2.6/logging/__init__.py", line 835, in __init__ > [2016-03-30 17:14:54.215327] E [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> StreamHandler.__init__(self, self._open()) > [2016-03-30 17:14:54.215523] E [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> File "/usr/lib64/python2.6/logging/__init__.py", line 854, in _open > [2016-03-30 17:14:54.215703] E [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> stream = open(self.baseFilename, self.mode) > [2016-03-30 17:14:54.215883] E [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> IOError: [Errno 13] Permission denied: '/var/log/glusterfs/geo-replication-slaves/mbr/ef9ccae5-0def-4a47-9a96-881a1896755c:gluster%3A%2F%2F127.0.0.1%3Axfsvol2dr.log' > [2016-03-30 17:14:54.216063] E [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> failed with IOError. > [2016-03-30 17:14:54.216500] I [syncdutils(/mnt/brick10/xfsvol2):220:finalize] <top>: exiting. > [2016-03-30 17:14:54.218672] I [repce(agent):92:service_loop] RepceServer: terminating on reaching EOF. > [2016-03-30 17:14:54.219063] I [syncdutils(agent):220:finalize] <top>: exiting. > [2016-03-30 17:14:54.218930] I [monitor(monitor):274:monitor] Monitor: worker(/mnt/brick10/xfsvol2) died before establishing connection > > ?Bishoy > >> On Mar 29, 2016, at 1:05 AM, Aravinda <avishwan at redhat.com <mailto:avishwan at redhat.com>> wrote: >> >> Geo-replication command should be run as privileged user itself. >> >> gluster volume geo-replication <MASTERVOL> <SLAVEUSER>@<SLAVEHOST> start >> >> and then check the status, if it shows Faulty then please share the log files present in /var/log/glusterfs/geo-replication/<MASTERVOL>/*.log >> >> regards >> Aravinda >> On 03/29/2016 12:51 PM, Gmail wrote: >>> I?ve been trying to setup geo-replication using Gluster 3.7.3 on OEL 6.5 >>> It keeps giving me faulty session. >>> I?ve tried to use root user instead, it works fine! >>> >>> I?ve followed literally the documentation but no luck getting the unprivileged user working. >>> >>> I?ve tried running /usr/libexec/glusterfs/gsyncd on the slave node using the unprivileged user, and that?s what I get. >>> >>> /usr/libexec/glusterfs/gsyncd --session-owner ef9ccae5-0def-4a47-9a96-881a1896755c -N --listen --timeout 120 gluster://localhost:vol01dr <gluster://localhost:vol01dr> >>> [2016-03-29 00:52:49.058244] I [cli.c:720:main] 0-cli: Started running /usr/sbin/gluster with version 3.7.3 >>> [2016-03-29 00:52:49.058297] I [cli.c:608:cli_rpc_init] 0-cli: Connecting to remote glusterd at localhost >>> [2016-03-29 00:52:49.174686] I [MSGID: 101190] [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread with index 1 >>> [2016-03-29 00:52:49.174768] I [socket.c:2409:socket_event_handler] 0-transport: disconnecting now >>> [2016-03-29 00:52:49.177482] I [cli-rpc-ops.c:6230:gf_cli_getwd_cbk] 0-cli: Received resp to getwd >>> [2016-03-29 00:52:49.177545] I [input.c:36:cli_batch] 0-: Exiting with: 0 >>> ERROR:root:FAIL: >>> Traceback (most recent call last): >>> File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 165, in main >>> main_i() >>> File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 638, in main_i >>> startup(go_daemon=go_daemon, log_file=log_file, label=label) >>> File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 144, in startup >>> GLogger._gsyncd_loginit(**kw) >>> File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 107, in _gsyncd_loginit >>> cls.setup(label=kw.get('label'), **lkw) >>> File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 79, in setup >>> logging_handler = handlers.WatchedFileHandler(lprm['filename']) >>> File "/usr/lib64/python2.6/logging/handlers.py", line 377, in __init__ >>> logging.FileHandler.__init__(self, filename, mode, encoding, delay) >>> File "/usr/lib64/python2.6/logging/__init__.py", line 835, in __init__ >>> StreamHandler.__init__(self, self._open()) >>> File "/usr/lib64/python2.6/logging/__init__.py", line 854, in _open >>> stream = open(self.baseFilename, self.mode) >>> IOError: [Errno 13] Permission denied: '/var/log/glusterfs/geo-replication-slaves/mbr/ef9ccae5-0def-4a47-9a96-881a1896755c:gluster%3A%2F%2F127.0.0.1%3Avol01dr.log' >>> failed with IOError. >>> >>> >>> ? Bishoy >>> >>> >>> >>> _______________________________________________ >>> Gluster-users mailing list >>> Gluster-users at gluster.org <mailto:Gluster-users at gluster.org> >>> http://www.gluster.org/mailman/listinfo/gluster-users <http://www.gluster.org/mailman/listinfo/gluster-users> >-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160330/298761a6/attachment.html>
Saravanakumar Arumugam
2016-Mar-31 06:33 UTC
[Gluster-users] geo-replication unprivileged user error
Hi, Replies inline. Thanks, Saravana On 03/31/2016 04:00 AM, Gmail wrote:> I?ve rebuilt the cluster again, making a fresh installation. And now > the error is different. > > > > > > MASTER NODE MASTER VOL MASTER BRICK SLAVE USER > SLAVE SLAVE NODE STATUS CRAWL > STATUS LAST_SYNCED > ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- > master-host01.me.com <http://master-host01.me.com> geotest > /gpool/brick03/geotest guser guser at slave-host01::geotestdr N/A > Faulty N/A N/A > master-host02.me.com <http://master-host02.me.com> geotest > /gpool/brick03/geotest guser guser at slave-host01::geotestdr > slave-host01 Passive N/A N/A > master-host03.me.com <http://master-host03.me.com> geotest > /gpool/brick03/geotest guser guser at slave-host01::geotestdr > slave-host03 Passive N/A N/A >There seems to issue with geo-rep setup. - All the master bricks seems same..which should not be the case. What type of volume is this? Can you get "gluster volume status" and "gluster volume info" for both master and slave volume? Also, share all the commands you execute to setup this georep session ?> > > > > > [2016-03-30 22:09:31.326898] I [monitor(monitor):221:monitor] Monitor: > ------------------------------------------------------------ > [2016-03-30 22:09:31.327461] I [monitor(monitor):222:monitor] Monitor: > starting gsyncd worker > [2016-03-30 22:09:31.544631] I > [gsyncd(/gpool/brick03/geotest):649:main_i] <top>: syncing: > gluster://localhost:geotest -> > ssh://guser at slave-host02:gluster://localhost:geotestdr > [2016-03-30 22:09:31.547542] I [changelogagent(agent):75:__init__] > ChangelogAgent: Agent listining... > [2016-03-30 22:09:31.830554] E > [syncdutils(/gpool/brick03/geotest):252:log_raise_exception] <top>: > connection to peer is broken > [2016-03-30 22:09:31.831017] W > [syncdutils(/gpool/brick03/geotest):256:log_raise_exception] <top>: > !!!!!!!!!!!!! > [2016-03-30 22:09:31.831258] W > [syncdutils(/gpool/brick03/geotest):257:log_raise_exception] <top>: > !!! getting "No such file or directory" errors is most likely due to > MISCONFIGURATION, please consult > https://access.redhat.com/site/documentation/en-US/Red_Hat_Storage/2.1/html/Administration_Guide/chap-User_Guide-Geo_Rep-Preparation-Settingup_Environment.html > [2016-03-30 22:09:31.831502] W > [syncdutils(/gpool/brick03/geotest):265:log_raise_exception] <top>: > !!!!!!!!!!!!! > [2016-03-30 22:09:31.836395] E > [resource(/gpool/brick03/geotest):222:errlog] Popen: command "ssh > -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i > /var/lib/glusterd/geo-replication/secret.pem -oControlMaster=auto -S > /tmp/gsyncd-aux-ssh-SfXvbB/de372ce5774b5d259c58c5c9522ffc8f.sock > guser at slave-host02 /nonexistent/gsyncd --session-owner > ec473e17-b933-4bf7-9eed-4c393f7aaf5d -N --listen --timeout 120 > gluster://localhost:geotestdr" returned with 127, saying: > [2016-03-30 22:09:31.836694] E > [resource(/gpool/brick03/geotest):226:logerr] Popen: ssh> bash: > /nonexistent/gsyncd: No such file or directory > [2016-03-30 22:09:31.837193] I > [syncdutils(/gpool/brick03/geotest):220:finalize] <top>: exiting. > [2016-03-30 22:09:31.840569] I [repce(agent):92:service_loop] > RepceServer: terminating on reaching EOF. > [2016-03-30 22:09:31.840993] I [syncdutils(agent):220:finalize] <top>: > exiting. > [2016-03-30 22:09:31.840742] I [monitor(monitor):274:monitor] Monitor: > worker(/gpool/brick03/geotest) died before establishing connection > [2016-03-30 22:09:42.130866] I [monitor(monitor):221:monitor] Monitor: > ------------------------------------------------------------ > [2016-03-30 22:09:42.131448] I [monitor(monitor):222:monitor] Monitor: > starting gsyncd worker > [2016-03-30 22:09:42.348165] I > [gsyncd(/gpool/brick03/geotest):649:main_i] <top>: syncing: > gluster://localhost:geotest -> > ssh://guser at slave-host02:gluster://localhost:geotestdr > [2016-03-30 22:09:42.349118] I [changelogagent(agent):75:__init__] > ChangelogAgent: Agent listining... > [2016-03-30 22:09:42.653141] E > [syncdutils(/gpool/brick03/geotest):252:log_raise_exception] <top>: > connection to peer is broken > [2016-03-30 22:09:42.653656] W > [syncdutils(/gpool/brick03/geotest):256:log_raise_exception] <top>: > !!!!!!!!!!!!! > [2016-03-30 22:09:42.653898] W > [syncdutils(/gpool/brick03/geotest):257:log_raise_exception] <top>: > !!! getting "No such file or directory" errors is most likely due to > MISCONFIGURATION, please consult > https://access.redhat.com/site/documentation/en-US/Red_Hat_Storage/2.1/html/Administration_Guide/chap-User_Guide-Geo_Rep-Preparation-Settingup_Environment.html > [2016-03-30 22:09:42.654129] W > [syncdutils(/gpool/brick03/geotest):265:log_raise_exception] <top>: > !!!!!!!!!!!!! > [2016-03-30 22:09:42.659329] E > [resource(/gpool/brick03/geotest):222:errlog] Popen: command "ssh > -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i > /var/lib/glusterd/geo-replication/secret.pem -oControlMaster=auto -S > /tmp/gsyncd-aux-ssh-6r8rxx/de372ce5774b5d259c58c5c9522ffc8f.sock > guser at slave-host02 /nonexistent/gsyncd --session-owner > ec473e17-b933-4bf7-9eed-4c393f7aaf5d -N --listen --timeout 120 > gluster://localhost:geotestdr" returned with 127, saying: > [2016-03-30 22:09:42.659626] E > [resource(/gpool/brick03/geotest):226:logerr] Popen: ssh> bash: > /nonexistent/gsyncd: No such file or directory > [2016-03-30 22:09:42.660140] I > [syncdutils(/gpool/brick03/geotest):220:finalize] <top>: exiting. > [2016-03-30 22:09:42.662802] I [repce(agent):92:service_loop] > RepceServer: terminating on reaching EOF. > [2016-03-30 22:09:42.663197] I [syncdutils(agent):220:finalize] <top>: > exiting. > [2016-03-30 22:09:42.663024] I [monitor(monitor):274:monitor] Monitor: > worker(/gpool/brick03/geotest) died before establishing connection > > > ?Bishoy > >> On Mar 30, 2016, at 10:50 AM, Gmail <b.s.mikhael at gmail.com >> <mailto:b.s.mikhael at gmail.com>> wrote: >> >> I?ve tried changing the permissions to 777 on /var/log/glusterfs on >> all the slave nodes, but still no luck :( >> >> here is the log from the master node where I created and started the >> geo-replication session. >> >> [2016-03-30 17:14:53.463150] I [monitor(monitor):221:monitor] >> Monitor: ------------------------------------------------------------ >> [2016-03-30 17:14:53.463669] I [monitor(monitor):222:monitor] >> Monitor: starting gsyncd worker >> [2016-03-30 17:14:53.603774] I [changelogagent(agent):75:__init__] >> ChangelogAgent: Agent listining... >> [2016-03-30 17:14:53.604080] I >> [gsyncd(/mnt/brick10/xfsvol2):649:main_i] <top>: syncing: >> gluster://localhost:xfsvol2 -> >> ssh://guser at slave-host01:gluster://localhost:xfsvol2dr >> [2016-03-30 17:14:54.210602] E >> [syncdutils(/mnt/brick10/xfsvol2):252:log_raise_exception] <top>: >> connection to peer is broken >> [2016-03-30 17:14:54.211117] E >> [resource(/mnt/brick10/xfsvol2):222:errlog] Popen: command "ssh >> -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i >> /var/lib/glusterd/geo-replication/secret.pem -oControlMaster=auto -S >> /tmp/gsyncd-aux-ssh-evONxc/3bda60dc6e900c0833fed4e4fdfbd480.sock >> guser at slave-host01 /nonexistent/gsyncd --session-owner >> ef9ccae5-0def-4a47-9a96-881a1896755c -N --listen --timeout 120 >> gluster://localhost:xfsvol2dr" returned with 1, saying: >> [2016-03-30 17:14:54.211376] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> [2016-03-30 >> 17:14:53.933174] I [cli.c:720:main] 0-cli: Started running >> /usr/sbin/gluster with version 3.7.3 >> [2016-03-30 17:14:54.211631] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> [2016-03-30 >> 17:14:53.933225] I [cli.c:608:cli_rpc_init] 0-cli: Connecting to >> remote glusterd at localhost >> [2016-03-30 17:14:54.211828] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> [2016-03-30 >> 17:14:54.074207] I [MSGID: 101190] >> [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started >> thread with index 1 >> [2016-03-30 17:14:54.212017] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> [2016-03-30 >> 17:14:54.074302] I [socket.c:2409:socket_event_handler] 0-transport: >> disconnecting now >> [2016-03-30 17:14:54.212199] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> [2016-03-30 >> 17:14:54.077207] I [cli-rpc-ops.c:6230:gf_cli_getwd_cbk] 0-cli: >> Received resp to getwd >> [2016-03-30 17:14:54.212380] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> [2016-03-30 >> 17:14:54.077269] I [input.c:36:cli_batch] 0-: Exiting with: 0 >> [2016-03-30 17:14:54.212584] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> ERROR:root:FAIL: >> [2016-03-30 17:14:54.212774] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> Traceback >> (most recent call last): >> [2016-03-30 17:14:54.212954] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> File >> "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 165, in main >> [2016-03-30 17:14:54.213131] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> main_i() >> [2016-03-30 17:14:54.213308] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> File >> "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 638, in main_i >> [2016-03-30 17:14:54.213500] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> >> startup(go_daemon=go_daemon, log_file=log_file, label=label) >> [2016-03-30 17:14:54.213690] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> File >> "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 144, in >> startup >> [2016-03-30 17:14:54.213890] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> >> GLogger._gsyncd_loginit(**kw) >> [2016-03-30 17:14:54.214068] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> File >> "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 107, in >> _gsyncd_loginit >> [2016-03-30 17:14:54.214246] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> >> cls.setup(label=kw.get('label'), **lkw) >> [2016-03-30 17:14:54.214422] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> File >> "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 79, in setup >> [2016-03-30 17:14:54.214622] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> >> logging_handler = handlers.WatchedFileHandler(lprm['filename']) >> [2016-03-30 17:14:54.214802] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> File >> "/usr/lib64/python2.6/logging/handlers.py", line 377, in __init__ >> [2016-03-30 17:14:54.214977] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> >> logging.FileHandler.__init__(self, filename, mode, encoding, delay) >> [2016-03-30 17:14:54.215152] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> File >> "/usr/lib64/python2.6/logging/__init__.py", line 835, in __init__ >> [2016-03-30 17:14:54.215327] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> >> StreamHandler.__init__(self, self._open()) >> [2016-03-30 17:14:54.215523] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> File >> "/usr/lib64/python2.6/logging/__init__.py", line 854, in _open >> [2016-03-30 17:14:54.215703] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> stream = >> open(self.baseFilename, self.mode) >> [2016-03-30 17:14:54.215883] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> IOError: >> [Errno 13] Permission denied: >> '/var/log/glusterfs/geo-replication-slaves/mbr/ef9ccae5-0def-4a47-9a96-881a1896755c:gluster%3A%2F%2F127.0.0.1%3Axfsvol2dr.log' >> [2016-03-30 17:14:54.216063] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> failed with >> IOError. >> [2016-03-30 17:14:54.216500] I >> [syncdutils(/mnt/brick10/xfsvol2):220:finalize] <top>: exiting. >> [2016-03-30 17:14:54.218672] I [repce(agent):92:service_loop] >> RepceServer: terminating on reaching EOF. >> [2016-03-30 17:14:54.219063] I [syncdutils(agent):220:finalize] >> <top>: exiting. >> [2016-03-30 17:14:54.218930] I [monitor(monitor):274:monitor] >> Monitor: worker(/mnt/brick10/xfsvol2) died before establishing connection >> >> *?Bishoy* >> * >> * >>> On Mar 29, 2016, at 1:05 AM, Aravinda <avishwan at redhat.com >>> <mailto:avishwan at redhat.com>> wrote: >>> >>> Geo-replication command should be run as privileged user itself. >>> >>> gluster volume geo-replication <MASTERVOL> <SLAVEUSER>@<SLAVEHOST> start >>> >>> and then check the status, if it shows Faulty then please share the >>> log files present in >>> /var/log/glusterfs/geo-replication/<MASTERVOL>/*.log >>> >>> regards >>> Aravinda >>> On 03/29/2016 12:51 PM, Gmail wrote: >>>> I?ve been trying to setup geo-replication using Gluster 3.7.3 on >>>> OEL 6.5 >>>> It keeps giving me faulty session. >>>> I?ve tried to use root user instead, it works fine! >>>> >>>> I?ve followed literally the documentation but no luck getting the >>>> unprivileged user working. >>>> >>>> I?ve tried running /usr/libexec/glusterfs/gsyncd on the slave node >>>> using the unprivileged user, and that?s what I get. >>>> >>>> /usr/libexec/glusterfs/gsyncd --session-owner >>>> ef9ccae5-0def-4a47-9a96-881a1896755c -N --listen --timeout 120 >>>> gluster://localhost:vol01dr >>>> [2016-03-29 00:52:49.058244] I [cli.c:720:main] 0-cli: Started >>>> running /usr/sbin/gluster with version 3.7.3 >>>> [2016-03-29 00:52:49.058297] I [cli.c:608:cli_rpc_init] 0-cli: >>>> Connecting to remote glusterd at localhost >>>> [2016-03-29 00:52:49.174686] I [MSGID: 101190] >>>> [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started >>>> thread with index 1 >>>> [2016-03-29 00:52:49.174768] I [socket.c:2409:socket_event_handler] >>>> 0-transport: disconnecting now >>>> [2016-03-29 00:52:49.177482] I >>>> [cli-rpc-ops.c:6230:gf_cli_getwd_cbk] 0-cli: Received resp to getwd >>>> [2016-03-29 00:52:49.177545] I [input.c:36:cli_batch] 0-: Exiting >>>> with: 0 >>>> ERROR:root:FAIL: >>>> Traceback (most recent call last): >>>> File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line >>>> 165, in main >>>> main_i() >>>> File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line >>>> 638, in main_i >>>> startup(go_daemon=go_daemon, log_file=log_file, label=label) >>>> File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line >>>> 144, in startup >>>> GLogger._gsyncd_loginit(**kw) >>>> File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line >>>> 107, in _gsyncd_loginit >>>> cls.setup(label=kw.get('label'), **lkw) >>>> File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 79, >>>> in setup >>>> logging_handler = handlers.WatchedFileHandler(lprm['filename']) >>>> File "/usr/lib64/python2.6/logging/handlers.py", line 377, in __init__ >>>> logging.FileHandler.__init__(self, filename, mode, encoding, delay) >>>> File "/usr/lib64/python2.6/logging/__init__.py", line 835, in __init__ >>>> StreamHandler.__init__(self, self._open()) >>>> File "/usr/lib64/python2.6/logging/__init__.py", line 854, in _open >>>> stream = open(self.baseFilename, self.mode) >>>> IOError: [Errno 13] Permission denied: >>>> '/var/log/glusterfs/geo-replication-slaves/mbr/ef9ccae5-0def-4a47-9a96-881a1896755c:gluster%3A%2F%2F127.0.0.1%3Avol01dr.log' >>>> failed with IOError. >>>> >>>> >>>> *? Bishoy* >>>> >>>> >>>> >>>> _______________________________________________ >>>> Gluster-users mailing list >>>> Gluster-users at gluster.org >>>> http://www.gluster.org/mailman/listinfo/gluster-users >>> >> > > > > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://www.gluster.org/mailman/listinfo/gluster-users-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160331/2e3b9df0/attachment.html>
Hi, From the error I understood that SSH connection is failing. In slave-host02 extra entries present in /home/guser/.ssh/authorized_keys. In /home/guser/.ssh/authorized_keys Please delete extra lines which does not start with "command=". Then stop and start the Geo-replication. regards Aravinda On 03/31/2016 04:00 AM, Gmail wrote:> I?ve rebuilt the cluster again, making a fresh installation. And now > the error is different. > > > > > > MASTER NODE MASTER VOL MASTER BRICK SLAVE USER > SLAVE SLAVE NODE STATUS CRAWL > STATUS LAST_SYNCED > ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- > master-host01.me.com <http://master-host01.me.com> geotest > /gpool/brick03/geotest guser guser at slave-host01::geotestdr N/A > Faulty N/A N/A > master-host02.me.com <http://master-host02.me.com> geotest > /gpool/brick03/geotest guser guser at slave-host01::geotestdr > slave-host01 Passive N/A N/A > master-host03.me.com <http://master-host03.me.com> geotest > /gpool/brick03/geotest guser guser at slave-host01::geotestdr > slave-host03 Passive N/A N/A > > > > > > > [2016-03-30 22:09:31.326898] I [monitor(monitor):221:monitor] Monitor: > ------------------------------------------------------------ > [2016-03-30 22:09:31.327461] I [monitor(monitor):222:monitor] Monitor: > starting gsyncd worker > [2016-03-30 22:09:31.544631] I > [gsyncd(/gpool/brick03/geotest):649:main_i] <top>: syncing: > gluster://localhost:geotest -> > ssh://guser at slave-host02:gluster://localhost:geotestdr > [2016-03-30 22:09:31.547542] I [changelogagent(agent):75:__init__] > ChangelogAgent: Agent listining... > [2016-03-30 22:09:31.830554] E > [syncdutils(/gpool/brick03/geotest):252:log_raise_exception] <top>: > connection to peer is broken > [2016-03-30 22:09:31.831017] W > [syncdutils(/gpool/brick03/geotest):256:log_raise_exception] <top>: > !!!!!!!!!!!!! > [2016-03-30 22:09:31.831258] W > [syncdutils(/gpool/brick03/geotest):257:log_raise_exception] <top>: > !!! getting "No such file or directory" errors is most likely due to > MISCONFIGURATION, please consult > https://access.redhat.com/site/documentation/en-US/Red_Hat_Storage/2.1/html/Administration_Guide/chap-User_Guide-Geo_Rep-Preparation-Settingup_Environment.html > [2016-03-30 22:09:31.831502] W > [syncdutils(/gpool/brick03/geotest):265:log_raise_exception] <top>: > !!!!!!!!!!!!! > [2016-03-30 22:09:31.836395] E > [resource(/gpool/brick03/geotest):222:errlog] Popen: command "ssh > -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i > /var/lib/glusterd/geo-replication/secret.pem -oControlMaster=auto -S > /tmp/gsyncd-aux-ssh-SfXvbB/de372ce5774b5d259c58c5c9522ffc8f.sock > guser at slave-host02 /nonexistent/gsyncd --session-owner > ec473e17-b933-4bf7-9eed-4c393f7aaf5d -N --listen --timeout 120 > gluster://localhost:geotestdr" returned with 127, saying: > [2016-03-30 22:09:31.836694] E > [resource(/gpool/brick03/geotest):226:logerr] Popen: ssh> bash: > /nonexistent/gsyncd: No such file or directory > [2016-03-30 22:09:31.837193] I > [syncdutils(/gpool/brick03/geotest):220:finalize] <top>: exiting. > [2016-03-30 22:09:31.840569] I [repce(agent):92:service_loop] > RepceServer: terminating on reaching EOF. > [2016-03-30 22:09:31.840993] I [syncdutils(agent):220:finalize] <top>: > exiting. > [2016-03-30 22:09:31.840742] I [monitor(monitor):274:monitor] Monitor: > worker(/gpool/brick03/geotest) died before establishing connection > [2016-03-30 22:09:42.130866] I [monitor(monitor):221:monitor] Monitor: > ------------------------------------------------------------ > [2016-03-30 22:09:42.131448] I [monitor(monitor):222:monitor] Monitor: > starting gsyncd worker > [2016-03-30 22:09:42.348165] I > [gsyncd(/gpool/brick03/geotest):649:main_i] <top>: syncing: > gluster://localhost:geotest -> > ssh://guser at slave-host02:gluster://localhost:geotestdr > [2016-03-30 22:09:42.349118] I [changelogagent(agent):75:__init__] > ChangelogAgent: Agent listining... > [2016-03-30 22:09:42.653141] E > [syncdutils(/gpool/brick03/geotest):252:log_raise_exception] <top>: > connection to peer is broken > [2016-03-30 22:09:42.653656] W > [syncdutils(/gpool/brick03/geotest):256:log_raise_exception] <top>: > !!!!!!!!!!!!! > [2016-03-30 22:09:42.653898] W > [syncdutils(/gpool/brick03/geotest):257:log_raise_exception] <top>: > !!! getting "No such file or directory" errors is most likely due to > MISCONFIGURATION, please consult > https://access.redhat.com/site/documentation/en-US/Red_Hat_Storage/2.1/html/Administration_Guide/chap-User_Guide-Geo_Rep-Preparation-Settingup_Environment.html > [2016-03-30 22:09:42.654129] W > [syncdutils(/gpool/brick03/geotest):265:log_raise_exception] <top>: > !!!!!!!!!!!!! > [2016-03-30 22:09:42.659329] E > [resource(/gpool/brick03/geotest):222:errlog] Popen: command "ssh > -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i > /var/lib/glusterd/geo-replication/secret.pem -oControlMaster=auto -S > /tmp/gsyncd-aux-ssh-6r8rxx/de372ce5774b5d259c58c5c9522ffc8f.sock > guser at slave-host02 /nonexistent/gsyncd --session-owner > ec473e17-b933-4bf7-9eed-4c393f7aaf5d -N --listen --timeout 120 > gluster://localhost:geotestdr" returned with 127, saying: > [2016-03-30 22:09:42.659626] E > [resource(/gpool/brick03/geotest):226:logerr] Popen: ssh> bash: > /nonexistent/gsyncd: No such file or directory > [2016-03-30 22:09:42.660140] I > [syncdutils(/gpool/brick03/geotest):220:finalize] <top>: exiting. > [2016-03-30 22:09:42.662802] I [repce(agent):92:service_loop] > RepceServer: terminating on reaching EOF. > [2016-03-30 22:09:42.663197] I [syncdutils(agent):220:finalize] <top>: > exiting. > [2016-03-30 22:09:42.663024] I [monitor(monitor):274:monitor] Monitor: > worker(/gpool/brick03/geotest) died before establishing connection > > > ?Bishoy > >> On Mar 30, 2016, at 10:50 AM, Gmail <b.s.mikhael at gmail.com >> <mailto:b.s.mikhael at gmail.com>> wrote: >> >> I?ve tried changing the permissions to 777 on /var/log/glusterfs on >> all the slave nodes, but still no luck :( >> >> here is the log from the master node where I created and started the >> geo-replication session. >> >> [2016-03-30 17:14:53.463150] I [monitor(monitor):221:monitor] >> Monitor: ------------------------------------------------------------ >> [2016-03-30 17:14:53.463669] I [monitor(monitor):222:monitor] >> Monitor: starting gsyncd worker >> [2016-03-30 17:14:53.603774] I [changelogagent(agent):75:__init__] >> ChangelogAgent: Agent listining... >> [2016-03-30 17:14:53.604080] I >> [gsyncd(/mnt/brick10/xfsvol2):649:main_i] <top>: syncing: >> gluster://localhost:xfsvol2 -> >> ssh://guser at slave-host01:gluster://localhost:xfsvol2dr >> [2016-03-30 17:14:54.210602] E >> [syncdutils(/mnt/brick10/xfsvol2):252:log_raise_exception] <top>: >> connection to peer is broken >> [2016-03-30 17:14:54.211117] E >> [resource(/mnt/brick10/xfsvol2):222:errlog] Popen: command "ssh >> -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i >> /var/lib/glusterd/geo-replication/secret.pem -oControlMaster=auto -S >> /tmp/gsyncd-aux-ssh-evONxc/3bda60dc6e900c0833fed4e4fdfbd480.sock >> guser at slave-host01 /nonexistent/gsyncd --session-owner >> ef9ccae5-0def-4a47-9a96-881a1896755c -N --listen --timeout 120 >> gluster://localhost:xfsvol2dr" returned with 1, saying: >> [2016-03-30 17:14:54.211376] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> [2016-03-30 >> 17:14:53.933174] I [cli.c:720:main] 0-cli: Started running >> /usr/sbin/gluster with version 3.7.3 >> [2016-03-30 17:14:54.211631] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> [2016-03-30 >> 17:14:53.933225] I [cli.c:608:cli_rpc_init] 0-cli: Connecting to >> remote glusterd at localhost >> [2016-03-30 17:14:54.211828] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> [2016-03-30 >> 17:14:54.074207] I [MSGID: 101190] >> [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started >> thread with index 1 >> [2016-03-30 17:14:54.212017] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> [2016-03-30 >> 17:14:54.074302] I [socket.c:2409:socket_event_handler] 0-transport: >> disconnecting now >> [2016-03-30 17:14:54.212199] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> [2016-03-30 >> 17:14:54.077207] I [cli-rpc-ops.c:6230:gf_cli_getwd_cbk] 0-cli: >> Received resp to getwd >> [2016-03-30 17:14:54.212380] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> [2016-03-30 >> 17:14:54.077269] I [input.c:36:cli_batch] 0-: Exiting with: 0 >> [2016-03-30 17:14:54.212584] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> ERROR:root:FAIL: >> [2016-03-30 17:14:54.212774] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> Traceback >> (most recent call last): >> [2016-03-30 17:14:54.212954] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> File >> "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 165, in main >> [2016-03-30 17:14:54.213131] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> main_i() >> [2016-03-30 17:14:54.213308] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> File >> "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 638, in main_i >> [2016-03-30 17:14:54.213500] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> >> startup(go_daemon=go_daemon, log_file=log_file, label=label) >> [2016-03-30 17:14:54.213690] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> File >> "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 144, in >> startup >> [2016-03-30 17:14:54.213890] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> >> GLogger._gsyncd_loginit(**kw) >> [2016-03-30 17:14:54.214068] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> File >> "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 107, in >> _gsyncd_loginit >> [2016-03-30 17:14:54.214246] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> >> cls.setup(label=kw.get('label'), **lkw) >> [2016-03-30 17:14:54.214422] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> File >> "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 79, in setup >> [2016-03-30 17:14:54.214622] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> >> logging_handler = handlers.WatchedFileHandler(lprm['filename']) >> [2016-03-30 17:14:54.214802] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> File >> "/usr/lib64/python2.6/logging/handlers.py", line 377, in __init__ >> [2016-03-30 17:14:54.214977] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> >> logging.FileHandler.__init__(self, filename, mode, encoding, delay) >> [2016-03-30 17:14:54.215152] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> File >> "/usr/lib64/python2.6/logging/__init__.py", line 835, in __init__ >> [2016-03-30 17:14:54.215327] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> >> StreamHandler.__init__(self, self._open()) >> [2016-03-30 17:14:54.215523] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> File >> "/usr/lib64/python2.6/logging/__init__.py", line 854, in _open >> [2016-03-30 17:14:54.215703] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> stream = >> open(self.baseFilename, self.mode) >> [2016-03-30 17:14:54.215883] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> IOError: >> [Errno 13] Permission denied: >> '/var/log/glusterfs/geo-replication-slaves/mbr/ef9ccae5-0def-4a47-9a96-881a1896755c:gluster%3A%2F%2F127.0.0.1%3Axfsvol2dr.log' >> [2016-03-30 17:14:54.216063] E >> [resource(/mnt/brick10/xfsvol2):226:logerr] Popen: ssh> failed with >> IOError. >> [2016-03-30 17:14:54.216500] I >> [syncdutils(/mnt/brick10/xfsvol2):220:finalize] <top>: exiting. >> [2016-03-30 17:14:54.218672] I [repce(agent):92:service_loop] >> RepceServer: terminating on reaching EOF. >> [2016-03-30 17:14:54.219063] I [syncdutils(agent):220:finalize] >> <top>: exiting. >> [2016-03-30 17:14:54.218930] I [monitor(monitor):274:monitor] >> Monitor: worker(/mnt/brick10/xfsvol2) died before establishing connection >> >> *?Bishoy* >> * >> * >>> On Mar 29, 2016, at 1:05 AM, Aravinda <avishwan at redhat.com >>> <mailto:avishwan at redhat.com>> wrote: >>> >>> Geo-replication command should be run as privileged user itself. >>> >>> gluster volume geo-replication <MASTERVOL> <SLAVEUSER>@<SLAVEHOST> start >>> >>> and then check the status, if it shows Faulty then please share the >>> log files present in >>> /var/log/glusterfs/geo-replication/<MASTERVOL>/*.log >>> >>> regards >>> Aravinda >>> On 03/29/2016 12:51 PM, Gmail wrote: >>>> I?ve been trying to setup geo-replication using Gluster 3.7.3 on >>>> OEL 6.5 >>>> It keeps giving me faulty session. >>>> I?ve tried to use root user instead, it works fine! >>>> >>>> I?ve followed literally the documentation but no luck getting the >>>> unprivileged user working. >>>> >>>> I?ve tried running /usr/libexec/glusterfs/gsyncd on the slave node >>>> using the unprivileged user, and that?s what I get. >>>> >>>> /usr/libexec/glusterfs/gsyncd --session-owner >>>> ef9ccae5-0def-4a47-9a96-881a1896755c -N --listen --timeout 120 >>>> gluster://localhost:vol01dr >>>> [2016-03-29 00:52:49.058244] I [cli.c:720:main] 0-cli: Started >>>> running /usr/sbin/gluster with version 3.7.3 >>>> [2016-03-29 00:52:49.058297] I [cli.c:608:cli_rpc_init] 0-cli: >>>> Connecting to remote glusterd at localhost >>>> [2016-03-29 00:52:49.174686] I [MSGID: 101190] >>>> [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started >>>> thread with index 1 >>>> [2016-03-29 00:52:49.174768] I [socket.c:2409:socket_event_handler] >>>> 0-transport: disconnecting now >>>> [2016-03-29 00:52:49.177482] I >>>> [cli-rpc-ops.c:6230:gf_cli_getwd_cbk] 0-cli: Received resp to getwd >>>> [2016-03-29 00:52:49.177545] I [input.c:36:cli_batch] 0-: Exiting >>>> with: 0 >>>> ERROR:root:FAIL: >>>> Traceback (most recent call last): >>>> File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line >>>> 165, in main >>>> main_i() >>>> File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line >>>> 638, in main_i >>>> startup(go_daemon=go_daemon, log_file=log_file, label=label) >>>> File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line >>>> 144, in startup >>>> GLogger._gsyncd_loginit(**kw) >>>> File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line >>>> 107, in _gsyncd_loginit >>>> cls.setup(label=kw.get('label'), **lkw) >>>> File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 79, >>>> in setup >>>> logging_handler = handlers.WatchedFileHandler(lprm['filename']) >>>> File "/usr/lib64/python2.6/logging/handlers.py", line 377, in __init__ >>>> logging.FileHandler.__init__(self, filename, mode, encoding, delay) >>>> File "/usr/lib64/python2.6/logging/__init__.py", line 835, in __init__ >>>> StreamHandler.__init__(self, self._open()) >>>> File "/usr/lib64/python2.6/logging/__init__.py", line 854, in _open >>>> stream = open(self.baseFilename, self.mode) >>>> IOError: [Errno 13] Permission denied: >>>> '/var/log/glusterfs/geo-replication-slaves/mbr/ef9ccae5-0def-4a47-9a96-881a1896755c:gluster%3A%2F%2F127.0.0.1%3Avol01dr.log' >>>> failed with IOError. >>>> >>>> >>>> *? Bishoy* >>>> >>>> >>>> >>>> _______________________________________________ >>>> Gluster-users mailing list >>>> Gluster-users at gluster.org >>>> http://www.gluster.org/mailman/listinfo/gluster-users >>> >> >-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160331/5235734c/attachment.html>