Gilberto Ferreira
2025-Nov-24 18:16 UTC
[Gluster-users] Failed to Establish Geo-replication Session Please check gsync config file. Unable to get statefile's name
Hi there
FYI, I tried with the gluster deb packages from github, i.e. gluster
version 11.2, and could create the geo-rep session but got faulty status
again...
Some logs
[2025-11-24 18:13:33.110898] W [gsyncd(config-get):299:main] <top>:
Session
config file not exists, using the default config
[{path=/var/lib/glusterd/geo-replication/VMS_gluster3_VMS-REP/gsyncd.conf}]
[2025-11-24 18:13:35.199845] I
[subcmds(monitor-status):29:subcmd_monitor_status] <top>: Monitor Status
Change [{status=Created}]
[2025-11-24 18:14:39.398460] I
[gsyncdstatus(monitor):247:set_worker_status] GeorepStatus: Worker Status
Change [{status=Initializing...}]
[2025-11-24 18:14:39.398788] I [monitor(monitor):158:monitor] Monitor:
starting gsyncd worker [{brick=/mnt/pve/data1/vms},
{secondary_node=gluster4}]
[2025-11-24 18:14:39.491754] I [resource(worker
/mnt/pve/data1/vms):1388:connect_remote] SSH: Initializing SSH connection
between primary and secondary...
[2025-11-24 18:14:41.76723] I [resource(worker
/mnt/pve/data1/vms):1436:connect_remote] SSH: SSH connection between
primary and secondary established. [{duration=1.5848}]
[2025-11-24 18:14:41.76885] I [resource(worker
/mnt/pve/data1/vms):1117:connect] GLUSTER: Mounting gluster volume
locally...
[2025-11-24 18:14:42.108398] I [resource(worker
/mnt/pve/data1/vms):1139:connect] GLUSTER: Mounted gluster volume
[{duration=1.0314}]
[2025-11-24 18:14:42.108612] I [subcmds(worker
/mnt/pve/data1/vms):84:subcmd_worker] <top>: Worker spawn successful.
Acknowledging back to monitor
[2025-11-24 18:14:44.117002] I [primary(worker
/mnt/pve/data1/vms):1662:register] _GPrimary: Working dir
[{path=/var/lib/misc/gluster/gsyncd/VMS_gluster3_VMS-REP/mnt-pve-data1-vms}]
[2025-11-24 18:14:44.117345] I [resource(worker
/mnt/pve/data1/vms):1292:service_loop] GLUSTER: Register time
[{time=1764008084}]
[2025-11-24 18:14:44.124814] I [gsyncdstatus(worker
/mnt/pve/data1/vms):280:set_active] GeorepStatus: Worker Status Change
[{status=Active}]
[2025-11-24 18:14:44.218760] I [gsyncdstatus(worker
/mnt/pve/data1/vms):252:set_worker_crawl_status] GeorepStatus: Crawl Status
Change [{status=History Crawl}]
[2025-11-24 18:14:44.219035] I [primary(worker
/mnt/pve/data1/vms):1573:crawl] _GPrimary: starting history crawl
[{turns=1}, {stime=None}, {etime=1764008084}, {entry_stime=None}]
[2025-11-24 18:14:44.219154] I [resource(worker
/mnt/pve/data1/vms):1309:service_loop] GLUSTER: No stime available, using
xsync crawl
[2025-11-24 18:14:44.225654] I [primary(worker
/mnt/pve/data1/vms):1692:crawl] _GPrimary: starting hybrid crawl
[{stime=None}]
[2025-11-24 18:14:44.227337] I [gsyncdstatus(worker
/mnt/pve/data1/vms):252:set_worker_crawl_status] GeorepStatus: Crawl Status
Change [{status=Hybrid Crawl}]
[2025-11-24 18:14:45.227752] I [primary(worker
/mnt/pve/data1/vms):1703:crawl] _GPrimary: processing xsync changelog
[{path=/var/lib/misc/gluster/gsyncd/VMS_gluster3_VMS-REP/mnt-pve-data1-vms/xsync/XSYNC-CHANGELOG.1764008084}]
[2025-11-24 18:14:45.255496] I [primary(worker
/mnt/pve/data1/vms):1430:process] _GPrimary: Entry Time Taken [{UNL=0},
{RMD=0}, {CRE=0}, {MKN=0}, {MKD=1}, {REN=0}, {LIN=0}, {SYM=0},
{duration=0.0071}]
[2025-11-24 18:14:45.255639] I [primary(worker
/mnt/pve/data1/vms):1442:process] _GPrimary: Data/Metadata Time Taken
[{SETA=1}, {meta_duration=0.0078}, {SETX=0}, {XATT=0}, {DATA=0},
{data_duration=0.0010}]
[2025-11-24 18:14:45.255797] I [primary(worker
/mnt/pve/data1/vms):1452:process] _GPrimary: Batch Completed [{mode=xsync},
{duration=0.0277}, {changelog_start=1764008084},
{changelog_end=1764008084}, {num_changelogs=1}, {stime=None},
{entry_stime=None}]
[2025-11-24 18:14:45.260967] I [primary(worker
/mnt/pve/data1/vms):1699:crawl] _GPrimary: finished hybrid crawl
[{stime=(1764008084, 0)}]
[2025-11-24 18:14:45.266066] I [gsyncdstatus(worker
/mnt/pve/data1/vms):252:set_worker_crawl_status] GeorepStatus: Crawl Status
Change [{status=Changelog Crawl}]
[2025-11-24 18:14:55.280644] I [primary(worker
/mnt/pve/data1/vms):1525:crawl] _GPrimary: secondary's time
[{stime=(1764008084, 0)}]
[2025-11-24 18:14:55.645504] I [primary(worker
/mnt/pve/data1/vms):2010:syncjob] Syncer: Sync Time Taken [{job=1},
{num_files=2}, {return_code=12}, {duration=0.0327}]
[2025-11-24 18:14:55.645687] E [syncdutils(worker
/mnt/pve/data1/vms):845:errlog] Popen: command returned error [{cmd=rsync
-aR0 --inplace --files-from=- --super --stats --numeric-ids
--no-implied-dirs --existing --xattrs --acls --ignore-missing-args . -e ssh
-oPasswordAuthentication=no -oStrictHostKeyChecking=no -i
/var/lib/glusterd/geo-replication/secret.pem -p 22 -oControlMaster=auto -S
/tmp/gsyncd-aux-ssh-1u81eezq/dc97de7ffea2b18802fbdbd0b783902c.sock
-caes128-ctr gluster4:/proc/27188/cwd}, {error=12}]
[2025-11-24 18:14:56.111115] I [monitor(monitor):227:monitor] Monitor:
worker died in startup phase [{brick=/mnt/pve/data1/vms}]
[2025-11-24 18:14:56.122446] I
[gsyncdstatus(monitor):247:set_worker_status] GeorepStatus: Worker Status
Change [{status=Faulty}]
Em s?b., 22 de nov. de 2025 ?s 16:20, Gilberto Ferreira <
gilberto.nunes32 at gmail.com> escreveu:
> Here the log about the fault status
>
> [2025-11-22 19:18:56.478297] I [gsyncdstatus(worker
> /mnt/pve/data1/vms):252:set_worker_crawl_status] GeorepStatus: Crawl Status
> Change [{status=History Crawl}]
> [2025-11-22 19:18:56.478521] I [primary(worker
> /mnt/pve/data1/vms):1572:crawl] _GPrimary: starting history crawl
> [{turns=1}, {stime=(1763838427, 0)}, {etime=1763839136},
> {entry_stime=(1763838802, 0)}]
> [2025-11-22 19:18:57.479278] I [primary(worker
> /mnt/pve/data1/vms):1604:crawl] _GPrimary: secondary's time
> [{stime=(1763838427, 0)}]
> [2025-11-22 19:18:57.922752] I [primary(worker
> /mnt/pve/data1/vms):2009:syncjob] Syncer: Sync Time Taken [{job=1},
> {num_files=2}, {return_code=12}, {duration=0.0272}]
> [2025-11-22 19:18:57.922921] E [syncdutils(worker
> /mnt/pve/data1/vms):845:errlog] Popen: command returned error [{cmd=rsync
> -aR0 --inplace --files-from=- --super --stats --numeric-ids
> --no-implied-dirs --existing --xattrs --acls --ignore-missing-args . -e ssh
> -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i
> /var/lib/glusterd/geo-replication/secret.pem -p 22 -oControlMaster=auto -S
> /tmp/gsyncd-aux-ssh-c9h5okjo/5afb71218219138854b3c5a8eab300a4.sock
> -caes128-ctr gluster3:/proc/4004/cwd}, {error=12}]
> [2025-11-22 19:18:58.394410] I [monitor(monitor):227:monitor] Monitor:
> worker died in startup phase [{brick=/mnt/pve/data1/vms}]
> [2025-11-22 19:18:58.406208] I
> [gsyncdstatus(monitor):247:set_worker_status] GeorepStatus: Worker Status
> Change [{status=Faulty}]
> [2025-11-22 19:19:08.408871] I
> [gsyncdstatus(monitor):247:set_worker_status] GeorepStatus: Worker Status
> Change [{status=Initializing...}]
> [2025-11-22 19:19:08.408998] I [monitor(monitor):158:monitor] Monitor:
> starting gsyncd worker [{brick=/mnt/pve/data1/vms},
> {secondary_node=gluster3}]
> [2025-11-22 19:19:08.475625] I [resource(worker
> /mnt/pve/data1/vms):1388:connect_remote] SSH: Initializing SSH connection
> between primary and secondary...
> [2025-11-22 19:19:09.719658] I [resource(worker
> /mnt/pve/data1/vms):1436:connect_remote] SSH: SSH connection between
> primary and secondary established. [{duration=1.2439}]
> [2025-11-22 19:19:09.719800] I [resource(worker
> /mnt/pve/data1/vms):1117:connect] GLUSTER: Mounting gluster volume
> locally...
> [2025-11-22 19:19:10.740213] I [resource(worker
> /mnt/pve/data1/vms):1139:connect] GLUSTER: Mounted gluster volume
> [{duration=1.0203}]
> [2025-11-22 19:19:10.740427] I [subcmds(worker
> /mnt/pve/data1/vms):84:subcmd_worker] <top>: Worker spawn successful.
> Acknowledging back to monitor
> [2025-11-22 19:19:12.756579] I [primary(worker
> /mnt/pve/data1/vms):1661:register] _GPrimary: Working dir
>
[{path=/var/lib/misc/gluster/gsyncd/VMS_gluster3_VMS-REP/mnt-pve-data1-vms}]
> [2025-11-22 19:19:12.756854] I [resource(worker
> /mnt/pve/data1/vms):1292:service_loop] GLUSTER: Register time
> [{time=1763839152}]
> [2025-11-22 19:19:12.771767] I [gsyncdstatus(worker
> /mnt/pve/data1/vms):280:set_active] GeorepStatus: Worker Status Change
> [{status=Active}]
> [2025-11-22 19:19:12.834163] I [gsyncdstatus(worker
> /mnt/pve/data1/vms):252:set_worker_crawl_status] GeorepStatus: Crawl Status
> Change [{status=History Crawl}]
> [2025-11-22 19:19:12.834344] I [primary(worker
> /mnt/pve/data1/vms):1572:crawl] _GPrimary: starting history crawl
> [{turns=1}, {stime=(1763838427, 0)}, {etime=1763839152},
> {entry_stime=(1763838802, 0)}]
> [2025-11-22 19:19:13.835162] I [primary(worker
> /mnt/pve/data1/vms):1604:crawl] _GPrimary: secondary's time
> [{stime=(1763838427, 0)}]
> [2025-11-22 19:19:14.270295] I [primary(worker
> /mnt/pve/data1/vms):2009:syncjob] Syncer: Sync Time Taken [{job=1},
> {num_files=2}, {return_code=12}, {duration=0.0274}]
> [2025-11-22 19:19:14.270466] E [syncdutils(worker
> /mnt/pve/data1/vms):845:errlog] Popen: command returned error [{cmd=rsync
> -aR0 --inplace --files-from=- --super --stats --numeric-ids
> --no-implied-dirs --existing --xattrs --acls --ignore-missing-args . -e ssh
> -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i
> /var/lib/glusterd/geo-replication/secret.pem -p 22 -oControlMaster=auto -S
> /tmp/gsyncd-aux-ssh-1retjpoh/5afb71218219138854b3c5a8eab300a4.sock
> -caes128-ctr gluster3:/proc/4076/cwd}, {error=12}]
> [2025-11-22 19:19:14.741245] I [monitor(monitor):227:monitor] Monitor:
> worker died in startup phase [{brick=/mnt/pve/data1/vms}]
> [2025-11-22 19:19:14.752452] I
> [gsyncdstatus(monitor):247:set_worker_status] GeorepStatus: Worker Status
> Change [{status=Faulty}]
>
>
> It seems to me that it failed to open something via ssh session.
>
> I don't know... something like that.
>
>
> Em s?b., 22 de nov. de 2025 ?s 16:17, Gilberto Ferreira <
> gilberto.nunes32 at gmail.com> escreveu:
>
>> Hi
>>
>> I had succeeded in creating the session with the other side.
>> But I got a Faulty status
>> I the gsyncd.log I got this:
>>
>> Popen: command returned error [{cmd=rsync -aR0 --inplace --files-from=-
>> --super --stats --numeric-ids --no-implied-dirs --existing --xattrs
--acls
>> --ignore-missing-args . -e ssh -oPasswordAuthentication=no
>> -oStrictHostKeyChecking=no -i
/var/lib/glusterd/geo-replication/secret.pem
>> -p 22 -oControlMaster=auto -S
>> /tmp/gsyncd-aux-ssh-2j2yeofa/5afb71218219138854b3c5a8eab300a4.sock
>> -caes128-ctr gluster3:/proc/2662/cwd}, {error=12}]
>>
>>
>> This also happens with Gluster 12dev.
>>
>> After compiling the gluster 12dev, I successfully create a geo-rep
>> session but get the error above.
>>
>>
>> Any clue?
>>
>>
>> Best Regards
>>
>>
>>
>>
>>
>> Em s?b., 22 de nov. de 2025 ?s 15:12, Strahil Nikolov <
>> hunter86_bg at yahoo.com> escreveu:
>>
>>> Hi Gilberto,
>>>
>>> It should as long as it's the same problem.
>>>
>>> It will be nice to share your experience in the mailing list.
>>>
>>> Best Regards,
>>> Strahil Nikolov
>>>
>>> On Sat, Nov 22, 2025 at 18:13, Gilberto Ferreira
>>> <gilberto.nunes32 at gmail.com> wrote:
>>> Hi
>>>
>>> Should it work with Debian Trixie (13) as well?
>>>
>>> I will try it
>>>
>>>
>>>
>>> ---
>>>
>>>
>>> Gilberto Nunes Ferreira
>>> +55 (47) 99676-7530 - Whatsapp / Telegram
>>>
>>>
>>>
>>>
>>>
>>>
>>> Em s?b., 22 de nov. de 2025 ?s 12:37, Strahil Nikolov <
>>> hunter86_bg at yahoo.com> escreveu:
>>>
>>> Hi Gilberto,
>>>
>>> I think debian12 packages don't have
>>>
https://github.com/gluster/glusterfs/pull/4404/commits/c433a178e8208e1771fea4d61d0a22a95b8bc74b
>>>
>>> Run on source and destination this command and try again:
>>> sed -i 's/readfp/read_file/g'
>>>
/usr/lib/x86_64-linux-gnu/glusterfs/python/syncdaemon/gsyncdconfig.py
>>>
>>> In my test setup where source and destinations are each a single
debian
>>> 12 with a hackishl created gluster_shared_storage volume and
executing
>>> the sed, I got:
>>>
>>>
>>> # gluster volume geo-replication vol1 geoaccount at
gluster2::georep
>>> create push-pem
>>> Creating geo-replication session between vol1 & geoaccount at
gluster2::georep
>>> has been successful
>>>
>>>
>>> Best Regards,
>>> Strahil Nikolov
>>>
>>> ? ??????, 22 ??????? 2025 ?. ? 16:34:30 ?. ???????+2, Gilberto
Ferreira <
>>> gilberto.nunes32 at gmail.com> ??????:
>>>
>>>
>>> here the story about gluster 12dev faulty error:
>>>
>>> https://github.com/gluster/glusterfs/issues/4632
>>>
>>>
>>>
>>>
>>>
>>> Em s?b., 22 de nov. de 2025 ?s 10:59, Gilberto Ferreira <
>>> gilberto.nunes32 at gmail.com> escreveu:
>>>
>>> Hello there
>>> My testing was with proxmox 9 which has Debian 13.
>>> I tried with gluster 11.1 from Debian repo and then version 11.2
from
>>> git repo.
>>> I got statefile's name issue with both.
>>> Then I compiled the version 12dev and could create geo-replication
>>> session with success but got faulty status
>>>
>>> So that's it.
>>>
>>> ---
>>> Gilberto Nunes Ferreira
>>> +55 (47) 99676-7530
>>> Proxmox VE
>>>
>>> Em s?b., 22 de nov. de 2025, 10:14, Strahil Nikolov <
>>> hunter86_bg at yahoo.com> escreveu:
>>>
>>> Hi Gilberto,
>>>
>>> What version of os and gluster do you use exaclty ?
>>>
>>> Best Regards,
>>> Strahil Nikolov
>>>
>>> ? ?????, 21 ??????? 2025 ?. ? 14:08:19 ?. ???????+2, Gilberto
Ferreira <
>>> gilberto.nunes32 at gmail.com> ??????:
>>>
>>>
>>> Hello there
>>>
>>> If there is something else I could help, please let me know.
>>>
>>> Thanks
>>>
>>> Best Regards
>>>
>>>
>>>
>>>
>>>
>>>
>>> Em qua., 19 de nov. de 2025 ?s 15:21, Gilberto Ferreira <
>>> gilberto.nunes32 at gmail.com> escreveu:
>>>
>>> Hi there
>>>
>>> So there is no special script.
>>> First I try to using this:
>>> https://github.com/aravindavk/gluster-georep-tools, and then I
notice
>>> the issue.
>>> But after try do it by myself, I called for help.
>>> I tried:
>>>
>>> gluster volume geo-replication MASTERVOL root at
SLAVENODE::slavevol create push-pem
>>>
>>> gluster volume geo-replication MASTERVOL root at
SLAVENODE::slavevol start
>>>
>>>
>>> And got the issue.
>>>
>>>
>>> Thanks
>>>
>>>
>>> ---
>>>
>>>
>>> Gilberto Nunes Ferreira
>>> +55 (47) 99676-7530 - Whatsapp / Telegram
>>>
>>>
>>>
>>>
>>>
>>>
>>> Em qua., 19 de nov. de 2025 ?s 15:09, Strahil Nikolov <
>>> hunter86_bg at yahoo.com> escreveu:
>>>
>>> Hi Gilberto,
>>>
>>> I have no idea why my previous message was not sent (sorry about
that).
>>> I suspect it's a bug. If you have some script or ansible
playbook for
>>> the setup, it could help me reproduce it locally.
>>>
>>>
>>> Best Regards,
>>> Strahil Nikolov
>>>
>>> ? ???????, 11 ??????? 2025 ?. ? 16:55:59 ?. ???????+2, Gilberto
Ferreira
>>> <gilberto.nunes32 at gmail.com> ??????:
>>>
>>>
>>> Any clue about this issue?
>>>
>>>
>>>
>>>
>>>
>>>
>>> Em seg., 10 de nov. de 2025 ?s 14:47, Gilberto Ferreira <
>>> gilberto.nunes32 at gmail.com> escreveu:
>>>
>>> I still don't get it, because with Debian BookWorm and Gluster
10,
>>> geo-rep works perfectly.
>>> It's something about Trixie and Gluster 11.x.
>>>
>>>
>>>
>>>
>>>
>>> Em seg., 10 de nov. de 2025 ?s 14:45, Karl Kleinpaste <
>>> karl at kleinpaste.org> escreveu:
>>>
>>> On 11/10/25 12:21 PM, Gilberto Ferreira wrote:
>>>
>>> And yes. With gluster 11.2 from github repo, the very some error:
>>> gluster vol geo VMS gluster3::VMS-REP create push-pem
>>> Please check gsync config file. Unable to get statefile's name
>>> geo-replication command failed
>>>
>>>
>>> I had this problem a year ago, Aug 2024
>>>
<https://lists.gluster.org/pipermail/gluster-users/2024-August/040625.html>.
>>> I went rounds and rounds with Strahil for a week, trying to find
why I
>>> couldn't cross the finish line of successful georep. It always
ends in:
>>>
>>> Please check gsync config file. Unable to get statefile's name
>>> geo-replication command failed
>>>
>>> The volumes were set up properly, the commands for georep were done
>>> correctly, per guidelines, but georep was left forever in a state
of
>>> Created, never Active.
>>>
>>> Finally I just gave up. I can't use gluster if it won't
work with me. I
>>> found that gluster would not give enough adequate diagnostics to
provide a (
>>> *useful*!) explanation of what is actually wrong.
>>> ________
>>>
>>>
>>>
>>> Community Meeting Calendar:
>>>
>>> Schedule -
>>> Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
>>> Bridge: https://meet.google.com/cpu-eiue-hvk
>>> Gluster-users mailing list
>>> Gluster-users at gluster.org
>>> https://lists.gluster.org/mailman/listinfo/gluster-users
>>>
>>> ________
>>>
>>>
>>>
>>> Community Meeting Calendar:
>>>
>>> Schedule -
>>> Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
>>> Bridge: https://meet.google.com/cpu-eiue-hvk
>>> Gluster-users mailing list
>>> Gluster-users at gluster.org
>>> https://lists.gluster.org/mailman/listinfo/gluster-users
>>>
>>>
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
<http://lists.gluster.org/pipermail/gluster-users/attachments/20251124/b4b1efce/attachment.html>
Gilberto Ferreira
2025-Nov-24 20:41 UTC
[Gluster-users] Failed to Establish Geo-replication Session Please check gsync config file. Unable to get statefile's name
Look! Never mind, ok???? Sorry about that, but this geo-replication is crap. Tried with gluster10 with bookworm and did the first replication and then started to get faulty status. I had installed a VM with Debian using the gluster vol replicated to the other 2 nodes, and it went well. But when I lost the primary and tried to start the VM on secondary, the VM disk crashed. I simply give up And is so sad, because over the past 3 years I spent so much time saying gluster FS is reliable and it isn't. Sorry... I am really sorry. Em seg., 24 de nov. de 2025 ?s 15:16, Gilberto Ferreira < gilberto.nunes32 at gmail.com> escreveu:> Hi there > > FYI, I tried with the gluster deb packages from github, i.e. gluster > version 11.2, and could create the geo-rep session but got faulty status > again... > > Some logs > > > [2025-11-24 18:13:33.110898] W [gsyncd(config-get):299:main] <top>: > Session config file not exists, using the default config > [{path=/var/lib/glusterd/geo-replication/VMS_gluster3_VMS-REP/gsyncd.conf}] > [2025-11-24 18:13:35.199845] I > [subcmds(monitor-status):29:subcmd_monitor_status] <top>: Monitor Status > Change [{status=Created}] > [2025-11-24 18:14:39.398460] I > [gsyncdstatus(monitor):247:set_worker_status] GeorepStatus: Worker Status > Change [{status=Initializing...}] > [2025-11-24 18:14:39.398788] I [monitor(monitor):158:monitor] Monitor: > starting gsyncd worker [{brick=/mnt/pve/data1/vms}, > {secondary_node=gluster4}] > [2025-11-24 18:14:39.491754] I [resource(worker > /mnt/pve/data1/vms):1388:connect_remote] SSH: Initializing SSH connection > between primary and secondary... > [2025-11-24 18:14:41.76723] I [resource(worker > /mnt/pve/data1/vms):1436:connect_remote] SSH: SSH connection between > primary and secondary established. [{duration=1.5848}] > [2025-11-24 18:14:41.76885] I [resource(worker > /mnt/pve/data1/vms):1117:connect] GLUSTER: Mounting gluster volume > locally... > [2025-11-24 18:14:42.108398] I [resource(worker > /mnt/pve/data1/vms):1139:connect] GLUSTER: Mounted gluster volume > [{duration=1.0314}] > [2025-11-24 18:14:42.108612] I [subcmds(worker > /mnt/pve/data1/vms):84:subcmd_worker] <top>: Worker spawn successful. > Acknowledging back to monitor > [2025-11-24 18:14:44.117002] I [primary(worker > /mnt/pve/data1/vms):1662:register] _GPrimary: Working dir > [{path=/var/lib/misc/gluster/gsyncd/VMS_gluster3_VMS-REP/mnt-pve-data1-vms}] > [2025-11-24 18:14:44.117345] I [resource(worker > /mnt/pve/data1/vms):1292:service_loop] GLUSTER: Register time > [{time=1764008084}] > [2025-11-24 18:14:44.124814] I [gsyncdstatus(worker > /mnt/pve/data1/vms):280:set_active] GeorepStatus: Worker Status Change > [{status=Active}] > [2025-11-24 18:14:44.218760] I [gsyncdstatus(worker > /mnt/pve/data1/vms):252:set_worker_crawl_status] GeorepStatus: Crawl Status > Change [{status=History Crawl}] > [2025-11-24 18:14:44.219035] I [primary(worker > /mnt/pve/data1/vms):1573:crawl] _GPrimary: starting history crawl > [{turns=1}, {stime=None}, {etime=1764008084}, {entry_stime=None}] > [2025-11-24 18:14:44.219154] I [resource(worker > /mnt/pve/data1/vms):1309:service_loop] GLUSTER: No stime available, using > xsync crawl > [2025-11-24 18:14:44.225654] I [primary(worker > /mnt/pve/data1/vms):1692:crawl] _GPrimary: starting hybrid crawl > [{stime=None}] > [2025-11-24 18:14:44.227337] I [gsyncdstatus(worker > /mnt/pve/data1/vms):252:set_worker_crawl_status] GeorepStatus: Crawl Status > Change [{status=Hybrid Crawl}] > [2025-11-24 18:14:45.227752] I [primary(worker > /mnt/pve/data1/vms):1703:crawl] _GPrimary: processing xsync changelog > [{path=/var/lib/misc/gluster/gsyncd/VMS_gluster3_VMS-REP/mnt-pve-data1-vms/xsync/XSYNC-CHANGELOG.1764008084}] > [2025-11-24 18:14:45.255496] I [primary(worker > /mnt/pve/data1/vms):1430:process] _GPrimary: Entry Time Taken [{UNL=0}, > {RMD=0}, {CRE=0}, {MKN=0}, {MKD=1}, {REN=0}, {LIN=0}, {SYM=0}, > {duration=0.0071}] > [2025-11-24 18:14:45.255639] I [primary(worker > /mnt/pve/data1/vms):1442:process] _GPrimary: Data/Metadata Time Taken > [{SETA=1}, {meta_duration=0.0078}, {SETX=0}, {XATT=0}, {DATA=0}, > {data_duration=0.0010}] > [2025-11-24 18:14:45.255797] I [primary(worker > /mnt/pve/data1/vms):1452:process] _GPrimary: Batch Completed [{mode=xsync}, > {duration=0.0277}, {changelog_start=1764008084}, > {changelog_end=1764008084}, {num_changelogs=1}, {stime=None}, > {entry_stime=None}] > [2025-11-24 18:14:45.260967] I [primary(worker > /mnt/pve/data1/vms):1699:crawl] _GPrimary: finished hybrid crawl > [{stime=(1764008084, 0)}] > [2025-11-24 18:14:45.266066] I [gsyncdstatus(worker > /mnt/pve/data1/vms):252:set_worker_crawl_status] GeorepStatus: Crawl Status > Change [{status=Changelog Crawl}] > [2025-11-24 18:14:55.280644] I [primary(worker > /mnt/pve/data1/vms):1525:crawl] _GPrimary: secondary's time > [{stime=(1764008084, 0)}] > [2025-11-24 18:14:55.645504] I [primary(worker > /mnt/pve/data1/vms):2010:syncjob] Syncer: Sync Time Taken [{job=1}, > {num_files=2}, {return_code=12}, {duration=0.0327}] > [2025-11-24 18:14:55.645687] E [syncdutils(worker > /mnt/pve/data1/vms):845:errlog] Popen: command returned error [{cmd=rsync > -aR0 --inplace --files-from=- --super --stats --numeric-ids > --no-implied-dirs --existing --xattrs --acls --ignore-missing-args . -e ssh > -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i > /var/lib/glusterd/geo-replication/secret.pem -p 22 -oControlMaster=auto -S > /tmp/gsyncd-aux-ssh-1u81eezq/dc97de7ffea2b18802fbdbd0b783902c.sock > -caes128-ctr gluster4:/proc/27188/cwd}, {error=12}] > [2025-11-24 18:14:56.111115] I [monitor(monitor):227:monitor] Monitor: > worker died in startup phase [{brick=/mnt/pve/data1/vms}] > [2025-11-24 18:14:56.122446] I > [gsyncdstatus(monitor):247:set_worker_status] GeorepStatus: Worker Status > Change [{status=Faulty}] > > > > > > Em s?b., 22 de nov. de 2025 ?s 16:20, Gilberto Ferreira < > gilberto.nunes32 at gmail.com> escreveu: > >> Here the log about the fault status >> >> [2025-11-22 19:18:56.478297] I [gsyncdstatus(worker >> /mnt/pve/data1/vms):252:set_worker_crawl_status] GeorepStatus: Crawl Status >> Change [{status=History Crawl}] >> [2025-11-22 19:18:56.478521] I [primary(worker >> /mnt/pve/data1/vms):1572:crawl] _GPrimary: starting history crawl >> [{turns=1}, {stime=(1763838427, 0)}, {etime=1763839136}, >> {entry_stime=(1763838802, 0)}] >> [2025-11-22 19:18:57.479278] I [primary(worker >> /mnt/pve/data1/vms):1604:crawl] _GPrimary: secondary's time >> [{stime=(1763838427, 0)}] >> [2025-11-22 19:18:57.922752] I [primary(worker >> /mnt/pve/data1/vms):2009:syncjob] Syncer: Sync Time Taken [{job=1}, >> {num_files=2}, {return_code=12}, {duration=0.0272}] >> [2025-11-22 19:18:57.922921] E [syncdutils(worker >> /mnt/pve/data1/vms):845:errlog] Popen: command returned error [{cmd=rsync >> -aR0 --inplace --files-from=- --super --stats --numeric-ids >> --no-implied-dirs --existing --xattrs --acls --ignore-missing-args . -e ssh >> -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i >> /var/lib/glusterd/geo-replication/secret.pem -p 22 -oControlMaster=auto -S >> /tmp/gsyncd-aux-ssh-c9h5okjo/5afb71218219138854b3c5a8eab300a4.sock >> -caes128-ctr gluster3:/proc/4004/cwd}, {error=12}] >> [2025-11-22 19:18:58.394410] I [monitor(monitor):227:monitor] Monitor: >> worker died in startup phase [{brick=/mnt/pve/data1/vms}] >> [2025-11-22 19:18:58.406208] I >> [gsyncdstatus(monitor):247:set_worker_status] GeorepStatus: Worker Status >> Change [{status=Faulty}] >> [2025-11-22 19:19:08.408871] I >> [gsyncdstatus(monitor):247:set_worker_status] GeorepStatus: Worker Status >> Change [{status=Initializing...}] >> [2025-11-22 19:19:08.408998] I [monitor(monitor):158:monitor] Monitor: >> starting gsyncd worker [{brick=/mnt/pve/data1/vms}, >> {secondary_node=gluster3}] >> [2025-11-22 19:19:08.475625] I [resource(worker >> /mnt/pve/data1/vms):1388:connect_remote] SSH: Initializing SSH connection >> between primary and secondary... >> [2025-11-22 19:19:09.719658] I [resource(worker >> /mnt/pve/data1/vms):1436:connect_remote] SSH: SSH connection between >> primary and secondary established. [{duration=1.2439}] >> [2025-11-22 19:19:09.719800] I [resource(worker >> /mnt/pve/data1/vms):1117:connect] GLUSTER: Mounting gluster volume >> locally... >> [2025-11-22 19:19:10.740213] I [resource(worker >> /mnt/pve/data1/vms):1139:connect] GLUSTER: Mounted gluster volume >> [{duration=1.0203}] >> [2025-11-22 19:19:10.740427] I [subcmds(worker >> /mnt/pve/data1/vms):84:subcmd_worker] <top>: Worker spawn successful. >> Acknowledging back to monitor >> [2025-11-22 19:19:12.756579] I [primary(worker >> /mnt/pve/data1/vms):1661:register] _GPrimary: Working dir >> [{path=/var/lib/misc/gluster/gsyncd/VMS_gluster3_VMS-REP/mnt-pve-data1-vms}] >> [2025-11-22 19:19:12.756854] I [resource(worker >> /mnt/pve/data1/vms):1292:service_loop] GLUSTER: Register time >> [{time=1763839152}] >> [2025-11-22 19:19:12.771767] I [gsyncdstatus(worker >> /mnt/pve/data1/vms):280:set_active] GeorepStatus: Worker Status Change >> [{status=Active}] >> [2025-11-22 19:19:12.834163] I [gsyncdstatus(worker >> /mnt/pve/data1/vms):252:set_worker_crawl_status] GeorepStatus: Crawl Status >> Change [{status=History Crawl}] >> [2025-11-22 19:19:12.834344] I [primary(worker >> /mnt/pve/data1/vms):1572:crawl] _GPrimary: starting history crawl >> [{turns=1}, {stime=(1763838427, 0)}, {etime=1763839152}, >> {entry_stime=(1763838802, 0)}] >> [2025-11-22 19:19:13.835162] I [primary(worker >> /mnt/pve/data1/vms):1604:crawl] _GPrimary: secondary's time >> [{stime=(1763838427, 0)}] >> [2025-11-22 19:19:14.270295] I [primary(worker >> /mnt/pve/data1/vms):2009:syncjob] Syncer: Sync Time Taken [{job=1}, >> {num_files=2}, {return_code=12}, {duration=0.0274}] >> [2025-11-22 19:19:14.270466] E [syncdutils(worker >> /mnt/pve/data1/vms):845:errlog] Popen: command returned error [{cmd=rsync >> -aR0 --inplace --files-from=- --super --stats --numeric-ids >> --no-implied-dirs --existing --xattrs --acls --ignore-missing-args . -e ssh >> -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i >> /var/lib/glusterd/geo-replication/secret.pem -p 22 -oControlMaster=auto -S >> /tmp/gsyncd-aux-ssh-1retjpoh/5afb71218219138854b3c5a8eab300a4.sock >> -caes128-ctr gluster3:/proc/4076/cwd}, {error=12}] >> [2025-11-22 19:19:14.741245] I [monitor(monitor):227:monitor] Monitor: >> worker died in startup phase [{brick=/mnt/pve/data1/vms}] >> [2025-11-22 19:19:14.752452] I >> [gsyncdstatus(monitor):247:set_worker_status] GeorepStatus: Worker Status >> Change [{status=Faulty}] >> >> >> It seems to me that it failed to open something via ssh session. >> >> I don't know... something like that. >> >> >> Em s?b., 22 de nov. de 2025 ?s 16:17, Gilberto Ferreira < >> gilberto.nunes32 at gmail.com> escreveu: >> >>> Hi >>> >>> I had succeeded in creating the session with the other side. >>> But I got a Faulty status >>> I the gsyncd.log I got this: >>> >>> Popen: command returned error [{cmd=rsync -aR0 --inplace --files-from=- >>> --super --stats --numeric-ids --no-implied-dirs --existing --xattrs --acls >>> --ignore-missing-args . -e ssh -oPasswordAuthentication=no >>> -oStrictHostKeyChecking=no -i /var/lib/glusterd/geo-replication/secret.pem >>> -p 22 -oControlMaster=auto -S >>> /tmp/gsyncd-aux-ssh-2j2yeofa/5afb71218219138854b3c5a8eab300a4.sock >>> -caes128-ctr gluster3:/proc/2662/cwd}, {error=12}] >>> >>> >>> This also happens with Gluster 12dev. >>> >>> After compiling the gluster 12dev, I successfully create a geo-rep >>> session but get the error above. >>> >>> >>> Any clue? >>> >>> >>> Best Regards >>> >>> >>> >>> >>> >>> Em s?b., 22 de nov. de 2025 ?s 15:12, Strahil Nikolov < >>> hunter86_bg at yahoo.com> escreveu: >>> >>>> Hi Gilberto, >>>> >>>> It should as long as it's the same problem. >>>> >>>> It will be nice to share your experience in the mailing list. >>>> >>>> Best Regards, >>>> Strahil Nikolov >>>> >>>> On Sat, Nov 22, 2025 at 18:13, Gilberto Ferreira >>>> <gilberto.nunes32 at gmail.com> wrote: >>>> Hi >>>> >>>> Should it work with Debian Trixie (13) as well? >>>> >>>> I will try it >>>> >>>> >>>> >>>> --- >>>> >>>> >>>> Gilberto Nunes Ferreira >>>> +55 (47) 99676-7530 - Whatsapp / Telegram >>>> >>>> >>>> >>>> >>>> >>>> >>>> Em s?b., 22 de nov. de 2025 ?s 12:37, Strahil Nikolov < >>>> hunter86_bg at yahoo.com> escreveu: >>>> >>>> Hi Gilberto, >>>> >>>> I think debian12 packages don't have >>>> https://github.com/gluster/glusterfs/pull/4404/commits/c433a178e8208e1771fea4d61d0a22a95b8bc74b >>>> >>>> Run on source and destination this command and try again: >>>> sed -i 's/readfp/read_file/g' >>>> /usr/lib/x86_64-linux-gnu/glusterfs/python/syncdaemon/gsyncdconfig.py >>>> >>>> In my test setup where source and destinations are each a single debian >>>> 12 with a hackishl created gluster_shared_storage volume and executing >>>> the sed, I got: >>>> >>>> >>>> # gluster volume geo-replication vol1 geoaccount at gluster2::georep >>>> create push-pem >>>> Creating geo-replication session between vol1 & geoaccount at gluster2::georep >>>> has been successful >>>> >>>> >>>> Best Regards, >>>> Strahil Nikolov >>>> >>>> ? ??????, 22 ??????? 2025 ?. ? 16:34:30 ?. ???????+2, Gilberto Ferreira >>>> <gilberto.nunes32 at gmail.com> ??????: >>>> >>>> >>>> here the story about gluster 12dev faulty error: >>>> >>>> https://github.com/gluster/glusterfs/issues/4632 >>>> >>>> >>>> >>>> >>>> >>>> Em s?b., 22 de nov. de 2025 ?s 10:59, Gilberto Ferreira < >>>> gilberto.nunes32 at gmail.com> escreveu: >>>> >>>> Hello there >>>> My testing was with proxmox 9 which has Debian 13. >>>> I tried with gluster 11.1 from Debian repo and then version 11.2 from >>>> git repo. >>>> I got statefile's name issue with both. >>>> Then I compiled the version 12dev and could create geo-replication >>>> session with success but got faulty status >>>> >>>> So that's it. >>>> >>>> --- >>>> Gilberto Nunes Ferreira >>>> +55 (47) 99676-7530 >>>> Proxmox VE >>>> >>>> Em s?b., 22 de nov. de 2025, 10:14, Strahil Nikolov < >>>> hunter86_bg at yahoo.com> escreveu: >>>> >>>> Hi Gilberto, >>>> >>>> What version of os and gluster do you use exaclty ? >>>> >>>> Best Regards, >>>> Strahil Nikolov >>>> >>>> ? ?????, 21 ??????? 2025 ?. ? 14:08:19 ?. ???????+2, Gilberto Ferreira < >>>> gilberto.nunes32 at gmail.com> ??????: >>>> >>>> >>>> Hello there >>>> >>>> If there is something else I could help, please let me know. >>>> >>>> Thanks >>>> >>>> Best Regards >>>> >>>> >>>> >>>> >>>> >>>> >>>> Em qua., 19 de nov. de 2025 ?s 15:21, Gilberto Ferreira < >>>> gilberto.nunes32 at gmail.com> escreveu: >>>> >>>> Hi there >>>> >>>> So there is no special script. >>>> First I try to using this: >>>> https://github.com/aravindavk/gluster-georep-tools, and then I notice >>>> the issue. >>>> But after try do it by myself, I called for help. >>>> I tried: >>>> >>>> gluster volume geo-replication MASTERVOL root at SLAVENODE::slavevol create push-pem >>>> >>>> gluster volume geo-replication MASTERVOL root at SLAVENODE::slavevol start >>>> >>>> >>>> And got the issue. >>>> >>>> >>>> Thanks >>>> >>>> >>>> --- >>>> >>>> >>>> Gilberto Nunes Ferreira >>>> +55 (47) 99676-7530 - Whatsapp / Telegram >>>> >>>> >>>> >>>> >>>> >>>> >>>> Em qua., 19 de nov. de 2025 ?s 15:09, Strahil Nikolov < >>>> hunter86_bg at yahoo.com> escreveu: >>>> >>>> Hi Gilberto, >>>> >>>> I have no idea why my previous message was not sent (sorry about that). >>>> I suspect it's a bug. If you have some script or ansible playbook for >>>> the setup, it could help me reproduce it locally. >>>> >>>> >>>> Best Regards, >>>> Strahil Nikolov >>>> >>>> ? ???????, 11 ??????? 2025 ?. ? 16:55:59 ?. ???????+2, Gilberto >>>> Ferreira <gilberto.nunes32 at gmail.com> ??????: >>>> >>>> >>>> Any clue about this issue? >>>> >>>> >>>> >>>> >>>> >>>> >>>> Em seg., 10 de nov. de 2025 ?s 14:47, Gilberto Ferreira < >>>> gilberto.nunes32 at gmail.com> escreveu: >>>> >>>> I still don't get it, because with Debian BookWorm and Gluster 10, >>>> geo-rep works perfectly. >>>> It's something about Trixie and Gluster 11.x. >>>> >>>> >>>> >>>> >>>> >>>> Em seg., 10 de nov. de 2025 ?s 14:45, Karl Kleinpaste < >>>> karl at kleinpaste.org> escreveu: >>>> >>>> On 11/10/25 12:21 PM, Gilberto Ferreira wrote: >>>> >>>> And yes. With gluster 11.2 from github repo, the very some error: >>>> gluster vol geo VMS gluster3::VMS-REP create push-pem >>>> Please check gsync config file. Unable to get statefile's name >>>> geo-replication command failed >>>> >>>> >>>> I had this problem a year ago, Aug 2024 >>>> <https://lists.gluster.org/pipermail/gluster-users/2024-August/040625.html>. >>>> I went rounds and rounds with Strahil for a week, trying to find why I >>>> couldn't cross the finish line of successful georep. It always ends in: >>>> >>>> Please check gsync config file. Unable to get statefile's name >>>> geo-replication command failed >>>> >>>> The volumes were set up properly, the commands for georep were done >>>> correctly, per guidelines, but georep was left forever in a state of >>>> Created, never Active. >>>> >>>> Finally I just gave up. I can't use gluster if it won't work with me. I >>>> found that gluster would not give enough adequate diagnostics to provide a ( >>>> *useful*!) explanation of what is actually wrong. >>>> ________ >>>> >>>> >>>> >>>> Community Meeting Calendar: >>>> >>>> Schedule - >>>> Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC >>>> Bridge: https://meet.google.com/cpu-eiue-hvk >>>> Gluster-users mailing list >>>> Gluster-users at gluster.org >>>> https://lists.gluster.org/mailman/listinfo/gluster-users >>>> >>>> ________ >>>> >>>> >>>> >>>> Community Meeting Calendar: >>>> >>>> Schedule - >>>> Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC >>>> Bridge: https://meet.google.com/cpu-eiue-hvk >>>> Gluster-users mailing list >>>> Gluster-users at gluster.org >>>> https://lists.gluster.org/mailman/listinfo/gluster-users >>>> >>>>-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20251124/51dda777/attachment.html>