hi Andreas,
The directory is in split-brain. Do you have any
files/directories, that are in split-brain in the directory
'media/ga/live/a' ??
Could you give output of
"getfattr -d -m. -e hex media/ga/live/a/*" on both the bricks?
Pranith
On 01/04/2016 05:21 PM, Andreas Tsaridas wrote:> Hello,
>
> Please see below :
> -----
>
> web01 # getfattr -d -m . -e hex media/ga/live/a
> # file: media/ga/live/a
> trusted.afr.dirty=0x000000000000000000000000
> trusted.afr.remote1=0x000000000000000000000000
> trusted.afr.remote2=0x000000000000000000000005
> trusted.afr.share-client-0=0x000000000000000000000000
> trusted.afr.share-client-1=0x0000000000000000000000ee
> trusted.gfid=0xb13199a1464c44918464444b3f7eeee3
> trusted.glusterfs.dht=0x000000010000000000000000ffffffff
>
>
> ------
>
> web02 # getfattr -d -m . -e hex media/ga/live/a
> # file: media/ga/live/a
> trusted.afr.dirty=0x000000000000000000000000
> trusted.afr.remote1=0x000000000000000000000008
> trusted.afr.remote2=0x000000000000000000000000
> trusted.afr.share-client-0=0x000000000000000000000000
> trusted.afr.share-client-1=0x000000000000000000000000
> trusted.gfid=0xb13199a1464c44918464444b3f7eeee3
> trusted.glusterfs.dht=0x000000010000000000000000ffffffff
>
> ------
>
> Regards,
> AT
>
> On Mon, Jan 4, 2016 at 12:44 PM, Krutika Dhananjay
> <kdhananj at redhat.com <mailto:kdhananj at redhat.com>> wrote:
>
> Hi,
>
> Could you share the output of
> # getfattr -d -m . -e hex <abs-path-to-media/ga/live/a>
>
> from both the bricks?
>
> -Krutika
>
------------------------------------------------------------------------
>
> *From: *"Andreas Tsaridas" <andreas.tsaridas at
gmail.com
> <mailto:andreas.tsaridas at gmail.com>>
> *To: *gluster-users at gluster.org <mailto:gluster-users at
gluster.org>
> *Sent: *Monday, January 4, 2016 5:10:58 PM
> *Subject: *[Gluster-users] folder not being healed
>
>
> Hello,
>
> I have a cluster of two replicated nodes in glusterfs 3.6.3 in
> RedHat 6.6. Problem is that a specific folder is always trying
> to be healed but never gets healed. This has been going on for
> 2 weeks now.
>
> -----
>
> # gluster volume status
> Status of volume: share
> Gluster processPortOnlinePid
>
------------------------------------------------------------------------------
> Brick 172.16.4.1:/srv/share/glusterfs49152Y10416
> Brick 172.16.4.2:/srv/share/glusterfs49152Y19907
> NFS Server on localhost2049Y22664
> Self-heal Daemon on localhostN/AY22676
> NFS Server on 172.16.4.22049Y19923
> Self-heal Daemon on 172.16.4.2N/AY19937
>
> Task Status of Volume share
>
------------------------------------------------------------------------------
> There are no active volume tasks
>
> ------
>
> # gluster volume info
>
> Volume Name: share
> Type: Replicate
> Volume ID: 17224664-645c-48b7-bc3a-b8fc84c6ab30
> Status: Started
> Number of Bricks: 1 x 2 = 2
> Transport-type: tcp
> Bricks:
> Brick1: 172.16.4.1:/srv/share/glusterfs
> Brick2: 172.16.4.2:/srv/share/glusterfs
> Options Reconfigured:
> cluster.background-self-heal-count: 20
> cluster.heal-timeout: 2
> performance.normal-prio-threads: 64
> performance.high-prio-threads: 64
> performance.least-prio-threads: 64
> performance.low-prio-threads: 64
> performance.flush-behind: off
> performance.io-thread-count: 64
>
> ------
>
> # gluster volume heal share info
> Brick web01.rsdc:/srv/share/glusterfs/
> /media/ga/live/a - Possibly undergoing heal
>
> Number of entries: 1
>
> Brick web02.rsdc:/srv/share/glusterfs/
> Number of entries: 0
>
> -------
>
> # gluster volume heal share info split-brain
> Gathering list of split brain entries on volume share has been
> successful
>
> Brick 172.16.4.1:/srv/share/glusterfs
> Number of entries: 0
>
> Brick 172.16.4.2:/srv/share/glusterfs
> Number of entries: 0
>
> -------
>
> ==> /var/log/glusterfs/glustershd.log <=>
[2016-01-04 11:35:33.004831] I
> [afr-self-heal-entry.c:554:afr_selfheal_entry_do]
> 0-share-replicate-0: performing entry selfheal on
> b13199a1-464c-4491-8464-444b3f7eeee3
> [2016-01-04 11:36:07.449192] W
> [client-rpc-fops.c:2772:client3_3_lookup_cbk]
> 0-share-client-1: remote operation failed: No data available.
> Path: (null) (00000000-0000-0000-0000-000000000000)
> [2016-01-04 11:36:07.449706] W
> [client-rpc-fops.c:240:client3_3_mknod_cbk] 0-share-client-1:
> remote operation failed: File exists. Path: (null)
>
> Could you please advise ?
>
> Kind regards,
>
> AT
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org <mailto:Gluster-users at
gluster.org>
> http://www.gluster.org/mailman/listinfo/gluster-users
>
>
>
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
<http://www.gluster.org/pipermail/gluster-users/attachments/20160104/9fdf0a24/attachment.html>
Hello, Unfortunately I get : -bash: /usr/bin/getfattr: Argument list too long There are a lot of file in these directories and even ls takes a long time to show results. How would I be able to keep the copy from web01 and discard the other ? Thanks On Mon, Jan 4, 2016 at 3:59 PM, Pranith Kumar Karampuri <pkarampu at redhat.com> wrote:> hi Andreas, > The directory is in split-brain. Do you have any > files/directories, that are in split-brain in the directory > 'media/ga/live/a' ?? > > Could you give output of > "getfattr -d -m. -e hex media/ga/live/a/*" on both the bricks? > > Pranith > > > On 01/04/2016 05:21 PM, Andreas Tsaridas wrote: > > Hello, > > Please see below : > ----- > > web01 # getfattr -d -m . -e hex media/ga/live/a > # file: media/ga/live/a > trusted.afr.dirty=0x000000000000000000000000 > trusted.afr.remote1=0x000000000000000000000000 > trusted.afr.remote2=0x000000000000000000000005 > trusted.afr.share-client-0=0x000000000000000000000000 > trusted.afr.share-client-1=0x0000000000000000000000ee > trusted.gfid=0xb13199a1464c44918464444b3f7eeee3 > trusted.glusterfs.dht=0x000000010000000000000000ffffffff > > > ------ > > web02 # getfattr -d -m . -e hex media/ga/live/a > # file: media/ga/live/a > trusted.afr.dirty=0x000000000000000000000000 > trusted.afr.remote1=0x000000000000000000000008 > trusted.afr.remote2=0x000000000000000000000000 > trusted.afr.share-client-0=0x000000000000000000000000 > trusted.afr.share-client-1=0x000000000000000000000000 > trusted.gfid=0xb13199a1464c44918464444b3f7eeee3 > trusted.glusterfs.dht=0x000000010000000000000000ffffffff > > ------ > > Regards, > AT > > On Mon, Jan 4, 2016 at 12:44 PM, Krutika Dhananjay <kdhananj at redhat.com> > wrote: > >> Hi, >> >> Could you share the output of >> # getfattr -d -m . -e hex <abs-path-to-media/ga/live/a> >> >> from both the bricks? >> >> -Krutika >> ------------------------------ >> >> *From: *"Andreas Tsaridas" <andreas.tsaridas at gmail.com> >> *To: *gluster-users at gluster.org >> *Sent: *Monday, January 4, 2016 5:10:58 PM >> *Subject: *[Gluster-users] folder not being healed >> >> >> Hello, >> >> I have a cluster of two replicated nodes in glusterfs 3.6.3 in RedHat >> 6.6. Problem is that a specific folder is always trying to be healed but >> never gets healed. This has been going on for 2 weeks now. >> >> ----- >> >> # gluster volume status >> Status of volume: share >> Gluster process Port Online Pid >> >> ------------------------------------------------------------------------------ >> Brick 172.16.4.1:/srv/share/glusterfs 49152 Y 10416 >> Brick 172.16.4.2:/srv/share/glusterfs 49152 Y 19907 >> NFS Server on localhost 2049 Y 22664 >> Self-heal Daemon on localhost N/A Y 22676 >> NFS Server on 172.16.4.2 2049 Y 19923 >> Self-heal Daemon on 172.16.4.2 N/A Y 19937 >> >> Task Status of Volume share >> >> ------------------------------------------------------------------------------ >> There are no active volume tasks >> >> ------ >> >> # gluster volume info >> >> Volume Name: share >> Type: Replicate >> Volume ID: 17224664-645c-48b7-bc3a-b8fc84c6ab30 >> Status: Started >> Number of Bricks: 1 x 2 = 2 >> Transport-type: tcp >> Bricks: >> Brick1: 172.16.4.1:/srv/share/glusterfs >> Brick2: 172.16.4.2:/srv/share/glusterfs >> Options Reconfigured: >> cluster.background-self-heal-count: 20 >> cluster.heal-timeout: 2 >> performance.normal-prio-threads: 64 >> performance.high-prio-threads: 64 >> performance.least-prio-threads: 64 >> performance.low-prio-threads: 64 >> performance.flush-behind: off >> performance.io-thread-count: 64 >> >> ------ >> >> # gluster volume heal share info >> Brick web01.rsdc:/srv/share/glusterfs/ >> /media/ga/live/a - Possibly undergoing heal >> >> Number of entries: 1 >> >> Brick web02.rsdc:/srv/share/glusterfs/ >> Number of entries: 0 >> >> ------- >> >> # gluster volume heal share info split-brain >> Gathering list of split brain entries on volume share has been successful >> >> Brick 172.16.4.1:/srv/share/glusterfs >> Number of entries: 0 >> >> Brick 172.16.4.2:/srv/share/glusterfs >> Number of entries: 0 >> >> ------- >> >> ==> /var/log/glusterfs/glustershd.log <=>> [2016-01-04 11:35:33.004831] I >> [afr-self-heal-entry.c:554:afr_selfheal_entry_do] 0-share-replicate-0: >> performing entry selfheal on b13199a1-464c-4491-8464-444b3f7eeee3 >> [2016-01-04 11:36:07.449192] W >> [client-rpc-fops.c:2772:client3_3_lookup_cbk] 0-share-client-1: remote >> operation failed: No data available. Path: (null) >> (00000000-0000-0000-0000-000000000000) >> [2016-01-04 11:36:07.449706] W >> [client-rpc-fops.c:240:client3_3_mknod_cbk] 0-share-client-1: remote >> operation failed: File exists. Path: (null) >> >> Could you please advise ? >> >> Kind regards, >> >> AT >> >> _______________________________________________ >> Gluster-users mailing list >> Gluster-users at gluster.org >> http://www.gluster.org/mailman/listinfo/gluster-users >> >> >> > > > _______________________________________________ > Gluster-users mailing listGluster-users at gluster.orghttp://www.gluster.org/mailman/listinfo/gluster-users > > >-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160104/9481b367/attachment.html>