search for: maby

Displaying 20 results from an estimated 349 matches for "maby".

Did you mean: baby
2017 Aug 28
2
self-heal not working
Thank you for the command. I ran it on all my nodes and now finally the the self-heal daemon does not report any files to be healed. Hopefully this scenario can get handled properly in newer versions of GlusterFS. > -------- Original Message -------- > Subject: Re: [Gluster-users] self-heal not working > Local Time: August 28, 2017 10:41 AM > UTC Time: August 28, 2017 8:41 AM >
2017 Aug 28
3
self-heal not working
Excuse me for my naive questions but how do I reset the afr.dirty xattr on the file to be healed? and do I need to do that through a FUSE mount? or simply on every bricks directly? > -------- Original Message -------- > Subject: Re: [Gluster-users] self-heal not working > Local Time: August 28, 2017 5:58 AM > UTC Time: August 28, 2017 3:58 AM > From: ravishankar at redhat.com >
2017 Aug 28
0
self-heal not working
Great, can you raise a bug for the issue so that it is easier to keep track (plus you'll be notified if the patch is posted) of it? The general guidelines are @ https://gluster.readthedocs.io/en/latest/Contributors-Guide/Bug-Reporting-Guidelines but you just need to provide whatever you described in this email thread in the bug: i.e. volume info, heal info, getfattr and stat output of
2017 Aug 28
0
self-heal not working
On 08/28/2017 01:29 PM, mabi wrote: > Excuse me for my naive questions but how do I reset the afr.dirty > xattr on the file to be healed? and do I need to do that through a > FUSE mount? or simply on every bricks directly? > > Directly on the bricks: `setfattr -n trusted.afr.dirty -v 0x000000000000000000000000
2017 Aug 27
2
self-heal not working
----- Original Message ----- > From: "mabi" <mabi at protonmail.ch> > To: "Ravishankar N" <ravishankar at redhat.com> > Cc: "Ben Turner" <bturner at redhat.com>, "Gluster Users" <gluster-users at gluster.org> > Sent: Sunday, August 27, 2017 3:15:33 PM > Subject: Re: [Gluster-users] self-heal not working > >
2017 Aug 28
0
self-heal not working
On 08/28/2017 01:57 AM, Ben Turner wrote: > ----- Original Message ----- >> From: "mabi" <mabi at protonmail.ch> >> To: "Ravishankar N" <ravishankar at redhat.com> >> Cc: "Ben Turner" <bturner at redhat.com>, "Gluster Users" <gluster-users at gluster.org> >> Sent: Sunday, August 27, 2017 3:15:33 PM >>
2017 Aug 27
0
self-heal not working
Thanks Ravi for your analysis. So as far as I understand nothing to worry about but my question now would be: how do I get rid of this file from the heal info? > -------- Original Message -------- > Subject: Re: [Gluster-users] self-heal not working > Local Time: August 27, 2017 3:45 PM > UTC Time: August 27, 2017 1:45 PM > From: ravishankar at redhat.com > To: mabi <mabi at
2017 Aug 27
2
self-heal not working
Yes, the shds did pick up the file for healing (I saw messages like " got entry: 1985e233-d5ee-4e3e-a51a-cf0b5f9f2aea") but no error afterwards. Anyway I reproduced it by manually setting the afr.dirty bit for a zero byte file on all 3 bricks. Since there are no afr pending xattrs indicating good/bad copies and all files are zero bytes, the data self-heal algorithm just picks the
2017 Aug 24
2
self-heal not working
Thanks for confirming the command. I have now enabled DEBUG client-log-level, run a heal and then attached the glustershd log files of all 3 nodes in this mail. The volume concerned is called myvol-pro, the other 3 volumes have no problem so far. Also note that in the mean time it looks like the file has been deleted by the user and as such the heal info command does not show the file name
2017 Aug 23
2
self-heal not working
I just saw the following bug which was fixed in 3.8.15: https://bugzilla.redhat.com/show_bug.cgi?id=1471613 Is it possible that the problem I described in this post is related to that bug? > -------- Original Message -------- > Subject: Re: [Gluster-users] self-heal not working > Local Time: August 22, 2017 11:51 AM > UTC Time: August 22, 2017 9:51 AM > From: ravishankar at
2017 Aug 22
3
self-heal not working
Thanks for the additional hints, I have the following 2 questions first: - In order to launch the index heal is the following command correct: gluster volume heal myvolume - If I run a "volume start force" will it have any short disruptions on my clients which mount the volume through FUSE? If yes, how long? This is a production system that's why I am asking. > --------
2017 Aug 25
0
self-heal not working
Hi Ravi, Did you get a chance to have a look at the log files I have attached in my last mail? Best, Mabi > -------- Original Message -------- > Subject: Re: [Gluster-users] self-heal not working > Local Time: August 24, 2017 12:08 PM > UTC Time: August 24, 2017 10:08 AM > From: mabi at protonmail.ch > To: Ravishankar N <ravishankar at redhat.com> > Ben Turner
2017 Aug 22
0
self-heal not working
On 08/22/2017 02:30 PM, mabi wrote: > Thanks for the additional hints, I have the following 2 questions first: > > - In order to launch the index heal is the following command correct: > gluster volume heal myvolume > Yes > - If I run a "volume start force" will it have any short disruptions > on my clients which mount the volume through FUSE? If yes, how long?
2017 Aug 24
0
self-heal not working
Unlikely. In your case only the afr.dirty is set, not the afr.volname-client-xx xattr. `gluster volume set myvolume diagnostics.client-log-level DEBUG` is right. On 08/23/2017 10:31 PM, mabi wrote: > I just saw the following bug which was fixed in 3.8.15: > > https://bugzilla.redhat.com/show_bug.cgi?id=1471613 > > Is it possible that the problem I described in this post is
2017 Aug 04
1
Quotas not working after adding arbiter brick to replica 2
Thank you very much Sanoj, I ran your script once and it worked. I now have quotas again... Question: do you know in which release this issue will be fixed? > -------- Original Message -------- > Subject: Re: [Gluster-users] Quotas not working after adding arbiter brick to replica 2 > Local Time: August 4, 2017 3:28 PM > UTC Time: August 4, 2017 1:28 PM > From: sunnikri at
2017 Aug 03
2
Quotas not working after adding arbiter brick to replica 2
I tried to re-create manually my quotas but not even that works now. Running the "limit-usage" command as showed below returns success: $ sudo gluster volume quota myvolume limit-usage /userdirectory 50GB volume quota : success but when I list the quotas using "list" nothing appears. What can I do to fix that issue with the quotas? > -------- Original Message -------- >
2017 Aug 04
0
Quotas not working after adding arbiter brick to replica 2
Hi mabi, This is a likely issue where the last gfid entry in the quota.conf file is stale (because the directory was deleted with quota limit on it being removed) (https://review.gluster.org/#/c/16507/) To fix the issue, we need to remove the last entry (last 17 bytes/ 16bytes based on quota version) in the file. Please use the below work around for the same until next upgrade. you only need to
2017 Jul 31
2
Possible stale .glusterfs/indices/xattrop file?
Now I understand what you mean the the "-samefile" parameter of "find". As requested I have now run the following command on all 3 nodes with the ouput of all 3 nodes below: sudo find /data/myvolume/brick -samefile /data/myvolume/brick/.glusterfs/29/e0/29e0d13e-1217-41cc-9bda-1fbbf781c397 -ls node1: 8404683 0 lrwxrwxrwx 1 root root 66 Jul 27 15:43
2017 Aug 21
2
self-heal not working
Sure, it doesn't look like a split brain based on the output: Brick node1.domain.tld:/data/myvolume/brick Status: Connected Number of entries in split-brain: 0 Brick node2.domain.tld:/data/myvolume/brick Status: Connected Number of entries in split-brain: 0 Brick node3.domain.tld:/srv/glusterfs/myvolume/brick Status: Connected Number of entries in split-brain: 0 > -------- Original
2020 Jan 27
3
imap panic: file array.c: line 10 (array_idx_modifiable_i)
Hello, Does anyone one have any idea? Aki maybe? Regards, Mabi ??????? Original Message ??????? On Wednesday, January 22, 2020 11:37 AM, mabi <mabi at protonmail.ch> wrote: > Hello, > > Are there any chances that I could fix this issue by running a "doveadm force-rsync" (https://wiki2.dovecot.org/Tools/Doveadm/ForceResync) on this problematic mailbox? > > Best