Displaying 20 results from an estimated 6000 matches similar to: "Heal not working"
2014 Sep 05
2
glusterfs replica volume self heal dir very slow!!why?
Hi all?
I do the following test?
I create a glusterfs replica volume (replica count is 2 ) with two server node(server A and server B)? then mount the volume in client node?
then? I shut down the network of server A node? in client node? I copy a dir?which has a lot of small files?? the dir size is 2.9GByte?
when copy finish? I start the network of server A node? now?
2013 Jul 09
2
Gluster Self Heal
Hi,
I have a 2-node gluster with 3 TB storage.
1) I believe the "glusterfsd" is responsible for the self healing between the 2 nodes.
2) Due to some network error, the replication stopped for some reason but the application was accessing the data from node1. When I manually try to start "glusterfsd" service, its not starting.
Please advice on how I can maintain
2013 Nov 29
1
Self heal problem
Hi,
I have a glusterfs volume replicated on three nodes. I am planing to use
the volume as storage for vMware ESXi machines using NFS. The reason for
using tree nodes is to be able to configure Quorum and avoid
split-brains. However, during my initial testing when intentionally and
gracefully restart the node "ned", a split-brain/self-heal error
occurred.
The log on "todd"
2013 Dec 09
0
Gluster - replica - Unable to self-heal contents of '/' (possible split-brain)
Hello,
I''m trying to build a replica volume, on two servers.
The servers are: blade6 and blade7. (another blade1 in the peer, but with
no volumes)
The volume seems ok, but I cannot mount it from NFS.
Here are some logs:
[root@blade6 stor1]# df -h
/dev/mapper/gluster_stor1 882G 200M 837G 1% /gluster/stor1
[root@blade7 stor1]# df -h
/dev/mapper/gluster_fast
2017 Jul 11
1
Replica 3 with arbiter - heal error?
Hello,
I have a Gluster 3.8.13 with replica 3 arbiter volume mounted and run
there a following script:
while true; do echo "$(date)" >> a.txt; sleep 2; done
After few seconds I add a rule to the firewall on the client, that
blocks access to node specified during mount e.g. if volume is mounted
with:
mount -t glusterfs -o backupvolfile-server=10.0.0.2 10.0.0.1:/vol /mnt/vol
I
2013 Jul 24
2
Healing in glusterfs 3.3.1
Hi,
I have a glusterfs 3.3.1 setup with 2 servers and a replicated volume used
by 4 clients.
Sometimes from some clients I can't access some of the files. After I force
a full heal on the brick I see several files healed. Is this behavior
normal?
Thanks
--
Paulo Silva <paulojjs at gmail.com>
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
2013 Apr 30
1
Volume heal daemon 3.4alpha3
gluster> volume heal dyn_coldfusion
Self-heal daemon is not running. Check self-heal daemon log file.
gluster>
Is there a specific log? When i check /var/log/glusterfs/glustershd.log
glustershd.log:[2013-04-30 15:51:40.463259] E
[afr-self-heald.c:409:_crawl_proceed] 0-dyn_coldfusion-replicate-0:
Stopping crawl for dyn_coldfusion-client-1 , subvol went down
Is there a specific log? When
2013 May 10
2
Self-heal and high load
Hi all,
I'm pretty new to Gluster, and the company I work for uses it for storage
across 2 data centres. An issue has cropped up fairly recently with regards
to the self-heal mechanism.
Occasionally the connection between these 2 Gluster servers breaks or drops
momentarily. Due to the nature of the business it's highly likely that
files have been written during this time. When the
2017 Jul 25
1
stuck heal process
Good Morning!
We are running RedHat 7.3 with
glusterfs-server-3.8.4-18.el7rhgs.x86_64. Not sure if your able to
help with this version or not.
I have a 5 node setup with 1 node having no storage and only acting as
a quorum node. We have a mis of direct attached storage and iscsi SAN
storage.
We have distributed replica volumes created across all 4 nodes.
At some point last week one of the
2013 Oct 26
1
Crashing (signal received: 11)
I am seeing this crashing happening, I am working on the self healing errors as well, not sure if the two are related. I would appreciate any direction on trying to resolve the issue, I have clients dropping connection daily.
[2013-10-26 15:35:46.935903] E [afr-self-heal-common.c:2160:afr_self_heal_completion_cbk] 0-ENTV04EP-replicate-9: background meta-data self-heal failed on /
[2013-10-26
2006 Apr 28
1
errors while testing with imaptest
I get lots of these (in error log):
dovecot: Apr 28 14:50:03 Error: IMAP(jenslt): Corrupted index cache file
/home0/jenslt/.imap/INBOX/dovecot.index.cache: record points outside file
And some of these:
dovecot: Apr 28 14:50:03 Error: IMAP(jenslt): Corrupted index cache file
/home0/jenslt/.imap/INBOX/dovecot.index.cache: field index too large
(2037149797 >= 19)
Should I worry or not?
2013 Dec 03
3
Self Heal Issue GlusterFS 3.3.1
Hi,
I'm running glusterFS 3.3.1 on Centos 6.4.
? Gluster volume status
Status of volume: glustervol
Gluster process Port Online Pid
------------------------------------------------------------------------------
Brick KWTOCUATGS001:/mnt/cloudbrick 24009 Y 20031
Brick KWTOCUATGS002:/mnt/cloudbrick
2018 Jan 17
1
Gluster endless heal
Hi,
I have an issue with Gluster 3.8.14.
The cluster is 4 nodes with replica count 2, on of the nodes went offline for around 15 minutes, when it came back online, self heal triggered and it just did not stop afterward, it's been running for 3 days now, maxing the bricks utilization without actually healing anything.
The bricks are all SSDs, and the logs of the source node is spamming with
2018 Mar 20
0
Disperse volume recovery and healing
On Tue, Mar 20, 2018 at 5:26 AM, Victor T <hero_of_nothing_1 at hotmail.com>
wrote:
> That makes sense. In the case of "file damage," it would show up as files
> that could not be healed in logfiles or gluster volume heal [volume] info?
>
If the damage affects more bricks than the volume redundancy, then probably
yes. These files or directories will appear in
2017 Oct 26
0
not healing one file
Hi Richard,
Thanks for the informations. As you said there is gfid mismatch for the
file.
On brick-1 & brick-2 the gfids are same & on brick-3 the gfid is different.
This is not considered as split-brain because we have two good copies here.
Gluster 3.10 does not have a method to resolve this situation other than the
manual intervention [1]. Basically what you need to do is remove the
2023 Oct 24
0
Gluster heal script?
Hello all.
Is there a script to help healing files that remain in heal info even
after a pass with heal full ?
I recently (~august) restarted from scratch our Gluster cluster in
"replica 3 arbiter 1" but I already found some files that are not
healing and inaccessible (socket not connected) from the fuse mount.
volume info:
-8<--
Volume Name: cluster_data
Type:
2017 Sep 17
2
Volume Heal issue
Hi all,
I have a replica 3 with 1 arbiter.
I see the last days that one file at a volume is always showing as needing
healing:
gluster volume heal vms info
Brick gluster0:/gluster/vms/brick
Status: Connected
Number of entries: 0
Brick gluster1:/gluster/vms/brick
Status: Connected
Number of entries: 0
Brick gluster2:/gluster/vms/brick
*<gfid:66d3468e-00cf-44dc-a835-7624da0c5370>*
Status:
2017 Sep 17
0
Volume Heal issue
I am using gluster 3.8.12, the default on Centos 7.3
(I will update to 3.10 at some moment)
On Sun, Sep 17, 2017 at 11:30 AM, Alex K <rightkicktech at gmail.com> wrote:
> Hi all,
>
> I have a replica 3 with 1 arbiter.
>
> I see the last days that one file at a volume is always showing as needing
> healing:
>
> gluster volume heal vms info
> Brick
2018 Mar 18
1
Disperse volume recovery and healing
No. After bringing up one brick and before stopping the next one, you need to be sure that there are no damaged files. You shouldn't reboot a node if "gluster volume heal <volname> info" shows damaged files.
What happens in this case then? I'm thinking about a situation where the servers are kept in an environment that we don't control - i.e. the cloud. If the VMs are
2018 Feb 09
0
self-heal trouble after changing arbiter brick
Hey,
Did the heal completed and you still have some entries pending heal?
If yes then can you provide the following informations to debug the issue.
1. Which version of gluster you are running
2. gluster volume heal <volname> info summary or gluster volume heal
<volname> info
3. getfattr -d -e hex -m . <filepath-on-brick> output of any one of the
which is pending heal from all