Hello, I am using gluster 3.7.2 on docker 1.6.2 on centos 7.1. I had 2 brick replication gluster, but I lost one of them, so I added a new one and deleted from replica the old one. When I check the status shows this gluster volume info Volume Name: datastore Type: Replicate Volume ID: 4dd0a307-e44c-4655-9216-8d470f3a0d33 Status: Started Number of Bricks: 1 x 2 = 2 Transport-type: tcp Bricks: Brick1: gfs-2.xx.xx.xx:/raw Brick2: gfs-3.xx.xx.xx:/raw Options Reconfigured: cluster.self-heal-daemon: enable When I mount with the client from other servers to add data to gluster works fine, and the new data are replicated without problems. The issue is that I haven't the old data from gfs-2 on gfs-3. Both bricks are probed and on pool list. Connected and UUIDs are fine. When I check volume with status I have this: gluster volume status Status of volume: datastore Gluster process TCP Port RDMA Port Online Pid ------------------------------------------------------------------------------ Brick gfs-2.xx.xx.xx:/raw 49152 0 Y 116 Brick gfs-3.xx.xx.xx:/raw 49153 0 Y 132 NFS Server on localhost N/A N/A N N/A Self-heal Daemon on localhost N/A N/A Y 110 NFS Server on gfs-3.xx.xx.cc N/A N/A N N/A Self-heal Daemon on gfs-3.xx.xx.cc N/A N/A Y 122 Task Status of Volume datastore ------------------------------------------------------------------------------ There are no active volume tasks I don't know how to repair this, I tried with gluster volume heal datastore on gfs-3 but It gives a great out of files, what must I have to do? Go on with heal on gfs-3? Thank you in advance. Kind regards. -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150630/d18bc6c6/attachment.html>