Hi Aravinda,
Thanks for the reply. This test server is indeed the master server for
geo-replication to a slave.
I'm really surprised that geo-replication simply keeps writing logs until
all space is consumed, without cleaning them up itself. I didn't see any
warning about it in the geo-replication install documentation which is
unfortunate. We'll come up with a solution to delete log files older than
the LAST_SYNCED time in the geo-replication status. Is anyone aware of any
other potential gotchas like this?
Does anyone have an idea why in my previous note some space in the 2GB
GlusterFS partition apparently went missing? We had 0.47GB of data, 1GB
reported used by .glusterfs, which even if they were separate files would
only add up to 1.47GB used, meaning 0.53GB should have been left in the
partition. If less space is actually being used because of the hard links
then it's even harder to understand where the other 1.53GB went. So why
would GlusterFS report "No space left on device"?
Thanks again for any assistance.
On Thu, 5 Mar 2020 at 17:31, Aravinda VK <aravinda at kadalu.io> wrote:
> Hi David,
>
> Is this Volume is uses Geo-replication? Geo-replication feature enables
> Changelog to identify the latest changes happening in the GlusterFS volume.
>
> Content of .glusterfs directory also includes hardlinks to the actual
> data, so the size shown in .glusterfs is including data. Please refer the
> comment by Xavi
> https://github.com/gluster/glusterfs/issues/833#issuecomment-594436009
>
> If Changelogs files are causing issue, you can use archival tool to remove
> processed changelogs.
> https://github.com/aravindavk/archive_gluster_changelogs
>
> ?
> regards
> Aravinda Vishwanathapura
> https://kadalu.io
>
>
> On 05-Mar-2020, at 9:02 AM, David Cunningham <dcunningham at
voisonics.com>
> wrote:
>
> Hello,
>
> We are looking for some advice on disk use. This is on a single node
> GlusterFS test server.
>
> There's a 2GB partition for GlusterFS. Of that, 470MB is used for
actual
> data, and 1GB is used by the .glusterfs directory. The .glusterfs directory
> is mostly used by the two-character directories and the
"changelogs"
> directory. Why is so much used by .glusterfs, and can we reduce that
> overhead?
>
> We also have a problem with this test system where GlusterFS is giving
"No
> space left on device" errors. That's despite "df"
reporting only 54% used,
> and even if we add the 470MB to 1GB used above, that still comes out to
> less than the 2GB available, so there should be some spare.
>
> Would anyone be able to advise on these please? Thank you in advance.
>
> The GlusterFS version is 5.11 and here is the volume information:
>
> Volume Name: gvol0
> Type: Distribute
> Volume ID: 33ed309b-0e63-4f9a-8132-ab1b0fdcbc36
> Status: Started
> Snapshot Count: 0
> Number of Bricks: 1
> Transport-type: tcp
> Bricks:
> Brick1: myhost:/nodirectwritedata/gluster/gvol0
> Options Reconfigured:
> transport.address-family: inet
> nfs.disable: on
> geo-replication.indexing: on
> geo-replication.ignore-pid-check: on
> changelog.changelog: on
>
> --
> David Cunningham, Voisonics Limited
> http://voisonics.com/
> USA: +1 213 221 1092
> New Zealand: +64 (0)28 2558 3782
> ________
>
>
>
> Community Meeting Calendar:
>
> Schedule -
> Every Tuesday at 14:30 IST / 09:00 UTC
> Bridge: https://bluejeans.com/441850968
>
> Gluster-users mailing list
> Gluster-users at gluster.org
> https://lists.gluster.org/mailman/listinfo/gluster-users
>
>
>
>
>
>
>
--
David Cunningham, Voisonics Limited
http://voisonics.com/
USA: +1 213 221 1092
New Zealand: +64 (0)28 2558 3782
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
<http://lists.gluster.org/pipermail/gluster-users/attachments/20200305/08fbc536/attachment.html>