Displaying 20 results from an estimated 10000 matches similar to: "Disable write-back in tiering volume"
2018 Feb 01
0
Tiered volume performance degrades badly after a volume stop/start or system restart.
This problem appears to be related to the sqlite3 DB files
that are used for the tiering file access counters, stored on
each hot and cold tier brick in .glusterfs/<volname>.db.
When the tier is first created, these DB files do not exist,
they are created, and everything works fine.
On a stop/start or service restart, the .db files are already
present, albeit empty since I don't have
2018 Jan 31
1
Tiered volume performance degrades badly after a volume stop/start or system restart.
Tested it in two different environments lately with exactly same results.
Was trying to get better read performance from local mounts with
hundreds of thousands maildir email files by using SSD,
hoping that .gluster file stat read will improve which does migrate
to hot tire.
After seeing what you described for 24 hours and confirming all move
around on the tires is done - killed it.
Here are my
2018 Jan 30
2
Tiered volume performance degrades badly after a volume stop/start or system restart.
I am fighting this issue:
Bug 1540376 ? Tiered volume performance degrades badly after a
volume stop/start or system restart.
https://bugzilla.redhat.com/show_bug.cgi?id=1540376
Does anyone have any ideas on what might be causing this, and
what a fix or work-around might be?
Thanks!
~ Jeff Byers ~
Tiered volume performance degrades badly after a volume
stop/start or system restart.
The
2017 Oct 10
0
ZFS with SSD ZIL vs XFS
Last time I've read about tiering in gluster, there wasn't any performance
gain with VM workload and more over doesn't speed up writes...
Il 10 ott 2017 9:27 PM, "Bartosz Zi?ba" <kontakt at avatat.pl> ha scritto:
> Hi,
>
> Have you thought about using an SSD as a GlusterFS hot tiers?
>
> Regards,
> Bartosz
>
>
> On 10.10.2017 19:59, Gandalf
2018 Mar 05
0
tiering
Hi,
There isn't a way to replace the failing tier brick through a single
command as we don't have support for replace/ remove or add brick with
tier.
Once you bring the brick online(volume start force), the data in the
brick will be built by the self heal daemon (Done because its a
replicated tier).
But adding brick will still not work.
Else if you use the force option, it will work as
2018 Feb 27
1
On sharded tiered volume, only first shard of new file goes on hot tier.
Does anyone have any ideas about how to fix, or to work-around the
following issue?
Thanks!
Bug 1549714 - On sharded tiered volume, only first shard of new file
goes on hot tier.
https://bugzilla.redhat.com/show_bug.cgi?id=1549714
On sharded tiered volume, only first shard of new file goes on hot tier.
On a sharded tiered volume, only the first shard of a new file
goes on the hot tier, the rest
2017 Dec 18
0
Testing sharding on tiered volume
----- Original Message -----
> From: "Viktor Nosov" <vnosov at stonefly.com>
> To: gluster-users at gluster.org
> Cc: vnosov at stonefly.com
> Sent: Friday, December 8, 2017 5:45:25 PM
> Subject: [Gluster-users] Testing sharding on tiered volume
>
> Hi,
>
> I'm looking to use sharding on tiered volume. This is very attractive
> feature that could
2018 Mar 04
1
tiering
Hi,
Have a glusterfs 3.10.10 (tried 3.12.6 as well) volume on Ubuntu 16.04 with a 3 ssd tier where one ssd is bad.
Status of volume: labgreenbin
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Hot Bricks:
Brick labgfs81:/gfs/p1-tier/mount 49156 0 Y 4217
Brick
2017 Nov 03
1
Ignore failed connection messages during copying files with tiering
Hi, All,
We create a GlusterFS cluster with tiers. The hot tier is
distributed-replicated SSDs. The cold tier is a n*(6+2) disperse volume.
When copy millions of files to the cluster, we find these logs:
W [socket.c:3292:socket_connect] 0-tierd: Ignore failed connection attempt
on /var/run/gluster/39668fb028de4b1bb6f4880e7450c064.socket, (No such file
or directory)
W
2017 Nov 04
1
Fwd: Ignore failed connection messages during copying files with tiering
Hi,
We create a GlusterFS cluster with tiers. The hot tier is
distributed-replicated SSDs. The cold tier is a n*(6+2) disperse volume.
When copy millions of files to the cluster, we find these logs:
W [socket.c:3292:socket_connect] 0-tierd: Ignore failed connection attempt
on /var/run/gluster/39668fb028de4b1bb6f4880e7450c064.socket, (No such file
or directory)
W [socket.c:3292:socket_connect]
2017 Oct 22
1
gluster tiering errors
There are several messages "no space left on device". I would check first
that free disk space is available for the volume.
On Oct 22, 2017 18:42, "Milind Changire" <mchangir at redhat.com> wrote:
> Herb,
> What are the high and low watermarks for the tier set at ?
>
> # gluster volume get <vol> cluster.watermark-hi
>
> # gluster volume get
2017 Oct 27
0
gluster tiering errors
Herb,
I'm trying to weed out issues here.
So, I can see quota turned *on* and would like you to check the quota
settings and test to see system behavior *if quota is turned off*.
Although the file size that failed migration was 29K, I'm being a bit
paranoid while weeding out issues.
Are you still facing tiering errors ?
I can see your response to Alex with the disk space consumption and
2017 Aug 03
0
Hot Tier
Hi,
We will look into the " failed to get index" error.
It shouldn't affect the normal working. Do let us know if you face any
other issues.
Regards,
Hari.
On 02-Aug-2017 11:55 PM, "Dmitri Chebotarov" <4dimach at gmail.com> wrote:
Hello
I reattached hot tier to a new empty EC volume and started to copy data to
the volume.
Good news is I can see files now on SSD
2017 Oct 22
0
gluster tiering errors
Herb,
What are the high and low watermarks for the tier set at ?
# gluster volume get <vol> cluster.watermark-hi
# gluster volume get <vol> cluster.watermark-low
What is the size of the file that failed to migrate as per the following
tierd log:
[2017-10-19 17:52:07.519614] I [MSGID: 109038]
[tier.c:1169:tier_migrate_using_query_file] 0-<vol>-tier-dht: Promotion
failed for
2017 Oct 18
0
warning spam in the logs after tiering experiment
forgot to mention Gluster version 3.10.6
On 18 October 2017 at 13:26, Alastair Neil <ajneil.tech at gmail.com> wrote:
> a short while ago I experimented with tiering on one of my volumes. I
> decided it was not working out so I removed the tier. I now have spam in
> the glusterd.log evert 7 seconds:
>
> [2017-10-18 17:17:29.578327] W [socket.c:3207:socket_connect] 0-tierd:
2017 Oct 24
2
gluster tiering errors
Milind - Thank you for the response..
>> What are the high and low watermarks for the tier set at ?
# gluster volume get <vol> cluster.watermark-hi
Option Value
------ -----
cluster.watermark-hi 90
# gluster volume get <vol> cluster.watermark-low
Option
2017 Oct 18
2
warning spam in the logs after tiering experiment
a short while ago I experimented with tiering on one of my volumes. I
decided it was not working out so I removed the tier. I now have spam in
the glusterd.log evert 7 seconds:
[2017-10-18 17:17:29.578327] W [socket.c:3207:socket_connect] 0-tierd:
Ignore failed connection attempt on
/var/run/gluster/2e3df1c501d0a19e5076304179d1e43e.socket, (No such file or
directory)
[2017-10-18
2017 Oct 19
3
gluster tiering errors
All,
I am new to gluster and have some questions/concerns about some tiering
errors that I see in the log files.
OS: CentOs 7.3.1611
Gluster version: 3.10.5
Samba version: 4.6.2
I see the following (scrubbed):
Node 1 /var/log/glusterfs/tier/<vol>/tierd.log:
[2017-10-19 17:52:07.519614] I [MSGID: 109038]
[tier.c:1169:tier_migrate_using_query_file]
0-<vol>-tier-dht: Promotion failed
2018 Mar 08
0
Write speed of gluster volume reduced
http://lists.gluster.org/pipermail/gluster-users/2017-July/031788.html
http://lists.gluster.org/pipermail/gluster-users/2017-September/032385.html
also try disperse.eager-lock off
On Tue, Mar 6, 2018 at 7:40 AM, Sherin George <allmyforums at outlook.in> wrote:
> Hi Guys,
>
> I have a gluster volume with the following configuration.
> ~~~~~~~~~~~~~~~~~~~~~
> Number of Bricks:
2024 Feb 21
1
gluster volume tier option missing
good evening,
i am new to glusterFS and read about tiering option, BUT cannot find it on
current v10 ubuntu 22 lts version, my fault?
thank you.
https://www.gluster.org/automated-tiering-in-gluster-2/
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20240221/2e1fc4ab/attachment.html>