similar to: [ovirt-users] Re: Gluster problems, cluster performance issues

Displaying 20 results from an estimated 400 matches similar to: "[ovirt-users] Re: Gluster problems, cluster performance issues"

2018 Jun 01
0
[ovirt-users] Re: Gluster problems, cluster performance issues
On Thu, May 31, 2018 at 3:16 AM, Jim Kusznir <jim at palousetech.com> wrote: > I've been back at it, and still am unable to get more than one of my > physical nodes to come online in ovirt, nor am I able to get more than the > two gluster volumes (storage domains) to show online within ovirt. > > In Storage -> Volumes, they all show offline (many with one brick down,
2018 May 30
1
[ovirt-users] Re: Gluster problems, cluster performance issues
I've been back at it, and still am unable to get more than one of my physical nodes to come online in ovirt, nor am I able to get more than the two gluster volumes (storage domains) to show online within ovirt. In Storage -> Volumes, they all show offline (many with one brick down, which is correct: I have one server off) However, in Storage -> domains, they all show down (although
2018 May 30
1
[ovirt-users] Re: Gluster problems, cluster performance issues
Adding Ravi to look into the heal issue. As for the fsync hang and subsequent IO errors, it seems a lot like https://bugzilla.redhat.com/show_bug.cgi?id=1497156 and Paolo Bonzini from qemu had pointed out that this would be fixed by the following commit: commit e72c9a2a67a6400c8ef3d01d4c461dbbbfa0e1f0 Author: Paolo Bonzini <pbonzini at redhat.com> Date: Wed Jun 21 16:35:46 2017
2018 May 30
0
[ovirt-users] Re: Gluster problems, cluster performance issues
Hi all again: I'm now subscribed to gluster-users as well, so I should get any replies from that side too. At this point, I am seeing acceptable (although slower than I expect) performance much of the time, with periodic massive spikes in latency (occasionally so bad as to cause ovirt to detect a engine bad health status). Often, if I check the logs just then, I'll see those call traces
2018 May 30
2
[ovirt-users] Re: Gluster problems, cluster performance issues
The profile seems to suggest very high latencies on the brick at ovirt1.nwfiber.com:/gluster/brick1/engine ovirt2.* shows decent numbers. Is everything OK with the brick on ovirt1? Are the bricks of engine volume on both these servers identical in terms of their config? -Krutika On Wed, May 30, 2018 at 3:07 PM, Jim Kusznir <jim at palousetech.com> wrote: > Hi: > > Thank you. I
2018 May 29
0
[ovirt-users] Re: Gluster problems, cluster performance issues
I would check disks status and accessibility of mount points where your gluster volumes reside. On Tue, May 29, 2018, 22:28 Jim Kusznir <jim at palousetech.com> wrote: > On one ovirt server, I'm now seeing these messages: > [56474.239725] blk_update_request: 63 callbacks suppressed > [56474.239732] blk_update_request: I/O error, dev dm-2, sector 0 > [56474.240602]
2018 May 29
0
[ovirt-users] Gluster problems, cluster performance issues
[Adding gluster-users to look at the heal issue] On Tue, May 29, 2018 at 9:17 AM, Jim Kusznir <jim at palousetech.com> wrote: > Hello: > > I've been having some cluster and gluster performance issues lately. I > also found that my cluster was out of date, and was trying to apply updates > (hoping to fix some of these), and discovered the ovirt 4.1 repos were > taken
2018 May 29
0
[ovirt-users] Gluster problems, cluster performance issues
Do you see errors reported in the mount logs for the volume? If so, could you attach the logs? Any issues with your underlying disks. Can you also attach output of volume profiling? On Wed, May 30, 2018 at 12:13 AM, Jim Kusznir <jim at palousetech.com> wrote: > Ok, things have gotten MUCH worse this morning. I'm getting random errors > from VMs, right now, about a third of my
2017 Aug 16
1
[ovirt-users] Recovering from a multi-node failure
On Sun, Aug 6, 2017 at 4:42 AM, Jim Kusznir <jim at palousetech.com> wrote: > Well, after a very stressful weekend, I think I have things largely > working. Turns out that most of the above issues were caused by the linux > permissions of the exports for all three volumes (they had been reset to > 600; setting them to 774 or 770 fixed many of the issues). Of course, I >
2017 Jun 20
0
[ovirt-users] Very poor GlusterFS performance
Have you tried with: performance.strict-o-direct : off performance.strict-write-ordering : off They can be changed dynamically. On 20 June 2017 at 17:21, Sahina Bose <sabose at redhat.com> wrote: > [Adding gluster-users] > > On Mon, Jun 19, 2017 at 8:16 PM, Chris Boot <bootc at bootc.net> wrote: > >> Hi folks, >> >> I have 3x servers in a
2017 Jun 20
2
[ovirt-users] Very poor GlusterFS performance
[Adding gluster-users] On Mon, Jun 19, 2017 at 8:16 PM, Chris Boot <bootc at bootc.net> wrote: > Hi folks, > > I have 3x servers in a "hyper-converged" oVirt 4.1.2 + GlusterFS 3.10 > configuration. My VMs run off a replica 3 arbiter 1 volume comprised of > 6 bricks, which themselves live on two SSDs in each of the servers (one > brick per SSD). The bricks are
2017 Jun 20
0
[ovirt-users] Very poor GlusterFS performance
Dear Krutika, Sorry for asking so naively but can you tell me on what factor do you base that the client and server event-threads parameters for a volume should be set to 4? Is this metric for example based on the number of cores a GlusterFS server has? I am asking because I saw my GlusterFS volumes are set to 2 and would like to set these parameters to something meaningful for performance
2017 Jun 20
5
[ovirt-users] Very poor GlusterFS performance
Couple of things: 1. Like Darrell suggested, you should enable stat-prefetch and increase client and server event threads to 4. # gluster volume set <VOL> performance.stat-prefetch on # gluster volume set <VOL> client.event-threads 4 # gluster volume set <VOL> server.event-threads 4 2. Also glusterfs-3.10.1 and above has a shard performance bug fix -
2024 Jun 26
0
Trouble with data size in a Distributed-Replicated 2 x 3 = 6 volume
HI, In my Distriuted-Replicated volume [ 2 x 3 = 6] used as a data domain for oVirt hypervisor, I had a trouble with used space : It was a previous Replicated 1 x 3 = 3 volume extended by adding one brick (one disk) in each server to recharge a 2 x 3 = 6 volume. Df -h on the mount point give ? 98% used volume (see below) Filesystem Size Used Avail Use% Mounted on
2013 Feb 12
1
Replication Ok, or not?
Setup a DC using 4.0.3 - all appears to go fine... Setup a second DC and everything works fine to here...but I'm not sure if replication is actually working or not. Here's what I get from ./samba-tool drs showrepl I've also done. [./samba-tool drs kcc -Uadministrator dc2.samba.somedom.local] in an attempt to fix the replication problem. (or what I think is a problem.) [The outbound
2018 Jun 22
1
sievescript is not working
Benny Pedersen <me at mlfix.junc.eu> schrieb am 22.06.18 um 03:32:26 Uhr: > ratatouille skrev den 2018-06-21 20:25: > > > # sieve-test ./managesieve.sieve ./testfile -D > > sieve-test(andreas): Debug: sieve: Pigeonhole version 0.4.2 > > initializing > > sieve-test(andreas): Debug: sieve: include: sieve_global_dir is not > > set; it is currently not
2018 May 22
1
[SOLVED] [Nfs-ganesha-support] volume start: gv01: failed: Quorum not met. Volume operation not allowed.
Hey All, Appears I solved this one and NFS mounts now work on all my clients. No issues since fixing it a few hours back. RESOLUTION Auditd is to blame for the trouble. Noticed this in the logs on 2 of the 3 NFS servers (nfs01, nfs02, nfs03): type=AVC msg=audit(1526965320.850:4094): avc: denied { write } for pid=8714 comm="ganesha.nfsd" name="nfs_0"
2009 Apr 20
2
BUG: soft lockup - CPU#1 stuck for 61s
?i, I have a cluster with 5 nodes hosting web application. All web servers save log info into shared access.log file. There is awstats log analyzer on the first node. Sometimes this node fails with the following messages (captured on another server) Apr 20 17:31:16 um-be-2 [145813.022112] o2net: connection to node um-fe-1 (num 1) at 192.168.10.10:7777 has been idle for 30.0 seconds, shutting it
2002 Dec 10
0
Description of Shares in Windows
On my Windows XP and 200 boxes the shares that I have mapped in a login script have their descriptions set to the following: downloads on 'Samba Sever 2.2.6pre2' (M:) This would be fine if I had set server string = Samba Server %v in the smb.conf file. However I have this set to server string = Samba. When I browse the network on the Windows boxes the correct description appears for
2004 Nov 24
1
Problems with udev on FC3
I've been testing * on FC3. I have everything compiled and installed. However, when I do 'modprobe wcfxo' (I have an X100P clone), I get the following in /var/log/messages: Nov 24 10:23:40 jfd wait_for_sysfs[3366]: either wait_for_sysfs (udev 039) needs an update to handle the device '/class/zaptel/zaptimer' properly (no device symlink) or the sysfs-support of your