Krutika Dhananjay
2018-May-30  04:17 UTC
[Gluster-users] [ovirt-users] Re: Gluster problems, cluster performance issues
Adding Ravi to look into the heal issue.
As for the fsync hang and subsequent IO errors, it seems a lot like
https://bugzilla.redhat.com/show_bug.cgi?id=1497156 and Paolo Bonzini from
qemu had pointed out that this would be fixed by the following commit:
  commit e72c9a2a67a6400c8ef3d01d4c461dbbbfa0e1f0
    Author: Paolo Bonzini <pbonzini at redhat.com>
    Date:   Wed Jun 21 16:35:46 2017 +0200
    scsi: virtio_scsi: let host do exception handling
    virtio_scsi tries to do exception handling after the default 30 seconds
    timeout expires.  However, it's better to let the host control the
    timeout, otherwise with a heavy I/O load it is likely that an abort will
    also timeout.  This leads to fatal errors like filesystems going
    offline.
    Disable the 'sd' timeout and allow the host to do exception
handling,
    following the precedent of the storvsc driver.
    Hannes has a proposal to introduce timeouts in virtio, but this provides
    an immediate solution for stable kernels too.
    [mkp: fixed typo]
    Reported-by: Douglas Miller <dougmill at linux.vnet.ibm.com>
    Cc: "James E.J. Bottomley" <jejb at linux.vnet.ibm.com>
    Cc: "Martin K. Petersen" <martin.petersen at oracle.com>
    Cc: Hannes Reinecke <hare at suse.de>
    Cc: linux-scsi at vger.kernel.org
    Cc: stable at vger.kernel.org
    Signed-off-by: Paolo Bonzini <pbonzini at redhat.com>
    Signed-off-by: Martin K. Petersen <martin.petersen at oracle.com>
Adding Paolo/Kevin to comment.
As for the poor gluster performance, could you disable cluster.eager-lock
and see if that makes any difference:
# gluster volume set <VOL> cluster.eager-lock off
Do also capture the volume profile again if you still see performance
issues after disabling eager-lock.
-Krutika
On Wed, May 30, 2018 at 6:55 AM, Jim Kusznir <jim at palousetech.com>
wrote:
> I also finally found the following in my system log on one server:
>
> [10679.524491] INFO: task glusterclogro:14933 blocked for more than 120
> seconds.
> [10679.525826] "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs"
> disables this message.
> [10679.527144] glusterclogro   D ffff97209832bf40     0 14933      1
> 0x00000080
> [10679.527150] Call Trace:
> [10679.527161]  [<ffffffffb9913f79>] schedule+0x29/0x70
> [10679.527218]  [<ffffffffc060e388>] _xfs_log_force_lsn+0x2e8/0x340
[xfs]
> [10679.527225]  [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20
> [10679.527254]  [<ffffffffc05eeb97>] xfs_file_fsync+0x107/0x1e0 [xfs]
> [10679.527260]  [<ffffffffb944f0e7>] do_fsync+0x67/0xb0
> [10679.527268]  [<ffffffffb992076f>] ?
system_call_after_swapgs+0xbc/0x160
> [10679.527271]  [<ffffffffb944f3d0>] SyS_fsync+0x10/0x20
> [10679.527275]  [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21
> [10679.527279]  [<ffffffffb992077b>] ?
system_call_after_swapgs+0xc8/0x160
> [10679.527283] INFO: task glusterposixfsy:14941 blocked for more than 120
> seconds.
> [10679.528608] "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs"
> disables this message.
> [10679.529956] glusterposixfsy D ffff972495f84f10     0 14941      1
> 0x00000080
> [10679.529961] Call Trace:
> [10679.529966]  [<ffffffffb9913f79>] schedule+0x29/0x70
> [10679.530003]  [<ffffffffc060e388>] _xfs_log_force_lsn+0x2e8/0x340
[xfs]
> [10679.530008]  [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20
> [10679.530038]  [<ffffffffc05eeb97>] xfs_file_fsync+0x107/0x1e0 [xfs]
> [10679.530042]  [<ffffffffb944f0e7>] do_fsync+0x67/0xb0
> [10679.530046]  [<ffffffffb992076f>] ?
system_call_after_swapgs+0xbc/0x160
> [10679.530050]  [<ffffffffb944f3f3>] SyS_fdatasync+0x13/0x20
> [10679.530054]  [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21
> [10679.530058]  [<ffffffffb992077b>] ?
system_call_after_swapgs+0xc8/0x160
> [10679.530062] INFO: task glusteriotwr13:15486 blocked for more than 120
> seconds.
> [10679.531805] "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs"
> disables this message.
> [10679.533732] glusteriotwr13  D ffff9720a83f0000     0 15486      1
> 0x00000080
> [10679.533738] Call Trace:
> [10679.533747]  [<ffffffffb9913f79>] schedule+0x29/0x70
> [10679.533799]  [<ffffffffc060e388>] _xfs_log_force_lsn+0x2e8/0x340
[xfs]
> [10679.533806]  [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20
> [10679.533846]  [<ffffffffc05eeb97>] xfs_file_fsync+0x107/0x1e0 [xfs]
> [10679.533852]  [<ffffffffb944f0e7>] do_fsync+0x67/0xb0
> [10679.533858]  [<ffffffffb992076f>] ?
system_call_after_swapgs+0xbc/0x160
> [10679.533863]  [<ffffffffb944f3f3>] SyS_fdatasync+0x13/0x20
> [10679.533868]  [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21
> [10679.533873]  [<ffffffffb992077b>] ?
system_call_after_swapgs+0xc8/0x160
> [10919.512757] INFO: task glusterclogro:14933 blocked for more than 120
> seconds.
> [10919.514714] "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs"
> disables this message.
> [10919.516663] glusterclogro   D ffff97209832bf40     0 14933      1
> 0x00000080
> [10919.516677] Call Trace:
> [10919.516690]  [<ffffffffb9913f79>] schedule+0x29/0x70
> [10919.516696]  [<ffffffffb99118e9>] schedule_timeout+0x239/0x2c0
> [10919.516703]  [<ffffffffb951cc04>] ? blk_finish_plug+0x14/0x40
> [10919.516768]  [<ffffffffc05e9224>] ? _xfs_buf_ioapply+0x334/0x460
[xfs]
> [10919.516774]  [<ffffffffb991432d>] wait_for_completion+0xfd/0x140
> [10919.516782]  [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20
> [10919.516821]  [<ffffffffc05eb0a3>] ? _xfs_buf_read+0x23/0x40 [xfs]
> [10919.516859]  [<ffffffffc05eafa9>] xfs_buf_submit_wait+0xf9/0x1d0
[xfs]
> [10919.516902]  [<ffffffffc061b279>] ?
xfs_trans_read_buf_map+0x199/0x400
> [xfs]
> [10919.516940]  [<ffffffffc05eb0a3>] _xfs_buf_read+0x23/0x40 [xfs]
> [10919.516977]  [<ffffffffc05eb1b9>] xfs_buf_read_map+0xf9/0x160
[xfs]
> [10919.517022]  [<ffffffffc061b279>]
xfs_trans_read_buf_map+0x199/0x400
> [xfs]
> [10919.517057]  [<ffffffffc05c8d04>] xfs_da_read_buf+0xd4/0x100 [xfs]
> [10919.517091]  [<ffffffffc05c8d53>] xfs_da3_node_read+0x23/0xd0
[xfs]
> [10919.517126]  [<ffffffffc05c9fee>]
xfs_da3_node_lookup_int+0x6e/0x2f0
> [xfs]
> [10919.517160]  [<ffffffffc05d5a1d>] xfs_dir2_node_lookup+0x4d/0x170
[xfs]
> [10919.517194]  [<ffffffffc05ccf5d>] xfs_dir_lookup+0x1bd/0x1e0 [xfs]
> [10919.517233]  [<ffffffffc05fd8d9>] xfs_lookup+0x69/0x140 [xfs]
> [10919.517271]  [<ffffffffc05fa018>] xfs_vn_lookup+0x78/0xc0 [xfs]
> [10919.517278]  [<ffffffffb9425cf3>] lookup_real+0x23/0x60
> [10919.517283]  [<ffffffffb9426702>] __lookup_hash+0x42/0x60
> [10919.517288]  [<ffffffffb942d519>] SYSC_renameat2+0x3a9/0x5a0
> [10919.517296]  [<ffffffffb94d3753>] ? selinux_file_free_security+
> 0x23/0x30
> [10919.517304]  [<ffffffffb992077b>] ?
system_call_after_swapgs+0xc8/0x160
> [10919.517309]  [<ffffffffb992076f>] ?
system_call_after_swapgs+0xbc/0x160
> [10919.517313]  [<ffffffffb992077b>] ?
system_call_after_swapgs+0xc8/0x160
> [10919.517318]  [<ffffffffb992076f>] ?
system_call_after_swapgs+0xbc/0x160
> [10919.517323]  [<ffffffffb942e58e>] SyS_renameat2+0xe/0x10
> [10919.517328]  [<ffffffffb942e5ce>] SyS_rename+0x1e/0x20
> [10919.517333]  [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21
> [10919.517339]  [<ffffffffb992077b>] ?
system_call_after_swapgs+0xc8/0x160
> [11159.496095] INFO: task glusteriotwr9:15482 blocked for more than 120
> seconds.
> [11159.497546] "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs"
> disables this message.
> [11159.498978] glusteriotwr9   D ffff971fa0fa1fa0     0 15482      1
> 0x00000080
> [11159.498984] Call Trace:
> [11159.498995]  [<ffffffffb9911f00>] ? bit_wait+0x50/0x50
> [11159.498999]  [<ffffffffb9913f79>] schedule+0x29/0x70
> [11159.499003]  [<ffffffffb99118e9>] schedule_timeout+0x239/0x2c0
> [11159.499056]  [<ffffffffc05dd9b7>] ? xfs_iext_bno_to_ext+0xa7/0x1a0
[xfs]
> [11159.499082]  [<ffffffffc05dd43e>] ? xfs_iext_bno_to_irec+0x8e/0xd0
[xfs]
> [11159.499090]  [<ffffffffb92f7a12>] ? ktime_get_ts64+0x52/0xf0
> [11159.499093]  [<ffffffffb9911f00>] ? bit_wait+0x50/0x50
> [11159.499097]  [<ffffffffb991348d>] io_schedule_timeout+0xad/0x130
> [11159.499101]  [<ffffffffb9913528>] io_schedule+0x18/0x20
> [11159.499104]  [<ffffffffb9911f11>] bit_wait_io+0x11/0x50
> [11159.499107]  [<ffffffffb9911ac1>] __wait_on_bit_lock+0x61/0xc0
> [11159.499113]  [<ffffffffb9393634>] __lock_page+0x74/0x90
> [11159.499118]  [<ffffffffb92bc210>] ? wake_bit_function+0x40/0x40
> [11159.499121]  [<ffffffffb9394154>] __find_lock_page+0x54/0x70
> [11159.499125]  [<ffffffffb9394e85>]
grab_cache_page_write_begin+0x55/0xc0
> [11159.499130]  [<ffffffffb9484b76>] iomap_write_begin+0x66/0x100
> [11159.499135]  [<ffffffffb9484edf>] iomap_write_actor+0xcf/0x1d0
> [11159.499140]  [<ffffffffb9484e10>] ? iomap_write_end+0x80/0x80
> [11159.499144]  [<ffffffffb94854e7>] iomap_apply+0xb7/0x150
> [11159.499149]  [<ffffffffb9485621>]
iomap_file_buffered_write+0xa1/0xe0
> [11159.499153]  [<ffffffffb9484e10>] ? iomap_write_end+0x80/0x80
> [11159.499182]  [<ffffffffc05f025d>]
xfs_file_buffered_aio_write+0x12d/0x2c0
> [xfs]
> [11159.499213]  [<ffffffffc05f057d>] xfs_file_aio_write+0x18d/0x1b0
[xfs]
> [11159.499217]  [<ffffffffb941a533>] do_sync_write+0x93/0xe0
> [11159.499222]  [<ffffffffb941b010>] vfs_write+0xc0/0x1f0
> [11159.499225]  [<ffffffffb941c002>] SyS_pwrite64+0x92/0xc0
> [11159.499230]  [<ffffffffb992076f>] ?
system_call_after_swapgs+0xbc/0x160
> [11159.499234]  [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21
> [11159.499238]  [<ffffffffb992077b>] ?
system_call_after_swapgs+0xc8/0x160
> [11279.488720] INFO: task xfsaild/dm-10:1134 blocked for more than 120
> seconds.
> [11279.490197] "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs"
> disables this message.
> [11279.491665] xfsaild/dm-10   D ffff9720a8660fd0     0  1134      2
> 0x00000000
> [11279.491671] Call Trace:
> [11279.491682]  [<ffffffffb92a3a2e>] ?
try_to_del_timer_sync+0x5e/0x90
> [11279.491688]  [<ffffffffb9913f79>] schedule+0x29/0x70
> [11279.491744]  [<ffffffffc060de36>] _xfs_log_force+0x1c6/0x2c0 [xfs]
> [11279.491750]  [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20
> [11279.491783]  [<ffffffffc0619fec>] ? xfsaild+0x16c/0x6f0 [xfs]
> [11279.491817]  [<ffffffffc060df5c>] xfs_log_force+0x2c/0x70 [xfs]
> [11279.491849]  [<ffffffffc0619e80>] ?
xfs_trans_ail_cursor_first+0x90/0x90
> [xfs]
> [11279.491880]  [<ffffffffc0619fec>] xfsaild+0x16c/0x6f0 [xfs]
> [11279.491913]  [<ffffffffc0619e80>] ?
xfs_trans_ail_cursor_first+0x90/0x90
> [xfs]
> [11279.491919]  [<ffffffffb92bb161>] kthread+0xd1/0xe0
> [11279.491926]  [<ffffffffb92bb090>] ? insert_kthread_work+0x40/0x40
> [11279.491932]  [<ffffffffb9920677>]
ret_from_fork_nospec_begin+0x21/0x21
> [11279.491936]  [<ffffffffb92bb090>] ? insert_kthread_work+0x40/0x40
> [11279.491976] INFO: task glusterclogfsyn:14934 blocked for more than 120
> seconds.
> [11279.493466] "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs"
> disables this message.
> [11279.494952] glusterclogfsyn D ffff97209832af70     0 14934      1
> 0x00000080
> [11279.494957] Call Trace:
> [11279.494979]  [<ffffffffc0309839>] ?
__split_and_process_bio+0x2e9/0x520
> [dm_mod]
> [11279.494983]  [<ffffffffb9913f79>] schedule+0x29/0x70
> [11279.494987]  [<ffffffffb99118e9>] schedule_timeout+0x239/0x2c0
> [11279.494997]  [<ffffffffc0309d98>] ? dm_make_request+0x128/0x1a0
[dm_mod]
> [11279.495001]  [<ffffffffb991348d>] io_schedule_timeout+0xad/0x130
> [11279.495005]  [<ffffffffb99145ad>]
wait_for_completion_io+0xfd/0x140
> [11279.495010]  [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20
> [11279.495016]  [<ffffffffb951e574>] blkdev_issue_flush+0xb4/0x110
> [11279.495049]  [<ffffffffc06064b9>] xfs_blkdev_issue_flush+0x19/0x20
> [xfs]
> [11279.495079]  [<ffffffffc05eec40>] xfs_file_fsync+0x1b0/0x1e0 [xfs]
> [11279.495086]  [<ffffffffb944f0e7>] do_fsync+0x67/0xb0
> [11279.495090]  [<ffffffffb992076f>] ?
system_call_after_swapgs+0xbc/0x160
> [11279.495094]  [<ffffffffb944f3d0>] SyS_fsync+0x10/0x20
> [11279.495098]  [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21
> [11279.495102]  [<ffffffffb992077b>] ?
system_call_after_swapgs+0xc8/0x160
> [11279.495105] INFO: task glusterposixfsy:14941 blocked for more than 120
> seconds.
> [11279.496606] "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs"
> disables this message.
> [11279.498114] glusterposixfsy D ffff972495f84f10     0 14941      1
> 0x00000080
> [11279.498118] Call Trace:
> [11279.498134]  [<ffffffffc0309839>] ?
__split_and_process_bio+0x2e9/0x520
> [dm_mod]
> [11279.498138]  [<ffffffffb9913f79>] schedule+0x29/0x70
> [11279.498142]  [<ffffffffb99118e9>] schedule_timeout+0x239/0x2c0
> [11279.498152]  [<ffffffffc0309d98>] ? dm_make_request+0x128/0x1a0
[dm_mod]
> [11279.498156]  [<ffffffffb991348d>] io_schedule_timeout+0xad/0x130
> [11279.498160]  [<ffffffffb99145ad>]
wait_for_completion_io+0xfd/0x140
> [11279.498165]  [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20
> [11279.498169]  [<ffffffffb951e574>] blkdev_issue_flush+0xb4/0x110
> [11279.498202]  [<ffffffffc06064b9>] xfs_blkdev_issue_flush+0x19/0x20
> [xfs]
> [11279.498231]  [<ffffffffc05eec40>] xfs_file_fsync+0x1b0/0x1e0 [xfs]
> [11279.498238]  [<ffffffffb944f0e7>] do_fsync+0x67/0xb0
> [11279.498242]  [<ffffffffb992076f>] ?
system_call_after_swapgs+0xbc/0x160
> [11279.498246]  [<ffffffffb944f3f3>] SyS_fdatasync+0x13/0x20
> [11279.498250]  [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21
> [11279.498254]  [<ffffffffb992077b>] ?
system_call_after_swapgs+0xc8/0x160
> [11279.498257] INFO: task glusteriotwr1:14950 blocked for more than 120
> seconds.
> [11279.499789] "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs"
> disables this message.
> [11279.501343] glusteriotwr1   D ffff97208b6daf70     0 14950      1
> 0x00000080
> [11279.501348] Call Trace:
> [11279.501353]  [<ffffffffb9913f79>] schedule+0x29/0x70
> [11279.501390]  [<ffffffffc060e388>] _xfs_log_force_lsn+0x2e8/0x340
[xfs]
> [11279.501396]  [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20
> [11279.501428]  [<ffffffffc05eeb97>] xfs_file_fsync+0x107/0x1e0 [xfs]
> [11279.501432]  [<ffffffffb944ef3f>] generic_write_sync+0x4f/0x70
> [11279.501461]  [<ffffffffc05f0545>] xfs_file_aio_write+0x155/0x1b0
[xfs]
> [11279.501466]  [<ffffffffb941a533>] do_sync_write+0x93/0xe0
> [11279.501471]  [<ffffffffb941b010>] vfs_write+0xc0/0x1f0
> [11279.501475]  [<ffffffffb941c002>] SyS_pwrite64+0x92/0xc0
> [11279.501479]  [<ffffffffb992076f>] ?
system_call_after_swapgs+0xbc/0x160
> [11279.501483]  [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21
> [11279.501489]  [<ffffffffb992077b>] ?
system_call_after_swapgs+0xc8/0x160
> [11279.501493] INFO: task glusteriotwr4:14953 blocked for more than 120
> seconds.
> [11279.503047] "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs"
> disables this message.
> [11279.504630] glusteriotwr4   D ffff972499f2bf40     0 14953      1
> 0x00000080
> [11279.504635] Call Trace:
> [11279.504640]  [<ffffffffb9913f79>] schedule+0x29/0x70
> [11279.504676]  [<ffffffffc060e388>] _xfs_log_force_lsn+0x2e8/0x340
[xfs]
> [11279.504681]  [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20
> [11279.504710]  [<ffffffffc05eeb97>] xfs_file_fsync+0x107/0x1e0 [xfs]
> [11279.504714]  [<ffffffffb944f0e7>] do_fsync+0x67/0xb0
> [11279.504718]  [<ffffffffb992076f>] ?
system_call_after_swapgs+0xbc/0x160
> [11279.504722]  [<ffffffffb944f3d0>] SyS_fsync+0x10/0x20
> [11279.504725]  [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21
> [11279.504730]  [<ffffffffb992077b>] ?
system_call_after_swapgs+0xc8/0x160
> [12127.466494] perf: interrupt took too long (8263 > 8150), lowering
> kernel.perf_event_max_sample_rate to 24000
>
> --------------------
> I think this is the cause of the massive ovirt performance issues
> irrespective of gluster volume.  At the time this happened, I was also
> ssh'ed into the host, and was doing some rpm querry commands.  I had
just
> run rpm -qa |grep glusterfs (to verify what version was actually
> installed), and that command took almost 2 minutes to return!  Normally it
> takes less than 2 seconds.  That is all pure local SSD IO, too....
>
> I'm no expert, but its my understanding that anytime a software causes
> these kinds of issues, its a serious bug in the software, even if its
> mis-handled exceptions.  Is this correct?
>
> --Jim
>
> On Tue, May 29, 2018 at 3:01 PM, Jim Kusznir <jim at palousetech.com>
wrote:
>
>> I think this is the profile information for one of the volumes that
lives
>> on the SSDs and is fully operational with no down/problem disks:
>>
>> [root at ovirt2 yum.repos.d]# gluster volume profile data info
>> Brick: ovirt2.nwfiber.com:/gluster/brick2/data
>> ----------------------------------------------
>> Cumulative Stats:
>>    Block Size:                256b+                 512b+
>> 1024b+
>>  No. of Reads:                  983                  2696
>>   1059
>> No. of Writes:                    0                  1113
>>    302
>>
>>    Block Size:               2048b+                4096b+
>> 8192b+
>>  No. of Reads:                  852                 88608
>>  53526
>> No. of Writes:                  522                812340
>>  76257
>>
>>    Block Size:              16384b+               32768b+
>>  65536b+
>>  No. of Reads:                54351                241901
>>  15024
>> No. of Writes:                21636                  8656
>>   8976
>>
>>    Block Size:             131072b+
>>  No. of Reads:               524156
>> No. of Writes:               296071
>>  %-latency   Avg-latency   Min-Latency   Max-Latency   No. of calls
>>    Fop
>>  ---------   -----------   -----------   -----------   ------------
>>   ----
>>       0.00       0.00 us       0.00 us       0.00 us           4189
>>  RELEASE
>>       0.00       0.00 us       0.00 us       0.00 us           1257
>> RELEASEDIR
>>       0.00      46.19 us      12.00 us     187.00 us             69
>>  FLUSH
>>       0.00     147.00 us      78.00 us     367.00 us             86
>> REMOVEXATTR
>>       0.00     223.46 us      24.00 us    1166.00 us            149
>>  READDIR
>>       0.00     565.34 us      76.00 us    3639.00 us             88
>>  FTRUNCATE
>>       0.00     263.28 us      20.00 us   28385.00 us            228
>>     LK
>>       0.00      98.84 us       2.00 us     880.00 us           1198
>>  OPENDIR
>>       0.00      91.59 us      26.00 us   10371.00 us           3853
>> STATFS
>>       0.00     494.14 us      17.00 us  193439.00 us           1171
>> GETXATTR
>>       0.00     299.42 us      35.00 us    9799.00 us           2044
>> READDIRP
>>       0.00    1965.31 us     110.00 us  382258.00 us            321
>>  XATTROP
>>       0.01     113.40 us      24.00 us   61061.00 us           8134
>>   STAT
>>       0.01     755.38 us      57.00 us  607603.00 us           3196
>>  DISCARD
>>       0.05    2690.09 us      58.00 us 2704761.00 us           3206
>>   OPEN
>>       0.10  119978.25 us      97.00 us 9406684.00 us            154
>>  SETATTR
>>       0.18     101.73 us      28.00 us  700477.00 us         313379
>>  FSTAT
>>       0.23    1059.84 us      25.00 us 2716124.00 us          38255
>> LOOKUP
>>       0.47    1024.11 us      54.00 us 6197164.00 us          81455
>> FXATTROP
>>       1.72    2984.00 us      15.00 us 37098954.00 us         103020
>> FINODELK
>>       5.92   44315.32 us      51.00 us 24731536.00 us          23957
>>  FSYNC
>>      13.27    2399.78 us      25.00 us 22089540.00 us         991005
>>   READ
>>      37.00    5980.43 us      52.00 us 22099889.00 us        1108976
>>  WRITE
>>      41.04    5452.75 us      13.00 us 22102452.00 us        1349053
>>  INODELK
>>
>>     Duration: 10026 seconds
>>    Data Read: 80046027759 bytes
>> Data Written: 44496632320 bytes
>>
>> Interval 1 Stats:
>>    Block Size:                256b+                 512b+
>> 1024b+
>>  No. of Reads:                  983                  2696
>>   1059
>> No. of Writes:                    0                   838
>>    185
>>
>>    Block Size:               2048b+                4096b+
>> 8192b+
>>  No. of Reads:                  852                 85856
>>  51575
>> No. of Writes:                  382                705802
>>  57812
>>
>>    Block Size:              16384b+               32768b+
>>  65536b+
>>  No. of Reads:                52673                232093
>>  14984
>> No. of Writes:                13499                  4908
>>   4242
>>
>>    Block Size:             131072b+
>>  No. of Reads:               460040
>> No. of Writes:                 6411
>>  %-latency   Avg-latency   Min-Latency   Max-Latency   No. of calls
>>    Fop
>>  ---------   -----------   -----------   -----------   ------------
>>   ----
>>       0.00       0.00 us       0.00 us       0.00 us           2093
>>  RELEASE
>>       0.00       0.00 us       0.00 us       0.00 us           1093
>> RELEASEDIR
>>       0.00      53.38 us      26.00 us     111.00 us             16
>>  FLUSH
>>       0.00     145.14 us      78.00 us     367.00 us             71
>> REMOVEXATTR
>>       0.00     190.96 us     114.00 us     298.00 us             71
>>  SETATTR
>>       0.00     213.38 us      24.00 us    1145.00 us             90
>>  READDIR
>>       0.00     263.28 us      20.00 us   28385.00 us            228
>>     LK
>>       0.00     101.76 us       2.00 us     880.00 us           1093
>>  OPENDIR
>>       0.01      93.60 us      27.00 us   10371.00 us           3090
>> STATFS
>>       0.02     537.47 us      17.00 us  193439.00 us           1038
>> GETXATTR
>>       0.03     297.44 us      35.00 us    9799.00 us           1990
>> READDIRP
>>       0.03    2357.28 us     110.00 us  382258.00 us            253
>>  XATTROP
>>       0.04     385.93 us      58.00 us   47593.00 us           2091
>>   OPEN
>>       0.04     114.86 us      24.00 us   61061.00 us           7715
>>   STAT
>>       0.06     444.59 us      57.00 us  333240.00 us           3053
>>  DISCARD
>>       0.42     316.24 us      25.00 us  290728.00 us          29823
>> LOOKUP
>>       0.73     257.92 us      54.00 us  344812.00 us          63296
>> FXATTROP
>>       1.37      98.30 us      28.00 us   67621.00 us         313172
>>  FSTAT
>>       1.58    2124.69 us      51.00 us  849200.00 us          16717
>>  FSYNC
>>       5.73     162.46 us      52.00 us  748492.00 us         794079
>>  WRITE
>>       7.19    2065.17 us      16.00 us 37098954.00 us          78381
>> FINODELK
>>      36.44     886.32 us      25.00 us 2216436.00 us         925421
>>   READ
>>      46.30    1178.04 us      13.00 us 1700704.00 us         884635
>>  INODELK
>>
>>     Duration: 7485 seconds
>>    Data Read: 71250527215 bytes
>> Data Written: 5119903744 bytes
>>
>> Brick: ovirt3.nwfiber.com:/gluster/brick2/data
>> ----------------------------------------------
>> Cumulative Stats:
>>    Block Size:                  1b+
>>  No. of Reads:                    0
>> No. of Writes:              3264419
>>  %-latency   Avg-latency   Min-Latency   Max-Latency   No. of calls
>>    Fop
>>  ---------   -----------   -----------   -----------   ------------
>>   ----
>>       0.00       0.00 us       0.00 us       0.00 us             90
>> FORGET
>>       0.00       0.00 us       0.00 us       0.00 us           9462
>>  RELEASE
>>       0.00       0.00 us       0.00 us       0.00 us           4254
>> RELEASEDIR
>>       0.00      50.52 us      13.00 us     190.00 us             71
>>  FLUSH
>>       0.00     186.97 us      87.00 us     713.00 us             86
>> REMOVEXATTR
>>       0.00      79.32 us      33.00 us     189.00 us            228
>>     LK
>>       0.00     220.98 us     129.00 us     513.00 us             86
>>  SETATTR
>>       0.01     259.30 us      26.00 us    2632.00 us            137
>>  READDIR
>>       0.02     322.76 us     145.00 us    2125.00 us            321
>>  XATTROP
>>       0.03     109.55 us       2.00 us    1258.00 us           1193
>>  OPENDIR
>>       0.05      70.21 us      21.00 us     431.00 us           3196
>>  DISCARD
>>       0.05     169.26 us      21.00 us    2315.00 us           1545
>> GETXATTR
>>       0.12     176.85 us      63.00 us    2844.00 us           3206
>>   OPEN
>>       0.61     303.49 us      90.00 us    3085.00 us           9633
>>  FSTAT
>>       2.44     305.66 us      28.00 us    3716.00 us          38230
>> LOOKUP
>>       4.52     266.22 us      55.00 us   53424.00 us          81455
>> FXATTROP
>>       6.96    1397.99 us      51.00 us   64822.00 us          23889
>>  FSYNC
>>      16.48      84.74 us      25.00 us    6917.00 us         932592
>>  WRITE
>>      30.16     106.90 us      13.00 us 3920189.00 us        1353046
>>  INODELK
>>      38.55    1794.52 us      14.00 us 16210553.00 us         103039
>> FINODELK
>>
>>     Duration: 66562 seconds
>>    Data Read: 0 bytes
>> Data Written: 3264419 bytes
>>
>> Interval 1 Stats:
>>    Block Size:                  1b+
>>  No. of Reads:                    0
>> No. of Writes:               794080
>>  %-latency   Avg-latency   Min-Latency   Max-Latency   No. of calls
>>    Fop
>>  ---------   -----------   -----------   -----------   ------------
>>   ----
>>       0.00       0.00 us       0.00 us       0.00 us           2093
>>  RELEASE
>>       0.00       0.00 us       0.00 us       0.00 us           1093
>> RELEASEDIR
>>       0.00      70.31 us      26.00 us     125.00 us             16
>>  FLUSH
>>       0.00     193.10 us     103.00 us     713.00 us             71
>> REMOVEXATTR
>>       0.01     227.32 us     133.00 us     513.00 us             71
>>  SETATTR
>>       0.01      79.32 us      33.00 us     189.00 us            228
>>     LK
>>       0.01     259.83 us      35.00 us    1138.00 us             89
>>  READDIR
>>       0.03     318.26 us     145.00 us    2047.00 us            253
>>  XATTROP
>>       0.04     112.67 us       3.00 us    1258.00 us           1093
>>  OPENDIR
>>       0.06     167.98 us      23.00 us    1951.00 us           1014
>> GETXATTR
>>       0.08      70.97 us      22.00 us     431.00 us           3053
>>  DISCARD
>>       0.13     183.78 us      66.00 us    2844.00 us           2091
>>   OPEN
>>       1.01     303.82 us      90.00 us    3085.00 us           9610
>>  FSTAT
>>       3.27     316.59 us      30.00 us    3716.00 us          29820
>> LOOKUP
>>       5.83     265.79 us      59.00 us   53424.00 us          63296
>> FXATTROP
>>       7.95    1373.89 us      51.00 us   64822.00 us          16717
>>  FSYNC
>>      23.17     851.99 us      14.00 us 16210553.00 us          78555
>> FINODELK
>>      24.04      87.44 us      27.00 us    6917.00 us         794081
>>  WRITE
>>      34.36     111.91 us      14.00 us  984871.00 us         886790
>>  INODELK
>>
>>     Duration: 7485 seconds
>>    Data Read: 0 bytes
>> Data Written: 794080 bytes
>>
>>
>> -----------------------
>> Here is the data from the volume that is backed by the SHDDs and has
one
>> failed disk:
>> [root at ovirt2 yum.repos.d]# gluster volume profile data-hdd info
>> Brick: 172.172.1.12:/gluster/brick3/data-hdd
>> --------------------------------------------
>> Cumulative Stats:
>>    Block Size:                256b+                 512b+
>> 1024b+
>>  No. of Reads:                 1702                    86
>>     16
>> No. of Writes:                    0                   767
>>     71
>>
>>    Block Size:               2048b+                4096b+
>> 8192b+
>>  No. of Reads:                   19                 51841
>>   2049
>> No. of Writes:                   76                 60668
>>  35727
>>
>>    Block Size:              16384b+               32768b+
>>  65536b+
>>  No. of Reads:                 1744                   639
>>   1088
>> No. of Writes:                 8524                  2410
>>   1285
>>
>>    Block Size:             131072b+
>>  No. of Reads:               771999
>> No. of Writes:                29584
>>  %-latency   Avg-latency   Min-Latency   Max-Latency   No. of calls
>>    Fop
>>  ---------   -----------   -----------   -----------   ------------
>>   ----
>>       0.00       0.00 us       0.00 us       0.00 us           2902
>>  RELEASE
>>       0.00       0.00 us       0.00 us       0.00 us           1517
>> RELEASEDIR
>>       0.00     197.00 us     197.00 us     197.00 us              1
>>  FTRUNCATE
>>       0.00      70.24 us      16.00 us     758.00 us             51
>>  FLUSH
>>       0.00     143.93 us      82.00 us     305.00 us             57
>> REMOVEXATTR
>>       0.00     178.63 us     105.00 us     712.00 us             60
>>  SETATTR
>>       0.00      67.30 us      19.00 us     572.00 us            555
>>     LK
>>       0.00     322.80 us      23.00 us    4673.00 us            138
>>  READDIR
>>       0.00     336.56 us     106.00 us   11994.00 us            237
>>  XATTROP
>>       0.00      84.70 us      28.00 us    1071.00 us           3469
>> STATFS
>>       0.01     387.75 us       2.00 us  146017.00 us           1467
>>  OPENDIR
>>       0.01     148.59 us      21.00 us   64374.00 us           4454
>>   STAT
>>       0.02     783.02 us      16.00 us   93502.00 us           1902
>> GETXATTR
>>       0.03    1516.10 us      17.00 us  210690.00 us           1364
>>  ENTRYLK
>>       0.03    2555.47 us     300.00 us  674454.00 us           1064
>> READDIRP
>>       0.07      85.74 us      19.00 us   68340.00 us          62849
>>  FSTAT
>>       0.07    1978.12 us      59.00 us  202596.00 us           2729
>>   OPEN
>>       0.22     708.57 us      15.00 us  394799.00 us          25447
>> LOOKUP
>>       5.94    2331.74 us      15.00 us 1099530.00 us         207534
>> FINODELK
>>       7.31    8311.75 us      58.00 us 1800216.00 us          71668
>> FXATTROP
>>      12.49    7735.19 us      51.00 us 3595513.00 us         131642
>>  WRITE
>>      17.70     957.08 us      16.00 us 13700466.00 us        1508160
>>  INODELK
>>      24.55    2546.43 us      26.00 us 5077347.00 us         786060
>>   READ
>>      31.56   49699.15 us      47.00 us 3746331.00 us          51777
>>  FSYNC
>>
>>     Duration: 10101 seconds
>>    Data Read: 101562897361 bytes
>> Data Written: 4834450432 bytes
>>
>> Interval 0 Stats:
>>    Block Size:                256b+                 512b+
>> 1024b+
>>  No. of Reads:                 1702                    86
>>     16
>> No. of Writes:                    0                   767
>>     71
>>
>>    Block Size:               2048b+                4096b+
>> 8192b+
>>  No. of Reads:                   19                 51841
>>   2049
>> No. of Writes:                   76                 60668
>>  35727
>>
>>    Block Size:              16384b+               32768b+
>>  65536b+
>>  No. of Reads:                 1744                   639
>>   1088
>> No. of Writes:                 8524                  2410
>>   1285
>>
>>    Block Size:             131072b+
>>  No. of Reads:               771999
>> No. of Writes:                29584
>>  %-latency   Avg-latency   Min-Latency   Max-Latency   No. of calls
>>    Fop
>>  ---------   -----------   -----------   -----------   ------------
>>   ----
>>       0.00       0.00 us       0.00 us       0.00 us           2902
>>  RELEASE
>>       0.00       0.00 us       0.00 us       0.00 us           1517
>> RELEASEDIR
>>       0.00     197.00 us     197.00 us     197.00 us              1
>>  FTRUNCATE
>>       0.00      70.24 us      16.00 us     758.00 us             51
>>  FLUSH
>>       0.00     143.93 us      82.00 us     305.00 us             57
>> REMOVEXATTR
>>       0.00     178.63 us     105.00 us     712.00 us             60
>>  SETATTR
>>       0.00      67.30 us      19.00 us     572.00 us            555
>>     LK
>>       0.00     322.80 us      23.00 us    4673.00 us            138
>>  READDIR
>>       0.00     336.56 us     106.00 us   11994.00 us            237
>>  XATTROP
>>       0.00      84.70 us      28.00 us    1071.00 us           3469
>> STATFS
>>       0.01     387.75 us       2.00 us  146017.00 us           1467
>>  OPENDIR
>>       0.01     148.59 us      21.00 us   64374.00 us           4454
>>   STAT
>>       0.02     783.02 us      16.00 us   93502.00 us           1902
>> GETXATTR
>>       0.03    1516.10 us      17.00 us  210690.00 us           1364
>>  ENTRYLK
>>       0.03    2555.47 us     300.00 us  674454.00 us           1064
>> READDIRP
>>       0.07      85.73 us      19.00 us   68340.00 us          62849
>>  FSTAT
>>       0.07    1978.12 us      59.00 us  202596.00 us           2729
>>   OPEN
>>       0.22     708.57 us      15.00 us  394799.00 us          25447
>> LOOKUP
>>       5.94    2334.57 us      15.00 us 1099530.00 us         207534
>> FINODELK
>>       7.31    8311.49 us      58.00 us 1800216.00 us          71668
>> FXATTROP
>>      12.49    7735.32 us      51.00 us 3595513.00 us         131642
>>  WRITE
>>      17.71     957.08 us      16.00 us 13700466.00 us        1508160
>>  INODELK
>>      24.56    2546.42 us      26.00 us 5077347.00 us         786060
>>   READ
>>      31.54   49651.63 us      47.00 us 3746331.00 us          51777
>>  FSYNC
>>
>>     Duration: 10101 seconds
>>    Data Read: 101562897361 bytes
>> Data Written: 4834450432 bytes
>>
>>
>> On Tue, May 29, 2018 at 2:55 PM, Jim Kusznir <jim at
palousetech.com> wrote:
>>
>>> Thank you for your response.
>>>
>>> I have 4 gluster volumes.  3 are replica 2 + arbitrator.  replica
bricks
>>> are on ovirt1 and ovirt2, arbitrator on ovirt3.  The 4th volume is
replica
>>> 3, with a brick on all three ovirt machines.
>>>
>>> The first 3 volumes are on an SSD disk; the 4th is on a Seagate
SSHD
>>> (same in all three machines).  On ovirt3, the SSHD has reported
hard IO
>>> failures, and that brick is offline.  However, the other two
replicas are
>>> fully operational (although they still show contents in the heal
info
>>> command that won't go away, but that may be the case until I
replace the
>>> failed disk).
>>>
>>> What is bothering me is that ALL 4 gluster volumes are showing
horrible
>>> performance issues.  At this point, as the bad disk has been
completely
>>> offlined, I would expect gluster to perform at normal speed, but
that is
>>> definitely not the case.
>>>
>>> I've also noticed that the performance hits seem to come in
waves:
>>> things seem to work acceptably (but slow) for a while, then
suddenly, its
>>> as if all disk IO on all volumes (including non-gluster local OS
disk
>>> volumes for the hosts) pause for about 30 seconds, then IO resumes
again.
>>> During those times, I start getting VM not responding and host not
>>> responding notices as well as the applications having major issues.
>>>
>>> I've shut down most of my VMs and am down to just my essential
core VMs
>>> (shedded about 75% of my VMs).  I still am experiencing the same
issues.
>>>
>>> Am I correct in believing that once the failed disk was brought
offline
>>> that performance should return to normal?
>>>
>>> On Tue, May 29, 2018 at 1:27 PM, Alex K <rightkicktech at
gmail.com> wrote:
>>>
>>>> I would check disks status and accessibility of mount points
where your
>>>> gluster volumes reside.
>>>>
>>>> On Tue, May 29, 2018, 22:28 Jim Kusznir <jim at
palousetech.com> wrote:
>>>>
>>>>> On one ovirt server, I'm now seeing these messages:
>>>>> [56474.239725] blk_update_request: 63 callbacks suppressed
>>>>> [56474.239732] blk_update_request: I/O error, dev dm-2,
sector 0
>>>>> [56474.240602] blk_update_request: I/O error, dev dm-2,
sector
>>>>> 3905945472
>>>>> [56474.241346] blk_update_request: I/O error, dev dm-2,
sector
>>>>> 3905945584
>>>>> [56474.242236] blk_update_request: I/O error, dev dm-2,
sector 2048
>>>>> [56474.243072] blk_update_request: I/O error, dev dm-2,
sector
>>>>> 3905943424
>>>>> [56474.243997] blk_update_request: I/O error, dev dm-2,
sector
>>>>> 3905943536
>>>>> [56474.247347] blk_update_request: I/O error, dev dm-2,
sector 0
>>>>> [56474.248315] blk_update_request: I/O error, dev dm-2,
sector
>>>>> 3905945472
>>>>> [56474.249231] blk_update_request: I/O error, dev dm-2,
sector
>>>>> 3905945584
>>>>> [56474.250221] blk_update_request: I/O error, dev dm-2,
sector 2048
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> On Tue, May 29, 2018 at 11:59 AM, Jim Kusznir <jim at
palousetech.com>
>>>>> wrote:
>>>>>
>>>>>> I see in messages on ovirt3 (my 3rd machine, the one
upgraded to 4.2):
>>>>>>
>>>>>> May 29 11:54:41 ovirt3 ovs-vsctl:
ovs|00001|db_ctl_base|ERR|unix:/var/run/openvswitch/db.sock:
>>>>>> database connection failed (No such file or directory)
>>>>>> May 29 11:54:51 ovirt3 ovs-vsctl:
ovs|00001|db_ctl_base|ERR|unix:/var/run/openvswitch/db.sock:
>>>>>> database connection failed (No such file or directory)
>>>>>> May 29 11:55:01 ovirt3 ovs-vsctl:
ovs|00001|db_ctl_base|ERR|unix:/var/run/openvswitch/db.sock:
>>>>>> database connection failed (No such file or directory)
>>>>>> (appears a lot).
>>>>>>
>>>>>> I also found on the ssh session of that, some sysv
warnings about the
>>>>>> backing disk for one of the gluster volumes (straight
replica 3).  The
>>>>>> glusterfs process for that disk on that machine went
offline.  Its my
>>>>>> understanding that it should continue to work with the
other two machines
>>>>>> while I attempt to replace that disk, right?  Attempted
writes (touching an
>>>>>> empty file) can take 15 seconds, repeating it later
will be much faster.
>>>>>>
>>>>>> Gluster generates a bunch of different log files, I
don't know what
>>>>>> ones you want, or from which machine(s).
>>>>>>
>>>>>> How do I do "volume profiling"?
>>>>>>
>>>>>> Thanks!
>>>>>>
>>>>>> On Tue, May 29, 2018 at 11:53 AM, Sahina Bose
<sabose at redhat.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Do you see errors reported in the mount logs for
the volume? If so,
>>>>>>> could you attach the logs?
>>>>>>> Any issues with your underlying disks. Can you also
attach output of
>>>>>>> volume profiling?
>>>>>>>
>>>>>>> On Wed, May 30, 2018 at 12:13 AM, Jim Kusznir
<jim at palousetech.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Ok, things have gotten MUCH worse this morning.
I'm getting random
>>>>>>>> errors from VMs, right now, about a third of my
VMs have been paused due to
>>>>>>>> storage issues, and most of the remaining VMs
are not performing well.
>>>>>>>>
>>>>>>>> At this point, I am in full EMERGENCY mode, as
my production
>>>>>>>> services are now impacted, and I'm getting
calls coming in with problems...
>>>>>>>>
>>>>>>>> I'd greatly appreciate help...VMs are
running VERY slowly (when
>>>>>>>> they run), and they are steadily getting worse.
I don't know why.  I was
>>>>>>>> seeing CPU peaks (to 100%) on several VMs, in
perfect sync, for a few
>>>>>>>> minutes at a time (while the VM became
unresponsive and any VMs I was
>>>>>>>> logged into that were linux were giving me the
CPU stuck messages in my
>>>>>>>> origional post).  Is all this storage related?
>>>>>>>>
>>>>>>>> I also have two different gluster volumes for
VM storage, and only
>>>>>>>> one had the issues, but now VMs in both are
being affected at the same time
>>>>>>>> and same way.
>>>>>>>>
>>>>>>>> --Jim
>>>>>>>>
>>>>>>>> On Mon, May 28, 2018 at 10:50 PM, Sahina Bose
<sabose at redhat.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> [Adding gluster-users to look at the heal
issue]
>>>>>>>>>
>>>>>>>>> On Tue, May 29, 2018 at 9:17 AM, Jim
Kusznir <jim at palousetech.com>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>> Hello:
>>>>>>>>>>
>>>>>>>>>> I've been having some cluster and
gluster performance issues
>>>>>>>>>> lately.  I also found that my cluster
was out of date, and was trying to
>>>>>>>>>> apply updates (hoping to fix some of
these), and discovered the ovirt 4.1
>>>>>>>>>> repos were taken completely offline. 
So, I was forced to begin an upgrade
>>>>>>>>>> to 4.2.  According to docs I
found/read, I needed only add the new repo, do
>>>>>>>>>> a yum update, reboot, and be good on my
hosts (did the yum update, the
>>>>>>>>>> engine-setup on my hosted engine). 
Things seemed to work relatively well,
>>>>>>>>>> except for a gluster sync issue that
showed up.
>>>>>>>>>>
>>>>>>>>>> My cluster is a 3 node hyperconverged
cluster.  I upgraded the
>>>>>>>>>> hosted engine first, then engine 3. 
When engine 3 came back up, for some
>>>>>>>>>> reason one of my gluster volumes would
not sync.  Here's sample output:
>>>>>>>>>>
>>>>>>>>>> [root at ovirt3 ~]# gluster volume heal
data-hdd info
>>>>>>>>>> Brick
172.172.1.11:/gluster/brick3/data-hdd
>>>>>>>>>>
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/48d7ecb8-7ac5-4
>>>>>>>>>>
725-bca5-b3519681cf2f/0d6080b0-7018-4fa3-bb82-1dd9ef07d9b9
>>>>>>>>>>
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/647be733-f153-4
>>>>>>>>>>
cdc-85bd-ba72544c2631/b453a300-0602-4be1-8310-8bd5abe00971
>>>>>>>>>>
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/6da854d1-b6be-4
>>>>>>>>>>
46b-9bf0-90a0dbbea830/3c93bd1f-b7fa-4aa2-b445-6904e31839ba
>>>>>>>>>>
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/7f647567-d18c-4
>>>>>>>>>>
4f1-a58e-9b8865833acb/f9364470-9770-4bb1-a6b9-a54861849625
>>>>>>>>>>
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/f3c8e7aa-6ef2-4
>>>>>>>>>>
2a7-93d4-e0a4df6dd2fa/2eb0b1ad-2606-44ef-9cd3-ae59610a504b
>>>>>>>>>>
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/b1ea3f62-0f05-4
>>>>>>>>>>
ded-8c82-9c91c90e0b61/d5d6bf5a-499f-431d-9013-5453db93ed32
>>>>>>>>>>
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/8c8b5147-e9d6-4
>>>>>>>>>>
810-b45b-185e3ed65727/16f08231-93b0-489d-a2fd-687b6bf88eaa
>>>>>>>>>>
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/12924435-b9c2-4
>>>>>>>>>>
aab-ba19-1c1bc31310ef/07b3db69-440e-491e-854c-bbfa18a7cff2
>>>>>>>>>> Status: Connected
>>>>>>>>>> Number of entries: 8
>>>>>>>>>>
>>>>>>>>>> Brick
172.172.1.12:/gluster/brick3/data-hdd
>>>>>>>>>>
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/48d7ecb8-7ac5-4
>>>>>>>>>>
725-bca5-b3519681cf2f/0d6080b0-7018-4fa3-bb82-1dd9ef07d9b9
>>>>>>>>>>
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/647be733-f153-4
>>>>>>>>>>
cdc-85bd-ba72544c2631/b453a300-0602-4be1-8310-8bd5abe00971
>>>>>>>>>>
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/b1ea3f62-0f05-4
>>>>>>>>>>
ded-8c82-9c91c90e0b61/d5d6bf5a-499f-431d-9013-5453db93ed32
>>>>>>>>>>
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/6da854d1-b6be-4
>>>>>>>>>>
46b-9bf0-90a0dbbea830/3c93bd1f-b7fa-4aa2-b445-6904e31839ba
>>>>>>>>>>
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/7f647567-d18c-4
>>>>>>>>>>
4f1-a58e-9b8865833acb/f9364470-9770-4bb1-a6b9-a54861849625
>>>>>>>>>>
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/8c8b5147-e9d6-4
>>>>>>>>>>
810-b45b-185e3ed65727/16f08231-93b0-489d-a2fd-687b6bf88eaa
>>>>>>>>>>
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/12924435-b9c2-4
>>>>>>>>>>
aab-ba19-1c1bc31310ef/07b3db69-440e-491e-854c-bbfa18a7cff2
>>>>>>>>>>
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/f3c8e7aa-6ef2-4
>>>>>>>>>>
2a7-93d4-e0a4df6dd2fa/2eb0b1ad-2606-44ef-9cd3-ae59610a504b
>>>>>>>>>> Status: Connected
>>>>>>>>>> Number of entries: 8
>>>>>>>>>>
>>>>>>>>>> Brick
172.172.1.13:/gluster/brick3/data-hdd
>>>>>>>>>>
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/b1ea3f62-0f05-4
>>>>>>>>>>
ded-8c82-9c91c90e0b61/d5d6bf5a-499f-431d-9013-5453db93ed32
>>>>>>>>>>
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/8c8b5147-e9d6-4
>>>>>>>>>>
810-b45b-185e3ed65727/16f08231-93b0-489d-a2fd-687b6bf88eaa
>>>>>>>>>>
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/12924435-b9c2-4
>>>>>>>>>>
aab-ba19-1c1bc31310ef/07b3db69-440e-491e-854c-bbfa18a7cff2
>>>>>>>>>>
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/f3c8e7aa-6ef2-4
>>>>>>>>>>
2a7-93d4-e0a4df6dd2fa/2eb0b1ad-2606-44ef-9cd3-ae59610a504b
>>>>>>>>>>
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/647be733-f153-4
>>>>>>>>>>
cdc-85bd-ba72544c2631/b453a300-0602-4be1-8310-8bd5abe00971
>>>>>>>>>>
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/48d7ecb8-7ac5-4
>>>>>>>>>>
725-bca5-b3519681cf2f/0d6080b0-7018-4fa3-bb82-1dd9ef07d9b9
>>>>>>>>>>
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/6da854d1-b6be-4
>>>>>>>>>>
46b-9bf0-90a0dbbea830/3c93bd1f-b7fa-4aa2-b445-6904e31839ba
>>>>>>>>>>
/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/7f647567-d18c-4
>>>>>>>>>>
4f1-a58e-9b8865833acb/f9364470-9770-4bb1-a6b9-a54861849625
>>>>>>>>>> Status: Connected
>>>>>>>>>> Number of entries: 8
>>>>>>>>>>
>>>>>>>>>> ---------
>>>>>>>>>> Its been in this state for a couple
days now, and bandwidth
>>>>>>>>>> monitoring shows no appreciable data
moving.  I've tried repeatedly
>>>>>>>>>> commanding a full heal from all three
clusters in the node.  Its always the
>>>>>>>>>> same files that need healing.
>>>>>>>>>>
>>>>>>>>>> When running gluster volume heal
data-hdd statistics, I see
>>>>>>>>>> sometimes different information, but
always some number of "heal failed"
>>>>>>>>>> entries.  It shows 0 for split brain.
>>>>>>>>>>
>>>>>>>>>> I'm not quite sure what to do.  I
suspect it may be due to nodes
>>>>>>>>>> 1 and 2 still being on the older
ovirt/gluster release, but I'm afraid to
>>>>>>>>>> upgrade and reboot them until I have a
good gluster sync (don't need to
>>>>>>>>>> create a split brain issue).  How do I
proceed with this?
>>>>>>>>>>
>>>>>>>>>> Second issue: I've been
experiencing VERY POOR performance on
>>>>>>>>>> most of my VMs.  To the tune that
logging into a windows 10 vm via remote
>>>>>>>>>> desktop can take 5 minutes, launching
quickbooks inside said vm can easily
>>>>>>>>>> take 10 minutes.  On some linux VMs, I
get random messages like this:
>>>>>>>>>> Message from syslogd at unifi at May 28
20:39:23 ...
>>>>>>>>>>  kernel:[6171996.308904] NMI watchdog:
BUG: soft lockup - CPU#0
>>>>>>>>>> stuck for 22s! [mongod:14766]
>>>>>>>>>>
>>>>>>>>>> (the process and PID are often
different)
>>>>>>>>>>
>>>>>>>>>> I'm not quite sure what to do about
this either.  My initial
>>>>>>>>>> thought was upgrad everything to
current and see if its still there, but I
>>>>>>>>>> cannot move forward with that until my
gluster is healed...
>>>>>>>>>>
>>>>>>>>>> Thanks!
>>>>>>>>>> --Jim
>>>>>>>>>>
>>>>>>>>>>
_______________________________________________
>>>>>>>>>> Users mailing list -- users at
ovirt.org
>>>>>>>>>> To unsubscribe send an email to
users-leave at ovirt.org
>>>>>>>>>> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
>>>>>>>>>> oVirt Code of Conduct:
https://www.ovirt.org/communit
>>>>>>>>>> y/about/community-guidelines/
>>>>>>>>>> List Archives:
https://lists.ovirt.org/archiv
>>>>>>>>>> es/list/users at
ovirt.org/message/3LEV6ZQ3JV2XLAL7NYBTXOYMYUOTIRQF/
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>> _______________________________________________
>>>>> Users mailing list -- users at ovirt.org
>>>>> To unsubscribe send an email to users-leave at ovirt.org
>>>>> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
>>>>> oVirt Code of Conduct: https://www.ovirt.org/communit
>>>>> y/about/community-guidelines/
>>>>> List Archives: https://lists.ovirt.org/archiv
>>>>> es/list/users at
ovirt.org/message/ACO7RFSLBSRBAIONIC2HQ6Z24ZDES5MF/
>>>>>
>>>>
>>>
>>
>
> _______________________________________________
> Users mailing list -- users at ovirt.org
> To unsubscribe send an email to users-leave at ovirt.org
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct: https://www.ovirt.org/community/about/community-
> guidelines/
> List Archives: https://lists.ovirt.org/archives/list/users at ovirt.org/
> message/3DEQQLJM3WHQNZJ7KEMRZVFZ52MTIL74/
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
<http://lists.gluster.org/pipermail/gluster-users/attachments/20180530/bccc6e69/attachment.html>
Ravishankar N
2018-May-30  04:43 UTC
[Gluster-users] [ovirt-users] Re: Gluster problems, cluster performance issues
@Jim Kusznir For the heal issue, can you provide the getfattr output of one of the 8 files in question from all 3 bricks? Example: `getfattr -d -m . -e hex /gluster/brick3/data-hdd/cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/48d7ecb8-7ac5-4725-bca5-b3519681cf2f/0d6080b0-7018-4fa3-bb82-1dd9ef07d9b9` Also provide the stat output of the same file from all 3 bricks. Thanks, Ravi On 05/30/2018 09:47 AM, Krutika Dhananjay wrote:> Adding Ravi to look into the heal issue. > > As for the fsync hang and subsequent IO errors, it seems a lot like > https://bugzilla.redhat.com/show_bug.cgi?id=1497156 and Paolo Bonzini > from qemu had pointed out that this would be fixed by the following > commit: > > commit e72c9a2a67a6400c8ef3d01d4c461dbbbfa0e1f0 > Author: Paolo Bonzini <pbonzini at redhat.com <mailto:pbonzini at redhat.com>> > Date: Wed Jun 21 16:35:46 2017 +0200 > > scsi: virtio_scsi: let host do exception handling > > virtio_scsi tries to do exception handling after the default 30 seconds > timeout expires. However, it's better to let the host control the > timeout, otherwise with a heavy I/O load it is likely that an abort will > also timeout. This leads to fatal errors like filesystems going > offline. > > Disable the 'sd' timeout and allow the host to do exception handling, > following the precedent of the storvsc driver. > > Hannes has a proposal to introduce timeouts in virtio, but this provides > an immediate solution for stable kernels too. > > [mkp: fixed typo] > > Reported-by: Douglas Miller <dougmill at linux.vnet.ibm.com <mailto:dougmill at linux.vnet.ibm.com>> > Cc: "James E.J. Bottomley" <jejb at linux.vnet.ibm.com <mailto:jejb at linux.vnet.ibm.com>> > Cc: "Martin K. Petersen" <martin.petersen at oracle.com <mailto:martin.petersen at oracle.com>> > Cc: Hannes Reinecke <hare at suse.de <mailto:hare at suse.de>> > Cc:linux-scsi at vger.kernel.org <mailto:linux-scsi at vger.kernel.org> > Cc:stable at vger.kernel.org <mailto:stable at vger.kernel.org> > Signed-off-by: Paolo Bonzini <pbonzini at redhat.com <mailto:pbonzini at redhat.com>> > Signed-off-by: Martin K. Petersen <martin.petersen at oracle.com <mailto:martin.petersen at oracle.com>> > > Adding Paolo/Kevin to comment. > > As for the poor gluster performance, could you disable > cluster.eager-lock and see if that makes any difference: > > # gluster volume set <VOL> cluster.eager-lock off > > Do also capture the volume profile again if you still see performance > issues after disabling eager-lock. > > -Krutika > > > On Wed, May 30, 2018 at 6:55 AM, Jim Kusznir <jim at palousetech.com > <mailto:jim at palousetech.com>> wrote: > > I also finally found the following in my system log on one server: > > [10679.524491] INFO: task glusterclogro:14933 blocked for more > than 120 seconds. > [10679.525826] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" > disables this message. > [10679.527144] glusterclogro? ?D ffff97209832bf40 ? ?0 14933? ? ? > 1 0x00000080 > [10679.527150] Call Trace: > [10679.527161]? [<ffffffffb9913f79>] schedule+0x29/0x70 > [10679.527218]? [<ffffffffc060e388>] > _xfs_log_force_lsn+0x2e8/0x340 [xfs] > [10679.527225]? [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20 > [10679.527254]? [<ffffffffc05eeb97>] xfs_file_fsync+0x107/0x1e0 [xfs] > [10679.527260]? [<ffffffffb944f0e7>] do_fsync+0x67/0xb0 > [10679.527268]? [<ffffffffb992076f>] ? > system_call_after_swapgs+0xbc/0x160 > [10679.527271]? [<ffffffffb944f3d0>] SyS_fsync+0x10/0x20 > [10679.527275]? [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21 > [10679.527279]? [<ffffffffb992077b>] ? > system_call_after_swapgs+0xc8/0x160 > [10679.527283] INFO: task glusterposixfsy:14941 blocked for more > than 120 seconds. > [10679.528608] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" > disables this message. > [10679.529956] glusterposixfsy D ffff972495f84f10 ? ?0 14941? ? ? > 1 0x00000080 > [10679.529961] Call Trace: > [10679.529966]? [<ffffffffb9913f79>] schedule+0x29/0x70 > [10679.530003]? [<ffffffffc060e388>] > _xfs_log_force_lsn+0x2e8/0x340 [xfs] > [10679.530008]? [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20 > [10679.530038]? [<ffffffffc05eeb97>] xfs_file_fsync+0x107/0x1e0 [xfs] > [10679.530042]? [<ffffffffb944f0e7>] do_fsync+0x67/0xb0 > [10679.530046]? [<ffffffffb992076f>] ? > system_call_after_swapgs+0xbc/0x160 > [10679.530050]? [<ffffffffb944f3f3>] SyS_fdatasync+0x13/0x20 > [10679.530054]? [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21 > [10679.530058]? [<ffffffffb992077b>] ? > system_call_after_swapgs+0xc8/0x160 > [10679.530062] INFO: task glusteriotwr13:15486 blocked for more > than 120 seconds. > [10679.531805] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" > disables this message. > [10679.533732] glusteriotwr13? D ffff9720a83f0000 ? ?0 15486? ? ? > 1 0x00000080 > [10679.533738] Call Trace: > [10679.533747]? [<ffffffffb9913f79>] schedule+0x29/0x70 > [10679.533799]? [<ffffffffc060e388>] > _xfs_log_force_lsn+0x2e8/0x340 [xfs] > [10679.533806]? [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20 > [10679.533846]? [<ffffffffc05eeb97>] xfs_file_fsync+0x107/0x1e0 [xfs] > [10679.533852]? [<ffffffffb944f0e7>] do_fsync+0x67/0xb0 > [10679.533858]? [<ffffffffb992076f>] ? > system_call_after_swapgs+0xbc/0x160 > [10679.533863]? [<ffffffffb944f3f3>] SyS_fdatasync+0x13/0x20 > [10679.533868]? [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21 > [10679.533873]? [<ffffffffb992077b>] ? > system_call_after_swapgs+0xc8/0x160 > [10919.512757] INFO: task glusterclogro:14933 blocked for more > than 120 seconds. > [10919.514714] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" > disables this message. > [10919.516663] glusterclogro? ?D ffff97209832bf40 ? ?0 14933? ? ? > 1 0x00000080 > [10919.516677] Call Trace: > [10919.516690]? [<ffffffffb9913f79>] schedule+0x29/0x70 > [10919.516696]? [<ffffffffb99118e9>] schedule_timeout+0x239/0x2c0 > [10919.516703]? [<ffffffffb951cc04>] ? blk_finish_plug+0x14/0x40 > [10919.516768]? [<ffffffffc05e9224>] ? > _xfs_buf_ioapply+0x334/0x460 [xfs] > [10919.516774]? [<ffffffffb991432d>] wait_for_completion+0xfd/0x140 > [10919.516782]? [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20 > [10919.516821]? [<ffffffffc05eb0a3>] ? _xfs_buf_read+0x23/0x40 [xfs] > [10919.516859]? [<ffffffffc05eafa9>] > xfs_buf_submit_wait+0xf9/0x1d0 [xfs] > [10919.516902]? [<ffffffffc061b279>] ? > xfs_trans_read_buf_map+0x199/0x400 [xfs] > [10919.516940]? [<ffffffffc05eb0a3>] _xfs_buf_read+0x23/0x40 [xfs] > [10919.516977]? [<ffffffffc05eb1b9>] xfs_buf_read_map+0xf9/0x160 [xfs] > [10919.517022]? [<ffffffffc061b279>] > xfs_trans_read_buf_map+0x199/0x400 [xfs] > [10919.517057]? [<ffffffffc05c8d04>] xfs_da_read_buf+0xd4/0x100 [xfs] > [10919.517091]? [<ffffffffc05c8d53>] xfs_da3_node_read+0x23/0xd0 [xfs] > [10919.517126]? [<ffffffffc05c9fee>] > xfs_da3_node_lookup_int+0x6e/0x2f0 [xfs] > [10919.517160]? [<ffffffffc05d5a1d>] > xfs_dir2_node_lookup+0x4d/0x170 [xfs] > [10919.517194]? [<ffffffffc05ccf5d>] xfs_dir_lookup+0x1bd/0x1e0 [xfs] > [10919.517233]? [<ffffffffc05fd8d9>] xfs_lookup+0x69/0x140 [xfs] > [10919.517271]? [<ffffffffc05fa018>] xfs_vn_lookup+0x78/0xc0 [xfs] > [10919.517278]? [<ffffffffb9425cf3>] lookup_real+0x23/0x60 > [10919.517283]? [<ffffffffb9426702>] __lookup_hash+0x42/0x60 > [10919.517288]? [<ffffffffb942d519>] SYSC_renameat2+0x3a9/0x5a0 > [10919.517296]? [<ffffffffb94d3753>] ? > selinux_file_free_security+0x23/0x30 > [10919.517304]? [<ffffffffb992077b>] ? > system_call_after_swapgs+0xc8/0x160 > [10919.517309]? [<ffffffffb992076f>] ? > system_call_after_swapgs+0xbc/0x160 > [10919.517313]? [<ffffffffb992077b>] ? > system_call_after_swapgs+0xc8/0x160 > [10919.517318]? [<ffffffffb992076f>] ? > system_call_after_swapgs+0xbc/0x160 > [10919.517323]? [<ffffffffb942e58e>] SyS_renameat2+0xe/0x10 > [10919.517328]? [<ffffffffb942e5ce>] SyS_rename+0x1e/0x20 > [10919.517333]? [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21 > [10919.517339]? [<ffffffffb992077b>] ? > system_call_after_swapgs+0xc8/0x160 > [11159.496095] INFO: task glusteriotwr9:15482 blocked for more > than 120 seconds. > [11159.497546] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" > disables this message. > [11159.498978] glusteriotwr9? ?D ffff971fa0fa1fa0 ? ?0 15482? ? ? > 1 0x00000080 > [11159.498984] Call Trace: > [11159.498995]? [<ffffffffb9911f00>] ? bit_wait+0x50/0x50 > [11159.498999]? [<ffffffffb9913f79>] schedule+0x29/0x70 > [11159.499003]? [<ffffffffb99118e9>] schedule_timeout+0x239/0x2c0 > [11159.499056]? [<ffffffffc05dd9b7>] ? > xfs_iext_bno_to_ext+0xa7/0x1a0 [xfs] > [11159.499082]? [<ffffffffc05dd43e>] ? > xfs_iext_bno_to_irec+0x8e/0xd0 [xfs] > [11159.499090]? [<ffffffffb92f7a12>] ? ktime_get_ts64+0x52/0xf0 > [11159.499093]? [<ffffffffb9911f00>] ? bit_wait+0x50/0x50 > [11159.499097]? [<ffffffffb991348d>] io_schedule_timeout+0xad/0x130 > [11159.499101]? [<ffffffffb9913528>] io_schedule+0x18/0x20 > [11159.499104]? [<ffffffffb9911f11>] bit_wait_io+0x11/0x50 > [11159.499107]? [<ffffffffb9911ac1>] __wait_on_bit_lock+0x61/0xc0 > [11159.499113]? [<ffffffffb9393634>] __lock_page+0x74/0x90 > [11159.499118]? [<ffffffffb92bc210>] ? wake_bit_function+0x40/0x40 > [11159.499121]? [<ffffffffb9394154>] __find_lock_page+0x54/0x70 > [11159.499125]? [<ffffffffb9394e85>] > grab_cache_page_write_begin+0x55/0xc0 > [11159.499130]? [<ffffffffb9484b76>] iomap_write_begin+0x66/0x100 > [11159.499135]? [<ffffffffb9484edf>] iomap_write_actor+0xcf/0x1d0 > [11159.499140]? [<ffffffffb9484e10>] ? iomap_write_end+0x80/0x80 > [11159.499144]? [<ffffffffb94854e7>] iomap_apply+0xb7/0x150 > [11159.499149]? [<ffffffffb9485621>] > iomap_file_buffered_write+0xa1/0xe0 > [11159.499153]? [<ffffffffb9484e10>] ? iomap_write_end+0x80/0x80 > [11159.499182]? [<ffffffffc05f025d>] > xfs_file_buffered_aio_write+0x12d/0x2c0 [xfs] > [11159.499213]? [<ffffffffc05f057d>] > xfs_file_aio_write+0x18d/0x1b0 [xfs] > [11159.499217]? [<ffffffffb941a533>] do_sync_write+0x93/0xe0 > [11159.499222]? [<ffffffffb941b010>] vfs_write+0xc0/0x1f0 > [11159.499225]? [<ffffffffb941c002>] SyS_pwrite64+0x92/0xc0 > [11159.499230]? [<ffffffffb992076f>] ? > system_call_after_swapgs+0xbc/0x160 > [11159.499234]? [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21 > [11159.499238]? [<ffffffffb992077b>] ? > system_call_after_swapgs+0xc8/0x160 > [11279.488720] INFO: task xfsaild/dm-10:1134 blocked for more than > 120 seconds. > [11279.490197] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" > disables this message. > [11279.491665] xfsaild/dm-10? ?D ffff9720a8660fd0 ? ?0? 1134? ? ? > 2 0x00000000 > [11279.491671] Call Trace: > [11279.491682]? [<ffffffffb92a3a2e>] ? try_to_del_timer_sync+0x5e/0x90 > [11279.491688]? [<ffffffffb9913f79>] schedule+0x29/0x70 > [11279.491744]? [<ffffffffc060de36>] _xfs_log_force+0x1c6/0x2c0 [xfs] > [11279.491750]? [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20 > [11279.491783]? [<ffffffffc0619fec>] ? xfsaild+0x16c/0x6f0 [xfs] > [11279.491817]? [<ffffffffc060df5c>] xfs_log_force+0x2c/0x70 [xfs] > [11279.491849]? [<ffffffffc0619e80>] ? > xfs_trans_ail_cursor_first+0x90/0x90 [xfs] > [11279.491880]? [<ffffffffc0619fec>] xfsaild+0x16c/0x6f0 [xfs] > [11279.491913]? [<ffffffffc0619e80>] ? > xfs_trans_ail_cursor_first+0x90/0x90 [xfs] > [11279.491919]? [<ffffffffb92bb161>] kthread+0xd1/0xe0 > [11279.491926]? [<ffffffffb92bb090>] ? insert_kthread_work+0x40/0x40 > [11279.491932]? [<ffffffffb9920677>] > ret_from_fork_nospec_begin+0x21/0x21 > [11279.491936]? [<ffffffffb92bb090>] ? insert_kthread_work+0x40/0x40 > [11279.491976] INFO: task glusterclogfsyn:14934 blocked for more > than 120 seconds. > [11279.493466] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" > disables this message. > [11279.494952] glusterclogfsyn D ffff97209832af70 ? ?0 14934? ? ? > 1 0x00000080 > [11279.494957] Call Trace: > [11279.494979]? [<ffffffffc0309839>] ? > __split_and_process_bio+0x2e9/0x520 [dm_mod] > [11279.494983]? [<ffffffffb9913f79>] schedule+0x29/0x70 > [11279.494987]? [<ffffffffb99118e9>] schedule_timeout+0x239/0x2c0 > [11279.494997]? [<ffffffffc0309d98>] ? dm_make_request+0x128/0x1a0 > [dm_mod] > [11279.495001]? [<ffffffffb991348d>] io_schedule_timeout+0xad/0x130 > [11279.495005]? [<ffffffffb99145ad>] wait_for_completion_io+0xfd/0x140 > [11279.495010]? [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20 > [11279.495016]? [<ffffffffb951e574>] blkdev_issue_flush+0xb4/0x110 > [11279.495049]? [<ffffffffc06064b9>] > xfs_blkdev_issue_flush+0x19/0x20 [xfs] > [11279.495079]? [<ffffffffc05eec40>] xfs_file_fsync+0x1b0/0x1e0 [xfs] > [11279.495086]? [<ffffffffb944f0e7>] do_fsync+0x67/0xb0 > [11279.495090]? [<ffffffffb992076f>] ? > system_call_after_swapgs+0xbc/0x160 > [11279.495094]? [<ffffffffb944f3d0>] SyS_fsync+0x10/0x20 > [11279.495098]? [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21 > [11279.495102]? [<ffffffffb992077b>] ? > system_call_after_swapgs+0xc8/0x160 > [11279.495105] INFO: task glusterposixfsy:14941 blocked for more > than 120 seconds. > [11279.496606] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" > disables this message. > [11279.498114] glusterposixfsy D ffff972495f84f10 ? ?0 14941? ? ? > 1 0x00000080 > [11279.498118] Call Trace: > [11279.498134]? [<ffffffffc0309839>] ? > __split_and_process_bio+0x2e9/0x520 [dm_mod] > [11279.498138]? [<ffffffffb9913f79>] schedule+0x29/0x70 > [11279.498142]? [<ffffffffb99118e9>] schedule_timeout+0x239/0x2c0 > [11279.498152]? [<ffffffffc0309d98>] ? dm_make_request+0x128/0x1a0 > [dm_mod] > [11279.498156]? [<ffffffffb991348d>] io_schedule_timeout+0xad/0x130 > [11279.498160]? [<ffffffffb99145ad>] wait_for_completion_io+0xfd/0x140 > [11279.498165]? [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20 > [11279.498169]? [<ffffffffb951e574>] blkdev_issue_flush+0xb4/0x110 > [11279.498202]? [<ffffffffc06064b9>] > xfs_blkdev_issue_flush+0x19/0x20 [xfs] > [11279.498231]? [<ffffffffc05eec40>] xfs_file_fsync+0x1b0/0x1e0 [xfs] > [11279.498238]? [<ffffffffb944f0e7>] do_fsync+0x67/0xb0 > [11279.498242]? [<ffffffffb992076f>] ? > system_call_after_swapgs+0xbc/0x160 > [11279.498246]? [<ffffffffb944f3f3>] SyS_fdatasync+0x13/0x20 > [11279.498250]? [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21 > [11279.498254]? [<ffffffffb992077b>] ? > system_call_after_swapgs+0xc8/0x160 > [11279.498257] INFO: task glusteriotwr1:14950 blocked for more > than 120 seconds. > [11279.499789] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" > disables this message. > [11279.501343] glusteriotwr1? ?D ffff97208b6daf70 ? ?0 14950? ? ? > 1 0x00000080 > [11279.501348] Call Trace: > [11279.501353]? [<ffffffffb9913f79>] schedule+0x29/0x70 > [11279.501390]? [<ffffffffc060e388>] > _xfs_log_force_lsn+0x2e8/0x340 [xfs] > [11279.501396]? [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20 > [11279.501428]? [<ffffffffc05eeb97>] xfs_file_fsync+0x107/0x1e0 [xfs] > [11279.501432]? [<ffffffffb944ef3f>] generic_write_sync+0x4f/0x70 > [11279.501461]? [<ffffffffc05f0545>] > xfs_file_aio_write+0x155/0x1b0 [xfs] > [11279.501466]? [<ffffffffb941a533>] do_sync_write+0x93/0xe0 > [11279.501471]? [<ffffffffb941b010>] vfs_write+0xc0/0x1f0 > [11279.501475]? [<ffffffffb941c002>] SyS_pwrite64+0x92/0xc0 > [11279.501479]? [<ffffffffb992076f>] ? > system_call_after_swapgs+0xbc/0x160 > [11279.501483]? [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21 > [11279.501489]? [<ffffffffb992077b>] ? > system_call_after_swapgs+0xc8/0x160 > [11279.501493] INFO: task glusteriotwr4:14953 blocked for more > than 120 seconds. > [11279.503047] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" > disables this message. > [11279.504630] glusteriotwr4? ?D ffff972499f2bf40 ? ?0 14953? ? ? > 1 0x00000080 > [11279.504635] Call Trace: > [11279.504640]? [<ffffffffb9913f79>] schedule+0x29/0x70 > [11279.504676]? [<ffffffffc060e388>] > _xfs_log_force_lsn+0x2e8/0x340 [xfs] > [11279.504681]? [<ffffffffb92cf1b0>] ? wake_up_state+0x20/0x20 > [11279.504710]? [<ffffffffc05eeb97>] xfs_file_fsync+0x107/0x1e0 [xfs] > [11279.504714]? [<ffffffffb944f0e7>] do_fsync+0x67/0xb0 > [11279.504718]? [<ffffffffb992076f>] ? > system_call_after_swapgs+0xbc/0x160 > [11279.504722]? [<ffffffffb944f3d0>] SyS_fsync+0x10/0x20 > [11279.504725]? [<ffffffffb992082f>] system_call_fastpath+0x1c/0x21 > [11279.504730]? [<ffffffffb992077b>] ? > system_call_after_swapgs+0xc8/0x160 > [12127.466494] perf: interrupt took too long (8263 > 8150), > lowering kernel.perf_event_max_sample_rate to 24000 > > -------------------- > I think this is the cause of the massive ovirt performance issues > irrespective of gluster volume.? At the time this happened, I was > also ssh'ed into the host, and was doing some rpm querry > commands.? I had just run rpm -qa |grep glusterfs (to verify what > version was actually installed), and that command took almost 2 > minutes to return!? Normally it takes less than 2 seconds.? That > is all pure local SSD IO, too.... > > I'm no expert, but its my understanding that anytime a software > causes these kinds of issues, its a serious bug in the software, > even if its mis-handled exceptions.? Is this correct? > > --Jim > > On Tue, May 29, 2018 at 3:01 PM, Jim Kusznir <jim at palousetech.com > <mailto:jim at palousetech.com>> wrote: > > I think this is the profile information for one of the volumes > that lives on the SSDs and is fully operational with no > down/problem disks: > > [root at ovirt2 yum.repos.d]# gluster volume profile data info > Brick: ovirt2.nwfiber.com:/gluster/brick2/data > ---------------------------------------------- > Cumulative Stats: > ? ?Block Size:? ? ? ? ? ? ? ? 256b+ ? ? ? ? ? ?512b+? ? ? ? ? > ? ? ? 1024b+ > ?No. of Reads:? ? ? ? ? ? ? ? ? 983 ? ? ? ? ? ? 2696? ? ? ? ? > ? ? ? ? 1059 > No. of Writes:? ? ? ? ? ? ? ? ? ? 0 ? ? ? ? ? ? 1113? ? ? ? ? > ? ? ? ? ?302 > ? ?Block Size:? ? ? ? ? ? ? ?2048b+ ? ? ? ? ? 4096b+? ? ? ? ? > ? ? ? 8192b+ > ?No. of Reads:? ? ? ? ? ? ? ? ? 852 ? ? ? ? ? ?88608? ? ? ? ? > ? ? ? ?53526 > No. of Writes:? ? ? ? ? ? ? ? ? 522 ? ? ? ? ? 812340? ? ? ? ? > ? ? ? ?76257 > ? ?Block Size:? ? ? ? ? ? ? 16384b+ ? ? ? ? ?32768b+? ? ? ? ? > ? ? ?65536b+ > ?No. of Reads:? ? ? ? ? ? ? ? 54351 ? ? ? ? ? 241901? ? ? ? ? > ? ? ? ?15024 > No. of Writes:? ? ? ? ? ? ? ? 21636 ? ? ? ? ? ? 8656? ? ? ? ? > ? ? ? ? 8976 > ? ?Block Size:? ? ? ? ? ? ?131072b+ > ?No. of Reads:? ? ? ? ? ? ? ?524156 > No. of Writes:? ? ? ? ? ? ? ?296071 > ?%-latency? ?Avg-latency? ?Min-Latency ?Max-Latency? ?No. of > calls? ? ? ? ?Fop > ?---------? ?-----------? ?----------- ?-----------? > ?------------? ? ? ? ---- > ? ? ? 0.00? ? ? ?0.00 us? ? ? ?0.00 us ? ?0.00 us? ? ? ? ? > ?4189? ? ?RELEASE > ? ? ? 0.00? ? ? ?0.00 us? ? ? ?0.00 us ? ?0.00 us? ? ? ? ? > ?1257? RELEASEDIR > ? ? ? 0.00? ? ? 46.19 us? ? ? 12.00 us ?187.00 us? ? ? ? ? ? > ?69? ? ? ?FLUSH > ? ? ? 0.00? ? ?147.00 us? ? ? 78.00 us ?367.00 us? ? ? ? ? ? > ?86 REMOVEXATTR > ? ? ? 0.00? ? ?223.46 us? ? ? 24.00 us 1166.00 us? ? ? ? ? ? > 149? ? ?READDIR > ? ? ? 0.00? ? ?565.34 us? ? ? 76.00 us 3639.00 us? ? ? ? ? ? > ?88? ?FTRUNCATE > ? ? ? 0.00? ? ?263.28 us? ? ? 20.00 us ?28385.00 us? ? ? ? ? ? > 228? ? ? ? ? LK > ? ? ? 0.00? ? ? 98.84 us? ? ? ?2.00 us ?880.00 us? ? ? ? ? > ?1198? ? ?OPENDIR > ? ? ? 0.00? ? ? 91.59 us? ? ? 26.00 us ?10371.00 us? ? ? ? ? > ?3853? ? ? STATFS > ? ? ? 0.00? ? ?494.14 us? ? ? 17.00 us 193439.00 us? ? ? ? ? > ?1171? ? GETXATTR > ? ? ? 0.00? ? ?299.42 us? ? ? 35.00 us 9799.00 us? ? ? ? ? > ?2044? ? READDIRP > ? ? ? 0.00? ? 1965.31 us? ? ?110.00 us 382258.00 us? ? ? ? ? ? > 321? ? ?XATTROP > ? ? ? 0.01? ? ?113.40 us? ? ? 24.00 us ?61061.00 us? ? ? ? ? > ?8134? ? ? ? STAT > ? ? ? 0.01? ? ?755.38 us? ? ? 57.00 us 607603.00 us? ? ? ? ? > ?3196? ? ?DISCARD > ? ? ? 0.05? ? 2690.09 us? ? ? 58.00 us 2704761.00 us? ? ? ? ? > ?3206? ? ? ? OPEN > ? ? ? 0.10? 119978.25 us? ? ? 97.00 us 9406684.00 us? ? ? ? ? > ? 154? ? ?SETATTR > ? ? ? 0.18? ? ?101.73 us? ? ? 28.00 us 700477.00 us? ? ? ? > ?313379? ? ? ?FSTAT > ? ? ? 0.23? ? 1059.84 us? ? ? 25.00 us 2716124.00 us? ? ? ? ? > 38255? ? ? LOOKUP > ? ? ? 0.47? ? 1024.11 us? ? ? 54.00 us 6197164.00 us? ? ? ? ? > 81455? ? FXATTROP > ? ? ? 1.72? ? 2984.00 us? ? ? 15.00 us 37098954.00 us? ? ? ? > ?103020? ? FINODELK > ? ? ? 5.92? ?44315.32 us? ? ? 51.00 us 24731536.00 us? ? ? ? ? > 23957? ? ? ?FSYNC > ? ? ?13.27? ? 2399.78 us? ? ? 25.00 us 22089540.00 us? ? ? ? > ?991005? ? ? ? READ > ? ? ?37.00? ? 5980.43 us? ? ? 52.00 us 22099889.00 us? ? ? ? > 1108976? ? ? ?WRITE > ? ? ?41.04? ? 5452.75 us? ? ? 13.00 us 22102452.00 us? ? ? ? > 1349053? ? ?INODELK > ? ? Duration: 10026 seconds > ? ?Data Read: 80046027759 bytes > Data Written: 44496632320 bytes > Interval 1 Stats: > ? ?Block Size:? ? ? ? ? ? ? ? 256b+ ? ? ? ? ? ?512b+? ? ? ? ? > ? ? ? 1024b+ > ?No. of Reads:? ? ? ? ? ? ? ? ? 983 ? ? ? ? ? ? 2696? ? ? ? ? > ? ? ? ? 1059 > No. of Writes:? ? ? ? ? ? ? ? ? ? 0 ? ? ? ? ? ? ?838? ? ? ? ? > ? ? ? ? ?185 > ? ?Block Size:? ? ? ? ? ? ? ?2048b+ ? ? ? ? ? 4096b+? ? ? ? ? > ? ? ? 8192b+ > ?No. of Reads:? ? ? ? ? ? ? ? ? 852 ? ? ? ? ? ?85856? ? ? ? ? > ? ? ? ?51575 > No. of Writes:? ? ? ? ? ? ? ? ? 382 ? ? ? ? ? 705802? ? ? ? ? > ? ? ? ?57812 > ? ?Block Size:? ? ? ? ? ? ? 16384b+ ? ? ? ? ?32768b+? ? ? ? ? > ? ? ?65536b+ > ?No. of Reads:? ? ? ? ? ? ? ? 52673 ? ? ? ? ? 232093? ? ? ? ? > ? ? ? ?14984 > No. of Writes:? ? ? ? ? ? ? ? 13499 ? ? ? ? ? ? 4908? ? ? ? ? > ? ? ? ? 4242 > ? ?Block Size:? ? ? ? ? ? ?131072b+ > ?No. of Reads:? ? ? ? ? ? ? ?460040 > No. of Writes:? ? ? ? ? ? ? ? ?6411 > ?%-latency? ?Avg-latency? ?Min-Latency ?Max-Latency? ?No. of > calls? ? ? ? ?Fop > ?---------? ?-----------? ?----------- ?-----------? > ?------------? ? ? ? ---- > ? ? ? 0.00? ? ? ?0.00 us? ? ? ?0.00 us ? ?0.00 us? ? ? ? ? > ?2093? ? ?RELEASE > ? ? ? 0.00? ? ? ?0.00 us? ? ? ?0.00 us ? ?0.00 us? ? ? ? ? > ?1093? RELEASEDIR > ? ? ? 0.00? ? ? 53.38 us? ? ? 26.00 us ?111.00 us? ? ? ? ? ? > ?16? ? ? ?FLUSH > ? ? ? 0.00? ? ?145.14 us? ? ? 78.00 us ?367.00 us? ? ? ? ? ? > ?71 REMOVEXATTR > ? ? ? 0.00? ? ?190.96 us? ? ?114.00 us ?298.00 us? ? ? ? ? ? > ?71? ? ?SETATTR > ? ? ? 0.00? ? ?213.38 us? ? ? 24.00 us 1145.00 us? ? ? ? ? ? > ?90? ? ?READDIR > ? ? ? 0.00? ? ?263.28 us? ? ? 20.00 us ?28385.00 us? ? ? ? ? ? > 228? ? ? ? ? LK > ? ? ? 0.00? ? ?101.76 us? ? ? ?2.00 us ?880.00 us? ? ? ? ? > ?1093? ? ?OPENDIR > ? ? ? 0.01? ? ? 93.60 us? ? ? 27.00 us ?10371.00 us? ? ? ? ? > ?3090? ? ? STATFS > ? ? ? 0.02? ? ?537.47 us? ? ? 17.00 us 193439.00 us? ? ? ? ? > ?1038? ? GETXATTR > ? ? ? 0.03? ? ?297.44 us? ? ? 35.00 us 9799.00 us? ? ? ? ? > ?1990? ? READDIRP > ? ? ? 0.03? ? 2357.28 us? ? ?110.00 us 382258.00 us? ? ? ? ? ? > 253? ? ?XATTROP > ? ? ? 0.04? ? ?385.93 us? ? ? 58.00 us ?47593.00 us? ? ? ? ? > ?2091? ? ? ? OPEN > ? ? ? 0.04? ? ?114.86 us? ? ? 24.00 us ?61061.00 us? ? ? ? ? > ?7715? ? ? ? STAT > ? ? ? 0.06? ? ?444.59 us? ? ? 57.00 us 333240.00 us? ? ? ? ? > ?3053? ? ?DISCARD > ? ? ? 0.42? ? ?316.24 us? ? ? 25.00 us 290728.00 us? ? ? ? ? > 29823? ? ? LOOKUP > ? ? ? 0.73? ? ?257.92 us? ? ? 54.00 us 344812.00 us? ? ? ? ? > 63296? ? FXATTROP > ? ? ? 1.37? ? ? 98.30 us? ? ? 28.00 us ?67621.00 us? ? ? ? > ?313172? ? ? ?FSTAT > ? ? ? 1.58? ? 2124.69 us? ? ? 51.00 us 849200.00 us? ? ? ? ? > 16717? ? ? ?FSYNC > ? ? ? 5.73? ? ?162.46 us? ? ? 52.00 us 748492.00 us? ? ? ? > ?794079? ? ? ?WRITE > ? ? ? 7.19? ? 2065.17 us? ? ? 16.00 us 37098954.00 us? ? ? ? ? > 78381? ? FINODELK > ? ? ?36.44? ? ?886.32 us? ? ? 25.00 us 2216436.00 us? ? ? ? > ?925421? ? ? ? READ > ? ? ?46.30? ? 1178.04 us? ? ? 13.00 us 1700704.00 us? ? ? ? > ?884635? ? ?INODELK > ? ? Duration: 7485 seconds > ? ?Data Read: 71250527215 bytes > Data Written: 5119903744 bytes > Brick: ovirt3.nwfiber.com:/gluster/brick2/data > ---------------------------------------------- > Cumulative Stats: > ? ?Block Size:? ? ? ? ? ? ? ? ? 1b+ > ?No. of Reads:? ? ? ? ? ? ? ? ? ? 0 > No. of Writes:? ? ? ? ? ? ? 3264419 > ?%-latency? ?Avg-latency? ?Min-Latency ?Max-Latency? ?No. of > calls? ? ? ? ?Fop > ?---------? ?-----------? ?----------- ?-----------? > ?------------? ? ? ? ---- > ? ? ? 0.00? ? ? ?0.00 us? ? ? ?0.00 us ? ?0.00 us? ? ? ? ? ? > ?90? ? ? FORGET > ? ? ? 0.00? ? ? ?0.00 us? ? ? ?0.00 us ? ?0.00 us? ? ? ? ? > ?9462? ? ?RELEASE > ? ? ? 0.00? ? ? ?0.00 us? ? ? ?0.00 us ? ?0.00 us? ? ? ? ? > ?4254? RELEASEDIR > ? ? ? 0.00? ? ? 50.52 us? ? ? 13.00 us ?190.00 us? ? ? ? ? ? > ?71? ? ? ?FLUSH > ? ? ? 0.00? ? ?186.97 us? ? ? 87.00 us ?713.00 us? ? ? ? ? ? > ?86 REMOVEXATTR > ? ? ? 0.00? ? ? 79.32 us? ? ? 33.00 us ?189.00 us? ? ? ? ? ? > 228? ? ? ? ? LK > ? ? ? 0.00? ? ?220.98 us? ? ?129.00 us ?513.00 us? ? ? ? ? ? > ?86? ? ?SETATTR > ? ? ? 0.01? ? ?259.30 us? ? ? 26.00 us 2632.00 us? ? ? ? ? ? > 137? ? ?READDIR > ? ? ? 0.02? ? ?322.76 us? ? ?145.00 us 2125.00 us? ? ? ? ? ? > 321? ? ?XATTROP > ? ? ? 0.03? ? ?109.55 us? ? ? ?2.00 us 1258.00 us? ? ? ? ? > ?1193? ? ?OPENDIR > ? ? ? 0.05? ? ? 70.21 us? ? ? 21.00 us ?431.00 us? ? ? ? ? > ?3196? ? ?DISCARD > ? ? ? 0.05? ? ?169.26 us? ? ? 21.00 us 2315.00 us? ? ? ? ? > ?1545? ? GETXATTR > ? ? ? 0.12? ? ?176.85 us? ? ? 63.00 us 2844.00 us? ? ? ? ? > ?3206? ? ? ? OPEN > ? ? ? 0.61? ? ?303.49 us? ? ? 90.00 us 3085.00 us? ? ? ? ? > ?9633? ? ? ?FSTAT > ? ? ? 2.44? ? ?305.66 us? ? ? 28.00 us 3716.00 us? ? ? ? ? > 38230? ? ? LOOKUP > ? ? ? 4.52? ? ?266.22 us? ? ? 55.00 us ?53424.00 us? ? ? ? ? > 81455? ? FXATTROP > ? ? ? 6.96? ? 1397.99 us? ? ? 51.00 us ?64822.00 us? ? ? ? ? > 23889? ? ? ?FSYNC > ? ? ?16.48? ? ? 84.74 us? ? ? 25.00 us 6917.00 us? ? ? ? > ?932592? ? ? ?WRITE > ? ? ?30.16? ? ?106.90 us? ? ? 13.00 us 3920189.00 us? ? ? ? > 1353046? ? ?INODELK > ? ? ?38.55? ? 1794.52 us? ? ? 14.00 us 16210553.00 us? ? ? ? > ?103039? ? FINODELK > ? ? Duration: 66562 seconds > ? ?Data Read: 0 bytes > Data Written: 3264419 bytes > Interval 1 Stats: > ? ?Block Size:? ? ? ? ? ? ? ? ? 1b+ > ?No. of Reads:? ? ? ? ? ? ? ? ? ? 0 > No. of Writes:? ? ? ? ? ? ? ?794080 > ?%-latency? ?Avg-latency? ?Min-Latency ?Max-Latency? ?No. of > calls? ? ? ? ?Fop > ?---------? ?-----------? ?----------- ?-----------? > ?------------? ? ? ? ---- > ? ? ? 0.00? ? ? ?0.00 us? ? ? ?0.00 us ? ?0.00 us? ? ? ? ? > ?2093? ? ?RELEASE > ? ? ? 0.00? ? ? ?0.00 us? ? ? ?0.00 us ? ?0.00 us? ? ? ? ? > ?1093? RELEASEDIR > ? ? ? 0.00? ? ? 70.31 us? ? ? 26.00 us ?125.00 us? ? ? ? ? ? > ?16? ? ? ?FLUSH > ? ? ? 0.00? ? ?193.10 us? ? ?103.00 us ?713.00 us? ? ? ? ? ? > ?71 REMOVEXATTR > ? ? ? 0.01? ? ?227.32 us? ? ?133.00 us ?513.00 us? ? ? ? ? ? > ?71? ? ?SETATTR > ? ? ? 0.01? ? ? 79.32 us? ? ? 33.00 us ?189.00 us? ? ? ? ? ? > 228? ? ? ? ? LK > ? ? ? 0.01? ? ?259.83 us? ? ? 35.00 us 1138.00 us? ? ? ? ? ? > ?89? ? ?READDIR > ? ? ? 0.03? ? ?318.26 us? ? ?145.00 us 2047.00 us? ? ? ? ? ? > 253? ? ?XATTROP > ? ? ? 0.04? ? ?112.67 us? ? ? ?3.00 us 1258.00 us? ? ? ? ? > ?1093? ? ?OPENDIR > ? ? ? 0.06? ? ?167.98 us? ? ? 23.00 us 1951.00 us? ? ? ? ? > ?1014? ? GETXATTR > ? ? ? 0.08? ? ? 70.97 us? ? ? 22.00 us ?431.00 us? ? ? ? ? > ?3053? ? ?DISCARD > ? ? ? 0.13? ? ?183.78 us? ? ? 66.00 us 2844.00 us? ? ? ? ? > ?2091? ? ? ? OPEN > ? ? ? 1.01? ? ?303.82 us? ? ? 90.00 us 3085.00 us? ? ? ? ? > ?9610? ? ? ?FSTAT > ? ? ? 3.27? ? ?316.59 us? ? ? 30.00 us 3716.00 us? ? ? ? ? > 29820? ? ? LOOKUP > ? ? ? 5.83? ? ?265.79 us? ? ? 59.00 us ?53424.00 us? ? ? ? ? > 63296? ? FXATTROP > ? ? ? 7.95? ? 1373.89 us? ? ? 51.00 us ?64822.00 us? ? ? ? ? > 16717? ? ? ?FSYNC > ? ? ?23.17? ? ?851.99 us? ? ? 14.00 us 16210553.00 us? ? ? ? ? > 78555? ? FINODELK > ? ? ?24.04? ? ? 87.44 us? ? ? 27.00 us 6917.00 us? ? ? ? > ?794081? ? ? ?WRITE > ? ? ?34.36? ? ?111.91 us? ? ? 14.00 us 984871.00 us? ? ? ? > ?886790? ? ?INODELK > ? ? Duration: 7485 seconds > ? ?Data Read: 0 bytes > Data Written: 794080 bytes > > > ----------------------- > Here is the data from the volume that is backed by the SHDDs > and has one failed disk: > [root at ovirt2 yum.repos.d]# gluster volume profile data-hdd info > Brick: 172.172.1.12:/gluster/brick3/data-hdd > -------------------------------------------- > Cumulative Stats: > ? ?Block Size:? ? ? ? ? ? ? ? 256b+ ? ? ? ? ? ?512b+? ? ? ? ? > ? ? ? 1024b+ > ?No. of Reads:? ? ? ? ? ? ? ? ?1702 ? ? ? ? ? ? ? 86? ? ? ? ? > ? ? ? ? ? 16 > No. of Writes:? ? ? ? ? ? ? ? ? ? 0 ? ? ? ? ? ? ?767? ? ? ? ? > ? ? ? ? ? 71 > ? ?Block Size:? ? ? ? ? ? ? ?2048b+ ? ? ? ? ? 4096b+? ? ? ? ? > ? ? ? 8192b+ > ?No. of Reads:? ? ? ? ? ? ? ? ? ?19 ? ? ? ? ? ?51841? ? ? ? ? > ? ? ? ? 2049 > No. of Writes:? ? ? ? ? ? ? ? ? ?76 ? ? ? ? ? ?60668? ? ? ? ? > ? ? ? ?35727 > ? ?Block Size:? ? ? ? ? ? ? 16384b+ ? ? ? ? ?32768b+? ? ? ? ? > ? ? ?65536b+ > ?No. of Reads:? ? ? ? ? ? ? ? ?1744 ? ? ? ? ? ? ?639? ? ? ? ? > ? ? ? ? 1088 > No. of Writes:? ? ? ? ? ? ? ? ?8524 ? ? ? ? ? ? 2410? ? ? ? ? > ? ? ? ? 1285 > ? ?Block Size:? ? ? ? ? ? ?131072b+ > ?No. of Reads:? ? ? ? ? ? ? ?771999 > No. of Writes:? ? ? ? ? ? ? ? 29584 > ?%-latency? ?Avg-latency? ?Min-Latency ?Max-Latency? ?No. of > calls? ? ? ? ?Fop > ?---------? ?-----------? ?----------- ?-----------? > ?------------? ? ? ? ---- > ? ? ? 0.00? ? ? ?0.00 us? ? ? ?0.00 us ? ?0.00 us? ? ? ? ? > ?2902? ? ?RELEASE > ? ? ? 0.00? ? ? ?0.00 us? ? ? ?0.00 us ? ?0.00 us? ? ? ? ? > ?1517? RELEASEDIR > ? ? ? 0.00? ? ?197.00 us? ? ?197.00 us ?197.00 us? ? ? ? ? ? ? > 1? ?FTRUNCATE > ? ? ? 0.00? ? ? 70.24 us? ? ? 16.00 us ?758.00 us? ? ? ? ? ? > ?51? ? ? ?FLUSH > ? ? ? 0.00? ? ?143.93 us? ? ? 82.00 us ?305.00 us? ? ? ? ? ? > ?57 REMOVEXATTR > ? ? ? 0.00? ? ?178.63 us? ? ?105.00 us ?712.00 us? ? ? ? ? ? > ?60? ? ?SETATTR > ? ? ? 0.00? ? ? 67.30 us? ? ? 19.00 us ?572.00 us? ? ? ? ? ? > 555? ? ? ? ? LK > ? ? ? 0.00? ? ?322.80 us? ? ? 23.00 us 4673.00 us? ? ? ? ? ? > 138? ? ?READDIR > ? ? ? 0.00? ? ?336.56 us? ? ?106.00 us ?11994.00 us? ? ? ? ? ? > 237? ? ?XATTROP > ? ? ? 0.00? ? ? 84.70 us? ? ? 28.00 us 1071.00 us? ? ? ? ? > ?3469? ? ? STATFS > ? ? ? 0.01? ? ?387.75 us? ? ? ?2.00 us 146017.00 us? ? ? ? ? > ?1467? ? ?OPENDIR > ? ? ? 0.01? ? ?148.59 us? ? ? 21.00 us ?64374.00 us? ? ? ? ? > ?4454? ? ? ? STAT > ? ? ? 0.02? ? ?783.02 us? ? ? 16.00 us ?93502.00 us? ? ? ? ? > ?1902? ? GETXATTR > ? ? ? 0.03? ? 1516.10 us? ? ? 17.00 us 210690.00 us? ? ? ? ? > ?1364? ? ?ENTRYLK > ? ? ? 0.03? ? 2555.47 us? ? ?300.00 us 674454.00 us? ? ? ? ? > ?1064? ? READDIRP > ? ? ? 0.07? ? ? 85.74 us? ? ? 19.00 us ?68340.00 us? ? ? ? ? > 62849? ? ? ?FSTAT > ? ? ? 0.07? ? 1978.12 us? ? ? 59.00 us 202596.00 us? ? ? ? ? > ?2729? ? ? ? OPEN > ? ? ? 0.22? ? ?708.57 us? ? ? 15.00 us 394799.00 us? ? ? ? ? > 25447? ? ? LOOKUP > ? ? ? 5.94? ? 2331.74 us? ? ? 15.00 us 1099530.00 us? ? ? ? > ?207534? ? FINODELK > ? ? ? 7.31? ? 8311.75 us? ? ? 58.00 us 1800216.00 us? ? ? ? ? > 71668? ? FXATTROP > ? ? ?12.49? ? 7735.19 us? ? ? 51.00 us 3595513.00 us? ? ? ? > ?131642? ? ? ?WRITE > ? ? ?17.70? ? ?957.08 us? ? ? 16.00 us 13700466.00 us? ? ? ? > 1508160? ? ?INODELK > ? ? ?24.55? ? 2546.43 us? ? ? 26.00 us 5077347.00 us? ? ? ? > ?786060? ? ? ? READ > ? ? ?31.56? ?49699.15 us? ? ? 47.00 us 3746331.00 us? ? ? ? ? > 51777? ? ? ?FSYNC > ? ? Duration: 10101 seconds > ? ?Data Read: 101562897361 bytes > Data Written: 4834450432 bytes > Interval 0 Stats: > ? ?Block Size:? ? ? ? ? ? ? ? 256b+ ? ? ? ? ? ?512b+? ? ? ? ? > ? ? ? 1024b+ > ?No. of Reads:? ? ? ? ? ? ? ? ?1702 ? ? ? ? ? ? ? 86? ? ? ? ? > ? ? ? ? ? 16 > No. of Writes:? ? ? ? ? ? ? ? ? ? 0 ? ? ? ? ? ? ?767? ? ? ? ? > ? ? ? ? ? 71 > ? ?Block Size:? ? ? ? ? ? ? ?2048b+ ? ? ? ? ? 4096b+? ? ? ? ? > ? ? ? 8192b+ > ?No. of Reads:? ? ? ? ? ? ? ? ? ?19 ? ? ? ? ? ?51841? ? ? ? ? > ? ? ? ? 2049 > No. of Writes:? ? ? ? ? ? ? ? ? ?76 ? ? ? ? ? ?60668? ? ? ? ? > ? ? ? ?35727 > ? ?Block Size:? ? ? ? ? ? ? 16384b+ ? ? ? ? ?32768b+? ? ? ? ? > ? ? ?65536b+ > ?No. of Reads:? ? ? ? ? ? ? ? ?1744 ? ? ? ? ? ? ?639? ? ? ? ? > ? ? ? ? 1088 > No. of Writes:? ? ? ? ? ? ? ? ?8524 ? ? ? ? ? ? 2410? ? ? ? ? > ? ? ? ? 1285 > ? ?Block Size:? ? ? ? ? ? ?131072b+ > ?No. of Reads:? ? ? ? ? ? ? ?771999 > No. of Writes:? ? ? ? ? ? ? ? 29584 > ?%-latency? ?Avg-latency? ?Min-Latency ?Max-Latency? ?No. of > calls? ? ? ? ?Fop > ?---------? ?-----------? ?----------- ?-----------? > ?------------? ? ? ? ---- > ? ? ? 0.00? ? ? ?0.00 us? ? ? ?0.00 us ? ?0.00 us? ? ? ? ? > ?2902? ? ?RELEASE > ? ? ? 0.00? ? ? ?0.00 us? ? ? ?0.00 us ? ?0.00 us? ? ? ? ? > ?1517? RELEASEDIR > ? ? ? 0.00? ? ?197.00 us? ? ?197.00 us ?197.00 us? ? ? ? ? ? ? > 1? ?FTRUNCATE > ? ? ? 0.00? ? ? 70.24 us? ? ? 16.00 us ?758.00 us? ? ? ? ? ? > ?51? ? ? ?FLUSH > ? ? ? 0.00? ? ?143.93 us? ? ? 82.00 us ?305.00 us? ? ? ? ? ? > ?57 REMOVEXATTR > ? ? ? 0.00? ? ?178.63 us? ? ?105.00 us ?712.00 us? ? ? ? ? ? > ?60? ? ?SETATTR > ? ? ? 0.00? ? ? 67.30 us? ? ? 19.00 us ?572.00 us? ? ? ? ? ? > 555? ? ? ? ? LK > ? ? ? 0.00? ? ?322.80 us? ? ? 23.00 us 4673.00 us? ? ? ? ? ? > 138? ? ?READDIR > ? ? ? 0.00? ? ?336.56 us? ? ?106.00 us ?11994.00 us? ? ? ? ? ? > 237? ? ?XATTROP > ? ? ? 0.00? ? ? 84.70 us? ? ? 28.00 us 1071.00 us? ? ? ? ? > ?3469? ? ? STATFS > ? ? ? 0.01? ? ?387.75 us? ? ? ?2.00 us 146017.00 us? ? ? ? ? > ?1467? ? ?OPENDIR > ? ? ? 0.01? ? ?148.59 us? ? ? 21.00 us ?64374.00 us? ? ? ? ? > ?4454? ? ? ? STAT > ? ? ? 0.02? ? ?783.02 us? ? ? 16.00 us ?93502.00 us? ? ? ? ? > ?1902? ? GETXATTR > ? ? ? 0.03? ? 1516.10 us? ? ? 17.00 us 210690.00 us? ? ? ? ? > ?1364? ? ?ENTRYLK > ? ? ? 0.03? ? 2555.47 us? ? ?300.00 us 674454.00 us? ? ? ? ? > ?1064? ? READDIRP > ? ? ? 0.07? ? ? 85.73 us? ? ? 19.00 us ?68340.00 us? ? ? ? ? > 62849? ? ? ?FSTAT > ? ? ? 0.07? ? 1978.12 us? ? ? 59.00 us 202596.00 us? ? ? ? ? > ?2729? ? ? ? OPEN > ? ? ? 0.22? ? ?708.57 us? ? ? 15.00 us 394799.00 us? ? ? ? ? > 25447? ? ? LOOKUP > ? ? ? 5.94? ? 2334.57 us? ? ? 15.00 us 1099530.00 us? ? ? ? > ?207534? ? FINODELK > ? ? ? 7.31? ? 8311.49 us? ? ? 58.00 us 1800216.00 us? ? ? ? ? > 71668? ? FXATTROP > ? ? ?12.49? ? 7735.32 us? ? ? 51.00 us 3595513.00 us? ? ? ? > ?131642? ? ? ?WRITE > ? ? ?17.71? ? ?957.08 us? ? ? 16.00 us 13700466.00 us? ? ? ? > 1508160? ? ?INODELK > ? ? ?24.56? ? 2546.42 us? ? ? 26.00 us 5077347.00 us? ? ? ? > ?786060? ? ? ? READ > ? ? ?31.54? ?49651.63 us? ? ? 47.00 us 3746331.00 us? ? ? ? ? > 51777? ? ? ?FSYNC > ? ? Duration: 10101 seconds > ? ?Data Read: 101562897361 bytes > Data Written: 4834450432 bytes > > > On Tue, May 29, 2018 at 2:55 PM, Jim Kusznir > <jim at palousetech.com <mailto:jim at palousetech.com>> wrote: > > Thank you for your response. > > I have 4 gluster volumes.? 3 are replica 2 + arbitrator.? > replica bricks are on ovirt1 and ovirt2, arbitrator on > ovirt3.? The 4th volume is replica 3, with a brick on all > three ovirt machines. > > The first 3 volumes are on an SSD disk; the 4th is on a > Seagate SSHD (same in all three machines).? On ovirt3, the > SSHD has reported hard IO failures, and that brick is > offline.? However, the other two replicas are fully > operational (although they still show contents in the heal > info command that won't go away, but that may be the case > until I replace the failed disk). > > What is bothering me is that ALL 4 gluster volumes are > showing horrible performance issues.? At this point, as > the bad disk has been completely offlined, I would expect > gluster to perform at normal speed, but that is definitely > not the case. > > I've also noticed that the performance hits seem to come > in waves: things seem to work acceptably (but slow) for a > while, then suddenly, its as if all disk IO on all volumes > (including non-gluster local OS disk volumes for the > hosts) pause for about 30 seconds, then IO resumes again. > During those times, I start getting VM not responding and > host not responding notices as well as the applications > having major issues. > > I've shut down most of my VMs and am down to just my > essential core VMs (shedded about 75% of my VMs). I still > am experiencing the same issues. > > Am I correct in believing that once the failed disk was > brought offline that performance should return to normal? > > On Tue, May 29, 2018 at 1:27 PM, Alex K > <rightkicktech at gmail.com <mailto:rightkicktech at gmail.com>> > wrote: > > I would check disks status and accessibility of mount > points where your gluster volumes reside. > > On Tue, May 29, 2018, 22:28 Jim Kusznir > <jim at palousetech.com <mailto:jim at palousetech.com>> wrote: > > On one ovirt server, I'm now seeing these messages: > [56474.239725] blk_update_request: 63 callbacks > suppressed > [56474.239732] blk_update_request: I/O error, dev > dm-2, sector 0 > [56474.240602] blk_update_request: I/O error, dev > dm-2, sector 3905945472 > [56474.241346] blk_update_request: I/O error, dev > dm-2, sector 3905945584 > [56474.242236] blk_update_request: I/O error, dev > dm-2, sector 2048 > [56474.243072] blk_update_request: I/O error, dev > dm-2, sector 3905943424 > [56474.243997] blk_update_request: I/O error, dev > dm-2, sector 3905943536 > [56474.247347] blk_update_request: I/O error, dev > dm-2, sector 0 > [56474.248315] blk_update_request: I/O error, dev > dm-2, sector 3905945472 > [56474.249231] blk_update_request: I/O error, dev > dm-2, sector 3905945584 > [56474.250221] blk_update_request: I/O error, dev > dm-2, sector 2048 > > > > > On Tue, May 29, 2018 at 11:59 AM, Jim Kusznir > <jim at palousetech.com <mailto:jim at palousetech.com>> > wrote: > > I see in messages on ovirt3 (my 3rd machine, > the one upgraded to 4.2): > > May 29 11:54:41 ovirt3 ovs-vsctl: > ovs|00001|db_ctl_base|ERR|unix:/var/run/openvswitch/db.sock: > database connection failed (No such file or > directory) > May 29 11:54:51 ovirt3 ovs-vsctl: > ovs|00001|db_ctl_base|ERR|unix:/var/run/openvswitch/db.sock: > database connection failed (No such file or > directory) > May 29 11:55:01 ovirt3 ovs-vsctl: > ovs|00001|db_ctl_base|ERR|unix:/var/run/openvswitch/db.sock: > database connection failed (No such file or > directory) > (appears a lot). > > I also found on the ssh session of that, some > sysv warnings about the backing disk for one > of the gluster volumes (straight replica 3). > The glusterfs process for that disk on that > machine went offline. Its my understanding > that it should continue to work with the other > two machines while I attempt to replace that > disk, right? Attempted writes (touching an > empty file) can take 15 seconds, repeating it > later will be much faster. > > Gluster generates a bunch of different log > files, I don't know what ones you want, or > from which machine(s). > > How do I do "volume profiling"? > > Thanks! > > On Tue, May 29, 2018 at 11:53 AM, Sahina Bose > <sabose at redhat.com <mailto:sabose at redhat.com>> > wrote: > > Do you see errors reported in the mount > logs for the volume? If so, could you > attach the logs? > Any issues with your underlying disks. Can > you also attach output of volume profiling? > > On Wed, May 30, 2018 at 12:13 AM, Jim > Kusznir <jim at palousetech.com > <mailto:jim at palousetech.com>> wrote: > > Ok, things have gotten MUCH worse this > morning.? I'm getting random errors > from VMs, right now, about a third of > my VMs have been paused due to storage > issues, and most of the remaining VMs > are not performing well. > > At this point, I am in full EMERGENCY > mode, as my production services are > now impacted, and I'm getting calls > coming in with problems... > > I'd greatly appreciate help...VMs are > running VERY slowly (when they run), > and they are steadily getting worse.? > I don't know why.? I was seeing CPU > peaks (to 100%) on several VMs, in > perfect sync, for a few minutes at a > time (while the VM became unresponsive > and any VMs I was logged into that > were linux were giving me the CPU > stuck messages in my origional post).? > Is all this storage related? > > I also have two different gluster > volumes for VM storage, and only one > had the issues, but now VMs in both > are being affected at the same time > and same way. > > --Jim > > On Mon, May 28, 2018 at 10:50 PM, > Sahina Bose <sabose at redhat.com > <mailto:sabose at redhat.com>> wrote: > > [Adding gluster-users to look at > the heal issue] > > On Tue, May 29, 2018 at 9:17 AM, > Jim Kusznir <jim at palousetech.com > <mailto:jim at palousetech.com>> wrote: > > Hello: > > I've been having some cluster > and gluster performance issues > lately.? I also found that my > cluster was out of date, and > was trying to apply updates > (hoping to fix some of these), > and discovered the ovirt 4.1 > repos were taken completely > offline.? So, I was forced to > begin an upgrade to 4.2. > According to docs I > found/read, I needed only add > the new repo, do a yum update, > reboot, and be good on my > hosts (did the yum update, the > engine-setup on my hosted > engine). Things seemed to work > relatively well, except for a > gluster sync issue that showed up. > > My cluster is a 3 node > hyperconverged cluster.? I > upgraded the hosted engine > first, then engine 3. When > engine 3 came back up, for > some reason one of my gluster > volumes would not sync. Here's > sample output: > > [root at ovirt3 ~]# gluster > volume heal data-hdd info > Brick > 172.172.1.11:/gluster/brick3/data-hdd > /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/48d7ecb8-7ac5-4725-bca5-b3519681cf2f/0d6080b0-7018-4fa3-bb82-1dd9ef07d9b9 > > /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/647be733-f153-4cdc-85bd-ba72544c2631/b453a300-0602-4be1-8310-8bd5abe00971 > > /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/6da854d1-b6be-446b-9bf0-90a0dbbea830/3c93bd1f-b7fa-4aa2-b445-6904e31839ba > > /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/7f647567-d18c-44f1-a58e-9b8865833acb/f9364470-9770-4bb1-a6b9-a54861849625 > > /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/f3c8e7aa-6ef2-42a7-93d4-e0a4df6dd2fa/2eb0b1ad-2606-44ef-9cd3-ae59610a504b > > /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/b1ea3f62-0f05-4ded-8c82-9c91c90e0b61/d5d6bf5a-499f-431d-9013-5453db93ed32 > > /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/8c8b5147-e9d6-4810-b45b-185e3ed65727/16f08231-93b0-489d-a2fd-687b6bf88eaa > > /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/12924435-b9c2-4aab-ba19-1c1bc31310ef/07b3db69-440e-491e-854c-bbfa18a7cff2 > > Status: Connected > Number of entries: 8 > > Brick > 172.172.1.12:/gluster/brick3/data-hdd > /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/48d7ecb8-7ac5-4725-bca5-b3519681cf2f/0d6080b0-7018-4fa3-bb82-1dd9ef07d9b9 > > /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/647be733-f153-4cdc-85bd-ba72544c2631/b453a300-0602-4be1-8310-8bd5abe00971 > > /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/b1ea3f62-0f05-4ded-8c82-9c91c90e0b61/d5d6bf5a-499f-431d-9013-5453db93ed32 > > /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/6da854d1-b6be-446b-9bf0-90a0dbbea830/3c93bd1f-b7fa-4aa2-b445-6904e31839ba > > /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/7f647567-d18c-44f1-a58e-9b8865833acb/f9364470-9770-4bb1-a6b9-a54861849625 > > /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/8c8b5147-e9d6-4810-b45b-185e3ed65727/16f08231-93b0-489d-a2fd-687b6bf88eaa > > /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/12924435-b9c2-4aab-ba19-1c1bc31310ef/07b3db69-440e-491e-854c-bbfa18a7cff2 > > /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/f3c8e7aa-6ef2-42a7-93d4-e0a4df6dd2fa/2eb0b1ad-2606-44ef-9cd3-ae59610a504b > > Status: Connected > Number of entries: 8 > > Brick > 172.172.1.13:/gluster/brick3/data-hdd > /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/b1ea3f62-0f05-4ded-8c82-9c91c90e0b61/d5d6bf5a-499f-431d-9013-5453db93ed32 > > /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/8c8b5147-e9d6-4810-b45b-185e3ed65727/16f08231-93b0-489d-a2fd-687b6bf88eaa > > /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/12924435-b9c2-4aab-ba19-1c1bc31310ef/07b3db69-440e-491e-854c-bbfa18a7cff2 > > /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/f3c8e7aa-6ef2-42a7-93d4-e0a4df6dd2fa/2eb0b1ad-2606-44ef-9cd3-ae59610a504b > > /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/647be733-f153-4cdc-85bd-ba72544c2631/b453a300-0602-4be1-8310-8bd5abe00971 > > /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/48d7ecb8-7ac5-4725-bca5-b3519681cf2f/0d6080b0-7018-4fa3-bb82-1dd9ef07d9b9 > > /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/6da854d1-b6be-446b-9bf0-90a0dbbea830/3c93bd1f-b7fa-4aa2-b445-6904e31839ba > > /cc65f671-3377-494a-a7d4-1d9f7c3ae46c/images/7f647567-d18c-44f1-a58e-9b8865833acb/f9364470-9770-4bb1-a6b9-a54861849625 > > Status: Connected > Number of entries: 8 > > --------- > Its been in this state for a > couple days now, and bandwidth > monitoring shows no > appreciable data moving. I've > tried repeatedly commanding a > full heal from all three > clusters in the node.? Its > always the same files that > need healing. > > When running gluster volume > heal data-hdd statistics, I > see sometimes different > information, but always some > number of "heal failed" > entries.? It shows 0 for split > brain. > > I'm not quite sure what to > do.? I suspect it may be due > to nodes 1 and 2 still being > on the older ovirt/gluster > release, but I'm afraid to > upgrade and reboot them until > I have a good gluster sync > (don't need to create a split > brain issue).? How do I > proceed with this? > > Second issue: I've been > experiencing VERY POOR > performance on most of my > VMs.? To the tune that logging > into a windows 10 vm via > remote desktop can take 5 > minutes, launching quickbooks > inside said vm can easily take > 10 minutes.? On some linux > VMs, I get random messages > like this: > Message from syslogd at unifi at > May 28 20:39:23 ... > ?kernel:[6171996.308904] NMI > watchdog: BUG: soft lockup - > CPU#0 stuck for 22s! > [mongod:14766] > > (the process and PID are often > different) > > I'm not quite sure what to do > about this either.? My initial > thought was upgrad everything > to current and see if its > still there, but I cannot move > forward with that until my > gluster is healed... > > Thanks! > --Jim > > _______________________________________________ > Users mailing list -- > users at ovirt.org > <mailto:users at ovirt.org> > To unsubscribe send an email > to users-leave at ovirt.org > <mailto:users-leave at ovirt.org> > Privacy Statement: > https://www.ovirt.org/site/privacy-policy/ > <https://www.ovirt.org/site/privacy-policy/> > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > <https://www.ovirt.org/community/about/community-guidelines/> > List Archives: > https://lists.ovirt.org/archives/list/users at ovirt.org/message/3LEV6ZQ3JV2XLAL7NYBTXOYMYUOTIRQF/ > <https://lists.ovirt.org/archives/list/users at ovirt.org/message/3LEV6ZQ3JV2XLAL7NYBTXOYMYUOTIRQF/> > > > > > > > _______________________________________________ > Users mailing list -- users at ovirt.org > <mailto:users at ovirt.org> > To unsubscribe send an email to > users-leave at ovirt.org <mailto:users-leave at ovirt.org> > Privacy Statement: > https://www.ovirt.org/site/privacy-policy/ > <https://www.ovirt.org/site/privacy-policy/> > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > <https://www.ovirt.org/community/about/community-guidelines/> > List Archives: > https://lists.ovirt.org/archives/list/users at ovirt.org/message/ACO7RFSLBSRBAIONIC2HQ6Z24ZDES5MF/ > <https://lists.ovirt.org/archives/list/users at ovirt.org/message/ACO7RFSLBSRBAIONIC2HQ6Z24ZDES5MF/> > > > > > > _______________________________________________ > Users mailing list -- users at ovirt.org <mailto:users at ovirt.org> > To unsubscribe send an email to users-leave at ovirt.org > <mailto:users-leave at ovirt.org> > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > <https://www.ovirt.org/site/privacy-policy/> > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > <https://www.ovirt.org/community/about/community-guidelines/> > List Archives: > https://lists.ovirt.org/archives/list/users at ovirt.org/message/3DEQQLJM3WHQNZJ7KEMRZVFZ52MTIL74/ > <https://lists.ovirt.org/archives/list/users at ovirt.org/message/3DEQQLJM3WHQNZJ7KEMRZVFZ52MTIL74/> > >-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20180530/0704061f/attachment.html>
Seemingly Similar Threads
- [ovirt-users] Re: Gluster problems, cluster performance issues
- [ovirt-users] Re: Gluster problems, cluster performance issues
- [ovirt-users] Re: Gluster problems, cluster performance issues
- [ovirt-users] Re: Gluster problems, cluster performance issues
- [ovirt-users] Re: Gluster problems, cluster performance issues