Tiemen Ruiten
2015-Nov-13  08:56 UTC
[Gluster-users] rsync to gluster mount: self-heal and bad performance
Hello Ernie, list,
No, that's not the case. The volume is mounted through glusterfs-fuse - on
the same server running one of the bricks. The fstab:
# /etc/fstab
# Created by anaconda on Tue Aug 18 18:10:49 2015
#
# Accessible filesystems, by reference, are maintained under '/dev/disk'
# See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info
#
UUID=56778fed-bf3f-435e-8c32-edaa8c707f29 /                       xfs defaults
       0 0
UUID=a44e32ed-cfbe-4ba0-896f-1efff9397ba1 /boot                   xfs defaults
       0 0
UUID=a344d2bc-266d-4905-85b1-fbb7fe927659 swap
swap defaults
       0 0
/dev/vdb1  /data/brick   xfs defaults 1 2
iron2:/lpxassets  /mnt/lpxassets glusterfs _netdev,acl 0 0
On 12 November 2015 at 22:50, Ernie Dunbar <maillist at lightspeed.ca>
wrote:
> Hi Tiemen
>
> It sounds like you're trying to rsync files onto your Gluster server,
> rather than to the Gluster filesystem. You want to copy these files into
> the mounted filesystem (typically on some other system than the Gluster
> servers), because Gluster is designed to handle it that way.
>
> I can't remember the nitty gritty details about why this is, but
I've made
> this mistake before as well. Hope that helps. :)
>
>
> On 2015-11-12 11:31, Tiemen Ruiten wrote:
>
>> Hello,
>>
>> While rsyncing to a directory mounted through glusterfs fuse,
>> performance is very bad and it appears every synced file generates a
>> (metadata) self-heal.
>>
>> The volume is mounted with option acl and acl's are set on a
>> subdirectory.
>>
>> Setup is as follows:
>>
>> Two Centos 7 VM's (KVM), with Gluster 3.7.6 and one physical CentOS
6
>> node, also Gluster 3.7.6. Physical node functions as arbiter. So
it's
>> a replica 3 arbiter 1 volume. The bricks are LVM's with XFS
>> filesystem.
>>
>> While I don't think I should expect top performance for rsync on
>> Gluster, I wouldn't expect every file synced to trigger a
self-heal.
>> Anything I can do to improve this? Should I file a bug?
>>
>> Another thing that looks related, I see a lot of these messages,
>> especially when doing IO:
>>
>> [2015-11-12 19:25:42.185904] I [dict.c:473:dict_get]
>>
>>
(-->/usr/lib64/glusterfs/3.7.6/xlator/debug/io-stats.so(io_stats_lookup_cbk+0x121)
>> [0x7fdcc2d31161]
>>
>>
-->/usr/lib64/glusterfs/3.7.6/xlator/system/posix-acl.so(posix_acl_lookup_cbk+0x242)
>> [0x7fdcc2b1b212] -->/lib64/libglusterfs.so.0(dict_get+0xac)
>> [0x7fdcd5e770cc] ) 0-dict: !this || key=system.posix_acl_default
>> [Invalid argument]
>>
>> --
>>
>> Tiemen Ruiten
>> Systems Engineer
>> R&D Media
>>
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> http://www.gluster.org/mailman/listinfo/gluster-users
>>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
>
-- 
Tiemen Ruiten
Systems Engineer
R&D Media
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
<http://www.gluster.org/pipermail/gluster-users/attachments/20151113/77e487d1/attachment.html>
Tiemen Ruiten
2015-Nov-15  10:22 UTC
[Gluster-users] rsync to gluster mount: self-heal and bad performance
Any other suggestions? On 13 November 2015 at 09:56, Tiemen Ruiten <t.ruiten at rdmedia.com> wrote:> Hello Ernie, list, > > No, that's not the case. The volume is mounted through glusterfs-fuse - on > the same server running one of the bricks. The fstab: > > # /etc/fstab > # Created by anaconda on Tue Aug 18 18:10:49 2015 > # > # Accessible filesystems, by reference, are maintained under '/dev/disk' > # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info > # > UUID=56778fed-bf3f-435e-8c32-edaa8c707f29 / xfs defaults > 0 0 > UUID=a44e32ed-cfbe-4ba0-896f-1efff9397ba1 /boot xfs defaults > 0 0 > UUID=a344d2bc-266d-4905-85b1-fbb7fe927659 swap swap defaults > 0 0 > /dev/vdb1 /data/brick xfs defaults 1 2 > iron2:/lpxassets /mnt/lpxassets glusterfs _netdev,acl 0 0 > > > > > On 12 November 2015 at 22:50, Ernie Dunbar <maillist at lightspeed.ca> wrote: > >> Hi Tiemen >> >> It sounds like you're trying to rsync files onto your Gluster server, >> rather than to the Gluster filesystem. You want to copy these files into >> the mounted filesystem (typically on some other system than the Gluster >> servers), because Gluster is designed to handle it that way. >> >> I can't remember the nitty gritty details about why this is, but I've >> made this mistake before as well. Hope that helps. :) >> >> >> On 2015-11-12 11:31, Tiemen Ruiten wrote: >> >>> Hello, >>> >>> While rsyncing to a directory mounted through glusterfs fuse, >>> performance is very bad and it appears every synced file generates a >>> (metadata) self-heal. >>> >>> The volume is mounted with option acl and acl's are set on a >>> subdirectory. >>> >>> Setup is as follows: >>> >>> Two Centos 7 VM's (KVM), with Gluster 3.7.6 and one physical CentOS 6 >>> node, also Gluster 3.7.6. Physical node functions as arbiter. So it's >>> a replica 3 arbiter 1 volume. The bricks are LVM's with XFS >>> filesystem. >>> >>> While I don't think I should expect top performance for rsync on >>> Gluster, I wouldn't expect every file synced to trigger a self-heal. >>> Anything I can do to improve this? Should I file a bug? >>> >>> Another thing that looks related, I see a lot of these messages, >>> especially when doing IO: >>> >>> [2015-11-12 19:25:42.185904] I [dict.c:473:dict_get] >>> >>> (-->/usr/lib64/glusterfs/3.7.6/xlator/debug/io-stats.so(io_stats_lookup_cbk+0x121) >>> [0x7fdcc2d31161] >>> >>> -->/usr/lib64/glusterfs/3.7.6/xlator/system/posix-acl.so(posix_acl_lookup_cbk+0x242) >>> [0x7fdcc2b1b212] -->/lib64/libglusterfs.so.0(dict_get+0xac) >>> [0x7fdcd5e770cc] ) 0-dict: !this || key=system.posix_acl_default >>> [Invalid argument] >>> >>> -- >>> >>> Tiemen Ruiten >>> Systems Engineer >>> R&D Media >>> >>> _______________________________________________ >>> Gluster-users mailing list >>> Gluster-users at gluster.org >>> http://www.gluster.org/mailman/listinfo/gluster-users >>> >> >> _______________________________________________ >> Gluster-users mailing list >> Gluster-users at gluster.org >> http://www.gluster.org/mailman/listinfo/gluster-users >> > > > > -- > Tiemen Ruiten > Systems Engineer > R&D Media >-- Tiemen Ruiten Systems Engineer R&D Media -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20151115/cb18e045/attachment.html>