On 05/05/2017 08:42 PM, Pat Haley wrote:>
> Hi Pranith,
>
> I presume you are asking for some version of the profile data that
> just shows the dd test (or a repeat of the dd test). If yes, how do I
> extract just that data?
Yes, that is what he is asking for. Just clear the existing profile info
using `gluster volume profile volname clear` and run the dd test once.
Then when you run profile info again, it should just give you the stats
for the dd test.>
> Thanks
>
> Pat
>
>
>
> On 05/05/2017 10:58 AM, Pranith Kumar Karampuri wrote:
>> hi Pat,
>> Let us concentrate on the performance numbers part for now. We
>> will look at the permissions one after this?
>>
>> As per the profile info, only 2.6% of the work-load is writes. There
>> are too many Lookups.
>>
>> Would it be possible to get the data for just the dd test you were
>> doing earlier?
>>
>>
>> On Fri, May 5, 2017 at 8:14 PM, Pat Haley <phaley at mit.edu
>> <mailto:phaley at mit.edu>> wrote:
>>
>>
>> Hi Pranith & Ravi,
>>
>> A couple of quick questions
>>
>> We have profile turned on. Are there specific queries we should
>> make that would help debug our configuration? (The default
>> profile info was previously sent in
>>
http://lists.gluster.org/pipermail/gluster-users/2017-May/030840.html
>>
<http://lists.gluster.org/pipermail/gluster-users/2017-May/030840.html>
>> but I'm not sure if that is what you were looking for.)
>>
>> We also started to do a test on serving gluster over NFS. We
>> rediscovered an issue we previously reported (
>>
http://lists.gluster.org/pipermail/gluster-users/2016-September/028289.html
>>
<http://lists.gluster.org/pipermail/gluster-users/2016-September/028289.html>
>> ) in that the NFS mounted version was ignoring the group write
>> permissions. What specific information would be useful in
>> debugging this?
>>
>> Thanks
>>
>> Pat
>>
>>
>>
>> On 04/14/2017 03:01 AM, Ravishankar N wrote:
>>> On 04/14/2017 12:20 PM, Pranith Kumar Karampuri wrote:
>>>>
>>>>
>>>> On Sat, Apr 8, 2017 at 10:28 AM, Ravishankar N
>>>> <ravishankar at redhat.com <mailto:ravishankar at
redhat.com>> wrote:
>>>>
>>>> Hi Pat,
>>>>
>>>> I'm assuming you are using gluster native (fuse
mount). If
>>>> it helps, you could try mounting it via gluster NFS
(gnfs)
>>>> and then see if there is an improvement in speed. Fuse
>>>> mounts are slower than gnfs mounts but you get the
benefit
>>>> of avoiding a single point of failure. Unlike fuse
mounts,
>>>> if the gluster node containing the gnfs server goes
down,
>>>> all mounts done using that node will fail). For fuse
>>>> mounts, you could try tweaking the write-behind xlator
>>>> settings to see if it helps. See the
>>>> performance.write-behind and
>>>> performance.write-behind-window-size options in
`gluster
>>>> volume set help`. Of course, even for gnfs mounts, you
can
>>>> achieve fail-over by using CTDB.
>>>>
>>>>
>>>> Ravi,
>>>> Do you have any data that suggests fuse mounts are
slower
>>>> than gNFS servers?
>>> I have heard anecdotal evidence time and again on the ML and
>>> IRC, which is why I wanted to compare it with NFS numbers on
his
>>> setup.
>>>>
>>>> Pat,
>>>> I see that I am late to the thread, but do you happen
to
>>>> have "profile info" of the workload?
>>>>
>>>> You can follow
>>>>
https://gluster.readthedocs.io/en/latest/Administrator%20Guide/Monitoring%20Workload/
>>>>
<https://gluster.readthedocs.io/en/latest/Administrator%20Guide/Monitoring%20Workload/>
>>>> to get the information.
>>> Yeah, Let's see if profile info shows up anything
interesting.
>>> -Ravi
>>>>
>>>>
>>>> Thanks,
>>>> Ravi
>>>>
>>>>
>>>> On 04/08/2017 12:07 AM, Pat Haley wrote:
>>>>>
>>>>> Hi,
>>>>>
>>>>> We noticed a dramatic slowness when writing to a
gluster
>>>>> disk when compared to writing to an NFS disk.
Specifically
>>>>> when using dd (data duplicator) to write a 4.3 GB
file of
>>>>> zeros:
>>>>>
>>>>> * on NFS disk (/home): 9.5 Gb/s
>>>>> * on gluster disk (/gdata): 508 Mb/s
>>>>>
>>>>> The gluser disk is 2 bricks joined together, no
>>>>> replication or anything else. The hardware is
(literally)
>>>>> the same:
>>>>>
>>>>> * one server with 70 hard disks and a hardware
RAID card.
>>>>> * 4 disks in a RAID-6 group (the NFS disk)
>>>>> * 32 disks in a RAID-6 group (the max allowed by
the
>>>>> card, /mnt/brick1)
>>>>> * 32 disks in another RAID-6 group (/mnt/brick2)
>>>>> * 2 hot spare
>>>>>
>>>>> Some additional information and more tests results
(after
>>>>> changing the log level):
>>>>>
>>>>> glusterfs 3.7.11 built on Apr 27 2016 14:09:22
>>>>> CentOS release 6.8 (Final)
>>>>> RAID bus controller: LSI Logic / Symbios Logic
MegaRAID
>>>>> SAS-3 3108 [Invader] (rev 02)
>>>>>
>>>>>
>>>>>
>>>>> *Create the file to /gdata (gluster)*
>>>>> [root at mseas-data2 gdata]# dd if=/dev/zero
of=/gdata/zero1
>>>>> bs=1M count=1000
>>>>> 1000+0 records in
>>>>> 1000+0 records out
>>>>> 1048576000 bytes (1.0 GB) copied, 1.91876 s, *546
MB/s*
>>>>>
>>>>> *Create the file to /home (ext4)*
>>>>> [root at mseas-data2 gdata]# dd if=/dev/zero
of=/home/zero1
>>>>> bs=1M count=1000
>>>>> 1000+0 records in
>>>>> 1000+0 records out
>>>>> 1048576000 bytes (1.0 GB) copied, 0.686021 s, *1.5
GB/s -
>>>>> *3 times as fast*
>>>>>
>>>>>
>>>>> Copy from /gdata to /gdata (gluster to gluster)
>>>>> *[root at mseas-data2 gdata]# dd if=/gdata/zero1
of=/gdata/zero2
>>>>> 2048000+0 records in
>>>>> 2048000+0 records out
>>>>> 1048576000 bytes (1.0 GB) copied, 101.052 s, *10.4
MB/s* -
>>>>> realllyyy slooowww
>>>>>
>>>>>
>>>>> *Copy from /gdata to /gdata* *2nd time *(gluster to
gluster)**
>>>>> [root at mseas-data2 gdata]# dd if=/gdata/zero1
of=/gdata/zero2
>>>>> 2048000+0 records in
>>>>> 2048000+0 records out
>>>>> 1048576000 bytes (1.0 GB) copied, 92.4904 s, *11.3
MB/s* -
>>>>> realllyyy slooowww again
>>>>>
>>>>>
>>>>>
>>>>> *Copy from /home to /home (ext4 to ext4)*
>>>>> [root at mseas-data2 gdata]# dd if=/home/zero1
of=/home/zero2
>>>>> 2048000+0 records in
>>>>> 2048000+0 records out
>>>>> 1048576000 bytes (1.0 GB) copied, 3.53263 s, *297
MB/s *30
>>>>> times as fast
>>>>>
>>>>>
>>>>> *Copy from /home to /home (ext4 to ext4)*
>>>>> [root at mseas-data2 gdata]# dd if=/home/zero1
of=/home/zero3
>>>>> 2048000+0 records in
>>>>> 2048000+0 records out
>>>>> 1048576000 bytes (1.0 GB) copied, 4.1737 s, *251
MB/s* -
>>>>> 30 times as fast
>>>>>
>>>>>
>>>>> As a test, can we copy data directly to the xfs
mountpoint
>>>>> (/mnt/brick1) and bypass gluster?
>>>>>
>>>>>
>>>>> Any help you could give us would be appreciated.
>>>>>
>>>>> Thanks
>>>>>
>>>>> --
>>>>>
>>>>>
-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
>>>>> Pat Haley Email:phaley at
mit.edu <mailto:phaley at mit.edu>
>>>>> Center for Ocean Engineering Phone: (617)
253-6824
>>>>> Dept. of Mechanical Engineering Fax: (617)
253-8125
>>>>> MIT, Room 5-213http://web.mit.edu/phaley/www/
>>>>> 77 Massachusetts Avenue
>>>>> Cambridge, MA 02139-4301
>>>>>
>>>>> _______________________________________________
>>>>> Gluster-users mailing list
>>>>> Gluster-users at gluster.org
<mailto:Gluster-users at gluster.org>
>>>>>
http://lists.gluster.org/mailman/listinfo/gluster-users
>>>>>
<http://lists.gluster.org/mailman/listinfo/gluster-users>
>>>>
>>>> _______________________________________________
>>>> Gluster-users mailing list Gluster-users at gluster.org
>>>> <mailto:Gluster-users at gluster.org>
>>>> http://lists.gluster.org/mailman/listinfo/gluster-users
>>>>
<http://lists.gluster.org/mailman/listinfo/gluster-users>
>>>>
>>>> --
>>>> Pranith
>>>
>> --
>>
>> -=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
>> Pat Haley Email:phaley at mit.edu
<mailto:phaley at mit.edu>
>> Center for Ocean Engineering Phone: (617) 253-6824
>> Dept. of Mechanical Engineering Fax: (617) 253-8125
>> MIT, Room 5-213http://web.mit.edu/phaley/www/
>> 77 Massachusetts Avenue
>> Cambridge, MA 02139-4301
>>
>> --
>> Pranith
> --
>
> -=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
> Pat Haley Email:phaley at mit.edu
> Center for Ocean Engineering Phone: (617) 253-6824
> Dept. of Mechanical Engineering Fax: (617) 253-8125
> MIT, Room 5-213http://web.mit.edu/phaley/www/
> 77 Massachusetts Avenue
> Cambridge, MA 02139-4301
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
<http://lists.gluster.org/pipermail/gluster-users/attachments/20170505/2e52a62e/attachment.html>