doing the benchmark file size rage 128KB to 1GB .So up to file size 256MB I am getting buffer cache performance and file size 512MB ,1GB I am getting with in link speed .But in case of GlusterFS I not able to understand what is happening . Please any one can help me . NFS : iozone -Raceb ./perffinal.wks -y 4K -q 128K -n 128K -g 1G -i 0 -i 1 Reader 4 8 16 32 64 128 128 744701 727625 935039 633768 499971 391433 256 920892 1085148 1057519 931149 551834 380335 512 937558 1075517 1100810 904515 558917 368605 1024 974395 1072149 1094105 969724 555319 379390 2048 1026059 1125318 1137073 1005356 568252 375232 4096 1021220 1144780 1169589 1030467 578615 376367 8192 965366 1153315 1071693 1072681 607040 371771 16384 1008989 1133837 1163806 1046171 600500 376056 32768 1022692 1165701 1175739 1065870 630626 363563 65536 1005490 1152909 1168181 1048258 631148 374343 131072 1011405 1161491 1176534 1048509 637910 375741 262144 1011217 1130486 1118877 1075740 636433 375511 524288 9563 9562 9568 9551 9525 9562 1048576 9499 9520 9513 9535 9493 9469 GlusterFS: iozone -Raceb /root/glusterfs/perfgfs2.wks -y 4K -q 128K -n 128K -g 1G -i 0 -i 1 Reader Report 4 8 16 32 64 128 128 48834 50395 49785 48593 48450 47959 256 15276 15209 15210 15100 14998 14973 512 12343 12333 12340 12291 12202 12213 1024 11330 11334 11327 11303 11276 11283 2048 10875 10881 10877 10873 10857 10865 4096 10671 10670 9706 10673 9685 10640 8192 10572 10060 10571 10573 10555 10064 16384 10522 10523 10523 10522 10522 10263 32768 10494 10497 10495 10493 10497 10497 65536 10484 10483 10419 10483 10485 10485 131072 10419 10475 10477 10445 10445 10478 262144 10323 10241 10312 10226 10320 10237 524288 10074 9966 9707 8567 8213 9046 1048576 7440 7973 5737 7101 7678 5743 Any idea for this higher value in NFS test .some this is different . But I am not able to understand. Thanks for your time Mohan ------=_Part_259374_19852034.1230880209611 Content-Type: text/html; charset=ISO-8859-1 Content-Transfer-Encoding: quoted-printable Content-Disposition: inline we're conducting performance benchmark runs to evaluate Linux performance as NFS file servers.<br>It is observed that an unusual high percentage of benchmark time was spent in "read" operation.<br>A sampled workload consisting of 18% of read consumes 63% of total benchmark time. Did this<br> problem get analyzed before (or even better :)-is there a patch) ? We're on 2.4.19 kernel- NFS<br>V3 - UDP, with EXT3 as local file system.<br><br>Thanks in advance.<br><br><a href=3D"mailto:gluster-users at gluster.org">gluster-users at gluster.org</a><br> <br>Dear All,<br><br>we are currently using NFS to meet data sharing requirements.Now we are facing some performance and scalability problem ,so this form does not meet the requirements of our network(performance).So we are finding the possible solutions to increase the performance and scalability .To give very strong solution to NFS issue I have analysed two File System one is GlusterFS and another one is Red Hat GFS.we conclude that GlusterFS will increase the performance and scalability ,It has all the features we are looking .For the testing purpose I am benchmarking NFS and GlusterFS to get better performance .My benchmark result shows that GlusterFS give better performance ,but i am getting some unacceptable read performance . I am not able to understand how exactly the read operation performs NFS and GlusterFS .even I don't know anything i am doing wrong.here i am showing the benchmark result to get better idea of my read performance issuee .i have attached the result of NFS and GlusterFS read values .any one can please go thro this and give me some valuable guide .It will make my benchmarking very effective . <br> <br>This my server and client Hardware and software :<br><br>HARDWARE CONFIG:<br><br>Processor core speed : Intel(R) Celeron(R) CPU 1.70GHz<br><br>Number of cores : Single Core (not dual-core)<br><br>RAM size : 384MB(128MB+256MB)<br> <br>RAM type : DDR<br><br>RAM Speed : 266 MHz (3.8 ns)<br><br>Swap : 1027MB<br><br>Storage controller : ATA device<br><br>Disk model/size : SAMSUNG SV4012H /40 GB,2 MB Cache,<br><br>Storage speed : 52.4 MB/sec<br><br> Spindle Speed : 5400 rpm(Revolution per Minute)<br><br>NIC Type : VIA Rhine III chipset IRQ 18<br><br>NIC Speed : 100 Mbps/Full-Duplex Card<br><br>SOFTWARE:<br><br>Operation System : Fedora Core 9 GNU/Linux<br><br>Linux version : 2.6.9-42<br> <br>Local FS : Ext3<br><br>NFS version : 1.1.2<br><br>GlusterFS version: glusterfs 1.3.8 built on Feb 3 2008<br><br>Iozone : iozone-3-5.fc9.i386 (File System Benchmark Tool)<br><br>ttcp : ttcp-1.12-18.fc9.i386(RAW throughput measurement Tool)<br> <br>This is the server and client vol files i am using the benchmarking<br><br>#GlusterFS Server Volume Specification<br><br>volume brick<br> type storage/posix # POSIX FS translator<br> option directory /bench # /bench dir contains 25,000 files with size 10 KB 15KB<br> end-volume<br><br>volume iot<br> type performance/io-threads<br> option thread-count 4 <br> option cache-size 8MB<br> subvolumes brick<br>end-volume<br><br>volume server<br> type protocol/server<br> option transport-type tcp/server <br> subvolumes iot<br> option auth.ip.brick.allow * # Allow access to "brick" volume<br>end-volume<br><br><br><br># GlusterFS Client Volume Specification <br><br>volume client<br> type protocol/client<br> option transport-type tcp/client <br> option remote-host 192.xxx.x.xxx <br> option remote-subvolume brick <br>end-volume<br><br>volume readahead<br> type performance/read-ahead<br> option page-size 128KB # 256KB is the default option<br> option page-count 4 # cache per file =3D (page-count x page-size) 2 is default option<br> subvolumes client<br>end-volume<br><br>volume iocache<br> type performance/io-cache<br> #option page-size 128KB ## default is 32MB<br> option cache-size 256MB #128KB is default option<br> option page-count 4 <br> subvolumes readahead<br>end-volume<br><br>volume writeback<br> type performance/write-behind<br> option aggregate-size 128KB<br> option flush-behind on<br> subvolumes iocache <br>end-volume<br><br><br>I am confusing this result .I don't have idea how to trace and get good comparable result is read performance .I think I am miss understanding the buffer cache concepts .<br> <br>From attached NFS read result , I understand that I have 348MB RAM and I am doing the benchmark file size rage 128KB to 1GB .So up to file size 256MB I am getting buffer cache performance and file size 512MB ,1GB I am getting with in link speed .But in case of GlusterFS I not able to understand what is happening . <br> <br>Please any one can help me .<br><br>NFS :<br>iozone -Raceb ./perffinal.wks -y 4K -q 128K -n 128K -g 1G -i 0 -i 1 <br><br>Reader <br> 4 8 16 32 64 128<br> <br>128 744701 727625 935039 633768 499971 391433<br>256 920892 1085148 1057519 931149 551834 380335<br>512 937558 1075517 1100810 904515 558917 368605<br>1024 974395 1072149 1094105 969724 555319 379390<br> 2048 1026059 1125318 1137073 1005356 568252 375232<br>4096 1021220 1144780 1169589 1030467 578615 376367<br>8192 965366 1153315 1071693 1072681 607040 371771<br>16384 1008989 1133837 1163806 1046171 600500 376056<br> 32768 1022692 1165701 1175739 1065870 630626 363563<br>65536 1005490 1152909 1168181 1048258 631148 374343<br>131072 1011405 1161491 1176534 1048509 637910 375741<br>262144 1011217 1130486 1118877 1075740 636433 375511<br> 524288 9563 9562 9568 9551 9525 9562<br>1048576 9499 9520 9513 9535 9493 9469<br><br>GlusterFS:<br>iozone -Raceb /root/glusterfs/perfgfs2.wks -y 4K -q 128K -n 128K -g 1G -i 0 -i 1 <br> Reader Report <br> 4 8 16 32 64 128<br><br>128 48834 50395 49785 48593 48450 47959<br>256 15276 15209 15210 15100 14998 14973<br> 512 12343 12333 12340 12291 12202 12213<br>1024 11330 11334 11327 11303 11276 11283<br>2048 10875 10881 10877 10873 10857 10865<br>4096 10671 10670 9706 10673 9685 10640<br> 8192 10572 10060 10571 10573 10555 10064<br>16384 10522 10523 10523 10522 10522 10263<br>32768 10494 10497 10495 10493 10497 10497<br>65536 10484 10483 10419 10483 10485 10485<br> 131072 10419 10475 10477 10445 10445 10478<br>262144 10323 10241 10312 10226 10320 10237<br>524288 10074 9966 9707 8567 8213 9046<br>1048576 7440 7973 5737 7101 7678 5743<br> <br>Any idea for this higher value in NFS test .some this is different . But I am not able to understand.<br><br><br><br><br><br>Thanks for your time<br>Mohan<br><br> ------=_Part_259374_19852034.1230880209611--