doing the benchmark file size rage 128KB to 1GB .So up to file size 256MB I am getting buffer cache performance and file size 512MB ,1GB I am getting with in link speed .But in case of GlusterFS I not able to understand what is happening . Please any one can help me . NFS : iozone -Raceb ./perffinal.wks -y 4K -q 128K -n 128K -g 1G -i 0 -i 1 Reader 4 8 16 32 64 128 128 744701 727625 935039 633768 499971 391433 256 920892 1085148 1057519 931149 551834 380335 512 937558 1075517 1100810 904515 558917 368605 1024 974395 1072149 1094105 969724 555319 379390 2048 1026059 1125318 1137073 1005356 568252 375232 4096 1021220 1144780 1169589 1030467 578615 376367 8192 965366 1153315 1071693 1072681 607040 371771 16384 1008989 1133837 1163806 1046171 600500 376056 32768 1022692 1165701 1175739 1065870 630626 363563 65536 1005490 1152909 1168181 1048258 631148 374343 131072 1011405 1161491 1176534 1048509 637910 375741 262144 1011217 1130486 1118877 1075740 636433 375511 524288 9563 9562 9568 9551 9525 9562 1048576 9499 9520 9513 9535 9493 9469 GlusterFS: iozone -Raceb /root/glusterfs/perfgfs2.wks -y 4K -q 128K -n 128K -g 1G -i 0 -i 1 Reader Report 4 8 16 32 64 128 128 48834 50395 49785 48593 48450 47959 256 15276 15209 15210 15100 14998 14973 512 12343 12333 12340 12291 12202 12213 1024 11330 11334 11327 11303 11276 11283 2048 10875 10881 10877 10873 10857 10865 4096 10671 10670 9706 10673 9685 10640 8192 10572 10060 10571 10573 10555 10064 16384 10522 10523 10523 10522 10522 10263 32768 10494 10497 10495 10493 10497 10497 65536 10484 10483 10419 10483 10485 10485 131072 10419 10475 10477 10445 10445 10478 262144 10323 10241 10312 10226 10320 10237 524288 10074 9966 9707 8567 8213 9046 1048576 7440 7973 5737 7101 7678 5743 Any idea for this higher value in NFS test .some this is different . But I am not able to understand. Thanks for your time Mohan ------=_Part_259374_19852034.1230880209611 Content-Type: text/html; charset=ISO-8859-1 Content-Transfer-Encoding: quoted-printable Content-Disposition: inline we're conducting performance benchmark runs to evaluate Linux per= formance as NFS file servers.<br>It is observed that an unusual high percen= tage of benchmark time was spent in "read" operation.<br>A sample= d workload consisting of 18% of read consumes 63% of total benchmark time. = Did this<br> problem get analyzed before (or even better :)-is there a patch) ? We'r= e on 2.4.19 kernel- NFS<br>V3 - UDP, with EXT3 as local file system.<br><br= >Thanks in advance.<br><br><a href=3D"mailto:gluster-users at gluster.org">glu= ster-users at gluster.org</a><br> <br>Dear All,<br><br>we are currently using NFS to meet data sharing requir= ements.Now we are facing some performance and scalability problem ,so= this form does not meet the requirements of our network(performance).So we= are finding the possible solutions to increase the performance and scalabi= lity .To give very strong solution to NFS issue I have analysed two File Sy= stem one is GlusterFS and another one is Red Hat GFS.we conclude that Glust= erFS will increase the performance and scalability ,It has all the features= we are looking .For the testing purpose I am benchmarking NFS and GlusterF= S to get better performance .My benchmark result shows that GlusterFS give = better performance ,but i am getting some unacceptable read performance . I= am not able to understand how exactly the read operation performs NFS and = GlusterFS .even I don't know anything i am doing wrong.here i am showin= g the benchmark result to get better idea of my read performance issuee .i = have attached the result of NFS and GlusterFS read values .any one ca= n please go thro this and give me some valuable guide .It will make my benc= hmarking very effective . <br> <br>This my server and client Hardware and software :<br><br>HARDWARE CONFI= G:<br><br>Processor core speed : Intel(R) Celeron(R) CPU 1.70GHz<br><= br>Number of cores : Single Core (not dual-core)<br><br>RAM size = ; : 384MB(128MB+256MB)<br> <br>RAM type : DDR<br><br>RAM Speed : 266 MHz (3.8 ns)<br><br>S= wap : 1027MB<br><br>Storage controller : ATA device<br><br>Disk= model/size : SAMSUNG SV4012H /40 GB,2 MB Cache,<br><br>Storage speed= : 52.4 MB/sec<br><br> Spindle Speed : 5400 rpm(Revolution per Minute)<br><br>NIC Type = : VIA Rhine III chipset IRQ 18<br><br>NIC Speed : 100 Mbps/Full-Dupl= ex Card<br><br>SOFTWARE:<br><br>Operation System : Fedora Core 9 GNU/Linux<= br><br>Linux version : 2.6.9-42<br> <br>Local FS : Ext3<br><br>NFS version : 1.1.2<br><br>GlusterFS= version: glusterfs 1.3.8 built on Feb 3 2008<br><br>Iozone : iozone-= 3-5.fc9.i386 (File System Benchmark Tool)<br><br>ttcp : ttcp-1.12-18.= fc9.i386(RAW throughput measurement Tool)<br> <br>This is the server and client vol files i am using the benchmarking<br>= <br>#GlusterFS Server Volume Specification<br><br>volume brick<br> ty= pe storage/posix  = ; # POSIX FS translator<br>= option directory /bench # = /bench dir contains 25,000 files with size 10 KB 15KB<br> end-volume<br><br>volume iot<br> type performance/io-threads<br> = ; option thread-count 4 <br> option cache-size 8MB<br> subvolum= es brick<br>end-volume<br><br>volume server<br> type protocol/server<= br> option transport-type tcp/server <br> subvolumes iot<br> option auth.ip.brick.allow * # Allow access= to "brick" volume<br>end-volume<br><br><br><br># GlusterFS Clien= t Volume Specification <br><br>volume client<br> type protocol/client= <br> option transport-type tcp/client <br> option remote-host 192.xxx.x.xxx = <br> option remote-subvolume brick &nbs= p; <br>end-volume<br><br>volume readahead<br> type performance/read-a= head<br> option page-size 128KB # 256KB is th= e default option<br> option page-count 4 # ca= che per file =3D (page-count x page-size) 2 is default option<b= r> subvolumes client<br>end-volume<br><br>volume iocache<br> type= performance/io-cache<br> #option page-size 128KB ## defa= ult is 32MB<br> option cache-size 256MB #128KB is default optio= n<br> option page-count 4 <br> subvolumes readahead<br>end-volume<br><br>volume writeback<br> = type performance/write-behind<br> option aggregate-size 128KB<br>&nb= sp; option flush-behind on<br> subvolumes iocache <br>end-volum= e<br><br><br>I am confusing this result .I don't have idea how to trace= and get good comparable result is read performance .I think I am miss unde= rstanding the buffer cache concepts .<br> <br>From attached NFS read result , I understand that I have 348MB RAM = ; and I am doing the benchmark file size rage 128KB to 1GB .So up to file s= ize 256MB I am getting buffer cache performance and file size 512MB ,= 1GB I am getting with in link speed .But in case of GlusterFS I not a= ble to understand what is happening . <br> <br>Please any one can help me .<br><br>NFS :<br>iozone -Raceb ./perffinal.= wks -y 4K -q 128K -n 128K -g 1G -i 0 -i 1 &n= bsp; <br><br>Reader =  = ; <br> 4 &n= bsp; 8&nb= sp; 16 &nb= sp; 32 &nb= sp; 64 &nb= sp; 128<br> <br>128 744701 727625 = 935039 633768 499971 = 391433<br>256 920892 1085148 &nbs= p; 1057519 931149 551834 &nb= sp; 380335<br>512 937558 1075517&= nbsp; 1100810 904515 558917= 368605<br>1024 974395  = ; 1072149 1094105 969724 &nb= sp; 555319 379390<br> 2048 1026059 1125318 = 1137073 1005356 568252  = ; 375232<br>4096 1021220 1144780 &= nbsp; 1169589 1030467 578615 = ; 376367<br>8192 965366 115= 3315 1071693 1072681 = 607040 371771<br>16384 1008989 &nb= sp; 1133837 1163806 1046171 = 600500 376056<br> 32768 1022692 1165701 = 1175739 1065870 630626 &nbs= p; 363563<br>65536 1005490 1152909 = ; 1168181 1048258 631148&nb= sp; 374343<br>131072 1011405  = ; 1161491 1176534 1048509 &n= bsp; 637910 375741<br>262144 1011217&nb= sp; 1130486 1118877 1075740= 636433 375511<br> 524288 9563 9562 9568= 9551 9525 9562<br>10= 48576 9499 9520 9513&= nbsp; 9535 9493 9469<br><br= >GlusterFS:<br>iozone -Raceb /root/glusterfs/perfgfs2.wks -y 4K -q 128K -n = 128K -g 1G -i 0 -i 1 &nbs= p; <br> Reader Report  = ; <br> &= nbsp; 4 &n= bsp; 8  = ; 16 32 &= nbsp; 64 &nbs= p; 128<br><br>128 48834 50395 &nbs= p; 49785 48593 48450 &= nbsp; 47959<br>256 15276 15209 &nb= sp; 15210 15100 14998 = 14973<br> 512 12343 12333 12340= 12291 12202 12213<br= >1024 11330 11334 113= 27 11303 11276 11283<= br>2048 10875 10881 1= 0877 10873 10857 1086= 5<br>4096 10671 10670 = 9706 10673 9685 1064= 0<br> 8192 10572 10060 1057= 1 10573 10555 10064<b= r>16384 10522 10523 1= 0523 10522 10522 1026= 3<br>32768 10494 10497  = ; 10495 10493 10497 1= 0497<br>65536 10484 10483 &n= bsp; 10419 10483 10485  = ; 10485<br> 131072 10419 10475 10= 477 10445 10445 10478= <br>262144 10323 10241  = ; 10312 10226 10320 1= 0237<br>524288 10074 9966 &n= bsp; 9707 8567 8213 9= 046<br>1048576 7440 7973 &nb= sp; 5737 7101 7678 57= 43<br> <br>Any idea for this higher value in NFS test .some this is different . Bu= t I am not able to understand.<br><br><br><br><br><br>Thanks for your time<= br>Mohan<br><br> ------=_Part_259374_19852034.1230880209611--