Hello, Any updates on when SPC-1 is going to be available as part of fio ? Thanks, Kiran. On Wed, Nov 26, 2014 at 7:23 PM, Michael O'Sullivan <michael.osullivan@xxxxxxxxxxxxxx> wrote: > Hi Luis, > > We worked with Jens Axobe for a little bit to try and merge things but then just got busy testing distributed file systems as opposed to raw storage. > > We had an email in 2012 from > >>>I encountered a couple of segfaults when modifying the sample configuration file. >>> >>>I've thought to revamp it and make it more "fio" like, possibly turning SPC into a profile so that someone can just run "fio --profile=spc" > > But the person that emailed did not follow up. > > I think having an fio --profile=spc-1 would be great and I'd be happy to help get this working, but fio-type testing is not my core research area/area of expertise. We used fio+spc-1 to test disks in order to get inputs for optimal infrastructure design research (which is one mof my core research areas). That said I did a lot of the original development, so I can probably help people understand what the code is trying to do. > > I hope this helps. Please let me know if you'd like to revamp fio+spc-1 and if you need my help. > > Thanks, Mike > > -----Original Message----- > From: Luis Pabón [mailto:lpabon@xxxxxxxxxx] > Sent: Friday, 21 November 2014 3:24 a.m. > To: Michael O'Sullivan; Justin Clift > Cc: gluster-devel@xxxxxxxxxxx > Subject: Re: Open source SPC-1 Workload IO Pattern > > Hi Michael, > I noticed the code on the fio branch (that is where I grabbed the spc1.[hc] files :-) ). Do you know why that branch has not being merged to master? > > - Luis > > On 11/18/2014 11:56 PM, Michael O'Sullivan wrote: >> Hi Justin & Luis, >> >> We did a branch of fio that implemented this SPC-1 trace a few years ago. I can dig up the code and paper we wrote if it is useful? >> >> Cheers, Mike >> >>> On 19/11/2014, at 4:21 pm, "Justin Clift" <justin@xxxxxxxxxxx> wrote: >>> >>> Nifty. :) >>> >>> (Yeah, catching up on old unread email, as the wifi in this hotel is >>> so bad I can barely do anything else. 8-10 second ping times to >>> www.gluster.org. :/) >>> >>> As a thought, would there be useful analysis/visualisation >>> capabilities if you stored the data into a time series database (eg >>> InfluxDB) then used Grafana (http://grafana.org) on it? >>> >>> + Justin >>> >>> >>> On Fri, 07 Nov 2014 12:01:56 +0100 >>> Luis Pabón <lpabon@xxxxxxxxxx> wrote: >>> >>>> Hi guys, >>>> I created a simple test program to visualize the I/O pattern of >>>> NetApp's open source spc-1 workload generator. SPC-1 is an >>>> enterprise OLTP type workload created by the Storage Performance >>>> Council (http://www.storageperformance.org/results). Some of the >>>> results are published and available here: >>>> http://www.storageperformance.org/results/benchmark_results_spc1_active . >>>> >>>> NetApp created an open source version of this workload and described >>>> it in their publication "A portable, open-source implementation of >>>> the SPC-1 workload" ( >>>> http://www3.lrgl.uqam.ca/csdl/proceedings/iiswc/2005/9461/00/0152601 >>>> 4.pdf >>>> ) >>>> >>>> The code is available onGithub: https://github.com/lpabon/spc1 . >>>> All it does at the moment is capture the pattern, no real IO is >>>> generated. I will be working on a command line program to enable >>>> usage on real block storage systems. I may either extend fio or >>>> create a tool specifically tailored to the requirements needed to >>>> run this workload. >>>> >>>> On github, I have an example IO pattern for a simulation running 50 >>>> mil IOs using HRRW_V2. The simulation ran with an ASU1 (Data Store) >>>> size of 45GB, ASU2 (User Store) size of 45GB, and ASU3 (Log) size of >>>> 10GB. >>>> >>>> - Luis >>>> >>>> _______________________________________________ >>>> Gluster-devel mailing list >>>> Gluster-devel@xxxxxxxxxxx >>>> http://supercolony.gluster.org/mailman/listinfo/gluster-devel >>> >>> >>> -- >>> GlusterFS - http://www.gluster.org >>> >>> An open source, distributed file system scaling to several petabytes, >>> and handling thousands of clients. >>> >>> My personal twitter: twitter.com/realjustinclift >>> _______________________________________________ >>> Gluster-devel mailing list >>> Gluster-devel@xxxxxxxxxxx >>> http://supercolony.gluster.org/mailman/listinfo/gluster-devel > > _______________________________________________ > Gluster-devel mailing list > Gluster-devel@xxxxxxxxxxx > http://supercolony.gluster.org/mailman/listinfo/gluster-devel _______________________________________________ Gluster-devel mailing list Gluster-devel@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-devel