Re: [RFC] [PATCH 0/18] xfstests: move tests out of top level

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 8/23/12 12:00 PM, Ben Myers wrote:
> Dave,
> 
> On Thu, Aug 23, 2012 at 09:42:19AM +1000, Dave Chinner wrote:
>> On Wed, Aug 22, 2012 at 02:16:42PM -0500, Ben Myers wrote:
>>> On Wed, Aug 22, 2012 at 08:09:26AM +1000, Dave Chinner wrote:
>>>> On Tue, Aug 21, 2012 at 11:33:37AM -0500, Ben Myers wrote:
>>>>> On Tue, Aug 21, 2012 at 08:43:06AM +1000, Dave Chinner wrote:
>>>>>> On Mon, Aug 20, 2012 at 04:27:25PM -0500, Mark Tinguely wrote:
>>>>>>> On 07/26/12 04:27, Dave Chinner wrote:
>>>>>>>> Alt-Subject: Games with Sed, Grep and Awk.
>>>>>>>>
>>>>>>>> This series is based on top of the large filesystem test series.
>>>>>>>>
>>>>>>>> This moves all the tests into a ./tests subdirectory, and sorts them into
>>>>>>>> classes of related tests. Those are:
>>>>>>>>
>>>>>>>> 	tests/generic:	valid for all filesystems
>>>>>>>> 	tests/shared:	valid for a limited number of filesystems
>>>>>>>> 	tests/xfs:	xfs specific tests
>>>>>>>> 	tests/btrfs	btrfs specific tests
>>>>>>>> 	tests/ext4	ext4 specific tests
>>>>>>>> 	tests/udf	udf specific tests
>>>>>>>
>>>>>>> The SGI XFS group talked about your proposed changes to xfstests and
>>>>>>> the response is very positive.
>>>>>>>
>>>>>>> The couple concerns are:
>>>>>>>
>>>>>>> 1) There is a consensus in the group that the benchmark framework
>>>>>>>    should remain until there is a common benchmark available.
>>>>>>>
>>>>>>>    Could the benchmark infrastructure be placed into its own directory
>>>>>>>    until a new common benchmark framework has been adopted?
>>>>>>
>>>>>> Keeping it just complicates things. The benchmark infrastructure
>>>>>> is bitrotted and was largely just a hack tacked on to the side of
>>>>>> the regression test suite.
>>>>>>
>>>>>> For it to be useful in an automated test environment, it would need
>>>>>> to be re-implemented from scratch with reliable recording of results
>>>>>> and the ability to determine if a result is unusual or not. None of
>>>>>> this exists - it's just a framework to run a couple of benchmarks
>>>>>> and dump some output to stdout using the xfstests machine config
>>>>>> files....
>>>>>>
>>>>>> I have tried integrating other benchmarks into xfstests a while back
>>>>>> (e.g. compile bench, fsmark, etc) and using the results for some
>>>>>> kind of meaningful performance regression test. I rapidly came to
>>>>>> the conclusion that the infrastructure was not up to scratch and
>>>>>> that my simple handwritten standalone test scripts to iterate
>>>>>> through benchmarks and capture results was much easier to use and
>>>>>> modify than to jump through the weird bench infrastructure hoops.
>>>>>>
>>>>>> So, no, I don't think it's worth keeping at all.
>>>>>
>>>>> You've already made it clear that you feel the current bench implementation is
>>>>> not worth keeping.  Once a suitable replacement for the bench infrastructure
>>>>> has been implemented we can remove the old one.  Until then we prefer to keep
>>>>> what we have in the tree.
>>>>
>>>> That's not how the process works 
>>>
>>> That is exactly how the process works.  You posted an RFC and Mark and the XFS
>>> team at SGI walked through your patch set.  Mark subsequently posted the
>>> commentary in reply to your RFC.  Cruft or not, the removal of a feature goes
>>> through the same review process as everything else.
>>
>> Sure, but you need to justify your arguments for keeping something
>> with evidence and logic - handwaving about wanting something is, and
>> always has been, insufficient justification. That's the part of the
>> process I'm talking about - that statements of need require
>> evidence, especially when you agreed to the removal at LSF in San
>> Fransisco a few months ago. My arguments at the time were:
>>
>> 	a) nobody is actually using it,
>> 	b) it has effectively been unmaintained since 2003
>> 	c) it has no regression analysis or detection capability
>> 	d) it shares *very little* of xfstests
>> 	e) it gets in the way of cleaning up xfstests
>> 	f) there are far better workload generators that are being
>> 	actively maintained.
>>
>> And AFAIA, nothing has changed in the past few months.
> 
> "In this case, SGI would like to keep the benchmark capability in xfstests in
> order have a better chance of catching performance regressions."  There has
> been a been performance regression in the past few months (and there will be
> more in the future), we have had performance regressions internally too, and
> this has brought the value of having benchmarks in xfstests into sharp focus.

"xfs has had performance regressions; xfstests contains bitrotted perf code"

But that's not a justification for keeping bitrotted code.

I think you finally answered the basic question Dave asked, and we learned
that SGI is not using the code which he proposes removing.

<snip>

> I understand that bench is bitrotted, but it still has some value even today.

Not if nobody uses it.  If it really had value it would be in use.

> Phil has agreed to take this on as a project so the bitrot will be addressed.

How's that been going in the 6 months since this patchset stalled?

Can we get it moving again?  Ext4 folks would like to see these changes
proceed as well.  What issues remain, if any?

Thanks,
-Eric

_______________________________________________
xfs mailing list
xfs@xxxxxxxxxxx
http://oss.sgi.com/mailman/listinfo/xfs


[Index of Archives]     [Linux XFS Devel]     [Linux Filesystem Development]     [Filesystem Testing]     [Linux USB Devel]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux