Re: buffer-cache builds up with invalidate=1 too

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



> Il giorno 27 ott 2017, alle ore 16:21, Jens Axboe <axboe@xxxxxxxxx> ha scritto:
> 
> On 10/27/2017 12:52 AM, Paolo Valente wrote:
>> [RESENDING, BECAUSE REJECTED BY THE VGER]
>> 
>>> Il giorno 27 ott 2017, alle ore 08:22, Paolo Valente <paolo.valente@xxxxxxxxxx> ha scritto:
>>> 
>>> 
>>> 
>>> Il 26/ott/2017 06:32 AM, "Jens Axboe" <axboe@xxxxxxxxx> ha scritto:
>>> On 10/24/2017 08:10 AM, Paolo Valente wrote:
>>>> 
>>>>> Il giorno 24 ott 2017, alle ore 08:28, Sitsofe Wheeler <sitsofe@xxxxxxxxx> ha scritto:
>>>>> 
>>>>> Hi,
>>>>> 
>>>>> If memory serves it's actually slightly more complicated. If you are
>>>>> using loops=<number> then I *think* (you'll have to check) you will
>>>>> find that invalidation happens once per each loop start. However when
>>>>> you use time_based to do the repetition there is essentially only one
>>>>> "loop" (even though the job goes on forever) so loop actions only
>>>>> happen right at the start of the job with that option (that's why I
>>>>> put the scare quotes around "beginning" ;-).
>>>>> 
>>>> 
>>>> Thanks for this additional, useful piece of information.  Actually,
>>>> this further, possibly different caching behavior makes me think that
>>>> some extra comment in the manpage might be helpful.
>>> 
>>> Would probably make sense to change 'invalidate' to be a range of
>>> possible values:
>>> 
>>> 0       As it is now, never invalidate
>>> 1       As it is now, invalidate initially
>>> once    Same as '1', invalidate initially / once
>>> open    New value, invalidate on every open
>>> close   New value, invalidate on close
>>> 
>>> as I can definitely see reasons why you would want to invalidate every
>>> time you open the file.
>>> 
>>> To do that, the 'invalidate' option should be changed from a
>>> FIO_OPT_BOOL to a FIO_OPT_STR, and the above possible values should be
>>> added as posval[] for that option.
>>> 
>>> Compliment that with the an enum of ranges for the ovals:
>>> 
>>> enum {
>>>        FIO_FILE_INVALIDATE_OFF = 0,
>>>        FIO_FILE_INVALIDATE_ONCE,
>>>        FIO_FILE_INVALIDATE_OPEN,
>>>        FIO_FILE_INVALIDATE_CLOSE
>>> };
>>> 
>>> Hope this makes sense, should be trivial to add as most of the work is
>>> already documented in this email :-). The remaining bits is just calling
>>> file_invalidate_cache() in the proper locations,
>>> td_io_{open,close}_file() would be prime candidates.
>>> 
>>> IMO this solution would make things both clearer and more flexible
> 
> See my followup, fio already does invalidates for each open. The
> problem with time_based was that we just reset the file, we don't
> close and re-open it. That was fixed in git:
> 
> commit 0bcf41cdc22dfee6b3f3b2ba9a533b4b103c70c2
> Author: Jens Axboe <axboe@xxxxxxxxx>
> Date:   Thu Oct 26 12:08:20 2017 -0600
> 
>    io_u: re-invalidate cache when looping around without file open/close
> 
> so current git should work for your test case. Please test.
> 

Tested, it does solve the problem.  As a side note, and if useful for
you, the throughput is much higher with sequential reads and direct=0
(4.14-rc5, virtual disk on an SSD).  It happens because of merges,
which seem to not occur with direct=1.  I thought direct I/O skipped
buffering, but still enjoyed features as request merging, but probably
I'm just wrong.

Thanks for addressing this caching issue,
Paolo

> -- 
> Jens Axboe

--
To unsubscribe from this list: send the line "unsubscribe fio" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [Linux Kernel]     [Linux SCSI]     [Linux IDE]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux SCSI]

  Powered by Linux