Hi Hillf,
On 2023/4/23 16:12, Hillf Danton wrote:
On 23 Apr 2023 14:08:49 +0800 Gao Xiang <hsiangkao@xxxxxxxxxxxxxxxxx>
On 2023/4/22 13:18, Hillf Danton wrote:
On 21 Apr 2023 15:12:45 -0700 Douglas Anderson <dianders@xxxxxxxxxxxx>
Add a variant of folio_lock() that can timeout. This is useful to
avoid unbounded waits for the page lock in kcompactd.
Given no mutex_lock_timeout() (perhaps because timeout makes no sense for
spinlock), I suspect your fix lies in the right layer. If waiting for
page under IO causes trouble for you, another simpler option is make
IO faster (perhaps all you can do) for instance. If kcompactd is waken
up by kswapd, waiting for slow IO is the right thing to do.
A bit out of topic. That is almost our original inital use scenarios for
Thanks for taking a look.
EROFS [1] although we didn't actually test Chrome OS, there lies four
points:
1) 128kb compressed size unit is not suitable for memory constraint
workload, especially memory pressure scenarios, that amplify both I/Os
and memory footprints (EROFS was initially optimized with 4KiB
pclusters);
Feel free to take another one at 2M THP [1].
[1] https://lore.kernel.org/lkml/20230418191313.268131-1-hannes@xxxxxxxxxxx/
Honestly I don't catch your point here, does THP has some relationship with
this? Almost all smartphones (but I don't know Chromebook honestly) didn't
use THP at that time.
2) If you turn into a small compressed size (e.g. 4 KiB), some fs behaves
ineffective since its on-disk compressed index isn't designed to be
random accessed (another in-memory cache for random access) so you have
to count one by one to calculate physical data offset if cache miss;
3) compressed data needs to take extra memory during I/O (especially
low-ended devices) that makes the cases worse and our camera app
workloads once cannot be properly launched under heavy memory pressure,
but in order to keep user best experience we have to keep as many as
apps active so that it's hard to kill apps directly. So inplace I/O +
decompression is needed in addition to small compressed sizes for
overall performance.
Frankly nowadays I have no interest in running linux with <16M RAM for example.
Our cases are tested on 2016-2018 devices under 3 to 6 GB memory if you
take a glance at the original ATC paper, the page 9 (section 5.1) wrote:
"However, it costed too much CPU and memory resources, and when trying to
run a camera application, the phone froze for tens of seconds before it
finally failed."
I have no idea how 16M RAM here comes from but smartphones doesn't have
such limited memory. In brief, if you runs few app, you have few problem.
but as long as you keeps more apps in background (and running), then the
memory will eventually suffer pressure.
4) If considering real-time performance, some algorithms are not quite
suitable for extreme pressure cases;
Neither in considering any perf under extreme memory pressure (16M or 64G RAM)
because of crystally pure waste of time.
Personally I don't think so, if you'd like to land an effective compression
approach for end users and avoid user complaints (app lagging, app frozen,
etc). I think these all need to be considered in practice.
Thanks,
Gao Xiang
etc.
I could give more details on this year LSF/MM about this, although it's not
a new topic and I'm not a Android guy now.
Did you book the air ticket? How many bucks?
[1] https://www.usenix.org/conference/atc19/presentation/gao
Thanks,
Gao Xiang