On Tue, May 15, 2012 at 02:57:05PM +0900, Atsushi Kumagai wrote: > Hello Don, > > On Fri, 11 May 2012 09:26:01 -0400 > Don Zickus <dzickus at redhat.com> wrote: > > > > Thank you for your reply, Don, Vivek. > > > > > > As Don said, I tried to change the method to exclude free pages and > > > planed to resolve the memory consumption issue after it, because > > > parsing free list repeatedly may cause a performance issue. > > > > > > However, I'm thinking that to fix the size of memory consumption is more > > > important than to resolve a performance issue for large system. > > > > > > So I'm afraid that I would like to change the plan as: > > > > > > 1. Implement "iterating filtering processing" to fix the size of memory > > > consumption. At this stage, makedumpfile will parse free list repeatedly > > > even though it may cause a performance issue. > > > > > > 2. Take care of the performance issue after the 1st step. > > > > Hello Atsushi-san, > > > > Hmm. The problem with the free list is that the addresses are in random > > order, hence the reason to parse it repeatedly, correct? > > Yes. > > > I figured, now that you have a solution to parse the addresses in a linear > > way (the changes you made a couple of weeks ago), you would just continue > > with that. With that complete, we can look at the performance issues and > > solve them then. > > > > But it is up to you. You are willing to do the work, so I will defer to > > your judgement on how best to proceed. :-) > > What I wanted to tell you was I want to resolve the memory consumption issue > as soon as possible. In other words, I think the method to exclude free pages > is not so important. > I'll continue to work with the method which is easy to implement. Ok. I look forward to your results. Thanks for your effort. Cheers, Don