I didn't think of that! It would definitely be a neat hack if it works. My first question is (again) timing: The size of any particular night's download is unpredictable. If there are alot of updated headers and rpms to download, and the "leading" machine isn't finished downloading when one of the "trailing" machines starts up, then the trailing machine can get into trouble by not having a full deck to play with. My second question is whether it would work in the case where the machines have slightly different configurations. Yum only downloads what it needs. So if one of the trailing machines needs some rpm that isn't in the leading machine's cache (because the leading machine didn't need it for its configuration) what then? I still think it's a game for trained professionals. And then only if you're forced to play at the end of a long thin pipe. For the rest of us, Disk is cheap! Enjoy! Rick On Saturday, November 15, 2003, at 04:36 PM, Owyn wrote: > On Sat, 2003-11-15 at 12:40, Jeff Smith wrote: >> If I understood the suggestion correctly, he was saying to share the >> header cache via nfs, not the repos themselves. Each update on each >> machine will be trying to update the information in the cache. >> And hell >> could break loose if they start fighting each other for control. >> >> But for the sake of simplicity, I like Rick's suggestion. Damn the >> sharing and duplicate effort all over the place. :-) > > Would not "yum -C" eliminate all chances of multiple updates to the > cache. One machine could do the leading updates. The rest would > just use > the cache.