On 04/24/2015 10:01 AM, Christoph Lameter wrote: > On Thu, 23 Apr 2015, Paul E. McKenney wrote: > >>> As far as I know Jerome is talkeing about HPC loads and high performance >>> GPU processing. This is the same use case. >> >> The difference is sensitivity to latency. You have latency-sensitive >> HPC workloads, and Jerome is talking about HPC workloads that need >> high throughput, but are insensitive to latency. > > Those are correlated. > >>> What you are proposing for High Performacne Computing is reducing the >>> performance these guys trying to get. You cannot sell someone a Volkswagen >>> if he needs the Ferrari. >> >> You do need the low-latency Ferrari. But others are best served by a >> high-throughput freight train. > > The problem is that they want to run 2000 trains at the same time > and they all must arrive at the destination before they can be send on > their next trip. 1999 trains will be sitting idle because they need > to wait of the one train that was delayed. This reduces the troughput. > People really would like all 2000 trains to arrive on schedule so that > they get more performance. So you run 4000 or even 6000 trains, and have some subset of them run at full steam, while others are waiting on memory accesses. In reality the overcommit factor is likely much smaller, because the GPU threads run and block on memory in smaller, more manageable numbers, say a few dozen at a time. -- All rights reversed -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>