World's most popular travel blog for travel bloggers.

[Solved]: Regarding to the speed of supercomputer

, , No Comments
Problem Detail: 

There seems to be a sense of competition in creating the fastest computer in the world, and I feel like technically you just have to continue increasing the core number to gain more flops, which is exactly what some countries are doing, so how would you really bench mark the technological level of a supercomputer

for example

United States currently have two supercomputer on the top list with similar performance, however one has significantly less core than the other

Titan has 560000 cores 17590 Rmax Tflops with 8000 kW of power

Sequoia has 1570000 cores 17173 Rmax Tflops with 7800 kW of power

Does this mean that Titan is a much superior supercomputer than Sequoia as it only uses 1/3 of cores to reach a similar performance, and if it does, this will also imply that judging supercomputer by flops is completely meaningless?

Is there something I am missing here? Such as the increasing complexity of software as the amount of core gets larger and the synchronization of overhead to assign task to each core gets more sophisticated?

Any info would be much appreciated

Asked By : user1433153

Answered By : Wandering Logic

The only thing you are missing is the importance of the interconnect between cores/processors/racks and the design of the memory system for each processor.

I assume you are referring to the most recent top-500 supercomputer list. That list is sorted by Rmax, not Rpeak.

Rpeak is a measure of the theoretical maximum number of floating point operations that the system can perform. For Rpeak you are correct: just add more cores (or more vector units per core) and the Rpeak number will go up.

Rmax is (arguably) a little bit more meaningful. It is the maximal rate at which the system can run the Linpack benchmark. The Linpack benchmark requires the system to solve a huge, dense, system of linear equations using LU factorization with partial pivoting in double precision.

For solving a dense system of linear equations you need to carefully partition the problem among the cores, and there needs to be an efficient way for the cores to communicate results back and forth. So a big part of designing a computer that will get onto the Top500 list is careful design of the interconnect and the software used to send messages between processors.

On the top500 list they tell you what interconnect each system is using. Note that you don't find a system with Ethernet until you get to number 49. The first 48 all have infiniband, cray gemini or custom interconnects.

There is also a green500 list that measures the MFLOPS/Watt of systems while solving the Linpack benchmark. The difference between the top system and the 500th system on that list (from last November) is a factor of 75x.

Most people agree that there is a lot more that you want from a supercomputer in addition to being able to efficiently solve massive dense systems of linear equations. Jack Dongara (who originated the Linpack benchmark) has recently released a paper proposing a new benchmark that measures the speed at which a system can solve a sparse linear system. http://www.netlib.org/utk/people/JackDongarra/PAPERS/HPCG-Benchmark-utk.pdf.

Best Answer from StackOverflow

Question Source : http://cs.stackexchange.com/questions/12745

0 comments:

Post a Comment

Let us know your responses and feedback