Einstein@Home Set To Break Petaflops Barrier 96
hazeii writes "Einstein@home, the distributed computing project searching for the gravitational waves predicted to exist by Albert Einstein, looks set to breach the 1 Petaflops barrier around midnight UTC tonight. Put into context, if it was in the Top500 Supercomputers list, it would be in at number 24. I'm sure there are plenty of Slashdot readers who can contribute enough CPU and GPU cycles to push them well over 1,000 teraflops — and maybe even discover a pulsar in the process."
From their forums: "At 14:45 we had 989.2 TFLOPS with an increase of 1.3 TFLOPS/h. In principle that's enough to reach 1001.1 TFLOPS at midnight (UTC) but very often, like yesterday, between 22:45 and 22:50 there occurs a drop of about 5 TFLOPS. So we will have very likely hit 1 PFLOPS in the early morning tomorrow. "
folding@home (Score:2, Insightful)
wouldn't it be wise for practical* reasons for people to offer more power to folding@home instead of einstein@home?
* = has more chances to help humanity ( for curing diseases etc. )
Re:Top500 doesn't work that way (Score:4, Insightful)
You have to run the Linpack benchmark and report that.
And I guess no distributed computing platform is ever going to score in top 500 according to that benchmark. The communication performance between nodes is very important to most parallel algorithms. Any decent benchmark would take that into account. A real super computer has much faster communication between the nodes, than what you can achieve across the Internet. Both throughput and latency matters. There are some specific problems which can be split into parts that can be computed independently by nodes without communication between them, but most super computers are used for tasks, that do not fall into that class.
At some point I heard the rule of thumb, that when you are building a super computer, you divide your funds in three equal parts. One of those was to be spent on the interconnect. I don't recall what the other two were supposed to be spend on.