The debugging tool is a significant milestone in LLNL's multi-year collaboration with the University of Wisconsin (UW), Madison and the University of New Mexico (UNM) to ensure supercomputers run more efficiently.
Playing a significant role in scaling up the Sequoia supercomputer, STAT, a 2011 R&D 100 Award winner, has helped both early access users and system integrators quickly isolate a wide range of errors, including particularly perplexing issues that only manifested at extremely large scales up to 1,179,648 compute cores. During the Sequoia scale-up, bugs in applications as well as defects in system software and hardware have manifested themselves as failures in applications. It is important to quickly diagnose errors so they can be reported to experts who can analyze them in detail and ultimately solve the problem.
"STAT has been indispensable in this capacity, helping the multi-disciplined integration team keep pace with the aggressive system scale-up schedule," said LLNL computer scientist Greg Lee.
"While testing a subsystem of Blue/Gene Q, my test program consistently failed only when scaled to 1,179,648 MPI processes. Although the test program was simple, the sheer scale at which this program ran made debugging efforts highly challenging. But when I applied STAT, it quickly revealed that one particular rank process was consistently stuck in a system call," said Dong Ahn, a computer scientist in Livermore Computing.
Based on this finding, a system expert took a close look at the compute core on which this rank process was running and discovered a hardware defect. "Replacing the component suddenly got the entire Sequoia system back to life," Ahn said. "Putting this exercise into perspective, this error was due to a defect in a tiny hardware unit, the decrementor, of a single hardware thread out of a total of 4.7 million hardware threads. I felt it was like finding a needle in a haystack over a coffee break."
Sequoia delivers 20 petaflops of peak power and was ranked No. 1 in June of this year's TOP500 list. It is currently ranked No. 2, behind Oak Ridge National Laboratory's Titan.
LLNL plans to use Sequoia's impressive computational capability to advance understanding of fundamental physics and engineering questions that arise in the National Nuclear Security Administration's (NNSA) program to ensure the safety, security and effectiveness of the United States' nuclear deterrent without testing. Sequoia also will support NNSA/DOE programs at LLNL that focus on nonproliferation, counterterrorism, energy, security, health and climate change.
As LLNL takes delivery of the Sequoia system and works to move it into production, computer scientists will migrate applications that have been running on earlier systems to this newer architecture. This is a period of intense activity for LLNL's application teams as they gain experience with the new hardware and software environment.
"Having a highly effective debugging tool that scales to the full system is vital to the installation and acceptance process for Sequoia. It is critical that our development teams have a comprehensive parallel debugging tool set as they iron out the inevitable issues that come up with running on a new system like Sequoia," said Kim Cupps, leader of the Livermore Computing Division at LLNL.
STAT is particularly important for LLNL because supercomputer simulations are essential in virtually every mission area of the Laboratory. The tool also has been used at other sites and proved to be effective on a wide range of supercomputer platforms, including Linux clusters and Cray systems.
The team is actively pursuing further optimization of STAT technologies and is exploring commercialization strategies. More information about STAT, including a link to the source code, is available on the Web.More Information
LLNL news release, Nov. 9, 2012"Venturing into the heart of high-performance computing simulations"
Anne Stark | EurekAlert!
Laser process simulation available as app for first time
23.11.2015 | Fraunhofer-Institut für Lasertechnik ILT
Powering the next billion devices with Wi-Fi
19.11.2015 | University of Washington
Wissenschaftler aus Tübingen und Houston haben zahlreiche neue Zelltypen im Gehirn identifiziert. Im Neocortex von erwachsenen Mäusen führten sie erstmals eine...
Ende der 1980er Jahre erlebte die Erde eine dramatische Klimaveränderung. Sie umfasste die Tiefen der Ozeane ebenso wie die obere Atmosphäre und reichte vom Nord- bis zum Südpol. Ausgelöst durch den Ausbruch des Vulkans El Chichón in Mexico 1982 und verstärkt durch menschliches Handeln folgte daraus die größte Temperaturverschiebung der letzten 1.000 Jahre. Erstmals nachgewiesen hat dies ein internationales Forscherteam um Prof. Philip C. Reid von der Plymouth University und der Sir Alister Hardy Foundation for Ocean Science (UK). Die Ergebnisse wurden kürzlich in der Fachzeitschrift „Global Change Biology“ veröffentlicht.
Abrupte Klimaveränderungen haben oft dramatische Folgen für unseren Planeten. Dennoch sind sie in ihrer Art, ihrem Ausmaß und in ihrer Wirkungsweise meist nur...
Planet Earth experienced a global climate shift in the late 1980s on an unprecedented scale, fuelled by anthropogenic warming and a volcanic eruption, according to new research published this week.
Scientists say that a major step change, or ‘regime shift’, in the Earth’s biophysical systems, from the upper atmosphere to the depths of the ocean and from...
Fraunhofer ISE demonstriert neue Zell- und Modultechnologien an der Außenfassade eines Laborgebäudes
Das Fraunhofer-Institut für Solare Energiesysteme ISE hat die Außenfassade eines seiner Laborgebäude mit 70 Photovoltaik-Modulen ausgerüstet. Die Module...
The Fraunhofer Institute for Solar Energy Systems ISE has installed 70 photovoltaic modules on the outer façade of one of its lab buildings. The modules were...
27.11.2015 | Veranstaltungen
27.11.2015 | Veranstaltungen
26.11.2015 | Veranstaltungen
27.11.2015 | Förderungen Preise
27.11.2015 | Unternehmensmeldung
27.11.2015 | Energie und Elektrotechnik