When viewing the Technical Program schedule, on the far righthand side
is a column labeled "PLANNER." Use this planner to build your own
schedule. Once you select an event and want to add it to your personal
schedule, just click on the calendar icon of your choice (outlook
calendar, ical calendar or google calendar) and that event will be
stored there. As you select events in this manner, you will have your
own schedule to guide you through the week.
You can also create your personal schedule on the SC11 app (Boopsie) on your smartphone. Simply select a session you want to attend and "add" it to your plan. Continue in this manner until you have created your own personal schedule. All your events will appear under "My Event Planner" on your smartphone.
Co-visualization of full data and in situ data extracts from unstructured grid CFD at 160K cores
SESSION: Research Poster Reception
EVENT TYPE: ACM Student Research Competition Poster, Poster, Electronic Poster
TIME: 5:15PM - 7:00PM
SESSION CHAIR: Bernd Mohr
AUTHOR(S):Michel Rasquin, Patrick Marion, Venkatram Vishwanath, Benjamin Matthews, Mark Hereld, Kenneth Jansen, Raymond Loy, Andrew Bauer, Min Zhou, Onkar Sahni, Jing Fu, Ning Liu, Christopher Carothers, Mark Shephard, Michael Papka, Kalyan Kumaran, Berk Geveci
ROOM:WSCC North Galleria 2nd/3rd Floors
ABSTRACT: Scalability and time-to-solution studies have historically been focused on the size of the problem and run time. We consider a more strict definition of “solution” whereby a live data analysis provides continuous and reconfigurable insight into massively parallel simulations.
Specifically, we used the Argonne ALCF's BlueGene/P machine with 163,840 cores tightly linked through a high-speed network to 100 visualization nodes that share 200 GPUs.
Two meshes with 416M and 3.3B elements discretize the flow over a full swept wing with an unsteady synthetic jet. The 416M element mesh takes about 2 seconds per flow solve step including the extraction and rendering of a slice or a contour, slowing currently the simulation by only 10 and 15% respectively. The 3.3B element case proved scalable at about 15 seconds per time step.
This work will be presented with informational slides combined with a looping video and live demonstrations of our software.
Bernd Mohr (Chair) - Juelich Supercomputing Centre
Michel Rasquin - University of Colorado at Boulder
Patrick Marion - Kitware, Inc.
Venkatram Vishwanath - Argonne National Laboratory
Benjamin Matthews - University of Colorado at Boulder
Mark Hereld - Argonne National Laboratory
Kenneth Jansen - University of Colorado at Boulder
Raymond Loy - Argonne National Laboratory
Andrew Bauer - Kitware, Inc.
Min Zhou - Renssalaer Polytechnic University
Onkar Sahni - Renssalaer Polytechnic University
Jing Fu - Renssalaer Polytechnic University
Ning Liu - Renssalaer Polytechnic University
Christopher Carothers - Renssalaer Polytechnic University