Observations of an Accidental Computational Scientist

advertisement
Observations of an Accidental
Computational Scientist
SIAM/NSF/DOE CSME Workshop
25 March 2003
David Keyes
Department of Mathematics & Statistics
Old Dominion University
&
Institute for Scientific Computing Research
Lawrence Livermore National Laboratory
Academic and lab backgrounds

74-78: B.S.E., Aerospace and
Mechanical/Engineering Physics

78-84: M.S. & Ph.D., Applied
Mathematics

84-85: Post-doc, Computer
Science

86-93: Asst./Assoc. Prof.,
Mechanical Engineering

93-99: Assoc. Prof., Computer
Science



86-02: ICASE, NASA Langley
99-03: Prof., Mathematics &
Statistics

99- : ISCR, Lawrence
Livermore
03- : Prof., Applied Physics &
Applied Mathematics

03- : CDIC, Brookhaven
25 March 2003 SIAM/NSF/DOE CSMR Workshop
Computational Science & Engineering

A “multidiscipline” on the verge of full bloom





Envisioned by Von Neumann and others in the 1940’s
Undergirded by theory (numerical analysis) for the past
fifty years
Empowered by spectacular advances in computer
architecture over the last twenty years
Enabled by powerful programming paradigms in the last
decade
Adopted in industrial and government applications



Boeing 777’s computational design a renowned milestone
DOE NNSA’s “ASCI” (motivated by CTBT)
DOE SC’s “SciDAC” (motivated by Kyoto, etc.)
25 March 2003 SIAM/NSF/DOE CSMR Workshop
Niche for computational science

Has theoretical aspects (modeling)

Has experimental aspects (simulation)

Unifies theory and experiment by providing common
immersive environment for interacting with multiple
data sets of different sources

Provides “universal” tools, both hardware and software
Telescopes are for astronomers, microarray analyzers are
for biologists, spectrometers are for chemists, and
accelerators are for physicists, but computers are for
everyone!

Costs going down, capabilities going up every year
25 March 2003 SIAM/NSF/DOE CSMR Workshop
Simulation complements experimentation
Experiments
dangerous
Experiments
difficult to
instrument
Experiments
prohibited or
impossible
Engineering
electromagnetics
aerodynamics
Physics
cosmology
radiation transport
Environment
global climate
wildland firespread
Ex #3
Ex #2
Scientific
Simulation
Ex #1
Experiments
expensive
Energy
combustion
fusion
Ex #4
personal examples
25 March 2003 SIAM/NSF/DOE CSMR Workshop
Example #1: wildland firespread
Simulate fires at the wildland-urban interface, leading to strategies for
planning preventative burns, fire control, and evacuation
“It looks as if
all of Colorado
is burning” –
Bill Owens,
Governor
“About half of
the U.S. is in
altered fire
regimes” –
Ron Myers,
Nature Conservancy
Joint work between ODU, CMU, Rice, Sandia, and TRW
25 March 2003 SIAM/NSF/DOE CSMR Workshop
Example #1: wildland firespread, cont.



Objective
Develop mathematical models for tracking the evolution of
wildland fires and the capability to fit the model to fires of
different character (fuel density, moisture content, wind,
topography, etc.)
Accomplishment to date
Implemented firefront propagation with level set method with
empirical front advance function; working with firespread
experts to “tune” the resulting model
Significance
Wildland fires cost many lives and billions of dollars annually;
other fire models pursued at national labs are more detailed,
but too slow to be used in real time; one of our objectives is to
offer practical tools to firechiefs in the field
25 March 2003 SIAM/NSF/DOE CSMR Workshop
Example #2: aerodynamics
Simulate airflows over wings and streamlined bodies on highly
resolved grids leading to superior aerodynamic design
Joint work between ODU, Argonne, LLNL, and NASA-Langley
25 March 2003 SIAM/NSF/DOE CSMR Workshop
Example #2: aerodynamics, cont.



Objective
Develop analysis and optimization capability for compressible
and incompressible external aerodynamics
Accomplishment to date
Developed highly parallel nonlinear implicit solvers (NewtonKrylov-Schwarz) for unstructured grid CFD, implemented in
PETSc, demonstrated on a “workhorse” NASA code running
on the ASCI machines (up to 6,144 processors)
Significance
Windtunnel tests of aerodynamic bodies are expensive and
difficult to instrument; computational simulation and
optimization (as for the Boeing 777) will greatly reduce the
engineering risk of developing new fuel-efficient aircraft, cars,
etc.
25 March 2003 SIAM/NSF/DOE CSMR Workshop
Example #3: radiation transport
Simulate “flux-limited diffusion” transport of radiative energy in
inhomogeneous materials
Joint work between ODU, ICASE, and LLNL
25 March 2003 SIAM/NSF/DOE CSMR Workshop
Example #3: radiation transport, cont.



Objective
Enhance accuracy and reliability of analysis methods used in the
simulation of radiation transport in real materials
Accomplishment to date
Leveraged expertise and software (PETSc) developed for
aerodynamics simulations in a related physical application
domain, also governed by nonlinear PDEs discretized on
unstructured grids, where such methods were less developed
Significance
Under current stockpile stewardship policies, DOE must be able
to reliably predict the performance of high-energy devices
without full-scale physical experiments
25 March 2003 SIAM/NSF/DOE CSMR Workshop
Example #4: fusion energy
Simulate plasmas in tokomaks, leading to understanding of plasma
instability and (ultimately) new energy sources
Joint work between ODU, Argonne, LLNL, and PPPL
25 March 2003 SIAM/NSF/DOE CSMR Workshop
Example #4: fusion energy, cont.



Objective
Improve efficiency and therefore extend predictive capabilities of
Princeton’s leading magnetic fusion energy code “M3D” to
enable it to operate in regimes where practical sustained
controlled fusion occurs
Accomplishment to date
Augmented the implicit linear solver (taking up to 90% of
execution time) of original code with parallel algebraic
multigrid; new solvers are much faster and robust, and should
scale better to the finer mesh resolutions required for M3D
Significance
An M3D-like code will be used in DOE’s Integrated Simulation
and Optimization of Fusion Systems, and ITER collaborations,
with the goal of delivering cheap safe fusion energy devices by
early-to-mid 21st century
25 March 2003 SIAM/NSF/DOE CSMR Workshop
We lead the “TOPS” project
U.S. DOE has created the Terascale Optimal PDE Simulations
(TOPS) project within the Scientific Discovery through Advanced
Computing (SciDAC) initiative; nine partners in this 5-year, $17M
project, an “Integrated Software Infrastructure Center”
25 March 2003 SIAM/NSF/DOE CSMR Workshop
Toolchain for PDE Solvers in TOPS* project



Design and implementation of “solvers”
 Time integrators

(w/ sens. anal.)
Nonlinear solvers

(w/ sens. anal.)
Constrained optimizers
f ( x , x, t , p)  0
F ( x, p )  0
min  ( x, u ) s.t. F ( x, u )  0, u  0
u

Linear solvers

Eigensolvers
Ax  b
Optimizer
Sens. Analyzer
Time
integrator
Nonlinear
solver
Ax  Bx
Eigensolver
Linear
solver
Software integration
Performance optimization
Indicates
dependence
*Terascale Optimal PDE Simulations: www.tops-scidac.org
25 March 2003 SIAM/NSF/DOE CSMR Workshop
SciDAC apps and infrastructure
4 projects
in high
energy and
nuclear
physics
14 projects in
biological and
environmental
research
17 projects in
scientific
software and
network
infrastructure
5 projects
in fusion
energy
science
10 projects
in basic
energy
sciences
25 March 2003 SIAM/NSF/DOE CSMR Workshop
Optimal solvers
Convergence rate nearly
independent of discretization
parameters

Multilevel schemes for
linear and nonlinear
problems

Time to Solution

Newton-like schemes for
quadratic convergence of
nonlinear problems
700
600
500
400
200
100
0
3
60
50
150
40
100
30
10
100
12
27
procs
48
75
time
1000
Problem Size (increasing with number of
processors)
AMG shows perfect
iteration scaling, above,
in contrast to ASM, but
still needs performance
work to achieve
temporal scaling, below,
on CEMM fusion code,
M3D, though time is
halved (or better) for
large runs (all runs: 4K
dofs per processor)
ASM-GMRES
AMG-FMGRES
AMG inner
20
scalable
0
1
ASM-GMRES
AMG-FMGRES
300
200
50
iters
10
0
3
12
27
48
75
25 March 2003 SIAM/NSF/DOE CSMR Workshop
We have run on most ASCI platforms…
100+ Tflop / 30 TB
Livermore
Capability
50+ Tflop / 25 TB
30+ Tflop / 10 TB
White 10+ Tflop / 4 TB
Blue
Red
‘97
3+ Tflop / 1.5 TB
Plan
Develop
1+ Tflop / 0.5 TB
‘98
‘99
Use
‘00
Sandia
‘01
‘02
Time (CY)
‘03
‘04
Livermore
‘05 ‘06
Los Alamos
NNSA has roadmap to go
to 100 Tflop/s by 2006
www.llnl.gov/asci/platforms
…and now the SciDAC platforms
IBM Power3+ SMP
16 procs per node
208 nodes
24 Gflop/s per node
5 Tflop/s (doubled in February to 10)
Berkeley
IBM Power4 Regatta
32 procs per node
24 nodes
166 Gflop/s per node
4Tflop/s (10 in 2003)
Oak Ridge
25 March 2003 SIAM/NSF/DOE CSMR Workshop
Computational Science at Old Dominion

Launched in 1993 as “High Performance Computing”

Keyes appointed ‘93; Pothen early ’94

Major projects:

NSF Grand, National, and Multidisciplinary Challenges (19951998) [w/ ANL, Boeing, Boulder, ND, NYU]

DoEd Graduate Assistantships in Areas of National Need (19952001)

DOE Accelerated Strategic Computing Initiative “Level 2” (19982001) [w/ ICASE]

DOE Scientific Discovery through Advanced Computing (20012006) [w/ ANL, Berkeley, Boulder, CMU, LBNL, LLNL, NYU,
Tennessee]

NSF Information Technology Research (2001-2006) [w/ CMU,
Rice, Sandia, TRW]
25 March 2003 SIAM/NSF/DOE CSMR Workshop
CS&E at ODU today

Center for Computational Science at ODU established
8/2001; new 80,000 sq ft building (for Math, CS, Aero,
VMASC, CCS) opens 1/2004; finally getting local buy-in

ODU’s small program has placed five PhDs at DOE labs in
the past three years
25 March 2003 SIAM/NSF/DOE CSMR Workshop
Post-doctoral and student alumni
Linda Stals, ANU
Lois McInnes, ANL
Satish Balay, ANL Dinesh Kaushik, ANL
D. Karpeev, ANL David Hysom, LLNL Gary Kumfert, LLNL Florin Dobrian, ODU
25 March 2003 SIAM/NSF/DOE CSMR Workshop
<Begin> “pontification phase”
Five models that allow CS&E to prosper

Laboratory institutes (hosted at a lab)
ICASE, ISCR (more details to come)

National institutes (hosted at a university)
IMA, IPAM

Interdisciplinary centers
ASCI Alliances, SciDAC ISICs, SCCM, TICAM, CAAM, …

CS&E fellowship programs
CSGF, HPCF

Multi-agency funding (cyclical to be sure, but sometimes
collaborative)
DOD, DOE, NASA, NIH, NSF, …
25 March 2003 SIAM/NSF/DOE CSMR Workshop
LLNL’s ISCR fosters collaborations with
academe in computational science
Serves as lab’s point of contact for
computational science interests
Influences the external research community to
pursue laboratory-related interests
Manages LLNL’s ASCI Institute collaborations in
computer science and computational mathematics
Assists LLNL in technical workforce recruiting
and training
25 March 2003 SIAM/NSF/DOE CSMR Workshop
ISCR’s philosophy:
Science is borne by people

Be “eyes and ears” for LLNL by staying abreast of
advances in computer and computational science

Be “hands and feet” for LLNL by carrying those
advances into the laboratory

Three principal means for packaging scientific ideas for
transfer


papers

software

people
People are the most effective!
25 March 2003 SIAM/NSF/DOE CSMR Workshop
ISCR brings visitors to LLNL through a
variety of programs (FY 2002 data)
Seminars & Visitors
180 visits from 147 visitors
66 ISCR seminars
Summer Program
ISCR
B451
Postdocs & Faculty
9 postdoctoral researchers
3 faculty-in-residence
43 grad students
29 undergrads
24 faculty
Workshops & Tutorials
10 tutorial lectures
6 technical workshops
25 March 2003 SIAM/NSF/DOE CSMR Workshop
ISCR is the largest of LLNL’s six institutes

Founded in 1986

Under current leadership since June 1999
160
ISCR has grown with
LLNL’s increasing
reliance on simulation as
a predictive science
140
120
100
80
60
Seminars
Visitors
Students
40
20
0
FY 97 FY 98 FY 99 FY 00 FY 01 FY 02
25 March 2003 SIAM/NSF/DOE CSMR Workshop
Our academic collaborators are drawn from
all over

University of California
Berkeley
Davis
Irvine
Los Angeles
San Diego
Santa Barbara
Santa Cruz

ASCI ASAP-1 Centers
Caltech
Stanford University
University of Chicago
University of Illinois
University of Utah

Major European Centers
University of Bonn
University of Heidelberg

Other Universities
Carnegie Mellon
Florida State University
MIT
Ohio State University
Old Dominion University
RPI
Texas A&M University
University of Colorado
University of Kentucky
University of Minnesota
University of N. Carolina
University of Tennessee
University of Texas
University of Washington
Virginia Tech
and more!
25 March 2003 SIAM/NSF/DOE CSMR Workshop
Internships in Terascale Simulation
Technology (ITST) tutorials
Students in residence hear from enthusiastic members of lab divisions,
besides their own mentor, including five authors* of recent computational
science books, on a variety of computational science topics
Lecturers: David Brown, Eric Cantu-Paz*, Alej Garcia*, Van Henson*,
Chandrika Kamath, David Keyes, Alice Koniges*, Tanya Kostova, Gary
Kumfert, John May*, Garry Rodrigue
25 March 2003 SIAM/NSF/DOE CSMR Workshop
ISCR pipelines people between the
university and the laboratory
Universities
ISCR
Lab programs
Students
Faculty visit the ISCR, bringing students
Faculty
Most faculty return to university, with lab priorities
Lab Employees
Some students become lab employees
Some students become faculty, with lab priorities
A few faculty become lab employees
25 March 2003 SIAM/NSF/DOE CSMR Workshop
ISCR impact on DOE computational
science hiring

178 ISCR summer students in past five years (many repeaters)

51 have by now emerged from the academic pipeline

23 of these (~45%) are now working for the DOE


15 LLNL

3 each LANL and Sandia

1 each ANL and BNL
11 of these (~20%) are in their first academic appointment

In US: Duke, Stanford, U California, U Minnesota, U Montana,
U North Carolina, U Pennsylvania, U Utah, U Washington

Abroad: Swiss Federal Institute of Technology (ETH),
University of Toronto
25 March 2003 SIAM/NSF/DOE CSMR Workshop
ISCR sponsors and conducts meetings on
timely topics for lab missions

Bay Area NA Day

Common Component Architecture

Copper Mountain Multigrid Conference

DOE Computational Science Graduate Fellows

Hybrid Particle-Mesh AMR Methods

Mining Scientific Datasets

Large-scale Nonlinear Problems

Overset Grids & Solution Technology

Programming ASCI White

Sensitivity and Uncertainty Quantification
25 March 2003 SIAM/NSF/DOE CSMR Workshop
We hosted a “Power Programming” short
course to prepare LLNL for ASCI White

Steve White, IBM
ASCI White overview, POWER3 architecture, tuning for White

Larry Carter, UCSD/NPACI
designing kernels and data structures for scientific applications, cache
and TLB issues

David Culler, UC Berkeley
understanding performance thresholds

Clint Whalley, U Tennessee
coding for performance

Bill Gropp, Argonne National Lab
MPI-1, Parallel I/O, MPI/OpenMP tradeoffs
65 internal attendees
over 3 days
25 March 2003 SIAM/NSF/DOE CSMR Workshop
We launched the Terascale Simulation
Lecture Series to receptive audiences










Fred Brooks, UNC
Ingrid Daubechies, Princeton
David Johnson, AT&T
Peter Lax, NYU
Michael Norman, UCSD
Charlie Peskin, NYU
Gil Strang, MIT
Burton Smith, Cray
Eugene Spafford, Purdue
Andries Van Dam, Brown
25 March 2003 SIAM/NSF/DOE CSMR Workshop
<Continue> “pontification phase”
Concluding swipes

A curricular challenge for CS&E programs

Signs of the times for CS&E

“Red skies at morning” ( “sailers take warning”)

“Red skies at night” (“sailers delight”)

Opportunities in which CS&E will shine

A word to the sponsors
25 March 2003 SIAM/NSF/DOE CSMR Workshop
A curricular challenge




CS&E majors without a CS undergrad need to learn to compute!
Prerequisite or co-requisite to becoming useful interns at a lab
Suggest a “bootcamp” year-long course introducing:
 C/C++ and object-oriented program design
 Data structures for scientific computing
 Message passing (e.g., MPI) and multithreaded (e.g., OpenMP)
programming
 Scripting (e.g., Python)
 Linux clustering
 Scientific and performance visualization tools
 Profiling and debugging tools
NYU’s sequence G22.1133/G22.1144 is an example for CS
25 March 2003 SIAM/NSF/DOE CSMR Workshop
“Red skies at morning”

Difficult to get support for maintaining critical software
infrastructure and “benchmarking” activities

Difficult to get support for hardware that is designed with
computational science and engineering in mind

Difficult for pre-tenured faculty to find reward structures
conducive to interdisciplinary efforts

Unclear how stable is the market for CS&E graduates at the
entrance to a 5-year pipeline

Political necessity of creating new programs with each
change of administrations saps time and energy of managers
and community
25 March 2003 SIAM/NSF/DOE CSMR Workshop
“Red skies at night”

DOE’s SciDAC model being recognized and propagated

NSF’s DMS budgets on a multi-year roll

SIAM SIAG-CSE attracting members from outside of traditional
SIAM departments

CS&E programs beginning to exhibit “centripetal” potential in
traditionally fragmented research universities
e.g., SCCM’s “Advice” program

Computing at the large scale is weaning domain scientists from
“Numerical Recipes” and MATLAB and creating thirst for core
enabling technologies (NA, CS, Viz, …)

Cost effectiveness of computing, especially cluster computing, is
putting a premium on graduate students who have CS&E skills
25 March 2003 SIAM/NSF/DOE CSMR Workshop
Opportunity: nanoscience modeling

Jul 2002 report to DOE

Proposes $5M/year theory and
modeling initiative to accompany
the existing $50M/year
experimental initiative in nano
science

Report lays out research in
numerical algorithms and
optimization methods on the
critical path to progress in
nanotechnology
25 March 2003 SIAM/NSF/DOE CSMR Workshop
Opportunity: integrated fusion modeling

Dec 2002 report to DOE

Currently DOE supports 52 codes
in Fusion Energy Sciences

US contribution to ITER will
“major” in simulation

Initiative proposes to use advanced
computer science techniques and
numerical algorithms to improve
the US code base in magnetic fusion
energy and allow codes to
interoperate
25 March 2003 SIAM/NSF/DOE CSMR Workshop
A word to the sponsors



Don’t cut off the current good stuff to start the new stuff
Computational science & engineering workforce enters the pipeline
from a variety of conventional inlets (disciplinary first, then
interdisciplinary)
Personal debts:
 NSF HSSRP in Chemistry (SDSU)
 NSF URP in Computer Science (Brandeis) – precursor to
today’s REU
 NSF Graduate Fellowship in Applied Mathematics
 NSF individual PI grants in George Lea’s computational
engineering program – really built community (Benninghof,
Farhat, Ghattas, C. Mavriplis, Parsons, Powell + many others
active in CS&E at labs, agencies, and universities today) at
NSF-sponsored PI meetings, long before there was any
university support at all
25 March 2003 SIAM/NSF/DOE CSMR Workshop
Related URLs
 Personal
homepage: papers, talks, etc.
http://www.math.odu.edu/~keyes
 ISCR
(including annual report)
http://www.llnl.gov/casc/iscr
 SciDAC
initiative
http://www.science.doe.gov/scidac
 TOPS
software project
http://www.math.odu.edu/~keyes/scidac
25 March 2003 SIAM/NSF/DOE CSMR Workshop
The power of optimal algorithms

Advances in algorithmic efficiency rival advances in hardware
architecture

Consider Poisson’s equation on a cube of size N=n3
Year
Method
Reference
Storage
Flops
1947
GE (banded)
Von Neumann &
Goldstine
n5
n7
64
1950
Optimal SOR
Young
n3
n4 log n
2u=f
1971
CG
Reid
n3
n3.5 log n
1984
Full MG
Brandt
n3
n3

64
If n=64, this implies an overall reduction in flops of ~16 million
*On a 16 Mflop/s machine, six-months is reduced to 1 s
25 March 2003 SIAM/NSF/DOE CSMR Workshop
64
Algorithms and Moore’s Law


This advance took place over a span of about 36 years, or 24
doubling times for Moore’s Law
22416 million  the same as the factor from algorithms alone!
relative
speedup
year
25 March 2003 SIAM/NSF/DOE CSMR Workshop
The power of optimal algorithms

Since O(N) is already optimal, there is nowhere further
“upward” to go in efficiency, but one must extend optimality
“outward”, to more general problems

Hence, for instance, algebraic multigrid (AMG), obtaining
O(N) in anisotropic, inhomogeneous problems
R
n
error damped
by pointwise
relaxation
AMG Framework
algebraically
smooth error
Choose coarse grids,
transfer operators, etc. to
eliminate, based on
numerical weights,
heuristics
25 March 2003 SIAM/NSF/DOE CSMR Workshop
Download