PPT - Larry Smarr - California Institute for Telecommunications and

advertisement
“The Pacific Research Platform”
Briefing to The Quilt Visit
to Calit2’s Qualcomm Institute
University of California, San Diego
February 10, 2016
Dr. Larry Smarr
Director, California Institute for Telecommunications and Information Technology
Harry E. Gruber Professor,
Dept. of Computer Science and Engineering
Jacobs School of Engineering, UCSD
http://lsmarr.calit2.net
1
Vision: Creating a West Coast “Big Data Freeway”
Connected by CENIC/Pacific Wave
Use Lightpaths to Connect
All Data Generators and Consumers,
Creating a “Big Data” Freeway
Integrated With High Performance Global Networks
“The Bisection Bandwidth of a Cluster Interconnect,
but Deployed on a 20-Campus Scale.”
This Vision Has Been Building for Over a Decade
NSF’s OptIPuter Project: Demonstrating How SuperNetworks
Can Meet the Needs of Data-Intensive Researchers
LS Slide 2005
2003-2009
$13,500,000
OptIPortal–
Termination
Device
for the
OptIPuter
Global
Backplane
In August 2003,
Jason Leigh and his
students used
RBUDP to blast data
from NCSA to SDSC
over the
TeraGrid DTFnet,
achieving18Gbps file
transfer out of the
available 20Gbps
Calit2 (UCSD, UCI), SDSC, and UIC Leads—Larry Smarr PI
Univ. Partners: NCSA, USC, SDSU, NW, TA&M, UvA, SARA, KISTI, AIST
Industry: IBM, Sun, Telcordia, Chiaro, Calient, Glimmerglass, Lucent
DOE ESnet’s Science DMZ: A Scalable Network
Design Model for Optimizing Science Data Transfers
• A Science DMZ integrates 4 key concepts into a unified whole:
– A network architecture designed for high-performance applications,
with the science network distinct from the general-purpose network
– The use of dedicated systems for data transfer
– Performance measurement and network testing systems that are
regularly used to characterize and troubleshoot the network
– Security policies and enforcement mechanisms that are tailored for
high performance science environments
The DOE ESnet Science DMZ and the NSF “Campus Bridging” Taskforce Report Formed the Basis
for the NSF Campus Cyberinfrastructure Network Infrastructure and Engineering (CC-NIE) Program
Science DMZ
Coined 2010
http://fasterdata.es.net/science-dmz/
Based on Community Input and on ESnet’s Science DMZ Concept,
NSF Has Funded Over 100 Campuses to Build Local Big Data Freeways
Red 2012 CC-NIE Awardees
Yellow 2013 CC-NIE Awardees
Green 2014 CC*IIE Awardees
Blue 2015 CC*DNI Awardees
Purple Multiple Time Awardees
Source: NSF
The Pacific Research Platform: The Next Logical Step –
Connect Multiple Campus Science DMZs with 10-100Gbps Lightpaths
NSF CC*DNI Grant
$5M 10/2015-10/2020
PI: Larry Smarr, UC San Diego Calit2
Co-Pis:
• Camille Crittenden, UC Berkeley CITRIS,
• Tom DeFanti, UC San Diego Calit2,
• Philip Papadopoulos, UC San Diego SDSC,
• Frank Wuerthwein, UC San Diego Physics
and SDSC
FIONA – Flash I/O Network Appliance:
Termination Device for 10-100Gbps Flows
FIONAs Are
Science DMZ Data Transfer Nodes &
Optical Network Termination Devices
UCSD CC-NIE Prism Award & UCOP
Phil Papadopoulos & Tom DeFanti
Joe Keefe & John Graham
UCOP Rack-Mount Build:
Cost
$8,000
$20,000
Intel Xeon
Haswell Multicore
E5-1650 v3
6-Core
2x E5-2697 v3
14-Core
RAM
128 GB
256 GB
SSD
SATA 3.8 TB
SATA 3.8 TB
Network Interface
10/40GbE
Mellanox
2x40GbE
Chelsio+Mellanox
GPU
NVIDIA Tesla K80
RAID Drives 0 to 112TB (add ~$100/TB)
John Graham, Calit2’s QI
FIONAs as
Uniform DTN End Points
FIONA DTNs
Existing DTNs
As of October 2015
UC FIONAs Funded by
UCOP “Momentum” Grant
Ten Week Sprint to Demonstrate
the West Coast Big Data Freeway System: PRPv0
FIONA DTNs Now Deployed to All UC Campuses
And Most PRP Sites
Presented at CENIC 2015
March 9, 2015
Pacific Research Platform
Multi-Campus Science Driver Teams
• Jupyter Hub
• Biomedical
– Cancer Genomics Hub/Browser
– Microbiome and Integrative ‘Omics
– Integrative Structural Biology
• Earth Sciences
–
–
–
–
Data Analysis and Simulation for Earthquakes and Natural Disasters
Climate Modeling: NCAR/UCAR
California/Nevada Regional Climate Data Analysis
CO2 Subsurface Modeling
• Particle Physics
• Astronomy and Astrophysics
– Telescope Surveys
– Galaxy Evolution
– Gravitational Wave Astronomy
• Scalable Visualization, Virtual Reality, and Ultra-Resolution Video
10
PRP First Application: Distributed IPython/Jupyter Notebooks:
Cross-Platform, Browser-Based Application Interleaves Code, Text, & Images
IJulia
IHaskell
IFSharp
IRuby
IGo
IScala
IMathics
Ialdor
LuaJIT/Torch
Lua Kernel
IRKernel (for the R language)
IErlang
IOCaml
IForth
IPerl
IPerl6
Ioctave
Calico Project
• kernels implemented in Mono,
including Java, IronPython,
Boo, Logo, BASIC, and many
others
IScilab
IMatlab
ICSharp
Bash
Clojure Kernel
Hy Kernel
Redis Kernel
jove, a kernel for io.js
IJavascript
Calysto Scheme
Calysto Processing
idl_kernel
Mochi Kernel
Lua (used in Splash)
Spark Kernel
Skulpt Python Kernel
MetaKernel Bash
MetaKernel Python
Brython Kernel
IVisual VPython Kernel
Source: John Graham, QI
PRP UC-JupyterHub Backbone
Next Step: Deploy Across PRP
UC Berkeley
Source: John Graham, Calit2
UC San Diego
GPU JupyterHub:
GPU JupyterHub:
2 x 14-core CPUs
256GB RAM
1.2TB FLASH
3.8TB SSD
Nvidia K80 GPU
Dual 40GbE NICs
And a Trusted Platform
Module
1 x 18-core CPUs
128GB RAM
3.8TB SSD
Nvidia K80 GPU
Dual 40GbE NICs
And a Trusted Platform
Module
community resources. This facility depends on a range of common services, support activities, software,
and operational principles
thatFederates
coordinate the production
of scientific
knowledge
through the DHTC
OSG
Clusters
in 40/50
States:
model. In April 2012, the OSG project was extended until 2017; it is jointly funded by the Department of
Major XSEDE Resource
Energy and the National Science A
Foundation.
Source: Miron Livny, Frank Wuerthwein, OSG
Open Science Grid Has Had a Huge Growth Over the Last Decade Currently Federating Over 130 Clusters
Source: Miron Livny, Frank Wuerthwein, OSG
Crossed
100 Million
Core-Hours/Month
In Dec 2015
CMS
ATLAS
Supported Over
200 Million Jobs
In 2015
Over 1 Billion
Data Transfers
Moved
200 Petabytes
In 2015
PRP Prototype of LambdaGrid Aggregation of OSG Software & Services
Across California Universities in a Regional DMZ
• Aggregate Petabytes of Disk Space
& PetaFLOPs of Compute,
Connected at 10-100 Gbps
other sciences
life sciences
ATLAS
other physics
CMS
• Transparently Compute on Data
at Their Home Institutions & Systems
at SLAC, NERSC, Caltech, UCSD, & SDSC
UCD
SLAC
OSG Hours 2015
by Science Domain
UCSC
CSU Fresno
UCSB
Caltech
UCI
Source: Frank Wuerthwein,
UCSD Physics;
SDSC; co-PI PRP
UCSD
& SDSC
UCR
PRP Builds
on SDSC’s
LHC-UC Project
Two Automated Telescope Surveys
Creating Huge Datasets Will Drive PRP
Precursors to
LSST and NCSA
PRP Allows Researchers
to Bring Datasets from NERSC
to Their Local Clusters
for In-Depth Science Analysis
300 images per night.
100MB per raw image
250 images per night.
530MB per raw image
30GB per night
150 GB per night
120GB per night
When processed
at NERSC
Increased by 4x
Source: Peter Nugent, Division Deputy for Scientific Engagement, LBL
Professor of Astronomy, UC Berkeley
800GB per night
Global Scientific Instruments Will Produce Ultralarge Datasets Continuously
Requiring Dedicated Optic Fiber and Supercomputers
Square Kilometer Array
Large Synoptic Survey Telescope
Tracks ~40B Objects,
Creates 10M Alerts/Night
Within 1 Minute of Observing
2x40Gb/s
https://tnc15.terena.org/getfile/1939
www.lsst.org/sites/default/files/documents/DM%20Introduction%20-%20Kantor.pdf
PRP Will Support the Computation and Data Analysis
in the Search for Sources of Gravitational Radiation
Augment the aLIGO Data and Computing Systems
at Caltech, by connecting at 10Gb/s
to SDSC Comet supercomputer,
enabling LIGO computations
to enter via the same PRP “job cache” as for LHC.
HPWREN Users and Public Safety Clients
Gain Redundancy and Resilience from PRP Upgrade
• PRP CENIC 10G Link UCSD to SDSU
–
–
–
–
–
DTN FIONAs Endpoints
Data Redundancy
Disaster Recovery
High Availability
Network Redundancy
10X Increase During Wildfires
Data From Hans-Werner Braun
UCR
San Diego Countywide
Sensors and Camera
Resources
UCSD & SDSU
Source: Frank Vernon,
Greg Hidley, UCSD
Data & Compute
Resources
UCI
SDSU
UCSD
UCI & UCR
Data Replication
and PRP FIONA Anchors
as HPWREN Expands
Northward
PRP Backbone Sets Stage for 2016 Expansion
of HPWREN, Connected to CENIC, into Orange and Riverside Counties
•
Anchor to CENIC at UCI
UCR
– PRP FIONA Connects to
CalREN-HPR Network
– Data Replication Site
•
•
Potential Future UCR CENIC
Anchor
Camera and Relay Sites at:
–
–
–
–
–
–
–
–
UCI
Santiago Peak
Sierra Peak
Lake View
Bolero Peak
Modjeska Peak
Elsinore Peak
Sitton Peak
Via Marconi
UCSD
Collaborations through COAST –
County of Orange Safety Task Force
Source: Frank Vernon,
Greg Hidley, UCSD
SDSU
PRP Links FIONA Clusters
Creating Distributed Virtual Reality
PRP
20x40G PRP-connected
40G FIONAs
WAVE@UC San Diego
CAVE@UC Merced
PRP is NOT Just for Big Data Science and Engineering:
Linking Cultural Heritage and Archaeology Datasets
Building on CENIC’s Expansion
To Libraries, Museums,
and Cultural Sites
UW/
PNWGP
Seattle
UCD
Internet2
Seattle
Berkeley
“In an ideal world –
Extremely high bandwidth to
move large cultural heritage
datasets around the PRP cloud for
processing & viewing in CAVEs
around PRP with Unlimited Storage
for permanent archiving.”
-Tom Levy, UCSD
Esnet
DoE Labs
UCSF
Stanford
UCM
NASA
AMES/
NREN
Internet2
UCSC
UCSB
Los
Nettos
UCR
Caltech
* Institutions with
Active Archaeology Programs
USC
UCLA
UCI
Note: This diagram represents a subset of sites and connections.
UCSD
SDSU
Next Step: Global Research Platform
Building on CENIC/Pacific Wave and GLIF
Current
International
GRP Partners
Download