A.R.M.S. Active Resource Management Services Presentation One 2/21/2013

advertisement
A.R.M.S. Active Resource
Management Services
Presentation One
2/21/2013
1
Outline
Introductions
Societal Issue Examined
Michael Rajs
2/21/2013
2
Outline
•
•
•
•
•
•
•
•
•
Group Members and Roles:
slide 4
Introduce Mentor: slide 5
Societal Issue: slide 6
History: slides 7-11
Case Study: slides 12-16
Problem Statement: slide 18
Computer Components
Identified: slides 19 -21
Major Functional Component
Diagram: slide 22
Current Process Flow: slide
23
2/21/2013
•
•
•
•
•
•
•
•
Solution Statement: slide 25
Objectives: slide 26
Improved Process Flow:
slide 27
Competition Identified:
slides 28-30
Benefits of Solution: slide 32
Problems with Solution: slide
33
Conclusion: slide 34
References: slides 35-36
3
Group Members and Roles
• Michael Rajs (Group Manager)
• Adam Willis (Research Specialist)
• Sybil Acotanza (Visualization
Engineer)
• Scott Pardue (Team Leader)
• Jordan Heinrichs (Marketing Analyst)
• David Crook (Documentation
Specialist)
2/21/2013
4
Yaohang Li
•Is an Associate Professor
in the Department of
Computer Science at Old
Dominion University.
•His research interests are
in Computational
Biology, Markov Chain
Monte Carlo (MCMC)
methods and Parallel
Distributed Grid
Computing.
2/21/2013
5
What is the societal issue being faced?
How do researchers handle the
massive amounts of data they are
collecting?
2/21/2013
6
Historical Background
Adam Willis
2/21/2013
7
Collection of Data
• 1890 Census Recorded With an Electric Machine 1
• 1935 Social Security Act 2
• 1974 Privacy Act 3
• 1989 World Wide Web 4
• 1997 Big Data 5
• 2011 IBM’s Watson 6
• Now
“Every day, we create 2.5 quintillion bytes of data —
so much that 90% of the data in the world today has
been created in the last two years alone.”7
2/21/2013
8
Examples of Big Data
• Large Hadron Collider 8
– 150 million sensors report 40 million
times per second
• Facebook 9
– 2.5 billion – content items shared
– 2.7 billion – “Likes”
– 300 million – photos uploaded
• Walmart 8
– 1 million customer transactions
– 2.5 petabytes of data
2/21/2013
9
Big Data Analysis Hardware
• Cluster Computing 10
– A cluster consists of many nodes (computers).
– Big data can be generated and analyzed
quicker by spreading the workload amongst
the nodes.
2/21/2013
10
Managing the Cluster
• Distributed Resource
Management Systems (D-RMS)
–Job management subsystem
–Physical resource management
subsystem
–Scheduling and queuing
subsystem
2/21/2013
11
Case Study
Sybil Acotanza
2/21/2013
12
Dinosolve Case Study
• Bioinformatics
– Disulfide bond prediction program
2/21/2013
(Cronk, 2012)
13
Dinosolve Users
• Who will use it?
– Drug and antibody design
– Bio-energy development
– Genetic mapping11
• Why will they use it?
– 2% accuracy improvement12
2/21/2013
14
Dinosolve Web Site
2/21/2013 (Li & Yaseen, http://hpcr.cs.odu.edu/dinosolve/)
15
Dinosolve Possible Problems
• Hard resources for computation
– CPU cycles
– Memory
– Disk space
– Network bandwidth
• Server crashes
2/21/2013
16
Problem statement
Components of Hardware and Software
Current Process Flow
Scott Pardue
2/21/2013
17
What is the problem?
Processing time on big data sets is
computationally expensive and as
the volume of queries grows the
system will progressively drop in
performance until the system fails.
2/21/2013
18
What are the components of our current
system?
The current system uses the
following software and hardware.
2/21/2013
19
Software
• Unix operating system installed on
the dinosolve cluster
• Dinosolve algorithm
• Sun Grid Engine which will be our
Distributed Resource Management
System (D-RMS) installed on the
cluster.
• MySQL (database software)
• Web based user interface (website)
2/21/2013
20
Hardware
• MySQL database server
• A computer cluster to run the
dinosolve algorithm
• Web server for our web based
user interface
2/21/2013
21
Major Functional Component Diagram
2/21/2013
22
2/21/2013
23
Solution Statement
Objectives
Improved Process Flow
Competition Identified
Jordan Heinrichs
2/21/2013
24
How will we correct the problem?
We aim to configure a distributed
resource management system
(D-RMS), in this case Sun Grid
Engine (SGE), to handle resource
allocation on the dinosolve cluster.
2/21/2013
25
Objectives
• Interpret and visualize current
usage statistics
• Configure, utilize, and optimize
the SGE
• Aesthetically pleasing and
professional user interface
2/21/2013
26
Process Flow with Solution
2/21/2013
27
Competing Distributed Resource
Management Systems
• Sun Grid Engine (SGE)
• Portable Batch System (PBS)
• Load Sharing Facility (LSF)
2/21/2013
28
Competing Resource Management
Systems
Features of
systems
PBS
LSF
SGE
Supported
platforms
Unix
Unix & NT
Unix
Multi-cluster
support
Yes
Yes
No
System level
checkpoint restart
Yes
Yes
Yes
User level
checkpoint restart
No
Yes
Yes
Large
computational grid
support
No
No
No
Massive Scalability
Yes
Yes
Yes
Parallel job
support with Sun
HPC ClusterTools
Loose Integration
Tight Integration
Loose Integration
Distribution format
of end product
Source
Binary only
Binary and Source
Free?
Yes
No
Yes
Posix 1002.2d
compliance
Yes
No
Yes
2/21/2013
Reference 31
29
Competing Protein Prediction Servers
Dinosolve
DiANNA
Scrath Protein
Predictor
Accuracy
90.8%
81%
87%
Usability
X
X
X
508.22
compliance
percentage
67%
85%
67%
Professional
2/21/2013
Reference 19,20 and 21
30
Benefits of solution
Problems with solution
Conclusion
David Crook
2/21/2013
31
What benefits will come from attaining
our goals?
• Efficient utilization of available
resources
• Increased throughput of the cluster
• An intuitive and professional user
interface
• Rise in popularity due to excellent
accuracy, efficiency, and professional
design
2/21/2013
32
Problems with solution
• Improper synchronization of cluster
resources can lead to a deadlock in
the system
• Race conditions between the HPCR
cluster and the MySQL database
2/21/2013
33
Conclusion
With the updated user interface
and correctly configured Sun Grid
Engine we hope to establish a
reputable Disulfide Bonding
Prediction Server.
2/21/2013
34
References for history
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
http://www.columbia.edu/cu/computinghistory/hh/index.html
http://query.nytimes.com/gst/abstract.html?res=F50C11FE385D13728DDDAE0A
94DA415B868FF1D3
http://www.census.gov/history/pdf/kraus-natdatacenter.pdf
http://www.bbc.co.uk/history/historic_figures/berners_lee_tim.shtml
http://dl.acm.org/citation.cfm?id=266989.267068&coll=DL&dl=GUIDE
http://www.nytimes.com/2012/08/12/business/how-big-data-became-so-bigunboxed.html?_r=1
http://www-01.ibm.com/software/data/bigdata/
http://en.wikipedia.org/wiki/Big_data
http://techcrunch.com/2012/08/22/how-big-is-facebooks-data-2-5-billionpieces-of-content-and-500-terabytes-ingested-every-day/
http://en.wikipedia.org/wiki/Computer_cluster
2/21/2013
35
References for case study
11. Li, Y. (2010, September 1). CAREER: Novel Sampling Approaches for Protein
Modeling Applications [Abstract]. National Science Foundation Award
Abstract #1066471.
12. Li, Y., & Yaseen, A. (2012). Enhancing Protein Disulfide Bonding
Prediction Accuracy with Context-based Features. Biotechnology and
Bioinformatics Symposium
13. bioinformatics. 2011. In Merriam-Webster.com. Retrieved February 15, 2013,
from http://www.merriam-webster.com/dictionary/bioinformatics
14. Cronk, J. D. (2012). Disulfide Bond. Retrieved February 15, 2013, from
Biochemistry Dictionary:
http://guweb2.gonzaga.edu/faculty/cronk/biochem/Dindex.cfm?definition=disulfide_bond
15. Yan, Y., & Chapman, B. (2008). Comparative Study of Distributed Resource
Management Systems–SGE, LSF, PBS Pro, and LoadLeveler. Technical ReportCiteseerx.
16. Li, Y., & Yaseen, A. (2012). Dinosolve. Retrieved from
http://hpcr.cs.odu.edu/dinosolve/
2/21/2013
36
References for competition
17. Arvind Krishna, “Why Big Data? Why Now?”, IBM , 2011
URL: http://almaden.ibm.com/colloquium/resources/Why%20Big%20Data%20Krishna.PDF
18. Yonghong Yan, Barbara M. Chapman, Comparative Study of Distributed Resource Management
Systems - SGE, LSF, PBS Pro, and LoadLeveler, Department of Computer Science, University of Houston, May
2005 (pdf)
19. Dr. Li’s site
http://hpcr.cs.odu.edu/dinosolve/
20. Scratch Predictor
http://scratch.proteomics.ics.uci.edu/
21. DiANNA server
http://clavius.bc.edu/~clotelab/DiANNA/
Portable Batch System (PBS)
22. http://resources.altair.com/pbs/documentation/support/PBSProUserGuide12-2.pdf
23. http://www.pbsworks.com/SupportDocuments.aspx?AspxAutoDetectCookieSupport=1
24. http://resources.altair.com/pbs/documentation/support/PBSProRefGuide12-2.pdf
25.http://resources.altair.com/pbs/documentation/support/PBSProAdminGuide12-2.pdf
26.http://www.pbsworks.com/(S(tykrsyqbemmlf3o5zwrmjrgf))/images/solutions-en-US/PBS-Pro_DatasheetUSA_WEB.pdf
27.http://agendafisica.files.wordpress.com/2011/05/pbs.pdf
Moab HPC Suite
28.http://www.adaptivecomputing.com/publication/420/wppa_open/
IBM Platform LSF
29.http://public.dhe.ibm.com/common/ssi/ecm/en/dcd12354usen/DCD12354USEN.PDF
Apache Hadoop with Zookeeper
30. http://zookeeper.apache.org/doc/current/zookeeperOver.html
31. http://www.cloud-net.org/~swsellis/tech/solaris/performance/doc/blueprints/0102/jobsys.pdf
2/19/2013
References
37
Download