Alternative Response Technology API
API Study-Progressing Learnings
Michael J. Cortez
BP America-Oil Spill Technology Manager
RRT-3 Presentation- November 6, 2013
API Alternative Response Technology
Working Group Report
•
Commissioned by API as part of its Sept 2010 JITF Study
 Capture learnings from Macondo Incident
 Propose ARTES enhancements based on the learnings
•
Team composition: ARTES team members from Macondo
(USCG, OSPR, Obrien’s, various OSR consultants, BP) along
with NOAA, EPA and other industry entities
•
Kicked off study in Dec 2011; completed and endorsed July
2013
•
Proposed RRT & NRT presentations for late 2013 to review
the conclusions & recommendations
•
Request RRT-3 Endorsement of proposed changes to ICS
Pre – Macondo ARTES
•
•
•
•
ARTES within Environmental Unit
For non-conventional response tools
Evaluate one idea/tool at a time
Used in concert with NCP product
schedule
• Limited external public engagement
• Used sporadically, therefore never truly
embedded in ICS structure
Macondo experience
•
•
•
•
•
•
•
•
•
•
120,00+ total submissions
Multiple technical reviews required
Conventional & non-conventional ideas submitted
Submissions via phone, fax, e:mail, internet, walk-up
From 100 countries in 88 languages
Multiple submission channels (PIERS, EPA, IATAP, LABOEC)
Multiple Incident Command Posts and a Unified Area Command
Seek out Operational needs
Field Tested 100 new technologies; 45 were proved and implemented
Limited exposure within Planning cycle increased testing logistics difficulties
Post Macondo Study Recommendations
• Scalable Response Technology Evaluation
(RTE) Unit within the Planning Section.
• RTE Unit collects, evaluates, and tests all new
or improved technology ideas (conventional
and alternative).
• RTE Unit is the single ICS channel for public or
vendors submissions.
• Use of multi-stage, progressive technical
assessment for idea evaluation.
• The National Response Team should solicit for
the development of a database system.
• Use FEMA Responder Knowledge Base as the
“clearinghouse” to hold and disseminate RTE
assessment/test results.
Technical Work Group Members
James Fletcher - USCG
Kurt Hansen - USCG
Yvonne Addassi - OSPR
Ellen Faurot-Daniels - OSPR
Ed Levine - NOAA
Eric Koglin - EPA
Jim O’brien - Obrien’s Response
Steve Lehmann- NOAA
Ken Lukins - Lukins & Associates
Tommy Tomblin - Exxon-Mobil
Erik Demicco – Exxon-Mobil
Rene Bernier - Chevron
Hunter Rowe - BP
Mark Moran - BP
Mike Cortez - BP (Working Group lead)
Backup Slides
Current & Proposed ICS Structure
RTE Scalability Examples
RTE Unit positions added as needed,
written Incident Action Plans,
multiple operational periods (smaller
team)
Expanded RTE Unit Structure within
single ICP, multiple operational periods,
high submissions, field testing required
RTE Scalability Examples
RTE Unit Structure with multiple Incident Command Posts and when
an Area Command is established for a Tier 1 or SONS Event
Stage-Gate Review Process
DWH
ARTs
Process
•
•
•
•
•
Four-stage review and testing process (similar to diagram above)
Automated communication informing submitters of status
Scoring system suitable for ranking and prioritization
Review and scoring by technical specialists in RTE Team (separated from testing roles)
Test protocols written by technical specialists in RTE Team
Database Requirements
•
Maintain record of submittals including timestamps for all entries, changes, or
updates
•
Secure but accessible by multiple people at multiple locations
•
Tracks evaluation stages and retains testing records
•
Simple, web-based entry format
– Minimum /maximum information requirements for submissions
– Disclaimers for intellectual property (IP) issues, responsible-party (RP) issues, required
confidentiality clauses, contractual relationship
•
Automated email to submitters upon review stage decisions, to maintain
communication and establish transparency
•
Technology evaluation score archiving
•
Retention and linking of field-testing results to the submission record
Download

Alternative Response Technology API API Study