Alternative Response Technology API
API Study-Progressing Learnings
Michael J. Cortez
BP America-Oil Spill Technology Manager
RRT-3 Presentation- November 6, 2013
API Alternative Response Technology
Working Group Report
Commissioned by API as part of its Sept 2010 JITF Study
 Capture learnings from Macondo Incident
 Propose ARTES enhancements based on the learnings
Team composition: ARTES team members from Macondo
(USCG, OSPR, Obrien’s, various OSR consultants, BP) along
with NOAA, EPA and other industry entities
Kicked off study in Dec 2011; completed and endorsed July
Proposed RRT & NRT presentations for late 2013 to review
the conclusions & recommendations
Request RRT-3 Endorsement of proposed changes to ICS
Pre – Macondo ARTES
ARTES within Environmental Unit
For non-conventional response tools
Evaluate one idea/tool at a time
Used in concert with NCP product
• Limited external public engagement
• Used sporadically, therefore never truly
embedded in ICS structure
Macondo experience
120,00+ total submissions
Multiple technical reviews required
Conventional & non-conventional ideas submitted
Submissions via phone, fax, e:mail, internet, walk-up
From 100 countries in 88 languages
Multiple submission channels (PIERS, EPA, IATAP, LABOEC)
Multiple Incident Command Posts and a Unified Area Command
Seek out Operational needs
Field Tested 100 new technologies; 45 were proved and implemented
Limited exposure within Planning cycle increased testing logistics difficulties
Post Macondo Study Recommendations
• Scalable Response Technology Evaluation
(RTE) Unit within the Planning Section.
• RTE Unit collects, evaluates, and tests all new
or improved technology ideas (conventional
and alternative).
• RTE Unit is the single ICS channel for public or
vendors submissions.
• Use of multi-stage, progressive technical
assessment for idea evaluation.
• The National Response Team should solicit for
the development of a database system.
• Use FEMA Responder Knowledge Base as the
“clearinghouse” to hold and disseminate RTE
assessment/test results.
Technical Work Group Members
James Fletcher - USCG
Kurt Hansen - USCG
Yvonne Addassi - OSPR
Ellen Faurot-Daniels - OSPR
Ed Levine - NOAA
Eric Koglin - EPA
Jim O’brien - Obrien’s Response
Steve Lehmann- NOAA
Ken Lukins - Lukins & Associates
Tommy Tomblin - Exxon-Mobil
Erik Demicco – Exxon-Mobil
Rene Bernier - Chevron
Hunter Rowe - BP
Mark Moran - BP
Mike Cortez - BP (Working Group lead)
Backup Slides
Current & Proposed ICS Structure
RTE Scalability Examples
RTE Unit positions added as needed,
written Incident Action Plans,
multiple operational periods (smaller
Expanded RTE Unit Structure within
single ICP, multiple operational periods,
high submissions, field testing required
RTE Scalability Examples
RTE Unit Structure with multiple Incident Command Posts and when
an Area Command is established for a Tier 1 or SONS Event
Stage-Gate Review Process
Four-stage review and testing process (similar to diagram above)
Automated communication informing submitters of status
Scoring system suitable for ranking and prioritization
Review and scoring by technical specialists in RTE Team (separated from testing roles)
Test protocols written by technical specialists in RTE Team
Database Requirements
Maintain record of submittals including timestamps for all entries, changes, or
Secure but accessible by multiple people at multiple locations
Tracks evaluation stages and retains testing records
Simple, web-based entry format
– Minimum /maximum information requirements for submissions
– Disclaimers for intellectual property (IP) issues, responsible-party (RP) issues, required
confidentiality clauses, contractual relationship
Automated email to submitters upon review stage decisions, to maintain
communication and establish transparency
Technology evaluation score archiving
Retention and linking of field-testing results to the submission record

Alternative Response Technology API API Study