Topicality Topicality 1NC- Oceans A. Interpretation- the Earth’s Oceans are the 5 major oceans NALMS 14 – North American Lake Management Society, “WATER WORDS GLOSSARY”, http://www.nalms.org/home/publications/water-words-glossary/O.cmsx OCEAN Generally, the whole body of salt water which covers nearly three fourths of the surface of the globe. The average depth of the ocean is estimated to be about 13,000 feet (3,960 meters); the greatest reported depth is 34,218 feet (10,430 meters), north of Mindanao in the Western Pacific Ocean. The ocean bottom is a generally level or gently undulating plain, covered with a fine red or gray clay, or, in certain regions, with ooze of organic origin. The water, whose composition is fairly constant, contains on the average 3 percent of dissolved salts; of this solid portion, common salt forms about 78 percent, magnesium salts 15-16 percent, calcium salts 4 percent, with smaller amounts of various other substances. The density of ocean water is about 1.026 (relative to distilled water, or pure H2O). The oceans are divided into the Atlantic, Pacific, Indian, Arctic, and Antarctic Oceans. And, the federal government is the central government, distinguished from the states OED 89 (Oxford English Dictionary, 2ed. XIX, p. 795) b. Of or pertaining to the political unity so constituted, as distinguished from the separate states composing it. B. Violation- the IOOS is not limited to USFG action- it includes state, regional, and private sectors IOOS report to congress 13 [Official US IOOS report sent to congress. 2013, “U.S. Integrated Ocean Observing System (U.S. IOOS) 2013 Report to Congress,” http://www.ioos.noaa.gov/about/governance/ioos_report_congress2013.pdf //jweideman] U.S. IOOS works with its eyes on the future. The successes of U.S. IOOS are achieved through cooperation and coordination among Federal agencies, U.S. IOOS Regional Associations, State and regional agencies, and the private sector. This cooperation and coordination requires a sound governance and management structure. In 2011 and 2012, program milestones called for in U.S. IOOS legislation were achieved, laying the groundwork for more success in the future. First, the U.S. IOOS Advisory Committee was established. Second, the Independent Cost Estimate was delivered to Congress. As part of the estimate, each of the 11 U.S. IOOS Regional Associations completed 10-year build-out plans, describing services and products to address local user needs and outlining key assets required to meet the Nation’s greater ocean-observing needs. And, the IOOS also applies to the Great Lakes NOS and NOAA 14 [Federal Agency Name(s): National Ocean Service (NOS), National Oceanic and Atmospheric Administration (NOAA), Department of Commerce Funding Opportunity Title: FY2014 Marine Sensor and Other Advanced Observing Technologies Transition Project. “ANNOUNCEMENT OF FEDERAL FUNDING OPPORTUNITY EXECUTIVE SUMMARY,” http://www.ioos.noaa.gov/funding/fy14ffo_msi_noaa_nos_ioos_2014_2003854.pdf //jweideman] 1. Marine Sensor Transition Topic: U.S. IOOS seeks to increase the rate that new or existing marine sensor technologies are transitioned into operations mode in order to facilitate the efficient collection of ocean, coastal and Great Lakes observations. The Marine Sensor Transition topic is focused on transitioning marine sensors from research to operations mode to meet the demonstrated operational needs of end-users. Letters of Intent (LOIs) are being solicited for this topic with particular emphasis on a) projects comprised of multi-sector teams of partners, b) projects that will meet the demonstrated operational needs of end-users, and c) sensors that are at or above TRL 6. Applicants with sensors for ocean acidification that are at or above TRL 6 are also eligible to apply to this topic if they have strong commitments for operational transition C. Voting issue for fairness and ground- extra topicality forces the neg to waste time debating T just to get back to square one, and it allows the aff to gain extra advantages, counterplan answers, and link turns to disads 2NC Impact-Education Definitions are key to education about IOOS IOOS report to congress 13 [Official US IOOS report sent to congress. 2013, “U.S. Integrated Ocean Observing System (U.S. IOOS) 2013 Report to Congress,” http://www.ioos.noaa.gov/about/governance/ioos_report_congress2013.pdf //jweideman] The use of standard terms or vocabularies to describe ocean observations data is critical to facilitating broad sharing and integration of data. Working closely SECOORA and the community-based Marine Metadata Interoperability Project, U.S. IOOS has published nine recommended vocabularies over the past 12 months for review by the ocean observing community including lists for platforms, parameters, core variables, and biological terms. These efforts are helping lead the ocean observing community towards significantly improved levels of consistency via an improved semantic framework through which users can adopt recommended vocabularies or convert their vocabularies to terms that are perhaps used more widely. 2NC Cards- Oceans IOOS includes monitoring the Great Lakes IOOS report to congress 13 [Official US IOOS report sent to congress. 2013, “U.S. Integrated Ocean Observing System (U.S. IOOS) 2013 Report to Congress,” http://www.oos.noaa.gov/about/governance/ioos_report_congress2013.pdf //jweideman] The IOOC recognizes that U.S. IOOS must be responsive to environmental crises while maintaining the regular long-term ocean observation infrastructure required to support operational oceanography and climate research. As a source of our Nation’s ocean data and products, U.S. IOOS often serves as a resource for the development of targeted applications for a specific location or sector. At the same time, U.S. IOOS organizes data from across regions and sectors to foster the national and international application of local data and products broadly across oceans, coasts, and Great Lakes. Events over the last few years, including Hurricane Sandy and the Deep Water Horizon oil spill have awakened U.S. communities to the value and necessity of timely ocean information. IOOC commends U.S. IOOS for responsive and capable support to the Nation in these events in addition to diverse everyday support to the Nation’s maritime economy. We have much more work to do to build and organize the ocean-observing infrastructure of the Nateion and look forward to wrking with congress on this continuing challenge. Ocean exploration is distinct from Great Lakes observation COR 01 ~ Committee On Resources, “OCEAN EXPLORATION AND COASTAL AND OCEAN OBSERVING SYSTEMS”, Science Serial No. 107–26 Resources Serial No. 107–47, Accessed 7/3/14 //RJ On a summer day, our eyes and ears can sense an approaching thunderstorm. Our senses are extended by radar and satellites to detect advancing storm systems. Our senses are being extended yet again to anticipate changing states affecting coasts and oceans, our environment, and our climate. To truly understand the con- sequences of our actions on the environment and the environment’s impact on us, data obtained through ocean exploration, coastal observations , and ocean observa- tions will be critical. ‘‘Coastal observations’’ include observations in the Nation’s ports, bays, estuaries, Great Lakes, the waters of the EEZ, and adjacent land cover . Some of the properties measured in coastal zones, such as temperature and currents, are the same as those measured in the larger, basin-scale ocean observation systems. However, the users and applications of those data can be quite different. For those properties that are similar, there should be a consistent plan for deployment in the coastal and open ocean systems so that coastal observations represent a nested hierarchy of observa- tions collected at higher resolution than those from the open ocean. “Oceans” are only the 5 major bodies of water – landlocked and adjacent lakes and rivers are excluded. Rosenberg 14 ~ Matt Rosenberg, Master's in Geography from CSU, “Names for Water Bodies”, http://geography.about.com/od/physicalgeography/a/waterbodies.htm, accessed 7/3/14 //RJ Water bodies are described by a plethora of different names in English - rivers, streams, ponds, bays, gulfs, and seas, to name a few. Many of these terms' definitions overlap and thus become confusing when one attempts to pigeon-hole a type of water body. Read on to find out the similarities (and differences) between terms used to describe water bodies. We'll begin with the different forms of flowing water. The smallest water channels are often called brooks but creeks are often larger than brooks but may either be permanent or intermittent. Creeks are also sometimes known as streams but the word stream is quite a generic term for any body of flowing water. Streams can be intermittent or permanent and can be on the surface of the earth, underground, or even within an ocean (such as the Gulf Stream). A river is a large stream that flows over land. It is often a perennial water body and usually flows in a specific channel, with a considerable volume of water. The world's shortest river, the D River, in Oregon, is only 120 feet long and connects Devil's Lake directly to the Pacific Ocean. A pond is a small lake, most often in a natural depression. Like a stream, the word lake is quite a generic term - it refers to any accumulation of water surrounded by land - although it is often of a considerable size. A very large lake that contains salt water, is known as a sea (except the Sea of Galilee, which is actually a freshwater lake). A sea can also be attached to, or even part of, an ocean. For example, the Caspian Sea is a large saline lake surrounded by land, the Mediterranean Sea is attached to the Atlantic Ocean, and the Sargasso Sea is a portion of the Atlantic Ocean, surrounded by water. Oceans are the ultimate bodies of water and refers to the five oceans - Atlantic, Pacific, Arctic, Indian, and Southern . The equator divides the Atlantic Ocean and Pacific Oceans into the North and South Atlantic Ocean and the North and South Pacific Ocean. The plan explodes ground- includes the great lakes NOS and NOAA 14 [Federal Agency Name(s): National Ocean Service (NOS), National Oceanic and Atmospheric Administration (NOAA), Department of Commerce Funding Opportunity Title: FY2014 Marine Sensor and Other Advanced Observing Technologies Transition Project. “ANNOUNCEMENT OF FEDERAL FUNDING OPPORTUNITY EXECUTIVE SUMMARY,” http://www.ioos.noaa.gov/funding/fy14ffo_msi_noaa_nos_ioos_2014_2003854.pdf //jweideman] 1. Marine Sensor Transition Topic: U.S. IOOS seeks to increase the rate that new or existing marine sensor technologies are transitioned into operations mode in order to facilitate the efficient collection of ocean, coastal and Great Lakes observations. The Marine Sensor Transition topic is focused on transitioning marine sensors from research to operations mode to meet the demonstrated operational needs of end-users. Letters of Intent (LOIs) are being solicited for this topic with particular emphasis on a) projects comprised of multi-sector teams of partners, b) projects that will meet the demonstrated operational needs of end-users, and c) sensors that are at or above TRL 6. Applicants with sensors for ocean acidification that are at or above TRL 6 are also eligible to apply to this topic if they have strong commitments for operational transition 2NC Cards- USFG The data sharing components are the critical part of IOOS- they cant say they just don’t do the extra-topical parts IOOS report to congress 13 [Official US IOOS report sent to congress. 2013, “U.S. Integrated Ocean Observing System (U.S. IOOS) 2013 Report to Congress,” http://www.ioos.noaa.gov/about/governance/ioos_report_congress2013.pdf //jweideman] Observations are of little value if they cannot be found, accessed, and transformed into useful products. The U.S. IOOS Data Management and Communications subsystem, or “DMAC,” is the central operational infrastructure for assessing, disseminating, and integrating existing and future ocean observations data. As a core functional component for U.S. IOOS, establishing DMAC capabilities continues to be a principal focus for the program and a primary responsibility of the U.S. IOOS Program Office in NOAA. Importance and Objectives of DMAC Although DMAC implementation remains a work in progress, a fully implemented DMAC subsystem will be capable of delivering real-time, delayed-mode, and historical data. The data will include in situ and remotely sensed physical, chemical, and biological observations as well as modelgenerated outputs, including forecasts, to U.S. IOOS users and of delivering all forms of data to and from secure archive facilities. Achieving this requires a governance framework for recommending and promoting standards and policies to be implemented by data providers across the U.S. IOOS enterprise, to provide seamless long-term preservation and reuse of data across regional and national boundaries and across disciplines. The governance framework includes tools for data access, distribution, discovery, visualization, and analysis; standards for metadata, vocabularies, and quality control and quality assurance; and procedures for the entire ocean data life cycle. The DMAC design must be responsive to user needs and it must, at a minimum, make data and products discoverable and accessible, and provide essential metadata regarding sources, methods, and quality. The overall DMAC objectives are for U.S. IOOS data providers to develop and maintain capabilities to: • Deliver accurate and timely ocean observations and model outputs to a range of consumers; including government, academic, private sector users, and the general public; using specifications common across all providers • Deploy the information system components (including infrastructure and relevant personnel) for full life-cycle management of observations, from collection to product creation, public delivery, system documentation, and archiving • Establish robust data exchange responsive to variable customer requirements as well as routine feedback, which is not tightly bound to a specific application of the data or particular end-user decision support tool U.S. IOOS daia providers therefore are being the following DMAC- specific objectives: • A standards-based foundation for DMAC capabilities: U.S. IOOS partners must clearly demonstrate how they will ensure the establishment and maintenance of a standards- based approach for delivering their ocean observations data and associated products to users through local, regional and global/international data networks • Exposure of and access to coastal ocean observations: U.S. IOOS partners must describe how they will ensure coastal ocean observations are exposed to users via a service- oriented architecture and recommended data services that will ensure increased data interoperability including the use of improved metadata and uniform quality-control methods • Certification and governance of U.S. IOOS data and products: encouraged lo address U.S. IOOS partners must present a description of how they will participate in establishing an effective U.S. IOOS governance process for data certification standards and compliance procedures. This objective is part of an overall accreditation process which includes the other U.S. IOOS subsystems (observing, modeling and analysis, and governance) “Federal Government” means the United States government Black’s Law 99 (Dictionary, Seventh Edition, p.703) The U.S. government—also termed national government National government, not states or localities Black’s Law 99 (Dictionary, Seventh Edition, p.703) A national government that exercises some degree of control over smaller political units that have surrendered some degree of power in exchange for the right to participate in national political matters Central government AHD 92 (American Heritage Dictionary of the English Language, p. 647) federal—3. Of or relating to the central government of a federation as distinct from the governments of its member units. IOOS includes state, regional, and private sectors IOOS report to congress 13 [Official US IOOS report sent to congress. 2013, “U.S. Integrated Ocean Observing System (U.S. IOOS) 2013 Report to Congress,” http://www.ioos.noaa.gov/about/governance/ioos_report_congress2013.pdf //jweideman] U.S. IOOS works with its eyes on the future. The successes of U.S. IOOS are achieved through cooperation and coordination among Federal agencies, U.S. IOOS Regional Associations, State and regional agencies, and the private sector. This cooperation and coordination requires a sound governance and management structure. In 2011 and 2012, program milestones called for in U.S. IOOS legislation were achieved, laying the groundwork for more success in the future. First, the \U.S. IOOS Advisory Committee was established. Second, the Independent Cost Estimate was delivered to Congress. As part of the estimate, each of the 11 U.S. IOOS Regional Associations completed 10-year build-out plans, describing services and products to address local user needs and outlining key assets required to meet the Nation’s greater ocean-observing needs. Disads Politics 1NC – Politics Link Causes fights – anti-environmentalism Farr 2013 Sam, Member of the House of Representatives (D-CA) Chair of House Oceans Caucus, Review&Forecast Collaboration Helps to Understand And Adapt to Ocean, Climate Changes http://seatechnology.com/pdf/st_0113.pdf The past year in Washington, D.C., was rife with infighting between the two political parties. On issue after issue, the opposing sides were unable to reach a compromise on meaningful legislation for the American people. This division was most noticeable when dis- cussing the fate of our oceans. The widening chasm between the two political parties resulted in divergent paths for ocean policy: one with Presi- dent Barack Obama pushing the National Ocean Policy forward, and the other with U.S. House Republicans under- mining those efforts with opposing votes and funding cuts. Marine, Environmental Policy Regression The 112th Congress was called the "most antienviron- mental Congress in history" in a report published by House Democrats and has been credited for undermining the ma- jor environmental legislation of the past 40 years. After the Tea Party landslide in the congressional elections of 2010, conservatives on Capitol Hill began to flex their muscles to roll back environmental protections. Since taking power in January 2011, House Republicans held roughly 300 votes to undermine basic environmental protections that have existed for decades. To put that in per- spective, that was almost one in every five votes held in Congress during the past two years. These were votes to al- low additional oil and gas drilling in coastal waters, while simultaneously limiting the environmental review process for offshore drilling sites. There were repeal attempts to un- dermine the Clean Water Act and to roll back protections for threatened fish and other marine species. There were also attempts to block measures to address climate changes, ig- noring the consequences of inaction, such as sea level rise and ocean acidification. 2NC – Obama PC Link Costs political capital, requires Obama to push Farr 2013 Sam, Member of the House of Representatives (D-CA) Chair of House Oceans Caucus, Review&Forecast Collaboration Helps to Understand And Adapt to Ocean, Climate Changes http://seatechnology.com/pdf/st_0113.pdf Bui while the tide of sound ocean policy was retreating in Congress, it was rising on the other end of Pennsylvania Avenue. The National Ocean Policy Understanding the need for a federal policy to address the oceans after the Deepwater Horizon disaster, President Obama used an executive order in 2010 to create the Na- tional Ocean Policy. Modeled after the Oceans-21 legisla- tion I championed in Congress, this order meant that the U.S. would have a comprehensive plan to guide our stew- ardship of the oceans, coasts and Great Lakes. Despite push back from the House of Representatives, the president continued implementing his National Ocean Policy this past year through existing law, prioritizing ocean stewardship and promoting interagency collaboration to ad- dress the challenges facing our marine environment. For in- stance, the National Oceans Council created the www.data. gov/ocean portal, which went live just before the start of 2012 and has grown throughout the year to provide access to data and information about the oceans and coasts. The U.S. Geological Survey and other National Ocean Council agencies have been working with federal and state partners to identify the data sets needed by ocean users, such as fish- eries and living marine resources data, elevation and shore- line information, and ocean observations. 2NC – NOAA Link Congress hates increased funding for NOAA programs---the plan would be a fight Andrew Jensen 12, Peninsula Clarion, Apr 27 2012, “Congress takes another ax to NOAA budget,” http://peninsulaclarion.com/news/2012-04-27/congress-takes-another-ax-to-noaa-budget Frustrated senators from coastal states are wielding the power of the purse to rein in the N ational O ceanic and A tmospheric A dministration and refocus the agency's priorities on its core missions.¶ During recent appropriations subcommittee hearings April 17, Sen. Lisa Murkowski ensured no funds would be provided in fiscal year 2013 for coastal marine spatial planning, a key component of President Barack Obama's National Ocean Policy.¶ Murkowski also pushed for an additional $3 million for regional fishery management councils and secured $15 million for the Pacific Salmon Treaty that was in line to be cut by NOAA's proposed budget (for $65 million total).¶ On April 24, the full Senate Appropriations Committee approved the Commerce Department budget with language inserted by Sen. John Kerry, D-Mass., and Sen. Olympia Snowe, R-Maine, into NOAA's budget that would transfer $119 million currently unrestricted funds and require they be used for stock assessments, surveys and monitoring, cooperative research and fisheries grants.¶ The $119 million is derived from Saltonstall-Kennedy funds, which are levies collected on seafood imports by the Department of Agriculture. Thirty percent of the import levies are transferred to NOAA annually, and without Kerry's language there are no restrictions on how NOAA may use the funds.¶ In a Congress defined by fierce partisanship, no federal agency has drawn as much fire from both parties as NOAA and its Administrator Jane Lubchenco.¶ Sen. Scott Brown, R-Mass., has repeatedly demanded accountability for NOAA Office of Law Enforcement abuses uncovered by the Commerce Department Inspector General that included the use of fishermen's fines to purchase a luxury boat that was only used for joyriding around Puget Sound.¶ There is currently another Inspector General investigation under way into the regional fishery management council rulemaking process that was requested last August by Massachusetts Reps. John Tierney and Barney Frank, both Democrats.¶ In July 2010, both Frank and Tierney called for Lubchenco to step down, a remarkable statement for members of Obama's party to make about one of his top appointments.¶ Frank introduced companion legislation to Kerry's in the House earlier this year, where it should sail through in a body that has repeatedly stripped out tens of millions in budget requests for catch share programs. Catch share programs are Lubchenco's favored policy for fisheries management and have been widely panned after implementation in New England in 2010 resulted in massive consolidation of the groundfish catch onto the largest fishing vessels.¶ Another New England crisis this year with Gulf of Maine cod also drove Kerry's action after a two-year old stock assessment was revised sharply downward and threatened to close down the fishery. Unlike many fisheries in Alaska such as pollock, crab and halibut, there are not annual stock assessment surveys around the country.¶ Without a new stock assessment for Gulf of Maine cod, the 2013 season will be in jeopardy.¶ "I applaud Senator Kerry for his leadership on this issue and for making sure that this funding is used for its intended purpose - to help the fishing industry, not to cover NOAA's administrative overhead," Frank said in a statement. "We are at a critical juncture at which we absolutely must provide more funding for cooperative fisheries science so we can base management policies on sound data, and we should make good use of the world-class institutions in the Bay State which have special expertise in this area."¶ Alaska's Sen. Mark Begich and Murkowski, as well as Rep. Don Young have also denounced the National Ocean Policy as particularly misguided, not only for diverting core funding in a time of tightening budgets but for creating a massive new bureaucracy that threatens to overlap existing authorities for the regional fishery management councils and local governments.¶ The first 92 pages of the draft policy released Jan. 12 call for more than 50 actions, nine priorities, a new National Ocean Council, nine Regional Planning Bodies tasked with creating Coastal Marine Spatial Plans, several interagency committees and taskforces, pilot projects, training in ecosystem-based management for federal employees, new water quality standards and the incorporation of the policy into regulatory and permitting decisions.¶ Some of the action items call for the involvement of as many as 27 federal agencies. Another requires high-quality marine waters to be identified and new or modified water quality and monitoring protocols to be established.¶ Young hosted a field hearing of the House Natural Resources Committee in Anchorage April 3 where he blasted the administration for refusing to explain exactly how it is paying for implementing the National Ocean Policy.¶ "This National Ocean Policy is a bad idea," Young said. "It will create more uncertainty for businesses and will limit job growth. It will also compound the potential for litigation by groups that oppose human activities. To make matters worse, the administration refuses to tell Congress how much money it will be diverting from other uses to fund this new policy."¶ Natural Resources Committee Chairman Doc Hastings, R-Wash., sent a letter House Appropriations Committee Chairman Hal Rogers asking that every appropriations bill expressly prohibit any funds to be used for implementing the National Ocean Policy.¶ Another letter was sent April 12 to Rogers by more than 80 stakeholder groups from the Gulf of Mexico to the Bering Sea echoing the call to ban all federal funds for use in the policy implementation.¶ "The risk of unintended economic and societal consequences remains high, due in part to the unprecedented geographic scale under which the policy is to be established," the stakeholder letter states. "Concerns are further heightened because the policy has already been cited as justification in a federal decision restricting access to certain areas for commercial activity."¶ million in budget requests for NOAA in fiscal year 2012 implementation policy in January anyway. Congress refused to fund some $27 to implement the National Ocean Policy, but the administration released its draft Russian Oil Russian Oil DA 1NC Russia’s economy is at a tipping point- its holding on because of oil The economist 14 [The economist is an economic news source. May 3 2014, “The crisis in Ukraine is hurting an already weakening economy” http://www.economist.com/news/finance-and-economics/21601536-crisis-ukraine-hurtingalready-weakening-economy-tipping-scales //jweideman] WESTERN measures against Russia—asset freezes and visa restrictions aimed at people and firms close to Vladimir Putin—may be pinpricks, but the crisis in Ukraine has already taken its toll on Russia’s economy and financial markets. Capital flight in the first three months of 2014 is thought to exceed $60 billion. The stockmarket is down by 20% since the start of the year and the rouble has dropped by 8% against the dollar. Worries about the devaluation feeding through to consumer prices have prompted the central bank to yank up interest rates, from 5.5% at the start of March to 7.5%. The IMF reckons the economy is in recession; this week it cut its growth forecast for 2014 from 1.3% to 0.2%. Despite these upsets, Russia appears to hold strong economic as well as military cards. It provides 24% of the European Union’s gas and 30% of its oil. Its grip on Ukraine’s gas and oil consumption is tighter still. That makes it hard for the West to design sanctions that do not backfire. Satellite mapping finds oil- it increases production capabilities Short 11 [Nicholas M. Short, Sr is a geologist who received degrees in that field from St. Louis University (B.S.), Washington University (M.A.), and the Massachusetts Institute of Technology (Ph.D.); he also spent a year in graduate studies in the geosciences at The Pennsylvania State University. In his early postgraduate career, he worked for Gulf Research & Development Co., the Lawrence Livermore Laboratory, and the University of Houston. During the 1960s he specialized in the effects of underground nuclear explosions and asteroidal impacts on rocks (shock metamorphism), and was one of the original Principal Investigators of the Apollo 11 and 12 moon rocks. 2011, “Finding Oil and Gas from Space” https://apollomapping.com/wpcontent/user_uploads/2011/11/NASA_Remote_Sensing_Tutorial_Oil_and_Gas.pdf //jweideman] If precious metals are not your forte, then try the petroleum industry. Exploration for oil and gas has always depended on surface maps of rock types and structures that point directly to, or at least hint at, subsurface conditions favorable to accumulating oil and gas. Thus, looking at surfaces from satellites is a practical, cost-effective way to produce appropriate maps. But verifying the presence of hydrocarbons below surface requires two essential steps: 1) doing geophysical surveys; and 2) drilling into the subsurface to actually detect and extract oil or gas or both. This Tutorial website sponsored by the Society of Exploration Geophysicists is a simplified summary of the basics of hydrocarbon exploration. Oil and gas result from the decay of organisms - mostly marine plants (especially microscopic algae and similar free-floating vegetation) and small animals such as fish - that are buried in muds that convert to shale. Heating through burial and pressure from the overlying later sediments help in the process. (Coal forms from decay of buried plants that occur mainly in swamps and lagoons which are eventually buried by younger sediments.). The decaying liquids and gases from petroleum source beds, dominantly shales after muds convert to hard rock, migrate from their sources to become trapped in a variety of structural or stratigraphic conditions shown in this illustration:The oil and gas must migrate from deeper source beds into suitable reservoir rocks. These are usually porous sandstones, but limestones with solution cavities and even fractured igneous or metamorphic rocks can contain openings into which the petroleum products accumulate. An essential condition: the reservoir rocks must be surrounded (at least above) by impermeable (refers to minimal ability to allow flow through any openings - pores or fractures) rock, most commonly shales. The oil and gas, generally confined under some pressure, will escape to the surface - either naturally when the trap is intersected by downward moving erosional surfaces or by being penetrated by a drill. If pressure is high the oil and/or gas moves of its own accord to the surface but if pressure is initially low or drops over time, pumping is required. Exploration for new petroleum sources begins with a search for surface manifestations of suitable traps (but many times these are hidden by burial and other factors govern the decision to explore). Mapping of surface conditions begins with reconnaissance, and if that indicates the presence of hydrocarbons, then detailed mapping begins. Originally, both of these maps required field work. Often, the mapping job became easier by using aerial photos. After the mapping, much of the more intensive exploration depends on geophysical methods (principally, seismic) that can give 3-D constructions of subsurface structural and stratigraphic traps for the hydrocarbons. Then, the potential traps are sampled by exploratory drilling and their properties measured. Remote sensing from satellites or aircraft strives to find one or more indicators of surface anomalies. This diagram sets the framework for the approach used; this is the so-called microseepage model, which leads to specific geochemical anomalies: New US oil production collapses Russia’s economy Woodhill 14 [Louis, Contributor at forbes, entrepenour and investor. 3/3/14, “It's Time To Drive Russia Bankrupt – Again” http://www.forbes.com/sites/louiswoodhill/2014/03/03/its-time-to-drive-russia-bankrupt-again/ //jweideman] The high oil prices of 1980 were not real, and Reagan knew it. They were being caused by the weakness of the U.S. dollar, which had lost 94% of its value in terms of gold between 1969 and 1980. Reagan immediately decontrolled U.S. oil prices, to unleash the supply side of the U.S. economy. Even more importantly, Reagan backed Federal Reserve Chairman Paul Volcker’s campaign to strengthen and stabilize the U.S. dollar. By the end of Reagan’s two terms in office, real oil prices had plunged to $27.88/bbl. As Russia does today, the old USSR depended upon oil exports for most of its foreign exchange earnings, and much of its government revenue. The 68% reduction in real oil prices during the Reagan years drove the USSR bankrupt. In May 1990, Gorbachev called German Chancellor Helmut Kohl and begged him for a loan of $12 billion to stave off financial disaster. Kohl advanced only $3 billion. By August of 1990, Gorbachev was back, pleading for more loans. In December 1991, the Soviet Union collapsed. President Bill Clinton’s “strong dollar” policy (implemented via Federal Reserve Vice-Chairman Wayne Angell’s secret commodity price rule system) kept real oil prices low At real oil price levels like this, Russia is financially incapable of causing much trouble. It was George W. Bush and Barack Obama’s feckless “weak dollar” policy that let the Russian geopolitical genie out of the bottle. From the end of during the 1990s, despite rising world oil demand. Real crude oil prices during Clinton’s time in office averaged only $27.16/bbl. 2000 to the end of 2013, the gold value of the dollar fell by 77%, and real oil prices tripled, to $111.76/bbl. It is these artificially high oil prices that are fueling Putin’s mischief machine. The Russian government has approved a 2014 budget calling for revenues of $409.6 billion, spending of $419.6 billion, and a deficit of $10.0 billion, or 0.4% of expected GDP of $2.5 trillion. Unlike the U.S., which has deep financial markets and prints the world’s reserve currency, Russia cannot run large fiscal deficits without creating hyperinflation. Given that Russia expects to get about half of its revenue from taxes on its oil and gas industry, it is clear that it would not take much of a decline in world oil prices to create financial difficulties for Russia. Assuming year-end 2013 prices for crude oil ($111.76/bbl) and natural gas ($66.00/FOE* bbl) the total revenue of Russia’s petroleum industry is $662.3 billion (26.5% of GDP), and Russian’s oil and gas export earnings are $362.2 billion, or 14.5% of GDP. Obviously, a decline in world oil prices would cause the Russian economy and the Russian government significant financial pain. Nuclear war Filger ‘9 [Sheldon, founder of Global Economic Crisis, The Huffington Post,, 5.10.9, http://www.huffingtonpost.com/sheldon-filger/russian-economy-facesdis_b_201147.html // jweideman] In Russia, historically, economic health and political stability are intertwined to a degree that is rarely encountered in other major industrialized economies. It was the economic stagnation of the former Soviet Union that led to its political downfall. Similarly, Medvedev and Putin, both intimately acquainted with their nation's history, are unquestionably alarmed at the prospect that Russia's economic crisis will endanger the nation's political stability, achieved at great cost after years of chaos following the demise of the Soviet Union. Already, strikes and protests are occurring among rank and file workers facing unemployment or non-payment of their salaries. Recent polling demonstrates that the once supreme popularity ratings of Putin and Medvedev are eroding rapidly. Beyond the political elites are the financial oligarchs, who have been forced to deleverage, even unloading their yachts and executive jets in a desperate attempt to raise cash. Should the Russian economy deteriorate to the point where economic collapse is not out of the question, the impact will go far beyond the obvious accelerant such an outcome would be for the Global Economic Crisis. There is a geopolitical dimension that is even more relevant then the economic context . Despite its economic vulnerabilities and perceived decline from superpower status, Russia remains one of only two nations on earth with a nuclear arsenal of sufficient scope and capability to destroy the world as we know it. For that reason, it is not only President Medvedev and Prime Minister Putin who will be lying awake at nights over the prospect that a national economic crisis can transform itself into a virulent and destabilizing social and political upheaval. It just may be possible that U.S. UQ: econ Russia’s economy is weak, but growth is coming ITARR-TASS 7/3 [Russian news agency. 7/3/14, “Russia’s economic growth to accelerate to over 3% in 2017” http://en.itar-tass.com/economy/738849 //jweideman] The Russian government expects the national economy to grow, despite the unfavorable economic scenario for this year, Prime Minister Dmitry Medvedev said on Thursday. Medvedev chaired a government meeting on Thursday to discuss the federal budget parameters and state program financing until 2017. “This scenario (of Russia’s social and economic development) unfortunately envisages a general economic deterioration and slower economic growth this year. Then, the growth is expected to accelerate gradually to 2% next year and to over 3% in 2017,” Medvedev said. The 2015-2017 federal budget will set aside over 2.7 trillion rubles ($78 billion) for the payment of wages and salaries, the premier said, adding the government’s social obligations were a priority in the forthcoming budget period. More than a half of federal budget funds are spent through the mechanism of state programs. In 2015, the government intends to implement 46 targeted state programs to the amount of over 1 trillion rubles (about $30 billion) in budget financing, the premier said. The same budget financing principle will now be used by regional and local governments, including Russia’s two new constituent entities - the Republic of Crimea and the city of Sevastopol, the premier said. Russia will avoid recession AFP 7/10 [Agence France-Presse is an international news agency headquartered in Paris. It is the oldest news agency in the world and one of the largest. 7/10/14, “Russia avoids recession in second quarter” https://au.news.yahoo.com/thewest/business/world/a/24425122/russia-avoids-recession-in-second-quarter/ //jweideman] Moscow (AFP) - Russia managed to avoid sliding into a recession in the second quarter, its deputy economy minister said Wednesday, citing preliminary estimates showing zero growth for the three months ending June. The emerging giant was widely expected to post a technical recession with two consecutive quarters of contraction, after massive capital flight over the uncertainty caused by the Ukraine crisis. "We were expecting a possible technical recession in the second quarter," Russia's deputy economy minister Andrei Klepach said, according to Interfax news agency. But it now "appears that we will avoid recession, our preliminary forecast is one of zero growth after adjusting for seasonal variations," he said. Official data is due to be published by the statistics institute during the summer, although no specific date has been given. Russia's growth slowed down sharply in the first quarter as investors, fearing the impact of Western sanctions over Moscow's annexation of Crimea, withdrew billions worth of their assets. In the first three months of the year, output contracted by 0.3 percent compared to the previous quarter. Klepach estimated that on an annual basis, Russia's annual growth would better the 0.5 percent previously forecast. "The situation is unstable, but we think... that the trend would be positive despite everything," he said, adding however that it would be a "slight recovery rather than solid growth". The government has until now not issued any quarter to quarter growth forecast. Economy minister Alexei Ulyukayev on Monday issued a year-on-year forecast, saying output expanded by 1.2 percent in the second quarter of the year. Russia is experiencing growth AFP 7/7 [Agence France-Presse is an international news agency headquartered in Paris. It is the oldest news agency in the world and one of the largest. 7/7/14, “Russian output expands 1.2% in second quarter: economy minister” http://www.globalpost.com/dispatch/news/afp/140707/russian-output-expands-12-secondquarter-economy-minister-0 //jweideman] Russia's economy minister said Monday that output expanded by 1.2 percent in the second quarter compared to the same period last year, a preliminary figure that is "slightly better" than expected despite the Ukraine crisis. "The results are slightly better than we predicted, with the emphasis on 'slightly'," economy minister Alexei Ulyukayev said. He added that the "refined" official figure by Russia's statistics agency will be released later. The IMF said last month that Russia is already in recession, while the central bank said growth in 2014 is likely to slow to just 0.4 percent. Ulyukayev said that second quarter figure, up from 0.9 percent in the first quarter and 0.2 percent better than his ministry's expectations, was likely bolstered by industrial growth. "This year industry is becoming the main locomotive of growth," Ulyukayev told President Vladimir Putin in a meeting, citing industrial growth reaching 2.5 percent, up from 1.8 percent in the first quarter. UQ AT: sanctions There won’t be more sanctions Al Jazeera 14 [Al Jazeera also known as Aljazeera and JSC (Jazeera Satellite Channel), is a Doha-based broadcaster owned by the Al Jazeera Media Network, June 5 2014. “G7 holds off from further Russia sanctions” http://www.aljazeera.com/news/europe/2014/06/g7-hold-off-from-further-russia-sanctions20146545044590707.html //jweideman] Leaders of the G7 group of industrial nations have limited themselves to warning Russia of another round of sanctions as they urged President Vladimir Putin to stop destabilising Ukraine. On the first day of the group's summit in Brussels on Wednesday, the bloc said that Putin must pull Russian troops back from his country's border with Ukraine and stem the pro-Russian uprising in the east of the country."Actions to destabilise eastern Ukraine are unacceptable and must stop," said the group in a joint statement after the meeting. "We stand ready to intensify targeted sanctions and to implement significant additional restrictive measures to impose further costs on Russia should events so require." However, Angela Merkel, the German chancellor, said that further sanctions would take effect only if there had been "no progress whatsoever". Al Jazeera's James Bays, reporting from Brussels, said: "The G7 leader are talking here again about tougher sanctions on Russia. But they are talking about the same tougher sanctions they have been talking about for months now." Meanwhile, Putin reached out a hand towards dialogue, despite being banned from the summit following Russia's annexation of Crimea in March, saying that he was ready to meet Ukraine's president-elect Petro Poroshenko and US President Barack Obama. U-Production decline Oil production is declining Summers 14 [Dave, author for economonitor, july 7 14, “Oil Production Numbers Keep Going Down” http://www.economonitor.com/blog/2014/07/oilproduction-numbers-keep-going-down/ //jweideman] One problem with defining a peak in global oil production is that it is only really evident sometime after the event, when one can look in the rear view mirror and see the transition from a growing oil supply to one that is now declining. Before that relatively absolute point, there will likely come a time when global supply can no longer match the global demand for oil that exists at that price. We are beginning to approach the latter of these two conditions, with the former being increasingly probable in the non-too distant future. Rising prices continually change this latter condition, and may initially disguise the arrival of the peak, but it is becoming inevitable. Over the past two years there has been a steady growth in demand, which OPEC expects to continue at around the 1 mbd range, as has been the recent pattern. The challenge, on a global scale, has been to identify where the matching growth in supply will come from, given the declining production from older oilfields and the decline rate of most of the horizontal fracked wells in shale. At present the United States is sitting with folk being relatively complacent, anticipating that global oil supplies will remain sufficient, and that the availability of enough oil in the global market to supply that reducing volume of oil that the US cannot produce for itself will continue to exist. Increasingly over the next couple of years this is going to turn out to have created a false sense of security, and led to decisions on energy that will not easily be reversed. Consider that the Canadians have now decided to build their Pipeline to the Pacific. The Northern Gateway pipeline that Enbridge will build from the oil sands to the port of Kitimat. Oil production is decreasing off the coast Larino 14 [Jennifer, Staff writer for NOLA.com | The Times-Picayune covering energy, banking/finance and general business news in the greater New Orleans area. 7/8/14, “Oil, gas production declining in Gulf of Mexico federal waters, report says” http://www.nola.com/business/index.ssf/2014/07/oil_gas_production_in_federal.html //jweideman] Oil and gas found off the coast of Louisiana and other Gulf Coast states made up almost one quarter of all fossil fuel production on federal lands in 2013, reinforcing the region's role as a driving force in the U.S. energy industry, according to updated government data. But a closer look at the numbers shows the region's oil and gas production has been in steady decline for much of the past decade. A new U.S. Energy Information Administration report shows federal waters in the Gulf of Mexico in 2013 accounted for 23 percent of the 16.85 trillion British thermal units (Btu) of fossil fuels produced on land and water owned by the federal government. That was more than any other state or region aside from Wyoming, which has seen strong natural gas production in recent years. The report did not include data on oil and gas production on private lands, which makes up most production in many onshore oil and gas fields, including the Haynesville Shale in northwest Louisiana. But production in the offshore gulf has also fallen every year since 2003. According to the report, total fossil fuel production in the region is less than half of what it was a decade ago, down 49 percent from 7.57 trillion Btu in 2003 to 3.86 trillion in 2013. The report notes that the region has seen a sharp decline in natural gas production as older offshore fields dry out and more companies invest in newer gas finds onshore, where hydraulic fracturing has led to a boom production. Natural gas production in the offshore gulf was down 74 percent from 2003 to 2013. The region's oil production has declined, though less drastically. The offshore gulf produced about 447 million barrels of oil in 2013, down from a high of 584 million barrels in 2010. Still, the region accounted for 69 percent of all the crude oil produced on all federal lands and waters last year. Offshore well exploration is ineffective and expensive now API 14 [The American Petroleum Institute, commonly referred to as API, is the largest U.S trade association for the oil and natural gas industry. January 2014, “Offshore Access to Oil and Gas” http://www.api.org/policy-and-issues/policy-items/exploration/~/media/Files/Oil-and-Natural-Gas/Offshore/OffshoreAccessprimer-highres.pdf//jweideman] Sometimes when a lease is not producing, critics claim it is “idle.” Much more often than not, non-producing leases are not idle at all; they are under geological evaluation or in development and could become an important source of domestic supply. Companies purchase leases hoping they will hold enough oil or natural gas to benefit consumers and become economically viable for production. Companies can spend millions of dollars to purchase a lease and then explore and develop it, only to find that it does not contain oil and natural gas in commercial quantities. It is not unusual for a company to spend in excess of $100 million only to drill a dry hole. The reason is that a company usually only has limited knowledge of resource potential when it buys a lease. Only after the lease is acquired will the company be in a position to evaluate it, usually with a very costly seismic survey followed by an exploration well. If a company does not find oil or natural gas in commercial quantities, the company hands the lease back to the government, incurs the loss of invested money and moves on to more promising leases. If a company finds resources in commercial quantities, it will produce the lease. But there sometimes can be delays — often as long as ten years — for environmental and engineering studies, to acquire permits, to install production facilities (or platforms for offshore leases) and to build the necessary infrastructure to bring the resources to market. Litigation, landowner disputes and regulatory hurdles also can delay the process . Oil is getting harder to find Spak 11 [Kevin, staff writer for newser. 2/16/11, “Exxon: Oil Becoming Hard to Find” http://www.newser.com/story/112166/exxon-oil-becoming-hard-tofind.html //jweideman] Exxon Mobil Corp. is having trouble finding more oil, it revealed yesterday in its earnings call. The company said it was depleting its reserves, replacing only 95 out of every 100 barrels it pumps. Though it tried to put a positive spin on things by saying it had made up for the shortfall by bolstering its natural gas supplies, experts tell the Wall Street Journal that’s probably a shift born of grim necessity. Oil companies across the board are drifting toward natural gas, because oil is harder and harder to come by, according to the Journal. Most accessible fields are already tapped out, while new fields are either technically or politically difficult to exploit. “The good old days are gone and not to be repeated,” says one analyst. Natural gas “is not going to give you the same punch” as oil. 2NC Link Wall Satellites prevent rig destroying eddys- empirically solves production ESA 6 [European space agency, 1 March 2006, “SATELLITE DATA USED TO WARN OIL INDUSTRY OF POTENTIALLY DANGEROUS EDDY” http://www.esa.int/Our_Activities/Observing_the_Earth/Satellite_data_used_to_ warn_oil_industry_of_potentially_dangerous_eddy//jweideman] Ocean FOCUS began issuing forecasts on 16 February 2006 – just in time to warn oil production operators of a new warm eddy that has formed in the oil and gas-producing region of the Gulf of Mexico. These eddies, similar to underwater hurricanes, spin off the Loop Current – an intrusion of warm surface water that flows northward from the Caribbean Sea through the Yucatan Strait – from the Gulf Stream and can cause extensive and costly damage to underwater equipment due to the extensive deep water oil production activities in the region.The Ocean FOCUS service is a unique service that provides ocean current forecasts to the offshore oil production industry to give prior warning of the arrival of eddies. The service is based on a combination of state-of-the-art ocean models and satellite measurements. Oil companies require early warning of these eddies in order to minimise loss of production, optimise deep water drilling activities and prevent damage to critical equipment. The Loop Current and eddies shedding from it pose two types of problems for underwater production systems: direct force and induced vibrations, which create more stress than direct force and results in higher levels of fatigue and structural failure. The impact of these eddies can be very costly in terms of downtime in production and exploration and damage to sub sea components. IOOS leads to more effective production NOAA 13 [U.S. Department of Commerce www.ioos.noaa.gov May 2013 National Oceanic and Atmospheric Administration Integrated Ocean Observing System (IOOS®) Program. May 2013, “IOOS®: Partnerships Serving Lives and Livelihoods” http://www.ioos.noaa.gov/communications/handouts/partnerships_lives_livelihoods_flyer.pdf //jweideman] IOOS supplies critical information about our Nation's waters. Scientists working to understand climate change, governments adapting to changes in the Arctic, -nunicipalities monitoring local water quality, industries jnderstanding coastal and marine spatial planning all have the same need: reliable and timely access to data and information that informs decision making. Improving Lives and Livelihoods IOOS enhances our economy. Improved information allows offshore oil and gas platform and coastal operators, municipal Dlanners, and those with interests n the coastal zone to minimize impacts of natural hazards, sea level rise, and flooding. This infor- mation improves marine forecasts so mariners can optimize shipping routes, saving time and reducing fuel expenses - translating into cost savings for consumers.IOOS benefits our safety and environment. A network of water quality monitoring buoys on the Great Lakes makes beaches safer □y detecting and predicting the Dresence of bacteria (E. coli). Mariners use IOOS wave and surface current data to navigate ships safely under bridges and in narrow channels. Satellites are effective at finding new oil wells DTU 9 [Technical University of Denmark (DTU) 2/27/9, “New Oil Deposits Can Be Identified Through Satellite Images” http://www.sciencedaily.com/releases/2009/02/090226110812.htm //jweideman] A new map of the Earth’s gravitational force based on satellite measurements makes it much less resource intensive to find new oil deposits. The map will be particularly useful as the ice melts in the oil-rich Arctic regions. Ole Baltazar, senior scientist at the National Space Institute, Technical University of Denmark (DTU Space), headed the development of the map. The US company Fugro, one of the world’s leading oil exploration companies, is one of the companies that have already made use of the gravitational map. The company has now initiated a research partnership with DTU Space. “Ole Baltazar’s gravitational map is the most precise and has the widest coverage to date,” says Li Xiong, Vice President and Head Geophysicist with Fugro. “On account of its high resolution and accuracy, the map is particularly useful in coastal areas, where the majority of the oil is located.” Satellite measurements result in high precision Ole Baltazar’s map shows variations in gravitational force across the surface of the Earth and knowledge about these small variations is a valuable tool in oil exploration. Subterranean oil deposits are encapsulated in relatively light materials such as limestone and clay and because these materials are light, they have less gravitational force than the surrounding materials. Ole Baltazar’s map is based on satellite measurements and has a hitherto unseen level of detail and accuracy. With this map in your hands, it is, therefore, easier to find new deposits of oil underground. Climate change is revealing new sea regions The gravitational map from DTU Space is unique on account of its resolution of only 2 km and the fact that it covers both land and sea regions. Oil companies use the map in the first phases of oil exploration. Previously, interesting areas were typically selected using protracted, expensive measurements from planes or ships. The interesting areas appear clearly on the map and the companies can, therefore, plan their exploration much more efficiently. “The map will also be worth its weight in gold when the ice in the Arctic seriously begins to melt, revealing large sea regions where it is suspected that there are large deposits of oil underground. With our map, the companies can more quickly start to drill for oil in the right places without first having to go through a resource-intensive exploration process,” explains Ole Baltazar. Based on height measurements instead of direct gravitation measurements The success of the gravitational map is due in large part to the fact that it is not based on direct gravitation measurements but on observations of the height of the sea, which reflects the gravitation. “Height measurements have the advantage that it is possible to determine the gravitational field very locally and thus make a gravitational map with a resolution of a few km. For comparison, the resolution of satellite measurements of gravitational force is typically around 200 km. Satellite gravitation measurements are used, for example, to explore conditions in the deeper strata of the Earth, but are not well suited to our purposes,” Ole Baltazar explains. IL-Prices k2 russia’s economy High oil prices are key to political and economic stability in Russia Ioffe 12 [Julia, Foreign policy’s Moscow correspondant. 6/12/12, “What will it take to push russians over the edge” http://www.foreignpolicy.com/articles/2012/06/12/powder_keg?page=0,1 //jweideman] There is also the economic factor to consider. The Russian economy is currently growing at a relatively healthy 3.5 percent, but it's useful to recall the whopping growth rates Russia was posting just a few years ago. In 2007, the year before the world financial crisis hit Russia, Russia's GDP growth topped 8 percent. It had been growing at that pace, buoyed by soaring commodity prices, for almost a decade, and it was not accidental that this was the decade in which Putin made his pact with the people: You get financial and consumer comforts, and we get political power. It's hard to maintain such a pact when the goodies stop flowing. Which brings us to the looming issue of the Russian budget deficit. To keep the people happy and out of politics, the Russian government has promised a lot of things to a lot of people. (Putin's campaign promises alone are estimated by the Russian Central Bank to cost at least $170 billion.) To balance its budget with such magnanimity, Russia needs high oil prices, to the point where last month, the Ministry of Economic Development announced that an $80 barrel of oil would be a "crisis." Keeping in mind that oil is now about $98 a barrel, and that Russia used to be able to balance its budgets just fine with oil at a fraction of the price, this doesn't look too good for Putin. Factor in the worsening European crisis -- Europe is still Russia's biggest energy customer -- and the fact that the state has put off unpopular but increasingly necessary reforms, like raising utility prices, and you find yourself looking at a powder keg. Russia’s economy and regime is oil dependent- reverse causal Schuman 12 [Michael, is an American author and journalist who specializes in Asian economics, politics and history. He is currently the Asia business correspondent for TIME Magazine. July 5 2012, “Why Vladimir Putin Needs Higher Oil Prices” http://business.time.com/2012/07/05/why-vladimir-putin-needshigher-oil-prices/ //jweideman] Falling oil prices make just about everyone happy. For strapped consumers in struggling developed nations, lower oil prices mean a smaller payout at the pump, freeing up room in strained wallets to spend on other things and boosting economic growth. In the developing world, lower oil prices mean reduced inflationary pressures, which will give central bankers more room to stimulate sagging growth. With the global economy still climbing out of the 2008 financial crisis, policymakers around the world can welcome lower oil prices as a rare piece of helpful news. But Vladimir Putin is not one of them. The economy that the Russian President has built not only runs on oil, but runs on oil priced extremely high. Falling oil prices means rising problems for Russia – both for the strength of its economic performance, and possibly, the strength of Putin himself. Despite the fact that Russia has been labeled one of the world’s most promising emerging markets, often mentioned in the same breath as China and India, the Russian economy is actually quite different from the others. While India gains growth benefits from an expanding population, Russia, like much of Europe, is aging; while economists fret over China’s excessive dependence on investment, Russia badly needs more of it. Most of all, Russia is little more than an oil state in disguise. The country is the largest producer of oil in the world (yes, bigger even than Saudi Arabia), and Russia’s dependence on crude has been increasing. About a decade ago, oil and gas accounted for less than half of Russia’s exports; in recent years, that share has risen to two-thirds. Most of all, oil provides more than half of the federal government’s revenues. What’s more, the economic model Putin has designed in Russia relies heavily not just on oil, but high oil prices. Oil lubricates the Russian economy by making possible the increases in government largesse that have fueled Russian consumption. Budget spending reached 23.6% of GDP in the first quarter of 2012, up from 15.2% four years earlier. What that means is Putin requires a higher oil price to meet his spending requirements today than he did just a few years ago. Research firm Capital Economics figures that the government budget balanced at an oil price of $55 a barrel in 2008, but that now it balances at close to $120. Oil prices today have fallen far below that, with Brent near $100 and U.S. crude less than $90. The farther oil prices fall, the more pressure is placed on Putin’s budget, and the harder it is for him to keep spreading oil wealth to the greater population through the government. With a large swath of the populace angered by his re-election to the nation’s presidency in March, and protests erupting on the streets of Moscow, Putin can ill-afford a significant blow to the economy, or his ability to use government resources to firm up his popularity. That’s why Putin hasn’t been scaling back even as oil prices fall. His government is earmarking $40 billion to support the economy, if necessary, over the next two years. He does have financial wiggle room, even with oil prices falling. Moscow has wisely stashed away petrodollars into a rainy day fund it can tap to fill its budget needs. But Putin doesn’t have the flexibility he used to have. The fund has shrunk, from almost 8% of GDP in 2008 to a touch more than 3% today. The package, says Capital Economics, simply highlights the weaknesses of Russia’s economy: This cuts to the heart of a problem we have highlighted before – namely that Russia is now much more dependent on high and rising oil prices than in the past… The fact that the share of ‘permanent’ spending (e.g. on salaries and pensions) has increased…creates additional problems should oil prices drop back (and is also a concern from the perspective of medium-term growth)…The present growth model looks unsustainable unless oil prices remain at or above $120pb. IL-Production lowers prices Production lowers oil prices Leonardo 12 [Maugeri, Leonardo. “Global Oil Production is Surging: Implications for Prices, Geopolitics, and the Environment.” Policy Brief, Belfer Center for Science and International Affairs, Harvard Kennedy School, June 2012.http://belfercenter.ksg.harvard.edu/files/maugeri_policybrief.pdf //jweideman] Oil Prices May Collapse. Contrary to prevailing wisdom that increasing global demand for oil will increase prices, the report finds oil production capacity is growing at such an unprecedented level that supply might outpace consumption. When the glut of oil hits the market, it could trigger a collapse in oil prices. While the age of "cheap oil" may be ending, it is still uncertain what the future level of oil prices might be. Technology may turn today's expensive oil into tomorrow's cheap oil. The oil market will remain highly volatile until 2015 and prone to extreme movements in opposite directions, representing a challenge for investors. After 2015, however, most of the oil exploration and development projects analyzed in the report will advance significantly and contribute to a shoring up of the world's production capacity. This could provoke overproduction and lead to a significant, steady dip of oil prices, unless oil demand were to grow at a sustained yearly rate of at least 1.6 percent trough 2020. Shifting Market Has Geopolitical Consequences. The United States could conceivably produce up to 65 percent of its oil consumption needs domestically, and import the remainder from North American sources and thus dramatically affect the debate around dependence on foreign oil. However the reality will not change much, since there is one global oil market in which all countries are interdependent. A global oil market tempers the meaningfulness of self- sufficiency, and Canada, Venezuela, and Brazil may decide to export their oil and gas production to non- U.S. markets purely for commercial reasons. However, considering the recent political focus on U.S. energy security, even the spirit of oil self-sufficiency could have profound implications for domestic energy policy and foreign policy. While the unique conditions for the shale boom in the United States cannot be easily replicated in other parts of the world in the short-term, there are unknown and untapped resources around the globe and the results of future exploration development could be surprising. This combined with China's increasing influence in the Middle East oil realm will continue to alter the geopolitics of energy landscape for many decades. New production decreases global prices Tallet et al 14 [Harry vidas, Martin Tallett Tom O’Connor David Freyman William Pepper Briana Adams Thu Nguyen Robert Hugman Alanna Bock ICF International EnSys Energy. March 31, 2014. “The Impacts of U.S. Crude Oil Exports on Domestic Crude Production, GDP, Employment, Trade, and Consumer Costs” http://www.api.org/~/media/Files/Policy/LNG-Exports/LNG-primer/API-Crude-Exports-Study-by-ICF-3-31-2014.pdf //jweideman] The increase in U.S. crude production accompanied by a relaxation of crude export constraints would tend to increase the overall global supply of crude oil, thus pulling downward pressure global oil prices. Although the U.S. is the second largest oil producer in the world and could soon be the largest by 2015, according to the International Energy Agency (IEA)48, the price impact of crude exports is determined by the incremental production, rather than total production. For this study, ICF used Brent crude as proxy for the global crude price as affected by forces of global crude supply and demand. The impact of lifting crude exports on Brent prices, as shown in Exhibit 4-19, is relatively small, about $0.05 to $0.60/bbl in the Low- Differential Scenario and about $0.25 to $1,05/bbl in the High-Differential Scenario. It should be noted that Brent prices are affected by various factors such as emerging supply sources, OPEC responses to increasing U.S. and Canadian production, and geopolitical events. Changes in any of these factors could mean actual Brent prices would deviate significantly from our forecasts. However, in general, higher global production leads to lower crude prices, all other factors being equal.Allowing crude exports results in a Brent average price of $103.85/bbl over the 2015-2035 period, down S0.35/bbl from the Low-Differential Scenario without exports and down $0.75/bbl from the High-Differential Scenario without exports.While global crude prices drop, domestic crude prices gain strength when exports are allowed because lifting the restrictions helps relieve the U.S. crude oversupply situation and allows U.S. crudes to fully compete and achieve pricing in international markets close to those of similar crude types. Exhibit 4-20 shows WTI prices increase to an average of $98.95/bbl over the 2015-2035 period in the Low-Differential and High-Differential With Exports Cases as opposed to $96.70/bbl in the Low-Differential No Exports Case and $94.95/bbl in the High-Differential No Exports Case. The range of increase related to allowing exports is $2.25 to $4.00/bbl averaged over 2015 to 2035. Production decreases prices Alquist and Guenette 13 [Ron Alquist and Justin-Damien Guénette work for the bank of Canada. 2013, “A Blessing in Disguise: The Implications of High Global Oil Prices for the North American Market,” http://www.bankofcanada.ca/wp-content/uploads/2013/07/wp2013-23.pdf //jweideman] The presence of these unconventional sources of oil throughout the world and the ability to recover them makes a large expansion in the physical production of oil a possibility. Recent estimates suggest that about 3.2 trillion barrels of unconventional crude oil, including up to 240 billion barrels of tight oil, are available worldwide (IEA 2012a). By 2035, about 14 per cent of oil production will consist of unconventional oil, an increase of 9 percentage points. The potential for unconventional oil extraction around the world has led some oil industry analysts to describe scenarios in which the world experiences an oil glut and a decline in oil prices over the medium term (Maugeri 2012). Impact-Proliferaiton Russia’s economy is key to stop proliferation Bukharin 3 [Oleg, August, he is affiliated with Princeton University and received his Ph.D. in physics from the Moscow Institute of Physics and Technology “The Future of Russia: The Nuclear Factor”, http://www.princeton.edu/~lisd/publications/wp_russiaseries_bukharin.pdf,//jweideman] There are presently no definite answers about the future of the nuclear security agenda in Russia. The Russian nuclear legacy – its nuclear forces, the nuclear-weapons production and power- What is clear is that nuclear security and proliferation risks will be high as long as there remain their underlying causes: the oversized and underfunded nuclear complex, the economic turmoil, and wide-spread crime and corruption. The magnitude of the problem, however, could vary significantly depending on Russia’s progress in downsizing of its nuclear weapons complex; its ability to maintain core competence in the nuclear field; availability of funding for the nuclear industry and safeguards and security programs; political commitment by the Russian government to improve nuclear security; and international cooperation. Economically-prosperous Russia, the rule of law, and a smaller, safer and more secure nuclear complex would make nuclear risks manageable. An integration of the Russian nuclear complex into the world’s nuclear industry, increased transparency of nuclear operations, and generation complex, huge stocks of nuclear-useable highly enriched uranium and plutonium, and environmental clean-up problems – is not going to go away anytime soon. cooperative nuclear security relations with the United States and other western countries are also essential to reducing nuclear dangers and preventing catastrophic terrorism. Proliferation is the worst – it destabilizes hegemony and all other nations and escalates to nuclear war – deterrence doesn’t account for irrational actors Maass 10 Richard Maass, working for his Ph. D. in political science at Notre dame University, and currently teaches classes there on International Relations. 2010 “Nuclear Proliferation and Declining U.S. Hegemony” http://www.hamilton.edu/documents//levitt-center/Maass_article.pdf On August 29, 1949, The Soviet Union successfully tested its first nuclear fission bomb, signaling the end of U.S. hegemony in the international arena. On September 11th, 2001, the world’s single most powerful nation watched in awe as the very symbols of its prosperity fell to rubble in the streets of New York City. The United States undisputedly “has a greater share of world power than any other country in history” (Brooks and Wolforth, 2008, pg. 2). Yet even a global hegemon is ultimately fallible and vulnerable to rash acts of violence as it conducts itself in a rational manner and assumes the same from other states. Conventional strategic thought and military action no longer prevail in an era of increased globalization. Developing states and irrational actors play increasingly influential roles in the international arena. Beginning with the U.S.S.R. in 1949, nuclear proliferation has exponentially increased states’ relative military capabilities as well as global levels of political instability. Through ideas such as nuclear peace theory, liberal political scholars developed several models under which nuclear weapons not only maintain but increase global tranquility. These philosophies assume rationality on the part of political actors in an increasingly irrational world plagued by terrorism, despotic totalitarianism, geo-political instability and failed international institutionalism. Realistically, “proliferation of nuclear [weapons]…constitutes a threat to international peace and security” (UN Security Council, 2006, pg. 1). Nuclear security threats arise in four forms: the threat of existing arsenals, the emergence of new nuclear states, the collapse of international non-proliferation regimes and the rise of nuclear terrorism. Due to their asymmetric destabilizing and equalizing effects, nuclear weapons erode the unipolarity of the international system by balancing political actors’ relative military power and security. In the face of this inevitable nuclear proliferation and its effects on relative power, the United States must accept a position of declining hegemony. Despite nuclear proliferation’s controversial nature, states continue to develop the technologies requisite for constructing nuclear weapons. What motivates men to create “the most terrifying weapons ever created by human kind…unique in their destructive power and in their lack of direct military utility”(Cirincione, 2007, pg. 47)? Why then do states pursue the controversial and costly path of proliferation? To states, nuclear weapons comprise a symbolic asset of strength and “as a prerequisite for great power status” (Cirincione, 2007, pg. 47). On a simplistic level, nuclear weapons make states feel more powerful, respected and influential in world politics. When it is in their best interest, states develop nuclear capabilities to ensure their own sovereignty and to potentially deter other states from attacking. According to realist thinkers, nuclear weapons provide the “ultimate security guarantor” in an anarchic international system (Cirincione, 2007, pg. 51). Proliferation optimists and rational deterrence theorists, such as Kenneth Waltz, argue proliferation stabilizes international security and promotes peace. Rational deterrence theory states that nations refrain from nuclear conflict because of the extraordinarily high cost. Arguably the most powerful military technology ever developed by man, nuclear weapons have only twice been deployed in actual conflict, due to the devastation they incur. Nuclear weapons increase the potential damage of any given military conflict due to their immense destructive capabilities. Summarizing rational deterrence framework, Waltz asserts “states are deterred by the prospect of suffering severe damage and by their inability to do much to limit it” (Sagan and Waltz, 2003, pg 32). According to the rational deterrence framework, political actors refrain from both conventional and nuclear conflict because of the unacceptably high costs. Ultimately an assumption, rational deterrence theory lacks any empirically tested evidence. Nuclear proliferation exponentially increases the possibility of non-proliferation regime collapse and nuclear conflict, reducing all states’ relative power. Nuclear peace theory seems plausible, but like any mathematical model it may only marginally apply to world politics and the dynamics of nuclear proliferation, due to the fact that “international security is not reducible to the theory of mathematical games” (Bracken, 2002, pg. 403). Rather, the spread of nuclear weapons exponentially decreases the stability of regional and global politics by intensifying regional rivalries and political tensions, both of which may potentially catalyze a nuclear catastrophe. Frustrated with a lack of results through conventional conflict, desperate states may look to nuclear arsenals as a source of absolute resolution for any given conflict. The use of nuclear weapons, even in a limited theater, could plausibly trigger chain reactions rippling across the globe. With their interests and sovereignty threatened, other nuclear states will eventually use their own weapons in an effort to ensure national security. President Kennedy warned of the danger of nuclear proliferation in 1963: I ask you to stop and think for a moment what it would mean to have nuclear weapons in so many hands, in the hands of countries…there would be no rest for anyone then, no stability, no real security…there would only be the increased chance of accidental war, and an increased necessity for the great powers to involve themselves in what otherwise would be local conflicts (Cirincione, 2007, pg. 103). Proliferation decreases the relative security of all states not only through the possibility of direct conflict, but also by threatening foreign and domestic interests. As the sole international hegemon, the U.S. seeks to use its power to insure its security and influence international politics in a way that reflects its own interests and values (Huntington, 1993, pg. 70). In addition to creating a direct security threat, further proliferation jeopardizes the United States’ ability to project its primacy and promote its interests internationally Impact-Relations US-Russia tensions high now Labott 14 [Elise Labott, CNN Foreign Affairs Reporter. 3/11/14, “Ukraine impasse stirs U.S.-Russia tensions” http://www.cnn.com/2014/03/10/world/europe/ukraine-us-russia-tensions/ //jweideman] Tensions between the United States and Russia over the crisis in Crimea have exploded into an open row as Russia rejects U.S. diplomatic efforts to solve the impasse. Russian Foreign Minister Sergey Lavrov said Monday that U.S. Secretary of State John Kerry postponed a face-to-face meeting with Russian President Vladimir Putin to discuss American proposals, which Moscow has effectively rejected, on solving the crisis. The meeting, which Russia said was supposed to happen Monday, would have marked the highest-level contact between the two countries since Russian troops took up positions in Crimea, and would have come ahead of Sunday's potentially explosive vote on whether Crimea should split from Ukraine and join Russia. But Kerry told Lavrov he needed to know Moscow would engage seriously on a diplomatic solution before meeting with the Russian leader. He also wanted to see and end to Russia's "provocative steps" before traveling to Russia. Expert: We need a 'Plan B' for Ukraine Pro-Russian forces muscle into base What Bush admin. got wrong on Russia Relations between Russia and the West have grown increasingly tense since Russian soldiers seized effective control of the pro-Russian region. The United States and other European powers have threatened possible sanctions in response to Russia's moves, but Moscow has shown little sign of backing down. At: dutch disease Russia doesn’t have dutch disease, but oil still matters Adomanis 12 [Mark, contributer to forbes, specializes in Russian economics. 6/22/12, “Is Russia Suffering From Dutch Disease?” http://www.forbes.com/sites/markadomanis/2012/06/22/is-russia-suffering-from-dutch-disease///jweideman] Part of the reason that Russia is not experiencing Dutch Disease (which is something you would normally expect in a country that that the world economy has been in turmoil for most of the past 4 years: there has been a “flight” to quality in “safe” assets and currencies which has surely worked to weaken the ruble and depress its value. The “new normal” is actually a pretty bizarre state of affairs, and is characterized by any number of things, such as negative real interest rates on German Bunds and US treasuries, that ten years ago would have seemed impossible. Russia’s economy faces an awful lot of risks, and its over-dependence on natural resources is extremely dangerous, particularly at a time that global growth is slamming to a halt. Buckley is right that Russia needs to diversify, and that its government will find this process to be an extremely difficult and complicated one. But, at the present time, one of the very few economic risks that Russia doesn’t face is Dutch Disease: its currency isn’t overvalued and, if anything, is actually trending lower against the main has earned such an enormous pile of money from selling oil and natural gas) is reserve currencies. No Russian dutch disease Dobrynskaya and Turkisch 9 [Victoria Dobrynskaya & Edouard Turkisch CEPII. September 2009, “Is Russia sick with the Dutch disease?” http://www.cepii.com/PDF_PUB/wp/2009/wp2009-20.pdf//jweideman] The fear that the Russian economy may become too dependent on the energy sector and not sufficiently diversified has influenced monetary policy over the last ten years. This policy was aimed at preventing the nominal appreciation of the rouble in order to maintain industrial competitiveness. In this paper, using Rosstat and CHELEM databases, we study whether Russia suffered the Dutch disease in 1999-2007. We do find some symptoms of it in Russia: there was a strong real appreciation of the rouble, real wages increased, employment decreased in manufacturing industries and rose in services sector. However, there was no sign of a deindustrialisation, what contradicts the theory of the Dutch disease. Indeed, industrial production increased significantly. Furthermore, the symptoms present in Russia can be the consequences of other factors than the existence of natural resources. The appreciation of the rouble in real terms came partly from the Balassa-Samuelson effect. The quick development of the services was partly due to the fact that services were not put forward during the Soviet Union limes. The outflow of labour from the manufacturing industries resulted in inflow of labour in services sector rather than in the energy sector. AT: China turn The US and china cooperate over energy ECP 13 [US China Energy Cooperation Project. 2013-11-20, “U.S.-China Energy Cooperation Program (ECP) and Chinese Institute of Electronics (CIE) Launch Joint U.S.-China Green Data Center Industrial Initiative,” http://www.uschinaecp.org/en/News/NewsDetail.aspx?nid=f74ebbdf-f9be-418a-8147a12a7e31088b&cid=d952ba0f-3ba2-4372-8b26-ef635b67d638 //JWEIDEMAN] On November 20, 2013, with support of the Chinese Ministry of Industry and Information Technology (MIIT) and the U.S. Trade and Development Agency (USTDA), the U.S.-China Energy Cooperation Program (ECP) and Chinese Institute of Electronics (CIE) jointly launched the U.S.-China Green Data Center Industry Partnership at the US China Green Data Center Workshop at the Xiyuan Hotel in Beijing. The key related issues of the development of the energy efficient/green data center sector, which includes: market, needs, opportunities, challenges, technology solutions and best practices, evaluation methods and etc, have been addressed and discussed at the workshop. At the workshop, the two sides on behalf of the participating U.S. and Chinese industries, signed a MEMORANDUM OF UNDERSTANDING For Cooperation on Green Data Center to promote US-China Industry cooperation in China’s green data center sector. Senior officials from the Department of Energy Efficiency and Resources Utilization of MIIT, USTDA, Foreign Commercial Service and the U.S. Department of Energy China Office of the U.S. Embassy in Beijing witnessed the signing ceremony. Industry experts from Intel, UTC, Caterpillar, Cisco, Emerson, Fuxing Xiaocheng, NewCloud, Neusoft, Huawei, Inspur, ZTE attended the workshop. The three-year joint program between ECP and CIE aims to provide valuable reference and living best practices for green data center development in China through deeply cooperation between both US and China industries. Specifically, these include: 1. Green data center technical guideline, technology catalogue, and green data center best practice portfolio. 2. Green data center related incentive plan, business model, monitoring and evaluation method and system and etc. 3. Green data center energy saving demonstration projects, needs assessment, and industry entry study etc. 4. Capacity building: green data center expert committee establishment, practical training, and certificate training and study tour. The US and China are committed to cooperation over energy Department of energy 11 [United states department of energy. January 2011, “U.S.-China Clean Energy Cooperation A Progress rePort by the U.s. DePArtment of energy” http://www.us-china-cerc.org/pdfs/US_China_Clean_Energy_Progress_Report.pdf//jweideman] The United States and the People’s Republic of China have worked together on science and technology for more than 30 years. Under the Science and Technology Cooperation Agreement of 1979, signed soon after normalization of diplomatic relations, our two countries have cooperated in a diverse range of fields, including basic research in physics and chemistry, earth and atmospheric sciences, a variety of energy-related areas, environmental management, agriculture, fisheries, civil industrial technology, geology, health, and natural disaster planning. More recently, in the face of emerging global challenges such as energy security and climate change, the United States and China entered into a new phase of mutually beneficial cooperation. In June 2008, the U.S.-China Ten Year Framework for Cooperation on Energy and the Environment was created and today it includes action plans for cooperation on energy efficiency, electricity, transportation, air, water, wetlands, nature reserves and protected areas. In November 2009, President Barack Obama and President Hu Jintao announced seven new U.S.- China clean energy initiatives during their Beijing summit. In doing so, the leaders of the world’s two largest energy producers and consumers affirmed the importance of the transition to a clean and low-carbon economy—and the vast opportunities for citizens of both countries in that transition.the following joint initiatives were announced in november 2009: • U.S.-China Clean Energy Research Center. Scientists and engineers from both countries are working together to develop clean energy technologies, initially focusing on building energy efficiency, clean coal and clean vehicles. Both countries are contributing equally to $150 million in financial support from public and private sources over five years. • Electric Vehicles Initiative. This initiative includes the joint development of standards for charging plugs and testing protocols of batteries and other devices, demonstration projects in paired cities to collect and share data on charging patterns and consumer preferences, joint development of technical roadmaps, and public education projects. • Energy Efficiency Action Plan. Both governments are working together with the private sector to develop energy efficient building codes and rating systems, benchmark industrial energy efficiency, train building inspectors and energy efficiency auditors for industrial facilities, harmonize test procedures and performance metrics for energy-efficient consumer products, and exchange best practices in energy efficiency labeling systems. Energy innovation in one country accelerates clean energy deployment in all countries. And the combined research expertise and market size of the U.S. and China provide an unprecedented opportunity to develop clean energy solutions that will reduce pollution and improve energy security while enhancing economic growth globally. AT: Oil shocks Oil shocks don’t collapse the economy Fels, Pradhan, and Andreopoulos 11 [By Joachim Fels, Manoj Pradhan, Spyros Andreopoulos: Write for Morgan Stanley- economic institution. March 4, 2011. “This Time Is Different” http://www.morganstanley.com/views/gef/archive/2011/20110304-Fri.html //jweideman] Will elevated oil prices generate a sharp slowdown? The recent run-up in oil prices has investors worried that a sharp slowdown in global growth lies ahead. Worse, with the oil market increasingly pricing the risk that the shock may be permanent - the Dec 2012 futures contract trades at around US$109 per barrel - there is even talk of the possibility of a global recession. We think not, for two broad reasons. First, oil prices constitute wealth redistribution, not wealth destruction. And second, this time is different: as things stand, we think that the current oil shock is unlikely to harm growth much. Higher oil prices redistribute wealth rather than destroying it. From a global perspective, higher oil prices do not mean that wealth is destroyed. Rather, it is redistributed - from net oil importers to net exporters (see below on The Effects of Oil Price Shocks). While this wealth transfer can be substantial, much of it is, over time, recycled back into net oil-importer economies: •· Oil exporters will spend some of their newly gained oil wealth on imports from net importers; hence, increased import demand replaces some of the domestic demand lost due to the wealth transfer; •· Oil exporters will also purchase assets in net importer economies, providing support for asset markets there. Even so, oil shocks have always been redistributive - yet arguably they have done considerable damage to the global economy in the past. The 1970s, when oil price spikes preceded two successive recessions, are held up as the prime examples of the harm that oil can do to the economy. So, it is tempting to look at the effects of previous oil shocks in order to infer the consequences of the current one. We would urge caution, however, as all else is not equal. Many things are different this time, so a given increase in the price of oil should be less harmful now than it would have been in the 1970s: 1. Nature of the Shock It is worth bearing in mind that the increase in the oil price since September has been due to a mixture of demand and supply influences. Much of the rise in the price of Brent since September from a level of about US$80 per barrel to about US$100 was due to demand, in our view, with the oil market repricing the global growth and inflation trajectory. It is only the recent rise, due to events in the Middle East, that constitutes a supply shock. That is, supply has accounted for less than half of the oil price increase since September. Why does the distinction matter? Oil demand shocks due to strength in the real economy are endogenous; they are unlikely to derail growth. The price of oil merely acts like an elastic leash on the dog that is the global economy. If the dog surges ahead too quickly, the constricting effect of the leash will make sure it slows down - but it will not stop. Between 2003 and 2007, the real price of oil roughly doubled, with little evident harm to the global economy. But exogenous, supply-induced shocks tend to be stagflationary in nature. In the dog metaphor, an oil supply shock could act as a jerk on the leash, which could bring the dog to a halt. 2. Initial Conditions This recovery is still in relatively early stages and thus more robust than a late-cycle recovery. This is partly because personal savings rates are higher early in the cycle than late in the cycle, providing more of a cushion for consumer spending. Also, corporate profit margins are usually higher early on in the cycle than late in the cycle, and companies can thus more easily absorb the cost push from higher oil prices. Most importantly, global monetary conditions are very loose, providing a cushion to the real economy. 3. Structure of the Economy •· The oil intensity of global GDP is lower. That is, less oil is used today to produce a dollar of GDP. Hence, a given percentage increase in the inflation-adjusted price of oil will do less damage to the economy now. •· Labour markets are more flexible: The effect of an oil supply shock on employment and output is larger, the less real wages adjust in response. More flexible labour markets make real wages more responsive to supply shocks, thus dampening the effects on production and employment. •· Monetary policy is more credible: A more credible central bank needs to act less, all else equal, to achieve a given objective. In the case of an oil shock, the implication is that monetary policy needs to do less harm to the real economy to achieve a given dampening effect on inflation. 4. Policy Response in Net Importing Economies: ‘Rational Inaction' Means Central Banks Will Keep Rates on Hold The policy response to an oil shock is an important determinant of the overall economic effect. If central banks were to tighten in response to inflation, the dampening effect of the initial oil shock on the economy would be amplified. We have just argued, in very general terms, that a more credible monetary policy requires less of a response to achieve a desired effect. In this particular case, however, we do not expect much of a response in the first place: in our view, most central banks will not tighten policy aggressively in response to the recent surge in oil prices. This is likely to be true for both DM and EM monetary authorities. Again, this reaction - or rather, the lack of one - is rational: it is virtually impossible to forecast whether the current supply worries will abate or not, so a ‘wait and see' stance makes sense. 5. Oil Exporters' Behaviour May Be Materially Different •· Oil exporters will likely spend more of the wealth transfer than usual: With the risk of social unrest increasing, governments will be inclined to increase spending and transfers in order to maintain stability. This suggests that a larger part of the initial wealth transfer from net exporters to net importers will be reversed, over time, through goods imports of the former from the latter. •· At current prices, oil producers will probably want to make up a meaningful part of the Libyan production shortfall over the medium term - not least because doing so will generate additional revenue with which social stability can be bought. Our commodity strategy team thinks that there is enough spare capacity to do so (see Crude Oil: MENA Turm-OIL, February 25, 2011). Farms DA 1NC Water pollution from runoff on US farms has avoided regulation so far but continued exemption is not guaranteed David C. Roberts 9, assistant professor in the Department of Agribusiness and Applied Economics at North Dakota State University; Christopher D. Clark, associate professor in the Department of Agricultural Economics at the University of Tennessee; and Burton C. English, William M. Park and Roland K. Roberts, professors in the Department of Agricultural Economics at the University of Tennessee, “Estimating Annualized Riparian Buffer Costs for the Harpeth River Watershed,” Appl. Econ. Perspect. Pol. (Winter 2009) 31 (4): 894-913, oxford journals Since the passage of the Clean Water Act (CWA), largely unregulated nonpoint sources (NPS) of pollution have contributed to an increasing share of the nation’s water quality problems (Copeland). While the amount of agricultural land in the United States has remained relatively constant over the past thirty years, agricultural production has increased because of technological advances and increasingly fertilizer-intensive agricultural practices. These fertilizerintensive practices have been a major contributor to a threefold increase in the nitrate load in the Mississippi River entering the Gulf of Mexico (Greenhalgh and Sauer). When the states and territories surveyed the nation’s rivers and streams, they found 39% of the segments surveyed impaired for one or more uses. Agriculture was listed as a contributor to 48% of these impairments—more than twice as much as any other source (USEPA 2002).¶ Reducing environmental externalities associated with agriculture poses a number of difficult policy challenges. In the context of water quality, NPS runoff from agricultural lands has proven to be relatively resistant to regulation . the CWA imposes effluent limitations on point sources but not on NPS and explicitly exempts “agricultural stormwater discharges” from the definition of point sources (Houck).1 The void created by the absence of federal regulation has been filled with a variety of voluntary programs (Houck). More recently, it has been filled by the promotion of policies that would For example, allow point sources subject to the CWA’s effluent limitations to satisfy some of these restrictions by leasing effluent reductions from NPS. This has come to be known as “water quality trading” (USEPA 2003b, 1996). States, attracted by the possibility of reducing pollution abatement costs, have implemented a number of water quality trading (WQT) programs (Breetz et al.).¶ States have also stepped into the void by enacting statutes that address NPS pollution by requiring landowners to observe mandatory “setback” or “buffer” requirements along waterways. These statutes tend to be limited in their application to specific activities or waterways, or both. For example, while the Georgia Erosion and Sedimentation Control Act prohibits land-disturbing activities within 7.6 meters (25 feet) of the banks of any state water, it exempts a fairly extensive list of activities from the prohibition, including “agricultural practices, forestry land management practices, dairy operations, livestock and poultry management practices, construction of farm buildings…” (Georgia Department of Natural Resources). North Carolina has adopted regulations designed to protect a 15.2-meter (50-feet) riparian buffer along waterways in the Neuse and Tar-Pamlico River Basins (North Carolina Department of Environment and Natural Resources). Virginia’s Chesapeake Bay Preservation Act provides for 30.4-meter (100-feet) buffer areas along specifically designated “Resource Protection Areas,” but allows encroachments into buffer areas for agricultural and silvicultural activities under certain conditions (Chesapeake Bay Local Assistance Board).¶ Riparian buffer strips—areas of trees, shrubs, or other vegetation along surface water bodies—have also played a prominent role in federal and state voluntary programs to reduce NPS pollution. For example, the Conservation Reserve Program has supported the use of buffer strips since 1988 (Prato and Shi) and funds made available to states through the CWA’s Section 319 program are often used to subsidize buffer strip installation (Nakao and Sohngen). Buffer strips are proving an attractive policy option because of their effectiveness in intercepting and removing nutrients, sediment, organic matter, and other pollutants before these pollutants enter surface water and because of the other environmental benefits they provide, including improved terrestrial and aquatic habitat, flood control, stream bank stabilization, and esthetics (Qiu, Prato, and Boehm).¶ NPS pollution has typically avoided the type of regulation imposed on point sources. However, its prominent role in the impairment of surface and ground water quality means that water quality cannot be meaningfully improved without addressing agricultural NPS pollutant loadings. The possibility that the costs of abating pollution from NPS are lower than costs of abating pollution from point sources (Faeth) may also make NPS an attractive target . In any event, buffer strips are proving to be a relatively popular means of controlling NPS discharges. Although agriculture has often been excluded from regulation, there is no guarantee that the agricultural sector will continue to enjoy this treatment. For example, North Carolina’s Department of Water Quality drafted, but did not adopt , regulations that would have required all crop farms adjacent to streams to install buffer strips (Schwabe). In the absence of regulation of agricultural NPS, it seems likely that alternatives to regulation will continue, if not increase. Many of these will promote the installation of buffer strips through subsidies of one form or another. For example, the Conservation Reserve Enhancement Program, which subsidizes conservation-oriented practices by landowners, has established a goal of installing 157,000 miles of riparian buffers, filter strips, and wetlands in states in the Chesapeake Bay watershed (Bonham, Bosch, and Pease). IOOS leads to new policies to prevent harmful algal blooms---results in stricter regulation on non-point sources of nutrient pollution like agriculture Dr. Donald Anderson 8, Senior Scientist in the Biology Department of the Woods Hole Oceanographic Institution and Director of the U.S. National Office for Marine Biotoxins and Harmful Algal Blooms, Phd in Aquatic Sciences from MIT, Written testimony presented to the House Committee on Science and Technology, Subcommittee on Energy and Environment, July 10 2008, http://www.whoi.edu/page.do?pid=8915&tid=282&cid=46007 These are but a few of the advances in understanding that have accrued from ECOHAB regional funding. Equally important are the discoveries that provide management tools to reduce the Management options for dealing with the impacts of 8 HABs include reducing their incidence and extent (prevention), stopping or containing blooms (control), and minimizing impacts (mitigation). Where possible, it is preferable to prevent HABs rather than to treat their symptoms. Since increased pollution and nutrient loading may enhance the growth of some HAB species, these events may be prevented by reducing pollution inputs to coastal waters, particularly industrial, agricultural, and domestic effluents high in plant nutrients. This is especially important in shallow, poorly flushed coastal waters that are most susceptible to nutrient-related algal problems. As mentioned above, research on the links between certain HABs and nutrients has highlighted the importance of non-point sources of nutrients (e.g., from agricultural activities , fossil-fuel combustion, and animal feeding operations). ¶ The most effective HAB management tools are monitoring programs that involve sampling and testing of wild or cultured seafood products directly from the natural environment, as this allows unequivocal tracking of toxins to their site of origin and impacts of HABs on coastal resources. targeted regulatory action . Numerous monitoring programs of this type have been established in U.S. coastal waters, typically by state agencies. This monitoring has become quite expensive, however, due to the proliferation of toxins and potentially affected resources. States are faced with flat or declining budgets and yet need to monitor for a growing list of HAB toxins and potentially affected fisheries resources. Technologies are thus urgently needed to facilitate the detection and characterization of HAB cells and blooms.¶ One very useful technology that has been developed through recent HAB research relies on species- or strain-specific “probes” that can be used to label only the HAB cells of interest so they can then be detected visually, electronically, or chemically. Progress has been rapid and probes of several different types are now available for many of the harmful algae, along with techniques for their application in the rapid and accurate identification, enumeration, and isolation of individual species. One example of the direct application of this technology in operational HAB monitoring is for the New York and New Jersey brown tide organism, Aureococcus anophagefferens. The causative organism is so small and non-descript that it is virtually impossible to identify and count cells using traditional microscopic techniques. Antibody probes were developed that bind only to A. anophagefferens cells, and these are now used routinely in monitoring programs run by state and local authorities, greatly improving counting time and accuracy.¶ These probes are being incorporated into a variety of different assay systems, including some that can be mounted on buoys and left unattended while they robotically sample the water and test for HAB cells. Clustered with other instruments that measure the physical, chemical, and optical characteristics of the water column, information can be collected and used to make “algal forecasts” of impending toxicity. These instruments are taking advantage of advances in ocean optics, as well as the new A clear need has been identified for improved instrumentation for HAB cell and toxin detection, and additional resources are needed in this regard. This can be accomplished during development of the Integrated Ocean Observing System ( IOOS ) for U.S. coastal waters, and through a targeted research program on HAB prevention, control, and mitigation (see below). These are needed if we are to achieve our vision of future HAB monitoring and management programs – an integrated system that includes arrays of moored instruments as sentinels along the U.S. coastline, detecting HABs as they molecular and analytical methodologies that allow the toxic cells or chemicals (such as HAB toxins) to be detected with great sensitivity and specificity. develop and radioing the information to resource managers. Just as in weather forecasting, this information can be assimilated into numerical models to improve forecast accuracy. That wrecks farm productivity without reducing nutrient pollution Carl Shaffer 14, President of the Pennsylvania Farm Bureau, TESTIMONY TO THE HOUSE COMMITTEE ON TRANSPORTATION AND INFRASTRUCTURE SUBCOMMITTEE ON WATER RESOURCES AND ENVIRONMENT REGARDING NUTRIENT TRADING AND WATER QUALITY, March 25 2014, http://transportation.house.gov/uploadedfiles/2014-03-25-shaffer.pdf Fifth, the underlying assumption that it is easy and inexpensive for farmers and nonpoint sources to reduce nutrient loading is a myth. ¶ Concept of Water Quality Trading ¶ Farm Bureau policy supports the concept of water quality trading; implicit in that is the notion that participation by farmers is voluntary and that the system reflects the realities of agriculture. Farm Bureau has a long history of supporting market-based approaches to improving the environment. We have also encouraged states to include trading in their toolbox to help implement state water quality programs because trading and offsets can reduce costs associated with achieving environmental improvements. ¶ Even with that history of support, however, farmers and ranchers remain skeptical of trading programs in general and those associated with water quality specifically, and for good reason. Farmers grow things and understand that nutrient enrichment is a predictable outcome of all human activities – not just farming and ranching activities. Farmers understand that agricultural activities – like those in virtually every other part of life, such as shopping malls, golf courses, residential areas to name just a few – can affect the amount of nutrients that reach our waters. The fact is, each and every one of us plays a role in water quality; we all contribute to nutrient loading, either directly or indirectly, through the EPA’s environmental strategies too often focus more on affixing blame for problems or regulating some activity or person, rather than finding solutions that recognize and seek balance. food we consume, the products we purchase, and the way we live our lives. ¶ Unfortunately, EPA’s toolbox is both limited and dominated by an approach that focuses heavily on pollution prevention and reduction based on the concept of polluter pays. For conventional pollutants, this there is an ongoing misperception that agriculture chronically over-applies nutrients. Nutrients, however, are not conventional pollutants – they are a combination of pollutants from point sources and pollution from nonpoint sources . The fact is, nutrients are critical for optimal productivity approach has resulted in costly controls and restrictive permits on point sources. At the same time, in agriculture even though farmers and ranchers are striving for the best possible ecosystem function. Managing nutrients is extremely complicated because there is not one practice, technology or approach that can optimize nutrient utilization throughout the environment. Therefore, we need policy options that are balanced. We must develop solutions that optimize outcomes. We all want: 1) safe, affordable and abundant food, fiber and fuel; 2) vibrant and growing communities with jobs and expanding economic activity; and 3) fishable and swimmable waters. ¶ The challenges presented by trading and offset programs are the complex interplay of economic scenarios that could play out over time when such regulatory offsets programs are taken to their logical conclusions. For example, if are required for any new development or for expanding economic activity, one would expect a regulatory offsets process to trade low-value economic activity for high-value activity. In real life, however, such a program would not be likely to require an old home to be torn down before a new home could be constructed. Likewise, the construction and operation of a new manufacturing facility and the jobs inherent to that economic activity would not likely come at the expense result will undoubtedly be a shift in development activities out of lower value areas, likely rural areas and farmland , into high value urban areas. The downside of such an offsets of other high-value economic activity. But trading programs will allow “tradeoffs” and the program can be represented by simple math. For example, within an urban area, land suitable for building a factory could be valued at $100,000 or more per acre, while land in the same geographic area suitable to produce corn or soybeans could be valued at $10,000 per acre. In a market-based system, it would appear to be only rational to extinguish the environmental externalities generated by the farmland to offset the externalities associated with the higher value economic activity of manufacturing. While this may be an extreme example, the reality is The long-run reality for farmers and ranchers would be that, over time, rural areas will have fewer and fewer means to sustain themselves. ¶ Trading and Offsets are Creatures of State Statutes ¶ The Clean Water Act leaves the task of controlling water pollution largely to the states; it expressly recognizes, that the nation has never used water quality as a mechanism to cap or, in some cases like the Chesapeake Bay, reduce economic activity. preserves and protects “the primary responsibilities and rights of States to prevent, reduce, and eliminate pollution [and] to plan the development and use … of land and water resources.” Authorized federal involvement in state actions is carefully limited. Under no circumstances does the act authorize EPA to assume state responsibility to develop a planning process or a Total Maximum Daily Load (TMDL) implementation plan. It is within these contexts that trading programs are often contemplated. As such, states may implement trading and offsets programs established under state laws. In addition, states retain the flexibility to choose both if and how to use trading in the implementation of state water quality programs.¶ Nutrient Standards May Not be Affordable or Attainable Without Trading ¶ Optimizing nitrogen and phosphorus utilizations through trading may hold potential, but there are significant scientific, market and regulatory challenges. First, from a scientific standpoint, there is no direct relationship between agricultural nutrient management practices and nutrient loss. If the relationship were direct, trading would be straightforward, transparent and enable orderly operations of markets. ¶ Second, under the Clean Water Act, states establish and EPA approves water quality standards and criteria. States are currently feeling pressure from EPA to adopt default numeric nutrient standards and criteria based on the level of nutrients found in pristine waters. Such an approach holds If EPA is successful, cities, agriculture and other sources of nutrients will incur significant regulatory costs without any guarantee that water quality improvements will match the required level of investment. Restrictive state standards that are not based on reference waters can be unachievable and the prospect of establishing standards that force states to adopt costly control measures that, in the end, are not realistically attainable. require costly control and management measures. ¶ EPA and States Are Imposing Barriers for Markets and Trading ¶ Achieving the environmental and economic goals of point source nonpoint source (PS-NPS) water quality trading depends on having understandable rules that clearly define what is being traded and the parameters of the exchange. Trading rules and procedures establish who can trade, what is traded (credit definition), duration of a credit, baseline requirements (for calculating credits), accepted procedures for calculating credits, how the trade occurs, trading ratios, verification, liability rules, and enforcement procedures. ¶ In theory, trading assumes market participants have full information about the cost and effectiveness of their nutrient reduction options and can instantly and, at little-to-no-cost, obtain information on credit market prices and quantities. However, in the real world people are faced with limited time, resources, skills and acquaintance with markets. Complex rules and inadequate institutional design can result in poor buyer or seller participation, coordination failures and lack of desired outcomes. (Shortle, 2013). ¶ In fact, ex-post assessments of PS-NPS water quality trading programs already in existence have generally been negative about their performance. Most have seen little or no trading activity, with the expected role for nonpoint sources unrealized. A number of reasons have been presented including a lack of trading partners (due to limited regional scale or underlying economics), inadequate regulatory incentives, uncertainty about trading rules and practice performance, excessively high PS-NPS trading ratios (increasing the cost of nonpoint credits), legal and regulatory obstacles (including liability concerns), high transaction costs, and participant unfamiliarity and inexperience. ¶ Pennsylvania’s experience with water quality trading illustrates a number of the challenges I have mentioned. For example, the rules underlying Pennsylvania’s nutrient credit trading program, created in large part in response to an EPA mandate to reduce pollution in the Chesapeake Bay watershed, are the product of a multi-year stakeholder negotiation process that was codified in regulation in 2010. However, shortly thereafter, EPA announced that it would undertake a review of the offset and trading program in each Chesapeake Bay jurisdiction. EPA’s assessment included questions about whether or not Pennsylvania’s agricultural trading baseline met the requirements of TMDL—in spite of the fact that trades had already taken place under the program rules in place at the time. Further, the assessment included EPA’s expectations that Pennsylvania would demonstrate that the existing baseline was sufficient to meet the TMDL, or otherwise make “necessary adjustments” to the baseline acceptable to EPA. ¶ In response to EPA’s review, Pennsylvania has since proposed a number of possible changes to its trading program that have raised serious questions among existing and potential credit generators and users about what the rules governing the market for credits will look like going forward. Specifically, many are concerned about what happens to non-point source credit generators, primarily farmers, who have generated and sold credits under Pennsylvania's existing program, and who may have long-term commitments to provide credits for years into the future. The uncertainty is not conducive to sustaining a successful, transparent, long-term water quality ¶ trading program. ¶ The Myths – Easy Nor Inexpensive It’s Neither ¶ It is often assumed that agriculture can supply credits less expensively than other nonpoint and point sources. Whether or not this is true depends heavily on the trading rules and procedures described previously. Baseline requirements represent one trading rule that has an important impact on agriculture’s ability to be the low-price supplier of credits. ¶ Baseline requirements establish the level of stewardship farmers and ranchers perform on a parcel of land before they are eligible to participate in the trading program and actually produce credits for sale. Any abatement necessary to meet the baseline cannot be sold as credits, but is instead credited to meeting the load allocation for agriculture. When baselines are more stringent than current practices, a farmer would only be willing to create and sell credits if the expected credit price were high enough to cover the cost of meeting the baseline plus the cost of any measures taken to produce additional abatement. This increases the cost of supplying credits, and ¶ reduces the amount of credits purchased by point sources.¶ Current research suggests that concerns about baseline requirements are well founded. Stephenson et al. (2010) found that when the baseline is more stringent than current practices, agricultural credit costs for nitrogen can surpass costs per pound (for marginal abatement) for point sources because the baseline has claimed the lowest-cost pollutant reductions. Ghosh et al. (2011) found that Pennsylvania’s baseline requirements significantly increased the cost of entering a trading program, making it unlikely that nonpoint sources that could reduce nutrient losses for the lowest unit costs would enter the market. Wisconsin has expressed concern that EPA’s approach to defining baselines could obstruct agricultural sources’ participation in trading programs and possibly impede water quality improvements (Kramer, 2003). The impact of baseline requirements is a crucial matter and fundamental to the successful operation of any trading program, though its impact is not unique. Any trading rule or requirement that is incorrectly developed can have similar effects: fewer nonpoint source credits purchased by point sources, and total abatement costs for regulated sources higher than they could have been. ¶ As a regulatory agency, EPA appears to have difficulty appreciating the realities of how markets function. The agency is not necessarily tasked with creating private markets and most people would probably agree that the agency has difficulty appreciating the realities of how real markets function. As a result, environmental markets are suffering from a significant creditability crisis. This ultimately results in skeptical farmers and ranchers who then take a cautious approach to nutrient trading. if it were easy and inexpensive for farmers and ranchers to reduce nutrient loadings, they would have already figured out a way to capture the benefit associated with incremental nutrients lost to the environment. Farmers today use some of the most advanced technology in the world to optimize their productivity. From precision application using 4R nutrient stewardship to GPS technology, farmers and ranchers are committed to improving their production efficiencies, a fact that allows them in turn to reduce their environmental footprint. 4R nutrient stewardship is an effective concept that allows a farmer to use the right fertilizer source, at the right rate, at ¶ Regarding the cost of reducing nutrient loads, the right time and with the right placement to optimize nutrient utilization, while precision agriculture is a farming system that uses technology to allow closer, more site-specific management of the factors affecting crop production. ¶ For example, in precision agriculture, utilizing GPS and yield monitors, farmers can measure their output more precisely by matching yield data with the location in the field. Special computer-driven equipment can change the rate at which fertilizers, seeds, plant health products and other inputs are used, based on the needs of the soil and crop in a particular portion of a field. ¶ Farmers have embraced precision agriculture and the 4R philosophy because it is an innovative and science-based approach that enhances environmental protection, expands production, increases farmer profitability and improves sustainability. ¶ Conclusion ¶ Your constituents want affordable and abundant food , fiber and fuel, and the members of Farm Bureau want the chance to provide them. Farmers are concerned about the environment. As technology evolves so do farmers. We take advantage of technology, new practices and programs in order to not only provide safe, affordable and abundant food, fiber and fuel, but also to protect our land, water and air resources. Extinction Richard Lugar 2k, Chairman of the Senator Foreign Relations Committee and Member/Former Chair of the Senate Agriculture Committee “calls for a new green revolution to combat global warming and reduce world instability,” http://www.unep.org/OurPlanet/imgversn/143/lugar.html In a world confronted by global terrorism, turmoil in the Middle East, burgeoning nuclear threats and other crises, it is easy to lose sight of the long-range challenges. But we do so at our peril. One of the most daunting of them is meeting the world’s need for food and energy in this century. At stake is not only preventing starvation and saving the environment, but also world peace and security. History tells us that states may go to war over access to resources, and that poverty and famine have often bred fanaticism and terrorism. Working to feed the world will minimize factors that contribute to global instability and the proliferation of weapons of mass destruction . With the world population expected to grow from 6 billion people today to 9 billion by mid-century, the demand for affordable food will increase well beyond current international production levels. People in rapidly developing nations will have the means greatly to improve their standard of living and caloric intake. Inevitably, that means eating more meat. This will raise demand for feed grain at the same time that the growing world population will need vastly more basic food to eat. Complicating a solution to this problem is a dynamic that must be better understood in the West: developing countries often use limited arable land to expand cities to good land disappears, people destroy timber resources and even rainforests as they try to create more arable land to feed themselves. The long-term environmental consequences could be disastrous for the entire globe. ¶ Productivity revolution ¶ To meet the expected demand for food over the next 50 years, we in the United States will have to grow roughly three times more food on the land we have. That’s a tall order. My farm in Marion County, Indiana, for house their growing populations. As example, yields on average 8.3 to 8.6 tonnes of corn per hectare – typical for a farm in central Indiana. To triple our production by 2050, we will have to produce an annual average of 25 tonnes per hectare. Can we possibly boost output that much? Well, it’s been done before. Advances in the use of fertilizer and water, improved machinery and better tilling techniques combined to generate a threefold increase in yields since 1935 – on our farm back then, my dad produced 2.8 to 3 tonnes per hectare. Much US agriculture has seen similar increases. But of course there is no guarantee that we can achieve those results again. Given the urgency of expanding food production to meet world demand, we must invest much more in scientific research and target that money toward projects that promise to have significant national and global impact. For the United States, that will mean a major shift in the way we conduct and fund agricultural science. Fundamental research will generate the innovations that will be necessary to feed the world. The success at increasing United States can take a leading position in a productivity revolution. And our food production may play a decisive humanitarian role in the survival of billions of people and the health of our planet. AT Political Opposition Their evidence overestimates the farm lobby’s clout---the farm bill proves it’s too weak to influence policies Ron Nixon 13, NY Times, July 3 2013, “Farm Bill Defeat Shows Agriculture’s Waning Power,” http://www.nytimes.com/2013/07/03/us/politics/farm-bill-defeat-shows-agricultures-waningpower.html WASHINGTON — The startling failure of the farm bill last month reflects the declining clout of the farm lobby and the oncepowerful committees that have jurisdiction over agriculture policy, economists and political scientists said this week.¶ Although a number of factors contributed to the defeat of the bill — including Speaker John A. Boehner’s failure to rally enough Republican support and Democratic opposition to $20 billion in cuts to the food stamps program — analysts said the 234 to 195 vote also illustrated the shift in the American population and political power to more urban areas.¶ “There are a small number of Congressional districts where farming continues to carry much sway,” said Vincent H. Smith, a professor of agricultural economics at Montana State University. “ Especially in the House, the farm lobby has been substantially weakened .”¶ For much of American history, the agriculture sectors wielded tremendous political power. Farm groups were able to get key farm legislation passed by rallying millions of farmers in nearly every Congressional district. Influential farm state legislators like Representative Jamie L. Whitten of Mississippi, a Democrat who was chairman of the Appropriations Committee and its subcommittee on agriculture, brought billions in agriculture financing to their states and fought off attempts to cut subsidy programs despite pressure from both liberals and conservatives.¶ Mr. Whitten died in 1995 after 53 years in Congress. ¶ But as Americans have moved to the cities and suburbs, farmers and lawmakers representing districts largely dependent on agriculture have seen their political muscle steadily decline . Just 2.2 million people now work in farming in the United States, or about 2.5 percent of the total work force.¶ Farming now accounts for about 1 percent of gross national product, down from a high of about 9 percent in 1950. Only 40 lawmakers represent largely farming districts, according to research by Mr. Smith in 2006. He said that number was probably smaller today. Political opposition is insufficient to stop environmental regulation---greenhouse gas regulations prove Jonathan H. Adler 13, Prof of Law and Director of the Center for Business Law & Regulation at Case Western Reserve University School of Law, Oct 3 2013, “Conservatives and Environmental Regulation,” http://www.volokh.com/2013/10/03/conservatives-environmental-regulation/ Anti-regulatory rhetoric may be pervasive, but federal environmental regulation has continued to expand , under Democratic and Republican presidents alike. Anti-regulatory conservatives have been able to stem the tide of regulatory initiatives, but only for a time . The failure to develop and advance non-regulatory alternatives to environmental problems has compromised efforts to constrain the EPA’s regulatory authority. There are plenty of Americans who are suspicious of federal regulation, but they nonetheless prefer federal environmental regulation to no environmental protection at all.¶ The failure of antiregulatory conservatism is on display in the current fight over federal regulation of greenhouse gases. House Republicans oppose such regulation (and for good reason), but they have not put forward any alternatives (and many refuse to accept that climate change could be a problem). Nonetheless , federal regulation of GHGs marches on. The EPA just proposed another set of rules last month. Blanket opposition to federal GHG regulation failed to prevent (or even appreciably slow) such regulation , and time is running out. As GHG emission controls get imposed, and companies invest in the necessary control technologies, the political support for constraining EPA authority over GHGs will decline. Utilities may not want to give up coal or install costly carbon capture technologies, but once they’ve made these investments they will hardly want to see the regulations removed. If these rules are not stopped soon, it will be too late. This is a reason even those who refuse to accept the likelihood of climate change should consider alternatives to command and control regulation. Shouting “No” is insufficient to ensure success. AT Labor Shortages Farm labor shortages are a myth that repeatedly fail to pan out John Carney 12, CNBC, “More Data on The Phony Farm Labor Crisis,” 30 Aug 2012, http://www.cnbc.com/id/48847903 It’s become something of an annual tradition.¶ Every summer , newspapers around the country roll out stories of a labor shortage on farms. The fruit is going to rot in the orchards, crops will go unpicked, agricultural communities will be devastated unless something is done, the stories predict.¶ Here’s a pretty typical version of the story, as told by the San Francisco Chronicle:¶ But the American Farm Bureau Federation and its California chapter believe there is plenty of reason to worry.¶ "There have been instances in which growers had to disc up whole crops because they didn't have the workforce to harvest," said Kristi Boswell, the Farm Bureau's director of congressional relations. She points to Georgia, where whole tomato fields were plowed under last year. "The workforce has been decreasing in the last two to three years, but last year it was drastic."¶ And farmers are saying this year is even worse.¶ In recent years, we’ve seen You’d think the idea of a labor shortage in the midst of on ongoing job drought would be laughed out of town. (Read more: When Are You Retiring?...How Does Never Sound?)¶ But somehow the story keeps getting told and taken seriously. According to the Farm Labor Shortage Mythologists, American citizens are just too lazy and would rather go jobless than work on a farm.¶ Here’s how the Chronicle puts it:¶ Growers have tried to hire more domestic just how resilient this narrative really is. workers and had hoped that with unemployment rates so high, they'd find local labor.¶ "But domestic workers don't stick around to finish the job," Boswell said. "It's labor-intensive and often the flunked the test of reality . As it turned out, 2011 was one of involves hand-picking in grueling conditions."¶ Earlier this summer, as these stories started up once again, I tried to introduce some sanity into the discussion by pointing out that dire farm labor shortage forecasts of 2011 had absolutely the best years on record for American farms.¶ One of the criticisms of my reading of the farm data was that the data about national farm income might have been concealing more local crises. Perhaps the national numbers had come in well despite the alleged “farm crisis,” because fields of soy and corn in the Midwest had done well, while State-by-state data released by the Department of Agriculture Wednesday has pretty much destroyed that critique.¶ Let’s start with California. The San Francisco Chronicle had warned:¶ Farmers across California are experiencing the same problem: Seasonal workers who have been coming for decades to help with the harvest, planting and pruning have dropped off in recent years. With immigration crackdowns, an aging Mexican population, drug wars at the border and a weakened job market in the the apples of Washington, peanuts of Georgia, and California fruits and vegetables went unpicked.¶ Not quite. United States, the flow of migrants has stopped and may actually have reversed, according to the Pew Hispanic Center, a nonprofit, nonpartisan public policy research firm that has been So what happened?¶ Farm profits, what the Ag Department calls “net cash income,” in California rose from $11.1 billion in 2010 to $16.1 in 2011, an eye-popping 45 percent growth.¶ In Washington, the apple harvest was going to be devastated by a labor studying the trend.¶ Farm profits instead rose from $1.98 billion to $3.14 billion, a 58 percent rise.¶ Georgia was going to have a “rancid harvest” due to its farm labor shortage, according to The Georgia Report. I guess those peanuts picked themselves, because farm profits in the state rose from $2.5 billion to $2.6 billion.¶ The mythological Arizona farm labor shortage was supposedly “ destroying” its farm sector . shortage. Somehow or another , farm profits rose from $734 million to $1.3 billion.¶ Not every state saw profits boom. Farms in Arkansas, Delaware, Hawaii, Louisiana, New Hampshire, North Carolina, Rhode Island and South Carolina did less well in 2011 than 2010. Alabama’s farms saw net cash income fall off a cliff, from $1.2 billion to $773 million.¶ But clearly this national epidemic of farm labor shortages just never happened. Farmers can just raise wages to attract more workers John Carney 13, CNBC, 5/24/13, “Famers Solve Labor Shortage by Raising Pay,” http://www.thedailybeast.com/articles/2013/05/24/famers-solve-labor-shortage-by-raising-pay.html Farm owners have responded to fears of a labor shortage by actually raising wages by a little bit.¶ The Department of Agriculture reports:¶ Farm operators paid their hired workers an average wage of $11.91 per hour during the April 2013 reference week, up 4 percent from a year earlier. Field workers received an average of $10.92 per hour, up 4 percent from a year earlier. Livestock workers earned $11.46, up 51 cents. The field and livestock worker combined wage rate, at $11.10 per hour, was up 48 cents from a year earlier. Hired laborers worked an average of 40.3 hours during the April 2013 reference week, compared with 39.2 hours a year earlier.¶ Maybe someone should tell the Partnership for a New American Economy about this. It's just crazy enough that it may work!¶ By the way, don't worry about this bankrupting the great American farmer . Farm profits are expected to rise by more than 13 percent this year—to more than double what they were as recently as 2009. AT Brazil Solves Brazil doesn’t solve---lack of infrastructure Alistair Stewart 13, South America Correspondent for DTN/The Progressive Farmer, 8/6/13, “Brazilian Farming's Biggest Problem,” http://www.dtnprogressivefarmer.com/dtnag/view/blog/getBlog.do;jsessionid=2F48BC2D3422FCE86B5 624EA3DE307B4.agfreejvm1?blogHandle=southamerica&blogEntryId=8a82c0bc3e43976e014055b03e6 41466 Everybody knows the biggest problems facing Brazilian agribusiness. It's logistics .¶ When the cost of transporting soybeans from the fields in central Mato Grosso to a China-bound ship reach 40% of the commodity's value, you have a serious problem. ¶ Brazil more or less stopped heavy infrastructure investment in the 1970s. So when grain production began to ramp in the center-west, in areas thousands of miles from the sea, during the 1990s and 2000s, the roads, rail and port facilities eventually became overrun .¶ The situation gradually deteriorated before descending into chaos over the last year .¶ The freight cost from Sorriso (center-north Mato Grosso) to Paranagua port rose 50% this season, reaching $3.19 per bushel at the peak of the soybean harvest, while ships were waiting for up to 70 days to load beans at Paranagua. With demurrage costs (basically ship rental) at $15,000 to $20,000, that wait can be very expensive.¶ "While we don't resolve the logistics questions, we are inviting other players to enter into the game," said Julio Toledo Piza, chief executive at BrasilAgro, a corporate farm, at an industry conference to discuss logistics in Sao Paulo Monday. ¶ Brazil is the only major grain producer with substantial arable land available to exploit and so is in a great position to feed a growing world population over the next thirty years, as long as it works out a cheap way of delivering the grain, feed and meat . If not, alternative producers in South America, Africa and Asia will grow. ¶ "This is a huge opportunity that we can't let slip through our grasp," said Jorge Karl, director of OCEPAR, the farm coops association in Parana state. ¶ Governors have belatedly woken up to the logistical chaos that surrounds them and the federal government has recently given greater priority to infrastructure development. ¶ As a result, a series of plans to improve infrastructure have started moving forward and progress has been made on key grain corridors, such as the North-South railway, which will eventually connect the new eastern Cerrado soybean fields to ports in the North.¶ "The situation is vastly improved ... We hope that in six to seven years we can clear the logistics deficit," said Bernardo Figueiredo, head of EPL, the government's new infrastructure plan overseer. ¶ But farmer leaders point out that while more attention is now being paid to the infrastructural shortcomings, chronic delays are still the norm .¶ For example, farmers have been waiting for asphalt along the BR163 highway, which connects central Mato Grosso with northern Amazon ports, for a decade. Work has progressed over the last two years, leading many to believe the Transport Ministry forecast that it will be ready in 2014. However, just this week the Ministry admitted that the asphalt may only be completed in 2015, or after. ¶ Similarly, an East-West railway that will connect the soy and cotton fields of western Bahia to the sea is due to be complete in 2015, but work is yet to start on many stretches and, realistically, beans will flow along this route in 2018 at the earliest. ¶ Many other projects have been similarly delayed or suspended due to problems with implementation.¶ Faced with spiraling freight costs, the patience of farm leaders has been wearing thin for a while.¶ "We are told that things are improving but we can't wait. The inability of the government to deliver projects on time is unacceptable," according to Carlos Favaro, president of the Brazilian Soybean and Corn Growers Association (APROSOJA-MT).¶ None of the major new grain infrastructure projects will be ready for upcoming 2013-14 season, and with soybean area set to grow by 5% or more, the logistical chaos could deepen next year.¶ "We are going to have to muddle through next season," said Luis Carlos Correa Carvalho, president of the Brazilian Agribusiness Association (ABAG). ¶ "Realistically, the situation is only likely to improve in 2016-17," he added.¶ The high logistics costs will likely slow growth in Brazilian grain production over the next couple of years, but there remains so much pent up demand for logistics in Brazil that any new export corridor will be inundated as soon as it opens.¶ Speakers at the conference agreed that the fault for the current situation lies with government incompetence, as there are ample funds to invest. ¶ "Credit is everywhere. The money isn't the problem. The government has to allow it to be invested," said Ingo Ploger, president of the Business Council of Latin America (CEAL).¶ So why is government so sluggish?¶ For years, the problem was a lack of political will. Simply, farm infrastructure was not a vote winner. ¶ More recently, the realization that Brazil needs farm revenues to underpin the foreign accounts led President Dilma Rousseff to prioritize infrastructure. ¶ However, after 30 years of neglect, the know-how and systems to implement big rail, road and port projects just aren't there , explained EPL's Figueiredo.¶ "We suffer because plans drawn up are often of poor quality and environmental and other authorities don't have expertise in assess corruption remains rife in Brazil's construction industry. As a result, suspicion of graft is widespread. That means government auditors are quick to suspend projects when suspicions arise.¶ Until these problems are solved, Brazil will continue to have production costs 10% above its competitors in the U.S. and Argentina, noted CEAL's Ploger. them in a timely manner. That's a big reason why processes are delayed," he explained to the conference. ¶ Meanwhile, US is the world’s key food producer Brown 11 (The world is closer to a food crisis than most people realise Lester R. Brown guardian.co.uk, Tuesday 24 July 2012 07.21 EDT• Lester R. Brown is the president of the Earth Policy Institute and author of Full Planet, Empty Plates: The New Geopolitics of Food Scarcity, due to be published in October The United States is the leading producer and exporter of corn, the world's feedgrain. At home, corn accounts for four-fifths of the US grain harvest. Internationally, the US corn crop exceeds China's rice and wheat harvests combined. Among the big three grains – corn, wheat, and rice – corn is now the leader, with production well above that of wheat and nearly double that of rice. The corn plant is as sensitive as it is productive. Thirsty and fast-growing, it is vulnerable to both extreme heat and drought. At elevated temperatures, the corn plant, which is normally so productive, goes into thermal shock. As spring turned into summer, the thermometer began to rise across the corn belt. In St Louis, Missouri, in the southern corn belt, the temperature in late June and early July climbed to 100F or higher 10 days in a row. For the past several weeks, the corn belt has been blanketed with dehydrating heat. Weekly drought maps published by the University of Nebraska show the drought-stricken area spreading across more and more of the country until, by mid-July, it engulfed virtually the entire corn belt. Soil moisture readings in the corn belt are now among the lowest ever recorded. While temperature, rainfall, and drought serve as indirect indicators of crop growing conditions, each week the US Department of Agriculture releases a report on the actual state of the corn crop. This year the early reports were promising. On 21 May, 77% of the US corn crop was rated as good to excellent. The following week the share of the crop in this category dropped to 72%. Over the next eight weeks, it dropped to 26%, one of the lowest ratings on record. The other 74% is rated very poor to fair. And the crop is still deteriorating. Over a span of weeks, we have seen how the more extreme weather events that come with climate change can affect food security. Since the beginning of June, corn prices have increased by nearly one half, reaching an all-time high on 19 July. Although the world was hoping for a good US harvest to replenish dangerously low grain stocks, this is no longer on the cards. World carryover stocks of grain will fall further at the end of this crop year, making the food situation even more precarious. Food prices, already elevated, will follow the price of corn upward, quite possibly to record highs. Not only is the current food situation deteriorating, but so is the global food system itself. We saw early signs of the unraveling in 2008 following an abrupt doubling of world grain prices. As world food prices climbed, exporting countries began restricting grain exports to keep their domestic food prices down. In response, governments of importing countries panicked. Some of them turned to buying or leasing land in other countries on which to produce food for themselves. Welcome to the new geopolitics of food scarcity. As food supplies tighten, we are moving into a new food era, one in which it is every country for itself. The world is in serious trouble on the food front. But there is little evidence that political leaders have yet grasped the magnitude of what is happening. The progress in reducing hunger in recent decades has been reversed. Unless we move quickly to adopt new population, energy, and water policies, the goal of eradicating hunger will remain just that. Time is running out. The world may be much closer to an unmanageable food shortage – replete with soaring food prices, spreading food unrest, and ultimately political instability– than most people realise. Ext UQ Nutrient runoff from farms is currently unregulated Ken Kirk 12, Executive Director at the National Association of Clean Water Agencies, JD from Georgetown University Law Center and Master’s in Environmental Law from GWU Law School, June 20 2012, “Is the Clean Water Act Broken and Can We Fix It?” http://blog.nacwa.org/is-the-clean-water-actbroken-and-can-we-fix-it/ Much has been written and will continue to be written about the Clean Water Act this year as we celebrate the Act’s 40th anniversary. I don’t think anyone would argue the fact that the Act has been great for our waterways or that we would be much worse off without it.¶ But now we’ve entered a much more difficult and complex period in the history of the Clean Water Act. Federal money continues to dry up. Costs continue to increase. And requirements have become more stringent. We also have many new challenges to deal with including climate change, altered weather patterns, and population growth, to name just a few.¶ And we’re still struggling how to address wet weather and related runoff, particularly as this relates to with nutrients —arguably one of our biggest clean water issues right now. And yet, nonpoint sources remain unregulated . How has this been allowed to continue?¶ Let’s look back in time for a minute. Before the Clean Water Act, discharge was largely unregulated. We know this was not a good thing. Then, during the first 20 years of the Act, the focus was on wastewater treatment from domestic and industrial point sources and the broad use of secondary treatment to accomplish the Act’s goals. I believe this was the right thing to do at the time.¶ Unfortunately, this entire time, nonpoint sources— including agricultural operations—have completely avoided regulatory responsibility for their share of the water pollution problem. If we continue to ignore this very major source of water quality impairment, then it will come at great cost to all taxpayers. Despite water regulations, agricultural runoff is not restricted Neila Seaman 13, Director, Sierra Club Iowa Chapter, “IOWA LEADERS STILL NOT PROTECTING WATER QUALITY,” http://iowa.sierraclub.org/CAFOs/IALeadersStillNotProtectingWaterQuality.htm There are two sources of water pollution. Point sources are cities, towns and industries that treat their wastewater before it’s piped into a receiving water resource. They are heavily regulated. Non-point sources are primarily agricultural runoff from fertilizer and manure and there are basically no regulations on those sources . The point sources and non-point sources pointing their fingers at each other as the problem has created a huge logjam that has spanned decades. As a result, the point sources are now highly regulated and the non- point sources remain unregulated . Ext Food Insecurity -> War Food shortages cause nuclear world war 3 FDI 12, Future Directions International, a Research institute providing strategic analysis of Australia’s global interests; citing Lindsay Falvery, PhD in Agricultural Science and former Professor at the University of Melbourne’s Institute of Land and Environment, “Food and Water Insecurity: International Conflict Triggers & Potential Conflict Points,” http://www.futuredirections.org.au/workshoppapers/537-international-conflict-triggers-and-potential-conflict-points-resulting-from-food-and-waterinsecurity.html There is a growing appreciation that the conflicts in the next century will most likely be fought over a lack of resources.¶ Yet, in a sense, this is not new. Researchers point to the French and Russian revolutions as conflicts induced by a lack of food. More recently, Germany’s World War Two efforts are said to have been inspired, at least in part, by its perceived need to gain access to more food. Yet the general sense among those that attended FDI’s recent workshops, was that the scale of the problem in the future could be significantly greater as a result of population pressures, changing weather, urbanisation, migration, loss of arable land and other farm inputs, and increased affluence in the developing world.¶ In his book, Small Farmers Secure Food, Lindsay Falvey, a participant in FDI’s March 2012 workshop on the issue of food and conflict, clearly expresses the problem and why countries across the globe are starting to take note. .¶ He writes (p.36), “…if people are hungry, especially in cities, the state is not stable – riots, violence, breakdown of law and order and migration result.” ¶ “Hunger feeds anarchy.”¶ This view is also shared by Julian Cribb, who in his book, The Coming Famine, writes that if “large regions of the world run short of food, land or water in the decades that lie ahead, then wholesale, bloody wars are liable to follow.” ¶ He continues: “An increasingly credible scenario for World War 3 is not so much a confrontation of super powers and their allies, as a festering , self-perpetuating chain of resource conflicts.” He also says: “The wars of the 21st Century are less likely to be global conflicts with sharply defined sides and huge armies, than a scrappy mass of failed states, rebellions, civil strife, insurgencies, terrorism and genocides, sparked by bloody competition over dwindling resources.”¶ As another workshop participant put it, people do not go to war to kill; they go to war over resources, either to protect or to gain the resources for themselves.¶ Another observed that hunger results in passivity not conflict. Conflict is over resources, not because people are going hungry. ¶ A study by the I nternational P eace R esearch I nstitute indicates that where food security is an issue, it is more likely to result in some form of conflict. Darfur, Rwanda, Eritrea and the Balkans experienced such wars. Governments, especially in developed countries, are increasingly aware of this phenomenon. ¶ The UK Ministry of Defence, the CIA, the US C enter for S trategic and I nternational S tudies and the Oslo Peace Research Institute, all identify famine as a potential trigger for conflicts and possibly even nuclear war . Ext IOOS Causes Regulation Improved ocean observation would allow for regulation of agricultural runoff Mark J. Kaiser 4, PhD, professor and director of the Research & Development Division at the Center for Energy Studies at Louisiana State University, PhD and Allan G. Pulsipher, Professor of Energy Policy in the Center for Energy Studies and Professor in the Department of Environmental Studies at Louisiana State University, PhD in Economics from Tulane University, “The potential value of improved ocean observation systems in the Gulf of Mexico,” Marine Policy 28 (2004) 469–489, Science Direct Waterways draining into the GOM transport wastes from 75% of US farms and ranches, 80% of US cropland, hundreds of cities, and thousands of industries located upstream of the GOM coastal zone. Urban and agricultural runoff contributes large quantities of pesticides, nutrients , and fecal coliform bacteria. ¶ Activities that have contributed or are still contributing to the degradation of coastal water conditions along the Gulf Coast include the petrochemical, industrial, agricultural, power plants, pulp and paper mills, fish processing, municipal wastewater treatment, maritime shipping, and dredging. Nonpoint sources are difficult to regulate and currently have the greatest impact on the GOM coastal water quality. Nonpoint pollutant sources include agriculture , forestry, urban runoff, marinas, recreational boating, and atmospheric deposition. ¶ One of nutrients of the greatest concerns for GOM coastal water quality is an excess which can lead to noxious algal blooms, decreased seagrasses, fish kills, and oxygen-depletion events. Improved ocean observation systems is expected to allow environmental activity in the region to be better understood and monitored . Regulation is currently impossible because it’s hard to determine the cause of blooms--the plan allows for attribution that makes regulations possible David Caron 7, Professor of Biological Sciences at USC, PhD in Biological Oceanography from MIT and Woods Hole Oceanographic Inst., and Burt Jones, Adjunct Professor of Biological Sciences at USC, PhD in Biological Oceanography from Duke, “Making Use of Ocean Observing Systems: Applications to Marine Protected Areas and Water Quality,” Sept 25-26 2007, http://calost.org/pdf/resources/workshops/OceanObserving_Report.pdf Recent summaries of available information have indicated that coastal ecosystems have witnessed a general increase in the occurrence and severity of harmful and toxic algal blooms. The coastline of California is no exception, with powerful neurotoxins such as saxitoxin and domoic acid now commonly observed throughout the state. Numerous factors have been implicated as possible contributors to these coastal algal blooms. Harmful blooms can result from natural, seasonal supply of nutrients to coastal waters during upwelling and from anthropogenic inputs of nutrients in river discharges and land runoff.¶ Unfortunately, quantifying the contribution of the many potential sources of nutrients that can support algal blooms is a daunting task because of the difficulty of maintaining continuous observations in the ocean. Harmful algal blooms are ephemeral events that can develop quickly and dissipate before their causes can be adequately characterized. Efforts by municipalities, counties and the state to provide responsible environmental stewardship are often thwarted by the lack of sufficient observational capabilities to document water quality, let alone determine the cause(s) of adverse events such as harmful algal blooms. Partnerships are desperately needed between ocean observing programs, research/academic institutions, coastal managers and monitoring programs (including both government and nonprofit) to grapple with the increasing number of environmental influences on algal population growth and toxin production in coastal marine ecosystems. ¶ Ocean observing systems provide a tool for monitoring, understanding and ultimately predicting harmful algal blooms. Sophisticated sensing instruments installed on piers, deployed from ocean of coastal waters buoys, or carried by mobile surface and underwater vehicles provide a ‘constant presence’ in the ocean that assists scientists to detect impending or emerging events, and guide their sampling effort. Aquatic sensors can provide information on salinity (which can identify sources of freshwater input to the coastal ocean), temperature (which yields information on the physical structure of the water column, a primary factor affecting algal growth in nature), chlorophyll fluorescence (which documents the biomass of algae in the water) and dissolved oxygen (which indicates biological activity and ecosystem health). A variety of ever-improving nutrient sensors can quantify specific substances that may stimulate algal growth (e.g., ammonium, nitrate, phosphate). In addition, new sensors that can detect specific microorganisms and/or toxins produced by these species are under development and eventually will increase the capabilities of ocean observing systems. These observing systems fill a fundamental gap in our ability to document harmful or toxic events, and aid our attempts to attribute these events to specific environmental causes. Ext Regulations Hurt Farms Even a small link is enough---farms operate on thin margins that make the cost of regulation devastating Daniel R. Mandelker 89, Stamper Professor of Law, Washington University in St. Louis, June 1989, “Controlling Nonpoint Source Water Pollution: Can It Be Done,” Chicago-Kent Law Review Vol. 65, Issue 2, http://scholarship.kentlaw.iit.edu/cgi/viewcontent.cgi?article=2776&context=cklawreview Another obstacle to controlling nonpoint pollution is that the nonpoint source may be unable to internalize the cost of the control or pass it on to consumers. 53 This problem particularly arises with controls on agricultural nonpoint sources . These controls can be expensive in an industry marked by thin margins and low profitability . Nor are farmers, as an unorganized production group, able to pass the costs of these controls on to consumers. 54 In contrast, nonpoint source land use controls applied to urban development may not present this problem. Urban developers may be able to pass the cost of these controls on to their consumers,55 and local governments can use density bonuses to offset the cost of controls necessary to reduce nonpoint pollution. Regulations would be ineffective and wreck the economic viability of farms AFT 13, American Farmland Trust’s Center for Agriculture in the Environment, August 2013, “Controlling Nutrient Runoff on Farms,” http://www.farmland.org/documents/FINALControllingNutrientRunoffonFarms.pdf Direct regulation of nutrient runoff from farms is highly unlikely in the United States (Williams 2002). The geographic dimensions make “federally designed, nationally uniform technology based performance and emissions standards” difficult to implement without a marked increase in budgeting for individual farm permitting, monitoring and enforcement. Local variations in weather, soil salinity, and soil erosion potential, leaching potential, and freshwater availability present further challenges to an effective national regulatory regime. Variations in crop type, production practices, livestock type and concentration, use of irrigation, tillage practices, sediment runoff and fertilizer runoff all contribute to the difficulty of “ one size fits all” regulation . Social factors like proximity to metropolitan area, and surrounding land use also influence farm practices. EPA has noted that a program of this breadth would make it very difficult to implement and enforce regulations. ¶ The economic dimensions of agriculture also pose barriers to regulation. Agriculture in the U nited S tates has vast economic value , yet is dispersed widely across the country and by landowner. Faced with the rising costs of inputs and equipment, the farm industry is quickly consolidating. Increased environmental regulation of farms may reduce their economic viability due to compliance costs . And the political dimensions, mentioned earlier, that make regulation of agriculture difficult include a consolidated voting block, strong lobbying and political pressure. The cost of complying would be prohibitive John Leatherman 4, Associate Professor in the Dept of Agricultural Economics at Kansas State, PhD from the University of Wisconsin, Craig Smith, Graduate Research Assistant in the Dept of Agricultural Economics at Kansas State, and Jeffrey Peterson, Assistant Professor in the Dept of Agricultural Economics at Kansas State, Agu 19-20 2004, “An Introduction to Water Quality Trading,” http://www.agmanager.info/events/risk_profit/2004/LeathermanPeterson.pdf A significant share of the current water quality problems in Kansas stem from nonpoint source pollution, such as urban storm water and agricultural runoff (KDHE, 2002). Applying the same “command and control” regulatory approach to nonpoint source pollution would be problematic for a number of reasons. The widely distributed nature of nonpoint source pollution would make regulation and compliance costs frighteningly prohibitive . It’s also likely there would be significant public opposition to broad regulation of the many activities contributing nonpoint source pollution. Counterplans Europe 1NC Europe CP Counterplan: The European Union should fully develop and implement an integrated European Ocean Observation System. Integrated European observation is key to broader data – tech is world class, just a question of implementation. EMD 2014 REPORT FROM THE JOINT EUROGOOS/EMODNET/EMB/JRC WORKSHOP AT THE EUROPEAN MARITIME DAY IN BREMEN,The importance of an integrated end-to-end European Ocean Observing System: key message of EMD 2014 http://eurogoos.eu/2014/06/09/eoos-at-emd-bremen-2014/ Ocean observations are essential for marine science, operational services and systematic assessment of the marine environmental status. All types of activities in the marine environment require reliable data and information on the present and future conditions in which they operate. Many maritime economic sectors (e.g. oil and gas exploration, maritime transport, fisheries and aquaculture, maritime renewable energy) directly benefit from easily accessible marine data and information in several ways: improved planning of operations, risk minimization though increased safety, improved performance and overall reduced cost. Other activities, such as deep sea mining and marine biotechnology, also benefit from specialized deepsea observations that were not feasible until recently. The complexity and high density of human activities in European seas and oceans result in a high demand for marine knowledge in the form of data, products and services to support marine and maritime activities in Europe, stressing the need for an integrated European approach to ocean observation and marine data management (Navigating the Future IV, European Marine Board 2013). While Europe already has a relatively mature ocean observing and data management infrastructure capability , this is largely fragmented and currently not addressing the needs of multiple stakeholders. Mechanisms for coordinating existing and planned ocean observations using a system approach are needed for more integrated, efficient and sustained observations under the framework of a “European Ocean Observing System” (EOOS) following international practice (systems developed by USA, Australia and Canada) and the call of the EurOCEAN 2010 Conference Declaration . The integration of different national and local marine data systems into a coherent interconnected whole which provides free access to observations and data, as pursued by the European Marine Observation and Data Network (EMODnet) is of key importance for maritime sectors like fisheries, the environment, transport, research, enterprise and industry. However, much work still needs to be done in close collaboration with end-users, in particular industry, to further develop EMODnet into a fully functional, fit for purpose gateway to European marine data and data products taking into account requirements of multiple users. There is a need for scienceindustry partnerships to stimulate innovation and develop a successful EOOS that will further enhance the contribution of marine observations to economic activities relevant for Blue Growth in Europe. Innovative technologies, developed in collaboration between research given several solutions during the past years for more robust, multi-parametric and systematic observations. This, in turn, is leading to new and more reliable operational services that support a wide range of maritime economic activities: fisheries and aquaculture, offshore oil and gas, marine renewable energy, maritime transport, tourism etc. Other services address the sectors of marine safety, climate and weather applications, as well as marine environmental assessment. At the end of the marine observations, data to knowledge scientists and the industry, have cycle, activities and tools are needed to create added value products for specific stakeholders, including the wider public, such as the European Atlas of the Seas which allows professionals, students and anyone interested to explore Europe’s seas and coasts, their environment, related human activities and European policies. At the same time, it is critical to evaluate whether we are monitoring/observing what we actually need. Regional assessments such as performed by the newly established EMODnet sea-basin “checkpoints” could provide relevant information, among others to advise Member States about requirements for essential and optimal observation capability. Infrastructure for data sharing already exists Paris and Schneider et al 2013 Jean-Daniel Paris, Nadine Schneider Laboratoire des Sciences du Climat et de l’Environnement/IPSL Data sharing across the Atlantic: Gap analysis and development of a common understanding http://www.coopeus.eu/wp-content/uploads/2013/12/Carbon-data-sharing-gaps-06-12-2013-D31.pdf The global need to develop large, cross continental environmental datasets has been recognized*'. To address this issue, COOPEU5 is a joint US National Science Foundation (*4SF) and European Commission FP7 (in the frame of the European Strategy Forum on Research Infrastructures, ESFRI) supported project initiated in September 2012s. Its main goal is creating a framework to develop interoperable einfrastructures across several environmental and geoscience observatories in Europe and US. The National Ecological Observatory Network (NEON in the US, www.neoninc.org) and the Integrated Carbon Observatory System (ICOS in the EU, http://www.icos-infrastructure.eu/) are two of these governmental supported observatories. Here, the data products from these two observatories are centered around greenhouse gas (GHG) concentration, carbon and energy flux observations, and the surface micrometeorology surrounding these measurements. The objective of COOPEUS is to coordinate the integration plans for these carbon observations between Europe and the US. Even though both ICOS and NEON have the ability to collaborate on effective issues, we fully recognize that this effort cannot be effectively accomplished without the engagement of many other partners, such as; National Oceanic and Atmospheric Administration's Global Monitoring Division (US NOAA GMD), the Group on Earth Observations (GEO, www.earthobservations.org) and the Group of Earth Observations System of Systems (GEOSS), World Meteorological Organization (WMO, www.wmo.int), the Belmont Forum (www.igfagcr.org), ^SF- supported EarthCube (earthcube.ning.com)and DataOne (www.dataone.org) projects, and a wide variety of regional-pased flux networks (i.e., AmeriFlux, Fluxnet). Of course as COOPEUS continues to advance, this list of partners are not exclusive and are expected to increase. COOPEUS aims to strengthen and complement these partnerships through a variety of governance mechanisms, some informal and others formalized (e.g. Memorandum of Understanding), tailored to each individual organizational governance structure. Several of these organizations (mentioned above) have a history of collaborating and sharing of data. In this historical context, we also have recognized what has worked, exiting limitations, and what can be improved in terms of data sharing and interoperability. This COOPEUS work task is building upon these relationships and working history to strengthen these approaches and collaboration. Yes Data Sharing Europe will share data with the US – solves the aff COOPEUS 2013 Project CoopEUS WP 4 – Ocean Observation Deliverable D4.1- Fact Finding Report http://www.coopeus.eu/wp-content/uploads/2013/11/D4_1_Facts_Finding_10_03_FIN.pdf Sharing scientific data can be the 'playing-field' for an easy and profitable commencement of the scientific structured collaboration while complementing respective expertise. This approach also meet the recent trend of EU and US funding agencies moving towards more openness and more sharing. The analysis and definition of the basic principles and requirements for a shared data policy can be a first step for a long-term transatlantic cooperation and can pave the ground for the data integration on a global scale diminishing significant misalignment in the data issues. Besides, the adoption and implementation of common data standards can facilitate the development of infrastructure and data infrastructures that will be largely interoperable. The international ICT collaborative frame for the interoperability of scientific data and infrastructures has to be leveraged to take advantages of the continuous progresses and specialized skills. Parallel initiatives and projects at global and continental scale have to be exploited with the aim of supporting the on-going networking process of the Ocean Scientists and ICT engineers and specialists European ODIP, EUDAT iCORDI GENESI-DEC, GEOwow GPOD, TATOO and worldwide Research Data Alliance, DAITF( http://www.daitf.ore/ ) to mention some. The present development status of the research infrastructures for ocean observations 001 and EMSO enable, as they are now, to outline data sharing practices and protocols across borders and disciplines although with some limitations. Limiting factors for an easy and extensive data use out of the respective data production frame and user community can be misalignment of vocabularies and ontology, metadata Incompleteness and heterogeneities, non standardized QC/QA practices. These issues shall be the subjects of deepening discussion in CoopEUS to achieve reciprocal understanding and smooth data interoperability and exchange. EU Tech Good European sensors are cutting edge and real time – no distinction between them and the aff. COOPEUS 2013 Project CoopEUS WP 4 – Ocean Observation Deliverable D4.1- Fact Finding Report http://www.coopeus.eu/wp-content/uploads/2013/11/D4_1_Facts_Finding_10_03_FIN.pdf EMSO will pioneer in delivering multidisciplinary real-time data from the sea by providing data from the ocean surface through the water column to the benthos and sub-seafloor. It will be facilitated, in part, by advancements made on Eulerian (fixed) observatory infrastructures during the ESONET Network of Excellence, EuroSITES programs and the potential follow-on project Fixed- point Open Ocean Observatories (Fix03). And it will work alongside other key ESFRI infrastructures such as Euro-Argo, SIOS and EPOS. The implementation and management of EMSO infrastructure are based on an European transnational integration effort at EU Members State Governments level and on the commitment of research institutions from thirteen countries. The long-term sustainability and the close interaction with the reference European scientific community, coordinated through the ESONET- Vision upcoming association, is ensured through the establishment of a European Consortium (ERIC). Science objectives guide observatory design and dictate the ability to collect data without employing traditional means, such as research vessels. However, the latter are intended to complement the EMSO network, which will be serviced by a combination of research vessels and ROVs operations provided by EU Member States in a coordinated network established by EUROFLEETS. EMSO nodes include cabled and stand-alone observatories with moorings and benthic instruments, while communicating in real time or in delayed mode, and being serviced through periodic maintenance cruises. Mobile systems, such as ROVs and AUVs will improve the potential of some of the observatories by expanding their spatial extent. EMSO nodes are conceived as an integration of local and regional seafloor and water column in situ infrastructures equipped for both generic and science oriented approach (Ruhl et al., 2011). Generic sensor module While not all scientific questions can be addressed by each individual infrastructure, specific set of variables measured at all EMSO sites and depths are considered, including temperature, conductivity (salinity), pressure (depth), turbidity, dissolved oxygen, ocean currents, and passive acoustics. These generic sensors can be used to directly address a wide range of geo-hazard warning and scientific applications related to understanding natural and anthropogenic variation and the possible impacts of climate change. In the observatory setting, these data can then be relayed back to shore via seafloor cable (real-time) or satellite telemetry (delayed time). The use of salinity and conductivity sensors spaced regularly along strings and additional ADCP coverage can capture the themes related to ocean physics. These include understanding wind- driven and deep-ocean circulation, planetary waves, and interactions between the Benthic Boundary Layer and the seabed. Solves Environment Europe solves ecosystem management – academic, government, industry partnerships ESF 2014 European Science Foundation, Arctic 2050: Towards ecosystem-based management in a changing Arctic Ocean, March 12 2014, http://www.esf.org/media-centre/ext-single-news/article/arctic-2050-towardsecosystem-based-management-in-a-changing-arctic-ocean-1011.html About 150 scientists, policy makers and members of industry are gathering today at the 4th European Marine Board Forum in Brussels to discuss how best to manage the consequences of a changing Arctic Ocean for human health and well-being. The European Marine Board has convened this flagship event in collaboration with the European Polar Board, working in association with the European Science Foundation, in the knowledge that industry and science must work together to achieve sustainable management of resources such as fishing and oil and gas exploration while at the same time, protecting and conserving the Arctic environment. Dramatic changes, largely attributed to anthropogenic activity, have taken place in the Arctic in recent decades. These changes include melting of glaciers and sea ice, altered oceanic current patterns, movement and accumulation of contaminants and range shifts in many species. As a result of these changes the Arctic region is being transformed, with wide-ranging impacts and opportunities including the potential for ice-free shipping routes in the future, increased activity in oil and gas exploration, changes to Arctic fisheries and biodiversity, and impacts on residents’ livelihoods. “At present we are unprepared for the environmental and societal implications of increased human access to the Arctic that will come with the receding ice” explains Professor Peter Haugan from the University of Bergen and vice-Chair of the European Marine Board. “We have not fully anticipated the consequences of an increase in activities like hydrocarbon exploration, mineral extraction, bioprospecting and pelagic and demersal fisheries”. The 4th EMB Forum, recognized as an official ICARP III event, promotes the need for an ecosystem-based management approach in the Arctic Ocean, in order to adapt to and manage rapid environmental change and commercial exploitation, supporting a key recommendation of the recently published Arctic Biodiversity Assessment.[1] Moderated by David Shukman, BBC Science Editor, forum sessions include, ‘Living with a Changing Arctic Ocean’, ‘Utilizing and managing Arctic Ocean resources’ and a session on ‘Arctic Ocean Observation’, building on the European Marine Board call in 2013 for urgent action to increase our observational capacity across the entire Arctic Ocean (EMB, 2013).[2] Speakers will include industry representatives from Shell, the International Association of Oil and Gas Producers and the International Maritime Organisation. The forum provides a platform to address ecosystem-based management in the Arctic Ocean by stimulating dialogue across sectors to aid common understanding, collaborative actions and sustainability targets. Later today the forum will culminate with an open panel discussion on the roles of industry and science in achieving sustainable management of the Arctic Ocean. Solves Acidification Europe can develop acidification solutions – experimentation. Data is sufficient. Riebesell 2009 Ulf, professor of marine biogeochemistry and Head, biological Oceanography, leibniz institute of marine sciences, eXperimeNtAl ApprOAcHes iN OceAN AcidiFicAtiON reseArcHhttp://www.tos.org/oceanography/archive/22-4_gattuso.pdf Progress in our ability to make reliable predictions of the impacts of ocean acidification on marine biota critically depends on our capability to conduct experiments that cover the relevant temporal and spatial scales. One of the greatest challenges in this context will be scaling up biotic responses at the cellular and organismal level to the ecosystem level and their parameterization in regional ecosystem and global biogeochemical models. EPOCA employs a suite of experimental approaches to assess marine biota's ocean acidification sensitivities, ranging from single species culture experiments to field surveys of the distribution of potentially sensitive caxa In relation to seawacer carbonate chemistry (Figure BS). Each of these approaches has its distinct strengths and weaknesses. 3ottle and microcosm experiments allow for high replication of multiple CO?/pH treatments jndei well-controlled experimental conditions, thereby yielding high statistical power. However, they typically lack genetic and species diversity, competitive interac- tion, and the trophic complexity of natural systems, which complicates extrapolation of results to the real world. Field observations, on the other hand, cover the full range of biological Interactions and environmental complexities, but they generally provide only a snapshot In time, with little or no information on the history of the observed biota and environmental conditions prior to sampling. The interpretation of fi eld data In terms of dose/response relationships Is often obscured by multiple environmental factors simultaneously varying in time and space and by the lack of replication. Mesocosms, experimental enclosures that are designed to approximate natural conditions and that allow manipulation of environmental factors, provide a powerful tool to link small-scale single species laboratory experiments with observational and correla* tlve approaches applied In field surveys. A mesocosm study has an advantage over standard laboratory tests In that It maintains a natural community under close to natural selfsustaining conditions, taking Into account relevant aspects of natural systems such as indi- rect effects, biological compensation and recovery, and ecosystem resilience. Replicate enclosed populations can be experimentally manipulated and the same populations can be sampled repeatedly over time. Further advantages of flexible-wall. In situ enclosures are that a large volume of water and most of its associated organisms can be captured with minimal disturbance. The mesocosm approach is therefore often considered the experimental ecosystem closest to the real world, without losing the advantage of reliable reference conditions and replication. To Improve understanding of the under- lying mechanisms of observed responses, which are often difficult to infer from mesocosm results, and to facilitate scaling up mesocosm results, large-scale enclosure experiments should be closely integrated with well-controlled laboratory experiments and modeling of ecosystem responses. Taking advantage of a recently developed mobile, flexible wall mesocosm system. EPOCA will conduct a joint mesocosm CO, perturbation experiment In the high Arctic, involving marine and atmospheric chemists, molecular and cell biologists, marine ecolo- gists. and biogeochemists. A total of nine mesocosm units 2 m In diameter and lS-m deep, each containing approximately 45,000 liters of water, will be deployed In Kongsfjord off Ny-Alesund on Svalbard. The carbonate chemistry of the enclosed water will Initially be manipulated to cover a range of pCO_, levels from pre-lndustrlal to projected year 2100 values (and possibly beyond) and will be allowed to float freely during the course of the experiment to mimic varia- tions naturally occurring due to biological activity. The high level of scientific Integration and cross-disciplinary collaboration of this study Is expected to generate a comprehensive data set that lends itself to analyses of communitylevel ocean acidification sensitivities and ecosyscem/biogeochemical model parameterizations User Fees 1NC Commercialization CP Text The United States Federal Government should adopt a user fee system of commercialization for the Integrated Ocean Observing System. Observation One: Competition It’s is possessive-Commercialization is not USFG exploration or development Nguyen and Kim 08 (Ngan L.T. Nguyen, graduate from the University of Natural Sciences, Vietnam and pursuing Masters in Computer Sciences at the University of Tokyo, and Jin-Dong Kim, Senior research Associate at University of Manchester, Project Lecturer at University of Tokyo, PhD and Masters in Computer Seicnce at Korea University, Department of Computer Science, University of Tokyo. 2008, Licensed under the Creative CommonsAttribution-Noncommercial-Share Alike 3.0 Unported, Exploring Domain Differences for the Design of Pronoun Resolution Systems for Biomedical Text, page 631. http://www.nltg.brighton.ac.uk/home/Roger.Evans/private/coling2008/cdrom/PAPERS/pdf/PAPERS079. pdf) The combination of C netype and P semw features exploits the co-ocurrence of the semantic type of the candidate antecedent and the context word, which appears in some relationship with the pronoun. This combination feature uses the information similar to the semantic compatibility features proposed by Yang (Yang et al., 2005) and Bergsma (Bergsma and Lin, 2006). Depending on the pronoun type, the feature extractor decides which relationship is used. For example, the resolver successfully recognizes the antecedent of the pronoun its in this discourse: “HSF3 is constitutively expressed in the erythroblast cell line HD6 , the lymphoblast cell line MSB , and embryo fibroblasts , and yet its DNA-binding activity is induced only upon exposure of HD6 cells to heat shock ,” because HSF3 was detected as a Protein entity, which has a strong association with the governing head noun activity of the pronoun. Another example is the correct anaphora link between “it” and “the viral protein” in the following sentence, which the other features failed to detect. “Tax , the viral protein , is thought to be crucial in the development of the disease , since it transforms healthy T cells in vitro and induces tumors in transgenic animals.” The correct antecedent was recognized due to the bias given to the association of the Protein entity type, and the governing verb, “transform” of the pronoun. The experimental results show the contribution of the domain knowledge to the pronoun resolution, and the potential combination use of such knowledge with the syntactic features. Parse features (parg) The combinations of the primitive features of grammatical roles significantly improved the performance of our resolver. The following examples show the correct anaphora links resulting from using the parse features: • “By comparison, PMA is a very inefficient inducer of the jun gene family in Jurkat cells. Similar to its effect on the induction of AP1 by okadaic acid, PMA inhibits the induction of c-jun mRNA by okadaic acid.” In this example, the possessive pronoun “its” in the second sentence corefers to “PMA”, the subject of the preceding sentence. Among the combination features in this group, one noticeable feature is the combination of C parg, Sdist, and P type which contains the association of the grammatical role of the candidate, the sentence-based distance, and the pronoun type. The idea of adding this combination is based on the Centering theory (Walker et al., 1998), a theory of discourse successfully used in pronoun resolution. This simple feature shows the potential of encoding centering theory in the machine learning features, based on the parse information. Observation Two: Net Benefits-The CP solves the case avoids spending related arguments-Creating a system of user fees incentivizes commercialization Woll-OCEANS Conference-9 MTS/IEEE Biloxi - Marine Technology for Our Future: Global and Local Challenges The role of commercialization in IOOS and the regional associations http://ieeexplore.ieee.org.ezproxy.uky.edu/xpls/icp.jsp?arnumber=5422377#authors Since its inception, IOOS has operated in much the same way as federal government agencies traditionally have - set policy, identify requirements, prioritize within those requirements, secure federal funding , and allocate the funding via the RAs, which in turn fund the organizations that actually perform the required tasks. The organizations funded by the RAs to perform the desired tasks have included universities, research institutions, and private sector companies. This has resulted in a loose organizational structure, in which federal (and sometimes state) agencies provide policy, coordination, and funding; RAs and their member organizations conduct the science, process data, and operate observing and other systems; products and services are generated and provided (usually free of charge) to end users. Fig. 1 illustrates this functional structure. Ideally, funding for these programs follows the “traditional” trajectory: a large sum at the beginning to start the program up, and then some fraction of that annually to support ongoing Operations and Maintenance (O&M). Fig. 2 depicts this “traditional” funding profile. While significant progress has been made recently, the unfortunate fact is that funding for both initial start up costs and O&M have been less than anticipated. As a result, worthy new projects remain on the sidelines, and viable observing assets funded previously are beginning to be decommissioned or inactivated due to inadequate O&M funding. In response to this situation, some have called for a retreat from the IOOS organizational concept and a return to individual efforts by member organizations. Others retain faith in the IOOS concept, but argue that the enterprise simply needs to be more effective at articulating the requirements and hazards of inaction to the appropriate members of Congress. This paper proposes a third option - the embrace of commercialization as a program objective for IOOS the RAs, and the oceanography enterprise. The active participation of universities and research institutions in IOOS and the RAs has been a cornerstone of their success to date and is a logical extension of previous practices in the academic arena. The avoidance of redundancy in effort and the benefits of cooperation have long been recognized in establishing and operating successful programs in the academic community, and participation in IOOS by university and research institutions has been high. However, the participation of private sector businesses has been less widespread . A partial list of reasons would include the lack of a clear answer to why it is in a company's interest to participate, a business environment that is generally more competitive than cooperative, and a desire to protect proprietary information. Unfortunately, the current structure and functionality of the RAs does little to address or correct those concerns, and in many ways actually exacerbates them. The linchpins to this proposal are the embrace of commercialization using private sector businesses as an organizational goal and the imposition of appropriate fees on the end users for the products and services they receive . Aside from attracting more business members to IOOS and the RAs, the imposition of fees would immediately establish a two-way market feedback mechanism between the end users and the IOOS/RA community. This feedback mechanism is the most effective means available to help identify those projects that truly add value and address significant end user needs - if no one is willing to pay for a particular product or service (or pay enough to cover the costs of generation), then by definition it is not worth the money being spent on it. A2: CP Illegitimate ---Literature supports the counterplan and makes it predictable. Our Woll evidence contrast the plan with the CP which proves it’s germane and educational ---Test “Its” which is a core source of negative ground on the topic ---Net benefits check abuse and provide a germane policy warrant for voting negative. They shouldn’t get to pick and choose which parts of the plan to defend. ---Doesn’t waste the 1AC-The affirmative still sets the the initial grund for the debate. The negative must still find some aspect of the affirmative plan with which to compete. The negative isn’t obligated to run arguments that were preempted in the 1AC. It’s more fair to let both teams partially determine ground for the debate. ---Don’t trivialize the debate-If the difference between the plan and the counterplan is large enough to generate a net benefit, than it’s worth debating. This argument isn’t unique as many affirmatives are only a small departure from the status quo. ---Risk-Reward-The affirmative generates offense from the way their plan would be implemented. That reward should be offset by allowing the negative to test the plan. ---Punishment doesn’t fit the crime-The judge should evaluate theory like extratopicality. The counterplan should be judged outside their jurisdiction and a lost option for the negative to advocate. A2: Conditionality ---Real World-Policy makers do consider multiple options at once. Their argument guts one of the core elements of policy discussion. ---Best policy justifies-Multiple options make it more likeley that the best policy will be found. The role of the judge is to endorse the best policy at the end of the round. If a conditional counterplan has been proven to be the best policy, it’s perverse not to allow it to be endorsed. ---Education-Argument breadth has benefits. If depth were the only value, teams wouldn’t be allowed to debate more than one advantage or disadvanatge per round. Exploring the range of issues on a subject is also intellectualy important. ---Time limits aren’t an answer A. Time is finite in debate. Running one argument inherently trades off with another. B. Other arguments make this non-unique. Multipe topicality arguments, two card disads, or kritiks equally distort time. C. Creating time pressure and making time based decisions is an inherent part of debate strategy. It’s an acceptable part of all other debate arguments. ---Counterplans don’t introduce unique complexity into the round. The counterplan may just be a minor alteration of the plan. Disadvantage s also raise multiple issues. ---Permutations justify-Retaining the status quo as an option is reciprocal to the affirmative’s ability to advocate the plan or permutation. ---Conditionality is reciprocal to the affirmative’s ability to select a case . Since the affirmative selects the ground for the debate they enjoy a huge preparation advantage. Allowing hypothetical negative arguments helps to defeat this edge. ---Advocacy concerns aren’t decisive. A. In the real world, policies are attacked from avariety of perspectives. In debate there is only one negative team, so to encompass the true range of potential counter-affirmative advocacy, multiple positions must be allowed. B. Most debate practice isn’t consistent with the advocacy paradigm. Strategic concessions by the affirmative and permutations allow the affirmative to advocate multiple positions. ---Not a voting issue. Emphasis on punishment incentivizes a race to bottom discouraging substsantive debates. 2NC A2 Perm-Do the CP ---Severs it’s-Extend our Kim evidence-Commercialization is not USFG exploration or development-Severance permutations are illegitimate because no counterplan would compete if the 2AC could pick and choose which parts of the plan to defend. ---Severs increase-exploration doesn’t increase unless the private sector fills in via user fees. The CP doesn’t fiat an expansion of exploration. We read solvency evidence that commercialization would naturally expand the program ---Privatization is distinct from federal government Barbier 7 (Carl, US District Judge, TIEN VAN COA, ET AL VERSUS GREGORY WILSON, ET AL CIVIL ACTION NO: 07-7464 SECTION: J(1) UNITED STATES DISTRICT COURT FOR THE EASTERN DISTRICT OF LOUISIANA 2007 U.S. Dist. LEXIS 87653) However, in their motion to remand, Plaintiffs argue that as an independent contractor, P&J is not an employee of the federal government, and consequently does not enjoy derivative immunity and cannot invoke the FTCA. Plaintiffs cite United States v. New Mexico in support of the notion that private contractors, whether prime or subcontractors, are not government employees nor are they agents of the federal government. 455 U.S. 720, 102 S. Ct. 1373, 71 L. Ed. 2d 580 (1982). According to the Court, "[t]he congruence of professional interests between the contractors and the Federal Government is not complete" because "the contractors remained distinct entities pursuing private ends, and their actions remained [*4] commercial activities carried on for profit." Id. at 740; see also Powell v. U.S. Cartridge Co., 339 U.S. 497, 70 S. Ct. 755, 94 L. Ed. 1017 (1950). ---That explodes limits and allows for the thousands of private actors to become individual plan mechanisms – destroys core generics like agent counterplans or politics and explodes the neg research burden – that’s a key internal link to clash which allows for research skills and decisionmaking. ---Independently, The word “substantially” means that the government must play the main role. CFR No Date (Code of Federal Regulations, Subpart 3.1—Safeguards, http://www.acquisition.gov/far/html/Subpart%203_1.html) (3) “Participating substantially” means that the official’s involvement is of significance to the matter. Substantial participation requires more than official responsibility, knowledge, perfunctory involvement, or involvement on an administrative or peripheral issue. Participation may be substantial even though it is not determinative of the outcome of a particular matter. A finding of substantiality should be based not only on the effort devoted to a matter, but on the importance of the effort. While a series of peripheral involvements may be insubstantial, the single act of approving or participating in a critical step may be substantial. However, the review of procurement documents solely to determine compliance with regulatory, administrative, or budgetary procedures, does not constitute substantial participation in a procurement. A2: Perm-Do Both Links to our spending based net benefits or it severs out of it’s 2NC Solvency The CP solves the case A. Economic incentives-User fees encourage the private sector to take on a larger role for the financing of IOOS-fills in for current funding gaps-That’s Woll from the 1NC B. Re-allocation-Commercialization allows the government to spend scare resources on other IOOS related projects-solves better than the aff Woll-OCEANS Conference-9 MTS/IEEE Biloxi - Marine Technology for Our Future: Global and Local Challenges The role of commercialization in IOOS and the regional associations http://ieeexplore.ieee.org.ezproxy.uky.edu/xpls/icp.jsp?arnumber=5422377#authors A major side benefit to this construct is that it provides a second significant source of funding for the enterprise . Obviously, end users are not going to pay fees for a prospective product or service, so there will remain a need for government funding to start projects up. However, the fees should be set at a level that will provide a large portion ( ideally all) of the O&M funding required to sustain projects over time. This second source of funding has the net effect of freeing up previous government O&M funds to be re-applied to start other new projects - such that the government funding truly becomes a type of seed funding, helping worthy projects get off the ground and turning them over to operational organizations. Fig. 3 depicts how the funding profile could change over time under this proposal. C. This is the most effective funding strategy Woll-OCEANS Conference-9 MTS/IEEE Biloxi - Marine Technology for Our Future: Global and Local Challenges The role of commercialization in IOOS and the regional associations http://ieeexplore.ieee.org.ezproxy.uky.edu/xpls/icp.jsp?arnumber=5422377#authors SECTION IVCONCLUSION The IOOS program offers an unparalleled opportunity for all members of the oceanography enterprise to take part in a collaborative effort - one that provides the best possible stewardship of taxpayer dollars while helping advance practical scientific knowledge about the world's oceans. With the breadth and scope of challenges facing the ocean environment, such an approach is not only prudent - it offers the only reasonable chance of success in the future fiscal environment. Quite simply, the federal government will not be able to fund every worthy project, so it is incumbent on IOOS and the oceanography enterprise to make the most of every dollar. That means cooperating to make choices from an enterprise perspective, making the hard choices to eliminate once promising projects that have not panned out as expected, and being willing to challenge and change habits that have been successful in years past. D. Their authors support the CP U.S. IOOS SUMMIT REPORT 13 A New Decade for the Integrated Ocean Observing System http://www.ioos.noaa.gov/about/governance/summit2012/usioos_summit_report.pdf Step 8. Increase Advocacy. There is a need to develop and maintain advocacy for U.S. IOOS. Challenge. U.S. IOOS is a line item in the NOAA budget, but it remains a largely unfunded federal mandate. The existing approach of individual agency ocean observing programs addressing agency- unique missions with uncoordinated agency budgets is inadequate. This fragmented Federal approach .ies at the hea't of the U.S. IOOS challenge to thrive. Advocacy will develop naturally if stakeholders and users are actively engaged and their requirements are being met. But a proactive advocacy strategy is also needed. There are many users that have come to rely on U.S. IOOS products without providing any support for the system. The ocean observing community is not adept at turning that supportive relationship into advocacy for continued or expanded funding. Significant, well-qualified human resources are necessary to maintain effective user engagement. U.S. IOOS must recognize the importance of this process and support implementation of a user engagement infrastructure in order to meet the vision of U.S. IOOS for the next decade . A2: Obstacles to Commericalization The CPs restructuring solves Woll-OCEANS Conference-9 MTS/IEEE Biloxi - Marine Technology for Our Future: Global and Local Challenges The role of commercialization in IOOS and the regional associations http://ieeexplore.ieee.org.ezproxy.uky.edu/xpls/icp.jsp?arnumber=5422377#authors SECTION IIIADDITIONAL CHANGES TO OPTIMIZE ENTERPRISE EFFORT Commercializing the products and services provided to end users by IOOS and its member organizations would provide a clear definition of successful projects - those that can garner sustainment O&M funding from user fees. However, for the enterprise to be most effective and most efficient, one further step must be considered. Specifically, a realignment of organizational responsibilities to take maximum advantage of the strengths of the differing types of member organizations would yield additional benefits. There are four broad categories of IOOS participants: government, academic (university/research institution), business, and RA/IOOS staff. A summary of each of their strengths, weaknesses, and proposed roles is provided in Table 1. These proposed roles are reflected graphically in Fig. 4, which depicts a more structured and hierarchical organization than is currently in place. In this type of structure, the government actors (federal and state) would provide policy, coordination, and funding. The academic sector would focus on science, research, and providing high quality review and technical advice to regulators and other government actors. Businesses would focus on execution and serve as the bridge to the end user. The IOOS and RA staffs would coordinate between all of the levels, provide project oversight, and administer a data repository. This is admittedly a significant realignment - one which would ask some member organizations to forgo activities at which they have been very successful. However, when approached from an enterprise perspective, the benefits become clear. Table 2 summarizes some of the major cost benefit comparisons between the existing organizational structure and that proposed in this paper. Drones Drones/Robots Solvency T-rex AUV’s are effective, high tech, and solve data collection in all areas of the globe McGann et al 8 [Conor McGann, Frederic Py, Kanna Rajan, Hans Thomas, Richard Henthorn, Rob McEwen Monterey Bay Aquarium Research Institute. 2008, “A Deliberative Architecture for AUV Control,” http://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber=04543343 PDF //jweideman] Oceanography has traditionally relied on ship-based observations. These have recently been augmented by robotic platforms such as Autonomous Underwater Vehicles (AUV) [1-3], which are untethered powered mobile robots able to carry a range of payloads efficiently over large distances in the deep ocean. A common design relies on a modular tube-like structure with propulsion at the stern and various sensors, computers and batteries taking up the bulk of the tube (Fig. 1). AUVs have demonstrated their utility in oceanographic research in gathering time series data by repeated water-column surveys [4], detailed bathymetric maps of the ocean floor in areas of tectonic activity [5,6] and performed hazardous under-ice missions [7].Typically AUVs do not communicate with the support ship or shore while submerged and rely on limited stored battery packs while operating continuously for tens of hours. Current AUV control systems [8] are a variant of the behavior-based Subsumption architecture [9]. A behavior is a modular encapsulation of a specific control task and includes acquisition of a GPS fix, descent to a target depth, drive to a given waypoint, enforcement of a mission depth envelope etc. An operator defines each plan as a collection of behaviors with specific start and end times as well as maximum durations, which are scripted a priori using simple mission planning tools. In practice, missions predominantly consist of sequential behaviors with duration and task specific parameters equivalent to a linear plan with limited flexibility in task duration. Such an approach becomes less effective as mission uncertainty increases. Further, the architecture offers no support to manage the potentially complex interactions that may result amongst behaviors, pushing a greater cognitive burden on behavior developers and mission planners. This paper describes an automated onboard planning system to generate robust mission plans using system state and desired goals. By capturing explicit interactions between behaviors as plan constraints in the domain model and the use of goal oriented commanding, we expect this approach to reduce the cognitive burden on AUV operators. Our interest in the near term is to incorporate decision-making capability to deal with a range of dynamic and episodic ocean phenomenon that cannot be observed with scripted plans. The remainder of this paper is laid out as follows. Section II lays out the architecture of our autonomy system, section III details the experimental results to date, related work follows in section IV with concluding remarks in section V.II. THE T-REX ARCHITECTURE T-REX (TeleoReactive EXecutive) is a goal-oriented system, with embedded automated planning [14,15] and adaptive execution. It encapsulates the long-standing notion of a sense-deliberate-act cycle in what is typically considered a hybrid architecture where sensing, planning and execution are interleaved. In order to make embedded planning scalable the system enables the scope of deliberation to be partitioned functionally and temporally and to ensure the current state of the agent is kept consistent and complete during execution. While T-REX was built for a specific underwater robotics application, the principles behind its design are applicable in any domain where deliberation and execution are intertwined.Fig. 2 shows a conceptual view of a Teleo-Reactive Agent. An agent is viewed as the coordinator of a set of concurrent control loops. Each control loop is embodied in a Teleo-Reactor (or reactor for short) that encapsulates alldetails of how to accomplish its control objectives. Arrows represent a messaging protocol for exchanging facts and goals between reactors: thin arrows represent observations of current state; thick arrows represent goals to be accomplished. Reactors are differentiated in 3 ways: • Functional scope: indicating the state variables of concern for deliberation and action. • Temporal scope: indicating the look-ahead window over which to deliberate. • Timing requirements: the latency within which this component must deliberate for goals in its planning horizon. Fig. 2 for example, shows four different reactors; the Mission Manager provides high-level directives to satisfy the scientific and operational goals of the mission: its temporal scope is the whole mission, taking minutes to deliberate if necessary. The Navigator and Science Operator manage the execution of sub-goals generated by the Mission Manager. The temporal scope for both is in the order of a minute even as they differ in their functional scope. Each refines high-level directives into executable commands depending on current system state. The Science Operator is able to provide local directives to the Navigator. For example if it detects an ocean front it can request the navigation mode to switch from a Yo-Yo pattern in the vertical plane to a Zig-Zag pattern in the horizontal plane, to have better coverage of the area. Deliberation may safely occur at a latency of 1 second for these reactors. The Executive provides an interface to a modified version of the existing AUV functional layer. It encapsulates access to commands and vehicle state variables. The Executive is reasonably approximated as having zero latency within the timing model of our application since it will accomplish a goal received with no measurable delay, or not at all; in other words it does not deliberate. AUV tech provides effective mapping capabilities Yoerger et al. 7 [Dana R. Yoerger1, Michael Jakuba1, Albert M. Bradley1, and Brian Bingham2 1 Woods Hole Oceanographic Institution 2 Franklin W. Olin College of Engineering. 1/1/2007, “Techniques for Deep Sea Near Bottom Survey Using an Autonomous Underwater Vehicle,” http://business.highbeam.com/437280/article-1G1-156721474/techniques-deep-sea-near-bottom-survey-using-autonomous //jweideman] This paper reports navigation algorithms that enable an underwater vehicle to accomplish fully autonomous scientific surveys in the deep sea. These algorithms allow the vehicle to determine its position, to bottom-follow (maintain a constant height above sea oor terrain) and avoid obstacles, and to autonomously focus on the highest value parts of a survey. Scientific exploration of the deep sea has traditionally been performed using inhabited submersibles, towed vehicles, and tethered remotely operated vehicles (ROVs). Autonomous underwater vehicles (AUVs) have begun to replace these vehicles for mapping and survey missions. Autonomous vehicles complement the capabilities of these existing systems, offering superior mapping capabilities, improved logistics, and improved utilization of the surface support vessel. AUVs are particularly well suited to systematic preplanned surveys using sonars, in situ chemical sensors, and cameras in the rugged deep sea terrain that is the focus of many scienti c expeditions. Inhabited submersibles and ROVs remain the only option for manipulation tasks such as sampling, deploying and recovering experiments on the sea oor, detailed inspection, and servicing subsea instruments; however, high resolution maps from AUVs can facilitate these tasks.Figure 1 shows the Autonomous Benthic Explorer (ABE), a 6000 m autonomous underwater vehicle that our team has been developing and deploying for ne-scale quantitative survey and mapping of the sea floor. ABE can survey at constant depth or bottom-follow even in rugged terrain, and it can autonomously determine its position and drive track lines with a precision on the order of several meters. ABE carries a variety of sensors, including scanning and multibeam sonars; a magnetometer; a digital still camera; two sets of pumped conductivity and temperature probes; an acoustic Doppler current pro ler (ADCP); several chemical sensors for hydrothermal plume mapping; and occasional mission-specific instrumentation. ABE's shape and thruster placement allow it to maintain control over a wide range of speed, and to stop or back up if necessary to avoid obstacles. ABE descends to the sea oor with the aid of a descent weight. ABE glides in a controlled spiral trajectory to ensure that it reaches the desired starting point without consuming signi cant battery energy. After reaching the sea oor and performing a series of checks, ABE releases its descent weight to become neutrally buoyant and begins its survey. Throughout the dive, including descent, ABE uses acoustic long-baseline (LBL) transponder navigation and, when in range of the bottom (< 300 m), bottom-lock acoustic Doppler measurements to determine its position and velocity.A dive can consist of a mix of hydrothermal plume survey at constant depth, sonar and magnetics survey following the sea oor (at heights of 50{ 200 m), and digital photography (height of 5 m). ABE usually surveys until its batteries are depleted (between 15 and 30 hours depending on sensor payload and terrain). At the end of its dive, ABE releases its ascent weight to become positively buoyant and returns to the surface. The remainder of this report is organized as follows: Sect. 2 summarizes scienti c survey tasks that have motivated our AUV work, Sect. 3 reports an algorithm for acoustic positioning, Sect. 4 reports methods for terrainfollowing and obstacle avoidance, Sect. 5 reports a technique for automated nested survey, and Sect. 6 presents a brief summary and conclusion.2 Precisely Navigated, Coregistered AUV Surveys Proximity to the sea oor, precise navigation, robust control, and coregistered sensors permit an AUV to characterize the sea floor and the near-bottom environment with complementary sensing modalities on the meter-scale. This section summarizes scienti c work in which ABE-derived bathymetric maps, magnetics maps, digital photos, and hydrographic maps have played critical enabling roles. Meter-scale bathymetric and magnetic maps made using ABE have provided geologists and geophysicists with new perspectives on important sea oor processes. Combined magnetics and bathymetric maps show crustal magnetization, which permits the age and thickness of lava flows to be determined. Combined maps have also been used to identify volcanic features such as lava flow units [1], delimit their fronts, and estimate their thicknesses [2, 3]. Meter-scale bathymetric maps show tectonic features such as faults with great clarity, even enabling them to be resolved into multiple components [4]. In other cases, these maps have revealed the relationship between tectonic features and morphology, such as volcanic domes [3], and hydrothermal vents [1]. ABE bathymetric maps have proved to be of su cient detail and precision for one collaborator to reconstruct the tectonic history of a rift valley by computationally removing faults [5]. The result revealed a dome-like structure from which the valley evolved. On a recent cruise to the Atlantis Massif, detailed renderings of faults and the hydrothermal structures provided critical clues as to the mechanisms controlling the hydro-geology at the newly discovered Lost City hydrothermal vent site [6]. Digital photographs of the sea oor from ABE have provided details of lava ow types and e usion rates [3], sediment cover, and the distribution of benthic organisms. Water column data from ABE yields indications of hydrothermal plume activity and has been used to estimate heat ux from known hydrothermal vent sites, and to locate undiscovered sites on the sea oor. To estimate the heat ux from vent elds on the Juan de Fuca Ridge in the Northeast Paci c (47 540 N, 129 100 W) [7], ABE measured temperature, salinity, and threeaxis water velocity while repeatedly executing a tight grid pattern above the eld [8]. Recently ABE located and preliminarily characterized several previously unmapped hydrothermal sites on the Eastern Lau Spreading Center (ELSC) south of Tonga (21 080 S, 175 120 W) [9]; and on the Southern Mid Atlantic Ridge (SMAR) north of Ascension Island (7 570 S, 14 220 W) [10]. In each case, we started with clues provided by towed systems that indicated a vent site within several kilometers. ABE then executed a three-dive sequence [9, 10] of grid patterns at increasing ner scales and increasingly close to the sea oor. To plan each dive, the scienti c party carefully scrutinized the data from the previous dive along with any available ancillary data.These vent prospecting missions capitalized on ABE's ability to conduct precisely navigated surveys at scales O (m{km), to operate over rugged terrain, and relied on nearly all of ABE's sensing modalities. Figure 2 shows tracklines from sequence of dives designed to locate and survey a vent site on ELSC along with a sampling of the variety of data products acquired and used to plan each stage of the dive sequence. ABE mapped plume activity (temperature, optical backscatter, and reduction-oxidization potential (eH) [11]) to pinpoint the locations of plumes emanating from the eld, built ne-scale bathymetric maps of the vent elds and surrounding environment, and nally photographed the vent structures and animal populations. The remainder of this paper presents the underlying algorithms that enabled ABE to perform this work.This paper reports navigation algorithms that enable an underwater vehicle to accomplish fully autonomous scienti c surveys in the deep sea. These algorithms allow the vehicle to determine its position, to bottom-follow (maintain a constant height above sea oor terrain) and avoid obstacles, and to autonomously focus on the highest value parts of a survey. Scienti c exploration of the deep sea has traditionally been performed using inhabited submersibles, towed vehicles, and tethered remotely operated vehicles (ROVs). Autonomous underwater vehicles (AUVs) have begun to replace these vehicles for mapping and survey missions. Autonomous vehicles complement the capabilities of these existing systems, o ering superior mapping capabilities, improved logistics, and improved utilization of the surface support vessel. AUVs are particularly well suited to systematic preplanned surveys using sonars, in situ chemical sensors, and cameras in the rugged deep sea terrain that is the focus of many scienti c expeditions. Inhabited submersibles and ROVs remain the only option for manipulation tasks such as sampling, deploying and recovering experiments on the sea oor, detailed inspection, and servicing subsea instruments; however, high resolution maps from AUVs can facilitate these tasks.Figure 1 shows the Autonomous Benthic Explorer (ABE), a 6000 m autonomous underwater vehicle that our team has been developing and deploying for ne-scale quantitative survey and mapping of the seafloor. ABE can survey at constant depth or bottom-follow even in rugged terrain, and it can autonomously determine its position and drive tracklines with a precision on the order of several meters. ABE carries a variety of sensors, including scanning and multibeam sonars; a magnetometer; a digital still camera; two sets of pumped conductivity and temperature probes; an acoustic Doppler current pro ler (ADCP); several chemical sensors for hydrothermal plume mapping; and occasional mission-speci c instrumentation. ABE's shape and thruster placement allow it to maintain control over a wide range of speed, and to stop or back up if necessary to avoid obstacles. ABE descends to the sea oor with the aid of a descent weight. ABE glides in a controlled spiral trajectory to ensure that it reaches the desired starting point without consuming signi cant battery energy. After reaching the sea oor and performing a series of checks, ABE releases its descent weight to become neutrally buoyant and begins its survey. Throughout the dive, including descent, ABE uses acoustic long-baseline (LBL) transponder navigation and, when in range of the bottom (< 300 m), bottom-lock acoustic Doppler measurements to determine its position and velocity.A dive can consist of a mix of hydrothermal plume survey at constant depth, sonar and magnetics survey following the sea oor (at heights of 50{ 200 m), and digital photography (height of 5 m). ABE usually surveys until its batteries are depleted (between 15 and 30 hours depending on sensor payload and terrain). At the end of its dive, ABE releases its ascent weight to become positively buoyant and returns to the surface. The remainder of this report is organized as follows: Sect. 2 summarizes scienti c survey tasks that have motivated our AUV work, Sect. 3 reports an algorithm for acoustic positioning, Sect. 4 reports methods for terrainfollowing and obstacle avoidance, Sect. 5 reports a technique for automated nested survey, and Sect. 6 presents a brief summary and conclusion.2 Precisely Navigated, Coregistered AUV Surveys Proximity to the sea oor, precise navigation, robust control, and coregistered sensors permit an AUV to characterize the sea oor and the near-bottom environment with complementary sensing modalities on the meter-scale. This section summarizes scienti c work in which ABEderived bathymetric maps, magnetics maps, digital photos, and hydrographic maps have played critical enabling roles. Meter-scale bathymetric and magnetic maps made using ABE have provided geologists and geophysicists with new perspectives on important sea oor processes. Combined magnetics and bathymetric maps show crustal magnetization, which permits the age and thickness of lava ows to be determined. Combined maps have also been used to identify volcanic features such as lava ow units [1], delimit their fronts, and estimate their thicknesses [2, 3]. Meter-scale bathymetric maps show tectonic features such as faults with great clarity, even enabling them to be resolved into multiple components [4]. In other cases, these maps have revealed the relationship between tectonic features and morphology, such as volcanic domes [3], and hydrothermal vents [1]. ABE bathymetric maps have proved to be of sufficient detail and precision for one collaborator to reconstruct the tectonic history of a rift valley by computationally removing faults [5]. The result revealed a dome-like structure from which the valley evolved. On a recent cruise to the Atlantis Massif, detailed renderings of faults and the hydrothermal structures provided critical clues as to the mechanisms controlling the hydro-geology at the newly discovered Lost City hydrothermal vent site [6]. Digital photographs of the sea oor from ABE have provided details of lava ow types and e usion rates [3], sediment cover, and the distribution of benthic organisms. Water column data from ABE yields indications of hydrothermal plume activity and has been used to estimate heat ux from known hydrothermal vent sites, and to locate undiscovered sites on the sea oor. To estimate the heat ux from vent elds on the Juan de Fuca Ridge in the Northeast Paci c (47 540 N, 129 100 W) [7], ABE measured temperature, salinity, and threeaxis water velocity while repeatedly executing a tight grid pattern above the eld [8]. Recently ABE located and preliminarily characterized several previously unmapped hydrothermal sites on the Eastern Lau Spreading Center (ELSC) south of Tonga (21 080 S, 175 120 W) [9]; and on the Southern Mid Atlantic Ridge (SMAR) north of Ascension Island (7 570 S, 14 220 W) [10]. In each case, we started with clues provided by towed systems that indicated a vent site within several kilometers. ABE then executed a three-dive sequence [9, 10] of grid patterns at increasing ner scales and increasingly close to the sea oor. To plan each dive, the scienti c party carefully scrutinized the data from the previous dive along with any available ancillary data.These vent prospecting missions capitalized on ABE's ability to conduct precisely navigated surveys at scales O (m{km), to operate over rugged terrain, and relied on nearly all of ABE's sensing modalities. Figure 2 shows tracklines from sequence of dives designed to locate and survey a vent site on ELSC along with a sampling of the variety of data products acquired and used to plan each stage of the dive sequence. ABE mapped plume activity (temperature, optical backscatter, and reduction-oxidization potential (eH) [11]) to pinpoint the locations of plumes emanating from the eld, built ne-scale bathymetric maps of the vent elds and surrounding environment, and nally photographed the vent structures and animal populations. The remainder of this paper presents the underlying algorithms that enabled ABE to perform this work. Inherency SQ IOOS/Data solves IOOS already did enough mapping IOOS report to congress 13 [Official US IOOS report sent to congress. 2013, “U.S. Integrated Ocean Observing System (U.S. IOOS) 2013 Report to Congress,” http://www.ioos.noaa.gov/about/governance/ioos_report_congress2013.pdf //jweideman] Gliders are used to monitor water currents, temperature, and biological information such as dissolved oxygen and nitrate. This information offers a more complete picture of what is happening at and below the ocean surface, and may allow scientists to detect trends that otherwise might have gone undetected. Gliders are assuming a prominent and growing role in ocean science due to their unique capabilities for collecting data safely and at relatively low cost in remote locations, both in deep water and at the surface. An advantage of gliders is that they can be quickly deployed to areas of greatest need. A National Glider Asset Map was deployed by the U.S. IOOS program in 2012 and will include all historic and current glider flights once it is completed. The map shown on the right includes data from glider missions since 2005 from Southern California (SCCOOS), the Pacific Northwest (NANOOS), Central and Northern California (CeNCOOS) and the Mid- Atlantic (MARACOOS) regional glider operations. The glider asset map can be viewed at: http://www.ioos.noaa.gov/observ jng/observing assets/glider asset _map.html. Squo solves satellite and other environmental data Rogers and Gulledge 10 [Will Rogers is a Research Assistant at the Center for a New American Security. Dr. Jay Gulledge is a Non-Resident Senior Fellow at the Center for a New American Security and is the Senior Scientist and Science and Impacts Program Director at the Pew Center on Global Climate Change. April 2010. “Lost in Translation: Closing the Gap Between Climate Science and National Security Policy” http://www.cnas.org/files/documents/publications/Lost%20in%20Translation_Code406_Web_0.pdf //jweideman] The intelligence community continues to declassify one-meter resolution images taken from its satellite systems, giving climate scientists access to images 15 to 30 times sharper than the next- best systems controlled by NASA and commercial entities such as Google.25 These and a plethora of other advancements have produced a greater understanding of the Earth's climate system as well as the affects of human activities (or anthro- pogenic influences) on the climate system. The amount of observational data and output from climate models is growing quickly. For example, there are terabytes of highly credible climate change projections now available from the IPCC's Fourth Assessment Report that have never been examined in detail - particularly with regard to local and near-term projections, by decade, to the end of the century and beyond. The sheer vol- ume of available but unanalyzed data creates the potential for many policyrelevant questions to be answered today, if only decision makers were aware of the data, knew how to access it and could make sense of it and if more scientists understood the needs of decision makers and were motivated to provide it to them in a more useful form. In the future, as even more data the National Polarorbiting Operational Environmental Satellite System (NPOESS). Which will orbit the Earth every 100 minutes, "provid- ing global coverage, monitoring environmental conditions, collecting, disseminating and process- ing data about the Earth's weather, atmosphere, oceans, land and near-space environment."26 The private sector has started to contribute to the flow of new information as well. For example, there are new publicprivate partnerships to advance climate science data collection and analysis with new satel- lite systems.27 Meanwhile, other private companies arc embarking on similar solo endeavors, in part, in recognition of the likelihood that there will be a surge in become available, new efforts are emerging to handle the onslaught of information. NOAA is leading one public sector effort, the demand for collection and analysis of climate information. Given the proliferation of new tools (e.g., climate satellites and advanced computer models) and data acquisition systems, there will be no short- age of climate information (especially data related to present conditions and short-term trends). The question for the national security community is whether its unique needs will be met. Since the community has not traditionally based decisions on climate change projections or assessments, there are few processes in place to ensure that the necessary information will be available when it is needed and in a form that is useful. Solvency Satellites Can’t solve without new satellites Parthemore and Rogers 11 [Christine Parethemore, Senior Advisor at United States Department of Defense Adjunct Professor, Security Studies Program at Johns Hopkins University Past Fellow, CNAS at Center for a New American Security Research Assistant at Bob Woodward Education Georgetown University The Ohio State University. Will Rogers, works at cnas. July 2011, “Blinded: The Decline of U.S. Earth Monitoring Capabilities and Its Consequences for National Security” http://www.cnas.org/files/documents/publications/CNAS_Blinded_ParthemoreRogers_0.pdf //jweideman] Networks of satellites, ground-based sensors and unmanned aerial vehicles - the assets America uses to monitor and understand environmental change and its consequences - are going dark. By 2016, only seven of NASA's current 13 earth monitoring satellites are expected to be operational, leaving a crucial information gap that will hinder national security planning.1 Meanwhile, efforts to prevent this capability gap have been plagued by budget cuts, launch failures, technical deficiencies, chronic delays and poor interagency coordination. Without the information that these assets provide, core U.S. foreign policy and national security interests will be at risk. The United States depends on satellite systems for managing the unconventional challenges of the 21st century in ways that are rarely acknowledged. This is particularly true for satellites that monitor climate change and other environmental trends, which, in the words of the Department of Defense's (DOD's) 2010 Quadrennial Defense Review, "will shape the operating environment, roles, and missions" of DOD and "may act as an accelerant of instability or conflict, placing a burden to respond on civilian institutions and militaries around the world.'- Examples abound of how climate change is shap- ing the strategic environment and of why the U.S. government needs continued access to earth moni- toring data: • The opening of the Arctic is requiring the U.S. Navy and Coast Guard to execute new missions in the High North, including more frequent search and rescue missions. • The receding Himalayan glaciers and related reduced river flows to South Asia may shape the future relationship between India and Pakistan. Defense planners and diplomats will need to monitor changes in the glaciers that supply rivers in Pakistan in order to determine whether access to water will exacerbate existing military and diplomatic tensions between India and Pakistan - states that have longstanding grievances over how they share water. • In the South China Sea, changing ocean conditions are altering fish migration, leading neighboring countries to compete over access to billions of dol- lars in fish resources; this situation could escalate into serious conflict in contested territorial waters. • DOD and development agencies rely on earth monitoring systems to monitor urbanization, migration patterns and internal population displacement. Several government agencies also rely on earth monitoring capabilities to analyze compliance with deforestation and emissions measures in international climate change treaties, just as the government relies on space-based capabilities to monitor and verify compliance with non-proliferation treaties. Responding to environmental and climate change trends requires a steady stream of reliable information from earth monitoring satellites that is quickly becoming unavailable. Ideally, the U.S. government would replace its aging earth moni- toring satellites. Yet the current political and fiscal environments constrain available resources, making it less likely that Congress will appropri- ate funds to wholly replace old systems. Given this reality, U.S. policymakers should use existing systems more efficiently, improve information sharing among interagency partners and leverage international partners' investments in their own systems in order to bolster U.S. climate and envi- ronmental data collection capabilities. The Capability Gap Policymakers have known about the challenges stemming from America's declining earth moni- toring capabilities for years. In 2005, a report by the National Research Council warned that America's "system of environmental satellites is at risk of collapse."1 Key U.S. government agencies, including the Office of Science and Technology Policy (OSTP) and the Government Accountability Office (GAO), have recently reiter- ated those warnings. According to an April 2010 report by the GAO, "gaps in coverage ranging from 1 to 11 years are expected beginning as soon as 2015" and "are expected to affect the continuity of important climate and space weather measure- ments, such as our understanding of how weather cycles impact global food production."4 These gaps will include key environmental and climate monitoring functions, from radar altimeters that measure changes in land and ocean surfaces (such as sea level rise and desertification) to aerosol polarimetry sensors that can measure and distin- guish between sulfates, organic and black carbon and other atmospheric particles. "Meteorologists, oceanographers, and climatologists reported that these gaps will seriously impact ongoing and planned earth monitoring activities," according to the GAO.s One recent interagency effort to close such gaps has fallen short. The National Polar-orbiting Operational Environmental Satellite System (NPOESS) was designed to translate climate and environmental data (including data from exten- sive existing databases) into products and analysis for DOD, NASA and the National Oceanic and Atmospheric Administration (NOAA). However, after lone delays, cost overruns and inadequatecoordination among the partners in the interagency working group, the project was split into two components (as an alternative to being cancelled completely); DOD and the civilian agencies are moving forward separately with their own projects in order to sustain the capabilities that NPOESS was intended to provide. Tech can’t solve-political will Technology can’t resolve governance problems. Malone et al. 2014 Thomas C., University of Maryland Professor of Marine Biology, A global ocean observing system framework for sustainable development, Marine Policy 43 (2014) 262–272 As emphasized by Browman et al. [291, governance, not the science and technology behind ocean observing systems, is the weak link in the implementation chain for EBAs. Historically, governments have responded to changes in ecosystem states and their impacts in an ad hoc fashion focusing on sector by sector management (environmental protection and the management of fisheries, water, transportation, oil and gas extraction, wind energy, tourism and recreation, etc) rather than on an integrated, holistic strategy for managing human uses of ecosystem services [4262734]- Limitations of this sector-based approach to govern- ance are compounded by inherent complexities in managing a growing number of activities across different levels of government, across agencies (ministries) within governments, and (for multi- national regional entities such as the European Union) among governments. As a result, competition for funding among govern- ment agencies often inhibits needed collaborations and can result in policy choices that are detrimental to ecosystems and their services and dependent coastal communities. The limitations of sector-specific 'stove pipe" approaches to the stewardship of ecosystem services have been recognized for decades and are reflected in national and international calls for EBAs to managing human uses and adapting to climate change "3.4.28.153). An effective set of institutional responses to achieve sustainable development must address the reality that current governance arrangements are not well tuned to the challenges of formulating, implementing and improving EBAs that are guided by lEAs and the sustained provision of the required data on pressures, states and impacts. Conventional approaches to sustainable devel- opment incorporate current socio-political systems that limit the use of lEAs. Needed is the development of institutional frame- works that promote a shift from sector by sector management of human uses to a more holistic approach that considers the diversity of pressures on ecosystems, changes in ecosystem states and their impacts through integrated governance. Cross-cutting stove pipes must involve a system of integrated ocean governance that oversees and enables efficient linkages among three interdependent and sustained activities: (1) establish- ment of an integrated coastal ocean observing system to monitor pressures, states and impacts simultaneously; (2) routine and repeated lEAs based on these observations: and (3) the use of lEAs to help guide the sustainable implementation of EBAs (Fig. 31 The flow of data and information among these activities must enable an iterative process of evaluating performance against objectives that leads to more effective observing systems and EBAs. The primary impediments to institutionalizing these activities interactively are stable funding, capacity building, international collaboration, and timely data exchange among countries . Stable funding for sustained observations is critical to the establishment of EBAs for sustainable development Sustained capacity building is critical because most coastal waters are under the jurisdiction of developing countries and emerging economies. International col- laboration is essential to establish priorities based on issues of mutual concern, to attract funding, to effect timely data exchange among coastal states for comparative ecosystem analyses, and to develop systems that are interoperable globally (common stan- dards and protocols for comparative analyses of multidisciplinary data streams from sentinel sites). Of these, timely exchange of ecological data and information on pressures, states and impacts Is likely to be most challenging and will require international agreements that are codified in the UN Convention on the Law of the Sea. Holistic, integrated responses have the potential to effectively address issues related to ecosystem services and human well-being simultaneously. If there is any chance of developing and maintain- ing effective EBAs for sustainable development (and the sustained and integrated observing systems needed to inform them), the process of integrated ocean governance must be overseen by a diversity of stakeholders. These include integrated coastal zone managers, marine spatial planners and managers of CEOBON, LME programs. Marine Protected Areas 11511, Regional Seas Conventions |149|. and international maritime operations |154|. In addition to these practitioners, important stakeholders include data providers (scientists and technicians involved in ocean observing and pre- diction systems), analysts (natural, social and political scientists and economists), and journalists all working for the common good. This body of stakeholders might be called the "Integrated Govern- ance Forum for Sustaining Marine Ecosystem Services" and would function as a "Community of Practice* |155|. However, sustained engagement of such a diversity of stakeholders on a global scale to guide the establishment of EBAs on the spatial scales needed to sustain ecosystem services (Section 2) is not feasible. Developing COOS to provide the data and information needed to inform EBAs (Section 3) depends on enhancing the CCN locally and regionally based on priorities established by stakeholders and articulated in national and regional ocean policies |2S|. The E.U. Marine Strategy Framework Directive [156| and the U.S. National Policy for the Stewardship of the Ocean. Our Coasts, and the Great Lakes 1157] are important examples of emerging regional scale approaches to integrated ocean governance. How effective these policies will be remains to be seen. Warming Solvency No Solvency-Satellites Can’t solve without new satellites Parthemore and Rogers 11 (Christine Parthemore, Senior Advisor at United States Department of Defense Adjunct Professor, Security Studies Program at Johns Hopkins University, Will Rogers, a Research Assistant at the Center for a New American Security, “Blinded: The Decline of U.S. Earth Monitoring Capabilities and Its Consequences for National Security,” Center for a New American Security, July 2011, http://www.cnas.org/files/documents/publications/CNAS_Blinded_ParthemoreRogers_0.pdf) Networks of satellites, ground-based sensors and unmanned aerial vehicles - the assets America uses to monitor and understand environmental change and its consequences - are going dark. By 2016, only seven of NASA's current 13 earth monitoring satellites are expected to be operational, leaving a crucial information gap that will hinder national security planning.1 Meanwhile, efforts to prevent this capability gap have been plagued by budget cuts, launch failures, technical deficiencies, chronic delays and poor interagency coordination. Without the information that these assets provide, core U.S. foreign policy and national security interests will be at risk. The United States depends on satellite systems for managing the unconventional challenges of the 21st century in ways that are rarely acknowledged. This is particularly true for satellites that monitor climate change and other environmental trends, which, in the words of the Department of Defenses (DOD's) 2010 Quadrennial Defense Review, "will shape the operating environment, roles, and missions" of DOD and "may act as an accelerant of instability or conflict, placing a burden to respond on civilian institutions and militaries around the world."3 Examples abound of how climate change is shaping the strategic environment and of why the U.S. government needs continued access to earth monitoring data: • The opening of the Arctic is requiring the U.S. Navy and Coast Guard to execute new missions in the High North, including more frequent search and rescue missions. • The receding Himalayan glaciers and related reduced river flows to South Asia may shape the future relationship between India and Pakistan. Defense planners and diplomats will need to monitor changes in the glaciers that supply rivers in Pakistan in order to determine whether access to water will exacerbate existing military and diplomatic tensions between India and Pakistan - states that have longstanding grievances over how they share water. • In the South China Sea, changing ocean conditions are altering Ash migration, leading neighboring countries to compete over access to billions of dollars in fish resources; this situation could escalate into serious conflict in contested territorial waters. • DOD and development agencies rely on earth monitoring systems to monitor urbanization, migration patterns and internal population displacement. Several government agencies also rely on earth monitoring capabilities to analyze compliance with deforestation and emissions measures in international climate change treaties, just as the government relies on space-based capabilities to monitor and verify compliance with non-proliferation treaties. Political Will Trying to solve for warming does nothing. It’s politically controversial and links to politics Speath 12 (Ryu Speath, deputy editor at TheWeek.com, “Why it's probably too late to roll back global warming,” The Week, December 5, 2012, http://theweek.com/article/index/237392/why-its-probably-too-late-to-roll-back-global-warming) Two degrees Celsius. According to scientists, that's the rise in global temperature, measured against pre-industrial times, that could spark some of the most catastrophic effects of global warming. Preventing the two-degree bump has been the goal of every international treaty designed to reduce greenhouse gas emissions, including a new one currently being hammered out at a United Nations summit in Doha, Qatar. But a new study published by the journal Nature Climate Change shows that it's incredibly unlikely that global warming can be limited to two degrees. According to the study, the world in 2011 "pumped nearly 38.2 billion tons of carbon dioxide into the air from the burning of fossil fuels such as coal and oil," says Seth Borenstein at The Associated Press: The total amounts to more than 2.4 million pounds (1.1 million kilograms) of carbon dioxide released into the air every second. Because emissions of the key greenhouse gas have been rising steadily and most carbon stays in the air for a century, it is not just unlikely but "rather optimistic" to think that the world can limit future temperature increases to 2 degrees Celsius (3.6 degrees Fahrenheit), said the study's lead author, Glen Peters at the Center for International Climate and Environmental Research in Oslo, Norway. What happens when the two-degree threshold is crossed? Most notably, that's when the polar ice caps will begin to melt, leading to a dangerous rise in sea levels. Furthermore, the world's hottest regions will be unable to grow food, setting the stage for mass hunger and global food inflation. The rise in temperature would also likely exacerbate or cause extreme weather events, such as hurricanes and droughts. There is a very small chance that the world could pull back from the brink. The U.N. could still limit warming to two degrees if it adopts a "radical plan," says Peters' group. According to a PricewaterhouseCoopers study, such a plan would entail cutting carbon emissions "by 5.1 percent every year from now to 2050, essentially slamming the breaks on growth starting right now," says Coral Davenport at The National Journal, "and keeping the freeze on for 37 years." However, the U.N. has set a deadline of ratifying a new treaty by 2015, and implementing it by 2020, which means the world is already eight years behind that pace. There are still major disagreements between the U.S. and China over whether the developed world, which industrialized first, should bear the bulk of the cost of reducing carbon emissions. And there is, of course, a large contingent of Americans who don't even believe climate change exists, putting any treaty's ratification at risk. Climate change is so politically toxic in America that Congress has prioritized the fiscal cliff over — no exaggeration — untold suffering and the end of the world as we know it. In other words, it isn't happening. And if that's not bad enough, keep in mind that the two-degree mark is just the beginning, says Davenport: Michael Oppenheimer, a professor of geosciences and international affairs at Princeton University and a member of the Nobel Prize-winning U.N. Intergovernmental Panel on Climate Change, says that a 2degree rise is not itself that point, but rather the beginning of irreversible changes. "It starts to speed you toward a tipping point," he said. "It's driving toward a cliff at night with the headlights off. We don't know when we'll hit that cliff, but after 2 degrees, we're going faster, we have less control. After 3, 4, 5 degrees, you spiral out of control, you have even more irreversible change." Indeed, at the current emissions rate, the world is expected to broach the four-degree mark by 2100 — at which point, we can expect even worse environmental catastrophes. Some analysts say that the best possible scenario is preventing the Earth from warming up by three or four degrees. That means instead of focusing solely on preventing global warming, governments around the world should begin preparing for the major environmental upheavals, starting with protections for coastal cities. No Warming No warming Global warming theory is false – 5 warrants Hawkins 14 (John Hawkins, runs Right Wing News and Linkiest. He's also the co-owner of the The Looking Spoon. Additionally, he does weekly appearances on the #1 in it's market Jaz McKay show, writes two weekly columns for Townhall and a column for PJ Media. Additionally, his work has also been published at the Washington Examiner, The Hill, TPNN, Hot Air, The Huffington Post and at Human Events, “5 Scientific Reasons That Global Warming Isn't Happening,” Town Hall, February 18, 2014, http://townhall.com/columnists/johnhawkins/2014/02/18/5scientific-reasons-that-global-warming-isnt-happening-n1796423/page/full) How did global warming discussions end up hinging on what's happening with polar bears, unverifiable predictions of what will happen in a hundred years, and whether people are "climate deniers" or "global warming cultists?" If this is a scientific topic, why aren't we spending more time discussing the science involved? Why aren't we talking about the evidence and the actual data involved? Why aren't we looking at the predictions that were made and seeing if they match up to the results? If this is such an open and shut case, why are so many people who care about science skeptical? Many Americans have long since thought that the best scientific evidence available suggested that man wasn't causing any sort of global warming. However, now, we can go even further and suggest that the planet isn't warming at all. 1) There hasn't been any global warming since 1997: If nothing changes in the next year, we're going to have kids who graduate from high school who will have never seen any "global warming" during their lifetimes. That's right; the temperature of the planet has essentially been flat for 17 years. This isn't a controversial assertion either. Even the former Director of the Climate Research Unit (CRU) of the University of East Anglia, Phil Jones, admits that it's true. Since the planet was cooling from 1940-1975 and the upswing in temperature afterward only lasted 22 years, a 17 year pause is a big deal. It also begs an obvious question: How can we be experiencing global warming if there's no actual "global warming?" 2) There is no scientific consensus that global warming is occurring and caused by man: Questions are not decided by "consensus." In fact, many scientific theories that were once widely believed to be true were made irrelevant by new evidence. Just to name one of many, many examples, in the early seventies, scientists believed global cooling was occurring. However, once the planet started to warm up, they changed their minds. Yet, the primary "scientific" argument for global warming is that there is a "scientific consensus" that it's occurring. Setting aside the fact that's not a scientific argument, even if that ever was true (and it really wasn't), it's certainly not true anymore. Over 31,000 scientists have signed on to a petition saying humans aren't causing global warming. More than 1000 scientists signed on to another report saying there is no global warming at all. There are tens of thousands of well-educated, mainstream scientists who do not agree that global warming is occurring at all and people who share their opinion are taking a position grounded in science. 3) Arctic ice is up 50% since 2012: The loss of Arctic ice has been a big talking point for people who believe global warming is occurring. Some people have even predicted that all of the Arctic ice would melt by now because of global warming. Yet, Arctic ice is up 50% since 2012. How much Arctic ice really matters is an open question since the very limited evidence we have suggests that a few decades ago, there was less ice than there is today, but the same people who thought the drop in ice was noteworthy should at least agree that the increase is important as well. 4) Climate models showing global warming have been wrong over and over: These future projections of what global warming will do to the planet have been based on climate models. Essentially, scientists make assumptions about how much of an impact different factors will have; they guess how much of a change there will be and then they project changes over time. Unfortunately, almost all of these models showing huge temperature gains have turned out to be wrong. Former NASA scientist Dr. Roy Spencer says that climate models used by government agencies to create policies “have failed miserably.” Spencer analyzed 90 climate models against surface temperature and satellite temperature data, and found that more than 95 percent of the models “have over-forecast the warming trend since 1979, whether we use their own surface temperature dataset (HadCRUT4), or our satellite dataset of lower tropospheric temperatures (UAH).” There's an old saying in programming that goes, "Garbage in, garbage out." In other words, if the assumptions and data you put into the models are faulty, then the results will be worthless. If the climate models that show a dire impact because of global warming aren't reliable -- and they're not -- then the long term projections they make are meaningless. 5) Predictions about the impact of global warming have already been proven wrong: The debate over global warming has been going on long enough that we've had time to see whether some of the predictions people made about it have panned out in the real world. For example, Al Gore predicted all the Arctic ice would be gone by 2013. In 2005, the Independent ran an article saying that the Artic had entered a death spiral. Scientists fear that the Arctic has now entered an irreversible phase of warming which will accelerate the loss of the polar sea ice that has helped to keep the climate stable for thousands of years....The greatest fear is that the Arctic has reached a “tipping point” beyond which will raise sea levels dramatically. Of course, the highway is still there. Meanwhile, Arctic ice is up 50% since 2012. James Hansen of NASA fame predicted that the West Side Highway in New York would be under water by now because of global warming. If the climate models and the predictions about global warming aren't even close to being correct, wouldn't it be more scientific to reject hasty action based on faulty data so that we can further study the issue and find out what's really going on? nothing can reverse the continual loss of sea ice and with it the massive land glaciers of Greenland, which No warming – History and scientific study prove. Scientists exaggerate studies to get research grants. Jaworowski 08 (Prof. Zbigniew Jaworowski, Chairman of Scientific Council of the Central Laboratory for Radiological Protection in Warsaw, Chair of UN Scientific Committee on the Effects of Atomic Radiation, “Fear Propaganda” http://www.ourcivilisation.com/aginatur/cycles/chap3.htm) Doomsayers preaching the horrors of warming are not troubled by the fact that in the Middle Ages, when for a few hundred years it was warmer than it is now, neither the Maldive atolls nor the Pacific archipelagos were flooded. Global oceanic levels have been rising for some hundreds or thousands of years (the causes of this phenomenon are not clear). In the last 100 years, this increase amounted to 10 cm to 20 cm, (24) but it does not seem to be accelerated by the 20th Century warming. It turns out that in warmer climates, there is more water that evaporates from the ocean (and subsequently falls as snow on the Greenland and Antarctic ice caps) than there is water that flows to the seas from melting glaciers. (17) Since the 1970s, the glaciers of the Arctic, Greenland, and the Antarctic have ceased to retreat, and have started to grow. On January 18, 2002, the journal Science published the results of satellite-borne radar and ice core studies performed by scientists from CalTech's Jet Propulsion Laboratory and the University of California at Santa Cruz. These results indicate that the Antarctic ice flow has been slowed, and sometimes even stopped, and that this has resulted in the thickening of the continental glacier at a rate of 26.8 billion tons a year . (25) In 1999, a Polish Academy of Sciences paper was prepared as a source material for a report titled "Forecast of the Defense Conditions for the Republic of Poland in 2001-2020." The paper implied that the increase of atmospheric precipitation by 23 percent in Poland, which was presumed to be caused by global warming, would be detrimental. (Imagine stating this in a country where 38 percent of the area suffers from permanent surface water deficit!) The same paper also deemed an extension of the vegetation period by 60 to 120 days as a disaster. Truly, a possibility of doubling the crop rotation, or even prolonging by four months the harvest of radishes, makes for a horrific vision in the minds of the authors of this paper. Newspapers continuously write about the increasing frequency and power of the storms. The facts, however, speak otherwise. I cite here only some few data from Poland, but there are plenty of data from all over the world. In Cracow, in 1896-1995, the number of storms with hail and precipitation exceeding 20 millimeters has decreased continuously, and after 1930, the number of all storms decreased. (26) In 1813 to 1994, the frequency and magnitude of floods of Vistula River in Cracow not only did not increase but, since 1940, have significantly decreased. (27) Also, measurements in the Kolobrzeg Baltic Sea harbor indicate that the number of gales has not increased between 1901 and 1990. (28) Similar observations apply to the 20th Century hurricanes over the Atlantic Ocean (Figure 4,) and worldwide. Emissions are currently being released into space. Proves warming is actually reversing Taylor 11 (James Taylor, managing editor of Environment & Climate News, a national monthly publication devoted to sound science and free-market environmentalism. He is also senior fellow for The Heartland Institute, focusing on energy and environment issues, “New NASA Data Blow Gaping Hole In Global Warming Alarmism,” Forbes, July 27, 2011, http://www.forbes.com/sites/jamestaylor/2011/07/27/new-nasadata-blow-gaping-hold-in-global-warming-alarmism/http://www.forbes.com/sites/jamestaylor/2011/07/27/new-nasa-data-blow-gaping-holdin-global-warming-alarmism/) NASA satellite data from the years 2000 through 2011 show the Earth’s atmosphere is allowing far more heat to be released into space than alarmist computer models have predicted, reports a new study in the peer-reviewed science journal Remote Sensing. The study indicates far less future global warming will occur than United Nations computer models have predicted, and supports prior studies indicating increases in atmospheric carbon dioxide trap far less heat than alarmists have claimed. Study co-author Dr. Roy Spencer, a principal research scientist at the University of Alabama in Huntsville and U.S. Science Team Leader for the Advanced Microwave Scanning Radiometer flying on NASA’s Aqua satellite, reports that real-world data from NASA’s Terra satellite contradict multiple assumptions fed into alarmist computer models. “The satellite observations suggest there is much more energy lost to space during and after warming than the climate models show,” Spencer said in a July 26 University of Alabama press release. “There is a huge discrepancy between the data and the forecasts that is especially big over the oceans .” In addition to finding that far less heat is being trapped than alarmist computer models have predicted, the NASA satellite data show the atmosphere begins shedding heat into space long before United Nations computer models predicted. No Consensus 97 percent consensus claims are false Taylor 13, (James, Writer for forbes. 5/30/13, “Global Warming Alarmists Caught Doctoring '97-Percent Consensus' Claims,” http://www.forbes.com/sites/jamestaylor/2013/05/30/global-warming-alarmists-caught-doctoring-97-percent-consensus-claims/ //jweideman) Global warming alarmists and their allies in the liberal media have been caught doctoring the results of a widely cited paper asserting there is a 97-percent scientific consensus regarding human-caused global warming. After taking a closer look at the paper, investigative journalists report the authors’ claims of a 97-pecent consensus relied on the authors misclassifying the papers of some of the world’s most prominent global warming skeptics. At the same time, the authors deliberately presented a meaningless survey question so they could twist the responses to fit their own preconceived global warming alarmism. Global warming alarmist John Cook, founder of the misleadingly named blog site Skeptical Science, published a paper with several other global warming alarmists claiming they reviewed nearly 12,000 abstracts of studies published in the peer-reviewed climate literature. Cook reported that he and his colleagues found that 97 percent of the papers that expressed a position on human-caused global warming “endorsed the consensus position that humans are causing global warming.” As is the case with other ‘surveys’ alleging an overwhelming scientific consensus on global warming, the question surveyed had absolutely nothing to do with the issues of contention between global warming alarmists and global warming skeptics. The question Cook and his alarmist colleagues surveyed was simply whether humans have caused some global warming. The question is meaningless regarding the global warming debate because most skeptics as well as most alarmists believe humans have caused some global warming. The issue of contention dividing alarmists and skeptics is whether humans are causing global warming of such negative severity as to constitute a crisis demanding concerted action. Warming slow Worst case scenario warming will only be 1.5 degrees Freitas ‘2 (C. R., Associate Prof. in Geography and Enivonmental Science @ U. Aukland, Bulletin of Canadian Petroleum Geology, “Are observed changes in the concentration of carbon dioxide in the atmosphere really dangerous?” 50:2, GeoScienceWorld) In any analysis of CO2 it is important to differentiate between three quantities: 1) CO2 emissions, 2) atmospheric CO2 concentrations, and 3) greenhouse gas radiative forcing due to atmospheric CO2. As for the first, between 1980 and 2000 global CO2 emissions increased from 5.5 Gt C to about 6.5 Gt C, which amounts to an average annual increase of just over 1%. As regards the second, between 1980 and 2000 atmospheric CO2 concentrations increased by about 0.4 per cent per year. Concerning the third, between 1980 and 2000 greenhouse gas forcing increase due to CO2 has been about 0.25 W m–2 per decade (Hansen, 2000). Because of the logarithmic relationship between CO2 concentration and greenhouse gas forcing, even an exponential increase of atmospheric CO2 concentration translates into linear forcing and temperature increase; or, as CO2 gets higher, a constant annual increase of say 1.5 ppm has less and less effect on radiative forcing, as shown in Figure 3 . Leaving aside for the moment the satellite temperature data and using the surface data set, between 1980 and 2000 there has been this linear increase of both CO2 greenhouse gas forcing and temperature. If one extrapolates the rate of observed atmospheric CO2 increase into the future, the observed would only lead to a concentration of about 560 ppm in 2100, about double the concentration of the late 1800’s. That assumes a continuing increase in the CO2 emission rate of about 1% per year, and a carbon cycle leading to atmospheric concentrations observed in the past. If one assumes, in addition, that the increase of surface atmospheric CO2 increase temperatures in the last 20 years (about 0.3 °C) is entirely due to the increase in greenhouse gas forcing of all greenhouse gas, not just CO2, that would translate into a temperature increase of about 1.5 °C (or approximately 0.15 °C per decade). Using the satellite data, the temperature increase is correspondingly lower. Based on this, the temperature increase over the next 100 years might be less than 1.5 °C, as proposed in Figure 19. Too late Warming is too fast for adaption Willis et al 10 (K.J. Willis1,3,4 , R.M. Bailey2 , S.A. Bhagwat1,2 and H.J.B. Birks1,2,3 1 Institute of Biodiversity at the James Martin 21st Century School, University of Oxford, South Parks Road, Oxford OX1 3PS, UK 2School of Geography and the Environment, University of Oxford, Oxford OX1 3QY, UK 3Department of Biology, University of Bergen, Post Box 7803, N-5020 Bergen, Norway 4Department of Zoology, University of Oxford. 2010, “Biodiversity baselines, thresholds and resilience: testing predictions and assumptions using palaeoecological data” http://users.ugent.be/~everleye/Ecosystem%20Dynamics/papers%20assignment/Willis%20et%20al%20TREE%202010.pdf//jweideman) Another critical question relating to future climate change is whether rates of predicted change will be too rapid for ecological processes to react and so prevent species and communities persisting. There is also the question of whether species will be able to migrate quickly enough to new locations with a suitable climate. Studies based on data from extant populations and modelling suggest that rapid rates of change could pose a serious threat for many species and communities unable to track climate space quickly enough, resulting in extensive extinctions [9,31]. But it is also known from fossil records that there have been numerous previous intervals of abrupt climate change [32,33]. What were the responses of past biodiversity to these previous intervals of rapid climate change? The existence of highly detailed evidence from ice-core records (Box 2) spanning the last full glacial cycle provides an ideal opportunity to examine biodiversity responses to rapid climate change. For example, ice-cores indicate that temperatures in mid to high latitudes oscil- lated repeatedly by more than 4°C on timescales of decades or less 134) (Box 2). Numerous records of biodiversity response from North America and Europe across this time-interval reflect ecological changes with a decadal resolution [35-37). While they demonstrate clear evidence for rapid turnover of communities (e.g. Figure 2), novel assemblages, migrations and local extinctions, there is no evidence for the broad-scale extinctions predicted by mod- els; rather there is strong evidence for persistence |25). However, there is also evidence that some species expand- ed their range slowly or largely failed to expand from their refugia in response to this interval of rapid climate warm- ing 138). The challenge now is to determine which specific respect, in particular information on alternative stable states, rates of change, possible triggering mechanisms and systems that demonstrate resilience to thresholds. In a study from central Spain, for example, it has been demonstrated that over the past 9000 years, several threshold changes occurred, shifting from one stable forest type to another (pine to deciduous oak and then evergreen oak to pine) (52). The trigger appears to have been a combination of at least two abiotic variables; in the first shift, an interval of higher precipitation combined with less evaporation and in the second, increased aridity combined with increased fire frequency. A similar 'double-trigger' was also found to be important in regime shifts along the south-east coastline of Madagascar |53| where a threshold shift from closed littoral forest to open Erica- dominated heathland occurred in response to the combined effects of aridity and sea-level rise. Neither of the variables occurring alone resulted in a shift but the combination of the two did. Nothing can be done to stop global warming Skeptical Science 13 (Skeptical Science, “Global warming: Not reversible, but stoppable,” Raw Story, April 27, 2013, http://www.rawstory.com/rs/2013/04/27/global-warming-not-reversable-but-stoppable/) Global warming is not reversible but it is stoppable. Many people incorrectly assume that once we stop making greenhouse gas emissions, the CO2 will be drawn out of the air, the old equilibrium will be re-established and the climate of the planet will go back to the way it used to be; just like the way the acid rain problem was solved once scrubbers were put on smoke stacks, or the way lead pollution disappeared once we changed to unleaded gasoline. This misinterpretation can lead to complacency about the need to act now. In fact, global warming is, on human timescales, here forever. The truth is that the damage we have done—and continue to do—to the climate system cannot be undone. The second question reveals a different kind of misunderstanding: many mistakenly believe that the climate system is going to send more warming our way no matter what we choose to do. Taken to an extreme, that viewpoint can lead to a fatalistic approach, in which efforts to mitigate climate change by cutting emissions are seen as futile: we should instead begin planning for adaptation or, worse, start deliberately intervening through geoengineering. But this is wrong. The inertia is not in the physics of the climate system, but rather in the human economy. This is explained in a recent paper in Science Magazine (2013, paywalled but freely accessible here, scroll down to "Publications, 2013") by Damon Matthews and Susan Solomon: Irreversible Does Not Mean Unavoidable. Since the Industrial Revolution, CO2 from our burning of fossil fuels has been building up in the atmosphere. The concentration of CO2 is now approaching 400 parts per million (ppm), up from 280 ppm prior to 1800. If we were to stop all emissions immediately, the CO2 concentration would also start to decline immediately, with some of the gas continuing to be absorbed into the oceans and smaller amounts being taken up by carbon sinks on land. According to the models of the carbon cycle, the level of CO2 (the red line in Figure 1A) would have dropped to about 340 ppm by 2300, approximately the same level as it was in 1980. In the next 300 years, therefore, nature will have recouped the last 30 years of our emissions. So, does this mean that some of the climate change we have experienced so far would go into reverse, allowing, for example, the Arctic sea ice to freeze over again? Unfortunately, no. Today, because of the greenhouse gas build-up, there is more solar energy being trapped, which is warming the oceans, atmosphere, land and ice, a process that has been referred to as the Earth's energy imbalance. The energy flow will continue to be out of balance until the Earth warms up enough so that the amount of energy leaving the Earth matches the amount coming in. It takes time for the Earth to heat up, particularly the oceans, where approximately 90% of the thermal energy ends up. It just so happens that the delayed heating from this thermal inertia balances almost exactly with the drop in CO2 concentrations, meaning the temperature of the Earth would stay approximately constant from the minute we stopped adding more CO2, as shown in Figure 1C. There is bad news and good news in this. The bad news is that, once we have caused some warming, we can’t go back, at least not without huge and probably unaffordable efforts to put the CO2 back into the ground, or by making risky interventions by scattering tons of sulphate particles into the upper atmosphere, to shade us from the Sun. The good news is that, once we stop emissions, further warming will immediately cease; we are not on an unstoppable path to oblivion. The future is not out of our hands. Global warming is stoppable, even if it is not reversible. Too late to solve warming Adam 08 (David Adam, environment correspondent for the Guardian between 2005 and 2010, before which he was science correspondent for two years. He previously worked at the science journal Nature after a PhD in chemical engineering, “Too late? Why scientists say we should expect the worst,” The Guardian, December 8, 2008, http://www.theguardian.com/environment/2008/dec/09/poznancopenhagen-global-warming-targets-climate-change) Despite the political rhetoric, the scientific warnings, the media headlines and the corporate promises, he would say, carbon emissions were soaring way out of control - far above even the bleak scenarios considered by last year's report from the Intergovernmental Panel on Climate Change (IPCC) and the Stern review. The battle against dangerous climate change had been lost, and the world needed to prepare for things to get very, very bad. "As an academic I wanted to be told that it was a very good piece of work and that the conclusions were sound," Anderson said. "But as a human being I desperately wanted someone to point out a mistake, and to tell me we had got it completely wrong." Nobody did. The cream of the UK climate science community sat in stunned silence as Anderson pointed out that carbon emissions since 2000 have risen much faster than anyone thought possible, driven mainly by the coalfuelled economic boom in the developing world. So much extra pollution is being pumped out, he said, that most of the climate targets debated by politicians and campaigners are fanciful at best, and "dangerously misguided" at worst. In the jargon used to count the steady accumulation of carbon dioxide in the Earth's thin layer of atmosphere, he said it was "improbable" that levels could now be restricted to 650 parts per million (ppm). The CO2 level is currently over 380ppm, up from 280ppm at the time of the industrial revolution, and it rises by more than 2ppm each year. The government's official position is that the world should aim to cap this rise at 450ppm. Even a cut in emissions will accelerate global warming Dye 12 (Lee Dye, science expert for ABC News, “It May Be Too Late to Stop Global Warming,” ABC News, October 26, 2012, http://abcnews.go.com/Technology/late-stop-global-warming/story?id=17557814) Here's a dark secret about the earth's changing climate that many scientists believe, but few seem eager to discuss: It's too late to stop global warming . Greenhouse gasses pumped into the planet's atmosphere will continue to grow even if the industrialized nations cut their emissions down to the bone. Furthermore, the severe measures that would have to be taken to make those reductions stand about the same chance as that proverbial snowball in hell. Two scientists who believe we are on the wrong track argue in the current issue of the journal Nature Climate Change that global warming is inevitable and it's time to switch our focus from trying to stop it to figuring out how we are going to deal with its consequences. "At present, governments' attempts to limit greenhousegas emissions through carbon cap-and-trade schemes and to promote renewable and sustainable energy sources are probably too late to arrest the inevitable trend of global warming," Jasper Knight of Wits University in Johannesburg, South Africa, and Stephan Harrison of the University of Exeter in England argue in their study. Those efforts, they continue, "have little relationship to the real world." What is clear, they contend, is a profound lack of understanding about how we are going to deal with the loss of huge land areas, including some entire island nations, and massive migrations as humans flee areas no longer suitable for sustaining life, the inundation of coastal properties around the world, and so on ... and on ... and on. That doesn't mean nations should stop trying to reduce their carbon emissions, because any reduction could lessen the consequences. But the cold fact is no matter what Europe and the United States and other "developed" nations do, it's not going to curb global climate change, according to one scientist who was once highly skeptical of the entire issue of global warming. "Call me a converted skeptic," physicist Richard A. Muller says in an op-ed piece published in the New York Times last July. Muller's latest book, "Energy for Future Presidents," attempts to poke holes in nearly everything we've been told about energy and climate change, except the fact that "humans are almost entirely the cause" of global warming. Those of us who live in the "developed" world initiated it. Those who live in the "developing" world will sustain it as they strive for a standard of living equal to ours. "As far as global warming is concerned, the developed world is becoming irrelevant," Muller insists in his book. We could set an example by curbing our emissions, and thus claim in the future that "it wasn't our fault," but about the only thing that could stop it would be a complete economic collapse in China and the rest of the world's developing countries. We are already passed the tipping point Thomas 14 (Andrew Thomas, writer for the American Thinker, “The Global Warming Tipping Point is Near,” American Thinker, January 2, 2014, http://www.americanthinker.com/blog/2014/01/the_global_warming_tipping_point_is_near.html) Anthropogenic Global Warming (AGW) theory has been dominant for the past three decades as absolute fact in the public mind. In the last several years, however, cracks in the fortress of "settled science" have appeared, and public opinion has begun to shift. Increasingly, alarmist predictions have failed to come to fruition. In 2004, NASA's chief scientist James Hansen authoritatively announced that there is only a ten-year window to act on AGW (presumably by transferring mass quantities of taxpayer funds to global warmist causes) before climate Armageddon destroys humanity. Well, that window has now shut tight, and AGW is AWOL. Al Gore, the high priest of AGW theory, has closed all of his Alliance for Climate Protection field offices, and laid off 90% of his staff. Contributions have all but dried up since 2008. Australia's conservative government has severely curtailed the country's climate change initiatives and is in the process of repealing its business-killing carbon tax. A group of German scientists predicts dramatic global cooling over the next 90 years toward a new "little ice age." Of course, even many "low information" folks have an awareness of the record increase in Arctic sea ice, as well as the current highly-publicized predicament of the cadre of wealthy global warmists stuck in record-high sea ice while on a cruise to the Antarctic to prove the absence of sea ice. Now the UN's Intergovernmental Panel on Climate Change (IPCC) has quietly downgraded their prediction for global warming for the next 30 years in the final draft of their landmark "Fifth Assessment Report." The effect of this is that they are tacitly admitting that the computer models they have religiously relied-upon for decades as "proof" of AGW theory are dead wrong. The tipping point is near. I can smell it. Warming will never be reversible – vote NEG on presumption that the AFF creates worse that the status quo Romm 13 (Joe Romm, founding editor for Climate Change, “”The Dangerous Myth That Climate Change Is Reversible,” Climate Progress, March 17, 2013, http://thinkprogress.org/climate/2013/03/17/1731591/the-dangerous-myth-that-climate-change-is-reversible/) Memo to Nocera: As a NOAA-led paper explained 4 years ago, climate change is “largely irreversible for 1000 years.” This notion that we can reverse climate change by cutting emissions is one of the most commonly held myths — and one of the most dangerous, as explained in this 2007 MIT study, “Understanding Public Complacency About Climate Change: Adults’ mental models of climate change violate conservation of matter.” The fact is that, as RealClimate has explained, we would need “an immediate cut of around 60 to 70% globally and continued further cuts over time” merely to stabilize atmospheric concentrations of CO2 – and that would still leave us with a radiative imbalance that would lead to “an additional 0.3 to 0.8ºC warming over the 21st Century.” And that assumes no major carbon cycle feedbacks kick in, which seems highly unlikely. We’d have to drop total global emissions to zero now and for the rest of the century just to lower concentrations enough to stop temperatures from rising. Again, even in this implausible scenario, we still aren’t talking about reversing climate change, just stopping it — or, more technically, stopping the temperature rise. The great ice sheets might well continue to disintegrate, albeit slowly. Warming Natural Scientific study and consensus validates warming is a natural phenomenon McClintock 09 (Ian C. McClintock, has held many positions in the NFF and in the NSW Farmers Association where he currently serves on the Climate Change Task Force, the Business Economic & Trade Committee, and on the Executive Council. He has served on the NFF s Economic Committee, Conservation Committee, and as the NFF's representative on the National Farm Forestry Round Table, “Proof that CO2 is not the Cause of the Current Global Warming,” Lavoisier, June 2009, http://www.lavoisier.com.au/articles/greenhouse-science/climatechange/mcclintock-proofnotco2-2009.pdf) A careful study of these Reports reveals that there is in fact no empirical evidence that supports the main theory of anthropogenic global warming (AGW) (this has now been changed to the all inclusive terminology ‘climate change’2 ). Some 23 specialized computer models provide the only support for the IPCC theory. There is considerable uncertainty, as a result of a fundamental lack of knowledge about the causes of ‘climate change’, admitted by the IPCC. “There is still an incomplete physical understanding of many components of the climate system and their role in climate change. Key uncertainties include aspects of the roles played by clouds, the cryosphere, the oceans, land use and couplings between climate and biogeochemical cycles.” Significantly, our Sun provides 99.9% of the energy that drives the worlds climate systems and its varying Solar Wind protects us from galactic cosmic rays that have been shown to influence global cloud cover (and therefore temperature), and so this should be added to the list in the above quotation as this material is now largely ignored by the IPCC. It has been common in recent times to claim that there is a scientific consensus about the causes of climate change, however it has become evident that this is clearly not the case. Amongst many other examples there are now well over 31,000 scientists who have signed a petition disagreeing, on scientific grounds, with the IPCC’s claims. Science is never settled. The object of this paper is to cite a number of examples that conclusively demonstrate and prove that the claims that CO2 is the principle driver of climate change are false. Once this fact is accepted, it makes the whole objective of attempting to reduce CO2 in the atmosphere a total waste of time, effort, financial and other resources, all of which can be better directed towards tackling real problems. PROOF’S 1. LACK OF CORRELATION There is no correlation between levels of CO2 3 in the atmosphere and global temperature, in short, medium or long term historic time scales. There is a scientific truism that says, “Correlation can not prove causation, but a failure to correlate can prove non-causation”. If anthropogenic CO2 were the major climate driver (as claimed), warming would have continued to increase in synchronization with the increase in CO2 levels. (i) Short Term. The graph below indicates the present cooling period we are now experiencing. This shows the Hadley UK ground based (pink) and the satellite (blue) temperature record and the simultaneously increasing CO2 levels (green line). lack of correlation with CO2 increases and the inability to predict climate even in the very short term, let alone for 100 or more years into the future. With the Sun now entering a ‘quiet’ phase, it is anticipated that this cooling trend is likely to continue until Sun activity increases. (ii) Medium Term. This graph shows the temperature recovery (warming) that has occurred since the last gasp of the Little Ice Age (LIA) in the mid 1800’s. It shows three warmings and two coolings, in fact there are now three coolings with details of the latest one (shown on p4) not included on this graph. Again there is no correlation with CO2 levels in the atmosphere. It is worth noting that the rate of increase in the latest warming period (1975 – 1998) is similar to the rate of increase in the previous two warmings (sloping parallel bars), despite significantly higher levels of CO2 being present during the last warming. There appears to be a clear cyclical signature evident in these temperature perturbations that is absent in the CO2 data. It is very obvious that other influences are dominating (iii) Long Term. For the first 4 billion years of Earth history, the CO2 content of the atmosphere has been from 3 to 100 times the current level (Plimer. I. 2009), without any 'tipping point* leading to a runaway greenhouse effect, as predicted by the alarmists. The historical record indicates that CO2 levels have been significantly higher than they are now, being 25 times at 545 Ma, (million years ago) (Veizer. J. et al 2000). The killer proof that CO2 does not drive climate is to be found during the Ordovician- Silurian (450-420 Ma) and the Jurassic-Cretaceous periods (151-132 Ma), when CO> levels were greater than 4000 ppmv (parts per million by volume) and about 2000 ppmv respectively4. If the IPCC theory is correct there should have been runaway greenhouse induced global warming during these periods, but instead there was glaciation. This unequivocally proves that CO2 does not drive climate, it can only be a minor player, swamped by other far more powerful influences. This proves three things. Firstly, CO2 did not, and could not, have caused the warmings. Secondly, rising global temperatures were responsible for increasing levels of both CO2 and CH4 (Methane) and also Nitrous Oxide (N2O), not shown. The physics of how this might occur are well established and beyond dispute. Thirdly, increased levels of CO2 did not inhibit in any way the subsequent fall in temperatures that regularly occurred, plunging the world into another ice age. It occurs naturally Ferrara 12 (Peter Ferrara, senior fellow for entitlement and budget policy at The Heartland Institute, a senior fellow at the Social Security Institute, and the general counsel of the American Civil Rights Union, “Sorry Global Warming Alarmists, The Earth Is Cooling,” The Heartland Institute, June 1, 2012, http://blog.heartland.org/2012/06/sorry-global-warming-alarmists-the-earth-is-cooling/) Check out the 20th century temperature record, and you will find that its up and down pattern does not follow the industrial revolution’s upward march of atmospheric carbon dioxide (CO2), which is the supposed central culprit for man caused global warming (and has been much, much higher in the past). It follows instead the up and down pattern of naturally caused climate cycles . For example, temperatures dropped steadily from the late 1940s to the late 1970s. The popular press was even talking about a coming ice age. Ice ages have cyclically occurred roughly every 10,000 years, with a new one actually due around now. In the late 1970s, the natural cycles turned warm and temperatures rose until the late 1990s, a trend that political and economic interests have tried to milk mercilessly to their advantage. The incorruptible satellite measured global atmospheric temperatures show less warming during this period than the heavily manipulated land surface temperatures. Central to these natural cycles is the Pacific Decadal Oscillation (PDO). Every 25 to 30 years the oceans undergo a natural cycle where the colder water below churns to replace the warmer water at the surface, and that affects global temperatures by the fractions of a degree we have seen. The PDO was cold from the late 1940s to the late 1970s, and it was warm from the late 1970s to the late 1990s, similar to the Atlantic Multidecadal Oscillation (AMO). In 2000, the UN’s IPCC predicted that global temperatures would rise by 1 degree Celsius by 2010. Was that based on climate science, or political science to scare the public into accepting costly anti-industrial regulations and taxes? Don Easterbrook, Professor Emeritus of Geology at Western Washington University, knew the answer. He publicly predicted in 2000 that global temperatures would decline by 2010. He made that prediction because he knew the PDO had turned cold in 1999, something the political scientists at the UN’s IPCC did not know or did not think significant. Reducing emissions has no effects – empirics prove Bethel 05 (Tom Bethel, senior editor of The American Spectator, “THE FALSE ALERT OF GLOBAL WARMING,” The American Spectator, May 2005, http://spectator.org/articles/55208/false-alert-global-warming) But whether man-made carbon-dioxide emissions have caused measurable temperature increases over the last 30 years is debated. Carbon dioxide is itself a benign and essential substance, incidentally. Without it, plants would not grow, and without plant-life animals could not live. Any increase of carbon dioxide in the atmosphere causes plants, trees, and forests to grow more abundantly. It should be a treehugger's delight. The surface data suggest that man-made carbon dioxide has not in fact increased global temperatures. From 1940 to 1975, coal-fired plants emitted fumes with great abandon and without restraint by Greens. Yet the Earth cooled slightly in that time. And if man-made global warming is real, atmospheric as well as surface temperatures should have increased steadily. But they haven't. There was merely that one-time increase, possibly caused by a solar anomaly. In addition, an "urban heat island effect" has been identified. Build a tarmac runway near a weather station, and the nearby temperature readings will go up. Global warming became the focus of activism at the time of the Earth Summit in Rio, in 1992. Bush the elder signed a climate-change treaty, with signatories agreeing to reduce carbon dioxide emissions below 1990 levels. The details were worked out in Kyoto, Japan. But America was the principal target, everyone knew it, and Clinton didn't submit the treaty to the Senate for ratification. The 1990 date had been carefully chosen. Emissions in Germany and the Soviet Union were still high; Germany had just absorbed East Germany, then still using inefficient coal-fired plants. After they were modernized, Germany's emissions dropped, so the demand that they be reduced below 1990 levels had already been met and became an exercise in painless moralizing. The same was true for the Soviet Union. After its collapse, in 1991, economic activity fell by about one-third. As for France, most of its electricity comes from nuclear power, which has no global-warming effects but has been demonized for other reasons. If the enviros were serious about reducing carbon dioxide they would be urging us to build nuclear power plants, but that is not on their agenda. They want windmills (whether or not they kill golden eagles). Under the Kyoto Protocol, U.S. emissions would have to be cut so much that economic depression would have been the only certain outcome. We were expected to reduce energy use by about 35 percent within ten years, which might have meant eliminating one-third of all cars. You can see why the enviros fell in love with the idea. Impact Defense Adaption IPCC consensus proves we can adapt Rodgers 14 (Paul Rodgers, contributor of general sciences for Forbes, “Climate Change: We Can Adapt, Says IPCC,” Forbes, March 31, 2014, http://www.forbes.com/sites/paulrodgers/2014/03/31/climate-change-is-real-but-its-not-the-end-of-the-world-says-ipcc/) Yet for the first time, the IPCC is offering a glimmer of hope. It acknowledges that some of the changes will be beneficial – including higher crop yields in places like Canada, Europe and Central Asia – and that in others cases, people will be able to adapt to them. “The really big breakthrough in this report is the new idea of thinking about managing climate change,” said Dr Chris Field, the global ecology director at the Carnegie Institution in Washington and a co-chairman of the report. “We have a lot of the tools for dealing effectively with it. We just need to be smart about it. “Climate-change adaptation is not an exotic agenda that has never been tried,” said Field. “Governments, firms, and communities around the world are building experience with adaptation. This experience forms a starting point for bolder, more ambitious adaptations that will be important as climate and society continue to change.” Adaptations could include better flood defences or building houses that can withstand tropical cyclones. Vicente Barros, another co-chairman, said: “Investments in better preparation can pay dividends both for the present and for the future.” Cutting emissions will not help. Our only hope is to adapt to rising temperatures Roach 13 (John Roach, contributing writer for NBC News, “It's time to adapt to unstoppable global warming, scientists say,” NBC News, November 7, 2013, http://www.nbcnews.com/science/environment/its-time-adapt-unstoppable-global-warming-scientists-say-f8C11554338) Even if the world's 7 billion people magically stop burning fossil fuels and chopping down forests today, the greenhouse gases already emitted to the atmosphere will warm the planet by about 2 degrees Fahrenheit by the end of this century, according to scientists who are urging a focused scientific effort to help humanity adapt to the changing climate. And reality shows no sign of such a magic reduction in emissions. The amount of greenhouse gases in the atmosphere reached another new high in 2012, the World Meteorological Association announced Wednesday. In fact, concentrations of carbon dioxide, the most abundant planet warming gas, grew faster last year than its average growth rate of the past decade. "The fact is, we are not making a lot of progress in reducing emissions," Richard Moss, a senior scientist with the Pacific Northwest National Laboratory's Joint Global Change Research Institute at the University of Maryland, told NBC News. "So it seems like we really do need to face up to the fact that there is change that we can no longer avoid and we have to figure out how to manage." Moss is the lead author of an article in Thursday's issue of Science calling for the development of a field of climate science that provides relevant, tangible information to decision makers who are tasked to protect people and cities increasingly battered by extreme weather, flooded by rising seas, and nourished with food grown on drought-prone lands. Science which focuses on adapting to climate change — rather than just preventing it — is nothing new. It's the need for more information that field of science can yield that's increasingly vital. Superstorm Sandy and the onslaught of similar catastrophic events bear the fingerprint of climate change. Growing evidence that more of the same is on the way brings a new urgency for information that people can actually use to prepare, survive, and even thrive on a changing planet, Moss explained. Hope lost? The push for adaptation science also represents a shift in climate policy circles away from an agenda focused solely on cutting greenhouse gas emissions to reduce the impact of climate change, according to Adam Fenech, a climate scientist at the University of Prince Edward Island in Canada who has studied adaptation science for about 15 years. He did not contribute to the article in Science. For most of this time, "people wouldn't even consider adaptation. It was a bad word. You couldn't bring it up," he told NBC News. "But now it has been accepted probably because people have either abandoned completely or abandoned faith in the international climate change negotiations." In addition, Fenech said, people are coming to grips with the fact that the world is committed to a certain degree of climate change no matter what is done to cut emissions. The thinking goes " we're going to have to live with it anyway," he explained. "And that's adaptation." The Science article does not advocate abandoning mitigation efforts, but rather elucidates the need for adaptation, noted Joe Casola, director of the science and impacts program at the Center for Climate and Energy Solutions, an environmental think tank in Arlington, Va. "I think that both of them are going to be required," he told NBC News. "The more mitigation we do, the less adaptation we'll need," added Casola, who was not involved with the Science article. "If we were to do no mitigation, then our adaptation investments now are probably going to be made worthless or very, very limited as we have a lot of changes to deal with in the climate system." Institute of adaptation Key to making adaptation work is an acknowledgement that climate is just one factor planners are forced to wrestle with as they make decisions, noted Moss, whose lab is funded by the Department of Energy. Only adaptation solves IPCC 14 (IPCC, “Adaptation to Global Warming - IPCC Report,” Climate UU, 2014, http://climate.uu-uno.org/topics/view/23692/) Adaptation to climate change / global warming will be needed along with mitigation. Because impacts are already being observed and because the ocean takes a long time to heat up, mitigation - although crucial - will not be enough. Adaptation unfortunately will not be a simple matter. The human race is no longer composed of small groups of hunter-gatherers, but billions of people generally living in highly arranged societies with limited mobility. The worse the impacts of global warming, the more difficult adaptation will be. We should start planning for adaptation now, along with mitigation to try to lessen the impact of global warming. However, adaptation should not be used as an excuse not to mitigate global warming. The most complete compendium on adaptation is the 2014 IPCC Report, Vol WGII (Impacts, Adaptation and Vulnerability). Mitigation and adaptation are possible Pearson 11 (Charles S. Pearson, “Senior Adjunct Professor of International Economics and Environment at the Diplomatic Academy of Vienna and Professor Emeritus at the School of Advanced International Studies (SAIS), Johns Hopkins University, Washington, DC,” “Economics and the Challenge of Global Warming,” pg. 106, 2011, http://books.google.com/books?id=MDFQX2w7N5EC&pg=PA106&lpg=PA106&dq=%22global+warming%22+AND+%22no+adaptation%22&sou rce=bl&ots=Ki4SHTBz5-&sig=NQs_hBA1qjfowQITf0GsEZeXC_g&hl=en&sa=X&ei=qNu9U8riMIqpyASvYCADA&ved=0CEAQ6AEwBA#v=onepage&q=no%20adaptation&f=false) Bosello et al. (2010) also conclude that mitigation and adaptation are strategic complements. They fold into an optimization model three types of adaptation measures - anticipatory, reactive, and adaptation R&D - to sort out the relative contributions and timing of mitigation and adaptation expenditures in an optimal climate policy. The results appear to depend heavily on two factors: uncertainty regarding a catastrophic outcome, and the discount rate. The lower the dis- count rate and the greater the concern for catastrophic outcomes, the greater the role for mitigation relative to adaptation. These are not surprising because of the delay between expenditure and damages prevented with mitigation, and because adaptation copes poorly with catastrophic outcomes. They also conclude that expenditures should be tilled toward mitigation in the short run. These analyses are informative, but one should remember we are far from a world in which governments cooperate in a globally efficient mitigation effort, where all adaptation opportunities are recognized and acted on, and where allocations for mitigation and adaptation expenses are drawn from a single purse. No Extinction Warming won’t lead to complete extinction Green 11 (Roedy, PHD from British Colombia, “Extinction of Man”, http://mindprod.com/environment/extinction.html//umich-mp) Mankind is embarking on a strange ecological experiment. Over a couple of centuries, man is burning the carbon accumulated over millions of years by plants. The CO₂ levels are now at the level of the Permian extinction. There have been two mass extinctions in earth history, the Permian, 230 million years ago, was the worst. 70% of all species were lost. It was caused by natural global warming when volcanoes released greenhouse gases. (The other extinction event more familiar to most people was the more recent KT Cretaceous-Tertiary Mass Extinction event, 65 million years ago. It was caused when an asteroid plunged into the earth at Chicxulub Mexico wiping out the dinosaurs and half of earth’s species.) We are re-experiencing the same global warming conditions that triggered the more devastating Permian extinction, only this time it is man made. When it gets too hot, plants die. When it gets too hot and dry, massive fires ravage huge areas. When plants die, insects and herbivores die. When insects die, even heat-resistant plant’s don’t get pollinated and die. Birds die without insects to eat. Carnivores die without herbivores to eat, all triggered by what seems so innocuous — heat. Similarly, in the oceans, when they get just a few degrees too warm, corals expel their symbiotic algae and die soon thereafter. When coral reefs die, the fish that live on them die, triggering extinction chains. Satellites can chart the loss of vegetation over the planet. We are losing 4 species per hour, a rate on the same scale as the Permian and KT extinction events. Man has no ability to live without the support of other species. We are committing suicide and killing the family of life on earth along with us. The question is, will we wipe ourselves out along with the rest of the planet’s ecology? Man is very adaptable. He will destroy his food supply on land and in the oceans as a result, but some people will survive. That is not complete extinction. Their impacts are all empirically denied ---- past temperatures were substantially warmer than the present Idso and Idso in ‘7 (Sherwood, Research Physicist @ US Water Conservation laboratory, and Craig, President of Center for the Study of Carbon Dioxide and Global change and PhD in Geography, “Carbon Dioxide and Global Change: Separating Scientific Fact from Personal Opinion”, 6-6, http://www.co2science.org/education/reports/hansen/HansenTestimonyCritique.pdf) In an attempt to depict earth's current temperature as being extremely high and, therefore, extremely dangerous, Hansen focuses almost exclusively on a single point of the earth's surface in the Western Equatorial Pacific, for which he and others (Hansen et al., 2006) compared modern sea surface temperatures (SSTs) with paleo-SSTs that were derived by Medina-Elizade and Lea (2005) from the Mg/Ca ratios of shells of the surface-dwelling planktonic foraminifer Globigerinoides rubber concluded that “this critical ocean region, and probably the planet as a whole [our italics], is as warm now as at the Holocene maximum and within ~1°C of the maximum temperature of the past million years [our italics].” Is there any compelling reason to believe these claims of Hansen et al. about the entire planet? In a word, no, because there are a multitude of other single-point measurements that suggest something vastly different. Even in their own paper, Hansen that they obtained from an ocean sediment core. In doing so, they approximately et al. present data from the Indian Ocean that indicate, as best we can determine from their graph, that SSTs there were about 0.75°C warmer than they are currently some 125,000 years ago during the prior interglacial. Likewise, based on data obtained from the Vostok ice core in Antarctica, another of their graphs suggests that temperatures at that location some 125,000 years ago were about 1.8°C warmer than they are now; while data from two sites in the Eastern Equatorial Pacific indicate it was approximately 2.3 to 4.0°C warmer compared to the present at about that time. In fact, Petit et al.’s (1999) study of the Vostok ice core demonstrates that large periods of all four of the interglacials that preceded the Holocene were more than 2°C warmer we don’t have to go nearly so far back in time to demonstrate the nonuniqueness of current temperatures. Of the five SST records that Hansen et al. display, three of them indicate the mid-Holocene was also warmer than it is today. Indeed, it has been known for many years that the central portion of the current than the peak warmth of the current interglacial. But interglacial was much warmer than its latter stages have been. To cite just a few examples of pertinent work conducted in the 1970s and 80s – based on temperature reconstructions derived from studies of latitudinal displacements of terrestrial vegetation (Bernabo and Webb, 1977; Wijmstra, 1978; Davis et al., 1980; Ritchie et al., 1983; Overpeck, 1985) and vertical displacements of alpine plants (Kearney and Luckman, 1983) and mountain glaciers (Hope et al., 1976; Porter and Orombelli, 1985) – we note it was concluded by Webb et al. (1987) and the many COHMAP Members (1988) that mean annual temperatures in the Midwestern United States were about 2°C greater than those of the past few decades (Bartlein et al., 1984; Webb, 1985), that summer temperatures in Europe were 2°C warmer (Huntley and Prentice, 1988) – as they also were in New Guinea (Hope et al., 1976) – and that temperatures in the Alps were as much as 4°C warmer (Porter and Orombelli, 1985; Huntley and Prentice, 1988). Likewise, temperatures in the Russian Far East are reported to have been from 2°C (Velitchko and Klimanov, 1990) to as much as 46°C (Korotky et al., 1988) higher than they were in the 1970s and 80s; while the mean annual temperature of the Kuroshio Current between 22 and 35°N was 6°C warmer (Taira, 1975). Also, the southern boundary of the Pacific boreal region was positioned some 700 to 800 km north of its present location (Lutaenko, 1993). But we needn’t go back to even the mid-Holocene to the Medieval Warm Period, centered on about AD 1100, had lots of them. In fact, every single week since 1 Feb 2006, testifies to the existence of this severalcenturies-long period of notable warmth, in a feature we call our Medieval Warm Period Record of the Week. Also, whenever it has been possible to make encounter warmer-than-present temperatures, as we have featured on our website (www.co2science.org) a different peer-reviewed scientific journal article that either a quantitative or qualitative comparison between the peak temperature of the Medieval Warm Period (MWP) and the peak temperature of the Current Warm Period (CWP), we have a quick perusal of these ever-growing databases (reproduced below as of 23 May 2007) indicates that, in the overwhelming majority of cases, the peak warmth of the Medieval Warm Period was significantly greater than the peak warmth of the Current Warm Period. included those results in the appropriate quantitative or qualitative frequency distributions we have posted within this feature; and Global warming is nearing its end Solomon 09 (Lawrence Solomon, writer for the Financial Post, “The end is near; The media, polls and even scientists suggest the global warming scare is all over but the shouting,” National Post, October 3, 2009, http://www.lexisnexis.com/hottopics/lnacademic/) The great global warming scare is over -- it is well past its peak, very much a spent force, sputtering in fits and starts to a whimpering end. You may not know this yet. Or rather, you may know it but don't want to acknowledge it until every one else does, and that won't happen until the press, much of which also knows it, formally acknowledges it. I know that the global warming scare is over but for the shouting because that's what the polls show, at least those in the U.S., where unlike Canada the public is polled extensively on global warming. Most Americans don't blame humans for climate change -- they consider global warming to be a natural phenomenon. Even when the polls showed the public believed man was responsible for global warming, the public didn't take the scare seriously. When asked to rank global warming's importance compared to numerous other concerns -- unemployment, trade, health care, poverty, crime, and education among them -- global warming came in dead last. Fewer than 1% chose global warming as scare-worthy. The informed members of the media read those polls and know the global warming scare is over, too. Andrew Revkin, The New York Times reporter entrusted with the global warming scare beat, has for months lamented "the public's waning interest in global warming." His colleague at the Washington Post, Andrew Freedman, does his best to revive public fear, and to get politicians to act, by urging experts to up their hype so that the press will have scarier material to run with. The experts do their best to give us the willies. This week they offered up plagues of locusts in China and a warning that the 2016 Olympics "could be the last for mankind" because "the earth has passed the point of no return." But the press has also begun to tire of Armageddon All-The-Time, and (I believe) to position itself for its inevitable attack on the doomsters. In an online article in June entitled "Massive Estimates of Death are in Vogue for Copenhagen," Richard Cable of the BBC, until then the most stalwart of scaremongers, rattled off the global warnings du jour -they included a comparison of global warming to nuclear war and a report from the former Secretary General of the UN, Kofi Annan, to the effect that "every year climate change leaves over 300,000 people dead, 325-million people seriously affected, and economic losses of US $125 -billion." Cable's conclusion: "The problem is that once you've sat up and paid attention enough to examine them a bit more closely, you find that the means by which the figures were arrived at isn't very compelling... The report contains so many extrapolations derived from guesswork based on estimates inferred from unsuitable data." The scientist-scare-mongers, seeing the diminishing returns that come of their escalating claims of catastrophe, also know their stock is falling. Until now, they have all toughed it out when the data disagreed with their findings -as it does on every major climate issue, without exception. Some scientists, like Germany's Mojib Latif, have begun to break ranks. Frustrated by embarrassing questions about why the world hasn't seen any warming over the last decade, Latif, a tireless veteran of the public speaking circuits, now explains that global warming has paused, to resume in 2020 or perhaps 2030. "People understand what I'm saying but then basically wind up saying, 'We don't believe anything,'" he told The New York Times this week. And why should they believe anything that comes from the global warming camp? Not only has the globe not warmed over the last decade but the Arctic ice is returning , the Antarctic isn't shrinking, polar bear populations aren't diminishing, hurricanes aren't becoming more extreme. The only thing that's scary about the science is the frequency with which doomsayer data is hidden from public scrutiny, manipulated to mislead, or simply made up. None of this matters anymore, I recently heard at the Global Business Forum in Banff, where a fellow panelist from the Pew Centre on Global Climate Change told the audience that, while she couldn't dispute the claims I had made about the science being dubious, the rights and wrongs in the global warming debate are no longer relevant. "The train has left the station," she cheerily told the business audience, meaning that the debate is over, global warming regulations are coming in, and everyone in the room -- primarily business movers and shakers from Western Canada --had better learn to adapt. CO2 GOOD (AG DA) 1NC (USE AS CASE TURNS) High levels of CO2 are key to environmental prosperity Carlisle 01 (John Carlisle, national policy analyst, “Carbon Dioxide is Good for the Environment,” National Policy Analysis, April 2001, http://www.nationalcenter.org/NPA334.html) Carbon dioxide is good for the environment. That simple fact must be restated to counter environmentalists' baseless allegations that the accumulation of man-made carbon dioxide, produced by cars, power plants and other human activities, is causing dangerous global warming. Indeed, far from being a poisonous gas that will wreak havoc on the planet's ecosystem, carbon dioxide is arguably the Earth's best friend in that trees, wheat, peanuts, flowers, cotton and numerous other plants significantly benefit from increased levels of atmospheric carbon dioxide. Dr. Craig Idso of the Center for the Study of Carbon Dioxide and Global Change, one of the nation's leading carbon dioxide research centers, examined records of atmospheric carbon dioxide concentrations and air temperature over the last 250,000 years. There were three dramatic episodes of global warming that occurred at the end of the last three ice ages. Interestingly, temperatures started to rise during those warming periods well before the atmospheric carbon dioxide started to increase. In fact, the carbon dioxide levels did not begin to rise until 400 to 1,000 years after the planet began to warm. Concludes Dr. Idso, "Clearly, there is no way that these real-world observations can be construed to even hint at the possibility that a significant increase in atmospheric carbon dioxide will necessarily lead to any global warming."1 On the other hand, scientists have lots of evidence demonstrating that increased carbon dioxide levels leads to healthier plants. A team of scientists in Nevada conducted a five-year experiment in which they grew one group of ponderosa pine trees at the current carbon dioxide atmospheric level of about 360 parts per million (ppm) and another group of pines at 700 ppm. The doubled carbon dioxide level increased tree height by 43 percent and diameter by 24 percent. Similarly, a team of scientists from Virginia Tech University reported that growing loblolly pine trees in a greenhouse with a carbon dioxide concentration of 700 ppm increased average tree height 9 percent, diameter by 7 percent, needle biomass by 16 percent and root biomass by 33 percent.2 Increased atmospheric carbon dioxide doesn't just make a plant bigger. Carbon dioxide also makes plants more resistant to extreme weather conditions . In a study discussed in the journal Plant Ecology, a team of scientists subjected the Mojave Desert evergreen shrub to three different concentrations of carbon dioxide - the current level of 360 ppm and at 550 ppm and 700 ppm. The plants, which were being grown in simulated drought conditions, responded more favorably in the carbon dioxiderich environments. Photosynthetic activity doubled in the 550 ppm environment and tripled at 700 ppm. Increased photosynthetic activity enables plants to withstand drought better.3 Likewise, a team of biologists grew seedlings of three yucca plants in cooler greenhouse environments at the 360 ppm and 700 ppm concentrations. The yucca plants exposed to the enhanced carbon dioxide concentration showed a greater resistance to the colder temperatures. Dr. Robert Balling, a climatologist at Arizona State University, notes that by making plants healthier and more resistant to extreme weather conditions, higher levels of atmospheric carbon dioxide expands the habitat of many plants, improves rangeland in semi-arid areas and enhances agricultural productivity in arid areas.4 Another benefit of enhanced atmospheric carbon dioxide is that it helps the tropical rainforests. Scientists from Venezuela and the United Kingdom grew several species of tropical trees and other plants in greenhouse conditions at carbon dioxide concentrations double the current level. The plants responded favorably, showing an increase in photosynthetic activity. The scientists concluded that, "In a future atmosphere with a higher carbon dioxide concentration, these species should be able to show a higher productivity than today."5 Another team of British and New Zealand researchers grew tropical trees for 119 days at elevated levels of atmospheric carbon dioxide. They found that the enriched carbon dioxide environment stimulated the trees' root growth by 23 percent. Expanded root systems help tropical trees by increasing their ability to absorb water and nutrients.6 Bigger trees, increased resistance to bad weather, improved agricultural productivity and a boon to rainforests are just some of the many benefits that carbon dioxide bestows on the environment. With little evidence that carbon dioxide triggers dangerous global warming but lots of evidence showing how carbon dioxide helps the environment, environmentalists should be extolling the virtues of this benign greenhouse gas. CO2 good for the environment Idso and Idso 10 (Keith Idso, Vice President of the Center for the Study of Carbon Dioxide and Global Change, Craig Idso, founder, former president and current chairman of the board of the Center for the Study of Carbon Dioxide and Global Change, “Feeding the Future World ,” CO2 Science, September 29, 2010, http://www.co2science.org/articles/V13/N39/EDIT.php) In terms of confronting this daunting challenge, Zhu et al. say that "meeting future increases in demand will have to come from a near doubling of productivity on a land area basis," and they opine that "a large contribution will have to come from improved photosynthetic conversion efficiency," for which they estimate that "at least a 50% improvement will be required to double global production." The researchers' reason for focusing on photosynthetic conversion efficiency derives from the experimentally-observed facts that (1) increases in the atmosphere's CO2 concentration increase the photosynthetic rates of nearly all plants, and that (2) those rate increases generally lead to equivalent -- or only slightly smaller -- increases in plant productivity on a land area basis, thereby providing a solid foundation for their enthusiasm in this regard. In their review of the matter, however, they examine the prospects for boosting photosynthetic conversion efficiency in an entirely different way: by doing it genetically and without increasing the air's CO2 content. So what is the likelihood that their goal can be reached via this approach? High levels of atmospheric CO2 spur plant growth Idso and Idso 14 (Dr. Sherwood B. Idso, president of the Center for the Study of Carbon Dioxide and Global Change, Dr. Keith E. Idso, , Vice President of the Center for the Study of Carbon Dioxide and Global Change, “CO2 to the Rescue ... Again!,” CO2 Science, 2014, http://www.co2science.org/articles/V5/N22/COM.php) Atmospheric CO2 enrichment has long been known to help earth's plants withstand the debilitating effects of various environmental stresses, such as high temperature, excessive salinity levels and deleterious air pollution, as well as the negative consequences of certain resource limitations, such as less than optimal levels of light, water and nutrients (Idso and Idso, 1994). Now, in an important new study, Johnson et al. (2002) present evidence indicating that elevated levels of atmospheric CO2 do the same thing for soil microbes in the face of the enhanced receipt of solar ultraviolet-B radiation that would be expected to occur in response to a 15% depletion of the earth's stratospheric ozone layer. In addition, their study demonstrates that this phenomenon will likely have important consequences for soil carbon sequestration. Johnson et al. conducted their landmark work on experimental plots of subarctic heath located close to the Abisko Scientific Research Station in Swedish Lapland (68.35°N, 18.82°E). The plots they studied were composed of open canopies of Betula pubescens ssp. czerepanovii and dense dwarf-shrub layers containing scattered herbs and grasses. For a period of five years, the scientists exposed the plots to factorial combinations of UV-B radiation - ambient and that expected to result from a 15% stratospheric ozone depletion - and atmospheric CO2 concentration - ambient (around 365 ppm) and enriched (around 600 ppm) - after which they determined the amounts of microbial carbon (Cmic) and nitrogen (Nmic) in the soils of the plots. When the plots were exposed to the enhanced UV-B radiation level expected to result from a 15% depletion of the planet's stratospheric ozone layer, the researchers found that the amount of Cmic in the soil was reduced to only 37% of what it was at the ambient UV-B level when the air's CO2 content was maintained at the ambient concentration. When the UV-B increase was accompanied by the CO2 increase, only was there not a decrease in Cmic, there was an actual increase of fully 37%. The story with respect to Nmic was both similar and different at one and the same time. In this case, when the plots were exposed to the enhanced level of UV-B radiation, the amount of Nmic in the soil experienced a 69% increase when the air's CO2 content was maintained at the ambient concentration. When the UV-B increase was accompanied by the CO2 increase, however, Nmic rose even more, experiencing a whopping 138% increase. These findings, in the words of Johnson et al., "may have far-reaching implications ... because the productivity of many semihowever, not natural ecosystems is limited by N (Ellenberg, 1988)." Hence, the 138% increase in soil microbial N observed in this study to accompany a 15% reduction in stratospheric ozone and a concomitant 64% increase in atmospheric CO2concentration (experienced in going from 365 ppm to 600 ppm) should do wonders in enhancing the input of plant litter to the soils of these ecosystems, which phenomenon represents the first half of the carbon sequestration process, i.e., the carbon input stage. Biodiversity Mapping Can’t Solve Can’t solve, too many regional differences for one machine to cover Rosenfeld 12 (Dr. Leslie Rosenfeld is currently a Research Associate Professor at the Naval Postgraduate School and an oceanographic consultant. She has a Ph.D. in Physical Oceanography, and is a recognized expert on the California Current System, specializing in circulation over the continental shelf. December 2012. Synthesis of Regional IOOS Build-out Plans ¶ for the Next Decade from the Integrated Ocean Observing System Association. http://www.ioosassociation.org/sites/nfra/files/documents/ioos_documents/regional/BOP%20Synthesis%20Final.pdf July 6, 2014.) Types of hazards and needed products have many common elements throughout the country but ¶ also unique regional differences. For example, the Gulf of Mexico, the Caribbean and the ¶ southeastern U.S. experience more frequent and more intense hurricanes than other regions of ¶ the country. Substantial improvements in the NWS forecasts of storm intensity, track, and ¶ timing of passage are necessary for timely evacuations of communities and offshore facilities in ¶ the path of the storm, while avoiding evacuations that are unnecessary. Long-term plans for ¶ these regions include close coordination of data productS with the needs of hurricane modelers, ¶ including providing information on ocean heat content via air-deployed sensors during hurricane ¶ approach and passage, and autonomous underwater vehicles to monitor the water column. ¶ With the rapid loss of sea ice, Arctic weather and ocean conditions are increasingly endangering ¶ Alaska Native coastal communities. In a statewide assessment, flooding and erosion affects 184 ¶ out of 213 Native villages (GAO, 2003). This presents unique challenges in forecasting and ¶ effectively communicating conditions for small communities in isolated locations. These and ¶ many other types of regional differences must be considered when tailoring and refining ¶ common information needs. Regional Differences just make it to hard for the aff to solve in its entirety Rosenfeld 12 (Dr. Leslie Rosenfeld is currently a Research Associate Professor at the Naval Postgraduate School and an oceanographic consultant. She has a Ph.D. in Physical Oceanography, and is a recognized expert on the California Current System, specializing in circulation over the continental shelf. December 2012. Synthesis of Regional IOOS Build-out Plans ¶ for the Next Decade from the Integrated Ocean Observing System Association. http://www.ioosassociation.org/sites/nfra/files/documents/ioos_documents/regional/BOP%20Synthesis%20Final.pdf July 6, 2014.) Water quality issues also exhibit a variety of regionally unique differences. For example, water ¶ quality contamination in the Great Lakes is not only an issue for wildlife, but also for 40 million ¶ consumers of public drinking water in the region. To address this concern, GLOS will provide a ¶ decision support tool to track the movement of drinking water contaminants, and will provide 19 ¶ model output that helps county officials manage drinking water intake systems to avoid ¶ contamination. ¶ ¶ Tracking of water quality plumes and impacts also has unique challenges in the Gulf of Mexico. ¶ Drainage from urban development, industries and farmland comes into the Gulf from the ¶ enormous Mississippi River watershed, covering 1,245,000 square miles and 41% of the 48 ¶ contiguous states of the U.S. This drainage leads to hypoxic conditions (low dissolved oxygen ¶ content) in summertime over the TexasLouisiana shelf waters, and delivers large loads of ¶ sediment and associated pollutants to nearshore environments such as recreational waters and ¶ shellfish beds. Informed management decisions require a broad distribution of platforms, ¶ sensors and derived products to track this plume and its impacts through estuaries, nearshore and ¶ offshore habitats. Mapping can’t solve—too difficult to perfect and maintain Rosenfeld 12 (Dr. Leslie Rosenfeld is currently a Research Associate Professor at the Naval Postgraduate School and an oceanographic consultant. She has a Ph.D. in Physical Oceanography, and is a recognized expert on the California Current System, specializing in circulation over the continental shelf. December 2012. Synthesis of Regional IOOS Build-out Plans ¶ for the Next Decade from the Integrated Ocean Observing System Association. http://www.ioosassociation.org/sites/nfra/files/documents/ioos_documents/regional/BOP%20Synthesis%20Final.pdf July 6, 2014.) Effective utilization of the observing system requires translation of the data and model outputs ¶ into targeted products and decision support tools for a variety of users. The process of product ¶ development ultimately links management decisions to desired products, to the information ¶ needed to produce the products, to data and models required to create that information and ¶ finally to the essential observing system requirements. ¶ ¶ Successful engagement with users and the resultant product development is at the heart of ¶ ultimate success of the 10-year build-out plan. The 27 key products identified as part of the ¶ build-out plans include products that have already been successfully developed and utilized by ¶ one or more RAs as well as many cases where development is either not underway or is in the ¶ early stages. Successful completion of development for these 27 products for all the regions, and ¶ completion of additional unique regional products will involve multiple components, as outlined ¶ below. Sufficient funding needs to be identified to carry the process for a given product need ¶ through to completion, to ensure effective management and meeting of user expectations. ¶ ¶ Iterative two-way engagement with users is required for effective product development, often ¶ based on small group discussions with key decision makers and technical staff in the user ¶ community. It may also include surveys, user advisory panels, targeted workshops, and ¶ discussions at existing forums of the targeted users. This provides an understanding of the user’s ¶ objectives, the scope of the issue and the decision processes they use. Those user decisions or ¶ management needs that could most benefit from IOOS products can then be identified and ¶ prioritized, including products involving data, processed information, visualizations, models and ¶ decision support tools. ¶ ¶ Product development teams focused on specific priority needs should be created, including RA ¶ staff, key partners and users. The RAs should act as intermediaries that can translate between ¶ researchers and users, and should stay engaged throughout the process. After fully ¶ understanding user needs, the teams evaluate the variables, temporal and spatial scale and ¶ resolution, and data quality required to meet those needs. They evaluate current coastal IOOS, or ¶ other, products that could be used or modified, or new products that could be developed to meet ¶ needs. This should include evaluation across all 11 RAs and federal agencies to identify useful 26 ¶ building blocks and approaches for products. Identification of any gaps between the user’s ¶ information requirements and current coastal IOOS capabilities can then be identified, and, ¶ where possible, technical means to fill them can be developed. This should include engaging ¶ partners who may provide critical data or fulfill specific modeling needs, and pursuit of funding ¶ sources to fill gaps, if not already obtained. ¶ ¶ In addition to addressing technical issues, the team should identify any institutional issues among ¶ users that must be overcome to fully utilize IOOS products, and where possible work with them ¶ to develop approaches to overcome these barriers. Institutional issues include agency and ¶ industry policies and practices, regulations and permit procedures, perspectives and ¶ communication patterns, staff expertise, training and workload, and/or governance structures that ¶ could impede full use of IOOS information in decision-making. Addressing these issues will ¶ likely require engagement with management representatives of the user agency or organization. ¶ ¶ Development of the initial product includes software development to access and integrate data, ¶ and any necessary documentation and instructions, including description of the metadata. ¶ Testing and evaluation of the product must be conducted with a subset of the user group, ¶ followed by training for a larger user group, along with refinement of the product as needed. ¶ ¶ Product development also includes evaluation of the most effective means to package and ¶ distribute the information, e.g. through websites, e-mail notifications, use of existing channels ¶ already favored by the user, reports, etc. Product release should be prefaced by an evaluation to ¶ ensure that the product is ready to move from the developmental stage to full operational use and ¶ broad implementation, and a notification system for the targeted user group. Operational ¶ implementation of the product should include ongoing engagement with the users to adapt and ¶ modify products as needed as experience builds with product use, conditions or requirements ¶ change, or new data becomes available. ¶ ¶ As is evident from the above list, development of effective user products is a critical but costly ¶ and timeconsuming endeavor, and often includes work that is different from what is involved in ¶ the modeling and data management subsystems. Staff needs for product development and user ¶ engagement typically range from five to six full-time equivalents (FTE)s for each RA, and ¶ involve multiple types of staff including technical, outreach and management needed at various ¶ times during the activities outlined above. Meeting the need for adequate product development ¶ funding and staff time, and achieving coordination and synergism among the RAs in creating the ¶ common products will be an essential component of success in implementing the build-out plan ¶ over the next 10 years. SQ Solves-Mapping Data Insufficiency is being fixed in the Status Quo—the plan isn’t needed Piotroski 14 (Jon Piotroski is known for his articles on the science development website. He has written numerous articles pertaining to oceanography, climate change, and other environmental issues. 03/20/14 from SciDev. Tidal wave of ocean data leaves scientists swamped. http://www.scidev.net/global/data/news/tidal-wave-of-ocean-data-leaves-scientistsswamped.html July 3, 2014) A lack of data curators and managers capable of cleaning up observational measurements, particularly in less developed nations, is limiting the scale and scope of ocean research, researchers have said on the sidelines of an oceans science meeting.¶ To deal with the challenge, the Ocean Sciences Meeting in Hawaii last month (23-28 February) heard proposals for creating a comprehensive, quality controlled database.¶ This would help overcome limited scientific capacity and allow under-resourced researchers to better participate in efforts to understand pressing issues such as climate change, scientists said.¶ The complexities involved in transforming raw oceanographic data into a useable state means researchers with limited data skills or financial resources are less able to meet the necessary standard, said Stephen Diggs, a data manager at the Scripps Institution of Oceanography, in the United States.¶ The explosion of ocean data being collected through modern methods exceeds scientists’ capacity to deal with it, he told SciDev.Net on the fringes of the conference.¶ “There definitely needs to be workforce development to get more data curators and managers that come out of their training ready to deal with this,” he said.¶ Interdisciplinary work that is crucial for answering questions such as what effect climate change is having on marine environments is especially sensitive to reliability issues, he added.¶ This is because, when faced with data sets from various disciplines, researchers often lack the expertise to spot and correct all the biases created by the idiosyncrasies of different regions and collection methods, he said.¶ Providing researchers with rigorously quality controlled data from outside their immediate field was the motivation behind a proposed initiative, which called for scientists’ support at a town hall meeting during the conference.¶ The International Quality-Controlled Ocean Database (IQuOD) would combine automated, large-scale data cleaning done by computers with finer-grained expert analysis. The goal would be to produce a definitive database of the data from the 13 million locations with sea surface temperature records — some dating back to the eighteenth century — scattered across various institutions.¶ According to Diggs, this would significantly improve on the oceanographic database hosted at the US National Oceanic and Atmospheric Administration’s (NOAA’s) National Oceanographic Data Center. Even though NOAA's only conducts limited automated quality control its repository of marine datasets is the world’s most comprehensive.¶ Other variables, such as salinity and oxygen concentrations in the water, could be added to the temperature data as the IQuOD project progresses, said Diggs. IOOS and other tech already solve National Academy 11 (provide a public service by working outside the framework of government to ensure independent advice on matters of ¶ science, technology, and medicine. They enlist committees of the nation’s top scientists, engineers, and other experts— all of whom volunteer their time to study specific concerns. 08/24/2011. Ocean Exploration from the National Academies. http://oceanleadership.org/wp-content/uploads/2009/08/Ocean_Exploration.pdf July 6, 2014). ¶ U.S. and international efforts have made significant progress in recent years to-¶ ward establishing ocean observatories. The Integrated Ocean Observing System ¶ (IOOS), the U.S. contribution to the international Global Ocean Observing Sys-¶ tem, is designed as a large network that collects highresolution data along the ¶ entire U.S. coast. This information supports a wide range of operational services, ¶ including weather forecasts, coastal warning systems, and the monitoring of algal blooms. ¶ The Ocean Observatories Initiative (OOI), a National ¶ Science Foundation program, will enable landbased ¶ exploration and monitoring of processes throughout the ¶ seafloor, water column, and overlying atmosphere by ¶ real-time, remote interactions with arrays of sensors, in-¶ struments, and autonomous vehicles. For the first time, ¶ The world’s first underwater cabled observatory to span an entire ¶ plate will be installed as part of a collaborative effort between the ¶ U.S. Ocean Observatories Initiative (OOI) and U.S. and Canadian ¶ institutions. Fiberoptic cables will extend high bandwidth and power ¶ to a network of hundreds of sensors across, above, and below the ¶ seafloor, allowing in situ, interactive monitoring of the ocean and ¶ seafloor for the next 20 to 30 years. Image courtesy University of ¶ Washington Center for Environmental Visualization.¶ Discover scientists, educators and ¶ the public will have real-¶ time access to the data and ¶ imagery being collected, so that ¶ they can learn about events such as ¶ underwater volcanic eruptions and anoxia ¶ events as they happen, even if there are no ships ¶ in the area. SQ Conservation Solves Status Quo Solves -- US already implementing conservation programs to prevent impacts Tullo 14 (Michelle Tulo is a writer for various news magazines, in particular IPS. IPS is a reputable communication institution that specializes in global news. U.S. Turns Attention to Ocean Conservation, Food Security from IPS (Inter Press Service) on 06/19/14. http://www.ipsnews.net/2014/06/u-s-turns-attention-to-ocean-conservation-food-security/ July 7, 2014.) A first-time U.S.-hosted summit on protecting the oceans has resulted in pledges worth some 800 million dollars to be used for conservation efforts.¶ During the summit, held here in Washington, the administration of President Barack Obama pledged to massively expand U.S.-protected parts of the southern Pacific Ocean. In an effort to strengthen global food security, the president has also announced a major push against illegal fishing and to create a national strategic plan for aquaculture.¶ “If we drain our resources, we won’t just be squandering one of humanity’s greatest treasures, we’ll be cutting off one of the world’s leading sources of food and economic growth, including for the United States,” President Obama said via video Tuesday morning.¶ The “Our Ocean” conference, held Monday and Tuesday at the U.S. State Department, brought together ministers, heads of state, as well as civil society and private sector representatives from almost 90 countries. The summit, hosted by Secretary of State John Kerry, focused on overfishing, pollution and ocean acidification, all of which threaten global food security.¶ In his opening remarks, Kerry noted that ocean conservation constitutes a “great necessity” for food security. “More than three billion people, 50 percent of the people on this planet, in every corner of the world depend on fish as a significant source of protein,” he said.¶ Proponents hope that many of the solutions being used by U.S. scientists, policymakers and fishermen could serve to help international communities.¶ “There is increasing demand for seafood with diminished supply … We need to find ways to make seafood sustainable to rich and poor countries alike,” Danielle Nierenberg, the president of FoodTank, a Washington think tank, told IPS. No Impact-Species No species snowball Roger A Sedjo 2k, Sr. Fellow, Resources for the Future, Conserving Nature’s Biodiversity: insights from biology, ethics & economics, eds. Van Kooten, Bulte and Sinclair, p 114 As a critical input into the existence of humans and of life on earth, biodiversity obviously has a very high value (at least to humans). But, as with other resource questions, including public goods, biodiversity is not an either/or question, but rather a question of “how much.” Thus, we may argue as to how much biodiversity is desirable or is required for human life (threshold) and how much is desirable (insurance) and at what price, just as societies argue over the appropriate amount and cost of national defense. As discussed by Simpson, the value of water is small even though it is essential to human life, while diamonds are inessential but valuable to humans. The reason has to do with relative abundance and scarcity, with market value pertaining to the marginal unit. This water-diamond paradox can be applied to biodiversity. Although biological diversity is essential, a single species has only limited value, since the global system will continue to function without that species. Similarly, the value of a piece of biodiversity (e.g., 10 ha of tropical forest) is small to negligible since its contribution to the functioning of the global biodiversity is negligible. The global ecosystem can function with “somewhat more” or “somewhat less” biodiversity, since there have been larger amounts in times past and some losses in recent times. Therefore, in the absence of evidence to indicate that small habitat losses threaten the functioning of the global life support system, the value of these marginal habitats is negligible. The “value question” is that of how valuable to the life support function are species at the margin. While this, in principle, is an empirical question, in practice it is probably unknowable. However, thus far, biodiversity losses appear to have had little or no effect on the functioning of the earth’s life support system, presumably due to the resiliency of the system, which perhaps is due to the redundancy found in the system. Through most of its existence, earth has had far less biological diversity. Thus, as in the water-diamond paradox, the value of the marginal unit of biodiversity appears to be very small. No extinction Easterbrook 2003 Gregg, senior fellow at the New Republic, We're All Gonna Die! http://www.wired.com/wired/archive/11.07/doomsday.html?pg=2&topic=&topic_set= If we're talking about doomsday - the end of human civilization - many scenarios simply don't measure up. A single nuclear bomb ignited by terrorists, for example, would be awful beyond words, but life would go on. People and machines might converge in ways that you and I would find ghastly, but from the standpoint of the future, they would probably represent an adaptation. Environmental collapse might make parts of the globe unpleasant, but considering that the biosphere has survived ice ages, it wouldn't be the final curtain. Depression, which has become 10 times more prevalent in Western nations in the postwar era, might grow so widespread that vast numbers of people would refuse to get out of bed, a possibility that Petranek suggested in a doomsday talk at the Technology Entertainment Design conference in 2002. But Marcel Proust, as miserable as he was, wrote Remembrance of Things Past while lying in bed. No Impact-Exaggerated Environmental threats exaggerated Gordon 95 - a professor of mineral economics at Pennsylvania State University [Gordon, Richard, “Ecorealism Exposed,” Regulation, 1995, http://www.cato.org/pubs/regulation/regv18n3/reg18n3readings.html Easterbrook's argument is that although environmental problems deserve attention, the environmental movement has exaggerated the threats and ignored evidence of improvement. His discontent causes him to adopt and incessantly employ the pejoratively intended (and irritating) shorthand "enviros" to describe the leading environmental organizations and their admirers. He proposes-and overuses-an equally infelicitous alternative phrase, "ecorealism," that seems to mean that most environmental initiatives can be justifited by more moderate arguments. Given the mass, range, and defects of the book, any review of reasonable length must be selective. Easterbrook's critique begins with an overview of environmentalism from a global perspective. He then turns to a much longer (almost 500page) survey of many specific environmental issues. The overview section is a shorter, more devastating criticism, but it is also more speculative than the survey of specific issues. In essence, the overview argument is that human impacts on the environment are minor, easily correctable influences on a world affected by far more powerful forces. That is a more penetrating criticism than typically appears in works expressing skepticism about environmentalism. Easterbrook notes that mankind's effects on nature long predate industrialization or the white colonization of America, but still have had only minor impacts. We are then reminded of the vast, often highly destructive changes that occur naturally and the recuperative power of natural systems. Prefer our evidence – Environmental apocalypse scenarios are always overblown and recent human advancements solve. Ronald Bailey 2k, science correspondent, author of Earth Report 2000: Revisiting the True State of the Planet, former Brookes Fellow in Environmental Journalism at the Competitive Enterprise Institute, member of the Society of Environmental Journalists, adjunct scholar at the Cato Institute, May 2000, Reason Magazine, “Earth Day, Then and Now,” http://reason.com/0005/fe.rb.earth.shtml Earth Day 1970 provoked a torrent of apocalyptic predictions. “We have about five more years at the outside to do something,” ecologist Kenneth Watt declared to a Swarthmore College audience on April 19, 1970. Harvard biologist George Wald estimated that “civilization will end within 15 or 30 years unless immediate action is taken against problems facing mankind.” “We are in an environmental crisis which threatens the survival of this nation, and of the world as a suitable place of human habitation,” wrote Washington University biologist Barry Commoner in the Earth Day issue of the scholarly journal Environment. The day after Earth Day, even the staid New York Times editorial page warned, “Man must stop pollution and conserve his resources, not merely to enhance existence but to save the race from intolerable deterioration and possible extinction.” Very Apocalypse Now. Three decades later, of course, the world hasn’t come to an end; if anything, the planet’s ecological future has never looked so promising. With half a billion people suiting up around the globe for Earth Day 2000, now is a good time to look back on the predictions made at the first Earth Day and see how they’ve held up and what we can learn from them. The short answer: The prophets of doom were not simply wrong, but spectacularly wrong. More important, many contemporary environmental alarmists are similarly mistaken when they continue to insist that the Earth’s future remains an eco-tragedy that has already entered its final act. Such doomsters not only fail to appreciate the huge environmental gains made over the past 30 years, they ignore the simple fact that increased wealth, population, and technological innovation don’t degrade and destroy the environment. Rather, such developments preserve and enrich the environment. If it is impossible to predict fully the future, it is nonetheless possible to learn from the past. And the best lesson we can learn from revisiting the discourse surrounding the very first Earth Day is that passionate concern, however sincere, is no substitute for rational analysis. No Impact-Resilient No impact --- Ocean ecosystems are resilient CO2 Science 2008 Marine Ecosystem Response to "Ocean Acidification" Due to Atmospheric CO2 Enrichment, Vogt, M., Steinke, M., Turner, S., Paulino, A., Meyerhofer, M., Riebesell, U., LeQuere, C. and Liss, P. 2008. Dynamics of dimethylsulphoniopropionate and dimethylsulphide under different CO2 concentrations during a mesocosm experiment. Biogeosciences 5: 407-419, http://www.co2science.org/articles/V11/N29/B2.php Vogt et al. report that they detected no significant phytoplankton species shifts between treatments, and that " the ecosystem composition, bacterial and phytoplankton abundances and productivity, grazing rates and total grazer abundance and reproduction were not significantly affected by CO2 induced effects," citing in support of this statement the work of Riebesell et al. (2007), Riebesell et al. (2008), Egge et al. (2007), Paulino et al. (2007), Larsen et al. (2007), Suffrian et al. (2008) and Carotenuto et al. (2007). In addition, they say that "while DMS stayed elevated in the treatments with elevated CO2, we observed a steep decline in DMS concentration in the treatment with low CO2," i.e., the ambient CO2 treatment. What it means With respect to their many findings, the eight researchers say their observations suggest that "the system under study was surprisingly resilient to abrupt and large pH changes," which is just the opposite of what the world's climate alarmists characteristically predict about CO2-induced "ocean acidification." And that may be why Vogt et al. described the marine ecosystem they studied as "surprisingly resilient" to such change: it may have been a little unexpected. Nature sustains damage and recovers. Easterbrook 95 Distinguished Fellow, Fulbright Foundation (Gregg, A Moment on Earth) Nature is not ending, nor is human damage to the environment “unprecedented.” Nature has repelled forces of a magnitude many times greater than the worst human malfeasance. Nature is no ponderously slow. It’s just old. Old and slow are quite different concepts. That the living world can adjust with surprising alacrity is the reason nature has been able to get old. Most natural recoveries from ecological duress happen with amazing speed. Significant human tampering with the environment has been in progress for at least ten millennia and perhaps longer. If nature has been interacting with genus Homo for thousands of years, then the living things that made it to the present day may be ones whose genetic treasury renders them best suited to resist human mischief. This does not ensure any creature will continue to survive any clash with humankind. It does make survival more likely than doomsday orthodox asserts. If nature’s adjustment to the human presence began thousands of years ago, perhaps it will soon be complete. Far from reeling helplessly before a human onslaught, nature may be on the verge of reasserting itself. Nature still rules much more of the Earth than does genus Homo. To the statistical majority of natures creatures the arrival of men and women goes unnoticed. No Impact-long TF Their impact has a three hundred year timeframe CO2 Science 2005 The Fate of Fish in a High-CO2 World, Ishimatsu, A., Hayashi, M., Lee, K.-S., Kikkawa, T. and Kita, J. 2005. Physiological effects of fishes in a high-CO2 world. Journal of Geophysical Research 110: 10.1029/2004JC002564, http://www.co2science.org/articles/V8/N42/B3.php Although this conclusion sounds dire indeed, it represents an egregious flight of the imagination in terms of what could realistically be expected to happen anytime in earth's future. Ishimatsu et al. report, for example, that "predicted future CO2 concentrations in the atmosphere are lower than the known lethal concentrations for fish," noting that "the expected peak value is about 1.4 torr [just under 1850 ppm] around the year 2300 according to Caldeira and Wickett (2003)." So just how far below the lethal CO2 concentration for fish is 1.4 torr? In the case of short-term exposures on the order of a few days, the authors cite a number of studies that yield median lethal concentrations ranging from 37 to 50 torr, which values are 26 and 36 times greater than the maximum CO2 concentration expected some 300 years from now! No solvency-Past Tipping Point We’re passed the tipping point. AFP, Agence France Presse, September 15, 1999, “Outlook Grim For World’s Environment Says UN,” http://www.rense.com/earthchanges/grimoutlook_e.htm The United Nations warned Wednesday that the world’s environment was facing catastrophic damage as the new millennium nears, ranging from irreversible destruction to tropical rainforests to choking air pollution and a threat to the polar ice caps. In a lengthy report, the UN Environment Programme painted a grim tableau for the planet’s citizens in the next millennium, saying time was fast running out to devise a policy of sustainable human development. And for some fragile eco-systems and vulnerable species, it is already too late, warns the report, called GEO-2000. “Tropical forest destruction has gone too far to prevent irreversible damage. It would take many generations to replace the lost forests, and the cultures that have been lost with them can never be replaced,” it warns. “Many of the planet’s species have already been lost or condemned to extinction because of the slow response times of both the environment and policy-makers; it is too late to preserve all the bio-diversity the planet had.” Sounding the alarm, the UNEP said the planet now faced “full-scale emergencies” on several fronts, including these: -- it is probably too late to prevent global warming, a phenomenon whereby exhaust gases and other emissions will raise the temperature of the planet and wreak climate change. Indeed, many of the targets to reduce or stabilise emissions will not be met, the report says. -- urban air pollution problems are reaching “crisis dimensions” in the developing world’s mega-cities, inflicting damage to the health of their inhabitants. -- the seas are being “grossly over-exploited” and even with strenuous efforts will take a long time to recover. Too many alt causes to solve Pynn 07 – Journalist for the Vancouver Sun [Larry Pynn. “Global warming not biggest threat: expert.” The Vancouver Sun. January 27, 2007. http://www.canada.com/vancouversun/news/story.html?id=6e2988da-31ab-4697-810d7a008306d571] The biggest immediate threat to the vast majority of endangered species in Canada is not global warming, but issues such as habitat loss, pollution and overexploitation. "We all worry about climate change, as we should, but it doesn't mean we shouldn't worry about protecting habitat," says James Grant, a biology professor at Concordia University in Montreal and coauthor of a new report on threats to endangered species in Canada. "The really immediate causes right now for many species are things like farming, urbanization and habitat loss caused by the direct things we do." Research by Grant and his pupils shows the biggest threat is habitat loss at 84 percent, overexploitation 32 percent, native species interactions 31 percent, natural causes 27 percent, pollution 26 percent, and introduced species 22 percent. On average, species are threatened by at least two of the six categories. Human activities representing the biggest source of habitat loss and pollution are not industrial resource extraction, but agriculture at 46 per cent and urbanization at 44 per cent. "Farming is huge," Grant said in an interview. "The Prairies are one of the most affected habitats in the world. We've turned them into wheat fields."The southern Okanagan-Similkameen is another example, home to about one-third of species at risk in B.C. as well as a thriving agricultural industry, including vineyards, and increased urban development. No Solvency-Overpopulation swamps Population growth makes biodiversity loss inevitable Gaston 5 [Kevin J. Gaston Biodiversity and Macroecology Group, Department of Animal and Plant Sciences, University of Sheffield. Progress in Physical Geography 29, 2 (2005) pp. 239–247. “Biodiversity and extinction: species and people” http://www.epa.gov/ncer/biodiversity/pubs/ppg_vol29_239.pdf //jweideman] The human population is predicted to grow by 2 to 4 billion people by 2050 (United Nations, 2001). While it took until about 1800 to attain a global population of I billion people, a medium projection is that it may take just 13 to 14 years to add another billion to the pres- ent total (Cohen, 2003). All else remaining equal, which it seldom does, a number of pre- dictions would seem to follow from the work that has been conducted to date on the rela- tionships between human densities and the numbers of native species, numbers or pro- portions of threatened species, and the numbers or proportions of introduced species. First, the spatial scale at which relationships between overall numbers of native species and human density become hump-shaped or at least gain marked negative phases seems likely to increase, even when species numbers and human density are mapped at a low spa- tial resolution. Increased human densification will mean that the maintenance and conservation of tracts of natural or semi-natural vegetation will become more difficult in areas of higher human density. Secondly, the numbers and proportions of threatened species in different areas will tend to increase. McKee et al. (2003) have used existing relationships between numbers of threatened species and numbers of people in different areas to predict the consequences for biodiversity of continued increases in the human population. They found that the num- ber of threatened bird and mammal species across 114 continental nations is expected to increase in the average nation by 7% by 2020 and 14% by 2050, on the basis of human pop- ulation growth alone. Such aggregate estimates provide no indication of precisely what this is likely to do for the overall propor- tion of species that are globally threatened with extinction, but these increases can only serve to increase the 12% of bird species and the 23% of mammals currently listed as such (lUCN, 2003). Likewise, the proportion of species that have become globally extinct will increase. No Impact-Tech Solves Tech solves the impact Stossel 2007 John, Investigative reporter for Fox news, Environmental Alarmists Have It Backwards http://www.realclearpolitics.com/articles/2007/04/how_about_economic_progress_da.html Watching the media coverage, you'd think that the earth was in imminent danger -- that human life itself was on the verge of extinction. Technology is fingered as the perp. Nothing could be further from the truth. John Semmens of Arizona's Laissez Faire Institute points out that Earth Day misses an important point. In the April issue of The Freeman magazine, Semmens says the environmental movement overlooks how hospitable the earth has become -- thanks to technology. "The environmental alarmists have it backwards. If anything imperils the earth it is ignorant obstruction of science and progress. ... That technology provides the best option for serving human wants and conserving the environment should be evident in the progress made in environmental improvement in the United States. Virtually every measure shows that pollution is headed downward and that nature is making a comeback." (Carbon dioxide excepted, if it is really a pollutant.) Semmens describes his visit to historic Lexington and Concord in Massachusetts, an area "lush with trees and greenery." It wasn't always that way. In 1775, the land was cleared so it could be farmed. Today, technology makes farmers so efficient that only a fraction of the land is needed to produce much more food. As a result, "Massachusetts farmland has been allowed to revert back to forest." Human ingenuity and technology not only raised living standards, but also restored environmental amenities. How about a day to celebrate that? No Impact-Acidification No ocean acidification problem – it’s natural Idso et al 2009 Sherwood, founder and former President of the Center for the Study of Carbon Dioxide and Global Change and currently serves as Chairman of the Center's board of directors, The Ocean Acidification Fiction Volume 12, Number 22: 3 June 2009 There is considerable current concern that the ongoing rise in the air's CO2 content is causing a significant drop in the pH of the world's oceans in response to their absorption of a large fraction of each year's anthropogenic CO2 emissions. It has been estimated, for example, that the globe's seawater has been acidified (actually made less basic) by about 0.1 pH unit relative to what it was in pre-industrial times; and model calculations imply an additional 0.7-unit drop by the year 2300 (Caldeira and Wickett, 2003), which decline is hypothesized to cause great harm to calcifying marine life such as corals. But just how valid are these claims? Whenever the results of theoretical calculations are proposed as the basis for a crisis of some kind or other, it is always good to compare their predictions against what is known about the phenomenon in the real world. In the case of oceanic pH, for example, Liu et al. (2009) write in an important new paper that "the history of ocean pH variation during the current interglacial (Holocene) remains largely unknown," and that it "would provide critical insights on the possible impact of acidification on marine ecosystems." Hence, they set about to provide just such a context. Working with eighteen samples of fossil and modern Porites corals recovered from the South China Sea, the nine researchers employed 14C dating using the liquid scintillation counting method, along with positive thermal ionization mass spectrometry to generate high precision δ11B (boron) data, from which they reconstructed the paleo-pH record of the past 7000 years that is depicted in the figure below. As can be seen from this figure, there is nothing unusual, unnatural or unprecedented about the two most recent pH values. They are neither the lowest of the record, nor is the decline rate that led to them the greatest of the record. Hence, there is no compelling reason to believe they were influenced in any way by the nearly 40% increase in the air's CO2 concentration that has occurred to date over the course of the Industrial Revolution. As for the prior portion of the record, Liu et al. note that there is also "no correlation between the atmospheric CO2 concentration record from Antarctica ice cores and δ11B-reconstructed paleo-pH over the mid-late Holocene up to the Industrial Revolution." Further enlightenment comes from the earlier work of Pelejero et al. (2005), who developed a more refined history of seawater pH spanning the period 1708-1988 (depicted in the figure below), based on δ11B data obtained from a massive Porites coral from Flinders Reef in the western Coral Sea of the southwestern Pacific. These researchers also found that "there is no notable trend toward lower δ11B values." Instead, they discovered that "the dominant feature of the coral δ11B record is a clear interdecadal oscillation of pH, with δ11B values ranging between 23 and 25 per mil (7.9 and 8.2 pH units)," which they say "is synchronous with the Interdecadal Pacific Oscillation." Going one step further, Pelejero et al. also compared their results with coral extension and calcification rates obtained by Lough and Barnes (1997) over the same 1708-1988 time period; and as best we can determine from their graphical representations of these two coral growth parameters, extension rates over the last 50 years of this period were about 12% greater than they were over the first 50 years, while calcification rates were approximately 13% greater over the last 50 years. Most recently, Wei et al. (2009) derived the pH history of Arlington Reef (off the north-east coast of Australia) that is depicted in the figure below. As can be seen there, there was a ten-year pH minimum centered at about 1935 (which obviously was not CO2-induced) and a shorter more variable minimum at the end of the record (which also was not CO2-induced); and apart from these two non-CO2-related exceptions, the majority of the data once again fall within a band that exhibits no longterm trend, such as would be expected to have occurred if the gradual increase in atmospheric CO2 concentration since the inception of the Industrial Revolution were truly making the global ocean less basic. In light of these several diverse and independent assessments of the two major aspects of the ocean acidification hypothesis -- a CO2-induced decline in oceanic pH that leads to a concomitant decrease in coral growth rate -- it would appear that the catastrophe conjured up by the world's climate alarmists is but a wonderful work of fiction. Seagrasses solve the impact – acidification refuge. CO2 Science 2013 Seagrasses Enable Nearby Corals to Withstand Ocean Acidification, v16 n10 http://www.co2science.org/articles/V16/N10/B2.php Background The authors state that although many people expect future ocean acidification (OA) due to rising atmospheric CO2 concentrations to reduce the calcification rates of marine organisms, they say we have little understanding of how OA will manifest itself within dynamic, real-world systems, because, as they correctly note, "natural CO2, alkalinity, and salinity gradients can significantly alter local carbonate chemistry, and thereby create a range of susceptibility for different ecosystems to OA." What was done "To determine if photosynthetic CO2 uptake associated with seagrass beds has the potential to create OA refugia," as they describe it, Manzello et al. repeatedly measured carbonate chemistry across an inshore-to-offshore gradient in the upper, middle and lower Florida Reef Tract over a two-year period. What was learned During times of heightened oceanic vegetative productivity, the five U.S. researchers found "there is a net uptake of total CO2 which increases aragonite saturation state (Ωarag) values on inshore patch reefs of the upper Florida Reef Tract," and they say that "these waters can exhibit greater Ωarag than what has been modeled for the tropical surface ocean during preindustrial times, with mean Ωarag values in spring equaling 4.69 ± 0.10." At the same time, however, they report that Ωarag values on offshore reefs "generally represent oceanic carbonate chemistries consistent with present day tropical surface ocean conditions." What it means Manzello et al. hypothesize that the pattern described above "is caused by the photosynthetic uptake of total CO2 mainly by seagrasses and, to a lesser extent, macroalgae in the inshore waters of the Florida Reef Tract." And they therefore conclude that these inshore reef habitats are "potential acidification refugia that are defined not only in a spatial sense, but also in time, coinciding with seasonal productivity dynamics," which further implies that "coral reefs located within or immediately downstream of seagrass beds may find refuge from ocean acidification." And in further support of this conclusion, they cite the work of Palacios and Zimmerman (2007), which they describe as indicating that "seagrasses exposed to high-CO2 conditions for one year had increased reproduction, rhizome biomass, and vegetative growth of new shoots, which could represent a potential positive feedback to their ability to serve as ocean acidification refugia." Adaptation will solve acidification Knappenberger 10 (Chip Knappenberger is the assistant director of the Center for the Study of Science at the Cato Institute, and coordinates the scientific and outreach activities for the Center. He has over 20 years of experience in climate research and public outreach. January 6, 2010. Ocean Acidification: Another Failing Scare Story from Master Resource. http://www.masterresource.org/2010/01/ocean-acidification-another-failing-scare-story/ July 6, 2014.) The folks over at the Science and Public Policy Institute have taken it upon themselves to look a bit more closely into the ocean acidification story and see just what the scientific literature has to say about how organisms are actually responding to changes in ocean pH rather than just talk about how they may respond.¶ What SPPI finds is neither newsworthy nor catastrophic, simply that, as has always been the case, the aquatic organisms of the world’s oceans are very adaptable and are able to respond to environmental changes in such a way as to avoid large-scale detrimental impacts.¶ - Rising CO2 can’t cause acidification – 4 warrants and history proves SPPI 10 (The Science and Public Policy Institute provide research and educational materials dedicated to sound public policy based on sound science. Only through science and factual information, separating reality from rhetoric, can legislators develop beneficial policies without unintended consequences that might threaten the life, liberty, and prosperity of the citizenry. (SPPI). 01/05/10. A New Propaganda Film by Natl. Resources Defense Council Fails the Acid Test of Real World Data from SPPI. http://scienceandpublicpolicy.org/originals/acid_test.html July 6, 2014.) First, because it has not done so before. During the Cambrian era, 550 million years ago, there ¶ was 20 times as much CO2 in the atmosphere as there is today: yet that is when the calcite ¶ corals first achieved algal symbiosis. During the Jurassic era, 175 million years ago, there was ¶ again 20 times as much CO2 as there is today: yet that is when the delicate aragonite corals first ¶ came into being. ¶ ¶ Secondly, ocean acidification, as a notion, suffers from the same problem of scale as “global ¶ warming”. Just as the doubling of CO2 concentration expected this century will scarcely change ¶ global mean surface temperature because there is so little CO2 in the atmosphere in the first ¶ place, so it will scarcely change the acid-base balance of the ocean, because there is already 70 ¶ times as much CO2 in solution in the oceans as there is in the atmosphere. Even if all of the ¶ additional CO2 we emit were to end up not in the atmosphere (where it might in theory cause a 4 ¶ ¶ very little warming) but in the ocean (where it would cause none), the quantity of CO2 in the ¶ oceans would rise by little more than 1%, a trivial and entirely harmless change. ¶ ¶ Thirdly, to imagine that CO2 causes “ocean acidification” is to ignore the elementary chemistry ¶ of bicarbonate ions. Quantitatively, CO2 is only the seventh-largest of the substances in the ¶ oceans that could in theory alter the acid-base balance, so that in any event its effect on that ¶ balance would be minuscule. Qualitatively, however, CO2 is different from all the other ¶ substances in that it acts as the buffering mechanism for all of them, so that it does not itself ¶ alter the acid-base balance of the oceans at all. ¶ ¶ Fourthly, as Professor Ian Plimer points out in his excellent book Heaven and Earth (Quartet, ¶ London, 2009), the oceans slosh around over vast acreages of rock, and rocks are pronouncedly ¶ alkaline. Seen in a geological perspective, therefore, acidification of the oceans is impossible. SQ data solves acidification Status quo efforts solve ocean acid data OAP 12 [NOAA ocean acidification program. March 12, DATA COLLECTION AND MANAGEMENT http://oceanacidification.noaa.gov/AreasofFocus/DataCollectionandManagement.aspx//jweideman] OAP scientists collect a variety of data to understand changing ocean chemistry and its impacts on marine organisms and ecosystems. The National Oceanographic Data Center (NODC) serves as the NOAA Ocean Acidification data management focal point through its Ocean Acidification Data Stewardship (OADS) project. OA data will be archived at and available from an ocean acidification data stewardship system at NODC. Ocean acidification data can generally be classified as either physio-chemical or biological. Physio-chemical parameters include, among others, pCO2 (measurement of carbon dioxide gas both in the air and dissolved in seawater), pH, total alkalinity, total inorganic carbon, temperature, salinity, dissolved oxygen and current speed. Physio-chemical data from the field are collected either by remote observing methods (buoys, gliders) or through direct measurement from ships (hydrographic cruises or volunteer observing ships). Biological data from the field can be collected in similar ways, either by remote collection techniques or through direct measurement by scientists. For example, data about primary production (photosynthesizing activity) can be collected from buoys through measurement of chlorophyll a, nutrient levels and oxygen. Biologists have many techniques for collecting biological data in the field, both from ships and on shorelines. These collections can be instantaneous or from gear placed to collect organisms over time for later retrieval and analysis. During laboratory experiments on marine organisms, scientists can measure calcification rates, shell growth, behavior, otolith growth (for fish), metabolic rate, reproduction, among others parameters. Original datasets from all OAP research and the papers which analyze the data will be available through the OA Program and NODC. The National Oceanographic Data Center (NODC) is serving as the NOAA OA data management focal point by providing online data discovery, access to NODC-hosted and distributed data sources, and long-term archival for a diverse range of OA data. The OAP and NODC will build a collaborative relationship with shared responsibilities among scientists, data managers, and NODC staff towards the implementation of an OA data stewardship system (OADS). CURRENT EFFORTS In March 2012, the Ocean Acidification Program in collaboration with the University of Washington held an ocean acidification data management workshop in Seattle, WA which brought together a wide array of scientists and data managers from across US. Given that successful and integrated OA data management requires looking beyond just the NOAA program, we convened researchers and program managers from across the spectrum of US funded OA research. Representatives from NOAA, NSF, DOE, USGS and NASA attended as did academic researchers from a range of institutions. The workshop generated both a Declaration of Data Interdependence and a longer Draft Integrated Data Management Plan. No Impact-Coral Reefs Coral Reef’s Resilient as long as there is some Biodiversity—how much doesn’t matter Grimsditch and Salm 06 (Gabriel D Grimsditch holds the post of Programme Officer for oceans and climate change for the UNEP Marine and Coastal Ecosystems Branch in Nairobi, Kenya. Before joining UNEP, Gabriel worked for the IUCN Global Marine Programme where he coordinated the IUCN Climate Change and Coral Reefs Working Group. Rod Salm has his dissertation in marine ecology. He has 35 years experience in international marine conservation and ecotourism. 2006. Coral Reef Resilience and Resistance to Bleaching from The World Conservation Union. http://icriforum.org/sites/default/files/2006-042.pdf July 6, 2014) The main ecological factor that affects coral reef ¶ resilience to bleaching is a balanced biological and ¶ functional diversity (see Glossary) within the coral ¶ reef. It is essential to have a balanced ecological ¶ community with sufficient species interactions ¶ for coral reefs to recover from disturbances, and ¶ this applies not only to bleaching but to other ¶ disturbances as well (Nyström and Folke, 2001).¶ An especially important functional group (see ¶ Glossary) for coral reef resilience is grazing animals, comprising herbivorous fish and sea¶ urchins, among others. They enhance coral reef¶ resilience by preventing phase shifts from coral-¶ dominated reefs to algal-dominated reefs by ¶ keeping algal growth in check and allowing the¶ settlement of slower-growing coral recruits rather¶ than faster-growing algae. Their importance is¶ highlighted in a classic example from Jamaica, ¶ where the overfishing of predators and competitors ¶ (herbivorous fish) of the black-spined sea ¶ urchin Diadema antillarum led to an explosion ¶ in its population, and thus to a reduction of the ¶ diversity within the herbivorous functional group. ¶ Consequently, grazing by D. antillarum became ¶ the primary mechanism for algal control and crucial ¶ to coral reef recovery after Hurricane Allen in 1981. ¶ However, in 1983-1984, a pathogen killed off 95-¶ 99% of its population and enabled a phase shift ¶ as macroalgae out-competed coral. The extent to ¶ which this phase shift is irreversible is still unclear ¶ (Nyström et al, 2000). Adaptation and Resilience solve the coral reef scenario Griffin 13 (Catherine Griffin is an environmental and science journalist. 10/30/13. Coral Reefs Resist Climate Change with Genetic Adaptations from Science World Report. http://www.scienceworldreport.com/articles/10579/20131030/coral-reefs-resist-climate-change-genetic-adaptations.htm July 9, 2014) As our climate changes, coral reefs are becoming increasingly vulnerable to degradation. Now, though, scientists have discovered that corals may be more resilient than they thought. It turns out that coral reefs may be able to adapt to moderate climate warming, spelling new hope for those that hope to conserve these remarkable ecosystems.¶ Coral reefs are some of the most biodiverse areas in our oceans. They host thousands of species and act as nurseries for larger fish. Yet as temperatures warm, corals suffer from a process known as coral "bleaching." This occurs when reef-building corals eject algae living inside their tissues. Since these algae supply the coral with most of its food, bleaching can be fatal and leave vast swathes of a reef lifeless after one of these events.¶ Fortunately, though, it seems that the corals are adaptive. Scientists have discovered that corals possess a range of adaptations that can actually counteract the effects of warming. In order to come to these conclusions, though, the scientists used global sea surface temperature output from the NOAA/GFDL Earth System Model-2 for the pre-industrial period through 2100. This allowed them to project rates of coral bleaching. The initial results seemed to show that past temperature increases should have bleached reefs more often. The fact that they didn't seemed to indicate that corals are adapting. Overfishing Advantage Defense 1NC Regs Can’t Solve Literally no new regulations can solve overfishing- this card wrecks the advantage Bratspies 09 (Rebecca M. Bratspies, Professor of Law at the CUNY School of Law in New York, has taught environmental law, Natural Resources Law, International Environmental Law, Property, Administrative Law, and Lawyering, J.D. cum laude from the University of Pennsylvania Law School and a B.A. in Biology from Wesleyan University “Why the Free Market Can't Fix Overfishing”, http://www.theatlantic.com/health/archive/2009/08/why-the-free-market-cant-fix-overfishing/22524/, 8/9/09, SM) Catch shares" are the latest fix-all solution to the world's overfishing crisis, and that's too bad. The idea was recently promoted by Gregg Easterbrook here in The Atlantic, and NOAA Administrator Jane Lubchenco has committed her agency to "transitioning to catch shares as a solution to overfishing. Although it's tempting to think that property rights will transform fishing into an industry focused on long-term sustainability, catch-shares are really just a retread of the same "markets will fix everything" thinking that has been thoroughly discredited. Catch-shares allocate portions of the total catch within a fishery and give fishers a property-like right to a share of the fish. (These permits are freely tradable on the open market.) The alternative is open-access, in which anyone may catch fish until the total quota for the fishery is reached. Not surprisingly, open-access fisheries often involve a mad scramble to capture as large a share of fish as quickly as possible, the so-called "fisherman's dilemma." Catch-shares remove the incentive to catch all the fish immediately, but the real management challenge is deciding how many fish should be caught in the first place. Regardless of whether a fishery is open-access or allocates catch-shares, it is only sustainable if the total level of fishing is low enough to allow fish populations to regenerate. Unfortunately, total catch limits are often short-sighted. While setting the total allowable catch is theoretically a scientific decision, fishery managers face intense pressure to interpret every ambiguity in favor of allowing more, rather than less fishing. Sustainable levels of fishing fluctuate with environmental conditions: one year it might be fine to catch 100 tons of fish, while the next year even 10 tons might be too much. But the fishing industry depends on predictable levels of catch. Uncertainties inherent to estimating the "right" level of fishing makes it hard for managers to defend decisions to reduce fishing. That brings us to the real problem with fisheries: overcapacity. There are simply too many boats chasing too few fish . The catch-share approach tries to solve this problem by creating a permit thinking is that the permits will consolidate in the hands of "rational and efficient" producers who will voluntarily forego using a portion of their shares. That won't happen. The recent financial crisis ought to give anyone pause about the claim that markets inherently promote choices that are in everyone's long-term best interest . Regulators have sharply defined territorial jurisdictions, but fish trading market. The do not cooperate by staying in one place . Fish move between countries' Exclusive Economic Zones (waters under the effective control of a coastal state) and state's rigorously under the set catch limits the high seas. Boats on the high seas can undermine (Spain and Canada almost went to war over this in the 1990s). control of a single state, most a coastal Even when a fishery is governments don't have the capacity to make sure each boat takes only its allotted catch-share. Catch-shares also fail to address bycatch , the dirty little secret of the fishing industry. By most estimates, at least 40 percent of every catch is discarded as bycatch--fish other than the target species, including at least half a million endangered marine mammals and an unknown number of endangered sea turtles. Catch-shares will exacerbate this problem by creating a powerful incentive for fishing boats to discard not only unwanted or uncommercial fish, but also any fish potentially subject to someone else's share. Moreover, privatizing the ocean through catch-shares has troubling social justice implications. Catch-shares are typically awarded to boat owners on the basis of fishing capacity, rather than to each fisher on a per capita basis. Those with big boats benefit richly, but their workers are shut out of any ownership stake. This is the same "give to the rich" dynamic that distorts privatization schemes around the world. 2NC Can’t solve Regulations on overfishing are ineffective Boon 08 (Kristen E. Boon, Director of International Programs, specializes in public international law and international organization, Doctorate in law from Columbia Law School, a J.D. from New York University School of Law, M.A. in Political Science from McGill University “OVERFISHING OF BLUEFIN TUNA: INCENTIVIZING INCLUSIVE SOLUTIONS” http://www.louisvillelawreview.org/sites/louisvillelawreview.org/files/pdfs/printcontent/52/1/Boon.pdf, 2008) While my proposals would not preclude a property rights approach, this Article proceeded on the assumption that introducing property rights on the high seas would not be sufficient to address problems of scarcity or to inform institutional design. A final word justifying this presumption is in order. Advocates of property rights suggest that they would serve as a corrective force for the negative impacts of overfishing. Essentially, a property rights solution would involve the creation of a market for individual transferable quotas (ITQs). After a TAC has been set, either an RFMO like ICCAT could auction off quotas to interested rights holders, including corporations and states, or rights could be granted to individual fishermen directly who could then swap and sell their rights to others. Property rights to natural resources have never been introduced on the high seas, largely due to the difficulties of regulating species outside of national jurisdictions. When fisheries fall within one nation's enforcement jurisdiction, a public authority can institute a system of property rights based on catch or fishing effort. Thus a state can create, allocate, and enforce the allocation of property rights to fish on the basis of ITQs. This might involve setting a limit on the amount of fish that could be caught, and then auctioning off the fishing rights to those who want to purchase them. On the high seas, however, there is much more competition for the resource, and the role of property rights becomes considerably more complex. Moreover, the costs of implementing, maintaining, and enforcing a property rights solution becomes higher as the competition for resources increases. Can’t solve overfishing- Other nations won’t co-operate McDermott 10 (Mat McDermott, Writer about resource consumption for Treehugger, Masters from New York University’s Center for global affairs on environmental and energy policy “How Bad Is Overfishing & What Can We Do To Stop It?”, http://www.treehugger.com/greenfood/how-bad-is-overfishing-what-can-we-do-to-stop-it.html, August 16th 2010, Date Accessed: 7/3/14, SM) Enforcement is Key Though... The crucial variable in all this is enforcement. Though nations can control territorial waters--if they choose to, Libya for example actively turns a blind eye to overfishing in its waters, as do a number of other nations--in the open ocean things become much more difficult. Even with scientifically determined quotas, marine reserves and catch share systems in place, if the rules are not enforced, it all falls apart. Going back to the bluefin example, if the quota is 15,000 tons a year but the actual take, due to utter lack of enforcement, is 60,000 tons, you simply can't manage the fishery well. And when, as is the case with certain fisheries, you have organized crime syndicates involved, it all gets that much more complicated. Cant solve overfishing- Other countries won’t cooperate The House Of Ocean 13 (The House of ocean, “Big fishing nations that won’t stop overfishing”, http://houseofocean.org/2013/12/13/big-fishing-nations-that-wont-stop-overfishing/, 12/13/13 SM) A recent Guardian article exposes some of the figures behind industrial tuna fishing in the pacific. The article says that the US, China, South Korea, Japan, Indonesia and Taiwan are responsible for 80% of bigeye tuna caught each year. The remaining 20% is captured by vessels flagged to smaller fishing nations. Some of the smallest nations depend on their fisheries for basic survival. In 2012, 2.6m tonnes of tuna were extracted from the Pacific – 60% of the global total. Scientists are in agreement that tuna is being overfished at an alarming rate. Some species are practically on the brink, with bluefin tuna populations being currently just 4% of what they were before industrial fishing commenced. Yet, the organisation that has been entrusted by the international community to be the steward of tuna fisheries in the Pacific ocean, the Western and Central Pacific Fisheries Commission, has failed to protect the fish for yet another year. In spite of clear scientific advice regarding the need to reduce tuna quotas, the large fishing nations that currently haul the most, have point-blank refused to reduce their quota . Small Pacific nations have pointlessly warned of the consequences of overfishing – the big boys won’t budge. Unilateral fishing policies fail Schlick 09 (Katharina Schlick, “Fishery agreements in view of the South China Sea Conflict: A regional cooperative approach to maritime resource conflicts and territorial disputes”, http://www.victoria.ac.nz/chinaresearchcentre/publications/papers/19-katharina-schlick.pdf, 2009 SM) Overexploitation of marine living resources is a problem of both science and governance (Zha 2001: 577). As the example of Chinese domestic efforts in fisheries management show, many unilateral measures stay inefficient in view of the South China Sea ecosystem as a whole. On the one hand institutional weakness and poor implementation of regulations at the domestic level impairs any regional effort for sustainable resource management, on the other hand, diverging national, regional and international interests lead to different objectives and outcomes. The controversies over jurisdictional boundaries in many parts of the South China Sea and consequentially the absence of property rights over maritime living resources has created the image of an open access resource pool. In this situation, no one takes efforts to preserve the resource at a sustainable level as others will free-ride and enjoy the benefits from the resources at the expense of the others that put restrictions on their own use of the resource. The consequence is that the resource will gradually be overexploited and in view of fisheries, face danger of extinction. Given the migrator nature of many species of the sea, no single country would be able to manage or conserve these fish stocks (Wang 2001: 539-540). Conservation and management issues within the territorial waters and the high sea areas must therefore be compatible with each other. Nutall Goes neg- Can’t solve overfishing Nutall 06 (Nick Nuttall, Head of Media Services, United Nations Environment Program, “Overfishing: a threat to marine biodiversity”, http://www.un.org/events/tenstories/06/story.asp?storyID=800, 2006 ,SM) According to a Food and Agriculture Organization (FAO) estimate, over 70% of the world’s fish species are either fully exploited or depleted. The dramatic increase of destructive fishing techniques worldwide destroys marine mammals and entire ecosystems. FAO reports that illegal, unreported and unregulated fishing worldwide appears to be increasing as fishermen seek to avoid stricter rules in many places in response to shrinking catches and declining fish stocks. Few, if any, developing countries and only a limited number of developed ones are on track to put into effect by this year the International Plan of Action to Prevent, Deter and Eliminate Unreported and Unregulated Fishing. Despite that fact that each region has its Regional Sea Conventions, and some 108 governments and the European Commission have adopted the UNEP Global Programme of Action for the Protection of the Marine Environment from Land based Activities, oceans are cleared at twice the rate of forests. AT: Food prices I/L Food prices ARE increasing but because of biofuels Wahlberg 08 (Katarina Wahlberg, Advisor on Social and Economic Policy Program Coordinator, M.A. degree in political science from the University of Stockholm, “Between Soaring Food Prices and Food Aid Shortage” https://www.globalpolicy.org/component/content/article/217/46194.html, March 3rd 2008, SM) Food prices have soared because agricultural production has not kept up with the rising demand of cereals for food consumption, cattle feeding and biofuel production. For the first time in decades, worldwide scarcity of food is becoming a problem. Global cereal stocks are falling rapidly. Some predict that US wheat stocks will reach a 60-year low in 2008. Population growth in poor countries is boosting the grain demand for food consumption. But cereal demand for the feeding of cattle is increasing even more rapidly as consumers in both rich countries and fast growing economies are eating more dairy and meat. The most important factor behind the sudden spike in food prices, however, is the rapidly growing demand for biofuels, particularly in the EU and the US. Of total corn production, 12% is used to make biofuel, and that share is growing fast. Concerns about global climate change and soaring energy prices have boosted demand for biofuels. Until recently, few voices critical of biofuels were heard, but now an increasing number of policy makers and analysts strongly oppose converting food into fuel. In addition to directly threatening food security, there are alarming examples of how biofuel production causes environmental harm and speeds up global warming. US ethanol production uses large amounts of fuel, fertilizer, pesticides and water and most analysts consider its environmental impact quite negative. And in Indonesia, Malaysia and Brazil, companies have slashed thousands of hectares of rain forests to cultivate palm oil or sugarcane for biofuel production. According to the Food and Agricultural Organization in 2007. But, prices will remain high or grow even further in the coming years, as production is not growing fast enough to keep up with rising demand. Production is increasing mainly in the US, the EU, China and India. Not counting China and India, cereal production in poor countries decreased, due in part to climate change related emergencies such as droughts and floods. In addition to a tight supply and demand situation, soaring petroleum prices contribute to higher food prices by raising costs of transportation, fertilizers, and fuel for farm machinery. Moreover, financial investors speculating in commodity prices (FAO), world production of cereal, vegetables, fruit, meat and dairy increased aggravate prices and increase volatility in the market. Impact Defense Impact Defense Squo solves overfishing NOAA 11 (National Oceanic and Atmospheric Administration, “The Road to End Overfishing: 35 Years of Magnuson Act”, http://www.nmfs.noaa.gov/stories/2011/20110411roadendoverfishing.htm, 2011) I want to acknowledge and highlight the 35 th anniversary of the Magnuson-Stevens Fishery Conservation and Management Act. Simply called “the Magnuson Act”, this law, its regional framework and goal of sustainability, has proven to be a visionary force in natural resource management - both domestically and internationally. The Magnuson Act is, and will continue to be a key driver for NOAA as we deliver on our nation’s commitment to ocean stewardship, sustainable fisheries, and healthy marine ecosystems Because of the Magnuson Act, the U.S. is on track to end overfishing in federally-managed fisheries, rebuild stocks, and ensure conservation and sustainable use of our ocean resources. Fisheries harvested in the United States are scientifically monitored, regionally managed and legally enforced under 10 strict national standards of sustainability. This anniversary year marks a critical turning point in the Act’s history. By the end of 2011, we are on track to have an annual catch limit and accountability measures in place for all 528 federally-managed fish stocks and complexes. The dynamic, science-based management process envisioned by Congress is now in place, the rebuilding of our fisheries is underway, and we are beginning to see real benefits for fishermen, fishing communities and our commercial and recreational fishing industries. Squo solves overfishing- many species are recovering The Economist 08 (The economist, “Grabbing it all”, http://www.economist.com/node/12798494, December 30 th 2008) A variety of remedies have been tried, usually in combination. Thus regulations have been issued about the size and type of fish to be caught, the mesh of nets to be used, the number of days a month that boats may go to sea, the permissible weight of their catch and so on. In some countries fishermen are offered inducements to give up fishing altogether. Those that continue are, at least in theory, subject to monitoring both at sea and in port. Large areas are sometimes closed to fishing, to allow stocks to recover. Others have been designated as marine reserves akin to national parks. And some of the technology that fishermen use to find their prey is now used by inspectors to monitor the whereabouts of the hunters themselves. Most of these measures have helped, as the recovery of stocks in various places has shown. Striped bass and North Atlantic swordfish have returned along America's East Coast, for instance. Halibut have made a comeback in Alaska. Haddock, if not cod, have begun to recover in Georges Bank off Maine. And herring come and go off the coasts of Scotland. Those who doubt the value of government intervention have only to look at the waters off Somalia, a country that has been devoid of any government worth the name since 1991. The ensuing free-for-all has devastated the coastal stocks, ruining the livelihoods of local fishermen and encouraging them, it seems, to take up piracy instead. Climate change proves marine bioD are resilient Taylor 10 (James M. Taylor is a senior fellow of The Heartland Institute and managing editor of Environment and Climate News, “Ocean Acidification Scare Pushed at Copenhagen,” http://www.heartland.org/publications/environment%20climate/article/26815/Ocean_Acidification_Scare_Pushed_at_Copenhagen.html, Feb 10) With global temperatures continuing their decade-long decline and United Nationssponsored global warming talks falling apart in Copenhagen, alarmists at the U.N. talks spent considerable time claiming carbon dioxide emissions will cause catastrophic ocean acidification, regardless of whether temperatures rise. The latest scientific data, however, show no such catastrophe is likely to occur. Food Supply Risk Claimed The United Kingdom’s environment secretary, Hilary Benn, initiated the Copenhagen ocean scare with a high-profile speech and numerous media interviews claiming ocean acidification threatens the world’s food supply. “The fact is our seas absorb CO2. They absorb about a quarter of the total that we produce, but it is making our seas more acidic,” said Benn in his speech. “If this continues as a problem, then it can affect the one billion people who depend on fish as their principle source of protein, and we have to feed another 2½ to 3 billion people over the next 40 to 50 years.” Benn’s claim of oceans becoming “more acidic” is misleading, however. Water with a pH of 7.0 is considered neutral. pH values lower than 7.0 are considered acidic, while those higher than 7.0 are considered alkaline. The world’s oceans have a pH of 8.1, making them alkaline, not acidic. Increasing carbon dioxide concentrations would make the oceans less alkaline but not acidic. Since human industrial activity first began emitting carbon dioxide into the atmosphere a little more than 200 years ago, the pH of the oceans has fallen merely 0.1, from 8.2 to 8.1. Following Benn’s December 14 speech and public relations efforts, most of the world’s major media outlets produced stories claiming ocean acidification is threatening the world’s marine life. An Associated Press headline, for example, went so far as to call ocean acidification the “evil twin” of climate change. Studies Show CO2 Benefits Numerous recent scientific studies show higher carbon dioxide levels in the world’s oceans have the same beneficial effect on marine life as higher levels of atmospheric carbon dioxide have on terrestrial plant life. In a 2005 study published in the Journal of Geophysical Research, scientists examined trends in chlorophyll concentrations, critical building blocks in the oceanic food chain. The French and American scientists reported “an overall increase of the world ocean average chlorophyll concentration by about 22 percent” during the prior two decades of increasing carbon dioxide concentrations. In a 2006 study published in Global Change Biology, scientists observed higher CO2 levels are correlated with better growth conditions for oceanic life. The highest CO2 concentrations produced “higher growth rates and biomass yields” than the lower CO2 conditions. Higher CO2 levels may well fuel “subsequent primary production, phytoplankton blooms, and sustaining oceanic food-webs,” the study concluded. Ocean Life ‘Surprisingly Resilient’ In a 2008 study published in Biogeosciences, scientists subjected marine organisms to varying concentrations of CO2, including abrupt changes of CO2 concentration. The ecosystems were “surprisingly resilient” to changes in atmospheric CO2, and “the ecosystem composition, bacterial and phytoplankton abundances and productivity, grazing rates and total grazer abundance and reproduction were not significantly affected by CO2induced effects.” In a 2009 study published in Proceedings of the National Academy of Sciences, scientists reported, “Sea star growth and feeding rates increased with water temperature from 5ºC to 21ºC. A doubling of current [CO2] also increased growth rates both with and without a concurrent temperature increase from 12ºC to 15ºC.” Another False CO2 Scare “Far too many predictions of CO2-induced catastrophes are treated by alarmists as sure to occur, when real-world observations show these doomsday scenarios to be highly unlikely or even virtual impossibilities ,” said Craig Idso, Ph.D., author of the 2009 book CO2, Global Warming and Coral Reefs. “The phenomenon of CO2-induced ocean acidification appears to be no different. Science Leadership Tech Leadership High U.S. manufacturing dominates the market, manufacturing innovation solves tech output NIST 2009 National Institute of Standards and Technology, measurement standards laboratory, government agency, 2009, “The Facts About Modern Manufacturing”, http://www.nist.gov/mep/upload/FINAL_NAM_REPORT_PAGES.pdf The 8th edition of the Facts gives a snapshot of the state of U.S. manufacturing, and exhibits its strengths and challenges. The Facts clearly show that manufacturing continues to play a vital role in the U.S. economy. This edition illustrates that the quantity of manufactured goods produced in the United States has kept pace with overall economic growth since 1947, as both GDP and manufacturing have grown by about seven times (Figure 2). The United States still has the largest manufacturing sector in the world, and its market share (around 20 percent) has held steady for 30 years (Figure 131. One in six private sector jobs is still in or directly tied to manufacturing (Figure 8). Moreover, productivity growth is higher in manufacturing than in other sectors of the economy (Figure 25). Due largely to outstanding productivity growth, the prices of manufactured goods have declined since 1995 in contrast to inflation in most other sectors, with the result that manufacturers are contributing to a higher standard of living for U.S. consumers (Figure 12). Manufacturing still pays premium wages and benefits, and supports much more economic activity per dollar of production than other sectors (Figure 9). Another major indicator of the importance of manufacturing to the strength of the economy is its key role in driving innovation and technology. These are crucial components of a productivity-driven, global competitiveness agenda, and also help explain the steady rise in our standard of living. U.S. manufacturing accounts for 35 percent of value added in all of the world's high technology production, and enjoys a trade surplus in revenues from royalties from production processes and technology. U.S. inventors still account for more than one- half of all patents granted in the United States (Figure 26), and the nation outpaces its rivals in terms of industrial research and development. Technology has aided U.S. manufacturers to use less energy per unit or dollar of production (Figure 30) and to lead all other sectors in reducing C02 emissions in the last two decades (Figure 29). Finally, the technology and advanced processes developed in manufacturing consistently spill over into productivity growth in the service and agriculture sectors. For this reason, it is important to consider innovation in terms of processes as well as technologies. Tech leadership inevitable – solves the advantage Zakaria 2009 Fareed Zakaria, editor of Foreign Affairs, editor of Newsweek International, host of CNN's Fareed Zakaria GPS, 11/14/09, Newsweek, "Can America Still Innovate?", http://www.newsweek.com/id/222836 Government funding of basic research has been astonishingly productive. Over the past five decades it has led to the development of the Internet, lasers, global positioning satellites, magnetic resonance imaging, DNA sequencing, and hundreds of other technologies. Even when government was not the inventor, it was often the facilitator. One example: semiconductors. As a study by the Breakthrough Institute notes, after the microchip was invented in 1958 by an engineer at Texas Instruments, "the federal government bought virtually every microchip firms could produce." This was particularly true of the Air Force, which needed chips to guide the new Minuteman II missiles, and NASA, which required advanced chips for the on-board guidance computers on its Saturn rockets. "NASA bought so many [microchips] that manufacturers were able to achieve huge improvements in the production process—so much so, in fact, that the price of the Apollo microchip fell from $1,000 per unit to between $20 and $30 per unit in the span of a couple years." No Impact-Soft Power Soft power doesn’t solve — everything turns it Quinn 2011 Adam Quinn, Lecturer in International Studies at the University of Birmingham, previously worked at the University of Leicester and the University of Westminster, focuses on the role of national history and ideology in shaping US grand strategy, “The art of declining politely: Obama’s prudent presidency and the waning of American power”, International Affairs 87:4 (2011) 803–824 http://www.chathamhouse.org/sites/default/files/87_4quinn.pdf Nevertheless, this qualification demands two further qualifications of its own. The first is that if we consider ‘soft power’ as a national attribute then it is difficult to separate it with confidence from the economic and military dimensions of power. Is it really likely that America’s ideological and cultural influence will endure undiminished in the absence of the platform of military and economic primacy upon which it has been constructed? It may be overstatement to suggest that, borrowing Marxist terminology, hard power represents the ‘base’ and soft power mere ‘superstructure’. But one could plausibly argue that even America’s non-coercive power and political appeal are inextricably entwined with the status conferred upon it by possession of a preponderance of material resources. While vestigial soft power may delay or mitigate the consequences of relative material decline, it is surely unrealistic to expect it to override them such as to allow the US to continue to exercise the same influence in a multipolar or non-polar world as it did in a unipolar one. No Impact-Hegemony Hegemony fails – policy has shifted from helping other nations to focusing on America’s domestic interests – increasing wealth, military power and influence don’t solve Kagan 5/26 Robert Kagan, PhD in American History, senior fellow at the Brookings Institution and a member of the Council on Foreign Relations, 5/26/14, “Superpowers don’t get to retire”, http://www.newrepublic.com/article/117859/allure-normalcy-what-america-still-owes-world Almost 70 years ago, a new world order was born from the rubble of World War II, built by and around the power of the United States. Today that world order shows signs of cracking, and perhaps even collapsing. The Russia-Ukraine and Syria crises, and the world’s tepid response, the general upheaval in the greater Middle East and North Africa, the growing nationalist and great-power tensions in East Asia, the worldwide advance of autocracy and retreat of democracy—taken individually, these problems are neither unprecedented nor unmanageable. But collectively they are a sign that something is changing, and perhaps more quickly than we may imagine. They may signal a transition into a different world order or into a world disorder of a kind not seen since the 1930s. If a breakdown in the world order that America made is occurring, it is not because America’s power is declining—America’s wealth, power, and potential influence remain adequate to meet the present challenges. It is not because the world has become more complex and intractable—the world has always been complex and intractable. And it is not simply war-weariness. Strangely enough, it is an intellectual problem, a question of identity and purpose. Many Americans and their political leaders in both parties, including President Obama, have either forgotten or rejected the assumptions that undergirded American foreign policy for the past seven decades. In particular, American foreign policy may be moving away from the sense of global responsibility that equated American interests with the interests of many others around the world and back toward the defense of narrower, more parochial national interests. This is sometimes called “isolationism,” but that is not the right word. It may be more correctly described as a search for normalcy. At the core of American unease is a desire to shed the unusual burdens of responsibility that previous generations of Americans took on in World War II and throughout the cold war and to return to being a more normal kind of nation, more attuned to its own needs and less to those of the wider world. If this is indeed what a majority of Americans seek today, then the current period of retrenchment will not be a temporary pause before an inevitable return to global activism. It will mark a new phase in the evolution of America’s foreign policy. And because America’s role in shaping the world order has been so unusually powerful and pervasive, it will also begin a new phase in the international system, one that promises not to be marginally different but radically different from what we have known these past 70 years. Unless Americans can be led back to an understanding of their enlightened self-interest, to see again how their fate is entangled with that of the world, then the prospects for a peaceful twenty-first century in which Americans and American principles can thrive will be bleak. S&T Add on Neg S&T policy is effective now – US commitments and they’re legally binding Dolan 2012 Bridget M. Dolan PhD, Research Scholar for the American Association for the Advancement of Science, "Science and Technology Agreements as Tools for Science Diplomacy: A U.S. Case Study," Science & Diplomacy, Vol. 1, No. 4 (December 2012). http://www.sciencediplomacy.org/article/2012/science-andtechnology-agreements-tools-for- science-diplomacy. A Formal and Legally Binding Agreement International agreements to promote cooperation in scientific research and development can be bilateral or multilateral, government-wide or at the level of individual technical agencies (e.g., the National Science Foundation or the National Institutes of Health). The focus of this paper is on bilateral, government-wide agreements, also referred to as umbrella agreements, framework agreements, or simply S&T agreements. Scientific cooperation between the United States and other countries is undertaken using a variety of arrangements, from informal scientist-to-scientist collaborations to cooperation between research institutions to formal agreements between technical agencies. While S&T agreements are not necessary for these types of interactions, other nations often seek S&T agreements with the United States because they carry the weight of being legally binding and having been negotiated on behalf of the U.S. government. These agreements endeavor to establish a framework to foster international science collaboration while protecting intellectual property, establishing benefit sharing, and preventing taxation of research equipment. The contents of an S&T agreement usually include common features such as types of cooperative activities and ways to encourage access to facilities and personnel, as well as clarification that some information or equipment—such as those requiring protection for national security reasons—are not covered under the agreement. There are three areas where the agreement text often varies: (1) the preamble, which is not legally binding and is often used to highlight the public motivations behind the agreement; (2) the intellectual property rights annex, which delineates how the parties share and exploit intellectual property generated; and (3) the implementation plan, including whether to establish a joint committee that would meet regularly to review execution of the agreement. They can’t solve – funding Brown and Sarewitz 1998 George Brown Jr., former House representative, Daniel Sarewitz, CoDirector, Consortium for Science, Policy & Outcomes Associate Director, Fall 1998, Science and Technology, Volume 15, Issue 1, “U.S. failure in international scientific cooperation”, http://www.freepatentsonline.com/article/Issues-in-Science-Technology/53435944.html In August 1991, we traveled to Mexico to meet with policymakers and scientists about the establishment of a United States-Mexico science foundation devoted to supporting joint research on problems of mutual interest. We encountered enthusiasm and vision at every level, including an informal commitment by the Minister of Finance to match any U.S. contribution up to $20 million. At about this time, our article "Fiscal Alchemy: Transforming Debt into Research" (Issues, Fall 1991) sought to highlight three issues: 1) the pressing need for scientific partnerships between the United States and industrializing nations, 2) the mechanism of bilateral or multilateral foundations for funding such partnerships, and 3) the device of debt swaps for allowing debtor nations with limited foreign currency reserves to act as full partners in joint research ventures. We returned from our visit to Mexico flush with optimism about moving forward on all three fronts. Results, overall, have been disappointing. We had hoped that the debt-for-science concept would be adopted by philanthropic organizations and universities as a way to leverage the most bang for the research buck. This has not taken place. The complexity of negotiating debt swaps and the changing dynamics of the international economy may be inhibiting factors. But much more significant, in our view, is a general unwillingness in this nation to pursue substantive international scientific cooperation with industrializing and developing nations. Although the National Science Foundation and other agencies do fund U.S. scientists conducting research in the industrializing and developing world, this work does not support broader partnerships aimed at shared goals. Such partnerships can foster the local technological capacities that underlie economic growth and environmental stewardship; we also view them as key to successfully addressing a range of mutual problems, including transborder pollution, emerging diseases, and global climate change. Yet there is a conspicuous lack of attention to this approach at all levels of the administration; most important, the State Department continues to view scientific cooperation as a question of nothing more than diplomatic process. Incredibly, through 1995 (the latest year for which data are available) the United States has negotiated more than 800 bilateral and multilateral science and technology agreements (up from 668 in 1991), even though virtually none of these are backed by funding commitments. Nor is there any coordination among agencies regarding goals, implementation, redundancy, or follow-up. A report by the RAND Corporation, "International Cooperation in Research and Development," found little correlation between international agreements and actual research projects. Moreover, although there are few indications that these agreements have led to significant scientific partnerships with industrializing and developing nations, there is plenty of evidence that they support a healthy bureaucratic infrastructure in the U.S. government. We cannot help but think that a portion of the funds devoted to negotiating new agreements and maintaining existing ones might be better spent on cooperative science. No warming and not anthropogenic Ferrara, 2012 (Peter, Director of Entitlement and Budget Policy for the Heartland Institute, Senior Advisor for Entitlement Reform and Budget Policy at the National Tax Limitation Foundation, General Counsel for the American Civil Rights Union, and Senior Fellow at the National Center for Policy Analysis, served in the White House Office of Policy Development, graduate of Harvard College and Harvard Law School , 5/31/2012, "Sorry Global Warming Alarmists, The Earth Is Cooling," http://www.forbes.com/sites/peterferrara/2012/05/31/sorry-globalwarming-alarmists-the-earth-is-cooling/) Climate change itself is already in the process of definitively rebutting climate alarmists who think human use of fossil fuels is causing ultimately catastrophic global warming. That is because natural climate cycles have already turned from warming to cooling, global temperatures have already been declining for more than 10 years , and global temperatures will continue to decline for another two decades or more. That is one of the most interesting conclusions to come out of the seventh International Climate Change Conference sponsored by the Heartland Institute, held last week in Chicago. I attended, and served as one of the speakers, talking about The Economic Implications of High Cost Energy. The conference featured serious natural science, contrary to the self-interested political science you hear from government financed global warming alarmists seeking to justify widely expanded regulatory and taxation powers for government bodies, or government body wannabees, such as the United Nations. See for yourself, as the conference speeches are online. What you will see are calm, dispassionate presentations by serious, pedigreed scientists discussing and explaining reams of data. In sharp contrast to these climate realists, the climate alarmists have long admitted that they cannot defend their theory that humans are causing catastrophic global warming in public debate. With the conference presentations online, let’s see if the alarmists really do have any response. The Heartland Institute has effectively become the international headquarters of the climate realists, an analog to the UN’s Intergovernmental Panel on Climate Change (IPCC). It has achieved that status through these international climate conferences, and the publication of its Climate Change Reconsidered volumes, produced in conjunction with the Nongovernmental International Panel on Climate Change (NIPCC). Those Climate Change Reconsidered volumes are an equivalently thorough scientific rebuttal to the irregular Assessment Reports of the UN’s IPCC. You can ask any advocate of human caused catastrophic global warming what their response is to Climate Change Reconsidered. If they have none, they are not qualified to discuss the issue intelligently. Check out the 20th century temperature record, and you will find that and down pattern does not follow the industrial revolution’s upward march of atmospheric carbon dioxide (CO2), which is the supposed central culprit for man caused global warming (and has been much, much higher in the past). It follows its up instead the up and down pattern of naturally caused climate cycles. For example, temperatures dropped steadily from the late 1940s to the late 1970s. The popular press was even talking about a coming ice age. Ice ages have cyclically occurred roughly every 10,000 years, with a new one actually due around now. In the late 1970s, the natural cycles turned warm and temperatures rose until the late 1990s, a trend that political and economic interests have tried to milk mercilessly to their advantage. The incorruptible satellite measured global atmospheric temperatures show less warming during this period than the heavily manipulated land surface temperatures. Central to these natural cycles is the Pacific Decadal Oscillation (PDO). Every 25 to 30 years the oceans undergo a natural cycle where the colder water below churns to replace the warmer water at the surface, and that affects global temperatures by the fractions of a degree we have seen. The PDO was cold from the late 1940s to the late 1970s, and it was warm from the late 1970s to the late 1990s, similar to the Atlantic Multidecadal Oscillation (AMO). In 2000, the UN’s IPCC predicted that global temperatures would rise by 1 degree Celsius by 2010. Was that based on climate science, or political science to scare the public into accepting costly anti-industrial regulations and taxes? Don Easterbrook, Professor Emeritus of Geology at Western Washington University, knew the answer. He publicly predicted in 2000 that global temperatures would decline by 2010. He made that prediction because he knew the PDO had turned cold in 1999, something the political scientists at the UN’s IPCC did not know or did not think significant. Well, the results are in, and the winner is….Don Easterbrook. Easterbrook also spoke at the Heartland conference, with a presentation entitled “Are Forecasts of a 20-Year Cooling Trend Credible?” Watch that online and you will see how scientists are supposed to talk: cool, rational, logical analysis of the data, and full explanation of it. All I ever see from the global warming alarmists, by contrast, is political public relations, personal attacks, ad hominem arguments, and name calling, combined with admissions that they can’t defend their views in public debate. Easterbrook shows that by 2010 the 2000 prediction of the IPCC was wrong by well over a degree , and the gap was widening . That’s a big miss for a forecast just 10 years away, when the same folks expect us to take seriously their predictions for 100 years in the future. Howard Hayden, Professor of Physics Emeritus at the University of Connecticut showed in his presentation at the conference that based on the historical record a doubling of CO2 could be expected to produce a 2 degree C temperature increase. Such a doubling would take most of this century, and the temperature impact of increased concentrations of CO2 declines logarithmically. You can see Hayden’s presentation online as well. Because PDO cycles last 25 to 30 years, Easterbrook expects the cooling trend to continue for another 2 decades or so. Easterbrook, in fact, documents 40 such alternating periods of warming and cooling over the past 500 years, with similar data going back 15,000 years. He further expects the flipping of the ADO to add to the current downward trend. But that is not all. We are also currently experiencing a surprisingly long period with very low sunspot activity. That is associated in the earth’s history with even lower, colder temperatures. The pattern was seen during a period known as the Dalton Minimum from 1790 to 1830, which saw temperature readings decline by 2 degrees in a 20 year period, and the noted Year Without A Summer in 1816 (which may have had other contributing short term causes). Even worse was the period known as the Maunder Minimum from 1645 to 1715, which saw only about 50 sunspots during one 30 year period within the cycle, compared to a typical 40,000 to 50,000 sunspots during such periods in modern times. The Maunder Minimum coincided with the coldest part of the Little Ice Age, which the earth suffered from about 1350 to 1850. The Maunder Minimum saw sharply reduced agricultural output, and widespread human suffering, disease and premature death. Such impacts of the sun on the earth’s climate were discussed at the conference by astrophysicist and geoscientist Willie Soon, Nir J. Shaviv, of the Racah Institute of Physics in the Hebrew University of Jerusalem, and Sebastian Luning, co-author with leading German environmentalist Fritz Vahrenholt of The Cold Sun. Easterbrook suggests that the outstanding question is only how cold this present cold cycle will get. Will it be modest like the cooling from the late 1940s to late 1970s? Or will the paucity of sunspots drive us all the way down to the Dalton Minimum, or even the Maunder Minimum? He says it is impossible to know now. But based on experience, he will probably know before the UN and its politicized IPCC. STEM Add on Neg U.S. manufacturing dominates the market, manufacturing innovation solves tech output NIST 2009 National Institute of Standards and Technology, measurement standards laboratory, government agency, 2009, “The Facts About Modern Manufacturing”, http://www.nist.gov/mep/upload/FINAL_NAM_REPORT_PAGES.pdf The 8th edition of the Facts gives a snapshot of the state of U.S. manufacturing, and exhibits its strengths and challenges. The Facts clearly show that manufacturing continues to play a vital role in the U.S. economy. This edition illustrates that the quantity of manufactured goods produced in the United States has kept pace with overall economic growth since 1947, as both GDP and manufacturing have grown by about seven times (Figure 2). The United States still has the largest manufacturing sector in the world, and its market share (around 20 percent) has held steady for 30 years (Figure 131. One in six private sector jobs is still in or directly tied to manufacturing (Figure 8). Moreover, productivity growth is higher in manufacturing than in other sectors of the economy (Figure 25). Due largely to outstanding productivity growth, the prices of manufactured goods have declined since 1995 in contrast to inflation in most other sectors, with the result that manufacturers are contributing to a higher standard of living for U.S. consumers (Figure 12). Manufacturing still pays premium wages and benefits, and supports much more economic activity per dollar of production than other sectors (Figure 9). Another major indicator of the importance of manufacturing to the strength of the economy is its key role in driving innovation and technology. These are crucial components of a productivity-driven, global competitiveness agenda, and also help explain the steady rise in our standard of living. U.S. manufacturing accounts for 35 percent of value added in all of the world's high technology production, and enjoys a trade surplus in revenues from royalties from production processes and technology. U.S. inventors still account for more than one- half of all patents granted in the United States (Figure 26), and the nation outpaces its rivals in terms of industrial research and development. Technology has aided U.S. manufacturers to use less energy per unit or dollar of production (Figure 30) and to lead all other sectors in reducing C02 emissions in the last two decades (Figure 29). Finally, the technology and advanced processes developed in manufacturing consistently spill over into productivity growth in the service and agriculture sectors. For this reason, it is important to consider innovation in terms of processes as well as technologies. Ocean policy is increasing now – status quo solves the advantage Babb-Brott 2013 Deerin Babb-Brott, former director of the National Ocean Council, 7/19/13, “A Guide for Regional Marine Planning”,http://www.whitehouse.gov/blog/2013/07/19/guide-regionalmarine-planning Today, the National Ocean Council released a Marine Planning Handbook to support the efforts of regions that choose to engage marine industries, stakeholders, the public, and government to advance their economic development and conservation priorities. Each coastal region of the country has its own interests and ways of doing business, but all regions want to support their marine economies and coastal communities. Voluntary marine planning is a science-based tool that provides regionally tailored information that all ocean interests can use to reduce conflicts, grow ocean industries, and support the healthy natural resources that our economy and communities depend on. Federal, state and local governments have a variety of roles and responsibilities when it comes to the ocean, and make decisions every day that impact ocean resources, industries and coastal communities. Regions that choose to do marine planning are guaranteeing that the public and marine stakeholders will shape these decisions early on, promoting better outcomes for everyone. Regions can define what they want to address and how they do so, in ways that reflect their unique interests and priorities. At the same time, some components of planning – like making sure the public and stakeholders have a chance to engage – are common to all regions. The Handbook provides guidance on how regions can address their priorities through a bottom- up, transparent, science-based process. The Handbook reflects the extensive public and stakeholder input received in the development of the National Ocean Policy and its Implementation Plan. We will update it as needed to reflect the lessons learned in regions and ensure it continues to be a useful guide for successful, collaborative planning. New programs don’t generate interest and STEM is inevitable Spudis 2011 Paul Spudis, Senior Staff Scientist at the Lunar and Planetary Institute, 2011, The Once and Future Moon Blog, Smithsonian Air and Space Blog, 5/14/11, “Young Visitors Inspire Old Scientist,” http://blogs.airspacemag.com/moon/2011/05/young-visitors-inspire-old-scientist/ A perennial hand-wringing topic among policy geeks is America’s decline in math and science proficiency. This sentiment has been expressed the entire 30 years I’ve worked on space science and exploration – new generations don’t care about space, can’t do math and science, can’t think properly and the country’s going to hell in a hand basket. Complaint about the decline in our ability is something passed from one generation to the next. Today’s youth are being corrupted by iPods, Facebook and hip-hop; when I was a kid, it was Frisbees, MAD magazine and the Beatles. There is a continuous stream of doom-laden reports outlining the decline of American youth in the fields of science, technology, engineering and math (called STEM). In this country, most Ph.D.s in science and technology are now foreign-born (these reports don’t mention that most often, they stay here, adding to our technical knowledge base). Multiple factors are suggested as contributors to this decline, with the lack of an inspiring, exciting space program long believed to be important by many advocates. This meme, of long currency in space policy circles, has some flaws. Origins of the association between space exploration and science education go back to the days of Sputnik –the “ping” that shocked and alarmed the country. This event prompted loud public cries for somebody to “do something” about the educational decline of America’s youth (corrupted then by ’57 Chevys, hula-hoops and Elvis). Congress responded in the usual manner – they threw money at the problem. The National Defense Education Act of 1958 (interesting wording that) created a huge infrastructure largely dependent upon federal funding that directly answered the Soviet challenge in space. The number of science graduates exploded over the next couple of decades, leading many to conclude that 1) the excitement generated by the Apollo program inspired these students to aspire to careers in science; 2) huge amounts of federal money can solve any problem. Although Apollo is now a distant memory (or for many, a distant, confluence of space and education is taken as a historical event that they’ve only read about in third-hand accounts), the given by many in the business. NASA spends a designated fraction of their budget on a process called “EPO” (Education and Public Outreach), designed to inform and inspire the next generation of scientists and explorers. As you might expect, these efforts range from the interesting and innovative to the embarrassing (though well intentioned). A perception has emerged that the problem lies not with the methodology, but with the product – because we are not doing anything in space that is “exciting,” we aren’t producing quality scientists and engineers. This may well account for what sensible students – with an eye toward putting food on the table after they graduate – choose to study. Then too, perhaps there are too many in the field already. But with effort, excellence will find productive work; self-esteem and entitlement will not cut it in the long run, no matter what your field of endeavor. Recently, I had the opportunity to directly interact with students at two ends of the education pipeline and found the experience highly encouraging. In the first case, the father of a local second-grader asked if his son could visit and interview me. The boy had chosen to write a semester paper (in second grade??) about the Moon. The child was both well spoken and well informed. He asked relevant and very intelligent questions. What is the value of the Moon? What do we want to know about it and how do we find these things out? Can people live there? I found his questions and understanding of the problems and benefits of exploring the Moon to be at a very high level (much higher than many adult reporters who call me). Then he asked me an unexpected question: How fast does the Moon travel through space? After initially drawing a complete blank, I suggested that we find out together and went on to calculate it on the spot. We concluded that the Moon flies around the Earth at over 2200 miles per hour (much faster than he traveled down the freeway to visit me). He was delighted by this episode of “science in action.” I was delighted to be challenged by his understanding and his interest in the topic. Around the same time, a high school debate coach contacted me. He told me that next year’s debate question is asked if I could assist his team, as they were collecting information for their briefing books. Once again, I was pleasantly surprised by their level of knowledge and their understanding of complex issues. We reviewed the history of the space program, and why and how the current policy confusion has developed. These students were informed and sharp. They had already read and digested a great deal of information – drawing insight and conclusions about issues the space community is embroiled in. Their questions were both penetrating and logical, and sent a clear message of their desire to fully understand the technical and programmatic issues involved. What did I conclude from my encounter with a sample of today’s youth? Mostly, that reports of the demise of our Republic are premature. These kids were smart and well informed. They could assimilate new information and apply it to other topics in clever ways. They had an enthusiasm for their subject that was both gratifying and surprising. And, they are interested in space, regardless of the current “uninspiring” nature of the program. Inspiration is great, but “Should humans explore space?” and it’s a highly personal factor and its impact and importance are difficult to measure. The current STEM/outreach process at NASA conflates excitement and inspiration, but they are two different things. The circus entertains us but we find inspiration elsewhere. We need to focus on building a stable space program that will give us long-term benefits – a step-wise, incremental program that gradually increases the extent of our reach into space. Compared to the current policy chaos, it just might be inspirational too. Growth is not a strong enough predictor of war to matter Blackwill 2009 Robert Blackwill, former associate dean of the Kennedy School of Government and Deputy Assistant to the President and Deputy National Security Advisor for Strategic Planning, “The Geopolitical Consequences of the World Economic Recession—A Caution”, http://www.rand.org/pubs/occasional_papers/OP275.html Did the economic slump lead to strategic amendments in the way Japan sees the world? No. Did it slow the pace of India’s emergence as a rising great power? No. To the contrary, the new Congress-led government in New Delhi will accelerate that process. Did it alter Iran’s apparent determination to acquire a nuclear capability or something close to it? No. Was it a prime cause of the recent domestic crisis and instability in Iran after its 2009 presidential election? No. Did it slow or accelerate the moderate Arab states intent to move along the nuclear path? No. Did it affect North Korea’s destabilizing nuclear calculations? No. Did it importantly weaken political reconciliation in Iraq? No, because there is almost none in any case. Did it slow the Middle East peace process? No, not least because prospects for progress on issues between Israel and the Palestinians are the most unpromising in 25 years. Did it substantially affect the enormous internal and international challenges associated with the growth of Jihadiism in Pakistan? No. But at the same time, it is important to stress that Pakistan, quite apart from the global recession, is the epicenter of global terrorism and now represents potentially the most dangerous international situation since the 1962 Cuban Missile Crisis. Did the global economic downturn systemically affect the future of Afghanistan? No. The fact that the United States is doing badly in the war in Afghanistan has nothing to do with the economic deterioration. As Henry Kissinger observes, “The conventional army loses if it does not win. The guerrilla wins if he does not lose.” And NATO is not winning in Afghanistan. Did it change in a major way the future of the Mexican state? No. Did the downturn make Europe, because of its domestic politics, less willing and able over time to join the U.S. in effective alliance policies? No, there will likely be no basic variations in Europe’s external policies and no serious evolution in transatlantic relations. As President Obama is experiencing regarding Europe, the problems with European publics in this regard are civilizational in character, not especially tied to this recession—in general, European publics do not wish their nations to take on foreign missions that entail the use of force and possible loss of life. Did the downturn slow further EU integration? Perhaps, at the margin, but in any case one has to watch closely to see if EU integration moves like a turtle or like a rock. And so forth. To be clear, there will inevitably be major challenges in the international situation in the next five years. In fact, this will be the most dangerous and chaotic global period since before the 1973 Middle East war. But it is not obvious that these disturbing developments will be primarily a result of the global economic problems. It is, of course, important to be alert to primary and enduring international discontinuities. If such a convulsive geopolitical event is out there, what is it? One that comes to mind is another catastrophic attack on the American homeland. Another is the collapse of Pakistan and the loss of government control of its nuclear arsenal to Islamic extremists. But again, neither of these two geopolitical calamities would be connected to the current economic decline. Some argue that, even though geopolitical changes resulting from the current global economic tribulations are not yet apparent, they are occurring beneath the surface of the international system and will become manifest in the years to come. In short, causality not perceptible now will become so. This subterranean argument is difficult to rebut. To test that hypothesis, the obvious analytical method is to seek tangible data that demonstrates that it is so. In short, show A, B, and/or C (in this case, geopolitical transformations caused by the world slump) to have occurred, thus substantiating the contention. One could then examine said postulated evidence and come to a judgment regarding its validity. To instead contend that, even though no such data can be adduced, the assertion, nevertheless, is true because of presently invisible occurrences seems more in the realm of religious conviction than rigorous analysis. But it is worth asking, as the magisterial American soldier/statesman George Marshall often did, “Why might I be wrong?” If the global economic numbers continue to decline next year and the year after, one must wonder whether any region would remain stable— whether China would maintain internal stability, whether the United States would continue as the pillar of international order, and whether the European Union would hold together. In that same vein, it is unclear today what effect, if any, the reckless financial lending and huge public debt that the United States is accumulating, as well as current massive governmental fiscal and monetary intervention in the American economy, will have on U.S. economic dynamism, entrepreneurial creativity, and, consequently, power projection over the very long term. One can only speculate on that issue at present, but it is certainly worth worrying about, and it is the most important “known unknown”27 regarding this subject.28 In addition, perhaps the Chinese Communist Party’s grip on China is more fragile than posited here, and possibly Pakistan and Mexico are much more vulnerable to failed-state outcomes primarily because of the economic downturn than anticipated in this essay. While it seems unlikely that these worst-case scenarios will eventuate as a result of the world recession, they do illustrate again that crucial uncertainties in this analysis are the global downturn’s length and severity and the long-term effects of the Obama Administration’s policies on the U.S. economy. Finally, if not, why not? If the world is in the most severe international economic crisis since the 1930s, why is it not producing structural changes in the global order? A brief answer is that the transcendent geopolitical elements have not altered in substantial ways with regard to individual nations in the two years since the economic crisis began. What are those enduring geopolitical elements? For any given country, they include the following: • Geographic location, topography, and climate. As Robert Kaplan puts it, “to embrace geography is not to accept it as an implacable force against which humankind is powerless. Rather, it serves to qualify human freedom and choice with a modest acceptance of fate.”29 In this connection, see in particular the works of Sir Halford John Mackinder and his The Geographical Pivot of History (1904)30, and Alfred Thayer Mahan, The Influence of Sea Power upon History, 1660–1783 (1890).31 • Demography—the size, birth rate, growth, density, ethnicity, literacy, religions, migration/emigration/ assimilation/absorption, and industriousness of the population. • The histories, foreign and defense policy tendencies, cultural determinants, and domestic politics of individual countries. • The size and strength of the domestic economy. • The quality and pace of technology. • The presence of natural resources. • The character, capabilities, and policies of neighboring states. For the countries that matter most in the global order, perhaps unsurprisingly, none of these decisive variables have changed very much since the global downturn began, except for nations’ weaker economic performances. That single factor is not likely to trump all these other abiding geopolitical determinants and therefore produce international structural change. Moreover, the fundamental power relationships between and among the world’s foremost countries have also not altered, nor have those nations’ perceptions of their vital national interests and how best to promote and defend them. To sum up this pivotal concept, in the absence of war, revolution, or other extreme international or domestic disruptions, for nation-states, the powerful abiding conditions just listed do not evolve much except over the very long term, and thus neither do countries’ strategic intent and core external policies— even, as today, in the face of world economic trials. This point was made earlier about Russia’s enduring national security goals, which go back hundreds of years. Similarly, a Gulf monarch recently advised—with respect to Iran—not to fasten on the views of President Ahmadinejad or Supreme Leader Khamenei. Rather, he counseled that, to best understand contemporary Iranian policy, one should more usefully read the histories, objectives, and strategies of the Persian kings Cyrus, Darius, and Xerxes, who successively ruled a vast empire around 500 BC.32 The American filmmaker Orson Welles once opined that “To give an accurate description of what never happened is the proper occupation of the of pundits. ■ historian.” 33 Perhaps the same is occasionally true Disease No Solvency They can’t solve disease – anti-vaccination movement gives disease a foothold within communities and spreads from there Kollipara 5/5 Puneet Kollipara, Journalist for the Washington Post, 5/5/2014, “How the Anti-Vaccine Movement is Endangering Lives” http://www.washingtonpost.com/blogs/wonkblog/wp/2014/05/05/how-the-anti-vaccine-movement-isendangering-lives/ Infectious diseases that we normally think of as rare in the United States are making a comeback. In recent years, pertussis -- also known as whooping cough -- has returned to the headlines. A measles outbreak that struck a Texas megachurch community late last summer sickened 21 people. And just recently, at least 16 people got sick during a measles outbreak in Ohio. In fact, the Centers for Disease Control and Prevention recently reported 13 measles outbreaks so far in 2014 -- the most since 1996. The diseases are highly contagious, but they are also preventable; two of the many recommended childhood vaccinations protect against measles and pertussis. And measles had been considered effectively eliminated in the United States. What's going on, exactly? Here are some answers. Why are so many outbreaks happening? Although it's a complex problem, health officials say one key culprit is that more and more people are choosing not to get their kids vaccinated against these diseases. For instance, in California parents are increasingly seeking personal or religious exemptions from vaccination requirements for their kids to attend schools. Substandard vaccination rates create an opening for outbreaks, which often start when an unvaccinated person catches the disease while traveling abroad and spreads the illness to friends and family upon returning. But aren't overall vaccination rates really high? Nationally, yes, though it wasn't always this way. Before the 1990s, rates languished below 80 percent for most childhood vaccines at the time. In 1993, after the 1989-1991 measles outbreak, Congress enacted the Vaccines for Children (VFC) program to promote childhood vaccinations. CDC data show that vaccination rates are now above 90 percent range for several routine vaccines, including the measles-mumps-rubella (MMR) and whooping cough vaccines. Public health officials target a 90 percent vaccination rate for most routine childhood vaccines. Experts say that a population has "herd immunity" when enough people are vaccinated to prevent a disease from taking hold. This chart shows how vaccination rates climbed after VFC's enactment: If vaccination rates are high, why are we seeing so many outbreaks? That's because vaccination rates aren't geographically uniform. Public-health experts say that high non-vaccination or exemption rates can occur among pockets of people, particularly at the county or city level. And some research has found that outbreaks are far more likely to happen in these areas, such as during the recent whooping cough outbreaks in California. Why are people not vaccinating their kids? There are a number of factors at play. Many of the diseases we vaccinate against are so rare here now that the public's awareness of vaccination might have decreased. But the one reason that has most alarmed public-health experts lately has been the rise of the antivaccine movement. Groups and activists such as celebrity Jenny McCarthy have repeatedly claimed that vaccines cause autism. This vaccine-autism concern may be causing a drop in childhood vaccination rates in many communities, including in affluent, well-educated ones. Do vaccines cause autism? Science gives a resounding no. Anti-vaccine activists often hang their case on a study published in the British journal The Lancet in 1998. This study, which posited a link between the MMR vaccine and autism, was widely discredited by the scientific community and eventually retracted. But the anti-vaccine movement has still gained steam. Anti- vaccine activists have also raised concerns about vaccines made with the mercury-based preservative known as thimerosal, which they worry could cause brain damage in developing kids. It's true that vaccines once routinely contained thimerosal, which government officials recognized as generally safe. But this preservative has been phased out of nearly all vaccines as a precautionary measure. Anti-vaccine activists also worry that the CDC's recommended vaccination schedule could overwhelm infants' immune systems by packing too many doses into a short period of time. Although the number of vaccinations that kids receive now is higher than it used to be, the main ingredients in the vaccines have actually decreased in amount. Even if these ingredient amounts hadn't decreased, research has found no relationship between those amounts and autism risk. Vaccines do carry a risk of side effects, but they are usually minor. The CDC has concluded from reviewing the scientific evidence that there's no causal link between childhood vaccinations and autism. No Solvency-Cant contain The aff can’t solve, containment is impossible Blancou et al 2005 Jean Blancou, former General Director of OEI, Bruno Chomel, Researcher for WHO/PAHO Collaborating Center on New and Emerging Zoonoses, Albino Belotto, Researcher for Veterinary Public Health Unit, Pan American Health Organization, François Meslin, Researcher for Strategy Development and Monitoring of Zoonoses, Food-borne Diseases and Kinetoplastidae (ZFK), Communicable Diseases Control Prevention and Eradication “Emerging or re-emerging bacterial zoonoses: factors of emergence, surveillance and control,” Vet. Res. 36, pg 507-522, http://hal.archivesouvertes.fr/docs/00/90/29/77/PDF/hal-00902977.pdf The main obstacles that are encountered in the control of bacterial zoonoses are the same as those opposed to the control of any infectious disease, that is most often finan- cial and human obstacles rather than tech- nical limitations. The financial resources needed to effi- ciently fight against zoonotic agents are not available for all countries. Only the international community’s financial support, could, notably, allow developing countries to organize a proper control of zoonotic dis- eases, but it is rare that this is materialized as a financial gift and mobilization of specific funds, even by well-known interna- tional organizations (such as WHO, FAO, OIE), is limited for such diseases. Due to all these difficulties, many sanitary authorities of these countries have given up the estab- lishment of such prevention programs. Oth- ers manage, with a lot of perseverance, to elaborate complicated multilateral financial arrangements. This allows punctual projects to be realized, but rarely to establish the long-term prophylaxis plans that they really need. When financial and material problems are supposedly solved, humanrelated dif- ficulties should not be underestimated. These difficulties can originate within the services in charge of applying the national prophy- laxis plans, when these services are not themselves convinced of the good use of these plans, or when they do not seem to get specific benefits from it. The obstacles sometimes result from a lack of cooperation between specific professional categories, amongst which figure breeders, as well as livestock brokers or even veterinarians bothered by the application of certain pro- grams of control or the limited incentive given by the health authorities for perform- ing prophylaxis tasks. Finally, the obstacle to such plans may be caused by the active opposition of the public opinion to certain methods of control. This is notably the case for the hostility of some groups to the mass slaughtering of animals during epizootics, or to the use of vaccines issued from genetic engineering. By lack of an appropriate con- sensus, the control of some zoonotic dis- eases may simply be impossible in some countries. No Extinction ---Super viruses won’t cause extinction (A.) Burnout. Lafee 2009 Scott, Union-Tribune Staff Writer, “Viruses versus hosts: a battle as old as time”, May 3rd, http://www.signonsandiego.com/news/2009/may/03/1n3virus01745-viruses-versus-hosts-battle-old-time/?uniontrib Generally speaking, it's not in a virus's best interest to kill its host. Deadly viruses such as Ebola and SARS are self-limiting because they kill too effectively and quickly to spread widely. Flu viruses do kill, but they aren't considered especially deadly. The fatality rate of the 1918 “Spanish flu” pandemic was less than 2.5 percent, and most of those deaths are now attributed to secondary bacterial infections. The historic fatality rate for influenza pandemics is less than 0.1 percent. Humans make “imperfect hosts” for the nastiest flu viruses, Sette said. “From the point of view of the virus, infecting humans can be a dead end. We sicken and die too soon.” (B.) Genetic diversity. Sowell 2001 Thomas, Fellow @ Hoover Institution, Jewish World Review, “The Dangers of “Equality””, 3-5, http://www.jewishworldreview.com/cols/sowell030501.asp People have different vulnerabilities and resistances to a variety of diseases. That is why one disease is unlikely to wipe out the human species, even in one place. An epidemic that sweeps through an area may leave some people dying like flies while others remain as healthy as horses. (C.) Co-evolution. Posner 2005 Richard, Judge, 7th Circuit court of Appeals, Catastrophe: Risk and Response, pg. 22 AIDS illustrates the further point that despite the progress made by modern medicine in the diagnosis and treatment of diseases, developing a vaccine or cure for a new (or newly recognized or newly virulent) disease may be difficult, protracted, even impossible. Progress has been made in treating ATDS, but neither a cure nor a vaccine has yet been developed. And because the virus's mutation rate is high, the treatments may not work in the long run.7 Rapidly mutating viruses are difficult to vaccinate against, which is why there is no vaccine for the common cold and why flu vaccines provide only limited protection.8 Paradoxically, a treatment that is neither cure nor vaccine, but merely reduces the severity of a disease, may accelerate its spread by reducing the benefit from avoiding becoming infected. This is an important consideration with respect to AIDS, which is spread mainly by voluntary intimate contact with infected people. Yet the fact that Homo sapiens has managed to survive every disease to assail it in the 200,000 years or so of its existence is a source of genuine comfort, at least if the focus is on extinction events. There have been enormously destaictive plagues, such as the Black Death, smallpox, and now AIDS, but none has come close to destroying the entire human race. There is a biological reason. Natural selection favors germs of limited lethality; they are fitter in an evolutionary sense because their genes are more likely to be spread if the germs do not kill their hosts too quickly. The AIDS virus is an example of a lethal virus, wholly natural, that by lying dormant yet infectious in its host for years maximizes its spread. Yet there is no danger that AIDS will destroy the entire human race. The likelihood of a natural pandemic that would cause the extinction of the human race is probably even less today than in the past (except in prehistoric times, when people lived in small, scattered bands, which would have limited the spread of disease), despite wider human contacts that make it more difficult to localize an infectious disease. The reason is improvements in medical science. But the comfort is a small one. Pandemics can still impose enormous losses and resist prevention and cure: the lesson of the AIDS pandemic. And there is always a first time. No Impact-Zoonotic disease Zoonotic diseases are less threatening than others — they are contained now Torres 1999 Alfonso Torres, D.V.M., M.S., Ph.D., Deputy Administrator, USDA, Animal Plant and Health Inspection Service, Veterinary Services, 2/6/99, “International Economic Considerations Concerning Agricultural Diseases and Human Health Costs of Zoonotic Diseases,” Annals of the New York Academy of Sciences 894:80-82, http://onlinelibrary.wiley.com/doi/10.1111/j.1749-6632.1999.tb08047.x/abstract Animal diseases can negatively affect the number and availability of animals, their productivity, or their appearance. 1 A few centuries ago, animal diseases affected mostly individual owners or herdsmen, but did not have serious consequences on the larger community. A similar event today will not only have a negative impact on the animal owners, but more importantly, will significantly affect the general economy of the region, the entire nation, even a group of nations. The importance of animal diseases as an element affecting international trade of animals and animal products has reached its full impact level with the recent designation by the World Trade Organization (WTO) of the International Office of Epizootics (OIE) as the international agency in charge of establishing animal health standards upon which international commerce can institute restrictions to prevent the spread of animal diseases from one nation to another. It is important to point out that while the spread of human diseases around the world is due to the unrestricted movement of people across political boundaries, animal diseases are, for the most part, restricted to defined geographic areas of the world due to the implementation of animal importation requirements, quarantines, animal movement regulations, and by disease control measures that include mass vaccination campaigns and animal depopulation practices. A number of animal diseases have been eradicated from countries or even from continents around the world by aggressive, well-coordinated, long-term animal health campaigns. This is in contrast to the relatively few human diseases successfully eradicated from large areas of the world. Warming DA turns Disease Warming turns the diseases advantage – climate change allows diseases to spread more effectively NRDC 2011 National Resources Defense Council, environmental protection group, 8/3/11, “Infectious Diseases: Dengue Fever, West Nile Virus, and Lyme Disease”, http://www.nrdc.org/health/climate/disease.asp While many infectious diseases were once all but eliminated from the United States, there's evidence that climate change is a factor that could help them expand their range and make a comeback. Mosquitoes capable of carrying and transmitting diseases like Dengue Fever, for example, now live in at least 28 states. As temperatures increase and rainfall patterns change - and summers become longer - these insects can remain active for longer seasons and in wider areas, greatly increasing the risk for people who live there. The same is true on a global scale: increases in heat, precipitation, and humidity can allow tropical and subtropical insects to move from regions where infectious diseases thrive into new places. This, coupled with increased international travel to and from all 50 states, means that the U.S. is increasingly at risk for becoming home to these new diseases. Nearly 4,000 cases of imported and locally-transmitted Dengue Fever were reported in the U.S. between 1995 and 2005, and that number rises to 10,000 when cases in the Texas-Mexico border region are included. In Florida, 28 locally-transmitted cases were reported in a 2009-2010 outbreak, the first there in more than 40 years. Dengue Fever, also known as "Breakbone Fever", is characterized by high fever, headaches, bone and joint aches, and a rash. Recurrent infection can lead to bleeding, seizures, and death. Lyme disease - transmitted primarily through bites from certain tick species - could expand throughout the United States and northward into Canada, as temperatures warm, allowing ticks to move into new regions. West Nile virus, which first entered the U.S. in 1999, expanded rapidly westward across the country. By 2005, over 16,000 cases had been reported. Warmer temperatures, heavy rainfall and high humidity have reportedly increased the rate of human infection. Sea Power 1NC—Sea Power Now The navy is extremely powerful—the loss of information systems wouldn’t hurt it Mizokami 14(Kyle, June 6, The Five Most-Powerful Navies on the Planet, Kyle Mizokami is a writer based in San Francisco who has appeared in The Diplomat, Foreign Policy, War is Boring and The Daily Beast. In 2009 he cofounded the defense and security blog Japan Security Watch., The National Intrest, The Five Most-Powerful Navies on the Planet, http://nationalinterest.org/feature/the-five-most-powerful-navies-the-planet-10610) First place on the list is no surprise: the United States Navy. The U.S. Navy has the most ships by far of any navy worldwide. It also has the greatest diversity of missions and the largest area of responsibility.¶ No other navy has the global reach of the U.S. Navy, which regularly operates in the Pacific, Atlantic and Indian Oceans, as well as the Mediterranean, Persian Gulf and the Horn of Africa. The U.S. Navy also forward deploys ships to Japan, Europe and the Persian Gulf.¶ The U.S. Navy has 288 battle force ships, of which typically a third are underway at any given time. The U.S. Navy has 10 aircraft carriers, nine amphibious assault ships, 22 cruisers, 62 destroyers, 17 frigates and 72 submarines. In addition to ships, the U.S. Navy has 3,700 aircraft, making it the second largest air force in the world. At 323,000 active and 109,000 personnel, it is also the largest navy in terms of manpower.¶ What makes the U.S. Navy stand out the most is its 10 aircraft carriers—more than the rest of the world put together. Not only are there more of them, they’re also much bigger: a single Nimitz-class aircraft carrier can carry twice as many planes (72) as the next largest foreign carrier. Unlike the air wings of other countries, which typically concentrate on fighters, a typical U.S. carrier air wing is a balanced package capable of air superiority, strike, reconnaissance, anti-submarine warfare and humanitarian assistance/disaster relief missions.¶ The U.S. Navy’s 31 amphibious ships make it the largest “gator” fleet in the world, capable of transporting and landing on hostile beaches. The nine amphibious assault ships of the Tarawa and Wasp classes can carry helicopters to ferry troops or act as miniature aircraft carriers, equipped with AV-8B Harrier attack jets and soon F-35B fighter-bombers.¶ The U.S. Navy has 54 nuclear attack submarines, a mix of the Los Angeles, Seawolf, and Virginia classes. The U.S. Navy is also responsible for the United States’ strategic nuclear deterrent at sea, with 14 Ohio-class ballistic missile submarines equipped with a total of 336 Trident nuclear missiles. The USN also has four Ohio-class submarines stripped of nuclear missiles and modified to carry 154 Tomahawk land attack missiles.¶ The U.S. Navy has the additional roles of ballistic missile defense, space operations and humanitarian assistance/disaster relief. As of October 2013, 29 cruisers and destroyers were capable of intercepting ballistic missiles, with several forward deployed to Europe and Japan. It also monitors space in support of U.S. military forces, tracking the satellites of potential adversaries. Finally, the U.S. Navy’s existing aircraft carriers and amphibious vessels, plus the dedicated hospital ships USNS Mercy and USNS Comfort, constitute a disaster relief capability that has been deployed in recent years to Indonesia, Haiti, Japan and the Philippines. 2NC—Sea Power Now US naval power is decades ahead of the nearest competitor—plan doesn’t matter Bloomberg 14(April 9, China's Aircraft Carrier Is Nothing to Fear, http://www.bloombergview.com/articles/2014-04-09/china-s-aircraft-carrier-is-nothing-to-fear) China’s plans for a blue-water navy centered on aircraft carriers follow a course the U.S. Navy plotted more than half a century ago. This head start gives the U.S. an insurmountable lead in operational experience and military hardware. The Liaoningis a refurbished Soviet warship with a conventional (not nuclear) power plant and a tilted deck that severely limits the range and payload of its aircraft. It would be no match for any of the U.S. Navy's 11 nuclear-powered carriers, each carrying twice as many aircraft and stateof-the-art catapults for launching them.¶ China’s planned new carriers will be far more capable than the Liaoning -- but at a crippling price. The newest U.S. carrier will cost $13 billion. Add roughly 60 aircraft, 6,700 sailors, and other vessels to support and protect it, and a single U.S. carrier strike group costs roughly $6.5 million a day to operate. Such outlays are a questionable investment even for the U.S., let alone China. China's blue-water ambitions will drain a significant portion of its military budget to a sphere it cannot hope to dominate.¶ Better still, the strategic aims that China can reasonably advance with a carrier-based navy are relatively benign. Such a force could protect China’s energy lifeline of oil tankers stretching to the Middle East, which is occasionally threatened by pirates. It could more effectively respond to humanitarian disasters and, if the need arose, evacuate Chinese citizens working in far-flung locations.¶ In addition, of course, China's government wants its blue-water navy for prestige, as a symbol of the country's rise to great-power status. That's fine, too -- so long as it poses no threat to U.S. naval dominance. When it comes to Chinese military power, aircraft carriers are the least of the world's concerns. 1NC—Information Dominance Now Lockheed Martin solves data collection integration now Lockheed Martin 14(June 30, Lockheed Martin Awarded Contract to Enhance U.S. Navy C4ISR Collection and Dissemination Capabilities, http://www.lockheedmartin.com/us/news/press-releases/2014/june/isgs-navyC4ISR0630.html) Lockheed Martin [NYSE:LMT] will work to enhance how the Navy exchanges C4ISR data throughout the space, air, surface, subsurface, and unmanned sensor domains under a contract with Space and Naval Warfare Systems Center Pacific. This IDIQ contract has a ceiling value of $35 million over five years.¶ “For the Navy, every platform is a sensor, and every sensor must be networked,” said Dr. Rob Smith, vice president of C4ISR for Lockheed Martin Information Systems and Global Solutions. “We’ll leverage our more than 30 years developing and fielding signals intelligence systems to increase the Navy’s intelligence sharing capability across the full spectrum of maritime and littoral missions.”¶ Lockheed Martin codeveloped the Navy’s Distributed Information Operations-System, which addresses the Navy’s need for network-centric intelligence to improve interoperability and enhance battlespace awareness. For that effort, Lockheed Martin connected disparate Navy signals intelligence systems facilitating tactical data exchange and allowing commanders to better understand their operational environment.¶ Building upon those capabilities, Lockheed Martin will to continue to enhance the Navy’s signals intelligence collection, data fusion, and intelligence processing and dissemination capabilities. This could involve integrating and deploying capabilities that monitor the status of all sensors registered in the network; then displaying the input from those sensors in support of real-time planning. Network integration of sensors will be designed to accomplish cross-cueing, cooperative sensing and, where feasible and prudent, automated target recognition or classification. The workscope for this contract also includes analyzing ways to enhance the Navy’s use of Unmanned Aerial Vehicles (UAVs) for surface combatant land attacks. Lockheed Martin solves information dominance Tradingcharts.com 14(July 1, Lockheed Martin gets contract to enhance US Navy C4ISR collection, dissemination capabilities, http://futures.tradingcharts.com/news/futures/Lockheed_Martin_gets_contract_to_enhance_US_Navy_C4ISR_col lection__dissemination_capabilities_216288980.html) Lockheed Martin (NYSE: LMT) said it will work to enhance how the Navy exchanges C4ISR data throughout the space, air, surface, subsurface, and unmanned sensor domains under a contract with Space and Naval Warfare Systems Center Pacific. This IDIQ contract has a ceiling value of USD35m over five years.¶ Lockheed Martin co-developed the Navy's Distributed Information Operations-System, which addresses the Navy's need for network-centric intelligence to improve interoperability and enhance battlespace awareness. For that effort, Lockheed Martin connected disparate Navy signals intelligence systems facilitating tactical data exchange and allowing commanders to better understand their operational environment. 2NC—Information Dominance Now Lockheed Martin solves any problems with sensors Cheng 14(Joey, June 30, Navy awards $35M contract to boost C4ISR info sharing, Defense Systems, Joey Cheng is a writer for Defense Systems Magazine, http://defensesystems.com/articles/2014/06/30/spawar-dio-s-unpgradeslockheed.aspx) In an effort to improve information sharing from its multitude of sensors, the Navy has awarded Lockheed Martin a contract to enhance the service’s C4ISR data exchange capabilities.¶ The indefinite delivery/indefinite quantity contract with the Space and Naval Warfare Systems Center Pacific has a ceiling of up to $35 million over five years, and would upgrade how the Navy’s space, air, surface, subsurface and unmanned sensors would collect and disseminate data, according to a Lockheed Martin release. ¶ SPAWAR is the Navy’s Information Dominance system command, and is responsible for the development of communications and information capabilities for warfighters.¶ The Navy’s Distributed Information Operations-System, originally co-developed by Lockheed, was designed to improve interoperability and enhance battlespace awareness through network-centric intelligence, and connects a variety of signals intelligence systems for tactical data exchange.¶ Leveraging its experience with DIO-S, Lockheed may be working to implement a monitoring system capable of checking the statuses of all of the sensors registered in its network in the future. The same system would then display the input from the sensors for real-time planning.¶ Further integration of the sensors would allow cross-cueing and cooperative sensing, as well as automated target recognition. The scope of the contract also includes possible enhancements of the Navy’s use of unmanned aerial vehicles for ground attacks. ¶ Lockheed Martin also is a co-developer of the Distributed Common Ground System, which seeks to integrate multiple intelligence, surveillance and reconnaissance (ISR) sensors from different services and agencies into a common network. The Navy’s version, DCGS-N, is the service’s primary ISR&T (the “T” is for Targeting) support system and provides processing, exploitation, and dissemination capabilities for the operational and tactical level. These systems are designed to allow modularity, flexibility, and standardization for integrating data sources, transformation services, and user interfaces, according to Lockheed Martin.¶ “For the Navy, every platform is a sensor, and every sensor must be networked,” said Dr. Rob Smith, vice president of C4ISR for Lockheed Martin Information Systems and Global Solutions. “We’ll leverage our more than 30 years developing and fielding signals intelligence systems to increase the Navy’s intelligence sharing capability across the full spectrum of maritime and littoral missions.” Current navy sensors solve USNIDC 13(US Navy Information Dominance Corps, March, US Navy Dominance Information Dominance Roadmap 2013-2028, http://defenseinnovationmarketplace.mil/resources/Information_Dominance_Roadmap_March_2013.pdf) A number of changes are programmed to improve the Navy's ability to manage¶ collect, process and disseminate essential information to Navy units afloat and¶ ashore. Near-term BA capabilities to be addressed during this timeframe are as¶ follows:¶ ' Centralized fleet collection management (CM) and tasking to support both¶ standing and ad hoc requirements across diverse intelligence disciplines will¶ be developed and will include: integrated collection plans, a consolidated¶ collection requirements database, and the means to visualize available¶ collection assets;¶ ' Manned, unmanned and other diverse platforms will be fielded to extend¶ orgnic sensor capability and capacity; commercial space-based imaging and¶ Automatic lnformation Systems (AIS) collection systems will proliferate,¶ and an emergent capability to task, plan and direct organic and other¶ sensors will be developed;¶ ' The transformation to distributed network environments will begin to¶ emerge; high-performance computing will better understand and predict¶ the physical and virtual environments; automation of multiple intelligence¶ source (multi-INT) data/information fusion and correlation will evolve and¶ adopt common data models and standards; and service-based architectural¶ frameworks will be developed to enhance information sharing;¶ ' An emerging system-of-systems approach fiir providing a Common¶ Operational Picture (COP) and a Common Maritime Picture (CMP) will¶ begin to shape tactics, techniques and procedures to enhance multi-domain¶ awareness, and enterprise services and visualization tools will be developed¶ to help understand information, depict actions and trends in near real-time,¶ ' Data sharing will be enhanced, enterprise solutions will be pursued fiir¶ data purchasing and cross-domain solutions will be developed to begin¶ consolidating Top Secret and Secret networks into a single classified¶ domain;¶ MOCs will remain as the centers ofgravity for Fleet BA coordination and¶ will serve as central nodes for the C2 of Navy ISR assets and resultant in-¶ theater analysis, and fleet commanders and TAOs will be able to establish¶ and share a maritime COP that is periodically updated;¶ ' Planning tools with Theater Security Cooperation capabilities and mission¶ partner capability information sources will become better integrated;¶ ' Improved BA training fi)r all operators and watchstanders will be developed. 1NC—Information Overload IOOS fails—overloads system with too much data RAND 14(The RAND Corporation is a nonprofit institution that ¶ helps improve policy and decisionmaking through research and analysis, Data Flood Helping the Navy Address the Rising ¶ Tide of Sensor Information, http://www.rand.org/content/dam/rand/pubs/research_reports/RR300/RR315/RAND_RR315.pdf) ¶ Despite the bat"t1e—tested¶ value of ISR systems, however, the large amount of data diey gener-¶ ate has become overwhelming to Navy analysts. As the Intelligence¶ Science Board wrote in 2008, referring to d'1C entire Department of¶ Defense (DoD), “the number of images and signal intercepts are well¶ beyond the capacity of the existing analyst community, so there are¶ huge bacldogs For tra.nslators and image interpreters, and much of the¶ collected data are never reviewed." This is a good description of the¶ Navy’s big data challenge. 2NC—Information Overload IOOS doesn’t solve—information overload RAND 14(The RAND Corporation is a nonprofit institution that ¶ helps improve policy and decisionmaking through research and analysis, Data Flood Helping the Navy Address the Rising ¶ Tide of Sensor Information, http://www.rand.org/content/dam/rand/pubs/research_reports/RR300/RR315/RAND_RR315.pdf) ¶ U.S. Navy intelligence, surveillance, and reconnaissance (ISR) func-¶ tions have become critical to U.S. national security over the last two¶ decades.1 Within the Navy, there is a growing demand for ISR data¶ from drones and other souroes that provide situational awareness,¶ which helps Navy vessels avoid collisions, pinpoint targets, and per-¶ form a host of odier mission-critical tasks.¶ The amount of data generated by ISR systems has, however,¶ become overwhelming. All of the data collected by the Navy—and¶ available from other sources, both government and commercial—are¶ potentially useful, but processing them and deriving useful knowledge¶ from them are severely taxing die analytical capabilities of the Navy's¶ humans and networks. As the Navy aoquires and fields new and addi-¶ tional sensors for collecting data, this “big data” challenge will con-¶ tinue to grow. Indeed, if the Navy continues to field sensors as planned¶ but does not change the way it processes, exploits, and disseminates¶ information, it will reach an ISR “tipping point"—the point at which¶ intelligence analysts are no longer able to complete a minimum number¶ of exploitation tasks within given time constraints—as soon as 2016.1 IOOS would fail—too much data collection Sarkar 14(Dibya, May 21, Navy must deal with rising tide of sensor data, writer for GCN, http://gcn.com/Articles/2014/05/21/RAND-Navy-cloud.aspx?Page=1) “We’re only now at the point where we’re starting to put up new UAVs with incredible sensors, so the problem is only going to get worse,” said Isaac R. Porche III, a senior researcher with RAND and co-author of the report.¶ Porche said the Navy had argued for more manpower to deal with the growing volumes of data, but budgetary pressures forced it to seek other options to improve efficiency. RAND, which was hired to do a quantitative assessment, looked at the imagery, video and audio that the Navy was collecting from UAVs, studying how long it took for its analysts to process data from several siloed databases using different desktop applications.¶ The report, Data Flood: Helping the Navy Address the Rising Tide of Sensor Information, concluded that the Navy may reach a tipping point as early as 2016 when “analysts are no longer able to complete a minimum number of exploitation tasks within given time constraints.” In other words, analysts will be much less effective than they are now in finding value in the exponentially increasing data if the Navy doesn’t change the way it collects, processes, uses, and distributes that data. IOOS doesn’t solve—sensors collect too much data Ungerleider 14(Neal, May 23, Neal Ungerleider covers science and technology for Fast Company, Will Drones Make The U.S. Navy Migrate To The Cloud?, http://www.fastcolabs.com/3031082/will-drones-make-the-us-navymigrate-to-the-cloud) The U.S. Navy loves sensors. The same gyrometers and motion detectors that fuel smartphones and Microsoft’s Kinect also keep drones in the skies and aircraft carriers in sharp order. But the Navy--which is the size of a large global megacorporation and also saddled with a huge bureaucracy--also has problems dealing with the flood of data that sensors create. The RAND Corporation is arguing for something unorthodox to deal with that data influx: A cloud for the Navy (PDF).What do you think?¶ In a paper called, succinctly, “Data Flood,” think tank analyst Isaac R. Porche argues that the United States Navy needs a private cloud service to cope with the influx of data created by sensors on unmanned aerial vehicles (UAVs) and other smart military devices. Porche and his coauthors argue that there’s a precedent for creating a private cloud service for the military--the little-known fact that the NSA, CIA, and other American intelligence agencies are building their own cloud. In late October, Amazon Web Services reportedly sealed a $600 million, 10-year deal with the CIA to develop a custom cloud infrastructure for U.S. intelligence agents.¶ Porche and his team argue that if the Navy doesn’t adopt a cloud infrastructure, there will be a tipping point as early as 2016 where analysts will become less and less effective because bandwidth choke will prevent access to the information they need. “2016 is the year when the number of UAS (drone) systems will be high enough and generating enough sensor data/imagery to overwhelm human analysts,” Porche told Co.Labs by email.¶ Increasing the amount of information available does not increase military effectiveness USNI 9, US Naval institute blog, September 2009, “Information Dominance!? Say WHAT?” http://blog.usni.org/2009/09/18/information-dominance-say-what “ Information Dominance ?” Who was the Legion-of-Merit wearing 05 or 06 who thought THAT one up? Do we actually think we dominate the information spectrum?¶ Nothing against VADM Dorsett, but this type of terminology represents a badly flawed understanding of “information” as it pertains to warfighting. This seems like the Cult of Cebrowski. Network-Centric Warfare as a concept is found very much wanting where the rubber meets the road. It might work if the enemy we are fighting is the United States Navy, who has elevated information technology to a place on Olympus.¶ But for the rest of the world, our obsession with a flattened information hierarchy is much more of a vulnerability than an advantage. That obsession treats information like a drug. The more we have, the more we want. We are convinced that the key gem of information that will clear away all the fog MUST be coming in the next data dump. Except the data dump is increasingly difficult to sift through to find the critical information, if we are even aware of what is critical in light of so much unfiltered and unverified gatherings. The effort required to produce pertinent and actionable intelligence, or even timely and worthwhile suffers from such an approach to information and intelligence gathering.¶ Without launching a long dissertation regarding NCW and the problems created by “information inundation” resulting from a sensor/shooter marriage across a massive battlefield, I believe such a term as “Information Dominance” pays short shrift to a less sophisticated enemy whose extensive low-tech networks of informants are capable of negating billions of dollars of maritime “stealth” technology (or finding the right merchant ship in a 400-mile long shipping lane) by using the eyeball, the messenger, the motor scooter, or the sandal.¶ Such an enemy, I would argue, has a much clearer and simpler set of information requirements, and is far more skilled, more disciplined, and much more successful at meeting those requirements than we are. So, who is dominant in the information game? One could argue very effectively that it is not us.¶ Terms such as “Information Dominance” ring of a grandiose bit of self-delusion that is part plan and part capability, with a large measure of wishful thinking. In the words of a certain retired Marine Corps General, “words mean things”. They also information, oftentimes reveal a lot about who uses them, and how. The term “Information Dominance” gives me the willies. Can we call it something a little less presumtuous? 1NC—Sea Power Not K2 Deterrence Naval power is irrelevant—there are no emerging threats Gerson and Russell 11(Michael and Allison Lawler, Michael Gerson is a nationally syndicated columnist who appears twice weekly in the Washington Post. Michael Gerson is the Hastert Fellow at the J. Dennis Hastert Center for Economics, Government, and Public Policy at Wheaton College in Illinois., Russell is also an assistant professor of Political Science and International Studies at Merrimack College in North Andover, Massachusetts. Russell holds a Ph.D. from the Fletcher School of Law and Diplomacy at Tufts University, an M.A. in International Relations from American University in Washington, D.C., and a B.A. in Political Science from Boston College., American Grand Strategy and Sea Power, https://cna.org/sites/default/files/research/American%20Grand%20Strategy%20and%20Seapower%202011%20Conf erence%20Report%20CNA.pdf) According to Dr. John Mueller, the United States has been – and will continue to be – ¶ substantially free from threats that require a great deal of military preparedness. In his view, ¶ there are no major threats to U.S. security, and there have been none since the end of the ¶ Second World War. During the Cold War, the United States spent trillions of dollars to deter ¶ a direct military threat that did not exist, since the Soviet Union had no intention of ¶ launching an unprovoked attack on Europe or the United States. ¶ Despite the continued absence of significant threats today, the United States is still engaged ¶ in a number of conflicts in an effort to make the world look and act the way we want. In ¶ reality, however, most modern security issues are not really military in nature; rather, they are ¶ policing and diplomatic activities that do not require substantial U.S. military involvement. ¶ While isolationism is not a viable policy, the United States does not need to use its military to ¶ solve all of the problems in the world. 158¶ 17¶ Dr. Mueller argued that the absence of war among developed countries since 1945 is the ¶ greatest single development about war in history. The end of the Cold War ushered in a New ¶ World Order, and from 1989 to 2000 the United States was engaged in what Dr. Mueller ¶ called “policing wars.” There was very little domestic support for most of these ventures, ¶ however, because there was a strong U.S. public aversion to nation building, a low tolerance ¶ for casualties, and a lack of concrete political gains from success. Navy power is irrelevant—future conflicts will use ground forces Till 6(Geoffrey, December, NAVAL TRANSFORMATION, GROUND ¶ FORCES, AND THE EXPEDITIONARY IMPULSE:¶ THE SEA-BASING DEBATE, GEOFFREY TILL is the Professor of Maritime Studies ¶ at the Joint Services Command and Staff College, http://www.strategicstudiesinstitute.army.mil/pdffiles/pub743.pdf) Recent experience suggests that the lessons learned¶ with such difficulty by the allies when they confronted¶ the chaos ensuing from the sudden collapse of Nazi¶ Germany in 1945 were major casualties of the Cold¶ War.‘ In Afghanistan and Iraq, the United States and¶ its allies are being reminded painfully that stabilization¶ and reconstruction require larger numbers of troops¶ on the ground for much longer than preintervention¶ planners might have thought necessary, and that the¶ soldiers in question need much more than "mere"¶ warlighting skills. To relearn these lessons will likely¶ require a shift in the organizational cultures of the¶ armed services?¶ Navies and air forces around the world have¶ drawn from this experience the obvious conclusion that future defense priorities in countries with similar¶ interventionist aspirations are likely to reflect a grow-¶ ing relative emphasis on the provision of intelligently¶ trained and responsive "boots on the ground." With¶ resources being finite, defense expenditure on those¶ aspects of air and naval forces whose function seems¶ less than wholly related to this central aim seem likely¶ to be limited.3 2NC—Sea Power Not Key to Deterrence Naval power not key—ground forces are sufficient Rhodes 99(Edward, Spring, Naval War College Review, “. . . From the Sea”and Back Again Naval Power in the Second American Century, Dr. Rhodes is associate professor of international relations and director of the Center for Global Security and Democracy at Rutgers University.,Volume 42, No. 2, http://fas.org/man/dod101/sys/ship/docs/art1-sp9.htm#AUTHOR) On the other hand, it also suggests the possibility that the Army is right and that if forward presence is to matter it needs to be on the ground, that an offshore presence of a potent but limited force, with only the implicit threat of surged ground forces, is less likely to have an impact, at least if the potential aggressor has limited goals. It also suggests the possibility that the symbolism of naval forward presence, serving as a reminder of the full weight and power the United States could ultimately bring to bear, may not be that important. In war, the argument that forward naval forces operating with a littoral strategy can have an important impact in the initial phases of the conflict, thereby preparing the ground for later U.S. successes, is doubtless true. While true, however, it may well be relevant in only a limited range of cases. Most potential conflicts or contingencies involve adversaries who are too small for this effect to matter much. Short of a major regional conflict (¶ MRC¶ ), the superiority of U.S. military forces is sufficiently overwhelming that initial setbacks are not likely to be critically important. At the other extreme, in the case of a regional near-peer competitor—a Russia or a China—it is hard to imagine a littoral strategy having much of an impact: the amount of (nonnuclear)power that can be projected from the sea is trivial compared to the size of the adversary’s society or military establishment. Deterrence fails—multiple studies prove Rhodes 99(Edward, Spring, Naval War College Review, “. . . From the Sea”and Back Again Naval Power in the Second American Century, Dr. Rhodes is associate professor of international relations and director of the Center for Global Security and Democracy at Rutgers University.,Volume 42, No. 2, http://fas.org/man/dod101/sys/ship/docs/art1-sp9.htm#AUTHOR) In crisis, the forward-deployed capacity to project power "from the sea" is touted as having an immediate deterrent effect—that is, dissuading an adversary who is tentatively considering going to war from following through on that idea. Here we do have some evidence; at very best, however, it must be regarded as offering mixed support for the Navy's advocacy of a littoral approach. A variety of studies of conventional deterrence have been undertaken.60 While the research questions, underlying theoretical assumptions, and research methods have varied, several general findings emerge.¶ The principal one is that immediate extended deterrence with conventional means—that is, using threats of conventional response to deter an adversary who is considering aggression against a third party—regularly fails, even in cases where commitments are "clearly defined, repeatedly publicized and defensible, and the committed [gives] every indication of its intentions to defend them by force if necessary."61 Unlike nuclear deterrence, conventional deterrence does not appear to result in a robust, stable stalemate but in a fluid and competitive strategic interaction that, at best, buys time during which underlying disputes or antagonisms can be resolved. The possession of decisive conventional military superiority and the visible demonstration of a resolve will not necessarily permit the United States to deter attacks on friends and interests.¶ There are three reasons why immediate extended conventional deterrence is so problematic. First, potential aggressors are sometimes so strongly motivated to challenge the status quo that they are willing to run a high risk, or even the certainty, of paying the less-than-total costs of losing a war. Second, potential aggressors frequently conclude, correctly or incorrectly, that they have developed a military option that has politically or militarily "designed around" the deterrent threat. Third, there is considerable evidence that, particularly when they are under severe domestic stress, potential aggressors are unable to understand or respond rationally to deterrent threats. "Wishful thinking" by leaders who find themselves caught in a difficult situation appears to be an all-too-common pathology.¶ Further, and more germane to the issue of naval forward presence as a crisis deterrent tool, there is some evidence that because of the general insensitivity of potential aggressors to information, efforts to "signal" resolve through measures such as reinforcing or redeploying forces have limited effectiveness. If force movements are large enough to foreclose particular military options, they may forestall aggression. But as a means of indicating resolve and convincing an aggressor of the credibility of deterrent commitments, they do not generally appear to have an impact. Naval power irrelevant—it can’t effectively project power Rhodes 99(Edward, Spring, Naval War College Review, “. . . From the Sea”and Back Again Naval Power in the Second American Century, Dr. Rhodes is associate professor of international relations and director of the Center for Global Security and Democracy at Rutgers University.,Volume 42, No. 2, http://fas.org/man/dod101/sys/ship/docs/art1-sp9.htm#AUTHOR) The first and most immediate danger is from competitors to the littoral strategy: there are, as Army and Air Force voices have noted, a variety of ways besides projecting power "from the sea" to support a liberalinternationalist foreign policy and to fight a transoceanic-countermilitary war. While budgetary realities have stimulated this strategic competition between the services and are likely to continue to serve as the spur, it would be wrong to dismiss this challenge to the littoral strategy as mere interservice rivalry or budgetary gamesmanship. Rather, what has developed is a serious, if admittedly parochially grounded, intellectual debate over alternative national military strategies—over alternative ways to use America's military potential in support of "engagement and enlargement." While a littoral naval strategy is consistent with a liberalinternationalist vision of national security and a transoceanic-countermilitary image of war, it is not the only military strategy of which that can be said, and the Army and Air Force have successfully articulated alternative military strategies that call into question the need for significant naval effort in the littorals.¶ The second danger, linked to the first, is that the Navy may be unable to develop a workable operational concept for putting the littoral strategy into effect. Indeed, the Navy has found it remarkably difficult to script a convincing story about precisely how a littoral strategy works—that is, the Navy has had a hard time identifying what it is about naval operations in the littorals that yields political-military leverage and what forces and activities are therefore required. The failure of "Forward . . . from the Sea" to address the issue of alternative force packages is illustrative in this regard: continued insistence that carrier battle groups and amphibious ready groups are needed at all times in all theaters reflects the conceptual and bureaucratic difficulty of determining the actual requirements of a littoral strategy. Any decision to change deployment patterns, mixes, or timetables would at least implicitly require a prioritization of peacetime, crisis, and wartime duties; it would also represent a reallocation of resources within the service. But without a clear understanding of the process by which littoral operations generate the peacetime, crisis, and wartime outcomes sought, the Navy will find it impossible to make the difficult tradeoffs demanded by budgetary pressures. Indeed, as budgetary pressures, the need to moderate personnel and operational tempos, and the need to modernize become greater, the imperative for a clearer understanding of the relative value of (for example) forward peacetime presence, forward peacetime presence by carriers and amphibious forces, rapid crisis response, and massive wartime strike capacity will increase. Ultimately the danger is that a littoral strategy will become unworkable through an inability of the Navy to make the required tradeoffs, in which case it will find itself with forces that are too small, too overstretched, too poorly maintained, too poorly trained or manned, too obsolescent, or simply improperly configured to meet what prove to be the essential demands of a littoral strategy.¶ The third danger, more basic and more beyond the control of the Navy than the first two, is that the vision of warfare underlying the littoral strategy will be abandoned by the nation. The DESERT STORM image of war as a transoceanic countermilitary encounter is increasingly vulnerable, and as the elite and public begin to imagine war in other, more traditional terms, the attractiveness and importance of projecting power "from the sea" will become less apparent. To stay in harmony with national leadership and national strategy, the Navy will be called upon to offer a revised account of the utility of naval power. Large militaries are irrelevant in deterrence--empirics Mueller 11(John, “Embracing threatlessness: ¶ Reassessing U.S. military spending, John Mueller is Adjunct Professor of Political Science and Senior Research Scientist at the Mershon Center for International Security Studies., https://cna.org/sites/default/files/research/American%20Grand%20Strategy%20and%20Seapower%202011%20Conf erence%20Report%20CNA.pdf) Over the course of the last several decades, the United States has variously sensed threat from ¶ small counties led by people it found to be decidedly unpleasant. These rogue states (as they ¶ came to be called in the 1990s) were led by such devils de jour as Nasser, Sukarno, Castro, ¶ Qaddafi, Khomeini, Kim Il-Sung, and Saddam Hussein, all of whom have since faded into ¶ history’s dustbin. ¶ Today, such alarmed focus is directed at teetering Iran, and at North Korea, the most ¶ pathetic state on the planet. Except in worst-case fantasies, however, neither country presents ¶ a threat of direct military aggression – Iran, in fact, has eschewed the practice for several ¶ centuries. Nonetheless, it might make some sense to maintain a capacity to institute ¶ containment and deterrence efforts carried out in formal or informal coalition with ¶ concerned neighboring countries – and there are quite a few of these in each case. However, ¶ neither country is militarily impressive and the military requirements for effective ¶ containment are far from monumental and do not necessarily need large forces-in-being. ¶ Moreover, the Iraq syndrome seems already to be having its effect in this area. Despite nearly ¶ continuous concern about Iranian nuclear developments, proposals to use military force to ¶ undercut this progress have been persistently undercut. ¶ The Gulf War of 1991 is an example of military force being successfully applied to deal with a ¶ rogue venture – the conquest by Saddam Hussein’s Iraq of neighboring Kuwait. This ¶ experience does not necessarily justify the maintenance of substantial military forces, ¶ however. First, Iraq’s invasion was rare to the point of being unique: it has been the only case ¶ since World War II in which one UN country has invaded another with the intention of ¶ incorporating it into its own territory. As such, the experience seems to be much more of an ¶ aberration than a harbinger. Second, in a case such as that, countries do not need to have a ¶ large force-in-being because there is plenty of time to build a force should other measures to ¶ persuade the attacker to withdraw, such as economic sanctions and diplomatic forays, fail. ¶ And third, it certainly appears that Iraq’s pathetic forces – lacking strategy, tactics, leadership, ¶ and morale – needed the large force thrown at them in 1991 to decide to withdraw.18 Naval power fails Till 6(Geoffrey, December, NAVAL TRANSFORMATION, GROUND ¶ FORCES, AND THE EXPEDITIONARY IMPULSE:¶ THE SEA-BASING DEBATE, GEOFFREY TILL is the Professor of Maritime Studies ¶ at the Joint Services Command and Staff College, http://www.strategicstudiesinstitute.army.mil/pdffiles/pub743.pdf) Even so, at least three problems remain. First, there ¶ may be a tendency to focus force protection too much 1613¶ on the sea lines of communication and not enough ¶ on the security of sea ports, both of embarkation and ¶ arrival. There also is a natural tendency to concentrate ¶ on the safe and timely arrival of soldiers and their ¶ equipment in the theater, neglecting the dangers that, ¶ in these asymmetrical days, may be posed to their ¶ collection and dispatch at home. In the Iraq operation, ¶ Greenpeace attempted to interfere with the loading of ¶ military supplies for the British forces at the military ¶ port of Marchwood in Southampton Water. They were ¶ ineffective and harmless, but nonetheless represented ¶ a useful reminder of the vulnerability at the supplier’s ¶ end of the supply chain. ¶ Second, many would doubt the permanence of the ¶ shift of naval priorities away from oceanic sea control ¶ and towards its coastal force-protection variant. The ¶ emergence of new maritime powers such as Japan ¶ and China, or the recovery of Russia, might lead to ¶ a resurgence of peer competition and old-fashioned ¶ maritime rivalry on the high seas. The U.S. Navy’s ¶ current wariness about the prospective maritime ¶ expansion of China later in the century may be used to ¶ justify investment in more conventional forms of naval ¶ power. ¶ Third, the ability of naval forces to maintain an ¶ operational posture against relatively unsophisticated ¶ shore-based opposition can be exaggerated. The ¶ vulnerability of ships to coastal mines, small, quiet ¶ diesel submarines, terrorists on jet skis, or radical ¶ weaponry of the kind recently demonstrated by the ¶ Iranians, has to be taken seriously.24 1NC—Sea Power Not K2 Trade Naval power is not key to trade—there are no credible threats Mueller 11(John, “Embracing threatlessness: ¶ Reassessing U.S. military spending, John Mueller is Adjunct Professor of Political Science and Senior Research Scientist at the Mershon Center for International Security Studies., https://cna.org/sites/default/files/research/American%20Grand%20Strategy%20and%20Seapower%202011%20Conf erence%20Report%20CNA.pdf) Particularly in an age of globalization and expanding world trade, many, especially in the ¶ Navy, argue that a strong military force is needed to police what is portentously labeled the ¶ “global commons.” However, there seems to be no credible consequential threat, beyond ¶ those to worldwide shipping. There have been a few attacks by pirates off Somalia, exacting ¶ costs in the hundreds of millions of dollars a year from a multi-billion dollar industry which, ¶ surely, has the capacity to defend itself from such nuisances – perhaps by the application on ¶ decks of broken glass, a severe complication for barefoot predators. In the unlikely event that ¶ the problem should become severe, it would not need forces-inbeing; it could be dealt with ¶ by newly formulated forces designed for the specific purpose 2NC—Sea Power Not K2 Trade Naval power not key to trade Hultin & Blair 6(Jerry MacArthur HULTIN and Admiral Dennis BLAIR, Naval Power and Globalization: ¶ The Next Twenty Years in the Pacific, Jerry M. Hultin is former president of Polytechnic Institute of New York University From 1997 to 2000 Mr. Hultin served as Under Secretary of the Navy., Dennis Cutler Blair is the former United States Director of National Intelligence and is a retired United States Navy admiral who was the commander of U.S. forces in the Pacific region, http://engineering.nyu.edu/files/hultin%20naval%20power.pdf) Technology has fundamentally changed the relationship between national naval power ¶ and economic prosperity from the simplistic mercantilist tenets of Alfred Thayer Mahan almost ¶ two hundred years ago. No longer does trade simply follow the flag, no longer is a strong ¶ national Navy required to support strong national trade relationships. With the economies of ¶ many countries interdependent in sectors such as energy and raw materials, and with ¶ globalization increasing interdependence in manufacturing and services, military security is ¶ becoming increasingly a shared responsibility. 1NC—Alt Causes Alt causes to navy power decline that IOOS can’t solve for Hultin & Blair 6(Jerry MacArthur HULTIN and Admiral Dennis BLAIR, Naval Power and Globalization: ¶ The Next Twenty Years in the Pacific, Jerry M. Hultin is former president of Polytechnic Institute of New York University From 1997 to 2000 Mr. Hultin served as Under Secretary of the Navy., Dennis Cutler Blair is the former United States Director of National Intelligence and is a retired United States Navy admiral who was the commander of U.S. forces in the Pacific region, http://engineering.nyu.edu/files/hultin%20naval%20power.pdf) As successful as this transformation has been, there are new technological, geo-political, ¶ and warfighting challenges on the horizon that require solutions. The three most noteworthy are ¶ missile defense, irregular warfare – peace operations, counter-insurgency warfare, stability ¶ operations and military operations against non-state organizations using terrorist tactics – and the ¶ integration of all factors of political-military influence. ¶ ¶ First, missile defense: in the future increasing numbers of littoral countries, organizations ¶ such as Hezbollah that are sponsored by countries, and even unsponsored militant organizations ¶ will field ballistic and cruise missiles up to intermediate ranges. These weapons will threaten ¶ deployed US naval expeditionary forces, military forces and civilian populations of U.S. allies, ¶ and ultimately the United States itself. Thus, it is imperative that the United States and its allies ¶ develop and deploy a highly effective capacity to defend against such attacks8¶ . There is not a ¶ single system solution to this threat – the Navy must continue developing more capable sensors, ¶ communications networks, offensive weapons to take out enemy missiles and their support ¶ systems, along with both hard-kill and softkill defensive systems that can defeat enemy missiles ¶ once they have been launched. No uq to undersea warfare – shortage of subs and china military modernization Eaglen and Rodeback 2010 Mackenzie Eaglen Research Fellow for National Security Studies, Allison Center for Foreign Policy Studies Jon Rodeback Editor, Research Publications Heritage Submarine Arms Race in the Pacific: The Chinese Challenge to U.S. Undersea Supremacy http://www.heritage.org/research/reports/2010/02/submarinearms-race-in-the-pacific-the-chinese-challenge-to-us-undersea-supremacy The People's Republic of China (PRC) is rapidly emerging as a regional naval power and a potential global power, which "raises the prospect of intensifying security competition in East Asia both between the United States and China and between China and Japan."[2] Other Pacific countries in the region have also taken note of the changing security environment as evidenced in particular by their planned submarine acquisitions. Australia's military buildup includes doubling its submarine fleet from six submarines to 12 larger, more capable submarines.[3] In addition, "India, Malaysia, Pakistan, Indonesia, Singapore, Bangladesh and South Korea are planning to acquire modern, conventional submarines."[4] Both Australia and India have explicitly described their naval buildups as responses, at least in part, to China's naval buildup. In contrast, the U.S. submarine fleet is projected to continue shrinking through 2028, further limiting the U.S. ability to shape and influence events in the Pacific. The U.S. attack submarine serves an important part in establishing sea control and supremacy, and it is not interchangeable with other assets. Its unique capabilities make it a force multiplier and allow it to "punch above its weight." To protect U.S. interests in East Asia and the Pacific, and to support and reassure U.S. allies, the U.S. must halt and then reverse the decline of its submarine fleet as a critical component of a broader policy to maintain the military balance in the Pacific. Budget cuts destroy readiness Ron Ault 13, Executive Assistant to the President for Metal Trades Department, AFL-CIO, “Navy Budget Cuts Sinking America’s Military Readiness,” Feb 15 2013, http://www.metaltrades.org/index.cfm?zone=/unionactive/view_blog_post.cfm&blogID=1390&postID= 6618 While many in Washington are focused on the rapidly-approaching March 1st sequestration deadline, few seem to be paying much attention to the massive cuts that will go into effect this week across our military – cuts that will have an equally devastating impact on our economy and security.¶ As the result of the continuing resolution (CR) that was agreed upon in the final hours of 2012, a deal that averted the so-called fiscal cliff, the U.S. Navy was forced to accept $6.3 billion in cuts for the remainder of the force’s Fiscal Year 2013 (FY13) budget. If sequestration goes into effect, that number could grow to $10 billion.¶ As the representative of 80,000 workers employed at U.S. shipyards and naval bases – workers who take great pride in the role they play in supporting our military – I would like to shed some light on the severity and senselessness of the Navy budget cuts and implore the 113th Congress to take prudent and expeditious action to avoid further Defense spending cuts in FY13. ¶ Lawmakers have already taken a hatchet to the Defense budget; we simply cannot sustain any further cuts this year.¶ The $6.3 billion in Navy budget reductions cut too deeply . For example, Navy Fleet Commanders have been ordered to cancel 3rd and 4th Quarter surface ship maintenance. That’s a very tall order considering our shipyards are already fully loaded with high priority work on extremely tight schedules to support fleet deployment. Once scheduled maintenance on a ship is missed, there is no capacity to make it up later without causing delays elsewhere. The well-oiled machine breaks down. With the cuts contained in the CR alone, it is conceivable that aircraft carriers, destroyers, nuclear submarines, and other Navy vessels would be tied up at piers awaiting critical repairs without the ability to move , much less support our national defense. If we allow our fleet to collect barnacles in harbor for six months, we would significantly disrupt our military’s access to the ships they need to defend our country. Specifically wrecks undersea dominance Hugh Lessig 13, Daily Press, 9/12/13, “U.S. Navy: Budget challenges threaten submarine program,” http://articles.dailypress.com/2013-09-12/news/dp-nws-forbes-submarine-hearing-20130912_1_u-snavy-submarines-missile The U.S. military touts its submarine program as an unqualified success, yet the fleet is expected to drop by nearly 30 percent in coming years and one Hampton Roads lawmaker wants to pull money from outside the Navy's shipbuilding program to ease the pressure on one key program.¶ Two Navy leaders on Thursday told a panel chaired by Rep. Randy Forbes, RChesapeake, that more budget cuts would dull the edge of the submarine program , where the U.S. enjoys a distinct advantage over other superpowers. Forbes chairs the House Armed Service's subcommittee on sea power.¶ About 2,300 jobs at Newport News Shipbuilding are tied to the U.S. submarine program. The shipyard builds them in partnership with General Dynamics Electric Boat of Groton, Conn.¶ The U.S. military has three types of nuclear-powered submarines. First are the smaller fastattack submarines that fall primarily in two classes, the older Los Angeles class and the newer Virginia class. Last week, the Navy commissioned the newest Virginia class sub at Naval Station Norfolk.¶ The second type are Ohio class ballistic missile submarines that roam the seas and provide a nuclear strike capability.¶ The third type is an offshoot of the second: When the Cold War ended, the U.S. converted four of those ballistic missile submarines into guided-cruise missile submarines.¶ All three types are scheduled to drop , said Rear Adm. Richard P. Breckenridge and Rear Adm. David C. Johnson. who testified before the Forbes panel. The huge national debt means cuts will continue---guarantees an ineffective navy Seth Cropsey 10, senior fellow at the Hudson Institute, former Deputy Undersecretary of the Navy, “The US Navy in Distress,” Strategic Analysis Vol. 34, No. 1, January 2010, 35–45, http://navsci.berkeley.edu/ma20/Class%208/Class%2012.%20Sea%20Power.%20%20Cropsey_US_Navy _In_Distress.pdf The most tangible result is the continued withering of the US combat fleet which today numbers about 280 ships. This is less than half the size achieved towards the end of the Reagan administration buildup and 33 ships short of what the Navy says it needs to fulfill today’s commitments. Nothing suggests a substantive reversal . Most signs point to additional decline over the long term. Four years ago the Navy’s projected fleet size had dwindled to 313 ships. There it has stayed . . . until May of this year when a senior Navy budget official, commenting on the proposed 2010 budget, suggested that the new Quadrennial Review now underway at the Defense Department will likely result in a smaller projected fleet size. Huge increases in current and projected national debt and the vulnerability of the military budget to help offset it increase the chance that without compelling events the nation’s sea services will experience additional and perhaps drastic reductions . National indebtedness will grow from its current ratio of 40 per cent of GDP to 80 per cent of GDP in a decade. Servicing this will cripple the nation’s ability to modernize and increase a powerful world-class fleet or drive us deeper into a yawning financial hole. 1NC—No War (Escalation) No escalation to great power war--empirics Mueller 11(John, “Embracing threatlessness: ¶ Reassessing U.S. military spending, John Mueller is Adjunct Professor of Political Science and Senior Research Scientist at the Mershon Center for International Security Studies., https://cna.org/sites/default/files/research/American%20Grand%20Strategy%20and%20Seapower%202011%20Conf erence%20Report%20CNA.pdf) A sensible place to begin the consideration is with an examination of the prospects for a ¶ major war like World War II. ¶ Although there is no physical reason why such a war cannot recur, it has become fairly ¶ commonplace to regard such wars as obsolescent, if not completely obsolete.4¶ Leading or ¶ developed countries continue to have disputes, but, reversing the course of several millennia, ¶ none seems likely seriously to envision war as a sensible method for resolving any of these ¶ disputes. ¶ Europe, once the most warlike of continents, has taken the lead in this. It was on May 15, ¶ 1984, that the major countries of the developed world had managed to remain at peace with ¶ each other for the longest continuous stretch since the days of the Roman Empire.5¶ That ¶ rather amazing record has now been further extended, and today one has to go back more ¶ than two millennia to find a longer period in which the Rhine remained uncrossed by armies ¶ with hostile intent.6¶ ¶ “All historians agree,” observed Leo Tolstoy in War and Peace in 1869, “that states express ¶ their conflicts in wars and that as a direct result of greater or lesser success in war the political ¶ strength of states and nations increases or decreases.”7¶ Whatever historians may currently ¶ think, it certainly appears that this notion has become substantially obsolete. Prestige now ¶ 4¶ For an early examination of this proposition, see John Mueller, Retreat from Doomsday: The Obsolescence ¶ of Major War (New York: Free Press, 1989; reprinted and updated edition, edupublisher.com, 2009). ¶ See also Fettweis, Dangerous Times. ¶ 5¶ Paul Schroeder, “Does Murphy’s Law Apply to History?” Wilson Quarterly, New Year’s 1985: 88. The ¶ previous record, he notes, was chalked up during the period from the end of the Napoleonic Wars ¶ in 1815 to the effective beginning of the Crimean War in 1854. The period between the conclusion ¶ of the Franco-Prussian War in 1871 and the outbreak of World War I in 1914 – marred by a major ¶ war in Asia between Russia and Japan in 1904 – was an even longer era of peace among major ¶ European countries. That record was broken on November 8, 1988. On some of these issues, see ¶ also Evan Luard, War in International Society (New Haven, CT: Yale University Press, 1986), 395-99; ¶ and James J. Sheehan, Where Have All the Soldiers Gone? The Transformation of Modern Europe (Boston: ¶ Houghton Mifflin, 2008). ¶ 6¶ Bradford de Long, “Let Us Give Thanks (Wacht am Rhein Department), November 12, 2004, ¶ www.j-bradford-delong.net/movable_type/2004-2_arcives/000536.html. ¶ 7¶ (New York: Norton, 1966), 1145. 168¶ 49¶ comes from other factors, such as making economic progress and putting on a good ¶ Olympics. ¶ The Cold War did supply a set of crises and peripheral wars that engaged leading countries ¶ from time to time, and it was commonly envisioned that doom would inevitably emerge from ¶ the rivalry. Thus, political scientist Hans J. Morgenthau in 1979 said: “In my opinion the ¶ world is moving ineluctably towards a third world war – a strategic nuclear war. I do not ¶ believe that anything can be done to prevent it. The international system is simply too ¶ unstable to survive for long.”8¶ At about the same time, John Hackett penned his distinctly ¶ non-prescient book The Third World War: August 1985.¶ 9¶ Such anxieties obviously proved to be ¶ over-wrought, but to the degree that they were correctly focused on a potential cause of ¶ major war, that specific impetus no longer exists. ¶ World War III, then, continues to be the greatest nonevent in human history, and that happy ¶ condition seems very likely to continue. There have been wars throughout history, of course, ¶ but the remarkable absence of the species’ worst expression for twothirds of a century (and ¶ counting) strongly suggests that realities may have changed, and perhaps permanently. ¶ Accordingly it may be time to consider that spending a lot of money preparing for a ¶ “conceivable” eventuality – or fantasy – that is of ever-receding likelihood is a highly ¶ questionable undertaking. 1NC – South China Sea No South China conflict-engagement will check miscalc and mistrust Thayer, New South Wales emeritus professor, 2013 (Carlyle, “Why China and the US won’t go to war over the South China Sea”, 5-13, http://www.eastasiaforum.org/2013/05/13/why-china-and-the-us-wont-go-to-war-over-the-southchina-sea/, ldg) China’s increasing assertiveness in the South China Sea is challenging US primacy in the Asia Pacific. Even before Washington announced its official policy of rebalancing its force posture to the Asia Pacific, the United States had undertaken steps to strengthen its military posture by deploying more nuclear attack submarines to the region and negotiating arrangements with Australia to rotate Marines through Darwin.Since then, the United States has deployed Combat Littoral Ships to Singapore and is negotiating new arrangements for greater military access to the Philippines. But these developments do not presage armed conflict between China and the United States. The People’s Liberation Army Navy has been circumspect in its involvement in South China Sea territorial disputes, and the United States has been careful to avoid being entrapped by regional allies in their territorial disputes with China. Armed conflict between appears unlikely . Another, more probable, scenario is that both countries will find a modus vivendi enabling them to collaborate to maintain security in the South China Sea. The Obama administration has repeatedly emphasised that its policy of rebalancing to Asia is not directed at containing China. For example, Admiral Samuel China and the United States in the South China Sea J. Locklear III, Commander of the US Pacific Command, recently stated, ‘there has also been criticism that the Rebalance is a strategy of containment. This is not the case … it is a strategy of collaboration and cooperation’. However, a review of past US–China military-to-military interaction indicates that an agreement to jointly manage security in the South China Sea is unlikely because of continuing strategic mistrust between the two countries. This is also because the currents of regionalism are growing stronger. As such, a third scenario is more likely than the previous two: that China and the United States will maintain a relationship of cooperation and friction. In this scenario, both countries work separately to secure their interests through multilateral institutions such as the East Asia Summit, the ASEAN Defence Ministers’ Meeting Plus and the Enlarged ASEAN Maritime Forum. But they also continue to engage each other on points of mutual interest. The Pentagon has consistently sought to keep channels of communication open with China through three established bilateral mechanisms: Defense Consultative Talks, the Military Maritime Consultative Agreement (MMCA), and the Defense Policy Coordination Talks. On the one hand, these multilateral mechanisms reveal very little about US–China military relations. Military-to-military contacts between the two countries have gone through repeated cycles of cooperation and suspension, meaning that it has not been possible to isolate purely military-to-military contacts from their political and strategic settings. On the other hand, the channels have accomplished the following: continuing exchange visits by high-level defence officials; regular Defense Consultation Talks; continuing working-level discussions under the MMCA; agreement on the ‘7-point consensus’; and no serious naval incidents since the 2009 USNS Impeccable affair. They have also helped to ensure continuing exchange visits by senior military officers; the initiation of a Strategic Security Dialogue as part of the ministerial-level Strategic & Economic Dialogue process; agreement to hold meetings between coast guards; and agreement on a new working group to draft principles to establish a framework for military-to-military cooperation. So the bottom line is that, despite ongoing frictions in their relationship, the United States and China will continue engaging with each other. Both sides understand that military-to-military contacts are a critical component of bilateral engagement. Without such interaction there is a risk that mistrust between the two militaries could spill over and have a major negative impact on bilateral relations in general. But strategic mistrust will probably persist in the absence of greater transparency in military-to-military relations. In sum, Sino-American relations in the South China Sea are more likely to be characterised by cooperation and friction than scenario, armed conflict. a modus vivendi of collaboration or, a worst-case 2NC – South China Sea No south china sea conflict-china would never engage Carlson, Cornell government professor, 2013 (Allen, “China Keeps the Peace at Sea”, 2-21, http://www.foreignaffairs.com/articles/139024/allencarlson/china-keeps-the-peace-at-sea?page=show, ldg) The fundamentals of Deng's grand economic strategy are still revered in Beijing. But any war in the region would erode the hard-won, and precariously held, political capital that China has gained in the last several decades. It would also disrupt trade relations, complicate efforts to promote the yuan as an international currency, and send shock waves through the country's economic system at a time when it can ill afford them. There is thus little reason to think that China is readying for war with Japan. At the same time, the specter of rising Chinese nationalism, although often seen as a promoter of conflict, further limits the prospects for armed engagement. This is because Beijing will try to discourage nationalism if it fears it may lose control or be forced by popular sentiment to take an action it deems unwise. Ever since the Tiananmen Square massacre put questions about the Chinese Communist Party's right to govern before the population, successive generations of Chinese leaders have carefully negotiated a balance between promoting nationalist sentiment and preventing it from boiling over. In the process, they cemented the legitimacy of their rule. A war with Japan could easily upset that balance by inflaming nationalism that could blow back against China's leaders. Consider a hypothetical scenario in which a uniformed Chinese military member is killed during a firefight with Japanese soldiers. Regardless of the specific circumstances, the casualty would create a new martyr in China and, almost as quickly, catalyze popular protests against Japan. Demonstrators would call for blood, and if the government (fearing economic instability) did not extract enough, citizens would agitate against Beijing itself. Those in Zhongnanhai, the Chinese leadership compound in Beijing, would find themselves between a rock and a hard place. It is possible that Xi lost track of these basic facts during the fanfare of his rise to power and in the face of renewed Japanese assertiveness. It is also possible that the Chinese state is more rotten at the core than is understood. That is, party elites believe that a diversionary war is the only way to hold on to power -- damn the economic and social consequences. But Xi does not seem blind to the principles that have served Beijing so well over the last few decades. Indeed, although he recently warned unnamed others about infringing upon China's "national core interests" during a foreign policy speech to members of the Politburo, he also underscored China's commitment to "never pursue development at the cost of sacrificing other country's interests" and to never "benefit ourselves at others' expense or do harm to any neighbor." Of course, wars do happen -- and still could in the East China Sea. Should either side draw first blood through accident or an unexpected move, Sino-Japanese relations would be pushed into terrain that has not been charted since the middle of the last century. However, understanding that war would be a no-win situation, China has avoided rushing over the brink. This relative restraint seems to have surprised everyone. But it shouldn't. Beijing will continue to disagree with Tokyo over the sovereign status of the islands, and will not budge in its negotiating position over disputed territory. However, it cannot take the risk of going to war over a few rocks in the sea. On the contrary, in the coming months it will quietly seek a way to shelve the dispute in return for securing regional stability, facilitating economic development, and keeping a lid on the Pandora's box of rising nationalist sentiment. The ensuing peace, while unlikely to be deep, or especially conducive to improving Sino-Japanese relations, will be enduring. China won’t get in a naval conflict Holmes, Naval War College strategy professor, 2012 (James, “The Sino-Japanese Naval War of 2012”, 8-20, http://www.foreignpolicy.com/articles/2012/08/20/the_sino_japanese_naval_war_of_2012?page=0,1, DOA: 10-16-12, ldg) Whoever forges sea, land, and air forces into the sharpest weapon of sea combat stands a good chance of prevailing. That could be Japan if its political and military leaders think creatively, procure the right hardware, and arrange it on the map for maximum effect. After all, Japan doesn’t need to defeat China’s military in order to win a showdown at sea, because it already holds the contested real estate; all it needs to do is deny China access. If Northeast Asian seas became a no-man’s land but Japanese forces hung on, the political victory would be Tokyo’s. Japan also enjoys the luxury of concentrating its forces at home, whereas the PLA Navy is dispersed into three fleets spread along China’s lengthy coastline. Chinese commanders face a dilemma: If they concentrate forces to amass numerical superiority during hostilities with risk leaving other interests uncovered. It would hazardous for Beijing to leave, say, the South China Sea unguarded during a conflict in the northeast. And finally, Chinese leaders would be forced to consider how far a marine war would set back their sea-power project. China has staked its economic and diplomatic future in large part on a powerful oceangoing navy. In December 2006, President Hu Jintao ordered PLA commanders to construct “a Japan, they powerful people’s navy” that could defend the nation’s maritime lifelines — in particular sea lanes that connect Indian Ocean energy exporters with users in China takes lots of ships. If it lost much of the fleet in a Sino-Japanese clash — even in a winning effort — Beijing could see its momentum toward world-power status reversed in an afternoon. Here’s hoping China’s — “at any time.” That political and military leaders understand all this. If so, the Great Sino-Japanese Naval War of 2012 won’t be happening outside these pages. 1NC—No Taiwan War No China-Taiwan conflict—Chinese democratization and economic interdependence Mueller 11(John, “Embracing threatlessness: ¶ Reassessing U.S. military spending, John Mueller is Adjunct Professor of Political Science and Senior Research Scientist at the Mershon Center for International Security Studies., https://cna.org/sites/default/files/research/American%20Grand%20Strategy%20and%20Seapower%202011%20Conf erence%20Report%20CNA.pdf) China’s oft-stated desire to incorporate (or re-incorporate) Taiwan into its territory and its ¶ apparent design on other offshore areas do create problems – ones that world leaders ¶ elsewhere should sensibly keep their eyes on, and ones that could “conceivably” lead to an ¶ armed conflict to which American military forces might appear relevant. But it is also ¶ conceivable, and far more likely, that the whole problem will be worked out over the course ¶ of time without armed conflict. The Chinese strongly stress that their perspective on this ¶ issue is very long term and that they have a historic sense of patience. Indeed, if China ¶ eventually becomes a true democracy, Taiwan might even join up voluntarily or, failing that, ¶ some sort of legalistic face-saving agreement might eventually be worked out. Above all, ¶ China is increasing becoming a trading state, in Richard Rosecrance’s phrase.11 Its integration ¶ into the world economy and its increasing dependence on it for economic development and ¶ for the consequent acquiescent contentment of the Chinese people is likely to keep the ¶ country reasonable. Armed conflict over the issue would be extremely – even overwhelmingly ¶ – costly to the country, and, in particular, to the regime in charge, and Chinese leaders seem ¶ to realize this. 2NC—No Taiwan War China would not attack Taiwan—it destroys is position globally and there is no certainty that it would be successful Keck 13(Zachary, December 24, Why China Won't Attack Taiwan, Zachary Keck is Associate Editor of The Diplomat., The Diplomat, http://thediplomat.com/2013/12/why-china-wont-attack-taiwan/) Although the trend lines are undoubtedly working in China’s favor, it is ultimately extremely unlikely that China will try to seize Taiwan by force. Furthermore, should it try to do this, it is unlikely to succeed.¶ Even assuming China’s military capabilities are great enough to prevent the U.S. from intervening, there are two forces that would likely be sufficient to deter China from invading Taiwan. The first and least important is the dramatic impact this would have on how countries in the region and around the world would view such a move. Globally, China seizing Taiwan would result in it being permanently viewed as a malicious nation. Regionally, China’s invasion of Taiwan would diminish any lingering debate over how Beijing will use its growing power. Every regional power would see its own fate in Taiwan. Although Beijing would try to reassure countries by claiming that Taiwan was part of China already, and thus the operation was a domestic stability one, this narrative would be convincing to none of China’s neighbors. Consequently, Beijing would face an environment in which each state was dedicated to cooperating with others to balance against Chinese power.¶ But the more important deterrent for China would be the uncertainty of success. No escalation to China-US war Lowther 14(William, Febuary 20, Independence ‘will not happen’, William Lowther is a staff writer for the Taipei Times, http://www.taipeitimes.com/News/taiwan/archives/2014/02/20/2003583939/1) In a fight over Taiwan, Mearsheimer says, US policymakers would be reluctant to launch major strikes against Chinese forces in China, for fear it might precipitate nuclear escalation.¶ “The US is not going to escalate to the nuclear level if Taiwan is being overrun by China. The stakes are not high enough to risk a general thermonuclear war. Taiwan is not Japan, or even South Korea,” he says.¶ He says Taiwan is an especially dangerous flashpoint, which could easily precipitate a Sino-US war that is not in the US’ interest.¶ US policymakers understand that the fate of Taiwan is a matter of great concern to China and there is a “reasonable chance” US policymakers will eventually conclude that it makes good strategic sense to abandon Taiwan and allow China to coerce it into accepting unification, Mearsheimer says. Homeland Security Terrorism Better detection fails- terrorists can hide their HEU from radiation detectors. de Rugy 05 (Veronique, senior research fellow at the Mercatus Center at George Mason University, “Is Port Security Spending Making Us Safer?,” http://cip.management.dal.ca/publications/Is%20Port%20Security%20Spending%20Making%20Us%20Safer.pdf) Considering the extreme difficulty of interdicting drug smugglers, it seems that determined smugglers with a nuclear device would have little trouble circumventing the nation’s border protection and control, particularly because they would be able to leverage the techniques used successfully by drug smugglers. Further lowering their probability of being caught is the fact that, according to a series of experts testifying 11 before Congress in July 2005, terrorists could easily shield the highly enriched uranium and avoid detection from radiation detectors. 34 Port security fails to solve terrorism a. Other targets besides ports Shie 04, Tamara Renee Shie, [former visiting fellow at the Pacific Forum CSIS in Honolulu, research assistant at the Institute for National Strategic Studies (INSS) at the National Defense University (NDU) in Washington, D.C. ], “Ships and Terrorists – Thinking Beyond Port Security”, http://csis.org/files/media/csis/pubs/pac0445a.pdf Unfortunately, relying on these initiatives alone creates a false sense of security. They are inadequate to deter terrorists from pursuing many maritime targets. The principal limitation of these two initiatives is their specific focus on the security of major transshipment ports. Though these are essential to international trade, securing only these ports will not protect them or a region from terrorist attacks. First, the emphasis on upgrading the security of major ports neglects the fact that these represent only a single link in the transportation chain. A shipping container may pass through some 15 physical locations and some two dozen individuals and/or companies while traveling from departure point to destination. Because containers are only searched at the major port, there is no guarantee they cannot be waylaid in route after that point. Second, the CSI conducts security checks only on U.S. bound containers. Therefore even if a tampered container arrives at a major port, if it is destined for a port other than the U.S., it is more likely to escape notice. Containers between the major ports of Singapore and Shenzhen or Pusan and Hong Kong are not subject to CSI requirements. Yet terrorist assaults on U.S. ships or interests can occur outside the U.S. Third, as major ports increase security, terrorists will look for other maritime targets or other means to target those ports. Terrorists are increasingly aiming at soft targets. Attacking maritime targets has never been particularly easy, often requiring a greater sophistication in planning, training, and coordination than those aimed at many land-based facilities. This is why maritime terrorism is rather rare, and why terrorists are less likely to attack a more secure major port. Yet in considering maritime terrorist threat scenarios – using a ship to smuggle goods or weapons, sinking a vessel in a major shipping thoroughfare, using a ship as a weapon, or even targeting maritime vessels – none require access to a major port or a shipping container to carry out a strike. There remain numerous small ports and small vessels not covered under the new security initiatives. The ISPS Code for instance only covers ships of 500 tons or more and port facilities that serve large international-bound vessels. The Code would not have protected the USS Cole. b. Effective nuclear detection is impossible Rugy 07, Veronique de Rugy, [senior research fellow at the Mercatus Center at George Mason University],“Is Port Security Funding Making Us Safer?”, MIT Center for International Studies, November 2007, http://web.mit.edu/cis/pdf/Audit_11_07_derugy.pdf. The Domestic Nuclear Detection Office (DNDO) received $535 million in 2007.14 DNDO’s mission addresses a broad spectrum of radiological and nuclear protective measures, but is focused exclusively on domestic nuclear detection.15 The fundamental problem is that DNDO relies on radiation portal monitors that have been proven unable to detect shielded nuclear material essentially rendering them useless.16 Besides, even if the system could detect every dangerous item, it is ineffective unless the nuclear material is brought through the fixed ports of entry where the monitors are located. With thousand of miles of unguarded borders—and no cost effective way to address the issue—smugglers can easily find positions to bring illicit goods inside the country. Consider the country’s long standing War on Drugs and the inability to stop the flow of illegal drugs into the country. c. Least cost-effective way Rugy 07, Veronique de Rugy, [senior research fellow at the Mercatus Center at George Mason University],“Is Port Security Funding Making Us Safer?”, MIT Center for International Studies, November 2007, http://web.mit.edu/cis/pdf/Audit_11_07_derugy.pdf. A close look at port security allocation decisions indicates that spending occurs without regard for risk analysis let alone cost-benefit analysis, leading to a large array of misallocated spending. For instance, why should the highest priorities—preventing terrorists from acquiring nuclear devices and material— receive less money than much less cost-effective policies such as nuclear detection in the ports or postdisaster response activities. Because it rests mainly on domestic detection of WMD in ports—a task that is not clear could be achieved—the port security model offers almost no value to the nation.6 Even if we could seal our ports, America wouldn’t be safe. The only effective way to prevent nuclear attacks is to deny terrorists access to weapons and material. Without nuclear materials there can be no nuclear bombs. d. Vulnerability between stops Jon D. Haveman and Howard J. Shatz 2006- Director of the Economy Program at the Public Policy Institute of California and Senior Economist; Professor, Pardee RAND Graduate School (Protecting the Nation’s Seaports: Balancing Security and Cost, Pg 30-31, http://www.ppic.org/content/pubs/report/r_606jhr.pdf) In Chapter 4, Stephen S. Cohen considers the security threat that¶ the container creates for the maritime transportation system. Each day,¶ tens of thousands of containers flow through U.S. ports, largely¶ undisturbed in their trip from one part of the world to another. In¶ general, containers are loaded and sealed, or perhaps only closed and not¶ sealed, well inland of a port. They are then transferred by truck, or truck¶ and train, to a seaport, where they are loaded onto a ship. Following the¶ sea journey, they are transferred to another truck, and perhaps another¶ train, for a further journey over land to the ultimate destination.¶ Each container trip to the United States has, on average, 17 different¶ stops, or points at which the container’s journey temporarily halts.14¶ The adage “goods at rest are goods at risk” readily applies to the terrorist¶ threat. The container will be at rest at any point in the journey that¶ involves a change in mode of transportation. While at rest, the container¶ is vulnerable to thieves and terrorists alike. Providing port security¶ therefore involves closely scrutinizing activities not only at the port but at¶ points all along the shipping chain. The truck driver picking up the¶ container at the U.S. port, often poorly paid and possibly an illegal¶ immigrant not well integrated into U.S. society, may himself represent a¶ vulnerability in the system.¶ The issue is not merely that something could be put in a container¶ illicitly for an attack on the port where it is unloaded but that nuclear weapons or radiological material could be inserted, shipped to the¶ United States, moved inland without inspection, and then unloaded into¶ the hands of terrorists. These objects could then be transported for use¶ in major population centers—perhaps better targets than a port complex.¶ Likewise, explosive material could be put in several containers and then¶ detonated at or near port complexes around the same time, leading to a¶ security reaction that could shut down the entire maritime transportation¶ system until officials, and port workers and management, were certain¶ the threat had passed.¶ There is no way to completely inspect all of the millions of¶ containers entering the United States. They are about as large as a fullsize¶ moving van and are often tightly packed. Inspecting each¶ thoroughly would bring commerce to a halt, exactly the kind of reaction¶ that terrorists hope to generate. The risk of nuclear terrorism is vanishingly small -- terrorists must succeed at each of twenty plus stages -- failing at one means zero risk Mueller, ‘10 [John, Woody Hayes Chair of National Security Studies at the Mershon Center for International Security Studies and a Professor of Political Science at The Ohio State University, A.B. from the University of Chicago, M.A. and Ph.D. -- UCLA, “Atomic Obsession – Nuclear Alarmism from Hiroshima to Al-Qaeda,” Oxford University Press] .Even those who decidedly disagree with such scary-sounding, if somewhat elusive, prognostications about nuclear terrorism often come out seeming like they more or less agree. In his Atomic Bazaar, William Langewiesche spends a great deal of time and effort assessing the process by means of which a terrorist group could come up with a bomb. Unlike Allison—and, for that matter, the considerable bulk of accepted opinion —he concludes that it "remains very, very unlikely . It's a possibility, but unlikely." Also: The best information is that no one has gotten anywhere near this. I mean, if you look carefully and practically at this process, you see that it is an enormous undertaking full of risks for the would-be terrorists. And so far there is no public case, at least known, of any appreciable amount of weapons-grade HEU [highly enriched uranium] disappearing. And that's the first step. If you don't have that, you don't have anything. The first of these bold and unconventional declarations comes from a book discussion telecast in June 2007 on C-SPAN and the second from an interview on National Public Radio. Judgments in the book itself, however, while consistent with such conclusions, are expressed more ambiguously, even coyly: "at the extreme is the possibility, entirely real, that one or two nuclear weapons will pass into the hands of the new stateless guerrillas, the jihad-ists, who offer none of the retaliatory targets that have so far underlain the nuclear peace" or "if a would-be nuclear terrorist calculated the odds, he would have to admit that they are stacked against^ffen," but they are "not impossible."5 The previous chapter arrayed a lengthy set of obstacles confront-: v ,„ ing the would-be atomic terrorist—often making use in the process of Langewlesche's excellent reporting. Those who warn about the likelihood of a terrorist bomb contend that a terrorist group could, if often with great difficulty, surmount each obstacle—that doing so in each case is, in Langewiesche's phrase, "not impossible."6 But it is vital to point out that, while it may be "not impossible" to surmount each individual step, the likelihood that a group the odds are "stacked against" the terrorists, as well as other material, helps us evaluate the many ways such a quest—in his words, "an enormous undertaking full of risks"— could fail. The odds, indeed, are stacked against the terrorists, perhaps massively so. In fact, the likelihood a terrorist group will come up with an atomic bomb seems to be vanishingly small . ARRAYING THE BARRIERS Assuming terrorists have some desire for the bomb (an assumption ques-tioned in the next chapter), fulfillment of that desire is obviously another matter. Even the very alarmed Matthew Bunn and Anthony Wier contend that the atomic terrorists' task "would clearly be among the most difficult types of attack to carry out" or "one of the most difficult missions a terrorist group could could surmount a series of them could quickly approach impossibility. If what are they? Lange-wiesche's discussion, hope to try" But, stresses the CIA's George Tenet, a terrorist atomic bomb is "possible" or "not beyond the realm of possibility." In his excellent discussion of the issue, Michael Levi ably catalogues a wide array of difficulties confronting the would-be atomic terrorist, adroitly points out that "terrorists must succeed at every stage, but the defense needs to succeed only once ," sensibly warns against preoccupation with worst-case scenarios, and pointedly formulates "Murphy's Law of Nuclear Terrorism: What can go wrong might go wrong." Nevertheless, he holds nuclear terrorism to be a "genuine possibility," and concludes that a good defensive strategy can merely "tilt the odds in our favor."7 Accordingly, it might be useful to take a stab at estimating just how "difficult" or "not impossible" the atomic terrorists' task, in aggregate, is— that is, how far from the fringe of the "realm of possibility" it might be, how "genuine" the possibilities are, how tilted the odds actually are. After all, lots of things are "not impossible." It is "not impossible" that those legendary monkeys with typewriters could eventually output Shakespeare.8 Or it is "not impossible"—that is, there is a "genuine possibility"— that a colliding meteor or comet could destroy the earth, that Vladimir Putin or the British could decide one morning to launch a few nuclear weapons at Ohio, that an underwater volcano could erupt to cause a civilization-ending tidal wave, or that Osama bin Laden could convert to Judaism, declare himself to be the Messiah, and fly in a gaggle of mafioso hit men from Rome to have himself publicly crucified.9 As suggested, most discussions of atomic terrorism deal in a rather piecemeal fashion with the subject—focusing separately on individual tasks such as procuring HEU or assembling a device or transporting it. However, as the Gilmore Commission, a special advisory panel to the president and Congress, stresses, setting off a nuclear device capable of producing mass destruction presents "Herculean challenges," requiring that a whole series of steps be accomplished: obtaining enough fissile material, designing a weapon "that will bring that mass together in a tiny fraction of a second" and figuring out some way to deliver the thing. And it emphasizes that these merely constitute "the minimum requirements." If each is not fully met, the result is not simply a less powerful weapon, but one that can't produce any significant nuclear yield at all or can't be delivered.10 Following this perspective, an approach that seems appropriate is to catalogue the barriers that must be overcome by a terrorist group in order to carry out the task of producing, transporting, and then successfully detonating an improvised nuclear device, an explosive that, as Allison acknowledges, would be "large, cumbersome, unsafe, unreliable, unpredictable, and inefficient." Table 13.1 attempts to do this, and it arrays some 20 of these— all of which must be surmounted by the atomic aspirant. Actually, it would be quite possible to come up with a longer list: in the interests of keeping the catalogue of hurdles down to a reasonable number, some of the entries are actually collections of tasks and could be divided into two or three or more. For example, number 5 on the list requires that heisted highly enriched uranium be neither a scam nor part of a sting nor of inadequate quality due to insider incompetence, but this hurdle could as readily be rendered as three separate ones. In contemplating the task before them, woixftlsbe atomic terrorists effectively must go through an exercise that looks much like this. If and when they do so, they are likely to find the prospects daunting and accordingly uninspiring or even terminally dispiriting. " Economy Terrorist attacks have a minimal effect on the economy a. Ports aren’t key and empirics disprove Jon D. Haveman and Howard J. Shatz 2006- Director of the Economy Program at the Public Policy Institute of California and Senior Economist; Professor, Pardee RAND Graduate School (Protecting the Nation’s Seaports: Balancing Security and Cost, Pg 30-31, http://www.ppic.org/content/pubs/report/r_606jhr.pdf) In Chapter 2, Edward E. Leamer and Christopher Thornberg argue¶ that the actual costs of an attack on the Los Angeles–Long Beach port¶ complex may not be as high as many fear. For example, if a port is¶ closed, many shippers will reroute their shipments through other ports.¶ In addition, displaced workers will seek alternative employment. As a result, the economy will adjust. Some output will be lost, but it may be¶ so small in magnitude that it will not reveal itself in data that track¶ national or even regional macroeconomic trends.¶ The authors provide examples of other disruptions that might have¶ caused severe economic damage but did not, such as the terrorist attacks¶ of September 11, 2001. Consumer spending fell immediately after the¶ attacks but then rebounded sharply at the end of 2001, growing at an¶ unprecedented, seasonally adjusted annual rate of 7 percent. Likewise,¶ although retail sales fell immediately after the attacks, they returned to¶ trend in November, only two months later. Some sectors did suffer,¶ most notably the airline industry, which had already been in deep¶ trouble before the end of the technology boom in early 2001. But¶ consumer spending actually increased, suggesting that people reallocated¶ the money that they would have spent on airline travel to other forms of¶ consumption. Similarly, the authors argue that other disruptions such as¶ hurricanes, earthquakes, and even labor disputes at seaports did have¶ immediate negative economic effects but that these effects dissipated¶ quickly as the economy adjusted. The message in this is that most such¶ disruptions lead to business being delayed rather than business being¶ cancelled, which in turn results in much less economic harm than would¶ be expected. b. The economy’s resilience and excess physical capacity checks back Jon D. Haveman and Howard J. Shatz 2006- Director of the Economy Program at the Public Policy Institute of California and Senior Economist; Professor, Pardee RAND Graduate School (Protecting the Nation’s Seaports: Balancing Security and Cost, Pg 67-69, http://www.ppic.org/content/pubs/report/r_606jhr.pdf) A cursory look would seem to portend a dramatic, dangerous¶ scenario, but a closer look at the facts suggests otherwise. From an¶ input-output perspective, a wide variety of holes would be quickly¶ created in the flow of production that would seem to lead to a very sharp¶ downturn in economic activity. But our economy is not a mechanical¶ system; it is an organic self-healing system, much like that of a human¶ being: Large injuries take time to heal, but for the most part they do¶ eventually heal. To continue the analogy, a port attack is only a cut on the arm—quickly healed with little noticeable effect on the day-to-day¶ functioning of the person.¶ Although the ports of Los Angeles and Long Beach certainly¶ represent a primary infrastructure target in the United States, a complete¶ shutdown of the ports is highly unlikely as a direct result of some¶ physical attack. There are two reasons for this: the sheer physical scale of¶ the facilities and the large amount of excess physical capacity (as opposed¶ to human capital capacity) currently in place. As shown in the port map¶ on p. xxiii, the two facilities take up approximately 12 square miles of¶ space in a six-by-four-mile area. The complex is broken into a number¶ of separate yards, each completely controlled by a number of¶ independent, competing major shipping lines, each of which have¶ substantial investment in the physical cranes and equipment on their¶ property. Some of these yards are on Terminal Island, connected to the¶ mainland by three road bridges and a railroad; others are on the¶ mainland itself. There are multiple access points into the area as the map¶ shows, including two highways. Even if these roads were shut down, it¶ would be relatively simple to construct a temporary bridge to the island,¶ and although it might have some implications for the movement of¶ ships, no yard would be effectively isolated.3¶ Conventional weapons would be able to damage, at best, only a¶ small portion of the complex, and would be unable to isolate a¶ substantial portion of the port given the multiple access routes into and¶ out of the area. Even a so-called “dirty bomb” could cover only one or¶ two square miles of area with radioactivity. Given the location on the¶ water, winds would quickly blow most of the radioactive materials away,¶ leaving even most of the initially affected area quickly reusable. The only¶ known weapon that could take out an area of this size for an extended¶ period of time would be a nuclear weapon. It seems more likely that the¶ target of such a horrific device would be a densely populated area, not a¶ port. Terrorism doesn’t affect growth Haveman and Shatz 06 (Jon and Howard, Public Policy Institute of California, “Protecting the Nation’s Seaports: Balancing Security and Cost” http://www.ppic.org/content/pubs/report/R_606JHR.pdf) September 11 Did Not Cause the 2001 Recession There is a strong tendency to blame too many secondary effects on disasters. A good example of this phenomenon is found in the September 11 attacks on New York and Washington, D.C. In the days after the attacks, the rhetoric regarding the potential effect on the national economy was both loud and wrong. The theory proposed by many analysts and journalists was that psychologically fragile consumers in the United States would suffer a crisis and stop spending, driving the economy into a deeper recession. Support for this theory came from the first Gulf War, which supposedly caused a similar consumer crisis of confidence that in turn drove us into a recession in 1990. For example, the Wall Street Journal reported on September 13, 2001: “Past shocks to America’s sense of security, such as the Oklahoma City bombing or the Gulf War, have prompted consumers to pull back temporarily on major purchases and other discretionary spending,” said Richard Curtin, director of surveys of consumers at the University of Michigan. He expects a similar reaction now, which could mean a rough time in the next several weeks for the economy, which was already struggling with rising jobless rates and high consumer debt burdens. “We were teetering on the edge, and this might well push us over,” said Mr. Curtin. This hypothesis ignores the facts and completely overstates the psychological fragility of American consumers. The 1990 recession was not caused by the first Gulf War at all. Residential investment and expenditures on consumer durables typically are leading indicators of the economy. When spending on these items begins to fall as a percentage of gross domestic product (GDP), this is a strong indication of an underlying weakness in the economy that will create a recession. Expenditures in these two sectors had dropped from 14 percent of GDP to below 12 percent of GDP in the three years preceding the 1990 downturn—and before the Gulf war. There has never been such a drop that did not eventually lead to a recession, with one exception—in 1967, when the economy was wobbling, appearing to be on the verge of recession, the sharp increase in spending for the Vietnam War propelled the economy forward. This was just the reverse of what Mr. Curtin suggested. Economic growth causes extinction – we should allow the collapse to occur Barry 08 [Dr. Glen Barry – PhD in Land Resources from Wisconsin-Madison University, January 14, 2008, “Economic Collapse and Global Ecology”, http://www.countercurrents.org/barry140108.htm] Bright greens take the continued existence of a habitable Earth with viable, sustainable populations of all species including humans as the ultimate truth and the meaning of life. Whether this is possible in a time of economic collapse is crucially dependent upon whether enough ecosystems and resources remain post collapse to allow humanity to recover and reconstitute sustainable, relocalized societies.It may be better for the Earth and humanity's future that economic collapse comes sooner rather than later, while more ecosystems and opportunities to return to nature's fold exist. Economic collapse will be deeply wrenching -- part Great Depression, part African famine. There will be starvation and civil strife, and a long period of suffering and turmoil.Many will be killed as balance returns to the Earth. Most people have forgotten how to grow food and that their identity is more than what they own. Yet there is some justice, in that those who have lived most lightly upon the land will have an easier time of it, even as those super-consumers living in massive cities finally learn where their food comes from and that ecology is the meaning of life. Economic collapse now means humanity and the Earth ultimately survive to prosper again.¶ Human suffering -- already the norm for many, but hitting the currently materially affluent -- is inevitable given the degree to which the planet's carrying capacity has been exceeded. We are a couple decades at most away from societal strife of a much greater magnitude as the Earth's biosphere fails. Humanity can take the bitter medicine now, and recover while emerging better for it; or our total collapse can be a final, fatal death swoon. Economic decline doesn’t cause war Tir 10 [Jaroslav Tir - Ph.D. in Political Science, University of Illinois at Urbana-Champaign and is an Associate Professor in the Department of International Affairs at the University of Georgia, “Territorial Diversion: Diversionary Theory of War and Territorial Conflict”, The Journal of Politics, 2010, Volume 72: 413-425] Empirical support for the economic growth rate is much weaker. The finding that poor economic performance is associated with a higher likelihood of territorial conflict initiation is significant only in Models 3–4.14 The weak results are not altogether surprising given the findings from prior literature. In accordance with the insignificant relationships of Models 1–2 and 5–6, Ostrom and Job (1986), for example, note that the likelihood that a U.S. President will use force is uncertain, as the bad economy might create incentives both to divert the public’s attention with a foreign adventure and to focus on solving the economic problem, thus reducing the inclination to act abroad. Similarly, Fordham (1998a, 1998b), DeRouen (1995), and Gowa (1998) find no relation between a poor economy and U.S. use of force. Furthermore, Leeds and Davis (1997) conclude that the conflict-initiating behavior of 18 industrialized democracies is unrelated to economic conditions as do Pickering and Kisangani (2005) and Russett and Oneal (2001) in global studies. In contrast and more in line with my findings of a significant relationship (in Models 3–4), Hess and Orphanides (1995), for example, argue that economic recessions are linked with forceful action by an incumbent U.S. president. Furthermore, Fordham’s (2002) revision of Gowa’s (1998) analysis shows some effect of a bad economy and DeRouen and Peake (2002) report that U.S. use of force diverts the public’s attention from a poor economy. Among cross-national studies, Oneal and Russett (1997) report that slow growth increases the incidence of militarized disputes, as does Russett (1990)—but only for the United States; slow growth does not affect the behavior of other countries. Kisangani and Pickering (2007) report some significant associations, but they are sensitive to model specification, while Tir and Jasinski (2008) find a clearer link between economic underperformance and increased attacks on domestic ethnic minorities. While none of these works has focused on territorial diversions, my own inconsistent findings for economic growth fit well with the mixed results reported in the literature.15 Hypothesis 1 thus receives strong support via the unpopularity variable but only weak support via the economic growth variable. These results suggest that embattled leaders are much more likely to respond with territorial diversions to direct signs of their unpopularity (e.g., strikes, protests, riots) than to general background conditions such as economic malaise. Presumably, protesters can be distracted via territorial diversions while fixing the economy would take a more concerted and prolonged policy effort. Bad economic conditions seem to motivate only the most serious, fatal territorial confrontations. This implies that leaders may be reserving the most high-profile and risky diversions for the times when they are the most desperate, that is when their power is threatened both by signs of discontent with their rule and by more systemic problems plaguing the country (i.e., an underperforming economy). No impact to economic decline D. Scott Bennett and Timothy Nordstrom, February 2k. Department of Political Science Professors at Pennsylvania State. “Foreign Policy Substitutability and Internal Economic Problems in Enduring Rivalries,” Journal of Conflict Resolution, Ebsco. In this analysis, we focus on using economic conditions to understand when rivalries are likely to escalate or end. Rivalries are an appropriate set of cases to use when examining substitutability both because leaders in rival states have clearly substitutable choices and because rivalries are a set of cases in which externalization is a particularly plausible policy option.7 In particular, when confronted with domestic problems, leaders in a rivalry have the clear alternatives of escalating the conflict with the rival to divert attention or to work to settle the rivalry as a means of freeing up a substantial amount of resources that can be directed toward solving internal problems. In the case of the diversion option, rivals provide logical, believable actors for leaders to target; the presence of a clear rival may offer unstable elites a particularly inviting target for hostile statements or actual conflict as necessary. The public and relevant elites already consider the rival a threat or else the rivalry would not have continued for an extended period; the presence of disputed issues also provides a casus belli with the rival that is always present. Rivals also may provide a target where the possible costs and risks of externalization are relatively controlled. If the goal is diversion, leaders willwant to divert attention without provoking an actual (and expensive)war. Over the course of many confrontations, rival states may learn to anticipate response patterns, leading to safer disputes or at least to leaders believing that they can control the risks of conflict when they initiate a new confrontation. In sum, rivals provide good targets for domestically challenged political leaders. This leads to our first hypothesis, which is as follows: Hypothesis 1: Poor economic conditions lead to diversionary actions against the rival. Conflict settlement is also a distinct route to dealing with internal problems that leaders in rivalries may pursue when faced with internal problems. Military competition between states requires large amounts of resources, and rivals require even more attention. Leaders may choose to negotiate a settlement that ends a rivalry to free up important resources that may be reallocated to the domestic economy. In a “guns versus butter” world of economic trade-offs, when a state can no longer afford to pay the expenses associated with competition in a rivalry, it is quite rational for leaders to reduce costs by ending a rivalry. This gain (a peace dividend) could be achieved at any time by ending a rivalry. However, such a gain is likely to be most important and attractive to leaders when internal conditions are bad and the leader is seeking ways to alleviate active problems. Support for policy change away from continued rivalry is more likely to develop when the economic situation sours and elites and masses are looking for ways to improve a worsening situation. It is at these times that the pressure to cut military investment will be greatest and that state leaders will be forced to recognize the difficulty of continuing to pay for a rivalry. Among other things, this argument also encompasses the view that the cold war ended because the Union of Soviet Socialist Republics could no longer compete economically with the United States. Hypothesis 2: Poor economic conditions increase the probability of rivalry termination. Hypotheses 1 and 2 posit opposite behaviors in response to a single cause (internal economic problems). As such, they demand a research design that can account for substitutability between them. Water Water Wars-Solvency Scarcity drives middle eastern cooperation Sikimic 11 [Simona, author for the daily star, Lebanon news. Jan. 21, 2011, “Water scarcity can be source for regional cooperation: report” http://www.dailystar.com.lb/News/Lebanon-News/2011/Jan-21/61157-water-scarcity-can-be-source-for-regional-cooperation-report.ashx#axzz36jugkuUs //jweideman] BEIRUT: Water scarcity in the region can be channeled for a common good and used to reduce, rather than ignite conflict, an environmental report released Thursday has claimed. “The Blue Peace: Rethinking Middle East Water” launched at the Beirut-based Carnegie Middle East Center, proposes radical cooperation between the six concerned states: Lebanon, Israel, the Palestinian Territories, Jordan, Iraq and Turkey, and envisages the neighbors setting up a mutual monitoring system to guarantee collaboration and more equal partitioning of resources. “The social and economic development of nations depends on water availability in terms of quantity and quality,” said Fadi Comair, the president of Mediterranean Network of River Basin Organizations. “It holds a major place on the diplomatic agenda of the [six] governments,” Comair said. River flows in Turkey, Syria, Iraq, Lebanon and Jordan have been depleted by 50 to 90 percent in the last 50 years alone, while the vital Jordan River, which acts as a water source for five of the concerned countries, has decreased its discharge by over 90 percent from 1960, the report said. “This is a serious problem,” said Sundeep Waslekar, the president of the India-based think tank Strategic Foresight Group, which coordinated the report’s compilation. “Especially as water demand is rising and consumption has gone up from [an estimated] 10-15 percent 50 years ago to 37 percent now.” With consumer requirements predicted to increase to 50-60 percent over the next decade, further pressure will be put on ever-dwindling supplies, he said. But “hydrodiplomacy” as outlined in the report has the potential to alleviate “conflicts on trans-boundary watercourse between riparian states [which] will intensify more and more, especially in the Middle East,” Comair added. Some moves toward cooperation have already been made, especially between Syria and Lebanon, which have signed two important accords to partition resources from the Orontes and Nahr al-Kabir rivers. Trade No Solvency-Alt cause Too many alt causes – trade is declining in the SQ IEC 09 (The International Economy is a specialized quarterly magazine covering global financial policy, economic trends, and international trade. Spring of 2009. “Collapse in World Trade” from The Magazine of International Economic Policy. http://www.international-economy.com/TIE_Sp09_WorldTrade.pdf July 8, 2014) The collapse of trade since the summer of 2008 has¶ been absolutely terrifying, more so insofar as we lack¶ an adequate understanding of its causes. Murky protectionism has played a role. Disruptions to the supply of¶ trade credit from international banks in particular have¶ negatively impacted some countries. The most important¶ factor is probably the growth of global supply chains,¶ which has magnified the impact of declining final demand¶ on trade. When a U.S. household decides not to buy a¶ $40,000 Cayenne sport utility vehicle from Germany,¶ German exports to the United States go down by $40,000,¶ but Slovakian exports to Germany go down by perhaps¶ half that amount, since while the final assembly is done in¶ Leipzig, the coachwork is done in Bratislava. All this said, it really is the case that we don’t fully understand the rel-¶ ative importance of the effects.¶ If it is any consolation, there are signs that trade will¶ rise with recovery every bit as fast as it fell with the onset¶ of the crisis. Too many other reasons that prohibit economic growth. IEC 09 (The International Economy is a specialized quarterly magazine covering global financial policy, economic trends, and international trade. Spring of 2009. “Collapse in World Trade” from The Magazine of International Economic Policy. http://www.international-economy.com/TIE_Sp09_WorldTrade.pdf July 8, 2014) There are several reasons why international trade has collapsed faster than global GDP. First, the decline in manufacturing output has exceeded the decline in real GDP. For example, U.S. industrial pro- duction in the first quarter of 2009 was 14 percent lower than its peak in the fourth quarter of 2007. In contrast, the level of real GDP in the United States in the first quarter of 2009 was only 3.2 percent below its peak. Demand for tradable goods is largely derived from demand for manufactured goods, so the collapse in trade reflects the collapse in manufacturing. Second, the decline in global trade has been exacerbated by business efforts to conserve cash rather than build up inventories to prior levels. Trade flows are being buffeted not only by a general decline in activity, but also by this temporary move to lower inventories. Third, the decline in trade reflects the crisis in finance. Credit plays a critical role in international trade, and the disruption in global credit markets has restricted flows of credit needed to support trade. Finally, the boom in commodities-related trade has been replaced by gloom. Prices for many commodities are falling and ports are packed with imports (think Houston and oil-country tubular goods) for which there is currently limited demand. This new reality is forcing even hyperactive exporters to cut back shipments dramatically. No Trade War Trade war won’t happen. Fletcher 2010 Ian, Senior Economist of the Coalition for a Prosperous America, former Research Fellow at the U.S. Business and Industry Council, The Mythical Concept of Trade War, April 2nd 2010, http://www.huffingtonpost.com/ian-fletcher/the-mythical-concept-of-t_b_523864.html As Americans ponder how to get the U.S. out of its current trade mess, we are constantly warned to do nothing - like impose a tariff to neutralize Chinese currency manipulation - that would trigger a "trade war." Supposedly, no matter how bad our problems with our trading partners get, they are less bad than the spiraling catastrophe that would ensue if we walked a single inch away from our current policy of unilateral free trade. But the curious thing about the concept of trade war is that, unlike actual shooting war, it has no historical precedent . In fact, the reality is that there has never been a significant trade war, "significant" in the sense of having done serious economic damage. All history records are minor skirmishes at best. The standard example free traders give is that America's Smoot-Hawley tariff of 1930 either caused the Great Depression or made it spread around the world. But this canard does not survive serious examination, and has actually been denied by almost every economist who has actually researched the question in depth--a group ranging from Paul Krugman on the left to Milton Friedman on the right. The Depression's cause was monetary. The Fed allowed the money supply to balloon during the late 1920s, piling up in the stock market as a bubble. It then panicked, miscalculated, and let it collapse by a third by 1933, depriving the economy of the liquidity it needed to breathe. Trade had nothing to do with it. As for the charge that Smoot caused the Depression to spread worldwide: it was too small a change to have plausibly so large an effect. For a start, it only applied to about one-third of America's trade: about 1.3 percent of our GDP. Our average tariff on dutiable goods went from 44.6 to 53.2 percent--not a terribly big jump. Tariffs were higher in almost every year from 1821 to 1914. Our tariff went up in 1861, 1864, 1890, and 1922 without producing global depressions, and the recessions of 1873 and 1893 managed to spread worldwide without tariff increases. Neither does the myth of a death spiral of retaliation by foreign nations hold water. According to the official State Department report on this question in 1931: With the exception of discriminations in France, the extent of discrimination against American commerce is very slight...By far the largest number of countries do not discriminate against the commerce of the United States in any way. "Notorious" Smoot-Hawley is a deliberately fabricated myth, plain and simple. There is a basic unresolved paradox at the bottom of the very concept of trade war. If, as free traders insist, free trade is beneficial whether or not one's trading partners reciprocate, then why would any rational nation start one, no matter how provoked? The only way to explain this is to assume that major national governments like the Chinese and the U.S.--governments which, whatever bad things they may have done, have managed to hold nuclear weapons for decades without nuking each other over trivial spats--are not players of realpolitik, but schoolchildren. When the moneymen in Beijing, Tokyo, Berlin, and the other nations currently running trade surpluses against the U.S. start to ponder the financial realpolitik of exaggerated retaliation against the U.S. for any measures we may employ to bring our trade back into balance, they will discover the advantage is with us, not them. Because they are the ones with trade surpluses to lose, not us. So our position of weakness is actually a position of strength. Supposedly, China can suddenly stop buying our Treasury Debt if we rock the boat. But this would immediately reduce the value of the trillion or so they already hold--not to mention destroying, by making their hostility overt, the fragile (and desperatelytended) delusion in the U.S. that America and China are still benign economic "partners" in a win-win economic relationship. At the end of the day, China cannot force us to do anything economically that we don't choose to. America is still a nuclear power. We can--an irresponsible but not impossible scenario--repudiate our debt to them (or stop paying the interest) as the ultimate countermove to anything they might contemplate. More plausibly, we might simply restore the tax on the interest on foreign-held bonds that was repealed in 1984 thanks to Treasury Secretary Donald Regan. A certain amount of back-and-forth token retaliation (and loud squealing) is indeed likely if America starts defending its interests in trade as diligently as our trading partners have been defending theirs, but that's it. After all, the world trading system has survived their trade barriers long enough without collapsing. Won’t be hot wars Erixon and Sally, directors-ECIPE, 10 (Fredrik and Razeen, European Centre for International Political Economy, TRADE, GLOBALISATION AND EMERGING PROTECTIONISM SINCE THE CRISIS, http://www.ecipe.org/media/publication_pdfs/trade-globalisation-and-emerging-protectionism-since-thecrisis.pdf) But in one other important respect the comparison with the 1930s is highly misleading. Then, tit-for-tat trade protection rapidly followed the Wall Street Crash, and the world splintered into warring trade blocs . This has not happened today, and it is unlikely to happen anytime soon. As we will discuss in the next section, crisis-related protectionism today is remarkably restrained . Multilateral trade rules, international policy cooperation and market-led globalisation provide defences against a headlong descent into 1930s-style protectionism. No retaliation Erixon and Sally, directors-ECIPE, 10 (Fredrik and Razeen, European Centre for International Political Economy, TRADE, GLOBALISATION AND EMERGING PROTECTIONISM SINCE THE CRISIS, http://www.ecipe.org/media/publication_pdfs/trade-globalisation-and-emerging-protectionism-since-thecrisis.pdf) Perhaps the biggest surprise is that the world has not hurtled into tit-for-tat protectionism. That spectre was real enough in late 2008, given the scale of growth contraction and deglobalisation. The good news is documented in WTO updates on new trade measures in 2009. New protectionist measures have appeared – what the WTO refers to as “policy slippage” – but they are remarkably mild . They affect a maximum of 1 per cent of world trade in goods, and protectionism in trade in services has not increased noticeably. New protectionism is concentrated in sectors that have long been protected: textiles, clothing, footwear, iron, steel, consumer electronics and agriculture. Trade remedies have increased, as they do in economic downturns. New antidumping (AD) investigations increased by 15 per cent from mid 2008 to mid 2009, and there has been a marked pickup in new investigations for safeguards and countervailing duties. Signs are that trade-defence filings and investigations are mounting quarter-by-quarter, which points to a sharp increase in imposed duties in 2010 and 2011. Nevertheless, new investigations affect a tiny share of world trade. For example, new AD investigations affect 0.4 per cent of the value of US and EU imports. Finally, up to one-third of new trade measures have been liberalising. These include tariff reductions, removal of export restrictions and FDI liberalisation in several developing countries. 18 No Impact-Trade doesn’t solve war Interdependence doesn’t solve war Layne 8 (Christopher Layne is a professor at Texas A&M University’s George H. W. Bush School of Government and Public Service. He is author of The Peace of Illusions: American Grand Strategy from 1940 to the Present (Cornell University Press, 2006), and (with Bradley A. Thayer) American Empire: A Debate (Routledge, 2007). “China’s Challenge to US Hegemony” http://acme.highpoint.edu/~msetzler/IR/IRreadingsbank/chinauscontain.ch08.6.pdf //Donnie) Proponents of engagement have also argued that the United States can help foster political liberalization in China by integrating the country into the international economy and embedding it in the complex web of international institutional arrangements. A China so engaged, it is said, will have strong interests in cooperation and will not be inclined to pursue security competition with America or with its Asian neighbors. Engagement is a problematic strategy, however, because it rests on a shaky foundation. The conventional wisdom notwithstanding, there is little support in the historical record for the idea that economic interdependence leads to peace. After all, Europe never was more interdependent (not only economically but also, among the ruling elites, intellectually and culturally) than before World War I. It was famously predicted, on the eve of World War I, that the economic ties among Europe’s great powers had ushered in an era in which war among them was unthinkable. Yet, as we know, the prospect of forgoing the economic gains of trade did not stop Europe’s great powers from fighting a prolonged and devastating war. Beijing’s actual foreign policy furnishes a concrete reason to be skeptical of the argument that interdependence leads to peace. China’s behavior in the 1996 crisis with Taiwan (during which it conducted missile tests in waters surrounding the island in the run-up to Taiwan’s presidential election) suggested it was not constrained by fears that its muscular foreign policy would adversely affect its overseas trade. Of course, during the past decade, China has been mindful of its stake in international trade and investment. But this does not vindicate the us strategy of engagement. China’s current policy reflects the fact that, for now, Beijing recognizes its strategic interest in preserving peace in East Asia. Stability in the region, and in Sino-American relations, allows China to become richer and to catch up to the United States in relative power. For a state in China’s position vis-à-vis the United States, this is the optimal realpolitik strategy: buying time for its economy to grow so that the nation can openly balance against the United States militarily and establish its own regional hegemony in East Asia. Beijing is pursuing a peaceful policy today in order to strengthen itself to confront the United States tomorrow.