1nc – cloud computing - Millennial Speech & Debate

advertisement
1nc – cloud computing
NSA surveillance doesn’t undermine cloud computing
Henderson, 4/9/15 (Nicole, “Impact of NSA Surveillance on US Cloud Providers Not as Bad as
We Thought: Forrester” 4/9, http://www.thewhir.com/web-hosting-news/impact-nsasurveillance-us-cloud-providers-not-bad-thought-forrester
It’s been two years since Edward Snowden leaked details of the NSA’s PRISM surveillance
program, and although analysts predicted an exodus from US-based cloud and hosting services
in response to the revelations, it hasn’t exactly worked out that way, a new report finds.
Forrester released a new report last week that suggests concerns around international
customers severing ties with US-based hosting and cloud companies “were overblown.”
“Lost revenue from spending on cloud services and platforms comes to just over $500 million
between 2014 and 2016. While significant, these impacts are far less than speculated, as more
companies reported taking control of security and encryption instead of walking away from
US providers,” Forrester’s principal analyst serving security and risk professionals Edward
Ferrara said in a blog post.
Snowden recently told a crowd of cloud and hosting providers that use of encryption is growing,
and encrypted traffic has doubled since 2013.
In 2013, Forrester predicted that US cloud providers cloud lose up to $180 billion in business by
2016 due to concerns around the scope of NSA’s PRISM program.
According to NextGov, Forrester finds that 26 percent of enterprises based in Asia Pacific,
Canada, Europe and Latin America have stopped or reduced their spending with US-based firms
for Internet-based services. Thirty-four percent said these concerns were related to fears of US
surveillance, while others said they want to support businesses in their own country, or data
sovereignty rules prevent them from storing data abroad.
Forrester surveyed more than 3,000 businesses between June and July 2014.
More than half of respondents said that they did not trust US-based outsourcers to handle
sensitive information, with only 8 percent reporting to trust their company’s intellectual
property with a US-based outsourced company.
Ninety-percent of decision-makers have taken steps to encrypt their data, according to the
report.
Cloud computing not feasible – security hurdles
Xiao and Chen 15 – *professor at the Department of Software Engineering at
Hainan Software Profession Institute AND **Assistant Professor in Operations
Management at New York University, PhD (Ziqian and Jingyou, Cloud
Computing Security Issues and Countermeasures, Proceedings of the 4th
International Conference on Computer Engineering and Networks p. 731-737,
2015, http://link.springer.com/chapter/10.1007/978-3-319-11104-9_85)//JJ
Cloud Computing Security Challenges
New Risks Brought by Virtual Technologies
Virtualization brings new risks mainly in the virtual machine being abused , the virtual
machine escape , and multi-tenant isolation between the failures of security policy migration
of virtual machines.
Shared Data Security Environment
Under the cloud service model, users are very worried about whether the data stored in the
service provider will be compromised , tampered , or lost . Man-made threats facing the user
data mainly come from service providers , hackers , malicious neighboring tenants , and
subsequent tenants.
Cloud Platform Application Security
There are some application security problems existing in Cloud Computing Services, no matter
Saas, Paas or Iaas, mainly including three categories. The first one is the malicious program
review . The second one is the application interface security . The third one is code and test
safety .
Authentication and Access Control in the Cloud Service Model
Under the cloud service model, user authentication and access control face new challenges, for
example, the authentication and authorization of massive users, the rational division of access
rights, and the management of accounts, passwords, and keys. In dealing with massive users’
changeable business and their identification, the cloud service providers need to fully automate
users’ authentication and access management.
Cloud computing improvements now – new tech and legal measures
Rubinstein and Hoboken 14 – *Senior Fellow at the Information Law Institute
(ILI) and NYU School of Law, AND **Microsoft Research Fellow in the
Information Law Institute at New York University, PhD from the University of
Amsterdam (Ira and Joris Van, PRIVACY AND SECURITY IN THE CLOUD: SOME
REALISM ABOUT TECHNICAL SOLUTIONS TO TRANSNATIONAL SURVEILLANCE IN
THE POST- SNOWDEN ERA, 66 Maine L. Rev. 488, September 2014,
http://ssrn.com/abstract=2443604)//JJ
High-security demanding customers such as government agencies and
corporate and organizational users with particularly strict demands for
information security are likely to drive these market responses.214 Customers
will insist upon better guarantees of security and confidentiality and may refuse
to do business with popular, U.S.-based cloud services subject to far-reaching
government surveillance powers. Indeed, they may be barred from doing so
under new proposals in Europe and elsewhere requiring their citizens to rely on
local cloud services.215 In the market for individual users of cloud resources,
there may generally be an increasing demand for better security and privacy
safeguards as a result of the widely discussed examples of mass surveillance of
online interactions and communication. In addition, law and regulation may
increasingly require that certain types of disproportionate lawful access to
cloud data be excluded if cloud providers want unrestricted access to the
market.
Are these measures likely to be effective against intelligence agencies with the
skills and resources of NSA or GCHQ? The answer depends on a variety of
factors, which will be discussed further in this Section. One thing is clear: the
range of technical solutions described in Part III is not binary, and recent
announcements of ‘NSA-proof’ services seem highly oversimplified.
A better way of framing this topic is to ask a series of more nuanced questions
as follows: First, can technological and organizational design of services help to
protect against backdoor access of data in the cloud? Second, and related, can the
cloud industry help to prevent bulk and dragnet access to the data of their customers? Third,
to what extent can the technical and organizational design of cloud services
help to shape lawful access dynamics, such as where and how lawful access
takes place (i.e., which entity and in which geographical location)? And, finally, to
what extent can government agencies armed with surveillance orders counter the design
choices of industry players when new technologies undermine lawful access to data in the cloud
the government is seeking?
Based on the analysis outlined herein, the first question should be answered
positively. As cloud services roll out new security and encryption measures with the
goal of preventing bulk data collection by surreptitious means, this will
undoubtedly interfere with large scale intelligence gathering, such as the
interception of client-server and server-server data streams . Firms like Google,
Microsoft, Yahoo, and Facebook have already begun to implement wellestablished techniques such as TLS/SSL and perfect forward secrecy, just as various
security organizations have begun to review how they develop cryptographic
standards.216 At the end of the day, the protection against backdoor access is
also a matter of resources, however. Certain technological solutions may
prevent effective bulk collection through specific intelligence programs, but
intelligence agencies could in turn deploy targeted intelligence operations to
undo some of these protections implemented by cloud services.
The second question, which concerns the possibility of cloud firms preventing
dragnet surveillance, cannot generally be answered affirmatively. Technological
design may have some impact on front-door collection but where surveillance
regimes like Section 702 of the FAA authorize large scale transnational
surveillance directed at cloud services, industry has limited options. It may
oppose orders in court, or it may take a public stance to the effect that certain
types of lawful access should not be legally permissible under current statutes
and strive for legal reforms that would enhance the privacy interests of cloud
customers.218
The third question must be answered positively also, at least in theory.
Technological and organizational design of services can help to shape lawful
access dynamics and could be used precisely to do so. While few cloud services
have actively implemented privacy-preserving encryption protocols, there is
reason to believe that this is changing . As discussed in the previous section, both
the cloud industry and the Internet security engineering community have taken
the first steps towards implementing technical and organizational measures to shape
the lawful access dynamics induced by the use of their services and further
innovations may be anticipated. The extent to which local jurisdictions may
force multinational cloud service providers to comply with domestic laws
notwithstanding these new security measures remains a particularly hotly
debated issue.
2nc – surveillance not hurt cloud
No significant impact on cloud computing
Weise, 4/7/15 (Elizabeth, “PRISM revelations didn't hit U.S. cloud computing as hard as
expected” 4/7, http://americasmarkets.usatoday.com/2015/04/07/prism-revelations-didnt-hitu-s-cloud-computing-as-hard-as-expected/
When Edward Snowden revealed the extent of the U.S. National Security Agency’s PRISM spying
program, there were concerns that American cloud, hosting and outsourcing businesses would
lose customers running to non-U.S.-based companies safe from NSA’s prying eyes.
“The assertion was that this would be a death blow to U.S. firms trying to operating in Europe
and Asia,” said Forrester Research analyst Ed Ferrara.
But two recent reports from Forrester find it was less catastrophic than expected.
That’s good news for companies like Box (BOX), DropBox and others that make their money by
selling U.S.-based data storage.
Forrester had originally predicted U.S. companies could lose as much as $180 billion in sales.
Instead, just 29% of technology decision-makers in Asia, Canada, Europe and Latin America
halted or reduced spending with U.S.-based firms offering Internet-based services due to the
PRISM scandal, Forrester’s Business Technographics Global Infrastructure Survey for 2014 found
“It’s a relatively small amount of data,” Ferrara said.
That’s because most of the companies didn’t need to move all their data, much of which was
stored in-house. Instead, only 33% of the data held by that 29% of companies was at a thirdparty data center or in a cloud system.
Forrester believes the overall loss to U.S. cloud providers for 2015 will be about $15 billion and
in 2016, $12 billion, a far cry from projections that were ten times that a year ago.
Forrester also found that companies are looking at other ways to protect the integrity of their
data, not just from the NSA but also from surveillance by other nations.
Chief among them was encryption. Eighty-four percent of the companies said they’re using
various encryption methods to protect sensitive material.
The survey’s definition of cloud providers is broad, and includes both platform as a service,
infrastructure as a service and software as a service companies, said Ferrara.
2nc - cloud not feasible
Tons of alt. causes to cloud computing –
Castro and McQuinn 15 – * Vice President of the Information Technology and
Innovation Foundation and Director of the Center for Data Innovation, B.S. in
Foreign Service from Georgetown University and an M.S. in Information
Security Technology and Management from Carnegie Mellon University, AND
** Research Assistant with the Information Technology and Innovation
Foundation, B.S. in Public Relations and Political Communications from the
University of Texas (Daniel and Alan, Beyond the USA Freedom Act: How U.S.
Surveillance Still Subverts U.S. Competitiveness, Information Technology and
Innovation Foundation, June 2015, http://www2.itif.org/2015-beyond-usa-freedomact.pdf?_ga=1.33178294.940386433.1435342104)//JJ
In the short term, U.S. companies lose out on contracts, and over the long term,
other countries create protectionist policies that lock U.S. businesses out of
foreign markets. This not only hurt s U.S. technology companies, but costs
American jobs and weakens the U.S. trade balance. To reverse this trend, ITIF
surveillance activities both at home an
security by opposing any government efforts to introduce backdoors in software or
Strengthen U.S. mutual legal assistance treaties
Work to establish international legal standards for gover
Complete trade agreements like the Trans Pacific Partnership that ban digital
protectionism, and pressure nations that seek to erect protectionist barriers to
abandon those efforts .
Cloud computing not feasible – security hurdles
Xiao and Chen 15 – *professor at the Department of Software Engineering at
Hainan Software Profession Institute AND **Assistant Professor in Operations
Management at New York University, PhD (Ziqian and Jingyou, Cloud
Computing Security Issues and Countermeasures, Proceedings of the 4th
International Conference on Computer Engineering and Networks p. 731-737,
2015, http://link.springer.com/chapter/10.1007/978-3-319-11104-9_85)//JJ
Cloud Computing Security Challenges
New Risks Brought by Virtual Technologies
Virtualization brings new risks mainly in the virtual machine being abused , the virtual
machine escape , and multi-tenant isolation between the failures of security policy migration
of virtual machines.
Shared Data Security Environment
Under the cloud service model, users are very worried about whether the data stored in the
service provider will be compromised , tampered , or lost . Man-made threats facing the user
data mainly come from service providers , hackers , malicious neighboring tenants , and
subsequent tenants.
Cloud Platform Application Security
There are some application security problems existing in Cloud Computing Services, no matter
Saas, Paas or Iaas, mainly including three categories. The first one is the malicious program
review . The second one is the application interface security . The third one is code and test
safety .
Authentication and Access Control in the Cloud Service Model
Under the cloud service model, user authentication and access control face new challenges, for
example, the authentication and authorization of massive users, the rational division of access
rights, and the management of accounts, passwords, and keys. In dealing with massive users’
changeable business and their identification, the cloud service providers need to fully automate
users’ authentication and access management.
2nc – Squo solves
New protection standards and tech solve
Rubinstein and Hoboken 14 – *Senior Fellow at the Information Law Institute
(ILI) and NYU School of Law, AND **Microsoft Research Fellow in the
Information Law Institute at New York University, PhD from the University of
Amsterdam (Ira and Joris Van, PRIVACY AND SECURITY IN THE CLOUD: SOME
REALISM ABOUT TECHNICAL SOLUTIONS TO TRANSNATIONAL SURVEILLANCE IN
THE POST- SNOWDEN ERA, 66 Maine L. Rev. 488, September 2014,
http://ssrn.com/abstract=2443604)//JJ
V. CONCLUSION
This Article describes and places in a legal perspective the cloud industry’s
technological responses to the revelations about ongoing transnational
surveillance. By focusing on industry responses and exploring the ways in which
the technological design of cloud services could further address surveillance
concerns, we provide insights into the prospects of these services shaping
lawful government access to the cloud. This intersection of service design, on
the one hand, and government demands for access to data, on the other hand,
signals a dynamic new chapter in the ongoing debate between industry and
governments about the possibility and conditions of secure and privacy-friendly
information and communications technologies (ICTs) for global markets.
In particular, we have shown that it is helpful to distinguish between front-door
and backdoor access to data in the cloud. Our analysis of industry responses
has shown the cloud industry is moving quickly to address interception of their
customers’ data without their knowledge or involvement by adopting
technological solutions that limit lawful access (as far as possible) to legal
processes directed at the cloud service itself and/or its customers. Many of
these measures could have been implemented much earlier on. They are now
becoming industry norms . Industry standards like SSL/TLS and HTTPS, together
with a new generation of PETs offering “end-to-end” protection, can be
effective tools in preventing bulk acquisition through the targeting of the
worldwide communications infrastructure.
In short, technologies can help the industry shape lawful access even though
they do not change the legal framework, nor do they overcome the lack of progress in
reforming existing legal authorities ( such as Section 702 of the FAA ) to confine lawful
access to the front-door of service providers. We expect that this lack of
progress—with respect to transnational legal guarantees of privacy and
information security, not only in the U.S. but also elsewhere—will be a strong
driver for the wider adoption of more robust and comprehensive privacy
technologies in the cloud service context. And we argue that under current
conditions, the U.S. cloud industry will increasingly rely on technologies to
‘regulate’ government data access in an effort to enhance the privacy and
information security protections of their foreign customers.
This raises the pertinent question of how the U.S. government may respond to
increased resilience of cloud services against lawful surveillance. While FISA
and ECPA allow government agencies to obtain orders that ensure the
cooperation of providers notwithstanding strong technological protections,
existing law does not allow for unlimited bargaining room. Most of the services
in question are not subject to CALEA obligations and an extension of CALEA
seems neither warranted nor politically feasible under present conditions.
Moreover, most of these services have responded to the Snowden revelations
by implementing stronger privacy protections (and even some advanced
cryptographic protocols). No doubt they await the outcome of the ongoing
litigation in the Lavabit case, which may clarify the government’s power to
compel a service to break its security model in response to a valid surveillance
order. However, the Lavabit case does not yet present a scenario in which a
service’s use of advanced cryptography makes it impossible to comply with a
surveillance order by furnishing unencrypted data. 2014] PRIVACY AND
SECURITY IN THE CLOUD 533 A U.S. government win in the Lavabit case may
therefore be little more than a pyrrhic victory, for it could simply further
incentivize industry to adopt even stronger technological solutions against
surveillance, including both actively implemented and client-side encryption
protocols preserving privacy in the cloud .
Encryption solves – major companies prove
Rubinstein and Hoboken 14 – *Senior Fellow at the Information Law Institute
(ILI) and NYU School of Law, AND **Microsoft Research Fellow in the
Information Law Institute at New York University, PhD from the University of
Amsterdam (Ira and Joris Van, PRIVACY AND SECURITY IN THE CLOUD: SOME
REALISM ABOUT TECHNICAL SOLUTIONS TO TRANSNATIONAL SURVEILLANCE IN
THE POST- SNOWDEN ERA, 66 Maine L. Rev. 488, September 2014,
http://ssrn.com/abstract=2443604)//JJ
It is hardly surprising, then, that cloud firms like Microsoft have started taking steps to ensure
that governments use legal process rather than “technological brute force to access customer
data.” engineering effort to strengthen the encryption of customer data across [its] networks
and services.”159 This matches similar activity of Google, which had started to encrypt data
more comprehensively even before the specific revelations about the MUSCULAR program.160
As a Google security engineer explained shortly after these revelations, “the traffic shown in the
[MUSCULAR] slides below is now all encrypted and the work the NSA/GCHQ (U.K. Government
Communications Headquarters) staff did on understanding it, ruined .”161 Finally, Yahoo has
announced it will “[e]ncrypt all information that moves between [its] data centers by the end of
Q1 2014.” The encryption measures discussed above could help the cloud industry to
counteract programs like MUSCULAR and UPSTREAM, which rely on the bulk collection of data
by targeting communication links and the telecommunications infrastructure. Of course, this
assumes that the NSA does not seek to undermine these protections by relying on security
weaknesses in the implementation or use of SSL or the underlying encryption 158 Microsoft
recently announced “a comprehensive algorithms.
2nc – at: https encryption
HTTPS encryption protocols fail –
Rubinstein and Hoboken 14 – *Senior Fellow at the Information Law Institute
(ILI) and NYU School of Law, AND **Microsoft Research Fellow in the
Information Law Institute at New York University, PhD from the University of
Amsterdam (Ira and Joris Van, PRIVACY AND SECURITY IN THE CLOUD: SOME
REALISM ABOUT TECHNICAL SOLUTIONS TO TRANSNATIONAL SURVEILLANCE IN
THE POST- SNOWDEN ERA, 66 Maine L. Rev. 488, September 2014,
http://ssrn.com/abstract=2443604)//JJ
In terms of securing web-based communications, however, the HTTPS system is no panacea
against government surveillance. First, the protocol must be properly implemented .147
Second, there are known attacks on the use of encrypted web communications through
SSL.148 Third, intelligence agencies may work around the protections and attempt to secretly
install software on the computers of targeted users, thereby allowing them to capture their
communications before they are transmitted across an encrypted connection.149 Finally, and
most importantly, HTTPS is not designed to protect data at rest . Even if a cloud provider
properly implements this protocol, this does nothing to prevent a government agency from
obtaining the data it seeks by means of a compulsory order requiring the service provider to
furnish this data. Indeed, as Professor Peter Swire argues, the trend towards encrypting data in
transit between users and cloud services may well result in governments shifting their attention
from attacking the communication infrastructure to demanding that cloud service providers
hand over stored data after it has been securely transmitted.150 The Snowden revelations
already provide some evidence of this shift and the measures detailed in this Section could
accelerate this trend. To counter this trend, governments confronted with encrypted
communication channels could try to compel cloud providers to hand over their encryption
keys, enabling the continued effective interception over telecommunications infrastructure (an
option discussed further in Part IV).
2nc – at: pets
PETs fail – not technologically or economically feasible
Rubinstein and Hoboken 14 – *Senior Fellow at the Information Law Institute
(ILI) and NYU School of Law, AND **Microsoft Research Fellow in the
Information Law Institute at New York University, PhD from the University of
Amsterdam (Ira and Joris Van, PRIVACY AND SECURITY IN THE CLOUD: SOME
REALISM ABOUT TECHNICAL SOLUTIONS TO TRANSNATIONAL SURVEILLANCE IN
THE POST- SNOWDEN ERA, 66 Maine L. Rev. 488, September 2014,
http://ssrn.com/abstract=2443604)//JJ
It is important to emphasize that adoption of the solutions discussed remains
low even though some of them are ready for use. There are a number of
reasons for this. First, some of these solutions, such as FHE, are at the very early
stages of development.188 If service provision is limited to the mere storage of
data in the cloud, it may be technically feasible for the service provider to
anticipate and organize for encryption under the control of cloud users.
However, if the cloud provider also has to perform processing operations on
the encrypted data stored by its customers, the implementation of privacypreserving PETs in the cloud context is far more challenging and may even be
impossible for complex operations. 189
Second, many cloud providers lack the incentive to adopt and further develop
PETs based on advanced cryptographic solutions that would prevent them from
having access to user data. The reasons are obvious: many business models in
the cloud industry depend on generating revenue based on access to
customers’ data (e.g., profiling users for purposes of serving them targeted
ads).190 Thus, for many cloud service providers, the costs of implementing these
PETs (loss of profits) outweigh the potential benefits (improved security and privacy
guarantees for their customers).191 Arguably, the new emphasis on security
and privacy in the cloud in response to the Snowden revelations might
incentivize industry to consider developing and adopting similar measures.
Notwithstanding the current lack of adoption, the point this Article seeks to
emphasize is that if service providers were to deploy such measures, it would
interfere with lawful access requests to cloud providers in some obvious ways. For
example, a provider might simply be unable to share unencrypted customer
data with law enforcement or intelligence agencies notwithstanding a lawful
request for such access.192
Too many hurdles to client-side PETs – their ev. is theoretical
Rubinstein and Hoboken 14 – *Senior Fellow at the Information Law Institute
(ILI) and NYU School of Law, AND **Microsoft Research Fellow in the
Information Law Institute at New York University, PhD from the University of
Amsterdam (Ira and Joris Van, PRIVACY AND SECURITY IN THE CLOUD: SOME
REALISM ABOUT TECHNICAL SOLUTIONS TO TRANSNATIONAL SURVEILLANCE IN
THE POST- SNOWDEN ERA, 66 Maine L. Rev. 488, September 2014,
http://ssrn.com/abstract=2443604)//JJ
What happens if the government serves a lawful request for the content of
communications on a service provider whose customers utilize a client-side PET
for encrypted email or chat? At best, the service providers may hand over
encrypted data but these PETs prevent it from furnishing unencrypted data. On
the other hand, the provider may fully comply with requests for traffic data
unless the user combines a client-side PET with a collaborative PET like Tor.197
Cloud providers’ attitudes to these client-side PETs are likely to remain
ambivalent. On the one hand, they may decide to block their use because they
interfere with their business model and desired uses of the service ;198 on the
other hand, they may embrace PETs as proof of their good faith efforts to
ensure customer privacy in the cloud. By pointing out the possibility to adopt
end-to-end encryption solutions, companies could reassure users who are
rightly worried about the surveillance of their communications.199
Although the availability of encryption solutions may seem attractive for users,
they come with some well-documented downsides in terms of usability .200 As a
result, only dedicated or expert users tend to take advantage of them. In fact
this is another oft-cited reason for industry to shy away from promoting client-side
encryption solutions. In addition, the client-side approach to security tends to
rely on the free or open source software model, in which developers release
their source code, thereby allowing the security community to review the code
and determine that the software is indeed secure. From an ordinary user’s
perspective, this substitutes trust in a group of security experts in lieu of trusting
the third-party services. Finally, it is true that the implementation of end-to-end
encryption may help to protect against third party access to raw data through
the service provider. From the perspective of managing information security
more generally, however, many organizations and individuals may prefer
trusting a dedicated service provider over having to rely on their own expertise.
Of course, the Snowden revelations may boost the adoption of end-to-end
encryption as a way of limiting the widely publicized systematic monitoring of
global Internet communications. Certainly, the NSA’s targeting of major cloud service
providers through programs like PRISM has spiked interest in end-to-end encryption solutions,
at least according to all the hoopla in the popular press .201 For the moment, however,
there seems to be only a small niche market for services that cater to the demand
for properly implemented end-to-end security, as evidenced by services such as
Lavabit,202 Hushmail,203 Silent Circle,204 and Heml.is.205
2nc – at: hushmail/lavabit
Our ev accounts for Hushmail and Lavabit – s-quo progression of corporate
encryption solves
Rubinstein and Hoboken 14 – *Senior Fellow at the Information Law Institute
(ILI) and NYU School of Law, AND **Microsoft Research Fellow in the
Information Law Institute at New York University, PhD from the University of
Amsterdam (Ira and Joris Van, PRIVACY AND SECURITY IN THE CLOUD: SOME
REALISM ABOUT TECHNICAL SOLUTIONS TO TRANSNATIONAL SURVEILLANCE IN
THE POST- SNOWDEN ERA, 66 Maine L. Rev. 488, September 2014,
http://ssrn.com/abstract=2443604)//JJ
This may (or may not) be an accurate description of what happened in the
Hushmail case.273 Hushmail secure email service offers its customers two
options: a high-security option, which requires that users install and run a Javabased encryption applet and encrypts and decrypts email only on the
customer’s computer; and a low-security (non-Java) option, which is more
convenient but less secure because it handles encryption and decryption on
Hushmail’s web server.274 As a result, Hushmail retains the ability to decrypt
user’s emails when they select the low-security option (via an “insider attack”
like that against Lavabit) but no ability to do so when the customer selects the
high-security option.275 Of course, Hushmail’s design does not prevent the
company from modifying the Java applet so that it captures the user’s
passphrase and sends it to Hushmail, thereby enabling the company to decrypt
the email and share it with a third-party including the government. But it seems
unlikely that the company would destroy its own business by subverting its
software in this way and subject itself to a likely deceptive practice
enforcement action under Section 5 of the FTC Act.276 Unlike Lavabit, none of
the sealed documents in the Hushmail case have been leaked, so less
information is available. Also, it is not clear whether the 2007 court order
pertained to a high-security or a low-security user; or if Hushmail modified its
Java encryption engine; or if, in the interests of full disclosure, it merely pointed
out the possibility of doing so.277 In short, the Hushmail case exemplifies the
dilemmas that the government may begin to face if service providers take the
next logical step of adding government agencies to their threat models and
designing systems that protect against valid court orders. And while the
government has prevailed in its efforts to force niche players like Lavabit and Hushmail to
capitulate, it may face a much greater challenge if major Internet firms like Microsoft, Google,
and Facebook go down this path in response to the Snowden revelations.
AT: Space debris impact
Status quo solves space debris- NASA and NOAA prove
Haar and Leslie 14, Audrey Haar works at NASA's Goddard Space Flight Center and John
Lesilie works at the NOAA Office of Communications and External Affairs, (10/22/14, NASANOAA Suomi NPP Satellite Team Ward Off Recent Space Debris Threat,
https://www.nasa.gov/content/goddard/nasa-noaa-suomi-npp-satellite-team-ward-off-recentspace-debris-threat)//AK
While space debris was the uncontrolled adversary in the award-winning space thriller film
"Gravity," space debris, also known as "space junk," is an ongoing real-life concern for teams
managing satellites orbiting Earth, including NOAA-NASA's Suomi National Polar-orbiting
Partnership, or Suomi NPP, satellite. It is not unusual for satellites that have the capability of
maneuvering to be repositioned to avoid debris or to maintain the proper orbit.
On an otherwise quiet Sunday on September 28, the Suomi NPP mission team was monitoring a
possible close approach of a debris object. By early evening, the risk was assessed to be high
enough to start planning a spacecraft maneuver to put the satellite into a safer zone, out of the
path of the object classified in a size range of 4 inches up to 3.3 feet.
It was determined that the object (travelling at almost 17,000 mph) was approaching at a nearly
"head on" angle, and could potentially only miss the Suomi NPP satellite by approximately 300
feet on Tuesday, September 30, if no action was taken. With that knowledge, the decision was
made at 1:30 p.m. on Monday, September 29, for NOAA's Satellite Operations Facility, or NSOF,
in Suitland, Maryland, to reposition Suomi NPP. Operational control as well as planning and
execution of all Suomi NPP maneuvers take place at NSOF.
"Because Suomi NPP moves at a similar speed as the debris object, if there had been an impact,
it would have occurred at a combined speed of nearly 35,000 mph. This would have been
catastrophic not only to the satellite, but would result in thousands of pieces of new debris,"
said Harry Solomon, Mission Manager for Suomi NPP at NASA's Goddard Space Flight Center.
Space around Earth is littered with numerous man-made objects that could potentially collide
with operating spacecraft and each other (creating more debris). There are more than 20,000
objects being monitored by the U.S. Department of Defense for satellite managers around the
world.
Only about 1,000 of those 20,000 objects are operating spacecraft. The rest of the monitored
space debris ranges in size from the size of a softball, to massive rocket bodies, all orbiting
uncontrolled at relative speeds averaging about 22,300 mph in low-Earth orbit, where the
majority of the objects reside.
Yet it is the unknown, often smaller, untracked objects that pose the biggest threat. "If a
spacecraft is lost due to being hit by debris, the odds are the satellite will be hit by something
the trackers can't see," said Nicholas Johnson, NASA chief scientist (retired) for orbital debris at
Johnson Space Center in Houston.
That is exactly the scenario Solomon and his counterpart, Martin England, mission operations
engineering lead at NSOF hope will never happen.
Risk Team Monitors Unmanned Missions Threats for NOAA and NASA
While NASA's Johnson Space Center manages monitored debris threats for spacecraft related to
U.S. manned missions such as the International Space Station, the responsibility for unmanned
missions managed by NASA falls to the Conjunction Assessment Risk Analysis, or CARA, team
operating out of NASA Goddard.
About seven days before a potential threat, information from the Department of Defense is
analyzed by the CARA team to evaluate predicted close approaches. CARA monitors and
provides updated information about potential threats to satellite mission managers who then
make a decision about the need to reposition their satellites in a procedure known as a Risk
Mitigation Maneuver.
Since Suomi NPP's launch in October 2011, this recent reposition was the fourth Risk Mitigation
Maneuver to avoid space debris. In this case, the object was a section of a Thorad-Agena launch
vehicle used between 1966 and1972 primarily for Corona U.S. reconnaisssance satellites.
A previous Suomi NPP risk mitigation maneuver in January 2014 avoided a discarded booster
from a Delta 1 launch vehicle, a type of rocket made in the United States for a variety of space
missions from 1960 to 1990. There is also a significant amount of debris in Suomi NPP's orbit
from the Chinese Fengyun-1C, a meteorological satellite China destroyed in January 2007 in a
test of an anti-satellite missile. Another threat near Suomi NPP's orbit is the debris resulting
from a 2009 collision of a functioning commercial communications satellite and a defunct
Russian satellite.
Suomi NPP's job is to collect environmental observations of atmosphere, ocean and land for
both NOAA's weather and oceanography operational missions and NASA's research mission to
continue the long-term climate record to better understand the Earth's climate and long-term
trends.
To accomplish those goals, the satellite maintains a position on orbit such that the desired
path across the ground does not vary by more than 20 km (12 miles) on each side. This orbit is
adjusted with regular planned maneuvers to maintain the proper orbit and angles for best
information collection. But if a Risk Mitigation Maneuver to avoid space debris were to
necessitate moving out of that desired collection zone, then yet another maneuver would be
necessary to return to the optimum orbit position. These unplanned maneuvers tap into the
finite amount of fuel on satellites and could potentially shorten mission life of a spacecraft if
fuel is used more quickly than anticipated.
The amount of space debris is not constant. It generally increases every year, sometimes
generated from debris collisions, which can potentially create additional debris fragments. But
there are also debris reductions. One tracked object generally falls back to Earth daily,
sometimes burning up to nothing upon re-entry, or falling into water or the large areas of low
population density.
In addition, there are also natural events that help control debris. The sun is currently going
through a period known as solar maximum, the term for a high period of solar activity. The
increased number of sunspots and solar storms during solar maximum takes place
approximately every 11 years. During this period, the extent of Earth's atmosphere increases
due to solar heat generated by the increased amount of solar activity. As the atmosphere
extends to higher altitudes, debris at these altitudes are then subjected to increased friction,
known as drag, and as a result, space debris typically fall to Earth at a higher rate during solar
maximum.
The Suomi NPP mission is a bridge between NOAA and NASA legacy Earth observing missions
and NOAA's next-generation Joint Polar Satellite System, or JPSS. The next satellite, JPSS-1, is
targeted for launch in early 2017.
Status quo mechanisms being strengthened now to solves space debris threats
Bonard 14, expert on the ISS and a space analyst, (Michael, 11/10/14,
Commentary | Space Debris Mitigation: A New Hope for a Realistic Solution?,
http://spacenews.com/42511space-debris-mitigation-a-new-hope-for-arealistic-solution/)//AK
On Jan. 11, 2007, a Chinese antisatellite missile test completely fragmented a Chinese target
satellite into millions of pieces of debris — nearly 800 debris fragments 10 centimeters or
larger, nearly 40,000 debris fragments between 1 and 10 centimeters, and some 2 million
fragments of 1 millimeter or larger.
On Feb. 10, 2009, the operational Iridium 33 and decommissioned Kosmos-2251 satellites
collided at a speed of 42,120 kilometers per hour, destroying both satellites. In July 2011, more
than 2,000 large debris fragments resulting from this collision were detected.
The international space station is routinely dodging debris that are tracked by ground-based
radars.
Space debris constitutes a continuously growing threat to satellites and manned spacecraft.
Very small debris creates potentially nonthreatening damage. Large debris can be detected by
ground-based radars and avoided by spacecraft maneuvers. However, small- to medium-sized
debris in low or medium Earth orbits constitutes the biggest threat. These orbits have the
largest density of debris and the highest relative speeds, while the atmospheric drag is small
enough that it may take centuries to have the debris re-enter the atmosphere.
In 1978, NASA scientist Donald J. Kessler showed that if the density of space debris in low Earth
orbit is high enough, each collision generating space debris would increase the likelihood of
further collisions. One serious implication is that the multiplication of debris in orbit will render
space exploration, and even the use of satellites, increasingly dangerous and costly for many
generations.
Multiple solutions to remove space debris have been explored and published.
One of these solutions involves physical contact between debris and the spacecraft:
Shielding of in-orbit spacecraft has been considered. However, the satellite community has
recognized that the sheer weight of any reasonably efficient shielding would make launch not
economically viable. Furthermore, the speeds involved in physical contacts would generate a
cloud of additional debris.
“Catcher” spacecraft have also been proposed. Conceptually, highly mobile and agile spacecraft
equipped with a “catching device” like a net or a robotic arm could be launched from Earth to
intercept and catch debris. However, unless the catcher spacecraft are able to precisely match
the speed and direction of the debris, any high-speed physical contact between a component of
the catcher spacecraft and space debris will result in a collision, multiplying the debris. The cost
of designing, developing, testing and launching such a spacecraft, with sufficient fuel onboard to
repeatedly intercept multiple debris fragments at different speeds, orbits and altitudes, does
not seem to be economically viable.
Other solutions would use high-power lasers that could vaporize the surface of the debris in
space, deflecting it and possibly changing its orbit to intersect the atmosphere. These solutions
have the advantage of not requiring physical contact with the debris.
Space-based laser systems require designing, building, launching and operating a spacecraft
equipped with a very high-power laser system. Such a design is utterly complex and expensive
and very likely will not be economically viable.
Airborne laser systems are facing the same obstacles: The Boeing YAL-1 Airborne Laser Test Bed
program, which was designed as a missile defense system to destroy tactical ballistic missiles,
was terminated because of cost.
Ground-based laser systems are handicapped by the very long propagation distance,
atmospheric absorption and distortion of the laser beam. Such parameters make this solution
also not economically viable. Furthermore, being located in a single country, a ground-based
laser system would raise serious political issues within the international community because of
its implied antisatellite capability.
In summary, the cost/benefit ratio of the above solutions appears to be the main reason none
has been implemented to date to proactively mitigate the most dangerous debris.
A more affordable approach for cleaning low and medium Earth orbits of small- to mediumsized orbital debris may be achievable. This approach would use the principle of deflecting an
electrically charged, moving object in a magnetic field. The old television tube is probably the
most common example of this principle, where electrical charges (electrons) are deflected by
the magnetic fields generated by the tube deflection coils.
The application of this principle would use a space-based electron gun to generate an electron
beam directed at the orbital debris. The beam would remotely impart an electric charge to the
debris. Earth’s magnetic field would exert a force on the electric charge of such debris crossing
the magnetic field at high speed, modifying its orbit. Over time, the orbit would become highly
elliptical and would intersect the upper atmosphere, where the debris would vaporize or fall to
Earth. Preliminary calculations have shown that this concept is sound. The benefits include:
Cost: Lower cost is the major advantage of electromagnetic deflection.
Feasibility: There is no new or speculative technology to develop. Used in particle accelerators
and in millions of old-style television tubes, the electron gun technology is very mature. The
energy used to generate the electron beam is orders of magnitude lower than high-power
lasers.
Risk: It would reduce the probability of creating additional debris by avoiding any physical
contact.
The electron gun device could be integrated in an add-on module to the international space
station.
The ISS is already in space, and there would be no new spacecraft to develop and launch.
The ISS has a large power-generation capability, while the electron gun would require only
intermittent and modest amounts of energy to operate.
This solution would be more easily adopted by the international space community, since it does
not have the capability to damage or destroy a spacecraft. This feature would be expected to
encourage support and funding of the project by all the nations involved in space operations.
The electromagnetic deflection concept would best be implemented as an international
program, managed and coordinated by the space agencies of several countries.
As with any new technology development, there are still open questions associated with the
deployment of this concept. A formal study would have to be conducted by space specialists to
validate and test the concept and determine the optimum design parameters.
Areas that should be explored include:
The ability to precisely direct the electron beam at the debris. Although electrons can be sent at
near-light speed, they are also deflected by the very magnetic field that will act on the debris,
requiring precise aiming of the electron gun.
The ability of the target to store the electrons.
The retention of the charge by the target. Due to the constant bombardment of the target by
the solar wind that comprises ionized particles, it is expected that the charge of the target will
dissipate over time.
The dynamic response of the target trajectory under the influence of the deflecting force.
In conclusion, civil and government satellites as well as manned missions are currently
exposed to the growing risk of collisions with debris, which may result in costly incidents, or
accidents that could take human lives. It is essential to have a solution implemented as soon as
possible. As of today, the electromagnetic deflection approach seems to be one of the most cost
effective, most realistically achievable and least risky. It deserves to be further evaluated and
pursued.
Space debris not a threat to humans
Chun 11, space debris analyst and contributor at People’s Daily, (Yao, 9/28/11, Experts: No
need to worry about falling space debris, http://en.people.cn/202936/7606918.html) //AK
As more and more satellites are being launched into the space, will the debris of the failed
satellites bring disaster to earth? The experts from the Center for Space Science and Applied
Research (CSSAR) under the Chinese Academy of Sciences say: "Don't panic, space junk will not
fall on your head."
"Recently some reports may have caused certain panic in the public, who are worried that
space debris will threaten people's survival. But, in fact we can rest assured that space debris
will not hit people because the probability is minimal," said Gong Jiancun, deputy director of
CSSAR.
Space debris will not pose a threat to humans, he said. However, the real reason why scientists
are concerned about space debris is because of its potential to harm or hinder spacecraft.
Since 1957, when the first artificial satellite was launched into space, the amount of space debris
has increased year by year. As of this week, there are more than 16,000 pieces of debris with a
diameter of more than 10 centimeters in space, according to observation data from the United
States.
This debris is distributed in different earth orbits: low orbit, hundreds of kilometers away from
the earth; moderate-altitude orbit, thousands of kilometers away, and high orbit, tens of
thousands of kilometers away. Because of this, the debris is not concentrated in a dense region
of space.
Generally speaking, space debris is divided in three categories: large space debris, with a
diameter of more than 10 centimeters; small space debris, with a diameter of less than 1
millimeter, and dangerous debris, with a diameter between large and small debris.
"If the debris falls to the earth, most of it will be burned away by the high temperature of
thousands of degrees produced by the high-speed friction with the atmosphere. Even if a large
chunk of space debris penetrated the atmosphere and posed a threat to the earth, mankind
should be capable of defending against it," Gong said.
First, we can roughly estimate its orbit. With the estimation of its orbit, we can intercept it.
Gong said that the U.S. has successfully intercepted a failed satellite using a missile. That
satellite contained highly toxic substances. In order to prevent it from falling into the sea, the
U.S. destroyed the satellite by a missile launched from a warship. China also has similar
technologies and can disintegrate it in the space before it causes harm."
"Scientists also have come up with many other methods to clear the space debris. For
example, we can leave some fuel in satellites and control the satellite to fly out of the original
track," Gong said. "Some countries have developed passive technologies, such as launching a
spacecraft to catch space debris and take it away. Other countries are developing satellites
with mechanical arms, which not only can repair satellites but also can pull the failed satellites
out of the orbit."
However, these technologies are not very mature. It is still uncertain when they will come into
use, he said.
China turn
Chinese and U.S. tech industries are zero-sum – surveillance crowds out the U.S.
market
Castro and McQuinn 15 – * Vice President of the Information Technology and
Innovation Foundation and Director of the Center for Data Innovation, B.S. in
Foreign Service from Georgetown University and an M.S. in Information
Security Technology and Management from Carnegie Mellon University, AND
** Research Assistant with the Information Technology and Innovation
Foundation, B.S. in Public Relations and Political Communications from the
University of Texas (Daniel and Alan, Beyond the USA Freedom Act: How U.S.
Surveillance Still Subverts U.S. Competitiveness, Information Technology and
Innovation Foundation, June 2015, http://www2.itif.org/2015-beyond-usafreedom-act.pdf?_ga=1.33178294.940386433.1435342104)//JJ
Protectionist policies in China have further strained the U.S. tech industry . In
January 2015, the Chinese government adopted new regulations that forced
companies that sold equipment to Chinese banks to turn over secret source
code, submit to aggressive audits, and build en cryption keys into their
products. 38 While ostensibly an attempt to strengthen cybersecurity in critical
Chinese industries, many western tech companies saw these policies as a shot
across the bow trying to force them out of China’s markets . After all, the Chinese
government ha d already launched a “de - IOE” movement — IOE stands for
IBM, Oracle and EMC — to convince its state - owned banks to stop buying from
these U.S. tech giants . 39 To be sure, the Chinese government recently halted this
policy under U.S. pressur e. 40 However, the halted policy can be seen as a part
of a larger clash between China and the United States over trade and cybersecurity.
Indeed, these proposed barriers were in part a quid pro quo from China , after the
United States barred Huawei, a major Chinese computer maker, from selling its
products in the United States due to the fear that this equipment had “back
doors” for the Chinese government. 41 Since the Snowden revelations
essentially gave them cover, Chinese lawmakers have openly called for the use
of domestic tech products over foreign goods both to boost the Chinese economy and
in response to U.S. surveillance tactics. This system of retaliation has not only led
to a degradation of business interests for U.S. tech companies in China, but also
disrupted the dialogue between the U.S. government and China on
cybersecurity issues. 4
A2: Big Data/Cloud Computing Add-On
Reject internet doomsaying – no chance of collapse or a ton of other stuff would
cause it
Bernal 14 (Lecturer in Information Technology, Intellectual Property and Media Law at the University
of East Anglia Law School)
(Paul, So who’s breaking the internet this time?, November 11, 2014,
http://paulbernal.wordpress.com/2014/11/11/so-whos-breaking-the-internet-this-time/)
I’m not sure how many times I’ve been told that the internet is under dire threat over the last few
years. It sometimes seems as though there’s an apocalypse just around the corner pretty much all
the time. Something’s going to ‘break’ the internet unless we do something about it right
away. These last few weeks there seem to have been a particularly rich crop of apocalyptic
warnings – Obama’s proposal about net neutrality yesterday being the most recent. The internet
as we know it seems as though it’s always about to end.
Net neutrality will destroy us all…
If we are to believe the US cable companies, Obama’s proposals will pretty much break the internet, putting
development back 20 years. How many of us remember what the internet was like in 1994? Conversely, many have been
saying that if we don’t have net neutrality – and Obama’s proposals are pretty close to what most people I know would
understand by net neutrality – then the cable companies will break the internet. It’s apocalypse one way, and apocalypse
the other: no half measures here.
The cable companies are raising the spectre of government control of the net, something that has been a terror of internet
freedom activists for a very long time – in our internet law courses we start by looking at John Perry Barlow’s 1996
‘Declaration of the Independence of Cyberspace’, with its memorable opening:
“Governments of the Industrial World, you weary giants of flesh and steel, I come from Cyberspace, the new home of
Mind. On behalf of the future, I ask you of the past to leave us alone. You are not welcome among us. You have no
sovereignty where we gather.”
the UN, through the
International Telecommunication Union (ITU) was about to take over the internet,
crushing our freedom and ending the Internet as we know it. Anyone with real experience of
Another recent incarnation of this terror has been the formerly much hyped fear that
the way that UN bodies work would have realised this particular apocalypse had next-to-no chance of
every coming into fruition, and last week that must have become clear to most of even the more paranoid of internet
freedom fighters, as the ITU effectively resolved not to even try… Not that apocalypse, at least not now.
More dire warnings and apocalyptic worries have been circling about the notorious ‘ right to be forgotten’ –
either in its data protection reform version or in the Google Spain ruling back in May. The right to be forgotten, we were
told, is the biggest threat to freedom of speech in the coming decade, and will change the internet as we know it. Another
thing that’s going to break the internet. And yet, even though it’s now effectively in force in one particular way, there’s
not much sign that the internet is broken yet…
The deep, dark, disturbing web…
At times we’re also told that a lack of privacy will break the net – or that privacy itself will break
the net. Online behavioural advertisers have said that if they’re not allowed to track us, we’ll break the economic
model that sustains the net, so the net itself will break. We need to let ourselves be tracked, profiled and targeted or the
net itself will collapse. The authorities seem to have a similar view – recent pronouncements by Metropolitan Police
Commissioner Bernard Hogan-Howe and new head of GCHQ Robert Hannigan are decidedly apocalyptic, trying to
terrify us with the nightmares of what they seemingly interchangeably call the ‘dark’ web or the ‘deep’ web. Dark or
deep, it’s designed to disturb and frighten us – and warn us that if we keep on using encryption, claiming anonymity or
pseudonymity or, in practice, any kind of privacy, we’ll turn the internet into a paradise only for paedophiles, murderers,
terrorists and criminals. It’s the end of the internet as we know it, once more.
And of course there’s the converse view – that mass surveillance and intrusion by the NSA, GCHQ etc, as revealed by
Edward Snowden – is itself destroying the internet as we know it.
Money, money, money
Mind you, there are also dire threats from other directions. Internet freedom fighters have fought against things like
SOPA, PIPA and ACTA – ways in which the ‘copyright lobby’ sought to gain even more control over the internet.
Again, the arguments go both ways. The content industry suggest that uncontrolled piracy is breaking the net – while
those who fought against SOPA etc think that the iron fist of copyright enforcement is doing the same. And for those that
have read Zittrain’s ‘The Future of the Internet and How to Stop It’, it’s something else that’s breaking the net –
‘appliancization’ and ‘tethering’. To outrageously oversimplify, it’s the iPhone that’s breaking the net, turning it from a
place of freedom and creativity into a place for consumerist sheep.
It’s the end of the internet as we know it…..…or as we think we know it. We all have different visions of the internet,
some historical, some pretty much entirely imaginary, mowith elements of history and elements of wishful thinking. It’s
easy to become nostalgic about what we imagine was some golden age, and fearful about the future, without taking a
The internet was never a ‘wild west’ – and
even the ‘wild west’ itself was mostly mythical – and ‘freedom of speech’ has never been
as absolute as its most ardent advocates seem to believe. We’ve always had some control
and some freedom – but the thing about the internet is that, in reality, it’s pretty robust. We,
step back and wondering whether we’re really right.
as an internet community, are stronger and more wilful than some of those who wish to control it
might think. Attempts to rein it in often fail – either they’re opposed or they’re sidestepped, or they’re just absorbed into the new shape of the internet, because the internet
is always changing, and we need to understand that. The internet as we know it is always ending
– and the internet as we don’t know it is always beginning.
Download