“It is insufficient to protect ourselves with laws; we need to protect ourselves with mathematics.” ---Bruce Schneier in ‘Applied Cryptography’, pp xx 1 Security Planning A Revision Components of security planning: assessing the threat, writing a security policy: a statement of what is allowed and what is not allowed; assigning security responsibilities. Choosing the mechanism, tools and methodologies to implement the policy 2 Types of Attack A Revision Most Internet security problems are access control or authentication ones Denial of service is also popular, but mostly an annoyance Types of Attack • A Passive attack can only observe communications or data • An Active attack can actively modify communications or data • Often difficult to perform, but very powerful – Mail forgery/modification – TCP/IP spoofing/session hijacking • 3 Attackers External Attackers (through wired part of Internet): Class 1 External Attackers (through wireless part of Internet): Class 2 Internal Attackers (through wired segment of the LAN): Class 3 Internal Attackers (through wireless segment of the LAN): Class 4 4 5 Stages of an Attack The first three Reconnaissance: To find out about hosts, services and application versions: high probability of detection Exploitation: To enter the network to use the services (without legitimate authorization) or to subvert the services: medium probability of detection Reinforcement: To retrieve tools to elevate access rights and to hide the intrusion: If tools are encrypted difficult to detect; can be detected by keeping a watch on the outbound activity of servers 5 5 Stages of an Attack Consolidation: to communicate by using a secret channel (back-doors): profiling. The last two may be detected through traffic Pillage: to steal information or to damage the asset: profiling. may be detected through traffic Reference for the last three slides: Classification by Richard Bejtlich, “ The TAO of Network Security Monitoring”, Addison Wesley, 2005, pp45, pp 19 6 Additional terminology Shoulder Surfing: A hacker reads, when the user is writing on a paper or when he is typing on a keyboard. Pulsing zombie: A compromised computer (zombie), which is used for intermittently attacking other targets Snoop Server: a server put in a promiscuous mode for accessing all the data in each network packet; used for surveillance 7 Additional terminology continued Back Orifice: a window application, which allows a hacker at one computer to control a remote computer; written by a hackers’ group called “the Cult of the Dead Cow” War Driving: Unauthorized access into the wireless net of a company by parking a car outside the building of the company Smurf attack: DoS attack mounted through a ping addressed to an IP broadcast address; the resultant echo may flood the net 8 Additional terminology continued 2 Hacktivism: intrusion done as a protest; justified as free speech Rootkit: the tools installed on a computer to hide the presence of an intruder Symantec Definition: A rootkit is a component that uses stealth to maintain a persistent and undetectable presence on a computer. "Actions performed by a rootkit, such as installation and any form of code execution, are done without end-user consent or knowledge." -- Ryan Naraine,” When's a Rootkit Not a Rootkit? In Search of Definitions,” eWeek, Jan18, 2006 Pete Allor, director of operations, IT-ISAC ( Information Sharing and Analysis Center): working on a definition of Rootkit. 9 Security Theories Ref: Matt Bishop, “Computer Security: Art & Science,” Addison-Wesley 03 Given: A computing system ( with computers, networks etc) To Find: Is it (provably) secure? Answers: 1976: Harrison, Ruzzo and Ullman: In the most general abstract case, the security of computer systems was undecidable. Reference: M. Harrison, W. Ruzzo and J. Ullman, “Protection in Operating Systems,” Communications of the ACM 19 (8), pp.461-471 (Aug. 1976). 10 Security Theories: Answers … continued Jones, Lipton and Snyder: presented a specific system, in which security was decidable --- in a time period, which increased linearly with the size of the system. Reference: A. Jones, R. Lipton and L. Snyder, “A Linear-Time Algorithm for Deciding Security,” Proceedings of the 17th Symposium on the Foundations of Computer Science, pp.33-41 (Oct. 1976). 11 Security Theories: Answers … continued 2 Minsky: presented a model – to examine why in the general case the security was undecidable and in a specific case it was. Reference: N. Minsky, “Selective and Locally Controlled Transport of Priveleges,” ACM Transactions on Programming Languages and Systems 6 (4), pp.573-602 (Oct. 1984). Sandhu: Extended the Minsky model and presented further insights. Reference: R. Sandhu, “The Schematic Protection Model: Its Definition and Analysis for Acyclic Attenuating Schemes”, Journal of the ACM 35 (2), pp.404-432 (Apr. 1988). 12 Security Policy Study needs of an organization Security Policy Mechanism -- Procedural -- Technical -- Physical 13 Definitions Consider a computer system as a FINITE STATE AUTOMATON with Transition Functions that change state. Security Policy: A statement that partitions the system into sets of authorized or secure states; (called S in slide 38) unauthorized or secure states. (P – S) A Secure System: One that starts in an authorized state and cannot enter an unauthorized state. A Security Incident: When a system enters an unauthorized state. 14 Definitions: Confidentiality and Integrity X: a set of entities; or resource I: some information I has the property of confidentiality wrt X, if no member of X can obtain information about I. I has the property of integrity wrt X, if all members of X trust I. 15 TRUST Trust that Conveyance and storage of I does not change the information or its trustworthiness Data Integrity; I is correct and unchanged, if I is information about the origin of some thing or about identification of an entity Authentication The resource functions correctly, if I is a resource rather than information Assurance 16 Definitions: X: a set of entities; Availability I: some resource I has the property of availability wrt X, if all members of X can access it. Meaning of access: depends upon needs of members of X nature of resource use to which the resource is put 17 Security Policy: Confidentiality, Integrity The policy considers the issues of CIA as follows: confidentiality During information flow For environment, which changes with time ( Example: a contractor bound by non-disclosure agreement, during the period of contract) Integrity Authorized ways of altering information Entities authorized to alter it Principle of “separation of duties” 18 Security Policy: Availability Availability Services that must be provided Parameters within which the services will be accessible (Example: A browser may download web pages but not java applets.) QoS issues Assumptions: The context of the policy: laws, organizational policies and other environmental factors 19 Example: University Rule: No cheating is allowed. School of CS Procedures: Students should write the programs on the School computers and every such file should be read-protected so that other students are not able to read it. Example: A forgets to read-protect his file. B copies it. The copying is caught. Policy vs. Mechanism B claims: The policy does not prohibit copying of a file. So he is not guilty. The policy says that one should read-protect the file. So A is guilty. IS B GUILTY? A security mechanism: an entity or procedure to enforce some part of policy. 20 Security Models Security Model: represents a policy or a set of policies. Helps analyze specific characteristics of policies. No single non-trivial analysis can cover all policies. By restricting the class of policies, a meaningful analysis may be possible. 21 Confidentiality Policies: Bell-LaPadula Model Ref: D.Bell, L.LaPadula, “Secure Computer System: Mathematical Foundations,” Technical Report MTR-2547, Vol. I, MITRE Corporation, Bedford, MA (Mar. 1973) Confidentiality classification: linearly ordered sensitivity levels Subject: security clearance Object: security classification Goal: To prevent read access to objects at a security classification higher than the subject’s clearance. McLean’s questions about B-P model (and the B-P responses) essentially led to the IEEE Computer Security Fundamentals Workshops. 22 Research: the Theory of Security Systems June 1988: First IEEE Computer Security Foundations Workshop: held at The Franconia Inn, New Hampshire ( The Workshop: referred to as “Franconia” even today). (The preface of the Proceedings, written by the workshop Chair, Jonathan Millen, refers to another workshop on the “Foundations of Secure Computation” 1977.) 19th IEEE Computer Security Foundations Workshop (CSFW 19), July 5 - 7, 2006, Venice, Italy, sponsored by the Technical Committee on Security and Privacy of the IEEE Computer Society 23 Integrity Policies: Biba Integrity Model Ref: K.Biba, “Integrity Considerations for Secure Computer Systems,” Technical Report MTR-3153, MITRE Corporation, Bedford, MA (Apr. 1997). Goal of the model: To find answers to: “ Has the integrity of a piece of software or of data, on which the software relies, been compromised?” for software, that exhibit specific properties. Principle of separation of duties, wherever two or more steps are required for a critical function Principle of separation of functions (Ex.: Development, testing, deployment, certification) Requirements of auditing, extensive logging, recovery and accountability 24 The Biba Integrity Model S: a set of subjects; O: a set of objects; I: a set of integrity levels. s Є S can read o Є O, iff i(s) ≤ i(o). s Є S can write to o Є O, iff i(o) ≤ i(s). s1 Є S can execute s2 Є S , iff i(s2) ≤ i(s1). 25 Data Access Controls: Privacy issues Mandatory Access Controls (MACs) Discretionary Access Controls (DACs) Many questions?: Should MACs or DACs be exercised by the owner, the originator, the creator or all? Are temporal changes required in access rights? Conflict of Interest issues 26