Program Security - Computer Science

advertisement
CS 5950/6030 –
Computer Security and Information Assurance
Section 3: Program Security
Dr. Leszek Lilien
Department of Computer Science
Western Michigan University
Slides based on Security in Computing. Third Edition by Pfleeger and Pfleeger.
Using some slides courtesy of:
Prof. Aaron Striegel — course taught at U. of Notre Dame
Prof. Barbara Endicott-Popovsky and Prof. Deborah Frincke (U. Idaho) — taught at U. Washington
Prof. Jussipekka Leiwo — taught at Vrije Universiteit (Free U.), Amsterdam, The Netherlands
Slides not created by the above authors are © 2006 by Leszek T. Lilien
Requests to use original slides for non-profit purposes will be gladly granted upon a written request.
Program Security – Outline (1)
3.1. Secure Programs – Defining & Testing
a.
b.
c.
d.
e.
Introduction
Judging S/w Security by Fixing Faults
Judging S/w Security by Testing Pgm Behavior
Judging S/w Security by Pgm Security Analysis
Types of Pgm Flaws
3.2. Nonmalicious Program Errors
a.
b.
c.
d.
Buffer overflows
Incomplete mediation
Time-of-check to time-of-use errors
Combinations of nonmalicious program flaws
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
2
Program Security – Outline (2)
3.3. Malicious Code
3.3.1. General-Purpose Malicious Code incl.
Viruses
a.
b.
c.
d.
e.
f.
g.
h.
Introduction
Kinds of Malicious Code
How Viruses Work
Virus Signatures
Preventing Virus Infections
Seven Truths About Viruses
Case Studies
Virus Removal and System Recovery After Infection
3.3.2. Targeted Malicious Code
a. Trapdoors
b. Salami attack
c. Covert channels
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
3
Program Security – Outline (3)
3.4. Controls for Security
a. Introduction
b. Developmental controls for security
c. Operating System controls for security
d. Administratrive controls for security
e. Conclusions
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
4
3. Program Security (1)
 Program security –
Our first step on how to apply security to computing

Protecting programs is the heart of computer security



All kinds of programs, from apps via OS, DBMS, networks
Issues:
 How to keep pgms free from flaws
 How to protect computing resources from pgms with flaws
Issues of trust not considered:
 How trustworthy is a pgm you buy?
 How to use it in its most secure way?

Partial answers:
 Third-party evaluations
 Liability and s/w warranties
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
5
Program Security (2)
 Outline:
3.1. Secure Programs – Defining and Testing
3.2. Nonmalicious Program Errors
3.3. Malicious Code
3.3.1. General-Purpose Malicious Code incl. Viruses
3.3.2. Targeted Malicious Code
3.4. Controls Against Program Threats
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
6
3.1. Secure Programs - Defining & Testing

Outline
a. Introduction
b. Judging S/w Security by Fixing Faults
c. Judging S/w Security by Testing Pgm Behavior
d. Judging S/w Security by Pgm Security Analysis
e. Types of Pgm Flaws
Section 3 – Computer Security and Information Assurance – Spring 2006
[cf. B.
Endicott-Popovsky]
© 2006 by Leszek T. Lilien
7
a. Introduction (1)


Pgm is secure if we trust that it provides/enforces:
 Confidentiality
 Integrity
 Availability
What is „Program security?”
Depends on who you ask

user - fit for his task

programmer - passes all „her” tests

manager - conformance to all specs




Developmental criteria for program security include:
Correctness of security & other requirements
Correctness of implementation
Correctness of testing
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
8
Introduction (2)

Fault tolerance terminology:

Error - may lead to a fault

Fault - cause for deviation from intended function

Failure - system malfunction caused by fault

Note:
[cf. A. Striegel]
Faults - seen by „insiders” (e.g., programmers)
Failures - seen by „outsiders” (e.g., independent testers, users)
Error/fault/failure example:




Programmer’s indexing error, leads to buffer overflow fault
Buffer overflow fault causes system crash (a failure)
Two categories of faults w.r.t. duration
[cf. A. Striegel]

Permanent faults

Transient faults – can be much more difficult to diagnose
Section 3 – Computer Security and Information Assurance – Spring 2006
[cf.
A. Striegel]
© 2006 by Leszek T. Lilien
9
b. Judging S/w Security by Fixing Faults

An approach to judge s/w security:
penetrate and patch
 Red Team / Tiger Team tries to crack s/w
 If you withstand the attack => security is good
 Is this true? Rarely.

Too often developers try to quick-fix problems
discovered by Tiger Team
Quick patches often introduce new faults due to:
 Pressure – causing narrow focus on fault, not
context
 Non-obvious side effects
 System performance requirements not allowing
for security overhead
Section 3 – Computer Security and Information Assurance – Spring 2006
[cf.
A. Striegel]
© 2006 by Leszek T. Lilien
10
c. Judging S/w Security
by Testing Pgm Behavior (1)

Better approach to judging s/w security:
testing pgm behavior

Compare behavior vs. requirements (think testing/SW eng)

Program security flaw =
= inappropriate behavior caused by a pgm fault or failure

Flaw detected as a fault or a failure
Important: If flaw detected as a failure (an effect), look for
the underlying fault (the cause)

Recall: fault seen by insiders, failure – by outsiders

If possible, detect faults before they become failures
Note:
Texbook defines flaw-vulnerability-flaw in a circular way
– a terminology soup!
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
11
Judging S/w Security by Testing Pgm Behavior (2)
Any kind of fault/failure can cause a security incident





Misunderstood requirements /
error in coding / typing error
In a single pgm / interaction of k pgms
Intentional flaws or accidental (inadvertent) flaws
Therefore, we must consider security consequences for all
kinds of detected faults/failures
 Even inadvertent faults / failures
 Inadvertent faults are the biggest source of security
vulnerabilities exploited by attackers
 Even dormant faults
 Eventually can become failures harming users
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
12
Judging S/w Security by Testing Pgm Behavior (3)

Problems with pgm behavior testing
 Limitations of testing




Complexity – malicious attacker’s best friend




Can’t test exhaustively
Testing checks what the pgm should do
Can’t test what the pgm should not do
 i.e., can’t make sure that pgm does only what it should do –
nothing more
Too complex to model / to test
Exponential # of pgm states / data combinations
a faulty line hiding in 10 million lines of code
Evolving technology


New s/w technologies appear
Security techniques catching up with s/w technologies
Section 3 – Computer Security and Information Assurance – Spring 2006
[cf.
Striegel]
© A.
2006
by Leszek T. Lilien
13
d. Judging S/w Security
by Pgm Security Analysis
Best approach to judging s/w security:
pgm security analysis


Analyze what can go wrong

At every stage of program development!


After deployment


From requirement definition to testing
Configurations / policies / practices
Protect against security flaws

Specialized security methods and techniques

Specialized security tools

E.g., specialized security meth/tech/tools for switching s/w
Section 3 – Computer Security and Information Assurance – Spring 2006
[cf.©B.2006
Endicott-Popovsky]
by Leszek T. Lilien
14
e. Types of Pgm Flaws

Taxonomy of pgm flaws:

Intentional

Malicious

Nonmalicious

Inadvertent

Validation error (incomplete or inconsistent)


Domain error


e.g., incomplete or inconsistent input data
e.g., using a variable value outside of its domain
Serialization and aliasing


serialization – e.g., in DBMSs or OSs
aliasing - one variable or some reference, when changed, has an
indirect (usually unexpected) effect on some other data




Note: ‘Aliasing’ not in computer graphics sense!
Inadequate ID and authentication (Section 4—on OSs)
Boundary condition violation
Other exploitable logic errors
Section 3 – Computer Security and Information Assurance – Spring 2006
[cf. B.© Endicott-Popovsky]
2006 by Leszek T. Lilien
15
3.2. Nonmalicious Program Errors

Outline
a. Buffer overflows
b. Incomplete mediation
c. Time-of-check to time-of-use errors
d. Combinations of nonmalicious program flaws
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
16
a. Buffer Overflows (1)


Buffer overflow flaw — often inadvertent (=>nonmalicious)
but with serious security consequences
Many languages require buffer size declaration

C language statement: char sample[10];

Execute statement:
sample[i] = ‘A’; where i=10

Out of bounds (0-9) subscript – buffer overflow occurs

Some compilers don’t check for exceeding bounds

C does not perform array bounds checking.

Similar problem caused by pointers

No reasonable way to define limits for pointers
Section 3 – Computer Security and Information Assurance – Spring 2006
[cf. B.
Endicott-Popovsky]
© 2006 by Leszek T. Lilien
17
Buffer Overflows (2)


Where does ‘A’ go?

Depends on what is adjacent to ‘sample[10]’

Affects user’s data - overwrites user’s data

Affects users code - changes user’s instruction

Affects OS data
- overwrites OS data

Affects OS code
- changes OS instruction
This is a case of aliasing (cf. Slide 26)
Section 3 – Computer Security and Information Assurance – Spring 2006
[cf. ©
B.2006
Endicott-Popovsky]
by Leszek T. Lilien
18
Buffer Overflows (3)


Implications of buffer overflow:

Attacker can insert malicious data values/instruction
codes into „overflow space”
Supp. buffer overflow affects OS code area

Attacker code executed as if it were OS code

Attacker might need to experiment to see what
happens when he inserts A into OS code area


Can raise attacker’s privileges (to OS privilege level)

When A is an appropriate instruction
Attacker can gain full control of OS
Section 3 – Computer Security and Information Assurance – Spring 2006
[cf.©B.2006
Endicott-Popovsky]
by Leszek T. Lilien
19
Buffer Overflows (4)
Supp. buffer overflow affects a call stack area
A scenario:





Stack: [data][data][...]
Pgm executes a subroutine
=> return address pushed onto stack
(so subroutine knows where to return control to when finished)
Stack: [ret_addr][data][data][...]
Subroutine allocates dynamic buffer char sample[10]
=> buffer (10 empty spaces) pushed onto stack
Stack: [..........][ret_addr][data][data][...]
Subroutine executes: sample[i] = ‘A’ for i = 10
Stack: [..........][A][data][data][...]
Note: ret_address overwritten by A!
(Assumed: size of ret_address is 1 char)
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
20
Buffer Overflows (5)
Supp. buffer overflow affects a call stack area—CONT

Stack: [..........][A][data][data][...]

Subroutine finishes

Buffer for char sample[10] is deallocated
Stack: [A][data][data][...]

RET operation pops A from stack (considers it ret. addr.)
Stack: [data][data][...]

Pgm (which called the subroutine) jumps to A
=> shifts program control to where attacker wanted
Note: By playing with ones own pgm attacker can specify
any „return address” for his subroutine

Upon subroutine return, pgm transfers control to
attacker’s chosen address A (even in OS area)

Next instruction executed is the one at address A

Could be 1st instruction of pgm that grants
highest access privileges to its „executor”
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
21
Buffer Overflows (6)
Note:



[Wikipedia – aliasing]
C programming language specifications do not specify
how data is to be laid out in memory (incl. stack layout)
Some implementations of C may leave space between
arrays and variables on the stack, for instance, to
minimize possible aliasing effects.
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
22
Buffer Overflows (7)
Web server attack similar to buffer overflow attack:
pass very long string to web server (details: textbook, p.103)


Buffer overflows still common

Used by attackers

to crash systems

to exploit systems by taking over control

Large # of vulnerabilities due to buffer overflows
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
23
b. Incomplete Mediation (1)



Incomplete mediation flaw — often inadvertent (=>
nonmalicious) but with serious security consequences
Incomplete mediation:
Sensitive data are in exposed, uncontrolled condition
Example
 URL to be generated by client’s browser to access server,
e.g.:
http://www.things.com/order/final&custID=101&part=555A&qy=20
&price=10&ship=boat&shipcost=5&total=205

Instead, user edits URL directly, changing price and total
cost as follows:
http://www.things.com/order/final&custID=101&part=555A&qy=20
&price=1&ship=boat&shipcost=5&total=25

User uses forged URL to access server

The server takes 25 as the total cost
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
24
Incomplete Mediation (2)


Unchecked data are a serious vulnerability!
Possible solution: anticipate problems

Don’t let client return a sensitive result (like total)
that can be easily recomputed by server

Use drop-down boxes / choice lists for data input


Prevent user from editing input directly
Check validity of data values received from client
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
25
c. Time-of-check to Time-of-use Errors (1)
Time-of-check to time-of-use flaw — often inadvertent (=>
nonmalicious) but with serious security consequences

A.k.a. synchronization flaw / serialization flaw


TOCTTOU — mediation with “bait and switch” in the middle

Non-computing example:



Swindler shows buyer real Rolex watch (bait)
After buyer pays, switches real Rolex to a forged one
In computing:

Change of a resource (e.g., data) between time
access checked and time access used

Q: Any examples of TOCTTOU problems from
computing?
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
26
Time-of-check to Time-of-use Errors (2)


...
TOCTTOU — mediation with “bait and switch” in the middle

...

Q: Any examples of TOCTTOU problems from
computing?

A: E.g., DBMS/OS: serialization problem:
pgm1 reads value of X = 10
pgm1 adds X = X+ 5
 pgm2 reads X = 10, adds 3 to X, writes X = 13
pgm1 writes X = 15
X ends up with value 15 – should be X = 18
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
27
Time-of-check to Time-of-use Errors (3)

Prevention of TOCTTOU errors

Be aware of time lags

Use digital signatures and certificates to „lock” data
values after checking them

So nobody can modify them after check & before
use

Q: Any examples of preventing TOCTTOU from
DBMS/OS areas?
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
28
Time-of-check to Time-of-use Errors (4)

Prevention of TOCTTOU errors

...

Q: Any examples of preventing TOCTTOU from
DBMS/OS areas?

A1: E.g., DBMS: locking to enforce proper serialization
(locks need not use signatures—fully controlled by DBMS)
In the previous example:
will force writing X = 15 by pgm 1, before pgm2
reads X (so pgm 2 adds 3 to 15)
OR:
will force writing X = 13 by pgm 2, before pgm1
reads X (so pgm 1 adds 5 to 13)

A2: E.g., DBMS/OS: any other concurrency control
mechanism enforcing serializability
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
29
d. Combinations of Nonmal. Pgm Flaws


The above flaws can be exploited in multiple steps by a
concerted attack
Nonmalicious flaws can be exploited to plant malicious flaws
(next)
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
30
3.3. Malicious Code




Malicious code or rogue pgm is written to exploit flaws in pgms
Malicious code can do anything a pgm can
Malicious code can change
 data
 other programs
Malicious code has been „oficially” defined by Cohen in 1984
but virus behavior known since at least 1970 Ware’s study for
Defense Science Board (classified, made public in 1979)
Outline for this Subsection:
3.3.1. General-Purpose Malicious Code (incl. Viruses)
3.3.2. Targeted Malicious Code
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
31
3.3.1. General-Purpose Malicious Code
(incl. Viruses)

Outline
a. Introduction
b. Kinds of Malicious Code
c. How Viruses Work
d. Virus Signatures
e. Preventing Virus Infections
f. Seven Truths About Viruses
g. Case Studies
Section 3 – Computer Security and Information Assurance – Spring 2006
[cf. B.
Endicott-Popovsky]
© 2006 by Leszek T. Lilien
32
a. Introduction


Viruses are prominent example of general-purpose malicious
code
 Not „targeted” against any user
 Attacks anybody with a given app/system/config/...
Viruses
 Many kinds and varieties
 Benign or harmful
 Transferred even from trusted sources
 Also from „trusted” sources that are negligent to update
antiviral programs and check for viruses
Section 3 – Computer Security and Information Assurance – Spring 2006
[cf. ©B.2006
Endicott-Popovsky]
by Leszek T. Lilien
33
b. Kinds of Malicious Code (1)
[remember Introduction?]
Trapdoors
Trojan Horses
X
Files
Bacteria
Logic Bombs
Worms
Viruses
[cf. Barbara Edicott-Popovsky and Deborah Frincke, CSSE592/492, U. Washington]
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
34
b. Kinds of Malicious Code (2)


Trojan horse - A computer program that appears to have a
useful function, but also has a hidden and potentially
malicious function that evades security mechanisms,
sometimes by exploiting legitimate authorizations of a
system entity that invokes the program
Virus - A hidden, self-replicating section of computer
software, usually malicious logic, that propagates by
infecting (i.e., inserting a copy of itself into and becoming part of)
another program. A virus cannot run by itself; it requires
that its host program be run to make the virus active.

Worm - A computer program that can run independently,
can propagate a complete working version of itself onto
other hosts on a network, and may consume computer
resources destructively.
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
35
Kinds of Malicious Code (3)




Bacterium - A specialized form of virus which does not attach to a specific file.
Usage obscure.
Logic bomb - Malicious [program] logic that activates when
specified conditions are met. Usually intended to cause
denial of service or otherwise damage system resources.

Time bomb - activates when specified time occurs
Rabbit – A virus or worm that replicates itself without limit
to exhaust resource
Trapdoor / backdoor - A hidden computer flaw known to an
intruder, or a hidden computer mechanism (usually
software) installed by an intruder, who can activate the trap
door to gain access to the computer without being blocked
by security services or mechanisms.
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
36
Kinds of Malicious Code (4)



Above terms not always used consistently, esp. in popular
press
Combinations of the above kinds even more confusing

E.g., virus can be a time bomb
— spreads like virus, „explodes” when time occurs
Term „virus” often used to refer to any kind of malicious
code

When discussing malicious code, we’ll often say „virus”
for any malicious code
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
37
c. How Viruses Work (1)



Pgm containing virus must be executed to spread virus or
infect other pgms
 Even one pgm execution suffices to spread virus widely
Virus actions: spread / infect
Spreading – Example 1: Virus in a pgm on installation CD
 User activates pgm contaning virus when she runs
INSTALL or SETUP
 Virus installs itself in any/all executing pgms present in
memory
 Virus installs itself in pgms on hard disk

From now on virus spreads whenever any of the infected
pgms (from memory or hard disk) executes
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
38
How Viruses Work (2)



Spreading – Example 2: Virus in attachment to e-mail msg
 User activates pgm contaning virus (e.g. macro in MS
Word) by just opening the attachment
=> Disable automatic opening of attachments!!!
 Virus installs itself and spreads ... as in Example 1...
Spreading – Example 3: Virus in downloaded file
 File with pgm or document (.doc, .xls, .ppt, etc.)
 You know the rest by now...
Document virus

Spreads via picture, document, spreadsheet, slide
presentation, database, ...


E.g., via .jpg, via MS Office documents .doc, .xls, .ppt, .mdb
Currently most common!
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
39
How Viruses Work (3)

Kinds of viruses w.r.t. way of attaching to infected pgms
1) Appended viruses
 Appends to pgm
 Most often virus code precedes pgm code
 Inserts its code before the 1st pgm instruction in
executable pgm file
 Executes whenever program executed
2) Surrounding viruses
 Surronds program
 Executes before and after infected program
 Intercepts its input/output
 Erases its tracks
 The „after” part might be used to mask virus
existence
E.g. if surrounds „ls”, the „after” part removes listing of
virus file produced by „ls” so user can’t see it
... cont. ...
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
40
How Viruses Work (4)
... cont. ...
3) Integrating viruses
 Integrates into pgm code

Spread within infected pgms
4) Replacing viruses
 Entirely replaces code of infected pgm file
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
41
How Viruses Work (5)

(Replacing) virus V gains control over target pgm T by:
 Overwriting T on hard disk
OR
 Changing pointer to T with pointer to V
(textbook, Fig. 3-7)



OS has File Directory
File Directory has an entry that points to file with code for T
Virus replaces pointer to T’s file with pointer to V’s file
In both cases actions of V replace actions of T when user
executes what she thinks is „T”
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
42
How Viruses Work (6)

Characteristics of a ‘perfect’ virus (goals of virus writers)
 Hard to detect
 Not easily destroyed or deactivated
 Spreads infection widely
 Can reinfect programs
 Easy to create
 Machine and OS independent
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
43
How Viruses Work (7)

Virus hiding places
1) In bootstrap sector – best place for virus
 Bec. virus gains control early in the boot process
 Before detection tools are active!
Before infection:
After infection:
[Fig. cf. J. Leiwo & textbook]
2) In memory-resident pgms
 TSR pgms (TSR = terminate and stay resident)
 Most frequently used OS pgms or specialized user pgms
=> good place for viruses (activated very often)
...cont...
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
44
How Viruses Work (8)
...cont...
3) In application pgms
 Best for viruses: apps with macros
(MS Word, MS PowerPoint, MS Excel, MS Access, ...)
One macro: startup macro executed when app starts
Virus instructions attach to startup macro, infect
document files
Bec. doc files can include app macros (commands)
E.g., .doc file include macros for MS Word
Via data files infects other startup macros, etc. etc.
4) In libraries
 Libraries used/shared by many pgms => spread virus
 Execution of infected library pgm infects
5) In other widely shared pgms
 Compilers / loaders / linkers
 Runtime monitors
 Runtime debuggers
 Virus control pgms (!)
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
45
d. Virus Signatures (1)

Virus hides but can’t become invisible – leaves behind a virus
signature, defined by patterns:
1) Storage patterns : must be stored somewhere/somehow
(maybe in pieces)
2) Execution patterns: executes in a particular way
3) Distribution patterns: spreads in a certain way

Virus scanners use virus signatures to detect viruses
(in boot sectior, on hard disk, in memory)


Scanner can use file checksums to detect changes to files
Once scanner finds a virus, it tries to remove it


i.e., tries to remove all pieces of a virus V from target pgm T
Virus scanner and its database of virus signatures must be upto-date to be effective!

Update and run daily!
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
46
Virus Signatures (2)
Detecting Virus Signatures (1)

Difficulty 1 — in detecting execution patterns:
 Most of effects of virus execution (see next page) are
„invisible”

Bec. they are normal – any legitimate pgm could cause them
(hiding in a crowd)
=> can’t help in detecion
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
47
Virus Signatures (3)
Detecting Virus Signatures (2)
Virus Goal
How Achieved
Attach to executable
Attach to data/
control file
Remain in memory
Modify file directory / Write to executable pgm file
Modify directory / Rewrite data
Append to data / Append data to self
Intercept interrupt by modifying interrupt handler
address table / Load self in non-transient memory area
Intercept interrupt /Intercept OS call (e.g., to format
disk)
Modify system file / Modify ordinary executable pgm
Intercept system calls that would reveal self and falsify
results / Classify self as “hidden” file
Infect boot sector / Infect systems pgm
Infect ordinary pgm / Infect data ordinary pgm reads to
control its executable
Activate before deactivating pgmand block deactivation
Store copy to reinfect after deactivation
Infect disks
Conceal self
Spread self
Prevent deactivation
[cf. textbook & B. Endicott-Popovsky]
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
48
Virus Signatures (4)
Detecting Virus Signatures (3)


Difficulty 2 — in finding storage patterns:
 Polymorphic viruses:
changes from one „form” (storage pattern) to another
 Simple virus always recognizable by a certain char pattern
 Polymorphic virus mutates into variety of storage patterns
Examples of polymorphic virus mutations
 Randomly repositions all parts of itself and randomly
changes all fixed data within its code



Repositioning is easy since (infected) files stored as chains of data
blocks - chained with pointers
Randomly intersperses harmless instructions throughout its
code
(e.g., add 0, jump to next instruction)
Encrypting virus: Encrypts its object code (each time with a
different/random key), decrypts code to run
... More below ...
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
49
Virus Signatures (5)
Detecting Virus Signatures (4)

Encrypting virus structure
stored
encryp
-ted
(informal pseudo-code)
array decr_key;
procedure decrypt(virus_code, decr_key)
...
end /* decrypt */
begin /* virus V in target pgm T */
decrypt (V, decr_key);
infect: if infect_condition met then
find new target pgms NT to infect;
mutate V into V’ for copying;
encrypt V’ with random key into V”;
save new key in file for V”;
attach V” to NT;
hide modification of NT (with stealth
code of V);
damage: if damage_condition met then
execute damage_code of V
else start T
end /* virus V in target pgm T */
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
50
Virus Signatures (6)
Detecting Virus Signatures (5)

Encrypting virus: Encrypts its object code (each time with a
different/random key), decrypts code to run

Q: Is there any signature for encryption virus that a
scanner can see?
 Hint: consider 3 parts of encryption virus:
 „proper” virus code (infect/damage code)
 decr_key
 procedure decrypt
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
51
Virus Signatures (7)
Detecting Virus Signatures (6)
...
 Q: Q: Is there any signature for encryption virus that a
scanner can see?
 A: Lets’ see:
 „proper” virus code – encrypted with random key –
polymorphic
 decr_key – random key used to encrypt/decrypt –
polymorphic
 procedure decrypt (or a pointer to a library decrypt procedure)
– unencrypted, static
=> procedure decrypt of V is its signature
visible to a scanner
 But: Virus writer can use polymorphic techniques on
decryption code to make it „less visible” (to hide it)
Virus writers and scanner writers challenge each other
 An endless game?


Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
52
e. Preventing Virus Infections

Preventing Virus Infections
 Use commercial software from
trustworthy sources
But even this is not an absolute
guarantee of virus-free code!






Test new software on isolated computers
Open only safe attachments
Keep recoverable system image in safe place
Backup executable system files
Use virus scanners often (daily)
Update virus detectors daily
 Databases of virus signatures change very often
[cf. B. Endicott-Popovsky]

No absolute guarantees even if you follow all the rules –
just much better chances of preventing a virus
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
53
f. Seven Truths About Viruses





Viruses can infect any platform
Viruses can modify “hidden” / “read only” files
Viruses can appear anywhere in system
Viruses spread anywhere sharing occurs
Viruses cannot remain in memory aftera complete power
off/power on on reboot


Viruses infect software that runs hardware


But virus reappears if saved on disk (e.g., in the boot sector)
There are firmware viruses (if firmware writeable by s/w)
Viruses can be malevolent, benign, or benevolent

Hmmm...
Would you like a benevolent virus doing good things (like compressing
pgms to save storage) but without your knowledge?
Section 3 – Computer Security and Information Assurance – Spring 2006
[cf. B.
Endicott-Popovsky]
© 2006 by Leszek T. Lilien
54
g. Case Studies (1)

The Internet Worm
 Attacked on 11/2/1988
 Invaded VAX and Sun-3 computers running versions of
Berkeley UNIX
 Used their resources to attack still more computers
 Within hours spread across the U.S
 Infected hundreds / thousands of computers – serious
damage to Internet


Some uninfected networks were scared into disconnecting from
Internet => severed connections stopped necessary work
Made many computers unusable via resource exhaustion
Was a rabbit – supposedly by mistake unintended by its writer
Perpetrator was convicted in 1990 ($10,000 fine + 400 hrs of
community service + 3-year suspended jail sentence)



Caused forming Computer Emergency Response Team
(CERT) at CMU
Section 3 – Computer Security and Information Assurance – Spring 2006
[cf. textbook &
B. Endicott-Popovsky]
© 2006 by Leszek T. Lilien
55
Case Studies (2)

Other case studies [textbook – interesting reading]


The Brain (Pakistani) Virus (1986)
Code Red (2001)


Denial-of-service (DoS) attack on www.whitehouse.gov
Web Bugs (generic potentially malicious code on web
pages)





Placing a cookie on your hard drive
Cookie collects statistics on user’s surfing habits
Can be used to get your IP address, which can then be used to
target you for attack
Block cookies or delete cookies periodically (e.g., using browser
command; in MS IE: Tools>Internet Options-General:Delete
Cookies)
Tool: Bugnosis from Privacy Foundation – locates web bugs
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
56
h. Virus Removal and
System Recovery After Infection

Fixing a system after infection by virus V:
1) Disinfect (remove) viruses (using antivirus pgm)
 Can often remove V from infected file for T w/o
damaging T
 if V code can be separated from T code and V did
not corrupt T
 Have to delete T if can’t separate V from T code
2) Recover files:
- deleted by V
- modified by V
- deleted during disinfection (by antivirus pgm)
=> need file backups!
Make sure to have backups of (at least) important files
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
57
3.3.2. Targeted Malicious Code

Targeted = written to attack a particular system, a
particular application, and for a particular purpose

Many virus techniques apply
Some new techniques as well

Outline:
a. Trapdoors
b. Salami attack
c. Covert channels
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
58
a. Trapdoors (1)


Original def:
Trapdoor / backdoor - A hidden computer flaw known to an
intruder, or a hidden computer mechanism (usually
software) installed by an intruder, who can activate the trap
door to gain access to the computer without being blocked
by security services or mechanisms.
A broader definition:
Trapdoor – an undocumented entry point to a module
 Inserted during code development
 For testing
 As a hook for future extensions
 As emergency access in case of s/w failure
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
59
Trapdoors (2)

Testing:
 With stubs and drivers for unit testing (Fig. 3-10 p. 138)
 Testing with debugging code inserted into tested
modules


May allow programmer to modify internal module variables
Major sources of trapdoors:
 Left-over (purposely or not) stubs, drivers, debugging code
 Poor error checking


E.g., allowing for unacceptable input that causes buffer overflow

Some were used for testing, some random
Undefined opcodes in h/w processors
Not all trapdoors are bad


Some left purposely w/ good intentions
— facilitate system maintenance/audit/testing
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
60
b. Salami attack
Salami attack - merges bits of seemingly inconsequential
data to yield powerful results



Old example: interest calculation in a bank:
 Fractions of 1 ¢ „shaved off” n accounts and deposited in
attacker’s account
 Nobody notices/cares if 0.1 ¢ vanishes
 Can accumulate to a large sum
Easy target for salami attacks: Computer computations
combining large numbers with small numbers
 Require rounding and truncation of numbers
 Relatively small amounts of error from these op’s are
accepted as unavoidable – not checked unless a strong
suspicion
 Attacker can hide „salami slices” within the error margin
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
61
c. Covert Channels (CC) (1)

Outline:
i. Covert Channels - Definition and Examples
ii. Types of Covert Channels
iii. Storage Covert Channels
iv. Timing Covert Channels
v. Identifying Potential Covert Channels
vi. Covert Channels - Conclusions
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
62
i. CC – Definition and Examples (1)


So far: we looked at malicious pgms that perform wrong
actions
Now: pgms that disclose confidential/secret info
 They violate confidentiality, secrecy, or privacy of info
Covert channels = channels of unwelcome disclosure of info



Extract/leak data clandestinely
Examples
1) An old military radio communication network

The busiest node is most probably the command center
 Nobody is so naive nowadays
2) Secret ways spies recognize each other



Holding a certain magazine in hand
Exchanging a secret gesture when approaching each other
...
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
63
Covert Channels – Definition and Examples (2)
How programmers create covert channels?




Providing pgm with built-in Trojan horse
Uses covert channel to communicate extracted data
Example: pgm w/ Trojan horse using covert channel
 Should be:
Protected
Legitimate
Data
<------[ Service Pgm ]------> User

Is:
Protected
Legitimate
Data
<------[ Service Pgm ]------> User
[ w/ Trojan h. ]
covert channel
Spy
(Spy - e.g., programmer who put Trojan into pgm;
directly or via Spy Pgm)
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
64
Covert Channels – Definition and Examples (3)


How covert channels are created?
I.e., How leaked data are hidden?
Example: leaked data hidden in output reports (or displays)
 Different ‘marks’ in the report: (cf. Fig. 3-12, p.143)



Varying report format
Changing line length / changing nr of lines per page
Printing or not certain values, characters, or headings
- each ‘mark’ can convey one bit of info
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
65
Covert Channels – Definition and Examples (4)

Example – ctd.
 How Trojan within pgm can leak a 4-bit value of a
protected variable X?
cf. Fig. 3-12, p.143
Trojan signals value of X as follows:




Bit-1
Bit-2
Bit-3
Bit-4
=
=
=
=
1
1
1
1
if
if
if
if
>1 space follows ‘ACCOUNT CODE:’; 0 otherwise
last digit in ‘seconds’ field is >5; 0 otherwise
heading uses ‘TOTALS’; 0 otherwise (uses ‘TOTAL’)
no space follows subtotals line; 0 otherwise
=> For the values as in this Fig,
Trojan signaled and spy got: X = ‘1101’
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
66
ii. Types of Covert Channels
Types of covert channels


Storage covert channels
 Convey info by presence or absence of an object in
storage

Timing covert channels
 Convey info by varying the speed at which things
happen
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
67
iii. Storage Channels (1)

Example of storage channel: file lock covert channel
 Protected variable X has n bits: X1, ..., Xn
 Trojan within Service Pgm leaks value of X
 Trojan and Spy Pgm synchronized, so can „slice” time
into n intervals
 File FX (not used by anybody else)
 To signal that Xk=1, Trojan locks file FX for interval k (1≤
k ≤ n)
To signal that Xk=0, Trojan unlocks file FX for interval k
 Spy Pgm tries to lock FX during each interval
If it succeds during k-th interval, Xk = 0 (FX was unlocked)
Otherwise, Xk = 1 (FX was locked)
(see Fig. 3-13, 3-14 – p.144-145)

Q: Why FX should not be used by anybody else?
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
68
Storage Channels (2)

Example of storage channel: file lock covert channel
 ...


Q: Why FX should not be used by anybody else?
A: Any other user lockin/unlocking FX would interfere with
Trojan’s covert channel signaling.

Isn’t such bit-by-bit signaling too slow?
No – bec. computers are very fast!
 E.g., 10-100 bits/millisecond (10K – 100K b/s) is very slow
for computers
It still can leak entire P&P textbook in just minutes
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
69
Storage Channels (3)

Examples of covert storage channels (synchronized intervals!)
Covert channels can use:
 File locks (discussed above)

Disk storage quota
 To signal Xk=1, Trojan create enormous file (consuming
most of available disk space)

Spy Pgm attempts to create enormous file. If Spy fails
(bec. no disk space available), Xk = 1; otherwise, Xk = 0

Existence of a file
 To signal Xk=1, Trojan creates file FX (even empty file)
 Spy Pgm atempts to create file named FX. If Spy fails
(bec. FX already exists), Xk = 1; otherwise, Xk = 0

Other resources - similarly
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
70
Storage Channels (4)

Covert storage channels require:
 Shared resource
 To indicate Xk=1 or Xk=0
 Synchronized time
 To know which bit is signaled:
in interval k, Xk is signaled
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
71
iv. Timing Channels

Recall: Timing channels convey info by varying the speed
at which things happen

Simple example of timing channel:
 Multiprogramming system „slices” processor time for
programs running on the processor
 2 processes only: Trojan (Pgm w/ Trojan) and Spy Pgm
 Trojan receives all odd slices (unless abstains)
Spy Pgm receives all even slices (unless abstains)
 Trojan signals Xk=1 by using its time slice,
signals Xk=0 by abstaining from using its slice
see: Fig.3-15, p.147 – how ‘101’ is signaled
Details: Trojan takes Slice 1 (its 1st slice) signaling X1=1
Trojan abstains from taking Slice 3 (its 2nd slice) signaling X2=0
Trojan takes Slice 5 (its 3rd slice) signaling X3=1
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
72
v. Identifying Potential Covert Channels

Covert channels are not easy to identify
 Otherwise wouldn’t be covert, right?

Two techniques for locating covert channels:
1) Shared Resource Matrix
2) Information Flow Method
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
(1)
73
Identifying Potential Covert Channels (2)
1) The Shared Resource Matrix method


Shared resource is basis for a covert channel
=> identify shared resources and processes
reading/writing them
Step 1: Construct Shared Resource Matrix
Rows — resources
Columns — processes that access them:
R = observe resource M = modify/set/create/delete resource
Example
Lock on FX
X (confid.)
Process 1
Process 2
R, M
R
R, M
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
74
Identifying Potential Covert Channels (3)


...
Pgm 1
Pgm 2
Lock on FX
R, M
R, M
X (confid.)
R
Step 2: Look for pattern:
Meaning of this pattern:
Process Pj can get value of
Resource Rn via Process Pi
(and a covert channel)

Pi
Pj
Rm
M
R
Rn
R
Q: Do you see such a pattern in SRM above?
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
75
Identifying Potential Covert Channels (4)


...
Process 1
Process 2
Lock on FX
R, M
R, M
X (confid.)
R
Step 2: Look for pattern:
Meaning of this pattern:
Process Pj can get value of
Resource Rn via Process Pi
(and a covert channel)


i
j
m
M
R
n
R
Q: Do you see such a pattern in SRM above?
A: Yes. Process 2 can get value of X via Process 1
(no surprise: Proc. 1 & 2 are Trojan & Spy
from earlier example)
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
76
Identifying Potential Covert Channels (5)
2) Information Flow Method

Flow analysis of pgm’s syntax


Can be automated within a compiler
Identifies non-obvious flows of info between pgm
statements
Examples of flows of info between pgm stmts




B:= A – an explicit flow from A to B
B:= A; C:=B – an explicit flow from A to C (via B)
IF C=1 THEN B:=A
– an explicit flow from A to B
– an implicit flow from C to B (bec. B can change iff C=1)
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
77
Identifying Potential Covert Channels (6)

More examples of flows of info between pgm stmts
[textbook and J. Leiwo]
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
78
Identifying Potential Covert Channels (7)

Steps of Information Flow Method (IFM)
1) Analyze statements
2) Integrate results to see which outputs affected by which
inputs

Variants of IFM:
1) IFM during compilation
2) IFM on design specs
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
79
Covert Channels - Conclusions

Covert channels are a serious threat to confidentiality and
thus security
(„CIA” = security)

Any virus/Trojan horse can create a covert channel

In open systems — no way to prevent covert channels
 Very high security systems require a painstaking and
costly design preventing (some) covert channels
 Analysis must be performed periodically as high security
system evolves
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
80
3.4. Controls for Security


How to control security of pgms during their development
and maintenance
Outline:
a. Introduction
b. Developmental controls for security
c. Operating system controls for security
d. Administrative controls for security
e. Conclusions
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
81
a. Introduction



„Better to prevent than to cure”
Preventing security flaws
 We have seen a lot of possible security flaws
 How to prevent (some of) them?
 Software engineering concentrates on developing and
maintaining quality s/w
 We’ll take a look at some techniques useful specifically
for developing/ maintaining secure s/w
Three types of controls for security (against pgm flaws):
1) Developmental controls
2) OS controls
3) Administrative controls
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
82
b. Developmental Controls for Security (1)

Nature of s/w development
 Collaborative effort
 Team of developers, each involved in  1 of stages:









Requirement specification
 Regular req. specs: „do X”
 Security req. specs: „do X and nothing more”
Design
Implementation
Testing
Documenting at each stage
Reviewing at each stage
Managing system development thru all stages
Maintaining deployed system (updates, patches, new versions,
etc.)
Both product and process contribute to overall quality
— incl. security dimension of quality
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
83
Developmental Controls for Security (2)

Fundamental principles of s/w engineering
1) Modularity
2) Encapsulation
3) Info hiding
1) Modularity
 Modules should be:
 Single-purpose - logically/functionally
 Small - for a human to grasp
 Simple - for a human to grasp
 Independent – high cohesion, low coupling



High cohesion – highly focused on (single) purpose
Low coupling – free from interference from other modules
Modularity should improve correctness
 Fewer flaws => better security
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
84
Developmental Controls for Security (3)
2) Encapsulation
 Minimizing info sharing with other modules
=> Limited interfaces reduce # of covert channels
 Well documented interfaces
 „Hiding what should be hidden and showing what should
be visible.”
3) Information hiding



Module is a black box
 Well defined function and I/O
Easy to know what module does but not how it does it
Reduces complexity, interactions, covert channels, ...
=> better security
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
85
Developmental Controls for Security (4)

Techniques for building solid software
1) Peer reviews
2) Hazard analysis
3) Testing
4) Good design
5) Risk prediction & mangement
6) Static analysis
7) Configuration management
8) Additional developmental controls
... Please read on your own ...
..Also see slides—all discussed below ...
Section 3 – Computer Security and Information Assurance – Spring 2006
[cf. B.
Endicott-Popovsky]
© 2006 by Leszek T. Lilien
86
Developmental Controls for Security (5)
1) Peer reviews - three types
 Reviews
 Informal
 Team of reviewers
 Gain consensus on solutions
before development
 Walk-throughs
 Developer walks team through code/document
 Discover flaws in a single design document
 Inspection
 Formalized and detailed
 Statistical measures used

Various types of peer reviews can be highly effective
Section 3 – Computer Security and Information Assurance – Spring 2006
[cf. ©
B.2006
Endicott-Popovsky]
by Leszek T. Lilien
87
Developmental Controls for Security (6)
2) Hazard analysis
= systematic techniques to expose
potentially hazardous system states,
incl. security vulnerabilities


Components of HA
 Hazard lists
 What-if scenarios – identifies non-obvious hazards
 System-wide view (not just code)
 Begins Day 1
 Continues throughout SDLC (= s/w dev’t life cycle)
Techniques
 HAZOP – hazard and operability studies
 FMEA – failure modees and effects analysis
 FTA – fault tree analysis
Section 3 – Computer Security and Information Assurance – Spring 2006
[cf. B.
Endicott-Popovsky]
© 2006 by Leszek T. Lilien
88
Developmental Controls for Security (7)
3) Testing – phases:
 Module/component/unit testing of indiv. modules
 Integration testing of interacting (sub)system modules
 (System) function testing checking against requirement specs
 (System) performance testing
 (System) acceptance testing – with customer against
customer’s requirements — on seller’s or customer’s premises

(System) installation testing after installation on customer’s
system


Regression testing after updates/changes to s/w
Types of testing
 Black Box testing – testers can’t examine code
 White Box / Clear box testing – testers can examine design and
code, can see inside modules/system
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
89
Developmental Controls for Security (8)
4) Good design

Good design uses:
i. Modularity / encapsulation / info hiding
ii. Fault tolerance
iii. Consistent failure handling policies
iv. Design rationale and history
v. Design patterns
i. Using modularity / encapsulation / info hiding
- as discussed above
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
90
Developmental Controls for Security (9)
4) Good design – cont.1a
ii. Using fault tolerance for reliability and security
 System tolerates component failures
 System more reliable than any of its components


Different than for security, where system is as secure as its
weakest component
[cf. B. Endicott-Popovsky]
Fault-tolerant approach:
 Anticipate faults
(car: anticipate having a flat tire)
 Active fault detection rather than pasive fault detection
(e.g., by use of mutual suspicion: active input data checking)



Use redundancy
Isolate damage
Minimize disruption
(car: have a spare tire)
(car: replace flat tire, continue your trip)
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
91
Developmental Controls for Security (10)
4) Good design – cont.1b


Example 1: Majority voting (using h/w redundancy)
 3 processor running the same s/w
 E.g., in a spaceship
 Result accepted if results of  2 processors agree
Example 2: Recovery Block (using s/w redundancy)
Primary Code
e.g., Quick Sort
Secondary
Code
e.g., Bubble Sort
Acceptance Test
Section 3 – Computer Security and Information Assurance – Spring 2006
Quick Sort –
– new code (faster)
Bubble Sort –
– well-tested code
© 2006 by Leszek T. Lilien
92
Developmental Controls for Security (11)
4) Good design – cont.2
iii. Using consistent failure handling policies
 Each failure handled by one of 3 ways:
 Retrying


Correcting


Restore previous state, correct sth, run service using the
same code as before
Reporting


Restore previous state, redo service using different „path”
 E.g., use secondary code instead of primary code
Restore previous state, report failure to error handler, don’t
rerun service
Example — How fault-tolerance enhances security
 If security fault destroys important data (availability in CIA),
use f-t to revert to backup data set
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
93
Developmental Controls for Security (12)
4) Good design – cont.3
iv. Using design rationale and history

Knowing it (incl. knowing design rationale and history
for security mechanisms) helps developers modifying or
maintaining system
v. Using design patterns

Knowing it enables looking for patterns showing what
works best in which situation
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
94
Developmental Controls for Security (13)

Value of Good Design
 Easy maintenance
 Understandability
 Reuse
 Correctness
 Better testing
=> translates into (saving) BIG bucks !
Section 3 – Computer Security and Information Assurance – Spring 2006
[cf. ©
B.2006
Endicott-Popovsky]
by Leszek T. Lilien
95
Developmental Controls for Security (14)
5) Risk prediction & management
 Predict and manage risks involved in system development
and deployment
 Make plans to handle unwelcome events should they
occur
 Risk prediction/mgmt are esp. important for security
 Bec. unwelcome and rare events can have security
consequences
 Risk prediction/mgmt helps to select proper security
controls (e.g., proportional to risk)
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
96
Developmental Controls for Security (15)
6) Static analysis
 Before system is up and running, examine its design and
code to locate security flaws


More than peer review
Examines
 Control flow structure
(sequence in which instructions are
executed, incl. iterations and loops)
Data flow structure (trail of data)
 Data structures
Automated tools available


Section 3 – Computer Security and Information Assurance – Spring 2006
[cf. B.
Endicott-Popovsky]
© 2006 by Leszek T. Lilien
97
Developmental Controls for Security (16)
7) Configuration management
= process of controling system modifications during
development and maintenance
 Offers security benefits by scrutinizing new/changed code

Problems with system modifications
 One change interefering with other change


E.g., neutralizing it
Proliferation of different versions and releases



Older and newer
For different platforms
For different application environments (and/or customers
categories)
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
98
Developmental Controls for Security (17)

Reasons for software modification
 Corrective changes
 To maintain control of system’s day-to-day functions



Adaptive changes
 To maintain control over system’s modifications
Perfective changes
 To perfect existing acceptable system functions
Preventive changes
 To prevent system’s performance degradation to
unacceptable levels
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
99
Developmental Controls for Security (18)

Activities involved in configuration management process
(performed by reps from developers, customers, users, etc.)
1) Baseline identification
 Certain release/version (R/v) selected & frozen as
baseline
 Other R’s/v’s described as changes to the baseline
2) Configuration control and configuration management
 Coordinate separate but related v’s (versions) via:
 Separate files - separate files for each R or v
 Deltas - main v defined by „full files”
- other v’s defined by main v & deltas
(= difference files)
 Conditional compilation
- single source code file F for all v’s
uses begin_version_Vx / end_version_Vx brackets
or begin_not_version_Vx / end_not_version_Vx brackets
- compiler produces each v from F
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
100
Developmental Controls for Security (19)
3) Configuration auditing
 System must be audited regularly — to verify:
 Baseline completeness and accuracy
 Recording of changes
 Accuracy of software documentation for systems in
the field
 Peformed by independent parties
4) Status accounting
 Records info about system components
 Where they come from (purchased, reused, written
from scratch)
 Version
 Change history
 Pending change requests
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
101
Developmental Controls for Security (20)


All 4 activities performed by
Configuration Control Board (CCB)
 Includes reps from developers, customers, users
 Reviews proposed changes, approves/rejects
Security benefits of configuration mgmt
 Limits unintentional flaws
 Limits malicious modifications
by protecting integrity of pgms and documentation

Thanks to:
 careful reviewing/auditing, change mgmt
 preventing changes (e.g., trapdoors) to system w/o acceptance
by CCB
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
102
Developmental Controls for Security (21)
8) Additional developmental controls
8a) Learning from mistakes
 Avoiding such mistakes in the future enhances security
8b) Proofs of program correctness
 Formal methods to verify pgm correctness
 Logic analyzer shows that:
initial assertions about inputs...
... through implications of pgm statements...
... lead to the terminal condition (desired output)
 Problems with practical use of pgm correctness proofs
 Esp. for large pgms/systems
 Most successful for specific types of apps
 E.g. for communication protocols & security policies
Even with all these developmental controls (1-8) –
still no security guarantees! [cf. B. Endicott-Popovsky]
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
103
c. Operating System Controls for Security (1)
Developmental controls not always used
OR:
 Even if used, not foolproof
=> Need other, complementary controls, incl. OS controls


Such OS controls can protect against some pgm flaws
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
104
Operating System Controls for Security (2)

Trusted software
– code rigorously developed an analyzed so we can trust that
it does all and only what specs say
 Trusted code establishes foundation upon which untrusted
code runs
 Trusted code establishes security baseline for the whole
system
 In particular, OS can be trusted s/w
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
105
Operating System Controls for Security (3)

Key characteristics determining if OS code is trusted
1) Functional correctness
 OS code consistent with specs
2) Enforcement of integrity
 OS keeps integrity of its data and other resources even
if presented with flawed or unauthorized commands
3) Limited privileges
 OS minimizes access to secure data/resources
 Trusted pgms must have „need to access” and proper
access rights to use resources protected by OS
 Untrusted pgms can’t access resources protected by OS
4) Appropriate confidence level
 OS code examined and rated at appropriate trust level
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
106
Operating System Controls for Security (4)


Similar criteria used to establish if s/w other than OS can be
trusted
Ways of increasing security if untrusted pgms present:
1) Mutual suspicion
2) Confinement
3) Access log
1) Mutual suspicion between programs

Distrust other pgms – treat them as if they were
incorrect or malicious

Pgm protects its interface data

With data checks, etc.
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
107
Operating System Controls for Security (5)
2) Confinement


OS can confine access to resources by suspected pgm
Example 1: strict compartmentalization

Pgm can affect data and other pgms only within its
compartment

Example 2: sandbox for untrusted pgms

Can limit spread of viruses
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
108
Operating System Controls for Security (6)
3) Audit log / access log



Records who/when/how (e.g., for how long)
accessed/used which objects

Events logged: logins/logouts, file accesses, pgm
ecxecutions, device uses, failures, repeated
unsuccessful commands (e.g., many repeated failed
login attempts can indicate an attack)
Audit frequently for unusual events, suspicious patterns
Forensic measure not protective measure

Forensics – investigation to find who broke law,
policies, or rules
...Much more on OS controls soon...
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
109
d. Administrative Controls for Security (1)


They prohibit or demand certain human behavior via
policies, procedures, etc.
They include:
1) Standards of program development
2) Security audits
3) Separation of duties
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
110
Administrative Controls for Security (2)
1) Standards and guidelines for program development
 Capture experience and wisdom from previous projects
 Facilitate building higher-quality s/w (incl. more secure)
 They include:
 Design S&G – design tools, languages, methodologies
 S&G for documentation, language, and coding style
 Programming S&G - incl. reviews, audits
 Testing S&G
 Configuration mgmt S&G
2) Security audits
 Check compliance with S&G
 Scare potential dishonest programmer from including
illegitimate code (e.g., a trapdoor)
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
111
Administrative Controls for Security (3)
3) Separation of duties
 Break sensitive tasks into  2 pieces to be performed by
different people (learned from banks)
 Example 1: modularity
 Different developers for cooperating modules
 Example 2: independent testers
 Rather than developer testing her own code
...More (much) later...
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
112
e. Conclusions



(for Controls for Security)
Developmental / OS / administrative controls help
produce/maintain higher-quality (also more secure) s/w
Art and science - no „silver bullet” solutions
„A good developer who truly understands security will
incorporate security into all phases of development.”
[textbook, p. 172]

Summary:
Control
[cf. B. Endicott-Popovsky]
Purpose
Benefit
Developmental
Limit mistakes
Make malicious code difficult
Produce better software
Operating
System
Limit access to system
Promotes safe sharing of info
Administrative
Limit actions of people
Improve usability, reusability
and maintainability
Section 3 – Computer Security and Information Assurance – Spring 2006
© 2006 by Leszek T. Lilien
113
End of:
Section 3: Program Security
Download