Скачать презентацию Introduction to Computer Security Lecture 4 Security Policies Скачать презентацию Introduction to Computer Security Lecture 4 Security Policies

fd0f9ae94c274387abfe0d7b317825c1.ppt

  • Количество слайдов: 40

Introduction to Computer Security Lecture 4 Security Policies, Confidentiality and Integrity Policies September 27, Introduction to Computer Security Lecture 4 Security Policies, Confidentiality and Integrity Policies September 27, 2003 IS 2150/TEL 2810: Introduction of Computer Security 1

Security Policy l Defines what it means for a system to be secure l Security Policy l Defines what it means for a system to be secure l Formally: Partitions a system into ¡ Set of secure (authorized) states ¡ Set of non-secure (unauthorized) states l Secure system is one that ¡ Starts in authorized state ¡ Cannot enter unauthorized state IS 2150/TEL 2810: Introduction of Computer Security 2

Secure System - Example Unauthorized states A B C D Authorized states l Is Secure System - Example Unauthorized states A B C D Authorized states l Is this Finite State Machine Secure? ¡ ¡ ¡ A is start state ? B is start state ? C is start state ? How can this be made secure if not? Suppose A, B, and C are authorized states ? IS 2150/TEL 2810: Introduction of Computer Security 3

Additional Definitions: Security breach: system enters an unauthorized state l Let X be a Additional Definitions: Security breach: system enters an unauthorized state l Let X be a set of entities, I be information. l ¡ ¡ I has confidentiality with respect to X if no member of X can obtain information on I I has integrity with respect to X if all members of X trust I l l l ¡ Trust I, its conveyance and protection (data integrity) I maybe origin information or an identity (authentication) I is a resource – its integrity implies it functions as it should (assurance) I has availability with respect to X if all members of X can access I l Time limits (quality of service IS 2150/TEL 2810: Introduction of Computer Security 4

Confidentiality Policy l Also known as information flow ¡ Transfer of rights ¡ Transfer Confidentiality Policy l Also known as information flow ¡ Transfer of rights ¡ Transfer of information without transfer of rights ¡ Temporal context l Model often depends on trust ¡ Parts of system where information could flow ¡ Trusted entity must participate to enable flow l Highly developed in Military/Government IS 2150/TEL 2810: Introduction of Computer Security 5

Integrity Policy l Defines how information can be altered Entities allowed to alter data Integrity Policy l Defines how information can be altered Entities allowed to alter data ¡ Conditions under which data can be altered ¡ Limits to change of data ¡ l Examples: Purchase over $1000 requires signature ¡ Check over $10, 000 must be approved by one person and cashed by another ¡ l l Separation of duties : for preventing fraud Highly developed in commercial world IS 2150/TEL 2810: Introduction of Computer Security 6

Transaction-oriented Integrity l Begin in consistent state ¡ “Consistent” defined by specification l Perform Transaction-oriented Integrity l Begin in consistent state ¡ “Consistent” defined by specification l Perform series of actions (transaction) ¡ Actions cannot be interrupted ¡ If actions complete, system in consistent state ¡ If actions do not complete, system reverts to beginning (consistent) state IS 2150/TEL 2810: Introduction of Computer Security 7

Trust Theories and mechanisms rest on some trust assumptions l Administrator installs patch l Trust Theories and mechanisms rest on some trust assumptions l Administrator installs patch l 1. 2. 3. 4. Trusts patch came from vendor, not tampered with in transit Trusts vendor tested patch thoroughly Trusts vendor’s test environment corresponds to local environment Trusts patch is installed correctly IS 2150/TEL 2810: Introduction of Computer Security 8

Trust in Formal Verification l Formal verification provides a formal mathematical proof that given Trust in Formal Verification l Formal verification provides a formal mathematical proof that given input i, program P produces output o as specified l Suppose a security-related program S formally verified to work with operating system O l What are the assumptions? IS 2150/TEL 2810: Introduction of Computer Security 9

Trust in Formal Methods 1. Proof has no errors ¡ Bugs in automated theorem Trust in Formal Methods 1. Proof has no errors ¡ Bugs in automated theorem provers Preconditions hold in environment in which S is to be used 3. S transformed into executable S’ whose actions follow source code 2. ¡ 4. Compiler bugs, linker/loader/library problems Hardware executes S’ as intended ¡ Hardware bugs IS 2150/TEL 2810: Introduction of Computer Security 10

Security Mechanism l Policy describes what is allowed l Mechanism ¡ Is an entity/procedure Security Mechanism l Policy describes what is allowed l Mechanism ¡ Is an entity/procedure that enforces (part of) policy l Example Policy: homework ¡ Mechanism: other users Students should not copy Disallow access to files owned by l Does mechanism enforce policy? IS 2150/TEL 2810: Introduction of Computer Security 11

Security Model Security Policy: What is/isn’t authorized l Problem: Policy specification often informal l Security Model Security Policy: What is/isn’t authorized l Problem: Policy specification often informal l ¡ ¡ l Implicit vs. Explicit Ambiguity Security Model: Model that represents a particular policy (policies) ¡ ¡ ¡ Model must be explicit, unambiguous Abstract details for analysis HRU result suggests that no single nontrivial analysis can cover all policies, but restricting the class of security policies sufficiently allows meaningful analysis IS 2150/TEL 2810: Introduction of Computer Security 12

Common Mechanisms: Access Control l Discretionary Access Control (DAC) ¡ ¡ l Mandatory Access Common Mechanisms: Access Control l Discretionary Access Control (DAC) ¡ ¡ l Mandatory Access Control (MAC) ¡ ¡ l Rules specify granting of access Also called rule-based access control Originator Controlled Access Control (ORCON) ¡ ¡ l Owner determines access rights Typically identity-based access control: Owner specifies other users who have access Originator controls access Originator need not be owner! Role Based Access Control (RBAC) ¡ Identity governed by role user assumes IS 2150/TEL 2810: Introduction of Computer Security 13

Policy Languages l High-level: Independent of mechanisms Constraints expressed independent of enforcement mechanism ¡ Policy Languages l High-level: Independent of mechanisms Constraints expressed independent of enforcement mechanism ¡ Constraints restrict entities, actions ¡ Constraints expressed unambiguously ¡ l ¡ Requires a precise language, usually a mathematical, logical, or programming-like language Example: Domain-Type Enforcement Language l l l Subjects partitioned into domains Objects partitioned into types Each domain has set of rights over each type IS 2150/TEL 2810: Introduction of Computer Security 14

Example: Web Browser l Goal: restrict actions of Java programs that are downloaded and Example: Web Browser l Goal: restrict actions of Java programs that are downloaded and executed under control of web browser l Language specific to Java programs l Expresses constraints as conditions restricting invocation of entities IS 2150/TEL 2810: Introduction of Computer Security 15

Expressing Constraints l Entities are classes, methods Class: set of objects that an access Expressing Constraints l Entities are classes, methods Class: set of objects that an access constraint constrains ¡ Method: set of ways an operation can be invoked ¡ l Operations Instantiation: s creates instance of class c: s ├ c ¡ Invocation: s 1 executes object s 2: s 1 |→ s 2 ¡ l Access constraints deny(s op x) when b ¡ when b is true, subject s cannot perform op on (subject or class) x; empty s means all subjects ¡ IS 2150/TEL 2810: Introduction of Computer Security 16

Sample Constraints Downloaded program cannot access password database file on UNIX system l Program’s Sample Constraints Downloaded program cannot access password database file on UNIX system l Program’s class and methods for files: l class File { public file(String name); public String getfilename(); public char read(); …. l Constraint: deny(|→ file. read) when (file. getfilename() == “/etc/passwd”) IS 2150/TEL 2810: Introduction of Computer Security 17

Policy Languages l Low-level: close to mechanisms A set of inputs or arguments to Policy Languages l Low-level: close to mechanisms A set of inputs or arguments to commands that set, or check, constraints on a system ¡ Example: Tripwire: Flags what has changed ¡ l l Configuration file specifies settings to be checked History file keeps old (good) example IS 2150/TEL 2810: Introduction of Computer Security 18

Secure, Precise Mechanisms l Can one devise a procedure for developing a mechanism that Secure, Precise Mechanisms l Can one devise a procedure for developing a mechanism that is both secure and precise? Consider confidentiality policies only here ¡ Integrity policies produce same result ¡ l Program with multiple inputs and one output as an abstract function Let p be a function p: I 1 . . . In R. Then p is a program with n inputs ik Ik, 1 ≤ k ≤ n, and one output r R ¡ Goal: determine if P can violate a security requirement (confidentiality, integrity, etc. ) ¡ IS 2150/TEL 2810: Introduction of Computer Security 19

Programs and Postulates l Observability Postulate: ¡ the output of a function encodes all Programs and Postulates l Observability Postulate: ¡ the output of a function encodes all available information about its inputs l ¡ l Covert channels considered part of the output Output may contain things not normally thought of as part of function result Example: authentication function ¡ ¡ Inputs name, password; output Good or Bad If name invalid, print Bad; else access database Problem: time output of Bad, can determine if name valid This means timing is part of output IS 2150/TEL 2810: Introduction of Computer Security 20

Protection Mechanism . . . In R. A protection mechanism m is a function Protection Mechanism . . . In R. A protection mechanism m is a function m: I 1 . . . In R E for which, when ik Ik, 1 ≤ k ≤ n, either l Let p be a function p: I 1 ¡ m(i 1, . . . , in) = p(i 1, . . . , in) or ¡ m(i 1, . . . , in) E. l E is set of error outputs ¡ In above example, E = { “Password Database Missing”, “Password Database Locked” } IS 2150/TEL 2810: Introduction of Computer Security 21

Confidentiality Policy l Confidentiality policy for program p says which inputs can be revealed Confidentiality Policy l Confidentiality policy for program p says which inputs can be revealed ¡ ¡ l Formally, for p: I 1 . . . In R, it is a function c: I 1 . . . In A, where A I 1 . . . In A is set of inputs available to observer Security mechanism is function m: I 1 . . . In R E ¡ ¡ m secure iff m´: A R E such that, for all ik Ik, 1 ≤ k ≤ n, m(i 1, . . . , in) = m´(c(i 1, . . . , in)) m returns values consistent with c IS 2150/TEL 2810: Introduction of Computer Security 22

Examples l c(i 1, . . . , in) = C, a constant ¡ Examples l c(i 1, . . . , in) = C, a constant ¡ Deny observer any information (output does not vary with inputs) l c(i 1, . . . , in) = (i 1, . . . , in), and m´ = m ¡ Allow observer full access to information l c(i 1, . . . , in) = i 1 ¡ Allow observer information about first input but no information about other inputs. IS 2150/TEL 2810: Introduction of Computer Security 23

Precision l Security policy may be over-restrictive ¡ l Precision measures how over-restrictive m Precision l Security policy may be over-restrictive ¡ l Precision measures how over-restrictive m 1, m 2 distinct protection mechanisms for program p under policy c ¡ ¡ m 1 as precise as m 2 (m 1 m 2) if, for all inputs i 1, …, in - m 2(i 1, …, in) = p(i 1, …, in) - m 1(i 1, …, in) = p(i 1, …, in m 1 more precise than m 2 (m 1~m 2) if there is an input (i 1´, …, in´) such that - m 1(i 1´, …, in´) = p(i 1´, …, in´) and - m 2(i 1´, …, in´) ≠ p(i 1´, …, in´). IS 2150/TEL 2810: Introduction of Computer Security 24

Combining Mechanisms m 1, m 2 protection mechanisms l m 3 = m 1 Combining Mechanisms m 1, m 2 protection mechanisms l m 3 = m 1 m 2 defined as l p(i 1, …, in) when m 1(i 1, …, in) = p(i 1, …, in) or m 2(i 1, …, in) = p(i 1, …, in) ¡ else m 1(i 1, …, in) ¡ l Theorem: if m 1, m 2 secure, then m 3 secure m 1 m 2 secure ¡ m 1 m 2 ≈ m 1 and m 1 m 2 ≈ m 2 ¡ Proof follows from the definitions ¡ IS 2150/TEL 2810: Introduction of Computer Security 25

Modeling Secure/Precise: Confidentiality – existence theorem l Theorem: Given p and c, a precise, Modeling Secure/Precise: Confidentiality – existence theorem l Theorem: Given p and c, a precise, secure mechanism m* such that secure m for p and c, m* ≈ m ¡ Proof: Induction from previous theorem ¡ Maximally precise mechanism Ensures security Minimizes number of denials of legitimate actions ¡ ¡ l There is no effective procedure that determines a maximally precise, secure mechanism for any policy and program. IS 2150/TEL 2810: Introduction of Computer Security 26

Confidentiality Policies IS 2150/TEL 2810: Introduction of Computer Security 27 Confidentiality Policies IS 2150/TEL 2810: Introduction of Computer Security 27

Confidentiality Policy l Also known as information flow policy Integrity is secondary objective ¡ Confidentiality Policy l Also known as information flow policy Integrity is secondary objective ¡ Eg. Military mission “date” ¡ l Bell-La. Padula Model ¡ Formally models military requirements Information has sensitivity levels or classification l Subjects have clearance l Subjects with clearance are allowed access l ¡ Multi-level access control or mandatory access control IS 2150/TEL 2810: Introduction of Computer Security 28

Bell-La. Padula: Basics l Mandatory access control ¡ Entities are assigned security levels ¡ Bell-La. Padula: Basics l Mandatory access control ¡ Entities are assigned security levels ¡ Subject has security clearance L(s) = ls ¡ Object has security classification L(o) = lo ¡ Simplest case: Security levels are arranged in a linear order li < li+1 l Example Top secret > Secret > Confidential >Unclassified IS 2150/TEL 2810: Introduction of Computer Security 29

“No Read Up” l Information is allowed to flow up, not down l Simple “No Read Up” l Information is allowed to flow up, not down l Simple security property: ¡ s can read o if and only if l lo ≤ ls and l s has read access to o - Combines mandatory (security levels) and discretionary (permission required) - Prevents subjects from reading objects at higher levels (No Read Up rule) IS 2150/TEL 2810: Introduction of Computer Security 30

“No Write Down” l Information is allowed to flow up, not down l *property “No Write Down” l Information is allowed to flow up, not down l *property ¡ s can write o if and only if l ls ≤ lo and l s has write access to o - Combines mandatory (security levels) and discretionary (permission required) - Prevents subjects from writing to objects at lower levels (No Write Down rule) IS 2150/TEL 2810: Introduction of Computer Security 31

Example security level subject object Top Secret Tamara Personnel Files Secret Samuel E-Mail Files Example security level subject object Top Secret Tamara Personnel Files Secret Samuel E-Mail Files Confidential Claire Activity Logs Unclassified Ulaley Telephone Lists • Tamara can read which objects? And write? • Claire cannot read which objects? And write? • Ulaley can read which objects? And write? IS 2150/TEL 2810: Introduction of Computer Security 32

Access Rules l Secure system: ¡ One in which both the properties hold l Access Rules l Secure system: ¡ One in which both the properties hold l Theorem: Let Σ be a system with secure initial state σ0, T be a set of state transformations ¡ If every element of T follows rules, every state σi secure ¡ Proof - induction IS 2150/TEL 2810: Introduction of Computer Security 33

Categories l Total order of classifications not flexible enough ¡ l Alice cleared for Categories l Total order of classifications not flexible enough ¡ l Alice cleared for missiles; Bob cleared for warheads; Both cleared for targets Solution: Categories ¡ ¡ ¡ Use set of compartments (from power set of compartments) Enforce “need to know” principle Security levels (security level, category set) l l l (Top Secret, {Nuc, Eur, Asi}) (Top Secret, {Nuc, Asi}) Combining with clearance: ¡ ¡ (L, C) dominates (L’, C’) L’ ≤ L and C’ C Induces lattice of security levels IS 2150/TEL 2810: Introduction of Computer Security 34

Lattice of categories {Nuc, Eur, Us} l l Examples of levels ¡ (Top Secret, Lattice of categories {Nuc, Eur, Us} l l Examples of levels ¡ (Top Secret, {Nuc, Asi}) dom (Secret, {Nuc}) ¡ (Secret, {Nuc, Eur}) dom (Confidential, {Nuc, Eur}) ¡ (Top Secret, {Nuc}) dom (Confidential, {Eur}) {Nuc, Eur} {Nuc, Us} {Eur, Us} {Us} Bounds ¡ ¡ Greatest lower, Lowest upper glb of {X, Nuc, Us} & {X, Eur, Us}? lub of {X, Nuc, Us} & {X, Eur, Us}? IS 2150/TEL 2810: Introduction of Computer Security {} 35

Access Rules l Simple Security Condition: S can read O if and only if Access Rules l Simple Security Condition: S can read O if and only if ¡ ¡ l *-Property: S can write O if and only if ¡ ¡ l l S dominate O and S has read access to O O dom S and S has write access to O Secure system: One with above properties Theorem: Let Σ be a system with secure initial state σ0, T be a set of state transformations ¡ If every element of T follows rules, every state σi secure IS 2150/TEL 2810: Introduction of Computer Security 36

Problem: No write-down Cleared subject can’t communicate to non-cleared subject l Any write from Problem: No write-down Cleared subject can’t communicate to non-cleared subject l Any write from li to lk, i > k, would violate *property ¡ l Any read from lk to li, i > k, would violate simple security property ¡ l Subject at li can only write to li and above Subject at lk can only read from lk and below Subject at level i can’t write something readable by subject at k ¡ Not very practical IS 2150/TEL 2810: Introduction of Computer Security 37

Principle of Tranquility Should we change classification levels? l Raising object’s security level l Principle of Tranquility Should we change classification levels? l Raising object’s security level l ¡ ¡ ¡ l Information once available to some subjects is no longer available Usually assumes information has already been accessed Simple security property violated? Problem? Lowering object’s security level ¡ ¡ Simple security property violated? The declassification problem Essentially, a “write down” violating *-property Solution: define set of trusted subjects that sanitize or remove sensitive information before security level is lowered IS 2150/TEL 2810: Introduction of Computer Security 38

Types of Tranquility l Strong Tranquility ¡ l The clearances of subjects, and the Types of Tranquility l Strong Tranquility ¡ l The clearances of subjects, and the classifications of objects, do not change during the lifetime of the system Weak Tranquility ¡ The clearances of subjects, and the classifications of objects, do not change in a way that violates the simple security condition or the *-property during the lifetime of the system IS 2150/TEL 2810: Introduction of Computer Security 39

Example l DG/UX System ¡ Only a trusted user (security administrator) can lower object’s Example l DG/UX System ¡ Only a trusted user (security administrator) can lower object’s security level ¡ In general, process MAC labels cannot change If a user wants a new MAC label, needs to initiate new process l Cumbersome, so user can be designated as able to change process MAC label within a specified range l IS 2150/TEL 2810: Introduction of Computer Security 40