Knowledge Representation

advertisement
Knowledge Representation
The Edwin Smith papyrus

Title:


Symptoms:


If you examine a man with a fracture of the cheekbone, you will
find a salient and red fluxion, bordering the wound.
Diagnosis and prognosis:


Instructions for treating a fracture of the cheekbone.
Then you will tell your patient: "A fracture of the cheekbone. It is
an injury that I will cure."
Treatment:


You shall tend him with fresh meat the first day. The treatment
shall last until the fluxion resorbs.
Next you shall treat him with raspberry, honey, and bandages to
be renewed each day, until he is cured.
Searle’s Chinese Room



http://www.mind.ilstu.edu/curriculum/searle_chinese_room/searle_chinese_room.p
hp
Monolingual English speaker locked in a room,
Given





Those giving you the symbols call






a large batch of Chinese writing
a second batch of Chinese script
a set of rules in English for correlating the second batch with the first batch.
A third batch of Chinese symbols and more instructions in English enable you "to
correlate elements of this third batch with elements of the first two batches" and instruct
you, thereby, "to give back certain sorts of Chinese symbols with certain sorts of shapes in
response."
the first batch 'a script' [a data structure with natural language processing applications],
the second batch 'a story',
the third batch 'questions';
the symbols you give back the ‘answers to the questions’
the set of rules in English ‘the program‘
Can you be considered to understand Chinese ?
Approaches to Artificial Intelligence

Cognitive Scientists


Engineers


Want to build machines with human-like intelligence
Weak AI


want computers to do very smart things, quite independently of how
humans work
Strong AI


think AI is the only serious way of finding out how humans work
Want to build machines that exhibit intelligent like behaviour but
believe machines will always be intellectually inferior to humans
Computer as a metaphor for the mind has been the dominant
approach for the last 60 years
Weak Vs. Strong AI - Philosopher John
Searle

WEAK AI



like Cognitive Science above (I.e. about people)
uses machine representations and hypotheses to mimic human
mental function on a computer , but never ascribes those
mental properties to the machine.
STRONG AI


claim that machines programmed with the appropriate
behaviour, are having the same mental states as people would
who have the same behaviour
i.e. that machines can have MENTAL STATES.
What is Artificial Intelligence ?
Make machines behave as they do in the movies!
 About the emulation of human behaviour
 Make machines do things that would require intelligence if
done by humans
Boden, M.A. (1977). Artificial Intelligence and Natural Man. Basic
Books, New York.


Concerned with programming computers to perform tasks
presently done better by humans because they involve higher
mental processes such as perceptual learning, memory
organisation and judgemental reasoning
Minsky, M.L. and Papert, S.A. (1969). Perceptrons. MIT Press,
Cambridge, MA.
What is Artificial Intelligence ?

Agreement that it is concerned with two things


Studying human thought processes
Representing these processes via machines




Computers
Robots
Artificial Intelligence is behaviour by a machine which if
performed by a human would be considered intelligent
“Artificial Intelligence is the study of how to make
computers do things at which, at the moment, people are
better”
Elaine Rich, Artificial Intelligence, McGraw-Hill, 1983, p. 1
What are humans better at ?






Playing Games
Solving Puzzles
Common Sense Reasoning
Expert Reasoning
Understanding Language
Learning
AI
Psychology
Philosophy
AI
Anthropology
Linguistics
Computer Science
Neuro-Science
Intelligence



How does a human mind work ?
Can non-humans have minds ?
In terms of computing philosophy




Accept the idea that machines can do anything
Oppose this – machines incapable of sophisticated behaviour e.g.
love, creativity
OK in philosophy
How about in engineering/science terms ?
Intelligence


What does intelligence mean ?
Dictionary definition
1.
2.

Someone’s ability to understand and learn things
Intelligence is the ability to think and understand instead of
doing things by instinct or automatically
(Collins English Dictionary)
1st => possessed by humans Someone’s
2nd => some flexibility, does not specify someone
Intelligence – what is thinking ?

Thinking


=> have to have a brain



Activity of using your brain to consider a problem or create an
idea (Collins)
Organ that allows learning and understanding
Is it possible for machines to achieve this ?
Can machines think ?
Why build intelligent machines?





Cheaper to build and maintain
Offer new possibilities
Better solutions to problems
Software relatively cheap to develop
Software can be changed easily
Why is AI relevant to us ?

Ai is concerned with how





knowledge is acquired and used,
information is communicated,
collaboration is achieved,
how problems are solved,
languages are developed, etc.
History of AI (Classical Period or Dark Ages)
mid 1940’s – mid 1950’s





Game Playing & Theorem Proving
State Space Searching
Alan Turing
McCulloch & Pitts
Von Neumann
Turing




Proposed the concept of a universal machine
Mathematical Tool equivalent to Digital Computer
Takes input and computes output via a Finite State
Machine
Must construct a different machine for each computation
Turing



Enigma Machine
Wrote the first program capable of playing a complete
chess game;
Reflections on intelligence:





Is there thought without experience?
Is there mind without communication?
Is there language without living?
Is there intelligence without life?
i.e. can machines think?
Turing




Invented a game ‘Turning Imitation Game’
Can machines pass a behaviour test for intelligence
Defined the intelligent behaviour of a computer as the
ability to achieve the human-level performance in
cognitive tasks
Predicted that by 2000 a computer could be programmed
to have a conversation with a human interrogator for five
minutes and would have a 30 per cent chance of
deceiving the interrogator that it was a human
The Turing Test

Computer passes the test if interrogators cannot distinguish the machine from a
human on the basis of the answers to their questions.

Original Game:


First phase

Interrogator, a man and a woman are each placed in separate rooms and can communicate only via a
neutral medium such as a remote terminal.

Interrogator’s objective is to work out who is the man and who is the woman by questioning them.

Man should attempt to deceive the interrogator that he is the woman, while the woman has to
convince the interrogator that she is the woman.
Second phase

Man is replaced by a computer programmed to deceive the interrogator as the man did.

Programmed to make mistakes and provide fuzzy answers in the way a human would.

If the computer can fool the interrogator as often as the man did, we may say this computer has
passed the intelligent behaviour test.

Interrogator does not see, touch or hear the computer and is therefore not
influenced by its appearance or voice

Annually The Lobner Prize - http://www.loebner.net/Prizef/loebner-prize.html
McCulloch & Pitts


Proposed model of artificial neural networks in which
each neuron was postulated as being in binary state, that
is, in either on or off condition
Demonstrated that their neural network model was, in
fact, equivalent to the Turing machine, and proved that any
computable function could be computed by some
network of connected neurons
McCulloch & Pitts


Stimulated both theoretical and experimental work to
model the brain in the laboratory.
Experiments clearly demonstrated that the binary model
of neurons was not correct
Von Neumann


Part of the Manhattan Project
Adviser for the Electronic Numerical Integrator and
Calculator (ENIAC) project at the University of Pennsylvania


First general purpose computer
Helped to design the Electronic Discrete Variable Automatic
Computer (EDVAC), a stored program machine.

Binary rather than decimal
History of AI (Great Expectations)
(mid 50’s – late 60’s)

John McCarthy



Marvin Minsky





Continuing work on neural networks
Learning methods improved
Newell & Simon




Focus on formal logic
Developed anti-logic outlook on knowledge representation and reasoning
Frames
McCulloch & Pitts


Inventor of LISP
AdviceTaker – first complete knowledge-based system
General Problem Solver(GPS) – simulate human problem solving
Based on technique of means-end analysis
Choose and apply operators to achieve goal state
Focus on general problem solving, weak AI
History of AI (Great Expectations)
(mid 50’s – late 60’s)

Newell & Simon





Attempts to separate problem solving from data
Proposed that a problem to be solved could be defined in
terms of states.
Means-ends analysis was used to determine a difference
between the current state and the desirable state or the goal
state of the problem, and to choose and apply operators to
reach the goal state.
If the goal state could not be immediately reached from the
current state, a new state closer to the goal would be
established and the procedure repeated until the goal state was
reached.
The set of operators determined the solution plan.
History of AI (Great Expectations)
(mid 50’s – late 60’s)

Newell & Simon
 GPS failed to solve complicated problems.
 Program was based on formal logic and therefore could
generate an infinite number of possible operators, which
is inherently inefficient.
 The amount of computer time and memory that GPS
required to solve real-world problems led to the project
being abandoned.
History of AI (Reality Strikes)
(late 60’s – early 70’s)




AI researchers were developing general methods for
broad classes of problems
Programs contained little or no knowledge about
problem domain
Applied a search strategy by trying different combinations
of steps until right one found
Problems chosen too broad and too difficult
History of AI (Expert Systems)
(early 70’s – mid 80’s)


Realisation that problem domain must be restricted
Feigenbaum & Buchanan





DENDRAL program developed at Stanford to analyse chemicals
Incorporated knowledge of expert into program to perform at human
expert level
Shift from weak methods
Difficult – knowledge acquisition
Shortliffe


MYCIN – rule-based expert system for the diagnosis of infectious
diseases
Rules reflected uncertainty
History of AI (Making Machines Learn)
(mid 80’s - )


Expert Systems require more than rules
Rebirth of neural networks


Technology assisted
Evolutionary computing



Learning by doing
Ongoing since 70s
Natural intelligence is product of evolution



Based on computational models of natural selection and genetics
Simulate populate, evaluate performance, generate new population
Concept introduced by John Holland in 1975
History of AI (Making Machines Learn)
(1980’s onwards)






Knowledge Engineering
Computing with Words
Handling Uncertainty
Improved computational power
Improved cognitive modelling
The ability to represent multiple experts
Today





Topics in AI are much the same
Language now not so near the centre but it was at the
centre in the 70s
Roots now much further from logic and theorem proving
Neural nets and machine learning now more central
AI Approaches transitioned to main stream
What has AI achieved in real world ?







Robots in manufacturing
Diagnosis of illness: screen lab tests, diagnose blood
infections, identify tumors
Run airports: e.g. assign baggage gates, direct re-fuelling
Reasonable machine translation
Search systems like Google – efficient information
retrieval
Computer games
Deep Blue beat Kasparov in 1997
Key Lessons




Intelligence = ability to learn and understand, to solve
problems and to make decisions.
Goal of AI = making machines do things that would
require intelligence if done by humans.
A machine is thought intelligent if it can achieve humanlevel performance in some cognitive task.
To build an intelligent machine, we have to capture,
organise and use human expert knowledge in some
problem area.
Negnevitsky M 2005, Artificial Intelligence, A guide to
intelligent systems design, 2nd Edition, Addison Welsey
Why Representation?


Humans need words (or symbols) to communicate
efficiently
Mapping of words to things is only possible indirectly

Create concepts that refer to things
What is knowledge representation?

What is representation?


When do we need to represent?



Representation refers to a symbol or thing which represents
(’refers to’, ’stands for’) something else.
We need to represent a thing in the natural world when we
don’t have, for some reason, the possibility to use the original
’thing’.
Example: Planning ahead – how will our actions affect the
world, and how will we reach our goals?
The object of knowledge representation is to express the
problem in computer-understandable form
Aspects of KR

Syntactic




Semantic



Possible (allowed) constructions
Each individual representation is often called a sentence.
For example: color(my_car, red), my_car(red), red(my_car), etc.
What does the representation mean (maps the sentences to the
world)
For example:
color(my_car, red) → ??
‘my car is red’, ‘paint my car red’, etc.
Inferential



The interpreter
Decides what kind of conclusions can be drawn
For example: Modus ponens (P, P→Q, therefore Q)
Well-defined syntax/semantics


Knowledge representation languages should have precise
syntax and semantics.
You must know exactly what an expression means in
terms of objects in the real world.
Real World
Real World
Map to
KR language
Representation
of facts in the world
Map back to
real world
Inference
New
conclusions
Declarative vs. Procedural

Declarative knowledge (facts about the world)




A set of declarations or statements.
All facts stated in a knowledge base fall into this category of
knowledge.
In a sense, declarative knowledge tells us what a problem (or
problem domain) is all about
Procedural knowledge (how something is done)

Something that is not stated but which provides a mean of
dynamically (usually at run-time) arriving at new facts.
Declarative example

Information about items in a store
cheaper(coca_cola, pepsi)
tastier(coca_cola, pepsi)
if (cheaper(x,y) && (tastier(x,y) ) →
buy(x)
Procedural example

Shopping script:







Make a list of all items to buy
Walk to the shop
For each item on the list, get the item and add it to the
shopping basket
Walk to the checkout counter
Pack the items
Pay
Walk home
Types of knowledge

Domain knowledge:




What we reason about
Structural knowledge
 Organization of concepts
Relational knowledge
 How concepts relate
Strategic knowledge:


How we reason
At representation level, rather than at implementation level
 (e.g. at implementation level – control knowledge, for resolving
conflicting situations)
What is a Knowledge Representation?




“What is a Knowledge Representation?”
(Davis, Shrobe &Szolovits) AI Magazine, 14(1):17-33, 1993
http://groups.csail.mit.edu/medg/ftp/psz/k-rep.html
Defines the five roles the knowledge representation plays
Each role defines characteristics a KR should have
These roles provide a framework for comparison and
evaluating KRs
Role I: A KR is a Surrogate

A KR is used to model objects in the world.


Substitute for direct interaction with the world.
Cannot possibly represent everything in the world, a KR must
necessarily focus on certain objects and properties while
ignoring others.


As a result only objects and properties that are relevant to reasoning
are modeled.
Consequences:

Representation is not perfect


will have errors (at least by omission) and we may even introduce
new artifacts which not present
At least some unsound reasoning will occur
Role I: A KR is a Surrogate

The only complete accurate representation of an object is
the object itself.

All other representations are inaccurate.
Role II: A KR is a set of
Ontological Commitments



All representations are approximations to reality and they
are invariably imperfect.
Therefore we need to focus on only some parts of the
world, and ignore the others.
Ontological commitments determine what part of
the world we need to look at, and how to view it.
Role II: A KR is a set of
Ontological Commitments

The ontological commitments are accumulated in layers:

First layer – representation technologies.


For example, logic or semantic networks (entities and relations) vs.
frames (prototypes)
Second layer – how will we model the world.

Example from a frame-based system:



“The KB underlying INTERNIST system is composed of two basic types of
elements: disease entities and manifestations […] It also contains a hierarchy of
disease categories organised primarily around the concept of organ systems
having at the top level such categories as ’liver disease’, ’kidney disease’, etc”
Commits to model prototypical diseases which will be organised in a taxonomy
by organ failure
Third layer (conceptual) – which objects will be modelled.

What is considered a disease (abnormal state requiring cure), e.g.
alcoholism, chronic fatigue syndrome?
Role III : A KR is a Fragmentary
Theory of Intelligent Reasoning

“What is intelligent reasoning?”


The views of intelligence normally come from fields outside of
AI: mathematics, psychology, biology, statistics and economics.
Fragmentary


the representation typically incorporates only part of the
insight or belief that motivated it
that insight or belief is in turn only a part of the complex and
multi-faceted phenomenon of intelligent reasoning.
Role III : A KR is a Fragmentary
Theory of Intelligent Reasoning

There are three components:

the representation's fundamental conception of intelligent
inference
(What does it mean to reason intelligently?)

the set of inferences the representation sanctions
(What can we infer from what we know?)

the set of inferences it recommends
(What ought we to infer from what we know?)
Role IV: A KR is a medium for
efficient computation



The knowledge representation should make
recommended inferences efficient.
The information should be organized in such a way to
facilitate making those inferences.
There is usually a tradeoff between


the power of expression (how much can be expressed and
reasoned about in a language) and
how computationally efficient the language is.
Role V: A KR is a medium of
human expression

A representation is a language in which we communicate.





How well does the representation function as a medium of
expression?
How general is it?
How precise?
Does it provide expressive adequacy?
How well does it function as a medium of
communication?

How easy is it for us to ‘talk’ or think in that language?
Consequences of this KR

The spirit should be indulged, not overcome –


Representation and reasoning are intertwined


a recommended method of inference is needed to make sense
of a set of facts.
Some researchers claim equivalence between KRs, i.e.
“frames are just a new syntax for first-order logic”.


KRs should be used only in ways that they are intended to be
used, that is the source of their power.
However, such claims ignore the important ontological
commitments and computational properties of a
representation.
All five roles of a KR matter
Randall Davis, Howard Shrobe, Peter Szolovits MIT Lab
Requirements for KR languages

Representation adequacy


Inferential adequacy


inferences should be efficient
Clear syntax and semantics


should allow inferring new knowledge
Inferential efficiency


should to allow for representing all the required knowledge
unambiguous and well-defined syntax and semantics
Naturalness

easy to read and use
Some Knowledge Representation
Formalisms





Production systems, expert systems
Semantic networks
Frames
Case-based reasoning
Biologically inspired approaches


neural networks
genetic algorithms
Download