Introduction to the course, History of AI - clic

advertisement
INTRODUCTION TO ARTIFICIAL
INTELLIGENCE
Massimo Poesio
LECTURE 1: Intro to the course, History
of AI
ARTIFICIAL INTELLIGENCE:
A DEFINITION
• The branch of Computer Science and/or
Cognitive Science whose aim is
– ORIGINALLY: to develop machines able to display
intelligent behavior
– NOW: to study and design INTELLIGENT AGENTS
using computational methods
• Strong AI: developing machines all but
undistinguishable from human beings
• Weak AI: developing SIMULATIONS of human
intelligence
A BRIEF HISTORY OF AI
• Forerunners, I: logic and ontologies
• Forerunners, II: mechanical machines / robots
• The beginning of AI: Turing, Dartmouth,
Games, Search
• The role of Knowledge in Human Intelligence
• The role of Learning
• Modern AI
ARISTOTLE
• Aristotle developed the first theory of
knowledge and reasoning – his ideas
eventually developed into modern
– LOGIC
– ONTOLOGIES
ARISTOTLE: SYLLOGISM
(Prior Analytics)
The first attempt to develop a precise method for
reasoning about knowledge: identify VALID REASONING
PATTERNS, or SYLLOGISMS
BARBARA:
A: All animals are mortal
A: All men are animals.
A: Therefore, all men are mortal.
DARII:
A: All students in Fil., Logica & Informatica take Intro to AI
I: Some students in Filosofia take Fil., Logica & Informatica.
I: Therefore, some student in Filosofia takes Intro to AI.
FIRST ONTOLOGIES
LOGIC: BEYOND ARISTOTLE
• Ramon Lull (13th Century): first mechanical
devices for automatic reasoning (Lull’s disks)
• Leibniz (17th Century): Encoding for syllogisms
• Boole (19th Century): Boolean Algebra
• Frege (1879): Predicate calculus
BOOLEAN ALGEBRA
FORERUNNERS, 2
THE BEGINNINGS OF AI: 1930-1955
• The development of the first electronic
computing machines in the ‘30s also initiated
a lot of theoretical thinking about what these
machines could do
TURING
• Alan Turing is the father of both theoretical
computer science and of AI
• He was the first to imagine machines able of
intelligent behavior
• Devised an intelligence test, the TURING TEST:
replace the question “Can a machine be
endowed with intelligence” with the question:
– Can a machine display such human-like behavior
to convince a human observer that it is a human
being?
THE TURING TEST
• Alan M. Turing, “Computing Machinery and
Intelligence”, 1950 (in Boden, Philosophy of
AI)
• The question: what does it mean for a
machine to be intelligent?
Turing Test
Objections to the Turing Test
• Objections are basically of two forms:
– “No computer will ever be able to pass this test”
– “Even if a computer passed this test, it wouldn’t
be intelligent”
• Chinese Room argument (Searle, 1980),
responses, and counterresponses
– Robot reply
– Systems reply
20
“Machines can’t think”
• “It’s simply not possible, that’s all”
• Arguments from incompleteness theorems
– But people aren’t complete, are they?
• Machines can’t be conscious or feel emotions
– Reductionism doesn’t really answer the question: why
can’t machines be conscious or feel emotions??
• Machines don’t have Human Quality X
• Machines just do what we tell them to do
– Maybe people just do what their neurons tell them to do…
• Machines are digital; people are analog
21
DARTMOUTH
• In 1956 a group of researchers including J.
McCarthy, M. Minsky, C. Shannon, N.
Rochester organized a workshop at Dartmouth
to study the possibility of developing machine
intelligence
THE BEGINNINGS OF AI (1956-1966)
• Early AI researchers identified intelligence
with the kind of behavior that would be
considered intelligent when displayed by a
human, and tried to develop programs that
reproduced that behavior
• Examples:
– Games: checkers (Samuel), chess
– Theorem proving: Newell Shaw and Simon
HEURISTIC SEARCH
• This early research
focused on the
development of SEARCH
ALGORITHMS (A*) that
would allow computers to
explore a huge number of
alternatives very
efficiently
THE SUCCESS OF EARLY AI
In 1997, the chess-playing program DEEP BLUE developed by IBM
researchers led by Feng-hsiung Hsu, beat the chess world
champion Gary Kasparov over six games
THE LOGIC THEORIST
• Developed by Newell, Shaw and Simon in
1956
• Managed to prove 38 of the first 52 theorems
in the second chapter of Russell and
Whitehead’s Principia Mathematica
• (More on theorem proving later)
EARLY AI RUNS INTO TROUBLE (19661973)
• Soon however researchers realized that these
methods could not be applied to all problems
requiring intelligence, and that there were a
number of ‘simple’ problems that could not be
handled with these methods at all
– Example of the first: machine translation (the
ALPAC report)
– Example of the second: natural language, vision
COMMONSENSE KNOWLEDGE IN
LANGUAGE UNDERSTANDING
• Winograd (1974):
– The city council refused the women a
permit because they feared violence.
– The city council refused the women a
permit because they advocated violence
AI KEY DISCOVERIES, 1
• Performing even apparently simple tasks like
understanding natural language requires lots
of knowledge and reasoning
THE ‘KNOWLEDGE YEARS’ (1969-1985)
• Development of knowledge representation
techniques
• Development of EXPERT SYSTEMS
• Development of knowledge-based techniques
for
– Natural Language Understanding
– Vision
KNOWLEDGE REPRESENTATION
METHODS
• Logic is the older formalization of reasoning
• It was natural to think of logic as providing the
tools to develop theories of knowledge and its
use in natural language comprehension and
other tasks
• Great success in developing theorem provers
• But AI researchers quickly realized that the
form of logic required was not valid deduction
FROM LOGIC TO AUTOMATED REASONING
• Starting from the ‘50s AI researchers
developed techniques for automatic theorem
proving
• These techniques are still being developed
and have been used to prove non-trivial
theorems
RESOLUTION THEOREM PROVING
All Cretans are islanders.
All islanders are liars.
Therefore all Cretans are liars.
∀X C(X) implies I(X)
∀X I(X) implies L(X)
Therefore, ∀X C(X) implies L(X)
¬C(X) ∨ I(X)
¬I(Y) ∨ L(Y)
¬C(X) ∨ L(Y)
HIGH PERFORMANCE THEOREM
PROVING
• There are now a number of very efficient
theorem provers that can be used to
demonstrate sophisticated mathematical
theorems
– Otter
– Donner & Blitzen
THE FOUR-COLOR PROBLEM
• Conjecture: given a plane
divided in regions, it is
possible to color the regiones
in such a way that two
adjacent regions are always of
different colors using no more
than 4 colors
• This conjecture was
demonstrated by an automatic
theorem prover in 1997
AI KEY DISCOVERIES, 2
• Neither commonsense nor ‘expert’ reasoning
involve only valid inferences from certain
premisses:
– Commonsense reasoning often involves jumping
to plausible conclusions
– Expert reasoning involves making decisions with
uncertainty
COMMONSENSE KNOWLEDGE IN
LANGUAGE UNDERSTANDING
• Winograd (1974):
– The city council refused the women a
permit because they feared violence.
– The city council refused the women a
permit because they advocated violence
DIAGNOSTIC REASONING IN EXPERT
SYSTEMS
IF patient has loss of weight AND
anorexia AND
spleen is palpable AND
fever
THEN
patient MAY have Hodgkin’s disease
BEYOND VALID INFERENCE
• As a result of discovery number 2, AI
researchers quickly got involved with
developing better theories of commonsense
reasoning and uncertain reasoning
– Drawing inspiration from psychology (work on
semantic networks, scripts)
– Developing new inference formalisms
SOME RESULTS FROM COGNITIVE PSYCHOLOGY:
SEMANTIC NETWORKS
• Collins & Quillian, 1969: knowledge appears to be
organized around objects and in a taxonomic way
– A canary is yellow
– A canary has feathers
– A canary eats food
• Haviland & Clark 1974, Sanford & Garrod 1979:
‘associated’ knowledge available when concepts are
mentioned
– I looked around the house.
– The lounge was very spacious.
AN EXAMPLE OF SEMANTIC NETWORK
COMMONSENSE REASONING: SCRIPTS
(SCHANK & ABELSON, 1977)
John entered the restaurant.
The menu had changed, so he ordered one
of the new dishes.
The bill was very steep.
AN EXAMPLE OF SCRIPT
‘KNOWLEDGE-BASED’ SYSTEMS FOR
LANGUAGE UNDERSTANDING: SHRDLU
Winograd, 1974
A DIALOGUE WITH SHRDLU
Person: PICK UP A BIG RED BLOCK.
Computer: OK.
Person: GRASP THE PYRAMID.
Computer: I DON'T UNDERSTAND WHICH PYRAMID YOU MEAN.
Person (changing their mind): FIND A BLOCK WHICH IS TALLER THAN THE ONE YOU
ARE HOLDING AND PUT IT INTO THE BOX.
Computer: BY "IT", I ASSUME YOU MEAN THE BLOCK WHICH IS TALLER THAN THE
ONE I AM HOLDING.
Computer: OK.
Person: WHAT DOES THE BOX CONTAIN?
Computer: THE BLUE PYRAMID AND THE BLUE BLOCK.
Person: WHAT IS THE PYRAMID SUPPORTED BY?
Computer: THE BOX.
Person: HOW MANY BLOCKS ARE NOT IN THE BOX?
Computer: FOUR OF THEM.
Person: IS AT LEAST ONE OF THEM NARROWER THAN THE ONE WHICH I TOLD YOU
TO PICK UP?
Computer: YES, THE RED CUBE.
EXPERT SYSTEMS
• Beginning with the ’70s, AI researchers
developed systems that could be used to
diagnose
– Illnesses (MYCIN, DXPLAIN)
– Mechanical problems
• DELTA-CATS1, General Electric, Bonissone et al 1984
• IDEA, Centro Ricerche Fiat, ~1993
– etc
EXPERT SYSTEMS
A COMMONSENSE ENCYCLOPEDIA:
CYC
• A project initiated in 1984 by Doug Lenat. The goal: to
encode all of commonsense knowledge
• Changed the underlying formalism several times.
– These days: a logic-based representation
• Two versions available:
– OpenCyc (http://www.opencyc.org/)
• 50 000 concepts, 300 000 facts
• Can be downloaded / on the Web
– ResearchCyc (http://research.cyc.com/)
• 300 000 concepts, 3 million facts
KNOWLEDGE IN CYC
"Bill Clinton belongs to the class of US Presidents“
(#$isa #$BillClinton #$UnitedStatesPresident)
“All trees are plants”
(#$genls #$Tree-ThePlant #$Plant)
“Paris is the capital of France".
(#$capitalCity #$France #$Paris)
COMMONSENSE REASONING
• Modelling commonsense inference required
the development of entirely new paradigms
for inference beyond classical logic
• Non-monotonic reasoning
• Probabilistic models
AI RUNS INTO TROUBLE, AGAIN
• The CYC project started in 1984, and by common
opinion is nowhere near finished
– Hand-coding of commonsense FACTS is unfeasible
– (We will get back to this point later when talking
about socially constructed knowledge)
• Work on lower-level tasks such as speech
perception revealed the impossibility of handcoding commonsense RULES and assigning them
priorities
SPEECH
KEY AI DISCOVERIES, 3
• A theory of intelligence requires a theory of
how commonsense knowledge and cognitive
abilities are LEARNED
THE MACHINE LEARNING YEARS
(1985-PRESENT)
• The development of methods for learning
from evidence started even before Dartmouth
• But machine learning has now taken center
stage in AI
CYBERNETICS
• McCulloch, Pitts (1943): first artificial neurons
model (based on studies of real neurons)
KNOWLEDGE REPRESENTATION IN THE BRAIN
MODELS OF LEARNING BASED ON THE
BRAIN: THE PERCEPTRON
LEARNING TO CLASSIFY OBJECTS
ARTIFICIAL INTELLIGENCE TODAY
• Artificial intelligence as a science:
– Computer science vs. Cognitive Science
• Artificial Intelligence as a technology
COGNITION VS ENGINEERING
• Cognitive AI: use computer simulations to
study human cognitive abilities
– E.g., many neural net models of knowledge
• Engineering AI: develop machines able to
carry out intelligent tasks (possibly in a nonhuman way)
– E.g., Deep Blue
STRONG AI / WEAK AI
• Weak AI (Turing): developing machines that
are EXTERNALLY UNDISTINGUISHABLE from
human beings
– But whose internal behavior may be rather
different
• Strong AI: developing machines
undistinguishable from human beings IN ALL
RESPECTS
Three Schools of Artificial
Intelligence
•
symbolic functionalism
– intelligence represented in symbols and mutual
manipulations
– formal models of reasoning, knowledge based
•
connectionism
– inspired by natural processes
– emergency of intelligent behaviour
•
robotics functionalism
– computational implementation of behaviourism
MAIN SUBFIELDS
•
•
•
•
•
•
•
Reasoning and knowledge representation
Planning
Learning
Perception
Motion
Natural language processing
Social intelligence
MAIN SUBFIELDS
•
•
•
•
•
•
•
Reasoning and knowledge representation
Planning
Learning
Perception and Vision
Motion
Natural language processing
Social intelligence
AI INDUSTRY: GOOGLE
AI INDUSTRY: ROBOTICS
CONTENTS OF THE COURSE
• A very brief recap of knowledge
representation themes: logic, ontologies, etc
– (Assumption: you all followed Language Resources
and Ontologies)
• Machine Learning
– A reminder about statistics
– Supervised learning
– Unsupervised learning
– Applications
• Vision
PRACTICAL INFORMATION
• 42 hours / 6 credits
• Evaluation
– A PROJECT to be presented at the exam
• Web site:
http://clic.cimec.unitn.it/massimo/Teach/AI
READING MATERIAL
• Required:
– The course slides, available from the Web Site
– Other material downloadable from the Website
• Recommended readings:
– History of AI:
•
•
J Copeland, Artificial Intelligence: A Philosophical Perspective, Blackwell’s
M Boden (ed), The Philosophy of AI, Oxford
– Russell & Norvig, Artificial Intelligence: A Modern Approach (2nd ed), Prentice-Hall
– Bianchini, Gliozzo, Matteuzzi, Instrumentum vocale: intelligenza artificiale e linguaggio,
Bononia
• Supplementary on specific sub-areas of AI:
– KR:
•
John F. Sowa, Knowledge Representation, Brooks / Cole
– ML:
•
•
Mitchell – Machine Learning – Prentice-Hall
Witten, Frank, Hall – Data Mining – Morgan Kaufmann (3rd edition)
– NL:
•
•
Blackburn & Bos, Representation and Inference for Natural language, CSLI
Jurafsky & Martin, Speech and Language Processing, Pearson
READINGS
• This lecture:
– http://en.wikipedia.org/wiki/Timeline_of_artificial_intelligence
– John F. Sowa, Knowledge Representation, Brooks / Cole, ch. 1
Download