The Myth of Commonsense Knowledge

advertisement
The Myth of
Common Sense Knowledge
some exploratory ideas
Peter Clark
Boeing Phantom Works
The Myth of Common-Sense:
All you need is knowledge…
“We don’t believe that there’s any shortcut to being intelligent; the
“secret” is to have lots of knowledge.” Lenat & Guha ‘86
“Knowledge is the primary source of the intellectual power of
intelligent agents, both human and computer.” Feigenbaum ‘96
“Lenat and Feigenbaum suggest that just a million frames, massaged
by already-understood control structures, coupld intelligently manifest
the sum total of human knowledge.” Smith ‘91
The Myth of Common-Sense
• Common, implicit assumption (belief?) in AI:
– Knowledge is the key to intelligence
– Acquisition of knowledge is bottleneck
• Spawned from:
– ’70s experience with Expert Systems
• Feigenbaum’s “knowledge acquisition bottleneck”
– Introspection
• The assumption has fueled many programs…
A Rallying Cry….
“…What ontological categories would make up an
adequate set for carving up the universe? How are
they related? What are the important things most
humans today know about solid objects? And so on.
In short, we must bite the bullet.”
Lenat & Guha ‘86
Manifested in Multiple Programs:
Manual Knowledge-Base Construction
• Cyc
• Botany KB (Porter et al, 85-95)
• DARPA
– High Performance Knowledge Base (HPKB)
– Rapid Knowledge Formation (RKF)
• ThoughtTreasure
• Language-based knowledge resources
– WordNet
– Pangloss
– MicroKosmos
• And, of course, our own smallish endeavor here
Manifested in Multiple Programs:
Automated Knowledge-Base Construction
• Advances in NLP and statistical corpus-based
technologies have raised hopes of building KBs
automatically
– Far from abandoning the quest for knowledge,
researchers see new opportunities for acquiring it
– Recent boost from intelligence community
Manifested in Multiple Programs:
Automated Knowledge-Base Construction
• Dictionary Mining
– MindNet (Microsoft)
– Extended WordNet (HLI, Dallas)
• Corpus-based processing
– Schubert (Rochester); also us here!
– Girju (UT Dallas)
– Ontology learning technologies
– Hovy (ISI)
• convert the entire Web into a knowledge base
These techniques are rapidly overtaking the manual methods
(for certain types of knowledge)…
Manifested in Multiple Programs:
Community-Based Construction
• OpenMind
– community-based entry of knowledge
• The Semantic Web
– “knowledge processable by machines”
• semantic markup and annotation
– vision of a (very) large distributed KB
In fact…
• it’s not so difficult to build a large KB these days…
• …so where are the intelligent machines?
What Happens when you have lots of
knowledge?
(income GeorgeBush X)?
-> (residesIn GeorgeBush Shack)?
-> owns shack?
-> spouse resides in shack?
-> spouse?
-> person married on wedding day?
-> wedding day?
-> date on marriage certificate?
-> …
What Happens when you have lots of
knowledge?
Tweety is a penguin.
(residesInRegion Tweety X)?
-> is Tweety a citizen of city X?
-> is Tweety the head of state of X?
-> is Tweety the monarch of X?
-> …
What Happens when you have lots of
knowledge?
(maritalStatus BillClinton X)
-> age(BillClinton,A), A < 18? [-> single]
-> age of BillClinton?
-> does he have a twin?
-> is he a baby?
-> ….
What Happens when you have lots of
knowledge?
(volume USS-Enterprise X)
-> Is USS-Enterprise a human adult?
-> Isa Galaxy? Pebble? Boulder? Shampoo? Soup? …
“A man drives a car to the airport”…
What Happens when you have lots of
knowledge?
Far from expert systems “failing” due to their myopic
knowledge of the world, this is a primary reason for their
success.
Problems that the AI Community
have not taken Seriously Enough
•
•
•
•
Reasoning with “messy” knowledge
– Errors
– Approximations
– Different unstated assumptions
Facts may conflict in certain circumstances
Incompatible with the backward-chaining, deductive paradigm
Spawns into problems/topics of:
1. Elaboration
2. Coherence
3. Encoding possibilities
4. Relevance/viewpoints/Context
Reasoning with Messy Knowledge
• Not about deductive backward chaining…
• Rather is about creating coherent models
“Messy”
background
knowledge
Elaboration
Coherence
Possibilities
Relevance/
Context
?
?
Fragmented and
ambiguous input
Coherent Model
(situation-specific)
1. Elaboration
• Answering questions with messy knowledge
– not enough to backward chain
– need to create a coherent model
• need to elaborate possible components of the model
Age of
Bush?
Resides in
a shack?
President
Spouse resides
in a shack?
Resides in
WhiteHouse
Spouse?
1. Elaboration: LTI’s Illustration
“The kid hit the ball hard.”
1. Elaboration: Thought experiment
“Joe chopped down the tree.”
• A vivid picture comes to mind
– John: adult male, out in woods
– holding an axe (or chain saw?)
– Tree is ~30ft high pine tree
• or: a modification of that time I sawed a Christmas tree
• or: that documentary on logging in Canada
•
•
•
•
8:1 ratio of prior to explicit knowledge (Graesser, ’81)
Episodic/experiential knowledge plays a key role
Also core knowledge plays a key role
Not a deductive process!
2. Coherence: What Makes a
Representation (Model) Coherent?
• We don’t just blindly accept new knowledge:
– Minsky: We proactively ask a set of pertinent
questions about a scene, e.g., what is X for?
What are the goals? etc.
• What makes a representation coherent?
– Simple consistency (“The man fired the gnu.”)
– Purposefulness (for artifacts):
• “The engine contains a thrust reverser.”
• vs. “The engine contains an elephant.”
• vs. “The engine contains a book.”
• “Knowledge entry” is a serious misnomer!
– Really talking about Knowledge Integration
2. Coherence: A Catalog of Coherence Criterea
1. Volitional actions:
– Agents must be capable of an action
• legally, skill, fiscally, anatomically
– Action serves a broader purpose/goal
– Need equipment/resources/instruments,
instruments must be adequate
2. Non-volitional actions;
– There is a cause (inc. randomness)
– Spatial:
• statics: objects must be close
• dynamics: objects can move in the required way
– Temporal: objects exist at the same time
2. Coherence:
A Catalog of Coherence Criterea (cont)
3.
Objects:
– physically possible
•
•
•
parts connected together at appropriate places
materials are appropriate
suspension/tension etc., gravity
– physically normal/expected/standard
•
4.
need to know normal shapes, sizes, etc.
Artifacts:
– Purposefulness:
•
all parts play some role wrt. one of its intended functions (or
subtasks thereof). Expect design to be optimized.
– Could treat biological objects as “artifacts”
2. Coherence and Reasoning Engines
• Inference engines unable to tolerate incoherence:
– Current: “Error! Switching on the debugger…”
– Desired: “This representation is generally ok,
except this bit looks weird, and that bit conflicts
with this bit.”
• Problem compounded by long inference chains
• How could we change a reasoner to be more tolerant?
3. Representing Possibilities
• Represent things which “might” happen
– Not quite probabilities
– More, sanctioning certain facts
• Use this to guide search for coherence
People can walk
People can eat food
A person can be happy
A person can go to a shop
…
Semantics? Like a third truth value?
3. Automatically acquired possibilities…
Based on Schubert-style corpus analysis:
…
Banks may agree.
Banks may have borrowings.
Banks may seek approvals.
Banks may estimate shortfalls.
…
Bonds may be guaranteed.
Bonds may be listed.
Bonds may see gains.
…
Commissions may approve changes.
…
Companies may elect directors.
Companies may blame weather.
Companies may own shares.
…
Elections may be scheduled.
Elections can be annual.
Exports may rise.
…
Figures can be preliminary.
Figures may be revealed.
Figures may be stated.
Figures may exclude losses.
…
Shareholders may receive shares.
Shareholders may vote.
…
Shares may climb.
Shares may trade.
…
Spokesmen may decline comments.
…
Workers can be salaried.
…
4. Viewpoints and Context
• Far from expert systems “failing” due to their
myopic knowledge of the world, this is a primary
reason for their success.
• Can we model this?
– Contexts
– Selection of reasoning components/modules?
4. Viewpoints and Context
vs.
Reason over
Giant KB
Component
theories/
ontologies(?)
Problem-specific KB,
contains selected units
• Latter seems right, but:
– can a big KB really be partitioned like this?
(everything is connected!)
• Models may vary by:
– ignoring detail
– making different approximations
– using different ontologies
4. Viewpoints and Context
• e.g., DNA = sequence of different region types:
– intron-exon-intron-exon…
– promotor-gene-terminator
– nucleotide pair-nuc pair-nuc pair…
• Makes a difference:
– Given: “The polymerase attaches to the promotor,
and then moves down the strand.”
– then answer: “Where will the polymerase be?”
• nucleotide? gene? intron?
• The point: It’s not simply a matter of having all
viewpoints coexisting
• Another example: “A satellite sends
signals/messages/position information.”
Summary
• The myth of common-sense:
– Knowledge is the key to intelligence
• The truth:
– Knowledge is only part of the equation
• and less and less the most significant part
– Rather, radical revisions to reasoning are needed
• Messy knowledge is here to stay
• Old reasoning paradigms do not work effectively with this
• Need some new approaches, touching issues of elaboration,
coherence, possibilities, viewpoints
Download