Concurrent Reasoning with Inference Graphs Problem Summary Inference Graphs

advertisement
Concurrent Reasoning with Inference Graphs
Daniel R. Schlegel
Stuart C. Shapiro
Department of Computer Science and Engineering
Problem Summary
• Rise of multi-core computers, BUT:
• Lack of concurrent natural deduction systems.
Inference Capabilities
• Forward, backward, bi-directional, and focused
inference.
• Retains all derived formulas for later re-use.
• Propagates disbelief.
Only concurrent inference system with these
capabilities.
Propositional Graphs
• Directed acyclic graph
• Every well-formed expression is a node
• Individual constants
• Functional terms
• Atomic formulas
• Non-atomic formulas (“rules”)
• Each node has an identifier, either
• Symbol, or
• wfti[!]
• No two nodes with same identifier.
Example:
Propositional graph for the assertions that if a, b,
and c are true, then d is true, and if d or e are true,
then f is true.
Inference Graphs
Concurrency and Scheduling
• Extend Propositional Graphs
Example:
• Adds channels for information flow:
• i-channels report truth of an
antecedent to a rule node.
• u-channels report truth of
a consequent from a rule node.
Channels represented by dashed lines
• Channels contain valves.
are i-channels and are drawn from
• Hold messages back, or allow
antecedents to rule nodes. Channels
them through.
represented by dotted lines are u channels and are drawn from rule nodes
• Channels relay messages
to consequents.
• I-INFER (“I’ve been inferred”)
• U-INFER (“You’ve been inferred”)
• BACKWARD-INFER (“Open valves so messages that might infer me can arrive”)
• CANCEL-INFER (“Stop inferring me (close valves)”)
• UNASSERT (“I’m no longer believed”)
• Different message types have different relative priorities (important for scheduling).
The area between two valves is called an
inference segment.
Rule Node Inference
1.
2.
3.
4.
5.
Message arrives at node.
Message translated to a RUI, containing
positive and negative instances of antecedents
contained in the message.
New RUI combined with existing ones.
Output is a set of new RUIs which are used to
decide of the rule can fire.
When a rule fires, new messages are sent out.
Example: We assume backward inference has been
initiated, opening all the valves in the graph. First, in (a),
messages about the truth of a, b, and c flow through ichannels to wft1. Since wft1 is and-entailment, each of
its antecedents must be true for it to fire. Since they are,
in (b) the message that d is true flows through wft1’s uchannel. d becomes asserted and reports its new status
through its i-channel (c). In (d), wft2 receives this
information, and since it is an or-entailment rule and
requires only a single antecedent to be true for it to fire,
it reports to its consequents that they are now true, and
cancels inference in e. Finally, in (e), f is asserted, and
inference is complete.
When a message passes through a valve:
• A task is created with the same priority as the
message, and is the application of the
inference segment’s function to the
message.
• The task is added to a queue which puts
higher priority tasks towards its head.
A task only operates within a task segment.
1. tasks for relaying newly derived
information using segments to the right are
executed before those to the left, and
2. once a node is known to be true or false,
all tasks attempting to derive it (left of it in
the graph) are canceled, as long as their
results are not needed elsewhere.
There is minimal shared state between tasks,
allowing many tasks to operate concurrently.
Evaluation
Concurrency:
• Near linear performance improvement with the
number of processors
• Performance resilient to graph depth and
branching factor changes.
Scheduling Heuristics:
• Backward-inference with or-entailment shows
10x improvement over LIFO queues, and 20-40x
over FIFO queues.
See GKR paper (below) for more details.
References
• Daniel R. Schlegel and Stuart C. Shapiro, Concurrent Reasoning with
Inference Graphs. In Proceedings of the Third International IJCAI Workshop
on Graph Structures for Knowledge Representation and Reasoning (GKR
2013), 2013, in press.
This work has been supported by a Multidisciplinary University Research Initiative (MURI) grant (Number W911NF-09- 1-0392) for Unified Research on Network-based Hard/Soft Information Fusion, issued by the US Army Research Office (ARO) under the program management of Dr. John Lavery.
Download