Automated Briefing Production for Lessons Learned Systems

From: AAAI Technical Report WS-00-03. Compilation copyright © 2000, AAAI (www.aaai.org). All rights reserved.
AutomatedBriefing Production for Lessons LearnedSystems
Inderjeet
Mani and Kristian
Concepcion and Linda Van Guilder
The MITRECorporation, W640
11493Sunset Hills Road
Reston, Virginia 22090
{imani,kjc9,lcvg} @mitre.org
Abstract
Document
productionis an importantfunctionin many
organizations.In addition to instruction manuals,
courseware,
reports,systemdocumentation,
etc., briefings are a verycommon
type of document
product,often usedin slide formas a visual accompaniment
to a
talk. Thispaperdescribesa domain-independent
systemfor automatic
briefinggenerationfroma high-level
outline providedbythe user. Wediscusshowsomeof
the methodologies
usedin this systemmaybe leveraged
in intelligentlessonslearnedsystems.
Introduction
Document
productionis an important function in manyorganizations.In additionto instruction manuals,courseware,
reports, systemdocumentation,etc., briefings are a very
common
type of documentproduct, often used in slide form
as a visual accompaniment
to a talk. Since so muchtime is
spent by so manypeoplein producingbriefings, often under
serious time constraints, any methodto reduce the amount
of time spent on briefing productioncould yield great gains
in productivity. In this paper, weprovidean overviewof a
briefing generationsystem,followedby a brief discussionas
to howsomeof the methodologiesused in this systemmay
be leveragedin intelligent lessonslearnedsystems.
Many
briefings havea stereotypicalstructure, dictated in
part by the business rules of the organization. For example, a commander
maypresent a daily or weeklybrief to
her superiors, whichis morein the nature of a routine update of activities since the last briefing; or she mayprovide
an actionbrief, whichis triggeredby a particular situation,
and whichconsists of a situation update followedby argumentsrecommending
a particular courseof action. Further,
the process of constructing a briefing mayinvolve certain
stereotypicalactivities, includingculling informationfrom
particular sources,such as messages,news,webpage~s,previousbriefings, etc. Thus,whilepart of the briefingcontent
maybe created anewby the briefing author, other parts of
the briefing maybe constructed fromexisting information
sources. However,information in those sources need not
necessarilybe in the sameformas neededby the briefing.
Copyright©2000,American
Associationfor Artificial Intelligence(www.aaai.org).
All rights reserved.
43
Our workforms part of a larger DARPA-funded
project
(Project Genoa)aimedat improvinganalysis anddecisionmakingin crisis situations by providingtools tha t allow
analysts to collaborate to developstructured argumentsin
supportof particular conclusionsandto help predict likely
future scenarios. Thesetools leverage a corporate memory
that acts as a repository of knowledge
gained as a result
of analytical tasks. Thearguments,along with background
evidence,are packagedtogether as briefings to high-level
decision-makers.
Ourapproachneeds to allow the analyst to take on as
muchof the briefing authoringas she wantsto (e.g., it may
take time for her to adaptto or trust the machine,or she may
wantthe machineto present just part of the briefing). The
analyst’sorganizationusuallywill instantiate oneof several
templatesdictatingthe high-levelstructureof a briefing;for
example,a briefing mayalwayshaveto begin with an executive summary.
Themethodsused also needto be relatively
domain-independent,
since the subject matter of crises are
somewhat
unpredictable.
Giventhese task requirements, wehave adoptedan approachthat is flexible about accommodating
different degrees of authorinvolvement,
that is relatively neutral about
the rhetorical theoryunderlyingthe briefingstructure (since
a templatemaybe providedby others), and that is domainindependent.In our approach,the author creates the briefing outline, whichis then fleshedout further by the system
basedon informationin the outline. Thesystemfills out
somecontentby invokinginformationreductiontools called
summarization
filters; it also makesdecisions, whenneeded,
about outputmediatype; it introducesnarrative elementsto
improve
the coherenceof the briefing; andfinally, it assembles the final presentation, makingdecisionsaboutspatial
layoutin the process.
Abriefing is representedas a tree. Thestructure of the
tree representsthe rhetorical structure of the briefing. Each
nodehas a label, whichoffers a brief textual descriptionof
the node. Eachleaf node has an associated goal, which,
whenrealized, provides content for that node. There are
twokinds of goals: content-levelgoals and narrative-level
goals. Content-levelgoals are also of twokinds: retrieve
goals, whichretrieve existing mediaobjects of a particular
type (text, audio, image,audio, video)fromcorporatememory satisfying somedescription,andcreate goals, whichere-
ate newmediaobjects of these types using programs.The
programswehave used for create goals are summarization
filters, whichtake input informationandturn it into some
moreabstract anduseful representationtailored to the needs
of the end-user,filtering out unimportant
information.
Narrative-levelgoals introducedeseriptionsof contentat
other nodes:they include captions andrunningtext for media objects, andsegues,whichare rhetorical movesdescribing a transition to a node.Orderingrelations reflecting temporal andspatial layoutare definedon nodesin the tree. Two
coarse-grainedrelations, seq for precedence,andpar for simultaneity, are used to specify a temporalorderingon the
nodes in the tree. Thetree representation, along with the
temporalconstraints, can be rendered in text as XML;we
refer to the XML
representationas a script.
Theoverall architecture of our systemis nowdescribed.
Theuser creates the briefingoutline in the formof a script,
by using a GUI.Thebriefing generator takes the script as
input. TheScript Validator applies an XML
parser to the
script, to cheekfor syntacticcorrectness.It thenbuildsa tree
representationfor the script, whichrepresents the briefing
outline, with temporalconstraints attachedto the leaves of
the tree.
Next, a ContentCreatortakes the input tree andexpands
it by introducingnarrative-levelgoal coverseguesto nodes,
and runningtext and captions describing mediaobjects at
content nodes. Seguesare generatedbased on nodelabels,
distance betweennodes, and the height in the tree. Running
text andshort captions are generatedfrommeta-information
associatedwith mediaobjects. For example,in the case of a
create goal, a summarization
filter maytake in a collection
of documents
and producea graphof sometrend detected in
the collection, say, an associationbetweenpeopleandorganizationsover time. Or else, it maygeneratea biographyof
a person fromthe collection. Themeta-informationin the
structuredrepresentationof these outputsis usedto generate
the runningtext and captions, by using shallowtext generation methods(cannedtext). Wherethe mediaobjects are
foundby retrieve goals, the systemrelies on existing metainformation,
but is designed
not to fail if thereisn’t any- i.e.,
it will fail to providerunningtext.
The end result of content selection (which has an XML
representation called a groundscrip0 is that the complete
tree has beenfully specified, with all the create andretrieve goals fully specified, with all the outputmediatypes
decided. Whensummarizationfilters are used (for create
goals), the mediatypeof the outputis specifiedas a parameter to the filter. This mediatype maybe convertedto some
other type by the system(e.g., text to speechconversionusing Festival(Tayloret al. 98). Bydefault, all narrativenodes
attemptto realize their goals as a speechmediatype, using
rules basedon text length andtruncatability to decidewhen
to use text-to-speech.Captionsare alwaysrealized, in addition, as text (i.e., theyhavea text realizationanda possible
audiorealization).
Then,a ContentExecutorexecutesall the create and retrieve goals. Thisis a verysimplestep, resulting in the generation of all the mediaobjects in the presentation, except
for the audiofiles for speechto be synthesized.Thus,this
44
step results in realizationof the contentat the leavesof the
tree.
Finally, the PresentationGeneratortakes the tree which
is output fromContentExecution,along with its temporal
orderingconstraints, andgeneratesthe spatial layout of the
presentation.If no spatial layout constraints are specified
(the default is to not specify these), the systemallocates
space using a simple methodbased on the temporallayout
for nodes whichhave spatial manifestations. Speechsynthesis is also carded out here. Coherenceof a presented
segmentis influenced mainlyby the temporalconstraints
(whichhavebeenfleshed out by the ContentCreatorto include narrative nodes). Oncethe tree is augmentedwith
spatial layoutconstraints,it is translatedby the Presentation
Generator into SMIL1(SynchronizedMultimediaIntegration Language)(SMIL
99), a W3C-developed
extension
HTML
that can be played by standard multimediaplayers
(suchas (REAL
Player 99) and (Grins Player 99). This
thus presentsthe realizedcontent,synthesizingit into a multimediapresentationlaid out spatially andtemporally.
Relevance to Intelligent
Lessons Learned
Systems
Althoughthe briefing generation systemwasnot designed
specifically for LessonsLearned, the tools we havedeveloped use methodologiesthat are relevant to Lessons
Learned.
The Genoabriefing moduleis aimedat packagingargumentsin favor of a particular analysis, whichis presented
alongwith supportingevidenceto decision makers.In such
cases, the briefing outline can specify argumentsand their
supportingevidence(whichis filled in and reinforced by
create andretrieve goals, both of whichdrawandbuild from
data in corporate memory).Representationssuch as (Toulmin58) structures havebeen used in a variety of AI domains to conveyarguments(Clark 91) Oanssenand Sage
96). Toulminstructures identify the claim (the conclusion
reached),its grounds(facts on whichthe argument
is based),
its warrant(the inferencerules appliedto arrive at the conclusion), its qualifier (degreeof certainty), as wellas
backing(justification) anda possible rebuttal (exceptions
that invalidatethe claim).
Thesekinds of explanationscan be very useful in lessons
learnedas well. For example,the lesson learned, the experiencewhichgaverise to it, the impactit has on task performance,andthe specific processthat it correctsor reinforces
can be presentedin terms of arguments.Thebriefing generation modulecan help in presentingthe lessons learned, by
culling informationrelated to the argumentsandtheir supporting evidencefromweband Othersources, whichcan be
fleshedout bythe retrieval andcreationgoals. Thus,canonical presentationsof explanationsrelated to lessonslearned
can be carried out. Further, wherethe lessons learned are
fairly complex,summaries
of these lessonscan be presented,
allowinga drill-downcapability as needed.
Future Directions
There is a fair amountof work on automatic authoring of
multimediapresentations, e.g., (Wahlster et al. 92), (Dalai
et al. 96), (Mittal et al. 95), (Andreand Rist 97), generation
of documents from specification knowledge bases (Power
and Scott, 98), and explanatory caption generation (Mittal
et al. 95). Theseefforts differ fromours in two ways: first,
unlike us, they are not open-domain;and, second, they don’t
use summarization filters. The open-domainrequirement of
Project Genoahas required the design of a generic method.
However,in manydomains, especially for Lessons Learned
systems, domain knowledge bases may be available which
can be exploited by our approach to allow for more reasoning about the content of the presentation. This can be
viewed in part as an opportunity to leverage muchricher
meta-info specifications than the ones we currently use, ineluding representations of evidence and counter-evidence
for an argumentor point-of-view. Further, more effective
tailoring of presentations to individual users or students may
¯ be exploited, based on intensive domainmodeling(Daniel et
al. 99). Of course, we will need to carry out an evaluation
to assess the extent to which the automationdescribed here
provides efficiency gains in documentproduction.
A second direction for future work is the output format.
While multimedia players are popular, most presentations
we have comeacross are in Microsoft Powerpoint. In addition, if the user wantsto edit the output, Powerpointediting
is preferred to having to use a SMIL(i.e., HTML)
editor.
Weare therefore investigating the option of producing Powerpoint presentations.
Finally, we have begun the de.sign of the Script Creator
GUI(the only componentin Figure 1 remaining to be built).
This will allow the author to create ~ripts for the briefing
generator (instead of editing templates by hand), by laying
out icons for media objects in temporal order (seq of pars).
A user will be able to select a "standard" briefing template
from a menu,and then view it in a briefing/template structure editor. The user can then prOvide content by adding
annotations to any node in the briefing template. The user
has a choice of saving the edit version in template forma, or
in a particular output format e.g., (Powerpoint, SMIL).
References
Andre, E. and Rist, T. 1997. Towards a NewGeneration
of Hypermedia Systems: Extending Automated Presentation Design for Hypermedia.L. Dybkjaer, ed., Proce.e~lings of the Third Spoken Dialogue and Discourse Workshop, Topics in Natural Interactive Systems1. The Maersk
Me-KinneyMoiler Institute for Production Technology,
Odense University, Denmark,10-27.
Clark, P. 1991. A Modelof Argumentationand Its Application in a Cooperative Expert System. Ph.D. Thesis, Turing
Institute, University of Strathclyde, 1991.
Dalai, M., Feiner, S., McKeown,K., Pan, S., Zhou, M.,
I-Iollerer, T., Shaw,L, Feng, Y., and Fromer,J. 1996. Negotiation for AutomatedGeneration of TemporalMultimediaPresentations. Proceedings of ACM
Multimedia ’96.
45
Daniel, B. H., Bares, W.H., Callaway,C. B., I.e, ster, J.
C. 1999. Student-Sensitive MultimodalExplanation Generation for 3D Learning Environments. Proceedings of
AAAI/IAAI1999, 114-120.
Grins Player. http://www.oratrix.com
Janssen, T., and Sage, A. P. 1996. GroupDecision Support
Using ToulminArgumentStructure.s, II~.RR International
Conference on Systems, Man, and Cybernetics, Beijing,
China, 2704-2709.
Mittal, V., Roth, S., Moore,L, Mattis, L, and Carenini,
G. 1995. Generating Explanatory Captions for Information
Graphics. Proceedings of the International Joint Conference on Artificial Intelligence (IJCAr95), 1276-1283.
Power, R., and Scott, D. 1998. Multilingual Authoring using Feedback Texts. Proceedings of COLING’98,Monweal, 1053-1059, August 10-14.
Real Player. http://www.real.eom
SMIL. http:llwww.w3.orglAudio~deol
Taylor, P., Black, A., and Caley, R. 1998. The architecture
of the Festival Speech Synthesis System. Proceedings of
the Third ESCAWorkshopon Speech Synthesis, Jenolan
Caves, Australia, 147-151.
Toulmin,S. 1958. The Uses of Argument.University Press,
Cambridge.
Wahlster, W., Andre, E., Finkler, W., Profitlieh, I-I-L, and
Rist, T. 1993. Plan-BasedIntegration of Natural Language
and Graphics Generation. AI Journal, 63.