Intelligent Agents to Deliver Learning

advertisement
Intelligent
Agents to
Deliver
Learning
Materials
Leen-Kiat Soh
Computer Science & Engineering
University of Nebraska
Lincoln, NE 68588-0115
lksoh@cse.unl.edu
http://www.cse.unl.edu/agents
Future Problem
Solving Workshop—
Hastings
January 21, 2004
Challenges
• Customized education
– e.g., Modularized courseware to meet specific
requirements, deficiencies, and sequences
• Adapting education
– e.g., Flexible courseware that adapts in real time
to student behavior and aptitude
• Effective distance education
– e.g., tools with anytime, anywhere capabilities
– e.g., infrastructures to bring classroom experience
to distance learners
Problem
• Corbett et al. (1999):
– “the arsenal of sophisticated computational modules
inherited from AI produce learning gains of approximately .3
to 1.0 standard deviation units compared with students
learning the same content in a classroom.”
• Graesser et al. (2001):
– “Human tutors produce impressive learning gains (between
.4 and 2.3 standard deviation units over classroom
teachers), even though the vast majority of tutors in a
school’s system have modest domain knowledge, have no
training in pedagogical techniques, and rarely use the
sophisticated tutoring strategies of ITSs.”
ITS = Intelligent Tutoring System
Problem 2
• Woolf et al. (2002) also lists abilities that are
needed or present in tutors:
– Generative:
• Generates appropriate instructional material (problems,
hints, help) based on student performance
– Student modeling:
• Assesses the current state of a student’s knowledge and
does something instructionally useful based on the
assessment
– Expert modeling:
• Models expert performance and does something
instructionally useful based on domain knowledge
Problem 3
• Woolf et al. (2002) also lists abilities that are
needed or present in tutors, cont’d:
– Instructional modeling:
• Changes pedagogical strategies based on the changing
state of the student model, prescriptions of an expert
model, or both
– Mixed-initiative
• Human Computer Interaction (HCI)
– Self-improving:
• Capacity to monitor, evaluate, and improve its own
teaching performance as a function of experience
Problem 4
• Graesser et al. (2001) criticize the current
state of tutoring systems:
– If students merely keep guessing until they find an action
that gets positive feedback, they can learn to do the right
thing for the wrong reasons – shallow learning
– The tutor does not ask students to explain their actions –
multiple choice questions
– The user interface of tutoring systems requires students to
display many of the details of their reasoning – no stepping
back to see the “basic approach”
– When students learn quantitative skills (e.g., algebra or
physics problem solving), they are usually not encouraged to
see their work from a qualitative, semantic perspective
Solution Ideas
• Examples
– PACT (Koedinger et al. 1997): algebra, geometry, and
computer languages
– ANDES (Gertner and VanLehn 2000; VanLehn 1996): physics
• Used as an adjunct to college and high-school physics courses to help
students do their homework problems
• Has an immediate feedback to enhance learning
– SHERLOCK (Lesgold et al. 1992): electronics
– ATLAS (VanLehn et al. 2000):
• Model-tracing
• Students scored significantly higher than the ANDES students on a
conceptual post-test
Solution Ideas 2
• Examples
– AutoTutor (Graesser et al. 2001): introductory
course in computer literacy
• Fundamentals of computer hardware, OS, Internet, etc.
• Was designed to be a good conversational partner that
comprehends, speaks, points, and displays emotions, all
in a coordinated fashion
• Simulates a multi-turn conversation to extract more
information from the student and get the student to
articulate pieces of the answer
• Pumps the student for what s/he knows before drilling
down to specific pieces of the answer
• Uses Latent Semantic Analysis (LSA) to compute
matches between the student’s speech acts to the
expectations
Solution Ideas 3
• Examples
– AutoTutor (Graesser et al. 2001): introductory
course in computer literacy, cont’d:
• Feeds back to the student at three levels:
• (a) backchannel feedback that acknowledges the
learner’s input
• (b) evaluative pedagogical feedback on the learner’s
previous turn based on the LSA values of the learner’s
speech acts (negative, neutral negative, neutral positive,
positive)
• (c) corrective feedback that repairs bugs and
misconceptions that learners articulate
Solution Ideas 4
• Examples
– CIRCISM (Freedman and Evens 1996)
– BEE (Basic Electricity and Electronics) tutor (Rosé
et al. 1999)
– EVELYN Reading Coach and EMILY Reading Coach
(Mostow and Aist 2002): Project LISTEN
• Help students read by listening to children read aloud
– BELVEDERE (Suthers et al. 2002):
• Supports students in collaboratively solving ill-structured
problems in science and other areas (such as public
policy) as they develop critical inquiry skills
Solution Ideas 5
• Teachable Agents (Biswas et al. 2002)
– let students teach the agents to do things;
through teaching, the students learn
• Articulate Software (Forbus 2002); properties
are:
– It should be fluent (some understanding of the
subject being taught)
– It should be supportive (scaffolding)
– It should be generative (pose new questions)
– It should be customizable (manually)
Solution Ideas: Intelligent Agents
• What is an agent?
– An agent is an entity that takes sensory input from
its environment, makes autonomous decisions,
and carries out actions that affect the environment
– A thermostat is an agent
– A calculator is not an agent
Agent
sensory
input
think!
Environment
output
actions
Solution Ideas: Intelligent Agents 2
• What is an intelligent agent?
– An intelligent agent is one that is capable of flexible
autonomous actions in order to meet its design
objectives, where flexibility means:
• Reactivity: agents are able to perceive their environment, and
respond in a timely fashion to changes that occur in order to
satisfy their design objectives
• Pro-activeness: agents are able to exhibit goal-directed
behavior by taking the initiative in order to satisfy their design
objectives
• Social ability: agents are capable of interacting with other
agents (and possibly humans) in order to satisfy their design
objectives
(Wooldridge and Jennings 1995)
Solution Ideas: Intelligent Agents 3
• Machine Learning in AI says
The acquisition of new knowledge and motor and cognitive skills
and the incorporation of the acquired knowledge and skills in
future system activities, provided that this acquisition and
incorporation is conducted by the system itself and leads to an
improvement in its performance.
• Agents that learn are intelligent
• Not all agents are intelligent!
Solution Ideas: Agent Environment
• Inaccessible vs. accessible
– Incomplete vs. complete data
• Deterministic vs. non-deterministic
– Certainty vs. uncertainty
• Episodic vs. non-episodic
– Each episode is independent or not
• Static vs. dynamic
– Remain unchanged except by the performance
of actions by the agent?
• Discrete vs. continuous
– “Chess game” vs. “taxi driving”
Solution Ideas: Why Agents?
• If the system-to-be-built has, during the
execution of the system
Why does a person hire an agent?
– Incomplete data
– Uncertainty in the assessment/interaction of its
environment
– Inter-dependent episodes of events
– No full control over the events in the environment
– An “open world”, instead of a “closed world”
• In other words, agents are used when you need
to build a system that is adaptive to an uncertain,
dynamic, and at times unexpected environment
– So you can make full use of the autonomous property
of an agent
Evaluation Criteria
• Pre-Design
Some examples, not exhaustive
– Necessity
• Is the agent solution necessary? Or an overkill?
– Feasibility
• Does the agent solution make sense? Is it impossible to
implement? Do we have the resources? Will it work?
• During Design
– Modularity
• Are there different modules and components? Is the “data”
separated from the “brain”?
– Extensibility
• What if we want to apply it to another problem?
– Scalability
• What if we want to apply it to the same but bigger problem?
Evaluation Criteria 2
• Post-Design
Some examples, not exhaustive
– Correctness
• Are the algorithms correctly designed? Are the
solutions correct?
– Usefulness
• Is the solution useful? Does it actually address the
problem? Does it help the user?
– Reliability
• Does the solution work for all possible problems?
Will the performance deteriorate after a while?
– Adaptiveness/Learning
• Can the solution evolve by itself to solve new
problems or to become better at solving old
problems?
AI
ILMDA
• Intelligent Learning Materials Delivery
Agent
Bibliography
•
•
•
•
•
•
Allen, J. F., D. K. Byron, M. Dzikovska, G. Ferguson, L. Galescu, and A. Stent
(2001). Toward Conversational Human-Computer Interaction, AI Magazine,
22(4):27-37.
André, E. and T. Rist (2001). Controlling the Behavior of Animated
Presentation Agents in the Interface: Scripting versus Instructing, AI
Magazine, 22(4):53-66
Birmingham, W. P., et al (1994. The University of Michigan Digital Library:
This Is Not Your Father's Library, Proc. Digital Libraries 94, Hypermedia
Research Laboratory, Texas A&M University, College Station, Tex., pp. 53-60.
Biswas, G., D. Schwartz, J. Bransford, and the Teachable Agents Group at
Vanderbilt (2002). Technology Support for Complex Problem Solving: From
SAD Environments to AI, in Forbus, K. D. and P. J. Feltovich (2002). (Eds.)
Smart Machines in Education, Menlo Park, CA: AAAI Press, pp. 71-97.
Cassell, J. (2001). Embodied Conversational Agents: Representation and
Intelligence in User Interfaces, AI Magazine, 22(4):67-83.
Cassell, J. and H. Vilhjálmsson (1999). Fully Embodied Conversational
Avatars: Making Communicative Behaviors Autonomous, Autonomous Agents
and Multi-Agent Systems, 2:45-64.
Bibliography 2
•
•
•
•
•
•
Cohen, P. R. and S. L. Oviatt (1995). The Role of Voice Input for HumanMachine Communication, in Proceedings of the National Academy of
Sciences, 92(22):9921-9927.
Corbett, A., J. Anderson, A. Graesser, K. Koedinger, and K. VanLehn (1999).
Third Generation Computer Tutors: Learn from or Ignore Human Tutors? in
Proceedings of the 1999 Conference of Computer-Human Interaction, 8586.
D’Souza, A., J. Rickel, B. Herreros, and W. L. Johnson (2001). An
Automated Lab Instructor for Simulated Science Experiments, in J. D.
Moore, C. L. Radfield, and W. L. Johnson (eds.) Artificial Intelligence in
Education: AI-ED in the Wired and WirelessFuture, 65-76, Amsterdam: IOS.
Forbus, K. D. (2002). Articulate Software for Science and Engineering
Education, in Forbus, K. D. and P. J. Feltovich (2002). (Eds.) Smart
Machines in Education, Menlo Park, CA: AAAI Press, pp. 235-267.
Forbus, K. D. and P. J. Feltovich (2002). (Eds.) Smart Machines in
Education, Menlo Park, CA: AAAI Press.
Freedman, R. and M. W. Evens (1996).
Generating and Revising
Hierarchical MultiTurn Text Plans in an ITS, in Intelligent Tutoring Systems:
Proceedings of the 1996 Conference, 632-640.
Bibliography 3
•
•
•
•
•
Gertner, A. S. and K. VanLehn (2000). ANDES: A Coached Problem-Solving
Environment for Physics, in Intelligent Tutoring Systems: Fifth International
Conference, ITS 2000, eds. G. Gautheier, C. Frasson, and K. VanLehn, 133142.
Graesser, A. C., K. Wiemer-Hastings, P. Wiemer-Hastings, R. Kreuz, and the
Tutoring Research Group (1999). AutoTutor: A Simulation of a Human Tutor,
Journal of Cognitive Systems Research, 1(1):35-51.
Graesser, A. C., K. VanLehn, C. P. Rosé, P. W. Jordan, and D. Harter (2001).
Intelligent Tutoring Systems with Conversational Dialogue, AI Magazine,
22(4):39-51.
Herzog, M. and P. Petta (1995). Multimedia Information System in OpenWorld Domains, in Working Notes of 1st Int. Workshop on Multimedia
Information Systems, Arlington, VA, September 28-30, 13-16.
Herzog, M. and P. Petta (1996). Knowledge Representation in WERKL, an
Architecture for Intelligent Multimedia Information Systems, in Workshop
Notes of the ECAI'96 Workshop on Knowledge Representation for Interactive
Multimedia Systems: Research and Experience (KRIMS'96), ECAI'96,
•
Budapest, Hungary, August 12, 17-25.
Johnson, W. L. (2001). Pedagogical Agent Research at CARTE, AI Magazine,
22(4):85-94.
Bibliography 4
•
•
•
•
•
•
Johnson, W. L., J. W. Rickel, and J. C. Lester (2000). Animated Pedagogical
Agents: Face-to-Face Interaction in Interactive Learning Environments,
International Journal of Artificial Intelligence in Education, 11:47-78.
Koedinger, K. R., J. R. Anderson, W. H. Hadley, and M. A. Mark (1997).
Intelligent Tutoring Goes to School in the Big City, Journal of the Artificial
Intelligence in Education, 8(1):30-43.
Lesgold, A., S. Lajoie, M. Bunzo, and G. Eggan (1992). SHERLOCK: A
Coached Practice Environment for an Electronics Troubleshooting Job, in J. H.
Larkin and R. W. Chabay (eds.) Computer Assisted Instruction and Intelligent
Tutoring Systems, 201-238, Hillsdale, NJ: Lawrence Erlbaum.
Lester, J. C., C. B. Callaway, J. P. Grégoire, G. D. Stelling, S. G. Towns, and L.
S. Zettlemoyer (2002). Animated Pedagogical Agents in Knowledge-Based
Learning Environments, in Forbus, K. D. and P. J. Feltovich (2002). (Eds.)
Smart Machines in Education, Menlo Park, CA: AAAI Press, pp. 269-298.
Marsella, S. and W. L. Johnson (1998). An Intelligent Assistant for Team
Training in Dynamic Multi-Agent Virtual Worlds, in Proceedings of Intelligent
Tutoring Systems, 4th International Conference, 464-473.
Mostow, J. and G. Aist (2002). Evaluating Tutors that Listen: An Overview of
Project LISTEN, in Forbus, K. D. and P. J. Feltovich (2002). (Eds.) Smart
Machines in Education, Menlo Park, CA: AAAI Press, pp. 169-234.
Bibliography 5
•
•
•
•
•
•
Raines, T., M. Tambe, and S. Marsella (2000). Automated Assistants to Aid
Humans in Understanding Team Behaviors, in Proceedings of the Fourth
International Conference on Autonomous Agents, 419-426.
Rickel, J. and W. L. Johnson (1999). Animated Agents for Procedural
Training in Virtual Reality: Perception, Cognition, and Motor Control, Applied
Artificial Intelligence, 13(4):343-382.
Rickel, J., R. Ganeshan, C. Rich, C. L. Sidner, and N. Lesh (2000). TaskOriented Tutorial Dialogue: Issues and Agents, paper presented at the AAAI
Fall Symposium on Building Dialogue Systems for Tutorial Applications,
November 3-5, North Falmouth, MA.
Rosé, C. P., B. Di Eugenio, and J. Moore (1999). A Dialogue-Based Tutoring
System for Basic Electricity and Electronics, in S. P. Lajoie and M. Vivet (eds.)
Artificial Intelligence in Education, 759-761, Amsterdam: IOS.
Rusell, S. and P. Norvig (1995). Artificial Intelligence: A Modern Approach,
Prentice-Hall.
Schank, R. and A. Neaman (2002). Motivation and Failure in Educational
Simulation Design, in Forbus, K. D. and P. J. Feltovich (2002). (Eds.) Smart
Machines in Education, Menlo Park, CA: AAAI Press, pp. 37-69.
Bibliography 7
•
•
•
•
•
•
•
Shaw, E., W. L. Johnson, and R. Ganeshan (1999). Pedagogical Agents on
the Web, in Proceedings of the Third International Conference on
Autonomous Agents, 283-290.
Smith, J. (2000). GrandChair: Conversational Collection of Family Stories, M.
S. Thesis, Media Lab, Massachusetts Institute of Technology, Cambridge, MA.
VanLehn, K. (1996). Conceptual and Metalearning during Coached Problem
Solving, in Proceedings of the Third Intelligent Tutoring Systems Conference,
29-47, Berlin: Springer-Verlag.
Suthers, D., J. Connelly, A. Lesgold, M. Paolucci, E. E Toth, J. Toth, and A.
Weiner (2002). Representational and Advisory Guidance for Students
Learning Scientific Inquiry, in Forbus, K. D. and P. J. Feltovich (2002). (Eds.)
Smart Machines in Education, Menlo Park, CA: AAAI Press, pp. 7-35.
Weiss, G. (ed.) (1999). Multiagent Systems: A Modern Approach to
Distributed Artificial Intelligence, MIT Press.
Wooldridge, M. and N. R. Jennings (1995). Intelligent Agents: Theory and
Practice, The Knowledge Engineering Review, 10(2):115-152.
Woolf, B. P., J. Beck, C.Eliot, and M. Stern (2002). Growth and Maturity of
Intelligent Tutoring Systems: A Status Report, in Forbus, K. D. and P. J.
Feltovich (2002). (Eds.) Smart Machines in Education, Menlo Park, CA: AAAI
Press, pp. 99-144.
Latent Semantic Analysis
• Latent Semantic Analysis (LSA) is a theory and method for extracting
and representing the contextual-usage meaning of words by
statistical computations applied to a large corpus of text (Landauer
and Dumais, 1997). The underlying idea is that the aggregate of all
the word contexts in which a given word does and does not appear
provides a set of mutual constraints that largely determines the
similarity of meaning of words and sets of words to each other.
– Based on Landauer, T. K., P. W., and D. Laham (1998) Introduction to
Latent Semantic Analysis, Discourse Processes, 25:259-284.
Download