GOMS analysis

advertisement
GOMS Analysis &
Automated Usability Assessment
Melody Y. Ivory (UCB CS)
SIMS 213, UI Design & Development
March 8, 2001
GOMS Analysis Outline






2
GOMS at a glance
Model Human Processor
Original GOMS (CMN-GOMS)
Variants of GOMS
GOMS in practice
Summary
GOMS at a glance

Proposed by Card, Moran & Newell in 1983
–
apply psychology to CS

employ user model (MHP) to predict performance of tasks
in UI
–
–
applicable to


3
task completion time, short-term memory requirements
user interface design and evaluation
training and documentation
Model Human Processor (MHP)

Card, Moran & Newell (1983)
–
–
most influential model of user
interaction
3 interacting subsystems


cognitive, perceptual & motor
each with processor & memory
–

4
described by parameters
 e.g. capacity, cycle time
serial & parallel processing
Adapted from slide by Dan Glaser
MHP

Card, Moran & Newell (1983)
–
principles of operation

subsystem behavior under certain
conditions
–

5
e.g. Fitts’s Law, Power Law of
Practice
ten total
MHP Subsystems

Perceptual processor
–
–
–
6
sensory input (audio & visual)
code info symbolically
output into audio & visual image
storage (WM buffers)
MHP Subsystems

Cognitive processor
–
–
input from sensory buffers
access LTM to determine response

–
7
previously stored info
output response into WM
MHP Subsystems

Motor processor
–
–
8
input response from WM
carry out response
MHP Subsystem Interactions


Input/output
Processing
–
serial action

–
parallel perception

9
pressing key in response to light
driving, reading signs & hearing
MHP Parameters

Based on empirical data
–

Processors have
–

word processing in the ‘70s
cycle time ()
Memories have
–
–
–
storage capacity ()
decay time of an item ()
info code type ()

10
physical, acoustic, visual & semantic
Perceptual Subsystem Parameters

Processor
–

cycle time () = 100 msec
Visual Image Store
–
–
–
storage capacity () = 17
letters
decay time of an item () =
200 msec
info code type () = physical

physical properties of visual
stimulus
–
11
e.g. intensity, color,
curvature, length
One Principle of Operation

Power Law of Practice
–
task time on the nth trial follows a power law




12
Tn = T1 n-a, where a = .4
i.e., you get faster the more times you do it!
applies to skilled behavior (perceptual & motor)
does not apply to knowledge acquisition or quality
Original GOMS (CMN-GOMS)


Card, Moran & Newell (1983)
Engineering model of user interaction
–
task analysis (“how to” knowledge)

Goals - user’s intentions (tasks)
–

e.g. delete a file, edit text, assist a customer
Operators - actions to complete task
–
cognitive, perceptual & motor (MHP)
– low-level (e.g. move the mouse to menu)
13
CMN-GOMS

Engineering model of user interaction
–
task analysis (“how to” knowledge)

Methods - sequences of actions (operators)
–
–

Selections - rules for choosing appropriate method
–
–
method predicted based on context
explicit task structure

14
based on error-free expert
may be multiple methods for accomplishing same goal
 e.g. shortcut key or menu selection
hierarchy of goals & sub-goals
Text-Editing
Example
15
CMN-GOMS Analysis

Analysis of explicit task structure
–
add parameters for operators


–
predict user performance


–
16
approximations (MHP) or empirical data
single value or parameterized estimate
execution time (count statements in task structure)
short-term memory requirements (stacking depth of task
structure)
apply before user testing (reduce costs)
Limitations of CMN-GOMS

No directions for task analysis
–

Serial v.s. parallel perceptual processing
–


granularity (start & stop)
contrary to MHP
Only one active goal
Error-free expert performance
–
no problem solving or evaluation

17
Norman’s Action Cycle
Norman’s Action Cycle
Goals
Evaluation
Execution
Intention to act
Evaluation of interpretations
Sequence of actions
Interpreting the perception
Execution of sequence of actions
Perceiving the state of the world
GOMS
18
The World
Variants of GOMS

Keystroke-Level Model (KLM)
–
simpler than CMN-GOMS

six keystroke-level primitive operators
–
–
–
–
–
–


–
19
K - press a key or button
P - point with a mouse
H - home hands
D - draw a line segment
M - mentally prepare to do an action
R - system response time
no selections
five heuristic rules (mental operators)
still one goal activation
Text-Editing Example (KLM)
20
Variants of GOMS

Natural GOMS Language (NGOMSL)
–
more rigorous than CMN-GOMS

uses cognitive complexity theory (CCT)
–
user and system models
 mapping between user’s goals & system model
– user style rules (novice support)



task-analysis methodology
learning time predictions
flatten CMN-GOMS goal hierarchy
–
–
21
high-level notation (proceduralized actions) v.s. low-level operators
still one goal activation
Text-Editing Example (NGOMSL)
22
Variants of GOMS

Cognitive-Perceptual-Motor GOMS (CPMGOMS)
–
activation of several goals


–
23
uses schedule chart (PERT chart) to represent operators &
dependencies
critical path method for predictions
no selections
Text-Editing Ex. (CPM-GOMS)
24
GOMS in Practice





Mouse-driven text editor (KLM)
CAD system (KLM)
Television control system (NGOMSL)
Minimalist documentation (NGOMSL)
Telephone assistance operator workstation
(CMP-GOMS)
–
25
saved about $2 million a year
Summary


GOMS in general
–
The analysis of knowledge of how to do a task in terms of the components of
goals, operators, methods & selection rules. (John & Kieras 94)
–
CMN-GOMS, KLM, NGOMSL, CPM-GOMS
Analysis entails



task-analysis
parameterization of operators
predictions
–
26
execution time, learning time (NGOMSL), short-term memory
requirements
Automated Usability Assessment
Outline




27
Automated Usability Assessment?
Characterizing Automated Methods
Automated Assessment Methods
Summary
Automated Usability Assessment?



28
What does it mean to automate assessment?
How could this be done?
What does it require?
Characterizing Automated
Methods: Method Classes

Testing
–

Inspection
–

an evaluator uses a set of criteria or heuristics to identify
potential usability problems in an interface
Inquiry
–
29
an evaluator observes users interacting with an interface (i.e.,
completing tasks) to determine usability problems
users provide feedback on an interface via interviews, surveys,
etc.
Characterizing Automated
Methods: Method Classes

Analytical Modeling
–

Simulation
–
30
an evaluator employs user and interface models to
generate usability predictions
an evaluator employs user and interface models to
mimic a user interacting with an interface and report
the results of this interaction (e.g., simulated
activities, errors and other quantitative measures)
Characterizing Automated
Methods: Automation Types

None
–

Capture
–

software automatically identifies potential usability problems
Critique
–
31
software automatically records usability data (e.g., logging
interface usage)
Analysis
–

no level of automation supported (i.e.,evaluator performs all
aspects of the evaluation method)
software automates analysis and suggests improvements
Characterizing Automated
Methods: Effort Levels

Minimal Effort
–

Model Development
–

requires completion of freely chosen tasks (i.e., unconstrained
use by a user or evaluator)
Formal Use
–
32
requires the evaluator to develop a UI model and/or a user
model in order to employ the method
Informal Use
–

does not require interface usage or modeling
requires completion of specially selected tasks (i.e.,
constrained use by a user or evaluator)
Automated Assessment Methods
Method Class
Testing
Inspection
Inquiry
Analytical
Modeling
Simulation
33
Automation Type
Analysis
Log File Analysis
(FIM)
Capture
Critique
Performance
Measurement (F)
Remote Testing
(FI)
Cognitive
Guideline Review
Guideline
Walkthrough (F)
Review (M)
Questionnaires (FI)
GOMS Analysis (M)
Cognitive Task
Analysis (M)
Programmable User
Models (M)
Genetic Algorithm Information
Modeling
Processor Modeling
Information Scent (M)
Modeling (M)
Petri Net Modeling
(M)
Automated Assessment Methods:
Generating Usage Data

Simulation – Automated Capture
–

Mimic user and record activities for subsequent analysis
Genetic Algorithm Modeling
–
–
Script interacts with running interface (Motif-based UI)
Deviation points in script  behavior determined by genetic
algorithm

–
Inexpensively generate a large number of usage traces

–
34
Mimic novice user learning by exploration
Find weak spots, failures, usability problems, etc.
Requires manual evaluation of trace execution
Automated Assessment Methods:
Generating Usage Data

Information Scent Modeling
–
Mimic users navigating a Web site and record paths


Web site model – linking structure, usage data, and content
similarity
Considers information scent (common keywords between user
goals and link text) in choosing links
–


Does not consider impact of page elements, such as images,
reading complexity, etc.
Stopping criteria
–
–
Reach target pages or some threshold (e.g., maximum navigation
time)
Requires manual evaluation of navigation paths

35
Percentage of agents follow higher- and lower-scent links
Log file visualization tool (Dome-Tree Visualization)
Automated Assessment Methods:
Detecting Guideline Conformance

Inspection – Automated Analysis
–
–

Quantitative Screen Measures
–
–

Size of screen elements, alignment, balance, etc.
Possibly generate initial layouts (AIDE)
Interface Consistency (Sherlock)
–
–
36
Cannot automatically detect conformance for all guidelines
One study – 78% best case, 44% worst case
Same widget placement and terminology (Visual
Studies showed 10-25% speedup for consistent UIs
Automated Assessment Methods:
Detecting Guideline Conformance

Quantitative Web Measures
–
–
Words, links, graphics, page breadth & depth, etc.
(Rating Game, HyperAT, TANGO)
Most techniques not empirically-validated


HTML Analysis (WebSAT)
–
37
Web TANGO uses expert ratings to develop prediction
models
All images have alt tags, one outgoing link/page,
etc.
Automated Assessment Methods:
Detecting Guideline Conformance

Web Scanning Path
(Design Advisor)
–
Determine how users will
scan a page based on
attentional effects of
elements

38
motion, size, images, color,
text style, and position
Automated Assessment Methods:
Suggesting Improvements


Inspection – Automated Critique
Rule-based critique systems
–
Typically done within a user interface management system

–

–
Apply guidelines relevant to each graphical object
Widely applicable to Windows UIs
HTML Critique
–
39
X Window UIs (KRI/AG), control systems (SYNOP), space
systems (CHIMES)
Object-based critique systems (Ergoval)
–

Very limited application
–
Syntax, validation, accessibility (Bobby), and others
Although useful, not empirically validated
Automated Assessment Methods:
Modeling User Performance

Analytical Modeling – Automated Analysis
–
–

Predict user behavior, mainly execution time
No methods for Web interfaces
GOMS Analysis (previously discussed)
–
–
Generate predictions for GOMS models (CATHCI, QGOMS)
Generate model and predictions (USAGE, CRITIQUE)


Cognitive Task Analysis
–
Input interface parameters to an underlying theoretical model
(expert system)

–
40
UIs developed within user interface development environment
Do not construct new model for each task
Generate predictions based on parameters as well as
theoretical basis for predictions
Automated Assessment Methods:
Modeling User Performance

Programmable User Models
–
–
Cross between GOMS and CTA analyses
Program UI on a psychologically-constrained
architecture


41
Constraint violations suggest usability problems
Generate quantitative predictions
Automated Assessment Methods:
Simulating User Behavior


Simulation – Automated Analysis
Petri Net Modeling (AMME)
–
–
–

Construct petri net from logged interface usage
Simulates problem solving process (learning, decisions, and
task completion)
Outputs measure of behavior complexity
Information Processor Modeling (ACT-R, SOAR,
CCT,…)
–
Methods employ sophisticated cognitive architecture with
varying features

42
Modeled tasks and components, predictions, etc.
Automated Assessment Methods:
Simulating User Behavior

Web Site Navigation (WebCriteria)
–
Similar to GOMS Analysis

Constructs model of site and predicts navigation time for a
specified path
–
Based on idealized Web user (Max)
– Navigation time only for shortest path between endpoints
 Does not consider impact of page elements (e.g.,
colors,reading complexity, etc.)
– Reports on page freshness and composition of pages (text, image,
applets, etc.)
– Supports only a small fraction of analysis possible with guideline
review approaches
–
43
Used to compare sites (Industry Benchmarks)
Summary

Characterizing Automated Methods
–

Automated Methods
–
–
–
–

Mainly automated capture and analysis
Guideline review enables automated critique
Represented only 33% of 132 surveyed approaches
Most require formal or informal interface usage
More Information
–
–
–
44
Method Class, Automation Types, Effort Levels
webtango.berkeley.edu
Survey paper on automated methods
Papers on quantitative Web page analysis
Download