A Sketch Interface for Mobile Robots Marjorie Skubic Craig Bailey George Chronis Computational Intelligence Research Lab University of Missouri-Columbia Outline • • • • • Motivation and context Route maps The PDA sketch interface Experimental study and results Conclusions and future work Spatial Reasoning with Guinness References Acknowledgements Route Maps • Tversky’s work – Depictions vs. Descriptions – Extraction of route descriptions – 1 to 1 correlation • Michon and Denis – Landmarks and critical nodes The Sketch Interface • • • • • • • • Objects Labels Paths Delete Start Move Undo Send Objects • Closed Polygons • Any shape or size • Thresholds to determine gap closure • Feedback on recognition – Sound – Color Labels • Default numbering for object labels • Tap on screen to edit • Can use Palm OS Graffiti recognition or a software keyboard Paths • Limit of one • A minimum length required • Color Feedback Path Direction • Default direction is the direction the path is drawn • User can specify the direction with a sketched “blob” to denote the start of the path Recognized by – Number of points – Average distance of all points – Proximity to path endpoint Delete • An intuitive delete: cross out an object • Recognized by – Two consecutive strokes – Both lengths shorter than a path – The strokes cross • Color feedback • Search for closest object or path Determining Crossed Marks • • • • Use the slope equations of lines Endpoints of strokes determine the line A pair of decision parameters can be computed If both parameters lie between 0 and 1, then the two strokes must have an intersection (X1,Y1) (X3,Y3) ua= (Y4-Y3)(X2-X1) - (X4-X3)(Y2-Y1) ub= (X4,Y4) (X2,Y2) (X4-X3)(Y1-Y3) - (Y4-Y3)(X4-X3) (X2-X1)(Y1-Y3) - (Y2-Y1)(X1-X3) (Y4-Y3)(X2-X1) - (X4-X3)(Y2-Y1) IF (0 < ua < 1) AND (0 < ub < 1) THEN the lines intersect Menu Commands • Also accessible through graffiti • m Move • u Undo • c Clear • t Transmit • f Configure “Digitizing” the Sketch User Evaluation • Tested how well the interface performed with real users • Pre-experimental questionnaire • Tasks – Sketch tasks – Re-sketch tasks – Task scores • Post-experimental questionnaire • Questionnaires contain Lickert style statements (Lickert, 1932) along with several open-ended questions Statistical Analysis 2 groups, 2 scenes: • Compared by scene sketched • Compared by course level of participant • Means compared with the t test • Null Hypothesis: there are no differences when compared by sketched scene or course level Participants • • • • 26 students from CS courses One participant scores was not used Only 5 owned a PDA Students of Scene B rated themselves significantly better at giving directions (p = 0.02) • No differences when compared by course level Scene A Example Sketches of Scene A Scene B Example Sketches of Scene B Post-Experimental Survey: Landmark Scores 1 = very difficult; 5 = very easy • Creating Landmarks – 4.6 ± 0.6 • Deleting Landmarks – 4.2 ± 0.9 • Usefulness of Deleting – 4.7 ± 0.6 • Usefulness of Labeling – 4.8 ± 0.6 Post-Experimental Survey: Path Scores 1 = very difficult; 5 = very easy • Creating a path – 4.4 ± 1.0 • Deleting a path – 4.4 ± 1.0 • Usefulness of deleting – 4.7 ± 0.7 • Usefulness of the starting point – 4.2 ± 0.9 Post-Experimental Survey: Overall Scores • Usefulness of changing sketch – 4.8 ± 0.4 • Usefulness of deleting sketch – 4.2 ± 1.0 • How well sketch represents environment – 83.6 ± 7.4 • Overall ease of interface – 4.4 ± 0.6 Usability Results • Only two significant differences (p<=0.05) were found among the scores – Usefulness of deleting by scene (p=0.0) – Final sketch rating by scene (p=0.05) • In both cases, students in scene B rated higher • Same group that rated themselves better at giving directions • Differences were not found when compared by course level • The Null Hypothesis is accepted Task Score Results • Collected sketches were scored – – – – – – +1 for starting landmark +1 for each correct turn +1 for landmark at turn +1 for each correct straight segment + 1 for ending landmark -1 for extra turns or straight segments • No significant differences found (p=0.12) – Sketch Task Score – Re-sketchTask Score = 0.91 ± 0.11 = 0.82 ± 0.26 Conclusions • Created a new sketch based interface on a handheld computer • Intuitive and little reliance on traditional menus and icons • User evaluation finds the interface as easy to use as pencil and paper by 2:1 Future Work • • • • Continue integration into the Guinness system Recognition of more sketched symbols Recognition of turning rate Creation of 3D virtual environments with libraries of objects Email: SkubicM@missouri.edu Web: www.cecs.missouri.edu/~skubic funded by the Naval Research Lab ARCHITECTURE user commands and responses oldest short term map query & label SRserver SR & map info mapserver speech commands imageserver Cortex spatial behaviors robot commands continuous localization corrections pose gesture GUI(EUT) obstacle avoidance trulla PDA vfh robot speech user commands and responses short term map long term map sketch directives & feedback SRserver Behind the table User: Robot: User: User: Robot: User: How many objects do you see? I am sensing four objects. Object 2 is a table. Describe the scene. There are objects on my front right. The object number 4 is mostly in front of me. The table is behind me. Go behind the table. between object 1 and object 2 using the midpoint between closest points using the midpoint between centroids using the CFMD Image Server Understanding Sketched Route Maps PATH DESCRIPTION GENERATED FROM THE SKETCHED ROUTE MAP 1. When table is mostly on the right and door is mostly to the rear (and close) Then Move forward 2. When chair is in front or mostly in front Then Turn right 3. When table is mostly on the right and chair is to the left rear Then Move forward 4. When cabinet is mostly in front Then Turn left 5. When ATM is in front or mostly in front Then Move forward 6. When cabinet is mostly to the rear and tree is mostly on the left and ATM is mostly in front Then Stop References [1] [2] [3] [4] [5] M. Skubic, P. Matsakis, G. Chronis and J. Keller, "Generating MultiLevel Linguistic Spatial Descriptions from Range Sensor Readings Using the Histogram of Forces", Autonomous Robots, Vol. 14, No. 1, Jan., 2003, pp. 51-69. M. Skubic, D. Perzanowski, S. Blisard, A. Schultz, W. Adams, M. Bugajska and D. Brock “Spatial Language for Human-Robot Dialogs,” IEEE Transactions on SMC, Part C, to appear in the special issue on Human-Robot Interaction. M. Skubic, S. Blisard, C. Bailey, J.A. Adams and P. Matsakis, "Qualitative Analysis of Sketched Route Maps: Translating a Sketch into Linguistic Descriptions," IEEE Transactions on SMC Part B, to appear. G. Chronis and M. Skubic, “Sketch-Based Navigation for Mobile Robots,” In Proc. of the IEEE 2003 Intl. Conf. on Fuzzy Systems, May, 2003, St. Louis, MO. G. Scott, J.M. Keller, M. Skubic and R.H. Luke III, “Face Recognition for Homeland Security: A Computational Intelligence Approach,” In Proc. of the IEEE 2003 Intl. Conf. on Fuzzy Systems, May, 2003, St. Louis, MO. Guinness and Gang From left to right George Chronis, Grant Scott, Dr. Marge Skubic, Matt Williams, Craig Bailey, Bob Luke, Charlie Huggard and Sam Blisard Missing: Dr. Jim Keller Sketch-Based Navigation The sketched route map The robot traversing the sketched route Sketch-Based Navigation The digitized sketched route map The robot traversing the sketched route Acknowledgements This work has been supported by ONR and the U.S. Naval Research Lab. Natural language understanding is accomplished using a system developed by NRL, called Nautilus [Wauchope, 2000]. We also want to acknowledge the help of Dr. Pascal Matsakis. NRL’s Multimodal Robot Interface