A Neural Network Approach to UGV Reconnaissance in MOUT Environments Brandon S. Perelman Michigan Technological University Abstract One important trend in unmanned systems technology is the push for increased automation. Many UGVs are currently tele-operated, and attempts have been made to automate individual components of the work burden to improve the operator-UGV ratio. This paper contains a review of the literature regarding UGV history, the tasks that automation must accomplish, challenges for automation, finally a proposition for neural network-based UGV software that would allow a micro UGV, deployed by a controller during operations in urban environments, to explore buildings and encode the environment, transmit the information to the human controller, then return. Introduction Unmanned ground vehicles (UGVs), in a dictionary sense, are ground-based mechanized systems for carrying or transporting items. UGVs specifically do not carry human operators because they are often selected specifically for jobs with qualities contraindicating human presence, such as environmental hazards, size or weight constraints, or endurance requirements (Gage, 1995). Examples of applications outside the military domain include space exploration (e.g., planetary rovers) and inspecting and assessing damage in radiological disaster sites, such as the reactors at Chernobyl and Three Mile Island (Fong & Thorpe, 2001). Early research efforts at the Stanford Research Institute, funded by the Defense Advanced Research Products Agency (DARPA), produced the first “intelligent” mobile robot, Shakey, in the late 1960s. This unit was tele-operated via radio frequency, and contained a sensor package including a video camera, a rangefinder, and touch sensors (Nilsson, 1969). Much of the initial work in the field conducted by Hans Moravec, initially at Stanford and then at Carnegie Melon University, focused on producing automation (for a comprehensive review of the literature, see Gage, 1995). Importantly, into the late 80’s, while efforts to produce truly automated UGVs for military tasks, such as reconnaissance, surveillance, and target acquisition (RSTA), encountered funding obstacles, tele-operated UGV projects such as the Advanced Teleoperator Technology (ATT) dune buggy demonstrated to military shareholders that it was possible to remotely control ground vehicles for RSTA operations, and even remotely operate weapons systems attached to those vehicles (Hightower, Smith, & Wiker, 1986). Currently, the US military employs a number of tele-operated UGVs for a variety of missions. As of October 2006, US Army UGV systems were involved in over 11,000 incidents involving improvised explosive devices (IEDs; Nardi, 2009). iRobot’s PackBot system was used in Afghanistan as early as 2002 to explore cave complexes, and Iraq in 2003 to search for chemical weapons and conduct vehicle searches (Yamauchi, 2004). Tele-operated Military UGVs show promise for a number of tasks, including RSTA operations, countermine and EOD operations, chemical, biological, radiological, nuclear, and explosive (CBRNE) searches (Nardi, 2009), logistics (Gage, 1995), load carriage to support infantry operations (e.g., MULE and its child-robots, Boston Dynamics’ Legged Squad Support System; DARPA, 2012), and even offensive operations (e.g., the Marine Corps’ Gladiator UGV; Apostolopoulos, 2014). The remainder of this manuscript is dedicated to (1) a justification for pursuing autonomy in UGVs, a discussion of (2) approaches to automation, (3) challenges for automation, and finally (4) a proposed neural network-based system for UGV operation facilitating RSTA in military operations in urban terrain (MOUT). Why Autonomy in UGVs? Currently in the US military there is a strong push to increase the extent to which UGVs are capable of operating autonomously (Nardi, 2009). There are a number of logistical reasons for this drive. First, reducing (or eliminating) the number of human operators required to control a single system reduces cost, and potentially allows one human to control multiple systems. Second, the UGV should ideally be able to react immediately to dynamic tactical situations without taking a human controller out of the fight (Mills, 2007). Third, a human controller requires an interface with which to control the UGV, which increases costs. Fourth, perceptual data from the UGV must be transmitted to the controller, and commands must be then transmitted to the UGV. Signal lag in this loop and perceptual limitations imposed by the UGV’s sensor package necessarily inhibit the controller’s ability to make control decisions, compared with an autonomous system that can make decisions without the lag associated with transmitting and interpreting information. Finally, research in the unmanned aerial vehicle (UAV) domain (e.g., Mellinger, Michael, & Kumar, 2012) with highly agile UAVs indicates that autonomous systems are capable of highly aggressive maneuvers that are simply too fast and complex for even well trained humans to execute reliably using tele-operation. Tele-operation Interface Constraints Tele-operation interfaces may be classified into four systems. In direct interfaces, the operator controls the UGV from the “inside out,” controlling the UGV with hand controllers and receiving information from the UGV typically in the visual modality. This method of control requires very high bandwidth low delay communication because video data often contains a great deal of information. Multimodal / multisensor interfaces provide the operator with a variety of control modes (e.g., individual actuator control versus coordinated motion) and multiple displays (i.e., text and visual information). These interfaces combine information from multiple sensors or data sources to present a single integrated view, and are appropriate for applications where the remote task environment (e.g., satellite service in space, or volcano exploration) is too complex to perceive in a way that allows sound decision making. Multimodal / multisensory interfaces are used to control many High Altitude Long Endurance UAVs, such as Predator. Novel interfaces include all “other” types of interfaces, with examples including haptic control, gesture control, and hands-free brainwave-based methods. Finally, supervisory control describes a scheme in which the operator divides a control problem into a sequence of subtasks that the robot executes autonomously. Supervisory control may be thought of as a bridge between teleoperation and automation (Fong & Thorpe, 2001). Currently, the majority of UGVs in the US arsenal are tele-operated, and the US military’s tactics, techniques, and procedures (TTPs) for employing UGVs reflects this state (Maxwell, Larkin, & Lowrance, 2013). In order to facilitate effect tele-operation, research efforts have focused on improving sensor packages (i.e., perceptual systems) and controllers (i.e., executive systems). UGV sensors are generally regarded as range-limited to the extent that operators are seldom aware of events and objects outside the UGV’s immediate surroundings. Since the UGV cannot perceive remote events, it cannot respond to them and, therefore, cannot secure larger areas (Kogut, Blackburn, & Everett, 2003). Situation awareness (SA) and workload are serious concerns for UGV operators and the effects of various manipulations on these factors have been extensively studied. In their meta-analysis, Coovert, Prewett, Saboe, & Johnson (2010) illuminated a number of factors impacting aspects of UGV operation. Some of these factors are particularly pertinent to MOUT operations. For example, complex (e.g., urban) environments are demonstrably more difficult to navigate than sparse environments (e.g., the ocean floor). The literature concerning field of view (FOV), important in constricted urban environments, is particularly inconclusive. For example, wide FOV decreases workload, improves SA, and offers task-dependent performance benefits. However, wide FOV is associated with increased motion sickness and difficulty steering. In simulations, participants required maps to carry out coarse navigation. Maps of the environment are not always available and GPS is not always reliable, and using a map requires that the controller split his attention between the map and the UGV’s sensors. Tandem control schemes have been developed (e.g., Chen, 2010; Ha & Lee, 2013) that allow a UAV to increase the UGV operator’s SA without a need for GPS, but such schemes require direct line of sight between the unmanned systems, which would be obstructed in many urban environments, and the expense of fielding an additional unmanned system. Finally, workload issues are compounded when the UGV operator has duties beyond control, such as operating weapon systems (Irvin, Leo, & Kim, 2012). To address workload issues with controlling UGVs, researchers have tested different styles of joystick (Ögren, Svenmarck, Lif, Norberg, & Söderbäck, 2013) and smartphone control schemes (Walker, Miller, & Ling, 2013). Both of these types of controllers have associated costs and rewards, however, and neither addresses the workload burden of simultaneously controlling the UGV’s movement and performing other tasks (Irvin et al., 2012). In summary, tele-operation research generally indicates tradeoffs with any conceivable interface, at best, or unrealistic expectations, at worst. Providing the speed, tactical flexibility, and small “footprint” required for MOUT operations requires automation. Current Approaches to Automation One definition of the minimum requirement for true autonomy is the “capability for both planning the vehicle course and controlling the position of the vehicle on route.” (Childers, Bodt, & Camden, 2011) For the purposes of this manuscript, the definition of true autonomy is expanded to include an unmanned system’s ability to independently achieve stated mission objectives. For example, a UGV tasked with RSTA should be capable of scouting independently and recording the necessary information to provide its commander with a comprehensive account of the situation. More recently, the Army Research Laboratory (ARL) Robotics Collaborative Technology Alliance (RCTA) has made significant progress toward this goal, which is attributable to advances in technology. Specifically, these include perceptual advances such as improved laser detection and ranging (LADAR) and other sensory devices, perceptual algorithms usable by automatons to decipher this data, path planning algorithms, and improvements in networking amongst UGVs (Childers et al., 2011). The remainder of this section comprises a review of the literature of advances as they pertain to the RSTA mission. Environment Perception A UGV’s ability to perceive the environment takes on a different quality when the unit is automated versus tele-operated. For example, human factors considerations such as motion sickness prevention due to FOV characteristics (Coovert et al., 2010) are irrelevant in the context of automated agents. However, additional programming is required to allow the unit to perceive the information that it senses. UGVs typically use some combination of vision (i.e., digital camera), LADAR, and sonar sensors to gather environmental information. Sometimes this disparate information is aggregated into a single picture (e.g., Peynot & Kassir, 2010). Each of these sensor types has its own advantages. Vision and LADAR work well over moderate distances (around 200 m) in clear weather, and can be used for water and mud detection (both serious threats to UGVs; Rankin & Matthies, 2010; Rankin, Matthies, & Bellutta, 2011), but the quality of the sensory information is easily degraded by environmental factors such as precipitation, airborne particulate, and foliage. Sonar, despite its limited range, is capable of penetrating environmental debris (Yamauchi, 2010). As the UGV senses environment information, on-board algorithms create a perceptual picture, or model, of the environment using Simultaneous Localization and Mapping (SLAM) algorithms. It is worth noting that this information is both actionable by the UGV itself, and transmissible as reconnaissance information. LADAR is primarily used to create this model, and is usable in both rural (Zhou & Dai, 2010) and urban terrain (Whitty, Cossell, Dang, Guivant, & Katupitiya, 2010). In addition to mapping static terrain, LADAR data can be used to track moving objects such as people (Navarro-Serment, Mertz, & Hebert, 2010). Tactical Behavior UGVs may act on their environmental perceptions in a number of ways, of which navigation is perhaps the most common. In most contexts, because efficiency is the goal, shortest path algorithms are commonly used. Shakey, for example, used the A* algorithm, whereas many autonomous agents in both the UGV domain and cognitive modeling use Dijkstra’s algorithm or something similar. However, in military applications, UGVs may be targeted by the enemy. Therefore, tactical planning systems are required for self-preservation. One such system by Hussain, Vidaver, and Berliner (2005) enables a UGV to use tactical behaviors to evade mobile enemies and accomplish its goals. Specifically, the Advocates and Critics for Tactical Behaviors (ACTB) system generates lists of “advocates,” situational prompts for certain tactical behaviors, and selects from amongst them based on “critics,” reward plans based upon tactical goals (e.g., minimizing risk or maximizing speed). In addition to driving locomotion, the ACTB system includes logic for aiming cameras and could presumably adapt to control weapons systems. Challenges to Automation UGV barriers to automation are typically the result of hardware, rather than software, shortcomings. Specifically, unlike UAVs, UGVs sometimes operate in environments where compass and GPS instruments cannot be used for localization. In urban environments, environmental obstacles such as stairs pose serious threats to UGV mobility. Furthermore, because they travel on a surface rather than flying, UGVs must be able to physically traverse terrain, or at least recognize when traversing terrain would be difficult or impossible. Sensor Challenges GPS is often used by UGVs for localization, while a compass instrument is used to establish current facing. Maxwell et al. (2013) describe a scenario in which semi-autonomous unmanned air, sea, and land units defended a high value target attacked by an enemy which arrived by water and then proceeding over land. The unmanned systems were tasked with identifying and tracking the attacker. In this scenario, of the two UGVs, one performed acceptably while the other experienced difficulty navigating. Specifically, the UGV’s compass readings were disrupted by electromagnetic interference from that UGV’s range finder and perhaps also vibration associated with traversing outdoor terrain. UGVs operating in urban terrain will often need to traverse terrain such as stairs, and rely on LADAR for sensing the environment, so these concerns are particularly important. GPS units suffer from dilution of precision errors when UGV-satellite line of sight (LOS) is occluded. In urban environments, UGVs will often operate inside of buildings, or in areas where buildings are sufficiently high as to occlude LOS. Finally, accelerometers and gyroscopes used by UGVs for proprioceptive information are affected by environmental conditions, such as temperature (Durst & Goodin, 2012). In summary, compass, GPS, and proprioceptive information are not robust sources of data for localization. Environmental Challenges In addition to sensor issues, the environment poses challenges for UGVs. For example, mud and water threaten UGV mobility in rural as well as urban environments. Vision algorithms (e.g., Rankin & Matthies, 2010; Rankin, Matthies, & Bellutta, 2011, respectively) can detect these threats, but rely on cues that may not be available in urban environments, such as sky reflections off of the liquid’s surface, or information from multiple sensors (e.g., stereo, visible, and multi-wave IR) working together. Urban environments carry their own specific threats to UGVs. Research carried out as part of the Safe Operations of Unmanned systems for Reconnaissance in Complex Environments (SOURCE) Army Technology Objective (Grey, Karlsen, DiBerardino, Mottern, & Kott, 2012) identified a number of urban environment-specific threats to UGVs. However, most of these threats pertain to larger UGVs operating on roads. Comparatively little research has been conducted on UGVs operating inside structures, though the main challenges appear to be stairways. In its present incarnation, the proposed automated UGV system would not necessarily be required to traverse stairs. A Way Forward As UGVs are a developing technology, authors have propositioned numerous conceptualizations of how they might be used in future conflicts. Mills (2007) suggests that UGVs will likely support troops in combat operations. Mills suggests that future UGVs will be: Heavily armed and armored Automated, with manual control the exception rather than the rule Capable of functioning independently of GPS or way-points for pathfinding Equipped with excellent IFF capability and less than lethal options for force escalation Able to recognize obstacles (e.g., fences and jersey barriers) and hazards (e.g., mud and barbed wire) Quiet Lightweight Self-destructable or otherwise unusable by the enemy Reliable Cheap Modular – payloads are expected to evolve faster than the UGV itself Interoperable Capable of tactical behavior and quickly reacting to dynamic circumstances Mills (2007) presents a view of UGVs that is far more combat-oriented than traditional approaches to automation would allow, however. Contrary to this view, the role that we propose for UGVs in MOUT operations is less combat-oriented and more focused on information gathering. Lif, Jander, & Borgvall (2006) conducted an evaluation of UGVs employed for reconnaissance during MOUT training. Rather than focusing on the capabilities of the UGV, Lif et al. (2006) focus on the needs of the military unit and how the UGV meets or impedes those needs. For example, the authors sought to answer questions such as, “Does using the UGV necessitate changes in tactics, resource allocation, or unit organization?” The results of this study indicated that UGVs improve soldiers’ ability to gather reconnaissance, unit stealth, and the utility of information gathered (as leadership can view the images directly). However, using the UGV impeded forward movement – sending the UGV into a building and following with infantry is slower than simply advancing with infantry – and its use therefore requires a calculated tradeoff between safety and speed. Finally, the main human factors concern with the UGV was whether using the UGV is worth occupying the operator who could otherwise be fighting. Therefore, the UGV should at least incorporate a capability to “return to handler.” These needs form the basis for the system proposed in the remainder of this document. Neural Networks for UGV Reconnaissance While others (e.g., Mills, 2007) have envisioned armed UGVs patrolling alongside soldiers, here I propose an autonomous UGV, deployable in groups, capable of scouting ahead of infantry in urban terrain. In this section, I will discuss the current state of our modeling efforts, and potential methods for integrating other approaches to modeling with existing technology applied to machine cognition. This proposed system is inspired by neural networks of the hippocampus that are used to approximate navigation tasks and memory in animals and humans. A number of features make neural networks particularly attractive for this task. 1. Neural networks can be used as mechanisms for both encoding and retrieving information about the environment, allowing an agent to simultaneously direct itself and provide reconnaissance information using the same computations. 2. Neural networks in navigation are often used to model the behavior of animals that depend on proximal, rather than distal, cues in the environment. The models are capable of operating using sensory input from short range, highly reliable sources that have robust analogs in robot navigation. 3. Because they encode memories, neural networks do not rely on GPS for localization. Localization is achieved by examining the environment and comparing those features with the agent’s memory for the environment. Representing the Environment In neural network models of the hippocampus, the goal is to represent the environment using biologically plausible data structures. The most basic of these data structures are “place cells” (O’Keefe & Nadel, 1978) in the cornu ammonis (CA) area of the hippocampus, which fire as ensembles when the organism is in specific locations in the environment. These models are primarily concerned with CA layers 1 and 3 (CA1 and CA3; Levy, 1989; Levy, Colbert, & Desmond, 1990), with the CA1 layer containing the contextual cues required for localization, and the CA3 layer containing a network of interconnected cells forming a cognitive map (e.g., Mueller, Perelman, & Simpkins, 2013; Samsonovich & Ascoli, 2005). CA3-CA3 interconnections code for locations between which it is possible to move (see Figure 1). In order to navigate to a particular point, the location is chosen by activating the CA1 layer, which in turn activates its corresponding CA3 location. There are a number of mechanisms for pathfinding based on the CA3 representation, the most simple of which is perhaps a spreading activation followed by a “goal scene” (Mueller et al., 2013). Figure 1. The CA1 and CA3 matrices often used in hippocampus based neural networks. In the code (e.g., Mueller et al., 2013, adapted from code originally published by Samsonovich & Ascoli, 2005), place cell layers are structured as sets of matrices: an environment grid, a recency map, a clutter matrix, a CA1 matrix, and a CA3 layer comprised of four matrices – one for each cardinal direction. The cardinal direction matrices are necessary for experiential learning (see below). This matrix set can be applied to any environment of arbitrary size but can be scaled. That is to say that, in the current configuration, all matrices are 25 * 25 cells, but this could be adapted for higher or lower resolution as necessary. The agent “learns” the environment through exploration. The cardinal direction matrices represent known adjacency and accessibility amongst cells. As the agent explores the environment, these matrices (initially all set to zero) receive positive values where the agent can move between two cells. From the standpoint of biological plausibility, the associations here represent the strengthening of CA3-CA3 connections through exploration. For example, if the agent is in cell [3, 22] and moves north, then the north matrix for cell [3, 22] receives a value of 1. These matrices may or may not be treated as symmetrical (i.e., in such a scenario, the south matrix for cell [3, 23] would also receive a positive value as the agent had just come from that location). While currently the algorithm requires movement between the cells, it could easily be modified so that mere perception is sufficient to build these associations. Through this mechanism, the agent builds up a picture of where it (or other agents) can move in the environment. Purposive Navigation Movement throughout the space can be directed using a search algorithm. One such algorithm, novelty-seeking, relies upon the aforementioned recency map to push the agent to new locations. However, if goal-directed movement is desired, the agent uses interplay between the matrices comprising the CA3-CA3 structure and the data structure representing the CA1 layer. In the model used by Mueller et al. (2013), the CA1 layer was simply a matrix containing goal locations. In other models (e.g., Gorchetchnikov & Hasselmo, 2002; Koene, Gorchetchnikov, Cannon, & Hasselmo, 2003), the CA1 structure contains all features associated with that particular context or location. In these models, the agent selects or prioritizes goal locations based upon desired features, activating the CA1 locations that contain those features. One mechanism for directing navigation to goals is spreading activation, or “goal scent.” Once the agent has selected goals, and their corresponding CA1 cells are activated, the CA1 locations in turn activate corresponding CA3 locations. From those CA3 locations, activation spreads throughout the CA3 network through CA3-CA3 interconnections, coded for by the cardinal direction matrices. Spreading activation through grids is a biologically plausible mechanism that is also used by existing navigation algorithms (cf. Whitty et al., 2010), but it is important to note that the current neural network approach differs from these algorithms in a number of ways. First, existing algorithms use computationally expensive optimization whereas spreading activation is efficient. Second, the data structures in the aforementioned network, specifically those comprising the CA3 layer, allow for asymmetries that a robot might encounter in an urban environment; for example, an environment could contain a configuration where a UGV could safely traverse through a hole in a wall in one direction, but could not return through that hole due to height differences or debris. Third, the activation map created using the spreading activation algorithm, combined with the recency map, allows the agent to use heuristics for navigation such as hierarchical strategies for route planning. Finally, the present mechanism slots most or all of the information required for purposive navigation, such as the environment matrices, the goal locations, and the features in every location, into one list data structure. Since most of these matrices overlap, they are conceptually very easy to use. Regarding this final point, I have created a system (Perelman, 2013) whereby environmental features, contained within an array, can be indexed either alone or in combination, to feed the CA1 goal layer. In short, the array acts as a filter whereby goal sites are selected on the basis of the features they contain. This information can be used either for localization or purposive navigation. One example of an environment that necessitates this is navigation through a shopping mall. If a UGV wants to determine its current location, and has learned the entire layout of a mall including all of the stores, it can localize by comparing the combination of stores (i.e., features) in its current location with all store combinations in the mall. The system is highly versatile, as anything salient and perceptible can be coded as a feature. Encoding the Environment Many biologically inspired neural network models are used to understand animals, most typically the rat. There are distinct differences between the types of environmental information that rats encode versus, for example, primates. While primates can rely upon distal visual cues (Rolls, Robertson, & Georges-François, 1997), rats rely mainly upon proximal cues. The purpose of this section is to discuss how rodent-specific models handle environment encoding, and to draw parallels between these mechanisms and currently technology. Rodent models are ideal for our proposed UGV because they are parsimonious – the same mechanism that allows environment encoding also provides tactical movement advantages. Note that while all of the assumptions made herein are based upon established theory, some interpretations of the published literature are not necessarily supported by consensus. Rats cannot localize using GPS, and they presumably do not rely on an internal compass. Rather, theories of rat navigation indicate place cell activation when a boundary is at a particular distance and allocentric direction from the rat, a process afforded by boundary vector cells (BVCs; Hartley, Burgess, Lever, Cacucci, & O’Keefe, 2000). Each BVC monitors its own receptive field, a Gaussian field with specific shape and direction, and the receptive field can change based on the shape of the environment. The rat localizes by comparing BVC activation in its current location with BVC activations it has experienced in the past (a process that is theoretically congruent with the feature-based approach to localization described by Perelman, 2013). Interestingly, longer receptive fields are in fact broader, whereas receptive fields for detecting more proximal stimuli are tighter, potentially indicating (1) overlapping levels of precision amongst all BVCs responsible for a single area relative to the rat’s position, and (2) a potentially hierarchical structure where boundaries are scanned for at various resolution and distances. These BVCs perform a function essentially identical to proximal sensors described by (Yamauchi, 2010). Typically, the sensors involved with modeling environments, using point clouds for example (Whitty et al., 2010), are very expensive because they are intended to model large environments. Using the BVC metaphor, the UGV can establish point clouds over much shorter distances as it explores the environment. Therefore, it requires lower resolution sensors. Sensors such as sonar provide sufficient sensory capability to model an environment provided that the agent remains close to boundaries that it can use to “anchor” itself in the environment, which is presumably one reason that rats tend to stick to walls when moving (Hoffman, Timberlake, Leffel, & Gont, 1999). Moving along walls also affords tactical advantages such as reduced visibility and probability of detection. There is no suggestion that an autonomous UGV should travel along walls exclusively, only that such a preference may be advantageous. Design Guidelines – A small UGV for reconnaissance in MOUT environments The proposed autonomous UGV must meet existing design recommendations and exceed tele-operated systems in terms of mission requirement fulfillment. In this section, I will discuss the proposed hardware and software, and how such a system is superior to existing systems. The present system was conceived mainly using the mission requirements described by Lif et al. (2006). Specifically, the system is designed to address problems with tele-operation. A reconnaissance UGV should, according to those mission requirements, Complete its mission without burdening a human operator Provide information parsimoniously to commanders in real time Return to its handler, and be capable of navigating to a specific location for extraction Complete its mission quickly to minimize the extent to which the system slows the rate of advance In addition, though Mills’ (2007) recommendations pertained specifically to armed UGVs, the proposed system was inspired by some of that design criteria. Therefore, the UGV should be capable of, Automation, with affordances for manual control Functioning independently of GPS or way-points for pathfinding Able to recognize and model obstacles and hazards Quiet movement and operation Reliable Cheap The proposed system addresses these design considerations in a number of ways. The proposed system is nearly entirely autonomous, though the UGV can use waypoints or even search directed by certain features (e.g., preferentially search a space for humans or weapons). Therefore, the system does not require constant monitoring or control, allowing all human operators to remain “in the fight.” Partial or complete maps can be inspected by the operator in real time, or at will. The system relies on input from highly reliable and cheap low resolution sensors. Because it depends upon context interpretation for localization, and accomplishes its mission through experiential learning, the system does not rely on GPS. Even given a partial mapping of the environment, the UGV can be directed, via waypoints, to specific locations at will. The system is also very inexpensive. Sonar sensors are cheap (i.e., < $100 USD) compared to LIDAR sensors. A single gimbaled sonar sensor is sufficient to sweep the environment. A second advantage of sonar sensors is that they are very small. Ideally, the unit would be roughly the size of a football and ruggedized (see Figure 2), so that it can be thrown through windows. The unit should be capable of transmitting without line of sight in urban environments to its operator. Since it is intended to be thrown, the unit should ideally be capable of operating while inverted, or at least have a mechanism to right itself. Since the unit is entirely autonomous, after it is thrown into the structure, it immediately begins collecting data. Beginning with a novelty-seeking random walk, the unit begins mapping the environment by logging boundaries and the locations of objects in the room. As the unit explores, it creates both a picture of the environment that is interpretable by the operator, and a means by which it can navigate to a waypoint for extraction. Figure 2. The UGV. Not drawn to scale, the UGV is expected to be small, roughly the size of a football. Interface In the proposed system, the operator would receive information from the unit by means of a handheld display that updates in real time as the unit explores the environment. The display can be monitored as the unit explores, or monitored when the operator has time to spare. Since it is constantly transmitting, if the unit is destroyed or otherwise disrupted, no information will be lost. And, since the unit does not require GPS or tele-operation to accomplish its mission, it can continue logging information even if it temporarily moves outside of the range necessary to transmit to the operator. The operator should interface with the unit through a rugged handheld device (see Figure 3). The UGV unit should have only one button: a power button used to prime the unit. This same button should not be used to power down the unit, as it could be activated inadvertently as the unit explores. Regarding the interface device, touchscreen is contraindicated, as the operator may need to use the unit while wearing gloves. A small directional pad and cursor should provide sufficient precision to set waypoints and move the field of view around the space as it is constructed, two buttons should provide zoom-in and zoom-out functionality (for the display of the plotted space), and four separate numbered buttons should allow a single display device to receive and display data from up to four separate units. Those same buttons, when pressed in combination with the power button on each individual unit, should allow the operator to sync a unit with a specific number on the display device. Figure 3. Handset for interfacing with the UGV. The plotted space itself should display as a map of the information collected by the unit. A point could indicate where the unit was activated, so that the user has a bearing on the map’s orientation. As per Figure 3, the unit’s current location within its grid space should be visible along with the environment as it is constructed via sonar. If, for some reason, the agent’s feed (via radio) is terminated, its current position should change to indicate this state (becoming a hollow circle, for example) and the map should remain as it was last constructed. This requires that a representation of the map be stored on both the unit and the display, and necessitates memory on both devices. Example of Use Scenario A squad of soldiers is tasked with investigating a house where a “be on the lookout” (BOLO) suspect is believed to be hiding. The suspect is wanted for his role in manufacturing IEDs used against coalition soldiers. Because of the possibility of armed enemy inside, and the possibility of traps inside the house, they are authorized to first investigate the structure using a number of the automated UGVs. Arriving outside the house, a two story structure, the UGV operator activates two units and throws one unit through a downstairs window and another through the window of the upper floor. Immediately, the units begin a novelty-seeking random walk to map out the environment, preferring to stick to walls as they model the rooms. They transmit the information in real time to the UGV operator’s handset. While they are exploring, the UGV operator is free to provide perimeter security as the units transmit to his handset. After a couple of minutes, the operator examines the handset and, finding that the UGVs have mapped out the space, shows the display to the patrol leader. Finding no humans within the structure, the patrol leader calls explosive ordinance disposal (EOD) professionals to the site. The structure floor plan generated using the UGV sonar data provides the EOD personnel with information that will guide their search using either a teleoperated robot or by hand. References Apostolopoulos, D. (2014). Gladiator. Carnegie Mellon – The Robotics Institute. Retrieved from http://www.ri.cmu.edu/research_project_detail.html?project_id=566&menu_id=261, February, 2014. Chen, J. Y. C. (2010). UAV-guided navigation for ground robot tele-operation in a military reconnaissance environment. Ergonomics, 53, 940-950. Childers, M. A., Bodt, B. A., & Camden, R. (2011). Assessing unmanned ground vehicle tactical behaviors performance. International Journal of Intelligent Control and Systems, 16, 5266. Coovert, M. D., Prewett, M. S., Saboe, K. N., & Johnson, R. C. (2010). Development of Principles for Multimodal Displays in Army Human-Robot Operations (No. ARL-CR651). University of Florida: Tampa DARPA. (2012). DARPA’s four-legged robots walk out for capabilities demonstration. DARPA News. Retrieved from http://www.darpa.mil/NewsEvents/Releases/2012/09/10.aspx, February, 2014. Durst, P. J. & Goodin, C. (2012). High fidelity modelling and simulation of inertial sensors commonly used by autonomous mobile robots. World Journal of Modelling and Simulation, 8, 172-184. Fong, T. & Thorpe, C. (2001). Vehicle Teleoperation Interfaces. Autonomous Robots, 11, 9-18. Gage, D. W. (1995). UGV History 101: A brief history of unmanned ground vehicle (UGV) development efforts. Unmanned Systems Magazine, 13. Gorchetchnikov, A; Hasselmo, ME (2002). A model of hippocampal circuitry mediating goaldriven navigation in a familiar environment. Neurocomputing, 44, 423-427. doi: 10.1016/S0925-2312(02)00395-8 Gray, J. P., Karlsen, R. E., DiBerardino, C., Mottern, E., & Kott, N. J. (2012). Challenges to autonomous navigation in complex urban terrain. In SPIE Defense, Security, and Sensing (pp. 83870B-83870B). International Society for Optics and Photonics. Ha, C. & Lee, D. (2013). Vision-based teleoperation of unmanned aerial and ground vehicles. In Proceedings of the 2013 IEEE International Conference on Robotics and Automation (ICRA), Karlsruhe, Germany, May 6-10. Hartley, T., Burgess, N., Lever, C., Cacucci, F., & O’Keefe, J. (2000). Modeling place fields in terms of the cortical inputs to the hippocampus. Hippocampus, 10, 369-379. Hightower, J. D., Smith, D. C., & Wiker, S. F. (1986). Development of remote presence technology for teleoperator systems. In Proceedings of the 14th Meeting of the UJNR/MFP, September. Hoffman, C. M., Timberlake, W., Leffel, J., & Gont, R. (1999). How is radial arm maze behavior in rats related to locomotor search tactics? Animal Learning & Behavior, 27, 426-444. Hussain, T. S., Vidaver, G., & Berliner, J. (2005, May). Advocates and critics for tactical behaviors in UGV navigation. In Defense and Security (pp. 255-266). International Society for Optics and Photonics. Irvin, C., Leo, S., & Kim, J. (2012). GunBot: Design concept of a semi-autonomous UGV with omni-directional mobility and auto-target tracking. In Proceedings of the Florida Conference on Recent Advances in Robotics, Boca Raton, Florida, May 2012. Koene, R.A., Gorchetchnikov, A., Cannon, R.C. and Hasselmo M.E. (2003) Modeling goaldirected spatial navigation in the rat based on physiological data from the hippocampal formation. Neural Networks, 16, 577-84. Kogut, G., Blackburn, M., & Everett, H. R. (2003). Using video sensor networks to command and control unmanned ground vehicles. SPACE AND NAVAL WARFARE SYSTEMS CENTER SAN DIEGO CA. Levy, W.B. (1989). A computational approach to hippocampal function. In R. D. Hawkins & G. H. Bower (Eds.), Computational models of learning in simple neural systems (pp. 243305). New York, NY: Academic Press. Levy, W. B., Colbert, C. M., & Desmond, N. L. (1990). Elemental adaptive processes of neurons and synapses: a statistical/computational perspective. In M. A. Gluck & D. E. Rumelhart (Eds.), Neuroscience and connectionist models (pp. 187-235). Nillsdale, NJ: Lawrence Erlbaum Assoc., Inc. Lif, P., Jander, H., & Borgvall, J. (2006). Tactical evaluation of unmanned ground vehicle during a MOUT exercise. In Proceedings of the Human Factors and Ergonomics Society 50th Annual Meeting, 2557-2561. Maxwell, P., Larkin, D., & Lowrance, C. (2013). Turning Remote-Controlled Military Systems into Autonomous Force Multipliers. Potentials, IEEE, 32(6), 39-43. Mellinger, D., Michael, N., & Kumar, V. (2012). Trajectory generation and control for precise aggressive maneuvers with quadrotors. The International Journal of Robotics Research, 31(5), 664-674. Mills, M. E. (2007). Challenges to the acceptance and proliferation of tactical UGVs. RUSI Defence Systems, 10(2), 28-30. Mueller, S. T., Perelman, B. S., & Simpkins, B. (2013). Pathfinding in the cognitive map: Network models of mechanisms for search and planning. Biologically Inspired Cognitive Architectures, 5, 94-111. Nardi, G. J. (2009). Autonomy, Unmanned Ground Vehicles, and the US Army: Preparing for the Future by Examining the Past. ARMY COMMAND AND GENERAL STAFF COLL FORT LEAVENWORTH KS SCHOOL OF ADVANCED MILITARY STUDIES. Navarro-Serment, L. E., Mertz, C., & Hebert, M. (2010). Pedestrian detection and tracking using three-dimensional LADAR data. The International Journal of Robotics, 29, 1516-1528. Nilsson, N. J. (1969). A mobile automaton: An application of artificial intelligence techniques. In Proceedings of the First International Joint Conference on Artificial Intelligence, Washington, D.C., May, pp. 509-520. Ögren, P., Svenmarck, P., Lif, P., Norberg, M., & Söderbäck, N. E. (2013). Design and implementation of a new teleoperation control mode for differential drive UGVs. Autonomous Robots, 1-9. O'Keefe, J., & Nadel, L. (1978). The hippocampus as a cognitive map (Vol. 3, pp. 483-484). Oxford: Clarendon Press. Perelman, B. S. (2013). A simple feature based approach to context in neurocomputational models of navigation. Unpublished manuscript. Perelman, B. S. & Mueller, S. T. (2013). A Neurocomputational Approach to Modeling Human Behavior in Simulated Unmanned Aerial Search Tasks. In Proceedings of the 2013 International Conference on Cognitive Modeling. Peynot, T. & Kassir, A. (2010). Laser-camera data discrepancies and reliable perception in outdoor robotics. In Proceedings of the 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems, Taipei, Taiwan, October 18-22. Rankin, A. L. & Matthies, L. H. (2010). Passive sensor evaluation for unmanned ground vehicle mud detection. Journal of Field Robotics, 27, 473-490. Rankin, A. L., Matthies, L. H., & Bellutta, P. (2011). Daytime water detection based on sky reflections. In Proceedings of the 2011 IEEE International Conference on Robotics and Automation, Shanghai, China, May 9-13. Rolls, E.T., Robertson, R.G., & Georges-François, P. (1997). Spatial view cells in the primate hippocampus. European Journal of Neuroscience, 9, 1789–94 Sam, R. & Hattab, A. (2014). Improving the control behavior of unmanned ground vehicle (UGV) using virtual windows. Unpublished manuscript. Retrieved from http://www.ammarhattab.com/, January 31, 2014. Samsonovich, A. V., & Ascoli, G. A. (2005). A simple neural network model of the hippocampus suggesting its pathfinding role in episodic memory retrieval. Learning & Memory, 12, 193-208. doi:10.1101/lm.85205 Walker, A. M., Miller, D. P., & Ling, C. (2013). Spatial orientation aware smartphones for teleoperated robot control in military environments: A usability experiment. In Proceedings of the Human Factors and Ergonomics Society 57th Annual Meeting. Whitty, M., Cossell, S., Dang, K. S., Guivant, J., & Katupitiya, J. (2010). Autonomous navigation using a real-time 3D point cloud. In Proceedings of the Australasian Conference on Robotics and Automation. Yamauchi, B. (2004). PackBot: A versatile platform for military robotics. In Proceedings of the SPIE, 5422, 229. Yamauchi, B. (2010). All-weather perception for man-portable robots using ultra-wideband radar. In 2010 IEEE International Conference on Robotics and Automation, Anchorage, AK, May 3-8. Zhou, B. & Dai, X. Z. (2010). Set-membership based real-time terrain modeling of mobile robots with a laser scanner. In Proceedings of the 2010 IEEE International Conference on Mechatronics and Automation, Xi’an, China, August 4-7.