Intelligent and NoiseRobust Interfaces for MEMS Acoustic Sensors: Smart Microphone Cent er for Auditory and Acoust ic Research Inst it ut e f or Syst ems Research Universit y of Maryland UNIVERSITY OF MARYLAND Electrical and Computer Engineering & Psychology Departments Baras, Horiuchi, Krishnaprasad, Moss, Shamma THE JOHNS HOPKINS UNIVERSITY Electrical and Computer Engineering Department Andreou, Cauwenberghs, Etienne-Cummings UNIVERSITY OF SIDNEY Electrical Engineering Department van Schaik SIGNAL SYSTEMS CORPORATION Riddle, Murray COLLABORATIONS Institute for Neuroinformatics, ETH Army Research Laboratory Cent er for Auditory and Acoust ic Research Inst it ut e f or Syst ems Research Universit y of Maryland PROJECT GOALS AND MISSION Overall Mission Formulate, design, and implement signal processing systems and technology that can adapt, control, and utilize the noisy MEMS sensor signals Specific Approach Focus Area I Characterize and integrate M EMS sensors with aVLSI circuits that detect and receive the signals. Cent er for Auditory and Acoust ic Research Focus Area II Noise control in M EMS sensor arrays through design and fabrication of versatile analog VLSI M EMS interface and associated feature extraction and analysis stages Focus Area III Embedding and demonstrating the functional capability of the integrated M EMS/VLSI sensor and signal processing arrays in a moving robotic vehicle. Inst it ut e f or Syst ems Research Universit y of Maryland Specific Objectives Objective 1 Formulate strategies for interfacing with acoustic MEMS Objective 2 Develop and implement wind-noise and platform-noise reduction algorithms in VLSI Objective 3 Implement VLSI co chlear frequency analysis 2000 log f 1000 log u 500 log f log f u log f 250 125 100 log f Objective 4 Design and fabricate feature extraction algorithms 200 300 400 500 Time (ms) 600 700 800 900 1000 NSL TOOLBOX Cortical Decomposition of Sound QuickTime™ and a Photo - JPEG decompressor are needed to see this picture. Objective 5 Feature synthes is and recognition Objective 6 Technology transf er and demonstrations Cent er for Auditory and Acoust ic Research Inst it ut e f or Syst ems Research Universit y of Maryland Acoustic and Ultrasonic Transducers (AGA, REC) Prototype and Evaluate Various Types of MEMS Microphones/Speakers Custom Microphones/Speakers Cent er for Auditory and Acoust ic Research Commercial Transducers Inst it ut e f or Syst ems Research Universit y of Maryland Integrated MEMS Acoustic and Ultrasonic Arrays (AGA, REC) Prototype and Evaluate Various Types of MEMS Microphone/Speaker Arrays 2D Piezo Arrays Ceramics and Polymers Cent er for Auditory and Acoust ic Research 2D MEMS Arrays Capacitive Micro-Membranes Inst it ut e f or Syst ems Research Universit y of Maryland Vision: • A small, low power microphone interface for acoustic sensors that reduces turbulence and vibration induced noise on military platforms such as battlefield robotics Primary Microphone Port Polyurethane foam windscreen 1/2” Secondary Sensor Port for Wind Sensing Mounting plate Connector Microphones Preamps aVLSI noise reduction circuitry 1” Cent er for Auditory and Acoust ic Research Inst it ut e f or Syst ems Research Universit y of Maryland MEMS/VLSI Integration and Prototyping (REC,AGA) Develop Integrated Processing Electronics for Transducers Integrated Transducers and Electronics Cent er for Auditory and Acoust ic Research Transduction Electronics Inst it ut e f or Syst ems Research Universit y of Maryland • Approach: – Utilize multi-channel adaptive filtering modules based on aVLSI biomimetic technology • Analog filter banks with Independent Component Analysis (ICA) and Least Mean Squares (LMS) adaptation – Incorporate low noise preamps; acoustic and vibration sensors – Develop specification and prototypes – Demonstrate in acoustic duct and installed on unmanned land vehicle Acoustic Sensors Wind Noise Sensor Multi-Resolution Low noise acoustic signals Adaptive Filter Noise and Vibration Sensors Cent er for Auditory and Acoust ic Research Inst it ut e f or Syst ems Research Universit y of Maryland Cochlear Frequency Analysis • We will design a new silicon cochlea with the following features: – – – Increased robustness due to 2D design Integrated Inner Hair Cell Model Reproducible settings of the parameters Magnitude Response Phase Response Cent er for Auditory and Acoust ic Research Inst it ut e f or Syst ems Research Universit y of Maryland Stochastic Resonance Exploit stochastic resonance (noise-induced enhancement of spectral power amplification SPA) in conjunction with auditory-inspired (e.g. cochlear) sensor signal processing architectures External World Band-pass filter + Threshold detector Controlled noise generator circuit MEMS heater Cent er for Auditory and Acoust ic Research K Inst it ut e f or Syst ems Research Universit y of Maryland Adaptive Filtering and Blind Source Separation (GC, AGA) Static and Dynamic ICA (Independent Component Analysis) Adaptive Noise and Wind Cancellation without Need for Isolated Reference Dynamic ICA Array Processor Cent er for Auditory and Acoust ic Research Adaptive Cell Inst it ut e f or Syst ems Research Universit y of Maryland Filtering Auditory-Based Sound Separation X1 X2 X1 X2 A S1 S2 De-noising 1 H1 v v 1 1 H2 v v 1 2 w w ICA1 2 2 w w 2 ICA2 2 1 1 1 2 1 2 Grouping 2 And X1 X2 . . Competitive . . Learning . . . X1 X2 v v HN . N 1 N w w ICAN 2 w w 1 2 w sö y wx N 1 N 2 Early Auditory Models lo g f u lo g f S1 l gf u lo g f ICA lo g f u lo g f S2 Cent er for Auditory and Acoust ic Research l gf u lo g f Cochlea Hair cell Lateral Inhibition Inst it ut e f or Syst ems Research Universit y of Maryland Cochlear Feature Extraction (AGA, GC) Neuromorphic implementation with asynchronous “spiking” outputs BM in BM out auditory nerve hair cells 15 channels cochlea AM Energy TZC (single channel) TZC FM BM out Time Cent er for Auditory and Acoust ic Research Inst it ut e f or Syst ems Research Universit y of Maryland Early Auditory Processing Stages Analysis Transduction Reduction Cochlea r f ilt er s Hair ce lls Lat er al i nhib i t io n 2000 log f 1000 log u 500 log f log f u log f 250 125 100 log f eardrum / r i cochlea t basilar membrane filt ers a w a hair cell st a ges lat era l inhibit ory net work 200 300 400 500 Time (ms) 600 700 800 900 1000 Audit or y Spec t rogr am y / 4000 4000 250 250 Time ( ms ) Cent er for Auditory and Acoust ic Research 500 Time ( ms ) 500 Inst it ut e f or Syst ems Research Universit y of Maryland Multiresolution Preprocessor: Auditory Filtering Multiresolution cortical filter outputs Fast Rate Fine Scale Slow Rate Fine Scale Upward Moving Fast Rate Coarse Scale Slow Rate Coarse Scale Slow Rate Fine Scale Fast Rate Fine Scale Downward Moving Slow Rate Coarse Scale Fast Rate Coarse Scale • The second filter models the multiscale processing of the signal that happens in the auditory cortex • A Ripple Analysis Model, using a ripple filter bank, acts on the output of the inner ear to give multiscale spectra of the sound timbre (Wavelet Transform) Cent er for Auditory and Acoust ic Research Inst it ut e f or Syst ems Research Universit y of Maryland Auditory Processing of Vehicle Signals: Cortex 70 60 scale 50 40 30 20 10 0 20 40 60 80 100 120 Auditory frequency Example of multi-resolution representation from cortical module Cent er for Auditory and Acoust ic Research Inst it ut e f or Syst ems Research Universit y of Maryland Wavelet TSVQ Applied to Acoustic Vehicle Classification • Objective: a prototype vehicle acoustic signal classification system with low classification error and short search time • Biologically motivated feature extraction models: cochlear filter banks and A1-cortical wavelet transform • Vector Quantization (VQ) based classification algorithm. Including learning VQ (LVQ) and tree structured VQ (TSVQ) Acoustic Recording Preprocessing Peripheral auditory processing model Cortical processing model Feature extraction system VQ based Classification Algorithm Classification Result Algorithm Flowchart Cent er for Auditory and Acoust ic Research Inst it ut e f or Syst ems Research Universit y of Maryland Acoustic Transient Time-Frequency Analysis (GC, w/ APL) Models “Ripple” Dynamics of Cells Recorded in Auditory Cortex (Shamma) audio in Segmenter “Shelf” (class 10) “Tub” (class 11) Continuous Wavelet Filterbank 32 (freq) 32 (freq) X 64 (time) ATP Time-Frequency Template Correlator 16 (templ.) ... 16 (12 used) Digital Postprocessor Cent er for Auditory and Acoust ic Research Inst it ut e f or Syst ems Research Universit y of Maryland BINAURAL LOCALIZATION NM/NL (ITD processing) Cochlear filters ICc NA ICx VLVp (ILD processing) ABL Cochlear filters NM/NL (ITD processing) ICc NA ICx VLVp (ILD processing) Cent er for Auditory and Acoust ic Research Inst it ut e f or Syst ems Research Universit y of Maryland Stereausis: A Biologically Plausible Binaural Network. A binaural sound localization system will be developed using 2 silicon cochleas and an aVLSI implementation for ILD and ITD detection. Cent er for Auditory and Acoust ic Research Inst it ut e f or Syst ems Research Universit y of Maryland Cent er for Auditory and Acoust ic Research Inst it ut e f or Syst ems Research Universit y of Maryland Coc hlear filt ers A hair cell st ages lat eral Temporal inhibit ion Sharpening Coincidence C ij = z(t ; i).z(t; x=1 log f Mat rix j) Z(t; i) log f x=128 log f B r(t; x) .125 y(t; x) Z(t; j) z(t ; x) .125 .125 .2 5 .5 .25 1 1 1 2 2 2 .25 .5 200 .5 200 C Cent er for Auditory and Acoust ic Research Inst it ut e f or Syst ems Research Universit y of Maryland BATS Cent er for Auditory and Acoust ic Research Inst it ut e f or Syst ems Research Universit y of Maryland Steerable Range Gauging and Echolocation (REC) Develop Ranging Signal Processing Algorithms Cent er for Auditory and Acoust ic Research Inst it ut e f or Syst ems Research Universit y of Maryland Massively Parallel Kernel Learning “Machines” (GC) Kernel “machines” subsume LVQ, RBF and SVM classifiers Locally adaptive, distributed memory Scalable and modular Factor 100-10,000 more efficient than CPU or DSP Parallel vector quantizer Cent er for Auditory and Acoust ic Research Inst it ut e f or Syst ems Research Universit y of Maryland Integrate-and-Fire Address-Event VLSI Neural Networks (GC & AGA) T itle: c hip2.eps Creator: Xcircuit v2.0 Preview: T his EPS pic ture was not s aved with a preview i nc luded i n i t. Comment: T his EPS pic ture wi ll pri nt to a Pos tScri pt printer, but not to other types of printers. Scalable, multi-chip architecture for “neural” computations Address-event routing circuit provides for arbitrary interconnection topologies Analog-valued synaptic weights are implemented by probabilistically transmitting address-events Integrate-and-Fire Array Address-Event Transceiver Chip Cent er for Auditory and Acoust ic Research Inst it ut e f or Syst ems Research Universit y of Maryland Address-Event Asynchronous Communication and Computation (AGA, REC, GC) The multi-chip modular, scalable approach to system integration Cent er for Auditory and Acoust ic Research Inst it ut e f or Syst ems Research Universit y of Maryland New design: An AER cochlea chip • Currently in fabrication • 128 output channels • Both for sonar and audio • New silicon process (0.35um minimum feature size versus 2.0um) •AER makes inter-chip communication possible. •AER allows manipulation of output such that projective fields are readily implemented. Cent er for Auditory and Acoust ic Research Inst it ut e f or Syst ems Research Universit y of Maryland Cent er for Auditory and Acoust ic Research Inst it ut e f or Syst ems Research Universit y of Maryland 1-D Address-Event Transceiver with Diode-Capacitors Integrators (AGA) Address Event (AE) transceiver circuit is a modular element for future multidimensional communication between neuromorphic chips. Input AE data is processed by the sigmoid function of the nonlinear diode-capacitor integrators . The data is retransmitted using the AE protocol with an arbitrated queueing communication system. 1-D Address-Event Transceiver Chip Cent er for Auditory and Acoust ic Research Inst it ut e f or Syst ems Research Universit y of Maryland • Impact: – Enable effective acoustic surveillance for Future Combat Systems – Increase by 20 dB the turbulence induced noise rejection of acoustic sensors relative to passive treatments of the same size, using active noise control – Increase by 20 dB the platform noise rejection of acoustic sensors over existing methods Demo III Experimental Unmanned Vehicle Cent er for Auditory and Acoust ic Research Inst it ut e f or Syst ems Research Universit y of Maryland The Robots Microphones Speakers Wireless camera Sonar sensors Touch sensors Cent er for Auditory and Acoust ic Research Inst it ut e f or Syst ems Research Universit y of Maryland