Virtual Vision:

advertisement
Virtual Vision:
A Simulation Framework for
Camera Sensor Networks Research
Demetri Terzopoulos
UCLA
2007 PhD Thesis of Faisal Qureshi
University of Toronto
Publications:
•
•
•
•
•
2008 Proceedings of the IEEE
2008 11th Communications and Networking Simulation Symposium (CNS)
2008 ACM Symposium on Virtual Reality Software and Technology (VRST)
2007 First IEEE/ACM Intl. Conf. on Distributed Smart Cameras (ICDCS)
2007 IEEE/ACM Intl. Conf. on Distributed Computing in Sensor Systems
•
•
•
•
•
2007 IEEE Conf. on Computer Vision and Pattern Recognition (CVPR)
2006 ACM Intl. Wksp. on Distributed Smart Cameras (DSC)
2006 ACM Multimedia Systems Journal
2005 ACM Wksp. on Video Surveillance and Sensor Networks (VSSN)
2005 IEEE Intl. Wksp. on Visual Surveillance (VS-PETS)
(DCOSS)
1
Camera Networks and Smart Cameras
•
Visual surveillance is becoming ubiquitous
– London has roughly 4,000,000 cameras
•
Effective visual coverage of large spaces
require multi-camera systems
•
Operator monitoring is infeasible for large
networks
•
Need networks of smart cameras capable of autonomous operation
•
Smart cameras = visual sensor nodes
– Local on-board processing
– Communication with neighbors
Difficulty of Doing Large-Scale Visual Sensor Networks
Research
• Deploying large-scale camera networks in extensive public spaces
for research purposes:
– Hardware related technical challenges
– Privacy and legal issues
– Prohibitive cost
• Infeasible for most computer vision researchers
2
Virtual Camera Sensor Networks
Virtual Vision
Visually and behaviorally realistic simulators for designing and
evaluating machine vision systems
Camera Model:
Virtual
Camera
Environment Models:
geometry, texture,
illumination
Pedestrian Models:
appearance, movement,
behavior
Reality Emulator
Virtual Penn Station
pan, tilt, zoom,
camera jitter,
color response,
lens distortions,
etc.
High-Level Control
camera control,
assignment,
handover,
etc.
Virtual
Video
Machine Vision
tracking,
recognition,
etc.
3
Virtual Vision for Camera Networks Research
•
Emulates the characteristics of a physical vision system
•
Flexibility during system design and evolution
•
Readily available ground truth
•
Offline operation and testing
•
No legal impediments
•
No special hardware
•
Repeatability
•
Low cost
QuickTime™ and a
decompressor
are needed to see this picture.
Environmental Model of
Original Penn Station in NYC
[Shao & Terzopoulos, 2005]
Platforms
Concourses
Main
MainWaiting
WaitingRoom
Room
Concourses
Arcade
Arcade
Platforms
4
Architecture of the
Autonomous Pedestrian Model
Environment
& Interaction
Cognition
Behavior
Motion
Geometry
Pedestrian Simulation
5
Pedestrian Simulation
Human Activity in the Station
6
Virtual Vision
• Testbed for multi-camera sensor networks
Computer Vision Emulation Using Synthetic Video
• Pedestrian detection
– Background subtraction using a learned background model
QuickTime™ and a
decompressor
are needed to see this picture.
7
Low-Level Computer Vision
Emulation Using Synthetic Video
•
Pedestrian tracker
– Appearance based
Low-Level Computer Vision Emulation
Using Synthetic Video
• Appearance based pedestrian tracker
– [Swain & Ballard 91]
– Zoom invariant
Video frame
Signature
Backprojected image:
image
High intensity suggests
presence of target
8
Anatomy of a Camera Node
Decision
Decision logic
logic
Communication
Communication subsystem:
subsystem:
message
message passing
passing to
to
neighboring
neighboring nodes
nodes
Tracking
Task
Task “relevance”
“relevance” computation
computation
framework
framework
region of
interest
failure
reacquire
Vision
Vision routines:
routines: pedestrian
pedestrian tracking
tracking
Lost
Free
state management
timeout
y
Searching
timeout
left right
zoom
z
x
Image
Image driven
driven reactive
reactive behaviors:
behaviors:
fixation
fixation and
and zooming
zooming PD
PD controllers
controllers
up down
active pan-tilt-zoom camera
Virtual Vision:
A Tool for Visual Sensor Network Research
High-Level Camera Control
Visual Sensing
Visual Sensing
Synthetic Video
Real Video
Virtual Camera Network
Physical Camera Network
Synthetic World (Reality Emulator)
Real World
Virtual Vision
Real Vision
9
Camera Sensor Network #1
Active PTZ Camera Scheduling
Scheduling Active PTZ Cameras
• Number of Pedestrians
>
Number of active cameras
• Task active cameras to observe pedestrians in the scene
Active PTZ camera
Passive wide-FOV camera
10
Scheduling Active PTZ Cameras
• Scheduling problem
– Given n PTZ cameras, and m pedestrians, persistently observe
every pedestrian using one PTZ camera at a time
• Goal
– Observe as many pedestrians for as long as possible
Virtual Active Camera Scheduling Setup
• Passive wide-FOV cameras
– Calibrated
– Pedestrian localization through
triangulation
• Active PTZ cameras
– Un-calibrated
– Learn a coarse mapping between 3D
locations and internal pan-tilt settings
• Reliable pedestrian identification in
different cameras via appearance
based signatures
Calibrated passive cameras
at the four corners of
the waiting room in
the virtual train station
11
Distinguishing Features of Our Active Camera
Scheduling Strategy
•
Previous attempts
–
Costello et al., 2004
Costello et al., 2006
Lim et al., 2006
Huang and Trivedi, 2003
•
Exit times are not known a priori
•
Recording duration is not known a priori
•
Allow multiple observations of the same pedestrian
•
Multiple pan/tilt/zoom cameras
•
Tracking failures due to occlusions
Active Camera Scheduling Strategy
• Camera assignment via weighted round robin
• First come, first served tie breaking
• Multiple observation
• Preemption
Close-up snapshots captured
by active PTZ cameras
12
Active Camera Scheduling Results
Preemption (P)
No preemption (NP)
Single observation (SO)
Multiple observations (MO)
Single-class model (SC)
Multi-class model (MC)
Multiple observations, multi-class,
preemption scheduler outperforms other
variants
Up to 4 Cameras; 10, 20 Pedestrians
Camera Sensor Network #2
Active PTZ Camera Assignment and Grouping
13
Smart Camera Network
•
Cameras self-organize to observe pedestrians during their presence in the
region
– Smart, PTZ active cameras
– Un-calibrated
– Ad hoc deployment
– Camera additions & removal
– Handle camera failures
– Deal with imperfect communication
•
Previous work
– Simmons et al., 2006
Mallet, 2006
Park et al., 2006
Kulkarni et al., 2005
Javed et al., 2000;
Devarajan and Radke, 2006
Vision and Goal
• Ad hoc deployment
14
Vision and Goal
• Cameras work towards common sensing goals
Task assignment
Vision and Goal
• Cameras work towards common sensing goals
Group formation
15
Vision and Goal
• Cameras work towards common sensing goals
Group evolution
Vision and Goal
• Cameras work towards multiple common sensing goals
Group 1
Group 2
16
Vision and Goal
• Cameras work towards multiple common sensing goals
Camera failure
Vision and Goal
• Cameras work towards multiple common sensing goals
Task assignment
17
Vision and Goal
• Cameras work towards multiple common sensing goals
Conflict resolution
Camera Grouping and Reassignment
18
Active PTZ Camera Assignment and
Grouping
• Camera selection, grouping, and handoff via an auction model
– Announcement/Bidding/Selection
– ContractNet
• Smith, 1983
• Conflict resolution within a Constraint Satisfaction Problem
framework
– Partially distributed
• A Camera can only perform a single task at any given time
Camera Grouping: Announcement
•
Start with a single camera that is tasked to observe a person
Leader
19
Camera Grouping: Announcement
•
Seeks out other cameras in the vicinity to form a group to help
it with the observation task
Leader
Camera Grouping: Bidding
•
One or more cameras that receive the task announcement respond
with their relevance values
Never Received
the Message
Leader
Response: high
relevance
Response:
low relevance
Did not
respond
20
Camera Grouping: Bidding
•
One or more cameras that receive the task announcement respond
with their relevance values
•
Relevance encodes how successful a camera will be at an
observation task
Never Received
the Message
Leader
Response: high
relevance
Response:
low relevance
Did not
respond
Camera Grouping: Selection
•
After the leader gets relevance messages from neighboring
cameras, it selects suitable cameras to join the group
Join
Group
Leader
21
Conflict Detection
•
A conflict is detected when multiple tasks require the same
camera node to proceed successfully
Conflict Detection
•
A Red group member receives a recruit query from Green group
22
Conflict Detection
•
A Red group member receives a recruit query from Green group
Conflict Detection
•
A Red group member receives a recruit query from Green group
Conflict!
23
Conflict Detection
•
Nodes belonging to both groups send information to one of the
leaders
•
Leader selection
Centralization
Conflict Detection
•
The resulting node (camera) assignment is sent to the individual
nodes
Solution
24
Conflict Detection
•
The resulting node (camera) assignment is sent to the individual
nodes
Persistent Observation:
Virtual Active PTZ Camera Assignment and Grouping
25
Network Model
•
It does not require camera calibration or network calibration
– Can take advantage of calibration information, if available
– Ad hoc deployment
•
Camera grouping is a strictly local negotiation
– Typically camera groups are spatially local arrangements
•
Camera groups are dynamic arrangements
•
Camera handoffs occur naturally during negotiations
•
It can gracefully handle node and message failures
Network Model
• Camera nodes can be added or removed during the lifetime of a
task
• Even assuming perfect sensing, the proposed model can still fail if
– a significant number of messages are lost
– catastrophic node failure
– group evolution can’t keep up with a fast-changing observation task
• Scalability
– Small group sizes
– Conflict resolution is viable as long as the number of relevant sensors
for each task remains low ( < 10 )
– Optimal sensor assignment
26
Future Work
• Camera sensor network
– Consistent labeling of pedestrians
– Who, What, Where, When?
– Cognitive modeling
• Physical sensor networks
– Environment monitoring, urban sensing, intelligent environments, …
• Bigger, better reality emulators
– High-fidelity synthetic video
– An entire city …
Virtual LA
Urban Simulation Lab
Architecture Dept
UCLA
27
Virtual LA
• Building level of detail
Virtual LA
• Interior spaces
28
Strategy Looking Forward
• Collaborative engagement of the CV and CG communities
– Open source framework to develop and test state-of-the-art
vision algorithms and systems
– Modular www-accessible environment
• e.g., SRI’s Open Agent Architecture (OAA)
– “Any-world” emulator
– More realism (“The Matrix”)
Thank You !
29
Download