Both sides previous revisionPrevious revisionNext revision | Previous revision |
ece4580:module_surveillance [2017/02/21 23:00] – [Week #3: Optimization-Based Data Association] pvela | ece4580:module_surveillance [2024/08/20 21:38] (current) – external edit 127.0.0.1 |
---|
- Target recognition | - Target recognition |
| |
| ===== Learning Modules ===== |
| ---------------------------- |
The sequence below introduces one aspect of surveillance systems at a time. They direct you to Matlab code that sometimes implements multiple steps at a time. It is recommended that you implement each one individually to get a sense for what role it plays in the entire system, rather than just copy/paste the whole system. | The sequence below introduces one aspect of surveillance systems at a time. They direct you to Matlab code that sometimes implements multiple steps at a time. It is recommended that you implement each one individually to get a sense for what role it plays in the entire system, rather than just copy/paste the whole system. |
| |
===== Module #1: A Basic Surveillance System ===== | |
--------------------------------------------------- | |
==== Week #1: Setup, Data, and Basic ==== | |
Explore the [[ECE4580:Module_Surveillance:MatlabVideos|datasets available]] and select a couple of videos to use as the starting point for processing, testing, and debugging the surveillance system to be created as part of this learning module. | |
| |
Meanwhile, check out this review of three [[http://www.eetimes.com/document.asp?doc_id=1275604|simple background modeling methods]]. You are not advised to use the code provided when implementing the same in future activities. There are more efficient ways to do the same without resorting to as many for loops, plus the implementation needs to be packaged up for use by the overall system. Plus, the activities below utilize existing Matlab libraries to the extent possible or desirable. | Module Set #1: A Basic (Foreground Detection-Based) Surveillance System |
| - [[ECE4580:Module_Surveillance:M1W1|Week #1]]: Setup, Data, and Basics |
| - [[ECE4580:Module_Surveillance:M1W2|Week #2]]: Foreground Object Extraction |
| - [[ECE4580:Module_Surveillance:M1W3|Week #3]]: Optimization-Based Data Association |
| - [[ECE4580:Module_Surveillance:M1W4|Week #4]]: Adding Temporal Dynamics via a Kalman Filter |
| |
As a first step, obtain the [[http://pvela.gatech.edu/classes/files/ECE4580/Module_Surveillance/survSystem.m|surveillance system class stub]] and also the [[http://pvela.gatech.edu/classes/files/ECE4580/Module_Surveillance/mainLoop.m|main execution script]]. Modify the main loop code stub so it loads a video you've chosen, loops through the frames, displays the image associated the each frame, and quits when that's done. Naturally this code won't do any surveillance, but it will setup the system to do so. | Module Set #2: Target Modelling and Re-Identification |
| - [[ECE4580:Module_Surveillance:M2W1|Week #1]]: Differentiating People |
| - [[ECE4580:Module_Surveillance:M2W2|Week #2]]: Testing the Person Model |
| - [[ECE4580:Module_Surveillance:M2W4|Week #3]]: Re-Identification in Action |
| - [[ECE4580:Module_Surveillance:M2W3|Week #4]]: Enhancing Tracking |
| |
As a second step, implement the basic background modeling detection step. Matlab has implementation of the [[https://www.mathworks.com/help/vision/ref/vision.foregrounddetector-class.html|mixtures of Gaussians adaptive background estimation algorithm]]. Just perform the estimation part and retrieve the binary foreground image. Modify the ''displayState'' function to display this output. When run, the system should display the source video, plus a binary image sequence associated to the detected objects. | Module #3: Merging and Splitting |
| - TBD |
| |
//Explore & Deliverables:// How well is the background modeled? You can identify how well it works by examining the quality of the binary image sequence. Does it capture the target objects only? Are there more false positives or false negatives than you like? What did you do to get the best result possible (to what parameters)? You should turn in at least one image pair showing the input frame, plus the output frame after foreground detection (or with the mask as noted in the code stub). | Module #4: Tracking vs Detection |
| - TBD. |
==== Week #2: Foreground Object Extraction ==== | |
With the background modelling step done, we have a means to identify regions of the image that do not conform to the expected scene. We will presume that these are all objects of interest to track, e.g. //targets//. Advance beyond the current foreground modelling step to include processing of the binary foreground image for extraction of detected targets and their bounding boxes. | |
[[https://www.mathworks.com/help/vision/examples/detecting-cars-using-gaussian-mixture-models.html|Extracting the foreground objects]] is really performing blob extraction from the binary foreground image. While there are Matlab examples that go beyond simply detecting the bounding box, this activity is simply asking to place in the surveillance system the foreground detection, the blob extraction, and the visualization with track number. | |
| |
Flesh out the ''overlayState'' function so that it can overlay the surveillance system output over the current image frame. In addition to keeping track of the total number of detected objects, like in the example, for each box plotted, plot the detection index associated to the box like [[http://www.mathworks.com/help/examples/vision_product/multiObjectTracking_02.png|here]]. You will also need to modify the functions that are invoked within the ''process'' member function. You may find that some will not be modified due to how the Matlab implementation works. | |
| |
Apply the algorithm to the video [[http://www.cvg.reading.ac.uk/PETS2009/a.html | S2.L1 Walking, View 001]] from the PETS 2009 People Tracking Dataset. It may be a collection of images, in which case some modification of the main loop will be needed. I have a [[https://github.gatech.edu/ivaMatlibs/readers|github repository]] of reader functions/classes, one of which allows for reading from a directory with images. It is called ''impathreader'' and has an interface somewhat similar to how Matlab deals with video. You are free to use it. Also apply to the couple of videos that you've selected (here, couple = exactly two). If one of them is the S2.L1 PETS 2009 video, then select another one to process. | |
| |
//Explore & Deliverables:// Turn in state overlays of the processed video for the specified PETS 2009 dataset, plus for your two additional videos chosen. Discuss the performance of the foreground detection process, as well as the detection labelling for the targets over time. | |
* Are there any challenges that pop up regarding the box estimation process? | |
* How did you select your parameters? | |
* What are morphological operations? | |
* Explain the relationship between the erode, dilate, open, and close operations. | |
* One thing that the Matlab examples do that I requested not to include is to include some kind of spatio-temporal tracking module to maintain the track identify of the objects. Based on pure detection and the processing natively done in Matlab, how often do you see the numerical identity given to a detected person change? What happens as they come in and out? cross people? How often? | |
Naturally, you should be turning in your code in the submission document. The best is to provide code snippets of the class member functions that you modified as part of this activity. If you submit the entire class, then highlight the modified functions so that we can see what was done. | |
| |
==== Week #3: Optimization-Based Data Association ==== | |
Performing detection does provide a means to identify objects of interest versus the prevailing background image. However if we are interested in maintaining the identity of the objects, additional processing and logic is required. The simplest scheme simply considers the spatio-temporal history of the targets and tries to link the current detected objects to the previously detected objects. This form of data association is known as the //assignment problem.// | |
| |
The classic algorithms for performing the assignment are the www.mathworks.com/matlabcentral/fileexchange/20652-hungarian-algorithm-for-linear-assignment-problems--v2-3-|Hungarian or Munkres' algorithm]] and | |
[[https://www.mathworks.com/matlabcentral/fileexchange/26836-lapjv-jonker-volgenant-algorithm-for-linear-assignment-problem-v3-0|Jonker-Volgenant Algorithm]]. Another [[http://goldberg.berkeley.edu/pubs/acc-2012-visual-tracking-final.pdf|paper]] a few years ago looked at a different version of the assignment problem based on "stable marriage" selection. All of them seek a solution to the assignment problem, but do so in different ways. | |
| |
Incorporate the assignment component into your surveillance system. The net result should be similar to this Matlab demo on [[http://www.mathworks.com/help/vision/examples/motion-based-multiple-object-tracking.html|multiple object tracking]], which also implements far more than what is described. You should strip the extras. | |
| |
Run your surveillance system on the two videos selected from earlier weeks, both with and without the assignment problem. Turn in how many people there really were, plus how many the two versions of the system claimed there were at the end of the video, or at the end of the processing. Comment on whether the identity tracking appeared to improve the system's ability to maintain the identity of the targets. | |
| |
| |
====Week #4: Adding Temporal Dynamics via a Kalman Filter==== | |
We can do a better job handling things like occlusions, as well as improve the data association, by adding in a temporal filter. A simple, powerful method is to recursively estimate and correct the moving target dynamics through a Kalman filter. Matlab has a page on how to do so for | |
[[https://www.mathworks.com/help/vision/examples/using-kalman-filter-for-object-tracking.html|object tracking]] along with a more general [[https://www.mathworks.com/discovery/kalman-filter.html | intro page]] to additional documentation. | |
| |
* If you are interested in the details of the Kalman filter and their connection to code, this Matlab file exchange [[https://www.mathworks.com/matlabcentral/fileexchange/5377-learning-the-kalman-filter|code tutorial]] may be of assistance. | |
| |
===== Module #2: Going Further ===== | |
------------------------------------ | |
| |
__** Week #1: Differentiating People **__ | |
| |
Gaussian mixture model for targets using [[http://www.mathworks.com/matlabcentral/fileexchange/26184-em-algorithm-for-gaussian-mixture-model|Expectation Maximization algorithm]]. | |
| |
__** Week #2: Appearance-Based Data Association **__ | |
| |
Comparing models. | |
| |
__** Week #3: Re-Identification **__ | |
| |
Using comparison to know when same person has re-entered. | |
| |
===== Module #3: Merging and Splitting ===== | |
-------------------------- | |
| |
TBD. | |
| |
===== Additional Information ===== | ===== Additional Information ===== |