ece4580:module_recognition
Differences
This shows you the differences between two versions of the page.
Both sides previous revisionPrevious revisionNext revision | Previous revision | ||
ece4580:module_recognition [2017/01/24 05:10] – typos | ece4580:module_recognition [2024/08/20 21:38] (current) – external edit 127.0.0.1 | ||
---|---|---|---|
Line 1: | Line 1: | ||
====== Object Recognition ====== | ====== Object Recognition ====== | ||
+ | |||
+ | Object recognition is very similar to object classification, | ||
+ | |||
+ | So, here we study methods that deal with differentiating objects within a given class of objects (or so I believe ...). It may involve differentiation across different types of objects too. | ||
/* | /* | ||
Line 7: | Line 11: | ||
*/ | */ | ||
+ | ===== Module #1: Bag-of-Words ===== | ||
+ | Before people got into the auto-learning of feature spaces, it was common to try to hand craft a feature space, or come up with a mechanism for creating feature spaces that generalized. | ||
- | ===== Module | + | __**Week |
- | + | ||
- | Clustering | + | |
- Study [[https:// | - Study [[https:// | ||
- Download (or clone) the clustering skeleton code [[https:// | - Download (or clone) the clustering skeleton code [[https:// | ||
Line 19: | Line 23: | ||
- Comment on your different segmentation results. | - Comment on your different segmentation results. | ||
- | ----------------- | + | //Matlab Notes:// Matlab has several functions that can assist with the calculations so that you do not have to process the data in a for loops. |
- | ===== Module | + | __**Week |
- | Object Recognition | + | |
- Study the [[https:// | - Study the [[https:// | ||
Line 33: | Line 36: | ||
- Compute histogram of codewords using [[https:// | - Compute histogram of codewords using [[https:// | ||
- Use the rest of 50 images in both categories to test your implementation. | - Use the rest of 50 images in both categories to test your implementation. | ||
+ | - Report the accuracy and computation time with different k | ||
- | ----------------- | ||
- | ;#; | + | __**Week |
- | [[ECE4580:Modules|ECE4580 Learning Modules]] | + | Usually objects have different properties across the spatial scales, even though they may appear common at one given scale. |
- | ;#; | + | - Study [[https:// |
+ | - We will implement a simplified version of SPM based on your molude 2 | ||
+ | - First, for each traning image, divide it equally into a 2 × 2 spatial bin. | ||
+ | - Second, for each of the 4 bins, extract the SIFT features and compute the histograms of codewords as in module 2 | ||
+ | - Third, concatenate the 4 histogram vectors in a fixed order. (hint: the a vector has 4k dimension.) | ||
+ | - Forth, concatenate the vector you have in module 2 with this vector (both weighted by 0.5 before concatenated). | ||
+ | - Finally, use this 5k representation and re-run the training and testing again. | ||
+ | - Compare the results from module 3 and module 2. Explain what you observe. | ||
+ | __**Week #4: Base Feature Exploration**__\\ | ||
+ | At this point, you should have a good object recognition system in hand, capable of differentiating input objects through their //words//. To arrive at the system we employed the SIFT descriptor, which is but one such means to generate an identifier vector (one could even say a hash vector) from an image region. | ||
+ | - In modules 4, try to replace sift with other feature descriptor. For example, you may want to see how [[https:// | ||
+ | - Select one feature descriptor, rerun the training and testing, and show your comparison results and observations. | ||
+ | ------------------------- | ||
+ | ===== Module #2: Alternative Classifiers ===== | ||
+ | The vanilla Bag-of-Words algorithm utilizes the //k nearest neighbors// (kNN) algorithm to output the final decision. | ||
+ | Here we will explore support vector machines (SVM) as a means to perform the final decision. | ||
+ | |||
+ | |||
+ | |||
+ | __**Week #1: understanding SVM**__ \\ | ||
+ | - Go to [[https:// | ||
+ | - See square in the middle? Select a color by clicking on the " | ||
+ | - Here we would like to use SVM classifier to help us to train a model to predict if a coming image is a car or face by drawing a hyper-plane in feature space, so we no longer have to compare each new image to all the training image(k-nearest neighbor approach in module # | ||
+ | - Go to " | ||
+ | - Extract the compressed file, open your Matlab, and browse into the folder, say C:\libsvm | ||
+ | - Now type command >> mex -setup | ||
+ | - After you select your compiler for MEX-files, get into /matlab folder >> cd(' | ||
+ | - Compile it >> make | ||
+ | - Okay, now your should have SVM libraries in your computer! you can add path by >> addpath(' | ||
+ | - Now, we'd like to run some toy example from [[http:// | ||
+ | - >> labels = double(rand(10, | ||
+ | - >> data = rand(10,5); | ||
+ | - >> model = svmtrain(labels, | ||
+ | - Now you should be able to understand the basic usages. Please carefully read the README file in the folder and run the example. Output the result of accuracy_l and accuracy_p to demonstrate you've run the example yourself. It is very important your have correctly installed libSVM. We will use libSVM for the next task! | ||
+ | - (optional) anyone who is interested in SVM, I highly recommend [[https:// | ||
+ | |||
+ | __**Week #2: apply SVM to car and face dataset**__ \\ | ||
+ | - If you did the toy example in libSVM and understood commands correctly, now you are ready to apply this powerful library on our previous dataset - car and face!. | ||
+ | - In the toy example, data is actually a list of feature and it's true label. Now, we would like to use our bag-of-words features here. | ||
+ | - Generate bag-of-words feature for car and face as you did in your previous tasks (collect sift features, kmeans, and compute the histogram for each image. We use histogram as the feature for each image). | ||
+ | - Mimic the toy example, for each of the image, we set an label for car (and different label for face). Also for each of the image we have the histogram as its feature like you see in toy example. | ||
+ | - You have to generate two files, one for training and one for testing. | ||
+ | - Use the commands you learned last week, report your SVM accuracy. | ||
+ | |||
+ | |||
+ | __**Week #3: Kernel Trick**__ \\ | ||
+ | - So far we've applied SVM on our dataset to distinguish car and face. We are going to learn more about SVM to see how it works. | ||
+ | - If you played with the nice SVM GUI [[https:// | ||
+ | - Actually, one of the good properties of SVM is that you can apply different " | ||
+ | - Please read carefully about [[http:// | ||
+ | - Now, figure out how you can change different kernel in libSVM. Write down your answer. | ||
+ | - Here we provide an easier dataset [[https:// | ||
+ | - You can use the following command to visualize your data | ||
+ | - >> load mnist_49_3000; | ||
+ | - >> [d,n] = size(x); | ||
+ | - >> i = 1; % index of image to be visualized | ||
+ | - >> imagesc(reshape(x(:, | ||
+ | - Train SVM as you did before on this 4_9_3000 dataset with linear kernel. Compare the results with different kernels. | ||
+ | - Train SVM on your car and face dataset with linear kernel. Compare the results with different kernels. Which dataset has more improvements with different kernels. Why? | ||
+ | |||
+ | |||
+ | |||
+ | |||
+ | __**Week #4: Cross Validation**__ \\ | ||
+ | - Besides kernel, another important thing to apply SVM correctly is to select good hyper-parameters. | ||
+ | - If you checked SVM GUI [[https:// | ||
+ | - Please quickly review the SVM material again, and explain what is C parameter here? | ||
+ | - One way to select good hyper-parameter is to apply "cross validation" | ||
+ | - Please read carefully about [[https:// | ||
+ | - Now, apply cross validation by using 10% of your data for cross validation. Try it on mnist_49_3000. Test C = 10^-2, 10^-1, 1, 10^1, 10^2, which one gives you best performance? | ||
+ | - What is the difference between K-fold and LOO(leave one out)? | ||
+ | |||
+ | --------------- | ||
+ | ;#; | ||
+ | [[ECE4580: | ||
+ | ;#; |
ece4580/module_recognition.1485252631.txt.gz · Last modified: 2024/08/20 21:38 (external edit)