Mercurial > cortex
view org/ullman.org @ 379:f1b8727360fb
add images.
author | rlm |
---|---|
date | Wed, 10 Apr 2013 16:38:52 -0400 |
parents | |
children | 2d0afb231081 |
line wrap: on
line source
1 #+title: Ullman Literature Review2 #+author: Robert McIntyre3 #+email: rlm@mit.edu4 #+description: Review of some of the AI works of Professor Shimon Ullman.5 #+keywords: Shimon, Ullman, computer vision, artificial intelligence, literature review6 #+SETUPFILE: ../../aurellem/org/setup.org7 #+INCLUDE: ../../aurellem/org/level-0.org8 #+babel: :mkdirp yes :noweb yes :exports both11 * Ullman13 Actual code reuse!15 precision = fraction of retrieved instances that are relevant16 (true-postives/(true-positives+false-positives))18 recall = fraction of relevant instances that are retrieved19 (true-positives/total-in-class)21 cross-validation = train the model on two different sets to prevent22 overfitting.24 nifty, relevant, realistic ideas25 He doesn't confine himself to unplasaubile assumptions27 ** Our Reading29 *** 2002 Visual features of intermediate complexity and their use in classification34 ** Getting around the dumb "fixed training set" methods36 *** 2006 Learning to classify by ongoing feature selection38 Brings in the most informative features of a class, based on39 mutual information between that feature and all the examples40 encountered so far. To bound the running time, he uses only a41 fixed number of the most recent examples. He uses a replacement42 strategy to tell whether a new feature is better than one of the43 corrent features.45 *** 2009 Learning model complexity in an online environment47 Sort of like the heirichal baysean models of Tennanbaum, this48 system makes the model more and more complicated as it gets more49 and more training data. It does this by using two systems in50 parallell and then whenever the more complex one seems to be51 needed by the data, the less complex one is thrown out, and an52 even more complex model is initialized in its place.54 He uses a SVM with polynominal kernels of varying complexity. He55 gets good perfoemance on a handwriting classfication using a large56 range of training samples, since his model changes complexity57 depending on the number of training samples. The simpler models do58 better with few training points, and the more complex ones do59 better with many training points.61 The final model had intermediate complexity between published62 extremes.64 The more complex models must be able to be initialized efficiently65 from the less complex models which they replace!68 ** Non Parametric Models70 [[../images/viola-parzen-1.png]]71 [[../images/viola-parzen-2.png]]73 *** 2010 The chains model for detecting parts by their context75 Like the constelation method for rigid objects, but extended to76 non-rigid objects as well.78 Allows you to build a hand detector from a face detector. This is79 usefull because hands might be only a few pixels, and very80 ambiguous in an image, but if you are expecting them at the end of81 an arm, then they become easier to find.83 They make chains by using spatial proximity of features. That way,84 a hand can be idntified by chaining back from the head. If there85 is a good chain to the head, then it is more likely that there is86 a hand than if there isn't. Since there is some give in the87 proximity detection, the system can accomodate new poses that it88 has never seen before.90 Does not use any motion information.92 *** 2005 A Hierarchical Non-Parametric Method for Capturing Non-Rigid Deformations94 (relative dynamic programming [RDP])96 Goal is to match images, as in SIFT, but this time the images can97 be subject to non rigid transformations. They do this by finding98 small patches that look the same, then building up bigger99 patches. They get a tree of patches that describes each image, and100 find the edit distance between each tree. Editing operations101 involve a coherent shift of features, so they can accomodate local102 shifts of patches in any direction. They get some cool results103 over just straight correlation. Basically, they made an image104 comparor that is resistant to multiple independent deformations.106 !important small regions are treated the same as nonimportant107 small regions109 !no conception of shape111 quote:112 The dynamic programming procedure looks for an optimal113 transformation that aligns the patches of both images. This114 transformation is not a global transformation, but a composition115 of many local transformations of sub-patches at various sizes,116 performed one on top of the other.118 *** 2006 Satellite Features for the Classification of Visually Similar Classes120 Finds features that can distinguish subclasses of a class, by121 first finding a rigid set of anghor features that are common to122 both subclasses, then finding distinguishing features relative to123 those subfeatures. They keep things rigid because the satellite124 features don't have much information in and of themselves, and are125 only informative relative to other features.127 *** 2005 Learning a novel class from a single example by cross-generalization.129 Let's you use a vast visual experience to generate a classifier130 for a novel class by generating synthetic examples by replaceing131 features from the single example with features from similiar132 classes.134 quote: feature F is likely to be useful for class C if a similar135 feature F proved effective for a similar class C in the past.137 Allows you to trasfer the "gestalt" of a similiar class to a new138 class, by adapting all the features of the learned class that have139 correspondance to the new class.141 *** 2007 Semantic Hierarchies for Recognizing Objects and Parts143 Better learning of complex objects like faces by learning each144 piece (like nose, mouth, eye, etc) separately, then making sure145 that the features are in plausable positions.