view org/ullman.org @ 379:f1b8727360fb

add images.
author rlm
date Wed, 10 Apr 2013 16:38:52 -0400
parents
children 2d0afb231081
line wrap: on
line source
1 #+title: Ullman Literature Review
2 #+author: Robert McIntyre
3 #+email: rlm@mit.edu
4 #+description: Review of some of the AI works of Professor Shimon Ullman.
5 #+keywords: Shimon, Ullman, computer vision, artificial intelligence, literature review
6 #+SETUPFILE: ../../aurellem/org/setup.org
7 #+INCLUDE: ../../aurellem/org/level-0.org
8 #+babel: :mkdirp yes :noweb yes :exports both
11 * Ullman
13 Actual code reuse!
15 precision = fraction of retrieved instances that are relevant
16 (true-postives/(true-positives+false-positives))
18 recall = fraction of relevant instances that are retrieved
19 (true-positives/total-in-class)
21 cross-validation = train the model on two different sets to prevent
22 overfitting.
24 nifty, relevant, realistic ideas
25 He doesn't confine himself to unplasaubile assumptions
27 ** Our Reading
29 *** 2002 Visual features of intermediate complexity and their use in classification
34 ** Getting around the dumb "fixed training set" methods
36 *** 2006 Learning to classify by ongoing feature selection
38 Brings in the most informative features of a class, based on
39 mutual information between that feature and all the examples
40 encountered so far. To bound the running time, he uses only a
41 fixed number of the most recent examples. He uses a replacement
42 strategy to tell whether a new feature is better than one of the
43 corrent features.
45 *** 2009 Learning model complexity in an online environment
47 Sort of like the heirichal baysean models of Tennanbaum, this
48 system makes the model more and more complicated as it gets more
49 and more training data. It does this by using two systems in
50 parallell and then whenever the more complex one seems to be
51 needed by the data, the less complex one is thrown out, and an
52 even more complex model is initialized in its place.
54 He uses a SVM with polynominal kernels of varying complexity. He
55 gets good perfoemance on a handwriting classfication using a large
56 range of training samples, since his model changes complexity
57 depending on the number of training samples. The simpler models do
58 better with few training points, and the more complex ones do
59 better with many training points.
61 The final model had intermediate complexity between published
62 extremes.
64 The more complex models must be able to be initialized efficiently
65 from the less complex models which they replace!
68 ** Non Parametric Models
70 [[../images/viola-parzen-1.png]]
71 [[../images/viola-parzen-2.png]]
73 *** 2010 The chains model for detecting parts by their context
75 Like the constelation method for rigid objects, but extended to
76 non-rigid objects as well.
78 Allows you to build a hand detector from a face detector. This is
79 usefull because hands might be only a few pixels, and very
80 ambiguous in an image, but if you are expecting them at the end of
81 an arm, then they become easier to find.
83 They make chains by using spatial proximity of features. That way,
84 a hand can be idntified by chaining back from the head. If there
85 is a good chain to the head, then it is more likely that there is
86 a hand than if there isn't. Since there is some give in the
87 proximity detection, the system can accomodate new poses that it
88 has never seen before.
90 Does not use any motion information.
92 *** 2005 A Hierarchical Non-Parametric Method for Capturing Non-Rigid Deformations
94 (relative dynamic programming [RDP])
96 Goal is to match images, as in SIFT, but this time the images can
97 be subject to non rigid transformations. They do this by finding
98 small patches that look the same, then building up bigger
99 patches. They get a tree of patches that describes each image, and
100 find the edit distance between each tree. Editing operations
101 involve a coherent shift of features, so they can accomodate local
102 shifts of patches in any direction. They get some cool results
103 over just straight correlation. Basically, they made an image
104 comparor that is resistant to multiple independent deformations.
106 !important small regions are treated the same as nonimportant
107 small regions
109 !no conception of shape
111 quote:
112 The dynamic programming procedure looks for an optimal
113 transformation that aligns the patches of both images. This
114 transformation is not a global transformation, but a composition
115 of many local transformations of sub-patches at various sizes,
116 performed one on top of the other.
118 *** 2006 Satellite Features for the Classification of Visually Similar Classes
120 Finds features that can distinguish subclasses of a class, by
121 first finding a rigid set of anghor features that are common to
122 both subclasses, then finding distinguishing features relative to
123 those subfeatures. They keep things rigid because the satellite
124 features don't have much information in and of themselves, and are
125 only informative relative to other features.
127 *** 2005 Learning a novel class from a single example by cross-generalization.
129 Let's you use a vast visual experience to generate a classifier
130 for a novel class by generating synthetic examples by replaceing
131 features from the single example with features from similiar
132 classes.
134 quote: feature F is likely to be useful for class C if a similar
135 feature F proved effective for a similar class C in the past.
137 Allows you to trasfer the "gestalt" of a similiar class to a new
138 class, by adapting all the features of the learned class that have
139 correspondance to the new class.
141 *** 2007 Semantic Hierarchies for Recognizing Objects and Parts
143 Better learning of complex objects like faces by learning each
144 piece (like nose, mouth, eye, etc) separately, then making sure
145 that the features are in plausable positions.