comparison org/ullman.org @ 379:f1b8727360fb

add images.
author rlm
date Wed, 10 Apr 2013 16:38:52 -0400
parents
children 2d0afb231081
comparison
equal deleted inserted replaced
378:8e62bf52be59 379:f1b8727360fb
1 #+title: Ullman Literature Review
2 #+author: Robert McIntyre
3 #+email: rlm@mit.edu
4 #+description: Review of some of the AI works of Professor Shimon Ullman.
5 #+keywords: Shimon, Ullman, computer vision, artificial intelligence, literature review
6 #+SETUPFILE: ../../aurellem/org/setup.org
7 #+INCLUDE: ../../aurellem/org/level-0.org
8 #+babel: :mkdirp yes :noweb yes :exports both
9
10
11 * Ullman
12
13 Actual code reuse!
14
15 precision = fraction of retrieved instances that are relevant
16 (true-postives/(true-positives+false-positives))
17
18 recall = fraction of relevant instances that are retrieved
19 (true-positives/total-in-class)
20
21 cross-validation = train the model on two different sets to prevent
22 overfitting.
23
24 nifty, relevant, realistic ideas
25 He doesn't confine himself to unplasaubile assumptions
26
27 ** Our Reading
28
29 *** 2002 Visual features of intermediate complexity and their use in classification
30
31
32
33
34 ** Getting around the dumb "fixed training set" methods
35
36 *** 2006 Learning to classify by ongoing feature selection
37
38 Brings in the most informative features of a class, based on
39 mutual information between that feature and all the examples
40 encountered so far. To bound the running time, he uses only a
41 fixed number of the most recent examples. He uses a replacement
42 strategy to tell whether a new feature is better than one of the
43 corrent features.
44
45 *** 2009 Learning model complexity in an online environment
46
47 Sort of like the heirichal baysean models of Tennanbaum, this
48 system makes the model more and more complicated as it gets more
49 and more training data. It does this by using two systems in
50 parallell and then whenever the more complex one seems to be
51 needed by the data, the less complex one is thrown out, and an
52 even more complex model is initialized in its place.
53
54 He uses a SVM with polynominal kernels of varying complexity. He
55 gets good perfoemance on a handwriting classfication using a large
56 range of training samples, since his model changes complexity
57 depending on the number of training samples. The simpler models do
58 better with few training points, and the more complex ones do
59 better with many training points.
60
61 The final model had intermediate complexity between published
62 extremes.
63
64 The more complex models must be able to be initialized efficiently
65 from the less complex models which they replace!
66
67
68 ** Non Parametric Models
69
70 [[../images/viola-parzen-1.png]]
71 [[../images/viola-parzen-2.png]]
72
73 *** 2010 The chains model for detecting parts by their context
74
75 Like the constelation method for rigid objects, but extended to
76 non-rigid objects as well.
77
78 Allows you to build a hand detector from a face detector. This is
79 usefull because hands might be only a few pixels, and very
80 ambiguous in an image, but if you are expecting them at the end of
81 an arm, then they become easier to find.
82
83 They make chains by using spatial proximity of features. That way,
84 a hand can be idntified by chaining back from the head. If there
85 is a good chain to the head, then it is more likely that there is
86 a hand than if there isn't. Since there is some give in the
87 proximity detection, the system can accomodate new poses that it
88 has never seen before.
89
90 Does not use any motion information.
91
92 *** 2005 A Hierarchical Non-Parametric Method for Capturing Non-Rigid Deformations
93
94 (relative dynamic programming [RDP])
95
96 Goal is to match images, as in SIFT, but this time the images can
97 be subject to non rigid transformations. They do this by finding
98 small patches that look the same, then building up bigger
99 patches. They get a tree of patches that describes each image, and
100 find the edit distance between each tree. Editing operations
101 involve a coherent shift of features, so they can accomodate local
102 shifts of patches in any direction. They get some cool results
103 over just straight correlation. Basically, they made an image
104 comparor that is resistant to multiple independent deformations.
105
106 !important small regions are treated the same as nonimportant
107 small regions
108
109 !no conception of shape
110
111 quote:
112 The dynamic programming procedure looks for an optimal
113 transformation that aligns the patches of both images. This
114 transformation is not a global transformation, but a composition
115 of many local transformations of sub-patches at various sizes,
116 performed one on top of the other.
117
118 *** 2006 Satellite Features for the Classification of Visually Similar Classes
119
120 Finds features that can distinguish subclasses of a class, by
121 first finding a rigid set of anghor features that are common to
122 both subclasses, then finding distinguishing features relative to
123 those subfeatures. They keep things rigid because the satellite
124 features don't have much information in and of themselves, and are
125 only informative relative to other features.
126
127 *** 2005 Learning a novel class from a single example by cross-generalization.
128
129 Let's you use a vast visual experience to generate a classifier
130 for a novel class by generating synthetic examples by replaceing
131 features from the single example with features from similiar
132 classes.
133
134 quote: feature F is likely to be useful for class C if a similar
135 feature F proved effective for a similar class C in the past.
136
137 Allows you to trasfer the "gestalt" of a similiar class to a new
138 class, by adapting all the features of the learned class that have
139 correspondance to the new class.
140
141 *** 2007 Semantic Hierarchies for Recognizing Objects and Parts
142
143 Better learning of complex objects like faces by learning each
144 piece (like nose, mouth, eye, etc) separately, then making sure
145 that the features are in plausable positions.