diff org/ullman.org @ 379:f1b8727360fb

add images.
author rlm
date Wed, 10 Apr 2013 16:38:52 -0400
parents
children 2d0afb231081
line wrap: on
line diff
     1.1 --- /dev/null	Thu Jan 01 00:00:00 1970 +0000
     1.2 +++ b/org/ullman.org	Wed Apr 10 16:38:52 2013 -0400
     1.3 @@ -0,0 +1,145 @@
     1.4 +#+title: Ullman Literature Review
     1.5 +#+author: Robert McIntyre
     1.6 +#+email: rlm@mit.edu
     1.7 +#+description: Review of some of the AI works of Professor Shimon Ullman.
     1.8 +#+keywords: Shimon, Ullman, computer vision, artificial intelligence, literature review
     1.9 +#+SETUPFILE: ../../aurellem/org/setup.org
    1.10 +#+INCLUDE: ../../aurellem/org/level-0.org
    1.11 +#+babel: :mkdirp yes :noweb yes :exports both
    1.12 +
    1.13 +
    1.14 +* Ullman 
    1.15 +
    1.16 +Actual code reuse!
    1.17 +
    1.18 +precision = fraction of retrieved instances that are relevant
    1.19 +  (true-postives/(true-positives+false-positives))
    1.20 +
    1.21 +recall    =  fraction of relevant instances that are retrieved
    1.22 +  (true-positives/total-in-class)
    1.23 +
    1.24 +cross-validation = train the model on two different sets to prevent
    1.25 +overfitting. 
    1.26 +
    1.27 +nifty, relevant, realistic ideas
    1.28 +He doesn't confine himself to unplasaubile assumptions
    1.29 +
    1.30 +** Our Reading
    1.31 +
    1.32 +*** 2002 Visual features of intermediate complexity and their use in classification
    1.33 +
    1.34 +    
    1.35 +
    1.36 +
    1.37 +** Getting around the dumb "fixed training set" methods
    1.38 +
    1.39 +*** 2006 Learning to classify by ongoing feature selection
    1.40 +    
    1.41 +    Brings in the most informative features of a class, based on
    1.42 +    mutual information between that feature and all the examples
    1.43 +    encountered so far. To bound the running time, he uses only a
    1.44 +    fixed number of the most recent examples. He uses a replacement
    1.45 +    strategy to tell whether a new feature is better than one of the
    1.46 +    corrent features.
    1.47 +
    1.48 +*** 2009 Learning model complexity in an online environment
    1.49 +    
    1.50 +    Sort of like the heirichal baysean models of Tennanbaum, this
    1.51 +    system makes the model more and more complicated as it gets more
    1.52 +    and more training data. It does this by using two systems in
    1.53 +    parallell and then whenever the more complex one seems to be
    1.54 +    needed by the data, the less complex one is thrown out, and an
    1.55 +    even more complex model is initialized in its place.
    1.56 +
    1.57 +    He uses a SVM with polynominal kernels of varying complexity. He
    1.58 +    gets good perfoemance on a handwriting classfication using a large
    1.59 +    range of training samples, since his model changes complexity
    1.60 +    depending on the number of training samples. The simpler models do
    1.61 +    better with few training points, and the more complex ones do
    1.62 +    better with many training points.
    1.63 +
    1.64 +    The final model had intermediate complexity between published
    1.65 +    extremes. 
    1.66 +
    1.67 +    The more complex models must be able to be initialized efficiently
    1.68 +    from the less complex models which they replace!
    1.69 +
    1.70 +
    1.71 +** Non Parametric Models
    1.72 +
    1.73 +[[../images/viola-parzen-1.png]]
    1.74 +[[../images/viola-parzen-2.png]]
    1.75 +
    1.76 +*** 2010 The chains model for detecting parts by their context
    1.77 +
    1.78 +    Like the constelation method for rigid objects, but extended to
    1.79 +    non-rigid objects as well.
    1.80 +
    1.81 +    Allows you to build a hand detector from a face detector. This is
    1.82 +    usefull because hands might be only a few pixels, and very
    1.83 +    ambiguous in an image, but if you are expecting them at the end of
    1.84 +    an arm, then they become easier to find.
    1.85 +
    1.86 +    They make chains by using spatial proximity of features. That way,
    1.87 +    a hand can be idntified by chaining back from the head. If there
    1.88 +    is a good chain to the head, then it is more likely that there is
    1.89 +    a hand than if there isn't. Since there is some give in the
    1.90 +    proximity detection, the system can accomodate new poses that it
    1.91 +    has never seen before.
    1.92 +
    1.93 +    Does not use any motion information.
    1.94 +
    1.95 +*** 2005 A Hierarchical Non-Parametric Method for Capturing Non-Rigid Deformations
    1.96 +    
    1.97 +    (relative dynamic programming [RDP])
    1.98 +
    1.99 +    Goal is to match images, as in SIFT, but this time the images can
   1.100 +    be subject to non rigid transformations. They do this by finding
   1.101 +    small patches that look the same, then building up bigger
   1.102 +    patches. They get a tree of patches that describes each image, and
   1.103 +    find the edit distance between each tree. Editing operations
   1.104 +    involve a coherent shift of features, so they can accomodate local
   1.105 +    shifts of patches in any direction. They get some cool results
   1.106 +    over just straight correlation. Basically, they made an image
   1.107 +    comparor that is resistant to multiple independent deformations.
   1.108 +    
   1.109 +    !important small regions are treated the same as nonimportant
   1.110 +     small regions
   1.111 +     
   1.112 +    !no conception of shape
   1.113 +    
   1.114 +    quote:
   1.115 +    The dynamic programming procedure looks for an optimal
   1.116 +    transformation that aligns the patches of both images. This
   1.117 +    transformation is not a global transformation, but a composition
   1.118 +    of many local transformations of sub-patches at various sizes,
   1.119 +    performed one on top of the other.
   1.120 +
   1.121 +*** 2006 Satellite Features for the Classification of Visually Similar Classes
   1.122 +    
   1.123 +    Finds features that can distinguish subclasses of a class, by
   1.124 +    first finding a rigid set of anghor features that are common to
   1.125 +    both subclasses, then finding distinguishing features relative to
   1.126 +    those subfeatures. They keep things rigid because the satellite
   1.127 +    features don't have much information in and of themselves, and are
   1.128 +    only informative relative to other features.
   1.129 +
   1.130 +*** 2005 Learning a novel class from a single example by cross-generalization.
   1.131 +
   1.132 +    Let's you use a vast visual experience to generate a classifier
   1.133 +    for a novel class by generating synthetic examples by replaceing
   1.134 +    features from the single example with features from similiar
   1.135 +    classes.
   1.136 +
   1.137 +    quote: feature F is likely to be useful for class C if a similar
   1.138 +    feature F proved effective for a similar class C in the past.
   1.139 +
   1.140 +    Allows you to trasfer the "gestalt" of a similiar class to a new
   1.141 +    class, by adapting all the features of the learned class that have
   1.142 +    correspondance to the new class.
   1.143 +
   1.144 +*** 2007 Semantic Hierarchies for Recognizing Objects and Parts
   1.145 +
   1.146 +    Better learning of complex objects like faces by learning each
   1.147 +    piece (like nose, mouth, eye, etc) separately, then making sure
   1.148 +    that the features are in plausable positions.