diff org/literature-review.org @ 376:057d47fc4789

reviewing ullman's stuff.
author Robert McIntyre <rlm@mit.edu>
date Thu, 11 Apr 2013 05:40:23 +0000
parents 9c37a55e1cd2
children 80cd096682b2
line wrap: on
line diff
     1.1 --- a/org/literature-review.org	Tue Mar 12 04:34:37 2013 +0000
     1.2 +++ b/org/literature-review.org	Thu Apr 11 05:40:23 2013 +0000
     1.3 @@ -42,13 +42,13 @@
     1.4  
     1.5   - Zhang, Z., R. Deriche, O. Faugeras, Q.T. Luong, “A robust
     1.6    technique for matching two uncalibrated images through the recovery
     1.7 -  of the unknown epipolar geometry,” Artificial In- telligence, 78,
     1.8 +  of the unknown epipolar geometry,” Artificial Intelligence, 78,
     1.9    (1995), pp. 87-119.
    1.10  
    1.11  
    1.12  
    1.13  
    1.14 -
    1.15 +   
    1.16  * Alignment by Maximization of Mutual Information, Paul A. Viola
    1.17  
    1.18    PhD Thesis recommended by Winston. Describes a system that is able
    1.19 @@ -63,18 +63,198 @@
    1.20    - Differential entropy seems a bit odd -- you would think that it
    1.21      should be the same as normal entropy for a discrete distrubition
    1.22      embedded in continuous space. How do you measure the entropy of a
    1.23 -    half continuous, half discrete random variable?
    1.24 +    half continuous, half discrete random variable? Perhaps the
    1.25 +    problem is related to the delta function, and not the definition
    1.26 +    of differential entropy?
    1.27  
    1.28    - Expectation Maximation (Mixture of Gaussians cool stuff)
    1.29      (Dempster 1977)
    1.30  
    1.31    - Good introduction to Parzen Window Density Estimation. Parzen
    1.32      density functions trade construction time for evaulation
    1.33 -    time.(Pg. 41)
    1.34 +    time.(Pg. 41) They are a way to transform a sample into a
    1.35 +    distribution. They don't work very well in higher dimensions due
    1.36 +    to the thinning of sample points.
    1.37 +
    1.38 +  - Calculating the entropy of a Markov Model (or state machine,
    1.39 +    program, etc) seems like it would be very hard, since each trial
    1.40 +    would not be independent of the other trials. Yet, there are many
    1.41 +    common sense models that do need to have state to accurately model
    1.42 +    the world.
    1.43 +
    1.44 +  - "... there is no direct procedure for evaluating entropy from a
    1.45 +    sample. A common approach is to model the density from the sample,
    1.46 +    and then estimate the entropy from the density."
    1.47 +
    1.48 +  - pg. 55 he says that infinity minus infinity is zero lol.
    1.49 +
    1.50 +  - great idea on pg 62 about using random samples from images to
    1.51 +    speed up computation.
    1.52 +
    1.53 +  - practical way of terminating a random search: "A better idea is to
    1.54 +    reduce the learning rate until the parameters have a reasonable
    1.55 +    variance and then take the average parameters."
    1.56 +
    1.57 +  - p. 65 bullshit hack to make his parzen window estimates work.
    1.58 +
    1.59 +  - this alignment only works if the initial pose is not very far
    1.60 +    off. 
    1.61 +
    1.62  
    1.63    Occlusion? Seems a bit holistic.
    1.64  
    1.65 +** References
    1.66 + - "excellent" book on entropy (Cover & Thomas, 1991) [Elements of
    1.67 +   Information Theory.] 
    1.68 +
    1.69 + - Canny, J. (1986). A Computational Approach to Edge Detection. IEEE
    1.70 +   Transactions PAMI, PAMI-8(6):679{698
    1.71 +
    1.72 + - Chin, R. and Dyer, C. (1986). Model-Based Recognition in Robot
    1.73 +   Vision. Computing Surveys, 18:67-108.
    1.74 +
    1.75 + - Grimson, W., Lozano-Perez, T., Wells, W., et al. (1994). An
    1.76 +   Automatic Registration Method for Frameless Stereotaxy, Image
    1.77 +   Guided Surgery, and Enhanced Realigy Visualization. In Proceedings
    1.78 +   of the Computer Society Conference on Computer Vision and Pattern
    1.79 +   Recognition, Seattle, WA. IEEE.
    1.80 +
    1.81 + - Hill, D. L., Studholme, C., and Hawkes, D. J. (1994). Voxel
    1.82 +   Similarity Measures for Auto-mated Image Registration. In
    1.83 +   Proceedings of the Third Conference on Visualization in Biomedical
    1.84 +   Computing, pages 205 { 216. SPIE.
    1.85 +
    1.86 + - Kirkpatrick, S., Gelatt, C., and Vecch Optimization by Simulated
    1.87 +   Annealing. Science, 220(4598):671-680.
    1.88 +
    1.89 + - Jones, M. and Poggio, T. (1995). Model-based matching of line
    1.90 +   drawings by linear combin-ations of prototypes. Proceedings of the
    1.91 +   International Conference on Computer Vision
    1.92 +
    1.93 + - Ljung, L. and Soderstrom, T. (1983). Theory and Practice of
    1.94 +   Recursive Identi cation. MIT Press.
    1.95 +
    1.96 + - Shannon, C. E. (1948). A mathematical theory of communication. Bell
    1.97 +   Systems Technical Journal, 27:379-423 and 623-656.
    1.98 +
    1.99 + - Shashua, A. (1992). Geometry and Photometry in 3D Visual
   1.100 +   Recognition. PhD thesis, M.I.T Artificial Intelligence Laboratory,
   1.101 +   AI-TR-1401.
   1.102 +
   1.103 + - William H. Press, Brian P. Flannery, S. A. T. and Veterling,
   1.104 +   W. T. (1992). Numerical Recipes in C: The Art of Scienti c
   1.105 +   Computing. Cambridge University Press, Cambridge, England, second
   1.106 +   edition edition.
   1.107 +
   1.108 +* Semi-Automated Dialogue Act Classification for Situated Social Agents in Games, Deb Roy 
   1.109 +  
   1.110 +  Interesting attempt to learn "social scripts" related to resturant
   1.111 +  behaviour. The authors do this by creating a game which implements a
   1.112 +  virtual restruant, and recoding actual human players as they
   1.113 +  interact with the game. The learn scripts from annotated
   1.114 +  interactions and then use those scripts to label other
   1.115 +  interactions. They don't get very good results, but their
   1.116 +  methodology of creating a virtual world and recording
   1.117 +  low-dimensional actions is interesting.
   1.118 +
   1.119 +  - Torque 2D/3D looks like an interesting game engine.
   1.120 +
   1.121 +
   1.122 +* Face Recognition by Humans: Nineteen Results all Computer Vision Researchers should know, Sinha
   1.123 +  
   1.124 +  This is a summary of a lot of bio experiments on human face
   1.125 +  recognition.
   1.126 +  
   1.127 +  - They assert again that the internal gradients/structures of a face
   1.128 +    are more important than the edges.
   1.129 +
   1.130 +  - It's amazing to me that it takes about 10 years after birth for a
   1.131 +    human to get advanced adult-like face detection. They go through
   1.132 +    feature based processing to a holistic based approach during this
   1.133 +    time.
   1.134 +
   1.135 +  - Finally, color is a very important cue for identifying faces.
   1.136  
   1.137  ** References
   1.138 - - "excellent" book on entropy (Cover & Thomas, 1991)
   1.139 - 
   1.140 \ No newline at end of file
   1.141 +  - A. Freire, K. Lee, and L. A. Symons, BThe face-inversion effect as
   1.142 +    a deficit in the encoding of configural information: Direct
   1.143 +    evidence,[ Perception, vol. 29, no. 2, pp. 159–170, 2000.
   1.144 +  - M. B. Lewis, BThatcher’s children: Development and the Thatcher
   1.145 +    illusion,[Perception, vol. 32, pp. 1415–21, 2003.
   1.146 +  - E. McKone and N. Kanwisher, BDoes the human brain process objects
   1.147 +    of expertise like faces? A review of the evidence,[ in From Monkey
   1.148 +    Brain to Human Brain, S. Dehaene, J. R. Duhamel, M. Hauser, and
   1.149 +    G. Rizzolatti, Eds. Cambridge, MA: MIT Press, 2005.
   1.150 +
   1.151 +
   1.152 +
   1.153 +
   1.154 +heee~eeyyyy kids, time to get eagle'd!!!!
   1.155 +
   1.156 +
   1.157 +
   1.158 +
   1.159 +
   1.160 +* Ullman 
   1.161 +
   1.162 +Actual code reuse!
   1.163 +
   1.164 +precision = fraction of retrieved instances that are relevant
   1.165 +  (true-postives/(true-positives+false-positives))
   1.166 +
   1.167 +recall    =  fraction of relevant instances that are retrieved
   1.168 +  (true-positives/total-in-class)
   1.169 +
   1.170 +cross-validation = train the model on two different sets to prevent
   1.171 +overfitting. 
   1.172 +
   1.173 +
   1.174 +
   1.175 +
   1.176 +
   1.177 +** Getting around the dumb "fixed training set" methods
   1.178 +
   1.179 +*** 2006 Learning to classify by ongoing feature selection
   1.180 +    
   1.181 +    Brings in the most informative features of a class, based on
   1.182 +    mutual information between that feature and all the examples
   1.183 +    encountered so far. To bound the running time, he uses only a
   1.184 +    fixed number of the most recent examples. He uses a replacement
   1.185 +    strategy to tell whether a new feature is better than one of the
   1.186 +    corrent features.
   1.187 +
   1.188 +*** 2009 Learning model complexity in an online environment
   1.189 +    
   1.190 +    Sort of like the heirichal baysean models of Tennanbaum, this
   1.191 +    system makes the model more and more complicated as it gets more
   1.192 +    and more training data. It does this by using two systems in
   1.193 +    parallell and then whenever the more complex one seems to be
   1.194 +    needed by the data, the less complex one is thrown out, and an
   1.195 +    even more complex model is initialized in its place.
   1.196 +
   1.197 +    He uses a SVM with polynominal kernels of varying complexity. He
   1.198 +    gets good perfoemance on a handwriting classfication using a large
   1.199 +    range of training samples, since his model changes complexity
   1.200 +    depending on the number of training samples. The simpler models do
   1.201 +    better with few training points, and the more complex ones do
   1.202 +    better with many training points.
   1.203 +
   1.204 +    The more complex models must be able to be initialized efficiently
   1.205 +    from the less complex models which they replace!
   1.206 +
   1.207 +
   1.208 +** Non Parametric Models
   1.209 +
   1.210 +*** Visual features of intermediate complexity and their use in classification
   1.211 +
   1.212 +*** The chains model for detecting parts by their context
   1.213 +
   1.214 +    Like the constelation method for rigid objects, but extended to
   1.215 +    non-rigid objects as well.
   1.216 +
   1.217 +    Allows you to build a hand detector from a face detector. This is
   1.218 +    usefull because hands might be only a few pixels, and very
   1.219 +    ambiguous in an image, but if you are expecting them at the end of
   1.220 +    an arm, then they become easier to find.
   1.221 +
   1.222 +    
   1.223 \ No newline at end of file