# HG changeset patch
# User Robert McIntyre <rlm@mit.edu>
# Date 1398650306 14400
# Node ID d947636fe0eef5b2d332a50d60c60fa9d7ffbe24
# Parent  def6f62831df196eabb6015c6d3d3a50d32674a3
remove comment restrictions.

diff -r def6f62831df -r d947636fe0ee thesis/cortex.org
--- a/thesis/cortex.org	Sun Apr 27 21:57:05 2014 -0400
+++ b/thesis/cortex.org	Sun Apr 27 21:58:26 2014 -0400
@@ -41,7 +41,7 @@
     [[./images/aurellem-gray.png]]
 
 
-* COMMENT Empathy \& Embodiment: problem solving strategies
+* Empathy \& Embodiment: problem solving strategies
 
   By the end of this thesis, you will have seen a novel approach to
   interpreting video using embodiment and empathy. You will also see
@@ -415,7 +415,7 @@
    \end{sidewaysfigure}
 #+END_LaTeX
 
-* COMMENT Designing =CORTEX=
+* Designing =CORTEX=
 
   In this section, I outline the design decisions that went into
   making =CORTEX=, along with some details about its implementation.
@@ -3043,7 +3043,7 @@
    #+end_src
    #+end_listing
   
-** COMMENT =EMPATH= recognizes actions efficiently
+** =EMPATH= recognizes actions efficiently
    
    To use =EMPATH= with the worm, I first need to gather a set of
    experiences from the worm that includes the actions I want to
@@ -3213,7 +3213,7 @@
   identification accuracy of even higher than 95%. When I watch this
   system myself, I generally see no errors in action identification.
 
-** COMMENT Digression: Learning touch sensor layout through free play
+** Digression: Learning touch sensor layout through free play
 
    In the previous section I showed how to compute actions in terms of
    body-centered predicates, but some of those predicates relied on
@@ -3418,7 +3418,7 @@
 \appendix
 #+END_LaTeX
 
-* COMMENT Appendix: =CORTEX= User Guide
+* Appendix: =CORTEX= User Guide
 
   Those who write a thesis should endeavor to make their code not only
   accessible, but actually usable, as a way to pay back the community