# HG changeset patch # User Robert McIntyre # Date 1398650306 14400 # Node ID d947636fe0eef5b2d332a50d60c60fa9d7ffbe24 # Parent def6f62831df196eabb6015c6d3d3a50d32674a3 remove comment restrictions. diff -r def6f62831df -r d947636fe0ee thesis/cortex.org --- a/thesis/cortex.org Sun Apr 27 21:57:05 2014 -0400 +++ b/thesis/cortex.org Sun Apr 27 21:58:26 2014 -0400 @@ -41,7 +41,7 @@ [[./images/aurellem-gray.png]] -* COMMENT Empathy \& Embodiment: problem solving strategies +* Empathy \& Embodiment: problem solving strategies By the end of this thesis, you will have seen a novel approach to interpreting video using embodiment and empathy. You will also see @@ -415,7 +415,7 @@ \end{sidewaysfigure} #+END_LaTeX -* COMMENT Designing =CORTEX= +* Designing =CORTEX= In this section, I outline the design decisions that went into making =CORTEX=, along with some details about its implementation. @@ -3043,7 +3043,7 @@ #+end_src #+end_listing -** COMMENT =EMPATH= recognizes actions efficiently +** =EMPATH= recognizes actions efficiently To use =EMPATH= with the worm, I first need to gather a set of experiences from the worm that includes the actions I want to @@ -3213,7 +3213,7 @@ identification accuracy of even higher than 95%. When I watch this system myself, I generally see no errors in action identification. -** COMMENT Digression: Learning touch sensor layout through free play +** Digression: Learning touch sensor layout through free play In the previous section I showed how to compute actions in terms of body-centered predicates, but some of those predicates relied on @@ -3418,7 +3418,7 @@ \appendix #+END_LaTeX -* COMMENT Appendix: =CORTEX= User Guide +* Appendix: =CORTEX= User Guide Those who write a thesis should endeavor to make their code not only accessible, but actually usable, as a way to pay back the community