Mercurial > cortex
view thesis/abstract.org @ 479:a5480a06d5fe
first draft of proprioception complete.
author | Robert McIntyre <rlm@mit.edu> |
---|---|
date | Fri, 28 Mar 2014 23:16:32 -0400 |
parents | c20de2267d39 |
children | 8e52a2802821 |
line wrap: on
line source
1 Here I demonstrate the power of using embodied artificial intelligence2 to attack the /action recognition/ problem, which is the challenge of3 recognizing actions performed by a creature given limited data about4 the creature's actions, such as a video recording. I solve this5 problem in the case of a worm-like creature performing actions such as6 curling and wiggling.8 To attack the action recognition problem, I developed a computational9 model of empathy (=EMPATH=) which allows me to recognize actions using10 simple, embodied representations of actions (which require rich11 sensory data), even when that sensory data is not actually12 available. The missing sense data is ``imagined'' by the system by13 combining previous experiences gained from unsupervised free play.15 In order to build this empathic, action-recognizing system, I created16 a program called =CORTEX=, which is a complete platform for embodied17 AI research. It provides multiple senses for simulated creatures,18 including vision, touch, proprioception, muscle tension, and19 hearing. Each of these senses provides a wealth of parameters that are20 biologically inspired. =CORTEX= is able to simulate any number of21 creatures and senses, and provides facilities for easily modeling and22 creating new creatures. As a research platform it is more complete23 than any other system currently available.