rlm@436
|
1 Here I demonstrate the power of using embodied artificial intelligence
|
rlm@436
|
2 to attack the /action recognition/ problem, which is the challenge of
|
rlm@436
|
3 recognizing actions performed by a creature given limited data about
|
rlm@437
|
4 the creature's actions, such as a video recording. I solve this
|
rlm@437
|
5 problem in the case of a worm-like creature performing actions such as
|
rlm@524
|
6 curling and wiggling.
|
rlm@432
|
7
|
rlm@436
|
8 To attack the action recognition problem, I developed a computational
|
rlm@441
|
9 model of empathy (=EMPATH=) which allows me to recognize actions using
|
rlm@441
|
10 simple, embodied representations of actions (which require rich
|
rlm@524
|
11 sensory data), even when that sensory data is not actually available.
|
rlm@551
|
12 The missing sense data is imagined by combining previous experiences
|
rlm@551
|
13 gained from unsupervised free play. The worm is a five-segment
|
rlm@551
|
14 creature equipped with touch, proprioception, and muscle tension
|
rlm@551
|
15 senses. It recognizes actions using only proprioception data.
|
rlm@432
|
16
|
rlm@436
|
17 In order to build this empathic, action-recognizing system, I created
|
rlm@436
|
18 a program called =CORTEX=, which is a complete platform for embodied
|
rlm@436
|
19 AI research. It provides multiple senses for simulated creatures,
|
rlm@524
|
20 including vision, touch, proprioception, muscle tension, and hearing.
|
rlm@524
|
21 Each of these senses provides a wealth of parameters that are
|
rlm@436
|
22 biologically inspired. =CORTEX= is able to simulate any number of
|
rlm@436
|
23 creatures and senses, and provides facilities for easily modeling and
|
rlm@436
|
24 creating new creatures. As a research platform it is more complete
|
rlm@436
|
25 than any other system currently available.
|