view org/touch.org @ 560:1e5b5eb8c576

solved problem with biblography.
author Robert McIntyre <rlm@mit.edu>
date Fri, 02 May 2014 15:36:36 -0400
parents 763d13f77e03
children
line wrap: on
line source
1 #+title: Simulated Sense of Touch
2 #+author: Robert McIntyre
3 #+email: rlm@mit.edu
4 #+description: Simulated touch for AI research using JMonkeyEngine and clojure.
5 #+keywords: simulation, tactile sense, jMonkeyEngine3, clojure
6 #+SETUPFILE: ../../aurellem/org/setup.org
7 #+INCLUDE: ../../aurellem/org/level-0.org
9 * Touch
11 Touch is critical to navigation and spatial reasoning and as such I
12 need a simulated version of it to give to my AI creatures.
14 Human skin has a wide array of touch sensors, each of which specialize
15 in detecting different vibrational modes and pressures. These sensors
16 can integrate a vast expanse of skin (i.e. your entire palm), or a
17 tiny patch of skin at the tip of your finger. The hairs of the skin
18 help detect objects before they even come into contact with the skin
19 proper.
21 However, touch in my simulated world can not exactly correspond to
22 human touch because my creatures are made out of completely rigid
23 segments that don't deform like human skin.
25 Instead of measuring deformation or vibration, I surround each rigid
26 part with a plenitude of hair-like objects (/feelers/) which do not
27 interact with the physical world. Physical objects can pass through
28 them with no effect. The feelers are able to tell when other objects
29 pass through them, and they constantly report how much of their extent
30 is covered. So even though the creature's body parts do not deform,
31 the feelers create a margin around those body parts which achieves a
32 sense of touch which is a hybrid between a human's sense of
33 deformation and sense from hairs.
35 Implementing touch in jMonkeyEngine follows a different technical route
36 than vision and hearing. Those two senses piggybacked off
37 jMonkeyEngine's 3D audio and video rendering subsystems. To simulate
38 touch, I use jMonkeyEngine's physics system to execute many small
39 collision detections, one for each feeler. The placement of the
40 feelers is determined by a UV-mapped image which shows where each
41 feeler should be on the 3D surface of the body.
43 * Defining Touch Meta-Data in Blender
45 Each geometry can have a single UV map which describes the position of
46 the feelers which will constitute its sense of touch. This image path
47 is stored under the "touch" key. The image itself is black and white,
48 with black meaning a feeler length of 0 (no feeler is present) and
49 white meaning a feeler length of =scale=, which is a float stored
50 under the key "scale".
52 #+name: meta-data
53 #+begin_src clojure
54 (defn tactile-sensor-profile
55 "Return the touch-sensor distribution image in BufferedImage format,
56 or nil if it does not exist."
57 [#^Geometry obj]
58 (if-let [image-path (meta-data obj "touch")]
59 (load-image image-path)))
61 (defn tactile-scale
62 "Return the length of each feeler. Default scale is 0.01
63 jMonkeyEngine units."
64 [#^Geometry obj]
65 (if-let [scale (meta-data obj "scale")]
66 scale 0.1))
67 #+end_src
69 Here is an example of a UV-map which specifies the position of touch
70 sensors along the surface of the upper segment of the worm.
72 #+attr_html: width=755
73 #+caption: This is the tactile-sensor-profile for the upper segment of the worm. It defines regions of high touch sensitivity (where there are many white pixels) and regions of low sensitivity (where white pixels are sparse).
74 [[../images/finger-UV.png]]
76 * Implementation Summary
78 To simulate touch there are three conceptual steps. For each solid
79 object in the creature, you first have to get UV image and scale
80 parameter which define the position and length of the feelers. Then,
81 you use the triangles which comprise the mesh and the UV data stored in
82 the mesh to determine the world-space position and orientation of each
83 feeler. Then once every frame, update these positions and orientations
84 to match the current position and orientation of the object, and use
85 physics collision detection to gather tactile data.
87 Extracting the meta-data has already been described. The third step,
88 physics collision detection, is handled in =touch-kernel=.
89 Translating the positions and orientations of the feelers from the
90 UV-map to world-space is itself a three-step process.
92 - Find the triangles which make up the mesh in pixel-space and in
93 world-space. =triangles= =pixel-triangles=.
95 - Find the coordinates of each feeler in world-space. These are the
96 origins of the feelers. =feeler-origins=.
98 - Calculate the normals of the triangles in world space, and add
99 them to each of the origins of the feelers. These are the
100 normalized coordinates of the tips of the feelers. =feeler-tips=.
102 * Triangle Math
103 ** Shrapnel Conversion Functions
105 #+name: triangles-1
106 #+begin_src clojure
107 (defn vector3f-seq [#^Vector3f v]
108 [(.getX v) (.getY v) (.getZ v)])
110 (defn triangle-seq [#^Triangle tri]
111 [(vector3f-seq (.get1 tri))
112 (vector3f-seq (.get2 tri))
113 (vector3f-seq (.get3 tri))])
115 (defn ->vector3f
116 ([coords] (Vector3f. (nth coords 0 0)
117 (nth coords 1 0)
118 (nth coords 2 0))))
120 (defn ->triangle [points]
121 (apply #(Triangle. %1 %2 %3) (map ->vector3f points)))
122 #+end_src
124 It is convenient to treat a =Triangle= as a vector of vectors, and a
125 =Vector2f= or =Vector3f= as vectors of floats. (->vector3f) and
126 (->triangle) undo the operations of =vector3f-seq= and
127 =triangle-seq=. If these classes implemented =Iterable= then =seq=
128 would work on them automatically.
130 ** Decomposing a 3D shape into Triangles
132 The rigid objects which make up a creature have an underlying
133 =Geometry=, which is a =Mesh= plus a =Material= and other important
134 data involved with displaying the object.
136 A =Mesh= is composed of =Triangles=, and each =Triangle= has three
137 vertices which have coordinates in world space and UV space.
139 Here, =triangles= gets all the world-space triangles which comprise a
140 mesh, while =pixel-triangles= gets those same triangles expressed in
141 pixel coordinates (which are UV coordinates scaled to fit the height
142 and width of the UV image).
144 #+name: triangles-2
145 #+begin_src clojure
146 (in-ns 'cortex.touch)
147 (defn triangle
148 "Get the triangle specified by triangle-index from the mesh."
149 [#^Geometry geo triangle-index]
150 (triangle-seq
151 (let [scratch (Triangle.)]
152 (.getTriangle (.getMesh geo) triangle-index scratch) scratch)))
154 (defn triangles
155 "Return a sequence of all the Triangles which comprise a given
156 Geometry."
157 [#^Geometry geo]
158 (map (partial triangle geo) (range (.getTriangleCount (.getMesh geo)))))
160 (defn triangle-vertex-indices
161 "Get the triangle vertex indices of a given triangle from a given
162 mesh."
163 [#^Mesh mesh triangle-index]
164 (let [indices (int-array 3)]
165 (.getTriangle mesh triangle-index indices)
166 (vec indices)))
168 (defn vertex-UV-coord
169 "Get the UV-coordinates of the vertex named by vertex-index"
170 [#^Mesh mesh vertex-index]
171 (let [UV-buffer
172 (.getData
173 (.getBuffer
174 mesh
175 VertexBuffer$Type/TexCoord))]
176 [(.get UV-buffer (* vertex-index 2))
177 (.get UV-buffer (+ 1 (* vertex-index 2)))]))
179 (defn pixel-triangle [#^Geometry geo image index]
180 (let [mesh (.getMesh geo)
181 width (.getWidth image)
182 height (.getHeight image)]
183 (vec (map (fn [[u v]] (vector (* width u) (* height v)))
184 (map (partial vertex-UV-coord mesh)
185 (triangle-vertex-indices mesh index))))))
187 (defn pixel-triangles
188 "The pixel-space triangles of the Geometry, in the same order as
189 (triangles geo)"
190 [#^Geometry geo image]
191 (let [height (.getHeight image)
192 width (.getWidth image)]
193 (map (partial pixel-triangle geo image)
194 (range (.getTriangleCount (.getMesh geo))))))
195 #+end_src
196 ** The Affine Transform from one Triangle to Another
198 =pixel-triangles= gives us the mesh triangles expressed in pixel
199 coordinates and =triangles= gives us the mesh triangles expressed in
200 world coordinates. The tactile-sensor-profile gives the position of
201 each feeler in pixel-space. In order to convert pixel-space
202 coordinates into world-space coordinates we need something that takes
203 coordinates on the surface of one triangle and gives the corresponding
204 coordinates on the surface of another triangle.
206 Triangles are [[http://mathworld.wolfram.com/AffineTransformation.html ][affine]], which means any triangle can be transformed into
207 any other by a combination of translation, scaling, and
208 rotation. The affine transformation from one triangle to another
209 is readily computable if the triangle is expressed in terms of a $4x4$
210 matrix.
212 \begin{bmatrix}
213 x_1 & x_2 & x_3 & n_x \\
214 y_1 & y_2 & y_3 & n_y \\
215 z_1 & z_2 & z_3 & n_z \\
216 1 & 1 & 1 & 1
217 \end{bmatrix}
219 Here, the first three columns of the matrix are the vertices of the
220 triangle. The last column is the right-handed unit normal of the
221 triangle.
223 With two triangles $T_{1}$ and $T_{2}$ each expressed as a matrix like
224 above, the affine transform from $T_{1}$ to $T_{2}$ is
226 $T_{2}T_{1}^{-1}$
228 The clojure code below recapitulates the formulas above, using
229 jMonkeyEngine's =Matrix4f= objects, which can describe any affine
230 transformation.
232 #+name: triangles-3
233 #+begin_src clojure
234 (in-ns 'cortex.touch)
236 (defn triangle->matrix4f
237 "Converts the triangle into a 4x4 matrix: The first three columns
238 contain the vertices of the triangle; the last contains the unit
239 normal of the triangle. The bottom row is filled with 1s."
240 [#^Triangle t]
241 (let [mat (Matrix4f.)
242 [vert-1 vert-2 vert-3]
243 (mapv #(.get t %) (range 3))
244 unit-normal (do (.calculateNormal t)(.getNormal t))
245 vertices [vert-1 vert-2 vert-3 unit-normal]]
246 (dorun
247 (for [row (range 4) col (range 3)]
248 (do
249 (.set mat col row (.get (vertices row) col))
250 (.set mat 3 row 1)))) mat))
252 (defn triangles->affine-transform
253 "Returns the affine transformation that converts each vertex in the
254 first triangle into the corresponding vertex in the second
255 triangle."
256 [#^Triangle tri-1 #^Triangle tri-2]
257 (.mult
258 (triangle->matrix4f tri-2)
259 (.invert (triangle->matrix4f tri-1))))
260 #+end_src
261 ** Triangle Boundaries
263 For efficiency's sake I will divide the tactile-profile image into
264 small squares which inscribe each pixel-triangle, then extract the
265 points which lie inside the triangle and map them to 3D-space using
266 =triangle-transform= above. To do this I need a function,
267 =convex-bounds= which finds the smallest box which inscribes a 2D
268 triangle.
270 =inside-triangle?= determines whether a point is inside a triangle
271 in 2D pixel-space.
273 #+name: triangles-4
274 #+begin_src clojure
275 (defn convex-bounds
276 "Returns the smallest square containing the given vertices, as a
277 vector of integers [left top width height]."
278 [verts]
279 (let [xs (map first verts)
280 ys (map second verts)
281 x0 (Math/floor (apply min xs))
282 y0 (Math/floor (apply min ys))
283 x1 (Math/ceil (apply max xs))
284 y1 (Math/ceil (apply max ys))]
285 [x0 y0 (- x1 x0) (- y1 y0)]))
287 (defn same-side?
288 "Given the points p1 and p2 and the reference point ref, is point p
289 on the same side of the line that goes through p1 and p2 as ref is?"
290 [p1 p2 ref p]
291 (<=
292 0
293 (.dot
294 (.cross (.subtract p2 p1) (.subtract p p1))
295 (.cross (.subtract p2 p1) (.subtract ref p1)))))
297 (defn inside-triangle?
298 "Is the point inside the triangle?"
299 {:author "Dylan Holmes"}
300 [#^Triangle tri #^Vector3f p]
301 (let [[vert-1 vert-2 vert-3] [(.get1 tri) (.get2 tri) (.get3 tri)]]
302 (and
303 (same-side? vert-1 vert-2 vert-3 p)
304 (same-side? vert-2 vert-3 vert-1 p)
305 (same-side? vert-3 vert-1 vert-2 p))))
306 #+end_src
308 * Feeler Coordinates
310 The triangle-related functions above make short work of calculating
311 the positions and orientations of each feeler in world-space.
313 #+name: sensors
314 #+begin_src clojure
315 (in-ns 'cortex.touch)
317 (defn feeler-pixel-coords
318 "Returns the coordinates of the feelers in pixel space in lists, one
319 list for each triangle, ordered in the same way as (triangles) and
320 (pixel-triangles)."
321 [#^Geometry geo image]
322 (map
323 (fn [pixel-triangle]
324 (filter
325 (fn [coord]
326 (inside-triangle? (->triangle pixel-triangle)
327 (->vector3f coord)))
328 (white-coordinates image (convex-bounds pixel-triangle))))
329 (pixel-triangles geo image)))
331 (defn feeler-world-coords
332 "Returns the coordinates of the feelers in world space in lists, one
333 list for each triangle, ordered in the same way as (triangles) and
334 (pixel-triangles)."
335 [#^Geometry geo image]
336 (let [transforms
337 (map #(triangles->affine-transform
338 (->triangle %1) (->triangle %2))
339 (pixel-triangles geo image)
340 (triangles geo))]
341 (map (fn [transform coords]
342 (map #(.mult transform (->vector3f %)) coords))
343 transforms (feeler-pixel-coords geo image))))
345 (defn feeler-origins
346 "The world space coordinates of the root of each feeler."
347 [#^Geometry geo image]
348 (reduce concat (feeler-world-coords geo image)))
350 (defn feeler-tips
351 "The world space coordinates of the tip of each feeler."
352 [#^Geometry geo image]
353 (let [world-coords (feeler-world-coords geo image)
354 normals
355 (map
356 (fn [triangle]
357 (.calculateNormal triangle)
358 (.clone (.getNormal triangle)))
359 (map ->triangle (triangles geo)))]
361 (mapcat (fn [origins normal]
362 (map #(.add % normal) origins))
363 world-coords normals)))
365 (defn touch-topology
366 "touch-topology? is not a function."
367 [#^Geometry geo image]
368 (collapse (reduce concat (feeler-pixel-coords geo image))))
369 #+end_src
370 * Simulated Touch
372 =touch-kernel= generates functions to be called from within a
373 simulation that perform the necessary physics collisions to collect
374 tactile data, and =touch!= recursively applies it to every node in
375 the creature.
377 #+name: kernel
378 #+begin_src clojure
379 (in-ns 'cortex.touch)
381 (defn set-ray [#^Ray ray #^Matrix4f transform
382 #^Vector3f origin #^Vector3f tip]
383 ;; Doing everything locally reduces garbage collection by enough to
384 ;; be worth it.
385 (.mult transform origin (.getOrigin ray))
386 (.mult transform tip (.getDirection ray))
387 (.subtractLocal (.getDirection ray) (.getOrigin ray))
388 (.normalizeLocal (.getDirection ray)))
390 (import com.jme3.math.FastMath)
392 (defn touch-kernel
393 "Constructs a function which will return tactile sensory data from
394 'geo when called from inside a running simulation"
395 [#^Geometry geo]
396 (if-let
397 [profile (tactile-sensor-profile geo)]
398 (let [ray-reference-origins (feeler-origins geo profile)
399 ray-reference-tips (feeler-tips geo profile)
400 ray-length (tactile-scale geo)
401 current-rays (map (fn [_] (Ray.)) ray-reference-origins)
402 topology (touch-topology geo profile)
403 correction (float (* ray-length -0.2))]
405 ;; slight tolerance for very close collisions.
406 (dorun
407 (map (fn [origin tip]
408 (.addLocal origin (.mult (.subtract tip origin)
409 correction)))
410 ray-reference-origins ray-reference-tips))
411 (dorun (map #(.setLimit % ray-length) current-rays))
412 (fn [node]
413 (let [transform (.getWorldMatrix geo)]
414 (dorun
415 (map (fn [ray ref-origin ref-tip]
416 (set-ray ray transform ref-origin ref-tip))
417 current-rays ray-reference-origins
418 ray-reference-tips))
419 (vector
420 topology
421 (vec
422 (for [ray current-rays]
423 (do
424 (let [results (CollisionResults.)]
425 (.collideWith node ray results)
426 (let [touch-objects
427 (filter #(not (= geo (.getGeometry %)))
428 results)
429 limit (.getLimit ray)]
430 [(if (empty? touch-objects)
431 limit
432 (let [response
433 (apply min (map #(.getDistance %)
434 touch-objects))]
435 (FastMath/clamp
436 (float
437 (if (> response limit) (float 0.0)
438 (+ response correction)))
439 (float 0.0)
440 limit)))
441 limit])))))))))))
443 (defn touch!
444 "Endow the creature with the sense of touch. Returns a sequence of
445 functions, one for each body part with a tactile-sensor-profile,
446 each of which when called returns sensory data for that body part."
447 [#^Node creature]
448 (filter
449 (comp not nil?)
450 (map touch-kernel
451 (filter #(isa? (class %) Geometry)
452 (node-seq creature)))))
453 #+end_src
455 #+results: kernel
456 : #'cortex.touch/touch!
458 * Visualizing Touch
460 Each feeler is represented in the image as a single pixel. The
461 greyscale value of each pixel represents how deep the feeler
462 represented by that pixel is inside another object. Black means that
463 nothing is touching the feeler, while white means that the feeler is
464 completely inside another object, which is presumably flush with the
465 surface of the triangle from which the feeler originates.
467 #+name: visualization
468 #+begin_src clojure
469 (in-ns 'cortex.touch)
471 (defn touch->gray
472 "Convert a pair of [distance, max-distance] into a gray-scale pixel."
473 [distance max-distance]
474 (gray (- 255 (rem (int (* 255 (/ distance max-distance))) 256))))
476 (defn view-touch
477 "Creates a function which accepts a list of touch sensor-data and
478 displays each element to the screen."
479 []
480 (view-sense
481 (fn [[coords sensor-data]]
482 (let [image (points->image coords)]
483 (dorun
484 (for [i (range (count coords))]
485 (.setRGB image ((coords i) 0) ((coords i) 1)
486 (apply touch->gray (sensor-data i)))))
487 image))))
488 #+end_src
490 #+results: visualization
491 : #'cortex.touch/view-touch
493 * Basic Test of Touch
495 The worm's sense of touch is a bit complicated, so for this basic test
496 I'll use a new creature --- a simple cube which has touch sensors
497 evenly distributed along each of its sides.
499 #+name: test-touch-0
500 #+begin_src clojure
501 (in-ns 'cortex.test.touch)
503 (defn touch-cube []
504 (load-blender-model "Models/test-touch/touch-cube.blend"))
505 #+end_src
507 ** The Touch Cube
508 #+begin_html
509 <div class="figure">
510 <center>
511 <video controls="controls" width="500">
512 <source src="../video/touch-cube.ogg" type="video/ogg"
513 preload="none" poster="../images/aurellem-1280x480.png" />
514 </video>
515 <br> <a href="http://youtu.be/aEao4m8meAI"> YouTube </a>
516 </center>
517 <p>A simple creature with evenly distributed touch sensors.</p>
518 </div>
519 #+end_html
521 The tactile-sensor-profile image for this simple creature looks like
522 this:
524 #+attr_html: width=500
525 #+caption: The distribution of feelers along the touch-cube. The colors of the faces are irrelevant; only the white pixels specify feelers.
526 [[../images/touch-profile.png]]
528 #+name: test-touch-1
529 #+begin_src clojure
530 (in-ns 'cortex.test.touch)
532 (defn test-basic-touch
533 "Testing touch:
534 You should see a cube fall onto a table. There is a cross-shaped
535 display which reports the cube's sensation of touch. This display
536 should change when the cube hits the table, and whenever you hit
537 the cube with balls.
539 Keys:
540 <space> : fire ball"
541 ([] (test-basic-touch false))
542 ([record?]
543 (let [the-cube (doto (touch-cube) (body!))
544 touch (touch! the-cube)
545 touch-display (view-touch)]
546 (world
547 (nodify [the-cube
548 (box 10 1 10 :position (Vector3f. 0 -10 0)
549 :color ColorRGBA/Gray :mass 0)])
551 standard-debug-controls
553 (fn [world]
554 (let [timer (IsoTimer. 60)]
555 (.setTimer world timer)
556 (display-dilated-time world timer))
557 (if record?
558 (Capture/captureVideo
559 world
560 (File. "/home/r/proj/cortex/render/touch-cube/main-view/")))
561 (speed-up world)
562 (light-up-everything world))
564 (fn [world tpf]
565 (touch-display
566 (map #(% (.getRootNode world)) touch)
567 (if record?
568 (File. "/home/r/proj/cortex/render/touch-cube/touch/"))))))))
569 #+end_src
571 #+results: test-touch-1
572 : #'cortex.test.touch/test-basic-touch
574 ** Basic Touch Demonstration
576 #+begin_html
577 <div class="figure">
578 <center>
579 <video controls="controls" width="755">
580 <source src="../video/basic-touch.ogg" type="video/ogg"
581 preload="none" poster="../images/aurellem-1280x480.png" />
582 </video>
583 <br> <a href="http://youtu.be/8xNEtD-a8f0"> YouTube </a>
584 </center>
585 <p>The simple creature responds to touch.</p>
586 </div>
587 #+end_html
589 ** Generating the Basic Touch Video
590 #+name: magick4
591 #+begin_src clojure
592 (ns cortex.video.magick4
593 (:import java.io.File)
594 (:use clojure.java.shell))
596 (defn images [path]
597 (sort (rest (file-seq (File. path)))))
599 (def base "/home/r/proj/cortex/render/touch-cube/")
601 (defn pics [file]
602 (images (str base file)))
604 (defn combine-images []
605 (let [main-view (pics "main-view")
606 touch (pics "touch/0")
607 background (repeat 9001 (File. (str base "background.png")))
608 targets (map
609 #(File. (str base "out/" (format "%07d.png" %)))
610 (range 0 (count main-view)))]
611 (dorun
612 (pmap
613 (comp
614 (fn [[background main-view touch target]]
615 (println target)
616 (sh "convert"
617 touch
618 "-resize" "x300"
619 "-rotate" "180"
620 background
621 "-swap" "0,1"
622 "-geometry" "+776+129"
623 "-composite"
624 main-view "-geometry" "+66+21" "-composite"
625 target))
626 (fn [& args] (map #(.getCanonicalPath %) args)))
627 background main-view touch targets))))
628 #+end_src
630 #+begin_src sh :results silent
631 cd ~/proj/cortex/render/touch-cube/
632 ffmpeg -r 60 -i out/%07d.png -b:v 9000k -c:v libtheora basic-touch.ogg
633 #+end_src
635 #+begin_src sh :results silent
636 cd ~/proj/cortex/render/touch-cube/
637 ffmpeg -r 30 -i blender-intro/%07d.png -b:v 9000k -c:v libtheora touch-cube.ogg
638 #+end_src
640 * Adding Touch to the Worm
642 #+name: test-touch-2
643 #+begin_src clojure
644 (in-ns 'cortex.test.touch)
646 (defn test-worm-touch
647 "Testing touch:
648 You will see the worm fall onto a table. There is a display which
649 reports the worm's sense of touch. It should change when the worm
650 hits the table and when you hit it with balls.
652 Keys:
653 <space> : fire ball"
654 ([] (test-worm-touch false))
655 ([record?]
656 (let [the-worm (doto (worm) (body!))
657 touch (touch! the-worm)
658 touch-display (view-touch)]
659 (world
660 (nodify [the-worm (floor)])
661 standard-debug-controls
663 (fn [world]
664 (let [timer (IsoTimer. 60)]
665 (.setTimer world timer)
666 (display-dilated-time world timer))
667 (if record?
668 (Capture/captureVideo
669 world
670 (File. "/home/r/proj/cortex/render/worm-touch/main-view/")))
671 (speed-up world)
672 (light-up-everything world))
674 (fn [world tpf]
675 (touch-display
676 (map #(% (.getRootNode world)) touch)
677 (if record?
678 (File. "/home/r/proj/cortex/render/worm-touch/touch/"))))))))
679 #+end_src
681 #+results: test-touch-2
682 : #'cortex.test.touch/test-worm-touch
684 ** Worm Touch Demonstration
685 #+begin_html
686 <div class="figure">
687 <center>
688 <video controls="controls" width="550">
689 <source src="../video/worm-touch.ogg" type="video/ogg"
690 preload="none" poster="../images/aurellem-1280x480.png" />
691 </video>
692 <br> <a href="http://youtu.be/RHx2wqzNVcU"> YouTube </a>
693 </center>
694 <p>The worm responds to touch.</p>
695 </div>
696 #+end_html
699 ** Generating the Worm Touch Video
700 #+name: magick5
701 #+begin_src clojure
702 (ns cortex.video.magick5
703 (:import java.io.File)
704 (:use clojure.java.shell))
706 (defn images [path]
707 (sort (rest (file-seq (File. path)))))
709 (def base "/home/r/proj/cortex/render/worm-touch/")
711 (defn pics [file]
712 (images (str base file)))
714 (defn combine-images []
715 (let [main-view (pics "main-view")
716 touch (pics "touch/0")
717 targets (map
718 #(File. (str base "out/" (format "%07d.png" %)))
719 (range 0 (count main-view)))]
720 (dorun
721 (pmap
722 (comp
723 (fn [[ main-view touch target]]
724 (println target)
725 (sh "convert"
726 main-view
727 touch "-geometry" "+0+0" "-composite"
728 target))
729 (fn [& args] (map #(.getCanonicalPath %) args)))
730 main-view touch targets))))
731 #+end_src
733 #+begin_src sh :results silent
734 cd ~/proj/cortex/render/worm-touch
735 ffmpeg -r 60 -i out/%07d.png -b:v 9000k -c:v libtheora worm-touch.ogg
736 #+end_src
738 * Headers
740 #+name: touch-header
741 #+begin_src clojure
742 (ns cortex.touch
743 "Simulate the sense of touch in jMonkeyEngine3. Enables any Geometry
744 to be outfitted with touch sensors with density determined by a UV
745 image. In this way a Geometry can know what parts of itself are
746 touching nearby objects. Reads specially prepared blender files to
747 construct this sense automatically."
748 {:author "Robert McIntyre"}
749 (:use (cortex world util sense))
750 (:import (com.jme3.scene Geometry Node Mesh))
751 (:import com.jme3.collision.CollisionResults)
752 (:import com.jme3.scene.VertexBuffer$Type)
753 (:import (com.jme3.math Triangle Vector3f Vector2f Ray Matrix4f)))
754 #+end_src
756 #+name: test-touch-header
757 #+begin_src clojure
758 (ns cortex.test.touch
759 (:use (cortex world util sense body touch))
760 (:use cortex.test.body)
761 (:import (com.aurellem.capture Capture IsoTimer))
762 (:import java.io.File)
763 (:import (com.jme3.math Vector3f ColorRGBA)))
764 #+end_src
766 #+results: test-touch-header
767 : com.jme3.math.ColorRGBA
769 * Source Listing
770 - [[../src/cortex/touch.clj][cortex.touch]]
771 - [[../src/cortex/test/touch.clj][cortex.test.touch]]
772 - [[../src/cortex/video/magick4.clj][cortex.video.magick4]]
773 - [[../src/cortex/video/magick5.clj][cortex.video.magick5]]
774 - [[../assets/Models/test-touch/touch-cube.blend][touch-cube.blend]]
775 #+html: <ul> <li> <a href="../org/touch.org">This org file</a> </li> </ul>
776 - [[http://hg.bortreb.com ][source-repository]]
778 * Next
779 So far I've implemented simulated Vision, Hearing, and
780 Touch, the most obvious and prominent senses that humans
781 have. Smell and Taste shall remain unimplemented for
782 now. This accounts for the "five senses" that feature so
783 prominently in our lives. But humans have far more than the
784 five main senses. There are internal chemical senses, pain
785 (which is *not* the same as touch), heat sensitivity, and
786 our sense of balance, among others. One extra sense is so
787 important that I must implement it to have a hope of making
788 creatures that can gracefully control their own bodies. It
789 is Proprioception, which is the sense of the location of
790 each body part in relation to the other body parts.
792 Close your eyes, and touch your nose with your right index
793 finger. How did you do it? You could not see your hand, and
794 neither your hand nor your nose could use the sense of touch
795 to guide the path of your hand. There are no sound cues,
796 and Taste and Smell certainly don't provide any help. You
797 know where your hand is without your other senses because of
798 Proprioception.
800 Onward to [[./proprioception.org][proprioception]]!
802 * COMMENT Code Generation
803 #+begin_src clojure :tangle ../src/cortex/touch.clj
804 <<touch-header>>
805 <<meta-data>>
806 <<triangles-1>>
807 <<triangles-2>>
808 <<triangles-3>>
809 <<triangles-4>>
810 <<sensors>>
811 <<kernel>>
812 <<visualization>>
813 #+end_src
815 #+begin_src clojure :tangle ../src/cortex/test/touch.clj
816 <<test-touch-header>>
817 <<test-touch-0>>
818 <<test-touch-1>>
819 <<test-touch-2>>
820 #+end_src
822 #+begin_src clojure :tangle ../src/cortex/video/magick4.clj
823 <<magick4>>
824 #+end_src
826 #+begin_src clojure :tangle ../src/cortex/video/magick5.clj
827 <<magick5>>
828 #+end_src