Monthly Archives: June 2011
APRON was involved with several projects at Telluride 2010. In particular it was used for a sensory fusion project, where a robot had to learn the relationships between visual, audio and motor spaces. This short clip is the robot learning … Continue reading
A quick video showing some of the Playstation Eye capabilities being used in APRON. The screen capture software is a little slow (and consumes a great deal of CPU), but frame rates at QVGA can easily reach 180 FPS, even … Continue reading
I’ve added support for the Playstation EYE camera. Why? Good question! This cheap little camera actually packs a powerful punch, and could be compared with expensive industrial vision sensors. It can capture images of 320×240 at around 180FPS (!), it … Continue reading
APRON is capable of performing spiking neural network simulations. Here we see 6 interconnected layers of Izhikevich neurons, with various projections between the layers. The input stimulus is from a webcam. Although I’ve no idea what the model is actually … Continue reading
APRON can also handle colour images. Those extra two dimensions of data can be really handy for visualisation and segmentation. Check out the video below. Also shown are some more features of the APRON environment.
Here is a video of APRON executing a self organising map. It’s trivialised to highlight certain features of the APRON simulation environment. The main feature here, is that APRON can explode developing receptive fields implemented with LinkMaps, so you can … Continue reading
APRON is really good at array processing, and sometimes it can be used for applications other than image processing. Check out the video below, showing the iterative calculation of a Mandelbrot set.
Check out this video of real-time optic flow being calculated in the APRON environment. The approach is a basic block-matching algorithm, but the APRON environment allows you to interactively analyse and debug the running algorithm.