n this article, we apply a bio-inspired control architecture to a roving robot performing different tasks. The key of the control system is the perceptual core, where heterogeneous information coming from sensors is merged to build an internal portrait representing the current situation of the environment. The internal representation triggers an action as the response to the current stimuli, closing the loop between the agent and the external world. The robot's internal state is implemented through a nonlinear lattice of neuron cells, allowing the generation of a large amount of emergent steady-state solutions in the form of Turing patterns. These are incrementally shaped, through learning, so as to constitute a “mirror” of the environmental conditions. Reaction—diffusion cellular nonlinear networks were chosen to generate Turing patterns as internal representations of the robot surroundings. The associations between incoming sensations and the perceptual core, and between Turing patterns and actions to be performed, are driven by two reward-based learning mechanisms. We report on simulation results and experiments on a roving robot to show the suitability of the approach.
|Titolo:||Perception for action: Dynamic spatiotemporal patterns applied on a roving robot|
|Data di pubblicazione:||2008|
|Appare nelle tipologie:||1.1 Articolo in rivista|