A Neural Model for Hardware Plasticity in Artificial Vision Systems
Résumé
The advent of massively parallel many-core architectures on a chip can be considered as a good opportunity to rethink the classical computation model used for several decades and that now shows some limitations to follow both the potential and the usage of new technologies.
In this paper, the way explored to study new solutions is directly inspired from biology, and more precisely from neurosciences. This way could lead, for example, to best practices for dynamically partitioning application tasks onto a set of processing cores. Distributed load-balancing is known to bring more efficient utilization of resources in the case of regular applications. We propose in this paper a bioinspiredhardware substrate that brings a plasticity property into many-core architectures.
We describe a hardware controller in which a grid of processing elements will support a set of neurocognitive processes in order to drive a robot in different tasks. We propose an original distributed hardwareartificial neural network as support for this plasticity. It is inspired by Neural Fields equations that have shown self-organizing behaviors and can be suitable for this purpose. It can be used to take allocation decisions locally, taking into account the state of the whole system through the emergent behaviour of the network. This paper describes the neural model and focuses mainly on its implementation onto FPGA in the context of artificial vision.