Specialized visual sensor coupled to a dynamic neural field for embedded attentional process
Résumé
Machine learning has recently taken the leading role in machine vision through deep learning algorithms. It has brought the best results in object detection, recognition and tracking. Nevertheless, these systems are computationally expensive since they need to process the whole images from the camera for producing such results. Consequently, they require important hardware resources that limit their use for embedded applications. In the other hand, we find a more efficient mechanism in biological systems. The brain, indeed, enables an attentional process to focus on the relevant information from the environment, and hence process only a sub-part of the visual field at a time. In this work, we implement a brain-inspired attentional process through dynamic neural fields that is integrated in two types of specialized visual sensors: frame-based and event-based cameras. We compare the obtained results on tracking performances and power consumption in the context of embedded recognition and tracking.
Mots clés
specialized visual sensor
object detection
image sensors
cameras
image recognition
machine vision
embedded attentional process
dynamic neural field
neural nets
learning (artificial intelligence)
embedded applications
computer vision
brain-inspired attentional process
deep learning algorithms
machine learning
embedded systems
event-based cameras
tracking pe
biological systems