Conference paper Open Access
Ali Dabbous; Michele Mastella; Aishwarya Natarajan; Elisabetta Chicca; Maurizio Valle; Chiara Bartolozzi
Robots and users of hand prosthesis could easily manipulate objects if endowed with the sense of touch. Towards this goal, information about touched objects and surfaces has to be inferred from raw data coming from the sensors. An important cue for objects discrimination is the orientation of edges, that is used both in artificial vision and touch as pre-processing stage. We present a spiking neural network, inspired on the encoding of edges in human first order tactile afferents. The network uses three layers of Leaky Integrate and Fire neurons to distinguish different edge orientations of a bar pressed on the artificial skin of the iCub robot. The architecture is successfully able to discriminate eight different orientations (from 0 to 180), by implementing a structured model of overlapping receptive fields. We demonstrate that the network can learn the appropriate connectivity through unsupervised spike based learning, and that the number and spatial distribution of sensitive areas within the receptive fields are important in edge orientation discrimination.
|Data volume||25.8 MB|