Abstract
Speech recognition has become an important task to improve the human-machine interface. Taking into account the limitations of current automatic speech recognition systems, like non-real time cloud-based solutions or power demand, recent interest for neural networks and bio-inspired systems has motivated the implementation of new techniques. Among them, a combination of spiking neural networks and neuromorphic auditory sensors offer an alternative to carry out the human-like speech processing task. In this approach, a spiking convolutional neural network model was implemented, in which the weights of connections were calculated by training a convolutional neural network with specific activation functions, using firing rate-based static images with the spiking information obtained from a neuromorphic cochlea. The system was trained and tested with a large dataset that contains 'left' and 'right' speech commands, achieving 89.90% accuracy. A novel spiking neural network model has been proposed to adapt the network that has been trained with static images to a non-static processing approach, making it possible to classify audio signals and time series in real time.
Original language | English |
---|---|
Title of host publication | 2018 International Joint Conference on Neural Networks, IJCNN 2018 - Proceedings |
Publisher | IEEE |
Volume | 2018-July |
ISBN (Electronic) | 9781509060146 |
DOIs | |
Publication status | Published - 10 Oct 2018 |
Event | 2018 International Joint Conference on Neural Networks - Rio de Janeiro, Brazil Duration: 8 Jul 2018 → 13 Jul 2018 |
Conference
Conference | 2018 International Joint Conference on Neural Networks |
---|---|
Abbreviated title | IJCNN 2018 |
Country/Territory | Brazil |
City | Rio de Janeiro |
Period | 8/07/18 → 13/07/18 |
Keywords
- audio processing
- Convolutional Neural Networks
- deep learning
- neuromorphic hardware
- speech recognition
- Spiking Neural Networks