Delay-Based Reservoir Computing: Noise Effects in a Combined Analog and Digital Implementation

M.c. Soriano, Silvia Ortin, Lars Keuninckx, Lennert Appeltant, Jan Danckaert, L. Pesquera, Guy Van Der Sande

Research output: Contribution to journalArticlepeer-review

137 Citations (Scopus)

Abstract

Reservoir computing is a paradigm in machine learning whose processing capabilities rely on the dynamical behavior of recurrent neural networks. We present a mixed analog and digital implementation of this concept with a nonlinear analog electronic circuit as a main computational unit. In our approach, the reservoir network can be replaced by a single nonlinear element with delay via time-multiplexing. We analyze the influence of noise on the performance of the system for two benchmark tasks: 1) a classification problem and 2) a chaotic time-series prediction task. Special attention is given to the role of quantization noise, which is studied by varying the resolution in the conversion interface between the analog and digital worlds.
Original languageEnglish
Pages (from-to)388-393
JournalIEEE Transactions on Neural Networks and Learning Systems
Volume26
Issue number2
DOIs
Publication statusPublished - 2 Feb 2015

Keywords

  • delay systems
  • reservoir computing

Fingerprint

Dive into the research topics of 'Delay-Based Reservoir Computing: Noise Effects in a Combined Analog and Digital Implementation'. Together they form a unique fingerprint.

Cite this