Exploring unsupervised pre-training for echo state networks
Publikation: Beitrag in Fachzeitschrift › Forschungsartikel › Beigetragen › Begutachtung
Beitragende
Abstract
Echo State Networks (ESNs) are a special type of Recurrent Neural Networks (RNNs), in which the input and recurrent connections are traditionally generated randomly, and only the output weights are trained. However, recent publications have addressed the problem that a purely random initialization may not be ideal. Instead, a completely deterministic or data-driven initialized ESN structure was proposed. In this work, an unsupervised training methodology for the hidden components of an ESN is proposed. Motivated by traditional Hidden Markov Models (HMMs), which have been widely used for speech recognition for decades, we present an unsupervised pre-training method for the recurrent weights and bias weights of ESNs. This approach allows for using unlabeled data during the training procedure and shows superior results for continuous spoken phoneme recognition, as well as for a large variety of time-series classification datasets.
Details
Originalsprache | Englisch |
---|---|
Seiten (von - bis) | 24225–24242 |
Seitenumfang | 18 |
Fachzeitschrift | Neural Computing and Applications |
Jahrgang | 35 |
Ausgabenummer | 34 |
Publikationsstatus | Veröffentlicht - Dez. 2023 |
Peer-Review-Status | Ja |
Externe IDs
Scopus | 85169611449 |
---|---|
ORCID | /0000-0002-8149-2275/work/167217049 |
Schlagworte
ASJC Scopus Sachgebiete
Schlagwörter
- Clustering, ESN, RCN, State machine