TY - JOUR
T1 - Long short-term cognitive networks
AU - Nápoles, Gonzalo
AU - Grau, Isel
AU - Jastrzębska, Agnieszka
AU - Salgueiro, Yamisleydi
N1 - Funding Information:
Y. Salgueiro would like to acknowledge the support provided by the National Center for Artificial Intelligence CENIA FB210017, Basal ANID and the super-computing infrastructure of the NLHPC (ECM-02).
Publisher Copyright:
© 2022, The Author(s).
PY - 2022/10
Y1 - 2022/10
N2 - In this paper, we present a recurrent neural system named long short-term cognitive networks (LSTCNs) as a generalization of the short-term cognitive network (STCN) model. Such a generalization is motivated by the difficulty of forecasting very long time series efficiently. The LSTCN model can be defined as a collection of STCN blocks, each processing a specific time patch of the (multivariate) time series being modeled. In this neural ensemble, each block passes information to the subsequent one in the form of weight matrices representing the prior knowledge. As a second contribution, we propose a deterministic learning algorithm to compute the learnable weights while preserving the prior knowledge resulting from previous learning processes. As a third contribution, we introduce a feature influence score as a proxy to explain the forecasting process in multivariate time series. The simulations using three case studies show that our neural system reports small forecasting errors while being significantly faster than state-of-the-art recurrent models.
AB - In this paper, we present a recurrent neural system named long short-term cognitive networks (LSTCNs) as a generalization of the short-term cognitive network (STCN) model. Such a generalization is motivated by the difficulty of forecasting very long time series efficiently. The LSTCN model can be defined as a collection of STCN blocks, each processing a specific time patch of the (multivariate) time series being modeled. In this neural ensemble, each block passes information to the subsequent one in the form of weight matrices representing the prior knowledge. As a second contribution, we propose a deterministic learning algorithm to compute the learnable weights while preserving the prior knowledge resulting from previous learning processes. As a third contribution, we introduce a feature influence score as a proxy to explain the forecasting process in multivariate time series. The simulations using three case studies show that our neural system reports small forecasting errors while being significantly faster than state-of-the-art recurrent models.
KW - Interpretability
KW - Multivariate time Series
KW - Recurrent Neural Networks
KW - Short-term Cognitive Networks
UR - http://www.scopus.com/inward/record.url?scp=85130716461&partnerID=8YFLogxK
U2 - 10.1007/s00521-022-07348-5
DO - 10.1007/s00521-022-07348-5
M3 - Article
AN - SCOPUS:85130716461
SN - 0941-0643
VL - 34
SP - 16959
EP - 16971
JO - Neural Computing and Applications
JF - Neural Computing and Applications
IS - 19
ER -