Abstract
Modeling a real-world system by means of a neural model involves numerous challenges that range from formulating transparent knowledge representations to obtaining reliable simulation errors. However, that knowledge is often difficult to formalize in a precise way using crisp numbers. In this paper, we present the long-term grey cognitive networks which expands the recently proposed long-term cognitive networks (LTCNs) with grey numbers. One advantage of our neural system is that it allows embedding knowledge into the network using weights and constricted neurons. In addition, we propose two procedures to construct the network in situations where only historical data are available, and a regularization method that is coupled with a nonsynaptic backpropagation algorithm. The results have shown that our proposal outperforms the LTCN model and other state-of-the-art methods in terms of accuracy.
Original language | English |
---|---|
Article number | 8718506 |
Pages (from-to) | 686-695 |
Number of pages | 10 |
Journal | IEEE Transactions on Cybernetics |
Volume | 51 |
Issue number | 2 |
DOIs | |
Publication status | Published - Feb 2021 |
Externally published | Yes |
Keywords
- Error backpropagation
- grey systems
- neural cognitive modeling
- recurrent systems