Long-term Cognitive Network-based architecture for multi-label classification

Gonzalo Nápoles*, Marilyn Bello, Yamisleydi Salgueiro

*Corresponding author for this work

Research output: Contribution to journalArticleScientificpeer-review

Abstract

This paper presents a neural system to deal with multi-label classification problems that might involve sparse features. The architecture of this model involves three sequential blocks with well-defined functions. The first block consists of a multilayered feed-forward structure that extracts hidden features, thus reducing the problem dimensionality. This block is useful when dealing with sparse problems. The second block consists of a Long-term Cognitive Network-based model that operates on features extracted by the first block. The activation rule of this recurrent neural network is modified to prevent the vanishing of the input signal during the recurrent inference process. The modified activation rule combines the neurons’ state in the previous abstract layer (iteration) with the initial state. Moreover, we add a bias component to shift the transfer functions as needed to obtain good approximations. Finally, the third block consists of an output layer that adapts the second block’s outputs to the label space. We propose a backpropagation learning algorithm that uses a squared hinge loss function to maximize the margins between labels to train this network. The results show that our model outperforms the state-of-the-art algorithms in most datasets.

Keywords

  • Long-term cognitive networks
  • Recurrent neural networks
  • Backpropagation
  • Multi-label classification

Fingerprint Dive into the research topics of 'Long-term Cognitive Network-based architecture for multi-label classification'. Together they form a unique fingerprint.

Cite this