PruDent: A Pruned and Confident Stacking Approach for Multi-Label Classification

Abdulaziz Alali, Miroslav Kubat

Research output: Contribution to journalArticlepeer-review

20 Scopus citations


Over the past decade or so, several research groups have addressed the problem of multi-label classification where each example can belong to more than one class at the same time. A common approach, called Binary Relevance (BR), addresses this problem by inducing a separate classifier for each class. Research has shown that this framework can be improved if mutual class dependence is exploited: an example that belongs to class X is likely to belong also to class Y; conversely, belonging to X can make an example less likely to belong to Z. Several works sought to model this information by using the vector of class labels as additional example attributes. To fill the unknown values of these attributes during prediction, existing methods resort to using outputs of other classifiers, and this makes them prone to errors. This is where our paper wants to contribute. We identified two potential ways to prune unnecessary dependencies and to reduce error-propagation in our new classifier-stacking technique, which is named PruDent. Experimental results indicate that the classification performance of PruDent compares favorably with that of other state-of-the-art approaches over a broad range of testbeds. Moreover, its computational costs grow only linearly in the number of classes.

Original languageEnglish (US)
Article number7069233
Pages (from-to)2480-2493
Number of pages14
JournalIEEE Transactions on Knowledge and Data Engineering
Issue number9
StatePublished - Sep 1 2015


  • chaining
  • label dependence
  • Multi-label classification
  • stacking

ASJC Scopus subject areas

  • Computational Theory and Mathematics
  • Information Systems
  • Computer Science Applications


Dive into the research topics of 'PruDent: A Pruned and Confident Stacking Approach for Multi-Label Classification'. Together they form a unique fingerprint.

Cite this