*Result*: Explainable Model-specific Algorithm Selection for Multi-Label Classification

Title:
Explainable Model-specific Algorithm Selection for Multi-Label Classification
Contributors:
Recherche Opérationnelle (RO), LIP6, Sorbonne Université (SU)-Centre National de la Recherche Scientifique (CNRS)-Sorbonne Université (SU)-Centre National de la Recherche Scientifique (CNRS), ANR-22-ERCS-0003,VARIATION,Opérateurs de variation généralisés pour les heuristiques de recherche aléatoire(2022)
Source:
Proc. of 2022 IEEE Symposium Series on Computational Intelligence (SSCI) ; 2022 IEEE Symposium Series on Computational Intelligence (SSCI) ; https://hal.sorbonne-universite.fr/hal-04003128 ; 2022 IEEE Symposium Series on Computational Intelligence (SSCI), Dec 2022, Singapore, Singapore. pp.39-46, ⟨10.1109/SSCI51031.2022.10022177⟩
Publisher Information:
CCSD
IEEE
Publication Year:
2022
Subject Geographic:
Document Type:
*Conference* conference object
Language:
English
Relation:
info:eu-repo/semantics/altIdentifier/arxiv/2211.11227; ARXIV: 2211.11227
DOI:
10.1109/SSCI51031.2022.10022177
Rights:
info:eu-repo/semantics/OpenAccess
Accession Number:
edsbas.F191C58E
Database:
BASE

*Further Information*

*International audience ; Multi-label classification (MLC) is an ML task of predictive modeling in which a data instance can simultaneously belong to multiple classes. MLC is increasingly gaining interest in different application domains such as text mining, computer vision, and bioinformatics. Several MLC algorithms have been proposed in the literature, resulting in a meta-optimization problem that the user needs to address: which MLC approach to select for a given dataset? To address this algorithm selection problem, we investigate in this work the quality of an automated approach that uses characteristics of the datasets-so-called features-and a trained algorithm selector to choose which algorithm to apply for a given task. For our empirical evaluation, we use a portfolio of 38 datasets. We consider eight MLC algorithms, whose quality we evaluate using six different performance metrics. We show that our automated algorithm selector outperforms any of the single MLC algorithms, and this is for all evaluated performance measures. Our selection approach is explainable, a characteristic that we exploit to investigate which meta-features have the largest influence on the decisions made by the algorithm selector. Finally, we also quantify the importance of the most significant metafeatures for various domains.*