TY - JOUR
T1 - ECMEE
T2 - Expert Constrained Multi-Expert Ensembles with Category Entropy Minimization for Long-tailed Visual Recognition
AU - Fu, Yu
AU - Shang, Changjing
AU - Han, Jungong
AU - Shen, Qiang
N1 - Publisher Copyright:
© 2024 Elsevier B.V.
PY - 2024/4/1
Y1 - 2024/4/1
N2 - When the training dataset follows a long-tail distribution, models tend to prioritize the majority of the data, thus resulting in lower predictive accuracy for the minority data. Among existing methods, integrating multiple experts with different logit distributions has yielded promising results. However, the current state-of-the-art (SOTA) ensemble method, i.e., Self-supervised Aggregation of Diverse Experts, trains three expert models separately to favor the head, middle, and tail data, respectively, without imposing mutual constraints. Failure to constrain the magnitude of logits among experts may result in higher category entropy, making it difficult to achieve an optimal ensemble solution. To address this issue, we propose the Expert Constrained Multi-Expert Ensembles with Category Entropy Minimization method, which consists of two new strategies: (1) Confidence Enhancement Loss to constrain the expert models based on maximizing target and non-target logit margins, thereby minimizing category entropy; (2) Shot-aware Weights associated with expert models to accommodate the shot-headed characteristic of the experts. Experiments demonstrate that our method effectively reduces expert category entropy, improves integration effectiveness, and achieves SOTA results on three datasets in diverse test distributions.
AB - When the training dataset follows a long-tail distribution, models tend to prioritize the majority of the data, thus resulting in lower predictive accuracy for the minority data. Among existing methods, integrating multiple experts with different logit distributions has yielded promising results. However, the current state-of-the-art (SOTA) ensemble method, i.e., Self-supervised Aggregation of Diverse Experts, trains three expert models separately to favor the head, middle, and tail data, respectively, without imposing mutual constraints. Failure to constrain the magnitude of logits among experts may result in higher category entropy, making it difficult to achieve an optimal ensemble solution. To address this issue, we propose the Expert Constrained Multi-Expert Ensembles with Category Entropy Minimization method, which consists of two new strategies: (1) Confidence Enhancement Loss to constrain the expert models based on maximizing target and non-target logit margins, thereby minimizing category entropy; (2) Shot-aware Weights associated with expert models to accommodate the shot-headed characteristic of the experts. Experiments demonstrate that our method effectively reduces expert category entropy, improves integration effectiveness, and achieves SOTA results on three datasets in diverse test distributions.
KW - Long-tail
KW - Expert Ensemble
KW - Visual Recognition
KW - Entropy Minimization
KW - Self-supervised Learning
UR - http://www.scopus.com/inward/record.url?scp=85184077797&partnerID=8YFLogxK
U2 - 10.1016/j.neucom.2024.127357
DO - 10.1016/j.neucom.2024.127357
M3 - Article
AN - SCOPUS:85184077797
SN - 0925-2312
VL - 576
JO - Neurocomputing
JF - Neurocomputing
M1 - 127357
ER -