TY - JOUR
T1 - Knowledge Distillation Classifier Generation Network for Zero-Shot Learning
AU - Yu, Yunlong
AU - Li, Bin
AU - Ji, Zhong
AU - Han, Jungong
AU - Zhang, Zhongfei
N1 - Funding Information:
This work was supported in part by the National Natural Science Foundation of China (NSFC) under Grant 62002320, Grant U19B2043, and Grant 61672456; in part by the Fundamental Research Funds for the Central Universities under Grant 2020QNA5010; and in part by the Key Research and Development Program of Zhejiang Province, China, under Grant 2021C01119.
Publisher Copyright:
© 2012 IEEE.
PY - 2023/6/1
Y1 - 2023/6/1
N2 - In this article, we present a conceptually simple but effective framework called knowledge distillation classifier generation network (KDCGN) for zero-shot learning (ZSL), where the learning agent requires recognizing unseen classes that have no visual data for training. Different from the existing generative approaches that synthesize visual features for unseen classifiers' learning, the proposed framework directly generates classifiers for unseen classes conditioned on the corresponding class-level semantics. To ensure the generated classifiers to be discriminative to the visual features, we borrow the knowledge distillation idea to both supervise the classifier generation and distill the knowledge with, respectively, the visual classifiers and soft targets trained from a traditional classification network. Under this framework, we develop two, respectively, strategies, i.e., class augmentation and semantics guidance, to facilitate the supervision process from the perspectives of improving visual classifiers. Specifically, the class augmentation strategy incorporates some additional categories to train the visual classifiers, which regularizes the visual classifier weights to be compact, under supervision of which the generated classifiers will be more discriminative. The semantics-guidance strategy encodes the class semantics into the visual classifiers, which would facilitate the supervision process by minimizing the differences between the generated and the real-visual classifiers. To evaluate the effectiveness of the proposed framework, we have conducted extensive experiments on five datasets in image classification, i.e., AwA1, AwA2, CUB, FLO, and APY. Experimental results show that the proposed approach performs best in the traditional ZSL task and achieves a significant performance improvement on four out of the five datasets in the generalized ZSL task.
AB - In this article, we present a conceptually simple but effective framework called knowledge distillation classifier generation network (KDCGN) for zero-shot learning (ZSL), where the learning agent requires recognizing unseen classes that have no visual data for training. Different from the existing generative approaches that synthesize visual features for unseen classifiers' learning, the proposed framework directly generates classifiers for unseen classes conditioned on the corresponding class-level semantics. To ensure the generated classifiers to be discriminative to the visual features, we borrow the knowledge distillation idea to both supervise the classifier generation and distill the knowledge with, respectively, the visual classifiers and soft targets trained from a traditional classification network. Under this framework, we develop two, respectively, strategies, i.e., class augmentation and semantics guidance, to facilitate the supervision process from the perspectives of improving visual classifiers. Specifically, the class augmentation strategy incorporates some additional categories to train the visual classifiers, which regularizes the visual classifier weights to be compact, under supervision of which the generated classifiers will be more discriminative. The semantics-guidance strategy encodes the class semantics into the visual classifiers, which would facilitate the supervision process by minimizing the differences between the generated and the real-visual classifiers. To evaluate the effectiveness of the proposed framework, we have conducted extensive experiments on five datasets in image classification, i.e., AwA1, AwA2, CUB, FLO, and APY. Experimental results show that the proposed approach performs best in the traditional ZSL task and achieves a significant performance improvement on four out of the five datasets in the generalized ZSL task.
KW - Class-augmentation
KW - classifier generation
KW - knowledge distillation
KW - Knowledge engineering
KW - Learning systems
KW - Prototypes
KW - Semantics
KW - semantics-guidance (SG)
KW - Task analysis
KW - Training
KW - Visualization
KW - zero-shot learning (ZSL).
KW - zero-shot learning (ZSL)
UR - http://www.scopus.com/inward/record.url?scp=85118632554&partnerID=8YFLogxK
U2 - 10.1109/TNNLS.2021.3112229
DO - 10.1109/TNNLS.2021.3112229
M3 - Article
C2 - 34587096
AN - SCOPUS:85118632554
SN - 2162-237X
VL - 34
SP - 3183
EP - 3194
JO - IEEE Transactions on Neural Networks and Learning Systems
JF - IEEE Transactions on Neural Networks and Learning Systems
IS - 6
ER -