TY - GEN
T1 - Comparative Analysis of Catastrophic Forgetting in Metric Learning
AU - Huo, Jiahao
AU - Van Zyl, Terence L.
N1 - Publisher Copyright:
© 2020 IEEE.
PY - 2020/11/14
Y1 - 2020/11/14
N2 - Catastrophic forgetting in neural networks during incremental learning remains a challenging problem. Previous research investigated the catastrophic forgetting in fully connected networks with some earlier work exploring activation functions and learning algorithms. Applications of neural networks have been extended to include similarity and metric learning. It is of significant interest to understand how metric learning loss functions would be affected by catastrophic forgetting. Our research investigates catastrophic forgetting for four well-known metric-based loss functions during incremental class learning. The loss functions are angular, contrastive, center, and triplet loss. Our results show that the rate of forgetting is different across loss functions on multiple datasets. Triplet loss was least affected followed by contrastive, center, and angular loss. Center and angular loss produce better embeddings on difficult tasks when trained on all available training data, however, they are the least robust to forgetting during incremental class learning. We argue that triplet loss provides the ideal middle ground for future improvements.
AB - Catastrophic forgetting in neural networks during incremental learning remains a challenging problem. Previous research investigated the catastrophic forgetting in fully connected networks with some earlier work exploring activation functions and learning algorithms. Applications of neural networks have been extended to include similarity and metric learning. It is of significant interest to understand how metric learning loss functions would be affected by catastrophic forgetting. Our research investigates catastrophic forgetting for four well-known metric-based loss functions during incremental class learning. The loss functions are angular, contrastive, center, and triplet loss. Our results show that the rate of forgetting is different across loss functions on multiple datasets. Triplet loss was least affected followed by contrastive, center, and angular loss. Center and angular loss produce better embeddings on difficult tasks when trained on all available training data, however, they are the least robust to forgetting during incremental class learning. We argue that triplet loss provides the ideal middle ground for future improvements.
KW - Catastrophic forgetting
KW - convolutional neural network (CNN)
KW - incremental learning
KW - metric learning
UR - http://www.scopus.com/inward/record.url?scp=85100329147&partnerID=8YFLogxK
U2 - 10.1109/ISCMI51676.2020.9311580
DO - 10.1109/ISCMI51676.2020.9311580
M3 - Conference contribution
AN - SCOPUS:85100329147
T3 - 2020 7th International Conference on Soft Computing and Machine Intelligence, ISCMI 2020
SP - 68
EP - 72
BT - 2020 7th International Conference on Soft Computing and Machine Intelligence, ISCMI 2020
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 7th International Conference on Soft Computing and Machine Intelligence, ISCMI 2020
Y2 - 14 November 2020 through 15 November 2020
ER -