A Discriminant Correlation Neural Network for Feature Representation Learning
Issued Date
2025-01-01
Resource Type
ISSN
27704327
eISSN
27704319
Scopus ID
2-s2.0-105025009240
Journal Title
Proceedings of the International Conference on Multimedia Information Processing and Retrieval Mipr
Start Page
277
End Page
282
Rights Holder(s)
SCOPUS
Bibliographic Citation
Proceedings of the International Conference on Multimedia Information Processing and Retrieval Mipr (2025) , 277-282
Suggested Citation
Gao L., Liu K., Termritthikun C., Muneesawang P., Hnoohom N., Guan L. A Discriminant Correlation Neural Network for Feature Representation Learning. Proceedings of the International Conference on Multimedia Information Processing and Retrieval Mipr (2025) , 277-282. 282. doi:10.1109/MIPR67560.2025.00051 Retrieved from: https://repository.li.mahidol.ac.th/handle/123456789/113656
Title
A Discriminant Correlation Neural Network for Feature Representation Learning
Author(s)
Author's Affiliation
Corresponding Author(s)
Other Contributor(s)
Abstract
The progress of multimedia computing, particularly through deep neural networks (DNNs), has gained remarkable attention across a wide array of research domains and practical applications, such as image classification, video computing, natural language processing, among others. However, the inherent black-box characteristics and concern for sustainability of DNNs have emerged as significant challenges, especially when dealing with multi-view data. To tackle these challenges, alternative learning models have been actively explored in recent years, with green learning and interpretable learning standing out as two key directions. In this study, we explore one such model aimed at two-view feature representation learning. This proposed model leverages a perceptron-style neural network grounded in discriminant correlation analysis, which is named DC-PNN. The model incorporates statistical machine learning (SML) techniques to optimize network training. To evaluate its effectiveness and generalizability, we conduct experiments on audio emotion recognition and action recognition on two public data sets (i.e., RML emotional database and NTU RGB+D 120 database). Then, move forward to a real application, Glaucoma image classification. Experimental results demonstrate that the DC-PNN model outperforms state-of-the-art (SOTA) methods. In addition, the parameters and FLOPs analysis are implemented, validating the advantage of the model in terms of efficiency and resource requirements.
