Login | Register

Development of a Multi-Scale Model for Enhanced Performance by Using a Modified Soft-Margin SoftMax Loss and its Application for Facial Expression Recognition

Title:

Development of a Multi-Scale Model for Enhanced Performance by Using a Modified Soft-Margin SoftMax Loss and its Application for Facial Expression Recognition

Nabaei, Armin (2024) Development of a Multi-Scale Model for Enhanced Performance by Using a Modified Soft-Margin SoftMax Loss and its Application for Facial Expression Recognition. Masters thesis, Concordia University.

[thumbnail of NABAEI_MASc_F2024 .pdf]
Preview
Text (application/pdf)
NABAEI_MASc_F2024 .pdf - Accepted Version
Available under License Spectrum Terms of Access.
6MB

Abstract

This thesis introduces an improved Convolutional Neural Network (CNN) model that selectively excludes irrelevant elements from input vectors to extract distinct features. The model employs Convolution (Conv) to encode data information and Deconvolution (Deconv) layers to reconstruct the spatial dimensions of feature maps and utilizing shortcut connections to exploit applicable sparsity and capture comprehensive and detailed information. This research addresses the limitations of the traditional SoftMax Loss function, which tends to overfit by wrongly classifying due to shortage of discriminability through integrating a regularization technique to the conventional Cross-Entropy loss and introducing an adaptive-margin to the standard SoftMax function. The adjusted SoftMax Loss is designed to enhance the separation of different embedding vectors and tighten clusters of similar ones. This adjustment results in boosting both the diversity between different classes and the similarity within the same class. These modifications aim to elevate the proposed model’s accuracy, and evaluations are benchmarked against existing methods using well-known datasets such as CIFAR10 , MNIST, and SVHN. Lastly, the application of the model is fine-tunned in the domain of facial expression recognition (FER). This work asserts the model's advanced capabilities over state-of-the-art methods in FER and demonstrated its effectiveness through superior accuracy on three established datasets: FER-2013, RAF-DB, and CK+.

Divisions:Concordia University > Gina Cody School of Engineering and Computer Science > Electrical and Computer Engineering
Item Type:Thesis (Masters)
Authors:Nabaei, Armin
Institution:Concordia University
Degree Name:M.A. Sc.
Program:Electrical and Computer Engineering
Date:15 April 2024
Thesis Supervisor(s):Kabir, M. Zahangir
Keywords:Machine learning, Deep learning, Learning algorithm, Convolutional neural network
ID Code:993999
Deposited By: armin nabaei
Deposited On:24 Oct 2024 16:48
Last Modified:24 Oct 2024 16:48

References:

[1] Y. LeCun, L. Bottou, Y. Bengio, and P. Ha, “Gradient-Based Learning Applied to Document Recognition,” p. 46, 1998.
[2] “Generalized cross entropy loss for training deep neural networks with noisy labels | Proceedings of the 32nd International Conference on Neural Information Processing Systems.” Accessed: Apr. 23, 2024. [Online]. Available: https://dl.acm.org/doi/10.5555/3327546.3327555
[3] “U-Net and Its Variants for Medical Image Segmentation: A Review of Theory and Applications | IEEE Journals & Magazine | IEEE Xplore.” Accessed: Dec. 19, 2022. [Online]. Available: https://ieeexplore.ieee.org/abstract/document/9446143
[4] “(4) (PDF) A Multigrid Tutorial, 2nd Edition.” Accessed: Dec. 19, 2022. [Online]. Available: https://www.researchgate.net/publication/220690328_A_Multigrid_Tutorial_2nd_Edition
[5] R. Szeliski, “Fast surface interpolation using hierarchical basis functions,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 12, no. 6, pp. 513–528, Jun. 1990, doi: 10.1109/34.56188.
[6] O. Ronneberger, P. Fischer, and T. Brox, “U-Net: Convolutional Networks for Biomedical Image Segmentation,” in Medical Image Computing and Computer-Assisted Intervention – MICCAI 2015, N. Navab, J. Hornegger, W. M. Wells, and A. F. Frangi, Eds., in Lecture Notes in Computer Science. Cham: Springer International Publishing, 2015, pp. 234–241. doi: 10.1007/978-3-319-24574-4_28.
[7] “Deep Learning.” Accessed: Jan. 18, 2023. [Online]. Available: https://www.deeplearningbook.org/
[8] A. Dubey, O. Gupta, R. Raskar, and N. Naik, “Maximum-Entropy Fine Grained Classification,” in Advances in Neural Information Processing Systems, Curran Associates, Inc., 2018. Accessed: Mar. 19, 2024. [Online]. Available: https://papers.nips.cc/paper_files/paper/2018/hash/0c74b7f78409a4022a2c4c5a5ca3ee19-Abstract.html
[9] “Visualizing the loss landscape of neural nets | Proceedings of the 32nd International Conference on Neural Information Processing Systems.” Accessed: Mar. 20, 2024. [Online]. Available: https://dl.acm.org/doi/10.5555/3327345.3327535
[10] “The sample complexity of pattern classification with neural networks: the size of the weights is more important than the size of the network | IEEE Journals & Magazine | IEEE Xplore.” Accessed: Jan. 18, 2023. [Online]. Available: https://ieeexplore.ieee.org/document/661502
[11] “Angular Margin Softmax Loss and Its Variants for Double Compressed AMR Audio Detection | Proceedings of the 2021 ACM Workshop on Information Hiding and Multimedia Security.” Accessed: Feb. 01, 2023. [Online]. Available: https://dl.acm.org/doi/10.1145/3437880.3460414
[12] “[PDF] An Empirical Analysis of Deep Network Loss Surfaces | Semantic Scholar.” Accessed: Apr. 23, 2024. [Online]. Available: https://www.semanticscholar.org/paper/An-Empirical-Analysis-of-Deep-Network-Loss-Surfaces-Im-Tao/815ae2909fd7fc536afa9773228dab40872d5cb7
[13] “Focal Loss for Dense Object Detection | IEEE Journals & Magazine | IEEE Xplore.” Accessed: Apr. 23, 2024. [Online]. Available: https://ieeexplore.ieee.org/document/8417976
[14] D. Zhou et al., “Dynamic Margin Softmax Loss for Speaker Verification,” in Interspeech 2020, ISCA, Oct. 2020, pp. 3800–3804. doi: 10.21437/Interspeech.2020-1106.
[15] G. Hinton and S. Roweis, “Stochastic Neighbor Embedding”.
[16] S. Khan, H. Rahmani, S. A. A. Shah, and M. Bennamoun, A Guide to Convolutional Neural Networks for Computer Vision. in Synthesis Lectures on Computer Vision. Cham: Springer International Publishing, 2018. doi: 10.1007/978-3-031-01821-3.
[17] G. E. Hinton, N. Srivastava, A. Krizhevsky, I. Sutskever, and R. Salakhutdinov, “Improving neural networks by preventing co-adaptation of feature detectors,” CoRR, vol. abs/1207.0580, 2012, Accessed: Apr. 23, 2024.
[18] J. Donahue et al., “DeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition,” in Proceedings of the 31st International Conference on Machine Learning, PMLR, Jan. 2014, pp. 647–655. Accessed: Mar. 05, 2024. [Online]. Available: https://proceedings. Generalized Cross Entropy Loss for Training Deep Neural mlr.press/v32/donahue14.html
[19] “Visualizing and Understanding Convolutional Networks | SpringerLink.” Accessed: Apr. 23, 2024. [Online]. Available: https://link-springer-com.lib-ezproxy.concordia.ca/chapter/10.1007/978-3-319-10590-1_53
[20] N. Srivastava, G. Hinton, A. Krizhevsky, I. Sutskever, and R. Salakhutdinov, “Dropout: A Simple Way to Prevent Neural Networks from Overfitting,” J. Mach. Learn. Res., vol. 15, no. 56, pp. 1929–1958, 2014.
[21] J. Kukacka, V. Golkov, and D. Cremers, “Regularization for Deep Learning: A Taxonomy,” CoRR, vol. abs/1710.10686, 2017, Accessed: Apr. 16, 2024.
[22] “Training with Noise is Equivalent to Tikhonov Regularization | MIT Press Journals & Magazine | IEEE Xplore.” Accessed: Apr. 16, 2024. [Online]. Available: https://ieeexplore.ieee.org/abstract/document/6796505
[23] G. Huang, Z. Liu, L. V. D. Maaten, and K. Q. Weinberger, “Densely Connected Convolutional Networks,” presented at the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), IEEE Computer Society, Jul. 2017, pp. 2261–2269. doi: 10.1109/CVPR.2017.243.
[24] “Deep Residual Learning for Image Recognition | IEEE Conference Publication | IEEE Xplore.” Accessed: Mar. 05, 2024. [Online]. Available: https://ieeexplore.ieee.org/document/7780459
[25] C. Szegedy et al., “Going deeper with convolutions,” presented at the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), IEEE Computer Society, Jun. 2015, pp. 1–9. doi: 10.1109/CVPR.2015.7298594.
[26] “[PDF] Network In Network | Semantic Scholar.” Accessed: Apr. 23, 2024. [Online]. Available: https://www.semanticscholar.org/paper/Network-In-Network-Lin-Chen/5e83ab70d0cbc003471e87ec306d27d9c80ecb16
[27] C.-Y. Lee, S. Xie, P. Gallagher, Z. Zhang, and Z. Tu, “Deeply-Supervised Nets,” in Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics, PMLR, Feb. 2015, pp. 562–570. Accessed: Apr. 23, 2024. [Online]. Available: https://proceedings.mlr.press/v38/lee15a.html
[28] “Inception-v4, inception-ResNet and the impact of residual connections on learning | Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence.” Accessed: Apr. 23, 2024. [Online]. Available: https://dl.acm.org/doi/10.5555/3298023.3298188
[29] “Pattern Recognition and Neural Networks.” Accessed: Dec. 19, 2022. [Online]. Available: https://www.cambridge.org/core/books/pattern-recognition-and-neural-networks/4E038249C9BAA06C8F4EE6F044D09C5C
[30] “(PDF) Highway Networks.” Accessed: Apr. 23, 2024. [Online]. Available: https://www.researchgate.net/publication/275897262_Highway_Networks
[31] “Training very deep networks | Proceedings of the 28th International Conference on Neural Information Processing Systems - Volume 2.” Accessed: Apr. 23, 2024. [Online]. Available: https://dl.acm.org/doi/10.5555/2969442.2969505
[32] “Deep Residual Learning for Image Recognition | IEEE Conference Publication | IEEE Xplore.” Accessed: Mar. 05, 2024. [Online]. Available: https://ieeexplore.ieee.org/document/7780459
[33] “[PDF] Attention U-Net: Learning Where to Look for the Pancreas | Semantic Scholar.” Accessed: Apr. 23, 2024. [Online]. Available: https://www.semanticscholar.org/paper/Attention-U-Net%3A-Learning-Where-to-Look-for-the-Oktay-Schlemper/ae1c89817a3a239e5344293138bdd80293983460
[34] N. S. Punn and S. Agarwal, “Inception U-Net Architecture for Semantic Segmentation to Identify Nuclei in Microscopy Cell Images,” ACM Trans. Multimed. Comput. Commun. Appl., vol. 16, no. 1, p. 12:1-12:15, Feb. 2020, doi: 10.1145/3376922.
[35] H. Zhang and J. Ma, “Hartley Spectral Pooling for Deep Learning.” Oct. 08, 2020. doi: 10.4208/csiam-am.2020.
[36] Z. Zhang, Q. Liu, and Y. Wang, “Road Extraction by Deep Residual U-Net,” IEEE Geosci. Remote Sens. Lett., vol. 15, no. 5, pp. 749–753, May 2018, doi: 10.1109/LGRS.2018.2802944.
[37] Z. Zhou, M. M. Rahman Siddiquee, N. Tajbakhsh, and J. Liang, “UNet++: A Nested U-Net Architecture for Medical Image Segmentation,” in Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, D. Stoyanov, Z. Taylor, G. Carneiro, T. Syeda-Mahmood, A. Martel, L. Maier-Hein, J. M. R. S. Tavares, A. Bradley, J. P. Papa, V. Belagiannis, J. C. Nascimento, Z. Lu, S. Conjeti, M. Moradi, H. Greenspan, and A. Madabhushi, Eds., Cham: Springer International Publishing, 2018, pp. 3–11. doi: 10.1007/978-3-030-00889-5_1.
[38] “[PDF] OverFeat: Integrated Recognition, Localization and Detection using Convolutional Networks | Semantic Scholar.” Accessed: Apr. 23, 2024. [Online]. Available: https://www.semanticscholar.org/paper/OverFeat%3A-Integrated-Recognition%2C-Localization-and-Sermanet-Eigen/1109b663453e78a59e4f66446d71720ac58cec25
[39] “Ramp Loss Linear Programming Support Vector Machine.” Accessed: Mar. 17, 2024. [Online]. Available: https://www.jmlr.org/papers/v15/huang14a.html
[40] “On the dynamics under the unhinged loss and beyond | The Journal of Machine Learning Research.” Accessed: Mar. 17, 2024. [Online]. Available: https://dl.acm.org/doi/10.5555/3648699.3649075
[41] B. van Rooyen, A. Menon, and R. C. Williamson, “Learning with Symmetric Label Noise: The Importance of Being Unhinged,” in Advances in Neural Information Processing Systems, Curran Associates, Inc., 2015. Accessed: Jan. 18, 2023. [Online]. Available: https://papers.nips.cc/paper/2015/hash/45c48cce2e2d7fbdea1afc51c7c6ad26-Abstract.html
[42] C. Bishop, “Neural networks for pattern recognition,” 1995. Accessed: Jan. 18, 2023. [Online]. Available: https://www.semanticscholar.org/paper/Neural-networks-for-pattern-recognition-Bishop/b9b1b1654ce0eea729c4160bfedcbb3246460b1d
[43] “The Real-World-Weight Cross-Entropy Loss Function: Modeling the Costs of Mislabeling | IEEE Journals & Magazine | IEEE Xplore.” Accessed: Jan. 31, 2023. [Online]. Available: https://ieeexplore.ieee.org/document/8943952
[44] R. Machlev, Y. Levron, and Y. Beck, “Modified Cross-Entropy Method for Classification of Events in NILM Systems,” IEEE Trans. Smart Grid, vol. 10, no. 5, pp. 4962–4973, Sep. 2019, doi: 10.1109/TSG.2018.2871620.
[45] B. Shan and Y. Fang, “A Cross Entropy Based Deep Neural Network Model for Road Extraction from Satellite Images,” Entropy, vol. 22, p. 535, May 2020, doi: 10.3390/e22050535.
[46] P. Li et al., “An improved categorical cross entropy for remote sensing image classification based on noisy labels,” Expert Syst. Appl., vol. 205, p. 117296, Nov. 2022, doi: 10.1016/j.eswa.2022.117296.
[47] “Normalized loss functions for deep learning with noisy labels | Proceedings of the 37th International Conference on Machine Learning.” Accessed: Apr. 23, 2024. [Online]. Available: https://dl.acm.org/doi/abs/10.5555/3524938.3525545
[48] “[PDF] Mixed Cross Entropy Loss for Neural Machine Translation | Semantic Scholar.” Accessed: Apr. 23, 2024. [Online]. Available: https://www.semanticscholar.org/paper/Mixed-Cross-Entropy-Loss-for-Neural-Machine-Li-Lu/e8c7a43c7505d4990f4596b1ecc35f01e365a59e
[49] “Dynamically Weighted Balanced Loss: Class Imbalanced Learning and Confidence Calibration of Deep Neural Networks | IEEE Journals & Magazine | IEEE Xplore.” Accessed: Jan. 31, 2023. [Online]. Available: https://ieeexplore.ieee.org/document/9324926
[50] P. Goyal, “Shallow SegNet with bilinear interpolation and weighted cross-entropy loss for Semantic segmentation of brain tissue,” in 2022 IEEE International Conference on Signal Processing, Informatics, Communication and Energy Systems (SPICES), Mar. 2022, pp. 361–365. doi: 10.1109/SPICES52834.2022.9774193.
[51] K. R. M. Fernando and C. P. Tsokos, “Dynamically Weighted Balanced Loss: Class Imbalanced Learning and Confidence Calibration of Deep Neural Networks,” IEEE Trans. Neural Netw. Learn. Syst., vol. 33, no. 7, pp. 2940–2951, Jul. 2022, doi: 10.1109/TNNLS.2020.3047335.
[52] “Improved Categorical Cross-Entropy Loss for Training Deep Neural Networks with Noisy Labels | SpringerLink.” Accessed: Jan. 31, 2023. [Online]. Available: https://link.springer.com/chapter/10.1007/978-3-030-88013-2_7
[53] H. Takeda, S. Yoshida, and M. Muneyasu, “Learning from Noisy Labeled Data Using Symmetric Cross-Entropy Loss for Image Classification,” in 2020 IEEE 9th Global Conference on Consumer Electronics (GCCE), Oct. 2020, pp. 709–711. doi: 10.1109/GCCE50665.2020.9291873.
[54] “Large-margin softmax loss for convolutional neural networks | Proceedings of the 33rd International Conference on International Conference on Machine Learning - Volume 48.” Accessed: Apr. 06, 2024. [Online]. Available: https://dl.acm.org/doi/10.5555/3045390.3045445
[55] “[PDF] L2-constrained Softmax Loss for Discriminative Face Verification | Semantic Scholar.” Accessed: Apr. 06, 2024. [Online]. Available: https://www.semanticscholar.org/paper/L2-constrained-Softmax-Loss-for-Discriminative-Face-Ranjan-Castillo/57a807f65e61bbba0ea3ba02572cc5843ecef2b6
[56] “SphereFace: Deep Hypersphere Embedding for Face Recognition | IEEE Conference Publication | IEEE Xplore.” Accessed: Mar. 18, 2024. [Online]. Available: https://ieeexplore.ieee.org/document/8100196
[57] “Additive Margin Softmax for Face Verification | IEEE Journals & Magazine | IEEE Xplore.” Accessed: Feb. 01, 2023. [Online]. Available: https://ieeexplore.ieee.org/document/8331118
[58] T. Kobayashi, “Large Margin In Softmax Cross-Entropy Loss,” presented at the British Machine Vision Conference, 2019. Accessed: Feb. 01, 2023. [Online]. Available: https://www.semanticscholar.org/paper/Large-Margin-In-Softmax-Cross-Entropy-Loss-Kobayashi/0c88506b5e6091906e656dfeb9ee1060946d4aab
[59] X. Xiang, S. Wang, H. Huang, Y. Qian, and K. Yu, “Margin Matters: Towards More Discriminative Deep Neural Network Embeddings for Speaker Recognition,” 2019 Asia-Pac. Signal Inf. Process. Assoc. Annu. Summit Conf. APSIPA ASC, pp. 1652–1656, Nov. 2019, doi: 10.1109/APSIPAASC47483.2019.9023039.
[60] F. Gao, B. Li, L. Chen, Z. Shang, X. Wei, and C. He, “A softmax classifier for high-precision classification of ultrasonic similar signals,” Ultrasonics, vol. 112, p. 106344, Apr. 2021, doi: 10.1016/j.ultras.2020.106344.
[61] “ArcFace: Additive Angular Margin Loss for Deep Face Recognition | IEEE Conference Publication | IEEE Xplore.” Accessed: Feb. 19, 2024. [Online]. Available: https://ieeexplore.ieee.org/document/8953658
[62] H. Wang et al., “CosFace: Large Margin Cosine Loss for Deep Face Recognition,” presented at the Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 5265–5274. Accessed: Feb. 01, 2023. [Online]. Available: https://openaccess.thecvf.com/content_cvpr_2018/html/Wang_CosFace_Large_Margin_CVPR_2018_paper.html
[63] S. Zhou, C. Chen, G. Han, and X. Hou, “Double Additive Margin Softmax Loss for Face Recognition,” Appl. Sci., vol. 10, p. 60, Dec. 2019, doi: 10.3390/app10010060.
[64] H. Wang, H. Huang, Y. Hu, M. Anderson, P. Rollins, and F. Makedon, “Emotion detection via discriminative kernel method,” in Proceedings of the 3rd International Conference on PErvasive Technologies Related to Assistive Environments, in PETRA ’10. New York, NY, USA: Association for Computing Machinery, Jun. 2010, pp. 1–7. doi: 10.1145/1839294.1839303.
[65] F. Chen, Z. Wang, Z. Xu, J. Xiao, and G. Wang, “Facial Expression Recognition Using Wavelet Transform and Neural Network Ensemble,” in 2008 Second International Symposium on Intelligent Information Technology Application, Dec. 2008, pp. 871–875. doi: 10.1109/IITA.2008.24.
[66] N. Jain, S. Kumar, A. Kumar, P. Shamsolmoali, and M. Zareapoor, “Hybrid deep neural networks for face emotion recognition,” Pattern Recognit. Lett., vol. 115, pp. 101–106, Nov. 2018, doi: 10.1016/j.patrec.2018.04.010.
[67] L. Ma and K. Khorasani, “Facial expression recognition using constructive feedforward neural networks,” IEEE Trans. Syst. Man Cybern. Part B Cybern., vol. 34, no. 3, pp. 1588–1595, Jun. 2004, doi: 10.1109/TSMCB.2004.825930.
[68] “Papers with Code - FER2013 Benchmark (Facial Expression Recognition).” Accessed: Feb. 13, 2023. [Online]. Available: https://paperswithcode.com/sota/facial-expression-recognition-on-fer2013
[69] M.-I. Georgescu, R. T. Ionescu, and M. Popescu, “Local Learning with Deep and Handcrafted Features for Facial Expression Recognition,” IEEE Access, vol. 7, pp. 64827–64836, 2019, doi: 10.1109/ACCESS.2019.2917266.
[70] Y. Zhang, C. Wang, and W. Deng, “Relative Uncertainty Learning for Facial Expression Recognition,” in Advances in Neural Information Processing Systems, Curran Associates, Inc., 2021, pp. 17616–17627. Accessed: Nov. 13, 2022. [Online]. Available: https://proceedings.neurips.cc/paper/2021/hash/9332c513ef44b682e9347822c2e457ac-Abstract.html
[71] “Facial Expression Recognition in the Wild via Deep Attentive Center Loss | IEEE Conference Publication | IEEE Xplore.” Accessed: Nov. 13, 2022. [Online]. Available: https://ieeexplore.ieee.org/document/9423267
[72] “Learn from All: Erasing Attention Consistency for Noisy Label Facial Expression Recognition | Computer Vision – ECCV 2022.” Accessed: Apr. 23, 2024. [Online]. Available: https://dl.acm.org/doi/abs/10.1007/978-3-031-19809-0_24
[73] Z. Zhong, L. Zheng, G. Kang, S. Li, and Y. Yang, “Random Erasing Data Augmentation,” in The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, February 7-12, 2020, AAAI Press, 2020, pp. 13001–13008. doi: 10.1609/AAAI.V34I07.7000.

[74] A. P. Fard and M. H. Mahoor, “Ad-Corre: Adaptive Correlation-Based Loss for Facial Expression Recognition in the Wild,” IEEE Access, vol. 10, pp. 26756–26768, 2022, doi: 10.1109/ACCESS.2022.3156598.
[75] H. Ding, S. K. Zhou, and R. Chellappa, “FaceNet2ExpNet: Regularizing a Deep Face Recognition Net for Expression Recognition,” presented at the 2017 12th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2017), IEEE Computer Society, May 2017, pp. 118–126. doi: 10.1109/FG.2017.23.
[76] “Classifying emotions and engagement in online learning based on a single facial expression recognition neural network | IEEE Journals & Magazine | IEEE Xplore.” Accessed: Nov. 14, 2022. [Online]. Available: https://ieeexplore.ieee.org/document/9815154
[77] “[PDF] Facial Emotion Recognition: State of the Art Performance on FER2013 | Semantic Scholar.” Accessed: Apr. 23, 2024. [Online]. Available: https://www.semanticscholar.org/paper/Facial-Emotion-Recognition%3A-State-of-the-Art-on-Khaireddin-Chen/fac01fc5a47e92f54c7f719bf3c65d36e00c95ac
[78] “Convolutional Neural Network Hyperparameters optimization for Facial Emotion Recognition | IEEE Conference Publication | IEEE Xplore.” Accessed: Nov. 14, 2022. [Online]. Available: https://ieeexplore.ieee.org/document/9425073
[79] “IVADO Digital Futures 2024 | IVADO.” Accessed: Feb. 29, 2024. [Online]. Available: https://ivado.ca/en/events/ivado-digital-futures-2024/
[80] “17th ICCSIT 2024 | Computer Science and Information Technology.” Accessed: Mar. 11, 2024. [Online]. Available: https://www.iccsit.org/2022.html
[81] “GSR 2022.” Accessed: Mar. 11, 2024. [Online]. Available: https://sites.google.com/view/ece-gsr-2022/home
[82] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “ImageNet Classification with Deep Convolutional Neural Networks,” in Advances in Neural Information Processing Systems, Curran Associates, Inc., 2012. Accessed: Dec. 31, 2022. [Online]. Available: https://papers.nips.cc/paper/2012/hash/c399862d3b9d6b76c8436e924a68c45b-Abstract.html
[83] S. Ioffe and C. Szegedy, “Batch normalization: accelerating deep network training by reducing internal covariate shift,” in Proceedings of the 32nd International Conference on International Conference on Machine Learning - Volume 37, in ICML’15. Lille, France: JMLR.org, Jul. 2015, pp. 448–456.
[84] Y. Netzer, T. Wang, A. Coates, A. Bissacco, B. Wu, and A. Ng, “Reading Digits in Natural Images with Unsupervised Feature Learning,” 2011. Accessed: Feb. 28, 2024. [Online]. Available: https://www.semanticscholar.org/paper/Reading-Digits-in-Natural-Images-with-Unsupervised-Netzer-Wang/02227c94dd41fe0b439e050d377b0beb5d427cda
[85] R. R. Selvaraju, M. Cogswell, A. Das, R. Vedantam, D. Parikh, and D. Batra, “Grad-CAM: Visual Explanations from Deep Networks via Gradient-based Localization,” Int. J. Comput. Vis., vol. 128, no. 2, pp. 336–359, Feb. 2020, doi: 10.1007/s11263-019-01228-7.
[86] “Learning Deep Global Multi-Scale and Local Attention Features for Facial Expression Recognition in the Wild | IEEE Journals & Magazine | IEEE Xplore.” Accessed: Feb. 28, 2024. [Online]. Available: https://ieeexplore.ieee.org/document/9474949
[87] Z. Zhao, Q. Liu, and F. Zhou, “Robust Lightweight Facial Expression Recognition Network with Label Distribution Training,” Proc. AAAI Conf. Artif. Intell., vol. 35, no. 4, Art. no. 4, May 2021, doi: 10.1609/aaai.v35i4.16465.
[88] “EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks.” Accessed: Mar. 12, 2024. [Online]. Available: https://www.researchgate.net/publication/333444574_EfficientNet_Rethinking_Model_Scaling_for_Convolutional_Neural_Networks
[89] C. Feichtenhofer, H. Fan, B. Xiong, R. Girshick, and K. He, “A Large-Scale Study on Unsupervised Spatiotemporal Representation Learning,” 2021 IEEECVF Conf. Comput. Vis. Pattern Recognit. CVPR, pp. 3298–3308, Jun. 2021, doi: 10.1109/CVPR46437.2021.00331.
[90] “Trainable Activations for Image Classification[v1] | Preprints.org.” Accessed: Jan. 16, 2024. [Online]. Available: https://www.preprints.org/manuscript/202301.0463/v1
[91] S. Verma, A. Chug, and A. P. Singh, “Revisiting activation functions: empirical evaluation for image understanding and classification,” Multimed. Tools Appl., Jul. 2023, doi: 10.1007/s11042-023-16159-2.
[92] P. J. P. and A. Sethi, WaveMix: Resource-efficient Token Mixing for Images. 2022.
[93] M. Sandler, A. Howard, M. Zhu, A. Zhmoginov, and L.-C. Chen, “MobileNetV2: Inverted Residuals and Linear Bottlenecks,” presented at the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), IEEE Computer Society, Jun. 2018, pp. 4510–4520. doi: 10.1109/CVPR.2018.00474.
[94] G. Huang, Z. Liu, L. V. D. Maaten, and K. Q. Weinberger, “Densely Connected Convolutional Networks,” presented at the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), IEEE Computer Society, Jul. 2017, pp. 2261–2269. doi: 10.1109/CVPR.2017.243.
[95] “Tikhonov Regularization and Total Least Squares | SIAM Journal on Matrix Analysis and Applications.” Accessed: Mar. 19, 2024. [Online]. Available: https://epubs.siam.org/doi/abs/10.1137/S0895479897326432?journalCode=sjmael
[96] J. Bridle, “Training Stochastic Model Recognition Algorithms as Networks can Lead to Maximum Mutual Information Estimation of Parameters,” in Advances in Neural Information Processing Systems, Morgan-Kaufmann, 1989. Accessed: Jan. 18, 2023. [Online]. Available: https://proceedings.neurips.cc/paper/1989/hash/0336dcbab05b9d5ad24f4333c7658a0e-Abstract.html
[97] “[PDF] Gaussian Mixture Convolution Networks | Semantic Scholar.” Accessed: Apr. 23, 2024. [Online]. Available: https://www.semanticscholar.org/paper/Gaussian-Mixture-Convolution-Networks-Celarek-Hermosilla/ce108373d3458c2f27524483221aaad772c3edf3
[98] “An Introduction to Logistic Regression Analysis and Reporting: The Journal of Educational Research: Vol 96, No 1.” Accessed: Mar. 20, 2024. [Online]. Available: https://www.tandfonline.com/doi/abs/10.1080/00220670209598786
[99] E. Trentin, “Maximum-Likelihood Estimation of Neural Mixture Densities: Model, Algorithm, and Preliminary Experimental Evaluation,” 2018, pp. 178–189. doi: 10.1007/978-3-319-99978-4_14.
[100] “Soft-Margin Softmax for Deep Classification | SpringerLink.” Accessed: Feb. 01, 2023. [Online]. Available: https://link.springer.com/chapter/10.1007/978-3-319-70096-0_43
[101] “Large scale multi-output multi-class classification using Gaussian processes | Machine Learning.” Accessed: Mar. 20, 2024. [Online]. Available: https://link-springer-com.lib-ezproxy.concordia.ca/article/10.1007/s10994-022-06289-3
[102] T. Hastie, R. Tibshirani, and J. Friedman, The Elements of Statistical Learning. in Springer Series in Statistics. New York, NY: Springer, 2009. doi: 10.1007/978-0-387-84858-7.
[103] “[PDF] Striving for Simplicity: The All Convolutional Net | Semantic Scholar.” Accessed: Apr. 23, 2024. [Online]. Available: https://www.semanticscholar.org/paper/Striving-for-Simplicity%3A-The-All-Convolutional-Net-Springenberg Dosovitskiy/33af9298e5399269a12d4b9901492fe406af62b4
[104] “[PDF] Qualitatively characterizing neural network optimization problems | Semantic Scholar.” Accessed: Mar. 20, 2024. [Online]. Available: https://www.semanticscholar.org/paper/Qualitatively-characterizing-neural-network-Goodfellow-Vinyals/4d4d09ae8f6a11547441f7fee36405758102a801
[105] “[PDF] ARBEx: Attentive Feature Extraction with Reliability Balancing for Robust Facial Expression Learning | Semantic Scholar.” Accessed: Apr. 23, 2024. [Online]. Available: https://www.semanticscholar.org/paper/ARBEx%3A-Attentive-Feature-Extraction-with-Balancing-Wasi-vSerbetar/b468a27c1ea7692d4de00ff5a948f3bfeaa63a60
[106] D. Ruan, Y. Yan, S. Lai, Z. Chai, C. Shen, and H. Wang, “Feature Decomposition and Reconstruction Learning for Effective Facial Expression Recognition,” 2021 IEEECVF Conf. Comput. Vis. Pattern Recognit. CVPR, pp. 7656–7665, Jun. 2021, doi: 10.1109/CVPR46437.2021.00757.
[107] P. Lucey, J. F. Cohn, T. Kanade, J. Saragih, Z. Ambadar, and I. Matthews, “The Extended Cohn-Kanade Dataset (CK+): A complete dataset for action unit and emotion-specified expression,” in 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition - Workshops, Jun. 2010, pp. 94–101. doi: 10.1109/CVPRW.2010.5543262.
[108] S. Li, W. Deng, and J. Du, “Reliable Crowdsourcing and Deep Locality-Preserving Learning for Expression Recognition in the Wild,” in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Jul. 2017, pp. 2584–2593. doi: 10.1109/CVPR.2017.277.
[109] S. Chen, J. Wang, Y. Chen, Z. Shi, X. Geng, and Y. Rui, “Label Distribution Learning on Auxiliary Label Space Graphs for Facial Expression Recognition,” presented at the Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 13984–13993. Accessed: Apr. 07, 2024. [Online]. Available: https://openaccess.thecvf.com/content_CVPR_2020/html/Chen_Label_Distribution_Learning_on_Auxiliary_Label_Space_Graphs_for_Facial_CVPR_2020_paper.html
[110] “Pyramid With Super Resolution for In-the-Wild Facial Expression Recognition | IEEE Journals & Magazine | IEEE Xplore.” Accessed: Apr. 07, 2024. [Online]. Available: https://ieeexplore.ieee.org/document/9143068
[111] D. Zeng, Z. Lin, X. Yan, Y. Liu, F. Wang, and B. Tang, “Face2Exp: Combating Data Biases for Facial Expression Recognition,” presented at the Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 20291–20300. Accessed: Apr. 07, 2024. [Online]. Available: https://openaccess.thecvf.com/content/CVPR2022/html/Zeng_Face2Exp_Combating_Data_Biases_for_Facial_Expression_Recognition_CVPR_2022_paper.html
[112] C. L. Lisetti and D. E. Rumelhart, “Facial Expression Recognition,” p. 5.
[113] “[PDF] MViT: Mask Vision Transformer for Facial Expression Recognition in the wild | Semantic Scholar.” Accessed: Apr. 07, 2024. [Online]. Available: https://www.semanticscholar.org/paper/MViT%3A-Mask-Vision-Transformer-for-Facial-Expression-Li-Sui/070c917ab1a4d6b924a9613ca18443f260d8d5be
[114] “[PDF] TransFER: Learning Relation-aware Facial Expression Representations with Transformers | Semantic Scholar.” Accessed: Apr. 07, 2024. [Online]. Available: https://www.semanticscholar.org/paper/TransFER%3A-Learning-Relation-aware-Facial-Expression-Xue-Wang/ebf221bf7260e2d27b243b15909d89196f62f39b
[115] A. Vaswani et al., “Attention is All you Need,” presented at the Neural Information Processing Systems, Jun. 2017. Accessed: Apr. 07, 2024. [Online]. Available: https://www.semanticscholar.org/paper/Attention-is-All-you-Need-Vaswani-Shazeer/204e3073870fae3d05bcbc2f6a8e263d9b72e776
[116] A. H. Farzaneh and X. Qi, “Facial Expression Recognition in the Wild via Deep Attentive Center Loss,” in 2021 IEEE Winter Conference on Applications of Computer Vision (WACV), Jan. 2021, pp. 2401–2410. doi: 10.1109/WACV48630.2021.00245.
[117] “Relative Uncertainty Learning for Facial Expression Recognition.” Accessed: Apr. 07, 2024. [Online]. Available: https://proceedings.neurips.cc/paper/2021/hash/9332c513ef44b682e9347822c2e457ac-Abstract.html
[118] “[PDF] Deep-Emotion: Facial Expression Recognition Using Attentional Convolutional Network | Semantic Scholar.” Accessed: Apr. 07, 2024. [Online]. Available: https://www.semanticscholar.org/paper/Deep-Emotion%3A-Facial-Expression-Recognition-Using-Minaee-Abdolrashidi/361b59d962068d47cac30c6fabbf1b8a91efcb2a
[119] I. J. Goodfellow et al., “Challenges in Representation Learning: A Report on Three Machine Learning Contests,” in Neural Information Processing, M. Lee, A. Hirose, Z.-G. Hou, and R. M. Kil, Eds., in Lecture Notes in Computer Science. Berlin, Heidelberg: Springer, 2013, pp. 117–124. doi: 10.1007/978-3-642-42051-1_16.
[120] L. Pham, T. H. Vu, and T. A. Tran, “Facial Expression Recognition Using Residual Masking Network,” in 2020 25th International Conference on Pattern Recognition (ICPR), Jan. 2021, pp. 4513–4519. doi: 10.1109/ICPR48806.2021.9411919.
[121] “[PDF] Local Multi-Head Channel Self-Attention for Facial Expression Recognition | Semantic Scholar.” Accessed: Apr. 23, 2024. [Online]. Available: https://www.semanticscholar.org/paper/Local-Multi-Head-Channel-Self-Attention-for-Facial-Pecoraro-Basile/cde1045b1e9f3b3939ae54df6ae8a2e3079b6d15
[122] “Local Learning With Deep and Handcrafted Features for Facial Expression Recognition | IEEE Journals & Magazine | IEEE Xplore.” Accessed: Apr. 07, 2024. [Online]. Available: https://ieeexplore.ieee.org/document/8716652
[123] “[PDF] PAtt-Lite: Lightweight Patch and Attention MobileNet for Challenging Facial Expression Recognition | Semantic Scholar.” Accessed: Apr. 23, 2024. [Online]. Available: https://www.semanticscholar.org/paper/PAtt-Lite%3A-Lightweight-Patch-and-Attention-for-Ngwe-Lim/b0a3332bc56133fef59eba242f44add54fc2af9b
[124] “EmoNeXt: an Adapted ConvNeXt for Facial Emotion Recognition | IEEE Conference Publication | IEEE Xplore.” Accessed: Apr. 07, 2024. [Online]. Available: https://ieeexplore.ieee.org/abstract/document/10337732
[125] S. Vignesh, M. Savithadevi, M. Sridevi, and R. Sridhar, “A novel facial emotion recognition model using segmentation VGG-19 architecture,” Int. J. Inf. Technol., vol. 15, no. 4, pp. 1777–1787, Apr. 2023, doi: 10.1007/s41870-023-01184-z.
[126] “Fer2013 Recognition - ResNet18 With Tricks | Papers With Code.” Accessed: Aug. 14, 2023. [Online]. Available: https://paperswithcode.com/paper/fer2013-recognition-resnet18-with-tricks
[127] “Facial Expression Recognition via Deep Learning | IEEE Conference Publication | IEEE Xplore.” Accessed: Apr. 07, 2024. [Online]. Available: https://ieeexplore.ieee.org/document/8308363
[128] H. Vo, G.-S. Lee, H.-J. Yang, and S. H. Kim, “Pyramid with Super Resolution for In-The-Wild Facial Expression Recognition,” IEEE Access, vol. PP, pp. 1–1, Jul. 2020, doi: 10.1109/ACCESS.2020.3010018.
[129] Z. Wen, W. Lin, T. Wang, and G. Xu, “Distract Your Attention: Multi-Head Cross Attention Network for Facial Expression Recognition,” Biomim. Basel Switz., vol. 8, no. 2, p. 199, May 2023, doi: 10.3390/biomimetics8020199.
[130] S. Zhang, Y. Zhang, Y. Zhang, Y. Wang, and Z. Song, “A Dual-Direction Attention Mixed Feature Network for Facial Expression Recognition,” Electronics, vol. 12, no. 17, Art. no. 17, Jan. 2023, doi: 10.3390/electronics12173595.
[131] J. Li, J. Nie, D. Guo, R. Hong, and M. Wang, “Emotion Separation and Recognition from a Facial Expression by Generating the Poker Face with Vision Transformers.” arXiv, Jun. 09, 2023. doi: 10.48550/arXiv.2207.11081.
[132] F. Xue, Q. Wang, and G. Guo, “TransFER: Learning Relation-aware Facial Expression Representations with Transformers,” 2021 IEEECVF Int. Conf. Comput. Vis. ICCV, pp. 3581–3590, Oct. 2021, doi: 10.1109/ICCV48922.2021.00358.
[133] “MViT: Mask Vision Transformer for Facial Expression Recognition in the wild.” Accessed: Apr. 08, 2024. [Online]. Available: https://www.researchgate.net/publication/352244399_MViT_Mask_Vision_Transformer_for_Facial_Expression_Recognition_in_the_wild
[134] A. Psaroudakis and D. Kollias, “MixAugment & Mixup: Augmentation Methods for Facial Expression Recognition,” in 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), New Orleans, LA, USA: IEEE, Jun. 2022, pp. 2366–2374. doi: 10.1109/CVPRW56347.2022.00264.
[135] “[PDF] Suppressing Uncertainties for Large-Scale Facial Expression Recognition | Semantic Scholar.” Accessed: Apr. 08, 2024. [Online]. Available: https://www.semanticscholar.org/paper/Suppressing-Uncertainties-for-Large-Scale-Facial-Wang-Peng/92006a84036f9e37e11c773c19141ad8a6675e39
[136] K. Wang, X. Peng, J. Yang, D. Meng, and Y. Qiao, “Region Attention Networks for Pose and Occlusion Robust Facial Expression Recognition,” IEEE Trans. Image Process., vol. 29, pp. 4057–4069, 2020, doi: 10.1109/TIP.2019.2956143.
[137] “[PDF] Occlusion Aware Facial Expression Recognition Using CNN With Attention Mechanism | Semantic Scholar.” Accessed: Apr. 08, 2024. [Online]. Available: https://www.semanticscholar.org/paper/Occlusion-Aware-Facial-Expression-Recognition-Using-Li-Zeng/b5bb7e12a15b57b4d307e742da127a74596d0c7c
[138] “[PDF] Learning Vision Transformer with Squeeze and Excitation for Facial Expression Recognition | Semantic Scholar.” Accessed: Apr. 23, 2024. [Online]. Available: https://www.semanticscholar.org/paper/Learning-Vision-Transformer-with-Squeeze-and-for-Aouayeb-Hamidouche/7109de2f4c522af71098e99563c90db62c19ba55
[139] “Using Self-Supervised Auxiliary Tasks to Improve Fine-Grained Facial Representation | Semantic Scholar.” Accessed: Aug. 10, 2023. [Online]. Available: https://www.semanticscholar.org/paper/Using-Self-Supervised-Auxiliary-Tasks-to-Improve-Pourmirzaei-Esmaili/8c34613a9c123246802e079fa0feb0709f090087
[140] “Facial Expression Recognition Based on Deep Evolutional Spatial-Temporal Networks - PubMed.” Accessed: Apr. 08, 2024. [Online]. Available: https://pubmed.ncbi.nlm.nih.gov/28371777/
[141] “Spatial-Temporal Recurrent Neural Network for Emotion Recognition - PubMed.” Accessed: Oct. 18, 2023. [Online]. Available: https://pubmed.ncbi.nlm.nih.gov/29994572/
[142] “(PDF) Identity-Free Facial Expression Recognition Using Conditional Generative Adversarial Network.” Accessed: Apr. 08, 2024. [Online]. Available: https://www.researchgate.net/publication/351736453_IdentityFree_Facial_Expression_Recognition_Using_Conditional_Generative_Adversarial_Network
[143] “Peak-Piloted Deep Network for Facial Expression Recognition | SpringerLink.” Accessed: Apr. 23, 2024. [Online]. Available: https://link-springer-com.lib-ezproxy.concordia.ca/chapter/10.1007/978-3-319-46475-6_27
[144] H. Jung, S. Lee, J. Yim, S. Park, and J. Kim, “Joint Fine-Tuning in Deep Neural Networks for Facial Expression Recognition,” presented at the Proceedings of the IEEE International Conference on Computer Vision, 2015, pp. 2983–2991. Accessed: Oct. 18, 2023. [Online]. Available: https://openaccess.thecvf.com/content_iccv_2015/html/Jung_Joint_Fine-Tuning_in_ICCV_2015_paper.html
[145] “Occlusion Aware Facial Expression Recognition Using CNN With Attention Mechanism | IEEE Journals & Magazine | IEEE Xplore.” Accessed: Apr. 08, 2024. [Online]. Available: https://ieeexplore.ieee.org/document/8576656
[146] R. Cipolla, Y. Gal, and A. Kendall, “Multi-task Learning Using Uncertainty to Weigh Losses for Scene Geometry and Semantics,” in 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Jun. 2018, pp. 7482–7491. doi: 10.1109/CVPR.2018.00781.
[147] “[PDF] Multi-Task Learning with Deep Neural Networks: A Survey | Semantic Scholar.” Accessed: Apr. 08, 2024. [Online]. Available: https://www.semanticscholar.org/paper/Multi-Task-Learning-with-Deep-Neural-Networks%3A-A-Crawshaw/74f23063ca77f5b1caa3770a5957ae5fc565843e
All items in Spectrum are protected by copyright, with all rights reserved. The use of items is governed by Spectrum's terms of access.

Repository Staff Only: item control page

Downloads per month over past year

Research related to the current document (at the CORE website)
- Research related to the current document (at the CORE website)
Back to top Back to top