Abstract
Communication between people relies on human expression more than any other factor. Emotional identification is, right now, a hot study area. Typically, people will show seven distinct feelings (happy, angry, surprised, sad, disgusted, fearful, and neutral). Humans communicate their feelings via words, body language, and facial expressions. Humans’ ability to reveal inner thought processes via facial expressions is crucial to studying human behavior. Many fields use facial expression analysis, including those concerned with increased security, automated criminal identification, the diagnosis of mental illness, and communication between people and computers. Most studies on emotion detection have only used straightforward CNN and RNN models. However, training such a big data set may take a long because the models need a vast data set. To this end, we suggested a model that combines the Mobile Net-V2 model with the transfer learning strategy to speed up the generation time and improve the accuracy of emotion detection. In order to improve upon the current method, we have compiled a large CIFE data collection from the relevant literature. One of the best outcomes for emotion detection is attained by the suggested model architecture when tested on the data mentioned earlier. Based on what we found, the new system should be able to identify more accurately than the old one.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Sayeedunnisa, S., Hegde, N.P., Khan, K.U.R.: Sarcasm detection: a contemporary research affirmation of recent literature. ECS Trans. 107(1), 2419–2426 (2022). https://doi.org/10.1149/10701.2419ecst
McElroy-Heltzel, S.E.: Epistemic virtues and vices as attitudes: implications for empirical measures and virtue interventions. J. Philosoph. Res. 47, 83–94 (2022). https://doi.org/10.5840/jpr20221027195
Smith, L.: Heritage, the power of the past, and the politics of (mis)recognition. J. Theory Soc. Behav. 52(4), 623–642 (2022). https://doi.org/10.1111/jtsb.12353
Lozada, F.T., Riley, T.N., Catherine, E., Brown, D.W.: Black emotions matter: understanding the impact of racial oppression on Black youth’s emotional development: dismantling systems of racism and oppression during adolescence. J. Res. Adolescence 32(1), 13–33 (2022). https://doi.org/10.1111/jora.12699
Puckett, J.A., et al.: Transgender and gender diverse individuals’ daily experiences of rumination. Am. J. Orthopsych. (2022)
Araneta, J., Ingilan, S.: Pragmatic functions of formulaic expressions in Cebuano. Southeastern Philippines J. Res. Develop. 27(2), 61–83 (2022)
Argyriou, K.: Trajectories towards affirmation: gender identity in mental health services after the ICD-11. Quaderns de Psicologia 24(2), e1821–e1821 (2022)
Goldbach, C., et al.: Resisting trans medicalization: Body satisfaction and social contextual factors as predictors of sexual experiences among trans feminine and nonbinary individuals. J. Sex Res. 1–12 (2022)
Łakuta, P., et al.: Enhancing well-being and alleviating depressive symptoms in people with HIV/AIDS: An intervention based on if–then plans with self-affirming cognitions. Appl. Psychol. Health Well-Being (2022)
Palama, A., et al.: The cross-modal transfer of emotional information from voices to faces in 5-, 8-and 10-year-old children and adults: an eye-tracking study. Emotion 22.4, 725 (2022)
Peng, W., et al.: FADO: Feedback-Aware Double COntrolling Network for Emotional Support Conversation. arXiv preprint arXiv:2211.00250 (2022)
Carbon, C.-C., Held, M.J., Schütz, A.: Reading emotions in faces with and without masks is relatively independent of extended exposure and individual difference variables. Front. Psychol. 13 (2022). https://doi.org/10.3389/fpsyg.2022.856971
Qi, Y., Zhou, C., Chen, Y.: NA-Resnet: neighbor block and optimized attention module for global-local feature extraction in facial expression recognition. Multimedia Tools Appl. 82(11), 16375–16393 (2022). https://doi.org/10.1007/s11042-022-14191-2
Seager van Dyk, I., Aldao, A., Pachankis, J.E.: Coming out under fire: the role of minority stress and emotion regulation in sexual orientation disclosure. PLOS ONE 17(5), e0267810 (2022). https://doi.org/10.1371/journal.pone.0267810
Lazar, M., Miron-Spektor, E., Mueller, J.S.: Love at first insight: an attachment perspective on early-phase idea selection. Organ. Behav. Hum. Decis. Process. 172, 104168 (2022)
Zhao, D., et al.: An attentive and adaptive 3D CNN for automatic pulmonary nodule detection in CT image. Expert Syst. Appl. 211, 118672 (2023). https://doi.org/10.1016/j.eswa.2022.118672
Asia, A.-O., et al.: Detection of diabetic retinopathy in retinal fundus images using CNN classification models. Electronics 11(17), 2740 (2022). https://doi.org/10.3390/electronics11172740
Ouni, S., Fkih, F., Omri, M.N.: BERT- and CNN-based TOBEAT approach for unwelcome tweets detection. Soc. Network Anal. Mining 12(1) (2022). https://doi.org/10.1007/s13278-022-00970-0
Gonçalves, C.B., Souza, J.R., Fernandes, H.: CNN architecture optimization using bio-inspired algorithms for breast cancer detection in infrared images. Comput. Biol. Med. 142, 105205 (2022). https://doi.org/10.1016/j.compbiomed.2021.105205
Yu, K., et al.: Comparison of classical methods and mask R-CNN for automatic tree detection and mapping using UAV imagery. remote sensing 14.2, 295 (2022)
Kadek Sastrawan, I., Bayupati, I.P.A., Arsa, D.M.S.: Detection of fake news using deep learning CNN–RNN based methods. ICT Express 8(3), 396–408 (2022). https://doi.org/10.1016/j.icte.2021.10.003
Laguna, A.B., Krystian, M.: Key. Net: keypoint detection by handcrafted and learned CNN filters revisited. IEEE Trans. Pattern Anal. Mach. Intell. (2022)
Khan, A., et al.: CNN-based smoker classification and detection in smart city application. Sensors 22.3, 892 (2022)
Yang, J., et al.: CodnNet: a lightweight CNN architecture for detection of COVID-19 infection. Appl. Soft Comput. 130, 109656 (2022)
Agarwal, A., et al.: Enhanced iris presentation attack detection via contraction-expansion CNN. Pattern Recogn. Lett. 159, 61–69 (2022)
Hsia, C.-H., et al.: Mask R-CNN with new data augmentation features for smart detection of retail products. Appl. Sci. 12.6, 2902 (2022)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Ahammed, M.T., Ghosh, S., Rahman, M.A., Chandra, P., Shuvo, A.I., Balaji, P. (2023). Meta-transfer Learning for Contextual Emotion Detection in Face Affirmation. In: Tiwari, R.K., Sahoo, G. (eds) Recent Trends in Artificial Intelligence and IoT. ICAII 2023. Communications in Computer and Information Science, vol 1822. Springer, Cham. https://doi.org/10.1007/978-3-031-37303-9_9
Download citation
DOI: https://doi.org/10.1007/978-3-031-37303-9_9
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-37302-2
Online ISBN: 978-3-031-37303-9
eBook Packages: Computer ScienceComputer Science (R0)