User Responsive Automatic Method for Real Time Depression Detection Using Deep Neural Network

Authors

  • Jiger P. Acharya
  • Milind S. Shah

Keywords:

Mental health, Depression, SRI, PHQ-9, DMI-10, Visual behavioral analysis, Audio Video Emotion Challenge (AVEC), Visual features, DSM, Deep Neural Network (DNN) etc

Abstract

During last one decade the mental health is becoming a major concerning issue to society as number of various socio-economical, personnel and societal issues related to mental health   are increased in exponential manner. Depression is mood disorders which result in severe disabling conditions which affect person’s ability to cope with routine life and real-life challenges which are dynamic in nature. It may occur when person remain more than two weeks in negative state of mind continuously. Depression has observable behavioral symptoms related to affective and psychomotor domains which can be identified. Classical approaches majorly depend on person’s behavioral analysis and family observations during clinical interviews, which are effective if it can be precisely defined and assessed but persons have tendency to conceal it so they are less effective while proposed method with Deep Learning techniques perform the said task with accuracy of 79% and open a new era for health care domain.

Downloads

Download data is not yet available.

Metrics

Metrics Loading ...

References

P. Ekman, Handbook of Cognition and Emotion, JohnWiley& Sons Ltd, 2005, pp. 45-60.

H. Ellgring, R. J. Eiser, K. R. Scherer, "Non-verbal Communication in Depression," Cambridge University Press, 2007.

American Psychiatric Association (5th edition, 2013). Diagnostic and statistical manual of mental disorders. Washington: DC.

Gupta R, Malandrakis N, Xiao B, Guha T, Van Segbroeck M, Black MP,PotamianosA,Narayanan SS (2014). Multimodal prediction of affective dimensions and depression inhuman-computer interactions. In International Audio/Visual Emotion Challenge and Workshop.

A. Pampouchidou, P.G. Simos, K. Marias, F. Meriaudeau, F. Yang, M. Pediaditis, M. Tsiknakis, "Automatic Assessment of Depression Based on Visual Cues: A Systematic Review," IEEE Transactions on Affective Computing, vol. 10, no. 4, pp. 445-470, 2019.

Komal Anadkat & Hiteishi M. Diwanji “Unimodal to Multimodal Emotion Recognition: A systematic Review” Test Engineering and Management “ISSN: 0193 4120, Volume 83 PP- 6423-6427 August 2020 (SCOPUS Indexed).

Alghowinem S, Goecke R, Wagner M, Parker G, Breakspear M (2013). Head pose and movement analysis as an indicator of depression. In Humaine Association Conference on Affective Computing and Intelligent Interaction,283–288.

Alghowinem S, Goecke R, Wagner M, Parker G, Breakspear M (2013). Eye movement analysis for depression detection. In IEEE International Conference on Image Processing,4220–4224.

Jeffrey M Girard, Jeffrey F Cohn, Mohammad H Mahoor, S Mohammad Mavadati, ZakiaHammal, and Dean P Rosenwald, "Nonverbal social withdrawal in depression: Evidence from manual and automatic analyses," Image and vision computing, pp. 641-647, 2014.

Internet resource: https://www.verywellmind.com/colour-psychology-2795824

Morales, M., Scherer, S., & Levitan, R, "A Cross-modal Review of Indicators for Depression Detection Systems," Proceedings of the Fourth Workshop on Computational Linguistics and Clinical Psychology –- From Linguistic Signal to Clinical Reality, 2017.

Y. Zhou, S. Scherer, D. Devault, J. Gratch, G. Stratou, L.-P. Morency, and J. Cassell, "Multimodal Prediction of Psychological Disorders: Learning Verbal and Nonverbal Commonalities in Adjacency Pairs," 17thWorkshop on the Semantics and Pragmatics of Dialogue. SEMDIAL, pp. 160-169, 2013.

R.Verma,“FER2013,”Kaggle,26-May-2018.[Online].Available: https://www.kaggle.com/ datasets/deadskull7/fer2013.

Pampouchidou, Anastasia, Kostas Marias, ManolisTsiknakis, P.Simos, Fan Yang, and Fabrice Meriaudeau (2015). Designing a framework for assisting depression severity assessment from facial image analysis. In Signal and Image Processing Applications(ICSIPA), International Conference on, pp. 578-583, IEEE.

Patient Health Questionnaire-9,Kroenke, K., Spitzer, R. L., & Williams, J. B. W. (1999). Patient Health Questionnaire-9 (PHQ-9) [Database record]. APA PsycTests.

S. E. Kahou, X. Bouthillier, P. Lamblin, C. Gulcehre, V. Michalski, K. Konda, S. Jean, P. Froumenty, Y. Dauphin, N. Boulanger-Lewandowski, R. Chandias Ferrari, M. Mirza, D. Warde-Farley, A. Courville, P. Vincent, R. Memisevic, C. Pal, and Y. Bengio, “Emonets: Multimodal deep learning approaches for emotion recognition in video,” Journal on Multimodal User Interfaces, vol. 10, no. 2, pp. 99–111, 2015.

P. Tzirakis, G. Trigeorgis, M. A. Nicolaou, B. W. Schuller, and S. Zafeiriou, “End to-End Multimodal Emotion Recognition Using Deep Neural Networks,” IEEE Journal of Selected Topics in Signal Processing, vol. 11, no. 8, pp. 1301–1309, Dec. 2017, doi:10.1109/jstsp.2017.2764438.

S. Sahay, S. H. Kumar, R. Xia, J. Huang, and L. Nachman, “Multimodal relational tensor network for sentiment and emotion classification,” Proceedings of Grand Challenge and Workshop on Human Multimodal Language (Challenge-HML), 2018.

D. Hazarika, S. Poria, A. Zadeh, E. Cambria, L.-P. Morency, and R. Zimmermann, “Conversational memory network for emotion recognition in dyadic dialogue videos,” Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), 2018.

M. M. Hassan, Md. G. R. Alam, Md. Z. Uddin, S. Huda, A. Almogren, and G. Fortino, “Human emotion recognition using deep belief network architecture,” Information Fusion, vol. 10.1016/j.inffus.2018.10.009. 51, pp. 10–18, Nov. 2019, doi: 21. S. Siriwardhana, A. Reis, R. Weerasekera, and S. Nanayakkara, “Jointly fine-tuning ‘bert-like’ self supervised models to improve multimodal speech emotion recognition,” Interspeech 2020, 2020.

D. Priyasad, T. Fernando, S. Denman, S. Sridharan, and C. Fookes, “Attention driven fusion for multi-modal emotion recognition,” ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2020.

J. Njoku, Angela C. Caliwag, W. Lim, S. Kim, H.-J. Hwang, and J.-W. Jeong, “Deep Learning Based Data Fusion Methods for Multimodal Emotion Recognition,” The Journal of Korean Institute of Communications and Information Sciences, vol. 47, no. 1, pp. 79–87, Jan. 2022, doi: 10.7840/kics.2022.47.1.79.

Komal D. Anadkat,Dr. Hiteishi M. Diwanji & Dr. Shahid Modasiya “Effect Of Preprocessing In Human Emotion Analysis Using Social Media Status Dataset” Reliability: Theory and Application “ISSN 1932-2321, Volume 17 Issue 1 PP-67 March 2022 (SCOPUS Indexed).

Komal D. Anadkat & Dr. Hiteishi M. Diwanji “Effectof Activation Function in Speech Emotion Recognition On The Ravdess Dataset” Reliability: Theory and Application “ISSN 1932-2321, Volume 16 Issue 3 PP-63 September 2021 (SCOPUS Indexed).

Internet resources like: doc player.net ,journalofbigdata., springeropen.com, deepai.org, arxiv.org ,pubmed.ncbi.nlm.nih.gov

Downloads

Published

2025-07-08

How to Cite

1.
P. Acharya J, S. Shah M. User Responsive Automatic Method for Real Time Depression Detection Using Deep Neural Network. J Neonatal Surg [Internet]. 2025Jul.8 [cited 2025Sep.19];14(32S):4286-97. Available from: https://www.jneonatalsurg.com/index.php/jns/article/view/8117