Deep Learning based Classification of Depression and Suicidal Risk Among Normal Speakers using Delta-Spectral Cepstral Coefficients
Conference proceedings article
ผู้เขียน/บรรณาธิการ
กลุ่มสาขาการวิจัยเชิงกลยุทธ์
รายละเอียดสำหรับงานพิมพ์
รายชื่อผู้แต่ง: Pavat Ruckchopsanti, Nattarika Ngearnsajja, Pawat Isaraporn, Thanchanok Haruenputh, Thaweesak Yingthawornsuk
ปีที่เผยแพร่ (ค.ศ.): 2025
หน้าแรก: 159
หน้าสุดท้าย: 159
จำนวนหน้า: 1
ภาษา: English-United States (EN-US)
บทคัดย่อ
This study utilizes deep learning techniques to develop a precise model for identifying audio recordings, of female speakers for their mental health conditions. The female speakers were categorized into three volunteer groups, which are remitted (RMT) , depressed ( DPR), and high-risk suicidal (HRK) by psychiatrist. The various deep learning techniques such as CNN1D, CNN2D, SVM and LSTM were thoroughly trained and validated with our feature samples which are the Delta MFCC (∆MFCC), representing the vocal-tract frequency response associated with different categorized emotional illnesses of three studied volunteer groups. The experimental results show that the CNN2D model achieved the highest accuracies of 0.95 and 0.99 among studied models in classifying ∆MFCC samples between RMT and DPR groups, and between RMT and HRK groups, respectively. Moreover, precision and recall scores were also robust for the CNN models. In contrast, the SVM model achieved with accuracy of 0.83 as compare to CNN models while its precision and recall found to be adequately high. The findings can significantly contribute to understanding of affected vocal characteristics of spoken sound samples associated with mental illness conditions.
คำสำคัญ
ไม่พบข้อมูลที่เกี่ยวข้อง