2022
DOI: 10.1016/j.health.2022.100090
|View full text |Cite
|
Sign up to set email alerts
|

A multimodal computer-aided diagnostic system for depression relapse prediction using audiovisual cues: A proof of concept

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 16 publications
(4 citation statements)
references
References 32 publications
0
4
0
Order By: Relevance
“…In this manuscript, we proposed a multimodal deep convolution neural network to detect the severity of depression. Compared to an existing multimodal model ( 20 , 31 ), our model has the following improvements: (i) we integrated facial expression information that is an important feature in evaluating the severity of depression; (ii) we constructed the BDD metrics to quantify the severity of depression and achieved a good performance; (iii) we found that the information extracted from different modes, when integrated in appropriate proportions, can significantly improve the accuracy of the evaluation, which has not been reported in previous studies.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…In this manuscript, we proposed a multimodal deep convolution neural network to detect the severity of depression. Compared to an existing multimodal model ( 20 , 31 ), our model has the following improvements: (i) we integrated facial expression information that is an important feature in evaluating the severity of depression; (ii) we constructed the BDD metrics to quantify the severity of depression and achieved a good performance; (iii) we found that the information extracted from different modes, when integrated in appropriate proportions, can significantly improve the accuracy of the evaluation, which has not been reported in previous studies.…”
Section: Discussionmentioning
confidence: 99%
“…Most of the previous AI models focus on single-modal information of a certain feature of the patient. In recent years, some studies have shown that multi-modal information has a better prediction performance than single-modal data ( 20 , 31 ). In this paper, we proposed a multi-modal deep convolutional neural network (CNN) model based on facial expressions and body movements to evaluate the severity of depression.…”
Section: Introductionmentioning
confidence: 99%
“…Fully connected (FC) layers, multilayer perceptron (MLP), CNN, LSTM, BiLSTM, GRU, temporal convolutional network (TCN) [441] (with dilation for long sequences)-with activation function like Sigmoid, Softmax, ReLU, LeakyReLU, and GeLU Predict scores of assessment scales (regression) or probability distribution over classes (classification) [60,78,80,[84][85][86][87][88][90][91][92][93][94]96,98,105,111,113,117,131,133,135,136,[142][143][144]146,162,163,167,168,170,172,174,178,179,190,197,199,201,218,219,221,223,308] DCNN-DNN (combination of deep CNN and DNN), GCNN-LSTM (combination of gated convolutional neural network, which replaces a convolution block in CNN wi...…”
Section: Neural Networkmentioning
confidence: 99%
“…First, a disproportionately large number of models are derived from small datasets, such as the Distress Analysis Interview Corpus Wizard-of-Oz (DAIC-WOZ), which contains only 189 examples (Gratch et al, 2014). Because it is one of the only publicly available datasets that contains both audiovisual recordings and depression-scale scores, researchers continue to use it for training and validating extremely complex ML models (Othmani & Zeghina, 2022) that likely require exponentially larger sample sizes to achieve out-of-sample generalizability (McNamara et al, 2022). Moreover, models based on audiovisual features have typically not accounted for the confound that depression is more prevalent among women, and have inadvertently used "vocal biomarkers" linked to sex assigned at birth to artificially boost their accuracy at predicting depression (Bailey & Plumbley, 2021).…”
mentioning
confidence: 99%