Multi-lingual depression-level assessment from conversational speech using acoustic and text features
dc.contributor.author | Özkanca, Yasin Serdar | |
dc.contributor.author | Demiroğlu, Cenk | |
dc.contributor.author | Besirli, A. | |
dc.contributor.author | Çelik, S. | |
dc.date.accessioned | 2020-05-18T22:08:31Z | |
dc.date.available | 2020-05-18T22:08:31Z | |
dc.date.issued | 2018 | |
dc.identifier.isbn | 978-1-5108-7221-9 | |
dc.identifier.issn | 2308-457X | en_US |
dc.identifier.uri | http://hdl.handle.net/10679/6575 | |
dc.identifier.uri | https://www.isca-speech.org/archive/Interspeech_2018/abstracts/2169.html | |
dc.description.abstract | Depression is a common mental health problem around the world with a large burden on economies, well-being, hence productivity, of individuals. Its early diagnosis and treatment are critical to reduce the costs and even save lives. One key aspect to achieve that goal is to use voice technologies and monitor depression remotely and relatively inexpensively using automated agents. Although there has been efforts to automatically assess depression levels from audiovisual features, use of transcriptions along with the acoustic features has emerged as a more recent research venue. Moreover, difficulty in data collection and the limited amounts of data available for research are also challenges that are hampering the success of the algorithms. One of the novel contributions in this paper is to exploit the databases from multiple languages for feature selection. Since a large number of features can be extracted from speech and given the small amounts of training data available, effective data selection is critical for success. Our proposed multi-lingual method was effective at selecting better features and significantly improved the depression assessment accuracy. We also use text-based features for assessment and propose a novel strategy to fuse the text- and speech-based classifiers which further boosted the performance. | en_US |
dc.language.iso | eng | en_US |
dc.publisher | International Speech Communication Association | en_US |
dc.relation.ispartof | Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH | |
dc.rights | openAccess | |
dc.title | Multi-lingual depression-level assessment from conversational speech using acoustic and text features | en_US |
dc.type | Conference paper | en_US |
dc.description.version | Publisher version | |
dc.publicationstatus | Published | en_US |
dc.contributor.department | Özyeğin University | |
dc.contributor.authorID | (ORCID 0000-0002-6160-3169 & YÖK ID 144947) Demiroğlu, Cenk | |
dc.contributor.ozuauthor | Demiroğlu, Cenk | |
dc.identifier.startpage | 3398 | en_US |
dc.identifier.endpage | 3402 | en_US |
dc.identifier.wos | WOS:000465363900709 | |
dc.identifier.doi | 10.21437/Interspeech.2018-2169 | en_US |
dc.subject.keywords | Depression estimation | en_US |
dc.subject.keywords | Acoustic features | en_US |
dc.subject.keywords | Feature selection | en_US |
dc.subject.keywords | Multi-lingual applications | en_US |
dc.identifier.scopus | SCOPUS:2-s2.0-85055003235 | |
dc.contributor.ozugradstudent | Özkanca, Yasin Serdar | |
dc.contributor.authorMale | 2 | |
dc.relation.publicationcategory | Conference Paper - International - Institutional Academic Staff and Graduate Student |
Files in this item
This item appears in the following Collection(s)
Share this page