Acknowledgement
이 논문은 2022년도 정부(과학기술정보통신부)의 재원으로 국가과학기술연구회 창의형 융합연구사업(No.CAP21052-300)의 지원을 받아 수행된 연구임.
References
- J. W. Kim and H. H. Kim, "Communicative ability in normal aging: A Review," Korean Journal of Communication Disorders, Vol.14, No.4, pp.495-513, 2009.
- G. Gosztolya, V. Vincze, L. Toth, M. Pakaski, J. Kalman, and I. Hoffmann, "Identifying mild cognitive impairment and mild Alzheimer's disease based on spontaneous speech using ASR and linguistic features," Computer Speech & Language, Vol.53, pp.181-197, 2019. https://doi.org/10.1016/j.csl.2018.07.007
- M. R. Morales and R. Levitan, "Speech vs. text: A comparative analysis of features for depression detection systems," 2016 IEEE spoken language technology workshop (SLT), IEEE. 2016.
- I. Vigo, L. Coelho, and S. Reis. "Speech-and language-based classification of alzheimer's disease: A systematic review," Bioengineering, Vol.9, No.1, pp.27, 2022.
- M. Ehghaghi, F. Rudzicz, and J. Novikova, "Data-driven approach to differentiating between depression and dementia from noisy speech and language data," arXiv preprint arXiv:2210.03303, 2022.
- S. H. Han, S. H. Dong, and B. O. Kang, "Comparison of classification performance between adult and elderly using acoustic and linguistic features from spontaneous speech," in Proceedings of the Korea Conference on Software Engineering (KCSE) 2023, Vol.25, pp.117-118, 2023.
- F. Eyben, M. Wollmer, and B. Schuller, "Opensmile: The munich versatile and fast open-source audio feature extractor," Proceedings of the 18th ACM International Conference on Multimedia, pp.1459-1462, 2010.
- SKTBrain, "KoBERT", github repository, accessed Dec. 16, 2022, [Internet], https://github.com/SKTBrain/KoBERT
- F. Rangel, F. Celli, P. Rosso, M. Potthast, B. Stein, and W. Daelemans, "Overview of the 3rd Author Profiling Task at PAN 2015," Conference and Labs of the Evaluation Forum, 2015.
- A. Liesenfeld, G. Parti, Y. Y. Hsu, and C. R. Huang, "Predicting gender and age categories in English conversations using lexical, non-lexical, and turn-taking features," arXiv preprint arXiv:2102.13355, 2021.
- J. Devlin, M. W. Chang, K. Lee, and K. Toutanova, "Bert: Pre-training of deep bidirectional transformers for language understanding," arXiv preprint arXiv:1810.04805, 2018.
- B. P. R. Guda, A. Garimella, and N. Chhaya. "Empathbert: A bert-based framework for demographic-aware empathy prediction," arXiv preprint arXiv:2102.00272, 2021.
- B. Schuller et al., "Cross-corpus acoustic emotion recognition: Variances and strategies," IEEE Transactions on Affective Computing, Vol.1, No.2, pp.119-131, 2010. https://doi.org/10.1109/T-AFFC.2010.8
- F. Eyben et al., "The Geneva minimalistic acoustic parameter set (GeMAPS) for voice research and affective computing," IEEE Transactions on Affective Computing, Vol.7, No.2, pp.190-202, 2015. https://doi.org/10.1109/TAFFC.2015.2457417
- B. Schuller et al., "The INTERSPEECH 2010 paralinguistic challenge," 2010.
- F. Burkhardt, M. Bruckl, and B. Schuller, "Age classification: Comparison of human vs machine performance in prompted and spontaneous speech," Studientexte zur Sprachkommunikation: Elektronische Sprachsignalverarbeitung 2021, pp.35-42, 2021.
- T. Pires, E. Schlinger, and D. Garrette, "How multilingual is multilingual BERT?," arXiv preprint arXiv:1906.01502, 2019.
- S. Lee, H. Jang, Y. Baik, S. Park, and H. Shin, "Kr-bert: A small-scale korean-specific language model," arXiv preprint arXiv:2008.03979, 2020.
- "자유대화 음성(일반남여)," AI-Hub, accessed Dec. 16, 2022, https://aihub.or.kr/aihubdata/data/view.do?currMenu=115&topMenu=100&aihubDataSe=realm&dataSetSn=109
- "자유대화 음성(노인남여)," AI-Hub, accessed Dec. 16, 2022, https://aihub.or.kr/aihubdata/data/view.do?currMenu=115&topMenu=100&aihubDataSe=realm&dataSetSn=107