과제정보
본 연구는 방위산업기술지원센터의 지원으로 수행되었음(UC200019D).
참고문헌
- M. Zareapoor and K. R. Seeja, "Feature extraction or feature selection for text classification: A case study on phishing email detection," International Journal of Information Engineering and Electronic Business, Vol.7, No.2, pp.60, 2015.
- O. Bruna, H. Avetisyan, and J. Holub, "Emotion models for textual emotion classification," Journal of Physics: Conference Series, Vol.772, No.1, 2016.
- T. Moh, T.-S. Teng, and Z. Zhang, "Cross-lingual text classification with model translation and document translation," in Proceedings of the 50th Annual Southeast Regional Conference, 2012.
- R. Desai et al., "TextBrew: Automated Model Selection and Hyperparameter Optimization for Text Classification," International Journal of Advanced Computer Science and Applications, Vol.13, No.9, 2022.
- X. He, K. Zhao, and X. Chu, "AutoML: A survey of the state-of-the-art," Knowledge-Based Systems, Vol.212, pp. 106622, 2021.
- J. Han, K. S. Park, and K. M. Lee, "An automated machine learning platform for non-experts," in Proceedings of the International Conference on Research in Adaptive and Convergent Systems, 2020.
- E. LeDell and S. Poirier, "H2o automl: Scalable automatic machine learning," in Proceedings of the AutoML Workshop at ICML, Vol.2020, San Diego, CA, USA: ICML, 2020.
- S. Brandle et al., "Evaluation of representation models for text classification with AutoML tools," in Proceedings of the Future Technologies Conference (FTC) 2021, Vol.2, Springer International Publishing, 2022.
- K. Kowsari et al., "Text classification algorithms: A survey," Information, Vol.10, No.4, pp.150, Apr. 2019.
- M. M. Mironczuk and J. Protasiewicz, "A recent overview of the state-of-the-art elements of text classification," Expert Systems with Applications, Vol.106, pp.36-54, 2018.
- T. Mikolov, K. Chen, G. Corrado, and J. Dean, "Efficient estimation of word representations in vector space," arXiv preprint arXiv:1301.3781, 2013.
- J. Pennington, R. Socher, and C. D. Manning, "Glove: Global vectors for word representation," in Proc. 2014 Conf. Empirical Methods in Natural Language Processing (EMNLP), 2014.
- J. Devlin, M. W. Chang, K. Lee, and K. Toutanova, "BERT: Pre-training of deep bidirectional transformers for language understanding," arXiv preprint arXiv:1810.04805, 2018.
- R. Li, "A review of machine learning algorithms for text classification," Cyber Security, Vol.226, 2022.
- M. Lewis, "Bart: Denoising sequence-to-sequence pretraining for natural language generation, translation, and comprehension," arXiv preprint arXiv:1910.13461, 2019.
- T. B. Brown, "Language models are few-shot learners," arXiv preprint arXiv:2005.14165, 2020.
- Y. Chae and T. Davidson, "Large language models for text classification: From zero-shot learning to fine-tuning," Open Science Foundation, 2023.
- F. Stoica and L. F. Stoica, "AutoML Insights: Gaining Confidence to Operationalize Predictive Models," 2024.
- O. Levy, Y. Goldberg, and I. Dagan, "Improving distributional similarity with lessons learned from word embeddings," Transactions of the Association for Computational Linguistics, Vol.3, pp.211-225, 2015.
- T. Chen and C. Guestrin, "Xgboost: A scalable tree boosting system," in Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 2016.
- J. H. Friedman, "Greedy function approximation: a gradient boosting machine," Annals of Statistics, Vol.29, No.5, pp.1189-1232, 2001.
- N. E. Breslow, "Generalized linear models: checking assumptions and strengthening conclusions," Statistica Applicata, Vol.8, No.1, pp.23-41, 1996.
- A. Candel et al., "Deep learning with H2O," H2O.ai Inc., pp.1-21, 2016.
- P. Geurts, D. Ernst, and L. Wehenkel, "Extremely randomized trees," Machine Learning, Vol.63, pp.3-42, 2006.