참고문헌
- Benesty, J., Chen, J., Huang, Y., and Cohen, I. (2009). "Pearson correlation coefficient." In Noise reduction in speech processing, Springer, Berlin, Heidelberg, pp. 1-4.
- Bergstra, J., and Bengio, Y. (2012). "Random search for hyper-parameter optimization." The Journal of Machine Learning Research, 13(1), pp. 281-305.
- Choi, J., and Lee, S. (2020). "Short-term wind power forecast using hourly LSTM technique." Korean Institute of Electrical Engineers, 69(6), pp. 759-764. https://doi.org/10.5370/KIEE.2020.69.6.759
- Choi, M., and Kwon, O. (2008). "Construction material cost increase and countermeasures." Construction trend briefing by Korea Institute of Construction Industry, 6, pp. 2-34.
- Choi, Y., Yim, H., and Park, B. (2009). "Analysis on the Lotting Price Fluctuation of the Multi-Family Attached House According to the Construction Material Cost Variation." Journal of The Korean Society of Civil Engineers, 29(6D), pp. 753-760.
- Hen, X., Wei, L., and Xu, J. (2017). "House price prediction using lstm." arXiv preprint arXiv:1709.08432.
- Hochreiter, S., and Schmidhuber, J. (1997). "Long short-term memory." Neural computation, 9(8), pp. 1735-1780. https://doi.org/10.1162/neco.1997.9.8.1735
- Jeong, D. (2017). "Trend on Artificial Intelligence Technology and Its Related Industry." Korea Institute of Information Technology Magazine, 15(2), pp. 21-28. https://doi.org/10.14801/jkiit.2017.15.5.21
- Ji, S., Goo, Y., Baek, U., Park, J., and Yoon, S. (2019). "LSTM Learning Data Selection Technique for Number of Bitcoin Transactions Prediction." In KNOM Conference.
- Kim, B., Jung, S., Kim, M., Kim, J., Lee, H., and Kim, S. (2020a). "Solar Power Generation Forecasting based on LSTM considering Weather Conditions." Journal of Korean Institute of Intelligent Systems, 30(1), pp. 7-12. https://doi.org/10.5391/JKIIS.2020.30.1.7
- Kim, J., Lee, S.H., Choi, Y., and Woo, S. (2020b). "Long-term Settlement Prediction of Railway Concrete Track Based on Recurrent Neural Network (RNN)" Journal of the Korean Geotechnical Society, 36(3), pp. 5-14.
- Lahari, M.C., Ravi, D.H., and Bharathi, R. (2018). "Fuel Price Prediction Using RNN." In 2018 International Conference on Advances in Computing, Communications and Informatics (ICACCI), IEEE, pp. 1510-1514.
- Larochelle, H., Erhan, D., Courville, A., Bergstra, J., and Bengio, Y. (2007). "An empirical evaluation of deep architectures on problems with many factors of variation." In Proceedings of the 24th International Conference on Machine Learning, pp. 473-480.
- Lee, J., Yoo, J., Kim, C., Lee, G., and Lim, B. (2008). "How to calculate the order point considering the fluctuations in demand for materials at construction sites." Journal of the Architectural Institute of Korea-Structural System, 24 (10), pp. 117-125.
- Lee, Y., and Kim, K. (2020a). "Experimental Study on the Short-Term Prediction of Rebar Price using Bidirectional LSTM with Data Combination and Deep Learning Related Techniques." Korean Journal of Construction Engineering and Management, KICEM, 21(6), pp. 38-45. https://doi.org/10.6106/KJCEM.2020.21.6.038
- Lee, Y., and Kim, K. (2020b). "Experimental Study on the Expansion of the Short-term Prediction Range of Rebar Prices Using Deep Learning." Journal of The Architectural Institute of Korea, 36(12), pp. 265-272. https://doi.org/10.5659/JAIK.2020.36.12.265
- Nayak, S., Misra, B., and Behera, H. (2014). "Impact of data normalization on stock index forecasting." International Journal of Computer Information Systems and Industrial Management Applications, 6, pp. 257-269.
- Pawar, K., Jalem, R., and Tiwari, V. (2019). "Stock market price prediction using LSTM RNN." In Emerging Trends in Expert Applications and Security, pp. 493-503. Springer, Singapore.
- Rumelhart, D., Hinton, G., and Williams, R. (1985). Learning internal representations by error propagation (No. ICS-8506), California Univ San Diego La Jolla Inst for Cognitive Science, (No. ICS-8506), pp. 318-362.
- Schuster, M., and Paliwal, K. (1997). "Bidirectional recurrent neural networks." IEEE transactions on Signal Processing, 45(11), pp. 2673-2681. https://doi.org/10.1109/78.650093
- Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., and Salakhutdinov, R. (2014). "Dropout: a simple way to prevent neural networks from overfitting." The journal of machine learning research, 15(1), pp. 1929-1958.
- Taieb, S., Sorjamaa, A., and Bontempi, G. (2010). "Multiple-output modeling for multi-step-ahead time series forecasting." Neurocomputing, 73(10-12), pp. 1950-1957. https://doi.org/10.1016/j.neucom.2009.11.030
- Willmott, C., and Matsuura, K. (2005). "Advantages of the mean absolute error (MAE) over the root mean square error (RMSE) in assessing average model performance." Climate research, 30(1), pp. 79-82. https://doi.org/10.3354/cr030079
- Tensorflow.org. (2020). Overfitting and underfitting. accessed Sep 27, 2020, https://www.tensorflow.org/tutorials/keras/overfit_and_underfit.stand.