• Title/Summary/Keyword: group-specific time series regression analysis

Search Result 4, Processing Time 0.021 seconds

Time series regression model for forecasting the number of elementary school teachers (초등학교 교원 수 예측을 위한 시계열 회귀모형)

  • Ryu, Soo Rack;Kim, Jong Tae
    • Journal of the Korean Data and Information Science Society
    • /
    • v.24 no.2
    • /
    • pp.321-332
    • /
    • 2013
  • Because of the continuous low birthrates, the number of the elementary students will decrease by 17% in 2020 compared to 2011. The purpose of this study is to forecast the number of elementary school teachers until 2020. We used the data in education statistical year books from 1970 to 2010. We used the time-series regression model, time series grouped regression model and exponential smoothing model to predict the number of teachers for the next ten years. Consequently time-series grouped regression model is a better model for forecasting the number of elementary school teachers than other models.

Corporate Default Prediction Model Using Deep Learning Time Series Algorithm, RNN and LSTM (딥러닝 시계열 알고리즘 적용한 기업부도예측모형 유용성 검증)

  • Cha, Sungjae;Kang, Jungseok
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.4
    • /
    • pp.1-32
    • /
    • 2018
  • In addition to stakeholders including managers, employees, creditors, and investors of bankrupt companies, corporate defaults have a ripple effect on the local and national economy. Before the Asian financial crisis, the Korean government only analyzed SMEs and tried to improve the forecasting power of a default prediction model, rather than developing various corporate default models. As a result, even large corporations called 'chaebol enterprises' become bankrupt. Even after that, the analysis of past corporate defaults has been focused on specific variables, and when the government restructured immediately after the global financial crisis, they only focused on certain main variables such as 'debt ratio'. A multifaceted study of corporate default prediction models is essential to ensure diverse interests, to avoid situations like the 'Lehman Brothers Case' of the global financial crisis, to avoid total collapse in a single moment. The key variables used in corporate defaults vary over time. This is confirmed by Beaver (1967, 1968) and Altman's (1968) analysis that Deakins'(1972) study shows that the major factors affecting corporate failure have changed. In Grice's (2001) study, the importance of predictive variables was also found through Zmijewski's (1984) and Ohlson's (1980) models. However, the studies that have been carried out in the past use static models. Most of them do not consider the changes that occur in the course of time. Therefore, in order to construct consistent prediction models, it is necessary to compensate the time-dependent bias by means of a time series analysis algorithm reflecting dynamic change. Based on the global financial crisis, which has had a significant impact on Korea, this study is conducted using 10 years of annual corporate data from 2000 to 2009. Data are divided into training data, validation data, and test data respectively, and are divided into 7, 2, and 1 years respectively. In order to construct a consistent bankruptcy model in the flow of time change, we first train a time series deep learning algorithm model using the data before the financial crisis (2000~2006). The parameter tuning of the existing model and the deep learning time series algorithm is conducted with validation data including the financial crisis period (2007~2008). As a result, we construct a model that shows similar pattern to the results of the learning data and shows excellent prediction power. After that, each bankruptcy prediction model is restructured by integrating the learning data and validation data again (2000 ~ 2008), applying the optimal parameters as in the previous validation. Finally, each corporate default prediction model is evaluated and compared using test data (2009) based on the trained models over nine years. Then, the usefulness of the corporate default prediction model based on the deep learning time series algorithm is proved. In addition, by adding the Lasso regression analysis to the existing methods (multiple discriminant analysis, logit model) which select the variables, it is proved that the deep learning time series algorithm model based on the three bundles of variables is useful for robust corporate default prediction. The definition of bankruptcy used is the same as that of Lee (2015). Independent variables include financial information such as financial ratios used in previous studies. Multivariate discriminant analysis, logit model, and Lasso regression model are used to select the optimal variable group. The influence of the Multivariate discriminant analysis model proposed by Altman (1968), the Logit model proposed by Ohlson (1980), the non-time series machine learning algorithms, and the deep learning time series algorithms are compared. In the case of corporate data, there are limitations of 'nonlinear variables', 'multi-collinearity' of variables, and 'lack of data'. While the logit model is nonlinear, the Lasso regression model solves the multi-collinearity problem, and the deep learning time series algorithm using the variable data generation method complements the lack of data. Big Data Technology, a leading technology in the future, is moving from simple human analysis, to automated AI analysis, and finally towards future intertwined AI applications. Although the study of the corporate default prediction model using the time series algorithm is still in its early stages, deep learning algorithm is much faster than regression analysis at corporate default prediction modeling. Also, it is more effective on prediction power. Through the Fourth Industrial Revolution, the current government and other overseas governments are working hard to integrate the system in everyday life of their nation and society. Yet the field of deep learning time series research for the financial industry is still insufficient. This is an initial study on deep learning time series algorithm analysis of corporate defaults. Therefore it is hoped that it will be used as a comparative analysis data for non-specialists who start a study combining financial data and deep learning time series algorithm.

Liver Cancer Mortality Characteristics and Trends in China from 1991 to 2012

  • Fang, Jia-Ying;Wu, Ku-Sheng;Zeng, Yang;Tang, Wen-Rui;Du, Pei-Ling;Xu, Zhen-Xi;Xu, Xiao-Ling;Luo, Jia-Yi;Lin, Kun
    • Asian Pacific Journal of Cancer Prevention
    • /
    • v.16 no.5
    • /
    • pp.1959-1964
    • /
    • 2015
  • Purpose: To investigate the distribution of liver cancer mortality as well as its developing trend from 1991 to 2012, forecast the future five-year trend, and provide a basis for the comprehensive prevention and management. Materials and Methods: Mortality data for liver cancer in China from 1991 to 2012 were used to describe characteristics and distribution of liver cancer mortality. Trend surface analysis was used to study the geographical distribution of liver cancer mortality. Curve estimation, time series modeling, gray modeling (GM) and joinpoint regression were used to predict and forecast future trends. Results: The mortality rate of liver cancer has constantly increased in China since 1991. Rates in rural areas are higher than in urban areas, and in males are higher than in females. In addition, our data predicted that the trend will continue to increase in the next 5 years. The age-specific mortality of liver cancer increases with age and peaks in the group of 80-84 years old. Geographical analysis showed the liver mortality rate was higher in the southeast provinces, such as Jiangsu, Zhejiang and Guangdong, and southwest regions like Guangxi Province. Conclusions: The standardized mortality rate of liver cancer in China has consistently increased from 1991 to 2012, and the upward trend is predicted to continue in the future. Much better prevention and management of liver cancer is needed in high mortality areas (the southwestern and southeastern parts of China) and high mortality age groups (80- to 84-year-olds), especially in rural areas.

Analysis of Enactment and Utilization of Korean Industrial Standards(KS) by Time Series Data Mining (시계열 자료의 데이터마이닝을 통한 한국산업표준의 제정과 활용 분석)

  • Yoon, Jaekwon;Kim, Wan;Lee, Heesang
    • Journal of Technology Innovation
    • /
    • v.23 no.3
    • /
    • pp.225-253
    • /
    • 2015
  • The standard is a nation's one of the most important industrial issues that improve the social and economic efficiency and also the basis of the industrial development and trade liberalization. This research analyzes the enactment and the utilization of Korean industrial standards(KS) of various industries. This paper examines Korean industries' KS utilization status based on the KS possession, enactments and inquiry records. First, we implement multidimensional scaling method to visualize and group the KS possession records and the nation's institutional issues. We develop several hypothesis to find the decision factors of how each group's KS possession status impacts on the standard enactment activities of similar industry sectors, and analyzes the data by implementing regression analysis. The results show that the capital intensity, R&D activities and sales revenues affect standardization activities. It suggests that the government should encourage companies with high capital intensity, sales revenues to lead the industry's standard activities, and link the policies with the industry's standard and patent related activities from R&D. Second, we analyze the impacts of each KS data's inquiry records, the year of enactments, the form and the industrial segment on the utilization status by implementing statistical analysis and decision tree method. The results show that the enactment year has significant impact on the KS utilization status and some KSs of specific form and industrial segment have high utilization records despite of short enactment history. Our study suggests that government should make policies to utilize the low-utilized KSs and also consider the utilization of standards during the enactment processes.