• Title/Summary/Keyword: Variable Output

Search Result 1,169, Processing Time 0.025 seconds

A Case Study on the Exogenous Factors affecting Extra-large Egg Production in a Layer Farm in Korea (산란계 사육농장 특란 생산에 미치는 외부 요인 분석을 위한 사례 연구)

  • Lee, Hyun-Chang;Jang, Woo-Whan
    • Korean Journal of Poultry Science
    • /
    • v.41 no.2
    • /
    • pp.99-104
    • /
    • 2014
  • The objective of this study is to analyze the production of extra-large egg and assess the impacts of exogenous factors in feeding the layer chicken. The main results of this study are as follows; First, feeding rations on the basics of statistics, internal maximum and minimum temperature and, the age at first egg affect the production of extra-large egg. Second, implicating the standardized coefficients from the conclusion of regression model estimating suggest that the amount of feed has the greatest impact on production followed by the age at first egg. Third, by using the elasticity of output and the volatility in the production, the result suggest that among the independent variable factors in the external volatility, the biggest one goes to feed ration, and the age at first egg follows. In order to control the production volatility in the extra-large egg production of the farms, it is necessary to manage an efficient feeding based on feed ration, age at first egg and, the maximum and minimum temperature inside the farm. Taken together, the results demonstrates that it should be concentrated by controlling the exogenous factors affecting extra large egg production and the management system construct, to increase extra-large egg production and the income of farmers at the same time.

The Prediction of DEA based Efficiency Rating for Venture Business Using Multi-class SVM (다분류 SVM을 이용한 DEA기반 벤처기업 효율성등급 예측모형)

  • Park, Ji-Young;Hong, Tae-Ho
    • Asia pacific journal of information systems
    • /
    • v.19 no.2
    • /
    • pp.139-155
    • /
    • 2009
  • For the last few decades, many studies have tried to explore and unveil venture companies' success factors and unique features in order to identify the sources of such companies' competitive advantages over their rivals. Such venture companies have shown tendency to give high returns for investors generally making the best use of information technology. For this reason, many venture companies are keen on attracting avid investors' attention. Investors generally make their investment decisions by carefully examining the evaluation criteria of the alternatives. To them, credit rating information provided by international rating agencies, such as Standard and Poor's, Moody's and Fitch is crucial source as to such pivotal concerns as companies stability, growth, and risk status. But these types of information are generated only for the companies issuing corporate bonds, not venture companies. Therefore, this study proposes a method for evaluating venture businesses by presenting our recent empirical results using financial data of Korean venture companies listed on KOSDAQ in Korea exchange. In addition, this paper used multi-class SVM for the prediction of DEA-based efficiency rating for venture businesses, which was derived from our proposed method. Our approach sheds light on ways to locate efficient companies generating high level of profits. Above all, in determining effective ways to evaluate a venture firm's efficiency, it is important to understand the major contributing factors of such efficiency. Therefore, this paper is constructed on the basis of following two ideas to classify which companies are more efficient venture companies: i) making DEA based multi-class rating for sample companies and ii) developing multi-class SVM-based efficiency prediction model for classifying all companies. First, the Data Envelopment Analysis(DEA) is a non-parametric multiple input-output efficiency technique that measures the relative efficiency of decision making units(DMUs) using a linear programming based model. It is non-parametric because it requires no assumption on the shape or parameters of the underlying production function. DEA has been already widely applied for evaluating the relative efficiency of DMUs. Recently, a number of DEA based studies have evaluated the efficiency of various types of companies, such as internet companies and venture companies. It has been also applied to corporate credit ratings. In this study we utilized DEA for sorting venture companies by efficiency based ratings. The Support Vector Machine(SVM), on the other hand, is a popular technique for solving data classification problems. In this paper, we employed SVM to classify the efficiency ratings in IT venture companies according to the results of DEA. The SVM method was first developed by Vapnik (1995). As one of many machine learning techniques, SVM is based on a statistical theory. Thus far, the method has shown good performances especially in generalizing capacity in classification tasks, resulting in numerous applications in many areas of business, SVM is basically the algorithm that finds the maximum margin hyperplane, which is the maximum separation between classes. According to this method, support vectors are the closest to the maximum margin hyperplane. If it is impossible to classify, we can use the kernel function. In the case of nonlinear class boundaries, we can transform the inputs into a high-dimensional feature space, This is the original input space and is mapped into a high-dimensional dot-product space. Many studies applied SVM to the prediction of bankruptcy, the forecast a financial time series, and the problem of estimating credit rating, In this study we employed SVM for developing data mining-based efficiency prediction model. We used the Gaussian radial function as a kernel function of SVM. In multi-class SVM, we adopted one-against-one approach between binary classification method and two all-together methods, proposed by Weston and Watkins(1999) and Crammer and Singer(2000), respectively. In this research, we used corporate information of 154 companies listed on KOSDAQ market in Korea exchange. We obtained companies' financial information of 2005 from the KIS(Korea Information Service, Inc.). Using this data, we made multi-class rating with DEA efficiency and built multi-class prediction model based data mining. Among three manners of multi-classification, the hit ratio of the Weston and Watkins method is the best in the test data set. In multi classification problems as efficiency ratings of venture business, it is very useful for investors to know the class with errors, one class difference, when it is difficult to find out the accurate class in the actual market. So we presented accuracy results within 1-class errors, and the Weston and Watkins method showed 85.7% accuracy in our test samples. We conclude that the DEA based multi-class approach in venture business generates more information than the binary classification problem, notwithstanding its efficiency level. We believe this model can help investors in decision making as it provides a reliably tool to evaluate venture companies in the financial domain. For the future research, we perceive the need to enhance such areas as the variable selection process, the parameter selection of kernel function, the generalization, and the sample size of multi-class.

Evaluation of Methane Generation Rate Constant(k) by Estimating Greenhouse Gas Emission in Small Scale Landfill (소규모 매립지에 대한 메탄발생속도상수(k) 산출 및 온실가스 발생량 평가)

  • Lee, Wonjae;Kang, Byungwook;Cho, Byungyeol;Lee, Sangwoo;Yeon, Ikjun
    • Journal of the Korean GEO-environmental Society
    • /
    • v.15 no.5
    • /
    • pp.5-11
    • /
    • 2014
  • In this study, greenhouse gas emission for small scale landfill (H and Y landfill) was investigated to deduce special the methane generation rate constant(k). To achieve the purpose, the data of physical composition was collected and amount of LFG emission was calculated by using FOD method suggested in 2006 IPCC GL. Also, amount of LFG emission was directly measured in the active landfill sites. By comparing the results, the methane generation rate constant(k), which was used as input variable in FOD method suggested in 2006 IPCC GL, was deduced. From the results on the physical composition, it was shown that the ranges of DOC per year in H (1997~2011) and Y (1994~2011) landfill sites were 13.16 %~23.79 % ($16.52{\pm}3.84%$) and 7.24 %~34.67 % ($14.56{\pm}7.30%$), respectively. The DOC results showed the differences with the suggested values (= 18 %) in 2006 IPCC GL. The average values of methane generation rate constant(k) from each landfill site were $0.0413yr^{-1}$ and $0.0117yr^{-1}$. The results of methane generation rate constant(k) was shown big difference with 2006 IPCC GL defualt value (k = 0.09). It was confirmed that calculation results of greenhouse gas emission using default value in 2006 IPCC GL show excessive output.

An Empirical Comparison and Verification Study on the Seaport Clustering Measurement Using Meta-Frontier DEA and Integer Programming Models (메타프론티어 DEA모형과 정수계획모형을 이용한 항만클러스터링 측정에 대한 실증적 비교 및 검증연구)

  • Park, Ro-Kyung
    • Journal of Korea Port Economic Association
    • /
    • v.33 no.2
    • /
    • pp.53-82
    • /
    • 2017
  • The purpose of this study is to show the clustering trend and compare empirical results, as well as to choose the clustering ports for 3 Korean ports (Busan, Incheon, and Gwangyang) by using meta-frontier DEA (Data Envelopment Analysis) and integer models on 38 Asian container ports over the period 2005-2014. The models consider 4 input variables (birth length, depth, total area, and number of cranes) and 1 output variable (container TEU). The main empirical results of the study are as follows. First, the meta-frontier DEA for Chinese seaports identifies as most efficient ports (in decreasing order) Shanghai, Hongkong, Ningbo, Qingdao, and Guangzhou, while efficient Korean seaports are Busan, Incheon, and Gwangyang. Second, the clustering results of the integer model show that the Busan port should cluster with Dubai, Hongkong, Shanghai, Guangzhou, Ningbo, Qingdao, Singapore, and Kaosiung, while Incheon and Gwangyang should cluster with Shahid Rajaee, Haifa, Khor Fakkan, Tanjung Perak, Osaka, Keelong, and Bangkok ports. Third, clustering through the integer model sharply increases the group efficiency of Incheon (401.84%) and Gwangyang (354.25%), but not that of the Busan port. Fourth, the efficiency ranking comparison between the two models before and after the clustering using the Wilcoxon signed-rank test is matched with the average level of group efficiency (57.88 %) and the technology gap ratio (80.93%). The policy implication of this study is that Korean port policy planners should employ meta-frontier DEA, as well as integer models when clustering is needed among Asian container ports for enhancing the efficiency. In addition Korean seaport managers and port authorities should introduce port development and management plans accounting for the reference and clustered seaports after careful analysis.

Analyzing the Efficiency of Korean Rail Transit Properties using Data Envelopment Analysis (자료포락분석기법을 이용한 도시철도 운영기관의 효율성 분석)

  • 김민정;김성수
    • Journal of Korean Society of Transportation
    • /
    • v.21 no.4
    • /
    • pp.113-132
    • /
    • 2003
  • Using nonradial data envelopment analysis(DEA) under assumptions of strong disposability and variable returns scale, this paper annually estimates productive. technical and allocative efficiencies of three publicly-owned rail transit properties which are different in terms of organizational type: Seoul Subway Corporation(SSC, local public corporation), the Seoul Metropolitan Electrified Railways sector (SMESRS) of Korea National Railroad(the national railway operator controlled by the Ministry of Construction and Transportation(MOCT)), and Busan Urban Transit Authority (BUTA, the national authority controlled by MOCT). Using the estimation results of Tobit regression analysis. the paper next computes their true productive, true technical and true allocative efficiencies, which reflect only the impacts of internal factors such as production activity by removing the impacts of external factors such as an organizational type and a track utilization rate. And the paper also computes an organizational efficiency and annually gross efficiencies for each property. The paper then conceptualized that the property produces a single output(car-kilometers) using four inputs(labor, electricity, car & maintenance and track) and uses unbalanced panel data consisted of annual observations on SSC, SMESRS and BUTA. The results obtained from DEA show that, on an average, SSC is the most efficient property on the productive and allocative sides, while SMESRS is the most technically-efficient one. On the other hand. BUTA is the most efficient one on the truly-productive and allocative sides, while SMESRS on the truly-technical side. Another important result is that the differences in true efficiency estimates among the three properties are considerably smaller than those in efficiency estimates. Besides. the most cost-efficient organizational type appears to be a local public corporation represented by SSC, which is also the most grossly-efficient property. These results suggest that a measure to sort out the impacts of external factors on the efficiency of rail transit properties is required to assess fairly it, and that a measure to restructure (establish) an existing(a new) rail transit property into a local public corporation(or authority) is required to improve its cost efficiency.

An Empirical Comparative Study of the Seaport Clustering Measurement Using Bootstrapped DEA and Game Cross-efficiency Models (부트스트랩 DEA모형과 게임교차효율성모형을 이용한 항만클러스터링 측정에 대한 실증적 비교연구)

  • Park, Ro-Kyung
    • Journal of Korea Port Economic Association
    • /
    • v.32 no.1
    • /
    • pp.29-58
    • /
    • 2016
  • The purpose of this paper is to show the clustering trend and the comparison of empirical results and is to choose the clustering ports for 3 Korean ports(Busan, Incheon and Gwangyang Ports) by using the bootstrapped DEA(Data Envelopment Analysis) and game Cross-efficiency models for 38 Asian ports during the period 2003-2013 with 4 input variables(birth length, depth, total area, and number of cranes) and 1 output variable(container TEU). The main empirical results of this paper are as follows. First, bootstrapped DEA efficiency of SW and LT is 0.7660, 0.7341 respectively. Clustering results of the bootstrapped DEA analysis show that 3 Korean ports [ Busan (6.46%), Incheon (3.92%), and Gwangyang (2.78%)] can increase the efficiency in the SW model, but the LT model shows clustering values of -1.86%, -0.124%, and 2.11% for Busan, Gwangyang, and Incheon respectively. Second, the game cross-efficiency model suggests that Korean ports should be clustered with Hong Kong, Shanghi, Guangzhou, Ningbo, Port Klang, Singapore, Kaosiung, Keelong, and Bangkok ports. This clustering enhances the efficiency of Gwangyang by 0.131%, and decreases that of Busan by-1.08%, and that of Incheon by -0.009%. Third, the efficiency ranking comparison between the two models using the Wilcoxon Signed-rank Test was matched with the average level of SW (72.83 %) and LT (68.91%). The policy implication of this paper is that Korean port policy planners should introduce the bootstrapped DEA, and game cross-efficiency models when clustering is needed among Asian ports for enhancing the efficiency of inputs and outputs. Also, the results of SWOT(Strength, Weakness, Opportunity, and Threat) analysis among the clustering ports should be considered.

Study on water quality prediction in water treatment plants using AI techniques (AI 기법을 활용한 정수장 수질예측에 관한 연구)

  • Lee, Seungmin;Kang, Yujin;Song, Jinwoo;Kim, Juhwan;Kim, Hung Soo;Kim, Soojun
    • Journal of Korea Water Resources Association
    • /
    • v.57 no.3
    • /
    • pp.151-164
    • /
    • 2024
  • In water treatment plants supplying potable water, the management of chlorine concentration in water treatment processes involving pre-chlorination or intermediate chlorination requires process control. To address this, research has been conducted on water quality prediction techniques utilizing AI technology. This study developed an AI-based predictive model for automating the process control of chlorine disinfection, targeting the prediction of residual chlorine concentration downstream of sedimentation basins in water treatment processes. The AI-based model, which learns from past water quality observation data to predict future water quality, offers a simpler and more efficient approach compared to complex physicochemical and biological water quality models. The model was tested by predicting the residual chlorine concentration downstream of the sedimentation basins at Plant, using multiple regression models and AI-based models like Random Forest and LSTM, and the results were compared. For optimal prediction of residual chlorine concentration, the input-output structure of the AI model included the residual chlorine concentration upstream of the sedimentation basin, turbidity, pH, water temperature, electrical conductivity, inflow of raw water, alkalinity, NH3, etc. as independent variables, and the desired residual chlorine concentration of the effluent from the sedimentation basin as the dependent variable. The independent variables were selected from observable data at the water treatment plant, which are influential on the residual chlorine concentration downstream of the sedimentation basin. The analysis showed that, for Plant, the model based on Random Forest had the lowest error compared to multiple regression models, neural network models, model trees, and other Random Forest models. The optimal predicted residual chlorine concentration downstream of the sedimentation basin presented in this study is expected to enable real-time control of chlorine dosing in previous treatment stages, thereby enhancing water treatment efficiency and reducing chemical costs.

Feasibility of Deep Learning Algorithms for Binary Classification Problems (이진 분류문제에서의 딥러닝 알고리즘의 활용 가능성 평가)

  • Kim, Kitae;Lee, Bomi;Kim, Jong Woo
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.1
    • /
    • pp.95-108
    • /
    • 2017
  • Recently, AlphaGo which is Bakuk (Go) artificial intelligence program by Google DeepMind, had a huge victory against Lee Sedol. Many people thought that machines would not be able to win a man in Go games because the number of paths to make a one move is more than the number of atoms in the universe unlike chess, but the result was the opposite to what people predicted. After the match, artificial intelligence technology was focused as a core technology of the fourth industrial revolution and attracted attentions from various application domains. Especially, deep learning technique have been attracted as a core artificial intelligence technology used in the AlphaGo algorithm. The deep learning technique is already being applied to many problems. Especially, it shows good performance in image recognition field. In addition, it shows good performance in high dimensional data area such as voice, image and natural language, which was difficult to get good performance using existing machine learning techniques. However, in contrast, it is difficult to find deep leaning researches on traditional business data and structured data analysis. In this study, we tried to find out whether the deep learning techniques have been studied so far can be used not only for the recognition of high dimensional data but also for the binary classification problem of traditional business data analysis such as customer churn analysis, marketing response prediction, and default prediction. And we compare the performance of the deep learning techniques with that of traditional artificial neural network models. The experimental data in the paper is the telemarketing response data of a bank in Portugal. It has input variables such as age, occupation, loan status, and the number of previous telemarketing and has a binary target variable that records whether the customer intends to open an account or not. In this study, to evaluate the possibility of utilization of deep learning algorithms and techniques in binary classification problem, we compared the performance of various models using CNN, LSTM algorithm and dropout, which are widely used algorithms and techniques in deep learning, with that of MLP models which is a traditional artificial neural network model. However, since all the network design alternatives can not be tested due to the nature of the artificial neural network, the experiment was conducted based on restricted settings on the number of hidden layers, the number of neurons in the hidden layer, the number of output data (filters), and the application conditions of the dropout technique. The F1 Score was used to evaluate the performance of models to show how well the models work to classify the interesting class instead of the overall accuracy. The detail methods for applying each deep learning technique in the experiment is as follows. The CNN algorithm is a method that reads adjacent values from a specific value and recognizes the features, but it does not matter how close the distance of each business data field is because each field is usually independent. In this experiment, we set the filter size of the CNN algorithm as the number of fields to learn the whole characteristics of the data at once, and added a hidden layer to make decision based on the additional features. For the model having two LSTM layers, the input direction of the second layer is put in reversed position with first layer in order to reduce the influence from the position of each field. In the case of the dropout technique, we set the neurons to disappear with a probability of 0.5 for each hidden layer. The experimental results show that the predicted model with the highest F1 score was the CNN model using the dropout technique, and the next best model was the MLP model with two hidden layers using the dropout technique. In this study, we were able to get some findings as the experiment had proceeded. First, models using dropout techniques have a slightly more conservative prediction than those without dropout techniques, and it generally shows better performance in classification. Second, CNN models show better classification performance than MLP models. This is interesting because it has shown good performance in binary classification problems which it rarely have been applied to, as well as in the fields where it's effectiveness has been proven. Third, the LSTM algorithm seems to be unsuitable for binary classification problems because the training time is too long compared to the performance improvement. From these results, we can confirm that some of the deep learning algorithms can be applied to solve business binary classification problems.

Application of Support Vector Regression for Improving the Performance of the Emotion Prediction Model (감정예측모형의 성과개선을 위한 Support Vector Regression 응용)

  • Kim, Seongjin;Ryoo, Eunchung;Jung, Min Kyu;Kim, Jae Kyeong;Ahn, Hyunchul
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.3
    • /
    • pp.185-202
    • /
    • 2012
  • .Since the value of information has been realized in the information society, the usage and collection of information has become important. A facial expression that contains thousands of information as an artistic painting can be described in thousands of words. Followed by the idea, there has recently been a number of attempts to provide customers and companies with an intelligent service, which enables the perception of human emotions through one's facial expressions. For example, MIT Media Lab, the leading organization in this research area, has developed the human emotion prediction model, and has applied their studies to the commercial business. In the academic area, a number of the conventional methods such as Multiple Regression Analysis (MRA) or Artificial Neural Networks (ANN) have been applied to predict human emotion in prior studies. However, MRA is generally criticized because of its low prediction accuracy. This is inevitable since MRA can only explain the linear relationship between the dependent variables and the independent variable. To mitigate the limitations of MRA, some studies like Jung and Kim (2012) have used ANN as the alternative, and they reported that ANN generated more accurate prediction than the statistical methods like MRA. However, it has also been criticized due to over fitting and the difficulty of the network design (e.g. setting the number of the layers and the number of the nodes in the hidden layers). Under this background, we propose a novel model using Support Vector Regression (SVR) in order to increase the prediction accuracy. SVR is an extensive version of Support Vector Machine (SVM) designated to solve the regression problems. The model produced by SVR only depends on a subset of the training data, because the cost function for building the model ignores any training data that is close (within a threshold ${\varepsilon}$) to the model prediction. Using SVR, we tried to build a model that can measure the level of arousal and valence from the facial features. To validate the usefulness of the proposed model, we collected the data of facial reactions when providing appropriate visual stimulating contents, and extracted the features from the data. Next, the steps of the preprocessing were taken to choose statistically significant variables. In total, 297 cases were used for the experiment. As the comparative models, we also applied MRA and ANN to the same data set. For SVR, we adopted '${\varepsilon}$-insensitive loss function', and 'grid search' technique to find the optimal values of the parameters like C, d, ${\sigma}^2$, and ${\varepsilon}$. In the case of ANN, we adopted a standard three-layer backpropagation network, which has a single hidden layer. The learning rate and momentum rate of ANN were set to 10%, and we used sigmoid function as the transfer function of hidden and output nodes. We performed the experiments repeatedly by varying the number of nodes in the hidden layer to n/2, n, 3n/2, and 2n, where n is the number of the input variables. The stopping condition for ANN was set to 50,000 learning events. And, we used MAE (Mean Absolute Error) as the measure for performance comparison. From the experiment, we found that SVR achieved the highest prediction accuracy for the hold-out data set compared to MRA and ANN. Regardless of the target variables (the level of arousal, or the level of positive / negative valence), SVR showed the best performance for the hold-out data set. ANN also outperformed MRA, however, it showed the considerably lower prediction accuracy than SVR for both target variables. The findings of our research are expected to be useful to the researchers or practitioners who are willing to build the models for recognizing human emotions.