• Title/Summary/Keyword: Information Management Structure

Search Result 2,757, Processing Time 0.04 seconds

Priority Analysis of Cause Factors of Safety Valve Failure Mode Using Analytical Hierarchy Process (AHP를 활용한 안전밸브(PSV) 고장모드의 Cause Factors 우선순위 분석)

  • Kim, Myung Chul;Lee, Mi Jeong;Lee, Dong Geon;Baek, Jong-Bae
    • Korean Chemical Engineering Research
    • /
    • v.60 no.3
    • /
    • pp.347-355
    • /
    • 2022
  • The safety valve (PSV) is a safety device that automatically releases a spring when the pressure generated by various causes reaches the set pressure, and is restored to a normal state when the pressure falls below a certain level. Periodic inspection and monitoring of safety valves are essential so that they can operate normally in abnormal conditions such as pressure rise. However, as the current safety inspection is performed only at a set period, it is difficult to ensure the safety of normal operation. Therefore, evaluation items were developed by finding failure modes and causative factors of safety valves required for safety management. In addition, it is intended to provide decision-making information for securing safety by deriving the priority of items. To this end, a Delphi survey was conducted three times to derive evaluation factors that were judged to be important in relation to the Failure Mode Cause Factor (FMCFs) of the safety valve (PSV) targeting 15 experts. As a result, 6 failure modes of the safety valve and 22 evaluation factors of its sub-factors were selected. In order to analyze the priorities of the evaluation factors selected in this way, the hierarchical structure was schematized, and the hierarchical decision-making method (AHP) was applied to the priority calculation. As a result of the analysis, the failure mode priorities of FMCFs were 'Leakage' (0.226), 'Fail to open' (0.201), 'Fail to relieve req'd capacity' (0.152), 'Open above set pressure' (0.149), 'Spuriously' 'open' (0.146) and 'Stuck open' (0.127) were confirmed in the order. The lower priority of FMCFs is 'PSV component rupture' (0.109), 'Fail to PSV size calculation' (0.068), 'PSV Spring aging' (0.065), 'Erratic opening' (0.059), 'Damage caused by improper installation and handling' (0.058), 'Fail to spring' (0.053), etc. were checked in the order. It is expected that through efficient management of FMCFs that have been prioritized, it will be possible to identify vulnerabilities of safety valves and contribute to improving safety.

Development of a complex failure prediction system using Hierarchical Attention Network (Hierarchical Attention Network를 이용한 복합 장애 발생 예측 시스템 개발)

  • Park, Youngchan;An, Sangjun;Kim, Mintae;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.4
    • /
    • pp.127-148
    • /
    • 2020
  • The data center is a physical environment facility for accommodating computer systems and related components, and is an essential foundation technology for next-generation core industries such as big data, smart factories, wearables, and smart homes. In particular, with the growth of cloud computing, the proportional expansion of the data center infrastructure is inevitable. Monitoring the health of these data center facilities is a way to maintain and manage the system and prevent failure. If a failure occurs in some elements of the facility, it may affect not only the relevant equipment but also other connected equipment, and may cause enormous damage. In particular, IT facilities are irregular due to interdependence and it is difficult to know the cause. In the previous study predicting failure in data center, failure was predicted by looking at a single server as a single state without assuming that the devices were mixed. Therefore, in this study, data center failures were classified into failures occurring inside the server (Outage A) and failures occurring outside the server (Outage B), and focused on analyzing complex failures occurring within the server. Server external failures include power, cooling, user errors, etc. Since such failures can be prevented in the early stages of data center facility construction, various solutions are being developed. On the other hand, the cause of the failure occurring in the server is difficult to determine, and adequate prevention has not yet been achieved. In particular, this is the reason why server failures do not occur singularly, cause other server failures, or receive something that causes failures from other servers. In other words, while the existing studies assumed that it was a single server that did not affect the servers and analyzed the failure, in this study, the failure occurred on the assumption that it had an effect between servers. In order to define the complex failure situation in the data center, failure history data for each equipment existing in the data center was used. There are four major failures considered in this study: Network Node Down, Server Down, Windows Activation Services Down, and Database Management System Service Down. The failures that occur for each device are sorted in chronological order, and when a failure occurs in a specific equipment, if a failure occurs in a specific equipment within 5 minutes from the time of occurrence, it is defined that the failure occurs simultaneously. After configuring the sequence for the devices that have failed at the same time, 5 devices that frequently occur simultaneously within the configured sequence were selected, and the case where the selected devices failed at the same time was confirmed through visualization. Since the server resource information collected for failure analysis is in units of time series and has flow, we used Long Short-term Memory (LSTM), a deep learning algorithm that can predict the next state through the previous state. In addition, unlike a single server, the Hierarchical Attention Network deep learning model structure was used in consideration of the fact that the level of multiple failures for each server is different. This algorithm is a method of increasing the prediction accuracy by giving weight to the server as the impact on the failure increases. The study began with defining the type of failure and selecting the analysis target. In the first experiment, the same collected data was assumed as a single server state and a multiple server state, and compared and analyzed. The second experiment improved the prediction accuracy in the case of a complex server by optimizing each server threshold. In the first experiment, which assumed each of a single server and multiple servers, in the case of a single server, it was predicted that three of the five servers did not have a failure even though the actual failure occurred. However, assuming multiple servers, all five servers were predicted to have failed. As a result of the experiment, the hypothesis that there is an effect between servers is proven. As a result of this study, it was confirmed that the prediction performance was superior when the multiple servers were assumed than when the single server was assumed. In particular, applying the Hierarchical Attention Network algorithm, assuming that the effects of each server will be different, played a role in improving the analysis effect. In addition, by applying a different threshold for each server, the prediction accuracy could be improved. This study showed that failures that are difficult to determine the cause can be predicted through historical data, and a model that can predict failures occurring in servers in data centers is presented. It is expected that the occurrence of disability can be prevented in advance using the results of this study.

The Improvement of Real-time Updating Methods of the National Base Map Using Building Layout Drawing (건물배치도를 이용한 국가기본도 수시수정 방법 개선)

  • Shin, Chang Soo;Park, Moon Jae;Choi, Yun Soo;Baek, kyu Yeong;Kim, Jaemyeong
    • Journal of Cadastre & Land InformatiX
    • /
    • v.48 no.1
    • /
    • pp.139-151
    • /
    • 2018
  • The National Base Map construction consists of the regular correction work of dividing the whole country into two regions and carrying out the modification Plotting by aerial photographs every two years as well as the real time updating work of correcting the major change feature within two weeks by the field survey and the As-Built Drawing. In the case of the Building Layout Drawing of Korea Real estate Administration intelligence System(KRAS) used for real time updating work of the National base map, the coordinate transformation error is included in the positional error when applied to the National Base Map based on the World Geodetic Reference System as the coordinate system based on the Regional Geodetic Reference System. In addition, National Base Map is registered based on the outline(eaves line) of the building in the Digital Topographic Map, and the Cadastral and Architecture are registered based on the building center line. Therefore, the Building Object management standard is inconsistent. In order to investigate the improvement method, the network RTK survey was conducted directly on a location of the Building Layout Drawing of Korea Real estate Administration intelligence System(KRAS) and the problems were analyzed by comparing with the plane plotting position reference in National Base Map. In the case of the general structure with the difference on the Building center line and the eaves line, beside the location information was different also the difference in the ratio of the building object was different between Building center line and the eave. In conclusion, it is necessary to provide the Base data of the double layer of the Building center line and the outline of the building(eaves line) in order to utilize the Building Layout Drawing of Korea Real estate Administration intelligence System(KRAS). In addition, it is necessary to study an organic map update process that can acquire the up-to-dateness and the accuracy at the same time.

A Study on Collection and Usage of Panel Data on On-board Job Taking and Separation of Korean Seafarers (한국선원의 승선과 이직에 대한 패널자료 구축과 활용방안)

  • Park, Yong-An
    • Journal of Korea Port Economic Association
    • /
    • v.32 no.4
    • /
    • pp.149-163
    • /
    • 2016
  • Seafarers are an essential resource in maritime industries, which provide navigation skills, vessel maneuvering skills and fishing skills in the fishery industry. They also work as a driving force in pilotage, port operation, vessel traffic service, and marine safety. Other areas in maritime services, which rely on seafarer include safety management of ships, supervisory activities, and maritime accident assessment. In these ways, Korean seafarers have contributed to the growth of Korean economy. However, there have been issues of high separation rate, shortage of supply, multi-nationality, multiplicity of culture caused by employment of foreign seafarers, and aging. The present paper finds that maritime officers and fishery officers demonstrate differences in the statistics of on-board job taking and separation: the separation rate of fishery officers is higher than that of maritime officers. The existing data and statistics by the Korea Seafarer's Welfare & Employment Center could be improved by changing its structure from time series to panel data. The Korea Seafarer's Welfare & Employment Center is the ideal institution for collecting the panel data, as it has already accumulated and published relevant statistics regarding seafarer. The basic design method of the panel data is to adopt and improve it by including the information on ratings of maritime and fishery industries, ranks in a ship, personal information, family life, and career goal. Panel data are useful in short- and long-term forecasts of supply of Korean seafarers; demand evaluation of education, training, and reeducation of the seafarers; demographical dynamic analysis on Korean seafarers; inducement policy of long-term on board job taking in harmony with man-power demands in marine industries such as pilotage service; implementation of job attractiveness policy on Korean seafarers; and employment stabilization of Korean seafarers.

Perception and Appraisal of Urban Park Users Using Text Mining of Google Maps Review - Cases of Seoul Forest, Boramae Park, Olympic Park - (구글맵리뷰 텍스트마이닝을 활용한 공원 이용자의 인식 및 평가 - 서울숲, 보라매공원, 올림픽공원을 대상으로 -)

  • Lee, Ju-Kyung;Son, Yong-Hoon
    • Journal of the Korean Institute of Landscape Architecture
    • /
    • v.49 no.4
    • /
    • pp.15-29
    • /
    • 2021
  • The study aims to grasp the perception and appraisal of urban park users through text analysis. This study used Google review data provided by Google Maps. Google Maps Review is an online review platform that provides information evaluating locations through social media and provides an understanding of locations from the perspective of general reviewers and regional guides who are registered as members of Google Maps. The study determined if the Google Maps Reviews were useful for extracting meaningful information about the user perceptions and appraisals for parks management plans. The study chose three urban parks in Seoul, South Korea; Seoul Forest, Boramae Park, and Olympic Park. Review data for each of these three parks were collected via web crawling using Python. Through text analysis, the keywords and network structure characteristics for each park were analyzed. The text was analyzed, as were park ratings, and the analysis compared the reviews of residents and foreign tourists. The common keywords found in the review comments for the three parks were "walking", "bicycle", "rest" and "picnic" for activities, "family", "child" and "dogs" for accompanying types, and "playground" and "walking trail" for park facilities. Looking at the characteristics of each park, Seoul Forest shows many outdoor activities based on nature, while the lack of parking spaces and congestion on weekends negatively impacted users. Boramae Park has the appearance of a city park, with various facilities providing numerous activities, but reviewers often cited the park's complexity and the negative aspects in terms of dog walking groups. At Olympic Park, large-scale complex facilities and cultural events were frequently mentioned, emphasizing its entertainment functions. Google Maps Review can function as useful data to identify parks' overall users' experiences and general feelings. Compared to data from other social media sites, Google Maps Review's data provides ratings and understanding factors, including user satisfaction and dissatisfaction.

Prediction of Distribution Changes of Carpinus laxiflora and C. tschonoskii Based on Climate Change Scenarios Using MaxEnt Model (MaxEnt 모델링을 이용한 기후변화 시나리오에 따른 서어나무 (Carpinus laxiflora)와 개서어나무 (C. tschonoskii)의 분포변화 예측)

  • Lee, Min-Ki;Chun, Jung-Hwa;Lee, Chang-Bae
    • Korean Journal of Agricultural and Forest Meteorology
    • /
    • v.23 no.1
    • /
    • pp.55-67
    • /
    • 2021
  • Hornbeams (Carpinus spp.), which are widely distributed in South Korea, are recognized as one of the most abundant species at climax stage in the temperate forests. Although the distribution and vegetation structure of the C. laxiflora community have been reported, little ecological information of C. tschonoskii is available. Little effort was made to examine the distribution shift of these species under the future climate conditions. This study was conducted to predict potential shifts in the distribution of C. laxiflora and C. tschonoskii in 2050s and 2090s under the two sets of climate change scenarios, RCP4.5 and RCP8.5. The MaxEnt model was used to predict the spatial distribution of two species using the occurrence data derived from the 6th National Forest Inventory data as well as climate and topography data. It was found that the main factors for the distribution of C. laxiflora were elevation, temperature seasonality, and mean annual precipitation. The distribution of C. tschonoskii, was influenced by temperature seasonality, mean annual precipitation, and mean diurnal rang. It was projected that the total habitat area of the C. laxiflora could increase by 1.05% and 1.11% under RCP 4.5 and RCP 8.5 scenarios, respectively. It was also predicted that the distributional area of C. tschonoskii could expand under the future climate conditions. These results highlighted that the climate change would have considerable impact on the spatial distribution of C. laxiflora and C. tschonoskii. These also suggested that ecological information derived from climate change impact assessment study can be used to develop proper forest management practices in response to climate change.

Analysis of Fire Occurrence Characteristics According to Ignition Heat Sources (발화열원에 따른 화재발생 특성 분석)

  • Lee, Kyung-Su;Kim, Tae-Hyeung;Lee, Jae-Ou
    • Journal of the Society of Disaster Information
    • /
    • v.18 no.2
    • /
    • pp.280-289
    • /
    • 2022
  • Purpose: In this study, the characteristics of fire occurrence according to ignition heat sources such as operating equipment, cigarette/lighter fire, and flame/fire were analyzed. Method: One-way ANOVA and cross-analysis were used to analyze the characteristics of fire occurrence by verifying the difference between the ignition environment, fire damage status and scale, and cause of ignition according to the ignition heat source. Result: The fire occurrence characteristics were analyzed through As a result of the analysis, it was found that fires caused by operating devices occurred more frequently on weekdays than other ignition heat sources, and the number of victims and the number of victims were the highest, so mobilization of firefighting power and property damage were the greatest. The initial ignition was generated by electric and electronic devices, and the combustion was expanded by the synthetic resin. For fires caused by cigarette and lighter fires, the most fires occurred on Saturdays and Sundays, and the mobilization of the police force was more characteristic than the mobilization of the firefighting force. In particular, it was found that the initial ignition and combustion expansion were caused by paper, wood, and hay. Fires caused by sparks and sparks occurred most frequently on Saturdays and Sundays, and initial ignition and combustion expansion were found to be caused by paper, wood, and hay. In particular, it showed the characteristic that it occurred in the place farthest from the fire station. The common characteristic of all ignition heat sources was that the fire occurred most frequently in the afternoon time, and the fire type was predominantly the building structure fire, and only the ignition point was burned the most. Conclusion: In order to prevent fire and minimize damage, it is necessary to analyze the tendency of fire occurrence and to prepare appropriate preparations according to the fire occurrence factors. In order to analyze the characteristics of fire occurrence using public data in the future, it is necessary to standardize disaster data and to open and activate data.

Design and Implementation of MongoDB-based Unstructured Log Processing System over Cloud Computing Environment (클라우드 환경에서 MongoDB 기반의 비정형 로그 처리 시스템 설계 및 구현)

  • Kim, Myoungjin;Han, Seungho;Cui, Yun;Lee, Hanku
    • Journal of Internet Computing and Services
    • /
    • v.14 no.6
    • /
    • pp.71-84
    • /
    • 2013
  • Log data, which record the multitude of information created when operating computer systems, are utilized in many processes, from carrying out computer system inspection and process optimization to providing customized user optimization. In this paper, we propose a MongoDB-based unstructured log processing system in a cloud environment for processing the massive amount of log data of banks. Most of the log data generated during banking operations come from handling a client's business. Therefore, in order to gather, store, categorize, and analyze the log data generated while processing the client's business, a separate log data processing system needs to be established. However, the realization of flexible storage expansion functions for processing a massive amount of unstructured log data and executing a considerable number of functions to categorize and analyze the stored unstructured log data is difficult in existing computer environments. Thus, in this study, we use cloud computing technology to realize a cloud-based log data processing system for processing unstructured log data that are difficult to process using the existing computing infrastructure's analysis tools and management system. The proposed system uses the IaaS (Infrastructure as a Service) cloud environment to provide a flexible expansion of computing resources and includes the ability to flexibly expand resources such as storage space and memory under conditions such as extended storage or rapid increase in log data. Moreover, to overcome the processing limits of the existing analysis tool when a real-time analysis of the aggregated unstructured log data is required, the proposed system includes a Hadoop-based analysis module for quick and reliable parallel-distributed processing of the massive amount of log data. Furthermore, because the HDFS (Hadoop Distributed File System) stores data by generating copies of the block units of the aggregated log data, the proposed system offers automatic restore functions for the system to continually operate after it recovers from a malfunction. Finally, by establishing a distributed database using the NoSQL-based Mongo DB, the proposed system provides methods of effectively processing unstructured log data. Relational databases such as the MySQL databases have complex schemas that are inappropriate for processing unstructured log data. Further, strict schemas like those of relational databases cannot expand nodes in the case wherein the stored data are distributed to various nodes when the amount of data rapidly increases. NoSQL does not provide the complex computations that relational databases may provide but can easily expand the database through node dispersion when the amount of data increases rapidly; it is a non-relational database with an appropriate structure for processing unstructured data. The data models of the NoSQL are usually classified as Key-Value, column-oriented, and document-oriented types. Of these, the representative document-oriented data model, MongoDB, which has a free schema structure, is used in the proposed system. MongoDB is introduced to the proposed system because it makes it easy to process unstructured log data through a flexible schema structure, facilitates flexible node expansion when the amount of data is rapidly increasing, and provides an Auto-Sharding function that automatically expands storage. The proposed system is composed of a log collector module, a log graph generator module, a MongoDB module, a Hadoop-based analysis module, and a MySQL module. When the log data generated over the entire client business process of each bank are sent to the cloud server, the log collector module collects and classifies data according to the type of log data and distributes it to the MongoDB module and the MySQL module. The log graph generator module generates the results of the log analysis of the MongoDB module, Hadoop-based analysis module, and the MySQL module per analysis time and type of the aggregated log data, and provides them to the user through a web interface. Log data that require a real-time log data analysis are stored in the MySQL module and provided real-time by the log graph generator module. The aggregated log data per unit time are stored in the MongoDB module and plotted in a graph according to the user's various analysis conditions. The aggregated log data in the MongoDB module are parallel-distributed and processed by the Hadoop-based analysis module. A comparative evaluation is carried out against a log data processing system that uses only MySQL for inserting log data and estimating query performance; this evaluation proves the proposed system's superiority. Moreover, an optimal chunk size is confirmed through the log data insert performance evaluation of MongoDB for various chunk sizes.

Development of Predictive Models for Rights Issues Using Financial Analysis Indices and Decision Tree Technique (경영분석지표와 의사결정나무기법을 이용한 유상증자 예측모형 개발)

  • Kim, Myeong-Kyun;Cho, Yoonho
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.4
    • /
    • pp.59-77
    • /
    • 2012
  • This study focuses on predicting which firms will increase capital by issuing new stocks in the near future. Many stakeholders, including banks, credit rating agencies and investors, performs a variety of analyses for firms' growth, profitability, stability, activity, productivity, etc., and regularly report the firms' financial analysis indices. In the paper, we develop predictive models for rights issues using these financial analysis indices and data mining techniques. This study approaches to building the predictive models from the perspective of two different analyses. The first is the analysis period. We divide the analysis period into before and after the IMF financial crisis, and examine whether there is the difference between the two periods. The second is the prediction time. In order to predict when firms increase capital by issuing new stocks, the prediction time is categorized as one year, two years and three years later. Therefore Total six prediction models are developed and analyzed. In this paper, we employ the decision tree technique to build the prediction models for rights issues. The decision tree is the most widely used prediction method which builds decision trees to label or categorize cases into a set of known classes. In contrast to neural networks, logistic regression and SVM, decision tree techniques are well suited for high-dimensional applications and have strong explanation capabilities. There are well-known decision tree induction algorithms such as CHAID, CART, QUEST, C5.0, etc. Among them, we use C5.0 algorithm which is the most recently developed algorithm and yields performance better than other algorithms. We obtained data for the rights issue and financial analysis from TS2000 of Korea Listed Companies Association. A record of financial analysis data is consisted of 89 variables which include 9 growth indices, 30 profitability indices, 23 stability indices, 6 activity indices and 8 productivity indices. For the model building and test, we used 10,925 financial analysis data of total 658 listed firms. PASW Modeler 13 was used to build C5.0 decision trees for the six prediction models. Total 84 variables among financial analysis data are selected as the input variables of each model, and the rights issue status (issued or not issued) is defined as the output variable. To develop prediction models using C5.0 node (Node Options: Output type = Rule set, Use boosting = false, Cross-validate = false, Mode = Simple, Favor = Generality), we used 60% of data for model building and 40% of data for model test. The results of experimental analysis show that the prediction accuracies of data after the IMF financial crisis (59.04% to 60.43%) are about 10 percent higher than ones before IMF financial crisis (68.78% to 71.41%). These results indicate that since the IMF financial crisis, the reliability of financial analysis indices has increased and the firm intention of rights issue has been more obvious. The experiment results also show that the stability-related indices have a major impact on conducting rights issue in the case of short-term prediction. On the other hand, the long-term prediction of conducting rights issue is affected by financial analysis indices on profitability, stability, activity and productivity. All the prediction models include the industry code as one of significant variables. This means that companies in different types of industries show their different types of patterns for rights issue. We conclude that it is desirable for stakeholders to take into account stability-related indices and more various financial analysis indices for short-term prediction and long-term prediction, respectively. The current study has several limitations. First, we need to compare the differences in accuracy by using different data mining techniques such as neural networks, logistic regression and SVM. Second, we are required to develop and to evaluate new prediction models including variables which research in the theory of capital structure has mentioned about the relevance to rights issue.

A Hybrid SVM Classifier for Imbalanced Data Sets (불균형 데이터 집합의 분류를 위한 하이브리드 SVM 모델)

  • Lee, Jae Sik;Kwon, Jong Gu
    • Journal of Intelligence and Information Systems
    • /
    • v.19 no.2
    • /
    • pp.125-140
    • /
    • 2013
  • We call a data set in which the number of records belonging to a certain class far outnumbers the number of records belonging to the other class, 'imbalanced data set'. Most of the classification techniques perform poorly on imbalanced data sets. When we evaluate the performance of a certain classification technique, we need to measure not only 'accuracy' but also 'sensitivity' and 'specificity'. In a customer churn prediction problem, 'retention' records account for the majority class, and 'churn' records account for the minority class. Sensitivity measures the proportion of actual retentions which are correctly identified as such. Specificity measures the proportion of churns which are correctly identified as such. The poor performance of the classification techniques on imbalanced data sets is due to the low value of specificity. Many previous researches on imbalanced data sets employed 'oversampling' technique where members of the minority class are sampled more than those of the majority class in order to make a relatively balanced data set. When a classification model is constructed using this oversampled balanced data set, specificity can be improved but sensitivity will be decreased. In this research, we developed a hybrid model of support vector machine (SVM), artificial neural network (ANN) and decision tree, that improves specificity while maintaining sensitivity. We named this hybrid model 'hybrid SVM model.' The process of construction and prediction of our hybrid SVM model is as follows. By oversampling from the original imbalanced data set, a balanced data set is prepared. SVM_I model and ANN_I model are constructed using the imbalanced data set, and SVM_B model is constructed using the balanced data set. SVM_I model is superior in sensitivity and SVM_B model is superior in specificity. For a record on which both SVM_I model and SVM_B model make the same prediction, that prediction becomes the final solution. If they make different prediction, the final solution is determined by the discrimination rules obtained by ANN and decision tree. For a record on which SVM_I model and SVM_B model make different predictions, a decision tree model is constructed using ANN_I output value as input and actual retention or churn as target. We obtained the following two discrimination rules: 'IF ANN_I output value <0.285, THEN Final Solution = Retention' and 'IF ANN_I output value ${\geq}0.285$, THEN Final Solution = Churn.' The threshold 0.285 is the value optimized for the data used in this research. The result we present in this research is the structure or framework of our hybrid SVM model, not a specific threshold value such as 0.285. Therefore, the threshold value in the above discrimination rules can be changed to any value depending on the data. In order to evaluate the performance of our hybrid SVM model, we used the 'churn data set' in UCI Machine Learning Repository, that consists of 85% retention customers and 15% churn customers. Accuracy of the hybrid SVM model is 91.08% that is better than that of SVM_I model or SVM_B model. The points worth noticing here are its sensitivity, 95.02%, and specificity, 69.24%. The sensitivity of SVM_I model is 94.65%, and the specificity of SVM_B model is 67.00%. Therefore the hybrid SVM model developed in this research improves the specificity of SVM_B model while maintaining the sensitivity of SVM_I model.