• Title/Summary/Keyword: Success Prediction

Search Result 199, Processing Time 0.023 seconds

Extracting the Distribution Potential Area of Debris Landform Using a Fuzzy Set Model (퍼지집합 모델을 이용한 암설지형 분포 가능지 추출 연구)

  • Wi, Nun-Sol;JANG, Dong-Ho
    • Journal of The Geomorphological Association of Korea
    • /
    • v.24 no.1
    • /
    • pp.77-91
    • /
    • 2017
  • Many debris landforms in the mountains of Korea have formed in the periglacial environment during the last glacial stage when the generation of sediments was active. Because these landforms are generally located on steep slopes and mostly covered by vegetation, however, it is difficult to observe and access them through field investigation. A scientific method is required to reduce the survey range before performing field investigation and to save time and cost. For this purpose, the use of remote sensing and GIS technologies is essential. This study has extracted the potential area of debris landform formation using a fuzzy set model as a mathematical data integration method. The first step was to obtain information about the location of debris landforms and their related factors. This information was verified through field observation and then used to build a database. In the second step, we conducted the fuzzy set modeling to generate a map, which classified the study area based on the possibility of debris formation. We then applied a cross-validation technique in order to evaluate the map. For a quantitative analysis, the calculated potential rate of debris formation was evaluated by plotting SRC(Success Rate Curve) and calculating AUC(Area Under the Curve). The prediction accuracy of the model was found to be 83.1%. We posit that the model is accurate and reliable enough to contribute to efficient field investigation and debris landform management.

Image Augmentation of Paralichthys Olivaceus Disease Using SinGAN Deep Learning Model (SinGAN 딥러닝 모델을 이용한 넙치 질병 이미지 증강)

  • Son, Hyun Seung;Choi, Han Suk
    • The Journal of the Korea Contents Association
    • /
    • v.21 no.12
    • /
    • pp.322-330
    • /
    • 2021
  • In modern aquaculture, mass mortality is a very important issue that determines the success of aquaculture business. If a fish disease is not detected at an early stage in the farm, the disease spreads quickly because the farm is a closed environment. Therefore, early detection of diseases is crucial to prevent mass mortality of fish raised in farms. Recently deep learning-based automatic identification of fish diseases has been widely used, but there are many difficulties in identifying objects due to insufficient images of fish diseases. Therefore, this paper suggests a method to generate a large number of fish disease images by synthesizing normal images and disease images using SinGAN deep learning model in order to to solve the lack of fish disease images. We generate images from the three most frequently occurring Paralichthys Olivaceus diseases such as Scuticociliatida, Vibriosis, and Lymphocytosis and compare them with the original image. In this study, a total of 330 sheets of scutica disease, 110 sheets of vibrioemia, and 110 sheets of limphosis were made by synthesizing 10 disease patterns with 11 normal halibut images, and 1,320 images were produced by quadrupling the images.

A gene expression programming-based model to predict water inflow into tunnels

  • Arsalan Mahmoodzadeh;Hawkar Hashim Ibrahim;Laith R. Flaih;Abed Alanazi;Abdullah Alqahtani;Shtwai Alsubai;Nabil Ben Kahla;Adil Hussein Mohammed
    • Geomechanics and Engineering
    • /
    • v.37 no.1
    • /
    • pp.65-72
    • /
    • 2024
  • Water ingress poses a common and intricate geological hazard with profound implications for tunnel construction's speed and safety. The project's success hinges significantly on the precision of estimating water inflow during excavation, a critical factor in early-stage decision-making during conception and design. This article introduces an optimized model employing the gene expression programming (GEP) approach to forecast tunnel water inflow. The GEP model was refined by developing an equation that best aligns with predictive outcomes. The equation's outputs were compared with measured data and assessed against practical scenarios to validate its potential applicability in calculating tunnel water input. The optimized GEP model excelled in forecasting tunnel water inflow, outperforming alternative machine learning algorithms like SVR, GPR, DT, and KNN. This positions the GEP model as a leading choice for accurate and superior predictions. A state-of-the-art machine learning-based graphical user interface (GUI) was innovatively crafted for predicting and visualizing tunnel water inflow. This cutting-edge tool leverages ML algorithms, marking a substantial advancement in tunneling prediction technologies, providing accuracy and accessibility in water inflow projections.

A Hybrid Multi-Level Feature Selection Framework for prediction of Chronic Disease

  • G.S. Raghavendra;Shanthi Mahesh;M.V.P. Chandrasekhara Rao
    • International Journal of Computer Science & Network Security
    • /
    • v.23 no.12
    • /
    • pp.101-106
    • /
    • 2023
  • Chronic illnesses are among the most common serious problems affecting human health. Early diagnosis of chronic diseases can assist to avoid or mitigate their consequences, potentially decreasing mortality rates. Using machine learning algorithms to identify risk factors is an exciting strategy. The issue with existing feature selection approaches is that each method provides a distinct set of properties that affect model correctness, and present methods cannot perform well on huge multidimensional datasets. We would like to introduce a novel model that contains a feature selection approach that selects optimal characteristics from big multidimensional data sets to provide reliable predictions of chronic illnesses without sacrificing data uniqueness.[1] To ensure the success of our proposed model, we employed balanced classes by employing hybrid balanced class sampling methods on the original dataset, as well as methods for data pre-processing and data transformation, to provide credible data for the training model. We ran and assessed our model on datasets with binary and multivalued classifications. We have used multiple datasets (Parkinson, arrythmia, breast cancer, kidney, diabetes). Suitable features are selected by using the Hybrid feature model consists of Lassocv, decision tree, random forest, gradient boosting,Adaboost, stochastic gradient descent and done voting of attributes which are common output from these methods.Accuracy of original dataset before applying framework is recorded and evaluated against reduced data set of attributes accuracy. The results are shown separately to provide comparisons. Based on the result analysis, we can conclude that our proposed model produced the highest accuracy on multi valued class datasets than on binary class attributes.[1]

Development of a Stock Trading System Using M & W Wave Patterns and Genetic Algorithms (M&W 파동 패턴과 유전자 알고리즘을 이용한 주식 매매 시스템 개발)

  • Yang, Hoonseok;Kim, Sunwoong;Choi, Heung Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.1
    • /
    • pp.63-83
    • /
    • 2019
  • Investors prefer to look for trading points based on the graph shown in the chart rather than complex analysis, such as corporate intrinsic value analysis and technical auxiliary index analysis. However, the pattern analysis technique is difficult and computerized less than the needs of users. In recent years, there have been many cases of studying stock price patterns using various machine learning techniques including neural networks in the field of artificial intelligence(AI). In particular, the development of IT technology has made it easier to analyze a huge number of chart data to find patterns that can predict stock prices. Although short-term forecasting power of prices has increased in terms of performance so far, long-term forecasting power is limited and is used in short-term trading rather than long-term investment. Other studies have focused on mechanically and accurately identifying patterns that were not recognized by past technology, but it can be vulnerable in practical areas because it is a separate matter whether the patterns found are suitable for trading. When they find a meaningful pattern, they find a point that matches the pattern. They then measure their performance after n days, assuming that they have bought at that point in time. Since this approach is to calculate virtual revenues, there can be many disparities with reality. The existing research method tries to find a pattern with stock price prediction power, but this study proposes to define the patterns first and to trade when the pattern with high success probability appears. The M & W wave pattern published by Merrill(1980) is simple because we can distinguish it by five turning points. Despite the report that some patterns have price predictability, there were no performance reports used in the actual market. The simplicity of a pattern consisting of five turning points has the advantage of reducing the cost of increasing pattern recognition accuracy. In this study, 16 patterns of up conversion and 16 patterns of down conversion are reclassified into ten groups so that they can be easily implemented by the system. Only one pattern with high success rate per group is selected for trading. Patterns that had a high probability of success in the past are likely to succeed in the future. So we trade when such a pattern occurs. It is a real situation because it is measured assuming that both the buy and sell have been executed. We tested three ways to calculate the turning point. The first method, the minimum change rate zig-zag method, removes price movements below a certain percentage and calculates the vertex. In the second method, high-low line zig-zag, the high price that meets the n-day high price line is calculated at the peak price, and the low price that meets the n-day low price line is calculated at the valley price. In the third method, the swing wave method, the high price in the center higher than n high prices on the left and right is calculated as the peak price. If the central low price is lower than the n low price on the left and right, it is calculated as valley price. The swing wave method was superior to the other methods in the test results. It is interpreted that the transaction after checking the completion of the pattern is more effective than the transaction in the unfinished state of the pattern. Genetic algorithms(GA) were the most suitable solution, although it was virtually impossible to find patterns with high success rates because the number of cases was too large in this simulation. We also performed the simulation using the Walk-forward Analysis(WFA) method, which tests the test section and the application section separately. So we were able to respond appropriately to market changes. In this study, we optimize the stock portfolio because there is a risk of over-optimized if we implement the variable optimality for each individual stock. Therefore, we selected the number of constituent stocks as 20 to increase the effect of diversified investment while avoiding optimization. We tested the KOSPI market by dividing it into six categories. In the results, the portfolio of small cap stock was the most successful and the high vol stock portfolio was the second best. This shows that patterns need to have some price volatility in order for patterns to be shaped, but volatility is not the best.

A Study on the Intelligent Quick Response System for Fast Fashion(IQRS-FF) (패스트 패션을 위한 지능형 신속대응시스템(IQRS-FF)에 관한 연구)

  • Park, Hyun-Sung;Park, Kwang-Ho
    • Journal of Intelligence and Information Systems
    • /
    • v.16 no.3
    • /
    • pp.163-179
    • /
    • 2010
  • Recentlythe concept of fast fashion is drawing attention as customer needs are diversified and supply lead time is getting shorter in fashion industry. It is emphasized as one of the critical success factors in the fashion industry how quickly and efficiently to satisfy the customer needs as the competition has intensified. Because the fast fashion is inherently susceptible to trend, it is very important for fashion retailers to make quick decisions regarding items to launch, quantity based on demand prediction, and the time to respond. Also the planning decisions must be executed through the business processes of procurement, production, and logistics in real time. In order to adapt to this trend, the fashion industry urgently needs supports from intelligent quick response(QR) system. However, the traditional functions of QR systems have not been able to completely satisfy such demands of the fast fashion industry. This paper proposes an intelligent quick response system for the fast fashion(IQRS-FF). Presented are models for QR process, QR principles and execution, and QR quantity and timing computation. IQRS-FF models support the decision makers by providing useful information with automated and rule-based algorithms. If the predefined conditions of a rule are satisfied, the actions defined in the rule are automatically taken or informed to the decision makers. In IQRS-FF, QRdecisions are made in two stages: pre-season and in-season. In pre-season, firstly master demand prediction is performed based on the macro level analysis such as local and global economy, fashion trends and competitors. The prediction proceeds to the master production and procurement planning. Checking availability and delivery of materials for production, decision makers must make reservations or request procurements. For the outsourcing materials, they must check the availability and capacity of partners. By the master plans, the performance of the QR during the in-season is greatly enhanced and the decision to select the QR items is made fully considering the availability of materials in warehouse as well as partners' capacity. During in-season, the decision makers must find the right time to QR as the actual sales occur in stores. Then they are to decide items to QRbased not only on the qualitative criteria such as opinions from sales persons but also on the quantitative criteria such as sales volume, the recent sales trend, inventory level, the remaining period, the forecast for the remaining period, and competitors' performance. To calculate QR quantity in IQRS-FF, two calculation methods are designed: QR Index based calculation and attribute similarity based calculation using demographic cluster. In the early period of a new season, the attribute similarity based QR amount calculation is better used because there are not enough historical sales data. By analyzing sales trends of the categories or items that have similar attributes, QR quantity can be computed. On the other hand, in case of having enough information to analyze the sales trends or forecasting, the QR Index based calculation method can be used. Having defined the models for decision making for QR, we design KPIs(Key Performance Indicators) to test the reliability of the models in critical decision makings: the difference of sales volumebetween QR items and non-QR items; the accuracy rate of QR the lead-time spent on QR decision-making. To verify the effectiveness and practicality of the proposed models, a case study has been performed for a representative fashion company which recently developed and launched the IQRS-FF. The case study shows that the average sales rateof QR items increased by 15%, the differences in sales rate between QR items and non-QR items increased by 10%, the QR accuracy was 70%, the lead time for QR dramatically decreased from 120 hours to 8 hours.

Development of New Variables Affecting Movie Success and Prediction of Weekly Box Office Using Them Based on Machine Learning (영화 흥행에 영향을 미치는 새로운 변수 개발과 이를 이용한 머신러닝 기반의 주간 박스오피스 예측)

  • Song, Junga;Choi, Keunho;Kim, Gunwoo
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.4
    • /
    • pp.67-83
    • /
    • 2018
  • The Korean film industry with significant increase every year exceeded the number of cumulative audiences of 200 million people in 2013 finally. However, starting from 2015 the Korean film industry entered a period of low growth and experienced a negative growth after all in 2016. To overcome such difficulty, stakeholders like production company, distribution company, multiplex have attempted to maximize the market returns using strategies of predicting change of market and of responding to such market change immediately. Since a film is classified as one of experiential products, it is not easy to predict a box office record and the initial number of audiences before the film is released. And also, the number of audiences fluctuates with a variety of factors after the film is released. So, the production company and distribution company try to be guaranteed the number of screens at the opining time of a newly released by multiplex chains. However, the multiplex chains tend to open the screening schedule during only a week and then determine the number of screening of the forthcoming week based on the box office record and the evaluation of audiences. Many previous researches have conducted to deal with the prediction of box office records of films. In the early stage, the researches attempted to identify factors affecting the box office record. And nowadays, many studies have tried to apply various analytic techniques to the factors identified previously in order to improve the accuracy of prediction and to explain the effect of each factor instead of identifying new factors affecting the box office record. However, most of previous researches have limitations in that they used the total number of audiences from the opening to the end as a target variable, and this makes it difficult to predict and respond to the demand of market which changes dynamically. Therefore, the purpose of this study is to predict the weekly number of audiences of a newly released film so that the stakeholder can flexibly and elastically respond to the change of the number of audiences in the film. To that end, we considered the factors used in the previous studies affecting box office and developed new factors not used in previous studies such as the order of opening of movies, dynamics of sales. Along with the comprehensive factors, we used the machine learning method such as Random Forest, Multi Layer Perception, Support Vector Machine, and Naive Bays, to predict the number of cumulative visitors from the first week after a film release to the third week. At the point of the first and the second week, we predicted the cumulative number of visitors of the forthcoming week for a released film. And at the point of the third week, we predict the total number of visitors of the film. In addition, we predicted the total number of cumulative visitors also at the point of the both first week and second week using the same factors. As a result, we found the accuracy of predicting the number of visitors at the forthcoming week was higher than that of predicting the total number of them in all of three weeks, and also the accuracy of the Random Forest was the highest among the machine learning methods we used. This study has implications in that this study 1) considered various factors comprehensively which affect the box office record and merely addressed by other previous researches such as the weekly rating of audiences after release, the weekly rank of the film after release, and the weekly sales share after release, and 2) tried to predict and respond to the demand of market which changes dynamically by suggesting models which predicts the weekly number of audiences of newly released films so that the stakeholders can flexibly and elastically respond to the change of the number of audiences in the film.

Empirical Forecast of Corotating Interacting Regions and Geomagnetic Storms Based on Coronal Hole Information (코로나 홀을 이용한 CIR과 지자기 폭풍의 경험적 예보 연구)

  • Lee, Ji-Hye;Moon, Yong-Jae;Choi, Yun-Hee;Yoo, Kye-Hwa
    • Journal of Astronomy and Space Sciences
    • /
    • v.26 no.3
    • /
    • pp.305-316
    • /
    • 2009
  • In this study, we suggest an empirical forecast of CIR (Corotating Interaction Regions) and geomagnetic storm based on the information of coronal holes (CH). For this we used CH data obtained from He I $10830{\AA}$ maps at National Solar Observatory-Kitt Peak from January 1996 to November 2003 and the CIR and storm data that Choi et al. (2009) identified. Considering the relationship among coronal holes, CIRs, and geomagnetic storms (Choi et al. 2009), we propose the criteria for geoeffective coronal holes; the center of CH is located between $N40^{\circ}$ and $S40^{\circ}$ and between $E40^{\circ}$ and $W20^{\circ}$, and its area in percentage of solar hemispheric area is larger than the following areas: (1) case 1: 0.36%, (2) case 2: 0.66%, (3) case 3: 0.36% for 1996-2000, and 0.66% for 2001-2003. Then we present contingency tables between prediction and observation for three cases and their dependence on solar cycle phase. From the contingency tables, we determined several statistical parameters for forecast evaluation such as PODy (the probability of detection yes), FAR (the false alarm ratio), Bias (the ratio of "yes" predictions to "yes" observations) and CSI (critical success index). Considering the importance of PODy and CSI, we found that the best criterion is case 3; CH-CIR: PODy=0.77, FAR=0.66, Bias=2.28, CSI=0.30. CH-storm: PODy=0.81, FAR=0.84, Bias=5.00, CSI=0.16. It is also found that the parameters after the solar maximum are much better than those before the solar maximum. Our results show that the forecasting of CIR based on coronal hole information is meaningful but the forecast of goemagnetic storm is challenging.

The Study of the Effects of the Enterprise Mobile Social Network Service on User Satisfaction and the Continuous Use Intention (기업 모바일 소셜네트워크서비스 특성요인이 사용자 만족과 지속적 사용의도에 미치는 영향에 관한 연구)

  • Kim, Joon-Hee;Ha, Kyu-Soo
    • Journal of Digital Convergence
    • /
    • v.10 no.8
    • /
    • pp.135-148
    • /
    • 2012
  • This work is intended to investigate how the factors of enterprise mobile SNS affect user satisfaction and continuous use intention through technology acceptance model proposed by Davis. To achieve the purpose, this researcher explored Information Systems Success model proposed by DeLone & McLean, Technology Acceptance Model proposed by Davis, and Model after Acceptance, and on the basis of the investigation, performed a study. For the data of this work, 9 enterprises, each of which has more than 100 employees and is located in Seoul, were chosen, and a questionnaire survey was conducted on their 276 employees who experienced enterprise mobile SNS. As a data collection tool, a structured self-administered questionnaire was used. For data analysis, SPSS 18.0 and AMOS 18.0 were used for applying Structural Equation modelling. According to the results of this work, three factors of enterprise mobile SNS-systematic factor (system quality, information quality, and service quality), user factor (personal innovation and personal familiarity), social factor (social effects and social interaction)-affected user satisfaction and continuous use intention through perceived availability, perceived easiness, and perceived enjoyment. Also, it was found that the direction of effects matched a theoretical prediction. And, it was revealed that the decision variables and mediating variables significantly affected user satisfaction and continuous use intention. Theoretical and practical meanings were discussed for the study result, and some suggestions were made for the issues of this work and future studies.

The Box-office Success Factors of Films Utilizing Big Data-Focus on Laugh and Tear of Film Factors (빅데이터를 활용한 영화 흥행 분석 -천만 영화의 웃음과 눈물 요소를 중심으로)

  • Hwang, Young-mee;Park, Jin-tae;Moon, Il-young;Kim, Kwang-sun;Kwon, Oh-young
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.20 no.6
    • /
    • pp.1087-1095
    • /
    • 2016
  • The study aims to analyze factors of box office utilizing big data. The film industry has been increasing in the scale, but the discussion on analysis and prediction of box-office hit has not secured reliability because of failing in including all relevant data. 13 films have sold 10 million tickets until the present in Korea. The study demonstrated laughs and tears as an main interior factors of box-office hit films which showed more than 10 milling tickets power. First, the study collected terms relevant to laugh and tear. Next, it schematizes how frequently laugh and tear factors could be found along the 5-film-stage (exposition - Rising action - crisis - climax - ending) and revealed box-office hit films by genre. The results of the analysis would contribute to the construction of comprehensive database for the box office predictions on future scenarios.