• Title/Summary/Keyword: multiple-decision method

Search Result 459, Processing Time 0.033 seconds

Machine learning application to seismic site classification prediction model using Horizontal-to-Vertical Spectral Ratio (HVSR) of strong-ground motions

  • Francis G. Phi;Bumsu Cho;Jungeun Kim;Hyungik Cho;Yun Wook Choo;Dookie Kim;Inhi Kim
    • Geomechanics and Engineering
    • /
    • v.37 no.6
    • /
    • pp.539-554
    • /
    • 2024
  • This study explores development of prediction model for seismic site classification through the integration of machine learning techniques with horizontal-to-vertical spectral ratio (HVSR) methodologies. To improve model accuracy, the research employs outlier detection methods and, synthetic minority over-sampling technique (SMOTE) for data balance, and evaluates using seven machine learning models using seismic data from KiK-net. Notably, light gradient boosting method (LGBM), gradient boosting, and decision tree models exhibit improved performance when coupled with SMOTE, while Multiple linear regression (MLR) and Support vector machine (SVM) models show reduced efficacy. Outlier detection techniques significantly enhance accuracy, particularly for LGBM, gradient boosting, and voting boosting. The ensemble of LGBM with the isolation forest and SMOTE achieves the highest accuracy of 0.91, with LGBM and local outlier factor yielding the highest F1-score of 0.79. Consistently outperforming other models, LGBM proves most efficient for seismic site classification when supported by appropriate preprocessing procedures. These findings show the significance of outlier detection and data balancing for precise seismic soil classification prediction, offering insights and highlighting the potential of machine learning in optimizing site classification accuracy.

Ensemble of Nested Dichotomies for Activity Recognition Using Accelerometer Data on Smartphone (Ensemble of Nested Dichotomies 기법을 이용한 스마트폰 가속도 센서 데이터 기반의 동작 인지)

  • Ha, Eu Tteum;Kim, Jeongmin;Ryu, Kwang Ryel
    • Journal of Intelligence and Information Systems
    • /
    • v.19 no.4
    • /
    • pp.123-132
    • /
    • 2013
  • As the smartphones are equipped with various sensors such as the accelerometer, GPS, gravity sensor, gyros, ambient light sensor, proximity sensor, and so on, there have been many research works on making use of these sensors to create valuable applications. Human activity recognition is one such application that is motivated by various welfare applications such as the support for the elderly, measurement of calorie consumption, analysis of lifestyles, analysis of exercise patterns, and so on. One of the challenges faced when using the smartphone sensors for activity recognition is that the number of sensors used should be minimized to save the battery power. When the number of sensors used are restricted, it is difficult to realize a highly accurate activity recognizer or a classifier because it is hard to distinguish between subtly different activities relying on only limited information. The difficulty gets especially severe when the number of different activity classes to be distinguished is very large. In this paper, we show that a fairly accurate classifier can be built that can distinguish ten different activities by using only a single sensor data, i.e., the smartphone accelerometer data. The approach that we take to dealing with this ten-class problem is to use the ensemble of nested dichotomy (END) method that transforms a multi-class problem into multiple two-class problems. END builds a committee of binary classifiers in a nested fashion using a binary tree. At the root of the binary tree, the set of all the classes are split into two subsets of classes by using a binary classifier. At a child node of the tree, a subset of classes is again split into two smaller subsets by using another binary classifier. Continuing in this way, we can obtain a binary tree where each leaf node contains a single class. This binary tree can be viewed as a nested dichotomy that can make multi-class predictions. Depending on how a set of classes are split into two subsets at each node, the final tree that we obtain can be different. Since there can be some classes that are correlated, a particular tree may perform better than the others. However, we can hardly identify the best tree without deep domain knowledge. The END method copes with this problem by building multiple dichotomy trees randomly during learning, and then combining the predictions made by each tree during classification. The END method is generally known to perform well even when the base learner is unable to model complex decision boundaries As the base classifier at each node of the dichotomy, we have used another ensemble classifier called the random forest. A random forest is built by repeatedly generating a decision tree each time with a different random subset of features using a bootstrap sample. By combining bagging with random feature subset selection, a random forest enjoys the advantage of having more diverse ensemble members than a simple bagging. As an overall result, our ensemble of nested dichotomy can actually be seen as a committee of committees of decision trees that can deal with a multi-class problem with high accuracy. The ten classes of activities that we distinguish in this paper are 'Sitting', 'Standing', 'Walking', 'Running', 'Walking Uphill', 'Walking Downhill', 'Running Uphill', 'Running Downhill', 'Falling', and 'Hobbling'. The features used for classifying these activities include not only the magnitude of acceleration vector at each time point but also the maximum, the minimum, and the standard deviation of vector magnitude within a time window of the last 2 seconds, etc. For experiments to compare the performance of END with those of other methods, the accelerometer data has been collected at every 0.1 second for 2 minutes for each activity from 5 volunteers. Among these 5,900 ($=5{\times}(60{\times}2-2)/0.1$) data collected for each activity (the data for the first 2 seconds are trashed because they do not have time window data), 4,700 have been used for training and the rest for testing. Although 'Walking Uphill' is often confused with some other similar activities, END has been found to classify all of the ten activities with a fairly high accuracy of 98.4%. On the other hand, the accuracies achieved by a decision tree, a k-nearest neighbor, and a one-versus-rest support vector machine have been observed as 97.6%, 96.5%, and 97.6%, respectively.

Design of method to analyze UI structure of contents based on the Morphology (형태적 관점의 콘텐츠 UI구조 분석 방법 설계)

  • Yun, Bong Shik
    • Smart Media Journal
    • /
    • v.8 no.4
    • /
    • pp.58-63
    • /
    • 2019
  • The growth of the mobile device market has changed the education market and led to the quantitative growth of various media education. In particular, smart devices, which have better interaction than existing PCs or consoles, can develop more user-friendly content, allowing various types of educational content and inducing changes in traditional education methods for consumers. Although many researchers recently suggest viable development methods or marketing elements of contents, development companies, and developers, until now, merely rely on the human senses. Therefore, it is necessary to study the actual user's smart-device based usability and experience environment. This study aims to propose an intuitive statistical processing method for analyzing the usability of game-type educational contents in terms of form, for popular games that have been released as a basis for analyzing the user experience environment. In particular, because the game industry has a sufficient number of similar examples, it is possible to conduct research based on big data and to use them for immediate decision-making between multiple co-developers through the analysis method proposed by the research. It is expected to become an analytical model that can communicate with other industries because it is effective in securing data sources.

Scalp reconstruction: A 10-year experience

  • Jang, Hyeon Uk;Choi, Young Woong
    • Archives of Craniofacial Surgery
    • /
    • v.21 no.4
    • /
    • pp.237-243
    • /
    • 2020
  • Background: The scalp is an important functional and aesthetic structure that protects the cranial bone. Due to its inelastic characteristics, soft-tissue defects of the scalp make reconstruction surgery difficult. This study aims to provide an improved scalp reconstruction decision making algorithm for surgeons. Methods: This study examined patients who underwent scalp reconstruction within the last 10 years. The study evaluated several factors that surgeons use to select a given reconstruction method such as etiology, defect location, size, depth, and complications. An algorithmic approach was then suggested based on an analysis of these factors. Results: Ninety-four patients were selected in total and 98 cases, including revision surgery, were performed for scalp reconstruction. Scalp reconstruction was performed by primary closure (36.73%), skin graft (27.55%), local flap (17.34%), pedicled regional flap (15.30%), and free flap (3.06%). The ratio of primary closure to more complex procedure on loose scalps (51.11%) was significantly higher than on tight scalps (24.52%) (p=0.011). The choice of scalp reconstruction method was affected significantly by the defect size (R=0.479, p<0.001) and depth (p<0.001). There were five major complications which were three cases of flap necrosis and two cases of skin necrosis. Hematoma was the most common of the 29 minor complications reported, followed by skin necrosis. Conclusion: There are multiple factors affecting the choice of scalp reconstruction method. We suggest an algorithm based on 10 years of experience that will help surgeons establish successful surgical management for their patients.

Finding Alternative Routes in Inter Modal Transit Networks Considering Constraints (제약조건을 고려한 통합대중교통망의 대안경로탐색)

  • Lee Mee-Young;Baek Nam-Chul
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.4 no.1 s.6
    • /
    • pp.1-13
    • /
    • 2005
  • For integrated transit networks operated with diverse transit modes, it is difficult to find related literatures on the provision of multiple travel time routes information. This study proposes a methodology on how to select K number of least time transit routes in the integrated transit network with two modes, bus and subway. In the Proposed method, three constraints usually (will be) encountered for decision by users-(1) service time constraints of travel modes, (2) maximum payment fare constraints under the integrated distance based fare system, and (3) maximum number of transfer constraints - are taken into consideration in the searched routes. The experimental studies shows that the proposed method properly provides K number of least time routes. In addition, it somehow proves that by combinatorially considering three constraints the proposed method can be evaluated as to enlarge the quality as well as the diversity of route information demanded by users.

  • PDF

Candidate Marker Identification from Gene Expression Data with Attribute Value Discretization and Negation (속성값 이산화 및 부정값 허용을 하는 의사결정트리 기반의 유전자 발현 데이터의 마커 후보 식별)

  • Lee, Kyung-Mi;Lee, Keon-Myung
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.21 no.5
    • /
    • pp.575-580
    • /
    • 2011
  • With the increasing expectation on personalized medicine, it is getting importance to analyze medical information in molecular biology perspective. Gene expression data are one of representative ones to show the microscopic phenomena of biological activities. In gene expression data analysis, one of major concerns is to identify markers which can be used to predict disease occurrence, progression or recurrence in the molecular level. Existing markers candidate identification methods mainly depend on statistical hypothesis test methods. This paper proposes a search method based decision tree induction to identify candidate markers which consist of multiple genes. The propose method discretizes numeric expression level into three categorical values and allows candidate markers' genes to be expressed by their negation as well as categorical values. It is desirable to have some number of genes to be included in markers. Hence the method is devised to try to find candidate markers with restricted number of genes.

Sensing Model for Reducing Power Consumption for Indoor/Outdoor Context Transition (실내/실외 컨텍스트 전이를 고려한 저전력 센싱 모델)

  • Kim, Deok-Ki;Park, Jae-Hyeon;Lee, Jung-Won
    • Journal of KIISE
    • /
    • v.43 no.7
    • /
    • pp.763-772
    • /
    • 2016
  • With the spread of smartphones containing multiple on-board sensors, the market for context aware applications have grown. However, due to the limited power capacity of a smartphone, users feel discontented QoS. Additionally, context aware applications require the utilization of many forms of context and sensing information. If context transition has occurred, types of needed sensors must be changed and each sensor modules need to turn on/off. In addition, excessive sensing has been found when the context decision is ambiguous. In this paper, we focus on power consumption associated with the context transition that occurs during indoor/outdoor detection, modeling the activities of the sensor associated with these contexts. And we suggest a freezing algorithm that reduces power consumption in context transition. We experiment with a commercial application that service is indoor/outdoor location tracking, measure power consumption in context transition with and without the utilization of the proposed method. We find that proposed method reduces power consumption about 20% during context transition.

Searching an Efficient frontier in the DEA Model based on the Reference Point Method (참조점 방법을 이용한 DEA모형의 프론티어 탐구)

  • 오동일
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.1 no.1
    • /
    • pp.83-90
    • /
    • 2000
  • DEA is a newly developed analyzing tool to measure efficiency evaluation of decision making units (DMU). It compares DMU by radial Projection on the efficient frontier. The purpose of this study is to show reference point approach used for searching solution in multiple objective linear Programming can be usefully used to determine flexible efficient frontier of each DMU In reference point approach, the minimization of ASF Produces an efficient points in frontier and enhances the usefulness of DEA by Providing flexibility in DEA and optimally allocating resources to DMU. Various DEA models can be supported by reference point method by changing the projection direction in order to choose the targets units, standards costs and management benching-marking.

  • PDF

A study on forecasting attendance rate of reserve forces training based on Data Mining (데이터마이닝에 기반한 예비군훈련 입소율 예측에 관한 연구)

  • Cho, Sangjoon;Ma, Jungmok
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.22 no.3
    • /
    • pp.261-267
    • /
    • 2021
  • The mission of the reserve forces unit is to prepare good training for reserve forces during peacetime. For good training, units require proper organization support agents, but they have difficulties due to a lack of unit members. For that reason, the units forecast the monthly attendance rate of reserve forces (using the x-1 year's result) to organize support agents and unit schedule. On the other hand, the existing planning method can have more errors compared to the actual result of the attendance rate. This problem has a negative effect on the training performance. Therefore, it requires more accurate forecast models to reduce attendance rate errors. This paper proposes an attendance rate forecast model using data mining. To verify the proposed data mining based model, the existing planning method was compared with the proposed model using real data. The results showed that the proposed model outperforms the existing planning method.

Integrating physics-based fragility for hierarchical spectral clustering for resilience assessment of power distribution systems under extreme winds

  • Jintao Zhang;Wei Zhang;William Hughes;Amvrossios C. Bagtzoglou
    • Wind and Structures
    • /
    • v.39 no.1
    • /
    • pp.1-14
    • /
    • 2024
  • Widespread damages from extreme winds have attracted lots of attentions of the resilience assessment of power distribution systems. With many related environmental parameters as well as numerous power infrastructure components, such as poles and wires, the increased challenge of power asset management before, during and after extreme events have to be addressed to prevent possible cascading failures in the power distribution system. Many extreme winds from weather events, such as hurricanes, generate widespread damages in multiple areas such as the economy, social security, and infrastructure management. The livelihoods of residents in the impaired areas are devastated largely due to the paucity of vital utilities, such as electricity. To address the challenge of power grid asset management, power system clustering is needed to partition a complex power system into several stable clusters to prevent the cascading failure from happening. Traditionally, system clustering uses the Binary Decision Diagram (BDD) to derive the clustering result, which is time-consuming and inefficient. Meanwhile, the previous studies considering the weather hazards did not include any detailed weather-related meteorologic parameters which is not appropriate as the heterogeneity of the parameters could largely affect the system performance. Therefore, a fragility-based network hierarchical spectral clustering method is proposed. In the present paper, the fragility curve and surfaces for a power distribution subsystem are obtained first. The fragility of the subsystem under typical failure mechanisms is calculated as a function of wind speed and pole characteristic dimension (diameter or span length). Secondly, the proposed fragility-based hierarchical spectral clustering method (F-HSC) integrates the physics-based fragility analysis into Hierarchical Spectral Clustering (HSC) technique from graph theory to achieve the clustering result for the power distribution system under extreme weather events. From the results of vulnerability analysis, it could be seen that the system performance after clustering is better than before clustering. With the F-HSC method, the impact of the extreme weather events could be considered with topology to cluster different power distribution systems to prevent the system from experiencing power blackouts.