• Title/Summary/Keyword: Multiple Machine Learning

Search Result 356, Processing Time 0.024 seconds

A Study on the Application of Measurement Data Using Machine Learning Regression Models

  • Yun-Seok Seo;Young-Gon Kim
    • International journal of advanced smart convergence
    • /
    • v.12 no.2
    • /
    • pp.47-55
    • /
    • 2023
  • The automotive industry is undergoing a paradigm shift due to the convergence of IT and rapid digital transformation. Various components, including embedded structures and systems with complex architectures that incorporate IC semiconductors, are being integrated and modularized. As a result, there has been a significant increase in vehicle defects, raising expectations for the quality of automotive parts. As more and more data is being accumulated, there is an active effort to go beyond traditional reliability analysis methods and apply machine learning models based on the accumulated big data. However, there are still not many cases where machine learning is used in product development to identify factors of defects in performance and durability of products and incorporate feedback into the design to improve product quality. In this paper, we applied a prediction algorithm to the defects of automotive door devices equipped with automatic responsive sensors, which are commonly installed in recent electric and hydrogen vehicles. To do so, we selected test items, built a measurement emulation system for data acquisition, and conducted comparative evaluations by applying different machine learning algorithms to the measured data. The results in terms of R2 score were as follows: Ordinary multiple regression 0.96, Ridge regression 0.95, Lasso regression 0.89, Elastic regression 0.91.

Data-Driven-Based Beam Selection for Hybrid Beamforming in Ultra-Dense Networks

  • Ju, Sang-Lim;Kim, Kyung-Seok
    • International journal of advanced smart convergence
    • /
    • v.9 no.2
    • /
    • pp.58-67
    • /
    • 2020
  • In this paper, we propose a data-driven-based beam selection scheme for massive multiple-input and multiple-output (MIMO) systems in ultra-dense networks (UDN), which is capable of addressing the problem of high computational cost of conventional coordinated beamforming approaches. We consider highly dense small-cell scenarios with more small cells than mobile stations, in the millimetre-wave band. The analog beam selection for hybrid beamforming is a key issue in realizing millimetre-wave UDN MIMO systems. To reduce the computation complexity for the analog beam selection, in this paper, two deep neural network models are used. The channel samples, channel gains, and radio frequency beamforming vectors between the access points and mobile stations are collected at the central/cloud unit that is connected to all the small-cell access points, and are used to train the networks. The proposed machine-learning-based scheme provides an approach for the effective implementation of massive MIMO system in UDN environment.

An Estimation Model of Fine Dust Concentration Using Meteorological Environment Data and Machine Learning (기상환경데이터와 머신러닝을 활용한 미세먼지농도 예측 모델)

  • Lim, Joon-Mook
    • Journal of Information Technology Services
    • /
    • v.18 no.1
    • /
    • pp.173-186
    • /
    • 2019
  • Recently, as the amount of fine dust has risen rapidly, our interest is increasing day by day. It is virtually impossible to remove fine dust. However, it is best to predict the concentration of fine dust and minimize exposure to it. In this study, we developed a mathematical model that can predict the concentration of fine dust using various information related to the weather and air quality, which is provided in real time in 'Air Korea (http://www.airkorea.or.kr/)' and 'Weather Data Open Portal (https://data.kma.go.kr/).' In the mathematical model, various domestic seasonal variables and atmospheric state variables are extracted by multiple regression analysis. The parameters that have significant influence on the fine dust concentration are extracted, and using ANN (Artificial Neural Network) and SVM (Support Vector Machine), which are machine learning techniques, we proposed a prediction model. The proposed model can verify its effectiveness by using past dust and weather big data.

Systematic Research on Privacy-Preserving Distributed Machine Learning (프라이버시를 보호하는 분산 기계 학습 연구 동향)

  • Min Seob Lee;Young Ah Shin;Ji Young Chun
    • The Transactions of the Korea Information Processing Society
    • /
    • v.13 no.2
    • /
    • pp.76-90
    • /
    • 2024
  • Although artificial intelligence (AI) can be utilized in various domains such as smart city, healthcare, it is limited due to concerns about the exposure of personal and sensitive information. In response, the concept of distributed machine learning has emerged, wherein learning occurs locally before training a global model, mitigating the concentration of data on a central server. However, overall learning phase in a collaborative way among multiple participants poses threats to data privacy. In this paper, we systematically analyzes recent trends in privacy protection within the realm of distributed machine learning, considering factors such as the presence of a central server, distribution environment of the training datasets, and performance variations among participants. In particular, we focus on key distributed machine learning techniques, including horizontal federated learning, vertical federated learning, and swarm learning. We examine privacy protection mechanisms within these techniques and explores potential directions for future research.

Machine Learning Model to Predict Osteoporotic Spine with Hounsfield Units on Lumbar Computed Tomography

  • Nam, Kyoung Hyup;Seo, Il;Kim, Dong Hwan;Lee, Jae Il;Choi, Byung Kwan;Han, In Ho
    • Journal of Korean Neurosurgical Society
    • /
    • v.62 no.4
    • /
    • pp.442-449
    • /
    • 2019
  • Objective : Bone mineral density (BMD) is an important consideration during fusion surgery. Although dual X-ray absorptiometry is considered as the gold standard for assessing BMD, quantitative computed tomography (QCT) provides more accurate data in spine osteoporosis. However, QCT has the disadvantage of additional radiation hazard and cost. The present study was to demonstrate the utility of artificial intelligence and machine learning algorithm for assessing osteoporosis using Hounsfield units (HU) of preoperative lumbar CT coupling with data of QCT. Methods : We reviewed 70 patients undergoing both QCT and conventional lumbar CT for spine surgery. The T-scores of 198 lumbar vertebra was assessed in QCT and the HU of vertebral body at the same level were measured in conventional CT by the picture archiving and communication system (PACS) system. A multiple regression algorithm was applied to predict the T-score using three independent variables (age, sex, and HU of vertebral body on conventional CT) coupling with T-score of QCT. Next, a logistic regression algorithm was applied to predict osteoporotic or non-osteoporotic vertebra. The Tensor flow and Python were used as the machine learning tools. The Tensor flow user interface developed in our institute was used for easy code generation. Results : The predictive model with multiple regression algorithm estimated similar T-scores with data of QCT. HU demonstrates the similar results as QCT without the discordance in only one non-osteoporotic vertebra that indicated osteoporosis. From the training set, the predictive model classified the lumbar vertebra into two groups (osteoporotic vs. non-osteoporotic spine) with 88.0% accuracy. In a test set of 40 vertebrae, classification accuracy was 92.5% when the learning rate was 0.0001 (precision, 0.939; recall, 0.969; F1 score, 0.954; area under the curve, 0.900). Conclusion : This study is a simple machine learning model applicable in the spine research field. The machine learning model can predict the T-score and osteoporotic vertebrae solely by measuring the HU of conventional CT, and this would help spine surgeons not to under-estimate the osteoporotic spine preoperatively. If applied to a bigger data set, we believe the predictive accuracy of our model will further increase. We propose that machine learning is an important modality of the medical research field.

Multiple Discriminative DNNs for I-Vector Based Open-Set Language Recognition (I-벡터 기반 오픈세트 언어 인식을 위한 다중 판별 DNN)

  • Kang, Woo Hyun;Cho, Won Ik;Kang, Tae Gyoon;Kim, Nam Soo
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.41 no.8
    • /
    • pp.958-964
    • /
    • 2016
  • In this paper, we propose an i-vector based language recognition system to identify the spoken language of the speaker, which uses multiple discriminative deep neural network (DNN) models analogous to the multi-class support vector machine (SVM) classification system. The proposed model was trained and tested using the i-vectors included in the NIST 2015 i-vector Machine Learning Challenge database, and shown to outperform the conventional language recognition methods such as cosine distance, SVM and softmax NN classifier in open-set experiments.

From dark matter to baryons in a simulated universe via machine learning

  • Jo, Yongseok
    • The Bulletin of The Korean Astronomical Society
    • /
    • v.45 no.1
    • /
    • pp.50.2-50.2
    • /
    • 2020
  • The dark matter (DM) only simulations have been exploited to study e.g. the large scale structures and properties of a halo. In a baryon side, the high-resolution hydrodynamic simulation such as IllustrisTNG has helped extend the physics of gas along with stars and DM. However, the expansive computational cost of hydrodynamic simulations limits the size of a simulated universe whereas DM-only simulations can generate the universe of the cosmological horizon size approximately. I will introduce a pipeline to estimate baryonic properties of a galaxy inside a dark matter (DM) halo in DM-only simulations using a machine trained on high-resolution hydrodynamic simulations. An extremely randomized tree (ERT) algorithm is used together with multiple novel improvements such as a refined error function in machine training and two-stage learning. By applying our machine to the DM-only simulation of a large volume, I then validate the pipeline that rapidly generates a galaxy catalog from a DM halo catalog using the correlations the machine found in hydrodynamic simulations. I will discuss the benefits that machine-based approaches like this entail, as well as suggestions to raise the scientific potential of such approaches.

  • PDF

Estimation of regional flow duration curve applicable to ungauged areas using machine learning technique (머신러닝 기법을 이용한 미계측 유역에 적용 가능한 지역화 유황곡선 산정)

  • Jeung, Se Jin;Lee, Seung Pil;Kim, Byung Sik
    • Journal of Korea Water Resources Association
    • /
    • v.54 no.spc1
    • /
    • pp.1183-1193
    • /
    • 2021
  • Low flow affects various fields such as river water supply management and planning, and irrigation water. A sufficient period of flow data is required to calculate the Flow Duration Curve. However, in order to calculate the Flow Duration Curve, it is essential to secure flow data for more than 30 years. However, in the case of rivers below the national river unit, there is no long-term flow data or there are observed data missing for a certain period in the middle, so there is a limit to calculating the Flow Duration Curve for each river. In the past, statistical-based methods such as Multiple Regression Analysis and ARIMA models were used to predict sulfur in the unmeasured watershed, but recently, the demand for machine learning and deep learning models is increasing. Therefore, in this study, we present the DNN technique, which is a machine learning technique that fits the latest paradigm. The DNN technique is a method that compensates for the shortcomings of the ANN technique, such as difficult to find optimal parameter values in the learning process and slow learning time. Therefore, in this study, the Flow Duration Curve applicable to the unmeasured watershed is calculated using the DNN model. First, the factors affecting the Flow Duration Curve were collected and statistically significant variables were selected through multicollinearity analysis between the factors, and input data were built into the machine learning model. The effectiveness of machine learning techniques was reviewed through statistical verification.

Comparative Application of Various Machine Learning Techniques for Lithology Predictions (다양한 기계학습 기법의 암상예측 적용성 비교 분석)

  • Jeong, Jina;Park, Eungyu
    • Journal of Soil and Groundwater Environment
    • /
    • v.21 no.3
    • /
    • pp.21-34
    • /
    • 2016
  • In the present study, we applied various machine learning techniques comparatively for prediction of subsurface structures based on multiple secondary information (i.e., well-logging data). The machine learning techniques employed in this study are Naive Bayes classification (NB), artificial neural network (ANN), support vector machine (SVM) and logistic regression classification (LR). As an alternative model, conventional hidden Markov model (HMM) and modified hidden Markov model (mHMM) are used where additional information of transition probability between primary properties is incorporated in the predictions. In the comparisons, 16 boreholes consisted with four different materials are synthesized, which show directional non-stationarity in upward and downward directions. Futhermore, two types of the secondary information that is statistically related to each material are generated. From the comparative analysis with various case studies, the accuracies of the techniques become degenerated with inclusion of additive errors and small amount of the training data. For HMM predictions, the conventional HMM shows the similar accuracies with the models that does not relies on transition probability. However, the mHMM consistently shows the highest prediction accuracy among the test cases, which can be attributed to the consideration of geological nature in the training of the model.

Application of machine learning for merging multiple satellite precipitation products

  • Van, Giang Nguyen;Jung, Sungho;Lee, Giha
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2021.06a
    • /
    • pp.134-134
    • /
    • 2021
  • Precipitation is a crucial component of water cycle and play a key role in hydrological processes. Traditionally, gauge-based precipitation is the main method to achieve high accuracy of rainfall estimation, but its distribution is sparsely in mountainous areas. Recently, satellite-based precipitation products (SPPs) provide grid-based precipitation with spatio-temporal variability, but SPPs contain a lot of uncertainty in estimated precipitation, and the spatial resolution quite coarse. To overcome these limitations, this study aims to generate new grid-based daily precipitation using Automatic weather system (AWS) in Korea and multiple SPPs(i.e. CHIRPSv2, CMORPH, GSMaP, TRMMv7) during the period of 2003-2017. And this study used a machine learning based Random Forest (RF) model for generating new merging precipitation. In addition, several statistical linear merging methods are used to compare with the results of the RF model. In order to investigate the efficiency of RF, observed data from 64 observed Automated Synoptic Observation System (ASOS) were collected to evaluate the accuracy of the products through Kling-Gupta efficiency (KGE), probability of detection (POD), false alarm rate (FAR), and critical success index (CSI). As a result, the new precipitation generated through the random forest model showed higher accuracy than each satellite rainfall product and spatio-temporal variability was better reflected than other statistical merging methods. Therefore, a random forest-based ensemble satellite precipitation product can be efficiently used for hydrological simulations in ungauged basins such as the Mekong River.

  • PDF