• Title/Summary/Keyword: MachineLearning

Search Result 5,654, Processing Time 0.03 seconds

Performance Comparison of Machine-learning Models for Analyzing Weather and Traffic Accident Correlations

  • Li Zi Xuan;Hyunho Yang
    • Journal of information and communication convergence engineering
    • /
    • v.21 no.3
    • /
    • pp.225-232
    • /
    • 2023
  • Owing to advancements in intelligent transportation systems (ITS) and artificial-intelligence technologies, various machine-learning models can be employed to simulate and predict the number of traffic accidents under different weather conditions. Furthermore, we can analyze the relationship between weather and traffic accidents, allowing us to assess whether the current weather conditions are suitable for travel, which can significantly reduce the risk of traffic accidents. In this study, we analyzed 30000 traffic flow data points collected by traffic cameras at nearby intersections in Washington, D.C., USA from October 2012 to May 2017, using Pearson's heat map. We then predicted, analyzed, and compared the performance of the correlation between continuous features by applying several machine-learning algorithms commonly used in ITS, including random forest, decision tree, gradient-boosting regression, and support vector regression. The experimental results indicated that the gradient-boosting regression machine-learning model had the best performance.

Compact Modeling for Nanosheet FET Based on TCAD-Machine Learning (TCAD-머신러닝 기반 나노시트 FETs 컴팩트 모델링)

  • Junhyeok Song;Wonbok Lee;Jonghwan Lee
    • Journal of the Semiconductor & Display Technology
    • /
    • v.22 no.4
    • /
    • pp.136-141
    • /
    • 2023
  • The continuous shrinking of transistors in integrated circuits leads to difficulties in improving performance, resulting in the emerging transistors such as nanosheet field-effect transistors. In this paper, we propose a TCAD-machine learning framework of nanosheet FETs to model the current-voltage characteristics. Sentaurus TCAD simulations of nanosheet FETs are performed to obtain a large amount of device data. A machine learning model of I-V characteristics is trained using the multi-layer perceptron from these TCAD data. The weights and biases obtained from multi-layer perceptron are implemented in a PSPICE netlist to verify the accuracy of I-V and the DC transfer characteristics of a CMOS inverter. It is found that the proposed machine learning model is applicable to the prediction of nanosheet field-effect transistors device and circuit performance.

  • PDF

Selecting Machine Learning Model Based on Natural Language Processing for Shanghanlun Diagnostic System Classification (자연어 처리 기반 『상한론(傷寒論)』 변병진단체계(辨病診斷體系) 분류를 위한 기계학습 모델 선정)

  • Young-Nam Kim
    • 대한상한금궤의학회지
    • /
    • v.14 no.1
    • /
    • pp.41-50
    • /
    • 2022
  • Objective : The purpose of this study is to explore the most suitable machine learning model algorithm for Shanghanlun diagnostic system classification using natural language processing (NLP). Methods : A total of 201 data items were collected from 『Shanghanlun』 and 『Clinical Shanghanlun』, 'Taeyangbyeong-gyeolhyung' and 'Eumyangyeokchahunobokbyeong' were excluded to prevent oversampling or undersampling. Data were pretreated using a twitter Korean tokenizer and trained by logistic regression, ridge regression, lasso regression, naive bayes classifier, decision tree, and random forest algorithms. The accuracy of the models were compared. Results : As a result of machine learning, ridge regression and naive Bayes classifier showed an accuracy of 0.843, logistic regression and random forest showed an accuracy of 0.804, and decision tree showed an accuracy of 0.745, while lasso regression showed an accuracy of 0.608. Conclusions : Ridge regression and naive Bayes classifier are suitable NLP machine learning models for the Shanghanlun diagnostic system classification.

  • PDF

Machine Learning Techniques for Diabetic Retinopathy Detection: A Review

  • Rachna Kumari;Sanjeev Kumar;Sunila Godara
    • International Journal of Computer Science & Network Security
    • /
    • v.24 no.4
    • /
    • pp.67-76
    • /
    • 2024
  • Diabetic retinopathy is a threatening complication of diabetes, caused by damaged blood vessels of light sensitive areas of retina. DR leads to total or partial blindness if left untreated. DR does not give any symptoms at early stages so earlier detection of DR is a big challenge for proper treatment of diseases. With advancement of technology various computer-aided diagnostic programs using image processing and machine learning approaches are designed for early detection of DR so that proper treatment can be provided to the patients for preventing its harmful effects. Now a day machine learning techniques are widely applied for image processing. These techniques also provide amazing result in this field also. In this paper we discuss various machine learning and deep learning based techniques developed for automatic detection of Diabetic Retinopathy.

A Study of Machine Learning-Based Scheduling Strategy for Fuzzing (기계학습 기반 스케줄링 전략을 적용한 최신 퍼징 연구)

  • Jeewoo Jung;Taeho Kim;Taekyoung Kwon
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.34 no.5
    • /
    • pp.973-980
    • /
    • 2024
  • Fuzzing is an automated testing technique that generates a lot of testcases and monitors for exceptions to test a program. Recently, fuzzing research using machine learning has been actively proposed to solve various problems in the fuzzing process, but a comprehensive evaluation of fuzzing research using machine learning is lacking. In this paper, we analyze recent research that applies machine learning to scheduling techniques for fuzzing, categorizing them into reinforcement learning-based and supervised learning-based fuzzers. We evaluated the coverage performance of the analyzed machine learning-based fuzzers against real-world programs with four different file formats and bug detection performance against the LAVA-M dataset. The results showed that AFL-HIER, which applied seed clustering and seed scheduling with reinforcement learning outperformed in coverage and bug detection. In the case of supervised learning, it showed high coverage on tcpdumps with high code complexity, and its superior bug detection performance when applied to hybrid fuzzing. This research shows that performance of machine learning-based fuzzer is better when both machine learning and additional fuzzing techniques are used to optimize the fuzzing process. Future research is needed on practical and robust machine learning-based fuzzing techniques that can be effectively applied to programs that handle various input formats.

Deep Structured Learning: Architectures and Applications

  • Lee, Soowook
    • International Journal of Advanced Culture Technology
    • /
    • v.6 no.4
    • /
    • pp.262-265
    • /
    • 2018
  • Deep learning, a sub-field of machine learning changing the prospects of artificial intelligence (AI) because of its recent advancements and application in various field. Deep learning deals with algorithms inspired by the structure and function of the brain called artificial neural networks. This works reviews basic architecture and recent advancement of deep structured learning. It also describes contemporary applications of deep structured learning and its advantages over the treditional learning in artificial interlligence. This study is useful for the general readers and students who are in the early stage of deep learning studies.

Price Determinant Factors of Artworks and Prediction Model Based on Machine Learning (작품 가격 추정을 위한 기계 학습 기법의 응용 및 가격 결정 요인 분석)

  • Jang, Dongryul;Park, Minjae
    • Journal of Korean Society for Quality Management
    • /
    • v.47 no.4
    • /
    • pp.687-700
    • /
    • 2019
  • Purpose: The purpose of this study is to investigate the interaction effects between price determinants of artworks. We expand the methodology in art market by applying machine learning techniques to estimate the price of artworks and compare linear regression and machine learning in terms of prediction accuracy. Methods: Moderated regression analysis was performed to verify the interaction effects of artistic characteristics on price. The moderating effects were studied by confirming the significance level of the interaction terms of the derived regression equation. In order to derive price estimation model, we use multiple linear regression analysis, which is a parametric statistical technique, and k-nearest neighbor (kNN) regression, which is a nonparametric statistical technique in machine learning methods. Results: Mostly, the influences of the price determinants of art are different according to the auction types and the artist 's reputation. However, the auction type did not control the influence of the genre of the work on the price. As a result of the analysis, the kNN regression was superior to the linear regression analysis based on the prediction accuracy. Conclusion: It provides a theoretical basis for the complexity that exists between pricing determinant factors of artworks. In addition, the nonparametric models and machine learning techniques as well as existing parameter models are implemented to estimate the artworks' price.

Machine Learning Based Malware Detection Using API Call Time Interval (API Call Time Interval을 활용한 머신러닝 기반의 악성코드 탐지)

  • Cho, Young Min;Kwon, Hun Yeong
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.30 no.1
    • /
    • pp.51-58
    • /
    • 2020
  • The use of malware in cyber threats continues to be used in all ages, and will continue to be a major attack method even if IT technology advances. Therefore, researches for detecting such malicious codes are constantly tried in various ways. Recently, with the development of AI-related technology, many researches related to machine learning have been conducted to detect malware. In this paper, we propose a method to detect malware using machine learning. For machine learning detection, we create a feature around each call interval, ie Time Interval, in which API calls occur among dynamic analysis data, and then apply the result to machine learning techniques.

A Study on Ontology Generation by Machine Learning in Big Data (빅 데이터에서 기계학습을 통한 온톨로지 생성에 관한 연구)

  • Hwang, Chi-Gon;Yoon, Chang-Pyo
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2018.10a
    • /
    • pp.645-646
    • /
    • 2018
  • Recently, the concept of machine learning has been introduced as a decision making method through data processing. Machine learning uses the results of running based on existing data as a means of decision making. The data generated by the development of technology is vast. This data is called big data. It is important to extract the necessary data from these data. In this paper, we propose a method for extracting related data for constructing an ontology through machine learning. The results of machine learning can be given a relationship from a semantic perspective. it can be added to the ontology to support relationships depending on the needs of the application.

  • PDF

Short-term Wind Power Prediction Based on Empirical Mode Decomposition and Improved Extreme Learning Machine

  • Tian, Zhongda;Ren, Yi;Wang, Gang
    • Journal of Electrical Engineering and Technology
    • /
    • v.13 no.5
    • /
    • pp.1841-1851
    • /
    • 2018
  • For the safe and stable operation of the power system, accurate wind power prediction is of great significance. A wind power prediction method based on empirical mode decomposition and improved extreme learning machine is proposed in this paper. Firstly, wind power time series is decomposed into several components with different frequency by empirical mode decomposition, which can reduce the non-stationary of time series. The components after decomposing remove the long correlation and promote the different local characteristics of original wind power time series. Secondly, an improved extreme learning machine prediction model is introduced to overcome the sample data updating disadvantages of standard extreme learning machine. Different improved extreme learning machine prediction model of each component is established. Finally, the prediction value of each component is superimposed to obtain the final result. Compared with other prediction models, the simulation results demonstrate that the proposed prediction method has better prediction accuracy for wind power.