• Title/Summary/Keyword: machine learning applications

Search Result 538, Processing Time 0.028 seconds

Advanced Feature Selection Method on Android Malware Detection by Machine Learning (악성 안드로이드 앱 탐지를 위한 개선된 특성 선택 모델)

  • Boo, Joo-hun;Lee, Kyung-ho
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.30 no.3
    • /
    • pp.357-367
    • /
    • 2020
  • According to Symantec's 2018 internet security threat report, The number of new mobile malware variants increased by 54 percent in 2017, as compared to 2016. And last year, there were an average of 24,000 malicious mobile applications blocked each day. Existing signature-based technologies of malware detection have limitations. So, malware detection technique through machine learning is being researched to detect malware variant. However, even in the case of applying machine learning, if the proper features of the malware are not properly selected, the machine learning cannot be shown correctly. We are focusing on feature selection method to find the features of malware variant in this research.

Feature Selection and Hyper-Parameter Tuning for Optimizing Decision Tree Algorithm on Heart Disease Classification

  • Tsehay Admassu Assegie;Sushma S.J;Bhavya B.G;Padmashree S
    • International Journal of Computer Science & Network Security
    • /
    • v.24 no.2
    • /
    • pp.150-154
    • /
    • 2024
  • In recent years, there are extensive researches on the applications of machine learning to the automation and decision support for medical experts during disease detection. However, the performance of machine learning still needs improvement so that machine learning model produces result that is more accurate and reliable for disease detection. Selecting the hyper-parameter that could produce the possible maximum classification accuracy on medical dataset is the most challenging task in developing decision support systems with machine learning algorithms for medical dataset classification. Moreover, selecting the features that best characterizes a disease is another challenge in developing machine-learning model with better classification accuracy. In this study, we have proposed an optimized decision tree model for heart disease classification by using heart disease dataset collected from kaggle data repository. The proposed model is evaluated and experimental test reveals that the performance of decision tree improves when an optimal number of features are used for training. Overall, the accuracy of the proposed decision tree model is 98.2% for heart disease classification.

A Prediction Triage System for Emergency Department During Hajj Period using Machine Learning Models

  • Huda N. Alhazmi
    • International Journal of Computer Science & Network Security
    • /
    • v.24 no.7
    • /
    • pp.11-23
    • /
    • 2024
  • Triage is a practice of accurately prioritizing patients in emergency department (ED) based on their medical condition to provide them with proper treatment service. The variation in triage assessment among medical staff can cause mis-triage which affect the patients negatively. Developing ED triage system based on machine learning (ML) techniques can lead to accurate and efficient triage outcomes. This study aspires to develop a triage system using machine learning techniques to predict ED triage levels using patients' information. We conducted a retrospective study using Security Forces Hospital ED data, from 2021 through 2023 during Hajj period in Saudia Arabi. Using demographics, vital signs, and chief complaints as predictors, two machine learning models were investigated, naming gradient boosted decision tree (XGB) and deep neural network (DNN). The models were trained to predict ED triage levels and their predictive performance was evaluated using area under the receiver operating characteristic curve (AUC) and confusion matrix. A total of 11,584 ED visits were collected and used in this study. XGB and DNN models exhibit high abilities in the predicting performance with AUC-ROC scores 0.85 and 0.82, respectively. Compared to the traditional approach, our proposed system demonstrated better performance and can be implemented in real-world clinical settings. Utilizing ML applications can power the triage decision-making, clinical care, and resource utilization.

Development of Galaxy Image Classification Based on Hand-crafted Features and Machine Learning (Hand-crafted 특징 및 머신 러닝 기반의 은하 이미지 분류 기법 개발)

  • Oh, Yoonju;Jung, Heechul
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.16 no.1
    • /
    • pp.17-27
    • /
    • 2021
  • In this paper, we develop a galaxy image classification method based on hand-crafted features and machine learning techniques. Additionally, we provide an empirical analysis to reveal which combination of the techniques is effective for galaxy image classification. To achieve this, we developed a framework which consists of four modules such as preprocessing, feature extraction, feature post-processing, and classification. Finally, we found that the best technique for galaxy image classification is a method to use a median filter, ORB vector features and a voting classifier based on RBF SVM, random forest and logistic regression. The final method is efficient so we believe that it is applicable to embedded environments.

An investigation into the effects of lime-stabilization on soil-geosynthetic interface behavior

  • Khadije Mahmoodi;Nazanin Mahbubi Motlagh;Ahmad-Reza Mahboubi Ardakani
    • Geomechanics and Engineering
    • /
    • v.38 no.3
    • /
    • pp.231-247
    • /
    • 2024
  • The use of lime stabilization and geosynthetic reinforcement is a common approach to improve the performance of fine-grained soils in geotechnical applications. However, the impact of this combination on the soil-geosynthetic interaction remains unclear. This study addresses this gap by evaluating the interface efficiency and soil-geosynthetic interaction parameters of lime-stabilized clay (2%, 4%, 6%, and 8% lime content) reinforced with geotextile or geogrid using direct shear tests at various curing times (1, 7, 14, and 28 days). Additionally, machine learning algorithms (Support Vector Machine and Artificial Neural Network) were employed to predict soil shear strength. Findings revealed that lime stabilization significantly increased soil shear strength and interaction parameters, particularly at the optimal lime content (4%). Notably, stabilization improved the performance of soil-geogrid interfaces but had an adverse effect on soil-geotextile interfaces. Furthermore, machine learning algorithms effectively predicted soil shear strength, with sensitivity analysis highlighting lime percentage and geosynthetic type as the most significant influencing factors.

FedGCD: Federated Learning Algorithm with GNN based Community Detection for Heterogeneous Data

  • Wooseok Shin;Jitae Shin
    • Journal of Internet Computing and Services
    • /
    • v.24 no.6
    • /
    • pp.1-11
    • /
    • 2023
  • Federated learning (FL) is a ground breaking machine learning paradigm that allow smultiple participants to collaboratively train models in a cloud environment, all while maintaining the privacy of their raw data. This approach is in valuable in applications involving sensitive or geographically distributed data. However, one of the challenges in FL is dealing with heterogeneous and non-independent and identically distributed (non-IID) data across participants, which can result in suboptimal model performance compared to traditionalmachine learning methods. To tackle this, we introduce FedGCD, a novel FL algorithm that employs Graph Neural Network (GNN)-based community detection to enhance model convergence in federated settings. In our experiments, FedGCD consistently outperformed existing FL algorithms in various scenarios: for instance, in a non-IID environment, it achieved an accuracy of 0.9113, a precision of 0.8798,and an F1-Score of 0.8972. In a semi-IID setting, it demonstrated the highest accuracy at 0.9315 and an impressive F1-Score of 0.9312. We also introduce a new metric, nonIIDness, to quantitatively measure the degree of data heterogeneity. Our results indicate that FedGCD not only addresses the challenges of data heterogeneity and non-IIDness but also sets new benchmarks for FL algorithms. The community detection approach adopted in FedGCD has broader implications, suggesting that it could be adapted for other distributed machine learning scenarios, thereby improving model performance and convergence across a range of applications.

Performance Comparison Analysis of AI Supervised Learning Methods of Tensorflow and Scikit-Learn in the Writing Digit Data (필기숫자 데이터에 대한 텐서플로우와 사이킷런의 인공지능 지도학습 방식의 성능비교 분석)

  • Jo, Jun-Mo
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.14 no.4
    • /
    • pp.701-706
    • /
    • 2019
  • The advent of the AI(: Artificial Intelligence) has applied to many industrial and general applications have havingact on our lives these days. Various types of machine learning methods are supported in this field. The supervised learning method of the machine learning has features and targets as an input in the learning process. There are many supervised learning methods as well and their performance varies depends on the characteristics and states of the big data type as an input data. Therefore, in this paper, in order to compare the performance of the various supervised learning method with a specific big data set, the supervised learning methods supported in the Tensorflow and the Sckit-Learn are simulated and analyzed in the Jupyter Notebook environment with python.

Selection of Machine Learning Techniques for Network Lifetime Parameters and Synchronization Issues in Wireless Networks

  • Srilakshmi, Nimmagadda;Sangaiah, Arun Kumar
    • Journal of Information Processing Systems
    • /
    • v.15 no.4
    • /
    • pp.833-852
    • /
    • 2019
  • In real time applications, due to their effective cost and small size, wireless networks play an important role in receiving particular data and transmitting it to a base station for analysis, a process that can be easily deployed. Due to various internal and external factors, networks can change dynamically, which impacts the localisation of nodes, delays, routing mechanisms, geographical coverage, cross-layer design, the quality of links, fault detection, and quality of service, among others. Conventional methods were programmed, for static networks which made it difficult for networks to respond dynamically. Here, machine learning strategies can be applied for dynamic networks effecting self-learning and developing tools to react quickly and efficiently, with less human intervention and reprogramming. In this paper, we present a wireless networks survey based on different machine learning algorithms and network lifetime parameters, and include the advantages and drawbacks of such a system. Furthermore, we present learning algorithms and techniques for congestion, synchronisation, energy harvesting, and for scheduling mobile sinks. Finally, we present a statistical evaluation of the survey, the motive for choosing specific techniques to deal with wireless network problems, and a brief discussion on the challenges inherent in this area of research.

A Survey of Transfer and Multitask Learning in Bioinformatics

  • Xu, Qian;Yang, Qiang
    • Journal of Computing Science and Engineering
    • /
    • v.5 no.3
    • /
    • pp.257-268
    • /
    • 2011
  • Machine learning and data mining have found many applications in biological domains, where we look to build predictive models based on labeled training data. However, in practice, high quality labeled data is scarce, and to label new data incurs high costs. Transfer and multitask learning offer an attractive alternative, by allowing useful knowledge to be extracted and transferred from data in auxiliary domains helps counter the lack of data problem in the target domain. In this article, we survey recent advances in transfer and multitask learning for bioinformatics applications. In particular, we survey several key bioinformatics application areas, including sequence classification, gene expression data analysis, biological network reconstruction and biomedical applications.

Deep Learning-Based Occupancy Detection and Visualization for Architecture and Urban Data - Towards Augmented Reality and GIS Integration for Improved Safety and Emergency Response Modeling - (건물 내 재실자 감지 및 시각화를 위한 딥러닝 모델 - 증강현실 및 GIS 통합을 통한 안전 및 비상 대응 개선모델 프로토타이핑 -)

  • Shin, Dongyoun
    • Journal of KIBIM
    • /
    • v.13 no.2
    • /
    • pp.29-36
    • /
    • 2023
  • This study explores the potential of utilizing video-based data analysis and machine learning techniques to estimate the number of occupants within a building. The research methodology involves developing a sophisticated counting system capable of detecting and tracking individuals' entry and exit patterns. The proposed method demonstrates promising results in various scenarios; however, it also identifies the need for improvements in camera performance and external environmental conditions, such as lighting. The study emphasizes the significance of incorporating machine learning in architectural and urban planning applications, offering valuable insights for the field. In conclusion, the research calls for further investigation to address the limitations and enhance the system's accuracy, ultimately contributing to the development of a more robust and reliable solution for building occupancy estimation.