• Title/Summary/Keyword: size classification

Search Result 1,475, Processing Time 0.033 seconds

Design and Implementation of a Body Fat Classification Model using Human Body Size Data

  • Taejun Lee;Hakseong Kim;Hoekyung Jung
    • Journal of information and communication convergence engineering
    • /
    • v.21 no.2
    • /
    • pp.110-116
    • /
    • 2023
  • Recently, as various examples of machine learning have been applied in the healthcare field, deep learning technology has been applied to various tasks, such as electrocardiogram examination and body composition analysis using wearable devices such as smart watches. To utilize deep learning, securing data is the most important procedure, where human intervention, such as data classification, is required. In this study, we propose a model that uses a clustering algorithm, namely, the K-means clustering, to label body fat according to gender and age considering body size aspects, such as chest circumference and waist circumference, and classifies body fat into five groups from high risk to low risk using a convolutional neural network (CNN). As a result of model validation, accuracy, precision, and recall results of more than 95% were obtained. Thus, rational decision making can be made in the field of healthcare or obesity analysis using the proposed method.

Document Classification of Small Size Documents Using Extended Relief-F Algorithm (확장된 Relief-F 알고리즘을 이용한 소규모 크기 문서의 자동분류)

  • Park, Heum
    • The KIPS Transactions:PartB
    • /
    • v.16B no.3
    • /
    • pp.233-238
    • /
    • 2009
  • This paper presents an approach to the classifications of small size document using the instance-based feature filtering Relief-F algorithm. In the document classifications, we have not always good classification performances of small size document included a few features. Because total number of feature in the document set is large, but feature count of each document is very small relatively, so the similarities between documents are very low when we use general assessment of similarity and classifiers. Specially, in the cases of the classification of web document in the directory service and the classification of the sectors that cannot connect with the original file after recovery hard-disk, we have not good classification performances. Thus, we propose the Extended Relief-F(ERelief-F) algorithm using instance-based feature filtering algorithm Relief-F to solve problems of Relief-F as preprocess of classification. For the performance comparison, we tested information gain, odds ratio and Relief-F for feature filtering and getting those feature values, and used kNN and SVM classifiers. In the experimental results, the Extended Relief-F(ERelief-F) algorithm, compared with the others, performed best for all of the datasets and reduced many irrelevant features from document sets.

Discriminant Analysis of Binary Data by Using the Maximum Entropy Distribution

  • Lee, Jung Jin;Hwang, Joon
    • Communications for Statistical Applications and Methods
    • /
    • v.10 no.3
    • /
    • pp.909-917
    • /
    • 2003
  • Although many classification models have been used to classify binary data, none of the classification models dominates all varying circumstances depending on the number of variables and the size of data(Asparoukhov and Krzanowski (2001)). This paper proposes a classification model which uses information on marginal distributions of sub-variables and its maximum entropy distribution. Classification experiments by using simulation are discussed.

Enhancing Gene Expression Classification of Support Vector Machines with Generative Adversarial Networks

  • Huynh, Phuoc-Hai;Nguyen, Van Hoa;Do, Thanh-Nghi
    • Journal of information and communication convergence engineering
    • /
    • v.17 no.1
    • /
    • pp.14-20
    • /
    • 2019
  • Currently, microarray gene expression data take advantage of the sufficient classification of cancers, which addresses the problems relating to cancer causes and treatment regimens. However, the sample size of gene expression data is often restricted, because the price of microarray technology on studies in humans is high. We propose enhancing the gene expression classification of support vector machines with generative adversarial networks (GAN-SVMs). A GAN that generates new data from original training datasets was implemented. The GAN was used in conjunction with nonlinear SVMs that efficiently classify gene expression data. Numerical test results on 20 low-sample-size and very high-dimensional microarray gene expression datasets from the Kent Ridge Biomedical and Array Expression repositories indicate that the model is more accurate than state-of-the-art classifying models.

A Through-focus Scanning Optical Microscopy Dimensional Measurement Method based on a Deep-learning Regression Model (딥 러닝 회귀 모델 기반의 TSOM 계측)

  • Jeong, Jun Hee;Cho, Joong Hwee
    • Journal of the Semiconductor & Display Technology
    • /
    • v.21 no.1
    • /
    • pp.108-113
    • /
    • 2022
  • The deep-learning-based measurement method with the through-focus scanning optical microscopy (TSOM) estimated the size of the object using the classification. However, the measurement performance of the method depends on the number of subdivided classes, and it is practically difficult to prepare data at regular intervals for training each class. We propose an approach to measure the size of an object in the TSOM image using the deep-learning regression model instead of using classification. We attempted our proposed method to estimate the top critical dimension (TCD) of through silicon via (TSV) holes with 2461 TSOM images and the results were compared with the existing method. As a result of our experiment, the average measurement error of our method was within 30 nm (1σ) which is 1/13.5 of the sampling distance of the applied microscope. Measurement errors decreased by 31% compared to the classification result. This result proves that the proposed method is more effective and practical than the classification method.

Seafloor Classification Using Fuzzy Logic (퍼지 이론을 이용한 해저면 분류 기법)

  • 윤관섭;박순식;나정열;석동우;주진용;조진석
    • The Journal of the Acoustical Society of Korea
    • /
    • v.23 no.4
    • /
    • pp.296-302
    • /
    • 2004
  • Acoustic experiments are performed for a seafloor classification from 19 May to 25 May 2003. The six different sites of bottom composition are settled and the bottom reflection losses with frequencies (30, 50, 80. 100, 120 kHz) are measured. Sediment samples were collected using gravity core and the sample was extracted for grain size analysis. The fuzzy logic is used to classify the seabed. In the fuzzy logic. Bottom 1083 model of frequency dependence is used as the input membership functions and the output membership functions are composed of the Wentworth grain size of the bottom. The possibility of the seafloor classification is verified comparing the inversed mean grain size using fuzzy logic with the results of the coring.

A Study on Optimal Shape-Size Index Extraction for Classification of High Resolution Satellite Imagery (고해상도 영상의 분류결과 개선을 위한 최적의 Shape-Size Index 추출에 관한 연구)

  • Han, You-Kyung;Kim, Hye-Jin;Choi, Jae-Wan;Kim, Yong-Il
    • Korean Journal of Remote Sensing
    • /
    • v.25 no.2
    • /
    • pp.145-154
    • /
    • 2009
  • High spatial resolution satellite image classification has a limitation when only using the spectral information due to the complex spatial arrangement of features and spectral heterogeneity within each class. Therefore, the extraction of the spatial information is one of the most important steps in high resolution satellite image classification. This study proposes a new spatial feature extraction method, named SSI(Shape-Size Index). SSI uses a simple region-growing based image segmentation and allocates spatial property value in each segment. The extracted feature is integrated with spectral bands to improve overall classification accuracy. The classification is achieved by applying a SVM(Support Vector Machines) classifier. In order to evaluate the proposed feature extraction method, KOMPSAT-2 and QuickBird-2 data are used for experiments. It is demonstrated that proposed SSI algorithm leads to a notable increase in classification accuracy.

An Analytical Study on Performance Factors of Automatic Classification based on Machine Learning (기계학습에 기초한 자동분류의 성능 요소에 관한 연구)

  • Kim, Pan Jun
    • Journal of the Korean Society for information Management
    • /
    • v.33 no.2
    • /
    • pp.33-59
    • /
    • 2016
  • This study examined the factors affecting the performance of automatic classification for the domestic conference papers based on machine learning techniques. In particular, In view of the classification performance that assigning automatically the class labels to the papers in Proceedings of the Conference of Korean Society for Information Management using Rocchio algorithm, I investigated the characteristics of the key factors (classifier formation methods, training set size, weighting schemes, label assigning methods) through the diversified experiments. Consequently, It is more effective that apply proper parameters (${\beta}$, ${\lambda}$) and training set size (more than 5 years) according to the classification environments and properties of the document set. and If the performance is equivalent, I discovered that the use of the more simple methods (single weighting schemes) is very efficient. Also, because the classification of domestic papers is corresponding with multi-label classification which assigning more than one label to an article, it is necessary to develop the optimum classification model based on the characteristics of the key factors in consideration of this environment.

TCAM Partitioning for High-Performance Packet Classification (고성능 패킷 분류를 위한 TCAM 분할)

  • Kim Kyu-Ho;Kang Seok-Min;Song Il-Seop;Kwon Teack-Geun
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.31 no.2B
    • /
    • pp.91-97
    • /
    • 2006
  • As increasing the network bandwidth, the threat of a network also increases with emerging various new services. For a high-performance network security, It is generally used that high-speed packet classification methods which employ hardware like TCAM. There needs an method using these devices efficiently because they are expensive and their capacity is not sufficient. In this paper, we propose an efficient packet classification using a Ternary-CAM(TCAM) which is widely used device for high-speed packet classification in which we have applied Snort rule set for the well-known intrusion detection system. In order to save the size of an expensive TCAM, we have eliminated duplicated IP addresses and port numbers in the rule according to the partitioning of a table in the TCAM, and we have represented negation and range rules with reduced TCAM size. We also keep advantages of low TCAM capacity consumption and reduce the number of TCAM lookups by decreasing the TCAM partitioning using combining port numbers. According to simulation results on our TCAM partitioning, the size of a TCAM can be reduced by upto 98$\%$ and the performance does not degrade significantly for high-speed packet classification with a large amount of rules.

A Deep Learning-based Automatic Modulation Classification Method on SDR Platforms (SDR 플랫폼을 위한 딥러닝 기반의 무선 자동 변조 분류 기술 연구)

  • Jung-Ik, Jang;Jaehyuk, Choi;Young-Il, Yoon
    • Journal of IKEEE
    • /
    • v.26 no.4
    • /
    • pp.568-576
    • /
    • 2022
  • Automatic modulation classification(AMC) is a core technique in Software Defined Radio(SDR) platform that enables smart and flexible spectrum sensing and access in a wide frequency band. In this study, we propose a simple yet accurate deep learning-based method that allows AMC for variable-size radio signals. To this end, we design a classification architecture consisting of two Convolutional Neural Network(CNN)-based models, namely main and small models, which were trained on radio signal datasets with two different signal sizes, respectively. Then, for a received signal input with an arbitrary length, modulation classification is performed by augmenting the input samples using a self-replicating padding technique to fit the input layer size of our model. Experiments using the RadioML 2018.01A dataset demonstrated that the proposed method provides higher accuracy than the existing methods in all signal-to-noise ratio(SNR) domains with less computation overhead.