• 제목/요약/키워드: Network Feature Extraction

검색결과 491건 처리시간 0.03초

웨이브렛 계수에 근거한 Fuzzy-ART 네트워크를 이용한 PVC 분류 (Classification of the PVC Using The Fuzzy-ART Network Based on Wavelet Coefficient)

  • 박광리;이경중;이윤선;윤형로
    • 대한의용생체공학회:의공학회지
    • /
    • 제20권4호
    • /
    • pp.435-442
    • /
    • 1999
  • 본 연구에서는 PVC를 분류하기 위하여 웨이브렛 계수를 기반으로 하는 fuzzy-ART 네트워크를 설계하였다. 설계된 네트워크는 feature를 추출하는 부분과 fuzzy-ART 네트워크를 학습시키는 부분으로 구성된다. 우선 feature의 문턱치 구간을 설정하기 위하여 심전도 신호의 QRS를 검출하였고, 검출된 QRS는 Haar 웨이브렛을 이용한 웨이브렛 변환에 의해 주파수 분할하였다. 분할된 주파수 중에서 입력 feature를 추출하기 위하여 저주파 영역의 6번째 계수(D6)만을 선택하였다. D6신호는 입력 feature를 구성하기 위한 문턱치를 적용하여 fuzzy-ART 네트워크의 2진수 입력 feature로 전환하였고, PVC를 분류하기 위하여 fuzzy-ART네트워크를 학습시켰다. 본 연구의 성능을 평가하기 위하여 PVC가 포함된 MIT/BIH 데이터 베이스가 사용되었으며, fuzzy-ART 네트워크의 분류성능은 96.25%이었다.

  • PDF

Cooperative network와 MLP를 이용한 PSRI 특징추출 및 자동표적인식 (A PSRI Feature Extraction and Automatic Target Recognition Using a Cooperative Network and an MLP.)

  • 전준형;김진호;최흥문
    • 전자공학회논문지B
    • /
    • 제33B권6호
    • /
    • pp.198-207
    • /
    • 1996
  • A PSRI (position, scale, and rotation invariant ) feature extraction and automatic target recognition system using a cooperative network and an MLP is proposed. We can extract position invarient features by obtaining the target center using the projection and the moment in preprocessing stage. The scale and rotation invariant features are extracted from the contour projection of the number of edge pixels on each of the concentric circles, which is input to the cooperative network. By extracting the representative PSRI features form the features and their differentiations using max-net and min-net, we can rdduce the number of input neurons of the MLP, and make the resulted automatic target recognition system less sensitive to input variances. Experiments are conduted on various complex images which are shifted, rotated, or scaled, and the results show that the proposed system is very efficient for PSRI feature extractions and automatic target recognitions.

  • PDF

GAN-based Color Palette Extraction System by Chroma Fine-tuning with Reinforcement Learning

  • Kim, Sanghyuk;Kang, Suk-Ju
    • Journal of Semiconductor Engineering
    • /
    • 제2권1호
    • /
    • pp.125-129
    • /
    • 2021
  • As the interest of deep learning, techniques to control the color of images in image processing field are evolving together. However, there is no clear standard for color, and it is not easy to find a way to represent only the color itself like the color-palette. In this paper, we propose a novel color palette extraction system by chroma fine-tuning with reinforcement learning. It helps to recognize the color combination to represent an input image. First, we use RGBY images to create feature maps by transferring the backbone network with well-trained model-weight which is verified at super resolution convolutional neural networks. Second, feature maps are trained to 3 fully connected layers for the color-palette generation with a generative adversarial network (GAN). Third, we use the reinforcement learning method which only changes chroma information of the GAN-output by slightly moving each Y component of YCbCr color gamut of pixel values up and down. The proposed method outperforms existing color palette extraction methods as given the accuracy of 0.9140.

컨볼루션 신경망의 특징맵을 사용한 객체 추적 (Object Tracking using Feature Map from Convolutional Neural Network)

  • 임수창;김도연
    • 한국멀티미디어학회논문지
    • /
    • 제20권2호
    • /
    • pp.126-133
    • /
    • 2017
  • The conventional hand-crafted features used to track objects have limitations in object representation. Convolutional neural networks, which show good performance results in various areas of computer vision, are emerging as new ways to break through the limitations of feature extraction. CNN extracts the features of the image through layers of multiple layers, and learns the kernel used for feature extraction by itself. In this paper, we use the feature map extracted from the convolution layer of the convolution neural network to create an outline model of the object and use it for tracking. We propose a method to adaptively update the outline model to cope with various environment change factors affecting the tracking performance. The proposed algorithm evaluated the validity test based on the 11 environmental change attributes of the CVPR2013 tracking benchmark and showed excellent results in six attributes.

Comparison of Machine Learning Techniques for Cyberbullying Detection on YouTube Arabic Comments

  • Alsubait, Tahani;Alfageh, Danyah
    • International Journal of Computer Science & Network Security
    • /
    • 제21권1호
    • /
    • pp.1-5
    • /
    • 2021
  • Cyberbullying is a problem that is faced in many cultures. Due to their popularity and interactive nature, social media platforms have also been affected by cyberbullying. Social media users from Arab countries have also reported being a target of cyberbullying. Machine learning techniques have been a prominent approach used by scientists to detect and battle this phenomenon. In this paper, we compare different machine learning algorithms for their performance in cyberbullying detection based on a labeled dataset of Arabic YouTube comments. Three machine learning models are considered, namely: Multinomial Naïve Bayes (MNB), Complement Naïve Bayes (CNB), and Linear Regression (LR). In addition, we experiment with two feature extraction methods, namely: Count Vectorizer and Tfidf Vectorizer. Our results show that, using count vectroizer feature extraction, the Logistic Regression model can outperform both Multinomial and Complement Naïve Bayes models. However, when using Tfidf vectorizer feature extraction, Complement Naive Bayes model can outperform the other two models.

정서재활 바이오피드백을 위한 얼굴 영상 기반 정서인식 연구 (Study of Emotion Recognition based on Facial Image for Emotional Rehabilitation Biofeedback)

  • 고광은;심귀보
    • 제어로봇시스템학회논문지
    • /
    • 제16권10호
    • /
    • pp.957-962
    • /
    • 2010
  • If we want to recognize the human's emotion via the facial image, first of all, we need to extract the emotional features from the facial image by using a feature extraction algorithm. And we need to classify the emotional status by using pattern classification method. The AAM (Active Appearance Model) is a well-known method that can represent a non-rigid object, such as face, facial expression. The Bayesian Network is a probability based classifier that can represent the probabilistic relationships between a set of facial features. In this paper, our approach to facial feature extraction lies in the proposed feature extraction method based on combining AAM with FACS (Facial Action Coding System) for automatically modeling and extracting the facial emotional features. To recognize the facial emotion, we use the DBNs (Dynamic Bayesian Networks) for modeling and understanding the temporal phases of facial expressions in image sequences. The result of emotion recognition can be used to rehabilitate based on biofeedback for emotional disabled.

전처리와 특징 추출이 CNN기반 화재 탐지 성능에 미치는 효과 (Effects of Preprocessing and Feature Extraction on CNN-based Fire Detection Performance)

  • 이정환;김병만;신윤식
    • 한국산업정보학회논문지
    • /
    • 제23권4호
    • /
    • pp.41-53
    • /
    • 2018
  • 최근 들어 머신 러닝 기술의 발달로 기존 영상 기반의 응용시스템에 딥러닝 기술을 적용하는 사례들이 늘고 있다. 이러한 맥락에서 화재 감지 분야에서도 CNN (Convolutional Neural Network)을 적용하는 시도들이 이루어지고 있다. 본 논문에서는 기존 전처리 방법과 특징 추출 방법이 CNN과 결합되었을 때 화재 탐지에 어떤 효과를 유발하는지를 검증하기 위해 인식 성능과 학습 시간을 평가해 보았다. VGG19 CNN 구조를 변경, 즉 컨볼루션층을 조금씩 늘리면서 실험을 진행한 결과, 일반적으로 전처리하지 않는 이미지를 사용한 경우가 성능이 훨씬 좋음을 확인할 수 있었다. 또한 성능적인 측면에서는 전처리 방법과 특징 추출 방법이 부정적인 영향을 미치지만 학습속도 측면에서는 많은 이득이 있음을 확인할 수 있었다.

Text-Independent Speaker Identification System Based On Vowel And Incremental Learning Neural Networks

  • Heo, Kwang-Seung;Lee, Dong-Wook;Sim, Kwee-Bo
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 제어로봇시스템학회 2003년도 ICCAS
    • /
    • pp.1042-1045
    • /
    • 2003
  • In this paper, we propose the speaker identification system that uses vowel that has speaker's characteristic. System is divided to speech feature extraction part and speaker identification part. Speech feature extraction part extracts speaker's feature. Voiced speech has the characteristic that divides speakers. For vowel extraction, formants are used in voiced speech through frequency analysis. Vowel-a that different formants is extracted in text. Pitch, formant, intensity, log area ratio, LP coefficients, cepstral coefficients are used by method to draw characteristic. The cpestral coefficients that show the best performance in speaker identification among several methods are used. Speaker identification part distinguishes speaker using Neural Network. 12 order cepstral coefficients are used learning input data. Neural Network's structure is MLP and learning algorithm is BP (Backpropagation). Hidden nodes and output nodes are incremented. The nodes in the incremental learning neural network are interconnected via weighted links and each node in a layer is generally connected to each node in the succeeding layer leaving the output node to provide output for the network. Though the vowel extract and incremental learning, the proposed system uses low learning data and reduces learning time and improves identification rate.

  • PDF

Enhanced CT-image for Covid-19 classification using ResNet 50

  • Lobna M. Abouelmagd;Manal soubhy Ali Elbelkasy
    • International Journal of Computer Science & Network Security
    • /
    • 제24권1호
    • /
    • pp.119-126
    • /
    • 2024
  • Disease caused by the coronavirus (COVID-19) is sweeping the globe. There are numerous methods for identifying this disease using a chest imaging. Computerized Tomography (CT) chest scans are used in this study to detect COVID-19 disease using a pretrain Convolutional Neural Network (CNN) ResNet50. This model is based on image dataset taken from two hospitals and used to identify Covid-19 illnesses. The pre-train CNN (ResNet50) architecture was used for feature extraction, and then fully connected layers were used for classification, yielding 97%, 96%, 96%, 96% for accuracy, precision, recall, and F1-score, respectively. When combining the feature extraction techniques with the Back Propagation Neural Network (BPNN), it produced accuracy, precision, recall, and F1-scores of 92.5%, 83%, 92%, and 87.3%. In our suggested approach, we use a preprocessing phase to improve accuracy. The image was enhanced using the Contrast Limited Adaptive Histogram Equalization (CLAHE) algorithm, which was followed by cropping the image before feature extraction with ResNet50. Finally, a fully connected layer was added for classification, with results of 99.1%, 98.7%, 99%, 98.8% in terms of accuracy, precision, recall, and F1-score.

GPR B-scan 회색조 이미지의 싱크홀 특성추출 최적 컨볼루션 신경망 백본 연구 (A Study on the Optimal Convolution Neural Network Backbone for Sinkhole Feature Extraction of GPR B-scan Grayscale Images)

  • 박영훈
    • 대한토목학회논문집
    • /
    • 제44권3호
    • /
    • pp.385-396
    • /
    • 2024
  • GPR을 활용한 싱크홀 감지 정확도 강화를 위하여 본 연구에서는 GPR B-scan 회색조 이미지의 싱크홀 특성을 최적으로 추출할 수 있는 컨볼루션 신경망을 도출하였다. 사전 훈련된 컨볼루션 신경망이 바닐라 컨볼루션 신경망보다 2배 이상의 효용성을 가지는 것으로 평가되었다. 사전 훈련된 컨볼루션 신경망에 있어서 빠른 특성 추출이 특성 추출보다 낮은 과대적합을 발생시키는 것으로 나타났다. 아키텍처 종류와 시뮬레이션 조건에 따라 top-1 검증 정확도 크기와 발생 조건 및 연산 시간이 상이한 것으로 분석되어, 사전 훈련된 컨볼루션 신경망 중 InceptionV3가 GPR B-scan 회색조 이미지의 싱크홀 감지에 가장 강건한 것으로 평가되었다. Top-1 검증 정확도와 아키텍처 효율 지수를 동시에 고려할 경우 VGG19와 VGG16가 GPR B-scan 회색조 이미지의 싱크홀 특성 추출 백본으로 높은 효율성을 가지는 것으로 분석되었으며, GPR 장비에 탑재하여 실시간으로 싱크홀 특성 추출을 할 경우에는 MobileNetV3-Large 백본이 적합한 것으로 나타났다.