• Title/Summary/Keyword: Gradient feature

Search Result 279, Processing Time 0.024 seconds

The Study of Support Vector Machine-based HOG (Histogram of Oriented Gradients) Feature Vector for Recognition by Numerical Sign Language (숫자 수화 인식을 위한 서포트 벡터 머신 기반의 HOG(Histogram of Oriented Gradients) 특징 벡터 연구)

  • Lee, SeungHwan;Yoo, JaeChern
    • Proceedings of the Korean Society of Computer Information Conference
    • /
    • 2019.07a
    • /
    • pp.271-272
    • /
    • 2019
  • 현재 4차 산업혁명으로 인해 많은 이들의 삶의 질이 이전보다 개선되었음에도 불구하고, 소외된 계층을 위한 개발은 타 분야에 비해서 더뎌지고 있는 실정이다. 현대의 청각 장애인과 언어 장애인들은 시각 언어인 수화를 이용하여 의사소통을 한다. 그러나 수화는 진입 장벽이 높기 때문에, 이를 사용하지 않는 사람들은 청각 장애인 및 언어 장애인과 의사소통을 하는데 어려움을 겪는다. 본 논문은 이러한 불편함을 줄이기 위해 서포트 벡터 머신(Support Vector Machine, SVM) 기반의 HOG(Histogram of Oriented Gradients) 특징 벡터를 이용하여 수화의 기본인 숫자를 분류할 수 있는 시스템을 구현하여 수화를 번역할 수 있는 가능성을 제안한다.

  • PDF

Anomaly-based Alzheimer's disease detection using entropy-based probability Positron Emission Tomography images

  • Husnu Baris Baydargil;Jangsik Park;Ibrahim Furkan Ince
    • ETRI Journal
    • /
    • v.46 no.3
    • /
    • pp.513-525
    • /
    • 2024
  • Deep neural networks trained on labeled medical data face major challenges owing to the economic costs of data acquisition through expensive medical imaging devices, expert labor for data annotation, and large datasets to achieve optimal model performance. The heterogeneity of diseases, such as Alzheimer's disease, further complicates deep learning because the test cases may substantially differ from the training data, possibly increasing the rate of false positives. We propose a reconstruction-based self-supervised anomaly detection model to overcome these challenges. It has a dual-subnetwork encoder that enhances feature encoding augmented by skip connections to the decoder for improving the gradient flow. The novel encoder captures local and global features to improve image reconstruction. In addition, we introduce an entropy-based image conversion method. Extensive evaluations show that the proposed model outperforms benchmark models in anomaly detection and classification using an encoder. The supervised and unsupervised models show improved performances when trained with data preprocessed using the proposed image conversion method.

Incorporating BERT-based NLP and Transformer for An Ensemble Model and its Application to Personal Credit Prediction

  • Sophot Ky;Ju-Hong Lee;Kwangtek Na
    • Smart Media Journal
    • /
    • v.13 no.4
    • /
    • pp.9-15
    • /
    • 2024
  • Tree-based algorithms have been the dominant methods used build a prediction model for tabular data. This also includes personal credit data. However, they are limited to compatibility with categorical and numerical data only, and also do not capture information of the relationship between other features. In this work, we proposed an ensemble model using the Transformer architecture that includes text features and harness the self-attention mechanism to tackle the feature relationships limitation. We describe a text formatter module, that converts the original tabular data into sentence data that is fed into FinBERT along with other text features. Furthermore, we employed FT-Transformer that train with the original tabular data. We evaluate this multi-modal approach with two popular tree-based algorithms known as, Random Forest and Extreme Gradient Boosting, XGBoost and TabTransformer. Our proposed method shows superior Default Recall, F1 score and AUC results across two public data sets. Our results are significant for financial institutions to reduce the risk of financial loss regarding defaulters.

An Effective Feature Extraction Method for Fault Diagnosis of Induction Motors (유도전동기의 고장 진단을 위한 효과적인 특징 추출 방법)

  • Nguyen, Hung N.;Kim, Jong-Myon
    • Journal of the Korea Society of Computer and Information
    • /
    • v.18 no.7
    • /
    • pp.23-35
    • /
    • 2013
  • This paper proposes an effective technique that is used to automatically extract feature vectors from vibration signals for fault classification systems. Conventional mel-frequency cepstral coefficients (MFCCs) are sensitive to noise of vibration signals, degrading classification accuracy. To solve this problem, this paper proposes spectral envelope cepstral coefficients (SECC) analysis, where a 4-step filter bank based on spectral envelopes of vibration signals is used: (1) a linear predictive coding (LPC) algorithm is used to specify spectral envelopes of all faulty vibration signals, (2) all envelopes are averaged to get general spectral shape, (3) a gradient descent method is used to find extremes of the average envelope and its frequencies, (4) a non-overlapped filter is used to have centers calculated from distances between valley frequencies of the envelope. This 4-step filter bank is then used in cepstral coefficients computation to extract feature vectors. Finally, a multi-layer support vector machine (MLSVM) with various sigma values uses these special parameters to identify faulty types of induction motors. Experimental results indicate that the proposed extraction method outperforms other feature extraction algorithms, yielding more than about 99.65% of classification accuracy.

Comparative Study of GDPA and Hough Transformation for Linear Feature Extraction using Space-borne Imagery (위성 영상정보를 이용한 선형 지형지물 추출에서의 GDPA와 Hough 변환 처리결과 비교연구)

  • Lee Kiwon;Ryu Hee-Young;Kwon Byung-Doo
    • Korean Journal of Remote Sensing
    • /
    • v.20 no.4
    • /
    • pp.261-274
    • /
    • 2004
  • The feature extraction using remotely sensed imagery has been recognized one of the important tasks in remote sensing applications. As the high-resolution imagery are widely used to the engineering purposes, need of more accurate feature information also is increasing. Especially, in case of the automatic extraction of linear feature such as road using mid or low-resolution imagery, several techniques was developed and applied in the mean time. But quantitatively comparative analysis of techniques and case studies for high-resolution imagery is rare. In this study, we implemented a computer program to perform and compare GDPA (Gradient Direction Profile Analysis) algorithm and Hough transformation. Also the results of applying two techniques to some images were compared with road centerline layers and boundary layers of digital map and presented. For quantitative comparison, the ranking method using commission error and omission error was used. As results, Hough transform had high accuracy over 20% on the average. As for execution speed, GDPA shows main advantage over Hough transform. But the accuracy was not remarkable difference between GDPA and Hough transform, when the noise removal was app]ied to the result of GDPA. In conclusion, it is expected that GDPA have more advantage than Hough transform in the application side.

Prediction of Residual Axillary Nodal Metastasis Following Neoadjuvant Chemotherapy for Breast Cancer: Radiomics Analysis Based on Chest Computed Tomography

  • Hyo-jae Lee;Anh-Tien Nguyen;Myung Won Song;Jong Eun Lee;Seol Bin Park;Won Gi Jeong;Min Ho Park;Ji Shin Lee;Ilwoo Park;Hyo Soon Lim
    • Korean Journal of Radiology
    • /
    • v.24 no.6
    • /
    • pp.498-511
    • /
    • 2023
  • Objective: To evaluate the diagnostic performance of chest computed tomography (CT)-based qualitative and radiomics models for predicting residual axillary nodal metastasis after neoadjuvant chemotherapy (NAC) for patients with clinically node-positive breast cancer. Materials and Methods: This retrospective study included 226 women (mean age, 51.4 years) with clinically node-positive breast cancer treated with NAC followed by surgery between January 2015 and July 2021. Patients were randomly divided into the training and test sets (4:1 ratio). The following predictive models were built: a qualitative CT feature model using logistic regression based on qualitative imaging features of axillary nodes from the pooled data obtained using the visual interpretations of three radiologists; three radiomics models using radiomics features from three (intranodal, perinodal, and combined) different regions of interest (ROIs) delineated on pre-NAC CT and post-NAC CT using a gradient-boosting classifier; and fusion models integrating clinicopathologic factors with the qualitative CT feature model (referred to as clinical-qualitative CT feature models) or with the combined ROI radiomics model (referred to as clinical-radiomics models). The area under the curve (AUC) was used to assess and compare the model performance. Results: Clinical N stage, biological subtype, and primary tumor response indicated by imaging were associated with residual nodal metastasis during the multivariable analysis (all P < 0.05). The AUCs of the qualitative CT feature model and radiomics models (intranodal, perinodal, and combined ROI models) according to post-NAC CT were 0.642, 0.812, 0.762, and 0.832, respectively. The AUCs of the clinical-qualitative CT feature model and clinical-radiomics model according to post-NAC CT were 0.740 and 0.866, respectively. Conclusion: CT-based predictive models showed good diagnostic performance for predicting residual nodal metastasis after NAC. Quantitative radiomics analysis may provide a higher level of performance than qualitative CT features models. Larger multicenter studies should be conducted to confirm their performance.

Multi-view Image Generation from Stereoscopic Image Features and the Occlusion Region Extraction (가려짐 영역 검출 및 스테레오 영상 내의 특징들을 이용한 다시점 영상 생성)

  • Lee, Wang-Ro;Ko, Min-Soo;Um, Gi-Mun;Cheong, Won-Sik;Hur, Nam-Ho;Yoo, Ji-Sang
    • Journal of Broadcast Engineering
    • /
    • v.17 no.5
    • /
    • pp.838-850
    • /
    • 2012
  • In this paper, we propose a novel algorithm that generates multi-view images by using various image features obtained from the given stereoscopic images. In the proposed algorithm, we first create an intensity gradient saliency map from the given stereo images. And then we calculate a block-based optical flow that represents the relative movement(disparity) of each block with certain size between left and right images. And we also obtain the disparities of feature points that are extracted by SIFT(scale-invariant We then create a disparity saliency map by combining these extracted disparity features. Disparity saliency map is refined through the occlusion detection and removal of false disparities. Thirdly, we extract straight line segments in order to minimize the distortion of straight lines during the image warping. Finally, we generate multi-view images by grid mesh-based image warping algorithm. Extracted image features are used as constraints during grid mesh-based image warping. The experimental results show that the proposed algorithm performs better than the conventional DIBR algorithm in terms of visual quality.

Forecasting the Busan Container Volume Using XGBoost Approach based on Machine Learning Model (기계 학습 모델을 통해 XGBoost 기법을 활용한 부산 컨테이너 물동량 예측)

  • Nguyen Thi Phuong Thanh;Gyu Sung Cho
    • Journal of Internet of Things and Convergence
    • /
    • v.10 no.1
    • /
    • pp.39-45
    • /
    • 2024
  • Container volume is a very important factor in accurate evaluation of port performance, and accurate prediction of effective port development and operation strategies is essential. However, it is difficult to improve the accuracy of container volume prediction due to rapid changes in the marine industry. To solve this problem, it is necessary to analyze the impact on port performance using the Internet of Things (IoT) and apply it to improve the competitiveness and efficiency of Busan Port. Therefore, this study aims to develop a prediction model for predicting the future container volume of Busan Port, and through this, focuses on improving port productivity and making improved decision-making by port management agencies. In order to predict port container volume, this study introduced the Extreme Gradient Boosting (XGBoost) technique of a machine learning model. XGBoost stands out of its higher accuracy, faster learning and prediction than other algorithms, preventing overfitting, along with providing Feature Importance. Especially, XGBoost can be used directly for regression predictive modelling, which helps improve the accuracy of the volume prediction model presented in previous studies. Through this, this study can accurately and reliably predict container volume by the proposed method with a 4.3% MAPE (Mean absolute percentage error) value, highlighting its high forecasting accuracy. It is believed that the accuracy of Busan container volume can be increased through the methodology presented in this study.

Multiresolutional Reconstruction from Contours (윤곽선을 이용한 다중해상도적 복원)

  • 민경하;이인권
    • Journal of KIISE:Computer Systems and Theory
    • /
    • v.30 no.11
    • /
    • pp.629-654
    • /
    • 2003
  • A new multiresolutional scheme that reconstructs a polygonal mesh from the set of contours is presented. In the first step, we apply a radial gradient method to extract the contours on the sampled slices from a volume data. After classifying the types of the edges on the contours, we represent the contour using the context-free grammar. The polygons between two neighboring contours are generated through the traversal of the derivation trees of the context-free grammar. The polygonal surface of the coarsest resolution is refined through the refinement of the contours, which is executed by casting more rays on the slices. The topologies between the polygonal surfaces of various resolutions are maintained from the fact that the radial gradient method preserves the topologies of the contours of various resolutions. The proposed scheme provides efficient computation and compression methods for the tiling procedure with the feature preservation.

On Robust Principal Component using Analysis Neural Networks (신경망을 이용한 로버스트 주성분 분석에 관한 연구)

  • Kim, Sang-Min;Oh, Kwang-Sik;Park, Hee-Joo
    • Journal of the Korean Data and Information Science Society
    • /
    • v.7 no.1
    • /
    • pp.113-118
    • /
    • 1996
  • Principal component analysis(PCA) is an essential technique for data compression and feature extraction, and has been widely used in statistical data analysis, communication theory, pattern recognition, and image processing. Oja(1992) found that a linear neuron with constrained Hebbian learning rule can extract the principal component by using stochastic gradient ascent method. In practice real data often contain some outliers. These outliers will significantly deteriorate the performances of the PCA algorithms. In order to make PCA robust, Xu & Yuille(1995) applied statistical physics to the problem of robust principal component analysis(RPCA). Devlin et.al(1981) obtained principal components by using techniques such as M-estimation. The propose of this paper is to investigate from the statistical point of view how Xu & Yuille's(1995) RPCA works under the same simulation condition as in Devlin et.al(1981).

  • PDF