• Title/Summary/Keyword: deep convolution neural network

Search Result 267, Processing Time 0.031 seconds

Interpolation based Single-path Sub-pixel Convolution for Super-Resolution Multi-Scale Networks

  • Alao, Honnang;Kim, Jin-Sung;Kim, Tae Sung;Oh, Juhyen;Lee, Kyujoong
    • Journal of Multimedia Information System
    • /
    • v.8 no.4
    • /
    • pp.203-210
    • /
    • 2021
  • Deep leaning convolutional neural networks (CNN) have successfully been applied to image super-resolution (SR). Despite their great performances, SR techniques tend to focus on a certain upscale factor when training a particular model. Algorithms for single model multi-scale networks can easily be constructed if images are upscaled prior to input, but sub-pixel convolution upsampling works differently for each scale factor. Recent SR methods employ multi-scale and multi-path learning as a solution. However, this causes unshared parameters and unbalanced parameter distribution across various scale factors. We present a multi-scale single-path upsample module as a solution by exploiting the advantages of sub-pixel convolution and interpolation algorithms. The proposed model employs sub-pixel convolution for the highest scale factor among the learning upscale factors, and then utilize 1-dimension interpolation, compressing the learned features on the channel axis to match the desired output image size. Experiments are performed for the single-path upsample module, and compared to the multi-path upsample module. Based on the experimental results, the proposed algorithm reduces the upsample module's parameters by 24% and presents slightly to better performance compared to the previous algorithm.

Using 3D Deep Convolutional Neural Network with MRI Biomarker patch Images for Alzheimer's Disease Diagnosis (치매 진단을 위한 MRI 바이오마커 패치 영상 기반 3차원 심층합성곱신경망 분류 기술)

  • Yun, Joo Young;Kim, Kyeong Tae;Choi, Jae Young
    • Journal of Korea Multimedia Society
    • /
    • v.23 no.8
    • /
    • pp.940-952
    • /
    • 2020
  • The Alzheimer's disease (AD) is a neurodegenerative disease commonly found in the elderly individuals. It is one of the most common forms of dementia; patients with AD suffer from a degradation of cognitive abilities over time. To correctly diagnose AD, compuated-aided system equipped with automatic classification algorithm is of great importance. In this paper, we propose a novel deep learning based classification algorithm that takes advantage of MRI biomarker images including brain areas of hippocampus and cerebrospinal fluid for the purpose of improving the AD classification performance. In particular, we develop a new approach that effectively applies MRI biomarker patch images as input to 3D Deep Convolution Neural Network. To integrate multiple classification results from multiple biomarker patch images, we proposed the effective confidence score fusion that combine classification scores generated from soft-max layer. Experimental results show that AD classification performance can be considerably enhanced by using our proposed approach. Compared to the conventional AD classification approach relying on entire MRI input, our proposed method can improve AD classification performance of up to 10.57% thanks to using biomarker patch images. Moreover, the proposed method can attain better or comparable AD classification performances, compared to state-of-the-art methods.

A Deep Learning-based Automatic Modulation Classification Method on SDR Platforms (SDR 플랫폼을 위한 딥러닝 기반의 무선 자동 변조 분류 기술 연구)

  • Jung-Ik, Jang;Jaehyuk, Choi;Young-Il, Yoon
    • Journal of IKEEE
    • /
    • v.26 no.4
    • /
    • pp.568-576
    • /
    • 2022
  • Automatic modulation classification(AMC) is a core technique in Software Defined Radio(SDR) platform that enables smart and flexible spectrum sensing and access in a wide frequency band. In this study, we propose a simple yet accurate deep learning-based method that allows AMC for variable-size radio signals. To this end, we design a classification architecture consisting of two Convolutional Neural Network(CNN)-based models, namely main and small models, which were trained on radio signal datasets with two different signal sizes, respectively. Then, for a received signal input with an arbitrary length, modulation classification is performed by augmenting the input samples using a self-replicating padding technique to fit the input layer size of our model. Experiments using the RadioML 2018.01A dataset demonstrated that the proposed method provides higher accuracy than the existing methods in all signal-to-noise ratio(SNR) domains with less computation overhead.

A Study on Deep Learning Binary Classification of Prostate Pathological Images Using Multiple Image Enhancement Techniques (다양한 이미지 향상 기법을 사용한 전립선 병리영상 딥러닝 이진 분류 연구)

  • Park, Hyeon-Gyun;Bhattacharjee, Subrata;Deekshitha, Prakash;Kim, Cho-Hee;Choi, Heung-Kook
    • Journal of Korea Multimedia Society
    • /
    • v.23 no.4
    • /
    • pp.539-548
    • /
    • 2020
  • Deep learning technology is currently being used and applied in many different fields. Convolution neural network (CNN) is a method of artificial neural networks in deep learning, which is commonly used for analyzing different types of images through classification. In the conventional classification of histopathology images of prostate carcinomas, the rating of cancer is classified by human subjective observation. However, this approach has produced to some misdiagnosing of cancer grading. To solve this problem, CNN based classification method is proposed in this paper, to train the histological images and classify the prostate cancer grading into two classes of the benign and malignant. The CNN architecture used in this paper is based on the VGG models, which is specialized for image classification. However, color normalization was performed based on the contrast enhancement technique, and the normalized images were used for CNN training, to compare the classification results of both original and normalized images. In all cases, accuracy was over 90%, accuracy of the original was 96%, accuracy of other cases was higher, and loss was the lowest with 9%.

A Real-Time Hardware Design of CNN for Vehicle Detection (차량 검출용 CNN 분류기의 실시간 처리를 위한 하드웨어 설계)

  • Bang, Ji-Won;Jeong, Yong-Jin
    • Journal of IKEEE
    • /
    • v.20 no.4
    • /
    • pp.351-360
    • /
    • 2016
  • Recently, machine learning algorithms, especially deep learning-based algorithms, have been receiving attention due to its high classification performance. Among the algorithms, Convolutional Neural Network(CNN) is known to be efficient for image processing tasks used for Advanced Driver Assistance Systems(ADAS). However, it is difficult to achieve real-time processing for CNN in vehicle embedded software environment due to the repeated operations contained in each layer of CNN. In this paper, we propose a hardware accelerator which enhances the execution time of CNN by parallelizing the repeated operations such as convolution. Xilinx ZC706 evaluation board is used to verify the performance of the proposed accelerator. For $36{\times}36$ input images, the hardware execution time of CNN is 2.812ms in 100MHz clock frequency and shows that our hardware can be executed in real-time.

Lip Reading Method Using CNN for Utterance Period Detection (발화구간 검출을 위해 학습된 CNN 기반 입 모양 인식 방법)

  • Kim, Yong-Ki;Lim, Jong Gwan;Kim, Mi-Hye
    • Journal of Digital Convergence
    • /
    • v.14 no.8
    • /
    • pp.233-243
    • /
    • 2016
  • Due to speech recognition problems in noisy environment, Audio Visual Speech Recognition (AVSR) system, which combines speech information and visual information, has been proposed since the mid-1990s,. and lip reading have played significant role in the AVSR System. This study aims to enhance recognition rate of utterance word using only lip shape detection for efficient AVSR system. After preprocessing for lip region detection, Convolution Neural Network (CNN) techniques are applied for utterance period detection and lip shape feature vector extraction, and Hidden Markov Models (HMMs) are then used for the recognition. As a result, the utterance period detection results show 91% of success rates, which are higher performance than general threshold methods. In the lip reading recognition, while user-dependent experiment records 88.5%, user-independent experiment shows 80.2% of recognition rates, which are improved results compared to the previous studies.

Artificial intelligence application UX/UI study for language learning of children with articulation disorder (조음장애 아동의 언어학습을 위한 인공지능 애플리케이션 UX/UI 연구)

  • Yang, Eun-mi;Park, Dea-woo
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.05a
    • /
    • pp.174-176
    • /
    • 2022
  • In this paper, we present a mobile application for 'personalized customized learning' for children with articulation disorders using an artificial intelligence (AI) algorithm. A dataset (Data Set) to analyze, judge, and predict the learner's articulation situation and degree. In particular, we designed a prototype model by looking at how AI can be improved and advanced compared to existing applications from the UX/UI (GUI) aspect. So far, the focus has been on visual experience, but now it is an important time to process data and provide a UX/UI (GUI) experience to users. The UX/UI (GUI) of the proposed mobile application was to be provided according to the learner's articulation level and situation by using CRNN (Convolution Recurrent Neural Network) of DeepLearning and Auto Encoder GPT-3 (Generative Pretrained Transformer). The use of artificial intelligence algorithms will provide a learning environment with a high degree of perfection to children with articulation disorders, thereby enhancing the learning effect. I hope that you do not have any fear or discomfort in conversation by improving the perfection of articulation with 'personalized and customized learning'.

  • PDF

A Novel Text to Image Conversion Method Using Word2Vec and Generative Adversarial Networks

  • LIU, XINRUI;Joe, Inwhee
    • Annual Conference of KIPS
    • /
    • 2019.05a
    • /
    • pp.401-403
    • /
    • 2019
  • In this paper, we propose a generative adversarial networks (GAN) based text-to-image generating method. In many natural language processing tasks, which word expressions are determined by their term frequency -inverse document frequency scores. Word2Vec is a type of neural network model that, in the case of an unlabeled corpus, produces a vector that expresses semantics for words in the corpus and an image is generated by GAN training according to the obtained vector. Thanks to the understanding of the word we can generate higher and more realistic images. Our GAN structure is based on deep convolution neural networks and pixel recurrent neural networks. Comparing the generated image with the real image, we get about 88% similarity on the Oxford-102 flowers dataset.

A Method of License Plate Location and Character Recognition based on CNN

  • Fang, Wei;Yi, Weinan;Pang, Lin;Hou, Shuonan
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.8
    • /
    • pp.3488-3500
    • /
    • 2020
  • At the present time, the economy continues to flourish, and private cars have become the means of choice for most people. Therefore, the license plate recognition technology has become an indispensable part of intelligent transportation, with research and application value. In recent years, the convolution neural network for image classification is an application of deep learning on image processing. This paper proposes a strategy to improve the YOLO model by studying the deep learning convolutional neural network (CNN) and related target detection methods, and combines the OpenCV and TensorFlow frameworks to achieve efficient recognition of license plate characters. The experimental results show that target detection method based on YOLO is beneficial to shorten the training process and achieve a good level of accuracy.

Effects of Preprocessing and Feature Extraction on CNN-based Fire Detection Performance (전처리와 특징 추출이 CNN기반 화재 탐지 성능에 미치는 효과)

  • Lee, JeongHwan;Kim, Byeong Man;Shin, Yoon Sik
    • Journal of Korea Society of Industrial Information Systems
    • /
    • v.23 no.4
    • /
    • pp.41-53
    • /
    • 2018
  • Recently, the development of machine learning technology has led to the application of deep learning technology to existing image based application systems. In this context, some researches have been made to apply CNN (Convolutional Neural Network) to the field of fire detection. To verify the effects of existing preprocessing and feature extraction methods on fire detection when combined with CNN, in this paper, the recognition performance and learning time are evaluated by changing the VGG19 CNN structure while gradually increasing the convolution layer. In general, the accuracy is better when the image is not preprocessed. Also it's shown that the preprocessing method and the feature extraction method have many benefits in terms of learning speed.