• Title/Summary/Keyword: network recognition memory

Search Result 122, Processing Time 0.023 seconds

A Study on an Image Classifier using Multi-Neural Networks (다중 신경망을 이용한 영상 분류기에 관한 연구)

  • Park, Soo-Bong;Park, Jong-An
    • The Journal of the Acoustical Society of Korea
    • /
    • v.14 no.1
    • /
    • pp.13-21
    • /
    • 1995
  • In this paper, we improve an image classifier algorithm based on neural network learning. It consists of two steps. The first is input pattern generation and the second, the global neural network implementation using an improved back-propagation algorithm. The feature vector for pattern recognition consists of the codebook data obtained from self-organization feature map learning. It decreases the input neuron number as well as the computational cost. The global neural network algorithm which is used in classifier inserts a control part and an address memory part to the back-propagation algorithm to control weights and unit-offsets. The simulation results show that it does not fall into the local minima and can implement easily the large-scale neural network. And it decreases largely the learning time.

  • PDF

Flower Recognition System Using OpenCV on Android Platform (OpenCV를 이용한 안드로이드 플랫폼 기반 꽃 인식 시스템)

  • Kim, Kangchul;Yu, Cao
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.21 no.1
    • /
    • pp.123-129
    • /
    • 2017
  • New mobile phones with high tech-camera and a large size memory have been recently launched and people upload pictures of beautiful scenes or unknown flowers in SNS. This paper develops a flower recognition system that can get information on flowers in the place where mobile communication is not even available. It consists of a registration part for reference flowers and a recognition part based on OpenCV for Android platform. A new color classification method using RGB color channel and K-means clustering is proposed to reduce the recognition processing time. And ORB for feature extraction and Brute-Force Hamming algorithm for matching are used. We use 12 kinds of flowers with four color groups, and 60 images are applied for reference DB design and 60 images for test. Simulation results show that the success rate is 83.3% and the average recognition time is 2.58 s on Huawei ALEUL00 and the proposed system is suitable for a mobile phone without a network.

Implementation of handwritten digit recognition CNN structure using GPGPU and Combined Layer (GPGPU와 Combined Layer를 이용한 필기체 숫자인식 CNN구조 구현)

  • Lee, Sangil;Nam, Kihun;Jung, Jun Mo
    • The Journal of the Convergence on Culture Technology
    • /
    • v.3 no.4
    • /
    • pp.165-169
    • /
    • 2017
  • CNN(Convolutional Nerual Network) is one of the algorithms that show superior performance in image recognition and classification among machine learning algorithms. CNN is simple, but it has a large amount of computation and it takes a lot of time. Consequently, in this paper we performed an parallel processing unit for the convolution layer, pooling layer and the fully connected layer, which consumes a lot of handling time in the process of CNN, through the SIMT(Single Instruction Multiple Thread)'s structure of GPGPU(General-Purpose computing on Graphics Processing Units).And we also expect to improve performance by reducing the number of memory accesses and directly using the output of convolution layer not storing it in pooling layer. In this paper, we use MNIST dataset to verify this experiment and confirm that the proposed CNN structure is 12.38% better than existing structure.

Deep Learning Architectures and Applications (딥러닝의 모형과 응용사례)

  • Ahn, SungMahn
    • Journal of Intelligence and Information Systems
    • /
    • v.22 no.2
    • /
    • pp.127-142
    • /
    • 2016
  • Deep learning model is a kind of neural networks that allows multiple hidden layers. There are various deep learning architectures such as convolutional neural networks, deep belief networks and recurrent neural networks. Those have been applied to fields like computer vision, automatic speech recognition, natural language processing, audio recognition and bioinformatics where they have been shown to produce state-of-the-art results on various tasks. Among those architectures, convolutional neural networks and recurrent neural networks are classified as the supervised learning model. And in recent years, those supervised learning models have gained more popularity than unsupervised learning models such as deep belief networks, because supervised learning models have shown fashionable applications in such fields mentioned above. Deep learning models can be trained with backpropagation algorithm. Backpropagation is an abbreviation for "backward propagation of errors" and a common method of training artificial neural networks used in conjunction with an optimization method such as gradient descent. The method calculates the gradient of an error function with respect to all the weights in the network. The gradient is fed to the optimization method which in turn uses it to update the weights, in an attempt to minimize the error function. Convolutional neural networks use a special architecture which is particularly well-adapted to classify images. Using this architecture makes convolutional networks fast to train. This, in turn, helps us train deep, muti-layer networks, which are very good at classifying images. These days, deep convolutional networks are used in most neural networks for image recognition. Convolutional neural networks use three basic ideas: local receptive fields, shared weights, and pooling. By local receptive fields, we mean that each neuron in the first(or any) hidden layer will be connected to a small region of the input(or previous layer's) neurons. Shared weights mean that we're going to use the same weights and bias for each of the local receptive field. This means that all the neurons in the hidden layer detect exactly the same feature, just at different locations in the input image. In addition to the convolutional layers just described, convolutional neural networks also contain pooling layers. Pooling layers are usually used immediately after convolutional layers. What the pooling layers do is to simplify the information in the output from the convolutional layer. Recent convolutional network architectures have 10 to 20 hidden layers and billions of connections between units. Training deep learning networks has taken weeks several years ago, but thanks to progress in GPU and algorithm enhancement, training time has reduced to several hours. Neural networks with time-varying behavior are known as recurrent neural networks or RNNs. A recurrent neural network is a class of artificial neural network where connections between units form a directed cycle. This creates an internal state of the network which allows it to exhibit dynamic temporal behavior. Unlike feedforward neural networks, RNNs can use their internal memory to process arbitrary sequences of inputs. Early RNN models turned out to be very difficult to train, harder even than deep feedforward networks. The reason is the unstable gradient problem such as vanishing gradient and exploding gradient. The gradient can get smaller and smaller as it is propagated back through layers. This makes learning in early layers extremely slow. The problem actually gets worse in RNNs, since gradients aren't just propagated backward through layers, they're propagated backward through time. If the network runs for a long time, that can make the gradient extremely unstable and hard to learn from. It has been possible to incorporate an idea known as long short-term memory units (LSTMs) into RNNs. LSTMs make it much easier to get good results when training RNNs, and many recent papers make use of LSTMs or related ideas.

Toward Optimal FPGA Implementation of Deep Convolutional Neural Networks for Handwritten Hangul Character Recognition

  • Park, Hanwool;Yoo, Yechan;Park, Yoonjin;Lee, Changdae;Lee, Hakkyung;Kim, Injung;Yi, Kang
    • Journal of Computing Science and Engineering
    • /
    • v.12 no.1
    • /
    • pp.24-35
    • /
    • 2018
  • Deep convolutional neural network (DCNN) is an advanced technology in image recognition. Because of extreme computing resource requirements, DCNN implementation with software alone cannot achieve real-time requirement. Therefore, the need to implement DCNN accelerator hardware is increasing. In this paper, we present a field programmable gate array (FPGA)-based hardware accelerator design of DCNN targeting handwritten Hangul character recognition application. Also, we present design optimization techniques in SDAccel environments for searching the optimal FPGA design space. The techniques we used include memory access optimization and computing unit parallelism, and data conversion. We achieved about 11.19 ms recognition time per character with Xilinx FPGA accelerator. Our design optimization was performed with Xilinx HLS and SDAccel environment targeting Kintex XCKU115 FPGA from Xilinx. Our design outperforms CPU in terms of energy efficiency (the number of samples per unit energy) by 5.88 times, and GPGPU in terms of energy efficiency by 5 times. We expect the research results will be an alternative to GPGPU solution for real-time applications, especially in data centers or server farms where energy consumption is a critical problem.

Lightweight Speaker Recognition for Pet Robots using Residuals Neural Network (잔차 신경망을 활용한 펫 로봇용 화자인식 경량화)

  • Seong-Hyun Kang;Tae-Hee Lee;Myung-Ryul Choi
    • Journal of IKEEE
    • /
    • v.28 no.2
    • /
    • pp.168-173
    • /
    • 2024
  • Speaker recognition refers to a technology that analyzes voice frequencies that are different for each individual and compares them with pre-stored voices to determine the identity of the person. Deep learning-based speaker recognition is being applied to many fields, and pet robots are one of them. However, the hardware performance of pet robots is very limited in terms of the large memory space and calculations of deep learning technology. This is an important problem that pet robots must solve in real-time interaction with users. Lightening deep learning models has become an important way to solve the above problems, and a lot of research is being done recently. In this paper, we describe the results of research on lightweight speaker recognition for pet robots by constructing a voice data set for pet robots, which is a specific command type, and comparing the results of models using residuals. In the conclusion, we present the results of the proposed method and Future research plans are described.

Development of an HTM Network Training System for Recognition of Molding Parts (부품 이미지 인식을 위한 HTM 네트워크 훈련 시스템 개발)

  • Lee, Dae-Han;Bae, Sun-Gap;Seo, Dae-Ho;Kang, Hyun-Syug;Bae, Jong-Min
    • Journal of Korea Multimedia Society
    • /
    • v.13 no.11
    • /
    • pp.1643-1656
    • /
    • 2010
  • It is necessary to develop a system to judge inferiority of goods to minimize the loss at small factories in which produces various kinds of goods with small amounts. That system can be developed based on HTM theory. HTM is a model to apply the operation principles of the neocortex in human brain to the machine learning. We have to build the trained HTM network to use the HTM-based machine learning system. It requires the knowledge for the HTM theory. This paper presents the design and implementation of the training system to support the development of HTM networks which recognize the molding parts to judge its badness. This training system allows field technicians to train the HTM network with high accuracy without the knowledge of the HTM theory. It also can be applied to any kind of the HTM-based judging systems for molding parts.

A deep learning method for the automatic modulation recognition of received radio signals (수신된 전파신호의 자동 변조 인식을 위한 딥러닝 방법론)

  • Kim, Hanjin;Kim, Hyeockjin;Je, Junho;Kim, Kyungsup
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.23 no.10
    • /
    • pp.1275-1281
    • /
    • 2019
  • The automatic modulation recognition of a radio signal is a major task of an intelligent receiver, with various civilian and military applications. In this paper, we propose a method to recognize the modulation of radio signals in wireless communication based on the deep neural network. We classify the modulation pattern of radio signal by using the LSTM model, which can catch the long-term pattern for the sequential data as the input data of the deep neural network. The amplitude and phase of the modulated signal, the in-phase carrier, and the quadrature-phase carrier are used as input data in the LSTM model. In order to verify the performance of the proposed learning method, we use a large dataset for training and test, including the ten types of modulation signal under various signal-to-noise ratios.

A Dynamic Programming Neural Network to find the Safety Distance of Industrial Field (산업 현장의 안전거리 계측을 위한 동적 계획 신경회로망)

  • Kim, Jong-Man;Kim, Won-Sub;Kim, Yeong-Min;Hwang, Jong-Sun;Park, Hyun-Chul
    • Proceedings of the Korean Institute of Electrical and Electronic Material Engineers Conference
    • /
    • 2001.09a
    • /
    • pp.23-27
    • /
    • 2001
  • Making the safety situation from the various work system is very important in the industrial fields. The proposed neural network technique is the real titre computation method based theory of inter-node diffusion for searching the safety distances from the sudden appearance-objests during the work driving. The main steps of the distance computation using the theory of stereo vision like the eyes of man is following steps. One is the processing for finding the corresponding points of stereo images and the other is the interpolation processing of full image data from nonlinear image data of obejects. All of them request much memory space and titre. Therefore the most reliable neural-network algorithm is drived for real time recognition of obejects, which is composed of a dynamic programming algorithm based on sequence matching techniques. And the real time reconstruction of nonlinear image information is processed through several simulations. I-D LIPN hardware has been composed, and the real time reconstruction is verified through the various experiments.

  • PDF

Trends in Deep-neural-network-based Dialogue Systems (심층 신경망 기반 대화처리 기술 동향)

  • Kwon, O.W.;Hong, T.G.;Huang, J.X.;Roh, Y.H.;Choi, S.K.;Kim, H.Y.;Kim, Y.K.;Lee, Y.K.
    • Electronics and Telecommunications Trends
    • /
    • v.34 no.4
    • /
    • pp.55-64
    • /
    • 2019
  • In this study, we introduce trends in neural-network-based deep learning research applied to dialogue systems. Recently, end-to-end trainable goal-oriented dialogue systems using long short-term memory, sequence-to-sequence models, among others, have been studied to overcome the difficulties of domain adaptation and error recognition and recovery in traditional pipeline goal-oriented dialogue systems. In addition, some research has been conducted on applying reinforcement learning to end-to-end trainable goal-oriented dialogue systems to learn dialogue strategies that do not appear in training corpora. Recent neural network models for end-to-end trainable chit-chat systems have been improved using dialogue context as well as personal and topic information to produce a more natural human conversation. Unlike previous studies that have applied different approaches to goal-oriented dialogue systems and chit-chat systems respectively, recent studies have attempted to apply end-to-end trainable approaches based on deep neural networks in common to them. Acquiring dialogue corpora for training is now necessary. Therefore, future research will focus on easily and cheaply acquiring dialogue corpora and training with small annotated dialogue corpora and/or large raw dialogues.