• Title/Summary/Keyword: pruning method

Search Result 169, Processing Time 0.028 seconds

An Adaptive Pruning Threshold Algorithm for the Korean Address Speech Recognition (한국어 주소 음성인식의 고속화를 위한 적응 프루닝 문턱치 알고리즘)

  • 황철준;오세진;김범국;정호열;정현열
    • The Journal of the Acoustical Society of Korea
    • /
    • v.20 no.7
    • /
    • pp.55-62
    • /
    • 2001
  • In this paper, we propose a new adaptative pruning algorithm which effectively reduces the search space during the recognition process. As maximum probabilities between neighbor frames are highly interrelated, an efficient pruning threshold value can be obtained from the maximum probabilities of previous frames. The main idea is to update threshold at the present frame by a combination of previous maximum probability and hypotheses probabilities. As present threshold is obtained in on-going recognition process, the algorithm does not need any pre-experiments to find threshold values even when recognition tasks are changed. In addition, the adaptively selected threshold allows an improvement of recognition speed under different environments. The proposed algorithm has been applied to a Korean Address recognition system. Experimental results show that the proposed algorithm reduces the search space of average 14.4% and 9.14% respectively while preserving the recognition accuracy, compared to the previous method of using fixed pruning threshold values and variable pruning threshold values.

  • PDF

Application and Performance Analysis of Double Pruning Method for Deep Neural Networks (심층신경망의 더블 프루닝 기법의 적용 및 성능 분석에 관한 연구)

  • Lee, Seon-Woo;Yang, Ho-Jun;Oh, Seung-Yeon;Lee, Mun-Hyung;Kwon, Jang-Woo
    • Journal of Convergence for Information Technology
    • /
    • v.10 no.8
    • /
    • pp.23-34
    • /
    • 2020
  • Recently, the artificial intelligence deep learning field has been hard to commercialize due to the high computing power and the price problem of computing resources. In this paper, we apply a double pruning techniques to evaluate the performance of the in-depth neural network and various datasets. Double pruning combines basic Network-slimming and Parameter-prunning. Our proposed technique has the advantage of reducing the parameters that are not important to the existing learning and improving the speed without compromising the learning accuracy. After training various datasets, the pruning ratio was increased to reduce the size of the model.We confirmed that MobileNet-V3 showed the highest performance as a result of NetScore performance analysis. We confirmed that the performance after pruning was the highest in MobileNet-V3 consisting of depthwise seperable convolution neural networks in the Cifar 10 dataset, and VGGNet and ResNet in traditional convolutional neural networks also increased significantly.

An Approach to Combining Classifier with MIMO Fuzzy Model

  • Kim, Do-Wan;Park, Jin-Bae;Lee, Yeon-Woo;Joo, Young-Hoon
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 2003.05a
    • /
    • pp.182-185
    • /
    • 2003
  • This paper presents a new design algorithm for the combination with the fuzzy classifier and the Bayesian classifier. Only few attempts have so far been made at providing an effective design algorithm combining the advantages and removing the disadvantages of two classifiers. Specifically, the suggested algorithms are composed of three steps: the combining, the fuzzy-set-based pruning, and the fuzzy set tuning. In the combining, the multi-inputs and multi-outputs (MIMO) fuzzy model is used to combine two classifiers. In the fuzzy-set-based pruning, to effectively decrease the complexity of the fuzzy-Bayesian classifier and the risk of the overfitting, the analysis method of the fuzzy set and the recursive pruning method are proposesd. In the fuzzy set tuning for the misclassified feature vectors, the premise parameters are adjusted by using the gradient decent algorithm. Finally, to show the feasibility and the validity of the proposed algorithm, a computer simulation is provided.

  • PDF

Training for Huge Data set with On Line Pruning Regression by LS-SVM

  • Kim, Dae-Hak;Shim, Joo-Yong;Oh, Kwang-Sik
    • Proceedings of the Korean Statistical Society Conference
    • /
    • 2003.10a
    • /
    • pp.137-141
    • /
    • 2003
  • LS-SVM(least squares support vector machine) is a widely applicable and useful machine learning technique for classification and regression analysis. LS-SVM can be a good substitute for statistical method but computational difficulties are still remained to operate the inversion of matrix of huge data set. In modern information society, we can easily get huge data sets by on line or batch mode. For these kind of huge data sets, we suggest an on line pruning regression method by LS-SVM. With relatively small number of pruned support vectors, we can have almost same performance as regression with full data set.

  • PDF

Structure Minimization using Impact Factor in Neural Networks

  • Seo, Kap-Ho;Song, Jae-Su;Lee, Ju-Jang
    • 제어로봇시스템학회:학술대회논문집
    • /
    • 2000.10a
    • /
    • pp.484-484
    • /
    • 2000
  • The problem of determining the proper size of an neural network is recognized to be crucial, especially for its practical implications in such important issues as learning and generalization. Unfortunately, it usually is not obvious what size is best: a system that is too snail will not be able to learn the data while one that is just big enough may learn the slowly and be very sensitive to initial conditions and learning parameters. One popular technique is commonly known as pruning and consists of training a larger than necessary network and then removing unnecessary weights/nodes. In this paper, a new pruning method is developed, based on the penalty-term methods. This method makes the neural network good for the generalization and reduces the retraining time after pruning weights/nodes.

  • PDF

AutoCor: A Query Based Automatic Acquisition of Corpora of Closely-related Languages

  • Dimalen, Davis Muhajereen D.;Roxas, Rachel Edita O.
    • Proceedings of the Korean Society for Language and Information Conference
    • /
    • 2007.11a
    • /
    • pp.146-154
    • /
    • 2007
  • AutoCor is a method for the automatic acquisition and classification of corpora of documents in closely-related languages. It is an extension and enhancement of CorpusBuilder, a system that automatically builds specific minority language corpora from a closed corpus, since some Tagalog documents retrieved by CorpusBuilder are actually documents in other closely-related Philippine languages. AutoCor used the query generation method odds ratio, and introduced the concept of common word pruning to differentiate between documents of closely-related Philippine languages and Tagalog. The performance of the system using with and without pruning are compared, and common word pruning was found to improve the precision of the system.

  • PDF

An Efficient and Accurate Artificial Neural Network through Induced Learning Retardation and Pruning Training Methods Sequence

  • Bandibas, Joel;Kohyama, Kazunori;Wakita, Koji
    • Proceedings of the KSRS Conference
    • /
    • 2003.11a
    • /
    • pp.429-431
    • /
    • 2003
  • The induced learning retardation method involves the temporary inhibition of the artificial neural network’s active units from participating in the error reduction process during training. This stimulates the less active units to contribute significantly to reduce the network error. However, some less active units are not sensitive to stimulation making them almost useless. The network can then be pruned by removing the less active units to make it smaller and more efficient. This study focuses on making the network more efficient and accurate by developing the induced learning retardation and pruning sequence training method. The developed procedure results to faster learning and more accurate artificial neural network for satellite image classification.

  • PDF

An Efficient Pruning Method for Subspace Skyline Queries of Moving Objects (이동 객체의 부분차원 스카이라인 질의를 위한 효율적인 가지치기 기법)

  • Kim, Jin-Ho;Park, Young-Bae
    • Journal of KIISE:Databases
    • /
    • v.35 no.2
    • /
    • pp.182-191
    • /
    • 2008
  • Most of previous works for skyline queries have focused only on static attributes of target objects. With the advance in mobile applications, however, the need of continuous skyline queries for moving objects has been increasing. Even though several techniques to process continuous skyline queries have been proposed recently, they cannot process subspace queries, which use only the subset of attribute dimensions. Therefore it is not feasible to utilize those methods for mobile applications which must consider moving objects and subspaces simultaneously. In this paper, we propose a dominant object-based pruning method to compute subspace skyline of moving objects efficiently at query time and present the experimental results to show the effectiveness of the proposed method.

An Learning Algorithm to find the Optimized Network Structure in an Incremental Model (점증적 모델에서 최적의 네트워크 구조를 구하기 위한 학습 알고리즘)

  • Lee Jong-Chan;Cho Sang-Yeop
    • Journal of Internet Computing and Services
    • /
    • v.4 no.5
    • /
    • pp.69-76
    • /
    • 2003
  • In this paper we show a new learning algorithm for pattern classification. This algorithm considered a scheme to find a solution to a problem of incremental learning algorithm when the structure becomes too complex by noise patterns included in learning data set. Our approach for this problem uses a pruning method which terminates the learning process with a predefined criterion. In this process, an iterative model with 3 layer feedforward structure is derived from the incremental model by an appropriate manipulations. Notice that this network structure is not full-connected between upper and lower layers. To verify the effectiveness of pruning method, this network is retrained by EBP. From this results, we can find out that the proposed algorithm is effective, as an aspect of a system performence and the node number included in network structure.

  • PDF

Improved LC-trie for Efficient IP Address Lookup (효율적인 IP 주소 검색을 위한 개선된 LC-trie)

  • Kim, Jin-Soo;Kim, Jung-Hwan
    • The Journal of the Korea Contents Association
    • /
    • v.7 no.3
    • /
    • pp.50-59
    • /
    • 2007
  • IP address lookup is one of the most important and complex functions in the router. In this paper, we propose an improved technique of LC-trie to increase the performance of IP address lookup in the high performance router. We effectively apply the prefix pruning method, which is used for the compression of the forwarding table in TCAM((Ternary Content Addressable Memory), to the LC-trie. This technique can decrease the number of memory accesses and upgrade the lookup speed. Moreover, through the real forwarding table and the real traffic distribution, we evaluate the performance of our scheme in terms of the lookup time and the number of memory access, comparing with that of the previous LC-trie.