• 제목/요약/키워드: steepest gradient descent

검색결과 34건 처리시간 0.019초

Optimizing Feature Extractioin for Multiclass problems Based on Classification Error (다중 클래스 데이터를 위한 분류오차 최소화기반 특징추출 기법)

  • Choi, Eui-Sun;Lee, Chul-Hee
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • 제37권2호
    • /
    • pp.39-49
    • /
    • 2000
  • In this paper, we propose an optimizing feature extraction method for multiclass problems assuming normal distributions. Initially, We start with an arbitrary feature vector Assuming that the feature vector is used for classification, we compute the classification error Then we move the feature vector slightly in the direction so that classification error decreases most rapidly This can be done by taking gradient We propose two search methods, sequential search and global search In the sequential search, an additional feature vector is selected so that it provides the best accuracy along with the already chosen feature vectors In the global search, we are not constrained to use the chosen feature vectors Experimental results show that the proposed algorithm provides a favorable performance.

  • PDF

Analysis of Microwave Inverse Scattering Using the Broadband Electromagnetic Waves (광대역 전자파를 이용한 역산란 해석 연구)

  • Lee Jung-Hoon;Chung Young-Seek;So Joon-Ho;Kim Junyeon;Jang Won
    • The Journal of Korean Institute of Electromagnetic Engineering and Science
    • /
    • 제17권2호
    • /
    • pp.158-164
    • /
    • 2006
  • In this paper, we proposed a new algorithm of the inverse scattering for the reconstruction of unknown dielectric scatterers using the finite-difference time-domain method and the design sensitivity analysis. We introduced the design sensitivity analysis based on the gradient information for the fast convergence of the reconstruction. By introducing the adjoint variable method for the efficient calculation, we derived the adjoint variable equation. As an optimal algorithm, we used the steepest descent method and reconstructed the dielectric targets using the iterative estimation. To verify our algorithm, we will show the numerical examples for the two-dimensional $TM^2$ cases.

Time-domain Seismic Waveform Inversion for Anisotropic media (이방성을 고려한 탄성매질에서의 시간영역 파형역산)

  • Lee, Ho-Yong;Min, Dong-Joo;Kwon, Byung-Doo;Yoo, Hai-Soo
    • 한국지구물리탐사학회:학술대회논문집
    • /
    • 한국지구물리탐사학회 2008년도 공동학술대회
    • /
    • pp.51-56
    • /
    • 2008
  • The waveform inversion for isotropic media has ever been studied since the 1980s, but there has been few studies for anisotropic media. We present a seismic waveform inversion algorithm for 2-D heterogeneous transversely isotropic structures. A cell-based finite difference algorithm for anisotropic media in time domain is adopted. The steepest descent during the non-linear iterative inversion approach is obtained by backpropagating residual errors using a reverse time migration technique. For scaling the gradient of a misfit function, we use the pseudo Hessian matrix which is assumed to neglect the zero-lag auto-correlation terms of impulse responses in the approximate Hessian matrix of the Gauss-Newton method. We demonstrate the use of these waveform inversion algorithm by applying them to a two layer model and the anisotropic Marmousi model data. With numerical examples, we show that it's difficult to converge to the true model when we assumed that anisotropic media are isotropic. Therefore, it is expected that our waveform inversion algorithm for anisotropic media is adequate to interpret real seismic exploration data.

  • PDF

A Fast-Loaming Algorithm for MLP in Pattern Recognition (패턴인식의 MLP 고속학습 알고리즘)

  • Lee, Tae-Seung;Choi, Ho-Jin
    • Journal of KIISE:Computing Practices and Letters
    • /
    • 제8권3호
    • /
    • pp.344-355
    • /
    • 2002
  • Having a variety of good characteristics against other pattern recognition techniques, Multilayer Perceptron (MLP) has been used in wide applications. But, it is known that Error Backpropagation (EBP) algorithm which MLP uses in learning has a defect that requires relatively long leaning time. Because learning data in pattern recognition contain abundant redundancies, in order to increase learning speed it is very effective to use online-based teaming methods, which update parameters of MLP pattern by pattern. Typical online EBP algorithm applies fixed learning rate for each update of parameters. Though a large amount of speedup with online EBP can be obtained by choosing an appropriate fixed rate, fixing the rate leads to the problem that the algorithm cannot respond effectively to different leaning phases as the phases change and the learning pattern areas vary. To solve this problem, this paper defines learning as three phases and proposes a Instant Learning by Varying Rate and Skipping (ILVRS) method to reflect only necessary patterns when learning phases change. The basic concept of ILVRS is as follows. To discriminate and use necessary patterns which change as learning proceeds, (1) ILVRS uses a variable learning rate which is an error calculated from each pattern and is suppressed within a proper range, and (2) ILVRS bypasses unnecessary patterns in loaming phases. In this paper, an experimentation is conducted for speaker verification as an application of pattern recognition, and the results are presented to verify the performance of ILVRS.