• Title/Summary/Keyword: 다층신경망

Search Result 347, Processing Time 0.025 seconds

Prediction of Groundwater Level in Jeju Island Using Deep Learning Algorithm MLP and LSTM (딥러닝 알고리즘 MLP 및 LSTM을 활용한 제주도 지하수위 예측)

  • Kang, Dayoung;Byun, Kyuhyun
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2022.05a
    • /
    • pp.206-206
    • /
    • 2022
  • 제주도는 투수성이 좋은 대수층이 발달한 화산섬으로 지하수가 가장 중요한 수자원이다. 인위적 요인과 기후변화로 인해 제주도의 지하수위가 저하하는 추세를 보이고 있음에 따라 지하수의 적정 관리를 위해 지하수위의 정확하고 장기적인 예측이 매우 중요하다. 다양한 환경적인 요인이 지하수의 함양 및 수위에 영향을 미치는 것으로 알려져 있지만, 제주도의 특징적인 기상인자가 지하수 시스템에 어떻게 영향을 미치는지를 파악하기 위한 연구는 거의 진행되지 않았다. 지하수위측에 있어서 물리적 모델을 이용한 방안은 다양한 조건에 의해 변화하는 지하수위의 정확하고 빠른 예측에 한계가 있는 것으로 알려져 있다. 이에 본 연구에서는 제주도 애월읍과 남원읍에 위치한 지하수위 관측정의 일 수위자료와 강수량, 온도, 강설량, 풍속, VPD의 다양한 기상 자료를 대상으로 인공신경망 알고리즘인 다층 퍼셉트론(MLP)와 Long Short Term Memory(LSTM)에 기반한 표준지하수지수(SGI) 예측 모델을 개발하였다. MLP와 LSTM의 표준지하수지수(SGI) 예측결과가 상당히 유사한 것으로 나타났으며 MLP과 LSTM 예측모델의 결정계수(R2)는 애월읍의 경우 각각 0.98, 남원읍의 경우 각각 0.96으로 높은 값을 보였다. 본 연구에서 개발한 지하수위 예측모델을 통해 효율적인 운영과 정밀한 지하수위 예측이 가능해질 것이며 기후변화 대응을 위한 지속가능한 지하수자원 관리 방안 마련에 도움을 줄 것이라 판단된다.

  • PDF

Prediction of Music Generation on Time Series Using Bi-LSTM Model (Bi-LSTM 모델을 이용한 음악 생성 시계열 예측)

  • Kwangjin, Kim;Chilwoo, Lee
    • Smart Media Journal
    • /
    • v.11 no.10
    • /
    • pp.65-75
    • /
    • 2022
  • Deep learning is used as a creative tool that could overcome the limitations of existing analysis models and generate various types of results such as text, image, and music. In this paper, we propose a method necessary to preprocess audio data using the Niko's MIDI Pack sound source file as a data set and to generate music using Bi-LSTM. Based on the generated root note, the hidden layers are composed of multi-layers to create a new note suitable for the musical composition, and an attention mechanism is applied to the output gate of the decoder to apply the weight of the factors that affect the data input from the encoder. Setting variables such as loss function and optimization method are applied as parameters for improving the LSTM model. The proposed model is a multi-channel Bi-LSTM with attention that applies notes pitch generated from separating treble clef and bass clef, length of notes, rests, length of rests, and chords to improve the efficiency and prediction of MIDI deep learning process. The results of the learning generate a sound that matches the development of music scale distinct from noise, and we are aiming to contribute to generating a harmonistic stable music.

Automatic Interpretation of F-18-FDG Brain PET Using Artificial Neural Network: Discrimination of Medial and Lateral Temporal Lobe Epilepsy (인공신경회로망을 이용한 뇌 F-18-FDG PET 자동 해석: 내.외측 측두엽간질의 감별)

  • Lee, Jae-Sung;Lee, Dong-Soo;Kim, Seok-Ki;Park, Kwang-Suk;Lee, Sang-Kun;Chung, June-Key;Lee, Myung-Chul
    • The Korean Journal of Nuclear Medicine
    • /
    • v.38 no.3
    • /
    • pp.233-240
    • /
    • 2004
  • Purpose: We developed a computer-aided classifier using artificial neural network (ANN) to discriminate the cerebral metabolic pattern of medial and lateral temporal lobe epilepsy (TLE). Materials and Methods: We studied brain F-18-FDG PET images of 113 epilepsy patients sugically and pathologically proven as medial TLE (left 41, right 42) or lateral TLE (left 14, right 16). PET images were spatially transformed onto a standard template and normalized to the mean counts of cortical regions. Asymmetry indices for predefined 17 mirrored regions to hemispheric midline and those for medial and lateral temporal lobes were used as input features for ANN. ANN classifier was composed of 3 independent multi-layered perceptrons (1 for left/right lateralization and 2 for medial/lateral discrimination) and trained to interpret metabolic patterns and produce one of 4 diagnoses (L/R medial TLE or L/R lateral TLE). Randomly selected 8 images from each group were used to train the ANN classifier and remaining 51 images were used as test sets. The accuracy of the diagnosis with ANN was estimated by averaging the agreement rates of independent 50 trials and compared to that of nuclear medicine experts. Results: The accuracy in lateralization was 89% by the human experts and 90% by the ANN classifier Overall accuracy in localization of epileptogenic zones by the ANN classifier was 69%, which was comparable to that by the human experts (72%). Conclusion: We conclude that ANN classifier performed as well as human experts and could be potentially useful supporting tool for the differential diagnosis of TLE.

Face Recognition based on Hybrid Classifiers with Virtual Samples (가상 데이터와 융합 분류기에 기반한 얼굴인식)

  • 류연식;오세영
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.40 no.1
    • /
    • pp.19-29
    • /
    • 2003
  • This paper presents a novel hybrid classifier for face recognition with artificially generated virtual training samples. We utilize both the nearest neighbor approach in feature angle space and a connectionist model to obtain a synergy effect by combining the results of two heterogeneous classifiers. First, a classifier called the nearest feature angle (NFA), based on angular information, finds the most similar feature to the query from a given training set. Second, a classifier has been developed based on the recall of stored frontal projection of the query feature. It uses a frontal recall network (FRN) that finds the most similar frontal one among the stored frontal feature set. For FRN, we used an ensemble neural network consisting of multiple multiplayer perceptrons (MLPs), each of which is trained independently to enhance generalization capability. Further, both classifiers used the virtual training set generated adaptively, according to the spatial distribution of each person's training samples. Finally, the results of the two classifiers are combined to comprise the best matching class, and a corresponding similarit measure is used to make the final decision. The proposed classifier achieved an average classification rate of 96.33% against a large group of different test sets of images, and its average error rate is 61.5% that of the nearest feature line (NFL) method, and achieves a more robust classification performance.

A Study on the Spoken Korean Citynames Using Multi-Layered Perceptron of Back-Propagation Algorithm (오차 역전파 알고리즘을 갖는 MLP를 이용한 한국 지명 인식에 대한 연구)

  • Song, Do-Sun;Lee, Jae-Gheon;Kim, Seok-Dong;Lee, Haing-Sei
    • The Journal of the Acoustical Society of Korea
    • /
    • v.13 no.6
    • /
    • pp.5-14
    • /
    • 1994
  • This paper is about an experiment of speaker-independent automatic Korean spoken words recognition using Multi-Layered Perceptron and Error Back-propagation algorithm. The object words are 50 citynames of D.D.D local numbers. 43 of those are 2 syllables and the rest 7 are 3 syllables. The words were not segmented into syllables or phonemes, and some feature components extracted from the words in equal gap were applied to the neural network. That led independent result on the speech duration, and the PARCOR coefficients calculated from the frames using linear predictive analysis were employed as feature components. This paper tried to find out the optimum conditions through 4 differerent experiments which are comparison between total and pre-classified training, dependency of recognition rate on the number of frames and PAROCR order, recognition change due to the number of neurons in the hidden layer, and the comparison of the output pattern composition method of output neurons. As a result, the recognition rate of $89.6\%$ is obtaimed through the research.

  • PDF

Steganalysis Using Joint Moment of Wavelet Subbands (웨이블렛 부밴드의 조인트 모멘트를 이용한 스테그분석)

  • Park, Tae-Hee;Hyun, Seung-Hwa;Kim, Jae-Ho;Eom, Il-Kyu
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.48 no.3
    • /
    • pp.71-78
    • /
    • 2011
  • This paper propose image steganalysis scheme based on independence between parent and child subband on the multi-layer wavelet domain. The proposed method decompose cover and stego images into 12 subbands by applying 3-level Haar UWT(Undecimated Wavelet Transform), analyze statistical independency between parent and child subband. Because this independency is appeared more difference in stego image than in cover image, we can use it as feature to differenciate between cover and stego image. Therefore we extract 72D features by calculation first 3 order statistical moments from joint characteristic function between parent and child subband. Multi-layer perceptron(MLP) is applied as classifier to discriminate between cover and stego image. We test the performance of proposed scheme over various embedding rates by the LSB, SS, BSS embedding method. The proposed scheme outperforms the previous schemes in detection rate to existence of hidden message as well as exactness of discrimination.

A Survey on Oil Spill and Weather Forecast Using Machine Learning Based on Neural Networks and Statistical Methods (신경망 및 통계 기법 기반의 기계학습을 이용한 유류유출 및 기상 예측 연구 동향)

  • Kim, Gyoung-Do;Kim, Yong-Hyuk
    • Journal of the Korea Convergence Society
    • /
    • v.8 no.10
    • /
    • pp.1-8
    • /
    • 2017
  • Accurate forecasting enables to effectively prepare for future phenomenon. Especially, meteorological phenomenon is closely related with human life, and it can prevent from damage such as human life and property through forecasting of weather and disaster that can occur. To respond quickly and effectively to oil spill accidents, it is important to accurately predict the movement of oil spills and the weather in the surrounding waters. In this paper, we selected four representative machine learning techniques: support vector machine, Gaussian process, multilayer perceptron, and radial basis function network that have shown good performance and predictability in the previous studies related to oil spill detection and prediction in meteorology such as wind, rainfall and ozone. we suggest the applicability of oil spill prediction model based on machine learning.

Gaze Detection by Computing Facial Rotation and Translation (얼굴의 회전 및 이동 분석에 의한 응시 위치 파악)

  • Lee, Jeong-Jun;Park, Kang-Ryoung;Kim, Jai-Hie
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.39 no.5
    • /
    • pp.535-543
    • /
    • 2002
  • In this paper, we propose a new gaze detection method using 2-D facial images captured by a camera on top of the monitor. We consider only the facial rotation and translation and not the eyes' movements. The proposed method computes the gaze point caused by the facial rotation and the amount of the facial translation respectively, and by combining these two the final gaze point on a monitor screen can be obtained. We detected the gaze point caused by the facial rotation by using a neural network(a multi-layered perceptron) whose inputs are the 2-D geometric changes of the facial features' points and estimated the amount of the facial translation by image processing algorithms in real time. Experimental results show that the gaze point detection accuracy between the computed positions and the real ones is about 2.11 inches in RMS error when the distance between the user and a 19-inch monitor is about 50~70cm. The processing time is about 0.7 second with a Pentium PC(233MHz) and 320${\times}$240 pixel-size images.

Dynamic Gesture Recognition for the Remote Camera Robot Control (원격 카메라 로봇 제어를 위한 동적 제스처 인식)

  • Lee Ju-Won;Lee Byung-Ro
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.8 no.7
    • /
    • pp.1480-1487
    • /
    • 2004
  • This study is proposed the novel gesture recognition method for the remote camera robot control. To recognize the dynamics gesture, the preprocessing step is the image segmentation. The conventional methods for the effectively object segmentation has need a lot of the cole. information about the object(hand) image. And these methods in the recognition step have need a lot of the features with the each object. To improve the problems of the conventional methods, this study proposed the novel method to recognize the dynamic hand gesture such as the MMS(Max-Min Search) method to segment the object image, MSM(Mean Space Mapping) method and COG(Conte. Of Gravity) method to extract the features of image, and the structure of recognition MLPNN(Multi Layer Perceptron Neural Network) to recognize the dynamic gestures. In the results of experiment, the recognition rate of the proposed method appeared more than 90[%], and this result is shown that is available by HCI(Human Computer Interface) device for .emote robot control.

Gaze Detection System by IR-LED based Camera (적외선 조명 카메라를 이용한 시선 위치 추적 시스템)

  • 박강령
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.29 no.4C
    • /
    • pp.494-504
    • /
    • 2004
  • The researches about gaze detection have been much developed with many applications. Most previous researches only rely on image processing algorithm, so they take much processing time and have many constraints. In our work, we implement it with a computer vision system setting a IR-LED based single camera. To detect the gaze position, we locate facial features, which is effectively performed with IR-LED based camera and SVM(Support Vector Machine). When a user gazes at a position of monitor, we can compute the 3D positions of those features based on 3D rotation and translation estimation and affine transform. Finally, the gaze position by the facial movements is computed from the normal vector of the plane determined by those computed 3D positions of features. In addition, we use a trained neural network to detect the gaze position by eye's movement. As experimental results, we can obtain the facial and eye gaze position on a monitor and the gaze position accuracy between the computed positions and the real ones is about 4.2 cm of RMS error.