• Title/Summary/Keyword: long memory process

Search Result 161, Processing Time 0.03 seconds

An Empirical Study on the Relationships among Safeguarding Mechanism, Relationship Learning, and Relationship Performance in Technology Cooperation Network by Applying Resource Based Theory (자원기반이론을 적용한 기술협력 네트워크에서 보호 메커니즘, 관계학습, 관계성과의 관계에 대한 실증연구)

  • Kang, Seok-Min
    • Management & Information Systems Review
    • /
    • v.35 no.2
    • /
    • pp.45-66
    • /
    • 2016
  • Firms can make scale of economy and scope of economy by internalizing and using new advanced technology and knowledge from technology cooperation network, decrease risk and cost with partner firm of technology cooperation network, and increase market advantage of product & strengthen firms' position in the market. Due to the advantages of technology cooperation network, the related studies have focused on the positive effect of technology cooperation network. However, the related studies investigating the relationship between technology cooperation network and firm performance have only examined the role of technology cooperation network. Safeguarding mechanism, relationship learning, and relationship performance are categorized into the process of technology cooperation network, and this categorization is applied as resources, capability, and performance by resource based view. The empirical results are reported as belows. First, relationship specific investment and relationship capital positively affect on relationship learning as capability. Second, information sharing, common information understanding, and relationship specific memory development positively affect on long-term orientation, but information sharing has no impact on efficiency and effectiveness. Third, relationship specific investment positively affects on relationship capital and efficiency and effectiveness have positive effects on long-term orientation. Applying technology cooperation network in asymmetric technology dependency with resource based theory, this study suggested the importance of both safeguarding and relationship learning by investigating the relationship among safeguarding, relationship learning, and relationship performance. And it is worthy that this study investigated how firms' behavior change affects relationship performance in the relationship of technology cooperation partner.

  • PDF

Korean Sentence Generation Using Phoneme-Level LSTM Language Model (한국어 음소 단위 LSTM 언어모델을 이용한 문장 생성)

  • Ahn, SungMahn;Chung, Yeojin;Lee, Jaejoon;Yang, Jiheon
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.2
    • /
    • pp.71-88
    • /
    • 2017
  • Language models were originally developed for speech recognition and language processing. Using a set of example sentences, a language model predicts the next word or character based on sequential input data. N-gram models have been widely used but this model cannot model the correlation between the input units efficiently since it is a probabilistic model which are based on the frequency of each unit in the training set. Recently, as the deep learning algorithm has been developed, a recurrent neural network (RNN) model and a long short-term memory (LSTM) model have been widely used for the neural language model (Ahn, 2016; Kim et al., 2016; Lee et al., 2016). These models can reflect dependency between the objects that are entered sequentially into the model (Gers and Schmidhuber, 2001; Mikolov et al., 2010; Sundermeyer et al., 2012). In order to learning the neural language model, texts need to be decomposed into words or morphemes. Since, however, a training set of sentences includes a huge number of words or morphemes in general, the size of dictionary is very large and so it increases model complexity. In addition, word-level or morpheme-level models are able to generate vocabularies only which are contained in the training set. Furthermore, with highly morphological languages such as Turkish, Hungarian, Russian, Finnish or Korean, morpheme analyzers have more chance to cause errors in decomposition process (Lankinen et al., 2016). Therefore, this paper proposes a phoneme-level language model for Korean language based on LSTM models. A phoneme such as a vowel or a consonant is the smallest unit that comprises Korean texts. We construct the language model using three or four LSTM layers. Each model was trained using Stochastic Gradient Algorithm and more advanced optimization algorithms such as Adagrad, RMSprop, Adadelta, Adam, Adamax, and Nadam. Simulation study was done with Old Testament texts using a deep learning package Keras based the Theano. After pre-processing the texts, the dataset included 74 of unique characters including vowels, consonants, and punctuation marks. Then we constructed an input vector with 20 consecutive characters and an output with a following 21st character. Finally, total 1,023,411 sets of input-output vectors were included in the dataset and we divided them into training, validation, testsets with proportion 70:15:15. All the simulation were conducted on a system equipped with an Intel Xeon CPU (16 cores) and a NVIDIA GeForce GTX 1080 GPU. We compared the loss function evaluated for the validation set, the perplexity evaluated for the test set, and the time to be taken for training each model. As a result, all the optimization algorithms but the stochastic gradient algorithm showed similar validation loss and perplexity, which are clearly superior to those of the stochastic gradient algorithm. The stochastic gradient algorithm took the longest time to be trained for both 3- and 4-LSTM models. On average, the 4-LSTM layer model took 69% longer training time than the 3-LSTM layer model. However, the validation loss and perplexity were not improved significantly or became even worse for specific conditions. On the other hand, when comparing the automatically generated sentences, the 4-LSTM layer model tended to generate the sentences which are closer to the natural language than the 3-LSTM model. Although there were slight differences in the completeness of the generated sentences between the models, the sentence generation performance was quite satisfactory in any simulation conditions: they generated only legitimate Korean letters and the use of postposition and the conjugation of verbs were almost perfect in the sense of grammar. The results of this study are expected to be widely used for the processing of Korean language in the field of language processing and speech recognition, which are the basis of artificial intelligence systems.

Meaning and Value Analysis of Records of Laos Renewable Energy Support Activities Collection (라오스 재생가능에너지 지원활동 컬렉션의 의미와 가치 연구)

  • Ju, Hyun Mi;Yim, Jin Hee
    • The Korean Journal of Archival Studies
    • /
    • no.51
    • /
    • pp.45-87
    • /
    • 2017
  • In recent years, there were some who have conducted research on deriving the social and historical meanings of records through the analysis of specific records collections. This trend is an attempt to pay attention to human actions in the society and to read the society again through the records created by such actions. In this study, I derive various meanings and values of these records through the analysis of the "Laos Renewable Energy Support Activities" collection. Moreover, I study how the collection was reconstructed by the Human and Memory Archives. The "Laos Renewable Energy Support Activities" is the personal record of the donor who led the project, and contains the process and results of the project. Through this collection, I was able to look at the life of the donor as a foreign aid activist in Laos and realized his values. Furthermore, through the business process record, I was able to discover the implications of climate change response overseas aid projects. In addition, I was able to look at the culture and environment of Laos through the eyes of the donor who has been residing there for a long time.

Water temperature prediction of Daecheong Reservoir by a process-guided deep learning model (역학적 모델과 딥러닝 모델을 융합한 대청호 수온 예측)

  • Kim, Sung Jin;Park, Hyungseok;Lee, Gun Ho;Chung, Se Woong
    • Proceedings of the Korea Water Resources Association Conference
    • /
    • 2021.06a
    • /
    • pp.88-88
    • /
    • 2021
  • 최근 수자원과 수질관리 분야에 자료기반 머신러닝 모델과 딥러닝 모델의 활용이 급증하고 있다. 그러나 딥러닝 모델은 Blackbox 모델의 특성상 고전적인 질량, 운동량, 에너지 보존법칙을 고려하지 않고, 데이터에 내재된 패턴과 관계를 해석하기 때문에 물리적 법칙을 만족하지 않는 예측결과를 가져올 수 있다. 또한, 딥러닝 모델의 예측 성능은 학습데이터의 양과 변수 선정에 크게 영향을 받는 모델이기 때문에 양질의 데이터가 제공되지 않으면 모델의 bias와 variation이 클 수 있으며 정확도 높은 예측이 어렵다. 최근 이러한 자료기반 모델링 방법의 단점을 보완하기 위해 프로세스 기반 수치모델과 딥러닝 모델을 결합하여 두 모델링 방법의 장점을 활용하는 연구가 활발히 진행되고 있다(Read et al., 2019). Process-Guided Deep Learning (PGDL) 방법은 물리적 법칙을 반영하여 딥러닝 모델을 훈련시킴으로써 순수한 딥러닝 모델의 물리적 법칙 결여성 문제를 해결할 수 있는 대안으로 활용되고 있다. PGDL 모델은 딥러닝 모델에 물리적인 법칙을 해석할 수 있는 추가변수를 도입하며, 딥러닝 모델의 매개변수 최적화 과정에서 Cost 함수에 물리적 법칙을 위반하는 경우 Penalty를 추가하는 알고리즘을 도입하여 물리적 보존법칙을 만족하도록 모델을 훈련시킨다. 본 연구의 목적은 대청호의 수심별 수온을 예측하기 위해 역학적 모델과 딥러닝 모델을 융합한 PGDL 모델을 개발하고 적용성을 평가하는데 있다. 역학적 모델은 2차원 횡방향 평균 수리·수질 모델인 CE-QUAL-W2을 사용하였으며, 대청호를 대상으로 2017년부터 2018년까지 총 2년간 수온과 에너지 수지를 모의하였다. 기상(기온, 이슬점온도, 풍향, 풍속, 운량), 수문(저수위, 유입·유출 유량), 수온자료를 수집하여 CE-QUAL-W2 모델을 구축하고 보정하였으며, 모델은 저수위 변화, 수온의 수심별 시계열 변동 특성을 적절하게 재현하였다. 또한, 동일기간 대청호 수심별 수온 예측을 위한 순환 신경망 모델인 LSTM(Long Short-Term Memory)을 개발하였으며, 종속변수는 수온계 체인을 통해 수집한 수심별 고빈도 수온 자료를 사용하고 독립 변수는 기온, 풍속, 상대습도, 강수량, 단파복사에너지, 장파복사에너지를 사용하였다. LSTM 모델의 매개변수 최적화는 지도학습을 통해 예측값과 실측값의 RMSE가 최소화 되로록 훈련하였다. PGDL 모델은 동일 기간 LSTM 모델과 동일 입력 자료를 사용하여 구축하였으며, 역학적 모델에서 얻은 에너지 수지를 만족하지 않는 경우 Cost Function에 Penalty를 추가하여 물리적 보존법칙을 만족하도록 훈련하고 수심별 수온 예측결과를 비교·분석하였다.

  • PDF

What Changed and Unchanged After Science Class: Analyzing High School Student's Conceptual Change on Circular Motion Based on Mental Model Theory (과학수업 후 변하는 것과 변하지 않는 것: 정신모형 이론을 중심으로 한 고등학생의 원운동 개념변화 사례 분석)

  • Park, Ji-Yeon;Lee, Gyoung-Ho;Shin, Jong-Ho;Song, Sang-Ho
    • Journal of The Korean Association For Science Education
    • /
    • v.26 no.4
    • /
    • pp.475-491
    • /
    • 2006
  • In physics education, the research on students' conceptions has developed in the discussion on the nature and the difficulty of conceptual change. Recently, mental models have been a theoretical background in concrete arguments on "how students' conceptions are constructed or created." Mental models that integrate information in the presented problem and individual knowledge in their long-term memory have important information about not only expressed ideas but also in the thinking process behind the expressed ideas. The purpose of this study is to investigate the forming process and the characteristics of high school student's mental models about circular motion, and how they were changed by instruction. We used the think-aloud method based on the instrument for identifying student's mental models about circular motion, pretest of physics concept, mind map and interview for investigating student's characteristics. The results of the study showed that instructions based on the mental model theory facilitated scientific expressed model, but several factors that affected forming mental models like epistemological belief didn't change scientifically after 3 lessons.

Deep Learning Architectures and Applications (딥러닝의 모형과 응용사례)

  • Ahn, SungMahn
    • Journal of Intelligence and Information Systems
    • /
    • v.22 no.2
    • /
    • pp.127-142
    • /
    • 2016
  • Deep learning model is a kind of neural networks that allows multiple hidden layers. There are various deep learning architectures such as convolutional neural networks, deep belief networks and recurrent neural networks. Those have been applied to fields like computer vision, automatic speech recognition, natural language processing, audio recognition and bioinformatics where they have been shown to produce state-of-the-art results on various tasks. Among those architectures, convolutional neural networks and recurrent neural networks are classified as the supervised learning model. And in recent years, those supervised learning models have gained more popularity than unsupervised learning models such as deep belief networks, because supervised learning models have shown fashionable applications in such fields mentioned above. Deep learning models can be trained with backpropagation algorithm. Backpropagation is an abbreviation for "backward propagation of errors" and a common method of training artificial neural networks used in conjunction with an optimization method such as gradient descent. The method calculates the gradient of an error function with respect to all the weights in the network. The gradient is fed to the optimization method which in turn uses it to update the weights, in an attempt to minimize the error function. Convolutional neural networks use a special architecture which is particularly well-adapted to classify images. Using this architecture makes convolutional networks fast to train. This, in turn, helps us train deep, muti-layer networks, which are very good at classifying images. These days, deep convolutional networks are used in most neural networks for image recognition. Convolutional neural networks use three basic ideas: local receptive fields, shared weights, and pooling. By local receptive fields, we mean that each neuron in the first(or any) hidden layer will be connected to a small region of the input(or previous layer's) neurons. Shared weights mean that we're going to use the same weights and bias for each of the local receptive field. This means that all the neurons in the hidden layer detect exactly the same feature, just at different locations in the input image. In addition to the convolutional layers just described, convolutional neural networks also contain pooling layers. Pooling layers are usually used immediately after convolutional layers. What the pooling layers do is to simplify the information in the output from the convolutional layer. Recent convolutional network architectures have 10 to 20 hidden layers and billions of connections between units. Training deep learning networks has taken weeks several years ago, but thanks to progress in GPU and algorithm enhancement, training time has reduced to several hours. Neural networks with time-varying behavior are known as recurrent neural networks or RNNs. A recurrent neural network is a class of artificial neural network where connections between units form a directed cycle. This creates an internal state of the network which allows it to exhibit dynamic temporal behavior. Unlike feedforward neural networks, RNNs can use their internal memory to process arbitrary sequences of inputs. Early RNN models turned out to be very difficult to train, harder even than deep feedforward networks. The reason is the unstable gradient problem such as vanishing gradient and exploding gradient. The gradient can get smaller and smaller as it is propagated back through layers. This makes learning in early layers extremely slow. The problem actually gets worse in RNNs, since gradients aren't just propagated backward through layers, they're propagated backward through time. If the network runs for a long time, that can make the gradient extremely unstable and hard to learn from. It has been possible to incorporate an idea known as long short-term memory units (LSTMs) into RNNs. LSTMs make it much easier to get good results when training RNNs, and many recent papers make use of LSTMs or related ideas.

Automatic gasometer reading system using selective optical character recognition (관심 문자열 인식 기술을 이용한 가스계량기 자동 검침 시스템)

  • Lee, Kyohyuk;Kim, Taeyeon;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.2
    • /
    • pp.1-25
    • /
    • 2020
  • In this paper, we suggest an application system architecture which provides accurate, fast and efficient automatic gasometer reading function. The system captures gasometer image using mobile device camera, transmits the image to a cloud server on top of private LTE network, and analyzes the image to extract character information of device ID and gas usage amount by selective optical character recognition based on deep learning technology. In general, there are many types of character in an image and optical character recognition technology extracts all character information in an image. But some applications need to ignore non-of-interest types of character and only have to focus on some specific types of characters. For an example of the application, automatic gasometer reading system only need to extract device ID and gas usage amount character information from gasometer images to send bill to users. Non-of-interest character strings, such as device type, manufacturer, manufacturing date, specification and etc., are not valuable information to the application. Thus, the application have to analyze point of interest region and specific types of characters to extract valuable information only. We adopted CNN (Convolutional Neural Network) based object detection and CRNN (Convolutional Recurrent Neural Network) technology for selective optical character recognition which only analyze point of interest region for selective character information extraction. We build up 3 neural networks for the application system. The first is a convolutional neural network which detects point of interest region of gas usage amount and device ID information character strings, the second is another convolutional neural network which transforms spatial information of point of interest region to spatial sequential feature vectors, and the third is bi-directional long short term memory network which converts spatial sequential information to character strings using time-series analysis mapping from feature vectors to character strings. In this research, point of interest character strings are device ID and gas usage amount. Device ID consists of 12 arabic character strings and gas usage amount consists of 4 ~ 5 arabic character strings. All system components are implemented in Amazon Web Service Cloud with Intel Zeon E5-2686 v4 CPU and NVidia TESLA V100 GPU. The system architecture adopts master-lave processing structure for efficient and fast parallel processing coping with about 700,000 requests per day. Mobile device captures gasometer image and transmits to master process in AWS cloud. Master process runs on Intel Zeon CPU and pushes reading request from mobile device to an input queue with FIFO (First In First Out) structure. Slave process consists of 3 types of deep neural networks which conduct character recognition process and runs on NVidia GPU module. Slave process is always polling the input queue to get recognition request. If there are some requests from master process in the input queue, slave process converts the image in the input queue to device ID character string, gas usage amount character string and position information of the strings, returns the information to output queue, and switch to idle mode to poll the input queue. Master process gets final information form the output queue and delivers the information to the mobile device. We used total 27,120 gasometer images for training, validation and testing of 3 types of deep neural network. 22,985 images were used for training and validation, 4,135 images were used for testing. We randomly splitted 22,985 images with 8:2 ratio for training and validation respectively for each training epoch. 4,135 test image were categorized into 5 types (Normal, noise, reflex, scale and slant). Normal data is clean image data, noise means image with noise signal, relfex means image with light reflection in gasometer region, scale means images with small object size due to long-distance capturing and slant means images which is not horizontally flat. Final character string recognition accuracies for device ID and gas usage amount of normal data are 0.960 and 0.864 respectively.

A Study of the Planning Process, Design Idea and Implementation of the Gwanghwamun Plaza (광화문광장 조성과정 및 설계 연구)

  • Shin, Hyun-Don;Zoh, Kyung-Jin
    • Journal of the Korean Institute of Landscape Architecture
    • /
    • v.41 no.4
    • /
    • pp.24-41
    • /
    • 2013
  • This study aims to investigate the complex planning processes and design ideas of the Gwanghwamun Plaza which was opened in 2009. The opening of the plaza was significant as it was built in order to restore the symbolic meaning of axis in the historical Joseon Dynasty that was distorted during the Japanese Imperialism. The plaza itself attracts many citizens and tourists by providing the empty ground carrying historical ambiences around. In this paper, the story of the Gwanghwamun plaza will be summarized. Particularly, the background of promoting the project will be discussed and the whole planning process will be dealt with. The plaza was realized through several stages. First, the planning stage will be reviewed. The planning stage had been a quite long process since the initial idea was discussed. Since the early 90s, the political decision of making the plaza was made through the change of people's understanding toward public space. At this stage, the city government worked together with diverse citizens and professionals to share the vision and to realize the right decision in making the plaza. Second, the design stage will be elaborated in detail. This is the second design stage. The former was the idea competition and the latter was the turn-key base. The final design scheme emphasized the restoration of symbolic axis and the forgotten Yukjo Street. The scheme consists of four zones such as history restoration zone, prospect and history representation zone, culture zone, and the urban zone. Through the whole design process, the original idea remained as it was. The design concept was "a place of memory and prospect." It emphasizes the history representation, view corridor, cultural activities platform as well as the emptiness and flexibility of the basic premise of the plaza. Finally, the construction stage will be discussed. There were some additions and omissions in the construction process. The design chances in the construction stage will be reviewed in detail. After the opening of the plaza, there were some changes in the detail design. We will discuss how and why these modifications were made. In the end, the social and cultural implication of the plaza will be discussed. The storytelling of the Gwanghwamun Plaza will contribute to the clear understanding of planning and design process of pubic places. Based on this reflection, we are able to think about some suggestions of public projects for the future.

Proposal of a Step-by-Step Optimized Campus Power Forecast Model using CNN-LSTM Deep Learning (CNN-LSTM 딥러닝 기반 캠퍼스 전력 예측 모델 최적화 단계 제시)

  • Kim, Yein;Lee, Seeun;Kwon, Youngsung
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.21 no.10
    • /
    • pp.8-15
    • /
    • 2020
  • A forecasting method using deep learning does not have consistent results due to the differences in the characteristics of the dataset, even though they have the same forecasting models and parameters. For example, the forecasting model X optimized with dataset A would not produce the optimized result with another dataset B. The forecasting model with the characteristics of the dataset needs to be optimized to increase the accuracy of the forecasting model. Therefore, this paper proposes novel optimization steps for outlier removal, dataset classification, and a CNN-LSTM-based hyperparameter tuning process to forecast the daily power usage of a university campus based on the hourly interval. The proposing model produces high forecasting accuracy with a 2% of MAPE with a single power input variable. The proposing model can be used in EMS to suggest improved strategies to users and consequently to improve the power efficiency.

An Analysis on the Asymmetric Time Varying Spillover Effect between Capesize and Panamax Markets (케이프사이즈와 파나막스 시장간의 비대칭 시간가변 파급효과에 관한 분석)

  • Chung, Sang-Kuck
    • Journal of Korea Port Economic Association
    • /
    • v.27 no.3
    • /
    • pp.41-64
    • /
    • 2011
  • This article investigates the interrelationships in daily returns using fractionally integrated error correction term and volatilities using constant conditional correlation and dynamic conditional correlation GARCH with asymmetries between Capesize and Panamax markets. Our findings are as follows. First, for the fractionally cointegrated error correction model, there is a unidirectional relationship in returns from the Panamax market to the Capesize market, but a bidirectional causal relationship prevails for the traditional error correction models. Second, the coefficients for the error correction term are all statistically significant. Of particular interest are the signs of the estimates for the error correction term, which are all negative for the Capesize return equation and all positive for the Panamax return. Third, there are bidirectional volatility spillovers between both markets and the direction of the information flow seems to be stronger from Panamax to Capesize. Fourth, the coefficients for the asymmetric term are all significantly positive in the Capesize market, but the Panamax market does not have a significant effect. However, the coefficients for the asymmetric term are all significant, implying that the leverage effect does exist in the Capesize and Panamax markets.