• Title/Summary/Keyword: Data & Knowledge Engineering

Search Result 1,311, Processing Time 0.035 seconds

A Study on the Automatic Detection of Railroad Power Lines Using LiDAR Data and RANSAC Algorithm (LiDAR 데이터와 RANSAC 알고리즘을 이용한 철도 전력선 자동탐지에 관한 연구)

  • Jeon, Wang Gyu;Choi, Byoung Gil
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.31 no.4
    • /
    • pp.331-339
    • /
    • 2013
  • LiDAR has been one of the widely used and important technologies for 3D modeling of ground surface and objects because of its ability to provide dense and accurate range measurement. The objective of this research is to develop a method for automatic detection and modeling of railroad power lines using high density LiDAR data and RANSAC algorithms. For detecting railroad power lines, multi-echoes properties of laser data and shape knowledge of railroad power lines were employed. Cuboid analysis for detecting seed line segments, tracking lines, connecting and labeling are the main processes. For modeling railroad power lines, iterative RANSAC and least square adjustment were carried out to estimate the lines parameters. The validation of the result is very challenging due to the difficulties in determining the actual references on the ground surface. Standard deviations of 8cm and 5cm for x-y and z coordinates, respectively are satisfactory outcomes. In case of completeness, the result of visual inspection shows that all the lines are detected and modeled well as compare with the original point clouds. The overall processes are fully automated and the methods manage any state of railroad wires efficiently.

GIS-based Spatial Zonations for Regional Estimation of Site-specific Seismic Response in Seoul Metropolis (대도시 서울에서의 부지고유 지진 응답의 지역적 예측을 위한 GIS 기반의 공간 구역화)

  • Sun, Chang-Guk;Chun, Sung-Ho;Chung, Choong-Ki
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.30 no.1C
    • /
    • pp.65-76
    • /
    • 2010
  • Recent earthquake events revealed that severe seismic damages were concentrated mostly at sites composed of soil sediments rather than firm rock. This indicates that the site effects inducing the amplification of earthquake ground motion are associated mainly with the spatial distribution and dynamic properties of the soils overlying bedrock. In this study, an integrated GIS-based information system for geotechnical data was constructed to establish a regional counterplan against ground motions at a representative metropolitan area, Seoul, in Korea. To implement the GIS-based geotechnical information system for the Seoul area, existing geotechnical investigation data were collected in and around the study area and additionally a walkover site survey was carried out to acquire surface geo-knowledge data. For practical application of the geotechnical information system used to estimate the site effects at the area of interest, seismic zoning maps of geotechnical earthquake engineering parameters, such as the depth to bedrock and the site period, were created and presented as regional synthetic strategy for earthquake-induced hazards prediction. In addition, seismic zonation of site classification was also performed to determine the site amplification coefficients for seismic design at any site and administrative sub-unit in the Seoul area. Based on the case study on seismic zonations for Seoul, it was verified that the GIS-based geotechnical information system was very useful for the regional prediction of seismic hazards and also the decision support for seismic hazard mitigation particularly at the metropolitan area.

Development of Korean Tissue Probability Map from 3D Magnetic Resonance Images (3차원 MR 영상으로부터의 한국인 뇌조직확률지도 개발)

  • Jung Hyun, Kim;Jong-Min, Lee;Uicheul, Yoon;Hyun-Pil, Kim;Bang Bon, Koo;In Young, Kim;Dong Soo, Lee;Jun Soo, Kwon;Sun I., Kim
    • Journal of Biomedical Engineering Research
    • /
    • v.25 no.5
    • /
    • pp.323-328
    • /
    • 2004
  • The development of group-specific tissue probability maps (TPM) provides a priori knowledge for better result of cerebral tissue classification with regard to the inter-ethnic differences of inter-subject variability. We present sequential procedures of group-specific TPM and evaluate the age effects in the structural differences of TPM. We investigated 100 healthy volunteers with high resolution MRI scalming. The subjects were classified into young (60, 25.92+4.58) and old groups (40, 58.83${\pm}$8.10) according to the age. To avoid any bias from random selected single subject and improve registration robustness, average atlas as target for TPM was constructed from skull-stripped whole data using linear and nonlinear registration of AIR. Each subject was segmented into binary images of gray matter, white matter, and cerebrospinal fluid using fuzzy clustering and normalized into the space of average atlas. The probability images were the means of these binary images, and contained values in the range of zero to one. A TPM of a given tissue is a spatial probability distribution representing a certain subject population. In the spatial distribution of tissue probability according to the threshold of probability, the old group exhibited enlarged ventricles and overall GM atrophy as age-specific changes, compared to the young group. Our results are generally consistent with the few published studies on age differences in the brain morphology. The more similar the morphology of the subject is to the average of the population represented by the TPM, the better the entire classification procedure should work. Therefore, we suggest that group-specific TPM should be used as a priori information for the cerebral tissue classification.

Study on Recognitions of Department of Technology Education Professors Introducing 'Creative Engineering Design' Unit of Highschool Technology.Home Economics (고등학교 기술.가정 교과 '창의공학설계' 단원 신설에 따른 기술교육과 교수의 인식 분석)

  • Kim, Seong-Il
    • 대한공업교육학회지
    • /
    • v.39 no.1
    • /
    • pp.128-142
    • /
    • 2014
  • The purpose of this study was to provide a basis information for the education direction of 'creative engineering design' lesson to highschool students and pre-technology teachers by analyzing pre-experienced the d epartment of technology professors' recognitions, according to the newly opened unit 'creative engineering design' in 2009 revised highschool technology home economics education curriculum. The survey questionnaire was composed of 34 questions and collected from 16 department of technology education professors. The data was analyzed by SPSS program. The results of study were as follows : First, in 'creative engineering design' educational objectives, the highest average level(M) of response was problem solving ability, the opportunity of engineering experience and creativity improvement were followed. In the response about 'creative engineering design' educational contents, the average level(M) of the creative thinking method education was the highest(3.94). and the following important level was the write training education for presentation. Second, in the 'creative engineering design' educational areas, the professors preferred the idea conception, design, and production manufacturing area. In the teaching learning method, the average level(M) of design-based learning method was the first. and PBL, problem solving method were favored Third, in the 'creative engineering design' lesson manage, the secure of practice room and material cost, and the check of the progress situations were needed preferentially. In the education assessments, various and combined assessment preferred such as production, portfolio, and presentation. Fourth, in the pre-technology teachers' lesson of the university course, the product manufacturing level was preferred production to improve uncomfortable thing in life. The major difficulty in product manufacturing was the lack of major knowledge and product conception.

A Big Data Based Random Motif Frequency Method for Analyzing Human Proteins (인간 단백질 분석을 위한 빅 데이타 기반 RMF 방법)

  • Kim, Eun-Mi;Jeong, Jong-Cheol;Lee, Bae-Ho
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.13 no.6
    • /
    • pp.1397-1404
    • /
    • 2018
  • Due to the technical difficulties and high cost for obtaining 3-dimensional structure data, sequence-based approaches in proteins have not been widely acknowledged. A motif can be defined as any segments in protein or gene sequences. With this simplicity, motifs have been actively and widely used in various areas. However, the motif itself has not been studied comprehensively. The value of this study can be categorized in three fields in order to analyze the human proteins using artificial intelligence method: (1) Based on our best knowledge, this research is the first comprehensive motif analysis by analyzing motifs with all human proteins in Protein Data Bank (PDB) associated with the database of Enzyme Commission (EC) number and Structural Classification of Proteins (SCOP). (2) We deeply analyze the motif in three different categories: pattern, statistical, and functional analysis of clusters. (3) At the last and most importantly, we proposed random motif frequency(RMF) matric that can efficiently distinct the characteristics of proteins by identifying interface residues from non-interface residues and clustering protein functions based on big data while varying the size of random motif.

Frequent Origin-Destination Sequence Pattern Analysis from Taxi Trajectories (택시 기종점 빈번 순차 패턴 분석)

  • Lee, Tae Young;Jeon, Seung Bae;Jeong, Myeong Hun;Choi, Yun Woong
    • KSCE Journal of Civil and Environmental Engineering Research
    • /
    • v.39 no.3
    • /
    • pp.461-467
    • /
    • 2019
  • Advances in location-aware and IoT (Internet of Things) technology increase the rapid generation of massive movement data. Knowledge discovery from massive movement data helps us to understand the urban flow and traffic management. This paper proposes a method to analyze frequent origin-destination sequence patterns from irregular spatiotemporal taxi pick-up locations. The proposed method starts by conducting cluster analysis and then run a frequent sequence pattern analysis based on identified clusters as a base unit. The experimental data is Seoul taxi trajectory data between 7 a.m. and 9 a.m. during one week. The experimental results present that significant frequent sequence patterns occur within Gangnam. The significant frequent sequence patterns of different regions are identified between Gangnam and Seoul City Hall area. Further, this study uses administrative boundaries as a base unit. The results based on administrative boundaries fails to detect the frequent sequence patterns between different regions. The proposed method can be applied to decrease not only taxis' empty-loaded rate, but also improve urban flow management.

Comparison of Association Rule Learning and Subgroup Discovery for Mining Traffic Accident Data (교통사고 데이터의 마이닝을 위한 연관규칙 학습기법과 서브그룹 발견기법의 비교)

  • Kim, Jeongmin;Ryu, Kwang Ryel
    • Journal of Intelligence and Information Systems
    • /
    • v.21 no.4
    • /
    • pp.1-16
    • /
    • 2015
  • Traffic accident is one of the major cause of death worldwide for the last several decades. According to the statistics of world health organization, approximately 1.24 million deaths occurred on the world's roads in 2010. In order to reduce future traffic accident, multipronged approaches have been adopted including traffic regulations, injury-reducing technologies, driving training program and so on. Records on traffic accidents are generated and maintained for this purpose. To make these records meaningful and effective, it is necessary to analyze relationship between traffic accident and related factors including vehicle design, road design, weather, driver behavior etc. Insight derived from these analysis can be used for accident prevention approaches. Traffic accident data mining is an activity to find useful knowledges about such relationship that is not well-known and user may interested in it. Many studies about mining accident data have been reported over the past two decades. Most of studies mainly focused on predict risk of accident using accident related factors. Supervised learning methods like decision tree, logistic regression, k-nearest neighbor, neural network are used for these prediction. However, derived prediction model from these algorithms are too complex to understand for human itself because the main purpose of these algorithms are prediction, not explanation of the data. Some of studies use unsupervised clustering algorithm to dividing the data into several groups, but derived group itself is still not easy to understand for human, so it is necessary to do some additional analytic works. Rule based learning methods are adequate when we want to derive comprehensive form of knowledge about the target domain. It derives a set of if-then rules that represent relationship between the target feature with other features. Rules are fairly easy for human to understand its meaning therefore it can help provide insight and comprehensible results for human. Association rule learning methods and subgroup discovery methods are representing rule based learning methods for descriptive task. These two algorithms have been used in a wide range of area from transaction analysis, accident data analysis, detection of statistically significant patient risk groups, discovering key person in social communities and so on. We use both the association rule learning method and the subgroup discovery method to discover useful patterns from a traffic accident dataset consisting of many features including profile of driver, location of accident, types of accident, information of vehicle, violation of regulation and so on. The association rule learning method, which is one of the unsupervised learning methods, searches for frequent item sets from the data and translates them into rules. In contrast, the subgroup discovery method is a kind of supervised learning method that discovers rules of user specified concepts satisfying certain degree of generality and unusualness. Depending on what aspect of the data we are focusing our attention to, we may combine different multiple relevant features of interest to make a synthetic target feature, and give it to the rule learning algorithms. After a set of rules is derived, some postprocessing steps are taken to make the ruleset more compact and easier to understand by removing some uninteresting or redundant rules. We conducted a set of experiments of mining our traffic accident data in both unsupervised mode and supervised mode for comparison of these rule based learning algorithms. Experiments with the traffic accident data reveals that the association rule learning, in its pure unsupervised mode, can discover some hidden relationship among the features. Under supervised learning setting with combinatorial target feature, however, the subgroup discovery method finds good rules much more easily than the association rule learning method that requires a lot of efforts to tune the parameters.

Change Acceptable In-Depth Searching in LOD Cloud for Efficient Knowledge Expansion (효과적인 지식확장을 위한 LOD 클라우드에서의 변화수용적 심층검색)

  • Kim, Kwangmin;Sohn, Yonglak
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.2
    • /
    • pp.171-193
    • /
    • 2018
  • LOD(Linked Open Data) cloud is a practical implementation of semantic web. We suggested a new method that provides identity links conveniently in LOD cloud. It also allows changes in LOD to be reflected to searching results without any omissions. LOD provides detail descriptions of entities to public in RDF triple form. RDF triple is composed of subject, predicates, and objects and presents detail description for an entity. Links in LOD cloud, named identity links, are realized by asserting entities of different RDF triples to be identical. Currently, the identity link is provided with creating a link triple explicitly in which associates its subject and object with source and target entities. Link triples are appended to LOD. With identity links, a knowledge achieves from an LOD can be expanded with different knowledge from different LODs. The goal of LOD cloud is providing opportunity of knowledge expansion to users. Appending link triples to LOD, however, has serious difficulties in discovering identity links between entities one by one notwithstanding the enormous scale of LOD. Newly added entities cannot be reflected to searching results until identity links heading for them are serialized and published to LOD cloud. Instead of creating enormous identity links, we propose LOD to prepare its own link policy. The link policy specifies a set of target LODs to link and constraints necessary to discover identity links to entities on target LODs. On searching, it becomes possible to access newly added entities and reflect them to searching results without any omissions by referencing the link policies. Link policy specifies a set of predicate pairs for discovering identity between associated entities in source and target LODs. For the link policy specification, we have suggested a set of vocabularies that conform to RDFS and OWL. Identity between entities is evaluated in accordance with a similarity of the source and the target entities' objects which have been associated with the predicates' pair in the link policy. We implemented a system "Change Acceptable In-Depth Searching System(CAIDS)". With CAIDS, user's searching request starts from depth_0 LOD, i.e. surface searching. Referencing the link policies of LODs, CAIDS proceeds in-depth searching, next LODs of next depths. To supplement identity links derived from the link policies, CAIDS uses explicit link triples as well. Following the identity links, CAIDS's in-depth searching progresses. Content of an entity obtained from depth_0 LOD expands with the contents of entities of other LODs which have been discovered to be identical to depth_0 LOD entity. Expanding content of depth_0 LOD entity without user's cognition of such other LODs is the implementation of knowledge expansion. It is the goal of LOD cloud. The more identity links in LOD cloud, the wider content expansions in LOD cloud. We have suggested a new way to create identity links abundantly and supply them to LOD cloud. Experiments on CAIDS performed against DBpedia LODs of Korea, France, Italy, Spain, and Portugal. They present that CAIDS provides appropriate expansion ratio and inclusion ratio as long as degree of similarity between source and target objects is 0.8 ~ 0.9. Expansion ratio, for each depth, depicts the ratio of the entities discovered at the depth to the entities of depth_0 LOD. For each depth, inclusion ratio illustrates the ratio of the entities discovered only with explicit links to the entities discovered only with link policies. In cases of similarity degrees with under 0.8, expansion becomes excessive and thus contents become distorted. Similarity degree of 0.8 ~ 0.9 provides appropriate amount of RDF triples searched as well. Experiments have evaluated confidence degree of contents which have been expanded in accordance with in-depth searching. Confidence degree of content is directly coupled with identity ratio of an entity, which means the degree of identity to the entity of depth_0 LOD. Identity ratio of an entity is obtained by multiplying source LOD's confidence and source entity's identity ratio. By tracing the identity links in advance, LOD's confidence is evaluated in accordance with the amount of identity links incoming to the entities in the LOD. While evaluating the identity ratio, concept of identity agreement, which means that multiple identity links head to a common entity, has been considered. With the identity agreement concept, experimental results present that identity ratio decreases as depth deepens, but rebounds as the depth deepens more. For each entity, as the number of identity links increases, identity ratio rebounds early and reaches at 1 finally. We found out that more than 8 identity links for each entity would lead users to give their confidence to the contents expanded. Link policy based in-depth searching method, we proposed, is expected to contribute to abundant identity links provisions to LOD cloud.

Improving Bidirectional LSTM-CRF model Of Sequence Tagging by using Ontology knowledge based feature (온톨로지 지식 기반 특성치를 활용한 Bidirectional LSTM-CRF 모델의 시퀀스 태깅 성능 향상에 관한 연구)

  • Jin, Seunghee;Jang, Heewon;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.1
    • /
    • pp.253-266
    • /
    • 2018
  • This paper proposes a methodology applying sequence tagging methodology to improve the performance of NER(Named Entity Recognition) used in QA system. In order to retrieve the correct answers stored in the database, it is necessary to switch the user's query into a language of the database such as SQL(Structured Query Language). Then, the computer can recognize the language of the user. This is the process of identifying the class or data name contained in the database. The method of retrieving the words contained in the query in the existing database and recognizing the object does not identify the homophone and the word phrases because it does not consider the context of the user's query. If there are multiple search results, all of them are returned as a result, so there can be many interpretations on the query and the time complexity for the calculation becomes large. To overcome these, this study aims to solve this problem by reflecting the contextual meaning of the query using Bidirectional LSTM-CRF. Also we tried to solve the disadvantages of the neural network model which can't identify the untrained words by using ontology knowledge based feature. Experiments were conducted on the ontology knowledge base of music domain and the performance was evaluated. In order to accurately evaluate the performance of the L-Bidirectional LSTM-CRF proposed in this study, we experimented with converting the words included in the learned query into untrained words in order to test whether the words were included in the database but correctly identified the untrained words. As a result, it was possible to recognize objects considering the context and can recognize the untrained words without re-training the L-Bidirectional LSTM-CRF mode, and it is confirmed that the performance of the object recognition as a whole is improved.

Characterization of Water Quality and the Aerobic Bacterial Population in Leachate Derived from Animal Carcass Disposal (가축 매몰지 침출수에 대한 수질 특성 및 호기성 미생물 분포에 관한 연구)

  • Choi, Nag-Choul;Choi, Eun-Ju;Kim, Bong-Ju;Park, Jeong-Ann;Kim, Song-Bae;Park, Cheon-Young
    • The Journal of Engineering Geology
    • /
    • v.23 no.1
    • /
    • pp.37-46
    • /
    • 2013
  • Leakage of leachate from animal carcass disposal is a significant issue because disease can easily spread to humans and other livestock. In this study, we analyzed the physicochemical properties of leachate and tested for the presence of aerobic bacteria in leachate using molecular biology methods, for 16 animal carcass disposals in the first stage (after burial for 5 months). Leachate physicochemical analysis revealed higher total coliforms, TOC, $NH^{4+}$, and $NO^{3-}$ concentrations compared with previously published data. In most leachate samples, the concentrations of $NH^{4+}$ and $NO^{3-}$ exceeded the Korean guideline values for drinking water. In 16S rRNA sequence analysis of the distribution of leachate under aerobic conditions, Bacillus pumilus, Lysinibacillus sphaericus, and B. sphaericus were observed with high frequency, whereas no food-poisoning-related bacteria such as B. cereus or Salmonella were detected. The present findings improve our knowledge of the transport of leachate from animal carcass disposal sites through geologic media, and are useful in risk analysis and for subsequent studies.