• Title/Summary/Keyword: convert

Search Result 2,467, Processing Time 0.033 seconds

A Study for Planning Optimal Location of Solar Photovoltaic Facilities using GIS (GIS를 이용한 태양광시설 설치를 위한 적정지역 선정에 관한 연구)

  • Yun, Sung-Wook;Paek, Yee;Jang, Jae-Kyung;Choi, Duk-Kyu;Kang, Donghyeon;Son, Jinkwan;Park, Min-Jung;Kang, Suk-Won;Gwon, Jin-Kyung
    • Journal of Bio-Environment Control
    • /
    • v.28 no.3
    • /
    • pp.243-254
    • /
    • 2019
  • With the recent accelerated policy-making and interests in new renewable energy, plans to develop and supply the new renewable energy have been devised across multiple regions in Korea. Solar energy, in particular, is being applied to small-scale power supply in provincial areas, as solar cells are used to convert solar energy into electric energy to produce electric power. Nonetheless, in the case of solar power plants, the need for a large stretch of land and considerable sum of financial support implies that the planning step should take into consideration the most suitable meteorological and geographical factors. In this study, the proxy variables of meteorological and geographical factors associated with solar energy were considered in analyzing the vulnerable areas regarding the photovoltaic power generation facility across the nation. GIS was used in the spatial analysis to develop a map for assessing the optimal location for photovoltaic power generation facility. The final vulnerability map developed in this study did not reveal any areas that exhibit vulnerability level 5 (very high) or 1 (very low). Jeollanam-do showed the largest value of vulnerability level 4 (high), while a large value of vulnerability level 3 (moderate) was shown by several administrative districts including Gwangju metropolitan city, Jeollabuk-do, Chungcheongbuk-do, and Gangwon-do. A value of vulnerability level 2 (low) was shown by the metropolitan cities including Daegu, Ulsan, and Incheon. When the 30 currently operating solar power plants were compared and reviewed, most were found to be in an area of vulnerability level 2 or 3, indicating that the locations were relatively suitable for solar energy. However, the limited data quantity for solar power plants, which is the limitation of this study, prevents the accuracy of the findings to be clearly established. Nevertheless, the significance of this study lies in that an attempt has been made to assess the vulnerability map for photovoltaic power generation facility targeting various regions across the nation, through the use of the GIS-based spatial analysis technique that takes into account the diverse meteorological and geographical factors. Furthermore, by presenting the data obtained for all regions across the nation, the findings of this study are likely to prove useful as the basic data in fields related to the photovoltaic power generation.

Status and Preservation of Cultural Relics in the Demilitarized Zone (비무장지대(DMZ) 문화유적 현황과 보전방안)

  • Lee, Jae
    • Korean Journal of Heritage: History & Science
    • /
    • v.52 no.1
    • /
    • pp.216-241
    • /
    • 2019
  • There are 35 cultural properties of fourteen kinds in the Demilitarized Zone known so far, but this number is expected to increase in the future. Among them, Cheolwon-Doseong and Jeongol-Chong of Gimhwa should be the first step toward conservation efforts by conducting a joint investigation through the collaboration of North and South Korea. In particular, the joint investigation of Cheolwon-Doseong will not only remind the North and South that they are the same people who have had common history and cultural traditions for a long time, but will also give symbolic meaning to convert the demilitarized zone into a stage for peace. Since Jeongol-Chong is a mass grave of the fallen soldiers of Pyeongan Province who fought against the invasion of the Qing of China, it should be managed as a national designated cultural asset through joint investigation. In addition, the Demilitarized Zone should become a World Heritage Site because of its importance to the legacy of the Korean War, an international war caused by an ideological confrontation. Furthermore, it has more than 6,000 kinds of temperate forests in addition to 100 species of endangered species and natural monuments. The DMZ is very qualified to be a World Natural Heritage Site, and should be included as a World Complex Cultural Heritage Site that qualifies as a World Heritage and World Natural Heritage Site. In the Demilitarized Zone, we can also find numerous highlands, tunnels and posts used during the Korean War, as well as surveillance posts, a military demarcation line, barbed wire fences, and Panmunjom, which were created by the armistice agreement. it would be desirable to select some of its sections and war facilities and to register them as modern cultural heritage assets. Finally, it is necessary to reconstruct the Dorasan Signal Fire Site, which was the communication facility of a traditional era which connected the South (Dorasan) and North (Gaesong). This would symbolize smooth communication between the two Koreas. In order to prepare for the reckless development of the Demilitarized Zone due to the upcoming cease-fire, the government and cultural asset experts will have to work hard to identify and preserve the cultural properties of the Demilitarized Zone, and they will also have to maintain consistent control over matters such as indiscriminate investigation and mine clearance.

Development of Carbon Dioxide Emission Factor from Resource Recovery Facility (폐기물자원회수시설의 이산화탄소 배출계수 개발)

  • Kim, Seungjin;Im, Gikyo;Yi, Chi-Yeong;Lee, Seehyung;Sa, Jae-Hwan;Jeon, Eui-Chan
    • Journal of Climate Change Research
    • /
    • v.4 no.1
    • /
    • pp.51-61
    • /
    • 2013
  • To address the problems associated with climate change and energy shortage, Korea has been making efforts to turn waste materials into usable energy. Due to the ongoing efforts to convert waste materials into energy, waste incineration is expanding to utilize the heat generated, and the subsequent greenhouse gas emissions from these waste material incineration are expected to increase. In this study, a municipal waste incineration plant that generates heat and electricity through heat recovery was selected as a subject facility. Methods for estimating the greenhouse gas emissions in the municipal waste incineration plant that was selected as a subject plant were sought, and the greenhouse gas emissions and emission factor were estimated. The $CO_2$ concentrations in discharge gas from the subject facility were on average 6.99%, and the result from calculating this into greenhouse gas emissions showed that the total amount of emissions was $254.60ton\;CO_2/day$. The net emissions, excluding the amount of greenhouse gas emitted from biomass incineration, was shown to be $110.59ton\;CO_2/day$. In addition, after estimating the emissions by separating the heat and electricity generated in the incineration facility, greenhouse gas emission factors were calculated using the greenhouse gas emissions produced per each unit of output. The estimated emission factor for heat was found to be $0.047ton\;CO_2/GJ$ and the emission factor for electricity was found to be $0.652ton\;CO_2/MWh$. The estimated emission factor was shown to be about 17% lower than the $0.783ton\;CO_2/MWh$ emission factor for thermal power plants that use fossil fuels. Waste material types and fossil carbon contents were evaluated as being the factors that have major effects on the greenhouse gas emissions and emission factor.

Hydrolysis of Non-digestible Components of Soybean Meal by α-Galactosidase from Bacillus coagulans NRR1207 (Bacillus coagulans NRR1207이 생산하는 α-galactosidase에 의한 대두박 비소화성분의 가수분해)

  • Ra, Seok Han;Renchinkhand, Gereltuya;Park, Min-gil;Kim, Woan-sub;Paik, Seung-Hee;Nam, Myoung Soo
    • Journal of Life Science
    • /
    • v.28 no.11
    • /
    • pp.1347-1353
    • /
    • 2018
  • The fermentation of non-digestible soy meal can convert polysaccharides into many compounds that have a wide variety of biological functions. Bacillus strains are capable of hydrolyzing non-digestible saccharides, such as melibiose, raffinose, and stachyose, found in soy meal components. A highly active ${\alpha}$-galactosidase (${\alpha}$-d-galactoside galactohydrolase, EC 3.2.1.22) was isolated from a bacterium in a traditional Korean fermented medicinal herb preparation. The isolate, T2-16, was identified as Bacillus coagulans based on its 16S rRNA sequence and biochemical properties, and the strain was named Bacillus coagulans NRR-1207. When incubated in 10%(w/v) skim milk, Bacillus coagulans NRR1207 caused a decrease in the pH of the culture medium, as well as an increase in titratable acidity and viable cell counts. This strain also showed higher activities of ${\alpha}$-galactosidase, ${\beta}$-galactosidase, ${\alpha}$-glucosidase, naphthol-AS-BO-phosphohydrolase, and acid phosphatase when compared to other enzymes. It hydrolyzed oligomeric substrates, such as raffinose and stachyose, and liberated galactose, indicating that the Bacillus coagulans NRR1207 ${\alpha}$-galactosidase hydrolyzed the ${\alpha}$-1,6 glycoside linkage. These results suggest that the decreased stachyose and raffinose contents observed in fermented soy meal are due to this ${\alpha}$-galactosidase activity. Bacillus coagulans NRR1207 therefore has potential probiotic activity and could be utilized in feed manufacturing, as well as for hydrolyzing non-digestible soy meal components.

Multi-Vector Document Embedding Using Semantic Decomposition of Complex Documents (복합 문서의 의미적 분해를 통한 다중 벡터 문서 임베딩 방법론)

  • Park, Jongin;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.3
    • /
    • pp.19-41
    • /
    • 2019
  • According to the rapidly increasing demand for text data analysis, research and investment in text mining are being actively conducted not only in academia but also in various industries. Text mining is generally conducted in two steps. In the first step, the text of the collected document is tokenized and structured to convert the original document into a computer-readable form. In the second step, tasks such as document classification, clustering, and topic modeling are conducted according to the purpose of analysis. Until recently, text mining-related studies have been focused on the application of the second steps, such as document classification, clustering, and topic modeling. However, with the discovery that the text structuring process substantially influences the quality of the analysis results, various embedding methods have actively been studied to improve the quality of analysis results by preserving the meaning of words and documents in the process of representing text data as vectors. Unlike structured data, which can be directly applied to a variety of operations and traditional analysis techniques, Unstructured text should be preceded by a structuring task that transforms the original document into a form that the computer can understand before analysis. It is called "Embedding" that arbitrary objects are mapped to a specific dimension space while maintaining algebraic properties for structuring the text data. Recently, attempts have been made to embed not only words but also sentences, paragraphs, and entire documents in various aspects. Particularly, with the demand for analysis of document embedding increases rapidly, many algorithms have been developed to support it. Among them, doc2Vec which extends word2Vec and embeds each document into one vector is most widely used. However, the traditional document embedding method represented by doc2Vec generates a vector for each document using the whole corpus included in the document. This causes a limit that the document vector is affected by not only core words but also miscellaneous words. Additionally, the traditional document embedding schemes usually map each document into a single corresponding vector. Therefore, it is difficult to represent a complex document with multiple subjects into a single vector accurately using the traditional approach. In this paper, we propose a new multi-vector document embedding method to overcome these limitations of the traditional document embedding methods. This study targets documents that explicitly separate body content and keywords. In the case of a document without keywords, this method can be applied after extract keywords through various analysis methods. However, since this is not the core subject of the proposed method, we introduce the process of applying the proposed method to documents that predefine keywords in the text. The proposed method consists of (1) Parsing, (2) Word Embedding, (3) Keyword Vector Extraction, (4) Keyword Clustering, and (5) Multiple-Vector Generation. The specific process is as follows. all text in a document is tokenized and each token is represented as a vector having N-dimensional real value through word embedding. After that, to overcome the limitations of the traditional document embedding method that is affected by not only the core word but also the miscellaneous words, vectors corresponding to the keywords of each document are extracted and make up sets of keyword vector for each document. Next, clustering is conducted on a set of keywords for each document to identify multiple subjects included in the document. Finally, a Multi-vector is generated from vectors of keywords constituting each cluster. The experiments for 3.147 academic papers revealed that the single vector-based traditional approach cannot properly map complex documents because of interference among subjects in each vector. With the proposed multi-vector based method, we ascertained that complex documents can be vectorized more accurately by eliminating the interference among subjects.

Knowledge Extraction Methodology and Framework from Wikipedia Articles for Construction of Knowledge-Base (지식베이스 구축을 위한 한국어 위키피디아의 학습 기반 지식추출 방법론 및 플랫폼 연구)

  • Kim, JaeHun;Lee, Myungjin
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.1
    • /
    • pp.43-61
    • /
    • 2019
  • Development of technologies in artificial intelligence has been rapidly increasing with the Fourth Industrial Revolution, and researches related to AI have been actively conducted in a variety of fields such as autonomous vehicles, natural language processing, and robotics. These researches have been focused on solving cognitive problems such as learning and problem solving related to human intelligence from the 1950s. The field of artificial intelligence has achieved more technological advance than ever, due to recent interest in technology and research on various algorithms. The knowledge-based system is a sub-domain of artificial intelligence, and it aims to enable artificial intelligence agents to make decisions by using machine-readable and processible knowledge constructed from complex and informal human knowledge and rules in various fields. A knowledge base is used to optimize information collection, organization, and retrieval, and recently it is used with statistical artificial intelligence such as machine learning. Recently, the purpose of the knowledge base is to express, publish, and share knowledge on the web by describing and connecting web resources such as pages and data. These knowledge bases are used for intelligent processing in various fields of artificial intelligence such as question answering system of the smart speaker. However, building a useful knowledge base is a time-consuming task and still requires a lot of effort of the experts. In recent years, many kinds of research and technologies of knowledge based artificial intelligence use DBpedia that is one of the biggest knowledge base aiming to extract structured content from the various information of Wikipedia. DBpedia contains various information extracted from Wikipedia such as a title, categories, and links, but the most useful knowledge is from infobox of Wikipedia that presents a summary of some unifying aspect created by users. These knowledge are created by the mapping rule between infobox structures and DBpedia ontology schema defined in DBpedia Extraction Framework. In this way, DBpedia can expect high reliability in terms of accuracy of knowledge by using the method of generating knowledge from semi-structured infobox data created by users. However, since only about 50% of all wiki pages contain infobox in Korean Wikipedia, DBpedia has limitations in term of knowledge scalability. This paper proposes a method to extract knowledge from text documents according to the ontology schema using machine learning. In order to demonstrate the appropriateness of this method, we explain a knowledge extraction model according to the DBpedia ontology schema by learning Wikipedia infoboxes. Our knowledge extraction model consists of three steps, document classification as ontology classes, proper sentence classification to extract triples, and value selection and transformation into RDF triple structure. The structure of Wikipedia infobox are defined as infobox templates that provide standardized information across related articles, and DBpedia ontology schema can be mapped these infobox templates. Based on these mapping relations, we classify the input document according to infobox categories which means ontology classes. After determining the classification of the input document, we classify the appropriate sentence according to attributes belonging to the classification. Finally, we extract knowledge from sentences that are classified as appropriate, and we convert knowledge into a form of triples. In order to train models, we generated training data set from Wikipedia dump using a method to add BIO tags to sentences, so we trained about 200 classes and about 2,500 relations for extracting knowledge. Furthermore, we evaluated comparative experiments of CRF and Bi-LSTM-CRF for the knowledge extraction process. Through this proposed process, it is possible to utilize structured knowledge by extracting knowledge according to the ontology schema from text documents. In addition, this methodology can significantly reduce the effort of the experts to construct instances according to the ontology schema.

Impact of Lambertian Cloud Top Pressure Error on Ozone Profile Retrieval Using OMI (램버시안 구름 모델의 운정기압 오차가 OMI 오존 프로파일 산출에 미치는 영향)

  • Nam, Hyeonshik;Kim, Jae Hawn;Shin, Daegeun;Baek, Kanghyun
    • Korean Journal of Remote Sensing
    • /
    • v.35 no.3
    • /
    • pp.347-358
    • /
    • 2019
  • Lambertian cloud model (Lambertian Cloud Model) is the simplified cloud model which is used to effectively retrieve the vertical ozone distribution of the atmosphere where the clouds exist. By using the Lambertian cloud model, the optical characteristics of clouds required for radiative transfer simulation are parametrized by Optical Centroid Cloud Pressure (OCCP) and Effective Cloud Fraction (ECF), and the accuracy of each parameter greatly affects the radiation simulation accuracy. However, it is very difficult to generalize the vertical ozone error due to the OCCP error because it varies depending on the radiation environment and algorithm setting. In addition, it is also difficult to analyze the effect of OCCP error because it is mixed with other errors that occur in the vertical ozone calculation process. This study analyzed the ozone retrieval error due to OCCP error using two methods. First, we simulated the impact of OCCP error on ozone retrieval based on Optimal Estimation. Using LIDORT radiation model, the radiation error due to the OCCP error is calculated. In order to convert the radiation error to the ozone calculation error, the radiation error is assigned to the conversion equation of the optimal estimation method. The results show that when the OCCP error occurs by 100 hPa, the total ozone is overestimated by 2.7%. Second, a case analysis is carried out to find the ozone retrieval error due to OCCP error. For the case analysis, the ozone retrieval error is simulated assuming OCCP error and compared with the ozone error in the case of PROFOZ 2005-2006, an OMI ozone profile product. In order to define the ozone error in the case, we assumed an ideal assumption. Considering albedo, and the horizontal change of ozone for satisfying the assumption, the 49 cases are selected. As a result, 27 out of 49 cases(about 55%)showed a correlation of 0.5 or more. This result show that the error of OCCP has a significant influence on the accuracy of ozone profile calculation.

Development of a Retrieval Algorithm for Adjustment of Satellite-viewed Cloudiness (위성관측운량 보정을 위한 알고리즘의 개발)

  • Son, Jiyoung;Lee, Yoon-Kyoung;Choi, Yong-Sang;Ok, Jung;Kim, Hye-Sil
    • Korean Journal of Remote Sensing
    • /
    • v.35 no.3
    • /
    • pp.415-431
    • /
    • 2019
  • The satellite-viewed cloudiness, a ratio of cloudy pixels to total pixels ($C_{sat,\;prev}$), inevitably differs from the "ground-viewed" cloudiness ($C_{grd}$) due to different viewpoints. Here we develop an algorithm to retrieve the satellite-viewed, but adjusted cloudiness to $C_{grd} (C_{sat,\;adj})$. The key process of the algorithm is to convert the cloudiness projected on the plane surface into the cloudiness on the celestial hemisphere from the observer. For this conversion, the supplementary satellite retrievals such as cloud detection and cloud top pressure are used as they provide locations of cloudy pixels and cloud base height information, respectively. The algorithm is tested for Himawari-8 level 1B data. The $C_{sat,\;adj}$ and $C_{sat,\;prev}$ are retrieved and validated with $C_{grd}$ of SYNOP station over Korea (22 stations) and China (724 stations) during only daytime for the first seven days of every month from July 2016 to June 2017. As results, the mean error of $C_{sat,\;adj}$ (0.61) is less that than that of $C_{sat,\;prev}$ (1.01). The percent of detection for 'Cloudy' scenario of $C_{sat,\;adj}$ (73%) is higher than that of $C_{sat,\;prev}$ (60%) The percent of correction, the accuracy, of $C_{sat,\;adj}$ is 61%, while that of $C_{sat,\;prev}$ is 55% for all seasons. For the December-January-February period when cloudy pixels are readily overestimated, the proportion of correction of $C_{sat,\;adj$ is 60%, while that of $C_{sat,\;prev}$ is 56%. Therefore, we conclude that the present algorithm can effectively get the satellite cloudiness near to the ground-viewed cloudiness.

Cellulose degrading basidiomycetes yeast isolated from the gut of grasshopper in Korea (한국의 메뚜기의 장에서 분리된 Cellulose를 분해하는 담자균 효모)

  • Kim, Ju-Young;Jang, Jun Hwee;Park, Ji-Hyun;Jung, Hee-Young;Park, Jong-Seok;Cho, Sung-Jin;Lee, Hoon Bok;Limtong, Savitree;Subramani, Gayathri;Sung, Gi-Ho;Kim, Myung Kyum
    • Korean Journal of Microbiology
    • /
    • v.54 no.4
    • /
    • pp.362-368
    • /
    • 2018
  • Grasshoppers play vital role in the digestion of photosynthetically fixed carbons. With the aid of intestinal microflora, the grasshopper can degrade leaves constituents such as cellulose and hemicellulose. The purpose of this study was to examine cellulolytic yeast isolates from the gut of grasshoppers collected in Gyeonggi Province, South Korea. Among the yeast isolates, ON2, ON17 (two strains), and ON6 (one strain) showed positive cellulolytic activity in the CMC-plate assay. The sequence analyses of D1/D2 domains of the large subunit rDNA gene and the internal transcribed spacer (ITS) regions revealed that the strains ON2 and ON17 were most closely related to Papiliotrema aspenensis CBS $13867^T$ (100%, sequence similarity in D1/D2 domains; 99.4% sequence similarity in ITS) and strain ON6 related to Saitozyma flava (100% in D1/D2 domains; 99.0% in ITS). All these three yeast strains are capable of degrading cellulose; therefore, the members of endosymbiotic yeasts may produce their own enzymes for carbohydrate degradation and convert mobilized sugar monomers to volatile fatty acids. Thus, the endosymbiotic yeast strains ON2, ON17 (represents the genus Papilioterma) and ON6 (Saitozyma) belonging to the family Tremellomycetes, are unreported strains in Korea.

Conditional Generative Adversarial Network based Collaborative Filtering Recommendation System (Conditional Generative Adversarial Network(CGAN) 기반 협업 필터링 추천 시스템)

  • Kang, Soyi;Shin, Kyung-shik
    • Journal of Intelligence and Information Systems
    • /
    • v.27 no.3
    • /
    • pp.157-173
    • /
    • 2021
  • With the development of information technology, the amount of available information increases daily. However, having access to so much information makes it difficult for users to easily find the information they seek. Users want a visualized system that reduces information retrieval and learning time, saving them from personally reading and judging all available information. As a result, recommendation systems are an increasingly important technologies that are essential to the business. Collaborative filtering is used in various fields with excellent performance because recommendations are made based on similar user interests and preferences. However, limitations do exist. Sparsity occurs when user-item preference information is insufficient, and is the main limitation of collaborative filtering. The evaluation value of the user item matrix may be distorted by the data depending on the popularity of the product, or there may be new users who have not yet evaluated the value. The lack of historical data to identify consumer preferences is referred to as data sparsity, and various methods have been studied to address these problems. However, most attempts to solve the sparsity problem are not optimal because they can only be applied when additional data such as users' personal information, social networks, or characteristics of items are included. Another problem is that real-world score data are mostly biased to high scores, resulting in severe imbalances. One cause of this imbalance distribution is the purchasing bias, in which only users with high product ratings purchase products, so those with low ratings are less likely to purchase products and thus do not leave negative product reviews. Due to these characteristics, unlike most users' actual preferences, reviews by users who purchase products are more likely to be positive. Therefore, the actual rating data is over-learned in many classes with high incidence due to its biased characteristics, distorting the market. Applying collaborative filtering to these imbalanced data leads to poor recommendation performance due to excessive learning of biased classes. Traditional oversampling techniques to address this problem are likely to cause overfitting because they repeat the same data, which acts as noise in learning, reducing recommendation performance. In addition, pre-processing methods for most existing data imbalance problems are designed and used for binary classes. Binary class imbalance techniques are difficult to apply to multi-class problems because they cannot model multi-class problems, such as objects at cross-class boundaries or objects overlapping multiple classes. To solve this problem, research has been conducted to convert and apply multi-class problems to binary class problems. However, simplification of multi-class problems can cause potential classification errors when combined with the results of classifiers learned from other sub-problems, resulting in loss of important information about relationships beyond the selected items. Therefore, it is necessary to develop more effective methods to address multi-class imbalance problems. We propose a collaborative filtering model using CGAN to generate realistic virtual data to populate the empty user-item matrix. Conditional vector y identify distributions for minority classes and generate data reflecting their characteristics. Collaborative filtering then maximizes the performance of the recommendation system via hyperparameter tuning. This process should improve the accuracy of the model by addressing the sparsity problem of collaborative filtering implementations while mitigating data imbalances arising from real data. Our model has superior recommendation performance over existing oversampling techniques and existing real-world data with data sparsity. SMOTE, Borderline SMOTE, SVM-SMOTE, ADASYN, and GAN were used as comparative models and we demonstrate the highest prediction accuracy on the RMSE and MAE evaluation scales. Through this study, oversampling based on deep learning will be able to further refine the performance of recommendation systems using actual data and be used to build business recommendation systems.