• Title/Summary/Keyword: 과학적 데이터 분석 방법론

Search Result 134, Processing Time 0.034 seconds

Evaluating Global Container Ports' Performance Considering the Port Calls' Attractiveness (기항 매력도를 고려한 세계 컨테이너 항만의 성과 평가)

  • Park, Byungin
    • Journal of Korea Port Economic Association
    • /
    • v.38 no.3
    • /
    • pp.105-131
    • /
    • 2022
  • Even after the improvement in 2019, UNCTAD's Liner Shipping Connectivity Index (LSCI), which evaluates the performance of the global container port market, has limited use. In particular, since the liner shipping connectivity index evaluates the performance based only on the distance of the relationship, the performance index combining the port attractiveness of calling would be more efficient. This study used the modified Huff model, the hub-authority algorithm and the eigenvector centrality of social network analysis, and correlation analysis for 2007, 2017, and 2019 data of Ocean-Commerce, Japan. The findings are as follows: Firstly, the port attractiveness of calling and the overall performance of the port did not always match. However, according to the analysis of the attractiveness of a port calling, Busan remained within the top 10. Still, the attractiveness among other Korean ports improved slowly from the low level during the study period. Secondly, Global container ports are generally specialized for long-term specialized inbound and outbound ports by the route and grow while maintaining professionalism throughout the entire period. The Korean ports continue to change roles from analysis period to period. Lastly, the volume of cargo by period and the extended port connectivity index (EPCI) presented in this study showed a correlation from 0.77 to 0.85. Even though the Atlantic data is excluded from the analysis and the ship's operable capacity is used instead of the port throughput volume, it shows a high correlation. The study result would help evaluate and analyze global ports. According to the study, Korean ports need a long-term strategy to improve performance while maintaining professionalism. In order to maintain and develop the port's desirable role, it is necessary to utilize cooperation and partnerships with the complimentary port and attract shipping companies' services calling to the complementary port. Although this study carried out a complex analysis using a lot of data and methodologies for an extended period, it is necessary to conduct a study covering ports around the world, a long-term panel analysis, and a scientific parameter estimation study of the attractiveness analysis.

A Study on Web-based Technology Valuation System (웹기반 지능형 기술가치평가 시스템에 관한 연구)

  • Sung, Tae-Eung;Jun, Seung-Pyo;Kim, Sang-Gook;Park, Hyun-Woo
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.1
    • /
    • pp.23-46
    • /
    • 2017
  • Although there have been cases of evaluating the value of specific companies or projects which have centralized on developed countries in North America and Europe from the early 2000s, the system and methodology for estimating the economic value of individual technologies or patents has been activated on and on. Of course, there exist several online systems that qualitatively evaluate the technology's grade or the patent rating of the technology to be evaluated, as in 'KTRS' of the KIBO and 'SMART 3.1' of the Korea Invention Promotion Association. However, a web-based technology valuation system, referred to as 'STAR-Value system' that calculates the quantitative values of the subject technology for various purposes such as business feasibility analysis, investment attraction, tax/litigation, etc., has been officially opened and recently spreading. In this study, we introduce the type of methodology and evaluation model, reference information supporting these theories, and how database associated are utilized, focusing various modules and frameworks embedded in STAR-Value system. In particular, there are six valuation methods, including the discounted cash flow method (DCF), which is a representative one based on the income approach that anticipates future economic income to be valued at present, and the relief-from-royalty method, which calculates the present value of royalties' where we consider the contribution of the subject technology towards the business value created as the royalty rate. We look at how models and related support information (technology life, corporate (business) financial information, discount rate, industrial technology factors, etc.) can be used and linked in a intelligent manner. Based on the classification of information such as International Patent Classification (IPC) or Korea Standard Industry Classification (KSIC) for technology to be evaluated, the STAR-Value system automatically returns meta data such as technology cycle time (TCT), sales growth rate and profitability data of similar company or industry sector, weighted average cost of capital (WACC), indices of industrial technology factors, etc., and apply adjustment factors to them, so that the result of technology value calculation has high reliability and objectivity. Furthermore, if the information on the potential market size of the target technology and the market share of the commercialization subject refers to data-driven information, or if the estimated value range of similar technologies by industry sector is provided from the evaluation cases which are already completed and accumulated in database, the STAR-Value is anticipated that it will enable to present highly accurate value range in real time by intelligently linking various support modules. Including the explanation of the various valuation models and relevant primary variables as presented in this paper, the STAR-Value system intends to utilize more systematically and in a data-driven way by supporting the optimal model selection guideline module, intelligent technology value range reasoning module, and similar company selection based market share prediction module, etc. In addition, the research on the development and intelligence of the web-based STAR-Value system is significant in that it widely spread the web-based system that can be used in the validation and application to practices of the theoretical feasibility of the technology valuation field, and it is expected that it could be utilized in various fields of technology commercialization.

A Study on Robust Optimal Sensor Placement for Real-time Monitoring of Containment Buildings in Nuclear Power Plants (원전 격납 건물의 실시간 모니터링을 위한 강건한 최적 센서배치 연구)

  • Chanwoo Lee;Youjin Kim;Hyung-jo Jung
    • Journal of the Computational Structural Engineering Institute of Korea
    • /
    • v.36 no.3
    • /
    • pp.155-163
    • /
    • 2023
  • Real-time monitoring technology is critical for ensuring the safety and reliability of nuclear power plant structures. However, the current seismic monitoring system has limited system identification capabilities such as modal parameter estimation. To obtain global behavior data and dynamic characteristics, multiple sensors must be optimally placed. Although several studies on optimal sensor placement have been conducted, they have primarily focused on civil and mechanical structures. Nuclear power plant structures require robust signals, even at low signal-to-noise ratios, and the robustness of each mode must be assessed separately. This is because the mode contributions of nuclear power plant containment buildings are concentrated in low-order modes. Therefore, this study proposes an optimal sensor placement methodology that can evaluate robustness against noise and the effects of each mode. Indicators, such as auto modal assurance criterion (MAC), cross MAC, and mode shape distribution by node were analyzed, and the suitability of the methodology was verified through numerical analysis.

A New Correction Method for Ship's Viscous Magnetization Effect on Shipboard Three-component Magnetic Data Using a Total Field Magnetometer (총자력계를 이용한 선상 삼성분 자기 데이터의 선박 점성 자화 효과에 대한 새로운 보정 방법 연구)

  • Hanjin Choe;Nobukazu Seama
    • Geophysics and Geophysical Exploration
    • /
    • v.27 no.2
    • /
    • pp.119-128
    • /
    • 2024
  • Marine magnetic surveys provide a rapid and cost-effective method for pioneer geophysical survey for many purposes. Sea-surface magnetometers offer high accuracy but are limited to measuring the scalar total magnetic field and require dedicated cruise missions. Shipboard three-component magnetometers, on the other hand, can collect vector three components and applicable to any cruise missions. However, correcting for the ship's magnetic field, particularly viscous magnetization, still remains a challenge. This study proposes a new additional correction method for ship's viscous magnetization effect in vector data acquired by shipboard three-component magnetometer. This method utilizes magnetic data collected simultaneously with a sea-surface magnetometer providing total magnetic field measurements. Our method significantly reduces deviations between the two datasets, resulting in corrected vector anomalies with errors as low as 7-25 nT. These tiny errors are possibly caused by the vector magnetic anomaly and its related viscous magnetization. This method is expected to significantly improve the accuracy of shipborne magnetic surveys by providing corrected vector components. This will enhance magnetic interpretations and might be useful for understanding plate tectonics, geological structures, hydrothermal deposits, and more.

Real-time Natural Disaster Failure Analysis Information System Development using GIS Environment (GIS환경의 실시간 자연재해정보를 연계한 재해고장분석시스템 개발)

  • Ahn, Yeon-S.
    • Journal of Digital Contents Society
    • /
    • v.10 no.4
    • /
    • pp.639-648
    • /
    • 2009
  • Earth's environment issues are introduced recently and every year the social loss have been occurred by the impact of various disaster. This kind of disaster and weather problems are the increasing reason of electricity transmission network equipment's failures because of exposing by the natural environment. The emergency and abnormal status of electricity equipment make the power outage of manufacturing plant and discomfort of people's lives. So, to protect the electricity equipment from the natural disasters and to supply the power to customer as stable, the supporting systems are required. In this paper, the research results are described the development process and the outcomes of the real-time natural disaster failure analysis information system including the describing about the impact of disaster and weather change, making the natural weather information, and linking the realtime monitoring system. As of development process, according to application development methodology, techniques are enumerated including the real time interface with related systems, the analysing the geographic information on the digital map using GIS application technology to extract the malfunction equipment potentially and to manage the equipments efficiently. Through this system makes remarkable performance it minimize the failures of the equipments, the increasing the efficiency of the equipment operation, the support of scientific information related on the mid-term enhancement plan, the savings on equipment investment, the quality upgrading of electricity supply, and the various supports in the field.

  • PDF

Reliability-Based Design Optimization of 130m Class Fixed-Type Offshore Platform (신뢰성 기반 최적설계를 이용한 130m급 고정식 해양구조물 최적설계 개발)

  • Kim, Hyun-Seok;Kim, Hyun-Sung;Park, Byoungjae;Lee, Kangsu
    • Journal of the Computational Structural Engineering Institute of Korea
    • /
    • v.34 no.5
    • /
    • pp.263-270
    • /
    • 2021
  • In this study, a reliability-based design optimization of a 130-m class fixed-type offshore platform, to be installed in the North Sea, was carried out, while considering environmental, material, and manufacturing uncertainties to enhance its structural safety and economic aspects. For the reliability analysis, and reliability-based design optimization of the structural integrity, unity check values (defined as the ratio between working and allowable stress, for axial, bending, and shear stresses), of the members of the offshore platform were considered as constraints. Weight of the supporting jacket structure was minimized to reduce the manufacturing cost of the offshore platform. Statistical characteristics of uncertainties were defined based on observed and measured data references. Reliability analysis and reliability-based design optimization of a jacket-type offshore structure were computationally burdensome due to the large number of members; therefore, we suggested a method for variable screening, based on the importance of their output responses, to reduce the dimension of the problem. Furthermore, a deterministic design optimization was carried out prior to the reliability-based design optimization, to improve overall computational efficiency. Finally, the optimal design obtained was compared with the conventional rule-based offshore platform design in terms of safety and cost.

Label Embedding for Improving Classification Accuracy UsingAutoEncoderwithSkip-Connections (다중 레이블 분류의 정확도 향상을 위한 스킵 연결 오토인코더 기반 레이블 임베딩 방법론)

  • Kim, Museong;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.27 no.3
    • /
    • pp.175-197
    • /
    • 2021
  • Recently, with the development of deep learning technology, research on unstructured data analysis is being actively conducted, and it is showing remarkable results in various fields such as classification, summary, and generation. Among various text analysis fields, text classification is the most widely used technology in academia and industry. Text classification includes binary class classification with one label among two classes, multi-class classification with one label among several classes, and multi-label classification with multiple labels among several classes. In particular, multi-label classification requires a different training method from binary class classification and multi-class classification because of the characteristic of having multiple labels. In addition, since the number of labels to be predicted increases as the number of labels and classes increases, there is a limitation in that performance improvement is difficult due to an increase in prediction difficulty. To overcome these limitations, (i) compressing the initially given high-dimensional label space into a low-dimensional latent label space, (ii) after performing training to predict the compressed label, (iii) restoring the predicted label to the high-dimensional original label space, research on label embedding is being actively conducted. Typical label embedding techniques include Principal Label Space Transformation (PLST), Multi-Label Classification via Boolean Matrix Decomposition (MLC-BMaD), and Bayesian Multi-Label Compressed Sensing (BML-CS). However, since these techniques consider only the linear relationship between labels or compress the labels by random transformation, it is difficult to understand the non-linear relationship between labels, so there is a limitation in that it is not possible to create a latent label space sufficiently containing the information of the original label. Recently, there have been increasing attempts to improve performance by applying deep learning technology to label embedding. Label embedding using an autoencoder, a deep learning model that is effective for data compression and restoration, is representative. However, the traditional autoencoder-based label embedding has a limitation in that a large amount of information loss occurs when compressing a high-dimensional label space having a myriad of classes into a low-dimensional latent label space. This can be found in the gradient loss problem that occurs in the backpropagation process of learning. To solve this problem, skip connection was devised, and by adding the input of the layer to the output to prevent gradient loss during backpropagation, efficient learning is possible even when the layer is deep. Skip connection is mainly used for image feature extraction in convolutional neural networks, but studies using skip connection in autoencoder or label embedding process are still lacking. Therefore, in this study, we propose an autoencoder-based label embedding methodology in which skip connections are added to each of the encoder and decoder to form a low-dimensional latent label space that reflects the information of the high-dimensional label space well. In addition, the proposed methodology was applied to actual paper keywords to derive the high-dimensional keyword label space and the low-dimensional latent label space. Using this, we conducted an experiment to predict the compressed keyword vector existing in the latent label space from the paper abstract and to evaluate the multi-label classification by restoring the predicted keyword vector back to the original label space. As a result, the accuracy, precision, recall, and F1 score used as performance indicators showed far superior performance in multi-label classification based on the proposed methodology compared to traditional multi-label classification methods. This can be seen that the low-dimensional latent label space derived through the proposed methodology well reflected the information of the high-dimensional label space, which ultimately led to the improvement of the performance of the multi-label classification itself. In addition, the utility of the proposed methodology was identified by comparing the performance of the proposed methodology according to the domain characteristics and the number of dimensions of the latent label space.

Topic Model Augmentation and Extension Method using LDA and BERTopic (LDA와 BERTopic을 이용한 토픽모델링의 증강과 확장 기법 연구)

  • Kim, SeonWook;Yang, Kiduk
    • Journal of the Korean Society for information Management
    • /
    • v.39 no.3
    • /
    • pp.99-132
    • /
    • 2022
  • The purpose of this study is to propose AET (Augmented and Extended Topics), a novel method of synthesizing both LDA and BERTopic results, and to analyze the recently published LIS articles as an experimental approach. To achieve the purpose of this study, 55,442 abstracts from 85 LIS journals within the WoS database, which spans from January 2001 to October 2021, were analyzed. AET first constructs a WORD2VEC-based cosine similarity matrix between LDA and BERTopic results, extracts AT (Augmented Topics) by repeating the matrix reordering and segmentation procedures as long as their semantic relations are still valid, and finally determines ET (Extended Topics) by removing any LDA related residual subtopics from the matrix and ordering the rest of them by F1 (BERTopic topic size rank, Inverse cosine similarity rank). AET, by comparing with the baseline LDA result, shows that AT has effectively concretized the original LDA topic model and ET has discovered new meaningful topics that LDA didn't. When it comes to the qualitative performance evaluation, AT performs better than LDA while ET shows similar performances except in a few cases.

Predicting stock movements based on financial news with systematic group identification (시스템적인 군집 확인과 뉴스를 이용한 주가 예측)

  • Seong, NohYoon;Nam, Kihwan
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.3
    • /
    • pp.1-17
    • /
    • 2019
  • Because stock price forecasting is an important issue both academically and practically, research in stock price prediction has been actively conducted. The stock price forecasting research is classified into using structured data and using unstructured data. With structured data such as historical stock price and financial statements, past studies usually used technical analysis approach and fundamental analysis. In the big data era, the amount of information has rapidly increased, and the artificial intelligence methodology that can find meaning by quantifying string information, which is an unstructured data that takes up a large amount of information, has developed rapidly. With these developments, many attempts with unstructured data are being made to predict stock prices through online news by applying text mining to stock price forecasts. The stock price prediction methodology adopted in many papers is to forecast stock prices with the news of the target companies to be forecasted. However, according to previous research, not only news of a target company affects its stock price, but news of companies that are related to the company can also affect the stock price. However, finding a highly relevant company is not easy because of the market-wide impact and random signs. Thus, existing studies have found highly relevant companies based primarily on pre-determined international industry classification standards. However, according to recent research, global industry classification standard has different homogeneity within the sectors, and it leads to a limitation that forecasting stock prices by taking them all together without considering only relevant companies can adversely affect predictive performance. To overcome the limitation, we first used random matrix theory with text mining for stock prediction. Wherever the dimension of data is large, the classical limit theorems are no longer suitable, because the statistical efficiency will be reduced. Therefore, a simple correlation analysis in the financial market does not mean the true correlation. To solve the issue, we adopt random matrix theory, which is mainly used in econophysics, to remove market-wide effects and random signals and find a true correlation between companies. With the true correlation, we perform cluster analysis to find relevant companies. Also, based on the clustering analysis, we used multiple kernel learning algorithm, which is an ensemble of support vector machine to incorporate the effects of the target firm and its relevant firms simultaneously. Each kernel was assigned to predict stock prices with features of financial news of the target firm and its relevant firms. The results of this study are as follows. The results of this paper are as follows. (1) Following the existing research flow, we confirmed that it is an effective way to forecast stock prices using news from relevant companies. (2) When looking for a relevant company, looking for it in the wrong way can lower AI prediction performance. (3) The proposed approach with random matrix theory shows better performance than previous studies if cluster analysis is performed based on the true correlation by removing market-wide effects and random signals. The contribution of this study is as follows. First, this study shows that random matrix theory, which is used mainly in economic physics, can be combined with artificial intelligence to produce good methodologies. This suggests that it is important not only to develop AI algorithms but also to adopt physics theory. This extends the existing research that presented the methodology by integrating artificial intelligence with complex system theory through transfer entropy. Second, this study stressed that finding the right companies in the stock market is an important issue. This suggests that it is not only important to study artificial intelligence algorithms, but how to theoretically adjust the input values. Third, we confirmed that firms classified as Global Industrial Classification Standard (GICS) might have low relevance and suggested it is necessary to theoretically define the relevance rather than simply finding it in the GICS.

Analyzing Research Trends in Blockchain Studies in South Korea Using Dynamic Topic Modeling and Network Analysis (다이나믹 토픽모델링 및 네트워크 분석 기법을 통한 블록체인 관련 국내 연구 동향 분석)

  • Kim, Donghun;Oh, Chanhee;Zhu, Yongjun
    • Journal of the Korean Society for information Management
    • /
    • v.38 no.3
    • /
    • pp.23-39
    • /
    • 2021
  • This study aims to explore research trends in Blockchain studies in South Korea using dynamic topic modeling and network analysis. To achieve this goal, we conducted the university & institute collaboration network analysis, the keyword co-occurrence network analysis, and times series topic analysis using dynamic topic modeling. Through the university & institute collaboration network analysis, we found major universities such as Soongsil University, Soonchunhyang University, Korea University, Korea Advanced Institute of Science and Technology (KAIST) and major institutes such as Ministry of National Defense, Korea Railroad Research Institute, Samil PricewaterhouseCoopers, Electronics and Telecommunications Research Institute that led collaborative research. Next, through the analysis of the keyword co-occurrence network, we found major research keywords including virtual assets (Cryptocurrency, Bitcoin, Ethereum, Virtual currency), blockchain technology (Distributed ledger, Distributed ledger technology), finance (Smart contract), and information security (Security, privacy, Personal information). Smart contracts showed the highest scores in all network centrality measures showing its importance in the field. Finally, through the time series topic analysis, we identified five major topics including blockchain technology, blockchain ecosystem, blockchain application 1 (trade, online voting, real estate), blockchain application 2 (food, tourism, distribution, media), and blockchain application 3 (economy, finance). Changes of topics were also investigated by exploring proportions of representative keywords for each topic. The study is the first of its kind to attempt to conduct university & institute collaboration networks analysis and dynamic topic modeling-based times series topic analysis for exploring research trends in Blockchain studies in South Korea. Our results can be used by government agencies, universities, and research institutes to develop effective strategies of promoting university & institutes collaboration and interdisciplinary research in the field.