• Title/Summary/Keyword: 데이터 비식별화

Search Result 101, Processing Time 0.026 seconds

A Study on the Optimization of Fire Awareness Model Based on Convolutional Neural Network: Layer Importance Evaluation-Based Approach (합성곱 신경망 기반 화재 인식 모델 최적화 연구: Layer Importance Evaluation 기반 접근법)

  • Won Jin;Mi-Hwa Song
    • The Transactions of the Korea Information Processing Society
    • /
    • v.13 no.9
    • /
    • pp.444-452
    • /
    • 2024
  • This study proposes a deep learning architecture optimized for fire detection derived through Layer Importance Evaluation. In order to solve the problem of unnecessary complexity and operation of the existing Convolutional Neural Network (CNN)-based fire detection system, the operation of the inner layer of the model based on the weight and activation values was analyzed through the Layer Importance Evaluation technique, the layer with a high contribution to fire detection was identified, and the model was reconstructed only with the identified layer, and the performance indicators were compared and analyzed with the existing model. After learning the fire data using four transfer learning models: Xception, VGG19, ResNet, and EfficientNetB5, the Layer Importance Evaluation technique was applied to analyze the weight and activation value of each layer, and then a new model was constructed by selecting the top rank layers with the highest contribution. As a result of the study, it was confirmed that the implemented architecture maintains the same performance with parameters that are about 80% lighter than the existing model, and can contribute to increasing the efficiency of fire monitoring equipment by outputting the same performance in accuracy, loss, and confusion matrix indicators compared to conventional complex transfer learning models while having a learning speed of about 3 to 5 times faster.

Managing Duplicate Memberships of Websites : An Approach of Social Network Analysis (웹사이트 중복회원 관리 : 소셜 네트워크 분석 접근)

  • Kang, Eun-Young;Kwahk, Kee-Young
    • Journal of Intelligence and Information Systems
    • /
    • v.17 no.1
    • /
    • pp.153-169
    • /
    • 2011
  • Today using Internet environment is considered absolutely essential for establishing corporate marketing strategy. Companies have promoted their products and services through various ways of on-line marketing activities such as providing gifts and points to customers in exchange for participating in events, which is based on customers' membership data. Since companies can use these membership data to enhance their marketing efforts through various data analysis, appropriate website membership management may play an important role in increasing the effectiveness of on-line marketing campaign. Despite the growing interests in proper membership management, however, there have been difficulties in identifying inappropriate members who can weaken on-line marketing effectiveness. In on-line environment, customers tend to not reveal themselves clearly compared to off-line market. Customers who have malicious intent are able to create duplicate IDs by using others' names illegally or faking login information during joining membership. Since the duplicate members are likely to intercept gifts and points that should be sent to appropriate customers who deserve them, this can result in ineffective marketing efforts. Considering that the number of website members and its related marketing costs are significantly increasing, it is necessary for companies to find efficient ways to screen and exclude unfavorable troublemakers who are duplicate members. With this motivation, this study proposes an approach for managing duplicate membership based on the social network analysis and verifies its effectiveness using membership data gathered from real websites. A social network is a social structure made up of actors called nodes, which are tied by one or more specific types of interdependency. Social networks represent the relationship between the nodes and show the direction and strength of the relationship. Various analytical techniques have been proposed based on the social relationships, such as centrality analysis, structural holes analysis, structural equivalents analysis, and so on. Component analysis, one of the social network analysis techniques, deals with the sub-networks that form meaningful information in the group connection. We propose a method for managing duplicate memberships using component analysis. The procedure is as follows. First step is to identify membership attributes that will be used for analyzing relationship patterns among memberships. Membership attributes include ID, telephone number, address, posting time, IP address, and so on. Second step is to compose social matrices based on the identified membership attributes and aggregate the values of each social matrix into a combined social matrix. The combined social matrix represents how strong pairs of nodes are connected together. When a pair of nodes is strongly connected, we expect that those nodes are likely to be duplicate memberships. The combined social matrix is transformed into a binary matrix with '0' or '1' of cell values using a relationship criterion that determines whether the membership is duplicate or not. Third step is to conduct a component analysis for the combined social matrix in order to identify component nodes and isolated nodes. Fourth, identify the number of real memberships and calculate the reliability of website membership based on the component analysis results. The proposed procedure was applied to three real websites operated by a pharmaceutical company. The empirical results showed that the proposed method was superior to the traditional database approach using simple address comparison. In conclusion, this study is expected to shed some light on how social network analysis can enhance a reliable on-line marketing performance by efficiently and effectively identifying duplicate memberships of websites.

Computation of Maintainability Index Using SysML-Based M&S Technique for Improved Weapon Systems Development (SysML 기반 모델링 및 시뮬레이션 기법을 활용한 무기체계 정비도 지수 산출)

  • Yoo, Yeon-Yong;Lee, Jae-Chon
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.19 no.11
    • /
    • pp.88-95
    • /
    • 2018
  • Maintainability indicates how easily a system can be restored to the normal state when a system failure occurs. Systems developed to have high maintainability can be competitive due to reduced maintenance time, workforce and resources. Quantification of the maintainability is possible in many ways, but only after prototype production or with historical data. As such, the graph theory and 3D model data have been used, but there are limitations in management efficiency and early use. To solve this problem, we studied the maintainability index of weapon systems using SysML-based modeling and simulation technique. A SysML structure diagram was generated to simultaneously model the system design and maintainability of system components by reflecting the maintainability attributes acquired from the system engineering tool. Then, a SysML parametric diagram was created to quantify the maintainability through simulation linked with MATLAB. As a result, an integrated model to account for system design and maintainability simultaneously has been presented. The model can be used from early design stages to identify components with low maintainability index. The design of such components can be changed to improve maintainability and thus to reduce the risks of cost overruns and time delays due to belated design changes.

The Information Modeling Method based on Extended IFC for Alignment-based Objects of Railway Track (선형중심 객체 관리를 위한 확장된 IFC 기반 철도 궤도부 정보모델링 방안)

  • Kwon, Tae Ho;Park, Sang I.;Seo, Kyung-Wan;Lee, Sang-Ho
    • Journal of the Computational Structural Engineering Institute of Korea
    • /
    • v.31 no.6
    • /
    • pp.339-346
    • /
    • 2018
  • An Industry Foundation Classes(IFC), which is a data schema developed focusing on architecture, is being expanded to civil engineering structures. However, it is difficult to create an information model based on extended IFC since the BIM software cannot provide support functions. To manage a railway track based on the extended IFC, this paper proposed a method to create an alignment-centered separated railway track model and convert it to an extended IFC-based information model. First, railway track elements have been classified into continuous and discontinuous structures. The continuous structures were created by an alignment-based software, and discontinuous structures were created as independent objects through linkage of the discretized alignment. Second, a classification system and extended IFC schema for railway track have been proposed. Finally, the semantic information was identified by using the property of classification code and user interface. The availability of the methods was verified by developing an extended IFC-based information model of the Osong railway site.

Assessment of the Impact Factor and the Stress Histogram of Railway Bridges in Korea (국내 강철도 교량의 충격계수 및 응력빈도분포의 평가)

  • Choi, Jun Hyeok;Cho, Sun Kyu;Chang, Dong Il
    • Journal of Korean Society of Steel Construction
    • /
    • v.9 no.4 s.33
    • /
    • pp.489-500
    • /
    • 1997
  • The number of railway bridges in Korea reaches to about 3,000. Among them, the ratio of steel bridges is about 47%. Most of the long span railway bridges take the truss type, its number is 15 in Korea. These bridges have more than 30 years service life, it is expected that the damages of cracks and corrosion is going. Therefore, the estimation of fatigue life with random stress is considered as significant subject in maintenance. In this paper, the random stress was measured in bridges and counted the stress cycles using the rainflow counting method So, the stress range frequencies and their equivalent stress are obtained and the amplification factor for the span length, bridge types and traveling speed is assessed by the dynamic testing. From the results, the stress range of the stringer and the lower chord is higher than the other members and the characteristics of the stress histogram is varied to the loading system and the amount of the passing through trains. And the impact effects are depended on the traveling speed as well as the span length.

  • PDF

A Reduction Method of Over-Segmented Regions at Image Segmentation based on Homogeneity Threshold (동질성 문턱 값 기반 영상분할에서 과분할 영역 축소 방법)

  • Han, Gi-Tae
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.1 no.1
    • /
    • pp.55-68
    • /
    • 2012
  • In this paper, we propose a novel method to solve the problem of excessive segmentation out of the method of segmenting regions from an image using Homogeneity Threshold($H_T$). The algorithm of the previous image segmentation based on $H_T$ was carried out region growth by using only the center pixel of selected window. Therefore it was caused resulting in excessive segmented regions. However, before carrying region growth, the proposed method first of all finds out whether the selected window is homogeneity or not. Subsequently, if the selected window is homogeneity it carries out region growth using the total pixels of selected window. But if the selected window is not homogeneity, it carries out region growth using only the center pixel of selected window. So, the method can reduce remarkably the number of excessive segmented regions of image segmentation based on $H_T$. In order to show the validity of the proposed method, we carried out multiple experiments to compare the proposed method with previous method in same environment and conditions. As the results, the proposed method can reduce the number of segmented regions above 40% and doesn't make any difference in the quality of visual image when we compare with previous method. Especially, when we compare the image united with regions of descending order by size of segmented regions in experimentation with the previous method, even though the united image has regions more than 1,000, we can't recognize what the image means. However, in the proposed method, even though image is united by segmented regions less than 10, we can recognize what the image is. For these reason, we expect that the proposed method will be utilized in various fields, such as the extraction of objects, the retrieval of informations from the image, research for anatomy, biology, image visualization, and animation and so on.

Development of LiDAR-Based MRM Algorithm for LKS System (LKS 시스템을 위한 라이다 기반 MRM 알고리즘 개발)

  • Son, Weon Il;Oh, Tae Young;Park, Kihong
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.20 no.1
    • /
    • pp.174-192
    • /
    • 2021
  • The LIDAR sensor, which provides higher cognitive performance than cameras and radar, is difficult to apply to ADAS or autonomous driving because of its high price. On the other hand, as the price is decreasing rapidly, expectations are rising to improve existing autonomous driving functions by taking advantage of the LIDAR sensor. In level 3 autonomous vehicles, when a dangerous situation in the cognitive module occurs due to a sensor defect or sensor limit, the driver must take control of the vehicle for manual driving. If the driver does not respond to the request, the system must automatically kick in and implement a minimum risk maneuver to maintain the risk within a tolerable level. In this study, based on this background, a LIDAR-based LKS MRM algorithm was developed for the case when the normal operation of LKS was not possible due to troubles in the cognitive system. From point cloud data collected by LIDAR, the algorithm generates the trajectory of the vehicle in front through object clustering and converts it to the target waypoints of its own. Hence, if the camera-based LKS is not operating normally, LIDAR-based path tracking control is performed as MRM. The HAZOP method was used to identify the risk sources in the LKS cognitive systems. B, and based on this, test scenarios were derived and used in the validation process by simulation. The simulation results indicated that the LIDAR-based LKS MRM algorithm of this study prevents lane departure in dangerous situations caused by various problems or difficulties in the LKS cognitive systems and could prevent possible traffic accidents.

Technique to Reduce Container Restart for Improving Execution Time of Container Workflow in Kubernetes Environments (쿠버네티스 환경에서 컨테이너 워크플로의 실행 시간 개선을 위한 컨테이너 재시작 감소 기법)

  • Taeshin Kang;Heonchang Yu
    • The Transactions of the Korea Information Processing Society
    • /
    • v.13 no.3
    • /
    • pp.91-101
    • /
    • 2024
  • The utilization of container virtualization technology ensures the consistency and portability of data-intensive and memory volatile workflows. Kubernetes serves as the de facto standard for orchestrating these container applications. Cloud users often overprovision container applications to avoid container restarts caused by resource shortages. However, overprovisioning results in decreased CPU and memory resource utilization. To address this issue, oversubscription of container resources is commonly employed, although excessive oversubscription of memory resources can lead to a cascade of container restarts due to node memory scarcity. Container restarts can reset operations and impose substantial overhead on containers with high memory volatility that include numerous stateful applications. This paper proposes a technique to mitigate container restarts in a memory oversubscription environment based on Kubernetes. The proposed technique involves identifying containers that are likely to request memory allocation on nodes experiencing high memory usage and temporarily pausing these containers. By significantly reducing the CPU usage of containers, an effect similar to a paused state is achieved. The suspension of the identified containers is released once it is determined that the corresponding node's memory usage has been reduced. The average number of container restarts was reduced by an average of 40% and a maximum of 58% when executing a high memory volatile workflow in a Kubernetes environment with the proposed method compared to its absence. Furthermore, the total execution time of a container workflow is decreased by an average of 7% and a maximum of 13% due to the reduced frequency of container restarts.

A Study on the Application of Outlier Analysis for Fraud Detection: Focused on Transactions of Auction Exception Agricultural Products (부정 탐지를 위한 이상치 분석 활용방안 연구 : 농수산 상장예외품목 거래를 대상으로)

  • Kim, Dongsung;Kim, Kitae;Kim, Jongwoo;Park, Steve
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.3
    • /
    • pp.93-108
    • /
    • 2014
  • To support business decision making, interests and efforts to analyze and use transaction data in different perspectives are increasing. Such efforts are not only limited to customer management or marketing, but also used for monitoring and detecting fraud transactions. Fraud transactions are evolving into various patterns by taking advantage of information technology. To reflect the evolution of fraud transactions, there are many efforts on fraud detection methods and advanced application systems in order to improve the accuracy and ease of fraud detection. As a case of fraud detection, this study aims to provide effective fraud detection methods for auction exception agricultural products in the largest Korean agricultural wholesale market. Auction exception products policy exists to complement auction-based trades in agricultural wholesale market. That is, most trades on agricultural products are performed by auction; however, specific products are assigned as auction exception products when total volumes of products are relatively small, the number of wholesalers is small, or there are difficulties for wholesalers to purchase the products. However, auction exception products policy makes several problems on fairness and transparency of transaction, which requires help of fraud detection. In this study, to generate fraud detection rules, real huge agricultural products trade transaction data from 2008 to 2010 in the market are analyzed, which increase more than 1 million transactions and 1 billion US dollar in transaction volume. Agricultural transaction data has unique characteristics such as frequent changes in supply volumes and turbulent time-dependent changes in price. Since this was the first trial to identify fraud transactions in this domain, there was no training data set for supervised learning. So, fraud detection rules are generated using outlier detection approach. We assume that outlier transactions have more possibility of fraud transactions than normal transactions. The outlier transactions are identified to compare daily average unit price, weekly average unit price, and quarterly average unit price of product items. Also quarterly averages unit price of product items of the specific wholesalers are used to identify outlier transactions. The reliability of generated fraud detection rules are confirmed by domain experts. To determine whether a transaction is fraudulent or not, normal distribution and normalized Z-value concept are applied. That is, a unit price of a transaction is transformed to Z-value to calculate the occurrence probability when we approximate the distribution of unit prices to normal distribution. The modified Z-value of the unit price in the transaction is used rather than using the original Z-value of it. The reason is that in the case of auction exception agricultural products, Z-values are influenced by outlier fraud transactions themselves because the number of wholesalers is small. The modified Z-values are called Self-Eliminated Z-scores because they are calculated excluding the unit price of the specific transaction which is subject to check whether it is fraud transaction or not. To show the usefulness of the proposed approach, a prototype of fraud transaction detection system is developed using Delphi. The system consists of five main menus and related submenus. First functionalities of the system is to import transaction databases. Next important functions are to set up fraud detection parameters. By changing fraud detection parameters, system users can control the number of potential fraud transactions. Execution functions provide fraud detection results which are found based on fraud detection parameters. The potential fraud transactions can be viewed on screen or exported as files. The study is an initial trial to identify fraud transactions in Auction Exception Agricultural Products. There are still many remained research topics of the issue. First, the scope of analysis data was limited due to the availability of data. It is necessary to include more data on transactions, wholesalers, and producers to detect fraud transactions more accurately. Next, we need to extend the scope of fraud transaction detection to fishery products. Also there are many possibilities to apply different data mining techniques for fraud detection. For example, time series approach is a potential technique to apply the problem. Even though outlier transactions are detected based on unit prices of transactions, however it is possible to derive fraud detection rules based on transaction volumes.

A Study on Intelligent Value Chain Network System based on Firms' Information (기업정보 기반 지능형 밸류체인 네트워크 시스템에 관한 연구)

  • Sung, Tae-Eung;Kim, Kang-Hoe;Moon, Young-Su;Lee, Ho-Shin
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.3
    • /
    • pp.67-88
    • /
    • 2018
  • Until recently, as we recognize the significance of sustainable growth and competitiveness of small-and-medium sized enterprises (SMEs), governmental support for tangible resources such as R&D, manpower, funds, etc. has been mainly provided. However, it is also true that the inefficiency of support systems such as underestimated or redundant support has been raised because there exist conflicting policies in terms of appropriateness, effectiveness and efficiency of business support. From the perspective of the government or a company, we believe that due to limited resources of SMEs technology development and capacity enhancement through collaboration with external sources is the basis for creating competitive advantage for companies, and also emphasize value creation activities for it. This is why value chain network analysis is necessary in order to analyze inter-company deal relationships from a series of value chains and visualize results through establishing knowledge ecosystems at the corporate level. There exist Technology Opportunity Discovery (TOD) system that provides information on relevant products or technology status of companies with patents through retrievals over patent, product, or company name, CRETOP and KISLINE which both allow to view company (financial) information and credit information, but there exists no online system that provides a list of similar (competitive) companies based on the analysis of value chain network or information on potential clients or demanders that can have business deals in future. Therefore, we focus on the "Value Chain Network System (VCNS)", a support partner for planning the corporate business strategy developed and managed by KISTI, and investigate the types of embedded network-based analysis modules, databases (D/Bs) to support them, and how to utilize the system efficiently. Further we explore the function of network visualization in intelligent value chain analysis system which becomes the core information to understand industrial structure ystem and to develop a company's new product development. In order for a company to have the competitive superiority over other companies, it is necessary to identify who are the competitors with patents or products currently being produced, and searching for similar companies or competitors by each type of industry is the key to securing competitiveness in the commercialization of the target company. In addition, transaction information, which becomes business activity between companies, plays an important role in providing information regarding potential customers when both parties enter similar fields together. Identifying a competitor at the enterprise or industry level by using a network map based on such inter-company sales information can be implemented as a core module of value chain analysis. The Value Chain Network System (VCNS) combines the concepts of value chain and industrial structure analysis with corporate information simply collected to date, so that it can grasp not only the market competition situation of individual companies but also the value chain relationship of a specific industry. Especially, it can be useful as an information analysis tool at the corporate level such as identification of industry structure, identification of competitor trends, analysis of competitors, locating suppliers (sellers) and demanders (buyers), industry trends by item, finding promising items, finding new entrants, finding core companies and items by value chain, and recognizing the patents with corresponding companies, etc. In addition, based on the objectivity and reliability of the analysis results from transaction deals information and financial data, it is expected that value chain network system will be utilized for various purposes such as information support for business evaluation, R&D decision support and mid-term or short-term demand forecasting, in particular to more than 15,000 member companies in Korea, employees in R&D service sectors government-funded research institutes and public organizations. In order to strengthen business competitiveness of companies, technology, patent and market information have been provided so far mainly by government agencies and private research-and-development service companies. This service has been presented in frames of patent analysis (mainly for rating, quantitative analysis) or market analysis (for market prediction and demand forecasting based on market reports). However, there was a limitation to solving the lack of information, which is one of the difficulties that firms in Korea often face in the stage of commercialization. In particular, it is much more difficult to obtain information about competitors and potential candidates. In this study, the real-time value chain analysis and visualization service module based on the proposed network map and the data in hands is compared with the expected market share, estimated sales volume, contact information (which implies potential suppliers for raw material / parts, and potential demanders for complete products / modules). In future research, we intend to carry out the in-depth research for further investigating the indices of competitive factors through participation of research subjects and newly developing competitive indices for competitors or substitute items, and to additively promoting with data mining techniques and algorithms for improving the performance of VCNS.