• 제목/요약/키워드: multi-time scale

Search Result 522, Processing Time 0.031 seconds

Analysis of Spatial Characteristics of Old Building Districts to Evaluate Fire Risk Factors (화재 위험요소의 도출을 위한 노후건축지구의 공간구성 특성분석)

  • Son, Byeung-Hun;Kang, Kyung-Ha;Ryu, Jung-Rim;Roh, Seung-Jun
    • Journal of the Korea Institute of Building Construction
    • /
    • v.22 no.1
    • /
    • pp.69-80
    • /
    • 2022
  • The proportion of buildings over 30 years old in Korea has increased, from 29.0% in 2005 to 37.8% in 2019. These old buildings were built during a time in which there was a lack of building-related safety standards in areas such as fire safety performance. Worse, during their years of use, many such old buildings have had illegal changes and extensions made, making them more vulnerable in terms of safety. Fire safety investigations are being conducted to prevent large-scale disasters in multi-use buildings, but no investigation has been conducted at the regional district level, where small-scale old buildings are concentrated. Therefore, to identify fire risk factors in the old building district where old buildings are concentrated, the composition characteristics of the buildings were first analyzed. To examine the spatial characteristics of old building districts in order to derive fire risk factors, the results of this analysis based on the structure, use, roof type, and year of approval for use are as follows. It was found through our analysis that as of the date of approval for the use of the building, the main structure of the building has the greatest impact.

A Cluster Based Energy Efficient Tree Routing Protocol in Wireless Sensor Networks (광역 WSN 을 위한 클러스팅 트리 라우팅 프로토콜)

  • Nurhayati, Nurhayati;Choi, Sung-Hee;Lee, Kyung-Oh
    • Annual Conference of KIPS
    • /
    • 2011.04a
    • /
    • pp.576-579
    • /
    • 2011
  • Wireless sensor network are widely all over different fields. Because of its distinguished characteristics, we must take account of the factor of energy consumed when designing routing protocol. Wireless sensor networks consist of small battery powered devices with limited energy resources. Once deployed, the small sensor nodes are usually inaccessible to the user, and thus replacement of the energy source is not feasible. Hence, energy efficiency is a key design issue that needs to be enhanced in order to improve the life span of the network. In BCDCP, all sensors sends data from the CH (Cluster Head) and then to the BS (Base Station). BCDCP works well in a smallscale network however is not preferred in a large scale network since it uses much energy for long distance wireless communication. TBRP can be used for large scale network, but it weakness lies on the fact that the nodedry out of energy easily since it uses multi-hops transmission data to the Base Station. Here, we proposed a routing protocol. A Cluster Based Energy Efficient Tree Routing Protocol (CETRP) in Wireless Sensor Networks (WSNs) to prolong network life time through the balanced energy consumption. CETRP selects Cluster Head of cluster tree shape and uses maximum two hops data transmission to the Cluster Head in every level. We show CETRP outperforms BCDCP and TBRP with several experiments.

The way to make training data for deep learning model to recognize keywords in product catalog image at E-commerce (온라인 쇼핑몰에서 상품 설명 이미지 내의 키워드 인식을 위한 딥러닝 훈련 데이터 자동 생성 방안)

  • Kim, Kitae;Oh, Wonseok;Lim, Geunwon;Cha, Eunwoo;Shin, Minyoung;Kim, Jongwoo
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.1
    • /
    • pp.1-23
    • /
    • 2018
  • From the 21st century, various high-quality services have come up with the growth of the internet or 'Information and Communication Technologies'. Especially, the scale of E-commerce industry in which Amazon and E-bay are standing out is exploding in a large way. As E-commerce grows, Customers could get what they want to buy easily while comparing various products because more products have been registered at online shopping malls. However, a problem has arisen with the growth of E-commerce. As too many products have been registered, it has become difficult for customers to search what they really need in the flood of products. When customers search for desired products with a generalized keyword, too many products have come out as a result. On the contrary, few products have been searched if customers type in details of products because concrete product-attributes have been registered rarely. In this situation, recognizing texts in images automatically with a machine can be a solution. Because bulk of product details are written in catalogs as image format, most of product information are not searched with text inputs in the current text-based searching system. It means if information in images can be converted to text format, customers can search products with product-details, which make them shop more conveniently. There are various existing OCR(Optical Character Recognition) programs which can recognize texts in images. But existing OCR programs are hard to be applied to catalog because they have problems in recognizing texts in certain circumstances, like texts are not big enough or fonts are not consistent. Therefore, this research suggests the way to recognize keywords in catalog with the Deep Learning algorithm which is state of the art in image-recognition area from 2010s. Single Shot Multibox Detector(SSD), which is a credited model for object-detection performance, can be used with structures re-designed to take into account the difference of text from object. But there is an issue that SSD model needs a lot of labeled-train data to be trained, because of the characteristic of deep learning algorithms, that it should be trained by supervised-learning. To collect data, we can try labelling location and classification information to texts in catalog manually. But if data are collected manually, many problems would come up. Some keywords would be missed because human can make mistakes while labelling train data. And it becomes too time-consuming to collect train data considering the scale of data needed or costly if a lot of workers are hired to shorten the time. Furthermore, if some specific keywords are needed to be trained, searching images that have the words would be difficult, as well. To solve the data issue, this research developed a program which create train data automatically. This program can make images which have various keywords and pictures like catalog and save location-information of keywords at the same time. With this program, not only data can be collected efficiently, but also the performance of SSD model becomes better. The SSD model recorded 81.99% of recognition rate with 20,000 data created by the program. Moreover, this research had an efficiency test of SSD model according to data differences to analyze what feature of data exert influence upon the performance of recognizing texts in images. As a result, it is figured out that the number of labeled keywords, the addition of overlapped keyword label, the existence of keywords that is not labeled, the spaces among keywords and the differences of background images are related to the performance of SSD model. This test can lead performance improvement of SSD model or other text-recognizing machine based on deep learning algorithm with high-quality data. SSD model which is re-designed to recognize texts in images and the program developed for creating train data are expected to contribute to improvement of searching system in E-commerce. Suppliers can put less time to register keywords for products and customers can search products with product-details which is written on the catalog.

Product Recommender Systems using Multi-Model Ensemble Techniques (다중모형조합기법을 이용한 상품추천시스템)

  • Lee, Yeonjeong;Kim, Kyoung-Jae
    • Journal of Intelligence and Information Systems
    • /
    • v.19 no.2
    • /
    • pp.39-54
    • /
    • 2013
  • Recent explosive increase of electronic commerce provides many advantageous purchase opportunities to customers. In this situation, customers who do not have enough knowledge about their purchases, may accept product recommendations. Product recommender systems automatically reflect user's preference and provide recommendation list to the users. Thus, product recommender system in online shopping store has been known as one of the most popular tools for one-to-one marketing. However, recommender systems which do not properly reflect user's preference cause user's disappointment and waste of time. In this study, we propose a novel recommender system which uses data mining and multi-model ensemble techniques to enhance the recommendation performance through reflecting the precise user's preference. The research data is collected from the real-world online shopping store, which deals products from famous art galleries and museums in Korea. The data initially contain 5759 transaction data, but finally remain 3167 transaction data after deletion of null data. In this study, we transform the categorical variables into dummy variables and exclude outlier data. The proposed model consists of two steps. The first step predicts customers who have high likelihood to purchase products in the online shopping store. In this step, we first use logistic regression, decision trees, and artificial neural networks to predict customers who have high likelihood to purchase products in each product group. We perform above data mining techniques using SAS E-Miner software. In this study, we partition datasets into two sets as modeling and validation sets for the logistic regression and decision trees. We also partition datasets into three sets as training, test, and validation sets for the artificial neural network model. The validation dataset is equal for the all experiments. Then we composite the results of each predictor using the multi-model ensemble techniques such as bagging and bumping. Bagging is the abbreviation of "Bootstrap Aggregation" and it composite outputs from several machine learning techniques for raising the performance and stability of prediction or classification. This technique is special form of the averaging method. Bumping is the abbreviation of "Bootstrap Umbrella of Model Parameter," and it only considers the model which has the lowest error value. The results show that bumping outperforms bagging and the other predictors except for "Poster" product group. For the "Poster" product group, artificial neural network model performs better than the other models. In the second step, we use the market basket analysis to extract association rules for co-purchased products. We can extract thirty one association rules according to values of Lift, Support, and Confidence measure. We set the minimum transaction frequency to support associations as 5%, maximum number of items in an association as 4, and minimum confidence for rule generation as 10%. This study also excludes the extracted association rules below 1 of lift value. We finally get fifteen association rules by excluding duplicate rules. Among the fifteen association rules, eleven rules contain association between products in "Office Supplies" product group, one rules include the association between "Office Supplies" and "Fashion" product groups, and other three rules contain association between "Office Supplies" and "Home Decoration" product groups. Finally, the proposed product recommender systems provides list of recommendations to the proper customers. We test the usability of the proposed system by using prototype and real-world transaction and profile data. For this end, we construct the prototype system by using the ASP, Java Script and Microsoft Access. In addition, we survey about user satisfaction for the recommended product list from the proposed system and the randomly selected product lists. The participants for the survey are 173 persons who use MSN Messenger, Daum Caf$\acute{e}$, and P2P services. We evaluate the user satisfaction using five-scale Likert measure. This study also performs "Paired Sample T-test" for the results of the survey. The results show that the proposed model outperforms the random selection model with 1% statistical significance level. It means that the users satisfied the recommended product list significantly. The results also show that the proposed system may be useful in real-world online shopping store.

Development and Performance Evaluation of Multi-sensor Module for Use in Disaster Sites of Mobile Robot (조사로봇의 재난현장 활용을 위한 다중센서모듈 개발 및 성능평가에 관한 연구)

  • Jung, Yonghan;Hong, Junwooh;Han, Soohee;Shin, Dongyoon;Lim, Eontaek;Kim, Seongsam
    • Korean Journal of Remote Sensing
    • /
    • v.38 no.6_3
    • /
    • pp.1827-1836
    • /
    • 2022
  • Disasters that occur unexpectedly are difficult to predict. In addition, the scale and damage are increasing compared to the past. Sometimes one disaster can develop into another disaster. Among the four stages of disaster management, search and rescue are carried out in the response stage when an emergency occurs. Therefore, personnel such as firefighters who are put into the scene are put in at a lot of risk. In this respect, in the initial response process at the disaster site, robots are a technology with high potential to reduce damage to human life and property. In addition, Light Detection And Ranging (LiDAR) can acquire a relatively wide range of 3D information using a laser. Due to its high accuracy and precision, it is a very useful sensor when considering the characteristics of a disaster site. Therefore, in this study, development and experiments were conducted so that the robot could perform real-time monitoring at the disaster site. Multi-sensor module was developed by combining LiDAR, Inertial Measurement Unit (IMU) sensor, and computing board. Then, this module was mounted on the robot, and a customized Simultaneous Localization and Mapping (SLAM) algorithm was developed. A method for stably mounting a multi-sensor module to a robot to maintain optimal accuracy at disaster sites was studied. And to check the performance of the module, SLAM was tested inside the disaster building, and various SLAM algorithms and distance comparisons were performed. As a result, PackSLAM developed in this study showed lower error compared to other algorithms, showing the possibility of application in disaster sites. In the future, in order to further enhance usability at disaster sites, various experiments will be conducted by establishing a rough terrain environment with many obstacles.

Acoustic technology-assisted rapid proteolysis for high-throughput proteome analysis (대량 발굴 프로테옴 분석을 위한 어쿠스틱 기술 기반 고속 단백질 절편화)

  • Kim, Bo-Ra;Huyen, Trang Tran;Han, Na-Young;Park, Jong-Moon;Yu, Ung-Sik;Lee, Hoo-Keun
    • Analytical Science and Technology
    • /
    • v.24 no.6
    • /
    • pp.510-518
    • /
    • 2011
  • Recent developments and improvements of multiple technological elements including mass spectrometry (MS) instrument, multi-dimensional chromatographic separation, and software tools processing MS data resulted in benefits of large scale proteomics analysis. However, its throughput is limited by the speed and reproducibility of the protein digestion process. In this study, we demonstrated a new method for rapid proteolytic digestion of proteins using acoustic technology. Tryptic digests of BSA prepared at various conditions by super acoustic for optimization time and intensity were analyzed by LC-MS/MS showed higher sequence coverage in compared with traditional 16 hrs digestion method. The method was applied successfully for complex proteins of a breast cancer cells at 30 min of digestion at intensity 2. This new application reduces time-consuming of sample preparation with better efficiency, even with large amount of proteins, and increases high-throughput process in sample preparation state.

Removal of Nitrate-Nitrogen in Pickling Acid Wastewater from Stainless Steel Industry Using Electrodialysis and Ion Exchange Resin (전기투석과 이온교환수지를 이용한 스테인레스 산업의 산세폐수 내 질산성 질소의 제거)

  • Yun, Young-Ki;Park, Yeon-Jin;Oh, Sang-Hwa;Shin, Won-Sik;Choi, Sang-June;Ryu, Seung-Ki
    • Journal of Environmental Science International
    • /
    • v.18 no.6
    • /
    • pp.645-654
    • /
    • 2009
  • Lab-scale Electrodialysis(ED) system with different membranes combined with before or after pyroma process were carried out to remove nitrate from two pickling acid wastewater containing high concentrations of $NO_3\;^-$(${\approx}$150,000 mg/L) and F($({\approx}$ 160,000 mg/L) and some heavy metals(Fe, Ti, and Cr). The ED system before Pyroma process(Sample A) was not successful in $NO_3\;^-$ removal due to cation membrane fouling by the heavy metals, whereas, in the ED system after Pyroma process(Sample B), about 98% of nitrate was removed because of relatively low $NO_3\;^-$ concentration (about 30,000 mg/L) and no heavy metals. Mono-selective membranes(CIMS/ACS) in ED system have no selectivity for nitrate compared to divalent-selective membranes(CMX/AMX). The operation time for nitrate removal time decreased with increasing the applied voltage from 10V to 15V with no difference in the nitrate removal rate between both voltages. Nitrate adsorption of a strong-base anion exchange resin of $Cl\;^-$ type was also conducted. The Freundlich model($R^2$ > 0.996) was fitted better than Langmuir mode($R^2$ > 0.984) to the adsorption data. The maximum adsorption capacity ($Q^0$) was 492 mg/g for Sample A and 111 mg/g for Sample B due to the difference in initial nitrate concentrations between the two wastewater samples. In the regeneration of ion exchange resins, the nitrate removal rate in the pickling acid wastewater decreased as the adsorption step was repeated because certain amount of adsorbed $NO_3\;^-$ remained in the resins in spite of several desorption steps for regeneration. In conclusion, the optimum system configuration to treat pickling acid wastewater from stainless-steel industry is the multi-processes of the Pyroma-Electrodialysis-Ion exchange.

Use of a Temperature as a Tracer to Study Stream-groundwater Exchange in the Hyporheic Zone (열추적자를 이용한 지하수-하천수 혼합대 연구)

  • Kim, Kue-Young;Chon, Chul-Min;Kim, Tae-Hee;Oh, Jun-Ho;Jeoung, Jae-Hoon;Park, Seung-Ki
    • Economic and Environmental Geology
    • /
    • v.39 no.5 s.180
    • /
    • pp.525-535
    • /
    • 2006
  • A study on stream-groundwater exchange was performed using head and temperature data of stream water, streambed, and groundwater. Groundwater level and temperature were obtained from multi-depth monitoring wells in small-scale watershed. During the summer and winter season, time series of temperature data at streambed and groundwater were monitored for six months. In the winter time, we measured the temperature gradient between stream water and streambed. The observed data showed three typical types of temperature characteristics. First, the temperature of streambed was lower than that of stream water; second, the temperature of streambed and stream water was similar; and the last, the temperature of streambed was higher than that of stream water. The interconnections between the stream and the streambed were not homogeneously distributed due to weakly developed sediments and heterogeneous bedrock exposed as bed of the stream. The temperature data may be used in formal solutions of the inverse problems to estimate groundwater flow and hydraulic conductivity.

Research on ITB Contract Terms Classification Model for Risk Management in EPC Projects: Deep Learning-Based PLM Ensemble Techniques (EPC 프로젝트의 위험 관리를 위한 ITB 문서 조항 분류 모델 연구: 딥러닝 기반 PLM 앙상블 기법 활용)

  • Hyunsang Lee;Wonseok Lee;Bogeun Jo;Heejun Lee;Sangjin Oh;Sangwoo You;Maru Nam;Hyunsik Lee
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.12 no.11
    • /
    • pp.471-480
    • /
    • 2023
  • The Korean construction order volume in South Korea grew significantly from 91.3 trillion won in public orders in 2013 to a total of 212 trillion won in 2021, particularly in the private sector. As the size of the domestic and overseas markets grew, the scale and complexity of EPC (Engineering, Procurement, Construction) projects increased, and risk management of project management and ITB (Invitation to Bid) documents became a critical issue. The time granted to actual construction companies in the bidding process following the EPC project award is not only limited, but also extremely challenging to review all the risk terms in the ITB document due to manpower and cost issues. Previous research attempted to categorize the risk terms in EPC contract documents and detect them based on AI, but there were limitations to practical use due to problems related to data, such as the limit of labeled data utilization and class imbalance. Therefore, this study aims to develop an AI model that can categorize the contract terms based on the FIDIC Yellow 2017(Federation Internationale Des Ingenieurs-Conseils Contract terms) standard in detail, rather than defining and classifying risk terms like previous research. A multi-text classification function is necessary because the contract terms that need to be reviewed in detail may vary depending on the scale and type of the project. To enhance the performance of the multi-text classification model, we developed the ELECTRA PLM (Pre-trained Language Model) capable of efficiently learning the context of text data from the pre-training stage, and conducted a four-step experiment to validate the performance of the model. As a result, the ensemble version of the self-developed ITB-ELECTRA model and Legal-BERT achieved the best performance with a weighted average F1-Score of 76% in the classification of 57 contract terms.

A Study on Design of Agent based Nursing Records System in Attending System (에이전트기반 개방병원 간호기록시스템 설계에 관한 연구)

  • Kim, Kyoung-Hwan
    • Journal of Intelligence and Information Systems
    • /
    • v.16 no.2
    • /
    • pp.73-94
    • /
    • 2010
  • The attending system is a medical system that allows doctors in clinics to use the extra equipment in hospitals-beds, laboratory, operating room, etc-for their patient's care under a contract between the doctors and hospitals. Therefore, the system is very beneficial in terms of the efficiency of the usage of medical resources. However, it is necessary to develop a strong support system to strengthen its weaknesses and supplement its merits. If doctors use hospital beds under the attending system of hospitals, they would be able to check a patient's condition often and provide them with nursing care services. However, the current attending system lacks delivery and assistance support. Thus, for the successful performance of the attending system, a networking system should be developed to facilitate communication between the doctors and nurses. In particular, the nursing records in the attending system could help doctors monitor the patient's condition and provision of nursing care services. A nursing record is the formal documentation associated with nursing care. It is merely a data repository that helps nurses to track their activities; nursing records thus represent a resource of primary information that can be reused. In order to maximize their usefulness, nursing records have been introduced as part of computerized patient records. However, nursing records are internal data that are not disclosed by hospitals. Moreover, the lack of standardization of the record list makes it difficult to share nursing records. Under the attending system, nurses would want to minimize the amount of effort they have to put in for the maintenance of additional records. Hence, they would try to maintain the current level of nursing records in the form of record lists and record attributes, while doctors would require more detailed and real-time information about their patients in order to monitor their condition. Therefore, this study developed a system for assisting in the maintenance and sharing of the nursing records under the attending system. In contrast to previous research on the functionality of computer-based nursing records, we have emphasized the practical usefulness of nursing records from the viewpoint of the actual implementation of the attending system. We suggested that nurses could design a nursing record dictionary for their convenience, and that doctors and nurses could confirm the definitions that they looked up in the dictionary through negotiations with intelligent agents. Such an agent-based system could facilitate networking among medical institutes. Multi-agent systems are a widely accepted paradigm for the distribution and sharing of computation workloads in the scientific community. Agent-based systems have been developed with differences in functional cooperation, coordination, and negotiation. To increase such communication, a framework for a multi-agent based system is proposed in this study. The agent-based approach is useful for developing a system that promotes trade-offs between transactions involving multiple attributes. A brief summary of our contributions follows. First, we propose an efficient and accurate utility representation and acquisition mechanism based on a preference scale while minimizing user interactions with the agent. Trade-offs between various transaction attributes can also be easily computed. Second, by providing a multi-attribute negotiation framework based on the attribute utility evaluation mechanism, we allow both the doctors in charge and nurses to negotiate over various transaction attributes in the nursing record lists that are defined by the latter. Third, we have designed the architecture of the nursing record management server and a system of agents that provides support to the doctors and nurses with regard to the framework and mechanisms proposed above. A formal protocol has also been developed to create and control the communication required for negotiations. We verified the realization of the system by developing a web-based prototype. The system was implemented using ASP and IIS5.1.