• Title/Summary/Keyword: Multiple Sensor Systems

Search Result 373, Processing Time 0.021 seconds

Development of Insole for AI-Based Diagnosis of Diabetic Foot Ulcers in IoT Environment (IoT 환경에서 AI 기반의 당뇨발 진단을 위한 깔창 개발)

  • Choi, Won Hoo;Chung, Tai Myoung;Park, Ji Ung;Lee, Seo Hu
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.11 no.3
    • /
    • pp.83-90
    • /
    • 2022
  • Diabetes is a common disease today, and there are also many cases of developing into serious complications called Diabetic Foot Ulcers(DFU). Diagnosis and prevention of DFU in advance is an important task, and this paper proposes the method. Based on existing studies introduced in the paper, it can be seen that foot pressure and temperature information are deeply correlated with DFU. Introduce the process and architecture of SmarTinsole, an IoT device that measures these indicators. Also, the paper describes the preprocessing process for AI-based diagnosis of DFU. Through the comparison of the measured pressure graph and the actual human step distribution, it presents the results that multiple information collected in real-time from SmarTinsole are more efficient and reliable than the previous study.

Structural health monitoring data anomaly detection by transformer enhanced densely connected neural networks

  • Jun, Li;Wupeng, Chen;Gao, Fan
    • Smart Structures and Systems
    • /
    • v.30 no.6
    • /
    • pp.613-626
    • /
    • 2022
  • Guaranteeing the quality and integrity of structural health monitoring (SHM) data is very important for an effective assessment of structural condition. However, sensory system may malfunction due to sensor fault or harsh operational environment, resulting in multiple types of data anomaly existing in the measured data. Efficiently and automatically identifying anomalies from the vast amounts of measured data is significant for assessing the structural conditions and early warning for structural failure in SHM. The major challenges of current automated data anomaly detection methods are the imbalance of dataset categories. In terms of the feature of actual anomalous data, this paper proposes a data anomaly detection method based on data-level and deep learning technique for SHM of civil engineering structures. The proposed method consists of a data balancing phase to prepare a comprehensive training dataset based on data-level technique, and an anomaly detection phase based on a sophisticatedly designed network. The advanced densely connected convolutional network (DenseNet) and Transformer encoder are embedded in the specific network to facilitate extraction of both detail and global features of response data, and to establish the mapping between the highest level of abstractive features and data anomaly class. Numerical studies on a steel frame model are conducted to evaluate the performance and noise immunity of using the proposed network for data anomaly detection. The applicability of the proposed method for data anomaly classification is validated with the measured data of a practical supertall structure. The proposed method presents a remarkable performance on data anomaly detection, which reaches a 95.7% overall accuracy with practical engineering structural monitoring data, which demonstrates the effectiveness of data balancing and the robust classification capability of the proposed network.

CNN based data anomaly detection using multi-channel imagery for structural health monitoring

  • Shajihan, Shaik Althaf V.;Wang, Shuo;Zhai, Guanghao;Spencer, Billie F. Jr.
    • Smart Structures and Systems
    • /
    • v.29 no.1
    • /
    • pp.181-193
    • /
    • 2022
  • Data-driven structural health monitoring (SHM) of civil infrastructure can be used to continuously assess the state of a structure, allowing preemptive safety measures to be carried out. Long-term monitoring of large-scale civil infrastructure often involves data-collection using a network of numerous sensors of various types. Malfunctioning sensors in the network are common, which can disrupt the condition assessment and even lead to false-negative indications of damage. The overwhelming size of the data collected renders manual approaches to ensure data quality intractable. The task of detecting and classifying an anomaly in the raw data is non-trivial. We propose an approach to automate this task, improving upon the previously developed technique of image-based pre-processing on one-dimensional (1D) data by enriching the features of the neural network input data with multiple channels. In particular, feature engineering is employed to convert the measured time histories into a 3-channel image comprised of (i) the time history, (ii) the spectrogram, and (iii) the probability density function representation of the signal. To demonstrate this approach, a CNN model is designed and trained on a dataset consisting of acceleration records of sensors installed on a long-span bridge, with the goal of fault detection and classification. The effect of imbalance in anomaly patterns observed is studied to better account for unseen test cases. The proposed framework achieves high overall accuracy and recall even when tested on an unseen dataset that is much larger than the samples used for training, offering a viable solution for implementation on full-scale structures where limited labeled-training data is available.

Computer vision and deep learning-based post-earthquake intelligent assessment of engineering structures: Technological status and challenges

  • T. Jin;X.W. Ye;W.M. Que;S.Y. Ma
    • Smart Structures and Systems
    • /
    • v.31 no.4
    • /
    • pp.311-323
    • /
    • 2023
  • Ever since ancient times, earthquakes have been a major threat to the civil infrastructures and the safety of human beings. The majority of casualties in earthquake disasters are caused by the damaged civil infrastructures but not by the earthquake itself. Therefore, the efficient and accurate post-earthquake assessment of the conditions of structural damage has been an urgent need for human society. Traditional ways for post-earthquake structural assessment rely heavily on field investigation by experienced experts, yet, it is inevitably subjective and inefficient. Structural response data are also applied to assess the damage; however, it requires mounted sensor networks in advance and it is not intuitional. As many types of damaged states of structures are visible, computer vision-based post-earthquake structural assessment has attracted great attention among the engineers and scholars. With the development of image acquisition sensors, computing resources and deep learning algorithms, deep learning-based post-earthquake structural assessment has gradually shown potential in dealing with image acquisition and processing tasks. This paper comprehensively reviews the state-of-the-art studies of deep learning-based post-earthquake structural assessment in recent years. The conventional way of image processing and machine learning-based structural assessment are presented briefly. The workflow of the methodology for computer vision and deep learning-based post-earthquake structural assessment was introduced. Then, applications of assessment for multiple civil infrastructures are presented in detail. Finally, the challenges of current studies are summarized for reference in future works to improve the efficiency, robustness and accuracy in this field.

Experimental Implementation of a Cableless Seismic Data Acquisition Module Using Arduino (아두이노를 활용한 무선 탄성파 자료취득 모듈 구현 실험)

  • Chanil Kim;Sangin Cho;Sukjoon Pyun
    • Geophysics and Geophysical Exploration
    • /
    • v.26 no.3
    • /
    • pp.103-113
    • /
    • 2023
  • In the oil and gas exploration market, various cableless seismic systems have been developed as an alternative to improve data acquisition efficiency. However, developing such equipment at a small scale for academic research is not available owing to highly priced commercial products. Fortunately, building and experimenting with open-source hardware enable the academic utilization of cableless seismic equipment with relatively low cost. This study aims to develop a cableless seismic acquisition module using Arduino. A cableless seismic system requires the combination of signal sensing, simple pre-processing, and data storage in a single device. A conventional geophone is used as the sensor that detects the seismic wave signal. In addition, it is connected to an Arduino circuit that plays a role in implementing the processing and storing module for the detected signals. Three main functions are implemented in the Arduino module: preprocessing, A/D conversion, and data storage. The developed single-channel module can acquire a common receiver gather from multiple source experiments.

Research of the Delivery Autonomy and Vision-based Landing Algorithm for Last-Mile Service using a UAV (무인기를 이용한 Last-Mile 서비스를 위한 배송 자동화 및 영상기반 착륙 알고리즘 연구)

  • Hanseob Lee;Hoon Jung
    • Journal of Korean Society of Industrial and Systems Engineering
    • /
    • v.46 no.2
    • /
    • pp.160-167
    • /
    • 2023
  • This study focuses on the development of a Last-Mile delivery service using unmanned vehicles to deliver goods directly to the end consumer utilizing drones to perform autonomous delivery missions and an image-based precision landing algorithm for handoff to a robot in an intermediate facility. As the logistics market continues to grow rapidly, parcel volumes increase exponentially each year. However, due to low delivery fees, the workload of delivery personnel is increasing, resulting in a decrease in the quality of delivery services. To address this issue, the research team conducted a study on a Last-Mile delivery service using unmanned vehicles and conducted research on the necessary technologies for drone-based goods transportation in this paper. The flight scenario begins with the drone carrying the goods from a pickup location to the rooftop of a building where the final delivery destination is located. There is a handoff facility on the rooftop of the building, and a marker on the roof must be accurately landed upon. The mission is complete once the goods are delivered and the drone returns to its original location. The research team developed a mission planning algorithm to perform the above scenario automatically and constructed an algorithm to recognize the marker through a camera sensor and achieve a precision landing. The performance of the developed system has been verified through multiple trial operations within ETRI.

IoT-Based Automatic Water Quality Monitoring System with Optimized Neural Network

  • Anusha Bamini A M;Chitra R;Saurabh Agarwal;Hyunsung Kim;Punitha Stephan;Thompson Stephan
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.18 no.1
    • /
    • pp.46-63
    • /
    • 2024
  • One of the biggest dangers in the globe is water contamination. Water is a necessity for human survival. In most cities, the digging of borewells is restricted. In some cities, the borewell is allowed for only drinking water. Hence, the scarcity of drinking water is a vital issue for industries and villas. Most of the water sources in and around the cities are also polluted, and it will cause significant health issues. Real-time quality observation is necessary to guarantee a secure supply of drinking water. We offer a model of a low-cost system of monitoring real-time water quality using IoT to address this issue. The potential for supporting the real world has expanded with the introduction of IoT and other sensors. Multiple sensors make up the suggested system, which is utilized to identify the physical and chemical features of the water. Various sensors can measure the parameters such as temperature, pH, and turbidity. The core controller can process the values measured by sensors. An Arduino model is implemented in the core controller. The sensor data is forwarded to the cloud database using a WI-FI setup. The observed data will be transferred and stored in a cloud-based database for further processing. It wasn't easy to analyze the water quality every time. Hence, an Optimized Neural Network-based automation system identifies water quality from remote locations. The performance of the feed-forward neural network classifier is further enhanced with a hybrid GA- PSO algorithm. The optimized neural network outperforms water quality prediction applications and yields 91% accuracy. The accuracy of the developed model is increased by 20% because of optimizing network parameters compared to the traditional feed-forward neural network. Significant improvement in precision and recall is also evidenced in the proposed work.

Status and Prospect of Unmanned, Global Ocean Observations Network (글로벌 무인해양관측 네트워크 현황과 전망)

  • Nam, Sunghyun;Kim, Yun-Bae;Park, Jong Jin;Chang, Kyung-Il
    • The Sea:JOURNAL OF THE KOREAN SOCIETY OF OCEANOGRAPHY
    • /
    • v.19 no.3
    • /
    • pp.202-214
    • /
    • 2014
  • We introduce status and prospect of increasingly utilizing, unmanned, global ocean observing systems, and the global network to integrate, coordinate, and manage the systems. Platforms of the ocean observing system are diversified in order to resolve/monitor the variability occurring at multiple scales in both three-dimensional space and time. Here purpose, development history, and current status of the systems in two kinds - mobile (surface drifter, subsurface float, underwater glider) and fixed platforms (surface and subsurface moorings, bottom mounts), are examined and the increased future uses to produce synergies are envisioned. Simultaneous use of various mobile and fixed platforms is suggested to more effectively design the observing system, with an example of the NSF-funded OOI (Ocean Observations Initiative) program. Efforts are suggested 1) to fill the data gap existing in the deep sea and the Southern Ocean, and toward 2) new global network for oceanic boundary currents, 3) new technologies for existing and new sensors including biogeochemical, acoustic, and optical sensors, 3) data standardization, and 4) sensor calibration and data quality control.

Energy-Efficient Division Protocol for Mobile Sink Groups in Wireless Sensor Network (무선 센서 네트워크에서 이동 싱크 그룹의 분리를 지원하기 위한 라우팅 프로토콜)

  • Jang, Jaeyoung;Lee, Euisin
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.6 no.1
    • /
    • pp.1-8
    • /
    • 2017
  • Communications for mobile sink groups such as rescue teams or platoons bring about a new challenging issue for handling mobility in wireless sensor networks. To do this, many studies have been proposed to support mobile sink groups. When closely looking at mobile sink groups, they can be divided into (multiple) small groups according to the property of applications. For example, a platoon can be divided into multiple squads to carry out its mission in the battle field. However, the previous studies cannot efficiently support the division of mobile sink groups because they do not address three challenging issues engendered by the mobile sink group division. The first issue is to select a leader sink for a new small mobile sink group. The efficient data delivery from a source to small mobile sink groups is the second issue. Last, the third issue is to share data between leader sinks of small mobile sink groups. Thus, this paper proposes a routing protocol to efficiently support the division of mobile sink groups by solving the three challenging issues. For the first issue, the proposed protocol selects a leader sink of a new small mobile sink group which provide a minimum summation of the distance between the new leader sink and the previous leader sink and the distance from the new leader sink to all of its member sinks. For the efficient data delivery from a source to small mobile sink groups in the second issue, the proposed protocol determines the path to minimize the data dissemination distance from source to small mobile sink group by calculating with the location information of both the source and the leader sinks. With regard to the third issue, the proposed protocol exploits member sinks located among leader sinks to provide efficient data sharing among leaders sinks by considering the location information of member sinks. Simulation results verified that the proposed protocol is superior to the previous protocol in terms of the energy consumption.

The Individual Discrimination Location Tracking Technology for Multimodal Interaction at the Exhibition (전시 공간에서 다중 인터랙션을 위한 개인식별 위치 측위 기술 연구)

  • Jung, Hyun-Chul;Kim, Nam-Jin;Choi, Lee-Kwon
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.2
    • /
    • pp.19-28
    • /
    • 2012
  • After the internet era, we are moving to the ubiquitous society. Nowadays the people are interested in the multimodal interaction technology, which enables audience to naturally interact with the computing environment at the exhibitions such as gallery, museum, and park. Also, there are other attempts to provide additional service based on the location information of the audience, or to improve and deploy interaction between subjects and audience by analyzing the using pattern of the people. In order to provide multimodal interaction service to the audience at the exhibition, it is important to distinguish the individuals and trace their location and route. For the location tracking on the outside, GPS is widely used nowadays. GPS is able to get the real time location of the subjects moving fast, so this is one of the important technologies in the field requiring location tracking service. However, as GPS uses the location tracking method using satellites, the service cannot be used on the inside, because it cannot catch the satellite signal. For this reason, the studies about inside location tracking are going on using very short range communication service such as ZigBee, UWB, RFID, as well as using mobile communication network and wireless lan service. However these technologies have shortcomings in that the audience needs to use additional sensor device and it becomes difficult and expensive as the density of the target area gets higher. In addition, the usual exhibition environment has many obstacles for the network, which makes the performance of the system to fall. Above all these things, the biggest problem is that the interaction method using the devices based on the old technologies cannot provide natural service to the users. Plus the system uses sensor recognition method, so multiple users should equip the devices. Therefore, there is the limitation in the number of the users that can use the system simultaneously. In order to make up for these shortcomings, in this study we suggest a technology that gets the exact location information of the users through the location mapping technology using Wi-Fi and 3d camera of the smartphones. We applied the signal amplitude of access point using wireless lan, to develop inside location tracking system with lower price. AP is cheaper than other devices used in other tracking techniques, and by installing the software to the user's mobile device it can be directly used as the tracking system device. We used the Microsoft Kinect sensor for the 3D Camera. Kinect is equippedwith the function discriminating the depth and human information inside the shooting area. Therefore it is appropriate to extract user's body, vector, and acceleration information with low price. We confirm the location of the audience using the cell ID obtained from the Wi-Fi signal. By using smartphones as the basic device for the location service, we solve the problems of additional tagging device and provide environment that multiple users can get the interaction service simultaneously. 3d cameras located at each cell areas get the exact location and status information of the users. The 3d cameras are connected to the Camera Client, calculate the mapping information aligned to each cells, get the exact information of the users, and get the status and pattern information of the audience. The location mapping technique of Camera Client decreases the error rate that occurs on the inside location service, increases accuracy of individual discrimination in the area through the individual discrimination based on body information, and establishes the foundation of the multimodal interaction technology at the exhibition. Calculated data and information enables the users to get the appropriate interaction service through the main server.