• Title/Summary/Keyword: Visual Sensor Network

Search Result 64, Processing Time 0.02 seconds

On Addressing Network Synchronization in Object Tracking with Multi-modal Sensors

  • Jung, Sang-Kil;Lee, Jin-Seok;Hong, Sang-Jin
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.3 no.4
    • /
    • pp.344-365
    • /
    • 2009
  • The performance of a tracking system is greatly increased if multiple types of sensors are combined to achieve the objective of the tracking instead of relying on single type of sensor. To conduct the multi-modal tracking, we have previously developed a multi-modal sensor-based tracking model where acoustic sensors mainly track the objects and visual sensors compensate the tracking errors [1]. In this paper, we find a network synchronization problem appearing in the developed tracking system. The problem is caused by the different location and traffic characteristics of multi-modal sensors and non-synchronized arrival of the captured sensor data at a processing server. To effectively deliver the sensor data, we propose a time-based packet aggregation algorithm where the acoustic sensor data are aggregated based on the sampling time and sent to the server. The delivered acoustic sensor data is then compensated by visual images to correct the tracking errors and such a compensation process improves the tracking accuracy in ideal case. However, in real situations, the tracking improvement from visual compensation can be severely degraded due to the aforementioned network synchronization problem, the impact of which is analyzed by simulations in this paper. To resolve the network synchronization problem, we differentiate the service level of sensor traffic based on Weight Round Robin (WRR) scheduling at the routers. The weighting factor allocated to each queue is calculated by a proposed Delay-based Weight Allocation (DWA) algorithm. From the simulations, we show the traffic differentiation model can mitigate the non-synchronization of sensor data. Finally, we analyze expected traffic behaviors of the tracking system in terms of acoustic sampling interval and visual image size.

An Efficient Implementation of Key Frame Extraction and Sharing in Android for Wireless Video Sensor Network

  • Kim, Kang-Wook
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.9 no.9
    • /
    • pp.3357-3376
    • /
    • 2015
  • Wireless sensor network is an important research topic that has attracted a lot of attention in recent years. However, most of the interest has focused on wireless sensor network to gather scalar data such as temperature, humidity and vibration. Scalar data are insufficient for diverse applications such as video surveillance, target recognition and traffic monitoring. However, if we use camera sensors in wireless sensor network to collect video data which are vast in information, they can provide important visual information. Video sensor networks continue to gain interest due to their ability to collect video information for a wide range of applications in the past few years. However, how to efficiently store the massive data that reflect environmental state of different times in video sensor network and how to quickly search interested information from them are challenging issues in current research, especially when the sensor network environment is complicated. Therefore, in this paper, we propose a fast algorithm for extracting key frames from video and describe the design and implementation of key frame extraction and sharing in Android for wireless video sensor network.

Simulation for sensor network packet routing protocols based on distributed-event (분산이벤트 기반 센서네트워크 패킷 라우팅 프로토콜 시뮬레이션)

  • Chung, Kyung-Yul;Lim, Byung-Ju;Lee, Hoo-Rock;Rhyu, Keel-Soo
    • Journal of Advanced Marine Engineering and Technology
    • /
    • v.37 no.4
    • /
    • pp.418-424
    • /
    • 2013
  • Both simulation and physical implementation are valuable tasks in sensor network routing protocols. In this paper, we propose an efficient underground utilities monitoring method within several constraints using wireless sensor network. First, in order to physically implement protocol of network which is applied, the distributed event-based simulation, which applies an existing nesC codes of sensor network routing protocols, is implemented and analyzed. Also, we have performed the simulation and analyzed the execution results for application model of routing protocols for monitoring underground utilities in the VIPTOS(Visual Ptolemy and TinyOS) environments which combine TOSSIM and Ptolemy II based on distributed event.

Visual Navigation by Neural Network Learning (신경망 학습에 의한 영상처리 네비게이션)

  • Shin, Suk-Young;Hoon Kang
    • Proceedings of the Korean Institute of Intelligent Systems Conference
    • /
    • 2001.12a
    • /
    • pp.263-266
    • /
    • 2001
  • It has been integrated into several navigation systems. This paper shows that system recognizes difficult indoor roads and open area without any specific mark such as painted guide line or tape. In this method, Robot navigates with visual sensors, which uses visual information to navigate itself along the road. An Artificial Neural Network System was used to decide where to move. It is designed with USB web camera as visual sensor.

  • PDF

Traffic Estimation Method for Visual Sensor Networks (비쥬얼 센서 네트워크에서 트래픽 예측 방법)

  • Park, Sang-Hyun
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.11 no.11
    • /
    • pp.1069-1076
    • /
    • 2016
  • Recent development in visual sensor technologies has encouraged various researches on adding imaging capabilities to sensor networks. Video data are bigger than other sensor data, so it is essential to manage the amount of image data efficiently. In this paper, a new method of video traffic estimation is proposed for efficient traffic management of visual sensor networks. In the proposed method, a first order autoregressive model is used for modeling the traffic with the consideration of the characteristics of video traffics acquired from visual sensors, and a Kalman filter algorithm is used to estimate the amount of video traffics. The proposed method is computationally simple, so it is proper to be applied to sensor nodes. It is shown by experimental results that the proposed method is simple but estimate the video traffics exactly by less than 1% of the average.

Simultaneous Localization and Mobile Robot Navigation using a Sensor Network

  • Jin Tae-Seok;Bashimoto Hideki
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.6 no.2
    • /
    • pp.161-166
    • /
    • 2006
  • Localization of mobile agent within a sensing network is a fundamental requirement for many applications, using networked navigating systems such as the sonar-sensing system or the visual-sensing system. To fully utilize the strengths of both the sonar and visual sensing systems, This paper describes a networked sensor-based navigation method in an indoor environment for an autonomous mobile robot which can navigate and avoid obstacle. In this method, the self-localization of the robot is done with a model-based vision system using networked sensors, and nonstop navigation is realized by a Kalman filter-based STSF(Space and Time Sensor Fusion) method. Stationary obstacles and moving obstacles are avoided with networked sensor data such as CCD camera and sonar ring. We will report on experiments in a hallway using the Pioneer-DX robot. In addition to that, the localization has inevitable uncertainties in the features and in the robot position estimation. Kalman filter scheme is used for the estimation of the mobile robot localization. And Extensive experiments with a robot and a sensor network confirm the validity of the approach.

Neural Network Approach to Sensor Fusion System for Improving the Recognition Performance of 3D Objects (3차원 물체의 인식 성능 향상을 위한 감각 융합 신경망 시스템)

  • Dong Sung Soo;Lee Chong Ho;Kim Ji Kyoung
    • The Transactions of the Korean Institute of Electrical Engineers D
    • /
    • v.54 no.3
    • /
    • pp.156-165
    • /
    • 2005
  • Human being recognizes the physical world by integrating a great variety of sensory inputs, the information acquired by their own action, and their knowledge of the world using hierarchically parallel-distributed mechanism. In this paper, authors propose the sensor fusion system that can recognize multiple 3D objects from 2D projection images and tactile informations. The proposed system focuses on improving recognition performance of 3D objects. Unlike the conventional object recognition system that uses image sensor alone, the proposed method uses tactual sensors in addition to visual sensor. Neural network is used to fuse the two sensory signals. Tactual signals are obtained from the reaction force of the pressure sensors at the fingertips when unknown objects are grasped by four-fingered robot hand. The experiment evaluates the recognition rate and the number of learning iterations of various objects. The merits of the proposed systems are not only the high performance of the learning ability but also the reliability of the system with tactual information for recognizing various objects even though the visual sensory signals get defects. The experimental results show that the proposed system can improve recognition rate and reduce teeming time. These results verify the effectiveness of the proposed sensor fusion system as recognition scheme for 3D objects.

Contrast Enhancement Method for Images from Visual Sensors (비주얼 센서 영상에 대한 대비 개선 방법)

  • Park, Sang-Hyun
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.13 no.3
    • /
    • pp.525-532
    • /
    • 2018
  • Recently, due to the advancements of sensor network technologies and camera technologies, there are increasing needs to effectively monitor the environment in a region that is difficult to access by using the visual sensor network that combines these two technologies. Since the image captured by the visual sensor reflects the natural phenomenon as it is, the quality of the image may deteriorate depending on the weather or time. In this paper, we propose an algorithm to improve the contrast of images using the characteristics of images obtained from visual sensors. In the proposed method, we first set the region of interest and then analyzes the change of the color value of the region of interest according to the brightness value of the image. The contrast of an image is improved by using the high contrast image of the same object and the analysis information. It is shown by experimental results that the proposed method improves the contrast of an image by restoring the color components of the low contrast image simply and accurately.

Development of Road-Following Controller for Autonomous Vehicle using Relative Similarity Modular Network (상대분할 신경회로망에 의한 자율주행차량 도로추적 제어기의 개발)

  • Ryoo, Young-Jae;Lim, Young-Cheol
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.5 no.5
    • /
    • pp.550-557
    • /
    • 1999
  • This paper describes a road-following controller using the proposed neural network for autonomous vehicle. Road-following with visual sensor like camera requires intelligent control algorithm because analysis of relation from road image to steering control is complex. The proposed neural network, relative similarity modular network(RSMN), is composed of some learning networks and a partitioniing network. The partitioning network divides input space into multiple sections by similarity of input data. Because divided section has simlar input patterns, RSMN can learn nonlinear relation such as road-following with visual control easily. Visual control uses two criteria on road image from camera; one is position of vanishing point of road, the other is slope of vanishing line of road. The controller using neural network has input of two criteria and output of steering angle. To confirm performance of the proposed neural network controller, a software is developed to simulate vehicle dynamics, camera image generation, visual control, and road-following. Also, prototype autonomous electric vehicle is developed, and usefulness of the controller is verified by physical driving test.

  • PDF

Prototype for the Weather Monitoring System with Web - Based Data Management - Construction and Operation

  • Kim, Jinwoo;Kim, Jin-Young;Oh, Jai-Ho;Kim, Do-Yong
    • Atmosphere
    • /
    • v.20 no.2
    • /
    • pp.153-160
    • /
    • 2010
  • In this paper, an attempt has been made to build and test self-configuring weather sensor networks and internet based observation system to gather atmospheric data. The aim is to provide integrated or real-time weather information in standard form using network data access protocol. This system was successfully developed to record weather information both digital as well as visual using sensor network and web-enabled surveillance cameras. These data were transformed by network based data access protocol to access and utilize for public domain. The competed system has been successfully utilized to monitor different types of weather. The results show that this is one of the most useful weather monitoring system.