• Title/Summary/Keyword: camera image

Search Result 4,918, Processing Time 0.043 seconds

Realtime 3D Human Full-Body Convergence Motion Capture using a Kinect Sensor (Kinect Sensor를 이용한 실시간 3D 인체 전신 융합 모션 캡처)

  • Kim, Sung-Ho
    • Journal of Digital Convergence
    • /
    • v.14 no.1
    • /
    • pp.189-194
    • /
    • 2016
  • Recently, there is increasing demand for image processing technology while activated the use of equipments such as camera, camcorder and CCTV. In particular, research and development related to 3D image technology using the depth camera such as Kinect sensor has been more activated. Kinect sensor is a high-performance camera that can acquire a 3D human skeleton structure via a RGB, skeleton and depth image in real-time frame-by-frame. In this paper, we develop a system. This system captures the motion of a 3D human skeleton structure using the Kinect sensor. And this system can be stored by selecting the motion file format as trc and bvh that is used for general purposes. The system also has a function that converts TRC motion captured format file into BVH format. Finally, this paper confirms visually through the motion capture data viewer that motion data captured using the Kinect sensor is captured correctly.

Tracking of Walking Human Based on Position Uncertainty of Dynamic Vision Sensor of Quadcopter UAV (UAV기반 동적영상센서의 위치불확실성을 통한 보행자 추정)

  • Lee, Junghyun;Jin, Taeseok
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.22 no.1
    • /
    • pp.24-30
    • /
    • 2016
  • The accuracy of small and low-cost CCD cameras is insufficient to provide data for precisely tracking unmanned aerial vehicles (UAVs). This study shows how a quad rotor UAV can hover on a human targeted tracking object by using data from a CCD camera rather than imprecise GPS data. To realize this, quadcopter UAVs need to recognize their position and posture in known environments as well as unknown environments. Moreover, it is necessary for their localization to occur naturally. It is desirable for UAVs to estimate their position by solving uncertainty for quadcopter UAV hovering, as this is one of the most important problems. In this paper, we describe a method for determining the altitude of a quadcopter UAV using image information of a moving object like a walking human. This method combines the observed position from GPS sensors and the estimated position from images captured by a fixed camera to localize a UAV. Using the a priori known path of a quadcopter UAV in the world coordinates and a perspective camera model, we derive the geometric constraint equations that represent the relation between image frame coordinates for a moving object and the estimated quadcopter UAV's altitude. Since the equations are based on the geometric constraint equation, measurement error may exist all the time. The proposed method utilizes the error between the observed and estimated image coordinates to localize the quadcopter UAV. The Kalman filter scheme is applied for this method. Its performance is verified by a computer simulation and experiments.

Biomimetic approach object detection sensors using multiple imaging (다중 영상을 이용한 생체모방형 물체 접근 감지 센서)

  • Choi, Myoung Hoon;Kim, Min;Jeong, Jae-Hoon;Park, Won-Hyeon;Lee, Dong Heon;Byun, Gi-Sik;Kim, Gwan-Hyung
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2016.05a
    • /
    • pp.91-93
    • /
    • 2016
  • From the 2-D image extracting three-dimensional information as the latter is in the bilateral sibeop using two camera method and when using a monocular camera as a very important step generally as "stereo vision". There in today's CCTV and automatic object tracking system used in many medium much to know the site conditions or work developed more clearly by using a stereo camera that mimics the eyes of humans to maximize the efficiency of avoidance / control start and multiple jobs can do. Object tracking system of the existing 2D image will have but can not recognize the distance to the transition could not be recognized by the observer display using a parallax of a stereo image, and the object can be more effectively controlled.

  • PDF

Real-time monitoring system with Kinect v2 using notifications on mobile devices (Kinect V2를 이용한 모바일 장치 실시간 알림 모니터링 시스템)

  • Eric, Niyonsaba;Jang, Jong Wook
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2016.05a
    • /
    • pp.277-280
    • /
    • 2016
  • Real-time remote monitoring system has an important value in many surveillance situations. It allows someone to be informed of what is happening in his monitoring locations. Kinect v2 is a new kind of camera which gives computers eyes and can generate different data such as color and depth images, audio input and skeletal data. In this paper, using Kinect v2 sensor with its depth image, we present a monitoring system in a space covered by Kinect. Therefore, based on space covered by Kinect camera, we define a target area to monitor using depth range by setting minimum and maximum distances. With computer vision library (Emgu CV), if there is an object tracked in the target space, kinect camera captures the whole image color and sends it in database and user gets at the same time a notification on his mobile device wherever he is with internet access.

  • PDF

Real-time Eye Contact System Using a Kinect Depth Camera for Realistic Telepresence (Kinect 깊이 카메라를 이용한 실감 원격 영상회의의 시선 맞춤 시스템)

  • Lee, Sang-Beom;Ho, Yo-Sung
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.37 no.4C
    • /
    • pp.277-282
    • /
    • 2012
  • In this paper, we present a real-time eye contact system for realistic telepresence using a Kinect depth camera. In order to generate the eye contact image, we capture a pair of color and depth video. Then, the foreground single user is separated from the background. Since the raw depth data includes several types of noises, we perform a joint bilateral filtering method. We apply the discontinuity-adaptive depth filter to the filtered depth map to reduce the disocclusion area. From the color image and the preprocessed depth map, we construct a user mesh model at the virtual viewpoint. The entire system is implemented through GPU-based parallel programming for real-time processing. Experimental results have shown that the proposed eye contact system is efficient in realizing eye contact, providing the realistic telepresence.

An Accurate Moving Distance Measurement Using the Rear-View Images in Parking Assistant Systems (후방영상 기반 주차 보조 시스템에서 정밀 이동거리 추출 기법)

  • Kim, Ho-Young;Lee, Seong-Won
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.37C no.12
    • /
    • pp.1271-1280
    • /
    • 2012
  • In the recent parking assistant systems, finding out the distance to the object behind a car is often performed by the range sensors such as ultrasonic sensors, radars. However, the installation of additional sensors on the used vehicle could be difficult and require extra cost. On the other hand, the motion stereo technique that extracts distance information using only an image sensor was also proposed. However, In the stereo rectification step, the motion stereo requires good features and exacts matching result. In this paper, we propose a fast algorithm that extracts the accurate distance information for the parallel parking situation using the consecutive images that is acquired by a rear-view camera. The proposed algorithm uses the quadrangle transform of the image, the horizontal line integral projection, and the blocking-based correlation measurement. In the experiment with the magna parallel test sequence, the result shows that the line-accurate distance measurement with the image sequence from the rear-view camera is possible.

Image Based Human Action Recognition System to Support the Blind (시각장애인 보조를 위한 영상기반 휴먼 행동 인식 시스템)

  • Ko, ByoungChul;Hwang, Mincheol;Nam, Jae-Yeal
    • Journal of KIISE
    • /
    • v.42 no.1
    • /
    • pp.138-143
    • /
    • 2015
  • In this paper we develop a novel human action recognition system based on communication between an ear-mounted Bluetooth camera and an action recognition server to aid scene recognition for the blind. First, if the blind capture an image of a specific location using the ear-mounted camera, the captured image is transmitted to the recognition server using a smartphone that is synchronized with the camera. The recognition server sequentially performs human detection, object detection and action recognition by analyzing human poses. The recognized action information is retransmitted to the smartphone and the user can hear the action information through the text-to-speech (TTS). Experimental results using the proposed system showed a 60.7% action recognition performance on the test data captured in indoor and outdoor environments.

Design of a scintillator-based prompt gamma camera for boron-neutron capture therapy: Comparison of SrI2 and GAGG using Monte-Carlo simulation

  • Kim, Minho;Hong, Bong Hwan;Cho, Ilsung;Park, Chawon;Min, Sun-Hong;Hwang, Won Taek;Lee, Wonho;Kim, Kyeong Min
    • Nuclear Engineering and Technology
    • /
    • v.53 no.2
    • /
    • pp.626-636
    • /
    • 2021
  • Boron-neutron capture therapy (BNCT) is a cancer treatment method that exploits the high neutron reactivity of boron. Monitoring the prompt gamma rays (PGs) produced during neutron irradiation is essential for ensuring the accuracy and safety of BNCT. We investigate the imaging of PGs produced by the boron-neutron capture reaction through Monte Carlo simulations of a gamma camera with a SrI2 scintillator and parallel-hole collimator. GAGG scintillator is also used for a comparison. The simulations allow the shapes of the energy spectra, which exhibit a peak at 478 keV, to be determined along with the PG images from a boron-water phantom. It is found that increasing the size of the water phantom results in a greater number of image counts and lower contrast. Additionally, a higher septal penetration ratio results in poorer image quality, and a SrI2 scintillator results in higher image contrast. Thus, we can simulate the BNCT process and obtain an energy spectrum with a reasonable shape, as well as suitable PG images. Both GAGG and SrI2 crystals are suitable for PG imaging during BNCT. However, for higher imaging quality, SrI2 and a collimator with a lower septal penetration ratio should be utilized.

3D feature point extraction technique using a mobile device (모바일 디바이스를 이용한 3차원 특징점 추출 기법)

  • Kim, Jin-Kyum;Seo, Young-Ho
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.10a
    • /
    • pp.256-257
    • /
    • 2022
  • In this paper, we introduce a method of extracting three-dimensional feature points through the movement of a single mobile device. Using a monocular camera, a 2D image is acquired according to the camera movement and a baseline is estimated. Perform stereo matching based on feature points. A feature point and a descriptor are acquired, and the feature point is matched. Using the matched feature points, the disparity is calculated and a depth value is generated. The 3D feature point is updated according to the camera movement. Finally, the feature point is reset at the time of scene change by using scene change detection. Through the above process, an average of 73.5% of additional storage space can be secured in the key point database. By applying the algorithm proposed to the depth ground truth value of the TUM Dataset and the RGB image, it was confirmed that the\re was an average distance difference of 26.88mm compared with the 3D feature point result.

  • PDF

Study on the Method of Diagnosing the Individuals Crop Growth Using by Multi-Spectral Images

  • Dongwon Kwon;Jaekyeong Baek;Wangyu Sang;Sungyul Chang;Jung-Il Cho;Ho-young Ban;HyeokJin Bak
    • Proceedings of the Korean Society of Crop Science Conference
    • /
    • 2022.10a
    • /
    • pp.108-108
    • /
    • 2022
  • In this study, multispectral images of wheat according to soil water state were collected, compared, and analyzed to measure the physiological response of crops to environmental stress at the individual level. CMS-V multi-spectral camera(Silios Technologies) was used for image acquisition. The camera lens consists of eight spectral bands between 550nm and 830nm. Light Reflective information collected in each band sensor and stored in digital values, and it is converted into a reflectance for calculating the vegetation index and used. According to the camera manual, the NDVI(Normalized Difference vegetation index) value was calculated using 628 nm and 752 nm bands. Image measurement was conducted under natural light conditions, and reflectance standards(Labsphere) were captured with plants for reflectance calculation. The wheat variety used Gosomil, and the wheat grown in the field was transplanted into a pot after heading date and measured. Three treatments were performed so that the soil volumetric water content of the pot was 13~17%, 20~23%, and 25%, and the growth response of wheat according to each treatment was compared using the NDVI value. In the first measurement after port transplantation, the difference in NDVI value according to treatment was not significant, but in the subsequent measurement, the NDVI value of the treatment with a water content of 13 to 17% was lowest and was the highest at 20 to 23%. The NDVI values decreased compared to the first measurement in all treatment, and the decrease was the largest at 13-17% water content and the smallest at 20-23%. Although the difference in NDVI values could be confirmed, it would be difficult to directly relate it to the water stress of plants, and further research on the response of crops to environmental stress and the analysis of multi-spectral image will be needed.

  • PDF