• Title/Summary/Keyword: camera image

Search Result 4,918, Processing Time 0.034 seconds

Distance Measurement of Small Moving Object using Infrared Stereo Camera (적외선 스테레오 카메라를 이용한 소형 이동체의 거리 측정)

  • Oh, Jun-Ho;Lee, Sang-Hwa;Lee, Boo-Hwan;Park, Jong-Il
    • Journal of the Institute of Electronics Engineers of Korea SC
    • /
    • v.49 no.3
    • /
    • pp.53-61
    • /
    • 2012
  • This paper proposes a real-time distance measurement system of high temperature and high speed target using infrared stereo camera. We construct an infrared stereo camera system that measure the difference between target and background temperatures for automatic target measurement. First, the proposed method detects target region based on target motion and intensity variation of local region using difference between target and background temperatures. Second, stereo matching by left and right target information is used to estimate disparity about real-time distance of target. In the proposed method using infrared stereo camera system, we compare distances in three dimension trajectory measuring instrument and in infrared stereo camera measurement. In this experiment from three video data, the result shows an average 9.68% distance error rate. The proposed method is suitable for distance and position measurement of varied targets using infrared stereo system.

Research for Calibration and Correction of Multi-Spectral Aerial Photographing System(PKNU 3) (다중분광 항공촬영 시스템(PKNU 3) 검정 및 보정에 관한 연구)

  • Lee, Eun Kyung;Choi, Chul Uong
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.7 no.4
    • /
    • pp.143-154
    • /
    • 2004
  • The researchers, who seek geological and environmental information, depend on the remote sensing and aerial photographic datum from various commercial satellites and aircraft. However, the adverse weather conditions and the expensive equipment can restrict that the researcher can collect their data anywhere and any time. To allow for better flexibility, we have developed a compact, a multi-spectral automatic Aerial photographic system(PKNU 2). This system's Multi-spectral camera can catch the visible(RGB) and infrared(NIR) bands($3032{\times}2008$ pixels) image. Visible and infrared bands images were obtained from each camera respectively and produced Color-infrared composite images to be analyzed in the purpose of the environment monitor but that was not very good data. Moreover, it has a demerit that the stereoscopic overlap area is not satisfied with 60% due to the 12s storage time of each data, while it was possible that PKNU 2 system photographed photos of great capacity. Therefore, we have been developing the advanced PKNU 2(PKNU 3) that consists of color-infrared spectral camera can photograph the visible and near infrared bands data using one sensor at once, thermal infrared camera, two of 40 G computers to store images, and MPEG board to compress and transfer data to the computer at the real time and can attach and detach itself to a helicopter. Verification and calibration of each sensor(REDLAKE MS 4000, Raytheon IRPro) were conducted before we took the aerial photographs for obtaining more valuable data. Corrections for the spectral characteristics and radial lens distortions of sensor were carried out.

  • PDF

FMD response cow hooves and temperature detection algorithm using a thermal imaging camera (열화상 카메라를 이용한 구제역 대응 소 발굽 온도 검출 알고리즘 개발)

  • Yu, Chan-Ju;Kim, Jeong-Jun
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.17 no.9
    • /
    • pp.292-301
    • /
    • 2016
  • Because damages arising from the occurrence of foot-and-mouth disease (FMD) are very great, it is essential to make a preemptive diagnosis to cope with it in order to minimize those damages. The main symptoms of foot-and-mouth disease are body temperature increase, loss of appetite, formation of blisters in the mouth, on hooves and breasts, etc. in a cow or a bull, among which the body temperature check is the easiest and quickest way to detect the disease. In this paper, an algorithm to detect FMD from the hooves of cattle was developed and implemented for preemptive coping with foot-and-mouth disease, and a hoof check test is conducted after the installation of a high-resolution camera module, a thermo-graphic camera, and a temperature/humidity module in the cattle shed. Through the algorithm and system developed in this study, it is possible to cope with an early-stage situation in which cattle are suspected as suffering from foot-and-mouth disease, creating an optimized growth environment for cattle. In particular, in this study, the system to cope with FMD does not use a portable thermo-graphic camera, but a fixed camera attached to the cattle shed. It does not need additional personnel, has a function to measure the temperature of cattle hooves automatically through an image algorithm, and includes an automated alarm for a smart phone. This system enables the prediction of a possible occurrence of foot-and-mouth disease on a real-time basis, and also enables initial-stage disinfection to be performed to cope with the disease without needing extra personnel.

Conversion of Camera Lens Distortions between Photogrammetry and Computer Vision (사진측량과 컴퓨터비전 간의 카메라 렌즈왜곡 변환)

  • Hong, Song Pyo;Choi, Han Seung;Kim, Eui Myoung
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.37 no.4
    • /
    • pp.267-277
    • /
    • 2019
  • Photogrammetry and computer vision are identical in determining the three-dimensional coordinates of images taken with a camera, but the two fields are not directly compatible with each other due to differences in camera lens distortion modeling methods and camera coordinate systems. In general, data processing of drone images is performed by bundle block adjustments using computer vision-based software, and then the plotting of the image is performed by photogrammetry-based software for mapping. In this case, we are faced with the problem of converting the model of camera lens distortions into the formula used in photogrammetry. Therefore, this study described the differences between the coordinate systems and lens distortion models used in photogrammetry and computer vision, and proposed a methodology for converting them. In order to verify the conversion formula of the camera lens distortion models, first, lens distortions were added to the virtual coordinates without lens distortions by using the computer vision-based lens distortion models. Then, the distortion coefficients were determined using photogrammetry-based lens distortion models, and the lens distortions were removed from the photo coordinates and compared with the virtual coordinates without the original distortions. The results showed that the root mean square distance was good within 0.5 pixels. In addition, epipolar images were generated to determine the accuracy by applying lens distortion coefficients for photogrammetry. The calculated root mean square error of y-parallax was found to be within 0.3 pixels.

A New Calibration of 3D Point Cloud using 3D Skeleton (3D 스켈레톤을 이용한 3D 포인트 클라우드의 캘리브레이션)

  • Park, Byung-Seo;Kang, Ji-Won;Lee, Sol;Park, Jung-Tak;Choi, Jang-Hwan;Kim, Dong-Wook;Seo, Young-Ho
    • Journal of Broadcast Engineering
    • /
    • v.26 no.3
    • /
    • pp.247-257
    • /
    • 2021
  • This paper proposes a new technique for calibrating a multi-view RGB-D camera using a 3D (dimensional) skeleton. In order to calibrate a multi-view camera, consistent feature points are required. In addition, it is necessary to acquire accurate feature points in order to obtain a high-accuracy calibration result. We use the human skeleton as a feature point to calibrate a multi-view camera. The human skeleton can be easily obtained using state-of-the-art pose estimation algorithms. We propose an RGB-D-based calibration algorithm that uses the joint coordinates of the 3D skeleton obtained through the posture estimation algorithm as a feature point. Since the human body information captured by the multi-view camera may be incomplete, the skeleton predicted based on the image information acquired through it may be incomplete. After efficiently integrating a large number of incomplete skeletons into one skeleton, multi-view cameras can be calibrated by using the integrated skeleton to obtain a camera transformation matrix. In order to increase the accuracy of the calibration, multiple skeletons are used for optimization through temporal iterations. We demonstrate through experiments that a multi-view camera can be calibrated using a large number of incomplete skeletons.

Scaling Attack Method for Misalignment Error of Camera-LiDAR Calibration Model (카메라-라이다 융합 모델의 오류 유발을 위한 스케일링 공격 방법)

  • Yi-ji Im;Dae-seon Choi
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.33 no.6
    • /
    • pp.1099-1110
    • /
    • 2023
  • The recognition system of autonomous driving and robot navigation performs vision work such as object recognition, tracking, and lane detection after multi-sensor fusion to improve performance. Currently, research on a deep learning model based on the fusion of a camera and a lidar sensor is being actively conducted. However, deep learning models are vulnerable to adversarial attacks through modulation of input data. Attacks on the existing multi-sensor-based autonomous driving recognition system are focused on inducing obstacle detection by lowering the confidence score of the object recognition model.However, there is a limitation that an attack is possible only in the target model. In the case of attacks on the sensor fusion stage, errors in vision work after fusion can be cascaded, and this risk needs to be considered. In addition, an attack on LIDAR's point cloud data, which is difficult to judge visually, makes it difficult to determine whether it is an attack. In this study, image scaling-based camera-lidar We propose an attack method that reduces the accuracy of LCCNet, a fusion model (camera-LiDAR calibration model). The proposed method is to perform a scaling attack on the point of the input lidar. As a result of conducting an attack performance experiment by size with a scaling algorithm, an average of more than 77% of fusion errors were caused.

Development of a Deep Learning Network for Quality Inspection in a Multi-Camera Inline Inspection System for Pharmaceutical Containers (의약 용기의 다중 카메라 인라인 검사 시스템에서의 품질 검사를 위한 딥러닝 네트워크 개발)

  • Tae-Yoon Lee;Seok-Moon Yoon;Seung-Ho Lee
    • Journal of IKEEE
    • /
    • v.28 no.3
    • /
    • pp.474-478
    • /
    • 2024
  • In this paper, we proposes a deep learning network for quality inspection in a multi-camera inline inspection system for pharmaceutical containers. The proposed deep learning network is specifically designed for pharmaceutical containers by using data produced in real manufacturing environments, leading to more accurate quality inspection. Additionally, the use of an inline-capable deep learning network allows for an increase in inspection speed. The development of the deep learning network for quality inspection in the multi-camera inline inspection system consists of three steps. First, a dataset of approximately 10,000 images is constructed from the production site using one line camera for foreign substance inspection and three area cameras for dimensional inspection. Second, the pharmaceutical container data is preprocessed by designating regions of interest (ROI) in areas where defects are likely to occur, tailored for foreign substance and dimensional inspections. Third, the preprocessed data is used to train the deep learning network. The network improves inference speed by reducing the number of channels and eliminating the use of linear layers, while accuracy is enhanced by applying PReLU and residual learning. This results in the creation of four deep learning modules tailored to the dataset built from the four cameras. The performance of the proposed deep learning network for quality inspection in the multi-camera inline inspection system for pharmaceutical containers was evaluated through experiments conducted by a certified testing agency. The results show that the deep learning modules achieved a classification accuracy of 99.4%, exceeding the world-class level of 95%, and an average classification speed of 0.947 seconds, which is superior to the world-class level of 1 second. Therefore, the effectiveness of the proposed deep learning network for quality inspection in a multi-camera inline inspection system for pharmaceutical containers has been demonstrated.

A Study of Experimental Image Direction for Short Animation Movies -focusing in short film and (단편애니메이션의 실험적 영상연출 연구 -<탱고>와 <페스트 필름>을 중심으로)

  • Choi, Don-Ill
    • Cartoon and Animation Studies
    • /
    • s.36
    • /
    • pp.375-391
    • /
    • 2014
  • Animation movie is a non-photorealistic animated art that consists of formative language forming a frame based on a story and cuts describing frames that form the cuts. Therefore, in expressing an image, artistic expression methods and devices for a formative space are should be provided in a frame while cuts have the images between frames faithfully. Short animation movie is produced by various image experiments with unique image expressions rather than narration for expressing subjective discourse of a writer. Therefore, image style that forms unique images and various image directions are important factors. This study compared the experimental image directions of and , both of which showed a production method of film manipulation. First, while uses pixilation that produces images obtained from live images through painting and many optical disclosure process on a cell mat, was made with diverse collage techniques such as tearing, cutting, pasting, and folding hundreds of scenes from action movies. Second, expresses non-causal relationship of characters by their repetitive behaviors and circulatory image structure through a fixed camera angle, resisting typical scene transition. On the other hand, has an advancing structure that progresses antagonistic relationship of characters through diverse camera angles and scene transition of unique images. Third, in terms of editing, uses a long-take short cut technique in which the whole image consists of one short cut, though it seems to be many scenes with the appearance of various characters. On the other hand, maximizes visual fun and commitment by image reconstruction with hundreds of various short cuts. That is, both works have common features of an experimental work that shows expansion of animated image expressions through film manipulation that is different form general animation productions. On top of that, delivers routine life of diverse human beings without clear narration through image of conceptualized spaces. expresses it in a new image space through image reconstruction with collage technique and speedy progress, setting a binary opposition structure.

Investigation of the Signal Characteristics of a Small Gamma Camera System Using NaI(Tl)-Position Sensitive Photomultiplier Tube (NaI(Tl) 섬광결정과 위치민감형 광전자증배관을 이용한 소형 감마카메라의 신호 특성 고찰)

  • Choi, Yong;Kim, Jong-Ho;Kim, Joon-Young;Im, Ki-Chun;Kim, Sang-Eun;Choe, Yearn-Seong;Lee, Kyung-Han;Joo, Koan-Sik;Kim, Byung-Tae
    • The Korean Journal of Nuclear Medicine
    • /
    • v.34 no.1
    • /
    • pp.82-93
    • /
    • 2000
  • Purpose: We characterized the signals obtained from the components of a small gamma camera using Nal(Tl)-position sensitive photomultiplier tube (PSPMT) and optimized the parameters employed in the modules of the system. Materials and Methods: The small gamma camera system consists of a Nal(Tl) crystal ($60{\times}60{\times}6mm^3$) coupled with a Hamamatsu R3941 PSPMT, a resister chain circuit, preamplifiers, nuclear instrument modules (NIMs), an analog to digital converter and a personal computer for control and display. The PSPMT was read out using a resistive charge division circuit which multiplexes the 34 cross wire anode channels into 4 signals (X+, X-, Y+, Y -). Those signals were individually amplified by four preamplifiers and then, shaped and amplified by amplifiers. The signals were discriminated and digitized via triggering signal and used to localize the position of an event by applying the Anger logic. The gamma camera control and image display was performed by a program implemented using a graphic software. Results: The characteristics of signal and the parameters employed in each module of the system were presented. The intrinsic sensitivity of the system was approximately $8{\times}10^3$ counts/sec/${\mu}Ci$. The intrinsic energy resolution of the system was 18% FWHM at 140 keV. The spatial resolution obtained using a line-slit mask and $^{99m}Tc$ point source were, respectively, 2.2 and 2.3 mm FWHM in X and Y directions. Breast phantom containing $2{\sim}7mm$ diameter spheres was successfully imaged with a parallel hole collimator. The image displayed accurate size and activity distribution over the imaging field of view Conclusion: We proposed a simple method for development of a small gamma camera and presented the characteristics of the signals from the system and the optimized parameters used in the modules of the small gamma camera.

  • PDF

Development of a Small Gamma Camera Using NaI(T1)-Position Sensitive Photomultiplier Tube for Breast Imaging (NaI (T1) 섬광결정과 위치민감형 광전자증배관을 이용한 유방암 진단용 소형 감마카메라 개발)

  • Kim, Jong-Ho;Choi, Yong;Kwon, Hong-Seong;Kim, Hee-Joung;Kim, Sang-Eun;Choe, Yearn-Seong;Lee, Kyung-Han;Kim, Moon-Hae;Joo, Koan-Sik;Kim, Byuug-Tae
    • The Korean Journal of Nuclear Medicine
    • /
    • v.32 no.4
    • /
    • pp.365-373
    • /
    • 1998
  • Purpose: The conventional gamma camera is not ideal for scintimammography because of its large detector size (${\sim}500mm$ in width) causing high cost and low image quality. We are developing a small gamma camera dedicated for breast imaging. Materials and Methods: The small gamma camera system consists of a NaI (T1) crystal ($60 mm{\times}60 mm{\times}6 mm$) coupled with a Hamamatsu R3941 Position Sensitive Photomultiplier Tube (PSPMT), a resister chain circuit, preamplifiers, nuclear instrument modules, an analog to digital converter and a personal computer for control and display. The PSPMT was read out using a standard resistive charge division which multiplexes the 34 cross wire anode channels into 4 signals ($X^+,\;X^-,\;Y^+,\;Y^-$). Those signals were individually amplified by four preamplifiers and then, shaped and amplified by amplifiers. The signals were discriminated ana digitized via triggering signal and used to localize the position of an event by applying the Anger logic. Results: The intrinsic sensitivity of the system was approximately 8,000 counts/sec/${\mu}Ci$. High quality flood and hole mask images were obtained. Breast phantom containing $2{\sim}7 mm$ diameter spheres was successfully imaged with a parallel hole collimator The image displayed accurate size and activity distribution over the imaging field of view Conclusion: We have succesfully developed a small gamma camera using NaI(T1)-PSPMT and nuclear Instrument modules. The small gamma camera developed in this study might improve the diagnostic accuracy of scintimammography by optimally imaging the breast.

  • PDF