• Title/Summary/Keyword: camera image

Search Result 4,917, Processing Time 0.031 seconds

Image Processing and Deep Learning Techniques for Fast Pig's Posture Determining and Head Removal (돼지의 빠른 자세 결정과 머리 제거를 위한 영상처리 및 딥러닝 기법)

  • Ahn, Hanse;Choi, Wonseok;Park, Sunhwa;Chung, Yongwha;Park, Daihee
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.8 no.11
    • /
    • pp.457-464
    • /
    • 2019
  • The weight of pig is one of the main factors in determining the health and growth state of pigs, their shipment, the breeding environment, and the ration of feed, and thus measuring the pig's weight is an important issue in productivity perspective. In order to estimate the pig's weight by using the number of pig's pixels from images, acquired from a Top-view camera, the posture determining and the head removal from images are necessary to measure the accurate number of pixels. In this research, we propose the fast and accurate method to determine the pig's posture by using a fast image processing technique, find the head location by using a fast deep learning technique, and remove pig's head by using light weighted image processing technique. First, we determine the pig's posture by comparing the length from the center of the pig's body to the outline of the pig in the binary image. Then, we train the location of pig's head, body, and hip in images using YOLO(one of the fast deep learning based object detector), and then we obtain the location of pig's head and remove an outside area of head by using head location. Finally, we find the boundary of head and body by using Convex-hull, and we remove pig's head. In the Experiment result, we confirmed that the pig's posture was determined with an accuracy of 0.98 and a processing speed of 250.00fps, and the pig's head was removed with an accuracy of 0.96 and a processing speed of 48.97fps.

Improved Performance of Image Semantic Segmentation using NASNet (NASNet을 이용한 이미지 시맨틱 분할 성능 개선)

  • Kim, Hyoung Seok;Yoo, Kee-Youn;Kim, Lae Hyun
    • Korean Chemical Engineering Research
    • /
    • v.57 no.2
    • /
    • pp.274-282
    • /
    • 2019
  • In recent years, big data analysis has been expanded to include automatic control through reinforcement learning as well as prediction through modeling. Research on the utilization of image data is actively carried out in various industrial fields such as chemical, manufacturing, agriculture, and bio-industry. In this paper, we applied NASNet, which is an AutoML reinforced learning algorithm, to DeepU-Net neural network that modified U-Net to improve image semantic segmentation performance. We used BRATS2015 MRI data for performance verification. Simulation results show that DeepU-Net has more performance than the U-Net neural network. In order to improve the image segmentation performance, remove dropouts that are typically applied to neural networks, when the number of kernels and filters obtained through reinforcement learning in DeepU-Net was selected as a hyperparameter of neural network. The results show that the training accuracy is 0.5% and the verification accuracy is 0.3% better than DeepU-Net. The results of this study can be applied to various fields such as MRI brain imaging diagnosis, thermal imaging camera abnormality diagnosis, Nondestructive inspection diagnosis, chemical leakage monitoring, and monitoring forest fire through CCTV.

Accuracy Assessment of Aerial Triangulation of Network RTK UAV (네트워크 RTK 무인기의 항공삼각측량 정확도 평가)

  • Han, Soohee;Hong, Chang-Ki
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.38 no.6
    • /
    • pp.663-670
    • /
    • 2020
  • In the present study, we assessed the accuracy of aerial triangulation using a UAV (Unmanned Aerial Vehicle) capable of network RTK (Real-Time Kinematic) survey in a disaster situation that may occur in a semi-urban area mixed with buildings. For a reliable survey of check points, they were installed on the roofs of buildings, and static GNSS (Global Navigation Satellite System) survey was conducted for more than four hours. For objective accuracy assessment, coded aerial targets were installed on the check points to be automatically recognized by software. At the instance of image acquisition, the 3D coordinates of the UAV camera were measured using VRS (Virtual Reference Station) method, as a kind of network RTK survey, and the 3-axial angles were achieved using IMU (Inertial Measurement Unit) and gimbal rotation measurement. As a result of estimation and update of the interior and exterior orientation parameters using Agisoft Metashape, the 3D RMSE (Root Mean Square Error) of aerial triangulation ranged from 0.153 m to 0.102 m according to the combination of the image overlap and the angle of the image acquisition. To get higher aerial triangulation accuracy, it was proved to be effective to incorporate oblique images, though it is common to increase the overlap of vertical images. Therefore, to conduct a UAV mapping in an urgent disaster site, it is necessary to acquire oblique images together rather than improving image overlap.

A Study on Development of Portable Concrete Crack Measurement Device Using Image Processing Technique and Laser Sensors (이미지 처리기법 및 레이저 센서를 이용한 휴대용 콘크리트 균열 측정 장치 개발에 관한 연구)

  • Seo, Seunghwan;Ohn, Syng-Yup;Kim, Dong-Hyun;Kwak, Kiseok;Chung, Moonkyung
    • Journal of the Korean Geosynthetics Society
    • /
    • v.19 no.4
    • /
    • pp.41-50
    • /
    • 2020
  • Since cracks in concrete structures expedite corrosion of reinforced concrete over a long period of time, regular on-site inspections are essential to ensure structural usability and prevent degradation. Most of the safety inspections of facilities rely on visual inspection with naked eye, so cost and time consuming are severe, and the reliability of results differs depending on the inspector. In this study, a portable measuring device that can be used for safety diagnosis and maintenance was developed as a device that measures the width and length of concrete cracks through image analysis of cracks photographed with a camera. This device captures the cracks found within a close distance (3 m), and accurately calculates the unit pixel size by laser distance measurement, and automatically calculates the crack length and width with the image processing algorithm developed in this study. In measurement results using the crack image applied to the experiment, the measurement of the length of a 0.3 mm crack within a distance of 3 m was possible with a range of about 10% error. The crack width showed a tendency to be overestimated by detecting surrounding pixels due to vibration and blurring effect during the binarization process, but it could be effectively corrected by applying the crack width reduction function.

Analysis of UAV-based Multispectral Reflectance Variability for Agriculture Monitoring (농업관측을 위한 다중분광 무인기 반사율 변동성 분석)

  • Ahn, Ho-yong;Na, Sang-il;Park, Chan-won;Hong, Suk-young;So, Kyu-ho;Lee, Kyung-do
    • Korean Journal of Remote Sensing
    • /
    • v.36 no.6_1
    • /
    • pp.1379-1391
    • /
    • 2020
  • UAV in the agricultural application are capable of collecting ultra-high resolution image. It is possible to obtain timeliness images for phenological phases of the crop. However, the UAV uses a variety of sensors and multi-temporal images according to the environment. Therefore, it is essential to use normalized image data for time series image application for crop monitoring. This study analyzed the variability of UAV reflectance and vegetation index according to Aviation Image Making Environment to utilize the UAV multispectral image for agricultural monitoring time series. The variability of the reflectance according to environmental factors such as altitude, direction, time, and cloud was very large, ranging from 8% to 11%, but the vegetation index variability was stable, ranging from 1% to 5%. This phenomenon is believed to have various causes such as the characteristics of the UAV multispectral sensor and the normalization of the post-processing program. In order to utilize the time series of unmanned aerial vehicles, it is recommended to use the same ratio function as the vegetation index, and it is recommended to minimize the variability of time series images by setting the same time, altitude and direction as possible.

A Study of the Characteristics of Highly Spatially Resolved CW-laser-based Aerosol Lidar (고공간분해능 연속 광원을 이용한 미세먼지 라이다의 신호 특성에 관한 연구)

  • Sim, Juhyeon;Kim, Taekeong;Ju, Sohee;Noh, Youngmin;Kim, Dukhyeon
    • Korean Journal of Optics and Photonics
    • /
    • v.33 no.1
    • /
    • pp.1-10
    • /
    • 2022
  • In this study we introduce a new method for high-spatial-resolution continuous wave (CW) aerosol lidar that has a high spatial resolution in the near field and a low spatial resolution at long distances. A normal lidar system uses a nanosecond-pulse laser and measures the round-trip TOF between the aerosol and laser to obtain range resolution. In this study, however, we propose a new type of spatially resolving aerosol lidar that uses laser-scattering images. Using a laser-light-scattering image, we have calculated the distance of each scattering aerosol image for a given pixel, and recovered the short-range aerosol extinction. For this purpose, we have calculated the distance image and the contribution range of the aerosol to the given one-pixel image, and finally we have calculated the extinction coefficients of the aerosol with range-resolved information. In the case of traditional aerosol lidar, we can only obtain the aerosol extinction coefficients above 400 m. Using our suggested method, it was possible to extend the range of the extinction coefficient lower then several tens of meters. Finally, we can remove the unknown short-range region of pulsed aerosol lidar using our method.

Optimized Hardware Design using Sobel and Median Filters for Lane Detection

  • Lee, Chang-Yong;Kim, Young-Hyung;Lee, Yong-Hwan
    • Journal of Advanced Information Technology and Convergence
    • /
    • v.9 no.1
    • /
    • pp.115-125
    • /
    • 2019
  • In this paper, the image is received from the camera and the lane is sensed. There are various ways to detect lanes. Generally, the method of detecting edges uses a lot of the Sobel edge detection and the Canny edge detection. The minimum use of multiplication and division is used when designing for the hardware configuration. The images are tested using a black box image mounted on the vehicle. Because the top of the image of the used the black box is mostly background, the calculation process is excluded. Also, to speed up, YCbCr is calculated from the image and only the data for the desired color, white and yellow lane, is obtained to detect the lane. The median filter is used to remove noise from images. Intermediate filters excel at noise rejection, but they generally take a long time to compare all values. In this paper, by using addition, the time can be shortened by obtaining and using the result value of the median filter. In case of the Sobel edge detection, the speed is faster and noise sensitive compared to the Canny edge detection. These shortcomings are constructed using complementary algorithms. It also organizes and processes data into parallel processing pipelines. To reduce the size of memory, the system does not use memory to store all data at each step, but stores it using four line buffers. Three line buffers perform mask operations, and one line buffer stores new data at the same time as the operation. Through this work, memory can use six times faster the processing speed and about 33% greater quantity than other methods presented in this paper. The target operating frequency is designed so that the system operates at 50MHz. It is possible to use 2157fps for the images of 640by360 size based on the target operating frequency, 540fps for the HD images and 240fps for the Full HD images, which can be used for most images with 30fps as well as 60fps for the images with 60fps. The maximum operating frequency can be used for larger amounts of the frame processing.

A Study on Atmospheric Turbulence-Induced Errors in Vision Sensor based Structural Displacement Measurement (대기외란시 비전센서를 활용한 구조물 동적 변위 측정 성능에 관한 연구)

  • Junho Gong
    • Journal of the Korea institute for structural maintenance and inspection
    • /
    • v.28 no.3
    • /
    • pp.1-9
    • /
    • 2024
  • This study proposes a multi-scale template matching technique with image pyramids (TMI) to measure structural dynamic displacement using a vision sensor under atmospheric turbulence conditions and evaluates its displacement measurement performance. To evaluate displacement measurement performance according to distance, the three-story shear structure was designed, and an FHD camera was prepared to measure structural response. The initial measurement distance was set at 10m, and increased with an increment of 10m up to 40m. The atmospheric disturbance was generated using a heating plate under indoor illuminance condition, and the image was distorted by the optical turbulence. Through preliminary experiments, the feasibility of displacement measurement of the feature point-based displacement measurement method and the proposed method during atmospheric disturbances were compared and verified, and the verification results showed a low measurement error rate of the proposed method. As a result of evaluating displacement measurement performance in an atmospheric disturbance environment, there was no significant difference in displacement measurement performance for TMI using an artificial target depending on the presence or absence of atmospheric disturbance. However, when natural targets were used, RMSE increased significantly at shooting distances of 20 m or more, showing the operating limitations of the proposed technique. This indicates that the resolution of the natural target decreases as the shooting distance increases, and image distortion due to atmospheric disturbance causes errors in template image estimation, resulting in a high displacement measurement error.

Pseudo Image Composition and Sensor Models Analysis of SPOT Satellite Imagery of Non-Accessible Area (비접근 지역에 대한 SPOT 위성영상의 Pseudo영상 구성 및 센서모델 분석)

  • 방기인;조우석
    • Proceedings of the KSRS Conference
    • /
    • 2001.03a
    • /
    • pp.140-148
    • /
    • 2001
  • The satellite sensor model is typically established using ground control points acquired by ground survey Of existing topographic maps. In some cases where the targeted area can't be accessed and the topographic maps are not available, it is difficult to obtain ground control points so that geospatial information could not be obtained from satellite image. The paper presents several satellite sensor models and satellite image decomposition methods for non-accessible area where ground control points can hardly acquired in conventional ways. First, 10 different satellite sensor models, which were extended from collinearity condition equations, were developed and then the behavior of each sensor model was investigated. Secondly, satellite images were decomposed and also pseudo images were generated. The satellite sensor model extended from collinearity equations was represented by the six exterior orientation parameters in 1$^{st}$, 2$^{nd}$ and 3$^{rd}$ order function of satellite image row. Among them, the rotational angle parameters such as $\omega$(omega) and $\phi$(phi) correlated highly with positional parameters could be assigned to constant values. For non-accessible area, satellite images were decomposed, which means that two consecutive images were combined as one image. The combined image consists of one satellite image with ground control points and the other without ground control points. In addition, a pseudo image which is an imaginary image, was prepared from one satellite image with ground control points and the other without ground control points. In other words, the pseudo image is an arbitrary image bridging two consecutive images. For the experiments, SPOT satellite images exposed to the similar area in different pass were used. Conclusively, it was found that 10 different satellite sensor models and 5 different decomposed methods delivered different levels of accuracy. Among them, the satellite camera model with 1$^{st}$ order function of image row for positional orientation parameters and rotational angle parameter of kappa, and constant rotational angle parameter omega and phi provided the best 60m maximum error at check point with pseudo images arrangement.

  • PDF

Detection of Gaze Direction for the Hearing-impaired in the Intelligent Space (지능형 공간에서 청각장애인의 시선 방향 검출)

  • Oh, Young-Joon;Hong, Kwang-Jin;Kim, Jong-In;Jung, Kee-Chul
    • The KIPS Transactions:PartB
    • /
    • v.18B no.6
    • /
    • pp.333-340
    • /
    • 2011
  • The Human-Computer Interaction(HCI) is a study of the method for interaction between human and computers that merges the ergonomics and the information technology. The intelligent space, which is a part of the HCI, is an important area to provide effective user interface for the disabled, who are alienated from the information-oriented society. In the intelligent space for the disabled, the method supporting information depends on types of disability. In this paper, we only support the hearing-impaired. It is material to the gaze direction detection method because it is very efficient information provide method to present information on gazing direction point, except for the information provide location perception method through directly contact with the hearing-impaired. We proposed the gaze direction detection method must be necessary in order to provide the residence life application to the hearing-impaired like this. The proposed method detects the region of the user from multi-view camera images, generates candidates for directions of gaze for horizontal and vertical from each camera, and calculates the gaze direction of the user through the comparison with the size of each candidate. In experimental results, the proposed method showed high detection rate with gaze direction and foot sensing rate with user's position, and showed the performance possibility of the scenario for the disabled.