• Title/Summary/Keyword: Image Edge

Search Result 2,464, Processing Time 0.03 seconds

Development of the Algorithm for Traffic Accident Auto-Detection in Signalized Intersection (신호교차로 내 실시간 교통사고 자동검지 알고리즘 개발)

  • O, Ju-Taek;Im, Jae-Geuk;Hwang, Bo-Hui
    • Journal of Korean Society of Transportation
    • /
    • v.27 no.5
    • /
    • pp.97-111
    • /
    • 2009
  • Image-based traffic information collection systems have entered widespread adoption and use in many countries since these systems are not only capable of replacing existing loop-based detectors which have limitations in management and administration, but are also capable of providing and managing a wide variety of traffic related information. In addition, these systems are expanding rapidly in terms of purpose and scope of use. Currently, the utilization of image processing technology in the field of traffic accident management is limited to installing surveillance cameras on locations where traffic accidents are expected to occur and digitalizing of recorded data. Accurately recording the sequence of situations around a traffic accident in a signal intersection and then objectively and clearly analyzing how such accident occurred is more urgent and important than anything else in resolving a traffic accident. Therefore, in this research, we intend to present a technology capable of overcoming problems in which advanced existing technologies exhibited limitations in handling real-time due to large data capacity such as object separation of vehicles and tracking, which pose difficulties due to environmental diversities and changes at a signal intersection with complex traffic situations, as pointed out by many past researches while presenting and implementing an active and environmentally adaptive methodology capable of effectively reducing false detection situations which frequently occur even with the Gaussian complex model analytical method which has been considered the best among well-known environmental obstacle reduction methods. To prove that the technology developed by this research has performance advantage over existing automatic traffic accident recording systems, a test was performed by entering image data from an actually operating crossroad online in real-time. The test results were compared with the performance of other existing technologies.

Container Image Recognition using Fuzzy-based Noise Removal Method and ART2-based Self-Organizing Supervised Learning Algorithm (퍼지 기반 잡음 제거 방법과 ART2 기반 자가 생성 지도 학습 알고리즘을 이용한 컨테이너 인식 시스템)

  • Kim, Kwang-Baek;Heo, Gyeong-Yong;Woo, Young-Woon
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.11 no.7
    • /
    • pp.1380-1386
    • /
    • 2007
  • This paper proposed an automatic recognition system of shipping container identifiers using fuzzy-based noise removal method and ART2-based self-organizing supervised learning algorithm. Generally, identifiers of a shipping container have a feature that the color of characters is blacker white. Considering such a feature, in a container image, all areas excepting areas with black or white colors are regarded as noises, and areas of identifiers and noises are discriminated by using a fuzzy-based noise detection method. Areas of identifiers are extracted by applying the edge detection by Sobel masking operation and the vertical and horizontal block extraction in turn to the noise-removed image. Extracted areas are binarized by using the iteration binarization algorithm, and individual identifiers are extracted by applying 8-directional contour tacking method. This paper proposed an ART2-based self-organizing supervised learning algorithm for the identifier recognition, which improves the performance of learning by applying generalized delta learning and Delta-bar-Delta algorithm. Experiments using real images of shipping containers showed that the proposed identifier extraction method and the ART2-based self-organizing supervised learning algorithm are more improved compared with the methods previously proposed.

PET System Design using a Scintillator with a Size of 0.8 mm to Improve Spatial Resolution (공간분해능 향상을 위한 0.8 mm 크기의 섬광체를 사용한 PET 시스템 설계)

  • Lee, Seung-Jae
    • Journal of the Korean Society of Radiology
    • /
    • v.16 no.5
    • /
    • pp.499-504
    • /
    • 2022
  • Positron emission tomography (PET) uses a very small scintillator to achieve exellent spatial resolution. Therefore, in this study, a PET system using a scintillator to 0.8 mm size was designed and the performance was evaluated. Anihilation radiation was generated from the center of the field of view (FOV) to the outskirts at intervals of 10 mm, and counted simultaneously. The image was reconstructed using the coincidence data, and the spatial resolution was calculated by acquiring the full width at half maximum through the profile. The spatial resolution at the center of the FOV was 1.02 mm, showing a very good result, and the spatial resolution decreased as it was located at the outer edge. To evaluate the phantom image, the Derenzo phantom was constructed to acquire the image, and the degree of classification between radiation sources was evaluated through profile analysis. The result showed that the distance between the radiation sources was larger than the spatial resolution of the radiation sources at each location, and it was confirmed that the radiation sources were distinguished through this. When the PET system designed in this study is applied to PET for small animals, it is considered that excellent performance can be secured through the characteristic of very good spatial resolution.

SAR(Synthetic Aperture Radar) 3-Dimensional Scatterers Point Cloud Target Model and Experiments on Bridge Area (영상레이더(SAR)용 3차원 산란점 점구름 표적모델의 교량 지역에 대한 적용)

  • Jong Hoo Park;Sang Chul Park
    • Journal of the Korea Society for Simulation
    • /
    • v.32 no.3
    • /
    • pp.1-8
    • /
    • 2023
  • Modeling of artificial targets in Synthetic Aperture radar (SAR) mainly simulates radar signals reflected from the faces and edges of the 3D Computer Aided Design (CAD) model with a ray-tracing method, and modeling of the clutter on the Earth's surface uses a method of distinguishing types with similar distribution characteristics through statistical analysis of the SAR image itself. In this paper, man-made targets on the surface and background clutter on the terrain are integrated and made into a three-dimensional (3D) point cloud scatterer model, and SAR image were created through computational signal processing. The results of the SAR Stripmap image generation of the actual automobile based SAR radar system and the results analyzed using EM modeling or statistical distribution models are compared with this 3D point cloud scatterer model. The modeling target is selected as an bridge because it has the characteristic of having both water surface and ground terrain around the bridge and is also a target of great interest in both military and civilian use.

The Mirror-based real-time dynamic projection mapping design and dynamic object detection system research (미러 방식의 실시간 동적 프로젝션 매핑 설계 및 동적 사물 검출 시스템 연구)

  • Soe-Young Ahn;Bum-Suk Seo;Sung Dae Hong
    • Journal of Internet of Things and Convergence
    • /
    • v.10 no.2
    • /
    • pp.85-91
    • /
    • 2024
  • In this paper, we studied projection mapping, which is being utilized as a digital canvas beyond space and time for theme parks, mega events, and exhibition performances. Since the existing projection technology used for fixed objects has the limitation that it is difficult to map moving objects in terms of utilization, it is urgent to develop a technology that can track and map moving objects and a real-time dynamic projection mapping system based on dynamically moving objects so that it can respond to various markets such as performances, exhibitions, and theme parks. In this paper, we propose a system that can track real-time objects in real time and eliminate the delay phenomenon by developing hardware and performing high-speed image processing. Specifically, we develop a real-time object image analysis and projection focusing control unit, an integrated operating system for a real-time object tracking system, and an image processing library for projection mapping. This research is expected to have a wide range of applications in the technology-intensive industry that utilizes real-time vision machine-based detection technology, as well as in the industry where cutting-edge science and technology are converged and produced.

A study on the design of an efficient hardware and software mixed-mode image processing system for detecting patient movement (환자움직임 감지를 위한 효율적인 하드웨어 및 소프트웨어 혼성 모드 영상처리시스템설계에 관한 연구)

  • Seungmin Jung;Euisung Jung;Myeonghwan Kim
    • Journal of Internet Computing and Services
    • /
    • v.25 no.1
    • /
    • pp.29-37
    • /
    • 2024
  • In this paper, we propose an efficient image processing system to detect and track the movement of specific objects such as patients. The proposed system extracts the outline area of an object from a binarized difference image by applying a thinning algorithm that enables more precise detection compared to previous algorithms and is advantageous for mixed-mode design. The binarization and thinning steps, which require a lot of computation, are designed based on RTL (Register Transfer Level) and replaced with optimized hardware blocks through logic circuit synthesis. The designed binarization and thinning block was synthesized into a logic circuit using the standard 180n CMOS library and its operation was verified through simulation. To compare software-based performance, performance analysis of binary and thinning operations was also performed by applying sample images with 640 × 360 resolution in a 32-bit FPGA embedded system environment. As a result of verification, it was confirmed that the mixed-mode design can improve the processing speed by 93.8% in the binary and thinning stages compared to the previous software-only processing speed. The proposed mixed-mode system for object recognition is expected to be able to efficiently monitor patient movements even in an edge computing environment where artificial intelligence networks are not applied.

A Road Luminance Measurement Application based on Android (안드로이드 기반의 도로 밝기 측정 어플리케이션 구현)

  • Choi, Young-Hwan;Kim, Hongrae;Hong, Min
    • Journal of Internet Computing and Services
    • /
    • v.16 no.2
    • /
    • pp.49-55
    • /
    • 2015
  • According to the statistics of traffic accidents over recent 5 years, traffic accidents during the night times happened more than the day times. There are various causes to occur traffic accidents and the one of the major causes is inappropriate or missing street lights that make driver's sight confused and causes the traffic accidents. In this paper, with smartphones, we designed and implemented a lane luminance measurement application which stores the information of driver's location, driving, and lane luminance into database in real time to figure out the inappropriate street light facilities and the area that does not have any street lights. This application is implemented under Native C/C++ environment using android NDK and it improves the operation speed than code written in Java or other languages. To measure the luminance of road, the input image with RGB color space is converted to image with YCbCr color space and Y value returns the luminance of road. The application detects the road lane and calculates the road lane luminance into the database sever. Also this application receives the road video image using smart phone's camera and improves the computational cost by allocating the ROI(Region of interest) of input images. The ROI of image is converted to Grayscale image and then applied the canny edge detector to extract the outline of lanes. After that, we applied hough line transform method to achieve the candidated lane group. The both sides of lane is selected by lane detection algorithm that utilizes the gradient of candidated lanes. When the both lanes of road are detected, we set up a triangle area with a height 20 pixels down from intersection of lanes and the luminance of road is estimated from this triangle area. Y value is calculated from the extracted each R, G, B value of pixels in the triangle. The average Y value of pixels is ranged between from 0 to 100 value to inform a luminance of road and each pixel values are represented with color between black and green. We store car location using smartphone's GPS sensor into the database server after analyzing the road lane video image with luminance of road about 60 meters ahead by wireless communication every 10 minutes. We expect that those collected road luminance information can warn drivers about safe driving or effectively improve the renovation plans of road luminance management.

Spectral Band Selection for Detecting Fire Blight Disease in Pear Trees by Narrowband Hyperspectral Imagery (초분광 이미지를 이용한 배나무 화상병에 대한 최적 분광 밴드 선정)

  • Kang, Ye-Seong;Park, Jun-Woo;Jang, Si-Hyeong;Song, Hye-Young;Kang, Kyung-Suk;Ryu, Chan-Seok;Kim, Seong-Heon;Jun, Sae-Rom;Kang, Tae-Hwan;Kim, Gul-Hwan
    • Korean Journal of Agricultural and Forest Meteorology
    • /
    • v.23 no.1
    • /
    • pp.15-33
    • /
    • 2021
  • In this study, the possibility of discriminating Fire blight (FB) infection tested using the hyperspectral imagery. The reflectance of healthy and infected leaves and branches was acquired with 5 nm of full width at high maximum (FWHM) and then it was standardized to 10 nm, 25 nm, 50 nm, and 80 nm of FWHM. The standardized samples were divided into training and test sets at ratios of 7:3, 5:5 and 3:7 to find the optimal bands of FWHM by the decision tree analysis. Classification accuracy was evaluated using overall accuracy (OA) and kappa coefficient (KC). The hyperspectral reflectance of infected leaves and branches was significantly lower than those of healthy green, red-edge (RE) and near infrared (NIR) regions. The bands selected for the first node were generally 750 and 800 nm; these were used to identify the infection of leaves and branches, respectively. The accuracy of the classifier was higher in the 7:3 ratio. Four bands with 50 nm of FWHM (450, 650, 750, and 950 nm) might be reasonable because the difference in the recalculated accuracy between 8 bands with 10 nm of FWHM (440, 580, 640, 660, 680, 710, 730, and 740 nm) and 4 bands was only 1.8% for OA and 4.1% for KC, respectively. Finally, adding two bands (550 nm and 800 nm with 25 nm of FWHM) in four bands with 50 nm of FWHM have been proposed to improve the usability of multispectral image sensors with performing various roles in agriculture as well as detecting FB with other combinations of spectral bands.

Low-Power Discrete-Event SoC for 3DTV Active Shutter Glasses (3DTV 엑티브 셔터 안경을 위한 저전력 이산-사건 SoC)

  • Park, Dae-Jin;Kwak, Sung-Ho;Kim, Chang-Min;Kim, Tag-Gon
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.48 no.6
    • /
    • pp.18-26
    • /
    • 2011
  • Debates concerning the competitive edge of leading 3DTV technology of the shutter glasses (SG) 3D and the film-type patterned retarder (FPR) are flaring up. Although SG technology enables Full-HD 3D vision, it requires complex systems including the sync transmitter (emitter), the sync processor chip, and the LCD lens in the active shutter glasses. In addition, the transferred sync-signal is easily affected by the external noise and a 3DTV viewer may feel flicker-effect caused by cross-talk of the left and right image. The operating current of the sync processor in the 3DTV active shutter glasses is gradually increasing to compensate the sync reconstruction error. The proposed chip is a low-power hardware sync processor based discrete-event SoC(system on a chip) designed specifically for the 3DTV active shutter glasses. This processor implements the newly designed power-saving techniques targeted for low-power operation in a noisy environment between 3DTV and the active shutter glasses. This design includes a hardware pre-processor based on a universal edge tracer and provides a perfect sync reconstruction based on a floating-point timer to advance the prior commercial 3DTV shutter glasses in terms of their power consumption. These two techniques enable an accurate sync reconstruction in the slow clock frequency of the synchronization timer and reduce the power consumption to less than about a maximum of 20% compared with other major commercial processors. This article describes the system's architecture and the details of the proposed techniques, also identifying the key concepts and functions.

A Fast Iris Region Finding Algorithm for Iris Recognition (홍채 인식을 위한 고속 홍채 영역 추출 방법)

  • 송선아;김백섭;송성호
    • Journal of KIISE:Software and Applications
    • /
    • v.30 no.9
    • /
    • pp.876-884
    • /
    • 2003
  • It is essential to identify both the pupil and iris boundaries for iris recognition. The circular edge detector proposed by Daugman is the most common and powerful method for the iris region extraction. The method is accurate but requires lots of computational time since it is based on the exhaustive search. Some heuristic methods have been proposed to reduce the computational time, but they are not as accurate as that of Daugman. In this paper, we propose a pupil and iris boundary finding algorithm which is faster than and as accurate as that of Daugman. The proposed algorithm searches the boundaries using the Daugman's circular edge detector, but reduces the search region using the problem domain knowledge. In order to find the pupil boundary, the search region is restricted in the maximum and minimum bounding circles in which the pupil resides. The bounding circles are obtained from the binarized pupil image. Two iris boundary points are obtained from the horizontal line passing through the center of the pupil region obtained above. These initial boundary points, together with the pupil point comprise two bounding circles. The iris boundary is searched in this bounding circles. Experiments show that the proposed algorithm is faster than that of Daugman and more accurate than the conventional heuristic methods.