• Title/Summary/Keyword: Video Scene Detection

Search Result 190, Processing Time 0.027 seconds

Lane Detection Based on a Cumulative Distribution function of Edge Direction (에지 방향의 누적분포함수에 기반한 차선인식)

  • Yi, Un-Kun;Baek, Kwang-Ryul;Lee, Joon-Woong
    • Proceedings of the KIEE Conference
    • /
    • 2000.07d
    • /
    • pp.2814-2818
    • /
    • 2000
  • This paper describes an image processing algorithm capable of recognizing the road lane using a CDF (Cumulative Distribution Function). which is designed for the model function of the road lane. The CDF has distinctive peak points at the vicinity of the lane direction because of the directional and positional continuities of the lane. We construct a scatter diagram by collecting the edge pixels with the direction corresponding to the peak point of the CDF and carry out the principal axis-based line fitting for the scatter diagram to obtain the lane information. As noises play the role of making a lot of similar features to the lane appear and disappear in the image we introduce a recursive estimator of the function to reduce the noise effect and a scene understanding index (SUI) formulated by statistical parameters of the CDF to prevent a false alarm or miss detection. The proposed algorithm has been implemented in a real time on the video data obtained from a test vehicle driven in a typical highway.

  • PDF

Road-Lane Detection Based on a Cumulative Distribution Function of Edge Direction

  • Yi, Un-Kun;Lee, Joon-Woong;Baek, Kwang-Ryul
    • Journal of KIEE
    • /
    • v.11 no.1
    • /
    • pp.69-77
    • /
    • 2001
  • This paper describes an image processing algorithm capable of recognizing road lanes by using a CDF(cumulative distribution function). The CDF is designed for the model function of road lanes. Based on the assumptions that there are no abrupt changes in the direction and location of road lanes and that the intensity of lane boundaries differs from that of the background, we formulated the CDF, which accumulates the edge magnitude for edge directions. The CDF has distinctive peak points at the vicinity of lane directions due to the directional and the positional continuities of a lane. To obtain lane-related information a scatter diagram was constructed by collecting edge pixels, of which the direction corresponds to the peak point of the CDF, then the principal axis-based line fitting was performed for the scatter diagram. Noises can cause many similar features to appear and to disappear in an image. Therefore, to reduce the noise effect a recursive estimator of the CDF was introduced, and also to prevent false alarms or miss detection a scene understanding index (DUI) was formulated by the statistical parameters of the CDF. The proposed algorithm has been implemented in real time on video data obtained from a test vehicle driven on a typical highway.

  • PDF

Content based Video Segmentation Algorithm using Comparison of Pattern Similarity (장면의 유사도 패턴 비교를 이용한 내용기반 동영상 분할 알고리즘)

  • Won, In-Su;Cho, Ju-Hee;Na, Sang-Il;Jin, Ju-Kyong;Jeong, Jae-Hyup;Jeong, Dong-Seok
    • Journal of Korea Multimedia Society
    • /
    • v.14 no.10
    • /
    • pp.1252-1261
    • /
    • 2011
  • In this paper, we propose the comparison method of pattern similarity for video segmentation algorithm. The shot boundary type is categorized as 2 types, abrupt change and gradual change. The representative examples of gradual change are dissolve, fade-in, fade-out or wipe transition. The proposed method consider the problem to detect shot boundary as 2-class problem. We concentrated if the shot boundary event happens or not. It is essential to define similarity between frames for shot boundary detection. We proposed 2 similarity measures, within similarity and between similarity. The within similarity is defined by feature comparison between frames belong to same shot. The between similarity is defined by feature comparison between frames belong to different scene. Finally we calculated the statistical patterns comparison between the within similarity and between similarity. Because this measure is robust to flash light or object movement, our proposed algorithm make contribution towards reducing false positive rate. We employed color histogram and mean of sub-block on frame image as frame feature. We performed the experimental evaluation with video dataset including set of TREC-2001 and TREC-2002. The proposed algorithm shows the performance, 91.84% recall and 86.43% precision in experimental circumstance.

A 3-D Vision Sensor Implementation on Multiple DSPs TMS320C31 (다중 TMS320C31 DSP를 사용한 3-D 비젼센서 Implementation)

  • Oksenhendler, V.;Bensrhair, Abdelaziz;Miche, Pierre;Lee, Sang-Goog
    • Journal of Sensor Science and Technology
    • /
    • v.7 no.2
    • /
    • pp.124-130
    • /
    • 1998
  • High-speed 3D vision systems are essential for autonomous robot or vehicle control applications. In our study, a stereo vision process has been developed. It consists of three steps : extraction of edges in right and left images, matching corresponding edges and calculation of the 3D map. This process is implemented in a VME 150/40 Imaging Technology vision system. It is a modular system composed by a display, an acquisition, a four Mbytes image frame memory, and three computational cards. Programmable accelerator computational modules are running at 40 MHz and are based on TMS320C31 DSP with a $64{\times}32$ bit instruction cache and two $1024{\times}32$ bit internal RAMs. Each is equipped with 512 Kbytes static RAM, 4 Mbytes image memory, 1 Mbytes flash EEPROM and a serial port. Data transfers and communications between modules are provided by three 8 bit global video bus, and three local configurable pipeline 8 bit video bus. The VME bus is dedicated to system management. Tasks between DSPs are distributed as follows: two DSPs are used to edges detection, one for the right image and the other for the left one. The last processor computes the matching process and the 3D calculation. With $512{\times}512$ pixels images, this sensor generates dense 3D maps at a rate of about 1 Hz depending of the scene complexity. Results can surely be improved by using a special suited multiprocessors cards.

  • PDF

A Study on Object Detection Algorithm for Abandoned and Removed Objects for Real-time Intelligent Surveillance System (실시간 지능형 감시 시스템을 위한 방치, 제거된 객체 검출에 관한 연구)

  • Jeon, Ji-Hye;Park, Jong-Hwa;Jeong, Cheol-Jun;Kang, In-Goo;An, Tae-Ki;Park, Goo-Man
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.35 no.1C
    • /
    • pp.24-32
    • /
    • 2010
  • In this paper we proposed an object tracking system that detects the abandoned and removed objects, which is to be used in the intelligent surveillance applications. After the GMM based background subtraction and by using histogram method, the static region is identified to detect abandoned and removed objects. Since the system is implemented on DSP chip, it operates in realtime and is programmable. The input videos used in the experiment contain various indoor and outdoor scenes, and they are categorized into three different complexities; low, midium and high. By 10 times of experiment, we obtained high detection ratio at low and medium complexity sequences. On the high complexity video, successful detection ratio was relatively low because the scene contains crowdedness and repeated occlusion. In the future work, these complicated situation should be solved.

Improvement of Character-net via Detection of Conversation Participant (대화 참여자 결정을 통한 Character-net의 개선)

  • Kim, Won-Taek;Park, Seung-Bo;Jo, Geun-Sik
    • Journal of the Korea Society of Computer and Information
    • /
    • v.14 no.10
    • /
    • pp.241-249
    • /
    • 2009
  • Recently, a number of researches related to video annotation and representation have been proposed to analyze video for searching and abstraction. In this paper, we have presented a method to provide the picture elements of conversational participants in video and the enhanced representation of the characters using those elements, collectively called Character-net. Because conversational participants are decided as characters detected in a script holding time, the previous Character-net suffers serious limitation that some listeners could not be detected as the participants. The participants who complete the story in video are very important factor to understand the context of the conversation. The picture elements for detecting the conversational participants consist of six elements as follows: subtitle, scene, the order of appearance, characters' eyes, patterns, and lip motion. In this paper, we present how to use those elements for detecting conversational participants and how to improve the representation of the Character-net. We can detect the conversational participants accurately when the proposed elements combine together and satisfy the special conditions. The experimental evaluation shows that the proposed method brings significant advantages in terms of both improving the detection of the conversational participants and enhancing the representation of Character-net.

A Low Cost 3D Skin Wrinkle Reconstruction System Based on Stereo Semi-Dense Matching (반 밀집 정합에 기반한 저가형 3차원 주름 데이터 복원)

  • Zhang, Qian;WhangBo, Taeg-Keun
    • Journal of Internet Computing and Services
    • /
    • v.10 no.4
    • /
    • pp.25-33
    • /
    • 2009
  • In the paper, we proposed a new system to retrieve 3D wrinkle data based on stereo images. Usually, 3D reconstruction based on stereo images or video is very popular and it is the research focus, which has been applied for culture heritage, building and other scene. The target is object measurement, the scene depth calculation and 3D data obtained. There are several challenges in our research. First, it is hard to take the full information wrinkle images by cameras because of light influence, skin with non-rigid object and camera performance. We design a particular computer vision system to take winkle images with a long length camera lens. Second, it is difficult to get the dense stereo data because of the hard skin texture image segmentation and corner detection. We focus on semi-dense stereo matching algorithm for the wrinkle depth. Compared with the 3D scanner, our system is much cheaper and compared with the physical modeling based method, our system is more flexible with high performance.

  • PDF

A Real-time Motion Object Detection based on Neighbor Foreground Pixel Propagation Algorithm (주변 전경 픽셀 전파 알고리즘 기반 실시간 이동 객체 검출)

  • Nguyen, Thanh Binh;Chung, Sun-Tae
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.47 no.1
    • /
    • pp.9-16
    • /
    • 2010
  • Moving object detection is to detect foreground object different from background scene in a new incoming image frame and is an essential ingredient process in some image processing applications such as intelligent visual surveillance, HCI, object-based video compression and etc. Most of previous object detection algorithms are still computationally heavy so that it is difficult to develop real-time multi-channel moving object detection in a workstation or even one-channel real-time moving object detection in an embedded system using them. Foreground mask correction necessary for a more precise object detection is usually accomplished using morphological operations like opening and closing. Morphological operations are not computationally cheap and moreover, they are difficult to be rendered to run simultaneously with the subsequent connected component labeling routine since they need quite different type of processing from what the connected component labeling does. In this paper, we first devise a fast and precise foreground mask correction algorithm, "Neighbor Foreground Pixel Propagation (NFPP)" which utilizes neighbor pixel checking employed in the connected component labeling. Next, we propose a novel moving object detection method based on the devised foreground mask correction algorithm, NFPP where the connected component labeling routine can be executed simultaneously with the foreground mask correction. Through experiments, it is verified that the proposed moving object detection method shows more precise object detection and more than 4 times faster processing speed for a image frame and videos in the given the experiments than the previous moving object detection method using morphological operations.

Effective Detection Techniques for Gradual Scene Change on MPEG Video (통신망 특성 파라미터 추출 및 Field data 분석)

  • 조영래
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.24 no.8A
    • /
    • pp.1208-1217
    • /
    • 1999
  • Being increased the errention and competition for communication business, the economical efficiency and the reliance of communication network were raised. The communication operators are sensitively reactioned to the data related communication. Especially, the trend of hazards related to communication networks and users is broadly deviede into 3 classes in the technical specificatons of foreigners. In this paper, the field data of parameters corresponding to 3 classes was statistically processed and compared to the specified values in the technical specification. The ststistical processes of field data play a role as an objective information for the service quality related to the communication network, and will be helpful for preparing the globalization of communication.

  • PDF

A Scene Boundary detection Scheme Using Video and Audio Information of MPEG Stream (MPEG 스트림에서의 비디오 및 오디오 정보를 이용한 신 경계 검출 방법)

  • 김재홍;강찬미;남종호;김경수;하명환;정경희
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2001.04b
    • /
    • pp.397-399
    • /
    • 2001
  • 본 논문에서는 MPEG 형식으로 압축된 동영상을 데이터에 대하여 비디오 및 오디오 정보를모두 이용하는 새로운 신 경계 검출방법을 제안하고 여러 실험을 통해서 그 유용성을 증명한다. 즉, 본 논문에서는 DC이미지 형태의 대표 프레임을 바탕으로 한 비디오 기반 신 경계 검출방법[8]과 dB값을 이용한 오디오 기반 신 경계 검출방법[9]을 결합하는 방법을 제안한다. 제안한 방법에서는 두 방법에서 모두 신으로 검출한 경계에 대하여서는 신으로 인정하고, 검출한 결과가 다른 경우에 대하여서는 각각의 경계 데이터를 좀 더 자세히 분석하여 신 경계를 검출하도록 한다. 비디오 기반 신 경계 검출방법에서만 검출된 신 경계에 대해서는 그 경계 데이터에 대해서 dB값의 차이를 해당 시간범위 내에서 다시 비교하여 신 경계 여부를 판단하고, 오디오 기반 신 경계 검출방법에서만 검출되 신 경계에 대해서는 그 경계 데이터에 대해서 샷의 유사도를 샷의 개수에 관계없이 시간의 임계치만 고려해서 비교한 다음 신 경계 여부를 판단하게 된다. 이러한 방법으로 신 경계를 검출한 결과를 살펴보면 Precision측면에서는 최고24%까지, Recall측면에서는 최고25%까지 효율을 높이고 있음을 알 수 있다. 이러한 알고리즘은 기존의 신 경계 검출 방법 보다 높은 효율을 제공하여 비디오 데이터를 사용하는 여러 응용분야에서의 프로그램 개발에 이용될 수 있을 것이다.

  • PDF