• Title/Summary/Keyword: 분할 차영상

Search Result 311, Processing Time 0.029 seconds

A Study on Image Segmentation and Tracking based on Fuzzy Method (퍼지기법을 이용한 영상분할 및 물체추적에 관한 연구)

  • Lee, Min-Jung;Jin, Tae-Seok;Hwang, Gi-Hyung
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.17 no.3
    • /
    • pp.368-373
    • /
    • 2007
  • In recent year s there have been increasing interests in real-time object tracking with image information. This dissertation presents a real-time object tracking method through the object recognition based on neural networks that have robust characteristics under various illuminations. This dissertation proposes a global search and a local search method to track the object in real-time. The global search recognizes a target object among the candidate objects through the entire image search, and the local search recognizes and track only the target object through the block search. This dissertation uses the object color and feature information to achieve fast object recognition. The experiment result shows the usefulness of the proposed method is verified.

Jpeg Sequence display system in Android (안드로이드 폰에서 Jpeg Sequence 출력 시스템)

  • Kang, Hyun-Min;Cho, Tai-Hoon
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2010.05a
    • /
    • pp.614-616
    • /
    • 2010
  • Recently, there is increasing concern about smart phone due to the release of iPhone and Android phone. So, There are many kinds of application in smart phone, and applications using multimedia contents is increasing. The famous codecs for Image and video are JPEG, MPEG and h.264, which is popular recently as a vedeo codec because of its high compression ratio compared with the other video codecs. In this paper, we will divide Jpeg Sequence to each Jpeg and will display that Jpeg looks like a video continuously in Android. And we will show you the simulation result.

  • PDF

Image Transmission Using Designed Source-Channel Combined Coder for Mobile Communication Systems (이동통신 시스템을 위한 소스코더와 결합된 채널코딩방법에 의한 영상전송)

  • Lee, Byung-Gil;Park, Pan-Jong;Cho, Hyun-Wook;Park, Gil-houm
    • Journal of the Institute of Electronics Engineers of Korea TC
    • /
    • v.37 no.1
    • /
    • pp.66-75
    • /
    • 2000
  • In this paper, we present the efficient image transmission system using designed source-channel combined coder in W-CDMA mobile communication system. In proposed schemes, we decompose the wavelet transformed hierarchical band-images into some types of different size blocks which have different properties in error sensitivity. The RS(Reed-Solomon) coder with different coding rate is used for each decomposed source blocks which has different importance. In addition, we combine retransmitted error frames in Truncated Hybrid Type I ARQ. The proposed algorithm shows efficient image transmission methods because it is not much degraded in PSNR compared with the existing not combined source-channel coder in erroneous wireless channel.

  • PDF

Image Restoration using Pattern of Non-noise Pixels in Impulse Noise Environments (임펄스 잡음 환경에서 비잡음 화소의 패턴을 사용한 영상복원)

  • Cheon, Bong-Won;Kim, Marn-Go;Kim, Nam-Ho
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.10a
    • /
    • pp.407-409
    • /
    • 2021
  • Under the influence of the 4th industrial revolution, various technologies such as artificial intelligence and automation are being grafted into industrial sites, and accordingly, the importance of data processing is increasing. Digital images may generate noise due to various reasons, and may affect various systems such as image recognition and classification and object tracking. To compensate for these shortcomings, we propose an image restoration algorithm based on pattern information of non-noise pixels. According to the distribution of non-noise pixels inside the filtering mask, the proposed algorithm switched the filtering process by dividing the interpolation method into a pattern that can be applied, a pattern based on region division, and a randomly arranged pixel pattern. preserves and restores the image. The proposed algorithm showed superior performance compared to the existing impulse noise removal algorithm.

  • PDF

HEVC Encoder Optimization using Depth Information (깊이정보를 이용한 HEVC의 인코더 고속화 방법)

  • Lee, Yoon Jin;Bae, Dong In;Park, Gwang Hoon
    • Journal of Broadcast Engineering
    • /
    • v.19 no.5
    • /
    • pp.640-655
    • /
    • 2014
  • Many of today's video systems have additional depth camera to provide extra features such as 3D support. Thanks to these changes made in multimedia system, it is now much easier to obtain depth information of the video. Depth information can be used in various areas such as object classification, background area recognition, and so on. With depth information, we can achieve even higher coding efficiency compared to only using conventional method. Thus, in this paper, we propose the 2D video coding algorithm which uses depth information on top of the next generation 2D video codec HEVC. Background area can be recognized with depth information and by performing HEVC with it, coding complexity can be reduced. If current CU is background area, we propose the following three methods, 1) Earlier stop split structure of CU with PU SKIP mode, 2) Limiting split structure of CU with CU information in temporal position, 3) Limiting the range of motion searching. We implement our proposal using HEVC HM 12.0 reference software. With these methods results shows that encoding complexity is reduced more than 40% with only 0.5% BD-Bitrate loss. Especially, in case of video acquired through the Kinect developed by Microsoft Corp., encoding complexity is reduced by max 53% without a loss of quality. So, it is expected that these techniques can apply real-time online communication, mobile or handheld video service and so on.

A Method of Hand Recognition for Virtual Hand Control of Virtual Reality Game Environment (가상 현실 게임 환경에서의 가상 손 제어를 위한 사용자 손 인식 방법)

  • Kim, Boo-Nyon;Kim, Jong-Ho;Kim, Tae-Young
    • Journal of Korea Game Society
    • /
    • v.10 no.2
    • /
    • pp.49-56
    • /
    • 2010
  • In this paper, we propose a control method of virtual hand by the recognition of a user's hand in the virtual reality game environment. We display virtual hand on the game screen after getting the information of the user's hand movement and the direction thru input images by camera. We can utilize the movement of a user's hand as an input interface for virtual hand to select and move the object. As a hand recognition method based on the vision technology, the proposed method transforms input image from RGB color space to HSV color space, then segments the hand area using double threshold of H, S value and connected component analysis. Next, The center of gravity of the hand area can be calculated by 0 and 1 moment implementation of the segmented area. Since the center of gravity is positioned onto the center of the hand, the further apart pixels from the center of the gravity among the pixels in the segmented image can be recognized as fingertips. Finally, the axis of the hand is obtained as the vector of the center of gravity and the fingertips. In order to increase recognition stability and performance the method using a history buffer and a bounding box is also shown. The experiments on various input images show that our hand recognition method provides high level of accuracy and relatively fast stable results.

A Study on Model for Drivable Area Segmentation based on Deep Learning (딥러닝 기반의 주행가능 영역 추출 모델에 관한 연구)

  • Jeon, Hyo-jin;Cho, Soo-sun
    • Journal of Internet Computing and Services
    • /
    • v.20 no.5
    • /
    • pp.105-111
    • /
    • 2019
  • Core technologies that lead the Fourth Industrial Revolution era, such as artificial intelligence, big data, and autonomous driving, are implemented and serviced through the rapid development of computing power and hyper-connected networks based on the Internet of Things. In this paper, we implement two different models for drivable area segmentation in various environment, and propose a better model by comparing the results. The models for drivable area segmentation are using DeepLab V3+ and Mask R-CNN, which have great performances in the field of image segmentation and are used in many studies in autonomous driving technology. For driving information in various environment, we use BDD dataset which provides driving videos and images in various weather conditions and day&night time. The result of two different models shows that Mask R-CNN has higher performance with 68.33% IoU than DeepLab V3+ with 48.97% IoU. In addition, the result of visual inspection of drivable area segmentation on driving image, the accuracy of Mask R-CNN is 83% and DeepLab V3+ is 69%. It indicates Mask R-CNN is more efficient than DeepLab V3+ in drivable area segmentation.

A Robust Object Extraction Method for Immersive Video Conferencing (몰입형 화상 회의를 위한 강건한 객체 추출 방법)

  • Ahn, Il-Koo;Oh, Dae-Young;Kim, Jae-Kwang;Kim, Chang-Ick
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.48 no.2
    • /
    • pp.11-23
    • /
    • 2011
  • In this paper, an accurate and fully automatic video object segmentation method is proposed for video conferencing systems in which the real-time performance is required. The proposed method consists of two steps: 1) accurate object extraction on the initial frame, 2) real-time object extraction from the next frame using the result of the first step. Object extraction on the initial frame starts with generating a cumulative edge map obtained from frame differences in the beginning. This is because we can estimate the initial shape of the foreground object from the cumulative motion. This estimated shape is used to assign the seeds for both object and background, which are needed for Graph-Cut segmentation. Once the foreground object is extracted by Graph-Cut segmentation, real-time object extraction is conducted using the extracted object and the double edge map obtained from the difference between two successive frames. Experimental results show that the proposed method is suitable for real-time processing even in VGA resolution videos contrary to previous methods, being a useful tool for immersive video conferencing systems.

Automatic Text Extraction from News Video using Morphology and Text Shape (형태학과 문자의 모양을 이용한 뉴스 비디오에서의 자동 문자 추출)

  • Jang, In-Young;Ko, Byoung-Chul;Kim, Kil-Cheon;Byun, Hye-Ran
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.8 no.4
    • /
    • pp.479-488
    • /
    • 2002
  • In recent years the amount of digital video used has risen dramatically to keep pace with the increasing use of the Internet and consequently an automated method is needed for indexing digital video databases. Textual information, both superimposed and embedded scene texts, appearing in a digital video can be a crucial clue for helping the video indexing. In this paper, a new method is presented to extract both superimposed and embedded scene texts in a freeze-frame of news video. The algorithm is summarized in the following three steps. For the first step, a color image is converted into a gray-level image and applies contrast stretching to enhance the contrast of the input image. Then, a modified local adaptive thresholding is applied to the contrast-stretched image. The second step is divided into three processes: eliminating text-like components by applying erosion, dilation, and (OpenClose+CloseOpen)/2 morphological operations, maintaining text components using (OpenClose+CloseOpen)/2 operation with a new Geo-correction method, and subtracting two result images for eliminating false-positive components further. In the third filtering step, the characteristics of each component such as the ratio of the number of pixels in each candidate component to the number of its boundary pixels and the ratio of the minor to the major axis of each bounding box are used. Acceptable results have been obtained using the proposed method on 300 news images with a recognition rate of 93.6%. Also, my method indicates a good performance on all the various kinds of images by adjusting the size of the structuring element.

A Study on Class Sample Extraction Technique Using Histogram Back-Projection for Object-Based Image Classification (객체 기반 영상 분류를 위한 히스토그램 역투영을 이용한 클래스 샘플 추출 기법에 관한 연구)

  • Chul-Soo Ye
    • Korean Journal of Remote Sensing
    • /
    • v.39 no.2
    • /
    • pp.157-168
    • /
    • 2023
  • Image segmentation and supervised classification techniques are widely used to monitor the ground surface using high-resolution remote sensing images. In order to classify various objects, a process of defining a class corresponding to each object and selecting samples belonging to each class is required. Existing methods for extracting class samples should select a sufficient number of samples having similar intensity characteristics for each class. This process depends on the user's visual identification and takes a lot of time. Representative samples of the class extracted are likely to vary depending on the user, and as a result, the classification performance is greatly affected by the class sample extraction result. In this study, we propose an image classification technique that minimizes user intervention when extracting class samples by applying the histogram back-projection technique and has consistent intensity characteristics of samples belonging to classes. The proposed classification technique using histogram back-projection showed improved classification accuracy in both the experiment using hue subchannels of the hue saturation value transformed image from Compact Advanced Satellite 500-1 imagery and the experiment using the original image compared to the technique that did not use histogram back-projection.