• Title/Summary/Keyword: Hand segmentation

Search Result 119, Processing Time 0.026 seconds

Deep Multi-task Network for Simultaneous Hazy Image Semantic Segmentation and Dehazing (안개영상의 의미론적 분할 및 안개제거를 위한 심층 멀티태스크 네트워크)

  • Song, Taeyong;Jang, Hyunsung;Ha, Namkoo;Yeon, Yoonmo;Kwon, Kuyong;Sohn, Kwanghoon
    • Journal of Korea Multimedia Society
    • /
    • v.22 no.9
    • /
    • pp.1000-1010
    • /
    • 2019
  • Image semantic segmentation and dehazing are key tasks in the computer vision. In recent years, researches in both tasks have achieved substantial improvements in performance with the development of Convolutional Neural Network (CNN). However, most of the previous works for semantic segmentation assume the images are captured in clear weather and show degraded performance under hazy images with low contrast and faded color. Meanwhile, dehazing aims to recover clear image given observed hazy image, which is an ill-posed problem and can be alleviated with additional information about the image. In this work, we propose a deep multi-task network for simultaneous semantic segmentation and dehazing. The proposed network takes single haze image as input and predicts dense semantic segmentation map and clear image. The visual information getting refined during the dehazing process can help the recognition task of semantic segmentation. On the other hand, semantic features obtained during the semantic segmentation process can provide cues for color priors for objects, which can help dehazing process. Experimental results demonstrate the effectiveness of the proposed multi-task approach, showing improved performance compared to the separate networks.

A methodology for spatial distribution of grain and voids in self compacting concrete using digital image processing methods

  • Onal, Okan;Ozden, Gurkan;Felekoglu, Burak
    • Computers and Concrete
    • /
    • v.5 no.1
    • /
    • pp.61-74
    • /
    • 2008
  • Digital image processing algorithms for the analysis and characterization of grains and voids in cemented materials were developed using toolbox functions of a mathematical software package. Utilization of grayscale, color and watershed segmentation algorithms and their performances were demonstrated on artificially prepared self-compacting concrete (SCC) samples. It has been found that color segmentation was more advantageous over the gray scale segmentation for the detection of voids whereas the latter method provided satisfying results for the aggregate grains due to the sharp contrast between their colors and the cohesive matrix. The watershed segmentation method, on the other hand, appeared to be very efficient while separating touching objects in digital images.

The Error Pattern Analysis of the HMM-Based Automatic Phoneme Segmentation (HMM기반 자동음소분할기의 음소분할 오류 유형 분석)

  • Kim Min-Je;Lee Jung-Chul;Kim Jong-Jin
    • The Journal of the Acoustical Society of Korea
    • /
    • v.25 no.5
    • /
    • pp.213-221
    • /
    • 2006
  • Phone segmentation of speech waveform is especially important for concatenative text to speech synthesis which uses segmented corpora for the construction of synthetic units. because the quality of synthesized speech depends critically on the accuracy of the segmentation. In the beginning. the phone segmentation was manually performed. but it brings the huge effort and the large time delay. HMM-based approaches adopted from automatic speech recognition are most widely used for automatic segmentation in speech synthesis, providing a consistent and accurate phone labeling scheme. Even the HMM-based approach has been successful, it may locate a phone boundary at a different position than expected. In this paper. we categorized adjacent phoneme pairs and analyzed the mismatches between hand-labeled transcriptions and HMM-based labels. Then we described the dominant error patterns that must be improved for the speech synthesis. For the experiment. hand labeled standard Korean speech DB from ETRI was used as a reference DB. Time difference larger than 20ms between hand-labeled phoneme boundary and auto-aligned boundary is treated as an automatic segmentation error. Our experimental results from female speaker revealed that plosive-vowel, affricate-vowel and vowel-liquid pairs showed high accuracies, 99%, 99.5% and 99% respectively. But stop-nasal, stop-liquid and nasal-liquid pairs showed very low accuracies, 45%, 50% and 55%. And these from male speaker revealed similar tendency.

Hand Gesture Segmentation Method using a Wrist-Worn Wearable Device

  • Lee, Dong-Woo;Son, Yong-Ki;Kim, Bae-Sun;Kim, Minkyu;Jeong, Hyun-Tae;Cho, Il-Yeon
    • Journal of the Ergonomics Society of Korea
    • /
    • v.34 no.5
    • /
    • pp.541-548
    • /
    • 2015
  • Objective: We introduce a hand gesture segmentation method using a wrist-worn wearable device which can recognize simple gestures of clenching and unclenching ones' fist. Background: There are many types of smart watches and fitness bands in the markets. And most of them already adopt a gesture interaction to provide ease of use. However, there are many cases in which the malfunction is difficult to distinguish between the user's gesture commands and user's daily life motion. It is needed to develop a simple and clear gesture segmentation method to improve the gesture interaction performance. Method: At first, we defined the gestures of making a fist (start of gesture command) and opening one's fist (end of gesture command) as segmentation gestures to distinguish a gesture. The gestures of clenching and unclenching one's fist are simple and intuitive. And we also designed a single gesture consisting of a set of making a fist, a command gesture, and opening one's fist in order. To detect segmentation gestures at the bottom of the wrist, we used a wrist strap on which an array of infrared sensors (emitters and receivers) were mounted. When a user takes gestures of making a fist and opening one's a fist, this changes the shape of the bottom of the wrist, and simultaneously changes the reflected amount of the infrared light detected by the receiver sensor. Results: An experiment was conducted in order to evaluate gesture segmentation performance. 12 participants took part in the experiment: 10 males, and 2 females with an average age of 38. The recognition rates of the segmentation gestures, clenching and unclenching one's fist, are 99.58% and 100%, respectively. Conclusion: Through the experiment, we have evaluated gesture segmentation performance and its usability. The experimental results show a potential for our suggested segmentation method in the future. Application: The results of this study can be used to develop guidelines to prevent injury in auto workers at mission assembly plants.

Controlling Slides using Hand tracking and Gesture Recognition (손의 추적과 제스쳐 인식에 의한 슬라이드 제어)

  • Fayyaz, Rabia;Rhee, Eun Joo
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2012.04a
    • /
    • pp.436-439
    • /
    • 2012
  • The work is to the control the desktop Computers based on hand gesture recognition. This paper is worked en real time tracking and recognizes the hand gesture for controlling the slides based on hand direction such as right and left using a real time camera.

An Efficient Data Augmentation for 3D Medical Image Segmentation (3차원 의료 영상의 영역 분할을 위한 효율적인 데이터 보강 방법)

  • Park, Sangkun
    • Journal of Institute of Convergence Technology
    • /
    • v.11 no.1
    • /
    • pp.1-5
    • /
    • 2021
  • Deep learning based methods achieve state-of-the-art accuracy, however, they typically rely on supervised training with large labeled datasets. It is known in many medical applications that labeling medical images requires significant expertise and much time, and typical hand-tuned approaches for data augmentation fail to capture the complex variations in such images. This paper proposes a 3D image augmentation method to overcome these difficulties. It allows us to enrich diversity of training data samples that is essential in medical image segmentation tasks, thus reducing the data overfitting problem caused by the fact the scale of medical image dataset is typically smaller. Our numerical experiments demonstrate that the proposed approach provides significant improvements over state-of-the-art methods for 3D medical image segmentation.

Improvement of an Automatic Segmentation for TTS Using Voiced/Unvoiced/Silence Information (유/무성/묵음 정보를 이용한 TTS용 자동음소분할기 성능향상)

  • Kim Min-Je;Lee Jung-Chul;Kim Jong-Jin
    • MALSORI
    • /
    • no.58
    • /
    • pp.67-81
    • /
    • 2006
  • For a large corpus of time-aligned data, HMM based approaches are most widely used for automatic segmentation, providing a consistent and accurate phone labeling scheme. There are two methods for training in HMM. Flat starting method has a property that human interference is minimized but it has low accuracy. Bootstrap method has a high accuracy, but it has a defect that manual segmentation is required In this paper, a new algorithm is proposed to minimize manual work and to improve the performance of automatic segmentation. At first phase, voiced, unvoiced and silence classification is performed for each speech data frame. At second phase, the phoneme sequence is aligned dynamically to the voiced/unvoiced/silence sequence according to the acoustic phonetic rules. Finally, using these segmented speech data as a bootstrap, phoneme model parameters based on HMM are trained. For the performance test, hand labeled ETRI speech DB was used. The experiment results showed that our algorithm achieved 10% improvement of segmentation accuracy within 20 ms tolerable error range. Especially for the unvoiced consonants, it showed 30% improvement.

  • PDF

Interactive visual knowledge acquisition for hand-gesture recognition (손 제스쳐 인식을 위한 상호작용 시각정보 추출)

  • 양선옥;최형일
    • Journal of the Korean Institute of Telematics and Electronics B
    • /
    • v.33B no.9
    • /
    • pp.88-96
    • /
    • 1996
  • Computer vision-based gesture recognition systems consist of image segmentation, object tracking and decision. However, it is difficult to segment an object from image for gesture in computer systems because of vaious illuminations and backgrounds. In this paper, we describe a method to learn features for segmentation, which improves the performance of computer vision-based hand-gesture recognition systems. Systems interact with a user to acquire exact training data and segment information according to a predefined plan. System provides some models to the user, takes pictures of the user's response and then analyzes the pictures with models and a prior knowledge. The system sends messages to the user and operates learning module to extract information with the analyzed result.

  • PDF

Segmentation and 3D Visualization of Medical Image : An Overview

  • Kang, Jiwoo;Kim, Doyoung;Lee, Sanghoon
    • Journal of International Society for Simulation Surgery
    • /
    • v.1 no.1
    • /
    • pp.27-31
    • /
    • 2014
  • In this paper, an overview of segmentation and 3D visualization methods are presented. Commonly, the two kinds of methods are used to visualize organs and vessels into 3D from medical images such as CT(A) and MRI - Direct Volume Rendering (DVR) and Iso-surface Rendering (IR). DVR can be applied directly to a volume. It directly penetrates through the volume while it determines which voxels are visualizedbased on a transfer function. On the other hand, IR requires a series of processes such as segmentation, polygonization and visualization. To extract a region of interest (ROI) from the medical volume image via the segmentation, some regions of an object and a background are required, which are typically obtained from the user. To visualize the extracted regions, the boundary points of the regions should be polygonized. In other words, the boundary surface composed of polygons such as a triangle and a rectangle should be required to visualize the regions into 3D because illumination effects, which makes the object shaded and seen in 3D, cannot be applied directly to the points.

Segmentation of Bacterial Cells Based on a Hybrid Feature Generation and Deep Learning (하이브리드 피처 생성 및 딥 러닝 기반 박테리아 세포의 세분화)

  • Lim, Seon-Ja;Vununu, Caleb;Kwon, Ki-Ryong;Youn, Sung-Dae
    • Journal of Korea Multimedia Society
    • /
    • v.23 no.8
    • /
    • pp.965-976
    • /
    • 2020
  • We present in this work a segmentation method of E. coli bacterial images generated via phase contrast microscopy using a deep learning based hybrid feature generation. Unlike conventional machine learning methods that use the hand-crafted features, we adopt the denoising autoencoder in order to generate a precise and accurate representation of the pixels. We first construct a hybrid vector that combines original image, difference of Gaussians and image gradients. The created hybrid features are then given to a deep autoencoder that learns the pixels' internal dependencies and the cells' shape and boundary information. The latent representations learned by the autoencoder are used as the inputs of a softmax classification layer and the direct outputs from the classifier represent the coarse segmentation mask. Finally, the classifier's outputs are used as prior information for a graph partitioning based fine segmentation. We demonstrate that the proposed hybrid vector representation manages to preserve the global shape and boundary information of the cells, allowing to retrieve the majority of the cellular patterns without the need of any post-processing.