• Title/Summary/Keyword: information cues

Search Result 221, Processing Time 0.03 seconds

2D-to-3D Conversion System using Depth Map Enhancement

  • Chen, Ju-Chin;Huang, Meng-yuan
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.10 no.3
    • /
    • pp.1159-1181
    • /
    • 2016
  • This study introduces an image-based 2D-to-3D conversion system that provides significant stereoscopic visual effects for humans. The linear and atmospheric perspective cues that compensate each other are employed to estimate depth information. Rather than retrieving a precise depth value for pixels from the depth cues, a direction angle of the image is estimated and then the depth gradient, in accordance with the direction angle, is integrated with superpixels to obtain the depth map. However, stereoscopic effects of synthesized views obtained from this depth map are limited and dissatisfy viewers. To obtain impressive visual effects, the viewer's main focus is considered, and thus salient object detection is performed to explore the significance region for visual attention. Then, the depth map is refined by locally modifying the depth values within the significance region. The refinement process not only maintains global depth consistency by correcting non-uniform depth values but also enhances the visual stereoscopic effect. Experimental results show that in subjective evaluation, the subjectively evaluated degree of satisfaction with the proposed method is approximately 7% greater than both existing commercial conversion software and state-of-the-art approach.

Effects of phonological and phonetic information of vowels on perception of prosodic prominence in English

  • Suyeon Im
    • Phonetics and Speech Sciences
    • /
    • v.15 no.3
    • /
    • pp.1-7
    • /
    • 2023
  • This study investigates how the phonological and phonetic information of vowels influences prosodic prominence among linguistically untrained listeners using public speech in American English. We first examined the speech material's phonetic realization of vowels (i.e., maximum F0, F0 range, phone rate [as a measure of duration considering the speech rate of the utterance], and mean intensity). Results showed that the high vowels /i/ and /u/ likely had the highest max F0, while the low vowels /æ/ and /ɑ/ tended to have the highest mean intensity. Both high and low vowels had similarly high phone rates. Next, we examined the effects of the vowels' phonological and phonetic information on listeners' perceptions of prosodic prominence. The results showed that vowels significantly affected the likelihood of perceived prominence independent of acoustic cues. The high and low vowels affected probability of perceived prominence less than the mid vowels /ɛ/ and /ʌ/, although the former two were more likely to be phonetically enhanced in the speech than the latter. Overall, these results suggest that perceptions of prosodic prominence in English are not directly influenced by signal-driven factors (i.e., vowels' acoustic information) but are mediated by expectation-driven factors (e.g., vowels' phonological information).

An Efficient Method for Korean Noun Extraction Using Noun Patterns (명사 출현 특성을 이용한 효율적인 한국어 명사 추출 방법)

  • 이도길;이상주;임해창
    • Journal of KIISE:Software and Applications
    • /
    • v.30 no.1_2
    • /
    • pp.173-183
    • /
    • 2003
  • Morphological analysis is the most widely used method for extracting nouns from Korean texts. For every Eojeol, in order to extract nouns from it, a morphological analyzer performs frequent dictionary lookup and applies many morphonological rules, therefore it requires many operations. Moreover, a morphological analyzer generates all the possible morphological interpretations (sequences of morphemes) of a given Eojeol, which may by unnecessary from the noun extraction`s point of view. To reduce unnecessary computation of morphological analysis from the noun extraction`s point of view, this paper proposes a method for Korean noun extraction considering noun occurrence characteristics. Noun patterns denote conditions on which nouns are included in an Eojeol or not, which are positive cues or negative cues, respectively. When using the exclusive information as the negative cues, it is possible to reduce the search space of morphological analysis by ignoring Eojeols not including nouns. Post-noun syllable sequences(PNSS) as the positive cues can simply extract nouns by checking the part of the Eojeol preceding the PNSS and can guess unknown nouns. In addition, morphonological information is used instead of many morphonological rules in order to recover the lexical form from its altered surface form. Experimental results show that the proposed method can speed up without losing accuracy compared with other systems based on morphological analysis.

A WWW Images Automatic Annotation Based On Multi-cues Integration (멀티-큐 통합을 기반으로 WWW 영상의 자동 주석)

  • Shin, Seong-Yoon;Moon, Hyung-Yoon;Rhee, Yang-Won
    • Journal of the Korea Society of Computer and Information
    • /
    • v.13 no.4
    • /
    • pp.79-86
    • /
    • 2008
  • As the rapid development of the Internet, the embedded images in HTML web pages nowadays become predominant. For its amazing function in describing the content and attracting attention, images become substantially important in web pages. All these images consist a considerable database. What's more, the semantic meanings of images are well presented by the surrounding text and links. But only a small minority of these images have precise assigned keyphrases. and manually assigning keyphrases to existing images is very laborious. Therefore it is highly desirable to automate the keyphrases extraction process. In this paper, we first introduce WWW image annotation methods, based on low level features, page tags, overall word frequency and local word frequency. Then we put forward our method of multi-cues integration image annotation. Also, show multi-cue image annotation method is more superior than other method through an experiment.

  • PDF

Dual-stream Co-enhanced Network for Unsupervised Video Object Segmentation

  • Hongliang Zhu;Hui Yin;Yanting Liu;Ning Chen
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.18 no.4
    • /
    • pp.938-958
    • /
    • 2024
  • Unsupervised Video Object Segmentation (UVOS) is a highly challenging problem in computer vision as the annotation of the target object in the testing video is unknown at all. The main difficulty is to effectively handle the complicated and changeable motion state of the target object and the confusion of similar background objects in video sequence. In this paper, we propose a novel deep Dual-stream Co-enhanced Network (DC-Net) for UVOS via bidirectional motion cues refinement and multi-level feature aggregation, which can fully take advantage of motion cues and effectively integrate different level features to produce high-quality segmentation mask. DC-Net is a dual-stream architecture where the two streams are co-enhanced by each other. One is a motion stream with a Motion-cues Refine Module (MRM), which learns from bidirectional optical flow images and produces fine-grained and complete distinctive motion saliency map, and the other is an appearance stream with a Multi-level Feature Aggregation Module (MFAM) and a Context Attention Module (CAM) which are designed to integrate the different level features effectively. Specifically, the motion saliency map obtained by the motion stream is fused with each stage of the decoder in the appearance stream to improve the segmentation, and in turn the segmentation loss in the appearance stream feeds back into the motion stream to enhance the motion refinement. Experimental results on three datasets (Davis2016, VideoSD, SegTrack-v2) demonstrate that DC-Net has achieved comparable results with some state-of-the-art methods.

THE EFFECTIVENESS AND CHARACTERISTICS OF 3 POINT TASK ANALYSIS AS A NEW ERGONOMIC AND KANSEI DESIGN METHOD

  • Yamaoka, Toshiki;Matsunobe, Takuo
    • Proceedings of the Korean Society for Emotion and Sensibility Conference
    • /
    • 2001.05a
    • /
    • pp.15-19
    • /
    • 2001
  • This paper describes effectiveness and characteristics of 3 P(point) task analysis as a new Ergonomic and Kansei design method for extracting user demand especially. The key point in 3 P task analysis is to describe the flow of tasks and extract any problems in each task. A solution of a problem means a user demand. 3 P task analysis cal eliminate an oversight of check items by examining the users' information processing level. The suers' information processing level was divided into the following three stages for problem extraction: acquirement of information ---> understanding and judgment ---> operation. Three stages has fourteenth cues such as difficulty of seeing, no emphasis, mapping for extracting problems. To link analysis results to the formulation of a product concept. I added a column on the right side of the table for writing the requirements (user demand) to resolve the problems extracted from each task. The requirements are extracted by using seventh cues. Finally 3 P task analysis was compared with group interview to make the characteristics of 3 P task analysis, especially extracting user demand, clear.

  • PDF

Perceptual Cues for Korean Affricate vs. Fricative Distinction (한국어 마찰음과 파찰음의 변별 지각 단서)

  • Park, Soon-Boak;Yi, Bong-Won;Shin, Ji-Young;Kim, Kee-Ho
    • Speech Sciences
    • /
    • v.4 no.1
    • /
    • pp.47-58
    • /
    • 1998
  • This paper tests whether frication duration plays an important role in perceiving fricatives and affricates. In other words, whether the frication duration is an acoustic cue for distinguishing fricatives from affricates. For this purpose two types of experiments are carried out. In the first experiment two sets of stimuli are created by decreasing the frication duration of fricatives by 10ms, and in the second experiment two steps of stimuli are created by increasing the frication duration of affricates by 10 ms. The results of the present study show that frication duration is the primary acoustic cue in perceiving fricative/affricate distinction in Korean. In addition, the amplitude rise time and burst information appear to play secondary roles in perceiving the distinction by a series of perception tests.

  • PDF

A Study of Factors Affecting Group Polarization in Online Communication : Based on Anonymity (온라인 커뮤니케이션에서 집단극화 현상에 영향을 미치는 요인에 관한 연구: 익명성 관점에서)

  • Suh, Eung-Kyo
    • Journal of Distribution Science
    • /
    • v.13 no.2
    • /
    • pp.75-83
    • /
    • 2015
  • Purpose - This study aims to identify the effects of communication cues, anonymity, and social presence on group polarization in computer-mediated communication (CMC) settings. Extant literature has introduced some theoretical backgrounds of social presence and SIDE (Social Identity model of Deindividuation Effects) to explain the effects of communication cues and anonymity. The concept of social presence emphasized the mediating role on communication cues and anonymity. However, most literature did not measure social presence and compare group polarization of all condition groups. This does not sufficiently explain the result of group polarization. Research design, data, and methodology - We believe that the direct impact of anonymity on group polarization can provide a more admissible and clearer explanation for the results. In addition, this study categorizes anonymity into two levels, as anonymity of group and anonymity of self. To justify the anonymity view, a laboratory experiment was conducted. The experiment was conducted in communication cues settings (visual cue; without visual cue) and anonymity settings (identified; anonymous). Each of the four settings has 10 groups consisting of five subjects each (total 200 subjects). The subjects are undergraduates from a large university, majoring in business. All experimental procedures and calculations of choice shift and preference change follow the literature. Results - First, the removal of visual cues does not produce a significant impact on group polarization, which cannot be explained by the social presence view. Second, the anonymous condition does not significantly affect group polarization, which also cannot be explained by the social presence view. However, the anonymous condition directly affects group polarization. Specifically, anonymity of self has a stronger effect on group polarization than anonymity of group. The result explains about the leading factor affecting group polarization. This study examines another view of how computer-mediated communication may be associated with group polarization. The process and outcome data from the experiment reveal that group polarization is not affected by level of social presence, but by level of anonymity. Group discussions conducted with visual cue CMC setting and identified CMC setting result in weaker group polarization. Conversely, group discussions conducted without visual cue CMC setting and anonymous CMC setting lead to stronger group polarization. The results of the study have the following implications. First, they provide clues for business organizations to design the most appropriate media conditions and preemptive social conditions to implement when making group decisions through CMC, to maximize achievements, generate amicable agreements, or actively share information. Second, this study can be useful in analyzing different adverse effects generated through Internet use. Conclusions - This research can help explain discussions and decision-making actions on Internet forums, which have recently increased, as well as providing a foundational basis in newly establishing policies for the forums. Finally, it should be noted that many other factors such as group size, topics, and group history may affect group polarization. These should be examined in future studies.

Adaptive Depth Fusion based on Reliability of Depth Cues for 2D-to-3D Video Conversion (2차원 동영상의 3차원 변환을 위한 깊이 단서의 신뢰성 기반 적응적 깊이 융합)

  • Han, Chan-Hee;Choi, Hae-Chul;Lee, Si-Woong
    • The Journal of the Korea Contents Association
    • /
    • v.12 no.12
    • /
    • pp.1-13
    • /
    • 2012
  • 3D video is regarded as the next generation contents in numerous applications. The 2D-to-3D video conversion technologies are strongly required to resolve a lack of 3D videos during the period of transition to the full ripe 3D video era. In 2D-to-3D conversion methods, after the depth image of each scene in 2D video is estimated, stereoscopic video is synthesized using DIBR (Depth Image Based Rendering) technologies. This paper proposes a novel depth fusion algorithm that integrates multiple depth cues contained in 2D video to generate stereoscopic video. For the proper depth fusion, it is checked whether some cues are reliable or not in current scene. Based on the result of the reliability tests, current scene is classified into one of 4 scene types and scene-adaptive depth fusion is applied to combine those reliable depth cues to generate the final depth information. Simulation results show that each depth cue is reasonably utilized according to scene types and final depth is generated by cues which can effectively represent the current scene.