• Title/Summary/Keyword: Visual Signal

Search Result 548, Processing Time 0.026 seconds

A Case of Unusual Pituitary Apoplexy Presented as Aseptic Meningitis (무균성 수막염으로 발현한 흔하지 않은 뇌하수체졸중)

  • Park, Kang Min;Kim, Yeon Mee;Kim, Si Eun;Shin, Kyong Jin;Ha, Sam Yeol;Park, Jinse;Kim, Sung Eun
    • Annals of Clinical Neurophysiology
    • /
    • v.15 no.1
    • /
    • pp.24-26
    • /
    • 2013
  • We encountered a case of pituitary apoplexy who presented with isolated headache and vomiting without visual disturbance or ophthalmoplegia. The cerebrospinal fluid examination was compatible with aseptic meningitis. A computed tomography revealed slightly high density in the pituitary fossa and suprasella area, but the signal change was very faint. Our case suggests that clinicians should take into account the possibility of pituitary apoplexy without visual disturbance or ophthalmoplegia, when aseptic meningitis is suspected.

A Robust Wavelet-Based Digital Watermarking Using Statistical Characteristic of Image and Human Visual System

  • Kim, Bong-Seok;Kwon, Kee-Koo;Kwon, Seong-Geun;Park, Kyung-Nam
    • Proceedings of the IEEK Conference
    • /
    • 2002.07b
    • /
    • pp.1019-1022
    • /
    • 2002
  • The current paper proposes a wavelet-based digital watermarking algorithm using statistical characteristic of image and human visual system (HVS). The original image is decomposed into 4-level using a discrete wavelet transform (DWT), then the watermark is embedded into the perceptually significant coefficients (PSCs) of the image. In general, the baseband of a wavelet-decomposed image includes most of the energy of the original image, thereby having a crucial effect on the image quality. As such, to retain invisibility, the proposed algorithm does not utilize the baseband. Plus, the wavelet coefficients on the lowest level are also excluded in the watermark-embedding step, because these coefficients call be easily eliminated and modified by lossy compression and common signal processing. As such, the PSCs are selected from all subbands, except for the baseband and subbands on the lowest level. Finally, using the selected PSCs, the watermark is then embedded based on spatial masking of the wavelet coefficients so as to provide invisibility and robustness. Computer simulation results confirmed that the proposed watermarking algorithm was more invisible and robust than conventional algorithms.

  • PDF

Temporal Anti-aliasing of a Stereoscopic 3D Video

  • Kim, Wook-Joong;Kim, Seong-Dae;Hur, Nam-Ho;Kim, Jin-Woong
    • ETRI Journal
    • /
    • v.31 no.1
    • /
    • pp.1-9
    • /
    • 2009
  • Frequency domain analysis is a fundamental procedure for understanding the characteristics of visual data. Several studies have been conducted with 2D videos, but analysis of stereoscopic 3D videos is rarely carried out. In this paper, we derive the Fourier transform of a simplified 3D video signal and analyze how a 3D video is influenced by disparity and motion in terms of temporal aliasing. It is already known that object motion affects temporal frequency characteristics of a time-varying image sequence. In our analysis, we show that a 3D video is influenced not only by motion but also by disparity. Based on this conclusion, we present a temporal anti-aliasing filter for a 3D video. Since the human process of depth perception mainly determines the quality of a reproduced 3D image, 2D image processing techniques are not directly applicable to 3D images. The analysis presented in this paper will be useful for reducing undesirable visual artifacts in 3D video as well as for assisting the development of relevant technologies.

  • PDF

Optimizing Bit Rate Control for Realtime TV Broadcasting Transmission using LTE Network (LTE 무선통신을 활용한 TV 생방송 중계화면 안정화 비트레이트 조정 연구)

  • Kwon, Mahnwoo;Lim, Hyunchan
    • Journal of Korea Multimedia Society
    • /
    • v.21 no.3
    • /
    • pp.415-422
    • /
    • 2018
  • Advances of telecommunication technology bring various changes in journalism field. Reporters started to gather, edit, and transmit content to main server in media company using hand-held smart media and notebook computer. This paper tried to testify valid bit-rate of visual news content using LTE network and mobile phone. Field news like natural disasters need real-time transmission of video content. But broadcasting company normally use heavy ENG system and transmission satellite trucks. We prepared and experimented different types of visual content that has different bit-rates. Transmission tool was LU-60HD mobile system of LiveU Corporation. Transmission result shows that bit-rate of 2Mbps news content is not suitable for broadcasting and VBR (Variable Bit Rate) transmission has better definition quality than CBR (Constant Bit Rate) method. Three different bit-rate of VBR transmission result shows that 5Mbps clip has better quality than 1Mbps and 3Mbps. The higher bit-rate, the better video quality. But if the content has much movements, that cause delay and abnormal quality of video. So optimizing the balance between stability of signal and quality of bit-rate is crucial factor of real-time broadcasting news gathering business.

Image Cryptographic Algorithm Based on the Property of Wavelet Packet Transform (웨이브렛 패킷 변환의 특성을 이용한 영상 암호화 알고리즘)

  • Shin, Jonghong
    • Journal of Korea Society of Digital Industry and Information Management
    • /
    • v.14 no.2
    • /
    • pp.49-59
    • /
    • 2018
  • Encryption of digital images has been requested various fields. In the meantime, many algorithms based on a text - based encryption algorithm have been proposed. In this paper, we propose a method of encryption in wavelet transform domain to utilize the characteristics of digital image. In particular, wavelet transform is used to reduce the association between the encrypted image and the original image. Wavelet packet transformations can be decomposed into more subband images than wavelet transform, and various position permutation, numerical transformation, and visual transformation are performed on the coefficients of this subband image. As a result, this paper proposes a method that satisfies the characteristics of high encryption strength than the conventional wavelet transform and reversibility. This method also satisfies the lossless symmetric key encryption and decryption algorithm. The performance of the proposed method is confirmed by visual and quantitative. Experimental results show that the visually encrypted image is seen as a completely different signal from the original image. We also confirmed that the proposed method shows lower values of cross correlation than conventional wavelet transform. And PSNR has a sufficiently high value in terms of decoding performance of the proposed method. In this paper, we also proposed that the degree of correlation of the encrypted image can be controlled by adjusting the number of wavelet transform steps according to the characteristics of the image.

Comparison of Fusion Methods for Generating 250m MODIS Image

  • Kim, Sun-Hwa;Kang, Sung-Jin;Lee, Kyu-Sung
    • Korean Journal of Remote Sensing
    • /
    • v.26 no.3
    • /
    • pp.305-316
    • /
    • 2010
  • The MODerate Resolution Imaging Spectroradiometer (MODIS) sensor has 36 bands at 250m, 500m, 1km spatial resolution. However, 500m or 1km MODIS data exhibits a few limitations when low resolution data is applied at small areas that possess complex land cover types. In this study, we produce seven 250m spectral bands by fusing two MODIS 250m bands into five 500m bands. In order to recommend the best fusion method by which one acquires MODIS data, we compare seven fusion methods including the Brovey transform, principle components algorithm (PCA) fusion method, the Gram-Schmidt fusion method, the least mean and variance matching method, the least square fusion method, the discrete wavelet fusion method, and the wavelet-PCA fusion method. Results of the above fusion methods are compared using various evaluation indicators such as correlation, relative difference of mean, relative variation, deviation index, peak signal-to-noise ratio index and universal image quality index, as well as visual interpretation method. Among various fusion methods, the local mean and variance matching method provides the best fusion result for the visual interpretation and the evaluation indicators. The fusion algorithm of 250m MODIS data may be used to effectively improve the accuracy of various MODIS land products.

JPEG Pleno: Providing representation interoperability for holographic applications and devices

  • Schelkens, Peter;Ebrahimi, Touradj;Gilles, Antonin;Gioia, Patrick;Oh, Kwan-Jung;Pereira, Fernando;Perra, Cristian;Pinheiro, Antonio M.G.
    • ETRI Journal
    • /
    • v.41 no.1
    • /
    • pp.93-108
    • /
    • 2019
  • Guaranteeing interoperability between devices and applications is the core role of standards organizations. Since its first JPEG standard in 1992, the Joint Photographic Experts Group (JPEG) has published several image coding standards that have been successful in a plethora of imaging markets. Recently, these markets have become subject to potentially disruptive innovations owing to the rise of new imaging modalities such as light fields, point clouds, and holography. These so-called plenoptic modalities hold the promise of facilitating a more efficient and complete representation of 3D scenes when compared to classic 2D modalities. However, due to the heterogeneity of plenoptic products that will hit the market, serious interoperability concerns have arisen. In this paper, we particularly focus on the holographic modality and outline how the JPEG committee has addressed these tremendous challenges. We discuss the main use cases and provide a preliminary list of requirements. In addition, based on the discussion of real-valued and complex data representations, we elaborate on potential coding technologies that range from approaches utilizing classical 2D coding technologies to holographic content-aware coding solutions. Finally, we address the problem of visual quality assessment of holographic data covering both visual quality metrics and subjective assessment methodologies.

Multimodal audiovisual speech recognition architecture using a three-feature multi-fusion method for noise-robust systems

  • Sanghun Jeon;Jieun Lee;Dohyeon Yeo;Yong-Ju Lee;SeungJun Kim
    • ETRI Journal
    • /
    • v.46 no.1
    • /
    • pp.22-34
    • /
    • 2024
  • Exposure to varied noisy environments impairs the recognition performance of artificial intelligence-based speech recognition technologies. Degraded-performance services can be utilized as limited systems that assure good performance in certain environments, but impair the general quality of speech recognition services. This study introduces an audiovisual speech recognition (AVSR) model robust to various noise settings, mimicking human dialogue recognition elements. The model converts word embeddings and log-Mel spectrograms into feature vectors for audio recognition. A dense spatial-temporal convolutional neural network model extracts features from log-Mel spectrograms, transformed for visual-based recognition. This approach exhibits improved aural and visual recognition capabilities. We assess the signal-to-noise ratio in nine synthesized noise environments, with the proposed model exhibiting lower average error rates. The error rate for the AVSR model using a three-feature multi-fusion method is 1.711%, compared to the general 3.939% rate. This model is applicable in noise-affected environments owing to its enhanced stability and recognition rate.

Development of Processing Program for Audio-vision System Based on Auditory Input (청각을 이용한 시각 재현장치의 분석프로그램 개발)

  • Heo, Se-Jin;Bang, Sung-Sik;Seo, Jee-Hye;Choi, Hyun-Woo;Kim, Tae-Ho;Lee, Na-Hee;Lee, Yu-Jin;Park, Ji-Won;Lee, Hui-Joong;Won, Chul-Ho;Lee, Jong-Min
    • Journal of Korea Multimedia Society
    • /
    • v.13 no.1
    • /
    • pp.58-65
    • /
    • 2010
  • The final goal of our research is developing not a simple collision a1ann equipment for the blinded walkers, but the apparatus (Audio- Vision System) which can simulate vision based on auditory information so that the blinds can figure the three dimensional space in front of them. On the way to the final goal, in this study, simulation software was developed and verified. Thirty normal volunteers were included in the subject group and the average age Was 25.8 years old. After being accustomed to the system by evaluating 10 blinded virtual spaces, the volunteers performed test using another set of 10 blinded virtual spaces. The results of test were scored by shape, center, margin, and gradient surface of objects in virtual space. The score of each checking point ranged from 1 to 5, and the full score was converted to 100. As results of this study, the total score ranged from 77 to 97 with the average of 88.7. In this study, a simulation software was developed and verified to have acceptable success rale. By combining to visual sensors, the vision-reconstruction system based on auditory signal (Audio-vision System) may be developed.

DECODE: A Novel Method of DEep CNN-based Object DEtection using Chirps Emission and Echo Signals in Indoor Environment (실내 환경에서 Chirp Emission과 Echo Signal을 이용한 심층신경망 기반 객체 감지 기법)

  • Nam, Hyunsoo;Jeong, Jongpil
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.21 no.3
    • /
    • pp.59-66
    • /
    • 2021
  • Humans mainly recognize surrounding objects using visual and auditory information among the five senses (sight, hearing, smell, touch, taste). Major research related to the latest object recognition mainly focuses on analysis using image sensor information. In this paper, after emitting various chirp audio signals into the observation space, collecting echoes through a 2-channel receiving sensor, converting them into spectral images, an object recognition experiment in 3D space was conducted using an image learning algorithm based on deep learning. Through this experiment, the experiment was conducted in a situation where there is noise and echo generated in a general indoor environment, not in the ideal condition of an anechoic room, and the object recognition through echo was able to estimate the position of the object with 83% accuracy. In addition, it was possible to obtain visual information through sound through learning of 3D sound by mapping the inference result to the observation space and the 3D sound spatial signal and outputting it as sound. This means that the use of various echo information along with image information is required for object recognition research, and it is thought that this technology can be used for augmented reality through 3D sound.