• Title/Summary/Keyword: 청각특성

Search Result 331, Processing Time 0.205 seconds

Speech Signal Processing using Pitch Synchronous Multi-Spectra and DSP System Design in Cochlear Implant (피치동기 다중 스펙트럼을 이용한 청각보철장치의 음성신호처리 및 DSP 시스템 설계)

  • Shin, J. I.;Park, S. J.;Shin, D. K.;Lee, J. H.;Park, S. H.
    • Journal of Biomedical Engineering Research
    • /
    • v.20 no.4
    • /
    • pp.495-502
    • /
    • 1999
  • We propose efficient speech signal processing algorithms and a system for cochlear implant in this paper. The outer and the middle car which perform amplifying, lowpass filtering and AGC, are modeled by an analog system, and the inner ear acting as a time-delayed multi filter and the transducer is implemented by the DSP circuit which enables real-time processing. Especially, the basilar membrane characteristic of the inner ear is modeled by a nonlinear filter bank, and then tonotopy and periodicity of the auditory system is satisfied by using a pitch-synchronous multi-spectra(PSMS) method. Moreover, most of the speech processing is performed by S/W so the system can be easily modified. And as our program is written in C-language, it can be easily transplanted to the system using other processors.

  • PDF

Automatic Vowel Onset Point Detection Based on Auditory Frequency Response (청각 주파수 응답에 기반한 자동 모음 개시 지점 탐지)

  • Zang, Xian;Kim, Hag-Tae;Chong, Kil-To
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.13 no.1
    • /
    • pp.333-342
    • /
    • 2012
  • This paper presents a vowel onset point (VOP) detection method based on the human auditory system. This method maps the "perceptual" frequency scale, i.e. Mel scale onto a linear acoustic frequency, and then establishes a series of Triangular Mel-weighted Filter Bank simulate the function of band pass filtering in human ear. This nonlinear critical-band filter bank helps greatly reduce the data dimensionality, and eliminate the effect of harmonic waves to make the formants more prominent in the nonlinear spaced Mel spectrum. The sum of mel spectrum peaks energy is extracted as feature for each frame, and the instinct at which the energy amplitude starts rising sharply is detected as VOP, by convolving with Gabor window. For the single-word database which contains 12 vowels articulated with different kinds of consonants, the experimental results showed a good average detection rate of 72.73%, higher than other vowel detection methods based on short-time energy and zero-crossing rate.

Comparing the Effects of Visual and Visual-auditory Feedback on Eco-driving and Driving Workload (시각적 피드백과 시각-청각적 피드백이 에코 드라이빙과 운전부하에 미치는 상대적 효과)

  • Lee, Kye hoon;Lim, Sung jun;Oah, She zeen
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.16 no.3
    • /
    • pp.120-131
    • /
    • 2017
  • Recent studies have suggested that providing in-vehicle feedback on various driving behaviors promote eco-friendly driving behaviors. However, there was relatively little interest in cognitive overload that can be caused by the provision of information. Thus, the goal of this study was to investigate the relative effects of two types of feedback(visual feedback vs. visual-auditory feedback) to increase eco-driving performance while minimizing driving workload. Also, in this study, the complexity of the driving task was distinguished (secondary vs. tertiary task) in order to reflect the actual driving situation. The study adopted a counterbalancing design in which the two feedback types were delivered in a different order under the two different task conditions. Results showed that providing the visual-auditory feedback was more effective than the visual only feedback in both promoting eco-friendly driving behaviors and minimizing driving workload under both task conditions.

Chaotic Dynamics in EEG Signal Responding to Auditory Stimulus with Various Sound-Cutting Frequencies. (단속 주파수를 변화시킨 청각자극에 반응하는 뇌전위신호의 카오스 분석)

  • Choe, Jeong-Mi;Bae, Byeong-Hun;Kim, Su-Yong
    • Journal of Biomedical Engineering Research
    • /
    • v.15 no.3
    • /
    • pp.237-244
    • /
    • 1994
  • We investigated the qualitive and quantitative properties in EEG signal which responds to auditory stimulus with increaing the sound-cutting frequency from 2 Hz to 20 Hz by 2 Hz step units, by chaotic dynamics. To bigin with, general chaotic properties such as fractal mechanism, 1 If frequency spectrum and positive Lyapunov exponent are discussed in EEG signal. For evoked potential with given auditory stimulus, the route to chaos by bifurcation diagram and the changes in geometrical property of Poincare sections of 2-dimensional psedophase space is observed. For that containing spontaneous potential, seen as the random background signal, the chaotic attractors in 3-dimensional phase space are found containing the same infomation as the above mentioned evoked potential. Finally the chinges of Lyapunov exponent by various sound-cutting frequencies of stimulus and by the various spatial positions (occipital region) in a brain surface to be measured, are illustrated meaningfully.

  • PDF

Analysis of Objective Sound Quality Features for Vacuum Cleaner Noise (청소기 소음 측정을 위한 객관적 음질 특성 분석)

  • Lee, Sang-Wook;Cho, Youn;Park, Jong-Geun;Hwang, Dae-Sun;Song, Chi-Mun;Lee, Chul-Hee
    • The Journal of the Acoustical Society of Korea
    • /
    • v.29 no.4
    • /
    • pp.258-264
    • /
    • 2010
  • In this paper, we propose an objective quality feature which is based on the human auditory system to measure vacuum cleaner noise. It is observed that some frequency bands are more sensitive to the human auditory system. Therefore, we divided the audible frequency range of vacuum cleaner noise into a number of frequency bands and the average energy of these bands was calculated. Among a number of average energies, an average energy of a frequency band was selected as the proposed feature. In order to test the performance of the proposed feature, fourteen vacuum cleaners were chosen and the noise was recorded in an anechoic-chamber. Then we performed subjective tests to obtain subjective scores of the noise data using the PCM (paired comparison method) and ACR (absolute category rating) subjective methods. The proposed objective quality feature shows high correlation with the subjective scores.

Spike Feature Extraction for Emotion Recognition based on Deep Neural Network (심층 신경망 기반 감정 인식을 위한 스파이크 특성 추출 기술)

  • An, Soonho;Kim, Jaewon;Han, Seokhyeon;Shin, Seonghyeon;Park, Hochong
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2019.06a
    • /
    • pp.158-159
    • /
    • 2019
  • 본 논문에서는 심층 신경망을 기반으로 하는 감정 인식을 위해 스파이크 특성을 추출하는 기술을 제안한다. 기존의 심층 신경망을 이용한 감정 인식 기술은 대부분 MFCC를 특성 백터를 사용한다. 그러나 프레임 단위의 연산인 MFCC는 높은 시간 해상도를 확보하기 어려워 시간적 특성의 영향을 받는 감정 인식에 한계가 있다. 이를 해결하기 위해 본 논문에서는 인간의 청각 필터를 모델링한 ERB에 따라 샘플 단위로 주파수의 특성을 나타내는 스파이크그램을 이용한 감정 인식 기술을 제안한다. 제안하는 방법이 감정 인식의 대표적 특성인 MFCC보다 높은 인식률을 제공하는 것을 확인하였다.

  • PDF

A Study of Acoustic Masking Effect from Formant Enhancement in Digital Hearing Aid (디지털 보청기에서의 포먼트 강조에 의한 마스킹 효과 연구)

  • Jeon, Yu-Yong;Kil, Se-Kee;Yoon, Kwang-Sub;Lee, Sang-Min
    • Journal of the Institute of Electronics Engineers of Korea SC
    • /
    • v.45 no.5
    • /
    • pp.13-20
    • /
    • 2008
  • Although digital hearing aid algorithms have been developed to compensate hearing loss and to help hearing impaired people to communicate with others, digital hearing aid user still complain about difficulty of hearing the speech. The reason could be the quality of speech through digital hearing aid is insufficient to understand the speech caused by feedback, residual noise and etc. And another thing is masking effect among formants that makes sound quality low. In this study, we measured the masking characteristics of normal listeners and hearing impaired listeners having presbyacusis to confirm masking effect in speech itself. The experiment is composed of 5 tests; pure tone test, speech reception threshold (SRT) test, word recognition score (WRS) test, puretone masking test and speech masking test. In speech masking test, there are 25 speeches in each speech set. And log likelihood ratio (LLR) is introduced to evaluate the distortion of each speech objectively. As a result, the speech perception became lower by increasing the quantity of formant enhancement. And each enhanced speech in a speech set has statistically similar LLR, however speech perception is not. It means that acoustic masking effect rather than distortion influences speech perception. In actuality, according to the result of frequency analysis of the speech that people can not answer correctly, level difference between first formant and second formant is about 35dB, and it is similar to result of pure tone masking test(normal hearing subject:36.36dB, hearing impaired subject:32.86dB). Characteristics of masking effect is not similar between normal listeners and hearing impaired listeners. So it is required to check the characteristics of masking effect before wearing a hearing aid and to apply this characteristics to fitting.

Improvement of 3D Sound Using Psychoacoustic Characteristics (인간의 청각 특성을 이용한 입체음향의 방향감 개선)

  • Koo, Kyo-Sik;Cha, Hyung-Tai
    • The Journal of the Acoustical Society of Korea
    • /
    • v.30 no.5
    • /
    • pp.255-264
    • /
    • 2011
  • The Head Related Transfer Function (HRTF) means a process related to acoustic transmission from 3d space to the listener's ear. In other words, it contains the information that human can perceive locations of sound sources. So, we make virtual 3d sound using HRTF, despite it doesn't actually exist. But, it can deteriorate some three-dimensional effect by the confusion between front and back directions due to the non-individual HRTF depending on each listener. In this paper, we proposed the new algorithm to reduce the confusion of sound image localization using human's acoustic characteristics. The frequency spectrum and global masking threshold of 3d sounds using HRTF are used to calculate the psychoacoustical differences among each directions. And perceptible cues in each critical band are boosted to create effective 3d sound. As a result, we can make the improved 3d sound, and the performances are much better than conventional methods.

The Analysis of EEG Signal Responding to the Pure Tone Auditory Stimulus (청각자극의 반송 주파수에 따른 뇌전위 신호의 해석)

  • Choe, Jeong-Mi;Bae, Byeong-Hun;Kim, Su-Yong
    • Journal of Biomedical Engineering Research
    • /
    • v.15 no.4
    • /
    • pp.383-388
    • /
    • 1994
  • Chaotic analysis of EEG signal responding to auditory stimulus with various carrier frequency and constant triggering frequency is given in this paper. The EEG signal is obtained from the digital 12channel EEG system made in our laboratory. The carrier frequency is varied from 1 kHz to 3 kHz by 0.5 kHz step. Chaos analysis such as pseudo phase space portrait, Lyapunov exponent, and so on is done on the auditory stimulated evoked potential. This result is found to be quite consistent with the well known results from the psychological perception theory.

  • PDF

Matching Pursuit based on Frequency Characteristics of Heart Sounds for Phonocardiogram Analysis (심음 분석을 위한 주파수 특성에 기반한 매칭퍼슈잇 방법)

  • Kim, Il-Dong;Jeong, Gyu-Hyeok;Lim, Jong-Ha;Lee, In-Sung
    • Journal of the Institute of Electronics Engineers of Korea SC
    • /
    • v.47 no.1
    • /
    • pp.47-55
    • /
    • 2010
  • Although heart auscultation is an important tool for the diagnosis of heart disease, it is seen that the human ear is poorly suited for cardiac auscultation such as the paradoxical split of S2, the split of S2, and the summation of S2 and S3 gallop. In this paper, the analysis-synthesis method that is suitable for the characteristics of PCG is proposed to classify the heart sound. In analyzing the PCG according to matching pursuit method and synthesizing the PCG in each bandwidth based on the frequency characteristics, it sorts normal hearts with abnormal hearts that is difficult to classify as cardiac auscultation. The result shows that it qualifies the appearance of split in S2 and the paradoxical split of S2 and also can estimate the delay between the two components.