• Title/Summary/Keyword: digital speech signal

Search Result 136, Processing Time 0.028 seconds

A Feedback and Noise Cancellation Algorithm of Hearing Aids Using Adaptive Beamforming Method (적응 빔형성기법을 이용한 보청기의 궤환 및 잡음제거 알고리즘)

  • Lee, Haeng-Woo
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.35 no.1C
    • /
    • pp.96-102
    • /
    • 2010
  • This paper proposes a new adaptive algorithm to cancel the acoustic feedback and noise signals in the digital hearing aids. The proposed algorithm improves its convergence performances by canceling the speech signal from the residual signal using two microphones. The feedback canceller firstly cancels the feedback signal among the mic signal, and then it is reduced the noise using the beamforming method. To verify the performances of the proposed algorithm, the simulations were carried out for some cases. As the results of simulations, it was proved that the feedback canceller and the noise canceller advance about 14.43 dB for SFR, 10.19 dB for SNR respectively during speech, in the case of using the new algorithm.

A Study on Performance Improvement of Modified Window Function (변형된 창함수의 성능향상에 관한 연구)

  • Lee, Kyung-Hyo;Kim, Nam-Ho
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2008.05a
    • /
    • pp.925-928
    • /
    • 2008
  • With basis of the development of information communication techniques in recent year, the digital processing techniquy also has been growed fast. The digital processing technique have used signals - speech and image processing- for processing of transmission and analysis. After we get and save the signals. Effective signal processing techniques have varied filters and typical digital filters are FIR filter and IIR filter. The FIR digital filter is more secure because phase response characteristics have linear phase. But, FIR digital filters have a problem to product the Gibbs phenomenon generating around a discontinuous point. A propose of filer is to remove the problem. Therefore, in this paper I was proposed a method using FIR digital filter applied a modified window function and the method was compared with conventional methods.

  • PDF

Design of pitch parameter search architecture for a speech coder using dual MACs (Dual MAC을 이용한 음성 부호화기용 피치 매개변수 검색 구조 설계)

  • 박주현;심재술;김영민
    • Journal of the Korean Institute of Telematics and Electronics A
    • /
    • v.33A no.5
    • /
    • pp.172-179
    • /
    • 1996
  • In the paper, QCELP (qualcomm code excited linear predictive), CDMA (code division multiple access)'s vocoder algorithm, was analyzed. And then, a ptich parameter seaarch architecture for 16-bit programmable DSP(digital signal processor) for QCELP was designed. Because we speed up the parameter search through high speed DSP using two MACs, we can satisfy speech codec specifiction for the digital celluar. Also, we implemented in FIFO(first-in first-out) memory using register file to increase the access time of data. This DSP was designed using COMPASS, ASIC design tool, by top-down design methodology. Therefore, it is possible to cope with rapid change at mobile communication market.

  • PDF

Implementation of a Real-time SIFT Pitch Detector (실시간 SIFT 기본주파수 검출기의 구현)

  • Lee, Jong Seok;Lee, Sang Uk
    • Journal of the Korean Institute of Telematics and Electronics
    • /
    • v.23 no.1
    • /
    • pp.101-113
    • /
    • 1986
  • In this paper, a real-time pitch detector LPC vocoder as implemented on a high speed digital signal processor, NEC 7720, is described. The pitch detector was based mainly on the SIFT algorithm. The SIFT pitch detector consists primarily of a digital low pass filter, inverse filter, computation of autocorrelation, a peak picker, interpolation, V/UV defcision and a final pitch smoother. In our approach, modification, mainly on the V/UV decision and a final pitch smoother, was made to estimate more accurate pitches. An 16-bit fixed-point aithmatic was employed for all necessary computation and the simulated results were compared with the eye detected pitches obtained from real speech data. The pitch detector occupies 98.8% of the instruction ROM, 37% of the data ROM, and 94% of internal RAM and takes 15.2ms to estimate a pitch when an analysis frame is consisted of 128 sampled speech data. It is observed that the tested results were well agreed with the computer simulation results.

  • PDF

An Interdisciplinary Study of A Leaders' Voice Characteristics: Acoustical Analysis and Members' Cognition

  • Hahm, SangWoo;Park, Hyungwoo
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.12
    • /
    • pp.4849-4865
    • /
    • 2020
  • The traditional roles of leaders are to influence members and motivate them to achieve shared goals in organizations. However, leaders such as top managers and chief executive officers, in practice, do not always directly meet or influence other company members. In fact, they tend to have the greatest impact on their members through formal speeches, company procedures, and the like. As such, official speech is directly related to the motivation of company employees. In an official speech, not only the contents of the speech, but also the voice characteristics of the speaker have an important influence on listeners, as the different vocal characteristics of a person can have different effects on the listener. Therefore, according to the voice characteristics of a leader, the cognition of the members may change, and, the degree to which the members are influenced and motivated will be different. This study identifies how members may perceive a speech differently according to the different voice characteristics of leaders in formal speeches. Further, different perceptions about voices will influence members' cognition of the leader, for example, in how trustworthy they appear. The study analyzed recorded speeches of leaders, and extracted features of their speaking style through digital speech signal analysis. Then, parameters were extracted and analyzed by the time domain, frequency domain, and spectrogram domain methods. We also analyzed the parameters for use in Natural Language Processing. We investigated which leader's voice characteristics had more influence on members or were more effective on them. A person's voice characteristics can be changed. Therefore, leaders who seek to influence members in formal speeches should have effective voice characteristics to motivate followers.

Implementation of Speaker Independent Speech Recognition System Using Independent Component Analysis based on DSP (독립성분분석을 이용한 DSP 기반의 화자 독립 음성 인식 시스템의 구현)

  • 김창근;박진영;박정원;이광석;허강인
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.8 no.2
    • /
    • pp.359-364
    • /
    • 2004
  • In this paper, we implemented real-time speaker undependent speech recognizer that is robust in noise environment using DSP(Digital Signal Processor). Implemented system is composed of TMS320C32 that is floating-point DSP of Texas Instrument Inc. and CODEC for real-time speech input. Speech feature parameter of the speech recognizer used robust feature parameter in noise environment that is transformed feature space of MFCC(met frequency cepstral coefficient) using ICA(Independent Component Analysis) on behalf of MFCC. In recognition result in noise environment, we hew that recognition performance of ICA feature parameter is superior than that of MFCC.

A Study on LMS-MPC Method Considering Low Bit Rate (Low Bit Rate을 고려한 LMS-MPC 방식에 관한 연구)

  • Lee, See-Woo
    • Journal of Digital Convergence
    • /
    • v.10 no.5
    • /
    • pp.233-238
    • /
    • 2012
  • In a speech coding system using excitation source of voiced and unvoiced, it would be a distortion of speech waveform in case of exist a voiced and an unvoiced consonants in a frame. To solve this problem, this paper present a method of LMS-MPC uses individual pitch and LMS(Least Mean Square). I evaluate the MPC and LMS-MPC using LMS. As a result, SNRseg of LMS-MPC was improved 1.5dB for female voice and 1.3dB for male voice respectively. Compared to the MPC, SNRseg of LMS-MPC has been improved that I was able to control the distortion of the speech waveform finally. And so, I expect to be able to this method for cellular phone and smart phone using excitation source of low bit rate.

Semantic Ontology Speech Recognition Performance Improvement using ERB Filter (ERB 필터를 이용한 시맨틱 온톨로지 음성 인식 성능 향상)

  • Lee, Jong-Sub
    • Journal of Digital Convergence
    • /
    • v.12 no.10
    • /
    • pp.265-270
    • /
    • 2014
  • Existing speech recognition algorithm have a problem with not distinguish the order of vocabulary, and the voice detection is not the accurate of noise in accordance with recognized environmental changes, and retrieval system, mismatches to user's request are problems because of the various meanings of keywords. In this article, we proposed to event based semantic ontology inference model, and proposed system have a model to extract the speech recognition feature extract using ERB filter. The proposed model was used to evaluate the performance of the train station, train noise. Noise environment of the SNR-10dB, -5dB in the signal was performed to remove the noise. Distortion measure results confirmed the improved performance of 2.17dB, 1.31dB.

A PERFORMANCE STUDY OF SPEECH CODERS FOR TELEPHONE CONFERENCING IN DIGITAL MOBILE COMMUNICATION NETWORKS

  • Lee, M.S.;Lee, G.C.;Kim, K.C.;Lee, H.S.;Lyu, D.S.;Shin, D.J.;Lee, Hun
    • Proceedings of the Acoustical Society of Korea Conference
    • /
    • 1994.06a
    • /
    • pp.899-903
    • /
    • 1994
  • This paper describes two methods to assess the output speech, quality of vocoders for telephone conferencing in digital mobile communication networks. The proposed methods are the sentence discrimiantion method and the modified degraded mean opinion score (MDMOS) test. We apply these two methods to Qualcomm code excited linear prediction (QCELP), vector sum excited linear prediction (VSELP) and regular pulse excited-long term predictin (RPE-LTD) voceders to evaluate which vocoding algorithm can process mixed voice signal from two speakers better for telephone conferencing. From the experiments we obtain that the VSELP vocoding algorithm reveals superior output speech quality to the other two.

  • PDF

Investigating the Effects of Hearing Loss and Hearing Aid Digital Delay on Sound-Induced Flash Illusion

  • Moradi, Vahid;Kheirkhah, Kiana;Farahani, Saeid;Kavianpour, Iman
    • Korean Journal of Audiology
    • /
    • v.24 no.4
    • /
    • pp.174-179
    • /
    • 2020
  • Background and Objectives: The integration of auditory-visual speech information improves speech perception; however, if the auditory system input is disrupted due to hearing loss, auditory and visual inputs cannot be fully integrated. Additionally, temporal coincidence of auditory and visual input is a significantly important factor in integrating the input of these two senses. Time delayed acoustic pathway caused by the signal passing through digital signal processing. Therefore, this study aimed to investigate the effects of hearing loss and hearing aid digital delay circuit on sound-induced flash illusion. Subjects and Methods: A total of 13 adults with normal hearing, 13 with mild to moderate hearing loss, and 13 with moderate to severe hearing loss were enrolled in this study. Subsequently, the sound-induced flash illusion test was conducted, and the results were analyzed. Results: The results showed that hearing aid digital delay and hearing loss had no detrimental effect on sound-induced flash illusion. Conclusions: Transmission velocity and neural transduction rate of the auditory inputs decreased in patients with hearing loss. Hence, the integrating auditory and visual sensory cannot be combined completely. Although the transmission rate of the auditory sense input was approximately normal when the hearing aid was prescribed. Thus, it can be concluded that the processing delay in the hearing aid circuit is insufficient to disrupt the integration of auditory and visual information.