• Title/Summary/Keyword: sound based information

Search Result 637, Processing Time 0.031 seconds

Implementation of an Intelligent Audio Graphic Equalizer System (지능형 오디오 그래픽 이퀄라이저 시스템 구현)

  • Lee Kang-Kyu;Cho Youn-Ho;Park Kyu-Sik
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.43 no.3 s.309
    • /
    • pp.76-83
    • /
    • 2006
  • A main objective of audio equalizer is for user to tailor acoustic frequency response to increase sound comfort and example applications of audio equalizer includes large-scale audio system to portable audio such as mobile MP3 player. Up to now, all the audio equalizer requires manual setting to equalize frequency bands to create suitable sound quality for each genre of music. In this paper, we propose an intelligent audio graphic equalizer system that automatically classifies the music genre using music content analysis and then the music sound is boosted with the given frequency gains according to the classified musical genre when playback. In order to reproduce comfort sound, the musical genre is determined based on two-step hierarchical algorithm - coarse-level and fine-level classification. It can prevent annoying sound reproduction due to the sudden change of the equalizer gains at the beginning of the music playback. Each stage of the music classification experiments shows at least 80% of success with complete genre classification and equalizer operation within 2 sec. Simple S/W graphical user interface of 3-band automatic equalizer is implemented using visual C on personal computer.

A method of the cross-talk cancellation for an sound reproduction of 5.1 channel speaker system (5.1 채널 스피커 시스템 음향재생을 위한 크로스토크 제거방법)

  • Lee, Soo-Jeong;Cho, Gab-Ken;Kim, Soon-Hyob
    • Journal of the Institute of Electronics Engineers of Korea SP
    • /
    • v.42 no.4 s.304
    • /
    • pp.159-166
    • /
    • 2005
  • This thesis deals with a method to deliver more realistic sound by cancelling the cross-talk which is inherent to the 5.1 channel speaker system. First, the cross-talk cancellation method that eliminates cross-talks on the path from left speaker to right ear and from right speaker to left ear is explained. Then the application and replaying method using the cross-talk cancellation explained here is introduced. The acoustical model for cross-talk cancellation is the free field model This model minimizes distortion of sound. Many experts also make studies on this model. I used the bark scale sound quality compensation based on psycho-acoustic. For the surround channels, band-limited sound quality compensation is performed in the frequency domain.

A Study on the Sound Quality Improvement Using the Equal Compensation Filter in Bark-scale for the Cross-talk Cancellation (크로스토크 제거를 위한 바크스케일 등가 보상 필터를 이용한 음질 향상에 관한 연구)

  • Kim, Hack-Jin;Kim, Soon-Hyub
    • The KIPS Transactions:PartB
    • /
    • v.11B no.3
    • /
    • pp.345-352
    • /
    • 2004
  • This paper deals a method to deliver more realistic sound by cancelling the cross-talk which is inherent to the 5.1 channel speaker system. The acoustical model for cross-talk cancellation is the free field model. This model minimizes distortion of sound. 1 used the bark scale sound quality compensation which based on psycho-acoustic. For the surround channels, band-limited sound quality compensation is performed in the frequency domain. I also performed the sound qualify assessment test on the traditional 2 channel stereo and 5.1 channel system. This test is performed in the tort chamber which satisfies the ITU-R specifications. 1 uses the IACC(Inter-Aural Cross-Correlation) to determine the preferences of the amateur and the golden ear experts to asses the trans-aural filter. According to the result from the proposed method, I got more the 38dB separation rates with the Dolby standard speaker array. The results on the diffusion by the subjective test with the experts shows 0.4∼0.5 point Increased then before.

Yoga of Consilience through Immersive Sound Experience (실감음향 체험을 통한 통섭의 요가)

  • Hyon, Jinoh
    • Journal of Broadcast Engineering
    • /
    • v.26 no.5
    • /
    • pp.643-651
    • /
    • 2021
  • Most people acquire information visually. Screens of computers, smart phones, etc. constantly stimulate people's eyes, increasing fatigue. In this social phenomenon, the realistic and rich sound of the 21st century's state-of-art sound system can affect people's bodies and minds in various ways. Through sound, human beings are given space to calm and observe themselves. The purpose of this paper is to introduce immersive yoga training based on 3D sound conducted together by ALgruppe & Rory's PranaLab and to promote the understanding of immersive audio system. As a result, people, experienced immersive yoga, not only enjoy the effect of sound, but also receive a powerful energy that gives them a sense of inner self-awareness. This is a response to multidisciplinary exchange required by the knowledge of modern society, and at the same time, informs the possibility of new cultural contents.

Indoor Scene Classification based on Color and Depth Images for Automated Reverberation Sound Editing (자동 잔향 편집을 위한 컬러 및 깊이 정보 기반 실내 장면 분류)

  • Jeong, Min-Heuk;Yu, Yong-Hyun;Park, Sung-Jun;Hwang, Seung-Jun;Baek, Joong-Hwan
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.24 no.3
    • /
    • pp.384-390
    • /
    • 2020
  • The reverberation effect on the sound when producing movies or VR contents is a very important factor in the realism and liveliness. The reverberation time depending the space is recommended in a standard called RT60(Reverberation Time 60 dB). In this paper, we propose a scene recognition technique for automatic reverberation editing. To this end, we devised a classification model that independently trains color images and predicted depth images in the same model. Indoor scene classification is limited only by training color information because of the similarity of internal structure. Deep learning based depth information extraction technology is used to use spatial depth information. Based on RT60, 10 scene classes were constructed and model training and evaluation were conducted. Finally, the proposed SCR + DNet (Scene Classification for Reverb + Depth Net) classifier achieves higher performance than conventional CNN classifiers with 92.4% accuracy.

Ranging Algorithm of Underwater Acoustic Wave with Look-up Table (Look-up table을 이용한 수중 음향파 거리 추정 알고리즘)

  • Cheon, Ju-Hyun;Moon, Seung-Hyun;Lee, Ho-Kyoung
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.52 no.4
    • /
    • pp.23-29
    • /
    • 2015
  • In this paper, we introduce a underwater ranging algorithm with Look-up Table (LUT) by modifying the existing method which is using the changes of angles of accoustic rays with SSP (Sound Speed Profile). We compare the horizontal distance errors and the calculation times. Our new algorithm exploits Time of Arriva l(ToA) - horizontal distance table based on SSP. This algorithm offers faster calculation speed than the previous one with the slight increase of the distance estimation error.

Context-Awareness Cat Behavior Captioning System (반려묘의 상황인지형 행동 캡셔닝 시스템)

  • Chae, Heechan;Choi, Yoona;Lee, Jonguk;Park, Daihee;Chung, Yongwha
    • Journal of Korea Multimedia Society
    • /
    • v.24 no.1
    • /
    • pp.21-29
    • /
    • 2021
  • With the recent increase in the number of households raising pets, various engineering studies have been underway for pets. The final purpose of this study is to automatically generate situation-sensitive captions that can express implicit intentions based on the behavior and sound of cats by embedding the already mature behavioral detection technology of pets as basic element technology in the video capturing research. As a pilot project to this end, this paper proposes a high-level capturing system using optical-flow, RGB, and sound information of cat videos. That is, the proposed system uses video datasets collected in an actual breeding environment to extract feature vectors from the video and sound, then through hierarchical LSTM encoder and decoder, to identify the cat's behavior and its implicit intentions, and to perform learning to create context-sensitive captions. The performance of the proposed system was verified experimentally by utilizing video data collected in the environment where actual cats are raised.

Efficient and Secure Sound-Based Hybrid Authentication Factor with High Usability

  • Mohinder Singh B;Jaisankar N.
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.17 no.10
    • /
    • pp.2844-2861
    • /
    • 2023
  • Internet is the most prevailing word being used nowadays. Over the years, people are becoming more dependent on the internet as it makes their job easier. This became a part of everyone's life as a means of communication in almost every area like financial transactions, education, and personal-health operations. A lot of data is being converted to digital and made online. Many researchers have proposed different authentication factors - biometric and/or non-biometric authentication factors - as the first line of defense to secure online data. Among all those factors, passwords and passphrases are being used by many users around the world. However, the usability of these factors is low. Also, the passwords are easily susceptible to brute force and dictionary attacks. This paper proposes the generation of a novel passcode from the hybrid authentication factor - sound. The proposed passcode is evaluated for its strength to resist brute-force and dictionary attacks using the Shannon entropy and Passcode (or password) entropy formulae. Also, the passcode is evaluated for its usability. The entropy value of the proposed is 658.2. This is higher than that of other authentication factors. Like, for a 6-digit pin - the entropy value was 13.2, 101.4 for Password with Passphrase combined with Keystroke dynamics and 193 for fingerprint, and 30 for voice biometrics. The proposed novel passcode is far much better than other authentication factors when compared with their corresponding strength and usability values.

Software Architecture of Contents-based Control for Co-operative Remote Manipulation of Multi-Robots

  • Thuy, Dinh Trong;Kang, Soon-Ju
    • Annual Conference of KIPS
    • /
    • 2008.05a
    • /
    • pp.508-511
    • /
    • 2008
  • In this paper, we propose software architecture for conveying contents-based OpenSound Control (OSC) packet from manipulation user interface to cooperative remote multi-robots. The Flash application is used as a controlling user interface and the physical prototyping of multi-robots were developed using physical prototyping toolkit.

Ensuring Sound Numerical Simulation of Hybrid Automata

  • Hur, Yerang;Sim, Jae-Hwan;Kim, Je-Sung;Chai, Jin-Young
    • Journal of Computing Science and Engineering
    • /
    • v.3 no.2
    • /
    • pp.73-87
    • /
    • 2009
  • A hybrid system is a dynamical system in which states can be changed continuously and discretely. Simulation based on numerical methods is the widely used technique for analyzing complicated hybrid systems. Numerical simulation of hybrid systems, however, is subject to two types of numerical errors: truncation error and round-off error. The effect of such errors can make an impossible transition step to become possible during simulation, and thus, to generate a simulation behavior that is not allowed by the model. The possibility of an incorrect simulation behavior reduces con.dence in simulation-based analysis since it is impossible to know whether a particular simulation trace is allowed by the model or not. To address this problem, we define the notion of Instrumented Hybrid Automata (IHA), which considers the effect of accumulated numerical errors on discrete transition steps. We then show how to convert Hybrid Automata (HA) to IRA and prove that every simulation behavior of IHA preserves the discrete transition steps of some behavior in HA; that is, simulation of IHA is sound with respect to HA.