• Title/Summary/Keyword: multimodal system

Search Result 227, Processing Time 0.025 seconds

Speaker Identification Using an Ensemble of Feature Enhancement Methods (특징 강화 방법의 앙상블을 이용한 화자 식별)

  • Yang, IL-Ho;Kim, Min-Seok;So, Byung-Min;Kim, Myung-Jae;Yu, Ha-Jin
    • Phonetics and Speech Sciences
    • /
    • v.3 no.2
    • /
    • pp.71-78
    • /
    • 2011
  • In this paper, we propose an approach which constructs classifier ensembles of various channel compensation and feature enhancement methods. CMN and CMVN are used as channel compensation methods. PCA, kernel PCA, greedy kernel PCA, and kernel multimodal discriminant analysis are used as feature enhancement methods. The proposed ensemble system is constructed with the combination of 15 classifiers which include three channel compensation methods (including 'without compensation') and five feature enhancement methods (including 'without enhancement'). Experimental results show that the proposed ensemble system gives highest average speaker identification rate in various environments (channels, noises, and sessions).

  • PDF

Data model of Multimodal Visual Interface (멀티모달 비주얼 인터페이스의 테이터형)

  • Malyanov, Ilya;d'Auriol, Brian J.;Lee, Sung-Young;Lee, Young-Koo
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2011.06b
    • /
    • pp.240-241
    • /
    • 2011
  • Contemporary electronic healthcare systems are getting more and more complex, providing users a broad functionality, but often fail to have accessible interfaces. However, the importance of a good interface is nearly as great as of the rest of the system. Development of an intuitive multimodal interface for a healthcare system is the goal of our research work. This paper discusses data model of the interface.

Multimodal Parametric Fusion for Emotion Recognition

  • Kim, Jonghwa
    • International journal of advanced smart convergence
    • /
    • v.9 no.1
    • /
    • pp.193-201
    • /
    • 2020
  • The main objective of this study is to investigate the impact of additional modalities on the performance of emotion recognition using speech, facial expression and physiological measurements. In order to compare different approaches, we designed a feature-based recognition system as a benchmark which carries out linear supervised classification followed by the leave-one-out cross-validation. For the classification of four emotions, it turned out that bimodal fusion in our experiment improves recognition accuracy of unimodal approach, while the performance of trimodal fusion varies strongly depending on the individual. Furthermore, we experienced extremely high disparity between single class recognition rates, while we could not observe a best performing single modality in our experiment. Based on these observations, we developed a novel fusion method, called parametric decision fusion (PDF), which lies in building emotion-specific classifiers and exploits advantage of a parametrized decision process. By using the PDF scheme we achieved 16% improvement in accuracy of subject-dependent recognition and 10% for subject-independent recognition compared to the best unimodal results.

A multimodal adaptive evolution of the N1 method for assessment and design of r.c. framed structures

  • Lenza, Pietro;Ghersi, Aurelio;Marino, Edoardo M.;Pellecchia, Marcello
    • Earthquakes and Structures
    • /
    • v.12 no.3
    • /
    • pp.271-284
    • /
    • 2017
  • This paper presents a multimodal adaptive nonlinear static method of analysis that, differently from the nonlinear static methods suggested in seismic codes, does not require the definition of the equivalent Single-Degree-Of-Freedom (SDOF) system to evaluate the seismic response of structures. First, the proposed method is formulated for the assessment of r.c. plane frames and then it is extended to 3D framed structures. Furthermore, the proposed nonlinear static approach is re-elaborated as a displacement-based design method that does not require the use of the behaviour factor and takes into account explicitly the plastic deformation capacity of the structure. Numerical applications to r.c. plane frames and to a 3D framed structure with inplan irregularity are carried out to illustrate the attractive features as well as the limitations of the proposed method. Furthermore, the numerical applications evidence the uncertainty about the suitability of the displacement demand prediction obtained by the nonlinear static methods commonly adopted.

An Watermarking Algorithm for Multimodal Biometric Systems (다중 생체인식 시스템에 적합한 워터마킹 알고리즘)

  • Moon, Dae-Sung;Jung, Seung-Hwan;Kim, Tae-Hae;Chung, Yong-Wha;Moon, Ki-Young
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.15 no.4
    • /
    • pp.93-100
    • /
    • 2005
  • In this paper, we describe biometric watermarking techniques for secure user verification on the remote, multimodal biometric system employing both fingerprint and face information, and compare their effects on verification accuracy quantitatively. To hide biometric data with watermarking techniques, we first consider possible two scenarios. In the scenario 1, we use a fingerprint image as a cover work and hide facial features into it. On the contrary, we hide fingerprint features into a facial image in the Scenario 2. Based on the experimental results, we confirm that the Scenario 2 is superior to the Scenario 1 in terms of the verification accuracy of the watermarking image.

Analysis on Barriers and Resolution Priority of Sea-Rail Multimodal Logistics among Korea and Eurasia Nations (한국-유라시아간 해륙복합운송 문제점 및 해결 우선순위 분석)

  • Lee, Eon-Kyung;Lee, Suyoung;Kim, Bokyung;Euh, Seungseob
    • Journal of Korea Port Economic Association
    • /
    • v.35 no.2
    • /
    • pp.109-126
    • /
    • 2019
  • The Panmunjom Declaration adopted by the leaders of South and North Korea on April 27, 2018, has created an environment conducive for peace and cooperation in the Korean Peninsula. In the June of last year, South Korea has joined the Organization for Cooperation between Railways (OSJD). The membership of OSJD has established a solid foundation for restoring a multimodal logistics system that connects the Korean peninsula to Eurasia countries, including China and Russia. In this paper, a questionnaire survey targeting working-level experts was conducted to find the barriers in constructing multimodal logistics that efficiently connect the port-continental railways of the Korean peninsula and the Eurasian nations. Survey items were divided into five categories-border crossing procedures, technology, facilities, operation, and government support. As a result, among the most important problems of international multimodal logistics in Eurasia that need to be solved on priority include improving transshipment facilities, eliminating inspection carried out at every country for transit, simplifying documents for customs clearance, and minimizing the changes in freight rates. In conclusion, for vitalizing the connection between the Korean peninsula and the continental railways, it is necessary to develop a transshipment system to facilitate the changes in tracks at the borders by making a joint effort with the international community. Second, railway and operational systems in South Korea, North Korea, China, and Russia should be standardized. Third, international cooperation among South Korea, North Korea, China, and Russia is essential for simplifying customs clearance at borders, priority departure of domestic cargo, sharing information about the changes in freight rates, and so on. Finally, the government should come up with measures to secure the quantity of cargo required to form block trains, while developing new business models.

Impact Analysis of nonverbal multimodals for recognition of emotion expressed virtual humans (가상 인간의 감정 표현 인식을 위한 비언어적 다중모달 영향 분석)

  • Kim, Jin Ok
    • Journal of Internet Computing and Services
    • /
    • v.13 no.5
    • /
    • pp.9-19
    • /
    • 2012
  • Virtual human used as HCI in digital contents expresses his various emotions across modalities like facial expression and body posture. However, few studies considered combinations of such nonverbal multimodal in emotion perception. Computational engine models have to consider how a combination of nonverbal modal like facial expression and body posture will be perceived by users to implement emotional virtual human, This paper proposes the impacts of nonverbal multimodal in design of emotion expressed virtual human. First, the relative impacts are analysed between different modals by exploring emotion recognition of modalities for virtual human. Then, experiment evaluates the contribution of the facial and postural congruent expressions to recognize basic emotion categories, as well as the valence and activation dimensions. Measurements are carried out to the impact of incongruent expressions of multimodal on the recognition of superposed emotions which are known to be frequent in everyday life. Experimental results show that the congruence of facial and postural expression of virtual human facilitates perception of emotion categories and categorical recognition is influenced by the facial expression modality, furthermore, postural modality are preferred to establish a judgement about level of activation dimension. These results will be used to implementation of animation engine system and behavior syncronization for emotion expressed virtual human.

Nano Bio Imaging for NT and BT

  • Moon, DaeWon
    • Proceedings of the Korean Vacuum Society Conference
    • /
    • 2015.08a
    • /
    • pp.51.2-51.2
    • /
    • 2015
  • Understanding interfacial phenomena has been one of the main research issues not only in semiconductors but only in life sciences. I have been trying to meet the atomic scale surface and interface analysis challenges from semiconductor industries and furthermore to extend the application scope to biomedical areas. Optical imaing has been most widely and successfully used for biomedical imaging but complementary ion beam imaging techniques based on mass spectrometry and ion scattering can provide more detailed molecular specific and nanoscale information In this presentation, I will review the 27 years history of medium energy ion scattering (MEIS) development at KRISS and DGIST for nanoanalysis. A electrostatic MEIS system constructed at KRISS after the FOM, Netherland design had been successfully applied for the gate oxide analysis and quantitative surface analysis. Recenlty, we developed time-of-flight (TOF) MEIS system, for the first time in the world. With TOF-MEIS, we reported quantitative compositional profiling with single atomic layer resolution for 0.5~3 nm CdSe/ZnS conjugated QDs and ultra shallow junctions and FINFET's of As implanted Si. With this new TOF-MEIS nano analysis technique, details of nano-structured materials could be measured quantitatively. Progresses in TOF-MEIS analysis in various nano & bio technology will be discussed. For last 10 years, I have been trying to develop multimodal nanobio imaging techniques for cardiovascular and brain tissues. Firstly, in atherosclerotic plaque imaging, using, coherent anti-stokes raman scattering (CARS) and time-of-flight secondary ion mass spectrometry (TOF-SIMS) multimodal analysis showed that increased cholesterol palmitate may contribute to the formation of a necrotic core by increasing cell death. Secondly, surface plasmon resonance imaging ellipsometry (SPRIE) was developed for cell biointerface imaging of cell adhesion, migration, and infiltration dynamics for HUVEC, CASMC, and T cells. Thirdly, we developed an ambient mass spectrometric imaging system for live cells and tissues. Preliminary results on mouse brain hippocampus and hypotahlamus will be presented. In conclusions, multimodal optical and mass spectrometric imaging privides overall structural and morphological information with complementary molecular specific information, which can be a useful methodology for biomedical studies. Future challenges in optical and mass spectrometric imaging for new biomedical applications will be discussed.

  • PDF

Multimedia Information and Authoring for Personalized Media Networks

  • Choi, Insook;Bargar, Robin
    • Journal of Multimedia Information System
    • /
    • v.4 no.3
    • /
    • pp.123-144
    • /
    • 2017
  • Personalized media includes user-targeted and user-generated content (UGC) exchanged through social media and interactive applications. The increased consumption of UGC presents challenges and opportunities to multimedia information systems. We work towards modeling a deep structure for content networks. To gain insights, a hybrid practice with Media Framework (MF) is presented for network creation of personalized media, which leverages the authoring methodology with user-generated semantics. The system's vertical integration allows users to audition their personalized media networks in the context of a global system network. A navigation scheme with dynamic GUI shifts the interaction paradigm for content query and sharing. MF adopts a multimodal architecture anticipating emerging use cases and genres. To model diversification of platforms, information processing is robust across multiple technology configurations. Physical and virtual networks are integrated with distributed services and transactions, IoT, and semantic networks representing media content. MF applies spatiotemporal and semantic signal processing to differentiate action responsiveness and information responsiveness. The extension of multimedia information processing into authoring enables generating interactive and impermanent media on computationally enabled devices. The outcome of this integrated approach with presented methodologies demonstrates a paradigmatic shift of the concept of UGC as personalized media network, which is dynamical and evolvable.