• Title/Summary/Keyword: conversational AI

Search Result 46, Processing Time 0.022 seconds

The Effect of Interjection in Conversational Interaction with the AI Agent: In the Context of Self-Driving Car (인공지능 에이전트 대화형 인터랙션에서의 감탄사 효과: 자율주행 맥락에서)

  • Lee, Sooji;Seo, Jeeyoon;Choi, Junho
    • The Journal of the Convergence on Culture Technology
    • /
    • v.8 no.1
    • /
    • pp.551-563
    • /
    • 2022
  • This study aims to identify the effect on the user experiences when the embodied agent in a self-driving car interacts with emotional expressions by using 'interjection'. An experimental study was designed with two conditions: the inclusion of injections in the agent's conversation feedbacks (with interjections vs. without interjections) and the type of conversation (task-oriented conversation vs. social-oriented conversation). The online experiment was conducted with the four video clips of conversation scenario treatments and measured intimacy, likability, trust, social presence, perceived anthropomorphism, and future intention to use. The result showed that when the agent used interjection, the main effect on social presence was found in both conversation types. When the agent did not use interjection in the task-oriented conversation, trust and future intention to use were higher than when the agent talked with emotional expressions. In the context of the conversation with the AI agent in a self-driving car, we found only the effect of adding emotional expression by using interjection on the enhancing social presence, but no effect on the other user experience factors.

Developing a New Algorithm for Conversational Agent to Detect Recognition Error and Neologism Meaning: Utilizing Korean Syllable-based Word Similarity (대화형 에이전트 인식오류 및 신조어 탐지를 위한 알고리즘 개발: 한글 음절 분리 기반의 단어 유사도 활용)

  • Jung-Won Lee;Il Im
    • Journal of Intelligence and Information Systems
    • /
    • v.29 no.3
    • /
    • pp.267-286
    • /
    • 2023
  • The conversational agents such as AI speakers utilize voice conversation for human-computer interaction. Voice recognition errors often occur in conversational situations. Recognition errors in user utterance records can be categorized into two types. The first type is misrecognition errors, where the agent fails to recognize the user's speech entirely. The second type is misinterpretation errors, where the user's speech is recognized and services are provided, but the interpretation differs from the user's intention. Among these, misinterpretation errors require separate error detection as they are recorded as successful service interactions. In this study, various text separation methods were applied to detect misinterpretation. For each of these text separation methods, the similarity of consecutive speech pairs using word embedding and document embedding techniques, which convert words and documents into vectors. This approach goes beyond simple word-based similarity calculation to explore a new method for detecting misinterpretation errors. The research method involved utilizing real user utterance records to train and develop a detection model by applying patterns of misinterpretation error causes. The results revealed that the most significant analysis result was obtained through initial consonant extraction for detecting misinterpretation errors caused by the use of unregistered neologisms. Through comparison with other separation methods, different error types could be observed. This study has two main implications. First, for misinterpretation errors that are difficult to detect due to lack of recognition, the study proposed diverse text separation methods and found a novel method that improved performance remarkably. Second, if this is applied to conversational agents or voice recognition services requiring neologism detection, patterns of errors occurring from the voice recognition stage can be specified. The study proposed and verified that even if not categorized as errors, services can be provided according to user-desired results.

Performance Evaluation of Pre-trained Language Models in Multi-Goal Conversational Recommender Systems (다중목표 대화형 추천시스템을 위한 사전 학습된 언어모델들에 대한 성능 평가)

  • Taeho Kim;Hyung-Jun Jang;Sang-Wook Kim
    • Smart Media Journal
    • /
    • v.12 no.6
    • /
    • pp.35-40
    • /
    • 2023
  • In this study paper, we examine pre-trained language models used in Multi-Goal Conversational Recommender Systems (MG-CRS), comparing and analyzing their performances of various pre-trained language models. Specifically, we investigates the impact of the sizes of language models on the performance of MG-CRS. The study targets three types of language models - of BERT, GPT2, and BART, and measures and compares their accuracy in two tasks of 'type prediction' and 'topic prediction' on the MG-CRS dataset, DuRecDial 2.0. Experimental results show that all models demonstrated excellent performance in the type prediction task, but there were notable provide significant performance differences in performance depending on among the models or based on their sizes in the topic prediction task. Based on these findings, the study provides directions for improving the performance of MG-CRS.

Users' Attachment Styles and ChatGPT Interaction: Revealing Insights into User Experiences

  • I-Tsen Hsieh;Chang-Hoon Oh
    • Journal of the Korea Society of Computer and Information
    • /
    • v.29 no.3
    • /
    • pp.21-41
    • /
    • 2024
  • This study explores the relationship between users' attachment styles and their interactions with ChatGPT (Chat Generative Pre-trained Transformer), an advanced language model developed by OpenAI. As artificial intelligence (AI) becomes increasingly integrated into everyday life, it is essential to understand how individuals with different attachment styles engage with AI chatbots in order to build a better user experience that meets specific user needs and interacts with users in the most ideal way. Grounded in attachment theory from psychology, we are exploring the influence of attachment style on users' interaction with ChatGPT, bridging a significant gap in understanding human-AI interaction. Contrary to expectations, attachment styles did not have a significant impact on ChatGPT usage or reasons for engagement. Regardless of their attachment styles, hesitated to fully trust ChatGPT with critical information, emphasizing the need to address trust issues in AI systems. Additionally, this study uncovers complex patterns of attachment styles, demonstrating their influence on interaction patterns between users and ChatGPT. By focusing on the distinctive dynamics between users and ChatGPT, our aim is to uncover how attachment styles influence these interactions, guiding the development of AI chatbots for personalized user experiences. The introduction of the Perceived Partner Responsiveness Scale serves as a valuable tool to evaluate users' perceptions of ChatGPT's role, shedding light on the anthropomorphism of AI. This study contributes to the wider discussion on human-AI relationships, emphasizing the significance of incorporating emotional intelligence into AI systems for a user-centered future.

Research on Developing a Conversational AI Callbot Solution for Medical Counselling

  • Won Ro LEE;Jeong Hyon CHOI;Min Soo KANG
    • Korean Journal of Artificial Intelligence
    • /
    • v.11 no.4
    • /
    • pp.9-13
    • /
    • 2023
  • In this study, we explored the potential of integrating interactive AI callbot technology into the medical consultation domain as part of a broader service development initiative. Aimed at enhancing patient satisfaction, the AI callbot was designed to efficiently address queries from hospitals' primary users, especially the elderly and those using phone services. By incorporating an AI-driven callbot into the hospital's customer service center, routine tasks such as appointment modifications and cancellations were efficiently managed by the AI Callbot Agent. On the other hand, tasks requiring more detailed attention or specialization were addressed by Human Agents, ensuring a balanced and collaborative approach. The deep learning model for voice recognition for this study was based on the Transformer model and fine-tuned to fit the medical field using a pre-trained model. Existing recording files were converted into learning data to perform SSL(self-supervised learning) Model was implemented. The ANN (Artificial neural network) neural network model was used to analyze voice signals and interpret them as text, and after actual application, the intent was enriched through reinforcement learning to continuously improve accuracy. In the case of TTS(Text To Speech), the Transformer model was applied to Text Analysis, Acoustic model, and Vocoder, and Google's Natural Language API was applied to recognize intent. As the research progresses, there are challenges to solve, such as interconnection issues between various EMR providers, problems with doctor's time slots, problems with two or more hospital appointments, and problems with patient use. However, there are specialized problems that are easy to make reservations. Implementation of the callbot service in hospitals appears to be applicable immediately.

Evaluating the Current State of ChatGPT and Its Disruptive Potential: An Empirical Study of Korean Users

  • Jiwoong Choi;Jinsoo Park;Jihae Suh
    • Asia pacific journal of information systems
    • /
    • v.33 no.4
    • /
    • pp.1058-1092
    • /
    • 2023
  • This study investigates the perception and adoption of ChatGPT (a large language model (LLM)-based chatbot created by OpenAI) among Korean users and assesses its potential as the next disruptive innovation. Drawing on previous literature, the study proposes perceived intelligence and perceived anthropomorphism as key differentiating factors of ChatGPT from earlier AI-based chatbots. Four individual motives (i.e., perceived usefulness, ease of use, enjoyment, and trust) and two societal motives (social influence and AI anxiety) were identified as antecedents of ChatGPT acceptance. A survey was conducted within two Korean online communities related to artificial intelligence, the findings of which confirm that ChatGPT is being used for both utilitarian and hedonic purposes, and that perceived usefulness and enjoyment positively impact the behavioral intention to adopt the chatbot. However, unlike prior expectations, perceived ease-of-use was not shown to exert significant influence on behavioral intention. Moreover, trust was not found to be a significant influencer to behavioral intention, and while social influence played a substantial role in adoption intention and perceived usefulness, AI anxiety did not show a significant effect. The study confirmed that perceived intelligence and perceived anthropomorphism are constructs that influence the individual factors that influence behavioral intention to adopt and highlights the need for future research to deconstruct and explore the factors that make ChatGPT "enjoyable" and "easy to use" and to better understand its potential as a disruptive technology. Service developers and LLM providers are advised to design user-centric applications, focus on user-friendliness, acknowledge that building trust takes time, and recognize the role of social influence in adoption.

A Study on College Students' Perceptions of ChatGPT (ChatGPT에 대한 대학생의 인식에 관한 연구)

  • Rhee, Jung-uk;Kim, Hee Ra;Shin, Hye Won
    • Journal of Korean Home Economics Education Association
    • /
    • v.35 no.4
    • /
    • pp.1-12
    • /
    • 2023
  • At a time when interest in the educational use of ChatGPT is increasing, it is necessary to investigate the perception of ChatGPT among college students. A survey was conducted to compare the current status of internet and interactive artificial intelligence use and perceptions of ChatGPT after using it in the following courses in Spring 2023; 'Family Life and Culture', 'Fashion and Museums', and 'Fashion in Movies' in the first semester of 2023. We also looked at comparative analysis reports and reflection diaries. Information for coursework was mainly obtained through internet searches and articles, but only 9.84% used interactive AI, showing that its application to learning is still insufficient. ChatGPT was first used in the Spring semester of 2023, and ChatGPT was mainly used among conversational AI. ChatGPT is a bit lacking in terms of information accuracy and reliability, but it is convenient because it allows students to find information while interacting easily and quickly, and the satisfaction level was high, so there was a willingness to use ChatGPT more actively in the future. Regarding the impact of ChatGPT on education, students said that it was positive that they were self-directed and that they set up a cooperative class process to verify information through group discussions and problem-solving attitudes through questions. However, problems were recognized that lowered trust, such as plagiarism, copyright, data bias, lack of up-to-date data learning, and generation of inaccurate or incorrect information, which need to be improved.

Generative Interactive Psychotherapy Expert (GIPE) Bot

  • Ayesheh Ahrari Khalaf;Aisha Hassan Abdalla Hashim;Akeem Olowolayemo;Rashidah Funke Olanrewaju
    • International Journal of Computer Science & Network Security
    • /
    • v.23 no.4
    • /
    • pp.15-24
    • /
    • 2023
  • One of the objectives and aspirations of scientists and engineers ever since the development of computers has been to interact naturally with machines. Hence features of artificial intelligence (AI) like natural language processing and natural language generation were developed. The field of AI that is thought to be expanding the fastest is interactive conversational systems. Numerous businesses have created various Virtual Personal Assistants (VPAs) using these technologies, including Apple's Siri, Amazon's Alexa, and Google Assistant, among others. Even though many chatbots have been introduced through the years to diagnose or treat psychological disorders, we are yet to have a user-friendly chatbot available. A smart generative cognitive behavioral therapy with spoken dialogue systems support was then developed using a model Persona Perception (P2) bot with Generative Pre-trained Transformer-2 (GPT-2). The model was then implemented using modern technologies in VPAs like voice recognition, Natural Language Understanding (NLU), and text-to-speech. This system is a magnificent device to help with voice-based systems because it can have therapeutic discussions with the users utilizing text and vocal interactive user experience.

Persona-based Korean Conversational Model (페르소나 기반 한국어 대화 모델)

  • Jang, Yoonna;Lim, Jungwoo;Hur, Yuna;Yang, Kisu;Park, Chanjun;Seo, Jaehyung;Lee, Seungjun;Lim, Heuiseok
    • Annual Conference on Human and Language Technology
    • /
    • 2021.10a
    • /
    • pp.453-456
    • /
    • 2021
  • 대화형 에이전트가 일관성 없는 답변, 재미 없는 답변을 하는 문제를 해결하기 위하여 최근 페르소나 기반의 대화 분야의 연구가 활발히 진행되고 있다. 그러나 한국어로 구축된 페르소나 대화 데이터는 아직 구축되지 않은 상황이다. 이에 본 연구에서는 영어 원본 데이터에서 한국어로 번역된 데이터를 활용하여 최초의 페르소나 기반 한국어 대화 모델을 제안한다. 전처리를 통하여 번역 품질을 향상시킨 데이터에 사전 학습 된 한국어 모델인 KoBERT와 KoELECTRA를 미세조정(fine-tuning) 시킴으로써 모델에게 주어진 페르소나와 대화 맥락을 고려하여 올바른 답변을 선택하는 모델을 학습한다. 실험 결과 KoELECTRA-base 모델이 가장 높은 성능을 보이는 것을 확인하였으며, 단순하게 사용자의 발화만을 주는 것 보다 이전 대화 이력이 추가적으로 주어졌을 때 더 좋은 성능을 보이는 것을 확인할 수 있었다.

  • PDF

Action-Based Audit with Relational Rules to Avatar Interactions for Metaverse Ethics

  • Bang, Junseong;Ahn, Sunghee
    • Smart Media Journal
    • /
    • v.11 no.6
    • /
    • pp.51-63
    • /
    • 2022
  • Metaverse provides a simulated environment where a large number of users can participate in various activities. In order for Metaverse to be sustainable, it is necessary to study ethics that can be applied to a Metaverse service platform. In this paper, Metaverse ethics and the rules for applying to the platform are explored. And, in order to judge the ethicality of avatar actions in social Metaverse, the identity, interaction, and relationship of an avatar are investigated. Then, an action-based audit approach to avatar interactions (e.g., dialogues, gestures, facial expressions) is introduced in two cases that an avatar enters a digital world and that an avatar requests the auditing to subjects, e.g., avatars controlled by human users, artificial intelligence (AI) avatars (e.g., as conversational bots), and virtual objects. Pseudocodes for performing the two cases in a system are presented and they are examined based on the description of the avatars' actions.