• Title/Summary/Keyword: large-language model

Search Result 302, Processing Time 0.031 seconds

A study on performance improvement considering the balance between corpus in Neural Machine Translation (인공신경망 기계번역에서 말뭉치 간의 균형성을 고려한 성능 향상 연구)

  • Park, Chanjun;Park, Kinam;Moon, Hyeonseok;Eo, Sugyeong;Lim, Heuiseok
    • Journal of the Korea Convergence Society
    • /
    • v.12 no.5
    • /
    • pp.23-29
    • /
    • 2021
  • Recent deep learning-based natural language processing studies are conducting research to improve performance by training large amounts of data from various sources together. However, there is a possibility that the methodology of learning by combining data from various sources into one may prevent performance improvement. In the case of machine translation, data deviation occurs due to differences in translation(liberal, literal), style(colloquial, written, formal, etc.), domains, etc. Combining these corpora into one for learning can adversely affect performance. In this paper, we propose a new Corpus Weight Balance(CWB) method that considers the balance between parallel corpora in machine translation. As a result of the experiment, the model trained with balanced corpus showed better performance than the existing model. In addition, we propose an additional corpus construction process that enables coexistence with the human translation market, which can build high-quality parallel corpus even with a monolingual corpus.

Emergency dispatching based on automatic speech recognition (음성인식 기반 응급상황관제)

  • Lee, Kyuwhan;Chung, Jio;Shin, Daejin;Chung, Minhwa;Kang, Kyunghee;Jang, Yunhee;Jang, Kyungho
    • Phonetics and Speech Sciences
    • /
    • v.8 no.2
    • /
    • pp.31-39
    • /
    • 2016
  • In emergency dispatching at 119 Command & Dispatch Center, some inconsistencies between the 'standard emergency aid system' and 'dispatch protocol,' which are both mandatory to follow, cause inefficiency in the dispatcher's performance. If an emergency dispatch system uses automatic speech recognition (ASR) to process the dispatcher's protocol speech during the case registration, it instantly extracts and provides the required information specified in the 'standard emergency aid system,' making the rescue command more efficient. For this purpose, we have developed a Korean large vocabulary continuous speech recognition system for 400,000 words to be used for the emergency dispatch system. The 400,000 words include vocabulary from news, SNS, blogs and emergency rescue domains. Acoustic model is constructed by using 1,300 hours of telephone call (8 kHz) speech, whereas language model is constructed by using 13 GB text corpus. From the transcribed corpus of 6,600 real telephone calls, call logs with emergency rescue command class and identified major symptom are extracted in connection with the rescue activity log and National Emergency Department Information System (NEDIS). ASR is applied to emergency dispatcher's repetition utterances about the patient information. Based on the Levenshtein distance between the ASR result and the template information, the emergency patient information is extracted. Experimental results show that 9.15% Word Error Rate of the speech recognition performance and 95.8% of emergency response detection performance are obtained for the emergency dispatch system.

The Parallel Corpus Approach to Building the Syntactic Tree Transfer Set in the English-to- Vietnamese Machine Translation

  • Dien Dinh;Ngan Thuy;Quang Xuan;Nam Chi
    • Proceedings of the IEEK Conference
    • /
    • summer
    • /
    • pp.382-386
    • /
    • 2004
  • Recently, with the machine learning trend, most of the machine translation systems on over the world use two syntax tree sets of two relevant languages to learn syntactic tree transfer rules. However, for the English-Vietnamese language pair, this approach is impossible because until now we have not had a Vietnamese syntactic tree set which is correspondent to English one. Building of a very large correspondent Vietnamese syntactic tree set (thousands of trees) requires so much work and take the investment of specialists in linguistics. To take advantage from our available English-Vietnamese Corpus (EVC) which was tagged in word alignment, we choose the SITG (Stochastic Inversion Transduction Grammar) model to construct English- Vietnamese syntactic tree sets automatically. This model is used to parse two languages at the same time and then carry out the syntactic tree transfer. This English-Vietnamese bilingual syntactic tree set is the basic training data to carry out transferring automatically from English syntactic trees to Vietnamese ones by machine learning models. We tested the syntax analysis by comparing over 10,000 sentences in the amount of 500,000 sentences of our English-Vietnamese bilingual corpus and first stage got encouraging result $(analyzed\;about\;80\%)[5].$ We have made use the TBL algorithm (Transformation Based Learning) to carry out automatic transformations from English syntactic trees to Vietnamese ones based on that parallel syntactic tree transfer set[6].

  • PDF

Design and Implementation of Peer-to-Peer Electronic Commerce Systems based on the File Sharing Method between Users (이용자간 파일공유방식에 기반한 P2P 전자상거래 시스템 설계 및 구현)

  • Kim Chang-Su;Seo Young-Suk
    • The Journal of Information Systems
    • /
    • v.15 no.1
    • /
    • pp.1-20
    • /
    • 2006
  • Peer-to-peer systems (P2P) are rapidly growing in importance on the Internet environment, quickly extending the range of their usage. However, peer-to-peer systems have not been widely applied in electronic commerce because they have not been established as an appropriate business model. Therefore, we firstly review the previous research relevant to peer-to-peer systems, and then analyze the business models for P2P systems presented by previous researchers. Furthermore, this study categorizes major issues in terms of the technical and business model aspects. On the basis of these reviews, we develop P2P electronic commerce systems based on the file sharing method between users, focusing on user interface friendliness. A developed P2P electronic commerce systems are programmed by using the C# based on the Microsoft.net solution. A database is implemented using the MSSQL2000. A main application technology is designed that P2P electronic commerce systems make it possible. for user to extend into BtoB Solution by using WSDL (Web Services Description Language), UDDI (Universal Description, Discovery, and Integration) and the XML that is a document for users. User interface is made as form of Internet messenger for a user's convenience and is possible to develop into a commodity transaction system based on XML. In this study, it is possible for the P2P electronic commerce system to have extended application to fields such as Internet shopping mall and property transaction in a nonprofit organization, a public institution and a large scale nonprofit institution that have a similar structure as compared with a structure of a nonprofit educational institution.

  • PDF

Preliminary design of control software for SDSS-V Local Volume Mapper Instrument

  • Kim, Changgon;Ji, Tae-geun;Ahn, Hojae;Yang, Mingyeong;Lee, Sumin;Kim, Taeeun;Pak, Soojong;Konidaris, Nicholas P.;Drory, Niv;Froning, Cynthia S.;Hebert, Anthony;Bilgi, Pavan;Blanc, Guillermo A.;Lanz, Alicia E.;Hull, Charles L;Kollmeier, Juna A.;Ramirez, Solange;Wachter, Stefanie;Kreckel, Kathryn;Pellegrini, Eric;Almeida, Andr'es;Case, Scott;Zhelem, Ross;Feger, Tobias;Lawrence, Jon;Lesser, Michael;Herbst, Tom;Sanchez-Gallego, Jose;Bershady, Matthew A;Chattopadhyay, Sabyasachi;Hauser, Andrew;Smith, Michael;Wolf, Marsha J;Yan, Renbin
    • The Bulletin of The Korean Astronomical Society
    • /
    • v.46 no.1
    • /
    • pp.39.1-39.1
    • /
    • 2021
  • The Local Volume Mapper(LVM) project in the fifth iteration of the Sloan Digital Sky Survey (SDSS-V) will produce large integral-field spectroscopic survey data to understand the physical conditions of the interstellar medium in the Milky Way, the Magellanic Clouds, and other local-volume galaxies. We are developing the LVM Instrument control software. The architecture design of the software follows a hierarchical structure in which the high-level software packages interact with the low-level and mid-level software and hardware components. We adopt the spiral software development model in which the software evolves by iteration of sequential processes, i.e., software requirement analysis, design, code generation, and testing. This spiral model ensures that even after being commissioned, the software can be revised according to new operational requirements. We designed the software by using the Unified Modeling Language, which can visualize functional interactions in structure diagrams. We plan to use the SDSS software framework CLU for the interaction between components, based on the RabbitMQ that implemented the Advanced Message Queuing Protocol (AMQP).

  • PDF

Evaluating the Current State of ChatGPT and Its Disruptive Potential: An Empirical Study of Korean Users

  • Jiwoong Choi;Jinsoo Park;Jihae Suh
    • Asia pacific journal of information systems
    • /
    • v.33 no.4
    • /
    • pp.1058-1092
    • /
    • 2023
  • This study investigates the perception and adoption of ChatGPT (a large language model (LLM)-based chatbot created by OpenAI) among Korean users and assesses its potential as the next disruptive innovation. Drawing on previous literature, the study proposes perceived intelligence and perceived anthropomorphism as key differentiating factors of ChatGPT from earlier AI-based chatbots. Four individual motives (i.e., perceived usefulness, ease of use, enjoyment, and trust) and two societal motives (social influence and AI anxiety) were identified as antecedents of ChatGPT acceptance. A survey was conducted within two Korean online communities related to artificial intelligence, the findings of which confirm that ChatGPT is being used for both utilitarian and hedonic purposes, and that perceived usefulness and enjoyment positively impact the behavioral intention to adopt the chatbot. However, unlike prior expectations, perceived ease-of-use was not shown to exert significant influence on behavioral intention. Moreover, trust was not found to be a significant influencer to behavioral intention, and while social influence played a substantial role in adoption intention and perceived usefulness, AI anxiety did not show a significant effect. The study confirmed that perceived intelligence and perceived anthropomorphism are constructs that influence the individual factors that influence behavioral intention to adopt and highlights the need for future research to deconstruct and explore the factors that make ChatGPT "enjoyable" and "easy to use" and to better understand its potential as a disruptive technology. Service developers and LLM providers are advised to design user-centric applications, focus on user-friendliness, acknowledge that building trust takes time, and recognize the role of social influence in adoption.

ChatGPT-based Software Requirements Engineering (ChatGPT 기반 소프트웨어 요구공학)

  • Jongmyung Choi
    • Journal of Internet of Things and Convergence
    • /
    • v.9 no.6
    • /
    • pp.45-50
    • /
    • 2023
  • In software development, the elicitation and analysis of requirements is a crucial phase, and it involves considerable time and effort due to the involvement of various stakeholders. ChatGPT, having been trained on a diverse array of documents, is a large language model that possesses not only the ability to generate code and perform debugging but also the capability to be utilized in the domain of software analysis and design. This paper proposes a method of requirements engineering that leverages ChatGPT's capabilities for eliciting software requirements, analyzing them to align with system goals, and documenting them in the form of use cases. In software requirements engineering, it suggests that stakeholders, analysts, and ChatGPT should engage in a collaborative model. The process should involve using the outputs of ChatGPT as initial requirements, which are then reviewed and augmented by analysts and stakeholders. As ChatGPT's capability improves, it is anticipated that the accuracy of requirements elicitation and analysis will increase, leading to time and cost savings in the field of software requirements engineering.

Fine-tuning Method to Improve Sentiment Classification Perfoimance of Review Data (리뷰 데이터 감성 분류 성능 향상을 위한 Fine-tuning 방법)

  • Jung II Park;Myimg Jin Lim;Pan Koo Kim
    • Smart Media Journal
    • /
    • v.13 no.6
    • /
    • pp.44-53
    • /
    • 2024
  • Companies in modern society are increasingly recognizing sentiment classification as a crucial task, emphasizing the importance of accurately understanding consumer opinions opinions across various platforms such as social media, product reviews, and customer feedback for competitive success. Extensive research is being conducted on sentiment classification as it helps improve products or services by identifying the diverse opinions and emotions of consumers. In sentiment classification, fine-tuning with large-scale datasets and pre-trained language models is essential for enhancing performance. Recent advancements in artificial intelligence have led to high-performing sentiment classification models, with the ELECTRA model standing out due to its efficient learning methods and minimal computing resource requirements. Therefore, this paper proposes a method to enhance sentiment classification performance through efficient fine-tuning of various datasets using the KoELECTRA model, specifically trained for Korean.

A Design and Implementation of The Deep Learning-Based Senior Care Service Application Using AI Speaker

  • Mun Seop Yun;Sang Hyuk Yoon;Ki Won Lee;Se Hoon Kim;Min Woo Lee;Ho-Young Kwak;Won Joo Lee
    • Journal of the Korea Society of Computer and Information
    • /
    • v.29 no.4
    • /
    • pp.23-30
    • /
    • 2024
  • In this paper, we propose a deep learning-based personalized senior care service application. The proposed application uses Speech to Text technology to convert the user's speech into text and uses it as input to Autogen, an interactive multi-agent large-scale language model developed by Microsoft, for user convenience. Autogen uses data from previous conversations between the senior and ChatBot to understand the other user's intent and respond to the response, and then uses a back-end agent to create a wish list, a shared calendar, and a greeting message with the other user's voice through a deep learning model for voice cloning. Additionally, the application can perform home IoT services with SKT's AI speaker (NUGU). The proposed application is expected to contribute to future AI-based senior care technology.

A Study of Automatic Deep Learning Data Generation by Considering Private Information Protection (개인정보 보호를 고려한 딥러닝 데이터 자동 생성 방안 연구)

  • Sung-Bong Jang
    • The Journal of the Convergence on Culture Technology
    • /
    • v.10 no.1
    • /
    • pp.435-441
    • /
    • 2024
  • In order for the large amount of collected data sets to be used as deep learning training data, sensitive personal information such as resident registration number and disease information must be changed or encrypted to prevent it from being exposed to hackers, and the data must be reconstructed to match the structure of the built deep learning model. Currently, these tasks are performed manually by experts, which takes a lot of time and money. To solve these problems, this paper proposes a technique that can automatically perform data processing tasks to protect personal information during the deep learning process. In the proposed technique, privacy protection tasks are performed based on data generalization and data reconstruction tasks are performed using circular queues. To verify the validity of the proposed technique, it was directly implemented using C language. As a result of the verification, it was confirmed that data generalization was performed normally and data reconstruction suitable for the deep learning model was performed properly.